.. | .. |
---|
522 | 522 | { |
---|
523 | 523 | struct skcipher_request *req; |
---|
524 | 524 | struct scatterlist *dst; |
---|
525 | | - dma_addr_t addr; |
---|
526 | 525 | |
---|
527 | 526 | req = skcipher_request_cast(pd_uinfo->async_req); |
---|
528 | 527 | |
---|
.. | .. |
---|
531 | 530 | req->cryptlen, req->dst); |
---|
532 | 531 | } else { |
---|
533 | 532 | dst = pd_uinfo->dest_va; |
---|
534 | | - addr = dma_map_page(dev->core_dev->device, sg_page(dst), |
---|
535 | | - dst->offset, dst->length, DMA_FROM_DEVICE); |
---|
| 533 | + dma_unmap_page(dev->core_dev->device, pd->dest, dst->length, |
---|
| 534 | + DMA_FROM_DEVICE); |
---|
536 | 535 | } |
---|
537 | 536 | |
---|
538 | 537 | if (pd_uinfo->sa_va->sa_command_0.bf.save_iv == SA_SAVE_IV) { |
---|
.. | .. |
---|
557 | 556 | struct ahash_request *ahash_req; |
---|
558 | 557 | |
---|
559 | 558 | ahash_req = ahash_request_cast(pd_uinfo->async_req); |
---|
560 | | - ctx = crypto_tfm_ctx(ahash_req->base.tfm); |
---|
| 559 | + ctx = crypto_ahash_ctx(crypto_ahash_reqtfm(ahash_req)); |
---|
561 | 560 | |
---|
562 | | - crypto4xx_copy_digest_to_dst(ahash_req->result, pd_uinfo, |
---|
563 | | - crypto_tfm_ctx(ahash_req->base.tfm)); |
---|
| 561 | + crypto4xx_copy_digest_to_dst(ahash_req->result, pd_uinfo, ctx); |
---|
564 | 562 | crypto4xx_ret_sg_desc(dev, pd_uinfo); |
---|
565 | 563 | |
---|
566 | 564 | if (pd_uinfo->state & PD_ENTRY_BUSY) |
---|