| .. | .. |
|---|
| 522 | 522 | { |
|---|
| 523 | 523 | struct skcipher_request *req; |
|---|
| 524 | 524 | struct scatterlist *dst; |
|---|
| 525 | | - dma_addr_t addr; |
|---|
| 526 | 525 | |
|---|
| 527 | 526 | req = skcipher_request_cast(pd_uinfo->async_req); |
|---|
| 528 | 527 | |
|---|
| .. | .. |
|---|
| 531 | 530 | req->cryptlen, req->dst); |
|---|
| 532 | 531 | } else { |
|---|
| 533 | 532 | dst = pd_uinfo->dest_va; |
|---|
| 534 | | - addr = dma_map_page(dev->core_dev->device, sg_page(dst), |
|---|
| 535 | | - dst->offset, dst->length, DMA_FROM_DEVICE); |
|---|
| 533 | + dma_unmap_page(dev->core_dev->device, pd->dest, dst->length, |
|---|
| 534 | + DMA_FROM_DEVICE); |
|---|
| 536 | 535 | } |
|---|
| 537 | 536 | |
|---|
| 538 | 537 | if (pd_uinfo->sa_va->sa_command_0.bf.save_iv == SA_SAVE_IV) { |
|---|
| .. | .. |
|---|
| 557 | 556 | struct ahash_request *ahash_req; |
|---|
| 558 | 557 | |
|---|
| 559 | 558 | ahash_req = ahash_request_cast(pd_uinfo->async_req); |
|---|
| 560 | | - ctx = crypto_tfm_ctx(ahash_req->base.tfm); |
|---|
| 559 | + ctx = crypto_ahash_ctx(crypto_ahash_reqtfm(ahash_req)); |
|---|
| 561 | 560 | |
|---|
| 562 | | - crypto4xx_copy_digest_to_dst(ahash_req->result, pd_uinfo, |
|---|
| 563 | | - crypto_tfm_ctx(ahash_req->base.tfm)); |
|---|
| 561 | + crypto4xx_copy_digest_to_dst(ahash_req->result, pd_uinfo, ctx); |
|---|
| 564 | 562 | crypto4xx_ret_sg_desc(dev, pd_uinfo); |
|---|
| 565 | 563 | |
|---|
| 566 | 564 | if (pd_uinfo->state & PD_ENTRY_BUSY) |
|---|