| .. | .. | 
|---|
| 522 | 522 |  { | 
|---|
| 523 | 523 |  	struct skcipher_request *req; | 
|---|
| 524 | 524 |  	struct scatterlist *dst; | 
|---|
| 525 |  | -	dma_addr_t addr;  | 
|---|
| 526 | 525 |   | 
|---|
| 527 | 526 |  	req = skcipher_request_cast(pd_uinfo->async_req); | 
|---|
| 528 | 527 |   | 
|---|
| .. | .. | 
|---|
| 531 | 530 |  					  req->cryptlen, req->dst); | 
|---|
| 532 | 531 |  	} else { | 
|---|
| 533 | 532 |  		dst = pd_uinfo->dest_va; | 
|---|
| 534 |  | -		addr = dma_map_page(dev->core_dev->device, sg_page(dst),  | 
|---|
| 535 |  | -				    dst->offset, dst->length, DMA_FROM_DEVICE);  | 
|---|
 | 533 | +		dma_unmap_page(dev->core_dev->device, pd->dest, dst->length,  | 
|---|
 | 534 | +			       DMA_FROM_DEVICE);  | 
|---|
| 536 | 535 |  	} | 
|---|
| 537 | 536 |   | 
|---|
| 538 | 537 |  	if (pd_uinfo->sa_va->sa_command_0.bf.save_iv == SA_SAVE_IV) { | 
|---|
| .. | .. | 
|---|
| 557 | 556 |  	struct ahash_request *ahash_req; | 
|---|
| 558 | 557 |   | 
|---|
| 559 | 558 |  	ahash_req = ahash_request_cast(pd_uinfo->async_req); | 
|---|
| 560 |  | -	ctx  = crypto_tfm_ctx(ahash_req->base.tfm);  | 
|---|
 | 559 | +	ctx = crypto_ahash_ctx(crypto_ahash_reqtfm(ahash_req));  | 
|---|
| 561 | 560 |   | 
|---|
| 562 |  | -	crypto4xx_copy_digest_to_dst(ahash_req->result, pd_uinfo,  | 
|---|
| 563 |  | -				     crypto_tfm_ctx(ahash_req->base.tfm));  | 
|---|
 | 561 | +	crypto4xx_copy_digest_to_dst(ahash_req->result, pd_uinfo, ctx);  | 
|---|
| 564 | 562 |  	crypto4xx_ret_sg_desc(dev, pd_uinfo); | 
|---|
| 565 | 563 |   | 
|---|
| 566 | 564 |  	if (pd_uinfo->state & PD_ENTRY_BUSY) | 
|---|