| .. | .. |
|---|
| 102 | 102 | struct snic_req_info *rqi = NULL; |
|---|
| 103 | 103 | unsigned long flags; |
|---|
| 104 | 104 | |
|---|
| 105 | | - pci_unmap_single(snic->pdev, buf->dma_addr, buf->len, PCI_DMA_TODEVICE); |
|---|
| 105 | + dma_unmap_single(&snic->pdev->dev, buf->dma_addr, buf->len, |
|---|
| 106 | + DMA_TO_DEVICE); |
|---|
| 106 | 107 | |
|---|
| 107 | 108 | rqi = req_to_rqi(req); |
|---|
| 108 | 109 | spin_lock_irqsave(&snic->spl_cmd_lock, flags); |
|---|
| .. | .. |
|---|
| 172 | 173 | snic_print_desc(__func__, os_buf, len); |
|---|
| 173 | 174 | |
|---|
| 174 | 175 | /* Map request buffer */ |
|---|
| 175 | | - pa = pci_map_single(snic->pdev, os_buf, len, PCI_DMA_TODEVICE); |
|---|
| 176 | | - if (pci_dma_mapping_error(snic->pdev, pa)) { |
|---|
| 176 | + pa = dma_map_single(&snic->pdev->dev, os_buf, len, DMA_TO_DEVICE); |
|---|
| 177 | + if (dma_mapping_error(&snic->pdev->dev, pa)) { |
|---|
| 177 | 178 | SNIC_HOST_ERR(snic->shost, "qdesc: PCI DMA Mapping Fail.\n"); |
|---|
| 178 | 179 | |
|---|
| 179 | 180 | return -ENOMEM; |
|---|
| .. | .. |
|---|
| 186 | 187 | spin_lock_irqsave(&snic->wq_lock[q_num], flags); |
|---|
| 187 | 188 | desc_avail = snic_wqdesc_avail(snic, q_num, req->hdr.type); |
|---|
| 188 | 189 | if (desc_avail <= 0) { |
|---|
| 189 | | - pci_unmap_single(snic->pdev, pa, len, PCI_DMA_TODEVICE); |
|---|
| 190 | + dma_unmap_single(&snic->pdev->dev, pa, len, DMA_TO_DEVICE); |
|---|
| 190 | 191 | req->req_pa = 0; |
|---|
| 191 | 192 | spin_unlock_irqrestore(&snic->wq_lock[q_num], flags); |
|---|
| 192 | 193 | atomic64_inc(&snic->s_stats.misc.wq_alloc_fail); |
|---|
| .. | .. |
|---|
| 350 | 351 | |
|---|
| 351 | 352 | if (rqi->abort_req) { |
|---|
| 352 | 353 | if (rqi->abort_req->req_pa) |
|---|
| 353 | | - pci_unmap_single(snic->pdev, |
|---|
| 354 | + dma_unmap_single(&snic->pdev->dev, |
|---|
| 354 | 355 | rqi->abort_req->req_pa, |
|---|
| 355 | 356 | sizeof(struct snic_host_req), |
|---|
| 356 | | - PCI_DMA_TODEVICE); |
|---|
| 357 | + DMA_TO_DEVICE); |
|---|
| 357 | 358 | |
|---|
| 358 | 359 | mempool_free(rqi->abort_req, snic->req_pool[SNIC_REQ_TM_CACHE]); |
|---|
| 359 | 360 | } |
|---|
| 360 | 361 | |
|---|
| 361 | 362 | if (rqi->dr_req) { |
|---|
| 362 | 363 | if (rqi->dr_req->req_pa) |
|---|
| 363 | | - pci_unmap_single(snic->pdev, |
|---|
| 364 | + dma_unmap_single(&snic->pdev->dev, |
|---|
| 364 | 365 | rqi->dr_req->req_pa, |
|---|
| 365 | 366 | sizeof(struct snic_host_req), |
|---|
| 366 | | - PCI_DMA_TODEVICE); |
|---|
| 367 | + DMA_TO_DEVICE); |
|---|
| 367 | 368 | |
|---|
| 368 | 369 | mempool_free(rqi->dr_req, snic->req_pool[SNIC_REQ_TM_CACHE]); |
|---|
| 369 | 370 | } |
|---|
| 370 | 371 | |
|---|
| 371 | 372 | if (rqi->req->req_pa) |
|---|
| 372 | | - pci_unmap_single(snic->pdev, |
|---|
| 373 | + dma_unmap_single(&snic->pdev->dev, |
|---|
| 373 | 374 | rqi->req->req_pa, |
|---|
| 374 | 375 | rqi->req_len, |
|---|
| 375 | | - PCI_DMA_TODEVICE); |
|---|
| 376 | + DMA_TO_DEVICE); |
|---|
| 376 | 377 | |
|---|
| 377 | 378 | mempool_free(rqi, snic->req_pool[rqi->rq_pool_type]); |
|---|
| 378 | 379 | } |
|---|
| .. | .. |
|---|
| 384 | 385 | |
|---|
| 385 | 386 | sgd = req_to_sgl(rqi_to_req(rqi)); |
|---|
| 386 | 387 | SNIC_BUG_ON(sgd[0].addr == 0); |
|---|
| 387 | | - pci_unmap_single(snic->pdev, |
|---|
| 388 | + dma_unmap_single(&snic->pdev->dev, |
|---|
| 388 | 389 | le64_to_cpu(sgd[0].addr), |
|---|
| 389 | 390 | le32_to_cpu(sgd[0].len), |
|---|
| 390 | | - PCI_DMA_FROMDEVICE); |
|---|
| 391 | + DMA_FROM_DEVICE); |
|---|
| 391 | 392 | } |
|---|
| 392 | 393 | |
|---|
| 393 | 394 | /* |
|---|