.. | .. |
---|
105 | 105 | #define NIXGE_MAX_JUMBO_FRAME_SIZE \ |
---|
106 | 106 | (NIXGE_JUMBO_MTU + NIXGE_HDR_SIZE + NIXGE_TRL_SIZE) |
---|
107 | 107 | |
---|
| 108 | +enum nixge_version { |
---|
| 109 | + NIXGE_V2, |
---|
| 110 | + NIXGE_V3, |
---|
| 111 | + NIXGE_VERSION_COUNT |
---|
| 112 | +}; |
---|
| 113 | + |
---|
108 | 114 | struct nixge_hw_dma_bd { |
---|
109 | | - u32 next; |
---|
110 | | - u32 reserved1; |
---|
111 | | - u32 phys; |
---|
112 | | - u32 reserved2; |
---|
| 115 | + u32 next_lo; |
---|
| 116 | + u32 next_hi; |
---|
| 117 | + u32 phys_lo; |
---|
| 118 | + u32 phys_hi; |
---|
113 | 119 | u32 reserved3; |
---|
114 | 120 | u32 reserved4; |
---|
115 | 121 | u32 cntrl; |
---|
.. | .. |
---|
119 | 125 | u32 app2; |
---|
120 | 126 | u32 app3; |
---|
121 | 127 | u32 app4; |
---|
122 | | - u32 sw_id_offset; |
---|
123 | | - u32 reserved5; |
---|
| 128 | + u32 sw_id_offset_lo; |
---|
| 129 | + u32 sw_id_offset_hi; |
---|
124 | 130 | u32 reserved6; |
---|
125 | 131 | }; |
---|
| 132 | + |
---|
| 133 | +#ifdef CONFIG_PHYS_ADDR_T_64BIT |
---|
| 134 | +#define nixge_hw_dma_bd_set_addr(bd, field, addr) \ |
---|
| 135 | + do { \ |
---|
| 136 | + (bd)->field##_lo = lower_32_bits((addr)); \ |
---|
| 137 | + (bd)->field##_hi = upper_32_bits((addr)); \ |
---|
| 138 | + } while (0) |
---|
| 139 | +#else |
---|
| 140 | +#define nixge_hw_dma_bd_set_addr(bd, field, addr) \ |
---|
| 141 | + ((bd)->field##_lo = lower_32_bits((addr))) |
---|
| 142 | +#endif |
---|
| 143 | + |
---|
| 144 | +#define nixge_hw_dma_bd_set_phys(bd, addr) \ |
---|
| 145 | + nixge_hw_dma_bd_set_addr((bd), phys, (addr)) |
---|
| 146 | + |
---|
| 147 | +#define nixge_hw_dma_bd_set_next(bd, addr) \ |
---|
| 148 | + nixge_hw_dma_bd_set_addr((bd), next, (addr)) |
---|
| 149 | + |
---|
| 150 | +#define nixge_hw_dma_bd_set_offset(bd, addr) \ |
---|
| 151 | + nixge_hw_dma_bd_set_addr((bd), sw_id_offset, (addr)) |
---|
| 152 | + |
---|
| 153 | +#ifdef CONFIG_PHYS_ADDR_T_64BIT |
---|
| 154 | +#define nixge_hw_dma_bd_get_addr(bd, field) \ |
---|
| 155 | + (dma_addr_t)((((u64)(bd)->field##_hi) << 32) | ((bd)->field##_lo)) |
---|
| 156 | +#else |
---|
| 157 | +#define nixge_hw_dma_bd_get_addr(bd, field) \ |
---|
| 158 | + (dma_addr_t)((bd)->field##_lo) |
---|
| 159 | +#endif |
---|
126 | 160 | |
---|
127 | 161 | struct nixge_tx_skb { |
---|
128 | 162 | struct sk_buff *skb; |
---|
.. | .. |
---|
176 | 210 | writel(val, priv->dma_regs + offset); |
---|
177 | 211 | } |
---|
178 | 212 | |
---|
| 213 | +static void nixge_dma_write_desc_reg(struct nixge_priv *priv, off_t offset, |
---|
| 214 | + dma_addr_t addr) |
---|
| 215 | +{ |
---|
| 216 | + writel(lower_32_bits(addr), priv->dma_regs + offset); |
---|
| 217 | +#ifdef CONFIG_PHYS_ADDR_T_64BIT |
---|
| 218 | + writel(upper_32_bits(addr), priv->dma_regs + offset + 4); |
---|
| 219 | +#endif |
---|
| 220 | +} |
---|
| 221 | + |
---|
179 | 222 | static u32 nixge_dma_read_reg(const struct nixge_priv *priv, off_t offset) |
---|
180 | 223 | { |
---|
181 | 224 | return readl(priv->dma_regs + offset); |
---|
.. | .. |
---|
202 | 245 | static void nixge_hw_dma_bd_release(struct net_device *ndev) |
---|
203 | 246 | { |
---|
204 | 247 | struct nixge_priv *priv = netdev_priv(ndev); |
---|
| 248 | + dma_addr_t phys_addr; |
---|
| 249 | + struct sk_buff *skb; |
---|
205 | 250 | int i; |
---|
206 | 251 | |
---|
207 | | - for (i = 0; i < RX_BD_NUM; i++) { |
---|
208 | | - dma_unmap_single(ndev->dev.parent, priv->rx_bd_v[i].phys, |
---|
209 | | - NIXGE_MAX_JUMBO_FRAME_SIZE, DMA_FROM_DEVICE); |
---|
210 | | - dev_kfree_skb((struct sk_buff *) |
---|
211 | | - (priv->rx_bd_v[i].sw_id_offset)); |
---|
212 | | - } |
---|
| 252 | + if (priv->rx_bd_v) { |
---|
| 253 | + for (i = 0; i < RX_BD_NUM; i++) { |
---|
| 254 | + phys_addr = nixge_hw_dma_bd_get_addr(&priv->rx_bd_v[i], |
---|
| 255 | + phys); |
---|
213 | 256 | |
---|
214 | | - if (priv->rx_bd_v) |
---|
| 257 | + dma_unmap_single(ndev->dev.parent, phys_addr, |
---|
| 258 | + NIXGE_MAX_JUMBO_FRAME_SIZE, |
---|
| 259 | + DMA_FROM_DEVICE); |
---|
| 260 | + |
---|
| 261 | + skb = (struct sk_buff *)(uintptr_t) |
---|
| 262 | + nixge_hw_dma_bd_get_addr(&priv->rx_bd_v[i], |
---|
| 263 | + sw_id_offset); |
---|
| 264 | + dev_kfree_skb(skb); |
---|
| 265 | + } |
---|
| 266 | + |
---|
215 | 267 | dma_free_coherent(ndev->dev.parent, |
---|
216 | 268 | sizeof(*priv->rx_bd_v) * RX_BD_NUM, |
---|
217 | 269 | priv->rx_bd_v, |
---|
218 | 270 | priv->rx_bd_p); |
---|
| 271 | + } |
---|
219 | 272 | |
---|
220 | 273 | if (priv->tx_skb) |
---|
221 | 274 | devm_kfree(ndev->dev.parent, priv->tx_skb); |
---|
.. | .. |
---|
231 | 284 | { |
---|
232 | 285 | struct nixge_priv *priv = netdev_priv(ndev); |
---|
233 | 286 | struct sk_buff *skb; |
---|
| 287 | + dma_addr_t phys; |
---|
234 | 288 | u32 cr; |
---|
235 | 289 | int i; |
---|
236 | 290 | |
---|
.. | .. |
---|
240 | 294 | priv->rx_bd_ci = 0; |
---|
241 | 295 | |
---|
242 | 296 | /* Allocate the Tx and Rx buffer descriptors. */ |
---|
243 | | - priv->tx_bd_v = dma_zalloc_coherent(ndev->dev.parent, |
---|
244 | | - sizeof(*priv->tx_bd_v) * TX_BD_NUM, |
---|
245 | | - &priv->tx_bd_p, GFP_KERNEL); |
---|
| 297 | + priv->tx_bd_v = dma_alloc_coherent(ndev->dev.parent, |
---|
| 298 | + sizeof(*priv->tx_bd_v) * TX_BD_NUM, |
---|
| 299 | + &priv->tx_bd_p, GFP_KERNEL); |
---|
246 | 300 | if (!priv->tx_bd_v) |
---|
247 | 301 | goto out; |
---|
248 | 302 | |
---|
.. | .. |
---|
252 | 306 | if (!priv->tx_skb) |
---|
253 | 307 | goto out; |
---|
254 | 308 | |
---|
255 | | - priv->rx_bd_v = dma_zalloc_coherent(ndev->dev.parent, |
---|
256 | | - sizeof(*priv->rx_bd_v) * RX_BD_NUM, |
---|
257 | | - &priv->rx_bd_p, GFP_KERNEL); |
---|
| 309 | + priv->rx_bd_v = dma_alloc_coherent(ndev->dev.parent, |
---|
| 310 | + sizeof(*priv->rx_bd_v) * RX_BD_NUM, |
---|
| 311 | + &priv->rx_bd_p, GFP_KERNEL); |
---|
258 | 312 | if (!priv->rx_bd_v) |
---|
259 | 313 | goto out; |
---|
260 | 314 | |
---|
261 | 315 | for (i = 0; i < TX_BD_NUM; i++) { |
---|
262 | | - priv->tx_bd_v[i].next = priv->tx_bd_p + |
---|
263 | | - sizeof(*priv->tx_bd_v) * |
---|
264 | | - ((i + 1) % TX_BD_NUM); |
---|
| 316 | + nixge_hw_dma_bd_set_next(&priv->tx_bd_v[i], |
---|
| 317 | + priv->tx_bd_p + |
---|
| 318 | + sizeof(*priv->tx_bd_v) * |
---|
| 319 | + ((i + 1) % TX_BD_NUM)); |
---|
265 | 320 | } |
---|
266 | 321 | |
---|
267 | 322 | for (i = 0; i < RX_BD_NUM; i++) { |
---|
268 | | - priv->rx_bd_v[i].next = priv->rx_bd_p + |
---|
269 | | - sizeof(*priv->rx_bd_v) * |
---|
270 | | - ((i + 1) % RX_BD_NUM); |
---|
| 323 | + nixge_hw_dma_bd_set_next(&priv->rx_bd_v[i], |
---|
| 324 | + priv->rx_bd_p |
---|
| 325 | + + sizeof(*priv->rx_bd_v) * |
---|
| 326 | + ((i + 1) % RX_BD_NUM)); |
---|
271 | 327 | |
---|
272 | 328 | skb = netdev_alloc_skb_ip_align(ndev, |
---|
273 | 329 | NIXGE_MAX_JUMBO_FRAME_SIZE); |
---|
274 | 330 | if (!skb) |
---|
275 | 331 | goto out; |
---|
276 | 332 | |
---|
277 | | - priv->rx_bd_v[i].sw_id_offset = (u32)skb; |
---|
278 | | - priv->rx_bd_v[i].phys = |
---|
279 | | - dma_map_single(ndev->dev.parent, |
---|
280 | | - skb->data, |
---|
281 | | - NIXGE_MAX_JUMBO_FRAME_SIZE, |
---|
282 | | - DMA_FROM_DEVICE); |
---|
| 333 | + nixge_hw_dma_bd_set_offset(&priv->rx_bd_v[i], (uintptr_t)skb); |
---|
| 334 | + phys = dma_map_single(ndev->dev.parent, skb->data, |
---|
| 335 | + NIXGE_MAX_JUMBO_FRAME_SIZE, |
---|
| 336 | + DMA_FROM_DEVICE); |
---|
| 337 | + |
---|
| 338 | + nixge_hw_dma_bd_set_phys(&priv->rx_bd_v[i], phys); |
---|
| 339 | + |
---|
283 | 340 | priv->rx_bd_v[i].cntrl = NIXGE_MAX_JUMBO_FRAME_SIZE; |
---|
284 | 341 | } |
---|
285 | 342 | |
---|
.. | .. |
---|
312 | 369 | /* Populate the tail pointer and bring the Rx Axi DMA engine out of |
---|
313 | 370 | * halted state. This will make the Rx side ready for reception. |
---|
314 | 371 | */ |
---|
315 | | - nixge_dma_write_reg(priv, XAXIDMA_RX_CDESC_OFFSET, priv->rx_bd_p); |
---|
| 372 | + nixge_dma_write_desc_reg(priv, XAXIDMA_RX_CDESC_OFFSET, priv->rx_bd_p); |
---|
316 | 373 | cr = nixge_dma_read_reg(priv, XAXIDMA_RX_CR_OFFSET); |
---|
317 | 374 | nixge_dma_write_reg(priv, XAXIDMA_RX_CR_OFFSET, |
---|
318 | 375 | cr | XAXIDMA_CR_RUNSTOP_MASK); |
---|
319 | | - nixge_dma_write_reg(priv, XAXIDMA_RX_TDESC_OFFSET, priv->rx_bd_p + |
---|
| 376 | + nixge_dma_write_desc_reg(priv, XAXIDMA_RX_TDESC_OFFSET, priv->rx_bd_p + |
---|
320 | 377 | (sizeof(*priv->rx_bd_v) * (RX_BD_NUM - 1))); |
---|
321 | 378 | |
---|
322 | 379 | /* Write to the RS (Run-stop) bit in the Tx channel control register. |
---|
323 | 380 | * Tx channel is now ready to run. But only after we write to the |
---|
324 | 381 | * tail pointer register that the Tx channel will start transmitting. |
---|
325 | 382 | */ |
---|
326 | | - nixge_dma_write_reg(priv, XAXIDMA_TX_CDESC_OFFSET, priv->tx_bd_p); |
---|
| 383 | + nixge_dma_write_desc_reg(priv, XAXIDMA_TX_CDESC_OFFSET, priv->tx_bd_p); |
---|
327 | 384 | cr = nixge_dma_read_reg(priv, XAXIDMA_TX_CR_OFFSET); |
---|
328 | 385 | nixge_dma_write_reg(priv, XAXIDMA_TX_CR_OFFSET, |
---|
329 | 386 | cr | XAXIDMA_CR_RUNSTOP_MASK); |
---|
.. | .. |
---|
446 | 503 | return 0; |
---|
447 | 504 | } |
---|
448 | 505 | |
---|
449 | | -static int nixge_start_xmit(struct sk_buff *skb, struct net_device *ndev) |
---|
| 506 | +static netdev_tx_t nixge_start_xmit(struct sk_buff *skb, |
---|
| 507 | + struct net_device *ndev) |
---|
450 | 508 | { |
---|
451 | 509 | struct nixge_priv *priv = netdev_priv(ndev); |
---|
452 | 510 | struct nixge_hw_dma_bd *cur_p; |
---|
453 | 511 | struct nixge_tx_skb *tx_skb; |
---|
454 | | - dma_addr_t tail_p; |
---|
| 512 | + dma_addr_t tail_p, cur_phys; |
---|
455 | 513 | skb_frag_t *frag; |
---|
456 | 514 | u32 num_frag; |
---|
457 | 515 | u32 ii; |
---|
.. | .. |
---|
466 | 524 | return NETDEV_TX_OK; |
---|
467 | 525 | } |
---|
468 | 526 | |
---|
469 | | - cur_p->phys = dma_map_single(ndev->dev.parent, skb->data, |
---|
470 | | - skb_headlen(skb), DMA_TO_DEVICE); |
---|
471 | | - if (dma_mapping_error(ndev->dev.parent, cur_p->phys)) |
---|
| 527 | + cur_phys = dma_map_single(ndev->dev.parent, skb->data, |
---|
| 528 | + skb_headlen(skb), DMA_TO_DEVICE); |
---|
| 529 | + if (dma_mapping_error(ndev->dev.parent, cur_phys)) |
---|
472 | 530 | goto drop; |
---|
| 531 | + nixge_hw_dma_bd_set_phys(cur_p, cur_phys); |
---|
473 | 532 | |
---|
474 | 533 | cur_p->cntrl = skb_headlen(skb) | XAXIDMA_BD_CTRL_TXSOF_MASK; |
---|
475 | 534 | |
---|
476 | 535 | tx_skb->skb = NULL; |
---|
477 | | - tx_skb->mapping = cur_p->phys; |
---|
| 536 | + tx_skb->mapping = cur_phys; |
---|
478 | 537 | tx_skb->size = skb_headlen(skb); |
---|
479 | 538 | tx_skb->mapped_as_page = false; |
---|
480 | 539 | |
---|
.. | .. |
---|
485 | 544 | tx_skb = &priv->tx_skb[priv->tx_bd_tail]; |
---|
486 | 545 | frag = &skb_shinfo(skb)->frags[ii]; |
---|
487 | 546 | |
---|
488 | | - cur_p->phys = skb_frag_dma_map(ndev->dev.parent, frag, 0, |
---|
489 | | - skb_frag_size(frag), |
---|
490 | | - DMA_TO_DEVICE); |
---|
491 | | - if (dma_mapping_error(ndev->dev.parent, cur_p->phys)) |
---|
| 547 | + cur_phys = skb_frag_dma_map(ndev->dev.parent, frag, 0, |
---|
| 548 | + skb_frag_size(frag), |
---|
| 549 | + DMA_TO_DEVICE); |
---|
| 550 | + if (dma_mapping_error(ndev->dev.parent, cur_phys)) |
---|
492 | 551 | goto frag_err; |
---|
| 552 | + nixge_hw_dma_bd_set_phys(cur_p, cur_phys); |
---|
493 | 553 | |
---|
494 | 554 | cur_p->cntrl = skb_frag_size(frag); |
---|
495 | 555 | |
---|
496 | 556 | tx_skb->skb = NULL; |
---|
497 | | - tx_skb->mapping = cur_p->phys; |
---|
| 557 | + tx_skb->mapping = cur_phys; |
---|
498 | 558 | tx_skb->size = skb_frag_size(frag); |
---|
499 | 559 | tx_skb->mapped_as_page = true; |
---|
500 | 560 | } |
---|
.. | .. |
---|
506 | 566 | |
---|
507 | 567 | tail_p = priv->tx_bd_p + sizeof(*priv->tx_bd_v) * priv->tx_bd_tail; |
---|
508 | 568 | /* Start the transfer */ |
---|
509 | | - nixge_dma_write_reg(priv, XAXIDMA_TX_TDESC_OFFSET, tail_p); |
---|
| 569 | + nixge_dma_write_desc_reg(priv, XAXIDMA_TX_TDESC_OFFSET, tail_p); |
---|
510 | 570 | ++priv->tx_bd_tail; |
---|
511 | 571 | priv->tx_bd_tail %= TX_BD_NUM; |
---|
512 | 572 | |
---|
.. | .. |
---|
537 | 597 | struct nixge_priv *priv = netdev_priv(ndev); |
---|
538 | 598 | struct sk_buff *skb, *new_skb; |
---|
539 | 599 | struct nixge_hw_dma_bd *cur_p; |
---|
540 | | - dma_addr_t tail_p = 0; |
---|
| 600 | + dma_addr_t tail_p = 0, cur_phys = 0; |
---|
541 | 601 | u32 packets = 0; |
---|
542 | 602 | u32 length = 0; |
---|
543 | 603 | u32 size = 0; |
---|
.. | .. |
---|
549 | 609 | tail_p = priv->rx_bd_p + sizeof(*priv->rx_bd_v) * |
---|
550 | 610 | priv->rx_bd_ci; |
---|
551 | 611 | |
---|
552 | | - skb = (struct sk_buff *)(cur_p->sw_id_offset); |
---|
| 612 | + skb = (struct sk_buff *)(uintptr_t) |
---|
| 613 | + nixge_hw_dma_bd_get_addr(cur_p, sw_id_offset); |
---|
553 | 614 | |
---|
554 | 615 | length = cur_p->status & XAXIDMA_BD_STS_ACTUAL_LEN_MASK; |
---|
555 | 616 | if (length > NIXGE_MAX_JUMBO_FRAME_SIZE) |
---|
556 | 617 | length = NIXGE_MAX_JUMBO_FRAME_SIZE; |
---|
557 | 618 | |
---|
558 | | - dma_unmap_single(ndev->dev.parent, cur_p->phys, |
---|
| 619 | + dma_unmap_single(ndev->dev.parent, |
---|
| 620 | + nixge_hw_dma_bd_get_addr(cur_p, phys), |
---|
559 | 621 | NIXGE_MAX_JUMBO_FRAME_SIZE, |
---|
560 | 622 | DMA_FROM_DEVICE); |
---|
561 | 623 | |
---|
.. | .. |
---|
579 | 641 | if (!new_skb) |
---|
580 | 642 | return packets; |
---|
581 | 643 | |
---|
582 | | - cur_p->phys = dma_map_single(ndev->dev.parent, new_skb->data, |
---|
583 | | - NIXGE_MAX_JUMBO_FRAME_SIZE, |
---|
584 | | - DMA_FROM_DEVICE); |
---|
585 | | - if (dma_mapping_error(ndev->dev.parent, cur_p->phys)) { |
---|
| 644 | + cur_phys = dma_map_single(ndev->dev.parent, new_skb->data, |
---|
| 645 | + NIXGE_MAX_JUMBO_FRAME_SIZE, |
---|
| 646 | + DMA_FROM_DEVICE); |
---|
| 647 | + if (dma_mapping_error(ndev->dev.parent, cur_phys)) { |
---|
586 | 648 | /* FIXME: bail out and clean up */ |
---|
587 | 649 | netdev_err(ndev, "Failed to map ...\n"); |
---|
588 | 650 | } |
---|
| 651 | + nixge_hw_dma_bd_set_phys(cur_p, cur_phys); |
---|
589 | 652 | cur_p->cntrl = NIXGE_MAX_JUMBO_FRAME_SIZE; |
---|
590 | 653 | cur_p->status = 0; |
---|
591 | | - cur_p->sw_id_offset = (u32)new_skb; |
---|
| 654 | + nixge_hw_dma_bd_set_offset(cur_p, (uintptr_t)new_skb); |
---|
592 | 655 | |
---|
593 | 656 | ++priv->rx_bd_ci; |
---|
594 | 657 | priv->rx_bd_ci %= RX_BD_NUM; |
---|
.. | .. |
---|
599 | 662 | ndev->stats.rx_bytes += size; |
---|
600 | 663 | |
---|
601 | 664 | if (tail_p) |
---|
602 | | - nixge_dma_write_reg(priv, XAXIDMA_RX_TDESC_OFFSET, tail_p); |
---|
| 665 | + nixge_dma_write_desc_reg(priv, XAXIDMA_RX_TDESC_OFFSET, tail_p); |
---|
603 | 666 | |
---|
604 | 667 | return packets; |
---|
605 | 668 | } |
---|
.. | .. |
---|
637 | 700 | struct nixge_priv *priv = netdev_priv(_ndev); |
---|
638 | 701 | struct net_device *ndev = _ndev; |
---|
639 | 702 | unsigned int status; |
---|
| 703 | + dma_addr_t phys; |
---|
640 | 704 | u32 cr; |
---|
641 | 705 | |
---|
642 | 706 | status = nixge_dma_read_reg(priv, XAXIDMA_TX_SR_OFFSET); |
---|
.. | .. |
---|
650 | 714 | return IRQ_NONE; |
---|
651 | 715 | } |
---|
652 | 716 | if (status & XAXIDMA_IRQ_ERROR_MASK) { |
---|
| 717 | + phys = nixge_hw_dma_bd_get_addr(&priv->tx_bd_v[priv->tx_bd_ci], |
---|
| 718 | + phys); |
---|
| 719 | + |
---|
653 | 720 | netdev_err(ndev, "DMA Tx error 0x%x\n", status); |
---|
654 | | - netdev_err(ndev, "Current BD is at: 0x%x\n", |
---|
655 | | - (priv->tx_bd_v[priv->tx_bd_ci]).phys); |
---|
| 721 | + netdev_err(ndev, "Current BD is at: 0x%llx\n", (u64)phys); |
---|
656 | 722 | |
---|
657 | 723 | cr = nixge_dma_read_reg(priv, XAXIDMA_TX_CR_OFFSET); |
---|
658 | 724 | /* Disable coalesce, delay timer and error interrupts */ |
---|
.. | .. |
---|
678 | 744 | struct nixge_priv *priv = netdev_priv(_ndev); |
---|
679 | 745 | struct net_device *ndev = _ndev; |
---|
680 | 746 | unsigned int status; |
---|
| 747 | + dma_addr_t phys; |
---|
681 | 748 | u32 cr; |
---|
682 | 749 | |
---|
683 | 750 | status = nixge_dma_read_reg(priv, XAXIDMA_RX_SR_OFFSET); |
---|
.. | .. |
---|
697 | 764 | return IRQ_NONE; |
---|
698 | 765 | } |
---|
699 | 766 | if (status & XAXIDMA_IRQ_ERROR_MASK) { |
---|
| 767 | + phys = nixge_hw_dma_bd_get_addr(&priv->rx_bd_v[priv->rx_bd_ci], |
---|
| 768 | + phys); |
---|
700 | 769 | netdev_err(ndev, "DMA Rx error 0x%x\n", status); |
---|
701 | | - netdev_err(ndev, "Current BD is at: 0x%x\n", |
---|
702 | | - (priv->rx_bd_v[priv->rx_bd_ci]).phys); |
---|
| 770 | + netdev_err(ndev, "Current BD is at: 0x%llx\n", (u64)phys); |
---|
703 | 771 | |
---|
704 | 772 | cr = nixge_dma_read_reg(priv, XAXIDMA_TX_CR_OFFSET); |
---|
705 | 773 | /* Disable coalesce, delay timer and error interrupts */ |
---|
.. | .. |
---|
720 | 788 | return IRQ_HANDLED; |
---|
721 | 789 | } |
---|
722 | 790 | |
---|
723 | | -static void nixge_dma_err_handler(unsigned long data) |
---|
| 791 | +static void nixge_dma_err_handler(struct tasklet_struct *t) |
---|
724 | 792 | { |
---|
725 | | - struct nixge_priv *lp = (struct nixge_priv *)data; |
---|
| 793 | + struct nixge_priv *lp = from_tasklet(lp, t, dma_err_tasklet); |
---|
726 | 794 | struct nixge_hw_dma_bd *cur_p; |
---|
727 | 795 | struct nixge_tx_skb *tx_skb; |
---|
728 | 796 | u32 cr, i; |
---|
.. | .. |
---|
735 | 803 | tx_skb = &lp->tx_skb[i]; |
---|
736 | 804 | nixge_tx_skb_unmap(lp, tx_skb); |
---|
737 | 805 | |
---|
738 | | - cur_p->phys = 0; |
---|
| 806 | + nixge_hw_dma_bd_set_phys(cur_p, 0); |
---|
739 | 807 | cur_p->cntrl = 0; |
---|
740 | 808 | cur_p->status = 0; |
---|
741 | | - cur_p->sw_id_offset = 0; |
---|
| 809 | + nixge_hw_dma_bd_set_offset(cur_p, 0); |
---|
742 | 810 | } |
---|
743 | 811 | |
---|
744 | 812 | for (i = 0; i < RX_BD_NUM; i++) { |
---|
.. | .. |
---|
779 | 847 | /* Populate the tail pointer and bring the Rx Axi DMA engine out of |
---|
780 | 848 | * halted state. This will make the Rx side ready for reception. |
---|
781 | 849 | */ |
---|
782 | | - nixge_dma_write_reg(lp, XAXIDMA_RX_CDESC_OFFSET, lp->rx_bd_p); |
---|
| 850 | + nixge_dma_write_desc_reg(lp, XAXIDMA_RX_CDESC_OFFSET, lp->rx_bd_p); |
---|
783 | 851 | cr = nixge_dma_read_reg(lp, XAXIDMA_RX_CR_OFFSET); |
---|
784 | 852 | nixge_dma_write_reg(lp, XAXIDMA_RX_CR_OFFSET, |
---|
785 | 853 | cr | XAXIDMA_CR_RUNSTOP_MASK); |
---|
786 | | - nixge_dma_write_reg(lp, XAXIDMA_RX_TDESC_OFFSET, lp->rx_bd_p + |
---|
| 854 | + nixge_dma_write_desc_reg(lp, XAXIDMA_RX_TDESC_OFFSET, lp->rx_bd_p + |
---|
787 | 855 | (sizeof(*lp->rx_bd_v) * (RX_BD_NUM - 1))); |
---|
788 | 856 | |
---|
789 | 857 | /* Write to the RS (Run-stop) bit in the Tx channel control register. |
---|
790 | 858 | * Tx channel is now ready to run. But only after we write to the |
---|
791 | 859 | * tail pointer register that the Tx channel will start transmitting |
---|
792 | 860 | */ |
---|
793 | | - nixge_dma_write_reg(lp, XAXIDMA_TX_CDESC_OFFSET, lp->tx_bd_p); |
---|
| 861 | + nixge_dma_write_desc_reg(lp, XAXIDMA_TX_CDESC_OFFSET, lp->tx_bd_p); |
---|
794 | 862 | cr = nixge_dma_read_reg(lp, XAXIDMA_TX_CR_OFFSET); |
---|
795 | 863 | nixge_dma_write_reg(lp, XAXIDMA_TX_CR_OFFSET, |
---|
796 | 864 | cr | XAXIDMA_CR_RUNSTOP_MASK); |
---|
.. | .. |
---|
812 | 880 | phy_start(phy); |
---|
813 | 881 | |
---|
814 | 882 | /* Enable tasklets for Axi DMA error handling */ |
---|
815 | | - tasklet_init(&priv->dma_err_tasklet, nixge_dma_err_handler, |
---|
816 | | - (unsigned long)priv); |
---|
| 883 | + tasklet_setup(&priv->dma_err_tasklet, nixge_dma_err_handler); |
---|
817 | 884 | |
---|
818 | 885 | napi_enable(&priv->napi); |
---|
819 | 886 | |
---|
.. | .. |
---|
833 | 900 | err_rx_irq: |
---|
834 | 901 | free_irq(priv->tx_irq, ndev); |
---|
835 | 902 | err_tx_irq: |
---|
| 903 | + napi_disable(&priv->napi); |
---|
836 | 904 | phy_stop(phy); |
---|
837 | 905 | phy_disconnect(phy); |
---|
838 | 906 | tasklet_kill(&priv->dma_err_tasklet); |
---|
.. | .. |
---|
924 | 992 | struct ethtool_drvinfo *ed) |
---|
925 | 993 | { |
---|
926 | 994 | strlcpy(ed->driver, "nixge", sizeof(ed->driver)); |
---|
927 | | - strlcpy(ed->bus_info, "platform", sizeof(ed->driver)); |
---|
| 995 | + strlcpy(ed->bus_info, "platform", sizeof(ed->bus_info)); |
---|
928 | 996 | } |
---|
929 | 997 | |
---|
930 | 998 | static int nixge_ethtools_get_coalesce(struct net_device *ndev, |
---|
.. | .. |
---|
953 | 1021 | return -EBUSY; |
---|
954 | 1022 | } |
---|
955 | 1023 | |
---|
956 | | - if (ecoalesce->rx_coalesce_usecs || |
---|
957 | | - ecoalesce->rx_coalesce_usecs_irq || |
---|
958 | | - ecoalesce->rx_max_coalesced_frames_irq || |
---|
959 | | - ecoalesce->tx_coalesce_usecs || |
---|
960 | | - ecoalesce->tx_coalesce_usecs_irq || |
---|
961 | | - ecoalesce->tx_max_coalesced_frames_irq || |
---|
962 | | - ecoalesce->stats_block_coalesce_usecs || |
---|
963 | | - ecoalesce->use_adaptive_rx_coalesce || |
---|
964 | | - ecoalesce->use_adaptive_tx_coalesce || |
---|
965 | | - ecoalesce->pkt_rate_low || |
---|
966 | | - ecoalesce->rx_coalesce_usecs_low || |
---|
967 | | - ecoalesce->rx_max_coalesced_frames_low || |
---|
968 | | - ecoalesce->tx_coalesce_usecs_low || |
---|
969 | | - ecoalesce->tx_max_coalesced_frames_low || |
---|
970 | | - ecoalesce->pkt_rate_high || |
---|
971 | | - ecoalesce->rx_coalesce_usecs_high || |
---|
972 | | - ecoalesce->rx_max_coalesced_frames_high || |
---|
973 | | - ecoalesce->tx_coalesce_usecs_high || |
---|
974 | | - ecoalesce->tx_max_coalesced_frames_high || |
---|
975 | | - ecoalesce->rate_sample_interval) |
---|
976 | | - return -EOPNOTSUPP; |
---|
977 | 1024 | if (ecoalesce->rx_max_coalesced_frames) |
---|
978 | 1025 | priv->coalesce_count_rx = ecoalesce->rx_max_coalesced_frames; |
---|
979 | 1026 | if (ecoalesce->tx_max_coalesced_frames) |
---|
.. | .. |
---|
1017 | 1064 | } |
---|
1018 | 1065 | |
---|
1019 | 1066 | static const struct ethtool_ops nixge_ethtool_ops = { |
---|
| 1067 | + .supported_coalesce_params = ETHTOOL_COALESCE_MAX_FRAMES, |
---|
1020 | 1068 | .get_drvinfo = nixge_ethtools_get_drvinfo, |
---|
1021 | 1069 | .get_coalesce = nixge_ethtools_get_coalesce, |
---|
1022 | 1070 | .set_coalesce = nixge_ethtools_set_coalesce, |
---|
.. | .. |
---|
1165 | 1213 | return mac; |
---|
1166 | 1214 | } |
---|
1167 | 1215 | |
---|
| 1216 | +/* Match table for of_platform binding */ |
---|
| 1217 | +static const struct of_device_id nixge_dt_ids[] = { |
---|
| 1218 | + { .compatible = "ni,xge-enet-2.00", .data = (void *)NIXGE_V2 }, |
---|
| 1219 | + { .compatible = "ni,xge-enet-3.00", .data = (void *)NIXGE_V3 }, |
---|
| 1220 | + {}, |
---|
| 1221 | +}; |
---|
| 1222 | +MODULE_DEVICE_TABLE(of, nixge_dt_ids); |
---|
| 1223 | + |
---|
| 1224 | +static int nixge_of_get_resources(struct platform_device *pdev) |
---|
| 1225 | +{ |
---|
| 1226 | + const struct of_device_id *of_id; |
---|
| 1227 | + enum nixge_version version; |
---|
| 1228 | + struct resource *ctrlres; |
---|
| 1229 | + struct resource *dmares; |
---|
| 1230 | + struct net_device *ndev; |
---|
| 1231 | + struct nixge_priv *priv; |
---|
| 1232 | + |
---|
| 1233 | + ndev = platform_get_drvdata(pdev); |
---|
| 1234 | + priv = netdev_priv(ndev); |
---|
| 1235 | + of_id = of_match_node(nixge_dt_ids, pdev->dev.of_node); |
---|
| 1236 | + if (!of_id) |
---|
| 1237 | + return -ENODEV; |
---|
| 1238 | + |
---|
| 1239 | + version = (enum nixge_version)of_id->data; |
---|
| 1240 | + if (version <= NIXGE_V2) |
---|
| 1241 | + dmares = platform_get_resource(pdev, IORESOURCE_MEM, 0); |
---|
| 1242 | + else |
---|
| 1243 | + dmares = platform_get_resource_byname(pdev, IORESOURCE_MEM, |
---|
| 1244 | + "dma"); |
---|
| 1245 | + |
---|
| 1246 | + priv->dma_regs = devm_ioremap_resource(&pdev->dev, dmares); |
---|
| 1247 | + if (IS_ERR(priv->dma_regs)) { |
---|
| 1248 | + netdev_err(ndev, "failed to map dma regs\n"); |
---|
| 1249 | + return PTR_ERR(priv->dma_regs); |
---|
| 1250 | + } |
---|
| 1251 | + if (version <= NIXGE_V2) { |
---|
| 1252 | + priv->ctrl_regs = priv->dma_regs + NIXGE_REG_CTRL_OFFSET; |
---|
| 1253 | + } else { |
---|
| 1254 | + ctrlres = platform_get_resource_byname(pdev, IORESOURCE_MEM, |
---|
| 1255 | + "ctrl"); |
---|
| 1256 | + priv->ctrl_regs = devm_ioremap_resource(&pdev->dev, ctrlres); |
---|
| 1257 | + } |
---|
| 1258 | + if (IS_ERR(priv->ctrl_regs)) { |
---|
| 1259 | + netdev_err(ndev, "failed to map ctrl regs\n"); |
---|
| 1260 | + return PTR_ERR(priv->ctrl_regs); |
---|
| 1261 | + } |
---|
| 1262 | + return 0; |
---|
| 1263 | +} |
---|
| 1264 | + |
---|
1168 | 1265 | static int nixge_probe(struct platform_device *pdev) |
---|
1169 | 1266 | { |
---|
| 1267 | + struct device_node *mn, *phy_node; |
---|
1170 | 1268 | struct nixge_priv *priv; |
---|
1171 | 1269 | struct net_device *ndev; |
---|
1172 | | - struct resource *dmares; |
---|
1173 | 1270 | const u8 *mac_addr; |
---|
1174 | 1271 | int err; |
---|
1175 | 1272 | |
---|
.. | .. |
---|
1201 | 1298 | priv->dev = &pdev->dev; |
---|
1202 | 1299 | |
---|
1203 | 1300 | netif_napi_add(ndev, &priv->napi, nixge_poll, NAPI_POLL_WEIGHT); |
---|
1204 | | - |
---|
1205 | | - dmares = platform_get_resource(pdev, IORESOURCE_MEM, 0); |
---|
1206 | | - priv->dma_regs = devm_ioremap_resource(&pdev->dev, dmares); |
---|
1207 | | - if (IS_ERR(priv->dma_regs)) { |
---|
1208 | | - netdev_err(ndev, "failed to map dma regs\n"); |
---|
1209 | | - return PTR_ERR(priv->dma_regs); |
---|
1210 | | - } |
---|
1211 | | - priv->ctrl_regs = priv->dma_regs + NIXGE_REG_CTRL_OFFSET; |
---|
| 1301 | + err = nixge_of_get_resources(pdev); |
---|
| 1302 | + if (err) |
---|
| 1303 | + goto free_netdev; |
---|
1212 | 1304 | __nixge_hw_set_mac_address(ndev); |
---|
1213 | 1305 | |
---|
1214 | 1306 | priv->tx_irq = platform_get_irq_byname(pdev, "tx"); |
---|
1215 | 1307 | if (priv->tx_irq < 0) { |
---|
1216 | 1308 | netdev_err(ndev, "could not find 'tx' irq"); |
---|
1217 | | - return priv->tx_irq; |
---|
| 1309 | + err = priv->tx_irq; |
---|
| 1310 | + goto free_netdev; |
---|
1218 | 1311 | } |
---|
1219 | 1312 | |
---|
1220 | 1313 | priv->rx_irq = platform_get_irq_byname(pdev, "rx"); |
---|
1221 | 1314 | if (priv->rx_irq < 0) { |
---|
1222 | 1315 | netdev_err(ndev, "could not find 'rx' irq"); |
---|
1223 | | - return priv->rx_irq; |
---|
| 1316 | + err = priv->rx_irq; |
---|
| 1317 | + goto free_netdev; |
---|
1224 | 1318 | } |
---|
1225 | 1319 | |
---|
1226 | 1320 | priv->coalesce_count_rx = XAXIDMA_DFT_RX_THRESHOLD; |
---|
1227 | 1321 | priv->coalesce_count_tx = XAXIDMA_DFT_TX_THRESHOLD; |
---|
1228 | 1322 | |
---|
1229 | | - err = nixge_mdio_setup(priv, pdev->dev.of_node); |
---|
| 1323 | + mn = of_get_child_by_name(pdev->dev.of_node, "mdio"); |
---|
| 1324 | + if (mn) { |
---|
| 1325 | + err = nixge_mdio_setup(priv, mn); |
---|
| 1326 | + of_node_put(mn); |
---|
| 1327 | + if (err) { |
---|
| 1328 | + netdev_err(ndev, "error registering mdio bus"); |
---|
| 1329 | + goto free_netdev; |
---|
| 1330 | + } |
---|
| 1331 | + } |
---|
| 1332 | + |
---|
| 1333 | + err = of_get_phy_mode(pdev->dev.of_node, &priv->phy_mode); |
---|
1230 | 1334 | if (err) { |
---|
1231 | | - netdev_err(ndev, "error registering mdio bus"); |
---|
1232 | | - goto free_netdev; |
---|
1233 | | - } |
---|
1234 | | - |
---|
1235 | | - priv->phy_mode = of_get_phy_mode(pdev->dev.of_node); |
---|
1236 | | - if ((int)priv->phy_mode < 0) { |
---|
1237 | 1335 | netdev_err(ndev, "not find \"phy-mode\" property\n"); |
---|
1238 | | - err = -EINVAL; |
---|
1239 | 1336 | goto unregister_mdio; |
---|
1240 | 1337 | } |
---|
1241 | 1338 | |
---|
1242 | | - priv->phy_node = of_parse_phandle(pdev->dev.of_node, "phy-handle", 0); |
---|
1243 | | - if (!priv->phy_node) { |
---|
1244 | | - netdev_err(ndev, "not find \"phy-handle\" property\n"); |
---|
1245 | | - err = -EINVAL; |
---|
1246 | | - goto unregister_mdio; |
---|
| 1339 | + phy_node = of_parse_phandle(pdev->dev.of_node, "phy-handle", 0); |
---|
| 1340 | + if (!phy_node && of_phy_is_fixed_link(pdev->dev.of_node)) { |
---|
| 1341 | + err = of_phy_register_fixed_link(pdev->dev.of_node); |
---|
| 1342 | + if (err < 0) { |
---|
| 1343 | + netdev_err(ndev, "broken fixed-link specification\n"); |
---|
| 1344 | + goto unregister_mdio; |
---|
| 1345 | + } |
---|
| 1346 | + phy_node = of_node_get(pdev->dev.of_node); |
---|
1247 | 1347 | } |
---|
| 1348 | + priv->phy_node = phy_node; |
---|
1248 | 1349 | |
---|
1249 | 1350 | err = register_netdev(priv->ndev); |
---|
1250 | 1351 | if (err) { |
---|
1251 | 1352 | netdev_err(ndev, "register_netdev() error (%i)\n", err); |
---|
1252 | | - goto unregister_mdio; |
---|
| 1353 | + goto free_phy; |
---|
1253 | 1354 | } |
---|
1254 | 1355 | |
---|
1255 | 1356 | return 0; |
---|
1256 | 1357 | |
---|
| 1358 | +free_phy: |
---|
| 1359 | + if (of_phy_is_fixed_link(pdev->dev.of_node)) |
---|
| 1360 | + of_phy_deregister_fixed_link(pdev->dev.of_node); |
---|
| 1361 | + of_node_put(phy_node); |
---|
| 1362 | + |
---|
1257 | 1363 | unregister_mdio: |
---|
1258 | | - mdiobus_unregister(priv->mii_bus); |
---|
| 1364 | + if (priv->mii_bus) |
---|
| 1365 | + mdiobus_unregister(priv->mii_bus); |
---|
1259 | 1366 | |
---|
1260 | 1367 | free_netdev: |
---|
1261 | 1368 | free_netdev(ndev); |
---|
.. | .. |
---|
1270 | 1377 | |
---|
1271 | 1378 | unregister_netdev(ndev); |
---|
1272 | 1379 | |
---|
1273 | | - mdiobus_unregister(priv->mii_bus); |
---|
| 1380 | + if (of_phy_is_fixed_link(pdev->dev.of_node)) |
---|
| 1381 | + of_phy_deregister_fixed_link(pdev->dev.of_node); |
---|
| 1382 | + of_node_put(priv->phy_node); |
---|
| 1383 | + |
---|
| 1384 | + if (priv->mii_bus) |
---|
| 1385 | + mdiobus_unregister(priv->mii_bus); |
---|
1274 | 1386 | |
---|
1275 | 1387 | free_netdev(ndev); |
---|
1276 | 1388 | |
---|
1277 | 1389 | return 0; |
---|
1278 | 1390 | } |
---|
1279 | | - |
---|
1280 | | -/* Match table for of_platform binding */ |
---|
1281 | | -static const struct of_device_id nixge_dt_ids[] = { |
---|
1282 | | - { .compatible = "ni,xge-enet-2.00", }, |
---|
1283 | | - {}, |
---|
1284 | | -}; |
---|
1285 | | -MODULE_DEVICE_TABLE(of, nixge_dt_ids); |
---|
1286 | 1391 | |
---|
1287 | 1392 | static struct platform_driver nixge_driver = { |
---|
1288 | 1393 | .probe = nixge_probe, |
---|