| .. | .. |
|---|
| 105 | 105 | #define NIXGE_MAX_JUMBO_FRAME_SIZE \ |
|---|
| 106 | 106 | (NIXGE_JUMBO_MTU + NIXGE_HDR_SIZE + NIXGE_TRL_SIZE) |
|---|
| 107 | 107 | |
|---|
| 108 | +enum nixge_version { |
|---|
| 109 | + NIXGE_V2, |
|---|
| 110 | + NIXGE_V3, |
|---|
| 111 | + NIXGE_VERSION_COUNT |
|---|
| 112 | +}; |
|---|
| 113 | + |
|---|
| 108 | 114 | struct nixge_hw_dma_bd { |
|---|
| 109 | | - u32 next; |
|---|
| 110 | | - u32 reserved1; |
|---|
| 111 | | - u32 phys; |
|---|
| 112 | | - u32 reserved2; |
|---|
| 115 | + u32 next_lo; |
|---|
| 116 | + u32 next_hi; |
|---|
| 117 | + u32 phys_lo; |
|---|
| 118 | + u32 phys_hi; |
|---|
| 113 | 119 | u32 reserved3; |
|---|
| 114 | 120 | u32 reserved4; |
|---|
| 115 | 121 | u32 cntrl; |
|---|
| .. | .. |
|---|
| 119 | 125 | u32 app2; |
|---|
| 120 | 126 | u32 app3; |
|---|
| 121 | 127 | u32 app4; |
|---|
| 122 | | - u32 sw_id_offset; |
|---|
| 123 | | - u32 reserved5; |
|---|
| 128 | + u32 sw_id_offset_lo; |
|---|
| 129 | + u32 sw_id_offset_hi; |
|---|
| 124 | 130 | u32 reserved6; |
|---|
| 125 | 131 | }; |
|---|
| 132 | + |
|---|
| 133 | +#ifdef CONFIG_PHYS_ADDR_T_64BIT |
|---|
| 134 | +#define nixge_hw_dma_bd_set_addr(bd, field, addr) \ |
|---|
| 135 | + do { \ |
|---|
| 136 | + (bd)->field##_lo = lower_32_bits((addr)); \ |
|---|
| 137 | + (bd)->field##_hi = upper_32_bits((addr)); \ |
|---|
| 138 | + } while (0) |
|---|
| 139 | +#else |
|---|
| 140 | +#define nixge_hw_dma_bd_set_addr(bd, field, addr) \ |
|---|
| 141 | + ((bd)->field##_lo = lower_32_bits((addr))) |
|---|
| 142 | +#endif |
|---|
| 143 | + |
|---|
| 144 | +#define nixge_hw_dma_bd_set_phys(bd, addr) \ |
|---|
| 145 | + nixge_hw_dma_bd_set_addr((bd), phys, (addr)) |
|---|
| 146 | + |
|---|
| 147 | +#define nixge_hw_dma_bd_set_next(bd, addr) \ |
|---|
| 148 | + nixge_hw_dma_bd_set_addr((bd), next, (addr)) |
|---|
| 149 | + |
|---|
| 150 | +#define nixge_hw_dma_bd_set_offset(bd, addr) \ |
|---|
| 151 | + nixge_hw_dma_bd_set_addr((bd), sw_id_offset, (addr)) |
|---|
| 152 | + |
|---|
| 153 | +#ifdef CONFIG_PHYS_ADDR_T_64BIT |
|---|
| 154 | +#define nixge_hw_dma_bd_get_addr(bd, field) \ |
|---|
| 155 | + (dma_addr_t)((((u64)(bd)->field##_hi) << 32) | ((bd)->field##_lo)) |
|---|
| 156 | +#else |
|---|
| 157 | +#define nixge_hw_dma_bd_get_addr(bd, field) \ |
|---|
| 158 | + (dma_addr_t)((bd)->field##_lo) |
|---|
| 159 | +#endif |
|---|
| 126 | 160 | |
|---|
| 127 | 161 | struct nixge_tx_skb { |
|---|
| 128 | 162 | struct sk_buff *skb; |
|---|
| .. | .. |
|---|
| 176 | 210 | writel(val, priv->dma_regs + offset); |
|---|
| 177 | 211 | } |
|---|
| 178 | 212 | |
|---|
| 213 | +static void nixge_dma_write_desc_reg(struct nixge_priv *priv, off_t offset, |
|---|
| 214 | + dma_addr_t addr) |
|---|
| 215 | +{ |
|---|
| 216 | + writel(lower_32_bits(addr), priv->dma_regs + offset); |
|---|
| 217 | +#ifdef CONFIG_PHYS_ADDR_T_64BIT |
|---|
| 218 | + writel(upper_32_bits(addr), priv->dma_regs + offset + 4); |
|---|
| 219 | +#endif |
|---|
| 220 | +} |
|---|
| 221 | + |
|---|
| 179 | 222 | static u32 nixge_dma_read_reg(const struct nixge_priv *priv, off_t offset) |
|---|
| 180 | 223 | { |
|---|
| 181 | 224 | return readl(priv->dma_regs + offset); |
|---|
| .. | .. |
|---|
| 202 | 245 | static void nixge_hw_dma_bd_release(struct net_device *ndev) |
|---|
| 203 | 246 | { |
|---|
| 204 | 247 | struct nixge_priv *priv = netdev_priv(ndev); |
|---|
| 248 | + dma_addr_t phys_addr; |
|---|
| 249 | + struct sk_buff *skb; |
|---|
| 205 | 250 | int i; |
|---|
| 206 | 251 | |
|---|
| 207 | | - for (i = 0; i < RX_BD_NUM; i++) { |
|---|
| 208 | | - dma_unmap_single(ndev->dev.parent, priv->rx_bd_v[i].phys, |
|---|
| 209 | | - NIXGE_MAX_JUMBO_FRAME_SIZE, DMA_FROM_DEVICE); |
|---|
| 210 | | - dev_kfree_skb((struct sk_buff *) |
|---|
| 211 | | - (priv->rx_bd_v[i].sw_id_offset)); |
|---|
| 212 | | - } |
|---|
| 252 | + if (priv->rx_bd_v) { |
|---|
| 253 | + for (i = 0; i < RX_BD_NUM; i++) { |
|---|
| 254 | + phys_addr = nixge_hw_dma_bd_get_addr(&priv->rx_bd_v[i], |
|---|
| 255 | + phys); |
|---|
| 213 | 256 | |
|---|
| 214 | | - if (priv->rx_bd_v) |
|---|
| 257 | + dma_unmap_single(ndev->dev.parent, phys_addr, |
|---|
| 258 | + NIXGE_MAX_JUMBO_FRAME_SIZE, |
|---|
| 259 | + DMA_FROM_DEVICE); |
|---|
| 260 | + |
|---|
| 261 | + skb = (struct sk_buff *)(uintptr_t) |
|---|
| 262 | + nixge_hw_dma_bd_get_addr(&priv->rx_bd_v[i], |
|---|
| 263 | + sw_id_offset); |
|---|
| 264 | + dev_kfree_skb(skb); |
|---|
| 265 | + } |
|---|
| 266 | + |
|---|
| 215 | 267 | dma_free_coherent(ndev->dev.parent, |
|---|
| 216 | 268 | sizeof(*priv->rx_bd_v) * RX_BD_NUM, |
|---|
| 217 | 269 | priv->rx_bd_v, |
|---|
| 218 | 270 | priv->rx_bd_p); |
|---|
| 271 | + } |
|---|
| 219 | 272 | |
|---|
| 220 | 273 | if (priv->tx_skb) |
|---|
| 221 | 274 | devm_kfree(ndev->dev.parent, priv->tx_skb); |
|---|
| .. | .. |
|---|
| 231 | 284 | { |
|---|
| 232 | 285 | struct nixge_priv *priv = netdev_priv(ndev); |
|---|
| 233 | 286 | struct sk_buff *skb; |
|---|
| 287 | + dma_addr_t phys; |
|---|
| 234 | 288 | u32 cr; |
|---|
| 235 | 289 | int i; |
|---|
| 236 | 290 | |
|---|
| .. | .. |
|---|
| 240 | 294 | priv->rx_bd_ci = 0; |
|---|
| 241 | 295 | |
|---|
| 242 | 296 | /* Allocate the Tx and Rx buffer descriptors. */ |
|---|
| 243 | | - priv->tx_bd_v = dma_zalloc_coherent(ndev->dev.parent, |
|---|
| 244 | | - sizeof(*priv->tx_bd_v) * TX_BD_NUM, |
|---|
| 245 | | - &priv->tx_bd_p, GFP_KERNEL); |
|---|
| 297 | + priv->tx_bd_v = dma_alloc_coherent(ndev->dev.parent, |
|---|
| 298 | + sizeof(*priv->tx_bd_v) * TX_BD_NUM, |
|---|
| 299 | + &priv->tx_bd_p, GFP_KERNEL); |
|---|
| 246 | 300 | if (!priv->tx_bd_v) |
|---|
| 247 | 301 | goto out; |
|---|
| 248 | 302 | |
|---|
| .. | .. |
|---|
| 252 | 306 | if (!priv->tx_skb) |
|---|
| 253 | 307 | goto out; |
|---|
| 254 | 308 | |
|---|
| 255 | | - priv->rx_bd_v = dma_zalloc_coherent(ndev->dev.parent, |
|---|
| 256 | | - sizeof(*priv->rx_bd_v) * RX_BD_NUM, |
|---|
| 257 | | - &priv->rx_bd_p, GFP_KERNEL); |
|---|
| 309 | + priv->rx_bd_v = dma_alloc_coherent(ndev->dev.parent, |
|---|
| 310 | + sizeof(*priv->rx_bd_v) * RX_BD_NUM, |
|---|
| 311 | + &priv->rx_bd_p, GFP_KERNEL); |
|---|
| 258 | 312 | if (!priv->rx_bd_v) |
|---|
| 259 | 313 | goto out; |
|---|
| 260 | 314 | |
|---|
| 261 | 315 | for (i = 0; i < TX_BD_NUM; i++) { |
|---|
| 262 | | - priv->tx_bd_v[i].next = priv->tx_bd_p + |
|---|
| 263 | | - sizeof(*priv->tx_bd_v) * |
|---|
| 264 | | - ((i + 1) % TX_BD_NUM); |
|---|
| 316 | + nixge_hw_dma_bd_set_next(&priv->tx_bd_v[i], |
|---|
| 317 | + priv->tx_bd_p + |
|---|
| 318 | + sizeof(*priv->tx_bd_v) * |
|---|
| 319 | + ((i + 1) % TX_BD_NUM)); |
|---|
| 265 | 320 | } |
|---|
| 266 | 321 | |
|---|
| 267 | 322 | for (i = 0; i < RX_BD_NUM; i++) { |
|---|
| 268 | | - priv->rx_bd_v[i].next = priv->rx_bd_p + |
|---|
| 269 | | - sizeof(*priv->rx_bd_v) * |
|---|
| 270 | | - ((i + 1) % RX_BD_NUM); |
|---|
| 323 | + nixge_hw_dma_bd_set_next(&priv->rx_bd_v[i], |
|---|
| 324 | + priv->rx_bd_p |
|---|
| 325 | + + sizeof(*priv->rx_bd_v) * |
|---|
| 326 | + ((i + 1) % RX_BD_NUM)); |
|---|
| 271 | 327 | |
|---|
| 272 | 328 | skb = netdev_alloc_skb_ip_align(ndev, |
|---|
| 273 | 329 | NIXGE_MAX_JUMBO_FRAME_SIZE); |
|---|
| 274 | 330 | if (!skb) |
|---|
| 275 | 331 | goto out; |
|---|
| 276 | 332 | |
|---|
| 277 | | - priv->rx_bd_v[i].sw_id_offset = (u32)skb; |
|---|
| 278 | | - priv->rx_bd_v[i].phys = |
|---|
| 279 | | - dma_map_single(ndev->dev.parent, |
|---|
| 280 | | - skb->data, |
|---|
| 281 | | - NIXGE_MAX_JUMBO_FRAME_SIZE, |
|---|
| 282 | | - DMA_FROM_DEVICE); |
|---|
| 333 | + nixge_hw_dma_bd_set_offset(&priv->rx_bd_v[i], (uintptr_t)skb); |
|---|
| 334 | + phys = dma_map_single(ndev->dev.parent, skb->data, |
|---|
| 335 | + NIXGE_MAX_JUMBO_FRAME_SIZE, |
|---|
| 336 | + DMA_FROM_DEVICE); |
|---|
| 337 | + |
|---|
| 338 | + nixge_hw_dma_bd_set_phys(&priv->rx_bd_v[i], phys); |
|---|
| 339 | + |
|---|
| 283 | 340 | priv->rx_bd_v[i].cntrl = NIXGE_MAX_JUMBO_FRAME_SIZE; |
|---|
| 284 | 341 | } |
|---|
| 285 | 342 | |
|---|
| .. | .. |
|---|
| 312 | 369 | /* Populate the tail pointer and bring the Rx Axi DMA engine out of |
|---|
| 313 | 370 | * halted state. This will make the Rx side ready for reception. |
|---|
| 314 | 371 | */ |
|---|
| 315 | | - nixge_dma_write_reg(priv, XAXIDMA_RX_CDESC_OFFSET, priv->rx_bd_p); |
|---|
| 372 | + nixge_dma_write_desc_reg(priv, XAXIDMA_RX_CDESC_OFFSET, priv->rx_bd_p); |
|---|
| 316 | 373 | cr = nixge_dma_read_reg(priv, XAXIDMA_RX_CR_OFFSET); |
|---|
| 317 | 374 | nixge_dma_write_reg(priv, XAXIDMA_RX_CR_OFFSET, |
|---|
| 318 | 375 | cr | XAXIDMA_CR_RUNSTOP_MASK); |
|---|
| 319 | | - nixge_dma_write_reg(priv, XAXIDMA_RX_TDESC_OFFSET, priv->rx_bd_p + |
|---|
| 376 | + nixge_dma_write_desc_reg(priv, XAXIDMA_RX_TDESC_OFFSET, priv->rx_bd_p + |
|---|
| 320 | 377 | (sizeof(*priv->rx_bd_v) * (RX_BD_NUM - 1))); |
|---|
| 321 | 378 | |
|---|
| 322 | 379 | /* Write to the RS (Run-stop) bit in the Tx channel control register. |
|---|
| 323 | 380 | * Tx channel is now ready to run. But only after we write to the |
|---|
| 324 | 381 | * tail pointer register that the Tx channel will start transmitting. |
|---|
| 325 | 382 | */ |
|---|
| 326 | | - nixge_dma_write_reg(priv, XAXIDMA_TX_CDESC_OFFSET, priv->tx_bd_p); |
|---|
| 383 | + nixge_dma_write_desc_reg(priv, XAXIDMA_TX_CDESC_OFFSET, priv->tx_bd_p); |
|---|
| 327 | 384 | cr = nixge_dma_read_reg(priv, XAXIDMA_TX_CR_OFFSET); |
|---|
| 328 | 385 | nixge_dma_write_reg(priv, XAXIDMA_TX_CR_OFFSET, |
|---|
| 329 | 386 | cr | XAXIDMA_CR_RUNSTOP_MASK); |
|---|
| .. | .. |
|---|
| 446 | 503 | return 0; |
|---|
| 447 | 504 | } |
|---|
| 448 | 505 | |
|---|
| 449 | | -static int nixge_start_xmit(struct sk_buff *skb, struct net_device *ndev) |
|---|
| 506 | +static netdev_tx_t nixge_start_xmit(struct sk_buff *skb, |
|---|
| 507 | + struct net_device *ndev) |
|---|
| 450 | 508 | { |
|---|
| 451 | 509 | struct nixge_priv *priv = netdev_priv(ndev); |
|---|
| 452 | 510 | struct nixge_hw_dma_bd *cur_p; |
|---|
| 453 | 511 | struct nixge_tx_skb *tx_skb; |
|---|
| 454 | | - dma_addr_t tail_p; |
|---|
| 512 | + dma_addr_t tail_p, cur_phys; |
|---|
| 455 | 513 | skb_frag_t *frag; |
|---|
| 456 | 514 | u32 num_frag; |
|---|
| 457 | 515 | u32 ii; |
|---|
| .. | .. |
|---|
| 466 | 524 | return NETDEV_TX_OK; |
|---|
| 467 | 525 | } |
|---|
| 468 | 526 | |
|---|
| 469 | | - cur_p->phys = dma_map_single(ndev->dev.parent, skb->data, |
|---|
| 470 | | - skb_headlen(skb), DMA_TO_DEVICE); |
|---|
| 471 | | - if (dma_mapping_error(ndev->dev.parent, cur_p->phys)) |
|---|
| 527 | + cur_phys = dma_map_single(ndev->dev.parent, skb->data, |
|---|
| 528 | + skb_headlen(skb), DMA_TO_DEVICE); |
|---|
| 529 | + if (dma_mapping_error(ndev->dev.parent, cur_phys)) |
|---|
| 472 | 530 | goto drop; |
|---|
| 531 | + nixge_hw_dma_bd_set_phys(cur_p, cur_phys); |
|---|
| 473 | 532 | |
|---|
| 474 | 533 | cur_p->cntrl = skb_headlen(skb) | XAXIDMA_BD_CTRL_TXSOF_MASK; |
|---|
| 475 | 534 | |
|---|
| 476 | 535 | tx_skb->skb = NULL; |
|---|
| 477 | | - tx_skb->mapping = cur_p->phys; |
|---|
| 536 | + tx_skb->mapping = cur_phys; |
|---|
| 478 | 537 | tx_skb->size = skb_headlen(skb); |
|---|
| 479 | 538 | tx_skb->mapped_as_page = false; |
|---|
| 480 | 539 | |
|---|
| .. | .. |
|---|
| 485 | 544 | tx_skb = &priv->tx_skb[priv->tx_bd_tail]; |
|---|
| 486 | 545 | frag = &skb_shinfo(skb)->frags[ii]; |
|---|
| 487 | 546 | |
|---|
| 488 | | - cur_p->phys = skb_frag_dma_map(ndev->dev.parent, frag, 0, |
|---|
| 489 | | - skb_frag_size(frag), |
|---|
| 490 | | - DMA_TO_DEVICE); |
|---|
| 491 | | - if (dma_mapping_error(ndev->dev.parent, cur_p->phys)) |
|---|
| 547 | + cur_phys = skb_frag_dma_map(ndev->dev.parent, frag, 0, |
|---|
| 548 | + skb_frag_size(frag), |
|---|
| 549 | + DMA_TO_DEVICE); |
|---|
| 550 | + if (dma_mapping_error(ndev->dev.parent, cur_phys)) |
|---|
| 492 | 551 | goto frag_err; |
|---|
| 552 | + nixge_hw_dma_bd_set_phys(cur_p, cur_phys); |
|---|
| 493 | 553 | |
|---|
| 494 | 554 | cur_p->cntrl = skb_frag_size(frag); |
|---|
| 495 | 555 | |
|---|
| 496 | 556 | tx_skb->skb = NULL; |
|---|
| 497 | | - tx_skb->mapping = cur_p->phys; |
|---|
| 557 | + tx_skb->mapping = cur_phys; |
|---|
| 498 | 558 | tx_skb->size = skb_frag_size(frag); |
|---|
| 499 | 559 | tx_skb->mapped_as_page = true; |
|---|
| 500 | 560 | } |
|---|
| .. | .. |
|---|
| 506 | 566 | |
|---|
| 507 | 567 | tail_p = priv->tx_bd_p + sizeof(*priv->tx_bd_v) * priv->tx_bd_tail; |
|---|
| 508 | 568 | /* Start the transfer */ |
|---|
| 509 | | - nixge_dma_write_reg(priv, XAXIDMA_TX_TDESC_OFFSET, tail_p); |
|---|
| 569 | + nixge_dma_write_desc_reg(priv, XAXIDMA_TX_TDESC_OFFSET, tail_p); |
|---|
| 510 | 570 | ++priv->tx_bd_tail; |
|---|
| 511 | 571 | priv->tx_bd_tail %= TX_BD_NUM; |
|---|
| 512 | 572 | |
|---|
| .. | .. |
|---|
| 537 | 597 | struct nixge_priv *priv = netdev_priv(ndev); |
|---|
| 538 | 598 | struct sk_buff *skb, *new_skb; |
|---|
| 539 | 599 | struct nixge_hw_dma_bd *cur_p; |
|---|
| 540 | | - dma_addr_t tail_p = 0; |
|---|
| 600 | + dma_addr_t tail_p = 0, cur_phys = 0; |
|---|
| 541 | 601 | u32 packets = 0; |
|---|
| 542 | 602 | u32 length = 0; |
|---|
| 543 | 603 | u32 size = 0; |
|---|
| .. | .. |
|---|
| 549 | 609 | tail_p = priv->rx_bd_p + sizeof(*priv->rx_bd_v) * |
|---|
| 550 | 610 | priv->rx_bd_ci; |
|---|
| 551 | 611 | |
|---|
| 552 | | - skb = (struct sk_buff *)(cur_p->sw_id_offset); |
|---|
| 612 | + skb = (struct sk_buff *)(uintptr_t) |
|---|
| 613 | + nixge_hw_dma_bd_get_addr(cur_p, sw_id_offset); |
|---|
| 553 | 614 | |
|---|
| 554 | 615 | length = cur_p->status & XAXIDMA_BD_STS_ACTUAL_LEN_MASK; |
|---|
| 555 | 616 | if (length > NIXGE_MAX_JUMBO_FRAME_SIZE) |
|---|
| 556 | 617 | length = NIXGE_MAX_JUMBO_FRAME_SIZE; |
|---|
| 557 | 618 | |
|---|
| 558 | | - dma_unmap_single(ndev->dev.parent, cur_p->phys, |
|---|
| 619 | + dma_unmap_single(ndev->dev.parent, |
|---|
| 620 | + nixge_hw_dma_bd_get_addr(cur_p, phys), |
|---|
| 559 | 621 | NIXGE_MAX_JUMBO_FRAME_SIZE, |
|---|
| 560 | 622 | DMA_FROM_DEVICE); |
|---|
| 561 | 623 | |
|---|
| .. | .. |
|---|
| 579 | 641 | if (!new_skb) |
|---|
| 580 | 642 | return packets; |
|---|
| 581 | 643 | |
|---|
| 582 | | - cur_p->phys = dma_map_single(ndev->dev.parent, new_skb->data, |
|---|
| 583 | | - NIXGE_MAX_JUMBO_FRAME_SIZE, |
|---|
| 584 | | - DMA_FROM_DEVICE); |
|---|
| 585 | | - if (dma_mapping_error(ndev->dev.parent, cur_p->phys)) { |
|---|
| 644 | + cur_phys = dma_map_single(ndev->dev.parent, new_skb->data, |
|---|
| 645 | + NIXGE_MAX_JUMBO_FRAME_SIZE, |
|---|
| 646 | + DMA_FROM_DEVICE); |
|---|
| 647 | + if (dma_mapping_error(ndev->dev.parent, cur_phys)) { |
|---|
| 586 | 648 | /* FIXME: bail out and clean up */ |
|---|
| 587 | 649 | netdev_err(ndev, "Failed to map ...\n"); |
|---|
| 588 | 650 | } |
|---|
| 651 | + nixge_hw_dma_bd_set_phys(cur_p, cur_phys); |
|---|
| 589 | 652 | cur_p->cntrl = NIXGE_MAX_JUMBO_FRAME_SIZE; |
|---|
| 590 | 653 | cur_p->status = 0; |
|---|
| 591 | | - cur_p->sw_id_offset = (u32)new_skb; |
|---|
| 654 | + nixge_hw_dma_bd_set_offset(cur_p, (uintptr_t)new_skb); |
|---|
| 592 | 655 | |
|---|
| 593 | 656 | ++priv->rx_bd_ci; |
|---|
| 594 | 657 | priv->rx_bd_ci %= RX_BD_NUM; |
|---|
| .. | .. |
|---|
| 599 | 662 | ndev->stats.rx_bytes += size; |
|---|
| 600 | 663 | |
|---|
| 601 | 664 | if (tail_p) |
|---|
| 602 | | - nixge_dma_write_reg(priv, XAXIDMA_RX_TDESC_OFFSET, tail_p); |
|---|
| 665 | + nixge_dma_write_desc_reg(priv, XAXIDMA_RX_TDESC_OFFSET, tail_p); |
|---|
| 603 | 666 | |
|---|
| 604 | 667 | return packets; |
|---|
| 605 | 668 | } |
|---|
| .. | .. |
|---|
| 637 | 700 | struct nixge_priv *priv = netdev_priv(_ndev); |
|---|
| 638 | 701 | struct net_device *ndev = _ndev; |
|---|
| 639 | 702 | unsigned int status; |
|---|
| 703 | + dma_addr_t phys; |
|---|
| 640 | 704 | u32 cr; |
|---|
| 641 | 705 | |
|---|
| 642 | 706 | status = nixge_dma_read_reg(priv, XAXIDMA_TX_SR_OFFSET); |
|---|
| .. | .. |
|---|
| 650 | 714 | return IRQ_NONE; |
|---|
| 651 | 715 | } |
|---|
| 652 | 716 | if (status & XAXIDMA_IRQ_ERROR_MASK) { |
|---|
| 717 | + phys = nixge_hw_dma_bd_get_addr(&priv->tx_bd_v[priv->tx_bd_ci], |
|---|
| 718 | + phys); |
|---|
| 719 | + |
|---|
| 653 | 720 | netdev_err(ndev, "DMA Tx error 0x%x\n", status); |
|---|
| 654 | | - netdev_err(ndev, "Current BD is at: 0x%x\n", |
|---|
| 655 | | - (priv->tx_bd_v[priv->tx_bd_ci]).phys); |
|---|
| 721 | + netdev_err(ndev, "Current BD is at: 0x%llx\n", (u64)phys); |
|---|
| 656 | 722 | |
|---|
| 657 | 723 | cr = nixge_dma_read_reg(priv, XAXIDMA_TX_CR_OFFSET); |
|---|
| 658 | 724 | /* Disable coalesce, delay timer and error interrupts */ |
|---|
| .. | .. |
|---|
| 678 | 744 | struct nixge_priv *priv = netdev_priv(_ndev); |
|---|
| 679 | 745 | struct net_device *ndev = _ndev; |
|---|
| 680 | 746 | unsigned int status; |
|---|
| 747 | + dma_addr_t phys; |
|---|
| 681 | 748 | u32 cr; |
|---|
| 682 | 749 | |
|---|
| 683 | 750 | status = nixge_dma_read_reg(priv, XAXIDMA_RX_SR_OFFSET); |
|---|
| .. | .. |
|---|
| 697 | 764 | return IRQ_NONE; |
|---|
| 698 | 765 | } |
|---|
| 699 | 766 | if (status & XAXIDMA_IRQ_ERROR_MASK) { |
|---|
| 767 | + phys = nixge_hw_dma_bd_get_addr(&priv->rx_bd_v[priv->rx_bd_ci], |
|---|
| 768 | + phys); |
|---|
| 700 | 769 | netdev_err(ndev, "DMA Rx error 0x%x\n", status); |
|---|
| 701 | | - netdev_err(ndev, "Current BD is at: 0x%x\n", |
|---|
| 702 | | - (priv->rx_bd_v[priv->rx_bd_ci]).phys); |
|---|
| 770 | + netdev_err(ndev, "Current BD is at: 0x%llx\n", (u64)phys); |
|---|
| 703 | 771 | |
|---|
| 704 | 772 | cr = nixge_dma_read_reg(priv, XAXIDMA_TX_CR_OFFSET); |
|---|
| 705 | 773 | /* Disable coalesce, delay timer and error interrupts */ |
|---|
| .. | .. |
|---|
| 720 | 788 | return IRQ_HANDLED; |
|---|
| 721 | 789 | } |
|---|
| 722 | 790 | |
|---|
| 723 | | -static void nixge_dma_err_handler(unsigned long data) |
|---|
| 791 | +static void nixge_dma_err_handler(struct tasklet_struct *t) |
|---|
| 724 | 792 | { |
|---|
| 725 | | - struct nixge_priv *lp = (struct nixge_priv *)data; |
|---|
| 793 | + struct nixge_priv *lp = from_tasklet(lp, t, dma_err_tasklet); |
|---|
| 726 | 794 | struct nixge_hw_dma_bd *cur_p; |
|---|
| 727 | 795 | struct nixge_tx_skb *tx_skb; |
|---|
| 728 | 796 | u32 cr, i; |
|---|
| .. | .. |
|---|
| 735 | 803 | tx_skb = &lp->tx_skb[i]; |
|---|
| 736 | 804 | nixge_tx_skb_unmap(lp, tx_skb); |
|---|
| 737 | 805 | |
|---|
| 738 | | - cur_p->phys = 0; |
|---|
| 806 | + nixge_hw_dma_bd_set_phys(cur_p, 0); |
|---|
| 739 | 807 | cur_p->cntrl = 0; |
|---|
| 740 | 808 | cur_p->status = 0; |
|---|
| 741 | | - cur_p->sw_id_offset = 0; |
|---|
| 809 | + nixge_hw_dma_bd_set_offset(cur_p, 0); |
|---|
| 742 | 810 | } |
|---|
| 743 | 811 | |
|---|
| 744 | 812 | for (i = 0; i < RX_BD_NUM; i++) { |
|---|
| .. | .. |
|---|
| 779 | 847 | /* Populate the tail pointer and bring the Rx Axi DMA engine out of |
|---|
| 780 | 848 | * halted state. This will make the Rx side ready for reception. |
|---|
| 781 | 849 | */ |
|---|
| 782 | | - nixge_dma_write_reg(lp, XAXIDMA_RX_CDESC_OFFSET, lp->rx_bd_p); |
|---|
| 850 | + nixge_dma_write_desc_reg(lp, XAXIDMA_RX_CDESC_OFFSET, lp->rx_bd_p); |
|---|
| 783 | 851 | cr = nixge_dma_read_reg(lp, XAXIDMA_RX_CR_OFFSET); |
|---|
| 784 | 852 | nixge_dma_write_reg(lp, XAXIDMA_RX_CR_OFFSET, |
|---|
| 785 | 853 | cr | XAXIDMA_CR_RUNSTOP_MASK); |
|---|
| 786 | | - nixge_dma_write_reg(lp, XAXIDMA_RX_TDESC_OFFSET, lp->rx_bd_p + |
|---|
| 854 | + nixge_dma_write_desc_reg(lp, XAXIDMA_RX_TDESC_OFFSET, lp->rx_bd_p + |
|---|
| 787 | 855 | (sizeof(*lp->rx_bd_v) * (RX_BD_NUM - 1))); |
|---|
| 788 | 856 | |
|---|
| 789 | 857 | /* Write to the RS (Run-stop) bit in the Tx channel control register. |
|---|
| 790 | 858 | * Tx channel is now ready to run. But only after we write to the |
|---|
| 791 | 859 | * tail pointer register that the Tx channel will start transmitting |
|---|
| 792 | 860 | */ |
|---|
| 793 | | - nixge_dma_write_reg(lp, XAXIDMA_TX_CDESC_OFFSET, lp->tx_bd_p); |
|---|
| 861 | + nixge_dma_write_desc_reg(lp, XAXIDMA_TX_CDESC_OFFSET, lp->tx_bd_p); |
|---|
| 794 | 862 | cr = nixge_dma_read_reg(lp, XAXIDMA_TX_CR_OFFSET); |
|---|
| 795 | 863 | nixge_dma_write_reg(lp, XAXIDMA_TX_CR_OFFSET, |
|---|
| 796 | 864 | cr | XAXIDMA_CR_RUNSTOP_MASK); |
|---|
| .. | .. |
|---|
| 812 | 880 | phy_start(phy); |
|---|
| 813 | 881 | |
|---|
| 814 | 882 | /* Enable tasklets for Axi DMA error handling */ |
|---|
| 815 | | - tasklet_init(&priv->dma_err_tasklet, nixge_dma_err_handler, |
|---|
| 816 | | - (unsigned long)priv); |
|---|
| 883 | + tasklet_setup(&priv->dma_err_tasklet, nixge_dma_err_handler); |
|---|
| 817 | 884 | |
|---|
| 818 | 885 | napi_enable(&priv->napi); |
|---|
| 819 | 886 | |
|---|
| .. | .. |
|---|
| 833 | 900 | err_rx_irq: |
|---|
| 834 | 901 | free_irq(priv->tx_irq, ndev); |
|---|
| 835 | 902 | err_tx_irq: |
|---|
| 903 | + napi_disable(&priv->napi); |
|---|
| 836 | 904 | phy_stop(phy); |
|---|
| 837 | 905 | phy_disconnect(phy); |
|---|
| 838 | 906 | tasklet_kill(&priv->dma_err_tasklet); |
|---|
| .. | .. |
|---|
| 924 | 992 | struct ethtool_drvinfo *ed) |
|---|
| 925 | 993 | { |
|---|
| 926 | 994 | strlcpy(ed->driver, "nixge", sizeof(ed->driver)); |
|---|
| 927 | | - strlcpy(ed->bus_info, "platform", sizeof(ed->driver)); |
|---|
| 995 | + strlcpy(ed->bus_info, "platform", sizeof(ed->bus_info)); |
|---|
| 928 | 996 | } |
|---|
| 929 | 997 | |
|---|
| 930 | 998 | static int nixge_ethtools_get_coalesce(struct net_device *ndev, |
|---|
| .. | .. |
|---|
| 953 | 1021 | return -EBUSY; |
|---|
| 954 | 1022 | } |
|---|
| 955 | 1023 | |
|---|
| 956 | | - if (ecoalesce->rx_coalesce_usecs || |
|---|
| 957 | | - ecoalesce->rx_coalesce_usecs_irq || |
|---|
| 958 | | - ecoalesce->rx_max_coalesced_frames_irq || |
|---|
| 959 | | - ecoalesce->tx_coalesce_usecs || |
|---|
| 960 | | - ecoalesce->tx_coalesce_usecs_irq || |
|---|
| 961 | | - ecoalesce->tx_max_coalesced_frames_irq || |
|---|
| 962 | | - ecoalesce->stats_block_coalesce_usecs || |
|---|
| 963 | | - ecoalesce->use_adaptive_rx_coalesce || |
|---|
| 964 | | - ecoalesce->use_adaptive_tx_coalesce || |
|---|
| 965 | | - ecoalesce->pkt_rate_low || |
|---|
| 966 | | - ecoalesce->rx_coalesce_usecs_low || |
|---|
| 967 | | - ecoalesce->rx_max_coalesced_frames_low || |
|---|
| 968 | | - ecoalesce->tx_coalesce_usecs_low || |
|---|
| 969 | | - ecoalesce->tx_max_coalesced_frames_low || |
|---|
| 970 | | - ecoalesce->pkt_rate_high || |
|---|
| 971 | | - ecoalesce->rx_coalesce_usecs_high || |
|---|
| 972 | | - ecoalesce->rx_max_coalesced_frames_high || |
|---|
| 973 | | - ecoalesce->tx_coalesce_usecs_high || |
|---|
| 974 | | - ecoalesce->tx_max_coalesced_frames_high || |
|---|
| 975 | | - ecoalesce->rate_sample_interval) |
|---|
| 976 | | - return -EOPNOTSUPP; |
|---|
| 977 | 1024 | if (ecoalesce->rx_max_coalesced_frames) |
|---|
| 978 | 1025 | priv->coalesce_count_rx = ecoalesce->rx_max_coalesced_frames; |
|---|
| 979 | 1026 | if (ecoalesce->tx_max_coalesced_frames) |
|---|
| .. | .. |
|---|
| 1017 | 1064 | } |
|---|
| 1018 | 1065 | |
|---|
| 1019 | 1066 | static const struct ethtool_ops nixge_ethtool_ops = { |
|---|
| 1067 | + .supported_coalesce_params = ETHTOOL_COALESCE_MAX_FRAMES, |
|---|
| 1020 | 1068 | .get_drvinfo = nixge_ethtools_get_drvinfo, |
|---|
| 1021 | 1069 | .get_coalesce = nixge_ethtools_get_coalesce, |
|---|
| 1022 | 1070 | .set_coalesce = nixge_ethtools_set_coalesce, |
|---|
| .. | .. |
|---|
| 1165 | 1213 | return mac; |
|---|
| 1166 | 1214 | } |
|---|
| 1167 | 1215 | |
|---|
| 1216 | +/* Match table for of_platform binding */ |
|---|
| 1217 | +static const struct of_device_id nixge_dt_ids[] = { |
|---|
| 1218 | + { .compatible = "ni,xge-enet-2.00", .data = (void *)NIXGE_V2 }, |
|---|
| 1219 | + { .compatible = "ni,xge-enet-3.00", .data = (void *)NIXGE_V3 }, |
|---|
| 1220 | + {}, |
|---|
| 1221 | +}; |
|---|
| 1222 | +MODULE_DEVICE_TABLE(of, nixge_dt_ids); |
|---|
| 1223 | + |
|---|
| 1224 | +static int nixge_of_get_resources(struct platform_device *pdev) |
|---|
| 1225 | +{ |
|---|
| 1226 | + const struct of_device_id *of_id; |
|---|
| 1227 | + enum nixge_version version; |
|---|
| 1228 | + struct resource *ctrlres; |
|---|
| 1229 | + struct resource *dmares; |
|---|
| 1230 | + struct net_device *ndev; |
|---|
| 1231 | + struct nixge_priv *priv; |
|---|
| 1232 | + |
|---|
| 1233 | + ndev = platform_get_drvdata(pdev); |
|---|
| 1234 | + priv = netdev_priv(ndev); |
|---|
| 1235 | + of_id = of_match_node(nixge_dt_ids, pdev->dev.of_node); |
|---|
| 1236 | + if (!of_id) |
|---|
| 1237 | + return -ENODEV; |
|---|
| 1238 | + |
|---|
| 1239 | + version = (enum nixge_version)of_id->data; |
|---|
| 1240 | + if (version <= NIXGE_V2) |
|---|
| 1241 | + dmares = platform_get_resource(pdev, IORESOURCE_MEM, 0); |
|---|
| 1242 | + else |
|---|
| 1243 | + dmares = platform_get_resource_byname(pdev, IORESOURCE_MEM, |
|---|
| 1244 | + "dma"); |
|---|
| 1245 | + |
|---|
| 1246 | + priv->dma_regs = devm_ioremap_resource(&pdev->dev, dmares); |
|---|
| 1247 | + if (IS_ERR(priv->dma_regs)) { |
|---|
| 1248 | + netdev_err(ndev, "failed to map dma regs\n"); |
|---|
| 1249 | + return PTR_ERR(priv->dma_regs); |
|---|
| 1250 | + } |
|---|
| 1251 | + if (version <= NIXGE_V2) { |
|---|
| 1252 | + priv->ctrl_regs = priv->dma_regs + NIXGE_REG_CTRL_OFFSET; |
|---|
| 1253 | + } else { |
|---|
| 1254 | + ctrlres = platform_get_resource_byname(pdev, IORESOURCE_MEM, |
|---|
| 1255 | + "ctrl"); |
|---|
| 1256 | + priv->ctrl_regs = devm_ioremap_resource(&pdev->dev, ctrlres); |
|---|
| 1257 | + } |
|---|
| 1258 | + if (IS_ERR(priv->ctrl_regs)) { |
|---|
| 1259 | + netdev_err(ndev, "failed to map ctrl regs\n"); |
|---|
| 1260 | + return PTR_ERR(priv->ctrl_regs); |
|---|
| 1261 | + } |
|---|
| 1262 | + return 0; |
|---|
| 1263 | +} |
|---|
| 1264 | + |
|---|
| 1168 | 1265 | static int nixge_probe(struct platform_device *pdev) |
|---|
| 1169 | 1266 | { |
|---|
| 1267 | + struct device_node *mn, *phy_node; |
|---|
| 1170 | 1268 | struct nixge_priv *priv; |
|---|
| 1171 | 1269 | struct net_device *ndev; |
|---|
| 1172 | | - struct resource *dmares; |
|---|
| 1173 | 1270 | const u8 *mac_addr; |
|---|
| 1174 | 1271 | int err; |
|---|
| 1175 | 1272 | |
|---|
| .. | .. |
|---|
| 1201 | 1298 | priv->dev = &pdev->dev; |
|---|
| 1202 | 1299 | |
|---|
| 1203 | 1300 | netif_napi_add(ndev, &priv->napi, nixge_poll, NAPI_POLL_WEIGHT); |
|---|
| 1204 | | - |
|---|
| 1205 | | - dmares = platform_get_resource(pdev, IORESOURCE_MEM, 0); |
|---|
| 1206 | | - priv->dma_regs = devm_ioremap_resource(&pdev->dev, dmares); |
|---|
| 1207 | | - if (IS_ERR(priv->dma_regs)) { |
|---|
| 1208 | | - netdev_err(ndev, "failed to map dma regs\n"); |
|---|
| 1209 | | - return PTR_ERR(priv->dma_regs); |
|---|
| 1210 | | - } |
|---|
| 1211 | | - priv->ctrl_regs = priv->dma_regs + NIXGE_REG_CTRL_OFFSET; |
|---|
| 1301 | + err = nixge_of_get_resources(pdev); |
|---|
| 1302 | + if (err) |
|---|
| 1303 | + goto free_netdev; |
|---|
| 1212 | 1304 | __nixge_hw_set_mac_address(ndev); |
|---|
| 1213 | 1305 | |
|---|
| 1214 | 1306 | priv->tx_irq = platform_get_irq_byname(pdev, "tx"); |
|---|
| 1215 | 1307 | if (priv->tx_irq < 0) { |
|---|
| 1216 | 1308 | netdev_err(ndev, "could not find 'tx' irq"); |
|---|
| 1217 | | - return priv->tx_irq; |
|---|
| 1309 | + err = priv->tx_irq; |
|---|
| 1310 | + goto free_netdev; |
|---|
| 1218 | 1311 | } |
|---|
| 1219 | 1312 | |
|---|
| 1220 | 1313 | priv->rx_irq = platform_get_irq_byname(pdev, "rx"); |
|---|
| 1221 | 1314 | if (priv->rx_irq < 0) { |
|---|
| 1222 | 1315 | netdev_err(ndev, "could not find 'rx' irq"); |
|---|
| 1223 | | - return priv->rx_irq; |
|---|
| 1316 | + err = priv->rx_irq; |
|---|
| 1317 | + goto free_netdev; |
|---|
| 1224 | 1318 | } |
|---|
| 1225 | 1319 | |
|---|
| 1226 | 1320 | priv->coalesce_count_rx = XAXIDMA_DFT_RX_THRESHOLD; |
|---|
| 1227 | 1321 | priv->coalesce_count_tx = XAXIDMA_DFT_TX_THRESHOLD; |
|---|
| 1228 | 1322 | |
|---|
| 1229 | | - err = nixge_mdio_setup(priv, pdev->dev.of_node); |
|---|
| 1323 | + mn = of_get_child_by_name(pdev->dev.of_node, "mdio"); |
|---|
| 1324 | + if (mn) { |
|---|
| 1325 | + err = nixge_mdio_setup(priv, mn); |
|---|
| 1326 | + of_node_put(mn); |
|---|
| 1327 | + if (err) { |
|---|
| 1328 | + netdev_err(ndev, "error registering mdio bus"); |
|---|
| 1329 | + goto free_netdev; |
|---|
| 1330 | + } |
|---|
| 1331 | + } |
|---|
| 1332 | + |
|---|
| 1333 | + err = of_get_phy_mode(pdev->dev.of_node, &priv->phy_mode); |
|---|
| 1230 | 1334 | if (err) { |
|---|
| 1231 | | - netdev_err(ndev, "error registering mdio bus"); |
|---|
| 1232 | | - goto free_netdev; |
|---|
| 1233 | | - } |
|---|
| 1234 | | - |
|---|
| 1235 | | - priv->phy_mode = of_get_phy_mode(pdev->dev.of_node); |
|---|
| 1236 | | - if ((int)priv->phy_mode < 0) { |
|---|
| 1237 | 1335 | netdev_err(ndev, "not find \"phy-mode\" property\n"); |
|---|
| 1238 | | - err = -EINVAL; |
|---|
| 1239 | 1336 | goto unregister_mdio; |
|---|
| 1240 | 1337 | } |
|---|
| 1241 | 1338 | |
|---|
| 1242 | | - priv->phy_node = of_parse_phandle(pdev->dev.of_node, "phy-handle", 0); |
|---|
| 1243 | | - if (!priv->phy_node) { |
|---|
| 1244 | | - netdev_err(ndev, "not find \"phy-handle\" property\n"); |
|---|
| 1245 | | - err = -EINVAL; |
|---|
| 1246 | | - goto unregister_mdio; |
|---|
| 1339 | + phy_node = of_parse_phandle(pdev->dev.of_node, "phy-handle", 0); |
|---|
| 1340 | + if (!phy_node && of_phy_is_fixed_link(pdev->dev.of_node)) { |
|---|
| 1341 | + err = of_phy_register_fixed_link(pdev->dev.of_node); |
|---|
| 1342 | + if (err < 0) { |
|---|
| 1343 | + netdev_err(ndev, "broken fixed-link specification\n"); |
|---|
| 1344 | + goto unregister_mdio; |
|---|
| 1345 | + } |
|---|
| 1346 | + phy_node = of_node_get(pdev->dev.of_node); |
|---|
| 1247 | 1347 | } |
|---|
| 1348 | + priv->phy_node = phy_node; |
|---|
| 1248 | 1349 | |
|---|
| 1249 | 1350 | err = register_netdev(priv->ndev); |
|---|
| 1250 | 1351 | if (err) { |
|---|
| 1251 | 1352 | netdev_err(ndev, "register_netdev() error (%i)\n", err); |
|---|
| 1252 | | - goto unregister_mdio; |
|---|
| 1353 | + goto free_phy; |
|---|
| 1253 | 1354 | } |
|---|
| 1254 | 1355 | |
|---|
| 1255 | 1356 | return 0; |
|---|
| 1256 | 1357 | |
|---|
| 1358 | +free_phy: |
|---|
| 1359 | + if (of_phy_is_fixed_link(pdev->dev.of_node)) |
|---|
| 1360 | + of_phy_deregister_fixed_link(pdev->dev.of_node); |
|---|
| 1361 | + of_node_put(phy_node); |
|---|
| 1362 | + |
|---|
| 1257 | 1363 | unregister_mdio: |
|---|
| 1258 | | - mdiobus_unregister(priv->mii_bus); |
|---|
| 1364 | + if (priv->mii_bus) |
|---|
| 1365 | + mdiobus_unregister(priv->mii_bus); |
|---|
| 1259 | 1366 | |
|---|
| 1260 | 1367 | free_netdev: |
|---|
| 1261 | 1368 | free_netdev(ndev); |
|---|
| .. | .. |
|---|
| 1270 | 1377 | |
|---|
| 1271 | 1378 | unregister_netdev(ndev); |
|---|
| 1272 | 1379 | |
|---|
| 1273 | | - mdiobus_unregister(priv->mii_bus); |
|---|
| 1380 | + if (of_phy_is_fixed_link(pdev->dev.of_node)) |
|---|
| 1381 | + of_phy_deregister_fixed_link(pdev->dev.of_node); |
|---|
| 1382 | + of_node_put(priv->phy_node); |
|---|
| 1383 | + |
|---|
| 1384 | + if (priv->mii_bus) |
|---|
| 1385 | + mdiobus_unregister(priv->mii_bus); |
|---|
| 1274 | 1386 | |
|---|
| 1275 | 1387 | free_netdev(ndev); |
|---|
| 1276 | 1388 | |
|---|
| 1277 | 1389 | return 0; |
|---|
| 1278 | 1390 | } |
|---|
| 1279 | | - |
|---|
| 1280 | | -/* Match table for of_platform binding */ |
|---|
| 1281 | | -static const struct of_device_id nixge_dt_ids[] = { |
|---|
| 1282 | | - { .compatible = "ni,xge-enet-2.00", }, |
|---|
| 1283 | | - {}, |
|---|
| 1284 | | -}; |
|---|
| 1285 | | -MODULE_DEVICE_TABLE(of, nixge_dt_ids); |
|---|
| 1286 | 1391 | |
|---|
| 1287 | 1392 | static struct platform_driver nixge_driver = { |
|---|
| 1288 | 1393 | .probe = nixge_probe, |
|---|