1 /* 2 * CDDL HEADER START 3 * 4 * The contents of this file are subject to the terms of the 5 * Common Development and Distribution License (the "License"). 6 * You may not use this file except in compliance with the License. 7 * 8 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE 9 * or http://www.opensolaris.org/os/licensing. 10 * See the License for the specific language governing permissions 11 * and limitations under the License. 12 * 13 * When distributing Covered Code, include this CDDL HEADER in each 14 * file and include the License file at usr/src/OPENSOLARIS.LICENSE. 15 * If applicable, add the following below this CDDL HEADER, with the 16 * fields enclosed by brackets "[]" replaced with your own identifying 17 * information: Portions Copyright [yyyy] [name of copyright owner] 18 * 19 * CDDL HEADER END 20 */ 21 22 /* 23 * Copyright (c) 2009, 2010, Oracle and/or its affiliates. All rights reserved. 24 */ 25 #include <sys/stream.h> 26 #include <sys/strsun.h> 27 #include <sys/stat.h> 28 #include <sys/pci.h> 29 #include <sys/modctl.h> 30 #include <sys/kstat.h> 31 #include <sys/ethernet.h> 32 #include <sys/devops.h> 33 #include <sys/debug.h> 34 #include <sys/conf.h> 35 #include <sys/sysmacros.h> 36 #include <sys/dditypes.h> 37 #include <sys/ddi.h> 38 #include <sys/sunddi.h> 39 #include <sys/miiregs.h> 40 #include <sys/byteorder.h> 41 #include <sys/cyclic.h> 42 #include <sys/note.h> 43 #include <sys/crc32.h> 44 #include <sys/mac_provider.h> 45 #include <sys/mac_ether.h> 46 #include <sys/vlan.h> 47 #include <sys/errno.h> 48 #include <sys/sdt.h> 49 #include <sys/strsubr.h> 50 51 #include "bfe.h" 52 #include "bfe_hw.h" 53 54 55 /* 56 * Broadcom BCM4401 chipsets use two rings : 57 * 58 * - One TX : For sending packets down the wire. 59 * - One RX : For receving packets. 60 * 61 * Each ring can have any number of descriptors (configured during attach). 62 * As of now we configure only 128 descriptor per ring (TX/RX). Each descriptor 63 * has address (desc_addr) and control (desc_ctl) which holds a DMA buffer for 64 * the packet and control information (like start/end of frame or end of table). 65 * The descriptor table is allocated first and then a DMA buffer (for a packet) 66 * is allocated and linked to each descriptor. 67 * 68 * Each descriptor entry is bfe_desc_t structure in bfe. During TX/RX 69 * interrupt, the stat register will point to current descriptor being 70 * processed. 71 * 72 * Here's an example of TX and RX ring : 73 * 74 * TX: 75 * 76 * Base of the descriptor table is programmed using BFE_DMATX_CTRL control 77 * register. Each 'addr' points to DMA buffer (or packet data buffer) to 78 * be transmitted and 'ctl' has the length of the packet (usually MTU). 79 * 80 * ----------------------| 81 * | addr |Descriptor 0 | 82 * | ctl | | 83 * ----------------------| 84 * | addr |Descriptor 1 | SOF (start of the frame) 85 * | ctl | | 86 * ----------------------| 87 * | ... |Descriptor... | EOF (end of the frame) 88 * | ... | | 89 * ----------------------| 90 * | addr |Descritor 127 | 91 * | ctl | EOT | EOT (End of Table) 92 * ----------------------| 93 * 94 * 'r_curr_desc' : pointer to current descriptor which can be used to transmit 95 * a packet. 96 * 'r_avail_desc' : decremented whenever a packet is being sent. 97 * 'r_cons_desc' : incremented whenever a packet is sent down the wire and 98 * notified by an interrupt to bfe driver. 99 * 100 * RX: 101 * 102 * Base of the descriptor table is programmed using BFE_DMARX_CTRL control 103 * register. Each 'addr' points to DMA buffer (or packet data buffer). 'ctl' 104 * contains the size of the DMA buffer and all the DMA buffers are 105 * pre-allocated during attach and hence the maxmium size of the packet is 106 * also known (r_buf_len from the bfe_rint_t structure). During RX interrupt 107 * the packet length is embedded in bfe_header_t which is added by the 108 * chip in the beginning of the packet. 109 * 110 * ----------------------| 111 * | addr |Descriptor 0 | 112 * | ctl | | 113 * ----------------------| 114 * | addr |Descriptor 1 | 115 * | ctl | | 116 * ----------------------| 117 * | ... |Descriptor... | 118 * | ... | | 119 * ----------------------| 120 * | addr |Descriptor 127| 121 * | ctl | EOT | EOT (End of Table) 122 * ----------------------| 123 * 124 * 'r_curr_desc' : pointer to current descriptor while receving a packet. 125 * 126 */ 127 128 #define MODULE_NAME "bfe" 129 130 /* 131 * Used for checking PHY (link state, speed) 132 */ 133 #define BFE_TIMEOUT_INTERVAL (1000 * 1000 * 1000) 134 135 136 /* 137 * Chip restart action and reason for restart 138 */ 139 #define BFE_ACTION_RESTART 0x1 /* For restarting the chip */ 140 #define BFE_ACTION_RESTART_SETPROP 0x2 /* restart due to setprop */ 141 #define BFE_ACTION_RESTART_FAULT 0x4 /* restart due to fault */ 142 #define BFE_ACTION_RESTART_PKT 0x8 /* restart due to pkt timeout */ 143 144 static char bfe_ident[] = "bfe driver for Broadcom BCM4401 chipsets"; 145 146 /* 147 * Function Prototypes for bfe driver. 148 */ 149 static int bfe_check_link(bfe_t *); 150 static void bfe_report_link(bfe_t *); 151 static void bfe_chip_halt(bfe_t *); 152 static void bfe_chip_reset(bfe_t *); 153 static void bfe_tx_desc_init(bfe_ring_t *); 154 static void bfe_rx_desc_init(bfe_ring_t *); 155 static void bfe_set_rx_mode(bfe_t *); 156 static void bfe_enable_chip_intrs(bfe_t *); 157 static void bfe_chip_restart(bfe_t *); 158 static void bfe_init_vars(bfe_t *); 159 static void bfe_clear_stats(bfe_t *); 160 static void bfe_gather_stats(bfe_t *); 161 static void bfe_error(dev_info_t *, char *, ...); 162 static int bfe_mac_getprop(void *, const char *, mac_prop_id_t, uint_t, 163 void *); 164 static int bfe_mac_setprop(void *, const char *, mac_prop_id_t, uint_t, 165 const void *); 166 static int bfe_tx_reclaim(bfe_ring_t *); 167 int bfe_mac_set_ether_addr(void *, const uint8_t *); 168 169 170 /* 171 * Macros for ddi_dma_sync(). 172 */ 173 #define SYNC_DESC(r, s, l, d) \ 174 (void) ddi_dma_sync(r->r_desc_dma_handle, \ 175 (off_t)(s * sizeof (bfe_desc_t)), \ 176 (size_t)(l * sizeof (bfe_desc_t)), \ 177 d) 178 179 #define SYNC_BUF(r, s, b, l, d) \ 180 (void) ddi_dma_sync(r->r_buf_dma[s].handle, \ 181 (off_t)(b), (size_t)(l), d) 182 183 /* 184 * Supported Broadcom BCM4401 Cards. 185 */ 186 static bfe_cards_t bfe_cards[] = { 187 { 0x14e4, 0x170c, "BCM4401 100Base-TX"}, 188 }; 189 190 191 /* 192 * DMA attributes for device registers, packet data (buffer) and 193 * descriptor table. 194 */ 195 static struct ddi_device_acc_attr bfe_dev_attr = { 196 DDI_DEVICE_ATTR_V0, 197 DDI_STRUCTURE_LE_ACC, 198 DDI_STRICTORDER_ACC 199 }; 200 201 static struct ddi_device_acc_attr bfe_buf_attr = { 202 DDI_DEVICE_ATTR_V0, 203 DDI_NEVERSWAP_ACC, /* native endianness */ 204 DDI_STRICTORDER_ACC 205 }; 206 207 static ddi_dma_attr_t bfe_dma_attr_buf = { 208 DMA_ATTR_V0, /* dma_attr_version */ 209 0, /* dma_attr_addr_lo */ 210 BFE_PCI_DMA - 1, /* dma_attr_addr_hi */ 211 0x1fff, /* dma_attr_count_max */ 212 8, /* dma_attr_align */ 213 0, /* dma_attr_burstsizes */ 214 1, /* dma_attr_minxfer */ 215 0x1fff, /* dma_attr_maxxfer */ 216 BFE_PCI_DMA - 1, /* dma_attr_seg */ 217 1, /* dma_attr_sgllen */ 218 1, /* dma_attr_granular */ 219 0 /* dma_attr_flags */ 220 }; 221 222 static ddi_dma_attr_t bfe_dma_attr_desc = { 223 DMA_ATTR_V0, /* dma_attr_version */ 224 0, /* dma_attr_addr_lo */ 225 BFE_PCI_DMA - 1, /* dma_attr_addr_hi */ 226 BFE_PCI_DMA - 1, /* dma_attr_count_max */ 227 BFE_DESC_ALIGN, /* dma_attr_align */ 228 0, /* dma_attr_burstsizes */ 229 1, /* dma_attr_minxfer */ 230 BFE_PCI_DMA - 1, /* dma_attr_maxxfer */ 231 BFE_PCI_DMA - 1, /* dma_attr_seg */ 232 1, /* dma_attr_sgllen */ 233 1, /* dma_attr_granular */ 234 0 /* dma_attr_flags */ 235 }; 236 237 /* 238 * Ethernet broadcast addresses. 239 */ 240 static uchar_t bfe_broadcast[ETHERADDRL] = { 241 0xff, 0xff, 0xff, 0xff, 0xff, 0xff 242 }; 243 244 #define ASSERT_ALL_LOCKS(bfe) { \ 245 ASSERT(mutex_owned(&bfe->bfe_tx_ring.r_lock)); \ 246 ASSERT(rw_write_held(&bfe->bfe_rwlock)); \ 247 } 248 249 /* 250 * Debugging and error reproting code. 251 */ 252 static void 253 bfe_error(dev_info_t *dip, char *fmt, ...) 254 { 255 va_list ap; 256 char buf[256]; 257 258 va_start(ap, fmt); 259 (void) vsnprintf(buf, sizeof (buf), fmt, ap); 260 va_end(ap); 261 262 if (dip) { 263 cmn_err(CE_WARN, "%s%d: %s", 264 ddi_driver_name(dip), ddi_get_instance(dip), buf); 265 } else { 266 cmn_err(CE_WARN, "bfe: %s", buf); 267 } 268 } 269 270 /* 271 * Grabs all necessary locks to block any other operation on the chip. 272 */ 273 static void 274 bfe_grab_locks(bfe_t *bfe) 275 { 276 bfe_ring_t *tx = &bfe->bfe_tx_ring; 277 278 /* 279 * Grab all the locks. 280 * - bfe_rwlock : locks down whole chip including RX. 281 * - tx's r_lock : locks down only TX side. 282 */ 283 rw_enter(&bfe->bfe_rwlock, RW_WRITER); 284 mutex_enter(&tx->r_lock); 285 286 /* 287 * Note that we don't use RX's r_lock. 288 */ 289 } 290 291 /* 292 * Release lock on chip/drver. 293 */ 294 static void 295 bfe_release_locks(bfe_t *bfe) 296 { 297 bfe_ring_t *tx = &bfe->bfe_tx_ring; 298 299 /* 300 * Release all the locks in the order in which they were grabbed. 301 */ 302 mutex_exit(&tx->r_lock); 303 rw_exit(&bfe->bfe_rwlock); 304 } 305 306 307 /* 308 * It's used to make sure that the write to device register was successful. 309 */ 310 static int 311 bfe_wait_bit(bfe_t *bfe, uint32_t reg, uint32_t bit, 312 ulong_t t, const int clear) 313 { 314 ulong_t i; 315 uint32_t v; 316 317 for (i = 0; i < t; i++) { 318 v = INL(bfe, reg); 319 320 if (clear && !(v & bit)) 321 break; 322 323 if (!clear && (v & bit)) 324 break; 325 326 drv_usecwait(10); 327 } 328 329 /* if device still didn't see the value */ 330 if (i == t) 331 return (-1); 332 333 return (0); 334 } 335 336 /* 337 * PHY functions (read, write, stop, reset and startup) 338 */ 339 static int 340 bfe_read_phy(bfe_t *bfe, uint32_t reg) 341 { 342 OUTL(bfe, BFE_EMAC_ISTAT, BFE_EMAC_INT_MII); 343 OUTL(bfe, BFE_MDIO_DATA, (BFE_MDIO_SB_START | 344 (BFE_MDIO_OP_READ << BFE_MDIO_OP_SHIFT) | 345 (bfe->bfe_phy_addr << BFE_MDIO_PMD_SHIFT) | 346 (reg << BFE_MDIO_RA_SHIFT) | 347 (BFE_MDIO_TA_VALID << BFE_MDIO_TA_SHIFT))); 348 349 (void) bfe_wait_bit(bfe, BFE_EMAC_ISTAT, BFE_EMAC_INT_MII, 10, 0); 350 351 return ((INL(bfe, BFE_MDIO_DATA) & BFE_MDIO_DATA_DATA)); 352 } 353 354 static void 355 bfe_write_phy(bfe_t *bfe, uint32_t reg, uint32_t val) 356 { 357 OUTL(bfe, BFE_EMAC_ISTAT, BFE_EMAC_INT_MII); 358 OUTL(bfe, BFE_MDIO_DATA, (BFE_MDIO_SB_START | 359 (BFE_MDIO_OP_WRITE << BFE_MDIO_OP_SHIFT) | 360 (bfe->bfe_phy_addr << BFE_MDIO_PMD_SHIFT) | 361 (reg << BFE_MDIO_RA_SHIFT) | 362 (BFE_MDIO_TA_VALID << BFE_MDIO_TA_SHIFT) | 363 (val & BFE_MDIO_DATA_DATA))); 364 365 (void) bfe_wait_bit(bfe, BFE_EMAC_ISTAT, BFE_EMAC_INT_MII, 10, 0); 366 } 367 368 /* 369 * It resets the PHY layer. 370 */ 371 static int 372 bfe_reset_phy(bfe_t *bfe) 373 { 374 uint32_t i; 375 376 bfe_write_phy(bfe, MII_CONTROL, MII_CONTROL_RESET); 377 drv_usecwait(100); 378 for (i = 0; i < 10; i++) { 379 if (bfe_read_phy(bfe, MII_CONTROL) & 380 MII_CONTROL_RESET) { 381 drv_usecwait(500); 382 continue; 383 } 384 385 break; 386 } 387 388 if (i == 10) { 389 bfe_error(bfe->bfe_dip, "Timeout waiting for PHY to reset"); 390 bfe->bfe_phy_state = BFE_PHY_RESET_TIMEOUT; 391 return (BFE_FAILURE); 392 } 393 394 bfe->bfe_phy_state = BFE_PHY_RESET_DONE; 395 396 return (BFE_SUCCESS); 397 } 398 399 /* 400 * Make sure timer function is out of our way and especially during 401 * detach. 402 */ 403 static void 404 bfe_stop_timer(bfe_t *bfe) 405 { 406 if (bfe->bfe_periodic_id) { 407 ddi_periodic_delete(bfe->bfe_periodic_id); 408 bfe->bfe_periodic_id = NULL; 409 } 410 } 411 412 /* 413 * Stops the PHY 414 */ 415 static void 416 bfe_stop_phy(bfe_t *bfe) 417 { 418 bfe_write_phy(bfe, MII_CONTROL, MII_CONTROL_PWRDN | 419 MII_CONTROL_ISOLATE); 420 421 bfe->bfe_chip.link = LINK_STATE_UNKNOWN; 422 bfe->bfe_chip.speed = 0; 423 bfe->bfe_chip.duplex = LINK_DUPLEX_UNKNOWN; 424 425 bfe->bfe_phy_state = BFE_PHY_STOPPED; 426 427 /* 428 * Report the link status to MAC layer. 429 */ 430 if (bfe->bfe_machdl != NULL) 431 (void) bfe_report_link(bfe); 432 } 433 434 static int 435 bfe_probe_phy(bfe_t *bfe) 436 { 437 int phy; 438 uint32_t status; 439 440 if (bfe->bfe_phy_addr) { 441 status = bfe_read_phy(bfe, MII_STATUS); 442 if (status != 0xffff && status != 0) { 443 bfe_write_phy(bfe, MII_CONTROL, 0); 444 return (BFE_SUCCESS); 445 } 446 } 447 448 for (phy = 0; phy < 32; phy++) { 449 bfe->bfe_phy_addr = phy; 450 status = bfe_read_phy(bfe, MII_STATUS); 451 if (status != 0xffff && status != 0) { 452 bfe_write_phy(bfe, MII_CONTROL, 0); 453 return (BFE_SUCCESS); 454 } 455 } 456 457 return (BFE_FAILURE); 458 } 459 460 /* 461 * This timeout function fires at BFE_TIMEOUT_INTERVAL to check the link 462 * status. 463 */ 464 static void 465 bfe_timeout(void *arg) 466 { 467 bfe_t *bfe = (bfe_t *)arg; 468 int resched = 0; 469 470 /* 471 * We don't grab any lock because bfe can't go away. 472 * untimeout() will wait for this timeout instance to complete. 473 */ 474 if (bfe->bfe_chip_action & BFE_ACTION_RESTART) { 475 /* 476 * Restart the chip. 477 */ 478 bfe_grab_locks(bfe); 479 bfe_chip_restart(bfe); 480 bfe->bfe_chip_action &= ~BFE_ACTION_RESTART; 481 bfe->bfe_chip_action &= ~BFE_ACTION_RESTART_FAULT; 482 bfe->bfe_chip_action &= ~BFE_ACTION_RESTART_PKT; 483 bfe_release_locks(bfe); 484 mac_tx_update(bfe->bfe_machdl); 485 /* Restart will register a new timeout */ 486 return; 487 } 488 489 rw_enter(&bfe->bfe_rwlock, RW_READER); 490 491 if (bfe->bfe_chip_state == BFE_CHIP_ACTIVE) { 492 hrtime_t hr; 493 494 hr = gethrtime(); 495 if (bfe->bfe_tx_stall_time != 0 && 496 hr > bfe->bfe_tx_stall_time) { 497 DTRACE_PROBE2(chip__restart, int, bfe->bfe_unit, 498 char *, "pkt timeout"); 499 bfe->bfe_chip_action |= 500 (BFE_ACTION_RESTART | BFE_ACTION_RESTART_PKT); 501 bfe->bfe_tx_stall_time = 0; 502 } 503 } 504 505 if (bfe->bfe_phy_state == BFE_PHY_STARTED) { 506 /* 507 * Report the link status to MAC layer if link status changed. 508 */ 509 if (bfe_check_link(bfe)) { 510 bfe_report_link(bfe); 511 if (bfe->bfe_chip.link == LINK_STATE_UP) { 512 uint32_t val, flow; 513 514 val = INL(bfe, BFE_TX_CTRL); 515 val &= ~BFE_TX_DUPLEX; 516 if (bfe->bfe_chip.duplex == LINK_DUPLEX_FULL) { 517 val |= BFE_TX_DUPLEX; 518 flow = INL(bfe, BFE_RXCONF); 519 flow &= ~BFE_RXCONF_FLOW; 520 OUTL(bfe, BFE_RXCONF, flow); 521 522 flow = INL(bfe, BFE_MAC_FLOW); 523 flow &= ~(BFE_FLOW_RX_HIWAT); 524 OUTL(bfe, BFE_MAC_FLOW, flow); 525 } 526 527 resched = 1; 528 529 OUTL(bfe, BFE_TX_CTRL, val); 530 DTRACE_PROBE1(link__up, 531 int, bfe->bfe_unit); 532 } 533 } 534 } 535 536 rw_exit(&bfe->bfe_rwlock); 537 538 if (resched) 539 mac_tx_update(bfe->bfe_machdl); 540 } 541 542 /* 543 * Starts PHY layer. 544 */ 545 static int 546 bfe_startup_phy(bfe_t *bfe) 547 { 548 uint16_t bmsr, bmcr, anar; 549 int prog, s; 550 int phyid1, phyid2; 551 552 if (bfe_probe_phy(bfe) == BFE_FAILURE) { 553 bfe->bfe_phy_state = BFE_PHY_NOTFOUND; 554 return (BFE_FAILURE); 555 } 556 557 (void) bfe_reset_phy(bfe); 558 559 phyid1 = bfe_read_phy(bfe, MII_PHYIDH); 560 phyid2 = bfe_read_phy(bfe, MII_PHYIDL); 561 bfe->bfe_phy_id = (phyid1 << 16) | phyid2; 562 563 bmsr = bfe_read_phy(bfe, MII_STATUS); 564 anar = bfe_read_phy(bfe, MII_AN_ADVERT); 565 566 again: 567 anar &= ~(MII_ABILITY_100BASE_T4 | 568 MII_ABILITY_100BASE_TX_FD | MII_ABILITY_100BASE_TX | 569 MII_ABILITY_10BASE_T_FD | MII_ABILITY_10BASE_T); 570 571 /* 572 * Supported hardware modes are in bmsr. 573 */ 574 bfe->bfe_chip.bmsr = bmsr; 575 576 /* 577 * Assume no capabilities are supported in the hardware. 578 */ 579 bfe->bfe_cap_aneg = bfe->bfe_cap_100T4 = 580 bfe->bfe_cap_100fdx = bfe->bfe_cap_100hdx = 581 bfe->bfe_cap_10fdx = bfe->bfe_cap_10hdx = 0; 582 583 /* 584 * Assume property is set. 585 */ 586 s = 1; 587 if (!(bfe->bfe_chip_action & BFE_ACTION_RESTART_SETPROP)) { 588 /* 589 * Property is not set which means bfe_mac_setprop() 590 * is not called on us. 591 */ 592 s = 0; 593 } 594 595 bmcr = prog = 0; 596 597 if (bmsr & MII_STATUS_100_BASEX_FD) { 598 bfe->bfe_cap_100fdx = 1; 599 if (s == 0) { 600 anar |= MII_ABILITY_100BASE_TX_FD; 601 bfe->bfe_adv_100fdx = 1; 602 prog++; 603 } else if (bfe->bfe_adv_100fdx) { 604 anar |= MII_ABILITY_100BASE_TX_FD; 605 prog++; 606 } 607 } 608 609 if (bmsr & MII_STATUS_100_BASE_T4) { 610 bfe->bfe_cap_100T4 = 1; 611 if (s == 0) { 612 anar |= MII_ABILITY_100BASE_T4; 613 bfe->bfe_adv_100T4 = 1; 614 prog++; 615 } else if (bfe->bfe_adv_100T4) { 616 anar |= MII_ABILITY_100BASE_T4; 617 prog++; 618 } 619 } 620 621 if (bmsr & MII_STATUS_100_BASEX) { 622 bfe->bfe_cap_100hdx = 1; 623 if (s == 0) { 624 anar |= MII_ABILITY_100BASE_TX; 625 bfe->bfe_adv_100hdx = 1; 626 prog++; 627 } else if (bfe->bfe_adv_100hdx) { 628 anar |= MII_ABILITY_100BASE_TX; 629 prog++; 630 } 631 } 632 633 if (bmsr & MII_STATUS_10_FD) { 634 bfe->bfe_cap_10fdx = 1; 635 if (s == 0) { 636 anar |= MII_ABILITY_10BASE_T_FD; 637 bfe->bfe_adv_10fdx = 1; 638 prog++; 639 } else if (bfe->bfe_adv_10fdx) { 640 anar |= MII_ABILITY_10BASE_T_FD; 641 prog++; 642 } 643 } 644 645 if (bmsr & MII_STATUS_10) { 646 bfe->bfe_cap_10hdx = 1; 647 if (s == 0) { 648 anar |= MII_ABILITY_10BASE_T; 649 bfe->bfe_adv_10hdx = 1; 650 prog++; 651 } else if (bfe->bfe_adv_10hdx) { 652 anar |= MII_ABILITY_10BASE_T; 653 prog++; 654 } 655 } 656 657 if (bmsr & MII_STATUS_CANAUTONEG) { 658 bfe->bfe_cap_aneg = 1; 659 if (s == 0) { 660 bfe->bfe_adv_aneg = 1; 661 } 662 } 663 664 if (prog == 0) { 665 if (s == 0) { 666 bfe_error(bfe->bfe_dip, 667 "No valid link mode selected. Powering down PHY"); 668 bfe_stop_phy(bfe); 669 bfe_report_link(bfe); 670 return (BFE_FAILURE); 671 } 672 673 /* 674 * If property is set then user would have goofed up. So we 675 * go back to default properties. 676 */ 677 bfe->bfe_chip_action &= ~BFE_ACTION_RESTART_SETPROP; 678 goto again; 679 } 680 681 if (bfe->bfe_adv_aneg && (bmsr & MII_STATUS_CANAUTONEG)) { 682 bmcr = (MII_CONTROL_ANE | MII_CONTROL_RSAN); 683 } else { 684 if (bfe->bfe_adv_100fdx) 685 bmcr = (MII_CONTROL_100MB | MII_CONTROL_FDUPLEX); 686 else if (bfe->bfe_adv_100hdx) 687 bmcr = MII_CONTROL_100MB; 688 else if (bfe->bfe_adv_10fdx) 689 bmcr = MII_CONTROL_FDUPLEX; 690 else 691 bmcr = 0; /* 10HDX */ 692 } 693 694 if (prog) 695 bfe_write_phy(bfe, MII_AN_ADVERT, anar); 696 697 if (bmcr) 698 bfe_write_phy(bfe, MII_CONTROL, bmcr); 699 700 bfe->bfe_mii_anar = anar; 701 bfe->bfe_mii_bmcr = bmcr; 702 bfe->bfe_phy_state = BFE_PHY_STARTED; 703 704 if (bfe->bfe_periodic_id == NULL) { 705 bfe->bfe_periodic_id = ddi_periodic_add(bfe_timeout, 706 (void *)bfe, BFE_TIMEOUT_INTERVAL, DDI_IPL_0); 707 708 DTRACE_PROBE1(first__timeout, int, bfe->bfe_unit); 709 } 710 711 DTRACE_PROBE4(phy_started, int, bfe->bfe_unit, 712 int, bmsr, int, bmcr, int, anar); 713 714 return (BFE_SUCCESS); 715 } 716 717 /* 718 * Reports link status back to MAC Layer. 719 */ 720 static void 721 bfe_report_link(bfe_t *bfe) 722 { 723 mac_link_update(bfe->bfe_machdl, bfe->bfe_chip.link); 724 } 725 726 /* 727 * Reads PHY/MII registers and get the link status for us. 728 */ 729 static int 730 bfe_check_link(bfe_t *bfe) 731 { 732 uint16_t bmsr, bmcr, anar, anlpar; 733 int speed, duplex, link; 734 735 speed = bfe->bfe_chip.speed; 736 duplex = bfe->bfe_chip.duplex; 737 link = bfe->bfe_chip.link; 738 739 bmsr = bfe_read_phy(bfe, MII_STATUS); 740 bfe->bfe_mii_bmsr = bmsr; 741 742 bmcr = bfe_read_phy(bfe, MII_CONTROL); 743 744 anar = bfe_read_phy(bfe, MII_AN_ADVERT); 745 bfe->bfe_mii_anar = anar; 746 747 anlpar = bfe_read_phy(bfe, MII_AN_LPABLE); 748 bfe->bfe_mii_anlpar = anlpar; 749 750 bfe->bfe_mii_exp = bfe_read_phy(bfe, MII_AN_EXPANSION); 751 752 /* 753 * If exp register is not present in PHY. 754 */ 755 if (bfe->bfe_mii_exp == 0xffff) { 756 bfe->bfe_mii_exp = 0; 757 } 758 759 if ((bmsr & MII_STATUS_LINKUP) == 0) { 760 bfe->bfe_chip.link = LINK_STATE_DOWN; 761 bfe->bfe_chip.speed = 0; 762 bfe->bfe_chip.duplex = LINK_DUPLEX_UNKNOWN; 763 goto done; 764 } 765 766 bfe->bfe_chip.link = LINK_STATE_UP; 767 768 if (!(bmcr & MII_CONTROL_ANE)) { 769 /* Forced mode */ 770 if (bmcr & MII_CONTROL_100MB) 771 bfe->bfe_chip.speed = 100000000; 772 else 773 bfe->bfe_chip.speed = 10000000; 774 775 if (bmcr & MII_CONTROL_FDUPLEX) 776 bfe->bfe_chip.duplex = LINK_DUPLEX_FULL; 777 else 778 bfe->bfe_chip.duplex = LINK_DUPLEX_HALF; 779 780 } else if ((!(bmsr & MII_STATUS_CANAUTONEG)) || 781 (!(bmsr & MII_STATUS_ANDONE))) { 782 bfe->bfe_chip.speed = 0; 783 bfe->bfe_chip.duplex = LINK_DUPLEX_UNKNOWN; 784 } else if (anar & anlpar & MII_ABILITY_100BASE_TX_FD) { 785 bfe->bfe_chip.speed = 100000000; 786 bfe->bfe_chip.duplex = LINK_DUPLEX_FULL; 787 } else if (anar & anlpar & MII_ABILITY_100BASE_T4) { 788 bfe->bfe_chip.speed = 100000000; 789 bfe->bfe_chip.duplex = LINK_DUPLEX_HALF; 790 } else if (anar & anlpar & MII_ABILITY_100BASE_TX) { 791 bfe->bfe_chip.speed = 100000000; 792 bfe->bfe_chip.duplex = LINK_DUPLEX_HALF; 793 } else if (anar & anlpar & MII_ABILITY_10BASE_T_FD) { 794 bfe->bfe_chip.speed = 10000000; 795 bfe->bfe_chip.duplex = LINK_DUPLEX_FULL; 796 } else if (anar & anlpar & MII_ABILITY_10BASE_T) { 797 bfe->bfe_chip.speed = 10000000; 798 bfe->bfe_chip.duplex = LINK_DUPLEX_HALF; 799 } else { 800 bfe->bfe_chip.speed = 0; 801 bfe->bfe_chip.duplex = LINK_DUPLEX_UNKNOWN; 802 } 803 804 done: 805 /* 806 * If speed or link status or duplex mode changed then report to 807 * MAC layer which is done by the caller. 808 */ 809 if (speed != bfe->bfe_chip.speed || 810 duplex != bfe->bfe_chip.duplex || 811 link != bfe->bfe_chip.link) { 812 return (1); 813 } 814 815 return (0); 816 } 817 818 static void 819 bfe_cam_write(bfe_t *bfe, uchar_t *d, int index) 820 { 821 uint32_t v; 822 823 v = ((uint32_t)d[2] << 24); 824 v |= ((uint32_t)d[3] << 16); 825 v |= ((uint32_t)d[4] << 8); 826 v |= (uint32_t)d[5]; 827 828 OUTL(bfe, BFE_CAM_DATA_LO, v); 829 v = (BFE_CAM_HI_VALID | 830 (((uint32_t)d[0]) << 8) | 831 (((uint32_t)d[1]))); 832 833 OUTL(bfe, BFE_CAM_DATA_HI, v); 834 OUTL(bfe, BFE_CAM_CTRL, (BFE_CAM_WRITE | 835 ((uint32_t)index << BFE_CAM_INDEX_SHIFT))); 836 (void) bfe_wait_bit(bfe, BFE_CAM_CTRL, BFE_CAM_BUSY, 10, 1); 837 } 838 839 /* 840 * Chip related functions (halt, reset, start). 841 */ 842 static void 843 bfe_chip_halt(bfe_t *bfe) 844 { 845 /* 846 * Disables interrupts. 847 */ 848 OUTL(bfe, BFE_INTR_MASK, 0); 849 FLUSH(bfe, BFE_INTR_MASK); 850 851 OUTL(bfe, BFE_ENET_CTRL, BFE_ENET_DISABLE); 852 853 /* 854 * Wait until TX and RX finish their job. 855 */ 856 (void) bfe_wait_bit(bfe, BFE_ENET_CTRL, BFE_ENET_DISABLE, 20, 1); 857 858 /* 859 * Disables DMA engine. 860 */ 861 OUTL(bfe, BFE_DMARX_CTRL, 0); 862 OUTL(bfe, BFE_DMATX_CTRL, 0); 863 864 drv_usecwait(10); 865 866 bfe->bfe_chip_state = BFE_CHIP_HALT; 867 } 868 869 static void 870 bfe_chip_restart(bfe_t *bfe) 871 { 872 DTRACE_PROBE2(chip__restart, int, bfe->bfe_unit, 873 int, bfe->bfe_chip_action); 874 875 /* 876 * Halt chip and PHY. 877 */ 878 bfe_chip_halt(bfe); 879 bfe_stop_phy(bfe); 880 bfe->bfe_chip_state = BFE_CHIP_STOPPED; 881 882 /* 883 * Init variables. 884 */ 885 bfe_init_vars(bfe); 886 887 /* 888 * Reset chip and start PHY. 889 */ 890 bfe_chip_reset(bfe); 891 892 /* 893 * DMA descriptor rings. 894 */ 895 bfe_tx_desc_init(&bfe->bfe_tx_ring); 896 bfe_rx_desc_init(&bfe->bfe_rx_ring); 897 898 bfe->bfe_chip_state = BFE_CHIP_ACTIVE; 899 bfe_set_rx_mode(bfe); 900 bfe_enable_chip_intrs(bfe); 901 } 902 903 /* 904 * Disables core by stopping the clock. 905 */ 906 static void 907 bfe_core_disable(bfe_t *bfe) 908 { 909 if ((INL(bfe, BFE_SBTMSLOW) & BFE_RESET)) 910 return; 911 912 OUTL(bfe, BFE_SBTMSLOW, (BFE_REJECT | BFE_CLOCK)); 913 (void) bfe_wait_bit(bfe, BFE_SBTMSLOW, BFE_REJECT, 100, 0); 914 (void) bfe_wait_bit(bfe, BFE_SBTMSHIGH, BFE_BUSY, 100, 1); 915 OUTL(bfe, BFE_SBTMSLOW, (BFE_FGC | BFE_CLOCK | BFE_REJECT | BFE_RESET)); 916 FLUSH(bfe, BFE_SBTMSLOW); 917 drv_usecwait(10); 918 OUTL(bfe, BFE_SBTMSLOW, (BFE_REJECT | BFE_RESET)); 919 drv_usecwait(10); 920 } 921 922 /* 923 * Resets core. 924 */ 925 static void 926 bfe_core_reset(bfe_t *bfe) 927 { 928 uint32_t val; 929 930 /* 931 * First disable the core. 932 */ 933 bfe_core_disable(bfe); 934 935 OUTL(bfe, BFE_SBTMSLOW, (BFE_RESET | BFE_CLOCK | BFE_FGC)); 936 FLUSH(bfe, BFE_SBTMSLOW); 937 drv_usecwait(1); 938 939 if (INL(bfe, BFE_SBTMSHIGH) & BFE_SERR) 940 OUTL(bfe, BFE_SBTMSHIGH, 0); 941 942 val = INL(bfe, BFE_SBIMSTATE); 943 if (val & (BFE_IBE | BFE_TO)) 944 OUTL(bfe, BFE_SBIMSTATE, val & ~(BFE_IBE | BFE_TO)); 945 946 OUTL(bfe, BFE_SBTMSLOW, (BFE_CLOCK | BFE_FGC)); 947 FLUSH(bfe, BFE_SBTMSLOW); 948 drv_usecwait(1); 949 950 OUTL(bfe, BFE_SBTMSLOW, BFE_CLOCK); 951 FLUSH(bfe, BFE_SBTMSLOW); 952 drv_usecwait(1); 953 } 954 955 static void 956 bfe_setup_config(bfe_t *bfe, uint32_t cores) 957 { 958 uint32_t bar_orig, val; 959 960 /* 961 * Change bar0 window to map sbtopci registers. 962 */ 963 bar_orig = pci_config_get32(bfe->bfe_conf_handle, BFE_BAR0_WIN); 964 pci_config_put32(bfe->bfe_conf_handle, BFE_BAR0_WIN, BFE_REG_PCI); 965 966 /* Just read it and don't do anything */ 967 val = INL(bfe, BFE_SBIDHIGH) & BFE_IDH_CORE; 968 969 val = INL(bfe, BFE_SBINTVEC); 970 val |= cores; 971 OUTL(bfe, BFE_SBINTVEC, val); 972 973 val = INL(bfe, BFE_SSB_PCI_TRANS_2); 974 val |= BFE_SSB_PCI_PREF | BFE_SSB_PCI_BURST; 975 OUTL(bfe, BFE_SSB_PCI_TRANS_2, val); 976 977 /* 978 * Restore bar0 window mapping. 979 */ 980 pci_config_put32(bfe->bfe_conf_handle, BFE_BAR0_WIN, bar_orig); 981 } 982 983 /* 984 * Resets chip and starts PHY. 985 */ 986 static void 987 bfe_chip_reset(bfe_t *bfe) 988 { 989 uint32_t val; 990 991 /* Set the interrupt vector for the enet core */ 992 bfe_setup_config(bfe, BFE_INTVEC_ENET0); 993 994 /* check if core is up */ 995 val = INL(bfe, BFE_SBTMSLOW) & 996 (BFE_RESET | BFE_REJECT | BFE_CLOCK); 997 998 if (val == BFE_CLOCK) { 999 OUTL(bfe, BFE_RCV_LAZY, 0); 1000 OUTL(bfe, BFE_ENET_CTRL, BFE_ENET_DISABLE); 1001 (void) bfe_wait_bit(bfe, BFE_ENET_CTRL, 1002 BFE_ENET_DISABLE, 10, 1); 1003 OUTL(bfe, BFE_DMATX_CTRL, 0); 1004 FLUSH(bfe, BFE_DMARX_STAT); 1005 drv_usecwait(20000); /* 20 milli seconds */ 1006 if (INL(bfe, BFE_DMARX_STAT) & BFE_STAT_EMASK) { 1007 (void) bfe_wait_bit(bfe, BFE_DMARX_STAT, BFE_STAT_SIDLE, 1008 10, 0); 1009 } 1010 OUTL(bfe, BFE_DMARX_CTRL, 0); 1011 } 1012 1013 bfe_core_reset(bfe); 1014 bfe_clear_stats(bfe); 1015 1016 OUTL(bfe, BFE_MDIO_CTRL, 0x8d); 1017 val = INL(bfe, BFE_DEVCTRL); 1018 if (!(val & BFE_IPP)) 1019 OUTL(bfe, BFE_ENET_CTRL, BFE_ENET_EPSEL); 1020 else if (INL(bfe, BFE_DEVCTRL & BFE_EPR)) { 1021 OUTL_AND(bfe, BFE_DEVCTRL, ~BFE_EPR); 1022 drv_usecwait(20000); /* 20 milli seconds */ 1023 } 1024 1025 OUTL_OR(bfe, BFE_MAC_CTRL, BFE_CTRL_CRC32_ENAB | BFE_CTRL_LED); 1026 1027 OUTL_AND(bfe, BFE_MAC_CTRL, ~BFE_CTRL_PDOWN); 1028 1029 OUTL(bfe, BFE_RCV_LAZY, ((1 << BFE_LAZY_FC_SHIFT) & 1030 BFE_LAZY_FC_MASK)); 1031 1032 OUTL_OR(bfe, BFE_RCV_LAZY, 0); 1033 1034 OUTL(bfe, BFE_RXMAXLEN, bfe->bfe_rx_ring.r_buf_len); 1035 OUTL(bfe, BFE_TXMAXLEN, bfe->bfe_tx_ring.r_buf_len); 1036 1037 OUTL(bfe, BFE_TX_WMARK, 56); 1038 1039 /* Program DMA channels */ 1040 OUTL(bfe, BFE_DMATX_CTRL, BFE_TX_CTRL_ENABLE); 1041 1042 /* 1043 * DMA addresses need to be added to BFE_PCI_DMA 1044 */ 1045 OUTL(bfe, BFE_DMATX_ADDR, 1046 bfe->bfe_tx_ring.r_desc_cookie.dmac_laddress + BFE_PCI_DMA); 1047 1048 OUTL(bfe, BFE_DMARX_CTRL, (BFE_RX_OFFSET << BFE_RX_CTRL_ROSHIFT) 1049 | BFE_RX_CTRL_ENABLE); 1050 1051 OUTL(bfe, BFE_DMARX_ADDR, 1052 bfe->bfe_rx_ring.r_desc_cookie.dmac_laddress + BFE_PCI_DMA); 1053 1054 (void) bfe_startup_phy(bfe); 1055 1056 bfe->bfe_chip_state = BFE_CHIP_INITIALIZED; 1057 } 1058 1059 /* 1060 * It enables interrupts. Should be the last step while starting chip. 1061 */ 1062 static void 1063 bfe_enable_chip_intrs(bfe_t *bfe) 1064 { 1065 /* Enable the chip and core */ 1066 OUTL(bfe, BFE_ENET_CTRL, BFE_ENET_ENABLE); 1067 1068 /* Enable interrupts */ 1069 OUTL(bfe, BFE_INTR_MASK, BFE_IMASK_DEF); 1070 } 1071 1072 /* 1073 * Common code to take care of setting RX side mode (filter). 1074 */ 1075 static void 1076 bfe_set_rx_mode(bfe_t *bfe) 1077 { 1078 uint32_t val; 1079 int i; 1080 ether_addr_t mac[ETHERADDRL] = {{0, 0, 0, 0, 0, 0}}; 1081 1082 /* 1083 * We don't touch RX filter if we were asked to suspend. It's fine 1084 * if chip is not active (no interface is plumbed on us). 1085 */ 1086 if (bfe->bfe_chip_state == BFE_CHIP_SUSPENDED) 1087 return; 1088 1089 val = INL(bfe, BFE_RXCONF); 1090 1091 val &= ~BFE_RXCONF_PROMISC; 1092 val &= ~BFE_RXCONF_DBCAST; 1093 1094 if ((bfe->bfe_chip_mode & BFE_RX_MODE_ENABLE) == 0) { 1095 OUTL(bfe, BFE_CAM_CTRL, 0); 1096 FLUSH(bfe, BFE_CAM_CTRL); 1097 } else if (bfe->bfe_chip_mode & BFE_RX_MODE_PROMISC) { 1098 val |= BFE_RXCONF_PROMISC; 1099 val &= ~BFE_RXCONF_DBCAST; 1100 } else { 1101 if (bfe->bfe_chip_state == BFE_CHIP_ACTIVE) { 1102 /* Flush everything */ 1103 OUTL(bfe, BFE_RXCONF, val | 1104 BFE_RXCONF_PROMISC | BFE_RXCONF_ALLMULTI); 1105 FLUSH(bfe, BFE_RXCONF); 1106 } 1107 1108 /* Disable CAM */ 1109 OUTL(bfe, BFE_CAM_CTRL, 0); 1110 FLUSH(bfe, BFE_CAM_CTRL); 1111 1112 /* 1113 * We receive all multicast packets. 1114 */ 1115 val |= BFE_RXCONF_ALLMULTI; 1116 1117 for (i = 0; i < BFE_MAX_MULTICAST_TABLE - 1; i++) { 1118 bfe_cam_write(bfe, (uchar_t *)mac, i); 1119 } 1120 1121 bfe_cam_write(bfe, bfe->bfe_ether_addr, i); 1122 1123 /* Enable CAM */ 1124 OUTL_OR(bfe, BFE_CAM_CTRL, BFE_CAM_ENABLE); 1125 FLUSH(bfe, BFE_CAM_CTRL); 1126 } 1127 1128 DTRACE_PROBE2(rx__mode__filter, int, bfe->bfe_unit, 1129 int, val); 1130 1131 OUTL(bfe, BFE_RXCONF, val); 1132 FLUSH(bfe, BFE_RXCONF); 1133 } 1134 1135 /* 1136 * Reset various variable values to initial state. 1137 */ 1138 static void 1139 bfe_init_vars(bfe_t *bfe) 1140 { 1141 bfe->bfe_chip_mode = BFE_RX_MODE_ENABLE; 1142 1143 /* Initial assumption */ 1144 bfe->bfe_chip.link = LINK_STATE_UNKNOWN; 1145 bfe->bfe_chip.speed = 0; 1146 bfe->bfe_chip.duplex = LINK_DUPLEX_UNKNOWN; 1147 1148 bfe->bfe_periodic_id = NULL; 1149 bfe->bfe_chip_state = BFE_CHIP_UNINITIALIZED; 1150 1151 bfe->bfe_tx_stall_time = 0; 1152 } 1153 1154 /* 1155 * Initializes TX side descriptor entries (bfe_desc_t). Each descriptor entry 1156 * has control (desc_ctl) and address (desc_addr) member. 1157 */ 1158 static void 1159 bfe_tx_desc_init(bfe_ring_t *r) 1160 { 1161 int i; 1162 uint32_t v; 1163 1164 for (i = 0; i < r->r_ndesc; i++) { 1165 PUT_DESC(r, (uint32_t *)&(r->r_desc[i].desc_ctl), 1166 (r->r_buf_dma[i].len & BFE_DESC_LEN)); 1167 1168 /* 1169 * DMA addresses need to be added to BFE_PCI_DMA 1170 */ 1171 PUT_DESC(r, (uint32_t *)&(r->r_desc[i].desc_addr), 1172 (r->r_buf_dma[i].cookie.dmac_laddress + BFE_PCI_DMA)); 1173 } 1174 1175 v = GET_DESC(r, (uint32_t *)&(r->r_desc[i - 1].desc_ctl)); 1176 PUT_DESC(r, (uint32_t *)&(r->r_desc[i - 1].desc_ctl), 1177 v | BFE_DESC_EOT); 1178 1179 (void) SYNC_DESC(r, 0, r->r_ndesc, DDI_DMA_SYNC_FORDEV); 1180 1181 r->r_curr_desc = 0; 1182 r->r_avail_desc = TX_NUM_DESC; 1183 r->r_cons_desc = 0; 1184 } 1185 1186 /* 1187 * Initializes RX side descriptor entries (bfe_desc_t). Each descriptor entry 1188 * has control (desc_ctl) and address (desc_addr) member. 1189 */ 1190 static void 1191 bfe_rx_desc_init(bfe_ring_t *r) 1192 { 1193 int i; 1194 uint32_t v; 1195 1196 for (i = 0; i < r->r_ndesc; i++) { 1197 PUT_DESC(r, (uint32_t *)&(r->r_desc[i].desc_ctl), 1198 (r->r_buf_dma[i].len& BFE_DESC_LEN)); 1199 1200 PUT_DESC(r, (uint32_t *)&(r->r_desc[i].desc_addr), 1201 (r->r_buf_dma[i].cookie.dmac_laddress + BFE_PCI_DMA)); 1202 1203 /* Initialize rx header (len, flags) */ 1204 bzero(r->r_buf_dma[i].addr, sizeof (bfe_rx_header_t)); 1205 1206 (void) SYNC_BUF(r, i, 0, sizeof (bfe_rx_header_t), 1207 DDI_DMA_SYNC_FORDEV); 1208 } 1209 1210 v = GET_DESC(r, (uint32_t *)&(r->r_desc[i - 1].desc_ctl)); 1211 PUT_DESC(r, (uint32_t *)&(r->r_desc[i - 1].desc_ctl), 1212 v | BFE_DESC_EOT); 1213 1214 (void) SYNC_DESC(r, 0, r->r_ndesc, DDI_DMA_SYNC_FORDEV); 1215 1216 /* TAIL of RX Descriptor */ 1217 OUTL(r->r_bfe, BFE_DMARX_PTR, ((i) * sizeof (bfe_desc_t))); 1218 1219 r->r_curr_desc = 0; 1220 r->r_avail_desc = RX_NUM_DESC; 1221 } 1222 1223 static int 1224 bfe_chip_start(bfe_t *bfe) 1225 { 1226 ASSERT_ALL_LOCKS(bfe); 1227 1228 /* 1229 * Stop the chip first & then Reset the chip. At last enable interrupts. 1230 */ 1231 bfe_chip_halt(bfe); 1232 bfe_stop_phy(bfe); 1233 1234 /* 1235 * Reset chip and start PHY. 1236 */ 1237 bfe_chip_reset(bfe); 1238 1239 /* 1240 * Initailize Descriptor Rings. 1241 */ 1242 bfe_tx_desc_init(&bfe->bfe_tx_ring); 1243 bfe_rx_desc_init(&bfe->bfe_rx_ring); 1244 1245 bfe->bfe_chip_state = BFE_CHIP_ACTIVE; 1246 bfe->bfe_chip_mode |= BFE_RX_MODE_ENABLE; 1247 bfe_set_rx_mode(bfe); 1248 bfe_enable_chip_intrs(bfe); 1249 1250 /* Check link, speed and duplex mode */ 1251 (void) bfe_check_link(bfe); 1252 1253 return (DDI_SUCCESS); 1254 } 1255 1256 1257 /* 1258 * Clear chip statistics. 1259 */ 1260 static void 1261 bfe_clear_stats(bfe_t *bfe) 1262 { 1263 ulong_t r; 1264 1265 OUTL(bfe, BFE_MIB_CTRL, BFE_MIB_CLR_ON_READ); 1266 1267 /* 1268 * Stat registers are cleared by reading. 1269 */ 1270 for (r = BFE_TX_GOOD_O; r <= BFE_TX_PAUSE; r += 4) 1271 (void) INL(bfe, r); 1272 1273 for (r = BFE_RX_GOOD_O; r <= BFE_RX_NPAUSE; r += 4) 1274 (void) INL(bfe, r); 1275 } 1276 1277 /* 1278 * Collect chip statistics. 1279 */ 1280 static void 1281 bfe_gather_stats(bfe_t *bfe) 1282 { 1283 ulong_t r; 1284 uint32_t *v; 1285 uint32_t txerr = 0, rxerr = 0, coll = 0; 1286 1287 v = &bfe->bfe_hw_stats.tx_good_octets; 1288 for (r = BFE_TX_GOOD_O; r <= BFE_TX_PAUSE; r += 4) { 1289 *v += INL(bfe, r); 1290 v++; 1291 } 1292 1293 v = &bfe->bfe_hw_stats.rx_good_octets; 1294 for (r = BFE_RX_GOOD_O; r <= BFE_RX_NPAUSE; r += 4) { 1295 *v += INL(bfe, r); 1296 v++; 1297 } 1298 1299 /* 1300 * TX : 1301 * ------- 1302 * tx_good_octets, tx_good_pkts, tx_octets 1303 * tx_pkts, tx_broadcast_pkts, tx_multicast_pkts 1304 * tx_len_64, tx_len_65_to_127, tx_len_128_to_255 1305 * tx_len_256_to_511, tx_len_512_to_1023, tx_len_1024_to_max 1306 * tx_jabber_pkts, tx_oversize_pkts, tx_fragment_pkts 1307 * tx_underruns, tx_total_cols, tx_single_cols 1308 * tx_multiple_cols, tx_excessive_cols, tx_late_cols 1309 * tx_defered, tx_carrier_lost, tx_pause_pkts 1310 * 1311 * RX : 1312 * ------- 1313 * rx_good_octets, rx_good_pkts, rx_octets 1314 * rx_pkts, rx_broadcast_pkts, rx_multicast_pkts 1315 * rx_len_64, rx_len_65_to_127, rx_len_128_to_255 1316 * rx_len_256_to_511, rx_len_512_to_1023, rx_len_1024_to_max 1317 * rx_jabber_pkts, rx_oversize_pkts, rx_fragment_pkts 1318 * rx_missed_pkts, rx_crc_align_errs, rx_undersize 1319 * rx_crc_errs, rx_align_errs, rx_symbol_errs 1320 * rx_pause_pkts, rx_nonpause_pkts 1321 */ 1322 1323 bfe->bfe_stats.ether_stat_carrier_errors = 1324 bfe->bfe_hw_stats.tx_carrier_lost; 1325 1326 /* txerr += bfe->bfe_hw_stats.tx_carrier_lost; */ 1327 1328 bfe->bfe_stats.ether_stat_ex_collisions = 1329 bfe->bfe_hw_stats.tx_excessive_cols; 1330 txerr += bfe->bfe_hw_stats.tx_excessive_cols; 1331 coll += bfe->bfe_hw_stats.tx_excessive_cols; 1332 1333 bfe->bfe_stats.ether_stat_fcs_errors = 1334 bfe->bfe_hw_stats.rx_crc_errs; 1335 rxerr += bfe->bfe_hw_stats.rx_crc_errs; 1336 1337 bfe->bfe_stats.ether_stat_first_collisions = 1338 bfe->bfe_hw_stats.tx_single_cols; 1339 coll += bfe->bfe_hw_stats.tx_single_cols; 1340 bfe->bfe_stats.ether_stat_multi_collisions = 1341 bfe->bfe_hw_stats.tx_multiple_cols; 1342 coll += bfe->bfe_hw_stats.tx_multiple_cols; 1343 1344 bfe->bfe_stats.ether_stat_toolong_errors = 1345 bfe->bfe_hw_stats.rx_oversize_pkts; 1346 rxerr += bfe->bfe_hw_stats.rx_oversize_pkts; 1347 1348 bfe->bfe_stats.ether_stat_tooshort_errors = 1349 bfe->bfe_hw_stats.rx_undersize; 1350 rxerr += bfe->bfe_hw_stats.rx_undersize; 1351 1352 bfe->bfe_stats.ether_stat_tx_late_collisions += 1353 bfe->bfe_hw_stats.tx_late_cols; 1354 1355 bfe->bfe_stats.ether_stat_defer_xmts += 1356 bfe->bfe_hw_stats.tx_defered; 1357 1358 bfe->bfe_stats.ether_stat_macrcv_errors += rxerr; 1359 bfe->bfe_stats.ether_stat_macxmt_errors += txerr; 1360 1361 bfe->bfe_stats.collisions += coll; 1362 } 1363 1364 /* 1365 * Gets the state for dladm command and all. 1366 */ 1367 int 1368 bfe_mac_getstat(void *arg, uint_t stat, uint64_t *val) 1369 { 1370 bfe_t *bfe = (bfe_t *)arg; 1371 uint64_t v; 1372 int err = 0; 1373 1374 rw_enter(&bfe->bfe_rwlock, RW_READER); 1375 1376 1377 switch (stat) { 1378 default: 1379 err = ENOTSUP; 1380 break; 1381 1382 case MAC_STAT_IFSPEED: 1383 /* 1384 * MAC layer will ask for IFSPEED first and hence we 1385 * collect it only once. 1386 */ 1387 if (bfe->bfe_chip_state == BFE_CHIP_ACTIVE) { 1388 /* 1389 * Update stats from the hardware. 1390 */ 1391 bfe_gather_stats(bfe); 1392 } 1393 v = bfe->bfe_chip.speed; 1394 break; 1395 1396 case ETHER_STAT_ADV_CAP_100T4: 1397 v = bfe->bfe_adv_100T4; 1398 break; 1399 1400 case ETHER_STAT_ADV_CAP_100FDX: 1401 v = (bfe->bfe_mii_anar & MII_ABILITY_100BASE_TX_FD) != 0; 1402 break; 1403 1404 case ETHER_STAT_ADV_CAP_100HDX: 1405 v = (bfe->bfe_mii_anar & MII_ABILITY_100BASE_TX) != 0; 1406 break; 1407 1408 case ETHER_STAT_ADV_CAP_10FDX: 1409 v = (bfe->bfe_mii_anar & MII_ABILITY_10BASE_T_FD) != 0; 1410 break; 1411 1412 case ETHER_STAT_ADV_CAP_10HDX: 1413 v = (bfe->bfe_mii_anar & MII_ABILITY_10BASE_T) != 0; 1414 break; 1415 1416 case ETHER_STAT_ADV_CAP_ASMPAUSE: 1417 v = 0; 1418 break; 1419 1420 case ETHER_STAT_ADV_CAP_AUTONEG: 1421 v = bfe->bfe_adv_aneg; 1422 break; 1423 1424 case ETHER_STAT_ADV_CAP_PAUSE: 1425 v = (bfe->bfe_mii_anar & MII_ABILITY_PAUSE) != 0; 1426 break; 1427 1428 case ETHER_STAT_ADV_REMFAULT: 1429 v = (bfe->bfe_mii_anar & MII_AN_ADVERT_REMFAULT) != 0; 1430 break; 1431 1432 case ETHER_STAT_ALIGN_ERRORS: 1433 /* MIB */ 1434 v = bfe->bfe_stats.ether_stat_align_errors; 1435 break; 1436 1437 case ETHER_STAT_CAP_100T4: 1438 v = (bfe->bfe_mii_bmsr & MII_STATUS_100_BASE_T4) != 0; 1439 break; 1440 1441 case ETHER_STAT_CAP_100FDX: 1442 v = (bfe->bfe_mii_bmsr & MII_STATUS_100_BASEX_FD) != 0; 1443 break; 1444 1445 case ETHER_STAT_CAP_100HDX: 1446 v = (bfe->bfe_mii_bmsr & MII_STATUS_100_BASEX) != 0; 1447 break; 1448 1449 case ETHER_STAT_CAP_10FDX: 1450 v = (bfe->bfe_mii_bmsr & MII_STATUS_10_FD) != 0; 1451 break; 1452 1453 case ETHER_STAT_CAP_10HDX: 1454 v = (bfe->bfe_mii_bmsr & MII_STATUS_10) != 0; 1455 break; 1456 1457 case ETHER_STAT_CAP_ASMPAUSE: 1458 v = 0; 1459 break; 1460 1461 case ETHER_STAT_CAP_AUTONEG: 1462 v = ((bfe->bfe_mii_bmsr & MII_STATUS_CANAUTONEG) != 0); 1463 break; 1464 1465 case ETHER_STAT_CAP_PAUSE: 1466 v = 1; 1467 break; 1468 1469 case ETHER_STAT_CAP_REMFAULT: 1470 v = (bfe->bfe_mii_bmsr & MII_STATUS_REMFAULT) != 0; 1471 break; 1472 1473 case ETHER_STAT_CARRIER_ERRORS: 1474 v = bfe->bfe_stats.ether_stat_carrier_errors; 1475 break; 1476 1477 case ETHER_STAT_JABBER_ERRORS: 1478 err = ENOTSUP; 1479 break; 1480 1481 case ETHER_STAT_DEFER_XMTS: 1482 v = bfe->bfe_stats.ether_stat_defer_xmts; 1483 break; 1484 1485 case ETHER_STAT_EX_COLLISIONS: 1486 /* MIB */ 1487 v = bfe->bfe_stats.ether_stat_ex_collisions; 1488 break; 1489 1490 case ETHER_STAT_FCS_ERRORS: 1491 /* MIB */ 1492 v = bfe->bfe_stats.ether_stat_fcs_errors; 1493 break; 1494 1495 case ETHER_STAT_FIRST_COLLISIONS: 1496 /* MIB */ 1497 v = bfe->bfe_stats.ether_stat_first_collisions; 1498 break; 1499 1500 case ETHER_STAT_LINK_ASMPAUSE: 1501 v = 0; 1502 break; 1503 1504 case ETHER_STAT_LINK_AUTONEG: 1505 v = (bfe->bfe_mii_bmcr & MII_CONTROL_ANE) != 0 && 1506 (bfe->bfe_mii_bmsr & MII_STATUS_ANDONE) != 0; 1507 break; 1508 1509 case ETHER_STAT_LINK_DUPLEX: 1510 v = bfe->bfe_chip.duplex; 1511 break; 1512 1513 case ETHER_STAT_LP_CAP_100T4: 1514 v = (bfe->bfe_mii_anlpar & MII_ABILITY_100BASE_T4) != 0; 1515 break; 1516 1517 case ETHER_STAT_LP_CAP_100FDX: 1518 v = (bfe->bfe_mii_anlpar & MII_ABILITY_100BASE_TX_FD) != 0; 1519 break; 1520 1521 case ETHER_STAT_LP_CAP_100HDX: 1522 v = (bfe->bfe_mii_anlpar & MII_ABILITY_100BASE_TX) != 0; 1523 break; 1524 1525 case ETHER_STAT_LP_CAP_10FDX: 1526 v = (bfe->bfe_mii_anlpar & MII_ABILITY_10BASE_T_FD) != 0; 1527 break; 1528 1529 case ETHER_STAT_LP_CAP_10HDX: 1530 v = (bfe->bfe_mii_anlpar & MII_ABILITY_10BASE_T) != 0; 1531 break; 1532 1533 case ETHER_STAT_LP_CAP_ASMPAUSE: 1534 v = 0; 1535 break; 1536 1537 case ETHER_STAT_LP_CAP_AUTONEG: 1538 v = (bfe->bfe_mii_exp & MII_AN_EXP_LPCANAN) != 0; 1539 break; 1540 1541 case ETHER_STAT_LP_CAP_PAUSE: 1542 v = (bfe->bfe_mii_anlpar & MII_ABILITY_PAUSE) != 0; 1543 break; 1544 1545 case ETHER_STAT_LP_REMFAULT: 1546 v = (bfe->bfe_mii_anlpar & MII_STATUS_REMFAULT) != 0; 1547 break; 1548 1549 case ETHER_STAT_MACRCV_ERRORS: 1550 v = bfe->bfe_stats.ether_stat_macrcv_errors; 1551 break; 1552 1553 case ETHER_STAT_MACXMT_ERRORS: 1554 v = bfe->bfe_stats.ether_stat_macxmt_errors; 1555 break; 1556 1557 case ETHER_STAT_MULTI_COLLISIONS: 1558 v = bfe->bfe_stats.ether_stat_multi_collisions; 1559 break; 1560 1561 case ETHER_STAT_SQE_ERRORS: 1562 err = ENOTSUP; 1563 break; 1564 1565 case ETHER_STAT_TOOLONG_ERRORS: 1566 v = bfe->bfe_stats.ether_stat_toolong_errors; 1567 break; 1568 1569 case ETHER_STAT_TOOSHORT_ERRORS: 1570 v = bfe->bfe_stats.ether_stat_tooshort_errors; 1571 break; 1572 1573 case ETHER_STAT_TX_LATE_COLLISIONS: 1574 v = bfe->bfe_stats.ether_stat_tx_late_collisions; 1575 break; 1576 1577 case ETHER_STAT_XCVR_ADDR: 1578 v = bfe->bfe_phy_addr; 1579 break; 1580 1581 case ETHER_STAT_XCVR_ID: 1582 v = bfe->bfe_phy_id; 1583 break; 1584 1585 case MAC_STAT_BRDCSTRCV: 1586 v = bfe->bfe_stats.brdcstrcv; 1587 break; 1588 1589 case MAC_STAT_BRDCSTXMT: 1590 v = bfe->bfe_stats.brdcstxmt; 1591 break; 1592 1593 case MAC_STAT_MULTIXMT: 1594 v = bfe->bfe_stats.multixmt; 1595 break; 1596 1597 case MAC_STAT_COLLISIONS: 1598 v = bfe->bfe_stats.collisions; 1599 break; 1600 1601 case MAC_STAT_IERRORS: 1602 v = bfe->bfe_stats.ierrors; 1603 break; 1604 1605 case MAC_STAT_IPACKETS: 1606 v = bfe->bfe_stats.ipackets; 1607 break; 1608 1609 case MAC_STAT_MULTIRCV: 1610 v = bfe->bfe_stats.multircv; 1611 break; 1612 1613 case MAC_STAT_NORCVBUF: 1614 v = bfe->bfe_stats.norcvbuf; 1615 break; 1616 1617 case MAC_STAT_NOXMTBUF: 1618 v = bfe->bfe_stats.noxmtbuf; 1619 break; 1620 1621 case MAC_STAT_OBYTES: 1622 v = bfe->bfe_stats.obytes; 1623 break; 1624 1625 case MAC_STAT_OERRORS: 1626 /* MIB */ 1627 v = bfe->bfe_stats.ether_stat_macxmt_errors; 1628 break; 1629 1630 case MAC_STAT_OPACKETS: 1631 v = bfe->bfe_stats.opackets; 1632 break; 1633 1634 case MAC_STAT_RBYTES: 1635 v = bfe->bfe_stats.rbytes; 1636 break; 1637 1638 case MAC_STAT_UNDERFLOWS: 1639 v = bfe->bfe_stats.underflows; 1640 break; 1641 1642 case MAC_STAT_OVERFLOWS: 1643 v = bfe->bfe_stats.overflows; 1644 break; 1645 } 1646 1647 rw_exit(&bfe->bfe_rwlock); 1648 1649 *val = v; 1650 return (err); 1651 } 1652 1653 int 1654 bfe_mac_getprop(void *arg, const char *name, mac_prop_id_t num, uint_t sz, 1655 void *val) 1656 { 1657 bfe_t *bfe = (bfe_t *)arg; 1658 int err = 0; 1659 1660 switch (num) { 1661 case MAC_PROP_DUPLEX: 1662 ASSERT(sz >= sizeof (link_duplex_t)); 1663 bcopy(&bfe->bfe_chip.duplex, val, sizeof (link_duplex_t)); 1664 break; 1665 1666 case MAC_PROP_SPEED: 1667 ASSERT(sz >= sizeof (uint64_t)); 1668 bcopy(&bfe->bfe_chip.speed, val, sizeof (uint64_t)); 1669 break; 1670 1671 case MAC_PROP_AUTONEG: 1672 *(uint8_t *)val = bfe->bfe_adv_aneg; 1673 break; 1674 1675 case MAC_PROP_ADV_100FDX_CAP: 1676 *(uint8_t *)val = bfe->bfe_adv_100fdx; 1677 break; 1678 1679 case MAC_PROP_EN_100FDX_CAP: 1680 *(uint8_t *)val = bfe->bfe_adv_100fdx; 1681 break; 1682 1683 case MAC_PROP_ADV_100HDX_CAP: 1684 *(uint8_t *)val = bfe->bfe_adv_100hdx; 1685 break; 1686 1687 case MAC_PROP_EN_100HDX_CAP: 1688 *(uint8_t *)val = bfe->bfe_adv_100hdx; 1689 break; 1690 1691 case MAC_PROP_ADV_10FDX_CAP: 1692 *(uint8_t *)val = bfe->bfe_adv_10fdx; 1693 break; 1694 1695 case MAC_PROP_EN_10FDX_CAP: 1696 *(uint8_t *)val = bfe->bfe_adv_10fdx; 1697 break; 1698 1699 case MAC_PROP_ADV_10HDX_CAP: 1700 *(uint8_t *)val = bfe->bfe_adv_10hdx; 1701 break; 1702 1703 case MAC_PROP_EN_10HDX_CAP: 1704 *(uint8_t *)val = bfe->bfe_adv_10hdx; 1705 break; 1706 1707 case MAC_PROP_ADV_100T4_CAP: 1708 *(uint8_t *)val = bfe->bfe_adv_100T4; 1709 break; 1710 1711 case MAC_PROP_EN_100T4_CAP: 1712 *(uint8_t *)val = bfe->bfe_adv_100T4; 1713 break; 1714 1715 default: 1716 err = ENOTSUP; 1717 } 1718 1719 return (err); 1720 } 1721 1722 1723 static void 1724 bfe_mac_propinfo(void *arg, const char *name, mac_prop_id_t num, 1725 mac_prop_info_handle_t prh) 1726 { 1727 bfe_t *bfe = (bfe_t *)arg; 1728 1729 switch (num) { 1730 case MAC_PROP_DUPLEX: 1731 case MAC_PROP_SPEED: 1732 case MAC_PROP_ADV_100FDX_CAP: 1733 case MAC_PROP_ADV_100HDX_CAP: 1734 case MAC_PROP_ADV_10FDX_CAP: 1735 case MAC_PROP_ADV_10HDX_CAP: 1736 case MAC_PROP_ADV_100T4_CAP: 1737 case MAC_PROP_EN_100T4_CAP: 1738 mac_prop_info_set_perm(prh, MAC_PROP_PERM_READ); 1739 break; 1740 1741 case MAC_PROP_AUTONEG: 1742 mac_prop_info_set_default_uint8(prh, bfe->bfe_cap_aneg); 1743 break; 1744 1745 case MAC_PROP_EN_100FDX_CAP: 1746 mac_prop_info_set_default_uint8(prh, bfe->bfe_cap_100fdx); 1747 break; 1748 1749 case MAC_PROP_EN_100HDX_CAP: 1750 mac_prop_info_set_default_uint8(prh, bfe->bfe_cap_100hdx); 1751 break; 1752 1753 case MAC_PROP_EN_10FDX_CAP: 1754 mac_prop_info_set_default_uint8(prh, bfe->bfe_cap_10fdx); 1755 break; 1756 1757 case MAC_PROP_EN_10HDX_CAP: 1758 mac_prop_info_set_default_uint8(prh, bfe->bfe_cap_10hdx); 1759 break; 1760 } 1761 } 1762 1763 1764 /*ARGSUSED*/ 1765 int 1766 bfe_mac_setprop(void *arg, const char *name, mac_prop_id_t num, uint_t sz, 1767 const void *val) 1768 { 1769 bfe_t *bfe = (bfe_t *)arg; 1770 uint8_t *advp; 1771 uint8_t *capp; 1772 int r = 0; 1773 1774 switch (num) { 1775 case MAC_PROP_EN_100FDX_CAP: 1776 advp = &bfe->bfe_adv_100fdx; 1777 capp = &bfe->bfe_cap_100fdx; 1778 break; 1779 1780 case MAC_PROP_EN_100HDX_CAP: 1781 advp = &bfe->bfe_adv_100hdx; 1782 capp = &bfe->bfe_cap_100hdx; 1783 break; 1784 1785 case MAC_PROP_EN_10FDX_CAP: 1786 advp = &bfe->bfe_adv_10fdx; 1787 capp = &bfe->bfe_cap_10fdx; 1788 break; 1789 1790 case MAC_PROP_EN_10HDX_CAP: 1791 advp = &bfe->bfe_adv_10hdx; 1792 capp = &bfe->bfe_cap_10hdx; 1793 break; 1794 1795 case MAC_PROP_AUTONEG: 1796 advp = &bfe->bfe_adv_aneg; 1797 capp = &bfe->bfe_cap_aneg; 1798 break; 1799 1800 default: 1801 return (ENOTSUP); 1802 } 1803 1804 if (*capp == 0) 1805 return (ENOTSUP); 1806 1807 bfe_grab_locks(bfe); 1808 1809 if (*advp != *(const uint8_t *)val) { 1810 *advp = *(const uint8_t *)val; 1811 1812 bfe->bfe_chip_action = BFE_ACTION_RESTART_SETPROP; 1813 if (bfe->bfe_chip_state == BFE_CHIP_ACTIVE) { 1814 /* 1815 * We need to stop the timer before grabbing locks 1816 * otherwise we can land-up in deadlock with untimeout. 1817 */ 1818 bfe_stop_timer(bfe); 1819 1820 bfe->bfe_chip_action |= BFE_ACTION_RESTART; 1821 1822 bfe_chip_restart(bfe); 1823 1824 /* 1825 * We leave SETPROP because properties can be 1826 * temporary. 1827 */ 1828 bfe->bfe_chip_action &= ~(BFE_ACTION_RESTART); 1829 r = 1; 1830 } 1831 } 1832 1833 bfe_release_locks(bfe); 1834 1835 /* kick-off a potential stopped downstream */ 1836 if (r) 1837 mac_tx_update(bfe->bfe_machdl); 1838 1839 return (0); 1840 } 1841 1842 1843 int 1844 bfe_mac_set_ether_addr(void *arg, const uint8_t *ea) 1845 { 1846 bfe_t *bfe = (bfe_t *)arg; 1847 1848 bfe_grab_locks(bfe); 1849 bcopy(ea, bfe->bfe_ether_addr, ETHERADDRL); 1850 bfe_set_rx_mode(bfe); 1851 bfe_release_locks(bfe); 1852 return (0); 1853 } 1854 1855 int 1856 bfe_mac_start(void *arg) 1857 { 1858 bfe_t *bfe = (bfe_t *)arg; 1859 1860 bfe_grab_locks(bfe); 1861 if (bfe_chip_start(bfe) == DDI_FAILURE) { 1862 bfe_release_locks(bfe); 1863 return (EINVAL); 1864 } 1865 1866 bfe_release_locks(bfe); 1867 1868 mac_tx_update(bfe->bfe_machdl); 1869 1870 return (0); 1871 } 1872 1873 void 1874 bfe_mac_stop(void *arg) 1875 { 1876 bfe_t *bfe = (bfe_t *)arg; 1877 1878 /* 1879 * We need to stop the timer before grabbing locks otherwise 1880 * we can land-up in deadlock with untimeout. 1881 */ 1882 bfe_stop_timer(bfe); 1883 1884 bfe_grab_locks(bfe); 1885 1886 /* 1887 * First halt the chip by disabling interrupts. 1888 */ 1889 bfe_chip_halt(bfe); 1890 bfe_stop_phy(bfe); 1891 1892 bfe->bfe_chip_state = BFE_CHIP_STOPPED; 1893 1894 /* 1895 * This will leave the PHY running. 1896 */ 1897 bfe_chip_reset(bfe); 1898 1899 /* 1900 * Disable RX register. 1901 */ 1902 bfe->bfe_chip_mode &= ~BFE_RX_MODE_ENABLE; 1903 bfe_set_rx_mode(bfe); 1904 1905 bfe_release_locks(bfe); 1906 } 1907 1908 /* 1909 * Send a packet down the wire. 1910 */ 1911 static int 1912 bfe_send_a_packet(bfe_t *bfe, mblk_t *mp) 1913 { 1914 bfe_ring_t *r = &bfe->bfe_tx_ring; 1915 uint32_t cur = r->r_curr_desc; 1916 uint32_t next; 1917 size_t pktlen = msgsize(mp); 1918 uchar_t *buf; 1919 uint32_t v; 1920 1921 ASSERT(MUTEX_HELD(&r->r_lock)); 1922 ASSERT(mp != NULL); 1923 1924 if (pktlen > r->r_buf_len) { 1925 freemsg(mp); 1926 return (BFE_SUCCESS); 1927 } 1928 1929 /* 1930 * There is a big reason why we don't check for '0'. It becomes easy 1931 * for us to not roll over the ring since we are based on producer (tx) 1932 * and consumer (reclaim by an interrupt) model. Especially when we 1933 * run out of TX descriptor, chip will send a single interrupt and 1934 * both producer and consumer counter will be same. So we keep a 1935 * difference of 1 always. 1936 */ 1937 if (r->r_avail_desc <= 1) { 1938 bfe->bfe_stats.noxmtbuf++; 1939 bfe->bfe_tx_resched = 1; 1940 return (BFE_FAILURE); 1941 } 1942 1943 /* 1944 * Get the DMA buffer to hold packet. 1945 */ 1946 buf = (uchar_t *)r->r_buf_dma[cur].addr; 1947 1948 mcopymsg(mp, buf); /* it also frees mp */ 1949 1950 /* 1951 * Gather statistics. 1952 */ 1953 if (buf[0] & 0x1) { 1954 if (bcmp(buf, bfe_broadcast, ETHERADDRL) != 0) 1955 bfe->bfe_stats.multixmt++; 1956 else 1957 bfe->bfe_stats.brdcstxmt++; 1958 } 1959 bfe->bfe_stats.opackets++; 1960 bfe->bfe_stats.obytes += pktlen; 1961 1962 1963 /* 1964 * Program the DMA descriptor (start and end of frame are same). 1965 */ 1966 next = cur; 1967 v = (pktlen & BFE_DESC_LEN) | BFE_DESC_IOC | BFE_DESC_SOF | 1968 BFE_DESC_EOF; 1969 1970 if (cur == (TX_NUM_DESC - 1)) 1971 v |= BFE_DESC_EOT; 1972 1973 PUT_DESC(r, (uint32_t *)&(r->r_desc[cur].desc_ctl), v); 1974 1975 /* 1976 * DMA addresses need to be added to BFE_PCI_DMA 1977 */ 1978 PUT_DESC(r, (uint32_t *)&(r->r_desc[cur].desc_addr), 1979 (r->r_buf_dma[cur].cookie.dmac_laddress + BFE_PCI_DMA)); 1980 1981 /* 1982 * Sync the packet data for the device. 1983 */ 1984 (void) SYNC_BUF(r, cur, 0, pktlen, DDI_DMA_SYNC_FORDEV); 1985 1986 /* Move to next descriptor slot */ 1987 BFE_INC_SLOT(next, TX_NUM_DESC); 1988 1989 (void) SYNC_DESC(r, 0, r->r_ndesc, DDI_DMA_SYNC_FORDEV); 1990 1991 r->r_curr_desc = next; 1992 1993 /* 1994 * The order should be 1,2,3,... for BFE_DMATX_PTR if 0,1,2,3,... 1995 * descriptor slot are being programmed. 1996 */ 1997 OUTL(bfe, BFE_DMATX_PTR, next * sizeof (bfe_desc_t)); 1998 FLUSH(bfe, BFE_DMATX_PTR); 1999 2000 r->r_avail_desc--; 2001 2002 /* 2003 * Let timeout know that it must reset the chip if a 2004 * packet is not sent down the wire for more than 5 seconds. 2005 */ 2006 bfe->bfe_tx_stall_time = gethrtime() + (5 * 1000000000ULL); 2007 2008 return (BFE_SUCCESS); 2009 } 2010 2011 mblk_t * 2012 bfe_mac_transmit_packet(void *arg, mblk_t *mp) 2013 { 2014 bfe_t *bfe = (bfe_t *)arg; 2015 bfe_ring_t *r = &bfe->bfe_tx_ring; 2016 mblk_t *nmp; 2017 2018 mutex_enter(&r->r_lock); 2019 2020 if (bfe->bfe_chip_state != BFE_CHIP_ACTIVE) { 2021 DTRACE_PROBE1(tx__chip__not__active, int, bfe->bfe_unit); 2022 2023 freemsgchain(mp); 2024 mutex_exit(&r->r_lock); 2025 return (NULL); 2026 } 2027 2028 2029 while (mp != NULL) { 2030 nmp = mp->b_next; 2031 mp->b_next = NULL; 2032 2033 if (bfe_send_a_packet(bfe, mp) == BFE_FAILURE) { 2034 mp->b_next = nmp; 2035 break; 2036 } 2037 mp = nmp; 2038 } 2039 2040 mutex_exit(&r->r_lock); 2041 2042 return (mp); 2043 } 2044 2045 int 2046 bfe_mac_set_promisc(void *arg, boolean_t promiscflag) 2047 { 2048 bfe_t *bfe = (bfe_t *)arg; 2049 2050 bfe_grab_locks(bfe); 2051 if (bfe->bfe_chip_state != BFE_CHIP_ACTIVE) { 2052 bfe_release_locks(bfe); 2053 return (EIO); 2054 } 2055 2056 if (promiscflag) { 2057 /* Set Promiscous on */ 2058 bfe->bfe_chip_mode |= BFE_RX_MODE_PROMISC; 2059 } else { 2060 bfe->bfe_chip_mode &= ~BFE_RX_MODE_PROMISC; 2061 } 2062 2063 bfe_set_rx_mode(bfe); 2064 bfe_release_locks(bfe); 2065 2066 return (0); 2067 } 2068 2069 int 2070 bfe_mac_set_multicast(void *arg, boolean_t add, const uint8_t *macaddr) 2071 { 2072 /* 2073 * It was too much of pain to implement multicast in CAM. Instead 2074 * we never disable multicast filter. 2075 */ 2076 return (0); 2077 } 2078 2079 static mac_callbacks_t bfe_mac_callbacks = { 2080 MC_SETPROP | MC_GETPROP | MC_PROPINFO, 2081 bfe_mac_getstat, /* gets stats */ 2082 bfe_mac_start, /* starts mac */ 2083 bfe_mac_stop, /* stops mac */ 2084 bfe_mac_set_promisc, /* sets promisc mode for snoop */ 2085 bfe_mac_set_multicast, /* multicast implementation */ 2086 bfe_mac_set_ether_addr, /* sets ethernet address (unicast) */ 2087 bfe_mac_transmit_packet, /* transmits packet */ 2088 NULL, 2089 NULL, /* ioctl */ 2090 NULL, /* getcap */ 2091 NULL, /* open */ 2092 NULL, /* close */ 2093 bfe_mac_setprop, 2094 bfe_mac_getprop, 2095 bfe_mac_propinfo 2096 }; 2097 2098 static void 2099 bfe_error_handler(bfe_t *bfe, int intr_mask) 2100 { 2101 uint32_t v; 2102 2103 if (intr_mask & BFE_ISTAT_RFO) { 2104 bfe->bfe_stats.overflows++; 2105 bfe->bfe_chip_action |= 2106 (BFE_ACTION_RESTART | BFE_ACTION_RESTART_FAULT); 2107 goto action; 2108 } 2109 2110 if (intr_mask & BFE_ISTAT_TFU) { 2111 bfe->bfe_stats.underflows++; 2112 return; 2113 } 2114 2115 /* Descriptor Protocol Error */ 2116 if (intr_mask & BFE_ISTAT_DPE) { 2117 bfe_error(bfe->bfe_dip, 2118 "Descriptor Protocol Error. Halting Chip"); 2119 bfe->bfe_chip_action |= 2120 (BFE_ACTION_RESTART | BFE_ACTION_RESTART_FAULT); 2121 goto action; 2122 } 2123 2124 /* Descriptor Error */ 2125 if (intr_mask & BFE_ISTAT_DSCE) { 2126 bfe_error(bfe->bfe_dip, "Descriptor Error. Restarting Chip"); 2127 goto action; 2128 } 2129 2130 /* Receive Descr. Underflow */ 2131 if (intr_mask & BFE_ISTAT_RDU) { 2132 bfe_error(bfe->bfe_dip, 2133 "Receive Descriptor Underflow. Restarting Chip"); 2134 bfe->bfe_stats.ether_stat_macrcv_errors++; 2135 bfe->bfe_chip_action |= 2136 (BFE_ACTION_RESTART | BFE_ACTION_RESTART_FAULT); 2137 goto action; 2138 } 2139 2140 v = INL(bfe, BFE_DMATX_STAT); 2141 2142 /* Error while sending a packet */ 2143 if (v & BFE_STAT_EMASK) { 2144 bfe->bfe_stats.ether_stat_macxmt_errors++; 2145 bfe_error(bfe->bfe_dip, 2146 "Error while sending a packet. Restarting Chip"); 2147 } 2148 2149 /* Error while receiving a packet */ 2150 v = INL(bfe, BFE_DMARX_STAT); 2151 if (v & BFE_RX_FLAG_ERRORS) { 2152 bfe->bfe_stats.ierrors++; 2153 bfe_error(bfe->bfe_dip, 2154 "Error while receiving a packet. Restarting Chip"); 2155 } 2156 2157 2158 bfe->bfe_chip_action |= 2159 (BFE_ACTION_RESTART | BFE_ACTION_RESTART_FAULT); 2160 2161 action: 2162 bfe_chip_halt(bfe); 2163 } 2164 2165 /* 2166 * It will recycle a RX descriptor slot. 2167 */ 2168 static void 2169 bfe_rx_desc_buf_reinit(bfe_t *bfe, uint_t slot) 2170 { 2171 bfe_ring_t *r = &bfe->bfe_rx_ring; 2172 uint32_t v; 2173 2174 slot %= RX_NUM_DESC; 2175 2176 bzero(r->r_buf_dma[slot].addr, sizeof (bfe_rx_header_t)); 2177 2178 (void) SYNC_BUF(r, slot, 0, BFE_RX_OFFSET, DDI_DMA_SYNC_FORDEV); 2179 2180 v = r->r_buf_dma[slot].len & BFE_DESC_LEN; 2181 if (slot == (RX_NUM_DESC - 1)) 2182 v |= BFE_DESC_EOT; 2183 2184 PUT_DESC(r, (uint32_t *)&(r->r_desc[slot].desc_ctl), v); 2185 2186 /* 2187 * DMA addresses need to be added to BFE_PCI_DMA 2188 */ 2189 PUT_DESC(r, (uint32_t *)&(r->r_desc[slot].desc_addr), 2190 (r->r_buf_dma[slot].cookie.dmac_laddress + BFE_PCI_DMA)); 2191 } 2192 2193 /* 2194 * Gets called from interrupt context to handle RX interrupt. 2195 */ 2196 static mblk_t * 2197 bfe_receive(bfe_t *bfe, int intr_mask) 2198 { 2199 int rxstat, current; 2200 mblk_t *mp = NULL, *rx_head, *rx_tail; 2201 uchar_t *rx_header; 2202 uint16_t len; 2203 uchar_t *bp; 2204 bfe_ring_t *r = &bfe->bfe_rx_ring; 2205 int i; 2206 2207 rxstat = INL(bfe, BFE_DMARX_STAT); 2208 current = (rxstat & BFE_STAT_CDMASK) / sizeof (bfe_desc_t); 2209 i = r->r_curr_desc; 2210 2211 rx_head = rx_tail = NULL; 2212 2213 DTRACE_PROBE3(receive, int, bfe->bfe_unit, 2214 int, r->r_curr_desc, 2215 int, current); 2216 2217 for (i = r->r_curr_desc; i != current; 2218 BFE_INC_SLOT(i, RX_NUM_DESC)) { 2219 2220 /* 2221 * Sync the buffer associated with the descriptor table entry. 2222 */ 2223 (void) SYNC_BUF(r, i, 0, r->r_buf_dma[i].len, 2224 DDI_DMA_SYNC_FORKERNEL); 2225 2226 rx_header = (void *)r->r_buf_dma[i].addr; 2227 2228 /* 2229 * We do this to make sure we are endian neutral. Chip is 2230 * big endian. 2231 * 2232 * The header looks like :- 2233 * 2234 * Offset 0 -> uint16_t len 2235 * Offset 2 -> uint16_t flags 2236 * Offset 4 -> uint16_t pad[12] 2237 */ 2238 len = (rx_header[1] << 8) | rx_header[0]; 2239 len -= 4; /* CRC bytes need to be removed */ 2240 2241 /* 2242 * Don't receive this packet if pkt length is greater than 2243 * MTU + VLAN_TAGSZ. 2244 */ 2245 if (len > r->r_buf_len) { 2246 /* Recycle slot for later use */ 2247 bfe_rx_desc_buf_reinit(bfe, i); 2248 continue; 2249 } 2250 2251 if ((mp = allocb(len + VLAN_TAGSZ, BPRI_MED)) != NULL) { 2252 mp->b_rptr += VLAN_TAGSZ; 2253 bp = mp->b_rptr; 2254 mp->b_wptr = bp + len; 2255 2256 /* sizeof (bfe_rx_header_t) + 2 */ 2257 bcopy(r->r_buf_dma[i].addr + 2258 BFE_RX_OFFSET, bp, len); 2259 2260 mp->b_next = NULL; 2261 if (rx_tail == NULL) 2262 rx_head = rx_tail = mp; 2263 else { 2264 rx_tail->b_next = mp; 2265 rx_tail = mp; 2266 } 2267 2268 /* Number of packets received so far */ 2269 bfe->bfe_stats.ipackets++; 2270 2271 /* Total bytes of packets received so far */ 2272 bfe->bfe_stats.rbytes += len; 2273 2274 if (bcmp(mp->b_rptr, bfe_broadcast, ETHERADDRL) == 0) 2275 bfe->bfe_stats.brdcstrcv++; 2276 else 2277 bfe->bfe_stats.multircv++; 2278 } else { 2279 bfe->bfe_stats.norcvbuf++; 2280 /* Recycle the slot for later use */ 2281 bfe_rx_desc_buf_reinit(bfe, i); 2282 break; 2283 } 2284 2285 /* 2286 * Reinitialize the current descriptor slot's buffer so that 2287 * it can be reused. 2288 */ 2289 bfe_rx_desc_buf_reinit(bfe, i); 2290 } 2291 2292 r->r_curr_desc = i; 2293 2294 (void) SYNC_DESC(r, 0, r->r_ndesc, DDI_DMA_SYNC_FORDEV); 2295 2296 return (rx_head); 2297 } 2298 2299 static int 2300 bfe_tx_reclaim(bfe_ring_t *r) 2301 { 2302 uint32_t cur, start; 2303 uint32_t v; 2304 2305 cur = INL(r->r_bfe, BFE_DMATX_STAT) & BFE_STAT_CDMASK; 2306 cur = cur / sizeof (bfe_desc_t); 2307 2308 /* 2309 * Start with the last descriptor consumed by the chip. 2310 */ 2311 start = r->r_cons_desc; 2312 2313 DTRACE_PROBE3(tx__reclaim, int, r->r_bfe->bfe_unit, 2314 int, start, 2315 int, cur); 2316 2317 /* 2318 * There will be at least one descriptor to process. 2319 */ 2320 while (start != cur) { 2321 r->r_avail_desc++; 2322 v = r->r_buf_dma[start].len & BFE_DESC_LEN; 2323 if (start == (TX_NUM_DESC - 1)) 2324 v |= BFE_DESC_EOT; 2325 2326 PUT_DESC(r, (uint32_t *)&(r->r_desc[start].desc_ctl), v); 2327 PUT_DESC(r, (uint32_t *)&(r->r_desc[start].desc_addr), 2328 (r->r_buf_dma[start].cookie.dmac_laddress + BFE_PCI_DMA)); 2329 2330 /* Move to next descriptor in TX ring */ 2331 BFE_INC_SLOT(start, TX_NUM_DESC); 2332 } 2333 2334 (void) ddi_dma_sync(r->r_desc_dma_handle, 2335 0, (r->r_ndesc * sizeof (bfe_desc_t)), 2336 DDI_DMA_SYNC_FORDEV); 2337 2338 r->r_cons_desc = start; /* consumed pointer */ 2339 r->r_bfe->bfe_tx_stall_time = 0; 2340 2341 return (cur); 2342 } 2343 2344 static int 2345 bfe_tx_done(bfe_t *bfe, int intr_mask) 2346 { 2347 bfe_ring_t *r = &bfe->bfe_tx_ring; 2348 int resched = 0; 2349 2350 mutex_enter(&r->r_lock); 2351 (void) bfe_tx_reclaim(r); 2352 2353 if (bfe->bfe_tx_resched) { 2354 resched = 1; 2355 bfe->bfe_tx_resched = 0; 2356 } 2357 mutex_exit(&r->r_lock); 2358 2359 return (resched); 2360 } 2361 2362 /* 2363 * ISR for interrupt handling 2364 */ 2365 static uint_t 2366 bfe_interrupt(caddr_t arg1, caddr_t arg2) 2367 { 2368 bfe_t *bfe = (void *)arg1; 2369 uint32_t intr_stat; 2370 mblk_t *rx_head = NULL; 2371 int resched = 0; 2372 2373 /* 2374 * Grab the lock to avoid stopping the chip while this interrupt 2375 * is handled. 2376 */ 2377 rw_enter(&bfe->bfe_rwlock, RW_READER); 2378 2379 /* 2380 * It's necessary to read intr stat again because masking interrupt 2381 * register does not really mask interrupts coming from the chip. 2382 */ 2383 intr_stat = INL(bfe, BFE_INTR_STAT); 2384 intr_stat &= BFE_IMASK_DEF; 2385 OUTL(bfe, BFE_INTR_STAT, intr_stat); 2386 (void) INL(bfe, BFE_INTR_STAT); 2387 2388 if (intr_stat == 0) { 2389 rw_exit(&bfe->bfe_rwlock); 2390 return (DDI_INTR_UNCLAIMED); 2391 } 2392 2393 DTRACE_PROBE2(bfe__interrupt, int, bfe->bfe_unit, 2394 int, intr_stat); 2395 2396 if (bfe->bfe_chip_state != BFE_CHIP_ACTIVE) { 2397 /* 2398 * If chip is suspended then we just return. 2399 */ 2400 if (bfe->bfe_chip_state == BFE_CHIP_SUSPENDED) { 2401 rw_exit(&bfe->bfe_rwlock); 2402 DTRACE_PROBE1(interrupt__chip__is__suspend, int, 2403 bfe->bfe_unit); 2404 return (DDI_INTR_CLAIMED); 2405 } 2406 2407 /* 2408 * Halt the chip again i.e basically disable interrupts. 2409 */ 2410 bfe_chip_halt(bfe); 2411 rw_exit(&bfe->bfe_rwlock); 2412 DTRACE_PROBE1(interrupt__chip__not__active, int, 2413 bfe->bfe_unit); 2414 return (DDI_INTR_CLAIMED); 2415 } 2416 2417 /* A packet was received */ 2418 if (intr_stat & BFE_ISTAT_RX) { 2419 rx_head = bfe_receive(bfe, intr_stat); 2420 } 2421 2422 /* A packet was sent down the wire */ 2423 if (intr_stat & BFE_ISTAT_TX) { 2424 resched = bfe_tx_done(bfe, intr_stat); 2425 } 2426 2427 /* There was an error */ 2428 if (intr_stat & BFE_ISTAT_ERRORS) { 2429 bfe_error_handler(bfe, intr_stat); 2430 } 2431 2432 rw_exit(&bfe->bfe_rwlock); 2433 2434 /* 2435 * Pass the list of packets received from chip to MAC layer. 2436 */ 2437 if (rx_head) { 2438 mac_rx(bfe->bfe_machdl, 0, rx_head); 2439 } 2440 2441 /* 2442 * Let the MAC start sending pkts to a potential stopped stream. 2443 */ 2444 if (resched) 2445 mac_tx_update(bfe->bfe_machdl); 2446 2447 return (DDI_INTR_CLAIMED); 2448 } 2449 2450 /* 2451 * Removes registered interrupt handler. 2452 */ 2453 static void 2454 bfe_remove_intr(bfe_t *bfe) 2455 { 2456 (void) ddi_intr_remove_handler(bfe->bfe_intrhdl); 2457 (void) ddi_intr_free(bfe->bfe_intrhdl); 2458 } 2459 2460 /* 2461 * Add an interrupt for the driver. 2462 */ 2463 static int 2464 bfe_add_intr(bfe_t *bfe) 2465 { 2466 int nintrs = 1; 2467 int ret; 2468 2469 ret = ddi_intr_alloc(bfe->bfe_dip, &bfe->bfe_intrhdl, 2470 DDI_INTR_TYPE_FIXED, /* type */ 2471 0, /* inumber */ 2472 1, /* count */ 2473 &nintrs, /* actual nintrs */ 2474 DDI_INTR_ALLOC_STRICT); 2475 2476 if (ret != DDI_SUCCESS) { 2477 bfe_error(bfe->bfe_dip, "ddi_intr_alloc() failed" 2478 " : ret : %d", ret); 2479 return (DDI_FAILURE); 2480 } 2481 2482 ret = ddi_intr_add_handler(bfe->bfe_intrhdl, bfe_interrupt, bfe, NULL); 2483 if (ret != DDI_SUCCESS) { 2484 bfe_error(bfe->bfe_dip, "ddi_intr_add_handler() failed"); 2485 (void) ddi_intr_free(bfe->bfe_intrhdl); 2486 return (DDI_FAILURE); 2487 } 2488 2489 ret = ddi_intr_get_pri(bfe->bfe_intrhdl, &bfe->bfe_intrpri); 2490 if (ret != DDI_SUCCESS) { 2491 bfe_error(bfe->bfe_dip, "ddi_intr_get_pri() failed"); 2492 bfe_remove_intr(bfe); 2493 return (DDI_FAILURE); 2494 } 2495 2496 return (DDI_SUCCESS); 2497 } 2498 2499 2500 /* 2501 * Identify chipset family. 2502 */ 2503 static int 2504 bfe_identify_hardware(bfe_t *bfe) 2505 { 2506 uint16_t vid, did; 2507 int i; 2508 2509 vid = pci_config_get16(bfe->bfe_conf_handle, PCI_CONF_VENID); 2510 did = pci_config_get16(bfe->bfe_conf_handle, PCI_CONF_DEVID); 2511 2512 for (i = 0; i < (sizeof (bfe_cards) / sizeof (bfe_cards_t)); i++) { 2513 if (bfe_cards[i].vendor_id == vid && 2514 bfe_cards[i].device_id == did) { 2515 return (BFE_SUCCESS); 2516 } 2517 } 2518 2519 bfe_error(bfe->bfe_dip, "bfe driver is attaching to unknown pci%d,%d" 2520 " vendor/device-id card", vid, did); 2521 2522 return (BFE_SUCCESS); 2523 } 2524 2525 /* 2526 * Maps device registers. 2527 */ 2528 static int 2529 bfe_regs_map(bfe_t *bfe) 2530 { 2531 dev_info_t *dip = bfe->bfe_dip; 2532 int ret; 2533 2534 ret = ddi_regs_map_setup(dip, 1, &bfe->bfe_mem_regset.addr, 0, 0, 2535 &bfe_dev_attr, &bfe->bfe_mem_regset.hdl); 2536 2537 if (ret != DDI_SUCCESS) { 2538 bfe_error(bfe->bfe_dip, "ddi_regs_map_setup failed"); 2539 return (DDI_FAILURE); 2540 } 2541 2542 return (DDI_SUCCESS); 2543 } 2544 2545 static void 2546 bfe_unmap_regs(bfe_t *bfe) 2547 { 2548 ddi_regs_map_free(&bfe->bfe_mem_regset.hdl); 2549 } 2550 2551 static int 2552 bfe_get_chip_config(bfe_t *bfe) 2553 { 2554 uint32_t prom[BFE_EEPROM_SIZE]; 2555 int i; 2556 2557 /* 2558 * Read EEPROM in prom[] 2559 */ 2560 for (i = 0; i < BFE_EEPROM_SIZE; i++) { 2561 prom[i] = INL(bfe, BFE_EEPROM_BASE + i * sizeof (uint32_t)); 2562 } 2563 2564 bfe->bfe_dev_addr[0] = bfe->bfe_ether_addr[0] = 2565 INB(bfe, BFE_EEPROM_BASE + 79); 2566 2567 bfe->bfe_dev_addr[1] = bfe->bfe_ether_addr[1] = 2568 INB(bfe, BFE_EEPROM_BASE + 78); 2569 2570 bfe->bfe_dev_addr[2] = bfe->bfe_ether_addr[2] = 2571 INB(bfe, BFE_EEPROM_BASE + 81); 2572 2573 bfe->bfe_dev_addr[3] = bfe->bfe_ether_addr[3] = 2574 INB(bfe, BFE_EEPROM_BASE + 80); 2575 2576 bfe->bfe_dev_addr[4] = bfe->bfe_ether_addr[4] = 2577 INB(bfe, BFE_EEPROM_BASE + 83); 2578 2579 bfe->bfe_dev_addr[5] = bfe->bfe_ether_addr[5] = 2580 INB(bfe, BFE_EEPROM_BASE + 82); 2581 2582 bfe->bfe_phy_addr = -1; 2583 2584 return (DDI_SUCCESS); 2585 } 2586 2587 /* 2588 * Ring Management routines 2589 */ 2590 static int 2591 bfe_ring_buf_alloc(bfe_t *bfe, bfe_ring_t *r, int slot, int d) 2592 { 2593 int err; 2594 uint_t count = 0; 2595 2596 err = ddi_dma_alloc_handle(bfe->bfe_dip, 2597 &bfe_dma_attr_buf, DDI_DMA_SLEEP, NULL, 2598 &r->r_buf_dma[slot].handle); 2599 2600 if (err != DDI_SUCCESS) { 2601 bfe_error(bfe->bfe_dip, " bfe_ring_buf_alloc() :" 2602 " alloc_handle failed"); 2603 goto fail0; 2604 } 2605 2606 err = ddi_dma_mem_alloc(r->r_buf_dma[slot].handle, 2607 r->r_buf_len, &bfe_buf_attr, DDI_DMA_STREAMING, 2608 DDI_DMA_SLEEP, NULL, &r->r_buf_dma[slot].addr, 2609 &r->r_buf_dma[slot].len, 2610 &r->r_buf_dma[slot].acchdl); 2611 2612 if (err != DDI_SUCCESS) { 2613 bfe_error(bfe->bfe_dip, " bfe_ring_buf_alloc() :" 2614 " mem_alloc failed :%d", err); 2615 goto fail1; 2616 } 2617 2618 err = ddi_dma_addr_bind_handle(r->r_buf_dma[slot].handle, 2619 NULL, r->r_buf_dma[slot].addr, 2620 r->r_buf_dma[slot].len, 2621 (DDI_DMA_RDWR | DDI_DMA_STREAMING), 2622 DDI_DMA_SLEEP, NULL, 2623 &r->r_buf_dma[slot].cookie, 2624 &count); 2625 2626 if (err != DDI_DMA_MAPPED) { 2627 bfe_error(bfe->bfe_dip, " bfe_ring_buf_alloc() :" 2628 " bind_handle failed"); 2629 goto fail2; 2630 } 2631 2632 if (count > 1) { 2633 bfe_error(bfe->bfe_dip, " bfe_ring_buf_alloc() :" 2634 " more than one DMA cookie"); 2635 (void) ddi_dma_unbind_handle(r->r_buf_dma[slot].handle); 2636 goto fail2; 2637 } 2638 2639 return (DDI_SUCCESS); 2640 fail2: 2641 ddi_dma_mem_free(&r->r_buf_dma[slot].acchdl); 2642 fail1: 2643 ddi_dma_free_handle(&r->r_buf_dma[slot].handle); 2644 fail0: 2645 return (DDI_FAILURE); 2646 } 2647 2648 static void 2649 bfe_ring_buf_free(bfe_ring_t *r, int slot) 2650 { 2651 if (r->r_buf_dma == NULL) 2652 return; 2653 2654 (void) ddi_dma_unbind_handle(r->r_buf_dma[slot].handle); 2655 ddi_dma_mem_free(&r->r_buf_dma[slot].acchdl); 2656 ddi_dma_free_handle(&r->r_buf_dma[slot].handle); 2657 } 2658 2659 static void 2660 bfe_buffer_free(bfe_ring_t *r) 2661 { 2662 int i; 2663 2664 for (i = 0; i < r->r_ndesc; i++) { 2665 bfe_ring_buf_free(r, i); 2666 } 2667 } 2668 2669 static void 2670 bfe_ring_desc_free(bfe_ring_t *r) 2671 { 2672 (void) ddi_dma_unbind_handle(r->r_desc_dma_handle); 2673 ddi_dma_mem_free(&r->r_desc_acc_handle); 2674 ddi_dma_free_handle(&r->r_desc_dma_handle); 2675 kmem_free(r->r_buf_dma, r->r_ndesc * sizeof (bfe_dma_t)); 2676 2677 r->r_buf_dma = NULL; 2678 r->r_desc = NULL; 2679 } 2680 2681 2682 static int 2683 bfe_ring_desc_alloc(bfe_t *bfe, bfe_ring_t *r, int d) 2684 { 2685 int err, i, fail = 0; 2686 caddr_t ring; 2687 size_t size_krnl = 0, size_dma = 0, ring_len = 0; 2688 ddi_dma_cookie_t cookie; 2689 uint_t count = 0; 2690 2691 ASSERT(bfe != NULL); 2692 2693 size_krnl = r->r_ndesc * sizeof (bfe_dma_t); 2694 size_dma = r->r_ndesc * sizeof (bfe_desc_t); 2695 r->r_buf_dma = kmem_zalloc(size_krnl, KM_SLEEP); 2696 2697 2698 err = ddi_dma_alloc_handle(bfe->bfe_dip, &bfe_dma_attr_desc, 2699 DDI_DMA_SLEEP, NULL, &r->r_desc_dma_handle); 2700 2701 if (err != DDI_SUCCESS) { 2702 bfe_error(bfe->bfe_dip, "bfe_ring_desc_alloc() failed on" 2703 " ddi_dma_alloc_handle()"); 2704 kmem_free(r->r_buf_dma, size_krnl); 2705 return (DDI_FAILURE); 2706 } 2707 2708 2709 err = ddi_dma_mem_alloc(r->r_desc_dma_handle, 2710 size_dma, &bfe_buf_attr, 2711 DDI_DMA_CONSISTENT, DDI_DMA_SLEEP, NULL, 2712 &ring, &ring_len, &r->r_desc_acc_handle); 2713 2714 if (err != DDI_SUCCESS) { 2715 bfe_error(bfe->bfe_dip, "bfe_ring_desc_alloc() failed on" 2716 " ddi_dma_mem_alloc()"); 2717 ddi_dma_free_handle(&r->r_desc_dma_handle); 2718 kmem_free(r->r_buf_dma, size_krnl); 2719 return (DDI_FAILURE); 2720 } 2721 2722 err = ddi_dma_addr_bind_handle(r->r_desc_dma_handle, 2723 NULL, ring, ring_len, 2724 DDI_DMA_RDWR | DDI_DMA_CONSISTENT, 2725 DDI_DMA_SLEEP, NULL, 2726 &cookie, &count); 2727 2728 if (err != DDI_SUCCESS) { 2729 bfe_error(bfe->bfe_dip, "bfe_ring_desc_alloc() failed on" 2730 " ddi_dma_addr_bind_handle()"); 2731 ddi_dma_mem_free(&r->r_desc_acc_handle); 2732 ddi_dma_free_handle(&r->r_desc_dma_handle); 2733 kmem_free(r->r_buf_dma, size_krnl); 2734 return (DDI_FAILURE); 2735 } 2736 2737 /* 2738 * We don't want to have multiple cookies. Descriptor should be 2739 * aligned to PAGESIZE boundary. 2740 */ 2741 ASSERT(count == 1); 2742 2743 /* The actual descriptor for the ring */ 2744 r->r_desc_len = ring_len; 2745 r->r_desc_cookie = cookie; 2746 2747 r->r_desc = (void *)ring; 2748 2749 bzero(r->r_desc, size_dma); 2750 bzero(r->r_desc, ring_len); 2751 2752 /* For each descriptor, allocate a DMA buffer */ 2753 fail = 0; 2754 for (i = 0; i < r->r_ndesc; i++) { 2755 if (bfe_ring_buf_alloc(bfe, r, i, d) != DDI_SUCCESS) { 2756 i--; 2757 fail = 1; 2758 break; 2759 } 2760 } 2761 2762 if (fail) { 2763 while (i-- >= 0) { 2764 bfe_ring_buf_free(r, i); 2765 } 2766 2767 /* We don't need the descriptor anymore */ 2768 bfe_ring_desc_free(r); 2769 return (DDI_FAILURE); 2770 } 2771 2772 return (DDI_SUCCESS); 2773 } 2774 2775 static int 2776 bfe_rings_alloc(bfe_t *bfe) 2777 { 2778 /* TX */ 2779 mutex_init(&bfe->bfe_tx_ring.r_lock, NULL, MUTEX_DRIVER, NULL); 2780 bfe->bfe_tx_ring.r_lockp = &bfe->bfe_tx_ring.r_lock; 2781 bfe->bfe_tx_ring.r_buf_len = BFE_MTU + sizeof (struct ether_header) + 2782 VLAN_TAGSZ + ETHERFCSL; 2783 bfe->bfe_tx_ring.r_ndesc = TX_NUM_DESC; 2784 bfe->bfe_tx_ring.r_bfe = bfe; 2785 bfe->bfe_tx_ring.r_avail_desc = TX_NUM_DESC; 2786 2787 /* RX */ 2788 mutex_init(&bfe->bfe_rx_ring.r_lock, NULL, MUTEX_DRIVER, NULL); 2789 bfe->bfe_rx_ring.r_lockp = &bfe->bfe_rx_ring.r_lock; 2790 bfe->bfe_rx_ring.r_buf_len = BFE_MTU + sizeof (struct ether_header) + 2791 VLAN_TAGSZ + ETHERFCSL + RX_HEAD_ROOM; 2792 bfe->bfe_rx_ring.r_ndesc = RX_NUM_DESC; 2793 bfe->bfe_rx_ring.r_bfe = bfe; 2794 bfe->bfe_rx_ring.r_avail_desc = RX_NUM_DESC; 2795 2796 /* Allocate TX Ring */ 2797 if (bfe_ring_desc_alloc(bfe, &bfe->bfe_tx_ring, 2798 DDI_DMA_WRITE) != DDI_SUCCESS) 2799 return (DDI_FAILURE); 2800 2801 /* Allocate RX Ring */ 2802 if (bfe_ring_desc_alloc(bfe, &bfe->bfe_rx_ring, 2803 DDI_DMA_READ) != DDI_SUCCESS) { 2804 cmn_err(CE_NOTE, "RX ring allocation failed"); 2805 bfe_ring_desc_free(&bfe->bfe_tx_ring); 2806 return (DDI_FAILURE); 2807 } 2808 2809 bfe->bfe_tx_ring.r_flags = BFE_RING_ALLOCATED; 2810 bfe->bfe_rx_ring.r_flags = BFE_RING_ALLOCATED; 2811 2812 return (DDI_SUCCESS); 2813 } 2814 2815 static int 2816 bfe_resume(dev_info_t *dip) 2817 { 2818 bfe_t *bfe; 2819 int err = DDI_SUCCESS; 2820 2821 if ((bfe = ddi_get_driver_private(dip)) == NULL) { 2822 bfe_error(dip, "Unexpected error (no driver private data)" 2823 " while resume"); 2824 return (DDI_FAILURE); 2825 } 2826 2827 /* 2828 * Grab all the locks first. 2829 */ 2830 bfe_grab_locks(bfe); 2831 bfe->bfe_chip_state = BFE_CHIP_RESUME; 2832 2833 bfe_init_vars(bfe); 2834 /* PHY will also start running */ 2835 bfe_chip_reset(bfe); 2836 if (bfe_chip_start(bfe) == DDI_FAILURE) { 2837 bfe_error(dip, "Could not resume chip"); 2838 err = DDI_FAILURE; 2839 } 2840 2841 bfe_release_locks(bfe); 2842 2843 if (err == DDI_SUCCESS) 2844 mac_tx_update(bfe->bfe_machdl); 2845 2846 return (err); 2847 } 2848 2849 static int 2850 bfe_attach(dev_info_t *dip, ddi_attach_cmd_t cmd) 2851 { 2852 int unit; 2853 bfe_t *bfe; 2854 mac_register_t *macreg; 2855 int ret; 2856 2857 switch (cmd) { 2858 case DDI_RESUME: 2859 return (bfe_resume(dip)); 2860 2861 case DDI_ATTACH: 2862 break; 2863 2864 default: 2865 return (DDI_FAILURE); 2866 } 2867 2868 2869 unit = ddi_get_instance(dip); 2870 2871 bfe = kmem_zalloc(sizeof (bfe_t), KM_SLEEP); 2872 bfe->bfe_dip = dip; 2873 bfe->bfe_unit = unit; 2874 2875 if (pci_config_setup(dip, &bfe->bfe_conf_handle) != DDI_SUCCESS) { 2876 bfe_error(dip, "pci_config_setup failed"); 2877 goto fail0; 2878 } 2879 2880 /* 2881 * Enable IO space, Bus Master and Memory Space accessess. 2882 */ 2883 ret = pci_config_get16(bfe->bfe_conf_handle, PCI_CONF_COMM); 2884 pci_config_put16(bfe->bfe_conf_handle, PCI_CONF_COMM, 2885 PCI_COMM_IO | PCI_COMM_MAE | PCI_COMM_ME | ret); 2886 2887 ddi_set_driver_private(dip, bfe); 2888 2889 /* Identify hardware */ 2890 if (bfe_identify_hardware(bfe) == BFE_FAILURE) { 2891 bfe_error(dip, "Could not identify device"); 2892 goto fail1; 2893 } 2894 2895 if (bfe_regs_map(bfe) != DDI_SUCCESS) { 2896 bfe_error(dip, "Could not map device registers"); 2897 goto fail1; 2898 } 2899 2900 (void) bfe_get_chip_config(bfe); 2901 2902 /* 2903 * Register with MAC layer 2904 */ 2905 if ((macreg = mac_alloc(MAC_VERSION)) == NULL) { 2906 bfe_error(dip, "mac_alloc() failed"); 2907 goto fail2; 2908 } 2909 2910 macreg->m_type_ident = MAC_PLUGIN_IDENT_ETHER; 2911 macreg->m_driver = bfe; 2912 macreg->m_dip = dip; 2913 macreg->m_instance = unit; 2914 macreg->m_src_addr = bfe->bfe_ether_addr; 2915 macreg->m_callbacks = &bfe_mac_callbacks; 2916 macreg->m_min_sdu = 0; 2917 macreg->m_max_sdu = ETHERMTU; 2918 macreg->m_margin = VLAN_TAGSZ; 2919 2920 if ((ret = mac_register(macreg, &bfe->bfe_machdl)) != 0) { 2921 bfe_error(dip, "mac_register() failed with %d error", ret); 2922 mac_free(macreg); 2923 goto fail2; 2924 } 2925 2926 mac_free(macreg); 2927 2928 rw_init(&bfe->bfe_rwlock, NULL, RW_DRIVER, 2929 DDI_INTR_PRI(bfe->bfe_intrpri)); 2930 2931 if (bfe_add_intr(bfe) != DDI_SUCCESS) { 2932 bfe_error(dip, "Could not add interrupt"); 2933 goto fail3; 2934 } 2935 2936 if (bfe_rings_alloc(bfe) != DDI_SUCCESS) { 2937 bfe_error(dip, "Could not allocate TX/RX Ring"); 2938 goto fail4; 2939 } 2940 2941 /* Init and then reset the chip */ 2942 bfe->bfe_chip_action = 0; 2943 bfe_init_vars(bfe); 2944 2945 /* PHY will also start running */ 2946 bfe_chip_reset(bfe); 2947 2948 /* 2949 * Even though we enable the interrupts here but chip's interrupt 2950 * is not enabled yet. It will be enabled once we plumb the interface. 2951 */ 2952 if (ddi_intr_enable(bfe->bfe_intrhdl) != DDI_SUCCESS) { 2953 bfe_error(dip, "Could not enable interrupt"); 2954 goto fail4; 2955 } 2956 2957 return (DDI_SUCCESS); 2958 2959 fail4: 2960 bfe_remove_intr(bfe); 2961 fail3: 2962 (void) mac_unregister(bfe->bfe_machdl); 2963 fail2: 2964 bfe_unmap_regs(bfe); 2965 fail1: 2966 pci_config_teardown(&bfe->bfe_conf_handle); 2967 fail0: 2968 kmem_free(bfe, sizeof (bfe_t)); 2969 return (DDI_FAILURE); 2970 } 2971 2972 static int 2973 bfe_detach(dev_info_t *devinfo, ddi_detach_cmd_t cmd) 2974 { 2975 bfe_t *bfe; 2976 2977 bfe = ddi_get_driver_private(devinfo); 2978 2979 switch (cmd) { 2980 case DDI_DETACH: 2981 /* 2982 * We need to stop the timer before grabbing locks otherwise 2983 * we can land-up in deadlock with untimeout. 2984 */ 2985 bfe_stop_timer(bfe); 2986 2987 /* 2988 * First unregister with MAC layer before stopping DMA 2989 * engine. 2990 */ 2991 if (mac_unregister(bfe->bfe_machdl) != DDI_SUCCESS) 2992 return (DDI_FAILURE); 2993 2994 bfe->bfe_machdl = NULL; 2995 2996 /* 2997 * Quiesce the chip first. 2998 */ 2999 bfe_grab_locks(bfe); 3000 bfe_chip_halt(bfe); 3001 bfe_stop_phy(bfe); 3002 bfe_release_locks(bfe); 3003 3004 (void) ddi_intr_disable(bfe->bfe_intrhdl); 3005 3006 /* Make sure timer is gone. */ 3007 bfe_stop_timer(bfe); 3008 3009 /* 3010 * Free the DMA resources for buffer and then descriptors 3011 */ 3012 if (bfe->bfe_tx_ring.r_flags == BFE_RING_ALLOCATED) { 3013 /* TX */ 3014 bfe_buffer_free(&bfe->bfe_tx_ring); 3015 bfe_ring_desc_free(&bfe->bfe_tx_ring); 3016 } 3017 3018 if (bfe->bfe_rx_ring.r_flags == BFE_RING_ALLOCATED) { 3019 /* RX */ 3020 bfe_buffer_free(&bfe->bfe_rx_ring); 3021 bfe_ring_desc_free(&bfe->bfe_rx_ring); 3022 } 3023 3024 bfe_remove_intr(bfe); 3025 bfe_unmap_regs(bfe); 3026 pci_config_teardown(&bfe->bfe_conf_handle); 3027 3028 mutex_destroy(&bfe->bfe_tx_ring.r_lock); 3029 mutex_destroy(&bfe->bfe_rx_ring.r_lock); 3030 rw_destroy(&bfe->bfe_rwlock); 3031 3032 kmem_free(bfe, sizeof (bfe_t)); 3033 3034 ddi_set_driver_private(devinfo, NULL); 3035 return (DDI_SUCCESS); 3036 3037 case DDI_SUSPEND: 3038 /* 3039 * We need to stop the timer before grabbing locks otherwise 3040 * we can land-up in deadlock with untimeout. 3041 */ 3042 bfe_stop_timer(bfe); 3043 3044 /* 3045 * Grab all the locks first. 3046 */ 3047 bfe_grab_locks(bfe); 3048 bfe_chip_halt(bfe); 3049 bfe_stop_phy(bfe); 3050 bfe->bfe_chip_state = BFE_CHIP_SUSPENDED; 3051 bfe_release_locks(bfe); 3052 3053 return (DDI_SUCCESS); 3054 3055 default: 3056 return (DDI_FAILURE); 3057 } 3058 } 3059 3060 /* 3061 * Quiesce the card for fast reboot 3062 */ 3063 int 3064 bfe_quiesce(dev_info_t *dev_info) 3065 { 3066 bfe_t *bfe; 3067 3068 bfe = ddi_get_driver_private(dev_info); 3069 3070 bfe_chip_halt(bfe); 3071 bfe_stop_phy(bfe); 3072 bfe->bfe_chip_state = BFE_CHIP_QUIESCED; 3073 3074 return (DDI_SUCCESS); 3075 } 3076 3077 static struct cb_ops bfe_cb_ops = { 3078 nulldev, /* cb_open */ 3079 nulldev, /* cb_close */ 3080 nodev, /* cb_strategy */ 3081 nodev, /* cb_print */ 3082 nodev, /* cb_dump */ 3083 nodev, /* cb_read */ 3084 nodev, /* cb_write */ 3085 nodev, /* cb_ioctl */ 3086 nodev, /* cb_devmap */ 3087 nodev, /* cb_mmap */ 3088 nodev, /* cb_segmap */ 3089 nochpoll, /* cb_chpoll */ 3090 ddi_prop_op, /* cb_prop_op */ 3091 NULL, /* cb_stream */ 3092 D_MP | D_HOTPLUG, /* cb_flag */ 3093 CB_REV, /* cb_rev */ 3094 nodev, /* cb_aread */ 3095 nodev /* cb_awrite */ 3096 }; 3097 3098 static struct dev_ops bfe_dev_ops = { 3099 DEVO_REV, /* devo_rev */ 3100 0, /* devo_refcnt */ 3101 NULL, /* devo_getinfo */ 3102 nulldev, /* devo_identify */ 3103 nulldev, /* devo_probe */ 3104 bfe_attach, /* devo_attach */ 3105 bfe_detach, /* devo_detach */ 3106 nodev, /* devo_reset */ 3107 &bfe_cb_ops, /* devo_cb_ops */ 3108 NULL, /* devo_bus_ops */ 3109 ddi_power, /* devo_power */ 3110 bfe_quiesce /* devo_quiesce */ 3111 }; 3112 3113 static struct modldrv bfe_modldrv = { 3114 &mod_driverops, 3115 bfe_ident, 3116 &bfe_dev_ops 3117 }; 3118 3119 static struct modlinkage modlinkage = { 3120 MODREV_1, { (void *)&bfe_modldrv, NULL } 3121 }; 3122 3123 int 3124 _info(struct modinfo *modinfop) 3125 { 3126 return (mod_info(&modlinkage, modinfop)); 3127 } 3128 3129 int 3130 _init(void) 3131 { 3132 int status; 3133 3134 mac_init_ops(&bfe_dev_ops, MODULE_NAME); 3135 status = mod_install(&modlinkage); 3136 if (status == DDI_FAILURE) 3137 mac_fini_ops(&bfe_dev_ops); 3138 return (status); 3139 } 3140 3141 int 3142 _fini(void) 3143 { 3144 int status; 3145 3146 status = mod_remove(&modlinkage); 3147 if (status == 0) { 3148 mac_fini_ops(&bfe_dev_ops); 3149 } 3150 return (status); 3151 }