1 /* 2 * sfe.c : DP83815/DP83816/SiS900 Fast Ethernet MAC driver for Solaris 3 * 4 * Copyright (c) 2002-2008 Masayuki Murayama. All rights reserved. 5 * 6 * Redistribution and use in source and binary forms, with or without 7 * modification, are permitted provided that the following conditions are met: 8 * 9 * 1. Redistributions of source code must retain the above copyright notice, 10 * this list of conditions and the following disclaimer. 11 * 12 * 2. Redistributions in binary form must reproduce the above copyright notice, 13 * this list of conditions and the following disclaimer in the documentation 14 * and/or other materials provided with the distribution. 15 * 16 * 3. Neither the name of the author nor the names of its contributors may be 17 * used to endorse or promote products derived from this software without 18 * specific prior written permission. 19 * 20 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS 21 * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT 22 * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS 23 * FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE 24 * COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, 25 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, 26 * BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS 27 * OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED 28 * AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, 29 * OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT 30 * OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH 31 * DAMAGE. 32 */ 33 34 /* Avoid undefined symbol for non IA architectures */ 35 #pragma weak inb 36 #pragma weak outb 37 38 /* 39 * Copyright 2009 Sun Microsystems, Inc. All rights reserved. 40 * Use is subject to license terms. 41 */ 42 43 /* 44 * System Header files. 45 */ 46 #include <sys/types.h> 47 #include <sys/conf.h> 48 #include <sys/debug.h> 49 #include <sys/kmem.h> 50 #include <sys/modctl.h> 51 #include <sys/errno.h> 52 #include <sys/ddi.h> 53 #include <sys/sunddi.h> 54 #include <sys/byteorder.h> 55 #include <sys/ethernet.h> 56 #include <sys/pci.h> 57 58 #include "sfe_mii.h" 59 #include "sfe_util.h" 60 #include "sfereg.h" 61 62 char ident[] = "sis900/dp83815 driver v" "2.6.1t30os"; 63 64 /* Debugging support */ 65 #ifdef DEBUG_LEVEL 66 static int sfe_debug = DEBUG_LEVEL; 67 #if DEBUG_LEVEL > 4 68 #define CONS "^" 69 #else 70 #define CONS "!" 71 #endif 72 #define DPRINTF(n, args) if (sfe_debug > (n)) cmn_err args 73 #else 74 #define CONS "!" 75 #define DPRINTF(n, args) 76 #endif 77 78 /* 79 * Useful macros and typedefs 80 */ 81 #define ONESEC (drv_usectohz(1*1000000)) 82 #define ROUNDUP2(x, a) (((x) + (a) - 1) & ~((a) - 1)) 83 84 /* 85 * Our configuration 86 */ 87 #define MAXTXFRAGS 1 88 #define MAXRXFRAGS 1 89 90 #ifndef TX_BUF_SIZE 91 #define TX_BUF_SIZE 64 92 #endif 93 #ifndef TX_RING_SIZE 94 #if MAXTXFRAGS == 1 95 #define TX_RING_SIZE TX_BUF_SIZE 96 #else 97 #define TX_RING_SIZE (TX_BUF_SIZE * 4) 98 #endif 99 #endif 100 101 #ifndef RX_BUF_SIZE 102 #define RX_BUF_SIZE 256 103 #endif 104 #ifndef RX_RING_SIZE 105 #define RX_RING_SIZE RX_BUF_SIZE 106 #endif 107 108 #define OUR_INTR_BITS \ 109 (ISR_DPERR | ISR_SSERR | ISR_RMABT | ISR_RTABT | ISR_RXSOVR | \ 110 ISR_TXURN | ISR_TXDESC | ISR_TXERR | \ 111 ISR_RXORN | ISR_RXIDLE | ISR_RXOK | ISR_RXERR) 112 113 #define USE_MULTICAST_HASHTBL 114 115 static int sfe_tx_copy_thresh = 256; 116 static int sfe_rx_copy_thresh = 256; 117 118 /* special PHY registers for SIS900 */ 119 #define MII_CONFIG1 0x0010 120 #define MII_CONFIG2 0x0011 121 #define MII_MASK 0x0013 122 #define MII_RESV 0x0014 123 124 #define PHY_MASK 0xfffffff0 125 #define PHY_SIS900_INTERNAL 0x001d8000 126 #define PHY_ICS1893 0x0015f440 127 128 129 #define SFE_DESC_SIZE 16 /* including pads rounding up to power of 2 */ 130 131 /* 132 * Supported chips 133 */ 134 struct chip_info { 135 uint16_t venid; 136 uint16_t devid; 137 char *chip_name; 138 int chip_type; 139 #define CHIPTYPE_DP83815 0 140 #define CHIPTYPE_SIS900 1 141 }; 142 143 /* 144 * Chip dependent MAC state 145 */ 146 struct sfe_dev { 147 /* misc HW information */ 148 struct chip_info *chip; 149 uint32_t our_intr_bits; 150 uint32_t isr_pended; 151 uint32_t cr; 152 uint_t tx_drain_threshold; 153 uint_t tx_fill_threshold; 154 uint_t rx_drain_threshold; 155 uint_t rx_fill_threshold; 156 uint8_t revid; /* revision from PCI configuration */ 157 boolean_t (*get_mac_addr)(struct gem_dev *); 158 uint8_t mac_addr[ETHERADDRL]; 159 uint8_t bridge_revid; 160 }; 161 162 /* 163 * Hardware information 164 */ 165 struct chip_info sfe_chiptbl[] = { 166 { 0x1039, 0x0900, "SiS900", CHIPTYPE_SIS900, }, 167 { 0x100b, 0x0020, "DP83815/83816", CHIPTYPE_DP83815, }, 168 { 0x1039, 0x7016, "SiS7016", CHIPTYPE_SIS900, }, 169 }; 170 #define CHIPTABLESIZE (sizeof (sfe_chiptbl)/sizeof (struct chip_info)) 171 172 /* ======================================================== */ 173 174 /* mii operations */ 175 static void sfe_mii_sync_dp83815(struct gem_dev *); 176 static void sfe_mii_sync_sis900(struct gem_dev *); 177 static uint16_t sfe_mii_read_dp83815(struct gem_dev *, uint_t); 178 static uint16_t sfe_mii_read_sis900(struct gem_dev *, uint_t); 179 static void sfe_mii_write_dp83815(struct gem_dev *, uint_t, uint16_t); 180 static void sfe_mii_write_sis900(struct gem_dev *, uint_t, uint16_t); 181 static void sfe_set_eq_sis630(struct gem_dev *dp); 182 /* nic operations */ 183 static int sfe_reset_chip_sis900(struct gem_dev *); 184 static int sfe_reset_chip_dp83815(struct gem_dev *); 185 static int sfe_init_chip(struct gem_dev *); 186 static int sfe_start_chip(struct gem_dev *); 187 static int sfe_stop_chip(struct gem_dev *); 188 static int sfe_set_media(struct gem_dev *); 189 static int sfe_set_rx_filter_dp83815(struct gem_dev *); 190 static int sfe_set_rx_filter_sis900(struct gem_dev *); 191 static int sfe_get_stats(struct gem_dev *); 192 static int sfe_attach_chip(struct gem_dev *); 193 194 /* descriptor operations */ 195 static int sfe_tx_desc_write(struct gem_dev *dp, int slot, 196 ddi_dma_cookie_t *dmacookie, int frags, uint64_t flags); 197 static void sfe_tx_start(struct gem_dev *dp, int startslot, int nslot); 198 static void sfe_rx_desc_write(struct gem_dev *dp, int slot, 199 ddi_dma_cookie_t *dmacookie, int frags); 200 static uint_t sfe_tx_desc_stat(struct gem_dev *dp, int slot, int ndesc); 201 static uint64_t sfe_rx_desc_stat(struct gem_dev *dp, int slot, int ndesc); 202 203 static void sfe_tx_desc_init(struct gem_dev *dp, int slot); 204 static void sfe_rx_desc_init(struct gem_dev *dp, int slot); 205 static void sfe_tx_desc_clean(struct gem_dev *dp, int slot); 206 static void sfe_rx_desc_clean(struct gem_dev *dp, int slot); 207 208 /* interrupt handler */ 209 static uint_t sfe_interrupt(struct gem_dev *dp); 210 211 /* ======================================================== */ 212 213 /* mapping attributes */ 214 /* Data access requirements. */ 215 static struct ddi_device_acc_attr sfe_dev_attr = { 216 DDI_DEVICE_ATTR_V0, 217 DDI_STRUCTURE_LE_ACC, 218 DDI_STRICTORDER_ACC 219 }; 220 221 /* On sparc, Buffers should be native endian for speed */ 222 static struct ddi_device_acc_attr sfe_buf_attr = { 223 DDI_DEVICE_ATTR_V0, 224 DDI_NEVERSWAP_ACC, /* native endianness */ 225 DDI_STRICTORDER_ACC 226 }; 227 228 static ddi_dma_attr_t sfe_dma_attr_buf = { 229 DMA_ATTR_V0, /* dma_attr_version */ 230 0, /* dma_attr_addr_lo */ 231 0xffffffffull, /* dma_attr_addr_hi */ 232 0x00000fffull, /* dma_attr_count_max */ 233 0, /* patched later */ /* dma_attr_align */ 234 0x000003fc, /* dma_attr_burstsizes */ 235 1, /* dma_attr_minxfer */ 236 0x00000fffull, /* dma_attr_maxxfer */ 237 0xffffffffull, /* dma_attr_seg */ 238 0, /* patched later */ /* dma_attr_sgllen */ 239 1, /* dma_attr_granular */ 240 0 /* dma_attr_flags */ 241 }; 242 243 static ddi_dma_attr_t sfe_dma_attr_desc = { 244 DMA_ATTR_V0, /* dma_attr_version */ 245 16, /* dma_attr_addr_lo */ 246 0xffffffffull, /* dma_attr_addr_hi */ 247 0xffffffffull, /* dma_attr_count_max */ 248 16, /* dma_attr_align */ 249 0x000003fc, /* dma_attr_burstsizes */ 250 1, /* dma_attr_minxfer */ 251 0xffffffffull, /* dma_attr_maxxfer */ 252 0xffffffffull, /* dma_attr_seg */ 253 1, /* dma_attr_sgllen */ 254 1, /* dma_attr_granular */ 255 0 /* dma_attr_flags */ 256 }; 257 258 uint32_t sfe_use_pcimemspace = 0; 259 260 /* ======================================================== */ 261 /* 262 * HW manipulation routines 263 */ 264 /* ======================================================== */ 265 266 #define SFE_EEPROM_DELAY(dp) \ 267 { (void) INL(dp, EROMAR); (void) INL(dp, EROMAR); } 268 #define EE_CMD_READ 6 269 #define EE_CMD_SHIFT 6 270 271 static uint16_t 272 sfe_read_eeprom(struct gem_dev *dp, uint_t offset) 273 { 274 int eedi; 275 int i; 276 uint16_t ret; 277 278 /* ensure de-assert chip select */ 279 OUTL(dp, EROMAR, 0); 280 SFE_EEPROM_DELAY(dp); 281 OUTL(dp, EROMAR, EROMAR_EESK); 282 SFE_EEPROM_DELAY(dp); 283 284 /* assert chip select */ 285 offset |= EE_CMD_READ << EE_CMD_SHIFT; 286 287 for (i = 8; i >= 0; i--) { 288 /* make command */ 289 eedi = ((offset >> i) & 1) << EROMAR_EEDI_SHIFT; 290 291 /* send 1 bit */ 292 OUTL(dp, EROMAR, EROMAR_EECS | eedi); 293 SFE_EEPROM_DELAY(dp); 294 OUTL(dp, EROMAR, EROMAR_EECS | eedi | EROMAR_EESK); 295 SFE_EEPROM_DELAY(dp); 296 } 297 298 OUTL(dp, EROMAR, EROMAR_EECS); 299 300 ret = 0; 301 for (i = 0; i < 16; i++) { 302 /* Get 1 bit */ 303 OUTL(dp, EROMAR, EROMAR_EECS); 304 SFE_EEPROM_DELAY(dp); 305 OUTL(dp, EROMAR, EROMAR_EECS | EROMAR_EESK); 306 SFE_EEPROM_DELAY(dp); 307 308 ret = (ret << 1) | ((INL(dp, EROMAR) >> EROMAR_EEDO_SHIFT) & 1); 309 } 310 311 OUTL(dp, EROMAR, 0); 312 SFE_EEPROM_DELAY(dp); 313 314 return (ret); 315 } 316 #undef SFE_EEPROM_DELAY 317 318 static boolean_t 319 sfe_get_mac_addr_dp83815(struct gem_dev *dp) 320 { 321 uint8_t *mac; 322 uint_t val; 323 int i; 324 325 #define BITSET(p, ix, v) (p)[(ix)/8] |= ((v) ? 1 : 0) << ((ix) & 0x7) 326 327 DPRINTF(4, (CE_CONT, CONS "%s: %s: called", dp->name, __func__)); 328 329 mac = dp->dev_addr.ether_addr_octet; 330 331 /* first of all, clear MAC address buffer */ 332 bzero(mac, ETHERADDRL); 333 334 /* get bit 0 */ 335 val = sfe_read_eeprom(dp, 0x6); 336 BITSET(mac, 0, val & 1); 337 338 /* get bit 1 - 16 */ 339 val = sfe_read_eeprom(dp, 0x7); 340 for (i = 0; i < 16; i++) { 341 BITSET(mac, 1 + i, val & (1 << (15 - i))); 342 } 343 344 /* get bit 17 - 32 */ 345 val = sfe_read_eeprom(dp, 0x8); 346 for (i = 0; i < 16; i++) { 347 BITSET(mac, 17 + i, val & (1 << (15 - i))); 348 } 349 350 /* get bit 33 - 47 */ 351 val = sfe_read_eeprom(dp, 0x9); 352 for (i = 0; i < 15; i++) { 353 BITSET(mac, 33 + i, val & (1 << (15 - i))); 354 } 355 356 return (B_TRUE); 357 #undef BITSET 358 } 359 360 static boolean_t 361 sfe_get_mac_addr_sis900(struct gem_dev *dp) 362 { 363 uint_t val; 364 int i; 365 uint8_t *mac; 366 367 mac = dp->dev_addr.ether_addr_octet; 368 369 for (i = 0; i < ETHERADDRL/2; i++) { 370 val = sfe_read_eeprom(dp, 0x8 + i); 371 *mac++ = (uint8_t)val; 372 *mac++ = (uint8_t)(val >> 8); 373 } 374 375 return (B_TRUE); 376 } 377 378 static dev_info_t * 379 sfe_search_pci_dev_subr(dev_info_t *cur_node, int vendor_id, int device_id) 380 { 381 dev_info_t *child_id; 382 dev_info_t *ret; 383 int vid, did; 384 385 if (cur_node == NULL) { 386 return (NULL); 387 } 388 389 /* check brothers */ 390 do { 391 vid = ddi_prop_get_int(DDI_DEV_T_ANY, cur_node, 392 DDI_PROP_DONTPASS, "vendor-id", -1); 393 did = ddi_prop_get_int(DDI_DEV_T_ANY, cur_node, 394 DDI_PROP_DONTPASS, "device-id", -1); 395 396 if (vid == vendor_id && did == device_id) { 397 /* found */ 398 return (cur_node); 399 } 400 401 /* check children */ 402 if ((child_id = ddi_get_child(cur_node)) != NULL) { 403 if ((ret = sfe_search_pci_dev_subr(child_id, 404 vendor_id, device_id)) != NULL) { 405 return (ret); 406 } 407 } 408 409 } while ((cur_node = ddi_get_next_sibling(cur_node)) != NULL); 410 411 /* not found */ 412 return (NULL); 413 } 414 415 static dev_info_t * 416 sfe_search_pci_dev(int vendor_id, int device_id) 417 { 418 return (sfe_search_pci_dev_subr(ddi_root_node(), vendor_id, device_id)); 419 } 420 421 static boolean_t 422 sfe_get_mac_addr_sis630e(struct gem_dev *dp) 423 { 424 int i; 425 dev_info_t *isa_bridge; 426 ddi_acc_handle_t isa_handle; 427 int reg; 428 429 if (inb == NULL || outb == NULL) { 430 /* this is not IA architecture */ 431 return (B_FALSE); 432 } 433 434 if ((isa_bridge = sfe_search_pci_dev(0x1039, 0x8)) == NULL) { 435 cmn_err(CE_WARN, "%s: failed to find isa-bridge pci1039,8", 436 dp->name); 437 return (B_FALSE); 438 } 439 440 if (pci_config_setup(isa_bridge, &isa_handle) != DDI_SUCCESS) { 441 cmn_err(CE_WARN, "%s: ddi_regs_map_setup failed", 442 dp->name); 443 return (B_FALSE); 444 } 445 446 /* enable to access CMOS RAM */ 447 reg = pci_config_get8(isa_handle, 0x48); 448 pci_config_put8(isa_handle, 0x48, reg | 0x40); 449 450 for (i = 0; i < ETHERADDRL; i++) { 451 outb(0x70, 0x09 + i); 452 dp->dev_addr.ether_addr_octet[i] = inb(0x71); 453 } 454 455 /* disable to access CMOS RAM */ 456 pci_config_put8(isa_handle, 0x48, reg); 457 pci_config_teardown(&isa_handle); 458 459 return (B_TRUE); 460 } 461 462 static boolean_t 463 sfe_get_mac_addr_sis635(struct gem_dev *dp) 464 { 465 int i; 466 uint32_t rfcr; 467 uint16_t v; 468 struct sfe_dev *lp = dp->private; 469 470 DPRINTF(2, (CE_CONT, CONS "%s: %s: called", dp->name, __func__)); 471 rfcr = INL(dp, RFCR); 472 473 OUTL(dp, CR, lp->cr | CR_RELOAD); 474 OUTL(dp, CR, lp->cr); 475 476 /* disable packet filtering before reading filter */ 477 OUTL(dp, RFCR, rfcr & ~RFCR_RFEN); 478 479 /* load MAC addr from filter data register */ 480 for (i = 0; i < ETHERADDRL; i += 2) { 481 OUTL(dp, RFCR, 482 (RFADDR_MAC_SIS900 + (i/2)) << RFCR_RFADDR_SHIFT_SIS900); 483 v = INL(dp, RFDR); 484 dp->dev_addr.ether_addr_octet[i] = (uint8_t)v; 485 dp->dev_addr.ether_addr_octet[i+1] = (uint8_t)(v >> 8); 486 } 487 488 /* re-enable packet filtering */ 489 OUTL(dp, RFCR, rfcr | RFCR_RFEN); 490 491 return (B_TRUE); 492 } 493 494 static boolean_t 495 sfe_get_mac_addr_sis962(struct gem_dev *dp) 496 { 497 boolean_t ret; 498 int i; 499 500 ret = B_FALSE; 501 502 /* rise request signal to access EEPROM */ 503 OUTL(dp, MEAR, EROMAR_EEREQ); 504 for (i = 0; (INL(dp, MEAR) & EROMAR_EEGNT) == 0; i++) { 505 if (i > 200) { 506 /* failed to acquire eeprom */ 507 cmn_err(CE_NOTE, 508 CONS "%s: failed to access eeprom", dp->name); 509 goto x; 510 } 511 drv_usecwait(10); 512 } 513 ret = sfe_get_mac_addr_sis900(dp); 514 x: 515 /* release EEPROM */ 516 OUTL(dp, MEAR, EROMAR_EEDONE); 517 518 return (ret); 519 } 520 521 static int 522 sfe_reset_chip_sis900(struct gem_dev *dp) 523 { 524 int i; 525 uint32_t done; 526 uint32_t val; 527 struct sfe_dev *lp = dp->private; 528 529 DPRINTF(4, (CE_CONT, CONS "%s: %s called", dp->name, __func__)); 530 531 /* invalidate mac addr cache */ 532 bzero(lp->mac_addr, sizeof (lp->mac_addr)); 533 534 lp->cr = 0; 535 536 /* inhibit interrupt */ 537 OUTL(dp, IMR, 0); 538 lp->isr_pended |= INL(dp, ISR) & lp->our_intr_bits; 539 540 OUTLINL(dp, RFCR, 0); 541 542 OUTL(dp, CR, CR_RST | CR_TXR | CR_RXR); 543 drv_usecwait(10); 544 545 done = 0; 546 for (i = 0; done != (ISR_TXRCMP | ISR_RXRCMP); i++) { 547 if (i > 1000) { 548 cmn_err(CE_WARN, "%s: chip reset timeout", dp->name); 549 return (GEM_FAILURE); 550 } 551 done |= INL(dp, ISR) & (ISR_TXRCMP | ISR_RXRCMP); 552 drv_usecwait(10); 553 } 554 555 if (lp->revid == SIS630ET_900_REV) { 556 lp->cr |= CR_ACCESSMODE; 557 OUTL(dp, CR, lp->cr | INL(dp, CR)); 558 } 559 560 /* Configuration register: enable PCI parity */ 561 DPRINTF(2, (CE_CONT, CONS "%s: cfg:%b", 562 dp->name, INL(dp, CFG), CFG_BITS_SIS900)); 563 val = 0; 564 if (lp->revid >= SIS635A_900_REV || 565 lp->revid == SIS900B_900_REV) { 566 /* what is this ? */ 567 val |= CFG_RND_CNT; 568 } 569 OUTL(dp, CFG, val); 570 DPRINTF(2, (CE_CONT, CONS "%s: cfg:%b", dp->name, 571 INL(dp, CFG), CFG_BITS_SIS900)); 572 573 return (GEM_SUCCESS); 574 } 575 576 static int 577 sfe_reset_chip_dp83815(struct gem_dev *dp) 578 { 579 int i; 580 uint32_t val; 581 struct sfe_dev *lp = dp->private; 582 583 DPRINTF(4, (CE_CONT, CONS "%s: %s called", dp->name, __func__)); 584 585 /* invalidate mac addr cache */ 586 bzero(lp->mac_addr, sizeof (lp->mac_addr)); 587 588 lp->cr = 0; 589 590 /* inhibit interrupts */ 591 OUTL(dp, IMR, 0); 592 lp->isr_pended |= INL(dp, ISR) & lp->our_intr_bits; 593 594 OUTL(dp, RFCR, 0); 595 596 OUTL(dp, CR, CR_RST); 597 drv_usecwait(10); 598 599 for (i = 0; INL(dp, CR) & CR_RST; i++) { 600 if (i > 100) { 601 cmn_err(CE_WARN, "!%s: chip reset timeout", dp->name); 602 return (GEM_FAILURE); 603 } 604 drv_usecwait(10); 605 } 606 DPRINTF(0, (CE_CONT, "!%s: chip reset in %duS", dp->name, i*10)); 607 608 OUTL(dp, CCSR, CCSR_PMESTS); 609 OUTL(dp, CCSR, 0); 610 611 /* Configuration register: enable PCI parity */ 612 DPRINTF(2, (CE_CONT, CONS "%s: cfg:%b", 613 dp->name, INL(dp, CFG), CFG_BITS_DP83815)); 614 val = INL(dp, CFG) & (CFG_ANEG_SEL | CFG_PHY_CFG); 615 OUTL(dp, CFG, val | CFG_PAUSE_ADV); 616 DPRINTF(2, (CE_CONT, CONS "%s: cfg:%b", dp->name, 617 INL(dp, CFG), CFG_BITS_DP83815)); 618 619 return (GEM_SUCCESS); 620 } 621 622 static int 623 sfe_init_chip(struct gem_dev *dp) 624 { 625 /* Configuration register: have been set up in sfe_chip_reset */ 626 627 /* PCI test control register: do nothing */ 628 629 /* Interrupt status register : do nothing */ 630 631 /* Interrupt mask register: clear, but leave lp->our_intr_bits */ 632 OUTL(dp, IMR, 0); 633 634 /* Enhanced PHY Access register (sis900): do nothing */ 635 636 /* Transmit Descriptor Pointer register: base addr of TX ring */ 637 OUTL(dp, TXDP, dp->tx_ring_dma); 638 639 /* Receive descriptor pointer register: base addr of RX ring */ 640 OUTL(dp, RXDP, dp->rx_ring_dma); 641 642 return (GEM_SUCCESS); 643 } 644 645 static uint_t 646 sfe_mcast_hash(struct gem_dev *dp, uint8_t *addr) 647 { 648 return (gem_ether_crc_be(addr, ETHERADDRL)); 649 } 650 651 #ifdef DEBUG_LEVEL 652 static void 653 sfe_rxfilter_dump(struct gem_dev *dp, int start, int end) 654 { 655 int i; 656 int j; 657 uint16_t ram[0x10]; 658 659 cmn_err(CE_CONT, "!%s: rx filter ram dump:", dp->name); 660 #define WORDS_PER_LINE 4 661 for (i = start; i < end; i += WORDS_PER_LINE*2) { 662 for (j = 0; j < WORDS_PER_LINE; j++) { 663 OUTL(dp, RFCR, RFADDR_MAC_DP83815 + i + j*2); 664 ram[j] = INL(dp, RFDR); 665 } 666 667 cmn_err(CE_CONT, "!0x%02x: 0x%04x 0x%04x 0x%04x 0x%04x", 668 i, ram[0], ram[1], ram[2], ram[3]); 669 } 670 671 #undef WORDS_PER_LINE 672 } 673 #endif 674 675 static uint_t sfe_rf_perfect_base_dp83815[] = { 676 RFADDR_PMATCH0_DP83815, 677 RFADDR_PMATCH1_DP83815, 678 RFADDR_PMATCH2_DP83815, 679 RFADDR_PMATCH3_DP83815, 680 }; 681 682 static int 683 sfe_set_rx_filter_dp83815(struct gem_dev *dp) 684 { 685 int i; 686 int j; 687 uint32_t mode; 688 uint8_t *mac = dp->cur_addr.ether_addr_octet; 689 uint16_t hash_tbl[32]; 690 struct sfe_dev *lp = dp->private; 691 692 DPRINTF(1, (CE_CONT, CONS "%s: %s: called, mc_count:%d, mode:0x%b", 693 dp->name, __func__, dp->mc_count, dp->rxmode, RXMODE_BITS)); 694 695 #if DEBUG_LEVEL > 0 696 for (i = 0; i < dp->mc_count; i++) { 697 cmn_err(CE_CONT, 698 "!%s: adding mcast(%d) %02x:%02x:%02x:%02x:%02x:%02x", 699 dp->name, i, 700 dp->mc_list[i].addr.ether_addr_octet[0], 701 dp->mc_list[i].addr.ether_addr_octet[1], 702 dp->mc_list[i].addr.ether_addr_octet[2], 703 dp->mc_list[i].addr.ether_addr_octet[3], 704 dp->mc_list[i].addr.ether_addr_octet[4], 705 dp->mc_list[i].addr.ether_addr_octet[5]); 706 } 707 #endif 708 if ((dp->rxmode & RXMODE_ENABLE) == 0) { 709 /* disable rx filter */ 710 OUTL(dp, RFCR, 0); 711 return (GEM_SUCCESS); 712 } 713 714 /* 715 * Set Receive filter control register 716 */ 717 if (dp->rxmode & RXMODE_PROMISC) { 718 /* all broadcast, all multicast, all physical */ 719 mode = RFCR_AAB | RFCR_AAM | RFCR_AAP; 720 } else if ((dp->rxmode & RXMODE_ALLMULTI) || dp->mc_count > 16*32/2) { 721 /* all broadcast, all multicast, physical for the chip */ 722 mode = RFCR_AAB | RFCR_AAM | RFCR_APM_DP83815; 723 } else if (dp->mc_count > 4) { 724 /* 725 * Use multicast hash table, 726 * accept all broadcast and physical for the chip. 727 */ 728 mode = RFCR_AAB | RFCR_MHEN_DP83815 | RFCR_APM_DP83815; 729 730 bzero(hash_tbl, sizeof (hash_tbl)); 731 for (i = 0; i < dp->mc_count; i++) { 732 j = dp->mc_list[i].hash >> (32 - 9); 733 hash_tbl[j / 16] |= 1 << (j % 16); 734 } 735 } else { 736 /* 737 * Use pattern mach filter for multicast address, 738 * accept all broadcast and physical for the chip 739 */ 740 /* need to enable corresponding pattern registers */ 741 mode = RFCR_AAB | RFCR_APM_DP83815 | 742 (((1 << dp->mc_count) - 1) << RFCR_APAT_SHIFT); 743 } 744 745 #if DEBUG_LEVEL > 1 746 cmn_err(CE_CONT, 747 "!%s: mac %02x:%02x:%02x:%02x:%02x:%02x" 748 " cache %02x:%02x:%02x:%02x:%02x:%02x", 749 dp->name, mac[0], mac[1], mac[2], mac[3], mac[4], mac[5], 750 lp->mac_addr[0], lp->mac_addr[1], 751 lp->mac_addr[2], lp->mac_addr[3], 752 lp->mac_addr[4], lp->mac_addr[5]); 753 #endif 754 if (bcmp(mac, lp->mac_addr, ETHERADDRL) != 0) { 755 /* 756 * XXX - need to *disable* rx filter to load mac address for 757 * the chip. otherwise, we cannot setup rxfilter correctly. 758 */ 759 /* setup perfect match register for my station address */ 760 for (i = 0; i < ETHERADDRL; i += 2) { 761 OUTL(dp, RFCR, RFADDR_MAC_DP83815 + i); 762 OUTL(dp, RFDR, (mac[i+1] << 8) | mac[i]); 763 } 764 765 bcopy(mac, lp->mac_addr, ETHERADDRL); 766 } 767 768 #if DEBUG_LEVEL > 3 769 /* clear pattern ram */ 770 for (j = 0x200; j < 0x380; j += 2) { 771 OUTL(dp, RFCR, j); 772 OUTL(dp, RFDR, 0); 773 } 774 #endif 775 if (mode & RFCR_APAT_DP83815) { 776 /* setup multicast address into pattern match registers */ 777 for (j = 0; j < dp->mc_count; j++) { 778 mac = &dp->mc_list[j].addr.ether_addr_octet[0]; 779 for (i = 0; i < ETHERADDRL; i += 2) { 780 OUTL(dp, RFCR, 781 sfe_rf_perfect_base_dp83815[j] + i*2); 782 OUTL(dp, RFDR, (mac[i+1] << 8) | mac[i]); 783 } 784 } 785 786 /* setup pattern count registers */ 787 OUTL(dp, RFCR, RFADDR_PCOUNT01_DP83815); 788 OUTL(dp, RFDR, (ETHERADDRL << 8) | ETHERADDRL); 789 OUTL(dp, RFCR, RFADDR_PCOUNT23_DP83815); 790 OUTL(dp, RFDR, (ETHERADDRL << 8) | ETHERADDRL); 791 } 792 793 if (mode & RFCR_MHEN_DP83815) { 794 /* Load Multicast hash table */ 795 for (i = 0; i < 32; i++) { 796 /* for DP83815, index is in byte */ 797 OUTL(dp, RFCR, RFADDR_MULTICAST_DP83815 + i*2); 798 OUTL(dp, RFDR, hash_tbl[i]); 799 } 800 } 801 #if DEBUG_LEVEL > 2 802 sfe_rxfilter_dump(dp, 0, 0x10); 803 sfe_rxfilter_dump(dp, 0x200, 0x380); 804 #endif 805 /* Set rx filter mode and enable rx filter */ 806 OUTL(dp, RFCR, RFCR_RFEN | mode); 807 808 return (GEM_SUCCESS); 809 } 810 811 static int 812 sfe_set_rx_filter_sis900(struct gem_dev *dp) 813 { 814 int i; 815 uint32_t mode; 816 uint16_t hash_tbl[16]; 817 uint8_t *mac = dp->cur_addr.ether_addr_octet; 818 int hash_size; 819 int hash_shift; 820 struct sfe_dev *lp = dp->private; 821 822 DPRINTF(4, (CE_CONT, CONS "%s: %s: called", dp->name, __func__)); 823 824 if ((dp->rxmode & RXMODE_ENABLE) == 0) { 825 /* disable rx filter */ 826 OUTLINL(dp, RFCR, 0); 827 return (GEM_SUCCESS); 828 } 829 830 /* 831 * determine hardware hash table size in word. 832 */ 833 hash_shift = 25; 834 if (lp->revid >= SIS635A_900_REV || lp->revid == SIS900B_900_REV) { 835 hash_shift = 24; 836 } 837 hash_size = (1 << (32 - hash_shift)) / 16; 838 bzero(hash_tbl, sizeof (hash_tbl)); 839 840 /* Set Receive filter control register */ 841 842 if (dp->rxmode & RXMODE_PROMISC) { 843 /* all broadcast, all multicast, all physical */ 844 mode = RFCR_AAB | RFCR_AAM | RFCR_AAP; 845 } else if ((dp->rxmode & RXMODE_ALLMULTI) || 846 dp->mc_count > hash_size*16/2) { 847 /* all broadcast, all multicast, physical for the chip */ 848 mode = RFCR_AAB | RFCR_AAM; 849 } else { 850 /* all broadcast, physical for the chip */ 851 mode = RFCR_AAB; 852 } 853 854 /* make hash table */ 855 for (i = 0; i < dp->mc_count; i++) { 856 uint_t h; 857 h = dp->mc_list[i].hash >> hash_shift; 858 hash_tbl[h / 16] |= 1 << (h % 16); 859 } 860 861 if (bcmp(mac, lp->mac_addr, ETHERADDRL) != 0) { 862 /* Disable Rx filter and load mac address */ 863 for (i = 0; i < ETHERADDRL/2; i++) { 864 /* For sis900, index is in word */ 865 OUTLINL(dp, RFCR, 866 (RFADDR_MAC_SIS900+i) << RFCR_RFADDR_SHIFT_SIS900); 867 OUTLINL(dp, RFDR, (mac[i*2+1] << 8) | mac[i*2]); 868 } 869 870 bcopy(mac, lp->mac_addr, ETHERADDRL); 871 } 872 873 /* Load Multicast hash table */ 874 for (i = 0; i < hash_size; i++) { 875 /* For sis900, index is in word */ 876 OUTLINL(dp, RFCR, 877 (RFADDR_MULTICAST_SIS900 + i) << RFCR_RFADDR_SHIFT_SIS900); 878 OUTLINL(dp, RFDR, hash_tbl[i]); 879 } 880 881 /* Load rx filter mode and enable rx filter */ 882 OUTLINL(dp, RFCR, RFCR_RFEN | mode); 883 884 return (GEM_SUCCESS); 885 } 886 887 static int 888 sfe_start_chip(struct gem_dev *dp) 889 { 890 struct sfe_dev *lp = dp->private; 891 892 DPRINTF(4, (CE_CONT, CONS "%s: %s: called", dp->name, __func__)); 893 894 /* 895 * setup interrupt mask, which shouldn't include ISR_TOK 896 * to improve performance. 897 */ 898 lp->our_intr_bits = OUR_INTR_BITS; 899 900 /* enable interrupt */ 901 if ((dp->misc_flag & GEM_NOINTR) == 0) { 902 OUTL(dp, IER, 1); 903 OUTL(dp, IMR, lp->our_intr_bits); 904 } 905 906 /* Kick RX */ 907 OUTL(dp, CR, lp->cr | CR_RXE); 908 909 return (GEM_SUCCESS); 910 } 911 912 /* 913 * Stop nic core gracefully. 914 */ 915 static int 916 sfe_stop_chip(struct gem_dev *dp) 917 { 918 struct sfe_dev *lp = dp->private; 919 uint32_t done; 920 int i; 921 uint32_t val; 922 923 DPRINTF(4, (CE_CONT, CONS "%s: %s: called", dp->name, __func__)); 924 925 /* 926 * Although we inhibit interrupt here, we don't clear soft copy of 927 * interrupt mask to avoid bogus interrupts. 928 */ 929 OUTL(dp, IMR, 0); 930 931 /* stop TX and RX immediately */ 932 OUTL(dp, CR, lp->cr | CR_TXR | CR_RXR); 933 934 done = 0; 935 for (i = 0; done != (ISR_RXRCMP | ISR_TXRCMP); i++) { 936 if (i > 1000) { 937 /* 938 * As gem layer will call sfe_reset_chip(), 939 * we don't neet to reset futher 940 */ 941 cmn_err(CE_NOTE, "!%s: %s: Tx/Rx reset timeout", 942 dp->name, __func__); 943 944 return (GEM_FAILURE); 945 } 946 val = INL(dp, ISR); 947 done |= val & (ISR_RXRCMP | ISR_TXRCMP); 948 lp->isr_pended |= val & lp->our_intr_bits; 949 drv_usecwait(10); 950 } 951 952 return (GEM_SUCCESS); 953 } 954 955 #ifndef __sparc 956 /* 957 * Stop nic core gracefully for quiesce 958 */ 959 static int 960 sfe_stop_chip_quiesce(struct gem_dev *dp) 961 { 962 struct sfe_dev *lp = dp->private; 963 uint32_t done; 964 int i; 965 uint32_t val; 966 967 /* 968 * Although we inhibit interrupt here, we don't clear soft copy of 969 * interrupt mask to avoid bogus interrupts. 970 */ 971 OUTL(dp, IMR, 0); 972 973 /* stop TX and RX immediately */ 974 OUTL(dp, CR, CR_TXR | CR_RXR); 975 976 done = 0; 977 for (i = 0; done != (ISR_RXRCMP | ISR_TXRCMP); i++) { 978 if (i > 1000) { 979 /* 980 * As gem layer will call sfe_reset_chip(), 981 * we don't neet to reset futher 982 */ 983 984 return (DDI_FAILURE); 985 } 986 val = INL(dp, ISR); 987 done |= val & (ISR_RXRCMP | ISR_TXRCMP); 988 lp->isr_pended |= val & lp->our_intr_bits; 989 drv_usecwait(10); 990 } 991 return (DDI_SUCCESS); 992 } 993 #endif 994 995 /* 996 * Setup media mode 997 */ 998 static uint_t 999 sfe_mxdma_value[] = { 512, 4, 8, 16, 32, 64, 128, 256, }; 1000 1001 static uint_t 1002 sfe_encode_mxdma(uint_t burstsize) 1003 { 1004 int i; 1005 1006 if (burstsize > 256) { 1007 /* choose 512 */ 1008 return (0); 1009 } 1010 1011 for (i = 1; i < 8; i++) { 1012 if (burstsize <= sfe_mxdma_value[i]) { 1013 break; 1014 } 1015 } 1016 return (i); 1017 } 1018 1019 static int 1020 sfe_set_media(struct gem_dev *dp) 1021 { 1022 uint32_t txcfg; 1023 uint32_t rxcfg; 1024 uint32_t pcr; 1025 uint32_t val; 1026 uint32_t txmxdma; 1027 uint32_t rxmxdma; 1028 struct sfe_dev *lp = dp->private; 1029 #ifdef DEBUG_LEVEL 1030 extern int gem_speed_value[]; 1031 #endif 1032 DPRINTF(2, (CE_CONT, CONS "%s: %s: %s duplex, %d Mbps", 1033 dp->name, __func__, 1034 dp->full_duplex ? "full" : "half", gem_speed_value[dp->speed])); 1035 1036 /* initialize txcfg and rxcfg */ 1037 txcfg = TXCFG_ATP; 1038 if (dp->full_duplex) { 1039 txcfg |= (TXCFG_CSI | TXCFG_HBI); 1040 } 1041 rxcfg = RXCFG_AEP | RXCFG_ARP; 1042 if (dp->full_duplex) { 1043 rxcfg |= RXCFG_ATX; 1044 } 1045 1046 /* select txmxdma and rxmxdma, maxmum burst length */ 1047 if (lp->chip->chip_type == CHIPTYPE_SIS900) { 1048 #ifdef DEBUG_SIS900_EDB 1049 val = CFG_EDB_MASTER; 1050 #else 1051 val = INL(dp, CFG) & CFG_EDB_MASTER; 1052 #endif 1053 if (val) { 1054 /* 1055 * sis900 built-in cores: 1056 * max burst length must be fixed to 64 1057 */ 1058 txmxdma = 64; 1059 rxmxdma = 64; 1060 } else { 1061 /* 1062 * sis900 pci chipset: 1063 * the vendor recommended to fix max burst length 1064 * to 512 1065 */ 1066 txmxdma = 512; 1067 rxmxdma = 512; 1068 } 1069 } else { 1070 /* 1071 * NS dp83815/816: 1072 * use user defined or default for tx/rx max burst length 1073 */ 1074 txmxdma = max(dp->txmaxdma, 256); 1075 rxmxdma = max(dp->rxmaxdma, 256); 1076 } 1077 1078 1079 /* tx high water mark */ 1080 lp->tx_drain_threshold = ROUNDUP2(dp->txthr, TXCFG_FIFO_UNIT); 1081 1082 /* determine tx_fill_threshold accroding drain threshold */ 1083 lp->tx_fill_threshold = 1084 TXFIFOSIZE - lp->tx_drain_threshold - TXCFG_FIFO_UNIT; 1085 1086 /* tune txmxdma not to exceed tx_fill_threshold */ 1087 for (; ; ) { 1088 /* normalize txmxdma requested */ 1089 val = sfe_encode_mxdma(txmxdma); 1090 txmxdma = sfe_mxdma_value[val]; 1091 1092 if (txmxdma <= lp->tx_fill_threshold) { 1093 break; 1094 } 1095 /* select new txmxdma */ 1096 txmxdma = txmxdma / 2; 1097 } 1098 txcfg |= val << TXCFG_MXDMA_SHIFT; 1099 1100 /* encode rxmxdma, maxmum burst length for rx */ 1101 val = sfe_encode_mxdma(rxmxdma); 1102 rxcfg |= val << RXCFG_MXDMA_SHIFT; 1103 rxmxdma = sfe_mxdma_value[val]; 1104 1105 /* receive starting threshold - it have only 5bit-wide field */ 1106 val = ROUNDUP2(max(dp->rxthr, ETHERMIN), RXCFG_FIFO_UNIT); 1107 lp->rx_drain_threshold = 1108 min(val, (RXCFG_DRTH >> RXCFG_DRTH_SHIFT) * RXCFG_FIFO_UNIT); 1109 1110 DPRINTF(0, (CE_CONT, 1111 "%s: %s: tx: drain:%d(rest %d) fill:%d mxdma:%d," 1112 " rx: drain:%d mxdma:%d", 1113 dp->name, __func__, 1114 lp->tx_drain_threshold, TXFIFOSIZE - lp->tx_drain_threshold, 1115 lp->tx_fill_threshold, txmxdma, 1116 lp->rx_drain_threshold, rxmxdma)); 1117 1118 ASSERT(lp->tx_drain_threshold < 64*TXCFG_FIFO_UNIT); 1119 ASSERT(lp->tx_fill_threshold < 64*TXCFG_FIFO_UNIT); 1120 ASSERT(lp->rx_drain_threshold < 32*RXCFG_FIFO_UNIT); 1121 1122 txcfg |= ((lp->tx_fill_threshold/TXCFG_FIFO_UNIT) << TXCFG_FLTH_SHIFT) 1123 | (lp->tx_drain_threshold/TXCFG_FIFO_UNIT); 1124 OUTL(dp, TXCFG, txcfg); 1125 1126 rxcfg |= ((lp->rx_drain_threshold/RXCFG_FIFO_UNIT) << RXCFG_DRTH_SHIFT); 1127 if (lp->chip->chip_type == CHIPTYPE_DP83815) { 1128 rxcfg |= RXCFG_ALP_DP83815; 1129 } 1130 OUTL(dp, RXCFG, rxcfg); 1131 1132 DPRINTF(0, (CE_CONT, CONS "%s: %s: txcfg:%b rxcfg:%b", 1133 dp->name, __func__, 1134 txcfg, TXCFG_BITS, rxcfg, RXCFG_BITS)); 1135 1136 /* Flow control */ 1137 if (lp->chip->chip_type == CHIPTYPE_DP83815) { 1138 pcr = INL(dp, PCR); 1139 switch (dp->flow_control) { 1140 case FLOW_CONTROL_SYMMETRIC: 1141 case FLOW_CONTROL_RX_PAUSE: 1142 OUTL(dp, PCR, pcr | PCR_PSEN | PCR_PS_MCAST); 1143 break; 1144 1145 default: 1146 OUTL(dp, PCR, 1147 pcr & ~(PCR_PSEN | PCR_PS_MCAST | PCR_PS_DA)); 1148 break; 1149 } 1150 DPRINTF(2, (CE_CONT, CONS "%s: PCR: %b", dp->name, 1151 INL(dp, PCR), PCR_BITS)); 1152 1153 } else if (lp->chip->chip_type == CHIPTYPE_SIS900) { 1154 switch (dp->flow_control) { 1155 case FLOW_CONTROL_SYMMETRIC: 1156 case FLOW_CONTROL_RX_PAUSE: 1157 OUTL(dp, FLOWCTL, FLOWCTL_FLOWEN); 1158 break; 1159 default: 1160 OUTL(dp, FLOWCTL, 0); 1161 break; 1162 } 1163 DPRINTF(2, (CE_CONT, CONS "%s: FLOWCTL: %b", 1164 dp->name, INL(dp, FLOWCTL), FLOWCTL_BITS)); 1165 } 1166 return (GEM_SUCCESS); 1167 } 1168 1169 static int 1170 sfe_get_stats(struct gem_dev *dp) 1171 { 1172 /* do nothing */ 1173 return (GEM_SUCCESS); 1174 } 1175 1176 /* 1177 * descriptor manipulations 1178 */ 1179 static int 1180 sfe_tx_desc_write(struct gem_dev *dp, int slot, 1181 ddi_dma_cookie_t *dmacookie, int frags, uint64_t flags) 1182 { 1183 uint32_t mark; 1184 struct sfe_desc *tdp; 1185 ddi_dma_cookie_t *dcp; 1186 uint32_t tmp0; 1187 #if DEBUG_LEVEL > 2 1188 int i; 1189 1190 cmn_err(CE_CONT, 1191 CONS "%s: time:%d %s seqnum: %d, slot %d, frags: %d flags: %llx", 1192 dp->name, ddi_get_lbolt(), __func__, 1193 dp->tx_desc_tail, slot, frags, flags); 1194 1195 for (i = 0; i < frags; i++) { 1196 cmn_err(CE_CONT, CONS "%d: addr: 0x%x, len: 0x%x", 1197 i, dmacookie[i].dmac_address, dmacookie[i].dmac_size); 1198 } 1199 #endif 1200 /* 1201 * write tx descriptor in reversed order. 1202 */ 1203 #if DEBUG_LEVEL > 3 1204 flags |= GEM_TXFLAG_INTR; 1205 #endif 1206 mark = (flags & GEM_TXFLAG_INTR) 1207 ? (CMDSTS_OWN | CMDSTS_INTR) : CMDSTS_OWN; 1208 1209 ASSERT(frags == 1); 1210 dcp = &dmacookie[0]; 1211 if (flags & GEM_TXFLAG_HEAD) { 1212 mark &= ~CMDSTS_OWN; 1213 } 1214 1215 tdp = (void *)&dp->tx_ring[SFE_DESC_SIZE * slot]; 1216 tmp0 = (uint32_t)dcp->dmac_address; 1217 mark |= (uint32_t)dcp->dmac_size; 1218 tdp->d_bufptr = LE_32(tmp0); 1219 tdp->d_cmdsts = LE_32(mark); 1220 1221 return (frags); 1222 } 1223 1224 static void 1225 sfe_tx_start(struct gem_dev *dp, int start_slot, int nslot) 1226 { 1227 uint_t tx_ring_size = dp->gc.gc_tx_ring_size; 1228 struct sfe_desc *tdp; 1229 struct sfe_dev *lp = dp->private; 1230 1231 if (nslot > 1) { 1232 gem_tx_desc_dma_sync(dp, 1233 SLOT(start_slot + 1, tx_ring_size), 1234 nslot - 1, DDI_DMA_SYNC_FORDEV); 1235 } 1236 1237 tdp = (void *)&dp->tx_ring[SFE_DESC_SIZE * start_slot]; 1238 tdp->d_cmdsts |= LE_32(CMDSTS_OWN); 1239 1240 gem_tx_desc_dma_sync(dp, start_slot, 1, DDI_DMA_SYNC_FORDEV); 1241 1242 /* 1243 * Let the Transmit Buffer Manager Fill state machine active. 1244 */ 1245 if (dp->mac_active) { 1246 OUTL(dp, CR, lp->cr | CR_TXE); 1247 } 1248 } 1249 1250 static void 1251 sfe_rx_desc_write(struct gem_dev *dp, int slot, 1252 ddi_dma_cookie_t *dmacookie, int frags) 1253 { 1254 struct sfe_desc *rdp; 1255 uint32_t tmp0; 1256 uint32_t tmp1; 1257 #if DEBUG_LEVEL > 2 1258 int i; 1259 1260 ASSERT(frags == 1); 1261 1262 cmn_err(CE_CONT, CONS 1263 "%s: %s seqnum: %d, slot %d, frags: %d", 1264 dp->name, __func__, dp->rx_active_tail, slot, frags); 1265 for (i = 0; i < frags; i++) { 1266 cmn_err(CE_CONT, CONS " frag: %d addr: 0x%llx, len: 0x%lx", 1267 i, dmacookie[i].dmac_address, dmacookie[i].dmac_size); 1268 } 1269 #endif 1270 /* for the last slot of the packet */ 1271 rdp = (void *)&dp->rx_ring[SFE_DESC_SIZE * slot]; 1272 1273 tmp0 = (uint32_t)dmacookie->dmac_address; 1274 tmp1 = CMDSTS_INTR | (uint32_t)dmacookie->dmac_size; 1275 rdp->d_bufptr = LE_32(tmp0); 1276 rdp->d_cmdsts = LE_32(tmp1); 1277 } 1278 1279 static uint_t 1280 sfe_tx_desc_stat(struct gem_dev *dp, int slot, int ndesc) 1281 { 1282 uint_t tx_ring_size = dp->gc.gc_tx_ring_size; 1283 struct sfe_desc *tdp; 1284 uint32_t status; 1285 int cols; 1286 struct sfe_dev *lp = dp->private; 1287 #ifdef DEBUG_LEVEL 1288 int i; 1289 clock_t delay; 1290 #endif 1291 /* check status of the last descriptor */ 1292 tdp = (void *) 1293 &dp->tx_ring[SFE_DESC_SIZE * SLOT(slot + ndesc - 1, tx_ring_size)]; 1294 1295 /* 1296 * Don't use LE_32() directly to refer tdp->d_cmdsts. 1297 * It is not atomic for big endian cpus. 1298 */ 1299 status = tdp->d_cmdsts; 1300 status = LE_32(status); 1301 1302 DPRINTF(2, (CE_CONT, CONS "%s: time:%ld %s: slot:%d, status:0x%b", 1303 dp->name, ddi_get_lbolt(), __func__, 1304 slot, status, TXSTAT_BITS)); 1305 1306 if (status & CMDSTS_OWN) { 1307 /* 1308 * not yet transmitted 1309 */ 1310 /* workaround for tx hang */ 1311 if (lp->chip->chip_type == CHIPTYPE_DP83815 && 1312 dp->mac_active) { 1313 OUTL(dp, CR, lp->cr | CR_TXE); 1314 } 1315 return (0); 1316 } 1317 1318 if (status & CMDSTS_MORE) { 1319 /* XXX - the hardware problem but don't panic the system */ 1320 /* avoid lint bug for %b format string including 32nd bit */ 1321 cmn_err(CE_NOTE, CONS 1322 "%s: tx status bits incorrect: slot:%d, status:0x%x", 1323 dp->name, slot, status); 1324 } 1325 1326 #if DEBUG_LEVEL > 3 1327 delay = (ddi_get_lbolt() - dp->tx_buf_head->txb_stime) * 10; 1328 if (delay >= 50) { 1329 DPRINTF(0, (CE_NOTE, "%s: tx deferred %d mS: slot %d", 1330 dp->name, delay, slot)); 1331 } 1332 #endif 1333 1334 #if DEBUG_LEVEL > 3 1335 for (i = 0; i < nfrag-1; i++) { 1336 uint32_t s; 1337 int n; 1338 1339 n = SLOT(slot + i, tx_ring_size); 1340 s = LE_32( 1341 ((struct sfe_desc *)((void *) 1342 &dp->tx_ring[SFE_DESC_SIZE * n]))->d_cmdsts); 1343 1344 ASSERT(s & CMDSTS_MORE); 1345 ASSERT((s & CMDSTS_OWN) == 0); 1346 } 1347 #endif 1348 1349 /* 1350 * collect statistics 1351 */ 1352 if ((status & CMDSTS_OK) == 0) { 1353 1354 /* failed to transmit the packet */ 1355 1356 DPRINTF(0, (CE_CONT, CONS "%s: Transmit error, Tx status %b", 1357 dp->name, status, TXSTAT_BITS)); 1358 1359 dp->stats.errxmt++; 1360 1361 if (status & CMDSTS_TFU) { 1362 dp->stats.underflow++; 1363 } else if (status & CMDSTS_CRS) { 1364 dp->stats.nocarrier++; 1365 } else if (status & CMDSTS_OWC) { 1366 dp->stats.xmtlatecoll++; 1367 } else if ((!dp->full_duplex) && (status & CMDSTS_EC)) { 1368 dp->stats.excoll++; 1369 dp->stats.collisions += 16; 1370 } else { 1371 dp->stats.xmit_internal_err++; 1372 } 1373 } else if (!dp->full_duplex) { 1374 cols = (status >> CMDSTS_CCNT_SHIFT) & CCNT_MASK; 1375 1376 if (cols > 0) { 1377 if (cols == 1) { 1378 dp->stats.first_coll++; 1379 } else /* (cols > 1) */ { 1380 dp->stats.multi_coll++; 1381 } 1382 dp->stats.collisions += cols; 1383 } else if (status & CMDSTS_TD) { 1384 dp->stats.defer++; 1385 } 1386 } 1387 return (GEM_TX_DONE); 1388 } 1389 1390 static uint64_t 1391 sfe_rx_desc_stat(struct gem_dev *dp, int slot, int ndesc) 1392 { 1393 struct sfe_desc *rdp; 1394 uint_t len; 1395 uint_t flag; 1396 uint32_t status; 1397 1398 flag = GEM_RX_DONE; 1399 1400 /* Dont read ISR because we cannot ack only to rx interrupt. */ 1401 1402 rdp = (void *)&dp->rx_ring[SFE_DESC_SIZE * slot]; 1403 1404 /* 1405 * Don't use LE_32() directly to refer rdp->d_cmdsts. 1406 * It is not atomic for big endian cpus. 1407 */ 1408 status = rdp->d_cmdsts; 1409 status = LE_32(status); 1410 1411 DPRINTF(2, (CE_CONT, CONS "%s: time:%ld %s: slot:%d, status:0x%b", 1412 dp->name, ddi_get_lbolt(), __func__, 1413 slot, status, RXSTAT_BITS)); 1414 1415 if ((status & CMDSTS_OWN) == 0) { 1416 /* 1417 * No more received packets because 1418 * this buffer is owned by NIC. 1419 */ 1420 return (0); 1421 } 1422 1423 #define RX_ERR_BITS \ 1424 (CMDSTS_RXA | CMDSTS_RXO | CMDSTS_LONG | CMDSTS_RUNT | \ 1425 CMDSTS_ISE | CMDSTS_CRCE | CMDSTS_FAE | CMDSTS_MORE) 1426 1427 if (status & RX_ERR_BITS) { 1428 /* 1429 * Packet with error received 1430 */ 1431 DPRINTF(0, (CE_CONT, CONS "%s: Corrupted packet " 1432 "received, buffer status: %b", 1433 dp->name, status, RXSTAT_BITS)); 1434 1435 /* collect statistics information */ 1436 dp->stats.errrcv++; 1437 1438 if (status & CMDSTS_RXO) { 1439 dp->stats.overflow++; 1440 } else if (status & (CMDSTS_LONG | CMDSTS_MORE)) { 1441 dp->stats.frame_too_long++; 1442 } else if (status & CMDSTS_RUNT) { 1443 dp->stats.runt++; 1444 } else if (status & (CMDSTS_ISE | CMDSTS_FAE)) { 1445 dp->stats.frame++; 1446 } else if (status & CMDSTS_CRCE) { 1447 dp->stats.crc++; 1448 } else { 1449 dp->stats.rcv_internal_err++; 1450 } 1451 1452 return (flag | GEM_RX_ERR); 1453 } 1454 1455 /* 1456 * this packet was received without errors 1457 */ 1458 if ((len = (status & CMDSTS_SIZE)) >= ETHERFCSL) { 1459 len -= ETHERFCSL; 1460 } 1461 1462 #if DEBUG_LEVEL > 10 1463 { 1464 int i; 1465 uint8_t *bp = dp->rx_buf_head->rxb_buf; 1466 1467 cmn_err(CE_CONT, CONS "%s: len:%d", dp->name, len); 1468 1469 for (i = 0; i < 60; i += 10) { 1470 cmn_err(CE_CONT, CONS 1471 "%02x %02x %02x %02x %02x %02x %02x %02x %02x %02x", 1472 bp[0], bp[1], bp[2], bp[3], bp[4], 1473 bp[5], bp[6], bp[7], bp[8], bp[9]); 1474 } 1475 bp += 10; 1476 } 1477 #endif 1478 return (flag | (len & GEM_RX_LEN)); 1479 } 1480 1481 static void 1482 sfe_tx_desc_init(struct gem_dev *dp, int slot) 1483 { 1484 uint_t tx_ring_size = dp->gc.gc_tx_ring_size; 1485 struct sfe_desc *tdp; 1486 uint32_t here; 1487 1488 tdp = (void *)&dp->tx_ring[SFE_DESC_SIZE * slot]; 1489 1490 /* don't clear d_link field, which have a valid pointer */ 1491 tdp->d_cmdsts = 0; 1492 1493 /* make a link to this from the previous descriptor */ 1494 here = ((uint32_t)dp->tx_ring_dma) + SFE_DESC_SIZE*slot; 1495 1496 tdp = (void *) 1497 &dp->tx_ring[SFE_DESC_SIZE * SLOT(slot - 1, tx_ring_size)]; 1498 tdp->d_link = LE_32(here); 1499 } 1500 1501 static void 1502 sfe_rx_desc_init(struct gem_dev *dp, int slot) 1503 { 1504 uint_t rx_ring_size = dp->gc.gc_rx_ring_size; 1505 struct sfe_desc *rdp; 1506 uint32_t here; 1507 1508 rdp = (void *)&dp->rx_ring[SFE_DESC_SIZE * slot]; 1509 1510 /* don't clear d_link field, which have a valid pointer */ 1511 rdp->d_cmdsts = LE_32(CMDSTS_OWN); 1512 1513 /* make a link to this from the previous descriptor */ 1514 here = ((uint32_t)dp->rx_ring_dma) + SFE_DESC_SIZE*slot; 1515 1516 rdp = (void *) 1517 &dp->rx_ring[SFE_DESC_SIZE * SLOT(slot - 1, rx_ring_size)]; 1518 rdp->d_link = LE_32(here); 1519 } 1520 1521 static void 1522 sfe_tx_desc_clean(struct gem_dev *dp, int slot) 1523 { 1524 struct sfe_desc *tdp; 1525 1526 tdp = (void *)&dp->tx_ring[SFE_DESC_SIZE * slot]; 1527 tdp->d_cmdsts = 0; 1528 } 1529 1530 static void 1531 sfe_rx_desc_clean(struct gem_dev *dp, int slot) 1532 { 1533 struct sfe_desc *rdp; 1534 1535 rdp = (void *)&dp->rx_ring[SFE_DESC_SIZE * slot]; 1536 rdp->d_cmdsts = LE_32(CMDSTS_OWN); 1537 } 1538 1539 /* 1540 * Device depend interrupt handler 1541 */ 1542 static uint_t 1543 sfe_interrupt(struct gem_dev *dp) 1544 { 1545 uint_t rx_ring_size = dp->gc.gc_rx_ring_size; 1546 uint32_t isr; 1547 uint32_t isr_bogus; 1548 uint_t flags = 0; 1549 boolean_t need_to_reset = B_FALSE; 1550 struct sfe_dev *lp = dp->private; 1551 1552 /* read reason and clear interrupt */ 1553 isr = INL(dp, ISR); 1554 1555 isr_bogus = lp->isr_pended; 1556 lp->isr_pended = 0; 1557 1558 if (((isr | isr_bogus) & lp->our_intr_bits) == 0) { 1559 /* we are not the interrupt source */ 1560 return (DDI_INTR_UNCLAIMED); 1561 } 1562 1563 DPRINTF(3, (CE_CONT, 1564 CONS "%s: time:%ld %s:called: isr:0x%b rx_active_head: %d", 1565 dp->name, ddi_get_lbolt(), __func__, 1566 isr, INTR_BITS, dp->rx_active_head)); 1567 1568 if (!dp->mac_active) { 1569 /* the device is going to stop */ 1570 lp->our_intr_bits = 0; 1571 return (DDI_INTR_CLAIMED); 1572 } 1573 1574 isr &= lp->our_intr_bits; 1575 1576 if (isr & (ISR_RXSOVR | ISR_RXORN | ISR_RXIDLE | ISR_RXERR | 1577 ISR_RXDESC | ISR_RXOK)) { 1578 (void) gem_receive(dp); 1579 1580 if (isr & (ISR_RXSOVR | ISR_RXORN)) { 1581 DPRINTF(0, (CE_CONT, 1582 CONS "%s: rx fifo overrun: isr %b", 1583 dp->name, isr, INTR_BITS)); 1584 /* no need restart rx */ 1585 dp->stats.overflow++; 1586 } 1587 1588 if (isr & ISR_RXIDLE) { 1589 DPRINTF(0, (CE_CONT, 1590 CONS "%s: rx buffer ran out: isr %b", 1591 dp->name, isr, INTR_BITS)); 1592 1593 dp->stats.norcvbuf++; 1594 1595 /* 1596 * Make RXDP points the head of receive 1597 * buffer list. 1598 */ 1599 OUTL(dp, RXDP, dp->rx_ring_dma + 1600 SFE_DESC_SIZE * 1601 SLOT(dp->rx_active_head, rx_ring_size)); 1602 1603 /* Restart the receive engine */ 1604 OUTL(dp, CR, lp->cr | CR_RXE); 1605 } 1606 } 1607 1608 if (isr & (ISR_TXURN | ISR_TXERR | ISR_TXDESC | 1609 ISR_TXIDLE | ISR_TXOK)) { 1610 /* need to reclaim tx buffers */ 1611 if (gem_tx_done(dp)) { 1612 flags |= INTR_RESTART_TX; 1613 } 1614 /* 1615 * XXX - tx error statistics will be counted in 1616 * sfe_tx_desc_stat() and no need to restart tx on errors. 1617 */ 1618 } 1619 1620 if (isr & (ISR_DPERR | ISR_SSERR | ISR_RMABT | ISR_RTABT)) { 1621 cmn_err(CE_WARN, "%s: ERROR interrupt: isr %b.", 1622 dp->name, isr, INTR_BITS); 1623 need_to_reset = B_TRUE; 1624 } 1625 reset: 1626 if (need_to_reset) { 1627 (void) gem_restart_nic(dp, GEM_RESTART_KEEP_BUF); 1628 flags |= INTR_RESTART_TX; 1629 } 1630 1631 DPRINTF(5, (CE_CONT, CONS "%s: %s: return: isr: %b", 1632 dp->name, __func__, isr, INTR_BITS)); 1633 1634 return (DDI_INTR_CLAIMED | flags); 1635 } 1636 1637 /* ======================================================== */ 1638 /* 1639 * HW depend MII routine 1640 */ 1641 /* ======================================================== */ 1642 1643 /* 1644 * MII routines for NS DP83815 1645 */ 1646 static void 1647 sfe_mii_sync_dp83815(struct gem_dev *dp) 1648 { 1649 /* do nothing */ 1650 } 1651 1652 static uint16_t 1653 sfe_mii_read_dp83815(struct gem_dev *dp, uint_t offset) 1654 { 1655 DPRINTF(4, (CE_CONT, CONS"%s: %s: offset 0x%x", 1656 dp->name, __func__, offset)); 1657 return ((uint16_t)INL(dp, MII_REGS_BASE + offset*4)); 1658 } 1659 1660 static void 1661 sfe_mii_write_dp83815(struct gem_dev *dp, uint_t offset, uint16_t val) 1662 { 1663 DPRINTF(4, (CE_CONT, CONS"%s: %s: offset 0x%x 0x%x", 1664 dp->name, __func__, offset, val)); 1665 OUTL(dp, MII_REGS_BASE + offset*4, val); 1666 } 1667 1668 static int 1669 sfe_mii_config_dp83815(struct gem_dev *dp) 1670 { 1671 uint32_t srr; 1672 1673 srr = INL(dp, SRR) & SRR_REV; 1674 1675 DPRINTF(0, (CE_CONT, CONS "%s: srr:0x%04x %04x %04x %04x %04x %04x", 1676 dp->name, srr, 1677 INW(dp, 0x00cc), /* PGSEL */ 1678 INW(dp, 0x00e4), /* PMDCSR */ 1679 INW(dp, 0x00fc), /* TSTDAT */ 1680 INW(dp, 0x00f4), /* DSPCFG */ 1681 INW(dp, 0x00f8))); /* SDCFG */ 1682 1683 if (srr == SRR_REV_DP83815CVNG) { 1684 /* 1685 * NS datasheet says that DP83815CVNG needs following 1686 * registers to be patched for optimizing its performance. 1687 * A report said that CRC errors on RX disappeared 1688 * with the patch. 1689 */ 1690 OUTW(dp, 0x00cc, 0x0001); /* PGSEL */ 1691 OUTW(dp, 0x00e4, 0x189c); /* PMDCSR */ 1692 OUTW(dp, 0x00fc, 0x0000); /* TSTDAT */ 1693 OUTW(dp, 0x00f4, 0x5040); /* DSPCFG */ 1694 OUTW(dp, 0x00f8, 0x008c); /* SDCFG */ 1695 OUTW(dp, 0x00cc, 0x0000); /* PGSEL */ 1696 1697 DPRINTF(0, (CE_CONT, 1698 CONS "%s: PHY patched %04x %04x %04x %04x %04x", 1699 dp->name, 1700 INW(dp, 0x00cc), /* PGSEL */ 1701 INW(dp, 0x00e4), /* PMDCSR */ 1702 INW(dp, 0x00fc), /* TSTDAT */ 1703 INW(dp, 0x00f4), /* DSPCFG */ 1704 INW(dp, 0x00f8))); /* SDCFG */ 1705 } else if (((srr ^ SRR_REV_DP83815DVNG) & 0xff00) == 0 || 1706 ((srr ^ SRR_REV_DP83816AVNG) & 0xff00) == 0) { 1707 /* 1708 * Additional packets for later chipset 1709 */ 1710 OUTW(dp, 0x00cc, 0x0001); /* PGSEL */ 1711 OUTW(dp, 0x00e4, 0x189c); /* PMDCSR */ 1712 OUTW(dp, 0x00cc, 0x0000); /* PGSEL */ 1713 1714 DPRINTF(0, (CE_CONT, 1715 CONS "%s: PHY patched %04x %04x", 1716 dp->name, 1717 INW(dp, 0x00cc), /* PGSEL */ 1718 INW(dp, 0x00e4))); /* PMDCSR */ 1719 } 1720 1721 return (gem_mii_config_default(dp)); 1722 } 1723 1724 static int 1725 sfe_mii_probe_dp83815(struct gem_dev *dp) 1726 { 1727 uint32_t val; 1728 1729 /* try external phy first */ 1730 DPRINTF(0, (CE_CONT, CONS "%s: %s: trying external phy", 1731 dp->name, __func__)); 1732 dp->mii_phy_addr = 0; 1733 dp->gc.gc_mii_sync = &sfe_mii_sync_sis900; 1734 dp->gc.gc_mii_read = &sfe_mii_read_sis900; 1735 dp->gc.gc_mii_write = &sfe_mii_write_sis900; 1736 1737 val = INL(dp, CFG) & (CFG_ANEG_SEL | CFG_PHY_CFG); 1738 OUTL(dp, CFG, val | CFG_EXT_PHY | CFG_PHY_DIS); 1739 1740 if (gem_mii_probe_default(dp) == GEM_SUCCESS) { 1741 return (GEM_SUCCESS); 1742 } 1743 1744 /* switch to internal phy */ 1745 DPRINTF(0, (CE_CONT, CONS "%s: %s: switching to internal phy", 1746 dp->name, __func__)); 1747 dp->mii_phy_addr = -1; 1748 dp->gc.gc_mii_sync = &sfe_mii_sync_dp83815; 1749 dp->gc.gc_mii_read = &sfe_mii_read_dp83815; 1750 dp->gc.gc_mii_write = &sfe_mii_write_dp83815; 1751 1752 val = INL(dp, CFG) & (CFG_ANEG_SEL | CFG_PHY_CFG); 1753 OUTL(dp, CFG, val | CFG_PAUSE_ADV | CFG_PHY_RST); 1754 drv_usecwait(100); /* keep to assert RST bit for a while */ 1755 OUTL(dp, CFG, val | CFG_PAUSE_ADV); 1756 1757 /* wait for PHY reset */ 1758 delay(drv_usectohz(10000)); 1759 1760 return (gem_mii_probe_default(dp)); 1761 } 1762 1763 static int 1764 sfe_mii_init_dp83815(struct gem_dev *dp) 1765 { 1766 uint32_t val; 1767 1768 val = INL(dp, CFG) & (CFG_ANEG_SEL | CFG_PHY_CFG); 1769 1770 if (dp->mii_phy_addr == -1) { 1771 /* select internal phy */ 1772 OUTL(dp, CFG, val | CFG_PAUSE_ADV); 1773 } else { 1774 /* select external phy */ 1775 OUTL(dp, CFG, val | CFG_EXT_PHY | CFG_PHY_DIS); 1776 } 1777 1778 return (GEM_SUCCESS); 1779 } 1780 1781 /* 1782 * MII routines for SiS900 1783 */ 1784 #define MDIO_DELAY(dp) {(void) INL(dp, MEAR); (void) INL(dp, MEAR); } 1785 static void 1786 sfe_mii_sync_sis900(struct gem_dev *dp) 1787 { 1788 int i; 1789 1790 /* send 32 ONE's to make MII line idle */ 1791 for (i = 0; i < 32; i++) { 1792 OUTL(dp, MEAR, MEAR_MDDIR | MEAR_MDIO); 1793 MDIO_DELAY(dp); 1794 OUTL(dp, MEAR, MEAR_MDDIR | MEAR_MDIO | MEAR_MDC); 1795 MDIO_DELAY(dp); 1796 } 1797 } 1798 1799 static int 1800 sfe_mii_config_sis900(struct gem_dev *dp) 1801 { 1802 struct sfe_dev *lp = dp->private; 1803 1804 /* Do chip depend setup */ 1805 if ((dp->mii_phy_id & PHY_MASK) == PHY_ICS1893) { 1806 /* workaround for ICS1893 PHY */ 1807 gem_mii_write(dp, 0x0018, 0xD200); 1808 } 1809 1810 if (lp->revid == SIS630E_900_REV) { 1811 /* 1812 * SiS 630E has bugs on default values 1813 * of PHY registers 1814 */ 1815 gem_mii_write(dp, MII_AN_ADVERT, 0x05e1); 1816 gem_mii_write(dp, MII_CONFIG1, 0x0022); 1817 gem_mii_write(dp, MII_CONFIG2, 0xff00); 1818 gem_mii_write(dp, MII_MASK, 0xffc0); 1819 } 1820 sfe_set_eq_sis630(dp); 1821 1822 return (gem_mii_config_default(dp)); 1823 } 1824 1825 static uint16_t 1826 sfe_mii_read_sis900(struct gem_dev *dp, uint_t reg) 1827 { 1828 uint32_t cmd; 1829 uint16_t ret; 1830 int i; 1831 uint32_t data; 1832 1833 cmd = MII_READ_CMD(dp->mii_phy_addr, reg); 1834 1835 for (i = 31; i >= 18; i--) { 1836 data = ((cmd >> i) & 1) << MEAR_MDIO_SHIFT; 1837 OUTL(dp, MEAR, data | MEAR_MDDIR); 1838 MDIO_DELAY(dp); 1839 OUTL(dp, MEAR, data | MEAR_MDDIR | MEAR_MDC); 1840 MDIO_DELAY(dp); 1841 } 1842 1843 /* turn around cycle */ 1844 OUTL(dp, MEAR, 0); 1845 MDIO_DELAY(dp); 1846 1847 /* get response from PHY */ 1848 OUTL(dp, MEAR, MEAR_MDC); 1849 MDIO_DELAY(dp); 1850 1851 OUTL(dp, MEAR, 0); 1852 #if DEBUG_LEBEL > 0 1853 (void) INL(dp, MEAR); /* delay */ 1854 if (INL(dp, MEAR) & MEAR_MDIO) { 1855 cmn_err(CE_WARN, "%s: PHY@%d not responded", 1856 dp->name, dp->mii_phy_addr); 1857 } 1858 #else 1859 MDIO_DELAY(dp); 1860 #endif 1861 /* terminate response cycle */ 1862 OUTL(dp, MEAR, MEAR_MDC); 1863 MDIO_DELAY(dp); 1864 1865 ret = 0; /* to avoid lint errors */ 1866 for (i = 16; i > 0; i--) { 1867 OUTL(dp, MEAR, 0); 1868 (void) INL(dp, MEAR); /* delay */ 1869 ret = (ret << 1) | ((INL(dp, MEAR) >> MEAR_MDIO_SHIFT) & 1); 1870 OUTL(dp, MEAR, MEAR_MDC); 1871 MDIO_DELAY(dp); 1872 } 1873 1874 /* send two idle(Z) bits to terminate the read cycle */ 1875 for (i = 0; i < 2; i++) { 1876 OUTL(dp, MEAR, 0); 1877 MDIO_DELAY(dp); 1878 OUTL(dp, MEAR, MEAR_MDC); 1879 MDIO_DELAY(dp); 1880 } 1881 1882 return (ret); 1883 } 1884 1885 static void 1886 sfe_mii_write_sis900(struct gem_dev *dp, uint_t reg, uint16_t val) 1887 { 1888 uint32_t cmd; 1889 int i; 1890 uint32_t data; 1891 1892 cmd = MII_WRITE_CMD(dp->mii_phy_addr, reg, val); 1893 1894 for (i = 31; i >= 0; i--) { 1895 data = ((cmd >> i) & 1) << MEAR_MDIO_SHIFT; 1896 OUTL(dp, MEAR, data | MEAR_MDDIR); 1897 MDIO_DELAY(dp); 1898 OUTL(dp, MEAR, data | MEAR_MDDIR | MEAR_MDC); 1899 MDIO_DELAY(dp); 1900 } 1901 1902 /* send two idle(Z) bits to terminate the write cycle. */ 1903 for (i = 0; i < 2; i++) { 1904 OUTL(dp, MEAR, 0); 1905 MDIO_DELAY(dp); 1906 OUTL(dp, MEAR, MEAR_MDC); 1907 MDIO_DELAY(dp); 1908 } 1909 } 1910 #undef MDIO_DELAY 1911 1912 static void 1913 sfe_set_eq_sis630(struct gem_dev *dp) 1914 { 1915 uint16_t reg14h; 1916 uint16_t eq_value; 1917 uint16_t max_value; 1918 uint16_t min_value; 1919 int i; 1920 uint8_t rev; 1921 struct sfe_dev *lp = dp->private; 1922 1923 rev = lp->revid; 1924 1925 if (!(rev == SIS630E_900_REV || rev == SIS630EA1_900_REV || 1926 rev == SIS630A_900_REV || rev == SIS630ET_900_REV)) { 1927 /* it doesn't have a internal PHY */ 1928 return; 1929 } 1930 1931 if (dp->mii_state == MII_STATE_LINKUP) { 1932 reg14h = gem_mii_read(dp, MII_RESV); 1933 gem_mii_write(dp, MII_RESV, (0x2200 | reg14h) & 0xBFFF); 1934 1935 eq_value = (0x00f8 & gem_mii_read(dp, MII_RESV)) >> 3; 1936 max_value = min_value = eq_value; 1937 for (i = 1; i < 10; i++) { 1938 eq_value = (0x00f8 & gem_mii_read(dp, MII_RESV)) >> 3; 1939 max_value = max(eq_value, max_value); 1940 min_value = min(eq_value, min_value); 1941 } 1942 1943 /* for 630E, rule to determine the equalizer value */ 1944 if (rev == SIS630E_900_REV || rev == SIS630EA1_900_REV || 1945 rev == SIS630ET_900_REV) { 1946 if (max_value < 5) { 1947 eq_value = max_value; 1948 } else if (5 <= max_value && max_value < 15) { 1949 eq_value = 1950 max(max_value + 1, 1951 min_value + 2); 1952 } else if (15 <= max_value) { 1953 eq_value = 1954 max(max_value + 5, 1955 min_value + 6); 1956 } 1957 } 1958 /* for 630B0&B1, rule to determine the equalizer value */ 1959 else 1960 if (rev == SIS630A_900_REV && 1961 (lp->bridge_revid == SIS630B0 || 1962 lp->bridge_revid == SIS630B1)) { 1963 1964 if (max_value == 0) { 1965 eq_value = 3; 1966 } else { 1967 eq_value = (max_value + min_value + 1)/2; 1968 } 1969 } 1970 /* write equalizer value and setting */ 1971 reg14h = gem_mii_read(dp, MII_RESV) & ~0x02f8; 1972 reg14h |= 0x6000 | (eq_value << 3); 1973 gem_mii_write(dp, MII_RESV, reg14h); 1974 } else { 1975 reg14h = (gem_mii_read(dp, MII_RESV) & ~0x4000) | 0x2000; 1976 if (rev == SIS630A_900_REV && 1977 (lp->bridge_revid == SIS630B0 || 1978 lp->bridge_revid == SIS630B1)) { 1979 1980 reg14h |= 0x0200; 1981 } 1982 gem_mii_write(dp, MII_RESV, reg14h); 1983 } 1984 } 1985 1986 /* ======================================================== */ 1987 /* 1988 * OS depend (device driver) routine 1989 */ 1990 /* ======================================================== */ 1991 static void 1992 sfe_chipinfo_init_sis900(struct gem_dev *dp) 1993 { 1994 int rev; 1995 struct sfe_dev *lp = (struct sfe_dev *)dp->private; 1996 1997 rev = lp->revid; 1998 1999 if (rev == SIS630E_900_REV /* 0x81 */) { 2000 /* sis630E */ 2001 lp->get_mac_addr = &sfe_get_mac_addr_sis630e; 2002 } else if (rev > 0x81 && rev <= 0x90) { 2003 /* 630S, 630EA1, 630ET, 635A */ 2004 lp->get_mac_addr = &sfe_get_mac_addr_sis635; 2005 } else if (rev == SIS962_900_REV /* 0x91 */) { 2006 /* sis962 or later */ 2007 lp->get_mac_addr = &sfe_get_mac_addr_sis962; 2008 } else { 2009 /* sis900 */ 2010 lp->get_mac_addr = &sfe_get_mac_addr_sis900; 2011 } 2012 2013 lp->bridge_revid = 0; 2014 2015 if (rev == SIS630E_900_REV || rev == SIS630EA1_900_REV || 2016 rev == SIS630A_900_REV || rev == SIS630ET_900_REV) { 2017 /* 2018 * read host bridge revision 2019 */ 2020 dev_info_t *bridge; 2021 ddi_acc_handle_t bridge_handle; 2022 2023 if ((bridge = sfe_search_pci_dev(0x1039, 0x630)) == NULL) { 2024 cmn_err(CE_WARN, 2025 "%s: cannot find host bridge (pci1039,630)", 2026 dp->name); 2027 return; 2028 } 2029 2030 if (pci_config_setup(bridge, &bridge_handle) != DDI_SUCCESS) { 2031 cmn_err(CE_WARN, "%s: pci_config_setup failed", 2032 dp->name); 2033 return; 2034 } 2035 2036 lp->bridge_revid = 2037 pci_config_get8(bridge_handle, PCI_CONF_REVID); 2038 pci_config_teardown(&bridge_handle); 2039 } 2040 } 2041 2042 static int 2043 sfe_attach_chip(struct gem_dev *dp) 2044 { 2045 struct sfe_dev *lp = (struct sfe_dev *)dp->private; 2046 2047 DPRINTF(4, (CE_CONT, CONS "!%s: %s called", dp->name, __func__)); 2048 2049 /* setup chip-depend get_mac_address function */ 2050 if (lp->chip->chip_type == CHIPTYPE_SIS900) { 2051 sfe_chipinfo_init_sis900(dp); 2052 } else { 2053 lp->get_mac_addr = &sfe_get_mac_addr_dp83815; 2054 } 2055 2056 /* read MAC address */ 2057 if (!(lp->get_mac_addr)(dp)) { 2058 cmn_err(CE_WARN, 2059 "!%s: %s: failed to get factory mac address" 2060 " please specify a mac address in sfe.conf", 2061 dp->name, __func__); 2062 return (GEM_FAILURE); 2063 } 2064 2065 if (lp->chip->chip_type == CHIPTYPE_DP83815) { 2066 dp->mii_phy_addr = -1; /* no need to scan PHY */ 2067 dp->misc_flag |= GEM_VLAN_SOFT; 2068 dp->txthr += 4; /* VTAG_SIZE */ 2069 } 2070 dp->txthr = min(dp->txthr, TXFIFOSIZE - 2); 2071 2072 return (GEM_SUCCESS); 2073 } 2074 2075 static int 2076 sfeattach(dev_info_t *dip, ddi_attach_cmd_t cmd) 2077 { 2078 int unit; 2079 const char *drv_name; 2080 int i; 2081 ddi_acc_handle_t conf_handle; 2082 uint16_t vid; 2083 uint16_t did; 2084 uint8_t rev; 2085 #ifdef DEBUG_LEVEL 2086 uint32_t iline; 2087 uint8_t latim; 2088 #endif 2089 struct chip_info *p; 2090 struct gem_dev *dp; 2091 struct sfe_dev *lp; 2092 caddr_t base; 2093 ddi_acc_handle_t regs_ha; 2094 struct gem_conf *gcp; 2095 2096 unit = ddi_get_instance(dip); 2097 drv_name = ddi_driver_name(dip); 2098 2099 DPRINTF(3, (CE_CONT, CONS "%s%d: sfeattach: called", drv_name, unit)); 2100 2101 /* 2102 * Common codes after power-up 2103 */ 2104 if (pci_config_setup(dip, &conf_handle) != DDI_SUCCESS) { 2105 cmn_err(CE_WARN, "%s%d: ddi_regs_map_setup failed", 2106 drv_name, unit); 2107 goto err; 2108 } 2109 2110 vid = pci_config_get16(conf_handle, PCI_CONF_VENID); 2111 did = pci_config_get16(conf_handle, PCI_CONF_DEVID); 2112 rev = pci_config_get16(conf_handle, PCI_CONF_REVID); 2113 #ifdef DEBUG_LEVEL 2114 iline = pci_config_get32(conf_handle, PCI_CONF_ILINE); 2115 latim = pci_config_get8(conf_handle, PCI_CONF_LATENCY_TIMER); 2116 #endif 2117 #ifdef DEBUG_BUILT_IN_SIS900 2118 rev = SIS630E_900_REV; 2119 #endif 2120 for (i = 0, p = sfe_chiptbl; i < CHIPTABLESIZE; i++, p++) { 2121 if (p->venid == vid && p->devid == did) { 2122 /* found */ 2123 goto chip_found; 2124 } 2125 } 2126 2127 /* Not found */ 2128 cmn_err(CE_WARN, 2129 "%s%d: sfe_attach: wrong PCI venid/devid (0x%x, 0x%x)", 2130 drv_name, unit, vid, did); 2131 pci_config_teardown(&conf_handle); 2132 goto err; 2133 2134 chip_found: 2135 pci_config_put16(conf_handle, PCI_CONF_COMM, 2136 PCI_COMM_IO | PCI_COMM_MAE | PCI_COMM_ME | 2137 pci_config_get16(conf_handle, PCI_CONF_COMM)); 2138 2139 /* ensure D0 mode */ 2140 (void) gem_pci_set_power_state(dip, conf_handle, PCI_PMCSR_D0); 2141 2142 pci_config_teardown(&conf_handle); 2143 2144 switch (cmd) { 2145 case DDI_RESUME: 2146 return (gem_resume(dip)); 2147 2148 case DDI_ATTACH: 2149 2150 DPRINTF(0, (CE_CONT, 2151 CONS "%s%d: ilr 0x%08x, latency_timer:0x%02x", 2152 drv_name, unit, iline, latim)); 2153 2154 /* 2155 * Map in the device registers. 2156 */ 2157 if (gem_pci_regs_map_setup(dip, 2158 (sfe_use_pcimemspace && p->chip_type == CHIPTYPE_DP83815) 2159 ? PCI_ADDR_MEM32 : PCI_ADDR_IO, PCI_ADDR_MASK, 2160 &sfe_dev_attr, &base, ®s_ha) != DDI_SUCCESS) { 2161 cmn_err(CE_WARN, 2162 "%s%d: ddi_regs_map_setup failed", 2163 drv_name, unit); 2164 goto err; 2165 } 2166 2167 /* 2168 * construct gem configuration 2169 */ 2170 gcp = kmem_zalloc(sizeof (*gcp), KM_SLEEP); 2171 2172 /* name */ 2173 (void) sprintf(gcp->gc_name, "%s%d", drv_name, unit); 2174 2175 /* consistency on tx and rx */ 2176 gcp->gc_tx_buf_align = sizeof (uint8_t) - 1; 2177 gcp->gc_tx_max_frags = MAXTXFRAGS; 2178 gcp->gc_tx_max_descs_per_pkt = gcp->gc_tx_max_frags; 2179 gcp->gc_tx_desc_unit_shift = 4; /* 16 byte */ 2180 gcp->gc_tx_buf_size = TX_BUF_SIZE; 2181 gcp->gc_tx_buf_limit = gcp->gc_tx_buf_size; 2182 gcp->gc_tx_ring_size = TX_RING_SIZE; 2183 gcp->gc_tx_ring_limit = gcp->gc_tx_ring_size; 2184 gcp->gc_tx_auto_pad = B_TRUE; 2185 gcp->gc_tx_copy_thresh = sfe_tx_copy_thresh; 2186 gcp->gc_tx_desc_write_oo = B_TRUE; 2187 2188 gcp->gc_rx_buf_align = sizeof (uint8_t) - 1; 2189 gcp->gc_rx_max_frags = MAXRXFRAGS; 2190 gcp->gc_rx_desc_unit_shift = 4; 2191 gcp->gc_rx_ring_size = RX_RING_SIZE; 2192 gcp->gc_rx_buf_max = RX_BUF_SIZE; 2193 gcp->gc_rx_copy_thresh = sfe_rx_copy_thresh; 2194 2195 /* map attributes */ 2196 gcp->gc_dev_attr = sfe_dev_attr; 2197 gcp->gc_buf_attr = sfe_buf_attr; 2198 gcp->gc_desc_attr = sfe_buf_attr; 2199 2200 /* dma attributes */ 2201 gcp->gc_dma_attr_desc = sfe_dma_attr_desc; 2202 2203 gcp->gc_dma_attr_txbuf = sfe_dma_attr_buf; 2204 gcp->gc_dma_attr_txbuf.dma_attr_align = gcp->gc_tx_buf_align+1; 2205 gcp->gc_dma_attr_txbuf.dma_attr_sgllen = gcp->gc_tx_max_frags; 2206 2207 gcp->gc_dma_attr_rxbuf = sfe_dma_attr_buf; 2208 gcp->gc_dma_attr_rxbuf.dma_attr_align = gcp->gc_rx_buf_align+1; 2209 gcp->gc_dma_attr_rxbuf.dma_attr_sgllen = gcp->gc_rx_max_frags; 2210 2211 /* time out parameters */ 2212 gcp->gc_tx_timeout = 3*ONESEC; 2213 gcp->gc_tx_timeout_interval = ONESEC; 2214 if (p->chip_type == CHIPTYPE_DP83815) { 2215 /* workaround for tx hang */ 2216 gcp->gc_tx_timeout_interval = ONESEC/20; /* 50mS */ 2217 } 2218 2219 /* MII timeout parameters */ 2220 gcp->gc_mii_link_watch_interval = ONESEC; 2221 gcp->gc_mii_an_watch_interval = ONESEC/5; 2222 gcp->gc_mii_reset_timeout = MII_RESET_TIMEOUT; /* 1 sec */ 2223 gcp->gc_mii_an_timeout = MII_AN_TIMEOUT; /* 5 sec */ 2224 gcp->gc_mii_an_wait = 0; 2225 gcp->gc_mii_linkdown_timeout = MII_LINKDOWN_TIMEOUT; 2226 2227 /* setting for general PHY */ 2228 gcp->gc_mii_an_delay = 0; 2229 gcp->gc_mii_linkdown_action = MII_ACTION_RSA; 2230 gcp->gc_mii_linkdown_timeout_action = MII_ACTION_RESET; 2231 gcp->gc_mii_dont_reset = B_FALSE; 2232 2233 2234 /* I/O methods */ 2235 2236 /* mac operation */ 2237 gcp->gc_attach_chip = &sfe_attach_chip; 2238 if (p->chip_type == CHIPTYPE_DP83815) { 2239 gcp->gc_reset_chip = &sfe_reset_chip_dp83815; 2240 } else { 2241 gcp->gc_reset_chip = &sfe_reset_chip_sis900; 2242 } 2243 gcp->gc_init_chip = &sfe_init_chip; 2244 gcp->gc_start_chip = &sfe_start_chip; 2245 gcp->gc_stop_chip = &sfe_stop_chip; 2246 #ifdef USE_MULTICAST_HASHTBL 2247 gcp->gc_multicast_hash = &sfe_mcast_hash; 2248 #endif 2249 if (p->chip_type == CHIPTYPE_DP83815) { 2250 gcp->gc_set_rx_filter = &sfe_set_rx_filter_dp83815; 2251 } else { 2252 gcp->gc_set_rx_filter = &sfe_set_rx_filter_sis900; 2253 } 2254 gcp->gc_set_media = &sfe_set_media; 2255 gcp->gc_get_stats = &sfe_get_stats; 2256 gcp->gc_interrupt = &sfe_interrupt; 2257 2258 /* descriptor operation */ 2259 gcp->gc_tx_desc_write = &sfe_tx_desc_write; 2260 gcp->gc_tx_start = &sfe_tx_start; 2261 gcp->gc_rx_desc_write = &sfe_rx_desc_write; 2262 gcp->gc_rx_start = NULL; 2263 2264 gcp->gc_tx_desc_stat = &sfe_tx_desc_stat; 2265 gcp->gc_rx_desc_stat = &sfe_rx_desc_stat; 2266 gcp->gc_tx_desc_init = &sfe_tx_desc_init; 2267 gcp->gc_rx_desc_init = &sfe_rx_desc_init; 2268 gcp->gc_tx_desc_clean = &sfe_tx_desc_clean; 2269 gcp->gc_rx_desc_clean = &sfe_rx_desc_clean; 2270 2271 /* mii operations */ 2272 if (p->chip_type == CHIPTYPE_DP83815) { 2273 gcp->gc_mii_probe = &sfe_mii_probe_dp83815; 2274 gcp->gc_mii_init = &sfe_mii_init_dp83815; 2275 gcp->gc_mii_config = &sfe_mii_config_dp83815; 2276 gcp->gc_mii_sync = &sfe_mii_sync_dp83815; 2277 gcp->gc_mii_read = &sfe_mii_read_dp83815; 2278 gcp->gc_mii_write = &sfe_mii_write_dp83815; 2279 gcp->gc_mii_tune_phy = NULL; 2280 gcp->gc_flow_control = FLOW_CONTROL_NONE; 2281 } else { 2282 gcp->gc_mii_probe = &gem_mii_probe_default; 2283 gcp->gc_mii_init = NULL; 2284 gcp->gc_mii_config = &sfe_mii_config_sis900; 2285 gcp->gc_mii_sync = &sfe_mii_sync_sis900; 2286 gcp->gc_mii_read = &sfe_mii_read_sis900; 2287 gcp->gc_mii_write = &sfe_mii_write_sis900; 2288 gcp->gc_mii_tune_phy = &sfe_set_eq_sis630; 2289 gcp->gc_flow_control = FLOW_CONTROL_RX_PAUSE; 2290 } 2291 2292 lp = kmem_zalloc(sizeof (*lp), KM_SLEEP); 2293 lp->chip = p; 2294 lp->revid = rev; 2295 lp->our_intr_bits = 0; 2296 lp->isr_pended = 0; 2297 2298 cmn_err(CE_CONT, CONS "%s%d: chip:%s rev:0x%02x", 2299 drv_name, unit, p->chip_name, rev); 2300 2301 dp = gem_do_attach(dip, 0, gcp, base, ®s_ha, 2302 lp, sizeof (*lp)); 2303 kmem_free(gcp, sizeof (*gcp)); 2304 2305 if (dp == NULL) { 2306 goto err_freelp; 2307 } 2308 2309 return (DDI_SUCCESS); 2310 2311 err_freelp: 2312 kmem_free(lp, sizeof (struct sfe_dev)); 2313 err: 2314 return (DDI_FAILURE); 2315 } 2316 return (DDI_FAILURE); 2317 } 2318 2319 static int 2320 sfedetach(dev_info_t *dip, ddi_detach_cmd_t cmd) 2321 { 2322 switch (cmd) { 2323 case DDI_SUSPEND: 2324 return (gem_suspend(dip)); 2325 2326 case DDI_DETACH: 2327 return (gem_do_detach(dip)); 2328 } 2329 return (DDI_FAILURE); 2330 } 2331 2332 /* 2333 * quiesce(9E) entry point. 2334 * 2335 * This function is called when the system is single-threaded at high 2336 * PIL with preemption disabled. Therefore, this function must not be 2337 * blocked. 2338 * 2339 * This function returns DDI_SUCCESS on success, or DDI_FAILURE on failure. 2340 * DDI_FAILURE indicates an error condition and should almost never happen. 2341 */ 2342 #ifdef __sparc 2343 #define sfe_quiesce ddi_quiesce_not_supported 2344 #else 2345 static int 2346 sfe_quiesce(dev_info_t *dip) 2347 { 2348 struct gem_dev *dp; 2349 int ret = 0; 2350 2351 dp = GEM_GET_DEV(dip); 2352 2353 if (dp == NULL) 2354 return (DDI_FAILURE); 2355 2356 ret = sfe_stop_chip_quiesce(dp); 2357 2358 return (ret); 2359 } 2360 #endif 2361 2362 /* ======================================================== */ 2363 /* 2364 * OS depend (loadable streams driver) routine 2365 */ 2366 /* ======================================================== */ 2367 DDI_DEFINE_STREAM_OPS(sfe_ops, nulldev, nulldev, sfeattach, sfedetach, 2368 nodev, NULL, D_MP, NULL, sfe_quiesce); 2369 2370 static struct modldrv modldrv = { 2371 &mod_driverops, /* Type of module. This one is a driver */ 2372 ident, 2373 &sfe_ops, /* driver ops */ 2374 }; 2375 2376 static struct modlinkage modlinkage = { 2377 MODREV_1, { &modldrv, NULL } 2378 }; 2379 2380 /* ======================================================== */ 2381 /* 2382 * Loadable module support 2383 */ 2384 /* ======================================================== */ 2385 int 2386 _init(void) 2387 { 2388 int status; 2389 2390 DPRINTF(2, (CE_CONT, CONS "sfe: _init: called")); 2391 gem_mod_init(&sfe_ops, "sfe"); 2392 status = mod_install(&modlinkage); 2393 if (status != DDI_SUCCESS) { 2394 gem_mod_fini(&sfe_ops); 2395 } 2396 return (status); 2397 } 2398 2399 /* 2400 * _fini : done 2401 */ 2402 int 2403 _fini(void) 2404 { 2405 int status; 2406 2407 DPRINTF(2, (CE_CONT, CONS "sfe: _fini: called")); 2408 status = mod_remove(&modlinkage); 2409 if (status == DDI_SUCCESS) { 2410 gem_mod_fini(&sfe_ops); 2411 } 2412 return (status); 2413 } 2414 2415 int 2416 _info(struct modinfo *modinfop) 2417 { 2418 return (mod_info(&modlinkage, modinfop)); 2419 }