1 /* 2 * CDDL HEADER START 3 * 4 * The contents of this file are subject to the terms of the 5 * Common Development and Distribution License (the "License"). 6 * You may not use this file except in compliance with the License. 7 * 8 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE 9 * or http://www.opensolaris.org/os/licensing. 10 * See the License for the specific language governing permissions 11 * and limitations under the License. 12 * 13 * When distributing Covered Code, include this CDDL HEADER in each 14 * file and include the License file at usr/src/OPENSOLARIS.LICENSE. 15 * If applicable, add the following below this CDDL HEADER, with the 16 * fields enclosed by brackets "[]" replaced with your own identifying 17 * information: Portions Copyright [yyyy] [name of copyright owner] 18 * 19 * CDDL HEADER END 20 */ 21 /* 22 * Copyright (c) 2006, 2010, Oracle and/or its affiliates. All rights reserved. 23 * Copyright 2016 OmniTI Computer Consulting, Inc. All rights reserved. 24 */ 25 26 /* 27 * SunOs MT STREAMS NIU/Neptune 10Gb Ethernet Device Driver. 28 */ 29 #include <sys/nxge/nxge_impl.h> 30 #include <sys/nxge/nxge_hio.h> 31 #include <sys/nxge/nxge_rxdma.h> 32 #include <sys/pcie.h> 33 34 uint32_t nxge_use_partition = 0; /* debug partition flag */ 35 uint32_t nxge_dma_obp_props_only = 1; /* use obp published props */ 36 uint32_t nxge_use_rdc_intr = 1; /* debug to assign rdc intr */ 37 /* 38 * PSARC/2007/453 MSI-X interrupt limit override 39 */ 40 uint32_t nxge_msi_enable = 2; 41 42 /* 43 * Software workaround for a Neptune (PCI-E) 44 * hardware interrupt bug which the hardware 45 * may generate spurious interrupts after the 46 * device interrupt handler was removed. If this flag 47 * is enabled, the driver will reset the 48 * hardware when devices are being detached. 49 */ 50 uint32_t nxge_peu_reset_enable = 0; 51 52 /* 53 * Software workaround for the hardware 54 * checksum bugs that affect packet transmission 55 * and receive: 56 * 57 * Usage of nxge_cksum_offload: 58 * 59 * (1) nxge_cksum_offload = 0 (default): 60 * - transmits packets: 61 * TCP: uses the hardware checksum feature. 62 * UDP: driver will compute the software checksum 63 * based on the partial checksum computed 64 * by the IP layer. 65 * - receives packets 66 * TCP: marks packets checksum flags based on hardware result. 67 * UDP: will not mark checksum flags. 68 * 69 * (2) nxge_cksum_offload = 1: 70 * - transmit packets: 71 * TCP/UDP: uses the hardware checksum feature. 72 * - receives packets 73 * TCP/UDP: marks packet checksum flags based on hardware result. 74 * 75 * (3) nxge_cksum_offload = 2: 76 * - The driver will not register its checksum capability. 77 * Checksum for both TCP and UDP will be computed 78 * by the stack. 79 * - The software LSO is not allowed in this case. 80 * 81 * (4) nxge_cksum_offload > 2: 82 * - Will be treated as it is set to 2 83 * (stack will compute the checksum). 84 * 85 * (5) If the hardware bug is fixed, this workaround 86 * needs to be updated accordingly to reflect 87 * the new hardware revision. 88 */ 89 uint32_t nxge_cksum_offload = 0; 90 91 /* 92 * Globals: tunable parameters (/etc/system or adb) 93 * 94 */ 95 uint32_t nxge_rbr_size = NXGE_RBR_RBB_DEFAULT; 96 uint32_t nxge_rbr_spare_size = 0; 97 uint32_t nxge_rcr_size = NXGE_RCR_DEFAULT; 98 uint16_t nxge_rdc_buf_offset = SW_OFFSET_NO_OFFSET; 99 uint32_t nxge_tx_ring_size = NXGE_TX_RING_DEFAULT; 100 boolean_t nxge_no_msg = B_TRUE; /* control message display */ 101 uint32_t nxge_no_link_notify = 0; /* control DL_NOTIFY */ 102 uint32_t nxge_bcopy_thresh = TX_BCOPY_MAX; 103 uint32_t nxge_dvma_thresh = TX_FASTDVMA_MIN; 104 uint32_t nxge_dma_stream_thresh = TX_STREAM_MIN; 105 uint32_t nxge_jumbo_mtu = TX_JUMBO_MTU; 106 nxge_tx_mode_t nxge_tx_scheme = NXGE_USE_SERIAL; 107 108 /* MAX LSO size */ 109 #define NXGE_LSO_MAXLEN 65535 110 uint32_t nxge_lso_max = NXGE_LSO_MAXLEN; 111 112 113 /* 114 * Add tunable to reduce the amount of time spent in the 115 * ISR doing Rx Processing. 116 */ 117 uint32_t nxge_max_rx_pkts = 1024; 118 119 /* 120 * Tunables to manage the receive buffer blocks. 121 * 122 * nxge_rx_threshold_hi: copy all buffers. 123 * nxge_rx_bcopy_size_type: receive buffer block size type. 124 * nxge_rx_threshold_lo: copy only up to tunable block size type. 125 */ 126 nxge_rxbuf_threshold_t nxge_rx_threshold_hi = NXGE_RX_COPY_6; 127 nxge_rxbuf_type_t nxge_rx_buf_size_type = RCR_PKTBUFSZ_0; 128 nxge_rxbuf_threshold_t nxge_rx_threshold_lo = NXGE_RX_COPY_3; 129 130 /* Use kmem_alloc() to allocate data buffers. */ 131 #if defined(__sparc) 132 uint32_t nxge_use_kmem_alloc = 1; 133 #elif defined(__i386) 134 uint32_t nxge_use_kmem_alloc = 0; 135 #else 136 uint32_t nxge_use_kmem_alloc = 1; 137 #endif 138 139 rtrace_t npi_rtracebuf; 140 141 /* 142 * The hardware sometimes fails to allow enough time for the link partner 143 * to send an acknowledgement for packets that the hardware sent to it. The 144 * hardware resends the packets earlier than it should be in those instances. 145 * This behavior caused some switches to acknowledge the wrong packets 146 * and it triggered the fatal error. 147 * This software workaround is to set the replay timer to a value 148 * suggested by the hardware team. 149 * 150 * PCI config space replay timer register: 151 * The following replay timeout value is 0xc 152 * for bit 14:18. 153 */ 154 #define PCI_REPLAY_TIMEOUT_CFG_OFFSET 0xb8 155 #define PCI_REPLAY_TIMEOUT_SHIFT 14 156 157 uint32_t nxge_set_replay_timer = 1; 158 uint32_t nxge_replay_timeout = 0xc; 159 160 /* 161 * The transmit serialization sometimes causes 162 * longer sleep before calling the driver transmit 163 * function as it sleeps longer than it should. 164 * The performace group suggests that a time wait tunable 165 * can be used to set the maximum wait time when needed 166 * and the default is set to 1 tick. 167 */ 168 uint32_t nxge_tx_serial_maxsleep = 1; 169 170 #if defined(sun4v) 171 /* 172 * Hypervisor N2/NIU services information. 173 */ 174 /* 175 * The following is the default API supported: 176 * major 1 and minor 1. 177 * 178 * Please update the MAX_NIU_MAJORS, 179 * MAX_NIU_MINORS, and minor number supported 180 * when the newer Hypervior API interfaces 181 * are added. Also, please update nxge_hsvc_register() 182 * if needed. 183 */ 184 static hsvc_info_t niu_hsvc = { 185 HSVC_REV_1, NULL, HSVC_GROUP_NIU, NIU_MAJOR_VER, 186 NIU_MINOR_VER, "nxge" 187 }; 188 189 static int nxge_hsvc_register(p_nxge_t); 190 #endif 191 192 /* 193 * Function Prototypes 194 */ 195 static int nxge_attach(dev_info_t *, ddi_attach_cmd_t); 196 static int nxge_detach(dev_info_t *, ddi_detach_cmd_t); 197 static void nxge_unattach(p_nxge_t); 198 static int nxge_quiesce(dev_info_t *); 199 200 #if NXGE_PROPERTY 201 static void nxge_remove_hard_properties(p_nxge_t); 202 #endif 203 204 /* 205 * These two functions are required by nxge_hio.c 206 */ 207 extern int nxge_m_mmac_remove(void *arg, int slot); 208 extern void nxge_grp_cleanup(p_nxge_t nxge); 209 210 static nxge_status_t nxge_setup_system_dma_pages(p_nxge_t); 211 212 static nxge_status_t nxge_setup_mutexes(p_nxge_t); 213 static void nxge_destroy_mutexes(p_nxge_t); 214 215 static nxge_status_t nxge_map_regs(p_nxge_t nxgep); 216 static void nxge_unmap_regs(p_nxge_t nxgep); 217 #ifdef NXGE_DEBUG 218 static void nxge_test_map_regs(p_nxge_t nxgep); 219 #endif 220 221 static nxge_status_t nxge_add_intrs(p_nxge_t nxgep); 222 static void nxge_remove_intrs(p_nxge_t nxgep); 223 224 static nxge_status_t nxge_add_intrs_adv(p_nxge_t nxgep); 225 static nxge_status_t nxge_add_intrs_adv_type(p_nxge_t, uint32_t); 226 static nxge_status_t nxge_add_intrs_adv_type_fix(p_nxge_t, uint32_t); 227 static void nxge_intrs_enable(p_nxge_t nxgep); 228 static void nxge_intrs_disable(p_nxge_t nxgep); 229 230 static void nxge_suspend(p_nxge_t); 231 static nxge_status_t nxge_resume(p_nxge_t); 232 233 static nxge_status_t nxge_setup_dev(p_nxge_t); 234 static void nxge_destroy_dev(p_nxge_t); 235 236 static nxge_status_t nxge_alloc_mem_pool(p_nxge_t); 237 static void nxge_free_mem_pool(p_nxge_t); 238 239 nxge_status_t nxge_alloc_rx_mem_pool(p_nxge_t); 240 static void nxge_free_rx_mem_pool(p_nxge_t); 241 242 nxge_status_t nxge_alloc_tx_mem_pool(p_nxge_t); 243 static void nxge_free_tx_mem_pool(p_nxge_t); 244 245 static nxge_status_t nxge_dma_mem_alloc(p_nxge_t, dma_method_t, 246 struct ddi_dma_attr *, 247 size_t, ddi_device_acc_attr_t *, uint_t, 248 p_nxge_dma_common_t); 249 250 static void nxge_dma_mem_free(p_nxge_dma_common_t); 251 static void nxge_dma_free_rx_data_buf(p_nxge_dma_common_t); 252 253 static nxge_status_t nxge_alloc_rx_buf_dma(p_nxge_t, uint16_t, 254 p_nxge_dma_common_t *, size_t, size_t, uint32_t *); 255 static void nxge_free_rx_buf_dma(p_nxge_t, p_nxge_dma_common_t, uint32_t); 256 257 static nxge_status_t nxge_alloc_rx_cntl_dma(p_nxge_t, uint16_t, 258 p_nxge_dma_common_t *, size_t); 259 static void nxge_free_rx_cntl_dma(p_nxge_t, p_nxge_dma_common_t); 260 261 extern nxge_status_t nxge_alloc_tx_buf_dma(p_nxge_t, uint16_t, 262 p_nxge_dma_common_t *, size_t, size_t, uint32_t *); 263 static void nxge_free_tx_buf_dma(p_nxge_t, p_nxge_dma_common_t, uint32_t); 264 265 extern nxge_status_t nxge_alloc_tx_cntl_dma(p_nxge_t, uint16_t, 266 p_nxge_dma_common_t *, 267 size_t); 268 static void nxge_free_tx_cntl_dma(p_nxge_t, p_nxge_dma_common_t); 269 270 static int nxge_init_common_dev(p_nxge_t); 271 static void nxge_uninit_common_dev(p_nxge_t); 272 extern int nxge_param_set_mac(p_nxge_t, queue_t *, mblk_t *, 273 char *, caddr_t); 274 #if defined(sun4v) 275 extern nxge_status_t nxge_hio_rdc_enable(p_nxge_t nxgep); 276 extern nxge_status_t nxge_hio_rdc_intr_arm(p_nxge_t nxge, boolean_t arm); 277 #endif 278 279 /* 280 * The next declarations are for the GLDv3 interface. 281 */ 282 static int nxge_m_start(void *); 283 static void nxge_m_stop(void *); 284 static int nxge_m_multicst(void *, boolean_t, const uint8_t *); 285 static int nxge_m_promisc(void *, boolean_t); 286 static void nxge_m_ioctl(void *, queue_t *, mblk_t *); 287 nxge_status_t nxge_mac_register(p_nxge_t); 288 static int nxge_altmac_set(p_nxge_t nxgep, uint8_t *mac_addr, 289 int slot, int rdctbl, boolean_t usetbl); 290 void nxge_mmac_kstat_update(p_nxge_t nxgep, int slot, 291 boolean_t factory); 292 293 static void nxge_m_getfactaddr(void *, uint_t, uint8_t *); 294 static boolean_t nxge_m_getcapab(void *, mac_capab_t, void *); 295 static int nxge_m_setprop(void *, const char *, mac_prop_id_t, 296 uint_t, const void *); 297 static int nxge_m_getprop(void *, const char *, mac_prop_id_t, 298 uint_t, void *); 299 static void nxge_m_propinfo(void *, const char *, mac_prop_id_t, 300 mac_prop_info_handle_t); 301 static void nxge_priv_propinfo(const char *, mac_prop_info_handle_t); 302 static int nxge_set_priv_prop(nxge_t *, const char *, uint_t, 303 const void *); 304 static int nxge_get_priv_prop(nxge_t *, const char *, uint_t, void *); 305 static void nxge_fill_ring(void *, mac_ring_type_t, const int, const int, 306 mac_ring_info_t *, mac_ring_handle_t); 307 static void nxge_group_add_ring(mac_group_driver_t, mac_ring_driver_t, 308 mac_ring_type_t); 309 static void nxge_group_rem_ring(mac_group_driver_t, mac_ring_driver_t, 310 mac_ring_type_t); 311 312 static void nxge_niu_peu_reset(p_nxge_t nxgep); 313 static void nxge_set_pci_replay_timeout(nxge_t *); 314 315 char *nxge_priv_props[] = { 316 "_adv_10gfdx_cap", 317 "_adv_pause_cap", 318 "_function_number", 319 "_fw_version", 320 "_port_mode", 321 "_hot_swap_phy", 322 "_rxdma_intr_time", 323 "_rxdma_intr_pkts", 324 "_class_opt_ipv4_tcp", 325 "_class_opt_ipv4_udp", 326 "_class_opt_ipv4_ah", 327 "_class_opt_ipv4_sctp", 328 "_class_opt_ipv6_tcp", 329 "_class_opt_ipv6_udp", 330 "_class_opt_ipv6_ah", 331 "_class_opt_ipv6_sctp", 332 "_soft_lso_enable", 333 NULL 334 }; 335 336 #define NXGE_NEPTUNE_MAGIC 0x4E584745UL 337 #define MAX_DUMP_SZ 256 338 339 #define NXGE_M_CALLBACK_FLAGS \ 340 (MC_IOCTL | MC_GETCAPAB | MC_SETPROP | MC_GETPROP | MC_PROPINFO) 341 342 mac_callbacks_t nxge_m_callbacks = { 343 NXGE_M_CALLBACK_FLAGS, 344 nxge_m_stat, 345 nxge_m_start, 346 nxge_m_stop, 347 nxge_m_promisc, 348 nxge_m_multicst, 349 NULL, 350 NULL, 351 NULL, 352 nxge_m_ioctl, 353 nxge_m_getcapab, 354 NULL, 355 NULL, 356 nxge_m_setprop, 357 nxge_m_getprop, 358 nxge_m_propinfo 359 }; 360 361 void 362 nxge_err_inject(p_nxge_t, queue_t *, mblk_t *); 363 364 /* PSARC/2007/453 MSI-X interrupt limit override. */ 365 #define NXGE_MSIX_REQUEST_10G 8 366 #define NXGE_MSIX_REQUEST_1G 2 367 static int nxge_create_msi_property(p_nxge_t); 368 /* 369 * For applications that care about the 370 * latency, it was requested by PAE and the 371 * customers that the driver has tunables that 372 * allow the user to tune it to a higher number 373 * interrupts to spread the interrupts among 374 * multiple channels. The DDI framework limits 375 * the maximum number of MSI-X resources to allocate 376 * to 8 (ddi_msix_alloc_limit). If more than 8 377 * is set, ddi_msix_alloc_limit must be set accordingly. 378 * The default number of MSI interrupts are set to 379 * 8 for 10G and 2 for 1G link. 380 */ 381 #define NXGE_MSIX_MAX_ALLOWED 32 382 uint32_t nxge_msix_10g_intrs = NXGE_MSIX_REQUEST_10G; 383 uint32_t nxge_msix_1g_intrs = NXGE_MSIX_REQUEST_1G; 384 385 /* 386 * These global variables control the message 387 * output. 388 */ 389 out_dbgmsg_t nxge_dbgmsg_out = DBG_CONSOLE | STR_LOG; 390 uint64_t nxge_debug_level; 391 392 /* 393 * This list contains the instance structures for the Neptune 394 * devices present in the system. The lock exists to guarantee 395 * mutually exclusive access to the list. 396 */ 397 void *nxge_list = NULL; 398 void *nxge_hw_list = NULL; 399 nxge_os_mutex_t nxge_common_lock; 400 nxge_os_mutex_t nxgedebuglock; 401 402 extern uint64_t npi_debug_level; 403 404 extern nxge_status_t nxge_ldgv_init(p_nxge_t, int *, int *); 405 extern nxge_status_t nxge_ldgv_init_n2(p_nxge_t, int *, int *); 406 extern nxge_status_t nxge_ldgv_uninit(p_nxge_t); 407 extern nxge_status_t nxge_intr_ldgv_init(p_nxge_t); 408 extern void nxge_fm_init(p_nxge_t, 409 ddi_device_acc_attr_t *, 410 ddi_dma_attr_t *); 411 extern void nxge_fm_fini(p_nxge_t); 412 extern npi_status_t npi_mac_altaddr_disable(npi_handle_t, uint8_t, uint8_t); 413 414 /* 415 * Count used to maintain the number of buffers being used 416 * by Neptune instances and loaned up to the upper layers. 417 */ 418 uint32_t nxge_mblks_pending = 0; 419 420 /* 421 * Device register access attributes for PIO. 422 */ 423 static ddi_device_acc_attr_t nxge_dev_reg_acc_attr = { 424 DDI_DEVICE_ATTR_V1, 425 DDI_STRUCTURE_LE_ACC, 426 DDI_STRICTORDER_ACC, 427 DDI_DEFAULT_ACC 428 }; 429 430 /* 431 * Device descriptor access attributes for DMA. 432 */ 433 static ddi_device_acc_attr_t nxge_dev_desc_dma_acc_attr = { 434 DDI_DEVICE_ATTR_V0, 435 DDI_STRUCTURE_LE_ACC, 436 DDI_STRICTORDER_ACC 437 }; 438 439 /* 440 * Device buffer access attributes for DMA. 441 */ 442 static ddi_device_acc_attr_t nxge_dev_buf_dma_acc_attr = { 443 DDI_DEVICE_ATTR_V0, 444 DDI_STRUCTURE_BE_ACC, 445 DDI_STRICTORDER_ACC 446 }; 447 448 ddi_dma_attr_t nxge_desc_dma_attr = { 449 DMA_ATTR_V0, /* version number. */ 450 0, /* low address */ 451 0xffffffffffffffff, /* high address */ 452 0xffffffffffffffff, /* address counter max */ 453 #ifndef NIU_PA_WORKAROUND 454 0x100000, /* alignment */ 455 #else 456 0x2000, 457 #endif 458 0xfc00fc, /* dlim_burstsizes */ 459 0x1, /* minimum transfer size */ 460 0xffffffffffffffff, /* maximum transfer size */ 461 0xffffffffffffffff, /* maximum segment size */ 462 1, /* scatter/gather list length */ 463 (unsigned int) 1, /* granularity */ 464 0 /* attribute flags */ 465 }; 466 467 ddi_dma_attr_t nxge_tx_dma_attr = { 468 DMA_ATTR_V0, /* version number. */ 469 0, /* low address */ 470 0xffffffffffffffff, /* high address */ 471 0xffffffffffffffff, /* address counter max */ 472 #if defined(_BIG_ENDIAN) 473 0x2000, /* alignment */ 474 #else 475 0x1000, /* alignment */ 476 #endif 477 0xfc00fc, /* dlim_burstsizes */ 478 0x1, /* minimum transfer size */ 479 0xffffffffffffffff, /* maximum transfer size */ 480 0xffffffffffffffff, /* maximum segment size */ 481 5, /* scatter/gather list length */ 482 (unsigned int) 1, /* granularity */ 483 0 /* attribute flags */ 484 }; 485 486 ddi_dma_attr_t nxge_rx_dma_attr = { 487 DMA_ATTR_V0, /* version number. */ 488 0, /* low address */ 489 0xffffffffffffffff, /* high address */ 490 0xffffffffffffffff, /* address counter max */ 491 0x2000, /* alignment */ 492 0xfc00fc, /* dlim_burstsizes */ 493 0x1, /* minimum transfer size */ 494 0xffffffffffffffff, /* maximum transfer size */ 495 0xffffffffffffffff, /* maximum segment size */ 496 1, /* scatter/gather list length */ 497 (unsigned int) 1, /* granularity */ 498 DDI_DMA_RELAXED_ORDERING /* attribute flags */ 499 }; 500 501 ddi_dma_lim_t nxge_dma_limits = { 502 (uint_t)0, /* dlim_addr_lo */ 503 (uint_t)0xffffffff, /* dlim_addr_hi */ 504 (uint_t)0xffffffff, /* dlim_cntr_max */ 505 (uint_t)0xfc00fc, /* dlim_burstsizes for 32 and 64 bit xfers */ 506 0x1, /* dlim_minxfer */ 507 1024 /* dlim_speed */ 508 }; 509 510 dma_method_t nxge_force_dma = DVMA; 511 512 /* 513 * dma chunk sizes. 514 * 515 * Try to allocate the largest possible size 516 * so that fewer number of dma chunks would be managed 517 */ 518 #ifdef NIU_PA_WORKAROUND 519 size_t alloc_sizes [] = {0x2000}; 520 #else 521 size_t alloc_sizes [] = {0x1000, 0x2000, 0x4000, 0x8000, 522 0x10000, 0x20000, 0x40000, 0x80000, 523 0x100000, 0x200000, 0x400000, 0x800000, 524 0x1000000, 0x2000000, 0x4000000}; 525 #endif 526 527 /* 528 * Translate "dev_t" to a pointer to the associated "dev_info_t". 529 */ 530 531 extern void nxge_get_environs(nxge_t *); 532 533 static int 534 nxge_attach(dev_info_t *dip, ddi_attach_cmd_t cmd) 535 { 536 p_nxge_t nxgep = NULL; 537 int instance; 538 int status = DDI_SUCCESS; 539 uint8_t portn; 540 nxge_mmac_t *mmac_info; 541 542 NXGE_DEBUG_MSG((nxgep, DDI_CTL, "==> nxge_attach")); 543 544 /* 545 * Get the device instance since we'll need to setup 546 * or retrieve a soft state for this instance. 547 */ 548 instance = ddi_get_instance(dip); 549 550 switch (cmd) { 551 case DDI_ATTACH: 552 NXGE_DEBUG_MSG((nxgep, DDI_CTL, "doing DDI_ATTACH")); 553 break; 554 555 case DDI_RESUME: 556 NXGE_DEBUG_MSG((nxgep, DDI_CTL, "doing DDI_RESUME")); 557 nxgep = (p_nxge_t)ddi_get_soft_state(nxge_list, instance); 558 if (nxgep == NULL) { 559 status = DDI_FAILURE; 560 break; 561 } 562 if (nxgep->dip != dip) { 563 status = DDI_FAILURE; 564 break; 565 } 566 if (nxgep->suspended == DDI_PM_SUSPEND) { 567 status = ddi_dev_is_needed(nxgep->dip, 0, 1); 568 } else { 569 status = nxge_resume(nxgep); 570 } 571 goto nxge_attach_exit; 572 573 case DDI_PM_RESUME: 574 NXGE_DEBUG_MSG((nxgep, DDI_CTL, "doing DDI_PM_RESUME")); 575 nxgep = (p_nxge_t)ddi_get_soft_state(nxge_list, instance); 576 if (nxgep == NULL) { 577 status = DDI_FAILURE; 578 break; 579 } 580 if (nxgep->dip != dip) { 581 status = DDI_FAILURE; 582 break; 583 } 584 status = nxge_resume(nxgep); 585 goto nxge_attach_exit; 586 587 default: 588 NXGE_DEBUG_MSG((nxgep, DDI_CTL, "doing unknown")); 589 status = DDI_FAILURE; 590 goto nxge_attach_exit; 591 } 592 593 594 if (ddi_soft_state_zalloc(nxge_list, instance) == DDI_FAILURE) { 595 status = DDI_FAILURE; 596 goto nxge_attach_exit; 597 } 598 599 nxgep = ddi_get_soft_state(nxge_list, instance); 600 if (nxgep == NULL) { 601 status = NXGE_ERROR; 602 goto nxge_attach_fail2; 603 } 604 605 nxgep->nxge_magic = NXGE_MAGIC; 606 607 nxgep->drv_state = 0; 608 nxgep->dip = dip; 609 nxgep->instance = instance; 610 nxgep->p_dip = ddi_get_parent(dip); 611 nxgep->nxge_debug_level = nxge_debug_level; 612 npi_debug_level = nxge_debug_level; 613 614 /* Are we a guest running in a Hybrid I/O environment? */ 615 nxge_get_environs(nxgep); 616 617 status = nxge_map_regs(nxgep); 618 619 if (status != NXGE_OK) { 620 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, "nxge_map_regs failed")); 621 goto nxge_attach_fail3; 622 } 623 624 nxge_fm_init(nxgep, &nxge_dev_reg_acc_attr, &nxge_rx_dma_attr); 625 626 /* Create & initialize the per-Neptune data structure */ 627 /* (even if we're a guest). */ 628 status = nxge_init_common_dev(nxgep); 629 if (status != NXGE_OK) { 630 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, 631 "nxge_init_common_dev failed")); 632 goto nxge_attach_fail4; 633 } 634 635 /* 636 * Software workaround: set the replay timer. 637 */ 638 if (nxgep->niu_type != N2_NIU) { 639 nxge_set_pci_replay_timeout(nxgep); 640 } 641 642 #if defined(sun4v) 643 /* This is required by nxge_hio_init(), which follows. */ 644 if ((status = nxge_hsvc_register(nxgep)) != DDI_SUCCESS) 645 goto nxge_attach_fail4; 646 #endif 647 648 if ((status = nxge_hio_init(nxgep)) != NXGE_OK) { 649 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, 650 "nxge_hio_init failed")); 651 goto nxge_attach_fail4; 652 } 653 654 if (nxgep->niu_type == NEPTUNE_2_10GF) { 655 if (nxgep->function_num > 1) { 656 NXGE_DEBUG_MSG((nxgep, DDI_CTL, "Unsupported" 657 " function %d. Only functions 0 and 1 are " 658 "supported for this card.", nxgep->function_num)); 659 status = NXGE_ERROR; 660 goto nxge_attach_fail4; 661 } 662 } 663 664 if (isLDOMguest(nxgep)) { 665 /* 666 * Use the function number here. 667 */ 668 nxgep->mac.portnum = nxgep->function_num; 669 nxgep->mac.porttype = PORT_TYPE_LOGICAL; 670 671 /* XXX We'll set the MAC address counts to 1 for now. */ 672 mmac_info = &nxgep->nxge_mmac_info; 673 mmac_info->num_mmac = 1; 674 mmac_info->naddrfree = 1; 675 } else { 676 portn = NXGE_GET_PORT_NUM(nxgep->function_num); 677 nxgep->mac.portnum = portn; 678 if ((portn == 0) || (portn == 1)) 679 nxgep->mac.porttype = PORT_TYPE_XMAC; 680 else 681 nxgep->mac.porttype = PORT_TYPE_BMAC; 682 /* 683 * Neptune has 4 ports, the first 2 ports use XMAC (10G MAC) 684 * internally, the rest 2 ports use BMAC (1G "Big" MAC). 685 * The two types of MACs have different characterizations. 686 */ 687 mmac_info = &nxgep->nxge_mmac_info; 688 if (nxgep->function_num < 2) { 689 mmac_info->num_mmac = XMAC_MAX_ALT_ADDR_ENTRY; 690 mmac_info->naddrfree = XMAC_MAX_ALT_ADDR_ENTRY; 691 } else { 692 mmac_info->num_mmac = BMAC_MAX_ALT_ADDR_ENTRY; 693 mmac_info->naddrfree = BMAC_MAX_ALT_ADDR_ENTRY; 694 } 695 } 696 /* 697 * Setup the Ndd parameters for the this instance. 698 */ 699 nxge_init_param(nxgep); 700 701 /* 702 * Setup Register Tracing Buffer. 703 */ 704 npi_rtrace_buf_init((rtrace_t *)&npi_rtracebuf); 705 706 /* init stats ptr */ 707 nxge_init_statsp(nxgep); 708 709 /* 710 * Copy the vpd info from eeprom to a local data 711 * structure, and then check its validity. 712 */ 713 if (!isLDOMguest(nxgep)) { 714 int *regp; 715 uint_t reglen; 716 int rv; 717 718 nxge_vpd_info_get(nxgep); 719 720 /* Find the NIU config handle. */ 721 rv = ddi_prop_lookup_int_array(DDI_DEV_T_ANY, 722 ddi_get_parent(nxgep->dip), DDI_PROP_DONTPASS, 723 "reg", ®p, ®len); 724 725 if (rv != DDI_PROP_SUCCESS) { 726 goto nxge_attach_fail5; 727 } 728 /* 729 * The address_hi, that is the first int, in the reg 730 * property consists of config handle, but need to remove 731 * the bits 28-31 which are OBP specific info. 732 */ 733 nxgep->niu_cfg_hdl = (*regp) & 0xFFFFFFF; 734 ddi_prop_free(regp); 735 } 736 737 /* 738 * Set the defaults for the MTU size. 739 */ 740 nxge_hw_id_init(nxgep); 741 742 if (isLDOMguest(nxgep)) { 743 uchar_t *prop_val; 744 uint_t prop_len; 745 uint32_t max_frame_size; 746 747 extern void nxge_get_logical_props(p_nxge_t); 748 749 nxgep->statsp->mac_stats.xcvr_inuse = LOGICAL_XCVR; 750 nxgep->mac.portmode = PORT_LOGICAL; 751 (void) ddi_prop_update_string(DDI_DEV_T_NONE, nxgep->dip, 752 "phy-type", "virtual transceiver"); 753 754 nxgep->nports = 1; 755 nxgep->board_ver = 0; /* XXX What? */ 756 757 /* 758 * local-mac-address property gives us info on which 759 * specific MAC address the Hybrid resource is associated 760 * with. 761 */ 762 if (ddi_prop_lookup_byte_array(DDI_DEV_T_ANY, nxgep->dip, 0, 763 "local-mac-address", &prop_val, 764 &prop_len) != DDI_PROP_SUCCESS) { 765 goto nxge_attach_fail5; 766 } 767 if (prop_len != ETHERADDRL) { 768 ddi_prop_free(prop_val); 769 goto nxge_attach_fail5; 770 } 771 ether_copy(prop_val, nxgep->hio_mac_addr); 772 ddi_prop_free(prop_val); 773 nxge_get_logical_props(nxgep); 774 775 /* 776 * Enable Jumbo property based on the "max-frame-size" 777 * property value. 778 */ 779 max_frame_size = ddi_prop_get_int(DDI_DEV_T_ANY, 780 nxgep->dip, DDI_PROP_DONTPASS | DDI_PROP_NOTPROM, 781 "max-frame-size", NXGE_MTU_DEFAULT_MAX); 782 if ((max_frame_size > NXGE_MTU_DEFAULT_MAX) && 783 (max_frame_size <= TX_JUMBO_MTU)) { 784 nxgep->mac.is_jumbo = B_TRUE; 785 nxgep->mac.maxframesize = (uint16_t)max_frame_size; 786 nxgep->mac.default_mtu = nxgep->mac.maxframesize - 787 NXGE_EHEADER_VLAN_CRC; 788 } 789 } else { 790 status = nxge_xcvr_find(nxgep); 791 792 if (status != NXGE_OK) { 793 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, "nxge_attach: " 794 " Couldn't determine card type" 795 " .... exit ")); 796 goto nxge_attach_fail5; 797 } 798 799 status = nxge_get_config_properties(nxgep); 800 801 if (status != NXGE_OK) { 802 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, 803 "get_hw create failed")); 804 goto nxge_attach_fail; 805 } 806 } 807 808 /* 809 * Setup the Kstats for the driver. 810 */ 811 nxge_setup_kstats(nxgep); 812 813 if (!isLDOMguest(nxgep)) 814 nxge_setup_param(nxgep); 815 816 status = nxge_setup_system_dma_pages(nxgep); 817 if (status != NXGE_OK) { 818 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, "set dma page failed")); 819 goto nxge_attach_fail; 820 } 821 822 823 if (!isLDOMguest(nxgep)) 824 nxge_hw_init_niu_common(nxgep); 825 826 status = nxge_setup_mutexes(nxgep); 827 if (status != NXGE_OK) { 828 NXGE_DEBUG_MSG((nxgep, DDI_CTL, "set mutex failed")); 829 goto nxge_attach_fail; 830 } 831 832 #if defined(sun4v) 833 if (isLDOMguest(nxgep)) { 834 /* Find our VR & channel sets. */ 835 status = nxge_hio_vr_add(nxgep); 836 if (status != DDI_SUCCESS) { 837 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, 838 "nxge_hio_vr_add failed")); 839 (void) hsvc_unregister(&nxgep->niu_hsvc); 840 nxgep->niu_hsvc_available = B_FALSE; 841 goto nxge_attach_fail; 842 } 843 goto nxge_attach_exit; 844 } 845 #endif 846 847 status = nxge_setup_dev(nxgep); 848 if (status != DDI_SUCCESS) { 849 NXGE_DEBUG_MSG((nxgep, DDI_CTL, "set dev failed")); 850 goto nxge_attach_fail; 851 } 852 853 status = nxge_add_intrs(nxgep); 854 if (status != DDI_SUCCESS) { 855 NXGE_DEBUG_MSG((nxgep, DDI_CTL, "add_intr failed")); 856 goto nxge_attach_fail; 857 } 858 859 /* If a guest, register with vio_net instead. */ 860 if ((status = nxge_mac_register(nxgep)) != NXGE_OK) { 861 NXGE_DEBUG_MSG((nxgep, DDI_CTL, 862 "unable to register to mac layer (%d)", status)); 863 goto nxge_attach_fail; 864 } 865 866 mac_link_update(nxgep->mach, LINK_STATE_UNKNOWN); 867 868 NXGE_DEBUG_MSG((nxgep, DDI_CTL, 869 "registered to mac (instance %d)", instance)); 870 871 /* nxge_link_monitor calls xcvr.check_link recursively */ 872 (void) nxge_link_monitor(nxgep, LINK_MONITOR_START); 873 874 goto nxge_attach_exit; 875 876 nxge_attach_fail: 877 nxge_unattach(nxgep); 878 goto nxge_attach_fail1; 879 880 nxge_attach_fail5: 881 /* 882 * Tear down the ndd parameters setup. 883 */ 884 nxge_destroy_param(nxgep); 885 886 /* 887 * Tear down the kstat setup. 888 */ 889 nxge_destroy_kstats(nxgep); 890 891 nxge_attach_fail4: 892 if (nxgep->nxge_hw_p) { 893 nxge_uninit_common_dev(nxgep); 894 nxgep->nxge_hw_p = NULL; 895 } 896 897 nxge_attach_fail3: 898 /* 899 * Unmap the register setup. 900 */ 901 nxge_unmap_regs(nxgep); 902 903 nxge_fm_fini(nxgep); 904 905 nxge_attach_fail2: 906 ddi_soft_state_free(nxge_list, nxgep->instance); 907 908 nxge_attach_fail1: 909 if (status != NXGE_OK) 910 status = (NXGE_ERROR | NXGE_DDI_FAILED); 911 nxgep = NULL; 912 913 nxge_attach_exit: 914 NXGE_DEBUG_MSG((nxgep, DDI_CTL, "<== nxge_attach status = 0x%08x", 915 status)); 916 917 return (status); 918 } 919 920 static int 921 nxge_detach(dev_info_t *dip, ddi_detach_cmd_t cmd) 922 { 923 int status = DDI_SUCCESS; 924 int instance; 925 p_nxge_t nxgep = NULL; 926 927 NXGE_DEBUG_MSG((nxgep, DDI_CTL, "==> nxge_detach")); 928 instance = ddi_get_instance(dip); 929 nxgep = ddi_get_soft_state(nxge_list, instance); 930 if (nxgep == NULL) { 931 status = DDI_FAILURE; 932 goto nxge_detach_exit; 933 } 934 935 switch (cmd) { 936 case DDI_DETACH: 937 NXGE_DEBUG_MSG((nxgep, DDI_CTL, "doing DDI_DETACH")); 938 break; 939 940 case DDI_PM_SUSPEND: 941 NXGE_DEBUG_MSG((nxgep, DDI_CTL, "doing DDI_PM_SUSPEND")); 942 nxgep->suspended = DDI_PM_SUSPEND; 943 nxge_suspend(nxgep); 944 break; 945 946 case DDI_SUSPEND: 947 NXGE_DEBUG_MSG((nxgep, DDI_CTL, "doing DDI_SUSPEND")); 948 if (nxgep->suspended != DDI_PM_SUSPEND) { 949 nxgep->suspended = DDI_SUSPEND; 950 nxge_suspend(nxgep); 951 } 952 break; 953 954 default: 955 status = DDI_FAILURE; 956 } 957 958 if (cmd != DDI_DETACH) 959 goto nxge_detach_exit; 960 961 /* 962 * Stop the xcvr polling. 963 */ 964 nxgep->suspended = cmd; 965 966 (void) nxge_link_monitor(nxgep, LINK_MONITOR_STOP); 967 968 if (nxgep->mach && (status = mac_unregister(nxgep->mach)) != 0) { 969 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, 970 "<== nxge_detach status = 0x%08X", status)); 971 return (DDI_FAILURE); 972 } 973 974 NXGE_DEBUG_MSG((nxgep, DDI_CTL, 975 "<== nxge_detach (mac_unregister) status = 0x%08X", status)); 976 977 nxge_unattach(nxgep); 978 nxgep = NULL; 979 980 nxge_detach_exit: 981 NXGE_DEBUG_MSG((nxgep, DDI_CTL, "<== nxge_detach status = 0x%08X", 982 status)); 983 984 return (status); 985 } 986 987 static void 988 nxge_unattach(p_nxge_t nxgep) 989 { 990 NXGE_DEBUG_MSG((nxgep, DDI_CTL, "==> nxge_unattach")); 991 992 if (nxgep == NULL || nxgep->dev_regs == NULL) { 993 return; 994 } 995 996 nxgep->nxge_magic = 0; 997 998 if (nxgep->nxge_timerid) { 999 nxge_stop_timer(nxgep, nxgep->nxge_timerid); 1000 nxgep->nxge_timerid = 0; 1001 } 1002 1003 /* 1004 * If this flag is set, it will affect the Neptune 1005 * only. 1006 */ 1007 if ((nxgep->niu_type != N2_NIU) && nxge_peu_reset_enable) { 1008 nxge_niu_peu_reset(nxgep); 1009 } 1010 1011 #if defined(sun4v) 1012 if (isLDOMguest(nxgep)) { 1013 (void) nxge_hio_vr_release(nxgep); 1014 } 1015 #endif 1016 1017 if (nxgep->nxge_hw_p) { 1018 nxge_uninit_common_dev(nxgep); 1019 nxgep->nxge_hw_p = NULL; 1020 } 1021 1022 #if defined(sun4v) 1023 if (nxgep->niu_type == N2_NIU && nxgep->niu_hsvc_available == B_TRUE) { 1024 (void) hsvc_unregister(&nxgep->niu_hsvc); 1025 nxgep->niu_hsvc_available = B_FALSE; 1026 } 1027 #endif 1028 /* 1029 * Stop any further interrupts. 1030 */ 1031 nxge_remove_intrs(nxgep); 1032 1033 /* 1034 * Stop the device and free resources. 1035 */ 1036 if (!isLDOMguest(nxgep)) { 1037 nxge_destroy_dev(nxgep); 1038 } 1039 1040 /* 1041 * Tear down the ndd parameters setup. 1042 */ 1043 nxge_destroy_param(nxgep); 1044 1045 /* 1046 * Tear down the kstat setup. 1047 */ 1048 nxge_destroy_kstats(nxgep); 1049 1050 /* 1051 * Free any memory allocated for PHY properties 1052 */ 1053 if (nxgep->phy_prop.cnt > 0) { 1054 KMEM_FREE(nxgep->phy_prop.arr, 1055 sizeof (nxge_phy_mdio_val_t) * nxgep->phy_prop.cnt); 1056 nxgep->phy_prop.cnt = 0; 1057 } 1058 1059 /* 1060 * Destroy all mutexes. 1061 */ 1062 nxge_destroy_mutexes(nxgep); 1063 1064 /* 1065 * Remove the list of ndd parameters which 1066 * were setup during attach. 1067 */ 1068 if (nxgep->dip) { 1069 NXGE_DEBUG_MSG((nxgep, OBP_CTL, 1070 " nxge_unattach: remove all properties")); 1071 1072 (void) ddi_prop_remove_all(nxgep->dip); 1073 } 1074 1075 #if NXGE_PROPERTY 1076 nxge_remove_hard_properties(nxgep); 1077 #endif 1078 1079 /* 1080 * Unmap the register setup. 1081 */ 1082 nxge_unmap_regs(nxgep); 1083 1084 nxge_fm_fini(nxgep); 1085 1086 ddi_soft_state_free(nxge_list, nxgep->instance); 1087 1088 NXGE_DEBUG_MSG((NULL, DDI_CTL, "<== nxge_unattach")); 1089 } 1090 1091 #if defined(sun4v) 1092 int 1093 nxge_hsvc_register(nxge_t *nxgep) 1094 { 1095 nxge_status_t status; 1096 int i, j; 1097 1098 NXGE_DEBUG_MSG((nxgep, DDI_CTL, "==> nxge_hsvc_register")); 1099 if (nxgep->niu_type != N2_NIU) { 1100 NXGE_DEBUG_MSG((nxgep, DDI_CTL, "<== nxge_hsvc_register")); 1101 return (DDI_SUCCESS); 1102 } 1103 1104 /* 1105 * Currently, the NIU Hypervisor API supports two major versions: 1106 * version 1 and 2. 1107 * If Hypervisor introduces a higher major or minor version, 1108 * please update NIU_MAJOR_HI and NIU_MINOR_HI accordingly. 1109 */ 1110 nxgep->niu_hsvc_available = B_FALSE; 1111 bcopy(&niu_hsvc, &nxgep->niu_hsvc, 1112 sizeof (hsvc_info_t)); 1113 1114 for (i = NIU_MAJOR_HI; i > 0; i--) { 1115 nxgep->niu_hsvc.hsvc_major = i; 1116 for (j = NIU_MINOR_HI; j >= 0; j--) { 1117 nxgep->niu_hsvc.hsvc_minor = j; 1118 NXGE_DEBUG_MSG((nxgep, DDI_CTL, 1119 "nxge_hsvc_register: %s: negotiating " 1120 "hypervisor services revision %d " 1121 "group: 0x%lx major: 0x%lx " 1122 "minor: 0x%lx", 1123 nxgep->niu_hsvc.hsvc_modname, 1124 nxgep->niu_hsvc.hsvc_rev, 1125 nxgep->niu_hsvc.hsvc_group, 1126 nxgep->niu_hsvc.hsvc_major, 1127 nxgep->niu_hsvc.hsvc_minor, 1128 nxgep->niu_min_ver)); 1129 1130 if ((status = hsvc_register(&nxgep->niu_hsvc, 1131 &nxgep->niu_min_ver)) == 0) { 1132 /* Use the supported minor */ 1133 nxgep->niu_hsvc.hsvc_minor = nxgep->niu_min_ver; 1134 NXGE_DEBUG_MSG((nxgep, DDI_CTL, 1135 "nxge_hsvc_register: %s: negotiated " 1136 "hypervisor services revision %d " 1137 "group: 0x%lx major: 0x%lx " 1138 "minor: 0x%lx (niu_min_ver 0x%lx)", 1139 nxgep->niu_hsvc.hsvc_modname, 1140 nxgep->niu_hsvc.hsvc_rev, 1141 nxgep->niu_hsvc.hsvc_group, 1142 nxgep->niu_hsvc.hsvc_major, 1143 nxgep->niu_hsvc.hsvc_minor, 1144 nxgep->niu_min_ver)); 1145 1146 nxgep->niu_hsvc_available = B_TRUE; 1147 NXGE_DEBUG_MSG((nxgep, DDI_CTL, 1148 "<== nxge_hsvc_register: " 1149 "NIU Hypervisor service enabled")); 1150 return (DDI_SUCCESS); 1151 } 1152 1153 NXGE_DEBUG_MSG((nxgep, DDI_CTL, 1154 "nxge_hsvc_register: %s: negotiated failed - " 1155 "try lower major number " 1156 "hypervisor services revision %d " 1157 "group: 0x%lx major: 0x%lx minor: 0x%lx " 1158 "errno: %d", 1159 nxgep->niu_hsvc.hsvc_modname, 1160 nxgep->niu_hsvc.hsvc_rev, 1161 nxgep->niu_hsvc.hsvc_group, 1162 nxgep->niu_hsvc.hsvc_major, 1163 nxgep->niu_hsvc.hsvc_minor, status)); 1164 } 1165 } 1166 1167 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, 1168 "nxge_hsvc_register: %s: cannot negotiate " 1169 "hypervisor services revision %d group: 0x%lx " 1170 "major: 0x%lx minor: 0x%lx errno: %d", 1171 niu_hsvc.hsvc_modname, niu_hsvc.hsvc_rev, 1172 niu_hsvc.hsvc_group, niu_hsvc.hsvc_major, 1173 niu_hsvc.hsvc_minor, status)); 1174 1175 NXGE_DEBUG_MSG((nxgep, DDI_CTL, 1176 "<== nxge_hsvc_register: Register to NIU Hypervisor failed")); 1177 1178 return (DDI_FAILURE); 1179 } 1180 #endif 1181 1182 static char n2_siu_name[] = "niu"; 1183 1184 static nxge_status_t 1185 nxge_map_regs(p_nxge_t nxgep) 1186 { 1187 int ddi_status = DDI_SUCCESS; 1188 p_dev_regs_t dev_regs; 1189 char buf[MAXPATHLEN + 1]; 1190 char *devname; 1191 #ifdef NXGE_DEBUG 1192 char *sysname; 1193 #endif 1194 off_t regsize; 1195 nxge_status_t status = NXGE_OK; 1196 #if !defined(_BIG_ENDIAN) 1197 off_t pci_offset; 1198 uint16_t pcie_devctl; 1199 #endif 1200 1201 if (isLDOMguest(nxgep)) { 1202 return (nxge_guest_regs_map(nxgep)); 1203 } 1204 1205 NXGE_DEBUG_MSG((nxgep, DDI_CTL, "==> nxge_map_regs")); 1206 nxgep->dev_regs = NULL; 1207 dev_regs = KMEM_ZALLOC(sizeof (dev_regs_t), KM_SLEEP); 1208 dev_regs->nxge_regh = NULL; 1209 dev_regs->nxge_pciregh = NULL; 1210 dev_regs->nxge_msix_regh = NULL; 1211 dev_regs->nxge_vir_regh = NULL; 1212 dev_regs->nxge_vir2_regh = NULL; 1213 nxgep->niu_type = NIU_TYPE_NONE; 1214 1215 devname = ddi_pathname(nxgep->dip, buf); 1216 ASSERT(strlen(devname) > 0); 1217 NXGE_DEBUG_MSG((nxgep, DDI_CTL, 1218 "nxge_map_regs: pathname devname %s", devname)); 1219 1220 /* 1221 * The driver is running on a N2-NIU system if devname is something 1222 * like "/niu@80/network@0" 1223 */ 1224 if (strstr(devname, n2_siu_name)) { 1225 /* N2/NIU */ 1226 nxgep->niu_type = N2_NIU; 1227 NXGE_DEBUG_MSG((nxgep, DDI_CTL, 1228 "nxge_map_regs: N2/NIU devname %s", devname)); 1229 /* 1230 * Get function number: 1231 * - N2/NIU: "/niu@80/network@0" and "/niu@80/network@1" 1232 */ 1233 nxgep->function_num = 1234 (devname[strlen(devname) -1] == '1' ? 1 : 0); 1235 NXGE_DEBUG_MSG((nxgep, DDI_CTL, 1236 "nxge_map_regs: N2/NIU function number %d", 1237 nxgep->function_num)); 1238 } else { 1239 int *prop_val; 1240 uint_t prop_len; 1241 uint8_t func_num; 1242 1243 if (ddi_prop_lookup_int_array(DDI_DEV_T_ANY, nxgep->dip, 1244 0, "reg", 1245 &prop_val, &prop_len) != DDI_PROP_SUCCESS) { 1246 NXGE_DEBUG_MSG((nxgep, VPD_CTL, 1247 "Reg property not found")); 1248 ddi_status = DDI_FAILURE; 1249 goto nxge_map_regs_fail0; 1250 1251 } else { 1252 func_num = (prop_val[0] >> 8) & 0x7; 1253 NXGE_DEBUG_MSG((nxgep, DDI_CTL, 1254 "Reg property found: fun # %d", 1255 func_num)); 1256 nxgep->function_num = func_num; 1257 if (isLDOMguest(nxgep)) { 1258 nxgep->function_num /= 2; 1259 return (NXGE_OK); 1260 } 1261 ddi_prop_free(prop_val); 1262 } 1263 } 1264 1265 switch (nxgep->niu_type) { 1266 default: 1267 (void) ddi_dev_regsize(nxgep->dip, 0, ®size); 1268 NXGE_DEBUG_MSG((nxgep, DDI_CTL, 1269 "nxge_map_regs: pci config size 0x%x", regsize)); 1270 1271 ddi_status = ddi_regs_map_setup(nxgep->dip, 0, 1272 (caddr_t *)&(dev_regs->nxge_pciregp), 0, 0, 1273 &nxge_dev_reg_acc_attr, &dev_regs->nxge_pciregh); 1274 if (ddi_status != DDI_SUCCESS) { 1275 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, 1276 "ddi_map_regs, nxge bus config regs failed")); 1277 goto nxge_map_regs_fail0; 1278 } 1279 NXGE_DEBUG_MSG((nxgep, DDI_CTL, 1280 "nxge_map_reg: PCI config addr 0x%0llx " 1281 " handle 0x%0llx", dev_regs->nxge_pciregp, 1282 dev_regs->nxge_pciregh)); 1283 /* 1284 * IMP IMP 1285 * workaround for bit swapping bug in HW 1286 * which ends up in no-snoop = yes 1287 * resulting, in DMA not synched properly 1288 */ 1289 #if !defined(_BIG_ENDIAN) 1290 /* workarounds for x86 systems */ 1291 pci_offset = 0x80 + PCIE_DEVCTL; 1292 pcie_devctl = pci_config_get16(dev_regs->nxge_pciregh, 1293 pci_offset); 1294 pcie_devctl &= ~PCIE_DEVCTL_ENABLE_NO_SNOOP; 1295 pcie_devctl |= PCIE_DEVCTL_RO_EN; 1296 pci_config_put16(dev_regs->nxge_pciregh, pci_offset, 1297 pcie_devctl); 1298 #endif 1299 1300 (void) ddi_dev_regsize(nxgep->dip, 1, ®size); 1301 NXGE_DEBUG_MSG((nxgep, DDI_CTL, 1302 "nxge_map_regs: pio size 0x%x", regsize)); 1303 /* set up the device mapped register */ 1304 ddi_status = ddi_regs_map_setup(nxgep->dip, 1, 1305 (caddr_t *)&(dev_regs->nxge_regp), 0, 0, 1306 &nxge_dev_reg_acc_attr, &dev_regs->nxge_regh); 1307 if (ddi_status != DDI_SUCCESS) { 1308 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, 1309 "ddi_map_regs for Neptune global reg failed")); 1310 goto nxge_map_regs_fail1; 1311 } 1312 1313 /* set up the msi/msi-x mapped register */ 1314 (void) ddi_dev_regsize(nxgep->dip, 2, ®size); 1315 NXGE_DEBUG_MSG((nxgep, DDI_CTL, 1316 "nxge_map_regs: msix size 0x%x", regsize)); 1317 ddi_status = ddi_regs_map_setup(nxgep->dip, 2, 1318 (caddr_t *)&(dev_regs->nxge_msix_regp), 0, 0, 1319 &nxge_dev_reg_acc_attr, &dev_regs->nxge_msix_regh); 1320 if (ddi_status != DDI_SUCCESS) { 1321 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, 1322 "ddi_map_regs for msi reg failed")); 1323 goto nxge_map_regs_fail2; 1324 } 1325 1326 /* set up the vio region mapped register */ 1327 (void) ddi_dev_regsize(nxgep->dip, 3, ®size); 1328 NXGE_DEBUG_MSG((nxgep, DDI_CTL, 1329 "nxge_map_regs: vio size 0x%x", regsize)); 1330 ddi_status = ddi_regs_map_setup(nxgep->dip, 3, 1331 (caddr_t *)&(dev_regs->nxge_vir_regp), 0, 0, 1332 &nxge_dev_reg_acc_attr, &dev_regs->nxge_vir_regh); 1333 1334 if (ddi_status != DDI_SUCCESS) { 1335 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, 1336 "ddi_map_regs for nxge vio reg failed")); 1337 goto nxge_map_regs_fail3; 1338 } 1339 nxgep->dev_regs = dev_regs; 1340 1341 NPI_PCI_ACC_HANDLE_SET(nxgep, dev_regs->nxge_pciregh); 1342 NPI_PCI_ADD_HANDLE_SET(nxgep, 1343 (npi_reg_ptr_t)dev_regs->nxge_pciregp); 1344 NPI_MSI_ACC_HANDLE_SET(nxgep, dev_regs->nxge_msix_regh); 1345 NPI_MSI_ADD_HANDLE_SET(nxgep, 1346 (npi_reg_ptr_t)dev_regs->nxge_msix_regp); 1347 1348 NPI_ACC_HANDLE_SET(nxgep, dev_regs->nxge_regh); 1349 NPI_ADD_HANDLE_SET(nxgep, (npi_reg_ptr_t)dev_regs->nxge_regp); 1350 1351 NPI_REG_ACC_HANDLE_SET(nxgep, dev_regs->nxge_regh); 1352 NPI_REG_ADD_HANDLE_SET(nxgep, 1353 (npi_reg_ptr_t)dev_regs->nxge_regp); 1354 1355 NPI_VREG_ACC_HANDLE_SET(nxgep, dev_regs->nxge_vir_regh); 1356 NPI_VREG_ADD_HANDLE_SET(nxgep, 1357 (npi_reg_ptr_t)dev_regs->nxge_vir_regp); 1358 1359 break; 1360 1361 case N2_NIU: 1362 NXGE_DEBUG_MSG((nxgep, DDI_CTL, "ddi_map_regs, NIU")); 1363 /* 1364 * Set up the device mapped register (FWARC 2006/556) 1365 * (changed back to 1: reg starts at 1!) 1366 */ 1367 (void) ddi_dev_regsize(nxgep->dip, 1, ®size); 1368 NXGE_DEBUG_MSG((nxgep, DDI_CTL, 1369 "nxge_map_regs: dev size 0x%x", regsize)); 1370 ddi_status = ddi_regs_map_setup(nxgep->dip, 1, 1371 (caddr_t *)&(dev_regs->nxge_regp), 0, 0, 1372 &nxge_dev_reg_acc_attr, &dev_regs->nxge_regh); 1373 1374 if (ddi_status != DDI_SUCCESS) { 1375 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, 1376 "ddi_map_regs for N2/NIU, global reg failed ")); 1377 goto nxge_map_regs_fail1; 1378 } 1379 1380 /* set up the first vio region mapped register */ 1381 (void) ddi_dev_regsize(nxgep->dip, 2, ®size); 1382 NXGE_DEBUG_MSG((nxgep, DDI_CTL, 1383 "nxge_map_regs: vio (1) size 0x%x", regsize)); 1384 ddi_status = ddi_regs_map_setup(nxgep->dip, 2, 1385 (caddr_t *)&(dev_regs->nxge_vir_regp), 0, 0, 1386 &nxge_dev_reg_acc_attr, &dev_regs->nxge_vir_regh); 1387 1388 if (ddi_status != DDI_SUCCESS) { 1389 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, 1390 "ddi_map_regs for nxge vio reg failed")); 1391 goto nxge_map_regs_fail2; 1392 } 1393 /* set up the second vio region mapped register */ 1394 (void) ddi_dev_regsize(nxgep->dip, 3, ®size); 1395 NXGE_DEBUG_MSG((nxgep, DDI_CTL, 1396 "nxge_map_regs: vio (3) size 0x%x", regsize)); 1397 ddi_status = ddi_regs_map_setup(nxgep->dip, 3, 1398 (caddr_t *)&(dev_regs->nxge_vir2_regp), 0, 0, 1399 &nxge_dev_reg_acc_attr, &dev_regs->nxge_vir2_regh); 1400 1401 if (ddi_status != DDI_SUCCESS) { 1402 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, 1403 "ddi_map_regs for nxge vio2 reg failed")); 1404 goto nxge_map_regs_fail3; 1405 } 1406 nxgep->dev_regs = dev_regs; 1407 1408 NPI_ACC_HANDLE_SET(nxgep, dev_regs->nxge_regh); 1409 NPI_ADD_HANDLE_SET(nxgep, (npi_reg_ptr_t)dev_regs->nxge_regp); 1410 1411 NPI_REG_ACC_HANDLE_SET(nxgep, dev_regs->nxge_regh); 1412 NPI_REG_ADD_HANDLE_SET(nxgep, 1413 (npi_reg_ptr_t)dev_regs->nxge_regp); 1414 1415 NPI_VREG_ACC_HANDLE_SET(nxgep, dev_regs->nxge_vir_regh); 1416 NPI_VREG_ADD_HANDLE_SET(nxgep, 1417 (npi_reg_ptr_t)dev_regs->nxge_vir_regp); 1418 1419 NPI_V2REG_ACC_HANDLE_SET(nxgep, dev_regs->nxge_vir2_regh); 1420 NPI_V2REG_ADD_HANDLE_SET(nxgep, 1421 (npi_reg_ptr_t)dev_regs->nxge_vir2_regp); 1422 1423 break; 1424 } 1425 1426 NXGE_DEBUG_MSG((nxgep, DDI_CTL, "nxge_map_reg: hardware addr 0x%0llx " 1427 " handle 0x%0llx", dev_regs->nxge_regp, dev_regs->nxge_regh)); 1428 1429 goto nxge_map_regs_exit; 1430 nxge_map_regs_fail3: 1431 if (dev_regs->nxge_msix_regh) { 1432 ddi_regs_map_free(&dev_regs->nxge_msix_regh); 1433 } 1434 if (dev_regs->nxge_vir_regh) { 1435 ddi_regs_map_free(&dev_regs->nxge_regh); 1436 } 1437 nxge_map_regs_fail2: 1438 if (dev_regs->nxge_regh) { 1439 ddi_regs_map_free(&dev_regs->nxge_regh); 1440 } 1441 nxge_map_regs_fail1: 1442 if (dev_regs->nxge_pciregh) { 1443 ddi_regs_map_free(&dev_regs->nxge_pciregh); 1444 } 1445 nxge_map_regs_fail0: 1446 NXGE_DEBUG_MSG((nxgep, DDI_CTL, "Freeing register set memory")); 1447 kmem_free(dev_regs, sizeof (dev_regs_t)); 1448 1449 nxge_map_regs_exit: 1450 if (ddi_status != DDI_SUCCESS) 1451 status |= (NXGE_ERROR | NXGE_DDI_FAILED); 1452 NXGE_DEBUG_MSG((nxgep, DDI_CTL, "<== nxge_map_regs")); 1453 return (status); 1454 } 1455 1456 static void 1457 nxge_unmap_regs(p_nxge_t nxgep) 1458 { 1459 NXGE_DEBUG_MSG((nxgep, DDI_CTL, "==> nxge_unmap_regs")); 1460 1461 if (isLDOMguest(nxgep)) { 1462 nxge_guest_regs_map_free(nxgep); 1463 return; 1464 } 1465 1466 if (nxgep->dev_regs) { 1467 if (nxgep->dev_regs->nxge_pciregh) { 1468 NXGE_DEBUG_MSG((nxgep, DDI_CTL, 1469 "==> nxge_unmap_regs: bus")); 1470 ddi_regs_map_free(&nxgep->dev_regs->nxge_pciregh); 1471 nxgep->dev_regs->nxge_pciregh = NULL; 1472 } 1473 if (nxgep->dev_regs->nxge_regh) { 1474 NXGE_DEBUG_MSG((nxgep, DDI_CTL, 1475 "==> nxge_unmap_regs: device registers")); 1476 ddi_regs_map_free(&nxgep->dev_regs->nxge_regh); 1477 nxgep->dev_regs->nxge_regh = NULL; 1478 } 1479 if (nxgep->dev_regs->nxge_msix_regh) { 1480 NXGE_DEBUG_MSG((nxgep, DDI_CTL, 1481 "==> nxge_unmap_regs: device interrupts")); 1482 ddi_regs_map_free(&nxgep->dev_regs->nxge_msix_regh); 1483 nxgep->dev_regs->nxge_msix_regh = NULL; 1484 } 1485 if (nxgep->dev_regs->nxge_vir_regh) { 1486 NXGE_DEBUG_MSG((nxgep, DDI_CTL, 1487 "==> nxge_unmap_regs: vio region")); 1488 ddi_regs_map_free(&nxgep->dev_regs->nxge_vir_regh); 1489 nxgep->dev_regs->nxge_vir_regh = NULL; 1490 } 1491 if (nxgep->dev_regs->nxge_vir2_regh) { 1492 NXGE_DEBUG_MSG((nxgep, DDI_CTL, 1493 "==> nxge_unmap_regs: vio2 region")); 1494 ddi_regs_map_free(&nxgep->dev_regs->nxge_vir2_regh); 1495 nxgep->dev_regs->nxge_vir2_regh = NULL; 1496 } 1497 1498 kmem_free(nxgep->dev_regs, sizeof (dev_regs_t)); 1499 nxgep->dev_regs = NULL; 1500 } 1501 1502 NXGE_DEBUG_MSG((nxgep, DDI_CTL, "<== nxge_unmap_regs")); 1503 } 1504 1505 static nxge_status_t 1506 nxge_setup_mutexes(p_nxge_t nxgep) 1507 { 1508 int ddi_status = DDI_SUCCESS; 1509 nxge_status_t status = NXGE_OK; 1510 nxge_classify_t *classify_ptr; 1511 int partition; 1512 1513 NXGE_DEBUG_MSG((nxgep, DDI_CTL, "==> nxge_setup_mutexes")); 1514 1515 /* 1516 * Get the interrupt cookie so the mutexes can be 1517 * Initialized. 1518 */ 1519 if (isLDOMguest(nxgep)) { 1520 nxgep->interrupt_cookie = 0; 1521 } else { 1522 ddi_status = ddi_get_iblock_cookie(nxgep->dip, 0, 1523 &nxgep->interrupt_cookie); 1524 1525 if (ddi_status != DDI_SUCCESS) { 1526 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, 1527 "<== nxge_setup_mutexes: failed 0x%x", 1528 ddi_status)); 1529 goto nxge_setup_mutexes_exit; 1530 } 1531 } 1532 1533 cv_init(&nxgep->poll_cv, NULL, CV_DRIVER, NULL); 1534 MUTEX_INIT(&nxgep->poll_lock, NULL, 1535 MUTEX_DRIVER, (void *)nxgep->interrupt_cookie); 1536 1537 /* 1538 * Initialize mutexes for this device. 1539 */ 1540 MUTEX_INIT(nxgep->genlock, NULL, 1541 MUTEX_DRIVER, (void *)nxgep->interrupt_cookie); 1542 MUTEX_INIT(&nxgep->ouraddr_lock, NULL, 1543 MUTEX_DRIVER, (void *)nxgep->interrupt_cookie); 1544 MUTEX_INIT(&nxgep->mif_lock, NULL, 1545 MUTEX_DRIVER, (void *)nxgep->interrupt_cookie); 1546 MUTEX_INIT(&nxgep->group_lock, NULL, 1547 MUTEX_DRIVER, (void *)nxgep->interrupt_cookie); 1548 RW_INIT(&nxgep->filter_lock, NULL, 1549 RW_DRIVER, (void *)nxgep->interrupt_cookie); 1550 1551 classify_ptr = &nxgep->classifier; 1552 /* 1553 * FFLP Mutexes are never used in interrupt context 1554 * as fflp operation can take very long time to 1555 * complete and hence not suitable to invoke from interrupt 1556 * handlers. 1557 */ 1558 MUTEX_INIT(&classify_ptr->tcam_lock, NULL, 1559 NXGE_MUTEX_DRIVER, (void *)nxgep->interrupt_cookie); 1560 if (NXGE_IS_VALID_NEPTUNE_TYPE(nxgep)) { 1561 MUTEX_INIT(&classify_ptr->fcram_lock, NULL, 1562 NXGE_MUTEX_DRIVER, (void *)nxgep->interrupt_cookie); 1563 for (partition = 0; partition < MAX_PARTITION; partition++) { 1564 MUTEX_INIT(&classify_ptr->hash_lock[partition], NULL, 1565 NXGE_MUTEX_DRIVER, (void *)nxgep->interrupt_cookie); 1566 } 1567 } 1568 1569 nxge_setup_mutexes_exit: 1570 NXGE_DEBUG_MSG((nxgep, DDI_CTL, 1571 "<== nxge_setup_mutexes status = %x", status)); 1572 1573 if (ddi_status != DDI_SUCCESS) 1574 status |= (NXGE_ERROR | NXGE_DDI_FAILED); 1575 1576 return (status); 1577 } 1578 1579 static void 1580 nxge_destroy_mutexes(p_nxge_t nxgep) 1581 { 1582 int partition; 1583 nxge_classify_t *classify_ptr; 1584 1585 NXGE_DEBUG_MSG((nxgep, DDI_CTL, "==> nxge_destroy_mutexes")); 1586 RW_DESTROY(&nxgep->filter_lock); 1587 MUTEX_DESTROY(&nxgep->group_lock); 1588 MUTEX_DESTROY(&nxgep->mif_lock); 1589 MUTEX_DESTROY(&nxgep->ouraddr_lock); 1590 MUTEX_DESTROY(nxgep->genlock); 1591 1592 classify_ptr = &nxgep->classifier; 1593 MUTEX_DESTROY(&classify_ptr->tcam_lock); 1594 1595 /* Destroy all polling resources. */ 1596 MUTEX_DESTROY(&nxgep->poll_lock); 1597 cv_destroy(&nxgep->poll_cv); 1598 1599 /* free data structures, based on HW type */ 1600 if (NXGE_IS_VALID_NEPTUNE_TYPE(nxgep)) { 1601 MUTEX_DESTROY(&classify_ptr->fcram_lock); 1602 for (partition = 0; partition < MAX_PARTITION; partition++) { 1603 MUTEX_DESTROY(&classify_ptr->hash_lock[partition]); 1604 } 1605 } 1606 1607 NXGE_DEBUG_MSG((nxgep, DDI_CTL, "<== nxge_destroy_mutexes")); 1608 } 1609 1610 nxge_status_t 1611 nxge_init(p_nxge_t nxgep) 1612 { 1613 nxge_status_t status = NXGE_OK; 1614 1615 NXGE_DEBUG_MSG((nxgep, STR_CTL, "==> nxge_init")); 1616 1617 if (nxgep->drv_state & STATE_HW_INITIALIZED) { 1618 return (status); 1619 } 1620 1621 /* 1622 * Allocate system memory for the receive/transmit buffer blocks 1623 * and receive/transmit descriptor rings. 1624 */ 1625 status = nxge_alloc_mem_pool(nxgep); 1626 if (status != NXGE_OK) { 1627 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, "alloc mem failed\n")); 1628 goto nxge_init_fail1; 1629 } 1630 1631 if (!isLDOMguest(nxgep)) { 1632 /* 1633 * Initialize and enable the TXC registers. 1634 * (Globally enable the Tx controller, 1635 * enable the port, configure the dma channel bitmap, 1636 * configure the max burst size). 1637 */ 1638 status = nxge_txc_init(nxgep); 1639 if (status != NXGE_OK) { 1640 NXGE_ERROR_MSG((nxgep, 1641 NXGE_ERR_CTL, "init txc failed\n")); 1642 goto nxge_init_fail2; 1643 } 1644 } 1645 1646 /* 1647 * Initialize and enable TXDMA channels. 1648 */ 1649 status = nxge_init_txdma_channels(nxgep); 1650 if (status != NXGE_OK) { 1651 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, "init txdma failed\n")); 1652 goto nxge_init_fail3; 1653 } 1654 1655 /* 1656 * Initialize and enable RXDMA channels. 1657 */ 1658 status = nxge_init_rxdma_channels(nxgep); 1659 if (status != NXGE_OK) { 1660 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, "init rxdma failed\n")); 1661 goto nxge_init_fail4; 1662 } 1663 1664 /* 1665 * The guest domain is now done. 1666 */ 1667 if (isLDOMguest(nxgep)) { 1668 nxgep->drv_state |= STATE_HW_INITIALIZED; 1669 goto nxge_init_exit; 1670 } 1671 1672 /* 1673 * Initialize TCAM and FCRAM (Neptune). 1674 */ 1675 status = nxge_classify_init(nxgep); 1676 if (status != NXGE_OK) { 1677 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, "init classify failed\n")); 1678 goto nxge_init_fail5; 1679 } 1680 1681 /* 1682 * Initialize ZCP 1683 */ 1684 status = nxge_zcp_init(nxgep); 1685 if (status != NXGE_OK) { 1686 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, "init ZCP failed\n")); 1687 goto nxge_init_fail5; 1688 } 1689 1690 /* 1691 * Initialize IPP. 1692 */ 1693 status = nxge_ipp_init(nxgep); 1694 if (status != NXGE_OK) { 1695 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, "init IPP failed\n")); 1696 goto nxge_init_fail5; 1697 } 1698 1699 /* 1700 * Initialize the MAC block. 1701 */ 1702 status = nxge_mac_init(nxgep); 1703 if (status != NXGE_OK) { 1704 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, "init MAC failed\n")); 1705 goto nxge_init_fail5; 1706 } 1707 1708 /* 1709 * Enable the interrrupts for DDI. 1710 */ 1711 nxge_intrs_enable(nxgep); 1712 1713 nxgep->drv_state |= STATE_HW_INITIALIZED; 1714 1715 goto nxge_init_exit; 1716 1717 nxge_init_fail5: 1718 nxge_uninit_rxdma_channels(nxgep); 1719 nxge_init_fail4: 1720 nxge_uninit_txdma_channels(nxgep); 1721 nxge_init_fail3: 1722 if (!isLDOMguest(nxgep)) { 1723 (void) nxge_txc_uninit(nxgep); 1724 } 1725 nxge_init_fail2: 1726 nxge_free_mem_pool(nxgep); 1727 nxge_init_fail1: 1728 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, 1729 "<== nxge_init status (failed) = 0x%08x", status)); 1730 return (status); 1731 1732 nxge_init_exit: 1733 NXGE_DEBUG_MSG((nxgep, DDI_CTL, "<== nxge_init status = 0x%08x", 1734 status)); 1735 return (status); 1736 } 1737 1738 1739 timeout_id_t 1740 nxge_start_timer(p_nxge_t nxgep, fptrv_t func, int msec) 1741 { 1742 if ((nxgep->suspended == 0) || (nxgep->suspended == DDI_RESUME)) { 1743 return (timeout(func, (caddr_t)nxgep, 1744 drv_usectohz(1000 * msec))); 1745 } 1746 return (NULL); 1747 } 1748 1749 /*ARGSUSED*/ 1750 void 1751 nxge_stop_timer(p_nxge_t nxgep, timeout_id_t timerid) 1752 { 1753 if (timerid) { 1754 (void) untimeout(timerid); 1755 } 1756 } 1757 1758 void 1759 nxge_uninit(p_nxge_t nxgep) 1760 { 1761 NXGE_DEBUG_MSG((nxgep, DDI_CTL, "==> nxge_uninit")); 1762 1763 if (!(nxgep->drv_state & STATE_HW_INITIALIZED)) { 1764 NXGE_DEBUG_MSG((nxgep, DDI_CTL, 1765 "==> nxge_uninit: not initialized")); 1766 NXGE_DEBUG_MSG((nxgep, DDI_CTL, 1767 "<== nxge_uninit")); 1768 return; 1769 } 1770 1771 if (!isLDOMguest(nxgep)) { 1772 /* 1773 * Reset the receive MAC side. 1774 */ 1775 (void) nxge_rx_mac_disable(nxgep); 1776 1777 /* 1778 * Drain the IPP. 1779 */ 1780 (void) nxge_ipp_drain(nxgep); 1781 } 1782 1783 /* stop timer */ 1784 if (nxgep->nxge_timerid) { 1785 nxge_stop_timer(nxgep, nxgep->nxge_timerid); 1786 nxgep->nxge_timerid = 0; 1787 } 1788 1789 (void) nxge_link_monitor(nxgep, LINK_MONITOR_STOP); 1790 (void) nxge_intr_hw_disable(nxgep); 1791 1792 1793 /* Disable and soft reset the IPP */ 1794 if (!isLDOMguest(nxgep)) 1795 (void) nxge_ipp_disable(nxgep); 1796 1797 /* Free classification resources */ 1798 (void) nxge_classify_uninit(nxgep); 1799 1800 /* 1801 * Reset the transmit/receive DMA side. 1802 */ 1803 (void) nxge_txdma_hw_mode(nxgep, NXGE_DMA_STOP); 1804 (void) nxge_rxdma_hw_mode(nxgep, NXGE_DMA_STOP); 1805 1806 nxge_uninit_txdma_channels(nxgep); 1807 nxge_uninit_rxdma_channels(nxgep); 1808 1809 /* 1810 * Reset the transmit MAC side. 1811 */ 1812 (void) nxge_tx_mac_disable(nxgep); 1813 1814 nxge_free_mem_pool(nxgep); 1815 1816 /* 1817 * Start the timer if the reset flag is not set. 1818 * If this reset flag is set, the link monitor 1819 * will not be started in order to stop furthur bus 1820 * activities coming from this interface. 1821 * The driver will start the monitor function 1822 * if the interface was initialized again later. 1823 */ 1824 if (!nxge_peu_reset_enable) { 1825 (void) nxge_link_monitor(nxgep, LINK_MONITOR_START); 1826 } 1827 1828 nxgep->drv_state &= ~STATE_HW_INITIALIZED; 1829 1830 NXGE_DEBUG_MSG((nxgep, DDI_CTL, "<== nxge_uninit: " 1831 "nxge_mblks_pending %d", nxge_mblks_pending)); 1832 } 1833 1834 void 1835 nxge_get64(p_nxge_t nxgep, p_mblk_t mp) 1836 { 1837 uint64_t reg; 1838 uint64_t regdata; 1839 int i, retry; 1840 1841 bcopy((char *)mp->b_rptr, (char *)®, sizeof (uint64_t)); 1842 regdata = 0; 1843 retry = 1; 1844 1845 for (i = 0; i < retry; i++) { 1846 NXGE_REG_RD64(nxgep->npi_handle, reg, ®data); 1847 } 1848 bcopy((char *)®data, (char *)mp->b_rptr, sizeof (uint64_t)); 1849 } 1850 1851 void 1852 nxge_put64(p_nxge_t nxgep, p_mblk_t mp) 1853 { 1854 uint64_t reg; 1855 uint64_t buf[2]; 1856 1857 bcopy((char *)mp->b_rptr, (char *)&buf[0], 2 * sizeof (uint64_t)); 1858 reg = buf[0]; 1859 1860 NXGE_NPI_PIO_WRITE64(nxgep->npi_handle, reg, buf[1]); 1861 } 1862 1863 /*ARGSUSED*/ 1864 /*VARARGS*/ 1865 void 1866 nxge_debug_msg(p_nxge_t nxgep, uint64_t level, char *fmt, ...) 1867 { 1868 char msg_buffer[1048]; 1869 char prefix_buffer[32]; 1870 int instance; 1871 uint64_t debug_level; 1872 int cmn_level = CE_CONT; 1873 va_list ap; 1874 1875 if (nxgep && nxgep->nxge_debug_level != nxge_debug_level) { 1876 /* In case a developer has changed nxge_debug_level. */ 1877 if (nxgep->nxge_debug_level != nxge_debug_level) 1878 nxgep->nxge_debug_level = nxge_debug_level; 1879 } 1880 1881 debug_level = (nxgep == NULL) ? nxge_debug_level : 1882 nxgep->nxge_debug_level; 1883 1884 if ((level & debug_level) || 1885 (level == NXGE_NOTE) || 1886 (level == NXGE_ERR_CTL)) { 1887 /* do the msg processing */ 1888 MUTEX_ENTER(&nxgedebuglock); 1889 1890 if ((level & NXGE_NOTE)) { 1891 cmn_level = CE_NOTE; 1892 } 1893 1894 if (level & NXGE_ERR_CTL) { 1895 cmn_level = CE_WARN; 1896 } 1897 1898 va_start(ap, fmt); 1899 (void) vsprintf(msg_buffer, fmt, ap); 1900 va_end(ap); 1901 if (nxgep == NULL) { 1902 instance = -1; 1903 (void) sprintf(prefix_buffer, "%s :", "nxge"); 1904 } else { 1905 instance = nxgep->instance; 1906 (void) sprintf(prefix_buffer, 1907 "%s%d :", "nxge", instance); 1908 } 1909 1910 MUTEX_EXIT(&nxgedebuglock); 1911 cmn_err(cmn_level, "!%s %s\n", 1912 prefix_buffer, msg_buffer); 1913 1914 } 1915 } 1916 1917 char * 1918 nxge_dump_packet(char *addr, int size) 1919 { 1920 uchar_t *ap = (uchar_t *)addr; 1921 int i; 1922 static char etherbuf[1024]; 1923 char *cp = etherbuf; 1924 char digits[] = "0123456789abcdef"; 1925 1926 if (!size) 1927 size = 60; 1928 1929 if (size > MAX_DUMP_SZ) { 1930 /* Dump the leading bytes */ 1931 for (i = 0; i < MAX_DUMP_SZ/2; i++) { 1932 if (*ap > 0x0f) 1933 *cp++ = digits[*ap >> 4]; 1934 *cp++ = digits[*ap++ & 0xf]; 1935 *cp++ = ':'; 1936 } 1937 for (i = 0; i < 20; i++) 1938 *cp++ = '.'; 1939 /* Dump the last MAX_DUMP_SZ/2 bytes */ 1940 ap = (uchar_t *)(addr + (size - MAX_DUMP_SZ/2)); 1941 for (i = 0; i < MAX_DUMP_SZ/2; i++) { 1942 if (*ap > 0x0f) 1943 *cp++ = digits[*ap >> 4]; 1944 *cp++ = digits[*ap++ & 0xf]; 1945 *cp++ = ':'; 1946 } 1947 } else { 1948 for (i = 0; i < size; i++) { 1949 if (*ap > 0x0f) 1950 *cp++ = digits[*ap >> 4]; 1951 *cp++ = digits[*ap++ & 0xf]; 1952 *cp++ = ':'; 1953 } 1954 } 1955 *--cp = 0; 1956 return (etherbuf); 1957 } 1958 1959 #ifdef NXGE_DEBUG 1960 static void 1961 nxge_test_map_regs(p_nxge_t nxgep) 1962 { 1963 ddi_acc_handle_t cfg_handle; 1964 p_pci_cfg_t cfg_ptr; 1965 ddi_acc_handle_t dev_handle; 1966 char *dev_ptr; 1967 ddi_acc_handle_t pci_config_handle; 1968 uint32_t regval; 1969 int i; 1970 1971 NXGE_DEBUG_MSG((nxgep, DDI_CTL, "==> nxge_test_map_regs")); 1972 1973 dev_handle = nxgep->dev_regs->nxge_regh; 1974 dev_ptr = (char *)nxgep->dev_regs->nxge_regp; 1975 1976 if (NXGE_IS_VALID_NEPTUNE_TYPE(nxgep)) { 1977 cfg_handle = nxgep->dev_regs->nxge_pciregh; 1978 cfg_ptr = (void *)nxgep->dev_regs->nxge_pciregp; 1979 1980 NXGE_DEBUG_MSG((nxgep, DDI_CTL, 1981 "Neptune PCI regp cfg_ptr 0x%llx", (char *)cfg_ptr)); 1982 NXGE_DEBUG_MSG((nxgep, DDI_CTL, 1983 "Neptune PCI cfg_ptr vendor id ptr 0x%llx", 1984 &cfg_ptr->vendorid)); 1985 NXGE_DEBUG_MSG((nxgep, DDI_CTL, 1986 "\tvendorid 0x%x devid 0x%x", 1987 NXGE_PIO_READ16(cfg_handle, &cfg_ptr->vendorid, 0), 1988 NXGE_PIO_READ16(cfg_handle, &cfg_ptr->devid, 0))); 1989 NXGE_DEBUG_MSG((nxgep, DDI_CTL, 1990 "PCI BAR: base 0x%x base14 0x%x base 18 0x%x " 1991 "bar1c 0x%x", 1992 NXGE_PIO_READ32(cfg_handle, &cfg_ptr->base, 0), 1993 NXGE_PIO_READ32(cfg_handle, &cfg_ptr->base14, 0), 1994 NXGE_PIO_READ32(cfg_handle, &cfg_ptr->base18, 0), 1995 NXGE_PIO_READ32(cfg_handle, &cfg_ptr->base1c, 0))); 1996 NXGE_DEBUG_MSG((nxgep, DDI_CTL, 1997 "\nNeptune PCI BAR: base20 0x%x base24 0x%x " 1998 "base 28 0x%x bar2c 0x%x\n", 1999 NXGE_PIO_READ32(cfg_handle, &cfg_ptr->base20, 0), 2000 NXGE_PIO_READ32(cfg_handle, &cfg_ptr->base24, 0), 2001 NXGE_PIO_READ32(cfg_handle, &cfg_ptr->base28, 0), 2002 NXGE_PIO_READ32(cfg_handle, &cfg_ptr->base2c, 0))); 2003 NXGE_DEBUG_MSG((nxgep, DDI_CTL, 2004 "\nNeptune PCI BAR: base30 0x%x\n", 2005 NXGE_PIO_READ32(cfg_handle, &cfg_ptr->base30, 0))); 2006 2007 cfg_handle = nxgep->dev_regs->nxge_pciregh; 2008 cfg_ptr = (void *)nxgep->dev_regs->nxge_pciregp; 2009 NXGE_DEBUG_MSG((nxgep, DDI_CTL, 2010 "first 0x%llx second 0x%llx third 0x%llx " 2011 "last 0x%llx ", 2012 NXGE_PIO_READ64(dev_handle, 2013 (uint64_t *)(dev_ptr + 0), 0), 2014 NXGE_PIO_READ64(dev_handle, 2015 (uint64_t *)(dev_ptr + 8), 0), 2016 NXGE_PIO_READ64(dev_handle, 2017 (uint64_t *)(dev_ptr + 16), 0), 2018 NXGE_PIO_READ64(cfg_handle, 2019 (uint64_t *)(dev_ptr + 24), 0))); 2020 } 2021 } 2022 2023 #endif 2024 2025 static void 2026 nxge_suspend(p_nxge_t nxgep) 2027 { 2028 NXGE_DEBUG_MSG((nxgep, DDI_CTL, "==> nxge_suspend")); 2029 2030 nxge_intrs_disable(nxgep); 2031 nxge_destroy_dev(nxgep); 2032 2033 NXGE_DEBUG_MSG((nxgep, DDI_CTL, "<== nxge_suspend")); 2034 } 2035 2036 static nxge_status_t 2037 nxge_resume(p_nxge_t nxgep) 2038 { 2039 nxge_status_t status = NXGE_OK; 2040 2041 NXGE_DEBUG_MSG((nxgep, DDI_CTL, "==> nxge_resume")); 2042 2043 nxgep->suspended = DDI_RESUME; 2044 (void) nxge_link_monitor(nxgep, LINK_MONITOR_START); 2045 (void) nxge_rxdma_hw_mode(nxgep, NXGE_DMA_START); 2046 (void) nxge_txdma_hw_mode(nxgep, NXGE_DMA_START); 2047 (void) nxge_rx_mac_enable(nxgep); 2048 (void) nxge_tx_mac_enable(nxgep); 2049 nxge_intrs_enable(nxgep); 2050 nxgep->suspended = 0; 2051 2052 NXGE_DEBUG_MSG((nxgep, DDI_CTL, 2053 "<== nxge_resume status = 0x%x", status)); 2054 return (status); 2055 } 2056 2057 static nxge_status_t 2058 nxge_setup_dev(p_nxge_t nxgep) 2059 { 2060 nxge_status_t status = NXGE_OK; 2061 2062 NXGE_DEBUG_MSG((nxgep, DDI_CTL, "==> nxge_setup_dev port %d", 2063 nxgep->mac.portnum)); 2064 2065 status = nxge_link_init(nxgep); 2066 2067 if (fm_check_acc_handle(nxgep->dev_regs->nxge_regh) != DDI_FM_OK) { 2068 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, 2069 "port%d Bad register acc handle", nxgep->mac.portnum)); 2070 status = NXGE_ERROR; 2071 } 2072 2073 if (status != NXGE_OK) { 2074 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, 2075 " nxge_setup_dev status " 2076 "(xcvr init 0x%08x)", status)); 2077 goto nxge_setup_dev_exit; 2078 } 2079 2080 nxge_setup_dev_exit: 2081 NXGE_DEBUG_MSG((nxgep, DDI_CTL, 2082 "<== nxge_setup_dev port %d status = 0x%08x", 2083 nxgep->mac.portnum, status)); 2084 2085 return (status); 2086 } 2087 2088 static void 2089 nxge_destroy_dev(p_nxge_t nxgep) 2090 { 2091 NXGE_DEBUG_MSG((nxgep, DDI_CTL, "==> nxge_destroy_dev")); 2092 2093 (void) nxge_link_monitor(nxgep, LINK_MONITOR_STOP); 2094 2095 (void) nxge_hw_stop(nxgep); 2096 2097 NXGE_DEBUG_MSG((nxgep, DDI_CTL, "<== nxge_destroy_dev")); 2098 } 2099 2100 static nxge_status_t 2101 nxge_setup_system_dma_pages(p_nxge_t nxgep) 2102 { 2103 int ddi_status = DDI_SUCCESS; 2104 uint_t count; 2105 ddi_dma_cookie_t cookie; 2106 uint_t iommu_pagesize; 2107 nxge_status_t status = NXGE_OK; 2108 2109 NXGE_ERROR_MSG((nxgep, DDI_CTL, "==> nxge_setup_system_dma_pages")); 2110 nxgep->sys_page_sz = ddi_ptob(nxgep->dip, (ulong_t)1); 2111 if (nxgep->niu_type != N2_NIU) { 2112 iommu_pagesize = dvma_pagesize(nxgep->dip); 2113 NXGE_DEBUG_MSG((nxgep, DDI_CTL, 2114 " nxge_setup_system_dma_pages: page %d (ddi_ptob %d) " 2115 " default_block_size %d iommu_pagesize %d", 2116 nxgep->sys_page_sz, 2117 ddi_ptob(nxgep->dip, (ulong_t)1), 2118 nxgep->rx_default_block_size, 2119 iommu_pagesize)); 2120 2121 if (iommu_pagesize != 0) { 2122 if (nxgep->sys_page_sz == iommu_pagesize) { 2123 if (iommu_pagesize > 0x4000) 2124 nxgep->sys_page_sz = 0x4000; 2125 } else { 2126 if (nxgep->sys_page_sz > iommu_pagesize) 2127 nxgep->sys_page_sz = iommu_pagesize; 2128 } 2129 } 2130 } 2131 nxgep->sys_page_mask = ~(nxgep->sys_page_sz - 1); 2132 NXGE_DEBUG_MSG((nxgep, DDI_CTL, 2133 "==> nxge_setup_system_dma_pages: page %d (ddi_ptob %d) " 2134 "default_block_size %d page mask %d", 2135 nxgep->sys_page_sz, 2136 ddi_ptob(nxgep->dip, (ulong_t)1), 2137 nxgep->rx_default_block_size, 2138 nxgep->sys_page_mask)); 2139 2140 2141 switch (nxgep->sys_page_sz) { 2142 default: 2143 nxgep->sys_page_sz = 0x1000; 2144 nxgep->sys_page_mask = ~(nxgep->sys_page_sz - 1); 2145 nxgep->rx_default_block_size = 0x1000; 2146 nxgep->rx_bksize_code = RBR_BKSIZE_4K; 2147 break; 2148 case 0x1000: 2149 nxgep->rx_default_block_size = 0x1000; 2150 nxgep->rx_bksize_code = RBR_BKSIZE_4K; 2151 break; 2152 case 0x2000: 2153 nxgep->rx_default_block_size = 0x2000; 2154 nxgep->rx_bksize_code = RBR_BKSIZE_8K; 2155 break; 2156 case 0x4000: 2157 nxgep->rx_default_block_size = 0x4000; 2158 nxgep->rx_bksize_code = RBR_BKSIZE_16K; 2159 break; 2160 case 0x8000: 2161 nxgep->rx_default_block_size = 0x8000; 2162 nxgep->rx_bksize_code = RBR_BKSIZE_32K; 2163 break; 2164 } 2165 2166 #ifndef USE_RX_BIG_BUF 2167 nxge_rx_dma_attr.dma_attr_align = nxgep->sys_page_sz; 2168 #else 2169 nxgep->rx_default_block_size = 0x2000; 2170 nxgep->rx_bksize_code = RBR_BKSIZE_8K; 2171 #endif 2172 /* 2173 * Get the system DMA burst size. 2174 */ 2175 ddi_status = ddi_dma_alloc_handle(nxgep->dip, &nxge_tx_dma_attr, 2176 DDI_DMA_DONTWAIT, 0, 2177 &nxgep->dmasparehandle); 2178 if (ddi_status != DDI_SUCCESS) { 2179 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, 2180 "ddi_dma_alloc_handle: failed " 2181 " status 0x%x", ddi_status)); 2182 goto nxge_get_soft_properties_exit; 2183 } 2184 2185 ddi_status = ddi_dma_addr_bind_handle(nxgep->dmasparehandle, NULL, 2186 (caddr_t)nxgep->dmasparehandle, 2187 sizeof (nxgep->dmasparehandle), 2188 DDI_DMA_RDWR | DDI_DMA_CONSISTENT, 2189 DDI_DMA_DONTWAIT, 0, 2190 &cookie, &count); 2191 if (ddi_status != DDI_DMA_MAPPED) { 2192 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, 2193 "Binding spare handle to find system" 2194 " burstsize failed.")); 2195 ddi_status = DDI_FAILURE; 2196 goto nxge_get_soft_properties_fail1; 2197 } 2198 2199 nxgep->sys_burst_sz = ddi_dma_burstsizes(nxgep->dmasparehandle); 2200 (void) ddi_dma_unbind_handle(nxgep->dmasparehandle); 2201 2202 nxge_get_soft_properties_fail1: 2203 ddi_dma_free_handle(&nxgep->dmasparehandle); 2204 2205 nxge_get_soft_properties_exit: 2206 2207 if (ddi_status != DDI_SUCCESS) 2208 status |= (NXGE_ERROR | NXGE_DDI_FAILED); 2209 2210 NXGE_DEBUG_MSG((nxgep, DDI_CTL, 2211 "<== nxge_setup_system_dma_pages status = 0x%08x", status)); 2212 return (status); 2213 } 2214 2215 static nxge_status_t 2216 nxge_alloc_mem_pool(p_nxge_t nxgep) 2217 { 2218 nxge_status_t status = NXGE_OK; 2219 2220 NXGE_DEBUG_MSG((nxgep, DDI_CTL, "==> nxge_alloc_mem_pool")); 2221 2222 status = nxge_alloc_rx_mem_pool(nxgep); 2223 if (status != NXGE_OK) { 2224 return (NXGE_ERROR); 2225 } 2226 2227 status = nxge_alloc_tx_mem_pool(nxgep); 2228 if (status != NXGE_OK) { 2229 nxge_free_rx_mem_pool(nxgep); 2230 return (NXGE_ERROR); 2231 } 2232 2233 NXGE_DEBUG_MSG((nxgep, DDI_CTL, "<== nxge_alloc_mem_pool")); 2234 return (NXGE_OK); 2235 } 2236 2237 static void 2238 nxge_free_mem_pool(p_nxge_t nxgep) 2239 { 2240 NXGE_DEBUG_MSG((nxgep, MEM_CTL, "==> nxge_free_mem_pool")); 2241 2242 nxge_free_rx_mem_pool(nxgep); 2243 nxge_free_tx_mem_pool(nxgep); 2244 2245 NXGE_DEBUG_MSG((nxgep, MEM_CTL, "<== nxge_free_mem_pool")); 2246 } 2247 2248 nxge_status_t 2249 nxge_alloc_rx_mem_pool(p_nxge_t nxgep) 2250 { 2251 uint32_t rdc_max; 2252 p_nxge_dma_pt_cfg_t p_all_cfgp; 2253 p_nxge_hw_pt_cfg_t p_cfgp; 2254 p_nxge_dma_pool_t dma_poolp; 2255 p_nxge_dma_common_t *dma_buf_p; 2256 p_nxge_dma_pool_t dma_cntl_poolp; 2257 p_nxge_dma_common_t *dma_cntl_p; 2258 uint32_t *num_chunks; /* per dma */ 2259 nxge_status_t status = NXGE_OK; 2260 2261 uint32_t nxge_port_rbr_size; 2262 uint32_t nxge_port_rbr_spare_size; 2263 uint32_t nxge_port_rcr_size; 2264 uint32_t rx_cntl_alloc_size; 2265 2266 NXGE_DEBUG_MSG((nxgep, DMA_CTL, "==> nxge_alloc_rx_mem_pool")); 2267 2268 p_all_cfgp = (p_nxge_dma_pt_cfg_t)&nxgep->pt_config; 2269 p_cfgp = (p_nxge_hw_pt_cfg_t)&p_all_cfgp->hw_config; 2270 rdc_max = NXGE_MAX_RDCS; 2271 2272 /* 2273 * Allocate memory for the common DMA data structures. 2274 */ 2275 dma_poolp = (p_nxge_dma_pool_t)KMEM_ZALLOC(sizeof (nxge_dma_pool_t), 2276 KM_SLEEP); 2277 dma_buf_p = (p_nxge_dma_common_t *)KMEM_ZALLOC( 2278 sizeof (p_nxge_dma_common_t) * rdc_max, KM_SLEEP); 2279 2280 dma_cntl_poolp = (p_nxge_dma_pool_t) 2281 KMEM_ZALLOC(sizeof (nxge_dma_pool_t), KM_SLEEP); 2282 dma_cntl_p = (p_nxge_dma_common_t *)KMEM_ZALLOC( 2283 sizeof (p_nxge_dma_common_t) * rdc_max, KM_SLEEP); 2284 2285 num_chunks = (uint32_t *)KMEM_ZALLOC( 2286 sizeof (uint32_t) * rdc_max, KM_SLEEP); 2287 2288 /* 2289 * Assume that each DMA channel will be configured with 2290 * the default block size. 2291 * rbr block counts are modulo the batch count (16). 2292 */ 2293 nxge_port_rbr_size = p_all_cfgp->rbr_size; 2294 nxge_port_rcr_size = p_all_cfgp->rcr_size; 2295 2296 if (!nxge_port_rbr_size) { 2297 nxge_port_rbr_size = NXGE_RBR_RBB_DEFAULT; 2298 } 2299 if (nxge_port_rbr_size % NXGE_RXDMA_POST_BATCH) { 2300 nxge_port_rbr_size = (NXGE_RXDMA_POST_BATCH * 2301 (nxge_port_rbr_size / NXGE_RXDMA_POST_BATCH + 1)); 2302 } 2303 2304 p_all_cfgp->rbr_size = nxge_port_rbr_size; 2305 nxge_port_rbr_spare_size = nxge_rbr_spare_size; 2306 2307 if (nxge_port_rbr_spare_size % NXGE_RXDMA_POST_BATCH) { 2308 nxge_port_rbr_spare_size = (NXGE_RXDMA_POST_BATCH * 2309 (nxge_port_rbr_spare_size / NXGE_RXDMA_POST_BATCH + 1)); 2310 } 2311 if (nxge_port_rbr_size > RBR_DEFAULT_MAX_BLKS) { 2312 NXGE_DEBUG_MSG((nxgep, MEM_CTL, 2313 "nxge_alloc_rx_mem_pool: RBR size too high %d, " 2314 "set to default %d", 2315 nxge_port_rbr_size, RBR_DEFAULT_MAX_BLKS)); 2316 nxge_port_rbr_size = RBR_DEFAULT_MAX_BLKS; 2317 } 2318 if (nxge_port_rcr_size > RCR_DEFAULT_MAX) { 2319 NXGE_DEBUG_MSG((nxgep, MEM_CTL, 2320 "nxge_alloc_rx_mem_pool: RCR too high %d, " 2321 "set to default %d", 2322 nxge_port_rcr_size, RCR_DEFAULT_MAX)); 2323 nxge_port_rcr_size = RCR_DEFAULT_MAX; 2324 } 2325 2326 /* 2327 * N2/NIU has limitation on the descriptor sizes (contiguous 2328 * memory allocation on data buffers to 4M (contig_mem_alloc) 2329 * and little endian for control buffers (must use the ddi/dki mem alloc 2330 * function). 2331 */ 2332 #if defined(sun4v) && defined(NIU_LP_WORKAROUND) 2333 if (nxgep->niu_type == N2_NIU) { 2334 nxge_port_rbr_spare_size = 0; 2335 if ((nxge_port_rbr_size > NXGE_NIU_CONTIG_RBR_MAX) || 2336 (!ISP2(nxge_port_rbr_size))) { 2337 nxge_port_rbr_size = NXGE_NIU_CONTIG_RBR_MAX; 2338 } 2339 if ((nxge_port_rcr_size > NXGE_NIU_CONTIG_RCR_MAX) || 2340 (!ISP2(nxge_port_rcr_size))) { 2341 nxge_port_rcr_size = NXGE_NIU_CONTIG_RCR_MAX; 2342 } 2343 } 2344 #endif 2345 2346 /* 2347 * Addresses of receive block ring, receive completion ring and the 2348 * mailbox must be all cache-aligned (64 bytes). 2349 */ 2350 rx_cntl_alloc_size = nxge_port_rbr_size + nxge_port_rbr_spare_size; 2351 rx_cntl_alloc_size *= (sizeof (rx_desc_t)); 2352 rx_cntl_alloc_size += (sizeof (rcr_entry_t) * nxge_port_rcr_size); 2353 rx_cntl_alloc_size += sizeof (rxdma_mailbox_t); 2354 2355 NXGE_DEBUG_MSG((nxgep, MEM2_CTL, "==> nxge_alloc_rx_mem_pool: " 2356 "nxge_port_rbr_size = %d nxge_port_rbr_spare_size = %d " 2357 "nxge_port_rcr_size = %d " 2358 "rx_cntl_alloc_size = %d", 2359 nxge_port_rbr_size, nxge_port_rbr_spare_size, 2360 nxge_port_rcr_size, 2361 rx_cntl_alloc_size)); 2362 2363 #if defined(sun4v) && defined(NIU_LP_WORKAROUND) 2364 if (nxgep->niu_type == N2_NIU) { 2365 uint32_t rx_buf_alloc_size = (nxgep->rx_default_block_size * 2366 (nxge_port_rbr_size + nxge_port_rbr_spare_size)); 2367 2368 if (!ISP2(rx_buf_alloc_size)) { 2369 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, 2370 "==> nxge_alloc_rx_mem_pool: " 2371 " must be power of 2")); 2372 status |= (NXGE_ERROR | NXGE_DDI_FAILED); 2373 goto nxge_alloc_rx_mem_pool_exit; 2374 } 2375 2376 if (rx_buf_alloc_size > (1 << 22)) { 2377 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, 2378 "==> nxge_alloc_rx_mem_pool: " 2379 " limit size to 4M")); 2380 status |= (NXGE_ERROR | NXGE_DDI_FAILED); 2381 goto nxge_alloc_rx_mem_pool_exit; 2382 } 2383 2384 if (rx_cntl_alloc_size < 0x2000) { 2385 rx_cntl_alloc_size = 0x2000; 2386 } 2387 } 2388 #endif 2389 nxgep->nxge_port_rbr_size = nxge_port_rbr_size; 2390 nxgep->nxge_port_rcr_size = nxge_port_rcr_size; 2391 nxgep->nxge_port_rbr_spare_size = nxge_port_rbr_spare_size; 2392 nxgep->nxge_port_rx_cntl_alloc_size = rx_cntl_alloc_size; 2393 2394 dma_poolp->ndmas = p_cfgp->max_rdcs; 2395 dma_poolp->num_chunks = num_chunks; 2396 dma_poolp->buf_allocated = B_TRUE; 2397 nxgep->rx_buf_pool_p = dma_poolp; 2398 dma_poolp->dma_buf_pool_p = dma_buf_p; 2399 2400 dma_cntl_poolp->ndmas = p_cfgp->max_rdcs; 2401 dma_cntl_poolp->buf_allocated = B_TRUE; 2402 nxgep->rx_cntl_pool_p = dma_cntl_poolp; 2403 dma_cntl_poolp->dma_buf_pool_p = dma_cntl_p; 2404 2405 /* Allocate the receive rings, too. */ 2406 nxgep->rx_rbr_rings = 2407 KMEM_ZALLOC(sizeof (rx_rbr_rings_t), KM_SLEEP); 2408 nxgep->rx_rbr_rings->rbr_rings = 2409 KMEM_ZALLOC(sizeof (p_rx_rbr_ring_t) * rdc_max, KM_SLEEP); 2410 nxgep->rx_rcr_rings = 2411 KMEM_ZALLOC(sizeof (rx_rcr_rings_t), KM_SLEEP); 2412 nxgep->rx_rcr_rings->rcr_rings = 2413 KMEM_ZALLOC(sizeof (p_rx_rcr_ring_t) * rdc_max, KM_SLEEP); 2414 nxgep->rx_mbox_areas_p = 2415 KMEM_ZALLOC(sizeof (rx_mbox_areas_t), KM_SLEEP); 2416 nxgep->rx_mbox_areas_p->rxmbox_areas = 2417 KMEM_ZALLOC(sizeof (p_rx_mbox_t) * rdc_max, KM_SLEEP); 2418 2419 nxgep->rx_rbr_rings->ndmas = nxgep->rx_rcr_rings->ndmas = 2420 p_cfgp->max_rdcs; 2421 2422 NXGE_DEBUG_MSG((nxgep, DMA_CTL, 2423 "<== nxge_alloc_rx_mem_pool:status 0x%08x", status)); 2424 2425 nxge_alloc_rx_mem_pool_exit: 2426 return (status); 2427 } 2428 2429 /* 2430 * nxge_alloc_rxb 2431 * 2432 * Allocate buffers for an RDC. 2433 * 2434 * Arguments: 2435 * nxgep 2436 * channel The channel to map into our kernel space. 2437 * 2438 * Notes: 2439 * 2440 * NPI function calls: 2441 * 2442 * NXGE function calls: 2443 * 2444 * Registers accessed: 2445 * 2446 * Context: 2447 * 2448 * Taking apart: 2449 * 2450 * Open questions: 2451 * 2452 */ 2453 nxge_status_t 2454 nxge_alloc_rxb( 2455 p_nxge_t nxgep, 2456 int channel) 2457 { 2458 size_t rx_buf_alloc_size; 2459 nxge_status_t status = NXGE_OK; 2460 2461 nxge_dma_common_t **data; 2462 nxge_dma_common_t **control; 2463 uint32_t *num_chunks; 2464 2465 NXGE_DEBUG_MSG((nxgep, DMA_CTL, "==> nxge_alloc_rbb")); 2466 2467 /* 2468 * Allocate memory for the receive buffers and descriptor rings. 2469 * Replace these allocation functions with the interface functions 2470 * provided by the partition manager if/when they are available. 2471 */ 2472 2473 /* 2474 * Allocate memory for the receive buffer blocks. 2475 */ 2476 rx_buf_alloc_size = (nxgep->rx_default_block_size * 2477 (nxgep->nxge_port_rbr_size + nxgep->nxge_port_rbr_spare_size)); 2478 2479 data = &nxgep->rx_buf_pool_p->dma_buf_pool_p[channel]; 2480 num_chunks = &nxgep->rx_buf_pool_p->num_chunks[channel]; 2481 2482 if ((status = nxge_alloc_rx_buf_dma( 2483 nxgep, channel, data, rx_buf_alloc_size, 2484 nxgep->rx_default_block_size, num_chunks)) != NXGE_OK) { 2485 return (status); 2486 } 2487 2488 NXGE_DEBUG_MSG((nxgep, MEM2_CTL, "<== nxge_alloc_rxb(): " 2489 "dma %d dma_buf_p %llx &dma_buf_p %llx", channel, *data, data)); 2490 2491 /* 2492 * Allocate memory for descriptor rings and mailbox. 2493 */ 2494 control = &nxgep->rx_cntl_pool_p->dma_buf_pool_p[channel]; 2495 2496 if ((status = nxge_alloc_rx_cntl_dma( 2497 nxgep, channel, control, nxgep->nxge_port_rx_cntl_alloc_size)) 2498 != NXGE_OK) { 2499 nxge_free_rx_cntl_dma(nxgep, *control); 2500 (*data)->buf_alloc_state |= BUF_ALLOCATED_WAIT_FREE; 2501 nxge_free_rx_buf_dma(nxgep, *data, *num_chunks); 2502 return (status); 2503 } 2504 2505 NXGE_DEBUG_MSG((nxgep, DMA_CTL, 2506 "<== nxge_alloc_rx_mem_pool:status 0x%08x", status)); 2507 2508 return (status); 2509 } 2510 2511 void 2512 nxge_free_rxb( 2513 p_nxge_t nxgep, 2514 int channel) 2515 { 2516 nxge_dma_common_t *data; 2517 nxge_dma_common_t *control; 2518 uint32_t num_chunks; 2519 2520 NXGE_DEBUG_MSG((nxgep, DMA_CTL, "==> nxge_alloc_rbb")); 2521 2522 data = nxgep->rx_buf_pool_p->dma_buf_pool_p[channel]; 2523 num_chunks = nxgep->rx_buf_pool_p->num_chunks[channel]; 2524 nxge_free_rx_buf_dma(nxgep, data, num_chunks); 2525 2526 nxgep->rx_buf_pool_p->dma_buf_pool_p[channel] = 0; 2527 nxgep->rx_buf_pool_p->num_chunks[channel] = 0; 2528 2529 control = nxgep->rx_cntl_pool_p->dma_buf_pool_p[channel]; 2530 nxge_free_rx_cntl_dma(nxgep, control); 2531 2532 nxgep->rx_cntl_pool_p->dma_buf_pool_p[channel] = 0; 2533 2534 KMEM_FREE(data, sizeof (nxge_dma_common_t) * NXGE_DMA_BLOCK); 2535 KMEM_FREE(control, sizeof (nxge_dma_common_t)); 2536 2537 NXGE_DEBUG_MSG((nxgep, DMA_CTL, "<== nxge_alloc_rbb")); 2538 } 2539 2540 static void 2541 nxge_free_rx_mem_pool(p_nxge_t nxgep) 2542 { 2543 int rdc_max = NXGE_MAX_RDCS; 2544 2545 NXGE_DEBUG_MSG((nxgep, MEM2_CTL, "==> nxge_free_rx_mem_pool")); 2546 2547 if (!nxgep->rx_buf_pool_p || !nxgep->rx_buf_pool_p->buf_allocated) { 2548 NXGE_DEBUG_MSG((nxgep, MEM2_CTL, 2549 "<== nxge_free_rx_mem_pool " 2550 "(null rx buf pool or buf not allocated")); 2551 return; 2552 } 2553 if (!nxgep->rx_cntl_pool_p || !nxgep->rx_cntl_pool_p->buf_allocated) { 2554 NXGE_DEBUG_MSG((nxgep, MEM2_CTL, 2555 "<== nxge_free_rx_mem_pool " 2556 "(null rx cntl buf pool or cntl buf not allocated")); 2557 return; 2558 } 2559 2560 KMEM_FREE(nxgep->rx_cntl_pool_p->dma_buf_pool_p, 2561 sizeof (p_nxge_dma_common_t) * rdc_max); 2562 KMEM_FREE(nxgep->rx_cntl_pool_p, sizeof (nxge_dma_pool_t)); 2563 2564 KMEM_FREE(nxgep->rx_buf_pool_p->num_chunks, 2565 sizeof (uint32_t) * rdc_max); 2566 KMEM_FREE(nxgep->rx_buf_pool_p->dma_buf_pool_p, 2567 sizeof (p_nxge_dma_common_t) * rdc_max); 2568 KMEM_FREE(nxgep->rx_buf_pool_p, sizeof (nxge_dma_pool_t)); 2569 2570 nxgep->rx_buf_pool_p = 0; 2571 nxgep->rx_cntl_pool_p = 0; 2572 2573 KMEM_FREE(nxgep->rx_rbr_rings->rbr_rings, 2574 sizeof (p_rx_rbr_ring_t) * rdc_max); 2575 KMEM_FREE(nxgep->rx_rbr_rings, sizeof (rx_rbr_rings_t)); 2576 KMEM_FREE(nxgep->rx_rcr_rings->rcr_rings, 2577 sizeof (p_rx_rcr_ring_t) * rdc_max); 2578 KMEM_FREE(nxgep->rx_rcr_rings, sizeof (rx_rcr_rings_t)); 2579 KMEM_FREE(nxgep->rx_mbox_areas_p->rxmbox_areas, 2580 sizeof (p_rx_mbox_t) * rdc_max); 2581 KMEM_FREE(nxgep->rx_mbox_areas_p, sizeof (rx_mbox_areas_t)); 2582 2583 nxgep->rx_rbr_rings = 0; 2584 nxgep->rx_rcr_rings = 0; 2585 nxgep->rx_mbox_areas_p = 0; 2586 2587 NXGE_DEBUG_MSG((nxgep, MEM2_CTL, "<== nxge_free_rx_mem_pool")); 2588 } 2589 2590 2591 static nxge_status_t 2592 nxge_alloc_rx_buf_dma(p_nxge_t nxgep, uint16_t dma_channel, 2593 p_nxge_dma_common_t *dmap, 2594 size_t alloc_size, size_t block_size, uint32_t *num_chunks) 2595 { 2596 p_nxge_dma_common_t rx_dmap; 2597 nxge_status_t status = NXGE_OK; 2598 size_t total_alloc_size; 2599 size_t allocated = 0; 2600 int i, size_index, array_size; 2601 boolean_t use_kmem_alloc = B_FALSE; 2602 2603 NXGE_DEBUG_MSG((nxgep, DMA_CTL, "==> nxge_alloc_rx_buf_dma")); 2604 2605 rx_dmap = (p_nxge_dma_common_t) 2606 KMEM_ZALLOC(sizeof (nxge_dma_common_t) * NXGE_DMA_BLOCK, 2607 KM_SLEEP); 2608 2609 NXGE_DEBUG_MSG((nxgep, MEM2_CTL, 2610 " alloc_rx_buf_dma rdc %d asize %x bsize %x bbuf %llx ", 2611 dma_channel, alloc_size, block_size, dmap)); 2612 2613 total_alloc_size = alloc_size; 2614 2615 #if defined(RX_USE_RECLAIM_POST) 2616 total_alloc_size = alloc_size + alloc_size/4; 2617 #endif 2618 2619 i = 0; 2620 size_index = 0; 2621 array_size = sizeof (alloc_sizes)/sizeof (size_t); 2622 while ((size_index < array_size) && 2623 (alloc_sizes[size_index] < alloc_size)) 2624 size_index++; 2625 if (size_index >= array_size) { 2626 size_index = array_size - 1; 2627 } 2628 2629 /* For Neptune, use kmem_alloc if the kmem flag is set. */ 2630 if (nxgep->niu_type != N2_NIU && nxge_use_kmem_alloc) { 2631 use_kmem_alloc = B_TRUE; 2632 #if defined(__i386) || defined(__amd64) 2633 size_index = 0; 2634 #endif 2635 NXGE_DEBUG_MSG((nxgep, MEM2_CTL, 2636 "==> nxge_alloc_rx_buf_dma: " 2637 "Neptune use kmem_alloc() - size_index %d", 2638 size_index)); 2639 } 2640 2641 while ((allocated < total_alloc_size) && 2642 (size_index >= 0) && (i < NXGE_DMA_BLOCK)) { 2643 rx_dmap[i].dma_chunk_index = i; 2644 rx_dmap[i].block_size = block_size; 2645 rx_dmap[i].alength = alloc_sizes[size_index]; 2646 rx_dmap[i].orig_alength = rx_dmap[i].alength; 2647 rx_dmap[i].nblocks = alloc_sizes[size_index] / block_size; 2648 rx_dmap[i].dma_channel = dma_channel; 2649 rx_dmap[i].contig_alloc_type = B_FALSE; 2650 rx_dmap[i].kmem_alloc_type = B_FALSE; 2651 rx_dmap[i].buf_alloc_type = DDI_MEM_ALLOC; 2652 2653 /* 2654 * N2/NIU: data buffers must be contiguous as the driver 2655 * needs to call Hypervisor api to set up 2656 * logical pages. 2657 */ 2658 if ((nxgep->niu_type == N2_NIU) && (NXGE_DMA_BLOCK == 1)) { 2659 rx_dmap[i].contig_alloc_type = B_TRUE; 2660 rx_dmap[i].buf_alloc_type = CONTIG_MEM_ALLOC; 2661 } else if (use_kmem_alloc) { 2662 /* For Neptune, use kmem_alloc */ 2663 NXGE_DEBUG_MSG((nxgep, MEM2_CTL, 2664 "==> nxge_alloc_rx_buf_dma: " 2665 "Neptune use kmem_alloc()")); 2666 rx_dmap[i].kmem_alloc_type = B_TRUE; 2667 rx_dmap[i].buf_alloc_type = KMEM_ALLOC; 2668 } 2669 2670 NXGE_DEBUG_MSG((nxgep, MEM2_CTL, 2671 "alloc_rx_buf_dma rdc %d chunk %d bufp %llx size %x " 2672 "i %d nblocks %d alength %d", 2673 dma_channel, i, &rx_dmap[i], block_size, 2674 i, rx_dmap[i].nblocks, 2675 rx_dmap[i].alength)); 2676 status = nxge_dma_mem_alloc(nxgep, nxge_force_dma, 2677 &nxge_rx_dma_attr, 2678 rx_dmap[i].alength, 2679 &nxge_dev_buf_dma_acc_attr, 2680 DDI_DMA_READ | DDI_DMA_STREAMING, 2681 (p_nxge_dma_common_t)(&rx_dmap[i])); 2682 if (status != NXGE_OK) { 2683 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, 2684 "nxge_alloc_rx_buf_dma: Alloc Failed: " 2685 "dma %d size_index %d size requested %d", 2686 dma_channel, 2687 size_index, 2688 rx_dmap[i].alength)); 2689 size_index--; 2690 } else { 2691 rx_dmap[i].buf_alloc_state = BUF_ALLOCATED; 2692 NXGE_DEBUG_MSG((nxgep, MEM2_CTL, 2693 " nxge_alloc_rx_buf_dma DONE alloc mem: " 2694 "dma %d dma_buf_p $%p kaddrp $%p alength %d " 2695 "buf_alloc_state %d alloc_type %d", 2696 dma_channel, 2697 &rx_dmap[i], 2698 rx_dmap[i].kaddrp, 2699 rx_dmap[i].alength, 2700 rx_dmap[i].buf_alloc_state, 2701 rx_dmap[i].buf_alloc_type)); 2702 NXGE_DEBUG_MSG((nxgep, MEM2_CTL, 2703 " alloc_rx_buf_dma allocated rdc %d " 2704 "chunk %d size %x dvma %x bufp %llx kaddrp $%p", 2705 dma_channel, i, rx_dmap[i].alength, 2706 rx_dmap[i].ioaddr_pp, &rx_dmap[i], 2707 rx_dmap[i].kaddrp)); 2708 i++; 2709 allocated += alloc_sizes[size_index]; 2710 } 2711 } 2712 2713 if (allocated < total_alloc_size) { 2714 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, 2715 "==> nxge_alloc_rx_buf_dma: not enough for channel %d " 2716 "allocated 0x%x requested 0x%x", 2717 dma_channel, 2718 allocated, total_alloc_size)); 2719 status = NXGE_ERROR; 2720 goto nxge_alloc_rx_mem_fail1; 2721 } 2722 2723 NXGE_DEBUG_MSG((nxgep, MEM2_CTL, 2724 "==> nxge_alloc_rx_buf_dma: Allocated for channel %d " 2725 "allocated 0x%x requested 0x%x", 2726 dma_channel, 2727 allocated, total_alloc_size)); 2728 2729 NXGE_DEBUG_MSG((nxgep, DMA_CTL, 2730 " alloc_rx_buf_dma rdc %d allocated %d chunks", 2731 dma_channel, i)); 2732 *num_chunks = i; 2733 *dmap = rx_dmap; 2734 2735 goto nxge_alloc_rx_mem_exit; 2736 2737 nxge_alloc_rx_mem_fail1: 2738 KMEM_FREE(rx_dmap, sizeof (nxge_dma_common_t) * NXGE_DMA_BLOCK); 2739 2740 nxge_alloc_rx_mem_exit: 2741 NXGE_DEBUG_MSG((nxgep, DMA_CTL, 2742 "<== nxge_alloc_rx_buf_dma status 0x%08x", status)); 2743 2744 return (status); 2745 } 2746 2747 /*ARGSUSED*/ 2748 static void 2749 nxge_free_rx_buf_dma(p_nxge_t nxgep, p_nxge_dma_common_t dmap, 2750 uint32_t num_chunks) 2751 { 2752 int i; 2753 2754 NXGE_DEBUG_MSG((nxgep, MEM2_CTL, 2755 "==> nxge_free_rx_buf_dma: # of chunks %d", num_chunks)); 2756 2757 if (dmap == 0) 2758 return; 2759 2760 for (i = 0; i < num_chunks; i++) { 2761 NXGE_DEBUG_MSG((nxgep, MEM2_CTL, 2762 "==> nxge_free_rx_buf_dma: chunk %d dmap 0x%llx", 2763 i, dmap)); 2764 nxge_dma_free_rx_data_buf(dmap++); 2765 } 2766 2767 NXGE_DEBUG_MSG((nxgep, MEM2_CTL, "==> nxge_free_rx_buf_dma")); 2768 } 2769 2770 /*ARGSUSED*/ 2771 static nxge_status_t 2772 nxge_alloc_rx_cntl_dma(p_nxge_t nxgep, uint16_t dma_channel, 2773 p_nxge_dma_common_t *dmap, size_t size) 2774 { 2775 p_nxge_dma_common_t rx_dmap; 2776 nxge_status_t status = NXGE_OK; 2777 2778 NXGE_DEBUG_MSG((nxgep, DMA_CTL, "==> nxge_alloc_rx_cntl_dma")); 2779 2780 rx_dmap = (p_nxge_dma_common_t) 2781 KMEM_ZALLOC(sizeof (nxge_dma_common_t), KM_SLEEP); 2782 2783 rx_dmap->contig_alloc_type = B_FALSE; 2784 rx_dmap->kmem_alloc_type = B_FALSE; 2785 2786 status = nxge_dma_mem_alloc(nxgep, nxge_force_dma, 2787 &nxge_desc_dma_attr, 2788 size, 2789 &nxge_dev_desc_dma_acc_attr, 2790 DDI_DMA_RDWR | DDI_DMA_CONSISTENT, 2791 rx_dmap); 2792 if (status != NXGE_OK) { 2793 goto nxge_alloc_rx_cntl_dma_fail1; 2794 } 2795 2796 *dmap = rx_dmap; 2797 goto nxge_alloc_rx_cntl_dma_exit; 2798 2799 nxge_alloc_rx_cntl_dma_fail1: 2800 KMEM_FREE(rx_dmap, sizeof (nxge_dma_common_t)); 2801 2802 nxge_alloc_rx_cntl_dma_exit: 2803 NXGE_DEBUG_MSG((nxgep, DMA_CTL, 2804 "<== nxge_alloc_rx_cntl_dma status 0x%08x", status)); 2805 2806 return (status); 2807 } 2808 2809 /*ARGSUSED*/ 2810 static void 2811 nxge_free_rx_cntl_dma(p_nxge_t nxgep, p_nxge_dma_common_t dmap) 2812 { 2813 NXGE_DEBUG_MSG((nxgep, DMA_CTL, "==> nxge_free_rx_cntl_dma")); 2814 2815 if (dmap == 0) 2816 return; 2817 2818 nxge_dma_mem_free(dmap); 2819 2820 NXGE_DEBUG_MSG((nxgep, DMA_CTL, "<== nxge_free_rx_cntl_dma")); 2821 } 2822 2823 typedef struct { 2824 size_t tx_size; 2825 size_t cr_size; 2826 size_t threshhold; 2827 } nxge_tdc_sizes_t; 2828 2829 static 2830 nxge_status_t 2831 nxge_tdc_sizes( 2832 nxge_t *nxgep, 2833 nxge_tdc_sizes_t *sizes) 2834 { 2835 uint32_t threshhold; /* The bcopy() threshhold */ 2836 size_t tx_size; /* Transmit buffer size */ 2837 size_t cr_size; /* Completion ring size */ 2838 2839 /* 2840 * Assume that each DMA channel will be configured with the 2841 * default transmit buffer size for copying transmit data. 2842 * (If a packet is bigger than this, it will not be copied.) 2843 */ 2844 if (nxgep->niu_type == N2_NIU) { 2845 threshhold = TX_BCOPY_SIZE; 2846 } else { 2847 threshhold = nxge_bcopy_thresh; 2848 } 2849 tx_size = nxge_tx_ring_size * threshhold; 2850 2851 cr_size = nxge_tx_ring_size * sizeof (tx_desc_t); 2852 cr_size += sizeof (txdma_mailbox_t); 2853 2854 #if defined(sun4v) && defined(NIU_LP_WORKAROUND) 2855 if (nxgep->niu_type == N2_NIU) { 2856 if (!ISP2(tx_size)) { 2857 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, 2858 "==> nxge_tdc_sizes: Tx size" 2859 " must be power of 2")); 2860 return (NXGE_ERROR); 2861 } 2862 2863 if (tx_size > (1 << 22)) { 2864 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, 2865 "==> nxge_tdc_sizes: Tx size" 2866 " limited to 4M")); 2867 return (NXGE_ERROR); 2868 } 2869 2870 if (cr_size < 0x2000) 2871 cr_size = 0x2000; 2872 } 2873 #endif 2874 2875 sizes->threshhold = threshhold; 2876 sizes->tx_size = tx_size; 2877 sizes->cr_size = cr_size; 2878 2879 return (NXGE_OK); 2880 } 2881 /* 2882 * nxge_alloc_txb 2883 * 2884 * Allocate buffers for an TDC. 2885 * 2886 * Arguments: 2887 * nxgep 2888 * channel The channel to map into our kernel space. 2889 * 2890 * Notes: 2891 * 2892 * NPI function calls: 2893 * 2894 * NXGE function calls: 2895 * 2896 * Registers accessed: 2897 * 2898 * Context: 2899 * 2900 * Taking apart: 2901 * 2902 * Open questions: 2903 * 2904 */ 2905 nxge_status_t 2906 nxge_alloc_txb( 2907 p_nxge_t nxgep, 2908 int channel) 2909 { 2910 nxge_dma_common_t **dma_buf_p; 2911 nxge_dma_common_t **dma_cntl_p; 2912 uint32_t *num_chunks; 2913 nxge_status_t status = NXGE_OK; 2914 2915 nxge_tdc_sizes_t sizes; 2916 2917 NXGE_DEBUG_MSG((nxgep, DMA_CTL, "==> nxge_alloc_tbb")); 2918 2919 if (nxge_tdc_sizes(nxgep, &sizes) != NXGE_OK) 2920 return (NXGE_ERROR); 2921 2922 /* 2923 * Allocate memory for transmit buffers and descriptor rings. 2924 * Replace these allocation functions with the interface functions 2925 * provided by the partition manager Real Soon Now. 2926 */ 2927 dma_buf_p = &nxgep->tx_buf_pool_p->dma_buf_pool_p[channel]; 2928 num_chunks = &nxgep->tx_buf_pool_p->num_chunks[channel]; 2929 2930 dma_cntl_p = &nxgep->tx_cntl_pool_p->dma_buf_pool_p[channel]; 2931 2932 /* 2933 * Allocate memory for transmit buffers and descriptor rings. 2934 * Replace allocation functions with interface functions provided 2935 * by the partition manager when it is available. 2936 * 2937 * Allocate memory for the transmit buffer pool. 2938 */ 2939 NXGE_DEBUG_MSG((nxgep, DMA_CTL, 2940 "sizes: tx: %ld, cr:%ld, th:%ld", 2941 sizes.tx_size, sizes.cr_size, sizes.threshhold)); 2942 2943 *num_chunks = 0; 2944 status = nxge_alloc_tx_buf_dma(nxgep, channel, dma_buf_p, 2945 sizes.tx_size, sizes.threshhold, num_chunks); 2946 if (status != NXGE_OK) { 2947 cmn_err(CE_NOTE, "nxge_alloc_tx_buf_dma failed!"); 2948 return (status); 2949 } 2950 2951 /* 2952 * Allocate memory for descriptor rings and mailbox. 2953 */ 2954 status = nxge_alloc_tx_cntl_dma(nxgep, channel, dma_cntl_p, 2955 sizes.cr_size); 2956 if (status != NXGE_OK) { 2957 nxge_free_tx_buf_dma(nxgep, *dma_buf_p, *num_chunks); 2958 cmn_err(CE_NOTE, "nxge_alloc_tx_cntl_dma failed!"); 2959 return (status); 2960 } 2961 2962 return (NXGE_OK); 2963 } 2964 2965 void 2966 nxge_free_txb( 2967 p_nxge_t nxgep, 2968 int channel) 2969 { 2970 nxge_dma_common_t *data; 2971 nxge_dma_common_t *control; 2972 uint32_t num_chunks; 2973 2974 NXGE_DEBUG_MSG((nxgep, DMA_CTL, "==> nxge_free_txb")); 2975 2976 data = nxgep->tx_buf_pool_p->dma_buf_pool_p[channel]; 2977 num_chunks = nxgep->tx_buf_pool_p->num_chunks[channel]; 2978 nxge_free_tx_buf_dma(nxgep, data, num_chunks); 2979 2980 nxgep->tx_buf_pool_p->dma_buf_pool_p[channel] = 0; 2981 nxgep->tx_buf_pool_p->num_chunks[channel] = 0; 2982 2983 control = nxgep->tx_cntl_pool_p->dma_buf_pool_p[channel]; 2984 nxge_free_tx_cntl_dma(nxgep, control); 2985 2986 nxgep->tx_cntl_pool_p->dma_buf_pool_p[channel] = 0; 2987 2988 KMEM_FREE(data, sizeof (nxge_dma_common_t) * NXGE_DMA_BLOCK); 2989 KMEM_FREE(control, sizeof (nxge_dma_common_t)); 2990 2991 NXGE_DEBUG_MSG((nxgep, DMA_CTL, "<== nxge_free_txb")); 2992 } 2993 2994 /* 2995 * nxge_alloc_tx_mem_pool 2996 * 2997 * This function allocates all of the per-port TDC control data structures. 2998 * The per-channel (TDC) data structures are allocated when needed. 2999 * 3000 * Arguments: 3001 * nxgep 3002 * 3003 * Notes: 3004 * 3005 * Context: 3006 * Any domain 3007 */ 3008 nxge_status_t 3009 nxge_alloc_tx_mem_pool(p_nxge_t nxgep) 3010 { 3011 nxge_hw_pt_cfg_t *p_cfgp; 3012 nxge_dma_pool_t *dma_poolp; 3013 nxge_dma_common_t **dma_buf_p; 3014 nxge_dma_pool_t *dma_cntl_poolp; 3015 nxge_dma_common_t **dma_cntl_p; 3016 uint32_t *num_chunks; /* per dma */ 3017 int tdc_max; 3018 3019 NXGE_DEBUG_MSG((nxgep, MEM_CTL, "==> nxge_alloc_tx_mem_pool")); 3020 3021 p_cfgp = &nxgep->pt_config.hw_config; 3022 tdc_max = NXGE_MAX_TDCS; 3023 3024 /* 3025 * Allocate memory for each transmit DMA channel. 3026 */ 3027 dma_poolp = (p_nxge_dma_pool_t)KMEM_ZALLOC(sizeof (nxge_dma_pool_t), 3028 KM_SLEEP); 3029 dma_buf_p = (p_nxge_dma_common_t *)KMEM_ZALLOC( 3030 sizeof (p_nxge_dma_common_t) * tdc_max, KM_SLEEP); 3031 3032 dma_cntl_poolp = (p_nxge_dma_pool_t) 3033 KMEM_ZALLOC(sizeof (nxge_dma_pool_t), KM_SLEEP); 3034 dma_cntl_p = (p_nxge_dma_common_t *)KMEM_ZALLOC( 3035 sizeof (p_nxge_dma_common_t) * tdc_max, KM_SLEEP); 3036 3037 if (nxge_tx_ring_size > TDC_DEFAULT_MAX) { 3038 NXGE_DEBUG_MSG((nxgep, MEM_CTL, 3039 "nxge_alloc_tx_mem_pool: TDC too high %d, " 3040 "set to default %d", 3041 nxge_tx_ring_size, TDC_DEFAULT_MAX)); 3042 nxge_tx_ring_size = TDC_DEFAULT_MAX; 3043 } 3044 3045 #if defined(sun4v) && defined(NIU_LP_WORKAROUND) 3046 /* 3047 * N2/NIU has limitation on the descriptor sizes (contiguous 3048 * memory allocation on data buffers to 4M (contig_mem_alloc) 3049 * and little endian for control buffers (must use the ddi/dki mem alloc 3050 * function). The transmit ring is limited to 8K (includes the 3051 * mailbox). 3052 */ 3053 if (nxgep->niu_type == N2_NIU) { 3054 if ((nxge_tx_ring_size > NXGE_NIU_CONTIG_TX_MAX) || 3055 (!ISP2(nxge_tx_ring_size))) { 3056 nxge_tx_ring_size = NXGE_NIU_CONTIG_TX_MAX; 3057 } 3058 } 3059 #endif 3060 3061 nxgep->nxge_port_tx_ring_size = nxge_tx_ring_size; 3062 3063 num_chunks = (uint32_t *)KMEM_ZALLOC( 3064 sizeof (uint32_t) * tdc_max, KM_SLEEP); 3065 3066 dma_poolp->ndmas = p_cfgp->tdc.owned; 3067 dma_poolp->num_chunks = num_chunks; 3068 dma_poolp->dma_buf_pool_p = dma_buf_p; 3069 nxgep->tx_buf_pool_p = dma_poolp; 3070 3071 dma_poolp->buf_allocated = B_TRUE; 3072 3073 dma_cntl_poolp->ndmas = p_cfgp->tdc.owned; 3074 dma_cntl_poolp->dma_buf_pool_p = dma_cntl_p; 3075 nxgep->tx_cntl_pool_p = dma_cntl_poolp; 3076 3077 dma_cntl_poolp->buf_allocated = B_TRUE; 3078 3079 nxgep->tx_rings = 3080 KMEM_ZALLOC(sizeof (tx_rings_t), KM_SLEEP); 3081 nxgep->tx_rings->rings = 3082 KMEM_ZALLOC(sizeof (p_tx_ring_t) * tdc_max, KM_SLEEP); 3083 nxgep->tx_mbox_areas_p = 3084 KMEM_ZALLOC(sizeof (tx_mbox_areas_t), KM_SLEEP); 3085 nxgep->tx_mbox_areas_p->txmbox_areas_p = 3086 KMEM_ZALLOC(sizeof (p_tx_mbox_t) * tdc_max, KM_SLEEP); 3087 3088 nxgep->tx_rings->ndmas = p_cfgp->tdc.owned; 3089 3090 NXGE_DEBUG_MSG((nxgep, MEM_CTL, 3091 "==> nxge_alloc_tx_mem_pool: ndmas %d poolp->ndmas %d", 3092 tdc_max, dma_poolp->ndmas)); 3093 3094 return (NXGE_OK); 3095 } 3096 3097 nxge_status_t 3098 nxge_alloc_tx_buf_dma(p_nxge_t nxgep, uint16_t dma_channel, 3099 p_nxge_dma_common_t *dmap, size_t alloc_size, 3100 size_t block_size, uint32_t *num_chunks) 3101 { 3102 p_nxge_dma_common_t tx_dmap; 3103 nxge_status_t status = NXGE_OK; 3104 size_t total_alloc_size; 3105 size_t allocated = 0; 3106 int i, size_index, array_size; 3107 3108 NXGE_DEBUG_MSG((nxgep, DMA_CTL, "==> nxge_alloc_tx_buf_dma")); 3109 3110 tx_dmap = (p_nxge_dma_common_t) 3111 KMEM_ZALLOC(sizeof (nxge_dma_common_t) * NXGE_DMA_BLOCK, 3112 KM_SLEEP); 3113 3114 total_alloc_size = alloc_size; 3115 i = 0; 3116 size_index = 0; 3117 array_size = sizeof (alloc_sizes) / sizeof (size_t); 3118 while ((size_index < array_size) && 3119 (alloc_sizes[size_index] < alloc_size)) 3120 size_index++; 3121 if (size_index >= array_size) { 3122 size_index = array_size - 1; 3123 } 3124 3125 while ((allocated < total_alloc_size) && 3126 (size_index >= 0) && (i < NXGE_DMA_BLOCK)) { 3127 3128 tx_dmap[i].dma_chunk_index = i; 3129 tx_dmap[i].block_size = block_size; 3130 tx_dmap[i].alength = alloc_sizes[size_index]; 3131 tx_dmap[i].orig_alength = tx_dmap[i].alength; 3132 tx_dmap[i].nblocks = alloc_sizes[size_index] / block_size; 3133 tx_dmap[i].dma_channel = dma_channel; 3134 tx_dmap[i].contig_alloc_type = B_FALSE; 3135 tx_dmap[i].kmem_alloc_type = B_FALSE; 3136 3137 /* 3138 * N2/NIU: data buffers must be contiguous as the driver 3139 * needs to call Hypervisor api to set up 3140 * logical pages. 3141 */ 3142 if ((nxgep->niu_type == N2_NIU) && (NXGE_DMA_BLOCK == 1)) { 3143 tx_dmap[i].contig_alloc_type = B_TRUE; 3144 } 3145 3146 status = nxge_dma_mem_alloc(nxgep, nxge_force_dma, 3147 &nxge_tx_dma_attr, 3148 tx_dmap[i].alength, 3149 &nxge_dev_buf_dma_acc_attr, 3150 DDI_DMA_WRITE | DDI_DMA_STREAMING, 3151 (p_nxge_dma_common_t)(&tx_dmap[i])); 3152 if (status != NXGE_OK) { 3153 size_index--; 3154 } else { 3155 i++; 3156 allocated += alloc_sizes[size_index]; 3157 } 3158 } 3159 3160 if (allocated < total_alloc_size) { 3161 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, 3162 "==> nxge_alloc_tx_buf_dma: not enough channel %d: " 3163 "allocated 0x%x requested 0x%x", 3164 dma_channel, 3165 allocated, total_alloc_size)); 3166 status = NXGE_ERROR; 3167 goto nxge_alloc_tx_mem_fail1; 3168 } 3169 3170 NXGE_DEBUG_MSG((nxgep, MEM2_CTL, 3171 "==> nxge_alloc_tx_buf_dma: Allocated for channel %d: " 3172 "allocated 0x%x requested 0x%x", 3173 dma_channel, 3174 allocated, total_alloc_size)); 3175 3176 *num_chunks = i; 3177 *dmap = tx_dmap; 3178 NXGE_DEBUG_MSG((nxgep, DMA_CTL, 3179 "==> nxge_alloc_tx_buf_dma dmap 0x%016llx num chunks %d", 3180 *dmap, i)); 3181 goto nxge_alloc_tx_mem_exit; 3182 3183 nxge_alloc_tx_mem_fail1: 3184 KMEM_FREE(tx_dmap, sizeof (nxge_dma_common_t) * NXGE_DMA_BLOCK); 3185 3186 nxge_alloc_tx_mem_exit: 3187 NXGE_DEBUG_MSG((nxgep, DMA_CTL, 3188 "<== nxge_alloc_tx_buf_dma status 0x%08x", status)); 3189 3190 return (status); 3191 } 3192 3193 /*ARGSUSED*/ 3194 static void 3195 nxge_free_tx_buf_dma(p_nxge_t nxgep, p_nxge_dma_common_t dmap, 3196 uint32_t num_chunks) 3197 { 3198 int i; 3199 3200 NXGE_DEBUG_MSG((nxgep, MEM_CTL, "==> nxge_free_tx_buf_dma")); 3201 3202 if (dmap == 0) 3203 return; 3204 3205 for (i = 0; i < num_chunks; i++) { 3206 nxge_dma_mem_free(dmap++); 3207 } 3208 3209 NXGE_DEBUG_MSG((nxgep, MEM_CTL, "<== nxge_free_tx_buf_dma")); 3210 } 3211 3212 /*ARGSUSED*/ 3213 nxge_status_t 3214 nxge_alloc_tx_cntl_dma(p_nxge_t nxgep, uint16_t dma_channel, 3215 p_nxge_dma_common_t *dmap, size_t size) 3216 { 3217 p_nxge_dma_common_t tx_dmap; 3218 nxge_status_t status = NXGE_OK; 3219 3220 NXGE_DEBUG_MSG((nxgep, DMA_CTL, "==> nxge_alloc_tx_cntl_dma")); 3221 tx_dmap = (p_nxge_dma_common_t) 3222 KMEM_ZALLOC(sizeof (nxge_dma_common_t), KM_SLEEP); 3223 3224 tx_dmap->contig_alloc_type = B_FALSE; 3225 tx_dmap->kmem_alloc_type = B_FALSE; 3226 3227 status = nxge_dma_mem_alloc(nxgep, nxge_force_dma, 3228 &nxge_desc_dma_attr, 3229 size, 3230 &nxge_dev_desc_dma_acc_attr, 3231 DDI_DMA_RDWR | DDI_DMA_CONSISTENT, 3232 tx_dmap); 3233 if (status != NXGE_OK) { 3234 goto nxge_alloc_tx_cntl_dma_fail1; 3235 } 3236 3237 *dmap = tx_dmap; 3238 goto nxge_alloc_tx_cntl_dma_exit; 3239 3240 nxge_alloc_tx_cntl_dma_fail1: 3241 KMEM_FREE(tx_dmap, sizeof (nxge_dma_common_t)); 3242 3243 nxge_alloc_tx_cntl_dma_exit: 3244 NXGE_DEBUG_MSG((nxgep, DMA_CTL, 3245 "<== nxge_alloc_tx_cntl_dma status 0x%08x", status)); 3246 3247 return (status); 3248 } 3249 3250 /*ARGSUSED*/ 3251 static void 3252 nxge_free_tx_cntl_dma(p_nxge_t nxgep, p_nxge_dma_common_t dmap) 3253 { 3254 NXGE_DEBUG_MSG((nxgep, DMA_CTL, "==> nxge_free_tx_cntl_dma")); 3255 3256 if (dmap == 0) 3257 return; 3258 3259 nxge_dma_mem_free(dmap); 3260 3261 NXGE_DEBUG_MSG((nxgep, DMA_CTL, "<== nxge_free_tx_cntl_dma")); 3262 } 3263 3264 /* 3265 * nxge_free_tx_mem_pool 3266 * 3267 * This function frees all of the per-port TDC control data structures. 3268 * The per-channel (TDC) data structures are freed when the channel 3269 * is stopped. 3270 * 3271 * Arguments: 3272 * nxgep 3273 * 3274 * Notes: 3275 * 3276 * Context: 3277 * Any domain 3278 */ 3279 static void 3280 nxge_free_tx_mem_pool(p_nxge_t nxgep) 3281 { 3282 int tdc_max = NXGE_MAX_TDCS; 3283 3284 NXGE_DEBUG_MSG((nxgep, MEM2_CTL, "==> nxge_free_tx_mem_pool")); 3285 3286 if (!nxgep->tx_buf_pool_p || !nxgep->tx_buf_pool_p->buf_allocated) { 3287 NXGE_DEBUG_MSG((nxgep, MEM2_CTL, 3288 "<== nxge_free_tx_mem_pool " 3289 "(null tx buf pool or buf not allocated")); 3290 return; 3291 } 3292 if (!nxgep->tx_cntl_pool_p || !nxgep->tx_cntl_pool_p->buf_allocated) { 3293 NXGE_DEBUG_MSG((nxgep, MEM2_CTL, 3294 "<== nxge_free_tx_mem_pool " 3295 "(null tx cntl buf pool or cntl buf not allocated")); 3296 return; 3297 } 3298 3299 /* 1. Free the mailboxes. */ 3300 KMEM_FREE(nxgep->tx_mbox_areas_p->txmbox_areas_p, 3301 sizeof (p_tx_mbox_t) * tdc_max); 3302 KMEM_FREE(nxgep->tx_mbox_areas_p, sizeof (tx_mbox_areas_t)); 3303 3304 nxgep->tx_mbox_areas_p = 0; 3305 3306 /* 2. Free the transmit ring arrays. */ 3307 KMEM_FREE(nxgep->tx_rings->rings, 3308 sizeof (p_tx_ring_t) * tdc_max); 3309 KMEM_FREE(nxgep->tx_rings, sizeof (tx_rings_t)); 3310 3311 nxgep->tx_rings = 0; 3312 3313 /* 3. Free the completion ring data structures. */ 3314 KMEM_FREE(nxgep->tx_cntl_pool_p->dma_buf_pool_p, 3315 sizeof (p_nxge_dma_common_t) * tdc_max); 3316 KMEM_FREE(nxgep->tx_cntl_pool_p, sizeof (nxge_dma_pool_t)); 3317 3318 nxgep->tx_cntl_pool_p = 0; 3319 3320 /* 4. Free the data ring data structures. */ 3321 KMEM_FREE(nxgep->tx_buf_pool_p->num_chunks, 3322 sizeof (uint32_t) * tdc_max); 3323 KMEM_FREE(nxgep->tx_buf_pool_p->dma_buf_pool_p, 3324 sizeof (p_nxge_dma_common_t) * tdc_max); 3325 KMEM_FREE(nxgep->tx_buf_pool_p, sizeof (nxge_dma_pool_t)); 3326 3327 nxgep->tx_buf_pool_p = 0; 3328 3329 NXGE_DEBUG_MSG((nxgep, MEM2_CTL, "<== nxge_free_tx_mem_pool")); 3330 } 3331 3332 /*ARGSUSED*/ 3333 static nxge_status_t 3334 nxge_dma_mem_alloc(p_nxge_t nxgep, dma_method_t method, 3335 struct ddi_dma_attr *dma_attrp, 3336 size_t length, ddi_device_acc_attr_t *acc_attr_p, uint_t xfer_flags, 3337 p_nxge_dma_common_t dma_p) 3338 { 3339 caddr_t kaddrp; 3340 int ddi_status = DDI_SUCCESS; 3341 boolean_t contig_alloc_type; 3342 boolean_t kmem_alloc_type; 3343 3344 contig_alloc_type = dma_p->contig_alloc_type; 3345 3346 if (contig_alloc_type && (nxgep->niu_type != N2_NIU)) { 3347 /* 3348 * contig_alloc_type for contiguous memory only allowed 3349 * for N2/NIU. 3350 */ 3351 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, 3352 "nxge_dma_mem_alloc: alloc type not allowed (%d)", 3353 dma_p->contig_alloc_type)); 3354 return (NXGE_ERROR | NXGE_DDI_FAILED); 3355 } 3356 3357 dma_p->dma_handle = NULL; 3358 dma_p->acc_handle = NULL; 3359 dma_p->kaddrp = dma_p->last_kaddrp = NULL; 3360 dma_p->first_ioaddr_pp = dma_p->last_ioaddr_pp = NULL; 3361 ddi_status = ddi_dma_alloc_handle(nxgep->dip, dma_attrp, 3362 DDI_DMA_DONTWAIT, NULL, &dma_p->dma_handle); 3363 if (ddi_status != DDI_SUCCESS) { 3364 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, 3365 "nxge_dma_mem_alloc:ddi_dma_alloc_handle failed.")); 3366 return (NXGE_ERROR | NXGE_DDI_FAILED); 3367 } 3368 3369 kmem_alloc_type = dma_p->kmem_alloc_type; 3370 3371 switch (contig_alloc_type) { 3372 case B_FALSE: 3373 switch (kmem_alloc_type) { 3374 case B_FALSE: 3375 ddi_status = ddi_dma_mem_alloc(dma_p->dma_handle, 3376 length, 3377 acc_attr_p, 3378 xfer_flags, 3379 DDI_DMA_DONTWAIT, 0, &kaddrp, &dma_p->alength, 3380 &dma_p->acc_handle); 3381 if (ddi_status != DDI_SUCCESS) { 3382 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, 3383 "nxge_dma_mem_alloc: " 3384 "ddi_dma_mem_alloc failed")); 3385 ddi_dma_free_handle(&dma_p->dma_handle); 3386 dma_p->dma_handle = NULL; 3387 return (NXGE_ERROR | NXGE_DDI_FAILED); 3388 } 3389 if (dma_p->alength < length) { 3390 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, 3391 "nxge_dma_mem_alloc:di_dma_mem_alloc " 3392 "< length.")); 3393 ddi_dma_mem_free(&dma_p->acc_handle); 3394 ddi_dma_free_handle(&dma_p->dma_handle); 3395 dma_p->acc_handle = NULL; 3396 dma_p->dma_handle = NULL; 3397 return (NXGE_ERROR); 3398 } 3399 3400 ddi_status = ddi_dma_addr_bind_handle(dma_p->dma_handle, 3401 NULL, 3402 kaddrp, dma_p->alength, xfer_flags, 3403 DDI_DMA_DONTWAIT, 3404 0, &dma_p->dma_cookie, &dma_p->ncookies); 3405 if (ddi_status != DDI_DMA_MAPPED) { 3406 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, 3407 "nxge_dma_mem_alloc: ddi_dma_addr_bind " 3408 "failed " 3409 "(staus 0x%x ncookies %d.)", ddi_status, 3410 dma_p->ncookies)); 3411 if (dma_p->acc_handle) { 3412 ddi_dma_mem_free(&dma_p->acc_handle); 3413 dma_p->acc_handle = NULL; 3414 } 3415 ddi_dma_free_handle(&dma_p->dma_handle); 3416 dma_p->dma_handle = NULL; 3417 return (NXGE_ERROR | NXGE_DDI_FAILED); 3418 } 3419 3420 if (dma_p->ncookies != 1) { 3421 NXGE_DEBUG_MSG((nxgep, DMA_CTL, 3422 "nxge_dma_mem_alloc:ddi_dma_addr_bind " 3423 "> 1 cookie" 3424 "(staus 0x%x ncookies %d.)", ddi_status, 3425 dma_p->ncookies)); 3426 (void) ddi_dma_unbind_handle(dma_p->dma_handle); 3427 if (dma_p->acc_handle) { 3428 ddi_dma_mem_free(&dma_p->acc_handle); 3429 dma_p->acc_handle = NULL; 3430 } 3431 ddi_dma_free_handle(&dma_p->dma_handle); 3432 dma_p->dma_handle = NULL; 3433 dma_p->acc_handle = NULL; 3434 return (NXGE_ERROR); 3435 } 3436 break; 3437 3438 case B_TRUE: 3439 kaddrp = KMEM_ALLOC(length, KM_NOSLEEP); 3440 if (kaddrp == NULL) { 3441 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, 3442 "nxge_dma_mem_alloc:ddi_dma_mem_alloc " 3443 "kmem alloc failed")); 3444 return (NXGE_ERROR); 3445 } 3446 3447 dma_p->alength = length; 3448 ddi_status = ddi_dma_addr_bind_handle(dma_p->dma_handle, 3449 NULL, kaddrp, dma_p->alength, xfer_flags, 3450 DDI_DMA_DONTWAIT, 0, 3451 &dma_p->dma_cookie, &dma_p->ncookies); 3452 if (ddi_status != DDI_DMA_MAPPED) { 3453 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, 3454 "nxge_dma_mem_alloc:ddi_dma_addr_bind: " 3455 "(kmem_alloc) failed kaddrp $%p length %d " 3456 "(staus 0x%x (%d) ncookies %d.)", 3457 kaddrp, length, 3458 ddi_status, ddi_status, dma_p->ncookies)); 3459 KMEM_FREE(kaddrp, length); 3460 dma_p->acc_handle = NULL; 3461 ddi_dma_free_handle(&dma_p->dma_handle); 3462 dma_p->dma_handle = NULL; 3463 dma_p->kaddrp = NULL; 3464 return (NXGE_ERROR | NXGE_DDI_FAILED); 3465 } 3466 3467 if (dma_p->ncookies != 1) { 3468 NXGE_DEBUG_MSG((nxgep, DMA_CTL, 3469 "nxge_dma_mem_alloc:ddi_dma_addr_bind " 3470 "(kmem_alloc) > 1 cookie" 3471 "(staus 0x%x ncookies %d.)", ddi_status, 3472 dma_p->ncookies)); 3473 (void) ddi_dma_unbind_handle(dma_p->dma_handle); 3474 KMEM_FREE(kaddrp, length); 3475 ddi_dma_free_handle(&dma_p->dma_handle); 3476 dma_p->dma_handle = NULL; 3477 dma_p->acc_handle = NULL; 3478 dma_p->kaddrp = NULL; 3479 return (NXGE_ERROR); 3480 } 3481 3482 dma_p->kaddrp = kaddrp; 3483 3484 NXGE_DEBUG_MSG((nxgep, NXGE_ERR_CTL, 3485 "nxge_dma_mem_alloc: kmem_alloc dmap $%p " 3486 "kaddr $%p alength %d", 3487 dma_p, 3488 kaddrp, 3489 dma_p->alength)); 3490 break; 3491 } 3492 break; 3493 3494 #if defined(sun4v) && defined(NIU_LP_WORKAROUND) 3495 case B_TRUE: 3496 kaddrp = (caddr_t)contig_mem_alloc(length); 3497 if (kaddrp == NULL) { 3498 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, 3499 "nxge_dma_mem_alloc:contig_mem_alloc failed.")); 3500 ddi_dma_free_handle(&dma_p->dma_handle); 3501 return (NXGE_ERROR | NXGE_DDI_FAILED); 3502 } 3503 3504 dma_p->alength = length; 3505 ddi_status = ddi_dma_addr_bind_handle(dma_p->dma_handle, NULL, 3506 kaddrp, dma_p->alength, xfer_flags, DDI_DMA_DONTWAIT, 0, 3507 &dma_p->dma_cookie, &dma_p->ncookies); 3508 if (ddi_status != DDI_DMA_MAPPED) { 3509 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, 3510 "nxge_dma_mem_alloc:di_dma_addr_bind failed " 3511 "(status 0x%x ncookies %d.)", ddi_status, 3512 dma_p->ncookies)); 3513 3514 NXGE_DEBUG_MSG((nxgep, DMA_CTL, 3515 "==> nxge_dma_mem_alloc: (not mapped)" 3516 "length %lu (0x%x) " 3517 "free contig kaddrp $%p " 3518 "va_to_pa $%p", 3519 length, length, 3520 kaddrp, 3521 va_to_pa(kaddrp))); 3522 3523 3524 contig_mem_free((void *)kaddrp, length); 3525 ddi_dma_free_handle(&dma_p->dma_handle); 3526 3527 dma_p->dma_handle = NULL; 3528 dma_p->acc_handle = NULL; 3529 dma_p->alength = NULL; 3530 dma_p->kaddrp = NULL; 3531 3532 return (NXGE_ERROR | NXGE_DDI_FAILED); 3533 } 3534 3535 if (dma_p->ncookies != 1 || 3536 (dma_p->dma_cookie.dmac_laddress == NULL)) { 3537 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, 3538 "nxge_dma_mem_alloc:di_dma_addr_bind > 1 " 3539 "cookie or " 3540 "dmac_laddress is NULL $%p size %d " 3541 " (status 0x%x ncookies %d.)", 3542 ddi_status, 3543 dma_p->dma_cookie.dmac_laddress, 3544 dma_p->dma_cookie.dmac_size, 3545 dma_p->ncookies)); 3546 3547 contig_mem_free((void *)kaddrp, length); 3548 (void) ddi_dma_unbind_handle(dma_p->dma_handle); 3549 ddi_dma_free_handle(&dma_p->dma_handle); 3550 3551 dma_p->alength = 0; 3552 dma_p->dma_handle = NULL; 3553 dma_p->acc_handle = NULL; 3554 dma_p->kaddrp = NULL; 3555 3556 return (NXGE_ERROR | NXGE_DDI_FAILED); 3557 } 3558 break; 3559 3560 #else 3561 case B_TRUE: 3562 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, 3563 "nxge_dma_mem_alloc: invalid alloc type for !sun4v")); 3564 return (NXGE_ERROR | NXGE_DDI_FAILED); 3565 #endif 3566 } 3567 3568 dma_p->kaddrp = kaddrp; 3569 dma_p->last_kaddrp = (unsigned char *)kaddrp + 3570 dma_p->alength - RXBUF_64B_ALIGNED; 3571 #if defined(__i386) 3572 dma_p->ioaddr_pp = 3573 (unsigned char *)(uint32_t)dma_p->dma_cookie.dmac_laddress; 3574 #else 3575 dma_p->ioaddr_pp = (unsigned char *)dma_p->dma_cookie.dmac_laddress; 3576 #endif 3577 dma_p->last_ioaddr_pp = 3578 #if defined(__i386) 3579 (unsigned char *)(uint32_t)dma_p->dma_cookie.dmac_laddress + 3580 #else 3581 (unsigned char *)dma_p->dma_cookie.dmac_laddress + 3582 #endif 3583 dma_p->alength - RXBUF_64B_ALIGNED; 3584 3585 NPI_DMA_ACC_HANDLE_SET(dma_p, dma_p->acc_handle); 3586 3587 #if defined(sun4v) && defined(NIU_LP_WORKAROUND) 3588 dma_p->orig_ioaddr_pp = 3589 (unsigned char *)dma_p->dma_cookie.dmac_laddress; 3590 dma_p->orig_alength = length; 3591 dma_p->orig_kaddrp = kaddrp; 3592 dma_p->orig_vatopa = (uint64_t)va_to_pa(kaddrp); 3593 #endif 3594 3595 NXGE_DEBUG_MSG((nxgep, DMA_CTL, "<== nxge_dma_mem_alloc: " 3596 "dma buffer allocated: dma_p $%p " 3597 "return dmac_ladress from cookie $%p cookie dmac_size %d " 3598 "dma_p->ioaddr_p $%p " 3599 "dma_p->orig_ioaddr_p $%p " 3600 "orig_vatopa $%p " 3601 "alength %d (0x%x) " 3602 "kaddrp $%p " 3603 "length %d (0x%x)", 3604 dma_p, 3605 dma_p->dma_cookie.dmac_laddress, dma_p->dma_cookie.dmac_size, 3606 dma_p->ioaddr_pp, 3607 dma_p->orig_ioaddr_pp, 3608 dma_p->orig_vatopa, 3609 dma_p->alength, dma_p->alength, 3610 kaddrp, 3611 length, length)); 3612 3613 return (NXGE_OK); 3614 } 3615 3616 static void 3617 nxge_dma_mem_free(p_nxge_dma_common_t dma_p) 3618 { 3619 if (dma_p->dma_handle != NULL) { 3620 if (dma_p->ncookies) { 3621 (void) ddi_dma_unbind_handle(dma_p->dma_handle); 3622 dma_p->ncookies = 0; 3623 } 3624 ddi_dma_free_handle(&dma_p->dma_handle); 3625 dma_p->dma_handle = NULL; 3626 } 3627 3628 if (dma_p->acc_handle != NULL) { 3629 ddi_dma_mem_free(&dma_p->acc_handle); 3630 dma_p->acc_handle = NULL; 3631 NPI_DMA_ACC_HANDLE_SET(dma_p, NULL); 3632 } 3633 3634 #if defined(sun4v) && defined(NIU_LP_WORKAROUND) 3635 if (dma_p->contig_alloc_type && 3636 dma_p->orig_kaddrp && dma_p->orig_alength) { 3637 NXGE_DEBUG_MSG((NULL, DMA_CTL, "nxge_dma_mem_free: " 3638 "kaddrp $%p (orig_kaddrp $%p)" 3639 "mem type %d ", 3640 "orig_alength %d " 3641 "alength 0x%x (%d)", 3642 dma_p->kaddrp, 3643 dma_p->orig_kaddrp, 3644 dma_p->contig_alloc_type, 3645 dma_p->orig_alength, 3646 dma_p->alength, dma_p->alength)); 3647 3648 contig_mem_free(dma_p->orig_kaddrp, dma_p->orig_alength); 3649 dma_p->orig_alength = NULL; 3650 dma_p->orig_kaddrp = NULL; 3651 dma_p->contig_alloc_type = B_FALSE; 3652 } 3653 #endif 3654 dma_p->kaddrp = NULL; 3655 dma_p->alength = NULL; 3656 } 3657 3658 static void 3659 nxge_dma_free_rx_data_buf(p_nxge_dma_common_t dma_p) 3660 { 3661 uint64_t kaddr; 3662 uint32_t buf_size; 3663 3664 NXGE_DEBUG_MSG((NULL, DMA_CTL, "==> nxge_dma_free_rx_data_buf")); 3665 3666 if (dma_p->dma_handle != NULL) { 3667 if (dma_p->ncookies) { 3668 (void) ddi_dma_unbind_handle(dma_p->dma_handle); 3669 dma_p->ncookies = 0; 3670 } 3671 ddi_dma_free_handle(&dma_p->dma_handle); 3672 dma_p->dma_handle = NULL; 3673 } 3674 3675 if (dma_p->acc_handle != NULL) { 3676 ddi_dma_mem_free(&dma_p->acc_handle); 3677 dma_p->acc_handle = NULL; 3678 NPI_DMA_ACC_HANDLE_SET(dma_p, NULL); 3679 } 3680 3681 NXGE_DEBUG_MSG((NULL, DMA_CTL, 3682 "==> nxge_dma_free_rx_data_buf: dmap $%p buf_alloc_state %d", 3683 dma_p, 3684 dma_p->buf_alloc_state)); 3685 3686 if (!(dma_p->buf_alloc_state & BUF_ALLOCATED_WAIT_FREE)) { 3687 NXGE_DEBUG_MSG((NULL, DMA_CTL, 3688 "<== nxge_dma_free_rx_data_buf: " 3689 "outstanding data buffers")); 3690 return; 3691 } 3692 3693 #if defined(sun4v) && defined(NIU_LP_WORKAROUND) 3694 if (dma_p->contig_alloc_type && 3695 dma_p->orig_kaddrp && dma_p->orig_alength) { 3696 NXGE_DEBUG_MSG((NULL, DMA_CTL, "nxge_dma_free_rx_data_buf: " 3697 "kaddrp $%p (orig_kaddrp $%p)" 3698 "mem type %d ", 3699 "orig_alength %d " 3700 "alength 0x%x (%d)", 3701 dma_p->kaddrp, 3702 dma_p->orig_kaddrp, 3703 dma_p->contig_alloc_type, 3704 dma_p->orig_alength, 3705 dma_p->alength, dma_p->alength)); 3706 3707 kaddr = (uint64_t)dma_p->orig_kaddrp; 3708 buf_size = dma_p->orig_alength; 3709 nxge_free_buf(CONTIG_MEM_ALLOC, kaddr, buf_size); 3710 dma_p->orig_alength = NULL; 3711 dma_p->orig_kaddrp = NULL; 3712 dma_p->contig_alloc_type = B_FALSE; 3713 dma_p->kaddrp = NULL; 3714 dma_p->alength = NULL; 3715 return; 3716 } 3717 #endif 3718 3719 if (dma_p->kmem_alloc_type) { 3720 NXGE_DEBUG_MSG((NULL, DMA_CTL, 3721 "nxge_dma_free_rx_data_buf: free kmem " 3722 "kaddrp $%p (orig_kaddrp $%p)" 3723 "alloc type %d " 3724 "orig_alength %d " 3725 "alength 0x%x (%d)", 3726 dma_p->kaddrp, 3727 dma_p->orig_kaddrp, 3728 dma_p->kmem_alloc_type, 3729 dma_p->orig_alength, 3730 dma_p->alength, dma_p->alength)); 3731 #if defined(__i386) 3732 kaddr = (uint64_t)(uint32_t)dma_p->kaddrp; 3733 #else 3734 kaddr = (uint64_t)dma_p->kaddrp; 3735 #endif 3736 buf_size = dma_p->orig_alength; 3737 NXGE_DEBUG_MSG((NULL, DMA_CTL, 3738 "nxge_dma_free_rx_data_buf: free dmap $%p " 3739 "kaddr $%p buf_size %d", 3740 dma_p, 3741 kaddr, buf_size)); 3742 nxge_free_buf(KMEM_ALLOC, kaddr, buf_size); 3743 dma_p->alength = 0; 3744 dma_p->orig_alength = 0; 3745 dma_p->kaddrp = NULL; 3746 dma_p->kmem_alloc_type = B_FALSE; 3747 } 3748 3749 NXGE_DEBUG_MSG((NULL, DMA_CTL, "<== nxge_dma_free_rx_data_buf")); 3750 } 3751 3752 /* 3753 * nxge_m_start() -- start transmitting and receiving. 3754 * 3755 * This function is called by the MAC layer when the first 3756 * stream is open to prepare the hardware ready for sending 3757 * and transmitting packets. 3758 */ 3759 static int 3760 nxge_m_start(void *arg) 3761 { 3762 p_nxge_t nxgep = (p_nxge_t)arg; 3763 3764 NXGE_DEBUG_MSG((nxgep, NXGE_CTL, "==> nxge_m_start")); 3765 3766 /* 3767 * Are we already started? 3768 */ 3769 if (nxgep->nxge_mac_state == NXGE_MAC_STARTED) { 3770 return (0); 3771 } 3772 3773 if (nxge_peu_reset_enable && !nxgep->nxge_link_poll_timerid) { 3774 (void) nxge_link_monitor(nxgep, LINK_MONITOR_START); 3775 } 3776 3777 /* 3778 * Make sure RX MAC is disabled while we initialize. 3779 */ 3780 if (!isLDOMguest(nxgep)) { 3781 (void) nxge_rx_mac_disable(nxgep); 3782 } 3783 3784 /* 3785 * Grab the global lock. 3786 */ 3787 MUTEX_ENTER(nxgep->genlock); 3788 3789 /* 3790 * Initialize the driver and hardware. 3791 */ 3792 if (nxge_init(nxgep) != NXGE_OK) { 3793 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, 3794 "<== nxge_m_start: initialization failed")); 3795 MUTEX_EXIT(nxgep->genlock); 3796 return (EIO); 3797 } 3798 3799 /* 3800 * Start timer to check the system error and tx hangs 3801 */ 3802 if (!isLDOMguest(nxgep)) 3803 nxgep->nxge_timerid = nxge_start_timer(nxgep, 3804 nxge_check_hw_state, NXGE_CHECK_TIMER); 3805 #if defined(sun4v) 3806 else 3807 nxge_hio_start_timer(nxgep); 3808 #endif 3809 3810 nxgep->link_notify = B_TRUE; 3811 nxgep->link_check_count = 0; 3812 nxgep->nxge_mac_state = NXGE_MAC_STARTED; 3813 3814 /* 3815 * Let the global lock go, since we are intialized. 3816 */ 3817 MUTEX_EXIT(nxgep->genlock); 3818 3819 /* 3820 * Let the MAC start receiving packets, now that 3821 * we are initialized. 3822 */ 3823 if (!isLDOMguest(nxgep)) { 3824 if (nxge_rx_mac_enable(nxgep) != NXGE_OK) { 3825 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, 3826 "<== nxge_m_start: enable of RX mac failed")); 3827 return (EIO); 3828 } 3829 3830 /* 3831 * Enable hardware interrupts. 3832 */ 3833 nxge_intr_hw_enable(nxgep); 3834 } 3835 #if defined(sun4v) 3836 else { 3837 /* 3838 * In guest domain we enable RDCs and their interrupts as 3839 * the last step. 3840 */ 3841 if (nxge_hio_rdc_enable(nxgep) != NXGE_OK) { 3842 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, 3843 "<== nxge_m_start: enable of RDCs failed")); 3844 return (EIO); 3845 } 3846 3847 if (nxge_hio_rdc_intr_arm(nxgep, B_TRUE) != NXGE_OK) { 3848 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, 3849 "<== nxge_m_start: intrs enable for RDCs failed")); 3850 return (EIO); 3851 } 3852 } 3853 #endif 3854 NXGE_DEBUG_MSG((nxgep, NXGE_CTL, "<== nxge_m_start")); 3855 return (0); 3856 } 3857 3858 static boolean_t 3859 nxge_check_groups_stopped(p_nxge_t nxgep) 3860 { 3861 int i; 3862 3863 for (i = 0; i < NXGE_MAX_RDC_GROUPS; i++) { 3864 if (nxgep->rx_hio_groups[i].started) 3865 return (B_FALSE); 3866 } 3867 3868 return (B_TRUE); 3869 } 3870 3871 /* 3872 * nxge_m_stop(): stop transmitting and receiving. 3873 */ 3874 static void 3875 nxge_m_stop(void *arg) 3876 { 3877 p_nxge_t nxgep = (p_nxge_t)arg; 3878 boolean_t groups_stopped; 3879 3880 NXGE_DEBUG_MSG((nxgep, NXGE_CTL, "==> nxge_m_stop")); 3881 3882 /* 3883 * Are the groups stopped? 3884 */ 3885 groups_stopped = nxge_check_groups_stopped(nxgep); 3886 ASSERT(groups_stopped == B_TRUE); 3887 if (!groups_stopped) { 3888 cmn_err(CE_WARN, "nxge(%d): groups are not stopped!\n", 3889 nxgep->instance); 3890 return; 3891 } 3892 3893 if (!isLDOMguest(nxgep)) { 3894 /* 3895 * Disable the RX mac. 3896 */ 3897 (void) nxge_rx_mac_disable(nxgep); 3898 3899 /* 3900 * Wait for the IPP to drain. 3901 */ 3902 (void) nxge_ipp_drain(nxgep); 3903 3904 /* 3905 * Disable hardware interrupts. 3906 */ 3907 nxge_intr_hw_disable(nxgep); 3908 } 3909 #if defined(sun4v) 3910 else { 3911 (void) nxge_hio_rdc_intr_arm(nxgep, B_FALSE); 3912 } 3913 #endif 3914 3915 /* 3916 * Grab the global lock. 3917 */ 3918 MUTEX_ENTER(nxgep->genlock); 3919 3920 nxgep->nxge_mac_state = NXGE_MAC_STOPPING; 3921 if (nxgep->nxge_timerid) { 3922 nxge_stop_timer(nxgep, nxgep->nxge_timerid); 3923 nxgep->nxge_timerid = 0; 3924 } 3925 3926 /* 3927 * Clean up. 3928 */ 3929 nxge_uninit(nxgep); 3930 3931 nxgep->nxge_mac_state = NXGE_MAC_STOPPED; 3932 3933 /* 3934 * Let go of the global lock. 3935 */ 3936 MUTEX_EXIT(nxgep->genlock); 3937 NXGE_DEBUG_MSG((nxgep, NXGE_CTL, "<== nxge_m_stop")); 3938 } 3939 3940 static int 3941 nxge_m_multicst(void *arg, boolean_t add, const uint8_t *mca) 3942 { 3943 p_nxge_t nxgep = (p_nxge_t)arg; 3944 struct ether_addr addrp; 3945 3946 NXGE_DEBUG_MSG((nxgep, MAC_CTL, 3947 "==> nxge_m_multicst: add %d", add)); 3948 3949 bcopy(mca, (uint8_t *)&addrp, ETHERADDRL); 3950 if (add) { 3951 if (nxge_add_mcast_addr(nxgep, &addrp)) { 3952 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, 3953 "<== nxge_m_multicst: add multicast failed")); 3954 return (EINVAL); 3955 } 3956 } else { 3957 if (nxge_del_mcast_addr(nxgep, &addrp)) { 3958 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, 3959 "<== nxge_m_multicst: del multicast failed")); 3960 return (EINVAL); 3961 } 3962 } 3963 3964 NXGE_DEBUG_MSG((nxgep, MAC_CTL, "<== nxge_m_multicst")); 3965 3966 return (0); 3967 } 3968 3969 static int 3970 nxge_m_promisc(void *arg, boolean_t on) 3971 { 3972 p_nxge_t nxgep = (p_nxge_t)arg; 3973 3974 NXGE_DEBUG_MSG((nxgep, MAC_CTL, 3975 "==> nxge_m_promisc: on %d", on)); 3976 3977 if (nxge_set_promisc(nxgep, on)) { 3978 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, 3979 "<== nxge_m_promisc: set promisc failed")); 3980 return (EINVAL); 3981 } 3982 3983 NXGE_DEBUG_MSG((nxgep, MAC_CTL, 3984 "<== nxge_m_promisc: on %d", on)); 3985 3986 return (0); 3987 } 3988 3989 static void 3990 nxge_m_ioctl(void *arg, queue_t *wq, mblk_t *mp) 3991 { 3992 p_nxge_t nxgep = (p_nxge_t)arg; 3993 struct iocblk *iocp; 3994 boolean_t need_privilege; 3995 int err; 3996 int cmd; 3997 3998 NXGE_DEBUG_MSG((nxgep, NXGE_CTL, "==> nxge_m_ioctl")); 3999 4000 iocp = (struct iocblk *)mp->b_rptr; 4001 iocp->ioc_error = 0; 4002 need_privilege = B_TRUE; 4003 cmd = iocp->ioc_cmd; 4004 NXGE_DEBUG_MSG((nxgep, NXGE_CTL, "==> nxge_m_ioctl: cmd 0x%08x", cmd)); 4005 switch (cmd) { 4006 default: 4007 miocnak(wq, mp, 0, EINVAL); 4008 NXGE_DEBUG_MSG((nxgep, NXGE_CTL, "<== nxge_m_ioctl: invalid")); 4009 return; 4010 4011 case LB_GET_INFO_SIZE: 4012 case LB_GET_INFO: 4013 case LB_GET_MODE: 4014 need_privilege = B_FALSE; 4015 break; 4016 case LB_SET_MODE: 4017 break; 4018 4019 4020 case NXGE_GET_MII: 4021 case NXGE_PUT_MII: 4022 case NXGE_GET64: 4023 case NXGE_PUT64: 4024 case NXGE_GET_TX_RING_SZ: 4025 case NXGE_GET_TX_DESC: 4026 case NXGE_TX_SIDE_RESET: 4027 case NXGE_RX_SIDE_RESET: 4028 case NXGE_GLOBAL_RESET: 4029 case NXGE_RESET_MAC: 4030 case NXGE_TX_REGS_DUMP: 4031 case NXGE_RX_REGS_DUMP: 4032 case NXGE_INT_REGS_DUMP: 4033 case NXGE_VIR_INT_REGS_DUMP: 4034 case NXGE_PUT_TCAM: 4035 case NXGE_GET_TCAM: 4036 case NXGE_RTRACE: 4037 case NXGE_RDUMP: 4038 case NXGE_RX_CLASS: 4039 case NXGE_RX_HASH: 4040 4041 need_privilege = B_FALSE; 4042 break; 4043 case NXGE_INJECT_ERR: 4044 cmn_err(CE_NOTE, "!nxge_m_ioctl: Inject error\n"); 4045 nxge_err_inject(nxgep, wq, mp); 4046 break; 4047 } 4048 4049 if (need_privilege) { 4050 err = secpolicy_net_config(iocp->ioc_cr, B_FALSE); 4051 if (err != 0) { 4052 miocnak(wq, mp, 0, err); 4053 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, 4054 "<== nxge_m_ioctl: no priv")); 4055 return; 4056 } 4057 } 4058 4059 switch (cmd) { 4060 4061 case LB_GET_MODE: 4062 case LB_SET_MODE: 4063 case LB_GET_INFO_SIZE: 4064 case LB_GET_INFO: 4065 nxge_loopback_ioctl(nxgep, wq, mp, iocp); 4066 break; 4067 4068 case NXGE_GET_MII: 4069 case NXGE_PUT_MII: 4070 case NXGE_PUT_TCAM: 4071 case NXGE_GET_TCAM: 4072 case NXGE_GET64: 4073 case NXGE_PUT64: 4074 case NXGE_GET_TX_RING_SZ: 4075 case NXGE_GET_TX_DESC: 4076 case NXGE_TX_SIDE_RESET: 4077 case NXGE_RX_SIDE_RESET: 4078 case NXGE_GLOBAL_RESET: 4079 case NXGE_RESET_MAC: 4080 case NXGE_TX_REGS_DUMP: 4081 case NXGE_RX_REGS_DUMP: 4082 case NXGE_INT_REGS_DUMP: 4083 case NXGE_VIR_INT_REGS_DUMP: 4084 NXGE_DEBUG_MSG((nxgep, NXGE_CTL, 4085 "==> nxge_m_ioctl: cmd 0x%x", cmd)); 4086 nxge_hw_ioctl(nxgep, wq, mp, iocp); 4087 break; 4088 case NXGE_RX_CLASS: 4089 if (nxge_rxclass_ioctl(nxgep, wq, mp->b_cont) < 0) 4090 miocnak(wq, mp, 0, EINVAL); 4091 else 4092 miocack(wq, mp, sizeof (rx_class_cfg_t), 0); 4093 break; 4094 case NXGE_RX_HASH: 4095 4096 if (nxge_rxhash_ioctl(nxgep, wq, mp->b_cont) < 0) 4097 miocnak(wq, mp, 0, EINVAL); 4098 else 4099 miocack(wq, mp, sizeof (cfg_cmd_t), 0); 4100 break; 4101 } 4102 4103 NXGE_DEBUG_MSG((nxgep, NXGE_CTL, "<== nxge_m_ioctl")); 4104 } 4105 4106 extern void nxge_rx_hw_blank(void *arg, time_t ticks, uint_t count); 4107 4108 void 4109 nxge_mmac_kstat_update(p_nxge_t nxgep, int slot, boolean_t factory) 4110 { 4111 p_nxge_mmac_stats_t mmac_stats; 4112 int i; 4113 nxge_mmac_t *mmac_info; 4114 4115 mmac_info = &nxgep->nxge_mmac_info; 4116 4117 mmac_stats = &nxgep->statsp->mmac_stats; 4118 mmac_stats->mmac_max_cnt = mmac_info->num_mmac; 4119 mmac_stats->mmac_avail_cnt = mmac_info->naddrfree; 4120 4121 for (i = 0; i < ETHERADDRL; i++) { 4122 if (factory) { 4123 mmac_stats->mmac_avail_pool[slot-1].ether_addr_octet[i] 4124 = mmac_info->factory_mac_pool[slot][ 4125 (ETHERADDRL-1) - i]; 4126 } else { 4127 mmac_stats->mmac_avail_pool[slot-1].ether_addr_octet[i] 4128 = mmac_info->mac_pool[slot].addr[ 4129 (ETHERADDRL - 1) - i]; 4130 } 4131 } 4132 } 4133 4134 /* 4135 * nxge_altmac_set() -- Set an alternate MAC address 4136 */ 4137 static int 4138 nxge_altmac_set(p_nxge_t nxgep, uint8_t *maddr, int slot, 4139 int rdctbl, boolean_t usetbl) 4140 { 4141 uint8_t addrn; 4142 uint8_t portn; 4143 npi_mac_addr_t altmac; 4144 hostinfo_t mac_rdc; 4145 p_nxge_class_pt_cfg_t clscfgp; 4146 4147 4148 altmac.w2 = ((uint16_t)maddr[0] << 8) | ((uint16_t)maddr[1] & 0x0ff); 4149 altmac.w1 = ((uint16_t)maddr[2] << 8) | ((uint16_t)maddr[3] & 0x0ff); 4150 altmac.w0 = ((uint16_t)maddr[4] << 8) | ((uint16_t)maddr[5] & 0x0ff); 4151 4152 portn = nxgep->mac.portnum; 4153 addrn = (uint8_t)slot - 1; 4154 4155 if (npi_mac_altaddr_entry(nxgep->npi_handle, OP_SET, 4156 nxgep->function_num, addrn, &altmac) != NPI_SUCCESS) 4157 return (EIO); 4158 4159 /* 4160 * Set the rdc table number for the host info entry 4161 * for this mac address slot. 4162 */ 4163 clscfgp = (p_nxge_class_pt_cfg_t)&nxgep->class_config; 4164 mac_rdc.value = 0; 4165 if (usetbl) 4166 mac_rdc.bits.w0.rdc_tbl_num = rdctbl; 4167 else 4168 mac_rdc.bits.w0.rdc_tbl_num = 4169 clscfgp->mac_host_info[addrn].rdctbl; 4170 mac_rdc.bits.w0.mac_pref = clscfgp->mac_host_info[addrn].mpr_npr; 4171 4172 if (npi_mac_hostinfo_entry(nxgep->npi_handle, OP_SET, 4173 nxgep->function_num, addrn, &mac_rdc) != NPI_SUCCESS) { 4174 return (EIO); 4175 } 4176 4177 /* 4178 * Enable comparison with the alternate MAC address. 4179 * While the first alternate addr is enabled by bit 1 of register 4180 * BMAC_ALTAD_CMPEN, it is enabled by bit 0 of register 4181 * XMAC_ADDR_CMPEN, so slot needs to be converted to addrn 4182 * accordingly before calling npi_mac_altaddr_entry. 4183 */ 4184 if (portn == XMAC_PORT_0 || portn == XMAC_PORT_1) 4185 addrn = (uint8_t)slot - 1; 4186 else 4187 addrn = (uint8_t)slot; 4188 4189 if (npi_mac_altaddr_enable(nxgep->npi_handle, 4190 nxgep->function_num, addrn) != NPI_SUCCESS) { 4191 return (EIO); 4192 } 4193 4194 return (0); 4195 } 4196 4197 /* 4198 * nxeg_m_mmac_add_g() - find an unused address slot, set the address 4199 * value to the one specified, enable the port to start filtering on 4200 * the new MAC address. Returns 0 on success. 4201 */ 4202 int 4203 nxge_m_mmac_add_g(void *arg, const uint8_t *maddr, int rdctbl, 4204 boolean_t usetbl) 4205 { 4206 p_nxge_t nxgep = arg; 4207 int slot; 4208 nxge_mmac_t *mmac_info; 4209 int err; 4210 nxge_status_t status; 4211 4212 mutex_enter(nxgep->genlock); 4213 4214 /* 4215 * Make sure that nxge is initialized, if _start() has 4216 * not been called. 4217 */ 4218 if (!(nxgep->drv_state & STATE_HW_INITIALIZED)) { 4219 status = nxge_init(nxgep); 4220 if (status != NXGE_OK) { 4221 mutex_exit(nxgep->genlock); 4222 return (ENXIO); 4223 } 4224 } 4225 4226 mmac_info = &nxgep->nxge_mmac_info; 4227 if (mmac_info->naddrfree == 0) { 4228 mutex_exit(nxgep->genlock); 4229 return (ENOSPC); 4230 } 4231 4232 /* 4233 * Search for the first available slot. Because naddrfree 4234 * is not zero, we are guaranteed to find one. 4235 * Each of the first two ports of Neptune has 16 alternate 4236 * MAC slots but only the first 7 (of 15) slots have assigned factory 4237 * MAC addresses. We first search among the slots without bundled 4238 * factory MACs. If we fail to find one in that range, then we 4239 * search the slots with bundled factory MACs. A factory MAC 4240 * will be wasted while the slot is used with a user MAC address. 4241 * But the slot could be used by factory MAC again after calling 4242 * nxge_m_mmac_remove and nxge_m_mmac_reserve. 4243 */ 4244 for (slot = 0; slot <= mmac_info->num_mmac; slot++) { 4245 if (!(mmac_info->mac_pool[slot].flags & MMAC_SLOT_USED)) 4246 break; 4247 } 4248 4249 ASSERT(slot <= mmac_info->num_mmac); 4250 4251 if ((err = nxge_altmac_set(nxgep, (uint8_t *)maddr, slot, rdctbl, 4252 usetbl)) != 0) { 4253 mutex_exit(nxgep->genlock); 4254 return (err); 4255 } 4256 4257 bcopy(maddr, mmac_info->mac_pool[slot].addr, ETHERADDRL); 4258 mmac_info->mac_pool[slot].flags |= MMAC_SLOT_USED; 4259 mmac_info->mac_pool[slot].flags &= ~MMAC_VENDOR_ADDR; 4260 mmac_info->naddrfree--; 4261 nxge_mmac_kstat_update(nxgep, slot, B_FALSE); 4262 4263 mutex_exit(nxgep->genlock); 4264 return (0); 4265 } 4266 4267 /* 4268 * Remove the specified mac address and update the HW not to filter 4269 * the mac address anymore. 4270 */ 4271 int 4272 nxge_m_mmac_remove(void *arg, int slot) 4273 { 4274 p_nxge_t nxgep = arg; 4275 nxge_mmac_t *mmac_info; 4276 uint8_t addrn; 4277 uint8_t portn; 4278 int err = 0; 4279 nxge_status_t status; 4280 4281 mutex_enter(nxgep->genlock); 4282 4283 /* 4284 * Make sure that nxge is initialized, if _start() has 4285 * not been called. 4286 */ 4287 if (!(nxgep->drv_state & STATE_HW_INITIALIZED)) { 4288 status = nxge_init(nxgep); 4289 if (status != NXGE_OK) { 4290 mutex_exit(nxgep->genlock); 4291 return (ENXIO); 4292 } 4293 } 4294 4295 mmac_info = &nxgep->nxge_mmac_info; 4296 if (slot < 1 || slot > mmac_info->num_mmac) { 4297 mutex_exit(nxgep->genlock); 4298 return (EINVAL); 4299 } 4300 4301 portn = nxgep->mac.portnum; 4302 if (portn == XMAC_PORT_0 || portn == XMAC_PORT_1) 4303 addrn = (uint8_t)slot - 1; 4304 else 4305 addrn = (uint8_t)slot; 4306 4307 if (mmac_info->mac_pool[slot].flags & MMAC_SLOT_USED) { 4308 if (npi_mac_altaddr_disable(nxgep->npi_handle, portn, addrn) 4309 == NPI_SUCCESS) { 4310 mmac_info->naddrfree++; 4311 mmac_info->mac_pool[slot].flags &= ~MMAC_SLOT_USED; 4312 /* 4313 * Regardless if the MAC we just stopped filtering 4314 * is a user addr or a facory addr, we must set 4315 * the MMAC_VENDOR_ADDR flag if this slot has an 4316 * associated factory MAC to indicate that a factory 4317 * MAC is available. 4318 */ 4319 if (slot <= mmac_info->num_factory_mmac) { 4320 mmac_info->mac_pool[slot].flags 4321 |= MMAC_VENDOR_ADDR; 4322 } 4323 /* 4324 * Clear mac_pool[slot].addr so that kstat shows 0 4325 * alternate MAC address if the slot is not used. 4326 * (But nxge_m_mmac_get returns the factory MAC even 4327 * when the slot is not used!) 4328 */ 4329 bzero(mmac_info->mac_pool[slot].addr, ETHERADDRL); 4330 nxge_mmac_kstat_update(nxgep, slot, B_FALSE); 4331 } else { 4332 err = EIO; 4333 } 4334 } else { 4335 err = EINVAL; 4336 } 4337 4338 mutex_exit(nxgep->genlock); 4339 return (err); 4340 } 4341 4342 /* 4343 * The callback to query all the factory addresses. naddr must be the same as 4344 * the number of factory addresses (returned by MAC_CAPAB_MULTIFACTADDR), and 4345 * mcm_addr is the space allocated for keep all the addresses, whose size is 4346 * naddr * MAXMACADDRLEN. 4347 */ 4348 static void 4349 nxge_m_getfactaddr(void *arg, uint_t naddr, uint8_t *addr) 4350 { 4351 nxge_t *nxgep = arg; 4352 nxge_mmac_t *mmac_info; 4353 int i; 4354 4355 mutex_enter(nxgep->genlock); 4356 4357 mmac_info = &nxgep->nxge_mmac_info; 4358 ASSERT(naddr == mmac_info->num_factory_mmac); 4359 4360 for (i = 0; i < naddr; i++) { 4361 bcopy(mmac_info->factory_mac_pool[i + 1], 4362 addr + i * MAXMACADDRLEN, ETHERADDRL); 4363 } 4364 4365 mutex_exit(nxgep->genlock); 4366 } 4367 4368 4369 static boolean_t 4370 nxge_m_getcapab(void *arg, mac_capab_t cap, void *cap_data) 4371 { 4372 nxge_t *nxgep = arg; 4373 uint32_t *txflags = cap_data; 4374 4375 switch (cap) { 4376 case MAC_CAPAB_HCKSUM: 4377 NXGE_DEBUG_MSG((nxgep, NXGE_CTL, 4378 "==> nxge_m_getcapab: checksum %d", nxge_cksum_offload)); 4379 if (nxge_cksum_offload <= 1) { 4380 *txflags = HCKSUM_INET_PARTIAL; 4381 } 4382 break; 4383 4384 case MAC_CAPAB_MULTIFACTADDR: { 4385 mac_capab_multifactaddr_t *mfacp = cap_data; 4386 4387 if (!isLDOMguest(nxgep)) { 4388 mutex_enter(nxgep->genlock); 4389 mfacp->mcm_naddr = 4390 nxgep->nxge_mmac_info.num_factory_mmac; 4391 mfacp->mcm_getaddr = nxge_m_getfactaddr; 4392 mutex_exit(nxgep->genlock); 4393 } 4394 break; 4395 } 4396 4397 case MAC_CAPAB_LSO: { 4398 mac_capab_lso_t *cap_lso = cap_data; 4399 4400 if (nxgep->soft_lso_enable) { 4401 if (nxge_cksum_offload <= 1) { 4402 cap_lso->lso_flags = LSO_TX_BASIC_TCP_IPV4; 4403 if (nxge_lso_max > NXGE_LSO_MAXLEN) { 4404 nxge_lso_max = NXGE_LSO_MAXLEN; 4405 } 4406 cap_lso->lso_basic_tcp_ipv4.lso_max = 4407 nxge_lso_max; 4408 } 4409 break; 4410 } else { 4411 return (B_FALSE); 4412 } 4413 } 4414 4415 case MAC_CAPAB_RINGS: { 4416 mac_capab_rings_t *cap_rings = cap_data; 4417 p_nxge_hw_pt_cfg_t p_cfgp = &nxgep->pt_config.hw_config; 4418 4419 mutex_enter(nxgep->genlock); 4420 if (cap_rings->mr_type == MAC_RING_TYPE_RX) { 4421 if (isLDOMguest(nxgep)) { 4422 cap_rings->mr_group_type = 4423 MAC_GROUP_TYPE_STATIC; 4424 cap_rings->mr_rnum = 4425 NXGE_HIO_SHARE_MAX_CHANNELS; 4426 cap_rings->mr_rget = nxge_fill_ring; 4427 cap_rings->mr_gnum = 1; 4428 cap_rings->mr_gget = nxge_hio_group_get; 4429 cap_rings->mr_gaddring = NULL; 4430 cap_rings->mr_gremring = NULL; 4431 } else { 4432 /* 4433 * Service Domain. 4434 */ 4435 cap_rings->mr_group_type = 4436 MAC_GROUP_TYPE_DYNAMIC; 4437 cap_rings->mr_rnum = p_cfgp->max_rdcs; 4438 cap_rings->mr_rget = nxge_fill_ring; 4439 cap_rings->mr_gnum = p_cfgp->max_rdc_grpids; 4440 cap_rings->mr_gget = nxge_hio_group_get; 4441 cap_rings->mr_gaddring = nxge_group_add_ring; 4442 cap_rings->mr_gremring = nxge_group_rem_ring; 4443 } 4444 4445 NXGE_DEBUG_MSG((nxgep, RX_CTL, 4446 "==> nxge_m_getcapab: rx nrings[%d] ngroups[%d]", 4447 p_cfgp->max_rdcs, p_cfgp->max_rdc_grpids)); 4448 } else { 4449 /* 4450 * TX Rings. 4451 */ 4452 if (isLDOMguest(nxgep)) { 4453 cap_rings->mr_group_type = 4454 MAC_GROUP_TYPE_STATIC; 4455 cap_rings->mr_rnum = 4456 NXGE_HIO_SHARE_MAX_CHANNELS; 4457 cap_rings->mr_rget = nxge_fill_ring; 4458 cap_rings->mr_gnum = 0; 4459 cap_rings->mr_gget = NULL; 4460 cap_rings->mr_gaddring = NULL; 4461 cap_rings->mr_gremring = NULL; 4462 } else { 4463 /* 4464 * Service Domain. 4465 */ 4466 cap_rings->mr_group_type = 4467 MAC_GROUP_TYPE_DYNAMIC; 4468 cap_rings->mr_rnum = p_cfgp->tdc.count; 4469 cap_rings->mr_rget = nxge_fill_ring; 4470 4471 /* 4472 * Share capable. 4473 * 4474 * Do not report the default group: hence -1 4475 */ 4476 cap_rings->mr_gnum = 4477 NXGE_MAX_TDC_GROUPS / nxgep->nports - 1; 4478 cap_rings->mr_gget = nxge_hio_group_get; 4479 cap_rings->mr_gaddring = nxge_group_add_ring; 4480 cap_rings->mr_gremring = nxge_group_rem_ring; 4481 } 4482 4483 NXGE_DEBUG_MSG((nxgep, TX_CTL, 4484 "==> nxge_m_getcapab: tx rings # of rings %d", 4485 p_cfgp->tdc.count)); 4486 } 4487 mutex_exit(nxgep->genlock); 4488 break; 4489 } 4490 4491 #if defined(sun4v) 4492 case MAC_CAPAB_SHARES: { 4493 mac_capab_share_t *mshares = (mac_capab_share_t *)cap_data; 4494 4495 /* 4496 * Only the service domain driver responds to 4497 * this capability request. 4498 */ 4499 mutex_enter(nxgep->genlock); 4500 if (isLDOMservice(nxgep)) { 4501 mshares->ms_snum = 3; 4502 mshares->ms_handle = (void *)nxgep; 4503 mshares->ms_salloc = nxge_hio_share_alloc; 4504 mshares->ms_sfree = nxge_hio_share_free; 4505 mshares->ms_sadd = nxge_hio_share_add_group; 4506 mshares->ms_sremove = nxge_hio_share_rem_group; 4507 mshares->ms_squery = nxge_hio_share_query; 4508 mshares->ms_sbind = nxge_hio_share_bind; 4509 mshares->ms_sunbind = nxge_hio_share_unbind; 4510 mutex_exit(nxgep->genlock); 4511 } else { 4512 mutex_exit(nxgep->genlock); 4513 return (B_FALSE); 4514 } 4515 break; 4516 } 4517 #endif 4518 default: 4519 return (B_FALSE); 4520 } 4521 return (B_TRUE); 4522 } 4523 4524 static boolean_t 4525 nxge_param_locked(mac_prop_id_t pr_num) 4526 { 4527 /* 4528 * All adv_* parameters are locked (read-only) while 4529 * the device is in any sort of loopback mode ... 4530 */ 4531 switch (pr_num) { 4532 case MAC_PROP_ADV_1000FDX_CAP: 4533 case MAC_PROP_EN_1000FDX_CAP: 4534 case MAC_PROP_ADV_1000HDX_CAP: 4535 case MAC_PROP_EN_1000HDX_CAP: 4536 case MAC_PROP_ADV_100FDX_CAP: 4537 case MAC_PROP_EN_100FDX_CAP: 4538 case MAC_PROP_ADV_100HDX_CAP: 4539 case MAC_PROP_EN_100HDX_CAP: 4540 case MAC_PROP_ADV_10FDX_CAP: 4541 case MAC_PROP_EN_10FDX_CAP: 4542 case MAC_PROP_ADV_10HDX_CAP: 4543 case MAC_PROP_EN_10HDX_CAP: 4544 case MAC_PROP_AUTONEG: 4545 case MAC_PROP_FLOWCTRL: 4546 return (B_TRUE); 4547 } 4548 return (B_FALSE); 4549 } 4550 4551 /* 4552 * callback functions for set/get of properties 4553 */ 4554 static int 4555 nxge_m_setprop(void *barg, const char *pr_name, mac_prop_id_t pr_num, 4556 uint_t pr_valsize, const void *pr_val) 4557 { 4558 nxge_t *nxgep = barg; 4559 p_nxge_param_t param_arr = nxgep->param_arr; 4560 p_nxge_stats_t statsp = nxgep->statsp; 4561 int err = 0; 4562 4563 NXGE_DEBUG_MSG((nxgep, NXGE_CTL, "==> nxge_m_setprop")); 4564 4565 mutex_enter(nxgep->genlock); 4566 if (statsp->port_stats.lb_mode != nxge_lb_normal && 4567 nxge_param_locked(pr_num)) { 4568 /* 4569 * All adv_* parameters are locked (read-only) 4570 * while the device is in any sort of loopback mode. 4571 */ 4572 NXGE_DEBUG_MSG((nxgep, NXGE_CTL, 4573 "==> nxge_m_setprop: loopback mode: read only")); 4574 mutex_exit(nxgep->genlock); 4575 return (EBUSY); 4576 } 4577 4578 switch (pr_num) { 4579 case MAC_PROP_EN_1000FDX_CAP: 4580 nxgep->param_en_1000fdx = 4581 param_arr[param_anar_1000fdx].value = *(uint8_t *)pr_val; 4582 goto reprogram; 4583 4584 case MAC_PROP_EN_100FDX_CAP: 4585 nxgep->param_en_100fdx = 4586 param_arr[param_anar_100fdx].value = *(uint8_t *)pr_val; 4587 goto reprogram; 4588 4589 case MAC_PROP_EN_10FDX_CAP: 4590 nxgep->param_en_10fdx = 4591 param_arr[param_anar_10fdx].value = *(uint8_t *)pr_val; 4592 goto reprogram; 4593 4594 case MAC_PROP_AUTONEG: 4595 param_arr[param_autoneg].value = *(uint8_t *)pr_val; 4596 goto reprogram; 4597 4598 case MAC_PROP_MTU: { 4599 uint32_t cur_mtu, new_mtu, old_framesize; 4600 4601 cur_mtu = nxgep->mac.default_mtu; 4602 ASSERT(pr_valsize >= sizeof (new_mtu)); 4603 bcopy(pr_val, &new_mtu, sizeof (new_mtu)); 4604 4605 NXGE_DEBUG_MSG((nxgep, NXGE_CTL, 4606 "==> nxge_m_setprop: set MTU: %d is_jumbo %d", 4607 new_mtu, nxgep->mac.is_jumbo)); 4608 4609 if (new_mtu == cur_mtu) { 4610 err = 0; 4611 break; 4612 } 4613 4614 if (nxgep->nxge_mac_state == NXGE_MAC_STARTED) { 4615 err = EBUSY; 4616 break; 4617 } 4618 4619 if ((new_mtu < NXGE_DEFAULT_MTU) || 4620 (new_mtu > NXGE_MAXIMUM_MTU)) { 4621 err = EINVAL; 4622 break; 4623 } 4624 4625 old_framesize = (uint32_t)nxgep->mac.maxframesize; 4626 nxgep->mac.maxframesize = (uint16_t) 4627 (new_mtu + NXGE_EHEADER_VLAN_CRC); 4628 if (nxge_mac_set_framesize(nxgep)) { 4629 nxgep->mac.maxframesize = 4630 (uint16_t)old_framesize; 4631 err = EINVAL; 4632 break; 4633 } 4634 4635 nxgep->mac.default_mtu = new_mtu; 4636 nxgep->mac.is_jumbo = (new_mtu > NXGE_DEFAULT_MTU); 4637 4638 NXGE_DEBUG_MSG((nxgep, NXGE_CTL, 4639 "==> nxge_m_setprop: set MTU: %d maxframe %d", 4640 new_mtu, nxgep->mac.maxframesize)); 4641 break; 4642 } 4643 4644 case MAC_PROP_FLOWCTRL: { 4645 link_flowctrl_t fl; 4646 4647 ASSERT(pr_valsize >= sizeof (fl)); 4648 bcopy(pr_val, &fl, sizeof (fl)); 4649 4650 switch (fl) { 4651 case LINK_FLOWCTRL_NONE: 4652 param_arr[param_anar_pause].value = 0; 4653 break; 4654 4655 case LINK_FLOWCTRL_RX: 4656 param_arr[param_anar_pause].value = 1; 4657 break; 4658 4659 case LINK_FLOWCTRL_TX: 4660 case LINK_FLOWCTRL_BI: 4661 err = EINVAL; 4662 break; 4663 default: 4664 err = EINVAL; 4665 break; 4666 } 4667 reprogram: 4668 if ((err == 0) && !isLDOMguest(nxgep)) { 4669 if (!nxge_param_link_update(nxgep)) { 4670 err = EINVAL; 4671 } 4672 } else { 4673 err = EINVAL; 4674 } 4675 break; 4676 } 4677 4678 case MAC_PROP_PRIVATE: 4679 NXGE_DEBUG_MSG((nxgep, NXGE_CTL, 4680 "==> nxge_m_setprop: private property")); 4681 err = nxge_set_priv_prop(nxgep, pr_name, pr_valsize, pr_val); 4682 break; 4683 4684 default: 4685 err = ENOTSUP; 4686 break; 4687 } 4688 4689 mutex_exit(nxgep->genlock); 4690 4691 NXGE_DEBUG_MSG((nxgep, NXGE_CTL, 4692 "<== nxge_m_setprop (return %d)", err)); 4693 return (err); 4694 } 4695 4696 static int 4697 nxge_m_getprop(void *barg, const char *pr_name, mac_prop_id_t pr_num, 4698 uint_t pr_valsize, void *pr_val) 4699 { 4700 nxge_t *nxgep = barg; 4701 p_nxge_param_t param_arr = nxgep->param_arr; 4702 p_nxge_stats_t statsp = nxgep->statsp; 4703 4704 NXGE_DEBUG_MSG((nxgep, NXGE_CTL, 4705 "==> nxge_m_getprop: pr_num %d", pr_num)); 4706 4707 switch (pr_num) { 4708 case MAC_PROP_DUPLEX: 4709 *(uint8_t *)pr_val = statsp->mac_stats.link_duplex; 4710 break; 4711 4712 case MAC_PROP_SPEED: { 4713 uint64_t val = statsp->mac_stats.link_speed * 1000000ull; 4714 4715 ASSERT(pr_valsize >= sizeof (val)); 4716 bcopy(&val, pr_val, sizeof (val)); 4717 break; 4718 } 4719 4720 case MAC_PROP_STATUS: { 4721 link_state_t state = statsp->mac_stats.link_up ? 4722 LINK_STATE_UP : LINK_STATE_DOWN; 4723 4724 ASSERT(pr_valsize >= sizeof (state)); 4725 bcopy(&state, pr_val, sizeof (state)); 4726 break; 4727 } 4728 4729 case MAC_PROP_AUTONEG: 4730 *(uint8_t *)pr_val = param_arr[param_autoneg].value; 4731 break; 4732 4733 case MAC_PROP_FLOWCTRL: { 4734 link_flowctrl_t fl = param_arr[param_anar_pause].value != 0 ? 4735 LINK_FLOWCTRL_RX : LINK_FLOWCTRL_NONE; 4736 4737 ASSERT(pr_valsize >= sizeof (fl)); 4738 bcopy(&fl, pr_val, sizeof (fl)); 4739 break; 4740 } 4741 4742 case MAC_PROP_ADV_1000FDX_CAP: 4743 *(uint8_t *)pr_val = param_arr[param_anar_1000fdx].value; 4744 break; 4745 4746 case MAC_PROP_EN_1000FDX_CAP: 4747 *(uint8_t *)pr_val = nxgep->param_en_1000fdx; 4748 break; 4749 4750 case MAC_PROP_ADV_100FDX_CAP: 4751 *(uint8_t *)pr_val = param_arr[param_anar_100fdx].value; 4752 break; 4753 4754 case MAC_PROP_EN_100FDX_CAP: 4755 *(uint8_t *)pr_val = nxgep->param_en_100fdx; 4756 break; 4757 4758 case MAC_PROP_ADV_10FDX_CAP: 4759 *(uint8_t *)pr_val = param_arr[param_anar_10fdx].value; 4760 break; 4761 4762 case MAC_PROP_EN_10FDX_CAP: 4763 *(uint8_t *)pr_val = nxgep->param_en_10fdx; 4764 break; 4765 4766 case MAC_PROP_PRIVATE: 4767 return (nxge_get_priv_prop(nxgep, pr_name, pr_valsize, 4768 pr_val)); 4769 4770 default: 4771 return (ENOTSUP); 4772 } 4773 4774 return (0); 4775 } 4776 4777 static void 4778 nxge_m_propinfo(void *barg, const char *pr_name, mac_prop_id_t pr_num, 4779 mac_prop_info_handle_t prh) 4780 { 4781 nxge_t *nxgep = barg; 4782 p_nxge_stats_t statsp = nxgep->statsp; 4783 4784 /* 4785 * By default permissions are read/write unless specified 4786 * otherwise by the driver. 4787 */ 4788 4789 switch (pr_num) { 4790 case MAC_PROP_DUPLEX: 4791 case MAC_PROP_SPEED: 4792 case MAC_PROP_STATUS: 4793 case MAC_PROP_EN_1000HDX_CAP: 4794 case MAC_PROP_EN_100HDX_CAP: 4795 case MAC_PROP_EN_10HDX_CAP: 4796 case MAC_PROP_ADV_1000FDX_CAP: 4797 case MAC_PROP_ADV_1000HDX_CAP: 4798 case MAC_PROP_ADV_100FDX_CAP: 4799 case MAC_PROP_ADV_100HDX_CAP: 4800 case MAC_PROP_ADV_10FDX_CAP: 4801 case MAC_PROP_ADV_10HDX_CAP: 4802 /* 4803 * Note that read-only properties don't need to 4804 * provide default values since they cannot be 4805 * changed by the administrator. 4806 */ 4807 mac_prop_info_set_perm(prh, MAC_PROP_PERM_READ); 4808 break; 4809 4810 case MAC_PROP_EN_1000FDX_CAP: 4811 case MAC_PROP_EN_100FDX_CAP: 4812 case MAC_PROP_EN_10FDX_CAP: 4813 mac_prop_info_set_default_uint8(prh, 1); 4814 break; 4815 4816 case MAC_PROP_AUTONEG: 4817 mac_prop_info_set_default_uint8(prh, 1); 4818 break; 4819 4820 case MAC_PROP_FLOWCTRL: 4821 mac_prop_info_set_default_link_flowctrl(prh, LINK_FLOWCTRL_RX); 4822 break; 4823 4824 case MAC_PROP_MTU: 4825 mac_prop_info_set_range_uint32(prh, 4826 NXGE_DEFAULT_MTU, NXGE_MAXIMUM_MTU); 4827 break; 4828 4829 case MAC_PROP_PRIVATE: 4830 nxge_priv_propinfo(pr_name, prh); 4831 break; 4832 } 4833 4834 mutex_enter(nxgep->genlock); 4835 if (statsp->port_stats.lb_mode != nxge_lb_normal && 4836 nxge_param_locked(pr_num)) { 4837 /* 4838 * Some properties are locked (read-only) while the 4839 * device is in any sort of loopback mode. 4840 */ 4841 mac_prop_info_set_perm(prh, MAC_PROP_PERM_READ); 4842 } 4843 mutex_exit(nxgep->genlock); 4844 } 4845 4846 static void 4847 nxge_priv_propinfo(const char *pr_name, mac_prop_info_handle_t prh) 4848 { 4849 char valstr[64]; 4850 4851 bzero(valstr, sizeof (valstr)); 4852 4853 if (strcmp(pr_name, "_function_number") == 0 || 4854 strcmp(pr_name, "_fw_version") == 0 || 4855 strcmp(pr_name, "_port_mode") == 0 || 4856 strcmp(pr_name, "_hot_swap_phy") == 0) { 4857 mac_prop_info_set_perm(prh, MAC_PROP_PERM_READ); 4858 4859 } else if (strcmp(pr_name, "_rxdma_intr_time") == 0) { 4860 (void) snprintf(valstr, sizeof (valstr), 4861 "%d", RXDMA_RCR_TO_DEFAULT); 4862 4863 } else if (strcmp(pr_name, "_rxdma_intr_pkts") == 0) { 4864 (void) snprintf(valstr, sizeof (valstr), 4865 "%d", RXDMA_RCR_PTHRES_DEFAULT); 4866 4867 } else if (strcmp(pr_name, "_class_opt_ipv4_tcp") == 0 || 4868 strcmp(pr_name, "_class_opt_ipv4_udp") == 0 || 4869 strcmp(pr_name, "_class_opt_ipv4_ah") == 0 || 4870 strcmp(pr_name, "_class_opt_ipv4_sctp") == 0 || 4871 strcmp(pr_name, "_class_opt_ipv6_tcp") == 0 || 4872 strcmp(pr_name, "_class_opt_ipv6_udp") == 0 || 4873 strcmp(pr_name, "_class_opt_ipv6_ah") == 0 || 4874 strcmp(pr_name, "_class_opt_ipv6_sctp") == 0) { 4875 (void) snprintf(valstr, sizeof (valstr), "%x", 4876 NXGE_CLASS_FLOW_GEN_SERVER); 4877 4878 } else if (strcmp(pr_name, "_soft_lso_enable") == 0) { 4879 (void) snprintf(valstr, sizeof (valstr), "%d", 0); 4880 4881 } else if (strcmp(pr_name, "_adv_10gfdx_cap") == 0) { 4882 (void) snprintf(valstr, sizeof (valstr), "%d", 1); 4883 4884 } else if (strcmp(pr_name, "_adv_pause_cap") == 0) { 4885 (void) snprintf(valstr, sizeof (valstr), "%d", 1); 4886 } 4887 4888 if (strlen(valstr) > 0) 4889 mac_prop_info_set_default_str(prh, valstr); 4890 } 4891 4892 /* ARGSUSED */ 4893 static int 4894 nxge_set_priv_prop(p_nxge_t nxgep, const char *pr_name, uint_t pr_valsize, 4895 const void *pr_val) 4896 { 4897 p_nxge_param_t param_arr = nxgep->param_arr; 4898 int err = 0; 4899 long result; 4900 4901 NXGE_DEBUG_MSG((nxgep, NXGE_CTL, 4902 "==> nxge_set_priv_prop: name %s", pr_name)); 4903 4904 /* Blanking */ 4905 if (strcmp(pr_name, "_rxdma_intr_time") == 0) { 4906 err = nxge_param_rx_intr_time(nxgep, NULL, NULL, 4907 (char *)pr_val, 4908 (caddr_t)¶m_arr[param_rxdma_intr_time]); 4909 if (err) { 4910 NXGE_DEBUG_MSG((nxgep, NXGE_CTL, 4911 "<== nxge_set_priv_prop: " 4912 "unable to set (%s)", pr_name)); 4913 err = EINVAL; 4914 } else { 4915 err = 0; 4916 NXGE_DEBUG_MSG((nxgep, NXGE_CTL, 4917 "<== nxge_set_priv_prop: " 4918 "set (%s)", pr_name)); 4919 } 4920 4921 NXGE_DEBUG_MSG((nxgep, NXGE_CTL, 4922 "<== nxge_set_priv_prop: name %s (value %d)", 4923 pr_name, result)); 4924 4925 return (err); 4926 } 4927 4928 if (strcmp(pr_name, "_rxdma_intr_pkts") == 0) { 4929 err = nxge_param_rx_intr_pkts(nxgep, NULL, NULL, 4930 (char *)pr_val, 4931 (caddr_t)¶m_arr[param_rxdma_intr_pkts]); 4932 if (err) { 4933 NXGE_DEBUG_MSG((nxgep, NXGE_CTL, 4934 "<== nxge_set_priv_prop: " 4935 "unable to set (%s)", pr_name)); 4936 err = EINVAL; 4937 } else { 4938 err = 0; 4939 NXGE_DEBUG_MSG((nxgep, NXGE_CTL, 4940 "<== nxge_set_priv_prop: " 4941 "set (%s)", pr_name)); 4942 } 4943 4944 NXGE_DEBUG_MSG((nxgep, NXGE_CTL, 4945 "<== nxge_set_priv_prop: name %s (value %d)", 4946 pr_name, result)); 4947 4948 return (err); 4949 } 4950 4951 /* Classification */ 4952 if (strcmp(pr_name, "_class_opt_ipv4_tcp") == 0) { 4953 if (pr_val == NULL) { 4954 err = EINVAL; 4955 return (err); 4956 } 4957 (void) ddi_strtol(pr_val, (char **)NULL, 0, &result); 4958 4959 err = nxge_param_set_ip_opt(nxgep, NULL, 4960 NULL, (char *)pr_val, 4961 (caddr_t)¶m_arr[param_class_opt_ipv4_tcp]); 4962 4963 NXGE_DEBUG_MSG((nxgep, NXGE_CTL, 4964 "<== nxge_set_priv_prop: name %s (value 0x%x)", 4965 pr_name, result)); 4966 4967 return (err); 4968 } 4969 4970 if (strcmp(pr_name, "_class_opt_ipv4_udp") == 0) { 4971 if (pr_val == NULL) { 4972 err = EINVAL; 4973 return (err); 4974 } 4975 (void) ddi_strtol(pr_val, (char **)NULL, 0, &result); 4976 4977 err = nxge_param_set_ip_opt(nxgep, NULL, 4978 NULL, (char *)pr_val, 4979 (caddr_t)¶m_arr[param_class_opt_ipv4_udp]); 4980 4981 NXGE_DEBUG_MSG((nxgep, NXGE_CTL, 4982 "<== nxge_set_priv_prop: name %s (value 0x%x)", 4983 pr_name, result)); 4984 4985 return (err); 4986 } 4987 if (strcmp(pr_name, "_class_opt_ipv4_ah") == 0) { 4988 if (pr_val == NULL) { 4989 err = EINVAL; 4990 return (err); 4991 } 4992 (void) ddi_strtol(pr_val, (char **)NULL, 0, &result); 4993 4994 err = nxge_param_set_ip_opt(nxgep, NULL, 4995 NULL, (char *)pr_val, 4996 (caddr_t)¶m_arr[param_class_opt_ipv4_ah]); 4997 4998 NXGE_DEBUG_MSG((nxgep, NXGE_CTL, 4999 "<== nxge_set_priv_prop: name %s (value 0x%x)", 5000 pr_name, result)); 5001 5002 return (err); 5003 } 5004 if (strcmp(pr_name, "_class_opt_ipv4_sctp") == 0) { 5005 if (pr_val == NULL) { 5006 err = EINVAL; 5007 return (err); 5008 } 5009 (void) ddi_strtol(pr_val, (char **)NULL, 0, &result); 5010 5011 err = nxge_param_set_ip_opt(nxgep, NULL, 5012 NULL, (char *)pr_val, 5013 (caddr_t)¶m_arr[param_class_opt_ipv4_sctp]); 5014 5015 NXGE_DEBUG_MSG((nxgep, NXGE_CTL, 5016 "<== nxge_set_priv_prop: name %s (value 0x%x)", 5017 pr_name, result)); 5018 5019 return (err); 5020 } 5021 5022 if (strcmp(pr_name, "_class_opt_ipv6_tcp") == 0) { 5023 if (pr_val == NULL) { 5024 err = EINVAL; 5025 return (err); 5026 } 5027 (void) ddi_strtol(pr_val, (char **)NULL, 0, &result); 5028 5029 err = nxge_param_set_ip_opt(nxgep, NULL, 5030 NULL, (char *)pr_val, 5031 (caddr_t)¶m_arr[param_class_opt_ipv6_tcp]); 5032 5033 NXGE_DEBUG_MSG((nxgep, NXGE_CTL, 5034 "<== nxge_set_priv_prop: name %s (value 0x%x)", 5035 pr_name, result)); 5036 5037 return (err); 5038 } 5039 5040 if (strcmp(pr_name, "_class_opt_ipv6_udp") == 0) { 5041 if (pr_val == NULL) { 5042 err = EINVAL; 5043 return (err); 5044 } 5045 (void) ddi_strtol(pr_val, (char **)NULL, 0, &result); 5046 5047 err = nxge_param_set_ip_opt(nxgep, NULL, 5048 NULL, (char *)pr_val, 5049 (caddr_t)¶m_arr[param_class_opt_ipv6_udp]); 5050 5051 NXGE_DEBUG_MSG((nxgep, NXGE_CTL, 5052 "<== nxge_set_priv_prop: name %s (value 0x%x)", 5053 pr_name, result)); 5054 5055 return (err); 5056 } 5057 if (strcmp(pr_name, "_class_opt_ipv6_ah") == 0) { 5058 if (pr_val == NULL) { 5059 err = EINVAL; 5060 return (err); 5061 } 5062 (void) ddi_strtol(pr_val, (char **)NULL, 0, &result); 5063 5064 err = nxge_param_set_ip_opt(nxgep, NULL, 5065 NULL, (char *)pr_val, 5066 (caddr_t)¶m_arr[param_class_opt_ipv6_ah]); 5067 5068 NXGE_DEBUG_MSG((nxgep, NXGE_CTL, 5069 "<== nxge_set_priv_prop: name %s (value 0x%x)", 5070 pr_name, result)); 5071 5072 return (err); 5073 } 5074 if (strcmp(pr_name, "_class_opt_ipv6_sctp") == 0) { 5075 if (pr_val == NULL) { 5076 err = EINVAL; 5077 return (err); 5078 } 5079 (void) ddi_strtol(pr_val, (char **)NULL, 0, &result); 5080 5081 err = nxge_param_set_ip_opt(nxgep, NULL, 5082 NULL, (char *)pr_val, 5083 (caddr_t)¶m_arr[param_class_opt_ipv6_sctp]); 5084 5085 NXGE_DEBUG_MSG((nxgep, NXGE_CTL, 5086 "<== nxge_set_priv_prop: name %s (value 0x%x)", 5087 pr_name, result)); 5088 5089 return (err); 5090 } 5091 5092 if (strcmp(pr_name, "_soft_lso_enable") == 0) { 5093 if (pr_val == NULL) { 5094 NXGE_DEBUG_MSG((nxgep, NXGE_CTL, 5095 "==> nxge_set_priv_prop: name %s (null)", pr_name)); 5096 err = EINVAL; 5097 return (err); 5098 } 5099 5100 (void) ddi_strtol(pr_val, (char **)NULL, 0, &result); 5101 NXGE_DEBUG_MSG((nxgep, NXGE_CTL, 5102 "<== nxge_set_priv_prop: name %s " 5103 "(lso %d pr_val %s value %d)", 5104 pr_name, nxgep->soft_lso_enable, pr_val, result)); 5105 5106 if (result > 1 || result < 0) { 5107 err = EINVAL; 5108 } else { 5109 if (nxgep->soft_lso_enable == (uint32_t)result) { 5110 NXGE_DEBUG_MSG((nxgep, NXGE_CTL, 5111 "no change (%d %d)", 5112 nxgep->soft_lso_enable, result)); 5113 return (0); 5114 } 5115 } 5116 5117 nxgep->soft_lso_enable = (int)result; 5118 5119 NXGE_DEBUG_MSG((nxgep, NXGE_CTL, 5120 "<== nxge_set_priv_prop: name %s (value %d)", 5121 pr_name, result)); 5122 5123 return (err); 5124 } 5125 /* 5126 * Commands like "ndd -set /dev/nxge0 adv_10gfdx_cap 1" cause the 5127 * following code to be executed. 5128 */ 5129 if (strcmp(pr_name, "_adv_10gfdx_cap") == 0) { 5130 err = nxge_param_set_mac(nxgep, NULL, NULL, (char *)pr_val, 5131 (caddr_t)¶m_arr[param_anar_10gfdx]); 5132 return (err); 5133 } 5134 if (strcmp(pr_name, "_adv_pause_cap") == 0) { 5135 err = nxge_param_set_mac(nxgep, NULL, NULL, (char *)pr_val, 5136 (caddr_t)¶m_arr[param_anar_pause]); 5137 return (err); 5138 } 5139 5140 return (ENOTSUP); 5141 } 5142 5143 static int 5144 nxge_get_priv_prop(p_nxge_t nxgep, const char *pr_name, uint_t pr_valsize, 5145 void *pr_val) 5146 { 5147 p_nxge_param_t param_arr = nxgep->param_arr; 5148 char valstr[MAXNAMELEN]; 5149 int err = ENOTSUP; 5150 uint_t strsize; 5151 5152 NXGE_DEBUG_MSG((nxgep, NXGE_CTL, 5153 "==> nxge_get_priv_prop: property %s", pr_name)); 5154 5155 /* function number */ 5156 if (strcmp(pr_name, "_function_number") == 0) { 5157 (void) snprintf(valstr, sizeof (valstr), "%d", 5158 nxgep->function_num); 5159 NXGE_DEBUG_MSG((nxgep, NXGE_CTL, 5160 "==> nxge_get_priv_prop: name %s " 5161 "(value %d valstr %s)", 5162 pr_name, nxgep->function_num, valstr)); 5163 5164 err = 0; 5165 goto done; 5166 } 5167 5168 /* Neptune firmware version */ 5169 if (strcmp(pr_name, "_fw_version") == 0) { 5170 (void) snprintf(valstr, sizeof (valstr), "%s", 5171 nxgep->vpd_info.ver); 5172 NXGE_DEBUG_MSG((nxgep, NXGE_CTL, 5173 "==> nxge_get_priv_prop: name %s " 5174 "(value %d valstr %s)", 5175 pr_name, nxgep->vpd_info.ver, valstr)); 5176 5177 err = 0; 5178 goto done; 5179 } 5180 5181 /* port PHY mode */ 5182 if (strcmp(pr_name, "_port_mode") == 0) { 5183 switch (nxgep->mac.portmode) { 5184 case PORT_1G_COPPER: 5185 (void) snprintf(valstr, sizeof (valstr), "1G copper %s", 5186 nxgep->hot_swappable_phy ? 5187 "[Hot Swappable]" : ""); 5188 break; 5189 case PORT_1G_FIBER: 5190 (void) snprintf(valstr, sizeof (valstr), "1G fiber %s", 5191 nxgep->hot_swappable_phy ? 5192 "[hot swappable]" : ""); 5193 break; 5194 case PORT_10G_COPPER: 5195 (void) snprintf(valstr, sizeof (valstr), 5196 "10G copper %s", 5197 nxgep->hot_swappable_phy ? 5198 "[hot swappable]" : ""); 5199 break; 5200 case PORT_10G_FIBER: 5201 (void) snprintf(valstr, sizeof (valstr), "10G fiber %s", 5202 nxgep->hot_swappable_phy ? 5203 "[hot swappable]" : ""); 5204 break; 5205 case PORT_10G_SERDES: 5206 (void) snprintf(valstr, sizeof (valstr), 5207 "10G serdes %s", nxgep->hot_swappable_phy ? 5208 "[hot swappable]" : ""); 5209 break; 5210 case PORT_1G_SERDES: 5211 (void) snprintf(valstr, sizeof (valstr), "1G serdes %s", 5212 nxgep->hot_swappable_phy ? 5213 "[hot swappable]" : ""); 5214 break; 5215 case PORT_1G_TN1010: 5216 (void) snprintf(valstr, sizeof (valstr), 5217 "1G TN1010 copper %s", nxgep->hot_swappable_phy ? 5218 "[hot swappable]" : ""); 5219 break; 5220 case PORT_10G_TN1010: 5221 (void) snprintf(valstr, sizeof (valstr), 5222 "10G TN1010 copper %s", nxgep->hot_swappable_phy ? 5223 "[hot swappable]" : ""); 5224 break; 5225 case PORT_1G_RGMII_FIBER: 5226 (void) snprintf(valstr, sizeof (valstr), 5227 "1G rgmii fiber %s", nxgep->hot_swappable_phy ? 5228 "[hot swappable]" : ""); 5229 break; 5230 case PORT_HSP_MODE: 5231 (void) snprintf(valstr, sizeof (valstr), 5232 "phy not present[hot swappable]"); 5233 break; 5234 default: 5235 (void) snprintf(valstr, sizeof (valstr), "unknown %s", 5236 nxgep->hot_swappable_phy ? 5237 "[hot swappable]" : ""); 5238 break; 5239 } 5240 5241 NXGE_DEBUG_MSG((nxgep, NXGE_CTL, 5242 "==> nxge_get_priv_prop: name %s (value %s)", 5243 pr_name, valstr)); 5244 5245 err = 0; 5246 goto done; 5247 } 5248 5249 /* Hot swappable PHY */ 5250 if (strcmp(pr_name, "_hot_swap_phy") == 0) { 5251 (void) snprintf(valstr, sizeof (valstr), "%s", 5252 nxgep->hot_swappable_phy ? 5253 "yes" : "no"); 5254 5255 NXGE_DEBUG_MSG((nxgep, NXGE_CTL, 5256 "==> nxge_get_priv_prop: name %s " 5257 "(value %d valstr %s)", 5258 pr_name, nxgep->hot_swappable_phy, valstr)); 5259 5260 err = 0; 5261 goto done; 5262 } 5263 5264 5265 /* Receive Interrupt Blanking Parameters */ 5266 if (strcmp(pr_name, "_rxdma_intr_time") == 0) { 5267 err = 0; 5268 (void) snprintf(valstr, sizeof (valstr), "%d", 5269 nxgep->intr_timeout); 5270 NXGE_DEBUG_MSG((nxgep, NXGE_CTL, 5271 "==> nxge_get_priv_prop: name %s (value %d)", 5272 pr_name, 5273 (uint32_t)nxgep->intr_timeout)); 5274 goto done; 5275 } 5276 5277 if (strcmp(pr_name, "_rxdma_intr_pkts") == 0) { 5278 err = 0; 5279 (void) snprintf(valstr, sizeof (valstr), "%d", 5280 nxgep->intr_threshold); 5281 NXGE_DEBUG_MSG((nxgep, NXGE_CTL, 5282 "==> nxge_get_priv_prop: name %s (value %d)", 5283 pr_name, (uint32_t)nxgep->intr_threshold)); 5284 5285 goto done; 5286 } 5287 5288 /* Classification and Load Distribution Configuration */ 5289 if (strcmp(pr_name, "_class_opt_ipv4_tcp") == 0) { 5290 err = nxge_dld_get_ip_opt(nxgep, 5291 (caddr_t)¶m_arr[param_class_opt_ipv4_tcp]); 5292 5293 (void) snprintf(valstr, sizeof (valstr), "%x", 5294 (int)param_arr[param_class_opt_ipv4_tcp].value); 5295 5296 NXGE_DEBUG_MSG((nxgep, NXGE_CTL, 5297 "==> nxge_get_priv_prop: %s", valstr)); 5298 goto done; 5299 } 5300 5301 if (strcmp(pr_name, "_class_opt_ipv4_udp") == 0) { 5302 err = nxge_dld_get_ip_opt(nxgep, 5303 (caddr_t)¶m_arr[param_class_opt_ipv4_udp]); 5304 5305 (void) snprintf(valstr, sizeof (valstr), "%x", 5306 (int)param_arr[param_class_opt_ipv4_udp].value); 5307 5308 NXGE_DEBUG_MSG((nxgep, NXGE_CTL, 5309 "==> nxge_get_priv_prop: %s", valstr)); 5310 goto done; 5311 } 5312 if (strcmp(pr_name, "_class_opt_ipv4_ah") == 0) { 5313 err = nxge_dld_get_ip_opt(nxgep, 5314 (caddr_t)¶m_arr[param_class_opt_ipv4_ah]); 5315 5316 (void) snprintf(valstr, sizeof (valstr), "%x", 5317 (int)param_arr[param_class_opt_ipv4_ah].value); 5318 5319 NXGE_DEBUG_MSG((nxgep, NXGE_CTL, 5320 "==> nxge_get_priv_prop: %s", valstr)); 5321 goto done; 5322 } 5323 5324 if (strcmp(pr_name, "_class_opt_ipv4_sctp") == 0) { 5325 err = nxge_dld_get_ip_opt(nxgep, 5326 (caddr_t)¶m_arr[param_class_opt_ipv4_sctp]); 5327 5328 (void) snprintf(valstr, sizeof (valstr), "%x", 5329 (int)param_arr[param_class_opt_ipv4_sctp].value); 5330 5331 NXGE_DEBUG_MSG((nxgep, NXGE_CTL, 5332 "==> nxge_get_priv_prop: %s", valstr)); 5333 goto done; 5334 } 5335 5336 if (strcmp(pr_name, "_class_opt_ipv6_tcp") == 0) { 5337 err = nxge_dld_get_ip_opt(nxgep, 5338 (caddr_t)¶m_arr[param_class_opt_ipv6_tcp]); 5339 5340 (void) snprintf(valstr, sizeof (valstr), "%x", 5341 (int)param_arr[param_class_opt_ipv6_tcp].value); 5342 5343 NXGE_DEBUG_MSG((nxgep, NXGE_CTL, 5344 "==> nxge_get_priv_prop: %s", valstr)); 5345 goto done; 5346 } 5347 5348 if (strcmp(pr_name, "_class_opt_ipv6_udp") == 0) { 5349 err = nxge_dld_get_ip_opt(nxgep, 5350 (caddr_t)¶m_arr[param_class_opt_ipv6_udp]); 5351 5352 (void) snprintf(valstr, sizeof (valstr), "%x", 5353 (int)param_arr[param_class_opt_ipv6_udp].value); 5354 5355 NXGE_DEBUG_MSG((nxgep, NXGE_CTL, 5356 "==> nxge_get_priv_prop: %s", valstr)); 5357 goto done; 5358 } 5359 5360 if (strcmp(pr_name, "_class_opt_ipv6_ah") == 0) { 5361 err = nxge_dld_get_ip_opt(nxgep, 5362 (caddr_t)¶m_arr[param_class_opt_ipv6_ah]); 5363 5364 (void) snprintf(valstr, sizeof (valstr), "%x", 5365 (int)param_arr[param_class_opt_ipv6_ah].value); 5366 5367 NXGE_DEBUG_MSG((nxgep, NXGE_CTL, 5368 "==> nxge_get_priv_prop: %s", valstr)); 5369 goto done; 5370 } 5371 5372 if (strcmp(pr_name, "_class_opt_ipv6_sctp") == 0) { 5373 err = nxge_dld_get_ip_opt(nxgep, 5374 (caddr_t)¶m_arr[param_class_opt_ipv6_sctp]); 5375 5376 (void) snprintf(valstr, sizeof (valstr), "%x", 5377 (int)param_arr[param_class_opt_ipv6_sctp].value); 5378 5379 NXGE_DEBUG_MSG((nxgep, NXGE_CTL, 5380 "==> nxge_get_priv_prop: %s", valstr)); 5381 goto done; 5382 } 5383 5384 /* Software LSO */ 5385 if (strcmp(pr_name, "_soft_lso_enable") == 0) { 5386 (void) snprintf(valstr, sizeof (valstr), 5387 "%d", nxgep->soft_lso_enable); 5388 err = 0; 5389 NXGE_DEBUG_MSG((nxgep, NXGE_CTL, 5390 "==> nxge_get_priv_prop: name %s (value %d)", 5391 pr_name, nxgep->soft_lso_enable)); 5392 5393 goto done; 5394 } 5395 if (strcmp(pr_name, "_adv_10gfdx_cap") == 0) { 5396 err = 0; 5397 if (nxgep->param_arr[param_anar_10gfdx].value != 0) { 5398 (void) snprintf(valstr, sizeof (valstr), "%d", 1); 5399 goto done; 5400 } else { 5401 (void) snprintf(valstr, sizeof (valstr), "%d", 0); 5402 goto done; 5403 } 5404 } 5405 if (strcmp(pr_name, "_adv_pause_cap") == 0) { 5406 err = 0; 5407 if (nxgep->param_arr[param_anar_pause].value != 0) { 5408 (void) snprintf(valstr, sizeof (valstr), "%d", 1); 5409 goto done; 5410 } else { 5411 (void) snprintf(valstr, sizeof (valstr), "%d", 0); 5412 goto done; 5413 } 5414 } 5415 5416 done: 5417 if (err == 0) { 5418 strsize = (uint_t)strlen(valstr); 5419 if (pr_valsize < strsize) { 5420 err = ENOBUFS; 5421 } else { 5422 (void) strlcpy(pr_val, valstr, pr_valsize); 5423 } 5424 } 5425 5426 NXGE_DEBUG_MSG((nxgep, NXGE_CTL, 5427 "<== nxge_get_priv_prop: return %d", err)); 5428 return (err); 5429 } 5430 5431 /* 5432 * Module loading and removing entry points. 5433 */ 5434 5435 DDI_DEFINE_STREAM_OPS(nxge_dev_ops, nulldev, nulldev, nxge_attach, nxge_detach, 5436 nodev, NULL, D_MP, NULL, nxge_quiesce); 5437 5438 #define NXGE_DESC_VER "Sun NIU 10Gb Ethernet" 5439 5440 /* 5441 * Module linkage information for the kernel. 5442 */ 5443 static struct modldrv nxge_modldrv = { 5444 &mod_driverops, 5445 NXGE_DESC_VER, 5446 &nxge_dev_ops 5447 }; 5448 5449 static struct modlinkage modlinkage = { 5450 MODREV_1, { (void *) &nxge_modldrv, NULL } 5451 }; 5452 5453 int 5454 _init(void) 5455 { 5456 int status; 5457 5458 MUTEX_INIT(&nxgedebuglock, NULL, MUTEX_DRIVER, NULL); 5459 5460 NXGE_DEBUG_MSG((NULL, MOD_CTL, "==> _init")); 5461 5462 mac_init_ops(&nxge_dev_ops, "nxge"); 5463 5464 status = ddi_soft_state_init(&nxge_list, sizeof (nxge_t), 0); 5465 if (status != 0) { 5466 NXGE_ERROR_MSG((NULL, NXGE_ERR_CTL, 5467 "failed to init device soft state")); 5468 goto _init_exit; 5469 } 5470 5471 status = mod_install(&modlinkage); 5472 if (status != 0) { 5473 ddi_soft_state_fini(&nxge_list); 5474 NXGE_ERROR_MSG((NULL, NXGE_ERR_CTL, "Mod install failed")); 5475 goto _init_exit; 5476 } 5477 5478 MUTEX_INIT(&nxge_common_lock, NULL, MUTEX_DRIVER, NULL); 5479 5480 NXGE_DEBUG_MSG((NULL, MOD_CTL, "<== _init status = 0x%X", status)); 5481 return (status); 5482 5483 _init_exit: 5484 NXGE_DEBUG_MSG((NULL, MOD_CTL, "<== _init status = 0x%X", status)); 5485 MUTEX_DESTROY(&nxgedebuglock); 5486 return (status); 5487 } 5488 5489 int 5490 _fini(void) 5491 { 5492 int status; 5493 5494 NXGE_DEBUG_MSG((NULL, MOD_CTL, "==> _fini")); 5495 NXGE_DEBUG_MSG((NULL, MOD_CTL, "==> _fini: mod_remove")); 5496 5497 if (nxge_mblks_pending) 5498 return (EBUSY); 5499 5500 status = mod_remove(&modlinkage); 5501 if (status != DDI_SUCCESS) { 5502 NXGE_DEBUG_MSG((NULL, MOD_CTL, 5503 "Module removal failed 0x%08x", 5504 status)); 5505 goto _fini_exit; 5506 } 5507 5508 mac_fini_ops(&nxge_dev_ops); 5509 5510 ddi_soft_state_fini(&nxge_list); 5511 5512 NXGE_DEBUG_MSG((NULL, MOD_CTL, "<== _fini status = 0x%08x", status)); 5513 5514 MUTEX_DESTROY(&nxge_common_lock); 5515 MUTEX_DESTROY(&nxgedebuglock); 5516 return (status); 5517 5518 _fini_exit: 5519 NXGE_DEBUG_MSG((NULL, MOD_CTL, "<== _fini status = 0x%08x", status)); 5520 return (status); 5521 } 5522 5523 int 5524 _info(struct modinfo *modinfop) 5525 { 5526 int status; 5527 5528 NXGE_DEBUG_MSG((NULL, MOD_CTL, "==> _info")); 5529 status = mod_info(&modlinkage, modinfop); 5530 NXGE_DEBUG_MSG((NULL, MOD_CTL, " _info status = 0x%X", status)); 5531 5532 return (status); 5533 } 5534 5535 /*ARGSUSED*/ 5536 static int 5537 nxge_tx_ring_start(mac_ring_driver_t rdriver, uint64_t mr_gen_num) 5538 { 5539 p_nxge_ring_handle_t rhp = (p_nxge_ring_handle_t)rdriver; 5540 p_nxge_t nxgep = rhp->nxgep; 5541 uint32_t channel; 5542 p_tx_ring_t ring; 5543 5544 channel = nxgep->pt_config.hw_config.tdc.start + rhp->index; 5545 ring = nxgep->tx_rings->rings[channel]; 5546 5547 MUTEX_ENTER(&ring->lock); 5548 ASSERT(ring->tx_ring_handle == NULL); 5549 ring->tx_ring_handle = rhp->ring_handle; 5550 MUTEX_EXIT(&ring->lock); 5551 5552 return (0); 5553 } 5554 5555 static void 5556 nxge_tx_ring_stop(mac_ring_driver_t rdriver) 5557 { 5558 p_nxge_ring_handle_t rhp = (p_nxge_ring_handle_t)rdriver; 5559 p_nxge_t nxgep = rhp->nxgep; 5560 uint32_t channel; 5561 p_tx_ring_t ring; 5562 5563 channel = nxgep->pt_config.hw_config.tdc.start + rhp->index; 5564 ring = nxgep->tx_rings->rings[channel]; 5565 5566 MUTEX_ENTER(&ring->lock); 5567 ASSERT(ring->tx_ring_handle != NULL); 5568 ring->tx_ring_handle = (mac_ring_handle_t)NULL; 5569 MUTEX_EXIT(&ring->lock); 5570 } 5571 5572 int 5573 nxge_rx_ring_start(mac_ring_driver_t rdriver, uint64_t mr_gen_num) 5574 { 5575 p_nxge_ring_handle_t rhp = (p_nxge_ring_handle_t)rdriver; 5576 p_nxge_t nxgep = rhp->nxgep; 5577 uint32_t channel; 5578 p_rx_rcr_ring_t ring; 5579 int i; 5580 5581 channel = nxgep->pt_config.hw_config.start_rdc + rhp->index; 5582 ring = nxgep->rx_rcr_rings->rcr_rings[channel]; 5583 5584 MUTEX_ENTER(&ring->lock); 5585 5586 if (ring->started) { 5587 ASSERT(ring->started == B_FALSE); 5588 MUTEX_EXIT(&ring->lock); 5589 return (0); 5590 } 5591 5592 /* set rcr_ring */ 5593 for (i = 0; i < nxgep->ldgvp->maxldvs; i++) { 5594 if ((nxgep->ldgvp->ldvp[i].is_rxdma) && 5595 (nxgep->ldgvp->ldvp[i].channel == channel)) { 5596 ring->ldvp = &nxgep->ldgvp->ldvp[i]; 5597 ring->ldgp = nxgep->ldgvp->ldvp[i].ldgp; 5598 } 5599 } 5600 5601 ring->rcr_mac_handle = rhp->ring_handle; 5602 ring->rcr_gen_num = mr_gen_num; 5603 ring->started = B_TRUE; 5604 rhp->ring_gen_num = mr_gen_num; 5605 MUTEX_EXIT(&ring->lock); 5606 5607 return (0); 5608 } 5609 5610 static void 5611 nxge_rx_ring_stop(mac_ring_driver_t rdriver) 5612 { 5613 p_nxge_ring_handle_t rhp = (p_nxge_ring_handle_t)rdriver; 5614 p_nxge_t nxgep = rhp->nxgep; 5615 uint32_t channel; 5616 p_rx_rcr_ring_t ring; 5617 5618 channel = nxgep->pt_config.hw_config.start_rdc + rhp->index; 5619 ring = nxgep->rx_rcr_rings->rcr_rings[channel]; 5620 5621 MUTEX_ENTER(&ring->lock); 5622 ASSERT(ring->started == B_TRUE); 5623 ring->rcr_mac_handle = NULL; 5624 ring->ldvp = NULL; 5625 ring->ldgp = NULL; 5626 ring->started = B_FALSE; 5627 MUTEX_EXIT(&ring->lock); 5628 } 5629 5630 static int 5631 nxge_ring_get_htable_idx(p_nxge_t nxgep, mac_ring_type_t type, uint32_t channel) 5632 { 5633 int i; 5634 5635 #if defined(sun4v) 5636 if (isLDOMguest(nxgep)) { 5637 return (nxge_hio_get_dc_htable_idx(nxgep, 5638 (type == MAC_RING_TYPE_TX) ? VP_BOUND_TX : VP_BOUND_RX, 5639 channel)); 5640 } 5641 #endif 5642 5643 ASSERT(nxgep->ldgvp != NULL); 5644 5645 switch (type) { 5646 case MAC_RING_TYPE_TX: 5647 for (i = 0; i < nxgep->ldgvp->maxldvs; i++) { 5648 if ((nxgep->ldgvp->ldvp[i].is_txdma) && 5649 (nxgep->ldgvp->ldvp[i].channel == channel)) { 5650 return ((int) 5651 nxgep->ldgvp->ldvp[i].ldgp->htable_idx); 5652 } 5653 } 5654 break; 5655 5656 case MAC_RING_TYPE_RX: 5657 for (i = 0; i < nxgep->ldgvp->maxldvs; i++) { 5658 if ((nxgep->ldgvp->ldvp[i].is_rxdma) && 5659 (nxgep->ldgvp->ldvp[i].channel == channel)) { 5660 return ((int) 5661 nxgep->ldgvp->ldvp[i].ldgp->htable_idx); 5662 } 5663 } 5664 } 5665 5666 return (-1); 5667 } 5668 5669 /* 5670 * Callback funtion for MAC layer to register all rings. 5671 */ 5672 static void 5673 nxge_fill_ring(void *arg, mac_ring_type_t rtype, const int rg_index, 5674 const int index, mac_ring_info_t *infop, mac_ring_handle_t rh) 5675 { 5676 p_nxge_t nxgep = (p_nxge_t)arg; 5677 p_nxge_hw_pt_cfg_t p_cfgp = &nxgep->pt_config.hw_config; 5678 p_nxge_intr_t intrp; 5679 uint32_t channel; 5680 int htable_idx; 5681 p_nxge_ring_handle_t rhandlep; 5682 5683 ASSERT(nxgep != NULL); 5684 ASSERT(p_cfgp != NULL); 5685 ASSERT(infop != NULL); 5686 5687 NXGE_DEBUG_MSG((nxgep, DDI_CTL, 5688 "==> nxge_fill_ring 0x%x index %d", rtype, index)); 5689 5690 5691 switch (rtype) { 5692 case MAC_RING_TYPE_TX: { 5693 mac_intr_t *mintr = &infop->mri_intr; 5694 5695 NXGE_DEBUG_MSG((nxgep, TX_CTL, 5696 "==> nxge_fill_ring (TX) 0x%x index %d ntdcs %d", 5697 rtype, index, p_cfgp->tdc.count)); 5698 5699 ASSERT((index >= 0) && (index < p_cfgp->tdc.count)); 5700 rhandlep = &nxgep->tx_ring_handles[index]; 5701 rhandlep->nxgep = nxgep; 5702 rhandlep->index = index; 5703 rhandlep->ring_handle = rh; 5704 5705 channel = nxgep->pt_config.hw_config.tdc.start + index; 5706 rhandlep->channel = channel; 5707 intrp = (p_nxge_intr_t)&nxgep->nxge_intr_type; 5708 htable_idx = nxge_ring_get_htable_idx(nxgep, rtype, 5709 channel); 5710 if (htable_idx >= 0) 5711 mintr->mi_ddi_handle = intrp->htable[htable_idx]; 5712 else 5713 mintr->mi_ddi_handle = NULL; 5714 5715 infop->mri_driver = (mac_ring_driver_t)rhandlep; 5716 infop->mri_start = nxge_tx_ring_start; 5717 infop->mri_stop = nxge_tx_ring_stop; 5718 infop->mri_tx = nxge_tx_ring_send; 5719 infop->mri_stat = nxge_tx_ring_stat; 5720 infop->mri_flags = MAC_RING_TX_SERIALIZE; 5721 break; 5722 } 5723 5724 case MAC_RING_TYPE_RX: { 5725 mac_intr_t nxge_mac_intr; 5726 int nxge_rindex; 5727 p_nxge_intr_t intrp; 5728 5729 intrp = (p_nxge_intr_t)&nxgep->nxge_intr_type; 5730 5731 NXGE_DEBUG_MSG((nxgep, RX_CTL, 5732 "==> nxge_fill_ring (RX) 0x%x index %d nrdcs %d", 5733 rtype, index, p_cfgp->max_rdcs)); 5734 5735 /* 5736 * 'index' is the ring index within the group. 5737 * Find the ring index in the nxge instance. 5738 */ 5739 nxge_rindex = nxge_get_rxring_index(nxgep, rg_index, index); 5740 channel = nxgep->pt_config.hw_config.start_rdc + index; 5741 intrp = (p_nxge_intr_t)&nxgep->nxge_intr_type; 5742 5743 ASSERT((nxge_rindex >= 0) && (nxge_rindex < p_cfgp->max_rdcs)); 5744 rhandlep = &nxgep->rx_ring_handles[nxge_rindex]; 5745 rhandlep->nxgep = nxgep; 5746 rhandlep->index = nxge_rindex; 5747 rhandlep->ring_handle = rh; 5748 rhandlep->channel = channel; 5749 5750 /* 5751 * Entrypoint to enable interrupt (disable poll) and 5752 * disable interrupt (enable poll). 5753 */ 5754 bzero(&nxge_mac_intr, sizeof (nxge_mac_intr)); 5755 nxge_mac_intr.mi_handle = (mac_intr_handle_t)rhandlep; 5756 nxge_mac_intr.mi_enable = (mac_intr_enable_t)nxge_disable_poll; 5757 nxge_mac_intr.mi_disable = (mac_intr_disable_t)nxge_enable_poll; 5758 5759 htable_idx = nxge_ring_get_htable_idx(nxgep, rtype, 5760 channel); 5761 if (htable_idx >= 0) 5762 nxge_mac_intr.mi_ddi_handle = intrp->htable[htable_idx]; 5763 else 5764 nxge_mac_intr.mi_ddi_handle = NULL; 5765 5766 infop->mri_driver = (mac_ring_driver_t)rhandlep; 5767 infop->mri_start = nxge_rx_ring_start; 5768 infop->mri_stop = nxge_rx_ring_stop; 5769 infop->mri_intr = nxge_mac_intr; 5770 infop->mri_poll = nxge_rx_poll; 5771 infop->mri_stat = nxge_rx_ring_stat; 5772 infop->mri_flags = MAC_RING_RX_ENQUEUE; 5773 break; 5774 } 5775 5776 default: 5777 break; 5778 } 5779 5780 NXGE_DEBUG_MSG((nxgep, DDI_CTL, "<== nxge_fill_ring 0x%x", rtype)); 5781 } 5782 5783 static void 5784 nxge_group_add_ring(mac_group_driver_t gh, mac_ring_driver_t rh, 5785 mac_ring_type_t type) 5786 { 5787 nxge_ring_group_t *rgroup = (nxge_ring_group_t *)gh; 5788 nxge_ring_handle_t *rhandle = (nxge_ring_handle_t *)rh; 5789 nxge_t *nxge; 5790 nxge_grp_t *grp; 5791 nxge_rdc_grp_t *rdc_grp; 5792 uint16_t channel; /* device-wise ring id */ 5793 int dev_gindex; 5794 int rv; 5795 5796 nxge = rgroup->nxgep; 5797 5798 switch (type) { 5799 case MAC_RING_TYPE_TX: 5800 /* 5801 * nxge_grp_dc_add takes a channel number which is a 5802 * "devise" ring ID. 5803 */ 5804 channel = nxge->pt_config.hw_config.tdc.start + rhandle->index; 5805 5806 /* 5807 * Remove the ring from the default group 5808 */ 5809 if (rgroup->gindex != 0) { 5810 (void) nxge_grp_dc_remove(nxge, VP_BOUND_TX, channel); 5811 } 5812 5813 /* 5814 * nxge->tx_set.group[] is an array of groups indexed by 5815 * a "port" group ID. 5816 */ 5817 grp = nxge->tx_set.group[rgroup->gindex]; 5818 rv = nxge_grp_dc_add(nxge, grp, VP_BOUND_TX, channel); 5819 if (rv != 0) { 5820 NXGE_ERROR_MSG((nxge, NXGE_ERR_CTL, 5821 "nxge_group_add_ring: nxge_grp_dc_add failed")); 5822 } 5823 break; 5824 5825 case MAC_RING_TYPE_RX: 5826 /* 5827 * nxge->rx_set.group[] is an array of groups indexed by 5828 * a "port" group ID. 5829 */ 5830 grp = nxge->rx_set.group[rgroup->gindex]; 5831 5832 dev_gindex = nxge->pt_config.hw_config.def_mac_rxdma_grpid + 5833 rgroup->gindex; 5834 rdc_grp = &nxge->pt_config.rdc_grps[dev_gindex]; 5835 5836 /* 5837 * nxge_grp_dc_add takes a channel number which is a 5838 * "devise" ring ID. 5839 */ 5840 channel = nxge->pt_config.hw_config.start_rdc + rhandle->index; 5841 rv = nxge_grp_dc_add(nxge, grp, VP_BOUND_RX, channel); 5842 if (rv != 0) { 5843 NXGE_ERROR_MSG((nxge, NXGE_ERR_CTL, 5844 "nxge_group_add_ring: nxge_grp_dc_add failed")); 5845 } 5846 5847 rdc_grp->map |= (1 << channel); 5848 rdc_grp->max_rdcs++; 5849 5850 (void) nxge_init_fzc_rdc_tbl(nxge, rdc_grp, rgroup->rdctbl); 5851 break; 5852 } 5853 } 5854 5855 static void 5856 nxge_group_rem_ring(mac_group_driver_t gh, mac_ring_driver_t rh, 5857 mac_ring_type_t type) 5858 { 5859 nxge_ring_group_t *rgroup = (nxge_ring_group_t *)gh; 5860 nxge_ring_handle_t *rhandle = (nxge_ring_handle_t *)rh; 5861 nxge_t *nxge; 5862 uint16_t channel; /* device-wise ring id */ 5863 nxge_rdc_grp_t *rdc_grp; 5864 int dev_gindex; 5865 5866 nxge = rgroup->nxgep; 5867 5868 switch (type) { 5869 case MAC_RING_TYPE_TX: 5870 dev_gindex = nxge->pt_config.hw_config.def_mac_txdma_grpid + 5871 rgroup->gindex; 5872 channel = nxge->pt_config.hw_config.tdc.start + rhandle->index; 5873 nxge_grp_dc_remove(nxge, VP_BOUND_TX, channel); 5874 5875 /* 5876 * Add the ring back to the default group 5877 */ 5878 if (rgroup->gindex != 0) { 5879 nxge_grp_t *grp; 5880 grp = nxge->tx_set.group[0]; 5881 (void) nxge_grp_dc_add(nxge, grp, VP_BOUND_TX, channel); 5882 } 5883 break; 5884 5885 case MAC_RING_TYPE_RX: 5886 dev_gindex = nxge->pt_config.hw_config.def_mac_rxdma_grpid + 5887 rgroup->gindex; 5888 rdc_grp = &nxge->pt_config.rdc_grps[dev_gindex]; 5889 channel = rdc_grp->start_rdc + rhandle->index; 5890 nxge_grp_dc_remove(nxge, VP_BOUND_RX, channel); 5891 5892 rdc_grp->map &= ~(1 << channel); 5893 rdc_grp->max_rdcs--; 5894 5895 (void) nxge_init_fzc_rdc_tbl(nxge, rdc_grp, rgroup->rdctbl); 5896 break; 5897 } 5898 } 5899 5900 5901 /*ARGSUSED*/ 5902 static nxge_status_t 5903 nxge_add_intrs(p_nxge_t nxgep) 5904 { 5905 5906 int intr_types; 5907 int type = 0; 5908 int ddi_status = DDI_SUCCESS; 5909 nxge_status_t status = NXGE_OK; 5910 5911 NXGE_DEBUG_MSG((nxgep, DDI_CTL, "==> nxge_add_intrs")); 5912 5913 nxgep->nxge_intr_type.intr_registered = B_FALSE; 5914 nxgep->nxge_intr_type.intr_enabled = B_FALSE; 5915 nxgep->nxge_intr_type.msi_intx_cnt = 0; 5916 nxgep->nxge_intr_type.intr_added = 0; 5917 nxgep->nxge_intr_type.niu_msi_enable = B_FALSE; 5918 nxgep->nxge_intr_type.intr_type = 0; 5919 5920 if (nxgep->niu_type == N2_NIU) { 5921 nxgep->nxge_intr_type.niu_msi_enable = B_TRUE; 5922 } else if (nxge_msi_enable) { 5923 nxgep->nxge_intr_type.niu_msi_enable = B_TRUE; 5924 } 5925 5926 /* Get the supported interrupt types */ 5927 if ((ddi_status = ddi_intr_get_supported_types(nxgep->dip, &intr_types)) 5928 != DDI_SUCCESS) { 5929 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, "<== nxge_add_intrs: " 5930 "ddi_intr_get_supported_types failed: status 0x%08x", 5931 ddi_status)); 5932 return (NXGE_ERROR | NXGE_DDI_FAILED); 5933 } 5934 nxgep->nxge_intr_type.intr_types = intr_types; 5935 5936 NXGE_DEBUG_MSG((nxgep, DDI_CTL, "==> nxge_add_intrs: " 5937 "ddi_intr_get_supported_types: 0x%08x", intr_types)); 5938 5939 /* 5940 * Solaris MSIX is not supported yet. use MSI for now. 5941 * nxge_msi_enable (1): 5942 * 1 - MSI 2 - MSI-X others - FIXED 5943 */ 5944 switch (nxge_msi_enable) { 5945 default: 5946 type = DDI_INTR_TYPE_FIXED; 5947 NXGE_DEBUG_MSG((nxgep, INT_CTL, "==> nxge_add_intrs: " 5948 "use fixed (intx emulation) type %08x", 5949 type)); 5950 break; 5951 5952 case 2: 5953 NXGE_DEBUG_MSG((nxgep, INT_CTL, "==> nxge_add_intrs: " 5954 "ddi_intr_get_supported_types: 0x%08x", intr_types)); 5955 if (intr_types & DDI_INTR_TYPE_MSIX) { 5956 type = DDI_INTR_TYPE_MSIX; 5957 NXGE_DEBUG_MSG((nxgep, DDI_CTL, "==> nxge_add_intrs: " 5958 "ddi_intr_get_supported_types: MSIX 0x%08x", 5959 type)); 5960 } else if (intr_types & DDI_INTR_TYPE_MSI) { 5961 type = DDI_INTR_TYPE_MSI; 5962 NXGE_DEBUG_MSG((nxgep, DDI_CTL, "==> nxge_add_intrs: " 5963 "ddi_intr_get_supported_types: MSI 0x%08x", 5964 type)); 5965 } else if (intr_types & DDI_INTR_TYPE_FIXED) { 5966 type = DDI_INTR_TYPE_FIXED; 5967 NXGE_DEBUG_MSG((nxgep, INT_CTL, "==> nxge_add_intrs: " 5968 "ddi_intr_get_supported_types: MSXED0x%08x", 5969 type)); 5970 } 5971 break; 5972 5973 case 1: 5974 if (intr_types & DDI_INTR_TYPE_MSI) { 5975 type = DDI_INTR_TYPE_MSI; 5976 NXGE_DEBUG_MSG((nxgep, INT_CTL, "==> nxge_add_intrs: " 5977 "ddi_intr_get_supported_types: MSI 0x%08x", 5978 type)); 5979 } else if (intr_types & DDI_INTR_TYPE_MSIX) { 5980 type = DDI_INTR_TYPE_MSIX; 5981 NXGE_DEBUG_MSG((nxgep, DDI_CTL, "==> nxge_add_intrs: " 5982 "ddi_intr_get_supported_types: MSIX 0x%08x", 5983 type)); 5984 } else if (intr_types & DDI_INTR_TYPE_FIXED) { 5985 type = DDI_INTR_TYPE_FIXED; 5986 NXGE_DEBUG_MSG((nxgep, DDI_CTL, "==> nxge_add_intrs: " 5987 "ddi_intr_get_supported_types: MSXED0x%08x", 5988 type)); 5989 } 5990 } 5991 5992 nxgep->nxge_intr_type.intr_type = type; 5993 if ((type == DDI_INTR_TYPE_MSIX || type == DDI_INTR_TYPE_MSI || 5994 type == DDI_INTR_TYPE_FIXED) && 5995 nxgep->nxge_intr_type.niu_msi_enable) { 5996 if ((status = nxge_add_intrs_adv(nxgep)) != DDI_SUCCESS) { 5997 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, 5998 " nxge_add_intrs: " 5999 " nxge_add_intrs_adv failed: status 0x%08x", 6000 status)); 6001 return (status); 6002 } else { 6003 NXGE_DEBUG_MSG((nxgep, DDI_CTL, "==> nxge_add_intrs: " 6004 "interrupts registered : type %d", type)); 6005 nxgep->nxge_intr_type.intr_registered = B_TRUE; 6006 6007 NXGE_DEBUG_MSG((nxgep, DDI_CTL, 6008 "\nAdded advanced nxge add_intr_adv " 6009 "intr type 0x%x\n", type)); 6010 6011 return (status); 6012 } 6013 } 6014 6015 if (!nxgep->nxge_intr_type.intr_registered) { 6016 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, "==> nxge_add_intrs: " 6017 "failed to register interrupts")); 6018 return (NXGE_ERROR | NXGE_DDI_FAILED); 6019 } 6020 6021 NXGE_DEBUG_MSG((nxgep, DDI_CTL, "<== nxge_add_intrs")); 6022 return (status); 6023 } 6024 6025 static nxge_status_t 6026 nxge_add_intrs_adv(p_nxge_t nxgep) 6027 { 6028 int intr_type; 6029 p_nxge_intr_t intrp; 6030 6031 NXGE_DEBUG_MSG((nxgep, DDI_CTL, "==> nxge_add_intrs_adv")); 6032 6033 intrp = (p_nxge_intr_t)&nxgep->nxge_intr_type; 6034 intr_type = intrp->intr_type; 6035 NXGE_DEBUG_MSG((nxgep, DDI_CTL, "==> nxge_add_intrs_adv: type 0x%x", 6036 intr_type)); 6037 6038 switch (intr_type) { 6039 case DDI_INTR_TYPE_MSI: /* 0x2 */ 6040 case DDI_INTR_TYPE_MSIX: /* 0x4 */ 6041 return (nxge_add_intrs_adv_type(nxgep, intr_type)); 6042 6043 case DDI_INTR_TYPE_FIXED: /* 0x1 */ 6044 return (nxge_add_intrs_adv_type_fix(nxgep, intr_type)); 6045 6046 default: 6047 return (NXGE_ERROR); 6048 } 6049 } 6050 6051 6052 /*ARGSUSED*/ 6053 static nxge_status_t 6054 nxge_add_intrs_adv_type(p_nxge_t nxgep, uint32_t int_type) 6055 { 6056 dev_info_t *dip = nxgep->dip; 6057 p_nxge_ldg_t ldgp; 6058 p_nxge_intr_t intrp; 6059 uint_t *inthandler; 6060 void *arg1, *arg2; 6061 int behavior; 6062 int nintrs, navail, nrequest; 6063 int nactual, nrequired; 6064 int inum = 0; 6065 int x, y; 6066 int ddi_status = DDI_SUCCESS; 6067 nxge_status_t status = NXGE_OK; 6068 6069 NXGE_DEBUG_MSG((nxgep, INT_CTL, "==> nxge_add_intrs_adv_type")); 6070 intrp = (p_nxge_intr_t)&nxgep->nxge_intr_type; 6071 intrp->start_inum = 0; 6072 6073 ddi_status = ddi_intr_get_nintrs(dip, int_type, &nintrs); 6074 if ((ddi_status != DDI_SUCCESS) || (nintrs == 0)) { 6075 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, 6076 "ddi_intr_get_nintrs() failed, status: 0x%x%, " 6077 "nintrs: %d", ddi_status, nintrs)); 6078 return (NXGE_ERROR | NXGE_DDI_FAILED); 6079 } 6080 6081 ddi_status = ddi_intr_get_navail(dip, int_type, &navail); 6082 if ((ddi_status != DDI_SUCCESS) || (navail == 0)) { 6083 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, 6084 "ddi_intr_get_navail() failed, status: 0x%x%, " 6085 "nintrs: %d", ddi_status, navail)); 6086 return (NXGE_ERROR | NXGE_DDI_FAILED); 6087 } 6088 6089 NXGE_DEBUG_MSG((nxgep, INT_CTL, 6090 "ddi_intr_get_navail() returned: nintrs %d, navail %d", 6091 nintrs, navail)); 6092 6093 /* PSARC/2007/453 MSI-X interrupt limit override */ 6094 if (int_type == DDI_INTR_TYPE_MSIX) { 6095 nrequest = nxge_create_msi_property(nxgep); 6096 if (nrequest < navail) { 6097 navail = nrequest; 6098 NXGE_DEBUG_MSG((nxgep, INT_CTL, 6099 "nxge_add_intrs_adv_type: nintrs %d " 6100 "navail %d (nrequest %d)", 6101 nintrs, navail, nrequest)); 6102 } 6103 } 6104 6105 if (int_type == DDI_INTR_TYPE_MSI && !ISP2(navail)) { 6106 /* MSI must be power of 2 */ 6107 if ((navail & 16) == 16) { 6108 navail = 16; 6109 } else if ((navail & 8) == 8) { 6110 navail = 8; 6111 } else if ((navail & 4) == 4) { 6112 navail = 4; 6113 } else if ((navail & 2) == 2) { 6114 navail = 2; 6115 } else { 6116 navail = 1; 6117 } 6118 NXGE_DEBUG_MSG((nxgep, INT_CTL, 6119 "ddi_intr_get_navail(): (msi power of 2) nintrs %d, " 6120 "navail %d", nintrs, navail)); 6121 } 6122 6123 behavior = ((int_type == DDI_INTR_TYPE_FIXED) ? DDI_INTR_ALLOC_STRICT : 6124 DDI_INTR_ALLOC_NORMAL); 6125 intrp->intr_size = navail * sizeof (ddi_intr_handle_t); 6126 intrp->htable = kmem_alloc(intrp->intr_size, KM_SLEEP); 6127 ddi_status = ddi_intr_alloc(dip, intrp->htable, int_type, inum, 6128 navail, &nactual, behavior); 6129 if (ddi_status != DDI_SUCCESS || nactual == 0) { 6130 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, 6131 " ddi_intr_alloc() failed: %d", 6132 ddi_status)); 6133 kmem_free(intrp->htable, intrp->intr_size); 6134 return (NXGE_ERROR | NXGE_DDI_FAILED); 6135 } 6136 6137 if ((ddi_status = ddi_intr_get_pri(intrp->htable[0], 6138 (uint_t *)&intrp->pri)) != DDI_SUCCESS) { 6139 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, 6140 " ddi_intr_get_pri() failed: %d", 6141 ddi_status)); 6142 /* Free already allocated interrupts */ 6143 for (y = 0; y < nactual; y++) { 6144 (void) ddi_intr_free(intrp->htable[y]); 6145 } 6146 6147 kmem_free(intrp->htable, intrp->intr_size); 6148 return (NXGE_ERROR | NXGE_DDI_FAILED); 6149 } 6150 6151 nrequired = 0; 6152 switch (nxgep->niu_type) { 6153 default: 6154 status = nxge_ldgv_init(nxgep, &nactual, &nrequired); 6155 break; 6156 6157 case N2_NIU: 6158 status = nxge_ldgv_init_n2(nxgep, &nactual, &nrequired); 6159 break; 6160 } 6161 6162 if (status != NXGE_OK) { 6163 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, 6164 "nxge_add_intrs_adv_typ:nxge_ldgv_init " 6165 "failed: 0x%x", status)); 6166 /* Free already allocated interrupts */ 6167 for (y = 0; y < nactual; y++) { 6168 (void) ddi_intr_free(intrp->htable[y]); 6169 } 6170 6171 kmem_free(intrp->htable, intrp->intr_size); 6172 return (status); 6173 } 6174 6175 ldgp = nxgep->ldgvp->ldgp; 6176 for (x = 0; x < nrequired; x++, ldgp++) { 6177 ldgp->vector = (uint8_t)x; 6178 ldgp->intdata = SID_DATA(ldgp->func, x); 6179 arg1 = ldgp->ldvp; 6180 arg2 = nxgep; 6181 if (ldgp->nldvs == 1) { 6182 inthandler = (uint_t *)ldgp->ldvp->ldv_intr_handler; 6183 NXGE_DEBUG_MSG((nxgep, INT_CTL, 6184 "nxge_add_intrs_adv_type: " 6185 "arg1 0x%x arg2 0x%x: " 6186 "1-1 int handler (entry %d intdata 0x%x)\n", 6187 arg1, arg2, 6188 x, ldgp->intdata)); 6189 } else if (ldgp->nldvs > 1) { 6190 inthandler = (uint_t *)ldgp->sys_intr_handler; 6191 NXGE_DEBUG_MSG((nxgep, INT_CTL, 6192 "nxge_add_intrs_adv_type: " 6193 "arg1 0x%x arg2 0x%x: " 6194 "nldevs %d int handler " 6195 "(entry %d intdata 0x%x)\n", 6196 arg1, arg2, 6197 ldgp->nldvs, x, ldgp->intdata)); 6198 } 6199 6200 NXGE_DEBUG_MSG((nxgep, INT_CTL, 6201 "==> nxge_add_intrs_adv_type: ddi_add_intr(inum) #%d " 6202 "htable 0x%llx", x, intrp->htable[x])); 6203 6204 if ((ddi_status = ddi_intr_add_handler(intrp->htable[x], 6205 (ddi_intr_handler_t *)inthandler, arg1, arg2)) 6206 != DDI_SUCCESS) { 6207 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, 6208 "==> nxge_add_intrs_adv_type: failed #%d " 6209 "status 0x%x", x, ddi_status)); 6210 for (y = 0; y < intrp->intr_added; y++) { 6211 (void) ddi_intr_remove_handler( 6212 intrp->htable[y]); 6213 } 6214 /* Free already allocated intr */ 6215 for (y = 0; y < nactual; y++) { 6216 (void) ddi_intr_free(intrp->htable[y]); 6217 } 6218 kmem_free(intrp->htable, intrp->intr_size); 6219 6220 (void) nxge_ldgv_uninit(nxgep); 6221 6222 return (NXGE_ERROR | NXGE_DDI_FAILED); 6223 } 6224 6225 ldgp->htable_idx = x; 6226 intrp->intr_added++; 6227 } 6228 6229 intrp->msi_intx_cnt = nactual; 6230 6231 NXGE_DEBUG_MSG((nxgep, DDI_CTL, 6232 "Requested: %d, Allowed: %d msi_intx_cnt %d intr_added %d", 6233 navail, nactual, 6234 intrp->msi_intx_cnt, 6235 intrp->intr_added)); 6236 6237 (void) ddi_intr_get_cap(intrp->htable[0], &intrp->intr_cap); 6238 6239 (void) nxge_intr_ldgv_init(nxgep); 6240 6241 NXGE_DEBUG_MSG((nxgep, INT_CTL, "<== nxge_add_intrs_adv_type")); 6242 6243 return (status); 6244 } 6245 6246 /*ARGSUSED*/ 6247 static nxge_status_t 6248 nxge_add_intrs_adv_type_fix(p_nxge_t nxgep, uint32_t int_type) 6249 { 6250 dev_info_t *dip = nxgep->dip; 6251 p_nxge_ldg_t ldgp; 6252 p_nxge_intr_t intrp; 6253 uint_t *inthandler; 6254 void *arg1, *arg2; 6255 int behavior; 6256 int nintrs, navail; 6257 int nactual, nrequired; 6258 int inum = 0; 6259 int x, y; 6260 int ddi_status = DDI_SUCCESS; 6261 nxge_status_t status = NXGE_OK; 6262 6263 NXGE_DEBUG_MSG((nxgep, INT_CTL, "==> nxge_add_intrs_adv_type_fix")); 6264 intrp = (p_nxge_intr_t)&nxgep->nxge_intr_type; 6265 intrp->start_inum = 0; 6266 6267 ddi_status = ddi_intr_get_nintrs(dip, int_type, &nintrs); 6268 if ((ddi_status != DDI_SUCCESS) || (nintrs == 0)) { 6269 NXGE_DEBUG_MSG((nxgep, INT_CTL, 6270 "ddi_intr_get_nintrs() failed, status: 0x%x%, " 6271 "nintrs: %d", status, nintrs)); 6272 return (NXGE_ERROR | NXGE_DDI_FAILED); 6273 } 6274 6275 ddi_status = ddi_intr_get_navail(dip, int_type, &navail); 6276 if ((ddi_status != DDI_SUCCESS) || (navail == 0)) { 6277 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, 6278 "ddi_intr_get_navail() failed, status: 0x%x%, " 6279 "nintrs: %d", ddi_status, navail)); 6280 return (NXGE_ERROR | NXGE_DDI_FAILED); 6281 } 6282 6283 NXGE_DEBUG_MSG((nxgep, INT_CTL, 6284 "ddi_intr_get_navail() returned: nintrs %d, naavail %d", 6285 nintrs, navail)); 6286 6287 behavior = ((int_type == DDI_INTR_TYPE_FIXED) ? DDI_INTR_ALLOC_STRICT : 6288 DDI_INTR_ALLOC_NORMAL); 6289 intrp->intr_size = navail * sizeof (ddi_intr_handle_t); 6290 intrp->htable = kmem_alloc(intrp->intr_size, KM_SLEEP); 6291 ddi_status = ddi_intr_alloc(dip, intrp->htable, int_type, inum, 6292 navail, &nactual, behavior); 6293 if (ddi_status != DDI_SUCCESS || nactual == 0) { 6294 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, 6295 " ddi_intr_alloc() failed: %d", 6296 ddi_status)); 6297 kmem_free(intrp->htable, intrp->intr_size); 6298 return (NXGE_ERROR | NXGE_DDI_FAILED); 6299 } 6300 6301 if ((ddi_status = ddi_intr_get_pri(intrp->htable[0], 6302 (uint_t *)&intrp->pri)) != DDI_SUCCESS) { 6303 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, 6304 " ddi_intr_get_pri() failed: %d", 6305 ddi_status)); 6306 /* Free already allocated interrupts */ 6307 for (y = 0; y < nactual; y++) { 6308 (void) ddi_intr_free(intrp->htable[y]); 6309 } 6310 6311 kmem_free(intrp->htable, intrp->intr_size); 6312 return (NXGE_ERROR | NXGE_DDI_FAILED); 6313 } 6314 6315 nrequired = 0; 6316 switch (nxgep->niu_type) { 6317 default: 6318 status = nxge_ldgv_init(nxgep, &nactual, &nrequired); 6319 break; 6320 6321 case N2_NIU: 6322 status = nxge_ldgv_init_n2(nxgep, &nactual, &nrequired); 6323 break; 6324 } 6325 6326 if (status != NXGE_OK) { 6327 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, 6328 "nxge_add_intrs_adv_type_fix:nxge_ldgv_init " 6329 "failed: 0x%x", status)); 6330 /* Free already allocated interrupts */ 6331 for (y = 0; y < nactual; y++) { 6332 (void) ddi_intr_free(intrp->htable[y]); 6333 } 6334 6335 kmem_free(intrp->htable, intrp->intr_size); 6336 return (status); 6337 } 6338 6339 ldgp = nxgep->ldgvp->ldgp; 6340 for (x = 0; x < nrequired; x++, ldgp++) { 6341 ldgp->vector = (uint8_t)x; 6342 if (nxgep->niu_type != N2_NIU) { 6343 ldgp->intdata = SID_DATA(ldgp->func, x); 6344 } 6345 6346 arg1 = ldgp->ldvp; 6347 arg2 = nxgep; 6348 if (ldgp->nldvs == 1) { 6349 inthandler = (uint_t *)ldgp->ldvp->ldv_intr_handler; 6350 NXGE_DEBUG_MSG((nxgep, INT_CTL, 6351 "nxge_add_intrs_adv_type_fix: " 6352 "1-1 int handler(%d) ldg %d ldv %d " 6353 "arg1 $%p arg2 $%p\n", 6354 x, ldgp->ldg, ldgp->ldvp->ldv, 6355 arg1, arg2)); 6356 } else if (ldgp->nldvs > 1) { 6357 inthandler = (uint_t *)ldgp->sys_intr_handler; 6358 NXGE_DEBUG_MSG((nxgep, INT_CTL, 6359 "nxge_add_intrs_adv_type_fix: " 6360 "shared ldv %d int handler(%d) ldv %d ldg %d" 6361 "arg1 0x%016llx arg2 0x%016llx\n", 6362 x, ldgp->nldvs, ldgp->ldg, ldgp->ldvp->ldv, 6363 arg1, arg2)); 6364 } 6365 6366 if ((ddi_status = ddi_intr_add_handler(intrp->htable[x], 6367 (ddi_intr_handler_t *)inthandler, arg1, arg2)) 6368 != DDI_SUCCESS) { 6369 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, 6370 "==> nxge_add_intrs_adv_type_fix: failed #%d " 6371 "status 0x%x", x, ddi_status)); 6372 for (y = 0; y < intrp->intr_added; y++) { 6373 (void) ddi_intr_remove_handler( 6374 intrp->htable[y]); 6375 } 6376 for (y = 0; y < nactual; y++) { 6377 (void) ddi_intr_free(intrp->htable[y]); 6378 } 6379 /* Free already allocated intr */ 6380 kmem_free(intrp->htable, intrp->intr_size); 6381 6382 (void) nxge_ldgv_uninit(nxgep); 6383 6384 return (NXGE_ERROR | NXGE_DDI_FAILED); 6385 } 6386 6387 ldgp->htable_idx = x; 6388 intrp->intr_added++; 6389 } 6390 6391 intrp->msi_intx_cnt = nactual; 6392 6393 (void) ddi_intr_get_cap(intrp->htable[0], &intrp->intr_cap); 6394 6395 status = nxge_intr_ldgv_init(nxgep); 6396 NXGE_DEBUG_MSG((nxgep, INT_CTL, "<== nxge_add_intrs_adv_type_fix")); 6397 6398 return (status); 6399 } 6400 6401 static void 6402 nxge_remove_intrs(p_nxge_t nxgep) 6403 { 6404 int i, inum; 6405 p_nxge_intr_t intrp; 6406 6407 NXGE_DEBUG_MSG((nxgep, INT_CTL, "==> nxge_remove_intrs")); 6408 intrp = (p_nxge_intr_t)&nxgep->nxge_intr_type; 6409 if (!intrp->intr_registered) { 6410 NXGE_DEBUG_MSG((nxgep, INT_CTL, 6411 "<== nxge_remove_intrs: interrupts not registered")); 6412 return; 6413 } 6414 6415 NXGE_DEBUG_MSG((nxgep, INT_CTL, "==> nxge_remove_intrs:advanced")); 6416 6417 if (intrp->intr_cap & DDI_INTR_FLAG_BLOCK) { 6418 (void) ddi_intr_block_disable(intrp->htable, 6419 intrp->intr_added); 6420 } else { 6421 for (i = 0; i < intrp->intr_added; i++) { 6422 (void) ddi_intr_disable(intrp->htable[i]); 6423 } 6424 } 6425 6426 for (inum = 0; inum < intrp->intr_added; inum++) { 6427 if (intrp->htable[inum]) { 6428 (void) ddi_intr_remove_handler(intrp->htable[inum]); 6429 } 6430 } 6431 6432 for (inum = 0; inum < intrp->msi_intx_cnt; inum++) { 6433 if (intrp->htable[inum]) { 6434 NXGE_DEBUG_MSG((nxgep, DDI_CTL, 6435 "nxge_remove_intrs: ddi_intr_free inum %d " 6436 "msi_intx_cnt %d intr_added %d", 6437 inum, 6438 intrp->msi_intx_cnt, 6439 intrp->intr_added)); 6440 6441 (void) ddi_intr_free(intrp->htable[inum]); 6442 } 6443 } 6444 6445 kmem_free(intrp->htable, intrp->intr_size); 6446 intrp->intr_registered = B_FALSE; 6447 intrp->intr_enabled = B_FALSE; 6448 intrp->msi_intx_cnt = 0; 6449 intrp->intr_added = 0; 6450 6451 (void) nxge_ldgv_uninit(nxgep); 6452 6453 (void) ddi_prop_remove(DDI_DEV_T_NONE, nxgep->dip, 6454 "#msix-request"); 6455 6456 NXGE_DEBUG_MSG((nxgep, INT_CTL, "<== nxge_remove_intrs")); 6457 } 6458 6459 /*ARGSUSED*/ 6460 static void 6461 nxge_intrs_enable(p_nxge_t nxgep) 6462 { 6463 p_nxge_intr_t intrp; 6464 int i; 6465 int status; 6466 6467 NXGE_DEBUG_MSG((nxgep, INT_CTL, "==> nxge_intrs_enable")); 6468 6469 intrp = (p_nxge_intr_t)&nxgep->nxge_intr_type; 6470 6471 if (!intrp->intr_registered) { 6472 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, "<== nxge_intrs_enable: " 6473 "interrupts are not registered")); 6474 return; 6475 } 6476 6477 if (intrp->intr_enabled) { 6478 NXGE_DEBUG_MSG((nxgep, INT_CTL, 6479 "<== nxge_intrs_enable: already enabled")); 6480 return; 6481 } 6482 6483 if (intrp->intr_cap & DDI_INTR_FLAG_BLOCK) { 6484 status = ddi_intr_block_enable(intrp->htable, 6485 intrp->intr_added); 6486 NXGE_DEBUG_MSG((nxgep, INT_CTL, "==> nxge_intrs_enable " 6487 "block enable - status 0x%x total inums #%d\n", 6488 status, intrp->intr_added)); 6489 } else { 6490 for (i = 0; i < intrp->intr_added; i++) { 6491 status = ddi_intr_enable(intrp->htable[i]); 6492 NXGE_DEBUG_MSG((nxgep, INT_CTL, "==> nxge_intrs_enable " 6493 "ddi_intr_enable:enable - status 0x%x " 6494 "total inums %d enable inum #%d\n", 6495 status, intrp->intr_added, i)); 6496 if (status == DDI_SUCCESS) { 6497 intrp->intr_enabled = B_TRUE; 6498 } 6499 } 6500 } 6501 6502 NXGE_DEBUG_MSG((nxgep, INT_CTL, "<== nxge_intrs_enable")); 6503 } 6504 6505 /*ARGSUSED*/ 6506 static void 6507 nxge_intrs_disable(p_nxge_t nxgep) 6508 { 6509 p_nxge_intr_t intrp; 6510 int i; 6511 6512 NXGE_DEBUG_MSG((nxgep, INT_CTL, "==> nxge_intrs_disable")); 6513 6514 intrp = (p_nxge_intr_t)&nxgep->nxge_intr_type; 6515 6516 if (!intrp->intr_registered) { 6517 NXGE_DEBUG_MSG((nxgep, INT_CTL, "<== nxge_intrs_disable: " 6518 "interrupts are not registered")); 6519 return; 6520 } 6521 6522 if (intrp->intr_cap & DDI_INTR_FLAG_BLOCK) { 6523 (void) ddi_intr_block_disable(intrp->htable, 6524 intrp->intr_added); 6525 } else { 6526 for (i = 0; i < intrp->intr_added; i++) { 6527 (void) ddi_intr_disable(intrp->htable[i]); 6528 } 6529 } 6530 6531 intrp->intr_enabled = B_FALSE; 6532 NXGE_DEBUG_MSG((nxgep, INT_CTL, "<== nxge_intrs_disable")); 6533 } 6534 6535 nxge_status_t 6536 nxge_mac_register(p_nxge_t nxgep) 6537 { 6538 mac_register_t *macp; 6539 int status; 6540 6541 NXGE_DEBUG_MSG((nxgep, DDI_CTL, "==> nxge_mac_register")); 6542 6543 if ((macp = mac_alloc(MAC_VERSION)) == NULL) 6544 return (NXGE_ERROR); 6545 6546 macp->m_type_ident = MAC_PLUGIN_IDENT_ETHER; 6547 macp->m_driver = nxgep; 6548 macp->m_dip = nxgep->dip; 6549 if (!isLDOMguest(nxgep)) { 6550 macp->m_src_addr = nxgep->ouraddr.ether_addr_octet; 6551 } else { 6552 macp->m_src_addr = KMEM_ZALLOC(MAXMACADDRLEN, KM_SLEEP); 6553 macp->m_dst_addr = KMEM_ZALLOC(MAXMACADDRLEN, KM_SLEEP); 6554 (void) memset(macp->m_src_addr, 0xff, sizeof (MAXMACADDRLEN)); 6555 } 6556 macp->m_callbacks = &nxge_m_callbacks; 6557 macp->m_min_sdu = 0; 6558 nxgep->mac.default_mtu = nxgep->mac.maxframesize - 6559 NXGE_EHEADER_VLAN_CRC; 6560 macp->m_max_sdu = nxgep->mac.default_mtu; 6561 macp->m_margin = VLAN_TAGSZ; 6562 macp->m_priv_props = nxge_priv_props; 6563 if (isLDOMguest(nxgep)) 6564 macp->m_v12n = MAC_VIRT_LEVEL1; 6565 else 6566 macp->m_v12n = MAC_VIRT_HIO | MAC_VIRT_LEVEL1; 6567 6568 NXGE_DEBUG_MSG((nxgep, MAC_CTL, 6569 "==> nxge_mac_register: instance %d " 6570 "max_sdu %d margin %d maxframe %d (header %d)", 6571 nxgep->instance, 6572 macp->m_max_sdu, macp->m_margin, 6573 nxgep->mac.maxframesize, 6574 NXGE_EHEADER_VLAN_CRC)); 6575 6576 status = mac_register(macp, &nxgep->mach); 6577 if (isLDOMguest(nxgep)) { 6578 KMEM_FREE(macp->m_src_addr, MAXMACADDRLEN); 6579 KMEM_FREE(macp->m_dst_addr, MAXMACADDRLEN); 6580 } 6581 mac_free(macp); 6582 6583 if (status != 0) { 6584 cmn_err(CE_WARN, 6585 "!nxge_mac_register failed (status %d instance %d)", 6586 status, nxgep->instance); 6587 return (NXGE_ERROR); 6588 } 6589 6590 NXGE_DEBUG_MSG((nxgep, DDI_CTL, "<== nxge_mac_register success " 6591 "(instance %d)", nxgep->instance)); 6592 6593 return (NXGE_OK); 6594 } 6595 6596 void 6597 nxge_err_inject(p_nxge_t nxgep, queue_t *wq, mblk_t *mp) 6598 { 6599 ssize_t size; 6600 mblk_t *nmp; 6601 uint8_t blk_id; 6602 uint8_t chan; 6603 uint32_t err_id; 6604 err_inject_t *eip; 6605 6606 NXGE_DEBUG_MSG((nxgep, STR_CTL, "==> nxge_err_inject")); 6607 6608 size = 1024; 6609 nmp = mp->b_cont; 6610 eip = (err_inject_t *)nmp->b_rptr; 6611 blk_id = eip->blk_id; 6612 err_id = eip->err_id; 6613 chan = eip->chan; 6614 cmn_err(CE_NOTE, "!blk_id = 0x%x\n", blk_id); 6615 cmn_err(CE_NOTE, "!err_id = 0x%x\n", err_id); 6616 cmn_err(CE_NOTE, "!chan = 0x%x\n", chan); 6617 switch (blk_id) { 6618 case MAC_BLK_ID: 6619 break; 6620 case TXMAC_BLK_ID: 6621 break; 6622 case RXMAC_BLK_ID: 6623 break; 6624 case MIF_BLK_ID: 6625 break; 6626 case IPP_BLK_ID: 6627 nxge_ipp_inject_err(nxgep, err_id); 6628 break; 6629 case TXC_BLK_ID: 6630 nxge_txc_inject_err(nxgep, err_id); 6631 break; 6632 case TXDMA_BLK_ID: 6633 nxge_txdma_inject_err(nxgep, err_id, chan); 6634 break; 6635 case RXDMA_BLK_ID: 6636 nxge_rxdma_inject_err(nxgep, err_id, chan); 6637 break; 6638 case ZCP_BLK_ID: 6639 nxge_zcp_inject_err(nxgep, err_id); 6640 break; 6641 case ESPC_BLK_ID: 6642 break; 6643 case FFLP_BLK_ID: 6644 break; 6645 case PHY_BLK_ID: 6646 break; 6647 case ETHER_SERDES_BLK_ID: 6648 break; 6649 case PCIE_SERDES_BLK_ID: 6650 break; 6651 case VIR_BLK_ID: 6652 break; 6653 } 6654 6655 nmp->b_wptr = nmp->b_rptr + size; 6656 NXGE_DEBUG_MSG((nxgep, STR_CTL, "<== nxge_err_inject")); 6657 6658 miocack(wq, mp, (int)size, 0); 6659 } 6660 6661 static int 6662 nxge_init_common_dev(p_nxge_t nxgep) 6663 { 6664 p_nxge_hw_list_t hw_p; 6665 dev_info_t *p_dip; 6666 6667 ASSERT(nxgep != NULL); 6668 6669 NXGE_DEBUG_MSG((nxgep, MOD_CTL, "==> nxge_init_common_device")); 6670 6671 p_dip = nxgep->p_dip; 6672 MUTEX_ENTER(&nxge_common_lock); 6673 NXGE_DEBUG_MSG((nxgep, MOD_CTL, 6674 "==> nxge_init_common_dev:func # %d", 6675 nxgep->function_num)); 6676 /* 6677 * Loop through existing per neptune hardware list. 6678 */ 6679 for (hw_p = nxge_hw_list; hw_p; hw_p = hw_p->next) { 6680 NXGE_DEBUG_MSG((nxgep, MOD_CTL, 6681 "==> nxge_init_common_device:func # %d " 6682 "hw_p $%p parent dip $%p", 6683 nxgep->function_num, 6684 hw_p, 6685 p_dip)); 6686 if (hw_p->parent_devp == p_dip) { 6687 nxgep->nxge_hw_p = hw_p; 6688 hw_p->ndevs++; 6689 hw_p->nxge_p[nxgep->function_num] = nxgep; 6690 NXGE_DEBUG_MSG((nxgep, MOD_CTL, 6691 "==> nxge_init_common_device:func # %d " 6692 "hw_p $%p parent dip $%p " 6693 "ndevs %d (found)", 6694 nxgep->function_num, 6695 hw_p, 6696 p_dip, 6697 hw_p->ndevs)); 6698 break; 6699 } 6700 } 6701 6702 if (hw_p == NULL) { 6703 6704 char **prop_val; 6705 uint_t prop_len; 6706 int i; 6707 6708 NXGE_DEBUG_MSG((nxgep, MOD_CTL, 6709 "==> nxge_init_common_device:func # %d " 6710 "parent dip $%p (new)", 6711 nxgep->function_num, 6712 p_dip)); 6713 hw_p = kmem_zalloc(sizeof (nxge_hw_list_t), KM_SLEEP); 6714 hw_p->parent_devp = p_dip; 6715 hw_p->magic = NXGE_NEPTUNE_MAGIC; 6716 nxgep->nxge_hw_p = hw_p; 6717 hw_p->ndevs++; 6718 hw_p->nxge_p[nxgep->function_num] = nxgep; 6719 hw_p->next = nxge_hw_list; 6720 if (nxgep->niu_type == N2_NIU) { 6721 hw_p->niu_type = N2_NIU; 6722 hw_p->platform_type = P_NEPTUNE_NIU; 6723 hw_p->tcam_size = TCAM_NIU_TCAM_MAX_ENTRY; 6724 } else { 6725 hw_p->niu_type = NIU_TYPE_NONE; 6726 hw_p->platform_type = P_NEPTUNE_NONE; 6727 hw_p->tcam_size = TCAM_NXGE_TCAM_MAX_ENTRY; 6728 } 6729 6730 hw_p->tcam = KMEM_ZALLOC(sizeof (tcam_flow_spec_t) * 6731 hw_p->tcam_size, KM_SLEEP); 6732 6733 MUTEX_INIT(&hw_p->nxge_cfg_lock, NULL, MUTEX_DRIVER, NULL); 6734 MUTEX_INIT(&hw_p->nxge_tcam_lock, NULL, MUTEX_DRIVER, NULL); 6735 MUTEX_INIT(&hw_p->nxge_vlan_lock, NULL, MUTEX_DRIVER, NULL); 6736 MUTEX_INIT(&hw_p->nxge_mdio_lock, NULL, MUTEX_DRIVER, NULL); 6737 6738 nxge_hw_list = hw_p; 6739 6740 if (ddi_prop_lookup_string_array(DDI_DEV_T_ANY, nxgep->dip, 0, 6741 "compatible", &prop_val, &prop_len) == DDI_PROP_SUCCESS) { 6742 for (i = 0; i < prop_len; i++) { 6743 if ((strcmp((caddr_t)prop_val[i], 6744 NXGE_ROCK_COMPATIBLE) == 0)) { 6745 hw_p->platform_type = P_NEPTUNE_ROCK; 6746 NXGE_DEBUG_MSG((nxgep, MOD_CTL, 6747 "ROCK hw_p->platform_type %d", 6748 hw_p->platform_type)); 6749 break; 6750 } 6751 NXGE_DEBUG_MSG((nxgep, MOD_CTL, 6752 "nxge_init_common_dev: read compatible" 6753 " property[%d] val[%s]", 6754 i, (caddr_t)prop_val[i])); 6755 } 6756 } 6757 6758 ddi_prop_free(prop_val); 6759 6760 (void) nxge_scan_ports_phy(nxgep, nxge_hw_list); 6761 } 6762 6763 MUTEX_EXIT(&nxge_common_lock); 6764 6765 nxgep->platform_type = hw_p->platform_type; 6766 NXGE_DEBUG_MSG((nxgep, MOD_CTL, "nxgep->platform_type %d", 6767 nxgep->platform_type)); 6768 if (nxgep->niu_type != N2_NIU) { 6769 nxgep->niu_type = hw_p->niu_type; 6770 } 6771 6772 NXGE_DEBUG_MSG((nxgep, MOD_CTL, 6773 "==> nxge_init_common_device (nxge_hw_list) $%p", 6774 nxge_hw_list)); 6775 NXGE_DEBUG_MSG((nxgep, MOD_CTL, "<== nxge_init_common_device")); 6776 6777 return (NXGE_OK); 6778 } 6779 6780 static void 6781 nxge_uninit_common_dev(p_nxge_t nxgep) 6782 { 6783 p_nxge_hw_list_t hw_p, h_hw_p; 6784 p_nxge_dma_pt_cfg_t p_dma_cfgp; 6785 p_nxge_hw_pt_cfg_t p_cfgp; 6786 dev_info_t *p_dip; 6787 6788 ASSERT(nxgep != NULL); 6789 6790 NXGE_DEBUG_MSG((nxgep, MOD_CTL, "==> nxge_uninit_common_device")); 6791 if (nxgep->nxge_hw_p == NULL) { 6792 NXGE_DEBUG_MSG((nxgep, MOD_CTL, 6793 "<== nxge_uninit_common_device (no common)")); 6794 return; 6795 } 6796 6797 MUTEX_ENTER(&nxge_common_lock); 6798 h_hw_p = nxge_hw_list; 6799 for (hw_p = nxge_hw_list; hw_p; hw_p = hw_p->next) { 6800 p_dip = hw_p->parent_devp; 6801 if (nxgep->nxge_hw_p == hw_p && 6802 p_dip == nxgep->p_dip && 6803 nxgep->nxge_hw_p->magic == NXGE_NEPTUNE_MAGIC && 6804 hw_p->magic == NXGE_NEPTUNE_MAGIC) { 6805 6806 NXGE_DEBUG_MSG((nxgep, MOD_CTL, 6807 "==> nxge_uninit_common_device:func # %d " 6808 "hw_p $%p parent dip $%p " 6809 "ndevs %d (found)", 6810 nxgep->function_num, 6811 hw_p, 6812 p_dip, 6813 hw_p->ndevs)); 6814 6815 /* 6816 * Release the RDC table, a shared resoruce 6817 * of the nxge hardware. The RDC table was 6818 * assigned to this instance of nxge in 6819 * nxge_use_cfg_dma_config(). 6820 */ 6821 if (!isLDOMguest(nxgep)) { 6822 p_dma_cfgp = 6823 (p_nxge_dma_pt_cfg_t)&nxgep->pt_config; 6824 p_cfgp = 6825 (p_nxge_hw_pt_cfg_t)&p_dma_cfgp->hw_config; 6826 (void) nxge_fzc_rdc_tbl_unbind(nxgep, 6827 p_cfgp->def_mac_rxdma_grpid); 6828 6829 /* Cleanup any outstanding groups. */ 6830 nxge_grp_cleanup(nxgep); 6831 } 6832 6833 if (hw_p->ndevs) { 6834 hw_p->ndevs--; 6835 } 6836 hw_p->nxge_p[nxgep->function_num] = NULL; 6837 if (!hw_p->ndevs) { 6838 KMEM_FREE(hw_p->tcam, 6839 sizeof (tcam_flow_spec_t) * 6840 hw_p->tcam_size); 6841 MUTEX_DESTROY(&hw_p->nxge_vlan_lock); 6842 MUTEX_DESTROY(&hw_p->nxge_tcam_lock); 6843 MUTEX_DESTROY(&hw_p->nxge_cfg_lock); 6844 MUTEX_DESTROY(&hw_p->nxge_mdio_lock); 6845 NXGE_DEBUG_MSG((nxgep, MOD_CTL, 6846 "==> nxge_uninit_common_device: " 6847 "func # %d " 6848 "hw_p $%p parent dip $%p " 6849 "ndevs %d (last)", 6850 nxgep->function_num, 6851 hw_p, 6852 p_dip, 6853 hw_p->ndevs)); 6854 6855 nxge_hio_uninit(nxgep); 6856 6857 if (hw_p == nxge_hw_list) { 6858 NXGE_DEBUG_MSG((nxgep, MOD_CTL, 6859 "==> nxge_uninit_common_device:" 6860 "remove head func # %d " 6861 "hw_p $%p parent dip $%p " 6862 "ndevs %d (head)", 6863 nxgep->function_num, 6864 hw_p, 6865 p_dip, 6866 hw_p->ndevs)); 6867 nxge_hw_list = hw_p->next; 6868 } else { 6869 NXGE_DEBUG_MSG((nxgep, MOD_CTL, 6870 "==> nxge_uninit_common_device:" 6871 "remove middle func # %d " 6872 "hw_p $%p parent dip $%p " 6873 "ndevs %d (middle)", 6874 nxgep->function_num, 6875 hw_p, 6876 p_dip, 6877 hw_p->ndevs)); 6878 h_hw_p->next = hw_p->next; 6879 } 6880 6881 nxgep->nxge_hw_p = NULL; 6882 KMEM_FREE(hw_p, sizeof (nxge_hw_list_t)); 6883 } 6884 break; 6885 } else { 6886 h_hw_p = hw_p; 6887 } 6888 } 6889 6890 MUTEX_EXIT(&nxge_common_lock); 6891 NXGE_DEBUG_MSG((nxgep, MOD_CTL, 6892 "==> nxge_uninit_common_device (nxge_hw_list) $%p", 6893 nxge_hw_list)); 6894 6895 NXGE_DEBUG_MSG((nxgep, MOD_CTL, "<= nxge_uninit_common_device")); 6896 } 6897 6898 /* 6899 * Determines the number of ports from the niu_type or the platform type. 6900 * Returns the number of ports, or returns zero on failure. 6901 */ 6902 6903 int 6904 nxge_get_nports(p_nxge_t nxgep) 6905 { 6906 int nports = 0; 6907 6908 switch (nxgep->niu_type) { 6909 case N2_NIU: 6910 case NEPTUNE_2_10GF: 6911 nports = 2; 6912 break; 6913 case NEPTUNE_4_1GC: 6914 case NEPTUNE_2_10GF_2_1GC: 6915 case NEPTUNE_1_10GF_3_1GC: 6916 case NEPTUNE_1_1GC_1_10GF_2_1GC: 6917 case NEPTUNE_2_10GF_2_1GRF: 6918 nports = 4; 6919 break; 6920 default: 6921 switch (nxgep->platform_type) { 6922 case P_NEPTUNE_NIU: 6923 case P_NEPTUNE_ATLAS_2PORT: 6924 nports = 2; 6925 break; 6926 case P_NEPTUNE_ATLAS_4PORT: 6927 case P_NEPTUNE_MARAMBA_P0: 6928 case P_NEPTUNE_MARAMBA_P1: 6929 case P_NEPTUNE_ROCK: 6930 case P_NEPTUNE_ALONSO: 6931 nports = 4; 6932 break; 6933 default: 6934 break; 6935 } 6936 break; 6937 } 6938 6939 return (nports); 6940 } 6941 6942 /* 6943 * The following two functions are to support 6944 * PSARC/2007/453 MSI-X interrupt limit override. 6945 */ 6946 static int 6947 nxge_create_msi_property(p_nxge_t nxgep) 6948 { 6949 int nmsi; 6950 extern int ncpus; 6951 6952 NXGE_DEBUG_MSG((nxgep, MOD_CTL, "==>nxge_create_msi_property")); 6953 6954 switch (nxgep->mac.portmode) { 6955 case PORT_10G_COPPER: 6956 case PORT_10G_FIBER: 6957 case PORT_10G_TN1010: 6958 (void) ddi_prop_create(DDI_DEV_T_NONE, nxgep->dip, 6959 DDI_PROP_CANSLEEP, "#msix-request", NULL, 0); 6960 /* 6961 * The maximum MSI-X requested will be 8. 6962 * If the # of CPUs is less than 8, we will request 6963 * # MSI-X based on the # of CPUs (default). 6964 */ 6965 NXGE_DEBUG_MSG((nxgep, MOD_CTL, 6966 "==>nxge_create_msi_property (10G): nxge_msix_10g_intrs %d", 6967 nxge_msix_10g_intrs)); 6968 if ((nxge_msix_10g_intrs == 0) || 6969 (nxge_msix_10g_intrs > NXGE_MSIX_MAX_ALLOWED)) { 6970 nmsi = NXGE_MSIX_REQUEST_10G; 6971 NXGE_DEBUG_MSG((nxgep, MOD_CTL, 6972 "==>nxge_create_msi_property (10G): reset to 8")); 6973 } else { 6974 nmsi = nxge_msix_10g_intrs; 6975 } 6976 6977 /* 6978 * If # of interrupts requested is 8 (default), 6979 * the checking of the number of cpus will be 6980 * be maintained. 6981 */ 6982 if ((nmsi == NXGE_MSIX_REQUEST_10G) && 6983 (ncpus < nmsi)) { 6984 NXGE_DEBUG_MSG((nxgep, MOD_CTL, 6985 "==>nxge_create_msi_property (10G): reset to 8")); 6986 nmsi = ncpus; 6987 } 6988 NXGE_DEBUG_MSG((nxgep, MOD_CTL, 6989 "==>nxge_create_msi_property(10G): exists 0x%x (nmsi %d)", 6990 ddi_prop_exists(DDI_DEV_T_NONE, nxgep->dip, 6991 DDI_PROP_CANSLEEP, "#msix-request"), nmsi)); 6992 break; 6993 6994 default: 6995 (void) ddi_prop_create(DDI_DEV_T_NONE, nxgep->dip, 6996 DDI_PROP_CANSLEEP, "#msix-request", NULL, 0); 6997 NXGE_DEBUG_MSG((nxgep, MOD_CTL, 6998 "==>nxge_create_msi_property (1G): nxge_msix_1g_intrs %d", 6999 nxge_msix_1g_intrs)); 7000 if ((nxge_msix_1g_intrs == 0) || 7001 (nxge_msix_1g_intrs > NXGE_MSIX_MAX_ALLOWED)) { 7002 nmsi = NXGE_MSIX_REQUEST_1G; 7003 NXGE_DEBUG_MSG((nxgep, MOD_CTL, 7004 "==>nxge_create_msi_property (1G): reset to 2")); 7005 } else { 7006 nmsi = nxge_msix_1g_intrs; 7007 } 7008 NXGE_DEBUG_MSG((nxgep, MOD_CTL, 7009 "==>nxge_create_msi_property(1G): exists 0x%x (nmsi %d)", 7010 ddi_prop_exists(DDI_DEV_T_NONE, nxgep->dip, 7011 DDI_PROP_CANSLEEP, "#msix-request"), nmsi)); 7012 break; 7013 } 7014 7015 NXGE_DEBUG_MSG((nxgep, MOD_CTL, "<==nxge_create_msi_property")); 7016 return (nmsi); 7017 } 7018 7019 /* 7020 * The following is a software around for the Neptune hardware's 7021 * interrupt bugs; The Neptune hardware may generate spurious interrupts when 7022 * an interrupr handler is removed. 7023 */ 7024 #define NXGE_PCI_PORT_LOGIC_OFFSET 0x98 7025 #define NXGE_PIM_RESET (1ULL << 29) 7026 #define NXGE_GLU_RESET (1ULL << 30) 7027 #define NXGE_NIU_RESET (1ULL << 31) 7028 #define NXGE_PCI_RESET_ALL (NXGE_PIM_RESET | \ 7029 NXGE_GLU_RESET | \ 7030 NXGE_NIU_RESET) 7031 7032 #define NXGE_WAIT_QUITE_TIME 200000 7033 #define NXGE_WAIT_QUITE_RETRY 40 7034 #define NXGE_PCI_RESET_WAIT 1000000 /* one second */ 7035 7036 static void 7037 nxge_niu_peu_reset(p_nxge_t nxgep) 7038 { 7039 uint32_t rvalue; 7040 p_nxge_hw_list_t hw_p; 7041 p_nxge_t fnxgep; 7042 int i, j; 7043 7044 NXGE_DEBUG_MSG((nxgep, NXGE_ERR_CTL, "==> nxge_niu_peu_reset")); 7045 if ((hw_p = nxgep->nxge_hw_p) == NULL) { 7046 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, 7047 "==> nxge_niu_peu_reset: NULL hardware pointer")); 7048 return; 7049 } 7050 7051 NXGE_DEBUG_MSG((nxgep, NXGE_ERR_CTL, 7052 "==> nxge_niu_peu_reset: flags 0x%x link timer id %d timer id %d", 7053 hw_p->flags, nxgep->nxge_link_poll_timerid, 7054 nxgep->nxge_timerid)); 7055 7056 MUTEX_ENTER(&hw_p->nxge_cfg_lock); 7057 /* 7058 * Make sure other instances from the same hardware 7059 * stop sending PIO and in quiescent state. 7060 */ 7061 for (i = 0; i < NXGE_MAX_PORTS; i++) { 7062 fnxgep = hw_p->nxge_p[i]; 7063 NXGE_DEBUG_MSG((nxgep, NXGE_ERR_CTL, 7064 "==> nxge_niu_peu_reset: checking entry %d " 7065 "nxgep $%p", i, fnxgep)); 7066 #ifdef NXGE_DEBUG 7067 if (fnxgep) { 7068 NXGE_DEBUG_MSG((nxgep, NXGE_ERR_CTL, 7069 "==> nxge_niu_peu_reset: entry %d (function %d) " 7070 "link timer id %d hw timer id %d", 7071 i, fnxgep->function_num, 7072 fnxgep->nxge_link_poll_timerid, 7073 fnxgep->nxge_timerid)); 7074 } 7075 #endif 7076 if (fnxgep && fnxgep != nxgep && 7077 (fnxgep->nxge_timerid || fnxgep->nxge_link_poll_timerid)) { 7078 NXGE_DEBUG_MSG((nxgep, NXGE_ERR_CTL, 7079 "==> nxge_niu_peu_reset: checking $%p " 7080 "(function %d) timer ids", 7081 fnxgep, fnxgep->function_num)); 7082 for (j = 0; j < NXGE_WAIT_QUITE_RETRY; j++) { 7083 NXGE_DEBUG_MSG((nxgep, NXGE_ERR_CTL, 7084 "==> nxge_niu_peu_reset: waiting")); 7085 NXGE_DELAY(NXGE_WAIT_QUITE_TIME); 7086 if (!fnxgep->nxge_timerid && 7087 !fnxgep->nxge_link_poll_timerid) { 7088 break; 7089 } 7090 } 7091 NXGE_DELAY(NXGE_WAIT_QUITE_TIME); 7092 if (fnxgep->nxge_timerid || 7093 fnxgep->nxge_link_poll_timerid) { 7094 MUTEX_EXIT(&hw_p->nxge_cfg_lock); 7095 NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, 7096 "<== nxge_niu_peu_reset: cannot reset " 7097 "hardware (devices are still in use)")); 7098 return; 7099 } 7100 } 7101 } 7102 7103 if ((hw_p->flags & COMMON_RESET_NIU_PCI) != COMMON_RESET_NIU_PCI) { 7104 hw_p->flags |= COMMON_RESET_NIU_PCI; 7105 rvalue = pci_config_get32(nxgep->dev_regs->nxge_pciregh, 7106 NXGE_PCI_PORT_LOGIC_OFFSET); 7107 NXGE_DEBUG_MSG((nxgep, NXGE_ERR_CTL, 7108 "nxge_niu_peu_reset: read offset 0x%x (%d) " 7109 "(data 0x%x)", 7110 NXGE_PCI_PORT_LOGIC_OFFSET, 7111 NXGE_PCI_PORT_LOGIC_OFFSET, 7112 rvalue)); 7113 7114 rvalue |= NXGE_PCI_RESET_ALL; 7115 pci_config_put32(nxgep->dev_regs->nxge_pciregh, 7116 NXGE_PCI_PORT_LOGIC_OFFSET, rvalue); 7117 NXGE_DEBUG_MSG((nxgep, NXGE_ERR_CTL, 7118 "nxge_niu_peu_reset: RESETTING NIU: write NIU reset 0x%x", 7119 rvalue)); 7120 7121 NXGE_DELAY(NXGE_PCI_RESET_WAIT); 7122 } 7123 7124 MUTEX_EXIT(&hw_p->nxge_cfg_lock); 7125 NXGE_DEBUG_MSG((nxgep, NXGE_ERR_CTL, "<== nxge_niu_peu_reset")); 7126 } 7127 7128 static void 7129 nxge_set_pci_replay_timeout(p_nxge_t nxgep) 7130 { 7131 p_dev_regs_t dev_regs; 7132 uint32_t value; 7133 7134 NXGE_DEBUG_MSG((nxgep, DDI_CTL, "==> nxge_set_pci_replay_timeout")); 7135 7136 if (!nxge_set_replay_timer) { 7137 NXGE_DEBUG_MSG((nxgep, DDI_CTL, 7138 "==> nxge_set_pci_replay_timeout: will not change " 7139 "the timeout")); 7140 return; 7141 } 7142 7143 dev_regs = nxgep->dev_regs; 7144 NXGE_DEBUG_MSG((nxgep, DDI_CTL, 7145 "==> nxge_set_pci_replay_timeout: dev_regs 0x%p pcireg 0x%p", 7146 dev_regs, dev_regs->nxge_pciregh)); 7147 7148 if (dev_regs == NULL || (dev_regs->nxge_pciregh == NULL)) { 7149 NXGE_DEBUG_MSG((nxgep, DDI_CTL, 7150 "==> nxge_set_pci_replay_timeout: NULL dev_regs $%p or " 7151 "no PCI handle", 7152 dev_regs)); 7153 return; 7154 } 7155 value = (pci_config_get32(dev_regs->nxge_pciregh, 7156 PCI_REPLAY_TIMEOUT_CFG_OFFSET) | 7157 (nxge_replay_timeout << PCI_REPLAY_TIMEOUT_SHIFT)); 7158 7159 NXGE_DEBUG_MSG((nxgep, DDI_CTL, 7160 "nxge_set_pci_replay_timeout: replay timeout value before set 0x%x " 7161 "(timeout value to set 0x%x at offset 0x%x) value 0x%x", 7162 pci_config_get32(dev_regs->nxge_pciregh, 7163 PCI_REPLAY_TIMEOUT_CFG_OFFSET), nxge_replay_timeout, 7164 PCI_REPLAY_TIMEOUT_CFG_OFFSET, value)); 7165 7166 pci_config_put32(dev_regs->nxge_pciregh, PCI_REPLAY_TIMEOUT_CFG_OFFSET, 7167 value); 7168 7169 NXGE_DEBUG_MSG((nxgep, DDI_CTL, 7170 "nxge_set_pci_replay_timeout: replay timeout value after set 0x%x", 7171 pci_config_get32(dev_regs->nxge_pciregh, 7172 PCI_REPLAY_TIMEOUT_CFG_OFFSET))); 7173 7174 NXGE_DEBUG_MSG((nxgep, DDI_CTL, "<== nxge_set_pci_replay_timeout")); 7175 } 7176 7177 /* 7178 * quiesce(9E) entry point. 7179 * 7180 * This function is called when the system is single-threaded at high 7181 * PIL with preemption disabled. Therefore, this function must not be 7182 * blocked. 7183 * 7184 * This function returns DDI_SUCCESS on success, or DDI_FAILURE on failure. 7185 * DDI_FAILURE indicates an error condition and should almost never happen. 7186 */ 7187 static int 7188 nxge_quiesce(dev_info_t *dip) 7189 { 7190 int instance = ddi_get_instance(dip); 7191 p_nxge_t nxgep = (p_nxge_t)ddi_get_soft_state(nxge_list, instance); 7192 7193 if (nxgep == NULL) 7194 return (DDI_FAILURE); 7195 7196 /* Turn off debugging */ 7197 nxge_debug_level = NO_DEBUG; 7198 nxgep->nxge_debug_level = NO_DEBUG; 7199 npi_debug_level = NO_DEBUG; 7200 7201 /* 7202 * Stop link monitor only when linkchkmod is interrupt based 7203 */ 7204 if (nxgep->mac.linkchkmode == LINKCHK_INTR) { 7205 (void) nxge_link_monitor(nxgep, LINK_MONITOR_STOP); 7206 } 7207 7208 (void) nxge_intr_hw_disable(nxgep); 7209 7210 /* 7211 * Reset the receive MAC side. 7212 */ 7213 (void) nxge_rx_mac_disable(nxgep); 7214 7215 /* Disable and soft reset the IPP */ 7216 if (!isLDOMguest(nxgep)) 7217 (void) nxge_ipp_disable(nxgep); 7218 7219 /* 7220 * Reset the transmit/receive DMA side. 7221 */ 7222 (void) nxge_txdma_hw_mode(nxgep, NXGE_DMA_STOP); 7223 (void) nxge_rxdma_hw_mode(nxgep, NXGE_DMA_STOP); 7224 7225 /* 7226 * Reset the transmit MAC side. 7227 */ 7228 (void) nxge_tx_mac_disable(nxgep); 7229 7230 return (DDI_SUCCESS); 7231 }