1 /* 2 * CDDL HEADER START 3 * 4 * The contents of this file are subject to the terms of the 5 * Common Development and Distribution License (the "License"). 6 * You may not use this file except in compliance with the License. 7 * 8 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE 9 * or http://www.opensolaris.org/os/licensing. 10 * See the License for the specific language governing permissions 11 * and limitations under the License. 12 * 13 * When distributing Covered Code, include this CDDL HEADER in each 14 * file and include the License file at usr/src/OPENSOLARIS.LICENSE. 15 * If applicable, add the following below this CDDL HEADER, with the 16 * fields enclosed by brackets "[]" replaced with your own identifying 17 * information: Portions Copyright [yyyy] [name of copyright owner] 18 * 19 * CDDL HEADER END 20 */ 21 /* 22 * Copyright 2008 Sun Microsystems, Inc. All rights reserved. 23 * Use is subject to license terms. 24 * Copyright (c) 2016 by Delphix. All rights reserved. 25 */ 26 27 /* 28 * SunOS 5.x Multithreaded STREAMS DLPI FCIP Module 29 * This is a pseudo driver module to handle encapsulation of IP and ARP 30 * datagrams over FibreChannel interfaces. FCIP is a cloneable STREAMS 31 * driver module which interfaces with IP/ARP using DLPI. This module 32 * is a Style-2 DLS provider. 33 * 34 * The implementation of this module is based on RFC 2625 which gives 35 * details on the encapsulation of IP/ARP data over FibreChannel. 36 * The fcip module needs to resolve an IP address to a port address before 37 * sending data to a destination port. A FC device port has 2 addresses 38 * associated with it: A 8 byte World Wide unique Port Name and a 3 byte 39 * volatile Port number or Port_ID. 40 * 41 * The mapping between a IP address and the World Wide Port Name is handled 42 * by the ARP layer since the IP over FC draft requires the MAC address to 43 * be the least significant six bytes of the WorldWide Port Names. The 44 * fcip module however needs to identify the destination port uniquely when 45 * the destination FC device has multiple FC ports. 46 * 47 * The FC layer mapping between the World Wide Port Name and the Port_ID 48 * will be handled through the use of a fabric name server or through the 49 * use of the FARP ELS command as described in the draft. Since the Port_IDs 50 * are volatile, the mapping between the World Wide Port Name and Port_IDs 51 * must be maintained and validated before use each time a datagram 52 * needs to be sent to the destination ports. The FC transport module 53 * informs the fcip module of all changes to states of ports on the 54 * fabric through registered callbacks. This enables the fcip module 55 * to maintain the WW_PN to Port_ID mappings current. 56 * 57 * For details on how this module interfaces with the FibreChannel Transport 58 * modules, refer to PSARC/1997/385. Chapter 3 of the FibreChannel Transport 59 * Programming guide details the APIs between ULPs and the Transport. 60 * 61 * Now for some Caveats: 62 * 63 * RFC 2625 requires that a FibreChannel Port name (the Port WWN) have 64 * the NAA bits set to '0001' indicating a IEEE 48bit address which 65 * corresponds to a ULA (Universal LAN MAC address). But with FibreChannel 66 * adapters containing 2 or more ports, IEEE naming cannot identify the 67 * ports on an adapter uniquely so we will in the first implementation 68 * be operating only on Port 0 of each adapter. 69 */ 70 71 #include <sys/types.h> 72 #include <sys/errno.h> 73 #include <sys/debug.h> 74 #include <sys/time.h> 75 #include <sys/sysmacros.h> 76 #include <sys/systm.h> 77 #include <sys/user.h> 78 #include <sys/stropts.h> 79 #include <sys/stream.h> 80 #include <sys/strlog.h> 81 #include <sys/strsubr.h> 82 #include <sys/cmn_err.h> 83 #include <sys/cpu.h> 84 #include <sys/kmem.h> 85 #include <sys/conf.h> 86 #include <sys/ddi.h> 87 #include <sys/sunddi.h> 88 #include <sys/ksynch.h> 89 #include <sys/stat.h> 90 #include <sys/kstat.h> 91 #include <sys/vtrace.h> 92 #include <sys/strsun.h> 93 #include <sys/varargs.h> 94 #include <sys/modctl.h> 95 #include <sys/thread.h> 96 #include <sys/var.h> 97 #include <sys/proc.h> 98 #include <inet/common.h> 99 #include <netinet/ip6.h> 100 #include <inet/ip.h> 101 #include <inet/arp.h> 102 #include <inet/mi.h> 103 #include <inet/nd.h> 104 #include <sys/dlpi.h> 105 #include <sys/ethernet.h> 106 #include <sys/file.h> 107 #include <sys/syslog.h> 108 #include <sys/disp.h> 109 #include <sys/taskq.h> 110 111 /* 112 * Leadville includes 113 */ 114 115 #include <sys/fibre-channel/fc.h> 116 #include <sys/fibre-channel/impl/fc_ulpif.h> 117 #include <sys/fibre-channel/ulp/fcip.h> 118 119 /* 120 * TNF Probe/trace facility include 121 */ 122 #if defined(lint) || defined(FCIP_TNF_ENABLED) 123 #include <sys/tnf_probe.h> 124 #endif 125 126 #define FCIP_ESBALLOC 127 128 /* 129 * Function prototypes 130 */ 131 132 /* standard loadable modules entry points */ 133 static int fcip_attach(dev_info_t *, ddi_attach_cmd_t); 134 static int fcip_detach(dev_info_t *, ddi_detach_cmd_t); 135 static void fcip_dodetach(struct fcipstr *slp); 136 static int fcip_getinfo(dev_info_t *dip, ddi_info_cmd_t cmd, 137 void *arg, void **result); 138 139 140 /* streams specific */ 141 static void fcip_setipq(struct fcip *fptr); 142 static int fcip_wput(queue_t *, mblk_t *); 143 static int fcip_wsrv(queue_t *); 144 static void fcip_proto(queue_t *, mblk_t *); 145 static void fcip_ioctl(queue_t *, mblk_t *); 146 static int fcip_open(queue_t *wq, dev_t *devp, int flag, 147 int sflag, cred_t *credp); 148 static int fcip_close(queue_t *rq, int flag, int otyp, cred_t *credp); 149 static int fcip_start(queue_t *wq, mblk_t *mp, struct fcip *fptr, 150 struct fcip_dest *fdestp, int flags); 151 static void fcip_sendup(struct fcip *fptr, mblk_t *mp, 152 struct fcipstr *(*acceptfunc)()); 153 static struct fcipstr *fcip_accept(struct fcipstr *slp, struct fcip *fptr, 154 int type, la_wwn_t *dhostp); 155 static mblk_t *fcip_addudind(struct fcip *fptr, mblk_t *mp, 156 fcph_network_hdr_t *nhdr, int type); 157 static int fcip_setup_mac_addr(struct fcip *fptr); 158 static void fcip_kstat_init(struct fcip *fptr); 159 static int fcip_stat_update(kstat_t *, int); 160 161 162 /* dlpi specific */ 163 static void fcip_spareq(queue_t *wq, mblk_t *mp); 164 static void fcip_pareq(queue_t *wq, mblk_t *mp); 165 static void fcip_ubreq(queue_t *wq, mblk_t *mp); 166 static void fcip_breq(queue_t *wq, mblk_t *mp); 167 static void fcip_dreq(queue_t *wq, mblk_t *mp); 168 static void fcip_areq(queue_t *wq, mblk_t *mp); 169 static void fcip_udreq(queue_t *wq, mblk_t *mp); 170 static void fcip_ireq(queue_t *wq, mblk_t *mp); 171 static void fcip_dl_ioc_hdr_info(queue_t *wq, mblk_t *mp); 172 173 174 /* solaris sundry, DR/CPR etc */ 175 static int fcip_cache_constructor(void *buf, void *arg, int size); 176 static void fcip_cache_destructor(void *buf, void *size); 177 static int fcip_handle_suspend(fcip_port_info_t *fport, fc_detach_cmd_t cmd); 178 static int fcip_handle_resume(fcip_port_info_t *fport, 179 fc_ulp_port_info_t *port_info, fc_attach_cmd_t cmd); 180 static fcip_port_info_t *fcip_softstate_free(fcip_port_info_t *fport); 181 static int fcip_port_attach_handler(struct fcip *fptr); 182 183 184 /* 185 * ulp - transport interface function prototypes 186 */ 187 static int fcip_port_attach(opaque_t ulp_handle, fc_ulp_port_info_t *, 188 fc_attach_cmd_t cmd, uint32_t sid); 189 static int fcip_port_detach(opaque_t ulp_handle, fc_ulp_port_info_t *, 190 fc_detach_cmd_t cmd); 191 static int fcip_port_ioctl(opaque_t ulp_handle, opaque_t port_handle, 192 dev_t dev, int cmd, intptr_t data, int mode, cred_t *credp, int *rval, 193 uint32_t claimed); 194 static void fcip_statec_cb(opaque_t ulp_handle, opaque_t phandle, 195 uint32_t port_state, uint32_t port_top, fc_portmap_t changelist[], 196 uint32_t listlen, uint32_t sid); 197 static int fcip_els_cb(opaque_t ulp_handle, opaque_t phandle, 198 fc_unsol_buf_t *buf, uint32_t claimed); 199 static int fcip_data_cb(opaque_t ulp_handle, opaque_t phandle, 200 fc_unsol_buf_t *payload, uint32_t claimed); 201 202 203 /* Routing table specific */ 204 static void fcip_handle_topology(struct fcip *fptr); 205 static int fcip_init_port(struct fcip *fptr); 206 struct fcip_routing_table *fcip_lookup_rtable(struct fcip *fptr, 207 la_wwn_t *pwwn, int matchflag); 208 static void fcip_rt_update(struct fcip *fptr, fc_portmap_t *devlist, 209 uint32_t listlen); 210 static void fcip_rt_flush(struct fcip *fptr); 211 static void fcip_rte_remove_deferred(void *arg); 212 static int fcip_do_plogi(struct fcip *fptr, struct fcip_routing_table *frp); 213 214 215 /* dest table specific */ 216 static struct fcip_dest *fcip_get_dest(struct fcip *fptr, 217 la_wwn_t *dlphys); 218 static struct fcip_dest *fcip_add_dest(struct fcip *fptr, 219 struct fcip_routing_table *frp); 220 static int fcip_dest_add_broadcast_entry(struct fcip *fptr, int new_flag); 221 static uint32_t fcip_get_broadcast_did(struct fcip *fptr); 222 static void fcip_cleanup_dest(struct fcip *fptr); 223 224 225 /* helper functions */ 226 static fcip_port_info_t *fcip_get_port(opaque_t phandle); 227 static int fcip_wwn_compare(la_wwn_t *wwn1, la_wwn_t *wwn2, int flag); 228 static void fcip_ether_to_str(struct ether_addr *e, caddr_t s); 229 static int fcip_port_get_num_pkts(struct fcip *fptr); 230 static int fcip_check_port_busy(struct fcip *fptr); 231 static void fcip_check_remove_minor_node(void); 232 static int fcip_set_wwn(la_wwn_t *pwwn); 233 static int fcip_plogi_in_progress(struct fcip *fptr); 234 static int fcip_check_port_exists(struct fcip *fptr); 235 static int fcip_is_supported_fc_topology(int fc_topology); 236 237 238 /* pkt specific */ 239 static fcip_pkt_t *fcip_pkt_alloc(struct fcip *fptr, mblk_t *bp, 240 int flags, int datalen); 241 static void fcip_pkt_free(struct fcip_pkt *fcip_pkt, int flags); 242 static fcip_pkt_t *fcip_ipkt_alloc(struct fcip *fptr, int cmdlen, 243 int resplen, opaque_t pd, int flags); 244 static void fcip_ipkt_free(fcip_pkt_t *fcip_pkt); 245 static void fcip_ipkt_callback(fc_packet_t *fc_pkt); 246 static void fcip_free_pkt_dma(fcip_pkt_t *fcip_pkt); 247 static void fcip_pkt_callback(fc_packet_t *fc_pkt); 248 static void fcip_init_unicast_pkt(fcip_pkt_t *fcip_pkt, fc_portid_t sid, 249 fc_portid_t did, void (*comp) ()); 250 static int fcip_transport(fcip_pkt_t *fcip_pkt); 251 static void fcip_pkt_timeout(void *arg); 252 static void fcip_timeout(void *arg); 253 static void fcip_fdestp_enqueue_pkt(struct fcip_dest *fdestp, 254 fcip_pkt_t *fcip_pkt); 255 static int fcip_fdestp_dequeue_pkt(struct fcip_dest *fdestp, 256 fcip_pkt_t *fcip_pkt); 257 static int fcip_sendup_constructor(void *buf, void *arg, int flags); 258 static void fcip_sendup_thr(void *arg); 259 static int fcip_sendup_alloc_enque(struct fcip *ftpr, mblk_t *mp, 260 struct fcipstr *(*f)()); 261 262 /* 263 * zero copy inbound data handling 264 */ 265 #ifdef FCIP_ESBALLOC 266 static void fcip_ubfree(char *arg); 267 #endif /* FCIP_ESBALLOC */ 268 269 #if !defined(FCIP_ESBALLOC) 270 static void *fcip_allocb(size_t size, uint_t pri); 271 #endif 272 273 274 /* FCIP FARP support functions */ 275 static struct fcip_dest *fcip_do_farp(struct fcip *fptr, la_wwn_t *pwwn, 276 char *ip_addr, size_t ip_addr_len, int flags); 277 static void fcip_init_broadcast_pkt(fcip_pkt_t *fcip_pkt, void (*comp) (), 278 int is_els); 279 static int fcip_handle_farp_request(struct fcip *fptr, la_els_farp_t *fcmd); 280 static int fcip_handle_farp_response(struct fcip *fptr, la_els_farp_t *fcmd); 281 static void fcip_cache_arp_broadcast(struct fcip *ftpr, fc_unsol_buf_t *buf); 282 static void fcip_port_ns(void *arg); 283 284 #ifdef DEBUG 285 286 #include <sys/debug.h> 287 288 #define FCIP_DEBUG_DEFAULT 0x1 289 #define FCIP_DEBUG_ATTACH 0x2 290 #define FCIP_DEBUG_INIT 0x4 291 #define FCIP_DEBUG_DETACH 0x8 292 #define FCIP_DEBUG_DLPI 0x10 293 #define FCIP_DEBUG_ELS 0x20 294 #define FCIP_DEBUG_DOWNSTREAM 0x40 295 #define FCIP_DEBUG_UPSTREAM 0x80 296 #define FCIP_DEBUG_MISC 0x100 297 298 #define FCIP_DEBUG_STARTUP (FCIP_DEBUG_ATTACH|FCIP_DEBUG_INIT) 299 #define FCIP_DEBUG_DATAOUT (FCIP_DEBUG_DLPI|FCIP_DEBUG_DOWNSTREAM) 300 #define FCIP_DEBUG_DATAIN (FCIP_DEBUG_ELS|FCIP_DEBUG_UPSTREAM) 301 302 static int fcip_debug = FCIP_DEBUG_DEFAULT; 303 304 #define FCIP_DEBUG(level, args) \ 305 if (fcip_debug & (level)) cmn_err args; 306 307 #else /* DEBUG */ 308 309 #define FCIP_DEBUG(level, args) /* do nothing */ 310 311 #endif /* DEBUG */ 312 313 #define KIOIP KSTAT_INTR_PTR(fcip->fcip_intrstats) 314 315 /* 316 * Endian independent ethernet to WWN copy 317 */ 318 #define ether_to_wwn(E, W) \ 319 bzero((void *)(W), sizeof (la_wwn_t)); \ 320 bcopy((void *)(E), (void *)&((W)->raw_wwn[2]), ETHERADDRL); \ 321 (W)->raw_wwn[0] |= 0x10 322 323 /* 324 * wwn_to_ether : Endian independent, copies a WWN to struct ether_addr. 325 * The args to the macro are pointers to WWN and ether_addr structures 326 */ 327 #define wwn_to_ether(W, E) \ 328 bcopy((void *)&((W)->raw_wwn[2]), (void *)E, ETHERADDRL) 329 330 /* 331 * The module_info structure contains identification and limit values. 332 * All queues associated with a certain driver share the same module_info 333 * structures. This structure defines the characteristics of that driver/ 334 * module's queues. The module name must be unique. The max and min packet 335 * sizes limit the no. of characters in M_DATA messages. The Hi and Lo 336 * water marks are for flow control when a module has a service procedure. 337 */ 338 static struct module_info fcipminfo = { 339 FCIPIDNUM, /* mi_idnum : Module ID num */ 340 FCIPNAME, /* mi_idname: Module Name */ 341 FCIPMINPSZ, /* mi_minpsz: Min packet size */ 342 FCIPMAXPSZ, /* mi_maxpsz: Max packet size */ 343 FCIPHIWAT, /* mi_hiwat : High water mark */ 344 FCIPLOWAT /* mi_lowat : Low water mark */ 345 }; 346 347 /* 348 * The qinit structres contain the module put, service. open and close 349 * procedure pointers. All modules and drivers with the same streamtab 350 * file (i.e same fmodsw or cdevsw entry points) point to the same 351 * upstream (read) and downstream (write) qinit structs. 352 */ 353 static struct qinit fcip_rinit = { 354 NULL, /* qi_putp */ 355 NULL, /* qi_srvp */ 356 fcip_open, /* qi_qopen */ 357 fcip_close, /* qi_qclose */ 358 NULL, /* qi_qadmin */ 359 &fcipminfo, /* qi_minfo */ 360 NULL /* qi_mstat */ 361 }; 362 363 static struct qinit fcip_winit = { 364 fcip_wput, /* qi_putp */ 365 fcip_wsrv, /* qi_srvp */ 366 NULL, /* qi_qopen */ 367 NULL, /* qi_qclose */ 368 NULL, /* qi_qadmin */ 369 &fcipminfo, /* qi_minfo */ 370 NULL /* qi_mstat */ 371 }; 372 373 /* 374 * streamtab contains pointers to the read and write qinit structures 375 */ 376 377 static struct streamtab fcip_info = { 378 &fcip_rinit, /* st_rdinit */ 379 &fcip_winit, /* st_wrinit */ 380 NULL, /* st_muxrinit */ 381 NULL, /* st_muxwrinit */ 382 }; 383 384 static struct cb_ops fcip_cb_ops = { 385 nodev, /* open */ 386 nodev, /* close */ 387 nodev, /* strategy */ 388 nodev, /* print */ 389 nodev, /* dump */ 390 nodev, /* read */ 391 nodev, /* write */ 392 nodev, /* ioctl */ 393 nodev, /* devmap */ 394 nodev, /* mmap */ 395 nodev, /* segmap */ 396 nochpoll, /* poll */ 397 ddi_prop_op, /* cb_prop_op */ 398 &fcip_info, /* streamtab */ 399 D_MP | D_HOTPLUG, /* Driver compatibility flag */ 400 CB_REV, /* rev */ 401 nodev, /* int (*cb_aread)() */ 402 nodev /* int (*cb_awrite)() */ 403 }; 404 405 /* 406 * autoconfiguration routines. 407 */ 408 static struct dev_ops fcip_ops = { 409 DEVO_REV, /* devo_rev, */ 410 0, /* refcnt */ 411 fcip_getinfo, /* info */ 412 nulldev, /* identify */ 413 nulldev, /* probe */ 414 fcip_attach, /* attach */ 415 fcip_detach, /* detach */ 416 nodev, /* RESET */ 417 &fcip_cb_ops, /* driver operations */ 418 NULL, /* bus operations */ 419 ddi_power /* power management */ 420 }; 421 422 #define FCIP_VERSION "1.61" 423 #define FCIP_NAME "SunFC FCIP v" FCIP_VERSION 424 425 #define PORT_DRIVER "fp" 426 427 #define GETSTRUCT(struct, number) \ 428 ((struct *)kmem_zalloc((size_t)(sizeof (struct) * (number)), \ 429 KM_SLEEP)) 430 431 static struct modldrv modldrv = { 432 &mod_driverops, /* Type of module - driver */ 433 FCIP_NAME, /* Name of module */ 434 &fcip_ops, /* driver ops */ 435 }; 436 437 static struct modlinkage modlinkage = { 438 MODREV_1, (void *)&modldrv, NULL 439 }; 440 441 442 /* 443 * Now for some global statics 444 */ 445 static uint32_t fcip_ub_nbufs = FCIP_UB_NBUFS; 446 static uint32_t fcip_ub_size = FCIP_UB_SIZE; 447 static int fcip_pkt_ttl_ticks = FCIP_PKT_TTL; 448 static int fcip_tick_incr = 1; 449 static int fcip_wait_cmds = FCIP_WAIT_CMDS; 450 static int fcip_num_attaching = 0; 451 static int fcip_port_attach_pending = 0; 452 static int fcip_create_nodes_on_demand = 1; /* keep it similar to fcp */ 453 static int fcip_cache_on_arp_broadcast = 0; 454 static int fcip_farp_supported = 0; 455 static int fcip_minor_node_created = 0; 456 457 /* 458 * Supported FCAs 459 */ 460 #define QLC_PORT_1_ID_BITS 0x100 461 #define QLC_PORT_2_ID_BITS 0x101 462 #define QLC_PORT_NAA 0x2 463 #define QLC_MODULE_NAME "qlc" 464 #define IS_QLC_PORT(port_dip) \ 465 (strcmp(ddi_driver_name(ddi_get_parent((port_dip))),\ 466 QLC_MODULE_NAME) == 0) 467 468 469 /* 470 * fcip softstate structures head. 471 */ 472 473 static void *fcip_softp = NULL; 474 475 /* 476 * linked list of active (inuse) driver streams 477 */ 478 479 static int fcip_num_instances = 0; 480 static dev_info_t *fcip_module_dip = (dev_info_t *)0; 481 482 483 /* 484 * Ethernet broadcast address: Broadcast addressing in IP over fibre 485 * channel should be the IEEE ULA (also the low 6 bytes of the Port WWN). 486 * 487 * The broadcast addressing varies for differing topologies a node may be in: 488 * - On a private loop the ARP broadcast is a class 3 sequence sent 489 * using OPNfr (Open Broadcast Replicate primitive) followed by 490 * the ARP frame to D_ID 0xFFFFFF 491 * 492 * - On a public Loop the broadcast sequence is sent to AL_PA 0x00 493 * (no OPNfr primitive). 494 * 495 * - For direct attach and point to point topologies we just send 496 * the frame to D_ID 0xFFFFFF 497 * 498 * For public loop the handling would probably be different - for now 499 * I'll just declare this struct - It can be deleted if not necessary. 500 * 501 */ 502 503 504 /* 505 * DL_INFO_ACK template for the fcip module. The dl_info_ack_t structure is 506 * returned as a part of an DL_INFO_ACK message which is a M_PCPROTO message 507 * returned in response to a DL_INFO_REQ message sent to us from a DLS user 508 * Let us fake an ether header as much as possible. 509 * 510 * dl_addr_length is the Provider's DLSAP addr which is SAP addr + 511 * Physical addr of the provider. We set this to 512 * ushort_t + sizeof (la_wwn_t) for Fibre Channel ports. 513 * dl_mac_type Lets just use DL_ETHER - we can try using DL_IPFC, a new 514 * dlpi.h define later. 515 * dl_sap_length -2 indicating the SAP address follows the Physical addr 516 * component in the DLSAP addr. 517 * dl_service_mode: DLCLDS - connectionless data link service. 518 * 519 */ 520 521 static dl_info_ack_t fcip_infoack = { 522 DL_INFO_ACK, /* dl_primitive */ 523 FCIPMTU, /* dl_max_sdu */ 524 0, /* dl_min_sdu */ 525 FCIPADDRL, /* dl_addr_length */ 526 DL_ETHER, /* dl_mac_type */ 527 0, /* dl_reserved */ 528 0, /* dl_current_state */ 529 -2, /* dl_sap_length */ 530 DL_CLDLS, /* dl_service_mode */ 531 0, /* dl_qos_length */ 532 0, /* dl_qos_offset */ 533 0, /* dl_range_length */ 534 0, /* dl_range_offset */ 535 DL_STYLE2, /* dl_provider_style */ 536 sizeof (dl_info_ack_t), /* dl_addr_offset */ 537 DL_VERSION_2, /* dl_version */ 538 ETHERADDRL, /* dl_brdcst_addr_length */ 539 sizeof (dl_info_ack_t) + FCIPADDRL, /* dl_brdcst_addr_offset */ 540 0 /* dl_growth */ 541 }; 542 543 /* 544 * FCIP broadcast address definition. 545 */ 546 static struct ether_addr fcipnhbroadcastaddr = { 547 0xff, 0xff, 0xff, 0xff, 0xff, 0xff 548 }; 549 550 /* 551 * RFC2625 requires the broadcast ARP address in the ARP data payload to 552 * be set to 0x00 00 00 00 00 00 for ARP broadcast packets 553 */ 554 static struct ether_addr fcip_arpbroadcast_addr = { 555 0x00, 0x00, 0x00, 0x00, 0x00, 0x00 556 }; 557 558 559 #define ether_bcopy(src, dest) bcopy((src), (dest), ETHERADDRL); 560 561 /* 562 * global kernel locks 563 */ 564 static kcondvar_t fcip_global_cv; 565 static kmutex_t fcip_global_mutex; 566 567 /* 568 * fctl external defines 569 */ 570 extern int fc_ulp_add(fc_ulp_modinfo_t *); 571 572 /* 573 * fctl data structures 574 */ 575 576 #define FCIP_REV 0x07 577 578 /* linked list of port info structures */ 579 static fcip_port_info_t *fcip_port_head = NULL; 580 581 /* linked list of fcip structures */ 582 static struct fcipstr *fcipstrup = NULL; 583 static krwlock_t fcipstruplock; 584 585 586 /* 587 * Module information structure. This structure gives the FC Transport modules 588 * information about an ULP that registers with it. 589 */ 590 static fc_ulp_modinfo_t fcip_modinfo = { 591 0, /* for xref checks? */ 592 FCTL_ULP_MODREV_4, /* FCIP revision */ 593 FC_TYPE_IS8802_SNAP, /* type 5 for SNAP encapsulated datagrams */ 594 FCIP_NAME, /* module name as in the modldrv struct */ 595 0x0, /* get all statec callbacks for now */ 596 fcip_port_attach, /* port attach callback */ 597 fcip_port_detach, /* port detach callback */ 598 fcip_port_ioctl, /* port ioctl callback */ 599 fcip_els_cb, /* els callback */ 600 fcip_data_cb, /* data callback */ 601 fcip_statec_cb /* state change callback */ 602 }; 603 604 605 /* 606 * Solaris 9 and up, the /kernel/drv/fp.conf file will have the following entry 607 * 608 * ddi-forceattach=1; 609 * 610 * This will ensure that fp is loaded at bootup. No additional checks are needed 611 */ 612 int 613 _init(void) 614 { 615 int rval; 616 617 FCIP_TNF_LOAD(); 618 619 /* 620 * Initialize the mutexs used by port attach and other callbacks. 621 * The transport can call back into our port_attach_callback 622 * routine even before _init() completes and bad things can happen. 623 */ 624 mutex_init(&fcip_global_mutex, NULL, MUTEX_DRIVER, NULL); 625 cv_init(&fcip_global_cv, NULL, CV_DRIVER, NULL); 626 rw_init(&fcipstruplock, NULL, RW_DRIVER, NULL); 627 628 mutex_enter(&fcip_global_mutex); 629 fcip_port_attach_pending = 1; 630 mutex_exit(&fcip_global_mutex); 631 632 /* 633 * Now attempt to register fcip with the transport. 634 * If fc_ulp_add fails, fcip module will not be loaded. 635 */ 636 rval = fc_ulp_add(&fcip_modinfo); 637 if (rval != FC_SUCCESS) { 638 mutex_destroy(&fcip_global_mutex); 639 cv_destroy(&fcip_global_cv); 640 rw_destroy(&fcipstruplock); 641 switch (rval) { 642 case FC_ULP_SAMEMODULE: 643 FCIP_DEBUG(FCIP_DEBUG_DEFAULT, (CE_WARN, 644 "!fcip: module is already registered with" 645 " transport")); 646 rval = EEXIST; 647 break; 648 case FC_ULP_SAMETYPE: 649 FCIP_DEBUG(FCIP_DEBUG_DEFAULT, (CE_WARN, 650 "!fcip: Another module of the same ULP type 0x%x" 651 " is already registered with the transport", 652 fcip_modinfo.ulp_type)); 653 rval = EEXIST; 654 break; 655 case FC_BADULP: 656 FCIP_DEBUG(FCIP_DEBUG_DEFAULT, (CE_WARN, 657 "!fcip: Current fcip version 0x%x does not match" 658 " fctl version", 659 fcip_modinfo.ulp_rev)); 660 rval = ENODEV; 661 break; 662 default: 663 FCIP_DEBUG(FCIP_DEBUG_DEFAULT, (CE_WARN, 664 "!fcip: fc_ulp_add failed with status 0x%x", rval)); 665 rval = ENODEV; 666 break; 667 } 668 FCIP_TNF_UNLOAD(&modlinkage); 669 return (rval); 670 } 671 672 if ((rval = ddi_soft_state_init(&fcip_softp, sizeof (struct fcip), 673 FCIP_NUM_INSTANCES)) != 0) { 674 mutex_destroy(&fcip_global_mutex); 675 cv_destroy(&fcip_global_cv); 676 rw_destroy(&fcipstruplock); 677 (void) fc_ulp_remove(&fcip_modinfo); 678 FCIP_TNF_UNLOAD(&modlinkage); 679 return (rval); 680 } 681 682 if ((rval = mod_install(&modlinkage)) != 0) { 683 FCIP_TNF_UNLOAD(&modlinkage); 684 (void) fc_ulp_remove(&fcip_modinfo); 685 mutex_destroy(&fcip_global_mutex); 686 cv_destroy(&fcip_global_cv); 687 rw_destroy(&fcipstruplock); 688 ddi_soft_state_fini(&fcip_softp); 689 } 690 return (rval); 691 } 692 693 /* 694 * Unload the port driver if this was the only ULP loaded and then 695 * deregister with the transport. 696 */ 697 int 698 _fini(void) 699 { 700 int rval; 701 int rval1; 702 703 /* 704 * Do not permit the module to be unloaded before a port 705 * attach callback has happened. 706 */ 707 mutex_enter(&fcip_global_mutex); 708 if (fcip_num_attaching || fcip_port_attach_pending) { 709 mutex_exit(&fcip_global_mutex); 710 return (EBUSY); 711 } 712 mutex_exit(&fcip_global_mutex); 713 714 if ((rval = mod_remove(&modlinkage)) != 0) { 715 return (rval); 716 } 717 718 /* 719 * unregister with the transport layer 720 */ 721 rval1 = fc_ulp_remove(&fcip_modinfo); 722 723 /* 724 * If the ULP was not registered with the transport, init should 725 * have failed. If transport has no knowledge of our existence 726 * we should simply bail out and succeed 727 */ 728 #ifdef DEBUG 729 if (rval1 == FC_BADULP) { 730 FCIP_DEBUG(FCIP_DEBUG_DEFAULT, (CE_WARN, 731 "fcip: ULP was never registered with the transport")); 732 rval = ENODEV; 733 } else if (rval1 == FC_BADTYPE) { 734 FCIP_DEBUG(FCIP_DEBUG_DEFAULT, (CE_WARN, 735 "fcip: No ULP of this type 0x%x was registered with " 736 "transport", fcip_modinfo.ulp_type)); 737 rval = ENODEV; 738 } 739 #endif /* DEBUG */ 740 741 mutex_destroy(&fcip_global_mutex); 742 rw_destroy(&fcipstruplock); 743 cv_destroy(&fcip_global_cv); 744 ddi_soft_state_fini(&fcip_softp); 745 746 FCIP_TNF_UNLOAD(&modlinkage); 747 748 return (rval); 749 } 750 751 /* 752 * Info about this loadable module 753 */ 754 int 755 _info(struct modinfo *modinfop) 756 { 757 return (mod_info(&modlinkage, modinfop)); 758 } 759 760 /* 761 * The port attach callback is invoked by the port driver when a FCA 762 * port comes online and binds with the transport layer. The transport 763 * then callsback into all ULP modules registered with it. The Port attach 764 * call back will also provide the ULP module with the Port's WWN and S_ID 765 */ 766 /* ARGSUSED */ 767 static int 768 fcip_port_attach(opaque_t ulp_handle, fc_ulp_port_info_t *port_info, 769 fc_attach_cmd_t cmd, uint32_t sid) 770 { 771 int rval = FC_FAILURE; 772 int instance; 773 struct fcip *fptr; 774 fcip_port_info_t *fport = NULL; 775 fcip_port_info_t *cur_fport; 776 fc_portid_t src_id; 777 778 switch (cmd) { 779 case FC_CMD_ATTACH: { 780 la_wwn_t *ww_pn = NULL; 781 /* 782 * It was determined that, as per spec, the lower 48 bits of 783 * the port-WWN will always be unique. This will make the MAC 784 * address (i.e the lower 48 bits of the WWN), that IP/ARP 785 * depend on, unique too. Hence we should be able to remove the 786 * restriction of attaching to only one of the ports of 787 * multi port FCAs. 788 * 789 * Earlier, fcip used to attach only to qlc module and fail 790 * silently for attach failures resulting from unknown FCAs or 791 * unsupported FCA ports. Now, we'll do no such checks. 792 */ 793 ww_pn = &port_info->port_pwwn; 794 795 FCIP_TNF_PROBE_2((fcip_port_attach, "fcip io", /* CSTYLED */, 796 tnf_string, msg, "port id bits", 797 tnf_opaque, nport_id, ww_pn->w.nport_id)); 798 FCIP_DEBUG(FCIP_DEBUG_ATTACH, (CE_NOTE, 799 "port id bits: 0x%x", ww_pn->w.nport_id)); 800 /* 801 * A port has come online 802 */ 803 mutex_enter(&fcip_global_mutex); 804 fcip_num_instances++; 805 fcip_num_attaching++; 806 807 if (fcip_port_head == NULL) { 808 /* OK to sleep here ? */ 809 fport = kmem_zalloc(sizeof (fcip_port_info_t), 810 KM_NOSLEEP); 811 if (fport == NULL) { 812 fcip_num_instances--; 813 fcip_num_attaching--; 814 ASSERT(fcip_num_attaching >= 0); 815 mutex_exit(&fcip_global_mutex); 816 rval = FC_FAILURE; 817 cmn_err(CE_WARN, "!fcip(%d): port attach " 818 "failed: alloc failed", 819 ddi_get_instance(port_info->port_dip)); 820 goto done; 821 } 822 fcip_port_head = fport; 823 } else { 824 /* 825 * traverse the port list and also check for 826 * duplicate port attaches - Nothing wrong in being 827 * paranoid Heh Heh. 828 */ 829 cur_fport = fcip_port_head; 830 while (cur_fport != NULL) { 831 if (cur_fport->fcipp_handle == 832 port_info->port_handle) { 833 fcip_num_instances--; 834 fcip_num_attaching--; 835 ASSERT(fcip_num_attaching >= 0); 836 mutex_exit(&fcip_global_mutex); 837 FCIP_DEBUG(FCIP_DEBUG_ATTACH, (CE_WARN, 838 "!fcip(%d): port already " 839 "attached!!", ddi_get_instance( 840 port_info->port_dip))); 841 rval = FC_FAILURE; 842 goto done; 843 } 844 cur_fport = cur_fport->fcipp_next; 845 } 846 fport = kmem_zalloc(sizeof (fcip_port_info_t), 847 KM_NOSLEEP); 848 if (fport == NULL) { 849 rval = FC_FAILURE; 850 fcip_num_instances--; 851 fcip_num_attaching--; 852 ASSERT(fcip_num_attaching >= 0); 853 mutex_exit(&fcip_global_mutex); 854 cmn_err(CE_WARN, "!fcip(%d): port attach " 855 "failed: alloc failed", 856 ddi_get_instance(port_info->port_dip)); 857 goto done; 858 } 859 fport->fcipp_next = fcip_port_head; 860 fcip_port_head = fport; 861 } 862 863 mutex_exit(&fcip_global_mutex); 864 865 /* 866 * now fill in the details about the port itself 867 */ 868 fport->fcipp_linkage = *port_info->port_linkage; 869 fport->fcipp_handle = port_info->port_handle; 870 fport->fcipp_dip = port_info->port_dip; 871 fport->fcipp_topology = port_info->port_flags; 872 fport->fcipp_pstate = port_info->port_state; 873 fport->fcipp_naa = port_info->port_pwwn.w.naa_id; 874 bcopy(&port_info->port_pwwn, &fport->fcipp_pwwn, 875 sizeof (la_wwn_t)); 876 bcopy(&port_info->port_nwwn, &fport->fcipp_nwwn, 877 sizeof (la_wwn_t)); 878 fport->fcipp_fca_pkt_size = port_info->port_fca_pkt_size; 879 fport->fcipp_cmd_dma_attr = *port_info->port_cmd_dma_attr; 880 fport->fcipp_resp_dma_attr = *port_info->port_resp_dma_attr; 881 fport->fcipp_fca_acc_attr = *port_info->port_acc_attr; 882 src_id.port_id = sid; 883 src_id.priv_lilp_posit = 0; 884 fport->fcipp_sid = src_id; 885 886 /* 887 * allocate soft state for this instance 888 */ 889 instance = ddi_get_instance(fport->fcipp_dip); 890 if (ddi_soft_state_zalloc(fcip_softp, 891 instance) != DDI_SUCCESS) { 892 rval = FC_FAILURE; 893 cmn_err(CE_WARN, "!fcip(%d): port attach failed: " 894 "soft state alloc failed", instance); 895 goto failure; 896 } 897 898 fptr = ddi_get_soft_state(fcip_softp, instance); 899 900 if (fptr == NULL) { 901 rval = FC_FAILURE; 902 cmn_err(CE_WARN, "!fcip(%d): port attach failed: " 903 "failure to get soft state", instance); 904 goto failure; 905 } 906 907 /* 908 * initialize all mutexes and locks required for this module 909 */ 910 mutex_init(&fptr->fcip_mutex, NULL, MUTEX_DRIVER, NULL); 911 mutex_init(&fptr->fcip_ub_mutex, NULL, MUTEX_DRIVER, NULL); 912 mutex_init(&fptr->fcip_rt_mutex, NULL, MUTEX_DRIVER, NULL); 913 mutex_init(&fptr->fcip_dest_mutex, NULL, MUTEX_DRIVER, NULL); 914 mutex_init(&fptr->fcip_sendup_mutex, NULL, MUTEX_DRIVER, NULL); 915 cv_init(&fptr->fcip_farp_cv, NULL, CV_DRIVER, NULL); 916 cv_init(&fptr->fcip_sendup_cv, NULL, CV_DRIVER, NULL); 917 cv_init(&fptr->fcip_ub_cv, NULL, CV_DRIVER, NULL); 918 919 mutex_enter(&fptr->fcip_mutex); 920 921 fptr->fcip_dip = fport->fcipp_dip; /* parent's dip */ 922 fptr->fcip_instance = instance; 923 fptr->fcip_ub_upstream = 0; 924 925 if (FC_PORT_STATE_MASK(port_info->port_state) == 926 FC_STATE_ONLINE) { 927 fptr->fcip_port_state = FCIP_PORT_ONLINE; 928 if (fptr->fcip_flags & FCIP_LINK_DOWN) { 929 fptr->fcip_flags &= ~FCIP_LINK_DOWN; 930 } 931 } else { 932 fptr->fcip_port_state = FCIP_PORT_OFFLINE; 933 } 934 935 fptr->fcip_flags |= FCIP_ATTACHING; 936 fptr->fcip_port_info = fport; 937 938 /* 939 * Extract our MAC addr from our port's WWN. The lower 48 940 * bits will be our MAC address 941 */ 942 wwn_to_ether(&fport->fcipp_nwwn, &fptr->fcip_macaddr); 943 944 fport->fcipp_fcip = fptr; 945 946 FCIP_DEBUG(FCIP_DEBUG_ATTACH, 947 (CE_NOTE, "fcipdest : 0x%lx, rtable : 0x%lx", 948 (long)(sizeof (fptr->fcip_dest)), 949 (long)(sizeof (fptr->fcip_rtable)))); 950 951 bzero(fptr->fcip_dest, sizeof (fptr->fcip_dest)); 952 bzero(fptr->fcip_rtable, sizeof (fptr->fcip_rtable)); 953 954 /* 955 * create a taskq to handle sundry jobs for the driver 956 * This way we can have jobs run in parallel 957 */ 958 fptr->fcip_tq = taskq_create("fcip_tasks", 959 FCIP_NUM_THREADS, MINCLSYSPRI, FCIP_MIN_TASKS, 960 FCIP_MAX_TASKS, TASKQ_PREPOPULATE); 961 962 mutex_exit(&fptr->fcip_mutex); 963 964 /* 965 * create a separate thread to handle all unsolicited 966 * callback handling. This is because unsolicited_callback 967 * can happen from an interrupt context and the upstream 968 * modules can put new messages right back in the same 969 * thread context. This usually works fine, but sometimes 970 * we may have to block to obtain the dest struct entries 971 * for some remote ports. 972 */ 973 mutex_enter(&fptr->fcip_sendup_mutex); 974 if (thread_create(NULL, DEFAULTSTKSZ, 975 (void (*)())fcip_sendup_thr, (caddr_t)fptr, 0, &p0, 976 TS_RUN, minclsyspri) == NULL) { 977 mutex_exit(&fptr->fcip_sendup_mutex); 978 cmn_err(CE_WARN, 979 "!unable to create fcip sendup thread for " 980 " instance: 0x%x", instance); 981 rval = FC_FAILURE; 982 goto done; 983 } 984 fptr->fcip_sendup_thr_initted = 1; 985 fptr->fcip_sendup_head = fptr->fcip_sendup_tail = NULL; 986 mutex_exit(&fptr->fcip_sendup_mutex); 987 988 989 /* Let the attach handler do the rest */ 990 if (fcip_port_attach_handler(fptr) != FC_SUCCESS) { 991 /* 992 * We have already cleaned up so return 993 */ 994 rval = FC_FAILURE; 995 cmn_err(CE_WARN, "!fcip(%d): port attach failed", 996 instance); 997 goto done; 998 } 999 1000 FCIP_DEBUG(FCIP_DEBUG_ATTACH, (CE_CONT, 1001 "!fcip attach for port instance (0x%x) successful", 1002 instance)); 1003 1004 rval = FC_SUCCESS; 1005 goto done; 1006 } 1007 case FC_CMD_POWER_UP: 1008 /* FALLTHROUGH */ 1009 case FC_CMD_RESUME: 1010 mutex_enter(&fcip_global_mutex); 1011 fport = fcip_port_head; 1012 while (fport != NULL) { 1013 if (fport->fcipp_handle == port_info->port_handle) { 1014 break; 1015 } 1016 fport = fport->fcipp_next; 1017 } 1018 if (fport == NULL) { 1019 rval = FC_SUCCESS; 1020 mutex_exit(&fcip_global_mutex); 1021 goto done; 1022 } 1023 rval = fcip_handle_resume(fport, port_info, cmd); 1024 mutex_exit(&fcip_global_mutex); 1025 goto done; 1026 1027 default: 1028 FCIP_TNF_PROBE_2((fcip_port_attach, "fcip io", /* CSTYLED */, 1029 tnf_string, msg, "unknown command type", 1030 tnf_uint, cmd, cmd)); 1031 FCIP_DEBUG(FCIP_DEBUG_ATTACH, (CE_WARN, 1032 "unknown cmd type 0x%x in port_attach", cmd)); 1033 rval = FC_FAILURE; 1034 goto done; 1035 } 1036 1037 failure: 1038 if (fport) { 1039 mutex_enter(&fcip_global_mutex); 1040 fcip_num_attaching--; 1041 ASSERT(fcip_num_attaching >= 0); 1042 (void) fcip_softstate_free(fport); 1043 fcip_port_attach_pending = 0; 1044 mutex_exit(&fcip_global_mutex); 1045 } 1046 return (rval); 1047 1048 done: 1049 mutex_enter(&fcip_global_mutex); 1050 fcip_port_attach_pending = 0; 1051 mutex_exit(&fcip_global_mutex); 1052 return (rval); 1053 } 1054 1055 /* 1056 * fcip_port_attach_handler : Completes the port attach operation after 1057 * the ulp_port_attach routine has completed its ground work. The job 1058 * of this function among other things is to obtain and handle topology 1059 * specifics, initialize a port, setup broadcast address entries in 1060 * the fcip tables etc. This routine cleans up behind itself on failures. 1061 * Returns FC_SUCCESS or FC_FAILURE. 1062 */ 1063 static int 1064 fcip_port_attach_handler(struct fcip *fptr) 1065 { 1066 fcip_port_info_t *fport = fptr->fcip_port_info; 1067 int rval = FC_FAILURE; 1068 1069 ASSERT(fport != NULL); 1070 1071 mutex_enter(&fcip_global_mutex); 1072 1073 FCIP_DEBUG(FCIP_DEBUG_ATTACH, (CE_NOTE, 1074 "fcip module dip: %p instance: %d", 1075 (void *)fcip_module_dip, ddi_get_instance(fptr->fcip_dip))); 1076 1077 if (fcip_module_dip == NULL) { 1078 clock_t fcip_lbolt; 1079 1080 fcip_lbolt = ddi_get_lbolt(); 1081 /* 1082 * we need to use the fcip devinfo for creating 1083 * the clone device node, but the fcip attach 1084 * (from its conf file entry claiming to be a 1085 * child of pseudo) may not have happened yet. 1086 * wait here for 10 seconds and fail port attach 1087 * if the fcip devinfo is not attached yet 1088 */ 1089 fcip_lbolt += drv_usectohz(FCIP_INIT_DELAY); 1090 1091 FCIP_DEBUG(FCIP_DEBUG_ATTACH, 1092 (CE_WARN, "cv_timedwait lbolt %lx", fcip_lbolt)); 1093 1094 (void) cv_timedwait(&fcip_global_cv, &fcip_global_mutex, 1095 fcip_lbolt); 1096 1097 if (fcip_module_dip == NULL) { 1098 mutex_exit(&fcip_global_mutex); 1099 1100 FCIP_DEBUG(FCIP_DEBUG_ATTACH, (CE_WARN, 1101 "fcip attach did not happen")); 1102 goto port_attach_cleanup; 1103 } 1104 } 1105 1106 if ((!fcip_minor_node_created) && 1107 fcip_is_supported_fc_topology(fport->fcipp_topology)) { 1108 /* 1109 * Checking for same topologies which are considered valid 1110 * by fcip_handle_topology(). Dont create a minor node if 1111 * nothing is hanging off the FC port. 1112 */ 1113 if (ddi_create_minor_node(fcip_module_dip, "fcip", S_IFCHR, 1114 ddi_get_instance(fptr->fcip_dip), DDI_PSEUDO, 1115 CLONE_DEV) == DDI_FAILURE) { 1116 mutex_exit(&fcip_global_mutex); 1117 FCIP_DEBUG(FCIP_DEBUG_ATTACH, (CE_WARN, 1118 "failed to create minor node for fcip(%d)", 1119 ddi_get_instance(fptr->fcip_dip))); 1120 goto port_attach_cleanup; 1121 } 1122 fcip_minor_node_created++; 1123 } 1124 mutex_exit(&fcip_global_mutex); 1125 1126 /* 1127 * initialize port for traffic 1128 */ 1129 if (fcip_init_port(fptr) != FC_SUCCESS) { 1130 /* fcip_init_port has already cleaned up its stuff */ 1131 1132 mutex_enter(&fcip_global_mutex); 1133 1134 if ((fcip_num_instances == 1) && 1135 (fcip_minor_node_created == 1)) { 1136 /* Remove minor node iff this is the last instance */ 1137 ddi_remove_minor_node(fcip_module_dip, NULL); 1138 } 1139 1140 mutex_exit(&fcip_global_mutex); 1141 1142 goto port_attach_cleanup; 1143 } 1144 1145 mutex_enter(&fptr->fcip_mutex); 1146 fptr->fcip_flags &= ~FCIP_ATTACHING; 1147 fptr->fcip_flags |= FCIP_INITED; 1148 fptr->fcip_timeout_ticks = 0; 1149 1150 /* 1151 * start the timeout threads 1152 */ 1153 fptr->fcip_timeout_id = timeout(fcip_timeout, fptr, 1154 drv_usectohz(1000000)); 1155 1156 mutex_exit(&fptr->fcip_mutex); 1157 mutex_enter(&fcip_global_mutex); 1158 fcip_num_attaching--; 1159 ASSERT(fcip_num_attaching >= 0); 1160 mutex_exit(&fcip_global_mutex); 1161 rval = FC_SUCCESS; 1162 return (rval); 1163 1164 port_attach_cleanup: 1165 mutex_enter(&fcip_global_mutex); 1166 (void) fcip_softstate_free(fport); 1167 fcip_num_attaching--; 1168 ASSERT(fcip_num_attaching >= 0); 1169 mutex_exit(&fcip_global_mutex); 1170 rval = FC_FAILURE; 1171 return (rval); 1172 } 1173 1174 1175 /* 1176 * Handler for DDI_RESUME operations. Port must be ready to restart IP 1177 * traffic on resume 1178 */ 1179 static int 1180 fcip_handle_resume(fcip_port_info_t *fport, fc_ulp_port_info_t *port_info, 1181 fc_attach_cmd_t cmd) 1182 { 1183 int rval = FC_SUCCESS; 1184 struct fcip *fptr = fport->fcipp_fcip; 1185 struct fcipstr *tslp; 1186 int index; 1187 1188 1189 ASSERT(fptr != NULL); 1190 1191 mutex_enter(&fptr->fcip_mutex); 1192 1193 if (cmd == FC_CMD_POWER_UP) { 1194 fptr->fcip_flags &= ~(FCIP_POWER_DOWN); 1195 if (fptr->fcip_flags & FCIP_SUSPENDED) { 1196 mutex_exit(&fptr->fcip_mutex); 1197 return (FC_SUCCESS); 1198 } 1199 } else if (cmd == FC_CMD_RESUME) { 1200 fptr->fcip_flags &= ~(FCIP_SUSPENDED); 1201 } else { 1202 mutex_exit(&fptr->fcip_mutex); 1203 return (FC_FAILURE); 1204 } 1205 1206 /* 1207 * set the current port state and topology 1208 */ 1209 fport->fcipp_topology = port_info->port_flags; 1210 fport->fcipp_pstate = port_info->port_state; 1211 1212 rw_enter(&fcipstruplock, RW_READER); 1213 for (tslp = fcipstrup; tslp; tslp = tslp->sl_nextp) { 1214 if (tslp->sl_fcip == fptr) { 1215 break; 1216 } 1217 } 1218 rw_exit(&fcipstruplock); 1219 1220 /* 1221 * No active streams on this port 1222 */ 1223 if (tslp == NULL) { 1224 rval = FC_SUCCESS; 1225 goto done; 1226 } 1227 1228 mutex_enter(&fptr->fcip_rt_mutex); 1229 for (index = 0; index < FCIP_RT_HASH_ELEMS; index++) { 1230 struct fcip_routing_table *frp; 1231 1232 frp = fptr->fcip_rtable[index]; 1233 while (frp) { 1234 uint32_t did; 1235 /* 1236 * Mark the broadcast RTE available again. It 1237 * was marked SUSPENDED during SUSPEND. 1238 */ 1239 did = fcip_get_broadcast_did(fptr); 1240 if (frp->fcipr_d_id.port_id == did) { 1241 frp->fcipr_state = 0; 1242 index = FCIP_RT_HASH_ELEMS; 1243 break; 1244 } 1245 frp = frp->fcipr_next; 1246 } 1247 } 1248 mutex_exit(&fptr->fcip_rt_mutex); 1249 1250 /* 1251 * fcip_handle_topology will update the port entries in the 1252 * routing table. 1253 * fcip_handle_topology also takes care of resetting the 1254 * fcipr_state field in the routing table structure. The entries 1255 * were set to RT_INVALID during suspend. 1256 */ 1257 fcip_handle_topology(fptr); 1258 1259 done: 1260 /* 1261 * Restart the timeout thread 1262 */ 1263 fptr->fcip_timeout_id = timeout(fcip_timeout, fptr, 1264 drv_usectohz(1000000)); 1265 mutex_exit(&fptr->fcip_mutex); 1266 return (rval); 1267 } 1268 1269 1270 /* 1271 * Insert a destination port entry into the routing table for 1272 * this port 1273 */ 1274 static void 1275 fcip_rt_update(struct fcip *fptr, fc_portmap_t *devlist, uint32_t listlen) 1276 { 1277 struct fcip_routing_table *frp; 1278 fcip_port_info_t *fport = fptr->fcip_port_info; 1279 int hash_bucket, i; 1280 fc_portmap_t *pmap; 1281 char wwn_buf[20]; 1282 1283 FCIP_TNF_PROBE_2((fcip_rt_update, "fcip io", /* CSTYLED */, 1284 tnf_string, msg, "enter", 1285 tnf_int, listlen, listlen)); 1286 1287 ASSERT(!mutex_owned(&fptr->fcip_mutex)); 1288 mutex_enter(&fptr->fcip_rt_mutex); 1289 1290 for (i = 0; i < listlen; i++) { 1291 pmap = &(devlist[i]); 1292 1293 frp = fcip_lookup_rtable(fptr, &(pmap->map_pwwn), 1294 FCIP_COMPARE_PWWN); 1295 /* 1296 * If an entry for a port in the devlist exists in the 1297 * in the per port routing table, make sure the data 1298 * is current. We need to do this irrespective of the 1299 * underlying port topology. 1300 */ 1301 switch (pmap->map_type) { 1302 /* FALLTHROUGH */ 1303 case PORT_DEVICE_NOCHANGE: 1304 /* FALLTHROUGH */ 1305 case PORT_DEVICE_USER_LOGIN: 1306 /* FALLTHROUGH */ 1307 case PORT_DEVICE_CHANGED: 1308 /* FALLTHROUGH */ 1309 case PORT_DEVICE_NEW: 1310 if (frp == NULL) { 1311 goto add_new_entry; 1312 } else if (frp) { 1313 goto update_entry; 1314 } else { 1315 continue; 1316 } 1317 1318 case PORT_DEVICE_OLD: 1319 /* FALLTHROUGH */ 1320 case PORT_DEVICE_USER_LOGOUT: 1321 /* 1322 * Mark entry for removal from Routing Table if 1323 * one exists. Let the timeout thread actually 1324 * remove the entry after we've given up hopes 1325 * of the port ever showing up. 1326 */ 1327 if (frp) { 1328 uint32_t did; 1329 1330 /* 1331 * Mark the routing table as invalid to bail 1332 * the packets early that are in transit 1333 */ 1334 did = fptr->fcip_broadcast_did; 1335 if (frp->fcipr_d_id.port_id != did) { 1336 frp->fcipr_pd = NULL; 1337 frp->fcipr_state = FCIP_RT_INVALID; 1338 frp->fcipr_invalid_timeout = 1339 fptr->fcip_timeout_ticks + 1340 FCIP_RTE_TIMEOUT; 1341 } 1342 } 1343 continue; 1344 1345 default: 1346 FCIP_DEBUG(FCIP_DEBUG_INIT, (CE_WARN, 1347 "unknown map flags in rt_update")); 1348 continue; 1349 } 1350 add_new_entry: 1351 ASSERT(frp == NULL); 1352 hash_bucket = FCIP_RT_HASH(pmap->map_pwwn.raw_wwn); 1353 1354 ASSERT(hash_bucket < FCIP_RT_HASH_ELEMS); 1355 1356 FCIP_TNF_PROBE_2((fcip_rt_update, "cfip io", /* CSTYLED */, 1357 tnf_string, msg, 1358 "add new entry", 1359 tnf_int, hashbucket, hash_bucket)); 1360 1361 frp = (struct fcip_routing_table *) 1362 kmem_zalloc(sizeof (struct fcip_routing_table), KM_SLEEP); 1363 /* insert at beginning of hash bucket */ 1364 frp->fcipr_next = fptr->fcip_rtable[hash_bucket]; 1365 fptr->fcip_rtable[hash_bucket] = frp; 1366 fc_wwn_to_str(&pmap->map_pwwn, wwn_buf); 1367 FCIP_DEBUG(FCIP_DEBUG_ATTACH, (CE_NOTE, 1368 "added entry for pwwn %s and d_id 0x%x", 1369 wwn_buf, pmap->map_did.port_id)); 1370 update_entry: 1371 bcopy((void *)&pmap->map_pwwn, 1372 (void *)&frp->fcipr_pwwn, sizeof (la_wwn_t)); 1373 bcopy((void *)&pmap->map_nwwn, (void *)&frp->fcipr_nwwn, 1374 sizeof (la_wwn_t)); 1375 frp->fcipr_d_id = pmap->map_did; 1376 frp->fcipr_state = pmap->map_state; 1377 frp->fcipr_pd = pmap->map_pd; 1378 1379 /* 1380 * If there is no pd for a destination port that is not 1381 * a broadcast entry, the port is pretty much unusable - so 1382 * mark the port for removal so we can try adding back the 1383 * entry again. 1384 */ 1385 if ((frp->fcipr_pd == NULL) && 1386 (frp->fcipr_d_id.port_id != fptr->fcip_broadcast_did)) { 1387 frp->fcipr_state = PORT_DEVICE_INVALID; 1388 frp->fcipr_invalid_timeout = fptr->fcip_timeout_ticks + 1389 (FCIP_RTE_TIMEOUT / 2); 1390 } 1391 frp->fcipr_fca_dev = 1392 fc_ulp_get_fca_device(fport->fcipp_handle, pmap->map_did); 1393 1394 /* 1395 * login to the remote port. Don't worry about 1396 * plogi failures for now 1397 */ 1398 if (pmap->map_pd != NULL) { 1399 (void) fcip_do_plogi(fptr, frp); 1400 } else if (FC_TOP_EXTERNAL(fport->fcipp_topology)) { 1401 fc_wwn_to_str(&frp->fcipr_pwwn, wwn_buf); 1402 FCIP_DEBUG(FCIP_DEBUG_MISC, (CE_NOTE, 1403 "logging into pwwn %s, d_id 0x%x", 1404 wwn_buf, frp->fcipr_d_id.port_id)); 1405 (void) fcip_do_plogi(fptr, frp); 1406 } 1407 1408 FCIP_TNF_BYTE_ARRAY(fcip_rt_update, "fcip io", "detail", 1409 "new wwn in rt", pwwn, 1410 &frp->fcipr_pwwn, sizeof (la_wwn_t)); 1411 } 1412 mutex_exit(&fptr->fcip_rt_mutex); 1413 } 1414 1415 1416 /* 1417 * return a matching routing table entry for a given fcip instance 1418 */ 1419 struct fcip_routing_table * 1420 fcip_lookup_rtable(struct fcip *fptr, la_wwn_t *wwn, int matchflag) 1421 { 1422 struct fcip_routing_table *frp = NULL; 1423 int hash_bucket; 1424 1425 1426 FCIP_TNF_PROBE_1((fcip_lookup_rtable, "fcip io", /* CSTYLED */, 1427 tnf_string, msg, "enter")); 1428 FCIP_TNF_BYTE_ARRAY(fcip_lookup_rtable, "fcip io", "detail", 1429 "rtable lookup for", wwn, 1430 &wwn->raw_wwn, sizeof (la_wwn_t)); 1431 FCIP_TNF_PROBE_2((fcip_lookup_rtable, "fcip io", /* CSTYLED */, 1432 tnf_string, msg, "match by", 1433 tnf_int, matchflag, matchflag)); 1434 1435 ASSERT(mutex_owned(&fptr->fcip_rt_mutex)); 1436 1437 hash_bucket = FCIP_RT_HASH(wwn->raw_wwn); 1438 frp = fptr->fcip_rtable[hash_bucket]; 1439 while (frp != NULL) { 1440 1441 FCIP_TNF_BYTE_ARRAY(fcip_lookup_rtable, "fcip io", "detail", 1442 "rtable entry", nwwn, 1443 &(frp->fcipr_nwwn.raw_wwn), sizeof (la_wwn_t)); 1444 1445 if (fcip_wwn_compare(&frp->fcipr_pwwn, wwn, matchflag) == 0) { 1446 break; 1447 } 1448 1449 frp = frp->fcipr_next; 1450 } 1451 FCIP_TNF_PROBE_2((fcip_lookup_rtable, "fcip io", /* CSTYLED */, 1452 tnf_string, msg, "lookup result", 1453 tnf_opaque, frp, frp)); 1454 return (frp); 1455 } 1456 1457 /* 1458 * Attach of fcip under pseudo. The actual setup of the interface 1459 * actually happens in fcip_port_attach on a callback from the 1460 * transport. The port_attach callback however can proceed only 1461 * after the devinfo for fcip has been created under pseudo 1462 */ 1463 static int 1464 fcip_attach(dev_info_t *dip, ddi_attach_cmd_t cmd) 1465 { 1466 switch ((int)cmd) { 1467 1468 case DDI_ATTACH: { 1469 ASSERT(fcip_module_dip == NULL); 1470 fcip_module_dip = dip; 1471 1472 /* 1473 * this call originates as a result of fcip's conf 1474 * file entry and will result in a fcip instance being 1475 * a child of pseudo. We should ensure here that the port 1476 * driver (fp) has been loaded and initted since we would 1477 * never get a port attach callback without fp being loaded. 1478 * If we are unable to succesfully load and initalize fp - 1479 * just fail this attach. 1480 */ 1481 mutex_enter(&fcip_global_mutex); 1482 1483 FCIP_DEBUG(FCIP_DEBUG_ATTACH, 1484 (CE_WARN, "global cv - signaling")); 1485 1486 cv_signal(&fcip_global_cv); 1487 1488 FCIP_DEBUG(FCIP_DEBUG_ATTACH, 1489 (CE_WARN, "global cv - signaled")); 1490 mutex_exit(&fcip_global_mutex); 1491 return (DDI_SUCCESS); 1492 } 1493 case DDI_RESUME: 1494 /* 1495 * Resume appears trickier 1496 */ 1497 return (DDI_SUCCESS); 1498 default: 1499 return (DDI_FAILURE); 1500 } 1501 } 1502 1503 1504 /* 1505 * The detach entry point to permit unloading fcip. We make sure 1506 * there are no active streams before we proceed with the detach 1507 */ 1508 /* ARGSUSED */ 1509 static int 1510 fcip_detach(dev_info_t *dip, ddi_detach_cmd_t cmd) 1511 { 1512 struct fcip *fptr; 1513 fcip_port_info_t *fport; 1514 int detached; 1515 1516 switch (cmd) { 1517 case DDI_DETACH: { 1518 /* 1519 * If we got here, any active streams should have been 1520 * unplumbed but check anyway 1521 */ 1522 mutex_enter(&fcip_global_mutex); 1523 if (fcipstrup != NULL) { 1524 mutex_exit(&fcip_global_mutex); 1525 return (DDI_FAILURE); 1526 } 1527 1528 if (fcip_port_head != NULL) { 1529 /* 1530 * Check to see if we have unattached/unbound 1531 * ports. If all the ports are unattached/unbound go 1532 * ahead and unregister with the transport 1533 */ 1534 fport = fcip_port_head; 1535 while (fport != NULL) { 1536 fptr = fport->fcipp_fcip; 1537 if (fptr == NULL) { 1538 continue; 1539 } 1540 mutex_enter(&fptr->fcip_mutex); 1541 fptr->fcip_flags |= FCIP_DETACHING; 1542 if (fptr->fcip_ipq || 1543 fptr->fcip_flags & (FCIP_IN_TIMEOUT | 1544 FCIP_IN_CALLBACK | FCIP_ATTACHING | 1545 FCIP_SUSPENDED | FCIP_POWER_DOWN | 1546 FCIP_REG_INPROGRESS)) { 1547 FCIP_TNF_PROBE_1((fcip_detach, 1548 "fcip io", /* CSTYLED */, 1549 tnf_string, msg, 1550 "fcip instance busy")); 1551 1552 mutex_exit(&fptr->fcip_mutex); 1553 FCIP_DEBUG(FCIP_DEBUG_DETACH, (CE_WARN, 1554 "fcip instance busy")); 1555 break; 1556 } 1557 /* 1558 * Check for any outstanding pkts. If yes 1559 * fail the detach 1560 */ 1561 mutex_enter(&fptr->fcip_dest_mutex); 1562 if (fcip_port_get_num_pkts(fptr) > 0) { 1563 mutex_exit(&fptr->fcip_dest_mutex); 1564 mutex_exit(&fptr->fcip_mutex); 1565 FCIP_DEBUG(FCIP_DEBUG_DETACH, (CE_WARN, 1566 "fcip instance busy - pkts " 1567 "pending")); 1568 break; 1569 } 1570 mutex_exit(&fptr->fcip_dest_mutex); 1571 1572 mutex_enter(&fptr->fcip_rt_mutex); 1573 if (fcip_plogi_in_progress(fptr)) { 1574 mutex_exit(&fptr->fcip_rt_mutex); 1575 mutex_exit(&fptr->fcip_mutex); 1576 FCIP_DEBUG(FCIP_DEBUG_DETACH, (CE_WARN, 1577 "fcip instance busy - plogi in " 1578 "progress")); 1579 break; 1580 } 1581 mutex_exit(&fptr->fcip_rt_mutex); 1582 1583 mutex_exit(&fptr->fcip_mutex); 1584 fport = fport->fcipp_next; 1585 } 1586 /* 1587 * if fport is non NULL - we have active ports 1588 */ 1589 if (fport != NULL) { 1590 /* 1591 * Remove the DETACHING flags on the ports 1592 */ 1593 fport = fcip_port_head; 1594 while (fport != NULL) { 1595 fptr = fport->fcipp_fcip; 1596 mutex_enter(&fptr->fcip_mutex); 1597 fptr->fcip_flags &= ~(FCIP_DETACHING); 1598 mutex_exit(&fptr->fcip_mutex); 1599 fport = fport->fcipp_next; 1600 } 1601 mutex_exit(&fcip_global_mutex); 1602 return (DDI_FAILURE); 1603 } 1604 } 1605 1606 /* 1607 * free up all softstate structures 1608 */ 1609 fport = fcip_port_head; 1610 while (fport != NULL) { 1611 detached = 1; 1612 1613 fptr = fport->fcipp_fcip; 1614 if (fptr) { 1615 mutex_enter(&fptr->fcip_mutex); 1616 /* 1617 * Check to see if somebody beat us to the 1618 * punch 1619 */ 1620 detached = fptr->fcip_flags & FCIP_DETACHED; 1621 fptr->fcip_flags &= ~(FCIP_DETACHING); 1622 fptr->fcip_flags |= FCIP_DETACHED; 1623 mutex_exit(&fptr->fcip_mutex); 1624 } 1625 1626 if (!detached) { 1627 fport = fcip_softstate_free(fport); 1628 } else { 1629 /* 1630 * If the port was marked as detached 1631 * but it was still in the list, that 1632 * means another thread has marked it 1633 * but we got in while it released the 1634 * fcip_global_mutex in softstate_free. 1635 * Given that, we're still safe to use 1636 * fport->fcipp_next to find out what 1637 * the next port on the list is. 1638 */ 1639 fport = fport->fcipp_next; 1640 } 1641 1642 FCIP_DEBUG(FCIP_DEBUG_DETACH, 1643 (CE_NOTE, "detaching port")); 1644 1645 FCIP_TNF_PROBE_1((fcip_detach, 1646 "fcip io", /* CSTYLED */, tnf_string, 1647 msg, "detaching port")); 1648 } 1649 1650 /* 1651 * If we haven't removed all the port structures, we 1652 * aren't yet ready to be detached. 1653 */ 1654 if (fcip_port_head != NULL) { 1655 mutex_exit(&fcip_global_mutex); 1656 return (DDI_FAILURE); 1657 } 1658 1659 fcip_num_instances = 0; 1660 mutex_exit(&fcip_global_mutex); 1661 fcip_module_dip = NULL; 1662 return (DDI_SUCCESS); 1663 } 1664 case DDI_SUSPEND: 1665 return (DDI_SUCCESS); 1666 default: 1667 return (DDI_FAILURE); 1668 } 1669 } 1670 1671 /* 1672 * The port_detach callback is called from the transport when a 1673 * FC port is being removed from the transport's control. This routine 1674 * provides fcip with an opportunity to cleanup all activities and 1675 * structures on the port marked for removal. 1676 */ 1677 /* ARGSUSED */ 1678 static int 1679 fcip_port_detach(opaque_t ulp_handle, fc_ulp_port_info_t *port_info, 1680 fc_detach_cmd_t cmd) 1681 { 1682 int rval = FC_FAILURE; 1683 fcip_port_info_t *fport; 1684 struct fcip *fptr; 1685 struct fcipstr *strp; 1686 1687 switch (cmd) { 1688 case FC_CMD_DETACH: { 1689 mutex_enter(&fcip_global_mutex); 1690 1691 if (fcip_port_head == NULL) { 1692 /* 1693 * we are all done but our fini has not been 1694 * called yet!! Let's hope we have no active 1695 * fcip instances here. - strange secnario but 1696 * no harm in having this return a success. 1697 */ 1698 fcip_check_remove_minor_node(); 1699 1700 mutex_exit(&fcip_global_mutex); 1701 return (FC_SUCCESS); 1702 } else { 1703 /* 1704 * traverse the port list 1705 */ 1706 fport = fcip_port_head; 1707 while (fport != NULL) { 1708 if (fport->fcipp_handle == 1709 port_info->port_handle) { 1710 fptr = fport->fcipp_fcip; 1711 1712 /* 1713 * Fail the port detach if there is 1714 * still an attached, bound stream on 1715 * this interface. 1716 */ 1717 1718 rw_enter(&fcipstruplock, RW_READER); 1719 1720 for (strp = fcipstrup; strp != NULL; 1721 strp = strp->sl_nextp) { 1722 if (strp->sl_fcip == fptr) { 1723 rw_exit(&fcipstruplock); 1724 mutex_exit( 1725 &fcip_global_mutex); 1726 return (FC_FAILURE); 1727 } 1728 } 1729 1730 rw_exit(&fcipstruplock); 1731 1732 /* 1733 * fail port detach if we are in 1734 * the middle of a deferred port attach 1735 * or if the port has outstanding pkts 1736 */ 1737 if (fptr != NULL) { 1738 mutex_enter(&fptr->fcip_mutex); 1739 if (fcip_check_port_busy 1740 (fptr) || 1741 (fptr->fcip_flags & 1742 FCIP_DETACHED)) { 1743 mutex_exit( 1744 &fptr->fcip_mutex); 1745 mutex_exit( 1746 &fcip_global_mutex); 1747 return (FC_FAILURE); 1748 } 1749 1750 fptr->fcip_flags |= 1751 FCIP_DETACHED; 1752 mutex_exit(&fptr->fcip_mutex); 1753 } 1754 (void) fcip_softstate_free(fport); 1755 1756 fcip_check_remove_minor_node(); 1757 mutex_exit(&fcip_global_mutex); 1758 return (FC_SUCCESS); 1759 } 1760 fport = fport->fcipp_next; 1761 } 1762 ASSERT(fport == NULL); 1763 } 1764 mutex_exit(&fcip_global_mutex); 1765 break; 1766 } 1767 case FC_CMD_POWER_DOWN: 1768 /* FALLTHROUGH */ 1769 case FC_CMD_SUSPEND: 1770 mutex_enter(&fcip_global_mutex); 1771 fport = fcip_port_head; 1772 while (fport != NULL) { 1773 if (fport->fcipp_handle == port_info->port_handle) { 1774 break; 1775 } 1776 fport = fport->fcipp_next; 1777 } 1778 if (fport == NULL) { 1779 mutex_exit(&fcip_global_mutex); 1780 break; 1781 } 1782 rval = fcip_handle_suspend(fport, cmd); 1783 mutex_exit(&fcip_global_mutex); 1784 break; 1785 default: 1786 FCIP_DEBUG(FCIP_DEBUG_DETACH, 1787 (CE_WARN, "unknown port detach command!!")); 1788 break; 1789 } 1790 return (rval); 1791 } 1792 1793 1794 /* 1795 * Returns 0 if the port is not busy, else returns non zero. 1796 */ 1797 static int 1798 fcip_check_port_busy(struct fcip *fptr) 1799 { 1800 int rval = 0, num_pkts = 0; 1801 1802 ASSERT(fptr != NULL); 1803 ASSERT(MUTEX_HELD(&fptr->fcip_mutex)); 1804 1805 mutex_enter(&fptr->fcip_dest_mutex); 1806 1807 if (fptr->fcip_flags & FCIP_PORT_BUSY || 1808 ((num_pkts = fcip_port_get_num_pkts(fptr)) > 0) || 1809 fptr->fcip_num_ipkts_pending) { 1810 rval = 1; 1811 FCIP_DEBUG(FCIP_DEBUG_DETACH, 1812 (CE_NOTE, "!fcip_check_port_busy: port is busy " 1813 "fcip_flags: 0x%x, num_pkts: 0x%x, ipkts_pending: 0x%lx!", 1814 fptr->fcip_flags, num_pkts, fptr->fcip_num_ipkts_pending)); 1815 } 1816 1817 mutex_exit(&fptr->fcip_dest_mutex); 1818 return (rval); 1819 } 1820 1821 /* 1822 * Helper routine to remove fcip's minor node 1823 * There is one minor node per system and it should be removed if there are no 1824 * other fcip instances (which has a 1:1 mapping for fp instances) present 1825 */ 1826 static void 1827 fcip_check_remove_minor_node(void) 1828 { 1829 ASSERT(MUTEX_HELD(&fcip_global_mutex)); 1830 1831 /* 1832 * If there are no more fcip (fp) instances, remove the 1833 * minor node for fcip. 1834 * Reset fcip_minor_node_created to invalidate it. 1835 */ 1836 if (fcip_num_instances == 0 && (fcip_module_dip != NULL)) { 1837 ddi_remove_minor_node(fcip_module_dip, NULL); 1838 fcip_minor_node_created = 0; 1839 } 1840 } 1841 1842 /* 1843 * This routine permits the suspend operation during a CPR/System 1844 * power management operation. The routine basically quiesces I/Os 1845 * on all active interfaces 1846 */ 1847 static int 1848 fcip_handle_suspend(fcip_port_info_t *fport, fc_detach_cmd_t cmd) 1849 { 1850 struct fcip *fptr = fport->fcipp_fcip; 1851 timeout_id_t tid; 1852 int index; 1853 int tryagain = 0; 1854 int count; 1855 struct fcipstr *tslp; 1856 1857 1858 ASSERT(fptr != NULL); 1859 mutex_enter(&fptr->fcip_mutex); 1860 1861 /* 1862 * Fail if we are in the middle of a callback. Don't use delay during 1863 * suspend since clock intrs are not available so busy wait 1864 */ 1865 count = 0; 1866 while (count++ < 15 && 1867 ((fptr->fcip_flags & FCIP_IN_CALLBACK) || 1868 (fptr->fcip_flags & FCIP_IN_TIMEOUT))) { 1869 mutex_exit(&fptr->fcip_mutex); 1870 drv_usecwait(1000000); 1871 mutex_enter(&fptr->fcip_mutex); 1872 } 1873 1874 if (fptr->fcip_flags & FCIP_IN_CALLBACK || 1875 fptr->fcip_flags & FCIP_IN_TIMEOUT) { 1876 mutex_exit(&fptr->fcip_mutex); 1877 return (FC_FAILURE); 1878 } 1879 1880 if (cmd == FC_CMD_POWER_DOWN) { 1881 if (fptr->fcip_flags & FCIP_SUSPENDED) { 1882 fptr->fcip_flags |= FCIP_POWER_DOWN; 1883 mutex_exit(&fptr->fcip_mutex); 1884 goto success; 1885 } else { 1886 fptr->fcip_flags |= FCIP_POWER_DOWN; 1887 } 1888 } else if (cmd == FC_CMD_SUSPEND) { 1889 fptr->fcip_flags |= FCIP_SUSPENDED; 1890 } else { 1891 mutex_exit(&fptr->fcip_mutex); 1892 return (FC_FAILURE); 1893 } 1894 1895 mutex_exit(&fptr->fcip_mutex); 1896 /* 1897 * If no streams are plumbed - its the easiest case - Just 1898 * bail out without having to do much 1899 */ 1900 1901 rw_enter(&fcipstruplock, RW_READER); 1902 for (tslp = fcipstrup; tslp; tslp = tslp->sl_nextp) { 1903 if (tslp->sl_fcip == fptr) { 1904 break; 1905 } 1906 } 1907 rw_exit(&fcipstruplock); 1908 1909 /* 1910 * No active streams on this port 1911 */ 1912 if (tslp == NULL) { 1913 goto success; 1914 } 1915 1916 /* 1917 * Walk through each Routing table structure and check if 1918 * the destination table has any outstanding commands. If yes 1919 * wait for the commands to drain. Since we go through each 1920 * routing table entry in succession, it may be wise to wait 1921 * only a few seconds for each entry. 1922 */ 1923 mutex_enter(&fptr->fcip_rt_mutex); 1924 while (!tryagain) { 1925 1926 tryagain = 0; 1927 for (index = 0; index < FCIP_RT_HASH_ELEMS; index++) { 1928 struct fcip_routing_table *frp; 1929 struct fcip_dest *fdestp; 1930 la_wwn_t *pwwn; 1931 int hash_bucket; 1932 1933 frp = fptr->fcip_rtable[index]; 1934 while (frp) { 1935 /* 1936 * Mark the routing table as SUSPENDED. Even 1937 * mark the broadcast entry SUSPENDED to 1938 * prevent any ARP or other broadcasts. We 1939 * can reset the state of the broadcast 1940 * RTE when we resume. 1941 */ 1942 frp->fcipr_state = FCIP_RT_SUSPENDED; 1943 pwwn = &frp->fcipr_pwwn; 1944 1945 /* 1946 * Get hold of destination pointer 1947 */ 1948 mutex_enter(&fptr->fcip_dest_mutex); 1949 1950 hash_bucket = FCIP_DEST_HASH(pwwn->raw_wwn); 1951 ASSERT(hash_bucket < FCIP_DEST_HASH_ELEMS); 1952 1953 fdestp = fptr->fcip_dest[hash_bucket]; 1954 while (fdestp != NULL) { 1955 mutex_enter(&fdestp->fcipd_mutex); 1956 if (fdestp->fcipd_rtable) { 1957 if (fcip_wwn_compare(pwwn, 1958 &fdestp->fcipd_pwwn, 1959 FCIP_COMPARE_PWWN) == 0) { 1960 mutex_exit( 1961 &fdestp->fcipd_mutex); 1962 break; 1963 } 1964 } 1965 mutex_exit(&fdestp->fcipd_mutex); 1966 fdestp = fdestp->fcipd_next; 1967 } 1968 1969 mutex_exit(&fptr->fcip_dest_mutex); 1970 if (fdestp == NULL) { 1971 frp = frp->fcipr_next; 1972 continue; 1973 } 1974 1975 /* 1976 * Wait for fcip_wait_cmds seconds for 1977 * the commands to drain. 1978 */ 1979 count = 0; 1980 mutex_enter(&fdestp->fcipd_mutex); 1981 while (fdestp->fcipd_ncmds && 1982 count < fcip_wait_cmds) { 1983 mutex_exit(&fdestp->fcipd_mutex); 1984 mutex_exit(&fptr->fcip_rt_mutex); 1985 drv_usecwait(1000000); 1986 mutex_enter(&fptr->fcip_rt_mutex); 1987 mutex_enter(&fdestp->fcipd_mutex); 1988 count++; 1989 } 1990 /* 1991 * Check if we were able to drain all cmds 1992 * successfully. Else continue with other 1993 * ports and try during the second pass 1994 */ 1995 if (fdestp->fcipd_ncmds) { 1996 tryagain++; 1997 } 1998 mutex_exit(&fdestp->fcipd_mutex); 1999 2000 frp = frp->fcipr_next; 2001 } 2002 } 2003 if (tryagain == 0) { 2004 break; 2005 } 2006 } 2007 mutex_exit(&fptr->fcip_rt_mutex); 2008 2009 if (tryagain) { 2010 mutex_enter(&fptr->fcip_mutex); 2011 fptr->fcip_flags &= ~(FCIP_SUSPENDED | FCIP_POWER_DOWN); 2012 mutex_exit(&fptr->fcip_mutex); 2013 return (FC_FAILURE); 2014 } 2015 2016 success: 2017 mutex_enter(&fptr->fcip_mutex); 2018 tid = fptr->fcip_timeout_id; 2019 fptr->fcip_timeout_id = NULL; 2020 mutex_exit(&fptr->fcip_mutex); 2021 2022 (void) untimeout(tid); 2023 2024 return (FC_SUCCESS); 2025 } 2026 2027 /* 2028 * the getinfo(9E) entry point 2029 */ 2030 /* ARGSUSED */ 2031 static int 2032 fcip_getinfo(dev_info_t *dip, ddi_info_cmd_t cmd, void *arg, void **result) 2033 { 2034 int rval = DDI_FAILURE; 2035 2036 switch (cmd) { 2037 case DDI_INFO_DEVT2DEVINFO: 2038 *result = fcip_module_dip; 2039 if (*result) 2040 rval = DDI_SUCCESS; 2041 break; 2042 2043 case DDI_INFO_DEVT2INSTANCE: 2044 *result = (void *)0; 2045 rval = DDI_SUCCESS; 2046 break; 2047 default: 2048 break; 2049 } 2050 2051 return (rval); 2052 } 2053 2054 /* 2055 * called from fcip_attach to initialize kstats for the link 2056 */ 2057 /* ARGSUSED */ 2058 static void 2059 fcip_kstat_init(struct fcip *fptr) 2060 { 2061 int instance; 2062 char buf[16]; 2063 struct fcipstat *fcipstatp; 2064 2065 ASSERT(mutex_owned(&fptr->fcip_mutex)); 2066 2067 instance = ddi_get_instance(fptr->fcip_dip); 2068 (void) sprintf(buf, "fcip%d", instance); 2069 2070 #ifdef kstat 2071 fptr->fcip_kstatp = kstat_create("fcip", instance, buf, "net", 2072 KSTAT_TYPE_NAMED, 2073 (sizeof (struct fcipstat)/ sizeof (kstat_named_t)), 2074 KSTAT_FLAG_PERSISTENT); 2075 #else 2076 fptr->fcip_kstatp = kstat_create("fcip", instance, buf, "net", 2077 KSTAT_TYPE_NAMED, 2078 (sizeof (struct fcipstat)/ sizeof (kstat_named_t)), 0); 2079 #endif 2080 if (fptr->fcip_kstatp == NULL) { 2081 FCIP_DEBUG(FCIP_DEBUG_INIT, (CE_WARN, "kstat created failed")); 2082 return; 2083 } 2084 2085 fcipstatp = (struct fcipstat *)fptr->fcip_kstatp->ks_data; 2086 kstat_named_init(&fcipstatp->fcips_ipackets, "ipackets", 2087 KSTAT_DATA_ULONG); 2088 kstat_named_init(&fcipstatp->fcips_ierrors, "ierrors", 2089 KSTAT_DATA_ULONG); 2090 kstat_named_init(&fcipstatp->fcips_opackets, "opackets", 2091 KSTAT_DATA_ULONG); 2092 kstat_named_init(&fcipstatp->fcips_oerrors, "oerrors", 2093 KSTAT_DATA_ULONG); 2094 kstat_named_init(&fcipstatp->fcips_collisions, "collisions", 2095 KSTAT_DATA_ULONG); 2096 kstat_named_init(&fcipstatp->fcips_nocanput, "nocanput", 2097 KSTAT_DATA_ULONG); 2098 kstat_named_init(&fcipstatp->fcips_allocbfail, "allocbfail", 2099 KSTAT_DATA_ULONG); 2100 2101 kstat_named_init(&fcipstatp->fcips_defer, "defer", 2102 KSTAT_DATA_ULONG); 2103 kstat_named_init(&fcipstatp->fcips_fram, "fram", 2104 KSTAT_DATA_ULONG); 2105 kstat_named_init(&fcipstatp->fcips_crc, "crc", 2106 KSTAT_DATA_ULONG); 2107 kstat_named_init(&fcipstatp->fcips_oflo, "oflo", 2108 KSTAT_DATA_ULONG); 2109 kstat_named_init(&fcipstatp->fcips_uflo, "uflo", 2110 KSTAT_DATA_ULONG); 2111 kstat_named_init(&fcipstatp->fcips_missed, "missed", 2112 KSTAT_DATA_ULONG); 2113 kstat_named_init(&fcipstatp->fcips_tlcol, "tlcol", 2114 KSTAT_DATA_ULONG); 2115 kstat_named_init(&fcipstatp->fcips_trtry, "trtry", 2116 KSTAT_DATA_ULONG); 2117 kstat_named_init(&fcipstatp->fcips_tnocar, "tnocar", 2118 KSTAT_DATA_ULONG); 2119 kstat_named_init(&fcipstatp->fcips_inits, "inits", 2120 KSTAT_DATA_ULONG); 2121 kstat_named_init(&fcipstatp->fcips_notbufs, "notbufs", 2122 KSTAT_DATA_ULONG); 2123 kstat_named_init(&fcipstatp->fcips_norbufs, "norbufs", 2124 KSTAT_DATA_ULONG); 2125 kstat_named_init(&fcipstatp->fcips_allocbfail, "allocbfail", 2126 KSTAT_DATA_ULONG); 2127 2128 /* 2129 * required by kstat for MIB II objects(RFC 1213) 2130 */ 2131 kstat_named_init(&fcipstatp->fcips_rcvbytes, "fcips_rcvbytes", 2132 KSTAT_DATA_ULONG); /* # octets received */ 2133 /* MIB - ifInOctets */ 2134 kstat_named_init(&fcipstatp->fcips_xmtbytes, "fcips_xmtbytes", 2135 KSTAT_DATA_ULONG); /* # octets xmitted */ 2136 /* MIB - ifOutOctets */ 2137 kstat_named_init(&fcipstatp->fcips_multircv, "fcips_multircv", 2138 KSTAT_DATA_ULONG); /* # multicast packets */ 2139 /* delivered to upper layer */ 2140 /* MIB - ifInNUcastPkts */ 2141 kstat_named_init(&fcipstatp->fcips_multixmt, "fcips_multixmt", 2142 KSTAT_DATA_ULONG); /* # multicast packets */ 2143 /* requested to be sent */ 2144 /* MIB - ifOutNUcastPkts */ 2145 kstat_named_init(&fcipstatp->fcips_brdcstrcv, "fcips_brdcstrcv", 2146 KSTAT_DATA_ULONG); /* # broadcast packets */ 2147 /* delivered to upper layer */ 2148 /* MIB - ifInNUcastPkts */ 2149 kstat_named_init(&fcipstatp->fcips_brdcstxmt, "fcips_brdcstxmt", 2150 KSTAT_DATA_ULONG); /* # broadcast packets */ 2151 /* requested to be sent */ 2152 /* MIB - ifOutNUcastPkts */ 2153 kstat_named_init(&fcipstatp->fcips_norcvbuf, "fcips_norcvbuf", 2154 KSTAT_DATA_ULONG); /* # rcv packets discarded */ 2155 /* MIB - ifInDiscards */ 2156 kstat_named_init(&fcipstatp->fcips_noxmtbuf, "fcips_noxmtbuf", 2157 KSTAT_DATA_ULONG); /* # xmt packets discarded */ 2158 2159 fptr->fcip_kstatp->ks_update = fcip_stat_update; 2160 fptr->fcip_kstatp->ks_private = (void *) fptr; 2161 kstat_install(fptr->fcip_kstatp); 2162 } 2163 2164 /* 2165 * Update the defined kstats for netstat et al to use 2166 */ 2167 /* ARGSUSED */ 2168 static int 2169 fcip_stat_update(kstat_t *fcip_statp, int val) 2170 { 2171 struct fcipstat *fcipstatp; 2172 struct fcip *fptr; 2173 2174 fptr = (struct fcip *)fcip_statp->ks_private; 2175 fcipstatp = (struct fcipstat *)fcip_statp->ks_data; 2176 2177 if (val == KSTAT_WRITE) { 2178 fptr->fcip_ipackets = fcipstatp->fcips_ipackets.value.ul; 2179 fptr->fcip_ierrors = fcipstatp->fcips_ierrors.value.ul; 2180 fptr->fcip_opackets = fcipstatp->fcips_opackets.value.ul; 2181 fptr->fcip_oerrors = fcipstatp->fcips_oerrors.value.ul; 2182 fptr->fcip_collisions = fcipstatp->fcips_collisions.value.ul; 2183 fptr->fcip_defer = fcipstatp->fcips_defer.value.ul; 2184 fptr->fcip_fram = fcipstatp->fcips_fram.value.ul; 2185 fptr->fcip_crc = fcipstatp->fcips_crc.value.ul; 2186 fptr->fcip_oflo = fcipstatp->fcips_oflo.value.ul; 2187 fptr->fcip_uflo = fcipstatp->fcips_uflo.value.ul; 2188 fptr->fcip_missed = fcipstatp->fcips_missed.value.ul; 2189 fptr->fcip_tlcol = fcipstatp->fcips_tlcol.value.ul; 2190 fptr->fcip_trtry = fcipstatp->fcips_trtry.value.ul; 2191 fptr->fcip_tnocar = fcipstatp->fcips_tnocar.value.ul; 2192 fptr->fcip_inits = fcipstatp->fcips_inits.value.ul; 2193 fptr->fcip_notbufs = fcipstatp->fcips_notbufs.value.ul; 2194 fptr->fcip_norbufs = fcipstatp->fcips_norbufs.value.ul; 2195 fptr->fcip_nocanput = fcipstatp->fcips_nocanput.value.ul; 2196 fptr->fcip_allocbfail = fcipstatp->fcips_allocbfail.value.ul; 2197 fptr->fcip_rcvbytes = fcipstatp->fcips_rcvbytes.value.ul; 2198 fptr->fcip_xmtbytes = fcipstatp->fcips_xmtbytes.value.ul; 2199 fptr->fcip_multircv = fcipstatp->fcips_multircv.value.ul; 2200 fptr->fcip_multixmt = fcipstatp->fcips_multixmt.value.ul; 2201 fptr->fcip_brdcstrcv = fcipstatp->fcips_brdcstrcv.value.ul; 2202 fptr->fcip_norcvbuf = fcipstatp->fcips_norcvbuf.value.ul; 2203 fptr->fcip_noxmtbuf = fcipstatp->fcips_noxmtbuf.value.ul; 2204 fptr->fcip_allocbfail = fcipstatp->fcips_allocbfail.value.ul; 2205 fptr->fcip_allocbfail = fcipstatp->fcips_allocbfail.value.ul; 2206 fptr->fcip_allocbfail = fcipstatp->fcips_allocbfail.value.ul; 2207 fptr->fcip_allocbfail = fcipstatp->fcips_allocbfail.value.ul; 2208 fptr->fcip_allocbfail = fcipstatp->fcips_allocbfail.value.ul; 2209 fptr->fcip_allocbfail = fcipstatp->fcips_allocbfail.value.ul; 2210 fptr->fcip_allocbfail = fcipstatp->fcips_allocbfail.value.ul; 2211 fptr->fcip_allocbfail = fcipstatp->fcips_allocbfail.value.ul; 2212 2213 } else { 2214 fcipstatp->fcips_ipackets.value.ul = fptr->fcip_ipackets; 2215 fcipstatp->fcips_ierrors.value.ul = fptr->fcip_ierrors; 2216 fcipstatp->fcips_opackets.value.ul = fptr->fcip_opackets; 2217 fcipstatp->fcips_oerrors.value.ul = fptr->fcip_oerrors; 2218 fcipstatp->fcips_collisions.value.ul = fptr->fcip_collisions; 2219 fcipstatp->fcips_nocanput.value.ul = fptr->fcip_nocanput; 2220 fcipstatp->fcips_allocbfail.value.ul = fptr->fcip_allocbfail; 2221 fcipstatp->fcips_defer.value.ul = fptr->fcip_defer; 2222 fcipstatp->fcips_fram.value.ul = fptr->fcip_fram; 2223 fcipstatp->fcips_crc.value.ul = fptr->fcip_crc; 2224 fcipstatp->fcips_oflo.value.ul = fptr->fcip_oflo; 2225 fcipstatp->fcips_uflo.value.ul = fptr->fcip_uflo; 2226 fcipstatp->fcips_missed.value.ul = fptr->fcip_missed; 2227 fcipstatp->fcips_tlcol.value.ul = fptr->fcip_tlcol; 2228 fcipstatp->fcips_trtry.value.ul = fptr->fcip_trtry; 2229 fcipstatp->fcips_tnocar.value.ul = fptr->fcip_tnocar; 2230 fcipstatp->fcips_inits.value.ul = fptr->fcip_inits; 2231 fcipstatp->fcips_norbufs.value.ul = fptr->fcip_norbufs; 2232 fcipstatp->fcips_notbufs.value.ul = fptr->fcip_notbufs; 2233 fcipstatp->fcips_rcvbytes.value.ul = fptr->fcip_rcvbytes; 2234 fcipstatp->fcips_xmtbytes.value.ul = fptr->fcip_xmtbytes; 2235 fcipstatp->fcips_multircv.value.ul = fptr->fcip_multircv; 2236 fcipstatp->fcips_multixmt.value.ul = fptr->fcip_multixmt; 2237 fcipstatp->fcips_brdcstrcv.value.ul = fptr->fcip_brdcstrcv; 2238 fcipstatp->fcips_brdcstxmt.value.ul = fptr->fcip_brdcstxmt; 2239 fcipstatp->fcips_norcvbuf.value.ul = fptr->fcip_norcvbuf; 2240 fcipstatp->fcips_noxmtbuf.value.ul = fptr->fcip_noxmtbuf; 2241 2242 } 2243 return (0); 2244 } 2245 2246 2247 /* 2248 * fcip_statec_cb: handles all required state change callback notifications 2249 * it receives from the transport 2250 */ 2251 /* ARGSUSED */ 2252 static void 2253 fcip_statec_cb(opaque_t ulp_handle, opaque_t phandle, 2254 uint32_t port_state, uint32_t port_top, fc_portmap_t changelist[], 2255 uint32_t listlen, uint32_t sid) 2256 { 2257 fcip_port_info_t *fport; 2258 struct fcip *fptr; 2259 struct fcipstr *slp; 2260 queue_t *wrq; 2261 int instance; 2262 int index; 2263 struct fcip_routing_table *frtp; 2264 2265 fport = fcip_get_port(phandle); 2266 2267 if (fport == NULL) { 2268 return; 2269 } 2270 2271 fptr = fport->fcipp_fcip; 2272 ASSERT(fptr != NULL); 2273 2274 if (fptr == NULL) { 2275 return; 2276 } 2277 2278 instance = ddi_get_instance(fport->fcipp_dip); 2279 2280 FCIP_TNF_PROBE_4((fcip_statec_cb, "fcip io", /* CSTYLED */, 2281 tnf_string, msg, "state change callback", 2282 tnf_uint, instance, instance, 2283 tnf_uint, S_ID, sid, 2284 tnf_int, count, listlen)); 2285 FCIP_DEBUG(FCIP_DEBUG_ELS, 2286 (CE_NOTE, "fcip%d, state change callback: state:0x%x, " 2287 "S_ID:0x%x, count:0x%x", instance, port_state, sid, listlen)); 2288 2289 mutex_enter(&fptr->fcip_mutex); 2290 2291 if ((fptr->fcip_flags & (FCIP_DETACHING | FCIP_DETACHED)) || 2292 (fptr->fcip_flags & (FCIP_SUSPENDED | FCIP_POWER_DOWN))) { 2293 mutex_exit(&fptr->fcip_mutex); 2294 return; 2295 } 2296 2297 /* 2298 * set fcip flags to indicate we are in the middle of a 2299 * state change callback so we can wait till the statechange 2300 * is handled before succeeding/failing the SUSPEND/POWER DOWN. 2301 */ 2302 fptr->fcip_flags |= FCIP_IN_SC_CB; 2303 2304 fport->fcipp_pstate = port_state; 2305 2306 /* 2307 * Check if topology changed. If Yes - Modify the broadcast 2308 * RTE entries to understand the new broadcast D_IDs 2309 */ 2310 if (fport->fcipp_topology != port_top && 2311 (port_top != FC_TOP_UNKNOWN)) { 2312 /* REMOVE later */ 2313 FCIP_DEBUG(FCIP_DEBUG_ELS, (CE_NOTE, 2314 "topology changed: Old topology: 0x%x New topology 0x%x", 2315 fport->fcipp_topology, port_top)); 2316 /* 2317 * If topology changed - attempt a rediscovery of 2318 * devices. Helps specially in Fabric/Public loops 2319 * and if on_demand_node_creation is disabled 2320 */ 2321 fport->fcipp_topology = port_top; 2322 fcip_handle_topology(fptr); 2323 } 2324 2325 mutex_exit(&fptr->fcip_mutex); 2326 2327 switch (FC_PORT_STATE_MASK(port_state)) { 2328 case FC_STATE_ONLINE: 2329 /* FALLTHROUGH */ 2330 case FC_STATE_LIP: 2331 /* FALLTHROUGH */ 2332 case FC_STATE_LIP_LBIT_SET: 2333 2334 /* 2335 * nothing to do here actually other than if we 2336 * were actually logged onto a port in the devlist 2337 * (which indicates active communication between 2338 * the host port and the port in the changelist). 2339 * If however we are in a private loop or point to 2340 * point mode, we need to check for any IP capable 2341 * ports and update our routing table. 2342 */ 2343 switch (port_top) { 2344 case FC_TOP_FABRIC: 2345 /* 2346 * This indicates a fabric port with a NameServer. 2347 * Check the devlist to see if we are in active 2348 * communication with a port on the devlist. 2349 */ 2350 FCIP_DEBUG(FCIP_DEBUG_ELS, (CE_NOTE, 2351 "Statec_cb: fabric topology")); 2352 fcip_rt_update(fptr, changelist, listlen); 2353 break; 2354 case FC_TOP_NO_NS: 2355 /* 2356 * No nameserver - so treat it like a Private loop 2357 * or point to point topology and get a map of 2358 * devices on the link and get IP capable ports to 2359 * to update the routing table. 2360 */ 2361 FCIP_DEBUG(FCIP_DEBUG_ELS, 2362 (CE_NOTE, "Statec_cb: NO_NS topology")); 2363 /* FALLTHROUGH */ 2364 case FC_TOP_PRIVATE_LOOP: 2365 FCIP_DEBUG(FCIP_DEBUG_ELS, (CE_NOTE, 2366 "Statec_cb: Pvt_Loop topology")); 2367 /* FALLTHROUGH */ 2368 case FC_TOP_PT_PT: 2369 /* 2370 * call get_port_map() and update routing table 2371 */ 2372 fcip_rt_update(fptr, changelist, listlen); 2373 break; 2374 default: 2375 FCIP_DEBUG(FCIP_DEBUG_ELS, 2376 (CE_NOTE, "Statec_cb: Unknown topology")); 2377 } 2378 2379 /* 2380 * We should now enable the Queues and permit I/Os 2381 * to flow through downstream. The update of routing 2382 * table should have flushed out any port entries that 2383 * don't exist or are not available after the state change 2384 */ 2385 mutex_enter(&fptr->fcip_mutex); 2386 fptr->fcip_port_state = FCIP_PORT_ONLINE; 2387 if (fptr->fcip_flags & FCIP_LINK_DOWN) { 2388 fptr->fcip_flags &= ~FCIP_LINK_DOWN; 2389 } 2390 mutex_exit(&fptr->fcip_mutex); 2391 2392 /* 2393 * Enable write queues 2394 */ 2395 rw_enter(&fcipstruplock, RW_READER); 2396 for (slp = fcipstrup; slp != NULL; slp = slp->sl_nextp) { 2397 if (slp && slp->sl_fcip == fptr) { 2398 wrq = WR(slp->sl_rq); 2399 if (wrq->q_flag & QFULL) { 2400 qenable(wrq); 2401 } 2402 } 2403 } 2404 rw_exit(&fcipstruplock); 2405 break; 2406 case FC_STATE_OFFLINE: 2407 /* 2408 * mark the port_state OFFLINE and wait for it to 2409 * become online. Any new messages in this state will 2410 * simply be queued back up. If the port does not 2411 * come online in a short while, we can begin failing 2412 * messages and flush the routing table 2413 */ 2414 mutex_enter(&fptr->fcip_mutex); 2415 fptr->fcip_mark_offline = fptr->fcip_timeout_ticks + 2416 FCIP_OFFLINE_TIMEOUT; 2417 fptr->fcip_port_state = FCIP_PORT_OFFLINE; 2418 mutex_exit(&fptr->fcip_mutex); 2419 2420 /* 2421 * Mark all Routing table entries as invalid to prevent 2422 * any commands from trickling through to ports that 2423 * have disappeared from under us 2424 */ 2425 mutex_enter(&fptr->fcip_rt_mutex); 2426 for (index = 0; index < FCIP_RT_HASH_ELEMS; index++) { 2427 frtp = fptr->fcip_rtable[index]; 2428 while (frtp) { 2429 frtp->fcipr_state = PORT_DEVICE_INVALID; 2430 frtp = frtp->fcipr_next; 2431 } 2432 } 2433 mutex_exit(&fptr->fcip_rt_mutex); 2434 2435 break; 2436 2437 case FC_STATE_RESET_REQUESTED: 2438 /* 2439 * Release all Unsolicited buffers back to transport/FCA. 2440 * This also means the port state is marked offline - so 2441 * we may have to do what OFFLINE state requires us to do. 2442 * Care must be taken to wait for any active unsolicited 2443 * buffer with the other Streams modules - so wait for 2444 * a freeb if the unsolicited buffer is passed back all 2445 * the way upstream. 2446 */ 2447 mutex_enter(&fptr->fcip_mutex); 2448 2449 #ifdef FCIP_ESBALLOC 2450 while (fptr->fcip_ub_upstream) { 2451 cv_wait(&fptr->fcip_ub_cv, &fptr->fcip_mutex); 2452 } 2453 #endif /* FCIP_ESBALLOC */ 2454 2455 fptr->fcip_mark_offline = fptr->fcip_timeout_ticks + 2456 FCIP_OFFLINE_TIMEOUT; 2457 fptr->fcip_port_state = FCIP_PORT_OFFLINE; 2458 mutex_exit(&fptr->fcip_mutex); 2459 break; 2460 2461 case FC_STATE_DEVICE_CHANGE: 2462 if (listlen) { 2463 fcip_rt_update(fptr, changelist, listlen); 2464 } 2465 break; 2466 case FC_STATE_RESET: 2467 /* 2468 * Not much to do I guess - wait for port to become 2469 * ONLINE. If the port doesn't become online in a short 2470 * while, the upper layers abort any request themselves. 2471 * We can just putback the messages in the streams queues 2472 * if the link is offline 2473 */ 2474 break; 2475 } 2476 mutex_enter(&fptr->fcip_mutex); 2477 fptr->fcip_flags &= ~(FCIP_IN_SC_CB); 2478 mutex_exit(&fptr->fcip_mutex); 2479 } 2480 2481 /* 2482 * Given a port handle, return the fcip_port_info structure corresponding 2483 * to that port handle. The transport allocates and communicates with 2484 * ULPs using port handles 2485 */ 2486 static fcip_port_info_t * 2487 fcip_get_port(opaque_t phandle) 2488 { 2489 fcip_port_info_t *fport; 2490 2491 ASSERT(phandle != NULL); 2492 2493 mutex_enter(&fcip_global_mutex); 2494 fport = fcip_port_head; 2495 2496 while (fport != NULL) { 2497 if (fport->fcipp_handle == phandle) { 2498 /* found */ 2499 break; 2500 } 2501 fport = fport->fcipp_next; 2502 } 2503 2504 mutex_exit(&fcip_global_mutex); 2505 2506 return (fport); 2507 } 2508 2509 /* 2510 * Handle inbound ELS requests received by the transport. We are only 2511 * intereseted in FARP/InARP mostly. 2512 */ 2513 /* ARGSUSED */ 2514 static int 2515 fcip_els_cb(opaque_t ulp_handle, opaque_t phandle, 2516 fc_unsol_buf_t *buf, uint32_t claimed) 2517 { 2518 fcip_port_info_t *fport; 2519 struct fcip *fptr; 2520 int instance; 2521 uchar_t r_ctl; 2522 uchar_t ls_code; 2523 la_els_farp_t farp_cmd; 2524 la_els_farp_t *fcmd; 2525 int rval = FC_UNCLAIMED; 2526 2527 fport = fcip_get_port(phandle); 2528 if (fport == NULL) { 2529 return (FC_UNCLAIMED); 2530 } 2531 2532 fptr = fport->fcipp_fcip; 2533 ASSERT(fptr != NULL); 2534 if (fptr == NULL) { 2535 return (FC_UNCLAIMED); 2536 } 2537 2538 instance = ddi_get_instance(fport->fcipp_dip); 2539 2540 mutex_enter(&fptr->fcip_mutex); 2541 if ((fptr->fcip_flags & (FCIP_DETACHING | FCIP_DETACHED)) || 2542 (fptr->fcip_flags & (FCIP_SUSPENDED | FCIP_POWER_DOWN))) { 2543 mutex_exit(&fptr->fcip_mutex); 2544 return (FC_UNCLAIMED); 2545 } 2546 2547 /* 2548 * set fcip flags to indicate we are in the middle of a 2549 * ELS callback so we can wait till the statechange 2550 * is handled before succeeding/failing the SUSPEND/POWER DOWN. 2551 */ 2552 fptr->fcip_flags |= FCIP_IN_ELS_CB; 2553 mutex_exit(&fptr->fcip_mutex); 2554 2555 FCIP_TNF_PROBE_2((fcip_els_cb, "fcip io", /* CSTYLED */, 2556 tnf_string, msg, "ELS callback", 2557 tnf_uint, instance, instance)); 2558 2559 FCIP_DEBUG(FCIP_DEBUG_ELS, 2560 (CE_NOTE, "fcip%d, ELS callback , ", instance)); 2561 2562 r_ctl = buf->ub_frame.r_ctl; 2563 switch (r_ctl & R_CTL_ROUTING) { 2564 case R_CTL_EXTENDED_SVC: 2565 if (r_ctl == R_CTL_ELS_REQ) { 2566 ls_code = buf->ub_buffer[0]; 2567 if (ls_code == LA_ELS_FARP_REQ) { 2568 /* 2569 * Inbound FARP broadcast request 2570 */ 2571 if (buf->ub_bufsize != sizeof (la_els_farp_t)) { 2572 FCIP_DEBUG(FCIP_DEBUG_ELS, (CE_WARN, 2573 "Invalid FARP req buffer size " 2574 "expected 0x%lx, got 0x%x", 2575 (long)(sizeof (la_els_farp_t)), 2576 buf->ub_bufsize)); 2577 rval = FC_UNCLAIMED; 2578 goto els_cb_done; 2579 } 2580 fcmd = (la_els_farp_t *)buf; 2581 if (fcip_wwn_compare(&fcmd->resp_nwwn, 2582 &fport->fcipp_nwwn, 2583 FCIP_COMPARE_NWWN) != 0) { 2584 rval = FC_UNCLAIMED; 2585 goto els_cb_done; 2586 } 2587 /* 2588 * copy the FARP request and release the 2589 * unsolicited buffer 2590 */ 2591 fcmd = &farp_cmd; 2592 bcopy((void *)buf, (void *)fcmd, 2593 sizeof (la_els_farp_t)); 2594 (void) fc_ulp_ubrelease(fport->fcipp_handle, 1, 2595 &buf->ub_token); 2596 2597 if (fcip_farp_supported && 2598 fcip_handle_farp_request(fptr, fcmd) == 2599 FC_SUCCESS) { 2600 /* 2601 * We successfully sent out a FARP 2602 * reply to the requesting port 2603 */ 2604 rval = FC_SUCCESS; 2605 goto els_cb_done; 2606 } else { 2607 rval = FC_UNCLAIMED; 2608 goto els_cb_done; 2609 } 2610 } 2611 } else if (r_ctl == R_CTL_ELS_RSP) { 2612 ls_code = buf->ub_buffer[0]; 2613 if (ls_code == LA_ELS_FARP_REPLY) { 2614 /* 2615 * We received a REPLY to our FARP request 2616 */ 2617 if (buf->ub_bufsize != sizeof (la_els_farp_t)) { 2618 FCIP_DEBUG(FCIP_DEBUG_ELS, (CE_WARN, 2619 "Invalid FARP req buffer size " 2620 "expected 0x%lx, got 0x%x", 2621 (long)(sizeof (la_els_farp_t)), 2622 buf->ub_bufsize)); 2623 rval = FC_UNCLAIMED; 2624 goto els_cb_done; 2625 } 2626 fcmd = &farp_cmd; 2627 bcopy((void *)buf, (void *)fcmd, 2628 sizeof (la_els_farp_t)); 2629 (void) fc_ulp_ubrelease(fport->fcipp_handle, 1, 2630 &buf->ub_token); 2631 if (fcip_farp_supported && 2632 fcip_handle_farp_response(fptr, fcmd) == 2633 FC_SUCCESS) { 2634 FCIP_DEBUG(FCIP_DEBUG_ELS, (CE_NOTE, 2635 "Successfully recevied a FARP " 2636 "response")); 2637 mutex_enter(&fptr->fcip_mutex); 2638 fptr->fcip_farp_rsp_flag = 1; 2639 cv_signal(&fptr->fcip_farp_cv); 2640 mutex_exit(&fptr->fcip_mutex); 2641 rval = FC_SUCCESS; 2642 goto els_cb_done; 2643 } else { 2644 FCIP_DEBUG(FCIP_DEBUG_ELS, (CE_WARN, 2645 "Unable to handle a FARP response " 2646 "receive")); 2647 rval = FC_UNCLAIMED; 2648 goto els_cb_done; 2649 } 2650 } 2651 } 2652 break; 2653 default: 2654 break; 2655 } 2656 els_cb_done: 2657 mutex_enter(&fptr->fcip_mutex); 2658 fptr->fcip_flags &= ~(FCIP_IN_ELS_CB); 2659 mutex_exit(&fptr->fcip_mutex); 2660 return (rval); 2661 } 2662 2663 2664 /* 2665 * Handle inbound FARP requests 2666 */ 2667 static int 2668 fcip_handle_farp_request(struct fcip *fptr, la_els_farp_t *fcmd) 2669 { 2670 fcip_pkt_t *fcip_pkt; 2671 fc_packet_t *fc_pkt; 2672 fcip_port_info_t *fport = fptr->fcip_port_info; 2673 int rval = FC_FAILURE; 2674 opaque_t fca_dev; 2675 fc_portmap_t map; 2676 struct fcip_routing_table *frp; 2677 struct fcip_dest *fdestp; 2678 2679 /* 2680 * Add an entry for the remote port into our routing and destination 2681 * tables. 2682 */ 2683 map.map_did = fcmd->req_id; 2684 map.map_hard_addr.hard_addr = fcmd->req_id.port_id; 2685 map.map_state = PORT_DEVICE_VALID; 2686 map.map_type = PORT_DEVICE_NEW; 2687 map.map_flags = 0; 2688 map.map_pd = NULL; 2689 bcopy((void *)&fcmd->req_pwwn, (void *)&map.map_pwwn, 2690 sizeof (la_wwn_t)); 2691 bcopy((void *)&fcmd->req_nwwn, (void *)&map.map_nwwn, 2692 sizeof (la_wwn_t)); 2693 fcip_rt_update(fptr, &map, 1); 2694 mutex_enter(&fptr->fcip_rt_mutex); 2695 frp = fcip_lookup_rtable(fptr, &fcmd->req_pwwn, FCIP_COMPARE_NWWN); 2696 mutex_exit(&fptr->fcip_rt_mutex); 2697 2698 fdestp = fcip_add_dest(fptr, frp); 2699 2700 fcip_pkt = fcip_ipkt_alloc(fptr, sizeof (la_els_farp_t), 2701 sizeof (la_els_farp_t), NULL, KM_SLEEP); 2702 if (fcip_pkt == NULL) { 2703 rval = FC_FAILURE; 2704 goto farp_done; 2705 } 2706 /* 2707 * Fill in our port's PWWN and NWWN 2708 */ 2709 fcmd->resp_pwwn = fport->fcipp_pwwn; 2710 fcmd->resp_nwwn = fport->fcipp_nwwn; 2711 2712 fcip_init_unicast_pkt(fcip_pkt, fport->fcipp_sid, 2713 fcmd->req_id, NULL); 2714 2715 fca_dev = 2716 fc_ulp_get_fca_device(fport->fcipp_handle, fcmd->req_id); 2717 fc_pkt = FCIP_PKT_TO_FC_PKT(fcip_pkt); 2718 fc_pkt->pkt_cmd_fhdr.r_ctl = R_CTL_ELS_RSP; 2719 fc_pkt->pkt_fca_device = fca_dev; 2720 fcip_pkt->fcip_pkt_dest = fdestp; 2721 2722 /* 2723 * Attempt a PLOGI again 2724 */ 2725 if (fcmd->resp_flags & FARP_INIT_P_LOGI) { 2726 if (fcip_do_plogi(fptr, frp) != FC_SUCCESS) { 2727 /* 2728 * Login to the remote port failed. There is no 2729 * point continuing with the FARP request further 2730 * so bail out here. 2731 */ 2732 frp->fcipr_state = PORT_DEVICE_INVALID; 2733 rval = FC_FAILURE; 2734 goto farp_done; 2735 } 2736 } 2737 2738 FCIP_CP_OUT(fcmd, fc_pkt->pkt_cmd, fc_pkt->pkt_cmd_acc, 2739 sizeof (la_els_farp_t)); 2740 2741 rval = fc_ulp_issue_els(fport->fcipp_handle, fc_pkt); 2742 if (rval != FC_SUCCESS) { 2743 FCIP_TNF_PROBE_2((fcip_handle_farp_request, "fcip io", 2744 /* CSTYLED */, tnf_string, msg, 2745 "fcip_transport of farp reply failed", 2746 tnf_uint, rval, rval)); 2747 FCIP_DEBUG(FCIP_DEBUG_ELS, (CE_WARN, 2748 "fcip_transport of farp reply failed 0x%x", rval)); 2749 } 2750 2751 farp_done: 2752 return (rval); 2753 } 2754 2755 2756 /* 2757 * Handle FARP responses to our FARP requests. When we receive a FARP 2758 * reply, we need to add the entry for the Port that replied into our 2759 * routing and destination hash tables. It is possible that the remote 2760 * port did not login into us (FARP responses can be received without 2761 * a PLOGI) 2762 */ 2763 static int 2764 fcip_handle_farp_response(struct fcip *fptr, la_els_farp_t *fcmd) 2765 { 2766 int rval = FC_FAILURE; 2767 fc_portmap_t map; 2768 struct fcip_routing_table *frp; 2769 struct fcip_dest *fdestp; 2770 2771 /* 2772 * Add an entry for the remote port into our routing and destination 2773 * tables. 2774 */ 2775 map.map_did = fcmd->dest_id; 2776 map.map_hard_addr.hard_addr = fcmd->dest_id.port_id; 2777 map.map_state = PORT_DEVICE_VALID; 2778 map.map_type = PORT_DEVICE_NEW; 2779 map.map_flags = 0; 2780 map.map_pd = NULL; 2781 bcopy((void *)&fcmd->resp_pwwn, (void *)&map.map_pwwn, 2782 sizeof (la_wwn_t)); 2783 bcopy((void *)&fcmd->resp_nwwn, (void *)&map.map_nwwn, 2784 sizeof (la_wwn_t)); 2785 fcip_rt_update(fptr, &map, 1); 2786 mutex_enter(&fptr->fcip_rt_mutex); 2787 frp = fcip_lookup_rtable(fptr, &fcmd->resp_pwwn, FCIP_COMPARE_NWWN); 2788 mutex_exit(&fptr->fcip_rt_mutex); 2789 2790 fdestp = fcip_add_dest(fptr, frp); 2791 2792 if (fdestp != NULL) { 2793 rval = FC_SUCCESS; 2794 } 2795 return (rval); 2796 } 2797 2798 2799 #define FCIP_HDRS_LENGTH \ 2800 sizeof (fcph_network_hdr_t)+sizeof (llc_snap_hdr_t)+sizeof (ipha_t) 2801 2802 /* 2803 * fcip_data_cb is the heart of most IP operations. This routine is called 2804 * by the transport when any unsolicited IP data arrives at a port (which 2805 * is almost all IP data). This routine then strips off the Network header 2806 * from the payload (after authenticating the received payload ofcourse), 2807 * creates a message blk and sends the data upstream. You will see ugly 2808 * #defines because of problems with using esballoc() as opposed to 2809 * allocb to prevent an extra copy of data. We should probably move to 2810 * esballoc entirely when the MTU eventually will be larger than 1500 bytes 2811 * since copies will get more expensive then. At 1500 byte MTUs, there is 2812 * no noticable difference between using allocb and esballoc. The other 2813 * caveat is that the qlc firmware still cannot tell us accurately the 2814 * no. of valid bytes in the unsol buffer it DMA'ed so we have to resort 2815 * to looking into the IP header and hoping that the no. of bytes speficified 2816 * in the header was actually received. 2817 */ 2818 /* ARGSUSED */ 2819 static int 2820 fcip_data_cb(opaque_t ulp_handle, opaque_t phandle, 2821 fc_unsol_buf_t *buf, uint32_t claimed) 2822 { 2823 fcip_port_info_t *fport; 2824 struct fcip *fptr; 2825 fcph_network_hdr_t *nhdr; 2826 llc_snap_hdr_t *snaphdr; 2827 mblk_t *bp; 2828 uint32_t len; 2829 uint32_t hdrlen; 2830 ushort_t type; 2831 ipha_t *iphdr; 2832 int rval; 2833 2834 #ifdef FCIP_ESBALLOC 2835 frtn_t *free_ubuf; 2836 struct fcip_esballoc_arg *fesb_argp; 2837 #endif /* FCIP_ESBALLOC */ 2838 2839 fport = fcip_get_port(phandle); 2840 if (fport == NULL) { 2841 return (FC_UNCLAIMED); 2842 } 2843 2844 fptr = fport->fcipp_fcip; 2845 ASSERT(fptr != NULL); 2846 2847 if (fptr == NULL) { 2848 return (FC_UNCLAIMED); 2849 } 2850 2851 mutex_enter(&fptr->fcip_mutex); 2852 if ((fptr->fcip_flags & (FCIP_DETACHING | FCIP_DETACHED)) || 2853 (fptr->fcip_flags & (FCIP_SUSPENDED | FCIP_POWER_DOWN))) { 2854 mutex_exit(&fptr->fcip_mutex); 2855 rval = FC_UNCLAIMED; 2856 goto data_cb_done; 2857 } 2858 2859 /* 2860 * set fcip flags to indicate we are in the middle of a 2861 * data callback so we can wait till the statechange 2862 * is handled before succeeding/failing the SUSPEND/POWER DOWN. 2863 */ 2864 fptr->fcip_flags |= FCIP_IN_DATA_CB; 2865 mutex_exit(&fptr->fcip_mutex); 2866 2867 FCIP_TNF_PROBE_2((fcip_data_cb, "fcip io", /* CSTYLED */, 2868 tnf_string, msg, "data callback", 2869 tnf_int, instance, ddi_get_instance(fport->fcipp_dip))); 2870 FCIP_DEBUG(FCIP_DEBUG_UPSTREAM, 2871 (CE_NOTE, "fcip%d, data callback", 2872 ddi_get_instance(fport->fcipp_dip))); 2873 2874 /* 2875 * get to the network and snap headers in the payload 2876 */ 2877 nhdr = (fcph_network_hdr_t *)buf->ub_buffer; 2878 snaphdr = (llc_snap_hdr_t *)(buf->ub_buffer + 2879 sizeof (fcph_network_hdr_t)); 2880 2881 hdrlen = sizeof (fcph_network_hdr_t) + sizeof (llc_snap_hdr_t); 2882 2883 /* 2884 * get the IP header to obtain the no. of bytes we need to read 2885 * off from the unsol buffer. This obviously is because not all 2886 * data fills up the unsol buffer completely and the firmware 2887 * doesn't tell us how many valid bytes are in there as well 2888 */ 2889 iphdr = (ipha_t *)(buf->ub_buffer + hdrlen); 2890 snaphdr->pid = BE_16(snaphdr->pid); 2891 type = snaphdr->pid; 2892 2893 FCIP_DEBUG(FCIP_DEBUG_UPSTREAM, 2894 (CE_CONT, "SNAPHDR: dsap %x, ssap %x, ctrl %x\n", 2895 snaphdr->dsap, snaphdr->ssap, snaphdr->ctrl)); 2896 2897 FCIP_DEBUG(FCIP_DEBUG_UPSTREAM, 2898 (CE_CONT, "oui[0] 0x%x oui[1] 0x%x oui[2] 0x%x pid 0x%x\n", 2899 snaphdr->oui[0], snaphdr->oui[1], snaphdr->oui[2], snaphdr->pid)); 2900 2901 /* Authneticate, Authenticate */ 2902 if (type == ETHERTYPE_IP) { 2903 len = hdrlen + BE_16(iphdr->ipha_length); 2904 } else if (type == ETHERTYPE_ARP) { 2905 len = hdrlen + 28; 2906 } else { 2907 len = buf->ub_bufsize; 2908 } 2909 2910 FCIP_DEBUG(FCIP_DEBUG_UPSTREAM, 2911 (CE_CONT, "effective packet length is %d bytes.\n", len)); 2912 2913 if (len < hdrlen || len > FCIP_UB_SIZE) { 2914 FCIP_DEBUG(FCIP_DEBUG_UPSTREAM, 2915 (CE_NOTE, "Incorrect buffer size %d bytes", len)); 2916 rval = FC_UNCLAIMED; 2917 goto data_cb_done; 2918 } 2919 2920 if (buf->ub_frame.type != FC_TYPE_IS8802_SNAP) { 2921 FCIP_DEBUG(FCIP_DEBUG_UPSTREAM, (CE_NOTE, "Not IP/ARP data")); 2922 rval = FC_UNCLAIMED; 2923 goto data_cb_done; 2924 } 2925 2926 FCIP_DEBUG(FCIP_DEBUG_UPSTREAM, (CE_NOTE, "checking wwn")); 2927 2928 if ((fcip_wwn_compare(&nhdr->net_dest_addr, &fport->fcipp_pwwn, 2929 FCIP_COMPARE_NWWN) != 0) && 2930 (!IS_BROADCAST_ADDR(&nhdr->net_dest_addr))) { 2931 rval = FC_UNCLAIMED; 2932 goto data_cb_done; 2933 } else if (fcip_cache_on_arp_broadcast && 2934 IS_BROADCAST_ADDR(&nhdr->net_dest_addr)) { 2935 fcip_cache_arp_broadcast(fptr, buf); 2936 } 2937 2938 FCIP_DEBUG(FCIP_DEBUG_UPSTREAM, (CE_NOTE, "Allocate streams block")); 2939 2940 /* 2941 * Using esballoc instead of allocb should be faster, atleast at 2942 * larger MTUs than 1500 bytes. Someday we'll get there :) 2943 */ 2944 #if defined(FCIP_ESBALLOC) 2945 /* 2946 * allocate memory for the frtn function arg. The Function 2947 * (fcip_ubfree) arg is a struct fcip_esballoc_arg type 2948 * which contains pointers to the unsol buffer and the 2949 * opaque port handle for releasing the unsol buffer back to 2950 * the FCA for reuse 2951 */ 2952 fesb_argp = (struct fcip_esballoc_arg *) 2953 kmem_zalloc(sizeof (struct fcip_esballoc_arg), KM_NOSLEEP); 2954 2955 if (fesb_argp == NULL) { 2956 FCIP_DEBUG(FCIP_DEBUG_UPSTREAM, 2957 (CE_WARN, "esballoc of mblk failed in data_cb")); 2958 rval = FC_UNCLAIMED; 2959 goto data_cb_done; 2960 } 2961 /* 2962 * Check with KM_NOSLEEP 2963 */ 2964 free_ubuf = (frtn_t *)kmem_zalloc(sizeof (frtn_t), KM_NOSLEEP); 2965 if (free_ubuf == NULL) { 2966 kmem_free(fesb_argp, sizeof (struct fcip_esballoc_arg)); 2967 FCIP_DEBUG(FCIP_DEBUG_UPSTREAM, 2968 (CE_WARN, "esballoc of mblk failed in data_cb")); 2969 rval = FC_UNCLAIMED; 2970 goto data_cb_done; 2971 } 2972 2973 fesb_argp->frtnp = free_ubuf; 2974 fesb_argp->buf = buf; 2975 fesb_argp->phandle = phandle; 2976 free_ubuf->free_func = fcip_ubfree; 2977 free_ubuf->free_arg = (char *)fesb_argp; 2978 if ((bp = (mblk_t *)esballoc((unsigned char *)buf->ub_buffer, 2979 len, BPRI_MED, free_ubuf)) == NULL) { 2980 kmem_free(fesb_argp, sizeof (struct fcip_esballoc_arg)); 2981 kmem_free(free_ubuf, sizeof (frtn_t)); 2982 FCIP_DEBUG(FCIP_DEBUG_UPSTREAM, 2983 (CE_WARN, "esballoc of mblk failed in data_cb")); 2984 rval = FC_UNCLAIMED; 2985 goto data_cb_done; 2986 } 2987 #elif !defined(FCIP_ESBALLOC) 2988 /* 2989 * allocate streams mblk and copy the contents of the 2990 * unsolicited buffer into this newly alloc'ed mblk 2991 */ 2992 if ((bp = (mblk_t *)fcip_allocb((size_t)len, BPRI_LO)) == NULL) { 2993 FCIP_DEBUG(FCIP_DEBUG_UPSTREAM, 2994 (CE_WARN, "alloc of mblk failed in data_cb")); 2995 rval = FC_UNCLAIMED; 2996 goto data_cb_done; 2997 } 2998 2999 /* 3000 * Unsolicited buffers handed up to us from the FCA must be 3001 * endian clean so just bcopy the data into our mblk. Else 3002 * we may have to either copy the data byte by byte or 3003 * use the ddi_rep_get* routines to do the copy for us. 3004 */ 3005 bcopy(buf->ub_buffer, bp->b_rptr, len); 3006 3007 /* 3008 * for esballoc'ed mblks - free the UB in the frtn function 3009 * along with the memory allocated for the function arg. 3010 * for allocb'ed mblk - release the unsolicited buffer here 3011 */ 3012 (void) fc_ulp_ubrelease(phandle, 1, &buf->ub_token); 3013 3014 #endif /* FCIP_ESBALLOC */ 3015 3016 bp->b_wptr = bp->b_rptr + len; 3017 fptr->fcip_ipackets++; 3018 3019 if (type == ETHERTYPE_IP) { 3020 mutex_enter(&fptr->fcip_mutex); 3021 fptr->fcip_ub_upstream++; 3022 mutex_exit(&fptr->fcip_mutex); 3023 bp->b_rptr += hdrlen; 3024 3025 /* 3026 * Check if ipq is valid in the sendup thread 3027 */ 3028 if (fcip_sendup_alloc_enque(fptr, bp, NULL) != FC_SUCCESS) { 3029 freemsg(bp); 3030 } 3031 } else { 3032 /* 3033 * We won't get ethernet 802.3 packets in FCIP but we may get 3034 * types other than ETHERTYPE_IP, such as ETHERTYPE_ARP. Let 3035 * fcip_sendup() do the matching. 3036 */ 3037 mutex_enter(&fptr->fcip_mutex); 3038 fptr->fcip_ub_upstream++; 3039 mutex_exit(&fptr->fcip_mutex); 3040 if (fcip_sendup_alloc_enque(fptr, bp, 3041 fcip_accept) != FC_SUCCESS) { 3042 freemsg(bp); 3043 } 3044 } 3045 3046 rval = FC_SUCCESS; 3047 3048 /* 3049 * Unset fcip_flags to indicate we are out of callback and return 3050 */ 3051 data_cb_done: 3052 mutex_enter(&fptr->fcip_mutex); 3053 fptr->fcip_flags &= ~(FCIP_IN_DATA_CB); 3054 mutex_exit(&fptr->fcip_mutex); 3055 return (rval); 3056 } 3057 3058 #if !defined(FCIP_ESBALLOC) 3059 /* 3060 * Allocate a message block for the inbound data to be sent upstream. 3061 */ 3062 static void * 3063 fcip_allocb(size_t size, uint_t pri) 3064 { 3065 mblk_t *mp; 3066 3067 if ((mp = allocb(size, pri)) == NULL) { 3068 return (NULL); 3069 } 3070 return (mp); 3071 } 3072 3073 #endif 3074 3075 /* 3076 * This helper routine kmem cache alloc's a sendup element for enquing 3077 * into the sendup list for callbacks upstream from the dedicated sendup 3078 * thread. We enque the msg buf into the sendup list and cv_signal the 3079 * sendup thread to finish the callback for us. 3080 */ 3081 static int 3082 fcip_sendup_alloc_enque(struct fcip *fptr, mblk_t *mp, struct fcipstr *(*f)()) 3083 { 3084 struct fcip_sendup_elem *msg_elem; 3085 int rval = FC_FAILURE; 3086 3087 FCIP_TNF_PROBE_1((fcip_sendup_alloc_enque, "fcip io", /* CSTYLED */, 3088 tnf_string, msg, "sendup msg enque")); 3089 msg_elem = kmem_cache_alloc(fptr->fcip_sendup_cache, KM_NOSLEEP); 3090 if (msg_elem == NULL) { 3091 /* drop pkt to floor - update stats */ 3092 rval = FC_FAILURE; 3093 goto sendup_alloc_done; 3094 } 3095 msg_elem->fcipsu_mp = mp; 3096 msg_elem->fcipsu_func = f; 3097 3098 mutex_enter(&fptr->fcip_sendup_mutex); 3099 if (fptr->fcip_sendup_head == NULL) { 3100 fptr->fcip_sendup_head = fptr->fcip_sendup_tail = msg_elem; 3101 } else { 3102 fptr->fcip_sendup_tail->fcipsu_next = msg_elem; 3103 fptr->fcip_sendup_tail = msg_elem; 3104 } 3105 fptr->fcip_sendup_cnt++; 3106 cv_signal(&fptr->fcip_sendup_cv); 3107 mutex_exit(&fptr->fcip_sendup_mutex); 3108 rval = FC_SUCCESS; 3109 3110 sendup_alloc_done: 3111 return (rval); 3112 } 3113 3114 /* 3115 * One of the ways of performing the WWN to D_ID mapping required for 3116 * IPFC data is to cache the unsolicited ARP broadcast messages received 3117 * and update the routing table to add entry for the destination port 3118 * if we are the intended recipient of the ARP broadcast message. This is 3119 * one of the methods recommended in the rfc to obtain the WWN to D_ID mapping 3120 * but is not typically used unless enabled. The driver prefers to use the 3121 * nameserver/lilp map to obtain this mapping. 3122 */ 3123 static void 3124 fcip_cache_arp_broadcast(struct fcip *fptr, fc_unsol_buf_t *buf) 3125 { 3126 fcip_port_info_t *fport; 3127 fcph_network_hdr_t *nhdr; 3128 struct fcip_routing_table *frp; 3129 fc_portmap_t map; 3130 3131 fport = fptr->fcip_port_info; 3132 if (fport == NULL) { 3133 return; 3134 } 3135 ASSERT(fport != NULL); 3136 3137 nhdr = (fcph_network_hdr_t *)buf->ub_buffer; 3138 3139 mutex_enter(&fptr->fcip_rt_mutex); 3140 frp = fcip_lookup_rtable(fptr, &nhdr->net_src_addr, FCIP_COMPARE_NWWN); 3141 mutex_exit(&fptr->fcip_rt_mutex); 3142 if (frp == NULL) { 3143 map.map_did.port_id = buf->ub_frame.s_id; 3144 map.map_hard_addr.hard_addr = buf->ub_frame.s_id; 3145 map.map_state = PORT_DEVICE_VALID; 3146 map.map_type = PORT_DEVICE_NEW; 3147 map.map_flags = 0; 3148 map.map_pd = NULL; 3149 bcopy((void *)&nhdr->net_src_addr, (void *)&map.map_pwwn, 3150 sizeof (la_wwn_t)); 3151 bcopy((void *)&nhdr->net_src_addr, (void *)&map.map_nwwn, 3152 sizeof (la_wwn_t)); 3153 fcip_rt_update(fptr, &map, 1); 3154 mutex_enter(&fptr->fcip_rt_mutex); 3155 frp = fcip_lookup_rtable(fptr, &nhdr->net_src_addr, 3156 FCIP_COMPARE_NWWN); 3157 mutex_exit(&fptr->fcip_rt_mutex); 3158 3159 (void) fcip_add_dest(fptr, frp); 3160 } 3161 3162 } 3163 3164 /* 3165 * This is a dedicated thread to do callbacks from fcip's data callback 3166 * routines into the modules upstream. The reason for this thread is 3167 * the data callback function can be called from an interrupt context and 3168 * the upstream modules *can* make calls downstream in the same thread 3169 * context. If the call is to a fabric port which is not yet in our 3170 * routing tables, we may have to query the nameserver/fabric for the 3171 * MAC addr to Port_ID mapping which may be blocking calls. 3172 */ 3173 static void 3174 fcip_sendup_thr(void *arg) 3175 { 3176 struct fcip *fptr = (struct fcip *)arg; 3177 struct fcip_sendup_elem *msg_elem; 3178 queue_t *ip4q = NULL; 3179 3180 CALLB_CPR_INIT(&fptr->fcip_cpr_info, &fptr->fcip_sendup_mutex, 3181 callb_generic_cpr, "fcip_sendup_thr"); 3182 3183 mutex_enter(&fptr->fcip_sendup_mutex); 3184 for (;;) { 3185 3186 while (fptr->fcip_sendup_thr_initted && 3187 fptr->fcip_sendup_head == NULL) { 3188 CALLB_CPR_SAFE_BEGIN(&fptr->fcip_cpr_info); 3189 cv_wait(&fptr->fcip_sendup_cv, 3190 &fptr->fcip_sendup_mutex); 3191 CALLB_CPR_SAFE_END(&fptr->fcip_cpr_info, 3192 &fptr->fcip_sendup_mutex); 3193 } 3194 3195 if (fptr->fcip_sendup_thr_initted == 0) { 3196 break; 3197 } 3198 3199 FCIP_TNF_PROBE_1((fcip_sendup_thr, "fcip io", /* CSTYLED */, 3200 tnf_string, msg, "fcip sendup thr - new msg")); 3201 3202 msg_elem = fptr->fcip_sendup_head; 3203 fptr->fcip_sendup_head = msg_elem->fcipsu_next; 3204 msg_elem->fcipsu_next = NULL; 3205 mutex_exit(&fptr->fcip_sendup_mutex); 3206 3207 if (msg_elem->fcipsu_func == NULL) { 3208 /* 3209 * Message for ipq. Check to see if the ipq is 3210 * is still valid. Since the thread is asynchronous, 3211 * there could have been a close on the stream 3212 */ 3213 mutex_enter(&fptr->fcip_mutex); 3214 if (fptr->fcip_ipq && canputnext(fptr->fcip_ipq)) { 3215 ip4q = fptr->fcip_ipq; 3216 mutex_exit(&fptr->fcip_mutex); 3217 putnext(ip4q, msg_elem->fcipsu_mp); 3218 } else { 3219 mutex_exit(&fptr->fcip_mutex); 3220 freemsg(msg_elem->fcipsu_mp); 3221 } 3222 } else { 3223 fcip_sendup(fptr, msg_elem->fcipsu_mp, 3224 msg_elem->fcipsu_func); 3225 } 3226 3227 #if !defined(FCIP_ESBALLOC) 3228 /* 3229 * for allocb'ed mblk - decrement upstream count here 3230 */ 3231 mutex_enter(&fptr->fcip_mutex); 3232 ASSERT(fptr->fcip_ub_upstream > 0); 3233 fptr->fcip_ub_upstream--; 3234 mutex_exit(&fptr->fcip_mutex); 3235 #endif /* FCIP_ESBALLOC */ 3236 3237 kmem_cache_free(fptr->fcip_sendup_cache, (void *)msg_elem); 3238 mutex_enter(&fptr->fcip_sendup_mutex); 3239 fptr->fcip_sendup_cnt--; 3240 } 3241 3242 3243 #ifndef __lock_lint 3244 CALLB_CPR_EXIT(&fptr->fcip_cpr_info); 3245 #else 3246 mutex_exit(&fptr->fcip_sendup_mutex); 3247 #endif /* __lock_lint */ 3248 3249 /* Wake up fcip detach thread by the end */ 3250 cv_signal(&fptr->fcip_sendup_cv); 3251 3252 thread_exit(); 3253 } 3254 3255 #ifdef FCIP_ESBALLOC 3256 3257 /* 3258 * called from the stream head when it is done using an unsolicited buffer. 3259 * We release this buffer then to the FCA for reuse. 3260 */ 3261 static void 3262 fcip_ubfree(char *arg) 3263 { 3264 struct fcip_esballoc_arg *fesb_argp = (struct fcip_esballoc_arg *)arg; 3265 fc_unsol_buf_t *ubuf; 3266 frtn_t *frtnp; 3267 fcip_port_info_t *fport; 3268 struct fcip *fptr; 3269 3270 3271 fport = fcip_get_port(fesb_argp->phandle); 3272 fptr = fport->fcipp_fcip; 3273 3274 ASSERT(fesb_argp != NULL); 3275 ubuf = fesb_argp->buf; 3276 frtnp = fesb_argp->frtnp; 3277 3278 3279 FCIP_DEBUG(FCIP_DEBUG_UPSTREAM, 3280 (CE_WARN, "freeing ubuf after esballoc in fcip_ubfree")); 3281 (void) fc_ulp_ubrelease(fesb_argp->phandle, 1, &ubuf->ub_token); 3282 3283 mutex_enter(&fptr->fcip_mutex); 3284 ASSERT(fptr->fcip_ub_upstream > 0); 3285 fptr->fcip_ub_upstream--; 3286 cv_signal(&fptr->fcip_ub_cv); 3287 mutex_exit(&fptr->fcip_mutex); 3288 3289 kmem_free(frtnp, sizeof (frtn_t)); 3290 kmem_free(fesb_argp, sizeof (struct fcip_esballoc_arg)); 3291 } 3292 3293 #endif /* FCIP_ESBALLOC */ 3294 3295 /* 3296 * handle data other than that of type ETHERTYPE_IP and send it on its 3297 * way upstream to the right streams module to handle 3298 */ 3299 static void 3300 fcip_sendup(struct fcip *fptr, mblk_t *mp, struct fcipstr *(*acceptfunc)()) 3301 { 3302 struct fcipstr *slp, *nslp; 3303 la_wwn_t *dhostp; 3304 mblk_t *nmp; 3305 uint32_t isgroupaddr; 3306 int type; 3307 uint32_t hdrlen; 3308 fcph_network_hdr_t *nhdr; 3309 llc_snap_hdr_t *snaphdr; 3310 3311 FCIP_TNF_PROBE_1((fcip_sendup, "fcip io", /* CSTYLED */, 3312 tnf_string, msg, "fcip sendup")); 3313 nhdr = (fcph_network_hdr_t *)mp->b_rptr; 3314 snaphdr = 3315 (llc_snap_hdr_t *)(mp->b_rptr + sizeof (fcph_network_hdr_t)); 3316 dhostp = &nhdr->net_dest_addr; 3317 type = snaphdr->pid; 3318 hdrlen = sizeof (fcph_network_hdr_t) + sizeof (llc_snap_hdr_t); 3319 3320 /* No group address with fibre channel */ 3321 isgroupaddr = 0; 3322 3323 /* 3324 * While holding a reader lock on the linked list of streams structures, 3325 * attempt to match the address criteria for each stream 3326 * and pass up the raw M_DATA ("fastpath") or a DL_UNITDATA_IND. 3327 */ 3328 3329 rw_enter(&fcipstruplock, RW_READER); 3330 3331 if ((slp = (*acceptfunc)(fcipstrup, fptr, type, dhostp)) == NULL) { 3332 rw_exit(&fcipstruplock); 3333 FCIP_TNF_PROBE_1((fcip_sendup, "fcip io", /* CSTYLED */, 3334 tnf_string, msg, "fcip sendup - no slp")); 3335 freemsg(mp); 3336 return; 3337 } 3338 3339 /* 3340 * Loop on matching open streams until (*acceptfunc)() returns NULL. 3341 */ 3342 for (; nslp = (*acceptfunc)(slp->sl_nextp, fptr, type, dhostp); 3343 slp = nslp) { 3344 if (canputnext(slp->sl_rq)) { 3345 if (nmp = dupmsg(mp)) { 3346 if ((slp->sl_flags & FCIP_SLFAST) && 3347 !isgroupaddr) { 3348 nmp->b_rptr += hdrlen; 3349 putnext(slp->sl_rq, nmp); 3350 } else if (slp->sl_flags & FCIP_SLRAW) { 3351 /* No headers when FCIP_SLRAW is set */ 3352 putnext(slp->sl_rq, nmp); 3353 } else if ((nmp = fcip_addudind(fptr, nmp, 3354 nhdr, type))) { 3355 putnext(slp->sl_rq, nmp); 3356 } 3357 } 3358 } 3359 } 3360 3361 /* 3362 * Do the last one. 3363 */ 3364 if (canputnext(slp->sl_rq)) { 3365 if (slp->sl_flags & FCIP_SLFAST) { 3366 mp->b_rptr += hdrlen; 3367 putnext(slp->sl_rq, mp); 3368 } else if (slp->sl_flags & FCIP_SLRAW) { 3369 putnext(slp->sl_rq, mp); 3370 } else if ((mp = fcip_addudind(fptr, mp, nhdr, type))) { 3371 putnext(slp->sl_rq, mp); 3372 } 3373 } else { 3374 freemsg(mp); 3375 } 3376 FCIP_TNF_PROBE_1((fcip_sendup, "fcip io", /* CSTYLED */, 3377 tnf_string, msg, "fcip sendup done")); 3378 3379 rw_exit(&fcipstruplock); 3380 } 3381 3382 /* 3383 * Match the stream based on type and wwn if necessary. 3384 * Destination wwn dhostp is passed to this routine is reserved 3385 * for future usage. We don't need to use it right now since port 3386 * to fcip instance mapping is unique and wwn is already validated when 3387 * packet comes to fcip. 3388 */ 3389 /* ARGSUSED */ 3390 static struct fcipstr * 3391 fcip_accept(struct fcipstr *slp, struct fcip *fptr, int type, la_wwn_t *dhostp) 3392 { 3393 t_uscalar_t sap; 3394 3395 FCIP_TNF_PROBE_1((fcip_accept, "fcip io", /* CSTYLED */, 3396 tnf_string, msg, "fcip accept")); 3397 3398 for (; slp; slp = slp->sl_nextp) { 3399 sap = slp->sl_sap; 3400 FCIP_DEBUG(FCIP_DEBUG_UPSTREAM, (CE_CONT, 3401 "fcip_accept: checking next sap = %x, type = %x", 3402 sap, type)); 3403 3404 if ((slp->sl_fcip == fptr) && (type == sap)) { 3405 return (slp); 3406 } 3407 } 3408 return (NULL); 3409 } 3410 3411 /* 3412 * Handle DL_UNITDATA_IND messages 3413 */ 3414 static mblk_t * 3415 fcip_addudind(struct fcip *fptr, mblk_t *mp, fcph_network_hdr_t *nhdr, 3416 int type) 3417 { 3418 dl_unitdata_ind_t *dludindp; 3419 struct fcipdladdr *dlap; 3420 mblk_t *nmp; 3421 int size; 3422 uint32_t hdrlen; 3423 struct ether_addr src_addr; 3424 struct ether_addr dest_addr; 3425 3426 3427 hdrlen = (sizeof (llc_snap_hdr_t) + sizeof (fcph_network_hdr_t)); 3428 mp->b_rptr += hdrlen; 3429 3430 FCIP_TNF_PROBE_1((fcip_addudind, "fcip io", /* CSTYLED */, 3431 tnf_string, msg, "fcip addudind")); 3432 3433 /* 3434 * Allocate an M_PROTO mblk for the DL_UNITDATA_IND. 3435 */ 3436 size = sizeof (dl_unitdata_ind_t) + FCIPADDRL + FCIPADDRL; 3437 if ((nmp = allocb(size, BPRI_LO)) == NULL) { 3438 fptr->fcip_allocbfail++; 3439 freemsg(mp); 3440 return (NULL); 3441 } 3442 DB_TYPE(nmp) = M_PROTO; 3443 nmp->b_wptr = nmp->b_datap->db_lim; 3444 nmp->b_rptr = nmp->b_wptr - size; 3445 3446 /* 3447 * Construct a DL_UNITDATA_IND primitive. 3448 */ 3449 dludindp = (dl_unitdata_ind_t *)nmp->b_rptr; 3450 dludindp->dl_primitive = DL_UNITDATA_IND; 3451 dludindp->dl_dest_addr_length = FCIPADDRL; 3452 dludindp->dl_dest_addr_offset = sizeof (dl_unitdata_ind_t); 3453 dludindp->dl_src_addr_length = FCIPADDRL; 3454 dludindp->dl_src_addr_offset = sizeof (dl_unitdata_ind_t) + FCIPADDRL; 3455 dludindp->dl_group_address = 0; /* not DL_MULTI */ 3456 3457 dlap = (struct fcipdladdr *)(nmp->b_rptr + sizeof (dl_unitdata_ind_t)); 3458 wwn_to_ether(&nhdr->net_dest_addr, &dest_addr); 3459 ether_bcopy(&dest_addr, &dlap->dl_phys); 3460 dlap->dl_sap = (uint16_t)type; 3461 3462 dlap = (struct fcipdladdr *)(nmp->b_rptr + sizeof (dl_unitdata_ind_t) 3463 + FCIPADDRL); 3464 wwn_to_ether(&nhdr->net_src_addr, &src_addr); 3465 ether_bcopy(&src_addr, &dlap->dl_phys); 3466 dlap->dl_sap = (uint16_t)type; 3467 3468 /* 3469 * Link the M_PROTO and M_DATA together. 3470 */ 3471 nmp->b_cont = mp; 3472 return (nmp); 3473 } 3474 3475 3476 /* 3477 * The open routine. For clone opens, we return the next available minor 3478 * no. for the stream to use 3479 */ 3480 /* ARGSUSED */ 3481 static int 3482 fcip_open(queue_t *rq, dev_t *devp, int flag, int sflag, cred_t *credp) 3483 { 3484 struct fcipstr *slp; 3485 struct fcipstr **prevslp; 3486 minor_t minor; 3487 3488 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, (CE_NOTE, "in fcip_open")); 3489 FCIP_TNF_PROBE_1((fcip_open, "fcip io", /* CSTYLED */, 3490 tnf_string, msg, "enter")); 3491 /* 3492 * We need to ensure that the port driver is loaded before 3493 * we proceed 3494 */ 3495 if (ddi_hold_installed_driver(ddi_name_to_major(PORT_DRIVER)) == NULL) { 3496 /* no port driver instances found */ 3497 FCIP_DEBUG(FCIP_DEBUG_STARTUP, (CE_WARN, 3498 "!ddi_hold_installed_driver of fp failed\n")); 3499 return (ENXIO); 3500 } 3501 /* serialize opens */ 3502 rw_enter(&fcipstruplock, RW_WRITER); 3503 3504 prevslp = &fcipstrup; 3505 if (sflag == CLONEOPEN) { 3506 minor = 0; 3507 for (; (slp = *prevslp) != NULL; prevslp = &slp->sl_nextp) { 3508 if (minor < slp->sl_minor) { 3509 break; 3510 } 3511 minor ++; 3512 } 3513 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, (CE_NOTE, 3514 "getmajor returns 0x%x", getmajor(*devp))); 3515 *devp = makedevice(getmajor(*devp), minor); 3516 } else { 3517 minor = getminor(*devp); 3518 } 3519 3520 /* 3521 * check if our qp's private area is already initialized. If yes 3522 * the stream is already open - just return 3523 */ 3524 if (rq->q_ptr) { 3525 goto done; 3526 } 3527 3528 slp = GETSTRUCT(struct fcipstr, 1); 3529 slp->sl_minor = minor; 3530 slp->sl_rq = rq; 3531 slp->sl_sap = 0; 3532 slp->sl_flags = 0; 3533 slp->sl_state = DL_UNATTACHED; 3534 slp->sl_fcip = NULL; 3535 3536 mutex_init(&slp->sl_lock, NULL, MUTEX_DRIVER, NULL); 3537 3538 /* 3539 * link this new stream entry into list of active streams 3540 */ 3541 slp->sl_nextp = *prevslp; 3542 *prevslp = slp; 3543 3544 rq->q_ptr = WR(rq)->q_ptr = (char *)slp; 3545 3546 /* 3547 * Disable automatic enabling of our write service procedures 3548 * we need to control this explicitly. This will prevent 3549 * anyone scheduling of our write service procedures. 3550 */ 3551 noenable(WR(rq)); 3552 3553 done: 3554 rw_exit(&fcipstruplock); 3555 /* 3556 * enable our put and service routines on the read side 3557 */ 3558 qprocson(rq); 3559 3560 /* 3561 * There is only one instance of fcip (instance = 0) 3562 * for multiple instances of hardware 3563 */ 3564 (void) qassociate(rq, 0); /* don't allow drcompat to be pushed */ 3565 return (0); 3566 } 3567 3568 /* 3569 * close an opened stream. The minor no. will then be available for 3570 * future opens. 3571 */ 3572 /* ARGSUSED */ 3573 static int 3574 fcip_close(queue_t *rq, int flag, int otyp, cred_t *credp) 3575 { 3576 struct fcipstr *slp; 3577 struct fcipstr **prevslp; 3578 3579 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, (CE_NOTE, "in fcip_close")); 3580 FCIP_TNF_PROBE_1((fcip_close, "fcip io", /* CSTYLED */, 3581 tnf_string, msg, "enter")); 3582 ASSERT(rq); 3583 /* we should also have the active stream pointer in q_ptr */ 3584 ASSERT(rq->q_ptr); 3585 3586 ddi_rele_driver(ddi_name_to_major(PORT_DRIVER)); 3587 /* 3588 * disable our put and service procedures. We had enabled them 3589 * on open 3590 */ 3591 qprocsoff(rq); 3592 slp = (struct fcipstr *)rq->q_ptr; 3593 3594 /* 3595 * Implicitly detach stream a stream from an interface. 3596 */ 3597 if (slp->sl_fcip) { 3598 fcip_dodetach(slp); 3599 } 3600 3601 (void) qassociate(rq, -1); /* undo association in open */ 3602 3603 rw_enter(&fcipstruplock, RW_WRITER); 3604 3605 /* 3606 * unlink this stream from the active stream list and free it 3607 */ 3608 for (prevslp = &fcipstrup; (slp = *prevslp) != NULL; 3609 prevslp = &slp->sl_nextp) { 3610 if (slp == (struct fcipstr *)rq->q_ptr) { 3611 break; 3612 } 3613 } 3614 3615 /* we should have found slp */ 3616 ASSERT(slp); 3617 3618 *prevslp = slp->sl_nextp; 3619 mutex_destroy(&slp->sl_lock); 3620 kmem_free(slp, sizeof (struct fcipstr)); 3621 rq->q_ptr = WR(rq)->q_ptr = NULL; 3622 3623 rw_exit(&fcipstruplock); 3624 return (0); 3625 } 3626 3627 /* 3628 * This is not an extension of the DDI_DETACH request. This routine 3629 * only detaches a stream from an interface 3630 */ 3631 static void 3632 fcip_dodetach(struct fcipstr *slp) 3633 { 3634 struct fcipstr *tslp; 3635 struct fcip *fptr; 3636 3637 FCIP_DEBUG(FCIP_DEBUG_DETACH, (CE_NOTE, "in fcip_dodetach")); 3638 FCIP_TNF_PROBE_1((fcip_dodetach, "fcip io", /* CSTYLED */, 3639 tnf_string, msg, "enter")); 3640 ASSERT(slp->sl_fcip != NULL); 3641 3642 fptr = slp->sl_fcip; 3643 slp->sl_fcip = NULL; 3644 3645 /* 3646 * we don't support promiscuous mode currently but check 3647 * for and disable any promiscuous mode operation 3648 */ 3649 if (slp->sl_flags & SLALLPHYS) { 3650 slp->sl_flags &= ~SLALLPHYS; 3651 } 3652 3653 /* 3654 * disable ALLMULTI mode if all mulitcast addr are ON 3655 */ 3656 if (slp->sl_flags & SLALLMULTI) { 3657 slp->sl_flags &= ~SLALLMULTI; 3658 } 3659 3660 /* 3661 * we are most likely going to perform multicast by 3662 * broadcasting to the well known addr (D_ID) 0xFFFFFF or 3663 * ALPA 0x00 in case of public loops 3664 */ 3665 3666 3667 /* 3668 * detach unit from device structure. 3669 */ 3670 for (tslp = fcipstrup; tslp != NULL; tslp = tslp->sl_nextp) { 3671 if (tslp->sl_fcip == fptr) { 3672 break; 3673 } 3674 } 3675 if (tslp == NULL) { 3676 FCIP_DEBUG(FCIP_DEBUG_DETACH, (CE_WARN, 3677 "fcip_dodeatch - active stream struct not found")); 3678 3679 /* unregister with Fabric nameserver?? */ 3680 } 3681 slp->sl_state = DL_UNATTACHED; 3682 3683 fcip_setipq(fptr); 3684 } 3685 3686 3687 /* 3688 * Set or clear device ipq pointer. 3689 * Walk thru all the streams on this device, if a ETHERTYPE_IP 3690 * stream is found, assign device ipq to its sl_rq. 3691 */ 3692 static void 3693 fcip_setipq(struct fcip *fptr) 3694 { 3695 struct fcipstr *slp; 3696 int ok = 1; 3697 queue_t *ipq = NULL; 3698 3699 FCIP_DEBUG(FCIP_DEBUG_INIT, (CE_NOTE, "entered fcip_setipq")); 3700 3701 rw_enter(&fcipstruplock, RW_READER); 3702 3703 for (slp = fcipstrup; slp != NULL; slp = slp->sl_nextp) { 3704 if (slp->sl_fcip == fptr) { 3705 if (slp->sl_flags & (SLALLPHYS|SLALLSAP)) { 3706 ok = 0; 3707 } 3708 if (slp->sl_sap == ETHERTYPE_IP) { 3709 if (ipq == NULL) { 3710 ipq = slp->sl_rq; 3711 } else { 3712 ok = 0; 3713 } 3714 } 3715 } 3716 } 3717 3718 rw_exit(&fcipstruplock); 3719 3720 if (fcip_check_port_exists(fptr)) { 3721 /* fptr passed to us is stale */ 3722 return; 3723 } 3724 3725 mutex_enter(&fptr->fcip_mutex); 3726 if (ok) { 3727 fptr->fcip_ipq = ipq; 3728 } else { 3729 fptr->fcip_ipq = NULL; 3730 } 3731 mutex_exit(&fptr->fcip_mutex); 3732 } 3733 3734 3735 /* ARGSUSED */ 3736 static void 3737 fcip_ioctl(queue_t *wq, mblk_t *mp) 3738 { 3739 struct iocblk *iocp = (struct iocblk *)mp->b_rptr; 3740 struct fcipstr *slp = (struct fcipstr *)wq->q_ptr; 3741 3742 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, 3743 (CE_NOTE, "in fcip ioctl : %d", iocp->ioc_cmd)); 3744 FCIP_TNF_PROBE_1((fcip_ioctl, "fcip io", /* CSTYLED */, 3745 tnf_string, msg, "enter")); 3746 3747 switch (iocp->ioc_cmd) { 3748 case DLIOCRAW: 3749 slp->sl_flags |= FCIP_SLRAW; 3750 miocack(wq, mp, 0, 0); 3751 break; 3752 3753 case DL_IOC_HDR_INFO: 3754 fcip_dl_ioc_hdr_info(wq, mp); 3755 break; 3756 3757 default: 3758 miocnak(wq, mp, 0, EINVAL); 3759 break; 3760 } 3761 } 3762 3763 /* 3764 * The streams 'Put' routine. 3765 */ 3766 /* ARGSUSED */ 3767 static int 3768 fcip_wput(queue_t *wq, mblk_t *mp) 3769 { 3770 struct fcipstr *slp = (struct fcipstr *)wq->q_ptr; 3771 struct fcip *fptr; 3772 struct fcip_dest *fdestp; 3773 fcph_network_hdr_t *headerp; 3774 3775 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, 3776 (CE_NOTE, "in fcip_wput :: type:%x", DB_TYPE(mp))); 3777 3778 switch (DB_TYPE(mp)) { 3779 case M_DATA: { 3780 3781 fptr = slp->sl_fcip; 3782 3783 if (((slp->sl_flags & (FCIP_SLFAST|FCIP_SLRAW)) == 0) || 3784 (slp->sl_state != DL_IDLE) || 3785 (fptr == NULL)) { 3786 /* 3787 * set error in the message block and send a reply 3788 * back upstream. Sun's merror routine does this 3789 * for us more cleanly. 3790 */ 3791 merror(wq, mp, EPROTO); 3792 break; 3793 } 3794 3795 /* 3796 * if any messages are already enqueued or if the interface 3797 * is in promiscuous mode, causing the packets to loop back 3798 * up, then enqueue the message. Otherwise just transmit 3799 * the message. putq() puts the message on fcip's 3800 * write queue and qenable() puts the queue (wq) on 3801 * the list of queues to be called by the streams scheduler. 3802 */ 3803 if (wq->q_first) { 3804 (void) putq(wq, mp); 3805 fptr->fcip_wantw = 1; 3806 qenable(wq); 3807 } else if (fptr->fcip_flags & FCIP_PROMISC) { 3808 /* 3809 * Promiscous mode not supported but add this code in 3810 * case it will be supported in future. 3811 */ 3812 (void) putq(wq, mp); 3813 qenable(wq); 3814 } else { 3815 3816 headerp = (fcph_network_hdr_t *)mp->b_rptr; 3817 fdestp = fcip_get_dest(fptr, &headerp->net_dest_addr); 3818 3819 if (fdestp == NULL) { 3820 merror(wq, mp, EPROTO); 3821 break; 3822 } 3823 3824 ASSERT(fdestp != NULL); 3825 3826 (void) fcip_start(wq, mp, fptr, fdestp, KM_SLEEP); 3827 } 3828 break; 3829 } 3830 case M_PROTO: 3831 case M_PCPROTO: 3832 /* 3833 * to prevent recursive calls into fcip_proto 3834 * (PROTO and PCPROTO messages are handled by fcip_proto) 3835 * let the service procedure handle these messages by 3836 * calling putq here. 3837 */ 3838 (void) putq(wq, mp); 3839 qenable(wq); 3840 break; 3841 3842 case M_IOCTL: 3843 fcip_ioctl(wq, mp); 3844 break; 3845 3846 case M_FLUSH: 3847 if (*mp->b_rptr & FLUSHW) { 3848 flushq(wq, FLUSHALL); 3849 *mp->b_rptr &= ~FLUSHW; 3850 } 3851 /* 3852 * we have both FLUSHW and FLUSHR set with FLUSHRW 3853 */ 3854 if (*mp->b_rptr & FLUSHR) { 3855 /* 3856 * send msg back upstream. qreply() takes care 3857 * of using the RD(wq) queue on its reply 3858 */ 3859 qreply(wq, mp); 3860 } else { 3861 freemsg(mp); 3862 } 3863 break; 3864 3865 default: 3866 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, 3867 (CE_NOTE, "default msg type: %x", DB_TYPE(mp))); 3868 freemsg(mp); 3869 break; 3870 } 3871 return (0); 3872 } 3873 3874 3875 /* 3876 * Handle M_PROTO and M_PCPROTO messages 3877 */ 3878 /* ARGSUSED */ 3879 static void 3880 fcip_proto(queue_t *wq, mblk_t *mp) 3881 { 3882 union DL_primitives *dlp; 3883 struct fcipstr *slp; 3884 t_uscalar_t prim; 3885 3886 slp = (struct fcipstr *)wq->q_ptr; 3887 dlp = (union DL_primitives *)mp->b_rptr; 3888 prim = dlp->dl_primitive; /* the DLPI command */ 3889 3890 FCIP_TNF_PROBE_5((fcip_proto, "fcip io", /* CSTYLED */, 3891 tnf_string, msg, "enter", 3892 tnf_opaque, wq, wq, 3893 tnf_opaque, mp, mp, 3894 tnf_opaque, MP_DB_TYPE, DB_TYPE(mp), 3895 tnf_opaque, dl_primitive, dlp->dl_primitive)); 3896 3897 FCIP_DEBUG(FCIP_DEBUG_INIT, (CE_NOTE, "dl_primitve : %x", prim)); 3898 3899 mutex_enter(&slp->sl_lock); 3900 3901 switch (prim) { 3902 case DL_UNITDATA_REQ: 3903 FCIP_TNF_PROBE_1((fcip_proto, "fcip io", /* CSTYLED */, 3904 tnf_string, msg, "unit data request")); 3905 FCIP_DEBUG(FCIP_DEBUG_DLPI, (CE_NOTE, "unit data request")); 3906 fcip_udreq(wq, mp); 3907 break; 3908 3909 case DL_ATTACH_REQ: 3910 FCIP_TNF_PROBE_1((fcip_proto, "fcip io", /* CSTYLED */, 3911 tnf_string, msg, "Attach request")); 3912 FCIP_DEBUG(FCIP_DEBUG_DLPI, (CE_NOTE, "Attach request")); 3913 fcip_areq(wq, mp); 3914 break; 3915 3916 case DL_DETACH_REQ: 3917 FCIP_TNF_PROBE_1((fcip_proto, "fcip io", /* CSTYLED */, 3918 tnf_string, msg, "Detach request")); 3919 FCIP_DEBUG(FCIP_DEBUG_DLPI, (CE_NOTE, "Detach request")); 3920 fcip_dreq(wq, mp); 3921 break; 3922 3923 case DL_BIND_REQ: 3924 FCIP_DEBUG(FCIP_DEBUG_DLPI, (CE_NOTE, "Bind request")); 3925 FCIP_TNF_PROBE_1((fcip_proto, "fcip io", /* CSTYLED */, 3926 tnf_string, msg, "Bind request")); 3927 fcip_breq(wq, mp); 3928 break; 3929 3930 case DL_UNBIND_REQ: 3931 FCIP_TNF_PROBE_1((fcip_proto, "fcip io", /* CSTYLED */, 3932 tnf_string, msg, "unbind request")); 3933 FCIP_DEBUG(FCIP_DEBUG_DLPI, (CE_NOTE, "unbind request")); 3934 fcip_ubreq(wq, mp); 3935 break; 3936 3937 case DL_INFO_REQ: 3938 FCIP_TNF_PROBE_1((fcip_proto, "fcip io", /* CSTYLED */, 3939 tnf_string, msg, "Info request")); 3940 FCIP_DEBUG(FCIP_DEBUG_DLPI, (CE_NOTE, "Info request")); 3941 fcip_ireq(wq, mp); 3942 break; 3943 3944 case DL_SET_PHYS_ADDR_REQ: 3945 FCIP_TNF_PROBE_1((fcip_proto, "fcip io", /* CSTYLED */, 3946 tnf_string, msg, "set phy addr request")); 3947 FCIP_DEBUG(FCIP_DEBUG_DLPI, 3948 (CE_NOTE, "set phy addr request")); 3949 fcip_spareq(wq, mp); 3950 break; 3951 3952 case DL_PHYS_ADDR_REQ: 3953 FCIP_TNF_PROBE_1((fcip_proto, "fcip io", /* CSTYLED */, 3954 tnf_string, msg, "phy addr request")); 3955 FCIP_DEBUG(FCIP_DEBUG_DLPI, (CE_NOTE, "phy addr request")); 3956 fcip_pareq(wq, mp); 3957 break; 3958 3959 case DL_ENABMULTI_REQ: 3960 FCIP_TNF_PROBE_1((fcip_proto, "fcip io", /* CSTYLED */, 3961 tnf_string, msg, "Enable Multicast request")); 3962 FCIP_DEBUG(FCIP_DEBUG_DLPI, 3963 (CE_NOTE, "Enable Multicast request")); 3964 dlerrorack(wq, mp, prim, DL_UNSUPPORTED, 0); 3965 break; 3966 3967 case DL_DISABMULTI_REQ: 3968 FCIP_TNF_PROBE_1((fcip_proto, "fcip io", /* CSTYLED */, 3969 tnf_string, msg, "Disable Multicast request")); 3970 FCIP_DEBUG(FCIP_DEBUG_DLPI, 3971 (CE_NOTE, "Disable Multicast request")); 3972 dlerrorack(wq, mp, prim, DL_UNSUPPORTED, 0); 3973 break; 3974 3975 case DL_PROMISCON_REQ: 3976 FCIP_TNF_PROBE_1((fcip_proto, "fcip io", /* CSTYLED */, 3977 tnf_string, msg, "Promiscuous mode ON request")); 3978 FCIP_DEBUG(FCIP_DEBUG_DLPI, 3979 (CE_NOTE, "Promiscuous mode ON request")); 3980 dlerrorack(wq, mp, prim, DL_UNSUPPORTED, 0); 3981 break; 3982 3983 case DL_PROMISCOFF_REQ: 3984 FCIP_TNF_PROBE_1((fcip_proto, "fcip io", /* CSTYLED */, 3985 tnf_string, msg, "Promiscuous mode OFF request")); 3986 FCIP_DEBUG(FCIP_DEBUG_DLPI, 3987 (CE_NOTE, "Promiscuous mode OFF request")); 3988 dlerrorack(wq, mp, prim, DL_UNSUPPORTED, 0); 3989 break; 3990 3991 default: 3992 FCIP_TNF_PROBE_1((fcip_proto, "fcip io", /* CSTYLED */, 3993 tnf_string, msg, "Unsupported request")); 3994 dlerrorack(wq, mp, prim, DL_UNSUPPORTED, 0); 3995 break; 3996 } 3997 mutex_exit(&slp->sl_lock); 3998 } 3999 4000 /* 4001 * Always enqueue M_PROTO and M_PCPROTO messages pn the wq and M_DATA 4002 * messages sometimes. Processing of M_PROTO and M_PCPROTO messages 4003 * require us to hold fcip's internal locks across (upstream) putnext 4004 * calls. Specifically fcip_intr could hold fcip_intrlock and fcipstruplock 4005 * when it calls putnext(). That thread could loop back around to call 4006 * fcip_wput and eventually fcip_init() to cause a recursive mutex panic 4007 * 4008 * M_DATA messages are enqueued only if we are out of xmit resources. Once 4009 * the transmit resources are available the service procedure is enabled 4010 * and an attempt is made to xmit all messages on the wq. 4011 */ 4012 /* ARGSUSED */ 4013 static int 4014 fcip_wsrv(queue_t *wq) 4015 { 4016 mblk_t *mp; 4017 struct fcipstr *slp; 4018 struct fcip *fptr; 4019 struct fcip_dest *fdestp; 4020 fcph_network_hdr_t *headerp; 4021 4022 slp = (struct fcipstr *)wq->q_ptr; 4023 fptr = slp->sl_fcip; 4024 4025 FCIP_TNF_PROBE_2((fcip_wsrv, "fcip io", /* CSTYLED */, 4026 tnf_string, msg, "enter", 4027 tnf_opaque, wq, wq)); 4028 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, (CE_NOTE, "fcip wsrv")); 4029 4030 while (mp = getq(wq)) { 4031 switch (DB_TYPE(mp)) { 4032 case M_DATA: 4033 if (fptr && mp) { 4034 headerp = (fcph_network_hdr_t *)mp->b_rptr; 4035 fdestp = fcip_get_dest(fptr, 4036 &headerp->net_dest_addr); 4037 if (fdestp == NULL) { 4038 freemsg(mp); 4039 goto done; 4040 } 4041 if (fcip_start(wq, mp, fptr, fdestp, 4042 KM_SLEEP)) { 4043 goto done; 4044 } 4045 } else { 4046 freemsg(mp); 4047 } 4048 break; 4049 4050 case M_PROTO: 4051 case M_PCPROTO: 4052 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, 4053 (CE_NOTE, "PROT msg in wsrv")); 4054 fcip_proto(wq, mp); 4055 break; 4056 default: 4057 break; 4058 } 4059 } 4060 done: 4061 return (0); 4062 } 4063 4064 4065 /* 4066 * This routine is called from fcip_wsrv to send a message downstream 4067 * on the fibre towards its destination. This routine performs the 4068 * actual WWN to D_ID mapping by looking up the routing and destination 4069 * tables. 4070 */ 4071 /* ARGSUSED */ 4072 static int 4073 fcip_start(queue_t *wq, mblk_t *mp, struct fcip *fptr, 4074 struct fcip_dest *fdestp, int flags) 4075 { 4076 int rval; 4077 int free; 4078 fcip_pkt_t *fcip_pkt; 4079 fc_packet_t *fc_pkt; 4080 fcip_port_info_t *fport = fptr->fcip_port_info; 4081 size_t datalen; 4082 4083 FCIP_TNF_PROBE_4((fcip_start, "fcip io", /* CSTYLED */, 4084 tnf_string, msg, "enter", tnf_opaque, wq, wq, 4085 tnf_opaque, mp, mp, 4086 tnf_opaque, MP_DB_TYPE, DB_TYPE(mp))); 4087 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, (CE_NOTE, "in fcipstart")); 4088 4089 ASSERT(fdestp != NULL); 4090 4091 /* 4092 * Only return if port has gone offline and not come back online 4093 * in a while 4094 */ 4095 if (fptr->fcip_flags & FCIP_LINK_DOWN) { 4096 freemsg(mp); 4097 return (0); 4098 } 4099 4100 /* 4101 * The message block coming in here already has the network and 4102 * llc_snap hdr stuffed in 4103 */ 4104 /* 4105 * Traditionally ethernet drivers at sun handle 3 cases here - 4106 * 1. messages with one mblk 4107 * 2. messages with 2 mblks 4108 * 3. messages with >2 mblks 4109 * For now lets handle all the 3 cases in a single case where we 4110 * put them together in one mblk that has all the data 4111 */ 4112 4113 if (mp->b_cont != NULL) { 4114 if (!pullupmsg(mp, -1)) { 4115 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, 4116 (CE_WARN, "failed to concat message")); 4117 freemsg(mp); 4118 return (1); 4119 } 4120 } 4121 4122 datalen = msgsize(mp); 4123 4124 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, (CE_NOTE, 4125 "msgsize with nhdr & llcsnap hdr in fcip_pkt_alloc 0x%lx", 4126 datalen)); 4127 4128 /* 4129 * We cannot have requests larger than FCIPMTU+Headers 4130 */ 4131 if (datalen > (FCIPMTU + sizeof (llc_snap_hdr_t) + 4132 sizeof (fcph_network_hdr_t))) { 4133 freemsg(mp); 4134 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, (CE_NOTE, 4135 "fcip_pkt_alloc: datalen is larger than " 4136 "max possible size.")); 4137 return (1); 4138 } 4139 4140 fcip_pkt = fcip_pkt_alloc(fptr, mp, flags, datalen); 4141 if (fcip_pkt == NULL) { 4142 (void) putbq(wq, mp); 4143 return (1); 4144 } 4145 4146 fcip_pkt->fcip_pkt_mp = mp; 4147 fcip_pkt->fcip_pkt_wq = wq; 4148 fc_pkt = FCIP_PKT_TO_FC_PKT(fcip_pkt); 4149 4150 mutex_enter(&fdestp->fcipd_mutex); 4151 /* 4152 * If the device dynamically disappeared, just fail the request. 4153 */ 4154 if (fdestp->fcipd_rtable == NULL) { 4155 mutex_exit(&fdestp->fcipd_mutex); 4156 fcip_pkt_free(fcip_pkt, 1); 4157 return (1); 4158 } 4159 4160 /* 4161 * Now that we've assigned pkt_pd, we can call fc_ulp_init_packet 4162 */ 4163 4164 fc_pkt->pkt_pd = fdestp->fcipd_pd; 4165 4166 if (fc_ulp_init_packet((opaque_t)fport->fcipp_handle, 4167 fc_pkt, flags) != FC_SUCCESS) { 4168 mutex_exit(&fdestp->fcipd_mutex); 4169 fcip_pkt_free(fcip_pkt, 1); 4170 return (1); 4171 } 4172 4173 fcip_fdestp_enqueue_pkt(fdestp, fcip_pkt); 4174 fcip_pkt->fcip_pkt_dest = fdestp; 4175 fc_pkt->pkt_fca_device = fdestp->fcipd_fca_dev; 4176 4177 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, (CE_NOTE, 4178 "setting cmdlen to 0x%x: rsp 0x%x : data 0x%x", 4179 fc_pkt->pkt_cmdlen, fc_pkt->pkt_rsplen, fc_pkt->pkt_datalen)); 4180 4181 fcip_init_unicast_pkt(fcip_pkt, fport->fcipp_sid, 4182 fdestp->fcipd_did, fcip_pkt_callback); 4183 4184 fdestp->fcipd_ncmds++; 4185 4186 mutex_exit(&fdestp->fcipd_mutex); 4187 if ((rval = fcip_transport(fcip_pkt)) == FC_SUCCESS) { 4188 fptr->fcip_opackets++; 4189 return (0); 4190 } 4191 4192 free = (rval == FC_STATEC_BUSY || rval == FC_OFFLINE || 4193 rval == FC_TRAN_BUSY) ? 0 : 1; 4194 4195 mutex_enter(&fdestp->fcipd_mutex); 4196 rval = fcip_fdestp_dequeue_pkt(fdestp, fcip_pkt); 4197 4198 if (!rval) { 4199 fcip_pkt = NULL; 4200 } else { 4201 fdestp->fcipd_ncmds--; 4202 } 4203 mutex_exit(&fdestp->fcipd_mutex); 4204 4205 if (fcip_pkt != NULL) { 4206 fcip_pkt_free(fcip_pkt, free); 4207 } 4208 4209 if (!free) { 4210 (void) putbq(wq, mp); 4211 } 4212 4213 return (1); 4214 } 4215 4216 4217 /* 4218 * This routine enqueus a packet marked to be issued to the 4219 * transport in the dest structure. This enables us to timeout any 4220 * request stuck with the FCA/transport for long periods of time 4221 * without a response. fcip_pkt_timeout will attempt to clean up 4222 * any packets hung in this state of limbo. 4223 */ 4224 static void 4225 fcip_fdestp_enqueue_pkt(struct fcip_dest *fdestp, fcip_pkt_t *fcip_pkt) 4226 { 4227 ASSERT(mutex_owned(&fdestp->fcipd_mutex)); 4228 FCIP_TNF_PROBE_1((fcip_fdestp_enqueue_pkt, "fcip io", /* CSTYLED */, 4229 tnf_string, msg, "destp enq pkt")); 4230 4231 /* 4232 * Just hang it off the head of packet list 4233 */ 4234 fcip_pkt->fcip_pkt_next = fdestp->fcipd_head; 4235 fcip_pkt->fcip_pkt_prev = NULL; 4236 fcip_pkt->fcip_pkt_flags |= FCIP_PKT_IN_LIST; 4237 4238 if (fdestp->fcipd_head != NULL) { 4239 ASSERT(fdestp->fcipd_head->fcip_pkt_prev == NULL); 4240 fdestp->fcipd_head->fcip_pkt_prev = fcip_pkt; 4241 } 4242 4243 fdestp->fcipd_head = fcip_pkt; 4244 } 4245 4246 4247 /* 4248 * dequeues any packets after the transport/FCA tells us it has 4249 * been successfully sent on its way. Ofcourse it doesn't mean that 4250 * the packet will actually reach its destination but its atleast 4251 * a step closer in that direction 4252 */ 4253 static int 4254 fcip_fdestp_dequeue_pkt(struct fcip_dest *fdestp, fcip_pkt_t *fcip_pkt) 4255 { 4256 fcip_pkt_t *fcipd_pkt; 4257 4258 ASSERT(mutex_owned(&fdestp->fcipd_mutex)); 4259 if (fcip_pkt->fcip_pkt_flags & FCIP_PKT_IN_TIMEOUT) { 4260 fcipd_pkt = fdestp->fcipd_head; 4261 while (fcipd_pkt) { 4262 if (fcipd_pkt == fcip_pkt) { 4263 fcip_pkt_t *pptr = NULL; 4264 4265 if (fcipd_pkt == fdestp->fcipd_head) { 4266 ASSERT(fcipd_pkt->fcip_pkt_prev == 4267 NULL); 4268 fdestp->fcipd_head = 4269 fcipd_pkt->fcip_pkt_next; 4270 } else { 4271 pptr = fcipd_pkt->fcip_pkt_prev; 4272 ASSERT(pptr != NULL); 4273 pptr->fcip_pkt_next = 4274 fcipd_pkt->fcip_pkt_next; 4275 } 4276 if (fcipd_pkt->fcip_pkt_next) { 4277 pptr = fcipd_pkt->fcip_pkt_next; 4278 pptr->fcip_pkt_prev = 4279 fcipd_pkt->fcip_pkt_prev; 4280 } 4281 fcip_pkt->fcip_pkt_flags &= ~FCIP_PKT_IN_LIST; 4282 break; 4283 } 4284 fcipd_pkt = fcipd_pkt->fcip_pkt_next; 4285 } 4286 } else { 4287 if (fcip_pkt->fcip_pkt_prev == NULL) { 4288 ASSERT(fdestp->fcipd_head == fcip_pkt); 4289 fdestp->fcipd_head = fcip_pkt->fcip_pkt_next; 4290 } else { 4291 fcip_pkt->fcip_pkt_prev->fcip_pkt_next = 4292 fcip_pkt->fcip_pkt_next; 4293 } 4294 4295 if (fcip_pkt->fcip_pkt_next) { 4296 fcip_pkt->fcip_pkt_next->fcip_pkt_prev = 4297 fcip_pkt->fcip_pkt_prev; 4298 } 4299 4300 fcipd_pkt = fcip_pkt; 4301 fcip_pkt->fcip_pkt_flags &= ~FCIP_PKT_IN_LIST; 4302 } 4303 4304 return (fcipd_pkt == fcip_pkt); 4305 } 4306 4307 /* 4308 * The transport routine - this is the routine that actually calls 4309 * into the FCA driver (through the transport ofcourse) to transmit a 4310 * datagram on the fibre. The dest struct assoicated with the port to 4311 * which the data is intended is already bound to the packet, this routine 4312 * only takes care of marking the packet a broadcast packet if it is 4313 * intended to be a broadcast request. This permits the transport to send 4314 * the packet down on the wire even if it doesn't have an entry for the 4315 * D_ID in its d_id hash tables. 4316 */ 4317 static int 4318 fcip_transport(fcip_pkt_t *fcip_pkt) 4319 { 4320 struct fcip *fptr; 4321 fc_packet_t *fc_pkt; 4322 fcip_port_info_t *fport; 4323 struct fcip_dest *fdestp; 4324 uint32_t did; 4325 int rval = FC_FAILURE; 4326 struct fcip_routing_table *frp = NULL; 4327 4328 FCIP_TNF_PROBE_1((fcip_transport, "fcip io", /* CSTYLED */, 4329 tnf_string, msg, "enter")); 4330 4331 fptr = fcip_pkt->fcip_pkt_fptr; 4332 fport = fptr->fcip_port_info; 4333 fc_pkt = FCIP_PKT_TO_FC_PKT(fcip_pkt); 4334 fdestp = fcip_pkt->fcip_pkt_dest; 4335 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, (CE_WARN, "fcip_transport called")); 4336 4337 did = fptr->fcip_broadcast_did; 4338 if (fc_pkt->pkt_cmd_fhdr.d_id == did && 4339 fc_pkt->pkt_tran_type != FC_PKT_BROADCAST) { 4340 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, 4341 (CE_NOTE, "trantype set to BROADCAST")); 4342 fc_pkt->pkt_tran_type = FC_PKT_BROADCAST; 4343 } 4344 4345 mutex_enter(&fptr->fcip_mutex); 4346 if ((fc_pkt->pkt_tran_type != FC_PKT_BROADCAST) && 4347 (fc_pkt->pkt_pd == NULL)) { 4348 mutex_exit(&fptr->fcip_mutex); 4349 FCIP_TNF_PROBE_1((fcip_transport, "fcip io", /* CSTYLED */, 4350 tnf_string, msg, "fcip transport no pd")); 4351 return (rval); 4352 } else if (fptr->fcip_port_state == FCIP_PORT_OFFLINE) { 4353 mutex_exit(&fptr->fcip_mutex); 4354 FCIP_TNF_PROBE_1((fcip_transport, "fcip io", /* CSTYLED */, 4355 tnf_string, msg, "fcip transport port offline")); 4356 return (FC_TRAN_BUSY); 4357 } 4358 mutex_exit(&fptr->fcip_mutex); 4359 4360 if (fdestp) { 4361 struct fcip_routing_table *frp; 4362 4363 frp = fdestp->fcipd_rtable; 4364 mutex_enter(&fptr->fcip_rt_mutex); 4365 mutex_enter(&fdestp->fcipd_mutex); 4366 if (fc_pkt->pkt_pd != NULL) { 4367 if ((frp == NULL) || 4368 (frp && FCIP_RTE_UNAVAIL(frp->fcipr_state))) { 4369 mutex_exit(&fdestp->fcipd_mutex); 4370 mutex_exit(&fptr->fcip_rt_mutex); 4371 if (frp && 4372 (frp->fcipr_state == FCIP_RT_INVALID)) { 4373 FCIP_TNF_PROBE_1((fcip_transport, 4374 "fcip io", /* CSTYLED */, 4375 tnf_string, msg, 4376 "fcip transport - TRANBUSY")); 4377 return (FC_TRAN_BUSY); 4378 } else { 4379 FCIP_TNF_PROBE_1((fcip_transport, 4380 "fcip io", /* CSTYLED */, 4381 tnf_string, msg, 4382 "fcip transport: frp unavailable")); 4383 return (rval); 4384 } 4385 } 4386 } 4387 mutex_exit(&fdestp->fcipd_mutex); 4388 mutex_exit(&fptr->fcip_rt_mutex); 4389 ASSERT(fcip_pkt->fcip_pkt_flags & FCIP_PKT_IN_LIST); 4390 } 4391 4392 /* Explicitly invalidate this field till fcip decides to use it */ 4393 fc_pkt->pkt_ulp_rscn_infop = NULL; 4394 4395 rval = fc_ulp_transport(fport->fcipp_handle, fc_pkt); 4396 if (rval == FC_STATEC_BUSY || rval == FC_OFFLINE) { 4397 /* 4398 * Need to queue up the command for retry 4399 */ 4400 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, 4401 (CE_WARN, "ulp_transport failed: 0x%x", rval)); 4402 } else if (rval == FC_LOGINREQ && (frp != NULL)) { 4403 (void) fcip_do_plogi(fptr, frp); 4404 } else if (rval == FC_BADPACKET && (frp != NULL)) { 4405 /* 4406 * There is a distinct possiblity in our scheme of things 4407 * that we have a routing table entry with a NULL pd struct. 4408 * Mark the routing table entry for removal if it is not a 4409 * broadcast entry 4410 */ 4411 if ((frp->fcipr_d_id.port_id != 0x0) && 4412 (frp->fcipr_d_id.port_id != 0xffffff)) { 4413 mutex_enter(&fptr->fcip_rt_mutex); 4414 frp->fcipr_pd = NULL; 4415 frp->fcipr_state = PORT_DEVICE_INVALID; 4416 mutex_exit(&fptr->fcip_rt_mutex); 4417 } 4418 } 4419 4420 FCIP_TNF_PROBE_1((fcip_transport, "fcip io", /* CSTYLED */, 4421 tnf_string, msg, "fcip transport done")); 4422 return (rval); 4423 } 4424 4425 /* 4426 * Call back routine. Called by the FCA/transport when the messages 4427 * has been put onto the wire towards its intended destination. We can 4428 * now free the fc_packet associated with the message 4429 */ 4430 static void 4431 fcip_pkt_callback(fc_packet_t *fc_pkt) 4432 { 4433 int rval; 4434 fcip_pkt_t *fcip_pkt; 4435 struct fcip_dest *fdestp; 4436 4437 fcip_pkt = (fcip_pkt_t *)fc_pkt->pkt_ulp_private; 4438 fdestp = fcip_pkt->fcip_pkt_dest; 4439 4440 /* 4441 * take the lock early so that we don't have a race condition 4442 * with fcip_timeout 4443 * 4444 * fdestp->fcipd_mutex isn't really intended to lock per 4445 * packet struct - see bug 5105592 for permanent solution 4446 */ 4447 mutex_enter(&fdestp->fcipd_mutex); 4448 4449 fcip_pkt->fcip_pkt_flags |= FCIP_PKT_RETURNED; 4450 fcip_pkt->fcip_pkt_flags &= ~FCIP_PKT_IN_ABORT; 4451 if (fcip_pkt->fcip_pkt_flags & FCIP_PKT_IN_TIMEOUT) { 4452 mutex_exit(&fdestp->fcipd_mutex); 4453 return; 4454 } 4455 4456 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, (CE_NOTE, "pkt callback")); 4457 4458 ASSERT(fdestp->fcipd_rtable != NULL); 4459 ASSERT(fcip_pkt->fcip_pkt_flags & FCIP_PKT_IN_LIST); 4460 rval = fcip_fdestp_dequeue_pkt(fdestp, fcip_pkt); 4461 fdestp->fcipd_ncmds--; 4462 mutex_exit(&fdestp->fcipd_mutex); 4463 4464 if (rval) { 4465 fcip_pkt_free(fcip_pkt, 1); 4466 } 4467 4468 FCIP_TNF_PROBE_1((fcip_pkt_callback, "fcip io", /* CSTYLED */, 4469 tnf_string, msg, "pkt callback done")); 4470 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, (CE_NOTE, "pkt callback done")); 4471 } 4472 4473 /* 4474 * Return 1 if the topology is supported, else return 0. 4475 * Topology support is consistent with what the whole 4476 * stack supports together. 4477 */ 4478 static int 4479 fcip_is_supported_fc_topology(int fc_topology) 4480 { 4481 switch (fc_topology) { 4482 4483 case FC_TOP_PRIVATE_LOOP : 4484 case FC_TOP_PUBLIC_LOOP : 4485 case FC_TOP_FABRIC : 4486 case FC_TOP_NO_NS : 4487 return (1); 4488 default : 4489 return (0); 4490 } 4491 } 4492 4493 /* 4494 * handle any topology specific initializations here 4495 * this routine must be called while holding fcip_mutex 4496 */ 4497 /* ARGSUSED */ 4498 static void 4499 fcip_handle_topology(struct fcip *fptr) 4500 { 4501 4502 fcip_port_info_t *fport = fptr->fcip_port_info; 4503 4504 ASSERT(mutex_owned(&fptr->fcip_mutex)); 4505 4506 /* 4507 * Since we know the port's topology - handle topology 4508 * specific details here. In Point to Point and Private Loop 4509 * topologies - we would probably not have a name server 4510 */ 4511 4512 FCIP_TNF_PROBE_3((fcip_handle_topology, "fcip io", /* CSTYLED */, 4513 tnf_string, msg, "enter", 4514 tnf_uint, port_state, fport->fcipp_pstate, 4515 tnf_uint, topology, fport->fcipp_topology)); 4516 FCIP_DEBUG(FCIP_DEBUG_INIT, (CE_NOTE, "port state: %x, topology %x", 4517 fport->fcipp_pstate, fport->fcipp_topology)); 4518 4519 fptr->fcip_broadcast_did = fcip_get_broadcast_did(fptr); 4520 mutex_exit(&fptr->fcip_mutex); 4521 (void) fcip_dest_add_broadcast_entry(fptr, 0); 4522 mutex_enter(&fptr->fcip_mutex); 4523 4524 if (!fcip_is_supported_fc_topology(fport->fcipp_topology)) { 4525 FCIP_DEBUG(FCIP_DEBUG_INIT, 4526 (CE_WARN, "fcip(0x%x): Unsupported port topology (0x%x)", 4527 fptr->fcip_instance, fport->fcipp_topology)); 4528 return; 4529 } 4530 4531 switch (fport->fcipp_topology) { 4532 case FC_TOP_PRIVATE_LOOP: { 4533 4534 fc_portmap_t *port_map; 4535 uint32_t listlen, alloclen; 4536 /* 4537 * we may have to maintain routing. Get a list of 4538 * all devices on this port that the transport layer is 4539 * aware of. Check if any of them is a IS8802 type port, 4540 * if yes get its WWN and DID mapping and cache it in 4541 * the purport routing table. Since there is no 4542 * State Change notification for private loop/point_point 4543 * topologies - this table may not be accurate. The static 4544 * routing table is updated on a state change callback. 4545 */ 4546 FCIP_DEBUG(FCIP_DEBUG_INIT, (CE_WARN, "port state valid!!")); 4547 fptr->fcip_port_state = FCIP_PORT_ONLINE; 4548 listlen = alloclen = FCIP_MAX_PORTS; 4549 port_map = (fc_portmap_t *) 4550 kmem_zalloc((FCIP_MAX_PORTS * sizeof (fc_portmap_t)), 4551 KM_SLEEP); 4552 if (fc_ulp_getportmap(fport->fcipp_handle, &port_map, 4553 &listlen, FC_ULP_PLOGI_PRESERVE) == FC_SUCCESS) { 4554 mutex_exit(&fptr->fcip_mutex); 4555 fcip_rt_update(fptr, port_map, listlen); 4556 mutex_enter(&fptr->fcip_mutex); 4557 } 4558 if (listlen > alloclen) { 4559 alloclen = listlen; 4560 } 4561 kmem_free(port_map, (alloclen * sizeof (fc_portmap_t))); 4562 /* 4563 * Now fall through and register with the transport 4564 * that this port is IP capable 4565 */ 4566 } 4567 /* FALLTHROUGH */ 4568 case FC_TOP_NO_NS: 4569 /* 4570 * If we don't have a nameserver, lets wait until we 4571 * have to send out a packet to a remote port and then 4572 * try and discover the port using ARP/FARP. 4573 */ 4574 /* FALLTHROUGH */ 4575 case FC_TOP_PUBLIC_LOOP: 4576 case FC_TOP_FABRIC: { 4577 fc_portmap_t *port_map; 4578 uint32_t listlen, alloclen; 4579 4580 /* FC_TYPE of 0x05 goes to word 0, LSB */ 4581 fptr->fcip_port_state = FCIP_PORT_ONLINE; 4582 4583 if (!(fptr->fcip_flags & FCIP_REG_INPROGRESS)) { 4584 fptr->fcip_flags |= FCIP_REG_INPROGRESS; 4585 if (taskq_dispatch(fptr->fcip_tq, fcip_port_ns, 4586 fptr, KM_NOSLEEP) == 0) { 4587 fptr->fcip_flags &= ~FCIP_REG_INPROGRESS; 4588 } 4589 } 4590 4591 /* 4592 * If fcip_create_nodes_on_demand is overridden to force 4593 * discovery of all nodes in Fabric/Public loop topologies 4594 * we need to query for and obtain all nodes and log into 4595 * them as with private loop devices 4596 */ 4597 if (!fcip_create_nodes_on_demand) { 4598 fptr->fcip_port_state = FCIP_PORT_ONLINE; 4599 listlen = alloclen = FCIP_MAX_PORTS; 4600 port_map = (fc_portmap_t *) 4601 kmem_zalloc((FCIP_MAX_PORTS * 4602 sizeof (fc_portmap_t)), KM_SLEEP); 4603 if (fc_ulp_getportmap(fport->fcipp_handle, &port_map, 4604 &listlen, FC_ULP_PLOGI_PRESERVE) == FC_SUCCESS) { 4605 mutex_exit(&fptr->fcip_mutex); 4606 fcip_rt_update(fptr, port_map, listlen); 4607 mutex_enter(&fptr->fcip_mutex); 4608 } 4609 if (listlen > alloclen) { 4610 alloclen = listlen; 4611 } 4612 kmem_free(port_map, 4613 (alloclen * sizeof (fc_portmap_t))); 4614 } 4615 break; 4616 } 4617 4618 default: 4619 break; 4620 } 4621 } 4622 4623 static void 4624 fcip_port_ns(void *arg) 4625 { 4626 struct fcip *fptr = (struct fcip *)arg; 4627 fcip_port_info_t *fport = fptr->fcip_port_info; 4628 fc_ns_cmd_t ns_cmd; 4629 uint32_t types[8]; 4630 ns_rfc_type_t rfc; 4631 4632 mutex_enter(&fptr->fcip_mutex); 4633 if ((fptr->fcip_flags & (FCIP_DETACHING | FCIP_DETACHED)) || 4634 (fptr->fcip_flags & (FCIP_SUSPENDED | FCIP_POWER_DOWN))) { 4635 fptr->fcip_flags &= ~FCIP_REG_INPROGRESS; 4636 mutex_exit(&fptr->fcip_mutex); 4637 return; 4638 } 4639 mutex_exit(&fptr->fcip_mutex); 4640 4641 /* 4642 * Prepare the Name server structure to 4643 * register with the transport in case of 4644 * Fabric configuration. 4645 */ 4646 bzero(&rfc, sizeof (rfc)); 4647 bzero(types, sizeof (types)); 4648 4649 types[FC4_TYPE_WORD_POS(FC_TYPE_IS8802_SNAP)] = (1 << 4650 FC4_TYPE_BIT_POS(FC_TYPE_IS8802_SNAP)); 4651 4652 rfc.rfc_port_id.port_id = fport->fcipp_sid.port_id; 4653 bcopy(types, rfc.rfc_types, sizeof (types)); 4654 4655 ns_cmd.ns_flags = 0; 4656 ns_cmd.ns_cmd = NS_RFT_ID; 4657 ns_cmd.ns_req_len = sizeof (rfc); 4658 ns_cmd.ns_req_payload = (caddr_t)&rfc; 4659 ns_cmd.ns_resp_len = 0; 4660 ns_cmd.ns_resp_payload = NULL; 4661 4662 /* 4663 * Perform the Name Server Registration for FC IS8802_SNAP Type. 4664 * We don't expect a reply for registering port type 4665 */ 4666 (void) fc_ulp_port_ns(fptr->fcip_port_info->fcipp_handle, 4667 (opaque_t)0, &ns_cmd); 4668 4669 mutex_enter(&fptr->fcip_mutex); 4670 fptr->fcip_flags &= ~FCIP_REG_INPROGRESS; 4671 mutex_exit(&fptr->fcip_mutex); 4672 } 4673 4674 /* 4675 * setup this instance of fcip. This routine inits kstats, allocates 4676 * unsolicited buffers, determines' this port's siblings and handles 4677 * topology specific details which includes registering with the name 4678 * server and also setting up the routing table for this port for 4679 * private loops and point to point topologies 4680 */ 4681 static int 4682 fcip_init_port(struct fcip *fptr) 4683 { 4684 int rval = FC_SUCCESS; 4685 fcip_port_info_t *fport = fptr->fcip_port_info; 4686 static char buf[64]; 4687 size_t tok_buf_size; 4688 4689 ASSERT(fport != NULL); 4690 4691 FCIP_TNF_PROBE_1((fcip_init_port, "fcip io", /* CSTYLED */, 4692 tnf_string, msg, "enter")); 4693 mutex_enter(&fptr->fcip_mutex); 4694 4695 /* 4696 * setup mac address for this port. Don't be too worried if 4697 * the WWN is zero, there is probably nothing attached to 4698 * to the port. There is no point allocating unsolicited buffers 4699 * for an unused port so return success if we don't have a MAC 4700 * address. Do the port init on a state change notification. 4701 */ 4702 if (fcip_setup_mac_addr(fptr) == FCIP_INVALID_WWN) { 4703 fptr->fcip_port_state = FCIP_PORT_OFFLINE; 4704 rval = FC_SUCCESS; 4705 goto done; 4706 } 4707 4708 /* 4709 * clear routing table hash list for this port 4710 */ 4711 fcip_rt_flush(fptr); 4712 4713 /* 4714 * init kstats for this instance 4715 */ 4716 fcip_kstat_init(fptr); 4717 4718 /* 4719 * Allocate unsolicited buffers 4720 */ 4721 fptr->fcip_ub_nbufs = fcip_ub_nbufs; 4722 tok_buf_size = sizeof (*fptr->fcip_ub_tokens) * fcip_ub_nbufs; 4723 4724 FCIP_TNF_PROBE_2((fcip_init_port, "fcip io", /* CSTYLED */, 4725 tnf_string, msg, "debug", 4726 tnf_int, tokBufsize, tok_buf_size)); 4727 4728 FCIP_DEBUG(FCIP_DEBUG_INIT, 4729 (CE_WARN, "tokBufsize: 0x%lx", tok_buf_size)); 4730 4731 fptr->fcip_ub_tokens = kmem_zalloc(tok_buf_size, KM_SLEEP); 4732 4733 if (fptr->fcip_ub_tokens == NULL) { 4734 rval = FC_FAILURE; 4735 FCIP_DEBUG(FCIP_DEBUG_INIT, 4736 (CE_WARN, "fcip(%d): failed to allocate unsol buf", 4737 fptr->fcip_instance)); 4738 goto done; 4739 } 4740 rval = fc_ulp_uballoc(fport->fcipp_handle, &fptr->fcip_ub_nbufs, 4741 fcip_ub_size, FC_TYPE_IS8802_SNAP, fptr->fcip_ub_tokens); 4742 4743 if (rval != FC_SUCCESS) { 4744 FCIP_DEBUG(FCIP_DEBUG_INIT, 4745 (CE_WARN, "fcip(%d): fc_ulp_uballoc failed with 0x%x!!", 4746 fptr->fcip_instance, rval)); 4747 } 4748 4749 switch (rval) { 4750 case FC_SUCCESS: 4751 break; 4752 4753 case FC_OFFLINE: 4754 fptr->fcip_port_state = FCIP_PORT_OFFLINE; 4755 rval = FC_FAILURE; 4756 goto done; 4757 4758 case FC_UB_ERROR: 4759 FCIP_TNF_PROBE_1((fcip_init_port, "fcip io", /* CSTYLED */, 4760 tnf_string, msg, "invalid ub alloc request")); 4761 FCIP_DEBUG(FCIP_DEBUG_INIT, 4762 (CE_WARN, "invalid ub alloc request !!")); 4763 rval = FC_FAILURE; 4764 goto done; 4765 4766 case FC_FAILURE: 4767 /* 4768 * requested bytes could not be alloced 4769 */ 4770 if (fptr->fcip_ub_nbufs != fcip_ub_nbufs) { 4771 cmn_err(CE_WARN, 4772 "!fcip(0x%x): Failed to alloc unsolicited bufs", 4773 ddi_get_instance(fport->fcipp_dip)); 4774 rval = FC_FAILURE; 4775 goto done; 4776 } 4777 break; 4778 4779 default: 4780 rval = FC_FAILURE; 4781 break; 4782 } 4783 4784 /* 4785 * Preallocate a Cache of fcip packets for transmit and receive 4786 * We don't want to be holding on to unsolicited buffers while 4787 * we transmit the message upstream 4788 */ 4789 FCIP_DEBUG(FCIP_DEBUG_INIT, (CE_NOTE, "allocating fcip_pkt cache")); 4790 4791 (void) sprintf(buf, "fcip%d_cache", fptr->fcip_instance); 4792 fptr->fcip_xmit_cache = kmem_cache_create(buf, 4793 (fport->fcipp_fca_pkt_size + sizeof (fcip_pkt_t)), 4794 8, fcip_cache_constructor, fcip_cache_destructor, 4795 NULL, (void *)fport, NULL, 0); 4796 4797 (void) sprintf(buf, "fcip%d_sendup_cache", fptr->fcip_instance); 4798 fptr->fcip_sendup_cache = kmem_cache_create(buf, 4799 sizeof (struct fcip_sendup_elem), 4800 8, fcip_sendup_constructor, NULL, NULL, (void *)fport, NULL, 0); 4801 4802 if (fptr->fcip_xmit_cache == NULL) { 4803 FCIP_TNF_PROBE_2((fcip_init_port, "fcip io", /* CSTYLED */, 4804 tnf_string, msg, "unable to allocate xmit cache", 4805 tnf_int, instance, fptr->fcip_instance)); 4806 FCIP_DEBUG(FCIP_DEBUG_INIT, 4807 (CE_WARN, "fcip%d unable to allocate xmit cache", 4808 fptr->fcip_instance)); 4809 rval = FC_FAILURE; 4810 goto done; 4811 } 4812 4813 /* 4814 * We may need to handle routing tables for point to point and 4815 * fcal topologies and register with NameServer for Fabric 4816 * topologies. 4817 */ 4818 fcip_handle_topology(fptr); 4819 mutex_exit(&fptr->fcip_mutex); 4820 if (fcip_dest_add_broadcast_entry(fptr, 1) != FC_SUCCESS) { 4821 FCIP_DEBUG(FCIP_DEBUG_INIT, 4822 (CE_WARN, "fcip(0x%x):add broadcast entry failed!!", 4823 fptr->fcip_instance)); 4824 mutex_enter(&fptr->fcip_mutex); 4825 rval = FC_FAILURE; 4826 goto done; 4827 } 4828 4829 rval = FC_SUCCESS; 4830 return (rval); 4831 4832 done: 4833 /* 4834 * we don't always come here from port_attach - so cleanup 4835 * anything done in the init_port routine 4836 */ 4837 if (fptr->fcip_kstatp) { 4838 kstat_delete(fptr->fcip_kstatp); 4839 fptr->fcip_kstatp = NULL; 4840 } 4841 4842 if (fptr->fcip_xmit_cache) { 4843 kmem_cache_destroy(fptr->fcip_xmit_cache); 4844 fptr->fcip_xmit_cache = NULL; 4845 } 4846 4847 if (fptr->fcip_sendup_cache) { 4848 kmem_cache_destroy(fptr->fcip_sendup_cache); 4849 fptr->fcip_sendup_cache = NULL; 4850 } 4851 4852 /* release unsolicited buffers */ 4853 if (fptr->fcip_ub_tokens) { 4854 uint64_t *tokens = fptr->fcip_ub_tokens; 4855 fptr->fcip_ub_tokens = NULL; 4856 4857 mutex_exit(&fptr->fcip_mutex); 4858 (void) fc_ulp_ubfree(fport->fcipp_handle, fptr->fcip_ub_nbufs, 4859 tokens); 4860 kmem_free(tokens, tok_buf_size); 4861 4862 } else { 4863 mutex_exit(&fptr->fcip_mutex); 4864 } 4865 4866 return (rval); 4867 } 4868 4869 /* 4870 * Sets up a port's MAC address from its WWN 4871 */ 4872 static int 4873 fcip_setup_mac_addr(struct fcip *fptr) 4874 { 4875 fcip_port_info_t *fport = fptr->fcip_port_info; 4876 4877 ASSERT(mutex_owned(&fptr->fcip_mutex)); 4878 4879 fptr->fcip_addrflags = 0; 4880 4881 /* 4882 * we cannot choose a MAC address for our interface - we have 4883 * to live with whatever node WWN we get (minus the top two 4884 * MSbytes for the MAC address) from the transport layer. We will 4885 * treat the WWN as our factory MAC address. 4886 */ 4887 4888 if ((fport->fcipp_nwwn.w.wwn_hi != 0) || 4889 (fport->fcipp_nwwn.w.wwn_lo != 0)) { 4890 char etherstr[ETHERSTRL]; 4891 4892 wwn_to_ether(&fport->fcipp_nwwn, &fptr->fcip_macaddr); 4893 fcip_ether_to_str(&fptr->fcip_macaddr, etherstr); 4894 FCIP_DEBUG(FCIP_DEBUG_INIT, 4895 (CE_NOTE, "setupmacaddr ouraddr %s", etherstr)); 4896 4897 fptr->fcip_addrflags = (FCIP_FACTADDR_PRESENT | 4898 FCIP_FACTADDR_USE); 4899 } else { 4900 /* 4901 * No WWN - just return failure - there's not much 4902 * we can do since we cannot set the WWN. 4903 */ 4904 FCIP_DEBUG(FCIP_DEBUG_INIT, 4905 (CE_WARN, "Port does not have a valid WWN")); 4906 return (FCIP_INVALID_WWN); 4907 } 4908 return (FC_SUCCESS); 4909 } 4910 4911 4912 /* 4913 * flush routing table entries 4914 */ 4915 static void 4916 fcip_rt_flush(struct fcip *fptr) 4917 { 4918 int index; 4919 4920 mutex_enter(&fptr->fcip_rt_mutex); 4921 for (index = 0; index < FCIP_RT_HASH_ELEMS; index++) { 4922 struct fcip_routing_table *frtp, *frtp_next; 4923 frtp = fptr->fcip_rtable[index]; 4924 while (frtp) { 4925 frtp_next = frtp->fcipr_next; 4926 kmem_free(frtp, sizeof (struct fcip_routing_table)); 4927 frtp = frtp_next; 4928 } 4929 fptr->fcip_rtable[index] = NULL; 4930 } 4931 mutex_exit(&fptr->fcip_rt_mutex); 4932 } 4933 4934 /* 4935 * Free up the fcip softstate and all allocated resources for the 4936 * fcip instance assoicated with a given port driver instance 4937 * 4938 * Given that the list of structures pointed to by fcip_port_head, 4939 * this function is called from multiple sources, and the 4940 * fcip_global_mutex that protects fcip_port_head must be dropped, 4941 * our best solution is to return a value that indicates the next 4942 * port in the list. This way the caller doesn't need to worry 4943 * about the race condition where it saves off a pointer to the 4944 * next structure in the list and by the time this routine returns, 4945 * that next structure has already been freed. 4946 */ 4947 static fcip_port_info_t * 4948 fcip_softstate_free(fcip_port_info_t *fport) 4949 { 4950 struct fcip *fptr = NULL; 4951 int instance; 4952 timeout_id_t tid; 4953 opaque_t phandle = NULL; 4954 fcip_port_info_t *prev_fport, *cur_fport, *next_fport = NULL; 4955 4956 ASSERT(MUTEX_HELD(&fcip_global_mutex)); 4957 4958 if (fport) { 4959 phandle = fport->fcipp_handle; 4960 fptr = fport->fcipp_fcip; 4961 } else { 4962 return (next_fport); 4963 } 4964 4965 if (fptr) { 4966 mutex_enter(&fptr->fcip_mutex); 4967 instance = ddi_get_instance(fptr->fcip_dip); 4968 4969 /* 4970 * dismantle timeout thread for this instance of fcip 4971 */ 4972 tid = fptr->fcip_timeout_id; 4973 fptr->fcip_timeout_id = NULL; 4974 4975 mutex_exit(&fptr->fcip_mutex); 4976 (void) untimeout(tid); 4977 mutex_enter(&fptr->fcip_mutex); 4978 4979 ASSERT(fcip_num_instances >= 0); 4980 fcip_num_instances--; 4981 4982 /* 4983 * stop sendup thread 4984 */ 4985 mutex_enter(&fptr->fcip_sendup_mutex); 4986 if (fptr->fcip_sendup_thr_initted) { 4987 fptr->fcip_sendup_thr_initted = 0; 4988 cv_signal(&fptr->fcip_sendup_cv); 4989 cv_wait(&fptr->fcip_sendup_cv, 4990 &fptr->fcip_sendup_mutex); 4991 } 4992 ASSERT(fptr->fcip_sendup_head == NULL); 4993 fptr->fcip_sendup_head = fptr->fcip_sendup_tail = NULL; 4994 mutex_exit(&fptr->fcip_sendup_mutex); 4995 4996 /* 4997 * dismantle taskq 4998 */ 4999 if (fptr->fcip_tq) { 5000 taskq_t *tq = fptr->fcip_tq; 5001 5002 fptr->fcip_tq = NULL; 5003 5004 mutex_exit(&fptr->fcip_mutex); 5005 taskq_destroy(tq); 5006 mutex_enter(&fptr->fcip_mutex); 5007 } 5008 5009 if (fptr->fcip_kstatp) { 5010 kstat_delete(fptr->fcip_kstatp); 5011 fptr->fcip_kstatp = NULL; 5012 } 5013 5014 /* flush the routing table entries */ 5015 fcip_rt_flush(fptr); 5016 5017 if (fptr->fcip_xmit_cache) { 5018 kmem_cache_destroy(fptr->fcip_xmit_cache); 5019 fptr->fcip_xmit_cache = NULL; 5020 } 5021 5022 if (fptr->fcip_sendup_cache) { 5023 kmem_cache_destroy(fptr->fcip_sendup_cache); 5024 fptr->fcip_sendup_cache = NULL; 5025 } 5026 5027 fcip_cleanup_dest(fptr); 5028 5029 /* release unsolicited buffers */ 5030 if (fptr->fcip_ub_tokens) { 5031 uint64_t *tokens = fptr->fcip_ub_tokens; 5032 5033 fptr->fcip_ub_tokens = NULL; 5034 mutex_exit(&fptr->fcip_mutex); 5035 if (phandle) { 5036 /* 5037 * release the global mutex here to 5038 * permit any data pending callbacks to 5039 * complete. Else we will deadlock in the 5040 * FCA waiting for all unsol buffers to be 5041 * returned. 5042 */ 5043 mutex_exit(&fcip_global_mutex); 5044 (void) fc_ulp_ubfree(phandle, 5045 fptr->fcip_ub_nbufs, tokens); 5046 mutex_enter(&fcip_global_mutex); 5047 } 5048 kmem_free(tokens, (sizeof (*tokens) * fcip_ub_nbufs)); 5049 } else { 5050 mutex_exit(&fptr->fcip_mutex); 5051 } 5052 5053 mutex_destroy(&fptr->fcip_mutex); 5054 mutex_destroy(&fptr->fcip_ub_mutex); 5055 mutex_destroy(&fptr->fcip_rt_mutex); 5056 mutex_destroy(&fptr->fcip_dest_mutex); 5057 mutex_destroy(&fptr->fcip_sendup_mutex); 5058 cv_destroy(&fptr->fcip_farp_cv); 5059 cv_destroy(&fptr->fcip_sendup_cv); 5060 cv_destroy(&fptr->fcip_ub_cv); 5061 5062 ddi_soft_state_free(fcip_softp, instance); 5063 } 5064 5065 /* 5066 * Now dequeue the fcip_port_info from the port list 5067 */ 5068 cur_fport = fcip_port_head; 5069 prev_fport = NULL; 5070 while (cur_fport != NULL) { 5071 if (cur_fport == fport) { 5072 break; 5073 } 5074 prev_fport = cur_fport; 5075 cur_fport = cur_fport->fcipp_next; 5076 } 5077 5078 /* 5079 * Assert that we found a port in our port list 5080 */ 5081 ASSERT(cur_fport == fport); 5082 5083 if (prev_fport) { 5084 /* 5085 * Not the first port in the port list 5086 */ 5087 prev_fport->fcipp_next = fport->fcipp_next; 5088 } else { 5089 /* 5090 * first port 5091 */ 5092 fcip_port_head = fport->fcipp_next; 5093 } 5094 next_fport = fport->fcipp_next; 5095 kmem_free(fport, sizeof (fcip_port_info_t)); 5096 5097 return (next_fport); 5098 } 5099 5100 5101 /* 5102 * This is called by transport for any ioctl operations performed 5103 * on the devctl or other transport minor nodes. It is currently 5104 * unused for fcip 5105 */ 5106 /* ARGSUSED */ 5107 static int 5108 fcip_port_ioctl(opaque_t ulp_handle, opaque_t port_handle, dev_t dev, 5109 int cmd, intptr_t data, int mode, cred_t *credp, int *rval, 5110 uint32_t claimed) 5111 { 5112 return (FC_UNCLAIMED); 5113 } 5114 5115 /* 5116 * DL_INFO_REQ - returns information about the DLPI stream to the DLS user 5117 * requesting information about this interface 5118 */ 5119 static void 5120 fcip_ireq(queue_t *wq, mblk_t *mp) 5121 { 5122 struct fcipstr *slp; 5123 struct fcip *fptr; 5124 dl_info_ack_t *dlip; 5125 struct fcipdladdr *dlap; 5126 la_wwn_t *ep; 5127 int size; 5128 char etherstr[ETHERSTRL]; 5129 5130 slp = (struct fcipstr *)wq->q_ptr; 5131 5132 fptr = slp->sl_fcip; 5133 5134 FCIP_DEBUG(FCIP_DEBUG_DLPI, 5135 (CE_NOTE, "fcip_ireq: info request req rcvd")); 5136 5137 FCIP_TNF_PROBE_1((fcip_ireq, "fcip io", /* CSTYLED */, 5138 tnf_string, msg, "fcip ireq entered")); 5139 5140 if (MBLKL(mp) < DL_INFO_REQ_SIZE) { 5141 dlerrorack(wq, mp, DL_INFO_REQ, DL_BADPRIM, 0); 5142 return; 5143 } 5144 5145 /* 5146 * Exchange current message for a DL_INFO_ACK 5147 */ 5148 size = sizeof (dl_info_ack_t) + FCIPADDRL + ETHERADDRL; 5149 if ((mp = mexchange(wq, mp, size, M_PCPROTO, DL_INFO_ACK)) == NULL) { 5150 return; 5151 } 5152 5153 /* 5154 * FILL in the DL_INFO_ACK fields and reply 5155 */ 5156 dlip = (dl_info_ack_t *)mp->b_rptr; 5157 *dlip = fcip_infoack; 5158 dlip->dl_current_state = slp->sl_state; 5159 dlap = (struct fcipdladdr *)(mp->b_rptr + dlip->dl_addr_offset); 5160 dlap->dl_sap = slp->sl_sap; 5161 5162 5163 if (fptr) { 5164 fcip_ether_to_str(&fptr->fcip_macaddr, etherstr); 5165 FCIP_DEBUG(FCIP_DEBUG_DLPI, 5166 (CE_NOTE, "ireq - our mac: %s", etherstr)); 5167 ether_bcopy(&fptr->fcip_macaddr, &dlap->dl_phys); 5168 } else { 5169 bzero((caddr_t)&dlap->dl_phys, ETHERADDRL); 5170 } 5171 5172 ep = (la_wwn_t *)(mp->b_rptr + dlip->dl_brdcst_addr_offset); 5173 ether_bcopy(&fcip_arpbroadcast_addr, ep); 5174 5175 FCIP_DEBUG(FCIP_DEBUG_DLPI, (CE_NOTE, "sending back info req..")); 5176 qreply(wq, mp); 5177 } 5178 5179 5180 /* 5181 * To handle DL_UNITDATA_REQ requests. 5182 */ 5183 5184 static void 5185 fcip_udreq(queue_t *wq, mblk_t *mp) 5186 { 5187 struct fcipstr *slp; 5188 struct fcip *fptr; 5189 fcip_port_info_t *fport; 5190 dl_unitdata_req_t *dludp; 5191 mblk_t *nmp; 5192 struct fcipdladdr *dlap; 5193 fcph_network_hdr_t *headerp; 5194 llc_snap_hdr_t *lsnap; 5195 t_uscalar_t off, len; 5196 struct fcip_dest *fdestp; 5197 la_wwn_t wwn; 5198 int hdr_size; 5199 5200 FCIP_DEBUG(FCIP_DEBUG_DLPI, (CE_NOTE, "inside fcip_udreq")); 5201 5202 FCIP_TNF_PROBE_1((fcip_udreq, "fcip io", /* CSTYLED */, 5203 tnf_string, msg, "fcip udreq entered")); 5204 5205 slp = (struct fcipstr *)wq->q_ptr; 5206 5207 if (slp->sl_state != DL_IDLE) { 5208 dlerrorack(wq, mp, DL_UNITDATA_REQ, DL_OUTSTATE, 0); 5209 return; 5210 } 5211 5212 fptr = slp->sl_fcip; 5213 5214 if (fptr == NULL) { 5215 dlerrorack(wq, mp, DL_UNITDATA_REQ, DL_OUTSTATE, 0); 5216 return; 5217 } 5218 5219 fport = fptr->fcip_port_info; 5220 5221 dludp = (dl_unitdata_req_t *)mp->b_rptr; 5222 off = dludp->dl_dest_addr_offset; 5223 len = dludp->dl_dest_addr_length; 5224 5225 /* 5226 * Validate destination address format 5227 */ 5228 if (!MBLKIN(mp, off, len) || (len != FCIPADDRL)) { 5229 dluderrorind(wq, mp, (mp->b_rptr + off), len, DL_BADADDR, 0); 5230 return; 5231 } 5232 5233 /* 5234 * Error if no M_DATA follows 5235 */ 5236 nmp = mp->b_cont; 5237 if (nmp == NULL) { 5238 dluderrorind(wq, mp, (mp->b_rptr + off), len, DL_BADDATA, 0); 5239 return; 5240 } 5241 dlap = (struct fcipdladdr *)(mp->b_rptr + off); 5242 5243 /* 5244 * Now get the destination structure for the remote NPORT 5245 */ 5246 ether_to_wwn(&dlap->dl_phys, &wwn); 5247 fdestp = fcip_get_dest(fptr, &wwn); 5248 5249 if (fdestp == NULL) { 5250 FCIP_DEBUG(FCIP_DEBUG_DLPI, (CE_NOTE, 5251 "udreq - couldn't find dest struct for remote port"); 5252 dluderrorind(wq, mp, (mp->b_rptr + off), len, DL_BADDATA, 0)); 5253 return; 5254 } 5255 5256 /* 5257 * Network header + SAP 5258 */ 5259 hdr_size = sizeof (fcph_network_hdr_t) + sizeof (llc_snap_hdr_t); 5260 5261 /* DB_REF gives the no. of msgs pointing to this block */ 5262 if ((DB_REF(nmp) == 1) && 5263 (MBLKHEAD(nmp) >= hdr_size) && 5264 (((uintptr_t)mp->b_rptr & 0x1) == 0)) { 5265 la_wwn_t wwn; 5266 nmp->b_rptr -= hdr_size; 5267 5268 /* first put the network header */ 5269 headerp = (fcph_network_hdr_t *)nmp->b_rptr; 5270 if (ether_cmp(&dlap->dl_phys, &fcip_arpbroadcast_addr) == 0) { 5271 ether_to_wwn(&fcipnhbroadcastaddr, &wwn); 5272 } else { 5273 ether_to_wwn(&dlap->dl_phys, &wwn); 5274 } 5275 bcopy(&wwn, &headerp->net_dest_addr, sizeof (la_wwn_t)); 5276 ether_to_wwn(&fptr->fcip_macaddr, &wwn); 5277 bcopy(&wwn, &headerp->net_src_addr, sizeof (la_wwn_t)); 5278 5279 /* Now the snap header */ 5280 lsnap = (llc_snap_hdr_t *)(nmp->b_rptr + 5281 sizeof (fcph_network_hdr_t)); 5282 lsnap->dsap = 0xAA; 5283 lsnap->ssap = 0xAA; 5284 lsnap->ctrl = 0x03; 5285 lsnap->oui[0] = 0x00; 5286 lsnap->oui[1] = 0x00; /* 80 */ 5287 lsnap->oui[2] = 0x00; /* C2 */ 5288 lsnap->pid = BE_16((dlap->dl_sap)); 5289 5290 freeb(mp); 5291 mp = nmp; 5292 5293 } else { 5294 la_wwn_t wwn; 5295 5296 DB_TYPE(mp) = M_DATA; 5297 headerp = (fcph_network_hdr_t *)mp->b_rptr; 5298 5299 /* 5300 * Only fill in the low 48bits of WWN for now - we can 5301 * fill in the NAA_ID after we find the port in the 5302 * routing tables 5303 */ 5304 if (ether_cmp(&dlap->dl_phys, &fcip_arpbroadcast_addr) == 0) { 5305 ether_to_wwn(&fcipnhbroadcastaddr, &wwn); 5306 } else { 5307 ether_to_wwn(&dlap->dl_phys, &wwn); 5308 } 5309 bcopy(&wwn, &headerp->net_dest_addr, sizeof (la_wwn_t)); 5310 /* need to send our PWWN */ 5311 bcopy(&fport->fcipp_pwwn, &headerp->net_src_addr, 5312 sizeof (la_wwn_t)); 5313 5314 lsnap = (llc_snap_hdr_t *)(nmp->b_rptr + 5315 sizeof (fcph_network_hdr_t)); 5316 lsnap->dsap = 0xAA; 5317 lsnap->ssap = 0xAA; 5318 lsnap->ctrl = 0x03; 5319 lsnap->oui[0] = 0x00; 5320 lsnap->oui[1] = 0x00; 5321 lsnap->oui[2] = 0x00; 5322 lsnap->pid = BE_16(dlap->dl_sap); 5323 5324 mp->b_wptr = mp->b_rptr + hdr_size; 5325 } 5326 5327 /* 5328 * Ethernet drivers have a lot of gunk here to put the Type 5329 * information (for Ethernet encapsulation (RFC 894) or the 5330 * Length (for 802.2/802.3) - I guess we'll just ignore that 5331 * here. 5332 */ 5333 5334 /* 5335 * Start the I/O on this port. If fcip_start failed for some reason 5336 * we call putbq in fcip_start so we don't need to check the 5337 * return value from fcip_start 5338 */ 5339 (void) fcip_start(wq, mp, fptr, fdestp, KM_SLEEP); 5340 } 5341 5342 /* 5343 * DL_ATTACH_REQ: attaches a PPA with a stream. ATTACH requets are needed 5344 * for style 2 DLS providers to identify the physical medium through which 5345 * the streams communication will happen 5346 */ 5347 static void 5348 fcip_areq(queue_t *wq, mblk_t *mp) 5349 { 5350 struct fcipstr *slp; 5351 union DL_primitives *dlp; 5352 fcip_port_info_t *fport; 5353 struct fcip *fptr; 5354 int ppa; 5355 5356 slp = (struct fcipstr *)wq->q_ptr; 5357 dlp = (union DL_primitives *)mp->b_rptr; 5358 5359 if (MBLKL(mp) < DL_ATTACH_REQ_SIZE) { 5360 dlerrorack(wq, mp, DL_ATTACH_REQ, DL_BADPRIM, 0); 5361 return; 5362 } 5363 5364 if (slp->sl_state != DL_UNATTACHED) { 5365 dlerrorack(wq, mp, DL_ATTACH_REQ, DL_OUTSTATE, 0); 5366 return; 5367 } 5368 5369 ppa = dlp->attach_req.dl_ppa; 5370 FCIP_DEBUG(FCIP_DEBUG_DLPI, (CE_NOTE, "attach req: ppa %x", ppa)); 5371 5372 /* 5373 * check if the PPA is valid 5374 */ 5375 5376 mutex_enter(&fcip_global_mutex); 5377 5378 for (fport = fcip_port_head; fport; fport = fport->fcipp_next) { 5379 if ((fptr = fport->fcipp_fcip) == NULL) { 5380 continue; 5381 } 5382 FCIP_DEBUG(FCIP_DEBUG_DLPI, (CE_NOTE, "ppa %x, inst %x", ppa, 5383 ddi_get_instance(fptr->fcip_dip))); 5384 5385 if (ppa == ddi_get_instance(fptr->fcip_dip)) { 5386 FCIP_DEBUG(FCIP_DEBUG_DLPI, 5387 (CE_NOTE, "ppa found %x", ppa)); 5388 break; 5389 } 5390 } 5391 5392 if (fport == NULL) { 5393 FCIP_DEBUG(FCIP_DEBUG_DLPI, 5394 (CE_NOTE, "dlerrorack coz fport==NULL")); 5395 5396 mutex_exit(&fcip_global_mutex); 5397 5398 if (fc_ulp_get_port_handle(ppa) == NULL) { 5399 dlerrorack(wq, mp, DL_ATTACH_REQ, DL_BADPPA, 0); 5400 return; 5401 } 5402 5403 /* 5404 * Wait for Port attach callback to trigger. If port_detach 5405 * got in while we were waiting, then ddi_get_soft_state 5406 * will return NULL, and we'll return error. 5407 */ 5408 5409 delay(drv_usectohz(FCIP_INIT_DELAY)); 5410 mutex_enter(&fcip_global_mutex); 5411 5412 fptr = ddi_get_soft_state(fcip_softp, ppa); 5413 if (fptr == NULL) { 5414 mutex_exit(&fcip_global_mutex); 5415 dlerrorack(wq, mp, DL_ATTACH_REQ, DL_BADPPA, 0); 5416 return; 5417 } 5418 } 5419 5420 /* 5421 * set link to device and update our state 5422 */ 5423 slp->sl_fcip = fptr; 5424 slp->sl_state = DL_UNBOUND; 5425 5426 mutex_exit(&fcip_global_mutex); 5427 5428 #ifdef DEBUG 5429 mutex_enter(&fptr->fcip_mutex); 5430 if (fptr->fcip_flags & FCIP_LINK_DOWN) { 5431 FCIP_DEBUG(FCIP_DEBUG_DLPI, (CE_WARN, "port not online yet")); 5432 } 5433 mutex_exit(&fptr->fcip_mutex); 5434 #endif 5435 5436 dlokack(wq, mp, DL_ATTACH_REQ); 5437 } 5438 5439 5440 /* 5441 * DL_DETACH request - detaches a PPA from a stream 5442 */ 5443 static void 5444 fcip_dreq(queue_t *wq, mblk_t *mp) 5445 { 5446 struct fcipstr *slp; 5447 5448 slp = (struct fcipstr *)wq->q_ptr; 5449 5450 if (MBLKL(mp) < DL_DETACH_REQ_SIZE) { 5451 dlerrorack(wq, mp, DL_DETACH_REQ, DL_BADPRIM, 0); 5452 return; 5453 } 5454 5455 if (slp->sl_state != DL_UNBOUND) { 5456 dlerrorack(wq, mp, DL_DETACH_REQ, DL_OUTSTATE, 0); 5457 return; 5458 } 5459 5460 fcip_dodetach(slp); 5461 dlokack(wq, mp, DL_DETACH_REQ); 5462 } 5463 5464 /* 5465 * DL_BIND request: requests a DLS provider to bind a DLSAP to the stream. 5466 * DLS users communicate with a physical interface through DLSAPs. Multiple 5467 * DLSAPs can be bound to the same stream (PPA) 5468 */ 5469 static void 5470 fcip_breq(queue_t *wq, mblk_t *mp) 5471 { 5472 struct fcipstr *slp; 5473 union DL_primitives *dlp; 5474 struct fcip *fptr; 5475 struct fcipdladdr fcipaddr; 5476 t_uscalar_t sap; 5477 int xidtest; 5478 5479 slp = (struct fcipstr *)wq->q_ptr; 5480 5481 if (MBLKL(mp) < DL_BIND_REQ_SIZE) { 5482 dlerrorack(wq, mp, DL_BIND_REQ, DL_BADPRIM, 0); 5483 return; 5484 } 5485 5486 if (slp->sl_state != DL_UNBOUND) { 5487 dlerrorack(wq, mp, DL_BIND_REQ, DL_OUTSTATE, 0); 5488 return; 5489 } 5490 5491 dlp = (union DL_primitives *)mp->b_rptr; 5492 fptr = slp->sl_fcip; 5493 5494 if (fptr == NULL) { 5495 dlerrorack(wq, mp, DL_BIND_REQ, DL_OUTSTATE, 0); 5496 return; 5497 } 5498 5499 sap = dlp->bind_req.dl_sap; 5500 FCIP_DEBUG(FCIP_DEBUG_DLPI, (CE_NOTE, "fcip_breq - sap: %x", sap)); 5501 xidtest = dlp->bind_req.dl_xidtest_flg; 5502 5503 if (xidtest) { 5504 dlerrorack(wq, mp, DL_BIND_REQ, DL_NOAUTO, 0); 5505 return; 5506 } 5507 5508 FCIP_DEBUG(FCIP_DEBUG_DLPI, (CE_NOTE, "DLBIND: sap : %x", sap)); 5509 5510 if (sap > ETHERTYPE_MAX) { 5511 dlerrorack(wq, mp, dlp->dl_primitive, DL_BADSAP, 0); 5512 return; 5513 } 5514 /* 5515 * save SAP for this stream and change the link state 5516 */ 5517 slp->sl_sap = sap; 5518 slp->sl_state = DL_IDLE; 5519 5520 fcipaddr.dl_sap = sap; 5521 ether_bcopy(&fptr->fcip_macaddr, &fcipaddr.dl_phys); 5522 dlbindack(wq, mp, sap, &fcipaddr, FCIPADDRL, 0, 0); 5523 5524 fcip_setipq(fptr); 5525 } 5526 5527 /* 5528 * DL_UNBIND request to unbind a previously bound DLSAP, from this stream 5529 */ 5530 static void 5531 fcip_ubreq(queue_t *wq, mblk_t *mp) 5532 { 5533 struct fcipstr *slp; 5534 5535 slp = (struct fcipstr *)wq->q_ptr; 5536 5537 if (MBLKL(mp) < DL_UNBIND_REQ_SIZE) { 5538 dlerrorack(wq, mp, DL_UNBIND_REQ, DL_BADPRIM, 0); 5539 return; 5540 } 5541 5542 if (slp->sl_state != DL_IDLE) { 5543 dlerrorack(wq, mp, DL_UNBIND_REQ, DL_OUTSTATE, 0); 5544 return; 5545 } 5546 5547 slp->sl_state = DL_UNBOUND; 5548 slp->sl_sap = 0; 5549 5550 (void) putnextctl1(RD(wq), M_FLUSH, FLUSHRW); 5551 dlokack(wq, mp, DL_UNBIND_REQ); 5552 5553 fcip_setipq(slp->sl_fcip); 5554 } 5555 5556 /* 5557 * Return our physical address 5558 */ 5559 static void 5560 fcip_pareq(queue_t *wq, mblk_t *mp) 5561 { 5562 struct fcipstr *slp; 5563 union DL_primitives *dlp; 5564 int type; 5565 struct fcip *fptr; 5566 fcip_port_info_t *fport; 5567 struct ether_addr addr; 5568 5569 slp = (struct fcipstr *)wq->q_ptr; 5570 5571 if (MBLKL(mp) < DL_PHYS_ADDR_REQ_SIZE) { 5572 dlerrorack(wq, mp, DL_PHYS_ADDR_REQ, DL_BADPRIM, 0); 5573 return; 5574 } 5575 5576 dlp = (union DL_primitives *)mp->b_rptr; 5577 type = dlp->physaddr_req.dl_addr_type; 5578 fptr = slp->sl_fcip; 5579 5580 if (fptr == NULL) { 5581 dlerrorack(wq, mp, DL_PHYS_ADDR_REQ, DL_OUTSTATE, 0); 5582 return; 5583 } 5584 5585 fport = fptr->fcip_port_info; 5586 5587 switch (type) { 5588 case DL_FACT_PHYS_ADDR: 5589 FCIP_DEBUG(FCIP_DEBUG_DLPI, 5590 (CE_NOTE, "returning factory phys addr")); 5591 wwn_to_ether(&fport->fcipp_pwwn, &addr); 5592 break; 5593 5594 case DL_CURR_PHYS_ADDR: 5595 FCIP_DEBUG(FCIP_DEBUG_DLPI, 5596 (CE_NOTE, "returning current phys addr")); 5597 ether_bcopy(&fptr->fcip_macaddr, &addr); 5598 break; 5599 5600 default: 5601 FCIP_DEBUG(FCIP_DEBUG_DLPI, 5602 (CE_NOTE, "Not known cmd type in phys addr")); 5603 dlerrorack(wq, mp, DL_PHYS_ADDR_REQ, DL_NOTSUPPORTED, 0); 5604 return; 5605 } 5606 dlphysaddrack(wq, mp, &addr, ETHERADDRL); 5607 } 5608 5609 /* 5610 * Set physical address DLPI request 5611 */ 5612 static void 5613 fcip_spareq(queue_t *wq, mblk_t *mp) 5614 { 5615 struct fcipstr *slp; 5616 union DL_primitives *dlp; 5617 t_uscalar_t off, len; 5618 struct ether_addr *addrp; 5619 la_wwn_t wwn; 5620 struct fcip *fptr; 5621 fc_ns_cmd_t fcip_ns_cmd; 5622 5623 slp = (struct fcipstr *)wq->q_ptr; 5624 5625 if (MBLKL(mp) < DL_SET_PHYS_ADDR_REQ_SIZE) { 5626 dlerrorack(wq, mp, DL_SET_PHYS_ADDR_REQ, DL_BADPRIM, 0); 5627 return; 5628 } 5629 5630 dlp = (union DL_primitives *)mp->b_rptr; 5631 len = dlp->set_physaddr_req.dl_addr_length; 5632 off = dlp->set_physaddr_req.dl_addr_offset; 5633 5634 if (!MBLKIN(mp, off, len)) { 5635 dlerrorack(wq, mp, DL_SET_PHYS_ADDR_REQ, DL_BADPRIM, 0); 5636 return; 5637 } 5638 5639 addrp = (struct ether_addr *)(mp->b_rptr + off); 5640 5641 /* 5642 * If the length of physical address is not correct or address 5643 * specified is a broadcast address or multicast addr - 5644 * return an error. 5645 */ 5646 if ((len != ETHERADDRL) || 5647 ((addrp->ether_addr_octet[0] & 01) == 1) || 5648 (ether_cmp(addrp, &fcip_arpbroadcast_addr) == 0)) { 5649 dlerrorack(wq, mp, DL_SET_PHYS_ADDR_REQ, DL_BADADDR, 0); 5650 return; 5651 } 5652 5653 /* 5654 * check if a stream is attached to this device. Else return an error 5655 */ 5656 if ((fptr = slp->sl_fcip) == NULL) { 5657 dlerrorack(wq, mp, DL_SET_PHYS_ADDR_REQ, DL_OUTSTATE, 0); 5658 return; 5659 } 5660 5661 /* 5662 * set the new interface local address. We request the transport 5663 * layer to change the Port WWN for this device - return an error 5664 * if we don't succeed. 5665 */ 5666 5667 ether_to_wwn(addrp, &wwn); 5668 if (fcip_set_wwn(&wwn) == FC_SUCCESS) { 5669 FCIP_DEBUG(FCIP_DEBUG_DLPI, 5670 (CE_WARN, "WWN changed in spareq")); 5671 } else { 5672 dlerrorack(wq, mp, DL_SET_PHYS_ADDR_REQ, DL_BADADDR, 0); 5673 } 5674 5675 /* 5676 * register The new Port WWN and Node WWN with the transport 5677 * and Nameserver. Hope the transport ensures all current I/O 5678 * has stopped before actually attempting to register a new 5679 * port and Node WWN else we are hosed. Maybe a Link reset 5680 * will get everyone's attention. 5681 */ 5682 fcip_ns_cmd.ns_flags = 0; 5683 fcip_ns_cmd.ns_cmd = NS_RPN_ID; 5684 fcip_ns_cmd.ns_req_len = sizeof (la_wwn_t); 5685 fcip_ns_cmd.ns_req_payload = (caddr_t)&wwn.raw_wwn[0]; 5686 fcip_ns_cmd.ns_resp_len = 0; 5687 fcip_ns_cmd.ns_resp_payload = (caddr_t)0; 5688 if (fc_ulp_port_ns(fptr->fcip_port_info->fcipp_handle, 5689 (opaque_t)0, &fcip_ns_cmd) != FC_SUCCESS) { 5690 FCIP_DEBUG(FCIP_DEBUG_DLPI, 5691 (CE_WARN, "setting Port WWN failed")); 5692 dlerrorack(wq, mp, DL_SET_PHYS_ADDR_REQ, DL_BADPRIM, 0); 5693 return; 5694 } 5695 5696 dlokack(wq, mp, DL_SET_PHYS_ADDR_REQ); 5697 } 5698 5699 /* 5700 * change our port's WWN if permitted by hardware 5701 */ 5702 /* ARGSUSED */ 5703 static int 5704 fcip_set_wwn(la_wwn_t *pwwn) 5705 { 5706 /* 5707 * We're usually not allowed to change the WWN of adapters 5708 * but some adapters do permit us to change the WWN - don't 5709 * permit setting of WWNs (yet?) - This behavior could be 5710 * modified if needed 5711 */ 5712 return (FC_FAILURE); 5713 } 5714 5715 5716 /* 5717 * This routine fills in the header for fastpath data requests. What this 5718 * does in simple terms is, instead of sending all data through the Unitdata 5719 * request dlpi code paths (which will then append the protocol specific 5720 * header - network and snap headers in our case), the upper layers issue 5721 * a M_IOCTL with a DL_IOC_HDR_INFO request and ask the streams endpoint 5722 * driver to give the header it needs appended and the upper layer 5723 * allocates and fills in the header and calls our put routine 5724 */ 5725 static void 5726 fcip_dl_ioc_hdr_info(queue_t *wq, mblk_t *mp) 5727 { 5728 mblk_t *nmp; 5729 struct fcipstr *slp; 5730 struct fcipdladdr *dlap; 5731 dl_unitdata_req_t *dlup; 5732 fcph_network_hdr_t *headerp; 5733 la_wwn_t wwn; 5734 llc_snap_hdr_t *lsnap; 5735 struct fcip *fptr; 5736 fcip_port_info_t *fport; 5737 t_uscalar_t off, len; 5738 size_t hdrlen; 5739 int error; 5740 5741 slp = (struct fcipstr *)wq->q_ptr; 5742 fptr = slp->sl_fcip; 5743 if (fptr == NULL) { 5744 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, 5745 (CE_NOTE, "dliochdr : returns EINVAL1")); 5746 miocnak(wq, mp, 0, EINVAL); 5747 return; 5748 } 5749 5750 error = miocpullup(mp, sizeof (dl_unitdata_req_t) + FCIPADDRL); 5751 if (error != 0) { 5752 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, 5753 (CE_NOTE, "dliochdr : returns %d", error)); 5754 miocnak(wq, mp, 0, error); 5755 return; 5756 } 5757 5758 fport = fptr->fcip_port_info; 5759 5760 /* 5761 * check if the DL_UNITDATA_REQ destination addr has valid offset 5762 * and length values 5763 */ 5764 dlup = (dl_unitdata_req_t *)mp->b_cont->b_rptr; 5765 off = dlup->dl_dest_addr_offset; 5766 len = dlup->dl_dest_addr_length; 5767 if (dlup->dl_primitive != DL_UNITDATA_REQ || 5768 !MBLKIN(mp->b_cont, off, len) || (len != FCIPADDRL)) { 5769 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, 5770 (CE_NOTE, "dliochdr : returns EINVAL2")); 5771 miocnak(wq, mp, 0, EINVAL); 5772 return; 5773 } 5774 5775 dlap = (struct fcipdladdr *)(mp->b_cont->b_rptr + off); 5776 5777 /* 5778 * Allocate a new mblk to hold the ether header 5779 */ 5780 5781 /* 5782 * setup space for network header 5783 */ 5784 hdrlen = (sizeof (llc_snap_hdr_t) + sizeof (fcph_network_hdr_t)); 5785 if ((nmp = allocb(hdrlen, BPRI_MED)) == NULL) { 5786 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, 5787 (CE_NOTE, "dliochdr : returns ENOMEM")); 5788 miocnak(wq, mp, 0, ENOMEM); 5789 return; 5790 } 5791 nmp->b_wptr += hdrlen; 5792 5793 /* 5794 * Fill in the Network Hdr and LLC SNAP header; 5795 */ 5796 headerp = (fcph_network_hdr_t *)nmp->b_rptr; 5797 /* 5798 * just fill in the Node WWN here - we can fill in the NAA_ID when 5799 * we search the routing table 5800 */ 5801 if (ether_cmp(&dlap->dl_phys, &fcip_arpbroadcast_addr) == 0) { 5802 ether_to_wwn(&fcipnhbroadcastaddr, &wwn); 5803 } else { 5804 ether_to_wwn(&dlap->dl_phys, &wwn); 5805 } 5806 bcopy(&wwn, &headerp->net_dest_addr, sizeof (la_wwn_t)); 5807 bcopy(&fport->fcipp_pwwn, &headerp->net_src_addr, sizeof (la_wwn_t)); 5808 lsnap = (llc_snap_hdr_t *)(nmp->b_rptr + sizeof (fcph_network_hdr_t)); 5809 lsnap->dsap = 0xAA; 5810 lsnap->ssap = 0xAA; 5811 lsnap->ctrl = 0x03; 5812 lsnap->oui[0] = 0x00; 5813 lsnap->oui[1] = 0x00; 5814 lsnap->oui[2] = 0x00; 5815 lsnap->pid = BE_16(dlap->dl_sap); 5816 5817 /* 5818 * Link new mblk in after the "request" mblks. 5819 */ 5820 linkb(mp, nmp); 5821 5822 slp->sl_flags |= FCIP_SLFAST; 5823 5824 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, 5825 (CE_NOTE, "dliochdr : returns success ")); 5826 miocack(wq, mp, msgsize(mp->b_cont), 0); 5827 } 5828 5829 5830 /* 5831 * Establish a kmem cache for fcip packets 5832 */ 5833 static int 5834 fcip_cache_constructor(void *buf, void *arg, int flags) 5835 { 5836 fcip_pkt_t *fcip_pkt = buf; 5837 fc_packet_t *fc_pkt; 5838 fcip_port_info_t *fport = (fcip_port_info_t *)arg; 5839 int (*cb) (caddr_t); 5840 struct fcip *fptr; 5841 5842 cb = (flags == KM_SLEEP) ? DDI_DMA_SLEEP : DDI_DMA_DONTWAIT; 5843 5844 ASSERT(fport != NULL); 5845 5846 fptr = fport->fcipp_fcip; 5847 5848 /* 5849 * we allocated space for our private area at the end of the 5850 * fc packet. Make sure we point to it correctly. Ideally we 5851 * should just push fc_packet_private to the beginning or end 5852 * of the fc_packet structure 5853 */ 5854 fcip_pkt->fcip_pkt_next = NULL; 5855 fcip_pkt->fcip_pkt_prev = NULL; 5856 fcip_pkt->fcip_pkt_dest = NULL; 5857 fcip_pkt->fcip_pkt_state = 0; 5858 fcip_pkt->fcip_pkt_reason = 0; 5859 fcip_pkt->fcip_pkt_flags = 0; 5860 fcip_pkt->fcip_pkt_fptr = fptr; 5861 fcip_pkt->fcip_pkt_dma_flags = 0; 5862 5863 fc_pkt = FCIP_PKT_TO_FC_PKT(fcip_pkt); 5864 fc_pkt->pkt_ulp_rscn_infop = NULL; 5865 5866 /* 5867 * We use pkt_cmd_dma for OUTBOUND requests. We don't expect 5868 * any responses for outbound IP data so no need to setup 5869 * response or data dma handles. 5870 */ 5871 if (ddi_dma_alloc_handle(fport->fcipp_dip, 5872 &fport->fcipp_cmd_dma_attr, cb, NULL, 5873 &fc_pkt->pkt_cmd_dma) != DDI_SUCCESS) { 5874 return (FCIP_FAILURE); 5875 } 5876 5877 fc_pkt->pkt_cmd_acc = fc_pkt->pkt_resp_acc = NULL; 5878 fc_pkt->pkt_fca_private = (opaque_t)((caddr_t)buf + 5879 sizeof (fcip_pkt_t)); 5880 fc_pkt->pkt_ulp_private = (opaque_t)fcip_pkt; 5881 5882 fc_pkt->pkt_cmd_cookie_cnt = fc_pkt->pkt_resp_cookie_cnt = 5883 fc_pkt->pkt_data_cookie_cnt = 0; 5884 fc_pkt->pkt_cmd_cookie = fc_pkt->pkt_resp_cookie = 5885 fc_pkt->pkt_data_cookie = NULL; 5886 5887 return (FCIP_SUCCESS); 5888 } 5889 5890 /* 5891 * destroy the fcip kmem cache 5892 */ 5893 static void 5894 fcip_cache_destructor(void *buf, void *arg) 5895 { 5896 fcip_pkt_t *fcip_pkt = (fcip_pkt_t *)buf; 5897 fc_packet_t *fc_pkt; 5898 fcip_port_info_t *fport = (fcip_port_info_t *)arg; 5899 struct fcip *fptr; 5900 5901 ASSERT(fport != NULL); 5902 5903 fptr = fport->fcipp_fcip; 5904 5905 ASSERT(fptr == fcip_pkt->fcip_pkt_fptr); 5906 fc_pkt = FCIP_PKT_TO_FC_PKT(fcip_pkt); 5907 5908 if (fc_pkt->pkt_cmd_dma) { 5909 ddi_dma_free_handle(&fc_pkt->pkt_cmd_dma); 5910 } 5911 } 5912 5913 /* 5914 * the fcip destination structure is hashed on Node WWN assuming 5915 * a NAA_ID of 0x1 (IEEE) 5916 */ 5917 static struct fcip_dest * 5918 fcip_get_dest(struct fcip *fptr, la_wwn_t *pwwn) 5919 { 5920 struct fcip_dest *fdestp = NULL; 5921 fcip_port_info_t *fport; 5922 int hash_bucket; 5923 opaque_t pd; 5924 int rval; 5925 struct fcip_routing_table *frp; 5926 la_wwn_t twwn; 5927 uint32_t *twwnp = (uint32_t *)&twwn; 5928 5929 hash_bucket = FCIP_DEST_HASH(pwwn->raw_wwn); 5930 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, 5931 (CE_NOTE, "get dest hashbucket : 0x%x", hash_bucket)); 5932 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, 5933 (CE_NOTE, "0x%x 0x%x 0x%x 0x%x 0x%x 0x%x", 5934 pwwn->raw_wwn[2], pwwn->raw_wwn[3], pwwn->raw_wwn[4], 5935 pwwn->raw_wwn[5], pwwn->raw_wwn[6], pwwn->raw_wwn[7])); 5936 5937 ASSERT(hash_bucket < FCIP_DEST_HASH_ELEMS); 5938 5939 if (fcip_check_port_exists(fptr)) { 5940 /* fptr is stale, return fdestp */ 5941 return (fdestp); 5942 } 5943 fport = fptr->fcip_port_info; 5944 5945 /* 5946 * First check if we have active I/Os going on with the 5947 * destination port (an entry would exist in fcip_dest hash table) 5948 */ 5949 mutex_enter(&fptr->fcip_dest_mutex); 5950 fdestp = fptr->fcip_dest[hash_bucket]; 5951 while (fdestp != NULL) { 5952 mutex_enter(&fdestp->fcipd_mutex); 5953 if (fdestp->fcipd_rtable) { 5954 if (fcip_wwn_compare(pwwn, &fdestp->fcipd_pwwn, 5955 FCIP_COMPARE_NWWN) == 0) { 5956 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, 5957 (CE_NOTE, "found fdestp")); 5958 mutex_exit(&fdestp->fcipd_mutex); 5959 mutex_exit(&fptr->fcip_dest_mutex); 5960 return (fdestp); 5961 } 5962 } 5963 mutex_exit(&fdestp->fcipd_mutex); 5964 fdestp = fdestp->fcipd_next; 5965 } 5966 mutex_exit(&fptr->fcip_dest_mutex); 5967 5968 /* 5969 * We did not find the destination port information in our 5970 * active port list so search for an entry in our routing 5971 * table. 5972 */ 5973 mutex_enter(&fptr->fcip_rt_mutex); 5974 frp = fcip_lookup_rtable(fptr, pwwn, FCIP_COMPARE_NWWN); 5975 mutex_exit(&fptr->fcip_rt_mutex); 5976 5977 if (frp == NULL || (frp && (!FCIP_RTE_UNAVAIL(frp->fcipr_state)) && 5978 frp->fcipr_state != PORT_DEVICE_LOGGED_IN) || 5979 (frp && frp->fcipr_pd == NULL)) { 5980 /* 5981 * No entry for the destination port in our routing 5982 * table too. First query the transport to see if it 5983 * already has structures for the destination port in 5984 * its hash tables. This must be done for all topologies 5985 * since we could have retired entries in the hash tables 5986 * which may have to be re-added without a statechange 5987 * callback happening. Its better to try and get an entry 5988 * for the destination port rather than simply failing a 5989 * request though it may be an overkill in private loop 5990 * topologies. 5991 * If a entry for the remote port exists in the transport's 5992 * hash tables, we are fine and can add the entry to our 5993 * routing and dest hash lists, Else for fabric configs we 5994 * query the nameserver if one exists or issue FARP ELS. 5995 */ 5996 5997 /* 5998 * We need to do a PortName based Nameserver 5999 * query operation. So get the right PortWWN 6000 * for the adapter. 6001 */ 6002 bcopy(pwwn, &twwn, sizeof (la_wwn_t)); 6003 6004 /* 6005 * Try IEEE Name (Format 1) first, this is the default and 6006 * Emulex uses this format. 6007 */ 6008 pd = fc_ulp_get_remote_port(fport->fcipp_handle, 6009 &twwn, &rval, 1); 6010 6011 if (rval != FC_SUCCESS) { 6012 /* 6013 * If IEEE Name (Format 1) query failed, try IEEE 6014 * Extended Name (Format 2) which Qlogic uses. 6015 * And try port 1 on Qlogic FC-HBA first. 6016 * Note: On x86, we need to byte swap the 32-bit 6017 * word first, after the modification, swap it back. 6018 */ 6019 *twwnp = BE_32(*twwnp); 6020 twwn.w.nport_id = QLC_PORT_1_ID_BITS; 6021 twwn.w.naa_id = QLC_PORT_NAA; 6022 *twwnp = BE_32(*twwnp); 6023 pd = fc_ulp_get_remote_port(fport->fcipp_handle, 6024 &twwn, &rval, 1); 6025 } 6026 6027 if (rval != FC_SUCCESS) { 6028 /* If still failed, try port 2 on Qlogic FC-HBA. */ 6029 *twwnp = BE_32(*twwnp); 6030 twwn.w.nport_id = QLC_PORT_2_ID_BITS; 6031 *twwnp = BE_32(*twwnp); 6032 pd = fc_ulp_get_remote_port(fport->fcipp_handle, 6033 &twwn, &rval, 1); 6034 } 6035 6036 if (rval == FC_SUCCESS) { 6037 fc_portmap_t map; 6038 /* 6039 * Add the newly found destination structure 6040 * to our routing table. Create a map with 6041 * the device we found. We could ask the 6042 * transport to give us the list of all 6043 * devices connected to our port but we 6044 * probably don't need to know all the devices 6045 * so let us just constuct a list with only 6046 * one device instead. 6047 */ 6048 6049 fc_ulp_copy_portmap(&map, pd); 6050 fcip_rt_update(fptr, &map, 1); 6051 6052 mutex_enter(&fptr->fcip_rt_mutex); 6053 frp = fcip_lookup_rtable(fptr, pwwn, 6054 FCIP_COMPARE_NWWN); 6055 mutex_exit(&fptr->fcip_rt_mutex); 6056 6057 fdestp = fcip_add_dest(fptr, frp); 6058 } else if (fcip_farp_supported && 6059 (FC_TOP_EXTERNAL(fport->fcipp_topology) || 6060 (fport->fcipp_topology == FC_TOP_PT_PT))) { 6061 /* 6062 * The Name server request failed so 6063 * issue an FARP 6064 */ 6065 fdestp = fcip_do_farp(fptr, pwwn, NULL, 6066 0, 0); 6067 } else { 6068 fdestp = NULL; 6069 } 6070 } else if (frp && frp->fcipr_state == PORT_DEVICE_LOGGED_IN) { 6071 /* 6072 * Prepare a dest structure to return to caller 6073 */ 6074 fdestp = fcip_add_dest(fptr, frp); 6075 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, 6076 (CE_NOTE, "in fcip get dest non fabric")); 6077 } 6078 return (fdestp); 6079 } 6080 6081 6082 /* 6083 * Endian clean WWN compare. 6084 * Returns 0 if they compare OK, else return non zero value. 6085 * flag can be bitwise OR of FCIP_COMPARE_NWWN, FCIP_COMPARE_PWWN, 6086 * FCIP_COMPARE_BROADCAST. 6087 */ 6088 static int 6089 fcip_wwn_compare(la_wwn_t *wwn1, la_wwn_t *wwn2, int flag) 6090 { 6091 int rval = 0; 6092 if ((wwn1->raw_wwn[2] != wwn2->raw_wwn[2]) || 6093 (wwn1->raw_wwn[3] != wwn2->raw_wwn[3]) || 6094 (wwn1->raw_wwn[4] != wwn2->raw_wwn[4]) || 6095 (wwn1->raw_wwn[5] != wwn2->raw_wwn[5]) || 6096 (wwn1->raw_wwn[6] != wwn2->raw_wwn[6]) || 6097 (wwn1->raw_wwn[7] != wwn2->raw_wwn[7])) { 6098 rval = 1; 6099 } else if ((flag == FCIP_COMPARE_PWWN) && 6100 (((wwn1->raw_wwn[0] & 0xf0) != (wwn2->raw_wwn[0] & 0xf0)) || 6101 (wwn1->raw_wwn[1] != wwn2->raw_wwn[1]))) { 6102 rval = 1; 6103 } 6104 return (rval); 6105 } 6106 6107 6108 /* 6109 * Add an entry for a remote port in the dest hash table. Dest hash table 6110 * has entries for ports in the routing hash table with which we decide 6111 * to establish IP communication with. The no. of entries in the dest hash 6112 * table must always be less than or equal to the entries in the routing 6113 * hash table. Every entry in the dest hash table ofcourse must have a 6114 * corresponding entry in the routing hash table 6115 */ 6116 static struct fcip_dest * 6117 fcip_add_dest(struct fcip *fptr, struct fcip_routing_table *frp) 6118 { 6119 struct fcip_dest *fdestp = NULL; 6120 la_wwn_t *pwwn; 6121 int hash_bucket; 6122 struct fcip_dest *fdest_new; 6123 6124 if (frp == NULL) { 6125 return (fdestp); 6126 } 6127 6128 pwwn = &frp->fcipr_pwwn; 6129 mutex_enter(&fptr->fcip_dest_mutex); 6130 hash_bucket = FCIP_DEST_HASH(pwwn->raw_wwn); 6131 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, 6132 (CE_NOTE, "add dest hash_bucket: 0x%x", hash_bucket)); 6133 6134 ASSERT(hash_bucket < FCIP_DEST_HASH_ELEMS); 6135 6136 fdestp = fptr->fcip_dest[hash_bucket]; 6137 while (fdestp != NULL) { 6138 mutex_enter(&fdestp->fcipd_mutex); 6139 if (fdestp->fcipd_rtable) { 6140 if (fcip_wwn_compare(pwwn, &fdestp->fcipd_pwwn, 6141 FCIP_COMPARE_PWWN) == 0) { 6142 mutex_exit(&fdestp->fcipd_mutex); 6143 mutex_exit(&fptr->fcip_dest_mutex); 6144 return (fdestp); 6145 } 6146 } 6147 mutex_exit(&fdestp->fcipd_mutex); 6148 fdestp = fdestp->fcipd_next; 6149 } 6150 6151 ASSERT(fdestp == NULL); 6152 6153 fdest_new = (struct fcip_dest *) 6154 kmem_zalloc(sizeof (struct fcip_dest), KM_SLEEP); 6155 6156 mutex_init(&fdest_new->fcipd_mutex, NULL, MUTEX_DRIVER, NULL); 6157 fdest_new->fcipd_next = fptr->fcip_dest[hash_bucket]; 6158 fdest_new->fcipd_refcnt = 0; 6159 fdest_new->fcipd_rtable = frp; 6160 fdest_new->fcipd_ncmds = 0; 6161 fptr->fcip_dest[hash_bucket] = fdest_new; 6162 fdest_new->fcipd_flags = FCIP_PORT_NOTLOGGED; 6163 6164 mutex_exit(&fptr->fcip_dest_mutex); 6165 return (fdest_new); 6166 } 6167 6168 /* 6169 * Cleanup the dest hash table and remove all entries 6170 */ 6171 static void 6172 fcip_cleanup_dest(struct fcip *fptr) 6173 { 6174 struct fcip_dest *fdestp = NULL; 6175 struct fcip_dest *fdest_delp = NULL; 6176 int i; 6177 6178 mutex_enter(&fptr->fcip_dest_mutex); 6179 6180 for (i = 0; i < FCIP_DEST_HASH_ELEMS; i++) { 6181 fdestp = fptr->fcip_dest[i]; 6182 while (fdestp != NULL) { 6183 mutex_destroy(&fdestp->fcipd_mutex); 6184 fdest_delp = fdestp; 6185 fdestp = fdestp->fcipd_next; 6186 kmem_free(fdest_delp, sizeof (struct fcip_dest)); 6187 fptr->fcip_dest[i] = NULL; 6188 } 6189 } 6190 mutex_exit(&fptr->fcip_dest_mutex); 6191 } 6192 6193 6194 /* 6195 * Send FARP requests for Fabric ports when we don't have the port 6196 * we wish to talk to in our routing hash table. FARP is specially required 6197 * to talk to FC switches for inband switch management. Most FC switches 6198 * today have a switch FC IP address for IP over FC inband switch management 6199 * but the WWN and Port_ID for this traffic is not available through the 6200 * Nameservers since the switch themeselves are transparent. 6201 */ 6202 /* ARGSUSED */ 6203 static struct fcip_dest * 6204 fcip_do_farp(struct fcip *fptr, la_wwn_t *pwwn, char *ip_addr, 6205 size_t ip_addr_len, int flags) 6206 { 6207 fcip_pkt_t *fcip_pkt; 6208 fc_packet_t *fc_pkt; 6209 fcip_port_info_t *fport = fptr->fcip_port_info; 6210 la_els_farp_t farp_cmd; 6211 la_els_farp_t *fcmd; 6212 struct fcip_dest *fdestp = NULL; 6213 int rval; 6214 clock_t farp_lbolt; 6215 la_wwn_t broadcast_wwn; 6216 struct fcip_dest *bdestp; 6217 struct fcip_routing_table *frp; 6218 6219 bdestp = fcip_get_dest(fptr, &broadcast_wwn); 6220 6221 if (bdestp == NULL) { 6222 return (fdestp); 6223 } 6224 6225 fcip_pkt = fcip_ipkt_alloc(fptr, sizeof (la_els_farp_t), 6226 sizeof (la_els_farp_t), bdestp->fcipd_pd, KM_SLEEP); 6227 6228 if (fcip_pkt == NULL) { 6229 return (fdestp); 6230 } 6231 6232 fc_pkt = FCIP_PKT_TO_FC_PKT(fcip_pkt); 6233 ether_to_wwn(&fcip_arpbroadcast_addr, &broadcast_wwn); 6234 6235 mutex_enter(&bdestp->fcipd_mutex); 6236 if (bdestp->fcipd_rtable == NULL) { 6237 mutex_exit(&bdestp->fcipd_mutex); 6238 fcip_ipkt_free(fcip_pkt); 6239 return (fdestp); 6240 } 6241 6242 fcip_pkt->fcip_pkt_dest = bdestp; 6243 fc_pkt->pkt_fca_device = bdestp->fcipd_fca_dev; 6244 6245 bdestp->fcipd_ncmds++; 6246 mutex_exit(&bdestp->fcipd_mutex); 6247 6248 fcip_init_broadcast_pkt(fcip_pkt, NULL, 1); 6249 fcip_pkt->fcip_pkt_flags |= FCIP_PKT_IN_LIST; 6250 6251 /* 6252 * Now initialize the FARP payload itself 6253 */ 6254 fcmd = &farp_cmd; 6255 fcmd->ls_code.ls_code = LA_ELS_FARP_REQ; 6256 fcmd->ls_code.mbz = 0; 6257 /* 6258 * for now just match the Port WWN since the other match addr 6259 * code points are optional. We can explore matching the IP address 6260 * if needed 6261 */ 6262 if (ip_addr) { 6263 fcmd->match_addr = FARP_MATCH_WW_PN_IPv4; 6264 } else { 6265 fcmd->match_addr = FARP_MATCH_WW_PN; 6266 } 6267 6268 /* 6269 * Request the responder port to log into us - that way 6270 * the Transport is aware of the remote port when we create 6271 * an entry for it in our tables 6272 */ 6273 fcmd->resp_flags = FARP_INIT_REPLY | FARP_INIT_P_LOGI; 6274 fcmd->req_id = fport->fcipp_sid; 6275 fcmd->dest_id.port_id = fc_pkt->pkt_cmd_fhdr.d_id; 6276 bcopy(&fport->fcipp_pwwn, &fcmd->req_pwwn, sizeof (la_wwn_t)); 6277 bcopy(&fport->fcipp_nwwn, &fcmd->req_nwwn, sizeof (la_wwn_t)); 6278 bcopy(pwwn, &fcmd->resp_pwwn, sizeof (la_wwn_t)); 6279 /* 6280 * copy in source IP address if we get to know it 6281 */ 6282 if (ip_addr) { 6283 bcopy(ip_addr, fcmd->resp_ip, ip_addr_len); 6284 } 6285 6286 fc_pkt->pkt_cmdlen = sizeof (la_els_farp_t); 6287 fc_pkt->pkt_rsplen = sizeof (la_els_farp_t); 6288 fc_pkt->pkt_tran_type = FC_PKT_EXCHANGE; 6289 fc_pkt->pkt_ulp_private = (opaque_t)fcip_pkt; 6290 6291 /* 6292 * Endian safe copy 6293 */ 6294 FCIP_CP_OUT(fcmd, fc_pkt->pkt_cmd, fc_pkt->pkt_cmd_acc, 6295 sizeof (la_els_farp_t)); 6296 6297 /* 6298 * send the packet in polled mode. 6299 */ 6300 rval = fc_ulp_issue_els(fport->fcipp_handle, fc_pkt); 6301 if (rval != FC_SUCCESS) { 6302 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, (CE_WARN, 6303 "fcip_transport of farp pkt failed 0x%x", rval)); 6304 fcip_pkt->fcip_pkt_flags &= ~FCIP_PKT_IN_LIST; 6305 fcip_ipkt_free(fcip_pkt); 6306 6307 mutex_enter(&bdestp->fcipd_mutex); 6308 bdestp->fcipd_ncmds--; 6309 mutex_exit(&bdestp->fcipd_mutex); 6310 6311 return (fdestp); 6312 } 6313 6314 farp_lbolt = ddi_get_lbolt(); 6315 farp_lbolt += drv_usectohz(FCIP_FARP_TIMEOUT); 6316 6317 mutex_enter(&fptr->fcip_mutex); 6318 fptr->fcip_farp_rsp_flag = 0; 6319 while (!fptr->fcip_farp_rsp_flag) { 6320 if (cv_timedwait(&fptr->fcip_farp_cv, &fptr->fcip_mutex, 6321 farp_lbolt) == -1) { 6322 /* 6323 * No FARP response from any destination port 6324 * so bail out. 6325 */ 6326 fptr->fcip_farp_rsp_flag = 1; 6327 } else { 6328 /* 6329 * We received a FARP response - check to see if the 6330 * response was in reply to our FARP request. 6331 */ 6332 6333 mutex_enter(&fptr->fcip_rt_mutex); 6334 frp = fcip_lookup_rtable(fptr, pwwn, FCIP_COMPARE_NWWN); 6335 mutex_exit(&fptr->fcip_rt_mutex); 6336 6337 if ((frp != NULL) && 6338 !FCIP_RTE_UNAVAIL(frp->fcipr_state)) { 6339 fdestp = fcip_get_dest(fptr, pwwn); 6340 } else { 6341 /* 6342 * Not our FARP response so go back and wait 6343 * again till FARP_TIMEOUT expires 6344 */ 6345 fptr->fcip_farp_rsp_flag = 0; 6346 } 6347 } 6348 } 6349 mutex_exit(&fptr->fcip_mutex); 6350 6351 fcip_pkt->fcip_pkt_flags |= FCIP_PKT_IN_LIST; 6352 fcip_ipkt_free(fcip_pkt); 6353 mutex_enter(&bdestp->fcipd_mutex); 6354 bdestp->fcipd_ncmds--; 6355 mutex_exit(&bdestp->fcipd_mutex); 6356 return (fdestp); 6357 } 6358 6359 6360 6361 /* 6362 * Helper routine to PLOGI to a remote port we wish to talk to. 6363 * This may not be required since the port driver does logins anyway, 6364 * but this can be required in fabric cases since FARP requests/responses 6365 * don't require you to be logged in? 6366 */ 6367 6368 /* ARGSUSED */ 6369 static int 6370 fcip_do_plogi(struct fcip *fptr, struct fcip_routing_table *frp) 6371 { 6372 fcip_pkt_t *fcip_pkt; 6373 fc_packet_t *fc_pkt; 6374 fcip_port_info_t *fport = fptr->fcip_port_info; 6375 la_els_logi_t logi; 6376 int rval; 6377 fc_frame_hdr_t *fr_hdr; 6378 6379 /* 6380 * Don't bother to login for broadcast RTE entries 6381 */ 6382 if ((frp->fcipr_d_id.port_id == 0x0) || 6383 (frp->fcipr_d_id.port_id == 0xffffff)) { 6384 return (FC_FAILURE); 6385 } 6386 6387 /* 6388 * We shouldn't pound in too many logins here 6389 * 6390 */ 6391 if (frp->fcipr_state == FCIP_RT_LOGIN_PROGRESS || 6392 frp->fcipr_state == PORT_DEVICE_LOGGED_IN) { 6393 return (FC_SUCCESS); 6394 } 6395 6396 fcip_pkt = fcip_ipkt_alloc(fptr, sizeof (la_els_logi_t), 6397 sizeof (la_els_logi_t), frp->fcipr_pd, KM_SLEEP); 6398 6399 if (fcip_pkt == NULL) { 6400 return (FC_FAILURE); 6401 } 6402 6403 /* 6404 * Update back pointer for login state update 6405 */ 6406 fcip_pkt->fcip_pkt_frp = frp; 6407 frp->fcipr_state = FCIP_RT_LOGIN_PROGRESS; 6408 6409 fc_pkt = FCIP_PKT_TO_FC_PKT(fcip_pkt); 6410 6411 /* 6412 * Initialize frame header for ELS 6413 */ 6414 fr_hdr = &fc_pkt->pkt_cmd_fhdr; 6415 fr_hdr->r_ctl = R_CTL_ELS_REQ; 6416 fr_hdr->type = FC_TYPE_EXTENDED_LS; 6417 fr_hdr->f_ctl = F_CTL_SEQ_INITIATIVE | F_CTL_FIRST_SEQ; 6418 fr_hdr->df_ctl = 0; 6419 fr_hdr->s_id = fport->fcipp_sid.port_id; 6420 fr_hdr->d_id = frp->fcipr_d_id.port_id; 6421 fr_hdr->seq_cnt = 0; 6422 fr_hdr->ox_id = 0xffff; 6423 fr_hdr->rx_id = 0xffff; 6424 fr_hdr->ro = 0; 6425 6426 fc_pkt->pkt_rsplen = sizeof (la_els_logi_t); 6427 fc_pkt->pkt_comp = fcip_ipkt_callback; 6428 fc_pkt->pkt_tran_type = FC_PKT_EXCHANGE; 6429 fc_pkt->pkt_timeout = 10; /* 10 seconds */ 6430 fcip_pkt->fcip_pkt_ttl = fptr->fcip_timeout_ticks + fc_pkt->pkt_timeout; 6431 fc_pkt->pkt_ulp_private = (opaque_t)fcip_pkt; 6432 6433 /* 6434 * Everybody does class 3, so let's just set it. If the transport 6435 * knows better, it will deal with the class appropriately. 6436 */ 6437 6438 fc_pkt->pkt_tran_flags = FC_TRAN_INTR | FC_TRAN_CLASS3; 6439 6440 /* 6441 * we need only fill in the ls_code and the cmd frame header 6442 */ 6443 bzero((void *)&logi, sizeof (la_els_logi_t)); 6444 logi.ls_code.ls_code = LA_ELS_PLOGI; 6445 logi.ls_code.mbz = 0; 6446 6447 FCIP_CP_OUT((uint8_t *)&logi, fc_pkt->pkt_cmd, fc_pkt->pkt_cmd_acc, 6448 sizeof (la_els_logi_t)); 6449 6450 rval = fc_ulp_login(fport->fcipp_handle, &fc_pkt, 1); 6451 if (rval != FC_SUCCESS) { 6452 cmn_err(CE_WARN, 6453 "!fc_ulp_login failed for d_id: 0x%x, rval: 0x%x", 6454 frp->fcipr_d_id.port_id, rval); 6455 fcip_ipkt_free(fcip_pkt); 6456 } 6457 return (rval); 6458 } 6459 6460 /* 6461 * The packet callback routine - called from the transport/FCA after 6462 * it is done DMA'ing/sending out the packet contents on the wire so 6463 * that the alloc'ed packet can be freed 6464 */ 6465 static void 6466 fcip_ipkt_callback(fc_packet_t *fc_pkt) 6467 { 6468 ls_code_t logi_req; 6469 ls_code_t logi_resp; 6470 fcip_pkt_t *fcip_pkt; 6471 fc_frame_hdr_t *fr_hdr; 6472 struct fcip *fptr; 6473 fcip_port_info_t *fport; 6474 struct fcip_routing_table *frp; 6475 6476 fr_hdr = &fc_pkt->pkt_cmd_fhdr; 6477 6478 FCIP_CP_IN(fc_pkt->pkt_resp, (uint8_t *)&logi_resp, 6479 fc_pkt->pkt_resp_acc, sizeof (logi_resp)); 6480 6481 FCIP_CP_IN(fc_pkt->pkt_cmd, (uint8_t *)&logi_req, fc_pkt->pkt_cmd_acc, 6482 sizeof (logi_req)); 6483 6484 fcip_pkt = (fcip_pkt_t *)fc_pkt->pkt_ulp_private; 6485 frp = fcip_pkt->fcip_pkt_frp; 6486 fptr = fcip_pkt->fcip_pkt_fptr; 6487 fport = fptr->fcip_port_info; 6488 6489 ASSERT(logi_req.ls_code == LA_ELS_PLOGI); 6490 6491 if (fc_pkt->pkt_state != FC_PKT_SUCCESS || 6492 logi_resp.ls_code != LA_ELS_ACC) { 6493 /* EMPTY */ 6494 6495 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, (CE_WARN, 6496 "opcode : 0x%x to d_id: 0x%x failed", 6497 logi_req.ls_code, fr_hdr->d_id)); 6498 6499 mutex_enter(&fptr->fcip_rt_mutex); 6500 frp->fcipr_state = PORT_DEVICE_INVALID; 6501 frp->fcipr_invalid_timeout = fptr->fcip_timeout_ticks + 6502 (FCIP_RTE_TIMEOUT / 2); 6503 mutex_exit(&fptr->fcip_rt_mutex); 6504 } else { 6505 fc_portid_t d_id; 6506 6507 d_id.port_id = fr_hdr->d_id; 6508 d_id.priv_lilp_posit = 0; 6509 6510 /* 6511 * Update PLOGI results; FCA Handle, and Port device handles 6512 */ 6513 mutex_enter(&fptr->fcip_rt_mutex); 6514 frp->fcipr_pd = fc_pkt->pkt_pd; 6515 frp->fcipr_fca_dev = 6516 fc_ulp_get_fca_device(fport->fcipp_handle, d_id); 6517 frp->fcipr_state = PORT_DEVICE_LOGGED_IN; 6518 mutex_exit(&fptr->fcip_rt_mutex); 6519 } 6520 6521 fcip_ipkt_free(fcip_pkt); 6522 } 6523 6524 6525 /* 6526 * pkt_alloc routine for outbound IP datagrams. The cache constructor 6527 * Only initializes the pkt_cmd_dma (which is where the outbound datagram 6528 * is stuffed) since we don't expect response 6529 */ 6530 static fcip_pkt_t * 6531 fcip_pkt_alloc(struct fcip *fptr, mblk_t *bp, int flags, int datalen) 6532 { 6533 fcip_pkt_t *fcip_pkt; 6534 fc_packet_t *fc_pkt; 6535 ddi_dma_cookie_t pkt_cookie; 6536 ddi_dma_cookie_t *cp; 6537 uint32_t cnt; 6538 fcip_port_info_t *fport = fptr->fcip_port_info; 6539 6540 fcip_pkt = kmem_cache_alloc(fptr->fcip_xmit_cache, flags); 6541 if (fcip_pkt == NULL) { 6542 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, (CE_WARN, 6543 "fcip_pkt_alloc: kmem_cache_alloc failed")); 6544 return (NULL); 6545 } 6546 6547 fc_pkt = FCIP_PKT_TO_FC_PKT(fcip_pkt); 6548 fcip_pkt->fcip_pkt_fcpktp = fc_pkt; 6549 fc_pkt->pkt_tran_flags = 0; 6550 fcip_pkt->fcip_pkt_dma_flags = 0; 6551 6552 /* 6553 * the cache constructor has allocated the dma handle 6554 */ 6555 fc_pkt->pkt_cmd = (caddr_t)bp->b_rptr; 6556 if (ddi_dma_addr_bind_handle(fc_pkt->pkt_cmd_dma, NULL, 6557 (caddr_t)bp->b_rptr, datalen, DDI_DMA_WRITE | DDI_DMA_CONSISTENT, 6558 DDI_DMA_DONTWAIT, NULL, &pkt_cookie, 6559 &fc_pkt->pkt_cmd_cookie_cnt) != DDI_DMA_MAPPED) { 6560 goto fail; 6561 } 6562 6563 fcip_pkt->fcip_pkt_dma_flags |= FCIP_CMD_DMA_BOUND; 6564 6565 if (fc_pkt->pkt_cmd_cookie_cnt > 6566 fport->fcipp_cmd_dma_attr.dma_attr_sgllen) { 6567 goto fail; 6568 } 6569 6570 ASSERT(fc_pkt->pkt_cmd_cookie_cnt != 0); 6571 6572 cp = fc_pkt->pkt_cmd_cookie = (ddi_dma_cookie_t *)kmem_alloc( 6573 fc_pkt->pkt_cmd_cookie_cnt * sizeof (pkt_cookie), 6574 KM_NOSLEEP); 6575 6576 if (cp == NULL) { 6577 goto fail; 6578 } 6579 6580 *cp = pkt_cookie; 6581 cp++; 6582 for (cnt = 1; cnt < fc_pkt->pkt_cmd_cookie_cnt; cnt++, cp++) { 6583 ddi_dma_nextcookie(fc_pkt->pkt_cmd_dma, &pkt_cookie); 6584 *cp = pkt_cookie; 6585 } 6586 6587 fc_pkt->pkt_cmdlen = datalen; 6588 6589 fcip_pkt->fcip_pkt_mp = NULL; 6590 fcip_pkt->fcip_pkt_wq = NULL; 6591 fcip_pkt->fcip_pkt_dest = NULL; 6592 fcip_pkt->fcip_pkt_next = NULL; 6593 fcip_pkt->fcip_pkt_prev = NULL; 6594 fcip_pkt->fcip_pkt_state = 0; 6595 fcip_pkt->fcip_pkt_reason = 0; 6596 fcip_pkt->fcip_pkt_flags = 0; 6597 fcip_pkt->fcip_pkt_frp = NULL; 6598 6599 return (fcip_pkt); 6600 fail: 6601 if (fcip_pkt) { 6602 fcip_pkt_free(fcip_pkt, 0); 6603 } 6604 return ((fcip_pkt_t *)0); 6605 } 6606 6607 /* 6608 * Free a packet and all its associated resources 6609 */ 6610 static void 6611 fcip_pkt_free(struct fcip_pkt *fcip_pkt, int free_mblk) 6612 { 6613 fc_packet_t *fc_pkt = FCIP_PKT_TO_FC_PKT(fcip_pkt); 6614 struct fcip *fptr = fcip_pkt->fcip_pkt_fptr; 6615 6616 if (fc_pkt->pkt_cmd_cookie != NULL) { 6617 kmem_free(fc_pkt->pkt_cmd_cookie, fc_pkt->pkt_cmd_cookie_cnt * 6618 sizeof (ddi_dma_cookie_t)); 6619 fc_pkt->pkt_cmd_cookie = NULL; 6620 } 6621 6622 fcip_free_pkt_dma(fcip_pkt); 6623 if (free_mblk && fcip_pkt->fcip_pkt_mp) { 6624 freemsg(fcip_pkt->fcip_pkt_mp); 6625 fcip_pkt->fcip_pkt_mp = NULL; 6626 } 6627 6628 (void) fc_ulp_uninit_packet(fptr->fcip_port_info->fcipp_handle, fc_pkt); 6629 6630 kmem_cache_free(fptr->fcip_xmit_cache, (void *)fcip_pkt); 6631 } 6632 6633 /* 6634 * Allocate a Packet for internal driver use. This is for requests 6635 * that originate from within the driver 6636 */ 6637 static fcip_pkt_t * 6638 fcip_ipkt_alloc(struct fcip *fptr, int cmdlen, int resplen, 6639 opaque_t pd, int flags) 6640 { 6641 fcip_pkt_t *fcip_pkt; 6642 fc_packet_t *fc_pkt; 6643 int (*cb)(caddr_t); 6644 fcip_port_info_t *fport = fptr->fcip_port_info; 6645 size_t real_len; 6646 uint_t held_here = 0; 6647 ddi_dma_cookie_t pkt_cookie; 6648 ddi_dma_cookie_t *cp; 6649 uint32_t cnt; 6650 6651 cb = (flags == KM_SLEEP) ? DDI_DMA_SLEEP : DDI_DMA_DONTWAIT; 6652 6653 fcip_pkt = kmem_zalloc((sizeof (fcip_pkt_t) + 6654 fport->fcipp_fca_pkt_size), flags); 6655 6656 if (fcip_pkt == NULL) { 6657 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, 6658 (CE_WARN, "pkt alloc of ineternal pkt failed")); 6659 goto fail; 6660 } 6661 6662 fcip_pkt->fcip_pkt_flags = FCIP_PKT_INTERNAL; 6663 fcip_pkt->fcip_pkt_fptr = fptr; 6664 fc_pkt = FCIP_PKT_TO_FC_PKT(fcip_pkt); 6665 fcip_pkt->fcip_pkt_fcpktp = fc_pkt; 6666 fc_pkt->pkt_tran_flags = 0; 6667 fc_pkt->pkt_cmdlen = 0; 6668 fc_pkt->pkt_rsplen = 0; 6669 fc_pkt->pkt_datalen = 0; 6670 fc_pkt->pkt_fca_private = (opaque_t)((caddr_t)fcip_pkt + 6671 sizeof (fcip_pkt_t)); 6672 fc_pkt->pkt_ulp_private = (opaque_t)fcip_pkt; 6673 6674 if (cmdlen) { 6675 if (ddi_dma_alloc_handle(fptr->fcip_dip, 6676 &fport->fcipp_cmd_dma_attr, cb, NULL, 6677 &fc_pkt->pkt_cmd_dma) != DDI_SUCCESS) { 6678 goto fail; 6679 } 6680 6681 if (ddi_dma_mem_alloc(fc_pkt->pkt_cmd_dma, cmdlen, 6682 &fport->fcipp_fca_acc_attr, DDI_DMA_CONSISTENT, 6683 cb, NULL, (caddr_t *)&fc_pkt->pkt_cmd, 6684 &real_len, &fc_pkt->pkt_cmd_acc) != DDI_SUCCESS) { 6685 goto fail; 6686 } 6687 6688 fcip_pkt->fcip_pkt_dma_flags |= FCIP_CMD_DMA_MEM; 6689 fc_pkt->pkt_cmdlen = cmdlen; 6690 6691 if (real_len < cmdlen) { 6692 goto fail; 6693 } 6694 6695 if (ddi_dma_addr_bind_handle(fc_pkt->pkt_cmd_dma, NULL, 6696 (caddr_t)fc_pkt->pkt_cmd, real_len, 6697 DDI_DMA_WRITE | DDI_DMA_CONSISTENT, cb, NULL, 6698 &pkt_cookie, &fc_pkt->pkt_cmd_cookie_cnt) != 6699 DDI_DMA_MAPPED) { 6700 goto fail; 6701 } 6702 6703 fcip_pkt->fcip_pkt_dma_flags |= FCIP_CMD_DMA_BOUND; 6704 6705 if (fc_pkt->pkt_cmd_cookie_cnt > 6706 fport->fcipp_cmd_dma_attr.dma_attr_sgllen) { 6707 goto fail; 6708 } 6709 6710 ASSERT(fc_pkt->pkt_cmd_cookie_cnt != 0); 6711 6712 cp = fc_pkt->pkt_cmd_cookie = (ddi_dma_cookie_t *)kmem_alloc( 6713 fc_pkt->pkt_cmd_cookie_cnt * sizeof (pkt_cookie), 6714 KM_NOSLEEP); 6715 6716 if (cp == NULL) { 6717 goto fail; 6718 } 6719 6720 *cp = pkt_cookie; 6721 cp++; 6722 for (cnt = 1; cnt < fc_pkt->pkt_cmd_cookie_cnt; cnt++, cp++) { 6723 ddi_dma_nextcookie(fc_pkt->pkt_cmd_dma, &pkt_cookie); 6724 *cp = pkt_cookie; 6725 } 6726 } 6727 6728 if (resplen) { 6729 if (ddi_dma_alloc_handle(fptr->fcip_dip, 6730 &fport->fcipp_resp_dma_attr, cb, NULL, 6731 &fc_pkt->pkt_resp_dma) != DDI_SUCCESS) { 6732 goto fail; 6733 } 6734 6735 if (ddi_dma_mem_alloc(fc_pkt->pkt_resp_dma, resplen, 6736 &fport->fcipp_fca_acc_attr, DDI_DMA_CONSISTENT, 6737 cb, NULL, (caddr_t *)&fc_pkt->pkt_resp, 6738 &real_len, &fc_pkt->pkt_resp_acc) != DDI_SUCCESS) { 6739 goto fail; 6740 } 6741 6742 fcip_pkt->fcip_pkt_dma_flags |= FCIP_RESP_DMA_MEM; 6743 6744 if (real_len < resplen) { 6745 goto fail; 6746 } 6747 6748 if (ddi_dma_addr_bind_handle(fc_pkt->pkt_resp_dma, NULL, 6749 (caddr_t)fc_pkt->pkt_resp, real_len, 6750 DDI_DMA_WRITE | DDI_DMA_CONSISTENT, cb, NULL, 6751 &pkt_cookie, &fc_pkt->pkt_resp_cookie_cnt) != 6752 DDI_DMA_MAPPED) { 6753 goto fail; 6754 } 6755 6756 fcip_pkt->fcip_pkt_dma_flags |= FCIP_RESP_DMA_BOUND; 6757 fc_pkt->pkt_rsplen = resplen; 6758 6759 if (fc_pkt->pkt_resp_cookie_cnt > 6760 fport->fcipp_resp_dma_attr.dma_attr_sgllen) { 6761 goto fail; 6762 } 6763 6764 ASSERT(fc_pkt->pkt_resp_cookie_cnt != 0); 6765 6766 cp = fc_pkt->pkt_resp_cookie = (ddi_dma_cookie_t *)kmem_alloc( 6767 fc_pkt->pkt_resp_cookie_cnt * sizeof (pkt_cookie), 6768 KM_NOSLEEP); 6769 6770 if (cp == NULL) { 6771 goto fail; 6772 } 6773 6774 *cp = pkt_cookie; 6775 cp++; 6776 for (cnt = 1; cnt < fc_pkt->pkt_resp_cookie_cnt; cnt++, cp++) { 6777 ddi_dma_nextcookie(fc_pkt->pkt_resp_dma, &pkt_cookie); 6778 *cp = pkt_cookie; 6779 } 6780 } 6781 6782 /* 6783 * Initialize pkt_pd prior to calling fc_ulp_init_packet 6784 */ 6785 6786 fc_pkt->pkt_pd = pd; 6787 6788 /* 6789 * Ask the FCA to bless the internal packet 6790 */ 6791 if (fc_ulp_init_packet((opaque_t)fport->fcipp_handle, 6792 fc_pkt, flags) != FC_SUCCESS) { 6793 goto fail; 6794 } 6795 6796 /* 6797 * Keep track of # of ipkts alloc-ed 6798 * This function can get called with mutex either held or not. So, we'll 6799 * grab mutex if it is not already held by this thread. 6800 * This has to be cleaned up someday. 6801 */ 6802 if (!MUTEX_HELD(&fptr->fcip_mutex)) { 6803 held_here = 1; 6804 mutex_enter(&fptr->fcip_mutex); 6805 } 6806 6807 fptr->fcip_num_ipkts_pending++; 6808 6809 if (held_here) 6810 mutex_exit(&fptr->fcip_mutex); 6811 6812 return (fcip_pkt); 6813 fail: 6814 if (fcip_pkt) { 6815 fcip_ipkt_free(fcip_pkt); 6816 } 6817 6818 return (NULL); 6819 } 6820 6821 /* 6822 * free up an internal IP packet (like a FARP pkt etc) 6823 */ 6824 static void 6825 fcip_ipkt_free(fcip_pkt_t *fcip_pkt) 6826 { 6827 fc_packet_t *fc_pkt; 6828 struct fcip *fptr = fcip_pkt->fcip_pkt_fptr; 6829 fcip_port_info_t *fport = fptr->fcip_port_info; 6830 6831 ASSERT(fptr != NULL); 6832 ASSERT(!mutex_owned(&fptr->fcip_mutex)); 6833 6834 /* One less ipkt to wait for */ 6835 mutex_enter(&fptr->fcip_mutex); 6836 if (fptr->fcip_num_ipkts_pending) /* Safety check */ 6837 fptr->fcip_num_ipkts_pending--; 6838 mutex_exit(&fptr->fcip_mutex); 6839 6840 fc_pkt = FCIP_PKT_TO_FC_PKT(fcip_pkt); 6841 6842 if (fc_pkt->pkt_cmd_cookie != NULL) { 6843 kmem_free(fc_pkt->pkt_cmd_cookie, fc_pkt->pkt_cmd_cookie_cnt * 6844 sizeof (ddi_dma_cookie_t)); 6845 fc_pkt->pkt_cmd_cookie = NULL; 6846 } 6847 6848 if (fc_pkt->pkt_resp_cookie != NULL) { 6849 kmem_free(fc_pkt->pkt_resp_cookie, fc_pkt->pkt_resp_cookie_cnt * 6850 sizeof (ddi_dma_cookie_t)); 6851 fc_pkt->pkt_resp_cookie = NULL; 6852 } 6853 6854 if (fc_ulp_uninit_packet(fport->fcipp_handle, fc_pkt) != FC_SUCCESS) { 6855 FCIP_DEBUG(FCIP_DEBUG_ELS, (CE_WARN, 6856 "fc_ulp_uninit_pkt failed for internal fc pkt 0x%p", 6857 (void *)fc_pkt)); 6858 } 6859 fcip_free_pkt_dma(fcip_pkt); 6860 kmem_free(fcip_pkt, (sizeof (fcip_pkt_t) + fport->fcipp_fca_pkt_size)); 6861 } 6862 6863 /* 6864 * initialize a unicast request. This is a misnomer because even the 6865 * broadcast requests are initialized with this routine 6866 */ 6867 static void 6868 fcip_init_unicast_pkt(fcip_pkt_t *fcip_pkt, fc_portid_t sid, fc_portid_t did, 6869 void (*comp) ()) 6870 { 6871 fc_packet_t *fc_pkt; 6872 fc_frame_hdr_t *fr_hdr; 6873 struct fcip *fptr = fcip_pkt->fcip_pkt_fptr; 6874 6875 fc_pkt = FCIP_PKT_TO_FC_PKT(fcip_pkt); 6876 fr_hdr = &fc_pkt->pkt_cmd_fhdr; 6877 6878 fr_hdr->r_ctl = R_CTL_DEVICE_DATA | R_CTL_UNSOL_DATA; 6879 fr_hdr->s_id = sid.port_id; 6880 fr_hdr->d_id = did.port_id; 6881 fr_hdr->type = FC_TYPE_IS8802_SNAP; 6882 fr_hdr->f_ctl = F_CTL_FIRST_SEQ | F_CTL_LAST_SEQ; 6883 fr_hdr->df_ctl = DF_CTL_NET_HDR; 6884 fr_hdr->seq_cnt = 0; 6885 fr_hdr->ox_id = 0xffff; 6886 fr_hdr->rx_id = 0xffff; 6887 fr_hdr->ro = 0; 6888 /* 6889 * reset all the length fields 6890 */ 6891 fc_pkt->pkt_rsplen = 0; 6892 fc_pkt->pkt_datalen = 0; 6893 fc_pkt->pkt_comp = comp; 6894 if (comp) { 6895 fc_pkt->pkt_tran_flags |= FC_TRAN_INTR; 6896 } else { 6897 fc_pkt->pkt_tran_flags |= FC_TRAN_NO_INTR; 6898 } 6899 fc_pkt->pkt_tran_type = FC_PKT_OUTBOUND | FC_PKT_IP_WRITE; 6900 fc_pkt->pkt_timeout = fcip_pkt_ttl_ticks; 6901 fcip_pkt->fcip_pkt_ttl = fptr->fcip_timeout_ticks + fc_pkt->pkt_timeout; 6902 } 6903 6904 6905 /* 6906 * Initialize a fcip_packet for broadcast data transfers 6907 */ 6908 static void 6909 fcip_init_broadcast_pkt(fcip_pkt_t *fcip_pkt, void (*comp) (), int is_els) 6910 { 6911 fc_packet_t *fc_pkt; 6912 fc_frame_hdr_t *fr_hdr; 6913 struct fcip *fptr = fcip_pkt->fcip_pkt_fptr; 6914 fcip_port_info_t *fport = fptr->fcip_port_info; 6915 uint32_t sid; 6916 uint32_t did; 6917 6918 FCIP_TNF_PROBE_1((fcip_init_broadcast_pkt, "fcip io", /* CSTYLED */, 6919 tnf_string, msg, "enter")); 6920 fc_pkt = FCIP_PKT_TO_FC_PKT(fcip_pkt); 6921 fr_hdr = &fc_pkt->pkt_cmd_fhdr; 6922 sid = fport->fcipp_sid.port_id; 6923 6924 if (is_els) { 6925 fr_hdr->r_ctl = R_CTL_ELS_REQ; 6926 } else { 6927 fr_hdr->r_ctl = R_CTL_DEVICE_DATA | R_CTL_UNSOL_DATA; 6928 } 6929 fr_hdr->s_id = sid; 6930 /* 6931 * The destination broadcast address depends on the topology 6932 * of the underlying port 6933 */ 6934 did = fptr->fcip_broadcast_did; 6935 /* 6936 * mark pkt a broadcast pkt 6937 */ 6938 fc_pkt->pkt_tran_type = FC_PKT_BROADCAST; 6939 6940 fr_hdr->d_id = did; 6941 fr_hdr->type = FC_TYPE_IS8802_SNAP; 6942 fr_hdr->f_ctl = F_CTL_FIRST_SEQ | F_CTL_LAST_SEQ | F_CTL_END_SEQ; 6943 fr_hdr->f_ctl &= ~(F_CTL_SEQ_INITIATIVE); 6944 fr_hdr->df_ctl = DF_CTL_NET_HDR; 6945 fr_hdr->seq_cnt = 0; 6946 fr_hdr->ox_id = 0xffff; 6947 fr_hdr->rx_id = 0xffff; 6948 fr_hdr->ro = 0; 6949 fc_pkt->pkt_comp = comp; 6950 6951 if (comp) { 6952 fc_pkt->pkt_tran_flags |= FC_TRAN_INTR; 6953 } else { 6954 fc_pkt->pkt_tran_flags |= FC_TRAN_NO_INTR; 6955 } 6956 6957 fc_pkt->pkt_tran_type = FC_PKT_BROADCAST; 6958 fc_pkt->pkt_timeout = fcip_pkt_ttl_ticks; 6959 fcip_pkt->fcip_pkt_ttl = fptr->fcip_timeout_ticks + fc_pkt->pkt_timeout; 6960 } 6961 6962 6963 6964 /* 6965 * Free up all DMA resources associated with an allocated packet 6966 */ 6967 static void 6968 fcip_free_pkt_dma(fcip_pkt_t *fcip_pkt) 6969 { 6970 fc_packet_t *fc_pkt; 6971 6972 fc_pkt = FCIP_PKT_TO_FC_PKT(fcip_pkt); 6973 6974 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, 6975 (CE_NOTE, "in freepktdma : flags 0x%x", 6976 fcip_pkt->fcip_pkt_dma_flags)); 6977 6978 if (fcip_pkt->fcip_pkt_dma_flags & FCIP_CMD_DMA_BOUND) { 6979 (void) ddi_dma_unbind_handle(fc_pkt->pkt_cmd_dma); 6980 } 6981 if (fcip_pkt->fcip_pkt_dma_flags & FCIP_CMD_DMA_MEM) { 6982 ddi_dma_mem_free(&fc_pkt->pkt_cmd_acc); 6983 } 6984 6985 if (fcip_pkt->fcip_pkt_dma_flags & FCIP_RESP_DMA_BOUND) { 6986 (void) ddi_dma_unbind_handle(fc_pkt->pkt_resp_dma); 6987 } 6988 if (fcip_pkt->fcip_pkt_dma_flags & FCIP_RESP_DMA_MEM) { 6989 ddi_dma_mem_free(&fc_pkt->pkt_resp_acc); 6990 } 6991 /* 6992 * for internal commands, we need to free up the dma handles too. 6993 * This is done in the cache destructor for non internal cmds 6994 */ 6995 if (fcip_pkt->fcip_pkt_flags & FCIP_PKT_INTERNAL) { 6996 if (fc_pkt->pkt_cmd_dma) { 6997 ddi_dma_free_handle(&fc_pkt->pkt_cmd_dma); 6998 } 6999 if (fc_pkt->pkt_resp_dma) { 7000 ddi_dma_free_handle(&fc_pkt->pkt_resp_dma); 7001 } 7002 } 7003 } 7004 7005 7006 /* 7007 * helper routine to generate a string, given an ether addr 7008 */ 7009 static void 7010 fcip_ether_to_str(struct ether_addr *e, caddr_t s) 7011 { 7012 int i; 7013 7014 for (i = 0; i < sizeof (struct ether_addr); i++, s += 2) { 7015 FCIP_DEBUG(FCIP_DEBUG_MISC, 7016 (CE_CONT, "0x%02X:", e->ether_addr_octet[i])); 7017 (void) sprintf(s, "%02X", e->ether_addr_octet[i]); 7018 } 7019 7020 *s = '\0'; 7021 } 7022 7023 /* 7024 * When a broadcast request comes from the upper streams modules, it 7025 * is ugly to look into every datagram to figure out if it is a broadcast 7026 * datagram or a unicast packet. Instead just add the broadcast entries 7027 * into our routing and dest tables and the standard hash table look ups 7028 * will find the entries. It is a lot cleaner this way. Also Solaris ifconfig 7029 * seems to be very ethernet specific and it requires broadcasts to the 7030 * ether broadcast addr of 0xffffffffff to succeed even though we specified 7031 * in the dl_info request that our broadcast MAC addr is 0x0000000000 7032 * (can't figure out why RFC2625 did this though). So add broadcast entries 7033 * for both MAC address 7034 */ 7035 static int 7036 fcip_dest_add_broadcast_entry(struct fcip *fptr, int new_flag) 7037 { 7038 fc_portmap_t map; 7039 struct fcip_routing_table *frp; 7040 uint32_t did; 7041 la_wwn_t broadcast_wwn; 7042 7043 /* 7044 * get port_id of destination for broadcast - this is topology 7045 * dependent 7046 */ 7047 did = fptr->fcip_broadcast_did; 7048 7049 ether_to_wwn(&fcip_arpbroadcast_addr, &broadcast_wwn); 7050 bcopy((void *)&broadcast_wwn, (void *)&map.map_pwwn, sizeof (la_wwn_t)); 7051 bcopy((void *)&broadcast_wwn, (void *)&map.map_nwwn, sizeof (la_wwn_t)); 7052 7053 map.map_did.port_id = did; 7054 map.map_hard_addr.hard_addr = did; 7055 map.map_state = PORT_DEVICE_VALID; 7056 if (new_flag) { 7057 map.map_type = PORT_DEVICE_NEW; 7058 } else { 7059 map.map_type = PORT_DEVICE_CHANGED; 7060 } 7061 map.map_flags = 0; 7062 map.map_pd = NULL; 7063 bzero(&map.map_fc4_types, sizeof (map.map_fc4_types)); 7064 fcip_rt_update(fptr, &map, 1); 7065 mutex_enter(&fptr->fcip_rt_mutex); 7066 frp = fcip_lookup_rtable(fptr, &broadcast_wwn, FCIP_COMPARE_NWWN); 7067 mutex_exit(&fptr->fcip_rt_mutex); 7068 if (frp == NULL) { 7069 return (FC_FAILURE); 7070 } 7071 (void) fcip_add_dest(fptr, frp); 7072 /* 7073 * The Upper IP layers expect the traditional broadcast MAC addr 7074 * of 0xff ff ff ff ff ff to work too if we want to plumb the fcip 7075 * stream through the /etc/hostname.fcipXX file. Instead of checking 7076 * each phys addr for a match with fcip's ARP header broadcast 7077 * addr (0x00 00 00 00 00 00), its simply easier to add another 7078 * broadcast entry for 0xff ff ff ff ff ff. 7079 */ 7080 ether_to_wwn(&fcipnhbroadcastaddr, &broadcast_wwn); 7081 bcopy((void *)&broadcast_wwn, (void *)&map.map_pwwn, sizeof (la_wwn_t)); 7082 bcopy((void *)&broadcast_wwn, (void *)&map.map_nwwn, sizeof (la_wwn_t)); 7083 fcip_rt_update(fptr, &map, 1); 7084 mutex_enter(&fptr->fcip_rt_mutex); 7085 frp = fcip_lookup_rtable(fptr, &broadcast_wwn, FCIP_COMPARE_NWWN); 7086 mutex_exit(&fptr->fcip_rt_mutex); 7087 if (frp == NULL) { 7088 return (FC_FAILURE); 7089 } 7090 (void) fcip_add_dest(fptr, frp); 7091 return (FC_SUCCESS); 7092 } 7093 7094 /* 7095 * We need to obtain the D_ID of the broadcast port for transmitting all 7096 * our broadcast (and multicast) requests. The broadcast D_ID as we know 7097 * is dependent on the link topology 7098 */ 7099 static uint32_t 7100 fcip_get_broadcast_did(struct fcip *fptr) 7101 { 7102 fcip_port_info_t *fport = fptr->fcip_port_info; 7103 uint32_t did = 0; 7104 uint32_t sid; 7105 7106 FCIP_TNF_PROBE_2((fcip_get_broadcast_did, "fcip io", /* CSTYLED */, 7107 tnf_string, msg, "enter", 7108 tnf_opaque, fptr, fptr)); 7109 7110 sid = fport->fcipp_sid.port_id; 7111 7112 switch (fport->fcipp_topology) { 7113 7114 case FC_TOP_PT_PT: { 7115 fc_portmap_t *port_map = NULL; 7116 uint32_t listlen = 0; 7117 7118 if (fc_ulp_getportmap(fport->fcipp_handle, &port_map, 7119 &listlen, FC_ULP_PLOGI_DONTCARE) == FC_SUCCESS) { 7120 FCIP_DEBUG(FCIP_DEBUG_INIT, (CE_NOTE, 7121 "fcip_gpmap: listlen : 0x%x", listlen)); 7122 if (listlen == 1) { 7123 did = port_map->map_did.port_id; 7124 } 7125 } 7126 if (port_map) { 7127 kmem_free(port_map, listlen * sizeof (fc_portmap_t)); 7128 } 7129 if (listlen != 1) { 7130 /* Dummy return value */ 7131 return (0x00FFFFFF); 7132 } 7133 break; 7134 } 7135 7136 case FC_TOP_NO_NS: 7137 /* FALLTHROUGH */ 7138 case FC_TOP_FABRIC: 7139 /* 7140 * The broadcast address is the same whether or not 7141 * the switch/fabric contains a Name service. 7142 */ 7143 did = 0x00FFFFFF; 7144 break; 7145 7146 case FC_TOP_PUBLIC_LOOP: 7147 /* 7148 * The open replicate primitive must not be used. The 7149 * broadcast sequence is simply sent to ALPA 0x00. The 7150 * fabric controller then propagates the broadcast to all 7151 * other ports. The fabric propagates the broadcast by 7152 * using the OPNfr primitive. 7153 */ 7154 did = 0x00; 7155 break; 7156 7157 case FC_TOP_PRIVATE_LOOP: 7158 /* 7159 * The source port for broadcast in private loop mode 7160 * must send an OPN(fr) signal forcing all ports in the 7161 * loop to replicate the frames that they receive. 7162 */ 7163 did = 0x00FFFFFF; 7164 break; 7165 7166 case FC_TOP_UNKNOWN: 7167 /* FALLTHROUGH */ 7168 default: 7169 did = sid; 7170 FCIP_DEBUG(FCIP_DEBUG_INIT, (CE_WARN, 7171 "fcip(0x%x):unknown topology in init_broadcast_pkt", 7172 fptr->fcip_instance)); 7173 break; 7174 } 7175 FCIP_TNF_PROBE_2((fcip_get_broadcast_did, "fcip io", /* CSTYLED */, 7176 tnf_string, msg, "return", 7177 tnf_opaque, did, did)); 7178 7179 return (did); 7180 } 7181 7182 7183 /* 7184 * fcip timeout performs 2 operations: 7185 * 1. timeout any packets sent to the FCA for which a callback hasn't 7186 * happened. If you are wondering why we need a callback since all 7187 * traffic in FCIP is unidirectional, hence all exchanges are unidirectional 7188 * but wait, we can only free up the resources after we know the FCA has 7189 * DMA'ed out the data. pretty obvious eh :) 7190 * 7191 * 2. Retire and routing table entries we marked up for retiring. This is 7192 * to give the link a chance to recover instead of marking a port down 7193 * when we have lost all communication with it after a link transition 7194 */ 7195 static void 7196 fcip_timeout(void *arg) 7197 { 7198 struct fcip *fptr = (struct fcip *)arg; 7199 int i; 7200 fcip_pkt_t *fcip_pkt; 7201 struct fcip_dest *fdestp; 7202 int index; 7203 struct fcip_routing_table *frtp; 7204 int dispatch_rte_removal = 0; 7205 7206 mutex_enter(&fptr->fcip_mutex); 7207 7208 fptr->fcip_flags |= FCIP_IN_TIMEOUT; 7209 fptr->fcip_timeout_ticks += fcip_tick_incr; 7210 7211 if (fptr->fcip_flags & (FCIP_DETACHED | FCIP_DETACHING | \ 7212 FCIP_SUSPENDED | FCIP_POWER_DOWN)) { 7213 fptr->fcip_flags &= ~(FCIP_IN_TIMEOUT); 7214 mutex_exit(&fptr->fcip_mutex); 7215 return; 7216 } 7217 7218 if (fptr->fcip_port_state == FCIP_PORT_OFFLINE) { 7219 if (fptr->fcip_timeout_ticks > fptr->fcip_mark_offline) { 7220 fptr->fcip_flags |= FCIP_LINK_DOWN; 7221 } 7222 } 7223 if (!fptr->fcip_flags & FCIP_RTE_REMOVING) { 7224 dispatch_rte_removal = 1; 7225 } 7226 mutex_exit(&fptr->fcip_mutex); 7227 7228 /* 7229 * Check if we have any Invalid routing table entries in our 7230 * hashtable we have marked off for deferred removal. If any, 7231 * we can spawn a taskq thread to do the cleanup for us. We 7232 * need to avoid cleanup in the timeout thread since we may 7233 * have to wait for outstanding commands to complete before 7234 * we retire a routing table entry. Also dispatch the taskq 7235 * thread only if we are already do not have a taskq thread 7236 * dispatched. 7237 */ 7238 if (dispatch_rte_removal) { 7239 mutex_enter(&fptr->fcip_rt_mutex); 7240 for (index = 0; index < FCIP_RT_HASH_ELEMS; index++) { 7241 frtp = fptr->fcip_rtable[index]; 7242 while (frtp) { 7243 if ((frtp->fcipr_state == FCIP_RT_INVALID) && 7244 (fptr->fcip_timeout_ticks > 7245 frtp->fcipr_invalid_timeout)) { 7246 /* 7247 * If we cannot schedule a task thread 7248 * let us attempt again on the next 7249 * tick rather than call 7250 * fcip_rte_remove_deferred() from here 7251 * directly since the routine can sleep. 7252 */ 7253 frtp->fcipr_state = FCIP_RT_RETIRED; 7254 7255 mutex_enter(&fptr->fcip_mutex); 7256 fptr->fcip_flags |= FCIP_RTE_REMOVING; 7257 mutex_exit(&fptr->fcip_mutex); 7258 7259 if (taskq_dispatch(fptr->fcip_tq, 7260 fcip_rte_remove_deferred, fptr, 7261 KM_NOSLEEP) == 0) { 7262 /* 7263 * failed - so mark the entry 7264 * as invalid again. 7265 */ 7266 frtp->fcipr_state = 7267 FCIP_RT_INVALID; 7268 7269 mutex_enter(&fptr->fcip_mutex); 7270 fptr->fcip_flags &= 7271 ~FCIP_RTE_REMOVING; 7272 mutex_exit(&fptr->fcip_mutex); 7273 } 7274 } 7275 frtp = frtp->fcipr_next; 7276 } 7277 } 7278 mutex_exit(&fptr->fcip_rt_mutex); 7279 } 7280 7281 mutex_enter(&fptr->fcip_dest_mutex); 7282 7283 /* 7284 * Now timeout any packets stuck with the transport/FCA for too long 7285 */ 7286 for (i = 0; i < FCIP_DEST_HASH_ELEMS; i++) { 7287 fdestp = fptr->fcip_dest[i]; 7288 while (fdestp != NULL) { 7289 mutex_enter(&fdestp->fcipd_mutex); 7290 for (fcip_pkt = fdestp->fcipd_head; fcip_pkt != NULL; 7291 fcip_pkt = fcip_pkt->fcip_pkt_next) { 7292 if (fcip_pkt->fcip_pkt_flags & 7293 (FCIP_PKT_RETURNED | FCIP_PKT_IN_TIMEOUT | 7294 FCIP_PKT_IN_ABORT)) { 7295 continue; 7296 } 7297 if (fptr->fcip_timeout_ticks > 7298 fcip_pkt->fcip_pkt_ttl) { 7299 fcip_pkt->fcip_pkt_flags |= 7300 FCIP_PKT_IN_TIMEOUT; 7301 7302 mutex_exit(&fdestp->fcipd_mutex); 7303 if (taskq_dispatch(fptr->fcip_tq, 7304 fcip_pkt_timeout, fcip_pkt, 7305 KM_NOSLEEP) == 0) { 7306 /* 7307 * timeout immediately 7308 */ 7309 fcip_pkt_timeout(fcip_pkt); 7310 } 7311 mutex_enter(&fdestp->fcipd_mutex); 7312 /* 7313 * The linked list is altered because 7314 * of one of the following reasons: 7315 * a. Timeout code dequeued a pkt 7316 * b. Pkt completion happened 7317 * 7318 * So restart the spin starting at 7319 * the head again; This is a bit 7320 * excessive, but okay since 7321 * fcip_timeout_ticks isn't incremented 7322 * for this spin, we will skip the 7323 * not-to-be-timedout packets quickly 7324 */ 7325 fcip_pkt = fdestp->fcipd_head; 7326 if (fcip_pkt == NULL) { 7327 break; 7328 } 7329 } 7330 } 7331 mutex_exit(&fdestp->fcipd_mutex); 7332 fdestp = fdestp->fcipd_next; 7333 } 7334 } 7335 mutex_exit(&fptr->fcip_dest_mutex); 7336 7337 /* 7338 * reschedule the timeout thread 7339 */ 7340 mutex_enter(&fptr->fcip_mutex); 7341 7342 fptr->fcip_timeout_id = timeout(fcip_timeout, fptr, 7343 drv_usectohz(1000000)); 7344 fptr->fcip_flags &= ~(FCIP_IN_TIMEOUT); 7345 mutex_exit(&fptr->fcip_mutex); 7346 } 7347 7348 7349 /* 7350 * This routine is either called from taskq or directly from fcip_timeout 7351 * does the actual job of aborting the packet 7352 */ 7353 static void 7354 fcip_pkt_timeout(void *arg) 7355 { 7356 fcip_pkt_t *fcip_pkt = (fcip_pkt_t *)arg; 7357 struct fcip_dest *fdestp; 7358 struct fcip *fptr; 7359 fc_packet_t *fc_pkt; 7360 fcip_port_info_t *fport; 7361 int rval; 7362 7363 fdestp = fcip_pkt->fcip_pkt_dest; 7364 fptr = fcip_pkt->fcip_pkt_fptr; 7365 fport = fptr->fcip_port_info; 7366 fc_pkt = FCIP_PKT_TO_FC_PKT(fcip_pkt); 7367 7368 /* 7369 * try to abort the pkt 7370 */ 7371 fcip_pkt->fcip_pkt_flags |= FCIP_PKT_IN_ABORT; 7372 rval = fc_ulp_abort(fport->fcipp_handle, fc_pkt, KM_NOSLEEP); 7373 7374 FCIP_DEBUG(FCIP_DEBUG_DOWNSTREAM, 7375 (CE_NOTE, "fc_ulp_abort returns: 0x%x", rval)); 7376 7377 if (rval == FC_SUCCESS) { 7378 ASSERT(fdestp != NULL); 7379 7380 /* 7381 * dequeue the pkt from the dest structure pkt list 7382 */ 7383 fcip_pkt->fcip_pkt_flags &= ~FCIP_PKT_IN_ABORT; 7384 mutex_enter(&fdestp->fcipd_mutex); 7385 rval = fcip_fdestp_dequeue_pkt(fdestp, fcip_pkt); 7386 ASSERT(rval == 1); 7387 mutex_exit(&fdestp->fcipd_mutex); 7388 7389 /* 7390 * Now cleanup the pkt and free the mblk 7391 */ 7392 fcip_pkt_free(fcip_pkt, 1); 7393 } else { 7394 /* 7395 * abort failed - just mark the pkt as done and 7396 * wait for it to complete in fcip_pkt_callback since 7397 * the pkt has already been xmitted by the FCA 7398 */ 7399 fcip_pkt->fcip_pkt_flags &= ~FCIP_PKT_IN_TIMEOUT; 7400 if (fcip_pkt->fcip_pkt_flags & FCIP_PKT_RETURNED) { 7401 fcip_pkt->fcip_pkt_flags &= ~FCIP_PKT_IN_ABORT; 7402 mutex_enter(&fdestp->fcipd_mutex); 7403 rval = fcip_fdestp_dequeue_pkt(fdestp, fcip_pkt); 7404 ASSERT(rval == 1); 7405 mutex_exit(&fdestp->fcipd_mutex); 7406 7407 fcip_pkt_free(fcip_pkt, 1); 7408 } 7409 return; 7410 } 7411 } 7412 7413 7414 /* 7415 * Remove a routing table entry marked for deferred removal. This routine 7416 * unlike fcip_pkt_timeout, is always called from a taskq context 7417 */ 7418 static void 7419 fcip_rte_remove_deferred(void *arg) 7420 { 7421 struct fcip *fptr = (struct fcip *)arg; 7422 int hash_bucket; 7423 struct fcip_dest *fdestp; 7424 la_wwn_t *pwwn; 7425 int index; 7426 struct fcip_routing_table *frtp, *frtp_next, *frtp_prev; 7427 7428 7429 mutex_enter(&fptr->fcip_rt_mutex); 7430 for (index = 0; index < FCIP_RT_HASH_ELEMS; index++) { 7431 frtp = fptr->fcip_rtable[index]; 7432 frtp_prev = NULL; 7433 while (frtp) { 7434 frtp_next = frtp->fcipr_next; 7435 7436 if (frtp->fcipr_state == FCIP_RT_RETIRED) { 7437 7438 pwwn = &frtp->fcipr_pwwn; 7439 /* 7440 * Get hold of destination pointer 7441 */ 7442 mutex_enter(&fptr->fcip_dest_mutex); 7443 7444 hash_bucket = FCIP_DEST_HASH(pwwn->raw_wwn); 7445 ASSERT(hash_bucket < FCIP_DEST_HASH_ELEMS); 7446 7447 fdestp = fptr->fcip_dest[hash_bucket]; 7448 while (fdestp != NULL) { 7449 mutex_enter(&fdestp->fcipd_mutex); 7450 if (fdestp->fcipd_rtable) { 7451 if (fcip_wwn_compare(pwwn, 7452 &fdestp->fcipd_pwwn, 7453 FCIP_COMPARE_PWWN) == 0) { 7454 mutex_exit( 7455 &fdestp->fcipd_mutex); 7456 break; 7457 } 7458 } 7459 mutex_exit(&fdestp->fcipd_mutex); 7460 fdestp = fdestp->fcipd_next; 7461 } 7462 7463 mutex_exit(&fptr->fcip_dest_mutex); 7464 if (fdestp == NULL) { 7465 frtp_prev = frtp; 7466 frtp = frtp_next; 7467 continue; 7468 } 7469 7470 mutex_enter(&fdestp->fcipd_mutex); 7471 if (fdestp->fcipd_ncmds) { 7472 /* 7473 * Instead of waiting to drain commands 7474 * let us revisit this RT entry in 7475 * the next pass. 7476 */ 7477 mutex_exit(&fdestp->fcipd_mutex); 7478 frtp_prev = frtp; 7479 frtp = frtp_next; 7480 continue; 7481 } 7482 7483 /* 7484 * We are clean, so remove the RTE 7485 */ 7486 fdestp->fcipd_rtable = NULL; 7487 mutex_exit(&fdestp->fcipd_mutex); 7488 7489 FCIP_TNF_PROBE_2((fcip_rte_remove_deferred, 7490 "fcip io", /* CSTYLED */, 7491 tnf_string, msg, 7492 "remove retired routing entry", 7493 tnf_int, index, index)); 7494 7495 if (frtp_prev == NULL) { 7496 /* first element */ 7497 fptr->fcip_rtable[index] = 7498 frtp->fcipr_next; 7499 } else { 7500 frtp_prev->fcipr_next = 7501 frtp->fcipr_next; 7502 } 7503 kmem_free(frtp, 7504 sizeof (struct fcip_routing_table)); 7505 7506 frtp = frtp_next; 7507 } else { 7508 frtp_prev = frtp; 7509 frtp = frtp_next; 7510 } 7511 } 7512 } 7513 mutex_exit(&fptr->fcip_rt_mutex); 7514 /* 7515 * Clear the RTE_REMOVING flag 7516 */ 7517 mutex_enter(&fptr->fcip_mutex); 7518 fptr->fcip_flags &= ~FCIP_RTE_REMOVING; 7519 mutex_exit(&fptr->fcip_mutex); 7520 } 7521 7522 /* 7523 * Walk through all the dest hash table entries and count up the total 7524 * no. of packets outstanding against a given port 7525 */ 7526 static int 7527 fcip_port_get_num_pkts(struct fcip *fptr) 7528 { 7529 int num_cmds = 0; 7530 int i; 7531 struct fcip_dest *fdestp; 7532 7533 ASSERT(mutex_owned(&fptr->fcip_dest_mutex)); 7534 7535 for (i = 0; i < FCIP_DEST_HASH_ELEMS; i++) { 7536 fdestp = fptr->fcip_dest[i]; 7537 while (fdestp != NULL) { 7538 mutex_enter(&fdestp->fcipd_mutex); 7539 7540 ASSERT(fdestp->fcipd_ncmds >= 0); 7541 7542 if (fdestp->fcipd_ncmds > 0) { 7543 num_cmds += fdestp->fcipd_ncmds; 7544 } 7545 mutex_exit(&fdestp->fcipd_mutex); 7546 fdestp = fdestp->fcipd_next; 7547 } 7548 } 7549 7550 return (num_cmds); 7551 } 7552 7553 7554 /* 7555 * Walk through the routing table for this state instance and see if there is a 7556 * PLOGI in progress for any of the entries. Return success even if we find one. 7557 */ 7558 static int 7559 fcip_plogi_in_progress(struct fcip *fptr) 7560 { 7561 int i; 7562 struct fcip_routing_table *frp; 7563 7564 ASSERT(mutex_owned(&fptr->fcip_rt_mutex)); 7565 7566 for (i = 0; i < FCIP_RT_HASH_ELEMS; i++) { 7567 frp = fptr->fcip_rtable[i]; 7568 while (frp) { 7569 if (frp->fcipr_state == FCIP_RT_LOGIN_PROGRESS) { 7570 /* Found an entry where PLOGI is in progress */ 7571 return (1); 7572 } 7573 frp = frp->fcipr_next; 7574 } 7575 } 7576 7577 return (0); 7578 } 7579 7580 /* 7581 * Walk through the fcip port global list and check if the given port exists in 7582 * the list. Returns "0" if port exists and "1" if otherwise. 7583 */ 7584 static int 7585 fcip_check_port_exists(struct fcip *fptr) 7586 { 7587 fcip_port_info_t *cur_fport; 7588 fcip_port_info_t *fport; 7589 7590 mutex_enter(&fcip_global_mutex); 7591 fport = fptr->fcip_port_info; 7592 cur_fport = fcip_port_head; 7593 while (cur_fport != NULL) { 7594 if (cur_fport == fport) { 7595 /* Found */ 7596 mutex_exit(&fcip_global_mutex); 7597 return (0); 7598 } else { 7599 cur_fport = cur_fport->fcipp_next; 7600 } 7601 } 7602 mutex_exit(&fcip_global_mutex); 7603 7604 return (1); 7605 } 7606 7607 /* 7608 * Constructor to initialize the sendup elements for callback into 7609 * modules upstream 7610 */ 7611 7612 /* ARGSUSED */ 7613 static int 7614 fcip_sendup_constructor(void *buf, void *arg, int flags) 7615 { 7616 struct fcip_sendup_elem *msg_elem = (struct fcip_sendup_elem *)buf; 7617 fcip_port_info_t *fport = (fcip_port_info_t *)arg; 7618 7619 ASSERT(fport != NULL); 7620 7621 msg_elem->fcipsu_mp = NULL; 7622 msg_elem->fcipsu_func = NULL; 7623 msg_elem->fcipsu_next = NULL; 7624 7625 return (FCIP_SUCCESS); 7626 }