1 /* 2 * CDDL HEADER START 3 * 4 * The contents of this file are subject to the terms of the 5 * Common Development and Distribution License (the "License"). 6 * You may not use this file except in compliance with the License. 7 * 8 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE 9 * or http://www.opensolaris.org/os/licensing. 10 * See the License for the specific language governing permissions 11 * and limitations under the License. 12 * 13 * When distributing Covered Code, include this CDDL HEADER in each 14 * file and include the License file at usr/src/OPENSOLARIS.LICENSE. 15 * If applicable, add the following below this CDDL HEADER, with the 16 * fields enclosed by brackets "[]" replaced with your own identifying 17 * information: Portions Copyright [yyyy] [name of copyright owner] 18 * 19 * CDDL HEADER END 20 */ 21 22 /* 23 * Copyright 2009 Sun Microsystems, Inc. All rights reserved. 24 * Use is subject to license terms. 25 */ 26 27 #include <sys/errno.h> 28 #include <sys/types.h> 29 #include <sys/conf.h> 30 #include <sys/kmem.h> 31 #include <sys/ddi.h> 32 #include <sys/stat.h> 33 #include <sys/sunddi.h> 34 #include <sys/file.h> 35 #include <sys/open.h> 36 #include <sys/modctl.h> 37 #include <sys/ddi_impldefs.h> 38 #include <vm/seg_kmem.h> 39 #include <sys/vmsystm.h> 40 #include <sys/sysmacros.h> 41 #include <sys/ddidevmap.h> 42 #include <sys/avl.h> 43 #ifdef __xpv 44 #include <sys/hypervisor.h> 45 #endif 46 47 #include <sys/xsvc.h> 48 49 /* total max memory which can be alloced with ioctl interface */ 50 uint64_t xsvc_max_memory = 10 * 1024 * 1024; 51 52 extern void i86_va_map(caddr_t vaddr, struct as *asp, caddr_t kaddr); 53 54 55 static int xsvc_open(dev_t *devp, int flag, int otyp, cred_t *cred); 56 static int xsvc_close(dev_t devp, int flag, int otyp, cred_t *cred); 57 static int xsvc_ioctl(dev_t dev, int cmd, intptr_t arg, int mode, cred_t *cred, 58 int *rval); 59 static int xsvc_devmap(dev_t dev, devmap_cookie_t dhp, offset_t off, size_t len, 60 size_t *maplen, uint_t model); 61 static int xsvc_attach(dev_info_t *devi, ddi_attach_cmd_t cmd); 62 static int xsvc_detach(dev_info_t *devi, ddi_detach_cmd_t cmd); 63 static int xsvc_getinfo(dev_info_t *dip, ddi_info_cmd_t cmd, void *arg, 64 void **result); 65 66 static struct cb_ops xsvc_cb_ops = { 67 xsvc_open, /* cb_open */ 68 xsvc_close, /* cb_close */ 69 nodev, /* cb_strategy */ 70 nodev, /* cb_print */ 71 nodev, /* cb_dump */ 72 nodev, /* cb_read */ 73 nodev, /* cb_write */ 74 xsvc_ioctl, /* cb_ioctl */ 75 xsvc_devmap, /* cb_devmap */ 76 NULL, /* cb_mmap */ 77 NULL, /* cb_segmap */ 78 nochpoll, /* cb_chpoll */ 79 ddi_prop_op, /* cb_prop_op */ 80 NULL, /* cb_stream */ 81 D_NEW | D_MP | D_64BIT | D_DEVMAP, /* cb_flag */ 82 CB_REV 83 }; 84 85 static struct dev_ops xsvc_dev_ops = { 86 DEVO_REV, /* devo_rev */ 87 0, /* devo_refcnt */ 88 xsvc_getinfo, /* devo_getinfo */ 89 nulldev, /* devo_identify */ 90 nulldev, /* devo_probe */ 91 xsvc_attach, /* devo_attach */ 92 xsvc_detach, /* devo_detach */ 93 nodev, /* devo_reset */ 94 &xsvc_cb_ops, /* devo_cb_ops */ 95 NULL, /* devo_bus_ops */ 96 NULL, /* power */ 97 ddi_quiesce_not_needed, /* quiesce */ 98 }; 99 100 static struct modldrv xsvc_modldrv = { 101 &mod_driverops, /* Type of module. This one is a driver */ 102 "xsvc driver", /* Name of the module. */ 103 &xsvc_dev_ops, /* driver ops */ 104 }; 105 106 static struct modlinkage xsvc_modlinkage = { 107 MODREV_1, 108 (void *) &xsvc_modldrv, 109 NULL 110 }; 111 112 113 static int xsvc_ioctl_alloc_memory(xsvc_state_t *state, void *arg, int mode); 114 static int xsvc_ioctl_flush_memory(xsvc_state_t *state, void *arg, int mode); 115 static int xsvc_ioctl_free_memory(xsvc_state_t *state, void *arg, int mode); 116 static int xsvc_mem_alloc(xsvc_state_t *state, uint64_t key, 117 xsvc_mem_t **mp); 118 static void xsvc_mem_free(xsvc_state_t *state, xsvc_mem_t *mp); 119 static xsvc_mem_t *xsvc_mem_lookup(xsvc_state_t *state, 120 uint64_t key); 121 static int xsvc_mnode_key_compare(const void *q, const void *e); 122 static int xsvc_umem_cookie_alloc(caddr_t kva, size_t size, int flags, 123 ddi_umem_cookie_t *cookiep); 124 static void xsvc_umem_cookie_free(ddi_umem_cookie_t *cookiep); 125 126 127 void *xsvc_statep; 128 129 static ddi_device_acc_attr_t xsvc_device_attr = { 130 DDI_DEVICE_ATTR_V0, 131 DDI_NEVERSWAP_ACC, 132 DDI_STRICTORDER_ACC 133 }; 134 135 static int xsvc_devmap_map(devmap_cookie_t dhp, dev_t dev, uint_t flags, 136 offset_t off, size_t len, void **pvtp); 137 static int xsvc_devmap_dup(devmap_cookie_t dhp, void *pvtp, 138 devmap_cookie_t new_dhp, void **new_pvtp); 139 static void xsvc_devmap_unmap(devmap_cookie_t dhp, void *pvtp, offset_t off, 140 size_t len, devmap_cookie_t new_dhp1, void **new_pvtp1, 141 devmap_cookie_t new_dhp2, void **new_pvtp2); 142 143 144 static struct devmap_callback_ctl xsvc_callbk = { 145 DEVMAP_OPS_REV, 146 xsvc_devmap_map, 147 NULL, 148 xsvc_devmap_dup, 149 xsvc_devmap_unmap 150 }; 151 152 153 /* 154 * _init() 155 * 156 */ 157 int 158 _init(void) 159 { 160 int err; 161 162 err = ddi_soft_state_init(&xsvc_statep, sizeof (xsvc_state_t), 1); 163 if (err != 0) { 164 return (err); 165 } 166 167 err = mod_install(&xsvc_modlinkage); 168 if (err != 0) { 169 ddi_soft_state_fini(&xsvc_statep); 170 return (err); 171 } 172 173 return (0); 174 } 175 176 /* 177 * _info() 178 * 179 */ 180 int 181 _info(struct modinfo *modinfop) 182 { 183 return (mod_info(&xsvc_modlinkage, modinfop)); 184 } 185 186 /* 187 * _fini() 188 * 189 */ 190 int 191 _fini(void) 192 { 193 int err; 194 195 err = mod_remove(&xsvc_modlinkage); 196 if (err != 0) { 197 return (err); 198 } 199 200 ddi_soft_state_fini(&xsvc_statep); 201 202 return (0); 203 } 204 205 /* 206 * xsvc_attach() 207 * 208 */ 209 static int 210 xsvc_attach(dev_info_t *dip, ddi_attach_cmd_t cmd) 211 { 212 xsvc_state_t *state; 213 int maxallocmem; 214 int instance; 215 int err; 216 217 218 switch (cmd) { 219 case DDI_ATTACH: 220 break; 221 222 case DDI_RESUME: 223 return (DDI_SUCCESS); 224 225 default: 226 return (DDI_FAILURE); 227 } 228 229 instance = ddi_get_instance(dip); 230 err = ddi_soft_state_zalloc(xsvc_statep, instance); 231 if (err != DDI_SUCCESS) { 232 return (DDI_FAILURE); 233 } 234 state = ddi_get_soft_state(xsvc_statep, instance); 235 if (state == NULL) { 236 goto attachfail_get_soft_state; 237 } 238 239 state->xs_dip = dip; 240 state->xs_instance = instance; 241 242 /* Initialize allocation count */ 243 mutex_init(&state->xs_mutex, NULL, MUTEX_DRIVER, NULL); 244 state->xs_currently_alloced = 0; 245 246 mutex_init(&state->xs_cookie_mutex, NULL, MUTEX_DRIVER, NULL); 247 248 /* create the minor node (for the ioctl) */ 249 err = ddi_create_minor_node(dip, "xsvc", S_IFCHR, instance, DDI_PSEUDO, 250 0); 251 if (err != DDI_SUCCESS) { 252 goto attachfail_minor_node; 253 } 254 255 /* 256 * the maxallocmem property will override the default (xsvc_max_memory). 257 * This is the maximum total memory the ioctl will allow to be alloced. 258 */ 259 maxallocmem = ddi_prop_get_int(DDI_DEV_T_ANY, state->xs_dip, 260 DDI_PROP_DONTPASS, "maxallocmem", -1); 261 if (maxallocmem >= 0) { 262 xsvc_max_memory = maxallocmem * 1024; 263 } 264 265 /* Initialize list of memory allocs */ 266 mutex_init(&state->xs_mlist.ml_mutex, NULL, MUTEX_DRIVER, NULL); 267 avl_create(&state->xs_mlist.ml_avl, xsvc_mnode_key_compare, 268 sizeof (xsvc_mnode_t), offsetof(xsvc_mnode_t, mn_link)); 269 270 /* Report that driver was loaded */ 271 ddi_report_dev(dip); 272 273 return (DDI_SUCCESS); 274 275 attachfail_minor_node: 276 mutex_destroy(&state->xs_cookie_mutex); 277 mutex_destroy(&state->xs_mutex); 278 attachfail_get_soft_state: 279 (void) ddi_soft_state_free(xsvc_statep, instance); 280 281 return (err); 282 } 283 284 /* 285 * xsvc_detach() 286 * 287 */ 288 static int 289 xsvc_detach(dev_info_t *dip, ddi_detach_cmd_t cmd) 290 { 291 xsvc_state_t *state; 292 xsvc_mnode_t *mnode; 293 xsvc_mem_t *mp; 294 int instance; 295 296 297 instance = ddi_get_instance(dip); 298 state = ddi_get_soft_state(xsvc_statep, instance); 299 if (state == NULL) { 300 return (DDI_FAILURE); 301 } 302 303 switch (cmd) { 304 case DDI_DETACH: 305 break; 306 307 case DDI_SUSPEND: 308 return (DDI_SUCCESS); 309 310 default: 311 return (DDI_FAILURE); 312 } 313 314 ddi_remove_minor_node(dip, NULL); 315 316 /* Free any memory on list */ 317 while ((mnode = avl_first(&state->xs_mlist.ml_avl)) != NULL) { 318 mp = mnode->mn_home; 319 xsvc_mem_free(state, mp); 320 } 321 322 /* remove list */ 323 avl_destroy(&state->xs_mlist.ml_avl); 324 mutex_destroy(&state->xs_mlist.ml_mutex); 325 326 mutex_destroy(&state->xs_cookie_mutex); 327 mutex_destroy(&state->xs_mutex); 328 (void) ddi_soft_state_free(xsvc_statep, state->xs_instance); 329 return (DDI_SUCCESS); 330 } 331 332 /* 333 * xsvc_getinfo() 334 * 335 */ 336 /*ARGSUSED*/ 337 static int 338 xsvc_getinfo(dev_info_t *dip, ddi_info_cmd_t cmd, void *arg, void **result) 339 { 340 xsvc_state_t *state; 341 int instance; 342 dev_t dev; 343 int err; 344 345 346 dev = (dev_t)arg; 347 instance = getminor(dev); 348 349 switch (cmd) { 350 case DDI_INFO_DEVT2DEVINFO: 351 state = ddi_get_soft_state(xsvc_statep, instance); 352 if (state == NULL) { 353 return (DDI_FAILURE); 354 } 355 *result = (void *)state->xs_dip; 356 err = DDI_SUCCESS; 357 break; 358 359 case DDI_INFO_DEVT2INSTANCE: 360 *result = (void *)(uintptr_t)instance; 361 err = DDI_SUCCESS; 362 break; 363 364 default: 365 err = DDI_FAILURE; 366 break; 367 } 368 369 return (err); 370 } 371 372 373 /* 374 * xsvc_open() 375 * 376 */ 377 /*ARGSUSED*/ 378 static int 379 xsvc_open(dev_t *devp, int flag, int otyp, cred_t *cred) 380 { 381 xsvc_state_t *state; 382 int instance; 383 384 instance = getminor(*devp); 385 state = ddi_get_soft_state(xsvc_statep, instance); 386 if (state == NULL) { 387 return (ENXIO); 388 } 389 390 return (0); 391 } 392 393 /* 394 * xsvc_close() 395 * 396 */ 397 /*ARGSUSED*/ 398 static int 399 xsvc_close(dev_t devp, int flag, int otyp, cred_t *cred) 400 { 401 return (0); 402 } 403 404 /* 405 * xsvc_ioctl() 406 * 407 */ 408 /*ARGSUSED*/ 409 static int 410 xsvc_ioctl(dev_t dev, int cmd, intptr_t arg, int mode, cred_t *cred, int *rval) 411 { 412 xsvc_state_t *state; 413 int instance; 414 int err; 415 416 417 err = drv_priv(cred); 418 if (err != 0) { 419 return (EPERM); 420 } 421 instance = getminor(dev); 422 if (instance == -1) { 423 return (EBADF); 424 } 425 state = ddi_get_soft_state(xsvc_statep, instance); 426 if (state == NULL) { 427 return (EBADF); 428 } 429 430 switch (cmd) { 431 case XSVC_ALLOC_MEM: 432 err = xsvc_ioctl_alloc_memory(state, (void *)arg, mode); 433 break; 434 435 case XSVC_FREE_MEM: 436 err = xsvc_ioctl_free_memory(state, (void *)arg, mode); 437 break; 438 439 case XSVC_FLUSH_MEM: 440 err = xsvc_ioctl_flush_memory(state, (void *)arg, mode); 441 break; 442 443 default: 444 err = ENXIO; 445 } 446 447 return (err); 448 } 449 450 /* 451 * xsvc_ioctl_alloc_memory() 452 * 453 */ 454 static int 455 xsvc_ioctl_alloc_memory(xsvc_state_t *state, void *arg, int mode) 456 { 457 xsvc_mem_req_32 params32; 458 xsvc_mloc_32 *usgl32; 459 xsvc_mem_req params; 460 xsvc_mloc_32 sgl32; 461 xsvc_mloc *usgl; 462 xsvc_mem_t *mp; 463 xsvc_mloc sgl; 464 uint64_t key; 465 size_t size; 466 int err; 467 int i; 468 469 470 /* Copy in the params, then get the size and key */ 471 if (ddi_model_convert_from(mode & FMODELS) == DDI_MODEL_ILP32) { 472 err = ddi_copyin(arg, ¶ms32, sizeof (xsvc_mem_req_32), 473 mode); 474 if (err != 0) { 475 return (EFAULT); 476 } 477 478 key = (uint64_t)params32.xsvc_mem_reqid; 479 size = P2ROUNDUP((size_t)params32.xsvc_mem_size, PAGESIZE); 480 } else { 481 err = ddi_copyin(arg, ¶ms, sizeof (xsvc_mem_req), mode); 482 if (err != 0) { 483 return (EFAULT); 484 } 485 key = (uint64_t)params.xsvc_mem_reqid; 486 size = P2ROUNDUP(params.xsvc_mem_size, PAGESIZE); 487 } 488 489 /* 490 * make sure this doesn't put us over the maximum allowed to be 491 * allocated 492 */ 493 mutex_enter(&state->xs_mutex); 494 if ((state->xs_currently_alloced + size) > xsvc_max_memory) { 495 mutex_exit(&state->xs_mutex); 496 return (EAGAIN); 497 } 498 state->xs_currently_alloced += size; 499 mutex_exit(&state->xs_mutex); 500 501 /* get state to track this memory */ 502 err = xsvc_mem_alloc(state, key, &mp); 503 if (err != 0) { 504 return (err); 505 } 506 mp->xm_size = size; 507 508 /* allocate and bind the memory */ 509 mp->xm_dma_attr.dma_attr_version = DMA_ATTR_V0; 510 mp->xm_dma_attr.dma_attr_count_max = (uint64_t)0xFFFFFFFF; 511 mp->xm_dma_attr.dma_attr_burstsizes = 1; 512 mp->xm_dma_attr.dma_attr_minxfer = 1; 513 mp->xm_dma_attr.dma_attr_maxxfer = (uint64_t)0xFFFFFFFF; 514 mp->xm_dma_attr.dma_attr_seg = (uint64_t)0xFFFFFFFF; 515 mp->xm_dma_attr.dma_attr_granular = 1; 516 mp->xm_dma_attr.dma_attr_flags = 0; 517 518 /* Finish converting params */ 519 if (ddi_model_convert_from(mode & FMODELS) == DDI_MODEL_ILP32) { 520 mp->xm_dma_attr.dma_attr_addr_lo = params32.xsvc_mem_addr_lo; 521 mp->xm_dma_attr.dma_attr_addr_hi = params32.xsvc_mem_addr_hi; 522 mp->xm_dma_attr.dma_attr_sgllen = params32.xsvc_mem_sgllen; 523 usgl32 = (xsvc_mloc_32 *)(uintptr_t)params32.xsvc_sg_list; 524 mp->xm_dma_attr.dma_attr_align = P2ROUNDUP( 525 params32.xsvc_mem_align, PAGESIZE); 526 } else { 527 mp->xm_dma_attr.dma_attr_addr_lo = params.xsvc_mem_addr_lo; 528 mp->xm_dma_attr.dma_attr_addr_hi = params.xsvc_mem_addr_hi; 529 mp->xm_dma_attr.dma_attr_sgllen = params.xsvc_mem_sgllen; 530 usgl = (xsvc_mloc *)(uintptr_t)params.xsvc_sg_list; 531 mp->xm_dma_attr.dma_attr_align = P2ROUNDUP( 532 params.xsvc_mem_align, PAGESIZE); 533 } 534 535 mp->xm_device_attr = xsvc_device_attr; 536 537 err = ddi_dma_alloc_handle(state->xs_dip, &mp->xm_dma_attr, 538 DDI_DMA_SLEEP, NULL, &mp->xm_dma_handle); 539 if (err != DDI_SUCCESS) { 540 err = EINVAL; 541 goto allocfail_alloc_handle; 542 } 543 544 /* don't sleep here so we don't get stuck in contig alloc */ 545 err = ddi_dma_mem_alloc(mp->xm_dma_handle, mp->xm_size, 546 &mp->xm_device_attr, DDI_DMA_CONSISTENT, DDI_DMA_DONTWAIT, NULL, 547 &mp->xm_addr, &mp->xm_real_length, &mp->xm_mem_handle); 548 if (err != DDI_SUCCESS) { 549 err = EINVAL; 550 goto allocfail_alloc_mem; 551 } 552 553 err = ddi_dma_addr_bind_handle(mp->xm_dma_handle, NULL, mp->xm_addr, 554 mp->xm_size, DDI_DMA_RDWR | DDI_DMA_CONSISTENT, DDI_DMA_SLEEP, 555 NULL, &mp->xm_cookie, &mp->xm_cookie_count); 556 if (err != DDI_DMA_MAPPED) { 557 err = EFAULT; 558 goto allocfail_bind; 559 } 560 561 /* return sgl */ 562 for (i = 0; i < mp->xm_cookie_count; i++) { 563 if (ddi_model_convert_from(mode & FMODELS) == DDI_MODEL_ILP32) { 564 sgl32.mloc_addr = mp->xm_cookie.dmac_laddress; 565 sgl32.mloc_size = mp->xm_cookie.dmac_size; 566 err = ddi_copyout(&sgl32, &usgl32[i], 567 sizeof (xsvc_mloc_32), mode); 568 if (err != 0) { 569 err = EFAULT; 570 goto allocfail_copyout; 571 } 572 } else { 573 sgl.mloc_addr = mp->xm_cookie.dmac_laddress; 574 sgl.mloc_size = mp->xm_cookie.dmac_size; 575 err = ddi_copyout(&sgl, &usgl[i], sizeof (xsvc_mloc), 576 mode); 577 if (err != 0) { 578 err = EFAULT; 579 goto allocfail_copyout; 580 } 581 } 582 ddi_dma_nextcookie(mp->xm_dma_handle, &mp->xm_cookie); 583 } 584 585 /* set the last sgl entry to 0 to indicate cookie count */ 586 if (ddi_model_convert_from(mode & FMODELS) == DDI_MODEL_ILP32) { 587 sgl32.mloc_addr = 0; 588 sgl32.mloc_size = 0; 589 err = ddi_copyout(&sgl32, &usgl32[i], sizeof (xsvc_mloc_32), 590 mode); 591 if (err != 0) { 592 err = EFAULT; 593 goto allocfail_copyout; 594 } 595 } else { 596 sgl.mloc_addr = 0; 597 sgl.mloc_size = 0; 598 err = ddi_copyout(&sgl, &usgl[i], sizeof (xsvc_mloc), mode); 599 if (err != 0) { 600 err = EFAULT; 601 goto allocfail_copyout; 602 } 603 } 604 605 return (0); 606 607 allocfail_copyout: 608 (void) ddi_dma_unbind_handle(mp->xm_dma_handle); 609 allocfail_bind: 610 ddi_dma_mem_free(&mp->xm_mem_handle); 611 allocfail_alloc_mem: 612 ddi_dma_free_handle(&mp->xm_dma_handle); 613 allocfail_alloc_handle: 614 mp->xm_dma_handle = NULL; 615 xsvc_mem_free(state, mp); 616 617 mutex_enter(&state->xs_mutex); 618 state->xs_currently_alloced = state->xs_currently_alloced - size; 619 mutex_exit(&state->xs_mutex); 620 621 return (err); 622 } 623 624 /* 625 * xsvc_ioctl_flush_memory() 626 * 627 */ 628 static int 629 xsvc_ioctl_flush_memory(xsvc_state_t *state, void *arg, int mode) 630 { 631 xsvc_mem_req_32 params32; 632 xsvc_mem_req params; 633 xsvc_mem_t *mp; 634 uint64_t key; 635 int err; 636 637 638 if (ddi_model_convert_from(mode & FMODELS) == DDI_MODEL_ILP32) { 639 err = ddi_copyin(arg, ¶ms32, sizeof (xsvc_mem_req_32), 640 mode); 641 if (err != 0) { 642 return (EFAULT); 643 } 644 key = (uint64_t)params32.xsvc_mem_reqid; 645 } else { 646 err = ddi_copyin(arg, ¶ms, sizeof (xsvc_mem_req), mode); 647 if (err != 0) { 648 return (EFAULT); 649 } 650 key = (uint64_t)params.xsvc_mem_reqid; 651 } 652 653 /* find the memory */ 654 mp = xsvc_mem_lookup(state, key); 655 if (mp == NULL) { 656 return (EINVAL); 657 } 658 659 (void) ddi_dma_sync(mp->xm_dma_handle, 0, 0, DDI_DMA_SYNC_FORCPU); 660 661 return (0); 662 } 663 664 665 /* 666 * xsvc_ioctl_free_memory() 667 * 668 */ 669 static int 670 xsvc_ioctl_free_memory(xsvc_state_t *state, void *arg, int mode) 671 { 672 xsvc_mem_req_32 params32; 673 xsvc_mem_req params; 674 xsvc_mem_t *mp; 675 uint64_t key; 676 int err; 677 678 679 if (ddi_model_convert_from(mode & FMODELS) == DDI_MODEL_ILP32) { 680 err = ddi_copyin(arg, ¶ms32, sizeof (xsvc_mem_req_32), 681 mode); 682 if (err != 0) { 683 return (EFAULT); 684 } 685 key = (uint64_t)params32.xsvc_mem_reqid; 686 } else { 687 err = ddi_copyin(arg, ¶ms, sizeof (xsvc_mem_req), mode); 688 if (err != 0) { 689 return (EFAULT); 690 } 691 key = (uint64_t)params.xsvc_mem_reqid; 692 } 693 694 /* find the memory */ 695 mp = xsvc_mem_lookup(state, key); 696 if (mp == NULL) { 697 return (EINVAL); 698 } 699 700 xsvc_mem_free(state, mp); 701 702 return (0); 703 } 704 705 /* 706 * xsvc_mem_alloc() 707 * 708 */ 709 static int 710 xsvc_mem_alloc(xsvc_state_t *state, uint64_t key, xsvc_mem_t **mp) 711 { 712 xsvc_mem_t *mem; 713 714 mem = xsvc_mem_lookup(state, key); 715 if (mem != NULL) { 716 xsvc_mem_free(state, mem); 717 } 718 719 *mp = kmem_alloc(sizeof (xsvc_mem_t), KM_SLEEP); 720 (*mp)->xm_mnode.mn_home = *mp; 721 (*mp)->xm_mnode.mn_key = key; 722 723 mutex_enter(&state->xs_mlist.ml_mutex); 724 avl_add(&state->xs_mlist.ml_avl, &(*mp)->xm_mnode); 725 mutex_exit(&state->xs_mlist.ml_mutex); 726 727 return (0); 728 } 729 730 /* 731 * xsvc_mem_free() 732 * 733 */ 734 static void 735 xsvc_mem_free(xsvc_state_t *state, xsvc_mem_t *mp) 736 { 737 if (mp->xm_dma_handle != NULL) { 738 (void) ddi_dma_unbind_handle(mp->xm_dma_handle); 739 ddi_dma_mem_free(&mp->xm_mem_handle); 740 ddi_dma_free_handle(&mp->xm_dma_handle); 741 742 mutex_enter(&state->xs_mutex); 743 state->xs_currently_alloced = state->xs_currently_alloced - 744 mp->xm_size; 745 mutex_exit(&state->xs_mutex); 746 } 747 748 mutex_enter(&state->xs_mlist.ml_mutex); 749 avl_remove(&state->xs_mlist.ml_avl, &mp->xm_mnode); 750 mutex_exit(&state->xs_mlist.ml_mutex); 751 752 kmem_free(mp, sizeof (*mp)); 753 } 754 755 /* 756 * xsvc_mem_lookup() 757 * 758 */ 759 static xsvc_mem_t * 760 xsvc_mem_lookup(xsvc_state_t *state, uint64_t key) 761 { 762 xsvc_mnode_t mnode; 763 xsvc_mnode_t *mnp; 764 avl_index_t where; 765 xsvc_mem_t *mp; 766 767 mnode.mn_key = key; 768 mutex_enter(&state->xs_mlist.ml_mutex); 769 mnp = avl_find(&state->xs_mlist.ml_avl, &mnode, &where); 770 mutex_exit(&state->xs_mlist.ml_mutex); 771 772 if (mnp != NULL) { 773 mp = mnp->mn_home; 774 } else { 775 mp = NULL; 776 } 777 778 return (mp); 779 } 780 781 /* 782 * xsvc_mnode_key_compare() 783 * 784 */ 785 static int 786 xsvc_mnode_key_compare(const void *q, const void *e) 787 { 788 xsvc_mnode_t *n1; 789 xsvc_mnode_t *n2; 790 791 n1 = (xsvc_mnode_t *)q; 792 n2 = (xsvc_mnode_t *)e; 793 794 if (n1->mn_key < n2->mn_key) { 795 return (-1); 796 } else if (n1->mn_key > n2->mn_key) { 797 return (1); 798 } else { 799 return (0); 800 } 801 } 802 803 /* 804 * xsvc_devmap() 805 * 806 */ 807 /*ARGSUSED*/ 808 static int 809 xsvc_devmap(dev_t dev, devmap_cookie_t dhp, offset_t off, size_t len, 810 size_t *maplen, uint_t model) 811 { 812 ddi_umem_cookie_t cookie; 813 xsvc_state_t *state; 814 offset_t off_align; 815 size_t npages; 816 caddr_t kvai; 817 size_t psize; 818 int instance; 819 caddr_t kva; 820 pfn_t pfn; 821 int err; 822 int i; 823 824 825 instance = getminor(dev); 826 state = ddi_get_soft_state(xsvc_statep, instance); 827 if (state == NULL) { 828 return (ENXIO); 829 } 830 831 /* 832 * On 64-bit kernels, if we have a 32-bit application doing a mmap(), 833 * smmap32 will sign extend the offset. We need to undo that since 834 * we are passed a physical address in off, not a offset. 835 */ 836 #if defined(__amd64) 837 if (((model & DDI_MODEL_MASK) == DDI_MODEL_ILP32) && 838 ((off & ~0xFFFFFFFFll) == ~0xFFFFFFFFll)) { 839 off = off & 0xFFFFFFFF; 840 } 841 #endif 842 843 #ifdef __xpv 844 /* 845 * we won't allow guest OSes to devmap mfn/pfns. Maybe we'll relax 846 * this some later when there is a good reason. 847 */ 848 if (!DOMAIN_IS_INITDOMAIN(xen_info)) { 849 return (-1); 850 } 851 852 /* we will always treat this as a foreign MFN */ 853 pfn = xen_assign_pfn(btop(off)); 854 #else 855 pfn = btop(off); 856 #endif 857 /* always work with whole pages */ 858 859 off_align = P2ALIGN(off, PAGESIZE); 860 psize = P2ROUNDUP(off + len, PAGESIZE) - off_align; 861 862 /* 863 * if this is memory we're trying to map into user space, we first 864 * need to map the PFNs into KVA, then build up a umem cookie, and 865 * finally do a umem_setup to map it in. 866 */ 867 if (pf_is_memory(pfn)) { 868 npages = btop(psize); 869 870 kva = vmem_alloc(heap_arena, psize, VM_SLEEP); 871 if (kva == NULL) { 872 return (-1); 873 } 874 875 kvai = kva; 876 for (i = 0; i < npages; i++) { 877 hat_devload(kas.a_hat, kvai, PAGESIZE, pfn, 878 PROT_READ | PROT_WRITE, HAT_LOAD_LOCK); 879 pfn++; 880 kvai = (caddr_t)((uintptr_t)kvai + PAGESIZE); 881 } 882 883 err = xsvc_umem_cookie_alloc(kva, psize, KM_SLEEP, &cookie); 884 if (err != 0) { 885 goto devmapfail_cookie_alloc; 886 } 887 888 if ((err = devmap_umem_setup(dhp, state->xs_dip, &xsvc_callbk, 889 cookie, 0, psize, PROT_ALL, 0, &xsvc_device_attr)) < 0) { 890 goto devmapfail_umem_setup; 891 } 892 *maplen = psize; 893 894 /* 895 * If this is not memory (or a foreign MFN in i86xpv), go through 896 * devmem_setup. 897 */ 898 } else { 899 if ((err = devmap_devmem_setup(dhp, state->xs_dip, NULL, 0, 900 off_align, psize, PROT_ALL, 0, &xsvc_device_attr)) < 0) { 901 return (err); 902 } 903 *maplen = psize; 904 } 905 906 return (0); 907 908 devmapfail_umem_setup: 909 xsvc_umem_cookie_free(&cookie); 910 911 devmapfail_cookie_alloc: 912 kvai = kva; 913 for (i = 0; i < npages; i++) { 914 hat_unload(kas.a_hat, kvai, PAGESIZE, 915 HAT_UNLOAD_UNLOCK); 916 kvai = (caddr_t)((uintptr_t)kvai + PAGESIZE); 917 } 918 vmem_free(heap_arena, kva, psize); 919 920 return (err); 921 } 922 923 /* 924 * xsvc_umem_cookie_alloc() 925 * 926 * allocate a umem cookie to be used in devmap_umem_setup using KVA already 927 * allocated. 928 */ 929 int 930 xsvc_umem_cookie_alloc(caddr_t kva, size_t size, int flags, 931 ddi_umem_cookie_t *cookiep) 932 { 933 struct ddi_umem_cookie *umem_cookiep; 934 935 umem_cookiep = kmem_zalloc(sizeof (struct ddi_umem_cookie), flags); 936 if (umem_cookiep == NULL) { 937 *cookiep = NULL; 938 return (-1); 939 } 940 941 umem_cookiep->cvaddr = kva; 942 umem_cookiep->type = KMEM_NON_PAGEABLE; 943 umem_cookiep->size = size; 944 *cookiep = (ddi_umem_cookie_t *)umem_cookiep; 945 946 return (0); 947 } 948 949 /* 950 * xsvc_umem_cookie_free() 951 * 952 */ 953 static void 954 xsvc_umem_cookie_free(ddi_umem_cookie_t *cookiep) 955 { 956 kmem_free(*cookiep, sizeof (struct ddi_umem_cookie)); 957 *cookiep = NULL; 958 } 959 960 961 /* 962 * xsvc_devmap_map() 963 * 964 */ 965 /*ARGSUSED*/ 966 static int 967 xsvc_devmap_map(devmap_cookie_t dhc, dev_t dev, uint_t flags, offset_t off, 968 size_t len, void **pvtp) 969 { 970 struct ddi_umem_cookie *cp; 971 devmap_handle_t *dhp; 972 xsvc_state_t *state; 973 int instance; 974 975 976 instance = getminor(dev); 977 state = ddi_get_soft_state(xsvc_statep, instance); 978 if (state == NULL) { 979 return (ENXIO); 980 } 981 982 dhp = (devmap_handle_t *)dhc; 983 /* This driver only supports MAP_SHARED, not MAP_PRIVATE */ 984 if (flags & MAP_PRIVATE) { 985 cmn_err(CE_WARN, "!xsvc driver doesn't support MAP_PRIVATE"); 986 return (EINVAL); 987 } 988 989 cp = (struct ddi_umem_cookie *)dhp->dh_cookie; 990 cp->cook_refcnt = 1; 991 992 *pvtp = state; 993 return (0); 994 } 995 996 997 /* 998 * xsvc_devmap_dup() 999 * 1000 * keep a reference count for forks so we don't unmap if we have multiple 1001 * mappings. 1002 */ 1003 /*ARGSUSED*/ 1004 static int 1005 xsvc_devmap_dup(devmap_cookie_t dhc, void *pvtp, devmap_cookie_t new_dhp, 1006 void **new_pvtp) 1007 { 1008 struct ddi_umem_cookie *cp; 1009 devmap_handle_t *dhp; 1010 xsvc_state_t *state; 1011 1012 1013 state = (xsvc_state_t *)pvtp; 1014 dhp = (devmap_handle_t *)dhc; 1015 1016 mutex_enter(&state->xs_cookie_mutex); 1017 cp = (struct ddi_umem_cookie *)dhp->dh_cookie; 1018 if (cp == NULL) { 1019 mutex_exit(&state->xs_cookie_mutex); 1020 return (ENOMEM); 1021 } 1022 1023 cp->cook_refcnt++; 1024 mutex_exit(&state->xs_cookie_mutex); 1025 1026 *new_pvtp = state; 1027 return (0); 1028 } 1029 1030 1031 /* 1032 * xsvc_devmap_unmap() 1033 * 1034 * This routine is only call if we were mapping in memory in xsvc_devmap(). 1035 * i.e. we only pass in xsvc_callbk to devmap_umem_setup if pf_is_memory() 1036 * was true. It would have been nice if devmap_callback_ctl had an args param. 1037 * We wouldn't have had to look into the devmap_handle and into the umem 1038 * cookie. 1039 */ 1040 /*ARGSUSED*/ 1041 static void 1042 xsvc_devmap_unmap(devmap_cookie_t dhc, void *pvtp, offset_t off, size_t len, 1043 devmap_cookie_t new_dhp1, void **new_pvtp1, devmap_cookie_t new_dhp2, 1044 void **new_pvtp2) 1045 { 1046 struct ddi_umem_cookie *ncp; 1047 struct ddi_umem_cookie *cp; 1048 devmap_handle_t *ndhp; 1049 devmap_handle_t *dhp; 1050 xsvc_state_t *state; 1051 size_t npages; 1052 caddr_t kvai; 1053 caddr_t kva; 1054 size_t size; 1055 int i; 1056 1057 1058 state = (xsvc_state_t *)pvtp; 1059 mutex_enter(&state->xs_cookie_mutex); 1060 1061 /* peek into the umem cookie to figure out what we need to free up */ 1062 dhp = (devmap_handle_t *)dhc; 1063 cp = (struct ddi_umem_cookie *)dhp->dh_cookie; 1064 ASSERT(cp != NULL); 1065 1066 if (new_dhp1 != NULL) { 1067 ndhp = (devmap_handle_t *)new_dhp1; 1068 ncp = (struct ddi_umem_cookie *)ndhp->dh_cookie; 1069 ncp->cook_refcnt++; 1070 *new_pvtp1 = state; 1071 } 1072 if (new_dhp2 != NULL) { 1073 ndhp = (devmap_handle_t *)new_dhp2; 1074 ncp = (struct ddi_umem_cookie *)ndhp->dh_cookie; 1075 ncp->cook_refcnt++; 1076 *new_pvtp2 = state; 1077 } 1078 1079 cp->cook_refcnt--; 1080 if (cp->cook_refcnt == 0) { 1081 kva = cp->cvaddr; 1082 size = cp->size; 1083 1084 /* 1085 * free up the umem cookie, then unmap all the pages what we 1086 * mapped in during devmap, then free up the kva space. 1087 */ 1088 npages = btop(size); 1089 xsvc_umem_cookie_free(&dhp->dh_cookie); 1090 kvai = kva; 1091 for (i = 0; i < npages; i++) { 1092 hat_unload(kas.a_hat, kvai, PAGESIZE, 1093 HAT_UNLOAD_UNLOCK); 1094 kvai = (caddr_t)((uintptr_t)kvai + PAGESIZE); 1095 } 1096 vmem_free(heap_arena, kva, size); 1097 } 1098 1099 mutex_exit(&state->xs_cookie_mutex); 1100 }