1 /*
   2  * CDDL HEADER START
   3  *
   4  * The contents of this file are subject to the terms of the
   5  * Common Development and Distribution License (the "License").
   6  * You may not use this file except in compliance with the License.
   7  *
   8  * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
   9  * or http://www.opensolaris.org/os/licensing.
  10  * See the License for the specific language governing permissions
  11  * and limitations under the License.
  12  *
  13  * When distributing Covered Code, include this CDDL HEADER in each
  14  * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
  15  * If applicable, add the following below this CDDL HEADER, with the
  16  * fields enclosed by brackets "[]" replaced with your own identifying
  17  * information: Portions Copyright [yyyy] [name of copyright owner]
  18  *
  19  * CDDL HEADER END
  20  */
  21 /*
  22  * Copyright (c) 2004, 2010, Oracle and/or its affiliates. All rights reserved.
  23  */
  24 
  25 /*
  26  * Copyright 2012 Joyent, Inc.  All rights reserved.
  27  */
  28 
  29 /* Copyright (c) 1984, 1986, 1987, 1988, 1989 AT&T */
  30 /*      All Rights Reserved   */
  31 
  32 /*
  33  * Portions of this source code were derived from Berkeley 4.3 BSD
  34  * under license from the Regents of the University of California.
  35  */
  36 
  37 
  38 /*
  39  * This file contains common functions to access and manage the page lists.
  40  * Many of these routines originated from platform dependent modules
  41  * (sun4/vm/vm_dep.c, i86pc/vm/vm_machdep.c) and modified to function in
  42  * a platform independent manner.
  43  *
  44  * vm/vm_dep.h provides for platform specific support.
  45  */
  46 
  47 #include <sys/types.h>
  48 #include <sys/debug.h>
  49 #include <sys/cmn_err.h>
  50 #include <sys/systm.h>
  51 #include <sys/atomic.h>
  52 #include <sys/sysmacros.h>
  53 #include <vm/as.h>
  54 #include <vm/page.h>
  55 #include <vm/seg_kmem.h>
  56 #include <vm/seg_vn.h>
  57 #include <sys/vmsystm.h>
  58 #include <sys/memnode.h>
  59 #include <vm/vm_dep.h>
  60 #include <sys/lgrp.h>
  61 #include <sys/mem_config.h>
  62 #include <sys/callb.h>
  63 #include <sys/mem_cage.h>
  64 #include <sys/sdt.h>
  65 #include <sys/dumphdr.h>
  66 #include <sys/swap.h>
  67 
  68 extern uint_t   vac_colors;
  69 
  70 #define MAX_PRAGMA_ALIGN        128
  71 
  72 /* vm_cpu_data0 for the boot cpu before kmem is initialized */
  73 
  74 #if L2CACHE_ALIGN_MAX <= MAX_PRAGMA_ALIGN
  75 #pragma align   L2CACHE_ALIGN_MAX(vm_cpu_data0)
  76 #else
  77 #pragma align   MAX_PRAGMA_ALIGN(vm_cpu_data0)
  78 #endif
  79 char            vm_cpu_data0[VM_CPU_DATA_PADSIZE];
  80 
  81 /*
  82  * number of page colors equivalent to reqested color in page_get routines.
  83  * If set, keeps large pages intact longer and keeps MPO allocation
  84  * from the local mnode in favor of acquiring the 'correct' page color from
  85  * a demoted large page or from a remote mnode.
  86  */
  87 uint_t  colorequiv;
  88 
  89 /*
  90  * color equivalency mask for each page size.
  91  * Mask is computed based on cpu L2$ way sizes and colorequiv global.
  92  * High 4 bits determine the number of high order bits of the color to ignore.
  93  * Low 4 bits determines number of low order bits of color to ignore (it's only
  94  * relevant for hashed index based page coloring).
  95  */
  96 uchar_t colorequivszc[MMU_PAGE_SIZES];
  97 
  98 /*
  99  * if set, specifies the percentage of large pages that are free from within
 100  * a large page region before attempting to lock those pages for
 101  * page_get_contig_pages processing.
 102  *
 103  * Should be turned on when kpr is available when page_trylock_contig_pages
 104  * can be more selective.
 105  */
 106 
 107 int     ptcpthreshold;
 108 
 109 /*
 110  * Limit page get contig page search based on failure cnts in pgcpfailcnt[].
 111  * Enabled by default via pgcplimitsearch.
 112  *
 113  * pgcpfailcnt[] is bounded by PGCPFAILMAX (>= 1/2 of installed
 114  * memory). When reached, pgcpfailcnt[] is reset to 1/2 of this upper
 115  * bound. This upper bound range guarantees:
 116  *    - all large page 'slots' will be searched over time
 117  *    - the minimum (1) large page candidates considered on each pgcp call
 118  *    - count doesn't wrap around to 0
 119  */
 120 pgcnt_t pgcpfailcnt[MMU_PAGE_SIZES];
 121 int     pgcplimitsearch = 1;
 122 
 123 #define PGCPFAILMAX             (1 << (highbit(physinstalled) - 1))
 124 #define SETPGCPFAILCNT(szc)                                             \
 125         if (++pgcpfailcnt[szc] >= PGCPFAILMAX)                               \
 126                 pgcpfailcnt[szc] = PGCPFAILMAX / 2;
 127 
 128 #ifdef VM_STATS
 129 struct vmm_vmstats_str  vmm_vmstats;
 130 
 131 #endif /* VM_STATS */
 132 
 133 #if defined(__sparc)
 134 #define LPGCREATE       0
 135 #else
 136 /* enable page_get_contig_pages */
 137 #define LPGCREATE       1
 138 #endif
 139 
 140 int pg_contig_disable;
 141 int pg_lpgcreate_nocage = LPGCREATE;
 142 
 143 /*
 144  * page_freelist_split pfn flag to signify no lo or hi pfn requirement.
 145  */
 146 #define PFNNULL         0
 147 
 148 /* Flags involved in promotion and demotion routines */
 149 #define PC_FREE         0x1     /* put page on freelist */
 150 #define PC_ALLOC        0x2     /* return page for allocation */
 151 
 152 /*
 153  * Flag for page_demote to be used with PC_FREE to denote that we don't care
 154  * what the color is as the color parameter to the function is ignored.
 155  */
 156 #define PC_NO_COLOR     (-1)
 157 
 158 /* mtype value for page_promote to use when mtype does not matter */
 159 #define PC_MTYPE_ANY    (-1)
 160 
 161 /*
 162  * page counters candidates info
 163  * See page_ctrs_cands comment below for more details.
 164  * fields are as follows:
 165  *      pcc_pages_free:         # pages which freelist coalesce can create
 166  *      pcc_color_free:         pointer to page free counts per color
 167  */
 168 typedef struct pcc_info {
 169         pgcnt_t pcc_pages_free;
 170         pgcnt_t *pcc_color_free;
 171         uint_t  pad[12];
 172 } pcc_info_t;
 173 
 174 /*
 175  * On big machines it can take a long time to check page_counters
 176  * arrays. page_ctrs_cands is a summary array whose elements are a dynamically
 177  * updated sum of all elements of the corresponding page_counters arrays.
 178  * page_freelist_coalesce() searches page_counters only if an appropriate
 179  * element of page_ctrs_cands array is greater than 0.
 180  *
 181  * page_ctrs_cands is indexed by mutex (i), region (r), mnode (m), mrange (g)
 182  */
 183 pcc_info_t **page_ctrs_cands[NPC_MUTEX][MMU_PAGE_SIZES];
 184 
 185 /*
 186  * Return in val the total number of free pages which can be created
 187  * for the given mnode (m), mrange (g), and region size (r)
 188  */
 189 #define PGCTRS_CANDS_GETVALUE(m, g, r, val) {                           \
 190         int i;                                                          \
 191         val = 0;                                                        \
 192         for (i = 0; i < NPC_MUTEX; i++) {                            \
 193             val += page_ctrs_cands[i][(r)][(m)][(g)].pcc_pages_free;    \
 194         }                                                               \
 195 }
 196 
 197 /*
 198  * Return in val the total number of free pages which can be created
 199  * for the given mnode (m), mrange (g), region size (r), and color (c)
 200  */
 201 #define PGCTRS_CANDS_GETVALUECOLOR(m, g, r, c, val) {                   \
 202         int i;                                                          \
 203         val = 0;                                                        \
 204         ASSERT((c) < PAGE_GET_PAGECOLORS(r));                                \
 205         for (i = 0; i < NPC_MUTEX; i++) {                            \
 206             val +=                                                      \
 207                 page_ctrs_cands[i][(r)][(m)][(g)].pcc_color_free[(c)];  \
 208         }                                                               \
 209 }
 210 
 211 /*
 212  * We can only allow a single thread to update a counter within the physical
 213  * range of the largest supported page size. That is the finest granularity
 214  * possible since the counter values are dependent on each other
 215  * as you move accross region sizes. PP_CTR_LOCK_INDX is used to determine the
 216  * ctr_mutex lock index for a particular physical range.
 217  */
 218 static kmutex_t *ctr_mutex[NPC_MUTEX];
 219 
 220 #define PP_CTR_LOCK_INDX(pp)                                            \
 221         (((pp)->p_pagenum >>                                           \
 222             (PAGE_BSZS_SHIFT(mmu_page_sizes - 1))) & (NPC_MUTEX - 1))
 223 
 224 #define INVALID_COLOR 0xffffffff
 225 #define INVALID_MASK  0xffffffff
 226 
 227 /*
 228  * Local functions prototypes.
 229  */
 230 
 231 void page_ctr_add(int, int, page_t *, int);
 232 void page_ctr_add_internal(int, int, page_t *, int);
 233 void page_ctr_sub(int, int, page_t *, int);
 234 void page_ctr_sub_internal(int, int, page_t *, int);
 235 void page_freelist_lock(int);
 236 void page_freelist_unlock(int);
 237 page_t *page_promote(int, pfn_t, uchar_t, int, int);
 238 page_t *page_demote(int, pfn_t, pfn_t, uchar_t, uchar_t, int, int);
 239 page_t *page_freelist_split(uchar_t,
 240     uint_t, int, int, pfn_t, pfn_t, page_list_walker_t *);
 241 page_t *page_get_mnode_cachelist(uint_t, uint_t, int, int);
 242 static int page_trylock_cons(page_t *pp, se_t se);
 243 
 244 /*
 245  * The page_counters array below is used to keep track of free contiguous
 246  * physical memory.  A hw_page_map_t will be allocated per mnode per szc.
 247  * This contains an array of counters, the size of the array, a shift value
 248  * used to convert a pagenum into a counter array index or vice versa, as
 249  * well as a cache of the last successful index to be promoted to a larger
 250  * page size.  As an optimization, we keep track of the last successful index
 251  * to be promoted per page color for the given size region, and this is
 252  * allocated dynamically based upon the number of colors for a given
 253  * region size.
 254  *
 255  * Conceptually, the page counters are represented as:
 256  *
 257  *      page_counters[region_size][mnode]
 258  *
 259  *      region_size:    size code of a candidate larger page made up
 260  *                      of contiguous free smaller pages.
 261  *
 262  *      page_counters[region_size][mnode].hpm_counters[index]:
 263  *              represents how many (region_size - 1) pages either
 264  *              exist or can be created within the given index range.
 265  *
 266  * Let's look at a sparc example:
 267  *      If we want to create a free 512k page, we look at region_size 2
 268  *      for the mnode we want.  We calculate the index and look at a specific
 269  *      hpm_counters location.  If we see 8 (FULL_REGION_CNT on sparc) at
 270  *      this location, it means that 8 64k pages either exist or can be created
 271  *      from 8K pages in order to make a single free 512k page at the given
 272  *      index.  Note that when a region is full, it will contribute to the
 273  *      counts in the region above it.  Thus we will not know what page
 274  *      size the free pages will be which can be promoted to this new free
 275  *      page unless we look at all regions below the current region.
 276  */
 277 
 278 /*
 279  * Note: hpmctr_t is defined in platform vm_dep.h
 280  * hw_page_map_t contains all the information needed for the page_counters
 281  * logic. The fields are as follows:
 282  *
 283  *      hpm_counters:   dynamically allocated array to hold counter data
 284  *      hpm_entries:    entries in hpm_counters
 285  *      hpm_shift:      shift for pnum/array index conv
 286  *      hpm_base:       PFN mapped to counter index 0
 287  *      hpm_color_current:      last index in counter array for this color at
 288  *                              which we successfully created a large page
 289  */
 290 typedef struct hw_page_map {
 291         hpmctr_t        *hpm_counters;
 292         size_t          hpm_entries;
 293         int             hpm_shift;
 294         pfn_t           hpm_base;
 295         size_t          *hpm_color_current[MAX_MNODE_MRANGES];
 296 #if defined(__sparc)
 297         uint_t          pad[4];
 298 #endif
 299 } hw_page_map_t;
 300 
 301 /*
 302  * Element zero is not used, but is allocated for convenience.
 303  */
 304 static hw_page_map_t *page_counters[MMU_PAGE_SIZES];
 305 
 306 /*
 307  * Cached value of MNODE_RANGE_CNT(mnode).
 308  * This is a function call in x86.
 309  */
 310 static int mnode_nranges[MAX_MEM_NODES];
 311 static int mnode_maxmrange[MAX_MEM_NODES];
 312 
 313 /*
 314  * The following macros are convenient ways to get access to the individual
 315  * elements of the page_counters arrays.  They can be used on both
 316  * the left side and right side of equations.
 317  */
 318 #define PAGE_COUNTERS(mnode, rg_szc, idx)                       \
 319         (page_counters[(rg_szc)][(mnode)].hpm_counters[(idx)])
 320 
 321 #define PAGE_COUNTERS_COUNTERS(mnode, rg_szc)                   \
 322         (page_counters[(rg_szc)][(mnode)].hpm_counters)
 323 
 324 #define PAGE_COUNTERS_SHIFT(mnode, rg_szc)                      \
 325         (page_counters[(rg_szc)][(mnode)].hpm_shift)
 326 
 327 #define PAGE_COUNTERS_ENTRIES(mnode, rg_szc)                    \
 328         (page_counters[(rg_szc)][(mnode)].hpm_entries)
 329 
 330 #define PAGE_COUNTERS_BASE(mnode, rg_szc)                       \
 331         (page_counters[(rg_szc)][(mnode)].hpm_base)
 332 
 333 #define PAGE_COUNTERS_CURRENT_COLOR_ARRAY(mnode, rg_szc, g)             \
 334         (page_counters[(rg_szc)][(mnode)].hpm_color_current[(g)])
 335 
 336 #define PAGE_COUNTERS_CURRENT_COLOR(mnode, rg_szc, color, mrange)       \
 337         (page_counters[(rg_szc)][(mnode)].                              \
 338         hpm_color_current[(mrange)][(color)])
 339 
 340 #define PNUM_TO_IDX(mnode, rg_szc, pnum)                        \
 341         (((pnum) - PAGE_COUNTERS_BASE((mnode), (rg_szc))) >>      \
 342                 PAGE_COUNTERS_SHIFT((mnode), (rg_szc)))
 343 
 344 #define IDX_TO_PNUM(mnode, rg_szc, index)                       \
 345         (PAGE_COUNTERS_BASE((mnode), (rg_szc)) +                \
 346                 ((index) << PAGE_COUNTERS_SHIFT((mnode), (rg_szc))))
 347 
 348 /*
 349  * Protects the hpm_counters and hpm_color_current memory from changing while
 350  * looking at page counters information.
 351  * Grab the write lock to modify what these fields point at.
 352  * Grab the read lock to prevent any pointers from changing.
 353  * The write lock can not be held during memory allocation due to a possible
 354  * recursion deadlock with trying to grab the read lock while the
 355  * write lock is already held.
 356  */
 357 krwlock_t page_ctrs_rwlock[MAX_MEM_NODES];
 358 
 359 
 360 /*
 361  * initialize cpu_vm_data to point at cache aligned vm_cpu_data_t.
 362  */
 363 void
 364 cpu_vm_data_init(struct cpu *cp)
 365 {
 366         if (cp == CPU0) {
 367                 cp->cpu_vm_data = (void *)&vm_cpu_data0;
 368         } else {
 369                 void    *kmptr;
 370                 int     align;
 371                 size_t  sz;
 372 
 373                 align = (L2CACHE_ALIGN) ? L2CACHE_ALIGN : L2CACHE_ALIGN_MAX;
 374                 sz = P2ROUNDUP(sizeof (vm_cpu_data_t), align) + align;
 375                 kmptr = kmem_zalloc(sz, KM_SLEEP);
 376                 cp->cpu_vm_data = (void *) P2ROUNDUP((uintptr_t)kmptr, align);
 377                 ((vm_cpu_data_t *)cp->cpu_vm_data)->vc_kmptr = kmptr;
 378                 ((vm_cpu_data_t *)cp->cpu_vm_data)->vc_kmsize = sz;
 379         }
 380 }
 381 
 382 /*
 383  * free cpu_vm_data
 384  */
 385 void
 386 cpu_vm_data_destroy(struct cpu *cp)
 387 {
 388         if (cp->cpu_seqid && cp->cpu_vm_data) {
 389                 ASSERT(cp != CPU0);
 390                 kmem_free(((vm_cpu_data_t *)cp->cpu_vm_data)->vc_kmptr,
 391                     ((vm_cpu_data_t *)cp->cpu_vm_data)->vc_kmsize);
 392         }
 393         cp->cpu_vm_data = NULL;
 394 }
 395 
 396 
 397 /*
 398  * page size to page size code
 399  */
 400 int
 401 page_szc(size_t pagesize)
 402 {
 403         int     i = 0;
 404 
 405         while (hw_page_array[i].hp_size) {
 406                 if (pagesize == hw_page_array[i].hp_size)
 407                         return (i);
 408                 i++;
 409         }
 410         return (-1);
 411 }
 412 
 413 /*
 414  * page size to page size code with the restriction that it be a supported
 415  * user page size.  If it's not a supported user page size, -1 will be returned.
 416  */
 417 int
 418 page_szc_user_filtered(size_t pagesize)
 419 {
 420         int szc = page_szc(pagesize);
 421         if ((szc != -1) && (SZC_2_USERSZC(szc) != -1)) {
 422                 return (szc);
 423         }
 424         return (-1);
 425 }
 426 
 427 /*
 428  * Return how many page sizes are available for the user to use.  This is
 429  * what the hardware supports and not based upon how the OS implements the
 430  * support of different page sizes.
 431  *
 432  * If legacy is non-zero, return the number of pagesizes available to legacy
 433  * applications. The number of legacy page sizes might be less than the
 434  * exported user page sizes. This is to prevent legacy applications that
 435  * use the largest page size returned from getpagesizes(3c) from inadvertantly
 436  * using the 'new' large pagesizes.
 437  */
 438 uint_t
 439 page_num_user_pagesizes(int legacy)
 440 {
 441         if (legacy)
 442                 return (mmu_legacy_page_sizes);
 443         return (mmu_exported_page_sizes);
 444 }
 445 
 446 uint_t
 447 page_num_pagesizes(void)
 448 {
 449         return (mmu_page_sizes);
 450 }
 451 
 452 /*
 453  * returns the count of the number of base pagesize pages associated with szc
 454  */
 455 pgcnt_t
 456 page_get_pagecnt(uint_t szc)
 457 {
 458         if (szc >= mmu_page_sizes)
 459                 panic("page_get_pagecnt: out of range %d", szc);
 460         return (hw_page_array[szc].hp_pgcnt);
 461 }
 462 
 463 size_t
 464 page_get_pagesize(uint_t szc)
 465 {
 466         if (szc >= mmu_page_sizes)
 467                 panic("page_get_pagesize: out of range %d", szc);
 468         return (hw_page_array[szc].hp_size);
 469 }
 470 
 471 /*
 472  * Return the size of a page based upon the index passed in.  An index of
 473  * zero refers to the smallest page size in the system, and as index increases
 474  * it refers to the next larger supported page size in the system.
 475  * Note that szc and userszc may not be the same due to unsupported szc's on
 476  * some systems.
 477  */
 478 size_t
 479 page_get_user_pagesize(uint_t userszc)
 480 {
 481         uint_t szc = USERSZC_2_SZC(userszc);
 482 
 483         if (szc >= mmu_page_sizes)
 484                 panic("page_get_user_pagesize: out of range %d", szc);
 485         return (hw_page_array[szc].hp_size);
 486 }
 487 
 488 uint_t
 489 page_get_shift(uint_t szc)
 490 {
 491         if (szc >= mmu_page_sizes)
 492                 panic("page_get_shift: out of range %d", szc);
 493         return (PAGE_GET_SHIFT(szc));
 494 }
 495 
 496 uint_t
 497 page_get_pagecolors(uint_t szc)
 498 {
 499         if (szc >= mmu_page_sizes)
 500                 panic("page_get_pagecolors: out of range %d", szc);
 501         return (PAGE_GET_PAGECOLORS(szc));
 502 }
 503 
 504 /*
 505  * this assigns the desired equivalent color after a split
 506  */
 507 uint_t
 508 page_correct_color(uchar_t szc, uchar_t nszc, uint_t color,
 509     uint_t ncolor, uint_t ceq_mask)
 510 {
 511         ASSERT(nszc > szc);
 512         ASSERT(szc < mmu_page_sizes);
 513         ASSERT(color < PAGE_GET_PAGECOLORS(szc));
 514         ASSERT(ncolor < PAGE_GET_PAGECOLORS(nszc));
 515 
 516         color &= ceq_mask;
 517         ncolor = PAGE_CONVERT_COLOR(ncolor, szc, nszc);
 518         return (color | (ncolor & ~ceq_mask));
 519 }
 520 
 521 /*
 522  * The interleaved_mnodes flag is set when mnodes overlap in
 523  * the physbase..physmax range, but have disjoint slices.
 524  * In this case hpm_counters is shared by all mnodes.
 525  * This flag is set dynamically by the platform.
 526  */
 527 int interleaved_mnodes = 0;
 528 
 529 /*
 530  * Called by startup().
 531  * Size up the per page size free list counters based on physmax
 532  * of each node and max_mem_nodes.
 533  *
 534  * If interleaved_mnodes is set we need to find the first mnode that
 535  * exists. hpm_counters for the first mnode will then be shared by
 536  * all other mnodes. If interleaved_mnodes is not set, just set
 537  * first=mnode each time. That means there will be no sharing.
 538  */
 539 size_t
 540 page_ctrs_sz(void)
 541 {
 542         int     r;              /* region size */
 543         int     mnode;
 544         int     firstmn;        /* first mnode that exists */
 545         int     nranges;
 546         pfn_t   physbase;
 547         pfn_t   physmax;
 548         uint_t  ctrs_sz = 0;
 549         int     i;
 550         pgcnt_t colors_per_szc[MMU_PAGE_SIZES];
 551 
 552         /*
 553          * We need to determine how many page colors there are for each
 554          * page size in order to allocate memory for any color specific
 555          * arrays.
 556          */
 557         for (i = 0; i < mmu_page_sizes; i++) {
 558                 colors_per_szc[i] = PAGE_GET_PAGECOLORS(i);
 559         }
 560 
 561         for (firstmn = -1, mnode = 0; mnode < max_mem_nodes; mnode++) {
 562 
 563                 pgcnt_t r_pgcnt;
 564                 pfn_t   r_base;
 565                 pgcnt_t r_align;
 566 
 567                 if (mem_node_config[mnode].exists == 0)
 568                         continue;
 569 
 570                 HPM_COUNTERS_LIMITS(mnode, physbase, physmax, firstmn);
 571                 nranges = MNODE_RANGE_CNT(mnode);
 572                 mnode_nranges[mnode] = nranges;
 573                 mnode_maxmrange[mnode] = MNODE_MAX_MRANGE(mnode);
 574 
 575                 /*
 576                  * determine size needed for page counter arrays with
 577                  * base aligned to large page size.
 578                  */
 579                 for (r = 1; r < mmu_page_sizes; r++) {
 580                         /* add in space for hpm_color_current */
 581                         ctrs_sz += sizeof (size_t) *
 582                             colors_per_szc[r] * nranges;
 583 
 584                         if (firstmn != mnode)
 585                                 continue;
 586 
 587                         /* add in space for hpm_counters */
 588                         r_align = page_get_pagecnt(r);
 589                         r_base = physbase;
 590                         r_base &= ~(r_align - 1);
 591                         r_pgcnt = howmany(physmax - r_base + 1, r_align);
 592 
 593                         /*
 594                          * Round up to always allocate on pointer sized
 595                          * boundaries.
 596                          */
 597                         ctrs_sz += P2ROUNDUP((r_pgcnt * sizeof (hpmctr_t)),
 598                             sizeof (hpmctr_t *));
 599                 }
 600         }
 601 
 602         for (r = 1; r < mmu_page_sizes; r++) {
 603                 ctrs_sz += (max_mem_nodes * sizeof (hw_page_map_t));
 604         }
 605 
 606         /* add in space for page_ctrs_cands and pcc_color_free */
 607         ctrs_sz += sizeof (pcc_info_t *) * max_mem_nodes *
 608             mmu_page_sizes * NPC_MUTEX;
 609 
 610         for (mnode = 0; mnode < max_mem_nodes; mnode++) {
 611 
 612                 if (mem_node_config[mnode].exists == 0)
 613                         continue;
 614 
 615                 nranges = mnode_nranges[mnode];
 616                 ctrs_sz += sizeof (pcc_info_t) * nranges *
 617                     mmu_page_sizes * NPC_MUTEX;
 618                 for (r = 1; r < mmu_page_sizes; r++) {
 619                         ctrs_sz += sizeof (pgcnt_t) * nranges *
 620                             colors_per_szc[r] * NPC_MUTEX;
 621                 }
 622         }
 623 
 624         /* ctr_mutex */
 625         ctrs_sz += (max_mem_nodes * NPC_MUTEX * sizeof (kmutex_t));
 626 
 627         /* size for page list counts */
 628         PLCNT_SZ(ctrs_sz);
 629 
 630         /*
 631          * add some slop for roundups. page_ctrs_alloc will roundup the start
 632          * address of the counters to ecache_alignsize boundary for every
 633          * memory node.
 634          */
 635         return (ctrs_sz + max_mem_nodes * L2CACHE_ALIGN);
 636 }
 637 
 638 caddr_t
 639 page_ctrs_alloc(caddr_t alloc_base)
 640 {
 641         int     mnode;
 642         int     mrange, nranges;
 643         int     r;              /* region size */
 644         int     i;
 645         int     firstmn;        /* first mnode that exists */
 646         pfn_t   physbase;
 647         pfn_t   physmax;
 648         pgcnt_t colors_per_szc[MMU_PAGE_SIZES];
 649 
 650         /*
 651          * We need to determine how many page colors there are for each
 652          * page size in order to allocate memory for any color specific
 653          * arrays.
 654          */
 655         for (i = 0; i < mmu_page_sizes; i++) {
 656                 colors_per_szc[i] = PAGE_GET_PAGECOLORS(i);
 657         }
 658 
 659         for (r = 1; r < mmu_page_sizes; r++) {
 660                 page_counters[r] = (hw_page_map_t *)alloc_base;
 661                 alloc_base += (max_mem_nodes * sizeof (hw_page_map_t));
 662         }
 663 
 664         /* page_ctrs_cands and pcc_color_free array */
 665         for (i = 0; i < NPC_MUTEX; i++) {
 666                 for (r = 1; r < mmu_page_sizes; r++) {
 667 
 668                         page_ctrs_cands[i][r] = (pcc_info_t **)alloc_base;
 669                         alloc_base += sizeof (pcc_info_t *) * max_mem_nodes;
 670 
 671                         for (mnode = 0; mnode < max_mem_nodes; mnode++) {
 672                                 pcc_info_t *pi;
 673 
 674                                 if (mem_node_config[mnode].exists == 0)
 675                                         continue;
 676 
 677                                 nranges = mnode_nranges[mnode];
 678 
 679                                 pi = (pcc_info_t *)alloc_base;
 680                                 alloc_base += sizeof (pcc_info_t) * nranges;
 681                                 page_ctrs_cands[i][r][mnode] = pi;
 682 
 683                                 for (mrange = 0; mrange < nranges; mrange++) {
 684                                         pi->pcc_color_free =
 685                                             (pgcnt_t *)alloc_base;
 686                                         alloc_base += sizeof (pgcnt_t) *
 687                                             colors_per_szc[r];
 688                                         pi++;
 689                                 }
 690                         }
 691                 }
 692         }
 693 
 694         /* ctr_mutex */
 695         for (i = 0; i < NPC_MUTEX; i++) {
 696                 ctr_mutex[i] = (kmutex_t *)alloc_base;
 697                 alloc_base += (max_mem_nodes * sizeof (kmutex_t));
 698         }
 699 
 700         /* initialize page list counts */
 701         PLCNT_INIT(alloc_base);
 702 
 703         for (firstmn = -1, mnode = 0; mnode < max_mem_nodes; mnode++) {
 704 
 705                 pgcnt_t r_pgcnt;
 706                 pfn_t   r_base;
 707                 pgcnt_t r_align;
 708                 int     r_shift;
 709                 int     nranges = mnode_nranges[mnode];
 710 
 711                 if (mem_node_config[mnode].exists == 0)
 712                         continue;
 713 
 714                 HPM_COUNTERS_LIMITS(mnode, physbase, physmax, firstmn);
 715 
 716                 for (r = 1; r < mmu_page_sizes; r++) {
 717                         /*
 718                          * the page_counters base has to be aligned to the
 719                          * page count of page size code r otherwise the counts
 720                          * will cross large page boundaries.
 721                          */
 722                         r_align = page_get_pagecnt(r);
 723                         r_base = physbase;
 724                         /* base needs to be aligned - lower to aligned value */
 725                         r_base &= ~(r_align - 1);
 726                         r_pgcnt = howmany(physmax - r_base + 1, r_align);
 727                         r_shift = PAGE_BSZS_SHIFT(r);
 728 
 729                         PAGE_COUNTERS_SHIFT(mnode, r) = r_shift;
 730                         PAGE_COUNTERS_ENTRIES(mnode, r) = r_pgcnt;
 731                         PAGE_COUNTERS_BASE(mnode, r) = r_base;
 732                         for (mrange = 0; mrange < nranges; mrange++) {
 733                                 PAGE_COUNTERS_CURRENT_COLOR_ARRAY(mnode,
 734                                     r, mrange) = (size_t *)alloc_base;
 735                                 alloc_base += sizeof (size_t) *
 736                                     colors_per_szc[r];
 737                         }
 738                         for (i = 0; i < colors_per_szc[r]; i++) {
 739                                 uint_t color_mask = colors_per_szc[r] - 1;
 740                                 pfn_t  pfnum = r_base;
 741                                 size_t idx;
 742                                 int mrange;
 743                                 MEM_NODE_ITERATOR_DECL(it);
 744 
 745                                 MEM_NODE_ITERATOR_INIT(pfnum, mnode, r, &it);
 746                                 if (pfnum == (pfn_t)-1) {
 747                                         idx = 0;
 748                                 } else {
 749                                         PAGE_NEXT_PFN_FOR_COLOR(pfnum, r, i,
 750                                             color_mask, color_mask, &it);
 751                                         idx = PNUM_TO_IDX(mnode, r, pfnum);
 752                                         idx = (idx >= r_pgcnt) ? 0 : idx;
 753                                 }
 754                                 for (mrange = 0; mrange < nranges; mrange++) {
 755                                         PAGE_COUNTERS_CURRENT_COLOR(mnode,
 756                                             r, i, mrange) = idx;
 757                                 }
 758                         }
 759 
 760                         /* hpm_counters may be shared by all mnodes */
 761                         if (firstmn == mnode) {
 762                                 PAGE_COUNTERS_COUNTERS(mnode, r) =
 763                                     (hpmctr_t *)alloc_base;
 764                                 alloc_base +=
 765                                     P2ROUNDUP((sizeof (hpmctr_t) * r_pgcnt),
 766                                     sizeof (hpmctr_t *));
 767                         } else {
 768                                 PAGE_COUNTERS_COUNTERS(mnode, r) =
 769                                     PAGE_COUNTERS_COUNTERS(firstmn, r);
 770                         }
 771 
 772                         /*
 773                          * Verify that PNUM_TO_IDX and IDX_TO_PNUM
 774                          * satisfy the identity requirement.
 775                          * We should be able to go from one to the other
 776                          * and get consistent values.
 777                          */
 778                         ASSERT(PNUM_TO_IDX(mnode, r,
 779                             (IDX_TO_PNUM(mnode, r, 0))) == 0);
 780                         ASSERT(IDX_TO_PNUM(mnode, r,
 781                             (PNUM_TO_IDX(mnode, r, r_base))) == r_base);
 782                 }
 783                 /*
 784                  * Roundup the start address of the page_counters to
 785                  * cache aligned boundary for every memory node.
 786                  * page_ctrs_sz() has added some slop for these roundups.
 787                  */
 788                 alloc_base = (caddr_t)P2ROUNDUP((uintptr_t)alloc_base,
 789                     L2CACHE_ALIGN);
 790         }
 791 
 792         /* Initialize other page counter specific data structures. */
 793         for (mnode = 0; mnode < MAX_MEM_NODES; mnode++) {
 794                 rw_init(&page_ctrs_rwlock[mnode], NULL, RW_DEFAULT, NULL);
 795         }
 796 
 797         return (alloc_base);
 798 }
 799 
 800 /*
 801  * Functions to adjust region counters for each size free list.
 802  * Caller is responsible to acquire the ctr_mutex lock if necessary and
 803  * thus can be called during startup without locks.
 804  */
 805 /* ARGSUSED */
 806 void
 807 page_ctr_add_internal(int mnode, int mtype, page_t *pp, int flags)
 808 {
 809         ssize_t         r;      /* region size */
 810         ssize_t         idx;
 811         pfn_t           pfnum;
 812         int             lckidx;
 813 
 814         ASSERT(mnode == PP_2_MEM_NODE(pp));
 815         ASSERT(mtype == PP_2_MTYPE(pp));
 816 
 817         ASSERT(pp->p_szc < mmu_page_sizes);
 818 
 819         PLCNT_INCR(pp, mnode, mtype, pp->p_szc, flags);
 820 
 821         /* no counter update needed for largest page size */
 822         if (pp->p_szc >= mmu_page_sizes - 1) {
 823                 return;
 824         }
 825 
 826         r = pp->p_szc + 1;
 827         pfnum = pp->p_pagenum;
 828         lckidx = PP_CTR_LOCK_INDX(pp);
 829 
 830         /*
 831          * Increment the count of free pages for the current
 832          * region. Continue looping up in region size incrementing
 833          * count if the preceeding region is full.
 834          */
 835         while (r < mmu_page_sizes) {
 836                 idx = PNUM_TO_IDX(mnode, r, pfnum);
 837 
 838                 ASSERT(idx < PAGE_COUNTERS_ENTRIES(mnode, r));
 839                 ASSERT(PAGE_COUNTERS(mnode, r, idx) < FULL_REGION_CNT(r));
 840 
 841                 if (++PAGE_COUNTERS(mnode, r, idx) != FULL_REGION_CNT(r)) {
 842                         break;
 843                 } else {
 844                         int root_mtype = PP_2_MTYPE(PP_GROUPLEADER(pp, r));
 845                         pcc_info_t *cand = &page_ctrs_cands[lckidx][r][mnode]
 846                             [MTYPE_2_MRANGE(mnode, root_mtype)];
 847 
 848                         cand->pcc_pages_free++;
 849                         cand->pcc_color_free[PP_2_BIN_SZC(pp, r)]++;
 850                 }
 851                 r++;
 852         }
 853 }
 854 
 855 void
 856 page_ctr_add(int mnode, int mtype, page_t *pp, int flags)
 857 {
 858         int             lckidx = PP_CTR_LOCK_INDX(pp);
 859         kmutex_t        *lock = &ctr_mutex[lckidx][mnode];
 860 
 861         mutex_enter(lock);
 862         page_ctr_add_internal(mnode, mtype, pp, flags);
 863         mutex_exit(lock);
 864 }
 865 
 866 void
 867 page_ctr_sub_internal(int mnode, int mtype, page_t *pp, int flags)
 868 {
 869         int             lckidx;
 870         ssize_t         r;      /* region size */
 871         ssize_t         idx;
 872         pfn_t           pfnum;
 873 
 874         ASSERT(mnode == PP_2_MEM_NODE(pp));
 875         ASSERT(mtype == PP_2_MTYPE(pp));
 876 
 877         ASSERT(pp->p_szc < mmu_page_sizes);
 878 
 879         PLCNT_DECR(pp, mnode, mtype, pp->p_szc, flags);
 880 
 881         /* no counter update needed for largest page size */
 882         if (pp->p_szc >= mmu_page_sizes - 1) {
 883                 return;
 884         }
 885 
 886         r = pp->p_szc + 1;
 887         pfnum = pp->p_pagenum;
 888         lckidx = PP_CTR_LOCK_INDX(pp);
 889 
 890         /*
 891          * Decrement the count of free pages for the current
 892          * region. Continue looping up in region size decrementing
 893          * count if the preceeding region was full.
 894          */
 895         while (r < mmu_page_sizes) {
 896                 idx = PNUM_TO_IDX(mnode, r, pfnum);
 897 
 898                 ASSERT(idx < PAGE_COUNTERS_ENTRIES(mnode, r));
 899                 ASSERT(PAGE_COUNTERS(mnode, r, idx) > 0);
 900 
 901                 if (--PAGE_COUNTERS(mnode, r, idx) != FULL_REGION_CNT(r) - 1) {
 902                         break;
 903                 } else {
 904                         int root_mtype = PP_2_MTYPE(PP_GROUPLEADER(pp, r));
 905                         pcc_info_t *cand = &page_ctrs_cands[lckidx][r][mnode]
 906                             [MTYPE_2_MRANGE(mnode, root_mtype)];
 907 
 908                         ASSERT(cand->pcc_pages_free != 0);
 909                         ASSERT(cand->pcc_color_free[PP_2_BIN_SZC(pp, r)] != 0);
 910 
 911                         cand->pcc_pages_free--;
 912                         cand->pcc_color_free[PP_2_BIN_SZC(pp, r)]--;
 913                 }
 914                 r++;
 915         }
 916 }
 917 
 918 void
 919 page_ctr_sub(int mnode, int mtype, page_t *pp, int flags)
 920 {
 921         int             lckidx = PP_CTR_LOCK_INDX(pp);
 922         kmutex_t        *lock = &ctr_mutex[lckidx][mnode];
 923 
 924         mutex_enter(lock);
 925         page_ctr_sub_internal(mnode, mtype, pp, flags);
 926         mutex_exit(lock);
 927 }
 928 
 929 /*
 930  * Adjust page counters following a memory attach, since typically the
 931  * size of the array needs to change, and the PFN to counter index
 932  * mapping needs to change.
 933  *
 934  * It is possible this mnode did not exist at startup. In that case
 935  * allocate pcc_info_t and pcc_color_free arrays. Also, allow for nranges
 936  * to change (a theoretical possibility on x86), which means pcc_color_free
 937  * arrays must be extended.
 938  */
 939 uint_t
 940 page_ctrs_adjust(int mnode)
 941 {
 942         pgcnt_t npgs;
 943         int     r;              /* region size */
 944         int     i;
 945         size_t  pcsz, old_csz;
 946         hpmctr_t *new_ctr, *old_ctr;
 947         pfn_t   oldbase, newbase;
 948         pfn_t   physbase, physmax;
 949         size_t  old_npgs;
 950         hpmctr_t *ctr_cache[MMU_PAGE_SIZES];
 951         size_t  size_cache[MMU_PAGE_SIZES];
 952         size_t  *color_cache[MMU_PAGE_SIZES][MAX_MNODE_MRANGES];
 953         size_t  *old_color_array[MAX_MNODE_MRANGES];
 954         pgcnt_t colors_per_szc[MMU_PAGE_SIZES];
 955         pcc_info_t **cands_cache;
 956         pcc_info_t *old_pi, *pi;
 957         pgcnt_t *pgcntp;
 958         int nr, old_nranges, mrange, nranges = MNODE_RANGE_CNT(mnode);
 959         int cands_cache_nranges;
 960         int old_maxmrange, new_maxmrange;
 961         int rc = 0;
 962         int oldmnode;
 963 
 964         cands_cache = kmem_zalloc(sizeof (pcc_info_t *) * NPC_MUTEX *
 965             MMU_PAGE_SIZES, KM_NOSLEEP);
 966         if (cands_cache == NULL)
 967                 return (ENOMEM);
 968 
 969         i = -1;
 970         HPM_COUNTERS_LIMITS(mnode, physbase, physmax, i);
 971 
 972         newbase = physbase & ~PC_BASE_ALIGN_MASK;
 973         npgs = roundup(physmax, PC_BASE_ALIGN) - newbase;
 974 
 975         /* prepare to free non-null pointers on the way out */
 976         cands_cache_nranges = nranges;
 977         bzero(ctr_cache, sizeof (ctr_cache));
 978         bzero(color_cache, sizeof (color_cache));
 979 
 980         /*
 981          * We need to determine how many page colors there are for each
 982          * page size in order to allocate memory for any color specific
 983          * arrays.
 984          */
 985         for (r = 0; r < mmu_page_sizes; r++) {
 986                 colors_per_szc[r] = PAGE_GET_PAGECOLORS(r);
 987         }
 988 
 989         /*
 990          * Preallocate all of the new hpm_counters arrays as we can't
 991          * hold the page_ctrs_rwlock as a writer and allocate memory.
 992          * If we can't allocate all of the arrays, undo our work so far
 993          * and return failure.
 994          */
 995         for (r = 1; r < mmu_page_sizes; r++) {
 996                 pcsz = npgs >> PAGE_BSZS_SHIFT(r);
 997                 size_cache[r] = pcsz;
 998                 ctr_cache[r] = kmem_zalloc(pcsz *
 999                     sizeof (hpmctr_t), KM_NOSLEEP);
1000                 if (ctr_cache[r] == NULL) {
1001                         rc = ENOMEM;
1002                         goto cleanup;
1003                 }
1004         }
1005 
1006         /*
1007          * Preallocate all of the new color current arrays as we can't
1008          * hold the page_ctrs_rwlock as a writer and allocate memory.
1009          * If we can't allocate all of the arrays, undo our work so far
1010          * and return failure.
1011          */
1012         for (r = 1; r < mmu_page_sizes; r++) {
1013                 for (mrange = 0; mrange < nranges; mrange++) {
1014                         color_cache[r][mrange] = kmem_zalloc(sizeof (size_t) *
1015                             colors_per_szc[r], KM_NOSLEEP);
1016                         if (color_cache[r][mrange] == NULL) {
1017                                 rc = ENOMEM;
1018                                 goto cleanup;
1019                         }
1020                 }
1021         }
1022 
1023         /*
1024          * Preallocate all of the new pcc_info_t arrays as we can't
1025          * hold the page_ctrs_rwlock as a writer and allocate memory.
1026          * If we can't allocate all of the arrays, undo our work so far
1027          * and return failure.
1028          */
1029         for (r = 1; r < mmu_page_sizes; r++) {
1030                 for (i = 0; i < NPC_MUTEX; i++) {
1031                         pi = kmem_zalloc(nranges * sizeof (pcc_info_t),
1032                             KM_NOSLEEP);
1033                         if (pi == NULL) {
1034                                 rc = ENOMEM;
1035                                 goto cleanup;
1036                         }
1037                         cands_cache[i * MMU_PAGE_SIZES + r] = pi;
1038 
1039                         for (mrange = 0; mrange < nranges; mrange++, pi++) {
1040                                 pgcntp = kmem_zalloc(colors_per_szc[r] *
1041                                     sizeof (pgcnt_t), KM_NOSLEEP);
1042                                 if (pgcntp == NULL) {
1043                                         rc = ENOMEM;
1044                                         goto cleanup;
1045                                 }
1046                                 pi->pcc_color_free = pgcntp;
1047                         }
1048                 }
1049         }
1050 
1051         /*
1052          * Grab the write lock to prevent others from walking these arrays
1053          * while we are modifying them.
1054          */
1055         PAGE_CTRS_WRITE_LOCK(mnode);
1056 
1057         /*
1058          * For interleaved mnodes, find the first mnode
1059          * with valid page counters since the current
1060          * mnode may have just been added and not have
1061          * valid page counters.
1062          */
1063         if (interleaved_mnodes) {
1064                 for (i = 0; i < max_mem_nodes; i++)
1065                         if (PAGE_COUNTERS_COUNTERS(i, 1) != NULL)
1066                                 break;
1067                 ASSERT(i < max_mem_nodes);
1068                 oldmnode = i;
1069         } else
1070                 oldmnode = mnode;
1071 
1072         old_nranges = mnode_nranges[mnode];
1073         cands_cache_nranges = old_nranges;
1074         mnode_nranges[mnode] = nranges;
1075         old_maxmrange = mnode_maxmrange[mnode];
1076         mnode_maxmrange[mnode] = MNODE_MAX_MRANGE(mnode);
1077         new_maxmrange = mnode_maxmrange[mnode];
1078 
1079         for (r = 1; r < mmu_page_sizes; r++) {
1080                 PAGE_COUNTERS_SHIFT(mnode, r) = PAGE_BSZS_SHIFT(r);
1081                 old_ctr = PAGE_COUNTERS_COUNTERS(oldmnode, r);
1082                 old_csz = PAGE_COUNTERS_ENTRIES(oldmnode, r);
1083                 oldbase = PAGE_COUNTERS_BASE(oldmnode, r);
1084                 old_npgs = old_csz << PAGE_COUNTERS_SHIFT(oldmnode, r);
1085                 for (mrange = 0; mrange < MAX_MNODE_MRANGES; mrange++) {
1086                         old_color_array[mrange] =
1087                             PAGE_COUNTERS_CURRENT_COLOR_ARRAY(mnode,
1088                             r, mrange);
1089                 }
1090 
1091                 pcsz = npgs >> PAGE_COUNTERS_SHIFT(mnode, r);
1092                 new_ctr = ctr_cache[r];
1093                 ctr_cache[r] = NULL;
1094                 if (old_ctr != NULL &&
1095                     (oldbase + old_npgs > newbase) &&
1096                     (newbase + npgs > oldbase)) {
1097                         /*
1098                          * Map the intersection of the old and new
1099                          * counters into the new array.
1100                          */
1101                         size_t offset;
1102                         if (newbase > oldbase) {
1103                                 offset = (newbase - oldbase) >>
1104                                     PAGE_COUNTERS_SHIFT(mnode, r);
1105                                 bcopy(old_ctr + offset, new_ctr,
1106                                     MIN(pcsz, (old_csz - offset)) *
1107                                     sizeof (hpmctr_t));
1108                         } else {
1109                                 offset = (oldbase - newbase) >>
1110                                     PAGE_COUNTERS_SHIFT(mnode, r);
1111                                 bcopy(old_ctr, new_ctr + offset,
1112                                     MIN(pcsz - offset, old_csz) *
1113                                     sizeof (hpmctr_t));
1114                         }
1115                 }
1116 
1117                 PAGE_COUNTERS_COUNTERS(mnode, r) = new_ctr;
1118                 PAGE_COUNTERS_ENTRIES(mnode, r) = pcsz;
1119                 PAGE_COUNTERS_BASE(mnode, r) = newbase;
1120 
1121                 /* update shared hpm_counters in other mnodes */
1122                 if (interleaved_mnodes) {
1123                         for (i = 0; i < max_mem_nodes; i++) {
1124                                 if ((i == mnode) ||
1125                                     (mem_node_config[i].exists == 0))
1126                                         continue;
1127                                 ASSERT(
1128                                     PAGE_COUNTERS_COUNTERS(i, r) == old_ctr ||
1129                                     PAGE_COUNTERS_COUNTERS(i, r) == NULL);
1130                                 PAGE_COUNTERS_COUNTERS(i, r) = new_ctr;
1131                                 PAGE_COUNTERS_ENTRIES(i, r) = pcsz;
1132                                 PAGE_COUNTERS_BASE(i, r) = newbase;
1133                         }
1134                 }
1135 
1136                 for (mrange = 0; mrange < MAX_MNODE_MRANGES; mrange++) {
1137                         PAGE_COUNTERS_CURRENT_COLOR_ARRAY(mnode, r, mrange) =
1138                             color_cache[r][mrange];
1139                         color_cache[r][mrange] = NULL;
1140                 }
1141                 /*
1142                  * for now, just reset on these events as it's probably
1143                  * not worthwhile to try and optimize this.
1144                  */
1145                 for (i = 0; i < colors_per_szc[r]; i++) {
1146                         uint_t color_mask = colors_per_szc[r] - 1;
1147                         int mlo = interleaved_mnodes ? 0 : mnode;
1148                         int mhi = interleaved_mnodes ? max_mem_nodes :
1149                             (mnode + 1);
1150                         int m;
1151                         pfn_t  pfnum;
1152                         size_t idx;
1153                         MEM_NODE_ITERATOR_DECL(it);
1154 
1155                         for (m = mlo; m < mhi; m++) {
1156                                 if (mem_node_config[m].exists == 0)
1157                                         continue;
1158                                 pfnum = newbase;
1159                                 MEM_NODE_ITERATOR_INIT(pfnum, m, r, &it);
1160                                 if (pfnum == (pfn_t)-1) {
1161                                         idx = 0;
1162                                 } else {
1163                                         PAGE_NEXT_PFN_FOR_COLOR(pfnum, r, i,
1164                                             color_mask, color_mask, &it);
1165                                         idx = PNUM_TO_IDX(m, r, pfnum);
1166                                         idx = (idx < pcsz) ? idx : 0;
1167                                 }
1168                                 for (mrange = 0; mrange < nranges; mrange++) {
1169                                         if (PAGE_COUNTERS_CURRENT_COLOR_ARRAY(m,
1170                                             r, mrange) != NULL)
1171                                                 PAGE_COUNTERS_CURRENT_COLOR(m,
1172                                                     r, i, mrange) = idx;
1173                                 }
1174                         }
1175                 }
1176 
1177                 /* cache info for freeing out of the critical path */
1178                 if ((caddr_t)old_ctr >= kernelheap &&
1179                     (caddr_t)old_ctr < ekernelheap) {
1180                         ctr_cache[r] = old_ctr;
1181                         size_cache[r] = old_csz;
1182                 }
1183                 for (mrange = 0; mrange < MAX_MNODE_MRANGES; mrange++) {
1184                         size_t *tmp = old_color_array[mrange];
1185                         if ((caddr_t)tmp >= kernelheap &&
1186                             (caddr_t)tmp < ekernelheap) {
1187                                 color_cache[r][mrange] = tmp;
1188                         }
1189                 }
1190                 /*
1191                  * Verify that PNUM_TO_IDX and IDX_TO_PNUM
1192                  * satisfy the identity requirement.
1193                  * We should be able to go from one to the other
1194                  * and get consistent values.
1195                  */
1196                 ASSERT(PNUM_TO_IDX(mnode, r,
1197                     (IDX_TO_PNUM(mnode, r, 0))) == 0);
1198                 ASSERT(IDX_TO_PNUM(mnode, r,
1199                     (PNUM_TO_IDX(mnode, r, newbase))) == newbase);
1200 
1201                 /* pcc_info_t and pcc_color_free */
1202                 for (i = 0; i < NPC_MUTEX; i++) {
1203                         pcc_info_t *epi;
1204                         pcc_info_t *eold_pi;
1205 
1206                         pi = cands_cache[i * MMU_PAGE_SIZES + r];
1207                         old_pi = page_ctrs_cands[i][r][mnode];
1208                         page_ctrs_cands[i][r][mnode] = pi;
1209                         cands_cache[i * MMU_PAGE_SIZES + r] = old_pi;
1210 
1211                         /* preserve old pcc_color_free values, if any */
1212                         if (old_pi == NULL)
1213                                 continue;
1214 
1215                         /*
1216                          * when/if x86 does DR, must account for
1217                          * possible change in range index when
1218                          * preserving pcc_info
1219                          */
1220                         epi = &pi[nranges];
1221                         eold_pi = &old_pi[old_nranges];
1222                         if (new_maxmrange > old_maxmrange) {
1223                                 pi += new_maxmrange - old_maxmrange;
1224                         } else if (new_maxmrange < old_maxmrange) {
1225                                 old_pi += old_maxmrange - new_maxmrange;
1226                         }
1227                         for (; pi < epi && old_pi < eold_pi; pi++, old_pi++) {
1228                                 pcc_info_t tmp = *pi;
1229                                 *pi = *old_pi;
1230                                 *old_pi = tmp;
1231                         }
1232                 }
1233         }
1234         PAGE_CTRS_WRITE_UNLOCK(mnode);
1235 
1236         /*
1237          * Now that we have dropped the write lock, it is safe to free all
1238          * of the memory we have cached above.
1239          * We come thru here to free memory when pre-alloc fails, and also to
1240          * free old pointers which were recorded while locked.
1241          */
1242 cleanup:
1243         for (r = 1; r < mmu_page_sizes; r++) {
1244                 if (ctr_cache[r] != NULL) {
1245                         kmem_free(ctr_cache[r],
1246                             size_cache[r] * sizeof (hpmctr_t));
1247                 }
1248                 for (mrange = 0; mrange < MAX_MNODE_MRANGES; mrange++) {
1249                         if (color_cache[r][mrange] != NULL) {
1250                                 kmem_free(color_cache[r][mrange],
1251                                     colors_per_szc[r] * sizeof (size_t));
1252                         }
1253                 }
1254                 for (i = 0; i < NPC_MUTEX; i++) {
1255                         pi = cands_cache[i * MMU_PAGE_SIZES + r];
1256                         if (pi == NULL)
1257                                 continue;
1258                         nr = cands_cache_nranges;
1259                         for (mrange = 0; mrange < nr; mrange++, pi++) {
1260                                 pgcntp = pi->pcc_color_free;
1261                                 if (pgcntp == NULL)
1262                                         continue;
1263                                 if ((caddr_t)pgcntp >= kernelheap &&
1264                                     (caddr_t)pgcntp < ekernelheap) {
1265                                         kmem_free(pgcntp,
1266                                             colors_per_szc[r] *
1267                                             sizeof (pgcnt_t));
1268                                 }
1269                         }
1270                         pi = cands_cache[i * MMU_PAGE_SIZES + r];
1271                         if ((caddr_t)pi >= kernelheap &&
1272                             (caddr_t)pi < ekernelheap) {
1273                                 kmem_free(pi, nr * sizeof (pcc_info_t));
1274                         }
1275                 }
1276         }
1277 
1278         kmem_free(cands_cache,
1279             sizeof (pcc_info_t *) * NPC_MUTEX * MMU_PAGE_SIZES);
1280         return (rc);
1281 }
1282 
1283 /*
1284  * Cleanup the hpm_counters field in the page counters
1285  * array.
1286  */
1287 void
1288 page_ctrs_cleanup(void)
1289 {
1290         int r;  /* region size */
1291         int i;  /* mnode index */
1292 
1293         /*
1294          * Get the page counters write lock while we are
1295          * setting the page hpm_counters field to NULL
1296          * for non-existent mnodes.
1297          */
1298         for (i = 0; i < max_mem_nodes; i++) {
1299                 PAGE_CTRS_WRITE_LOCK(i);
1300                 if (mem_node_config[i].exists) {
1301                         PAGE_CTRS_WRITE_UNLOCK(i);
1302                         continue;
1303                 }
1304                 for (r = 1; r < mmu_page_sizes; r++) {
1305                         PAGE_COUNTERS_COUNTERS(i, r) = NULL;
1306                 }
1307                 PAGE_CTRS_WRITE_UNLOCK(i);
1308         }
1309 }
1310 
1311 #ifdef DEBUG
1312 
1313 /*
1314  * confirm pp is a large page corresponding to szc
1315  */
1316 void
1317 chk_lpg(page_t *pp, uchar_t szc)
1318 {
1319         spgcnt_t npgs = page_get_pagecnt(pp->p_szc);
1320         uint_t noreloc;
1321 
1322         if (npgs == 1) {
1323                 ASSERT(pp->p_szc == 0);
1324                 ASSERT(pp->p_next == pp);
1325                 ASSERT(pp->p_prev == pp);
1326                 return;
1327         }
1328 
1329         ASSERT(pp->p_vpnext == pp || pp->p_vpnext == NULL);
1330         ASSERT(pp->p_vpprev == pp || pp->p_vpprev == NULL);
1331 
1332         ASSERT(IS_P2ALIGNED(pp->p_pagenum, npgs));
1333         ASSERT(pp->p_pagenum == (pp->p_next->p_pagenum - 1));
1334         ASSERT(pp->p_prev->p_pagenum == (pp->p_pagenum + (npgs - 1)));
1335         ASSERT(pp->p_prev == (pp + (npgs - 1)));
1336 
1337         /*
1338          * Check list of pages.
1339          */
1340         noreloc = PP_ISNORELOC(pp);
1341         while (npgs--) {
1342                 if (npgs != 0) {
1343                         ASSERT(pp->p_pagenum == pp->p_next->p_pagenum - 1);
1344                         ASSERT(pp->p_next == (pp + 1));
1345                 }
1346                 ASSERT(pp->p_szc == szc);
1347                 ASSERT(PP_ISFREE(pp));
1348                 ASSERT(PP_ISAGED(pp));
1349                 ASSERT(pp->p_vpnext == pp || pp->p_vpnext == NULL);
1350                 ASSERT(pp->p_vpprev == pp || pp->p_vpprev == NULL);
1351                 ASSERT(pp->p_vnode  == NULL);
1352                 ASSERT(PP_ISNORELOC(pp) == noreloc);
1353 
1354                 pp = pp->p_next;
1355         }
1356 }
1357 #endif /* DEBUG */
1358 
1359 void
1360 page_freelist_lock(int mnode)
1361 {
1362         int i;
1363         for (i = 0; i < NPC_MUTEX; i++) {
1364                 mutex_enter(FPC_MUTEX(mnode, i));
1365                 mutex_enter(CPC_MUTEX(mnode, i));
1366         }
1367 }
1368 
1369 void
1370 page_freelist_unlock(int mnode)
1371 {
1372         int i;
1373         for (i = 0; i < NPC_MUTEX; i++) {
1374                 mutex_exit(FPC_MUTEX(mnode, i));
1375                 mutex_exit(CPC_MUTEX(mnode, i));
1376         }
1377 }
1378 
1379 /*
1380  * add pp to the specified page list. Defaults to head of the page list
1381  * unless PG_LIST_TAIL is specified.
1382  */
1383 void
1384 page_list_add(page_t *pp, int flags)
1385 {
1386         page_t          **ppp;
1387         kmutex_t        *pcm;
1388         uint_t          bin, mtype;
1389         int             mnode;
1390 
1391         ASSERT(PAGE_EXCL(pp) || (flags & PG_LIST_ISINIT));
1392         ASSERT(PP_ISFREE(pp));
1393         ASSERT(!hat_page_is_mapped(pp));
1394         ASSERT(hat_page_getshare(pp) == 0);
1395 
1396         /*
1397          * Large pages should be freed via page_list_add_pages().
1398          */
1399         ASSERT(pp->p_szc == 0);
1400 
1401         /*
1402          * Don't need to lock the freelist first here
1403          * because the page isn't on the freelist yet.
1404          * This means p_szc can't change on us.
1405          */
1406 
1407         bin = PP_2_BIN(pp);
1408         mnode = PP_2_MEM_NODE(pp);
1409         mtype = PP_2_MTYPE(pp);
1410 
1411         if (flags & PG_LIST_ISINIT) {
1412                 /*
1413                  * PG_LIST_ISINIT is set during system startup (ie. single
1414                  * threaded), add a page to the free list and add to the
1415                  * the free region counters w/o any locking
1416                  */
1417                 ppp = &PAGE_FREELISTS(mnode, 0, bin, mtype);
1418 
1419                 /* inline version of page_add() */
1420                 if (*ppp != NULL) {
1421                         pp->p_next = *ppp;
1422                         pp->p_prev = (*ppp)->p_prev;
1423                         (*ppp)->p_prev = pp;
1424                         pp->p_prev->p_next = pp;
1425                 } else
1426                         *ppp = pp;
1427 
1428                 page_ctr_add_internal(mnode, mtype, pp, flags);
1429                 VM_STAT_ADD(vmm_vmstats.pladd_free[0]);
1430         } else {
1431                 pcm = PC_BIN_MUTEX(mnode, bin, flags);
1432 
1433                 if (flags & PG_FREE_LIST) {
1434                         VM_STAT_ADD(vmm_vmstats.pladd_free[0]);
1435                         ASSERT(PP_ISAGED(pp));
1436                         ppp = &PAGE_FREELISTS(mnode, 0, bin, mtype);
1437 
1438                 } else {
1439                         VM_STAT_ADD(vmm_vmstats.pladd_cache);
1440                         ASSERT(pp->p_vnode);
1441                         ASSERT((pp->p_offset & PAGEOFFSET) == 0);
1442                         ppp = &PAGE_CACHELISTS(mnode, bin, mtype);
1443                 }
1444                 mutex_enter(pcm);
1445                 page_add(ppp, pp);
1446 
1447                 if (flags & PG_LIST_TAIL)
1448                         *ppp = (*ppp)->p_next;
1449                 /*
1450                  * Add counters before releasing pcm mutex to avoid a race with
1451                  * page_freelist_coalesce and page_freelist_split.
1452                  */
1453                 page_ctr_add(mnode, mtype, pp, flags);
1454                 mutex_exit(pcm);
1455         }
1456 
1457 
1458 #if defined(__sparc)
1459         if (PP_ISNORELOC(pp)) {
1460                 kcage_freemem_add(1);
1461         }
1462 #endif
1463         /*
1464          * It is up to the caller to unlock the page!
1465          */
1466         ASSERT(PAGE_EXCL(pp) || (flags & PG_LIST_ISINIT));
1467 }
1468 
1469 
1470 #ifdef __sparc
1471 /*
1472  * This routine is only used by kcage_init during system startup.
1473  * It performs the function of page_list_sub/PP_SETNORELOC/page_list_add
1474  * without the overhead of taking locks and updating counters.
1475  */
1476 void
1477 page_list_noreloc_startup(page_t *pp)
1478 {
1479         page_t          **ppp;
1480         uint_t          bin;
1481         int             mnode;
1482         int             mtype;
1483         int             flags = 0;
1484 
1485         /*
1486          * If this is a large page on the freelist then
1487          * break it up into smaller pages.
1488          */
1489         if (pp->p_szc != 0)
1490                 page_boot_demote(pp);
1491 
1492         /*
1493          * Get list page is currently on.
1494          */
1495         bin = PP_2_BIN(pp);
1496         mnode = PP_2_MEM_NODE(pp);
1497         mtype = PP_2_MTYPE(pp);
1498         ASSERT(mtype == MTYPE_RELOC);
1499         ASSERT(pp->p_szc == 0);
1500 
1501         if (PP_ISAGED(pp)) {
1502                 ppp = &PAGE_FREELISTS(mnode, 0, bin, mtype);
1503                 flags |= PG_FREE_LIST;
1504         } else {
1505                 ppp = &PAGE_CACHELISTS(mnode, bin, mtype);
1506                 flags |= PG_CACHE_LIST;
1507         }
1508 
1509         ASSERT(*ppp != NULL);
1510 
1511         /*
1512          * Delete page from current list.
1513          */
1514         if (*ppp == pp)
1515                 *ppp = pp->p_next;           /* go to next page */
1516         if (*ppp == pp) {
1517                 *ppp = NULL;                    /* page list is gone */
1518         } else {
1519                 pp->p_prev->p_next = pp->p_next;
1520                 pp->p_next->p_prev = pp->p_prev;
1521         }
1522 
1523         /*
1524          * Decrement page counters
1525          */
1526         page_ctr_sub_internal(mnode, mtype, pp, flags);
1527 
1528         /*
1529          * Set no reloc for cage initted pages.
1530          */
1531         PP_SETNORELOC(pp);
1532 
1533         mtype = PP_2_MTYPE(pp);
1534         ASSERT(mtype == MTYPE_NORELOC);
1535 
1536         /*
1537          * Get new list for page.
1538          */
1539         if (PP_ISAGED(pp)) {
1540                 ppp = &PAGE_FREELISTS(mnode, 0, bin, mtype);
1541         } else {
1542                 ppp = &PAGE_CACHELISTS(mnode, bin, mtype);
1543         }
1544 
1545         /*
1546          * Insert page on new list.
1547          */
1548         if (*ppp == NULL) {
1549                 *ppp = pp;
1550                 pp->p_next = pp->p_prev = pp;
1551         } else {
1552                 pp->p_next = *ppp;
1553                 pp->p_prev = (*ppp)->p_prev;
1554                 (*ppp)->p_prev = pp;
1555                 pp->p_prev->p_next = pp;
1556         }
1557 
1558         /*
1559          * Increment page counters
1560          */
1561         page_ctr_add_internal(mnode, mtype, pp, flags);
1562 
1563         /*
1564          * Update cage freemem counter
1565          */
1566         atomic_inc_ulong(&kcage_freemem);
1567 }
1568 #else   /* __sparc */
1569 
1570 /* ARGSUSED */
1571 void
1572 page_list_noreloc_startup(page_t *pp)
1573 {
1574         panic("page_list_noreloc_startup: should be here only for sparc");
1575 }
1576 #endif
1577 
1578 void
1579 page_list_add_pages(page_t *pp, int flags)
1580 {
1581         kmutex_t *pcm;
1582         pgcnt_t pgcnt;
1583         uint_t  bin, mtype, i;
1584         int     mnode;
1585 
1586         /* default to freelist/head */
1587         ASSERT((flags & (PG_CACHE_LIST | PG_LIST_TAIL)) == 0);
1588 
1589         CHK_LPG(pp, pp->p_szc);
1590         VM_STAT_ADD(vmm_vmstats.pladd_free[pp->p_szc]);
1591 
1592         bin = PP_2_BIN(pp);
1593         mnode = PP_2_MEM_NODE(pp);
1594         mtype = PP_2_MTYPE(pp);
1595 
1596         if (flags & PG_LIST_ISINIT) {
1597                 ASSERT(pp->p_szc == mmu_page_sizes - 1);
1598                 page_vpadd(&PAGE_FREELISTS(mnode, pp->p_szc, bin, mtype), pp);
1599                 ASSERT(!PP_ISNORELOC(pp));
1600                 PLCNT_INCR(pp, mnode, mtype, pp->p_szc, flags);
1601         } else {
1602 
1603                 ASSERT(pp->p_szc != 0 && pp->p_szc < mmu_page_sizes);
1604 
1605                 pcm = PC_BIN_MUTEX(mnode, bin, PG_FREE_LIST);
1606 
1607                 mutex_enter(pcm);
1608                 page_vpadd(&PAGE_FREELISTS(mnode, pp->p_szc, bin, mtype), pp);
1609                 page_ctr_add(mnode, mtype, pp, PG_FREE_LIST);
1610                 mutex_exit(pcm);
1611 
1612                 pgcnt = page_get_pagecnt(pp->p_szc);
1613 #if defined(__sparc)
1614                 if (PP_ISNORELOC(pp))
1615                         kcage_freemem_add(pgcnt);
1616 #endif
1617                 for (i = 0; i < pgcnt; i++, pp++)
1618                         page_unlock_nocapture(pp);
1619         }
1620 }
1621 
1622 /*
1623  * During boot, need to demote a large page to base
1624  * pagesize pages for seg_kmem for use in boot_alloc()
1625  */
1626 void
1627 page_boot_demote(page_t *pp)
1628 {
1629         ASSERT(pp->p_szc != 0);
1630         ASSERT(PP_ISFREE(pp));
1631         ASSERT(PP_ISAGED(pp));
1632 
1633         (void) page_demote(PP_2_MEM_NODE(pp),
1634             PFN_BASE(pp->p_pagenum, pp->p_szc), 0, pp->p_szc, 0, PC_NO_COLOR,
1635             PC_FREE);
1636 
1637         ASSERT(PP_ISFREE(pp));
1638         ASSERT(PP_ISAGED(pp));
1639         ASSERT(pp->p_szc == 0);
1640 }
1641 
1642 /*
1643  * Take a particular page off of whatever freelist the page
1644  * is claimed to be on.
1645  *
1646  * NOTE: Only used for PAGESIZE pages.
1647  */
1648 void
1649 page_list_sub(page_t *pp, int flags)
1650 {
1651         int             bin;
1652         uint_t          mtype;
1653         int             mnode;
1654         kmutex_t        *pcm;
1655         page_t          **ppp;
1656 
1657         ASSERT(PAGE_EXCL(pp));
1658         ASSERT(PP_ISFREE(pp));
1659 
1660         /*
1661          * The p_szc field can only be changed by page_promote()
1662          * and page_demote(). Only free pages can be promoted and
1663          * demoted and the free list MUST be locked during these
1664          * operations. So to prevent a race in page_list_sub()
1665          * between computing which bin of the freelist lock to
1666          * grab and actually grabing the lock we check again that
1667          * the bin we locked is still the correct one. Notice that
1668          * the p_szc field could have actually changed on us but
1669          * if the bin happens to still be the same we are safe.
1670          */
1671 try_again:
1672         bin = PP_2_BIN(pp);
1673         mnode = PP_2_MEM_NODE(pp);
1674         pcm = PC_BIN_MUTEX(mnode, bin, flags);
1675         mutex_enter(pcm);
1676         if (PP_2_BIN(pp) != bin) {
1677                 mutex_exit(pcm);
1678                 goto try_again;
1679         }
1680         mtype = PP_2_MTYPE(pp);
1681 
1682         if (flags & PG_FREE_LIST) {
1683                 VM_STAT_ADD(vmm_vmstats.plsub_free[0]);
1684                 ASSERT(PP_ISAGED(pp));
1685                 ppp = &PAGE_FREELISTS(mnode, pp->p_szc, bin, mtype);
1686         } else {
1687                 VM_STAT_ADD(vmm_vmstats.plsub_cache);
1688                 ASSERT(!PP_ISAGED(pp));
1689                 ppp = &PAGE_CACHELISTS(mnode, bin, mtype);
1690         }
1691 
1692         /*
1693          * Common PAGESIZE case.
1694          *
1695          * Note that we locked the freelist. This prevents
1696          * any page promotion/demotion operations. Therefore
1697          * the p_szc will not change until we drop pcm mutex.
1698          */
1699         if (pp->p_szc == 0) {
1700                 page_sub(ppp, pp);
1701                 /*
1702                  * Subtract counters before releasing pcm mutex
1703                  * to avoid race with page_freelist_coalesce.
1704                  */
1705                 page_ctr_sub(mnode, mtype, pp, flags);
1706                 mutex_exit(pcm);
1707 
1708 #if defined(__sparc)
1709                 if (PP_ISNORELOC(pp)) {
1710                         kcage_freemem_sub(1);
1711                 }
1712 #endif
1713                 return;
1714         }
1715 
1716         /*
1717          * Large pages on the cache list are not supported.
1718          */
1719         if (flags & PG_CACHE_LIST)
1720                 panic("page_list_sub: large page on cachelist");
1721 
1722         /*
1723          * Slow but rare.
1724          *
1725          * Somebody wants this particular page which is part
1726          * of a large page. In this case we just demote the page
1727          * if it's on the freelist.
1728          *
1729          * We have to drop pcm before locking the entire freelist.
1730          * Once we have re-locked the freelist check to make sure
1731          * the page hasn't already been demoted or completely
1732          * freed.
1733          */
1734         mutex_exit(pcm);
1735         page_freelist_lock(mnode);
1736         if (pp->p_szc != 0) {
1737                 /*
1738                  * Large page is on freelist.
1739                  */
1740                 (void) page_demote(mnode, PFN_BASE(pp->p_pagenum, pp->p_szc),
1741                     0, pp->p_szc, 0, PC_NO_COLOR, PC_FREE);
1742         }
1743         ASSERT(PP_ISFREE(pp));
1744         ASSERT(PP_ISAGED(pp));
1745         ASSERT(pp->p_szc == 0);
1746 
1747         /*
1748          * Subtract counters before releasing pcm mutex
1749          * to avoid race with page_freelist_coalesce.
1750          */
1751         bin = PP_2_BIN(pp);
1752         mtype = PP_2_MTYPE(pp);
1753         ppp = &PAGE_FREELISTS(mnode, pp->p_szc, bin, mtype);
1754 
1755         page_sub(ppp, pp);
1756         page_ctr_sub(mnode, mtype, pp, flags);
1757         page_freelist_unlock(mnode);
1758 
1759 #if defined(__sparc)
1760         if (PP_ISNORELOC(pp)) {
1761                 kcage_freemem_sub(1);
1762         }
1763 #endif
1764 }
1765 
1766 void
1767 page_list_sub_pages(page_t *pp, uint_t szc)
1768 {
1769         kmutex_t *pcm;
1770         uint_t  bin, mtype;
1771         int     mnode;
1772 
1773         ASSERT(PAGE_EXCL(pp));
1774         ASSERT(PP_ISFREE(pp));
1775         ASSERT(PP_ISAGED(pp));
1776 
1777         /*
1778          * See comment in page_list_sub().
1779          */
1780 try_again:
1781         bin = PP_2_BIN(pp);
1782         mnode = PP_2_MEM_NODE(pp);
1783         pcm = PC_BIN_MUTEX(mnode, bin, PG_FREE_LIST);
1784         mutex_enter(pcm);
1785         if (PP_2_BIN(pp) != bin) {
1786                 mutex_exit(pcm);
1787                 goto    try_again;
1788         }
1789 
1790         /*
1791          * If we're called with a page larger than szc or it got
1792          * promoted above szc before we locked the freelist then
1793          * drop pcm and re-lock entire freelist. If page still larger
1794          * than szc then demote it.
1795          */
1796         if (pp->p_szc > szc) {
1797                 mutex_exit(pcm);
1798                 pcm = NULL;
1799                 page_freelist_lock(mnode);
1800                 if (pp->p_szc > szc) {
1801                         VM_STAT_ADD(vmm_vmstats.plsubpages_szcbig);
1802                         (void) page_demote(mnode,
1803                             PFN_BASE(pp->p_pagenum, pp->p_szc), 0,
1804                             pp->p_szc, szc, PC_NO_COLOR, PC_FREE);
1805                 }
1806                 bin = PP_2_BIN(pp);
1807         }
1808         ASSERT(PP_ISFREE(pp));
1809         ASSERT(PP_ISAGED(pp));
1810         ASSERT(pp->p_szc <= szc);
1811         ASSERT(pp == PP_PAGEROOT(pp));
1812 
1813         VM_STAT_ADD(vmm_vmstats.plsub_free[pp->p_szc]);
1814 
1815         mtype = PP_2_MTYPE(pp);
1816         if (pp->p_szc != 0) {
1817                 page_vpsub(&PAGE_FREELISTS(mnode, pp->p_szc, bin, mtype), pp);
1818                 CHK_LPG(pp, pp->p_szc);
1819         } else {
1820                 VM_STAT_ADD(vmm_vmstats.plsubpages_szc0);
1821                 page_sub(&PAGE_FREELISTS(mnode, pp->p_szc, bin, mtype), pp);
1822         }
1823         page_ctr_sub(mnode, mtype, pp, PG_FREE_LIST);
1824 
1825         if (pcm != NULL) {
1826                 mutex_exit(pcm);
1827         } else {
1828                 page_freelist_unlock(mnode);
1829         }
1830 
1831 #if defined(__sparc)
1832         if (PP_ISNORELOC(pp)) {
1833                 pgcnt_t pgcnt;
1834 
1835                 pgcnt = page_get_pagecnt(pp->p_szc);
1836                 kcage_freemem_sub(pgcnt);
1837         }
1838 #endif
1839 }
1840 
1841 /*
1842  * Add the page to the front of a linked list of pages
1843  * using the p_next & p_prev pointers for the list.
1844  * The caller is responsible for protecting the list pointers.
1845  */
1846 void
1847 mach_page_add(page_t **ppp, page_t *pp)
1848 {
1849         if (*ppp == NULL) {
1850                 pp->p_next = pp->p_prev = pp;
1851         } else {
1852                 pp->p_next = *ppp;
1853                 pp->p_prev = (*ppp)->p_prev;
1854                 (*ppp)->p_prev = pp;
1855                 pp->p_prev->p_next = pp;
1856         }
1857         *ppp = pp;
1858 }
1859 
1860 /*
1861  * Remove this page from a linked list of pages
1862  * using the p_next & p_prev pointers for the list.
1863  *
1864  * The caller is responsible for protecting the list pointers.
1865  */
1866 void
1867 mach_page_sub(page_t **ppp, page_t *pp)
1868 {
1869         ASSERT(PP_ISFREE(pp));
1870 
1871         if (*ppp == NULL || pp == NULL)
1872                 panic("mach_page_sub");
1873 
1874         if (*ppp == pp)
1875                 *ppp = pp->p_next;           /* go to next page */
1876 
1877         if (*ppp == pp)
1878                 *ppp = NULL;                    /* page list is gone */
1879         else {
1880                 pp->p_prev->p_next = pp->p_next;
1881                 pp->p_next->p_prev = pp->p_prev;
1882         }
1883         pp->p_prev = pp->p_next = pp;             /* make pp a list of one */
1884 }
1885 
1886 /*
1887  * Routine fsflush uses to gradually coalesce the free list into larger pages.
1888  */
1889 void
1890 page_promote_size(page_t *pp, uint_t cur_szc)
1891 {
1892         pfn_t pfn;
1893         int mnode;
1894         int idx;
1895         int new_szc = cur_szc + 1;
1896         int full = FULL_REGION_CNT(new_szc);
1897 
1898         pfn = page_pptonum(pp);
1899         mnode = PFN_2_MEM_NODE(pfn);
1900 
1901         page_freelist_lock(mnode);
1902 
1903         idx = PNUM_TO_IDX(mnode, new_szc, pfn);
1904         if (PAGE_COUNTERS(mnode, new_szc, idx) == full)
1905                 (void) page_promote(mnode, pfn, new_szc, PC_FREE, PC_MTYPE_ANY);
1906 
1907         page_freelist_unlock(mnode);
1908 }
1909 
1910 static uint_t page_promote_err;
1911 static uint_t page_promote_noreloc_err;
1912 
1913 /*
1914  * Create a single larger page (of szc new_szc) from smaller contiguous pages
1915  * for the given mnode starting at pfnum. Pages involved are on the freelist
1916  * before the call and may be returned to the caller if requested, otherwise
1917  * they will be placed back on the freelist.
1918  * If flags is PC_ALLOC, then the large page will be returned to the user in
1919  * a state which is consistent with a page being taken off the freelist.  If
1920  * we failed to lock the new large page, then we will return NULL to the
1921  * caller and put the large page on the freelist instead.
1922  * If flags is PC_FREE, then the large page will be placed on the freelist,
1923  * and NULL will be returned.
1924  * The caller is responsible for locking the freelist as well as any other
1925  * accounting which needs to be done for a returned page.
1926  *
1927  * RFE: For performance pass in pp instead of pfnum so
1928  *      we can avoid excessive calls to page_numtopp_nolock().
1929  *      This would depend on an assumption that all contiguous
1930  *      pages are in the same memseg so we can just add/dec
1931  *      our pp.
1932  *
1933  * Lock ordering:
1934  *
1935  *      There is a potential but rare deadlock situation
1936  *      for page promotion and demotion operations. The problem
1937  *      is there are two paths into the freelist manager and
1938  *      they have different lock orders:
1939  *
1940  *      page_create()
1941  *              lock freelist
1942  *              page_lock(EXCL)
1943  *              unlock freelist
1944  *              return
1945  *              caller drops page_lock
1946  *
1947  *      page_free() and page_reclaim()
1948  *              caller grabs page_lock(EXCL)
1949  *
1950  *              lock freelist
1951  *              unlock freelist
1952  *              drop page_lock
1953  *
1954  *      What prevents a thread in page_create() from deadlocking
1955  *      with a thread freeing or reclaiming the same page is the
1956  *      page_trylock() in page_get_freelist(). If the trylock fails
1957  *      it skips the page.
1958  *
1959  *      The lock ordering for promotion and demotion is the same as
1960  *      for page_create(). Since the same deadlock could occur during
1961  *      page promotion and freeing or reclaiming of a page on the
1962  *      cache list we might have to fail the operation and undo what
1963  *      have done so far. Again this is rare.
1964  */
1965 page_t *
1966 page_promote(int mnode, pfn_t pfnum, uchar_t new_szc, int flags, int mtype)
1967 {
1968         page_t          *pp, *pplist, *tpp, *start_pp;
1969         pgcnt_t         new_npgs, npgs;
1970         uint_t          bin;
1971         pgcnt_t         tmpnpgs, pages_left;
1972         uint_t          noreloc;
1973         int             which_list;
1974         ulong_t         index;
1975         kmutex_t        *phm;
1976 
1977         /*
1978          * General algorithm:
1979          * Find the starting page
1980          * Walk each page struct removing it from the freelist,
1981          * and linking it to all the other pages removed.
1982          * Once all pages are off the freelist,
1983          * walk the list, modifying p_szc to new_szc and what
1984          * ever other info needs to be done to create a large free page.
1985          * According to the flags, either return the page or put it
1986          * on the freelist.
1987          */
1988 
1989         start_pp = page_numtopp_nolock(pfnum);
1990         ASSERT(start_pp && (start_pp->p_pagenum == pfnum));
1991         new_npgs = page_get_pagecnt(new_szc);
1992         ASSERT(IS_P2ALIGNED(pfnum, new_npgs));
1993 
1994         /* don't return page of the wrong mtype */
1995         if (mtype != PC_MTYPE_ANY && mtype != PP_2_MTYPE(start_pp))
1996                         return (NULL);
1997 
1998         /*
1999          * Loop through smaller pages to confirm that all pages
2000          * give the same result for PP_ISNORELOC().
2001          * We can check this reliably here as the protocol for setting
2002          * P_NORELOC requires pages to be taken off the free list first.
2003          */
2004         noreloc = PP_ISNORELOC(start_pp);
2005         for (pp = start_pp + new_npgs; --pp > start_pp; ) {
2006                 if (noreloc != PP_ISNORELOC(pp)) {
2007                         page_promote_noreloc_err++;
2008                         page_promote_err++;
2009                         return (NULL);
2010                 }
2011         }
2012 
2013         pages_left = new_npgs;
2014         pplist = NULL;
2015         pp = start_pp;
2016 
2017         /* Loop around coalescing the smaller pages into a big page. */
2018         while (pages_left) {
2019                 /*
2020                  * Remove from the freelist.
2021                  */
2022                 ASSERT(PP_ISFREE(pp));
2023                 bin = PP_2_BIN(pp);
2024                 ASSERT(mnode == PP_2_MEM_NODE(pp));
2025                 mtype = PP_2_MTYPE(pp);
2026                 if (PP_ISAGED(pp)) {
2027 
2028                         /*
2029                          * PG_FREE_LIST
2030                          */
2031                         if (pp->p_szc) {
2032                                 page_vpsub(&PAGE_FREELISTS(mnode,
2033                                     pp->p_szc, bin, mtype), pp);
2034                         } else {
2035                                 mach_page_sub(&PAGE_FREELISTS(mnode, 0,
2036                                     bin, mtype), pp);
2037                         }
2038                         which_list = PG_FREE_LIST;
2039                 } else {
2040                         ASSERT(pp->p_szc == 0);
2041 
2042                         /*
2043                          * PG_CACHE_LIST
2044                          *
2045                          * Since this page comes from the
2046                          * cachelist, we must destroy the
2047                          * vnode association.
2048                          */
2049                         if (!page_trylock(pp, SE_EXCL)) {
2050                                 goto fail_promote;
2051                         }
2052 
2053                         /*
2054                          * We need to be careful not to deadlock
2055                          * with another thread in page_lookup().
2056                          * The page_lookup() thread could be holding
2057                          * the same phm that we need if the two
2058                          * pages happen to hash to the same phm lock.
2059                          * At this point we have locked the entire
2060                          * freelist and page_lookup() could be trying
2061                          * to grab a freelist lock.
2062                          */
2063                         index = PAGE_HASH_FUNC(pp->p_vnode, pp->p_offset);
2064                         phm = PAGE_HASH_MUTEX(index);
2065                         if (!mutex_tryenter(phm)) {
2066                                 page_unlock_nocapture(pp);
2067                                 goto fail_promote;
2068                         }
2069 
2070                         mach_page_sub(&PAGE_CACHELISTS(mnode, bin, mtype), pp);
2071                         page_hashout(pp, phm);
2072                         mutex_exit(phm);
2073                         PP_SETAGED(pp);
2074                         page_unlock_nocapture(pp);
2075                         which_list = PG_CACHE_LIST;
2076                 }
2077                 page_ctr_sub(mnode, mtype, pp, which_list);
2078 
2079                 /*
2080                  * Concatenate the smaller page(s) onto
2081                  * the large page list.
2082                  */
2083                 tmpnpgs = npgs = page_get_pagecnt(pp->p_szc);
2084                 pages_left -= npgs;
2085                 tpp = pp;
2086                 while (npgs--) {
2087                         tpp->p_szc = new_szc;
2088                         tpp = tpp->p_next;
2089                 }
2090                 page_list_concat(&pplist, &pp);
2091                 pp += tmpnpgs;
2092         }
2093         CHK_LPG(pplist, new_szc);
2094 
2095         /*
2096          * return the page to the user if requested
2097          * in the properly locked state.
2098          */
2099         if (flags == PC_ALLOC && (page_trylock_cons(pplist, SE_EXCL))) {
2100                 return (pplist);
2101         }
2102 
2103         /*
2104          * Otherwise place the new large page on the freelist
2105          */
2106         bin = PP_2_BIN(pplist);
2107         mnode = PP_2_MEM_NODE(pplist);
2108         mtype = PP_2_MTYPE(pplist);
2109         page_vpadd(&PAGE_FREELISTS(mnode, new_szc, bin, mtype), pplist);
2110 
2111         page_ctr_add(mnode, mtype, pplist, PG_FREE_LIST);
2112         return (NULL);
2113 
2114 fail_promote:
2115         /*
2116          * A thread must have still been freeing or
2117          * reclaiming the page on the cachelist.
2118          * To prevent a deadlock undo what we have
2119          * done sofar and return failure. This
2120          * situation can only happen while promoting
2121          * PAGESIZE pages.
2122          */
2123         page_promote_err++;
2124         while (pplist) {
2125                 pp = pplist;
2126                 mach_page_sub(&pplist, pp);
2127                 pp->p_szc = 0;
2128                 bin = PP_2_BIN(pp);
2129                 mtype = PP_2_MTYPE(pp);
2130                 mach_page_add(&PAGE_FREELISTS(mnode, 0, bin, mtype), pp);
2131                 page_ctr_add(mnode, mtype, pp, PG_FREE_LIST);
2132         }
2133         return (NULL);
2134 
2135 }
2136 
2137 /*
2138  * Break up a large page into smaller size pages.
2139  * Pages involved are on the freelist before the call and may
2140  * be returned to the caller if requested, otherwise they will
2141  * be placed back on the freelist.
2142  * The caller is responsible for locking the freelist as well as any other
2143  * accounting which needs to be done for a returned page.
2144  * If flags is not PC_ALLOC, the color argument is ignored, and thus
2145  * technically, any value may be passed in but PC_NO_COLOR is the standard
2146  * which should be followed for clarity's sake.
2147  * Returns a page whose pfn is < pfnmax
2148  */
2149 page_t *
2150 page_demote(int mnode, pfn_t pfnum, pfn_t pfnmax, uchar_t cur_szc,
2151     uchar_t new_szc, int color, int flags)
2152 {
2153         page_t  *pp, *pplist, *npplist;
2154         pgcnt_t npgs, n;
2155         uint_t  bin;
2156         uint_t  mtype;
2157         page_t  *ret_pp = NULL;
2158 
2159         ASSERT(cur_szc != 0);
2160         ASSERT(new_szc < cur_szc);
2161 
2162         pplist = page_numtopp_nolock(pfnum);
2163         ASSERT(pplist != NULL);
2164 
2165         ASSERT(pplist->p_szc == cur_szc);
2166 
2167         bin = PP_2_BIN(pplist);
2168         ASSERT(mnode == PP_2_MEM_NODE(pplist));
2169         mtype = PP_2_MTYPE(pplist);
2170         page_vpsub(&PAGE_FREELISTS(mnode, cur_szc, bin, mtype), pplist);
2171 
2172         CHK_LPG(pplist, cur_szc);
2173         page_ctr_sub(mnode, mtype, pplist, PG_FREE_LIST);
2174 
2175         /*
2176          * Number of PAGESIZE pages for smaller new_szc
2177          * page.
2178          */
2179         npgs = page_get_pagecnt(new_szc);
2180 
2181         while (pplist) {
2182                 pp = pplist;
2183 
2184                 ASSERT(pp->p_szc == cur_szc);
2185 
2186                 /*
2187                  * We either break it up into PAGESIZE pages or larger.
2188                  */
2189                 if (npgs == 1) {        /* PAGESIZE case */
2190                         mach_page_sub(&pplist, pp);
2191                         ASSERT(pp->p_szc == cur_szc);
2192                         ASSERT(new_szc == 0);
2193                         ASSERT(mnode == PP_2_MEM_NODE(pp));
2194                         pp->p_szc = new_szc;
2195                         bin = PP_2_BIN(pp);
2196                         if ((bin == color) && (flags == PC_ALLOC) &&
2197                             (ret_pp == NULL) && (pfnmax == 0 ||
2198                             pp->p_pagenum < pfnmax) &&
2199                             page_trylock_cons(pp, SE_EXCL)) {
2200                                 ret_pp = pp;
2201                         } else {
2202                                 mtype = PP_2_MTYPE(pp);
2203                                 mach_page_add(&PAGE_FREELISTS(mnode, 0, bin,
2204                                     mtype), pp);
2205                                 page_ctr_add(mnode, mtype, pp, PG_FREE_LIST);
2206                         }
2207                 } else {
2208                         page_t *try_to_return_this_page = NULL;
2209                         int count = 0;
2210 
2211                         /*
2212                          * Break down into smaller lists of pages.
2213                          */
2214                         page_list_break(&pplist, &npplist, npgs);
2215 
2216                         pp = pplist;
2217                         n = npgs;
2218                         while (n--) {
2219                                 ASSERT(pp->p_szc == cur_szc);
2220                                 /*
2221                                  * Check whether all the pages in this list
2222                                  * fit the request criteria.
2223                                  */
2224                                 if (pfnmax == 0 || pp->p_pagenum < pfnmax) {
2225                                         count++;
2226                                 }
2227                                 pp->p_szc = new_szc;
2228                                 pp = pp->p_next;
2229                         }
2230 
2231                         if (count == npgs &&
2232                             (pfnmax == 0 || pp->p_pagenum < pfnmax)) {
2233                                 try_to_return_this_page = pp;
2234                         }
2235 
2236                         CHK_LPG(pplist, new_szc);
2237 
2238                         bin = PP_2_BIN(pplist);
2239                         if (try_to_return_this_page)
2240                                 ASSERT(mnode ==
2241                                     PP_2_MEM_NODE(try_to_return_this_page));
2242                         if ((bin == color) && (flags == PC_ALLOC) &&
2243                             (ret_pp == NULL) && try_to_return_this_page &&
2244                             page_trylock_cons(try_to_return_this_page,
2245                             SE_EXCL)) {
2246                                 ret_pp = try_to_return_this_page;
2247                         } else {
2248                                 mtype = PP_2_MTYPE(pp);
2249                                 page_vpadd(&PAGE_FREELISTS(mnode, new_szc,
2250                                     bin, mtype), pplist);
2251 
2252                                 page_ctr_add(mnode, mtype, pplist,
2253                                     PG_FREE_LIST);
2254                         }
2255                         pplist = npplist;
2256                 }
2257         }
2258         return (ret_pp);
2259 }
2260 
2261 int mpss_coalesce_disable = 0;
2262 
2263 /*
2264  * Coalesce free pages into a page of the given szc and color if possible.
2265  * Return the pointer to the page created, otherwise, return NULL.
2266  *
2267  * If pfnhi is non-zero, search for large page with pfn range less than pfnhi.
2268  */
2269 page_t *
2270 page_freelist_coalesce(int mnode, uchar_t szc, uint_t color, uint_t ceq_mask,
2271     int mtype, pfn_t pfnhi)
2272 {
2273         int     r = szc;                /* region size */
2274         int     mrange;
2275         uint_t  full, bin, color_mask, wrap = 0;
2276         pfn_t   pfnum, lo, hi;
2277         size_t  len, idx, idx0;
2278         pgcnt_t cands = 0, szcpgcnt = page_get_pagecnt(szc);
2279         page_t  *ret_pp;
2280         MEM_NODE_ITERATOR_DECL(it);
2281 #if defined(__sparc)
2282         pfn_t pfnum0, nlo, nhi;
2283 #endif
2284 
2285         if (mpss_coalesce_disable) {
2286                 ASSERT(szc < MMU_PAGE_SIZES);
2287                 VM_STAT_ADD(vmm_vmstats.page_ctrs_coalesce[szc][0]);
2288                 return (NULL);
2289         }
2290 
2291         ASSERT(szc < mmu_page_sizes);
2292         color_mask = PAGE_GET_PAGECOLORS(szc) - 1;
2293         ASSERT(ceq_mask <= color_mask);
2294         ASSERT(color <= color_mask);
2295         color &= ceq_mask;
2296 
2297         /* Prevent page_counters dynamic memory from being freed */
2298         rw_enter(&page_ctrs_rwlock[mnode], RW_READER);
2299 
2300         mrange = MTYPE_2_MRANGE(mnode, mtype);
2301         ASSERT(mrange < mnode_nranges[mnode]);
2302         VM_STAT_ADD(vmm_vmstats.page_ctrs_coalesce[r][mrange]);
2303 
2304         /* get pfn range for mtype */
2305         len = PAGE_COUNTERS_ENTRIES(mnode, r);
2306         MNODETYPE_2_PFN(mnode, mtype, lo, hi);
2307         hi++;
2308 
2309         /* use lower limit if given */
2310         if (pfnhi != PFNNULL && pfnhi < hi)
2311                 hi = pfnhi;
2312 
2313         /* round to szcpgcnt boundaries */
2314         lo = P2ROUNDUP(lo, szcpgcnt);
2315         MEM_NODE_ITERATOR_INIT(lo, mnode, szc, &it);
2316         if (lo == (pfn_t)-1) {
2317                 rw_exit(&page_ctrs_rwlock[mnode]);
2318                 return (NULL);
2319         }
2320         hi = hi & ~(szcpgcnt - 1);
2321 
2322         /* set lo to the closest pfn of the right color */
2323         if (((PFN_2_COLOR(lo, szc, &it) ^ color) & ceq_mask) ||
2324             (interleaved_mnodes && PFN_2_MEM_NODE(lo) != mnode)) {
2325                 PAGE_NEXT_PFN_FOR_COLOR(lo, szc, color, ceq_mask, color_mask,
2326                     &it);
2327         }
2328 
2329         if (hi <= lo) {
2330                 rw_exit(&page_ctrs_rwlock[mnode]);
2331                 return (NULL);
2332         }
2333 
2334         full = FULL_REGION_CNT(r);
2335 
2336         /* calculate the number of page candidates and initial search index */
2337         bin = color;
2338         idx0 = (size_t)(-1);
2339         do {
2340                 pgcnt_t acand;
2341 
2342                 PGCTRS_CANDS_GETVALUECOLOR(mnode, mrange, r, bin, acand);
2343                 if (acand) {
2344                         idx = PAGE_COUNTERS_CURRENT_COLOR(mnode,
2345                             r, bin, mrange);
2346                         idx0 = MIN(idx0, idx);
2347                         cands += acand;
2348                 }
2349                 bin = ADD_MASKED(bin, 1, ceq_mask, color_mask);
2350         } while (bin != color);
2351 
2352         if (cands == 0) {
2353                 VM_STAT_ADD(vmm_vmstats.page_ctrs_cands_skip[r][mrange]);
2354                 rw_exit(&page_ctrs_rwlock[mnode]);
2355                 return (NULL);
2356         }
2357 
2358         pfnum = IDX_TO_PNUM(mnode, r, idx0);
2359         if (pfnum < lo || pfnum >= hi) {
2360                 pfnum = lo;
2361         } else {
2362                 MEM_NODE_ITERATOR_INIT(pfnum, mnode, szc, &it);
2363                 if (pfnum == (pfn_t)-1) {
2364                         pfnum = lo;
2365                         MEM_NODE_ITERATOR_INIT(pfnum, mnode, szc, &it);
2366                         ASSERT(pfnum != (pfn_t)-1);
2367                 } else if ((PFN_2_COLOR(pfnum, szc, &it) ^ color) & ceq_mask ||
2368                     (interleaved_mnodes && PFN_2_MEM_NODE(pfnum) != mnode)) {
2369                         /* invalid color, get the closest correct pfn */
2370                         PAGE_NEXT_PFN_FOR_COLOR(pfnum, szc, color, ceq_mask,
2371                             color_mask, &it);
2372                         if (pfnum >= hi) {
2373                                 pfnum = lo;
2374                                 MEM_NODE_ITERATOR_INIT(pfnum, mnode, szc, &it);
2375                         }
2376                 }
2377         }
2378 
2379         /* set starting index */
2380         idx0 = PNUM_TO_IDX(mnode, r, pfnum);
2381         ASSERT(idx0 < len);
2382 
2383 #if defined(__sparc)
2384         pfnum0 = pfnum;         /* page corresponding to idx0 */
2385         nhi = 0;                /* search kcage ranges */
2386 #endif
2387 
2388         for (idx = idx0; wrap == 0 || (idx < idx0 && wrap < 2); ) {
2389 
2390 #if defined(__sparc)
2391                 /*
2392                  * Find lowest intersection of kcage ranges and mnode.
2393                  * MTYPE_NORELOC means look in the cage, otherwise outside.
2394                  */
2395                 if (nhi <= pfnum) {
2396                         if (kcage_next_range(mtype == MTYPE_NORELOC, pfnum,
2397                             (wrap == 0 ? hi : pfnum0), &nlo, &nhi))
2398                                 goto wrapit;
2399 
2400                         /* jump to the next page in the range */
2401                         if (pfnum < nlo) {
2402                                 pfnum = P2ROUNDUP(nlo, szcpgcnt);
2403                                 MEM_NODE_ITERATOR_INIT(pfnum, mnode, szc, &it);
2404                                 idx = PNUM_TO_IDX(mnode, r, pfnum);
2405                                 if (idx >= len || pfnum >= hi)
2406                                         goto wrapit;
2407                                 if ((PFN_2_COLOR(pfnum, szc, &it) ^ color) &
2408                                     ceq_mask)
2409                                         goto next;
2410                                 if (interleaved_mnodes &&
2411                                     PFN_2_MEM_NODE(pfnum) != mnode)
2412                                         goto next;
2413                         }
2414                 }
2415 #endif
2416 
2417                 if (PAGE_COUNTERS(mnode, r, idx) != full)
2418                         goto next;
2419 
2420                 /*
2421                  * RFE: For performance maybe we can do something less
2422                  *      brutal than locking the entire freelist. So far
2423                  *      this doesn't seem to be a performance problem?
2424                  */
2425                 page_freelist_lock(mnode);
2426                 if (PAGE_COUNTERS(mnode, r, idx) == full) {
2427                         ret_pp =
2428                             page_promote(mnode, pfnum, r, PC_ALLOC, mtype);
2429                         if (ret_pp != NULL) {
2430                                 VM_STAT_ADD(vmm_vmstats.pfc_coalok[r][mrange]);
2431                                 PAGE_COUNTERS_CURRENT_COLOR(mnode, r,
2432                                     PFN_2_COLOR(pfnum, szc, &it), mrange) = idx;
2433                                 page_freelist_unlock(mnode);
2434                                 rw_exit(&page_ctrs_rwlock[mnode]);
2435 #if defined(__sparc)
2436                                 if (PP_ISNORELOC(ret_pp)) {
2437                                         pgcnt_t npgs;
2438 
2439                                         npgs = page_get_pagecnt(ret_pp->p_szc);
2440                                         kcage_freemem_sub(npgs);
2441                                 }
2442 #endif
2443                                 return (ret_pp);
2444                         }
2445                 } else {
2446                         VM_STAT_ADD(vmm_vmstats.page_ctrs_changed[r][mrange]);
2447                 }
2448 
2449                 page_freelist_unlock(mnode);
2450                 /*
2451                  * No point looking for another page if we've
2452                  * already tried all of the ones that
2453                  * page_ctr_cands indicated.  Stash off where we left
2454                  * off.
2455                  * Note: this is not exact since we don't hold the
2456                  * page_freelist_locks before we initially get the
2457                  * value of cands for performance reasons, but should
2458                  * be a decent approximation.
2459                  */
2460                 if (--cands == 0) {
2461                         PAGE_COUNTERS_CURRENT_COLOR(mnode, r, color, mrange) =
2462                             idx;
2463                         break;
2464                 }
2465 next:
2466                 PAGE_NEXT_PFN_FOR_COLOR(pfnum, szc, color, ceq_mask,
2467                     color_mask, &it);
2468                 idx = PNUM_TO_IDX(mnode, r, pfnum);
2469                 if (idx >= len || pfnum >= hi) {
2470 wrapit:
2471                         pfnum = lo;
2472                         MEM_NODE_ITERATOR_INIT(pfnum, mnode, szc, &it);
2473                         idx = PNUM_TO_IDX(mnode, r, pfnum);
2474                         wrap++;
2475 #if defined(__sparc)
2476                         nhi = 0;        /* search kcage ranges */
2477 #endif
2478                 }
2479         }
2480 
2481         rw_exit(&page_ctrs_rwlock[mnode]);
2482         VM_STAT_ADD(vmm_vmstats.page_ctrs_failed[r][mrange]);
2483         return (NULL);
2484 }
2485 
2486 /*
2487  * For the given mnode, promote as many small pages to large pages as possible.
2488  * mnode can be -1, which means do them all
2489  */
2490 void
2491 page_freelist_coalesce_all(int mnode)
2492 {
2493         int     r;              /* region size */
2494         int     idx, full;
2495         size_t  len;
2496         int doall = interleaved_mnodes || mnode < 0;
2497         int mlo = doall ? 0 : mnode;
2498         int mhi = doall ? max_mem_nodes : (mnode + 1);
2499 
2500         VM_STAT_ADD(vmm_vmstats.page_ctrs_coalesce_all);
2501 
2502         if (mpss_coalesce_disable) {
2503                 return;
2504         }
2505 
2506         /*
2507          * Lock the entire freelist and coalesce what we can.
2508          *
2509          * Always promote to the largest page possible
2510          * first to reduce the number of page promotions.
2511          */
2512         for (mnode = mlo; mnode < mhi; mnode++) {
2513                 rw_enter(&page_ctrs_rwlock[mnode], RW_READER);
2514                 page_freelist_lock(mnode);
2515         }
2516         for (r = mmu_page_sizes - 1; r > 0; r--) {
2517                 for (mnode = mlo; mnode < mhi; mnode++) {
2518                         pgcnt_t cands = 0;
2519                         int mrange, nranges = mnode_nranges[mnode];
2520 
2521                         for (mrange = 0; mrange < nranges; mrange++) {
2522                                 PGCTRS_CANDS_GETVALUE(mnode, mrange, r, cands);
2523                                 if (cands != 0)
2524                                         break;
2525                         }
2526                         if (cands == 0) {
2527                                 VM_STAT_ADD(vmm_vmstats.
2528                                     page_ctrs_cands_skip_all);
2529                                 continue;
2530                         }
2531 
2532                         full = FULL_REGION_CNT(r);
2533                         len  = PAGE_COUNTERS_ENTRIES(mnode, r);
2534 
2535                         for (idx = 0; idx < len; idx++) {
2536                                 if (PAGE_COUNTERS(mnode, r, idx) == full) {
2537                                         pfn_t pfnum =
2538                                             IDX_TO_PNUM(mnode, r, idx);
2539                                         int tmnode = interleaved_mnodes ?
2540                                             PFN_2_MEM_NODE(pfnum) : mnode;
2541 
2542                                         ASSERT(pfnum >=
2543                                             mem_node_config[tmnode].physbase &&
2544                                             pfnum <
2545                                             mem_node_config[tmnode].physmax);
2546 
2547                                         (void) page_promote(tmnode,
2548                                             pfnum, r, PC_FREE, PC_MTYPE_ANY);
2549                                 }
2550                         }
2551                         /* shared hpm_counters covers all mnodes, so we quit */
2552                         if (interleaved_mnodes)
2553                                 break;
2554                 }
2555         }
2556         for (mnode = mlo; mnode < mhi; mnode++) {
2557                 page_freelist_unlock(mnode);
2558                 rw_exit(&page_ctrs_rwlock[mnode]);
2559         }
2560 }
2561 
2562 /*
2563  * This is where all polices for moving pages around
2564  * to different page size free lists is implemented.
2565  * Returns 1 on success, 0 on failure.
2566  *
2567  * So far these are the priorities for this algorithm in descending
2568  * order:
2569  *
2570  *      1) When servicing a request try to do so with a free page
2571  *         from next size up. Helps defer fragmentation as long
2572  *         as possible.
2573  *
2574  *      2) Page coalesce on demand. Only when a freelist
2575  *         larger than PAGESIZE is empty and step 1
2576  *         will not work since all larger size lists are
2577  *         also empty.
2578  *
2579  * If pfnhi is non-zero, search for large page with pfn range less than pfnhi.
2580  */
2581 
2582 page_t *
2583 page_freelist_split(uchar_t szc, uint_t color, int mnode, int mtype,
2584     pfn_t pfnlo, pfn_t pfnhi, page_list_walker_t *plw)
2585 {
2586         uchar_t nszc = szc + 1;
2587         uint_t  bin, sbin, bin_prev;
2588         page_t  *pp, *firstpp;
2589         page_t  *ret_pp = NULL;
2590         uint_t  color_mask;
2591 
2592         if (nszc == mmu_page_sizes)
2593                 return (NULL);
2594 
2595         ASSERT(nszc < mmu_page_sizes);
2596         color_mask = PAGE_GET_PAGECOLORS(nszc) - 1;
2597         bin = sbin = PAGE_GET_NSZ_COLOR(szc, color);
2598         bin_prev = (plw->plw_bin_split_prev == color) ? INVALID_COLOR :
2599             PAGE_GET_NSZ_COLOR(szc, plw->plw_bin_split_prev);
2600 
2601         VM_STAT_ADD(vmm_vmstats.pfs_req[szc]);
2602         /*
2603          * First try to break up a larger page to fill current size freelist.
2604          */
2605         while (plw->plw_bins[nszc] != 0) {
2606 
2607                 ASSERT(nszc < mmu_page_sizes);
2608 
2609                 /*
2610                  * If page found then demote it.
2611                  */
2612                 if (PAGE_FREELISTS(mnode, nszc, bin, mtype)) {
2613                         page_freelist_lock(mnode);
2614                         firstpp = pp = PAGE_FREELISTS(mnode, nszc, bin, mtype);
2615 
2616                         /*
2617                          * If pfnhi is not PFNNULL, look for large page below
2618                          * pfnhi. PFNNULL signifies no pfn requirement.
2619                          */
2620                         if (pp &&
2621                             ((pfnhi != PFNNULL && pp->p_pagenum >= pfnhi) ||
2622                             (pfnlo != PFNNULL && pp->p_pagenum < pfnlo))) {
2623                                 do {
2624                                         pp = pp->p_vpnext;
2625                                         if (pp == firstpp) {
2626                                                 pp = NULL;
2627                                                 break;
2628                                         }
2629                                 } while ((pfnhi != PFNNULL &&
2630                                     pp->p_pagenum >= pfnhi) ||
2631                                     (pfnlo != PFNNULL &&
2632                                     pp->p_pagenum < pfnlo));
2633 
2634                                 if (pfnhi != PFNNULL && pp != NULL)
2635                                         ASSERT(pp->p_pagenum < pfnhi);
2636 
2637                                 if (pfnlo != PFNNULL && pp != NULL)
2638                                         ASSERT(pp->p_pagenum >= pfnlo);
2639                         }
2640                         if (pp) {
2641                                 uint_t ccolor = page_correct_color(szc, nszc,
2642                                     color, bin, plw->plw_ceq_mask[szc]);
2643 
2644                                 ASSERT(pp->p_szc == nszc);
2645                                 VM_STAT_ADD(vmm_vmstats.pfs_demote[nszc]);
2646                                 ret_pp = page_demote(mnode, pp->p_pagenum,
2647                                     pfnhi, pp->p_szc, szc, ccolor, PC_ALLOC);
2648                                 if (ret_pp) {
2649                                         page_freelist_unlock(mnode);
2650 #if defined(__sparc)
2651                                         if (PP_ISNORELOC(ret_pp)) {
2652                                                 pgcnt_t npgs;
2653 
2654                                                 npgs = page_get_pagecnt(
2655                                                     ret_pp->p_szc);
2656                                                 kcage_freemem_sub(npgs);
2657                                         }
2658 #endif
2659                                         return (ret_pp);
2660                                 }
2661                         }
2662                         page_freelist_unlock(mnode);
2663                 }
2664 
2665                 /* loop through next size bins */
2666                 bin = ADD_MASKED(bin, 1, plw->plw_ceq_mask[nszc], color_mask);
2667                 plw->plw_bins[nszc]--;
2668 
2669                 if (bin == sbin) {
2670                         uchar_t nnszc = nszc + 1;
2671 
2672                         /* we are done with this page size - check next */
2673                         if (plw->plw_bins[nnszc] == 0)
2674                                 /* we have already checked next size bins */
2675                                 break;
2676 
2677                         bin = sbin = PAGE_GET_NSZ_COLOR(nszc, bin);
2678                         if (bin_prev != INVALID_COLOR) {
2679                                 bin_prev = PAGE_GET_NSZ_COLOR(nszc, bin_prev);
2680                                 if (!((bin ^ bin_prev) &
2681                                     plw->plw_ceq_mask[nnszc]))
2682                                         break;
2683                         }
2684                         ASSERT(nnszc < mmu_page_sizes);
2685                         color_mask = PAGE_GET_PAGECOLORS(nnszc) - 1;
2686                         nszc = nnszc;
2687                         ASSERT(nszc < mmu_page_sizes);
2688                 }
2689         }
2690 
2691         return (ret_pp);
2692 }
2693 
2694 /*
2695  * Helper routine used only by the freelist code to lock
2696  * a page. If the page is a large page then it succeeds in
2697  * locking all the constituent pages or none at all.
2698  * Returns 1 on sucess, 0 on failure.
2699  */
2700 static int
2701 page_trylock_cons(page_t *pp, se_t se)
2702 {
2703         page_t  *tpp, *first_pp = pp;
2704 
2705         /*
2706          * Fail if can't lock first or only page.
2707          */
2708         if (!page_trylock(pp, se)) {
2709                 return (0);
2710         }
2711 
2712         /*
2713          * PAGESIZE: common case.
2714          */
2715         if (pp->p_szc == 0) {
2716                 return (1);
2717         }
2718 
2719         /*
2720          * Large page case.
2721          */
2722         tpp = pp->p_next;
2723         while (tpp != pp) {
2724                 if (!page_trylock(tpp, se)) {
2725                         /*
2726                          * On failure unlock what we have locked so far.
2727                          * We want to avoid attempting to capture these
2728                          * pages as the pcm mutex may be held which could
2729                          * lead to a recursive mutex panic.
2730                          */
2731                         while (first_pp != tpp) {
2732                                 page_unlock_nocapture(first_pp);
2733                                 first_pp = first_pp->p_next;
2734                         }
2735                         return (0);
2736                 }
2737                 tpp = tpp->p_next;
2738         }
2739         return (1);
2740 }
2741 
2742 /*
2743  * init context for walking page lists
2744  * Called when a page of the given szc in unavailable. Sets markers
2745  * for the beginning of the search to detect when search has
2746  * completed a full cycle. Sets flags for splitting larger pages
2747  * and coalescing smaller pages. Page walking procedes until a page
2748  * of the desired equivalent color is found.
2749  */
2750 void
2751 page_list_walk_init(uchar_t szc, uint_t flags, uint_t bin, int can_split,
2752     int use_ceq, page_list_walker_t *plw)
2753 {
2754         uint_t  nszc, ceq_mask, colors;
2755         uchar_t ceq = use_ceq ? colorequivszc[szc] : 0;
2756 
2757         ASSERT(szc < mmu_page_sizes);
2758         colors = PAGE_GET_PAGECOLORS(szc);
2759 
2760         plw->plw_colors = colors;
2761         plw->plw_color_mask = colors - 1;
2762         plw->plw_bin_marker = plw->plw_bin0 = bin;
2763         plw->plw_bin_split_prev = bin;
2764         plw->plw_bin_step = (szc == 0) ? vac_colors : 1;
2765 
2766         /*
2767          * if vac aliasing is possible make sure lower order color
2768          * bits are never ignored
2769          */
2770         if (vac_colors > 1)
2771                 ceq &= 0xf0;
2772 
2773         /*
2774          * calculate the number of non-equivalent colors and
2775          * color equivalency mask
2776          */
2777         plw->plw_ceq_dif = colors >> ((ceq >> 4) + (ceq & 0xf));
2778         ASSERT(szc > 0 || plw->plw_ceq_dif >= vac_colors);
2779         ASSERT(plw->plw_ceq_dif > 0);
2780         plw->plw_ceq_mask[szc] = (plw->plw_ceq_dif - 1) << (ceq & 0xf);
2781 
2782         if (flags & PG_MATCH_COLOR) {
2783                 if (cpu_page_colors <  0) {
2784                         /*
2785                          * this is a heterogeneous machine with different CPUs
2786                          * having different size e$ (not supported for ni2/rock
2787                          */
2788                         uint_t cpucolors = CPUSETSIZE() >> PAGE_GET_SHIFT(szc);
2789                         cpucolors = MAX(cpucolors, 1);
2790                         ceq_mask = plw->plw_color_mask & (cpucolors - 1);
2791                         plw->plw_ceq_mask[szc] =
2792                             MIN(ceq_mask, plw->plw_ceq_mask[szc]);
2793                 }
2794                 plw->plw_ceq_dif = 1;
2795         }
2796 
2797         /* we can split pages in the freelist, but not the cachelist */
2798         if (can_split) {
2799                 plw->plw_do_split = (szc + 1 < mmu_page_sizes) ? 1 : 0;
2800 
2801                 /* set next szc color masks and number of free list bins */
2802                 for (nszc = szc + 1; nszc < mmu_page_sizes; nszc++, szc++) {
2803                         plw->plw_ceq_mask[nszc] = PAGE_GET_NSZ_MASK(szc,
2804                             plw->plw_ceq_mask[szc]);
2805                         plw->plw_bins[nszc] = PAGE_GET_PAGECOLORS(nszc);
2806                 }
2807                 plw->plw_ceq_mask[nszc] = INVALID_MASK;
2808                 plw->plw_bins[nszc] = 0;
2809 
2810         } else {
2811                 ASSERT(szc == 0);
2812                 plw->plw_do_split = 0;
2813                 plw->plw_bins[1] = 0;
2814                 plw->plw_ceq_mask[1] = INVALID_MASK;
2815         }
2816 }
2817 
2818 /*
2819  * set mark to flag where next split should occur
2820  */
2821 #define PAGE_SET_NEXT_SPLIT_MARKER(szc, nszc, bin, plw) {                    \
2822         uint_t bin_nsz = PAGE_GET_NSZ_COLOR(szc, bin);                       \
2823         uint_t bin0_nsz = PAGE_GET_NSZ_COLOR(szc, plw->plw_bin0);         \
2824         uint_t neq_mask = ~plw->plw_ceq_mask[nszc] & plw->plw_color_mask;    \
2825         plw->plw_split_next =                                                     \
2826                 INC_MASKED(bin_nsz, neq_mask, plw->plw_color_mask);       \
2827         if (!((plw->plw_split_next ^ bin0_nsz) & plw->plw_ceq_mask[nszc])) { \
2828                 plw->plw_split_next =                                             \
2829                 INC_MASKED(plw->plw_split_next,                                   \
2830                     neq_mask, plw->plw_color_mask);                       \
2831         }                                                                    \
2832 }
2833 
2834 uint_t
2835 page_list_walk_next_bin(uchar_t szc, uint_t bin, page_list_walker_t *plw)
2836 {
2837         uint_t  neq_mask = ~plw->plw_ceq_mask[szc] & plw->plw_color_mask;
2838         uint_t  bin0_nsz, nbin_nsz, nbin0, nbin;
2839         uchar_t nszc = szc + 1;
2840 
2841         nbin = ADD_MASKED(bin,
2842             plw->plw_bin_step, neq_mask, plw->plw_color_mask);
2843 
2844         if (plw->plw_do_split) {
2845                 plw->plw_bin_split_prev = bin;
2846                 PAGE_SET_NEXT_SPLIT_MARKER(szc, nszc, bin, plw);
2847                 plw->plw_do_split = 0;
2848         }
2849 
2850         if (szc == 0) {
2851                 if (plw->plw_count != 0 || plw->plw_ceq_dif == vac_colors) {
2852                         if (nbin == plw->plw_bin0 &&
2853                             (vac_colors == 1 || nbin != plw->plw_bin_marker)) {
2854                                 nbin = ADD_MASKED(nbin, plw->plw_bin_step,
2855                                     neq_mask, plw->plw_color_mask);
2856                                 plw->plw_bin_split_prev = plw->plw_bin0;
2857                         }
2858 
2859                         if (vac_colors > 1 && nbin == plw->plw_bin_marker) {
2860                                 plw->plw_bin_marker =
2861                                     nbin = INC_MASKED(nbin, neq_mask,
2862                                     plw->plw_color_mask);
2863                                 plw->plw_bin_split_prev = plw->plw_bin0;
2864                                 /*
2865                                  * large pages all have the same vac color
2866                                  * so by now we should be done with next
2867                                  * size page splitting process
2868                                  */
2869                                 ASSERT(plw->plw_bins[1] == 0);
2870                                 plw->plw_do_split = 0;
2871                                 return (nbin);
2872                         }
2873 
2874                 } else {
2875                         uint_t bin_jump = (vac_colors == 1) ?
2876                             (BIN_STEP & ~3) - (plw->plw_bin0 & 3) : BIN_STEP;
2877 
2878                         bin_jump &= ~(vac_colors - 1);
2879 
2880                         nbin0 = ADD_MASKED(plw->plw_bin0, bin_jump, neq_mask,
2881                             plw->plw_color_mask);
2882 
2883                         if ((nbin0 ^ plw->plw_bin0) & plw->plw_ceq_mask[szc]) {
2884 
2885                                 plw->plw_bin_marker = nbin = nbin0;
2886 
2887                                 if (plw->plw_bins[nszc] != 0) {
2888                                         /*
2889                                          * check if next page size bin is the
2890                                          * same as the next page size bin for
2891                                          * bin0
2892                                          */
2893                                         nbin_nsz = PAGE_GET_NSZ_COLOR(szc,
2894                                             nbin);
2895                                         bin0_nsz = PAGE_GET_NSZ_COLOR(szc,
2896                                             plw->plw_bin0);
2897 
2898                                         if ((bin0_nsz ^ nbin_nsz) &
2899                                             plw->plw_ceq_mask[nszc])
2900                                                 plw->plw_do_split = 1;
2901                                 }
2902                                 return (nbin);
2903                         }
2904                 }
2905         }
2906 
2907         if (plw->plw_bins[nszc] != 0) {
2908                 nbin_nsz = PAGE_GET_NSZ_COLOR(szc, nbin);
2909                 if (!((plw->plw_split_next ^ nbin_nsz) &
2910                     plw->plw_ceq_mask[nszc]))
2911                         plw->plw_do_split = 1;
2912         }
2913 
2914         return (nbin);
2915 }
2916 
2917 page_t *
2918 page_get_mnode_freelist(int mnode, uint_t bin, int mtype, uchar_t szc,
2919     uint_t flags)
2920 {
2921         kmutex_t                *pcm;
2922         page_t                  *pp, *first_pp;
2923         uint_t                  sbin;
2924         int                     plw_initialized;
2925         page_list_walker_t      plw;
2926 
2927         ASSERT(szc < mmu_page_sizes);
2928 
2929         VM_STAT_ADD(vmm_vmstats.pgmf_alloc[szc]);
2930 
2931         MTYPE_START(mnode, mtype, flags);
2932         if (mtype < 0) {     /* mnode does not have memory in mtype range */
2933                 VM_STAT_ADD(vmm_vmstats.pgmf_allocempty[szc]);
2934                 return (NULL);
2935         }
2936 try_again:
2937 
2938         plw_initialized = 0;
2939         plw.plw_ceq_dif = 1;
2940 
2941         /*
2942          * Only hold one freelist lock at a time, that way we
2943          * can start anywhere and not have to worry about lock
2944          * ordering.
2945          */
2946         for (plw.plw_count = 0;
2947             plw.plw_count < plw.plw_ceq_dif; plw.plw_count++) {
2948                 sbin = bin;
2949                 do {
2950                         if (!PAGE_FREELISTS(mnode, szc, bin, mtype))
2951                                 goto bin_empty_1;
2952 
2953                         pcm = PC_BIN_MUTEX(mnode, bin, PG_FREE_LIST);
2954                         mutex_enter(pcm);
2955                         pp = PAGE_FREELISTS(mnode, szc, bin, mtype);
2956                         if (pp == NULL)
2957                                 goto bin_empty_0;
2958 
2959                         /*
2960                          * These were set before the page
2961                          * was put on the free list,
2962                          * they must still be set.
2963                          */
2964                         ASSERT(PP_ISFREE(pp));
2965                         ASSERT(PP_ISAGED(pp));
2966                         ASSERT(pp->p_vnode == NULL);
2967                         ASSERT(pp->p_hash == NULL);
2968                         ASSERT(pp->p_offset == (u_offset_t)-1);
2969                         ASSERT(pp->p_szc == szc);
2970                         ASSERT(PFN_2_MEM_NODE(pp->p_pagenum) == mnode);
2971 
2972                         /*
2973                          * Walk down the hash chain.
2974                          * 8k pages are linked on p_next
2975                          * and p_prev fields. Large pages
2976                          * are a contiguous group of
2977                          * constituent pages linked together
2978                          * on their p_next and p_prev fields.
2979                          * The large pages are linked together
2980                          * on the hash chain using p_vpnext
2981                          * p_vpprev of the base constituent
2982                          * page of each large page.
2983                          */
2984                         first_pp = pp;
2985                         while (IS_DUMP_PAGE(pp) || !page_trylock_cons(pp,
2986                             SE_EXCL)) {
2987                                 if (szc == 0) {
2988                                         pp = pp->p_next;
2989                                 } else {
2990                                         pp = pp->p_vpnext;
2991                                 }
2992 
2993                                 ASSERT(PP_ISFREE(pp));
2994                                 ASSERT(PP_ISAGED(pp));
2995                                 ASSERT(pp->p_vnode == NULL);
2996                                 ASSERT(pp->p_hash == NULL);
2997                                 ASSERT(pp->p_offset == (u_offset_t)-1);
2998                                 ASSERT(pp->p_szc == szc);
2999                                 ASSERT(PFN_2_MEM_NODE(pp->p_pagenum) == mnode);
3000 
3001                                 if (pp == first_pp)
3002                                         goto bin_empty_0;
3003                         }
3004 
3005                         ASSERT(pp != NULL);
3006                         ASSERT(mtype == PP_2_MTYPE(pp));
3007                         ASSERT(pp->p_szc == szc);
3008                         if (szc == 0) {
3009                                 page_sub(&PAGE_FREELISTS(mnode,
3010                                     szc, bin, mtype), pp);
3011                         } else {
3012                                 page_vpsub(&PAGE_FREELISTS(mnode,
3013                                     szc, bin, mtype), pp);
3014                                 CHK_LPG(pp, szc);
3015                         }
3016                         page_ctr_sub(mnode, mtype, pp, PG_FREE_LIST);
3017 
3018                         if ((PP_ISFREE(pp) == 0) || (PP_ISAGED(pp) == 0))
3019                                 panic("free page is not. pp %p", (void *)pp);
3020                         mutex_exit(pcm);
3021 
3022 #if defined(__sparc)
3023                         ASSERT(!kcage_on || PP_ISNORELOC(pp) ||
3024                             (flags & PG_NORELOC) == 0);
3025 
3026                         if (PP_ISNORELOC(pp))
3027                                 kcage_freemem_sub(page_get_pagecnt(szc));
3028 #endif
3029                         VM_STAT_ADD(vmm_vmstats.pgmf_allocok[szc]);
3030                         return (pp);
3031 
3032 bin_empty_0:
3033                         mutex_exit(pcm);
3034 bin_empty_1:
3035                         if (plw_initialized == 0) {
3036                                 page_list_walk_init(szc, flags, bin, 1, 1,
3037                                     &plw);
3038                                 plw_initialized = 1;
3039                                 ASSERT(plw.plw_colors <=
3040                                     PAGE_GET_PAGECOLORS(szc));
3041                                 ASSERT(plw.plw_colors > 0);
3042                                 ASSERT((plw.plw_colors &
3043                                     (plw.plw_colors - 1)) == 0);
3044                                 ASSERT(bin < plw.plw_colors);
3045                                 ASSERT(plw.plw_ceq_mask[szc] < plw.plw_colors);
3046                         }
3047                         /* calculate the next bin with equivalent color */
3048                         bin = ADD_MASKED(bin, plw.plw_bin_step,
3049                             plw.plw_ceq_mask[szc], plw.plw_color_mask);
3050                 } while (sbin != bin);
3051 
3052                 /*
3053                  * color bins are all empty if color match. Try and
3054                  * satisfy the request by breaking up or coalescing
3055                  * pages from a different size freelist of the correct
3056                  * color that satisfies the ORIGINAL color requested.
3057                  * If that fails then try pages of the same size but
3058                  * different colors assuming we are not called with
3059                  * PG_MATCH_COLOR.
3060                  */
3061                 if (plw.plw_do_split &&
3062                     (pp = page_freelist_split(szc, bin, mnode,
3063                     mtype, PFNNULL, PFNNULL, &plw)) != NULL)
3064                         return (pp);
3065 
3066                 if (szc > 0 && (pp = page_freelist_coalesce(mnode, szc,
3067                     bin, plw.plw_ceq_mask[szc], mtype, PFNNULL)) !=  NULL)
3068                         return (pp);
3069 
3070                 if (plw.plw_ceq_dif > 1)
3071                         bin = page_list_walk_next_bin(szc, bin, &plw);
3072         }
3073 
3074         /* if allowed, cycle through additional mtypes */
3075         MTYPE_NEXT(mnode, mtype, flags);
3076         if (mtype >= 0)
3077                 goto try_again;
3078 
3079         VM_STAT_ADD(vmm_vmstats.pgmf_allocfailed[szc]);
3080 
3081         return (NULL);
3082 }
3083 
3084 /*
3085  * Returns the count of free pages for 'pp' with size code 'szc'.
3086  * Note: This function does not return an exact value as the page freelist
3087  * locks are not held and thus the values in the page_counters may be
3088  * changing as we walk through the data.
3089  */
3090 static int
3091 page_freecnt(int mnode, page_t *pp, uchar_t szc)
3092 {
3093         pgcnt_t pgfree;
3094         pgcnt_t cnt;
3095         ssize_t r = szc;        /* region size */
3096         ssize_t idx;
3097         int     i;
3098         int     full, range;
3099 
3100         /* Make sure pagenum passed in is aligned properly */
3101         ASSERT((pp->p_pagenum & (PNUM_SIZE(szc) - 1)) == 0);
3102         ASSERT(szc > 0);
3103 
3104         /* Prevent page_counters dynamic memory from being freed */
3105         rw_enter(&page_ctrs_rwlock[mnode], RW_READER);
3106         idx = PNUM_TO_IDX(mnode, r, pp->p_pagenum);
3107         cnt = PAGE_COUNTERS(mnode, r, idx);
3108         pgfree = cnt << PNUM_SHIFT(r - 1);
3109         range = FULL_REGION_CNT(szc);
3110 
3111         /* Check for completely full region */
3112         if (cnt == range) {
3113                 rw_exit(&page_ctrs_rwlock[mnode]);
3114                 return (pgfree);
3115         }
3116 
3117         while (--r > 0) {
3118                 idx = PNUM_TO_IDX(mnode, r, pp->p_pagenum);
3119                 full = FULL_REGION_CNT(r);
3120                 for (i = 0; i < range; i++, idx++) {
3121                         cnt = PAGE_COUNTERS(mnode, r, idx);
3122                         /*
3123                          * If cnt here is full, that means we have already
3124                          * accounted for these pages earlier.
3125                          */
3126                         if (cnt != full) {
3127                                 pgfree += (cnt << PNUM_SHIFT(r - 1));
3128                         }
3129                 }
3130                 range *= full;
3131         }
3132         rw_exit(&page_ctrs_rwlock[mnode]);
3133         return (pgfree);
3134 }
3135 
3136 /*
3137  * Called from page_geti_contig_pages to exclusively lock constituent pages
3138  * starting from 'spp' for page size code 'szc'.
3139  *
3140  * If 'ptcpthreshold' is set, the number of free pages needed in the 'szc'
3141  * region needs to be greater than or equal to the threshold.
3142  */
3143 static int
3144 page_trylock_contig_pages(int mnode, page_t *spp, uchar_t szc, int flags)
3145 {
3146         pgcnt_t pgcnt = PNUM_SIZE(szc);
3147         pgcnt_t pgfree, i;
3148         page_t *pp;
3149 
3150         VM_STAT_ADD(vmm_vmstats.ptcp[szc]);
3151 
3152 
3153         if ((ptcpthreshold == 0) || (flags & PGI_PGCPHIPRI))
3154                 goto skipptcpcheck;
3155         /*
3156          * check if there are sufficient free pages available before attempting
3157          * to trylock. Count is approximate as page counters can change.
3158          */
3159         pgfree = page_freecnt(mnode, spp, szc);
3160 
3161         /* attempt to trylock if there are sufficient already free pages */
3162         if (pgfree < pgcnt/ptcpthreshold) {
3163                 VM_STAT_ADD(vmm_vmstats.ptcpfreethresh[szc]);
3164                 return (0);
3165         }
3166 
3167 skipptcpcheck:
3168 
3169         for (i = 0; i < pgcnt; i++) {
3170                 pp = &spp[i];
3171                 if (!page_trylock(pp, SE_EXCL)) {
3172                         VM_STAT_ADD(vmm_vmstats.ptcpfailexcl[szc]);
3173                         while (--i != (pgcnt_t)-1) {
3174                                 pp = &spp[i];
3175                                 ASSERT(PAGE_EXCL(pp));
3176                                 page_unlock_nocapture(pp);
3177                         }
3178                         return (0);
3179                 }
3180                 ASSERT(spp[i].p_pagenum == spp->p_pagenum + i);
3181                 if ((pp->p_szc > szc || (szc && pp->p_szc == szc)) &&
3182                     !PP_ISFREE(pp)) {
3183                         VM_STAT_ADD(vmm_vmstats.ptcpfailszc[szc]);
3184                         ASSERT(i == 0);
3185                         page_unlock_nocapture(pp);
3186                         return (0);
3187                 }
3188 
3189                 /*
3190                  * If a page has been marked non-relocatable or has been
3191                  * explicitly locked in memory, we don't want to relocate it;
3192                  * unlock the pages and fail the operation.
3193                  */
3194                 if (PP_ISNORELOC(pp) ||
3195                     pp->p_lckcnt != 0 || pp->p_cowcnt != 0) {
3196                         VM_STAT_ADD(vmm_vmstats.ptcpfailcage[szc]);
3197                         while (i != (pgcnt_t)-1) {
3198                                 pp = &spp[i];
3199                                 ASSERT(PAGE_EXCL(pp));
3200                                 page_unlock_nocapture(pp);
3201                                 i--;
3202                         }
3203                         return (0);
3204                 }
3205         }
3206         VM_STAT_ADD(vmm_vmstats.ptcpok[szc]);
3207         return (1);
3208 }
3209 
3210 /*
3211  * Claim large page pointed to by 'pp'. 'pp' is the starting set
3212  * of 'szc' constituent pages that had been locked exclusively previously.
3213  * Will attempt to relocate constituent pages in use.
3214  */
3215 static page_t *
3216 page_claim_contig_pages(page_t *pp, uchar_t szc, int flags)
3217 {
3218         spgcnt_t pgcnt, npgs, i;
3219         page_t *targpp, *rpp, *hpp;
3220         page_t *replpp = NULL;
3221         page_t *pplist = NULL;
3222 
3223         ASSERT(pp != NULL);
3224 
3225         pgcnt = page_get_pagecnt(szc);
3226         while (pgcnt) {
3227                 ASSERT(PAGE_EXCL(pp));
3228                 ASSERT(!PP_ISNORELOC(pp));
3229                 if (PP_ISFREE(pp)) {
3230                         /*
3231                          * If this is a PG_FREE_LIST page then its
3232                          * size code can change underneath us due to
3233                          * page promotion or demotion. As an optimzation
3234                          * use page_list_sub_pages() instead of
3235                          * page_list_sub().
3236                          */
3237                         if (PP_ISAGED(pp)) {
3238                                 page_list_sub_pages(pp, szc);
3239                                 if (pp->p_szc == szc) {
3240                                         return (pp);
3241                                 }
3242                                 ASSERT(pp->p_szc < szc);
3243                                 npgs = page_get_pagecnt(pp->p_szc);
3244                                 hpp = pp;
3245                                 for (i = 0; i < npgs; i++, pp++) {
3246                                         pp->p_szc = szc;
3247                                 }
3248                                 page_list_concat(&pplist, &hpp);
3249                                 pgcnt -= npgs;
3250                                 continue;
3251                         }
3252                         ASSERT(!PP_ISAGED(pp));
3253                         ASSERT(pp->p_szc == 0);
3254                         page_list_sub(pp, PG_CACHE_LIST);
3255                         page_hashout(pp, NULL);
3256                         PP_SETAGED(pp);
3257                         pp->p_szc = szc;
3258                         page_list_concat(&pplist, &pp);
3259                         pp++;
3260                         pgcnt--;
3261                         continue;
3262                 }
3263                 npgs = page_get_pagecnt(pp->p_szc);
3264 
3265                 /*
3266                  * page_create_wait freemem accounting done by caller of
3267                  * page_get_freelist and not necessary to call it prior to
3268                  * calling page_get_replacement_page.
3269                  *
3270                  * page_get_replacement_page can call page_get_contig_pages
3271                  * to acquire a large page (szc > 0); the replacement must be
3272                  * smaller than the contig page size to avoid looping or
3273                  * szc == 0 and PGI_PGCPSZC0 is set.
3274                  */
3275                 if (pp->p_szc < szc || (szc == 0 && (flags & PGI_PGCPSZC0))) {
3276                         replpp = page_get_replacement_page(pp, NULL, 0);
3277                         if (replpp) {
3278                                 npgs = page_get_pagecnt(pp->p_szc);
3279                                 ASSERT(npgs <= pgcnt);
3280                                 targpp = pp;
3281                         }
3282                 }
3283 
3284                 /*
3285                  * If replacement is NULL or do_page_relocate fails, fail
3286                  * coalescing of pages.
3287                  */
3288                 if (replpp == NULL || (do_page_relocate(&targpp, &replpp, 0,
3289                     &npgs, NULL) != 0)) {
3290                         /*
3291                          * Unlock un-processed target list
3292                          */
3293                         while (pgcnt--) {
3294                                 ASSERT(PAGE_EXCL(pp));
3295                                 page_unlock_nocapture(pp);
3296                                 pp++;
3297                         }
3298                         /*
3299                          * Free the processed target list.
3300                          */
3301                         while (pplist) {
3302                                 pp = pplist;
3303                                 page_sub(&pplist, pp);
3304                                 ASSERT(PAGE_EXCL(pp));
3305                                 ASSERT(pp->p_szc == szc);
3306                                 ASSERT(PP_ISFREE(pp));
3307                                 ASSERT(PP_ISAGED(pp));
3308                                 pp->p_szc = 0;
3309                                 page_list_add(pp, PG_FREE_LIST | PG_LIST_TAIL);
3310                                 page_unlock_nocapture(pp);
3311                         }
3312 
3313                         if (replpp != NULL)
3314                                 page_free_replacement_page(replpp);
3315 
3316                         return (NULL);
3317                 }
3318                 ASSERT(pp == targpp);
3319 
3320                 /* LINTED */
3321                 ASSERT(hpp = pp); /* That's right, it's an assignment */
3322 
3323                 pp += npgs;
3324                 pgcnt -= npgs;
3325 
3326                 while (npgs--) {
3327                         ASSERT(PAGE_EXCL(targpp));
3328                         ASSERT(!PP_ISFREE(targpp));
3329                         ASSERT(!PP_ISNORELOC(targpp));
3330                         PP_SETFREE(targpp);
3331                         ASSERT(PP_ISAGED(targpp));
3332                         ASSERT(targpp->p_szc < szc || (szc == 0 &&
3333                             (flags & PGI_PGCPSZC0)));
3334                         targpp->p_szc = szc;
3335                         targpp = targpp->p_next;
3336 
3337                         rpp = replpp;
3338                         ASSERT(rpp != NULL);
3339                         page_sub(&replpp, rpp);
3340                         ASSERT(PAGE_EXCL(rpp));
3341                         ASSERT(!PP_ISFREE(rpp));
3342                         page_unlock_nocapture(rpp);
3343                 }
3344                 ASSERT(targpp == hpp);
3345                 ASSERT(replpp == NULL);
3346                 page_list_concat(&pplist, &targpp);
3347         }
3348         CHK_LPG(pplist, szc);
3349         return (pplist);
3350 }
3351 
3352 /*
3353  * Trim kernel cage from pfnlo-pfnhi and store result in lo-hi. Return code
3354  * of 0 means nothing left after trim.
3355  */
3356 int
3357 trimkcage(struct memseg *mseg, pfn_t *lo, pfn_t *hi, pfn_t pfnlo, pfn_t pfnhi)
3358 {
3359         pfn_t   kcagepfn;
3360         int     decr;
3361         int     rc = 0;
3362 
3363         if (PP_ISNORELOC(mseg->pages)) {
3364                 if (PP_ISNORELOC(mseg->epages - 1) == 0) {
3365 
3366                         /* lower part of this mseg inside kernel cage */
3367                         decr = kcage_current_pfn(&kcagepfn);
3368 
3369                         /* kernel cage may have transitioned past mseg */
3370                         if (kcagepfn >= mseg->pages_base &&
3371                             kcagepfn < mseg->pages_end) {
3372                                 ASSERT(decr == 0);
3373                                 *lo = MAX(kcagepfn, pfnlo);
3374                                 *hi = MIN(pfnhi, (mseg->pages_end - 1));
3375                                 rc = 1;
3376                         }
3377                 }
3378                 /* else entire mseg in the cage */
3379         } else {
3380                 if (PP_ISNORELOC(mseg->epages - 1)) {
3381 
3382                         /* upper part of this mseg inside kernel cage */
3383                         decr = kcage_current_pfn(&kcagepfn);
3384 
3385                         /* kernel cage may have transitioned past mseg */
3386                         if (kcagepfn >= mseg->pages_base &&
3387                             kcagepfn < mseg->pages_end) {
3388                                 ASSERT(decr);
3389                                 *hi = MIN(kcagepfn, pfnhi);
3390                                 *lo = MAX(pfnlo, mseg->pages_base);
3391                                 rc = 1;
3392                         }
3393                 } else {
3394                         /* entire mseg outside of kernel cage */
3395                         *lo = MAX(pfnlo, mseg->pages_base);
3396                         *hi = MIN(pfnhi, (mseg->pages_end - 1));
3397                         rc = 1;
3398                 }
3399         }
3400         return (rc);
3401 }
3402 
3403 /*
3404  * called from page_get_contig_pages to search 'pfnlo' thru 'pfnhi' to claim a
3405  * page with size code 'szc'. Claiming such a page requires acquiring
3406  * exclusive locks on all constituent pages (page_trylock_contig_pages),
3407  * relocating pages in use and concatenating these constituent pages into a
3408  * large page.
3409  *
3410  * The page lists do not have such a large page and page_freelist_split has
3411  * already failed to demote larger pages and/or coalesce smaller free pages.
3412  *
3413  * 'flags' may specify PG_COLOR_MATCH which would limit the search of large
3414  * pages with the same color as 'bin'.
3415  *
3416  * 'pfnflag' specifies the subset of the pfn range to search.
3417  */
3418 
3419 static page_t *
3420 page_geti_contig_pages(int mnode, uint_t bin, uchar_t szc, int flags,
3421     pfn_t pfnlo, pfn_t pfnhi, pgcnt_t pfnflag)
3422 {
3423         struct memseg *mseg;
3424         pgcnt_t szcpgcnt = page_get_pagecnt(szc);
3425         pgcnt_t szcpgmask = szcpgcnt - 1;
3426         pfn_t   randpfn;
3427         page_t *pp, *randpp, *endpp;
3428         uint_t colors, ceq_mask;
3429         /* LINTED : set but not used in function */
3430         uint_t color_mask __unused;
3431         pfn_t hi, lo;
3432         uint_t skip;
3433         MEM_NODE_ITERATOR_DECL(it);
3434 
3435         ASSERT(szc != 0 || (flags & PGI_PGCPSZC0));
3436 
3437         pfnlo = P2ROUNDUP(pfnlo, szcpgcnt);
3438 
3439         if ((pfnhi - pfnlo) + 1 < szcpgcnt || pfnlo >= pfnhi)
3440                 return (NULL);
3441 
3442         ASSERT(szc < mmu_page_sizes);
3443 
3444         colors = PAGE_GET_PAGECOLORS(szc);
3445         color_mask = colors - 1;
3446         if ((colors > 1) && (flags & PG_MATCH_COLOR)) {
3447                 uchar_t ceq = colorequivszc[szc];
3448                 uint_t  ceq_dif = colors >> ((ceq >> 4) + (ceq & 0xf));
3449 
3450                 ASSERT(ceq_dif > 0);
3451                 ceq_mask = (ceq_dif - 1) << (ceq & 0xf);
3452         } else {
3453                 ceq_mask = 0;
3454         }
3455 
3456         ASSERT(bin < colors);
3457 
3458         /* clear "non-significant" color bits */
3459         bin &= ceq_mask;
3460 
3461         /*
3462          * trim the pfn range to search based on pfnflag. pfnflag is set
3463          * when there have been previous page_get_contig_page failures to
3464          * limit the search.
3465          *
3466          * The high bit in pfnflag specifies the number of 'slots' in the
3467          * pfn range and the remainder of pfnflag specifies which slot.
3468          * For example, a value of 1010b would mean the second slot of
3469          * the pfn range that has been divided into 8 slots.
3470          */
3471         if (pfnflag > 1) {
3472                 int     slots = 1 << (highbit(pfnflag) - 1);
3473                 int     slotid = pfnflag & (slots - 1);
3474                 pgcnt_t szcpages;
3475                 int     slotlen;
3476 
3477                 pfnhi = P2ALIGN((pfnhi + 1), szcpgcnt) - 1;
3478                 szcpages = ((pfnhi - pfnlo) + 1) / szcpgcnt;
3479                 slotlen = howmany(szcpages, slots);
3480                 /* skip if 'slotid' slot is empty */
3481                 if (slotid * slotlen >= szcpages)
3482                         return (NULL);
3483                 pfnlo = pfnlo + (((slotid * slotlen) % szcpages) * szcpgcnt);
3484                 ASSERT(pfnlo < pfnhi);
3485                 if (pfnhi > pfnlo + (slotlen * szcpgcnt))
3486                         pfnhi = pfnlo + (slotlen * szcpgcnt) - 1;
3487         }
3488 
3489         /*
3490          * This routine is can be called recursively so we shouldn't
3491          * acquire a reader lock if a write request is pending. This
3492          * could lead to a deadlock with the DR thread.
3493          *
3494          * Returning NULL informs the caller that we could not get
3495          * a contig page with the required characteristics.
3496          */
3497 
3498         if (!memsegs_trylock(0))
3499                 return (NULL);
3500 
3501         /*
3502          * loop through memsegs to look for contig page candidates
3503          */
3504 
3505         for (mseg = memsegs; mseg != NULL; mseg = mseg->next) {
3506                 if (pfnhi < mseg->pages_base || pfnlo >= mseg->pages_end) {
3507                         /* no overlap */
3508                         continue;
3509                 }
3510 
3511                 if (mseg->pages_end - mseg->pages_base < szcpgcnt)
3512                         /* mseg too small */
3513                         continue;
3514 
3515                 /*
3516                  * trim off kernel cage pages from pfn range and check for
3517                  * a trimmed pfn range returned that does not span the
3518                  * desired large page size.
3519                  */
3520                 if (kcage_on) {
3521                         if (trimkcage(mseg, &lo, &hi, pfnlo, pfnhi) == 0 ||
3522                             lo >= hi || ((hi - lo) + 1) < szcpgcnt)
3523                                 continue;
3524                 } else {
3525                         lo = MAX(pfnlo, mseg->pages_base);
3526                         hi = MIN(pfnhi, (mseg->pages_end - 1));
3527                 }
3528 
3529                 /* round to szcpgcnt boundaries */
3530                 lo = P2ROUNDUP(lo, szcpgcnt);
3531 
3532                 MEM_NODE_ITERATOR_INIT(lo, mnode, szc, &it);
3533                 hi = P2ALIGN((hi + 1), szcpgcnt) - 1;
3534 
3535                 if (hi <= lo)
3536                         continue;
3537 
3538                 /*
3539                  * set lo to point to the pfn for the desired bin. Large
3540                  * page sizes may only have a single page color
3541                  */
3542                 skip = szcpgcnt;
3543                 if (ceq_mask > 0 || interleaved_mnodes) {
3544                         /* set lo to point at appropriate color */
3545                         if (((PFN_2_COLOR(lo, szc, &it) ^ bin) & ceq_mask) ||
3546                             (interleaved_mnodes &&
3547                             PFN_2_MEM_NODE(lo) != mnode)) {
3548                                 PAGE_NEXT_PFN_FOR_COLOR(lo, szc, bin, ceq_mask,
3549                                     color_mask, &it);
3550                         }
3551                         if (hi <= lo)
3552                                 /* mseg cannot satisfy color request */
3553                                 continue;
3554                 }
3555 
3556                 /* randomly choose a point between lo and hi to begin search */
3557 
3558                 randpfn = (pfn_t)GETTICK();
3559                 randpfn = ((randpfn % (hi - lo)) + lo) & ~(skip - 1);
3560                 MEM_NODE_ITERATOR_INIT(randpfn, mnode, szc, &it);
3561                 if (ceq_mask || interleaved_mnodes || randpfn == (pfn_t)-1) {
3562                         if (randpfn != (pfn_t)-1) {
3563                                 PAGE_NEXT_PFN_FOR_COLOR(randpfn, szc, bin,
3564                                     ceq_mask, color_mask, &it);
3565                         }
3566                         if (randpfn >= hi) {
3567                                 randpfn = lo;
3568                                 MEM_NODE_ITERATOR_INIT(randpfn, mnode, szc,
3569                                     &it);
3570                         }
3571                 }
3572                 randpp = mseg->pages + (randpfn - mseg->pages_base);
3573 
3574                 ASSERT(randpp->p_pagenum == randpfn);
3575 
3576                 pp = randpp;
3577                 endpp =  mseg->pages + (hi - mseg->pages_base) + 1;
3578 
3579                 ASSERT(randpp + szcpgcnt <= endpp);
3580 
3581                 do {
3582                         ASSERT(!(pp->p_pagenum & szcpgmask));
3583                         ASSERT(((PP_2_BIN(pp) ^ bin) & ceq_mask) == 0);
3584 
3585                         if (page_trylock_contig_pages(mnode, pp, szc, flags)) {
3586                                 /* pages unlocked by page_claim on failure */
3587                                 if (page_claim_contig_pages(pp, szc, flags)) {
3588                                         memsegs_unlock(0);
3589                                         return (pp);
3590                                 }
3591                         }
3592 
3593                         if (ceq_mask == 0 && !interleaved_mnodes) {
3594                                 pp += skip;
3595                         } else {
3596                                 pfn_t pfn = pp->p_pagenum;
3597 
3598                                 PAGE_NEXT_PFN_FOR_COLOR(pfn, szc, bin,
3599                                     ceq_mask, color_mask, &it);
3600                                 if (pfn == (pfn_t)-1) {
3601                                         pp = endpp;
3602                                 } else {
3603                                         pp = mseg->pages +
3604                                             (pfn - mseg->pages_base);
3605                                 }
3606                         }
3607                         if (pp >= endpp) {
3608                                 /* start from the beginning */
3609                                 MEM_NODE_ITERATOR_INIT(lo, mnode, szc, &it);
3610                                 pp = mseg->pages + (lo - mseg->pages_base);
3611                                 ASSERT(pp->p_pagenum == lo);
3612                                 ASSERT(pp + szcpgcnt <= endpp);
3613                         }
3614                 } while (pp != randpp);
3615         }
3616         memsegs_unlock(0);
3617         return (NULL);
3618 }
3619 
3620 
3621 /*
3622  * controlling routine that searches through physical memory in an attempt to
3623  * claim a large page based on the input parameters.
3624  * on the page free lists.
3625  *
3626  * calls page_geti_contig_pages with an initial pfn range from the mnode
3627  * and mtype. page_geti_contig_pages will trim off the parts of the pfn range
3628  * that overlaps with the kernel cage or does not match the requested page
3629  * color if PG_MATCH_COLOR is set.  Since this search is very expensive,
3630  * page_geti_contig_pages may further limit the search range based on
3631  * previous failure counts (pgcpfailcnt[]).
3632  *
3633  * for PGI_PGCPSZC0 requests, page_get_contig_pages will relocate a base
3634  * pagesize page that satisfies mtype.
3635  */
3636 page_t *
3637 page_get_contig_pages(int mnode, uint_t bin, int mtype, uchar_t szc,
3638     uint_t flags)
3639 {
3640         pfn_t           pfnlo, pfnhi;   /* contig pages pfn range */
3641         page_t          *pp;
3642         pgcnt_t         pfnflag = 0;    /* no limit on search if 0 */
3643 
3644         VM_STAT_ADD(vmm_vmstats.pgcp_alloc[szc]);
3645 
3646         /* no allocations from cage */
3647         flags |= PGI_NOCAGE;
3648 
3649         /* LINTED */
3650         MTYPE_START(mnode, mtype, flags);
3651         if (mtype < 0) {     /* mnode does not have memory in mtype range */
3652                 VM_STAT_ADD(vmm_vmstats.pgcp_allocempty[szc]);
3653                 return (NULL);
3654         }
3655 
3656         ASSERT(szc > 0 || (flags & PGI_PGCPSZC0));
3657 
3658         /* do not limit search and ignore color if hi pri */
3659 
3660         if (pgcplimitsearch && ((flags & PGI_PGCPHIPRI) == 0))
3661                 pfnflag = pgcpfailcnt[szc];
3662 
3663         /* remove color match to improve chances */
3664 
3665         if (flags & PGI_PGCPHIPRI || pfnflag)
3666                 flags &= ~PG_MATCH_COLOR;
3667 
3668         do {
3669                 /* get pfn range based on mnode and mtype */
3670                 MNODETYPE_2_PFN(mnode, mtype, pfnlo, pfnhi);
3671 
3672                 ASSERT(pfnhi >= pfnlo);
3673 
3674                 pp = page_geti_contig_pages(mnode, bin, szc, flags,
3675                     pfnlo, pfnhi, pfnflag);
3676 
3677                 if (pp != NULL) {
3678                         pfnflag = pgcpfailcnt[szc];
3679                         if (pfnflag) {
3680                                 /* double the search size */
3681                                 pgcpfailcnt[szc] = pfnflag >> 1;
3682                         }
3683                         VM_STAT_ADD(vmm_vmstats.pgcp_allocok[szc]);
3684                         return (pp);
3685                 }
3686                 MTYPE_NEXT(mnode, mtype, flags);
3687         } while (mtype >= 0);
3688 
3689         VM_STAT_ADD(vmm_vmstats.pgcp_allocfailed[szc]);
3690         return (NULL);
3691 }
3692 
3693 #if defined(__i386) || defined(__amd64)
3694 /*
3695  * Determine the likelihood of finding/coalescing a szc page.
3696  * Return 0 if the likelihood is small otherwise return 1.
3697  *
3698  * For now, be conservative and check only 1g pages and return 0
3699  * if there had been previous coalescing failures and the szc pages
3700  * needed to satisfy request would exhaust most of freemem.
3701  */
3702 int
3703 page_chk_freelist(uint_t szc)
3704 {
3705         pgcnt_t         pgcnt;
3706 
3707         if (szc <= 1)
3708                 return (1);
3709 
3710         pgcnt = page_get_pagecnt(szc);
3711         if (pgcpfailcnt[szc] && pgcnt + throttlefree >= freemem) {
3712                 VM_STAT_ADD(vmm_vmstats.pcf_deny[szc]);
3713                 return (0);
3714         }
3715         VM_STAT_ADD(vmm_vmstats.pcf_allow[szc]);
3716         return (1);
3717 }
3718 #endif
3719 
3720 /*
3721  * Find the `best' page on the freelist for this (vp,off) (as,vaddr) pair.
3722  *
3723  * Does its own locking and accounting.
3724  * If PG_MATCH_COLOR is set, then NULL will be returned if there are no
3725  * pages of the proper color even if there are pages of a different color.
3726  *
3727  * Finds a page, removes it, THEN locks it.
3728  */
3729 
3730 /*ARGSUSED*/
3731 page_t *
3732 page_get_freelist(struct vnode *vp, u_offset_t off, struct seg *seg,
3733     caddr_t vaddr, size_t size, uint_t flags, struct lgrp *lgrp)
3734 {
3735         struct as       *as = seg->s_as;
3736         page_t          *pp = NULL;
3737         ulong_t         bin;
3738         uchar_t         szc;
3739         int             mnode;
3740         int             mtype;
3741         page_t          *(*page_get_func)(int, uint_t, int, uchar_t, uint_t);
3742         lgrp_mnode_cookie_t     lgrp_cookie;
3743 
3744         page_get_func = page_get_mnode_freelist;
3745 
3746         /*
3747          * If we aren't passed a specific lgroup, or passed a freed lgrp
3748          * assume we wish to allocate near to the current thread's home.
3749          */
3750         if (!LGRP_EXISTS(lgrp))
3751                 lgrp = lgrp_home_lgrp();
3752 
3753         if (kcage_on) {
3754                 if ((flags & (PG_NORELOC | PG_PANIC)) == PG_NORELOC &&
3755                     kcage_freemem < kcage_throttlefree + btop(size) &&
3756                     curthread != kcage_cageout_thread) {
3757                         /*
3758                          * Set a "reserve" of kcage_throttlefree pages for
3759                          * PG_PANIC and cageout thread allocations.
3760                          *
3761                          * Everybody else has to serialize in
3762                          * page_create_get_something() to get a cage page, so
3763                          * that we don't deadlock cageout!
3764                          */
3765                         return (NULL);
3766                 }
3767         } else {
3768                 flags &= ~PG_NORELOC;
3769                 flags |= PGI_NOCAGE;
3770         }
3771 
3772         /* LINTED */
3773         MTYPE_INIT(mtype, vp, vaddr, flags, size);
3774 
3775         /*
3776          * Convert size to page size code.
3777          */
3778         if ((szc = page_szc(size)) == (uchar_t)-1)
3779                 panic("page_get_freelist: illegal page size request");
3780         ASSERT(szc < mmu_page_sizes);
3781 
3782         VM_STAT_ADD(vmm_vmstats.pgf_alloc[szc]);
3783 
3784         /* LINTED */
3785         AS_2_BIN(as, seg, vp, vaddr, bin, szc);
3786 
3787         ASSERT(bin < PAGE_GET_PAGECOLORS(szc));
3788 
3789         /*
3790          * Try to get a local page first, but try remote if we can't
3791          * get a page of the right color.
3792          */
3793 pgretry:
3794         LGRP_MNODE_COOKIE_INIT(lgrp_cookie, lgrp, LGRP_SRCH_LOCAL);
3795         while ((mnode = lgrp_memnode_choose(&lgrp_cookie)) >= 0) {
3796                 pp = page_get_func(mnode, bin, mtype, szc, flags);
3797                 if (pp != NULL) {
3798                         VM_STAT_ADD(vmm_vmstats.pgf_allocok[szc]);
3799                         DTRACE_PROBE4(page__get,
3800                             lgrp_t *, lgrp,
3801                             int, mnode,
3802                             ulong_t, bin,
3803                             uint_t, flags);
3804                         return (pp);
3805                 }
3806         }
3807         ASSERT(pp == NULL);
3808 
3809         /*
3810          * for non-SZC0 PAGESIZE requests, check cachelist before checking
3811          * remote free lists.  Caller expected to call page_get_cachelist which
3812          * will check local cache lists and remote free lists.
3813          */
3814         if (szc == 0 && ((flags & PGI_PGCPSZC0) == 0)) {
3815                 VM_STAT_ADD(vmm_vmstats.pgf_allocdeferred);
3816                 return (NULL);
3817         }
3818 
3819         ASSERT(szc > 0 || (flags & PGI_PGCPSZC0));
3820 
3821         lgrp_stat_add(lgrp->lgrp_id, LGRP_NUM_ALLOC_FAIL, 1);
3822 
3823         if (!(flags & PG_LOCAL)) {
3824                 /*
3825                  * Try to get a non-local freelist page.
3826                  */
3827                 LGRP_MNODE_COOKIE_UPGRADE(lgrp_cookie);
3828                 while ((mnode = lgrp_memnode_choose(&lgrp_cookie)) >= 0) {
3829                         pp = page_get_func(mnode, bin, mtype, szc, flags);
3830                         if (pp != NULL) {
3831                                 DTRACE_PROBE4(page__get,
3832                                     lgrp_t *, lgrp,
3833                                     int, mnode,
3834                                     ulong_t, bin,
3835                                     uint_t, flags);
3836                                 VM_STAT_ADD(vmm_vmstats.pgf_allocokrem[szc]);
3837                                 return (pp);
3838                         }
3839                 }
3840                 ASSERT(pp == NULL);
3841         }
3842 
3843         /*
3844          * when the cage is off chances are page_get_contig_pages() will fail
3845          * to lock a large page chunk therefore when the cage is off it's not
3846          * called by default.  this can be changed via /etc/system.
3847          *
3848          * page_get_contig_pages() also called to acquire a base pagesize page
3849          * for page_create_get_something().
3850          */
3851         if (!(flags & PG_NORELOC) && (pg_contig_disable == 0) &&
3852             (kcage_on || pg_lpgcreate_nocage || szc == 0) &&
3853             (page_get_func != page_get_contig_pages)) {
3854 
3855                 VM_STAT_ADD(vmm_vmstats.pgf_allocretry[szc]);
3856                 page_get_func = page_get_contig_pages;
3857                 goto pgretry;
3858         }
3859 
3860         if (!(flags & PG_LOCAL) && pgcplimitsearch &&
3861             page_get_func == page_get_contig_pages)
3862                 SETPGCPFAILCNT(szc);
3863 
3864         VM_STAT_ADD(vmm_vmstats.pgf_allocfailed[szc]);
3865         return (NULL);
3866 }
3867 
3868 /*
3869  * Find the `best' page on the cachelist for this (vp,off) (as,vaddr) pair.
3870  *
3871  * Does its own locking.
3872  * If PG_MATCH_COLOR is set, then NULL will be returned if there are no
3873  * pages of the proper color even if there are pages of a different color.
3874  * Otherwise, scan the bins for ones with pages.  For each bin with pages,
3875  * try to lock one of them.  If no page can be locked, try the
3876  * next bin.  Return NULL if a page can not be found and locked.
3877  *
3878  * Finds a pages, trys to lock it, then removes it.
3879  */
3880 
3881 /*ARGSUSED*/
3882 page_t *
3883 page_get_cachelist(struct vnode *vp, u_offset_t off, struct seg *seg,
3884     caddr_t vaddr, uint_t flags, struct lgrp *lgrp)
3885 {
3886         page_t          *pp;
3887         struct as       *as = seg->s_as;
3888         ulong_t         bin;
3889         /*LINTED*/
3890         int             mnode;
3891         int             mtype;
3892         lgrp_mnode_cookie_t     lgrp_cookie;
3893 
3894         /*
3895          * If we aren't passed a specific lgroup, or pasased a freed lgrp
3896          * assume we wish to allocate near to the current thread's home.
3897          */
3898         if (!LGRP_EXISTS(lgrp))
3899                 lgrp = lgrp_home_lgrp();
3900 
3901         if (!kcage_on) {
3902                 flags &= ~PG_NORELOC;
3903                 flags |= PGI_NOCAGE;
3904         }
3905 
3906         if ((flags & (PG_NORELOC | PG_PANIC | PG_PUSHPAGE)) == PG_NORELOC &&
3907             kcage_freemem <= kcage_throttlefree) {
3908                 /*
3909                  * Reserve kcage_throttlefree pages for critical kernel
3910                  * threads.
3911                  *
3912                  * Everybody else has to go to page_create_get_something()
3913                  * to get a cage page, so we don't deadlock cageout.
3914                  */
3915                 return (NULL);
3916         }
3917 
3918         /* LINTED */
3919         AS_2_BIN(as, seg, vp, vaddr, bin, 0);
3920 
3921         ASSERT(bin < PAGE_GET_PAGECOLORS(0));
3922 
3923         /* LINTED */
3924         MTYPE_INIT(mtype, vp, vaddr, flags, MMU_PAGESIZE);
3925 
3926         VM_STAT_ADD(vmm_vmstats.pgc_alloc);
3927 
3928         /*
3929          * Try local cachelists first
3930          */
3931         LGRP_MNODE_COOKIE_INIT(lgrp_cookie, lgrp, LGRP_SRCH_LOCAL);
3932         while ((mnode = lgrp_memnode_choose(&lgrp_cookie)) >= 0) {
3933                 pp = page_get_mnode_cachelist(bin, flags, mnode, mtype);
3934                 if (pp != NULL) {
3935                         VM_STAT_ADD(vmm_vmstats.pgc_allocok);
3936                         DTRACE_PROBE4(page__get,
3937                             lgrp_t *, lgrp,
3938                             int, mnode,
3939                             ulong_t, bin,
3940                             uint_t, flags);
3941                         return (pp);
3942                 }
3943         }
3944 
3945         lgrp_stat_add(lgrp->lgrp_id, LGRP_NUM_ALLOC_FAIL, 1);
3946 
3947         /*
3948          * Try freelists/cachelists that are farther away
3949          * This is our only chance to allocate remote pages for PAGESIZE
3950          * requests.
3951          */
3952         LGRP_MNODE_COOKIE_UPGRADE(lgrp_cookie);
3953         while ((mnode = lgrp_memnode_choose(&lgrp_cookie)) >= 0) {
3954                 pp = page_get_mnode_freelist(mnode, bin, mtype,
3955                     0, flags);
3956                 if (pp != NULL) {
3957                         VM_STAT_ADD(vmm_vmstats.pgc_allocokdeferred);
3958                         DTRACE_PROBE4(page__get,
3959                             lgrp_t *, lgrp,
3960                             int, mnode,
3961                             ulong_t, bin,
3962                             uint_t, flags);
3963                         return (pp);
3964                 }
3965                 pp = page_get_mnode_cachelist(bin, flags, mnode, mtype);
3966                 if (pp != NULL) {
3967                         VM_STAT_ADD(vmm_vmstats.pgc_allocokrem);
3968                         DTRACE_PROBE4(page__get,
3969                             lgrp_t *, lgrp,
3970                             int, mnode,
3971                             ulong_t, bin,
3972                             uint_t, flags);
3973                         return (pp);
3974                 }
3975         }
3976 
3977         VM_STAT_ADD(vmm_vmstats.pgc_allocfailed);
3978         return (NULL);
3979 }
3980 
3981 page_t *
3982 page_get_mnode_cachelist(uint_t bin, uint_t flags, int mnode, int mtype)
3983 {
3984         kmutex_t                *pcm;
3985         page_t                  *pp, *first_pp;
3986         uint_t                  sbin;
3987         int                     plw_initialized;
3988         page_list_walker_t      plw;
3989 
3990         VM_STAT_ADD(vmm_vmstats.pgmc_alloc);
3991 
3992         /* LINTED */
3993         MTYPE_START(mnode, mtype, flags);
3994         if (mtype < 0) {     /* mnode does not have memory in mtype range */
3995                 VM_STAT_ADD(vmm_vmstats.pgmc_allocempty);
3996                 return (NULL);
3997         }
3998 
3999 try_again:
4000 
4001         plw_initialized = 0;
4002         plw.plw_ceq_dif = 1;
4003 
4004         /*
4005          * Only hold one cachelist lock at a time, that way we
4006          * can start anywhere and not have to worry about lock
4007          * ordering.
4008          */
4009 
4010         for (plw.plw_count = 0;
4011             plw.plw_count < plw.plw_ceq_dif; plw.plw_count++) {
4012                 sbin = bin;
4013                 do {
4014 
4015                         if (!PAGE_CACHELISTS(mnode, bin, mtype))
4016                                 goto bin_empty_1;
4017                         pcm = PC_BIN_MUTEX(mnode, bin, PG_CACHE_LIST);
4018                         mutex_enter(pcm);
4019                         pp = PAGE_CACHELISTS(mnode, bin, mtype);
4020                         if (pp == NULL)
4021                                 goto bin_empty_0;
4022 
4023                         first_pp = pp;
4024                         ASSERT(pp->p_vnode);
4025                         ASSERT(PP_ISAGED(pp) == 0);
4026                         ASSERT(pp->p_szc == 0);
4027                         ASSERT(PFN_2_MEM_NODE(pp->p_pagenum) == mnode);
4028                         while (IS_DUMP_PAGE(pp) || !page_trylock(pp, SE_EXCL)) {
4029                                 pp = pp->p_next;
4030                                 ASSERT(pp->p_szc == 0);
4031                                 if (pp == first_pp) {
4032                                         /*
4033                                          * We have searched the complete list!
4034                                          * And all of them (might only be one)
4035                                          * are locked. This can happen since
4036                                          * these pages can also be found via
4037                                          * the hash list. When found via the
4038                                          * hash list, they are locked first,
4039                                          * then removed. We give up to let the
4040                                          * other thread run.
4041                                          */
4042                                         pp = NULL;
4043                                         break;
4044                                 }
4045                                 ASSERT(pp->p_vnode);
4046                                 ASSERT(PP_ISFREE(pp));
4047                                 ASSERT(PP_ISAGED(pp) == 0);
4048                                 ASSERT(PFN_2_MEM_NODE(pp->p_pagenum) ==
4049                                     mnode);
4050                         }
4051 
4052                         if (pp) {
4053                                 page_t  **ppp;
4054                                 /*
4055                                  * Found and locked a page.
4056                                  * Pull it off the list.
4057                                  */
4058                                 ASSERT(mtype == PP_2_MTYPE(pp));
4059                                 ppp = &PAGE_CACHELISTS(mnode, bin, mtype);
4060                                 page_sub(ppp, pp);
4061                                 /*
4062                                  * Subtract counters before releasing pcm mutex
4063                                  * to avoid a race with page_freelist_coalesce
4064                                  * and page_freelist_split.
4065                                  */
4066                                 page_ctr_sub(mnode, mtype, pp, PG_CACHE_LIST);
4067                                 mutex_exit(pcm);
4068                                 ASSERT(pp->p_vnode);
4069                                 ASSERT(PP_ISAGED(pp) == 0);
4070 #if defined(__sparc)
4071                                 ASSERT(!kcage_on ||
4072                                     (flags & PG_NORELOC) == 0 ||
4073                                     PP_ISNORELOC(pp));
4074                                 if (PP_ISNORELOC(pp)) {
4075                                         kcage_freemem_sub(1);
4076                                 }
4077 #endif
4078                                 VM_STAT_ADD(vmm_vmstats. pgmc_allocok);
4079                                 return (pp);
4080                         }
4081 bin_empty_0:
4082                         mutex_exit(pcm);
4083 bin_empty_1:
4084                         if (plw_initialized == 0) {
4085                                 page_list_walk_init(0, flags, bin, 0, 1, &plw);
4086                                 plw_initialized = 1;
4087                         }
4088                         /* calculate the next bin with equivalent color */
4089                         bin = ADD_MASKED(bin, plw.plw_bin_step,
4090                             plw.plw_ceq_mask[0], plw.plw_color_mask);
4091                 } while (sbin != bin);
4092 
4093                 if (plw.plw_ceq_dif > 1)
4094                         bin = page_list_walk_next_bin(0, bin, &plw);
4095         }
4096 
4097         MTYPE_NEXT(mnode, mtype, flags);
4098         if (mtype >= 0)
4099                 goto try_again;
4100 
4101         VM_STAT_ADD(vmm_vmstats.pgmc_allocfailed);
4102         return (NULL);
4103 }
4104 
4105 #ifdef DEBUG
4106 #define REPL_PAGE_STATS
4107 #endif /* DEBUG */
4108 
4109 #ifdef REPL_PAGE_STATS
4110 struct repl_page_stats {
4111         uint_t  ngets;
4112         uint_t  ngets_noreloc;
4113         uint_t  npgr_noreloc;
4114         uint_t  nnopage_first;
4115         uint_t  nnopage;
4116         uint_t  nhashout;
4117         uint_t  nnofree;
4118         uint_t  nnext_pp;
4119 } repl_page_stats;
4120 #define REPL_STAT_INCR(v)       atomic_inc_32(&repl_page_stats.v)
4121 #else /* REPL_PAGE_STATS */
4122 #define REPL_STAT_INCR(v)
4123 #endif /* REPL_PAGE_STATS */
4124 
4125 int     pgrppgcp;
4126 
4127 /*
4128  * The freemem accounting must be done by the caller.
4129  * First we try to get a replacement page of the same size as like_pp,
4130  * if that is not possible, then we just get a set of discontiguous
4131  * PAGESIZE pages.
4132  */
4133 page_t *
4134 page_get_replacement_page(page_t *orig_like_pp, struct lgrp *lgrp_target,
4135     uint_t pgrflags)
4136 {
4137         page_t          *like_pp;
4138         page_t          *pp, *pplist;
4139         page_t          *pl = NULL;
4140         ulong_t         bin;
4141         int             mnode, page_mnode;
4142         int             szc;
4143         spgcnt_t        npgs, pg_cnt;
4144         pfn_t           pfnum;
4145         int             mtype;
4146         int             flags = 0;
4147         lgrp_mnode_cookie_t     lgrp_cookie;
4148         lgrp_t          *lgrp;
4149 
4150         REPL_STAT_INCR(ngets);
4151         like_pp = orig_like_pp;
4152         ASSERT(PAGE_EXCL(like_pp));
4153 
4154         szc = like_pp->p_szc;
4155         npgs = page_get_pagecnt(szc);
4156         /*
4157          * Now we reset like_pp to the base page_t.
4158          * That way, we won't walk past the end of this 'szc' page.
4159          */
4160         pfnum = PFN_BASE(like_pp->p_pagenum, szc);
4161         like_pp = page_numtopp_nolock(pfnum);
4162         ASSERT(like_pp->p_szc == szc);
4163 
4164         if (PP_ISNORELOC(like_pp)) {
4165                 ASSERT(kcage_on);
4166                 REPL_STAT_INCR(ngets_noreloc);
4167                 flags = PGI_RELOCONLY;
4168         } else if (pgrflags & PGR_NORELOC) {
4169                 ASSERT(kcage_on);
4170                 REPL_STAT_INCR(npgr_noreloc);
4171                 flags = PG_NORELOC;
4172         }
4173 
4174         /*
4175          * Kernel pages must always be replaced with the same size
4176          * pages, since we cannot properly handle demotion of kernel
4177          * pages.
4178          */
4179         if (PP_ISKAS(like_pp))
4180                 pgrflags |= PGR_SAMESZC;
4181 
4182         MTYPE_PGR_INIT(mtype, flags, like_pp, npgs);
4183 
4184         while (npgs) {
4185                 pplist = NULL;
4186                 for (;;) {
4187                         pg_cnt = page_get_pagecnt(szc);
4188                         bin = PP_2_BIN(like_pp);
4189                         ASSERT(like_pp->p_szc == orig_like_pp->p_szc);
4190                         ASSERT(pg_cnt <= npgs);
4191 
4192                         /*
4193                          * If an lgroup was specified, try to get the
4194                          * page from that lgroup.
4195                          * NOTE: Must be careful with code below because
4196                          *       lgroup may disappear and reappear since there
4197                          *       is no locking for lgroup here.
4198                          */
4199                         if (LGRP_EXISTS(lgrp_target)) {
4200                                 /*
4201                                  * Keep local variable for lgroup separate
4202                                  * from lgroup argument since this code should
4203                                  * only be exercised when lgroup argument
4204                                  * exists....
4205                                  */
4206                                 lgrp = lgrp_target;
4207 
4208                                 /* Try the lgroup's freelists first */
4209                                 LGRP_MNODE_COOKIE_INIT(lgrp_cookie, lgrp,
4210                                     LGRP_SRCH_LOCAL);
4211                                 while ((pplist == NULL) &&
4212                                     (mnode = lgrp_memnode_choose(&lgrp_cookie))
4213                                     != -1) {
4214                                         pplist =
4215                                             page_get_mnode_freelist(mnode, bin,
4216                                             mtype, szc, flags);
4217                                 }
4218 
4219                                 /*
4220                                  * Now try it's cachelists if this is a
4221                                  * small page. Don't need to do it for
4222                                  * larger ones since page_freelist_coalesce()
4223                                  * already failed.
4224                                  */
4225                                 if (pplist != NULL || szc != 0)
4226                                         break;
4227 
4228                                 /* Now try it's cachelists */
4229                                 LGRP_MNODE_COOKIE_INIT(lgrp_cookie, lgrp,
4230                                     LGRP_SRCH_LOCAL);
4231 
4232                                 while ((pplist == NULL) &&
4233                                     (mnode = lgrp_memnode_choose(&lgrp_cookie))
4234                                     != -1) {
4235                                         pplist =
4236                                             page_get_mnode_cachelist(bin, flags,
4237                                             mnode, mtype);
4238                                 }
4239                                 if (pplist != NULL) {
4240                                         page_hashout(pplist, NULL);
4241                                         PP_SETAGED(pplist);
4242                                         REPL_STAT_INCR(nhashout);
4243                                         break;
4244                                 }
4245                                 /* Done looking in this lgroup. Bail out. */
4246                                 break;
4247                         }
4248 
4249                         /*
4250                          * No lgroup was specified (or lgroup was removed by
4251                          * DR, so just try to get the page as close to
4252                          * like_pp's mnode as possible.
4253                          * First try the local freelist...
4254                          */
4255                         mnode = PP_2_MEM_NODE(like_pp);
4256                         pplist = page_get_mnode_freelist(mnode, bin,
4257                             mtype, szc, flags);
4258                         if (pplist != NULL)
4259                                 break;
4260 
4261                         REPL_STAT_INCR(nnofree);
4262 
4263                         /*
4264                          * ...then the local cachelist. Don't need to do it for
4265                          * larger pages cause page_freelist_coalesce() already
4266                          * failed there anyway.
4267                          */
4268                         if (szc == 0) {
4269                                 pplist = page_get_mnode_cachelist(bin, flags,
4270                                     mnode, mtype);
4271                                 if (pplist != NULL) {
4272                                         page_hashout(pplist, NULL);
4273                                         PP_SETAGED(pplist);
4274                                         REPL_STAT_INCR(nhashout);
4275                                         break;
4276                                 }
4277                         }
4278 
4279                         /* Now try remote freelists */
4280                         page_mnode = mnode;
4281                         lgrp =
4282                             lgrp_hand_to_lgrp(MEM_NODE_2_LGRPHAND(page_mnode));
4283                         LGRP_MNODE_COOKIE_INIT(lgrp_cookie, lgrp,
4284                             LGRP_SRCH_HIER);
4285                         while (pplist == NULL &&
4286                             (mnode = lgrp_memnode_choose(&lgrp_cookie))
4287                             != -1) {
4288                                 /*
4289                                  * Skip local mnode.
4290                                  */
4291                                 if ((mnode == page_mnode) ||
4292                                     (mem_node_config[mnode].exists == 0))
4293                                         continue;
4294 
4295                                 pplist = page_get_mnode_freelist(mnode,
4296                                     bin, mtype, szc, flags);
4297                         }
4298 
4299                         if (pplist != NULL)
4300                                 break;
4301 
4302 
4303                         /* Now try remote cachelists */
4304                         LGRP_MNODE_COOKIE_INIT(lgrp_cookie, lgrp,
4305                             LGRP_SRCH_HIER);
4306                         while (pplist == NULL && szc == 0) {
4307                                 mnode = lgrp_memnode_choose(&lgrp_cookie);
4308                                 if (mnode == -1)
4309                                         break;
4310                                 /*
4311                                  * Skip local mnode.
4312                                  */
4313                                 if ((mnode == page_mnode) ||
4314                                     (mem_node_config[mnode].exists == 0))
4315                                         continue;
4316 
4317                                 pplist = page_get_mnode_cachelist(bin,
4318                                     flags, mnode, mtype);
4319 
4320                                 if (pplist != NULL) {
4321                                         page_hashout(pplist, NULL);
4322                                         PP_SETAGED(pplist);
4323                                         REPL_STAT_INCR(nhashout);
4324                                         break;
4325                                 }
4326                         }
4327 
4328                         /*
4329                          * Break out of while loop under the following cases:
4330                          * - If we successfully got a page.
4331                          * - If pgrflags specified only returning a specific
4332                          *   page size and we could not find that page size.
4333                          * - If we could not satisfy the request with PAGESIZE
4334                          *   or larger pages.
4335                          */
4336                         if (pplist != NULL || szc == 0)
4337                                 break;
4338 
4339                         if ((pgrflags & PGR_SAMESZC) || pgrppgcp) {
4340                                 /* try to find contig page */
4341 
4342                                 LGRP_MNODE_COOKIE_INIT(lgrp_cookie, lgrp,
4343                                     LGRP_SRCH_HIER);
4344 
4345                                 while ((pplist == NULL) &&
4346                                     (mnode =
4347                                     lgrp_memnode_choose(&lgrp_cookie))
4348                                     != -1) {
4349                                         pplist = page_get_contig_pages(
4350                                             mnode, bin, mtype, szc,
4351                                             flags | PGI_PGCPHIPRI);
4352                                 }
4353                                 break;
4354                         }
4355 
4356                         /*
4357                          * The correct thing to do here is try the next
4358                          * page size down using szc--. Due to a bug
4359                          * with the processing of HAT_RELOAD_SHARE
4360                          * where the sfmmu_ttecnt arrays of all
4361                          * hats sharing an ISM segment don't get updated,
4362                          * using intermediate size pages for relocation
4363                          * can lead to continuous page faults.
4364                          */
4365                         szc = 0;
4366                 }
4367 
4368                 if (pplist != NULL) {
4369                         DTRACE_PROBE4(page__get,
4370                             lgrp_t *, lgrp,
4371                             int, mnode,
4372                             ulong_t, bin,
4373                             uint_t, flags);
4374 
4375                         while (pplist != NULL && pg_cnt--) {
4376                                 ASSERT(pplist != NULL);
4377                                 pp = pplist;
4378                                 page_sub(&pplist, pp);
4379                                 PP_CLRFREE(pp);
4380                                 PP_CLRAGED(pp);
4381                                 page_list_concat(&pl, &pp);
4382                                 npgs--;
4383                                 like_pp = like_pp + 1;
4384                                 REPL_STAT_INCR(nnext_pp);
4385                         }
4386                         ASSERT(pg_cnt == 0);
4387                 } else {
4388                         break;
4389                 }
4390         }
4391 
4392         if (npgs) {
4393                 /*
4394                  * We were unable to allocate the necessary number
4395                  * of pages.
4396                  * We need to free up any pl.
4397                  */
4398                 REPL_STAT_INCR(nnopage);
4399                 page_free_replacement_page(pl);
4400                 return (NULL);
4401         } else {
4402                 return (pl);
4403         }
4404 }
4405 
4406 /*
4407  * demote a free large page to it's constituent pages
4408  */
4409 void
4410 page_demote_free_pages(page_t *pp)
4411 {
4412 
4413         int mnode;
4414 
4415         ASSERT(pp != NULL);
4416         ASSERT(PAGE_LOCKED(pp));
4417         ASSERT(PP_ISFREE(pp));
4418         ASSERT(pp->p_szc != 0 && pp->p_szc < mmu_page_sizes);
4419 
4420         mnode = PP_2_MEM_NODE(pp);
4421         page_freelist_lock(mnode);
4422         if (pp->p_szc != 0) {
4423                 (void) page_demote(mnode, PFN_BASE(pp->p_pagenum,
4424                     pp->p_szc), 0, pp->p_szc, 0, PC_NO_COLOR, PC_FREE);
4425         }
4426         page_freelist_unlock(mnode);
4427         ASSERT(pp->p_szc == 0);
4428 }
4429 
4430 /*
4431  * Factor in colorequiv to check additional 'equivalent' bins.
4432  * colorequiv may be set in /etc/system
4433  */
4434 void
4435 page_set_colorequiv_arr(void)
4436 {
4437         if (colorequiv > 1) {
4438                 int i;
4439                 uint_t sv_a = lowbit(colorequiv) - 1;
4440 
4441                 if (sv_a > 15)
4442                         sv_a = 15;
4443 
4444                 for (i = 0; i < MMU_PAGE_SIZES; i++) {
4445                         uint_t colors;
4446                         uint_t a = sv_a;
4447 
4448                         if ((colors = hw_page_array[i].hp_colors) <= 1) {
4449                                 continue;
4450                         }
4451                         while ((colors >> a) == 0)
4452                                 a--;
4453                         if ((a << 4) > colorequivszc[i]) {
4454                                 colorequivszc[i] = (a << 4);
4455                         }
4456                 }
4457         }
4458 }