Print this page
restore sparc comments
de-linting of .s files
@@ -21,22 +21,16 @@
/*
* Copyright 2008 Sun Microsystems, Inc. All rights reserved.
* Use is subject to license terms.
*/
-#pragma ident "%Z%%M% %I% %E% SMI"
-
/*
* SFMMU primitives. These primitives should only be used by sfmmu
* routines.
*/
-#if defined(lint)
-#include <sys/types.h>
-#else /* lint */
#include "assym.h"
-#endif /* lint */
#include <sys/asm_linkage.h>
#include <sys/machtrap.h>
#include <sys/machasi.h>
#include <sys/sun4asi.h>
@@ -54,52 +48,11 @@
/*
* sfmmu related subroutines
*/
-#if defined (lint)
-
/*
- * sfmmu related subroutines
- */
-/* ARGSUSED */
-void
-sfmmu_raise_tsb_exception(uint64_t sfmmup, uint64_t rctx)
-{}
-
-/* ARGSUSED */
-void
-sfmmu_itlb_ld_kva(caddr_t vaddr, tte_t *tte)
-{}
-
-/* ARGSUSED */
-void
-sfmmu_dtlb_ld_kva(caddr_t vaddr, tte_t *tte)
-{}
-
-int
-sfmmu_getctx_pri()
-{ return(0); }
-
-int
-sfmmu_getctx_sec()
-{ return(0); }
-
-/* ARGSUSED */
-void
-sfmmu_setctx_sec(uint_t ctx)
-{}
-
-/* ARGSUSED */
-void
-sfmmu_load_mmustate(sfmmu_t *sfmmup)
-{
-}
-
-#else /* lint */
-
-/*
* Invalidate either the context of a specific victim or any process
* currently running on this CPU.
*
* %g1 = sfmmup whose ctx is being invalidated
* when called from sfmmu_wrap_around, %g1 == INVALID_CONTEXT
@@ -607,31 +560,22 @@
8:
retl
nop
SET_SIZE(sfmmu_load_mmustate)
-#endif /* lint */
-
-#if defined (lint)
/*
- * Invalidate all of the entries within the tsb, by setting the inv bit
+ * Invalidate all of the entries within the TSB, by setting the inv bit
* in the tte_tag field of each tsbe.
*
- * We take advantage of the fact TSBs are page aligned and a multiple of
- * PAGESIZE to use block stores.
+ * We take advantage of the fact that the TSBs are page aligned and a
+ * multiple of PAGESIZE to use ASI_BLK_INIT_xxx ASI.
*
* See TSB_LOCK_ENTRY and the miss handlers for how this works in practice
* (in short, we set all bits in the upper word of the tag, and we give the
* invalid bit precedence over other tag bits in both places).
*/
-/* ARGSUSED */
-void
-sfmmu_inv_tsb_fast(caddr_t tsb_base, uint_t tsb_bytes)
-{}
-#else /* lint */
-
#define VIS_BLOCKSIZE 64
ENTRY(sfmmu_inv_tsb_fast)
! Get space for aligned block of saved fp regs.
@@ -705,14 +649,10 @@
stb %l3, [THREAD_REG + T_PREEMPT]
ret
restore
SET_SIZE(sfmmu_inv_tsb_fast)
-#endif /* lint */
-
-#if defined(lint)
-
/*
* Prefetch "struct tsbe" while walking TSBs.
* prefetch 7 cache lines ahead of where we are at now.
* #n_reads is being used since #one_read only applies to
* floating point reads, and we are not doing floating point
@@ -719,29 +659,16 @@
* reads. However, this has the negative side effect of polluting
* the ecache.
* The 448 comes from (7 * 64) which is how far ahead of our current
* address, we want to prefetch.
*/
-/*ARGSUSED*/
-void
-prefetch_tsbe_read(struct tsbe *tsbep)
-{}
-
-/* Prefetch the tsbe that we are about to write */
-/*ARGSUSED*/
-void
-prefetch_tsbe_write(struct tsbe *tsbep)
-{}
-
-#else /* lint */
-
ENTRY(prefetch_tsbe_read)
retl
prefetch [%o0+448], #n_reads
SET_SIZE(prefetch_tsbe_read)
+/* Prefetch the tsbe that we are about to write */
ENTRY(prefetch_tsbe_write)
retl
prefetch [%o0], #n_writes
SET_SIZE(prefetch_tsbe_write)
-#endif /* lint */