Static'ify.

This commit is contained in:
thorpej 2006-04-02 03:41:32 +00:00
parent 999d9f8f32
commit b2bea1ed90
2 changed files with 87 additions and 89 deletions

View File

@ -1,4 +1,4 @@
/* $NetBSD: pmap.c,v 1.213 2005/12/24 20:06:46 perry Exp $ */
/* $NetBSD: pmap.c,v 1.214 2006/04/02 03:41:32 thorpej Exp $ */
/*-
* Copyright (c) 1998, 1999, 2000, 2001 The NetBSD Foundation, Inc.
@ -145,7 +145,7 @@
#include <sys/cdefs.h> /* RCS ID & Copyright macro defns */
__KERNEL_RCSID(0, "$NetBSD: pmap.c,v 1.213 2005/12/24 20:06:46 perry Exp $");
__KERNEL_RCSID(0, "$NetBSD: pmap.c,v 1.214 2006/04/02 03:41:32 thorpej Exp $");
#include <sys/param.h>
#include <sys/systm.h>
@ -191,7 +191,7 @@ int pmapdebug = PDB_PARANOIA;
* convert to an alpha protection code.
*/
#define pte_prot(m, p) (protection_codes[m == pmap_kernel() ? 0 : 1][p])
int protection_codes[2][8];
static int protection_codes[2][8];
/*
* kernel_lev1map:
@ -222,7 +222,7 @@ pt_entry_t *kernel_lev1map;
/*
* Virtual Page Table.
*/
pt_entry_t *VPT;
static pt_entry_t *VPT;
struct pmap kernel_pmap_store
[(PMAP_SIZEOF(ALPHA_MAXPROCS) + sizeof(struct pmap) - 1)
@ -232,7 +232,7 @@ paddr_t avail_start; /* PA of first available physical page */
paddr_t avail_end; /* PA of last available physical page */
static vaddr_t virtual_end; /* VA of last avail page (end of kernel AS) */
boolean_t pmap_initialized; /* Has pmap_init completed? */
static boolean_t pmap_initialized; /* Has pmap_init completed? */
u_long pmap_pages_stolen; /* instrumentation */
@ -241,7 +241,7 @@ u_long pmap_pages_stolen; /* instrumentation */
* space for when allocating the pmap structure. It is used to
* size a per-CPU array of ASN and ASN Generation number.
*/
u_long pmap_ncpuids;
static u_long pmap_ncpuids;
#ifndef PMAP_PV_LOWAT
#define PMAP_PV_LOWAT 16
@ -253,15 +253,15 @@ int pmap_pv_lowat = PMAP_PV_LOWAT;
* page tables are allocated. This list is kept LRU-ordered by
* pmap_activate().
*/
TAILQ_HEAD(, pmap) pmap_all_pmaps;
static TAILQ_HEAD(, pmap) pmap_all_pmaps;
/*
* The pools from which pmap structures and sub-structures are allocated.
*/
struct pool pmap_pmap_pool;
struct pool pmap_l1pt_pool;
struct pool_cache pmap_l1pt_cache;
struct pool pmap_pv_pool;
static struct pool pmap_pmap_pool;
static struct pool pmap_l1pt_pool;
static struct pool_cache pmap_l1pt_cache;
static struct pool pmap_pv_pool;
/*
* Address Space Numbers.
@ -320,9 +320,9 @@ struct pool pmap_pv_pool;
* the ASN generation in this particular case) to keep the logic sane
* in other parts of the code.
*/
u_int pmap_max_asn; /* max ASN supported by the system */
/* next ASN and current ASN generation */
struct pmap_asn_info pmap_asn_info[ALPHA_MAXPROCS];
static u_int pmap_max_asn; /* max ASN supported by the system */
/* next ASN and cur ASN generation */
static struct pmap_asn_info pmap_asn_info[ALPHA_MAXPROCS];
/*
* Locking:
@ -385,9 +385,9 @@ struct pmap_asn_info pmap_asn_info[ALPHA_MAXPROCS];
* with the pmap already locked by the caller (which will be
* an interface function).
*/
struct lock pmap_main_lock;
struct simplelock pmap_all_pmaps_slock;
struct simplelock pmap_growkernel_slock;
static struct lock pmap_main_lock;
static struct simplelock pmap_all_pmaps_slock;
static struct simplelock pmap_growkernel_slock;
#if defined(MULTIPROCESSOR) || defined(LOCKDEBUG)
#define PMAP_MAP_TO_HEAD_LOCK() \
@ -428,7 +428,7 @@ struct pmap_tlb_shootdown_job {
pt_entry_t pj_pte; /* the PTE bits */
};
struct pmap_tlb_shootdown_q {
static struct pmap_tlb_shootdown_q {
TAILQ_HEAD(, pmap_tlb_shootdown_job) pq_head;
int pq_pte; /* aggregate PTE bits */
int pq_count; /* number of pending requests */
@ -440,23 +440,23 @@ struct pmap_tlb_shootdown_q {
do { \
s = splvm(); \
simple_lock(&(pq)->pq_slock); \
} while (0)
} while (/*CONSTCOND*/0)
#define PSJQ_UNLOCK(pq, s) \
do { \
simple_unlock(&(pq)->pq_slock); \
splx(s); \
} while (0)
} while (/*CONSTCOND*/0)
/* If we have more pending jobs than this, we just nail the whole TLB. */
#define PMAP_TLB_SHOOTDOWN_MAXJOBS 6
struct pool pmap_tlb_shootdown_job_pool;
static struct pool pmap_tlb_shootdown_job_pool;
void pmap_tlb_shootdown_q_drain(struct pmap_tlb_shootdown_q *);
struct pmap_tlb_shootdown_job *pmap_tlb_shootdown_job_get
static void pmap_tlb_shootdown_q_drain(struct pmap_tlb_shootdown_q *);
static struct pmap_tlb_shootdown_job *pmap_tlb_shootdown_job_get
(struct pmap_tlb_shootdown_q *);
void pmap_tlb_shootdown_job_put(struct pmap_tlb_shootdown_q *,
static void pmap_tlb_shootdown_job_put(struct pmap_tlb_shootdown_q *,
struct pmap_tlb_shootdown_job *);
#endif /* MULTIPROCESSOR */
@ -465,42 +465,42 @@ void pmap_tlb_shootdown_job_put(struct pmap_tlb_shootdown_q *,
/*
* Internal routines
*/
void alpha_protection_init(void);
void pmap_do_remove(pmap_t, vaddr_t, vaddr_t, boolean_t);
boolean_t pmap_remove_mapping(pmap_t, vaddr_t, pt_entry_t *,
static void alpha_protection_init(void);
static void pmap_do_remove(pmap_t, vaddr_t, vaddr_t, boolean_t);
static boolean_t pmap_remove_mapping(pmap_t, vaddr_t, pt_entry_t *,
boolean_t, long);
void pmap_changebit(struct vm_page *, pt_entry_t, pt_entry_t, long);
static void pmap_changebit(struct vm_page *, pt_entry_t, pt_entry_t, long);
/*
* PT page management functions.
*/
int pmap_lev1map_create(pmap_t, long);
void pmap_lev1map_destroy(pmap_t, long);
int pmap_ptpage_alloc(pmap_t, pt_entry_t *, int);
void pmap_ptpage_free(pmap_t, pt_entry_t *);
void pmap_l3pt_delref(pmap_t, vaddr_t, pt_entry_t *, long);
void pmap_l2pt_delref(pmap_t, pt_entry_t *, pt_entry_t *, long);
void pmap_l1pt_delref(pmap_t, pt_entry_t *, long);
static int pmap_lev1map_create(pmap_t, long);
static void pmap_lev1map_destroy(pmap_t, long);
static int pmap_ptpage_alloc(pmap_t, pt_entry_t *, int);
static void pmap_ptpage_free(pmap_t, pt_entry_t *);
static void pmap_l3pt_delref(pmap_t, vaddr_t, pt_entry_t *, long);
static void pmap_l2pt_delref(pmap_t, pt_entry_t *, pt_entry_t *, long);
static void pmap_l1pt_delref(pmap_t, pt_entry_t *, long);
void *pmap_l1pt_alloc(struct pool *, int);
void pmap_l1pt_free(struct pool *, void *);
static void *pmap_l1pt_alloc(struct pool *, int);
static void pmap_l1pt_free(struct pool *, void *);
struct pool_allocator pmap_l1pt_allocator = {
static struct pool_allocator pmap_l1pt_allocator = {
pmap_l1pt_alloc, pmap_l1pt_free, 0,
};
int pmap_l1pt_ctor(void *, void *, int);
static int pmap_l1pt_ctor(void *, void *, int);
/*
* PV table management functions.
*/
int pmap_pv_enter(pmap_t, struct vm_page *, vaddr_t, pt_entry_t *,
static int pmap_pv_enter(pmap_t, struct vm_page *, vaddr_t, pt_entry_t *,
boolean_t);
void pmap_pv_remove(pmap_t, struct vm_page *, vaddr_t, boolean_t);
void *pmap_pv_page_alloc(struct pool *, int);
void pmap_pv_page_free(struct pool *, void *);
static void pmap_pv_remove(pmap_t, struct vm_page *, vaddr_t, boolean_t);
static void *pmap_pv_page_alloc(struct pool *, int);
static void pmap_pv_page_free(struct pool *, void *);
struct pool_allocator pmap_pv_page_allocator = {
static struct pool_allocator pmap_pv_page_allocator = {
pmap_pv_page_alloc, pmap_pv_page_free, 0,
};
@ -514,15 +514,15 @@ void pmap_pv_dump(paddr_t);
/*
* ASN management functions.
*/
void pmap_asn_alloc(pmap_t, long);
static void pmap_asn_alloc(pmap_t, long);
/*
* Misc. functions.
*/
boolean_t pmap_physpage_alloc(int, paddr_t *);
void pmap_physpage_free(paddr_t);
int pmap_physpage_addref(void *);
int pmap_physpage_delref(void *);
static boolean_t pmap_physpage_alloc(int, paddr_t *);
static void pmap_physpage_free(paddr_t);
static int pmap_physpage_addref(void *);
static int pmap_physpage_delref(void *);
/*
* Define PMAP_NO_LAZY_LEV1MAP in order to have a lev1map allocated
@ -606,7 +606,7 @@ do { \
panic("PMAP_ACTIVATE_ASN_SANITY"); \
} \
} \
} while (0)
} while (/*CONSTCOND*/0)
#else
#define PMAP_ACTIVATE_ASN_SANITY(pmap, cpu_id) /* nothing */
#endif
@ -637,7 +637,7 @@ do { \
*/ \
(void) alpha_pal_swpctx((u_long)l->l_md.md_pcbpaddr); \
} \
} while (0)
} while (/*CONSTCOND*/0)
#if defined(MULTIPROCESSOR)
/*
@ -679,13 +679,13 @@ do { \
do { \
alpha_pal_imb(); \
alpha_broadcast_ipi(ALPHA_IPI_IMB); \
} while (0)
} while (/*CONSTCOND*/0)
#define PMAP_SYNC_ISTREAM_USER(pmap) \
do { \
alpha_multicast_ipi((pmap)->pm_cpus, ALPHA_IPI_AST); \
/* for curcpu, will happen in userret() */ \
} while (0)
} while (/*CONSTCOND*/0)
#else
#define PMAP_SYNC_ISTREAM_KERNEL() alpha_pal_imb()
#define PMAP_SYNC_ISTREAM_USER(pmap) /* will happen in userret() */
@ -697,7 +697,7 @@ do { \
PMAP_SYNC_ISTREAM_KERNEL(); \
else \
PMAP_SYNC_ISTREAM_USER(pmap); \
} while (0)
} while (/*CONSTCOND*/0)
/*
* PMAP_INVALIDATE_ASN:
@ -715,7 +715,7 @@ do { \
#define PMAP_INVALIDATE_ASN(pmap, cpu_id) \
do { \
(pmap)->pm_asni[(cpu_id)].pma_asn = PMAP_ASN_RESERVED; \
} while (0)
} while (/*CONSTCOND*/0)
/*
* PMAP_INVALIDATE_TLB:
@ -745,7 +745,7 @@ do { \
* pmap becomes active on this processor, a new \
* ASN will be allocated anyway. \
*/ \
} while (0)
} while (/*CONSTCOND*/0)
/*
* PMAP_KERNEL_PTE:
@ -1317,7 +1317,7 @@ pmap_remove(pmap_t pmap, vaddr_t sva, vaddr_t eva)
* want to remove wired mappings) and pmap_remove() (does
* want to remove wired mappings).
*/
void
static void
pmap_do_remove(pmap_t pmap, vaddr_t sva, vaddr_t eva, boolean_t dowired)
{
pt_entry_t *l1pte, *l2pte, *l3pte;
@ -2553,7 +2553,7 @@ pmap_phys_address(int ppn)
*
* Note: no locking is necessary in this function.
*/
void
static void
alpha_protection_init(void)
{
int prot, *kp, *up;
@ -2600,7 +2600,7 @@ alpha_protection_init(void)
* Returns TRUE or FALSE, indicating if an I-stream sync needs
* to be initiated (for this CPU or for other CPUs).
*/
boolean_t
static boolean_t
pmap_remove_mapping(pmap_t pmap, vaddr_t va, pt_entry_t *pte,
boolean_t dolock, long cpu_id)
{
@ -2706,7 +2706,7 @@ pmap_remove_mapping(pmap_t pmap, vaddr_t va, pt_entry_t *pte,
* the caller has acquired a PV->pmap mutex so that we can lock
* the pmaps as we encounter them.
*/
void
static void
pmap_changebit(struct vm_page *pg, u_long set, u_long mask, long cpu_id)
{
pv_entry_t pv;
@ -2935,7 +2935,7 @@ vtophys(vaddr_t vaddr)
*
* Add a physical->virtual entry to the pv_table.
*/
int
static int
pmap_pv_enter(pmap_t pmap, struct vm_page *pg, vaddr_t va, pt_entry_t *pte,
boolean_t dolock)
{
@ -2986,7 +2986,7 @@ pmap_pv_enter(pmap_t pmap, struct vm_page *pg, vaddr_t va, pt_entry_t *pte,
*
* Remove a physical->virtual entry from the pv_table.
*/
void
static void
pmap_pv_remove(pmap_t pmap, struct vm_page *pg, vaddr_t va, boolean_t dolock)
{
pv_entry_t pv, *pvp;
@ -3020,7 +3020,7 @@ pmap_pv_remove(pmap_t pmap, struct vm_page *pg, vaddr_t va, boolean_t dolock)
*
* Allocate a page for the pv_entry pool.
*/
void *
static void *
pmap_pv_page_alloc(struct pool *pp, int flags)
{
paddr_t pg;
@ -3035,7 +3035,7 @@ pmap_pv_page_alloc(struct pool *pp, int flags)
*
* Free a pv_entry pool page.
*/
void
static void
pmap_pv_page_free(struct pool *pp, void *v)
{
@ -3050,7 +3050,7 @@ pmap_pv_page_free(struct pool *pp, void *v)
* Allocate a single page from the VM system and return the
* physical address for that page.
*/
boolean_t
static boolean_t
pmap_physpage_alloc(int usage, paddr_t *pap)
{
struct vm_page *pg;
@ -3086,7 +3086,7 @@ pmap_physpage_alloc(int usage, paddr_t *pap)
*
* Free the single page table page at the specified physical address.
*/
void
static void
pmap_physpage_free(paddr_t pa)
{
struct vm_page *pg;
@ -3109,7 +3109,7 @@ pmap_physpage_free(paddr_t pa)
*
* Add a reference to the specified special use page.
*/
int
static int
pmap_physpage_addref(void *kva)
{
struct vm_page *pg;
@ -3131,7 +3131,7 @@ pmap_physpage_addref(void *kva)
*
* Delete a reference to the specified special use page.
*/
int
static int
pmap_physpage_delref(void *kva)
{
struct vm_page *pg;
@ -3268,7 +3268,7 @@ pmap_growkernel(vaddr_t maxkvaddr)
*
* Note: growkernel and the pmap must already be locked.
*/
int
static int
pmap_lev1map_create(pmap_t pmap, long cpu_id)
{
pt_entry_t *l1pt;
@ -3324,7 +3324,7 @@ pmap_lev1map_create(pmap_t pmap, long cpu_id)
*
* Note: the pmap must already be locked.
*/
void
static void
pmap_lev1map_destroy(pmap_t pmap, long cpu_id)
{
pt_entry_t *l1pt = pmap->pm_lev1map;
@ -3375,7 +3375,7 @@ pmap_lev1map_destroy(pmap_t pmap, long cpu_id)
*
* Pool cache constructor for L1 PT pages.
*/
int
static int
pmap_l1pt_ctor(void *arg, void *object, int flags)
{
pt_entry_t *l1pt = object, pte;
@ -3408,7 +3408,7 @@ pmap_l1pt_ctor(void *arg, void *object, int flags)
*
* Page alloctaor for L1 PT pages.
*/
void *
static void *
pmap_l1pt_alloc(struct pool *pp, int flags)
{
paddr_t ptpa;
@ -3427,7 +3427,7 @@ pmap_l1pt_alloc(struct pool *pp, int flags)
*
* Page freer for L1 PT pages.
*/
void
static void
pmap_l1pt_free(struct pool *pp, void *v)
{
@ -3442,7 +3442,7 @@ pmap_l1pt_free(struct pool *pp, void *v)
*
* Note: the pmap must already be locked.
*/
int
static int
pmap_ptpage_alloc(pmap_t pmap, pt_entry_t *pte, int usage)
{
paddr_t ptpa;
@ -3471,7 +3471,7 @@ pmap_ptpage_alloc(pmap_t pmap, pt_entry_t *pte, int usage)
*
* Note: the pmap must already be locked.
*/
void
static void
pmap_ptpage_free(pmap_t pmap, pt_entry_t *pte)
{
paddr_t ptpa;
@ -3497,7 +3497,7 @@ pmap_ptpage_free(pmap_t pmap, pt_entry_t *pte)
*
* Note: the pmap must already be locked.
*/
void
static void
pmap_l3pt_delref(pmap_t pmap, vaddr_t va, pt_entry_t *l3pte, long cpu_id)
{
pt_entry_t *l1pte, *l2pte;
@ -3553,7 +3553,7 @@ pmap_l3pt_delref(pmap_t pmap, vaddr_t va, pt_entry_t *l3pte, long cpu_id)
*
* Note: the pmap must already be locked.
*/
void
static void
pmap_l2pt_delref(pmap_t pmap, pt_entry_t *l1pte, pt_entry_t *l2pte,
long cpu_id)
{
@ -3591,7 +3591,7 @@ pmap_l2pt_delref(pmap_t pmap, pt_entry_t *l1pte, pt_entry_t *l2pte,
*
* Note: the pmap must already be locked.
*/
void
static void
pmap_l1pt_delref(pmap_t pmap, pt_entry_t *l1pte, long cpu_id)
{
@ -3622,7 +3622,7 @@ pmap_l1pt_delref(pmap_t pmap, pt_entry_t *l1pte, long cpu_id)
* an interprocessor interrupt, and in that case, the sender of
* the IPI has the pmap lock.
*/
void
static void
pmap_asn_alloc(pmap_t pmap, long cpu_id)
{
struct pmap_asn_info *pma = &pmap->pm_asni[cpu_id];
@ -3914,7 +3914,7 @@ pmap_do_tlb_shootdown(struct cpu_info *ci, struct trapframe *framep)
*
* Note: We expect the queue to be locked.
*/
void
static void
pmap_tlb_shootdown_q_drain(struct pmap_tlb_shootdown_q *pq)
{
struct pmap_tlb_shootdown_job *pj;
@ -3934,7 +3934,7 @@ pmap_tlb_shootdown_q_drain(struct pmap_tlb_shootdown_q *pq)
*
* Note: We expect the queue to be locked.
*/
struct pmap_tlb_shootdown_job *
static struct pmap_tlb_shootdown_job *
pmap_tlb_shootdown_job_get(struct pmap_tlb_shootdown_q *pq)
{
struct pmap_tlb_shootdown_job *pj;
@ -3954,7 +3954,7 @@ pmap_tlb_shootdown_job_get(struct pmap_tlb_shootdown_q *pq)
*
* Note: We expect the queue to be locked.
*/
void
static void
pmap_tlb_shootdown_job_put(struct pmap_tlb_shootdown_q *pq,
struct pmap_tlb_shootdown_job *pj)
{

View File

@ -1,4 +1,4 @@
/* $NetBSD: pmap.h,v 1.66 2006/02/16 20:17:13 perry Exp $ */
/* $NetBSD: pmap.h,v 1.67 2006/04/02 03:41:32 thorpej Exp $ */
/*-
* Copyright (c) 1998, 1999, 2000, 2001 The NetBSD Foundation, Inc.
@ -236,8 +236,6 @@ pmap_remove_all(struct pmap *pmap)
#define pmap_is_modified(pg) \
(((pg)->mdpage.pvh_attrs & PGA_MODIFIED) != 0)
extern pt_entry_t *VPT; /* Virtual Page Table */
#define PMAP_STEAL_MEMORY /* enable pmap_steal_memory() */
#define PMAP_GROWKERNEL /* enable pmap_growkernel() */