Use ANSI function decls.

This commit is contained in:
thorpej 2005-06-27 02:19:48 +00:00
parent e4b2253c50
commit e569facced
25 changed files with 273 additions and 648 deletions

View File

@ -1,4 +1,4 @@
/* $NetBSD: uvm_amap.c,v 1.61 2005/05/17 13:55:33 yamt Exp $ */
/* $NetBSD: uvm_amap.c,v 1.62 2005/06/27 02:19:48 thorpej Exp $ */
/*
*
@ -42,7 +42,7 @@
*/
#include <sys/cdefs.h>
__KERNEL_RCSID(0, "$NetBSD: uvm_amap.c,v 1.61 2005/05/17 13:55:33 yamt Exp $");
__KERNEL_RCSID(0, "$NetBSD: uvm_amap.c,v 1.62 2005/06/27 02:19:48 thorpej Exp $");
#undef UVM_AMAP_INLINE /* enable/disable amap inlines */
@ -78,10 +78,6 @@ static LIST_HEAD(, vm_amap) amap_list;
* local functions
*/
static struct vm_amap *amap_alloc1(int, int, int);
static __inline void amap_list_insert(struct vm_amap *);
static __inline void amap_list_remove(struct vm_amap *);
static __inline void
amap_list_insert(struct vm_amap *amap)
{
@ -134,17 +130,13 @@ amap_list_remove(struct vm_amap *amap)
* here are some in-line functions to help us.
*/
static __inline void pp_getreflen(int *, int, int *, int *);
static __inline void pp_setreflen(int *, int, int, int);
/*
* pp_getreflen: get the reference and length for a specific offset
*
* => ppref's amap must be locked
*/
static __inline void
pp_getreflen(ppref, offset, refp, lenp)
int *ppref, offset, *refp, *lenp;
pp_getreflen(int *ppref, int offset, int *refp, int *lenp)
{
if (ppref[offset] > 0) { /* chunk size must be 1 */
@ -162,8 +154,7 @@ pp_getreflen(ppref, offset, refp, lenp)
* => ppref's amap must be locked
*/
static __inline void
pp_setreflen(ppref, offset, ref, len)
int *ppref, offset, ref, len;
pp_setreflen(int *ppref, int offset, int ref, int len)
{
if (len == 0)
return;
@ -174,7 +165,7 @@ pp_setreflen(ppref, offset, ref, len)
ppref[offset+1] = len;
}
}
#endif
#endif /* UVM_AMAP_PPREF */
/*
* amap_alloc1: internal function that allocates an amap, but does not
@ -183,8 +174,7 @@ pp_setreflen(ppref, offset, ref, len)
* => lock on returned amap is init'd
*/
static inline struct vm_amap *
amap_alloc1(slots, padslots, waitf)
int slots, padslots, waitf;
amap_alloc1(int slots, int padslots, int waitf)
{
struct vm_amap *amap;
int totalslots;
@ -250,9 +240,7 @@ fail1:
*/
struct vm_amap *
amap_alloc(sz, padsz, waitf)
vaddr_t sz, padsz;
int waitf;
amap_alloc(vaddr_t sz, vaddr_t padsz, int waitf)
{
struct vm_amap *amap;
int slots, padslots;
@ -280,8 +268,7 @@ amap_alloc(sz, padsz, waitf)
* => the amap should have a zero reference count and be empty
*/
void
amap_free(amap)
struct vm_amap *amap;
amap_free(struct vm_amap *amap)
{
UVMHIST_FUNC("amap_free"); UVMHIST_CALLED(maphist);
@ -309,10 +296,7 @@ amap_free(amap)
* one (thus it can't be shared)
*/
int
amap_extend(entry, addsize, flags)
struct vm_map_entry *entry;
vsize_t addsize;
int flags;
amap_extend(struct vm_map_entry *entry, vsize_t addsize, int flags)
{
struct vm_amap *amap = entry->aref.ar_amap;
int slotoff = entry->aref.ar_pageoff;
@ -633,9 +617,7 @@ amap_extend(entry, addsize, flags)
* => entry's map and amap must be locked by the caller
*/
void
amap_share_protect(entry, prot)
struct vm_map_entry *entry;
vm_prot_t prot;
amap_share_protect(struct vm_map_entry *entry, vm_prot_t prot)
{
struct vm_amap *amap = entry->aref.ar_amap;
int slots, lcv, slot, stop;
@ -676,8 +658,7 @@ amap_share_protect(entry, prot)
*/
void
amap_wipeout(amap)
struct vm_amap *amap;
amap_wipeout(struct vm_amap *amap)
{
int lcv, slot;
struct vm_anon *anon;
@ -747,12 +728,8 @@ amap_wipeout(amap)
*/
void
amap_copy(map, entry, waitf, canchunk, startva, endva)
struct vm_map *map;
struct vm_map_entry *entry;
int waitf;
boolean_t canchunk;
vaddr_t startva, endva;
amap_copy(struct vm_map *map, struct vm_map_entry *entry, int waitf,
boolean_t canchunk, vaddr_t startva, vaddr_t endva)
{
struct vm_amap *amap, *srcamap;
int slots, lcv;
@ -924,9 +901,7 @@ amap_copy(map, entry, waitf, canchunk, startva, endva)
*/
void
amap_cow_now(map, entry)
struct vm_map *map;
struct vm_map_entry *entry;
amap_cow_now(struct vm_map *map, struct vm_map_entry *entry)
{
struct vm_amap *amap = entry->aref.ar_amap;
int lcv, slot;
@ -1047,9 +1022,7 @@ ReStart:
* => origref->ar_amap should be unlocked (we will lock)
*/
void
amap_splitref(origref, splitref, offset)
struct vm_aref *origref, *splitref;
vaddr_t offset;
amap_splitref(struct vm_aref *origref, struct vm_aref *splitref, vaddr_t offset)
{
int leftslots;
@ -1089,9 +1062,7 @@ amap_splitref(origref, splitref, offset)
* => amap locked by caller
*/
void
amap_pp_establish(amap, offset)
struct vm_amap *amap;
vaddr_t offset;
amap_pp_establish(struct vm_amap *amap, vaddr_t offset)
{
amap->am_ppref = malloc(sizeof(int) * amap->am_maxslot,
M_UVMAMAP, M_NOWAIT);
@ -1119,11 +1090,7 @@ amap_pp_establish(amap, offset)
* => caller must check that ppref != PPREF_NONE before calling
*/
void
amap_pp_adjref(amap, curslot, slotlen, adjval)
struct vm_amap *amap;
int curslot;
vsize_t slotlen;
int adjval;
amap_pp_adjref(struct vm_amap *amap, int curslot, vsize_t slotlen, int adjval)
{
int stopslot, *ppref, lcv, prevlcv;
int ref, len, prevref, prevlen;
@ -1195,9 +1162,7 @@ amap_pp_adjref(amap, curslot, slotlen, adjval)
* => both map and amap must be locked by caller.
*/
void
amap_wiperange(amap, slotoff, slots)
struct vm_amap *amap;
int slotoff, slots;
amap_wiperange(struct vm_amap *amap, int slotoff, int slots)
{
int byanon, lcv, stop, curslot, ptr, slotend;
struct vm_anon *anon;

View File

@ -1,4 +1,4 @@
/* $NetBSD: uvm_amap_i.h,v 1.22 2005/05/11 13:02:25 yamt Exp $ */
/* $NetBSD: uvm_amap_i.h,v 1.23 2005/06/27 02:19:48 thorpej Exp $ */
/*
*
@ -52,9 +52,7 @@
* => amap should be locked by caller.
*/
AMAP_INLINE struct vm_anon *
amap_lookup(aref, offset)
struct vm_aref *aref;
vaddr_t offset;
amap_lookup(struct vm_aref *aref, vaddr_t offset)
{
int slot;
struct vm_amap *amap = aref->ar_amap;
@ -79,11 +77,8 @@ amap_lookup(aref, offset)
* => XXXCDC: this interface is biased toward array-based amaps. fix.
*/
AMAP_INLINE void
amap_lookups(aref, offset, anons, npages)
struct vm_aref *aref;
vaddr_t offset;
struct vm_anon **anons;
int npages;
amap_lookups(struct vm_aref *aref, vaddr_t offset, struct vm_anon **anons,
int npages)
{
int slot;
struct vm_amap *amap = aref->ar_amap;
@ -113,11 +108,8 @@ amap_lookups(aref, offset, anons, npages)
* pmap_page_protect on the anon's page.
*/
AMAP_INLINE void
amap_add(aref, offset, anon, replace)
struct vm_aref *aref;
vaddr_t offset;
struct vm_anon *anon;
boolean_t replace;
amap_add(struct vm_aref *aref, vaddr_t offset, struct vm_anon *anon,
boolean_t replace)
{
int slot;
struct vm_amap *amap = aref->ar_amap;
@ -162,9 +154,7 @@ amap_add(aref, offset, anon, replace)
* => caller must lock amap
*/
AMAP_INLINE void
amap_unadd(aref, offset)
struct vm_aref *aref;
vaddr_t offset;
amap_unadd(struct vm_aref *aref, vaddr_t offset)
{
int ptr, slot;
struct vm_amap *amap = aref->ar_amap;
@ -199,11 +189,7 @@ amap_unadd(aref, offset)
* => called at fork time to gain the child's reference
*/
AMAP_INLINE void
amap_ref(amap, offset, len, flags)
struct vm_amap *amap;
vaddr_t offset;
vsize_t len;
int flags;
amap_ref(struct vm_amap *amap, vaddr_t offset, vsize_t len, int flags)
{
UVMHIST_FUNC("amap_ref"); UVMHIST_CALLED(maphist);
@ -238,11 +224,7 @@ amap_ref(amap, offset, len, flags)
* => amap must be unlocked (we will lock it).
*/
AMAP_INLINE void
amap_unref(amap, offset, len, all)
struct vm_amap *amap;
vaddr_t offset;
vsize_t len;
boolean_t all;
amap_unref(struct vm_amap *amap, vaddr_t offset, vsize_t len, boolean_t all)
{
UVMHIST_FUNC("amap_unref"); UVMHIST_CALLED(maphist);

View File

@ -1,4 +1,4 @@
/* $NetBSD: uvm_anon.c,v 1.33 2005/05/11 13:02:25 yamt Exp $ */
/* $NetBSD: uvm_anon.c,v 1.34 2005/06/27 02:19:48 thorpej Exp $ */
/*
*
@ -37,7 +37,7 @@
*/
#include <sys/cdefs.h>
__KERNEL_RCSID(0, "$NetBSD: uvm_anon.c,v 1.33 2005/05/11 13:02:25 yamt Exp $");
__KERNEL_RCSID(0, "$NetBSD: uvm_anon.c,v 1.34 2005/06/27 02:19:48 thorpej Exp $");
#include "opt_uvmhist.h"
@ -61,7 +61,7 @@ static int uvm_anon_ctor(void *, void *, int);
* allocate anons
*/
void
uvm_anon_init()
uvm_anon_init(void)
{
pool_cache_init(&uvm_anon_pool_cache, &uvm_anon_pool,
@ -87,7 +87,7 @@ uvm_anon_ctor(void *arg, void *object, int flags)
* => new anon is returned locked!
*/
struct vm_anon *
uvm_analloc()
uvm_analloc(void)
{
struct vm_anon *anon;
@ -113,8 +113,7 @@ uvm_analloc()
*/
void
uvm_anfree(anon)
struct vm_anon *anon;
uvm_anfree(struct vm_anon *anon)
{
struct vm_page *pg;
UVMHIST_FUNC("uvm_anfree"); UVMHIST_CALLED(maphist);
@ -220,8 +219,7 @@ uvm_anfree(anon)
* => anon must be locked or have a reference count of 0.
*/
void
uvm_anon_dropswap(anon)
struct vm_anon *anon;
uvm_anon_dropswap(struct vm_anon *anon)
{
UVMHIST_FUNC("uvm_anon_dropswap"); UVMHIST_CALLED(maphist);
@ -252,8 +250,7 @@ uvm_anon_dropswap(anon)
* count.
*/
struct vm_page *
uvm_anon_lockloanpg(anon)
struct vm_anon *anon;
uvm_anon_lockloanpg(struct vm_anon *anon)
{
struct vm_page *pg;
boolean_t locked = FALSE;
@ -331,8 +328,7 @@ uvm_anon_lockloanpg(anon)
*/
boolean_t
uvm_anon_pagein(anon)
struct vm_anon *anon;
uvm_anon_pagein(struct vm_anon *anon)
{
struct vm_page *pg;
struct uvm_object *uobj;
@ -412,8 +408,7 @@ uvm_anon_pagein(anon)
*/
void
uvm_anon_release(anon)
struct vm_anon *anon;
uvm_anon_release(struct vm_anon *anon)
{
struct vm_page *pg = anon->an_page;

View File

@ -1,4 +1,4 @@
/* $NetBSD: uvm_aobj.c,v 1.66 2005/06/06 12:09:19 yamt Exp $ */
/* $NetBSD: uvm_aobj.c,v 1.67 2005/06/27 02:19:48 thorpej Exp $ */
/*
* Copyright (c) 1998 Chuck Silvers, Charles D. Cranor and
@ -43,7 +43,7 @@
*/
#include <sys/cdefs.h>
__KERNEL_RCSID(0, "$NetBSD: uvm_aobj.c,v 1.66 2005/06/06 12:09:19 yamt Exp $");
__KERNEL_RCSID(0, "$NetBSD: uvm_aobj.c,v 1.67 2005/06/27 02:19:48 thorpej Exp $");
#include "opt_uvmhist.h"
@ -224,10 +224,7 @@ static struct simplelock uao_list_lock;
*/
static struct uao_swhash_elt *
uao_find_swhash_elt(aobj, pageidx, create)
struct uvm_aobj *aobj;
int pageidx;
boolean_t create;
uao_find_swhash_elt(struct uvm_aobj *aobj, int pageidx, boolean_t create)
{
struct uao_swhash *swhash;
struct uao_swhash_elt *elt;
@ -271,9 +268,7 @@ uao_find_swhash_elt(aobj, pageidx, create)
*/
int
uao_find_swslot(uobj, pageidx)
struct uvm_object *uobj;
int pageidx;
uao_find_swslot(struct uvm_object *uobj, int pageidx)
{
struct uvm_aobj *aobj = (struct uvm_aobj *)uobj;
struct uao_swhash_elt *elt;
@ -314,9 +309,7 @@ uao_find_swslot(uobj, pageidx)
*/
int
uao_set_swslot(uobj, pageidx, slot)
struct uvm_object *uobj;
int pageidx, slot;
uao_set_swslot(struct uvm_object *uobj, int pageidx, int slot)
{
struct uvm_aobj *aobj = (struct uvm_aobj *)uobj;
struct uao_swhash_elt *elt;
@ -393,8 +386,7 @@ uao_set_swslot(uobj, pageidx, slot)
*/
static void
uao_free(aobj)
struct uvm_aobj *aobj;
uao_free(struct uvm_aobj *aobj)
{
int swpgonlydelta = 0;
@ -480,9 +472,7 @@ uao_free(aobj)
*/
struct uvm_object *
uao_create(size, flags)
vsize_t size;
int flags;
uao_create(vsize_t size, int flags)
{
static struct uvm_aobj kernel_object_store;
static int kobj_alloced = 0;
@ -588,8 +578,7 @@ uao_init(void)
*/
void
uao_reference(uobj)
struct uvm_object *uobj;
uao_reference(struct uvm_object *uobj)
{
simple_lock(&uobj->vmobjlock);
uao_reference_locked(uobj);
@ -606,8 +595,7 @@ uao_reference(uobj)
*/
void
uao_reference_locked(uobj)
struct uvm_object *uobj;
uao_reference_locked(struct uvm_object *uobj)
{
UVMHIST_FUNC("uao_reference"); UVMHIST_CALLED(maphist);
@ -631,8 +619,7 @@ uao_reference_locked(uobj)
*/
void
uao_detach(uobj)
struct uvm_object *uobj;
uao_detach(struct uvm_object *uobj)
{
simple_lock(&uobj->vmobjlock);
uao_detach_locked(uobj);
@ -648,8 +635,7 @@ uao_detach(uobj)
*/
void
uao_detach_locked(uobj)
struct uvm_object *uobj;
uao_detach_locked(struct uvm_object *uobj)
{
struct uvm_aobj *aobj = (struct uvm_aobj *)uobj;
struct vm_page *pg;
@ -751,10 +737,7 @@ uao_detach_locked(uobj)
*/
int
uao_put(uobj, start, stop, flags)
struct uvm_object *uobj;
voff_t start, stop;
int flags;
uao_put(struct uvm_object *uobj, voff_t start, voff_t stop, int flags)
{
struct uvm_aobj *aobj = (struct uvm_aobj *)uobj;
struct vm_page *pg, *nextpg, curmp, endmp;
@ -934,13 +917,8 @@ uao_put(uobj, start, stop, flags)
*/
static int
uao_get(uobj, offset, pps, npagesp, centeridx, access_type, advice, flags)
struct uvm_object *uobj;
voff_t offset;
struct vm_page **pps;
int *npagesp;
int centeridx, advice, flags;
vm_prot_t access_type;
uao_get(struct uvm_object *uobj, voff_t offset, struct vm_page **pps,
int *npagesp, int centeridx, vm_prot_t access_type, int advice, int flags)
{
struct uvm_aobj *aobj = (struct uvm_aobj *)uobj;
voff_t current_offset;
@ -1231,9 +1209,7 @@ gotpage:
*/
void
uao_dropswap(uobj, pageidx)
struct uvm_object *uobj;
int pageidx;
uao_dropswap(struct uvm_object *uobj, int pageidx)
{
int slot;
@ -1251,8 +1227,7 @@ uao_dropswap(uobj, pageidx)
*/
boolean_t
uao_swap_off(startslot, endslot)
int startslot, endslot;
uao_swap_off(int startslot, int endslot)
{
struct uvm_aobj *aobj, *nextaobj;
boolean_t rv;
@ -1327,9 +1302,7 @@ restart:
* => returns TRUE if pagein was aborted due to lack of memory.
*/
static boolean_t
uao_pagein(aobj, startslot, endslot)
struct uvm_aobj *aobj;
int startslot, endslot;
uao_pagein(struct uvm_aobj *aobj, int startslot, int endslot)
{
boolean_t rv;
@ -1408,9 +1381,7 @@ restart:
*/
static boolean_t
uao_pagein_page(aobj, pageidx)
struct uvm_aobj *aobj;
int pageidx;
uao_pagein_page(struct uvm_aobj *aobj, int pageidx)
{
struct vm_page *pg;
int rv, npages;

View File

@ -1,4 +1,4 @@
/* $NetBSD: uvm_bio.c,v 1.38 2005/06/06 12:09:19 yamt Exp $ */
/* $NetBSD: uvm_bio.c,v 1.39 2005/06/27 02:19:48 thorpej Exp $ */
/*
* Copyright (c) 1998 Chuck Silvers.
@ -34,7 +34,7 @@
*/
#include <sys/cdefs.h>
__KERNEL_RCSID(0, "$NetBSD: uvm_bio.c,v 1.38 2005/06/06 12:09:19 yamt Exp $");
__KERNEL_RCSID(0, "$NetBSD: uvm_bio.c,v 1.39 2005/06/27 02:19:48 thorpej Exp $");
#include "opt_uvmhist.h"
@ -53,9 +53,9 @@ __KERNEL_RCSID(0, "$NetBSD: uvm_bio.c,v 1.38 2005/06/06 12:09:19 yamt Exp $");
* local functions
*/
int ubc_fault(struct uvm_faultinfo *, vaddr_t, struct vm_page **, int,
int, vm_fault_t, vm_prot_t, int);
struct ubc_map *ubc_find_mapping(struct uvm_object *, voff_t);
static int ubc_fault(struct uvm_faultinfo *, vaddr_t, struct vm_page **,
int, int, vm_fault_t, vm_prot_t, int);
static struct ubc_map *ubc_find_mapping(struct uvm_object *, voff_t);
/*
* local data structues
@ -201,15 +201,10 @@ ubc_init(void)
* ubc_fault: fault routine for ubc mapping
*/
int
ubc_fault(ufi, ign1, ign2, ign3, ign4, fault_type, access_type, flags)
struct uvm_faultinfo *ufi;
vaddr_t ign1;
struct vm_page **ign2;
int ign3, ign4;
vm_fault_t fault_type;
vm_prot_t access_type;
int flags;
static int
ubc_fault(struct uvm_faultinfo *ufi, vaddr_t ign1, struct vm_page **ign2,
int ign3, int ign4, vm_fault_t fault_type, vm_prot_t access_type,
int flags)
{
struct uvm_object *uobj;
struct ubc_map *umap;
@ -364,10 +359,8 @@ again:
* local functions
*/
struct ubc_map *
ubc_find_mapping(uobj, offset)
struct uvm_object *uobj;
voff_t offset;
static struct ubc_map *
ubc_find_mapping(struct uvm_object *uobj, voff_t offset)
{
struct ubc_map *umap;
@ -389,11 +382,7 @@ ubc_find_mapping(uobj, offset)
*/
void *
ubc_alloc(uobj, offset, lenp, flags)
struct uvm_object *uobj;
voff_t offset;
vsize_t *lenp;
int flags;
ubc_alloc(struct uvm_object *uobj, voff_t offset, vsize_t *lenp, int flags)
{
vaddr_t slot_offset, va;
struct ubc_map *umap;
@ -501,9 +490,7 @@ out:
*/
void
ubc_release(va, flags)
void *va;
int flags;
ubc_release(void *va, int flags)
{
struct ubc_map *umap;
struct uvm_object *uobj;
@ -593,9 +580,7 @@ ubc_release(va, flags)
*/
void
ubc_flush(uobj, start, end)
struct uvm_object *uobj;
voff_t start, end;
ubc_flush(struct uvm_object *uobj, voff_t start, voff_t end)
{
struct ubc_map *umap;
vaddr_t va;

View File

@ -1,4 +1,4 @@
/* $NetBSD: uvm_device.c,v 1.43 2005/06/06 12:09:19 yamt Exp $ */
/* $NetBSD: uvm_device.c,v 1.44 2005/06/27 02:19:48 thorpej Exp $ */
/*
*
@ -39,7 +39,7 @@
*/
#include <sys/cdefs.h>
__KERNEL_RCSID(0, "$NetBSD: uvm_device.c,v 1.43 2005/06/06 12:09:19 yamt Exp $");
__KERNEL_RCSID(0, "$NetBSD: uvm_device.c,v 1.44 2005/06/27 02:19:48 thorpej Exp $");
#include "opt_uvmhist.h"
@ -112,11 +112,9 @@ udv_init(void)
*/
struct uvm_object *
udv_attach(arg, accessprot, off, size)
void *arg;
vm_prot_t accessprot;
voff_t off; /* used only for access check */
vsize_t size; /* used only for access check */
udv_attach(void *arg, vm_prot_t accessprot,
voff_t off, /* used only for access check */
vsize_t size /* used only for access check */)
{
dev_t device = *((dev_t *)arg);
struct uvm_device *udv, *lcv;
@ -270,8 +268,7 @@ udv_attach(arg, accessprot, off, size)
*/
static void
udv_reference(uobj)
struct uvm_object *uobj;
udv_reference(struct uvm_object *uobj)
{
UVMHIST_FUNC("udv_reference"); UVMHIST_CALLED(maphist);
@ -291,8 +288,7 @@ udv_reference(uobj)
*/
static void
udv_detach(uobj)
struct uvm_object *uobj;
udv_detach(struct uvm_object *uobj)
{
struct uvm_device *udv = (struct uvm_device *)uobj;
UVMHIST_FUNC("udv_detach"); UVMHIST_CALLED(maphist);
@ -352,13 +348,9 @@ again:
*/
static int
udv_fault(ufi, vaddr, pps, npages, centeridx, fault_type, access_type, flags)
struct uvm_faultinfo *ufi;
vaddr_t vaddr;
struct vm_page **pps;
int npages, centeridx, flags;
vm_fault_t fault_type;
vm_prot_t access_type;
udv_fault(struct uvm_faultinfo *ufi, vaddr_t vaddr, struct vm_page **pps,
int npages, int centeridx, vm_fault_t fault_type, vm_prot_t access_type,
int flags)
{
struct vm_map_entry *entry = ufi->entry;
struct uvm_object *uobj = entry->object.uvm_obj;

View File

@ -1,4 +1,4 @@
/* $NetBSD: uvm_fault.c,v 1.94 2005/05/11 13:02:25 yamt Exp $ */
/* $NetBSD: uvm_fault.c,v 1.95 2005/06/27 02:19:48 thorpej Exp $ */
/*
*
@ -39,7 +39,7 @@
*/
#include <sys/cdefs.h>
__KERNEL_RCSID(0, "$NetBSD: uvm_fault.c,v 1.94 2005/05/11 13:02:25 yamt Exp $");
__KERNEL_RCSID(0, "$NetBSD: uvm_fault.c,v 1.95 2005/06/27 02:19:48 thorpej Exp $");
#include "opt_uvmhist.h"
@ -167,7 +167,7 @@ struct uvm_advice {
* XXX: borrowed numbers from freebsd. do they work well for us?
*/
static struct uvm_advice uvmadvice[] = {
static const struct uvm_advice uvmadvice[] = {
{ MADV_NORMAL, 3, 4 },
{ MADV_RANDOM, 0, 0 },
{ MADV_SEQUENTIAL, 8, 7},
@ -179,9 +179,6 @@ static struct uvm_advice uvmadvice[] = {
* private prototypes
*/
static void uvmfault_amapcopy(struct uvm_faultinfo *);
static __inline void uvmfault_anonflush(struct vm_anon **, int);
/*
* inline functions
*/
@ -193,9 +190,7 @@ static __inline void uvmfault_anonflush(struct vm_anon **, int);
*/
static __inline void
uvmfault_anonflush(anons, n)
struct vm_anon **anons;
int n;
uvmfault_anonflush(struct vm_anon **anons, int n)
{
int lcv;
struct vm_page *pg;
@ -230,8 +225,7 @@ uvmfault_anonflush(anons, n)
*/
static void
uvmfault_amapcopy(ufi)
struct uvm_faultinfo *ufi;
uvmfault_amapcopy(struct uvm_faultinfo *ufi)
{
for (;;) {
@ -287,10 +281,8 @@ uvmfault_amapcopy(ufi)
*/
int
uvmfault_anonget(ufi, amap, anon)
struct uvm_faultinfo *ufi;
struct vm_amap *amap;
struct vm_anon *anon;
uvmfault_anonget(struct uvm_faultinfo *ufi, struct vm_amap *amap,
struct vm_anon *anon)
{
boolean_t we_own; /* we own anon's page? */
boolean_t locked; /* did we relock? */
@ -554,11 +546,8 @@ released:
~VM_PROT_WRITE : VM_PROT_ALL)
int
uvm_fault(orig_map, vaddr, fault_type, access_type)
struct vm_map *orig_map;
vaddr_t vaddr;
vm_fault_t fault_type;
vm_prot_t access_type;
uvm_fault(struct vm_map *orig_map, vaddr_t vaddr, vm_fault_t fault_type,
vm_prot_t access_type)
{
struct uvm_faultinfo ufi;
vm_prot_t enter_prot, check_prot;
@ -1726,11 +1715,8 @@ Case2:
*/
int
uvm_fault_wire(map, start, end, fault_type, access_type)
struct vm_map *map;
vaddr_t start, end;
vm_fault_t fault_type;
vm_prot_t access_type;
uvm_fault_wire(struct vm_map *map, vaddr_t start, vaddr_t end,
vm_fault_t fault_type, vm_prot_t access_type)
{
vaddr_t va;
int error;
@ -1766,9 +1752,7 @@ uvm_fault_wire(map, start, end, fault_type, access_type)
*/
void
uvm_fault_unwire(map, start, end)
struct vm_map *map;
vaddr_t start, end;
uvm_fault_unwire(struct vm_map *map, vaddr_t start, vaddr_t end)
{
vm_map_lock_read(map);
uvm_fault_unwire_locked(map, start, end);
@ -1782,9 +1766,7 @@ uvm_fault_unwire(map, start, end)
*/
void
uvm_fault_unwire_locked(map, start, end)
struct vm_map *map;
vaddr_t start, end;
uvm_fault_unwire_locked(struct vm_map *map, vaddr_t start, vaddr_t end)
{
struct vm_map_entry *entry;
pmap_t pmap = vm_map_pmap(map);

View File

@ -1,4 +1,4 @@
/* $NetBSD: uvm_fault_i.h,v 1.17 2004/03/24 07:55:01 junyoung Exp $ */
/* $NetBSD: uvm_fault_i.h,v 1.18 2005/06/27 02:19:48 thorpej Exp $ */
/*
*
@ -40,20 +40,13 @@
/*
* uvm_fault_i.h: fault inline functions
*/
static boolean_t uvmfault_lookup(struct uvm_faultinfo *, boolean_t);
static boolean_t uvmfault_relock(struct uvm_faultinfo *);
static void uvmfault_unlockall(struct uvm_faultinfo *, struct vm_amap *,
struct uvm_object *, struct vm_anon *);
static void uvmfault_unlockmaps(struct uvm_faultinfo *, boolean_t);
/*
* uvmfault_unlockmaps: unlock the maps
*/
static __inline void
uvmfault_unlockmaps(ufi, write_locked)
struct uvm_faultinfo *ufi;
boolean_t write_locked;
uvmfault_unlockmaps(struct uvm_faultinfo *ufi, boolean_t write_locked)
{
/*
* ufi can be NULL when this isn't really a fault,
@ -78,11 +71,8 @@ uvmfault_unlockmaps(ufi, write_locked)
*/
static __inline void
uvmfault_unlockall(ufi, amap, uobj, anon)
struct uvm_faultinfo *ufi;
struct vm_amap *amap;
struct uvm_object *uobj;
struct vm_anon *anon;
uvmfault_unlockall(struct uvm_faultinfo *ufi, struct vm_amap *amap,
struct uvm_object *uobj, struct vm_anon *anon)
{
if (anon)
@ -110,9 +100,7 @@ uvmfault_unlockall(ufi, amap, uobj, anon)
*/
static __inline boolean_t
uvmfault_lookup(ufi, write_lock)
struct uvm_faultinfo *ufi;
boolean_t write_lock;
uvmfault_lookup(struct uvm_faultinfo *ufi, boolean_t write_lock)
{
struct vm_map *tmpmap;
@ -197,8 +185,7 @@ uvmfault_lookup(ufi, write_lock)
*/
static __inline boolean_t
uvmfault_relock(ufi)
struct uvm_faultinfo *ufi;
uvmfault_relock(struct uvm_faultinfo *ufi)
{
/*
* ufi can be NULL when this isn't really a fault,

View File

@ -1,4 +1,4 @@
/* $NetBSD: uvm_glue.c,v 1.88 2005/06/10 05:10:13 matt Exp $ */
/* $NetBSD: uvm_glue.c,v 1.89 2005/06/27 02:19:48 thorpej Exp $ */
/*
* Copyright (c) 1997 Charles D. Cranor and Washington University.
@ -67,7 +67,7 @@
*/
#include <sys/cdefs.h>
__KERNEL_RCSID(0, "$NetBSD: uvm_glue.c,v 1.88 2005/06/10 05:10:13 matt Exp $");
__KERNEL_RCSID(0, "$NetBSD: uvm_glue.c,v 1.89 2005/06/27 02:19:48 thorpej Exp $");
#include "opt_kgdb.h"
#include "opt_kstack.h"
@ -112,10 +112,7 @@ static void uvm_uarea_free(vaddr_t);
*/
boolean_t
uvm_kernacc(addr, len, rw)
caddr_t addr;
size_t len;
int rw;
uvm_kernacc(caddr_t addr, size_t len, int rw)
{
boolean_t rv;
vaddr_t saddr, eaddr;
@ -144,10 +141,7 @@ uvm_kernacc(addr, len, rw)
* we can ensure the change takes place properly.
*/
void
uvm_chgkprot(addr, len, rw)
caddr_t addr;
size_t len;
int rw;
uvm_chgkprot(caddr_t addr, size_t len, int rw)
{
vm_prot_t prot;
paddr_t pa;
@ -175,11 +169,7 @@ uvm_chgkprot(addr, len, rw)
*/
int
uvm_vslock(p, addr, len, access_type)
struct proc *p;
caddr_t addr;
size_t len;
vm_prot_t access_type;
uvm_vslock(struct proc *p, caddr_t addr, size_t len, vm_prot_t access_type)
{
struct vm_map *map;
vaddr_t start, end;
@ -200,10 +190,7 @@ uvm_vslock(p, addr, len, access_type)
*/
void
uvm_vsunlock(p, addr, len)
struct proc *p;
caddr_t addr;
size_t len;
uvm_vsunlock(struct proc *p, caddr_t addr, size_t len)
{
uvm_fault_unwire(&p->p_vmspace->vm_map, trunc_page((vaddr_t)addr),
round_page((vaddr_t)addr + len));
@ -215,9 +202,7 @@ uvm_vsunlock(p, addr, len)
* - the address space is copied as per parent map's inherit values
*/
void
uvm_proc_fork(p1, p2, shared)
struct proc *p1, *p2;
boolean_t shared;
uvm_proc_fork(struct proc *p1, struct proc *p2, boolean_t shared)
{
if (shared == TRUE) {
@ -246,12 +231,8 @@ uvm_proc_fork(p1, p2, shared)
* than just hang
*/
void
uvm_lwp_fork(l1, l2, stack, stacksize, func, arg)
struct lwp *l1, *l2;
void *stack;
size_t stacksize;
void (*func)(void *);
void *arg;
uvm_lwp_fork(struct lwp *l1, struct lwp *l2, void *stack, size_t stacksize,
void (*func)(void *), void *arg)
{
struct user *up = l2->l_addr;
int error;
@ -375,8 +356,7 @@ uvm_uarea_drain(boolean_t empty)
*/
void
uvm_proc_exit(p)
struct proc *p;
uvm_proc_exit(struct proc *p)
{
struct lwp *l = curlwp; /* XXX */
struct vmspace *ovm;
@ -411,8 +391,7 @@ uvm_lwp_exit(struct lwp *l)
*/
void
uvm_init_limits(p)
struct proc *p;
uvm_init_limits(struct proc *p)
{
/*
@ -442,8 +421,7 @@ int swapdebug = 0;
*/
void
uvm_swapin(l)
struct lwp *l;
uvm_swapin(struct lwp *l)
{
vaddr_t addr;
int s, error;
@ -479,7 +457,7 @@ uvm_swapin(l)
*/
void
uvm_scheduler()
uvm_scheduler(void)
{
struct lwp *l, *ll;
int pri;
@ -577,7 +555,7 @@ loop:
*/
void
uvm_swapout_threads()
uvm_swapout_threads(void)
{
struct lwp *l;
struct lwp *outl, *outl2;
@ -655,8 +633,7 @@ uvm_swapout_threads()
*/
static void
uvm_swapout(l)
struct lwp *l;
uvm_swapout(struct lwp *l)
{
vaddr_t addr;
int s;
@ -706,11 +683,9 @@ uvm_swapout(l)
*/
int
uvm_coredump_walkmap(p, iocookie, func, cookie)
struct proc *p;
void *iocookie;
int (*func)(struct proc *, void *, struct uvm_coredump_state *);
void *cookie;
uvm_coredump_walkmap(struct proc *p, void *iocookie,
int (*func)(struct proc *, void *, struct uvm_coredump_state *),
void *cookie)
{
struct uvm_coredump_state state;
struct vmspace *vm = p->p_vmspace;

View File

@ -1,4 +1,4 @@
/* $NetBSD: uvm_init.c,v 1.22 2005/05/11 13:02:25 yamt Exp $ */
/* $NetBSD: uvm_init.c,v 1.23 2005/06/27 02:19:48 thorpej Exp $ */
/*
*
@ -39,7 +39,7 @@
*/
#include <sys/cdefs.h>
__KERNEL_RCSID(0, "$NetBSD: uvm_init.c,v 1.22 2005/05/11 13:02:25 yamt Exp $");
__KERNEL_RCSID(0, "$NetBSD: uvm_init.c,v 1.23 2005/06/27 02:19:48 thorpej Exp $");
#include <sys/param.h>
#include <sys/systm.h>
@ -70,7 +70,7 @@ struct uvmexp uvmexp; /* decl */
*/
void
uvm_init()
uvm_init(void)
{
vaddr_t kvm_start, kvm_end;

View File

@ -1,4 +1,4 @@
/* $NetBSD: uvm_io.c,v 1.20 2005/04/01 11:59:38 yamt Exp $ */
/* $NetBSD: uvm_io.c,v 1.21 2005/06/27 02:19:48 thorpej Exp $ */
/*
*
@ -39,7 +39,7 @@
*/
#include <sys/cdefs.h>
__KERNEL_RCSID(0, "$NetBSD: uvm_io.c,v 1.20 2005/04/01 11:59:38 yamt Exp $");
__KERNEL_RCSID(0, "$NetBSD: uvm_io.c,v 1.21 2005/06/27 02:19:48 thorpej Exp $");
#include <sys/param.h>
#include <sys/systm.h>
@ -62,9 +62,7 @@ __KERNEL_RCSID(0, "$NetBSD: uvm_io.c,v 1.20 2005/04/01 11:59:38 yamt Exp $");
*/
int
uvm_io(map, uio)
struct vm_map *map;
struct uio *uio;
uvm_io(struct vm_map *map, struct uio *uio)
{
vaddr_t baseva, endva, pageoffset, kva;
vsize_t chunksz, togo, sz;

View File

@ -1,4 +1,4 @@
/* $NetBSD: uvm_km.c,v 1.82 2005/05/29 21:06:33 christos Exp $ */
/* $NetBSD: uvm_km.c,v 1.83 2005/06/27 02:19:48 thorpej Exp $ */
/*
* Copyright (c) 1997 Charles D. Cranor and Washington University.
@ -130,7 +130,7 @@
*/
#include <sys/cdefs.h>
__KERNEL_RCSID(0, "$NetBSD: uvm_km.c,v 1.82 2005/05/29 21:06:33 christos Exp $");
__KERNEL_RCSID(0, "$NetBSD: uvm_km.c,v 1.83 2005/06/27 02:19:48 thorpej Exp $");
#include "opt_uvmhist.h"
@ -263,8 +263,7 @@ uvm_km_vacache_init(struct vm_map *map, const char *name, size_t size)
*/
void
uvm_km_init(start, end)
vaddr_t start, end;
uvm_km_init(vaddr_t start, vaddr_t end)
{
vaddr_t base = VM_MIN_KERNEL_ADDRESS;
@ -325,13 +324,9 @@ uvm_km_init(start, end)
*/
struct vm_map *
uvm_km_suballoc(map, vmin, vmax, size, flags, fixed, submap)
struct vm_map *map;
vaddr_t *vmin, *vmax; /* IN/OUT, OUT */
vsize_t size;
int flags;
boolean_t fixed;
struct vm_map_kernel *submap;
uvm_km_suballoc(struct vm_map *map, vaddr_t *vmin /* IN/OUT */,
vaddr_t *vmax /* OUT */, vsize_t size, int flags, boolean_t fixed,
struct vm_map_kernel *submap)
{
int mapflags = UVM_FLAG_NOMERGE | (fixed ? UVM_FLAG_FIXED : 0);
@ -386,8 +381,7 @@ uvm_km_suballoc(map, vmin, vmax, size, flags, fixed, submap)
*/
void
uvm_km_pgremove(startva, endva)
vaddr_t startva, endva;
uvm_km_pgremove(vaddr_t startva, vaddr_t endva)
{
struct uvm_object * const uobj = uvm.kernel_object;
const voff_t start = startva - vm_map_min(kernel_map);
@ -452,8 +446,7 @@ uvm_km_pgremove(startva, endva)
*/
void
uvm_km_pgremove_intrsafe(start, end)
vaddr_t start, end;
uvm_km_pgremove_intrsafe(vaddr_t start, vaddr_t end)
{
struct vm_page *pg;
paddr_t pa;
@ -516,11 +509,7 @@ uvm_km_check_empty(vaddr_t start, vaddr_t end, boolean_t intrsafe)
*/
vaddr_t
uvm_km_alloc(map, size, align, flags)
struct vm_map *map;
vsize_t size;
vsize_t align;
uvm_flag_t flags;
uvm_km_alloc(struct vm_map *map, vsize_t size, vsize_t align, uvm_flag_t flags)
{
vaddr_t kva, loopva;
vaddr_t offset;
@ -632,11 +621,7 @@ uvm_km_alloc(map, size, align, flags)
*/
void
uvm_km_free(map, addr, size, flags)
struct vm_map *map;
vaddr_t addr;
vsize_t size;
uvm_flag_t flags;
uvm_km_free(struct vm_map *map, vaddr_t addr, vsize_t size, uvm_flag_t flags)
{
KASSERT((flags & UVM_KMF_TYPEMASK) == UVM_KMF_WIRED ||
@ -672,9 +657,7 @@ uvm_km_free(map, addr, size, flags)
/* ARGSUSED */
vaddr_t
uvm_km_alloc_poolpage_cache(map, waitok)
struct vm_map *map;
boolean_t waitok;
uvm_km_alloc_poolpage_cache(struct vm_map *map, boolean_t waitok)
{
#if defined(PMAP_MAP_POOLPAGE)
return uvm_km_alloc_poolpage(map, waitok);
@ -719,9 +702,7 @@ again:
}
vaddr_t
uvm_km_alloc_poolpage(map, waitok)
struct vm_map *map;
boolean_t waitok;
uvm_km_alloc_poolpage(struct vm_map *map, boolean_t waitok)
{
#if defined(PMAP_MAP_POOLPAGE)
struct vm_page *pg;
@ -763,9 +744,7 @@ uvm_km_alloc_poolpage(map, waitok)
/* ARGSUSED */
void
uvm_km_free_poolpage_cache(map, addr)
struct vm_map *map;
vaddr_t addr;
uvm_km_free_poolpage_cache(struct vm_map *map, vaddr_t addr)
{
#if defined(PMAP_UNMAP_POOLPAGE)
uvm_km_free_poolpage(map, addr);
@ -797,9 +776,7 @@ uvm_km_free_poolpage_cache(map, addr)
/* ARGSUSED */
void
uvm_km_free_poolpage(map, addr)
struct vm_map *map;
vaddr_t addr;
uvm_km_free_poolpage(struct vm_map *map, vaddr_t addr)
{
#if defined(PMAP_UNMAP_POOLPAGE)
paddr_t pa;

View File

@ -1,4 +1,4 @@
/* $NetBSD: uvm_loan.c,v 1.53 2005/05/11 13:02:25 yamt Exp $ */
/* $NetBSD: uvm_loan.c,v 1.54 2005/06/27 02:19:48 thorpej Exp $ */
/*
*
@ -39,7 +39,7 @@
*/
#include <sys/cdefs.h>
__KERNEL_RCSID(0, "$NetBSD: uvm_loan.c,v 1.53 2005/05/11 13:02:25 yamt Exp $");
__KERNEL_RCSID(0, "$NetBSD: uvm_loan.c,v 1.54 2005/06/27 02:19:48 thorpej Exp $");
#include <sys/param.h>
#include <sys/systm.h>
@ -109,7 +109,6 @@ __KERNEL_RCSID(0, "$NetBSD: uvm_loan.c,v 1.53 2005/05/11 13:02:25 yamt Exp $");
static int uvm_loananon(struct uvm_faultinfo *, void ***,
int, struct vm_anon *);
static int uvm_loanentry(struct uvm_faultinfo *, void ***, int);
static int uvm_loanuobj(struct uvm_faultinfo *, void ***,
int, vaddr_t);
static int uvm_loanzero(struct uvm_faultinfo *, void ***, int);
@ -137,10 +136,7 @@ static int uvm_loanpage(struct vm_page **, int);
*/
static __inline int
uvm_loanentry(ufi, output, flags)
struct uvm_faultinfo *ufi;
void ***output;
int flags;
uvm_loanentry(struct uvm_faultinfo *ufi, void ***output, int flags)
{
vaddr_t curaddr = ufi->orig_rvaddr;
vsize_t togo = ufi->size;
@ -243,12 +239,7 @@ uvm_loanentry(ufi, output, flags)
*/
int
uvm_loan(map, start, len, v, flags)
struct vm_map *map;
vaddr_t start;
vsize_t len;
void *v;
int flags;
uvm_loan(struct vm_map *map, vaddr_t start, vsize_t len, void *v, int flags)
{
struct uvm_faultinfo ufi;
void **result, **output;
@ -354,11 +345,8 @@ fail:
*/
int
uvm_loananon(ufi, output, flags, anon)
struct uvm_faultinfo *ufi;
void ***output;
int flags;
struct vm_anon *anon;
uvm_loananon(struct uvm_faultinfo *ufi, void ***output, int flags,
struct vm_anon *anon)
{
struct vm_page *pg;
int error;
@ -465,9 +453,7 @@ uvm_loananon(ufi, output, flags, anon)
* => fail with EBUSY if meet a wired page.
*/
static int
uvm_loanpage(pgpp, npages)
struct vm_page **pgpp;
int npages;
uvm_loanpage(struct vm_page **pgpp, int npages)
{
int i;
int error = 0;
@ -529,11 +515,8 @@ uvm_loanpage(pgpp, npages)
* => fail with EBUSY if we meet a wired page.
*/
int
uvm_loanuobjpages(uobj, pgoff, orignpages, origpgpp)
struct uvm_object *uobj;
voff_t pgoff;
int orignpages;
struct vm_page **origpgpp;
uvm_loanuobjpages(struct uvm_object *uobj, voff_t pgoff, int orignpages,
struct vm_page **origpgpp)
{
int ndone; /* # of pages loaned out */
struct vm_page **pgpp;
@ -639,11 +622,7 @@ fail:
*/
static int
uvm_loanuobj(ufi, output, flags, va)
struct uvm_faultinfo *ufi;
void ***output;
int flags;
vaddr_t va;
uvm_loanuobj(struct uvm_faultinfo *ufi, void ***output, int flags, vaddr_t va)
{
struct vm_amap *amap = ufi->entry->aref.ar_amap;
struct uvm_object *uobj = ufi->entry->object.uvm_obj;
@ -852,10 +831,7 @@ fail:
static struct uvm_object uvm_loanzero_object;
static int
uvm_loanzero(ufi, output, flags)
struct uvm_faultinfo *ufi;
void ***output;
int flags;
uvm_loanzero(struct uvm_faultinfo *ufi, void ***output, int flags)
{
struct vm_anon *anon;
struct vm_page *pg;
@ -957,9 +933,7 @@ again:
*/
static void
uvm_unloananon(aloans, nanons)
struct vm_anon **aloans;
int nanons;
uvm_unloananon(struct vm_anon **aloans, int nanons)
{
struct vm_anon *anon;
@ -984,9 +958,7 @@ uvm_unloananon(aloans, nanons)
*/
static void
uvm_unloanpage(ploans, npages)
struct vm_page **ploans;
int npages;
uvm_unloanpage(struct vm_page **ploans, int npages)
{
struct vm_page *pg;
struct simplelock *slock;

View File

@ -1,4 +1,4 @@
/* $NetBSD: uvm_meter.c,v 1.34 2005/05/15 08:01:06 yamt Exp $ */
/* $NetBSD: uvm_meter.c,v 1.35 2005/06/27 02:19:48 thorpej Exp $ */
/*
* Copyright (c) 1997 Charles D. Cranor and Washington University.
@ -41,7 +41,7 @@
*/
#include <sys/cdefs.h>
__KERNEL_RCSID(0, "$NetBSD: uvm_meter.c,v 1.34 2005/05/15 08:01:06 yamt Exp $");
__KERNEL_RCSID(0, "$NetBSD: uvm_meter.c,v 1.35 2005/06/27 02:19:48 thorpej Exp $");
#include <sys/param.h>
#include <sys/proc.h>
@ -62,7 +62,7 @@ struct loadavg averunnable;
* 5 second intervals.
*/
static fixpt_t cexp[3] = {
static const fixpt_t cexp[3] = {
0.9200444146293232 * FSCALE, /* exp(-1/12) */
0.9834714538216174 * FSCALE, /* exp(-1/60) */
0.9944598480048967 * FSCALE, /* exp(-1/180) */
@ -79,7 +79,7 @@ static void uvm_total(struct vmtotal *);
* uvm_meter: calculate load average and wake up the swapper (if needed)
*/
void
uvm_meter()
uvm_meter(void)
{
if ((time.tv_sec % 5) == 0)
uvm_loadav(&averunnable);
@ -92,8 +92,7 @@ uvm_meter()
* 1, 5, and 15 minute internvals.
*/
static void
uvm_loadav(avg)
struct loadavg *avg;
uvm_loadav(struct loadavg *avg)
{
int i, nrun;
struct lwp *l;
@ -414,8 +413,7 @@ SYSCTL_SETUP(sysctl_vm_setup, "sysctl vm subtree setup")
* uvm_total: calculate the current state of the system.
*/
static void
uvm_total(totalp)
struct vmtotal *totalp;
uvm_total(struct vmtotal *totalp)
{
struct lwp *l;
#if 0

View File

@ -1,4 +1,4 @@
/* $NetBSD: uvm_page.c,v 1.104 2005/06/04 13:48:35 chs Exp $ */
/* $NetBSD: uvm_page.c,v 1.105 2005/06/27 02:19:48 thorpej Exp $ */
/*
* Copyright (c) 1997 Charles D. Cranor and Washington University.
@ -71,7 +71,7 @@
*/
#include <sys/cdefs.h>
__KERNEL_RCSID(0, "$NetBSD: uvm_page.c,v 1.104 2005/06/04 13:48:35 chs Exp $");
__KERNEL_RCSID(0, "$NetBSD: uvm_page.c,v 1.105 2005/06/27 02:19:48 thorpej Exp $");
#include "opt_uvmhist.h"
@ -168,9 +168,7 @@ static void uvm_pageremove(struct vm_page *);
*/
__inline static void
uvm_pageinsert_after(pg, where)
struct vm_page *pg;
struct vm_page *where;
uvm_pageinsert_after(struct vm_page *pg, struct vm_page *where)
{
struct pglist *buck;
struct uvm_object *uobj = pg->uobject;
@ -207,8 +205,7 @@ uvm_pageinsert_after(pg, where)
}
__inline static void
uvm_pageinsert(pg)
struct vm_page *pg;
uvm_pageinsert(struct vm_page *pg)
{
uvm_pageinsert_after(pg, NULL);
@ -222,8 +219,7 @@ uvm_pageinsert(pg)
*/
static __inline void
uvm_pageremove(pg)
struct vm_page *pg;
uvm_pageremove(struct vm_page *pg)
{
struct pglist *buck;
struct uvm_object *uobj = pg->uobject;
@ -275,8 +271,7 @@ uvm_page_init_buckets(struct pgfreelist *pgfl)
*/
void
uvm_page_init(kvm_startp, kvm_endp)
vaddr_t *kvm_startp, *kvm_endp;
uvm_page_init(vaddr_t *kvm_startp, vaddr_t *kvm_endp)
{
vsize_t freepages, pagecount, bucketcount, n;
struct pgflbucket *bucketarray;
@ -459,7 +454,7 @@ uvm_page_init(kvm_startp, kvm_endp)
*/
void
uvm_setpagesize()
uvm_setpagesize(void)
{
/*
@ -484,8 +479,7 @@ uvm_setpagesize()
*/
vaddr_t
uvm_pageboot_alloc(size)
vsize_t size;
uvm_pageboot_alloc(vsize_t size)
{
static boolean_t initialized = FALSE;
vaddr_t addr;
@ -584,9 +578,7 @@ uvm_pageboot_alloc(size)
static boolean_t uvm_page_physget_freelist(paddr_t *, int);
static boolean_t
uvm_page_physget_freelist(paddrp, freelist)
paddr_t *paddrp;
int freelist;
uvm_page_physget_freelist(paddr_t *paddrp, int freelist)
{
int lcv, x;
@ -676,8 +668,7 @@ uvm_page_physget_freelist(paddrp, freelist)
}
boolean_t
uvm_page_physget(paddrp)
paddr_t *paddrp;
uvm_page_physget(paddr_t *paddrp)
{
int i;
@ -699,9 +690,8 @@ uvm_page_physget(paddrp)
*/
void
uvm_page_physload(start, end, avail_start, avail_end, free_list)
paddr_t start, end, avail_start, avail_end;
int free_list;
uvm_page_physload(paddr_t start, paddr_t end, paddr_t avail_start,
paddr_t avail_end, int free_list)
{
int preload, lcv;
psize_t npages;
@ -836,7 +826,7 @@ uvm_page_physload(start, end, avail_start, avail_end, free_list)
*/
void
uvm_page_rehash()
uvm_page_rehash(void)
{
int freepages, lcv, bucketcount, oldcount;
struct pglist *newbuckets, *oldbuckets;
@ -1050,12 +1040,8 @@ uvm_pagealloc_pgfl(struct pgfreelist *pgfl, int try1, int try2,
*/
struct vm_page *
uvm_pagealloc_strat(obj, off, anon, flags, strat, free_list)
struct uvm_object *obj;
voff_t off;
int flags;
struct vm_anon *anon;
int strat, free_list;
uvm_pagealloc_strat(struct uvm_object *obj, voff_t off, struct vm_anon *anon,
int flags, int strat, int free_list)
{
int lcv, try1, try2, s, zeroit = 0, color;
struct vm_page *pg;
@ -1221,9 +1207,7 @@ uvm_pagealloc_strat(obj, off, anon, flags, strat, free_list)
*/
void
uvm_pagereplace(oldpg, newpg)
struct vm_page *oldpg;
struct vm_page *newpg;
uvm_pagereplace(struct vm_page *oldpg, struct vm_page *newpg)
{
KASSERT((oldpg->flags & PG_TABLED) != 0);
@ -1246,10 +1230,7 @@ uvm_pagereplace(oldpg, newpg)
*/
void
uvm_pagerealloc(pg, newobj, newoff)
struct vm_page *pg;
struct uvm_object *newobj;
voff_t newoff;
uvm_pagerealloc(struct vm_page *pg, struct uvm_object *newobj, voff_t newoff)
{
/*
* remove it from the old object
@ -1314,8 +1295,7 @@ uvm_pagezerocheck(struct vm_page *pg)
*/
void
uvm_pagefree(pg)
struct vm_page *pg;
uvm_pagefree(struct vm_page *pg)
{
int s;
struct pglist *pgfl;
@ -1451,9 +1431,7 @@ uvm_pagefree(pg)
*/
void
uvm_page_unbusy(pgs, npgs)
struct vm_page **pgs;
int npgs;
uvm_page_unbusy(struct vm_page **pgs, int npgs)
{
struct vm_page *pg;
int i;
@ -1501,9 +1479,7 @@ uvm_page_unbusy(pgs, npgs)
* => if "tag" is NULL then we are releasing page ownership
*/
void
uvm_page_own(pg, tag)
struct vm_page *pg;
const char *tag;
uvm_page_own(struct vm_page *pg, const char *tag)
{
KASSERT((pg->flags & (PG_PAGEOUT|PG_RELEASED)) == 0);
@ -1545,7 +1521,7 @@ uvm_page_own(pg, tag)
* there is a process ready to run.
*/
void
uvm_pageidlezero()
uvm_pageidlezero(void)
{
struct vm_page *pg;
struct pgfreelist *pgfl;

View File

@ -1,4 +1,4 @@
/* $NetBSD: uvm_page_i.h,v 1.22 2004/05/12 20:09:52 yamt Exp $ */
/* $NetBSD: uvm_page_i.h,v 1.23 2005/06/27 02:19:48 thorpej Exp $ */
/*
* Copyright (c) 1997 Charles D. Cranor and Washington University.
@ -88,7 +88,7 @@
*/
PAGE_INLINE int
uvm_lock_fpageq()
uvm_lock_fpageq(void)
{
int s;
@ -105,8 +105,7 @@ uvm_lock_fpageq()
*/
PAGE_INLINE void
uvm_unlock_fpageq(s)
int s;
uvm_unlock_fpageq(int s)
{
simple_unlock(&uvm.fpageqlock);
@ -121,9 +120,7 @@ uvm_unlock_fpageq(s)
*/
struct vm_page *
uvm_pagelookup(obj, off)
struct uvm_object *obj;
voff_t off;
uvm_pagelookup(struct uvm_object *obj, voff_t off)
{
struct vm_page *pg;
struct pglist *buck;
@ -149,8 +146,7 @@ uvm_pagelookup(obj, off)
*/
PAGE_INLINE void
uvm_pagewire(pg)
struct vm_page *pg;
uvm_pagewire(struct vm_page *pg)
{
UVM_LOCK_ASSERT_PAGEQ();
if (pg->wire_count == 0) {
@ -168,8 +164,7 @@ uvm_pagewire(pg)
*/
PAGE_INLINE void
uvm_pageunwire(pg)
struct vm_page *pg;
uvm_pageunwire(struct vm_page *pg)
{
UVM_LOCK_ASSERT_PAGEQ();
pg->wire_count--;
@ -191,8 +186,7 @@ uvm_pageunwire(pg)
*/
PAGE_INLINE void
uvm_pagedeactivate(pg)
struct vm_page *pg;
uvm_pagedeactivate(struct vm_page *pg)
{
UVM_LOCK_ASSERT_PAGEQ();
if (pg->pqflags & PQ_ACTIVE) {
@ -215,8 +209,7 @@ uvm_pagedeactivate(pg)
*/
PAGE_INLINE void
uvm_pageactivate(pg)
struct vm_page *pg;
uvm_pageactivate(struct vm_page *pg)
{
UVM_LOCK_ASSERT_PAGEQ();
uvm_pagedequeue(pg);
@ -232,8 +225,7 @@ uvm_pageactivate(pg)
*/
PAGE_INLINE void
uvm_pagedequeue(pg)
struct vm_page *pg;
uvm_pagedequeue(struct vm_page *pg)
{
if (pg->pqflags & PQ_ACTIVE) {
UVM_LOCK_ASSERT_PAGEQ();
@ -256,8 +248,7 @@ uvm_pagedequeue(pg)
*/
PAGE_INLINE void
uvm_pagezero(pg)
struct vm_page *pg;
uvm_pagezero(struct vm_page *pg)
{
pg->flags &= ~PG_CLEAN;
pmap_zero_page(VM_PAGE_TO_PHYS(pg));
@ -271,8 +262,7 @@ uvm_pagezero(pg)
*/
PAGE_INLINE void
uvm_pagecopy(src, dst)
struct vm_page *src, *dst;
uvm_pagecopy(struct vm_page *src, struct vm_page *dst)
{
dst->flags &= ~PG_CLEAN;
@ -284,8 +274,7 @@ uvm_pagecopy(src, dst)
*/
PAGE_INLINE int
uvm_page_lookup_freelist(pg)
struct vm_page *pg;
uvm_page_lookup_freelist(struct vm_page *pg)
{
int lcv;

View File

@ -1,4 +1,4 @@
/* $NetBSD: uvm_pager.c,v 1.66 2005/04/01 11:59:39 yamt Exp $ */
/* $NetBSD: uvm_pager.c,v 1.67 2005/06/27 02:19:48 thorpej Exp $ */
/*
*
@ -39,7 +39,7 @@
*/
#include <sys/cdefs.h>
__KERNEL_RCSID(0, "$NetBSD: uvm_pager.c,v 1.66 2005/04/01 11:59:39 yamt Exp $");
__KERNEL_RCSID(0, "$NetBSD: uvm_pager.c,v 1.67 2005/06/27 02:19:48 thorpej Exp $");
#include "opt_uvmhist.h"
@ -81,7 +81,7 @@ static boolean_t emerginuse;
*/
void
uvm_pager_init()
uvm_pager_init(void)
{
u_int lcv;
vaddr_t sva, eva;
@ -127,10 +127,7 @@ uvm_pager_init()
*/
vaddr_t
uvm_pagermapin(pps, npages, flags)
struct vm_page **pps;
int npages;
int flags;
uvm_pagermapin(struct vm_page **pps, int npages, int flags)
{
vsize_t size;
vaddr_t kva;
@ -205,9 +202,7 @@ enter:
*/
void
uvm_pagermapout(kva, npages)
vaddr_t kva;
int npages;
uvm_pagermapout(vaddr_t kva, int npages)
{
vsize_t size = npages << PAGE_SHIFT;
struct vm_map_entry *entries;
@ -250,8 +245,7 @@ uvm_pagermapout(kva, npages)
*/
void
uvm_aio_biodone1(bp)
struct buf *bp;
uvm_aio_biodone1(struct buf *bp)
{
struct buf *mbp = bp->b_private;
@ -275,8 +269,7 @@ uvm_aio_biodone1(bp)
*/
void
uvm_aio_biodone(bp)
struct buf *bp;
uvm_aio_biodone(struct buf *bp)
{
/* reset b_iodone for when this is a single-buf i/o. */
bp->b_iodone = uvm_aio_aiodone;
@ -293,8 +286,7 @@ uvm_aio_biodone(bp)
*/
void
uvm_aio_aiodone(bp)
struct buf *bp;
uvm_aio_aiodone(struct buf *bp)
{
int npages = bp->b_bufsize >> PAGE_SHIFT;
struct vm_page *pg, *pgs[npages];

View File

@ -1,4 +1,4 @@
/* $NetBSD: uvm_pager_i.h,v 1.12 2002/12/01 22:58:43 matt Exp $ */
/* $NetBSD: uvm_pager_i.h,v 1.13 2005/06/27 02:19:48 thorpej Exp $ */
/*
*
@ -53,8 +53,7 @@
*/
PAGER_INLINE struct vm_page *
uvm_pageratop(kva)
vaddr_t kva;
uvm_pageratop(vaddr_t kva)
{
struct vm_page *pg;
paddr_t pa;

View File

@ -1,4 +1,4 @@
/* $NetBSD: uvm_pdaemon.c,v 1.64 2005/05/11 13:02:26 yamt Exp $ */
/* $NetBSD: uvm_pdaemon.c,v 1.65 2005/06/27 02:19:48 thorpej Exp $ */
/*
* Copyright (c) 1997 Charles D. Cranor and Washington University.
@ -71,7 +71,7 @@
*/
#include <sys/cdefs.h>
__KERNEL_RCSID(0, "$NetBSD: uvm_pdaemon.c,v 1.64 2005/05/11 13:02:26 yamt Exp $");
__KERNEL_RCSID(0, "$NetBSD: uvm_pdaemon.c,v 1.65 2005/06/27 02:19:48 thorpej Exp $");
#include "opt_uvmhist.h"
@ -99,9 +99,9 @@ __KERNEL_RCSID(0, "$NetBSD: uvm_pdaemon.c,v 1.64 2005/05/11 13:02:26 yamt Exp $"
* local prototypes
*/
void uvmpd_scan(void);
void uvmpd_scan_inactive(struct pglist *);
void uvmpd_tune(void);
static void uvmpd_scan(void);
static void uvmpd_scan_inactive(struct pglist *);
static void uvmpd_tune(void);
/*
* XXX hack to avoid hangs when large processes fork.
@ -116,8 +116,7 @@ int uvm_extrapages;
*/
void
uvm_wait(wmsg)
const char *wmsg;
uvm_wait(const char *wmsg)
{
int timo = 0;
int s = splbio();
@ -168,7 +167,7 @@ uvm_wait(wmsg)
* => caller must call with page queues locked
*/
void
static void
uvmpd_tune(void)
{
UVMHIST_FUNC("uvmpd_tune"); UVMHIST_CALLED(pdhist);
@ -380,9 +379,8 @@ uvm_aiodone_daemon(void *arg)
* => we return TRUE if we are exiting because we met our target
*/
void
uvmpd_scan_inactive(pglst)
struct pglist *pglst;
static void
uvmpd_scan_inactive(struct pglist *pglst)
{
int error;
struct vm_page *p, *nextpg = NULL; /* Quell compiler warning */
@ -764,7 +762,7 @@ uvmpd_scan_inactive(pglst)
* => called with pageq's locked
*/
void
static void
uvmpd_scan(void)
{
int inactive_shortage, swap_shortage, pages_freed;

View File

@ -1,4 +1,4 @@
/* $NetBSD: uvm_pglist.c,v 1.32 2004/09/17 20:46:03 yamt Exp $ */
/* $NetBSD: uvm_pglist.c,v 1.33 2005/06/27 02:19:48 thorpej Exp $ */
/*-
* Copyright (c) 1997 The NetBSD Foundation, Inc.
@ -42,7 +42,7 @@
*/
#include <sys/cdefs.h>
__KERNEL_RCSID(0, "$NetBSD: uvm_pglist.c,v 1.32 2004/09/17 20:46:03 yamt Exp $");
__KERNEL_RCSID(0, "$NetBSD: uvm_pglist.c,v 1.33 2005/06/27 02:19:48 thorpej Exp $");
#include <sys/param.h>
#include <sys/systm.h>
@ -84,20 +84,8 @@ u_long uvm_pglistalloc_npages;
* power-of-two boundary (relative to zero).
*/
static void uvm_pglist_add(struct vm_page *, struct pglist *);
static int uvm_pglistalloc_c_ps(struct vm_physseg *, int, paddr_t, paddr_t,
paddr_t, paddr_t, struct pglist *);
static int uvm_pglistalloc_contig(int, paddr_t, paddr_t, paddr_t, paddr_t,
struct pglist *);
static int uvm_pglistalloc_s_ps(struct vm_physseg *, int, paddr_t, paddr_t,
struct pglist *);
static int uvm_pglistalloc_simple(int, paddr_t, paddr_t,
struct pglist *, int);
static void
uvm_pglist_add(pg, rlist)
struct vm_page *pg;
struct pglist *rlist;
uvm_pglist_add(struct vm_page *pg, struct pglist *rlist)
{
int free_list, color, pgflidx;
#ifdef DEBUG
@ -136,11 +124,8 @@ uvm_pglist_add(pg, rlist)
}
static int
uvm_pglistalloc_c_ps(ps, num, low, high, alignment, boundary, rlist)
struct vm_physseg *ps;
int num;
paddr_t low, high, alignment, boundary;
struct pglist *rlist;
uvm_pglistalloc_c_ps(struct vm_physseg *ps, int num, paddr_t low, paddr_t high,
paddr_t alignment, paddr_t boundary, struct pglist *rlist)
{
int try, limit, tryidx, end, idx;
struct vm_page *pgs;
@ -241,10 +226,8 @@ uvm_pglistalloc_c_ps(ps, num, low, high, alignment, boundary, rlist)
}
static int
uvm_pglistalloc_contig(num, low, high, alignment, boundary, rlist)
int num;
paddr_t low, high, alignment, boundary;
struct pglist *rlist;
uvm_pglistalloc_contig(int num, paddr_t low, paddr_t high, paddr_t alignment,
paddr_t boundary, struct pglist *rlist)
{
int fl, psi;
struct vm_physseg *ps;
@ -300,11 +283,8 @@ out:
}
static int
uvm_pglistalloc_s_ps(ps, num, low, high, rlist)
struct vm_physseg *ps;
int num;
paddr_t low, high;
struct pglist *rlist;
uvm_pglistalloc_s_ps(struct vm_physseg *ps, int num, paddr_t low, paddr_t high,
struct pglist *rlist)
{
int todo, limit, try;
struct vm_page *pg;
@ -343,11 +323,8 @@ uvm_pglistalloc_s_ps(ps, num, low, high, rlist)
}
static int
uvm_pglistalloc_simple(num, low, high, rlist, waitok)
int num;
paddr_t low, high;
struct pglist *rlist;
int waitok;
uvm_pglistalloc_simple(int num, paddr_t low, paddr_t high,
struct pglist *rlist, int waitok)
{
int fl, psi, s, error;
struct vm_physseg *ps;
@ -415,11 +392,8 @@ out:
}
int
uvm_pglistalloc(size, low, high, alignment, boundary, rlist, nsegs, waitok)
psize_t size;
paddr_t low, high, alignment, boundary;
struct pglist *rlist;
int nsegs, waitok;
uvm_pglistalloc(psize_t size, paddr_t low, paddr_t high, paddr_t alignment,
paddr_t boundary, struct pglist *rlist, int nsegs, int waitok)
{
int num, res;
@ -456,8 +430,7 @@ uvm_pglistalloc(size, low, high, alignment, boundary, rlist, nsegs, waitok)
*/
void
uvm_pglistfree(list)
struct pglist *list;
uvm_pglistfree(struct pglist *list)
{
struct vm_page *pg;
int s;

View File

@ -1,4 +1,4 @@
/* $NetBSD: uvm_stat.c,v 1.27 2005/05/15 08:01:06 yamt Exp $ */
/* $NetBSD: uvm_stat.c,v 1.28 2005/06/27 02:19:48 thorpej Exp $ */
/*
*
@ -39,7 +39,7 @@
*/
#include <sys/cdefs.h>
__KERNEL_RCSID(0, "$NetBSD: uvm_stat.c,v 1.27 2005/05/15 08:01:06 yamt Exp $");
__KERNEL_RCSID(0, "$NetBSD: uvm_stat.c,v 1.28 2005/06/27 02:19:48 thorpej Exp $");
#include "opt_uvmhist.h"
#include "opt_ddb.h"
@ -79,8 +79,7 @@ void uvmcnt_dump(void);
#ifdef UVMHIST
/* call this from ddb */
void
uvmhist_dump(l)
struct uvm_history *l;
uvmhist_dump(struct uvm_history *l)
{
int lcv, s;
@ -98,8 +97,7 @@ uvmhist_dump(l)
* print a merged list of uvm_history structures
*/
static void
uvmhist_dump_histories(hists)
struct uvm_history *hists[];
uvmhist_dump_histories(struct uvm_history *hists[])
{
struct timeval tv;
int cur[MAXHISTS];
@ -169,8 +167,7 @@ restart:
* merges the named histories.
*/
void
uvm_hist(bitmask)
u_int32_t bitmask; /* XXX only support 32 hists */
uvm_hist(u_int32_t bitmask) /* XXX only support 32 hists */
{
struct uvm_history *hists[MAXHISTS + 1];
int i = 0;

View File

@ -1,4 +1,4 @@
/* $NetBSD: uvm_swap.c,v 1.92 2005/05/29 21:06:33 christos Exp $ */
/* $NetBSD: uvm_swap.c,v 1.93 2005/06/27 02:19:48 thorpej Exp $ */
/*
* Copyright (c) 1995, 1996, 1997 Matthew R. Green
@ -32,7 +32,7 @@
*/
#include <sys/cdefs.h>
__KERNEL_RCSID(0, "$NetBSD: uvm_swap.c,v 1.92 2005/05/29 21:06:33 christos Exp $");
__KERNEL_RCSID(0, "$NetBSD: uvm_swap.c,v 1.93 2005/06/27 02:19:48 thorpej Exp $");
#include "fs_nfs.h"
#include "opt_uvmhist.h"
@ -257,7 +257,7 @@ const struct cdevsw swap_cdevsw = {
* are brought up (which happens after uvm_init())
*/
void
uvm_swap_init()
uvm_swap_init(void)
{
UVMHIST_FUNC("uvm_swap_init");
@ -307,10 +307,7 @@ uvm_swap_init()
* here while adding swap)
*/
static void
swaplist_insert(sdp, newspp, priority)
struct swapdev *sdp;
struct swappri *newspp;
int priority;
swaplist_insert(struct swapdev *sdp, struct swappri *newspp, int priority)
{
struct swappri *spp, *pspp;
UVMHIST_FUNC("swaplist_insert"); UVMHIST_CALLED(pdhist);
@ -362,9 +359,7 @@ swaplist_insert(sdp, newspp, priority)
* => we return the swapdev we found (and removed)
*/
static struct swapdev *
swaplist_find(vp, remove)
struct vnode *vp;
boolean_t remove;
swaplist_find(struct vnode *vp, boolean_t remove)
{
struct swapdev *sdp;
struct swappri *spp;
@ -396,7 +391,7 @@ swaplist_find(vp, remove)
* => caller must hold both swap_syscall_lock and uvm.swap_data_lock
*/
static void
swaplist_trim()
swaplist_trim(void)
{
struct swappri *spp, *nextspp;
@ -418,8 +413,7 @@ swaplist_trim()
* => caller must hold uvm.swap_data_lock
*/
static struct swapdev *
swapdrum_getsdp(pgno)
int pgno;
swapdrum_getsdp(int pgno)
{
struct swapdev *sdp;
struct swappri *spp;
@ -443,10 +437,7 @@ swapdrum_getsdp(pgno)
* [with two helper functions: swap_on and swap_off]
*/
int
sys_swapctl(l, v, retval)
struct lwp *l;
void *v;
register_t *retval;
sys_swapctl(struct lwp *l, void *v, register_t *retval)
{
struct sys_swapctl_args /* {
syscallarg(int) cmd;
@ -702,11 +693,7 @@ out:
* ensure it would fit in the stackgap in any case.
*/
void
uvm_swap_stats(cmd, sep, sec, retval)
int cmd;
struct swapent *sep;
int sec;
register_t *retval;
uvm_swap_stats(int cmd, struct swapent *sep, int sec, register_t *retval)
{
struct swappri *spp;
struct swapdev *sdp;
@ -763,9 +750,7 @@ uvm_swap_stats(cmd, sep, sec, retval)
* if needed.
*/
static int
swap_on(p, sdp)
struct proc *p;
struct swapdev *sdp;
swap_on(struct proc *p, struct swapdev *sdp)
{
struct vnode *vp;
int error, npages, nblocks, size;
@ -977,9 +962,7 @@ bad:
* => swap data should be locked, we will unlock.
*/
static int
swap_off(p, sdp)
struct proc *p;
struct swapdev *sdp;
swap_off(struct proc *p, struct swapdev *sdp)
{
int npages = sdp->swd_npages;
int error = 0;
@ -1056,10 +1039,7 @@ swap_off(p, sdp)
*/
/*ARGSUSED*/
int
swread(dev, uio, ioflag)
dev_t dev;
struct uio *uio;
int ioflag;
swread(dev_t dev, struct uio *uio, int ioflag)
{
UVMHIST_FUNC("swread"); UVMHIST_CALLED(pdhist);
@ -1072,10 +1052,7 @@ swread(dev, uio, ioflag)
*/
/*ARGSUSED*/
int
swwrite(dev, uio, ioflag)
dev_t dev;
struct uio *uio;
int ioflag;
swwrite(dev_t dev, struct uio *uio, int ioflag)
{
UVMHIST_FUNC("swwrite"); UVMHIST_CALLED(pdhist);
@ -1089,8 +1066,7 @@ swwrite(dev, uio, ioflag)
* => we must map the i/o request from the drum to the correct swapdev.
*/
void
swstrategy(bp)
struct buf *bp;
swstrategy(struct buf *bp)
{
struct swapdev *sdp;
struct vnode *vp;
@ -1177,10 +1153,7 @@ swstrategy(bp)
* sw_reg_strategy: handle swap i/o to regular files
*/
static void
sw_reg_strategy(sdp, bp, bn)
struct swapdev *sdp;
struct buf *bp;
int bn;
sw_reg_strategy(struct swapdev *sdp, struct buf *bp, int bn)
{
struct vnode *vp;
struct vndxfer *vnx;
@ -1330,8 +1303,7 @@ out: /* Arrive here at splbio */
* => reqs are sorted by b_rawblkno (above)
*/
static void
sw_reg_start(sdp)
struct swapdev *sdp;
sw_reg_start(struct swapdev *sdp)
{
struct buf *bp;
UVMHIST_FUNC("sw_reg_start"); UVMHIST_CALLED(pdhist);
@ -1365,8 +1337,7 @@ sw_reg_start(sdp)
* => note that we can recover the vndbuf struct by casting the buf ptr
*/
static void
sw_reg_iodone(bp)
struct buf *bp;
sw_reg_iodone(struct buf *bp)
{
struct vndbuf *vbp = (struct vndbuf *) bp;
struct vndxfer *vnx = vbp->vb_xfer;
@ -1443,9 +1414,7 @@ sw_reg_iodone(bp)
* => XXXMRG: "LESSOK" INTERFACE NEEDED TO EXTENT SYSTEM
*/
int
uvm_swap_alloc(nslots, lessok)
int *nslots; /* IN/OUT */
boolean_t lessok;
uvm_swap_alloc(int *nslots /* IN/OUT */, boolean_t lessok)
{
struct swapdev *sdp;
struct swappri *spp;
@ -1525,9 +1494,7 @@ uvm_swapisfull(void)
* => we lock uvm.swap_data_lock
*/
void
uvm_swap_markbad(startslot, nslots)
int startslot;
int nslots;
uvm_swap_markbad(int startslot, int nslots)
{
struct swapdev *sdp;
UVMHIST_FUNC("uvm_swap_markbad"); UVMHIST_CALLED(pdhist);
@ -1557,9 +1524,7 @@ uvm_swap_markbad(startslot, nslots)
* => we lock uvm.swap_data_lock
*/
void
uvm_swap_free(startslot, nslots)
int startslot;
int nslots;
uvm_swap_free(int startslot, int nslots)
{
struct swapdev *sdp;
UVMHIST_FUNC("uvm_swap_free"); UVMHIST_CALLED(pdhist);
@ -1599,11 +1564,7 @@ uvm_swap_free(startslot, nslots)
*/
int
uvm_swap_put(swslot, ppsp, npages, flags)
int swslot;
struct vm_page **ppsp;
int npages;
int flags;
uvm_swap_put(int swslot, struct vm_page **ppsp, int npages, int flags)
{
int error;
@ -1619,9 +1580,7 @@ uvm_swap_put(swslot, ppsp, npages, flags)
*/
int
uvm_swap_get(page, swslot, flags)
struct vm_page *page;
int swslot, flags;
uvm_swap_get(struct vm_page *page, int swslot, int flags)
{
int error;
@ -1652,9 +1611,7 @@ uvm_swap_get(page, swslot, flags)
*/
static int
uvm_swap_io(pps, startslot, npages, flags)
struct vm_page **pps;
int startslot, npages, flags;
uvm_swap_io(struct vm_page **pps, int startslot, int npages, int flags)
{
daddr_t startblk;
struct buf *bp;

View File

@ -1,4 +1,4 @@
/* $NetBSD: uvm_unix.c,v 1.30 2004/08/28 12:44:22 jdolecek Exp $ */
/* $NetBSD: uvm_unix.c,v 1.31 2005/06/27 02:19:48 thorpej Exp $ */
/*
* Copyright (c) 1997 Charles D. Cranor and Washington University.
@ -50,7 +50,7 @@
*/
#include <sys/cdefs.h>
__KERNEL_RCSID(0, "$NetBSD: uvm_unix.c,v 1.30 2004/08/28 12:44:22 jdolecek Exp $");
__KERNEL_RCSID(0, "$NetBSD: uvm_unix.c,v 1.31 2005/06/27 02:19:48 thorpej Exp $");
#include <sys/param.h>
#include <sys/systm.h>
@ -69,10 +69,7 @@ __KERNEL_RCSID(0, "$NetBSD: uvm_unix.c,v 1.30 2004/08/28 12:44:22 jdolecek Exp $
*/
int
sys_obreak(l, v, retval)
struct lwp *l;
void *v;
register_t *retval;
sys_obreak(struct lwp *l, void *v, register_t *retval)
{
struct sys_obreak_args /* {
syscallarg(char *) nsize;
@ -121,9 +118,7 @@ sys_obreak(l, v, retval)
*/
int
uvm_grow(p, sp)
struct proc *p;
vaddr_t sp;
uvm_grow(struct proc *p, vaddr_t sp)
{
struct vmspace *vm = p->p_vmspace;
vsize_t nss;
@ -156,10 +151,7 @@ uvm_grow(p, sp)
/* ARGSUSED */
int
sys_ovadvise(l, v, retval)
struct lwp *l;
void *v;
register_t *retval;
sys_ovadvise(struct lwp *l, void *v, register_t *retval)
{
#if 0
struct sys_ovadvise_args /* {

View File

@ -1,4 +1,4 @@
/* $NetBSD: uvm_user.c,v 1.11 2001/11/10 07:37:01 lukem Exp $ */
/* $NetBSD: uvm_user.c,v 1.12 2005/06/27 02:19:48 thorpej Exp $ */
/*
*
@ -39,7 +39,7 @@
*/
#include <sys/cdefs.h>
__KERNEL_RCSID(0, "$NetBSD: uvm_user.c,v 1.11 2001/11/10 07:37:01 lukem Exp $");
__KERNEL_RCSID(0, "$NetBSD: uvm_user.c,v 1.12 2005/06/27 02:19:48 thorpej Exp $");
#include <sys/param.h>
#include <sys/systm.h>
@ -52,10 +52,7 @@ __KERNEL_RCSID(0, "$NetBSD: uvm_user.c,v 1.11 2001/11/10 07:37:01 lukem Exp $");
*/
void
uvm_deallocate(map, start, size)
struct vm_map *map;
vaddr_t start;
vsize_t size;
uvm_deallocate(struct vm_map *map, vaddr_t start, vsize_t size)
{
if (size == 0)

View File

@ -1,4 +1,4 @@
/* $NetBSD: uvm_vnode.c,v 1.64 2005/01/09 16:42:44 chs Exp $ */
/* $NetBSD: uvm_vnode.c,v 1.65 2005/06/27 02:19:48 thorpej Exp $ */
/*
* Copyright (c) 1997 Charles D. Cranor and Washington University.
@ -50,7 +50,7 @@
*/
#include <sys/cdefs.h>
__KERNEL_RCSID(0, "$NetBSD: uvm_vnode.c,v 1.64 2005/01/09 16:42:44 chs Exp $");
__KERNEL_RCSID(0, "$NetBSD: uvm_vnode.c,v 1.65 2005/06/27 02:19:48 thorpej Exp $");
#include "fs_nfs.h"
#include "opt_uvmhist.h"
@ -118,9 +118,7 @@ struct uvm_pagerops uvm_vnodeops = {
*/
struct uvm_object *
uvn_attach(arg, accessprot)
void *arg;
vm_prot_t accessprot;
uvn_attach(void *arg, vm_prot_t accessprot)
{
struct vnode *vp = arg;
struct uvm_object *uobj = &vp->v_uobj;
@ -235,8 +233,7 @@ uvn_attach(arg, accessprot)
*/
void
uvn_reference(uobj)
struct uvm_object *uobj;
uvn_reference(struct uvm_object *uobj)
{
VREF((struct vnode *)uobj);
}
@ -251,8 +248,7 @@ uvn_reference(uobj)
*/
void
uvn_detach(uobj)
struct uvm_object *uobj;
uvn_detach(struct uvm_object *uobj)
{
vrele((struct vnode *)uobj);
}
@ -266,11 +262,7 @@ uvn_detach(uobj)
*/
int
uvn_put(uobj, offlo, offhi, flags)
struct uvm_object *uobj;
voff_t offlo;
voff_t offhi;
int flags;
uvn_put(struct uvm_object *uobj, voff_t offlo, voff_t offhi, int flags)
{
struct vnode *vp = (struct vnode *)uobj;
int error;
@ -294,14 +286,10 @@ uvn_put(uobj, offlo, offhi, flags)
*/
int
uvn_get(uobj, offset, pps, npagesp, centeridx, access_type, advice, flags)
struct uvm_object *uobj;
voff_t offset;
struct vm_page **pps; /* IN/OUT */
int *npagesp; /* IN (OUT if PGO_LOCKED) */
int centeridx;
vm_prot_t access_type;
int advice, flags;
uvn_get(struct uvm_object *uobj, voff_t offset,
struct vm_page **pps /* IN/OUT */,
int *npagesp /* IN (OUT if PGO_LOCKED)*/,
int centeridx, vm_prot_t access_type, int advice, int flags)
{
struct vnode *vp = (struct vnode *)uobj;
int error;
@ -322,12 +310,8 @@ uvn_get(uobj, offset, pps, npagesp, centeridx, access_type, advice, flags)
*/
int
uvn_findpages(uobj, offset, npagesp, pgs, flags)
struct uvm_object *uobj;
voff_t offset;
int *npagesp;
struct vm_page **pgs;
int flags;
uvn_findpages(struct uvm_object *uobj, voff_t offset, int *npagesp,
struct vm_page **pgs, int flags)
{
int i, count, found, npages, rv;
@ -359,11 +343,8 @@ uvn_findpages(uobj, offset, npagesp, pgs, flags)
}
int
uvn_findpage(uobj, offset, pgp, flags)
struct uvm_object *uobj;
voff_t offset;
struct vm_page **pgp;
int flags;
uvn_findpage(struct uvm_object *uobj, voff_t offset, struct vm_page **pgp,
int flags)
{
struct vm_page *pg;
boolean_t dirty;
@ -455,9 +436,7 @@ uvn_findpage(uobj, offset, pgp, flags)
*/
void
uvm_vnp_setsize(vp, newsize)
struct vnode *vp;
voff_t newsize;
uvm_vnp_setsize(struct vnode *vp, voff_t newsize)
{
struct uvm_object *uobj = &vp->v_uobj;
voff_t pgend = round_page(newsize);
@ -485,10 +464,7 @@ uvm_vnp_setsize(vp, newsize)
*/
void
uvm_vnp_zerorange(vp, off, len)
struct vnode *vp;
off_t off;
size_t len;
uvm_vnp_zerorange(struct vnode *vp, off_t off, size_t len)
{
void *win;
int flags;