NetBSD/sys/uvm/uvm_page_i.h

323 lines
8.1 KiB
C

/* $NetBSD: uvm_page_i.h,v 1.10 1999/05/24 19:10:57 thorpej Exp $ */
/*
* Copyright (c) 1997 Charles D. Cranor and Washington University.
* Copyright (c) 1991, 1993, The Regents of the University of California.
*
* All rights reserved.
*
* This code is derived from software contributed to Berkeley by
* The Mach Operating System project at Carnegie-Mellon University.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
* 1. Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
* 2. Redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in the
* documentation and/or other materials provided with the distribution.
* 3. All advertising materials mentioning features or use of this software
* must display the following acknowledgement:
* This product includes software developed by Charles D. Cranor,
* Washington University, the University of California, Berkeley and
* its contributors.
* 4. Neither the name of the University nor the names of its contributors
* may be used to endorse or promote products derived from this software
* without specific prior written permission.
*
* THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
* ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
* SUCH DAMAGE.
*
* @(#)vm_page.c 8.3 (Berkeley) 3/21/94
* from: Id: uvm_page_i.h,v 1.1.2.7 1998/01/05 00:26:02 chuck Exp
*
*
* Copyright (c) 1987, 1990 Carnegie-Mellon University.
* All rights reserved.
*
* Permission to use, copy, modify and distribute this software and
* its documentation is hereby granted, provided that both the copyright
* notice and this permission notice appear in all copies of the
* software, derivative works or modified versions, and any portions
* thereof, and that both notices appear in supporting documentation.
*
* CARNEGIE MELLON ALLOWS FREE USE OF THIS SOFTWARE IN ITS "AS IS"
* CONDITION. CARNEGIE MELLON DISCLAIMS ANY LIABILITY OF ANY KIND
* FOR ANY DAMAGES WHATSOEVER RESULTING FROM THE USE OF THIS SOFTWARE.
*
* Carnegie Mellon requests users of this software to return to
*
* Software Distribution Coordinator or Software.Distribution@CS.CMU.EDU
* School of Computer Science
* Carnegie Mellon University
* Pittsburgh PA 15213-3890
*
* any improvements or extensions that they make and grant Carnegie the
* rights to redistribute these changes.
*/
#ifndef _UVM_UVM_PAGE_I_H_
#define _UVM_UVM_PAGE_I_H_
/*
* uvm_page_i.h
*/
/*
* inline functions [maybe]
*/
#if defined(UVM_PAGE_INLINE) || defined(UVM_PAGE)
/*
* uvm_lock_fpageq: lock the free page queue
*
* => free page queue can be accessed in interrupt context, so this
* blocks all interrupts that can cause memory allocation, and
* returns the previous interrupt level.
*/
PAGE_INLINE int
uvm_lock_fpageq()
{
int s;
s = splimp();
simple_lock(&uvm.fpageqlock);
return (s);
}
/*
* uvm_unlock_fpageq: unlock the free page queue
*
* => caller must supply interrupt level returned by uvm_lock_fpageq()
* so that it may be restored.
*/
PAGE_INLINE void
uvm_unlock_fpageq(s)
int s;
{
simple_unlock(&uvm.fpageqlock);
splx(s);
}
/*
* uvm_pagelookup: look up a page
*
* => caller should lock object to keep someone from pulling the page
* out from under it
*/
struct vm_page *
uvm_pagelookup(obj, off)
struct uvm_object *obj;
vaddr_t off;
{
struct vm_page *pg;
struct pglist *buck;
int s;
buck = &uvm.page_hash[uvm_pagehash(obj,off)];
s = splimp();
simple_lock(&uvm.hashlock);
for (pg = buck->tqh_first ; pg != NULL ; pg = pg->hashq.tqe_next) {
if (pg->uobject == obj && pg->offset == off) {
simple_unlock(&uvm.hashlock);
splx(s);
return(pg);
}
}
simple_unlock(&uvm.hashlock);
splx(s);
return(NULL);
}
/*
* uvm_pagewire: wire the page, thus removing it from the daemon's grasp
*
* => caller must lock page queues
*/
PAGE_INLINE void
uvm_pagewire(pg)
struct vm_page *pg;
{
if (pg->wire_count == 0) {
if (pg->pqflags & PQ_ACTIVE) {
TAILQ_REMOVE(&uvm.page_active, pg, pageq);
pg->pqflags &= ~PQ_ACTIVE;
uvmexp.active--;
}
if (pg->pqflags & PQ_INACTIVE) {
if (pg->pqflags & PQ_SWAPBACKED)
TAILQ_REMOVE(&uvm.page_inactive_swp, pg, pageq);
else
TAILQ_REMOVE(&uvm.page_inactive_obj, pg, pageq);
pg->pqflags &= ~PQ_INACTIVE;
uvmexp.inactive--;
}
uvmexp.wired++;
}
pg->wire_count++;
}
/*
* uvm_pageunwire: unwire the page.
*
* => activate if wire count goes to zero.
* => caller must lock page queues
*/
PAGE_INLINE void
uvm_pageunwire(pg)
struct vm_page *pg;
{
pg->wire_count--;
if (pg->wire_count == 0) {
TAILQ_INSERT_TAIL(&uvm.page_active, pg, pageq);
uvmexp.active++;
pg->pqflags |= PQ_ACTIVE;
uvmexp.wired--;
}
}
/*
* uvm_pagedeactivate: deactivate page -- no pmaps have access to page
*
* => caller must lock page queues
* => caller must check to make sure page is not wired
* => object that page belongs to must be locked (so we can adjust pg->flags)
*/
PAGE_INLINE void
uvm_pagedeactivate(pg)
struct vm_page *pg;
{
if (pg->pqflags & PQ_ACTIVE) {
TAILQ_REMOVE(&uvm.page_active, pg, pageq);
pg->pqflags &= ~PQ_ACTIVE;
uvmexp.active--;
}
if ((pg->pqflags & PQ_INACTIVE) == 0) {
#ifdef DIAGNOSTIC
if (pg->wire_count)
panic("uvm_pagedeactivate: caller did not check "
"wire count");
#endif
if (pg->pqflags & PQ_SWAPBACKED)
TAILQ_INSERT_TAIL(&uvm.page_inactive_swp, pg, pageq);
else
TAILQ_INSERT_TAIL(&uvm.page_inactive_obj, pg, pageq);
pg->pqflags |= PQ_INACTIVE;
uvmexp.inactive++;
pmap_clear_reference(PMAP_PGARG(pg));
if (pmap_is_modified(PMAP_PGARG(pg)))
pg->flags &= ~PG_CLEAN;
}
}
/*
* uvm_pageactivate: activate page
*
* => caller must lock page queues
*/
PAGE_INLINE void
uvm_pageactivate(pg)
struct vm_page *pg;
{
if (pg->pqflags & PQ_INACTIVE) {
if (pg->pqflags & PQ_SWAPBACKED)
TAILQ_REMOVE(&uvm.page_inactive_swp, pg, pageq);
else
TAILQ_REMOVE(&uvm.page_inactive_obj, pg, pageq);
pg->pqflags &= ~PQ_INACTIVE;
uvmexp.inactive--;
}
if (pg->wire_count == 0) {
/*
* if page is already active, remove it from list so we
* can put it at tail. if it wasn't active, then mark
* it active and bump active count
*/
if (pg->pqflags & PQ_ACTIVE)
TAILQ_REMOVE(&uvm.page_active, pg, pageq);
else {
pg->pqflags |= PQ_ACTIVE;
uvmexp.active++;
}
TAILQ_INSERT_TAIL(&uvm.page_active, pg, pageq);
}
}
/*
* uvm_pagezero: zero fill a page
*
* => if page is part of an object then the object should be locked
* to protect pg->flags.
*/
PAGE_INLINE void
uvm_pagezero(pg)
struct vm_page *pg;
{
pg->flags &= ~PG_CLEAN;
pmap_zero_page(VM_PAGE_TO_PHYS(pg));
}
/*
* uvm_pagecopy: copy a page
*
* => if page is part of an object then the object should be locked
* to protect pg->flags.
*/
PAGE_INLINE void
uvm_pagecopy(src, dst)
struct vm_page *src, *dst;
{
dst->flags &= ~PG_CLEAN;
pmap_copy_page(VM_PAGE_TO_PHYS(src), VM_PAGE_TO_PHYS(dst));
}
/*
* uvm_page_lookup_freelist: look up the free list for the specified page
*/
PAGE_INLINE int
uvm_page_lookup_freelist(pg)
struct vm_page *pg;
{
int lcv;
lcv = vm_physseg_find(atop(VM_PAGE_TO_PHYS(pg)), NULL);
#ifdef DIAGNOSTIC
if (lcv == -1)
panic("uvm_page_lookup_freelist: unable to locate physseg");
#endif
return (vm_physmem[lcv].free_list);
}
#endif /* defined(UVM_PAGE_INLINE) || defined(UVM_PAGE) */
#endif /* _UVM_UVM_PAGE_I_H_ */