399 lines
11 KiB
C
399 lines
11 KiB
C
/*
|
|
* Copyright (c) 1991 Regents of the University of California.
|
|
* All rights reserved.
|
|
*
|
|
* This code is derived from software contributed to Berkeley by
|
|
* The Mach Operating System project at Carnegie-Mellon University.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions
|
|
* are met:
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
* documentation and/or other materials provided with the distribution.
|
|
* 3. All advertising materials mentioning features or use of this software
|
|
* must display the following acknowledgement:
|
|
* This product includes software developed by the University of
|
|
* California, Berkeley and its contributors.
|
|
* 4. Neither the name of the University nor the names of its contributors
|
|
* may be used to endorse or promote products derived from this software
|
|
* without specific prior written permission.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
|
|
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
* ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
|
|
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
|
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
|
|
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
|
|
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
|
|
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
|
|
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
|
|
* SUCH DAMAGE.
|
|
*
|
|
* from: @(#)vm_pageout.c 7.4 (Berkeley) 5/7/91
|
|
* $Id: vm_pageout.c,v 1.7 1993/12/20 12:40:21 cgd Exp $
|
|
*
|
|
*
|
|
* Copyright (c) 1987, 1990 Carnegie-Mellon University.
|
|
* All rights reserved.
|
|
*
|
|
* Authors: Avadis Tevanian, Jr., Michael Wayne Young
|
|
*
|
|
* Permission to use, copy, modify and distribute this software and
|
|
* its documentation is hereby granted, provided that both the copyright
|
|
* notice and this permission notice appear in all copies of the
|
|
* software, derivative works or modified versions, and any portions
|
|
* thereof, and that both notices appear in supporting documentation.
|
|
*
|
|
* CARNEGIE MELLON ALLOWS FREE USE OF THIS SOFTWARE IN ITS "AS IS"
|
|
* CONDITION. CARNEGIE MELLON DISCLAIMS ANY LIABILITY OF ANY KIND
|
|
* FOR ANY DAMAGES WHATSOEVER RESULTING FROM THE USE OF THIS SOFTWARE.
|
|
*
|
|
* Carnegie Mellon requests users of this software to return to
|
|
*
|
|
* Software Distribution Coordinator or Software.Distribution@CS.CMU.EDU
|
|
* School of Computer Science
|
|
* Carnegie Mellon University
|
|
* Pittsburgh PA 15213-3890
|
|
*
|
|
* any improvements or extensions that they make and grant Carnegie the
|
|
* rights to redistribute these changes.
|
|
*/
|
|
|
|
/*
|
|
* The proverbial page-out daemon.
|
|
*/
|
|
|
|
#include <sys/param.h>
|
|
#include <sys/vmmeter.h>
|
|
|
|
#include <vm/vm.h>
|
|
#include <vm/vm_page.h>
|
|
#include <vm/vm_pageout.h>
|
|
|
|
#include <machine/cpu.h>
|
|
|
|
int vm_pages_needed; /* Event on which pageout daemon sleeps */
|
|
int vm_pageout_free_min = 0; /* Stop pageout to wait for pagers at this free level */
|
|
|
|
int vm_page_free_min_sanity = 40;
|
|
|
|
int vm_page_pagesfreed; /* Pages freed by page daemon */
|
|
|
|
/*
|
|
* vm_pageout_scan does the dirty work for the pageout daemon.
|
|
*/
|
|
vm_pageout_scan()
|
|
{
|
|
register vm_page_t m;
|
|
register int page_shortage;
|
|
register int s;
|
|
register int pages_freed;
|
|
int free;
|
|
|
|
/*
|
|
* Only continue when we want more pages to be "free"
|
|
*/
|
|
|
|
s = splimp();
|
|
simple_lock(&vm_page_queue_free_lock);
|
|
free = vm_page_free_count;
|
|
simple_unlock(&vm_page_queue_free_lock);
|
|
splx(s);
|
|
|
|
if (free < vm_page_free_target) {
|
|
#ifdef OMIT
|
|
swapout_threads();
|
|
#endif /* OMIT*/
|
|
|
|
/*
|
|
* Be sure the pmap system is updated so
|
|
* we can scan the inactive queue.
|
|
*/
|
|
|
|
pmap_update();
|
|
}
|
|
|
|
/*
|
|
* Acquire the resident page system lock,
|
|
* as we may be changing what's resident quite a bit.
|
|
*/
|
|
VM_PAGE_LOCK_QUEUES();
|
|
|
|
/*
|
|
* Start scanning the inactive queue for pages we can free.
|
|
* We keep scanning until we have enough free pages or
|
|
* we have scanned through the entire queue. If we
|
|
* encounter dirty pages, we start cleaning them.
|
|
*/
|
|
|
|
pages_freed = 0;
|
|
m = (vm_page_t) queue_first(&vm_page_queue_inactive);
|
|
while (!queue_end(&vm_page_queue_inactive, (queue_entry_t) m)) {
|
|
vm_page_t next;
|
|
|
|
s = splimp();
|
|
simple_lock(&vm_page_queue_free_lock);
|
|
free = vm_page_free_count;
|
|
simple_unlock(&vm_page_queue_free_lock);
|
|
splx(s);
|
|
|
|
if (free >= vm_page_free_target)
|
|
break;
|
|
|
|
if (m->flags & PG_CLEAN) {
|
|
next = (vm_page_t) queue_next(&m->pageq);
|
|
if (pmap_is_referenced(VM_PAGE_TO_PHYS(m))) {
|
|
vm_page_activate(m);
|
|
vm_stat.reactivations++;
|
|
}
|
|
else {
|
|
register vm_object_t object;
|
|
object = m->object;
|
|
if (!vm_object_lock_try(object)) {
|
|
/*
|
|
* Can't lock object -
|
|
* skip page.
|
|
*/
|
|
m = next;
|
|
continue;
|
|
}
|
|
pmap_page_protect(VM_PAGE_TO_PHYS(m),
|
|
VM_PROT_NONE);
|
|
vm_page_free(m); /* will dequeue */
|
|
pages_freed++;
|
|
vm_object_unlock(object);
|
|
}
|
|
m = next;
|
|
}
|
|
else {
|
|
/*
|
|
* If a page is dirty, then it is either
|
|
* being washed (but not yet cleaned)
|
|
* or it is still in the laundry. If it is
|
|
* still in the laundry, then we start the
|
|
* cleaning operation.
|
|
*/
|
|
|
|
if (m->flags & PG_LAUNDRY) {
|
|
/*
|
|
* Clean the page and remove it from the
|
|
* laundry.
|
|
*
|
|
* We set the busy bit to cause
|
|
* potential page faults on this page to
|
|
* block.
|
|
*
|
|
* And we set pageout-in-progress to keep
|
|
* the object from disappearing during
|
|
* pageout. This guarantees that the
|
|
* page won't move from the inactive
|
|
* queue. (However, any other page on
|
|
* the inactive queue may move!)
|
|
*/
|
|
|
|
register vm_object_t object;
|
|
register vm_pager_t pager;
|
|
int pageout_status;
|
|
|
|
object = m->object;
|
|
if (!vm_object_lock_try(object)) {
|
|
/*
|
|
* Skip page if we can't lock
|
|
* its object
|
|
*/
|
|
m = (vm_page_t) queue_next(&m->pageq);
|
|
continue;
|
|
}
|
|
|
|
pmap_page_protect(VM_PAGE_TO_PHYS(m),
|
|
VM_PROT_NONE);
|
|
m->flags |= PG_BUSY;
|
|
vm_stat.pageouts++;
|
|
|
|
/*
|
|
* Try to collapse the object before
|
|
* making a pager for it. We must
|
|
* unlock the page queues first.
|
|
*/
|
|
VM_PAGE_UNLOCK_QUEUES();
|
|
|
|
vm_object_collapse(object);
|
|
|
|
object->paging_in_progress++;
|
|
vm_object_unlock(object);
|
|
|
|
/*
|
|
* Do a wakeup here in case the following
|
|
* operations block.
|
|
*/
|
|
thread_wakeup((int) &vm_page_free_count);
|
|
|
|
/*
|
|
* If there is no pager for the page,
|
|
* use the default pager. If there's
|
|
* no place to put the page at the
|
|
* moment, leave it in the laundry and
|
|
* hope that there will be paging space
|
|
* later.
|
|
*/
|
|
|
|
if ((pager = object->pager) == NULL) {
|
|
pager = vm_pager_allocate(PG_DFLT,
|
|
(caddr_t)0,
|
|
object->size,
|
|
VM_PROT_ALL,
|
|
0);
|
|
if (pager != NULL) {
|
|
vm_object_setpager(object,
|
|
pager, 0, FALSE);
|
|
}
|
|
}
|
|
pageout_status = pager ?
|
|
vm_pager_put(pager, m, FALSE) :
|
|
VM_PAGER_FAIL;
|
|
vm_object_lock(object);
|
|
VM_PAGE_LOCK_QUEUES();
|
|
next = (vm_page_t) queue_next(&m->pageq);
|
|
|
|
switch (pageout_status) {
|
|
case VM_PAGER_OK:
|
|
case VM_PAGER_PEND:
|
|
m->flags &= ~PG_LAUNDRY;
|
|
break;
|
|
case VM_PAGER_BAD:
|
|
/*
|
|
* Page outside of range of object.
|
|
* Right now we essentially lose the
|
|
* changes by pretending it worked.
|
|
* XXX dubious, what should we do?
|
|
*/
|
|
m->flags &= ~PG_LAUNDRY;
|
|
m->flags |= PG_CLEAN;
|
|
pmap_clear_modify(VM_PAGE_TO_PHYS(m));
|
|
break;
|
|
case VM_PAGER_FAIL:
|
|
/*
|
|
* If page couldn't be paged out, then
|
|
* reactivate the page so it doesn't
|
|
* clog the inactive list. (We will
|
|
* try paging out it again later).
|
|
*/
|
|
vm_page_activate(m);
|
|
break;
|
|
}
|
|
|
|
pmap_clear_reference(VM_PAGE_TO_PHYS(m));
|
|
|
|
/*
|
|
* If the operation is still going, leave
|
|
* the page busy to block all other accesses.
|
|
* Also, leave the paging in progress
|
|
* indicator set so that we don't attempt an
|
|
* object collapse.
|
|
*/
|
|
if (pageout_status != VM_PAGER_PEND) {
|
|
m->flags &= ~PG_BUSY;
|
|
PAGE_WAKEUP(m);
|
|
object->paging_in_progress--;
|
|
}
|
|
thread_wakeup((int) object);
|
|
vm_object_unlock(object);
|
|
m = next;
|
|
}
|
|
else
|
|
m = (vm_page_t) queue_next(&m->pageq);
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Compute the page shortage. If we are still very low on memory
|
|
* be sure that we will move a minimal amount of pages from active
|
|
* to inactive.
|
|
*/
|
|
|
|
page_shortage = vm_page_inactive_target - vm_page_inactive_count;
|
|
page_shortage -= vm_page_free_count;
|
|
|
|
if ((page_shortage <= 0) && (pages_freed == 0))
|
|
page_shortage = 1;
|
|
|
|
while (page_shortage > 0) {
|
|
/*
|
|
* Move some more pages from active to inactive.
|
|
*/
|
|
|
|
if (queue_empty(&vm_page_queue_active)) {
|
|
break;
|
|
}
|
|
m = (vm_page_t) queue_first(&vm_page_queue_active);
|
|
vm_page_deactivate(m);
|
|
page_shortage--;
|
|
}
|
|
|
|
vm_page_pagesfreed += pages_freed;
|
|
VM_PAGE_UNLOCK_QUEUES();
|
|
}
|
|
|
|
/*
|
|
* vm_pageout is the high level pageout daemon.
|
|
*/
|
|
|
|
void vm_pageout()
|
|
{
|
|
(void) spl0();
|
|
|
|
/*
|
|
* Initialize some paging parameters.
|
|
*/
|
|
|
|
if (vm_page_free_min == 0) {
|
|
vm_page_free_min = vm_page_free_count / 20;
|
|
if (vm_page_free_min < 3)
|
|
vm_page_free_min = 3;
|
|
|
|
if (vm_page_free_min > vm_page_free_min_sanity)
|
|
vm_page_free_min = vm_page_free_min_sanity;
|
|
}
|
|
|
|
if (vm_page_free_reserved == 0) {
|
|
if ((vm_page_free_reserved = vm_page_free_min / 2) < 10)
|
|
vm_page_free_reserved = 10;
|
|
}
|
|
if (vm_pageout_free_min == 0) {
|
|
if ((vm_pageout_free_min = vm_page_free_reserved / 2) > 10)
|
|
vm_pageout_free_min = 10;
|
|
}
|
|
|
|
if (vm_page_free_target == 0)
|
|
vm_page_free_target = (vm_page_free_min * 4) / 3;
|
|
|
|
if (vm_page_inactive_target == 0)
|
|
vm_page_inactive_target = vm_page_free_min * 2;
|
|
|
|
if (vm_page_free_target <= vm_page_free_min)
|
|
vm_page_free_target = vm_page_free_min + 1;
|
|
|
|
if (vm_page_inactive_target <= vm_page_free_target)
|
|
vm_page_inactive_target = vm_page_free_target + 1;
|
|
|
|
/*
|
|
* The pageout daemon is never done, so loop
|
|
* forever.
|
|
*/
|
|
|
|
simple_lock(&vm_pages_needed_lock);
|
|
while (TRUE) {
|
|
thread_sleep((int) &vm_pages_needed, &vm_pages_needed_lock,
|
|
FALSE);
|
|
cnt.v_scan++;
|
|
vm_pageout_scan();
|
|
vm_pager_sync();
|
|
simple_lock(&vm_pages_needed_lock);
|
|
thread_wakeup((int) &vm_page_free_count);
|
|
}
|
|
}
|