cbbf514e2c
use both types of list. - Make page coloring and idle zero state per-CPU. - Maintain per-CPU page freelists. When freeing, put pages onto the local CPU's lists and the global lists. When allocating, prefer to take pages from the local CPU. If none are available take from the global list as done now. Proposed on tech-kern@.
487 lines
12 KiB
C
487 lines
12 KiB
C
/* $NetBSD: uvm_meter.c,v 1.49 2008/06/04 12:45:28 ad Exp $ */
|
|
|
|
/*
|
|
* Copyright (c) 1997 Charles D. Cranor and Washington University.
|
|
* Copyright (c) 1982, 1986, 1989, 1993
|
|
* The Regents of the University of California.
|
|
*
|
|
* All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions
|
|
* are met:
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
* documentation and/or other materials provided with the distribution.
|
|
* 3. All advertising materials mentioning features or use of this software
|
|
* must display the following acknowledgement:
|
|
* This product includes software developed by Charles D. Cranor,
|
|
* Washington University, and the University of California, Berkeley
|
|
* and its contributors.
|
|
* 4. Neither the name of the University nor the names of its contributors
|
|
* may be used to endorse or promote products derived from this software
|
|
* without specific prior written permission.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
|
|
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
* ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
|
|
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
|
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
|
|
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
|
|
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
|
|
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
|
|
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
|
|
* SUCH DAMAGE.
|
|
*
|
|
* @(#)vm_meter.c 8.4 (Berkeley) 1/4/94
|
|
* from: Id: uvm_meter.c,v 1.1.2.1 1997/08/14 19:10:35 chuck Exp
|
|
*/
|
|
|
|
#include <sys/cdefs.h>
|
|
__KERNEL_RCSID(0, "$NetBSD: uvm_meter.c,v 1.49 2008/06/04 12:45:28 ad Exp $");
|
|
|
|
#include <sys/param.h>
|
|
#include <sys/proc.h>
|
|
#include <sys/systm.h>
|
|
#include <sys/kernel.h>
|
|
#include <sys/sysctl.h>
|
|
|
|
#include <uvm/uvm_extern.h>
|
|
#include <uvm/uvm_pdpolicy.h>
|
|
|
|
/*
|
|
* maxslp: ???? XXXCDC
|
|
*/
|
|
|
|
int maxslp = MAXSLP; /* patchable ... */
|
|
struct loadavg averunnable;
|
|
|
|
/*
|
|
* constants for averages over 1, 5, and 15 minutes when sampling at
|
|
* 5 second intervals.
|
|
*/
|
|
|
|
static const fixpt_t cexp[3] = {
|
|
0.9200444146293232 * FSCALE, /* exp(-1/12) */
|
|
0.9834714538216174 * FSCALE, /* exp(-1/60) */
|
|
0.9944598480048967 * FSCALE, /* exp(-1/180) */
|
|
};
|
|
|
|
/*
|
|
* prototypes
|
|
*/
|
|
|
|
static void uvm_loadav(struct loadavg *);
|
|
static void uvm_total(struct vmtotal *);
|
|
|
|
/*
|
|
* uvm_meter: calculate load average and wake up the swapper (if needed)
|
|
*/
|
|
void
|
|
uvm_meter(void)
|
|
{
|
|
static int count;
|
|
|
|
if (++count >= 5) {
|
|
count = 0;
|
|
uvm_loadav(&averunnable);
|
|
}
|
|
if (lwp0.l_slptime > (maxslp / 2))
|
|
uvm_kick_scheduler();
|
|
}
|
|
|
|
/*
|
|
* uvm_loadav: compute a tenex style load average of a quantity on
|
|
* 1, 5, and 15 minute intervals.
|
|
*/
|
|
static void
|
|
uvm_loadav(struct loadavg *avg)
|
|
{
|
|
int i, nrun;
|
|
struct lwp *l;
|
|
|
|
nrun = 0;
|
|
|
|
mutex_enter(proc_lock);
|
|
LIST_FOREACH(l, &alllwp, l_list) {
|
|
if ((l->l_flag & (LW_SINTR | LW_SYSTEM)) != 0)
|
|
continue;
|
|
switch (l->l_stat) {
|
|
case LSSLEEP:
|
|
if (l->l_slptime > 1)
|
|
continue;
|
|
/* fall through */
|
|
case LSRUN:
|
|
case LSONPROC:
|
|
case LSIDL:
|
|
nrun++;
|
|
}
|
|
}
|
|
mutex_exit(proc_lock);
|
|
|
|
for (i = 0; i < 3; i++)
|
|
avg->ldavg[i] = (cexp[i] * avg->ldavg[i] +
|
|
nrun * FSCALE * (FSCALE - cexp[i])) >> FSHIFT;
|
|
}
|
|
|
|
/*
|
|
* sysctl helper routine for the vm.vmmeter node.
|
|
*/
|
|
static int
|
|
sysctl_vm_meter(SYSCTLFN_ARGS)
|
|
{
|
|
struct sysctlnode node;
|
|
struct vmtotal vmtotals;
|
|
|
|
node = *rnode;
|
|
node.sysctl_data = &vmtotals;
|
|
uvm_total(&vmtotals);
|
|
|
|
return (sysctl_lookup(SYSCTLFN_CALL(&node)));
|
|
}
|
|
|
|
/*
|
|
* sysctl helper routine for the vm.uvmexp node.
|
|
*/
|
|
static int
|
|
sysctl_vm_uvmexp(SYSCTLFN_ARGS)
|
|
{
|
|
struct sysctlnode node;
|
|
|
|
node = *rnode;
|
|
if (oldp)
|
|
node.sysctl_size = min(*oldlenp, node.sysctl_size);
|
|
|
|
return (sysctl_lookup(SYSCTLFN_CALL(&node)));
|
|
}
|
|
|
|
static int
|
|
sysctl_vm_uvmexp2(SYSCTLFN_ARGS)
|
|
{
|
|
struct sysctlnode node;
|
|
struct uvmexp_sysctl u;
|
|
int active, inactive;
|
|
|
|
uvm_estimatepageable(&active, &inactive);
|
|
|
|
memset(&u, 0, sizeof(u));
|
|
|
|
/* Entries here are in order of uvmexp_sysctl, not uvmexp */
|
|
u.pagesize = uvmexp.pagesize;
|
|
u.pagemask = uvmexp.pagemask;
|
|
u.pageshift = uvmexp.pageshift;
|
|
u.npages = uvmexp.npages;
|
|
u.free = uvmexp.free;
|
|
u.active = active;
|
|
u.inactive = inactive;
|
|
u.paging = uvmexp.paging;
|
|
u.wired = uvmexp.wired;
|
|
u.zeropages = uvmexp.zeropages;
|
|
u.reserve_pagedaemon = uvmexp.reserve_pagedaemon;
|
|
u.reserve_kernel = uvmexp.reserve_kernel;
|
|
u.freemin = uvmexp.freemin;
|
|
u.freetarg = uvmexp.freetarg;
|
|
u.inactarg = 0; /* unused */
|
|
u.wiredmax = uvmexp.wiredmax;
|
|
u.nswapdev = uvmexp.nswapdev;
|
|
u.swpages = uvmexp.swpages;
|
|
u.swpginuse = uvmexp.swpginuse;
|
|
u.swpgonly = uvmexp.swpgonly;
|
|
u.nswget = uvmexp.nswget;
|
|
u.faults = uvmexp.faults;
|
|
u.traps = uvmexp.traps;
|
|
u.intrs = uvmexp.intrs;
|
|
u.swtch = uvmexp.swtch;
|
|
u.softs = uvmexp.softs;
|
|
u.syscalls = uvmexp.syscalls;
|
|
u.pageins = uvmexp.pageins;
|
|
u.swapins = uvmexp.swapins;
|
|
u.swapouts = uvmexp.swapouts;
|
|
u.pgswapin = uvmexp.pgswapin;
|
|
u.pgswapout = uvmexp.pgswapout;
|
|
u.forks = uvmexp.forks;
|
|
u.forks_ppwait = uvmexp.forks_ppwait;
|
|
u.forks_sharevm = uvmexp.forks_sharevm;
|
|
u.pga_zerohit = uvmexp.pga_zerohit;
|
|
u.pga_zeromiss = uvmexp.pga_zeromiss;
|
|
u.zeroaborts = uvmexp.zeroaborts;
|
|
u.fltnoram = uvmexp.fltnoram;
|
|
u.fltnoanon = uvmexp.fltnoanon;
|
|
u.fltpgwait = uvmexp.fltpgwait;
|
|
u.fltpgrele = uvmexp.fltpgrele;
|
|
u.fltrelck = uvmexp.fltrelck;
|
|
u.fltrelckok = uvmexp.fltrelckok;
|
|
u.fltanget = uvmexp.fltanget;
|
|
u.fltanretry = uvmexp.fltanretry;
|
|
u.fltamcopy = uvmexp.fltamcopy;
|
|
u.fltnamap = uvmexp.fltnamap;
|
|
u.fltnomap = uvmexp.fltnomap;
|
|
u.fltlget = uvmexp.fltlget;
|
|
u.fltget = uvmexp.fltget;
|
|
u.flt_anon = uvmexp.flt_anon;
|
|
u.flt_acow = uvmexp.flt_acow;
|
|
u.flt_obj = uvmexp.flt_obj;
|
|
u.flt_prcopy = uvmexp.flt_prcopy;
|
|
u.flt_przero = uvmexp.flt_przero;
|
|
u.pdwoke = uvmexp.pdwoke;
|
|
u.pdrevs = uvmexp.pdrevs;
|
|
u.pdswout = uvmexp.pdswout;
|
|
u.pdfreed = uvmexp.pdfreed;
|
|
u.pdscans = uvmexp.pdscans;
|
|
u.pdanscan = uvmexp.pdanscan;
|
|
u.pdobscan = uvmexp.pdobscan;
|
|
u.pdreact = uvmexp.pdreact;
|
|
u.pdbusy = uvmexp.pdbusy;
|
|
u.pdpageouts = uvmexp.pdpageouts;
|
|
u.pdpending = uvmexp.pdpending;
|
|
u.pddeact = uvmexp.pddeact;
|
|
u.anonpages = uvmexp.anonpages;
|
|
u.filepages = uvmexp.filepages;
|
|
u.execpages = uvmexp.execpages;
|
|
u.colorhit = uvmexp.colorhit;
|
|
u.colormiss = uvmexp.colormiss;
|
|
u.cpuhit = uvmexp.cpuhit;
|
|
u.cpumiss = uvmexp.cpumiss;
|
|
|
|
node = *rnode;
|
|
node.sysctl_data = &u;
|
|
node.sysctl_size = sizeof(u);
|
|
return (sysctl_lookup(SYSCTLFN_CALL(&node)));
|
|
}
|
|
|
|
/*
|
|
* sysctl helper routine for uvm_pctparam.
|
|
*/
|
|
static int
|
|
uvm_sysctlpctparam(SYSCTLFN_ARGS)
|
|
{
|
|
int t, error;
|
|
struct sysctlnode node;
|
|
struct uvm_pctparam *pct;
|
|
|
|
pct = rnode->sysctl_data;
|
|
t = pct->pct_pct;
|
|
|
|
node = *rnode;
|
|
node.sysctl_data = &t;
|
|
error = sysctl_lookup(SYSCTLFN_CALL(&node));
|
|
if (error || newp == NULL)
|
|
return error;
|
|
|
|
if (t < 0 || t > 100)
|
|
return EINVAL;
|
|
|
|
error = uvm_pctparam_check(pct, t);
|
|
if (error) {
|
|
return error;
|
|
}
|
|
uvm_pctparam_set(pct, t);
|
|
|
|
return (0);
|
|
}
|
|
|
|
/*
|
|
* uvm_sysctl: sysctl hook into UVM system.
|
|
*/
|
|
SYSCTL_SETUP(sysctl_vm_setup, "sysctl vm subtree setup")
|
|
{
|
|
|
|
sysctl_createv(clog, 0, NULL, NULL,
|
|
CTLFLAG_PERMANENT,
|
|
CTLTYPE_NODE, "vm", NULL,
|
|
NULL, 0, NULL, 0,
|
|
CTL_VM, CTL_EOL);
|
|
sysctl_createv(clog, 0, NULL, NULL,
|
|
CTLFLAG_PERMANENT,
|
|
CTLTYPE_STRUCT, "vmmeter",
|
|
SYSCTL_DESCR("Simple system-wide virtual memory "
|
|
"statistics"),
|
|
sysctl_vm_meter, 0, NULL, sizeof(struct vmtotal),
|
|
CTL_VM, VM_METER, CTL_EOL);
|
|
sysctl_createv(clog, 0, NULL, NULL,
|
|
CTLFLAG_PERMANENT,
|
|
CTLTYPE_STRUCT, "loadavg",
|
|
SYSCTL_DESCR("System load average history"),
|
|
NULL, 0, &averunnable, sizeof(averunnable),
|
|
CTL_VM, VM_LOADAVG, CTL_EOL);
|
|
sysctl_createv(clog, 0, NULL, NULL,
|
|
CTLFLAG_PERMANENT,
|
|
CTLTYPE_STRUCT, "uvmexp",
|
|
SYSCTL_DESCR("Detailed system-wide virtual memory "
|
|
"statistics"),
|
|
sysctl_vm_uvmexp, 0, &uvmexp, sizeof(uvmexp),
|
|
CTL_VM, VM_UVMEXP, CTL_EOL);
|
|
sysctl_createv(clog, 0, NULL, NULL,
|
|
CTLFLAG_PERMANENT,
|
|
CTLTYPE_INT, "nkmempages",
|
|
SYSCTL_DESCR("Default number of pages in kmem_map"),
|
|
NULL, 0, &nkmempages, 0,
|
|
CTL_VM, VM_NKMEMPAGES, CTL_EOL);
|
|
sysctl_createv(clog, 0, NULL, NULL,
|
|
CTLFLAG_PERMANENT,
|
|
CTLTYPE_STRUCT, "uvmexp2",
|
|
SYSCTL_DESCR("Detailed system-wide virtual memory "
|
|
"statistics (MI)"),
|
|
sysctl_vm_uvmexp2, 0, NULL, 0,
|
|
CTL_VM, VM_UVMEXP2, CTL_EOL);
|
|
sysctl_createv(clog, 0, NULL, NULL,
|
|
CTLFLAG_PERMANENT, CTLTYPE_INT, "maxslp",
|
|
SYSCTL_DESCR("Maximum process sleep time before being "
|
|
"swapped"),
|
|
NULL, 0, &maxslp, 0,
|
|
CTL_VM, VM_MAXSLP, CTL_EOL);
|
|
sysctl_createv(clog, 0, NULL, NULL,
|
|
CTLFLAG_PERMANENT|CTLFLAG_IMMEDIATE,
|
|
CTLTYPE_INT, "uspace",
|
|
SYSCTL_DESCR("Number of bytes allocated for a kernel "
|
|
"stack"),
|
|
NULL, USPACE, NULL, 0,
|
|
CTL_VM, VM_USPACE, CTL_EOL);
|
|
sysctl_createv(clog, 0, NULL, NULL,
|
|
CTLFLAG_PERMANENT|CTLFLAG_READWRITE,
|
|
CTLTYPE_INT, "idlezero",
|
|
SYSCTL_DESCR("Whether try to zero pages in idle loop"),
|
|
NULL, 0, &vm_page_zero_enable, 0,
|
|
CTL_VM, CTL_CREATE, CTL_EOL);
|
|
|
|
uvmpdpol_sysctlsetup();
|
|
}
|
|
|
|
/*
|
|
* uvm_total: calculate the current state of the system.
|
|
*/
|
|
static void
|
|
uvm_total(struct vmtotal *totalp)
|
|
{
|
|
struct lwp *l;
|
|
#if 0
|
|
struct vm_map_entry * entry;
|
|
struct vm_map *map;
|
|
int paging;
|
|
#endif
|
|
int active;
|
|
|
|
memset(totalp, 0, sizeof *totalp);
|
|
|
|
/*
|
|
* calculate process statistics
|
|
*/
|
|
mutex_enter(proc_lock);
|
|
LIST_FOREACH(l, &alllwp, l_list) {
|
|
if (l->l_proc->p_flag & PK_SYSTEM)
|
|
continue;
|
|
switch (l->l_stat) {
|
|
case 0:
|
|
continue;
|
|
|
|
case LSSLEEP:
|
|
case LSSTOP:
|
|
if (l->l_flag & LW_INMEM) {
|
|
if (lwp_eprio(l) <= PZERO)
|
|
totalp->t_dw++;
|
|
else if (l->l_slptime < maxslp)
|
|
totalp->t_sl++;
|
|
} else if (l->l_slptime < maxslp)
|
|
totalp->t_sw++;
|
|
if (l->l_slptime >= maxslp)
|
|
continue;
|
|
break;
|
|
|
|
case LSRUN:
|
|
case LSONPROC:
|
|
case LSIDL:
|
|
if (l->l_flag & LW_INMEM)
|
|
totalp->t_rq++;
|
|
else
|
|
totalp->t_sw++;
|
|
if (l->l_stat == LSIDL)
|
|
continue;
|
|
break;
|
|
}
|
|
/*
|
|
* note active objects
|
|
*/
|
|
#if 0
|
|
/*
|
|
* XXXCDC: BOGUS! rethink this. in the mean time
|
|
* don't do it.
|
|
*/
|
|
paging = 0;
|
|
vm_map_lock(map);
|
|
for (map = &p->p_vmspace->vm_map, entry = map->header.next;
|
|
entry != &map->header; entry = entry->next) {
|
|
if (entry->is_a_map || entry->is_sub_map ||
|
|
entry->object.uvm_obj == NULL)
|
|
continue;
|
|
/* XXX how to do this with uvm */
|
|
}
|
|
vm_map_unlock(map);
|
|
if (paging)
|
|
totalp->t_pw++;
|
|
#endif
|
|
}
|
|
mutex_exit(proc_lock);
|
|
|
|
/*
|
|
* Calculate object memory usage statistics.
|
|
*/
|
|
uvm_estimatepageable(&active, NULL);
|
|
totalp->t_free = uvmexp.free;
|
|
totalp->t_vm = uvmexp.npages - uvmexp.free + uvmexp.swpginuse;
|
|
totalp->t_avm = active + uvmexp.swpginuse; /* XXX */
|
|
totalp->t_rm = uvmexp.npages - uvmexp.free;
|
|
totalp->t_arm = active;
|
|
totalp->t_vmshr = 0; /* XXX */
|
|
totalp->t_avmshr = 0; /* XXX */
|
|
totalp->t_rmshr = 0; /* XXX */
|
|
totalp->t_armshr = 0; /* XXX */
|
|
}
|
|
|
|
void
|
|
uvm_pctparam_set(struct uvm_pctparam *pct, int val)
|
|
{
|
|
|
|
pct->pct_pct = val;
|
|
pct->pct_scaled = val * UVM_PCTPARAM_SCALE / 100;
|
|
}
|
|
|
|
int
|
|
uvm_pctparam_get(struct uvm_pctparam *pct)
|
|
{
|
|
|
|
return pct->pct_pct;
|
|
}
|
|
|
|
int
|
|
uvm_pctparam_check(struct uvm_pctparam *pct, int val)
|
|
{
|
|
|
|
if (pct->pct_check == NULL) {
|
|
return 0;
|
|
}
|
|
return (*pct->pct_check)(pct, val);
|
|
}
|
|
|
|
void
|
|
uvm_pctparam_init(struct uvm_pctparam *pct, int val,
|
|
int (*fn)(struct uvm_pctparam *, int))
|
|
{
|
|
|
|
pct->pct_check = fn;
|
|
uvm_pctparam_set(pct, val);
|
|
}
|
|
|
|
int
|
|
uvm_pctparam_createsysctlnode(struct uvm_pctparam *pct, const char *name,
|
|
const char *desc)
|
|
{
|
|
|
|
return sysctl_createv(NULL, 0, NULL, NULL,
|
|
CTLFLAG_PERMANENT|CTLFLAG_READWRITE,
|
|
CTLTYPE_INT, name, SYSCTL_DESCR(desc),
|
|
uvm_sysctlpctparam, 0, pct, 0, CTL_VM, CTL_CREATE, CTL_EOL);
|
|
}
|