Trailing whitespace
This commit is contained in:
parent
0f25b54f10
commit
4e62681fb0
|
@ -1,4 +1,4 @@
|
|||
/* $NetBSD: uvm_amap.c,v 1.120 2020/05/17 15:07:22 ad Exp $ */
|
||||
/* $NetBSD: uvm_amap.c,v 1.121 2020/07/08 13:26:22 skrll Exp $ */
|
||||
|
||||
/*
|
||||
* Copyright (c) 1997 Charles D. Cranor and Washington University.
|
||||
|
@ -35,7 +35,7 @@
|
|||
*/
|
||||
|
||||
#include <sys/cdefs.h>
|
||||
__KERNEL_RCSID(0, "$NetBSD: uvm_amap.c,v 1.120 2020/05/17 15:07:22 ad Exp $");
|
||||
__KERNEL_RCSID(0, "$NetBSD: uvm_amap.c,v 1.121 2020/07/08 13:26:22 skrll Exp $");
|
||||
|
||||
#include "opt_uvmhist.h"
|
||||
|
||||
|
@ -859,7 +859,7 @@ amap_copy(struct vm_map *map, struct vm_map_entry *entry, int flags,
|
|||
}
|
||||
|
||||
/*
|
||||
* First check and see if we are the only map entry referencing
|
||||
* First check and see if we are the only map entry referencing
|
||||
* he amap we currently have. If so, then just take it over instead
|
||||
* of copying it. Note that we are reading am_ref without lock held
|
||||
* as the value value can only be one if we have the only reference
|
||||
|
@ -1388,7 +1388,7 @@ amap_swap_off(int startslot, int endslot)
|
|||
if (am) {
|
||||
amap_unlock(am);
|
||||
}
|
||||
|
||||
|
||||
mutex_enter(&amap_list_lock);
|
||||
KASSERT(LIST_NEXT(&marker_prev, am_list) == &marker_next ||
|
||||
LIST_NEXT(LIST_NEXT(&marker_prev, am_list), am_list) ==
|
||||
|
|
|
@ -1,4 +1,4 @@
|
|||
/* $NetBSD: uvm_anon.c,v 1.77 2020/03/22 18:32:42 ad Exp $ */
|
||||
/* $NetBSD: uvm_anon.c,v 1.78 2020/07/08 13:26:22 skrll Exp $ */
|
||||
|
||||
/*
|
||||
* Copyright (c) 1997 Charles D. Cranor and Washington University.
|
||||
|
@ -30,7 +30,7 @@
|
|||
*/
|
||||
|
||||
#include <sys/cdefs.h>
|
||||
__KERNEL_RCSID(0, "$NetBSD: uvm_anon.c,v 1.77 2020/03/22 18:32:42 ad Exp $");
|
||||
__KERNEL_RCSID(0, "$NetBSD: uvm_anon.c,v 1.78 2020/07/08 13:26:22 skrll Exp $");
|
||||
|
||||
#include "opt_uvmhist.h"
|
||||
|
||||
|
@ -235,7 +235,7 @@ uvm_anon_lockloanpg(struct vm_anon *anon)
|
|||
/*
|
||||
* someone locking the object has a chance to
|
||||
* lock us right now
|
||||
*
|
||||
*
|
||||
* XXX Better than yielding but inadequate.
|
||||
*/
|
||||
mutex_exit(&pg->interlock);
|
||||
|
|
|
@ -1,4 +1,4 @@
|
|||
/* $NetBSD: uvm_aobj.c,v 1.147 2020/05/25 22:04:51 ad Exp $ */
|
||||
/* $NetBSD: uvm_aobj.c,v 1.148 2020/07/08 13:26:22 skrll Exp $ */
|
||||
|
||||
/*
|
||||
* Copyright (c) 1998 Chuck Silvers, Charles D. Cranor and
|
||||
|
@ -38,7 +38,7 @@
|
|||
*/
|
||||
|
||||
#include <sys/cdefs.h>
|
||||
__KERNEL_RCSID(0, "$NetBSD: uvm_aobj.c,v 1.147 2020/05/25 22:04:51 ad Exp $");
|
||||
__KERNEL_RCSID(0, "$NetBSD: uvm_aobj.c,v 1.148 2020/07/08 13:26:22 skrll Exp $");
|
||||
|
||||
#ifdef _KERNEL_OPT
|
||||
#include "opt_uvmhist.h"
|
||||
|
@ -965,7 +965,7 @@ uao_get(struct uvm_object *uobj, voff_t offset, struct vm_page **pps,
|
|||
}
|
||||
|
||||
/*
|
||||
* if swslot == 0, page hasn't existed before and is zeroed.
|
||||
* if swslot == 0, page hasn't existed before and is zeroed.
|
||||
* otherwise we have a "fake/busy/clean" page that we just
|
||||
* allocated. do the needed "i/o", reading from swap.
|
||||
*/
|
||||
|
|
|
@ -1,4 +1,4 @@
|
|||
/* $NetBSD: uvm_bio.c,v 1.118 2020/06/25 14:04:30 jdolecek Exp $ */
|
||||
/* $NetBSD: uvm_bio.c,v 1.119 2020/07/08 13:26:22 skrll Exp $ */
|
||||
|
||||
/*
|
||||
* Copyright (c) 1998 Chuck Silvers.
|
||||
|
@ -34,7 +34,7 @@
|
|||
*/
|
||||
|
||||
#include <sys/cdefs.h>
|
||||
__KERNEL_RCSID(0, "$NetBSD: uvm_bio.c,v 1.118 2020/06/25 14:04:30 jdolecek Exp $");
|
||||
__KERNEL_RCSID(0, "$NetBSD: uvm_bio.c,v 1.119 2020/07/08 13:26:22 skrll Exp $");
|
||||
|
||||
#include "opt_uvmhist.h"
|
||||
#include "opt_ubc.h"
|
||||
|
@ -747,12 +747,12 @@ ubc_uiomove(struct uvm_object *uobj, struct uio *uio, vsize_t todo, int advice,
|
|||
/*
|
||||
* during direct access pages need to be held busy to prevent them
|
||||
* changing identity, and therefore if we read or write an object
|
||||
* into a mapped view of same we could deadlock while faulting.
|
||||
* into a mapped view of same we could deadlock while faulting.
|
||||
*
|
||||
* avoid the problem by disallowing direct access if the object
|
||||
* might be visible somewhere via mmap().
|
||||
*
|
||||
* XXX concurrent reads cause thundering herd issues with PG_BUSY.
|
||||
* XXX concurrent reads cause thundering herd issues with PG_BUSY.
|
||||
* In the future enable by default for writes or if ncpu<=2, and
|
||||
* make the toggle override that.
|
||||
*/
|
||||
|
@ -914,7 +914,7 @@ again:
|
|||
* dirty.
|
||||
*/
|
||||
if ((flags & UBC_WRITE) != 0) {
|
||||
uvm_pagemarkdirty(pg, UVM_PAGE_STATUS_DIRTY);
|
||||
uvm_pagemarkdirty(pg, UVM_PAGE_STATUS_DIRTY);
|
||||
}
|
||||
}
|
||||
rw_exit(uobj->vmobjlock);
|
||||
|
|
|
@ -1,4 +1,4 @@
|
|||
/* $NetBSD: uvm_km.c,v 1.157 2020/03/14 20:23:51 ad Exp $ */
|
||||
/* $NetBSD: uvm_km.c,v 1.158 2020/07/08 13:26:22 skrll Exp $ */
|
||||
|
||||
/*
|
||||
* Copyright (c) 1997 Charles D. Cranor and Washington University.
|
||||
|
@ -152,7 +152,7 @@
|
|||
*/
|
||||
|
||||
#include <sys/cdefs.h>
|
||||
__KERNEL_RCSID(0, "$NetBSD: uvm_km.c,v 1.157 2020/03/14 20:23:51 ad Exp $");
|
||||
__KERNEL_RCSID(0, "$NetBSD: uvm_km.c,v 1.158 2020/07/08 13:26:22 skrll Exp $");
|
||||
|
||||
#include "opt_uvmhist.h"
|
||||
|
||||
|
@ -819,7 +819,7 @@ again:
|
|||
|
||||
#ifdef PMAP_GROWKERNEL
|
||||
/*
|
||||
* These VA allocations happen independently of uvm_map
|
||||
* These VA allocations happen independently of uvm_map
|
||||
* so this allocation must not extend beyond the current limit.
|
||||
*/
|
||||
KASSERTMSG(uvm_maxkaddr >= va + size,
|
||||
|
|
|
@ -1,4 +1,4 @@
|
|||
/* $NetBSD: uvm_pager.c,v 1.126 2020/06/25 09:58:44 jdolecek Exp $ */
|
||||
/* $NetBSD: uvm_pager.c,v 1.127 2020/07/08 13:26:22 skrll Exp $ */
|
||||
|
||||
/*
|
||||
* Copyright (c) 1997 Charles D. Cranor and Washington University.
|
||||
|
@ -32,7 +32,7 @@
|
|||
*/
|
||||
|
||||
#include <sys/cdefs.h>
|
||||
__KERNEL_RCSID(0, "$NetBSD: uvm_pager.c,v 1.126 2020/06/25 09:58:44 jdolecek Exp $");
|
||||
__KERNEL_RCSID(0, "$NetBSD: uvm_pager.c,v 1.127 2020/07/08 13:26:22 skrll Exp $");
|
||||
|
||||
#include "opt_uvmhist.h"
|
||||
#include "opt_readahead.h"
|
||||
|
@ -195,7 +195,7 @@ uvm_pagermapin(struct vm_page **pps, int npages, int flags)
|
|||
(uintptr_t)pps, npages, first_color, 0);
|
||||
|
||||
#ifdef PMAP_DIRECT
|
||||
/*
|
||||
/*
|
||||
* for a single page the direct mapped segment can be used.
|
||||
*/
|
||||
|
||||
|
@ -285,7 +285,7 @@ uvm_pagermapout(vaddr_t kva, int npages)
|
|||
UVMHIST_LOG(maphist, " (kva=%#jx, npages=%jd)", kva, npages,0,0);
|
||||
|
||||
#ifdef PMAP_DIRECT
|
||||
/*
|
||||
/*
|
||||
* solitary pages are mapped directly.
|
||||
*/
|
||||
|
||||
|
|
|
@ -1,4 +1,4 @@
|
|||
/* $NetBSD: uvm_swap.c,v 1.195 2020/06/29 23:40:28 riastradh Exp $ */
|
||||
/* $NetBSD: uvm_swap.c,v 1.196 2020/07/08 13:26:22 skrll Exp $ */
|
||||
|
||||
/*
|
||||
* Copyright (c) 1995, 1996, 1997, 2009 Matthew R. Green
|
||||
|
@ -30,7 +30,7 @@
|
|||
*/
|
||||
|
||||
#include <sys/cdefs.h>
|
||||
__KERNEL_RCSID(0, "$NetBSD: uvm_swap.c,v 1.195 2020/06/29 23:40:28 riastradh Exp $");
|
||||
__KERNEL_RCSID(0, "$NetBSD: uvm_swap.c,v 1.196 2020/07/08 13:26:22 skrll Exp $");
|
||||
|
||||
#include "opt_uvmhist.h"
|
||||
#include "opt_compat_netbsd.h"
|
||||
|
@ -553,7 +553,7 @@ sys_swapctl(struct lwp *l, const struct sys_swapctl_args *uap, register_t *retva
|
|||
NULL, sizeof(struct swapent), retval);
|
||||
UVMHIST_LOG(pdhist, "<- done SWAP_STATS", 0, 0, 0, 0);
|
||||
goto out;
|
||||
|
||||
|
||||
case SWAP_GETDUMPDEV:
|
||||
error = copyout(&dumpdev, SCARG(uap, arg), sizeof(dumpdev));
|
||||
goto out;
|
||||
|
|
Loading…
Reference in New Issue