NetBSD/sys/vm/vnode_pager.c

688 lines
18 KiB
C

/* $NetBSD: vnode_pager.c,v 1.35 1998/03/01 02:24:03 fvdl Exp $ */
/*
* Copyright (c) 1990 University of Utah.
* Copyright (c) 1991, 1993
* The Regents of the University of California. All rights reserved.
*
* This code is derived from software contributed to Berkeley by
* the Systems Programming Group of the University of Utah Computer
* Science Department.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
* 1. Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
* 2. Redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in the
* documentation and/or other materials provided with the distribution.
* 3. All advertising materials mentioning features or use of this software
* must display the following acknowledgement:
* This product includes software developed by the University of
* California, Berkeley and its contributors.
* 4. Neither the name of the University nor the names of its contributors
* may be used to endorse or promote products derived from this software
* without specific prior written permission.
*
* THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
* ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
* SUCH DAMAGE.
*
* @(#)vnode_pager.c 8.10 (Berkeley) 5/14/95
*/
/*
* Page to/from files (vnodes).
*
* TODO:
* pageouts
* fix credential use (uses current process credentials now)
*/
#include "fs_nfs.h"
#include <sys/param.h>
#include <sys/systm.h>
#include <sys/proc.h>
#include <sys/malloc.h>
#include <sys/vnode.h>
#include <sys/uio.h>
#include <sys/mount.h>
#include <vm/vm.h>
#include <vm/vm_page.h>
#include <vm/vnode_pager.h>
struct pagerlst vnode_pager_list; /* list of managed vnodes */
simple_lock_data_t vnode_pager_list_lock;
lock_data_t vnode_pager_sync_lock;
#ifdef DEBUG
int vpagerdebug = 0x00;
#define VDB_FOLLOW 0x01
#define VDB_INIT 0x02
#define VDB_IO 0x04
#define VDB_FAIL 0x08
#define VDB_ALLOC 0x10
#define VDB_SIZE 0x20
#endif
static vm_pager_t vnode_pager_alloc
__P((caddr_t, vm_size_t, vm_prot_t, vm_offset_t));
static void vnode_pager_cluster
__P((vm_pager_t, vm_offset_t,
vm_offset_t *, vm_offset_t *));
static void vnode_pager_dealloc __P((vm_pager_t));
static int vnode_pager_getpage
__P((vm_pager_t, vm_page_t *, int, boolean_t));
static boolean_t vnode_pager_haspage __P((vm_pager_t, vm_offset_t));
static void vnode_pager_init __P((void));
static int vnode_pager_io
__P((vn_pager_t, vm_page_t *, int,
boolean_t, enum uio_rw));
static boolean_t vnode_pager_putpage
__P((vm_pager_t, vm_page_t *, int, boolean_t));
struct pagerops vnodepagerops = {
vnode_pager_init,
vnode_pager_alloc,
vnode_pager_dealloc,
vnode_pager_getpage,
vnode_pager_putpage,
vnode_pager_haspage,
vnode_pager_cluster
};
static void
vnode_pager_init()
{
#ifdef DEBUG
if (vpagerdebug & VDB_FOLLOW)
printf("vnode_pager_init()\n");
#endif
TAILQ_INIT(&vnode_pager_list);
simple_lock_init(&vnode_pager_list_lock);
lockinit(&vnode_pager_sync_lock, PVM, "vnsync", 0, 0);
}
/*
* Allocate (or lookup) pager for a vnode.
* Handle is a vnode pointer.
*/
static vm_pager_t
vnode_pager_alloc(handle, size, prot, foff)
caddr_t handle;
vm_size_t size;
vm_prot_t prot;
vm_offset_t foff;
{
register vm_pager_t pager;
register vn_pager_t vnp;
vm_object_t object;
struct vattr vattr;
struct vnode *vp;
u_quad_t used_vnode_size;
struct proc *p = curproc; /* XXX */
#ifdef DEBUG
if (vpagerdebug & (VDB_FOLLOW|VDB_ALLOC))
printf("vnode_pager_alloc(%p, %lx, %x)\n", handle, size, prot);
#endif
/*
* Pageout to vnode, no can do yet.
*/
if (handle == NULL)
return(NULL);
/*
* Vnodes keep a pointer to any associated pager so no need to
* lookup with vm_pager_lookup.
*/
vp = (struct vnode *)handle;
pager = (vm_pager_t)vp->v_vmdata;
if (pager == NULL) {
/*
* Allocate pager structures
*/
pager = (vm_pager_t)malloc(sizeof *pager, M_VMPAGER, M_WAITOK);
if (pager == NULL)
return(NULL);
vnp = (vn_pager_t)malloc(sizeof *vnp, M_VMPGDATA, M_WAITOK);
if (vnp == NULL) {
free((caddr_t)pager, M_VMPAGER);
return(NULL);
}
/*
* And an object of the appropriate size
*/
if (VOP_GETATTR(vp, &vattr, p->p_ucred, p) != 0) {
free((caddr_t)vnp, M_VMPGDATA);
free((caddr_t)pager, M_VMPAGER);
return(NULL);
}
/* make sure mapping fits into numeric range,
truncate if necessary */
used_vnode_size = vattr.va_size;
if (used_vnode_size > (vm_offset_t)-PAGE_SIZE) {
#ifdef DEBUG
printf("vnode_pager_alloc: vn %p size truncated %qx->%lx\n",
vp, used_vnode_size, (vm_offset_t)-PAGE_SIZE);
#endif
used_vnode_size = (vm_offset_t)-PAGE_SIZE;
}
object = vm_object_allocate(round_page(used_vnode_size));
vm_object_enter(object, pager);
vm_object_setpager(object, pager, 0, TRUE);
/*
* Hold a reference to the vnode and initialize pager data.
*/
VREF(vp);
vnp->vnp_flags = 0;
vnp->vnp_vp = vp;
vnp->vnp_size = used_vnode_size;
pager->pg_handle = handle;
pager->pg_type = PG_VNODE;
pager->pg_flags = 0;
pager->pg_ops = &vnodepagerops;
pager->pg_data = vnp;
vp->v_vmdata = (caddr_t)pager;
simple_lock(&vnode_pager_list_lock);
TAILQ_INSERT_TAIL(&vnode_pager_list, pager, pg_list);
simple_unlock(&vnode_pager_list_lock);
} else {
/*
* vm_object_lookup() will remove the object from the
* cache if found and also gain a reference to the object.
*/
object = vm_object_lookup(pager);
#ifdef DEBUG
vnp = (vn_pager_t)pager->pg_data;
#endif
}
#ifdef DEBUG
if (vpagerdebug & VDB_ALLOC)
printf("vnode_pager_setup: vp %p sz %lx pager %p object %p\n",
vp, vnp->vnp_size, pager, object);
#endif
return(pager);
}
static void
vnode_pager_dealloc(pager)
vm_pager_t pager;
{
register vn_pager_t vnp = (vn_pager_t)pager->pg_data;
register struct vnode *vp;
#ifdef NOTDEF
struct proc *p = curproc; /* XXX */
#endif
#ifdef DEBUG
if (vpagerdebug & VDB_FOLLOW)
printf("vnode_pager_dealloc(%p)\n", pager);
#endif
simple_lock(&vnode_pager_list_lock);
TAILQ_REMOVE(&vnode_pager_list, pager, pg_list);
simple_unlock(&vnode_pager_list_lock);
if ((vp = vnp->vnp_vp) != NULL) {
vp->v_vmdata = NULL;
vp->v_flag &= ~VTEXT;
#if NOTDEF
/* can hang if done at reboot on NFS FS */
(void) VOP_FSYNC(vp, p->p_ucred, p);
#endif
vrele(vp);
}
free((caddr_t)vnp, M_VMPGDATA);
free((caddr_t)pager, M_VMPAGER);
}
static int
vnode_pager_getpage(pager, mlist, npages, sync)
vm_pager_t pager;
vm_page_t *mlist;
int npages;
boolean_t sync;
{
#ifdef DEBUG
if (vpagerdebug & VDB_FOLLOW)
printf("vnode_pager_getpage(%p, %p, %x, %x)\n",
pager, mlist, npages, sync);
#endif
return(vnode_pager_io((vn_pager_t)pager->pg_data,
mlist, npages, sync, UIO_READ));
}
static boolean_t
vnode_pager_putpage(pager, mlist, npages, sync)
vm_pager_t pager;
vm_page_t *mlist;
int npages;
boolean_t sync;
{
int err;
#ifdef DEBUG
if (vpagerdebug & VDB_FOLLOW)
printf("vnode_pager_putpage(%p, %p, %x, %x)\n",
pager, mlist, npages, sync);
#endif
if (pager == NULL)
return (FALSE); /* ??? */
err = vnode_pager_io((vn_pager_t)pager->pg_data,
mlist, npages, sync, UIO_WRITE);
/*
* If the operation was successful, mark the pages clean.
*/
if (err == VM_PAGER_OK) {
while (npages--) {
(*mlist)->flags |= PG_CLEAN;
pmap_clear_modify(VM_PAGE_TO_PHYS(*mlist));
mlist++;
}
}
return(err);
}
static boolean_t
vnode_pager_haspage(pager, offset)
vm_pager_t pager;
vm_offset_t offset;
{
register vn_pager_t vnp = (vn_pager_t)pager->pg_data;
daddr_t bn;
int err;
#ifdef DEBUG
if (vpagerdebug & VDB_FOLLOW)
printf("vnode_pager_haspage(%p, %lx)\n", pager, offset);
#endif
/*
* Offset beyond end of file, do not have the page
* Lock the vnode first to make sure we have the most recent
* version of the size.
*/
vn_lock(vnp->vnp_vp, LK_EXCLUSIVE | LK_RETRY);
if (offset >= vnp->vnp_size) {
VOP_UNLOCK(vnp->vnp_vp, 0);
#ifdef DEBUG
if (vpagerdebug & (VDB_FAIL|VDB_SIZE))
printf("vnode_pager_haspage: pg %p, off %lx, size %lx\n",
pager, offset, vnp->vnp_size);
#endif
return(FALSE);
}
/*
* Read the index to find the disk block to read
* from. If there is no block, report that we don't
* have this data.
*
* Assumes that the vnode has whole page or nothing.
*/
err = VOP_BMAP(vnp->vnp_vp,
offset / vnp->vnp_vp->v_mount->mnt_stat.f_iosize,
(struct vnode **)0, &bn, NULL);
VOP_UNLOCK(vnp->vnp_vp, 0);
if (err) {
#ifdef DEBUG
if (vpagerdebug & VDB_FAIL)
printf("vnode_pager_haspage: BMAP err %d, pg %p, off %lx\n",
err, pager, offset);
#endif
return(TRUE);
}
return((long)bn < 0 ? FALSE : TRUE);
}
static void
vnode_pager_cluster(pager, offset, loffset, hoffset)
vm_pager_t pager;
vm_offset_t offset;
vm_offset_t *loffset;
vm_offset_t *hoffset;
{
vn_pager_t vnp = (vn_pager_t)pager->pg_data;
vm_offset_t loff, hoff;
#ifdef DEBUG
if (vpagerdebug & VDB_FOLLOW)
printf("vnode_pager_cluster(%p, %lx) ", pager, offset);
#endif
loff = offset;
if (loff >= vnp->vnp_size)
panic("vnode_pager_cluster: bad offset");
/*
* XXX could use VOP_BMAP to get maxcontig value
*/
hoff = loff + MAXBSIZE;
if (hoff > round_page(vnp->vnp_size))
hoff = round_page(vnp->vnp_size);
*loffset = loff;
*hoffset = hoff;
#ifdef DEBUG
if (vpagerdebug & VDB_FOLLOW)
printf("returns [%lx-%lx]\n", loff, hoff);
#endif
}
/*
* (XXX)
* Lets the VM system know about a change in size for a file.
* If this vnode is mapped into some address space (i.e. we have a pager
* for it) we adjust our own internal size and flush any cached pages in
* the associated object that are affected by the size change.
*
* Note: this routine may be invoked as a result of a pager put
* operation (possibly at object termination time), so we must be careful.
*/
void
vnode_pager_setsize(vp, nsize)
struct vnode *vp;
u_quad_t nsize;
{
register vn_pager_t vnp;
register vm_object_t object;
vm_pager_t pager;
/*
* Not a mapped vnode
*/
if (vp == NULL || vp->v_type != VREG || vp->v_vmdata == NULL)
return;
/* make sure mapping fits into numeric range,
truncate if necessary */
if (nsize > (vm_offset_t)-PAGE_SIZE) {
#ifdef DEBUG
printf("vnode_pager_setsize: vn %p size truncated %qx->%lx\n",
vp, nsize, (vm_offset_t)-PAGE_SIZE);
#endif
nsize = (vm_offset_t)-PAGE_SIZE;
}
/*
* Hasn't changed size
*/
pager = (vm_pager_t)vp->v_vmdata;
vnp = (vn_pager_t)pager->pg_data;
if (nsize == vnp->vnp_size)
return;
/*
* No object.
* This can happen during object termination since
* vm_object_page_clean is called after the object
* has been removed from the hash table, and clean
* may cause vnode write operations which can wind
* up back here.
*/
object = vm_object_lookup(pager);
if (object == NULL)
return;
#ifdef DEBUG
if (vpagerdebug & (VDB_FOLLOW|VDB_SIZE))
printf("vnode_pager_setsize: vp %p obj %p osz %ld nsz %qd\n",
vp, object, vnp->vnp_size, nsize);
#endif
/*
* File has shrunk.
* Toss any cached pages beyond the new EOF.
*/
if (nsize < vnp->vnp_size) {
vm_object_lock(object);
vm_object_page_remove(object,
(vm_offset_t)nsize, vnp->vnp_size);
vm_object_unlock(object);
}
vnp->vnp_size = (vm_offset_t)nsize;
vm_object_deallocate(object);
}
void
vnode_pager_umount(mp)
register struct mount *mp;
{
register vm_pager_t pager, npager;
struct vnode *vp;
for (pager = vnode_pager_list.tqh_first; pager != NULL; pager = npager){
/*
* Save the next pointer now since uncaching may
* terminate the object and render pager invalid
*/
npager = pager->pg_list.tqe_next;
vp = ((vn_pager_t)pager->pg_data)->vnp_vp;
if (mp == (struct mount *)0 || vp->v_mount == mp) {
vn_lock(vp, LK_EXCLUSIVE | LK_RETRY);
(void) vnode_pager_uncache(vp);
VOP_UNLOCK(vp, 0);
}
}
}
/*
* Flush dirty pages in all vnode_pagers.
*/
void
vnode_pager_sync(mp)
struct mount *mp;
{
vm_pager_t pager;
struct vnode *vp;
vm_object_t object, next_object;
struct object_q object_list;
lockmgr(&vnode_pager_sync_lock, LK_EXCLUSIVE, (void *)0);
/*
* We do this in two passes:
* 1) We run through the list of pagers, making a list of the objects
* they back. This also gains a reference, preventing them from
* being deleted.
* 2) We then sync each object and deallocate it.
*
* We overload `cached_list', because it's convenient, and because
* the extra reference we hold to the object will prevent it from
* being on the cache list.
*/
TAILQ_INIT(&object_list);
simple_lock(&vnode_pager_list_lock);
for (pager = vnode_pager_list.tqh_first;
pager != NULL; pager = pager->pg_list.tqe_next) {
vp = ((vn_pager_t)pager->pg_data)->vnp_vp;
if (mp == (struct mount *)0 || vp->v_mount == mp) {
object = vm_object_lookup(pager);
if (object != NULL)
TAILQ_INSERT_TAIL(&object_list, object,
cached_list);
}
}
simple_unlock(&vnode_pager_list_lock);
for (object = object_list.tqh_first; object != NULL;
object = next_object) {
next_object = object->cached_list.tqe_next;
vm_object_lock(object);
/*
* If this object is already being deleted, or lost its last
* reference while we were sleeping, don't bother to clean it
* here; vm_object_terminate() will do it.
*/
if ((object->flags & OBJ_FADING) == 0 &&
object->ref_count > 1)
(void) vm_object_page_clean(object, 0, 0, FALSE, FALSE);
vm_object_unlock(object);
vm_object_deallocate(object);
}
lockmgr(&vnode_pager_sync_lock, LK_RELEASE, (void *)0);
}
/*
* Remove vnode associated object from the object cache.
*
* XXX unlock the vnode if it is currently locked.
* We must do this since uncaching the object may result in its
* destruction which may initiate paging activity which may necessitate
* re-locking the vnode.
*/
boolean_t
vnode_pager_uncache(vp)
register struct vnode *vp;
{
vm_object_t object;
boolean_t uncached;
vm_pager_t pager;
/*
* Not a mapped vnode
*/
pager = (vm_pager_t)vp->v_vmdata;
if (pager == NULL)
return (TRUE);
#ifdef DEBUG
if (!VOP_ISLOCKED(vp)) {
#ifdef NFS
extern int (**nfsv2_vnodeop_p) __P((void *));
extern int (**spec_nfsv2nodeop_p) __P((void *));
#ifdef FIFO
extern int (**fifo_nfsv2nodeop_p) __P((void *));
#endif
if (vp->v_op != nfsv2_vnodeop_p
&& vp->v_op != spec_nfsv2nodeop_p
#ifdef FIFO
&& vp->v_op != fifo_nfsv2nodeop_p
#endif
)
#endif
panic("vnode_pager_uncache: vnode not locked!");
}
#endif
/*
* Must use vm_object_lookup() as it actually removes
* the object from the cache list.
*/
object = vm_object_lookup(pager);
if (object) {
uncached = (object->ref_count <= 1);
VOP_UNLOCK(vp, 0);
pager_cache(object, FALSE);
vn_lock(vp, LK_EXCLUSIVE | LK_RETRY);
} else
uncached = TRUE;
return(uncached);
}
static int
vnode_pager_io(vnp, mlist, npages, sync, rw)
register vn_pager_t vnp;
vm_page_t *mlist;
int npages;
boolean_t sync;
enum uio_rw rw;
{
struct uio auio;
struct iovec aiov;
vm_offset_t kva, foff;
int error, size;
struct proc *p = curproc; /* XXX */
/* XXX */
vm_page_t m;
if (npages != 1)
panic("vnode_pager_io: cannot handle multiple pages");
m = *mlist;
/* XXX */
#ifdef DEBUG
if (vpagerdebug & VDB_FOLLOW)
printf("vnode_pager_io(%p, %p, %c): vnode %p\n",
vnp, m, rw == UIO_READ ? 'R' : 'W', vnp->vnp_vp);
#endif
foff = m->offset + m->object->paging_offset;
/*
* Allocate a kernel virtual address and initialize so that
* we can use VOP_READ/WRITE routines.
*/
kva = vm_pager_map_pages(mlist, npages, sync);
if (kva == NULL)
return(VM_PAGER_AGAIN);
/*
* After all of the potentially blocking operations have been
* performed, we can do the size checks:
* read beyond EOF (returns error)
* short read
*/
vn_lock(vnp->vnp_vp, LK_EXCLUSIVE | LK_RETRY);
if (foff >= vnp->vnp_size) {
VOP_UNLOCK(vnp->vnp_vp, 0);
vm_pager_unmap_pages(kva, npages);
#ifdef DEBUG
if (vpagerdebug & VDB_SIZE)
printf("vnode_pager_io: vp %p, off %ld size %ld\n",
vnp->vnp_vp, foff, vnp->vnp_size);
#endif
return(VM_PAGER_BAD);
}
if (foff + PAGE_SIZE > vnp->vnp_size)
size = vnp->vnp_size - foff;
else
size = PAGE_SIZE;
aiov.iov_base = (caddr_t)kva;
aiov.iov_len = size;
auio.uio_iov = &aiov;
auio.uio_iovcnt = 1;
auio.uio_offset = foff;
auio.uio_segflg = UIO_SYSSPACE;
auio.uio_rw = rw;
auio.uio_resid = size;
auio.uio_procp = (struct proc *)0;
#ifdef DEBUG
if (vpagerdebug & VDB_IO)
printf("vnode_pager_io: vp %p kva %lx foff %lx size %x",
vnp->vnp_vp, kva, foff, size);
#endif
if (rw == UIO_READ)
error = VOP_READ(vnp->vnp_vp, &auio, 0, p->p_ucred);
else
error = VOP_WRITE(vnp->vnp_vp, &auio, 0, p->p_ucred);
VOP_UNLOCK(vnp->vnp_vp, 0);
#ifdef DEBUG
if (vpagerdebug & VDB_IO) {
if (error || auio.uio_resid)
printf(" returns error %x, resid %x",
error, auio.uio_resid);
printf("\n");
}
#endif
if (!error) {
register int count = size - auio.uio_resid;
if (count == 0)
error = EINVAL;
else if (count != PAGE_SIZE && rw == UIO_READ)
bzero((void *)(kva + count), PAGE_SIZE - count);
}
vm_pager_unmap_pages(kva, npages);
return (error ? VM_PAGER_ERROR : VM_PAGER_OK);
}