/* $NetBSD: uvm_pager.h,v 1.6 1998/03/09 00:58:58 mrg Exp $ */ /* * XXXCDC: "ROUGH DRAFT" QUALITY UVM PRE-RELEASE FILE! * >>>USE AT YOUR OWN RISK, WORK IS NOT FINISHED<<< */ /* * * Copyright (c) 1997 Charles D. Cranor and Washington University. * All rights reserved. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions * are met: * 1. Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * 2. Redistributions in binary form must reproduce the above copyright * notice, this list of conditions and the following disclaimer in the * documentation and/or other materials provided with the distribution. * 3. All advertising materials mentioning features or use of this software * must display the following acknowledgement: * This product includes software developed by Charles D. Cranor and * Washington University. * 4. The name of the author may not be used to endorse or promote products * derived from this software without specific prior written permission. * * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. * * from: Id: uvm_pager.h,v 1.1.2.14 1998/01/13 19:00:50 chuck Exp */ #ifndef _UVM_UVM_PAGER_H_ #define _UVM_UVM_PAGER_H_ /* * uvm_pager.h */ /* * async pager i/o descriptor structure */ TAILQ_HEAD(uvm_aiohead, uvm_aiodesc); struct uvm_aiodesc { void (*aiodone) __P((struct uvm_aiodesc *)); /* aio done function */ vm_offset_t kva; /* KVA of mapped page(s) */ int npages; /* # of pages in I/O req */ void *pd_ptr; /* pager-dependent pointer */ TAILQ_ENTRY(uvm_aiodesc) aioq; /* linked list of aio's */ }; /* * pager ops */ struct uvm_pagerops { void (*pgo_init) __P((void));/* init pager */ struct uvm_object * (*pgo_attach) /* get uvm_object */ __P((void *, vm_prot_t)); void (*pgo_reference) /* add reference to obj */ __P((struct uvm_object *)); void (*pgo_detach) /* drop reference to obj */ __P((struct uvm_object *)); int (*pgo_fault) /* special nonstd fault fn */ __P((struct uvm_faultinfo *, vm_offset_t, vm_page_t *, int, int, vm_fault_t, vm_prot_t, int)); boolean_t (*pgo_flush) /* flush pages out of obj */ __P((struct uvm_object *, vm_offset_t, vm_offset_t, int)); int (*pgo_get) /* get/read page */ __P((struct uvm_object *, vm_offset_t, vm_page_t *, int *, int, vm_prot_t, int, int)); int (*pgo_asyncget) /* start async get */ __P((struct uvm_object *, vm_offset_t, int)); int (*pgo_put) /* put/write page */ __P((struct uvm_object *, vm_page_t *, int, boolean_t)); void (*pgo_cluster) /* return range of cluster */ __P((struct uvm_object *, vm_offset_t, vm_offset_t *, vm_offset_t *)); struct vm_page ** (*pgo_mk_pcluster) /* make "put" cluster */ __P((struct uvm_object *, struct vm_page **, int *, struct vm_page *, int, vm_offset_t, vm_offset_t)); void (*pgo_shareprot) /* share protect */ __P((vm_map_entry_t, vm_prot_t)); void (*pgo_aiodone) /* async iodone */ __P((struct uvm_aiodesc *)); boolean_t (*pgo_releasepg) /* release page */ __P((struct vm_page *, struct vm_page **)); }; /* pager flags [mostly for flush] */ #define PGO_CLEANIT 0x001 /* write dirty pages to backing store */ #define PGO_SYNCIO 0x002 /* if PGO_CLEAN: use sync I/O? */ /* * obviously if neither PGO_INVALIDATE or PGO_FREE are set then the pages * stay where they are. */ #define PGO_DEACTIVATE 0x004 /* deactivate flushed pages */ #define PGO_FREE 0x008 /* free flushed pages */ #define PGO_ALLPAGES 0x010 /* flush whole object/get all pages */ #define PGO_DOACTCLUST 0x020 /* flag to mk_pcluster to include active */ #define PGO_LOCKED 0x040 /* fault data structures are locked [get] */ #define PGO_PDFREECLUST 0x080 /* daemon's free cluster flag [uvm_pager_put] */ #define PGO_REALLOCSWAP 0x100 /* reallocate swap area [pager_dropcluster] */ /* page we are not interested in getting */ #define PGO_DONTCARE ((struct vm_page *) -1) /* [get only] */ /* * handle inline options */ #ifdef UVM_PAGER_INLINE #define PAGER_INLINE static __inline #else #define PAGER_INLINE /* nothing */ #endif /* UVM_PAGER_INLINE */ /* * prototypes */ void uvm_pager_dropcluster __P((struct uvm_object *, struct vm_page *, struct vm_page **, int *, int, int)); void uvm_pager_init __P((void)); int uvm_pager_put __P((struct uvm_object *, struct vm_page *, struct vm_page ***, int *, int, vm_offset_t, vm_offset_t)); PAGER_INLINE struct vm_page *uvm_pageratop __P((vm_offset_t)); vm_offset_t uvm_pagermapin __P((struct vm_page **, int, struct uvm_aiodesc **, int)); void uvm_pagermapout __P((vm_offset_t, int)); struct vm_page **uvm_mk_pcluster __P((struct uvm_object *, struct vm_page **, int *, struct vm_page *, int, vm_offset_t, vm_offset_t)); void uvm_shareprot __P((vm_map_entry_t, vm_prot_t)); #endif /* _UVM_UVM_PAGER_H_ */