/* $NetBSD: xenpmap.h,v 1.3 2004/04/26 22:05:05 cl Exp $ */ /* * * Copyright (c) 2004 Christian Limpach. * All rights reserved. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions * are met: * 1. Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * 2. Redistributions in binary form must reproduce the above copyright * notice, this list of conditions and the following disclaimer in the * documentation and/or other materials provided with the distribution. * 3. All advertising materials mentioning features or use of this software * must display the following acknowledgement: * This product includes software developed by Christian Limpach. * 4. The name of the author may not be used to endorse or promote products * derived from this software without specific prior written permission. * * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. */ #ifndef _XEN_XENPMAP_H_ #define _XEN_XENPMAP_H_ void xpq_queue_invlpg(vaddr_t); void xpq_queue_pde_update(pd_entry_t *, pd_entry_t); void xpq_queue_pte_update(pt_entry_t *, pt_entry_t); void xpq_queue_unchecked_pte_update(pt_entry_t *, pt_entry_t); void xpq_queue_pt_switch(paddr_t); void xpq_flush_queue(void); void xpq_queue_set_ldt(vaddr_t, uint32_t); void xpq_queue_tlb_flush(void); void xpq_queue_pin_table(paddr_t, int); void xpq_queue_unpin_table(paddr_t); extern paddr_t *xpmap_phys_to_machine_mapping; #define XPQ_PIN_L1_TABLE 1 #define XPQ_PIN_L2_TABLE 2 #ifndef XEN #define PDE_GET(_pdp) \ *(_pdp) #define PDE_SET(_pdp,_npde) \ *(_pdp) = (_npde) #define PDE_CLEAR(_pdp) \ *(_pdp) = 0 #define PTE_SET(_ptp,_npte) \ *(_ptp) = (_npte) #define PTE_CLEAR(_ptp) \ *(_ptp) = 0 #define PTE_ATOMIC_SET(_ptp,_npte,_opte) \ (_opte) = x86_atomic_testset_ul((_ptp), (_npte)) #define PTE_ATOMIC_CLEAR(_ptp,_opte) \ (_opte) = x86_atomic_testset_ul((_ptp), 0) #define PDE_CLEARBITS(_pdp,_bits) \ *(_pdp) &= ~(_bits) #define PTE_ATOMIC_CLEARBITS(_ptp,_bits) \ x86_atomic_clearbits_l((_ptp), (_bits)) #define PTE_SETBITS(_ptp,_bits) \ *(_ptp) |= (_bits) #define PTE_ATOMIC_SETBITS(_ptp,_bits) \ x86_atomic_setbits_l((_ptp), (_bits)) #else paddr_t *xpmap_phys_to_machine_mapping; #define PDE_GET(_pdp) \ (pmap_valid_entry(*(_pdp)) ? xpmap_mtop(*(_pdp)) : *(_pdp)) #define PDE_SET(_pdp,_npde) do { \ xpq_queue_pde_update((_pdp), xpmap_ptom((_npde))); \ xpq_flush_queue(); \ } while (/*CONSTCOND*/0) #define PDE_CLEAR(_pdp) do { \ xpq_queue_pde_update((_pdp), 0); \ xpq_flush_queue(); \ } while (/*CONSTCOND*/0) #define PTE_GET(_ptp) \ (pmap_valid_entry(*(_ptp)) ? xpmap_mtop(*(_ptp)) : *(_ptp)) #define PTE_GET_MA(_ptp) \ *(_ptp) #define PTE_SET(_ptp,_npte) do { \ xpq_queue_pte_update((_ptp), xpmap_ptom((_npte))); \ xpq_flush_queue(); \ } while (/*CONSTCOND*/0) #define PTE_SET_MA(_ptp,_npte) do { \ xpq_queue_pte_update((_ptp), (_npte)); \ xpq_flush_queue(); \ } while (/*CONSTCOND*/0) #define PTE_SET_MA_UNCHECKED(_ptp,_npte) do { \ xpq_queue_unchecked_pte_update((_ptp), (_npte)); \ xpq_flush_queue(); \ } while (/*CONSTCOND*/0) #define PTE_CLEAR(_ptp) do { \ xpq_queue_pte_update((_ptp), 0); \ xpq_flush_queue(); \ } while (/*CONSTCOND*/0) #define PTE_ATOMIC_SET(_ptp,_npte,_opte) do { \ (_opte) = PTE_GET(_ptp); \ xpq_queue_pte_update((_ptp), xpmap_ptom((_npte))); \ xpq_flush_queue(); \ } while (/*CONSTCOND*/0) #define PTE_ATOMIC_SET_MA(_ptp,_npte,_opte) do { \ (_opte) = *(_ptp); \ xpq_queue_pte_update((_ptp), (_npte)); \ xpq_flush_queue(); \ } while (/*CONSTCOND*/0) #define PTE_ATOMIC_CLEAR(_ptp,_opte) do { \ (_opte) = PTE_GET(_ptp); \ xpq_queue_pte_update((_ptp), 0); \ xpq_flush_queue(); \ } while (/*CONSTCOND*/0) #define PTE_ATOMIC_CLEAR_MA(_ptp,_opte) do { \ (_opte) = *(_ptp); \ xpq_queue_pte_update((_ptp), 0); \ xpq_flush_queue(); \ } while (/*CONSTCOND*/0) #define PDE_CLEARBITS(_pdp,_bits) do { \ xpq_queue_pte_update((_pdp), *(_pdp) & ~((_bits) & ~PG_FRAME)); \ xpq_flush_queue(); \ } while (/*CONSTCOND*/0) #define PTE_CLEARBITS(_ptp,_bits) do { \ xpq_queue_pte_update((_ptp), *(_ptp) & ~((_bits) & ~PG_FRAME)); \ xpq_flush_queue(); \ } while (/*CONSTCOND*/0) #define PDE_ATOMIC_CLEARBITS(_pdp,_bits) do { \ xpq_queue_pde_update((_pdp), *(_pdp) & ~((_bits) & ~PG_FRAME)); \ xpq_flush_queue(); \ } while (/*CONSTCOND*/0) #define PTE_ATOMIC_CLEARBITS(_ptp,_bits) do { \ xpq_queue_pte_update((_ptp), *(_ptp) & ~((_bits) & ~PG_FRAME)); \ xpq_flush_queue(); \ } while (/*CONSTCOND*/0) #define PTE_SETBITS(_ptp,_bits) do { \ xpq_queue_pte_update((_ptp), *(_ptp) | ((_bits) & ~PG_FRAME)); \ xpq_flush_queue(); \ } while (/*CONSTCOND*/0) #define PDE_ATOMIC_SETBITS(_pdp,_bits) do { \ xpq_queue_pde_update((_pdp), *(_pdp) | ((_bits) & ~PG_FRAME)); \ xpq_flush_queue(); \ } while (/*CONSTCOND*/0) #define PTE_ATOMIC_SETBITS(_ptp,_bits) do { \ xpq_queue_pde_update((_ptp), *(_ptp) | ((_bits) & ~PG_FRAME)); \ xpq_flush_queue(); \ } while (/*CONSTCOND*/0) #define PDE_COPY(_dpdp,_spdp) do { \ xpq_queue_pde_update((_dpdp), *(_spdp)); \ xpq_flush_queue(); \ } while (/*CONSTCOND*/0) #define PTE_UPDATES_FLUSH() do { \ xpq_flush_queue(); \ } while (/*CONSTCOND*/0) #endif static __inline paddr_t xpmap_mtop(paddr_t mpa) { return ((machine_to_phys_mapping[mpa >> PAGE_SHIFT] << PAGE_SHIFT) + (KERNTEXTOFF - KERNBASE_LOCORE)) | (mpa & ~PG_FRAME); } static __inline paddr_t xpmap_ptom(paddr_t ppa) { return (xpmap_phys_to_machine_mapping[(ppa - (KERNTEXTOFF - KERNBASE_LOCORE)) >> PAGE_SHIFT] << PAGE_SHIFT) | (ppa & ~PG_FRAME); } static __inline paddr_t xpmap_ptom_masked(paddr_t ppa) { return (xpmap_phys_to_machine_mapping[(ppa - (KERNTEXTOFF - KERNBASE_LOCORE)) >> PAGE_SHIFT] << PAGE_SHIFT); } #endif /* _XEN_XENPMAP_H_ */