diff --git a/sys/arch/sparc64/include/hypervisor.h b/sys/arch/sparc64/include/hypervisor.h new file mode 100644 index 000000000000..f811c728b845 --- /dev/null +++ b/sys/arch/sparc64/include/hypervisor.h @@ -0,0 +1,320 @@ +/* $OpenBSD: hypervisor.h,v 1.14 2011/06/26 17:23:46 kettenis Exp $ */ + +/* + * Copyright (c) 2008 Mark Kettenis + * + * Permission to use, copy, modify, and distribute this software for any + * purpose with or without fee is hereby granted, provided that the above + * copyright notice and this permission notice appear in all copies. + * + * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES + * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF + * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR + * ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES + * WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN + * ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF + * OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE. + */ + +/* + * UltraSPARC Hypervisor API. + */ + +/* + * API versioning + */ + +int64_t hv_api_get_version(uint64_t api_group, + uint64_t *major_number, uint64_t *minor_number); + +/* + * Domain services + */ + +int64_t hv_mach_desc(paddr_t buffer, psize_t *length); + +/* + * CPU services + */ + +void hv_cpu_yield(void); +int64_t hv_cpu_qconf(uint64_t queue, uint64_t base, uint64_t nentries); + +#define CPU_MONDO_QUEUE 0x3c +#define DEVICE_MONDO_QUEUE 0x3d + +int64_t hv_cpu_mondo_send(uint64_t ncpus, paddr_t cpulist, paddr_t data); +int64_t hv_cpu_myid(uint64_t *cpuid); + +/* + * MMU services + */ + +int64_t hv_mmu_demap_page(vaddr_t vaddr, uint64_t context, uint64_t flags); +int64_t hv_mmu_demap_ctx(uint64_t context, uint64_t flags); +int64_t hv_mmu_demap_all(uint64_t flags); +int64_t hv_mmu_map_perm_addr(vaddr_t vaddr, uint64_t tte, uint64_t flags); +int64_t hv_mmu_unmap_perm_addr(vaddr_t vaddr, uint64_t flags); +int64_t hv_mmu_map_addr(vaddr_t vaddr, uint64_t context, uint64_t tte, + uint64_t flags); +int64_t hv_mmu_unmap_addr(vaddr_t vaddr, uint64_t context, uint64_t flags); + +#define MAP_DTLB 0x1 +#define MAP_ITLB 0x2 + +struct tsb_desc { + uint16_t td_idxpgsz; + uint16_t td_assoc; + uint32_t td_size; + uint32_t td_ctxidx; + uint32_t td_pgsz; + paddr_t td_pa; + uint64_t td_reserved; +}; + +int64_t hv_mmu_tsb_ctx0(uint64_t ntsb, paddr_t tsbptr); +int64_t hv_mmu_tsb_ctxnon0(uint64_t ntsb, paddr_t tsbptr); + +/* + * Cache and memory services + */ + +int64_t hv_mem_scrub(paddr_t raddr, psize_t length); +int64_t hv_mem_sync(paddr_t raddr, psize_t length); + +/* + * Device interrupt services + */ + +int64_t hv_intr_devino_to_sysino(uint64_t devhandle, uint64_t devino, + uint64_t *sysino); +int64_t hv_intr_getenabled(uint64_t sysino, uint64_t *intr_enabled); +int64_t hv_intr_setenabled(uint64_t sysino, uint64_t intr_enabled); +int64_t hv_intr_getstate(uint64_t sysino, uint64_t *intr_state); +int64_t hv_intr_setstate(uint64_t sysino, uint64_t intr_state); +int64_t hv_intr_gettarget(uint64_t sysino, uint64_t *cpuid); +int64_t hv_intr_settarget(uint64_t sysino, uint64_t cpuid); + +#define INTR_DISABLED 0 +#define INTR_ENABLED 1 + +#define INTR_IDLE 0 +#define INTR_RECEIVED 1 +#define INTR_DELIVERED 2 + +int64_t hv_vintr_getcookie(uint64_t devhandle, uint64_t devino, + uint64_t *cookie_value); +int64_t hv_vintr_setcookie(uint64_t devhandle, uint64_t devino, + uint64_t cookie_value); +int64_t hv_vintr_getenabled(uint64_t devhandle, uint64_t devino, + uint64_t *intr_enabled); +int64_t hv_vintr_setenabled(uint64_t devhandle, uint64_t devino, + uint64_t intr_enabled); +int64_t hv_vintr_getstate(uint64_t devhandle, uint64_t devino, + uint64_t *intr_state); +int64_t hv_vintr_setstate(uint64_t devhandle, uint64_t devino, + uint64_t intr_state); +int64_t hv_vintr_gettarget(uint64_t devhandle, uint64_t devino, + uint64_t *cpuid); +int64_t hv_vintr_settarget(uint64_t devhandle, uint64_t devino, + uint64_t cpuid); + +/* + * Time of day services + */ + +int64_t hv_tod_get(uint64_t *tod); +int64_t hv_tod_set(uint64_t tod); + +/* + * Console services + */ + +int64_t hv_cons_getchar(int64_t *ch); +int64_t hv_cons_putchar(int64_t ch); +int64_t hv_api_putchar(int64_t ch); + +#define CONS_BREAK -1 +#define CONS_HUP -2 + +/* + * Domain state services + */ + +int64_t hv_soft_state_set(uint64_t software_state, + paddr_t software_description_ptr); + +#define SIS_NORMAL 0x1 +#define SIS_TRANSITION 0x2 + +/* + * PCI I/O services + */ + +int64_t hv_pci_iommu_map(uint64_t devhandle, uint64_t tsbid, + uint64_t nttes, uint64_t io_attributes, paddr_t io_page_list_p, + uint64_t *nttes_mapped); +int64_t hv_pci_iommu_demap(uint64_t devhandle, uint64_t tsbid, + uint64_t nttes, uint64_t *nttes_demapped); +int64_t hv_pci_iommu_getmap(uint64_t devhandle, uint64_t tsbid, + uint64_t *io_attributes, paddr_t *r_addr); +int64_t hv_pci_iommu_getbypass(uint64_t devhandle, paddr_t r_addr, + uint64_t io_attributes, uint64_t *io_addr); + +int64_t hv_pci_config_get(uint64_t devhandle, uint64_t pci_device, + uint64_t pci_config_offset, uint64_t size, + uint64_t *error_flag, uint64_t *data); +int64_t hv_pci_config_put(uint64_t devhandle, uint64_t pci_device, + uint64_t pci_config_offset, uint64_t size, uint64_t data, + uint64_t *error_flag); + +#define PCI_MAP_ATTR_READ 0x01 /* From memory */ +#define PCI_MAP_ATTR_WRITE 0x02 /* To memory */ + +/* + * PCI MSI services + */ + +int64_t hv_pci_msiq_conf(uint64_t devhandle, uint64_t msiqid, + uint64_t r_addr, uint64_t nentries); +int64_t hv_pci_msiq_info(uint64_t devhandle, uint64_t msiqid, + uint64_t *r_addr, uint64_t *nentries); + +int64_t hv_pci_msiq_getvalid(uint64_t devhandle, uint64_t msiqid, + uint64_t *msiqvalid); +int64_t hv_pci_msiq_setvalid(uint64_t devhandle, uint64_t msiqid, + uint64_t msiqvalid); + +#define PCI_MSIQ_INVALID 0 +#define PCI_MSIQ_VALID 1 + +int64_t hv_pci_msiq_getstate(uint64_t devhandle, uint64_t msiqid, + uint64_t *msiqstate); +int64_t hv_pci_msiq_setstate(uint64_t devhandle, uint64_t msiqid, + uint64_t msiqstate); + +#define PCI_MSIQSTATE_IDLE 0 +#define PCI_MSIQSTATE_ERROR 1 + +int64_t hv_pci_msiq_gethead(uint64_t devhandle, uint64_t msiqid, + uint64_t *msiqhead); +int64_t hv_pci_msiq_sethead(uint64_t devhandle, uint64_t msiqid, + uint64_t msiqhead); +int64_t hv_pci_msiq_gettail(uint64_t devhandle, uint64_t msiqid, + uint64_t *msiqtail); + +int64_t hv_pci_msi_getvalid(uint64_t devhandle, uint64_t msinum, + uint64_t *msivalidstate); +int64_t hv_pci_msi_setvalid(uint64_t devhandle, uint64_t msinum, + uint64_t msivalidstate); + +#define PCI_MSI_INVALID 0 +#define PCI_MSI_VALID 1 + +int64_t hv_pci_msi_getmsiq(uint64_t devhandle, uint64_t msinum, + uint64_t *msiqid); +int64_t hv_pci_msi_setmsiq(uint64_t devhandle, uint64_t msinum, + uint64_t msitype, uint64_t msiqid); + +int64_t hv_pci_msi_getstate(uint64_t devhandle, uint64_t msinum, + uint64_t *msistate); +int64_t hv_pci_msi_setstate(uint64_t devhandle, uint64_t msinum, + uint64_t msistate); + +#define PCI_MSISTATE_IDLE 0 +#define PCI_MSISTATE_DELIVERED 1 + +int64_t hv_pci_msg_getmsiq(uint64_t devhandle, uint64_t msg, + uint64_t *msiqid); +int64_t hv_pci_msg_setmsiq(uint64_t devhandle, uint64_t msg, + uint64_t msiqid); + +int64_t hv_pci_msg_getvalid(uint64_t devhandle, uint64_t msg, + uint64_t *msgvalidstate); +int64_t hv_pci_msg_setvalid(uint64_t devhandle, uint64_t msg, + uint64_t msgvalidstate); + +#define PCIE_MSG_INVALID 0 +#define PCIE_MSG_VALID 1 + +#define PCIE_PME_MSG 0x18 +#define PCIE_PME_ACK_MSG 0x1b +#define PCIE_CORR_MSG 0x30 +#define PCIE_NONFATAL_MSG 0x31 +#define PCIE_FATAL_MSG 0x32 + +/* + * Logical Domain Channel services + */ + +int64_t hv_ldc_tx_qconf(uint64_t ldc_id, paddr_t base_raddr, + uint64_t nentries); +int64_t hv_ldc_tx_qinfo(uint64_t ldc_id, paddr_t *base_raddr, + uint64_t *nentries); +int64_t hv_ldc_tx_get_state(uint64_t ldc_id, uint64_t *head_offset, + uint64_t *tail_offset, uint64_t *channel_state); +int64_t hv_ldc_tx_set_qtail(uint64_t ldc_id, uint64_t tail_offset); +int64_t hv_ldc_rx_qconf(uint64_t ldc_id, paddr_t base_raddr, + uint64_t nentries); +int64_t hv_ldc_rx_qinfo(uint64_t ldc_id, paddr_t *base_raddr, + uint64_t *nentries); +int64_t hv_ldc_rx_get_state(uint64_t ldc_id, uint64_t *head_offset, + uint64_t *tail_offset, uint64_t *channel_state); +int64_t hv_ldc_rx_set_qhead(uint64_t ldc_id, uint64_t head_offset); + +#define LDC_CHANNEL_DOWN 0 +#define LDC_CHANNEL_UP 1 +#define LDC_CHANNEL_RESET 2 + +int64_t hv_ldc_set_map_table(uint64_t ldc_id, paddr_t base_raddr, + uint64_t nentries); +int64_t hv_ldc_get_map_table(uint64_t ldc_id, paddr_t *base_raddr, + uint64_t *nentries); +int64_t hv_ldc_copy(uint64_t ldc_id, uint64_t flags, uint64_t cookie, + paddr_t raddr, psize_t length, psize_t *ret_length); + +#define LDC_COPY_IN 0 +#define LDC_COPY_OUT 1 + +int64_t hv_ldc_mapin(uint64_t ldc_id, uint64_t cookie, paddr_t *raddr, + uint64_t *perms); +int64_t hv_ldc_unmap(paddr_t raddr, uint64_t *perms); + +/* + * Cryptographic services + */ + +int64_t hv_rng_get_diag_control(void); +int64_t hv_rng_ctl_read(paddr_t raddr, uint64_t *state, uint64_t *delta); +int64_t hv_rng_ctl_write(paddr_t raddr, uint64_t state, uint64_t timeout, + uint64_t *delta); + +#define RNG_STATE_UNCONFIGURED 0 +#define RNG_STATE_CONFIGURED 1 +#define RNG_STATE_HEALTHCHECK 2 +#define RNG_STATE_ERROR 3 + +int64_t hv_rng_data_read_diag(paddr_t raddr, uint64_t size, uint64_t *delta); +int64_t hv_rng_data_read(paddr_t raddr, uint64_t *delta); + +/* + * Error codes + */ + +#define H_EOK 0 +#define H_ENOCPU 1 +#define H_ENORADDR 2 +#define H_ENOINTR 3 +#define H_EBADPGSZ 4 +#define H_EBADTSB 5 +#define H_EINVAL 6 +#define H_EBADTRAP 7 +#define H_EBADALIGN 8 +#define H_EWOULDBLOCK 9 +#define H_ENOACCESS 10 +#define H_EIO 11 +#define H_ECPUERROR 12 +#define H_ENOTSUPPORTED 13 +#define H_ENOMAP 14 +#define H_ETOOMANY 15 +#define H_ECHANNEL 16 diff --git a/sys/arch/sparc64/sparc64/hvcall.S b/sys/arch/sparc64/sparc64/hvcall.S new file mode 100644 index 000000000000..25e5ec60b294 --- /dev/null +++ b/sys/arch/sparc64/sparc64/hvcall.S @@ -0,0 +1,691 @@ +/* $OpenBSD: hvcall.S,v 1.10 2011/06/25 20:45:00 kettenis Exp $ */ + +/* + * Copyright (c) 2008 Mark Kettenis + * + * Permission to use, copy, modify, and distribute this software for any + * purpose with or without fee is hereby granted, provided that the above + * copyright notice and this permission notice appear in all copies. + * + * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES + * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF + * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR + * ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES + * WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN + * ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF + * OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE. + */ + +#if 0 +FIXME not necessary on NetBSD? fails building if present... +#include "assym.h" +#endif +#include + +#define FAST_TRAP 0x80 +#define MMU_MAP_ADDR 0x83 +#define MMU_UNMAP_ADDR 0x84 +#define CORE_TRAP 0xff + +#define MACH_EXIT 0x00 +#define MACH_DESC 0x01 +#define MACH_SIR 0x02 +#define MACH_SET_WATCHDOG 0x05 + +#define CPU_START 0x10 +#define CPU_STOP 0x11 +#define CPU_YIELD 0x12 +#define CPU_QCONF 0x14 +#define CPU_QINFO 0x15 +#define CPU_MYID 0x16 +#define CPU_STATE 0x17 +#define CPU_SET_RTBA 0x18 +#define CPU_GET_RTBA 0x19 + +#define MMU_TSB_CTX0 0x20 +#define MMU_TSB_CTXNON0 0x21 +#define MMU_DEMAP_PAGE 0x22 +#define MMU_DEMAP_CTX 0x23 +#define MMU_DEMAP_ALL 0x24 +#define MMU_MAP_PERM_ADDR 0x25 +#define MMU_FAULT_AREA_CONF 0x26 +#define MMU_ENABLE 0x27 +#define MMU_UNMAP_PERM_ADDR 0x28 +#define MMU_TSB_CTX0_INFO 0x29 +#define MMU_TSB_CTXNON0_INFO 0x2a +#define MMU_FAULT_AREA_INFO 0x2b + +#define MEM_SCRUB 0x30 +#define MEM_SYNC 0x31 + +#define CPU_MONDO_SEND 0x42 + +#define TOD_GET 0x50 +#define TOD_SET 0x51 + +#define CONS_GETCHAR 0x60 +#define CONS_PUTCHAR 0x61 + +#define SOFT_STATE_SET 0x70 +#define SOFT_STATE_GET 0x71 + +#define INTR_DEVINO2SYSINO 0xa0 +#define INTR_GETENABLED 0xa1 +#define INTR_SETENABLED 0xa2 +#define INTR_GETSTATE 0xa3 +#define INTR_SETSTATE 0xa4 +#define INTR_GETTARGET 0xa5 +#define INTR_SETTARGET 0xa6 + +#define VINTR_GETCOOKIE 0xa7 +#define VINTR_SETCOOKIE 0xa8 +#define VINTR_GETENABLED 0xa9 +#define VINTR_SETENABLED 0xaa +#define VINTR_GETSTATE 0xab +#define VINTR_SETSTATE 0xac +#define VINTR_GETTARGET 0xad +#define VINTR_SETTARGET 0xae + +#define PCI_IOMMU_MAP 0xb0 +#define PCI_IOMMU_DEMAP 0xb1 +#define PCI_IOMMU_GETMAP 0xb2 +#define PCI_IOMMU_GETBYPASS 0xb3 +#define PCI_CONFIG_GET 0xb4 +#define PCI_CONFIG_PUT 0xb5 + +#define PCI_MSIQ_CONF 0xc0 +#define PCI_MSIQ_INFO 0xc1 +#define PCI_MSIQ_GETVALID 0xc2 +#define PCI_MSIQ_SETVALID 0xc3 +#define PCI_MSIQ_GETSTATE 0xc4 +#define PCI_MSIQ_SETSTATE 0xc5 +#define PCI_MSIQ_GETHEAD 0xc6 +#define PCI_MSIQ_SETHEAD 0xc7 +#define PCI_MSIQ_GETTAIL 0xc8 +#define PCI_MSI_GETVALID 0xc9 +#define PCI_MSI_SETVALID 0xca +#define PCI_MSI_GETMSIQ 0xcb +#define PCI_MSI_SETMSIQ 0xcc +#define PCI_MSI_GETSTATE 0xcd +#define PCI_MSI_SETSTATE 0xce +#define PCI_MSG_GETMSIQ 0xd0 +#define PCI_MSG_SETMSIQ 0xd1 +#define PCI_MSG_GETSTATE 0xd2 +#define PCI_MSG_SETSTATE 0xd3 + +#define LDC_TX_QCONF 0xe0 +#define LDC_TX_QINFO 0xe1 +#define LDC_TX_GET_STATE 0xe2 +#define LDC_TX_SET_QTAIL 0xe3 +#define LDC_RX_QCONF 0xe4 +#define LDC_RX_QINFO 0xe5 +#define LDC_RX_GET_STATE 0xe6 +#define LDC_RX_SET_QHEAD 0xe7 + +#define LDC_SET_MAP_TABLE 0xea +#define LDC_GET_MAP_TABLE 0xeb +#define LDC_COPY 0xec + +#define LDC_MAPIN 0xed +#define LDC_UNMAP 0xee +#define LDC_REVOKE 0xef + +#define RNG_GET_DIAG_CONTROL 0x130 +#define RNG_CTL_READ 0x131 +#define RNG_CTL_WRITE 0x132 +#define RNG_DATA_READ_DIAG 0x133 +#define RNG_DATA_READ 0x134 + +#define API_SET_VERSION 0x00 +#define API_PUTCHAR 0x01 +#define API_EXIT 0x02 +#define API_GET_VERSION 0x03 + + +ENTRY(hv_api_putchar) + mov API_PUTCHAR, %o5 + ta CORE_TRAP + retl + nop + +ENTRY(hv_api_get_version) + mov %o2, %o4 + mov %o1, %o3 + mov API_GET_VERSION, %o5 + ta CORE_TRAP + stx %o1, [%o3] + retl + stx %o2, [%o4] + +ENTRY(hv_mach_desc) + mov %o1, %o2 + ldx [%o2], %o1 + mov MACH_DESC, %o5 + ta FAST_TRAP + retl + stx %o1, [%o2] + +ENTRY(hv_cpu_yield) + mov CPU_YIELD, %o5 + ta FAST_TRAP + retl + nop + +ENTRY(hv_cpu_qconf) + mov CPU_QCONF, %o5 + ta FAST_TRAP + retl + nop + +ENTRY(hv_cpu_mondo_send) + mov CPU_MONDO_SEND, %o5 + ta FAST_TRAP + retl + nop + +ENTRY(hv_cpu_myid) + mov %o0, %o2 + mov CPU_MYID, %o5 + ta FAST_TRAP + retl + stx %o1, [%o2] + +ENTRY(hv_mmu_tsb_ctx0) + mov MMU_TSB_CTX0, %o5 + ta FAST_TRAP + retl + nop + +ENTRY(hv_mmu_tsb_ctxnon0) + mov MMU_TSB_CTXNON0, %o5 + ta FAST_TRAP + retl + nop + +ENTRY(hv_mmu_demap_page) + mov %o2, %o4 + mov %o1, %o3 + mov %o0, %o2 + clr %o1 + clr %o0 + mov MMU_DEMAP_PAGE, %o5 + ta FAST_TRAP + retl + nop + +ENTRY(hv_mmu_demap_ctx) + mov %o1, %o3 + mov %o0, %o2 + clr %o1 + clr %o0 + mov MMU_DEMAP_CTX, %o5 + ta FAST_TRAP + retl + nop + +ENTRY(hv_mmu_demap_all) + mov %o1, %o3 + mov %o0, %o2 + clr %o1 + clr %o0 + mov MMU_DEMAP_CTX, %o5 + ta FAST_TRAP + retl + nop + +ENTRY(hv_mmu_map_perm_addr) + mov %o2, %o3 + mov %o1, %o2 + clr %o1 + mov MMU_MAP_PERM_ADDR, %o5 + ta FAST_TRAP + retl + nop + +ENTRY(hv_mmu_unmap_perm_addr) + mov %o1, %o2 + clr %o1 + mov MMU_UNMAP_PERM_ADDR, %o5 + ta FAST_TRAP + retl + nop + +ENTRY(hv_mmu_map_addr) + ta MMU_MAP_ADDR + retl + nop + +ENTRY(hv_mmu_unmap_addr) + ta MMU_UNMAP_ADDR + retl + nop + +ENTRY(hv_mem_scrub) + mov MEM_SCRUB, %o5 + ta FAST_TRAP + retl + nop + +ENTRY(hv_mem_sync) + mov MEM_SYNC, %o5 + ta FAST_TRAP + retl + nop + +ENTRY(hv_tod_get) + mov %o0, %o2 + mov TOD_GET, %o5 + ta FAST_TRAP + retl + stx %o1, [%o2] + +ENTRY(hv_tod_set) + mov TOD_SET, %o5 + ta FAST_TRAP + retl + nop + +ENTRY(hv_cons_getchar) + mov %o0, %o2 + mov CONS_GETCHAR, %o5 + ta FAST_TRAP + retl + stx %o1, [%o2] + +ENTRY(hv_cons_putchar) + mov CONS_PUTCHAR, %o5 + ta FAST_TRAP + retl + nop + +ENTRY(hv_soft_state_set) + mov SOFT_STATE_SET, %o5 + ta FAST_TRAP + retl + nop + +ENTRY(hv_intr_devino_to_sysino) + mov INTR_DEVINO2SYSINO, %o5 + ta FAST_TRAP + retl + stx %o1, [%o2] + +ENTRY(hv_intr_getenabled) + mov %o1, %o2 + mov INTR_GETENABLED, %o5 + ta FAST_TRAP + retl + stx %o1, [%o2] + +ENTRY(hv_intr_setenabled) + mov INTR_SETENABLED, %o5 + ta FAST_TRAP + retl + nop + +ENTRY(hv_intr_getstate) + mov %o1, %o2 + mov INTR_GETSTATE, %o5 + ta FAST_TRAP + retl + stx %o1, [%o2] + +ENTRY(hv_intr_setstate) + mov INTR_SETSTATE, %o5 + ta FAST_TRAP + retl + nop + +ENTRY(hv_intr_gettarget) + mov %o1, %o2 + mov INTR_GETTARGET, %o5 + ta FAST_TRAP + retl + stx %o1, [%o2] + +ENTRY(hv_intr_settarget) + mov INTR_SETTARGET, %o5 + ta FAST_TRAP + retl + nop + +ENTRY(hv_vintr_getcookie) + mov VINTR_GETCOOKIE, %o5 + ta FAST_TRAP + retl + stx %o1, [%o2] + +ENTRY(hv_vintr_setcookie) + mov VINTR_SETCOOKIE, %o5 + ta FAST_TRAP + retl + nop + +ENTRY(hv_vintr_getenabled) + mov VINTR_GETENABLED, %o5 + ta FAST_TRAP + retl + stx %o1, [%o2] + +ENTRY(hv_vintr_setenabled) + mov VINTR_SETENABLED, %o5 + ta FAST_TRAP + retl + nop + +ENTRY(hv_vintr_getstate) + mov VINTR_GETSTATE, %o5 + ta FAST_TRAP + retl + stx %o1, [%o2] + +ENTRY(hv_vintr_setstate) + mov VINTR_SETSTATE, %o5 + ta FAST_TRAP + retl + nop + +ENTRY(hv_vintr_gettarget) + mov VINTR_GETTARGET, %o5 + ta FAST_TRAP + retl + stx %o1, [%o2] + +ENTRY(hv_vintr_settarget) + mov VINTR_SETTARGET, %o5 + ta FAST_TRAP + retl + nop + +ENTRY(hv_pci_iommu_map) + mov %o5, %g5 + mov PCI_IOMMU_MAP, %o5 + ta FAST_TRAP + retl + stx %o1, [%g5] + +ENTRY(hv_pci_iommu_demap) + mov PCI_IOMMU_DEMAP, %o5 + ta FAST_TRAP + retl + stx %o1, [%o3] + +ENTRY(hv_pci_iommu_getmap) + mov %o2, %o4 + mov PCI_IOMMU_GETMAP, %o5 + ta FAST_TRAP + stx %o1, [%o4] + retl + stx %o2, [%o3] + +ENTRY(hv_pci_iommu_getbypass) + mov PCI_IOMMU_GETBYPASS, %o5 + ta FAST_TRAP + retl + stx %o1, [%o3] + +ENTRY(hv_pci_config_get) + mov %o5, %g5 + mov PCI_CONFIG_GET, %o5 + ta FAST_TRAP + stx %o1, [%o4] + retl + stx %o2, [%g5] + +ENTRY(hv_pci_config_put) + mov %o5, %g5 + mov PCI_CONFIG_PUT, %o5 + ta FAST_TRAP + retl + stx %o1, [%g5] + +ENTRY(hv_pci_msiq_conf) + mov PCI_MSIQ_CONF, %o5 + ta FAST_TRAP + retl + nop + +ENTRY(hv_pci_msiq_info) + mov %o2, %o4 + mov PCI_MSIQ_INFO, %o5 + ta FAST_TRAP + stx %o1, [%o4] + retl + stx %o2, [%o3] + +ENTRY(hv_pci_msiq_getvalid) + mov PCI_MSIQ_GETVALID, %o5 + ta FAST_TRAP + retl + stx %o1, [%o2] + +ENTRY(hv_pci_msiq_setvalid) + mov PCI_MSIQ_SETVALID, %o5 + ta FAST_TRAP + retl + nop + +ENTRY(hv_pci_msiq_getstate) + mov PCI_MSIQ_GETSTATE, %o5 + ta FAST_TRAP + retl + stx %o1, [%o2] + +ENTRY(hv_pci_msiq_setstate) + mov PCI_MSIQ_SETSTATE, %o5 + ta FAST_TRAP + retl + nop + +ENTRY(hv_pci_msiq_gethead) + mov PCI_MSIQ_GETHEAD, %o5 + ta FAST_TRAP + retl + stx %o1, [%o2] + +ENTRY(hv_pci_msiq_sethead) + mov PCI_MSIQ_SETHEAD, %o5 + ta FAST_TRAP + retl + nop + +ENTRY(hv_pci_msiq_gettail) + mov PCI_MSIQ_GETTAIL, %o5 + ta FAST_TRAP + retl + stx %o1, [%o2] + +ENTRY(hv_pci_msi_getvalid) + mov PCI_MSI_GETVALID, %o5 + ta FAST_TRAP + retl + stx %o1, [%o2] + +ENTRY(hv_pci_msi_setvalid) + mov PCI_MSI_SETVALID, %o5 + ta FAST_TRAP + retl + nop + +ENTRY(hv_pci_msi_getmsiq) + mov PCI_MSI_GETMSIQ, %o5 + ta FAST_TRAP + retl + stx %o1, [%o2] + +ENTRY(hv_pci_msi_setmsiq) + mov PCI_MSI_SETMSIQ, %o5 + ta FAST_TRAP + retl + nop + +ENTRY(hv_pci_msi_getstate) + mov PCI_MSI_GETSTATE, %o5 + ta FAST_TRAP + retl + stx %o1, [%o2] + +ENTRY(hv_pci_msi_setstate) + mov PCI_MSI_SETSTATE, %o5 + ta FAST_TRAP + retl + nop + +ENTRY(hv_pci_msg_getmsiq) + mov PCI_MSG_GETMSIQ, %o5 + ta FAST_TRAP + retl + stx %o1, [%o2] + +ENTRY(hv_pci_msg_setmsiq) + mov PCI_MSG_SETMSIQ, %o5 + ta FAST_TRAP + retl + nop + +ENTRY(hv_pci_msg_getstate) + mov PCI_MSG_GETSTATE, %o5 + ta FAST_TRAP + retl + stx %o1, [%o2] + +ENTRY(hv_pci_msg_setstate) + mov PCI_MSG_SETSTATE, %o5 + ta FAST_TRAP + retl + nop + +ENTRY(hv_ldc_tx_qconf) + mov LDC_TX_QCONF, %o5 + ta FAST_TRAP + retl + nop + +ENTRY(hv_ldc_tx_qinfo) + mov %o2, %o4 + mov %o1, %o3 + mov LDC_TX_QINFO, %o5 + ta FAST_TRAP + stx %o1, [%o3] + retl + stx %o2, [%o4] + +ENTRY(hv_ldc_tx_get_state) + mov %o3, %g5 + mov %o2, %g4 + mov %o1, %o4 + mov LDC_TX_GET_STATE, %o5 + ta FAST_TRAP + stx %o1, [%o4] + stx %o2, [%g4] + retl + stx %o3, [%g5] + +ENTRY(hv_ldc_tx_set_qtail) + mov LDC_TX_SET_QTAIL, %o5 + ta FAST_TRAP + retl + nop + +ENTRY(hv_ldc_rx_qconf) + mov LDC_RX_QCONF, %o5 + ta FAST_TRAP + retl + nop + +ENTRY(hv_ldc_rx_qinfo) + mov %o2, %o4 + mov %o1, %o3 + mov LDC_RX_QINFO, %o5 + ta FAST_TRAP + stx %o1, [%o3] + retl + stx %o2, [%o4] + +ENTRY(hv_ldc_rx_get_state) + mov %o3, %g5 + mov %o2, %g4 + mov %o1, %o4 + mov LDC_RX_GET_STATE, %o5 + ta FAST_TRAP + stx %o1, [%o4] + stx %o2, [%g4] + retl + stx %o3, [%g5] + +ENTRY(hv_ldc_rx_set_qhead) + mov LDC_RX_SET_QHEAD, %o5 + ta FAST_TRAP + retl + nop + +ENTRY(hv_ldc_set_map_table) + mov LDC_SET_MAP_TABLE, %o5 + ta FAST_TRAP + retl + nop + +ENTRY(hv_ldc_get_map_table) + mov %o2, %o4 + mov %o1, %o3 + mov LDC_GET_MAP_TABLE, %o5 + ta FAST_TRAP + stx %o1, [%o3] + retl + stx %o2, [%o4] + +ENTRY(hv_ldc_copy) + mov %o5, %g5 + mov LDC_COPY, %o5 + ta FAST_TRAP + retl + stx %o1, [%g5] + +ENTRY(hv_ldc_mapin) + mov %o2, %o4 + mov LDC_MAPIN, %o5 + ta FAST_TRAP + stx %o1, [%o4] + retl + stx %o2, [%o3] + +ENTRY(hv_ldc_unmap) + mov %o1, %o2 + mov LDC_UNMAP, %o5 + ta FAST_TRAP + retl + stx %o1, [%o2] + +ENTRY(hv_rng_get_diag_control) + mov RNG_GET_DIAG_CONTROL, %o5 + ta FAST_TRAP + retl + nop + +ENTRY(hv_rng_ctl_read) + mov %o2, %o4 + mov %o1, %o3 + mov RNG_CTL_READ, %o5 + ta FAST_TRAP + stx %o1, [%o3] + retl + stx %o2, [%o4] + +ENTRY(hv_rng_ctl_write) + mov RNG_CTL_WRITE, %o5 + ta FAST_TRAP + retl + stx %o1, [%o3] + +ENTRY(hv_rng_data_read_diag) + mov RNG_DATA_READ_DIAG, %o5 + ta FAST_TRAP + retl + stx %o1, [%o2] + +ENTRY(hv_rng_data_read) + mov %o1, %o2 + mov RNG_DATA_READ, %o5 + ta FAST_TRAP + retl + stx %o1, [%o2]