50c88402ca
There's already a partial set here; update them and pull in a more complete set. To start with, define __XEN_TOOLS__ in hw/xen/xen.h to ensure that any internal definitions needed by Xen toolstack libraries are present regardless of the order in which the headers are included. A reckoning will come later, once we make the PV backends work in emulation and untangle the headers for Xen-native vs. generic parts. Signed-off-by: Joao Martins <joao.m.martins@oracle.com> [dwmw2: Update to Xen public headers from 4.16.2 release, add some in io/, define __XEN_TOOLS__ in hw/xen/xen.h, move to hw/xen/interface/] Signed-off-by: David Woodhouse <dwmw@amazon.co.uk> Reviewed-by: Paul Durrant <paul@xen.org>
755 lines
26 KiB
C
755 lines
26 KiB
C
/******************************************************************************
|
|
* memory.h
|
|
*
|
|
* Memory reservation and information.
|
|
*
|
|
* Permission is hereby granted, free of charge, to any person obtaining a copy
|
|
* of this software and associated documentation files (the "Software"), to
|
|
* deal in the Software without restriction, including without limitation the
|
|
* rights to use, copy, modify, merge, publish, distribute, sublicense, and/or
|
|
* sell copies of the Software, and to permit persons to whom the Software is
|
|
* furnished to do so, subject to the following conditions:
|
|
*
|
|
* The above copyright notice and this permission notice shall be included in
|
|
* all copies or substantial portions of the Software.
|
|
*
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
|
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
|
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
|
|
* AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
|
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
|
|
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
|
|
* DEALINGS IN THE SOFTWARE.
|
|
*
|
|
* Copyright (c) 2005, Keir Fraser <keir@xensource.com>
|
|
*/
|
|
|
|
#ifndef __XEN_PUBLIC_MEMORY_H__
|
|
#define __XEN_PUBLIC_MEMORY_H__
|
|
|
|
#include "xen.h"
|
|
#include "physdev.h"
|
|
|
|
/*
|
|
* Increase or decrease the specified domain's memory reservation. Returns the
|
|
* number of extents successfully allocated or freed.
|
|
* arg == addr of struct xen_memory_reservation.
|
|
*/
|
|
#define XENMEM_increase_reservation 0
|
|
#define XENMEM_decrease_reservation 1
|
|
#define XENMEM_populate_physmap 6
|
|
|
|
#if __XEN_INTERFACE_VERSION__ >= 0x00030209
|
|
/*
|
|
* Maximum # bits addressable by the user of the allocated region (e.g., I/O
|
|
* devices often have a 32-bit limitation even in 64-bit systems). If zero
|
|
* then the user has no addressing restriction. This field is not used by
|
|
* XENMEM_decrease_reservation.
|
|
*/
|
|
#define XENMEMF_address_bits(x) (x)
|
|
#define XENMEMF_get_address_bits(x) ((x) & 0xffu)
|
|
/* NUMA node to allocate from. */
|
|
#define XENMEMF_node(x) (((x) + 1) << 8)
|
|
#define XENMEMF_get_node(x) ((((x) >> 8) - 1) & 0xffu)
|
|
/* Flag to populate physmap with populate-on-demand entries */
|
|
#define XENMEMF_populate_on_demand (1<<16)
|
|
/* Flag to request allocation only from the node specified */
|
|
#define XENMEMF_exact_node_request (1<<17)
|
|
#define XENMEMF_exact_node(n) (XENMEMF_node(n) | XENMEMF_exact_node_request)
|
|
/* Flag to indicate the node specified is virtual node */
|
|
#define XENMEMF_vnode (1<<18)
|
|
#endif
|
|
|
|
struct xen_memory_reservation {
|
|
|
|
/*
|
|
* XENMEM_increase_reservation:
|
|
* OUT: MFN (*not* GMFN) bases of extents that were allocated
|
|
* XENMEM_decrease_reservation:
|
|
* IN: GMFN bases of extents to free
|
|
* XENMEM_populate_physmap:
|
|
* IN: GPFN bases of extents to populate with memory
|
|
* OUT: GMFN bases of extents that were allocated
|
|
* (NB. This command also updates the mach_to_phys translation table)
|
|
* XENMEM_claim_pages:
|
|
* IN: must be zero
|
|
*/
|
|
XEN_GUEST_HANDLE(xen_pfn_t) extent_start;
|
|
|
|
/* Number of extents, and size/alignment of each (2^extent_order pages). */
|
|
xen_ulong_t nr_extents;
|
|
unsigned int extent_order;
|
|
|
|
#if __XEN_INTERFACE_VERSION__ >= 0x00030209
|
|
/* XENMEMF flags. */
|
|
unsigned int mem_flags;
|
|
#else
|
|
unsigned int address_bits;
|
|
#endif
|
|
|
|
/*
|
|
* Domain whose reservation is being changed.
|
|
* Unprivileged domains can specify only DOMID_SELF.
|
|
*/
|
|
domid_t domid;
|
|
};
|
|
typedef struct xen_memory_reservation xen_memory_reservation_t;
|
|
DEFINE_XEN_GUEST_HANDLE(xen_memory_reservation_t);
|
|
|
|
/*
|
|
* An atomic exchange of memory pages. If return code is zero then
|
|
* @out.extent_list provides GMFNs of the newly-allocated memory.
|
|
* Returns zero on complete success, otherwise a negative error code.
|
|
* On complete success then always @nr_exchanged == @in.nr_extents.
|
|
* On partial success @nr_exchanged indicates how much work was done.
|
|
*
|
|
* Note that only PV guests can use this operation.
|
|
*/
|
|
#define XENMEM_exchange 11
|
|
struct xen_memory_exchange {
|
|
/*
|
|
* [IN] Details of memory extents to be exchanged (GMFN bases).
|
|
* Note that @in.address_bits is ignored and unused.
|
|
*/
|
|
struct xen_memory_reservation in;
|
|
|
|
/*
|
|
* [IN/OUT] Details of new memory extents.
|
|
* We require that:
|
|
* 1. @in.domid == @out.domid
|
|
* 2. @in.nr_extents << @in.extent_order ==
|
|
* @out.nr_extents << @out.extent_order
|
|
* 3. @in.extent_start and @out.extent_start lists must not overlap
|
|
* 4. @out.extent_start lists GPFN bases to be populated
|
|
* 5. @out.extent_start is overwritten with allocated GMFN bases
|
|
*/
|
|
struct xen_memory_reservation out;
|
|
|
|
/*
|
|
* [OUT] Number of input extents that were successfully exchanged:
|
|
* 1. The first @nr_exchanged input extents were successfully
|
|
* deallocated.
|
|
* 2. The corresponding first entries in the output extent list correctly
|
|
* indicate the GMFNs that were successfully exchanged.
|
|
* 3. All other input and output extents are untouched.
|
|
* 4. If not all input exents are exchanged then the return code of this
|
|
* command will be non-zero.
|
|
* 5. THIS FIELD MUST BE INITIALISED TO ZERO BY THE CALLER!
|
|
*/
|
|
xen_ulong_t nr_exchanged;
|
|
};
|
|
typedef struct xen_memory_exchange xen_memory_exchange_t;
|
|
DEFINE_XEN_GUEST_HANDLE(xen_memory_exchange_t);
|
|
|
|
/*
|
|
* Returns the maximum machine frame number of mapped RAM in this system.
|
|
* This command always succeeds (it never returns an error code).
|
|
* arg == NULL.
|
|
*/
|
|
#define XENMEM_maximum_ram_page 2
|
|
|
|
struct xen_memory_domain {
|
|
/* [IN] Domain information is being queried for. */
|
|
domid_t domid;
|
|
};
|
|
|
|
/*
|
|
* Returns the current or maximum memory reservation, in pages, of the
|
|
* specified domain (may be DOMID_SELF). Returns -ve errcode on failure.
|
|
* arg == addr of struct xen_memory_domain.
|
|
*/
|
|
#define XENMEM_current_reservation 3
|
|
#define XENMEM_maximum_reservation 4
|
|
|
|
/*
|
|
* Returns the maximum GFN in use by the specified domain (may be DOMID_SELF).
|
|
* Returns -ve errcode on failure.
|
|
* arg == addr of struct xen_memory_domain.
|
|
*/
|
|
#define XENMEM_maximum_gpfn 14
|
|
|
|
/*
|
|
* Returns a list of MFN bases of 2MB extents comprising the machine_to_phys
|
|
* mapping table. Architectures which do not have a m2p table do not implement
|
|
* this command.
|
|
* arg == addr of xen_machphys_mfn_list_t.
|
|
*/
|
|
#define XENMEM_machphys_mfn_list 5
|
|
struct xen_machphys_mfn_list {
|
|
/*
|
|
* Size of the 'extent_start' array. Fewer entries will be filled if the
|
|
* machphys table is smaller than max_extents * 2MB.
|
|
*/
|
|
unsigned int max_extents;
|
|
|
|
/*
|
|
* Pointer to buffer to fill with list of extent starts. If there are
|
|
* any large discontiguities in the machine address space, 2MB gaps in
|
|
* the machphys table will be represented by an MFN base of zero.
|
|
*/
|
|
XEN_GUEST_HANDLE(xen_pfn_t) extent_start;
|
|
|
|
/*
|
|
* Number of extents written to the above array. This will be smaller
|
|
* than 'max_extents' if the machphys table is smaller than max_e * 2MB.
|
|
*/
|
|
unsigned int nr_extents;
|
|
};
|
|
typedef struct xen_machphys_mfn_list xen_machphys_mfn_list_t;
|
|
DEFINE_XEN_GUEST_HANDLE(xen_machphys_mfn_list_t);
|
|
|
|
/*
|
|
* For a compat caller, this is identical to XENMEM_machphys_mfn_list.
|
|
*
|
|
* For a non compat caller, this functions similarly to
|
|
* XENMEM_machphys_mfn_list, but returns the mfns making up the compatibility
|
|
* m2p table.
|
|
*/
|
|
#define XENMEM_machphys_compat_mfn_list 25
|
|
|
|
/*
|
|
* Returns the location in virtual address space of the machine_to_phys
|
|
* mapping table. Architectures which do not have a m2p table, or which do not
|
|
* map it by default into guest address space, do not implement this command.
|
|
* arg == addr of xen_machphys_mapping_t.
|
|
*/
|
|
#define XENMEM_machphys_mapping 12
|
|
struct xen_machphys_mapping {
|
|
xen_ulong_t v_start, v_end; /* Start and end virtual addresses. */
|
|
xen_ulong_t max_mfn; /* Maximum MFN that can be looked up. */
|
|
};
|
|
typedef struct xen_machphys_mapping xen_machphys_mapping_t;
|
|
DEFINE_XEN_GUEST_HANDLE(xen_machphys_mapping_t);
|
|
|
|
/* Source mapping space. */
|
|
/* ` enum phys_map_space { */
|
|
#define XENMAPSPACE_shared_info 0 /* shared info page */
|
|
#define XENMAPSPACE_grant_table 1 /* grant table page */
|
|
#define XENMAPSPACE_gmfn 2 /* GMFN */
|
|
#define XENMAPSPACE_gmfn_range 3 /* GMFN range, XENMEM_add_to_physmap only. */
|
|
#define XENMAPSPACE_gmfn_foreign 4 /* GMFN from another dom,
|
|
* XENMEM_add_to_physmap_batch only. */
|
|
#define XENMAPSPACE_dev_mmio 5 /* device mmio region
|
|
ARM only; the region is mapped in
|
|
Stage-2 using the Normal Memory
|
|
Inner/Outer Write-Back Cacheable
|
|
memory attribute. */
|
|
/* ` } */
|
|
|
|
/*
|
|
* Sets the GPFN at which a particular page appears in the specified guest's
|
|
* physical address space (translated guests only).
|
|
* arg == addr of xen_add_to_physmap_t.
|
|
*/
|
|
#define XENMEM_add_to_physmap 7
|
|
struct xen_add_to_physmap {
|
|
/* Which domain to change the mapping for. */
|
|
domid_t domid;
|
|
|
|
/* Number of pages to go through for gmfn_range */
|
|
uint16_t size;
|
|
|
|
unsigned int space; /* => enum phys_map_space */
|
|
|
|
#define XENMAPIDX_grant_table_status 0x80000000
|
|
|
|
/* Index into space being mapped. */
|
|
xen_ulong_t idx;
|
|
|
|
/* GPFN in domid where the source mapping page should appear. */
|
|
xen_pfn_t gpfn;
|
|
};
|
|
typedef struct xen_add_to_physmap xen_add_to_physmap_t;
|
|
DEFINE_XEN_GUEST_HANDLE(xen_add_to_physmap_t);
|
|
|
|
/* A batched version of add_to_physmap. */
|
|
#define XENMEM_add_to_physmap_batch 23
|
|
struct xen_add_to_physmap_batch {
|
|
/* IN */
|
|
/* Which domain to change the mapping for. */
|
|
domid_t domid;
|
|
uint16_t space; /* => enum phys_map_space */
|
|
|
|
/* Number of pages to go through */
|
|
uint16_t size;
|
|
|
|
#if __XEN_INTERFACE_VERSION__ < 0x00040700
|
|
domid_t foreign_domid; /* IFF gmfn_foreign. Should be 0 for other spaces. */
|
|
#else
|
|
union xen_add_to_physmap_batch_extra {
|
|
domid_t foreign_domid; /* gmfn_foreign */
|
|
uint16_t res0; /* All the other spaces. Should be 0 */
|
|
} u;
|
|
#endif
|
|
|
|
/* Indexes into space being mapped. */
|
|
XEN_GUEST_HANDLE(xen_ulong_t) idxs;
|
|
|
|
/* GPFN in domid where the source mapping page should appear. */
|
|
XEN_GUEST_HANDLE(xen_pfn_t) gpfns;
|
|
|
|
/* OUT */
|
|
|
|
/* Per index error code. */
|
|
XEN_GUEST_HANDLE(int) errs;
|
|
};
|
|
typedef struct xen_add_to_physmap_batch xen_add_to_physmap_batch_t;
|
|
DEFINE_XEN_GUEST_HANDLE(xen_add_to_physmap_batch_t);
|
|
|
|
#if __XEN_INTERFACE_VERSION__ < 0x00040400
|
|
#define XENMEM_add_to_physmap_range XENMEM_add_to_physmap_batch
|
|
#define xen_add_to_physmap_range xen_add_to_physmap_batch
|
|
typedef struct xen_add_to_physmap_batch xen_add_to_physmap_range_t;
|
|
DEFINE_XEN_GUEST_HANDLE(xen_add_to_physmap_range_t);
|
|
#endif
|
|
|
|
/*
|
|
* Unmaps the page appearing at a particular GPFN from the specified guest's
|
|
* physical address space (translated guests only).
|
|
* arg == addr of xen_remove_from_physmap_t.
|
|
*/
|
|
#define XENMEM_remove_from_physmap 15
|
|
struct xen_remove_from_physmap {
|
|
/* Which domain to change the mapping for. */
|
|
domid_t domid;
|
|
|
|
/* GPFN of the current mapping of the page. */
|
|
xen_pfn_t gpfn;
|
|
};
|
|
typedef struct xen_remove_from_physmap xen_remove_from_physmap_t;
|
|
DEFINE_XEN_GUEST_HANDLE(xen_remove_from_physmap_t);
|
|
|
|
/*** REMOVED ***/
|
|
/*#define XENMEM_translate_gpfn_list 8*/
|
|
|
|
/*
|
|
* Returns the pseudo-physical memory map as it was when the domain
|
|
* was started (specified by XENMEM_set_memory_map).
|
|
* arg == addr of xen_memory_map_t.
|
|
*/
|
|
#define XENMEM_memory_map 9
|
|
struct xen_memory_map {
|
|
/*
|
|
* On call the number of entries which can be stored in buffer. On
|
|
* return the number of entries which have been stored in
|
|
* buffer.
|
|
*/
|
|
unsigned int nr_entries;
|
|
|
|
/*
|
|
* Entries in the buffer are in the same format as returned by the
|
|
* BIOS INT 0x15 EAX=0xE820 call.
|
|
*/
|
|
XEN_GUEST_HANDLE(void) buffer;
|
|
};
|
|
typedef struct xen_memory_map xen_memory_map_t;
|
|
DEFINE_XEN_GUEST_HANDLE(xen_memory_map_t);
|
|
|
|
/*
|
|
* Returns the real physical memory map. Passes the same structure as
|
|
* XENMEM_memory_map.
|
|
* Specifying buffer as NULL will return the number of entries required
|
|
* to store the complete memory map.
|
|
* arg == addr of xen_memory_map_t.
|
|
*/
|
|
#define XENMEM_machine_memory_map 10
|
|
|
|
/*
|
|
* Set the pseudo-physical memory map of a domain, as returned by
|
|
* XENMEM_memory_map.
|
|
* arg == addr of xen_foreign_memory_map_t.
|
|
*/
|
|
#define XENMEM_set_memory_map 13
|
|
struct xen_foreign_memory_map {
|
|
domid_t domid;
|
|
struct xen_memory_map map;
|
|
};
|
|
typedef struct xen_foreign_memory_map xen_foreign_memory_map_t;
|
|
DEFINE_XEN_GUEST_HANDLE(xen_foreign_memory_map_t);
|
|
|
|
#define XENMEM_set_pod_target 16
|
|
#define XENMEM_get_pod_target 17
|
|
struct xen_pod_target {
|
|
/* IN */
|
|
uint64_t target_pages;
|
|
/* OUT */
|
|
uint64_t tot_pages;
|
|
uint64_t pod_cache_pages;
|
|
uint64_t pod_entries;
|
|
/* IN */
|
|
domid_t domid;
|
|
};
|
|
typedef struct xen_pod_target xen_pod_target_t;
|
|
|
|
#if defined(__XEN__) || defined(__XEN_TOOLS__)
|
|
|
|
#ifndef uint64_aligned_t
|
|
#define uint64_aligned_t uint64_t
|
|
#endif
|
|
|
|
/*
|
|
* Get the number of MFNs saved through memory sharing.
|
|
* The call never fails.
|
|
*/
|
|
#define XENMEM_get_sharing_freed_pages 18
|
|
#define XENMEM_get_sharing_shared_pages 19
|
|
|
|
#define XENMEM_paging_op 20
|
|
#define XENMEM_paging_op_nominate 0
|
|
#define XENMEM_paging_op_evict 1
|
|
#define XENMEM_paging_op_prep 2
|
|
|
|
struct xen_mem_paging_op {
|
|
uint8_t op; /* XENMEM_paging_op_* */
|
|
domid_t domain;
|
|
|
|
/* IN: (XENMEM_paging_op_prep) buffer to immediately fill page from */
|
|
XEN_GUEST_HANDLE_64(const_uint8) buffer;
|
|
/* IN: gfn of page being operated on */
|
|
uint64_aligned_t gfn;
|
|
};
|
|
typedef struct xen_mem_paging_op xen_mem_paging_op_t;
|
|
DEFINE_XEN_GUEST_HANDLE(xen_mem_paging_op_t);
|
|
|
|
#define XENMEM_access_op 21
|
|
#define XENMEM_access_op_set_access 0
|
|
#define XENMEM_access_op_get_access 1
|
|
/*
|
|
* XENMEM_access_op_enable_emulate and XENMEM_access_op_disable_emulate are
|
|
* currently unused, but since they have been in use please do not reuse them.
|
|
*
|
|
* #define XENMEM_access_op_enable_emulate 2
|
|
* #define XENMEM_access_op_disable_emulate 3
|
|
*/
|
|
#define XENMEM_access_op_set_access_multi 4
|
|
|
|
typedef enum {
|
|
XENMEM_access_n,
|
|
XENMEM_access_r,
|
|
XENMEM_access_w,
|
|
XENMEM_access_rw,
|
|
XENMEM_access_x,
|
|
XENMEM_access_rx,
|
|
XENMEM_access_wx,
|
|
XENMEM_access_rwx,
|
|
/*
|
|
* Page starts off as r-x, but automatically
|
|
* change to r-w on a write
|
|
*/
|
|
XENMEM_access_rx2rw,
|
|
/*
|
|
* Log access: starts off as n, automatically
|
|
* goes to rwx, generating an event without
|
|
* pausing the vcpu
|
|
*/
|
|
XENMEM_access_n2rwx,
|
|
/* Take the domain default */
|
|
XENMEM_access_default
|
|
} xenmem_access_t;
|
|
|
|
struct xen_mem_access_op {
|
|
/* XENMEM_access_op_* */
|
|
uint8_t op;
|
|
/* xenmem_access_t */
|
|
uint8_t access;
|
|
domid_t domid;
|
|
/*
|
|
* Number of pages for set op (or size of pfn_list for
|
|
* XENMEM_access_op_set_access_multi)
|
|
* Ignored on setting default access and other ops
|
|
*/
|
|
uint32_t nr;
|
|
/*
|
|
* First pfn for set op
|
|
* pfn for get op
|
|
* ~0ull is used to set and get the default access for pages
|
|
*/
|
|
uint64_aligned_t pfn;
|
|
/*
|
|
* List of pfns to set access for
|
|
* Used only with XENMEM_access_op_set_access_multi
|
|
*/
|
|
XEN_GUEST_HANDLE(const_uint64) pfn_list;
|
|
/*
|
|
* Corresponding list of access settings for pfn_list
|
|
* Used only with XENMEM_access_op_set_access_multi
|
|
*/
|
|
XEN_GUEST_HANDLE(const_uint8) access_list;
|
|
};
|
|
typedef struct xen_mem_access_op xen_mem_access_op_t;
|
|
DEFINE_XEN_GUEST_HANDLE(xen_mem_access_op_t);
|
|
|
|
#define XENMEM_sharing_op 22
|
|
#define XENMEM_sharing_op_nominate_gfn 0
|
|
#define XENMEM_sharing_op_nominate_gref 1
|
|
#define XENMEM_sharing_op_share 2
|
|
#define XENMEM_sharing_op_debug_gfn 3
|
|
#define XENMEM_sharing_op_debug_mfn 4
|
|
#define XENMEM_sharing_op_debug_gref 5
|
|
#define XENMEM_sharing_op_add_physmap 6
|
|
#define XENMEM_sharing_op_audit 7
|
|
#define XENMEM_sharing_op_range_share 8
|
|
#define XENMEM_sharing_op_fork 9
|
|
#define XENMEM_sharing_op_fork_reset 10
|
|
|
|
#define XENMEM_SHARING_OP_S_HANDLE_INVALID (-10)
|
|
#define XENMEM_SHARING_OP_C_HANDLE_INVALID (-9)
|
|
|
|
/* The following allows sharing of grant refs. This is useful
|
|
* for sharing utilities sitting as "filters" in IO backends
|
|
* (e.g. memshr + blktap(2)). The IO backend is only exposed
|
|
* to grant references, and this allows sharing of the grefs */
|
|
#define XENMEM_SHARING_OP_FIELD_IS_GREF_FLAG (xen_mk_ullong(1) << 62)
|
|
|
|
#define XENMEM_SHARING_OP_FIELD_MAKE_GREF(field, val) \
|
|
(field) = (XENMEM_SHARING_OP_FIELD_IS_GREF_FLAG | val)
|
|
#define XENMEM_SHARING_OP_FIELD_IS_GREF(field) \
|
|
((field) & XENMEM_SHARING_OP_FIELD_IS_GREF_FLAG)
|
|
#define XENMEM_SHARING_OP_FIELD_GET_GREF(field) \
|
|
((field) & (~XENMEM_SHARING_OP_FIELD_IS_GREF_FLAG))
|
|
|
|
struct xen_mem_sharing_op {
|
|
uint8_t op; /* XENMEM_sharing_op_* */
|
|
domid_t domain;
|
|
|
|
union {
|
|
struct mem_sharing_op_nominate { /* OP_NOMINATE_xxx */
|
|
union {
|
|
uint64_aligned_t gfn; /* IN: gfn to nominate */
|
|
uint32_t grant_ref; /* IN: grant ref to nominate */
|
|
} u;
|
|
uint64_aligned_t handle; /* OUT: the handle */
|
|
} nominate;
|
|
struct mem_sharing_op_share { /* OP_SHARE/ADD_PHYSMAP */
|
|
uint64_aligned_t source_gfn; /* IN: the gfn of the source page */
|
|
uint64_aligned_t source_handle; /* IN: handle to the source page */
|
|
uint64_aligned_t client_gfn; /* IN: the client gfn */
|
|
uint64_aligned_t client_handle; /* IN: handle to the client page */
|
|
domid_t client_domain; /* IN: the client domain id */
|
|
} share;
|
|
struct mem_sharing_op_range { /* OP_RANGE_SHARE */
|
|
uint64_aligned_t first_gfn; /* IN: the first gfn */
|
|
uint64_aligned_t last_gfn; /* IN: the last gfn */
|
|
uint64_aligned_t opaque; /* Must be set to 0 */
|
|
domid_t client_domain; /* IN: the client domain id */
|
|
uint16_t _pad[3]; /* Must be set to 0 */
|
|
} range;
|
|
struct mem_sharing_op_debug { /* OP_DEBUG_xxx */
|
|
union {
|
|
uint64_aligned_t gfn; /* IN: gfn to debug */
|
|
uint64_aligned_t mfn; /* IN: mfn to debug */
|
|
uint32_t gref; /* IN: gref to debug */
|
|
} u;
|
|
} debug;
|
|
struct mem_sharing_op_fork { /* OP_FORK */
|
|
domid_t parent_domain; /* IN: parent's domain id */
|
|
/* Only makes sense for short-lived forks */
|
|
#define XENMEM_FORK_WITH_IOMMU_ALLOWED (1u << 0)
|
|
/* Only makes sense for short-lived forks */
|
|
#define XENMEM_FORK_BLOCK_INTERRUPTS (1u << 1)
|
|
uint16_t flags; /* IN: optional settings */
|
|
uint32_t pad; /* Must be set to 0 */
|
|
} fork;
|
|
} u;
|
|
};
|
|
typedef struct xen_mem_sharing_op xen_mem_sharing_op_t;
|
|
DEFINE_XEN_GUEST_HANDLE(xen_mem_sharing_op_t);
|
|
|
|
/*
|
|
* Attempt to stake a claim for a domain on a quantity of pages
|
|
* of system RAM, but _not_ assign specific pageframes. Only
|
|
* arithmetic is performed so the hypercall is very fast and need
|
|
* not be preemptible, thus sidestepping time-of-check-time-of-use
|
|
* races for memory allocation. Returns 0 if the hypervisor page
|
|
* allocator has atomically and successfully claimed the requested
|
|
* number of pages, else non-zero.
|
|
*
|
|
* Any domain may have only one active claim. When sufficient memory
|
|
* has been allocated to resolve the claim, the claim silently expires.
|
|
* Claiming zero pages effectively resets any outstanding claim and
|
|
* is always successful.
|
|
*
|
|
* Note that a valid claim may be staked even after memory has been
|
|
* allocated for a domain. In this case, the claim is not incremental,
|
|
* i.e. if the domain's total page count is 3, and a claim is staked
|
|
* for 10, only 7 additional pages are claimed.
|
|
*
|
|
* Caller must be privileged or the hypercall fails.
|
|
*/
|
|
#define XENMEM_claim_pages 24
|
|
|
|
/*
|
|
* XENMEM_claim_pages flags - the are no flags at this time.
|
|
* The zero value is appropriate.
|
|
*/
|
|
|
|
/*
|
|
* With some legacy devices, certain guest-physical addresses cannot safely
|
|
* be used for other purposes, e.g. to map guest RAM. This hypercall
|
|
* enumerates those regions so the toolstack can avoid using them.
|
|
*/
|
|
#define XENMEM_reserved_device_memory_map 27
|
|
struct xen_reserved_device_memory {
|
|
xen_pfn_t start_pfn;
|
|
xen_ulong_t nr_pages;
|
|
};
|
|
typedef struct xen_reserved_device_memory xen_reserved_device_memory_t;
|
|
DEFINE_XEN_GUEST_HANDLE(xen_reserved_device_memory_t);
|
|
|
|
struct xen_reserved_device_memory_map {
|
|
#define XENMEM_RDM_ALL 1 /* Request all regions (ignore dev union). */
|
|
/* IN */
|
|
uint32_t flags;
|
|
/*
|
|
* IN/OUT
|
|
*
|
|
* Gets set to the required number of entries when too low,
|
|
* signaled by error code -ERANGE.
|
|
*/
|
|
unsigned int nr_entries;
|
|
/* OUT */
|
|
XEN_GUEST_HANDLE(xen_reserved_device_memory_t) buffer;
|
|
/* IN */
|
|
union {
|
|
physdev_pci_device_t pci;
|
|
} dev;
|
|
};
|
|
typedef struct xen_reserved_device_memory_map xen_reserved_device_memory_map_t;
|
|
DEFINE_XEN_GUEST_HANDLE(xen_reserved_device_memory_map_t);
|
|
|
|
#endif /* defined(__XEN__) || defined(__XEN_TOOLS__) */
|
|
|
|
/*
|
|
* Get the pages for a particular guest resource, so that they can be
|
|
* mapped directly by a tools domain.
|
|
*/
|
|
#define XENMEM_acquire_resource 28
|
|
struct xen_mem_acquire_resource {
|
|
/* IN - The domain whose resource is to be mapped */
|
|
domid_t domid;
|
|
/* IN - the type of resource */
|
|
uint16_t type;
|
|
|
|
#define XENMEM_resource_ioreq_server 0
|
|
#define XENMEM_resource_grant_table 1
|
|
#define XENMEM_resource_vmtrace_buf 2
|
|
|
|
/*
|
|
* IN - a type-specific resource identifier, which must be zero
|
|
* unless stated otherwise.
|
|
*
|
|
* type == XENMEM_resource_ioreq_server -> id == ioreq server id
|
|
* type == XENMEM_resource_grant_table -> id defined below
|
|
*/
|
|
uint32_t id;
|
|
|
|
#define XENMEM_resource_grant_table_id_shared 0
|
|
#define XENMEM_resource_grant_table_id_status 1
|
|
|
|
/*
|
|
* IN/OUT
|
|
*
|
|
* As an IN parameter number of frames of the resource to be mapped.
|
|
* This value may be updated over the course of the operation.
|
|
*
|
|
* When frame_list is NULL and nr_frames is 0, this is interpreted as a
|
|
* request for the size of the resource, which shall be returned in the
|
|
* nr_frames field.
|
|
*
|
|
* The size of a resource will never be zero, but a nonzero result doesn't
|
|
* guarantee that a subsequent mapping request will be successful. There
|
|
* are further type/id specific constraints which may change between the
|
|
* two calls.
|
|
*/
|
|
uint32_t nr_frames;
|
|
uint32_t pad;
|
|
/*
|
|
* IN - the index of the initial frame to be mapped. This parameter
|
|
* is ignored if nr_frames is 0. This value may be updated
|
|
* over the course of the operation.
|
|
*/
|
|
uint64_t frame;
|
|
|
|
#define XENMEM_resource_ioreq_server_frame_bufioreq 0
|
|
#define XENMEM_resource_ioreq_server_frame_ioreq(n) (1 + (n))
|
|
|
|
/*
|
|
* IN/OUT - If the tools domain is PV then, upon return, frame_list
|
|
* will be populated with the MFNs of the resource.
|
|
* If the tools domain is HVM then it is expected that, on
|
|
* entry, frame_list will be populated with a list of GFNs
|
|
* that will be mapped to the MFNs of the resource.
|
|
* If -EIO is returned then the frame_list has only been
|
|
* partially mapped and it is up to the caller to unmap all
|
|
* the GFNs.
|
|
* This parameter may be NULL if nr_frames is 0. This
|
|
* value may be updated over the course of the operation.
|
|
*/
|
|
XEN_GUEST_HANDLE(xen_pfn_t) frame_list;
|
|
};
|
|
typedef struct xen_mem_acquire_resource xen_mem_acquire_resource_t;
|
|
DEFINE_XEN_GUEST_HANDLE(xen_mem_acquire_resource_t);
|
|
|
|
/*
|
|
* XENMEM_get_vnumainfo used by guest to get
|
|
* vNUMA topology from hypervisor.
|
|
*/
|
|
#define XENMEM_get_vnumainfo 26
|
|
|
|
/* vNUMA node memory ranges */
|
|
struct xen_vmemrange {
|
|
uint64_t start, end;
|
|
unsigned int flags;
|
|
unsigned int nid;
|
|
};
|
|
typedef struct xen_vmemrange xen_vmemrange_t;
|
|
DEFINE_XEN_GUEST_HANDLE(xen_vmemrange_t);
|
|
|
|
/*
|
|
* vNUMA topology specifies vNUMA node number, distance table,
|
|
* memory ranges and vcpu mapping provided for guests.
|
|
* XENMEM_get_vnumainfo hypercall expects to see from guest
|
|
* nr_vnodes, nr_vmemranges and nr_vcpus to indicate available memory.
|
|
* After filling guests structures, nr_vnodes, nr_vmemranges and nr_vcpus
|
|
* copied back to guest. Domain returns expected values of nr_vnodes,
|
|
* nr_vmemranges and nr_vcpus to guest if the values where incorrect.
|
|
*/
|
|
struct xen_vnuma_topology_info {
|
|
/* IN */
|
|
domid_t domid;
|
|
uint16_t pad;
|
|
/* IN/OUT */
|
|
unsigned int nr_vnodes;
|
|
unsigned int nr_vcpus;
|
|
unsigned int nr_vmemranges;
|
|
/* OUT */
|
|
union {
|
|
XEN_GUEST_HANDLE(uint) h;
|
|
uint64_t pad;
|
|
} vdistance;
|
|
union {
|
|
XEN_GUEST_HANDLE(uint) h;
|
|
uint64_t pad;
|
|
} vcpu_to_vnode;
|
|
union {
|
|
XEN_GUEST_HANDLE(xen_vmemrange_t) h;
|
|
uint64_t pad;
|
|
} vmemrange;
|
|
};
|
|
typedef struct xen_vnuma_topology_info xen_vnuma_topology_info_t;
|
|
DEFINE_XEN_GUEST_HANDLE(xen_vnuma_topology_info_t);
|
|
|
|
/* Next available subop number is 29 */
|
|
|
|
#endif /* __XEN_PUBLIC_MEMORY_H__ */
|
|
|
|
/*
|
|
* Local variables:
|
|
* mode: C
|
|
* c-file-style: "BSD"
|
|
* c-basic-offset: 4
|
|
* tab-width: 4
|
|
* indent-tabs-mode: nil
|
|
* End:
|
|
*/
|