9118effd1b
In order to support grub2 in Xen PVH environment some additional Xen headers are needed as grub2 will be started in PVH mode requiring to use several HVM hypercalls and structures. Add the needed headers from Xen 4.10 being the first Xen version with full (not only experimental) PVH guest support. Signed-off-by: Juergen Gross <jgross@suse.com> Reviewed-by: Daniel Kiper <daniel.kiper@oracle.com> Tested-by: Hans van Kranenburg <hans@knorrie.org>
665 lines
22 KiB
C
665 lines
22 KiB
C
/******************************************************************************
|
|
* memory.h
|
|
*
|
|
* Memory reservation and information.
|
|
*
|
|
* Permission is hereby granted, free of charge, to any person obtaining a copy
|
|
* of this software and associated documentation files (the "Software"), to
|
|
* deal in the Software without restriction, including without limitation the
|
|
* rights to use, copy, modify, merge, publish, distribute, sublicense, and/or
|
|
* sell copies of the Software, and to permit persons to whom the Software is
|
|
* furnished to do so, subject to the following conditions:
|
|
*
|
|
* The above copyright notice and this permission notice shall be included in
|
|
* all copies or substantial portions of the Software.
|
|
*
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
|
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
|
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
|
|
* AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
|
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
|
|
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
|
|
* DEALINGS IN THE SOFTWARE.
|
|
*
|
|
* Copyright (c) 2005, Keir Fraser <keir@xensource.com>
|
|
*/
|
|
|
|
#ifndef __XEN_PUBLIC_MEMORY_H__
|
|
#define __XEN_PUBLIC_MEMORY_H__
|
|
|
|
#include "xen.h"
|
|
#include "physdev.h"
|
|
|
|
/*
|
|
* Increase or decrease the specified domain's memory reservation. Returns the
|
|
* number of extents successfully allocated or freed.
|
|
* arg == addr of struct xen_memory_reservation.
|
|
*/
|
|
#define XENMEM_increase_reservation 0
|
|
#define XENMEM_decrease_reservation 1
|
|
#define XENMEM_populate_physmap 6
|
|
|
|
#if __XEN_INTERFACE_VERSION__ >= 0x00030209
|
|
/*
|
|
* Maximum # bits addressable by the user of the allocated region (e.g., I/O
|
|
* devices often have a 32-bit limitation even in 64-bit systems). If zero
|
|
* then the user has no addressing restriction. This field is not used by
|
|
* XENMEM_decrease_reservation.
|
|
*/
|
|
#define XENMEMF_address_bits(x) (x)
|
|
#define XENMEMF_get_address_bits(x) ((x) & 0xffu)
|
|
/* NUMA node to allocate from. */
|
|
#define XENMEMF_node(x) (((x) + 1) << 8)
|
|
#define XENMEMF_get_node(x) ((((x) >> 8) - 1) & 0xffu)
|
|
/* Flag to populate physmap with populate-on-demand entries */
|
|
#define XENMEMF_populate_on_demand (1<<16)
|
|
/* Flag to request allocation only from the node specified */
|
|
#define XENMEMF_exact_node_request (1<<17)
|
|
#define XENMEMF_exact_node(n) (XENMEMF_node(n) | XENMEMF_exact_node_request)
|
|
/* Flag to indicate the node specified is virtual node */
|
|
#define XENMEMF_vnode (1<<18)
|
|
#endif
|
|
|
|
struct xen_memory_reservation {
|
|
|
|
/*
|
|
* XENMEM_increase_reservation:
|
|
* OUT: MFN (*not* GMFN) bases of extents that were allocated
|
|
* XENMEM_decrease_reservation:
|
|
* IN: GMFN bases of extents to free
|
|
* XENMEM_populate_physmap:
|
|
* IN: GPFN bases of extents to populate with memory
|
|
* OUT: GMFN bases of extents that were allocated
|
|
* (NB. This command also updates the mach_to_phys translation table)
|
|
* XENMEM_claim_pages:
|
|
* IN: must be zero
|
|
*/
|
|
XEN_GUEST_HANDLE(xen_pfn_t) extent_start;
|
|
|
|
/* Number of extents, and size/alignment of each (2^extent_order pages). */
|
|
xen_ulong_t nr_extents;
|
|
unsigned int extent_order;
|
|
|
|
#if __XEN_INTERFACE_VERSION__ >= 0x00030209
|
|
/* XENMEMF flags. */
|
|
unsigned int mem_flags;
|
|
#else
|
|
unsigned int address_bits;
|
|
#endif
|
|
|
|
/*
|
|
* Domain whose reservation is being changed.
|
|
* Unprivileged domains can specify only DOMID_SELF.
|
|
*/
|
|
domid_t domid;
|
|
};
|
|
typedef struct xen_memory_reservation xen_memory_reservation_t;
|
|
DEFINE_XEN_GUEST_HANDLE(xen_memory_reservation_t);
|
|
|
|
/*
|
|
* An atomic exchange of memory pages. If return code is zero then
|
|
* @out.extent_list provides GMFNs of the newly-allocated memory.
|
|
* Returns zero on complete success, otherwise a negative error code.
|
|
* On complete success then always @nr_exchanged == @in.nr_extents.
|
|
* On partial success @nr_exchanged indicates how much work was done.
|
|
*
|
|
* Note that only PV guests can use this operation.
|
|
*/
|
|
#define XENMEM_exchange 11
|
|
struct xen_memory_exchange {
|
|
/*
|
|
* [IN] Details of memory extents to be exchanged (GMFN bases).
|
|
* Note that @in.address_bits is ignored and unused.
|
|
*/
|
|
struct xen_memory_reservation in;
|
|
|
|
/*
|
|
* [IN/OUT] Details of new memory extents.
|
|
* We require that:
|
|
* 1. @in.domid == @out.domid
|
|
* 2. @in.nr_extents << @in.extent_order ==
|
|
* @out.nr_extents << @out.extent_order
|
|
* 3. @in.extent_start and @out.extent_start lists must not overlap
|
|
* 4. @out.extent_start lists GPFN bases to be populated
|
|
* 5. @out.extent_start is overwritten with allocated GMFN bases
|
|
*/
|
|
struct xen_memory_reservation out;
|
|
|
|
/*
|
|
* [OUT] Number of input extents that were successfully exchanged:
|
|
* 1. The first @nr_exchanged input extents were successfully
|
|
* deallocated.
|
|
* 2. The corresponding first entries in the output extent list correctly
|
|
* indicate the GMFNs that were successfully exchanged.
|
|
* 3. All other input and output extents are untouched.
|
|
* 4. If not all input exents are exchanged then the return code of this
|
|
* command will be non-zero.
|
|
* 5. THIS FIELD MUST BE INITIALISED TO ZERO BY THE CALLER!
|
|
*/
|
|
xen_ulong_t nr_exchanged;
|
|
};
|
|
typedef struct xen_memory_exchange xen_memory_exchange_t;
|
|
DEFINE_XEN_GUEST_HANDLE(xen_memory_exchange_t);
|
|
|
|
/*
|
|
* Returns the maximum machine frame number of mapped RAM in this system.
|
|
* This command always succeeds (it never returns an error code).
|
|
* arg == NULL.
|
|
*/
|
|
#define XENMEM_maximum_ram_page 2
|
|
|
|
/*
|
|
* Returns the current or maximum memory reservation, in pages, of the
|
|
* specified domain (may be DOMID_SELF). Returns -ve errcode on failure.
|
|
* arg == addr of domid_t.
|
|
*/
|
|
#define XENMEM_current_reservation 3
|
|
#define XENMEM_maximum_reservation 4
|
|
|
|
/*
|
|
* Returns the maximum GPFN in use by the guest, or -ve errcode on failure.
|
|
*/
|
|
#define XENMEM_maximum_gpfn 14
|
|
|
|
/*
|
|
* Returns a list of MFN bases of 2MB extents comprising the machine_to_phys
|
|
* mapping table. Architectures which do not have a m2p table do not implement
|
|
* this command.
|
|
* arg == addr of xen_machphys_mfn_list_t.
|
|
*/
|
|
#define XENMEM_machphys_mfn_list 5
|
|
struct xen_machphys_mfn_list {
|
|
/*
|
|
* Size of the 'extent_start' array. Fewer entries will be filled if the
|
|
* machphys table is smaller than max_extents * 2MB.
|
|
*/
|
|
unsigned int max_extents;
|
|
|
|
/*
|
|
* Pointer to buffer to fill with list of extent starts. If there are
|
|
* any large discontiguities in the machine address space, 2MB gaps in
|
|
* the machphys table will be represented by an MFN base of zero.
|
|
*/
|
|
XEN_GUEST_HANDLE(xen_pfn_t) extent_start;
|
|
|
|
/*
|
|
* Number of extents written to the above array. This will be smaller
|
|
* than 'max_extents' if the machphys table is smaller than max_e * 2MB.
|
|
*/
|
|
unsigned int nr_extents;
|
|
};
|
|
typedef struct xen_machphys_mfn_list xen_machphys_mfn_list_t;
|
|
DEFINE_XEN_GUEST_HANDLE(xen_machphys_mfn_list_t);
|
|
|
|
/*
|
|
* For a compat caller, this is identical to XENMEM_machphys_mfn_list.
|
|
*
|
|
* For a non compat caller, this functions similarly to
|
|
* XENMEM_machphys_mfn_list, but returns the mfns making up the compatibility
|
|
* m2p table.
|
|
*/
|
|
#define XENMEM_machphys_compat_mfn_list 25
|
|
|
|
/*
|
|
* Returns the location in virtual address space of the machine_to_phys
|
|
* mapping table. Architectures which do not have a m2p table, or which do not
|
|
* map it by default into guest address space, do not implement this command.
|
|
* arg == addr of xen_machphys_mapping_t.
|
|
*/
|
|
#define XENMEM_machphys_mapping 12
|
|
struct xen_machphys_mapping {
|
|
xen_ulong_t v_start, v_end; /* Start and end virtual addresses. */
|
|
xen_ulong_t max_mfn; /* Maximum MFN that can be looked up. */
|
|
};
|
|
typedef struct xen_machphys_mapping xen_machphys_mapping_t;
|
|
DEFINE_XEN_GUEST_HANDLE(xen_machphys_mapping_t);
|
|
|
|
/* Source mapping space. */
|
|
/* ` enum phys_map_space { */
|
|
#define XENMAPSPACE_shared_info 0 /* shared info page */
|
|
#define XENMAPSPACE_grant_table 1 /* grant table page */
|
|
#define XENMAPSPACE_gmfn 2 /* GMFN */
|
|
#define XENMAPSPACE_gmfn_range 3 /* GMFN range, XENMEM_add_to_physmap only. */
|
|
#define XENMAPSPACE_gmfn_foreign 4 /* GMFN from another dom,
|
|
* XENMEM_add_to_physmap_batch only. */
|
|
#define XENMAPSPACE_dev_mmio 5 /* device mmio region
|
|
ARM only; the region is mapped in
|
|
Stage-2 using the Normal Memory
|
|
Inner/Outer Write-Back Cacheable
|
|
memory attribute. */
|
|
/* ` } */
|
|
|
|
/*
|
|
* Sets the GPFN at which a particular page appears in the specified guest's
|
|
* pseudophysical address space.
|
|
* arg == addr of xen_add_to_physmap_t.
|
|
*/
|
|
#define XENMEM_add_to_physmap 7
|
|
struct xen_add_to_physmap {
|
|
/* Which domain to change the mapping for. */
|
|
domid_t domid;
|
|
|
|
/* Number of pages to go through for gmfn_range */
|
|
uint16_t size;
|
|
|
|
unsigned int space; /* => enum phys_map_space */
|
|
|
|
#define XENMAPIDX_grant_table_status 0x80000000
|
|
|
|
/* Index into space being mapped. */
|
|
xen_ulong_t idx;
|
|
|
|
/* GPFN in domid where the source mapping page should appear. */
|
|
xen_pfn_t gpfn;
|
|
};
|
|
typedef struct xen_add_to_physmap xen_add_to_physmap_t;
|
|
DEFINE_XEN_GUEST_HANDLE(xen_add_to_physmap_t);
|
|
|
|
/* A batched version of add_to_physmap. */
|
|
#define XENMEM_add_to_physmap_batch 23
|
|
struct xen_add_to_physmap_batch {
|
|
/* IN */
|
|
/* Which domain to change the mapping for. */
|
|
domid_t domid;
|
|
uint16_t space; /* => enum phys_map_space */
|
|
|
|
/* Number of pages to go through */
|
|
uint16_t size;
|
|
|
|
#if __XEN_INTERFACE_VERSION__ < 0x00040700
|
|
domid_t foreign_domid; /* IFF gmfn_foreign. Should be 0 for other spaces. */
|
|
#else
|
|
union xen_add_to_physmap_batch_extra {
|
|
domid_t foreign_domid; /* gmfn_foreign */
|
|
uint16_t res0; /* All the other spaces. Should be 0 */
|
|
} u;
|
|
#endif
|
|
|
|
/* Indexes into space being mapped. */
|
|
XEN_GUEST_HANDLE(xen_ulong_t) idxs;
|
|
|
|
/* GPFN in domid where the source mapping page should appear. */
|
|
XEN_GUEST_HANDLE(xen_pfn_t) gpfns;
|
|
|
|
/* OUT */
|
|
|
|
/* Per index error code. */
|
|
XEN_GUEST_HANDLE(int) errs;
|
|
};
|
|
typedef struct xen_add_to_physmap_batch xen_add_to_physmap_batch_t;
|
|
DEFINE_XEN_GUEST_HANDLE(xen_add_to_physmap_batch_t);
|
|
|
|
#if __XEN_INTERFACE_VERSION__ < 0x00040400
|
|
#define XENMEM_add_to_physmap_range XENMEM_add_to_physmap_batch
|
|
#define xen_add_to_physmap_range xen_add_to_physmap_batch
|
|
typedef struct xen_add_to_physmap_batch xen_add_to_physmap_range_t;
|
|
DEFINE_XEN_GUEST_HANDLE(xen_add_to_physmap_range_t);
|
|
#endif
|
|
|
|
/*
|
|
* Unmaps the page appearing at a particular GPFN from the specified guest's
|
|
* pseudophysical address space.
|
|
* arg == addr of xen_remove_from_physmap_t.
|
|
*/
|
|
#define XENMEM_remove_from_physmap 15
|
|
struct xen_remove_from_physmap {
|
|
/* Which domain to change the mapping for. */
|
|
domid_t domid;
|
|
|
|
/* GPFN of the current mapping of the page. */
|
|
xen_pfn_t gpfn;
|
|
};
|
|
typedef struct xen_remove_from_physmap xen_remove_from_physmap_t;
|
|
DEFINE_XEN_GUEST_HANDLE(xen_remove_from_physmap_t);
|
|
|
|
/*** REMOVED ***/
|
|
/*#define XENMEM_translate_gpfn_list 8*/
|
|
|
|
/*
|
|
* Returns the pseudo-physical memory map as it was when the domain
|
|
* was started (specified by XENMEM_set_memory_map).
|
|
* arg == addr of xen_memory_map_t.
|
|
*/
|
|
#define XENMEM_memory_map 9
|
|
struct xen_memory_map {
|
|
/*
|
|
* On call the number of entries which can be stored in buffer. On
|
|
* return the number of entries which have been stored in
|
|
* buffer.
|
|
*/
|
|
unsigned int nr_entries;
|
|
|
|
/*
|
|
* Entries in the buffer are in the same format as returned by the
|
|
* BIOS INT 0x15 EAX=0xE820 call.
|
|
*/
|
|
XEN_GUEST_HANDLE(void) buffer;
|
|
};
|
|
typedef struct xen_memory_map xen_memory_map_t;
|
|
DEFINE_XEN_GUEST_HANDLE(xen_memory_map_t);
|
|
|
|
/*
|
|
* Returns the real physical memory map. Passes the same structure as
|
|
* XENMEM_memory_map.
|
|
* Specifying buffer as NULL will return the number of entries required
|
|
* to store the complete memory map.
|
|
* arg == addr of xen_memory_map_t.
|
|
*/
|
|
#define XENMEM_machine_memory_map 10
|
|
|
|
/*
|
|
* Set the pseudo-physical memory map of a domain, as returned by
|
|
* XENMEM_memory_map.
|
|
* arg == addr of xen_foreign_memory_map_t.
|
|
*/
|
|
#define XENMEM_set_memory_map 13
|
|
struct xen_foreign_memory_map {
|
|
domid_t domid;
|
|
struct xen_memory_map map;
|
|
};
|
|
typedef struct xen_foreign_memory_map xen_foreign_memory_map_t;
|
|
DEFINE_XEN_GUEST_HANDLE(xen_foreign_memory_map_t);
|
|
|
|
#define XENMEM_set_pod_target 16
|
|
#define XENMEM_get_pod_target 17
|
|
struct xen_pod_target {
|
|
/* IN */
|
|
uint64_t target_pages;
|
|
/* OUT */
|
|
uint64_t tot_pages;
|
|
uint64_t pod_cache_pages;
|
|
uint64_t pod_entries;
|
|
/* IN */
|
|
domid_t domid;
|
|
};
|
|
typedef struct xen_pod_target xen_pod_target_t;
|
|
|
|
#if defined(__XEN__) || defined(__XEN_TOOLS__)
|
|
|
|
#ifndef uint64_aligned_t
|
|
#define uint64_aligned_t uint64_t
|
|
#endif
|
|
|
|
/*
|
|
* Get the number of MFNs saved through memory sharing.
|
|
* The call never fails.
|
|
*/
|
|
#define XENMEM_get_sharing_freed_pages 18
|
|
#define XENMEM_get_sharing_shared_pages 19
|
|
|
|
#define XENMEM_paging_op 20
|
|
#define XENMEM_paging_op_nominate 0
|
|
#define XENMEM_paging_op_evict 1
|
|
#define XENMEM_paging_op_prep 2
|
|
|
|
struct xen_mem_paging_op {
|
|
uint8_t op; /* XENMEM_paging_op_* */
|
|
domid_t domain;
|
|
|
|
/* PAGING_PREP IN: buffer to immediately fill page in */
|
|
uint64_aligned_t buffer;
|
|
/* Other OPs */
|
|
uint64_aligned_t gfn; /* IN: gfn of page being operated on */
|
|
};
|
|
typedef struct xen_mem_paging_op xen_mem_paging_op_t;
|
|
DEFINE_XEN_GUEST_HANDLE(xen_mem_paging_op_t);
|
|
|
|
#define XENMEM_access_op 21
|
|
#define XENMEM_access_op_set_access 0
|
|
#define XENMEM_access_op_get_access 1
|
|
/*
|
|
* XENMEM_access_op_enable_emulate and XENMEM_access_op_disable_emulate are
|
|
* currently unused, but since they have been in use please do not reuse them.
|
|
*
|
|
* #define XENMEM_access_op_enable_emulate 2
|
|
* #define XENMEM_access_op_disable_emulate 3
|
|
*/
|
|
#define XENMEM_access_op_set_access_multi 4
|
|
|
|
typedef enum {
|
|
XENMEM_access_n,
|
|
XENMEM_access_r,
|
|
XENMEM_access_w,
|
|
XENMEM_access_rw,
|
|
XENMEM_access_x,
|
|
XENMEM_access_rx,
|
|
XENMEM_access_wx,
|
|
XENMEM_access_rwx,
|
|
/*
|
|
* Page starts off as r-x, but automatically
|
|
* change to r-w on a write
|
|
*/
|
|
XENMEM_access_rx2rw,
|
|
/*
|
|
* Log access: starts off as n, automatically
|
|
* goes to rwx, generating an event without
|
|
* pausing the vcpu
|
|
*/
|
|
XENMEM_access_n2rwx,
|
|
/* Take the domain default */
|
|
XENMEM_access_default
|
|
} xenmem_access_t;
|
|
|
|
struct xen_mem_access_op {
|
|
/* XENMEM_access_op_* */
|
|
uint8_t op;
|
|
/* xenmem_access_t */
|
|
uint8_t access;
|
|
domid_t domid;
|
|
/*
|
|
* Number of pages for set op (or size of pfn_list for
|
|
* XENMEM_access_op_set_access_multi)
|
|
* Ignored on setting default access and other ops
|
|
*/
|
|
uint32_t nr;
|
|
/*
|
|
* First pfn for set op
|
|
* pfn for get op
|
|
* ~0ull is used to set and get the default access for pages
|
|
*/
|
|
uint64_aligned_t pfn;
|
|
/*
|
|
* List of pfns to set access for
|
|
* Used only with XENMEM_access_op_set_access_multi
|
|
*/
|
|
XEN_GUEST_HANDLE(const_uint64) pfn_list;
|
|
/*
|
|
* Corresponding list of access settings for pfn_list
|
|
* Used only with XENMEM_access_op_set_access_multi
|
|
*/
|
|
XEN_GUEST_HANDLE(const_uint8) access_list;
|
|
};
|
|
typedef struct xen_mem_access_op xen_mem_access_op_t;
|
|
DEFINE_XEN_GUEST_HANDLE(xen_mem_access_op_t);
|
|
|
|
#define XENMEM_sharing_op 22
|
|
#define XENMEM_sharing_op_nominate_gfn 0
|
|
#define XENMEM_sharing_op_nominate_gref 1
|
|
#define XENMEM_sharing_op_share 2
|
|
#define XENMEM_sharing_op_debug_gfn 3
|
|
#define XENMEM_sharing_op_debug_mfn 4
|
|
#define XENMEM_sharing_op_debug_gref 5
|
|
#define XENMEM_sharing_op_add_physmap 6
|
|
#define XENMEM_sharing_op_audit 7
|
|
#define XENMEM_sharing_op_range_share 8
|
|
|
|
#define XENMEM_SHARING_OP_S_HANDLE_INVALID (-10)
|
|
#define XENMEM_SHARING_OP_C_HANDLE_INVALID (-9)
|
|
|
|
/* The following allows sharing of grant refs. This is useful
|
|
* for sharing utilities sitting as "filters" in IO backends
|
|
* (e.g. memshr + blktap(2)). The IO backend is only exposed
|
|
* to grant references, and this allows sharing of the grefs */
|
|
#define XENMEM_SHARING_OP_FIELD_IS_GREF_FLAG (xen_mk_ullong(1) << 62)
|
|
|
|
#define XENMEM_SHARING_OP_FIELD_MAKE_GREF(field, val) \
|
|
(field) = (XENMEM_SHARING_OP_FIELD_IS_GREF_FLAG | val)
|
|
#define XENMEM_SHARING_OP_FIELD_IS_GREF(field) \
|
|
((field) & XENMEM_SHARING_OP_FIELD_IS_GREF_FLAG)
|
|
#define XENMEM_SHARING_OP_FIELD_GET_GREF(field) \
|
|
((field) & (~XENMEM_SHARING_OP_FIELD_IS_GREF_FLAG))
|
|
|
|
struct xen_mem_sharing_op {
|
|
uint8_t op; /* XENMEM_sharing_op_* */
|
|
domid_t domain;
|
|
|
|
union {
|
|
struct mem_sharing_op_nominate { /* OP_NOMINATE_xxx */
|
|
union {
|
|
uint64_aligned_t gfn; /* IN: gfn to nominate */
|
|
uint32_t grant_ref; /* IN: grant ref to nominate */
|
|
} u;
|
|
uint64_aligned_t handle; /* OUT: the handle */
|
|
} nominate;
|
|
struct mem_sharing_op_share { /* OP_SHARE/ADD_PHYSMAP */
|
|
uint64_aligned_t source_gfn; /* IN: the gfn of the source page */
|
|
uint64_aligned_t source_handle; /* IN: handle to the source page */
|
|
uint64_aligned_t client_gfn; /* IN: the client gfn */
|
|
uint64_aligned_t client_handle; /* IN: handle to the client page */
|
|
domid_t client_domain; /* IN: the client domain id */
|
|
} share;
|
|
struct mem_sharing_op_range { /* OP_RANGE_SHARE */
|
|
uint64_aligned_t first_gfn; /* IN: the first gfn */
|
|
uint64_aligned_t last_gfn; /* IN: the last gfn */
|
|
uint64_aligned_t opaque; /* Must be set to 0 */
|
|
domid_t client_domain; /* IN: the client domain id */
|
|
uint16_t _pad[3]; /* Must be set to 0 */
|
|
} range;
|
|
struct mem_sharing_op_debug { /* OP_DEBUG_xxx */
|
|
union {
|
|
uint64_aligned_t gfn; /* IN: gfn to debug */
|
|
uint64_aligned_t mfn; /* IN: mfn to debug */
|
|
uint32_t gref; /* IN: gref to debug */
|
|
} u;
|
|
} debug;
|
|
} u;
|
|
};
|
|
typedef struct xen_mem_sharing_op xen_mem_sharing_op_t;
|
|
DEFINE_XEN_GUEST_HANDLE(xen_mem_sharing_op_t);
|
|
|
|
/*
|
|
* Attempt to stake a claim for a domain on a quantity of pages
|
|
* of system RAM, but _not_ assign specific pageframes. Only
|
|
* arithmetic is performed so the hypercall is very fast and need
|
|
* not be preemptible, thus sidestepping time-of-check-time-of-use
|
|
* races for memory allocation. Returns 0 if the hypervisor page
|
|
* allocator has atomically and successfully claimed the requested
|
|
* number of pages, else non-zero.
|
|
*
|
|
* Any domain may have only one active claim. When sufficient memory
|
|
* has been allocated to resolve the claim, the claim silently expires.
|
|
* Claiming zero pages effectively resets any outstanding claim and
|
|
* is always successful.
|
|
*
|
|
* Note that a valid claim may be staked even after memory has been
|
|
* allocated for a domain. In this case, the claim is not incremental,
|
|
* i.e. if the domain's tot_pages is 3, and a claim is staked for 10,
|
|
* only 7 additional pages are claimed.
|
|
*
|
|
* Caller must be privileged or the hypercall fails.
|
|
*/
|
|
#define XENMEM_claim_pages 24
|
|
|
|
/*
|
|
* XENMEM_claim_pages flags - the are no flags at this time.
|
|
* The zero value is appropriate.
|
|
*/
|
|
|
|
/*
|
|
* With some legacy devices, certain guest-physical addresses cannot safely
|
|
* be used for other purposes, e.g. to map guest RAM. This hypercall
|
|
* enumerates those regions so the toolstack can avoid using them.
|
|
*/
|
|
#define XENMEM_reserved_device_memory_map 27
|
|
struct xen_reserved_device_memory {
|
|
xen_pfn_t start_pfn;
|
|
xen_ulong_t nr_pages;
|
|
};
|
|
typedef struct xen_reserved_device_memory xen_reserved_device_memory_t;
|
|
DEFINE_XEN_GUEST_HANDLE(xen_reserved_device_memory_t);
|
|
|
|
struct xen_reserved_device_memory_map {
|
|
#define XENMEM_RDM_ALL 1 /* Request all regions (ignore dev union). */
|
|
/* IN */
|
|
uint32_t flags;
|
|
/*
|
|
* IN/OUT
|
|
*
|
|
* Gets set to the required number of entries when too low,
|
|
* signaled by error code -ERANGE.
|
|
*/
|
|
unsigned int nr_entries;
|
|
/* OUT */
|
|
XEN_GUEST_HANDLE(xen_reserved_device_memory_t) buffer;
|
|
/* IN */
|
|
union {
|
|
struct physdev_pci_device pci;
|
|
} dev;
|
|
};
|
|
typedef struct xen_reserved_device_memory_map xen_reserved_device_memory_map_t;
|
|
DEFINE_XEN_GUEST_HANDLE(xen_reserved_device_memory_map_t);
|
|
|
|
#endif /* defined(__XEN__) || defined(__XEN_TOOLS__) */
|
|
|
|
/*
|
|
* XENMEM_get_vnumainfo used by guest to get
|
|
* vNUMA topology from hypervisor.
|
|
*/
|
|
#define XENMEM_get_vnumainfo 26
|
|
|
|
/* vNUMA node memory ranges */
|
|
struct xen_vmemrange {
|
|
uint64_t start, end;
|
|
unsigned int flags;
|
|
unsigned int nid;
|
|
};
|
|
typedef struct xen_vmemrange xen_vmemrange_t;
|
|
DEFINE_XEN_GUEST_HANDLE(xen_vmemrange_t);
|
|
|
|
/*
|
|
* vNUMA topology specifies vNUMA node number, distance table,
|
|
* memory ranges and vcpu mapping provided for guests.
|
|
* XENMEM_get_vnumainfo hypercall expects to see from guest
|
|
* nr_vnodes, nr_vmemranges and nr_vcpus to indicate available memory.
|
|
* After filling guests structures, nr_vnodes, nr_vmemranges and nr_vcpus
|
|
* copied back to guest. Domain returns expected values of nr_vnodes,
|
|
* nr_vmemranges and nr_vcpus to guest if the values where incorrect.
|
|
*/
|
|
struct xen_vnuma_topology_info {
|
|
/* IN */
|
|
domid_t domid;
|
|
uint16_t pad;
|
|
/* IN/OUT */
|
|
unsigned int nr_vnodes;
|
|
unsigned int nr_vcpus;
|
|
unsigned int nr_vmemranges;
|
|
/* OUT */
|
|
union {
|
|
XEN_GUEST_HANDLE(uint) h;
|
|
uint64_t pad;
|
|
} vdistance;
|
|
union {
|
|
XEN_GUEST_HANDLE(uint) h;
|
|
uint64_t pad;
|
|
} vcpu_to_vnode;
|
|
union {
|
|
XEN_GUEST_HANDLE(xen_vmemrange_t) h;
|
|
uint64_t pad;
|
|
} vmemrange;
|
|
};
|
|
typedef struct xen_vnuma_topology_info xen_vnuma_topology_info_t;
|
|
DEFINE_XEN_GUEST_HANDLE(xen_vnuma_topology_info_t);
|
|
|
|
/* Next available subop number is 28 */
|
|
|
|
#endif /* __XEN_PUBLIC_MEMORY_H__ */
|
|
|
|
/*
|
|
* Local variables:
|
|
* mode: C
|
|
* c-file-style: "BSD"
|
|
* c-basic-offset: 4
|
|
* tab-width: 4
|
|
* indent-tabs-mode: nil
|
|
* End:
|
|
*/
|