1 /******************************************************************************
4 * Memory reservation and information.
6 * Permission is hereby granted, free of charge, to any person obtaining a copy
7 * of this software and associated documentation files (the "Software"), to
8 * deal in the Software without restriction, including without limitation the
9 * rights to use, copy, modify, merge, publish, distribute, sublicense, and/or
10 * sell copies of the Software, and to permit persons to whom the Software is
11 * furnished to do so, subject to the following conditions:
13 * The above copyright notice and this permission notice shall be included in
14 * all copies or substantial portions of the Software.
16 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
17 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
18 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
19 * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
20 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
21 * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
22 * DEALINGS IN THE SOFTWARE.
24 * Copyright (c) 2005, Keir Fraser <keir@xensource.com>
27 #ifndef __XEN_PUBLIC_MEMORY_H__
28 #define __XEN_PUBLIC_MEMORY_H__
33 * Increase or decrease the specified domain's memory reservation. Returns the
34 * number of extents successfully allocated or freed.
35 * arg == addr of struct xen_memory_reservation.
37 #define XENMEM_increase_reservation 0
38 #define XENMEM_decrease_reservation 1
39 #define XENMEM_populate_physmap 6
41 #if __XEN_INTERFACE_VERSION__ >= 0x00030209
43 * Maximum # bits addressable by the user of the allocated region (e.g., I/O
44 * devices often have a 32-bit limitation even in 64-bit systems). If zero
45 * then the user has no addressing restriction. This field is not used by
46 * XENMEM_decrease_reservation.
48 #define XENMEMF_address_bits(x) (x)
49 #define XENMEMF_get_address_bits(x) ((x) & 0xffu)
50 /* NUMA node to allocate from. */
51 #define XENMEMF_node(x) (((x) + 1) << 8)
52 #define XENMEMF_get_node(x) ((((x) >> 8) - 1) & 0xffu)
53 /* Flag to populate physmap with populate-on-demand entries */
54 #define XENMEMF_populate_on_demand (1<<16)
55 /* Flag to request allocation only from the node specified */
56 #define XENMEMF_exact_node_request (1<<17)
57 #define XENMEMF_exact_node(n) (XENMEMF_node(n) | XENMEMF_exact_node_request)
60 struct xen_memory_reservation {
63 * XENMEM_increase_reservation:
64 * OUT: MFN (*not* GMFN) bases of extents that were allocated
65 * XENMEM_decrease_reservation:
66 * IN: GMFN bases of extents to free
67 * XENMEM_populate_physmap:
68 * IN: GPFN bases of extents to populate with memory
69 * OUT: GMFN bases of extents that were allocated
70 * (NB. This command also updates the mach_to_phys translation table)
72 XEN_GUEST_HANDLE(xen_pfn_t) extent_start;
74 /* Number of extents, and size/alignment of each (2^extent_order pages). */
75 xen_ulong_t nr_extents;
76 unsigned int extent_order;
78 #if __XEN_INTERFACE_VERSION__ >= 0x00030209
80 unsigned int mem_flags;
82 unsigned int address_bits;
86 * Domain whose reservation is being changed.
87 * Unprivileged domains can specify only DOMID_SELF.
91 typedef struct xen_memory_reservation xen_memory_reservation_t;
92 DEFINE_XEN_GUEST_HANDLE(xen_memory_reservation_t);
95 * An atomic exchange of memory pages. If return code is zero then
96 * @out.extent_list provides GMFNs of the newly-allocated memory.
97 * Returns zero on complete success, otherwise a negative error code.
98 * On complete success then always @nr_exchanged == @in.nr_extents.
99 * On partial success @nr_exchanged indicates how much work was done.
101 #define XENMEM_exchange 11
102 struct xen_memory_exchange {
104 * [IN] Details of memory extents to be exchanged (GMFN bases).
105 * Note that @in.address_bits is ignored and unused.
107 struct xen_memory_reservation in;
110 * [IN/OUT] Details of new memory extents.
112 * 1. @in.domid == @out.domid
113 * 2. @in.nr_extents << @in.extent_order ==
114 * @out.nr_extents << @out.extent_order
115 * 3. @in.extent_start and @out.extent_start lists must not overlap
116 * 4. @out.extent_start lists GPFN bases to be populated
117 * 5. @out.extent_start is overwritten with allocated GMFN bases
119 struct xen_memory_reservation out;
122 * [OUT] Number of input extents that were successfully exchanged:
123 * 1. The first @nr_exchanged input extents were successfully
125 * 2. The corresponding first entries in the output extent list correctly
126 * indicate the GMFNs that were successfully exchanged.
127 * 3. All other input and output extents are untouched.
128 * 4. If not all input exents are exchanged then the return code of this
129 * command will be non-zero.
130 * 5. THIS FIELD MUST BE INITIALISED TO ZERO BY THE CALLER!
132 xen_ulong_t nr_exchanged;
134 typedef struct xen_memory_exchange xen_memory_exchange_t;
135 DEFINE_XEN_GUEST_HANDLE(xen_memory_exchange_t);
138 * Returns the maximum machine frame number of mapped RAM in this system.
139 * This command always succeeds (it never returns an error code).
142 #define XENMEM_maximum_ram_page 2
145 * Returns the current or maximum memory reservation, in pages, of the
146 * specified domain (may be DOMID_SELF). Returns -ve errcode on failure.
147 * arg == addr of domid_t.
149 #define XENMEM_current_reservation 3
150 #define XENMEM_maximum_reservation 4
153 * Returns the maximum GPFN in use by the guest, or -ve errcode on failure.
155 #define XENMEM_maximum_gpfn 14
158 * Returns a list of MFN bases of 2MB extents comprising the machine_to_phys
159 * mapping table. Architectures which do not have a m2p table do not implement
161 * arg == addr of xen_machphys_mfn_list_t.
163 #define XENMEM_machphys_mfn_list 5
164 struct xen_machphys_mfn_list {
166 * Size of the 'extent_start' array. Fewer entries will be filled if the
167 * machphys table is smaller than max_extents * 2MB.
169 unsigned int max_extents;
172 * Pointer to buffer to fill with list of extent starts. If there are
173 * any large discontiguities in the machine address space, 2MB gaps in
174 * the machphys table will be represented by an MFN base of zero.
176 XEN_GUEST_HANDLE(xen_pfn_t) extent_start;
179 * Number of extents written to the above array. This will be smaller
180 * than 'max_extents' if the machphys table is smaller than max_e * 2MB.
182 unsigned int nr_extents;
184 typedef struct xen_machphys_mfn_list xen_machphys_mfn_list_t;
185 DEFINE_XEN_GUEST_HANDLE(xen_machphys_mfn_list_t);
188 * Returns the location in virtual address space of the machine_to_phys
189 * mapping table. Architectures which do not have a m2p table, or which do not
190 * map it by default into guest address space, do not implement this command.
191 * arg == addr of xen_machphys_mapping_t.
193 #define XENMEM_machphys_mapping 12
194 struct xen_machphys_mapping {
195 xen_ulong_t v_start, v_end; /* Start and end virtual addresses. */
196 xen_ulong_t max_mfn; /* Maximum MFN that can be looked up. */
198 typedef struct xen_machphys_mapping xen_machphys_mapping_t;
199 DEFINE_XEN_GUEST_HANDLE(xen_machphys_mapping_t);
201 #define XENMAPSPACE_shared_info 0 /* shared info page */
202 #define XENMAPSPACE_grant_table 1 /* grant table page */
203 #define XENMAPSPACE_gmfn 2 /* GMFN */
204 #define XENMAPSPACE_gmfn_range 3 /* GMFN range, XENMEM_add_to_physmap only. */
205 #define XENMAPSPACE_gmfn_foreign 4 /* GMFN from another dom,
206 * XENMEM_add_to_physmap_range only.
210 * Sets the GPFN at which a particular page appears in the specified guest's
211 * pseudophysical address space.
212 * arg == addr of xen_add_to_physmap_t.
214 #define XENMEM_add_to_physmap 7
215 struct xen_add_to_physmap {
216 /* Which domain to change the mapping for. */
219 /* Number of pages to go through for gmfn_range */
222 /* Source mapping space. */
223 #define XENMAPSPACE_shared_info 0 /* shared info page */
224 #define XENMAPSPACE_grant_table 1 /* grant table page */
225 #define XENMAPSPACE_gmfn 2 /* GMFN */
226 #define XENMAPSPACE_gmfn_range 3 /* GMFN range */
229 #define XENMAPIDX_grant_table_status 0x80000000
231 /* Index into source mapping space. */
234 /* GPFN where the source mapping page should appear. */
237 typedef struct xen_add_to_physmap xen_add_to_physmap_t;
238 DEFINE_XEN_GUEST_HANDLE(xen_add_to_physmap_t);
241 * Unmaps the page appearing at a particular GPFN from the specified guest's
242 * pseudophysical address space.
243 * arg == addr of xen_remove_from_physmap_t.
245 #define XENMEM_remove_from_physmap 15
246 struct xen_remove_from_physmap {
247 /* Which domain to change the mapping for. */
250 /* GPFN of the current mapping of the page. */
253 typedef struct xen_remove_from_physmap xen_remove_from_physmap_t;
254 DEFINE_XEN_GUEST_HANDLE(xen_remove_from_physmap_t);
257 /*#define XENMEM_translate_gpfn_list 8*/
259 #define XENMEM_add_to_physmap_range 23
260 struct xen_add_to_physmap_range {
262 /* Which domain to change the mapping for. */
264 uint16_t space; /* => enum phys_map_space */
266 /* Number of pages to go through */
268 domid_t foreign_domid; /* IFF gmfn_foreign */
270 /* Indexes into space being mapped. */
271 XEN_GUEST_HANDLE(xen_ulong_t) idxs;
273 /* GPFN in domid where the source mapping page should appear. */
274 XEN_GUEST_HANDLE(xen_pfn_t) gpfns;
278 /* Per index error code. */
279 XEN_GUEST_HANDLE(int) errs;
281 typedef struct xen_add_to_physmap_range xen_add_to_physmap_range_t;
282 DEFINE_XEN_GUEST_HANDLE(xen_add_to_physmap_range_t);
285 * Returns the pseudo-physical memory map as it was when the domain
286 * was started (specified by XENMEM_set_memory_map).
287 * arg == addr of xen_memory_map_t.
289 #define XENMEM_memory_map 9
290 struct xen_memory_map {
292 * On call the number of entries which can be stored in buffer. On
293 * return the number of entries which have been stored in
296 unsigned int nr_entries;
299 * Entries in the buffer are in the same format as returned by the
300 * BIOS INT 0x15 EAX=0xE820 call.
302 XEN_GUEST_HANDLE(void) buffer;
304 typedef struct xen_memory_map xen_memory_map_t;
305 DEFINE_XEN_GUEST_HANDLE(xen_memory_map_t);
308 * Returns the real physical memory map. Passes the same structure as
310 * arg == addr of xen_memory_map_t.
312 #define XENMEM_machine_memory_map 10
315 * Set the pseudo-physical memory map of a domain, as returned by
317 * arg == addr of xen_foreign_memory_map_t.
319 #define XENMEM_set_memory_map 13
320 struct xen_foreign_memory_map {
322 struct xen_memory_map map;
324 typedef struct xen_foreign_memory_map xen_foreign_memory_map_t;
325 DEFINE_XEN_GUEST_HANDLE(xen_foreign_memory_map_t);
327 #define XENMEM_set_pod_target 16
328 #define XENMEM_get_pod_target 17
329 struct xen_pod_target {
331 uint64_t target_pages;
334 uint64_t pod_cache_pages;
335 uint64_t pod_entries;
339 typedef struct xen_pod_target xen_pod_target_t;
341 #if defined(__XEN__) || defined(__XEN_TOOLS__)
343 #ifndef uint64_aligned_t
344 #define uint64_aligned_t uint64_t
348 * Get the number of MFNs saved through memory sharing.
349 * The call never fails.
351 #define XENMEM_get_sharing_freed_pages 18
352 #define XENMEM_get_sharing_shared_pages 19
354 #define XENMEM_paging_op 20
355 #define XENMEM_paging_op_nominate 0
356 #define XENMEM_paging_op_evict 1
357 #define XENMEM_paging_op_prep 2
359 #define XENMEM_access_op 21
360 #define XENMEM_access_op_resume 0
362 struct xen_mem_event_op {
363 uint8_t op; /* XENMEM_*_op_* */
367 /* PAGING_PREP IN: buffer to immediately fill page in */
368 uint64_aligned_t buffer;
370 uint64_aligned_t gfn; /* IN: gfn of page being operated on */
372 typedef struct xen_mem_event_op xen_mem_event_op_t;
373 DEFINE_XEN_GUEST_HANDLE(xen_mem_event_op_t);
375 #define XENMEM_sharing_op 22
376 #define XENMEM_sharing_op_nominate_gfn 0
377 #define XENMEM_sharing_op_nominate_gref 1
378 #define XENMEM_sharing_op_share 2
379 #define XENMEM_sharing_op_resume 3
380 #define XENMEM_sharing_op_debug_gfn 4
381 #define XENMEM_sharing_op_debug_mfn 5
382 #define XENMEM_sharing_op_debug_gref 6
383 #define XENMEM_sharing_op_add_physmap 7
384 #define XENMEM_sharing_op_audit 8
386 #define XENMEM_SHARING_OP_S_HANDLE_INVALID (-10)
387 #define XENMEM_SHARING_OP_C_HANDLE_INVALID (-9)
389 /* The following allows sharing of grant refs. This is useful
390 * for sharing utilities sitting as "filters" in IO backends
391 * (e.g. memshr + blktap(2)). The IO backend is only exposed
392 * to grant references, and this allows sharing of the grefs */
393 #define XENMEM_SHARING_OP_FIELD_IS_GREF_FLAG (1ULL << 62)
395 #define XENMEM_SHARING_OP_FIELD_MAKE_GREF(field, val) \
396 (field) = (XENMEM_SHARING_OP_FIELD_IS_GREF_FLAG | val)
397 #define XENMEM_SHARING_OP_FIELD_IS_GREF(field) \
398 ((field) & XENMEM_SHARING_OP_FIELD_IS_GREF_FLAG)
399 #define XENMEM_SHARING_OP_FIELD_GET_GREF(field) \
400 ((field) & (~XENMEM_SHARING_OP_FIELD_IS_GREF_FLAG))
402 struct xen_mem_sharing_op {
403 uint8_t op; /* XENMEM_sharing_op_* */
407 struct mem_sharing_op_nominate { /* OP_NOMINATE_xxx */
409 uint64_aligned_t gfn; /* IN: gfn to nominate */
410 uint32_t grant_ref; /* IN: grant ref to nominate */
412 uint64_aligned_t handle; /* OUT: the handle */
414 struct mem_sharing_op_share { /* OP_SHARE/ADD_PHYSMAP */
415 uint64_aligned_t source_gfn; /* IN: the gfn of the source page */
416 uint64_aligned_t source_handle; /* IN: handle to the source page */
417 uint64_aligned_t client_gfn; /* IN: the client gfn */
418 uint64_aligned_t client_handle; /* IN: handle to the client page */
419 domid_t client_domain; /* IN: the client domain id */
421 struct mem_sharing_op_debug { /* OP_DEBUG_xxx */
423 uint64_aligned_t gfn; /* IN: gfn to debug */
424 uint64_aligned_t mfn; /* IN: mfn to debug */
425 uint32_t gref; /* IN: gref to debug */
430 typedef struct xen_mem_sharing_op xen_mem_sharing_op_t;
431 DEFINE_XEN_GUEST_HANDLE(xen_mem_sharing_op_t);
433 #endif /* defined(__XEN__) || defined(__XEN_TOOLS__) */
435 #endif /* __XEN_PUBLIC_MEMORY_H__ */
443 * indent-tabs-mode: nil