2 * Copyright (c) 2011 NetApp, Inc.
5 * Redistribution and use in source and binary forms, with or without
6 * modification, are permitted provided that the following conditions
8 * 1. Redistributions of source code must retain the above copyright
9 * notice, this list of conditions and the following disclaimer.
10 * 2. Redistributions in binary form must reproduce the above copyright
11 * notice, this list of conditions and the following disclaimer in the
12 * documentation and/or other materials provided with the distribution.
14 * THIS SOFTWARE IS PROVIDED BY NETAPP, INC ``AS IS'' AND
15 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
16 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
17 * ARE DISCLAIMED. IN NO EVENT SHALL NETAPP, INC OR CONTRIBUTORS BE LIABLE
18 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
19 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
20 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
21 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
22 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
23 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
31 #include <sys/cdefs.h>
32 __FBSDID("$FreeBSD$");
34 #include <sys/param.h>
35 #include <sys/systm.h>
41 #include <machine/segments.h>
42 #include <machine/vmm.h>
44 #include "vmx_cpufunc.h"
54 vmcs_fix_regval(uint32_t encoding, uint64_t val)
59 val = vmx_fix_cr0(val);
62 val = vmx_fix_cr4(val);
71 vmcs_field_encoding(int ident)
74 case VM_REG_GUEST_CR0:
75 return (VMCS_GUEST_CR0);
76 case VM_REG_GUEST_CR3:
77 return (VMCS_GUEST_CR3);
78 case VM_REG_GUEST_CR4:
79 return (VMCS_GUEST_CR4);
80 case VM_REG_GUEST_DR7:
81 return (VMCS_GUEST_DR7);
82 case VM_REG_GUEST_RSP:
83 return (VMCS_GUEST_RSP);
84 case VM_REG_GUEST_RIP:
85 return (VMCS_GUEST_RIP);
86 case VM_REG_GUEST_RFLAGS:
87 return (VMCS_GUEST_RFLAGS);
89 return (VMCS_GUEST_ES_SELECTOR);
91 return (VMCS_GUEST_CS_SELECTOR);
93 return (VMCS_GUEST_SS_SELECTOR);
95 return (VMCS_GUEST_DS_SELECTOR);
97 return (VMCS_GUEST_FS_SELECTOR);
99 return (VMCS_GUEST_GS_SELECTOR);
100 case VM_REG_GUEST_TR:
101 return (VMCS_GUEST_TR_SELECTOR);
102 case VM_REG_GUEST_LDTR:
103 return (VMCS_GUEST_LDTR_SELECTOR);
104 case VM_REG_GUEST_EFER:
105 return (VMCS_GUEST_IA32_EFER);
106 case VM_REG_GUEST_PDPTE0:
107 return (VMCS_GUEST_PDPTE0);
108 case VM_REG_GUEST_PDPTE1:
109 return (VMCS_GUEST_PDPTE1);
110 case VM_REG_GUEST_PDPTE2:
111 return (VMCS_GUEST_PDPTE2);
112 case VM_REG_GUEST_PDPTE3:
113 return (VMCS_GUEST_PDPTE3);
121 vmcs_seg_desc_encoding(int seg, uint32_t *base, uint32_t *lim, uint32_t *acc)
125 case VM_REG_GUEST_ES:
126 *base = VMCS_GUEST_ES_BASE;
127 *lim = VMCS_GUEST_ES_LIMIT;
128 *acc = VMCS_GUEST_ES_ACCESS_RIGHTS;
130 case VM_REG_GUEST_CS:
131 *base = VMCS_GUEST_CS_BASE;
132 *lim = VMCS_GUEST_CS_LIMIT;
133 *acc = VMCS_GUEST_CS_ACCESS_RIGHTS;
135 case VM_REG_GUEST_SS:
136 *base = VMCS_GUEST_SS_BASE;
137 *lim = VMCS_GUEST_SS_LIMIT;
138 *acc = VMCS_GUEST_SS_ACCESS_RIGHTS;
140 case VM_REG_GUEST_DS:
141 *base = VMCS_GUEST_DS_BASE;
142 *lim = VMCS_GUEST_DS_LIMIT;
143 *acc = VMCS_GUEST_DS_ACCESS_RIGHTS;
145 case VM_REG_GUEST_FS:
146 *base = VMCS_GUEST_FS_BASE;
147 *lim = VMCS_GUEST_FS_LIMIT;
148 *acc = VMCS_GUEST_FS_ACCESS_RIGHTS;
150 case VM_REG_GUEST_GS:
151 *base = VMCS_GUEST_GS_BASE;
152 *lim = VMCS_GUEST_GS_LIMIT;
153 *acc = VMCS_GUEST_GS_ACCESS_RIGHTS;
155 case VM_REG_GUEST_TR:
156 *base = VMCS_GUEST_TR_BASE;
157 *lim = VMCS_GUEST_TR_LIMIT;
158 *acc = VMCS_GUEST_TR_ACCESS_RIGHTS;
160 case VM_REG_GUEST_LDTR:
161 *base = VMCS_GUEST_LDTR_BASE;
162 *lim = VMCS_GUEST_LDTR_LIMIT;
163 *acc = VMCS_GUEST_LDTR_ACCESS_RIGHTS;
165 case VM_REG_GUEST_IDTR:
166 *base = VMCS_GUEST_IDTR_BASE;
167 *lim = VMCS_GUEST_IDTR_LIMIT;
168 *acc = VMCS_INVALID_ENCODING;
170 case VM_REG_GUEST_GDTR:
171 *base = VMCS_GUEST_GDTR_BASE;
172 *lim = VMCS_GUEST_GDTR_LIMIT;
173 *acc = VMCS_INVALID_ENCODING;
183 vmcs_getreg(struct vmcs *vmcs, int running, int ident, uint64_t *retval)
189 * If we need to get at vmx-specific state in the VMCS we can bypass
190 * the translation of 'ident' to 'encoding' by simply setting the
191 * sign bit. As it so happens the upper 16 bits are reserved (i.e
192 * set to 0) in the encodings for the VMCS so we are free to use the
196 encoding = ident & 0x7fffffff;
198 encoding = vmcs_field_encoding(ident);
200 if (encoding == (uint32_t)-1)
206 error = vmread(encoding, retval);
215 vmcs_setreg(struct vmcs *vmcs, int running, int ident, uint64_t val)
221 encoding = ident & 0x7fffffff;
223 encoding = vmcs_field_encoding(ident);
225 if (encoding == (uint32_t)-1)
228 val = vmcs_fix_regval(encoding, val);
233 error = vmwrite(encoding, val);
242 vmcs_setdesc(struct vmcs *vmcs, int running, int seg, struct seg_desc *desc)
245 uint32_t base, limit, access;
247 error = vmcs_seg_desc_encoding(seg, &base, &limit, &access);
249 panic("vmcs_setdesc: invalid segment register %d", seg);
253 if ((error = vmwrite(base, desc->base)) != 0)
256 if ((error = vmwrite(limit, desc->limit)) != 0)
259 if (access != VMCS_INVALID_ENCODING) {
260 if ((error = vmwrite(access, desc->access)) != 0)
270 vmcs_getdesc(struct vmcs *vmcs, int running, int seg, struct seg_desc *desc)
273 uint32_t base, limit, access;
276 error = vmcs_seg_desc_encoding(seg, &base, &limit, &access);
278 panic("vmcs_getdesc: invalid segment register %d", seg);
282 if ((error = vmread(base, &u64)) != 0)
286 if ((error = vmread(limit, &u64)) != 0)
290 if (access != VMCS_INVALID_ENCODING) {
291 if ((error = vmread(access, &u64)) != 0)
302 vmcs_set_msr_save(struct vmcs *vmcs, u_long g_area, u_int g_count)
309 * Guest MSRs are saved in the VM-exit MSR-store area.
310 * Guest MSRs are loaded from the VM-entry MSR-load area.
311 * Both areas point to the same location in memory.
313 if ((error = vmwrite(VMCS_EXIT_MSR_STORE, g_area)) != 0)
315 if ((error = vmwrite(VMCS_EXIT_MSR_STORE_COUNT, g_count)) != 0)
318 if ((error = vmwrite(VMCS_ENTRY_MSR_LOAD, g_area)) != 0)
320 if ((error = vmwrite(VMCS_ENTRY_MSR_LOAD_COUNT, g_count)) != 0)
330 vmcs_init(struct vmcs *vmcs)
332 int error, codesel, datasel, tsssel;
333 u_long cr0, cr4, efer;
334 uint64_t pat, fsbase, idtrbase;
337 codesel = vmm_get_host_codesel();
338 datasel = vmm_get_host_datasel();
339 tsssel = vmm_get_host_tsssel();
342 * Make sure we have a "current" VMCS to work with.
346 /* Initialize guest IA32_PAT MSR with the default value */
347 pat = PAT_VALUE(0, PAT_WRITE_BACK) |
348 PAT_VALUE(1, PAT_WRITE_THROUGH) |
349 PAT_VALUE(2, PAT_UNCACHED) |
350 PAT_VALUE(3, PAT_UNCACHEABLE) |
351 PAT_VALUE(4, PAT_WRITE_BACK) |
352 PAT_VALUE(5, PAT_WRITE_THROUGH) |
353 PAT_VALUE(6, PAT_UNCACHED) |
354 PAT_VALUE(7, PAT_UNCACHEABLE);
355 if ((error = vmwrite(VMCS_GUEST_IA32_PAT, pat)) != 0)
360 /* Initialize host IA32_PAT MSR */
361 pat = vmm_get_host_pat();
362 if ((error = vmwrite(VMCS_HOST_IA32_PAT, pat)) != 0)
365 /* Load the IA32_EFER MSR */
366 efer = vmm_get_host_efer();
367 if ((error = vmwrite(VMCS_HOST_IA32_EFER, efer)) != 0)
370 /* Load the control registers */
372 cr0 = vmm_get_host_cr0();
373 if ((error = vmwrite(VMCS_HOST_CR0, cr0)) != 0)
376 cr4 = vmm_get_host_cr4() | CR4_VMXE;
377 if ((error = vmwrite(VMCS_HOST_CR4, cr4)) != 0)
380 /* Load the segment selectors */
381 if ((error = vmwrite(VMCS_HOST_ES_SELECTOR, datasel)) != 0)
384 if ((error = vmwrite(VMCS_HOST_CS_SELECTOR, codesel)) != 0)
387 if ((error = vmwrite(VMCS_HOST_SS_SELECTOR, datasel)) != 0)
390 if ((error = vmwrite(VMCS_HOST_DS_SELECTOR, datasel)) != 0)
393 if ((error = vmwrite(VMCS_HOST_FS_SELECTOR, datasel)) != 0)
396 if ((error = vmwrite(VMCS_HOST_GS_SELECTOR, datasel)) != 0)
399 if ((error = vmwrite(VMCS_HOST_TR_SELECTOR, tsssel)) != 0)
403 * Load the Base-Address for %fs and idtr.
405 * Note that we exclude %gs, tss and gdtr here because their base
406 * address is pcpu specific.
408 fsbase = vmm_get_host_fsbase();
409 if ((error = vmwrite(VMCS_HOST_FS_BASE, fsbase)) != 0)
412 idtrbase = vmm_get_host_idtrbase();
413 if ((error = vmwrite(VMCS_HOST_IDTR_BASE, idtrbase)) != 0)
416 /* instruction pointer */
417 if ((error = vmwrite(VMCS_HOST_RIP, (u_long)vmx_exit_guest)) != 0)
420 /* exception bitmap */
421 exc_bitmap = 1 << IDT_MC;
422 if ((error = vmwrite(VMCS_EXCEPTION_BITMAP, exc_bitmap)) != 0)
426 if ((error = vmwrite(VMCS_LINK_POINTER, ~0)) != 0)
434 extern int vmxon_enabled[];
436 DB_SHOW_COMMAND(vmcs, db_show_vmcs)
438 uint64_t cur_vmcs, val;
441 if (!vmxon_enabled[curcpu]) {
442 db_printf("VMX not enabled\n");
447 db_printf("Only current VMCS supported\n");
452 if (cur_vmcs == VMCS_INITIAL) {
453 db_printf("No current VM context\n");
456 db_printf("VMCS: %jx\n", cur_vmcs);
457 db_printf("VPID: %lu\n", vmcs_read(VMCS_VPID));
458 db_printf("Activity: ");
459 val = vmcs_read(VMCS_GUEST_ACTIVITY);
468 db_printf("Shutdown");
471 db_printf("Wait for SIPI");
474 db_printf("Unknown: %#lx", val);
477 exit = vmcs_read(VMCS_EXIT_REASON);
478 if (exit & 0x80000000)
479 db_printf("Entry Failure Reason: %u\n", exit & 0xffff);
481 db_printf("Exit Reason: %u\n", exit & 0xffff);
482 db_printf("Qualification: %#lx\n", vmcs_exit_qualification());
483 db_printf("Guest Linear Address: %#lx\n",
484 vmcs_read(VMCS_GUEST_LINEAR_ADDRESS));
485 switch (exit & 0x8000ffff) {
486 case EXIT_REASON_EXCEPTION:
487 case EXIT_REASON_EXT_INTR:
488 val = vmcs_read(VMCS_EXIT_INTR_INFO);
489 db_printf("Interrupt Type: ");
490 switch (val >> 8 & 0x7) {
492 db_printf("external");
498 db_printf("HW exception");
501 db_printf("SW exception");
504 db_printf("?? %lu", val >> 8 & 0x7);
507 db_printf(" Vector: %lu", val & 0xff);
509 db_printf(" Error Code: %lx",
510 vmcs_read(VMCS_EXIT_INTR_ERRCODE));
513 case EXIT_REASON_EPT_FAULT:
514 case EXIT_REASON_EPT_MISCONFIG:
515 db_printf("Guest Physical Address: %#lx\n",
516 vmcs_read(VMCS_GUEST_PHYSICAL_ADDRESS));
519 db_printf("VM-instruction error: %#lx\n", vmcs_instruction_error());