]> CyberLeo.Net >> Repos - FreeBSD/releng/10.2.git/blob - sys/x86/xen/hvm.c
- Copy stable/10@285827 to releng/10.2 in preparation for 10.2-RC1
[FreeBSD/releng/10.2.git] / sys / x86 / xen / hvm.c
1 /*
2  * Copyright (c) 2008, 2013 Citrix Systems, Inc.
3  * Copyright (c) 2012 Spectra Logic Corporation
4  * All rights reserved.
5  *
6  * Redistribution and use in source and binary forms, with or without
7  * modification, are permitted provided that the following conditions
8  * are met:
9  * 1. Redistributions of source code must retain the above copyright
10  *    notice, this list of conditions and the following disclaimer.
11  * 2. Redistributions in binary form must reproduce the above copyright
12  *    notice, this list of conditions and the following disclaimer in the
13  *    documentation and/or other materials provided with the distribution.
14  *
15  * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS AS IS'' AND
16  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
17  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
18  * ARE DISCLAIMED.  IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
19  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
20  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
21  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
22  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
23  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
24  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
25  * SUCH DAMAGE.
26  */
27
28 #include <sys/cdefs.h>
29 __FBSDID("$FreeBSD$");
30
31 #include <sys/param.h>
32 #include <sys/bus.h>
33 #include <sys/kernel.h>
34 #include <sys/malloc.h>
35 #include <sys/proc.h>
36 #include <sys/smp.h>
37 #include <sys/systm.h>
38
39 #include <vm/vm.h>
40 #include <vm/pmap.h>
41
42 #include <dev/pci/pcivar.h>
43
44 #include <machine/cpufunc.h>
45 #include <machine/cpu.h>
46 #include <machine/smp.h>
47
48 #include <x86/apicreg.h>
49
50 #include <xen/xen-os.h>
51 #include <xen/features.h>
52 #include <xen/gnttab.h>
53 #include <xen/hypervisor.h>
54 #include <xen/hvm.h>
55 #include <xen/xen_intr.h>
56
57 #include <xen/interface/hvm/params.h>
58 #include <xen/interface/vcpu.h>
59
60 /*--------------------------- Forward Declarations ---------------------------*/
61 #ifdef SMP
62 static driver_filter_t xen_smp_rendezvous_action;
63 static driver_filter_t xen_invltlb;
64 static driver_filter_t xen_invlpg;
65 static driver_filter_t xen_invlrng;
66 static driver_filter_t xen_invlcache;
67 #ifdef __i386__
68 static driver_filter_t xen_lazypmap;
69 #endif
70 static driver_filter_t xen_ipi_bitmap_handler;
71 static driver_filter_t xen_cpustop_handler;
72 static driver_filter_t xen_cpususpend_handler;
73 static driver_filter_t xen_cpustophard_handler;
74 static void xen_ipi_vectored(u_int vector, int dest);
75 static void xen_hvm_cpu_resume(void);
76 #endif
77 static void xen_hvm_cpu_init(void);
78
79 /*---------------------------- Extern Declarations ---------------------------*/
80 #ifdef __i386__
81 extern void pmap_lazyfix_action(void);
82 #endif
83 #ifdef __amd64__
84 extern int pmap_pcid_enabled;
85 #endif
86
87 /* Variables used by mp_machdep to perform the bitmap IPI */
88 extern volatile u_int cpu_ipi_pending[MAXCPU];
89
90 /*---------------------------------- Macros ----------------------------------*/
91 #define IPI_TO_IDX(ipi) ((ipi) - APIC_IPI_INTS)
92
93 /*-------------------------------- Local Types -------------------------------*/
94 enum xen_hvm_init_type {
95         XEN_HVM_INIT_COLD,
96         XEN_HVM_INIT_CANCELLED_SUSPEND,
97         XEN_HVM_INIT_RESUME
98 };
99
100 struct xen_ipi_handler
101 {
102         driver_filter_t *filter;
103         const char      *description;
104 };
105
106 /*-------------------------------- Global Data -------------------------------*/
107 enum xen_domain_type xen_domain_type = XEN_NATIVE;
108
109 #ifdef SMP
110 struct cpu_ops xen_hvm_cpu_ops = {
111         .ipi_vectored   = lapic_ipi_vectored,
112         .cpu_init       = xen_hvm_cpu_init,
113         .cpu_resume     = xen_hvm_cpu_resume
114 };
115 #endif
116
117 static MALLOC_DEFINE(M_XENHVM, "xen_hvm", "Xen HVM PV Support");
118
119 #ifdef SMP
120 static struct xen_ipi_handler xen_ipis[] = 
121 {
122         [IPI_TO_IDX(IPI_RENDEZVOUS)]    = { xen_smp_rendezvous_action,  "r"   },
123         [IPI_TO_IDX(IPI_INVLTLB)]       = { xen_invltlb,                "itlb"},
124         [IPI_TO_IDX(IPI_INVLPG)]        = { xen_invlpg,                 "ipg" },
125         [IPI_TO_IDX(IPI_INVLRNG)]       = { xen_invlrng,                "irg" },
126         [IPI_TO_IDX(IPI_INVLCACHE)]     = { xen_invlcache,              "ic"  },
127 #ifdef __i386__
128         [IPI_TO_IDX(IPI_LAZYPMAP)]      = { xen_lazypmap,               "lp"  },
129 #endif
130         [IPI_TO_IDX(IPI_BITMAP_VECTOR)] = { xen_ipi_bitmap_handler,     "b"   },
131         [IPI_TO_IDX(IPI_STOP)]          = { xen_cpustop_handler,        "st"  },
132         [IPI_TO_IDX(IPI_SUSPEND)]       = { xen_cpususpend_handler,     "sp"  },
133         [IPI_TO_IDX(IPI_STOP_HARD)]     = { xen_cpustophard_handler,    "sth" },
134 };
135 #endif
136
137 /**
138  * If non-zero, the hypervisor has been configured to use a direct
139  * IDT event callback for interrupt injection.
140  */
141 int xen_vector_callback_enabled;
142
143 /*------------------------------- Per-CPU Data -------------------------------*/
144 DPCPU_DEFINE(struct vcpu_info, vcpu_local_info);
145 DPCPU_DEFINE(struct vcpu_info *, vcpu_info);
146 #ifdef SMP
147 DPCPU_DEFINE(xen_intr_handle_t, ipi_handle[nitems(xen_ipis)]);
148 #endif
149
150 /*------------------ Hypervisor Access Shared Memory Regions -----------------*/
151 /** Hypercall table accessed via HYPERVISOR_*_op() methods. */
152 char *hypercall_stubs;
153 shared_info_t *HYPERVISOR_shared_info;
154
155 #ifdef SMP
156 /*---------------------------- XEN PV IPI Handlers ---------------------------*/
157 /*
158  * This are C clones of the ASM functions found in apic_vector.s
159  */
160 static int
161 xen_ipi_bitmap_handler(void *arg)
162 {
163         struct trapframe *frame;
164
165         frame = arg;
166         ipi_bitmap_handler(*frame);
167         return (FILTER_HANDLED);
168 }
169
170 static int
171 xen_smp_rendezvous_action(void *arg)
172 {
173 #ifdef COUNT_IPIS
174         (*ipi_rendezvous_counts[PCPU_GET(cpuid)])++;
175 #endif /* COUNT_IPIS */
176
177         smp_rendezvous_action();
178         return (FILTER_HANDLED);
179 }
180
181 static int
182 xen_invltlb(void *arg)
183 {
184
185         invltlb_handler();
186         return (FILTER_HANDLED);
187 }
188
189 #ifdef __amd64__
190 static int
191 xen_invltlb_pcid(void *arg)
192 {
193
194         invltlb_pcid_handler();
195         return (FILTER_HANDLED);
196 }
197 #endif
198
199 static int
200 xen_invlpg(void *arg)
201 {
202
203         invlpg_handler();
204         return (FILTER_HANDLED);
205 }
206
207 #ifdef __amd64__
208 static int
209 xen_invlpg_pcid(void *arg)
210 {
211
212         invlpg_pcid_handler();
213         return (FILTER_HANDLED);
214 }
215 #endif
216
217 static int
218 xen_invlrng(void *arg)
219 {
220
221         invlrng_handler();
222         return (FILTER_HANDLED);
223 }
224
225 static int
226 xen_invlcache(void *arg)
227 {
228
229         invlcache_handler();
230         return (FILTER_HANDLED);
231 }
232
233 #ifdef __i386__
234 static int
235 xen_lazypmap(void *arg)
236 {
237
238         pmap_lazyfix_action();
239         return (FILTER_HANDLED);
240 }
241 #endif
242
243 static int
244 xen_cpustop_handler(void *arg)
245 {
246
247         cpustop_handler();
248         return (FILTER_HANDLED);
249 }
250
251 static int
252 xen_cpususpend_handler(void *arg)
253 {
254
255         cpususpend_handler();
256         return (FILTER_HANDLED);
257 }
258
259 static int
260 xen_cpustophard_handler(void *arg)
261 {
262
263         ipi_nmi_handler();
264         return (FILTER_HANDLED);
265 }
266
267 /* Xen PV IPI sender */
268 static void
269 xen_ipi_vectored(u_int vector, int dest)
270 {
271         xen_intr_handle_t *ipi_handle;
272         int ipi_idx, to_cpu, self;
273
274         ipi_idx = IPI_TO_IDX(vector);
275         if (ipi_idx > nitems(xen_ipis))
276                 panic("IPI out of range");
277
278         switch(dest) {
279         case APIC_IPI_DEST_SELF:
280                 ipi_handle = DPCPU_GET(ipi_handle);
281                 xen_intr_signal(ipi_handle[ipi_idx]);
282                 break;
283         case APIC_IPI_DEST_ALL:
284                 CPU_FOREACH(to_cpu) {
285                         ipi_handle = DPCPU_ID_GET(to_cpu, ipi_handle);
286                         xen_intr_signal(ipi_handle[ipi_idx]);
287                 }
288                 break;
289         case APIC_IPI_DEST_OTHERS:
290                 self = PCPU_GET(cpuid);
291                 CPU_FOREACH(to_cpu) {
292                         if (to_cpu != self) {
293                                 ipi_handle = DPCPU_ID_GET(to_cpu, ipi_handle);
294                                 xen_intr_signal(ipi_handle[ipi_idx]);
295                         }
296                 }
297                 break;
298         default:
299                 to_cpu = apic_cpuid(dest);
300                 ipi_handle = DPCPU_ID_GET(to_cpu, ipi_handle);
301                 xen_intr_signal(ipi_handle[ipi_idx]);
302                 break;
303         }
304 }
305
306 /* XEN diverged cpu operations */
307 static void
308 xen_hvm_cpu_resume(void)
309 {
310         u_int cpuid = PCPU_GET(cpuid);
311
312         /*
313          * Reset pending bitmap IPIs, because Xen doesn't preserve pending
314          * event channels on migration.
315          */
316         cpu_ipi_pending[cpuid] = 0;
317
318         /* register vcpu_info area */
319         xen_hvm_cpu_init();
320 }
321
322 static void
323 xen_cpu_ipi_init(int cpu)
324 {
325         xen_intr_handle_t *ipi_handle;
326         const struct xen_ipi_handler *ipi;
327         device_t dev;
328         int idx, rc;
329
330         ipi_handle = DPCPU_ID_GET(cpu, ipi_handle);
331         dev = pcpu_find(cpu)->pc_device;
332         KASSERT((dev != NULL), ("NULL pcpu device_t"));
333
334         for (ipi = xen_ipis, idx = 0; idx < nitems(xen_ipis); ipi++, idx++) {
335
336                 if (ipi->filter == NULL) {
337                         ipi_handle[idx] = NULL;
338                         continue;
339                 }
340
341                 rc = xen_intr_alloc_and_bind_ipi(dev, cpu, ipi->filter,
342                     INTR_TYPE_TTY, &ipi_handle[idx]);
343                 if (rc != 0)
344                         panic("Unable to allocate a XEN IPI port");
345                 xen_intr_describe(ipi_handle[idx], "%s", ipi->description);
346         }
347 }
348
349 static void
350 xen_setup_cpus(void)
351 {
352         int i;
353
354         if (!xen_hvm_domain() || !xen_vector_callback_enabled)
355                 return;
356
357 #ifdef __amd64__
358         if (pmap_pcid_enabled) {
359                 xen_ipis[IPI_TO_IDX(IPI_INVLTLB)].filter = xen_invltlb_pcid;
360                 xen_ipis[IPI_TO_IDX(IPI_INVLPG)].filter = xen_invlpg_pcid;
361         }
362 #endif
363         CPU_FOREACH(i)
364                 xen_cpu_ipi_init(i);
365
366         /* Set the xen pv ipi ops to replace the native ones */
367         cpu_ops.ipi_vectored = xen_ipi_vectored;
368 }
369 #endif
370
371 /*---------------------- XEN Hypervisor Probe and Setup ----------------------*/
372 static uint32_t
373 xen_hvm_cpuid_base(void)
374 {
375         uint32_t base, regs[4];
376
377         for (base = 0x40000000; base < 0x40010000; base += 0x100) {
378                 do_cpuid(base, regs);
379                 if (!memcmp("XenVMMXenVMM", &regs[1], 12)
380                     && (regs[0] - base) >= 2)
381                         return (base);
382         }
383         return (0);
384 }
385
386 /*
387  * Allocate and fill in the hypcall page.
388  */
389 static int
390 xen_hvm_init_hypercall_stubs(void)
391 {
392         uint32_t base, regs[4];
393         int i;
394
395         base = xen_hvm_cpuid_base();
396         if (base == 0)
397                 return (ENXIO);
398
399         if (hypercall_stubs == NULL) {
400                 do_cpuid(base + 1, regs);
401                 printf("XEN: Hypervisor version %d.%d detected.\n",
402                     regs[0] >> 16, regs[0] & 0xffff);
403         }
404
405         /*
406          * Find the hypercall pages.
407          */
408         do_cpuid(base + 2, regs);
409         
410         if (hypercall_stubs == NULL) {
411                 size_t call_region_size;
412
413                 call_region_size = regs[0] * PAGE_SIZE;
414                 hypercall_stubs = malloc(call_region_size, M_XENHVM, M_NOWAIT);
415                 if (hypercall_stubs == NULL)
416                         panic("Unable to allocate Xen hypercall region");
417         }
418
419         for (i = 0; i < regs[0]; i++)
420                 wrmsr(regs[1], vtophys(hypercall_stubs + i * PAGE_SIZE) + i);
421
422         return (0);
423 }
424
425 static void
426 xen_hvm_init_shared_info_page(void)
427 {
428         struct xen_add_to_physmap xatp;
429
430         if (HYPERVISOR_shared_info == NULL) {
431                 HYPERVISOR_shared_info = malloc(PAGE_SIZE, M_XENHVM, M_NOWAIT);
432                 if (HYPERVISOR_shared_info == NULL)
433                         panic("Unable to allocate Xen shared info page");
434         }
435
436         xatp.domid = DOMID_SELF;
437         xatp.idx = 0;
438         xatp.space = XENMAPSPACE_shared_info;
439         xatp.gpfn = vtophys(HYPERVISOR_shared_info) >> PAGE_SHIFT;
440         if (HYPERVISOR_memory_op(XENMEM_add_to_physmap, &xatp))
441                 panic("HYPERVISOR_memory_op failed");
442 }
443
444 /*
445  * Tell the hypervisor how to contact us for event channel callbacks.
446  */
447 void
448 xen_hvm_set_callback(device_t dev)
449 {
450         struct xen_hvm_param xhp;
451         int irq;
452
453         if (xen_vector_callback_enabled)
454                 return;
455
456         xhp.domid = DOMID_SELF;
457         xhp.index = HVM_PARAM_CALLBACK_IRQ;
458         if (xen_feature(XENFEAT_hvm_callback_vector) != 0) {
459                 int error;
460
461                 xhp.value = HVM_CALLBACK_VECTOR(IDT_EVTCHN);
462                 error = HYPERVISOR_hvm_op(HVMOP_set_param, &xhp);
463                 if (error == 0) {
464                         xen_vector_callback_enabled = 1;
465                         return;
466                 }
467                 printf("Xen HVM callback vector registration failed (%d). "
468                     "Falling back to emulated device interrupt\n", error);
469         }
470         xen_vector_callback_enabled = 0;
471         if (dev == NULL) {
472                 /*
473                  * Called from early boot or resume.
474                  * xenpci will invoke us again later.
475                  */
476                 return;
477         }
478
479         irq = pci_get_irq(dev);
480         if (irq < 16) {
481                 xhp.value = HVM_CALLBACK_GSI(irq);
482         } else {
483                 u_int slot;
484                 u_int pin;
485
486                 slot = pci_get_slot(dev);
487                 pin = pci_get_intpin(dev) - 1;
488                 xhp.value = HVM_CALLBACK_PCI_INTX(slot, pin);
489         }
490
491         if (HYPERVISOR_hvm_op(HVMOP_set_param, &xhp) != 0)
492                 panic("Can't set evtchn callback");
493 }
494
495 #define XEN_MAGIC_IOPORT 0x10
496 enum {
497         XMI_MAGIC                        = 0x49d2,
498         XMI_UNPLUG_IDE_DISKS             = 0x01,
499         XMI_UNPLUG_NICS                  = 0x02,
500         XMI_UNPLUG_IDE_EXCEPT_PRI_MASTER = 0x04
501 };
502
503 static void
504 xen_hvm_disable_emulated_devices(void)
505 {
506         if (inw(XEN_MAGIC_IOPORT) != XMI_MAGIC)
507                 return;
508
509         if (bootverbose)
510                 printf("XEN: Disabling emulated block and network devices\n");
511         outw(XEN_MAGIC_IOPORT, XMI_UNPLUG_IDE_DISKS|XMI_UNPLUG_NICS);
512 }
513
514 static void
515 xen_hvm_init(enum xen_hvm_init_type init_type)
516 {
517         int error;
518         int i;
519
520         if (init_type == XEN_HVM_INIT_CANCELLED_SUSPEND)
521                 return;
522
523         error = xen_hvm_init_hypercall_stubs();
524
525         switch (init_type) {
526         case XEN_HVM_INIT_COLD:
527                 if (error != 0)
528                         return;
529
530                 setup_xen_features();
531 #ifdef SMP
532                 cpu_ops = xen_hvm_cpu_ops;
533 #endif
534                 vm_guest = VM_GUEST_XEN;
535                 break;
536         case XEN_HVM_INIT_RESUME:
537                 if (error != 0)
538                         panic("Unable to init Xen hypercall stubs on resume");
539
540                 /* Clear stale vcpu_info. */
541                 CPU_FOREACH(i)
542                         DPCPU_ID_SET(i, vcpu_info, NULL);
543                 break;
544         default:
545                 panic("Unsupported HVM initialization type");
546         }
547
548         xen_vector_callback_enabled = 0;
549         xen_domain_type = XEN_HVM_DOMAIN;
550         xen_hvm_init_shared_info_page();
551         xen_hvm_set_callback(NULL);
552         xen_hvm_disable_emulated_devices();
553
554
555 void
556 xen_hvm_suspend(void)
557 {
558 }
559
560 void
561 xen_hvm_resume(bool suspend_cancelled)
562 {
563
564         xen_hvm_init(suspend_cancelled ?
565             XEN_HVM_INIT_CANCELLED_SUSPEND : XEN_HVM_INIT_RESUME);
566
567         /* Register vcpu_info area for CPU#0. */
568         xen_hvm_cpu_init();
569 }
570  
571 static void
572 xen_hvm_sysinit(void *arg __unused)
573 {
574         xen_hvm_init(XEN_HVM_INIT_COLD);
575 }
576
577 static void
578 xen_set_vcpu_id(void)
579 {
580         struct pcpu *pc;
581         int i;
582
583         /* Set vcpu_id to acpi_id */
584         CPU_FOREACH(i) {
585                 pc = pcpu_find(i);
586                 pc->pc_vcpu_id = pc->pc_acpi_id;
587                 if (bootverbose)
588                         printf("XEN: CPU %u has VCPU ID %u\n",
589                                i, pc->pc_vcpu_id);
590         }
591 }
592
593 static void
594 xen_hvm_cpu_init(void)
595 {
596         struct vcpu_register_vcpu_info info;
597         struct vcpu_info *vcpu_info;
598         int cpu, rc;
599
600         if (!xen_domain())
601                 return;
602
603         if (DPCPU_GET(vcpu_info) != NULL) {
604                 /*
605                  * vcpu_info is already set.  We're resuming
606                  * from a failed migration and our pre-suspend
607                  * configuration is still valid.
608                  */
609                 return;
610         }
611
612         vcpu_info = DPCPU_PTR(vcpu_local_info);
613         cpu = PCPU_GET(vcpu_id);
614         info.mfn = vtophys(vcpu_info) >> PAGE_SHIFT;
615         info.offset = vtophys(vcpu_info) - trunc_page(vtophys(vcpu_info));
616
617         rc = HYPERVISOR_vcpu_op(VCPUOP_register_vcpu_info, cpu, &info);
618         if (rc != 0)
619                 DPCPU_SET(vcpu_info, &HYPERVISOR_shared_info->vcpu_info[cpu]);
620         else
621                 DPCPU_SET(vcpu_info, vcpu_info);
622 }
623
624 SYSINIT(xen_hvm_init, SI_SUB_HYPERVISOR, SI_ORDER_FIRST, xen_hvm_sysinit, NULL);
625 #ifdef SMP
626 SYSINIT(xen_setup_cpus, SI_SUB_SMP, SI_ORDER_FIRST, xen_setup_cpus, NULL);
627 #endif
628 SYSINIT(xen_hvm_cpu_init, SI_SUB_INTR, SI_ORDER_FIRST, xen_hvm_cpu_init, NULL);
629 SYSINIT(xen_set_vcpu_id, SI_SUB_CPU, SI_ORDER_ANY, xen_set_vcpu_id, NULL);