2 * Copyright (c) 2015 Nathan Whitehorn
3 * Copyright (c) 2017-2018 Semihalf
6 * Redistribution and use in source and binary forms, with or without
7 * modification, are permitted provided that the following conditions
10 * 1. Redistributions of source code must retain the above copyright
11 * notice, this list of conditions and the following disclaimer.
12 * 2. Redistributions in binary form must reproduce the above copyright
13 * notice, this list of conditions and the following disclaimer in the
14 * documentation and/or other materials provided with the distribution.
16 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
17 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
18 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
19 * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
20 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
21 * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
22 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
23 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
24 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
25 * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
28 #include <sys/cdefs.h>
29 __FBSDID("$FreeBSD$");
31 #include <sys/param.h>
32 #include <sys/systm.h>
33 #include <sys/kernel.h>
41 #include <machine/bus.h>
42 #include <machine/cpu.h>
43 #include <machine/hid.h>
44 #include <machine/platformvar.h>
45 #include <machine/pmap.h>
46 #include <machine/rtas.h>
47 #include <machine/smp.h>
48 #include <machine/spr.h>
49 #include <machine/trap.h>
51 #include <dev/ofw/openfirm.h>
52 #include <machine/ofw_machdep.h>
53 #include <powerpc/aim/mmu_oea64.h>
55 #include "platform_if.h"
62 void (*powernv_smp_ap_extra_init)(void);
64 static int powernv_probe(platform_t);
65 static int powernv_attach(platform_t);
66 void powernv_mem_regions(platform_t, struct mem_region *phys, int *physsz,
67 struct mem_region *avail, int *availsz);
68 static void powernv_numa_mem_regions(platform_t plat, struct numa_mem_region *phys, int *physsz);
69 static u_long powernv_timebase_freq(platform_t, struct cpuref *cpuref);
70 static int powernv_smp_first_cpu(platform_t, struct cpuref *cpuref);
71 static int powernv_smp_next_cpu(platform_t, struct cpuref *cpuref);
72 static int powernv_smp_get_bsp(platform_t, struct cpuref *cpuref);
73 static void powernv_smp_ap_init(platform_t);
75 static int powernv_smp_start_cpu(platform_t, struct pcpu *cpu);
76 static void powernv_smp_probe_threads(platform_t);
77 static struct cpu_group *powernv_smp_topo(platform_t plat);
79 static void powernv_reset(platform_t);
80 static void powernv_cpu_idle(sbintime_t sbt);
81 static int powernv_cpuref_init(void);
83 static platform_method_t powernv_methods[] = {
84 PLATFORMMETHOD(platform_probe, powernv_probe),
85 PLATFORMMETHOD(platform_attach, powernv_attach),
86 PLATFORMMETHOD(platform_mem_regions, powernv_mem_regions),
87 PLATFORMMETHOD(platform_numa_mem_regions, powernv_numa_mem_regions),
88 PLATFORMMETHOD(platform_timebase_freq, powernv_timebase_freq),
90 PLATFORMMETHOD(platform_smp_ap_init, powernv_smp_ap_init),
91 PLATFORMMETHOD(platform_smp_first_cpu, powernv_smp_first_cpu),
92 PLATFORMMETHOD(platform_smp_next_cpu, powernv_smp_next_cpu),
93 PLATFORMMETHOD(platform_smp_get_bsp, powernv_smp_get_bsp),
95 PLATFORMMETHOD(platform_smp_start_cpu, powernv_smp_start_cpu),
96 PLATFORMMETHOD(platform_smp_probe_threads, powernv_smp_probe_threads),
97 PLATFORMMETHOD(platform_smp_topo, powernv_smp_topo),
100 PLATFORMMETHOD(platform_reset, powernv_reset),
105 static platform_def_t powernv_platform = {
111 static struct cpuref platform_cpuref[MAXCPU];
112 static int platform_cpuref_cnt;
113 static int platform_cpuref_valid;
115 PLATFORM_DEF(powernv_platform);
117 static uint64_t powernv_boot_pir;
120 powernv_probe(platform_t plat)
122 if (opal_check() == 0)
123 return (BUS_PROBE_SPECIFIC);
129 powernv_attach(platform_t plat)
131 uint32_t nptlp, shift = 0, slb_encoding = 0;
132 int32_t lp_size, lp_encoding;
139 /* Ping OPAL again just to make sure */
142 #if BYTE_ORDER == LITTLE_ENDIAN
143 opal_call(OPAL_REINIT_CPUS, 2 /* Little endian */);
145 opal_call(OPAL_REINIT_CPUS, 1 /* Big endian */);
148 if (cpu_idle_hook == NULL)
149 cpu_idle_hook = powernv_cpu_idle;
151 powernv_boot_pir = mfspr(SPR_PIR);
153 /* LPID must not be altered when PSL_DR or PSL_IR is set */
155 mtmsr(msr & ~(PSL_DR | PSL_IR));
157 /* Direct interrupts to SRR instead of HSRR and reset LPCR otherwise */
161 if (cpu_features2 & PPC_FEATURE2_ARCH_3_00)
164 mtspr(SPR_LPCR, lpcr);
169 powernv_cpuref_init();
171 /* Set SLB count from device tree */
175 res = OF_getprop(cpu, "name", buf, sizeof(buf));
176 if (res > 0 && strcmp(buf, "cpus") == 0)
185 res = OF_getprop(cpu, "device_type", buf, sizeof(buf));
186 if (res > 0 && strcmp(buf, "cpu") == 0)
193 res = OF_getencprop(cpu, "ibm,slb-size", &prop, sizeof(prop));
198 * Scan the large page size property for PAPR compatible machines.
199 * See PAPR D.5 Changes to Section 5.1.4, 'CPU Node Properties'
200 * for the encoding of the property.
203 len = OF_getproplen(cpu, "ibm,segment-page-sizes");
206 * We have to use a variable length array on the stack
207 * since we have very limited stack space.
209 pcell_t arr[len/sizeof(cell_t)];
210 res = OF_getencprop(cpu, "ibm,segment-page-sizes", arr,
216 slb_encoding = arr[idx + 1];
217 nptlp = arr[idx + 2];
220 while (len > 0 && nptlp) {
222 lp_encoding = arr[idx+1];
223 if (slb_encoding == SLBV_L && lp_encoding == 0)
230 if (nptlp && slb_encoding == SLBV_L && lp_encoding == 0)
235 panic("Standard large pages (SLB[L] = 1, PTE[LP] = 0) "
236 "not supported by this system.");
238 moea64_large_page_shift = shift;
239 moea64_large_page_size = 1ULL << lp_size;
248 powernv_mem_regions(platform_t plat, struct mem_region *phys, int *physsz,
249 struct mem_region *avail, int *availsz)
252 ofw_mem_regions(phys, physsz, avail, availsz);
256 powernv_numa_mem_regions(platform_t plat, struct numa_mem_region *phys, int *physsz)
259 ofw_numa_mem_regions(phys, physsz);
263 powernv_timebase_freq(platform_t plat, struct cpuref *cpuref)
266 phandle_t cpu, dev, root;
271 dev = OF_child(root);
273 res = OF_getprop(dev, "name", buf, sizeof(buf));
274 if (res > 0 && strcmp(buf, "cpus") == 0)
279 for (cpu = OF_child(dev); cpu != 0; cpu = OF_peer(cpu)) {
280 res = OF_getprop(cpu, "device_type", buf, sizeof(buf));
281 if (res > 0 && strcmp(buf, "cpu") == 0)
287 OF_getencprop(cpu, "timebase-frequency", &ticks, sizeof(ticks));
290 panic("Unable to determine timebase frequency!");
297 powernv_cpuref_init(void)
301 int a, res, tmp_cpuref_cnt;
302 static struct cpuref tmp_cpuref[MAXCPU];
303 cell_t interrupt_servers[32];
306 if (platform_cpuref_valid)
312 res = OF_getprop(dev, "name", buf, sizeof(buf));
313 if (res > 0 && strcmp(buf, "cpus") == 0)
320 for (cpu = OF_child(dev); cpu != 0; cpu = OF_peer(cpu)) {
321 res = OF_getprop(cpu, "device_type", buf, sizeof(buf));
322 if (res > 0 && strcmp(buf, "cpu") == 0) {
323 res = OF_getproplen(cpu, "ibm,ppc-interrupt-server#s");
325 OF_getencprop(cpu, "ibm,ppc-interrupt-server#s",
326 interrupt_servers, res);
328 for (a = 0; a < res/sizeof(cell_t); a++) {
329 tmp_cpuref[tmp_cpuref_cnt].cr_hwref = interrupt_servers[a];
330 tmp_cpuref[tmp_cpuref_cnt].cr_cpuid = tmp_cpuref_cnt;
331 tmp_cpuref[tmp_cpuref_cnt].cr_domain = interrupt_servers[a] >> 11;
332 if (interrupt_servers[a] == (uint32_t)powernv_boot_pir)
333 bsp = tmp_cpuref_cnt;
341 /* Map IDs, so BSP has CPUID 0 regardless of hwref */
342 for (a = bsp; a < tmp_cpuref_cnt; a++) {
343 platform_cpuref[platform_cpuref_cnt].cr_hwref = tmp_cpuref[a].cr_hwref;
344 platform_cpuref[platform_cpuref_cnt].cr_cpuid = platform_cpuref_cnt;
345 platform_cpuref[platform_cpuref_cnt].cr_domain = tmp_cpuref[a].cr_domain;
346 platform_cpuref_cnt++;
348 for (a = 0; a < bsp; a++) {
349 platform_cpuref[platform_cpuref_cnt].cr_hwref = tmp_cpuref[a].cr_hwref;
350 platform_cpuref[platform_cpuref_cnt].cr_cpuid = platform_cpuref_cnt;
351 platform_cpuref[platform_cpuref_cnt].cr_domain = tmp_cpuref[a].cr_domain;
352 platform_cpuref_cnt++;
355 platform_cpuref_valid = 1;
361 powernv_smp_first_cpu(platform_t plat, struct cpuref *cpuref)
363 if (platform_cpuref_valid == 0)
366 cpuref->cr_cpuid = 0;
367 cpuref->cr_hwref = platform_cpuref[0].cr_hwref;
368 cpuref->cr_domain = platform_cpuref[0].cr_domain;
374 powernv_smp_next_cpu(platform_t plat, struct cpuref *cpuref)
378 if (platform_cpuref_valid == 0)
381 id = cpuref->cr_cpuid + 1;
382 if (id >= platform_cpuref_cnt)
385 cpuref->cr_cpuid = platform_cpuref[id].cr_cpuid;
386 cpuref->cr_hwref = platform_cpuref[id].cr_hwref;
387 cpuref->cr_domain = platform_cpuref[id].cr_domain;
393 powernv_smp_get_bsp(platform_t plat, struct cpuref *cpuref)
396 cpuref->cr_cpuid = platform_cpuref[0].cr_cpuid;
397 cpuref->cr_hwref = platform_cpuref[0].cr_hwref;
398 cpuref->cr_domain = platform_cpuref[0].cr_domain;
404 powernv_smp_start_cpu(platform_t plat, struct pcpu *pc)
411 result = opal_call(OPAL_START_CPU, pc->pc_hwref, EXC_RST);
412 if (result != OPAL_SUCCESS) {
413 printf("OPAL error (%d): unable to start AP %d\n",
414 result, (int)pc->pc_hwref);
422 powernv_smp_probe_threads(platform_t plat)
425 phandle_t cpu, dev, root;
430 dev = OF_child(root);
432 res = OF_getprop(dev, "name", buf, sizeof(buf));
433 if (res > 0 && strcmp(buf, "cpus") == 0)
439 for (cpu = OF_child(dev); cpu != 0; cpu = OF_peer(cpu)) {
440 res = OF_getprop(cpu, "device_type", buf, sizeof(buf));
441 if (res <= 0 || strcmp(buf, "cpu") != 0)
444 res = OF_getproplen(cpu, "ibm,ppc-interrupt-server#s");
447 nthreads = res / sizeof(cell_t);
453 smp_threads_per_core = nthreads;
454 if (mp_ncpus % nthreads == 0)
455 mp_ncores = mp_ncpus / nthreads;
458 static struct cpu_group *
459 powernv_smp_topo(platform_t plat)
461 if (mp_ncpus % smp_threads_per_core != 0) {
462 printf("WARNING: Irregular SMP topology. Performance may be "
463 "suboptimal (%d threads, %d on first core)\n",
464 mp_ncpus, smp_threads_per_core);
465 return (smp_topo_none());
468 /* Don't do anything fancier for non-threaded SMP */
469 if (smp_threads_per_core == 1)
470 return (smp_topo_none());
472 return (smp_topo_1level(CG_SHARE_L1, smp_threads_per_core,
479 powernv_reset(platform_t platform)
482 opal_call(OPAL_CEC_REBOOT);
486 powernv_smp_ap_init(platform_t platform)
489 if (powernv_smp_ap_extra_init != NULL)
490 powernv_smp_ap_extra_init();
494 powernv_cpu_idle(sbintime_t sbt)