2 * SPDX-License-Identifier: BSD-3-Clause
4 * Copyright (c) 1989, 1992, 1993
5 * The Regents of the University of California. All rights reserved.
7 * This code is derived from software developed by the Computer Systems
8 * Engineering group at Lawrence Berkeley Laboratory under DARPA contract
9 * BG 91-66 and contributed to Berkeley.
11 * Redistribution and use in source and binary forms, with or without
12 * modification, are permitted provided that the following conditions
14 * 1. Redistributions of source code must retain the above copyright
15 * notice, this list of conditions and the following disclaimer.
16 * 2. Redistributions in binary form must reproduce the above copyright
17 * notice, this list of conditions and the following disclaimer in the
18 * documentation and/or other materials provided with the distribution.
19 * 3. Neither the name of the University nor the names of its contributors
20 * may be used to endorse or promote products derived from this software
21 * without specific prior written permission.
23 * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
24 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
25 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
26 * ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
27 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
28 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
29 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
30 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
31 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
32 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
36 #include <sys/cdefs.h>
37 __FBSDID("$FreeBSD$");
39 #if defined(LIBC_SCCS) && !defined(lint)
41 static char sccsid[] = "@(#)kvm_hp300.c 8.1 (Berkeley) 6/4/93";
43 #endif /* LIBC_SCCS and not lint */
46 * AMD64 machine dependent routines for kvm. Hopefully, the forthcoming
47 * vm code will one day obsolete this module.
50 #include <sys/param.h>
51 #include <sys/endian.h>
61 #include "kvm_private.h"
62 #include "kvm_amd64.h"
71 * Translate a physical memory address to a file-offset in the crash-dump.
74 _kvm_pa2off(kvm_t *kd, uint64_t pa, off_t *ofs)
76 struct vmstate *vm = kd->vmst;
82 return (AMD64_PAGE_SIZE - (pa & AMD64_PAGE_MASK));
87 while (n && (pa < p->p_paddr || pa >= p->p_paddr + p->p_memsz))
91 *ofs = (pa - p->p_paddr) + p->p_offset;
92 return (AMD64_PAGE_SIZE - (pa & AMD64_PAGE_MASK));
96 _amd64_freevtop(kvm_t *kd)
98 struct vmstate *vm = kd->vmst;
108 _amd64_probe(kvm_t *kd)
111 return (_kvm_probe_elf_kernel(kd, ELFCLASS64, EM_X86_64) &&
112 !_kvm_is_minidump(kd));
116 _amd64_initvtop(kvm_t *kd)
118 struct kvm_nlist nl[2];
123 kd->vmst = (struct vmstate *)_kvm_malloc(kd, sizeof(*kd->vmst));
124 if (kd->vmst == NULL) {
125 _kvm_err(kd, kd->program, "cannot allocate vm");
130 if (kd->rawdump == 0) {
131 if (_kvm_read_core_phdrs(kd, &kd->vmst->phnum,
132 &kd->vmst->phdr) == -1)
136 nl[0].n_name = "kernbase";
139 if (kvm_nlist2(kd, nl) != 0) {
140 _kvm_err(kd, kd->program, "bad namelist - no kernbase");
143 kernbase = nl[0].n_value;
145 nl[0].n_name = "KPML4phys";
148 if (kvm_nlist2(kd, nl) != 0) {
149 _kvm_err(kd, kd->program, "bad namelist - no KPML4phys");
152 if (kvm_read2(kd, (nl[0].n_value - kernbase), &pa, sizeof(pa)) !=
154 _kvm_err(kd, kd->program, "cannot read KPML4phys");
158 PML4 = _kvm_malloc(kd, AMD64_PAGE_SIZE);
160 _kvm_err(kd, kd->program, "cannot allocate PML4");
163 if (kvm_read2(kd, pa, PML4, AMD64_PAGE_SIZE) != AMD64_PAGE_SIZE) {
164 _kvm_err(kd, kd->program, "cannot read KPML4phys");
168 kd->vmst->PML4 = PML4;
173 _amd64_vatop(kvm_t *kd, kvaddr_t va, off_t *pa)
176 amd64_physaddr_t offset;
177 amd64_physaddr_t pdpe_pa;
178 amd64_physaddr_t pde_pa;
179 amd64_physaddr_t pte_pa;
193 offset = va & AMD64_PAGE_MASK;
196 * If we are initializing (kernel page table descriptor pointer
197 * not yet set) then return pa == va to avoid infinite recursion.
199 if (vm->PML4 == NULL) {
200 s = _kvm_pa2off(kd, va, pa);
202 _kvm_err(kd, kd->program,
203 "_amd64_vatop: bootstrap data not in dump");
206 return (AMD64_PAGE_SIZE - offset);
209 pml4eindex = (va >> AMD64_PML4SHIFT) & (AMD64_NPML4EPG - 1);
210 pml4e = le64toh(vm->PML4[pml4eindex]);
211 if ((pml4e & AMD64_PG_V) == 0) {
212 _kvm_err(kd, kd->program, "_amd64_vatop: pml4e not valid");
216 pdpeindex = (va >> AMD64_PDPSHIFT) & (AMD64_NPDPEPG - 1);
217 pdpe_pa = (pml4e & AMD64_PG_FRAME) + (pdpeindex * sizeof(amd64_pdpe_t));
219 s = _kvm_pa2off(kd, pdpe_pa, &ofs);
220 if (s < sizeof(pdpe)) {
221 _kvm_err(kd, kd->program, "_amd64_vatop: pdpe_pa not found");
224 if (pread(kd->pmfd, &pdpe, sizeof(pdpe), ofs) != sizeof(pdpe)) {
225 _kvm_syserr(kd, kd->program, "_amd64_vatop: read pdpe");
228 pdpe = le64toh(pdpe);
229 if ((pdpe & AMD64_PG_V) == 0) {
230 _kvm_err(kd, kd->program, "_amd64_vatop: pdpe not valid");
234 if (pdpe & AMD64_PG_PS) {
236 * No next-level page table; pdpe describes one 1GB page.
238 a = (pdpe & AMD64_PG_1GB_FRAME) + (va & AMD64_PDPMASK);
239 s = _kvm_pa2off(kd, a, pa);
241 _kvm_err(kd, kd->program,
242 "_amd64_vatop: 1GB page address not in dump");
245 return (AMD64_NBPDP - (va & AMD64_PDPMASK));
248 pdeindex = (va >> AMD64_PDRSHIFT) & (AMD64_NPDEPG - 1);
249 pde_pa = (pdpe & AMD64_PG_FRAME) + (pdeindex * sizeof(amd64_pde_t));
251 s = _kvm_pa2off(kd, pde_pa, &ofs);
252 if (s < sizeof(pde)) {
253 _kvm_syserr(kd, kd->program, "_amd64_vatop: pde_pa not found");
256 if (pread(kd->pmfd, &pde, sizeof(pde), ofs) != sizeof(pde)) {
257 _kvm_syserr(kd, kd->program, "_amd64_vatop: read pde");
261 if ((pde & AMD64_PG_V) == 0) {
262 _kvm_err(kd, kd->program, "_amd64_vatop: pde not valid");
266 if (pde & AMD64_PG_PS) {
268 * No final-level page table; pde describes one 2MB page.
270 a = (pde & AMD64_PG_PS_FRAME) + (va & AMD64_PDRMASK);
271 s = _kvm_pa2off(kd, a, pa);
273 _kvm_err(kd, kd->program,
274 "_amd64_vatop: 2MB page address not in dump");
277 return (AMD64_NBPDR - (va & AMD64_PDRMASK));
280 pteindex = (va >> AMD64_PAGE_SHIFT) & (AMD64_NPTEPG - 1);
281 pte_pa = (pde & AMD64_PG_FRAME) + (pteindex * sizeof(amd64_pte_t));
283 s = _kvm_pa2off(kd, pte_pa, &ofs);
284 if (s < sizeof(pte)) {
285 _kvm_err(kd, kd->program, "_amd64_vatop: pte_pa not found");
288 if (pread(kd->pmfd, &pte, sizeof(pte), ofs) != sizeof(pte)) {
289 _kvm_syserr(kd, kd->program, "_amd64_vatop: read");
292 if ((pte & AMD64_PG_V) == 0) {
293 _kvm_err(kd, kd->program, "_amd64_vatop: pte not valid");
297 a = (pte & AMD64_PG_FRAME) + offset;
298 s = _kvm_pa2off(kd, a, pa);
300 _kvm_err(kd, kd->program, "_amd64_vatop: address not in dump");
303 return (AMD64_PAGE_SIZE - offset);
306 _kvm_err(kd, 0, "invalid address (0x%jx)", (uintmax_t)va);
311 _amd64_kvatop(kvm_t *kd, kvaddr_t va, off_t *pa)
315 _kvm_err(kd, 0, "kvm_kvatop called in live kernel!");
318 return (_amd64_vatop(kd, va, pa));
322 _amd64_native(kvm_t *kd __unused)
332 static struct kvm_arch kvm_amd64 = {
333 .ka_probe = _amd64_probe,
334 .ka_initvtop = _amd64_initvtop,
335 .ka_freevtop = _amd64_freevtop,
336 .ka_kvatop = _amd64_kvatop,
337 .ka_native = _amd64_native,