2 * SPDX-License-Identifier: BSD-3-Clause AND BSD-4-Clause
4 * Copyright (C) 2006 Semihalf, Marian Balakowicz <m8@semihalf.com>
7 * Adapted for Freescale's e500 core CPUs.
9 * Redistribution and use in source and binary forms, with or without
10 * modification, are permitted provided that the following conditions
12 * 1. Redistributions of source code must retain the above copyright
13 * notice, this list of conditions and the following disclaimer.
14 * 2. Redistributions in binary form must reproduce the above copyright
15 * notice, this list of conditions and the following disclaimer in the
16 * documentation and/or other materials provided with the distribution.
17 * 3. The name of the author may not be used to endorse or promote products
18 * derived from this software without specific prior written permission.
20 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
21 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
22 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN
23 * NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
24 * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED
25 * TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR
26 * PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF
27 * LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING
28 * NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
29 * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
34 * Copyright (C) 1995, 1996 Wolfgang Solfrank.
35 * Copyright (C) 1995, 1996 TooLs GmbH.
36 * All rights reserved.
38 * Redistribution and use in source and binary forms, with or without
39 * modification, are permitted provided that the following conditions
41 * 1. Redistributions of source code must retain the above copyright
42 * notice, this list of conditions and the following disclaimer.
43 * 2. Redistributions in binary form must reproduce the above copyright
44 * notice, this list of conditions and the following disclaimer in the
45 * documentation and/or other materials provided with the distribution.
46 * 3. All advertising materials mentioning features or use of this software
47 * must display the following acknowledgement:
48 * This product includes software developed by TooLs GmbH.
49 * 4. The name of TooLs GmbH may not be used to endorse or promote products
50 * derived from this software without specific prior written permission.
52 * THIS SOFTWARE IS PROVIDED BY TOOLS GMBH ``AS IS'' AND ANY EXPRESS OR
53 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
54 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
55 * IN NO EVENT SHALL TOOLS GMBH BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
56 * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO,
57 * PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS;
58 * OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY,
59 * WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR
60 * OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF
61 * ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
63 * from: $NetBSD: pmap.h,v 1.17 2000/03/30 16:18:24 jdolecek Exp $
66 #ifndef _MACHINE_PMAP_H_
67 #define _MACHINE_PMAP_H_
69 #include <sys/queue.h>
71 #include <sys/_cpuset.h>
72 #include <sys/_lock.h>
73 #include <sys/_mutex.h>
74 #include <machine/sr.h>
75 #include <machine/pte.h>
76 #include <machine/slb.h>
77 #include <machine/tlb.h>
78 #include <machine/vmparam.h>
81 typedef struct pmap *pmap_t;
83 #define PMAP_ENTER_QUICK_LOCKED 0x10000000
87 #endif /* !defined(NPMAPS) */
92 LIST_ENTRY(pvo_entry) pvo_vlink; /* Link to common virt page */
94 LIST_ENTRY(pvo_entry) pvo_olink; /* Link to overflow entry */
97 RB_ENTRY(pvo_entry) pvo_plink; /* Link to pmap entries */
98 SLIST_ENTRY(pvo_entry) pvo_dlink; /* Link to delete enty */
101 #ifndef __powerpc64__
110 pmap_t pvo_pmap; /* Owning pmap */
111 vm_offset_t pvo_vaddr; /* VA of entry */
112 uint64_t pvo_vpn; /* Virtual page number */
114 LIST_HEAD(pvo_head, pvo_entry);
115 SLIST_HEAD(pvo_dlist, pvo_entry);
116 RB_HEAD(pvo_tree, pvo_entry);
117 int pvo_vaddr_compare(struct pvo_entry *, struct pvo_entry *);
118 RB_PROTOTYPE(pvo_tree, pvo_entry, pvo_plink, pvo_vaddr_compare);
120 /* Used by 32-bit PMAP */
121 #define PVO_PTEGIDX_MASK 0x007UL /* which PTEG slot */
122 #define PVO_PTEGIDX_VALID 0x008UL /* slot is valid */
123 /* Used by 64-bit PMAP */
124 #define PVO_HID 0x008UL /* PVO entry in alternate hash*/
126 #define PVO_WIRED 0x010UL /* PVO entry is wired */
127 #define PVO_MANAGED 0x020UL /* PVO entry is managed */
128 #define PVO_BOOTSTRAP 0x080UL /* PVO entry allocated during
130 #define PVO_DEAD 0x100UL /* waiting to be deleted */
131 #define PVO_LARGE 0x200UL /* large page */
132 #define PVO_VADDR(pvo) ((pvo)->pvo_vaddr & ~ADDR_POFF)
133 #define PVO_PTEGIDX_GET(pvo) ((pvo)->pvo_vaddr & PVO_PTEGIDX_MASK)
134 #define PVO_PTEGIDX_ISSET(pvo) ((pvo)->pvo_vaddr & PVO_PTEGIDX_VALID)
135 #define PVO_PTEGIDX_CLR(pvo) \
136 ((void)((pvo)->pvo_vaddr &= ~(PVO_PTEGIDX_VALID|PVO_PTEGIDX_MASK)))
137 #define PVO_PTEGIDX_SET(pvo, i) \
138 ((void)((pvo)->pvo_vaddr |= (i)|PVO_PTEGIDX_VALID))
139 #define PVO_VSID(pvo) ((pvo)->pvo_vpn >> 16)
142 struct pmap_statistics pm_stats;
149 struct slbtnode *pm_slb_tree_root;
153 register_t pm_sr[16];
156 struct pmap *pmap_phys;
157 struct pvo_tree pmap_pvo;
160 /* TID to identify this pmap entries in TLB */
161 tlbtid_t pm_tid[MAXCPU];
165 * Page table directory,
166 * array of pointers to page directories.
171 * Page table directory,
172 * array of pointers to page tables.
176 /* List of allocated ptbl bufs (ptbl kva regions). */
177 TAILQ_HEAD(, ptbl_buf) pm_ptbl_list;
186 TAILQ_ENTRY(pv_entry) pv_link;
188 typedef struct pv_entry *pv_entry_t;
193 volatile int32_t mdpg_attrs;
194 vm_memattr_t mdpg_cache_attrs;
195 struct pvo_head mdpg_pvoh;
198 TAILQ_HEAD(, pv_entry) pv_list;
205 #define pmap_page_get_memattr(m) ((m)->md.mdpg_cache_attrs)
206 #define pmap_page_is_mapped(m) (!LIST_EMPTY(&(m)->md.mdpg_pvoh))
208 #define pmap_page_get_memattr(m) VM_MEMATTR_DEFAULT
209 #define pmap_page_is_mapped(m) (!TAILQ_EMPTY(&(m)->md.pv_list))
213 * Return the VSID corresponding to a given virtual address.
214 * If no VSID is currently defined, it will allocate one, and add
215 * it to a free slot if available.
217 * NB: The PMAP MUST be locked already.
219 uint64_t va_to_vsid(pmap_t pm, vm_offset_t va);
221 /* Lock-free, non-allocating lookup routines */
222 uint64_t kernel_va_to_slbv(vm_offset_t va);
223 struct slb *user_va_to_slb_entry(pmap_t pm, vm_offset_t va);
225 uint64_t allocate_user_vsid(pmap_t pm, uint64_t esid, int large);
226 void free_vsid(pmap_t pm, uint64_t esid, int large);
227 void slb_insert_user(pmap_t pm, struct slb *slb);
228 void slb_insert_kernel(uint64_t slbe, uint64_t slbv);
230 struct slbtnode *slb_alloc_tree(void);
231 void slb_free_tree(pmap_t pm);
232 struct slb **slb_alloc_user_cache(void);
233 void slb_free_user_cache(struct slb **);
235 extern struct pmap kernel_pmap_store;
236 #define kernel_pmap (&kernel_pmap_store)
240 #define PMAP_LOCK(pmap) mtx_lock(&(pmap)->pm_mtx)
241 #define PMAP_LOCK_ASSERT(pmap, type) \
242 mtx_assert(&(pmap)->pm_mtx, (type))
243 #define PMAP_LOCK_DESTROY(pmap) mtx_destroy(&(pmap)->pm_mtx)
244 #define PMAP_LOCK_INIT(pmap) mtx_init(&(pmap)->pm_mtx, \
245 (pmap == kernel_pmap) ? "kernelpmap" : \
246 "pmap", NULL, MTX_DEF)
247 #define PMAP_LOCKED(pmap) mtx_owned(&(pmap)->pm_mtx)
248 #define PMAP_MTX(pmap) (&(pmap)->pm_mtx)
249 #define PMAP_TRYLOCK(pmap) mtx_trylock(&(pmap)->pm_mtx)
250 #define PMAP_UNLOCK(pmap) mtx_unlock(&(pmap)->pm_mtx)
252 #define pmap_page_is_write_mapped(m) (((m)->a.flags & PGA_WRITEABLE) != 0)
254 void pmap_bootstrap(vm_offset_t, vm_offset_t);
255 void pmap_kenter(vm_offset_t va, vm_paddr_t pa);
256 void pmap_kenter_attr(vm_offset_t va, vm_paddr_t pa, vm_memattr_t);
257 void pmap_kremove(vm_offset_t);
258 void *pmap_mapdev(vm_paddr_t, vm_size_t);
259 void *pmap_mapdev_attr(vm_paddr_t, vm_size_t, vm_memattr_t);
260 void pmap_unmapdev(vm_offset_t, vm_size_t);
261 void pmap_page_set_memattr(vm_page_t, vm_memattr_t);
262 int pmap_change_attr(vm_offset_t, vm_size_t, vm_memattr_t);
263 int pmap_map_user_ptr(pmap_t pm, volatile const void *uaddr,
264 void **kaddr, size_t ulen, size_t *klen);
265 int pmap_decode_kernel_ptr(vm_offset_t addr, int *is_user,
266 vm_offset_t *decoded_addr);
267 void pmap_deactivate(struct thread *);
268 vm_paddr_t pmap_kextract(vm_offset_t);
269 int pmap_dev_direct_mapped(vm_paddr_t, vm_size_t);
270 boolean_t pmap_mmu_install(char *name, int prio);
271 const char *pmap_mmu_name(void);
273 void pmap_page_array_startup(long count);
275 #define vtophys(va) pmap_kextract((vm_offset_t)(va))
277 extern vm_offset_t virtual_avail;
278 extern vm_offset_t virtual_end;
279 extern caddr_t crashdumpmap;
281 extern vm_offset_t msgbuf_phys;
283 extern int pmap_bootstrapped;
285 vm_offset_t pmap_early_io_map(vm_paddr_t pa, vm_size_t size);
286 void pmap_early_io_unmap(vm_offset_t va, vm_size_t size);
287 void pmap_track_page(pmap_t pmap, vm_offset_t va);
290 pmap_vmspace_copy(pmap_t dst_pmap __unused, pmap_t src_pmap __unused)
298 #endif /* !_MACHINE_PMAP_H_ */