2 * Copyright (c) 2013 Dmitry Chagin
5 * Redistribution and use in source and binary forms, with or without
6 * modification, are permitted provided that the following conditions
8 * 1. Redistributions of source code must retain the above copyright
9 * notice, this list of conditions and the following disclaimer
10 * in this position and unchanged.
11 * 2. Redistributions in binary form must reproduce the above copyright
12 * notice, this list of conditions and the following disclaimer in the
13 * documentation and/or other materials provided with the distribution.
15 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
16 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
17 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
18 * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
19 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
20 * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
21 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
22 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
23 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
24 * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
27 #include <sys/cdefs.h>
28 __FBSDID("$FreeBSD$");
30 #include "opt_compat.h"
32 #define __ELF_WORD_SIZE 32
34 #include <sys/param.h>
35 #include <sys/systm.h>
37 #include <sys/kernel.h>
39 #include <sys/rwlock.h>
40 #include <sys/queue.h>
41 #include <sys/sysent.h>
44 #include <vm/vm_param.h>
46 #include <vm/vm_extern.h>
47 #include <vm/vm_kern.h>
48 #include <vm/vm_map.h>
49 #include <vm/vm_object.h>
50 #include <vm/vm_page.h>
51 #include <vm/vm_pager.h>
53 #include <compat/linux/linux_vdso.h>
55 SLIST_HEAD(, linux_vdso_sym) __elfN(linux_vdso_syms) =
56 SLIST_HEAD_INITIALIZER(__elfN(linux_vdso_syms));
58 static int __elfN(symtabindex);
59 static int __elfN(symstrindex);
62 __elfN(linux_vdso_lookup)(Elf_Ehdr *, struct linux_vdso_sym *);
66 __elfN(linux_vdso_sym_init)(struct linux_vdso_sym *s)
69 SLIST_INSERT_HEAD(&__elfN(linux_vdso_syms), s, sym);
73 __elfN(linux_shared_page_init)(char **mapping)
79 obj = vm_pager_allocate(OBJT_PHYS, 0, PAGE_SIZE,
80 VM_PROT_DEFAULT, 0, NULL);
82 m = vm_page_grab(obj, 0, VM_ALLOC_NOBUSY | VM_ALLOC_ZERO);
83 m->valid = VM_PAGE_BITS_ALL;
84 VM_OBJECT_WUNLOCK(obj);
85 addr = kva_alloc(PAGE_SIZE);
86 pmap_qenter(addr, &m, 1);
87 *mapping = (char *)addr;
92 __elfN(linux_shared_page_fini)(vm_object_t obj)
95 vm_object_deallocate(obj);
99 __elfN(linux_vdso_fixup)(struct sysentvec *sv)
105 ehdr = (Elf_Ehdr *) sv->sv_sigcode;
107 if (!IS_ELF(*ehdr) ||
108 ehdr->e_ident[EI_CLASS] != ELF_TARG_CLASS ||
109 ehdr->e_ident[EI_DATA] != ELF_TARG_DATA ||
110 ehdr->e_ident[EI_VERSION] != EV_CURRENT ||
111 ehdr->e_shoff == 0 ||
112 ehdr->e_shentsize != sizeof(Elf_Shdr))
113 panic("Linux invalid vdso header.\n");
115 if (ehdr->e_type != ET_DYN)
116 panic("Linux invalid vdso header.\n");
118 shdr = (Elf_Shdr *) ((caddr_t)ehdr + ehdr->e_shoff);
120 __elfN(symtabindex) = -1;
121 __elfN(symstrindex) = -1;
122 for (i = 0; i < ehdr->e_shnum; i++) {
123 if (shdr[i].sh_size == 0)
125 if (shdr[i].sh_type == SHT_DYNSYM) {
126 __elfN(symtabindex) = i;
127 __elfN(symstrindex) = shdr[i].sh_link;
131 if (__elfN(symtabindex) == -1 || __elfN(symstrindex) == -1)
132 panic("Linux invalid vdso header.\n");
134 ehdr->e_ident[EI_OSABI] = ELFOSABI_LINUX;
138 __elfN(linux_vdso_reloc)(struct sysentvec *sv, int vdso_adjust)
140 struct linux_vdso_sym *lsym;
148 ehdr = (Elf_Ehdr *) sv->sv_sigcode;
150 /* Adjust our so relative to the sigcode_base */
151 if (vdso_adjust != 0) {
152 ehdr->e_entry += vdso_adjust;
153 phdr = (Elf_Phdr *)((caddr_t)ehdr + ehdr->e_phoff);
156 for (i = 0; i < ehdr->e_phnum; i++) {
157 phdr[i].p_vaddr += vdso_adjust;
158 if (phdr[i].p_type != PT_DYNAMIC)
160 dyn = (Elf_Dyn *)((caddr_t)ehdr + phdr[i].p_offset);
161 for(; dyn->d_tag != DT_NULL; dyn++) {
162 switch (dyn->d_tag) {
176 case DT_ADDRRNGLO ... DT_ADDRRNGHI:
177 dyn->d_un.d_ptr += vdso_adjust;
179 case DT_ENCODING ... DT_LOOS-1:
180 case DT_LOOS ... DT_HIOS:
181 if (dyn->d_tag >= DT_ENCODING &&
182 (dyn->d_tag & 1) == 0)
183 dyn->d_un.d_ptr += vdso_adjust;
192 shdr = (Elf_Shdr *)((caddr_t)ehdr + ehdr->e_shoff);
193 for(i = 0; i < ehdr->e_shnum; i++) {
194 if (!(shdr[i].sh_flags & SHF_ALLOC))
196 shdr[i].sh_addr += vdso_adjust;
197 if (shdr[i].sh_type != SHT_SYMTAB &&
198 shdr[i].sh_type != SHT_DYNSYM)
201 sym = (Elf_Sym *)((caddr_t)ehdr + shdr[i].sh_offset);
202 symcnt = shdr[i].sh_size / sizeof(*sym);
204 for(i = 0; i < symcnt; i++, sym++) {
205 if (sym->st_shndx == SHN_UNDEF ||
206 sym->st_shndx == SHN_ABS)
208 sym->st_value += vdso_adjust;
213 SLIST_FOREACH(lsym, &__elfN(linux_vdso_syms), sym)
214 __elfN(linux_vdso_lookup)(ehdr, lsym);
218 __elfN(linux_vdso_lookup)(Elf_Ehdr *ehdr, struct linux_vdso_sym *vsym)
220 vm_offset_t strtab, symname;
225 shdr = (Elf_Shdr *) ((caddr_t)ehdr + ehdr->e_shoff);
227 strtab = (vm_offset_t)((caddr_t)ehdr +
228 shdr[__elfN(symstrindex)].sh_offset);
229 Elf_Sym *sym = (Elf_Sym *)((caddr_t)ehdr +
230 shdr[__elfN(symtabindex)].sh_offset);
231 symcnt = shdr[__elfN(symtabindex)].sh_size / sizeof(*sym);
233 for (i = 0; i < symcnt; ++i, ++sym) {
234 symname = strtab + sym->st_name;
235 if (strncmp(vsym->symname, (char *)symname, vsym->size) == 0) {
236 *vsym->ptr = (uintptr_t)sym->st_value;