/linux/arch/hexagon/kernel/ |
A D | vdso.c | 52 unsigned long vdso_base; in arch_setup_additional_pages() local 59 vdso_base = STACK_TOP; in arch_setup_additional_pages() 61 vdso_base = get_unmapped_area(NULL, vdso_base, PAGE_SIZE, 0, 0); in arch_setup_additional_pages() 62 if (IS_ERR_VALUE(vdso_base)) { in arch_setup_additional_pages() 63 ret = vdso_base; in arch_setup_additional_pages() 68 ret = install_special_mapping(mm, vdso_base, PAGE_SIZE, in arch_setup_additional_pages() 76 mm->context.vdso = (void *)vdso_base; in arch_setup_additional_pages()
|
/linux/arch/nds32/kernel/ |
A D | vdso.c | 117 unsigned long vdso_base, vdso_text_len, vdso_mapping_len; in arch_setup_additional_pages() local 137 vdso_base = get_unmapped_area(NULL, addr, vdso_mapping_len, 0, 0); in arch_setup_additional_pages() 138 if (IS_ERR_VALUE(vdso_base)) { in arch_setup_additional_pages() 139 ret = vdso_base; in arch_setup_additional_pages() 149 (vdso_base & aliasing_mask); in arch_setup_additional_pages() 150 vdso_base += page_colour_ofs & aliasing_mask; in arch_setup_additional_pages() 162 ret = io_remap_pfn_range(vma, vdso_base, in arch_setup_additional_pages() 169 vdso_base += PAGE_SIZE; in arch_setup_additional_pages() 177 vdso_base += PAGE_SIZE; in arch_setup_additional_pages() 178 mm->context.vdso = (void *)vdso_base; in arch_setup_additional_pages() [all …]
|
/linux/arch/csky/kernel/ |
A D | vdso.c | 59 unsigned long vdso_base, vdso_len; in arch_setup_additional_pages() local 65 vdso_base = get_unmapped_area(NULL, 0, vdso_len, 0, 0); in arch_setup_additional_pages() 66 if (IS_ERR_VALUE(vdso_base)) { in arch_setup_additional_pages() 67 ret = vdso_base; in arch_setup_additional_pages() 76 mm->context.vdso = (void *)vdso_base; in arch_setup_additional_pages() 79 install_special_mapping(mm, vdso_base, vdso_pages << PAGE_SHIFT, in arch_setup_additional_pages() 88 vdso_base += (vdso_pages << PAGE_SHIFT); in arch_setup_additional_pages() 89 ret = install_special_mapping(mm, vdso_base, PAGE_SIZE, in arch_setup_additional_pages()
|
/linux/arch/powerpc/kernel/ |
A D | vdso.c | 195 unsigned long vdso_size, vdso_base, mappings_size; in __arch_setup_additional_pages() local 204 vdso_base = VDSO32_MBASE; in __arch_setup_additional_pages() 213 vdso_base = 0; in __arch_setup_additional_pages() 225 vdso_base = get_unmapped_area(NULL, vdso_base, mappings_size, 0, 0); in __arch_setup_additional_pages() 226 if (IS_ERR_VALUE(vdso_base)) in __arch_setup_additional_pages() 227 return vdso_base; in __arch_setup_additional_pages() 230 vdso_base = ALIGN(vdso_base, VDSO_ALIGNMENT); in __arch_setup_additional_pages() 237 mm->context.vdso = (void __user *)vdso_base + vvar_size; in __arch_setup_additional_pages() 239 vma = _install_special_mapping(mm, vdso_base, vvar_size, in __arch_setup_additional_pages() 255 vma = _install_special_mapping(mm, vdso_base + vvar_size, vdso_size, in __arch_setup_additional_pages() [all …]
|
/linux/arch/riscv/kernel/ |
A D | vdso.c | 219 unsigned long vdso_base, vdso_text_len, vdso_mapping_len; in __setup_additional_pages() local 228 vdso_base = get_unmapped_area(NULL, 0, vdso_mapping_len, 0, 0); in __setup_additional_pages() 229 if (IS_ERR_VALUE(vdso_base)) { in __setup_additional_pages() 230 ret = ERR_PTR(vdso_base); in __setup_additional_pages() 234 ret = _install_special_mapping(mm, vdso_base, VVAR_SIZE, in __setup_additional_pages() 239 vdso_base += VVAR_SIZE; in __setup_additional_pages() 240 mm->context.vdso = (void *)vdso_base; in __setup_additional_pages() 242 _install_special_mapping(mm, vdso_base, vdso_text_len, in __setup_additional_pages()
|
/linux/arch/arm64/kernel/ |
A D | vdso.c | 219 unsigned long vdso_base, vdso_text_len, vdso_mapping_len; in __setup_additional_pages() local 229 vdso_base = get_unmapped_area(NULL, 0, vdso_mapping_len, 0, 0); in __setup_additional_pages() 230 if (IS_ERR_VALUE(vdso_base)) { in __setup_additional_pages() 231 ret = ERR_PTR(vdso_base); in __setup_additional_pages() 235 ret = _install_special_mapping(mm, vdso_base, VVAR_NR_PAGES * PAGE_SIZE, in __setup_additional_pages() 244 vdso_base += VVAR_NR_PAGES * PAGE_SIZE; in __setup_additional_pages() 245 mm->context.vdso = (void *)vdso_base; in __setup_additional_pages() 246 ret = _install_special_mapping(mm, vdso_base, vdso_text_len, in __setup_additional_pages()
|
/linux/arch/s390/include/asm/ |
A D | vdso.h | 14 #define VDSO64_SYMBOL(tsk, name) ((tsk)->mm->context.vdso_base + (vdso64_offset_##name)) 16 #define VDSO32_SYMBOL(tsk, name) ((tsk)->mm->context.vdso_base + (vdso32_offset_##name))
|
A D | mmu.h | 18 unsigned long vdso_base; member
|
A D | elf.h | 294 (unsigned long)current->mm->context.vdso_base); \
|
/linux/arch/powerpc/include/asm/ |
A D | mmu_context.h | 271 unsigned long vdso_base = (unsigned long)mm->context.vdso; in arch_unmap() local 273 if (start <= vdso_base && vdso_base < end) in arch_unmap()
|
/linux/arch/mips/kernel/ |
A D | vdso.c | 72 static unsigned long vdso_base(void) in vdso_base() function 131 base = get_unmapped_area(NULL, vdso_base(), size, 0, 0); in arch_setup_additional_pages()
|
/linux/arch/s390/kernel/ |
A D | vdso.c | 137 current->mm->context.vdso_base = vma->vm_start; in vdso_mremap() 205 current->mm->context.vdso_base = vdso_text_start; in arch_setup_additional_pages()
|