/linux/arch/nds32/mm/ |
A D | cacheflush.c | 34 cpu_cache_wbinval_page(kaddr, vma->vm_flags & VM_EXEC); in flush_icache_page() 68 (vma->vm_flags & VM_EXEC)) { in update_mmu_cache() 72 cpu_cache_wbinval_page(kaddr, vma->vm_flags & VM_EXEC); in update_mmu_cache() 139 if (vma->vm_flags & VM_EXEC) in flush_cache_range() 146 cpu_cache_wbinval_page(start, vma->vm_flags & VM_EXEC); in flush_cache_range() 160 cpu_cache_wbinval_page(vto, vma->vm_flags & VM_EXEC); in flush_cache_page() 274 if (vma->vm_flags & VM_EXEC) { in copy_to_user_page() 310 if (vma->vm_flags & VM_EXEC) in flush_anon_page()
|
A D | fault.c | 175 mask = VM_EXEC; in do_page_fault() 189 mask = VM_EXEC; in do_page_fault()
|
/linux/arch/nios2/mm/ |
A D | cacheflush.c | 137 if (vma == NULL || (vma->vm_flags & VM_EXEC)) in flush_cache_range() 157 if (vma->vm_flags & VM_EXEC) in flush_cache_page() 229 if (vma->vm_flags & VM_EXEC) in update_mmu_cache() 260 if (vma->vm_flags & VM_EXEC) in copy_from_user_page() 271 if (vma->vm_flags & VM_EXEC) in copy_to_user_page()
|
/linux/arch/arm/mm/ |
A D | cache-v4wt.S | 66 mov r2, #VM_EXEC 69 tst r2, #VM_EXEC 90 tst r2, #VM_EXEC
|
A D | cache-fa.S | 64 mov r2, #VM_EXEC 67 tst r2, #VM_EXEC 90 1: tst r2, #VM_EXEC 96 tst r2, #VM_EXEC
|
A D | flush.c | 94 if (vma->vm_flags & VM_EXEC) in flush_cache_range() 110 if (vma->vm_flags & VM_EXEC && icache_is_vivt_asid_tagged()) in flush_cache_page() 165 if (vma->vm_flags & VM_EXEC) in flush_ptrace_access()
|
A D | proc-arm922.S | 129 mov r2, #VM_EXEC 139 tst r2, #VM_EXEC 161 tst r2, #VM_EXEC 166 tst r2, #VM_EXEC
|
A D | proc-feroceon.S | 143 mov r2, #VM_EXEC 155 tst r2, #VM_EXEC 176 1: tst r2, #VM_EXEC 185 tst r2, #VM_EXEC 473 @ if r2 contains the VM_EXEC bit then the next 2 ops are done already 474 tst r2, #VM_EXEC
|
A D | proc-arm920.S | 127 mov r2, #VM_EXEC 137 tst r2, #VM_EXEC 159 tst r2, #VM_EXEC 164 tst r2, #VM_EXEC
|
A D | proc-arm946.S | 99 mov r2, #VM_EXEC 113 tst r2, #VM_EXEC 135 1: tst r2, #VM_EXEC 153 tst r2, #VM_EXEC
|
A D | proc-mohawk.S | 111 mov r2, #VM_EXEC 115 tst r2, #VM_EXEC 137 1: tst r2, #VM_EXEC 146 tst r2, #VM_EXEC
|
A D | proc-arm925.S | 162 mov r2, #VM_EXEC 174 tst r2, #VM_EXEC 194 1: tst r2, #VM_EXEC 212 tst r2, #VM_EXEC
|
A D | proc-arm926.S | 128 mov r2, #VM_EXEC 137 tst r2, #VM_EXEC 157 1: tst r2, #VM_EXEC 175 tst r2, #VM_EXEC
|
A D | cache-v4wb.S | 111 tst r2, #VM_EXEC @ executable region? 122 tst r2, #VM_EXEC
|
A D | proc-arm1020e.S | 137 mov r2, #VM_EXEC 150 tst r2, #VM_EXEC 179 tst r2, #VM_EXEC
|
/linux/arch/csky/abiv1/ |
A D | cacheflush.c | 54 if (vma->vm_flags & VM_EXEC) in update_mmu_cache() 64 if (vma->vm_flags & VM_EXEC) in flush_cache_range()
|
/linux/tools/testing/selftests/bpf/progs/ |
A D | bpf_iter_task_vma.c | 11 #define VM_EXEC 0x00000004 macro 40 perm_str[2] = (vma->vm_flags & VM_EXEC) ? 'x' : '-'; in proc_maps()
|
/linux/arch/alpha/include/asm/ |
A D | tlbflush.h | 43 if (vma->vm_flags & VM_EXEC) { in ev4_flush_tlb_current_page() 55 if (vma->vm_flags & VM_EXEC) in ev5_flush_tlb_current_page()
|
/linux/arch/xtensa/mm/ |
A D | cache.c | 243 && (vma->vm_flags & VM_EXEC) != 0) { in update_mmu_cache() 290 if ((vma->vm_flags & VM_EXEC) != 0) in copy_to_user_page() 294 } else if ((vma->vm_flags & VM_EXEC) != 0) { in copy_to_user_page()
|
A D | tlb.c | 107 if (vma->vm_flags & VM_EXEC) in local_flush_tlb_range() 141 if (vma->vm_flags & VM_EXEC) in local_flush_tlb_page()
|
/linux/arch/mips/kernel/ |
A D | vdso.c | 103 VM_READ | VM_EXEC | in arch_setup_additional_pages() 179 VM_READ | VM_EXEC | in arch_setup_additional_pages()
|
/linux/arch/arm64/kernel/ |
A D | vdso.c | 247 VM_READ|VM_EXEC|gp_flags| in __setup_additional_pages() 379 VM_READ | VM_EXEC | in aarch32_kuser_helpers_setup() 402 VM_READ | VM_EXEC | VM_MAYREAD | in aarch32_sigreturn_setup()
|
/linux/arch/sparc/mm/ |
A D | fault_64.c | 398 if ((fault_code & FAULT_CODE_ITLB) && !(vma->vm_flags & VM_EXEC)) { in do_sparc64_fault() 413 (vma->vm_flags & VM_EXEC) != 0 && in do_sparc64_fault() 421 if (!(vma->vm_flags & (VM_READ | VM_EXEC))) in do_sparc64_fault()
|
/linux/arch/sh/mm/ |
A D | cache-sh4.c | 252 if (vma->vm_flags & VM_EXEC) in sh4_flush_cache_page() 294 if (vma->vm_flags & VM_EXEC) in sh4_flush_cache_range()
|
/linux/arch/x86/entry/vsyscall/ |
A D | vsyscall_64.c | 313 .vm_flags = VM_READ | VM_EXEC, 394 gate_vma.vm_flags = VM_EXEC; in map_vsyscall()
|