/linux/arch/parisc/include/asm/ |
A D | cache.h | 43 ALTERNATIVE(ALT_COND_NO_SMP, INSN_PxTLB) \ 46 ALTERNATIVE(ALT_COND_NO_SMP, INSN_PxTLB) \ 47 ALTERNATIVE(ALT_COND_NO_SPLIT_TLB, INSN_NOP) \ 50 ALTERNATIVE(ALT_COND_NO_SMP, INSN_PxTLB) \ 54 ALTERNATIVE(ALT_COND_NO_DCACHE, INSN_NOP) \ 55 ALTERNATIVE(ALT_COND_NO_IOC_FDC, INSN_NOP) \ 58 ALTERNATIVE(ALT_COND_NO_DCACHE, INSN_NOP) \ 59 ALTERNATIVE(ALT_COND_NO_IOC_FDC, INSN_NOP) :::"memory")
|
A D | alternative.h | 36 #define ALTERNATIVE(cond, replacement) "!0:" \ macro 45 #define ALTERNATIVE(from, to, cond, replacement)\ macro
|
/linux/arch/x86/include/asm/ |
A D | smap.h | 25 ALTERNATIVE "", __ASM_CLAC, X86_FEATURE_SMAP 28 ALTERNATIVE "", __ASM_STAC, X86_FEATURE_SMAP 58 ALTERNATIVE("", "pushf; pop %0; " __ASM_CLAC "\n\t", in smap_save() 68 ALTERNATIVE("", "push %0; popf\n\t", in smap_restore() 75 ALTERNATIVE("", __ASM_CLAC, X86_FEATURE_SMAP) 77 ALTERNATIVE("", __ASM_STAC, X86_FEATURE_SMAP)
|
A D | barrier.h | 15 #define mb() asm volatile(ALTERNATIVE("lock; addl $0,-4(%%esp)", "mfence", \ 17 #define rmb() asm volatile(ALTERNATIVE("lock; addl $0,-4(%%esp)", "lfence", \ 19 #define wmb() asm volatile(ALTERNATIVE("lock; addl $0,-4(%%esp)", "sfence", \
|
A D | alternative.h | 160 #define ALTERNATIVE(oldinstr, newinstr, feature) \ macro 211 asm_inline volatile (ALTERNATIVE(oldinstr, newinstr, feature) : : : "memory") 228 asm_inline volatile (ALTERNATIVE(oldinstr, newinstr, feature) \ 247 asm_inline volatile (ALTERNATIVE(oldinstr, newinstr, feature) \ 252 asm_inline volatile (ALTERNATIVE("call %P[old]", "call %P[new]", feature) \ 327 .macro ALTERNATIVE oldinstr, newinstr, feature
|
A D | arch_hweight.h | 19 asm (ALTERNATIVE("call __sw_hweight32", "popcntl %1, %0", X86_FEATURE_POPCNT) in __arch_hweight32() 47 asm (ALTERNATIVE("call __sw_hweight64", "popcntq %1, %0", X86_FEATURE_POPCNT) in __arch_hweight64()
|
A D | nospec-branch.h | 109 ALTERNATIVE "jmp .Lskip_rsb_\@", "", \ftr 219 asm volatile(ALTERNATIVE("", "wrmsr", %c[feature])
|
/linux/arch/arm64/include/asm/ |
A D | irqflags.h | 35 asm volatile(ALTERNATIVE( in arch_local_irq_enable() 54 asm volatile(ALTERNATIVE( in arch_local_irq_disable() 70 asm volatile(ALTERNATIVE( in arch_local_save_flags() 85 asm volatile(ALTERNATIVE( in arch_irqs_disabled_flags() 122 asm volatile(ALTERNATIVE( in arch_local_irq_restore()
|
A D | arch_timer.h | 69 asm volatile(ALTERNATIVE("isb\n mrs %0, cntpct_el0", in arch_timer_read_cntpct_el0() 81 asm volatile(ALTERNATIVE("isb\n mrs %0, cntvct_el0", in arch_timer_read_cntvct_el0() 189 asm volatile(ALTERNATIVE("isb\n mrs %0, cntpct_el0", in __arch_counter_get_cntpct() 210 asm volatile(ALTERNATIVE("isb\n mrs %0, cntvct_el0", in __arch_counter_get_cntvct()
|
A D | io.h | 52 asm volatile(ALTERNATIVE("ldrb %w0, [%1]", in __raw_readb() 64 asm volatile(ALTERNATIVE("ldrh %w0, [%1]", in __raw_readw() 75 asm volatile(ALTERNATIVE("ldr %w0, [%1]", in __raw_readl() 86 asm volatile(ALTERNATIVE("ldr %0, [%1]", in __raw_readq()
|
A D | uaccess.h | 148 asm(ALTERNATIVE("nop", SET_PSTATE_PAN(0), ARM64_HAS_PAN, in __uaccess_disable_hw_pan() 154 asm(ALTERNATIVE("nop", SET_PSTATE_PAN(1), ARM64_HAS_PAN, in __uaccess_enable_hw_pan() 177 asm volatile(ALTERNATIVE("nop", SET_PSTATE_TCO(0), in __uaccess_disable_tco() 183 asm volatile(ALTERNATIVE("nop", SET_PSTATE_TCO(1), in __uaccess_enable_tco()
|
A D | kvm_hyp.h | 22 asm volatile(ALTERNATIVE(__mrs_s("%0", r##nvh), \ 32 asm volatile(ALTERNATIVE(__msr_s(r##nvh, "%x0"), \
|
A D | tlbflush.h | 33 ALTERNATIVE("nop\n nop", \ 41 ALTERNATIVE("nop\n nop", \
|
A D | lse.h | 37 ALTERNATIVE(llsc, __LSE_PREAMBLE lse, ARM64_HAS_LSE_ATOMICS)
|
/linux/arch/x86/entry/ |
A D | calling.h | 167 ALTERNATIVE "", "SET_NOFLUSH_BIT \reg", X86_FEATURE_PCID 173 ALTERNATIVE "jmp .Lend_\@", "", X86_FEATURE_PTI 184 ALTERNATIVE "jmp .Lend_\@", "", X86_FEATURE_PTI 187 ALTERNATIVE "jmp .Lwrcr3_\@", "", X86_FEATURE_PCID 224 ALTERNATIVE "jmp .Ldone_\@", "", X86_FEATURE_PTI 242 ALTERNATIVE "jmp .Lend_\@", "", X86_FEATURE_PTI 244 ALTERNATIVE "jmp .Lwrcr3_\@", "", X86_FEATURE_PCID 303 ALTERNATIVE "", "lfence", X86_FEATURE_FENCE_SWAPGS_USER 306 ALTERNATIVE "", "lfence", X86_FEATURE_FENCE_SWAPGS_KERNEL
|
A D | entry_64.S | 122 ALTERNATIVE "", "jmp swapgs_restore_regs_and_return_to_usermode", \ 143 ALTERNATIVE "shl $(64 - 48), %rcx; sar $(64 - 48), %rcx", \ 578 ALTERNATIVE "", "jmp xenpv_restore_regs_and_return_to_usermode", X86_FEATURE_XENPV 738 2: ALTERNATIVE "", "mfence", X86_BUG_SWAPGS_FENCE 755 ALTERNATIVE "", "ZAP_GS", X86_BUG_NULL_SEG 880 ALTERNATIVE "jmp .Lparanoid_entry_checkgs", "", X86_FEATURE_FSGSBASE 948 ALTERNATIVE "jmp .Lparanoid_exit_checkgs", "", X86_FEATURE_FSGSBASE 1375 ALTERNATIVE "jmp nmi_no_fsgsbase", "", X86_FEATURE_FSGSBASE
|
A D | entry_64_compat.S | 141 ALTERNATIVE "testl %eax, %eax; jz swapgs_restore_regs_and_return_to_usermode", \ 258 ALTERNATIVE "testl %eax, %eax; jz swapgs_restore_regs_and_return_to_usermode", \ 366 ALTERNATIVE "", "jmp .Lint80_keep_stack", X86_FEATURE_XENPV
|
A D | entry_32.S | 58 ALTERNATIVE "jmp .Lend_\@", "", X86_FEATURE_PTI 68 ALTERNATIVE "jmp .Lend_\@", "", X86_FEATURE_PTI 89 ALTERNATIVE "jmp .Lend_\@", "", X86_FEATURE_PTI 293 ALTERNATIVE "jmp .Lswitched_\@", "", X86_FEATURE_PTI 313 ALTERNATIVE "jmp .Lend_\@", "", X86_BUG_ESPFIX
|
/linux/arch/riscv/include/asm/ |
A D | errata_list.h | 20 ALTERNATIVE(__stringify(RISCV_PTR do_trap_insn_fault), \ 26 ALTERNATIVE(__stringify(RISCV_PTR do_page_fault), \ 33 asm(ALTERNATIVE("sfence.vma %0", "sfence.vma", SIFIVE_VENDOR_ID, \
|
/linux/arch/s390/kernel/ |
A D | entry.S | 56 ALTERNATIVE "", ".insn s,0xb2010000,\address", 193 60 ALTERNATIVE "", ".insn s,0xb2000000,\address", 193 64 ALTERNATIVE "b \lpswe", ".insn siy,0xeb0000000071,\address,0", 193 68 ALTERNATIVE "", __stringify(mvc __PT_LAST_BREAK(8,\reg),__LC_LAST_BREAK), 193 102 ALTERNATIVE ".insn s,0xb2050000,\savearea", \ 129 ALTERNATIVE "", ".long 0xb2e8c000", 82 133 ALTERNATIVE "", ".long 0xb2e8d000", 82 137 ALTERNATIVE "TSTMSK \tif_ptr,\tif_mask; jz .+8; .long 0xb2e8d000", \ 143 ALTERNATIVE "jz .+8; .long 0xb2e8c000", \ 235 ALTERNATIVE "", ".insn s,0xb2800000,_LPP_OFFSET", 40
|
/linux/arch/s390/include/asm/ |
A D | alternative.h | 109 #define ALTERNATIVE(oldinstr, altinstr, facility) \ macro 142 asm_inline volatile(ALTERNATIVE(oldinstr, altinstr, facility) : : : "memory") 150 asm_inline volatile (ALTERNATIVE(oldinstr, newinstr, feature) \ 155 asm_inline volatile(ALTERNATIVE(oldinstr, altinstr, facility) \
|
A D | nospec-insn.h | 111 # The ALTERNATIVE replacement code has a .+10 which targets 118 ALTERNATIVE "ex %r0,_LC_BR_R1", ".insn ril,0xc60000000000,0,.+10", 35
|
/linux/arch/parisc/kernel/ |
A D | pacache.S | 106 ALTERNATIVE(88b, fitdone, ALT_COND_NO_SPLIT_TLB, INSN_NOP) 244 89: ALTERNATIVE(88b, 89b, ALT_COND_NO_ICACHE, INSN_NOP) 306 89: ALTERNATIVE(88b, 89b, ALT_COND_NO_DCACHE, INSN_NOP) 570 ALTERNATIVE(0b, 0b+4, ALT_COND_NO_SMP, INSN_PxTLB) 571 ALTERNATIVE(1b, 1b+4, ALT_COND_NO_SMP, INSN_PxTLB) 709 ALTERNATIVE(0b, 0b+4, ALT_COND_NO_SMP, INSN_PxTLB) 785 ALTERNATIVE(0b, 0b+4, ALT_COND_NO_SMP, INSN_PxTLB) 817 89: ALTERNATIVE(88b, 89b, ALT_COND_NO_DCACHE, INSN_NOP) 844 ALTERNATIVE(0b, 0b+4, ALT_COND_NO_SMP, INSN_PxTLB) 909 ALTERNATIVE(0b, 0b+4, ALT_COND_NO_SMP, INSN_PxTLB) [all …]
|
/linux/arch/x86/lib/ |
A D | getuser.S | 38 #define ASM_BARRIER_NOSPEC ALTERNATIVE "", "lfence", X86_FEATURE_LFENCE_RDTSC 42 ALTERNATIVE __stringify(mov $((1 << 47) - 4096 - (n)),%rdx), \
|
A D | memmove_64.S | 42 ALTERNATIVE "cmp $0x20, %rdx; jb 1f", "", X86_FEATURE_FSRM 43 ALTERNATIVE "", "movq %rdx, %rcx; rep movsb; retq", X86_FEATURE_ERMS
|