Searched refs:MSR_LE (Results 1 – 22 of 22) sorted by relevance
/linux/arch/powerpc/kvm/ |
A D | book3s_hv_tm_builtin.c | 82 newmsr = (newmsr & ~MSR_LE) | (msr & MSR_LE); in kvmhv_p9_tm_emulation_early()
|
A D | book3s_hv_tm.c | 126 newmsr = (newmsr & ~MSR_LE) | (msr & MSR_LE); in kvmhv_p9_tm_emulation()
|
A D | book3s_pr.c | 232 smsr &= MSR_FE0 | MSR_FE1 | MSR_SF | MSR_SE | MSR_BE | MSR_LE | in kvmppc_recalc_shadow_msr() 235 smsr &= MSR_FE0 | MSR_FE1 | MSR_SF | MSR_SE | MSR_BE | MSR_LE; in kvmppc_recalc_shadow_msr() 1546 if (vcpu->arch.intr_msr & MSR_LE) in kvmppc_get_one_reg_pr() 1629 vcpu->arch.intr_msr |= MSR_LE; in kvmppc_set_lpcr_pr() 1631 vcpu->arch.intr_msr &= ~MSR_LE; in kvmppc_set_lpcr_pr() 1774 vcpu->arch.shadow_msr = MSR_USER64 & ~MSR_LE; in kvmppc_core_vcpu_create_pr()
|
A D | book3s_emulate.c | 247 if ((kvmppc_get_msr(vcpu) & MSR_LE) && in kvmppc_core_emulate_op_pr()
|
A D | powerpc.c | 185 if (vcpu->arch.intr_msr & MSR_LE) in kvmppc_kvm_pv()
|
A D | book3s_hv.c | 2002 vcpu->arch.intr_msr |= MSR_LE; in kvmppc_set_lpcr() 2004 vcpu->arch.intr_msr &= ~MSR_LE; in kvmppc_set_lpcr()
|
/linux/tools/testing/selftests/powerpc/tm/ |
A D | tm-trap.c | 48 #define MSR_LE 1UL macro 67 thread_endianness = MSR_LE & ucp->uc_mcontext.gp_regs[PT_MSR]; in trap_signal_handler()
|
/linux/arch/powerpc/kernel/ |
A D | signal_32.c | 482 regs_set_return_msr(regs, (regs->msr & ~MSR_LE) | (msr & MSR_LE)); in restore_user_regs() 583 regs_set_return_msr(regs, (regs->msr & ~MSR_LE) | (msr & MSR_LE)); in restore_tm_user_regs() 806 regs_set_return_msr(regs, (regs->msr & ~MSR_LE) | (MSR_KERNEL & MSR_LE)); in handle_rt_signal32() 894 regs_set_return_msr(regs, (regs->msr & ~MSR_LE) | (MSR_KERNEL & MSR_LE)); in handle_signal32()
|
A D | signal_64.c | 357 regs_set_return_msr(regs, (regs->msr & ~MSR_LE) | (msr & MSR_LE)); in __unsafe_restore_sigcontext() 471 regs_set_return_msr(regs, (regs->msr & ~MSR_LE) | (msr & MSR_LE)); in restore_tm_sigcontexts() 947 regs_set_return_msr(regs, (regs->msr & ~MSR_LE) | (MSR_KERNEL & MSR_LE)); in handle_rt_signal64()
|
A D | align.c | 314 if ((regs->msr & MSR_LE) != (MSR_KERNEL & MSR_LE)) { in fix_alignment()
|
A D | syscalls.c | 111 current->thread.regs->msr ^ MSR_LE); in SYSCALL_DEFINE0()
|
A D | entry_64.S | 339 ori r9,r9,MSR_IR|MSR_DR|MSR_FE0|MSR_FE1|MSR_FP|MSR_RI|MSR_LE 450 LOAD_REG_IMMEDIATE(r12, MSR_SF | MSR_LE)
|
A D | process.c | 1425 {MSR_LE, "LE"}, 2006 regs_set_return_msr(regs, regs->msr & ~MSR_LE); in set_endian() 2008 regs_set_return_msr(regs, regs->msr | MSR_LE); in set_endian() 2027 if (regs->msr & MSR_LE) { in get_endian()
|
A D | traps.c | 903 swap = (msr & MSR_LE) != (MSR_KERNEL & MSR_LE); in p9_hmi_special_emu()
|
A D | exceptions-64s.S | 1961 xori r12,r12,MSR_LE
|
/linux/arch/powerpc/boot/ |
A D | ppc_asm.h | 63 #define MSR_LE 0x0000000000000001 macro
|
A D | opal-calls.S | 43 li r11,MSR_LE
|
/linux/arch/powerpc/platforms/powernv/ |
A D | opal-wrappers.S | 27 li r0,MSR_IR|MSR_DR|MSR_LE
|
/linux/arch/powerpc/include/asm/ |
A D | kvm_book3s.h | 401 return (kvmppc_get_msr(vcpu) & MSR_LE) != (MSR_KERNEL & MSR_LE); in kvmppc_need_byteswap()
|
A D | reg.h | 113 #define MSR_LE __MASK(MSR_LE_LG) /* Little Endian */ macro 139 #define MSR_ (__MSR | MSR_LE) 140 #define MSR_IDLE (MSR_ME | MSR_SF | MSR_HV | MSR_LE)
|
/linux/arch/powerpc/platforms/pseries/ |
A D | ras.c | 488 (MSR_LE|MSR_RI|MSR_DR|MSR_IR|MSR_ME|MSR_PR| in pSeries_system_reset_exception()
|
/linux/arch/powerpc/lib/ |
A D | sstep.c | 1758 op->val = 0xffffffff & ~(MSR_ME | MSR_LE); in analyse_instr() 3336 cross_endian = (regs->msr & MSR_LE) != (MSR_KERNEL & MSR_LE); in emulate_loadstore() 3662 regs_set_return_msr(regs, regs->msr ^ MSR_LE); in emulate_step()
|
Completed in 52 milliseconds