Searched refs:MSR_VEC (Results 1 – 19 of 19) sorted by relevance
/linux/arch/powerpc/kernel/ |
A D | process.c | 245 msr &= ~MSR_VEC; in __giveup_altivec() 255 msr_check_and_set(MSR_VEC); in giveup_altivec() 257 msr_check_and_clear(MSR_VEC); in giveup_altivec() 318 if (msr & MSR_VEC) in __giveup_vsx() 416 msr_all_available |= MSR_VEC; in init_msr_all_available() 446 if (usermsr & MSR_VEC) in giveup_all() 536 new_msr |= MSR_VEC; in restore_math() 539 if (((msr | new_msr) & (MSR_FP | MSR_VEC)) == (MSR_FP | MSR_VEC)) in restore_math() 555 if (new_msr & MSR_VEC) in restore_math() 587 if (usermsr & MSR_VEC) in save_all() [all …]
|
A D | signal_64.c | 142 msr |= MSR_VEC; in __unsafe_setup_sigcontext() 232 msr |= tsk->thread.ckpt_regs.msr & (MSR_FP | MSR_VEC | MSR_VSX); in setup_tm_sigcontexts() 246 if (msr & MSR_VEC) in setup_tm_sigcontexts() 258 msr |= MSR_VEC; in setup_tm_sigcontexts() 266 if (msr & MSR_VEC) in setup_tm_sigcontexts() 379 regs_set_return_msr(regs, regs->msr & ~(MSR_FP | MSR_FE0 | MSR_FE1 | MSR_VEC | MSR_VSX)); in __unsafe_restore_sigcontext() 386 if (v_regs != NULL && (msr & MSR_VEC) != 0) { in __unsafe_restore_sigcontext() 510 if (v_regs != NULL && tm_v_regs != NULL && (msr & MSR_VEC) != 0) { in restore_tm_sigcontexts() 591 msr_check_and_set(msr & (MSR_FP | MSR_VEC)); in restore_tm_sigcontexts() 596 if (msr & MSR_VEC) { in restore_tm_sigcontexts() [all …]
|
A D | signal_32.c | 276 msr |= MSR_VEC; in __unsafe_save_user_regs() 380 if (msr & MSR_VEC) in save_tm_user_regs_unsafe() 392 msr |= MSR_VEC; in save_tm_user_regs_unsafe() 402 if (msr & MSR_VEC) in save_tm_user_regs_unsafe() 489 regs_set_return_msr(regs, regs->msr & ~MSR_VEC); in restore_user_regs() 490 if (msr & MSR_VEC) { in restore_user_regs() 585 regs_set_return_msr(regs, regs->msr & ~MSR_VEC); in restore_tm_user_regs() 586 if (msr & MSR_VEC) { in restore_tm_user_regs() 630 if (msr & MSR_VEC) in restore_tm_user_regs() 687 msr_check_and_set(msr & (MSR_FP | MSR_VEC)); in restore_tm_user_regs() [all …]
|
A D | vector.S | 50 oris r5,r5,MSR_VEC@h 69 oris r9,r9,MSR_VEC@h 73 oris r12,r12,MSR_VEC@h 126 andis. r5,r12,MSR_VEC@h
|
A D | interrupt.c | 373 mathflags |= MSR_VEC | MSR_VSX; in interrupt_exit_user_prepare_main() 375 mathflags |= MSR_VEC; in interrupt_exit_user_prepare_main()
|
A D | tm.S | 139 oris r15, r15, MSR_VEC@h 392 oris r5, r5, MSR_VEC@h
|
A D | traps.c | 961 msr_mask = MSR_VEC; in p9_hmi_special_emu()
|
/linux/arch/powerpc/include/asm/ |
A D | switch_to.h | 60 msr_check_and_clear(MSR_VEC); in disable_kernel_altivec() 72 msr_check_and_clear(MSR_FP|MSR_VEC|MSR_VSX); in disable_kernel_vsx()
|
A D | reg.h | 88 #define MSR_VEC __MASK(MSR_VEC_LG) /* Enable AltiVec */ macro
|
/linux/arch/powerpc/kvm/ |
A D | book3s_pr.c | 173 kvmppc_giveup_ext(vcpu, MSR_FP | MSR_VEC | MSR_VSX); in kvmppc_core_vcpu_put_pr() 362 (MSR_FP | MSR_VEC | MSR_VSX); in kvmppc_handle_lost_math_exts() 369 else if (ext_diff == MSR_VEC) in kvmppc_handle_lost_math_exts() 808 msr |= MSR_FP | MSR_VEC; in kvmppc_giveup_ext() 830 if (msr & MSR_VEC) { in kvmppc_giveup_ext() 831 if (current->thread.regs->msr & MSR_VEC) in kvmppc_giveup_ext() 889 msr = MSR_FP | MSR_VEC | MSR_VSX; in kvmppc_handle_ext() 910 if (msr & MSR_VEC) { in kvmppc_handle_ext() 948 if (lost_ext & MSR_VEC) { in kvmppc_handle_lost_ext() 1365 ext_msr = MSR_VEC; in kvmppc_handle_exit_pr() [all …]
|
A D | emulate_loadstore.c | 54 if (!(kvmppc_get_msr(vcpu) & MSR_VEC)) { in kvmppc_check_altivec_disabled() 270 MSR_VEC); in kvmppc_emulate_loadstore()
|
A D | tm.S | 45 oris r8, r8, (MSR_VEC | MSR_VSX)@h 239 oris r5, r5, (MSR_VEC | MSR_VSX)@h
|
A D | booke.c | 202 if (!(current->thread.regs->msr & MSR_VEC)) { in kvmppc_load_guest_altivec() 207 current->thread.regs->msr |= MSR_VEC; in kvmppc_load_guest_altivec() 221 if (current->thread.regs->msr & MSR_VEC) in kvmppc_save_guest_altivec()
|
A D | book3s_hv_rmhandlers.S | 2464 oris r8,r8,MSR_VEC@h 2499 oris r8,r8,MSR_VEC@h
|
A D | powerpc.c | 1205 vcpu->kvm->arch.kvm_ops->giveup_ext(vcpu, MSR_VEC); in kvmppc_complete_mmio_load()
|
A D | book3s_hv.c | 3919 msr_check_and_set(MSR_FP | MSR_VEC | MSR_VSX); in kvmhv_p9_guest_entry() 4028 msr_check_and_set(MSR_FP | MSR_VEC | MSR_VSX); in kvmhv_p9_guest_entry()
|
/linux/arch/powerpc/lib/ |
A D | ldstfp.S | 72 oris r7, r6, MSR_VEC@h 96 oris r7, r6, MSR_VEC@h
|
A D | sstep.c | 698 if (regs->msr & MSR_VEC) in do_vec_load() 721 if (regs->msr & MSR_VEC) in do_vec_store() 987 if (regs->msr & MSR_VEC) { in do_vsx_load() 1033 if (regs->msr & MSR_VEC) { in do_vsx_store() 3447 if (!(regs->msr & MSR_PR) && !(regs->msr & MSR_VEC)) in emulate_loadstore() 3461 msrbit = MSR_VEC; in emulate_loadstore() 3518 if (!(regs->msr & MSR_PR) && !(regs->msr & MSR_VEC)) in emulate_loadstore() 3532 msrbit = MSR_VEC; in emulate_loadstore()
|
A D | test_emulate_step.c | 74 regs->msr |= MSR_VEC; in init_pt_regs()
|
Completed in 53 milliseconds