/linux/arch/arm64/mm/ |
A D | fault.c | 70 if (esr & ESR_ELx_ISV) { in data_abort_decode() 94 ESR_ELx_EC(esr), esr_get_class_string(esr), in mem_abort_decode() 106 data_abort_decode(esr); in mem_abort_decode() 247 if (!is_el1_data_abort(esr) && !is_el1_instruction_abort(esr)) in is_el1_permission_fault() 261 unsigned int esr, in is_spurious_el1_translation_fault() argument 300 mem_abort_decode(esr); in die_kernel_fault() 303 die("Oops", regs, esr); in die_kernel_fault() 443 esr = 0; in set_thread_esr() 513 return (esr & ESR_ELx_WNR) && !(esr & ESR_ELx_CM); in is_write_abort() 566 addr, esr, regs); in do_page_fault() [all …]
|
/linux/arch/arm64/kernel/ |
A D | entry-common.c | 404 el1_pc(regs, esr); in el1h_64_sync_handler() 414 el1_dbg(regs, esr); in el1h_64_sync_handler() 624 el0_da(regs, esr); in el0t_64_sync_handler() 627 el0_ia(regs, esr); in el0t_64_sync_handler() 640 el0_sys(regs, esr); in el0t_64_sync_handler() 643 el0_sp(regs, esr); in el0t_64_sync_handler() 646 el0_pc(regs, esr); in el0t_64_sync_handler() 658 el0_dbg(regs, esr); in el0t_64_sync_handler() 749 el0_da(regs, esr); in el0t_32_sync_handler() 752 el0_ia(regs, esr); in el0t_32_sync_handler() [all …]
|
A D | traps.c | 256 if (esr) in arm64_show_signal() 257 pr_cont("%s, ESR 0x%08x, ", esr_get_class_string(esr), esr); in arm64_show_signal() 615 rt = ESR_ELx_SYS64_ISS_RT(esr); in mrs_handler() 684 if (!(esr & ESR_ELx_CV)) { in cp15_cond_valid() 754 switch (ESR_ELx_EC(esr)) { in do_cp15instr() 768 hook->handler(esr, regs); in do_cp15instr() 788 hook->handler(esr, regs); in do_sysinstr() 880 pr_emerg("ESR: 0x%08x -- %s\n", esr, esr_get_class_string(esr)); in panic_bad_stack() 906 smp_processor_id(), esr, esr_get_class_string(esr)); in arm64_serror_panic() 951 if (!arm64_is_ras_serror(esr) || arm64_is_fatal_ras_serror(regs, esr)) in do_serror() [all …]
|
A D | debug-monitors.c | 205 static int call_step_hook(struct pt_regs *regs, unsigned int esr) in call_step_hook() argument 218 retval = hook->fn(regs, esr); in call_step_hook() 241 static int single_step_handler(unsigned long unused, unsigned int esr, in single_step_handler() argument 253 if (!handler_found && call_step_hook(regs, esr) == DBG_HOOK_HANDLED) in single_step_handler() 302 static int call_break_hook(struct pt_regs *regs, unsigned int esr) in call_break_hook() argument 306 int (*fn)(struct pt_regs *regs, unsigned int esr) = NULL; in call_break_hook() 315 unsigned int comment = esr & ESR_ELx_BRK64_ISS_COMMENT_MASK; in call_break_hook() 321 return fn ? fn(regs, esr) : DBG_HOOK_ERROR; in call_break_hook() 325 static int brk_handler(unsigned long unused, unsigned int esr, in brk_handler() argument 328 if (call_break_hook(regs, esr) == DBG_HOOK_HANDLED) in brk_handler()
|
A D | fpsimd.c | 991 void do_sve_acc(unsigned int esr, struct pt_regs *regs) in do_sve_acc() argument 1033 void do_fpsimd_acc(unsigned int esr, struct pt_regs *regs) in do_fpsimd_acc() argument 1042 void do_fpsimd_exc(unsigned int esr, struct pt_regs *regs) in do_fpsimd_exc() argument 1046 if (esr & ESR_ELx_FP_EXC_TFV) { in do_fpsimd_exc() 1047 if (esr & FPEXC_IOF) in do_fpsimd_exc() 1049 else if (esr & FPEXC_DZF) in do_fpsimd_exc() 1051 else if (esr & FPEXC_OFF) in do_fpsimd_exc() 1053 else if (esr & FPEXC_UFF) in do_fpsimd_exc() 1055 else if (esr & FPEXC_IXF) in do_fpsimd_exc()
|
A D | kgdb.c | 235 static int kgdb_brk_fn(struct pt_regs *regs, unsigned int esr) in kgdb_brk_fn() argument 242 static int kgdb_compiled_brk_fn(struct pt_regs *regs, unsigned int esr) in NOKPROBE_SYMBOL() 251 static int kgdb_step_brk_fn(struct pt_regs *regs, unsigned int esr) in kgdb_step_brk_fn() argument
|
/linux/arch/arm64/include/asm/ |
A D | exception.h | 24 unsigned int esr = ESR_ELx_EC_SERROR << ESR_ELx_EC_SHIFT; in disr_to_esr() local 27 esr |= (disr & DISR_EL1_ESR_MASK); in disr_to_esr() 29 esr |= (disr & ESR_ELx_ISS_MASK); in disr_to_esr() 31 return esr; in disr_to_esr() 65 void do_fpsimd_acc(unsigned int esr, struct pt_regs *regs); 66 void do_sve_acc(unsigned int esr, struct pt_regs *regs); 67 void do_fpsimd_exc(unsigned int esr, struct pt_regs *regs); 68 void do_sysinstr(unsigned int esr, struct pt_regs *regs); 71 void do_cp15instr(unsigned int esr, struct pt_regs *regs); 74 void do_ptrauth_fault(struct pt_regs *regs, unsigned int esr); [all …]
|
A D | traps.h | 60 static inline bool arm64_is_ras_serror(u32 esr) in arm64_is_ras_serror() argument 64 if (esr & ESR_ELx_IDS) in arm64_is_ras_serror() 80 static inline u32 arm64_ras_serror_get_severity(u32 esr) in arm64_ras_serror_get_severity() argument 82 u32 aet = esr & ESR_ELx_AET; in arm64_ras_serror_get_severity() 84 if (!arm64_is_ras_serror(esr)) { in arm64_ras_serror_get_severity() 93 if ((esr & ESR_ELx_FSC) != ESR_ELx_FSC_SERROR) { in arm64_ras_serror_get_severity() 101 bool arm64_is_fatal_ras_serror(struct pt_regs *regs, unsigned int esr); 102 void __noreturn arm64_serror_panic(struct pt_regs *regs, u32 esr);
|
A D | esr.h | 73 #define ESR_ELx_EC(esr) (((esr) & ESR_ELx_EC_MASK) >> ESR_ELx_EC_SHIFT) argument 189 #define ESR_ELx_SYS64_ISS_RT(esr) \ argument 190 (((esr) & ESR_ELx_SYS64_ISS_RT_MASK) >> ESR_ELx_SYS64_ISS_RT_SHIFT) 333 static inline bool esr_is_data_abort(u32 esr) in esr_is_data_abort() argument 335 const u32 ec = ESR_ELx_EC(esr); in esr_is_data_abort() 340 const char *esr_get_class_string(u32 esr);
|
A D | kvm_emulate.h | 232 u32 esr = kvm_vcpu_get_esr(vcpu); in kvm_vcpu_get_condition() local 234 if (esr & ESR_ELx_CV) in kvm_vcpu_get_condition() 235 return (esr & ESR_ELx_COND_MASK) >> ESR_ELx_COND_SHIFT; in kvm_vcpu_get_condition() 363 u32 esr = kvm_vcpu_get_esr(vcpu); in kvm_vcpu_sys_get_rt() local 364 return ESR_ELx_SYS64_ISS_RT(esr); in kvm_vcpu_sys_get_rt()
|
A D | debug-monitors.h | 79 int (*fn)(struct pt_regs *regs, unsigned int esr); 90 int (*fn)(struct pt_regs *regs, unsigned int esr);
|
A D | kvm_ras.h | 17 static inline int kvm_handle_guest_sea(phys_addr_t addr, unsigned int esr) in kvm_handle_guest_sea() argument
|
/linux/arch/arm64/kvm/ |
A D | handle_exit.c | 29 static void kvm_handle_guest_serror(struct kvm_vcpu *vcpu, u32 esr) in kvm_handle_guest_serror() argument 31 if (!arm64_is_ras_serror(esr) || arm64_is_fatal_ras_serror(NULL, esr)) in kvm_handle_guest_serror() 121 u32 esr = kvm_vcpu_get_esr(vcpu); in kvm_handle_guest_debug() local 124 run->debug.arch.hsr = esr; in kvm_handle_guest_debug() 126 if (ESR_ELx_EC(esr) == ESR_ELx_EC_WATCHPT_LOW) in kvm_handle_guest_debug() 134 u32 esr = kvm_vcpu_get_esr(vcpu); in kvm_handle_unknown_ec() local 137 esr, esr_get_class_string(esr)); in kvm_handle_unknown_ec() 188 u32 esr = kvm_vcpu_get_esr(vcpu); in kvm_get_exit_handler() local 189 u8 esr_ec = ESR_ELx_EC(esr); in kvm_get_exit_handler() 297 } else if (ESR_ELx_EC(esr) == ESR_ELx_EC_BRK64 && in nvhe_hyp_panic_handler() [all …]
|
A D | inject_fault.c | 21 u32 esr = 0; in inject_abt64() local 34 esr |= ESR_ELx_IL; in inject_abt64() 41 esr |= (ESR_ELx_EC_IABT_LOW << ESR_ELx_EC_SHIFT); in inject_abt64() 43 esr |= (ESR_ELx_EC_IABT_CUR << ESR_ELx_EC_SHIFT); in inject_abt64() 46 esr |= ESR_ELx_EC_DABT_LOW << ESR_ELx_EC_SHIFT; in inject_abt64() 48 vcpu_write_sys_reg(vcpu, esr | ESR_ELx_FSC_EXTABT, ESR_EL1); in inject_abt64() 53 u32 esr = (ESR_ELx_EC_UNKNOWN << ESR_ELx_EC_SHIFT); in inject_undef64() local 64 esr |= ESR_ELx_IL; in inject_undef64() 66 vcpu_write_sys_reg(vcpu, esr, ESR_EL1); in inject_undef64() 163 void kvm_set_sei_esr(struct kvm_vcpu *vcpu, u64 esr) in kvm_set_sei_esr() argument [all …]
|
A D | sys_regs.h | 30 #define esr_sys64_to_params(esr) \ argument 31 ((struct sys_reg_params){ .Op0 = ((esr) >> 20) & 3, \ 32 .Op1 = ((esr) >> 14) & 0x7, \ 33 .CRn = ((esr) >> 10) & 0xf, \ 34 .CRm = ((esr) >> 1) & 0xf, \ 35 .Op2 = ((esr) >> 17) & 0x7, \ 36 .is_write = !((esr) & 1) })
|
/linux/arch/alpha/kernel/ |
A D | core_lca.c | 341 ((esr & ESR_CEE) ? "Correctable" : in mem_error() 342 (esr & ESR_UEE) ? "Uncorrectable" : "A"), in mem_error() 343 (esr & ESR_WRE) ? "write" : "read", in mem_error() 346 if (esr & ESR_CTE) { in mem_error() 349 if (esr & ESR_MSE) { in mem_error() 352 if (esr & ESR_MHE) { in mem_error() 355 if (esr & ESR_NXM) { in mem_error() 435 if (el.s->esr & ESR_EAV) { in lca_machine_check() 436 mem_error(el.s->esr, el.s->ear); in lca_machine_check() 450 if (el.l->esr & ESR_EAV) { in lca_machine_check() [all …]
|
/linux/arch/sh/boards/mach-dreamcast/ |
A D | irq.c | 90 __u32 esr = ESR_BASE + (LEVEL(irq) << 2); in mask_ack_systemasic_irq() local 92 outl((1 << EVENT_BIT(irq)), esr); in mask_ack_systemasic_irq() 107 __u32 emr, esr, status, level; in systemasic_irq_demux() local 124 esr = ESR_BASE + (level << 2); in systemasic_irq_demux() 127 status = inl(esr); in systemasic_irq_demux()
|
/linux/arch/arm64/kvm/hyp/include/hyp/ |
A D | fault.h | 44 static inline bool __get_fault_info(u64 esr, struct kvm_vcpu_fault_info *fault) in __get_fault_info() argument 61 if (!(esr & ESR_ELx_S1PTW) && in __get_fault_info() 63 (esr & ESR_ELx_FSC_TYPE) == FSC_PERM)) { in __get_fault_info()
|
/linux/drivers/net/ethernet/ibm/emac/ |
A D | mal.c | 220 u32 esr = get_mal_dcrn(mal, MAL_ESR); in mal_serr() local 223 set_mal_dcrn(mal, MAL_ESR, esr); in mal_serr() 225 MAL_DBG(mal, "SERR %08x" NL, esr); in mal_serr() 227 if (esr & MAL_ESR_EVB) { in mal_serr() 228 if (esr & MAL_ESR_DE) { in mal_serr() 235 if (esr & MAL_ESR_PEIN) { in mal_serr() 243 mal->index, esr); in mal_serr() 253 mal->index, esr); in mal_serr() 355 if (esr & MAL_ESR_EVB) { in mal_int() 357 if (esr & MAL_ESR_DE) { in mal_int() [all …]
|
/linux/sound/soc/fsl/ |
A D | fsl_esai.c | 102 u32 esr; in esai_isr() local 121 if (esr & ESAI_ESR_RFF_MASK) in esai_isr() 124 if (esr & ESAI_ESR_TFE_MASK) in esai_isr() 127 if (esr & ESAI_ESR_TLS_MASK) in esai_isr() 130 if (esr & ESAI_ESR_TDE_MASK) in esai_isr() 133 if (esr & ESAI_ESR_TED_MASK) in esai_isr() 136 if (esr & ESAI_ESR_TD_MASK) in esai_isr() 139 if (esr & ESAI_ESR_RLS_MASK) in esai_isr() 142 if (esr & ESAI_ESR_RDE_MASK) in esai_isr() 145 if (esr & ESAI_ESR_RED_MASK) in esai_isr() [all …]
|
/linux/arch/powerpc/platforms/44x/ |
A D | machine_check.c | 14 unsigned long reason = regs->esr; in machine_check_440A() 51 unsigned long reason = regs->esr; in machine_check_47x()
|
/linux/arch/arm64/kvm/hyp/nvhe/ |
A D | sys_regs.c | 35 u32 esr = (ESR_ELx_EC_UNKNOWN << ESR_ELx_EC_SHIFT); in inject_undef64() local 46 write_sysreg_el1(esr, SYS_ESR); in inject_undef64() 449 unsigned long esr = kvm_vcpu_get_esr(vcpu); in kvm_handle_pvm_sysreg() local 452 params = esr_sys64_to_params(esr); in kvm_handle_pvm_sysreg()
|
/linux/arch/arm64/kvm/hyp/ |
A D | vgic-v3-sr.c | 476 u32 esr = kvm_vcpu_get_esr(vcpu); in __vgic_v3_get_group() local 477 u8 crm = (esr & ESR_ELx_SYS64_ISS_CRM_MASK) >> ESR_ELx_SYS64_ISS_CRM_SHIFT; in __vgic_v3_get_group() 1016 u32 esr; in __vgic_v3_perform_cpuif_access() local 1022 esr = kvm_vcpu_get_esr(vcpu); in __vgic_v3_perform_cpuif_access() 1029 sysreg = esr_cp15_to_sysreg(esr); in __vgic_v3_perform_cpuif_access() 1031 sysreg = esr_sys64_to_sysreg(esr); in __vgic_v3_perform_cpuif_access() 1034 is_read = (esr & ESR_ELx_SYS64_ISS_DIR_MASK) == ESR_ELx_SYS64_ISS_DIR_READ; in __vgic_v3_perform_cpuif_access()
|
/linux/arch/mips/include/asm/octeon/ |
A D | cvmx-sli-defs.h | 107 __BITFIELD_FIELD(uint64_t esr:2, 119 __BITFIELD_FIELD(uint64_t esr:2,
|
/linux/arch/arm64/kernel/probes/ |
A D | uprobes.c | 169 unsigned int esr) in uprobe_breakpoint_handler() argument 178 unsigned int esr) in uprobe_single_step_handler() argument
|