Lines Matching refs:kp
125 static int can_optimize(struct kprobe *kp) in can_optimize() argument
127 if (kp->ainsn.stack_space < 0) in can_optimize()
133 if (kp->ainsn.stack_space > 255 - sizeof(struct pt_regs)) in can_optimize()
154 struct kprobe *p = &op->kp; in optimized_callback()
158 regs->ARM_pc = (unsigned long)op->kp.addr; in optimized_callback()
165 kprobes_inc_nmissed_count(&op->kp); in optimized_callback()
167 __this_cpu_write(current_kprobe, &op->kp); in optimized_callback()
169 opt_pre_handler(&op->kp, regs); in optimized_callback()
178 op->kp.ainsn.insn_singlestep(p->opcode, &p->ainsn, regs); in optimized_callback()
266 (unsigned long)(op->kp.addr) + 4); in NOKPROBE_SYMBOL()
297 WARN_ON(kprobe_disabled(&op->kp)); in arch_optimize_kprobes()
303 memcpy(op->optinsn.copied_insn, op->kp.addr, in arch_optimize_kprobes()
306 insn = arm_gen_branch((unsigned long)op->kp.addr, in arch_optimize_kprobes()
323 kprobes_remove_breakpoint(op->kp.addr, insn); in arch_optimize_kprobes()
331 arch_arm_kprobe(&op->kp); in arch_unoptimize_kprobe()
352 return (op->kp.addr <= addr && in arch_within_optimized_kprobe()
353 op->kp.addr + (RELATIVEJUMP_SIZE / sizeof(kprobe_opcode_t)) > addr); in arch_within_optimized_kprobe()