Lines Matching refs:kp
138 static int can_optimize(struct kprobe *kp) in can_optimize() argument
140 if (kp->ainsn.stack_space < 0) in can_optimize()
146 if (kp->ainsn.stack_space > 255 - sizeof(struct pt_regs)) in can_optimize()
167 struct kprobe *p = &op->kp; in optimized_callback()
171 regs->ARM_pc = (unsigned long)op->kp.addr; in optimized_callback()
177 kprobes_inc_nmissed_count(&op->kp); in optimized_callback()
179 __this_cpu_write(current_kprobe, &op->kp); in optimized_callback()
181 opt_pre_handler(&op->kp, regs); in optimized_callback()
190 op->kp.ainsn.insn_singlestep(p->opcode, &p->ainsn, regs); in optimized_callback()
277 (unsigned long)(op->kp.addr) + 4); in arch_prepare_optimized_kprobe()
308 WARN_ON(kprobe_disabled(&op->kp)); in arch_optimize_kprobes()
314 memcpy(op->optinsn.copied_insn, op->kp.addr, in arch_optimize_kprobes()
317 insn = arm_gen_branch((unsigned long)op->kp.addr, in arch_optimize_kprobes()
334 kprobes_remove_breakpoint(op->kp.addr, insn); in arch_optimize_kprobes()
342 arch_arm_kprobe(&op->kp); in arch_unoptimize_kprobe()
363 return ((unsigned long)op->kp.addr <= addr && in arch_within_optimized_kprobe()
364 (unsigned long)op->kp.addr + RELATIVEJUMP_SIZE > addr); in arch_within_optimized_kprobe()