/linux/arch/x86/kernel/ |
A D | espfix_64.c | 139 void *stack_page; in init_espfix_ap() local 150 stack_page = READ_ONCE(espfix_pages[page]); in init_espfix_ap() 151 if (likely(stack_page)) in init_espfix_ap() 157 stack_page = READ_ONCE(espfix_pages[page]); in init_espfix_ap() 158 if (stack_page) in init_espfix_ap() 189 stack_page = page_address(alloc_pages_node(node, GFP_KERNEL, 0)); in init_espfix_ap() 194 pte = __pte(__pa(stack_page) | ((__PAGE_KERNEL_RO | _PAGE_ENC) & ptemask)); in init_espfix_ap() 199 WRITE_ONCE(espfix_pages[page], stack_page); in init_espfix_ap() 205 per_cpu(espfix_waddr, cpu) = (unsigned long)stack_page in init_espfix_ap()
|
/linux/arch/mips/kernel/ |
A D | perf_event.c | 51 unsigned long stack_page = in perf_callchain_kernel() local 53 if (stack_page && sp >= stack_page && in perf_callchain_kernel() 54 sp <= stack_page + THREAD_SIZE - 32) in perf_callchain_kernel()
|
A D | stacktrace.c | 46 unsigned long stack_page = in save_context_stack() local 48 if (stack_page && sp >= stack_page && in save_context_stack() 49 sp <= stack_page + THREAD_SIZE - 32) in save_context_stack()
|
A D | process.c | 542 unsigned long notrace unwind_stack_by_address(unsigned long stack_page, in unwind_stack_by_address() argument 553 if (!stack_page) in unwind_stack_by_address() 560 low = stack_page; in unwind_stack_by_address() 562 high = stack_page + IRQ_STACK_START; in unwind_stack_by_address() 565 high = stack_page + THREAD_SIZE - 32; in unwind_stack_by_address() 637 unsigned long stack_page = 0; in unwind_stack() local 642 stack_page = (unsigned long)irq_stack[cpu]; in unwind_stack() 647 if (!stack_page) in unwind_stack() 648 stack_page = (unsigned long)task_stack_page(task); in unwind_stack() 650 return unwind_stack_by_address(stack_page, sp, pc, ra); in unwind_stack()
|
/linux/arch/hexagon/kernel/ |
A D | process.c | 136 unsigned long stack_page; in __get_wchan() local 139 stack_page = (unsigned long)task_stack_page(p); in __get_wchan() 142 if (fp < (stack_page + sizeof(struct thread_info)) || in __get_wchan() 143 fp >= (THREAD_SIZE - 8 + stack_page)) in __get_wchan()
|
/linux/arch/h8300/kernel/ |
A D | process.c | 134 unsigned long stack_page; in __get_wchan() local 137 stack_page = (unsigned long)p; in __get_wchan() 140 if (fp < stack_page+sizeof(struct thread_info) || in __get_wchan() 141 fp >= 8184+stack_page) in __get_wchan()
|
/linux/arch/powerpc/kernel/ |
A D | process.c | 2052 unsigned long stack_page; in valid_irq_stack() local 2055 stack_page = (unsigned long)hardirq_ctx[cpu]; in valid_irq_stack() 2056 if (sp >= stack_page && sp <= stack_page + THREAD_SIZE - nbytes) in valid_irq_stack() 2059 stack_page = (unsigned long)softirq_ctx[cpu]; in valid_irq_stack() 2060 if (sp >= stack_page && sp <= stack_page + THREAD_SIZE - nbytes) in valid_irq_stack() 2070 unsigned long stack_page; in valid_emergency_stack() local 2077 if (sp >= stack_page && sp <= stack_page + THREAD_SIZE - nbytes) in valid_emergency_stack() 2082 if (sp >= stack_page && sp <= stack_page + THREAD_SIZE - nbytes) in valid_emergency_stack() 2086 if (sp >= stack_page && sp <= stack_page + THREAD_SIZE - nbytes) in valid_emergency_stack() 2098 unsigned long stack_page = (unsigned long)task_stack_page(p); in validate_sp() local [all …]
|
A D | stacktrace.c | 70 unsigned long stack_page = (unsigned long)task_stack_page(task); in arch_stack_walk_reliable() local 75 stack_end = stack_page + THREAD_SIZE; in arch_stack_walk_reliable() 105 if (sp < stack_page + sizeof(struct thread_struct) || in arch_stack_walk_reliable()
|
/linux/arch/m68k/kernel/ |
A D | process.c | 269 unsigned long stack_page; in __get_wchan() local 272 stack_page = (unsigned long)task_stack_page(p); in __get_wchan() 275 if (fp < stack_page+sizeof(struct thread_info) || in __get_wchan() 276 fp >= 8184+stack_page) in __get_wchan()
|
/linux/arch/um/kernel/ |
A D | process.c | 369 unsigned long stack_page, sp, ip; in __get_wchan() local 372 stack_page = (unsigned long) task_stack_page(p); in __get_wchan() 374 if (stack_page == 0) in __get_wchan() 382 if (sp < stack_page) in __get_wchan() 385 while (sp < stack_page + THREAD_SIZE) { in __get_wchan()
|
/linux/arch/nios2/kernel/ |
A D | process.c | 223 unsigned long stack_page; in __get_wchan() local 226 stack_page = (unsigned long)p; in __get_wchan() 229 if (fp < stack_page+sizeof(struct task_struct) || in __get_wchan() 230 fp >= 8184+stack_page) /* ;dgt2;tmp */ in __get_wchan()
|
/linux/arch/xtensa/kernel/ |
A D | process.c | 325 unsigned long stack_page = (unsigned long) task_stack_page(p); in __get_wchan() local 332 if (sp < stack_page + sizeof(struct task_struct) || in __get_wchan() 333 sp >= (stack_page + THREAD_SIZE) || in __get_wchan()
|
/linux/arch/arm/kernel/ |
A D | process.c | 282 unsigned long stack_page; in __get_wchan() local 289 stack_page = (unsigned long)task_stack_page(p); in __get_wchan() 291 if (frame.sp < stack_page || in __get_wchan() 292 frame.sp >= stack_page + THREAD_SIZE || in __get_wchan()
|
/linux/arch/arm64/kernel/ |
A D | process.c | 534 unsigned long stack_page, ret = 0; in __get_wchan() local 537 stack_page = (unsigned long)try_get_task_stack(p); in __get_wchan() 538 if (!stack_page) in __get_wchan()
|
/linux/arch/arm64/kvm/ |
A D | arm.c | 1874 unsigned long stack_page; in init_hyp_mode() local 1876 stack_page = __get_free_page(GFP_KERNEL); in init_hyp_mode() 1877 if (!stack_page) { in init_hyp_mode() 1882 per_cpu(kvm_arm_hyp_stack_page, cpu) = stack_page; in init_hyp_mode() 1950 char *stack_page = (char *)per_cpu(kvm_arm_hyp_stack_page, cpu); in init_hyp_mode() local 1951 err = create_hyp_mappings(stack_page, stack_page + PAGE_SIZE, in init_hyp_mode()
|
/linux/arch/mips/include/asm/ |
A D | stacktrace.h | 13 extern unsigned long unwind_stack_by_address(unsigned long stack_page,
|