Lines Matching refs:rsp

92 	movq	%rsp, PER_CPU_VAR(cpu_tss_rw + TSS_sp2)
93 SWITCH_TO_KERNEL_CR3 scratch_reg=%rsp
94 movq PER_CPU_VAR(cpu_current_top_of_stack), %rsp
110 movq %rsp, %rdi
125 movq RCX(%rsp), %rcx
126 movq RIP(%rsp), %r11
154 cmpq $__USER_CS, CS(%rsp) /* CS must match SYSRET */
157 movq R11(%rsp), %r11
158 cmpq %r11, EFLAGS(%rsp) /* R11 == RFLAGS */
184 cmpq $__USER_DS, SS(%rsp) /* SS must match SYSRET */
199 movq %rsp, %rdi
200 movq PER_CPU_VAR(cpu_tss_rw + TSS_sp0), %rsp
215 popq %rsp
238 movq %rsp, TASK_threadsp(%rdi)
239 movq TASK_threadsp(%rsi), %rsp
287 movq %rsp, %rdi
301 movq $0, RAX(%rsp)
328 movq %rsp, %rdi /* pt_regs pointer into 1st argument*/
331 movq ORIG_RAX(%rsp), %rsi /* get error code into 2nd argument*/
332 movq $-1, ORIG_RAX(%rsp) /* no syscall to restart */
364 testb $3, CS-ORIG_RAX(%rsp)
367 pushq 5*8(%rsp)
428 testb $3, CS-ORIG_RAX(%rsp)
436 movq %rsp, %rdi /* pt_regs pointer */
481 testb $3, CS-ORIG_RAX(%rsp)
498 movq %rsp, %rdi /* pt_regs pointer */
500 movq %rax, %rsp /* Switch to new stack */
505 movq ORIG_RAX(%rsp), %rsi /* get error code into 2nd argument*/
506 movq $-1, ORIG_RAX(%rsp) /* no syscall to restart */
508 movq %rsp, %rdi /* pt_regs pointer */
542 movq %rsp, %rdi /* pt_regs pointer into first argument */
543 movq ORIG_RAX(%rsp), %rsi /* get error code into 2nd argument*/
544 movq $-1, ORIG_RAX(%rsp) /* no syscall to restart */
572 testb $3, CS(%rsp)
587 movq %rsp, %rdi
588 movq PER_CPU_VAR(cpu_tss_rw + TSS_sp0), %rsp
618 testb $3, CS(%rsp)
624 addq $8, %rsp /* skip regs->orig_ax */
638 testb $4, (SS-RIP)(%rsp)
680 movq (1*8)(%rsp), %rax /* user RIP */
682 movq (2*8)(%rsp), %rax /* user CS */
684 movq (3*8)(%rsp), %rax /* user RFLAGS */
686 movq (5*8)(%rsp), %rax /* user SS */
688 movq (4*8)(%rsp), %rax /* user RSP */
708 movq %rax, %rsp
785 movq %rdi, %rsp /* we don't return, adjust the stack frame */
809 cmpw %cx, 0x10(%rsp)
812 cmpw %cx, 0x18(%rsp)
815 cmpw %cx, 0x20(%rsp)
818 cmpw %cx, 0x28(%rsp)
821 movq (%rsp), %rcx
822 movq 8(%rsp), %r11
823 addq $0x30, %rsp
828 movq (%rsp), %rcx
829 movq 8(%rsp), %r11
830 addq $0x30, %rsp
972 testb $3, CS+8(%rsp)
987 movq %rsp, %rdi /* arg0 = pt_regs pointer */
989 movq %rax, %rsp /* switch stack */
1002 cmpq %rcx, RIP+8(%rsp)
1005 cmpq %rax, RIP+8(%rsp)
1007 cmpq $.Lgs_change, RIP+8(%rsp)
1027 movq %rcx, RIP+8(%rsp)
1043 mov %rsp, %rdi
1045 mov %rax, %rsp
1052 testb $3, CS(%rsp)
1111 testb $3, CS-RIP+8(%rsp)
1129 movq %rsp, %rdx
1130 movq PER_CPU_VAR(cpu_current_top_of_stack), %rsp
1148 movq %rsp, %rdi
1212 cmpq 8(%rsp), %rdx
1215 cmpq 8(%rsp), %rdx
1224 cmpl $1, -8(%rsp)
1239 lea 6*8(%rsp), %rdx
1241 cmpq %rdx, 4*8(%rsp)
1246 cmpq %rdx, 4*8(%rsp)
1252 testb $(X86_EFLAGS_DF >> 8), (3*8 + 1)(%rsp)
1262 subq $8, %rsp
1263 leaq -10*8(%rsp), %rdx
1271 addq $(6*8), %rsp
1281 movq (%rsp), %rdx
1287 subq $(5*8), %rsp
1291 pushq 11*8(%rsp)
1303 pushq %rsp /* RSP (minus 8 because of the previous push) */
1304 addq $8, (%rsp) /* Fix up RSP */
1329 movq $1, 10*8(%rsp) /* Set "NMI executing". */
1336 addq $(10*8), %rsp
1338 pushq -6*8(%rsp)
1340 subq $(5*8), %rsp
1360 movq %rsp, %rdi
1395 addq $6*8, %rsp
1407 movq $0, 5*8(%rsp) /* clear "NMI executing" */
1437 leaq -PTREGS_SIZE(%rax), %rsp