Lines Matching refs:rsp
66 * and does not change rsp.
93 movq %rsp, PER_CPU_VAR(cpu_tss_rw + TSS_sp2)
94 SWITCH_TO_KERNEL_CR3 scratch_reg=%rsp
95 movq PER_CPU_VAR(pcpu_hot + X86_top_of_stack), %rsp
112 movq %rsp, %rdi
132 movq RCX(%rsp), %rcx
133 movq RIP(%rsp), %r11
161 cmpq $__USER_CS, CS(%rsp) /* CS must match SYSRET */
164 movq R11(%rsp), %r11
165 cmpq %r11, EFLAGS(%rsp) /* R11 == RFLAGS */
191 cmpq $__USER_DS, SS(%rsp) /* SS must match SYSRET */
206 movq %rsp, %rdi
207 movq PER_CPU_VAR(cpu_tss_rw + TSS_sp0), %rsp
222 popq %rsp
251 movq %rsp, TASK_threadsp(%rdi)
252 movq TASK_threadsp(%rsi), %rsp
302 movq %rsp, %rsi /* regs */
353 ALTERNATIVE "call error_entry; movq %rax, %rsp", \
359 movq %rsp, %rdi /* pt_regs pointer into 1st argument*/
362 movq ORIG_RAX(%rsp), %rsi /* get error code into 2nd argument*/
363 movq $-1, ORIG_RAX(%rsp) /* no syscall to restart */
407 testb $3, CS-ORIG_RAX(%rsp)
410 pushq 5*8(%rsp)
473 testb $3, CS-ORIG_RAX(%rsp)
481 movq %rsp, %rdi /* pt_regs pointer */
528 testb $3, CS-ORIG_RAX(%rsp)
545 movq %rsp, %rdi /* pt_regs pointer */
547 movq %rax, %rsp /* Switch to new stack */
553 movq ORIG_RAX(%rsp), %rsi /* get error code into 2nd argument*/
554 movq $-1, ORIG_RAX(%rsp) /* no syscall to restart */
556 movq %rsp, %rdi /* pt_regs pointer */
592 movq %rsp, %rdi /* pt_regs pointer into first argument */
593 movq ORIG_RAX(%rsp), %rsi /* get error code into 2nd argument*/
594 movq $-1, ORIG_RAX(%rsp) /* no syscall to restart */
627 testb $3, CS(%rsp)
642 movq %rsp, %rdi
643 movq PER_CPU_VAR(cpu_tss_rw + TSS_sp0), %rsp
674 testb $3, CS(%rsp)
680 addq $8, %rsp /* skip regs->orig_ax */
699 testb $4, (SS-RIP)(%rsp)
742 movq (1*8)(%rsp), %rax /* user RIP */
744 movq (2*8)(%rsp), %rax /* user CS */
746 movq (3*8)(%rsp), %rax /* user RFLAGS */
748 movq (5*8)(%rsp), %rax /* user SS */
750 movq (4*8)(%rsp), %rax /* user RSP */
770 movq %rax, %rsp
850 movq %rdi, %rsp /* we don't return, adjust the stack frame */
876 cmpw %cx, 0x10(%rsp)
879 cmpw %cx, 0x18(%rsp)
882 cmpw %cx, 0x20(%rsp)
885 cmpw %cx, 0x28(%rsp)
888 movq (%rsp), %rcx
889 movq 8(%rsp), %r11
890 addq $0x30, %rsp
895 movq (%rsp), %rcx
896 movq 8(%rsp), %r11
897 addq $0x30, %rsp
1062 testb $3, CS+8(%rsp)
1076 leaq 8(%rsp), %rdi /* arg0 = pt_regs pointer */
1088 cmpq %rcx, RIP+8(%rsp)
1091 cmpq %rax, RIP+8(%rsp)
1093 cmpq $.Lgs_change, RIP+8(%rsp)
1110 leaq 8(%rsp), %rax /* return pt_regs pointer */
1116 movq %rcx, RIP+8(%rsp)
1134 leaq 8(%rsp), %rdi /* arg0 = pt_regs pointer */
1143 testb $3, CS(%rsp)
1204 testb $3, CS-RIP+8(%rsp)
1221 movq %rsp, %rdx
1222 movq PER_CPU_VAR(pcpu_hot + X86_top_of_stack), %rsp
1225 pushq 4*8(%rdx) /* pt_regs->rsp */
1243 movq %rsp, %rdi
1307 cmpq 8(%rsp), %rdx
1310 cmpq 8(%rsp), %rdx
1319 cmpl $1, -8(%rsp)
1334 lea 6*8(%rsp), %rdx
1335 /* Compare the NMI stack (rdx) with the stack we came from (4*8(%rsp)) */
1336 cmpq %rdx, 4*8(%rsp)
1341 cmpq %rdx, 4*8(%rsp)
1347 testb $(X86_EFLAGS_DF >> 8), (3*8 + 1)(%rsp)
1357 subq $8, %rsp
1358 leaq -10*8(%rsp), %rdx
1366 addq $(6*8), %rsp
1376 movq (%rsp), %rdx
1382 subq $(5*8), %rsp
1386 pushq 11*8(%rsp)
1398 pushq %rsp /* RSP (minus 8 because of the previous push) */
1399 addq $8, (%rsp) /* Fix up RSP */
1425 movq $1, 10*8(%rsp) /* Set "NMI executing". */
1432 addq $(10*8), %rsp
1434 pushq -6*8(%rsp)
1436 subq $(5*8), %rsp
1457 movq %rsp, %rdi
1495 addq $6*8, %rsp
1507 movq $0, 5*8(%rsp) /* clear "NMI executing" */
1546 leaq -PTREGS_SIZE(%rax), %rsp