/xen/xen/arch/x86/x86_64/ |
A D | entry.S | 18 testb $3, UREGS_cs(%rsp) 206 movq 32(%rsp),%rsp # RSP 218 addq $8,%rsp 270 mov %rsp, %rdi 619 movq %rsp,%rdi 699 1: movq %rsp,%rdi 727 movq %rsp,%rdi 742 testb $15,UREGS_rsp(%rsp) # return %rsp is now aligned? 744 movq %rsp,%rsi 745 subq $8,%rsp [all …]
|
A D | gdbstub.c | 35 GDB_REG64(regs->rsp); in gdb_arch_read_reg_array() 79 case 7: GDB_REG64(regs->rsp); break; in gdb_arch_read_reg() 118 case 7: regs->rsp = val; break; in gdb_arch_write_reg()
|
/xen/xen/include/asm-x86/ |
A D | asm_defns.h | 48 1: addq $8,%rsp; 142 orq %rsp, %r##reg 246 movq %rdi,UREGS_rdi(%rsp) 248 movq %rsi,UREGS_rsi(%rsp) 250 movq %rdx,UREGS_rdx(%rsp) 252 movq %rcx,UREGS_rcx(%rsp) 254 movq %rax,UREGS_rax(%rsp) 257 movq %r8,UREGS_r8(%rsp) 258 movq %r9,UREGS_r9(%rsp) 320 movq UREGS_r9(%rsp),%r9 [all …]
|
A D | spec_ctrl_asm.h | 99 mov %rsp, %\tmp /* Store the current %rsp */ 116 mov %\tmp, %rsp /* Restore old %rsp */ 147 andb $~SCF_use_shadow, CPUINFO_spec_ctrl_flags(%rsp) 150 movzbl CPUINFO_xen_spec_ctrl(%rsp), %eax 175 testb $3, UREGS_cs(%rsp) 181 andb $~SCF_use_shadow, CPUINFO_spec_ctrl_flags(%rsp) 182 movzbl CPUINFO_xen_spec_ctrl(%rsp), %eax 218 mov %eax, CPUINFO_shadow_spec_ctrl(%rsp) 221 orb $SCF_use_shadow, CPUINFO_spec_ctrl_flags(%rsp) 255 ALTERNATIVE "", __stringify(verw CPUINFO_verw_sel(%rsp)), \ [all …]
|
A D | vm_event.h | 49 vm_event_response_t *rsp); 51 void vm_event_register_write_resume(struct vcpu *v, vm_event_response_t *rsp); 53 void vm_event_emulate_check(struct vcpu *v, vm_event_response_t *rsp);
|
/xen/stubdom/ |
A D | vtpm-implicit-fallthrough.patch | 4 sizeof(rsp->auth2->nonceOdd.nonce)); 5 tpm_hmac_update(&hmac, (BYTE*)&rsp->auth2->continueAuthSession, 1); 6 tpm_hmac_final(&hmac, rsp->auth2->auth); 9 tpm_hmac_init(&hmac, rsp->auth1->secret, sizeof(rsp->auth1->secret)); 10 tpm_hmac_update(&hmac, rsp->auth1->digest, sizeof(rsp->auth1->digest));
|
A D | vtpm-deepquote.patch | 5 …-3343,6 +3343,39 @@ static TPM_RESULT execute_TPM_ParentSignEK(TPM_REQUEST *req, TPM_RESPONSE *rsp) 9 +static TPM_RESULT execute_TPM_DeepQuote(TPM_REQUEST *req, TPM_RESPONSE *rsp) 31 + rsp->paramSize = len = sigSize; 32 + rsp->param = ptr = tpm_malloc(len); 34 + tpm_free(rsp->param); 42 static void tpm_setup_rsp_auth(TPM_COMMAND_CODE ordinal, TPM_RESPONSE *rsp) 45 @@ -4098,6 +4131,11 @@ void tpm_execute_command(TPM_REQUEST *req, TPM_RESPONSE *rsp) 46 res = execute_TPM_ParentSignEK(req, rsp); 51 + res = execute_TPM_DeepQuote(req, rsp); 56 res = mtm_execute_command(req, rsp);
|
A D | vtpm-parent-sign-ek.patch | 5 …3312,6 +3312,37 @@ static TPM_RESULT execute_TPM_OwnerReadPubek(TPM_REQUEST *req, TPM_RESPONSE *rsp 9 +static TPM_RESULT execute_TPM_ParentSignEK(TPM_REQUEST *req, TPM_RESPONSE *rsp) 29 + rsp->paramSize = len = sigSize; 30 + rsp->param = ptr = tpm_malloc(len); 32 + tpm_free(rsp->param); 40 static void tpm_setup_rsp_auth(TPM_COMMAND_CODE ordinal, TPM_RESPONSE *rsp) 43 @@ -4062,6 +4093,11 @@ void tpm_execute_command(TPM_REQUEST *req, TPM_RESPONSE *rsp) 44 res = execute_TPM_OwnerReadPubek(req, rsp); 49 + res = execute_TPM_ParentSignEK(req, rsp); 54 res = mtm_execute_command(req, rsp);
|
A D | newlib.patch | 149 + subq $32, rsp 150 + movq rax, 24 (rsp) 151 + movq r12, 16 (rsp) 152 + movq r13, 8 (rsp) 153 + movq r14, 0 (rsp) 168 + movq 24 (rsp), rax 169 + movq 16 (rsp), r12 170 + movq 8 (rsp), r13 172 + addq $32, rsp 195 -#define rsp REG(rsp) [all …]
|
A D | vtpm-deepquote-anyloc.patch | 5 @@ -3347,12 +3347,13 @@ static TPM_RESULT execute_TPM_DeepQuote(TPM_REQUEST *req, TPM_RESPONSE *rsp) 21 @@ -3361,17 +3362,19 @@ static TPM_RESULT execute_TPM_DeepQuote(TPM_REQUEST *req, TPM_RESPONSE *rsp) 32 - rsp->paramSize = len = sigSize; 33 + rsp->paramSize = len = quote_blob_size; 34 rsp->param = ptr = tpm_malloc(len); 37 tpm_free(rsp->param);
|
/xen/xen/arch/x86/ |
A D | vm_event.c | 62 vm_event_response_t *rsp) in vm_event_toggle_singlestep() argument 73 if ( rsp->flags & VM_EVENT_FLAG_TOGGLE_SINGLESTEP ) in vm_event_toggle_singlestep() 81 if ( rsp->flags & VM_EVENT_FLAG_DENY ) in vm_event_register_write_resume() 93 switch ( rsp->reason ) in vm_event_register_write_resume() 99 switch ( rsp->u.write_ctrlreg.index ) in vm_event_register_write_resume() 120 v->arch.vm_event->gprs = rsp->data.regs.x86; in vm_event_set_registers() 213 req->data.regs.x86.rsp = regs->rsp; in vm_event_fill_regs() 259 if ( !(rsp->flags & VM_EVENT_FLAG_EMULATE) ) in vm_event_emulate_check() 265 switch ( rsp->reason ) in vm_event_emulate_check() 272 if ( p2m_mem_access_emulate_check(v, rsp) ) in vm_event_emulate_check() [all …]
|
A D | extable.c | 100 regs->rsp > (unsigned long)regs && in search_exception_table() 101 regs->rsp < (unsigned long)get_cpu_info() ) in search_exception_table() 103 unsigned long retptr = *(unsigned long *)regs->rsp; in search_exception_table() 120 *(unsigned long *)regs->rsp = token.raw; in search_exception_table()
|
/xen/xen/arch/x86/x86_64/compat/ |
A D | entry.S | 17 movl $HYPERCALL_VECTOR, 4(%rsp) 27 mov %rsp, %rdi 126 testb $3,UREGS_cs(%rsp) 211 movl $TRAP_syscall, 4(%rsp) 247 subl $2,UREGS_rip(%rsp) 290 testb $2,UREGS_cs+8(%rsp) 296 movl UREGS_rsp+8(%rsp),%eax 298 movl UREGS_ss+8(%rsp),%eax 303 .Lft4: mov UREGS_ss+8(%rsp),%fs 341 mov %fs,UREGS_ss+8(%rsp) [all …]
|
/xen/xen/arch/x86/pv/ |
A D | iret.c | 57 if ( unlikely(copy_from_user(&iret_saved, (void *)regs->rsp, in do_iret() 84 regs->rsp = iret_saved.rsp; in do_iret() 114 regs->rsp = (u32)regs->rsp; in compat_iret() 117 if ( unlikely(__get_user(regs->eax, (u32 *)regs->rsp)) ) in compat_iret() 125 unlikely(__get_user(regs->cs, (u32 *)regs->rsp + 2)) ) in compat_iret() 135 if ( unlikely(__get_user(eflags, (u32 *)regs->rsp + 3)) ) in compat_iret() 167 rc |= __get_user(x, (u32 *)regs->rsp + i); in compat_iret() 175 rc |= __get_user(x, (u32 *)regs->rsp + i); in compat_iret() 192 if ( unlikely(__put_user(0, (u32 *)regs->rsp)) ) in compat_iret() 208 else if ( __get_user(regs->ss, (u32 *)regs->rsp + 5) || in compat_iret() [all …]
|
/xen/tools/tests/xen-access/ |
A D | xen-access.c | 351 memcpy(RING_GET_RESPONSE(back_ring, rsp_prod), rsp, sizeof(*rsp)); in put_response() 380 vm_event_response_t rsp; in main() local 723 memset( &rsp, 0, sizeof (rsp) ); in main() 725 rsp.vcpu_id = req.vcpu_id; in main() 727 rsp.reason = req.reason; in main() 767 rsp.altp2m_idx = 0; in main() 782 rsp.u.mem_access = req.u.mem_access; in main() 808 rsp.data.regs.arm.pc += 4; in main() 863 rsp.data = req.data; in main() 874 rsp.flags |= VM_EVENT_FLAG_EMULATE; in main() [all …]
|
/xen/stubdom/vtpmmgr/ |
A D | manage-vtpmmgr.pl | 13 my $rsp; 16 sysread $tpm, $rsp, 4096; 17 print '>>', unpack('H*', $rsp), "\n" if $ENV{V}; 18 $rsp; 24 my $rsp = tpm_cmd_raw $head, $msg; 25 my($rtype, $len, $stat, $reply) = unpack 'nNNa*', $rsp;
|
/xen/xen/common/ |
A D | vm_event.c | 325 memcpy(rsp, RING_GET_RESPONSE(front_ring, rsp_cons), sizeof(*rsp)); in vm_event_get_response() 354 vm_event_response_t rsp; in vm_event_resume() local 370 while ( vm_event_get_response(d, ved, &rsp) ) in vm_event_resume() 374 if ( rsp.version != VM_EVENT_INTERFACE_VERSION ) in vm_event_resume() 381 v = domain_vcpu(d, rsp.vcpu_id); in vm_event_resume() 395 p2m_mem_paging_resume(d, &rsp); in vm_event_resume() 403 vm_event_emulate_check(v, &rsp); in vm_event_resume() 409 vm_event_register_write_resume(v, &rsp); in vm_event_resume() 415 vm_event_toggle_singlestep(d, v, &rsp); in vm_event_resume() 419 p2m_altp2m_check(v, rsp.altp2m_idx); in vm_event_resume() [all …]
|
/xen/xen/include/asm-arm/ |
A D | vm_event.h | 38 vm_event_response_t *rsp) in vm_event_toggle_singlestep() argument 44 void vm_event_register_write_resume(struct vcpu *v, vm_event_response_t *rsp) in vm_event_register_write_resume() argument 50 void vm_event_emulate_check(struct vcpu *v, vm_event_response_t *rsp) in vm_event_emulate_check() argument
|
/xen/xen/arch/x86/hvm/vmx/ |
A D | entry.S | 48 mov %rsp,%rdi 70 mov %rsp,%rdi 74 mov %rsp,%rdi 138 mov %rsp,%rdi
|
/xen/tools/xenpaging/ |
A D | xenpaging.c | 582 memcpy(RING_GET_RESPONSE(back_ring, rsp_prod), rsp, sizeof(*rsp)); in put_response() 672 put_response(&paging->vm_event, rsp); in xenpaging_resume_page() 682 policy_notify_paged_in(rsp->u.mem_paging.gfn); in xenpaging_resume_page() 684 policy_notify_paged_in_nomru(rsp->u.mem_paging.gfn); in xenpaging_resume_page() 856 vm_event_response_t rsp; in main() local 948 rsp.u.mem_paging.gfn = req.u.mem_paging.gfn; in main() 949 rsp.vcpu_id = req.vcpu_id; in main() 950 rsp.flags = req.flags; in main() 978 rsp.u.mem_paging.gfn = req.u.mem_paging.gfn; in main() 979 rsp.vcpu_id = req.vcpu_id; in main() [all …]
|
/xen/xen/arch/x86/hvm/svm/ |
A D | entry.S | 33 mov %rsp,%rdi 59 mov %rsp, %rdi 98 mov %rsp,%rdi
|
/xen/tools/ocaml/libs/xb/ |
A D | xs_ring_stubs.c | 150 memcpy(intf->rsp + MASK_XENSTORE_IDX(prod), buffer, len); in ml_interface_write() 153 memcpy(intf->rsp + MASK_XENSTORE_IDX(prod), buffer, space); in ml_interface_write() 154 memcpy(intf->rsp, buffer + space, len - space); in ml_interface_write() 193 intf->rsp[i] = 0xff; in ml_interface_close()
|
/xen/xen/arch/arm/ |
A D | vm_event.c | 35 void vm_event_set_registers(struct vcpu *v, vm_event_response_t *rsp) in vm_event_set_registers() argument 42 regs->pc = rsp->data.regs.arm.pc; in vm_event_set_registers()
|
/xen/xen/include/asm-x86/x86_64/ |
A D | elf.h | 24 unsigned long rsp; member 58 asm volatile("movq %%rsp,%0" : "=m"(core_regs->rsp)); in elf_core_save_regs()
|
/xen/xen/include/public/io/ |
A D | ring.h | 114 __rsp_t rsp; \ 217 unsigned int rsp = RING_SIZE(_r) - \ 219 req < rsp ? req : rsp; \ 249 (&((_r)->sring->ring[((_idx) & (RING_SIZE(_r) - 1))].rsp))
|