/linux/arch/x86/crypto/ |
A D | blake2s-core.S | 57 movq %rcx,%xmm15 68 leaq SIGMA(%rip),%rcx 70 movzbl (%rcx),%eax 72 movzbl 0x1(%rcx),%eax 74 movzbl 0x2(%rcx),%eax 76 movzbl 0x3(%rcx),%eax 91 movzbl 0x4(%rcx),%eax 93 movzbl 0x5(%rcx),%eax 160 addq $0x10,%rcx 161 cmpq %r8,%rcx [all …]
|
A D | chacha-avx512vl-x86_64.S | 115 cmp $0x10,%rcx 122 cmp $0x20,%rcx 129 cmp $0x30,%rcx 136 cmp $0x40,%rcx 144 cmp $0x50,%rcx 150 cmp $0x60,%rcx 156 cmp $0x70,%rcx 162 cmp $0x80,%rcx 173 mov %rcx,%rax 174 and $0xf,%rcx [all …]
|
A D | camellia-aesni-avx-asm_64.S | 194 %rcx, (%r9)); 727 leaq 8 * 16(%rax), %rcx; 731 %xmm15, %rax, %rcx); 735 %xmm15, %rax, %rcx, 0); 747 %xmm15, %rax, %rcx, 8); 767 vmovdqu 0 * 16(%rcx), %xmm8; 768 vmovdqu 1 * 16(%rcx), %xmm9; 815 leaq 8 * 16(%rax), %rcx; 819 %xmm15, %rax, %rcx); 854 vmovdqu 0 * 16(%rcx), %xmm8; [all …]
|
A D | camellia-aesni-avx2-asm_64.S | 228 %rcx, (%r9)); 763 leaq 8 * 32(%rax), %rcx; 767 %ymm15, %rax, %rcx); 771 %ymm15, %rax, %rcx, 0); 783 %ymm15, %rax, %rcx, 8); 803 vmovdqu 0 * 32(%rcx), %ymm8; 804 vmovdqu 1 * 32(%rcx), %ymm9; 851 leaq 8 * 32(%rax), %rcx; 855 %ymm15, %rax, %rcx); 890 vmovdqu 0 * 32(%rcx), %ymm8; [all …]
|
A D | sm4-aesni-avx-asm_64.S | 162 cmpq $2, %rcx; 167 cmpq $3, %rcx; 237 cmpq $2, %rcx; 242 cmpq $3, %rcx; 374 cmpq $5, %rcx; 389 cmpq $7, %rcx; 398 cmpq $6, %rcx; 408 cmpq $7, %rcx; 433 vmovdqu (%rcx), RA0; 558 vmovdqu (%rcx), RA0; [all …]
|
A D | poly1305-x86_64-cryptogams.pl | 267 cmovc %rcx,%r11 390 mov %r9,%rcx 1474 add %rcx,%rax 1488 mov %r10,%rcx 1491 and \$-4,%rcx 1492 add %rcx,%rax 1499 mov %r9,%rcx 2880 lea (%rcx,%rcx,4),%rcx # *5 3847 adc %rcx,%r9 3852 mov %r9,%rcx [all …]
|
/linux/arch/x86/kernel/cpu/sgx/ |
A D | encls.h | 102 #define __encls_ret_1(rax, rcx) \ argument 104 __encls_ret_N(rax, "c"(rcx)); \ 107 #define __encls_ret_2(rax, rbx, rcx) \ argument 109 __encls_ret_N(rax, "b"(rbx), "c"(rcx)); \ 112 #define __encls_ret_3(rax, rbx, rcx, rdx) \ argument 114 __encls_ret_N(rax, "b"(rbx), "c"(rcx), "d"(rdx)); \ 150 #define __encls_2(rax, rbx, rcx) \ argument 153 __encls_N(rax, ign_rbx_out, "b"(rbx), "c"(rcx)); \ 156 #define __encls_1_1(rax, data, rcx) \ argument 159 int ret = __encls_N(rax, rbx_out, "c"(rcx)); \
|
/linux/arch/x86/power/ |
A D | hibernate_asm_64.S | 37 movq %cr3, %rcx; # flush TLB 38 movq %rcx, %cr3 48 movq pt_regs_cx(%rax), %rcx 79 movq %rcx, pt_regs_cx(%rax) 117 jmpq *%rcx 125 movq %rbx, %rcx 126 andq $~(X86_CR4_PGE), %rcx 127 movq %rcx, %cr4; # turn off PGE 128 movq %cr3, %rcx; # flush TLB 129 movq %rcx, %cr3; [all …]
|
/linux/arch/x86/kernel/ |
A D | ftrace_64.S | 82 movq %rcx, RCX(%rsp) 125 movq RCX(%rsp), %rcx 145 movq %rcx, RSP(%rsp) 152 leaq (%rsp), %rcx 208 movq %rcx, EFLAGS(%rsp) 210 movq $__KERNEL_DS, %rcx 211 movq %rcx, SS(%rsp) 212 movq $__KERNEL_CS, %rcx 213 movq %rcx, CS(%rsp) 216 movq %rcx, RSP(%rsp) [all …]
|
A D | relocate_kernel_64.S | 159 movq %rcx, %r11 240 movq %rdi, %rcx /* Put the page_list in %rcx */ 247 movq (%rbx), %rcx 252 movq %rcx, %rdi 258 movq %rcx, %rbx 268 movq %rcx, %rsi /* For ever source page do a copy */
|
/linux/arch/x86/lib/ |
A D | memset_64.S | 32 movq %rdx,%rcx 34 shrq $3,%rcx 63 movq %rdx,%rcx 75 imulq %rcx,%rax 83 movq %rdx,%rcx 84 shrq $6,%rcx 89 decq %rcx
|
A D | csum-copy_64.S | 134 movq %rcx, %r10 219 subq $2, %rcx 229 subq $4, %rcx 242 leaq -1(%rcx, %rcx), %rcx 243 rorq $1, %rcx
|
A D | memmove_64.S | 83 movq %rdx, %rcx 86 shrq $3, %rcx 97 movq %rdx, %rcx 102 shrq $3, %rcx
|
/linux/tools/arch/x86/lib/ |
A D | memset_64.S | 32 movq %rdx,%rcx 34 shrq $3,%rcx 63 movq %rdx,%rcx 75 imulq %rcx,%rax 83 movq %rdx,%rcx 84 shrq $6,%rcx 89 decq %rcx
|
/linux/arch/x86/mm/ |
A D | mem_encrypt_boot.S | 37 movq %rcx, %rax /* Workarea stack page */ 49 movq $(.L__enc_copy_end - __enc_copy), %rcx /* Encryption routine length */ 56 movq %r12, %rcx /* Area length */ 108 movq %rcx, %r9 /* Save area length */ 132 movq %r12, %rcx 137 movq %r12, %rcx
|
/linux/arch/x86/xen/ |
A D | xen-asm.S | 42 push %rcx 58 pop %rcx 125 pop %rcx 165 pop %rcx 233 popq %rcx 252 popq %rcx 273 popq %rcx
|
A D | xen-head.S | 49 mov $__bss_stop, %rcx 50 sub %rdi, %rcx 51 shr $3, %rcx
|
/linux/arch/x86/entry/vdso/ |
A D | vsgx.S | 46 mov SGX_ENCLAVE_OFFSET_OF_RUN(%rbp), %rcx 51 cmpq $0, (%rcx, %rbx) 58 mov SGX_ENCLAVE_RUN_TCS(%rcx), %rbx 59 lea .Lasync_exit_pointer(%rip), %rcx 105 mov %rsp, %rcx
|
/linux/tools/testing/selftests/kvm/x86_64/ |
A D | userspace_io_test.c | 93 if (regs.rcx == 2) in main() 94 regs.rcx = 1; in main() 95 if (regs.rcx == 3) in main() 96 regs.rcx = 8192; in main()
|
/linux/arch/x86/entry/ |
A D | entry_64.S | 103 pushq %rcx /* pt_regs->ip */ 125 movq RCX(%rsp), %rcx 128 cmpq %rcx, %r11 /* SYSRET requires RCX == RIP */ 146 shl $(64 - (__VIRTUAL_MASK_SHIFT+1)), %rcx 147 sar $(64 - (__VIRTUAL_MASK_SHIFT+1)), %rcx 151 cmpq %rcx, %r11 821 movq (%rsp), %rcx 828 movq (%rsp), %rcx 1001 leaq native_irq_return_iret(%rip), %rcx 1002 cmpq %rcx, RIP+8(%rsp) [all …]
|
A D | entry_64_compat.S | 88 pushq %rcx /* pt_regs->cx */ 220 pushq %rcx /* pt_regs->ip */ 272 movq RIP(%rsp), %rcx /* pt_regs->ip (in rcx) */ 384 pushq %rcx /* pt_regs->cx */
|
/linux/arch/x86/purgatory/ |
A D | entry64.S | 40 movq rcx(%rip), %rcx 63 rcx: .quad 0x0 label
|
/linux/arch/x86/platform/efi/ |
A D | efi_stub_64.S | 22 mov %rcx, %r8 23 mov %rsi, %rcx
|
/linux/tools/testing/selftests/sgx/ |
A D | test_encl_bootstrap.S | 50 push %rcx # push the address after EENTER 58 xor %rcx, %rcx
|
/linux/arch/x86/um/ |
A D | stub_64.S | 41 pop %rcx 42 cmp %rcx, %rax
|