/sysdeps/x86_64/multiarch/ |
A D | memmove-vec-unaligned-erms.S | 800 LOAD_ONE_SET((%rsi), 0, %VEC(0), %VEC(1), %VEC(2), %VEC(3)) 801 LOAD_ONE_SET((%rsi), PAGE_SIZE, %VEC(4), %VEC(5), %VEC(6), %VEC(7)) 804 STORE_ONE_SET((%rdi), 0, %VEC(0), %VEC(1), %VEC(2), %VEC(3)) 805 STORE_ONE_SET((%rdi), PAGE_SIZE, %VEC(4), %VEC(5), %VEC(6), %VEC(7)) 871 LOAD_ONE_SET((%rsi), 0, %VEC(0), %VEC(1), %VEC(2), %VEC(3)) 872 LOAD_ONE_SET((%rsi), PAGE_SIZE, %VEC(4), %VEC(5), %VEC(6), %VEC(7)) 873 LOAD_ONE_SET((%rsi), PAGE_SIZE * 2, %VEC(8), %VEC(9), %VEC(10), %VEC(11)) 874 LOAD_ONE_SET((%rsi), PAGE_SIZE * 3, %VEC(12), %VEC(13), %VEC(14), %VEC(15)) 877 STORE_ONE_SET((%rdi), 0, %VEC(0), %VEC(1), %VEC(2), %VEC(3)) 878 STORE_ONE_SET((%rdi), PAGE_SIZE, %VEC(4), %VEC(5), %VEC(6), %VEC(7)) [all …]
|
A D | memset-vec-unaligned-erms.S | 137 VMOVU %VEC(0), -VEC_SIZE(%rdi,%rdx) 138 VMOVU %VEC(0), (%rdi) 193 VMOVU %VEC(0), (%rax) 194 VMOVU %VEC(0), -VEC_SIZE(%rax, %rdx) 204 VMOVU %VEC(0), (VEC_SIZE * -2)(%rdi) 205 VMOVU %VEC(0), (VEC_SIZE * -1)(%rdi) 238 vmovdqu8 %VEC(0), (%rax){%k1} 283 VMOVU %VEC(0), (%rax) 284 VMOVU %VEC(0), VEC_SIZE(%rax) 296 VMOVU %VEC(0), (VEC_SIZE * 2)(%rax) [all …]
|
A D | memset-avx512-unaligned-erms.S | 11 # define VEC(i) VEC##i macro
|
A D | memset-evex-unaligned-erms.S | 11 # define VEC(i) VEC##i macro
|
A D | memmove-avx512-unaligned-erms.S | 23 # define VEC(i) VEC##i macro
|
A D | memmove-evex-unaligned-erms.S | 23 # define VEC(i) VEC##i macro
|
A D | memmove-avx-unaligned-erms.S | 3 # define VEC(i) ymm##i macro
|
A D | memmove-avx-unaligned-erms-rtm.S | 3 # define VEC(i) ymm##i macro
|
A D | memset-avx2-unaligned-erms.S | 8 # define VEC(i) ymm##i macro
|
/sysdeps/x86_64/ |
A D | dl-trampoline.S | 57 #define VEC(i) zmm##i macro 61 #undef VEC 67 #define VEC(i) ymm##i macro 71 #undef VEC 78 #define VEC(i) xmm##i macro 83 #undef VEC
|
A D | dl-trampoline.h | 233 VMOVA %VEC(0), (LR_VECTOR_OFFSET)(%rsp) 234 VMOVA %VEC(1), (LR_VECTOR_OFFSET + VECTOR_SIZE)(%rsp) 235 VMOVA %VEC(2), (LR_VECTOR_OFFSET + VECTOR_SIZE*2)(%rsp) 236 VMOVA %VEC(3), (LR_VECTOR_OFFSET + VECTOR_SIZE*3)(%rsp) 237 VMOVA %VEC(4), (LR_VECTOR_OFFSET + VECTOR_SIZE*4)(%rsp) 238 VMOVA %VEC(5), (LR_VECTOR_OFFSET + VECTOR_SIZE*5)(%rsp) 286 2: VMOVA (LR_VECTOR_OFFSET)(%rsp), %VEC(0) 436 VMOVA %VEC(0), LRV_VECTOR0_OFFSET(%rcx) 437 VMOVA %VEC(1), LRV_VECTOR1_OFFSET(%rcx) 466 VMOVA LRV_VECTOR0_OFFSET(%rsp), %VEC(0) [all …]
|
A D | memset.S | 27 #define VEC(i) xmm##i macro
|
A D | memmove.S | 22 #define VEC(i) xmm##i macro
|