/sysdeps/unix/sysv/linux/mips/ |
A D | setcontext.S | 87 li v0, SYS_ify (rt_sigprocmask) 91 REG_L v0, A0OFF(sp) 95 l.d fs0, (24 * SZREG + MCONTEXT_FPREGS)(v0) 96 l.d fs1, (25 * SZREG + MCONTEXT_FPREGS)(v0) 97 l.d fs2, (26 * SZREG + MCONTEXT_FPREGS)(v0) 98 l.d fs3, (27 * SZREG + MCONTEXT_FPREGS)(v0) 99 l.d fs4, (28 * SZREG + MCONTEXT_FPREGS)(v0) 100 l.d fs5, (29 * SZREG + MCONTEXT_FPREGS)(v0) 101 l.d fs6, (30 * SZREG + MCONTEXT_FPREGS)(v0) 114 lw v1, MCONTEXT_FPC_CSR(v0) [all …]
|
A D | swapcontext.S | 138 li v0, SYS_ify (rt_sigprocmask) 142 REG_L v0, A1OFF(sp) 146 l.d fs0, (24 * SZREG + MCONTEXT_FPREGS)(v0) 147 l.d fs1, (25 * SZREG + MCONTEXT_FPREGS)(v0) 148 l.d fs2, (26 * SZREG + MCONTEXT_FPREGS)(v0) 149 l.d fs3, (27 * SZREG + MCONTEXT_FPREGS)(v0) 150 l.d fs4, (28 * SZREG + MCONTEXT_FPREGS)(v0) 151 l.d fs5, (29 * SZREG + MCONTEXT_FPREGS)(v0) 152 l.d fs6, (30 * SZREG + MCONTEXT_FPREGS)(v0) 165 lw v1, MCONTEXT_FPC_CSR(v0) [all …]
|
/sysdeps/powerpc/powerpc64/le/power10/ |
A D | memset.S | 38 mtvsrd v0+32,r4 39 vspltb v0,v0,7 57 stxvl v0+32,r3,r5 109 stxv v0+32,0(r6) 110 stxv v0+32,16(r6) 111 stxv v0+32,32(r6) 112 stxv v0+32,48(r6) 132 stxv v0+32,0(r6) 185 stxv v0+32,0(r6) 192 stxv v0+32,0(r6) [all …]
|
A D | memmove.S | 80 lxvl 32+v0,r4,r6 81 stxvl 32+v0,r11,r6 101 lxv 32+v0,0(r4) 106 stxv 32+v0,0(r11) 134 lxv 32+v0,0(r4) 139 stxv 32+v0,0(r11) 156 lxv 32+v0,0(r4) 169 stxv 32+v0,0(r11) 201 lxv 32+v0,0(r4) 218 lxv 32+v0,-16(r4) [all …]
|
/sysdeps/mips/ |
A D | strcmp.S | 124 move v0, zero 129 nor t1, v0, t9 131 xor t0, v0, v1 143 rotrv v0, v0, t1 145 and v0, v0, 0xff 148 subu v0, v0, v1 156 srl t8, v0, 8 163 srl t0, v0, 16 170 srl t8, v0, 24 185 srl t0, v0, 8 [all …]
|
/sysdeps/powerpc/powerpc64/le/power9/ |
A D | rawmemchr.S | 48 lvx v0,0,r3 50 vperm v0,v19,v0,v1 52 vcmpequb. v6,v0,v18 /* 0xff if byte matches, 0x00 otherwise */ 63 lxv v0+32,0(r3) 64 vcmpequb. v6,v0,v18 /* 0xff if byte matches, 0x00 otherwise */ 67 lxv v0+32,16(r3) 68 vcmpequb. v6,v0,v18 /* 0xff if byte matches, 0x00 otherwise */ 71 lxv v0+32,32(r3) 72 vcmpequb. v6,v0,v18 /* 0xff if byte matches, 0x00 otherwise */ 75 lxv v0+32,48(r3) [all …]
|
A D | strlen.S | 44 lvx v0,0,r3 46 vperm v0,v19,v0,v1 48 vcmpequb. v6,v0,v18 62 lxv v0+32,0(r4) 63 vcmpequb. v6,v0,v18 66 lxv v0+32,16(r4) 67 vcmpequb. v6,v0,v18 70 lxv v0+32,32(r4) 74 lxv v0+32,48(r4) 90 lxv v0+32,0(r4) [all …]
|
A D | strcpy.S | 69 lvx v0,0,r4 71 vperm v0,v19,v0,v1 80 stxvl 32+v0,r3,r10 /* Partial store */ 91 stxvl 32+v0,r3,r10 /* Partial store */ 95 CHECK16(v0,0,r5,tail1) 102 stxv 32+v0,0(r11) 129 stxv 32+v0,0(r11) 142 stxv 32+v0,0(r11) 156 stxv 32+v0,0(r11) 171 stxv 32+v0,0(r11) [all …]
|
A D | strncpy.S | 98 lvx v0,0,r4 100 vperm v0,v18,v0,v1 154 lxv 32+v0,0(r4) 170 stxv 32+v0,0(r11) 183 lxv 32+v0,0(r4) 212 stxv 32+v0,0(r11) 246 stxv 32+v0,0(r11) 263 stxv 32+v0,0(r11) 295 stxv 32+v0,0(r11) 313 stxv 32+v0,0(r11) [all …]
|
/sysdeps/unix/mips/ |
A D | sysdep.S | 45 REG_S v0, V0OFF(sp) 52 bne v0, EWOULDBLOCK_sys, L(skip) 54 li v0, EAGAIN 62 sw t0, 0(v0) 68 li v0, -1 85 bne v0, EWOULDBLOCK_sys, L(skip) 86 li v0, EAGAIN 90 sw v0, errno 93 li v0, -1
|
/sysdeps/alpha/ |
A D | htons.S | 33 extbl a0, 1, v0 # v0 = 00aa 34 bis v0, t1, v0 # v0 = bbaa
|
A D | strlen.S | 46 andnot a0, 7, v0 52 $loop: ldq t0, 8(v0) 53 addq v0, 8, v0 # addr += 8 67 addq v0, t4, v0 68 addq v0, t2, v0 71 subq v0, a0, v0
|
A D | rawmemchr.S | 38 andnot a0, 7, v0 # .. e1 : align our loop pointer 62 ldq t0, 8(v0) # e0 : 63 addq v0, 8, v0 # .. e1 : 82 addq v0, t4, v0 # .. e1 : 83 addq v0, t2, v0 # e0 :
|
A D | stpncpy.S | 44 andnot a0, 7, v0 48 addq v0, t3, v0 50 addq v0, t1, v0 57 addq v0, 1, t5 # e0 : 62 cmovne t4, t5, v0 # .. e1 : if last written wasnt null, inc v0 99 mov a0, v0
|
A D | strchr.S | 44 andnot a0, 7, v0 # .. e1 : align our loop pointer 60 $loop: ldq t0, 8(v0) # e0 : 61 addq v0, 8, v0 # .. e1 : 82 addq v0, t4, v0 # .. e1 : 83 addq v0, t2, v0 # e0 : 87 mov zero, v0 # e0 :
|
A D | div.S | 67 mov a0, v0 75 mov a0, v0 78 stl zero, 0(v0) 79 stl zero, 4(v0)
|
/sysdeps/powerpc/powerpc32/fpu/ |
A D | setjmp-common.S | 119 lvsr v0,0,r5 138 save_misaligned_vmx(v20,v2,v0,v3,r5,r6) 139 save_misaligned_vmx(v21,v20,v0,v3,r6,r5) 140 save_misaligned_vmx(v22,v21,v0,v3,r5,r6) 141 save_misaligned_vmx(v23,v22,v0,v3,r6,r5) 142 save_misaligned_vmx(v24,v23,v0,v3,r5,r6) 143 save_misaligned_vmx(v25,v24,v0,v3,r6,r5) 144 save_misaligned_vmx(v26,v25,v0,v3,r5,r6) 145 save_misaligned_vmx(v27,v26,v0,v3,r6,r5) 146 save_misaligned_vmx(v28,v27,v0,v3,r5,r6) [all …]
|
A D | __longjmp-common.S | 66 lvsl v0,0,r5 70 vperm v20,v1,v21,v0 75 load_misaligned_vmx_lo_loaded(v21,v22,v0,r5,r6) 76 load_misaligned_vmx_lo_loaded(v22,v23,v0,r6,r5) 77 load_misaligned_vmx_lo_loaded(v23,v24,v0,r5,r6) 78 load_misaligned_vmx_lo_loaded(v24,v25,v0,r6,r5) 79 load_misaligned_vmx_lo_loaded(v25,v26,v0,r5,r6) 80 load_misaligned_vmx_lo_loaded(v26,v27,v0,r6,r5) 81 load_misaligned_vmx_lo_loaded(v27,v28,v0,r5,r6) 82 load_misaligned_vmx_lo_loaded(v28,v29,v0,r6,r5) [all …]
|
/sysdeps/powerpc/powerpc64/multiarch/ |
A D | memcpy-power8-cached.S | 70 lxvd2x v0,0,r4 75 stxvd2x v0,0,r3 95 lxvd2x v0,r4,r7 100 stxvd2x v0,r9,r7 105 lxvd2x v0,r8,r7 124 lxvd2x v0,r4,r9 129 stxvd2x v0,r12,9 142 lxvd2x v0,r8,r9 150 lxvd2x v0,0,r4 152 stxvd2x v0,0,r3 [all …]
|
/sysdeps/unix/sysv/linux/alpha/ |
A D | clone.S | 50 ldiq v0, EINVAL 69 ldiq v0, __NR_clone 73 beq v0, thread_start 102 mov v0, a0 103 ldiq v0, __NR_exit
|
/sysdeps/unix/alpha/ |
A D | sysdep.S | 51 mov v0, t0 58 addq v0, t1, t1 59 lda v0, -1
|
/sysdeps/powerpc/powerpc64/ |
A D | __longjmp-common.S | 65 lvsl v0,0,r5 69 vperm v20,v1,v21,v0 74 load_misaligned_vmx_lo_loaded(v21,v22,v0,r5,r6) 75 load_misaligned_vmx_lo_loaded(v22,v23,v0,r6,r5) 76 load_misaligned_vmx_lo_loaded(v23,v24,v0,r5,r6) 77 load_misaligned_vmx_lo_loaded(v24,v25,v0,r6,r5) 78 load_misaligned_vmx_lo_loaded(v25,v26,v0,r5,r6) 79 load_misaligned_vmx_lo_loaded(v26,v27,v0,r6,r5) 80 load_misaligned_vmx_lo_loaded(v27,v28,v0,r5,r6) 81 load_misaligned_vmx_lo_loaded(v28,v29,v0,r6,r5) [all …]
|
A D | setjmp-common.S | 151 lvsr v0,0,r5 170 save_misaligned_vmx(v20,v2,v0,v3,r5,r6) 171 save_misaligned_vmx(v21,v20,v0,v3,r6,r5) 172 save_misaligned_vmx(v22,v21,v0,v3,r5,r6) 173 save_misaligned_vmx(v23,v22,v0,v3,r6,r5) 174 save_misaligned_vmx(v24,v23,v0,v3,r5,r6) 175 save_misaligned_vmx(v25,v24,v0,v3,r6,r5) 176 save_misaligned_vmx(v26,v25,v0,v3,r5,r6) 177 save_misaligned_vmx(v27,v26,v0,v3,r6,r5) 178 save_misaligned_vmx(v28,v27,v0,v3,r5,r6) [all …]
|
/sysdeps/alpha/alphaev67/ |
A D | stpncpy.S | 36 mov a0, v0 52 addq a0, t4, v0 56 addq v0, t5, v0 97 addq a0, t4, v0 101 addq v0, t5, v0
|
/sysdeps/powerpc/powerpc64/power8/ |
A D | strrchr.S | 47 vsumsws v2, v2, v0; 64 vsumsws v2, v2, v0; 228 vspltisb v0, 0 240 vcmpequb v2, v0, v4 241 vcmpequb v3, v0, v5 250 vcmpequb. v8, v0, v8 305 vor v6, v0, v7 322 vor v6, v0, v7 415 vspltisb v0, 0 425 vcmpequb v2, v0, v4 [all …]
|