/sysdeps/alpha/ |
A D | rawmemchr.S | 41 lda t4, -1 # .. e1 : build garbage mask 45 mskqh t4, a0, t4 # e0 : 48 cmpbge zero, t4, t4 # .. e1 : bits set iff byte is garbage 55 andnot t3, t4, t0 # e0 : clear garbage bits 76 and t0, 0xaa, t4 # e0 : 79 cmovne t4, 1, t4 # .. e1 : 82 addq v0, t4, v0 # .. e1 :
|
A D | strrchr.S | 50 lda t4, -1 # .. e1 : build garbage mask 53 mskqh t4, a0, t4 # e0 : 56 cmpbge zero, t4, t4 # .. e1 : bits set iff byte is garbage 58 andnot t1, t4, t1 # .. e1 : clear garbage from null test 59 andnot t3, t4, t3 # e0 : clear garbage from char test 75 negq t1, t4 # e0 : isolate first null byte match 76 and t1, t4, t4 # e1 : 77 subq t4, 1, t5 # e0 : build a mask of the bytes upto... 78 or t4, t5, t4 # e1 : ... and including the null 80 and t3, t4, t3 # e0 : mask out char matches after null
|
A D | strchr.S | 46 lda t4, -1 # .. e1 : build garbage mask 49 mskqh t4, a0, t4 # e0 : 52 cmpbge zero, t4, t4 # .. e1 : bits set iff byte is garbage 57 andnot t0, t4, t0 # e0 : clear garbage bits 77 and t0, 0xaa, t4 # e0 : 80 cmovne t4, 1, t4 # .. e1 : 82 addq v0, t4, v0 # .. e1 :
|
A D | divq.S | 138 stq t4, 8(sp) 161 ldq t4, 8(sp) 170 cfi_restore (t4) 213 subq R, SY, t4 216 cmovne AT, t4, R 258 negq Y, t4 259 cmovne AT, t4, Y 262 negq X, t4 263 cmovne AT, t4, X 265 negq RV, t4 [all …]
|
A D | remq.S | 139 stq t4, 8(sp) 162 ldq t4, 8(sp) 171 cfi_restore (t4) 214 subq R, SY, t4 217 cmovne AT, t4, R 255 negq Y, t4 256 cmovne AT, t4, Y 258 negq X, t4 259 cmovlbs t5, t4, X 260 negq RV, t4 [all …]
|
A D | strcat.S | 56 and t1, 0xaa, t4 59 cmovne t4, 1, t4 61 addq a0, t4, a0
|
A D | strlen.S | 62 and t1, 0xaa, t4 65 cmovne t4, 1, t4 67 addq v0, t4, v0
|
A D | strcmp.S | 76 and a0, 7, t4 # e0 : find s1 misalignment 78 subq a1, t4, a1 # e0 : 83 cmplt t4, t5, t8 # .. e1 : 112 extqh t2, a1, t4 # e0 : 114 or t1, t4, t1 # e0 : s2 word now complete 120 xor t0, t1, t4 # e0 : compare aligned words 121 bne t4, $wordcmp # .. e1 (zdb) 148 xor t0, t1, t4 # e0 : compare the words 149 bne t4, $wordcmp # .. e1 (zdb) 150 cmpbge zero, t2, t4 # e0 : find zero in next low bits [all …]
|
A D | strncat.S | 58 and t1, 0xaa, t4 61 cmovne t4, 1, t4 63 addq a0, t4, a0
|
A D | strncmp.S | 53 and a0, 7, t4 # find s1 misalignment 57 addq a2, t4, a2 # bias count by s1 misalignment 130 subq a1, t4, a1 # adjust s2 for s1 misalignment 131 cmpult t4, t5, t9 164 extqh t2, a1, t4 166 or t1, t4, t1 # s2 word now complete 173 xor t0, t1, t4 # compare aligned words 175 bne t4, $wordcmp 207 cmpbge zero, t2, t4 # .. e1 : eos in s2[1] 219 or t4, t5, t4 # e0 : eos or eoc in s2[1]. [all …]
|
A D | divqu.S | 154 stq t4, 8(sp) 158 cfi_rel_offset (t4, 8) 173 ldq t4, 8(sp) 186 cfi_restore (t4) 228 subq R, SY, t4 231 cmovne AT, t4, R
|
A D | remqu.S | 160 stq t4, 8(sp) 164 cfi_rel_offset (t4, 8) 179 ldq t4, 8(sp) 192 cfi_restore (t4) 234 subq R, SY, t4 237 cmovne AT, t4, R
|
A D | ldiv.S | 130 negq Q, t4 131 cmovlbs t5, t4, Q 170 subq R, SY, t4 173 cmovne AT, t4, R
|
A D | memset.S | 67 $tail: bne t4, 1f # is there a tail to do? 72 mskql a1, t4, t1 # .. e0 : 73 mskqh t0, t4, t0 # e0 : 96 and a2, 7, t4 # .. e1 : find number of bytes in tail
|
A D | bzero.S | 65 $tail: bne t4, 1f # is there a tail to do? 69 mskqh t0, t4, t0 # 85 and a1, 7, t4 # .. e1 : find number of bytes in tail
|
A D | _mcount.S | 62 stq t4, 0x68(sp) 89 ldq t4, 0x68(sp)
|
A D | stxcpy.S | 139 extqh t2, a1, t4 # e0 : 141 or t1, t4, t1 # .. e1 : 241 and a0, 7, t4 # .. e1 : find dest misalignment 249 beq t4, 1f # .. e1 : 254 subq a1, t4, a1 # .. e1 : sub dest misalignment from src addr 259 cmplt t4, t5, t8 # e0 :
|
A D | stxncpy.S | 167 extqh t2, a1, t4 # e0 : 168 or t1, t4, t1 # e1 : first aligned src word complete 299 and a0, 7, t4 # .. e1 : find dest misalignment 307 beq t4, 1f # .. e1 : 312 subq a1, t4, a1 # .. e1 : sub dest misalignment from src addr 317 cmplt t4, t5, t8 # e1 : 338 mskqh t1, t4, t1 # e0 :
|
/sysdeps/alpha/alphaev67/ |
A D | strrchr.S | 39 insbl a1, 1, t4 # U : 000000000000ch00 44 or t2, t4, a1 # E : 000000000000chch 51 sll a1, 48, t4 # U : chch000000000000 53 or t4, a1, a1 # E : chch00000000chch 59 mskqh t5, a0, t4 # E : Complete garbage mask 61 cmpbge zero, t4, t4 # E : bits set iff byte is garbage 64 andnot t1, t4, t1 # E : clear garbage from null test 65 andnot t3, t4, t3 # E : clear garbage from char test 87 negq t1, t4 # E : isolate first null byte match 88 and t1, t4, t4 # E : [all …]
|
A D | rawmemchr.S | 44 lda t4, -1 # E : build garbage mask 46 mskqh t4, a0, t4 # U : only want relevant part of first quad 53 cmpbge zero, t4, t4 # E : bits set iff byte is garbage 60 andnot t0, t4, t0 # E : clear garbage bits
|
A D | strchr.S | 45 lda t4, -1 # E : build garbage mask 47 mskqh t4, a0, t4 # U : only want relevant part of first quad 55 cmpbge zero, t4, t4 # E : bits set iff byte is garbage 63 andnot t0, t4, t0 # E : clear garbage bits
|
A D | stpncpy.S | 44 cttz t8, t4 52 addq a0, t4, v0 97 addq a0, t4, v0
|
/sysdeps/mips/sys/ |
A D | regdef.h | 52 #define t4 $12 macro 56 #define ta0 t4
|
/sysdeps/unix/sysv/linux/alpha/alpha/ |
A D | regdef.h | 10 #define t4 $5 macro
|
/sysdeps/ia64/fpu/ |
A D | libm_support.h | 580 t32,t1,t2,t3,t4,t5,t6) \ argument 581 t1=(x)*(t32); t3=x-t1; t3=t3+t1; t4=x-t3; \ 584 t2=(t3)*(t6)+(t4)*(t5); \ 586 rlo=t1-rhi; rlo=rlo+t2; rlo=rlo+(t4*t6); 595 t32,t1,t2,t3,t4,t5,t6,t7,t8) \ argument 596 __LIBM_MULL1_K80(t7,t8,xhi,yhi, t32,t1,t2,t3,t4,t5,t6) \ 611 t32,t1,t2,t3,t4,t5,t6,t7,t8) 629 __LIBM_MULL_K80(r,x,y, t32,t1,t2,t3,t4,t5,t6,t7,t8) 657 t32,t1,t2,t3,t4,t5,t6,t7,t8,t9) 675 __LIBM_DIVL_K80(r,x,y, t32,t1,t2,t3,t4,t5,t6,t7,t8,t9) [all …]
|