Home
last modified time | relevance | path

Searched refs:t4 (Results 1 – 25 of 33) sorted by relevance

12

/sysdeps/alpha/
A Drawmemchr.S41 lda t4, -1 # .. e1 : build garbage mask
45 mskqh t4, a0, t4 # e0 :
48 cmpbge zero, t4, t4 # .. e1 : bits set iff byte is garbage
55 andnot t3, t4, t0 # e0 : clear garbage bits
76 and t0, 0xaa, t4 # e0 :
79 cmovne t4, 1, t4 # .. e1 :
82 addq v0, t4, v0 # .. e1 :
A Dstrrchr.S50 lda t4, -1 # .. e1 : build garbage mask
53 mskqh t4, a0, t4 # e0 :
56 cmpbge zero, t4, t4 # .. e1 : bits set iff byte is garbage
58 andnot t1, t4, t1 # .. e1 : clear garbage from null test
59 andnot t3, t4, t3 # e0 : clear garbage from char test
75 negq t1, t4 # e0 : isolate first null byte match
76 and t1, t4, t4 # e1 :
77 subq t4, 1, t5 # e0 : build a mask of the bytes upto...
78 or t4, t5, t4 # e1 : ... and including the null
80 and t3, t4, t3 # e0 : mask out char matches after null
A Dstrchr.S46 lda t4, -1 # .. e1 : build garbage mask
49 mskqh t4, a0, t4 # e0 :
52 cmpbge zero, t4, t4 # .. e1 : bits set iff byte is garbage
57 andnot t0, t4, t0 # e0 : clear garbage bits
77 and t0, 0xaa, t4 # e0 :
80 cmovne t4, 1, t4 # .. e1 :
82 addq v0, t4, v0 # .. e1 :
A Ddivq.S138 stq t4, 8(sp)
161 ldq t4, 8(sp)
170 cfi_restore (t4)
213 subq R, SY, t4
216 cmovne AT, t4, R
258 negq Y, t4
259 cmovne AT, t4, Y
262 negq X, t4
263 cmovne AT, t4, X
265 negq RV, t4
[all …]
A Dremq.S139 stq t4, 8(sp)
162 ldq t4, 8(sp)
171 cfi_restore (t4)
214 subq R, SY, t4
217 cmovne AT, t4, R
255 negq Y, t4
256 cmovne AT, t4, Y
258 negq X, t4
259 cmovlbs t5, t4, X
260 negq RV, t4
[all …]
A Dstrcat.S56 and t1, 0xaa, t4
59 cmovne t4, 1, t4
61 addq a0, t4, a0
A Dstrlen.S62 and t1, 0xaa, t4
65 cmovne t4, 1, t4
67 addq v0, t4, v0
A Dstrcmp.S76 and a0, 7, t4 # e0 : find s1 misalignment
78 subq a1, t4, a1 # e0 :
83 cmplt t4, t5, t8 # .. e1 :
112 extqh t2, a1, t4 # e0 :
114 or t1, t4, t1 # e0 : s2 word now complete
120 xor t0, t1, t4 # e0 : compare aligned words
121 bne t4, $wordcmp # .. e1 (zdb)
148 xor t0, t1, t4 # e0 : compare the words
149 bne t4, $wordcmp # .. e1 (zdb)
150 cmpbge zero, t2, t4 # e0 : find zero in next low bits
[all …]
A Dstrncat.S58 and t1, 0xaa, t4
61 cmovne t4, 1, t4
63 addq a0, t4, a0
A Dstrncmp.S53 and a0, 7, t4 # find s1 misalignment
57 addq a2, t4, a2 # bias count by s1 misalignment
130 subq a1, t4, a1 # adjust s2 for s1 misalignment
131 cmpult t4, t5, t9
164 extqh t2, a1, t4
166 or t1, t4, t1 # s2 word now complete
173 xor t0, t1, t4 # compare aligned words
175 bne t4, $wordcmp
207 cmpbge zero, t2, t4 # .. e1 : eos in s2[1]
219 or t4, t5, t4 # e0 : eos or eoc in s2[1].
[all …]
A Ddivqu.S154 stq t4, 8(sp)
158 cfi_rel_offset (t4, 8)
173 ldq t4, 8(sp)
186 cfi_restore (t4)
228 subq R, SY, t4
231 cmovne AT, t4, R
A Dremqu.S160 stq t4, 8(sp)
164 cfi_rel_offset (t4, 8)
179 ldq t4, 8(sp)
192 cfi_restore (t4)
234 subq R, SY, t4
237 cmovne AT, t4, R
A Dldiv.S130 negq Q, t4
131 cmovlbs t5, t4, Q
170 subq R, SY, t4
173 cmovne AT, t4, R
A Dmemset.S67 $tail: bne t4, 1f # is there a tail to do?
72 mskql a1, t4, t1 # .. e0 :
73 mskqh t0, t4, t0 # e0 :
96 and a2, 7, t4 # .. e1 : find number of bytes in tail
A Dbzero.S65 $tail: bne t4, 1f # is there a tail to do?
69 mskqh t0, t4, t0 #
85 and a1, 7, t4 # .. e1 : find number of bytes in tail
A D_mcount.S62 stq t4, 0x68(sp)
89 ldq t4, 0x68(sp)
A Dstxcpy.S139 extqh t2, a1, t4 # e0 :
141 or t1, t4, t1 # .. e1 :
241 and a0, 7, t4 # .. e1 : find dest misalignment
249 beq t4, 1f # .. e1 :
254 subq a1, t4, a1 # .. e1 : sub dest misalignment from src addr
259 cmplt t4, t5, t8 # e0 :
A Dstxncpy.S167 extqh t2, a1, t4 # e0 :
168 or t1, t4, t1 # e1 : first aligned src word complete
299 and a0, 7, t4 # .. e1 : find dest misalignment
307 beq t4, 1f # .. e1 :
312 subq a1, t4, a1 # .. e1 : sub dest misalignment from src addr
317 cmplt t4, t5, t8 # e1 :
338 mskqh t1, t4, t1 # e0 :
/sysdeps/alpha/alphaev67/
A Dstrrchr.S39 insbl a1, 1, t4 # U : 000000000000ch00
44 or t2, t4, a1 # E : 000000000000chch
51 sll a1, 48, t4 # U : chch000000000000
53 or t4, a1, a1 # E : chch00000000chch
59 mskqh t5, a0, t4 # E : Complete garbage mask
61 cmpbge zero, t4, t4 # E : bits set iff byte is garbage
64 andnot t1, t4, t1 # E : clear garbage from null test
65 andnot t3, t4, t3 # E : clear garbage from char test
87 negq t1, t4 # E : isolate first null byte match
88 and t1, t4, t4 # E :
[all …]
A Drawmemchr.S44 lda t4, -1 # E : build garbage mask
46 mskqh t4, a0, t4 # U : only want relevant part of first quad
53 cmpbge zero, t4, t4 # E : bits set iff byte is garbage
60 andnot t0, t4, t0 # E : clear garbage bits
A Dstrchr.S45 lda t4, -1 # E : build garbage mask
47 mskqh t4, a0, t4 # U : only want relevant part of first quad
55 cmpbge zero, t4, t4 # E : bits set iff byte is garbage
63 andnot t0, t4, t0 # E : clear garbage bits
A Dstpncpy.S44 cttz t8, t4
52 addq a0, t4, v0
97 addq a0, t4, v0
/sysdeps/mips/sys/
A Dregdef.h52 #define t4 $12 macro
56 #define ta0 t4
/sysdeps/unix/sysv/linux/alpha/alpha/
A Dregdef.h10 #define t4 $5 macro
/sysdeps/ia64/fpu/
A Dlibm_support.h580 t32,t1,t2,t3,t4,t5,t6) \ argument
581 t1=(x)*(t32); t3=x-t1; t3=t3+t1; t4=x-t3; \
584 t2=(t3)*(t6)+(t4)*(t5); \
586 rlo=t1-rhi; rlo=rlo+t2; rlo=rlo+(t4*t6);
595 t32,t1,t2,t3,t4,t5,t6,t7,t8) \ argument
596 __LIBM_MULL1_K80(t7,t8,xhi,yhi, t32,t1,t2,t3,t4,t5,t6) \
611 t32,t1,t2,t3,t4,t5,t6,t7,t8)
629 __LIBM_MULL_K80(r,x,y, t32,t1,t2,t3,t4,t5,t6,t7,t8)
657 t32,t1,t2,t3,t4,t5,t6,t7,t8,t9)
675 __LIBM_DIVL_K80(r,x,y, t32,t1,t2,t3,t4,t5,t6,t7,t8,t9)
[all …]

Completed in 19 milliseconds

12