/linux/arch/powerpc/crypto/ |
A D | sha1-powerpc-asm.S | 58 xor r5,W((t)+4-3),W((t)+4-8); \ 60 xor W((t)+4),W((t)+4-16),W((t)+4-14); \ 62 xor W((t)+4),W((t)+4),r5; \ 64 rotlwi W((t)+4),W((t)+4),1 84 xor W((t)+4),W((t)+4-16),W((t)+4-14); \ 86 xor W((t)+4),W((t)+4),r5; \ 88 rotlwi W((t)+4),W((t)+4),1 97 xor r5,W((t)+4-3),W((t)+4-8); \ 99 xor W((t)+4),W((t)+4-16),W((t)+4-14); \ 103 xor W((t)+4),W((t)+4),r5; \ [all …]
|
/linux/lib/crypto/ |
A D | sha256.c | 62 W[I] = s1(W[I-2]) + W[I-7] + s0(W[I-15]) + W[I-16]; in BLEND_OP() 92 BLEND_OP(i + 0, W); in sha256_transform() 93 BLEND_OP(i + 1, W); in sha256_transform() 94 BLEND_OP(i + 2, W); in sha256_transform() 95 BLEND_OP(i + 3, W); in sha256_transform() 96 BLEND_OP(i + 4, W); in sha256_transform() 97 BLEND_OP(i + 5, W); in sha256_transform() 98 BLEND_OP(i + 6, W); in sha256_transform() 99 BLEND_OP(i + 7, W); in sha256_transform() 126 u32 W[64]; in sha256_update() local [all …]
|
/linux/arch/x86/kernel/ |
A D | uprobes.c | 89 W(0x00, 1, 1, 1, 1, 1, 1, 1, 0, 1, 1, 1, 1, 1, 1, 1, 1) | /* 00 */ 90 W(0x10, 1, 1, 1, 1, 1, 1, 1, 0, 1, 1, 1, 1, 1, 1, 1, 0) , /* 10 */ 91 W(0x20, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1) | /* 20 */ 92 W(0x30, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1) , /* 30 */ 93 W(0x40, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1) | /* 40 */ 94 W(0x50, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1) , /* 50 */ 95 W(0x60, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0) | /* 60 */ 96 W(0x70, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1) , /* 70 */ 97 W(0x80, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1) | /* 80 */ 98 W(0x90, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1) , /* 90 */ [all …]
|
/linux/arch/x86/crypto/ |
A D | sha1_ssse3_asm.S | 386 psrld $31, W 391 pslld $2, W 411 pxor W_minus_28, W # W is W_minus_32 before xor 418 psrld $30, W 502 vpxor W_minus_08, W, W 505 vpxor W_TMP1, W, W 509 vpsrld $31, W, W 525 vpxor W_minus_28, W, W # W is W_minus_32 before xor 528 vpxor W_TMP1, W, W 531 vpsrld $30, W, W [all …]
|
A D | sha512-ssse3-asm.S | 162 # Eg. XMM2=W[t-2] really means XMM2={W[t-2]|W[t-1]} 200 psrlq $(7-1), %xmm3 # XMM3 = ((W[t-15]>>1)^W[t-15])>>6 203 pxor %xmm2, %xmm0 # XMM0 = (((W[t-2]>>42)^W[t-2])>>13)^W[t-2] 206 pxor %xmm5, %xmm3 # XMM3 = (((W[t-15]>>1)^W[t-15])>>6)^W[t-15] 209 psrlq $6, %xmm0 # XMM0 = ((((W[t-2]>>42)^W[t-2])>>13)^W[t-2])>>6 212 psrlq $1, %xmm3 # XMM3 = (((W[t-15]>>1)^W[t-15])>>6)^W[t-15]>>1 228 pxor %xmm2, %xmm1 # XMM1 = (W[t-2] << 42)^W[t-2] 231 pxor %xmm5, %xmm4 # XMM4 = (W[t-15]<<7)^W[t-15] 248 paddq %xmm3, %xmm0 # XMM0 = s1(W[t-2]) + s0(W[t-15]) 251 paddq W_t(idx), %xmm0 # XMM0 = s1(W[t-2]) + s0(W[t-15]) + W[t-16] [all …]
|
A D | sha512-avx-asm.S | 77 # W[t] + K[t] | W[t+1] + K[t+1] 167 # Eg. XMM4=W[t-2] really means XMM4={W[t-2]|W[t-1]} 184 vpxor %xmm1, %xmm0, %xmm0 # XMM0 = W[t-2]>>61 ^ W[t-2]>>19 194 vpxor %xmm7, %xmm6, %xmm6 # XMM6 = W[t-15]>>1 ^ W[t-15]>>8 203 vpxor %xmm3, %xmm2, %xmm2 # XMM2 = W[t-2]>>6 ^ W[t-2]<<3 208 vpxor %xmm9, %xmm8, %xmm8 # XMM8 = W[t-15]>>7 ^ W[t-15]<<63 213 vpxor %xmm8, %xmm6, %xmm6 # XMM6 = W[t-15]>>1 ^ W[t-15]>>8 ^ 214 # W[t-15]>>7 ^ W[t-15]<<63 239 vpaddq %xmm6, %xmm0, %xmm0 # XMM0 = s1(W[t-2]) + W[t-16] + s0(W[t-15]) 241 vpaddq %xmm1, %xmm0, %xmm0 # XMM0 = W[t] = s1(W[t-2]) + W[t-7] + [all …]
|
A D | sha512-avx2-asm.S | 168 vpaddq Y_0, YTMP0, YTMP0 # YTMP0 = W[-7] + W[-16] 225 vpxor YTMP4, YTMP3, YTMP3 # YTMP3 = W[-15] ror 1 ^ W[-15] >> 7 230 vpaddq YTMP1, YTMP0, YTMP0 # YTMP0 = W[-16] + W[-7] + s0 232 vperm2f128 $0x0, YTMP0, YTMP0, Y_0 # Y_0 = W[-16] + W[-7] + s0 {BABA} 288 vpxor YTMP3, YTMP4, YTMP4 # YTMP4 = W[-2] ror 19 ^ W[-2] >> 6 {BABA} 293 # (W[-2] ror 61) ^ (W[-2] >> 6) {BABA} 296 vpaddq YTMP4, Y_0, Y_0 # Y_0 = {W[1], W[0], W[1], W[0]} 344 vpxor YTMP3, YTMP4, YTMP4 # YTMP4 = W[-2] ror 19 ^ W[-2] >> 6 {DC--} 349 # (W[-2] ror 61) ^ (W[-2] >> 6) {DC--} 353 vpaddq YTMP4, YTMP0, YTMP2 # YTMP2 = {W[3], W[2], --, --} [all …]
|
A D | sha256-ssse3-asm.S | 149 ## compute W[-16] + W[-7] 4 at a time 154 palignr $4, X2, XTMP0 # XTMP0 = W[-7] 162 paddd X0, XTMP0 # XTMP0 = W[-7] + W[-16] 167 palignr $4, X0, XTMP1 # XTMP1 = W[-15] 171 movdqa XTMP1, XTMP2 # XTMP2 = W[-15] 175 movdqa XTMP1, XTMP3 # XTMP3 = W[-15] 191 movdqa XTMP3, XTMP2 # XTMP2 = W[-15] 215 pxor XTMP2, XTMP1 # XTMP1 = W[-15] ror 7 ^ W[-15] ror 18 227 paddd XTMP1, XTMP0 # XTMP0 = W[-16] + W[-7] + s0 264 paddd XTMP4, XTMP0 # XTMP0 = {..., ..., W[1], W[0]} [all …]
|
A D | sha256-avx-asm.S | 155 ## compute W[-16] + W[-7] 4 at a time 160 vpalignr $4, X2, X3, XTMP0 # XTMP0 = W[-7] 167 vpaddd X0, XTMP0, XTMP0 # XTMP0 = W[-7] + W[-16] 172 vpalignr $4, X0, X1, XTMP1 # XTMP1 = W[-15] 189 vpor XTMP2, XTMP3, XTMP3 # XTMP1 = W[-15] MY_ROR 7 203 vpsrld $3, XTMP1, XTMP4 # XTMP4 = W[-15] >> 3 215 vpxor XTMP2, XTMP3, XTMP3 # XTMP1 = W[-15] MY_ROR 7 ^ W[-15] MY_ROR 224 vpshufd $0b11111010, X3, XTMP2 # XTMP2 = W[-2] {BBAA} 227 vpaddd XTMP1, XTMP0, XTMP0 # XTMP0 = W[-16] + W[-7] + s0 261 vpaddd XTMP4, XTMP0, XTMP0 # XTMP0 = {..., ..., W[1], W[0]} [all …]
|
/linux/tools/bpf/bpftool/bash-completion/ |
A D | bpftool | 22 COMPREPLY+=( $( compgen -W "$w" -- "$cur" ) ) 44 COMPREPLY+=( $( compgen -W "$*" -- "$cur" ) ) 49 COMPREPLY+=( $( compgen -W "$( bpftool -jp map 2>&1 | \ 57 COMPREPLY+=( $( compgen -W "$( bpftool -jp map 2>&1 | \ 64 COMPREPLY+=( $( compgen -W "$( bpftool -jp map 2>&1 | \ 72 COMPREPLY+=( $( compgen -W "$( bpftool -jp map 2>&1 | \ 79 COMPREPLY+=( $( compgen -W "$( bpftool -jp prog 2>&1 | \ 85 COMPREPLY+=( $( compgen -W "$( bpftool -jp prog 2>&1 | \ 97 COMPREPLY+=( $( compgen -W "$( bpftool -jp btf 2>&1 | \ 116 COMPREPLY+=( $( compgen -W "$maps" -- "$cur" ) ) [all …]
|
/linux/arch/arm/crypto/ |
A D | sha1-armv7-neon.S | 218 veor.32 W, W, W_m08; \ 222 veor W, W, tmp0; \ 225 vshl.u32 tmp0, W, #1; \ 229 vshr.u32 W, W, #31; \ 254 veor W, W_m28; \ 260 veor W, W_m16; \ 263 veor W, tmp0; \ 269 vshl.u32 tmp1, W, #2; \ 272 vshr.u32 tmp0, W, #30; \ 275 vorr W, tmp0, tmp1; \ [all …]
|
/linux/crypto/ |
A D | sha512_generic.c | 88 static inline void LOAD_OP(int I, u64 *W, const u8 *input) in LOAD_OP() argument 90 W[I] = get_unaligned_be64((__u64 *)input + I); in LOAD_OP() 93 static inline void BLEND_OP(int I, u64 *W) in BLEND_OP() argument 95 W[I & 15] += s1(W[(I-2) & 15]) + W[(I-7) & 15] + s0(W[(I-15) & 15]); in BLEND_OP() 104 u64 W[16]; in sha512_transform() local 118 LOAD_OP(i + j, W, input); in sha512_transform() 121 BLEND_OP(i + j, W); in sha512_transform() 126 t1 = h + e1(e) + Ch(e,f,g) + sha512_K[i ] + W[(i & 15)]; in sha512_transform() 128 t1 = g + e1(d) + Ch(d,e,f) + sha512_K[i+1] + W[(i & 15) + 1]; in sha512_transform() 130 t1 = f + e1(c) + Ch(c,d,e) + sha512_K[i+2] + W[(i & 15) + 2]; in sha512_transform() [all …]
|
/linux/tools/memory-model/ |
A D | linux-kernel.def | 34 cmpxchg(X,V,W) __cmpxchg{mb}(X,V,W) 35 cmpxchg_relaxed(X,V,W) __cmpxchg{once}(X,V,W) 36 cmpxchg_acquire(X,V,W) __cmpxchg{acquire}(X,V,W) 37 cmpxchg_release(X,V,W) __cmpxchg{release}(X,V,W) 108 atomic_cmpxchg(X,V,W) __cmpxchg{mb}(X,V,W) 109 atomic_cmpxchg_relaxed(X,V,W) __cmpxchg{once}(X,V,W) 110 atomic_cmpxchg_acquire(X,V,W) __cmpxchg{acquire}(X,V,W) 111 atomic_cmpxchg_release(X,V,W) __cmpxchg{release}(X,V,W)
|
/linux/arch/arm/lib/ |
A D | memmove.S | 86 6: W(nop) 87 W(ldr) r3, [r1, #-4]! 88 W(ldr) r4, [r1, #-4]! 89 W(ldr) r5, [r1, #-4]! 90 W(ldr) r6, [r1, #-4]! 91 W(ldr) r7, [r1, #-4]! 92 W(ldr) r8, [r1, #-4]! 93 W(ldr) lr, [r1, #-4]! 97 W(nop) 98 W(str) r3, [r0, #-4]! [all …]
|
/linux/arch/m68k/fpsp040/ |
A D | slogn.S | 436 |--LET V=U*U, W=V*V, CALCULATE 438 |--U + U*V*( [B1 + W*(B3 + W*B5)] + [V*(B2 + W*B4)] ) 443 fmulx %fp1,%fp1 | ...FP1 IS W 448 fmulx %fp1,%fp3 | ...W*B5 449 fmulx %fp1,%fp2 | ...W*B4 451 faddd LOGB3,%fp3 | ...B3+W*B5 452 faddd LOGB2,%fp2 | ...B2+W*B4 454 fmulx %fp3,%fp1 | ...W*(B3+W*B5), FP3 RELEASED 458 faddd LOGB1,%fp1 | ...B1+W*(B3+W*B5) 461 faddx %fp2,%fp1 | ...B1+W*(B3+W*B5) + V*(B2+W*B4), FP2 RELEASED [all …]
|
/linux/arch/arm/boot/compressed/ |
A D | head.S | 216 W(b) 1f 1017 W(b) __armv4_mmu_cache_on 1092 W(b) __fa526_cache_on 1094 W(b) __fa526_cache_flush 1423 W(b) . @ reset 1424 W(b) . @ undef 1428 W(b) . @ svc 1430 W(b) . @ pabort 1431 W(b) . @ dabort 1433 W(b) . @ irq [all …]
|
/linux/ |
A D | CREDITS | 339 W: http://www.mulix.org 576 W: https://pburton.com 1025 W: www.farsite.co.uk 1660 N: Rob W. W. Hooft 2009 W: http://bytesex.org 2043 W: http://www.xos.nl/ 2075 W: http://vtun.sf.net 2500 N: Mark W. McClelland 2703 W: http://namei.org/ 2863 W: https://ojeda.dev [all …]
|
/linux/arch/mips/n64/ |
A D | init.c | 51 #define W 320 macro 82 .width = W, in n64_platform_init() 84 .stride = W * 2, in n64_platform_init() 122 orig = kzalloc(W * H * 2 + 63, GFP_DMA | GFP_KERNEL); in n64_platform_init() 141 res[0].end = phys + W * H * 2 - 1; in n64_platform_init() 149 #undef W
|
/linux/tools/memory-model/Documentation/ |
A D | cheatsheet.txt | 3 C Self R W RMW Self R W DR DW RMW SV 11 Successful *_release() C Y Y Y W Y 13 smp_wmb() Y W Y Y W 28 W: Write, for example, WRITE_ONCE(), or write portion of RMW
|
/linux/drivers/atm/ |
A D | Kconfig | 86 when going from 8W to 16W bursts. 96 bool "Enable 8W TX bursts (recommended)" 103 bool "Enable 4W TX bursts (optional)" 107 this if you have disabled 8W bursts. Enabling 4W if 8W is also set 111 bool "Enable 2W TX bursts (optional)" 115 this if you have disabled 4W and 8W bursts. Enabling 2W if 4W or 8W 126 bool "Enable 8W RX bursts (discouraged)" 134 bool "Enable 4W RX bursts (recommended)" 138 default setting. Enabling 4W if 8W is also set may or may not 142 bool "Enable 2W RX bursts (optional)" [all …]
|
/linux/Documentation/translations/zh_TW/admin-guide/ |
A D | tainted-kernels.rst | 36 CPU: 0 PID: 4424 Comm: insmod Tainted: P W O 4.20.0-0.rc6.fc30 #1 44 Tainted: P W O 47 警告( ``W`` ),並且加載了外部構建的模塊( ``O`` ),所以內核早些時候受到 70 Raw taint value as int/string: 4609/'P W O ' 95 9 _/W 512 內核發出警告 143 9) ``W`` 內核之前已發出過警告(儘管有些警告可能會設置更具體的汙染標誌)。
|
/linux/Documentation/translations/zh_CN/admin-guide/ |
A D | tainted-kernels.rst | 33 CPU: 0 PID: 4424 Comm: insmod Tainted: P W O 4.20.0-0.rc6.fc30 #1 41 Tainted: P W O 44 警告( ``W`` ),并且加载了外部构建的模块( ``O`` ),所以内核早些时候受到 67 Raw taint value as int/string: 4609/'P W O ' 92 9 _/W 512 内核发出警告 140 9) ``W`` 内核之前已发出过警告(尽管有些警告可能会设置更具体的污染标志)。
|
/linux/Documentation/hwmon/ |
A D | sbrmi.rst | 19 specification show the address as 8 bits, left justified with the R/W 23 normally 78h(0111 100W) or 3Ch(011 1100) for socket 0 and 70h(0111 000W) 66 power1: 61.00 W (cap = 225.00 W) 70 power1: 28.39 W (cap = 224.77 W)
|
/linux/arch/x86/kernel/kprobes/ |
A D | core.c | 79 W(0x00, 0, 0, 1, 1, 0, 0, 1, 0, 1, 1, 0, 0, 0, 0, 0, 0) | /* 00 */ 80 W(0x10, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1) , /* 10 */ 81 W(0x20, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0) | /* 20 */ 82 W(0x30, 0, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0) , /* 30 */ 83 W(0x40, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1) | /* 40 */ 84 W(0x50, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0) , /* 50 */ 85 W(0x60, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 1, 1) | /* 60 */ 86 W(0x70, 0, 0, 0, 0, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 1, 1) , /* 70 */ 87 W(0x80, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0) | /* 80 */ 88 W(0x90, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1) , /* 90 */ [all …]
|
/linux/drivers/gpu/drm/i915/gt/ |
A D | intel_reset.h | 72 #define intel_wedge_on_timeout(W, GT, TIMEOUT) \ argument 73 for (__intel_init_wedge((W), (GT), (TIMEOUT), __func__); \ 74 (W)->gt; \ 75 __intel_fini_wedge((W)))
|