/linux-4.1.27/arch/x86/purgatory/ |
D | sha256.c | 28 static inline u32 Maj(u32 x, u32 y, u32 z) in Maj() function 68 t2 = e0(a) + Maj(a, b, c); d += t1; h = t1 + t2; in sha256_transform() 70 t2 = e0(h) + Maj(h, a, b); c += t1; g = t1 + t2; in sha256_transform() 72 t2 = e0(g) + Maj(g, h, a); b += t1; f = t1 + t2; in sha256_transform() 74 t2 = e0(f) + Maj(f, g, h); a += t1; e = t1 + t2; in sha256_transform() 76 t2 = e0(e) + Maj(e, f, g); h += t1; d = t1 + t2; in sha256_transform() 78 t2 = e0(d) + Maj(d, e, f); g += t1; c = t1 + t2; in sha256_transform() 80 t2 = e0(c) + Maj(c, d, e); f += t1; b = t1 + t2; in sha256_transform() 82 t2 = e0(b) + Maj(b, c, d); e += t1; a = t1 + t2; in sha256_transform() 85 t2 = e0(a) + Maj(a, b, c); d += t1; h = t1 + t2; in sha256_transform() [all …]
|
/linux-4.1.27/crypto/ |
D | sha256_generic.c | 35 static inline u32 Maj(u32 x, u32 y, u32 z) in Maj() function 75 t2 = e0(a) + Maj(a,b,c); d+=t1; h=t1+t2; in sha256_transform() 77 t2 = e0(h) + Maj(h,a,b); c+=t1; g=t1+t2; in sha256_transform() 79 t2 = e0(g) + Maj(g,h,a); b+=t1; f=t1+t2; in sha256_transform() 81 t2 = e0(f) + Maj(f,g,h); a+=t1; e=t1+t2; in sha256_transform() 83 t2 = e0(e) + Maj(e,f,g); h+=t1; d=t1+t2; in sha256_transform() 85 t2 = e0(d) + Maj(d,e,f); g+=t1; c=t1+t2; in sha256_transform() 87 t2 = e0(c) + Maj(c,d,e); f+=t1; b=t1+t2; in sha256_transform() 89 t2 = e0(b) + Maj(b,c,d); e+=t1; a=t1+t2; in sha256_transform() 92 t2 = e0(a) + Maj(a,b,c); d+=t1; h=t1+t2; in sha256_transform() [all …]
|
D | sha512_generic.c | 31 static inline u64 Maj(u64 x, u64 y, u64 z) in Maj() function 110 t2 = e0(a) + Maj(a,b,c); d+=t1; h=t1+t2; in sha512_transform() 112 t2 = e0(h) + Maj(h,a,b); c+=t1; g=t1+t2; in sha512_transform() 114 t2 = e0(g) + Maj(g,h,a); b+=t1; f=t1+t2; in sha512_transform() 116 t2 = e0(f) + Maj(f,g,h); a+=t1; e=t1+t2; in sha512_transform() 118 t2 = e0(e) + Maj(e,f,g); h+=t1; d=t1+t2; in sha512_transform() 120 t2 = e0(d) + Maj(d,e,f); g+=t1; c=t1+t2; in sha512_transform() 122 t2 = e0(c) + Maj(c,d,e); f+=t1; b=t1+t2; in sha512_transform() 124 t2 = e0(b) + Maj(b,c,d); e+=t1; a=t1+t2; in sha512_transform()
|
/linux-4.1.27/arch/arm/crypto/ |
D | sha256-core.S_shipped | 121 add r4,r4,r12 @ h+=Maj(a,b,c) from the past 128 add r4,r4,r12 @ h+=Maj(a,b,c) from the past 170 eor r3,r3,r5 @ Maj(a,b,c) 172 @ add r11,r11,r3 @ h+=Maj(a,b,c) 179 add r11,r11,r3 @ h+=Maj(a,b,c) from the past 186 add r11,r11,r3 @ h+=Maj(a,b,c) from the past 228 eor r12,r12,r4 @ Maj(a,b,c) 230 @ add r10,r10,r12 @ h+=Maj(a,b,c) 237 add r10,r10,r12 @ h+=Maj(a,b,c) from the past 244 add r10,r10,r12 @ h+=Maj(a,b,c) from the past [all …]
|
/linux-4.1.27/arch/x86/crypto/ |
D | sha512-avx-asm.S | 147 xor tmp0, T2 # T2 = ((a ^ c) & b) ^ (a & c) = Maj(a,b,c) 154 lea (T1, T2), h_64 # a(next_state) = T1 + Maj(a,b,c) 156 add tmp0, h_64 # a(next_state) = T1 + Maj(a,b,c) S0(a)
|
D | sha512-ssse3-asm.S | 140 xor tmp0, T2 # T2 = ((a ^ c) & b) ^ (a & c) = Maj(a,b,c) 147 lea (T1, T2), h_64 # a(next_state) = T1 + Maj(a,b,c) 149 add tmp0, h_64 # a(next_state) = T1 + Maj(a,b,c) S0(a)
|
/linux-4.1.27/drivers/staging/rtl8188eu/include/ |
D | rtw_security.h | 332 #define Maj(x, y, z) (((x | y) & z) | (x & y)) macro
|