Searched refs:rT1 (Results 1 - 6 of 6) sorted by relevance
/linux-4.1.27/arch/powerpc/crypto/ |
H A D | sha256-spe-asm.S | 44 #define rT1 r23 define 50 cmpwi rT1,0; 109 rotrwi rT1,e,11; /* 1: S1' = e rotr 11 */ \ 111 xor rT0,rT0,rT1; /* 1: S1 = S1 xor S1' */ \ 114 andc rT1,g,e; /* 1: ch' = ~e and g */ \ 116 xor rT3,rT3,rT1; /* 1: ch = ch xor ch' */ \ 121 rotrwi rT1,a,13; /* 1: S0' = a rotr 13 */ \ 124 xor rT0,rT0,rT1; /* 1: S0 = S0 xor S0' */ \ 129 and rT1,a,b; /* 1: maj' = a and b */ \ 132 or rT2,rT1,rT2; /* 1: maj = maj or maj' */ \ 135 rotrwi rT1,d,11; /* 2: S1' = e rotr 11 */ \ 138 xor rT0,rT0,rT1; /* 2: S1 = S1 xor S1' */ \ 141 andc rT1,f,d; /* 2: ch' = ~e and g */ \ 143 xor rT3,rT3,rT1; /* 2: ch = ch xor ch' */ \ 148 rotrwi rT1,h,13; /* 2: S0' = a rotr 13 */ \ 151 xor rT0,rT0,rT1; /* 2: S0 = S0 xor S0' */ \ 154 and rT1,h,a; /* 2: maj' = a and b */ \ 157 or rT2,rT1,rT2; /* 2: maj = maj or maj' */ \ 165 evsrwiu rT1,rT0,3; /* s0 = w[-15] >> 3 */ \ 169 evxor rT1,rT1,rT0; /* s0 = s0 xor s0' */ \ 173 evxor rT0,rT0,rT1; /* s0 = s0 xor s0' */ \ 179 evrlwi rT1,w7,15; /* s1' = w[-2] rotr 17 */ \ 181 evxor rT0,rT0,rT1; /* s1 = s1 xor s1' */ \ 183 evrlwi rT1,w7,13; /* s1' = w[-2] rotr 19 */ \ 185 evxor rT0,rT0,rT1; /* s1 = s1 xor s1' */ \ 187 evldw rT1,off(rKP); /* k */ \ 196 evaddw rT1,rT1,w0; /* wk = w + k */ \ 198 evmergehi rT0,rT1,rT1; /* wk1/wk2 */ \ 202 add g,g,rT1; /* 2: temp1 = temp1 + wk */ \ 206 rotrwi rT1,d,11; /* 2: S1' = e rotr 11 */ \ 208 xor rT0,rT0,rT1; /* 2: S1 = S1 xor S1' */ \ 211 andc rT1,f,d; /* 2: ch' = ~e and g */ \ 213 xor rT3,rT3,rT1; /* 2: ch = ch xor ch' */ \ 216 rotrwi rT1,h,13; /* 2: S0' = a rotr 13 */ \ 218 xor rT0,rT0,rT1; /* 2: S0 = S0 xor S0' */ \ 220 and rT1,h,a; /* 2: maj' = a and b */ \ 223 or rT2,rT1,rT2; /* 2: maj = maj or maj' */ \
|
H A D | sha1-spe-asm.S | 41 #define rT1 r0 /* 32 bit temporaries */ define 115 andc rT1,d,b; /* 1: F" = ~B and D */ \ 117 or rT2,rT2,rT1; /* 1: F = F' or F" */ \ 123 and rT1,a,b; /* 2: F' = B and C */ \ 127 or rT2,rT2,rT1; /* 2: F = F' or F" */ \ 138 andc rT1,d,b; /* 1: F" = ~B and D */ \ 140 or rT1,rT1,rT2; /* 1: F = F' or F" */ \ 142 add e,e,rT1; /* 1: E = E + F */ \ 150 evmergehi rT1,rT1,rT0; /* WK1/WK2 */ \ 152 add d,d,rT1; /* 2: E = E + WK */ \ 154 andc rT1,c,a; /* 2: F" = ~B and D */ \ 156 or rT1,rT1,rT2; /* 2: F = F' or F" */ \ 159 add d,d,rT1 /* 2: E = E + F */ 175 evmergehi rT1,rT1,rT0; /* WK1/WK2 */ \ 178 add d,d,rT1; /* 2: E = E + WK */ \ 188 or rT1,b,c; /* 1: F" = B or C */ \ 190 and rT1,d,rT1; /* 1: F" = F" and D */ \ 192 or rT2,rT2,rT1; /* 1: F = F' or F" */ \ 200 evmergehi rT1,rT1,rT0; /* WK1/WK2 */ \ 205 add d,d,rT1; /* 2: E = E + WK */ \
|
H A D | md5-asm.S | 41 #define rT1 r26 define 73 andc rT1,d,b; /* 1: f' = ~b and d */ \ 75 or rT0,rT0,rT1; /* 1: f = f or f' */ \ 86 andc rT1,c,a; /* 2: f' = ~b and d */ \ 87 or rT0,rT0,rT1; /* 2: f = f or f' */ \ 95 and rT1,b,d; /* 1: f' = b and d */ \ 97 or rT0,rT0,rT1; /* 1: f = f or f' */ \ 107 and rT1,a,c; /* 2: f' = b and d */ \ 108 or rT0,rT0,rT1; /* 2: f = f or f' */ \ 116 xor rT1,rT0,d; /* 1: f = f xor f' */ \ 118 add a,a,rT1; /* 1: a = a + f */ \ 125 xor rT1,rT0,a; /* 2: f = b xor f' */ \ 126 add d,d,rT1; /* 2: a = a + f */ \
|
H A D | aes-spe-regs.h | 22 #define rT1 r10 macro
|
H A D | aes-spe-modes.S | 200 addi rT1,rT0,4096 264 addi rT1,rT0,4096 346 li rT1,15 348 andc rLN,rLN,rT1 356 addi rT1,rT0,4096 575 addi rT1,rT0,4096
|
H A D | aes-spe-core.S | 25 rlwimi rT1,in,24-((bpos+3)%4)*8,24,31; 53 LBZ(out, rT1, 0) 56 LBZ(out, rT1, 0)
|
Completed in 95 milliseconds