Lines Matching refs:t2
61 t1x, t2, t2x, t3, t3x, t4, t5) \ argument
71 vinserti128 $1, t2x, t3, t2; /* ab: le0 ; cd: le1 */ \
72 vpshufb t1, t2, x0; \
75 add2_le128(t2, t0, t4, t3, t5); /* ab: le2 ; cd: le3 */ \
76 vpshufb t1, t2, x1; \
77 add2_le128(t2, t0, t4, t3, t5); \
78 vpshufb t1, t2, x2; \
79 add2_le128(t2, t0, t4, t3, t5); \
80 vpshufb t1, t2, x3; \
81 add2_le128(t2, t0, t4, t3, t5); \
82 vpshufb t1, t2, x4; \
83 add2_le128(t2, t0, t4, t3, t5); \
84 vpshufb t1, t2, x5; \
85 add2_le128(t2, t0, t4, t3, t5); \
86 vpshufb t1, t2, x6; \
87 add2_le128(t2, t0, t4, t3, t5); \
88 vpshufb t1, t2, x7; \
89 vextracti128 $1, t2, t2x; \
124 tivx, t0, t0x, t1, t1x, t2, t2x, t3, \ argument
133 vbroadcasti128 xts_gf128mul_and_shl1_mask_1, t2; \
139 gf128mul_x2_ble(tiv, t1, t2, t0, t3); \
143 gf128mul_x2_ble(tiv, t1, t2, t0, t3); \
147 gf128mul_x2_ble(tiv, t1, t2, t0, t3); \
151 gf128mul_x2_ble(tiv, t1, t2, t0, t3); \
155 gf128mul_x2_ble(tiv, t1, t2, t0, t3); \
159 gf128mul_x2_ble(tiv, t1, t2, t0, t3); \
163 gf128mul_x2_ble(tiv, t1, t2, t0, t3); \