Lines Matching refs:t4

66 #define roundsm32(x0, x1, x2, x3, x4, x5, x6, x7, t0, t1, t2, t3, t4, t5, t6, \  argument
71 vbroadcasti128 .Linv_shift_row, t4; \
79 vpshufb t4, x0, x0; \
80 vpshufb t4, x7, x7; \
81 vpshufb t4, x3, x3; \
82 vpshufb t4, x6, x6; \
83 vpshufb t4, x2, x2; \
84 vpshufb t4, x5, x5; \
85 vpshufb t4, x1, x1; \
86 vpshufb t4, x4, x4; \
90 filter_8bit(x0, t5, t6, t7, t4); \
91 filter_8bit(x7, t5, t6, t7, t4); \
94 filter_8bit(x3, t2, t3, t7, t4); \
95 filter_8bit(x6, t2, t3, t7, t4); \
98 filter_8bit(x2, t5, t6, t7, t4); \
99 filter_8bit(x5, t5, t6, t7, t4); \
100 filter_8bit(x1, t5, t6, t7, t4); \
101 filter_8bit(x4, t5, t6, t7, t4); \
103 vpxor t4##_x, t4##_x, t4##_x; \
108 vaesenclast t4##_x, x0##_x, x0##_x; \
109 vaesenclast t4##_x, t0##_x, t0##_x; \
111 vaesenclast t4##_x, x7##_x, x7##_x; \
112 vaesenclast t4##_x, t1##_x, t1##_x; \
114 vaesenclast t4##_x, x3##_x, x3##_x; \
115 vaesenclast t4##_x, t3##_x, t3##_x; \
117 vaesenclast t4##_x, x6##_x, x6##_x; \
118 vaesenclast t4##_x, t2##_x, t2##_x; \
124 vaesenclast t4##_x, x2##_x, x2##_x; \
125 vaesenclast t4##_x, t6##_x, t6##_x; \
127 vaesenclast t4##_x, x5##_x, x5##_x; \
128 vaesenclast t4##_x, t5##_x, t5##_x; \
130 vaesenclast t4##_x, x1##_x, x1##_x; \
131 vaesenclast t4##_x, t3##_x, t3##_x; \
133 vaesenclast t4##_x, x4##_x, x4##_x; \
134 vaesenclast t4##_x, t2##_x, t2##_x; \
146 vbroadcasti128 .Lpost_tf_lo_s2, t4; \
154 filter_8bit(x1, t4, t5, t7, t2); \
155 filter_8bit(x4, t4, t5, t7, t2); \
170 vpsrldq $4, t0, t4; \
173 vpshufb t7, t4, t4; \
209 vpxor t4, x3, x3; \