Lines Matching refs:reg

43 #define BLOCK(i, reg) \  in xor_avx_2()  argument
45 asm volatile("vmovdqa %0, %%ymm" #reg : : "m" (p1[i / sizeof(*p1)])); \ in xor_avx_2()
46 asm volatile("vxorps %0, %%ymm" #reg ", %%ymm" #reg : : \ in xor_avx_2()
48 asm volatile("vmovdqa %%ymm" #reg ", %0" : \ in xor_avx_2()
70 #define BLOCK(i, reg) \ in xor_avx_3() argument
72 asm volatile("vmovdqa %0, %%ymm" #reg : : "m" (p2[i / sizeof(*p2)])); \ in xor_avx_3()
73 asm volatile("vxorps %0, %%ymm" #reg ", %%ymm" #reg : : \ in xor_avx_3()
75 asm volatile("vxorps %0, %%ymm" #reg ", %%ymm" #reg : : \ in xor_avx_3()
77 asm volatile("vmovdqa %%ymm" #reg ", %0" : \ in xor_avx_3()
100 #define BLOCK(i, reg) \ in xor_avx_4() argument
102 asm volatile("vmovdqa %0, %%ymm" #reg : : "m" (p3[i / sizeof(*p3)])); \ in xor_avx_4()
103 asm volatile("vxorps %0, %%ymm" #reg ", %%ymm" #reg : : \ in xor_avx_4()
105 asm volatile("vxorps %0, %%ymm" #reg ", %%ymm" #reg : : \ in xor_avx_4()
107 asm volatile("vxorps %0, %%ymm" #reg ", %%ymm" #reg : : \ in xor_avx_4()
109 asm volatile("vmovdqa %%ymm" #reg ", %0" : \ in xor_avx_4()
133 #define BLOCK(i, reg) \ in xor_avx_5() argument
135 asm volatile("vmovdqa %0, %%ymm" #reg : : "m" (p4[i / sizeof(*p4)])); \ in xor_avx_5()
136 asm volatile("vxorps %0, %%ymm" #reg ", %%ymm" #reg : : \ in xor_avx_5()
138 asm volatile("vxorps %0, %%ymm" #reg ", %%ymm" #reg : : \ in xor_avx_5()
140 asm volatile("vxorps %0, %%ymm" #reg ", %%ymm" #reg : : \ in xor_avx_5()
142 asm volatile("vxorps %0, %%ymm" #reg ", %%ymm" #reg : : \ in xor_avx_5()
144 asm volatile("vmovdqa %%ymm" #reg ", %0" : \ in xor_avx_5()