Searched refs:XMM2 (Results 1 - 5 of 5) sorted by relevance

/linux-4.1.27/arch/x86/crypto/
H A Daesni-intel_avx-x86_64.S357 .macro INITIAL_BLOCKS_AVX num_initial_blocks T1 T2 T3 T4 T5 CTR XMM1 XMM2 XMM3 XMM4 XMM5 XMM6 XMM7 XMM8 T6 T_key ENC_DEC
491 vmovdqa \CTR, \XMM2
492 vpshufb SHUF_MASK(%rip), \XMM2, \XMM2 # perform a 16Byte swap
520 vpxor \T_key, \XMM2, \XMM2
533 vaesenc \T_key, \XMM2, \XMM2
547 vaesenclast \T_key, \XMM2, \XMM2
563 vpxor \T1, \XMM2, \XMM2
564 vmovdqu \XMM2, 16*1(arg2 , %r11)
566 vmovdqa \T1, \XMM2
615 vpshufb SHUF_MASK(%rip), \XMM2, \XMM2 # perform a 16Byte swap
633 .macro GHASH_8_ENCRYPT_8_PARALLEL_AVX T1 T2 T3 T4 T5 T6 CTR XMM1 XMM2 XMM3 XMM4 XMM5 XMM6 XMM7 XMM8 T7 loop_idx ENC_DEC
636 vmovdqa \XMM2, TMP2(%rsp)
646 vpaddd ONE(%rip), \XMM1, \XMM2
647 vpaddd ONE(%rip), \XMM2, \XMM3
656 vpshufb SHUF_MASK(%rip), \XMM2, \XMM2 # perform a 16Byte swap
665 vpaddd ONEf(%rip), \XMM1, \XMM2
666 vpaddd ONEf(%rip), \XMM2, \XMM3
680 vpxor \T1, \XMM2, \XMM2
696 vaesenc \T1, \XMM2, \XMM2
706 vaesenc \T1, \XMM2, \XMM2
729 vaesenc \T1, \XMM2, \XMM2
752 vaesenc \T1, \XMM2, \XMM2
777 vaesenc \T1, \XMM2, \XMM2
800 vaesenc \T1, \XMM2, \XMM2
824 vaesenc \T1, \XMM2, \XMM2
848 vaesenc \T1, \XMM2, \XMM2
873 vaesenc \T5, \XMM2, \XMM2
942 vmovdqu \XMM2, 16*1(arg2,%r11) # Write to the Ciphertext buffer
965 vpshufb SHUF_MASK(%rip), \XMM2, \XMM2 # perform a 16Byte swap
982 .macro GHASH_LAST_8_AVX T1 T2 T3 T4 T5 T6 T7 XMM1 XMM2 XMM3 XMM4 XMM5 XMM6 XMM7 XMM8
998 vpshufd $0b01001110, \XMM2, \T2
999 vpxor \XMM2, \T2, \T2
1001 vpclmulqdq $0x11, \T5, \XMM2, \T4
1004 vpclmulqdq $0x00, \T5, \XMM2, \T4
1616 .macro INITIAL_BLOCKS_AVX2 num_initial_blocks T1 T2 T3 T4 T5 CTR XMM1 XMM2 XMM3 XMM4 XMM5 XMM6 XMM7 XMM8 T6 T_key ENC_DEC VER
1751 vmovdqa \CTR, \XMM2
1752 vpshufb SHUF_MASK(%rip), \XMM2, \XMM2 # perform a 16Byte swap
1780 vpxor \T_key, \XMM2, \XMM2
1793 vaesenc \T_key, \XMM2, \XMM2
1807 vaesenclast \T_key, \XMM2, \XMM2
1823 vpxor \T1, \XMM2, \XMM2
1824 vmovdqu \XMM2, 16*1(arg2 , %r11)
1826 vmovdqa \T1, \XMM2
1876 vpshufb SHUF_MASK(%rip), \XMM2, \XMM2 # perform a 16Byte swap
1897 .macro GHASH_8_ENCRYPT_8_PARALLEL_AVX2 T1 T2 T3 T4 T5 T6 CTR XMM1 XMM2 XMM3 XMM4 XMM5 XMM6 XMM7 XMM8 T7 loop_idx ENC_DEC
1900 vmovdqa \XMM2, TMP2(%rsp)
1910 vpaddd ONE(%rip), \XMM1, \XMM2
1911 vpaddd ONE(%rip), \XMM2, \XMM3
1920 vpshufb SHUF_MASK(%rip), \XMM2, \XMM2 # perform a 16Byte swap
1929 vpaddd ONEf(%rip), \XMM1, \XMM2
1930 vpaddd ONEf(%rip), \XMM2, \XMM3
1944 vpxor \T1, \XMM2, \XMM2
1960 vaesenc \T1, \XMM2, \XMM2
1970 vaesenc \T1, \XMM2, \XMM2
1990 vaesenc \T1, \XMM2, \XMM2
2014 vaesenc \T1, \XMM2, \XMM2
2040 vaesenc \T1, \XMM2, \XMM2
2064 vaesenc \T1, \XMM2, \XMM2
2089 vaesenc \T1, \XMM2, \XMM2
2113 vaesenc \T1, \XMM2, \XMM2
2140 vaesenc \T5, \XMM2, \XMM2
2203 vmovdqu \XMM2, 16*1(arg2,%r11) # Write to the Ciphertext buffer
2225 vpshufb SHUF_MASK(%rip), \XMM2, \XMM2 # perform a 16Byte swap
2242 .macro GHASH_LAST_8_AVX2 T1 T2 T3 T4 T5 T6 T7 XMM1 XMM2 XMM3 XMM4 XMM5 XMM6 XMM7 XMM8
2261 vpshufd $0b01001110, \XMM2, \T2
2263 vpxor \XMM2, \T2, \T2
2266 vpclmulqdq $0x11, \T5, \XMM2, \T4
2269 vpclmulqdq $0x00, \T5, \XMM2, \T4
H A Daesni-intel_asm.S227 XMM2 XMM3 XMM4 XMMDst TMP6 TMP7 i i_seq operation label
336 MOVADQ \XMM0, \XMM2
337 PSHUFB_XMM %xmm14, \XMM2 # perform a 16 byte swap
349 pxor \TMP1, \XMM2
366 AESENC \TMP1, \XMM2
379 AESENC \TMP1, \XMM2
407 AESENCLAST \TMP2, \XMM2
415 pxor \TMP1, \XMM2
416 movdqu \XMM2, 16*1(%arg2 , %r11 , 1)
417 movdqa \TMP1, \XMM2
430 PSHUFB_XMM %xmm14, \XMM2 # perform a 16 byte swap
452 XMM2 XMM3 XMM4 XMMDst TMP6 TMP7 i i_seq operation label
557 MOVADQ \XMM0, \XMM2
558 PSHUFB_XMM %xmm14, \XMM2 # perform a 16 byte swap
570 pxor \TMP1, \XMM2
587 AESENC \TMP1, \XMM2
600 AESENC \TMP1, \XMM2
628 AESENCLAST \TMP2, \XMM2
634 pxor \TMP1, \XMM2
640 movdqu \XMM2, 16*1(%arg2 , %r11 , 1)
648 PSHUFB_XMM %xmm14, \XMM2 # perform a 16 byte swap
663 TMP6 XMM0 XMM1 XMM2 XMM3 XMM4 XMM5 XMM6 XMM7 XMM8 operation
666 movdqa \XMM2, \XMM6
681 movdqa \XMM0, \XMM2
688 PSHUFB_XMM %xmm15, \XMM2 # perform a 16 byte swap
693 pxor (%arg1), \XMM2
700 AESENC \TMP1, \XMM2
705 AESENC \TMP1, \XMM2
715 AESENC \TMP3, \XMM2
721 AESENC \TMP3, \XMM2
728 AESENC \TMP3, \XMM2
745 AESENC \TMP3, \XMM2
751 AESENC \TMP3, \XMM2
758 AESENC \TMP3, \XMM2
776 AESENC \TMP3, \XMM2
798 AESENCLAST \TMP3, \XMM2
806 pxor \TMP3, \XMM2 # Ciphertext/Plaintext XOR EK
812 movdqu \XMM2, 16(%arg2,%r11,1) # Write to the ciphertext buffer
816 PSHUFB_XMM %xmm15, \XMM2 # perform a 16 byte swap
871 TMP6 XMM0 XMM1 XMM2 XMM3 XMM4 XMM5 XMM6 XMM7 XMM8 operation
874 movdqa \XMM2, \XMM6
889 movdqa \XMM0, \XMM2
896 PSHUFB_XMM %xmm15, \XMM2 # perform a 16 byte swap
901 pxor (%arg1), \XMM2
908 AESENC \TMP1, \XMM2
913 AESENC \TMP1, \XMM2
923 AESENC \TMP3, \XMM2
929 AESENC \TMP3, \XMM2
936 AESENC \TMP3, \XMM2
953 AESENC \TMP3, \XMM2
959 AESENC \TMP3, \XMM2
966 AESENC \TMP3, \XMM2
984 AESENC \TMP3, \XMM2
1006 AESENCLAST \TMP3, \XMM2
1016 pxor \TMP3, \XMM2 # Ciphertext/Plaintext XOR EK
1017 movdqu \XMM2, 16(%arg2,%r11,1) # Write to plaintext buffer
1018 movdqa \TMP3, \XMM2
1028 PSHUFB_XMM %xmm15, \XMM2 # perform a 16 byte swap
1078 TMP7 XMM1 XMM2 XMM3 XMM4 XMMDst
1095 movdqa \XMM2, \TMP1
1096 pshufd $78, \XMM2, \TMP2
1097 pxor \XMM2, \TMP2
1100 PCLMULQDQ 0x00, \TMP5, \XMM2 # XMM2 = a0*b0
1104 pxor \XMM2, \XMMDst
H A Dsha512-ssse3-asm.S166 # Eg. XMM2=W[t-2] really means XMM2={W[t-2]|W[t-1]}
175 movdqa W_t(idx), %xmm2 # XMM2 = W[t-2]
H A Dsha512-avx-asm.S195 vpsrlq $6, %xmm4, %xmm2 # XMM2 = W[t-2]>>6
208 vpxor %xmm3, %xmm2, %xmm2 # XMM2 = W[t-2]>>6 ^ W[t-2]<<3
/linux-4.1.27/arch/x86/kvm/
H A Dcpuid.c307 F(FXSR) | F(XMM) | F(XMM2) | F(SELFSNOOP) | __do_cpuid_ent()

Completed in 138 milliseconds