Searched refs:XMM2 (Results 1 - 5 of 5) sorted by relevance
/linux-4.1.27/arch/x86/crypto/ |
H A D | aesni-intel_avx-x86_64.S | 357 .macro INITIAL_BLOCKS_AVX num_initial_blocks T1 T2 T3 T4 T5 CTR XMM1 XMM2 XMM3 XMM4 XMM5 XMM6 XMM7 XMM8 T6 T_key ENC_DEC 491 vmovdqa \CTR, \XMM2 492 vpshufb SHUF_MASK(%rip), \XMM2, \XMM2 # perform a 16Byte swap 520 vpxor \T_key, \XMM2, \XMM2 533 vaesenc \T_key, \XMM2, \XMM2 547 vaesenclast \T_key, \XMM2, \XMM2 563 vpxor \T1, \XMM2, \XMM2 564 vmovdqu \XMM2, 16*1(arg2 , %r11) 566 vmovdqa \T1, \XMM2 615 vpshufb SHUF_MASK(%rip), \XMM2, \XMM2 # perform a 16Byte swap 633 .macro GHASH_8_ENCRYPT_8_PARALLEL_AVX T1 T2 T3 T4 T5 T6 CTR XMM1 XMM2 XMM3 XMM4 XMM5 XMM6 XMM7 XMM8 T7 loop_idx ENC_DEC 636 vmovdqa \XMM2, TMP2(%rsp) 646 vpaddd ONE(%rip), \XMM1, \XMM2 647 vpaddd ONE(%rip), \XMM2, \XMM3 656 vpshufb SHUF_MASK(%rip), \XMM2, \XMM2 # perform a 16Byte swap 665 vpaddd ONEf(%rip), \XMM1, \XMM2 666 vpaddd ONEf(%rip), \XMM2, \XMM3 680 vpxor \T1, \XMM2, \XMM2 696 vaesenc \T1, \XMM2, \XMM2 706 vaesenc \T1, \XMM2, \XMM2 729 vaesenc \T1, \XMM2, \XMM2 752 vaesenc \T1, \XMM2, \XMM2 777 vaesenc \T1, \XMM2, \XMM2 800 vaesenc \T1, \XMM2, \XMM2 824 vaesenc \T1, \XMM2, \XMM2 848 vaesenc \T1, \XMM2, \XMM2 873 vaesenc \T5, \XMM2, \XMM2 942 vmovdqu \XMM2, 16*1(arg2,%r11) # Write to the Ciphertext buffer 965 vpshufb SHUF_MASK(%rip), \XMM2, \XMM2 # perform a 16Byte swap 982 .macro GHASH_LAST_8_AVX T1 T2 T3 T4 T5 T6 T7 XMM1 XMM2 XMM3 XMM4 XMM5 XMM6 XMM7 XMM8 998 vpshufd $0b01001110, \XMM2, \T2 999 vpxor \XMM2, \T2, \T2 1001 vpclmulqdq $0x11, \T5, \XMM2, \T4 1004 vpclmulqdq $0x00, \T5, \XMM2, \T4 1616 .macro INITIAL_BLOCKS_AVX2 num_initial_blocks T1 T2 T3 T4 T5 CTR XMM1 XMM2 XMM3 XMM4 XMM5 XMM6 XMM7 XMM8 T6 T_key ENC_DEC VER 1751 vmovdqa \CTR, \XMM2 1752 vpshufb SHUF_MASK(%rip), \XMM2, \XMM2 # perform a 16Byte swap 1780 vpxor \T_key, \XMM2, \XMM2 1793 vaesenc \T_key, \XMM2, \XMM2 1807 vaesenclast \T_key, \XMM2, \XMM2 1823 vpxor \T1, \XMM2, \XMM2 1824 vmovdqu \XMM2, 16*1(arg2 , %r11) 1826 vmovdqa \T1, \XMM2 1876 vpshufb SHUF_MASK(%rip), \XMM2, \XMM2 # perform a 16Byte swap 1897 .macro GHASH_8_ENCRYPT_8_PARALLEL_AVX2 T1 T2 T3 T4 T5 T6 CTR XMM1 XMM2 XMM3 XMM4 XMM5 XMM6 XMM7 XMM8 T7 loop_idx ENC_DEC 1900 vmovdqa \XMM2, TMP2(%rsp) 1910 vpaddd ONE(%rip), \XMM1, \XMM2 1911 vpaddd ONE(%rip), \XMM2, \XMM3 1920 vpshufb SHUF_MASK(%rip), \XMM2, \XMM2 # perform a 16Byte swap 1929 vpaddd ONEf(%rip), \XMM1, \XMM2 1930 vpaddd ONEf(%rip), \XMM2, \XMM3 1944 vpxor \T1, \XMM2, \XMM2 1960 vaesenc \T1, \XMM2, \XMM2 1970 vaesenc \T1, \XMM2, \XMM2 1990 vaesenc \T1, \XMM2, \XMM2 2014 vaesenc \T1, \XMM2, \XMM2 2040 vaesenc \T1, \XMM2, \XMM2 2064 vaesenc \T1, \XMM2, \XMM2 2089 vaesenc \T1, \XMM2, \XMM2 2113 vaesenc \T1, \XMM2, \XMM2 2140 vaesenc \T5, \XMM2, \XMM2 2203 vmovdqu \XMM2, 16*1(arg2,%r11) # Write to the Ciphertext buffer 2225 vpshufb SHUF_MASK(%rip), \XMM2, \XMM2 # perform a 16Byte swap 2242 .macro GHASH_LAST_8_AVX2 T1 T2 T3 T4 T5 T6 T7 XMM1 XMM2 XMM3 XMM4 XMM5 XMM6 XMM7 XMM8 2261 vpshufd $0b01001110, \XMM2, \T2 2263 vpxor \XMM2, \T2, \T2 2266 vpclmulqdq $0x11, \T5, \XMM2, \T4 2269 vpclmulqdq $0x00, \T5, \XMM2, \T4
|
H A D | aesni-intel_asm.S | 227 XMM2 XMM3 XMM4 XMMDst TMP6 TMP7 i i_seq operation label 336 MOVADQ \XMM0, \XMM2 337 PSHUFB_XMM %xmm14, \XMM2 # perform a 16 byte swap 349 pxor \TMP1, \XMM2 366 AESENC \TMP1, \XMM2 379 AESENC \TMP1, \XMM2 407 AESENCLAST \TMP2, \XMM2 415 pxor \TMP1, \XMM2 416 movdqu \XMM2, 16*1(%arg2 , %r11 , 1) 417 movdqa \TMP1, \XMM2 430 PSHUFB_XMM %xmm14, \XMM2 # perform a 16 byte swap 452 XMM2 XMM3 XMM4 XMMDst TMP6 TMP7 i i_seq operation label 557 MOVADQ \XMM0, \XMM2 558 PSHUFB_XMM %xmm14, \XMM2 # perform a 16 byte swap 570 pxor \TMP1, \XMM2 587 AESENC \TMP1, \XMM2 600 AESENC \TMP1, \XMM2 628 AESENCLAST \TMP2, \XMM2 634 pxor \TMP1, \XMM2 640 movdqu \XMM2, 16*1(%arg2 , %r11 , 1) 648 PSHUFB_XMM %xmm14, \XMM2 # perform a 16 byte swap 663 TMP6 XMM0 XMM1 XMM2 XMM3 XMM4 XMM5 XMM6 XMM7 XMM8 operation 666 movdqa \XMM2, \XMM6 681 movdqa \XMM0, \XMM2 688 PSHUFB_XMM %xmm15, \XMM2 # perform a 16 byte swap 693 pxor (%arg1), \XMM2 700 AESENC \TMP1, \XMM2 705 AESENC \TMP1, \XMM2 715 AESENC \TMP3, \XMM2 721 AESENC \TMP3, \XMM2 728 AESENC \TMP3, \XMM2 745 AESENC \TMP3, \XMM2 751 AESENC \TMP3, \XMM2 758 AESENC \TMP3, \XMM2 776 AESENC \TMP3, \XMM2 798 AESENCLAST \TMP3, \XMM2 806 pxor \TMP3, \XMM2 # Ciphertext/Plaintext XOR EK 812 movdqu \XMM2, 16(%arg2,%r11,1) # Write to the ciphertext buffer 816 PSHUFB_XMM %xmm15, \XMM2 # perform a 16 byte swap 871 TMP6 XMM0 XMM1 XMM2 XMM3 XMM4 XMM5 XMM6 XMM7 XMM8 operation 874 movdqa \XMM2, \XMM6 889 movdqa \XMM0, \XMM2 896 PSHUFB_XMM %xmm15, \XMM2 # perform a 16 byte swap 901 pxor (%arg1), \XMM2 908 AESENC \TMP1, \XMM2 913 AESENC \TMP1, \XMM2 923 AESENC \TMP3, \XMM2 929 AESENC \TMP3, \XMM2 936 AESENC \TMP3, \XMM2 953 AESENC \TMP3, \XMM2 959 AESENC \TMP3, \XMM2 966 AESENC \TMP3, \XMM2 984 AESENC \TMP3, \XMM2 1006 AESENCLAST \TMP3, \XMM2 1016 pxor \TMP3, \XMM2 # Ciphertext/Plaintext XOR EK 1017 movdqu \XMM2, 16(%arg2,%r11,1) # Write to plaintext buffer 1018 movdqa \TMP3, \XMM2 1028 PSHUFB_XMM %xmm15, \XMM2 # perform a 16 byte swap 1078 TMP7 XMM1 XMM2 XMM3 XMM4 XMMDst 1095 movdqa \XMM2, \TMP1 1096 pshufd $78, \XMM2, \TMP2 1097 pxor \XMM2, \TMP2 1100 PCLMULQDQ 0x00, \TMP5, \XMM2 # XMM2 = a0*b0 1104 pxor \XMM2, \XMMDst
|
H A D | sha512-ssse3-asm.S | 166 # Eg. XMM2=W[t-2] really means XMM2={W[t-2]|W[t-1]} 175 movdqa W_t(idx), %xmm2 # XMM2 = W[t-2]
|
H A D | sha512-avx-asm.S | 195 vpsrlq $6, %xmm4, %xmm2 # XMM2 = W[t-2]>>6 208 vpxor %xmm3, %xmm2, %xmm2 # XMM2 = W[t-2]>>6 ^ W[t-2]<<3
|
/linux-4.1.27/arch/x86/kvm/ |
H A D | cpuid.c | 307 F(FXSR) | F(XMM) | F(XMM2) | F(SELFSNOOP) | __do_cpuid_ent()
|
Completed in 138 milliseconds