/linux-4.1.27/arch/x86/crypto/ |
H A D | twofish-avx-x86_64-asm_64.S | 59 #define RA2 %xmm4 define 252 * RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2: blocks 254 * RC1, RD1, RA1, RB1, RC2, RD2, RA2, RB2: encrypted blocks 266 inpack_blocks(RA2, RB2, RC2, RD2, RK1, RX0, RY0, RK2); 285 outunpack_blocks(RC2, RD2, RA2, RB2, RK1, RX0, RY0, RK2); 294 * RC1, RD1, RA1, RB1, RC2, RD2, RA2, RB2: encrypted blocks 296 * RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2: decrypted blocks 307 inpack_blocks(RC2, RD2, RA2, RB2, RK1, RX0, RY0, RK2); 308 rotate_1l(RA2); 325 outunpack_blocks(RA2, RB2, RC2, RD2, RK1, RX0, RY0, RK2); 339 load_8way(%rdx, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2); 343 store_8way(%r11, RC1, RD1, RA1, RB1, RC2, RD2, RA2, RB2); 357 load_8way(%rdx, RC1, RD1, RA1, RB1, RC2, RD2, RA2, RB2); 361 store_8way(%r11, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2); 378 load_8way(%rdx, RC1, RD1, RA1, RB1, RC2, RD2, RA2, RB2); 382 store_cbc_8way(%r12, %r11, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2); 402 load_ctr_8way(%rcx, .Lbswap128_mask, RA1, RB1, RC1, RD1, RA2, RB2, RC2, 407 store_ctr_8way(%r12, %r11, RC1, RD1, RA1, RB1, RC2, RD2, RA2, RB2); 425 load_xts_8way(%rcx, %rdx, %rsi, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2, 431 store_xts_8way(%r11, RC1, RD1, RA1, RB1, RC2, RD2, RA2, RB2); 447 load_xts_8way(%rcx, %rdx, %rsi, RC1, RD1, RA1, RB1, RC2, RD2, RA2, RB2, 453 store_xts_8way(%r11, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2);
|
H A D | cast6-avx-x86_64-asm_64.S | 56 #define RA2 %xmm4 define 259 * RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2: blocks 261 * RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2: encrypted blocks 272 inpack_blocks(RA2, RB2, RC2, RD2, RTMP, RX, RKRF, RKM); 296 outunpack_blocks(RA2, RB2, RC2, RD2, RTMP, RX, RKRF, RKM); 305 * RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2: encrypted blocks 307 * RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2: decrypted blocks 318 inpack_blocks(RA2, RB2, RC2, RD2, RTMP, RX, RKRF, RKM); 341 outunpack_blocks(RA2, RB2, RC2, RD2, RTMP, RX, RKRF, RKM); 355 load_8way(%rdx, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2); 359 store_8way(%r11, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2); 373 load_8way(%rdx, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2); 377 store_8way(%r11, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2); 394 load_8way(%rdx, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2); 398 store_cbc_8way(%r12, %r11, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2); 418 load_ctr_8way(%rcx, .Lbswap128_mask, RA1, RB1, RC1, RD1, RA2, RB2, RC2, 423 store_ctr_8way(%r12, %r11, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2); 441 load_xts_8way(%rcx, %rdx, %rsi, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2, 447 store_xts_8way(%r11, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2); 463 load_xts_8way(%rcx, %rdx, %rsi, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2, 469 store_xts_8way(%r11, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2);
|
H A D | serpent-avx-x86_64-asm_64.S | 54 #define RA2 %xmm6 define 574 * RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2: blocks 576 * RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2: encrypted blocks 582 read_blocks(RA2, RB2, RC2, RD2, RK0, RK1, RK2); 619 write_blocks(RA2, RB2, RC2, RD2, RK0, RK1, RK2); 628 * RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2: encrypted blocks 636 read_blocks(RA2, RB2, RC2, RD2, RK0, RK1, RK2); 685 load_8way(%rdx, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2); 689 store_8way(%rsi, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2); 701 load_8way(%rdx, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2); 717 load_8way(%rdx, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2); 734 load_ctr_8way(%rcx, .Lbswap128_mask, RA1, RB1, RC1, RD1, RA2, RB2, RC2, 739 store_ctr_8way(%rdx, %rsi, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2); 753 load_xts_8way(%rcx, %rdx, %rsi, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2, 759 store_xts_8way(%rsi, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2); 773 load_xts_8way(%rcx, %rdx, %rsi, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2,
|
H A D | serpent-avx2-asm_64.S | 40 #define RA2 %ymm3 define 566 * RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2: plaintext 568 * RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2: ciphertext 574 read_blocks(RA2, RB2, RC2, RD2, RK0, RK1, RK2); 611 write_blocks(RA2, RB2, RC2, RD2, RK0, RK1, RK2); 620 * RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2: ciphertext 628 read_blocks(RA2, RB2, RC2, RD2, RK0, RK1, RK2); 679 load_16way(%rdx, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2); 683 store_16way(%rsi, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2); 699 load_16way(%rdx, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2); 719 load_16way(%rdx, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2); 741 load_ctr_16way(%rcx, .Lbswap128_mask, RA1, RB1, RC1, RD1, RA2, RB2, RC2, 747 store_ctr_16way(%rdx, %rsi, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2); 764 load_xts_16way(%rcx, %rdx, %rsi, RA1, RB1, RC1, RD1, RA2, RB2, RC2, 771 store_xts_16way(%rsi, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2); 788 load_xts_16way(%rcx, %rdx, %rsi, RA1, RB1, RC1, RD1, RA2, RB2, RC2,
|
H A D | serpent-sse2-x86_64-asm_64.S | 43 #define RA2 %xmm5 define 649 read_blocks(%rax, RA2, RB2, RC2, RD2, RK0, RK1, RK2); 691 write_blocks(%rax, RA2, RB2, RC2, RD2, RK0, RK1, RK2); 697 xor_blocks(%rax, RA2, RB2, RC2, RD2, RK0, RK1, RK2); 713 read_blocks(%rax, RA2, RB2, RC2, RD2, RK0, RK1, RK2);
|