Lines Matching refs:rsp

478         vmovdqa  \XMM8, TMP1(%rsp)
614 …vpxor TMP1(%rsp), \XMM1, \XMM1 # combine GHASHed value with the corresponding cipherte…
636 vmovdqa \XMM2, TMP2(%rsp)
637 vmovdqa \XMM3, TMP3(%rsp)
638 vmovdqa \XMM4, TMP4(%rsp)
639 vmovdqa \XMM5, TMP5(%rsp)
640 vmovdqa \XMM6, TMP6(%rsp)
641 vmovdqa \XMM7, TMP7(%rsp)
642 vmovdqa \XMM8, TMP8(%rsp)
737 vmovdqa TMP2(%rsp), \T1
762 vmovdqa TMP3(%rsp), \T1
785 vmovdqa TMP4(%rsp), \T1
809 vmovdqa TMP5(%rsp), \T1
832 vmovdqa TMP6(%rsp), \T1
856 vmovdqa TMP7(%rsp), \T1
881 vmovdqa TMP8(%rsp), \T1
1155 mov %rsp, %r14
1160 sub $VARIABLE_OFFSET, %rsp
1161 and $~63, %rsp # align rsp to 64 bytes
1326 movb %al, TMP1 (%rsp , %r11)
1331 vmovdqu TMP1(%rsp), %xmm1
1433 mov %r14, %rsp
1454 mov %rsp, %r14
1458 sub $VARIABLE_OFFSET, %rsp
1459 and $~63, %rsp # align rsp to 64 bytes
1483 mov %r14, %rsp
1738 vmovdqa \XMM8, TMP1(%rsp)
1874 vpxor TMP1(%rsp), \XMM1, \XMM1 # combine GHASHed value with
1900 vmovdqa \XMM2, TMP2(%rsp)
1901 vmovdqa \XMM3, TMP3(%rsp)
1902 vmovdqa \XMM4, TMP4(%rsp)
1903 vmovdqa \XMM5, TMP5(%rsp)
1904 vmovdqa \XMM6, TMP6(%rsp)
1905 vmovdqa \XMM7, TMP7(%rsp)
1906 vmovdqa \XMM8, TMP8(%rsp)
1998 vmovdqa TMP2(%rsp), \T1
2024 vmovdqa TMP3(%rsp), \T1
2048 vmovdqa TMP4(%rsp), \T1
2073 vmovdqa TMP5(%rsp), \T1
2097 vmovdqa TMP6(%rsp), \T1
2121 vmovdqa TMP7(%rsp), \T1
2148 vmovdqa TMP8(%rsp), \T1
2432 mov %rsp, %r14
2437 sub $VARIABLE_OFFSET, %rsp
2438 and $~63, %rsp # align rsp to 64 bytes
2603 movb %al, TMP1 (%rsp , %r11)
2608 vmovdqu TMP1(%rsp), %xmm1
2708 mov %r14, %rsp
2730 mov %rsp, %r14
2734 sub $VARIABLE_OFFSET, %rsp
2735 and $~63, %rsp # align rsp to 64 bytes
2759 mov %r14, %rsp