Lines Matching refs:r12
354 ## r10, r11, r12, rax are clobbered
362 mov arg7, %r12 # r12 = aadLen
365 mov %r12, %r11
375 sub $4, %r12
381 mov $16, %r12
385 sub $4, %r12
386 cmp %r11, %r12
1146 # clobbering r10, r11, r12, r13, r14, r15
1150 push %r12
1169 mov %r13, %r12
1170 shr $4, %r12
1171 and $7, %r12
1174 cmp $7, %r12
1176 cmp $6, %r12
1178 cmp $5, %r12
1180 cmp $4, %r12
1182 cmp $3, %r12
1184 cmp $2, %r12
1296 lea SHIFT_MASK+16(%rip), %r12
1297 sub %r13, %r12 # adjust the shuffle mask pointer to be
1300 vmovdqu (%r12), %xmm2 # get the appropriate shuffle mask
1319 lea SHIFT_MASK+16(%rip), %r12
1320 sub %r13, %r12 # adjust the shuffle mask pointer to be
1339 vmovdqu ALL_F-SHIFT_MASK(%r12), %xmm1 # get the appropriate mask to
1351 vmovdqu ALL_F-SHIFT_MASK(%r12), %xmm1 # get the appropriate mask to
1385 mov arg7, %r12 # r12 = aadLen (number of bytes)
1386 shl $3, %r12 # convert into number of bits
1438 pop %r12
1449 push %r12
1488 pop %r12
1613 ## r10, r11, r12, rax are clobbered
1621 mov arg7, %r12 # r12 = aadLen
1624 mov %r12, %r11
1634 sub $4, %r12
1640 mov $16, %r12
1644 sub $4, %r12
1645 cmp %r11, %r12
2423 # clobbering r10, r11, r12, r13, r14, r15
2427 push %r12
2446 mov %r13, %r12
2447 shr $4, %r12
2448 and $7, %r12
2451 cmp $7, %r12
2453 cmp $6, %r12
2455 cmp $5, %r12
2457 cmp $4, %r12
2459 cmp $3, %r12
2461 cmp $2, %r12
2573 lea SHIFT_MASK+16(%rip), %r12
2574 sub %r13, %r12 # adjust the shuffle mask pointer
2577 vmovdqu (%r12), %xmm2 # get the appropriate shuffle mask
2596 lea SHIFT_MASK+16(%rip), %r12
2597 sub %r13, %r12 # adjust the shuffle mask pointer to be
2616 …vmovdqu ALL_F-SHIFT_MASK(%r12), %xmm1 # get the appropriate mask to mask out top 16-r13 byt…
2627 …vmovdqu ALL_F-SHIFT_MASK(%r12), %xmm1 # get the appropriate mask to mask out top 16-r13 byt…
2660 mov arg7, %r12 # r12 = aadLen (number of bytes)
2661 shl $3, %r12 # convert into number of bits
2713 pop %r12
2725 push %r12
2764 pop %r12