Searched refs:r1 (Results 1 - 200 of 696) sorted by relevance

1234

/linux-4.1.27/arch/powerpc/platforms/ps3/
H A Dhvcall.S31 std r0, 16(r1); \
36 ld r0, 16(r1); \
53 std r0, 16(r1); \
55 stdu r3, -8(r1); \
60 addi r1, r1, 8; \
61 ld r11, -8(r1); \
64 ld r0, 16(r1); \
72 std r0, 16(r1); \
74 std r3, -8(r1); \
75 stdu r4, -16(r1); \
80 addi r1, r1, 16; \
81 ld r11, -8(r1); \
83 ld r11, -16(r1); \
86 ld r0, 16(r1); \
94 std r0, 16(r1); \
96 std r3, -8(r1); \
97 std r4, -16(r1); \
98 stdu r5, -24(r1); \
103 addi r1, r1, 24; \
104 ld r11, -8(r1); \
106 ld r11, -16(r1); \
108 ld r11, -24(r1); \
111 ld r0, 16(r1); \
119 std r0, 16(r1); \
121 std r3, -8(r1); \
122 std r4, -16(r1); \
123 std r5, -24(r1); \
124 std r6, -32(r1); \
125 std r7, -40(r1); \
126 std r8, -48(r1); \
127 stdu r9, -56(r1); \
132 addi r1, r1, 56; \
133 ld r11, -8(r1); \
135 ld r11, -16(r1); \
137 ld r11, -24(r1); \
139 ld r11, -32(r1); \
141 ld r11, -40(r1); \
143 ld r11, -48(r1); \
145 ld r11, -56(r1); \
148 ld r0, 16(r1); \
156 std r0, 16(r1); \
158 stdu r4, -8(r1); \
163 addi r1, r1, 8; \
164 ld r11, -8(r1); \
167 ld r0, 16(r1); \
175 std r0, 16(r1); \
177 std r4, -8(r1); \
178 stdu r5, -16(r1); \
183 addi r1, r1, 16; \
184 ld r11, -8(r1); \
186 ld r11, -16(r1); \
189 ld r0, 16(r1); \
197 std r0, 16(r1); \
199 std r4, -8(r1); \
200 std r5, -16(r1); \
201 stdu r6, -24(r1); \
206 addi r1, r1, 24; \
207 ld r11, -8(r1); \
209 ld r11, -16(r1); \
211 ld r11, -24(r1); \
214 ld r0, 16(r1); \
222 std r0, 16(r1); \
224 std r4, -8(r1); \
225 std r5, -16(r1); \
226 std r6, -24(r1); \
227 stdu r7, -32(r1); \
232 addi r1, r1, 32; \
233 ld r11, -8(r1); \
235 ld r11, -16(r1); \
237 ld r11, -24(r1); \
239 ld r11, -32(r1); \
242 ld r0, 16(r1); \
250 std r0, 16(r1); \
252 std r4, -8(r1); \
253 std r5, -16(r1); \
254 std r6, -24(r1); \
255 std r7, -32(r1); \
256 stdu r8, -40(r1); \
261 addi r1, r1, 40; \
262 ld r11, -8(r1); \
264 ld r11, -16(r1); \
266 ld r11, -24(r1); \
268 ld r11, -32(r1); \
270 ld r11, -40(r1); \
273 ld r0, 16(r1); \
281 std r0, 16(r1); \
283 std r4, -8(r1); \
284 std r5, -16(r1); \
285 std r6, -24(r1); \
286 std r7, -32(r1); \
287 std r8, -40(r1); \
288 stdu r9, -48(r1); \
293 addi r1, r1, 48; \
294 ld r11, -8(r1); \
296 ld r11, -16(r1); \
298 ld r11, -24(r1); \
300 ld r11, -32(r1); \
302 ld r11, -40(r1); \
304 ld r11, -48(r1); \
307 ld r0, 16(r1); \
315 std r0, 16(r1); \
317 std r4, -8(r1); \
318 std r5, -16(r1); \
319 std r6, -24(r1); \
320 std r7, -32(r1); \
321 std r8, -40(r1); \
322 std r9, -48(r1); \
323 stdu r10, -56(r1); \
328 addi r1, r1, 56; \
329 ld r11, -8(r1); \
331 ld r11, -16(r1); \
333 ld r11, -24(r1); \
335 ld r11, -32(r1); \
337 ld r11, -40(r1); \
339 ld r11, -48(r1); \
341 ld r11, -56(r1); \
344 ld r0, 16(r1); \
352 std r0, 16(r1); \
354 stdu r5, -8(r1); \
359 addi r1, r1, 8; \
360 ld r11, -8(r1); \
363 ld r0, 16(r1); \
371 std r0, 16(r1); \
373 std r5, -8(r1); \
374 stdu r6, -16(r1); \
379 addi r1, r1, 16; \
380 ld r11, -8(r1); \
382 ld r11, -16(r1); \
385 ld r0, 16(r1); \
393 std r0, 16(r1); \
395 std r5, -8(r1); \
396 std r6, -16(r1); \
397 stdu r7, -24(r1); \
402 addi r1, r1, 24; \
403 ld r11, -8(r1); \
405 ld r11, -16(r1); \
407 ld r11, -24(r1); \
410 ld r0, 16(r1); \
418 std r0, 16(r1); \
420 std r5, -8(r1); \
421 std r6, -16(r1); \
422 std r7, -24(r1); \
423 stdu r8, -32(r1); \
428 addi r1, r1, 32; \
429 ld r11, -8(r1); \
431 ld r11, -16(r1); \
433 ld r11, -24(r1); \
435 ld r11, -32(r1); \
438 ld r0, 16(r1); \
446 std r0, 16(r1); \
448 std r5, -8(r1); \
449 std r6, -16(r1); \
450 std r7, -24(r1); \
451 std r8, -32(r1); \
452 stdu r9, -40(r1); \
457 addi r1, r1, 40; \
458 ld r11, -8(r1); \
460 ld r11, -16(r1); \
462 ld r11, -24(r1); \
464 ld r11, -32(r1); \
466 ld r11, -40(r1); \
469 ld r0, 16(r1); \
477 std r0, 16(r1); \
479 stdu r6, -8(r1); \
484 addi r1, r1, 8; \
485 ld r11, -8(r1); \
488 ld r0, 16(r1); \
496 std r0, 16(r1); \
498 std r6, -8(r1); \
499 stdu r7, -16(r1); \
504 addi r1, r1, 16; \
505 ld r11, -8(r1); \
507 ld r11, -16(r1); \
510 ld r0, 16(r1); \
518 std r0, 16(r1); \
520 std r6, -8(r1); \
521 std r7, -16(r1); \
522 stdu r8, -24(r1); \
527 addi r1, r1, 24; \
528 ld r11, -8(r1); \
530 ld r11, -16(r1); \
532 ld r11, -24(r1); \
535 ld r0, 16(r1); \
543 std r0, 16(r1); \
545 stdu r7, -8(r1); \
550 addi r1, r1, 8; \
551 ld r11, -8(r1); \
554 ld r0, 16(r1); \
562 std r0, 16(r1); \
564 std r7, -8(r1); \
565 stdu r8, -16(r1); \
570 addi r1, r1, 16; \
571 ld r11, -8(r1); \
573 ld r11, -16(r1); \
576 ld r0, 16(r1); \
584 std r0, 16(r1); \
586 std r7, -8(r1); \
587 std r8, -16(r1); \
588 stdu r9, -24(r1); \
593 addi r1, r1, 24; \
594 ld r11, -8(r1); \
596 ld r11, -16(r1); \
598 ld r11, -24(r1); \
601 ld r0, 16(r1); \
609 std r0, 16(r1); \
611 stdu r8, -8(r1); \
616 addi r1, r1, 8; \
617 ld r11, -8(r1); \
620 ld r0, 16(r1); \
628 std r0, 16(r1); \
630 std r8, -8(r1); \
631 stdu r9, -16(r1); \
636 addi r1, r1, 16; \
637 ld r11, -8(r1); \
639 ld r11, -16(r1); \
642 ld r0, 16(r1); \
650 std r0, 16(r1); \
652 std r8, -8(r1); \
653 std r9, -16(r1); \
654 stdu r10, -24(r1); \
659 addi r1, r1, 24; \
660 ld r11, -8(r1); \
662 ld r11, -16(r1); \
664 ld r11, -24(r1); \
667 ld r0, 16(r1); \
675 std r0, 16(r1); \
677 stdu r9, -8(r1); \
682 addi r1, r1, 8; \
683 ld r11, -8(r1); \
686 ld r0, 16(r1); \
694 std r0, 16(r1); \
696 std r9, -8(r1); \
697 stdu r10, -16(r1); \
702 addi r1, r1, 16; \
703 ld r11, -8(r1); \
705 ld r11, -16(r1); \
708 ld r0, 16(r1); \
716 std r0, 16(r1); \
718 std r9, -8(r1); \
719 stdu r10, -16(r1); \
724 addi r1, r1, 16; \
725 ld r11, -8(r1); \
727 ld r11, -16(r1); \
729 ld r11, 48+8*8(r1); \
732 ld r0, 16(r1); \
740 std r0, 16(r1); \
742 stdu r10, -8(r1); \
747 addi r1, r1, 8; \
748 ld r11, -8(r1); \
751 ld r0, 16(r1); \
759 std r0, 16(r1); \
761 std r10, 48+8*7(r1); \
766 ld r11, 48+8*7(r1); \
768 ld r11, 48+8*8(r1); \
770 ld r11, 48+8*9(r1); \
772 ld r11, 48+8*10(r1); \
774 ld r11, 48+8*11(r1); \
776 ld r11, 48+8*12(r1); \
779 ld r0, 16(r1); \
787 std r0, 16(r1); \
792 ld r11, 48+8*8(r1); \
795 ld r0, 16(r1); \
/linux-4.1.27/tools/testing/selftests/powerpc/stringloops/asm/
H A Dppc_asm.h3 #ifndef r1
4 #define r1 sp macro
/linux-4.1.27/arch/arm/boot/compressed/
H A Ddebug.S9 addruart r1, r2, r3
10 waituart r3, r1
11 senduart r0, r1
12 busyuart r3, r1
19 adr r1, 1f
20 ldmia r1, {r2, r3}
21 add r2, r2, r1
22 ldr r1, [r2, r3]
23 strb r0, [r1]
H A Dll_char_wr.S32 * r1 = char
43 mov r1, r1, lsl #3
60 orr r1, r1, #7
61 ldrb r7, [r6, r1]
70 sub r1, r1, #1 @ avoid using r7 directly after
72 ldrb r7, [r6, r1]
75 tst r1, #7 @ avoid using r7 directly after
77 subne r1, r1, #1
78 ldrneb r7, [r6, r1]
92 sub r1, r1, #1 @ avoid ip
95 ldrb r7, [r6, r1]
102 tst r1, #7 @ avoid ip
105 subne r1, r1, #1
106 ldrneb r7, [r6, r1]
114 add r6, r6, r1
H A Dhead-sharpsl.S28 mov r1, #0x10000000 @ Base address of TC6393 chip
30 ldrh r3, [r1, #8] @ Load TC6393XB Revison: This is 0x0003
42 ldr r1, .W100ADDR @ Base address of w100 chip + regs offset
45 str r6, [r1, #0x280] @ to SCRATCH_UMSK
51 str r6, [r1, #0x280] @ to SCRATCH_UMSK
53 ldr r6, [r1, #0] @ Load Chip ID
83 ldr r1, .SCOOP2ADDR
86 strh r6, [r1]
87 ldrh r6, [r1]
125 * Corrupts: r1
128 mov r1, #0x0c000000 @ Base address of NAND chip
129 ldrb r3, [r1, #24] @ Load FLASHCTL
132 strb r3, [r1, #24] @ Save to FLASHCTL
134 strb r2, [r1, #20] @ Save to FLASHIO
137 strb r3, [r1, #24] @ Save to FLASHCTL
139 strb r2, [r1, #20] @ Save to FLASHIO
141 strb r3, [r1, #24] @ Save to FLASHCTL
143 ldrb r3, [r1, #24] @ Load FLASHCTL
146 ldrb r2, [r1, #20] @ NAND Manufacturer ID
147 ldrb r3, [r1, #20] @ NAND Chip ID
H A Dhead.S81 mov r1, #\len
148 mov r7, r1 @ save architecture ID
221 ldmia r0, {r1, r2, r3, r6, r10, r11, r12}
228 sub r0, r0, r1 @ calculate the delta offset
281 ldr r1, =0xedfe0dd0 @ sig is 0xd00dfeed big endian
283 ldr r1, =0xd00dfeed
285 cmp lr, r1
302 eor r1, r5, r5, ror #16
303 bic r1, r1, #0x00ff0000
305 eor r5, r5, r1, lsr #8
322 mov r1, r6
335 mov r1, r6
352 adr r1, wont_overwrite
353 sub r1, r6, r1
354 subs r1, r5, r1
355 addhi r9, r9, r1
361 eor r1, r5, r5, ror #16
362 bic r1, r1, #0x00ff0000
364 eor r5, r5, r1, lsr #8
468 orrs r1, r0, r5
487 1: ldr r1, [r11, #0] @ relocate entries in the GOT
488 add r1, r1, r0 @ This fixes up C references
489 cmp r1, r2 @ if entry >= bss_start &&
490 cmphs r3, r1 @ bss_end > entry
491 addhi r1, r1, r5 @ entry += dtb size
492 str r1, [r11], #4 @ next entry
506 1: ldr r1, [r11, #0] @ relocate entries in the GOT
507 cmp r1, r2 @ entry < bss_start ||
508 cmphs r3, r1 @ _end < entry
509 addlo r1, r1, r0 @ table. This fixes up the
510 str r1, [r11], #4 @ C references.
540 mov r1, sp @ malloc space above stack
546 mov r1, r7 @ restore architecture number
572 LC0: .word LC0 @ r1
604 * r0, r1, r2, r3, r9, r10, r12 corrupted
691 mov r1, #0x12 @ XN|U + section mapping
692 orr r1, r1, #3 << 10 @ AP=11
694 1: cmp r1, r9 @ if virt > start of RAM
695 cmphs r10, r1 @ && end of RAM > virt
696 bic r1, r1, #0x1c @ clear XN|U + C + B
697 orrlo r1, r1, #0x10 @ Set XN|U for non-RAM
698 orrhs r1, r1, r6 @ set RAM section settings
699 str r1, [r0], #4 @ 1:1 mapping
700 add r1, r1, #1048576
709 orr r1, r6, #0x04 @ ensure B is set for this
710 orr r1, r1, #3 << 10
713 orr r1, r1, r2, lsl #20
715 str r1, [r0], #4
716 add r1, r1, #1048576
717 str r1, [r0]
777 movne r1, #0xfffffffd @ domain 0 = client
781 mcrne p15, 0, r1, c3, c0, 0 @ load domain access control
811 mov r1, #-1
813 mcr p15, 0, r1, c3, c0, 0 @ load domain access control
830 * r1 = corrupted
853 1: ldr r1, [r12, #0] @ get value
855 eor r1, r1, r9 @ (real ^ match)
856 tst r1, r2 @ & mask
1037 * r0, r1, r2, r3, r9, r12 corrupted
1097 * r1, r2, r3, r9, r10, r11, r12 corrupted
1112 mov r1, #7 << 5 @ 8 segments
1113 1: orr r3, r1, #63 << 26 @ 64 entries
1117 subs r1, r1, #1 << 5
1128 mov r1, #0
1129 mcr p15, 0, r1, c7, c14, 0 @ clean and invalidate D cache
1130 mcr p15, 0, r1, c7, c5, 0 @ flush I cache
1131 mcr p15, 0, r1, c7, c10, 4 @ drain WB
1135 mov r1, #0
1137 mcreq p15, 0, r1, c7, c14, 0 @ clean+invalidate D
1138 mcr p15, 0, r1, c7, c5, 0 @ invalidate I+BTB
1139 mcreq p15, 0, r1, c7, c15, 0 @ clean+invalidate unified
1140 mcr p15, 0, r1, c7, c10, 4 @ drain WB
1162 mov r1, r0, lsr r2 @ extract cache type bits from clidr
1163 and r1, r1, #7 @ mask of the bits for current cache only
1164 cmp r1, #2 @ see what cache we have at this level
1168 mrc p15, 1, r1, c0, c0, 0 @ read the new csidr
1169 and r2, r1, #7 @ extract the length of the cache lines
1172 ands r4, r4, r1, lsr #3 @ find maximum number on the way size
1175 ands r7, r7, r1, lsr #13 @ extract max number of the index size
1222 mov r1, r3, lsr #18
1223 and r1, r1, #7
1225 mov r2, r2, lsl r1 @ base dcache size *2
1233 mov r1, pc
1234 bic r1, r1, #63 @ align to longest cache line
1235 add r2, r1, r2
1237 ARM( ldr r3, [r1], r11 ) @ s/w flush D cache
1238 THUMB( ldr r3, [r1] ) @ s/w flush D cache
1239 THUMB( add r1, r1, r11 )
1240 teq r1, r2
1243 mcr p15, 0, r1, c7, c5, 0 @ flush I cache
1244 mcr p15, 0, r1, c7, c6, 0 @ flush D cache
1245 mcr p15, 0, r1, c7, c10, 4 @ drain WB
1252 mov r1, #0
1253 mcr p15, 0, r1, c7, c0, 0 @ invalidate whole cache v3
1266 @ phex corrupts {r0, r1, r2, r3}
1269 strb r2, [r3, r1]
1270 1: subs r1, r1, #1
1278 strb r2, [r3, r1]
1281 @ puts corrupts {r0, r1, r2, r3}
1282 puts: loadsp r3, r1
1287 mov r1, #0x00020000
1288 3: subs r1, r1, #1
1296 @ putc corrupts {r0, r1, r2, r3}
1300 loadsp r3, r1
1303 @ memdump corrupts {r0, r1, r2, r3, r10, r11, r12, lr}
1309 mov r1, #8
1316 mov r1, #8
/linux-4.1.27/arch/microblaze/kernel/
H A Dmcount.S18 addik r1, r1, -120; \
19 swi r2, r1, 4; \
20 swi r3, r1, 8; \
21 swi r4, r1, 12; \
22 swi r5, r1, 116; \
23 swi r6, r1, 16; \
24 swi r7, r1, 20; \
25 swi r8, r1, 24; \
26 swi r9, r1, 28; \
27 swi r10, r1, 32; \
28 swi r11, r1, 36; \
29 swi r12, r1, 40; \
30 swi r13, r1, 44; \
31 swi r14, r1, 48; \
32 swi r16, r1, 52; \
33 swi r17, r1, 56; \
34 swi r18, r1, 60; \
35 swi r19, r1, 64; \
36 swi r20, r1, 68; \
37 swi r21, r1, 72; \
38 swi r22, r1, 76; \
39 swi r23, r1, 80; \
40 swi r24, r1, 84; \
41 swi r25, r1, 88; \
42 swi r26, r1, 92; \
43 swi r27, r1, 96; \
44 swi r28, r1, 100; \
45 swi r29, r1, 104; \
46 swi r30, r1, 108; \
47 swi r31, r1, 112;
50 lwi r2, r1, 4; \
51 lwi r3, r1, 8; \
52 lwi r4, r1, 12; \
53 lwi r5, r1, 116; \
54 lwi r6, r1, 16; \
55 lwi r7, r1, 20; \
56 lwi r8, r1, 24; \
57 lwi r9, r1, 28; \
58 lwi r10, r1, 32; \
59 lwi r11, r1, 36; \
60 lwi r12, r1, 40; \
61 lwi r13, r1, 44; \
62 lwi r14, r1, 48; \
63 lwi r16, r1, 52; \
64 lwi r17, r1, 56; \
65 lwi r18, r1, 60; \
66 lwi r19, r1, 64; \
67 lwi r20, r1, 68; \
68 lwi r21, r1, 72; \
69 lwi r22, r1, 76; \
70 lwi r23, r1, 80; \
71 lwi r24, r1, 84; \
72 lwi r25, r1, 88; \
73 lwi r26, r1, 92; \
74 lwi r27, r1, 96; \
75 lwi r28, r1, 100; \
76 lwi r29, r1, 104; \
77 lwi r30, r1, 108; \
78 lwi r31, r1, 112; \
79 addik r1, r1, 120;
93 swi r15, r1, 0;
112 addik r5, r1, 120; /* MS: load parent addr */
135 lwi r6, r1, 120; /* MS: load parent addr */
141 lwi r15, r1, 0;
152 swi r15, r1, 0;
H A Dentry-nommu.S54 swi r1, r0, PER_CPU(ENTRY_SP) /* save the current sp */
60 addik r1, r1, (-PT_SIZE) /* room for pt_regs (delay slot) */
62 lwi r1, r0, PER_CPU(CURRENT_SAVE) /* get the saved current */
63 lwi r1, r1, TS_THREAD_INFO /* get the thread info */
65 addik r1, r1, THREAD_SIZE - PT_SIZE
67 swi r11, r1, PT_MODE /* store the mode */
69 swi r2, r1, PT_R2
70 swi r3, r1, PT_R3
71 swi r4, r1, PT_R4
72 swi r5, r1, PT_R5
73 swi r6, r1, PT_R6
74 swi r7, r1, PT_R7
75 swi r8, r1, PT_R8
76 swi r9, r1, PT_R9
77 swi r10, r1, PT_R10
78 swi r11, r1, PT_R11
79 swi r12, r1, PT_R12
80 swi r13, r1, PT_R13
81 swi r14, r1, PT_R14
82 swi r14, r1, PT_PC
83 swi r15, r1, PT_R15
84 swi r16, r1, PT_R16
85 swi r17, r1, PT_R17
86 swi r18, r1, PT_R18
87 swi r19, r1, PT_R19
88 swi r20, r1, PT_R20
89 swi r21, r1, PT_R21
90 swi r22, r1, PT_R22
91 swi r23, r1, PT_R23
92 swi r24, r1, PT_R24
93 swi r25, r1, PT_R25
94 swi r26, r1, PT_R26
95 swi r27, r1, PT_R27
96 swi r28, r1, PT_R28
97 swi r29, r1, PT_R29
98 swi r30, r1, PT_R30
99 swi r31, r1, PT_R31
102 swi r11, r1, PT_MSR
104 swi r11, r1, PT_EAR
106 swi r11, r1, PT_ESR
108 swi r11, r1, PT_FSR
111 swi r11, r1, PT_R1
121 add r5, r0, r1
124 lwi r11, r1, PT_MODE
139 addk r5, r1, r0
149 lwi r11, r1, PT_MODE
156 lwi r11, r1, PT_FSR
158 lwi r11, r1, PT_ESR
160 lwi r11, r1, PT_EAR
162 lwi r11, r1, PT_MSR
165 lwi r31, r1, PT_R31
166 lwi r30, r1, PT_R30
167 lwi r29, r1, PT_R29
168 lwi r28, r1, PT_R28
169 lwi r27, r1, PT_R27
170 lwi r26, r1, PT_R26
171 lwi r25, r1, PT_R25
172 lwi r24, r1, PT_R24
173 lwi r23, r1, PT_R23
174 lwi r22, r1, PT_R22
175 lwi r21, r1, PT_R21
176 lwi r20, r1, PT_R20
177 lwi r19, r1, PT_R19
178 lwi r18, r1, PT_R18
179 lwi r17, r1, PT_R17
180 lwi r16, r1, PT_R16
181 lwi r15, r1, PT_R15
182 lwi r14, r1, PT_PC
183 lwi r13, r1, PT_R13
184 lwi r12, r1, PT_R12
185 lwi r11, r1, PT_R11
186 lwi r10, r1, PT_R10
187 lwi r9, r1, PT_R9
188 lwi r8, r1, PT_R8
189 lwi r7, r1, PT_R7
190 lwi r6, r1, PT_R6
191 lwi r5, r1, PT_R5
192 lwi r4, r1, PT_R4
193 lwi r3, r1, PT_R3
194 lwi r2, r1, PT_R2
195 lwi r1, r1, PT_R1
203 swi r1, r0, PER_CPU(ENTRY_SP) /* save the current sp */
209 addik r1, r1, (-PT_SIZE) /* Room for pt_regs (delay slot) */
211 lwi r1, r0, PER_CPU(CURRENT_SAVE) /* get the saved current */
212 lwi r1, r1, TS_THREAD_INFO /* get the thread info */
214 addik r1, r1, THREAD_SIZE - PT_SIZE
216 swi r11, r1, PT_MODE /* store the mode */
219 swi r2, r1, PT_R2
220 swi r3, r1, PT_R3 /* r3: _always_ in clobber list; see unistd.h */
221 swi r4, r1, PT_R4 /* r4: _always_ in clobber list; see unistd.h */
222 swi r5, r1, PT_R5
223 swi r6, r1, PT_R6
224 swi r7, r1, PT_R7
225 swi r8, r1, PT_R8
226 swi r9, r1, PT_R9
227 swi r10, r1, PT_R10
228 swi r11, r1, PT_R11
230 swi r12, r1, PT_R12
231 swi r13, r1, PT_R13
233 swi r14, r1, PT_R14
236 swi r14, r1, PT_PC /* increment by 4 and store in pc */
237 swi r15, r1, PT_R15
238 swi r16, r1, PT_R16
239 swi r17, r1, PT_R17
240 swi r18, r1, PT_R18
241 swi r19, r1, PT_R19
242 swi r20, r1, PT_R20
243 swi r21, r1, PT_R21
244 swi r22, r1, PT_R22
245 swi r23, r1, PT_R23
246 swi r24, r1, PT_R24
247 swi r25, r1, PT_R25
248 swi r26, r1, PT_R26
249 swi r27, r1, PT_R27
250 swi r28, r1, PT_R28
251 swi r29, r1, PT_R29
252 swi r30, r1, PT_R30
253 swi r31, r1, PT_R31
262 swi r11, r1, PT_MSR
264 swi r11, r1, PT_EAR
266 swi r11, r1, PT_ESR
268 swi r11, r1, PT_FSR
271 swi r11, r1, PT_R1
302 swi r1, r0, PER_CPU(ENTRY_SP) /* save the current sp */
303 lwi r1, r0, PER_CPU(CURRENT_SAVE) /* get the saved current */
304 lwi r1, r1, TS_THREAD_INFO /* get the thread info */
305 addik r1, r1, THREAD_SIZE - PT_SIZE /* get the kernel stack */
309 swi r11, r1, PT_MODE /* store the mode */
312 swi r2, r1, PT_R2
313 swi r3, r1, PT_R3 /* r3: _always_ in clobber list; see unistd.h */
314 swi r4, r1, PT_R4 /* r4: _always_ in clobber list; see unistd.h */
315 swi r5, r1, PT_R5
316 swi r6, r1, PT_R6
317 swi r7, r1, PT_R7
318 swi r8, r1, PT_R8
319 swi r9, r1, PT_R9
320 swi r10, r1, PT_R10
321 swi r11, r1, PT_R11
323 swi r12, r1, PT_R12
324 swi r13, r1, PT_R13
326 swi r14, r1, PT_R14
327 swi r14, r1, PT_PC /* Will return to interrupted instruction */
328 swi r15, r1, PT_R15
329 swi r16, r1, PT_R16
330 swi r17, r1, PT_R17
331 swi r18, r1, PT_R18
332 swi r19, r1, PT_R19
333 swi r20, r1, PT_R20
334 swi r21, r1, PT_R21
335 swi r22, r1, PT_R22
336 swi r23, r1, PT_R23
337 swi r24, r1, PT_R24
338 swi r25, r1, PT_R25
339 swi r26, r1, PT_R26
340 swi r27, r1, PT_R27
341 swi r28, r1, PT_R28
342 swi r29, r1, PT_R29
343 swi r30, r1, PT_R30
344 swi r31, r1, PT_R31
353 swi r11, r1, PT_MSR
355 swi r11, r1, PT_EAR
357 swi r11, r1, PT_ESR
359 swi r11, r1, PT_FSR
362 swi r11, r1, PT_R1
378 lwi r3, r1, PT_R3
379 lwi r4, r1, PT_R4
394 swi r1, r11, CC_R1
466 lwi r1, r11, CC_R1
475 swi r31, r1, PT_R31 /* save r31 in user context. */
490 lwi r11, r1, PT_MODE
514 swi r4, r1, PT_R4 /* return val */
515 swi r3, r1, PT_R3 /* return val */
527 lwi r18, r1, PT_MODE
531 lwi r18, r1, PT_FSR
533 lwi r18, r1, PT_ESR
535 lwi r18, r1, PT_EAR
537 lwi r18, r1, PT_MSR
540 lwi r31, r1, PT_R31
541 lwi r30, r1, PT_R30
542 lwi r29, r1, PT_R29
543 lwi r28, r1, PT_R28
544 lwi r27, r1, PT_R27
545 lwi r26, r1, PT_R26
546 lwi r25, r1, PT_R25
547 lwi r24, r1, PT_R24
548 lwi r23, r1, PT_R23
549 lwi r22, r1, PT_R22
550 lwi r21, r1, PT_R21
551 lwi r20, r1, PT_R20
552 lwi r19, r1, PT_R19
553 lwi r18, r1, PT_R18
554 lwi r17, r1, PT_R17
555 lwi r16, r1, PT_R16
556 lwi r15, r1, PT_R15
557 lwi r14, r1, PT_PC
558 lwi r13, r1, PT_R13
559 lwi r12, r1, PT_R12
560 lwi r11, r1, PT_R11
561 lwi r10, r1, PT_R10
562 lwi r9, r1, PT_R9
563 lwi r8, r1, PT_R8
564 lwi r7, r1, PT_R7
565 lwi r6, r1, PT_R6
566 lwi r5, r1, PT_R5
567 lwi r4, r1, PT_R4 /* return val */
568 lwi r3, r1, PT_R3 /* return val */
569 lwi r2, r1, PT_R2
570 lwi r1, r1, PT_R1
578 addk r5, r1, r0
H A Dentry.S178 swi r2, r1, PT_R2; /* Save SDA */ \
179 swi r3, r1, PT_R3; \
180 swi r4, r1, PT_R4; \
181 swi r5, r1, PT_R5; \
182 swi r6, r1, PT_R6; \
183 swi r7, r1, PT_R7; \
184 swi r8, r1, PT_R8; \
185 swi r9, r1, PT_R9; \
186 swi r10, r1, PT_R10; \
187 swi r11, r1, PT_R11; /* save clobbered regs after rval */\
188 swi r12, r1, PT_R12; \
189 swi r13, r1, PT_R13; /* Save SDA2 */ \
190 swi r14, r1, PT_PC; /* PC, before IRQ/trap */ \
191 swi r15, r1, PT_R15; /* Save LP */ \
192 swi r16, r1, PT_R16; \
193 swi r17, r1, PT_R17; \
194 swi r18, r1, PT_R18; /* Save asm scratch reg */ \
195 swi r19, r1, PT_R19; \
196 swi r20, r1, PT_R20; \
197 swi r21, r1, PT_R21; \
198 swi r22, r1, PT_R22; \
199 swi r23, r1, PT_R23; \
200 swi r24, r1, PT_R24; \
201 swi r25, r1, PT_R25; \
202 swi r26, r1, PT_R26; \
203 swi r27, r1, PT_R27; \
204 swi r28, r1, PT_R28; \
205 swi r29, r1, PT_R29; \
206 swi r30, r1, PT_R30; \
207 swi r31, r1, PT_R31; /* Save current task reg */ \
209 swi r11, r1, PT_MSR;
212 lwi r11, r1, PT_MSR; \
214 lwi r2, r1, PT_R2; /* restore SDA */ \
215 lwi r3, r1, PT_R3; \
216 lwi r4, r1, PT_R4; \
217 lwi r5, r1, PT_R5; \
218 lwi r6, r1, PT_R6; \
219 lwi r7, r1, PT_R7; \
220 lwi r8, r1, PT_R8; \
221 lwi r9, r1, PT_R9; \
222 lwi r10, r1, PT_R10; \
223 lwi r11, r1, PT_R11; /* restore clobbered regs after rval */\
224 lwi r12, r1, PT_R12; \
225 lwi r13, r1, PT_R13; /* restore SDA2 */ \
226 lwi r14, r1, PT_PC; /* RESTORE_LINK PC, before IRQ/trap */\
227 lwi r15, r1, PT_R15; /* restore LP */ \
228 lwi r16, r1, PT_R16; \
229 lwi r17, r1, PT_R17; \
230 lwi r18, r1, PT_R18; /* restore asm scratch reg */ \
231 lwi r19, r1, PT_R19; \
232 lwi r20, r1, PT_R20; \
233 lwi r21, r1, PT_R21; \
234 lwi r22, r1, PT_R22; \
235 lwi r23, r1, PT_R23; \
236 lwi r24, r1, PT_R24; \
237 lwi r25, r1, PT_R25; \
238 lwi r26, r1, PT_R26; \
239 lwi r27, r1, PT_R27; \
240 lwi r28, r1, PT_R28; \
241 lwi r29, r1, PT_R29; \
242 lwi r30, r1, PT_R30; \
243 lwi r31, r1, PT_R31; /* Restore cur task reg */
246 swi r1, r0, TOPHYS(PER_CPU(ENTRY_SP)); /* save stack */ \
248 mfs r1, rmsr; \
249 andi r1, r1, MSR_UMS; \
250 bnei r1, 1f; \
253 lwi r1, r0, TOPHYS(PER_CPU(ENTRY_SP)); \
255 /* tophys(r1,r1); */ \
256 /* addik r1, r1, -PT_SIZE; */ \
257 addik r1, r1, CONFIG_KERNEL_BASE_ADDR - CONFIG_KERNEL_START - PT_SIZE; \
260 swi r1, r1, PT_MODE; \
262 lwi r1, r0, TOPHYS(PER_CPU(CURRENT_SAVE)); /* get saved current */\
263 tophys(r1,r1); \
264 lwi r1, r1, TS_THREAD_INFO; /* get the thread info */ \
266 /* addik r1, r1, THREAD_SIZE; */ \
267 /* tophys(r1,r1); */ \
268 /* addik r1, r1, -PT_SIZE; */ \
269 addik r1, r1, THREAD_SIZE + CONFIG_KERNEL_BASE_ADDR - CONFIG_KERNEL_START - PT_SIZE; \
272 swi r11, r1, PT_R1; /* Store user SP. */ \
273 swi r0, r1, PT_MODE; /* Was in user-mode. */ \
293 swi r1, r0, TOPHYS(PER_CPU(ENTRY_SP)) /* save stack */
296 lwi r1, r0, TOPHYS(PER_CPU(CURRENT_SAVE)); /* get saved current */
297 tophys(r1,r1);
298 lwi r1, r1, TS_THREAD_INFO; /* get stack from task_struct */
300 addik r1, r1, THREAD_SIZE;
301 tophys(r1,r1);
303 addik r1, r1, -PT_SIZE; /* Make room on the stack. */
305 swi r0, r1, PT_R3
306 swi r0, r1, PT_R4
308 swi r0, r1, PT_MODE; /* Was in user-mode. */
310 swi r11, r1, PT_R1; /* Store user SP. */
314 swi r12, r1, PT_R0;
315 tovirt(r1,r1)
333 swi r3, r1, PT_R3
335 addik r5, r1, PT_R0
339 lwi r5, r1, PT_R5;
340 lwi r6, r1, PT_R6;
341 lwi r7, r1, PT_R7;
342 lwi r8, r1, PT_R8;
343 lwi r9, r1, PT_R9;
344 lwi r10, r1, PT_R10;
385 swi r3, r1, PT_R3
386 swi r4, r1, PT_R4
388 lwi r11, r1, PT_MODE;
400 addik r5, r1, PT_R0
419 addik r5, r1, 0; /* Arg 1: struct pt_regs *regs */
429 tophys(r1,r1);
431 addik r1, r1, PT_SIZE /* Clean up stack space. */
432 lwi r1, r1, PT_R1 - PT_SIZE;/* Restore user stack pointer. */
438 tophys(r1,r1);
440 addik r1, r1, PT_SIZE /* Clean up stack space. */
441 tovirt(r1,r1);
472 addik r5, r1, 0; /* add user context as 1st arg */
483 swi r17, r1, PT_PC;
484 tovirt(r1,r1)
493 addik r5, r1, 0 /* parameter struct pt_regs * regs */
519 swi r17, r1, PT_PC;
520 tovirt(r1,r1)
526 addik r7, r1, 0 /* parameter struct pt_regs * regs */
549 swi r17, r1, PT_PC;
550 tovirt(r1,r1)
556 addik r5, r1, 0 /* parameter struct pt_regs * regs */
561 swi r17, r1, PT_PC;
562 tovirt(r1,r1)
568 addik r5, r1, 0 /* parameter struct pt_regs * regs */
572 lwi r11, r1, PT_MODE;
604 addik r5, r1, 0; /* Arg 1: struct pt_regs *regs */
613 tophys(r1,r1);
616 addik r1, r1, PT_SIZE /* Clean up stack space. */
618 lwi r1, r1, PT_R1 - PT_SIZE; /* Restore user stack pointer. */
623 tophys(r1,r1);
625 addik r1, r1, PT_SIZE /* Clean up stack space. */
627 tovirt(r1,r1);
640 * The stack-pointer (r1) should have already been saved to the memory
646 swi r1, r0, TOPHYS(PER_CPU(ENTRY_SP))
648 mfs r1, rmsr
650 andi r1, r1, MSR_UMS
651 bnei r1, 1f
654 lwi r1, r0, TOPHYS(PER_CPU(ENTRY_SP))
655 tophys(r1,r1); /* MS: I have in r1 physical address where stack is */
658 addik r1, r1, -PT_SIZE;
661 swi r1, r1, PT_MODE; /* 0 - user mode, 1 - kernel mode */
665 lwi r1, r0, TOPHYS(PER_CPU(CURRENT_SAVE));
666 tophys(r1,r1);
667 lwi r1, r1, TS_THREAD_INFO;
668 addik r1, r1, THREAD_SIZE;
669 tophys(r1,r1);
671 addik r1, r1, -PT_SIZE;
674 swi r0, r1, PT_MODE;
676 swi r11, r1, PT_R1;
680 tovirt(r1,r1)
683 addik r5, r1, 0;
687 lwi r11, r1, PT_MODE;
703 addik r5, r1, 0; /* Arg 1: struct pt_regs *regs */
714 tophys(r1,r1);
716 addik r1, r1, PT_SIZE /* MS: Clean up stack space. */
717 lwi r1, r1, PT_R1 - PT_SIZE;
742 tophys(r1,r1)
744 addik r1, r1, PT_SIZE /* MS: Clean up stack space. */
745 tovirt(r1,r1);
757 swi r1, r0, TOPHYS(PER_CPU(ENTRY_SP))
759 mfs r1, rmsr
761 andi r1, r1, MSR_UMS
762 bnei r1, 1f
764 lwi r1, r0, TOPHYS(PER_CPU(ENTRY_SP)); /* Reload kernel stack-ptr*/
767 addik r1, r1, CONFIG_KERNEL_BASE_ADDR - CONFIG_KERNEL_START - PT_SIZE;
770 swi r0, r1, PT_R0; /* R0 must be saved too */
771 swi r14, r1, PT_R14 /* rewrite saved R14 value */
772 swi r16, r1, PT_PC; /* PC and r16 are the same */
775 swi r11, r1, PT_EAR;
777 swi r11, r1, PT_ESR;
779 swi r11, r1, PT_FSR;
783 addik r11, r1, CONFIG_KERNEL_START - CONFIG_KERNEL_BASE_ADDR + PT_SIZE;
784 swi r11, r1, PT_R1
786 tovirt(r1,r1)
788 addi r5, r1, 0 /* pass pt_reg address as the first arg */
798 1: lwi r1, r0, TOPHYS(PER_CPU(CURRENT_SAVE)); /* get saved current */
799 tophys(r1,r1);
800 lwi r1, r1, TS_THREAD_INFO; /* get the thread info */
801 addik r1, r1, THREAD_SIZE; /* calculate kernel stack pointer */
802 tophys(r1,r1);
804 addik r1, r1, -PT_SIZE; /* Make room on the stack. */
806 swi r16, r1, PT_PC; /* Save LP */
807 swi r0, r1, PT_MODE; /* Was in user-mode. */
809 swi r11, r1, PT_R1; /* Store user SP. */
811 tovirt(r1,r1)
813 addik r5, r1, 0;
821 lwi r11, r1, PT_MODE;
840 addik r5, r1, 0; /* Arg 1: struct pt_regs *regs */
848 tophys(r1,r1);
851 addik r1, r1, PT_SIZE /* Clean up stack space */
852 lwi r1, r1, PT_R1 - PT_SIZE; /* Restore user stack pointer */
859 tophys(r1,r1);
862 lwi r14, r1, PT_R14;
863 lwi r16, r1, PT_PC;
864 addik r1, r1, PT_SIZE; /* MS: Clean up stack space */
865 tovirt(r1,r1);
879 swi r1, r11, CC_R1
944 lwi r1, r11, CC_R1
/linux-4.1.27/arch/powerpc/include/asm/
H A Dftrace.h13 stwu r1,-48(r1); \
14 stw r3, 12(r1); \
15 stw r4, 16(r1); \
16 stw r5, 20(r1); \
17 stw r6, 24(r1); \
19 lwz r4, 52(r1); \
21 stw r7, 28(r1); \
22 stw r8, 32(r1); \
23 stw r9, 36(r1); \
24 stw r10,40(r1); \
25 stw r3, 44(r1); \
26 stw r5, 8(r1)
29 lwz r6, 8(r1); \
30 lwz r0, 44(r1); \
31 lwz r3, 12(r1); \
33 lwz r4, 16(r1); \
35 lwz r5, 20(r1); \
36 lwz r6, 24(r1); \
37 lwz r0, 52(r1); \
38 lwz r7, 28(r1); \
39 lwz r8, 32(r1); \
41 lwz r9, 36(r1); \
42 lwz r10,40(r1); \
43 addi r1, r1, 48
H A Dirqflags.h23 stdu r1, -STACK_FRAME_OVERHEAD(r1); \
24 std r0, 16(r1); \
25 stdu r1, -STACK_FRAME_OVERHEAD(r1); \
27 ld r1, 0(r1); \
28 ld r1, 0(r1);
/linux-4.1.27/arch/m32r/lib/
H A Dmemset.S10 * val: r1
35 stb r1, @r4 || addi r4, #1
41 and3 r1, r1, #0x00ff /* r1: abababab <-- ??????ab */
42 sll3 r3, r1, #8
43 or r1, r3 || addi r4, #-4
44 sll3 r3, r1, #16
45 or r1, r3 || addi r2, #-4
47 st r1, @+r4 || addi r2, #-4
50 st r1, @+r4
57 and3 r1, r1, #0x00ff /* r1: abababab <-- ??????ab */
58 sll3 r3, r1, #8
59 or r1, r3 || addi r4, #-4
60 sll3 r3, r1, #16
61 or r1, r3 || ldi r5, #16
64 st r1, @+r4 || addi r2, #-16
65 st r1, @+r4 || cmpu r2, r5
66 st r1, @+r4
67 st r1, @+r4
84 addi r2, #-1 || stb r1, @r4+
88 addi r2, #-1 || stb r1, @r4
113 stb r1, @r4
121 and3 r1, r1, #0x00ff /* r1: abababab <-- ??????ab */
122 sll3 r3, r1, #8
123 or r1, r3
124 sll3 r3, r1, #16
125 or r1, r3
129 st r1, @+r4
133 st r1, @+r4
140 and3 r1, r1, #0x00ff /* r1: abababab <-- ??????ab */
141 sll3 r3, r1, #8
142 or r1, r3
143 sll3 r3, r1, #16
144 or r1, r3
149 st r1, @+r4
150 st r1, @+r4
152 st r1, @+r4
153 st r1, @+r4
170 stb r1, @r4
H A Dashxdi3.S9 ; input (r0,r1) src
12 ; output (r0,r1)
27 mv r1, r0 || srai r0, #31
29 sra r1, r2
33 mv r3, r0 || srl r1, r2
36 or r1, r3 || jmp r14
47 mv r0, r1 || addi r2, #-32
48 sll r0, r2 || ldi r1, #0
52 mv r3, r1 || sll r0, r2
53 sll r1, r2 || neg r2, r2
64 mv r1, r0 || addi r2, #-32
65 ldi r0, #0 || srl r1, r2
69 mv r3, r0 || srl r1, r2
72 or r1, r3 || jmp r14
84 mv r0, r1 || srai r1, #31
90 mv r3, r1 || srl r0, r2
91 sra r1, r2 || neg r2, r2
104 mv r1, r0 || addi r2, #-32
105 sll r1, r2 || ldi r0, #0
109 mv r3, r0 || sll r1, r2
112 or r1, r3 || jmp r14
121 mv r0, r1 || addi r2, #-32
122 ldi r1, #0 || srl r0, r2
126 mv r3, r1 || srl r0, r2
127 srl r1, r2 || neg r2, r2
145 mv r1, r0
148 sra r1, r2
153 srl r1, r2
157 or r1, r3
171 mv r0, r1
174 ldi r1, #0
178 mv r3, r1
180 sll r1, r2
195 mv r1, r0
198 srl r1, r2
203 srl r1, r2
207 or r1, r3
222 mv r0, r1
223 srai r1, #31
229 mv r3, r1
231 sra r1, r2
248 mv r1, r0
250 sll r1, r2
256 sll r1, r2
260 or r1, r3
272 mv r0, r1
273 ldi r1, #0
279 mv r3, r1
281 srl r1, r2
H A Dmemcpy.S10 * src: r1
24 or r7, r1 || cmpz r2
25 jc r14 || cmpeq r0, r1 ; return if r2=0
26 jc r14 ; return if r0=r1
35 ld r7, @r1+ || addi r3, #-1
41 ldb r7, @r1 || addi r1, #1
46 ldb r7, @r1 || addi r1, #1
63 or r7, r1
64 beq r0, r1, end_memcopy
74 ld r7, @r1+
81 ldb r7, @r1
82 addi r1, #1
H A Dstrlen.S28 ld r1, @r6+ || not r4, r0
32 ld r0, @r6+ || not r4, r1
33 sub r1, r5 || and r4, r7
34 and r4, r1 || addi r2, #4
43 ldb r1, @r6 || addi r6, #1
44 beqz r1, strlen_exit
51 ldb r1, @r6 || addi r6, #1
52 addi r0, #-1 || cmpz r1
75 ld r1, @r6+
84 not r4, r1 ; NOTE: If a null char. exists, return 0.
85 sub r1, r5 ; if ((x - 0x01010101) & ~x & 0x80808080)
87 and r4, r1
93 ldb r1, @r6
95 beqz r1, strlen_exit
104 ldb r1, @r6
107 beqz r1, strlen_exit
/linux-4.1.27/arch/nios2/boot/compressed/
H A Dhead.S25 movia r1, NIOS2_ICACHE_SIZE
27 1: initi r1
28 sub r1, r1, r2
29 bgt r1, r0, 1b
31 movia r1, NIOS2_DCACHE_SIZE
33 1: initd 0(r1)
34 sub r1, r1, r2
35 bgt r1, r0, 1b
37 nextpc r1 /* Find out where we are */
40 beq r1, r2, finish_move /* We are running in correct address,
42 /* move code, r1: src, r2: dest, r3: last dest */
43 addi r1, r1, (_start - chkadr) /* Source */
46 1: ldw r8, 0(r1) /* load a word from [r1] */
48 addi r1, r1, 4 /* inc the src addr */
52 movia r1, NIOS2_DCACHE_SIZE
54 1: flushd 0(r1)
55 sub r1, r1, r2
56 bgt r1, r0, 1b
57 movia r1, finish_move
58 jmp r1 /* jmp to linked address */
63 movia r1, _end /* the .bss and _end. */
66 bne r1, r2, 1b
72 add sp, sp, r1
88 movia r1, NIOS2_DCACHE_SIZE
90 1: flushd 0(r1)
91 sub r1, r1, r2
92 bgt r1, r0, 1b
94 movia r1, NIOS2_ICACHE_SIZE
96 1: flushi r1
97 sub r1, r1, r2
98 bgt r1, r0, 1b
101 movia r1, (CONFIG_NIOS2_MEM_BASE | CONFIG_NIOS2_KERNEL_REGION_BASE)
102 jmp r1
/linux-4.1.27/arch/sh/kernel/cpu/sh2a/
H A Dex.S23 mov.l r1,@-sp
25 mov #no,r1
31 extu.b r1,r1
33 extu.w r1,r1
42 mov.l r1,@-sp
44 mov #no,r1
49 extu.b r1,r1
51 add r0,r1
54 extu.w r1,r1
H A Dentry.S41 ! r1
45 ! r1 = vector (pseudo EXPEVT / INTEVT / TRA)
61 mov.l r1,@-r15 ! TRA
75 mov r1,r9 ! r9 = interrupt vector
79 mov.l @r8+,r1
88 mov r1,r9 ! r9 = interrupt vector
92 mov.l @r8+,r1 ! old R1
163 mov.l @(OFF_SP,r0),r1
165 mov.l r3,@-r1
167 mov.l r3,@-r1
175 mov.l r1,@(8,r0)
213 mov.l @(OFF_SR,r0),r1
214 shll2 r1
215 shlr2 r1 ! clear MD bit
219 mov.l r1,@(4,r2) ! set sr
220 mov.l @(OFF_PC,r0),r1
221 mov.l r1,@r2 ! set pc
222 get_current_thread_info r0, r1
223 mov.l $current_thread_info,r1
224 mov.l r0,@r1
/linux-4.1.27/arch/arm/kernel/
H A Ddebug.S47 mov r1, #8
52 mov r1, #4
57 mov r1, #2
59 add r3, r2, r1
60 mov r1, #0
61 strb r1, [r3]
62 1: and r1, r0, #15
64 cmp r1, #10
65 addlt r1, r1, #'0'
66 addge r1, r1, #'a' - 10
67 strb r1, [r3, #-1]!
81 addruart_current r3, r1, r2
84 senduart r1, r3
86 teq r1, #'\n'
87 moveq r1, #'\r'
90 ldrneb r1, [r0], #1
91 teqne r1, #0
97 addruart_current r3, r1, r2
98 mov r1, r0
107 str r3, [r1]
115 mov r1, r0
123 adr r1, hexbuf
124 strb r0, [r1]
134 str r2, [r1]
H A Diwmmxt.S91 ldr r1, [r3] @ get current Concan owner
108 teq r1, #0 @ test for last ownership
119 wstrw wCSSF, [r1, #MMX_WCSSF]
120 wstrw wCASF, [r1, #MMX_WCASF]
121 wstrw wCGR0, [r1, #MMX_WCGR0]
122 wstrw wCGR1, [r1, #MMX_WCGR1]
123 wstrw wCGR2, [r1, #MMX_WCGR2]
124 wstrw wCGR3, [r1, #MMX_WCGR3]
130 wstrd wR0, [r1, #MMX_WR0]
131 wstrd wR1, [r1, #MMX_WR1]
132 wstrd wR2, [r1, #MMX_WR2]
133 wstrd wR3, [r1, #MMX_WR3]
134 wstrd wR4, [r1, #MMX_WR4]
135 wstrd wR5, [r1, #MMX_WR5]
136 wstrd wR6, [r1, #MMX_WR6]
137 wstrd wR7, [r1, #MMX_WR7]
138 wstrd wR8, [r1, #MMX_WR8]
139 wstrd wR9, [r1, #MMX_WR9]
140 wstrd wR10, [r1, #MMX_WR10]
141 wstrd wR11, [r1, #MMX_WR11]
142 wstrd wR12, [r1, #MMX_WR12]
143 wstrd wR13, [r1, #MMX_WR13]
144 wstrd wR14, [r1, #MMX_WR14]
145 wstrd wR15, [r1, #MMX_WR15]
178 @ clear CUP/MUP (only if r1 != 0)
179 teq r1, #0
205 ldr r1, [r3] @ get current Concan owner
206 teq r1, #0 @ any current owner?
209 teqne r1, r2 @ or specified one?
244 * r1 = memory address where to store Concan state
263 mov r0, r1
264 mov r1, r2
282 * r1 = memory address where to get Concan state from
307 mov r0, r1
308 mov r1, #0 @ don't clear CUP/MUP
325 XSC(mrc p15, 0, r1, c15, c1, 0)
326 PJ4(mrc p15, 0, r1, c1, c0, 2)
328 XSC(tst r1, #0x3)
329 PJ4(tst r1, #0xf)
339 XSC(eor r1, r1, #0x3)
340 XSC(mcr p15, 0, r1, c15, c1, 0)
341 PJ4(eor r1, r1, #0xf)
342 PJ4(mcr p15, 0, r1, c1, c0, 2)
344 mrc p15, 0, r1, c2, c0, 0
345 sub pc, lr, r1, lsr #32 @ cpwait and return
361 ldr r1, [r3] @ get current Concan owner
362 eors r0, r0, r1 @ if equal...
H A Dentry-v7m.S27 mrs r1, ipsr
46 ldr r1, =V7M_xPSR_EXCEPTIONNO
47 and r0, r1
49 mov r1, sp
51 @ routine called with r0 = irq number, r1 = struct pt_regs *
58 ldr r1, =BASEADDR_V7M_SCB
59 ldr r0, [r1, V7M_SCB_ICSR]
68 str r0, [r1, V7M_SCB_ICSR] @ raise PendSV
84 ldr r1, =BASEADDR_V7M_SCB
86 str r0, [r1, V7M_SCB_ICSR] @ clear PendSV
96 * r0 = previous task_struct, r1 = previous thread_info, r2 = next thread_info
102 add ip, r1, #TI_CPU_SAVE
109 mov r1, #THREAD_NOTIFY_SWITCH
/linux-4.1.27/arch/arc/lib/
H A Dmemcmp.S20 or r12,r0,r1
25 ld r5,[r1,0]
29 ld_s r12,[r1,4]
32 ld.a r5,[r1,8]
39 ld r5,[r1,4]
46 sub_s r1,r0,1
47 bic_s r1,r1,r0
48 norm r1,r1
50 and r1,r1,24
53 sub_s r1,r0,1
54 bic_s r1,r1,r0
55 norm r1,r1
57 and r1,r1,24
59 asl r2,r4,r1
60 asl r12,r5,r1
68 sub_s r1,r0,1
69 bic_s r1,r1,r0
70 norm r1,r1
72 and r1,r1,24
73 asl_s r2,r2,r1
74 asl_s r12,r12,r1
101 ldb r5,[r1,0]
105 ldb r12,[r1,1]
108 ldb.a r5,[r1,2]
114 ldb_s r12,[r1,1]
H A Dmemset.S17 extb_s r1,r1
18 asl r3,r1,8
20 or_s r1,r1,r3
23 stb r1,[r3,-1]
25 stw r1,[r3,-2]
29 stb.ab r1,[r4,1]
31 stw.ab r1,[r4,2]
34 asl r3,r1,16
36 or_s r1,r1,r3
38 st.ab r1,[r4,4]
46 stb.ab r1,[r4,1]
51 ; memzero: @r0 = mem, @r1 = size_t
52 ; memset: @r0 = mem, @r1 = char, @r2 = size_t
56 mov r2, r1
57 mov r1, 0
H A Dstrcpy-700.S22 or r2,r0,r1
26 ld_s r3,[r1,0]
28 bbit0.d r1,2,loop_start
37 ld.a r3,[r1,4]
40 ld.a r4,[r1,4]
52 r3z: bmsk.f r1,r3,7
55 r3z: lsr.f r1,r3,24
59 stb.ab r1,[r10,1]
64 ldb.ab r3,[r1,1]
/linux-4.1.27/arch/unicore32/kernel/
H A Ddebug.S30 mov r1, #8
35 mov r1, #4
40 mov r1, #2
42 add r3, r2, r1
43 mov r1, #0
44 stb r1, [r3]
45 1: and r1, r0, #15
47 csub.a r1, #10
49 add r1, r1, #'0' - 'a' + 10
50 2: add r1, r1, #'a' - 10
51 stb.w r1, [r3+], #-1
64 senduart r1, r3
66 cxor.a r1, #'\n'
67 cmoveq r1, #'\r'
71 ldb.w r1, [r0]+, #1
72 cxor.a r1, #0
79 mov r1, r0
H A Ddebug-macro.S55 @ We assume r1 and r2 can be clobbered.
58 mov r1, #0x80
59 str r1, [\rx, #UART_LCR_OFFSET]
60 and r1, r2, #0xff00
61 mov r1, r1, lsr #8
62 str r1, [\rx, #UART_DLH_OFFSET]
63 and r1, r2, #0xff
64 str r1, [\rx, #UART_DLL_OFFSET]
65 mov r1, #0x7
66 str r1, [\rx, #UART_FCR_OFFSET]
67 mov r1, #0x3
68 str r1, [\rx, #UART_LCR_OFFSET]
69 mov r1, #0x0
70 str r1, [\rx, #UART_IER_OFFSET]
H A Dsleep.S40 ldw r1, =sleep_save_sp
41 stw r0, [r1]
69 mov r1, #0
70 movc p0.c5, r1, #14
82 ldw r1, =(PKUNITY_PM_BASE)
86 stw r6, [r1+], #0x18
89 stw r6, [r1+], #0x1c
93 stw r8, [r1+], #0xc
97 stw r5, [r1+], #0x10
123 stw r6, [r1]
130 1: ldw r6, [r1+], #0x44
139 2: stw r7, [r1+], #0x4
141 stw r8, [r1]
170 mov r1, #0
171 movc p0.c6, r1, #6 @ invalidate I & D TLBs
172 movc p0.c5, r1, #28 @ invalidate I & D caches, BTB
H A Dentry.S71 ldw r1, [sp+], #\offset + S_PSR @ get calling asr
73 mov.a bsr, r1 @ save in bsr_priv
76 ldm.w (r1 - r15), [sp]+ @ get calling r1 - r15
147 mov r1, sp
149 @ routine called with r0 = irq number, r1 = struct pt_regs *
161 stm (r1 - r15), [sp]+
165 ldm (r1 - r3), [r0]+
169 stw.w r1, [sp+], #-4 @ save the "real" r0 copied
172 mov r1, lr
178 @ r1 - lr_priv
192 stm (r1 - r15), [sp+]
196 ldm (r1 - r3), [r0]+
200 stw r1, [sp] @ save the "real" r0 copied
234 stm (r1 - r15), [sp+]
235 add r1, sp, #S_R16
236 stm (r16 - r28, sp, lr), [r1]+
248 mov r1, asr
272 @ the fault status register in r1.
274 movc r1, p0.c3, #0 @ get FSR
318 mov r1, asr
339 @ r1 - pointer to registers on stack
342 mov r1, #5
415 cff r1, s31 @ get fpu FPSCR
416 andn r2, r1, #0x08000000
423 @ r1 holds the FPSCR value
435 @ the fault status register in r1.
437 movc r1, p0.c3, #0 @ get FSR
468 mov r1, asr
477 mov r1, #5
478 enable_irq r1 @ Enable interrupts
494 * r0 = previous task_struct, r1 = previous thread_info, r2 = next thread_info
498 add ip, r1, #TI_CPU_SAVE
503 add ip, r1, #TI_FPSTATE
531 disable_irq r1 @ disable interrupts
532 ldw r1, [tsk+], #TI_FLAGS
533 cand.a r1, #_TIF_WORK_MASK
545 cand.a r1, #_TIF_NEED_RESCHED
549 cand.a r1, #_TIF_SIGPENDING @ delivering a signal?
561 disable_irq r1 @ disable interrupts
563 ldw r1, [tsk+], #TI_FLAGS
564 cand.a r1, #_TIF_WORK_MASK
633 add r1, sp, #S_OFF
643 add r1, sp, #S_OFF
649 add r1, sp, #S_R0 + S_OFF @ pointer to regs
652 ldm (r0 - r3), [r1]+ @ have to reload r0 - r3
658 mov r1, sp
/linux-4.1.27/arch/tile/lib/
H A Dusercopy_32.S23 * strnlen_user_asm takes the pointer in r0, and the length bound in r1.
28 { bz r1, 2f; addi r3, r0, -1 } /* bias down to include NUL */
29 1: { lb_u r4, r0; addi r1, r1, -1 }
31 { bnzt r1, 1b; addi r0, r0, 1 }
45 * the userspace source pointer in r1, and the length bound (including
51 1: { lb_u r4, r1; addi r1, r1, 1; addi r2, r2, -1 }
70 * number of bytes to zero in r1.
76 { bz r1, 2f; or r2, r0, r1 }
79 1: { sb r0, zero; addi r0, r0, 1; addi r1, r1, -1 }
80 bnzt r1, 1b
81 2: { move r0, r1; jrp lr }
88 1: { sw r0, zero; addi r0, r0, 4; addi r1, r1, -4 }
89 bnzt r1, 1b
90 2: { move r0, r1; jrp lr }
99 * number of bytes to flush in r1.
103 bz r1, 2f
104 { movei r2, L2_CACHE_BYTES; add r1, r0, r1 }
105 { sub r2, zero, r2; addi r1, r1, L2_CACHE_BYTES-1 }
106 { and r0, r0, r2; and r1, r1, r2 }
107 { sub r1, r1, r0 }
108 1: { flush r0; addi r1, r1, -CHIP_FLUSH_STRIDE() }
109 { addi r0, r0, CHIP_FLUSH_STRIDE(); bnzt r1, 1b }
110 2: { move r0, r1; jrp lr }
119 * number of bytes to flush-invalidate in r1.
123 bz r1, 2f
124 { movei r2, L2_CACHE_BYTES; add r1, r0, r1 }
125 { sub r2, zero, r2; addi r1, r1, L2_CACHE_BYTES-1 }
126 { and r0, r0, r2; and r1, r1, r2 }
127 { sub r1, r1, r0 }
128 1: { finv r0; addi r1, r1, -CHIP_FINV_STRIDE() }
129 { addi r0, r0, CHIP_FINV_STRIDE(); bnzt r1, 1b }
130 2: { move r0, r1; jrp lr }
H A Dusercopy_64.S23 * strnlen_user_asm takes the pointer in r0, and the length bound in r1.
28 { beqz r1, 2f; addi r3, r0, -1 } /* bias down to include NUL */
29 1: { ld1u r4, r0; addi r1, r1, -1 }
31 { bnezt r1, 1b; addi r0, r0, 1 }
45 * the userspace source pointer in r1, and the length bound (including
51 1: { ld1u r4, r1; addi r1, r1, 1; addi r2, r2, -1 }
70 * number of bytes to zero in r1.
76 { beqz r1, 2f; or r2, r0, r1 }
79 1: { st1 r0, zero; addi r0, r0, 1; addi r1, r1, -1 }
80 bnezt r1, 1b
81 2: { move r0, r1; jrp lr }
88 1: { st r0, zero; addi r0, r0, 8; addi r1, r1, -8 }
89 bnezt r1, 1b
90 2: { move r0, r1; jrp lr }
99 * number of bytes to flush in r1.
103 beqz r1, 2f
104 { movei r2, L2_CACHE_BYTES; add r1, r0, r1 }
105 { sub r2, zero, r2; addi r1, r1, L2_CACHE_BYTES-1 }
106 { and r0, r0, r2; and r1, r1, r2 }
107 { sub r1, r1, r0 }
108 1: { flush r0; addi r1, r1, -CHIP_FLUSH_STRIDE() }
109 { addi r0, r0, CHIP_FLUSH_STRIDE(); bnezt r1, 1b }
110 2: { move r0, r1; jrp lr }
119 * number of bytes to flush-invalidate in r1.
123 beqz r1, 2f
124 { movei r2, L2_CACHE_BYTES; add r1, r0, r1 }
125 { sub r2, zero, r2; addi r1, r1, L2_CACHE_BYTES-1 }
126 { and r0, r0, r2; and r1, r1, r2 }
127 { sub r1, r1, r0 }
128 1: { finv r0; addi r1, r1, -CHIP_FINV_STRIDE() }
129 { addi r0, r0, CHIP_FINV_STRIDE(); bnezt r1, 1b }
130 2: { move r0, r1; jrp lr }
H A Dmemcpy_32.S46 * the user source in r1, and the bytes to copy in r2.
67 * the kernel source in r1, and the bytes to copy in r2.
88 /* r0 is the dest, r1 is the source, r2 is the size. */
91 { sw sp, lr; move r23, r0; or r4, r0, r1 }
97 { move r24, r1; move r25, r2 }
108 { andi r6, r1, 63; j .Lcopy_many }
118 EX: { lw r3, r1; addi r1, r1, 4; slti_u r8, r2, 16 }
119 EX: { lw r4, r1; addi r1, r1, 4 }
127 EX: { lw r3, r1; addi r1, r1, 4 }
144 /* Copy words until r1 is cache-line-aligned. */
146 EX: { lw r3, r1; addi r1, r1, 4 }
147 { andi r6, r1, 63 }
152 { addi r3, r1, 60; andi r9, r9, -64 }
169 * - r1 points to start of src line 0
171 * - r3 contains r1 + 128 + 60 [pointer to end of source line 2]
175 * - r5 contains *(r1 + 60) [i.e. last word of source line 0]
176 * - r6 contains *(r1 + 64 + 60) [i.e. last word of source line 1]
182 { jal .Lcopy_line2; add r15, r1, r2 }
186 EX: { move r12, r5; lw r16, r1 }
193 EX: { move r12, r6; lw r16, r1 }
200 EX: { move r12, r7; lw r16, r1 }
212 * - r1 points to the source line.
217 * - r17 == r1 + 16.
222 * - r1 is incremented by 64, unless that would point to a word
255 { add r15, r1, r2 }
257 { slt_u r13, r3, r15; addi r17, r1, 16 }
262 EX: { lw r17, r17; addi r1, r1, 48; mvz r3, r13, r1 } /* r17 = WORD_4 */
271 EX: { lw r4, r1; addi r1, r1, 4; addi r20, r1, 16 } /* r4 = WORD_12 */
272 EX: { lw r8, r1; addi r1, r1, 4; slt_u r13, r20, r15 }/* r8 = WORD_13 */
273 EX: { lw r11, r1; addi r1, r1, -52; mvz r20, r13, r1 } /* r11 = WORD_14 */
278 EX: { lw r13, r1; addi r1, r1, 4; move zero, r16 } /* r13 = WORD_1 */
279 EX: { lw r14, r1; addi r1, r1, 4 } /* r14 = WORD_2 */
280 EX: { lw r15, r1; addi r1, r1, 8; addi r10, r0, 60 } /* r15 = WORD_3 */
285 EX: { lw r19, r1; addi r1, r1, 4; move zero, r17 } /* r19 = WORD_5 */
293 EX: { lw r10, r1; addi r1, r1, 4 } /* r10 = WORD_6 */
294 EX: { lw r12, r1; addi r1, r1, 4 } /* r12 = WORD_7 */
297 EX: { lw r18, r1; addi r1, r1, 4 } /* r18 = WORD_8 */
310 EX: { lw r13, r1; addi r1, r1, 4; move zero, r18 } /* r13 = WORD_9 */
311 EX: { lw r14, r1; addi r1, r1, 4 } /* r14 = WORD_10 */
312 EX: { lw r15, r1; move r1, r20 } /* r15 = WORD_11 */
348 { bz r4, .Ldest_is_word_aligned; add r18, r1, r2 }
358 EX: { lb_u r3, r1; addi r1, r1, 1; addi r4, r4, -1 }
360 { bnzt r4, .Lalign_dest_loop; andi r3, r1, 3 }
367 EX: { andi r8, r0, 63; lwadd_na r6, r1, 4}
376 * - r1 points 4 bytes past the load address corresponding to r0.
381 EX: { lwadd_na r7, r1, 4; slti_u r8, r2, 4 + 4 }
383 { dword_align r6, r7, r1; slti_u r9, r2, 64 + 4 }
390 * - r1 points 4 bytes past the load address corresponding to r0.
407 { addi r3, r1, 63 - 4; addi r8, r1, 64 + 63 - 4 }
409 { mvz r3, r8, r1; addi r8, r3, 64 }
411 { mvz r3, r8, r1; movei r17, 0 }
415 { prefetch r3; addi r15, r1, 60; addi r3, r3, 64 }
419 EX: { mvz r3, r8, r1; wh64 r0 }
435 lwadd_na r7, r1, 16
437 EX: { lwadd_na r11, r1, 12 }
438 EX: { lwadd_na r14, r1, -24 }
439 EX: { lwadd_na r8, r1, 4 }
440 EX: { lwadd_na r9, r1, 4 }
442 lwadd_na r10, r1, 8
446 EX: { lwadd_na r12, r1, 4; addi r17, r17, 1 }
447 EX: { lwadd_na r13, r1, 8; dword_align r6, r7, r1 }
448 EX: { swadd r0, r6, 4; dword_align r7, r8, r1 }
449 EX: { swadd r0, r7, 4; dword_align r8, r9, r1 }
450 EX: { swadd r0, r8, 4; dword_align r9, r10, r1 }
451 EX: { swadd r0, r9, 4; dword_align r10, r11, r1 }
452 EX: { swadd r0, r10, 4; dword_align r11, r12, r1 }
453 EX: { swadd r0, r11, 4; dword_align r12, r13, r1 }
454 EX: { swadd r0, r12, 4; dword_align r13, r14, r1 }
462 * - r1 points 4 bytes past the load address corresponding to r0.
471 /* Move r1 back to the point where it corresponds to r0. */
472 { addi r1, r1, -4 }
483 EX: { lb_u r3, r1; addi r1, r1, 1 }
500 * r2 (num remaining) is correct, but r0 (dst) and r1 (src)
510 /* Add this to the original r0 and r1 to get their new values. */
511 { add r0, r23, r3; add r1, r24, r3 }
518 .Lcfu: { lb_u r3, r1; addi r1, r1, 1 }
533 { lb_u r3, r1; addi r1, r1, 1 }
544 { lb_u r3, r1; addi r1, r1, 1 }
/linux-4.1.27/arch/m32r/mm/
H A Dmmu.S31 st r1, @-sp
36 ld r1, @(MESTS_offset, r3) ; r1: status (MESTS reg.)
38 st r1, @(MESTS_offset, r3) ; clear status (MESTS reg.)
39 and3 r1, r1, #(MESTS_IT)
40 bnez r1, 1f ; instruction TLB miss?
45 ;; r1 - r3: free
48 ;; r1: TLB entry base address
56 ldi r1, #-8192
59 and r1, sp
60 ld r1, @(16, r1) ; current_thread_info->cpu
61 slli r1, #2
62 add r2, r1
64 seth r1, #high(DTLB_BASE)
65 or3 r1, r1, #low(DTLB_BASE)
74 ;; r1 - r3: free
77 ;; r1: TLB entry base address
82 mvfc r1, bpc
83 and r1, r3
84 or r0, r1 ; r0: PFN + ASID
89 ldi r1, #-8192
92 and r1, sp
93 ld r1, @(16, r1) ; current_thread_info->cpu
94 slli r1, #2
95 add r2, r1
97 seth r1, #high(ITLB_BASE)
98 or3 r1, r1, #low(ITLB_BASE)
105 ;; r1: TLB entry base address
110 ;; r1: TLB entry address
113 ld r3, @r2 || srli r1, #3
116 srli r1, #3
118 add r1, r3
123 st r3, @r2 || slli r1, #3
126 slli r1, #3
132 ;; r1: TLB entry address
136 ;; r1: TLB entry address
178 ;; r1: TLB entry address
181 st r0, @r1 ; set_tlb_tag(entry++, address);
182 st r2, @+r1 ; set_tlb_data(entry, pte_data);
188 ld r1, @sp+
197 ;; r1: TLB entry address
201 ;; r1: TLB entry address
215 st r1, @-sp
223 ld r1, @(MESTS_offset,r3) ; r1: status (MESTS reg.)
224 st r1, @(MESTS_offset,r3) ; clear status (MESTS reg.)
225 and3 r1, r1, #(MESTS_IT)
226 beqz r1, 1f ; data TLB miss?
234 seth r1, #high(ITLB_BASE)
235 or3 r1, r1, #low(ITLB_BASE)
236 add r2, r1 ; r2: entry
248 seth r1, #high(DTLB_BASE)
249 or3 r1, r1, #low(DTLB_BASE)
250 add r2, r1 ; r2: entry
258 ; r1,r3,r4: (free)
260 ld24 r1, #(-MPTB-1)
261 not r1, r1
262 ld r1, @r1
265 add r3, r1 ; r3: pgd
267 ld r1, @r3 ; r1: pmd
268 beqz r1, 3f ; pmd_none(*pmd) ?
270 and3 r1, r1, #0x3ff
272 bne r1, r4, 3f ; pmd_bad(*pmd) ?
286 ld r1, @r4 ; r1: pte_data
287 and3 r3, r1, #2 ; _PAGE_PRESENT(=2) check
292 ; r0: address, r1: pte_data, r2: entry
302 st r1, @(4,r2) ; set_tlb_data(entry, pte_data);
307 ld r1, @sp+
314 ldi r1, #2 ; r1: pte_data = 0 | _PAGE_PRESENT(=2)
325 ldi r1, #0
326 st r1, @(MPSZ_offset,r0) ; Set MPSZ Reg(Page size 4KB:0 16KB:1 64KB:2)
327 ldi r1, #0
328 st r1, @(MASID_offset,r0) ; Set ASID Zero
333 seth r1, #high(DTLB_BASE) ; Set DTLB_BASE higher
334 or3 r1, r1, #low(DTLB_BASE) ; Set DTLB_BASE lower
338 addi r1, #-4
342 st r2, @+r1 ; VPA <- 0
343 st r2, @+r1 ; PPA <- 0
H A Dpage.S28 ld r4, @r1+
29 ld r5, @r1+
30 ld r6, @r1+
31 ld r7, @r1+
38 ld r4, @r1+
40 ld r5, @r1+
41 ld r6, @r1+
42 ld r7, @r1+
/linux-4.1.27/arch/sh/lib/
H A Dmemmove.S30 mov #12,r1
31 cmp/gt r6,r1
44 mov r4,r1
46 and r2,r1
47 shll2 r1
50 add r1,r0
51 mov.l @r0,r1
52 jmp @r1
62 8: mov.b @r0+,r1
65 mov.b r1,@(r0,r4)
89 mov.b @r0+,r1
91 mov.b r1,@(r0,r4)
96 3: mov.l @r0+,r1
99 mov.l r1,@(r0,r4)
122 mov.b @r0+,r1
124 mov.b r1,@(r0,r4)
128 mov.l @(r0,r4),r1
133 shll8 r1
134 3: mov r1,r3 ! JIHG
136 mov.l @r0+,r1 ! NMLK
137 mov r1,r2
145 shlr8 r1
146 3: mov r1,r3 ! GHIJ
148 mov.l @r0+,r1 ! KLMN
149 mov r1,r2
175 mov.b @r0+,r1
176 mov.b r1,@(r0,r4)
182 3: mov.w @r0+,r1
185 mov.w r1,@(r0,r4)
192 mov.b @r0,r1
193 mov.b r1,@(r0,r4)
210 mov.b @r0+,r1
212 mov.b r1,@(r0,r4)
215 mov.l @(r0,r4),r1
220 shll16 r1
221 shll8 r1
222 3: mov r1,r3 ! JIHG
225 mov.l @r0+,r1 ! NMLK
226 mov r1,r2
233 shlr16 r1
234 shlr8 r1
235 3: mov r1,r3 ! GHIJ
238 mov.l @r0+,r1 ! KLMN
239 mov r1,r2
H A Dmemcpy.S21 mov #12,r1
22 cmp/gt r6,r1
34 mov r5,r1
36 and r2,r1
37 shll2 r1
40 add r1,r0
41 mov.l @r0,r1
42 jmp @r1
56 mov.b @(r0,r5),r1
58 mov.b r1,@-r0
75 mov.b @(r0,r5),r1
77 mov.b r1,@-r0
83 3: mov.l @(r0,r5),r1
86 mov.l r1,@-r0
106 mov.b @(r0,r5),r1
108 mov.b r1,@-r0
111 mov.l @(r0,r5),r1
117 3: mov r1,r3 ! RQPO
120 mov.l @(r0,r5),r1 ! NMLK
121 mov r1,r6
128 3: mov r1,r3 ! OPQR
131 mov.l @(r0,r5),r1 ! KLMN
132 mov r1,r6
155 mov.b @(r0,r5),r1
156 mov.b r1,@-r0
163 3: mov.w @(r0,r5),r1
166 mov.w r1,@-r0
172 mov.b @(r0,r5),r1
174 mov.b r1,@-r0
187 mov.b @(r0,r5),r1
189 mov.b r1,@-r0
193 mov.l @(r0,r5),r1
199 3: mov r1,r3 ! RQPO
201 mov.l @(r0,r5),r1 ! NMLK
202 mov r1,r6
210 3: mov r1,r3 ! OPQR
212 mov.l @(r0,r5),r1 ! KLMN
213 mov r1,r6
H A Dchecksum.S90 mov r5, r1
92 shld r0, r1
93 tst r1, r1
115 dt r1
118 ! here, we know r1==0
119 addc r1, r6 ! add carry to r6
126 mov r0, r1
127 shlr2 r1
133 dt r1
137 addc r1, r6 ! r1==0 here, so it means add carry-bit
144 mov #2, r1
145 cmp/hs r1, r5
149 cmp/eq r1, r5
221 mov r4,r1
222 and r0,r1
224 cmp/eq r1,r0
249 SRC( mov.b @r4+,r1 )
251 extu.b r1,r1
252 DST( mov.b r1,@r5 )
260 shll8 r1
262 or r1,r0
299 SRC( mov.l @r4+,r1 )
302 DST( mov.l r1,@(4,r5) )
303 addc r1,r7
306 SRC( mov.l @r4+,r1 )
309 DST( mov.l r1,@(12,r5) )
310 addc r1,r7
313 SRC( mov.l @r4+,r1 )
316 DST( mov.l r1,@(20,r5) )
317 addc r1,r7
320 SRC( mov.l @r4+,r1 )
323 DST( mov.l r1,@(28,r5) )
324 addc r1,r7
356 mov #2,r1
357 cmp/hs r1,r6
363 cmp/eq r1,r6
386 mov #-EFAULT,r1
387 mov.l r1,@r0
406 mov #-EFAULT,r1
407 mov.l r1,@r0
H A Dcopy_page.S20 * r0, r1, r2, r3, r4, r5, r6, r7 --- scratch
39 mov.l @r11+,r1
58 mov.l r1,@-r10
104 EX( mov.b @r5+,r1 )
107 EX( mov.b r1,@r4 )
112 2: mov #3,r1
114 and r4,r1
116 shll2 r1
118 mov.l @(r0,r1),r1
119 jmp @r1
154 mov.l 8000f,r1
156 jmp @r1
183 EX( mov.l @r5+,r1 )
185 EX( mov.l r1,@r4 )
191 EX( mov.l @r5+,r1 )
204 EX( mov.l r1,@(4,r4) )
221 EX( mov.l @r5+,r1 )
223 EX( mov.l r1,@r4 )
244 EX( mov.l @r5+,r1 )
250 xtrct r1,r0
251 xtrct r8,r1
256 EX( mov.l r1,@(4,r4) )
260 EX( mov.l @r5+,r1 )
263 xtrct r1,r10
264 xtrct r8,r1
268 EX( mov.l r1,@(20,r4) )
377 mov.l 8000f,r1
379 jmp @r1
H A Dudivsi3_i4i.S63 mov.w c128_w, r1
67 cmp/hi r1,r5
68 extu.w r5,r1
70 cmp/eq r5,r1
73 mov r5,r1
77 mov.l r1,@-r15
86 mov.b @(r0,r5),r1
90 mov.b @(r0,r5),r1
94 mov.l @(r0,r1),r1
98 dmulu.l r1,r4
99 mov.b @(r0,r5),r1
109 shld r1,r0
126 mov r5,r1
131 mov.l r1,@-r15
134 mov.l zero_l,r1
138 mov.l r1,@-r15
140 mov.w m256_w,r1
144 and r1,r0
160 mov r0,r1
161 div1 r5,r1
165 div1 r5,r1
167 rotcl r0; div1 r5,r1
171 div1 r4,r1
183 but we effectively clobber only r1. */
187 mov.w c128_w, r1
193 cmp/hi r1,r5
209 mov.l zero_l,r1
212 mov.l r1,@-r15
238 mov.b @(r0,r5),r1
241 mov.l @(r0,r1),r1
243 dmulu.l r1,r4
244 mov.b @(r0,r5),r1
251 shld r1,r0
258 cmp/hi r1,r5
269 mov.l zero_l,r1
272 mov.l r1,@-r15
288 extu.b r0,r1
290 or r4,r1
293 rotcl r1
295 neg r1,r0
301 mov.l zero_l,r1
305 mov.l r1,@-r15
307 mov.w m256_w,r1
311 and r1,r0
321 mov r4,r1
322 shll8 r1
325 rotcl r1; div1 r5,r0
328 rotcl r1
H A Dudiv_qrnnd.S41 /* r0: rn r1: qn */ /* r0: n1 r4: n0 r5: d r6: d1 */ /* r2: __m */
52 extu.w r0,r1
55 0: rotcl r1
56 mulu.w r1,r5
64 add #-1,r1
66 1: add #-1,r1
72 swap.w r4,r1
73 xtrct r0,r1
75 mov r1,r0
77 mov #-1,r1
79 shlr16 r1
H A Dmcount.S49 /* r1 = sp & (THREAD_SIZE - 1) */ \
50 mov #-1, r1; \
51 add r0, r1; \
52 and r15, r1; \
60 cmp/hi r2, r1; \
64 mov.l .L_ebss, r1; \
65 cmp/hi r1, r15; \
69 mov.l .L_init_thread_union, r1; \
70 cmp/hs r1, r15; \
74 add r0, r1; \
75 cmp/hs r1, r15; \
206 mov.l 2f, r1
207 jmp @r1
236 mov.l r1, @-r15
249 mov.l @r15+, r1
/linux-4.1.27/crypto/
H A Dserpent_generic.c237 u32 r0, r1, r2, r3, r4; __serpent_setkey() local
252 r1 = le32_to_cpu(k[4]); __serpent_setkey()
258 keyiter(le32_to_cpu(k[1]), r1, r0, r3, 1, 1); __serpent_setkey() local
259 keyiter(le32_to_cpu(k[2]), r2, r1, r4, 2, 2); __serpent_setkey() local
261 keyiter(le32_to_cpu(k[4]), r4, r3, r1, 4, 4); __serpent_setkey() local
263 keyiter(le32_to_cpu(k[6]), r1, r0, r3, 6, 6); __serpent_setkey() local
264 keyiter(le32_to_cpu(k[7]), r2, r1, r4, 7, 7); __serpent_setkey() local
267 keyiter(k[1], r4, r3, r1, 9, 9); __serpent_setkey()
269 keyiter(k[3], r1, r0, r3, 11, 11); __serpent_setkey()
270 keyiter(k[4], r2, r1, r4, 12, 12); __serpent_setkey()
272 keyiter(k[6], r4, r3, r1, 14, 14); __serpent_setkey()
274 keyiter(k[8], r1, r0, r3, 16, 16); __serpent_setkey()
275 keyiter(k[9], r2, r1, r4, 17, 17); __serpent_setkey()
277 keyiter(k[11], r4, r3, r1, 19, 19); __serpent_setkey()
279 keyiter(k[13], r1, r0, r3, 21, 21); __serpent_setkey()
280 keyiter(k[14], r2, r1, r4, 22, 22); __serpent_setkey()
282 keyiter(k[16], r4, r3, r1, 24, 24); __serpent_setkey()
284 keyiter(k[18], r1, r0, r3, 26, 26); __serpent_setkey()
285 keyiter(k[19], r2, r1, r4, 27, 27); __serpent_setkey()
287 keyiter(k[21], r4, r3, r1, 29, 29); __serpent_setkey()
289 keyiter(k[23], r1, r0, r3, 31, 31); __serpent_setkey()
293 keyiter(k[-26], r2, r1, r4, 32, -18); __serpent_setkey()
295 keyiter(k[-24], r4, r3, r1, 34, -16); __serpent_setkey()
297 keyiter(k[-22], r1, r0, r3, 36, -14); __serpent_setkey()
298 keyiter(k[-21], r2, r1, r4, 37, -13); __serpent_setkey()
300 keyiter(k[-19], r4, r3, r1, 39, -11); __serpent_setkey()
302 keyiter(k[-17], r1, r0, r3, 41, -9); __serpent_setkey()
303 keyiter(k[-16], r2, r1, r4, 42, -8); __serpent_setkey()
305 keyiter(k[-14], r4, r3, r1, 44, -6); __serpent_setkey()
307 keyiter(k[-12], r1, r0, r3, 46, -4); __serpent_setkey()
308 keyiter(k[-11], r2, r1, r4, 47, -3); __serpent_setkey()
310 keyiter(k[-9], r4, r3, r1, 49, -1); __serpent_setkey()
312 keyiter(k[-7], r1, r0, r3, 51, 1); __serpent_setkey()
313 keyiter(k[-6], r2, r1, r4, 52, 2); __serpent_setkey()
315 keyiter(k[-4], r4, r3, r1, 54, 4); __serpent_setkey()
317 keyiter(k[-2], r1, r0, r3, 56, 6); __serpent_setkey()
318 keyiter(k[-1], r2, r1, r4, 57, 7); __serpent_setkey()
320 keyiter(k[1], r4, r3, r1, 59, 9); __serpent_setkey()
322 keyiter(k[3], r1, r0, r3, 61, 11); __serpent_setkey()
323 keyiter(k[4], r2, r1, r4, 62, 12); __serpent_setkey()
325 keyiter(k[6], r4, r3, r1, 64, 14); __serpent_setkey()
327 keyiter(k[8], r1, r0, r3, 66, 16); __serpent_setkey()
328 keyiter(k[9], r2, r1, r4, 67, 17); __serpent_setkey()
330 keyiter(k[11], r4, r3, r1, 69, 19); __serpent_setkey()
332 keyiter(k[13], r1, r0, r3, 71, 21); __serpent_setkey()
333 keyiter(k[14], r2, r1, r4, 72, 22); __serpent_setkey()
335 keyiter(k[16], r4, r3, r1, 74, 24); __serpent_setkey()
337 keyiter(k[18], r1, r0, r3, 76, 26); __serpent_setkey()
338 keyiter(k[19], r2, r1, r4, 77, 27); __serpent_setkey()
340 keyiter(k[21], r4, r3, r1, 79, 29); __serpent_setkey()
342 keyiter(k[23], r1, r0, r3, 81, 31); __serpent_setkey()
346 keyiter(k[-26], r2, r1, r4, 82, -18); __serpent_setkey()
348 keyiter(k[-24], r4, r3, r1, 84, -16); __serpent_setkey()
350 keyiter(k[-22], r1, r0, r3, 86, -14); __serpent_setkey()
351 keyiter(k[-21], r2, r1, r4, 87, -13); __serpent_setkey()
353 keyiter(k[-19], r4, r3, r1, 89, -11); __serpent_setkey()
355 keyiter(k[-17], r1, r0, r3, 91, -9); __serpent_setkey()
356 keyiter(k[-16], r2, r1, r4, 92, -8); __serpent_setkey()
358 keyiter(k[-14], r4, r3, r1, 94, -6); __serpent_setkey()
360 keyiter(k[-12], r1, r0, r3, 96, -4); __serpent_setkey()
361 keyiter(k[-11], r2, r1, r4, 97, -3); __serpent_setkey()
363 keyiter(k[-9], r4, r3, r1, 99, -1); __serpent_setkey()
365 keyiter(k[-7], r1, r0, r3, 101, 1); __serpent_setkey()
366 keyiter(k[-6], r2, r1, r4, 102, 2); __serpent_setkey()
368 keyiter(k[-4], r4, r3, r1, 104, 4); __serpent_setkey()
370 keyiter(k[-2], r1, r0, r3, 106, 6); __serpent_setkey()
371 keyiter(k[-1], r2, r1, r4, 107, 7); __serpent_setkey()
373 keyiter(k[1], r4, r3, r1, 109, 9); __serpent_setkey()
375 keyiter(k[3], r1, r0, r3, 111, 11); __serpent_setkey()
376 keyiter(k[4], r2, r1, r4, 112, 12); __serpent_setkey()
378 keyiter(k[6], r4, r3, r1, 114, 14); __serpent_setkey()
380 keyiter(k[8], r1, r0, r3, 116, 16); __serpent_setkey()
381 keyiter(k[9], r2, r1, r4, 117, 17); __serpent_setkey()
383 keyiter(k[11], r4, r3, r1, 119, 19); __serpent_setkey()
385 keyiter(k[13], r1, r0, r3, 121, 21); __serpent_setkey()
386 keyiter(k[14], r2, r1, r4, 122, 22); __serpent_setkey()
388 keyiter(k[16], r4, r3, r1, 124, 24); __serpent_setkey()
390 keyiter(k[18], r1, r0, r3, 126, 26); __serpent_setkey()
391 keyiter(k[19], r2, r1, r4, 127, 27); __serpent_setkey()
393 keyiter(k[21], r4, r3, r1, 129, 29); __serpent_setkey()
395 keyiter(k[23], r1, r0, r3, 131, 31); __serpent_setkey()
399 S3(r3, r4, r0, r1, r2); store_and_load_keys(r1, r2, r4, r3, 28, 24); __serpent_setkey()
400 S4(r1, r2, r4, r3, r0); store_and_load_keys(r2, r4, r3, r0, 24, 20); __serpent_setkey()
401 S5(r2, r4, r3, r0, r1); store_and_load_keys(r1, r2, r4, r0, 20, 16); __serpent_setkey()
402 S6(r1, r2, r4, r0, r3); store_and_load_keys(r4, r3, r2, r0, 16, 12); __serpent_setkey()
403 S7(r4, r3, r2, r0, r1); store_and_load_keys(r1, r2, r0, r4, 12, 8); __serpent_setkey()
404 S0(r1, r2, r0, r4, r3); store_and_load_keys(r0, r2, r4, r1, 8, 4); __serpent_setkey()
405 S1(r0, r2, r4, r1, r3); store_and_load_keys(r3, r4, r1, r0, 4, 0); __serpent_setkey()
406 S2(r3, r4, r1, r0, r2); store_and_load_keys(r2, r4, r3, r0, 0, -4); __serpent_setkey()
407 S3(r2, r4, r3, r0, r1); store_and_load_keys(r0, r1, r4, r2, -4, -8); __serpent_setkey()
408 S4(r0, r1, r4, r2, r3); store_and_load_keys(r1, r4, r2, r3, -8, -12); __serpent_setkey()
409 S5(r1, r4, r2, r3, r0); store_and_load_keys(r0, r1, r4, r3, -12, -16); __serpent_setkey()
410 S6(r0, r1, r4, r3, r2); store_and_load_keys(r4, r2, r1, r3, -16, -20); __serpent_setkey()
411 S7(r4, r2, r1, r3, r0); store_and_load_keys(r0, r1, r3, r4, -20, -24); __serpent_setkey()
412 S0(r0, r1, r3, r4, r2); store_and_load_keys(r3, r1, r4, r0, -24, -28); __serpent_setkey()
414 S1(r3, r1, r4, r0, r2); store_and_load_keys(r2, r4, r0, r3, 22, 18); __serpent_setkey()
415 S2(r2, r4, r0, r3, r1); store_and_load_keys(r1, r4, r2, r3, 18, 14); __serpent_setkey()
416 S3(r1, r4, r2, r3, r0); store_and_load_keys(r3, r0, r4, r1, 14, 10); __serpent_setkey()
417 S4(r3, r0, r4, r1, r2); store_and_load_keys(r0, r4, r1, r2, 10, 6); __serpent_setkey()
418 S5(r0, r4, r1, r2, r3); store_and_load_keys(r3, r0, r4, r2, 6, 2); __serpent_setkey()
419 S6(r3, r0, r4, r2, r1); store_and_load_keys(r4, r1, r0, r2, 2, -2); __serpent_setkey()
420 S7(r4, r1, r0, r2, r3); store_and_load_keys(r3, r0, r2, r4, -2, -6); __serpent_setkey()
421 S0(r3, r0, r2, r4, r1); store_and_load_keys(r2, r0, r4, r3, -6, -10); __serpent_setkey()
422 S1(r2, r0, r4, r3, r1); store_and_load_keys(r1, r4, r3, r2, -10, -14); __serpent_setkey()
423 S2(r1, r4, r3, r2, r0); store_and_load_keys(r0, r4, r1, r2, -14, -18); __serpent_setkey()
424 S3(r0, r4, r1, r2, r3); store_and_load_keys(r2, r3, r4, r0, -18, -22); __serpent_setkey()
426 S4(r2, r3, r4, r0, r1); store_and_load_keys(r3, r4, r0, r1, 28, 24); __serpent_setkey()
427 S5(r3, r4, r0, r1, r2); store_and_load_keys(r2, r3, r4, r1, 24, 20); __serpent_setkey()
428 S6(r2, r3, r4, r1, r0); store_and_load_keys(r4, r0, r3, r1, 20, 16); __serpent_setkey()
429 S7(r4, r0, r3, r1, r2); store_and_load_keys(r2, r3, r1, r4, 16, 12); __serpent_setkey()
430 S0(r2, r3, r1, r4, r0); store_and_load_keys(r1, r3, r4, r2, 12, 8); __serpent_setkey()
431 S1(r1, r3, r4, r2, r0); store_and_load_keys(r0, r4, r2, r1, 8, 4); __serpent_setkey()
432 S2(r0, r4, r2, r1, r3); store_and_load_keys(r3, r4, r0, r1, 4, 0); __serpent_setkey()
433 S3(r3, r4, r0, r1, r2); storekeys(r1, r2, r4, r3, 0); __serpent_setkey()
450 u32 r0, r1, r2, r3, r4; __serpent_encrypt() local
458 r1 = le32_to_cpu(s[1]); __serpent_encrypt()
462 K(r0, r1, r2, r3, 0); __serpent_encrypt()
463 S0(r0, r1, r2, r3, r4); LK(r2, r1, r3, r0, r4, 1); __serpent_encrypt()
464 S1(r2, r1, r3, r0, r4); LK(r4, r3, r0, r2, r1, 2); __serpent_encrypt()
465 S2(r4, r3, r0, r2, r1); LK(r1, r3, r4, r2, r0, 3); __serpent_encrypt()
466 S3(r1, r3, r4, r2, r0); LK(r2, r0, r3, r1, r4, 4); __serpent_encrypt()
467 S4(r2, r0, r3, r1, r4); LK(r0, r3, r1, r4, r2, 5); __serpent_encrypt()
468 S5(r0, r3, r1, r4, r2); LK(r2, r0, r3, r4, r1, 6); __serpent_encrypt()
469 S6(r2, r0, r3, r4, r1); LK(r3, r1, r0, r4, r2, 7); __serpent_encrypt()
470 S7(r3, r1, r0, r4, r2); LK(r2, r0, r4, r3, r1, 8); __serpent_encrypt()
471 S0(r2, r0, r4, r3, r1); LK(r4, r0, r3, r2, r1, 9); __serpent_encrypt()
472 S1(r4, r0, r3, r2, r1); LK(r1, r3, r2, r4, r0, 10); __serpent_encrypt()
473 S2(r1, r3, r2, r4, r0); LK(r0, r3, r1, r4, r2, 11); __serpent_encrypt()
474 S3(r0, r3, r1, r4, r2); LK(r4, r2, r3, r0, r1, 12); __serpent_encrypt()
475 S4(r4, r2, r3, r0, r1); LK(r2, r3, r0, r1, r4, 13); __serpent_encrypt()
476 S5(r2, r3, r0, r1, r4); LK(r4, r2, r3, r1, r0, 14); __serpent_encrypt()
477 S6(r4, r2, r3, r1, r0); LK(r3, r0, r2, r1, r4, 15); __serpent_encrypt()
478 S7(r3, r0, r2, r1, r4); LK(r4, r2, r1, r3, r0, 16); __serpent_encrypt()
479 S0(r4, r2, r1, r3, r0); LK(r1, r2, r3, r4, r0, 17); __serpent_encrypt()
480 S1(r1, r2, r3, r4, r0); LK(r0, r3, r4, r1, r2, 18); __serpent_encrypt()
481 S2(r0, r3, r4, r1, r2); LK(r2, r3, r0, r1, r4, 19); __serpent_encrypt()
482 S3(r2, r3, r0, r1, r4); LK(r1, r4, r3, r2, r0, 20); __serpent_encrypt()
483 S4(r1, r4, r3, r2, r0); LK(r4, r3, r2, r0, r1, 21); __serpent_encrypt()
484 S5(r4, r3, r2, r0, r1); LK(r1, r4, r3, r0, r2, 22); __serpent_encrypt()
485 S6(r1, r4, r3, r0, r2); LK(r3, r2, r4, r0, r1, 23); __serpent_encrypt()
486 S7(r3, r2, r4, r0, r1); LK(r1, r4, r0, r3, r2, 24); __serpent_encrypt()
487 S0(r1, r4, r0, r3, r2); LK(r0, r4, r3, r1, r2, 25); __serpent_encrypt()
488 S1(r0, r4, r3, r1, r2); LK(r2, r3, r1, r0, r4, 26); __serpent_encrypt()
489 S2(r2, r3, r1, r0, r4); LK(r4, r3, r2, r0, r1, 27); __serpent_encrypt()
490 S3(r4, r3, r2, r0, r1); LK(r0, r1, r3, r4, r2, 28); __serpent_encrypt()
491 S4(r0, r1, r3, r4, r2); LK(r1, r3, r4, r2, r0, 29); __serpent_encrypt()
492 S5(r1, r3, r4, r2, r0); LK(r0, r1, r3, r2, r4, 30); __serpent_encrypt()
493 S6(r0, r1, r3, r2, r4); LK(r3, r4, r1, r2, r0, 31); __serpent_encrypt()
494 S7(r3, r4, r1, r2, r0); K(r0, r1, r2, r3, 32); __serpent_encrypt()
497 d[1] = cpu_to_le32(r1); __serpent_encrypt()
515 u32 r0, r1, r2, r3, r4; __serpent_decrypt() local
518 r1 = le32_to_cpu(s[1]); __serpent_decrypt()
522 K(r0, r1, r2, r3, 32); __serpent_decrypt()
523 SI7(r0, r1, r2, r3, r4); KL(r1, r3, r0, r4, r2, 31); __serpent_decrypt()
524 SI6(r1, r3, r0, r4, r2); KL(r0, r2, r4, r1, r3, 30); __serpent_decrypt()
525 SI5(r0, r2, r4, r1, r3); KL(r2, r3, r0, r4, r1, 29); __serpent_decrypt()
526 SI4(r2, r3, r0, r4, r1); KL(r2, r0, r1, r4, r3, 28); __serpent_decrypt()
527 SI3(r2, r0, r1, r4, r3); KL(r1, r2, r3, r4, r0, 27); __serpent_decrypt()
528 SI2(r1, r2, r3, r4, r0); KL(r2, r0, r4, r3, r1, 26); __serpent_decrypt()
529 SI1(r2, r0, r4, r3, r1); KL(r1, r0, r4, r3, r2, 25); __serpent_decrypt()
530 SI0(r1, r0, r4, r3, r2); KL(r4, r2, r0, r1, r3, 24); __serpent_decrypt()
531 SI7(r4, r2, r0, r1, r3); KL(r2, r1, r4, r3, r0, 23); __serpent_decrypt()
532 SI6(r2, r1, r4, r3, r0); KL(r4, r0, r3, r2, r1, 22); __serpent_decrypt()
533 SI5(r4, r0, r3, r2, r1); KL(r0, r1, r4, r3, r2, 21); __serpent_decrypt()
534 SI4(r0, r1, r4, r3, r2); KL(r0, r4, r2, r3, r1, 20); __serpent_decrypt()
535 SI3(r0, r4, r2, r3, r1); KL(r2, r0, r1, r3, r4, 19); __serpent_decrypt()
536 SI2(r2, r0, r1, r3, r4); KL(r0, r4, r3, r1, r2, 18); __serpent_decrypt()
537 SI1(r0, r4, r3, r1, r2); KL(r2, r4, r3, r1, r0, 17); __serpent_decrypt()
538 SI0(r2, r4, r3, r1, r0); KL(r3, r0, r4, r2, r1, 16); __serpent_decrypt()
539 SI7(r3, r0, r4, r2, r1); KL(r0, r2, r3, r1, r4, 15); __serpent_decrypt()
540 SI6(r0, r2, r3, r1, r4); KL(r3, r4, r1, r0, r2, 14); __serpent_decrypt()
541 SI5(r3, r4, r1, r0, r2); KL(r4, r2, r3, r1, r0, 13); __serpent_decrypt()
542 SI4(r4, r2, r3, r1, r0); KL(r4, r3, r0, r1, r2, 12); __serpent_decrypt()
543 SI3(r4, r3, r0, r1, r2); KL(r0, r4, r2, r1, r3, 11); __serpent_decrypt()
544 SI2(r0, r4, r2, r1, r3); KL(r4, r3, r1, r2, r0, 10); __serpent_decrypt()
545 SI1(r4, r3, r1, r2, r0); KL(r0, r3, r1, r2, r4, 9); __serpent_decrypt()
546 SI0(r0, r3, r1, r2, r4); KL(r1, r4, r3, r0, r2, 8); __serpent_decrypt()
547 SI7(r1, r4, r3, r0, r2); KL(r4, r0, r1, r2, r3, 7); __serpent_decrypt()
548 SI6(r4, r0, r1, r2, r3); KL(r1, r3, r2, r4, r0, 6); __serpent_decrypt()
549 SI5(r1, r3, r2, r4, r0); KL(r3, r0, r1, r2, r4, 5); __serpent_decrypt()
550 SI4(r3, r0, r1, r2, r4); KL(r3, r1, r4, r2, r0, 4); __serpent_decrypt()
551 SI3(r3, r1, r4, r2, r0); KL(r4, r3, r0, r2, r1, 3); __serpent_decrypt()
552 SI2(r4, r3, r0, r2, r1); KL(r3, r1, r2, r0, r4, 2); __serpent_decrypt()
553 SI1(r3, r1, r2, r0, r4); KL(r4, r1, r2, r0, r3, 1); __serpent_decrypt()
554 SI0(r4, r1, r2, r0, r3); K(r2, r3, r1, r4, 0); __serpent_decrypt()
558 d[2] = cpu_to_le32(r1); __serpent_decrypt()
/linux-4.1.27/arch/arm/lib/
H A Dbitops.h8 ands ip, r1, #3 variable
9 strneb r1, [ip] @ assert word-aligned variable
13 add r1, r1, r0, lsl #2 @ Get word offset variable
16 ALT_SMP(W(pldw) [r1])
20 1: ldrex r2, [r1]
22 strex r0, r2, [r1]
33 ands ip, r1, #3 variable
34 strneb r1, [ip] @ assert word-aligned variable
38 add r1, r1, r0, lsl #2 @ Get word offset variable
43 ALT_SMP(W(pldw) [r1])
46 1: ldrex r2, [r1]
49 strex ip, r2, [r1]
63 ands ip, r1, #3
64 strneb r1, [ip] @ assert word-aligned
70 ldr r2, [r1, r0, lsl #2]
72 str r2, [r1, r0, lsl #2]
90 ands ip, r1, #3
91 strneb r1, [ip] @ assert word-aligned
95 ldr r2, [r1, r0, lsl #2]!
99 \store r2, [r1]
H A Dio-writesw-armv4.S25 .Loutsw_align: movs ip, r1, lsl #31
28 ldrh r3, [r1], #2
35 ands r3, r1, #3
43 .Loutsw_8_lp: ldmia r1!, {r3, r4, r5, ip}
54 ldmia r1!, {r3, ip}
61 ldr r3, [r1], #4
64 .Lno_outsw_2: ldrneh r3, [r1]
78 ARM( ldr r3, [r1, -r3]! )
80 THUMB( ldr r3, [r1, r3] )
81 THUMB( sub r1, r3 )
90 ldr r3, [r1, #4]!
H A Dcopy_page.S29 PLD( pld [r1, #0] )
30 PLD( pld [r1, #L1_CACHE_BYTES] )
32 ldmia r1!, {r3, r4, ip, lr} @ 4+1
33 1: PLD( pld [r1, #2 * L1_CACHE_BYTES])
34 PLD( pld [r1, #3 * L1_CACHE_BYTES])
38 ldmia r1!, {r3, r4, ip, lr} @ 4
42 ldmgtia r1!, {r3, r4, ip, lr} @ 4
44 PLD( ldmeqia r1!, {r3, r4, ip, lr} )
H A Dio-readsw-armv4.S21 .Linsw_align: movs ip, r1, lsl #31
25 strh ip, [r1], #2
30 tst r1, #3
55 stmia r1!, {r3 - r5, ip}
69 stmia r1!, {r3, r4}
77 str r3, [r1], #4
80 strneh r3, [r1]
97 ldrccb ip, [r1, #-1]!
103 strb ip, [r1], #1
116 str ip, [r1], #4
123 strb ip, [r1], #1
126 strneb ip, [r1], #1
129 strneb ip, [r1]
H A Dcsumpartialcopyuser.S21 stmfd sp!, {r1, r2, r4 - r8, lr}
25 ldmfd sp!, {r1, r2, r4 - r8, pc}
56 * r0 = src, r1 = dst, r2 = len, r3 = sum, [sp] = *err_ptr
76 ldmia sp, {r1, r2} @ retrieve dst, len
77 add r2, r2, r1
79 9002: teq r2, r1
80 strneb r0, [r1], #1
H A Dlib1funcs.S212 subs r2, r1, #1
215 cmp r0, r1
217 tst r1, r2
220 ARM_DIV_BODY r0, r1, r2, r3
229 12: ARM_DIV2_ORDER r1, r2
241 subs r2, r1, #1 @ compare divisor with 1
243 cmpne r0, r1 @ compare dividend with divisor
245 tsthi r1, r2 @ see if divisor is power of 2
249 ARM_MOD_BODY r0, r1, r2, r3
260 cmp r1, #0
261 eor ip, r0, r1 @ save the sign of the result.
263 rsbmi r1, r1, #0 @ loops below use unsigned.
264 subs r2, r1, #1 @ division by 1 or -1 ?
268 cmp r3, r1
270 tst r1, r2 @ divisor is power of 2 ?
273 ARM_DIV_BODY r3, r1, r0, r2
288 12: ARM_DIV2_ORDER r1, r2
302 cmp r1, #0
304 rsbmi r1, r1, #0 @ loops below use unsigned.
307 subs r2, r1, #1 @ compare divisor with 1
308 cmpne r0, r1 @ compare dividend with divisor
310 tsthi r1, r2 @ see if divisor is power of 2
314 ARM_MOD_BODY r0, r1, r2, r3
327 UNWIND(.save {r0, r1, ip, lr} )
329 stmfd sp!, {r0, r1, ip, lr}
331 ldmfd sp!, {r1, r2, ip, lr}
333 sub r1, r1, r3
341 UNWIND(.save {r0, r1, ip, lr} )
342 stmfd sp!, {r0, r1, ip, lr}
344 ldmfd sp!, {r1, r2, ip, lr}
346 sub r1, r1, r3
H A Dbacktrace.S35 tst r1, #0x10 @ 26 or 32-bit mode?
43 adr r1, 1b
44 sub offset, r0, r1
76 ldr r1, [frame, #-4] @ get saved lr
78 bic r1, r1, mask @ mask PC/LR for the mode
81 ldr r1, [sv_pc, #-4] @ if stmfd sp!, {args} exists,
83 teq r3, r1, lsr #11
88 1004: ldr r1, [sv_pc, #0] @ if stmfd sp!, {..., fp, ip, lr, pc}
90 teq r3, r1, lsr #11
102 mov r1, frame
121 mov instr, r1
135 mov r1, reg
H A Dgetuser.S19 * r1 contains the address limit, which must be preserved
36 check_uaccess r0, 1, r1, r2, __get_user_bad
43 check_uaccess r0, 2, r1, r2, __get_user_bad
63 check_uaccess r0, 4, r1, r2, __get_user_bad
70 check_uaccess r0, 8, r1, r2, __get_user_bad
84 check_uaccess r0, 8, r1, r2, __get_user_bad
96 check_uaccess r0, 1, r1, r2, __get_user_bad8
103 check_uaccess r0, 2, r1, r2, __get_user_bad8
119 check_uaccess r0, 4, r1, r2, __get_user_bad8
H A Dcsumpartialcopy.S16 * Params : r0 = src, r1 = dst, r2 = len, r3 = checksum
21 stmfd sp!, {r1, r4 - r8, lr}
25 ldmfd sp!, {r1, r4 - r8, pc}
H A Decard.S26 mov r11, r1
27 mov r1, r0
42 add pc, r1, #8
/linux-4.1.27/arch/sh/include/asm/
H A Datomic-grb.h12 " mov r15, r1 \n\t" /* r1 = saved sp */ \
17 "1: mov r1, r15 \n\t" /* LOGOUT */ \
21 : "memory" , "r0", "r1"); \
32 " mov r15, r1 \n\t" /* r1 = saved sp */ \
37 "1: mov r1, r15 \n\t" /* LOGOUT */ \
41 : "memory" , "r0", "r1"); \
63 " mov r15, r1 \n\t" /* r1 = saved sp */ atomic_clear_mask()
68 "1: mov r1, r15 \n\t" /* LOGOUT */ atomic_clear_mask()
72 : "memory" , "r0", "r1"); atomic_clear_mask()
82 " mov r15, r1 \n\t" /* r1 = saved sp */ atomic_set_mask()
87 "1: mov r1, r15 \n\t" /* LOGOUT */ atomic_set_mask()
91 : "memory" , "r0", "r1"); atomic_set_mask()
H A Dbitops-grb.h16 " mov r15, r1 \n\t" /* r1 = saved sp */ set_bit()
21 "1: mov r1, r15 \n\t" /* LOGOUT */ set_bit()
25 : "memory" , "r0", "r1"); set_bit()
39 " mov r15, r1 \n\t" /* r1 = saved sp */ clear_bit()
44 "1: mov r1, r15 \n\t" /* LOGOUT */ clear_bit()
48 : "memory" , "r0", "r1"); clear_bit()
62 " mov r15, r1 \n\t" /* r1 = saved sp */ change_bit()
67 "1: mov r1, r15 \n\t" /* LOGOUT */ change_bit()
71 : "memory" , "r0", "r1"); change_bit()
86 " mov r15, r1 \n\t" /* r1 = saved sp */ test_and_set_bit()
95 "1: mov r1, r15 \n\t" /* LOGOUT */ test_and_set_bit()
100 : "memory" , "r0", "r1" ,"t"); test_and_set_bit()
119 " mov r15, r1 \n\t" /* r1 = saved sp */ test_and_clear_bit()
128 "1: mov r1, r15 \n\t" /* LOGOUT */ test_and_clear_bit()
134 : "memory" , "r0", "r1", "t"); test_and_clear_bit()
151 " mov r15, r1 \n\t" /* r1 = saved sp */ test_and_change_bit()
160 "1: mov r1, r15 \n\t" /* LOGOUT */ test_and_change_bit()
165 : "memory" , "r0", "r1", "t"); test_and_change_bit()
H A Dromimage-macros.h10 mov.l 1f, r1
12 mov.l r0, @r1
23 mov.l 1f, r1
25 mov.w r0, @r1
36 mov.l 1f, r1
38 mov.b r0, @r1
64 mov.l 1f, r1
65 mov.l @r1, r0
H A Dcmpxchg-grb.h12 " mov r15, r1 \n\t" /* r1 = saved sp */ xchg_u32()
16 "1: mov r1, r15 \n\t" /* LOGOUT */ xchg_u32()
21 : "memory", "r0", "r1"); xchg_u32()
33 " mov r15, r1 \n\t" /* r1 = saved sp */ xchg_u8()
38 "1: mov r1, r15 \n\t" /* LOGOUT */ xchg_u8()
43 : "memory" , "r0", "r1"); xchg_u8()
57 " mov r15, r1 \n\t" /* r1 = saved sp */ __cmpxchg_u32()
63 "1: mov r1, r15 \n\t" /* LOGOUT */ __cmpxchg_u32()
67 : "memory" , "r0", "r1", "t"); __cmpxchg_u32()
/linux-4.1.27/arch/arm/mach-tegra/
H A Dsleep-tegra30.S169 cpu_to_csr_reg r1, r3
170 add r1, r1, r12 @ virtual CSR address for this CPU
187 str r12, [r1]
195 ldr r3, [r1] @ read CSR
196 str r3, [r1] @ clear CSR
265 mov32 r1, tegra30_iram_start
266 sub r0, r0, r1
267 mov32 r1, TEGRA_IRAM_LPx_RESUME_AREA
268 add r0, r0, r1
329 mov r1, #(1 << 28)
330 str r1, [r0, #CLK_RESET_SCLK_BURST]
331 str r1, [r0, #CLK_RESET_CCLK_BURST]
332 mov r1, #0
333 str r1, [r0, #CLK_RESET_CCLK_DIVIDER]
334 str r1, [r0, #CLK_RESET_SCLK_DIVIDER]
340 pll_iddq_exit r1, r0, CLK_RESET_PLLM_MISC, CLK_RESET_PLLM_MISC_IDDQ
341 pll_iddq_exit r1, r0, CLK_RESET_PLLC_MISC, CLK_RESET_PLLC_MISC_IDDQ
342 pll_iddq_exit r1, r0, CLK_RESET_PLLX_MISC3, CLK_RESET_PLLX_MISC3_IDDQ
345 ldr r1, [r7]
346 add r1, r1, #2
347 wait_until r1, r7, r3
351 ldr r1, [r2, #PMC_PLLP_WB0_OVERRIDE]
352 orr r1, r1, #(1 << 12)
353 str r1, [r2, #PMC_PLLP_WB0_OVERRIDE]
355 pll_enable r1, r0, CLK_RESET_PLLM_BASE, 0
356 pll_enable r1, r0, CLK_RESET_PLLC_BASE, 0
357 pll_enable r1, r0, CLK_RESET_PLLX_BASE, 0
364 ldr r1, [r2, #PMC_PLLP_WB0_OVERRIDE]
365 orr r1, r1, #(1 << 12)
366 str r1, [r2, #PMC_PLLP_WB0_OVERRIDE]
368 pll_enable r1, r0, CLK_RESET_PLLM_BASE, CLK_RESET_PLLM_MISC
369 pll_enable r1, r0, CLK_RESET_PLLC_BASE, CLK_RESET_PLLC_MISC
370 pll_enable r1, r0, CLK_RESET_PLLX_BASE, CLK_RESET_PLLX_MISC
373 pll_enable r1, r0, CLK_RESET_PLLP_BASE, CLK_RESET_PLLP_MISC
374 pll_enable r1, r0, CLK_RESET_PLLA_BASE, CLK_RESET_PLLA_MISC
376 pll_locked r1, r0, CLK_RESET_PLLM_BASE
377 pll_locked r1, r0, CLK_RESET_PLLP_BASE
378 pll_locked r1, r0, CLK_RESET_PLLA_BASE
379 pll_locked r1, r0, CLK_RESET_PLLC_BASE
380 pll_locked r1, r0, CLK_RESET_PLLX_BASE
383 ldr r1, [r7]
384 add r1, r1, #LOCK_DELAY
385 wait_until r1, r7, r3
403 ldr r1, [r5, #0x14] @ PMC_IO_DPD_STATUS
404 mvn r1, r1
405 bic r1, r1, #(1 << 31)
406 orr r1, r1, #(1 << 30)
407 str r1, [r2, #PMC_IO_DPD_REQ] @ DPD_OFF
420 ldr r1, [r5, #0xC] @ restore EMC_XM2VTTGENPADCTRL
421 str r1, [r0, #EMC_XM2VTTGENPADCTRL]
422 ldr r1, [r5, #0x10] @ restore EMC_XM2VTTGENPADCTRL2
423 str r1, [r0, #EMC_XM2VTTGENPADCTRL2]
424 ldr r1, [r5, #0x8] @ restore EMC_AUTO_CAL_INTERVAL
425 str r1, [r0, #EMC_AUTO_CAL_INTERVAL]
428 ldr r1, [r0, #EMC_CFG_DIG_DLL]
429 orr r1, r1, #(1 << 30) @ set DLL_RESET
430 str r1, [r0, #EMC_CFG_DIG_DLL]
432 emc_timing_update r1, r0
435 movweq r1, #:lower16:TEGRA_EMC1_BASE
436 movteq r1, #:upper16:TEGRA_EMC1_BASE
437 cmpeq r0, r1
439 ldr r1, [r0, #EMC_AUTO_CAL_CONFIG]
440 orr r1, r1, #(1 << 31) @ set AUTO_CAL_ACTIVE
441 orreq r1, r1, #(1 << 27) @ set slave mode for channel 1
442 str r1, [r0, #EMC_AUTO_CAL_CONFIG]
445 ldr r1, [r0, #EMC_AUTO_CAL_STATUS]
446 tst r1, #(1 << 31) @ wait until AUTO_CAL_ACTIVE is cleared
449 ldr r1, [r0, #EMC_CFG]
450 bic r1, r1, #(1 << 31) @ disable DRAM_CLK_STOP_PD
451 str r1, [r0, #EMC_CFG]
453 mov r1, #0
454 str r1, [r0, #EMC_SELF_REF] @ take DRAM out of self refresh
455 mov r1, #1
457 streq r1, [r0, #EMC_NOP]
458 streq r1, [r0, #EMC_NOP]
459 streq r1, [r0, #EMC_REFRESH]
461 emc_device_mask r1, r0
465 ands r2, r2, r1
468 lsr r1, r1, #8 @ devSel, bit0:dev0, bit1:dev1
484 tst r1, #2
503 tst r1, #2
514 mov r1, #0 @ unstall all transactions
515 str r1, [r0, #EMC_REQ_CTRL]
516 ldr r1, [r5, #0x4] @ restore EMC_ZCAL_INTERVAL
517 str r1, [r0, #EMC_ZCAL_INTERVAL]
518 ldr r1, [r5, #0x0] @ restore EMC_CFG
519 str r1, [r0, #EMC_CFG]
524 mov32 r1, TEGRA_EMC1_BASE
525 cmp r0, r1
526 movne r0, r1
617 ldr r1, [r7]
618 add r1, r1, #2
619 wait_until r1, r7, r9
631 ldr r1, [r7]
632 add r1, r1, #2
633 wait_until r1, r7, r9
656 pll_iddq_entry r1, r5, CLK_RESET_PLLX_MISC3, CLK_RESET_PLLX_MISC3_IDDQ
674 cpu_id r1
676 cpu_to_csr_reg r2, r1
688 cpu_to_halt_reg r2, r1
731 ldr r1, [r0]
732 str r1, [r8, r9] @ save the content of the addr
750 mov r1, #0
751 str r1, [r0, #EMC_ZCAL_INTERVAL]
752 str r1, [r0, #EMC_AUTO_CAL_INTERVAL]
753 ldr r1, [r0, #EMC_CFG]
754 bic r1, r1, #(1 << 28)
755 bicne r1, r1, #(1 << 29)
756 str r1, [r0, #EMC_CFG] @ disable DYN_SELF_REF
758 emc_timing_update r1, r0
760 ldr r1, [r7]
761 add r1, r1, #5
762 wait_until r1, r7, r2
765 ldr r1, [r0, #EMC_AUTO_CAL_STATUS]
766 tst r1, #(1 << 31) @ wait until AUTO_CAL_ACTIVE is cleared
769 mov r1, #3
770 str r1, [r0, #EMC_REQ_CTRL] @ stall incoming DRAM requests
773 ldr r1, [r0, #EMC_EMC_STATUS]
774 tst r1, #4
777 mov r1, #1
778 str r1, [r0, #EMC_SELF_REF]
780 emc_device_mask r1, r0
784 and r2, r2, r1
785 cmp r2, r1
789 ldr r1, [r0, #EMC_XM2VTTGENPADCTRL]
791 and r1, r1, r2
792 str r1, [r0, #EMC_XM2VTTGENPADCTRL]
793 ldr r1, [r0, #EMC_XM2VTTGENPADCTRL2]
795 orreq r1, r1, #7 @ set E_NO_VTTGEN
796 orrne r1, r1, #0x3f
797 str r1, [r0, #EMC_XM2VTTGENPADCTRL2]
799 emc_timing_update r1, r0
804 mov32 r1, TEGRA_EMC1_BASE
805 cmp r0, r1
806 movne r0, r1
810 ldr r1, [r4, #PMC_CTRL]
811 tst r1, #PMC_CTRL_SIDE_EFFECT_LP0
817 mov32 r1, 0x8EC00000
818 str r1, [r4, #PMC_IO_DPD_REQ]
H A Dsleep-tegra20.S100 mov32 r1, TEGRA_IRAM_RESET_BASE_VIRT
103 strb r12, [r1, r2]
105 cpu_to_halt_reg r1, r0
108 str r2, [r3, r1] @ put flow controller in wait event mode
109 ldr r2, [r3, r1]
112 movw r1, 0x1011
113 mov r1, r1, lsl r0
115 str r1, [r3, #0x340] @ put slave CPU in reset
136 * SCRATCH37 = r1 = !turn (inverted from Peterson's algorithm)
150 add r1, r3, #PMC_SCRATCH37
160 str r12, [r1] @ !turn = cpu
164 ldreq r12, [r1]
190 mov32 r1, TEGRA_IRAM_RESET_BASE_VIRT
193 strb r12, [r1, r2]
204 mov32 r1, TEGRA_IRAM_RESET_BASE_VIRT
207 strb r12, [r1, r2]
218 mov32 r1, TEGRA_IRAM_RESET_BASE_VIRT
220 ldrb r12, [r1, r2]
241 mov32 r1, tegra20_iram_start
242 sub r0, r0, r1
243 mov32 r1, TEGRA_IRAM_LPx_RESUME_AREA
244 add r0, r0, r1
298 mov r1, #0
299 mcr p15, 0, r1, c8, c3, 0 @ invalidate shared TLBs
300 mcr p15, 0, r1, c7, c1, 6 @ invalidate shared BTAC
347 mov r1, #(1 << 28)
348 str r1, [r0, #CLK_RESET_SCLK_BURST]
349 str r1, [r0, #CLK_RESET_CCLK_BURST]
350 mov r1, #0
351 str r1, [r0, #CLK_RESET_CCLK_DIVIDER]
352 str r1, [r0, #CLK_RESET_SCLK_DIVIDER]
354 pll_enable r1, r0, CLK_RESET_PLLM_BASE
355 pll_enable r1, r0, CLK_RESET_PLLP_BASE
356 pll_enable r1, r0, CLK_RESET_PLLC_BASE
366 ldr r1, [r4, r5]
367 str r1, [r7] @ restore the value in pad_save
376 ldr r1, [r7]
377 add r1, r1, #0xff
378 wait_until r1, r7, r9
387 ldr r1, [r0, #EMC_CFG]
388 bic r1, r1, #(1 << 31) @ disable DRAM_CLK_STOP
389 str r1, [r0, #EMC_CFG]
391 mov r1, #0
392 str r1, [r0, #EMC_SELF_REF] @ take DRAM out of self refresh
393 mov r1, #1
394 str r1, [r0, #EMC_NOP]
395 str r1, [r0, #EMC_NOP]
396 str r1, [r0, #EMC_REFRESH]
398 emc_device_mask r1, r0
402 ands r2, r2, r1
405 mov r1, #0 @ unstall all transactions
406 str r1, [r0, #EMC_REQ_CTRL]
444 ldr r1, [r7]
445 add r1, r1, #2
446 wait_until r1, r7, r9
477 cpu_id r1
478 cpu_to_halt_reg r1, r1
479 str r0, [r6, r1]
481 ldr r0, [r6, r1] /* memory barrier */
497 mov32 r1, TEGRA_EMC_BASE @ r1 reserved for emc base addr
500 str r2, [r1, #EMC_REQ_CTRL] @ stall incoming DRAM requests
503 ldr r2, [r1, #EMC_EMC_STATUS]
508 str r2, [r1, #EMC_SELF_REF]
510 emc_device_mask r2, r1
513 ldr r3, [r1, #EMC_EMC_STATUS]
527 ldr r1, [r0]
528 str r1, [r4, r5] @ save the content of the addr
530 ldr r1, [r3, r5]
531 str r1, [r0] @ set the save val to the addr
H A Dreset-handler.S59 cpu_to_csr_reg r1, r0
61 ldr r1, [r2, r1]
63 orr r1, r1, \
67 bic r1, r1, r0
68 str r1, [r2]
77 ldr r1, [r0]
78 orr r1, r1, #1
79 str r1, [r0]
237 mov r1, r0, lsl r10
238 str r1, [r7, #0x340] @ CLK_RST_CPU_CMPLX_SET
245 moveq r1, #FLOW_CTRL_HALT_CPU0_EVENTS
247 movne r1, r10, lsl #3
248 addne r2, r1, #(FLOW_CTRL_CPU1_CSR-8)
249 addne r1, r1, #(FLOW_CTRL_HALT_CPU1_EVENTS-8)
260 str r0, [r6, +r1]
261 ldr r0, [r6, +r1] @ memory barrier
/linux-4.1.27/arch/powerpc/platforms/pseries/
H A DhvCall.S37 std r3,STK_PARAM(R3)(r1); \
38 std r4,STK_PARAM(R4)(r1); \
39 std r5,STK_PARAM(R5)(r1); \
40 std r6,STK_PARAM(R6)(r1); \
41 std r7,STK_PARAM(R7)(r1); \
42 std r8,STK_PARAM(R8)(r1); \
43 std r9,STK_PARAM(R9)(r1); \
44 std r10,STK_PARAM(R10)(r1); \
45 std r0,16(r1); \
46 addi r4,r1,STK_PARAM(FIRST_REG); \
47 stdu r1,-STACK_FRAME_OVERHEAD(r1); \
49 ld r3,STACK_FRAME_OVERHEAD+STK_PARAM(R3)(r1); \
50 ld r4,STACK_FRAME_OVERHEAD+STK_PARAM(R4)(r1); \
51 ld r5,STACK_FRAME_OVERHEAD+STK_PARAM(R5)(r1); \
52 ld r6,STACK_FRAME_OVERHEAD+STK_PARAM(R6)(r1); \
53 ld r7,STACK_FRAME_OVERHEAD+STK_PARAM(R7)(r1); \
54 ld r8,STACK_FRAME_OVERHEAD+STK_PARAM(R8)(r1); \
55 ld r9,STACK_FRAME_OVERHEAD+STK_PARAM(R9)(r1); \
56 ld r10,STACK_FRAME_OVERHEAD+STK_PARAM(R10)(r1)
63 ld r0,STACK_FRAME_OVERHEAD+STK_PARAM(R3)(r1); \
64 std r3,STACK_FRAME_OVERHEAD+STK_PARAM(R3)(r1); \
68 ld r0,STACK_FRAME_OVERHEAD+16(r1); \
69 addi r1,r1,STACK_FRAME_OVERHEAD; \
70 ld r3,STK_PARAM(R3)(r1); \
95 std r12,32(r1); \
112 stw r0,8(r1)
116 lwz r0,8(r1)
125 lwz r0,8(r1)
134 stw r0,8(r1)
138 std r4,STK_PARAM(R4)(r1) /* Save ret buffer */
149 ld r12,STK_PARAM(R4)(r1)
155 lwz r0,8(r1)
164 std r4,STK_PARAM(R4)(r1)
176 ld r12,STK_PARAM(R4)(r1)
184 lwz r0,8(r1)
200 stw r0,8(r1)
202 std r4,STK_PARAM(R4)(r1) /* Save ret buffer */
213 ld r12,STK_PARAM(R4)(r1)
219 lwz r0,8(r1)
228 stw r0,8(r1)
232 std r4,STK_PARAM(R4)(r1) /* Save ret buffer */
240 ld r10,STK_PARAM(R11)(r1) /* put arg7 in R10 */
241 ld r11,STK_PARAM(R12)(r1) /* put arg8 in R11 */
242 ld r12,STK_PARAM(R13)(r1) /* put arg9 in R12 */
247 ld r12,STK_PARAM(R4)(r1)
258 lwz r0,8(r1)
267 std r4,STK_PARAM(R4)(r1)
276 ld r10,STACK_FRAME_OVERHEAD+STK_PARAM(R11)(r1)
277 ld r11,STACK_FRAME_OVERHEAD+STK_PARAM(R12)(r1)
278 ld r12,STACK_FRAME_OVERHEAD+STK_PARAM(R13)(r1)
283 ld r12,STACK_FRAME_OVERHEAD+STK_PARAM(R4)(r1)
296 lwz r0,8(r1)
307 stw r0,8(r1)
309 std r4,STK_PARAM(R4)(r1) /* Save ret buffer */
317 ld r10,STK_PARAM(R11)(r1) /* put arg7 in R10 */
318 ld r11,STK_PARAM(R12)(r1) /* put arg8 in R11 */
319 ld r12,STK_PARAM(R13)(r1) /* put arg9 in R12 */
324 ld r12,STK_PARAM(R4)(r1)
335 lwz r0,8(r1)
/linux-4.1.27/arch/sh/lib64/
H A Dudivdi3.S10 sub r21,r5,r1
11 mmulfx.w r1,r1,r4
12 mshflo.w r1,r63,r1
17 msub.w r1,r4,r1
18 madd.w r1,r1,r1
19 mmulfx.w r1,r1,r4
25 msub.w r1,r4,r1
27 mulu.l r1,r7,r4
28 addi r1,-3,r5
30 sub r63,r4,r4 // Negate to make sure r1 ends up <= 1/r2
33 muls.l r1,r4,r4 /* leaving at least one sign bit. */
35 mshalds.l r1,r21,r1
38 add r1,r4,r1 // 31 bit unsigned reciprocal now in r1 (msb equiv. 0.5)
40 /* Can do second step of 64 : 32 div now, using r1 and the rest in r2. */
43 mulu.l r21,r1,r21
56 mulu.l r2,r1,r7
72 msub.w r1,r4,r1
74 mulu.l r1,r7,r4
75 addi r1,-3,r5
77 sub r63,r4,r4 // Negate to make sure r1 ends up <= 1/r2
80 muls.l r1,r4,r4 /* leaving at least one sign bit. */
83 mshalds.l r1,r21,r1
85 add r1,r4,r1 // 31 bit unsigned reciprocal now in r1 (msb equiv. 0.5)
87 /* Can do second step of 64 : 32 div now, using r1 and the rest in r25. */
90 mulu.l r21,r1,r21
H A Dsdivsi3.S8 /* clobbered: r1,r18,r19,r20,r21,r25,tr0 */
16 nsb r5, r1
17 shlld r5, r1, r25 /* normalize; [-2 ..1, 1..2) in s2.62 */
29 sub r63, r1, r1
30 addi r1, 92, r1
48 shard r21, r1, r21
H A Dstrlen.S25 ld.b r2, 0, r1
28 bnei/l r1, 0, tr0
/linux-4.1.27/arch/powerpc/lib/
H A Dldstfp.S80 PPC_STLU r1,-STKFRM(r1)
82 PPC_STL r0,STKFRM+PPC_LR_STKOFF(r1)
89 stfd fr0,STKFRM-16(r1)
95 lfd fr0,STKFRM-16(r1)
96 4: PPC_LL r0,STKFRM+PPC_LR_STKOFF(r1)
101 addi r1,r1,STKFRM
107 PPC_STLU r1,-STKFRM(r1)
109 PPC_STL r0,STKFRM+PPC_LR_STKOFF(r1)
116 stfd fr0,STKFRM-16(r1)
122 lfd fr0,STKFRM-16(r1)
123 4: PPC_LL r0,STKFRM+PPC_LR_STKOFF(r1)
128 addi r1,r1,STKFRM
134 PPC_STLU r1,-STKFRM(r1)
136 PPC_STL r0,STKFRM+PPC_LR_STKOFF(r1)
143 stfd fr0,STKFRM-16(r1)
149 lfd fr0,STKFRM-16(r1)
150 4: PPC_LL r0,STKFRM+PPC_LR_STKOFF(r1)
155 addi r1,r1,STKFRM
161 PPC_STLU r1,-STKFRM(r1)
163 PPC_STL r0,STKFRM+PPC_LR_STKOFF(r1)
170 stfd fr0,STKFRM-16(r1)
176 lfd fr0,STKFRM-16(r1)
177 4: PPC_LL r0,STKFRM+PPC_LR_STKOFF(r1)
182 addi r1,r1,STKFRM
227 PPC_STLU r1,-STKFRM(r1)
229 PPC_STL r0,STKFRM+PPC_LR_STKOFF(r1)
237 stvx v0,r1,r8
243 lvx v0,r1,r8
244 4: PPC_LL r0,STKFRM+PPC_LR_STKOFF(r1)
249 addi r1,r1,STKFRM
255 PPC_STLU r1,-STKFRM(r1)
257 PPC_STL r0,STKFRM+PPC_LR_STKOFF(r1)
265 stvx v0,r1,r8
271 lvx v0,r1,r8
272 4: PPC_LL r0,STKFRM+PPC_LR_STKOFF(r1)
277 addi r1,r1,STKFRM
323 PPC_STLU r1,-STKFRM(r1)
325 PPC_STL r0,STKFRM+PPC_LR_STKOFF(r1)
340 4: PPC_LL r0,STKFRM+PPC_LR_STKOFF(r1)
345 addi r1,r1,STKFRM
351 PPC_STLU r1,-STKFRM(r1)
353 PPC_STL r0,STKFRM+PPC_LR_STKOFF(r1)
368 4: PPC_LL r0,STKFRM+PPC_LR_STKOFF(r1)
373 addi r1,r1,STKFRM
H A Dcopyuser_power7.S65 ld r16,STK_REG(R16)(r1)
66 ld r15,STK_REG(R15)(r1)
67 ld r14,STK_REG(R14)(r1)
70 ld r0,STACKFRAMESIZE+16(r1)
76 ld r22,STK_REG(R22)(r1)
77 ld r21,STK_REG(R21)(r1)
78 ld r20,STK_REG(R20)(r1)
79 ld r19,STK_REG(R19)(r1)
80 ld r18,STK_REG(R18)(r1)
81 ld r17,STK_REG(R17)(r1)
82 ld r16,STK_REG(R16)(r1)
83 ld r15,STK_REG(R15)(r1)
84 ld r14,STK_REG(R14)(r1)
86 addi r1,r1,STACKFRAMESIZE
88 ld r3,-STACKFRAMESIZE+STK_REG(R31)(r1)
89 ld r4,-STACKFRAMESIZE+STK_REG(R30)(r1)
90 ld r5,-STACKFRAMESIZE+STK_REG(R29)(r1)
99 std r3,-STACKFRAMESIZE+STK_REG(R31)(r1)
100 std r4,-STACKFRAMESIZE+STK_REG(R30)(r1)
101 std r5,-STACKFRAMESIZE+STK_REG(R29)(r1)
108 std r3,-STACKFRAMESIZE+STK_REG(R31)(r1)
109 std r4,-STACKFRAMESIZE+STK_REG(R30)(r1)
110 std r5,-STACKFRAMESIZE+STK_REG(R29)(r1)
144 stdu r1,-STACKFRAMESIZE(r1)
145 std r14,STK_REG(R14)(r1)
146 std r15,STK_REG(R15)(r1)
147 std r16,STK_REG(R16)(r1)
148 std r17,STK_REG(R17)(r1)
149 std r18,STK_REG(R18)(r1)
150 std r19,STK_REG(R19)(r1)
151 std r20,STK_REG(R20)(r1)
152 std r21,STK_REG(R21)(r1)
153 std r22,STK_REG(R22)(r1)
154 std r0,STACKFRAMESIZE+16(r1)
200 ld r14,STK_REG(R14)(r1)
201 ld r15,STK_REG(R15)(r1)
202 ld r16,STK_REG(R16)(r1)
203 ld r17,STK_REG(R17)(r1)
204 ld r18,STK_REG(R18)(r1)
205 ld r19,STK_REG(R19)(r1)
206 ld r20,STK_REG(R20)(r1)
207 ld r21,STK_REG(R21)(r1)
208 ld r22,STK_REG(R22)(r1)
209 addi r1,r1,STACKFRAMESIZE
290 addi r1,r1,STACKFRAMESIZE
296 std r0,16(r1)
297 stdu r1,-STACKFRAMESIZE(r1)
300 ld r0,STACKFRAMESIZE+16(r1)
301 ld r3,STK_REG(R31)(r1)
302 ld r4,STK_REG(R30)(r1)
303 ld r5,STK_REG(R29)(r1)
419 std r14,STK_REG(R14)(r1)
420 std r15,STK_REG(R15)(r1)
421 std r16,STK_REG(R16)(r1)
456 ld r14,STK_REG(R14)(r1)
457 ld r15,STK_REG(R15)(r1)
458 ld r16,STK_REG(R16)(r1)
516 15: addi r1,r1,STACKFRAMESIZE
604 std r14,STK_REG(R14)(r1)
605 std r15,STK_REG(R15)(r1)
606 std r16,STK_REG(R16)(r1)
649 ld r14,STK_REG(R14)(r1)
650 ld r15,STK_REG(R15)(r1)
651 ld r16,STK_REG(R16)(r1)
719 15: addi r1,r1,STACKFRAMESIZE
H A Dmemcmp_64.S89 std r31,-8(r1)
90 std r30,-16(r1)
91 std r29,-24(r1)
92 std r28,-32(r1)
93 std r27,-40(r1)
182 ld r31,-8(r1)
183 ld r30,-16(r1)
184 ld r29,-24(r1)
185 ld r28,-32(r1)
186 ld r27,-40(r1)
228 ld r31,-8(r1)
229 ld r30,-16(r1)
230 ld r29,-24(r1)
231 ld r28,-32(r1)
232 ld r27,-40(r1)
H A Dcrtsavres.S321 std r14,-144(r1)
324 std r15,-136(r1)
327 std r16,-128(r1)
330 std r17,-120(r1)
333 std r18,-112(r1)
336 std r19,-104(r1)
339 std r20,-96(r1)
342 std r21,-88(r1)
345 std r22,-80(r1)
348 std r23,-72(r1)
351 std r24,-64(r1)
354 std r25,-56(r1)
357 std r26,-48(r1)
360 std r27,-40(r1)
363 std r28,-32(r1)
366 std r29,-24(r1)
369 std r30,-16(r1)
372 std r31,-8(r1)
373 std r0,16(r1)
378 ld r14,-144(r1)
381 ld r15,-136(r1)
384 ld r16,-128(r1)
387 ld r17,-120(r1)
390 ld r18,-112(r1)
393 ld r19,-104(r1)
396 ld r20,-96(r1)
399 ld r21,-88(r1)
402 ld r22,-80(r1)
405 ld r23,-72(r1)
408 ld r24,-64(r1)
411 ld r25,-56(r1)
414 ld r26,-48(r1)
417 ld r27,-40(r1)
420 ld r28,-32(r1)
423 ld r0,16(r1)
424 ld r29,-24(r1)
426 ld r30,-16(r1)
427 ld r31,-8(r1)
432 ld r30,-16(r1)
435 ld r0,16(r1)
436 ld r31,-8(r1)
H A Dmemcpy_power7.S36 std r3,-STACKFRAMESIZE+STK_REG(R31)(r1)
43 std r3,-STACKFRAMESIZE+STK_REG(R31)(r1)
77 stdu r1,-STACKFRAMESIZE(r1)
78 std r14,STK_REG(R14)(r1)
79 std r15,STK_REG(R15)(r1)
80 std r16,STK_REG(R16)(r1)
81 std r17,STK_REG(R17)(r1)
82 std r18,STK_REG(R18)(r1)
83 std r19,STK_REG(R19)(r1)
84 std r20,STK_REG(R20)(r1)
85 std r21,STK_REG(R21)(r1)
86 std r22,STK_REG(R22)(r1)
87 std r0,STACKFRAMESIZE+16(r1)
133 ld r14,STK_REG(R14)(r1)
134 ld r15,STK_REG(R15)(r1)
135 ld r16,STK_REG(R16)(r1)
136 ld r17,STK_REG(R17)(r1)
137 ld r18,STK_REG(R18)(r1)
138 ld r19,STK_REG(R19)(r1)
139 ld r20,STK_REG(R20)(r1)
140 ld r21,STK_REG(R21)(r1)
141 ld r22,STK_REG(R22)(r1)
142 addi r1,r1,STACKFRAMESIZE
219 15: ld r3,-STACKFRAMESIZE+STK_REG(R31)(r1)
223 addi r1,r1,STACKFRAMESIZE
229 std r4,-STACKFRAMESIZE+STK_REG(R30)(r1)
230 std r5,-STACKFRAMESIZE+STK_REG(R29)(r1)
231 std r0,16(r1)
232 stdu r1,-STACKFRAMESIZE(r1)
235 ld r0,STACKFRAMESIZE+16(r1)
236 ld r3,STK_REG(R31)(r1)
237 ld r4,STK_REG(R30)(r1)
238 ld r5,STK_REG(R29)(r1)
352 std r14,STK_REG(R14)(r1)
353 std r15,STK_REG(R15)(r1)
354 std r16,STK_REG(R16)(r1)
389 ld r14,STK_REG(R14)(r1)
390 ld r15,STK_REG(R15)(r1)
391 ld r16,STK_REG(R16)(r1)
449 15: addi r1,r1,STACKFRAMESIZE
450 ld r3,-STACKFRAMESIZE+STK_REG(R31)(r1)
538 std r14,STK_REG(R14)(r1)
539 std r15,STK_REG(R15)(r1)
540 std r16,STK_REG(R16)(r1)
583 ld r14,STK_REG(R14)(r1)
584 ld r15,STK_REG(R15)(r1)
585 ld r16,STK_REG(R16)(r1)
653 15: addi r1,r1,STACKFRAMESIZE
654 ld r3,-STACKFRAMESIZE+STK_REG(R31)(r1)
/linux-4.1.27/arch/openrisc/kernel/
H A Dentry.S54 l.lwz r3,PT_PC(r1) ;\
56 l.lwz r3,PT_SR(r1) ;\
58 l.lwz r2,PT_GPR2(r1) ;\
59 l.lwz r3,PT_GPR3(r1) ;\
60 l.lwz r4,PT_GPR4(r1) ;\
61 l.lwz r5,PT_GPR5(r1) ;\
62 l.lwz r6,PT_GPR6(r1) ;\
63 l.lwz r7,PT_GPR7(r1) ;\
64 l.lwz r8,PT_GPR8(r1) ;\
65 l.lwz r9,PT_GPR9(r1) ;\
66 l.lwz r10,PT_GPR10(r1) ;\
67 l.lwz r11,PT_GPR11(r1) ;\
68 l.lwz r12,PT_GPR12(r1) ;\
69 l.lwz r13,PT_GPR13(r1) ;\
70 l.lwz r14,PT_GPR14(r1) ;\
71 l.lwz r15,PT_GPR15(r1) ;\
72 l.lwz r16,PT_GPR16(r1) ;\
73 l.lwz r17,PT_GPR17(r1) ;\
74 l.lwz r18,PT_GPR18(r1) ;\
75 l.lwz r19,PT_GPR19(r1) ;\
76 l.lwz r20,PT_GPR20(r1) ;\
77 l.lwz r21,PT_GPR21(r1) ;\
78 l.lwz r22,PT_GPR22(r1) ;\
79 l.lwz r23,PT_GPR23(r1) ;\
80 l.lwz r24,PT_GPR24(r1) ;\
81 l.lwz r25,PT_GPR25(r1) ;\
82 l.lwz r26,PT_GPR26(r1) ;\
83 l.lwz r27,PT_GPR27(r1) ;\
84 l.lwz r28,PT_GPR28(r1) ;\
85 l.lwz r29,PT_GPR29(r1) ;\
86 l.lwz r30,PT_GPR30(r1) ;\
87 l.lwz r31,PT_GPR31(r1) ;\
88 l.lwz r1,PT_SP(r1) ;\
95 /* r1, EPCR, ESR a already saved */ ;\
96 l.sw PT_GPR2(r1),r2 ;\
97 l.sw PT_GPR3(r1),r3 ;\
99 l.sw PT_GPR5(r1),r5 ;\
100 l.sw PT_GPR6(r1),r6 ;\
101 l.sw PT_GPR7(r1),r7 ;\
102 l.sw PT_GPR8(r1),r8 ;\
103 l.sw PT_GPR9(r1),r9 ;\
105 l.sw PT_GPR11(r1),r11 ;\
107 l.sw PT_GPR13(r1),r13 ;\
108 l.sw PT_GPR14(r1),r14 ;\
109 l.sw PT_GPR15(r1),r15 ;\
110 l.sw PT_GPR16(r1),r16 ;\
111 l.sw PT_GPR17(r1),r17 ;\
112 l.sw PT_GPR18(r1),r18 ;\
113 l.sw PT_GPR19(r1),r19 ;\
114 l.sw PT_GPR20(r1),r20 ;\
115 l.sw PT_GPR21(r1),r21 ;\
116 l.sw PT_GPR22(r1),r22 ;\
117 l.sw PT_GPR23(r1),r23 ;\
118 l.sw PT_GPR24(r1),r24 ;\
119 l.sw PT_GPR25(r1),r25 ;\
120 l.sw PT_GPR26(r1),r26 ;\
121 l.sw PT_GPR27(r1),r27 ;\
122 l.sw PT_GPR28(r1),r28 ;\
123 l.sw PT_GPR29(r1),r29 ;\
125 /* l.sw PT_GPR30(r1),r30*/ ;\
126 l.sw PT_GPR31(r1),r31 ;\
129 l.sw PT_ORIG_GPR11(r1),r30
134 /* r1, EPCR, ESR already saved */ ;\
135 l.sw PT_GPR2(r1),r2 ;\
136 l.sw PT_GPR3(r1),r3 ;\
137 l.sw PT_GPR5(r1),r5 ;\
138 l.sw PT_GPR6(r1),r6 ;\
139 l.sw PT_GPR7(r1),r7 ;\
140 l.sw PT_GPR8(r1),r8 ;\
141 l.sw PT_GPR9(r1),r9 ;\
143 l.sw PT_GPR11(r1),r11 ;\
145 l.sw PT_GPR13(r1),r13 ;\
146 l.sw PT_GPR14(r1),r14 ;\
147 l.sw PT_GPR15(r1),r15 ;\
148 l.sw PT_GPR16(r1),r16 ;\
149 l.sw PT_GPR17(r1),r17 ;\
150 l.sw PT_GPR18(r1),r18 ;\
151 l.sw PT_GPR19(r1),r19 ;\
152 l.sw PT_GPR20(r1),r20 ;\
153 l.sw PT_GPR21(r1),r21 ;\
154 l.sw PT_GPR22(r1),r22 ;\
155 l.sw PT_GPR23(r1),r23 ;\
156 l.sw PT_GPR24(r1),r24 ;\
157 l.sw PT_GPR25(r1),r25 ;\
158 l.sw PT_GPR26(r1),r26 ;\
159 l.sw PT_GPR27(r1),r27 ;\
160 l.sw PT_GPR28(r1),r28 ;\
161 l.sw PT_GPR29(r1),r29 ;\
163 l.sw PT_GPR30(r1),r30 ;\
164 /* l.sw PT_GPR31(r1),r31 */ ;\
167 l.sw PT_ORIG_GPR11(r1),r30 ;\
168 l.addi r3,r1,0 ;\
198 l.addi r3,r1,0 /* pt_regs */
213 l.addi r3,r1,0 // pt_regs
293 l.addi r3,r1,0 // pt_regs
308 l.addi r3,r1,0 /* pt_regs */
318 l.addi r3,r1,0 /* pt_regs */
328 l.lwz r5,PT_PC(r1)
366 l.add r4,r4,r1 /* Load the jump register value from the stack */
374 l.sw PT_PC(r1),r5
399 l.add r4,r4,r1
409 l.add r4,r4,r1
425 l.add r4,r4,r1
441 l.add r4,r4,r1
448 l.add r4,r4,r1
458 l.add r4,r4,r1
479 l.addi r3,r1,0 /* pt_regs */
488 l.lwz r4,PT_SR(r1) // were interrupts enabled ?
494 l.addi r1,r1,-0x8
497 l.sw 0x0(r1),r3
499 l.sw 0x4(r1),r4
500 l.addi r1,r1,0x8
509 // l.sw PT_SR(r1),r4
512 l.addi r3,r1,0
552 /* r1, EPCR, ESR a already saved */
553 l.sw PT_GPR2(r1),r2
558 l.sw PT_GPR3(r1),r3
561 l.lwz r4,PT_GPR4(r1)
562 l.sw PT_GPR5(r1),r5
563 l.sw PT_GPR6(r1),r6
564 l.sw PT_GPR7(r1),r7
565 l.sw PT_GPR8(r1),r8
566 l.sw PT_GPR9(r1),r9
568 l.sw PT_GPR11(r1),r11
570 l.sw PT_ORIG_GPR11(r1),r11
582 /* l.sw PT_GPR30(r1),r30 */
612 l.sw PT_GPR11(r1),r11 // save return value
619 l.sw -4(r1),r27
620 l.sw -8(r1),r11
621 l.addi r1,r1,-8
626 l.addi r1,r1,8
656 l.sw PT_GPR14(r1),r14
657 l.sw PT_GPR16(r1),r16
658 l.sw PT_GPR18(r1),r18
659 l.sw PT_GPR20(r1),r20
660 l.sw PT_GPR22(r1),r22
661 l.sw PT_GPR24(r1),r24
662 l.sw PT_GPR26(r1),r26
663 l.sw PT_GPR28(r1),r28
693 l.lwz r2,PT_GPR2(r1)
699 l.lwz r3,PT_GPR3(r1)
700 l.lwz r4,PT_GPR4(r1)
701 l.lwz r5,PT_GPR5(r1)
702 l.lwz r6,PT_GPR6(r1)
703 l.lwz r7,PT_GPR7(r1)
704 l.lwz r8,PT_GPR8(r1)
706 l.lwz r9,PT_GPR9(r1)
707 l.lwz r10,PT_GPR10(r1)
708 l.lwz r11,PT_GPR11(r1)
711 l.lwz r30,PT_GPR30(r1)
714 l.lwz r13,PT_PC(r1)
715 l.lwz r15,PT_SR(r1)
716 l.lwz r1,PT_SP(r1)
738 l.addi r3,r1,0
744 l.lwz r11,PT_GPR11(r1)
745 l.lwz r3,PT_GPR3(r1)
746 l.lwz r4,PT_GPR4(r1)
747 l.lwz r5,PT_GPR5(r1)
748 l.lwz r6,PT_GPR6(r1)
749 l.lwz r7,PT_GPR7(r1)
752 l.lwz r8,PT_GPR8(r1)
756 l.addi r3,r1,0
781 l.addi r3,r1,0 /* pt_regs */
865 l.lwz r5,PT_ORIG_GPR11(r1)
872 l.ori r3,r1,0 /* pt_regs */
885 l.lwz r11,PT_ORIG_GPR11(r1)
887 l.lwz r3,PT_GPR3(r1)
888 l.lwz r4,PT_GPR4(r1)
889 l.lwz r5,PT_GPR5(r1)
890 l.lwz r6,PT_GPR6(r1)
891 l.lwz r7,PT_GPR7(r1)
893 l.lwz r8,PT_GPR8(r1)
902 l.lwz r4,PT_SR(r1)
925 l.lwz r11,PT_GPR11(r1)
932 l.lwz r12,PT_GPR12(r1)
933 l.lwz r14,PT_GPR14(r1)
934 l.lwz r16,PT_GPR16(r1)
935 l.lwz r18,PT_GPR18(r1)
936 l.lwz r20,PT_GPR20(r1)
937 l.lwz r22,PT_GPR22(r1)
938 l.lwz r24,PT_GPR24(r1)
939 l.lwz r26,PT_GPR26(r1)
940 l.lwz r28,PT_GPR28(r1)
990 l.addi r1,r1,-(INT_FRAME_SIZE)
992 /* No need to store r1/PT_SP as it goes into KSP below */
993 l.sw PT_GPR2(r1),r2
994 l.sw PT_GPR9(r1),r9
997 l.sw PT_GPR12(r1),r12
998 l.sw PT_GPR14(r1),r14
999 l.sw PT_GPR16(r1),r16
1000 l.sw PT_GPR18(r1),r18
1001 l.sw PT_GPR20(r1),r20
1002 l.sw PT_GPR22(r1),r22
1003 l.sw PT_GPR24(r1),r24
1004 l.sw PT_GPR26(r1),r26
1005 l.sw PT_GPR28(r1),r28
1006 l.sw PT_GPR30(r1),r30
1019 l.sw PT_SP(r1),r29
1022 l.sw TI_KSP(r10),r1 /* Save old stack pointer */
1024 l.lwz r1,TI_KSP(r10) /* Load new stack pointer */
1027 l.lwz r29,PT_SP(r1)
1033 l.lwz r2,PT_GPR2(r1)
1034 l.lwz r9,PT_GPR9(r1)
1040 l.lwz r12,PT_GPR12(r1)
1041 l.lwz r14,PT_GPR14(r1)
1042 l.lwz r16,PT_GPR16(r1)
1043 l.lwz r18,PT_GPR18(r1)
1044 l.lwz r20,PT_GPR20(r1)
1045 l.lwz r22,PT_GPR22(r1)
1046 l.lwz r24,PT_GPR24(r1)
1047 l.lwz r26,PT_GPR26(r1)
1048 l.lwz r28,PT_GPR28(r1)
1049 l.lwz r30,PT_GPR30(r1)
1052 l.addi r1,r1,(INT_FRAME_SIZE)
1081 l.sw PT_GPR14(r1),r14
1082 l.sw PT_GPR16(r1),r16
1083 l.sw PT_GPR18(r1),r18
1084 l.sw PT_GPR20(r1),r20
1085 l.sw PT_GPR22(r1),r22
1086 l.sw PT_GPR24(r1),r24
1087 l.sw PT_GPR26(r1),r26
1089 l.sw PT_GPR28(r1),r28
1095 l.addi r7,r1,0
1101 l.addi r3,r1,0
1105 l.addi r3,r1,0
/linux-4.1.27/arch/s390/kernel/
H A Dswsusp.S29 lgr %r1,%r15
31 stg %r1,__SF_BACKCHAIN(%r15)
43 lghi %r1,0x1000
49 mvc 0x318(4,%r1),__SF_EMPTY(%r15) /* move prefix to lowcore */
50 stfpc 0x31c(%r1) /* store fpu control */
51 std 0,0x200(%r1) /* store f0 */
52 std 1,0x208(%r1) /* store f1 */
53 std 2,0x210(%r1) /* store f2 */
54 std 3,0x218(%r1) /* store f3 */
55 std 4,0x220(%r1) /* store f4 */
56 std 5,0x228(%r1) /* store f5 */
57 std 6,0x230(%r1) /* store f6 */
58 std 7,0x238(%r1) /* store f7 */
59 std 8,0x240(%r1) /* store f8 */
60 std 9,0x248(%r1) /* store f9 */
61 std 10,0x250(%r1) /* store f10 */
62 std 11,0x258(%r1) /* store f11 */
63 std 12,0x260(%r1) /* store f12 */
64 std 13,0x268(%r1) /* store f13 */
65 std 14,0x270(%r1) /* store f14 */
66 std 15,0x278(%r1) /* store f15 */
67 stam %a0,%a15,0x340(%r1) /* store access registers */
68 stctg %c0,%c15,0x380(%r1) /* store control registers */
69 stmg %r0,%r15,0x280(%r1) /* store general registers */
71 stpt 0x328(%r1) /* store timer */
73 stckc 0x330(%r1) /* store clock comparator */
77 slg %r0,0x328(%r1)
80 mvc __LC_LAST_UPDATE_TIMER(8),0x328(%r1)
115 lghi %r1,0x1000
116 spx 0x318(%r1)
128 lgr %r1,%r15
130 stg %r1,__SF_BACKCHAIN(%r15)
144 larl %r1,restore_pblist
145 lg %r1,0(%r1)
146 ltgr %r1,%r1
149 lg %r2,8(%r1)
150 lg %r4,0(%r1)
157 lg %r2,8(%r1)
159 lg %r1,16(%r1)
160 ltgr %r1,%r1
166 larl %r1,restart_entry
168 og %r1,0(%r2)
169 stg %r1,0(%r0)
170 larl %r1,.Lnew_pgm_check_psw
172 stm %r2,%r3,0(%r1)
173 mvc __LC_PGM_NEW_PSW(16,%r0),0(%r1)
177 lhi %r1,1
178 sigp %r1,%r0,SIGP_SET_ARCHITECTURE
181 larl %r1,smp_cpu_mt_shift
182 icm %r1,15,0(%r1)
184 llgfr %r1,%r1
186 sigp %r1,%r0,SIGP_SET_MULTI_THREADING
191 larl %r1,.Lnew_pgm_check_psw
192 lpswe 0(%r1)
196 larl %r1,.Lresume_cpu /* Resume CPU address: r2 */
197 stap 0(%r1)
198 llgh %r2,0(%r1)
199 llgh %r1,__LC_EXT_CPU_ADDR(%r0) /* Suspend CPU address: r1 */
200 cgr %r1,%r2
201 je restore_registers /* r1 = r2 -> nothing to do */
205 sigp %r9,%r1,SIGP_INITIAL_CPU_RESET /* sigp initial cpu reset */
214 lghi %r1,0
216 sigp %r1,%r0,SIGP_SET_ARCHITECTURE
222 sigp %r9,%r1,SIGP_RESTART /* sigp restart to suspend CPU */
230 larl %r1,.Lresume_cpu
231 llgh %r2,0(%r1)
241 lghi %r13,0x1000 /* %r1 = pointer to save area */
244 llgf %r1,0x318(%r13)
245 stck __LC_LAST_UPDATE_CLOCK(%r1)
H A Dbase.S18 larl %r1,s390_base_mcck_handler_fn
19 lg %r1,0(%r1)
20 ltgr %r1,%r1
22 basr %r14,%r1
23 1: la %r1,4095
24 lmg %r0,%r15,__LC_GPREGS_SAVE_AREA-4095(%r1)
38 larl %r1,s390_base_ext_handler_fn
39 lg %r1,0(%r1)
40 ltgr %r1,%r1
42 basr %r14,%r1
58 larl %r1,s390_base_pgm_handler_fn
59 lg %r1,0(%r1)
60 ltgr %r1,%r1
62 basr %r14,%r1
97 lghi %r1,1
99 diag %r0,%r1,0x308
102 lhi %r1,2 # Use mode 2 = ESAME (dump)
103 sigp %r1,%r0,SIGP_SET_ARCHITECTURE # Switch to ESAME mode
H A Dmcount.S29 lgr %r1,%r15
34 stg %r1,__SF_BACKCHAIN(%r15)
35 stg %r1,(STACK_PTREGS_GPRS+15*8)(%r15)
41 lgrl %r1,ftrace_trace_function
47 larl %r1,ftrace_trace_function
48 lg %r1,0(%r1)
52 basr %r14,%r1
65 lg %r1,(STACK_PTREGS_PSW+8)(%r15)
67 br %r1
73 lgr %r1,%r15
75 stg %r1,__SF_BACKCHAIN(%r15)
H A Drelocate_kernel.S38 la %r1,load_psw-.base(%r13)
39 mvc 0(8,%r0),0(%r1)
46 lhi %r1,1 # mode 1 = esame
47 sigp %r1,%r0,SIGP_SET_ARCHITECTURE # switch to esame mode
96 sr %r1,%r1 # erase register r1
98 sigp %r1,%r2,SIGP_SET_ARCHITECTURE # set cpuid to zero
/linux-4.1.27/arch/unicore32/mm/
H A Dcache-ucv2.S58 sub r1, r1, r0
59 csub.a r1, #MAX_AREA_SIZE
62 andn r1, r1, #CACHE_LINESIZE - 1
63 add r1, r1, #CACHE_LINESIZE
68 sub.a r1, r1, #CACHE_LINESIZE
98 sub r1, r1, r0
99 csub.a r1, #MAX_AREA_SIZE
102 andn r1, r1, #CACHE_LINESIZE - 1
103 add r1, r1, #CACHE_LINESIZE
118 sub.a r1, r1, #CACHE_LINESIZE
152 sub r1, r1, r0
153 andn r1, r1, #CACHE_LINESIZE - 1
154 add r1, r1, #CACHE_LINESIZE
156 csub.a r1, #MAX_AREA_SIZE
171 sub.a r1, r1, #CACHE_LINESIZE
192 sub r1, r1, r0
193 andn r1, r1, #CACHE_LINESIZE - 1
194 add r1, r1, #CACHE_LINESIZE
196 csub.a r1, #MAX_AREA_SIZE
203 sub.a r1, r1, #CACHE_LINESIZE
/linux-4.1.27/arch/sh/boot/compressed/
H A Dhead_32.S15 mov.l init_sr, r1
16 ldc r1, sr
25 mov #0xffffffe0, r1
26 and r1, r0 ! align cache line
28 mov r0, r1
29 sub r2, r1
31 mov.l @r1, r4
32 mov.l @(4,r1), r5
33 mov.l @(8,r1), r6
34 mov.l @(12,r1), r7
35 mov.l @(16,r1), r8
36 mov.l @(20,r1), r9
37 mov.l @(24,r1), r10
38 mov.l @(28,r1), r11
53 add #-32, r1
66 mov.l end_addr, r1
70 mov.l r0, @-r1
71 cmp/eq r1,r2
/linux-4.1.27/arch/openrisc/lib/
H A Dstring.S37 l.addi r1,r1,-12
38 l.sw 0(r1),r6
39 l.sw 4(r1),r4
40 l.sw 8(r1),r3
54 l.lwz r6,0(r1)
55 l.lwz r4,4(r1)
56 l.lwz r3,8(r1)
58 l.addi r1,r1,12
78 l.addi r1,r1,-8
79 l.sw 0(r1),r4
80 l.sw 4(r1),r3
92 l.lwz r4,0(r1)
93 l.lwz r3,4(r1)
95 l.addi r1,r1,8
/linux-4.1.27/arch/arm/mm/
H A Dcopypage-xsc3.c25 * r1 = source
39 pld [r1, #0] \n\ xsc3_mc_copy_user_page()
40 pld [r1, #32] \n\ xsc3_mc_copy_user_page()
41 1: pld [r1, #64] \n\ xsc3_mc_copy_user_page()
42 pld [r1, #96] \n\ xsc3_mc_copy_user_page()
44 2: ldrd r2, [r1], #8 \n\ xsc3_mc_copy_user_page()
46 ldrd r4, [r1], #8 \n\ xsc3_mc_copy_user_page()
49 ldrd r2, [r1], #8 \n\ xsc3_mc_copy_user_page()
51 ldrd r4, [r1], #8 \n\ xsc3_mc_copy_user_page()
54 ldrd r2, [r1], #8 \n\ xsc3_mc_copy_user_page()
56 ldrd r4, [r1], #8 \n\ xsc3_mc_copy_user_page()
59 ldrd r2, [r1], #8 \n\ xsc3_mc_copy_user_page()
62 ldrd r4, [r1], #8 \n\ xsc3_mc_copy_user_page()
89 * r1 = virtual user address of ultimate destination page
95 mov r1, %2 \n\ xsc3_mc_clear_user_highpage()
103 subs r1, r1, #1 \n\ xsc3_mc_clear_user_highpage()
107 : "r1", "r2", "r3"); xsc3_mc_clear_user_highpage()
H A Dcopypage-xscale.c51 pld [r1, #0] \n\ mc_copy_user_page()
52 pld [r1, #32] \n\ mc_copy_user_page()
55 pld [r1, #64] \n\ mc_copy_user_page()
56 pld [r1, #96] \n\ mc_copy_user_page()
59 mov ip, r1 \n\ mc_copy_user_page()
60 strd r2, [r1], #8 \n\ mc_copy_user_page()
62 strd r4, [r1], #8 \n\ mc_copy_user_page()
64 strd r2, [r1], #8 \n\ mc_copy_user_page()
65 strd r4, [r1], #8 \n\ mc_copy_user_page()
70 mov ip, r1 \n\ mc_copy_user_page()
71 strd r2, [r1], #8 \n\ mc_copy_user_page()
73 strd r4, [r1], #8 \n\ mc_copy_user_page()
75 strd r2, [r1], #8 \n\ mc_copy_user_page()
76 strd r4, [r1], #8 \n\ mc_copy_user_page()
114 "mov r1, %2 \n\ xscale_mc_clear_user_highpage()
123 subs r1, r1, #1 \n\ xscale_mc_clear_user_highpage()
128 : "r1", "r2", "r3", "ip"); xscale_mc_clear_user_highpage()
H A Dcopypage-feroceon.c22 1: mov lr, r1 \n\ feroceon_copy_user_page()
23 ldmia r1!, {r2 - r9} \n\ feroceon_copy_user_page()
32 ldmia r1!, {r2 - r9} \n\ feroceon_copy_user_page()
36 ldmia r1!, {r2 - r9} \n\ feroceon_copy_user_page()
40 ldmia r1!, {r2 - r9} \n\ feroceon_copy_user_page()
44 ldmia r1!, {r2 - r9} \n\ feroceon_copy_user_page()
48 ldmia r1!, {r2 - r9} \n\ feroceon_copy_user_page()
52 ldmia r1!, {r2 - r9} \n\ feroceon_copy_user_page()
56 ldmia r1!, {r2 - r9} \n\ feroceon_copy_user_page()
87 mov r1, %2 \n\ feroceon_clear_user_highpage()
97 subs r1, r1, #1 \n\ feroceon_clear_user_highpage()
101 mcr p15, 0, r1, c7, c10, 4 @ drain WB" feroceon_clear_user_highpage()
104 : "r1", "r2", "r3", "r4", "r5", "r6", "r7", "ip", "lr"); feroceon_clear_user_highpage()
H A Dcopypage-v4wt.c29 ldmia r1!, {r3, r4, ip, lr} @ 4\n\ v4wt_copy_user_page()
31 ldmia r1!, {r3, r4, ip, lr} @ 4+1\n\ v4wt_copy_user_page()
33 ldmia r1!, {r3, r4, ip, lr} @ 4\n\ v4wt_copy_user_page()
35 ldmia r1!, {r3, r4, ip, lr} @ 4\n\ v4wt_copy_user_page()
38 ldmneia r1!, {r3, r4, ip, lr} @ 4\n\ v4wt_copy_user_page()
67 mov r1, %2 @ 1\n\ v4wt_clear_user_highpage()
76 subs r1, r1, #1 @ 1\n\ v4wt_clear_user_highpage()
81 : "r1", "r2", "r3", "ip", "lr"); v4wt_clear_user_highpage()
H A Dproc-v7-2level.S49 mmid r1, r1 @ get mm->context.id
55 bfi r1, r2, #8, #24 @ insert into new context ID
60 mcr p15, 0, r1, c13, c0, 1 @ set context ID
81 str r1, [r0] @ linux version
83 bic r3, r1, #0x000003f0
88 tst r1, #1 << 4
91 eor r1, r1, #L_PTE_DIRTY
92 tst r1, #L_PTE_RDONLY | L_PTE_DIRTY
95 tst r1, #L_PTE_USER
98 tst r1, #L_PTE_XN
101 tst r1, #L_PTE_YOUNG
102 tstne r1, #L_PTE_VALID
103 eorne r1, r1, #L_PTE_NONE
104 tstne r1, #L_PTE_NONE
H A Dcopypage-v4wb.c31 ldmia r1!, {r3, r4, ip, lr} @ 4\n\ v4wb_copy_user_page()
34 ldmia r1!, {r3, r4, ip, lr} @ 4+1\n\ v4wb_copy_user_page()
36 ldmia r1!, {r3, r4, ip, lr} @ 4\n\ v4wb_copy_user_page()
39 ldmia r1!, {r3, r4, ip, lr} @ 4\n\ v4wb_copy_user_page()
42 ldmneia r1!, {r3, r4, ip, lr} @ 4\n\ v4wb_copy_user_page()
44 mcr p15, 0, r1, c7, c10, 4 @ 1 drain WB\n\ v4wb_copy_user_page()
72 mov r1, %2 @ 1\n\ v4wb_clear_user_highpage()
83 subs r1, r1, #1 @ 1\n\ v4wb_clear_user_highpage()
85 mcr p15, 0, r1, c7, c10, 4 @ 1 drain WB" v4wb_clear_user_highpage()
88 : "r1", "r2", "r3", "ip", "lr"); v4wb_clear_user_highpage()
H A Dtlb-v6.S42 mov r1, r1, lsr #PAGE_SHIFT
45 mov r1, r1, lsl #PAGE_SHIFT
56 cmp r0, r1
73 mov r1, r1, lsr #PAGE_SHIFT
75 mov r1, r1, lsl #PAGE_SHIFT
84 cmp r0, r1
H A Dcopypage-fa.c26 1: ldmia r1!, {r3, r4, ip, lr} @ 4\n\ fa_copy_user_page()
30 ldmia r1!, {r3, r4, ip, lr} @ 4\n\ fa_copy_user_page()
63 mov r1, %2 @ 1\n\ fa_clear_user_highpage()
74 subs r1, r1, #1 @ 1\n\ fa_clear_user_highpage()
76 mcr p15, 0, r1, c7, c10, 4 @ 1 drain WB" fa_clear_user_highpage()
79 : "r1", "r2", "r3", "ip", "lr"); fa_clear_user_highpage()
H A Dcache-v4wb.S83 ldr r1, [r3, #0]
84 eor r1, r1, #CACHE_DSIZE
85 str r1, [r3, #0]
86 add r2, r1, #CACHE_DSIZE
87 1: ldr r3, [r1], #32
88 cmp r1, r2
92 sub r1, r2, #512 @ only 512 bytes
93 1: ldr r3, [r1], #32
94 cmp r1, r2
112 sub r3, r1, r0 @ calculate total size
122 cmp r0, r1
138 add r1, r0, r1
169 cmp r0, r1
192 tst r1, #CACHE_DLINESIZE - 1
193 mcrne p15, 0, r1, c7, c10, 1 @ clean D entry
196 cmp r0, r1
213 cmp r0, r1
238 add r1, r1, r0
H A Dcache-v6.S36 * r1 - corrupted
41 mrs r1, cpsr
47 msr cpsr_cx, r1 @ restore interrupts
139 cmp r0, r1
176 add r1, r0, r1
185 cmp r0, r1
216 tst r1, #D_CACHE_LINE_SIZE - 1
218 ldrneb r2, [r1, #-1] @ read for ownership
219 strneb r2, [r1, #-1] @ write for ownership
221 bic r1, r1, #D_CACHE_LINE_SIZE - 1
223 mcrne p15, 0, r1, c7, c14, 1 @ clean & invalidate D line
225 mcrne p15, 0, r1, c7, c15, 1 @ clean & invalidate unified line
234 cmp r0, r1
261 cmp r0, r1
285 cmp r0, r1
302 add r1, r1, r0
322 add r1, r1, r0
H A Dproc-arm1022.S146 mov r1, #(CACHE_DSEGMENTS - 1) << 5 @ 16 segments
147 1: orr r3, r1, #(CACHE_DENTRIES - 1) << 26 @ 64 entries
151 subs r1, r1, #1 << 5
173 sub r3, r1, r0 @ calculate total size
180 cmp r0, r1
224 cmp r0, r1
242 add r1, r0, r1
245 cmp r0, r1
270 tst r1, #CACHE_DLINESIZE - 1
271 mcrne p15, 0, r1, c7, c10, 1 @ clean D entry
274 cmp r0, r1
296 cmp r0, r1
316 cmp r0, r1
329 add r1, r1, r0
358 subs r1, r1, #CACHE_DLINESIZE
376 mov r1, #(CACHE_DSEGMENTS - 1) << 5 @ 16 segments
377 1: orr r3, r1, #(CACHE_DENTRIES - 1) << 26 @ 64 entries
381 subs r1, r1, #1 << 5
384 mov r1, #0
386 mcr p15, 0, r1, c7, c5, 0 @ invalidate I cache
388 mcr p15, 0, r1, c7, c10, 4 @ drain WB
390 mcr p15, 0, r1, c8, c7, 0 @ invalidate I & D TLBs
H A Dcache-v7.S39 movw r1, #0x7fff
40 and r2, r1, r0, lsr #13
42 movw r1, #0x3ff
44 and r3, r1, r0, lsr #3 @ NumWays - 1
50 clz r1, r3 @ WayShift
55 mov r5, r3, lsl r1
100 movw r1, #:lower16:(0x410fc090 >> 4) @ ID of ARM Cortex A9 r0p?
101 movt r1, #:upper16:(0x410fc090 >> 4)
102 teq r1, r2, lsr #4 @ test for errata affected core and if so...
128 mov r1, r0, lsr r2 @ extract cache type bits from clidr
129 and r1, r1, #7 @ mask of the bits for current cache only
130 cmp r1, #2 @ see what cache we have at this level
137 mrc p15, 1, r1, c0, c0, 0 @ read the new csidr
141 and r2, r1, #7 @ extract the length of the cache lines
144 ands r4, r4, r1, lsr #3 @ find maximum number on the way size
147 ands r7, r7, r1, lsr #13 @ extract max number of the index size
283 cmp r12, r1
292 cmp r12, r1
326 add r1, r0, r1
336 cmp r0, r1
363 tst r1, r3
364 bic r1, r1, r3
365 mcrne p15, 0, r1, c7, c14, 1 @ clean & invalidate D / U line
369 cmp r0, r1
391 cmp r0, r1
413 cmp r0, r1
426 add r1, r1, r0
439 add r1, r1, r0
H A Dproc-v6.S60 mrc p15, 0, r1, c1, c0, 0 @ ctrl register
61 bic r1, r1, #0x1 @ ...............m
62 mcr p15, 0, r1, c1, c0, 0 @ disable MMU
63 mov r1, #0
64 mcr p15, 0, r1, c7, c5, 4 @ ISB
77 mov r1, #0
78 mcr p15, 0, r1, c7, c10, 4 @ DWB - WFI may enter a low-power mode
79 mcr p15, 0, r1, c7, c0, 4 @ wait for interrupt
85 subs r1, r1, #D_CACHE_LINE_SIZE
102 mmid r1, r1 @ get mm->context.id
111 and r1, r1, #0xff
112 orr r1, r1, r2 @ insert into new context ID
114 mcr p15, 0, r1, c13, c0, 1 @ set context ID
165 ALT_SMP(orr r1, r1, #TTB_FLAGS_SMP)
166 ALT_UP(orr r1, r1, #TTB_FLAGS_UP)
167 mcr p15, 0, r1, c2, c0, 0 @ Translation table base 0
H A Dproc-arm940.S114 mov r1, #(CACHE_DSEGMENTS - 1) << 4 @ 4 segments
115 1: orr r3, r1, #(CACHE_DENTRIES - 1) << 26 @ 64 entries
119 subs r1, r1, #1 << 4
164 mov r1, #(CACHE_DSEGMENTS - 1) << 4 @ 4 segments
165 1: orr r3, r1, #(CACHE_DENTRIES - 1) << 26 @ 64 entries
169 subs r1, r1, #1 << 4
186 mov r1, #(CACHE_DSEGMENTS - 1) << 4 @ 4 segments
187 1: orr r3, r1, #(CACHE_DENTRIES - 1) << 26 @ 64 entries
191 subs r1, r1, #1 << 4
209 mov r1, #(CACHE_DSEGMENTS - 1) << 4 @ 4 segments
210 1: orr r3, r1, #(CACHE_DENTRIES - 1) << 26 @ 64 entries
214 subs r1, r1, #1 << 4
231 mov r1, #(CACHE_DSEGMENTS - 1) << 4 @ 4 segments
232 1: orr r3, r1, #(CACHE_DENTRIES - 1) << 26 @ 64 entries
241 subs r1, r1, #1 << 4
253 add r1, r1, r0
H A Dproc-feroceon.S61 ldr r1, __cache_params
71 stmia r1, {r2, r3}
159 ldr r1, __cache_params
160 ldmia r1, {r1, r3}
161 1: orr ip, r1, r3
165 subs r1, r1, #(1 << 5) @ next set
186 sub r3, r1, r0 @ calculate total size
196 cmp r0, r1
232 cmp r0, r1
249 add r1, r0, r1
252 cmp r0, r1
262 add r1, r0, #PAGE_SZ - CACHE_DLINESIZE @ top addr is inclusive
266 mcr p15, 5, r1, c15, c15, 1 @ D clean/inv range top
291 tst r1, #CACHE_DLINESIZE - 1
292 mcrne p15, 0, r1, c7, c10, 1 @ clean D entry
295 cmp r0, r1
305 tst r1, #CACHE_DLINESIZE - 1
306 mcrne p15, 0, r1, c7, c10, 1 @ clean D entry
307 cmp r1, r0
308 subne r1, r1, #1 @ top address is inclusive
312 mcr p15, 5, r1, c15, c14, 1 @ D inv range top
331 cmp r0, r1
339 cmp r1, r0
340 subne r1, r1, #1 @ top address is inclusive
344 mcr p15, 5, r1, c15, c13, 1 @ D clean range top
362 cmp r0, r1
370 cmp r1, r0
371 subne r1, r1, #1 @ top address is inclusive
375 mcr p15, 5, r1, c15, c15, 1 @ D clean/inv range top
387 add r1, r1, r0
401 add r1, r1, r0
450 mov r3, r1
454 subs r1, r1, #CACHE_DLINESIZE
537 mcr p15, 0, r1, c2, c0, 0 @ TTB address
H A Dproc-xscale.S118 mrc p15, 0, r1, c1, c0, 1
119 bic r1, r1, #1
120 mcr p15, 0, r1, c1, c0, 1
147 mov r1, #PSR_F_BIT|PSR_I_BIT|SVC_MODE
148 msr cpsr_c, r1 @ reset CPSR
149 mcr p15, 0, r1, c10, c4, 1 @ unlock I-TLB
150 mcr p15, 0, r1, c8, c5, 0 @ invalidate I-TLB
151 mrc p15, 0, r1, c1, c0, 0 @ ctrl register
152 bic r1, r1, #0x0086 @ ........B....CA.
153 bic r1, r1, #0x3900 @ ..VIZ..S........
156 mcr p15, 0, r1, c1, c0, 0 @ ctrl register
157 bic r1, r1, #0x0001 @ ...............M
159 mcr p15, 0, r1, c1, c0, 0 @ ctrl register
215 clean_d_cache r0, r1
234 sub r3, r1, r0 @ calculate total size
243 cmp r0, r1
267 cmp r0, r1
289 cmp r0, r1
306 add r1, r0, r1
310 cmp r0, r1
332 tst r1, #CACHELINESIZE - 1
333 mcrne p15, 0, r1, c7, c10, 1 @ clean D entry
336 cmp r0, r1
353 cmp r0, r1
371 cmp r0, r1
383 add r1, r1, r0
397 add r1, r1, r0
459 subs r1, r1, #CACHELINESIZE
474 clean_d_cache r1, r2
513 and ip, r1, #(L_PTE_MT_MASK | L_PTE_USER | L_PTE_RDONLY) & ~(4 << 2)
516 moveq r1, #L_PTE_MT_WRITETHROUGH
517 and r1, r1, #L_PTE_MT_MASK
519 ldr ip, [ip, r1]
554 mcr p15, 0, r1, c2, c0, 0 @ translation table base addr
H A Dproc-arm1020.S156 mov r1, #(CACHE_DSEGMENTS - 1) << 5 @ 16 segments
157 1: orr r3, r1, #(CACHE_DENTRIES - 1) << 26 @ 64 entries
162 subs r1, r1, #1 << 5
184 sub r3, r1, r0 @ calculate total size
193 cmp r0, r1
239 cmp r0, r1
257 add r1, r0, r1
261 cmp r0, r1
288 tst r1, #CACHE_DLINESIZE - 1
290 mcrne p15, 0, r1, c7, c10, 1 @ clean D entry
294 cmp r0, r1
317 cmp r0, r1
339 cmp r0, r1
352 add r1, r1, r0
382 subs r1, r1, #CACHE_DLINESIZE
401 mov r1, #0xF @ 16 segments
404 orr ip, ip, r1, LSL #5 @ shift in/up index
411 subs r1, r1, #1
412 cmp r1, #0
416 mov r1, #0
418 mcr p15, 0, r1, c7, c5, 0 @ invalidate I cache
420 mcr p15, 0, r1, c7, c10, 4 @ drain WB
422 mcr p15, 0, r1, c8, c7, 0 @ invalidate I & D TLBs
H A Dproc-arm1020e.S156 mov r1, #(CACHE_DSEGMENTS - 1) << 5 @ 16 segments
157 1: orr r3, r1, #(CACHE_DENTRIES - 1) << 26 @ 64 entries
161 subs r1, r1, #1 << 5
183 sub r3, r1, r0 @ calculate total size
190 cmp r0, r1
233 cmp r0, r1
251 add r1, r0, r1
254 cmp r0, r1
279 tst r1, #CACHE_DLINESIZE - 1
280 mcrne p15, 0, r1, c7, c10, 1 @ clean D entry
283 cmp r0, r1
305 cmp r0, r1
325 cmp r0, r1
338 add r1, r1, r0
367 subs r1, r1, #CACHE_DLINESIZE
386 mov r1, #0xF @ 16 segments
389 orr ip, ip, r1, LSL #5 @ shift in/up index
395 subs r1, r1, #1
396 cmp r1, #0
400 mov r1, #0
402 mcr p15, 0, r1, c7, c5, 0 @ invalidate I cache
404 mcr p15, 0, r1, c7, c10, 4 @ drain WB
406 mcr p15, 0, r1, c8, c7, 0 @ invalidate I & D TLBs
/linux-4.1.27/arch/sh/kernel/cpu/sh2/
H A Dex.S21 mov.l r1,@-sp
23 mov #no,r1
29 extu.b r1,r1
31 extu.w r1,r1
H A Dentry.S41 ! r1
45 ! r1 = vector (pseudo EXPEVT / INTEVT / TRA)
66 mov.l r1,@-r15 ! TRA
89 mov r1,r9 ! save TRA
95 mov.l @(12,r8),r1
96 mov.l r1,@-r15
117 mov.l r1,@-r2 ! TRA
138 mov r1,r9
140 mov.l @(OFF_R1,r15),r1
198 mov.l @(OFF_SP,r0),r1
201 mov.l r3,@-r1
204 mov.l r3,@-r1
212 mov.l r1,@(8,r0)
214 mov.l @r15+, r1
264 mov.l @(r0,r3),r1
266 and r1,r3 ! copy MD bit
268 shll2 r1 ! clear MD bit
269 shlr2 r1
273 mov.l r1,@(4,r2) ! set sr
275 mov.l @(r0,r3),r1
276 mov.l r1,@r2 ! set pc
277 get_current_thread_info r0, r1
278 mov.l $current_thread_info,r1
279 mov.l r0,@r1
281 mov.l @r15+,r1
/linux-4.1.27/arch/arm/mach-pxa/
H A Dstandby.S23 mov r1, #(PSSR_PH | PSSR_STS)
31 str r1, [r0] @ make sure PSSR_PH/STS are clear
51 mov r1, #0xf6000000 @ DMEMC_REG_BASE (PXA3_MDCNFG)
52 add r1, r1, #0x00100000
62 ldr r2, [r1] @ Dummy read PXA3_MDCNFG
69 ldr r0, [r1, #PXA3_DDR_HCAL] @ Clear (and wait for) HCEN
71 str r0, [r1, #PXA3_DDR_HCAL]
72 1: ldr r0, [r1, #PXA3_DDR_HCAL]
76 ldr r0, [r1, #PXA3_RCOMP] @ Initiate RCOMP
78 str r0, [r1, #PXA3_RCOMP]
81 str r0, [r1, #PXA3_DMCISR]
83 ldr r0, [r1, #PXA3_DMCIER] @ set DMIER[EDLP]
85 str r0, [r1, #PXA3_DMCIER]
87 ldr r0, [r1, #PXA3_DDR_HCAL] @ clear HCRNG, set HCPROG, HCEN
90 str r0, [r1, #PXA3_DDR_HCAL]
92 1: ldr r0, [r1, #PXA3_DMCISR]
96 ldr r0, [r1, #PXA3_MDCNFG] @ set PXA3_MDCNFG[DMCEN]
98 str r0, [r1, #PXA3_MDCNFG]
99 1: ldr r0, [r1, #PXA3_MDCNFG]
103 ldr r0, [r1, #PXA3_DDR_HCAL] @ set PXA3_DDR_HCAL[HCRNG]
105 str r0, [r1, #PXA3_DDR_HCAL]
107 ldr r0, [r1, #PXA3_DMCIER] @ Clear the interrupt
109 str r0, [r1, #PXA3_DMCIER]
/linux-4.1.27/arch/powerpc/platforms/powermac/
H A Dsleep.S62 stw r0,4(r1)
63 stwu r1,-SL_SIZE(r1)
65 stw r0,SL_CR(r1)
66 stw r2,SL_R2(r1)
67 stmw r12,SL_R12(r1)
71 stw r4,SL_MSR(r1)
73 stw r4,SL_SDR1(r1)
77 stw r4,SL_TB(r1)
79 stw r5,SL_TB+4(r1)
86 stw r4,SL_SPRG0(r1)
88 stw r4,SL_SPRG0+4(r1)
90 stw r4,SL_SPRG0+8(r1)
92 stw r4,SL_SPRG0+12(r1)
96 stw r4,SL_DBAT0(r1)
98 stw r4,SL_DBAT0+4(r1)
100 stw r4,SL_DBAT1(r1)
102 stw r4,SL_DBAT1+4(r1)
104 stw r4,SL_DBAT2(r1)
106 stw r4,SL_DBAT2+4(r1)
108 stw r4,SL_DBAT3(r1)
110 stw r4,SL_DBAT3+4(r1)
112 stw r4,SL_IBAT0(r1)
114 stw r4,SL_IBAT0+4(r1)
116 stw r4,SL_IBAT1(r1)
118 stw r4,SL_IBAT1+4(r1)
120 stw r4,SL_IBAT2(r1)
122 stw r4,SL_IBAT2+4(r1)
124 stw r4,SL_IBAT3(r1)
126 stw r4,SL_IBAT3+4(r1)
144 stw r5,SL_PC(r1)
146 tophys(r5,r1)
243 lwz r1,0(r3)
248 * r1 has the physical address of SL_PC(sp).
254 * we do any r1 memory access as we are not sure they
268 subi r1,r1,SL_PC
283 lwz r4,SL_SDR1(r1)
285 lwz r4,SL_SPRG0(r1)
287 lwz r4,SL_SPRG0+4(r1)
289 lwz r4,SL_SPRG0+8(r1)
291 lwz r4,SL_SPRG0+12(r1)
294 lwz r4,SL_DBAT0(r1)
296 lwz r4,SL_DBAT0+4(r1)
298 lwz r4,SL_DBAT1(r1)
300 lwz r4,SL_DBAT1+4(r1)
302 lwz r4,SL_DBAT2(r1)
304 lwz r4,SL_DBAT2+4(r1)
306 lwz r4,SL_DBAT3(r1)
308 lwz r4,SL_DBAT3+4(r1)
310 lwz r4,SL_IBAT0(r1)
312 lwz r4,SL_IBAT0+4(r1)
314 lwz r4,SL_IBAT1(r1)
316 lwz r4,SL_IBAT1+4(r1)
318 lwz r4,SL_IBAT2(r1)
320 lwz r4,SL_IBAT2+4(r1)
322 lwz r4,SL_IBAT3(r1)
324 lwz r4,SL_IBAT3+4(r1)
355 lwz r3,SL_MSR(r1)
359 tovirt(r1,r1)
364 lwz r3,SL_TB(r1)
365 lwz r4,SL_TB+4(r1)
370 lwz r0,SL_CR(r1)
372 lwz r2,SL_R2(r1)
373 lmw r12,SL_R12(r1)
374 addi r1,r1,SL_SIZE
375 lwz r0,4(r1)
/linux-4.1.27/arch/sh/boot/romimage/
H A Dhead.S47 mov.l extra_data_size, r1
48 add r1, r0
49 mov.l empty_zero_page_dst, r1
60 mov.l r4, @r1
61 mov.l r5, @(4, r1)
62 mov.l r6, @(8, r1)
63 mov.l r7, @(12, r1)
65 add #16,r1
70 mov #1, r1
71 shld r4, r1
73 add r1, r0
74 mov.l extra_data_size, r1
75 add r1, r0
/linux-4.1.27/arch/arm/mach-shmobile/
H A Dheadsmp.S28 ldr r1, 1f
29 bx r1
50 mrc p15, 0, r1, c0, c0, 5 @ r1 = MPIDR
51 and r0, r1, r0 @ r0 = cpu_logical_map() value
52 mov r1, #0 @ r1 = CPU index
58 ldr r8, [r5, r1, lsl #2]
62 ldr r9, [r6, r1, lsl #2]
67 add r1, r1, #1
68 cmp r1, #NR_CPUS
74 ldr r0, [r7, r1, lsl #2]
H A Dheadsmp-scu.S30 mrc p15, 0, r1, c0, c0, 5 @ read MIPDR
31 and r1, r1, #3 @ mask out cpu ID
32 lsl r1, r1, #3 @ we will shift by cpu_id * 8 bits
35 lsl r3, r3, r1
/linux-4.1.27/arch/blackfin/mach-bf561/
H A Datomic.S41 r1 = -L1_CACHE_BYTES; define
42 r1 = r0 & r1; define
51 p0 = r1;
87 * r1 = address of atomic data to flush and invalidate (32bit).
95 r1 = 0; define
96 [p0] = r1;
194 * r1 = address of atomic data
203 p1 = r1;
210 r1 = 1; define
211 r1 <<= r2;
212 r2 = ~r1;
214 r1 = [p1]; define
215 r1 >>= 28; /* CPU fingerprints are stored in the high nibble. */
216 r6 = r1 & r2;
217 r1 = [p1]; define
218 r1 <<= 4;
219 r1 >>= 4;
220 [p1] = r1;
227 r1 = 0; define
228 [p0] = r1;
254 * r1 = address of atomic data
260 p1 = r1;
263 r1 = 1; define
264 r1 <<= r2;
266 r2 = r1 | r2;
268 r1 = p1; define
285 r1 = p1; define
302 r1 = p1; define
337 r1 = p1; define
372 r1 = p1; define
392 r1 = [p1]; define
393 r1 += -1;
394 [p1] = r1;
395 cc = r1 < 0;
397 r1 = p1; define
407 r1 += 1;
408 [p1] = r1;
410 r1 = p1; define
415 r1 = [p1]; define
416 cc = r1 < 2;
430 r1 = [p1]; define
431 cc = r1 <= 0;
433 r1 += -1;
434 [p1] = r1;
435 r1 = p1; define
445 r1 = p1; define
465 r1 = [p1]; define
466 r1 += 1;
467 [p1] = r1;
468 r1 = p1; define
486 r1 = [p1]; define
487 r1 = r1 - r3; define
489 r2 = r1;
494 cc = r1 == 0;
497 [p1] = r1;
498 r1 = p1; define
508 r1 = p1; define
513 r1 = [p1]; define
515 r1 <<= 4;
516 r1 >>= 4;
518 cc = r1 == r3;
532 r1 = [p1]; define
535 cc = r1 == r2;
538 r1 >>= 28;
539 r1 <<= 28;
541 r1 = 0; define
543 [p1] = r1;
544 r1 = p1; define
555 r1 = p1; define
573 r1 = [p1]; define
574 r1 = r1 + r3; define
575 [p1] = r1;
576 r1 = p1; define
588 * r1 = value
595 r3 = r1;
601 r1 = p1; define
610 * r1 = mask
618 r3 = ~r1;
625 r1 = p1; define
634 * r1 = mask
642 r3 = r1;
649 r1 = p1; define
658 * r1 = mask
666 r3 = r1;
673 r1 = p1; define
682 * r1 = mask
692 r3 = r1;
693 r1 = -L1_CACHE_BYTES; define
694 r1 = r0 & r1; define
695 p0 = r1;
707 * r1 = value
714 r3 = r1; \
720 r1 = p1; \ define
740 * r1 = new
752 r3 = r1; \
760 r1 = p1; \ define
781 * r1 = bitnr
787 r2 = r1;
788 r1 = 1; define
789 r1 <<= r2;
795 * r1 = bitnr
801 r2 = r1;
802 r1 = 1; define
803 r1 <<= r2;
809 * r1 = bitnr
815 r2 = r1;
816 r1 = 1; define
817 r1 <<= r2;
823 * r1 = bitnr
830 [--sp] = r1;
832 r1 = [sp++]; define
834 r2 <<= r1;
846 * r1 = bitnr
853 [--sp] = r1;
855 r1 = [sp++]; define
857 r2 <<= r1;
869 * r1 = bitnr
877 [--sp] = r1;
879 r1 = [sp++]; define
881 r2 <<= r1;
893 * r1 = bitnr
902 r2 = r1;
903 r1 = 1; define
904 r1 <<= r2;
917 r1 = -L1_CACHE_BYTES; define
918 r1 = r0 & r1; define
919 p0 = r1;
/linux-4.1.27/arch/m32r/boot/compressed/
H A Dhead.S34 seth r1, #high(CONFIG_MEMORY_START + 0x00400000) /* Start address */
36 sub r12, r1 /* difference */
44 ldi r1, #low(got_len)
45 srli r1, #2
46 beqz r1, 2f
53 addi r1, #-1
54 bnez r1, 1b
72 mv r4, r3 || ldi r1, #0
80 st r1, @+r2 || addi r4, #-1
81 st r1, @+r2
82 st r1, @+r2
83 st r1, @+r2 || cmpeq r1, r4 ; R4 = 0?
90 stb r1, @r2 || addi r4, #-1
105 ldi r1, #0 ; clear R1 for longwords store
109 st r1, @+r2 ; yep, zero out another longword
116 seth r1, #high(end)
117 or3 r1, r1, #low(end)
118 add r1, r12
119 mv sp, r1
126 seth r1, #high(zimage_data)
127 or3 r1, r1, #low(zimage_data)
128 add r1, r12
138 ldi r1, 0xd0 ; invalidate i-cache, copy back d-cache
139 stb r1, @r0
143 ldi r1, 0x0100 ; invalidate
144 stb r1, @r0
148 ldi r1, 0x0700 ; invalidate i-cache, copy back d-cache
149 sth r1, @r0
158 seth r1, #high(CONFIG_MEMORY_START)
159 or r0, r1
/linux-4.1.27/arch/cris/arch-v32/mach-a3/
H A Ddram_init.S34 move.d REG_STATE(ddr2, rw_phy_cfg, en, yes), $r1
35 move.d $r1, [$r0]
45 REG_STATE(ddr2, rw_phy_ctrl, cal_rst, yes), $r1
46 move.d $r1, [$r0]
47 move.d REG_STATE(ddr2, rw_phy_ctrl, cal_start, yes), $r1
48 move.d $r1, [$r0]
59 movu.b [$r2+], $r1
62 lslq 16, $r1
63 or.d $r3, $r1
64 move.d $r1, [$r0]
75 move.d CONFIG_ETRAX_DDR2_TIMING, $r1
76 move.d $r1, [$r0]
80 move.d CONFIG_ETRAX_DDR2_LATENCY, $r1
81 move.d $r1, [$r0]
85 move.d CONFIG_ETRAX_DDR2_CONFIG, $r1
86 move.d $r1, [$r0]
/linux-4.1.27/arch/arm/mach-shmobile/include/mach/
H A Dzboot_macros.h11 LDR r1, 2f variable
12 STR r1, [r0] variable
22 LDR r1, 2f variable
23 STRH r1, [r0] variable
33 LDR r1, 2f variable
34 STRB r1, [r0] variable
43 LDR r1, 1f variable
45 STR r0, [r1]
47 LDR r0, [r1]
59 LDR r1, 1f variable
68 LDR r1, 3f variable
72 AND r3, r1, r3 variable
/linux-4.1.27/arch/arc/include/asm/
H A Dtlb-mmu1.h27 ; r1 = TLBPD0 from TLB_RELOAD above
34 ; as r0/r1 saves above
40 asr r0,r1,12 ; get set # <<1, note bit 12=R=0 variable
52 /* r1 = data TLBPD0 at this point */
54 xor r0,r0,r1 /* compare set # */
60 ; lr r1,[ARC_REG_TLBPD0] /* Data VPN+ASID - already in r1 from TLB_RELOAD*/ variable
61 and r1,r1,0xff /* Data ASID */ variable
62 or r0,r0,r1 /* Instruction address + Data ASID */
64 lr r1,[ARC_REG_TLBPD0] /* save TLBPD0 containing data TLB*/ variable
68 sr r1,[ARC_REG_TLBPD0] /* restore TLBPD0 */ variable
82 mov_s r3, r1 ; save PD0 prepared by TLB_RELOAD in r3 variable
85 bmsk r1,r3,7 /* Data ASID, bits 7-0 */ variable
86 or_s r0,r0,r1 /* Instruction address + Data ASID */
94 lr r1,[ARC_REG_TLBINDEX] /* r1 = index where MMU wants to put data */ variable
95 cmp r0,r1 /* if no match on indices, go around */
96 xor.eq r1,r1,1 /* flip bottom bit of data index */ variable
97 sr r1,[ARC_REG_TLBINDEX] /* and put it back */ variable
/linux-4.1.27/arch/s390/kernel/vdso64/
H A Dclock_gettime.S42 lg %r1,1(%r15)
43 sg %r1,__VDSO_XTIME_STAMP(%r5) /* TOD - cycle_last */
44 msgf %r1,__VDSO_TK_MULT(%r5) /* * tk->mult */
45 alg %r1,__VDSO_WTOM_NSEC(%r5)
46 srlg %r1,%r1,0(%r2) /* >> tk->shift */
50 1: clg %r1,0(%r5)
52 slg %r1,0(%r5)
56 stg %r1,8(%r3) /* store tp->tv_nsec */
66 lg %r1,__VDSO_WTOM_CRS_NSEC(%r5)
76 lg %r1,__VDSO_XTIME_CRS_NSEC(%r5)
87 lg %r1,1(%r15)
88 sg %r1,__VDSO_XTIME_STAMP(%r5) /* TOD - cycle_last */
89 msgf %r1,__VDSO_TK_MULT(%r5) /* * tk->mult */
90 alg %r1,__VDSO_XTIME_NSEC(%r5) /* + tk->xtime_nsec */
91 srlg %r1,%r1,0(%r2) /* >> tk->shift */
96 6: clg %r1,0(%r5)
98 slg %r1,0(%r5)
102 stg %r1,8(%r3) /* store tp->tv_nsec */
125 algr %r1,%r0 /* r1 = cputime as TOD value */
126 mghi %r1,1000 /* convert to nanoseconds */
127 srlg %r1,%r1,12 /* r1 = cputime in nanosec */
128 lgr %r4,%r1
130 srlg %r1,%r1,9 /* divide by 1000000000 */
142 12: lghi %r1,__NR_clock_gettime
H A Dgettimeofday.S33 lg %r1,1(%r15)
34 sg %r1,__VDSO_XTIME_STAMP(%r5) /* TOD - cycle_last */
35 msgf %r1,__VDSO_TK_MULT(%r5) /* * tk->mult */
36 alg %r1,__VDSO_XTIME_NSEC(%r5) /* + tk->xtime_nsec */
41 srlg %r1,%r1,0(%r5) /* >> tk->shift */
43 2: clg %r1,0(%r5)
45 slg %r1,0(%r5)
H A Dclock_getres.S22 larl %r1,4f
27 larl %r1,3f
41 lg %r0,0(%r1)
46 2: lghi %r1,__NR_clock_getres /* fallback to svc */
/linux-4.1.27/arch/nios2/kernel/
H A Dhead.S68 movia r1, NIOS2_ICACHE_SIZE
72 initi r1
73 sub r1, r1, r2
74 bgt r1, r0, icache_init
118 movia r1, NIOS2_DCACHE_SIZE
122 initd 0(r1)
123 sub r1, r1, r2
124 bgt r1, r0, dcache_init
126 nextpc r1 /* Find out where we are */
129 beq r1, r2,finish_move /* We are running in RAM done */
130 addi r1, r1,(_start - chkadr) /* Source */
134 loop_move: /* r1: src, r2: dest, r3: last dest */
135 ldw r8, 0(r1) /* load a word from [r1] */
138 addi r1, r1, 4 /* inc the src addr */
142 movia r1, finish_move /* VMA(_start)->l1 */
143 jmp r1 /* jmp to _start */
152 movia r1, __bss_stop
156 bne r1, r2, 1b
158 movia r1, init_thread_union /* set stack at top of the task union */
159 addi sp, r1, THREAD_SIZE
161 stw r1, 0(r2)
163 movia r1, nios2_boot_init /* save args r4-r7 passed from u-boot */
164 callr r1
166 movia r1, start_kernel /* call start_kernel as a subroutine */
167 callr r1
H A Dentry.S189 movui r1, __NR_syscalls
190 bgeu r2, r1, ret_invsyscall
191 slli r1, r2, 2
193 add r1, r1, r11
194 ldw r1, %lo(sys_call_table)(r1)
195 beq r1, r0, ret_invsyscall
203 callr r1
214 movi r1, 0
217 movi r1, 1
220 stw r1, PT_R7(sp)
224 ldw r1, PT_ESTATUS(sp)
226 TSTBNZ r1, r1, ESTATUS_EU, Luser_return
261 slli r1, r2, 2
263 add r1, r1, r11
264 ldw r1, %lo(sys_call_table)(r1)
266 callr r1
277 movi r1, 0
280 movi r1, 1
283 stw r1, PT_R7(sp)
295 BTBZ r1, r10, TIF_NEED_RESCHED, Lsignal_return
302 ANDI32 r1, r10, _TIF_SIGPENDING | _TIF_NOTIFY_RESUME
303 beq r1, r0, restore_all
365 ldw r1, PT_ESTATUS(sp) /* check if returning to kernel */
366 TSTBNZ r1, r1, ESTATUS_EU, Luser_return
369 GET_THREAD_INFO r1
370 ldw r4, TI_PREEMPT_COUNT(r1)
372 ldw r4, TI_FLAGS(r1) /* ? Need resched set */
488 GET_THREAD_INFO r1
489 stw r1, 0(r24)
/linux-4.1.27/arch/arm/mach-imx/
H A Dheadsmp.S21 ldr r1, [r0]
22 add r1, r1, r0 @ r1 = physical &g_diag_reg
23 ldr r0, [r1]
/linux-4.1.27/arch/powerpc/kernel/
H A Dentry_32.S96 rlwimi r0,r1,0,0,(31-THREAD_SHIFT)
120 rlwimi r0,r1,0,0,(31-THREAD_SHIFT)
151 addi r11,r1,STACK_FRAME_OVERHEAD
166 CURRENT_THREAD_INFO(r9, r1)
183 cmplw r1,r9 /* if r1 <= ksp_limit */
187 CURRENT_THREAD_INFO(r9, r1)
209 lwz r12,_MSR(r1)
221 * r3 can be different from GPR3(r1) at this point, r9 and r11
228 stwu r1,-32(r1)
229 stw r9,8(r1)
230 stw r11,12(r1)
231 stw r3,16(r1)
232 stw r4,20(r1)
233 stw r5,24(r1)
235 lwz r5,24(r1)
236 lwz r4,20(r1)
237 lwz r3,16(r1)
238 lwz r11,12(r1)
239 lwz r9,8(r1)
240 addi r1,r1,32
241 lwz r0,GPR0(r1)
242 lwz r6,GPR6(r1)
243 lwz r7,GPR7(r1)
244 lwz r8,GPR8(r1)
277 cmplw r1,r12
278 ble 5b /* r1 <= &_end is OK */
280 addi r3,r1,STACK_FRAME_OVERHEAD
281 lis r1,init_thread_union@ha
282 addi r1,r1,init_thread_union@l
283 addi r1,r1,THREAD_SIZE-STACK_FRAME_OVERHEAD
301 stw r3,ORIG_GPR3(r1)
303 stw r12,RESULT(r1)
304 lwz r11,_CCR(r1) /* Clear SO bit in CR */
306 stw r11,_CCR(r1)
321 lwz r0,GPR0(r1)
322 lwz r3,GPR3(r1)
323 lwz r4,GPR4(r1)
325 lwz r5,GPR5(r1)
326 lwz r6,GPR6(r1)
327 lwz r7,GPR7(r1)
328 lwz r8,GPR8(r1)
332 CURRENT_THREAD_INFO(r10, r1)
344 addi r9,r1,STACK_FRAME_OVERHEAD
350 CURRENT_THREAD_INFO(r12, r1)
362 lwz r11,_CCR(r1) /* Load CR */
365 stw r11,_CCR(r1)
367 lwz r8,_MSR(r1)
375 stw r3,GPR3(r1)
377 lwz r3,GPR3(r1)
397 lwarx r7,0,r1
399 stwcx. r0,0,r1 /* to clear the reservation */
400 lwz r4,_LINK(r1)
401 lwz r5,_CCR(r1)
404 lwz r7,_NIP(r1)
406 lwz r2,GPR2(r1)
407 lwz r1,GPR1(r1)
424 REST_NVGPRS(r1)
431 REST_NVGPRS(r1)
442 SAVE_NVGPRS(r1)
444 stw r0,_TRAP(r1)
445 addi r3,r1,STACK_FRAME_OVERHEAD
453 lwz r3,GPR3(r1)
454 lwz r4,GPR4(r1)
455 lwz r5,GPR5(r1)
456 lwz r6,GPR6(r1)
457 lwz r7,GPR7(r1)
458 lwz r8,GPR8(r1)
459 REST_NVGPRS(r1)
465 REST_NVGPRS(r1)
471 lwz r11,_CCR(r1) /* Load CR */
474 stw r11,_CCR(r1)
476 1: stw r6,RESULT(r1) /* Save result */
477 stw r3,GPR3(r1) /* Update return value */
506 lwz r4,_TRAP(r1)
509 SAVE_NVGPRS(r1)
511 stw r4,_TRAP(r1)
513 addi r3,r1,STACK_FRAME_OVERHEAD
524 SAVE_NVGPRS(r1)
525 lwz r0,_TRAP(r1)
527 stw r0,_TRAP(r1) /* register set saved */
532 SAVE_NVGPRS(r1)
533 lwz r0,_TRAP(r1)
535 stw r0,_TRAP(r1) /* register set saved */
540 SAVE_NVGPRS(r1)
541 lwz r0,_TRAP(r1)
543 stw r0,_TRAP(r1) /* register set saved */
548 SAVE_NVGPRS(r1)
549 lwz r0,_TRAP(r1)
551 stw r0,_TRAP(r1) /* register set saved */
562 stw r4,_DAR(r1)
563 addi r3,r1,STACK_FRAME_OVERHEAD
567 SAVE_NVGPRS(r1)
568 lwz r0,_TRAP(r1)
570 stw r0,_TRAP(r1)
572 addi r3,r1,STACK_FRAME_OVERHEAD
573 lwz r4,_DAR(r1)
598 stwu r1,-INT_FRAME_SIZE(r1)
600 stw r0,INT_FRAME_SIZE+4(r1)
602 SAVE_NVGPRS(r1)
603 stw r0,_NIP(r1) /* Return to switch caller */
625 1: stw r11,_MSR(r1)
627 stw r10,_CCR(r1)
628 stw r1,KSP(r3) /* Set old stack pointer */
641 lwz r1,KSP(r4) /* Load new stack pointer */
660 lwz r0,_CCR(r1)
663 REST_NVGPRS(r1)
665 lwz r4,_NIP(r1) /* Return to _switch caller in new task */
667 addi r1,r1,INT_FRAME_SIZE
725 addi r3,r1,STACK_FRAME_OVERHEAD
735 REST_NVGPRS(r1)
748 lwz r3,_MSR(r1) /* Returning to user mode? */
754 CURRENT_THREAD_INFO(r9, r1)
773 CURRENT_THREAD_INFO(r9, r1)
778 addi r8,r1,INT_FRAME_SIZE /* Get the kprobed function entry */
780 lwz r3,GPR1(r1)
782 mr r4,r1 /* src: current exception frame */
783 mr r1,r3 /* Reroute the trampoline frame to r1 */
795 lwz r5,GPR1(r1)
817 lwz r3,_MSR(r1)
828 CURRENT_THREAD_INFO(r9, r1)
856 lwz r9,_MSR(r1)
877 stwu r1,-32(r1)
879 stw r0,4(r1)
880 stwu r1,-32(r1)
882 lwz r1,0(r1)
883 lwz r1,0(r1)
884 lwz r9,_MSR(r1)
888 lwz r0,GPR0(r1)
889 lwz r2,GPR2(r1)
890 REST_4GPRS(3, r1)
891 REST_2GPRS(7, r1)
893 lwz r10,_XER(r1)
894 lwz r11,_CTR(r1)
898 PPC405_ERR77(0,r1)
900 lwarx r11,0,r1
902 stwcx. r0,0,r1 /* to clear the reservation */
908 lwz r10,_CCR(r1)
909 lwz r11,_LINK(r1)
927 lwz r12,_NIP(r1)
931 REST_4GPRS(9, r1)
932 lwz r1,GPR1(r1)
946 lwz r11,_LINK(r1)
948 lwz r10,_CCR(r1)
950 REST_2GPRS(9, r1)
953 lwz r11,_NIP(r1)
954 lwz r12,_MSR(r1)
958 REST_2GPRS(11, r1)
959 lwz r1,GPR1(r1)
972 * running with r1 pointing into critical_stack, not the current
986 tophys(r1, r1);
992 REST_NVGPRS(r1); \
993 lwz r3,_MSR(r1); \
997 lwz r0,GPR0(r1); \
998 lwz r2,GPR2(r1); \
999 REST_4GPRS(3, r1); \
1000 REST_2GPRS(7, r1); \
1001 lwz r10,_XER(r1); \
1002 lwz r11,_CTR(r1); \
1005 PPC405_ERR77(0,r1); \
1006 stwcx. r0,0,r1; /* to clear the reservation */ \
1007 lwz r11,_LINK(r1); \
1009 lwz r10,_CCR(r1); \
1012 lwz r9,_DEAR(r1); \
1013 lwz r10,_ESR(r1); \
1016 lwz r11,_NIP(r1); \
1017 lwz r12,_MSR(r1); \
1020 lwz r9,GPR9(r1); \
1021 lwz r12,GPR12(r1); \
1022 lwz r10,GPR10(r1); \
1023 lwz r11,GPR11(r1); \
1024 lwz r1,GPR1(r1); \
1030 lwz r9,_##exc_lvl_srr0(r1); \
1031 lwz r10,_##exc_lvl_srr1(r1); \
1038 lwz r11,MAS7(r1); \
1044 lwz r9,MAS0(r1); \
1045 lwz r10,MAS1(r1); \
1046 lwz r11,MAS2(r1); \
1048 lwz r9,MAS3(r1); \
1050 lwz r10,MAS6(r1); \
1057 lwz r9,MMUCR(r1); \
1084 lwz r10,SAVED_KSP_LIMIT(r1)
1093 lwz r10,SAVED_KSP_LIMIT(r1)
1096 CURRENT_THREAD_INFO(r10, r1)
1107 lwz r10,SAVED_KSP_LIMIT(r1)
1130 CURRENT_THREAD_INFO(r9, r1)
1171 CURRENT_THREAD_INFO(r9, r1)
1182 lwz r3,_TRAP(r1)
1185 SAVE_NVGPRS(r1)
1187 stw r3,_TRAP(r1)
1188 2: addi r3,r1,STACK_FRAME_OVERHEAD
1191 REST_NVGPRS(r1)
1219 lwz r3,_TRAP(r1)
1222 SAVE_NVGPRS(r1)
1224 stw r3,_TRAP(r1)
1225 4: addi r3,r1,STACK_FRAME_OVERHEAD
1247 stwu r1,-INT_FRAME_SIZE(r1)
1249 stw r0,INT_FRAME_SIZE+4(r1)
1254 tophys(r7,r1)
1258 stw r9,8(r1)
1268 1: tophys(r9,r1)
1272 addi r1,r1,INT_FRAME_SIZE
1299 lwz r0, 4(r1)
1347 lwz r4, 44(r1)
1351 lwz r3,52(r1)
1360 stw r3,52(r1)
1368 stwu r1, -32(r1)
1369 stw r3, 20(r1)
1370 stw r4, 16(r1)
1371 stw r31, 12(r1)
1372 mr r31, r1
1380 lwz r3, 20(r1)
1381 lwz r4, 16(r1)
1382 lwz r31,12(r1)
1383 lwz r1, 0(r1)
H A Dentry_64.S55 mr r10,r1
56 addi r1,r1,-INT_FRAME_SIZE
58 ld r1,PACAKSAVE(r13)
59 1: std r10,0(r1)
60 std r11,_NIP(r1)
61 std r12,_MSR(r1)
62 std r0,GPR0(r1)
63 std r10,GPR1(r1)
66 2: std r2,GPR2(r1)
67 std r3,GPR3(r1)
69 std r4,GPR4(r1)
70 std r5,GPR5(r1)
71 std r6,GPR6(r1)
72 std r7,GPR7(r1)
73 std r8,GPR8(r1)
75 std r11,GPR9(r1)
76 std r11,GPR10(r1)
77 std r11,GPR11(r1)
78 std r11,GPR12(r1)
79 std r11,_XER(r1)
80 std r11,_CTR(r1)
81 std r9,GPR13(r1)
89 std r10,_LINK(r1)
90 std r11,_TRAP(r1)
91 std r3,ORIG_GPR3(r1)
92 std r2,_CCR(r1)
94 addi r9,r1,STACK_FRAME_OVERHEAD
108 REST_GPR(0,r1)
109 REST_4GPRS(3,r1)
110 REST_2GPRS(7,r1)
111 addi r9,r1,STACK_FRAME_OVERHEAD
141 std r10,SOFTE(r1)
143 CURRENT_THREAD_INFO(r11, r1)
173 std r3,RESULT(r1)
174 CURRENT_THREAD_INFO(r12, r1)
176 ld r8,_MSR(r1)
207 ld r5,_CCR(r1)
210 ld r7,_NIP(r1)
212 stdcx. r0,0,r1 /* to clear the reservation */
215 ld r4,_LINK(r1)
220 ld r13,GPR13(r1) /* only restore r13 if returning to usermode */
221 1: ld r2,GPR2(r1)
222 ld r1,GPR1(r1)
233 std r5,_CCR(r1)
239 addi r3,r1,STACK_FRAME_OVERHEAD
247 ld r3,GPR3(r1)
248 ld r4,GPR4(r1)
249 ld r5,GPR5(r1)
250 ld r6,GPR6(r1)
251 ld r7,GPR7(r1)
252 ld r8,GPR8(r1)
253 addi r9,r1,STACK_FRAME_OVERHEAD
254 CURRENT_THREAD_INFO(r10, r1)
271 REST_NVGPRS(r1)
277 ld r5,_CCR(r1)
280 std r5,_CCR(r1)
281 1: std r3,GPR3(r1)
310 addi r3,r1,STACK_FRAME_OVERHEAD
316 ld r11,_TRAP(r1)
319 SAVE_NVGPRS(r1)
321 std r0,_TRAP(r1)
366 REST_NVGPRS(r1)
372 REST_NVGPRS(r1)
403 std r0,16(r1)
404 stdu r1,-SWITCH_FRAME_SIZE(r1)
406 SAVE_8GPRS(14, r1)
407 SAVE_10GPRS(22, r1)
428 1: std r20,_NIP(r1)
430 std r23,_CCR(r1)
431 std r1,KSP(r3) /* Set old stack pointer */
461 ldarx r6,0,r1
478 clrrdi r9,r1,28 /* get current sp ESID */
481 clrrdi r9,r1,40 /* get current sp 1T ESID */
529 mr r1,r8 /* start using new stack pointer */
573 ld r6,_CCR(r1)
577 REST_8GPRS(14, r1)
578 REST_10GPRS(22, r1)
582 ld r7,_NIP(r1) /* Return to _switch caller in new task */
584 addi r1,r1,SWITCH_FRAME_SIZE
589 ld r11,_TRAP(r1)
592 REST_NVGPRS(r1)
607 CURRENT_THREAD_INFO(r9, r1)
608 ld r3,_MSR(r1)
648 addi r3,r1,STACK_FRAME_OVERHEAD
661 addi r3,r1,STACK_FRAME_OVERHEAD
670 addi r8,r1,INT_FRAME_SIZE /* Get the kprobed function entry */
672 lwz r3,GPR1(r1)
674 mr r4,r1 /* src: current exception frame */
675 mr r1,r3 /* Reroute the trampoline frame to r1 */
687 lwz r5,GPR1(r1)
706 ld r0,SOFTE(r1)
719 CURRENT_THREAD_INFO(r9, r1)
745 ld r5,SOFTE(r1)
786 stdcx. r0,0,r1 /* to clear the reservation */
788 ldarx r4,0,r1
799 ld r3,_MSR(r1)
800 ld r4,_CTR(r1)
801 ld r0,_LINK(r1)
804 ld r4,_XER(r1)
807 REST_8GPRS(5, r1)
842 REST_GPR(13, r1)
846 ld r2,_CCR(r1)
848 ld r2,_NIP(r1)
851 ld r0,GPR0(r1)
852 ld r2,GPR2(r1)
853 ld r3,GPR3(r1)
854 ld r4,GPR4(r1)
855 ld r1,GPR1(r1)
870 ld r3,_MSR(r1)
902 ld r4,_TRAP(r1)
905 std r4,_TRAP(r1)
913 addi r3,r1,STACK_FRAME_OVERHEAD;
918 addi r3,r1,STACK_FRAME_OVERHEAD;
923 addi r3,r1,STACK_FRAME_OVERHEAD;
938 addi r3,r1,STACK_FRAME_OVERHEAD;
945 addi r3,r1,STACK_FRAME_OVERHEAD
960 std r0,16(r1)
961 stdu r1,-RTAS_FRAME_SIZE(r1) /* Save SP and create stack space. */
967 SAVE_GPR(2, r1) /* Save the TOC */
968 SAVE_GPR(13, r1) /* Save paca */
969 SAVE_8GPRS(14, r1) /* Save the non-volatiles */
970 SAVE_10GPRS(22, r1) /* ditto */
973 std r4,_CCR(r1)
975 std r5,_CTR(r1)
977 std r6,_XER(r1)
979 std r7,_DAR(r1)
981 std r8,_DSISR(r1)
1008 std r1,PACAR1(r13)
1053 ld r1,PACAR1(r4) /* Restore our SP */
1066 REST_GPR(2, r1) /* Restore the TOC */
1067 REST_GPR(13, r1) /* Restore paca */
1068 REST_8GPRS(14, r1) /* Restore the non-volatiles */
1069 REST_10GPRS(22, r1) /* ditto */
1073 ld r4,_CCR(r1)
1075 ld r5,_CTR(r1)
1077 ld r6,_XER(r1)
1079 ld r7,_DAR(r1)
1081 ld r8,_DSISR(r1)
1084 addi r1,r1,RTAS_FRAME_SIZE /* Unstack our frame */
1085 ld r0,16(r1) /* get return address */
1094 std r0,16(r1)
1095 stdu r1,-PROM_FRAME_SIZE(r1) /* Save SP and create stack space */
1101 SAVE_GPR(2, r1)
1102 SAVE_GPR(13, r1)
1103 SAVE_8GPRS(14, r1)
1104 SAVE_10GPRS(22, r1)
1107 std r10,_CCR(r1)
1108 std r11,_MSR(r1)
1135 /* Just make sure that r1 top 32 bits didn't get
1138 rldicl r1,r1,0,32
1141 ld r0,_MSR(r1)
1146 REST_GPR(2, r1)
1147 REST_GPR(13, r1)
1148 REST_8GPRS(14, r1)
1149 REST_10GPRS(22, r1)
1150 ld r4,_CCR(r1)
1153 addi r1,r1,PROM_FRAME_SIZE
1154 ld r0,16(r1)
1167 ld r11, 0(r1)
1168 stdu r1, -112(r1)
1169 std r3, 128(r1)
1182 ld r0, 128(r1)
1184 addi r1, r1, 112
1191 ld r11, 0(r1)
1192 stdu r1, -112(r1)
1193 std r3, 128(r1)
1208 ld r0, 128(r1)
1210 addi r1, r1, 112
1219 ld r4, 128(r1)
1223 ld r11, 112(r1)
1233 ld r11, 112(r1)
1236 ld r0, 128(r1)
1238 addi r1, r1, 112
1243 std r4, -32(r1)
1244 std r3, -24(r1)
1246 std r2, -16(r1)
1247 std r31, -8(r1)
1248 mr r31, r1
1249 stdu r1, -112(r1)
1263 ld r1, 0(r1)
1264 ld r4, -32(r1)
1265 ld r3, -24(r1)
1266 ld r2, -16(r1)
1267 ld r31, -8(r1)
H A Didle_power7.S44 std r0,0(r1); \
46 ld r0,0(r1); \
87 std r0,16(r1)
88 stdu r1,-INT_FRAME_SIZE(r1)
89 std r0,_LINK(r1)
90 std r0,_NIP(r1)
111 addi r1,r1,INT_FRAME_SIZE
112 ld r0,16(r1)
129 SAVE_GPR(2, r1)
130 SAVE_NVGPRS(r1)
132 std r4,_CCR(r1)
133 std r9,_MSR(r1)
134 std r1,PACAR1(r13)
221 std r3,_SDR1(r1)
223 std r3,_RPR(r1)
225 std r3,_SPURR(r1)
227 std r3,_PURR(r1)
229 std r3,_TSCR(r1)
231 std r3,_DSCR(r1)
233 std r3,_AMOR(r1)
235 std r3,_WORT(r1)
237 std r3,_WORC(r1)
278 ld r1,PACAR1(r13); \
279 std r3,ORIG_GPR3(r1); /* Save original r3 */ \
282 ld r3,ORIG_GPR3(r1); /* Restore original r3 */ \
288 ld r1,PACAR1(r13)
356 ld r4,_SDR1(r1)
358 ld r4,_RPR(r1)
360 ld r4,_AMOR(r1)
399 ld r4,_TSCR(r1)
401 ld r4,_WORC(r1)
437 ld r4,_SPURR(r1)
439 ld r4,_PURR(r1)
441 ld r4,_DSCR(r1)
443 ld r4,_WORT(r1)
464 REST_NVGPRS(r1)
465 REST_GPR(2, r1)
466 ld r3,_CCR(r1)
467 ld r4,_MSR(r1)
468 ld r5,_NIP(r1)
469 addi r1,r1,INT_FRAME_SIZE
488 ld r1,PACAR1(r13)
492 REST_NVGPRS(r1)
493 REST_GPR(2, r1)
494 ld r6,_CCR(r1)
495 ld r4,_MSR(r1)
496 ld r5,_NIP(r1)
497 addi r1,r1,INT_FRAME_SIZE
514 ld r1,PACAR1(r13)
515 ld r6,_CCR(r1)
516 ld r4,_MSR(r1)
517 ld r5,_NIP(r1)
518 addi r1,r1,INT_FRAME_SIZE
H A Didle_power4.S47 std r0,16(r1)
48 stdu r1,-128(r1)
50 addi r1,r1,128
51 ld r0,16(r1)
62 CURRENT_THREAD_INFO(r9, r1)
H A Dtm.S103 stw r6, 8(r1)
104 std r0, 16(r1)
105 std r2, STK_GOT(r1)
106 stdu r1, -TM_FRAME_SIZE(r1)
108 /* We've a struct pt_regs at [r1+STACK_FRAME_OVERHEAD]. */
110 std r3, STK_PARAM(R3)(r1)
111 SAVE_NVGPRS(r1)
133 std r14, TM_FRAME_L0(r1)
136 std r1, PACAR1(r13)
192 /* Stash the checkpointed r1 away in paca tm_scratch and get the real
195 std r1, PACATMSCRATCH(r13)
196 ld r1, PACAR1(r13)
199 std r11, GPR11(r1) /* Temporary stash */
204 std r7, GPR7(r1) /* Temporary stash */
205 std r12, GPR12(r1) /* '' '' '' */
206 ld r12, STK_PARAM(R3)(r1) /* Param 0, thread_struct * */
224 ld r3, PACATMSCRATCH(r13) /* user r1 */
225 ld r4, GPR7(r1) /* user r7 */
226 ld r5, GPR11(r1) /* user r11 */
227 ld r6, GPR12(r1) /* user r12 */
281 ld r14, TM_FRAME_L0(r1) /* Orig MSR */
286 REST_NVGPRS(r1)
288 addi r1, r1, TM_FRAME_SIZE
289 lwz r4, 8(r1)
290 ld r0, 16(r1)
293 ld r2, STK_GOT(r1)
313 stw r5, 8(r1)
314 std r0, 16(r1)
315 std r2, STK_GOT(r1)
316 stdu r1, -TM_FRAME_SIZE(r1)
318 /* We've a struct pt_regs at [r1+STACK_FRAME_OVERHEAD].
321 SAVE_NVGPRS(r1)
332 SET_SCRATCH0(r1)
419 * here before we load up the userspace r1 so any bugs we hit will get
460 GET_SCRATCH0(r1)
466 REST_NVGPRS(r1)
468 addi r1, r1, TM_FRAME_SIZE
469 lwz r4, 8(r1)
470 ld r0, 16(r1)
473 ld r2, STK_GOT(r1)
H A Didle_e500.S24 CURRENT_THREAD_INFO(r3, r1)
52 stwu r1,-16(r1)
54 stw r0,20(r1)
56 lwz r0,20(r1)
57 addi r1,r1,16
99 CURRENT_THREAD_INFO(r12, r1)
H A Didle_book3e.S28 std r0,16(r1)
45 stdu r1,-128(r1)
47 addi r1,r1,128
58 ld r0,16(r1)
65 CURRENT_THREAD_INFO(r11, r1)
H A Dexceptions-64e.S62 std reg, (SPECIAL_EXC_##name * 8 + SPECIAL_EXC_FRAME_OFFS)(r1)
65 ld reg, (SPECIAL_EXC_##name * 8 + SPECIAL_EXC_FRAME_OFFS)(r1)
75 ld r3,_MSR(r1)
82 CURRENT_THREAD_INFO(r12, r1)
144 ld r10,_NIP(r1)
146 ld r10,_MSR(r1)
152 ld r3,_MSR(r1)
180 REST_NVGPRS(r1)
210 ld r5,SOFTE(r1)
235 stdcx. r0,0,r1 /* to clear the reservation */
237 REST_4GPRS(2, r1)
238 REST_4GPRS(6, r1)
240 ld r10,_CTR(r1)
241 ld r11,_XER(r1)
250 ld r10,_LINK(r1)
251 ld r11,_CCR(r1)
252 ld r0,GPR13(r1)
256 ld r10,GPR10(r1)
257 ld r11,GPR11(r1)
258 ld r12,GPR12(r1)
263 ld r10,_NIP(r1)
264 ld r11,_MSR(r1)
265 ld r0,GPR0(r1)
266 ld r1,GPR1(r1)
293 std r1,PACA_EX##type+EX_R1(r13); /* save old r1 in the PACA */ \
297 ld r1,PACAKSAVE(r13); /* get kernel stack coming from usr */\
298 1: cmpdi cr1,r1,0; /* check if SP makes sense */ \
304 subi r1,r1,INT_FRAME_SIZE; /* alloc frame on kernel stack */
313 ld r1,PACA_CRIT_STACK(r13); \
314 subi r1,r1,SPECIAL_EXC_FRAME_SIZE
319 ld r1,PACA_DBG_STACK(r13); \
320 subi r1,r1,SPECIAL_EXC_FRAME_SIZE
325 ld r1,PACA_MC_STACK(r13); \
326 subi r1,r1,SPECIAL_EXC_FRAME_SIZE
381 std r0,GPR0(r1); /* save r0 in stackframe */ \
382 std r2,GPR2(r1); /* save r2 in stackframe */ \
383 SAVE_4GPRS(3, r1); /* save r3 - r6 in stackframe */ \
384 SAVE_2GPRS(7, r1); /* save r7, r8 in stackframe */ \
385 std r9,GPR9(r1); /* save r9 in stackframe */ \
386 std r10,_NIP(r1); /* save SRR0 to stackframe */ \
387 std r11,_MSR(r1); /* save SRR1 to stackframe */ \
393 std r12,GPR12(r1); /* save r12 in stackframe */ \
398 ld r9,excf+EX_R1(r13); /* load orig r1 back from PACA */ \
403 std r3,GPR10(r1); /* save r10 to stackframe */ \
404 std r4,GPR11(r1); /* save r11 to stackframe */ \
405 std r5,GPR13(r1); /* save it to stackframe */ \
406 std r6,_LINK(r1); \
407 std r7,_CTR(r1); \
408 std r8,_XER(r1); \
410 std r9,0(r1); /* store stack frame back link */ \
411 std r10,_CCR(r1); /* store orig CR in stackframe */ \
412 std r9,GPR1(r1); /* store stack frame back link */ \
413 std r11,SOFTE(r1); /* and save it to stackframe */ \
414 std r12,STACK_FRAME_OVERHEAD-16(r1); /* mark the frame */ \
415 std r3,_TRAP(r1); /* set trap number */ \
416 std r0,RESULT(r1); /* clear regs->result */
446 ld r11,_MSR(r1); \
452 li r1,(n); /* get exception number */ \
453 sth r1,PACA_TRAP_SAVE(r13); /* store trap */ \
482 CURRENT_THREAD_INFO(r11, r1); \
486 ld r8,_LINK(r1); \
488 std r8,_NIP(r1); \
500 addi r3,r1,STACK_FRAME_OVERHEAD; \
556 addi r3,r1,STACK_FRAME_OVERHEAD
568 addi r3,r1,STACK_FRAME_OVERHEAD
612 std r14,_DSISR(r1)
613 addi r3,r1,STACK_FRAME_OVERHEAD
625 ld r12,_MSR(r1)
632 addi r3,r1,STACK_FRAME_OVERHEAD
644 ld r12,_MSR(r1)
654 addi r3,r1,STACK_FRAME_OVERHEAD
666 addi r3,r1,STACK_FRAME_OVERHEAD
693 addi r3,r1,STACK_FRAME_OVERHEAD
716 addi r3,r1,STACK_FRAME_OVERHEAD
751 ld r1,PACA_EXCRIT+EX_R1(r13)
772 std r14,_DSISR(r1)
773 addi r3,r1,STACK_FRAME_OVERHEAD
815 ld r1,PACA_EXDBG+EX_R1(r13)
837 std r14,_DSISR(r1)
838 addi r3,r1,STACK_FRAME_OVERHEAD
852 addi r3,r1,STACK_FRAME_OVERHEAD
868 addi r3,r1,STACK_FRAME_OVERHEAD
880 addi r3,r1,STACK_FRAME_OVERHEAD
894 addi r3,r1,STACK_FRAME_OVERHEAD
903 addi r3,r1,STACK_FRAME_OVERHEAD
914 addi r3,r1,STACK_FRAME_OVERHEAD
925 addi r3,r1,STACK_FRAME_OVERHEAD
991 std r1,PACA_EXGEN+EX_R1(r13);
994 subi r1,r1,INT_FRAME_SIZE;
1010 std r14,_DAR(r1)
1011 std r15,_DSISR(r1)
1012 addi r3,r1,STACK_FRAME_OVERHEAD
1023 addi r3,r1,STACK_FRAME_OVERHEAD
1024 ld r4,_DAR(r1)
1033 std r14,_DAR(r1)
1034 std r15,_DSISR(r1)
1035 addi r3,r1,STACK_FRAME_OVERHEAD
1057 ld r10,_MSR(r1)
1058 REST_4GPRS(2, r1)
1060 REST_2GPRS(6, r1)
1063 ld r0,GPR13(r1)
1065 1: stdcx. r0,0,r1 /* to clear the reservation */
1067 ld r8,_CCR(r1)
1068 ld r9,_LINK(r1)
1069 ld r10,_CTR(r1)
1070 ld r11,_XER(r1)
1075 REST_2GPRS(8, r1)
1076 ld r10,GPR10(r1)
1077 ld r11,GPR11(r1)
1078 ld r12,GPR12(r1)
1083 ld r10,_NIP(r1)
1084 ld r11,_MSR(r1)
1085 ld r0,GPR0(r1)
1086 ld r1,GPR1(r1)
1135 ld r1,PACAEMERGSP(r13)
1136 subi r1,r1,64+INT_FRAME_SIZE
1137 std r10,_NIP(r1)
1138 std r11,_MSR(r1)
1141 std r10,GPR1(r1)
1142 std r11,_CCR(r1)
1145 std r10,_DAR(r1)
1146 std r11,_DSISR(r1)
1147 std r0,GPR0(r1); /* save r0 in stackframe */ \
1148 std r2,GPR2(r1); /* save r2 in stackframe */ \
1149 SAVE_4GPRS(3, r1); /* save r3 - r6 in stackframe */ \
1150 SAVE_2GPRS(7, r1); /* save r7, r8 in stackframe */ \
1151 std r9,GPR9(r1); /* save r9 in stackframe */ \
1155 std r3,GPR10(r1); /* save r10 to stackframe */ \
1156 std r4,GPR11(r1); /* save r11 to stackframe */ \
1157 std r12,GPR12(r1); /* save r12 in stackframe */ \
1158 std r5,GPR13(r1); /* save it to stackframe */ \
1162 std r10,_LINK(r1)
1163 std r11,_CTR(r1)
1164 std r12,_XER(r1)
1165 SAVE_10GPRS(14,r1)
1166 SAVE_8GPRS(24,r1)
1168 std r12,_TRAP(r1)
1169 addi r11,r1,INT_FRAME_SIZE
1170 std r11,0(r1)
1174 1: addi r3,r1,STACK_FRAME_OVERHEAD
H A Dswsusp_asm64.S90 SAVE_REGISTER(r1)
129 addi r1,r1,-128
137 addi r1,r1,128
197 ld r1, SL_TB(r11)
199 srdi r2, r1, 32
206 mttbl r1
209 RESTORE_REGISTER(r1)
262 addi r1,r1,-128
267 addi r1,r1,128
H A Dhead_32.S82 * pointer (r1) points to just below the end of the half-meg region
259 tophys(r11,r1); /* use tophys(r1) if kernel */ \
281 stw r1,GPR1(r11); \
282 stw r1,0(r11); \
283 tovirt(r1,r11); /* set new kernel sp */ \
297 * Note2: once we have set r1 we are in a position to take exceptions
309 addi r3,r1,STACK_FRAME_OVERHEAD; \
357 * (Other exception handlers assume that r1 is a valid kernel stack
373 addi r3,r1,STACK_FRAME_OVERHEAD
427 addi r3,r1,STACK_FRAME_OVERHEAD
448 1: addi r3,r1,STACK_FRAME_OVERHEAD
492 * r1: linux style pte ( later becomes ppc hardware pte )
498 lis r1,PAGE_OFFSET@h /* check if kernel address */
499 cmplw 0,r1,r3
501 li r1,_PAGE_USER|_PAGE_PRESENT /* low addresses tested as user */
505 rlwimi r1,r2,32-12,29,29 /* shift MSR_PR to _PAGE_USER posn */
515 andc. r1,r1,r0 /* check access & ~permission */
524 rlwinm r1,r0,32-10,31,31 /* _PAGE_RW -> PP lsb */
526 and r1,r1,r2 /* writable if _RW and _DIRTY */
529 ori r1,r1,0xe04 /* clear out reserved bits */
530 andc r1,r0,r1 /* PP = user? (rw&dirty? 2: 3): 0 */
532 rlwinm r1,r1,0,~_PAGE_COHERENT /* clear M (coherence not required) */
534 mtspr SPRN_RPA,r1
541 rlwinm r1,r3,9,6,6 /* Get load/store bit */
543 addis r1,r1,0x2000
544 mtspr SPRN_DSISR,r1 /* (shouldn't be needed) */
546 or r2,r2,r1
548 mfspr r1,SPRN_IMISS /* Get failing address */
551 xor r1,r1,r2
552 mtspr SPRN_DAR,r1 /* Set fault address */
566 * r1: linux style pte ( later becomes ppc hardware pte )
572 lis r1,PAGE_OFFSET@h /* check if kernel address */
573 cmplw 0,r1,r3
575 li r1,_PAGE_USER|_PAGE_PRESENT /* low addresses tested as user */
579 rlwimi r1,r2,32-12,29,29 /* shift MSR_PR to _PAGE_USER posn */
589 andc. r1,r1,r0 /* check access & ~permission */
598 rlwinm r1,r0,32-10,31,31 /* _PAGE_RW -> PP lsb */
600 and r1,r1,r2 /* writable if _RW and _DIRTY */
603 ori r1,r1,0xe04 /* clear out reserved bits */
604 andc r1,r0,r1 /* PP = user? (rw&dirty? 2: 3): 0 */
606 rlwinm r1,r1,0,~_PAGE_COHERENT /* clear M (coherence not required) */
608 mtspr SPRN_RPA,r1
613 mfspr r1,SPRN_SPRG_603_LRU
616 xor r1,r0,r1
617 srw r0,r1,r2
618 mtspr SPRN_SPRG_603_LRU,r1
627 rlwinm r1,r3,9,6,6 /* Get load/store bit */
628 addis r1,r1,0x2000
629 mtspr SPRN_DSISR,r1
632 mfspr r1,SPRN_DMISS /* Get failing address */
635 xori r1,r1,3
636 20: mtspr SPRN_DAR,r1 /* Set fault address */
650 * r1: linux style pte ( later becomes ppc hardware pte )
656 lis r1,PAGE_OFFSET@h /* check if kernel address */
657 cmplw 0,r1,r3
659 li r1,_PAGE_RW|_PAGE_USER|_PAGE_PRESENT /* access flags */
663 rlwimi r1,r2,32-12,29,29 /* shift MSR_PR to _PAGE_USER posn */
673 andc. r1,r1,r0 /* check access & ~permission */
683 li r1,0xe05 /* clear out reserved bits & PP lsb */
684 andc r1,r0,r1 /* PP = user? 2: 0 */
686 rlwinm r1,r1,0,~_PAGE_COHERENT /* clear M (coherence not required) */
688 mtspr SPRN_RPA,r1
693 mfspr r1,SPRN_SPRG_603_LRU
696 xor r1,r0,r1
697 srw r0,r1,r2
698 mtspr SPRN_SPRG_603_LRU,r1
752 1: addi r3,r1,STACK_FRAME_OVERHEAD
757 addi r3,r1,STACK_FRAME_OVERHEAD
847 lis r1,secondary_ti@ha
848 tophys(r1,r1)
849 lwz r1,secondary_ti@l(r1)
850 tophys(r2,r1)
854 addi r1,r1,THREAD_SIZE-STACK_FRAME_OVERHEAD
856 tophys(r3,r1)
958 lis r1,init_thread_union@ha
959 addi r1,r1,init_thread_union@l
961 stwu r0,THREAD_SIZE-STACK_FRAME_OVERHEAD(r1)
H A Dexceptions-64s.S472 * paca->mc_emergency_sp, otherwise r1 is already pointing to
479 mr r11,r1 /* Save r1 */
484 ld r1,PACAMCEMERGSP(r13) /* Use MC emergency stack */
485 0: subi r1,r1,INT_FRAME_SIZE /* alloc stack frame */
491 std r11,GPR1(r1) /* Save r1 on the stack. */
492 std r11,0(r1) /* make stack chain pointer */
494 std r11,_NIP(r1)
496 std r11,_MSR(r1)
498 std r11,_DAR(r1)
500 std r11,_DSISR(r1)
501 std r9,_CCR(r1) /* Save CR in stackframe */
518 addi r1,r1,INT_FRAME_SIZE /* go back to previous stack frame */
994 ld r12,_MSR(r1)
1010 addi r3,r1,STACK_FRAME_OVERHEAD
1019 ld r12,_MSR(r1)
1020 ld r3,_NIP(r1)
1076 std r4,_DAR(r1)
1077 std r5,_DSISR(r1)
1084 1: addi r3,r1,STACK_FRAME_OVERHEAD
1108 std r3,_DAR(r1)
1109 std r4,_DSISR(r1)
1111 addi r3,r1,STACK_FRAME_OVERHEAD
1125 std r3,_DAR(r1)
1126 std r4,_DSISR(r1)
1129 addi r3,r1,STACK_FRAME_OVERHEAD
1139 addi r3,r1,STACK_FRAME_OVERHEAD
1150 addi r3,r1,STACK_FRAME_OVERHEAD
1169 addi r3,r1,STACK_FRAME_OVERHEAD
1195 addi r3,r1,STACK_FRAME_OVERHEAD
1204 addi r3,r1,STACK_FRAME_OVERHEAD
1229 addi r3,r1,STACK_FRAME_OVERHEAD
1238 addi r3,r1,STACK_FRAME_OVERHEAD
1288 mr r10,r1 /* Save r1 */
1289 ld r1,PACAEMERGSP(r13) /* Use emergency stack */
1290 subi r1,r1,INT_FRAME_SIZE /* alloc stack frame */
1291 std r9,_CCR(r1) /* save CR in stackframe */
1293 std r11,_NIP(r1) /* save HSRR0 in stackframe */
1295 std r12,_MSR(r1) /* save SRR1 in stackframe */
1296 std r10,0(r1) /* make stack chain pointer */
1297 std r0,GPR0(r1) /* save r0 in stackframe */
1298 std r10,GPR1(r1) /* save r1 in stackframe */
1301 addi r3,r1,STACK_FRAME_OVERHEAD
1305 ld r9,_MSR(r1)
1307 ld r3,_NIP(r1)
1309 ld r9,_CTR(r1)
1311 ld r9,_XER(r1)
1313 ld r9,_LINK(r1)
1315 REST_GPR(0, r1)
1316 REST_8GPRS(2, r1)
1317 REST_GPR(10, r1)
1318 ld r11,_CCR(r1)
1320 REST_GPR(11, r1)
1321 REST_2GPRS(12, r1)
1322 /* restore original r1. */
1323 ld r1,GPR1(r1)
1343 ld r9,_MSR(r1); \
1345 ld r3,_NIP(r1); \
1347 ld r9,_CTR(r1); \
1349 ld r9,_XER(r1); \
1351 ld r9,_LINK(r1); \
1353 REST_GPR(0, r1); \
1354 REST_8GPRS(2, r1); \
1355 REST_GPR(10, r1); \
1356 ld r11,_CCR(r1); \
1362 REST_GPR(11, r1); \
1363 REST_2GPRS(12, r1); \
1364 /* restore original r1. */ \
1365 ld r1,GPR1(r1)
1374 std r0,GPR0(r1) /* Save r0 */
1377 addi r3,r1,STACK_FRAME_OVERHEAD
1379 std r3,RESULT(r1) /* Save result */
1380 ld r12,_MSR(r1)
1402 ld r1,PACAR1(r13)
1457 ld r3,RESULT(r1) /* Load result */
1476 addi r3,r1,STACK_FRAME_OVERHEAD
1482 1: addi r3,r1,STACK_FRAME_OVERHEAD
1543 1: addi r3,r1,STACK_FRAME_OVERHEAD
1552 ld r10,_LINK(r1) /* make idle task do the */
1553 std r10,_NIP(r1) /* equivalent of a blr */
1562 std r3,_DAR(r1)
1563 std r4,_DSISR(r1)
1569 CURRENT_THREAD_INFO(r11, r1)
1593 ld r6,_DSISR(r1)
1605 11: ld r4,_DAR(r1)
1606 ld r5,_DSISR(r1)
1607 addi r3,r1,STACK_FRAME_OVERHEAD
1613 addi r3,r1,STACK_FRAME_OVERHEAD
1614 lwz r4,_DAR(r1)
1621 ld r4,_DAR(r1)
1622 ld r5,_DSISR(r1)
1623 addi r3,r1,STACK_FRAME_OVERHEAD
1633 addi r3,r1,STACK_FRAME_OVERHEAD
1634 ld r4,_DAR(r1)
1647 addi r3,r1,STACK_FRAME_OVERHEAD
1661 ld r1,PACAEMERGSP(r13)
1662 subi r1,r1,64+INT_FRAME_SIZE
1663 std r9,_CCR(r1)
1664 std r10,GPR1(r1)
1665 std r11,_NIP(r1)
1666 std r12,_MSR(r1)
1669 std r11,_DAR(r1)
1670 std r12,_DSISR(r1)
1674 std r10,_LINK(r1)
1675 std r11,_CTR(r1)
1676 std r12,_XER(r1)
1677 SAVE_GPR(0,r1)
1678 SAVE_GPR(2,r1)
1680 std r10,GPR3(r1)
1681 SAVE_GPR(4,r1)
1682 SAVE_4GPRS(5,r1)
1685 SAVE_2GPRS(9,r1)
1689 std r9,GPR11(r1)
1690 std r10,GPR12(r1)
1691 std r11,GPR13(r1)
1694 std r10,ORIG_GPR3(r1)
1696 SAVE_8GPRS(14,r1)
1697 SAVE_10GPRS(22,r1)
1699 std r12,_TRAP(r1)
1700 addi r11,r1,INT_FRAME_SIZE
1701 std r11,0(r1)
1706 std r12,RESULT(r1)
1707 std r11,STACK_FRAME_OVERHEAD-16(r1)
1708 1: addi r3,r1,STACK_FRAME_OVERHEAD
/linux-4.1.27/arch/s390/kernel/vdso32/
H A Dclock_getres.S22 basr %r1,0
23 la %r1,4f-.(%r1)
28 la %r1,5f-4f(%r1)
35 1: l %r0,0(%r1)
40 3: lhi %r1,__NR_clock_getres /* fallback to svc */
H A Dclock_gettime.S39 lm %r0,%r1,1(%r15)
41 sl %r1,__VDSO_XTIME_STAMP+4(%r5)
47 ltr %r1,%r1
53 al %r1,__VDSO_WTOM_NSEC+4(%r5)
64 cl %r1,20f-6b(%r5)
67 sl %r1,20f-6b(%r5)
72 st %r1,4(%r3) /* store tp->tv_nsec */
82 l %r1,__VDSO_WTOM_CRS_NSEC+4(%r5)
92 l %r1,__VDSO_XTIME_CRS_NSEC+4(%r5)
102 lm %r0,%r1,1(%r15)
104 sl %r1,__VDSO_XTIME_STAMP+4(%r5)
110 ltr %r1,%r1
116 al %r1,__VDSO_XTIME_NSEC+4(%r5)
127 cl %r1,20f-15b(%r5)
130 sl %r1,20f-15b(%r5)
135 st %r1,4(%r3) /* store tp->tv_nsec */
141 19: lhi %r1,__NR_clock_gettime
H A Dgettimeofday.S34 lm %r0,%r1,1(%r15)
36 sl %r1,__VDSO_XTIME_STAMP+4(%r5)
42 ltr %r1,%r1
48 al %r1,__VDSO_XTIME_NSEC+4(%r5)
60 cl %r1,11f-6b(%r5)
63 sl %r1,11f-6b(%r5)
68 ltr %r1,%r1
/linux-4.1.27/arch/m32r/boot/
H A Dsetup.S67 ldi r1, #0x0101 ; cache on (with invalidation)
68 ; ldi r1, #0x00 ; cache off
69 sth r1, @r0
73 ldi r1, #0x73 ; cache on (with invalidation)
74 ; ldi r1, #0x00 ; cache off
75 st r1, @r0
78 ldi r1, #0x101 ; cache on (with invalidation)
79 ; ldi r1, #0x00 ; cache off
80 st r1, @r0
83 seth r1, #0x0060 ; from 0x00600000
84 or3 r1, r1, #0x0005 ; size 2MB
85 st r1, @r0
86 seth r1, #0x0100 ; from 0x01000000
87 or3 r1, r1, #0x0003 ; size 16MB
88 st r1, @+r0
89 seth r1, #0x0200 ; from 0x02000000
90 or3 r1, r1, #0x0002 ; size 32MB
91 st r1, @+r0
93 ldi r1, #0x703 ; cache on (with invalidation)
94 st r1, @r0
126 ldi r1, #0x01
127 st r1, @(MATM_offset,r0) ; Set MATM (T bit ON)
133 ld24 r1,#0x8080
134 st r1,@r0
/linux-4.1.27/arch/sh/kernel/cpu/sh3/
H A Dswsusp.S17 #define k1 r1
44 mov.l @r2+,r1 /* 16n+0 */
45 mov.l r1,@r5
47 mov.l @r2+,r1 /* 16n+4 */
48 mov.l r1,@r5
50 mov.l @r2+,r1 /* 16n+8 */
51 mov.l r1,@r5
53 mov.l @r2+,r1 /* 16n+12 */
54 mov.l r1,@r5
73 mov.l @r15+, r1
100 stc sr, r1
101 ldc r1, ssr ! save sr in ssr
102 mov.l 1f, r1
103 ldc r1, spc ! setup pc value for resuming
111 mov.l 7f, r1
112 jsr @r1 ! switch to bank1 and save bank1 r7->r0
125 stc ssr, r1
126 ldc r1, sr ! restore old sr
/linux-4.1.27/arch/hexagon/kernel/
H A Dhead.S53 r1.h = #HI(PAGE_OFFSET);
54 r1.l = #LO(PAGE_OFFSET);
55 r24 = sub(r24,r1); /* swapper_pg_dir - PAGE_OFFSET */
75 r1.l = #LO(_end);
80 r1.h = #HI(_end);
85 r1 = sub(r1, r2); define
87 } /* r1 = _end - stext */
88 r1 = add(r1, r3); /* + (4M-1) */ define
89 r26 = lsr(r1, #22); /* / 4M = # of entries */
91 r1 = r25; define
94 r1 = and(r1,r2); define
95 r2 = lsr(r1, #22) /* 4MB page number */
99 r1 = add(r1,r2) /* r1 = 4MB PTE for the first entry */ define
104 memw(r0 ++ #4) = r1
105 { r1 = add(r1, r2); } :endloop0
112 r0 = add(r1, r24); /* advance to 0xc0000000 entry */
113 r1 = r25; define
116 r1 = and(r1,r2); /* for huge page */ define
118 r1 = add(r1,r2); define
124 memw(r0 ++ #4) = r1;
125 { r1 = add(r1,r2); } :endloop0
135 r1 = #VM_TRANS_TYPE_TABLE define
153 r1.h = #0xffc0;
154 r1.l = #0x0000;
156 r2 = and(r1,r2);
158 r1.l = #lo(swapper_pg_dir)
159 r1.h = #hi(swapper_pg_dir)
162 r1 = add(r1,r2); define
171 r1 = #VM_TRANS_TYPE_TABLE define
208 { r0.L = #LO(__bss_start); r1 = #0; r2.l = #LO(__bss_stop); }
/linux-4.1.27/arch/microblaze/lib/
H A Ddivsi3.S14 .frame r1, 0, r15
16 addik r1, r1, -16
17 swi r28, r1, 0
18 swi r29, r1, 4
19 swi r30, r1, 8
20 swi r31, r1, 12
65 lwi r28, r1, 0
66 lwi r29, r1, 4
67 lwi r30, r1, 8
68 lwi r31, r1, 12
70 addik r1, r1, 16
H A Dmodsi3.S16 .frame r1, 0, r15
18 addik r1, r1, -16
19 swi r28, r1, 0
20 swi r29, r1, 4
21 swi r30, r1, 8
22 swi r31, r1, 12
65 lwi r28, r1, 0
66 lwi r29, r1, 4
67 lwi r30, r1, 8
68 lwi r31, r1, 12
70 addik r1, r1, 16
H A Duaccess_old.S192 addik r1, r1, -40
193 swi r5, r1, 0
194 swi r6, r1, 4
195 swi r7, r1, 8
196 swi r19, r1, 12
197 swi r20, r1, 16
198 swi r21, r1, 20
199 swi r22, r1, 24
200 swi r23, r1, 28
201 swi r24, r1, 32
202 swi r25, r1, 36
216 lwi r5, r1, 0
217 lwi r6, r1, 4
218 lwi r7, r1, 8
219 lwi r19, r1, 12
220 lwi r20, r1, 16
221 lwi r21, r1, 20
222 lwi r22, r1, 24
223 lwi r23, r1, 28
224 lwi r24, r1, 32
225 lwi r25, r1, 36
226 addik r1, r1, 40
236 lwi r5, r1, 0
237 lwi r6, r1, 4
238 lwi r7, r1, 8
239 lwi r19, r1, 12
240 lwi r20, r1, 16
241 lwi r21, r1, 20
242 lwi r22, r1, 24
243 lwi r23, r1, 28
244 lwi r24, r1, 32
245 lwi r25, r1, 36
246 addik r1, r1, 40
H A Dudivsi3.S17 .frame r1, 0, r15
19 addik r1, r1, -12
20 swi r29, r1, 0
21 swi r30, r1, 4
22 swi r31, r1, 8
77 lwi r29, r1, 0
78 lwi r30, r1, 4
79 lwi r31, r1, 8
81 addik r1, r1, 12
H A Dumodsi3.S16 .frame r1, 0, r15
18 addik r1, r1, -12
19 swi r29, r1, 0
20 swi r30, r1, 4
21 swi r31, r1, 8
79 lwi r29, r1, 0
80 lwi r30, r1, 4
81 lwi r31, r1, 8
83 addik r1, r1, 12
/linux-4.1.27/arch/powerpc/boot/
H A Dps3-hvcall.S37 * 5: 8(r1),12(r1) <-> r7
38 * 6: 16(r1),20(r1) <-> r8
39 * 7: 24(r1),28(r1) <-> r9
40 * 8: 32(r1),36(r1) <-> r10
64 stw r0, \offset(r1)
68 lwz r0, \offset(r1)
78 ld \target, \offset(r1)
138 lwz r11, 8(r1)
142 addi r1,r1,16
146 lwz r11, 16(r1)
148 lwz r11, 20(r1)
153 lwz r11, 24(r1)
H A Dcrt0.S136 lwz r1,0(r8)
138 stwu r0,-16(r1) /* establish a stack frame */
219 ld r1,0(r8)
221 stdu r0,-112(r1) /* establish a stack frame */
249 std r0,16(r1)
250 stdu r1,-PROM_FRAME_SIZE(r1) /* Save SP and create stack space */
252 SAVE_GPR(2, r1)
253 SAVE_GPR(13, r1)
254 SAVE_8GPRS(14, r1)
255 SAVE_10GPRS(22, r1)
257 std r10,8*32(r1)
259 std r10,8*33(r1)
281 rldicl r1,r1,0,32
284 ld r10,8*(33)(r1)
289 REST_GPR(2, r1)
290 REST_GPR(13, r1)
291 REST_8GPRS(14, r1)
292 REST_10GPRS(22, r1)
293 ld r10,8*32(r1)
296 addi r1,r1,PROM_FRAME_SIZE
297 ld r0,16(r1)
/linux-4.1.27/arch/arm/mach-exynos/
H A Dsleep.S46 ldr r1, =CPU_MASK
47 and r0, r0, r1
48 ldr r1, =CPU_CORTEX_A9
49 cmp r0, r1
59 ldr r1, =CPU_MASK
60 and r0, r0, r1
61 ldr r1, =CPU_CORTEX_A9
62 cmp r0, r1
66 ldr r1, [r0]
67 ldr r1, [r0, r1]
80 ldr r1, [r0, #L2X0_R_PHY_BASE]
81 teq r1, #0
85 ldr r2, [r1, #L2X0_CTRL]
89 ldr r1, [r0, #L2X0_R_TAG_LATENCY]
100 ldr r1, [r0, #L2X0_R_PWR_CTRL]
108 mov r1, #1
/linux-4.1.27/arch/powerpc/platforms/powernv/
H A Dopal-wrappers.S62 std r0,16(r1); \
66 stw r12,8(r1); \
67 std r1,PACAR1(r13); \
94 lwz r4,8(r1);
95 ld r5,16(r1);
104 stdu r1,-STACKFRAMESIZE(r1)
105 std r0,STK_REG(R23)(r1)
106 std r3,STK_REG(R24)(r1)
107 std r4,STK_REG(R25)(r1)
108 std r5,STK_REG(R26)(r1)
109 std r6,STK_REG(R27)(r1)
110 std r7,STK_REG(R28)(r1)
111 std r8,STK_REG(R29)(r1)
112 std r9,STK_REG(R30)(r1)
113 std r10,STK_REG(R31)(r1)
115 addi r4,r1,STK_REG(R24)
117 ld r0,STK_REG(R23)(r1)
118 ld r3,STK_REG(R24)(r1)
119 ld r4,STK_REG(R25)(r1)
120 ld r5,STK_REG(R26)(r1)
121 ld r6,STK_REG(R27)(r1)
122 ld r7,STK_REG(R28)(r1)
123 ld r8,STK_REG(R29)(r1)
124 ld r9,STK_REG(R30)(r1)
125 ld r10,STK_REG(R31)(r1)
128 std r11,16(r1)
129 stw r12,8(r1)
130 std r1,PACAR1(r13)
149 std r3,STK_REG(R31)(r1)
151 ld r0,STK_REG(R23)(r1)
153 ld r3,STK_REG(R31)(r1)
154 addi r1,r1,STACKFRAMESIZE
155 ld r0,16(r1)
165 * r1 - stack pointer
170 std r12,PPC_LR_STKOFF(r1)
193 ld r12,PPC_LR_STKOFF(r1)
/linux-4.1.27/arch/x86/crypto/
H A Daes-x86_64-asm_64.S52 #define prologue(FUNC,KEY,B128,B192,r1,r2,r3,r4,r5,r6,r7,r8,r9,r10,r11) \
54 movq r1,r2; \
59 movl 4(r7),r1 ## E; \
64 xorl -44(r9),r1 ## E; \
73 #define epilogue(FUNC,r1,r2,r3,r4,r5,r6,r7,r8,r9) \
74 movq r1,r2; \
83 #define round(TAB,OFFSET,r1,r2,r3,r4,r5,r6,r7,r8,ra,rb,rc,rd) \
97 movzbl r1 ## L,r7 ## E; \
98 movzbl r1 ## H,r4 ## E; \
100 movw r3 ## X,r1 ## X; \
101 roll $16,r1 ## E; \
108 movzbl r1 ## H,r7 ## E; \
109 movzbl r1 ## L,r3 ## E; \
110 shrl $16,r1 ## E; \
113 movzbl r1 ## H,r7 ## E; \
114 movzbl r1 ## L,r1 ## E; \
116 xorl TAB(,r1,4),r3 ## E; \
117 movzbl r2 ## H,r1 ## E; \
120 xorl TAB+3072(,r1,4),r3 ## E;\
122 movzbl r2 ## H,r1 ## E; \
126 xorl TAB+1024(,r1,4),r3 ## E;\
129 #define move_regs(r1,r2,r3,r4) \
130 movl r3 ## E,r1 ## E; \
H A Daes-i586-asm_32.S57 #define r1 ebx define
81 // output registers r0, r1, r4 or r5.
151 // the first previous round column values in r0,r1,r4,r5 and
156 // on entry: r0,r1,r4,r5
157 // on exit: r2,r1,r4,r5
159 save (0,r1); \
163 do_fcol(table, r2,r5,r4,r1, r0,r3, arg); /* idx=r0 */ \
164 do_col (table, r4,r1,r2,r5, r0,r3); /* idx=r4 */ \
166 do_col (table, r1,r2,r5,r4, r0,r3); /* idx=r1 */ \
168 do_col (table, r5,r4,r1,r2, r0,r3); /* idx=r5 */
171 // on entry: r2,r1,r4,r5
172 // on exit: r0,r1,r4,r5
174 save (0,r1); \
178 do_fcol(table, r0,r5,r4,r1, r2,r3, arg); /* idx=r2 */ \
179 do_col (table, r4,r1,r0,r5, r2,r3); /* idx=r4 */ \
181 do_col (table, r1,r0,r5,r4, r2,r3); /* idx=r1 */ \
183 do_col (table, r5,r4,r1,r0, r2,r3); /* idx=r5 */
186 // the first previous round column values in r0,r1,r4,r5 and
191 // on entry: r0,r1,r4,r5
192 // on exit: r2,r1,r4,r5
194 save (0,r1); \
198 do_icol(table, r2,r1,r4,r5, r0,r3, arg); /* idx=r0 */ \
199 do_col (table, r4,r5,r2,r1, r0,r3); /* idx=r4 */ \
201 do_col (table, r1,r4,r5,r2, r0,r3); /* idx=r1 */ \
203 do_col (table, r5,r2,r1,r4, r0,r3); /* idx=r5 */
206 // on entry: r2,r1,r4,r5
207 // on exit: r0,r1,r4,r5
209 save (0,r1); \
213 do_icol(table, r0,r1,r4,r5, r2,r3, arg); /* idx=r2 */ \
214 do_col (table, r4,r5,r0,r1, r2,r3); /* idx=r4 */ \
216 do_col (table, r1,r4,r5,r0, r2,r3); /* idx=r1 */ \
218 do_col (table, r5,r0,r1,r4, r2,r3); /* idx=r5 */
245 mov 4(%r2),%r1
249 xor 4(%ebp),%r1
285 mov %r1,4(%ebp)
317 mov 4(%r2),%r1
321 xor 4(%ebp),%r1
357 mov %r1,4(%ebp)
/linux-4.1.27/arch/arm/kvm/
H A Dinit.S31 * r1 = pointer to hyp vectors
48 * - Phase 1: r0 = 0, r1 = 0, r2,r3 contain the boot PGD.
50 * - Phase 2: r0 = ToS, r1 = vectors, r2,r3 contain the runtime PGD.
82 mrc p15, 0, r1, c2, c0, 2 @ TTBCR
83 and r1, r1, #(HTCR_MASK & ~TTBCR_T0SZ)
84 orr r0, r0, r1
87 mrc p15, 4, r1, c2, c1, 2 @ VTCR
89 bic r1, r1, r2
91 orr r1, r0, r1
92 orr r1, r1, #(KVM_VTCR_SL0 | KVM_VTCR_T0SZ | KVM_VTCR_S)
93 mcr p15, 4, r1, c2, c1, 2 @ VTCR
118 mrc p15, 0, r1, c1, c0, 0 @ SCTLR
120 and r1, r1, r2
123 orr r1, r1, r2
124 orr r0, r0, r1
136 mcr p15, 4, r1, c12, c0, 0 @ HVBAR
140 adr r1, target
141 bfi r0, r1, #0, #PAGE_SHIFT
H A Dinterrupts.S138 ldr r1, [vcpu, #VCPU_MIDR]
139 mcr p15, 4, r1, c0, c0, 0
142 ldr r1, [vcpu, #CP15_OFFSET(c0_MPIDR)]
143 mcr p15, 4, r1, c0, c0, 5
146 ldr r1, [vcpu, #VCPU_KVM]
147 add r1, r1, #KVM_VTTBR
148 ldrd r2, r3, [r1]
159 * guest r0, r1, r2 saved on the stack
161 * r1: exception code
212 mov r0, r1 @ Return the return code
213 mov r1, #0 @ Clear upper bits in return value
215 @ r1 already has return code
232 * passed as r0, r1, and r2 (a maximum of 3 arguments in addition to the
235 * passed in r0 and r1.
270 * - r1, r2, and r3 contain arguments to the above function.
271 * - The HYP function will be called with its arguments in r0, r1 and r2.
282 mrrc p15, 6, r0, r1, c2 @ Read VTTBR
283 lsr r1, r1, #16
284 ands r1, r1, #0xff
290 mrc p15, 4, r1, c6, c0, 0 @ HDFAR
292 str r1, [vcpu, #VCPU_HxFAR]
296 mrc p15, 4, r1, c6, c0, 2 @ HIFAR
298 str r1, [vcpu, #VCPU_HxFAR]
300 mov r1, #\exception_code
309 mrs r1, ELR_hyp
359 push {r0, r1, r2}
362 mrc p15, 4, r1, c5, c2, 0 @ HSR
363 lsr r0, r1, #HSR_EC_SHIFT
382 pop {r0, r1, r2}
394 mov r0, r1
395 mov r1, r2
408 str r1, [vcpu, #VCPU_HSR]
411 lsr r1, r1, #HSR_EC_SHIFT
412 cmp r1, #HSR_EC_IABT
415 cmp r1, #HSR_EC_DABT
435 mrc p15, 4, r1, c5, c2, 0 @ HSR
436 and r0, r1, #HSR_FSC_TYPE
438 tsteq r1, #(1 << 7) @ S1PTW
443 mrrc p15, 0, r0, r1, c7 @ PAR
444 push {r0, r1}
449 mrrc p15, 0, r0, r1, c7 @ PAR
454 orr r2, r2, r1, lsl #24
457 pop {r0, r1}
458 mcrr p15, 0, r0, r1, c7 @ PAR
463 1: mov r1, #ARM_EXCEPTION_HVC
466 4: pop {r0, r1} @ Failed translation, return to guest
467 mcrr p15, 0, r0, r1, c7 @ PAR
469 pop {r0, r1, r2}
501 push {r0, r1, r2}
502 mov r1, #ARM_EXCEPTION_IRQ
/linux-4.1.27/arch/arm/mach-mvebu/
H A Dcoherency_ll.S28 * Returns the coherency base address in r1 (r0 is untouched), or 0 if
32 mrc p15, 0, r1, c1, c0, 0
33 tst r1, #CR_M @ Check MMU bit enabled
42 ldr r1, =coherency_base
43 cmp r1, #0
45 adr r1, 3f
46 ldr r3, [r1]
47 ldr r1, [r1, r3]
54 ldr r1, =coherency_base
55 ldr r1, [r1]
97 cmp r1, #0
101 add r0, r1, #ARMADA_XP_CFB_CFG_REG_OFFSET
105 strex r1, r2, [r0]
106 cmp r1, #0
122 cmp r1, #0
126 add r0, r1, #ARMADA_XP_CFB_CTL_REG_OFFSET
130 strex r1, r2, [r0]
131 cmp r1, #0
149 cmp r1, #0
153 add r0, r1, #ARMADA_XP_CFB_CTL_REG_OFFSET
157 strex r1, r2, [r0]
158 cmp r1, #0
H A Dpmsu_ll.S17 mrc p15, 4, r1, c15, c0 @ get SCU base address
18 orr r1, r1, #0x8 @ SCU CPU Power Status Register
21 add r1, r1, r0
23 strb r0, [r1] @ switch SCU power state to Normal mode
38 mrc p15, 0, r1, c1, c0, 0
39 bic r1, #1
40 mcr p15, 0, r1, c1, c0, 0
/linux-4.1.27/arch/cris/arch-v32/mach-fs/
H A Ddram_init.S29 move.d CONFIG_ETRAX_SDRAM_GRP0_CONFIG, $r1
30 move.d $r1, [$r0]
32 move.d CONFIG_ETRAX_SDRAM_GRP1_CONFIG, $r1
33 move.d $r1, [$r0]
47 move.d CONFIG_ETRAX_SDRAM_TIMING, $r1
48 and.d 0x07, $r1 ; Get CAS latency
49 cmpq 2, $r1 ; CL = 2 ?
58 move.d CONFIG_ETRAX_SDRAM_GRP0_CONFIG, $r1
59 and.d 0x200, $r1 ; DRAM width is bit 9
66 move.d CONFIG_ETRAX_SDRAM_TIMING, $r1
67 and.d ~(3 << reg_bif_core_rw_sdram_timing___ref___lsb), $r1
69 move.d $r1, [$r0]
73 moveq regk_bif_core_nop, $r1
74 move.d $r1, [$r5]
97 move.d CONFIG_ETRAX_SDRAM_TIMING, $r1
99 move.d $r1, [$r0]
/linux-4.1.27/arch/arm/mach-ep93xx/
H A Dcrunch-bits.S73 ldr r1, [r8, #0x80]
74 tst r1, #0x00800000 @ access to crunch enabled?
78 orr r1, r1, #0x00800000 @ enable access to crunch
79 str r1, [r8, #0x80]
84 ldr r1, [r3] @ get current crunch owner
92 teq r1, #0 @ test for last ownership
97 cfstr64 mvdx0, [r1, #CRUNCH_MVDX0] @ save 64b registers
98 cfstr64 mvdx1, [r1, #CRUNCH_MVDX1]
99 cfstr64 mvdx2, [r1, #CRUNCH_MVDX2]
100 cfstr64 mvdx3, [r1, #CRUNCH_MVDX3]
101 cfstr64 mvdx4, [r1, #CRUNCH_MVDX4]
102 cfstr64 mvdx5, [r1, #CRUNCH_MVDX5]
103 cfstr64 mvdx6, [r1, #CRUNCH_MVDX6]
104 cfstr64 mvdx7, [r1, #CRUNCH_MVDX7]
105 cfstr64 mvdx8, [r1, #CRUNCH_MVDX8]
106 cfstr64 mvdx9, [r1, #CRUNCH_MVDX9]
107 cfstr64 mvdx10, [r1, #CRUNCH_MVDX10]
108 cfstr64 mvdx11, [r1, #CRUNCH_MVDX11]
109 cfstr64 mvdx12, [r1, #CRUNCH_MVDX12]
110 cfstr64 mvdx13, [r1, #CRUNCH_MVDX13]
111 cfstr64 mvdx14, [r1, #CRUNCH_MVDX14]
112 cfstr64 mvdx15, [r1, #CRUNCH_MVDX15]
119 cfstr32 mvfx0, [r1, #CRUNCH_MVAX0L]
121 cfstr32 mvfx0, [r1, #CRUNCH_MVAX0M]
123 cfstr32 mvfx0, [r1, #CRUNCH_MVAX0H]
125 cfstr32 mvfx0, [r1, #CRUNCH_MVAX1L]
127 cfstr32 mvfx0, [r1, #CRUNCH_MVAX1M]
129 cfstr32 mvfx0, [r1, #CRUNCH_MVAX1H]
131 cfstr32 mvfx0, [r1, #CRUNCH_MVAX2L]
133 cfstr32 mvfx0, [r1, #CRUNCH_MVAX2M]
135 cfstr32 mvfx0, [r1, #CRUNCH_MVAX2H]
137 cfstr32 mvfx0, [r1, #CRUNCH_MVAX3L]
139 cfstr32 mvfx0, [r1, #CRUNCH_MVAX3M]
141 cfstr32 mvfx0, [r1, #CRUNCH_MVAX3H]
144 cfstr64 mvdx0, [r1, #CRUNCH_DSPSC]
147 cfldr64eq mvdx0, [r1, #CRUNCH_MVDX0] @ mvdx0 was clobbered
220 ldr r1, [r3] @ get current crunch owner
221 teq r1, #0 @ any current owner?
224 teqne r1, r2 @ or specified one?
253 * r1 = memory address where to store crunch state
270 mov r0, r1
271 mov r1, r2
286 * r1 = memory address where to get crunch state from
308 mov r0, r1
309 mov r1, #0 @ nothing to save
/linux-4.1.27/sound/oss/
H A Dvidc_fill.S21 1: cmp r0, r1
34 1: cmp r0, r1
48 1: cmp r0, r1
60 1: cmp r0, r1
75 1: cmp r0, r1
81 cmp r0, r1
93 1: cmp r0, r1
97 cmp r0, r1
106 mov r1, #0
110 stmltia r2!, {r0, r1, r4, r5}
116 mov r1, #0
120 stmia r2!, {r0, r1}
126 * r1 = phys end
139 ldmia r8, {r0, r1, r2, r3, r4, r5}
140 teq r1, #0
152 add r1, r1, r0 @ End of virtual DMA buffer
155 sub r1, r1, r0 @ Remaining length
156 stmia r8, {r0, r1}
189 teq r1, #0 @ If we have no more
208 .long 0 @ r1
/linux-4.1.27/arch/cris/arch-v10/lib/
H A Ddram_init.S58 move.d CONFIG_ETRAX_DEF_R_SDRAM_TIMING, $r1
59 move.d $r1, $r3
60 and.d 0x03, $r1 ; Get CAS latency
65 cmp.d 0x00, $r1 ; CAS latency = 2?
72 cmp.d 0x01, $r1 ; CAS latency = 2?
77 move.d CONFIG_ETRAX_DEF_R_SDRAM_CONFIG, $r1
78 and.d 0x800000, $r1 ; DRAM width is bit 23
85 move.d CONFIG_ETRAX_DEF_R_SDRAM_TIMING, $r1
86 and.d 0x8000f9ff, $r1 ; Make sure mrs data and command is 0
87 or.d 0x80000000, $r1 ; Make sure sdram enable bit is set
88 move.d $r1, $r5
89 or.d 0x0000c000, $r1 ; ref = disable
91 or.d $r2, $r1
92 move.d $r1, [R_SDRAM_TIMING]
107 or.d $r1, $r4
/linux-4.1.27/arch/powerpc/net/
H A Dbpf_jit_asm.S98 * local variable space at r1+(BPF_PPC_STACK_BASIC).
104 PPC_STL r0, PPC_LR_STKOFF(r1); \
106 PPC_STL r_skb, (BPF_PPC_STACKFRAME+BPF_PPC_STACK_R3_OFF)(r1); \
107 PPC_STL r_A, (BPF_PPC_STACK_BASIC+(0*REG_SZ))(r1); \
108 PPC_STL r_X, (BPF_PPC_STACK_BASIC+(1*REG_SZ))(r1); \
109 addi r5, r1, BPF_PPC_STACK_BASIC+(2*REG_SZ); \
110 PPC_STLU r1, -BPF_PPC_SLOWPATH_FRAME(r1); \
117 addi r1, r1, BPF_PPC_SLOWPATH_FRAME; \
118 PPC_LL r0, PPC_LR_STKOFF(r1); \
119 PPC_LL r_A, (BPF_PPC_STACK_BASIC+(0*REG_SZ))(r1); \
120 PPC_LL r_X, (BPF_PPC_STACK_BASIC+(1*REG_SZ))(r1); \
124 PPC_LL r_skb, (BPF_PPC_STACKFRAME+BPF_PPC_STACK_R3_OFF)(r1); \
130 lwz r_A, BPF_PPC_STACK_BASIC+(2*REG_SZ)(r1)
135 lhz r_A, BPF_PPC_STACK_BASIC+(2*8)(r1)
140 lbz r_A, BPF_PPC_STACK_BASIC+(2*8)(r1)
145 lbz r_X, BPF_PPC_STACK_BASIC+(2*8)(r1)
151 * local variable space at r1+(BPF_PPC_STACK_BASIC).
157 PPC_STL r0, PPC_LR_STKOFF(r1); \
159 PPC_STL r_skb, (BPF_PPC_STACKFRAME+BPF_PPC_STACK_R3_OFF)(r1); \
160 PPC_STL r_A, (BPF_PPC_STACK_BASIC+(0*REG_SZ))(r1); \
161 PPC_STL r_X, (BPF_PPC_STACK_BASIC+(1*REG_SZ))(r1); \
162 PPC_STLU r1, -BPF_PPC_SLOWPATH_FRAME(r1); \
169 addi r1, r1, BPF_PPC_SLOWPATH_FRAME; \
170 PPC_LL r0, PPC_LR_STKOFF(r1); \
171 PPC_LL r_A, (BPF_PPC_STACK_BASIC+(0*REG_SZ))(r1); \
172 PPC_LL r_X, (BPF_PPC_STACK_BASIC+(1*REG_SZ))(r1); \
177 PPC_LL r_skb, (BPF_PPC_STACKFRAME+BPF_PPC_STACK_R3_OFF)(r1); \
/linux-4.1.27/arch/powerpc/platforms/pasemi/
H A Dpowersave.S40 std r3,8(r1); \
42 ld r3,8(r1); \
63 std r0, 16(r1)
64 stdu r1,-64(r1)
66 std r3, 48(r1)
73 ld r3, 48(r1)
85 1: addi r1,r1,64
86 ld r0,16(r1)
/linux-4.1.27/arch/blackfin/mach-common/
H A Dhead.S19 r2 = r2 - r1;
23 p1 = r1;
76 r1 = [p5 + (IMEM_CONTROL - COREMMR_BASE)]; define
77 BITCLR (r1, ENICPLB_P);
78 [p5 + (IMEM_CONTROL - COREMMR_BASE)] = r1;
82 r1 = [p5 + (DMEM_CONTROL - COREMMR_BASE)]; define
83 BITCLR (r1, ENDCPLB_P);
84 [p5 + (DMEM_CONTROL - COREMMR_BASE)] = r1;
98 r1 = [p0 + PDA_DF_DCPLB]; define
102 [p1 + PDA_INIT_DF_DCPLB] = r1;
123 r1.l = __sbss_l1;
124 r1.h = __sbss_l1;
130 r1.l = __sbss_b_l1;
131 r1.h = __sbss_b_l1;
137 r1.l = __sbss_l2;
138 r1.h = __sbss_l2;
143 r1.l = ___bss_start;
144 r1.h = ___bss_start;
/linux-4.1.27/tools/testing/selftests/powerpc/copyloops/
H A Dcopyuser_power7.S65 ld r16,STK_REG(R16)(r1)
66 ld r15,STK_REG(R15)(r1)
67 ld r14,STK_REG(R14)(r1)
70 ld r0,STACKFRAMESIZE+16(r1)
76 ld r22,STK_REG(R22)(r1)
77 ld r21,STK_REG(R21)(r1)
78 ld r20,STK_REG(R20)(r1)
79 ld r19,STK_REG(R19)(r1)
80 ld r18,STK_REG(R18)(r1)
81 ld r17,STK_REG(R17)(r1)
82 ld r16,STK_REG(R16)(r1)
83 ld r15,STK_REG(R15)(r1)
84 ld r14,STK_REG(R14)(r1)
86 addi r1,r1,STACKFRAMESIZE
88 ld r3,-STACKFRAMESIZE+STK_REG(R31)(r1)
89 ld r4,-STACKFRAMESIZE+STK_REG(R30)(r1)
90 ld r5,-STACKFRAMESIZE+STK_REG(R29)(r1)
99 std r3,-STACKFRAMESIZE+STK_REG(R31)(r1)
100 std r4,-STACKFRAMESIZE+STK_REG(R30)(r1)
101 std r5,-STACKFRAMESIZE+STK_REG(R29)(r1)
108 std r3,-STACKFRAMESIZE+STK_REG(R31)(r1)
109 std r4,-STACKFRAMESIZE+STK_REG(R30)(r1)
110 std r5,-STACKFRAMESIZE+STK_REG(R29)(r1)
144 stdu r1,-STACKFRAMESIZE(r1)
145 std r14,STK_REG(R14)(r1)
146 std r15,STK_REG(R15)(r1)
147 std r16,STK_REG(R16)(r1)
148 std r17,STK_REG(R17)(r1)
149 std r18,STK_REG(R18)(r1)
150 std r19,STK_REG(R19)(r1)
151 std r20,STK_REG(R20)(r1)
152 std r21,STK_REG(R21)(r1)
153 std r22,STK_REG(R22)(r1)
154 std r0,STACKFRAMESIZE+16(r1)
200 ld r14,STK_REG(R14)(r1)
201 ld r15,STK_REG(R15)(r1)
202 ld r16,STK_REG(R16)(r1)
203 ld r17,STK_REG(R17)(r1)
204 ld r18,STK_REG(R18)(r1)
205 ld r19,STK_REG(R19)(r1)
206 ld r20,STK_REG(R20)(r1)
207 ld r21,STK_REG(R21)(r1)
208 ld r22,STK_REG(R22)(r1)
209 addi r1,r1,STACKFRAMESIZE
290 addi r1,r1,STACKFRAMESIZE
296 std r0,16(r1)
297 stdu r1,-STACKFRAMESIZE(r1)
300 ld r0,STACKFRAMESIZE+16(r1)
301 ld r3,STK_REG(R31)(r1)
302 ld r4,STK_REG(R30)(r1)
303 ld r5,STK_REG(R29)(r1)
419 std r14,STK_REG(R14)(r1)
420 std r15,STK_REG(R15)(r1)
421 std r16,STK_REG(R16)(r1)
456 ld r14,STK_REG(R14)(r1)
457 ld r15,STK_REG(R15)(r1)
458 ld r16,STK_REG(R16)(r1)
516 15: addi r1,r1,STACKFRAMESIZE
604 std r14,STK_REG(R14)(r1)
605 std r15,STK_REG(R15)(r1)
606 std r16,STK_REG(R16)(r1)
649 ld r14,STK_REG(R14)(r1)
650 ld r15,STK_REG(R15)(r1)
651 ld r16,STK_REG(R16)(r1)
719 15: addi r1,r1,STACKFRAMESIZE
H A Dmemcpy_power7.S36 std r3,-STACKFRAMESIZE+STK_REG(R31)(r1)
43 std r3,-STACKFRAMESIZE+STK_REG(R31)(r1)
77 stdu r1,-STACKFRAMESIZE(r1)
78 std r14,STK_REG(R14)(r1)
79 std r15,STK_REG(R15)(r1)
80 std r16,STK_REG(R16)(r1)
81 std r17,STK_REG(R17)(r1)
82 std r18,STK_REG(R18)(r1)
83 std r19,STK_REG(R19)(r1)
84 std r20,STK_REG(R20)(r1)
85 std r21,STK_REG(R21)(r1)
86 std r22,STK_REG(R22)(r1)
87 std r0,STACKFRAMESIZE+16(r1)
133 ld r14,STK_REG(R14)(r1)
134 ld r15,STK_REG(R15)(r1)
135 ld r16,STK_REG(R16)(r1)
136 ld r17,STK_REG(R17)(r1)
137 ld r18,STK_REG(R18)(r1)
138 ld r19,STK_REG(R19)(r1)
139 ld r20,STK_REG(R20)(r1)
140 ld r21,STK_REG(R21)(r1)
141 ld r22,STK_REG(R22)(r1)
142 addi r1,r1,STACKFRAMESIZE
219 15: ld r3,-STACKFRAMESIZE+STK_REG(R31)(r1)
223 addi r1,r1,STACKFRAMESIZE
229 std r4,-STACKFRAMESIZE+STK_REG(R30)(r1)
230 std r5,-STACKFRAMESIZE+STK_REG(R29)(r1)
231 std r0,16(r1)
232 stdu r1,-STACKFRAMESIZE(r1)
235 ld r0,STACKFRAMESIZE+16(r1)
236 ld r3,STK_REG(R31)(r1)
237 ld r4,STK_REG(R30)(r1)
238 ld r5,STK_REG(R29)(r1)
352 std r14,STK_REG(R14)(r1)
353 std r15,STK_REG(R15)(r1)
354 std r16,STK_REG(R16)(r1)
389 ld r14,STK_REG(R14)(r1)
390 ld r15,STK_REG(R15)(r1)
391 ld r16,STK_REG(R16)(r1)
449 15: addi r1,r1,STACKFRAMESIZE
450 ld r3,-STACKFRAMESIZE+STK_REG(R31)(r1)
538 std r14,STK_REG(R14)(r1)
539 std r15,STK_REG(R15)(r1)
540 std r16,STK_REG(R16)(r1)
583 ld r14,STK_REG(R14)(r1)
584 ld r15,STK_REG(R15)(r1)
585 ld r16,STK_REG(R16)(r1)
653 15: addi r1,r1,STACKFRAMESIZE
654 ld r3,-STACKFRAMESIZE+STK_REG(R31)(r1)
/linux-4.1.27/arch/arm/mach-rockchip/
H A Dsleep.S29 setmode PSR_I_BIT | PSR_F_BIT | SVC_MODE, r1 @ set svc, irqs off
30 mrc p15, 0, r1, c0, c0, 5
31 and r1, r1, #0xf
32 cmp r1, #0
46 ldr r1, rkpm_bootdata_cpu_code
47 bx r1
/linux-4.1.27/arch/parisc/kernel/
H A Dsyscall.S109 mfsp %sr7,%r1 /* save user sr7 */
110 mtsp %r1,%sr3 /* and store it in sr3 */
119 ssm PSW_SM_W, %r1
120 extrd,u %r1,PSW_W_BIT,1,%r1
123 or,ev %r1,%r30,%r30
136 mfctl %cr30,%r1
137 xor %r1,%r30,%r30 /* ye olde xor trick */
138 xor %r1,%r30,%r1
139 xor %r1,%r30,%r30
150 STREGM %r1,FRAME_SIZE(%r30) /* save r1 (usp) here for now */
151 mfctl %cr30,%r1 /* get task ptr in %r1 */
152 LDREG TI_TASK(%r1),%r1
159 STREG %r0, TASK_PT_PSW(%r1)
160 STREG %r2, TASK_PT_GR2(%r1) /* preserve rp */
161 STREG %r19, TASK_PT_GR19(%r1)
169 std %r19,TASK_PT_PSW(%r1)
172 STREG %r2, TASK_PT_GR30(%r1) /* ... and save it */
174 STREG %r20, TASK_PT_GR20(%r1) /* Syscall number */
175 STREG %r21, TASK_PT_GR21(%r1)
176 STREG %r22, TASK_PT_GR22(%r1)
177 STREG %r23, TASK_PT_GR23(%r1) /* 4th argument */
178 STREG %r24, TASK_PT_GR24(%r1) /* 3rd argument */
179 STREG %r25, TASK_PT_GR25(%r1) /* 2nd argument */
180 STREG %r26, TASK_PT_GR26(%r1) /* 1st argument */
181 STREG %r27, TASK_PT_GR27(%r1) /* user dp */
182 STREG %r28, TASK_PT_GR28(%r1) /* return value 0 */
183 STREG %r0, TASK_PT_ORIG_R28(%r1) /* don't prohibit restarts */
184 STREG %r29, TASK_PT_GR29(%r1) /* return value 1 */
185 STREG %r31, TASK_PT_GR31(%r1) /* preserve syscall return ptr */
187 ldo TASK_PT_FR0(%r1), %r27 /* save fpregs from the kernel */
191 STREG %r27, TASK_PT_SAR(%r1)
206 mfctl %cr30, %r1
207 LDREG TI_FLAGS(%r1),%r1
209 and,COND(=) %r1, %r19, %r0
216 ldil L%sys_call_table, %r1
218 addil L%(sys_call_table64-sys_call_table), %r1
219 ldo R%sys_call_table(%r1), %r19
221 ldo R%sys_call_table64(%r1), %r19
223 ldil L%sys_call_table, %r1
224 ldo R%sys_call_table(%r1), %r19
255 ldil L%syscall_exit,%r1
256 be R%syscall_exit(%sr7,%r1)
273 ldo -THREAD_SZ_ALGN-FRAME_SIZE(%r30),%r1 /* get task ptr */
274 LDREG TI_TASK(%r1), %r1
276 STREG %r2,TASK_PT_PSW(%r1) /* Lower 8 bits only!! */
278 STREG %r2,TASK_PT_SR0(%r1)
280 STREG %r2,TASK_PT_SR1(%r1)
282 STREG %r2,TASK_PT_SR2(%r1)
284 STREG %r2,TASK_PT_SR3(%r1)
285 STREG %r2,TASK_PT_SR4(%r1)
286 STREG %r2,TASK_PT_SR5(%r1)
287 STREG %r2,TASK_PT_SR6(%r1)
288 STREG %r2,TASK_PT_SR7(%r1)
289 STREG %r2,TASK_PT_IASQ0(%r1)
290 STREG %r2,TASK_PT_IASQ1(%r1)
291 LDREG TASK_PT_GR31(%r1),%r2
292 STREG %r2,TASK_PT_IAOQ0(%r1)
294 STREG %r2,TASK_PT_IAOQ1(%r1)
295 ldo TASK_REGS(%r1),%r2
316 ldil L%do_syscall_trace_enter,%r1
318 be R%do_syscall_trace_enter(%sr7,%r1)
327 ldil L%sys_call_table,%r1
328 ldo R%sys_call_table(%r1), %r19
330 ldo -THREAD_SZ_ALGN-FRAME_SIZE(%r30),%r1 /* get task ptr */
331 LDREG TI_TASK(%r1), %r1
332 LDREG TASK_PT_GR26(%r1), %r26 /* Restore the users args */
333 LDREG TASK_PT_GR25(%r1), %r25
334 LDREG TASK_PT_GR24(%r1), %r24
335 LDREG TASK_PT_GR23(%r1), %r23
336 LDREG TASK_PT_GR22(%r1), %r22
337 LDREG TASK_PT_GR21(%r1), %r21
369 ldo -THREAD_SZ_ALGN-FRAME_SIZE(%r30),%r1 /* get task ptr */
370 LDREG TI_TASK(%r1), %r1
374 ldo TASK_REGS(%r1),%r26
376 STREG %r28,TASK_PT_GR28(%r1) /* save return value now */
377 ldo -THREAD_SZ_ALGN-FRAME_SIZE(%r30),%r1 /* get task ptr */
378 LDREG TI_TASK(%r1), %r1
379 LDREG TASK_PT_GR28(%r1), %r28 /* Restore return val. */
381 ldil L%syscall_exit,%r1
382 be,n R%syscall_exit(%sr7,%r1)
391 ldo -THREAD_SZ_ALGN-FRAME_SIZE(%r30),%r1 /* get task ptr */
392 LDREG TI_TASK(%r1), %r1
397 ldo TASK_REGS(%r1),%r26
399 ldil L%syscall_exit_rfi,%r1
400 be,n R%syscall_exit_rfi(%sr7,%r1)
422 - %r1, %r27, %r29
461 ssm PSW_SM_W, %r1
462 extrd,u %r1,PSW_W_BIT,1,%r1
465 or,ev %r1,%r30,%r30
476 mfsp %sr7,%r1 /* get userspace into sr3 */
477 mtsp %r1,%sr3
481 ldil L%lws_table, %r1
482 ldo R%lws_table(%r1), %r28 /* Scratch use of r28 */
498 extrd,u,*<> %r30,63,1,%r1
501 xor %r30,%r1,%r30
530 Scratch: r20, r28, r1
624 mfctl %cr27, %r1
625 stw %r1, 4(%sr2,%r20)
682 Scratch: r20, r22, r28, r29, r1, fr4 (32bit for 64bit CAS only)
H A Dentry.S62 /* Switch to virtual mapping, trashing only %r1 */
70 load32 KERNEL_PSW, %r1
75 mtctl %r1, %ipsw
76 load32 4f, %r1
77 mtctl %r1, %cr18 /* Set IIAOQ tail */
78 ldo 4(%r1), %r1
79 mtctl %r1, %cr18 /* Set IIAOQ head */
119 mfctl %cr30, %r1
120 ldo THREAD_SZ_ALGN(%r1), %r30
123 tophys %r1,%r9
124 LDREG TI_TASK(%r9), %r1 /* thread_info -> task_struct */
125 tophys %r1,%r9
139 copy %r30,%r1
141 STREG %r1,PT_GR30(%r9)
149 LDREG PT_GR1(%r29), %r1
763 LDREG TI_TASK-THREAD_SZ_ALGN-FRAME_SIZE(%r30), %r1
764 LDREG TASK_PT_GR25(%r1), %r26
766 LDREG TASK_PT_GR27(%r1), %r27
768 LDREG TASK_PT_GR26(%r1), %r1
769 ble 0(%sr7, %r1)
838 load32 USER_PSW_MASK,%r1
841 depd %r20,31,32,%r1
843 and %r19,%r1,%r19 /* Mask out bits that user shouldn't play with */
844 load32 USER_PSW,%r1
845 or %r19,%r1,%r19 /* Make sure default USER_PSW bits are set */
870 mfctl %cr30,%r1
871 LDREG TI_FLAGS(%r1),%r19 /* sched.h: TIF_NEED_RESCHED */
877 mfctl %cr30,%r1
878 LDREG TI_FLAGS(%r1),%r19
912 ldo PT_FR31(%r29),%r1
913 rest_fp %r1
927 * It also restores r1 and r30.
979 mfctl %cr30, %r1
980 LDREG TI_PRE_COUNT(%r1), %r19
1081 extrd,u %r16,63,SPACEID_SHIFT,%r1 /* get high bits from isr for ior */
1082 depd %r1,31,SPACEID_SHIFT,%r17 /* deposit them into ior */
1130 t0 = r1 /* temporary register 0 */
1347 cmpib,COND(=),n -1,%r1,nadtlb_fault /* have to use slow path */
1348 copy %r1,%r24
1351 cmpib,COND(=),n -1,%r1,nadtlb_fault /* have to use slow path */
1353 add,l %r1,%r24,%r1 /* doesn't affect c/b bits */
1383 cmpib,COND(=),n -1,%r1,nadtlb_fault /* have to use slow path */
1385 copy %r0,%r1 /* Write zero to target register */
1652 %r1 clobbered by system call macro in userspace
1712 LDREG TI_TASK-THREAD_SZ_ALGN-FRAME_SIZE(%r30), %r1
1713 ldo TASK_REGS(%r1),%r1
1714 reg_save %r1
1718 STREG %r28, PT_CR27(%r1)
1731 LDREG TI_TASK-THREAD_SZ_ALGN-FRAME_SIZE(%r30), %r1
1732 ldo TASK_REGS(%r1),%r1 /* get pt regs */
1734 LDREG PT_CR27(%r1), %r3
1736 reg_restore %r1
1759 LDREG TI_TASK-THREAD_SZ_ALGN-FRAME_SIZE(%r30),%r1
1760 ldo TASK_REGS(%r1),%r1 /* get pt regs */
1761 reg_restore %r1
1768 LDREG PT_GR28(%r1),%r28 /* reload original r28 for syscall_exit */
1779 mfctl %cr30, %r1
1780 LDREG TI_TASK(%r1),%r1
1781 STREG %r28,TASK_PT_GR28(%r1)
1808 LDREG TI_TASK-THREAD_SZ_ALGN-FRAME_SIZE(%r30),%r1
1809 ldo TASK_REGS(%r1), %r26 /* struct pt_regs *regs */
1819 LDREG TI_TASK-THREAD_SZ_ALGN-FRAME_SIZE(%r30),%r1
1820 ldo TASK_REGS(%r1), %r20 /* reload pt_regs */
1826 LDREG TI_TASK-THREAD_SZ_ALGN-FRAME_SIZE(%r30),%r1
1829 ldw TASK_FLAGS(%r1),%r19
1834 ldo TASK_PT_FR31(%r1),%r19 /* reload fpregs */
1837 LDREG TASK_PT_SAR(%r1),%r19 /* restore SAR */
1840 LDREG TASK_PT_GR2(%r1),%r2 /* restore user rp */
1841 LDREG TASK_PT_GR19(%r1),%r19
1842 LDREG TASK_PT_GR20(%r1),%r20
1843 LDREG TASK_PT_GR21(%r1),%r21
1844 LDREG TASK_PT_GR22(%r1),%r22
1845 LDREG TASK_PT_GR23(%r1),%r23
1846 LDREG TASK_PT_GR24(%r1),%r24
1847 LDREG TASK_PT_GR25(%r1),%r25
1848 LDREG TASK_PT_GR26(%r1),%r26
1849 LDREG TASK_PT_GR27(%r1),%r27 /* restore user dp */
1850 LDREG TASK_PT_GR28(%r1),%r28 /* syscall return value */
1851 LDREG TASK_PT_GR29(%r1),%r29
1852 LDREG TASK_PT_GR31(%r1),%r31 /* restore syscall rp */
1855 LDREG TASK_PT_GR30(%r1),%r1 /* Get user sp */
1857 copy %r1,%r30 /* Restore user sp */
1858 mfsp %sr3,%r1 /* Get user space id */
1859 mtsp %r1,%sr7 /* Restore sr7 */
1864 mtsp %r1,%sr4 /* Restore sr4 */
1865 mtsp %r1,%sr5 /* Restore sr5 */
1866 mtsp %r1,%sr6 /* Restore sr6 */
1875 extrd,u,*<> %r30,63,1,%r1
1878 xor %r30,%r1,%r30
1890 LDREG TASK_PT_PSW(%r1),%r2 /* Get old PSW */
1906 STREG %r20,TASK_PT_PSW(%r1)
1911 STREG %r25,TASK_PT_SR3(%r1)
1912 STREG %r25,TASK_PT_SR4(%r1)
1913 STREG %r25,TASK_PT_SR5(%r1)
1914 STREG %r25,TASK_PT_SR6(%r1)
1915 STREG %r25,TASK_PT_SR7(%r1)
1916 STREG %r25,TASK_PT_IASQ0(%r1)
1917 STREG %r25,TASK_PT_IASQ1(%r1)
1929 ldo TASK_REGS(%r1),%r25
1934 STREG %r2,TASK_PT_SR0(%r1)
1938 STREG %r2,TASK_PT_SR1(%r1)
1941 STREG %r0,TASK_PT_SR2(%r1)
1943 LDREG TASK_PT_GR31(%r1),%r2
1945 STREG %r2,TASK_PT_IAOQ0(%r1)
1947 STREG %r2,TASK_PT_IAOQ1(%r1)
1952 LDREG TASK_PT_IAOQ0(%r1),%r2
1954 STREG %r2,TASK_PT_IAOQ0(%r1)
1955 LDREG TASK_PT_IAOQ1(%r1),%r2
1957 STREG %r2,TASK_PT_IAOQ1(%r1)
2004 copy %sp, %r1
2018 STREG %r1, -136(%sp)
2026 STREG %r1, -68(%sp)
2046 * r1. This routine can't be used for shadowed registers, since
2048 * registers we put a -1 into r1 to indicate that the register
2056 copy %r0,%r1
2057 bv %r0(%r25) /* r1 - shadowed */
2058 ldi -1,%r1
2060 copy %r2,%r1
2062 copy %r3,%r1
2064 copy %r4,%r1
2066 copy %r5,%r1
2068 copy %r6,%r1
2070 copy %r7,%r1
2072 ldi -1,%r1
2074 ldi -1,%r1
2076 copy %r10,%r1
2078 copy %r11,%r1
2080 copy %r12,%r1
2082 copy %r13,%r1
2084 copy %r14,%r1
2086 copy %r15,%r1
2088 ldi -1,%r1
2090 ldi -1,%r1
2092 copy %r18,%r1
2094 copy %r19,%r1
2096 copy %r20,%r1
2098 copy %r21,%r1
2100 copy %r22,%r1
2102 copy %r23,%r1
2104 ldi -1,%r1
2106 ldi -1,%r1
2108 copy %r26,%r1
2110 copy %r27,%r1
2112 copy %r28,%r1
2114 copy %r29,%r1
2116 copy %r30,%r1
2118 copy %r31,%r1
2124 * copy the value of r1 into the general register specified in
2130 copy %r1,%r0
2131 bv %r0(%r25) /* r1 */
2132 copy %r1,%r1
2134 copy %r1,%r2
2136 copy %r1,%r3
2138 copy %r1,%r4
2140 copy %r1,%r5
2142 copy %r1,%r6
2144 copy %r1,%r7
2146 copy %r1,%r8
2148 copy %r1,%r9
2150 copy %r1,%r10
2152 copy %r1,%r11
2154 copy %r1,%r12
2156 copy %r1,%r13
2158 copy %r1,%r14
2160 copy %r1,%r15
2162 copy %r1,%r16
2164 copy %r1,%r17
2166 copy %r1,%r18
2168 copy %r1,%r19
2170 copy %r1,%r20
2172 copy %r1,%r21
2174 copy %r1,%r22
2176 copy %r1,%r23
2178 copy %r1,%r24
2180 copy %r1,%r25
2182 copy %r1,%r26
2184 copy %r1,%r27
2186 copy %r1,%r28
2188 copy %r1,%r29
2190 copy %r1,%r30
2192 copy %r1,%r31
H A Dpacache.S58 load32 PA(1f), %r1
68 mtctl %r1, %cr18 /* IIAOQ head */
69 ldo 4(%r1), %r1
70 mtctl %r1, %cr18 /* IIAOQ tail */
71 load32 REAL_MODE_PSW, %r1
72 mtctl %r1, %ipsw
76 1: load32 PA(cache_info), %r1
80 LDREG ITLB_SID_BASE(%r1), %r20
81 LDREG ITLB_SID_STRIDE(%r1), %r21
82 LDREG ITLB_SID_COUNT(%r1), %r22
83 LDREG ITLB_OFF_BASE(%r1), %arg0
84 LDREG ITLB_OFF_STRIDE(%r1), %arg1
85 LDREG ITLB_OFF_COUNT(%r1), %arg2
86 LDREG ITLB_LOOP(%r1), %arg3
123 LDREG DTLB_SID_BASE(%r1), %r20
124 LDREG DTLB_SID_STRIDE(%r1), %r21
125 LDREG DTLB_SID_COUNT(%r1), %r22
126 LDREG DTLB_OFF_BASE(%r1), %arg0
127 LDREG DTLB_OFF_STRIDE(%r1), %arg1
128 LDREG DTLB_OFF_COUNT(%r1), %arg2
129 LDREG DTLB_LOOP(%r1), %arg3
169 load32 2f, %r1
179 mtctl %r1, %cr18 /* IIAOQ head */
180 ldo 4(%r1), %r1
181 mtctl %r1, %cr18 /* IIAOQ tail */
182 load32 KERNEL_PSW, %r1
183 or %r1, %r19, %r1 /* I-bit to state on entry */
184 mtctl %r1, %ipsw /* restore I-bit (entire PSW) */
202 load32 cache_info, %r1
206 LDREG ICACHE_BASE(%r1), %arg0
207 LDREG ICACHE_STRIDE(%r1), %arg1
208 LDREG ICACHE_COUNT(%r1), %arg2
209 LDREG ICACHE_LOOP(%r1), %arg3
269 load32 cache_info, %r1
273 LDREG DCACHE_BASE(%r1), %arg0
274 LDREG DCACHE_STRIDE(%r1), %arg1
275 LDREG DCACHE_COUNT(%r1), %arg2
276 LDREG DCACHE_LOOP(%r1), %arg3
336 ldil L%pa_tlb_lock,%r1
337 ldo R%pa_tlb_lock(%r1),\la
367 ldi (PAGE_SIZE / 128), %r1
388 addib,COND(>),n -1, %r1, 1b
397 ldi (PAGE_SIZE / 64), %r1
417 addib,COND(>),n -1, %r1, 1b
441 ldi (PAGE_SIZE / 128), %r1
486 addib,COND(>),n -1, %r1, 1b
499 ldi (PAGE_SIZE / 64), %r1
535 addib,COND(>),n -1, %r1, 1b
596 ldil L%(__PAGE_OFFSET), %r1
597 sub %r26, %r1, %r26
598 sub %r25, %r1, %r23
639 ldi (PAGE_SIZE / 128), %r1
687 addib,COND(>),n -1, %r1, 1b /* bundle 10 */
691 ldi (PAGE_SIZE / 64), %r1
736 addib,COND(>) -1, %r1,1b
779 ldi (PAGE_SIZE / 128), %r1
800 addib,COND(>) -1, %r1, 1b
804 ldi (PAGE_SIZE / 64), %r1
822 addib,COND(>) -1, %r1, 1b
862 ldil L%dcache_stride, %r1
863 ldw R%dcache_stride(%r1), r31
938 ldil L%icache_stride, %r1
939 ldw R%icache_stride(%r1), %r31
992 ldil L%dcache_stride, %r1
993 ldw R%dcache_stride(%r1), %r23
1035 ldil L%dcache_stride, %r1
1036 ldw R%dcache_stride(%r1), %r23
1077 ldil L%dcache_stride, %r1
1078 ldw R%dcache_stride(%r1), %r23
1098 ldil L%dcache_stride, %r1
1099 ldw R%dcache_stride(%r1), %r23
1120 ldil L%icache_stride, %r1
1121 ldw R%icache_stride(%r1), %r23
1141 ldil L%icache_stride, %r1
1142 ldw R%icache_stride(%r1), %r23
1184 ldil L%icache_stride, %r1
1185 ldw R%icache_stride(%r1), %r23
1213 load32 PA(1f), %r1
1223 mtctl %r1, %cr18 /* IIAOQ head */
1224 ldo 4(%r1), %r1
1225 mtctl %r1, %cr18 /* IIAOQ tail */
1226 load32 REAL_MODE_PSW, %r1
1227 mtctl %r1, %ipsw
1269 load32 2f, %r1
1279 mtctl %r1, %cr18 /* IIAOQ head */
1280 ldo 4(%r1), %r1
1281 mtctl %r1, %cr18 /* IIAOQ tail */
1282 load32 KERNEL_PSW, %r1
1283 mtctl %r1, %ipsw
H A Dreal2.S90 b,l save_control_regs,%r2 /* modifies r1, r2, r28 */
105 b,l restore_control_regs, %r2 /* modifies r1, r2, r26 */
125 # define PUSH_CR(r, where) mfctl r, %r1 ! STREG,ma %r1, REG_SZ(where)
126 # define POP_CR(r, where) LDREG,mb -REG_SZ(where), %r1 ! mtctl %r1, r
165 load32 PA(rfi_v2r_1), %r1
175 mtctl %r1, %cr18 /* IIAOQ head */
176 ldo 4(%r1), %r1
177 mtctl %r1, %cr18 /* IIAOQ tail */
178 load32 REAL_MODE_PSW, %r1
179 mtctl %r1, %cr22
199 load32 (rfi_r2v_1), %r1
209 mtctl %r1, %cr18 /* IIAOQ head */
210 ldo 4(%r1), %r1
211 mtctl %r1, %cr18 /* IIAOQ tail */
212 load32 KERNEL_PSW, %r1
213 mtctl %r1, %cr22
268 b,l save_control_regs,%r2 /* modifies r1, r2, r28 */
276 b,l restore_control_regs, %r2 /* modifies r1, r2, r26 */
H A Dperf_asm.S114 ;* %r1 - scratch
150 extrd,u arg1,63,6,%r1 ; setup shift amount by bits to move
152 mtsar %r1
153 shladd arg0,2,%r0,%r1 ; %r1 = 4 * RDR number
154 blr %r1,%r0 ; branch to 8-instruction sequence
167 shrpd ret0,%r0,%sar,%r1
168 MTDIAG_1 (1) ; mtdiag %dr1, %r1
191 shrpd ret0,%r0,%sar,%r1
287 shrpd ret0,%r0,%sar,%r1
299 shrpd ret0,%r0,%sar,%r1
335 shrpd ret0,%r0,%sar,%r1
371 shrpd ret0,%r0,%sar,%r1
383 shrpd ret0,%r0,%sar,%r1
479 shrpd ret0,%r0,%sar,%r1
491 shrpd ret0,%r0,%sar,%r1
527 shrpd ret0,%r0,%sar,%r1
594 shladd arg0,2,%r0,%r1 ; %r1 = 4 * RDR number
595 blr %r1,%r0 ; branch to 8-instruction sequence
1013 ;* %r1 - scratch
1043 extrd,u arg1,63,6,%r1
1045 mtsar %r1
1046 shladd arg0,2,%r0,%r1 ; %r1 = 4 * RDR number
1047 blr %r1,%r0 ; branch to 8-instruction sequence
1056 shrpd ret0,%r0,%sar,%r1
1065 shrpd ret0,%r0,%sar,%r1
1137 shrpd ret0,%r0,%sar,%r1
1146 shrpd ret0,%r0,%sar,%r1
1155 shrpd ret0,%r0,%sar,%r1
1164 shrpd ret0,%r0,%sar,%r1
1173 shrpd ret0,%r0,%sar,%r1
1182 shrpd ret0,%r0,%sar,%r1
1191 shrpd ret0,%r0,%sar,%r1
1209 shrpd ret0,%r0,%sar,%r1
1218 shrpd ret0,%r0,%sar,%r1
1290 shrpd ret0,%r0,%sar,%r1
1299 shrpd ret0,%r0,%sar,%r1
1326 shrpd ret0,%r0,%sar,%r1
1335 shrpd ret0,%r0,%sar,%r1
1390 shladd arg0,2,%r0,%r1 ; %r1 = 4 * RDR number
1391 blr %r1,%r0 ; branch to 8-instruction sequence
/linux-4.1.27/arch/parisc/lib/
H A Dlusercopy.S49 mfctl %cr30,%r1
50 ldw TI_SEGMENT(%r1),%r22
51 mfsp %sr3,%r1
53 copy %r0,%r1
54 mtsp %r1,%sr1
58 ldil L%\lbl, %r1
59 ldo R%\lbl(%r1), %r1
60 bv %r0(%r1)
110 1: ldbs,ma 1(%sr1,%r26),%r1
112 comib,=,n 0,%r1,$lslen_done
114 2: ldbs,ma 1(%sr1,%r26),%r1
/linux-4.1.27/arch/arm/crypto/
H A Daes-armv4.S153 stmdb sp!,{r1,r4-r12,lr}
163 ldrb r1,[r12,#7]
169 orr r1,r1,r4,lsl#8
171 orr r1,r1,r5,lsl#16
173 orr r1,r1,r6,lsl#24
188 ldr r1,[r12,#4]
193 rev r1,r1
204 rev r1,r1
209 str r1,[r12,#4]
218 mov r4,r1,lsr#24
220 mov r5,r1,lsr#16
222 mov r6,r1,lsr#8
228 strb r1,[r12,#7]
252 eor r1,r1,r5
264 and r7,lr,r1,lsr#16 @ i0
266 and r8,lr,r1
268 and r9,lr,r1,lsr#8
270 mov r1,r1,lsr#24
276 ldr r1,[r10,r1,lsl#2] @ Te0[s1>>24]
283 eor r1,r1,r4,ror#24
291 eor r1,r1,r8,ror#8
303 eor r1,r1,r8,ror#16
313 eor r1,r1,r4
326 and r7,lr,r1,lsr#16 @ i0
328 and r8,lr,r1
330 and r9,lr,r1,lsr#8
332 mov r1,r1,lsr#24
338 ldrb r1,[r10,r1,lsl#2] @ Te4[s1>>24]
345 eor r1,r4,r1,lsl#24
353 eor r1,r1,r8,lsl#16
366 eor r1,r1,r8,lsl#8
374 eor r1,r1,r4
393 teq r1,#128
395 teq r1,#192
397 teq r1,#256
405 mov lr,r1 @ bits
414 ldrb r1,[r12,#7]
420 orr r1,r1,r4,lsl#8
422 orr r1,r1,r5,lsl#16
424 orr r1,r1,r6,lsl#24
437 str r1,[r11,#-12]
443 ldr r1,[r12,#4]
448 rev r1,r1
453 str r1,[r11,#-12]
480 eor r1,r1,r0 @ rk[5]=rk[1]^rk[4]
482 eor r2,r2,r1 @ rk[6]=rk[2]^rk[5]
483 str r1,[r11,#-12]
544 eor r1,r1,r0 @ rk[7]=rk[1]^rk[6]
546 eor r2,r2,r1 @ rk[8]=rk[2]^rk[7]
547 str r1,[r11,#-20]
613 eor r1,r1,r0 @ rk[9]=rk[1]^rk[8]
615 eor r2,r2,r1 @ rk[10]=rk[2]^rk[9]
616 str r1,[r11,#-28]
677 ldr r1,[r7,#4]
685 str r1,[r8,#16+4]
706 and r1,r0,r9
709 eor r1,r4,r1,lsl#1 @ tp2
711 and r4,r1,r7
712 and r2,r1,r9
723 eor r4,r1,r2
726 eor r4,r4,r1,ror#24
848 stmdb sp!,{r1,r4-r12,lr}
858 ldrb r1,[r12,#7]
864 orr r1,r1,r4,lsl#8
866 orr r1,r1,r5,lsl#16
868 orr r1,r1,r6,lsl#24
883 ldr r1,[r12,#4]
888 rev r1,r1
899 rev r1,r1
904 str r1,[r12,#4]
913 mov r4,r1,lsr#24
915 mov r5,r1,lsr#16
917 mov r6,r1,lsr#8
923 strb r1,[r12,#7]
947 eor r1,r1,r5
959 and r7,lr,r1 @ i0
961 and r8,lr,r1,lsr#16
963 and r9,lr,r1,lsr#8
965 mov r1,r1,lsr#24
971 ldr r1,[r10,r1,lsl#2] @ Td0[s1>>24]
978 eor r1,r1,r4,ror#8
986 eor r1,r1,r8,ror#24
998 eor r1,r1,r8,ror#16
1008 eor r1,r1,r4
1031 and r7,lr,r1 @ i0
1033 and r8,lr,r1,lsr#16
1035 and r9,lr,r1,lsr#8
1038 ARM( ldrb r1,[r10,r1,lsr#24] ) @ Td4[s1>>24]
1039 THUMB( add r1,r10,r1,lsr#24 ) @ Td4[s1>>24]
1040 THUMB( ldrb r1,[r1] )
1044 eor r1,r4,r1,lsl#8
1058 eor r1,r8,r1,lsl#16
1073 eor r1,r1,r8,lsl#8
1081 eor r1,r1,r4
/linux-4.1.27/arch/powerpc/crypto/
H A Dsha1-spe-asm.S62 stwu r1,-128(r1); /* create stack frame */ \
63 evstdw r14,8(r1); /* We must save non volatile */ \
64 evstdw r15,16(r1); /* registers. Take the chance */ \
65 evstdw r16,24(r1); /* and save the SPE part too */ \
66 evstdw r17,32(r1); \
67 evstdw r18,40(r1); \
68 evstdw r19,48(r1); \
69 evstdw r20,56(r1); \
70 evstdw r21,64(r1); \
71 evstdw r22,72(r1); \
72 evstdw r23,80(r1);
76 evldw r14,8(r1); /* restore SPE registers */ \
77 evldw r15,16(r1); \
78 evldw r16,24(r1); \
79 evldw r17,32(r1); \
80 evldw r18,40(r1); \
81 evldw r19,48(r1); \
82 evldw r20,56(r1); \
83 evldw r21,64(r1); \
84 evldw r22,72(r1); \
85 evldw r23,80(r1); \
87 stw r0,8(r1); /* Delete sensitive data */ \
88 stw r0,16(r1); /* that we might have pushed */ \
89 stw r0,24(r1); /* from other context that runs */ \
90 stw r0,32(r1); /* the same code. Assume that */ \
91 stw r0,40(r1); /* the lower part of the GPRs */ \
92 stw r0,48(r1); /* were already overwritten on */ \
93 stw r0,56(r1); /* the way down to here */ \
94 stw r0,64(r1); \
95 stw r0,72(r1); \
96 stw r0,80(r1); \
97 addi r1,r1,128; /* cleanup stack frame */
H A Dsha256-spe-asm.S53 stwu r1,-128(r1); /* create stack frame */ \
54 evstdw r14,8(r1); /* We must save non volatile */ \
55 evstdw r15,16(r1); /* registers. Take the chance */ \
56 evstdw r16,24(r1); /* and save the SPE part too */ \
57 evstdw r17,32(r1); \
58 evstdw r18,40(r1); \
59 evstdw r19,48(r1); \
60 evstdw r20,56(r1); \
61 evstdw r21,64(r1); \
62 evstdw r22,72(r1); \
63 evstdw r23,80(r1); \
64 stw r24,88(r1); /* save normal registers */ \
65 stw r25,92(r1);
69 evldw r14,8(r1); /* restore SPE registers */ \
70 evldw r15,16(r1); \
71 evldw r16,24(r1); \
72 evldw r17,32(r1); \
73 evldw r18,40(r1); \
74 evldw r19,48(r1); \
75 evldw r20,56(r1); \
76 evldw r21,64(r1); \
77 evldw r22,72(r1); \
78 evldw r23,80(r1); \
79 lwz r24,88(r1); /* restore normal registers */ \
80 lwz r25,92(r1); \
82 stw r0,8(r1); /* Delete sensitive data */ \
83 stw r0,16(r1); /* that we might have pushed */ \
84 stw r0,24(r1); /* from other context that runs */ \
85 stw r0,32(r1); /* the same code. Assume that */ \
86 stw r0,40(r1); /* the lower part of the GPRs */ \
87 stw r0,48(r1); /* was already overwritten on */ \
88 stw r0,56(r1); /* the way down to here */ \
89 stw r0,64(r1); \
90 stw r0,72(r1); \
91 stw r0,80(r1); \
92 addi r1,r1,128; /* cleanup stack frame */
H A Daes-spe-modes.S59 stw rI0,96(r1); /* save 32 bit registers */ \
60 stw rI1,100(r1); \
61 stw rI2,104(r1); \
62 stw rI3,108(r1);
65 lwz rI0,96(r1); /* restore 32 bit registers */ \
66 lwz rI1,100(r1); \
67 lwz rI2,104(r1); \
68 lwz rI3,108(r1);
72 stw rG0,112(r1); /* save 32 bit registers */ \
73 stw rG1,116(r1); \
74 stw rG2,120(r1); \
75 stw rG3,124(r1);
79 lwz rG0,112(r1); /* restore 32 bit registers */ \
80 lwz rG1,116(r1); \
81 lwz rG2,120(r1); \
82 lwz rG3,124(r1);
86 stwu r1,-160(r1); /* create stack frame */ \
88 stw r0,8(r1); /* save link register */ \
90 evstdw r14,16(r1); \
92 evstdw r15,24(r1); /* We must save non volatile */ \
93 evstdw r16,32(r1); /* registers. Take the chance */ \
94 evstdw r17,40(r1); /* and save the SPE part too */ \
95 evstdw r18,48(r1); \
96 evstdw r19,56(r1); \
97 evstdw r20,64(r1); \
98 evstdw r21,72(r1); \
99 evstdw r22,80(r1); \
100 evstdw r23,88(r1); \
104 lwz r0,8(r1); \
105 evldw r14,16(r1); /* restore SPE registers */ \
106 evldw r15,24(r1); \
107 evldw r16,32(r1); \
108 evldw r17,40(r1); \
109 evldw r18,48(r1); \
110 evldw r19,56(r1); \
111 evldw r20,64(r1); \
112 evldw r21,72(r1); \
113 evldw r22,80(r1); \
114 evldw r23,88(r1); \
118 stw r0,16(r1); /* delete sensitive data */ \
119 stw r0,24(r1); /* that we might have pushed */ \
120 stw r0,32(r1); /* from other context that runs */ \
121 stw r0,40(r1); /* the same code */ \
122 stw r0,48(r1); \
123 stw r0,56(r1); \
124 stw r0,64(r1); \
125 stw r0,72(r1); \
126 stw r0,80(r1); \
127 stw r0,88(r1); \
128 addi r1,r1,160; /* cleanup stack frame */
H A Dsha1-powerpc-asm.S116 PPC_STLU r1,-INT_FRAME_SIZE(r1)
117 SAVE_8GPRS(14, r1)
118 SAVE_10GPRS(22, r1)
176 REST_8GPRS(14, r1)
177 REST_10GPRS(22, r1)
178 addi r1,r1,INT_FRAME_SIZE
/linux-4.1.27/arch/cris/arch-v32/kernel/
H A Dhead.S74 | REG_FIELD(mmu, rw_mm_kbase_lo, base_0, 0), $r1
139 move $r1, $s1 ; kbase_lo.
148 move $r1, $s1 ; kbase_lo
204 move.d text_start, $r1 ; Destination.
207 sub.d $r1, $r4
209 move.w $r3, [$r1+]
210 cmp.d $r2, $r1
216 move.d romfs_length, $r1
217 move.d $r0, [$r1]
226 move.d romfs_length, $r1
227 move.d $r0, [$r1]
229 move.d romfs_start, $r1
230 move.d $r4, [$r1]
232 move.d romfs_in_flash, $r1
233 move.d $r0, [$r1]
244 move.d nand_boot, $r1 ; tell axisflashmap we're booting from NAND
245 move.d $r0, [$r1]
247 move.d romfs_in_flash, $r1 ; (directly accessed) flash
248 move.d $r0, [$r1]
257 move.d romfs_length, $r1
258 move.d $r0, [$r1]
277 move.d romfs_length, $r1
278 move.d $r0, [$r1]
280 move.d romfs_start, $r1
281 move.d $r9, [$r1]
283 move.d romfs_in_flash, $r1
284 move.d $r0, [$r1]
300 move.d [$r0], $r1
301 cmp.d CRAMFS_MAGIC, $r1 ; cramfs magic?
304 cmp.d JHEAD_MAGIC, $r1 ; jffs2 (jhead) magic?
319 move.d _end, $r1 ; start of cramfs -> r1
321 move.d $r1, [$r3] ; store at romfs_start (for axisflashmap)
326 add.d $r2, $r1
332 move.w $r3, [$r1]
334 subq 2, $r1
345 move.d romfs_in_flash, $r1
346 move.d $r0, [$r1]
348 move.d nand_boot, $r1
349 move.d $r0, [$r1]
365 move.b [$r11+], $r1
366 move.b $r1, [$r10+]
379 move.d etrax_irv, $r1 ; Set the exception base register and pointer.
380 move.d $r0, [$r1]
384 move.d _end, $r1
386 cmp.d $r1, $r0
/linux-4.1.27/arch/arm/include/uapi/asm/
H A Dswab.h41 t = x ^ ((x << 16) | (x >> 16)); /* eor r1,r0,r0,ror #16 */ __arch_swab32()
44 t &= ~0x00FF0000; /* bic r1,r1,#0x00FF0000 */ __arch_swab32()
45 x ^= (t >> 8); /* eor r0,r0,r1,lsr #8 */ __arch_swab32()
/linux-4.1.27/arch/tile/kernel/
H A Dentry.S31 { move r2, lr; lnk r1 }
32 { move r4, r52; addli r1, r1, dump_stack - . }
38 { move r2, lr; lnk r1 }
39 { move r4, r52; addli r1, r1, KBacktraceIterator_init_current - . }
61 movei r1, 1
63 mtspr INTERRUPT_CRITICAL_SECTION, r1
H A Dregs_32.S66 addli r4, r1, TASK_STRUCT_THREAD_KSP_OFFSET
79 addli r4, r1, TASK_STRUCT_THREAD_PC_OFFSET
108 .irp reg, r0, r1, r2, r3, r4, r5, r6, r7, \
124 lnk r1
126 sw r0, r1
129 mfspr r1, INTERRUPT_CRITICAL_SECTION
130 shli r1, r1, SPR_EX_CONTEXT_1_1__ICS_SHIFT
131 ori r1, r1, KERNEL_PL
133 sw r0, r1
H A Dregs_64.S65 addli r4, r1, TASK_STRUCT_THREAD_KSP_OFFSET
79 addli r4, r1, TASK_STRUCT_THREAD_PC_OFFSET
108 .irp reg, r0, r1, r2, r3, r4, r5, r6, r7, \
124 lnk r1
126 st r0, r1
129 mfspr r1, INTERRUPT_CRITICAL_SECTION
130 shli r1, r1, SPR_EX_CONTEXT_1_1__ICS_SHIFT
131 ori r1, r1, KERNEL_PL
133 st r0, r1
/linux-4.1.27/arch/arm/vfp/
H A Dvfphw.S42 mov r1, \arg
59 mov r1, \arg1
89 VFPFMRX r1, FPEXC @ Is the VFP enabled?
90 DBGSTR1 "fpexc %08x", r1
91 tst r1, #FPEXC_EN
96 orr r1, r1, #FPEXC_EN @ user FPEXC has the enable bit set
98 bic r5, r1, #FPEXC_EX @ make sure exceptions are disabled
119 tst r1, #FPEXC_EX @ is there additional state to save?
122 tst r1, #FPEXC_FP2V @ is there an FPINST2 to read?
127 stmia r4, {r1, r5, r6, r8} @ save FPEXC, FPSCR, FPINST, FPINST2
160 ldmia r10, {r1, r5, r6, r8} @ load FPEXC, FPSCR, FPINST, FPINST2
162 tst r1, #FPEXC_EX @ is there additional state to restore?
165 tst r1, #FPEXC_FP2V @ is there an FPINST2 to write?
174 tst r1, #FPEXC_EX
179 VFPFMXR FPEXC, r1 @ Restore FPEXC last
191 tst r1, #FPEXC_EX | FPEXC_DEX
202 orr r1, r1, #FPEXC_DEX
220 @ r1 holds the FPEXC value
231 @ r1 - FPEXC
235 tst r1, #FPEXC_EX @ is there additional state to save?
238 tst r1, #FPEXC_FP2V @ is there an FPINST2 to read?
242 stmia r0, {r1, r2, r3, r12} @ save FPEXC, FPSCR, FPINST, FPINST2
275 tbl_branch r1, r3, #3
289 1: fmrrd r0, r1, d\dr
296 1: mrrc p11, 3, r0, r1, c\dr @ fmrrd r0, r1, d\dr
304 mov r1, #0
311 1: fmdrr d\dr, r0, r1
318 1: mcrr p11, 3, r0, r1, c\dr @ fmdrr r0, r1, d\dr
/linux-4.1.27/arch/parisc/math-emu/
H A Dfpudispatch.c290 u_int r1,r2,t; /* operand register offsets */ decode_0c() local
303 r1 = extru(ir,fpr1pos,5) * sizeof(double)/sizeof(u_int); decode_0c()
304 if (r1 == 0) /* map fr0 source to constant zero */ decode_0c()
305 r1 = fpzeroreg; decode_0c()
323 r1 &= ~3; decode_0c()
324 fpregs[t+3] = fpregs[r1+3]; decode_0c()
325 fpregs[t+2] = fpregs[r1+2]; decode_0c()
327 fpregs[t+1] = fpregs[r1+1]; decode_0c()
329 fpregs[t] = fpregs[r1]; decode_0c()
338 r1 &= ~3; decode_0c()
339 fpregs[t+3] = fpregs[r1+3]; decode_0c()
340 fpregs[t+2] = fpregs[r1+2]; decode_0c()
342 fpregs[t+1] = fpregs[r1+1]; decode_0c()
345 fpregs[t] = fpregs[r1] & 0x7fffffff; decode_0c()
354 r1 &= ~3; decode_0c()
355 fpregs[t+3] = fpregs[r1+3]; decode_0c()
356 fpregs[t+2] = fpregs[r1+2]; decode_0c()
358 fpregs[t+1] = fpregs[r1+1]; decode_0c()
361 fpregs[t] = fpregs[r1] ^ 0x80000000; decode_0c()
370 r1 &= ~3; decode_0c()
371 fpregs[t+3] = fpregs[r1+3]; decode_0c()
372 fpregs[t+2] = fpregs[r1+2]; decode_0c()
374 fpregs[t+1] = fpregs[r1+1]; decode_0c()
377 fpregs[t] = fpregs[r1] | 0x80000000; decode_0c()
383 return(sgl_fsqrt(&fpregs[r1],0, decode_0c()
386 return(dbl_fsqrt(&fpregs[r1],0, decode_0c()
395 return(sgl_frnd(&fpregs[r1],0, decode_0c()
398 return(dbl_frnd(&fpregs[r1],0, decode_0c()
427 return(sgl_to_dbl_fcnvff(&fpregs[r1],0, decode_0c()
430 return(dbl_to_sgl_fcnvff(&fpregs[r1],0, decode_0c()
438 return(sgl_to_sgl_fcnvxf(&fpregs[r1],0, decode_0c()
441 return(sgl_to_dbl_fcnvxf(&fpregs[r1],0, decode_0c()
444 return(dbl_to_sgl_fcnvxf(&fpregs[r1],0, decode_0c()
447 return(dbl_to_dbl_fcnvxf(&fpregs[r1],0, decode_0c()
453 return(sgl_to_sgl_fcnvfx(&fpregs[r1],0, decode_0c()
456 return(sgl_to_dbl_fcnvfx(&fpregs[r1],0, decode_0c()
459 return(dbl_to_sgl_fcnvfx(&fpregs[r1],0, decode_0c()
462 return(dbl_to_dbl_fcnvfx(&fpregs[r1],0, decode_0c()
468 return(sgl_to_sgl_fcnvfxt(&fpregs[r1],0, decode_0c()
471 return(sgl_to_dbl_fcnvfxt(&fpregs[r1],0, decode_0c()
474 return(dbl_to_sgl_fcnvfxt(&fpregs[r1],0, decode_0c()
477 return(dbl_to_dbl_fcnvfxt(&fpregs[r1],0, decode_0c()
483 return(sgl_to_sgl_fcnvuf(&fpregs[r1],0, decode_0c()
486 return(sgl_to_dbl_fcnvuf(&fpregs[r1],0, decode_0c()
489 return(dbl_to_sgl_fcnvuf(&fpregs[r1],0, decode_0c()
492 return(dbl_to_dbl_fcnvuf(&fpregs[r1],0, decode_0c()
498 return(sgl_to_sgl_fcnvfu(&fpregs[r1],0, decode_0c()
501 return(sgl_to_dbl_fcnvfu(&fpregs[r1],0, decode_0c()
504 return(dbl_to_sgl_fcnvfu(&fpregs[r1],0, decode_0c()
507 return(dbl_to_dbl_fcnvfu(&fpregs[r1],0, decode_0c()
513 return(sgl_to_sgl_fcnvfut(&fpregs[r1],0, decode_0c()
516 return(sgl_to_dbl_fcnvfut(&fpregs[r1],0, decode_0c()
519 return(dbl_to_sgl_fcnvfut(&fpregs[r1],0, decode_0c()
522 return(dbl_to_dbl_fcnvfut(&fpregs[r1],0, decode_0c()
558 retval = sgl_fcmp(&fpregs[r1], decode_0c()
565 retval = dbl_fcmp(&fpregs[r1], decode_0c()
589 retval = sgl_fcmp(&fpregs[r1], decode_0c()
596 retval = dbl_fcmp(&fpregs[r1], decode_0c()
640 return(sgl_fadd(&fpregs[r1],&fpregs[r2], decode_0c()
643 return(dbl_fadd(&fpregs[r1],&fpregs[r2], decode_0c()
652 return(sgl_fsub(&fpregs[r1],&fpregs[r2], decode_0c()
655 return(dbl_fsub(&fpregs[r1],&fpregs[r2], decode_0c()
664 return(sgl_fmpy(&fpregs[r1],&fpregs[r2], decode_0c()
667 return(dbl_fmpy(&fpregs[r1],&fpregs[r2], decode_0c()
676 return(sgl_fdiv(&fpregs[r1],&fpregs[r2], decode_0c()
679 return(dbl_fdiv(&fpregs[r1],&fpregs[r2], decode_0c()
688 return(sgl_frem(&fpregs[r1],&fpregs[r2], decode_0c()
691 return(dbl_frem(&fpregs[r1],&fpregs[r2], decode_0c()
709 u_int r1,r2,t; /* operand register offsets */ decode_0e() local
718 r1 = ((extru(ir,fpr1pos,5)<<1)|(extru(ir,fpxr1pos,1))); decode_0e()
719 if (r1 == 0) decode_0e()
720 r1 = fpzeroreg; decode_0e()
734 r1 &= ~1; decode_0e()
751 fpregs[t+1] = fpregs[r1+1]; decode_0e()
753 fpregs[t] = fpregs[r1]; decode_0e()
762 fpregs[t+1] = fpregs[r1+1]; decode_0e()
764 fpregs[t] = fpregs[r1] & 0x7fffffff; decode_0e()
773 fpregs[t+1] = fpregs[r1+1]; decode_0e()
775 fpregs[t] = fpregs[r1] ^ 0x80000000; decode_0e()
784 fpregs[t+1] = fpregs[r1+1]; decode_0e()
786 fpregs[t] = fpregs[r1] | 0x80000000; decode_0e()
792 return(sgl_fsqrt(&fpregs[r1],0, decode_0e()
795 return(dbl_fsqrt(&fpregs[r1],0, decode_0e()
804 return(sgl_frnd(&fpregs[r1],0, decode_0e()
807 return(dbl_frnd(&fpregs[r1],0, decode_0e()
834 return(sgl_to_dbl_fcnvff(&fpregs[r1],0, decode_0e()
837 return(dbl_to_sgl_fcnvff(&fpregs[r1],0, decode_0e()
845 return(sgl_to_sgl_fcnvxf(&fpregs[r1],0, decode_0e()
848 return(sgl_to_dbl_fcnvxf(&fpregs[r1],0, decode_0e()
851 return(dbl_to_sgl_fcnvxf(&fpregs[r1],0, decode_0e()
854 return(dbl_to_dbl_fcnvxf(&fpregs[r1],0, decode_0e()
860 return(sgl_to_sgl_fcnvfx(&fpregs[r1],0, decode_0e()
863 return(sgl_to_dbl_fcnvfx(&fpregs[r1],0, decode_0e()
866 return(dbl_to_sgl_fcnvfx(&fpregs[r1],0, decode_0e()
869 return(dbl_to_dbl_fcnvfx(&fpregs[r1],0, decode_0e()
875 return(sgl_to_sgl_fcnvfxt(&fpregs[r1],0, decode_0e()
878 return(sgl_to_dbl_fcnvfxt(&fpregs[r1],0, decode_0e()
881 return(dbl_to_sgl_fcnvfxt(&fpregs[r1],0, decode_0e()
884 return(dbl_to_dbl_fcnvfxt(&fpregs[r1],0, decode_0e()
890 return(sgl_to_sgl_fcnvuf(&fpregs[r1],0, decode_0e()
893 return(sgl_to_dbl_fcnvuf(&fpregs[r1],0, decode_0e()
896 return(dbl_to_sgl_fcnvuf(&fpregs[r1],0, decode_0e()
899 return(dbl_to_dbl_fcnvuf(&fpregs[r1],0, decode_0e()
905 return(sgl_to_sgl_fcnvfu(&fpregs[r1],0, decode_0e()
908 return(sgl_to_dbl_fcnvfu(&fpregs[r1],0, decode_0e()
911 return(dbl_to_sgl_fcnvfu(&fpregs[r1],0, decode_0e()
914 return(dbl_to_dbl_fcnvfu(&fpregs[r1],0, decode_0e()
920 return(sgl_to_sgl_fcnvfut(&fpregs[r1],0, decode_0e()
923 return(sgl_to_dbl_fcnvfut(&fpregs[r1],0, decode_0e()
926 return(dbl_to_sgl_fcnvfut(&fpregs[r1],0, decode_0e()
929 return(dbl_to_dbl_fcnvfut(&fpregs[r1],0, decode_0e()
962 retval = sgl_fcmp(&fpregs[r1], decode_0e()
969 retval = dbl_fcmp(&fpregs[r1], decode_0e()
994 retval = sgl_fcmp(&fpregs[r1], decode_0e()
1001 retval = dbl_fcmp(&fpregs[r1], decode_0e()
1036 return(sgl_fadd(&fpregs[r1],&fpregs[r2], decode_0e()
1039 return(dbl_fadd(&fpregs[r1],&fpregs[r2], decode_0e()
1045 return(sgl_fsub(&fpregs[r1],&fpregs[r2], decode_0e()
1048 return(dbl_fsub(&fpregs[r1],&fpregs[r2], decode_0e()
1069 * impyu(&fpregs[r1],&fpregs[r2], decode_0e()
1080 return(sgl_fmpy(&fpregs[r1], decode_0e()
1083 return(dbl_fmpy(&fpregs[r1], decode_0e()
1090 return(sgl_fdiv(&fpregs[r1],&fpregs[r2], decode_0e()
1093 return(dbl_fdiv(&fpregs[r1],&fpregs[r2], decode_0e()
1099 return(sgl_frem(&fpregs[r1],&fpregs[r2], decode_0e()
1102 return(dbl_frem(&fpregs[r1],&fpregs[r2], decode_0e()
/linux-4.1.27/arch/arc/kernel/
H A Dfpu.c20 * dexcl1 0, r1, r0 ; where r1:r0 is the 64 bit val
24 * daddh11 r1, r3, r3 ; get "hi" into r1 (dpfp1 unchanged)
25 * dexcl1 r0, r1, r3 ; get "low" into r0 (dpfp1 low clobbered)
26 * dexcl1 0, r1, r0 ; restore dpfp1 to orig value
/linux-4.1.27/tools/testing/selftests/powerpc/pmu/ebb/
H A Debb_handler.S10 /* ppc-asm.h defines most of the reg aliases, but not r1/r2. */
11 #define r1 1 define
20 * Back chain ------+ <- r1 <-------+
37 * Back chain <+ <- updated r1 --------+
65 #define SAVE_GPR(n) std n,GPR_SAVE(n)(r1)
66 #define REST_GPR(n) ld n,GPR_SAVE(n)(r1)
69 #define SAVE_VSR(n, b) li b, VSR_SAVE(n); stxvd2x n,b,r1
70 #define LOAD_VSR(n, b) li b, VSR_SAVE(n); lxvd2x n,b,r1
103 stdu r1,-STACK_FRAME(r1)
106 std r0,LR_SAVE(r1)
108 std r0,CCR_SAVE(r1)
110 std r0,CTR_SAVE(r1)
112 std r0,XER_SAVE(r1)
145 stfd f0, FSCR_SAVE(r1)
147 stfd f0, VSCR_SAVE(r1)
255 lfd f0, FSCR_SAVE(r1)
257 lfd f0, VSCR_SAVE(r1)
324 ld r0,XER_SAVE(r1)
326 ld r0,CTR_SAVE(r1)
328 ld r0,LR_SAVE(r1)
330 ld r0,CCR_SAVE(r1)
363 addi r1,r1,STACK_FRAME
/linux-4.1.27/tools/testing/selftests/powerpc/stringloops/
H A Dmemcmp_64.S89 std r31,-8(r1)
90 std r30,-16(r1)
91 std r29,-24(r1)
92 std r28,-32(r1)
93 std r27,-40(r1)
182 ld r31,-8(r1)
183 ld r30,-16(r1)
184 ld r29,-24(r1)
185 ld r28,-32(r1)
186 ld r27,-40(r1)
228 ld r31,-8(r1)
229 ld r30,-16(r1)
230 ld r29,-24(r1)
231 ld r28,-32(r1)
232 ld r27,-40(r1)
/linux-4.1.27/arch/avr32/kernel/
H A Dentry-avr32b.S99 mfsr r1, SYSREG_PTBR
108 ld.w r3, r1[r2 << 2]
109 bfextu r1, r0, PAGE_SHIFT, PGDIR_SHIFT - PAGE_SHIFT
114 ld.w r2, r3[r1 << 2]
121 st.w r3[r1 << 2], r2
128 mfsr r1, SYSREG_MMUCR
135 1: bfins r1, r2, SYSREG_DRP_OFFSET, SYSREG_DRP_SIZE
136 mtsr SYSREG_MMUCR, r1
171 mov r1, lo(swapper_pg_dir)
172 orh r1, hi(swapper_pg_dir)
173 ld.w r3, r1[r2 << 2]
176 mfsr r1, SYSREG_PTBR
177 st.w r1[r2 << 2], r3
199 mfsr r1, SYSREG_RSR_SUP
204 stm --sp, r0-r1
208 ld.w r1, r0[TI_flags]
209 bld r1, TIF_SYSCALL_TRACE
231 ld.w r1, r0[TI_flags]
232 andl r1, _TIF_ALLWORK_MASK, COH
262 mov pc, r1
271 bld r1, TIF_SYSCALL_TRACE
276 ld.w r1, r0[TI_flags]
278 1: bld r1, TIF_NEED_RESCHED
283 ld.w r1, r0[TI_flags]
287 tst r1, r2
294 ld.w r1, r0[TI_flags]
297 3: bld r1, TIF_BREAKPOINT
383 ld.w r1, r0[TI_task]
384 ld.w r2, r1[TSK_active_mm]
392 mov r1, lr
490 mfsr r1, SYSREG_TLBEAR
492 lsr r2, r1, PGDIR_SHIFT
494 lsl r1, (32 - PGDIR_SHIFT)
495 lsr r1, (32 - PGDIR_SHIFT) + PAGE_SHIFT
500 add r2, r0, r1 << 2
534 ld.w r1, r0[TI_flags]
535 andl r1, _TIF_WORK_MASK, COH
553 ld.w r1, r0[TI_flags]
554 bld r1, TIF_NEED_RESCHED
565 mfsr r1, SYSREG_SR
582 ld.w r1, r0[TI_flags]
585 bld r1, TIF_NEED_RESCHED
590 ld.w r1, r0[TI_flags]
594 tst r1, r2
601 ld.w r1, r0[TI_flags]
604 2: bld r1, TIF_BREAKPOINT
633 ld.w r1, r0[TI_flags]
635 tst r1, r2
638 bld r1, TIF_SINGLE_STEP
686 mov r1, r2
691 mtsr SYSREG_SR, r1
767 ld.w r1, r0[TI_flags]
768 andl r1, _TIF_WORK_MASK, COH
793 ld.w r1, r0[TI_flags]
794 bld r1, TIF_CPU_GOING_TO_SLEEP
800 sub r1, pc, . - cpu_idle_skip_sleep
801 stdsp sp[REG_PC], r1
807 ld.w r1, r0[TI_flags]
808 bld r1, TIF_NEED_RESCHED
/linux-4.1.27/arch/powerpc/mm/
H A Dhash_low_64.S57 std r0,16(r1)
58 stdu r1,-STACKFRAMESIZE(r1)
60 std r6,STK_PARAM(R6)(r1)
61 std r8,STK_PARAM(R8)(r1)
62 std r9,STK_PARAM(R9)(r1)
71 std r27,STK_REG(R27)(r1)
72 std r28,STK_REG(R28)(r1)
73 std r29,STK_REG(R29)(r1)
74 std r30,STK_REG(R30)(r1)
75 std r31,STK_REG(R31)(r1)
169 std r3,STK_PARAM(R4)(r1)
199 ld r6,STK_PARAM(R4)(r1) /* Retrieve new pp bits */
204 ld r10,STK_PARAM(R9)(r1) /* segment size */
224 ld r6,STK_PARAM(R4)(r1) /* Retrieve new pp bits */
229 ld r10,STK_PARAM(R9)(r1) /* segment size */
267 ld r6,STK_PARAM(R6)(r1)
271 ld r27,STK_REG(R27)(r1)
272 ld r28,STK_REG(R28)(r1)
273 ld r29,STK_REG(R29)(r1)
274 ld r30,STK_REG(R30)(r1)
275 ld r31,STK_REG(R31)(r1)
276 addi r1,r1,STACKFRAMESIZE
277 ld r0,16(r1)
301 ld r8,STK_PARAM(R9)(r1) /* segment size */
302 ld r9,STK_PARAM(R8)(r1) /* get "flags" param */
326 ld r6,STK_PARAM(R6)(r1)
351 std r0,16(r1)
352 stdu r1,-STACKFRAMESIZE(r1)
354 std r6,STK_PARAM(R6)(r1)
355 std r8,STK_PARAM(R8)(r1)
356 std r9,STK_PARAM(R9)(r1)
367 std r25,STK_REG(R25)(r1)
368 std r26,STK_REG(R26)(r1)
369 std r27,STK_REG(R27)(r1)
370 std r28,STK_REG(R28)(r1)
371 std r29,STK_REG(R29)(r1)
372 std r30,STK_REG(R30)(r1)
373 std r31,STK_REG(R31)(r1)
485 std r3,STK_PARAM(R4)(r1)
506 ld r6,STK_PARAM(R6)(r1)
528 ld r6,STK_PARAM(R4)(r1) /* Retrieve new pp bits */
533 ld r10,STK_PARAM(R9)(r1) /* segment size */
557 ld r6,STK_PARAM(R4)(r1) /* Retrieve new pp bits */
562 ld r10,STK_PARAM(R9)(r1) /* segment size */
597 ld r7,STK_PARAM(R9)(r1) /* ssize */
598 ld r8,STK_PARAM(R8)(r1) /* flags */
614 ld r6,STK_PARAM(R6)(r1)
635 ld r25,STK_REG(R25)(r1)
636 ld r26,STK_REG(R26)(r1)
637 ld r27,STK_REG(R27)(r1)
638 ld r28,STK_REG(R28)(r1)
639 ld r29,STK_REG(R29)(r1)
640 ld r30,STK_REG(R30)(r1)
641 ld r31,STK_REG(R31)(r1)
642 addi r1,r1,STACKFRAMESIZE
643 ld r0,16(r1)
669 ld r8,STK_PARAM(R9)(r1) /* segment size */
670 ld r9,STK_PARAM(R8)(r1) /* get "flags" param */
684 ld r6,STK_PARAM(R6)(r1)
697 ld r6,STK_PARAM(R6)(r1)
714 std r0,16(r1)
715 stdu r1,-STACKFRAMESIZE(r1)
717 std r6,STK_PARAM(R6)(r1)
718 std r8,STK_PARAM(R8)(r1)
719 std r9,STK_PARAM(R9)(r1)
728 std r27,STK_REG(R27)(r1)
729 std r28,STK_REG(R28)(r1)
730 std r29,STK_REG(R29)(r1)
731 std r30,STK_REG(R30)(r1)
732 std r31,STK_REG(R31)(r1)
830 std r3,STK_PARAM(R4)(r1)
863 ld r6,STK_PARAM(R4)(r1) /* Retrieve new pp bits */
868 ld r10,STK_PARAM(R9)(r1) /* segment size */
888 ld r6,STK_PARAM(R4)(r1) /* Retrieve new pp bits */
893 ld r10,STK_PARAM(R9)(r1) /* segment size */
931 ld r6,STK_PARAM(R6)(r1)
935 ld r27,STK_REG(R27)(r1)
936 ld r28,STK_REG(R28)(r1)
937 ld r29,STK_REG(R29)(r1)
938 ld r30,STK_REG(R30)(r1)
939 ld r31,STK_REG(R31)(r1)
940 addi r1,r1,STACKFRAMESIZE
941 ld r0,16(r1)
965 ld r8,STK_PARAM(R9)(r1) /* segment size */
966 ld r9,STK_PARAM(R8)(r1) /* get "flags" param */
990 ld r6,STK_PARAM(R6)(r1)
/linux-4.1.27/drivers/block/paride/
H A Dfit2.c56 w0(0); a = r1(); fit2_read_regr()
57 w0(1); b = r1(); fit2_read_regr()
73 w0(0); a = r1(); w0(1); b = r1(); fit2_read_block()
74 w0(3); c = r1(); w0(2); d = r1(); fit2_read_block()
79 a = r1(); w0(3); b = r1(); fit2_read_block()
80 w0(1); c = r1(); w0(0); d = r1(); fit2_read_block()
H A Dcomm.c54 w2(6); l = r1(); w0(0x80); h = r1(); w2(4); comm_read_regr()
63 case 4: w3(r+0x20); (void)r1(); comm_read_regr()
85 case 4: w3(r); (void)r1(); w4(val); comm_write_regr()
116 w0(0); w2(6); l = r1(); comm_read_block()
117 w0(0x80); h = r1(); w2(4); comm_read_block()
129 case 2: w3(0x68); (void)r1(); w2(0x24); comm_read_block()
134 case 3: w3(0x68); (void)r1(); w2(0x24); comm_read_block()
139 case 4: w3(0x68); (void)r1(); w2(0x24); comm_read_block()
163 case 2: w3(0x48); (void)r1(); comm_write_block()
167 case 3: w3(0x48); (void)r1(); comm_write_block()
171 case 4: w3(0x48); (void)r1(); comm_write_block()
H A Dktti.c49 a = r1(); w2(0xc); b = r1(); w2(9); w2(0xc); w2(9); ktti_read_regr()
60 a = r1(); w2(0xc); b = r1(); w2(9); ktti_read_block()
62 a = r1(); w2(0xc); b = r1(); w2(9); ktti_read_block()
/linux-4.1.27/arch/score/include/asm/
H A Dasmmacro.h22 .set r1
23 sw r1, [r0, PT_R1]
92 .set r1
93 ldis r1, 0x00ff
94 and r30, r30, r1
95 not r1, r1 variable
97 and r31, r31, r1
119 .set r1
120 lw r1, [r0, PT_R1]
/linux-4.1.27/arch/arm/common/
H A Dvlock.S58 @ r1: CPU ID (0-based index within cluster)
60 add r1, r1, #VLOCK_VOTING_OFFSET
62 voting_begin r0, r1, r2
70 strb r1, [r0, #VLOCK_OWNER_OFFSET] @ submit my vote
72 voting_end r0, r1, r2 @ implies DMB
91 eor r0, r1, r2 @ zero if I won, else nonzero
95 voting_end r0, r1, r2
103 mov r1, #VLOCK_OWNER_NONE
104 strb r1, [r0, #VLOCK_OWNER_OFFSET]
/linux-4.1.27/arch/sh/kernel/
H A Drelocate_kernel.S56 mov.l r1, @-r15
71 mov.l r1, @-r15
98 mov.l @r15+, r1
112 mov.l @r15+, r1
156 mov #-16,r1
157 and r1,r2
194 mov.l @(0, r5), r1
196 mov.l r1, @(0, r2)
199 mov.l @(4, r5), r1
201 mov.l r1, @(4, r2)
204 mov.l @(8, r5), r1
206 mov.l r1, @(8, r2)
209 mov.l @(12, r5), r1
211 mov.l r1, @(12, r2)
/linux-4.1.27/arch/unicore32/boot/compressed/
H A Dhead.S29 ldm (r1, r2, r3, r5, r6, r7, r8), [r0]+
31 sub.a r0, r0, r1 @ calculate the delta offset
66 1001: ldw r1, [r7+], #0
67 add r1, r1, r0
68 stw.w r1, [r7]+, #4
100 mov r1, sp @ malloc space above stack
133 * r1: free_mem_ptr_p
162 LC0: .word LC0 @ r1
174 2001: ldb.w r1, [r0]+, #1
175 csub.a r1, #0
182 movc p1.c1, r1, #1
183 csub.a r1, #'\n'
184 cmoveq r1, #'\r'
/linux-4.1.27/arch/powerpc/kvm/
H A Dbooke_interrupts.S219 stw r1, VCPU_GPR(R1)(r4)
233 lwz r1, VCPU_HOST_STACK(r4)
251 lwz r3, HOST_RUN(r1)
252 lwz r2, HOST_R2(r1)
321 lwz r14, HOST_NV_GPR(R14)(r1)
322 lwz r15, HOST_NV_GPR(R15)(r1)
323 lwz r16, HOST_NV_GPR(R16)(r1)
324 lwz r17, HOST_NV_GPR(R17)(r1)
325 lwz r18, HOST_NV_GPR(R18)(r1)
326 lwz r19, HOST_NV_GPR(R19)(r1)
327 lwz r20, HOST_NV_GPR(R20)(r1)
328 lwz r21, HOST_NV_GPR(R21)(r1)
329 lwz r22, HOST_NV_GPR(R22)(r1)
330 lwz r23, HOST_NV_GPR(R23)(r1)
331 lwz r24, HOST_NV_GPR(R24)(r1)
332 lwz r25, HOST_NV_GPR(R25)(r1)
333 lwz r26, HOST_NV_GPR(R26)(r1)
334 lwz r27, HOST_NV_GPR(R27)(r1)
335 lwz r28, HOST_NV_GPR(R28)(r1)
336 lwz r29, HOST_NV_GPR(R29)(r1)
337 lwz r30, HOST_NV_GPR(R30)(r1)
338 lwz r31, HOST_NV_GPR(R31)(r1)
341 lwz r4, HOST_STACK_LR(r1)
342 lwz r5, HOST_CR(r1)
343 addi r1, r1, HOST_STACK_SIZE
355 stwu r1, -HOST_STACK_SIZE(r1)
356 stw r1, VCPU_HOST_STACK(r4) /* Save stack pointer to vcpu. */
359 stw r3, HOST_RUN(r1)
361 stw r3, HOST_STACK_LR(r1)
363 stw r5, HOST_CR(r1)
366 stw r14, HOST_NV_GPR(R14)(r1)
367 stw r15, HOST_NV_GPR(R15)(r1)
368 stw r16, HOST_NV_GPR(R16)(r1)
369 stw r17, HOST_NV_GPR(R17)(r1)
370 stw r18, HOST_NV_GPR(R18)(r1)
371 stw r19, HOST_NV_GPR(R19)(r1)
372 stw r20, HOST_NV_GPR(R20)(r1)
373 stw r21, HOST_NV_GPR(R21)(r1)
374 stw r22, HOST_NV_GPR(R22)(r1)
375 stw r23, HOST_NV_GPR(R23)(r1)
376 stw r24, HOST_NV_GPR(R24)(r1)
377 stw r25, HOST_NV_GPR(R25)(r1)
378 stw r26, HOST_NV_GPR(R26)(r1)
379 stw r27, HOST_NV_GPR(R27)(r1)
380 stw r28, HOST_NV_GPR(R28)(r1)
381 stw r29, HOST_NV_GPR(R29)(r1)
382 stw r30, HOST_NV_GPR(R30)(r1)
383 stw r31, HOST_NV_GPR(R31)(r1)
414 stw r2, HOST_R2(r1)
449 lwz r1, VCPU_GPR(R1)(r4)
H A Dbookehv_interrupts.S73 PPC_STL r1, VCPU_GPR(R1)(r4)
75 PPC_LL r1, VCPU_HOST_STACK(r4)
76 PPC_LL r2, HOST_R2(r1)
444 PPC_LL r3, HOST_RUN(r1)
481 PPC_LL r5, HOST_STACK_LR(r1)
482 lwz r6, HOST_CR(r1)
508 PPC_LL r14, HOST_NV_GPR(R14)(r1)
509 PPC_LL r15, HOST_NV_GPR(R15)(r1)
510 PPC_LL r16, HOST_NV_GPR(R16)(r1)
511 PPC_LL r17, HOST_NV_GPR(R17)(r1)
512 PPC_LL r18, HOST_NV_GPR(R18)(r1)
513 PPC_LL r19, HOST_NV_GPR(R19)(r1)
514 PPC_LL r20, HOST_NV_GPR(R20)(r1)
515 PPC_LL r21, HOST_NV_GPR(R21)(r1)
516 PPC_LL r22, HOST_NV_GPR(R22)(r1)
517 PPC_LL r23, HOST_NV_GPR(R23)(r1)
518 PPC_LL r24, HOST_NV_GPR(R24)(r1)
519 PPC_LL r25, HOST_NV_GPR(R25)(r1)
520 PPC_LL r26, HOST_NV_GPR(R26)(r1)
521 PPC_LL r27, HOST_NV_GPR(R27)(r1)
522 PPC_LL r28, HOST_NV_GPR(R28)(r1)
523 PPC_LL r29, HOST_NV_GPR(R29)(r1)
524 PPC_LL r30, HOST_NV_GPR(R30)(r1)
525 PPC_LL r31, HOST_NV_GPR(R31)(r1)
530 addi r1, r1, HOST_STACK_SIZE
539 stwu r1, -HOST_STACK_SIZE(r1)
540 PPC_STL r1, VCPU_HOST_STACK(r4) /* Save stack pointer to vcpu. */
543 PPC_STL r3, HOST_RUN(r1)
546 PPC_STL r3, HOST_STACK_LR(r1)
548 stw r5, HOST_CR(r1)
551 PPC_STL r14, HOST_NV_GPR(R14)(r1)
552 PPC_STL r15, HOST_NV_GPR(R15)(r1)
553 PPC_STL r16, HOST_NV_GPR(R16)(r1)
554 PPC_STL r17, HOST_NV_GPR(R17)(r1)
555 PPC_STL r18, HOST_NV_GPR(R18)(r1)
556 PPC_STL r19, HOST_NV_GPR(R19)(r1)
557 PPC_STL r20, HOST_NV_GPR(R20)(r1)
558 PPC_STL r21, HOST_NV_GPR(R21)(r1)
559 PPC_STL r22, HOST_NV_GPR(R22)(r1)
560 PPC_STL r23, HOST_NV_GPR(R23)(r1)
561 PPC_STL r24, HOST_NV_GPR(R24)(r1)
562 PPC_STL r25, HOST_NV_GPR(R25)(r1)
563 PPC_STL r26, HOST_NV_GPR(R26)(r1)
564 PPC_STL r27, HOST_NV_GPR(R27)(r1)
565 PPC_STL r28, HOST_NV_GPR(R28)(r1)
566 PPC_STL r29, HOST_NV_GPR(R29)(r1)
567 PPC_STL r30, HOST_NV_GPR(R30)(r1)
568 PPC_STL r31, HOST_NV_GPR(R31)(r1)
592 PPC_STL r2, HOST_R2(r1)
650 PPC_LL r1, VCPU_GPR(R1)(r4)
H A Dbook3s_interrupts.S76 PPC_STL r0,PPC_LR_STKOFF(r1)
79 PPC_STLU r1, -SWITCH_FRAME_SIZE(r1)
82 SAVE_2GPRS(3, r1)
85 SAVE_NVGPRS(r1)
89 stw r14, _CCR(r1)
92 PPC_STL r0, _LINK(r1)
102 REST_GPR(4, r1)
159 PPC_LL r3, GPR4(r1) /* vcpu pointer */
184 PPC_LL r7, GPR4(r1)
209 REST_2GPRS(3, r1)
221 PPC_LL r4, _LINK(r1)
224 lwz r14, _CCR(r1)
228 REST_NVGPRS(r1)
230 addi r1, r1, SWITCH_FRAME_SIZE
235 PPC_LL r4, _LINK(r1)
236 PPC_STL r4, (PPC_LR_STKOFF + SWITCH_FRAME_SIZE)(r1)
239 REST_2GPRS(3, r1)
250 REST_GPR(4, r1)
H A Dbook3s_hv_interrupts.S44 std r0,PPC_LR_STKOFF(r1)
47 stdu r1, -SWITCH_FRAME_SIZE(r1)
50 SAVE_NVGPRS(r1)
52 std r3, _CCR(r1)
151 REST_NVGPRS(r1)
152 ld r4, _CCR(r1)
155 addi r1, r1, SWITCH_FRAME_SIZE
156 ld r0, PPC_LR_STKOFF(r1)
/linux-4.1.27/arch/blackfin/lib/
H A Ddivsi3.S45 r1 = abs r1; /* now both positive, r3.30 means "negate result", define
48 cc = r0 < r1;
50 r2 = r1 >> 15;
53 r2 = r1 << 16;
76 r1 = r3 >> 31; /* add overflow issue back in */ define
77 r0 = r0 + r1;
78 r1 = -r0; define
80 if cc r0 = r1;
138 r1 = r3 >> 31; define
139 r2 = r2 + r1;
188 r1 = r3 >> 31; define
189 r0 = r0 + r1;
/linux-4.1.27/arch/sh/kernel/cpu/shmobile/
H A Dsleep.S25 #define k1 r1
131 mov #4, r1
132 mov.l r1, @r0
144 mov #0, r1
145 mov.l r1, @r0
165 mov #0x80, r1
174 mov.l @(SH_SLEEP_RESUME, r5), r1
175 mov.l r1, @r0
179 mov #0x20, r1
187 mov #0x10, r1
192 mov #0x00, r1
198 mov.l r1, @r0
206 mov.l @(r0, r5), r1
208 mov.l @r1, r1
210 mov.l r1, @(r0, r5)
236 mov.l r1, @-r15
375 mov.l @(r0, r5), r1
379 mov.l r1, @r0
393 mov.l @r15+, r1
/linux-4.1.27/arch/s390/boot/compressed/
H A Dhead.S31 la %r1,0x200
32 mvc 0(mover_end-mover,%r1),mover-.LPG1(%r13)
35 br %r1
/linux-4.1.27/arch/blackfin/kernel/
H A Dftrace-entry.S39 [--sp] = r1;
48 r1 = [sp + 16]; /* skip the 4 local regs on stack */ define
64 r1 = [sp++]; define
120 [--sp] = r1;
132 r1 = [sp + 16]; /* skip the 4 local regs on stack */ define
141 r1 = [sp++]; define
162 [--sp] = r1;
167 r1 = rets; /* unsigned long self_addr */ define
170 r1 = [sp]; /* unsigned long self_addr */ define
176 r1 += -MCOUNT_INSN_SIZE;
190 [--sp] = r1;
202 r1 = [sp++]; define
/linux-4.1.27/security/selinux/ss/
H A Dmls_types.h47 #define mls_range_contains(r1, r2) \
48 (mls_level_dom(&(r2).level[0], &(r1).level[0]) && \
49 mls_level_dom(&(r1).level[1], &(r2).level[1]))
/linux-4.1.27/kernel/bpf/
H A Dhelpers.c26 static u64 bpf_map_lookup_elem(u64 r1, u64 r2, u64 r3, u64 r4, u64 r5) bpf_map_lookup_elem() argument
32 struct bpf_map *map = (struct bpf_map *) (unsigned long) r1; bpf_map_lookup_elem()
54 static u64 bpf_map_update_elem(u64 r1, u64 r2, u64 r3, u64 r4, u64 r5) bpf_map_update_elem() argument
56 struct bpf_map *map = (struct bpf_map *) (unsigned long) r1; bpf_map_update_elem()
75 static u64 bpf_map_delete_elem(u64 r1, u64 r2, u64 r3, u64 r4, u64 r5) bpf_map_delete_elem() argument
77 struct bpf_map *map = (struct bpf_map *) (unsigned long) r1; bpf_map_delete_elem()
93 static u64 bpf_get_prandom_u32(u64 r1, u64 r2, u64 r3, u64 r4, u64 r5) bpf_get_prandom_u32() argument
104 static u64 bpf_get_smp_processor_id(u64 r1, u64 r2, u64 r3, u64 r4, u64 r5) bpf_get_smp_processor_id() argument
/linux-4.1.27/firmware/av7110/
H A DBoot.S61 ldr r1, wait_address
66 ldmia r1!, {r5-r12}
68 cmp r1, r2
72 wait: ldrh r1, [r4] // wait for flag!=0
73 cmp r1, #0
76 mov r1, r13 // buffer address
87 ldmia r1!, {r5-r12}
89 ldmia r1!, {r5-r12}
/linux-4.1.27/arch/unicore32/lib/
H A Dbacktrace.S40 adr r1, 1b
41 sub offset, r0, r1
74 ldw r1, [frame+], #-4 @ get saved lr
78 ldw r1, [sv_pc+], #-4 @ if stmfd sp, {args} exists,
80 cxor.a r3, r1 >> #14
86 1004: ldw r1, [sv_pc+], #0 @ if stmfd {, fp, ip, lr, pc}
88 cxor.a r3, r1 >> #14
101 mov r1, frame
121 mov instr, r1
134 cmoveq r1, #'\n'
135 cmovne r1, #' '
/linux-4.1.27/arch/ia64/kernel/
H A Dminstate.h50 mov r20=r1; /* A */ \
62 adds r1=-IA64_TASK_THREAD_ON_USTACK_OFFSET,r16 \
72 (pUStk) addl r22=IA64_RBS_OFFSET,r1; /* compute base of RBS */ \
73 (pKStk) mov r1=sp; /* get sp */ \
76 (pUStk) addl r1=IA64_STK_OFFSET-IA64_PT_REGS_SIZE,r1; /* compute base of memory stack */ \
80 (pKStk) addl r1=-IA64_PT_REGS_SIZE,r1; /* if in kernel mode, use sp (r12) */ \
84 adds r17=2*L1_CACHE_BYTES,r1; /* really: biggest cache-line size */ \
85 adds r16=PT(CR_IPSR),r1; \
95 adds r16=PT(R8),r1; /* initialize first base pointer */ \
96 adds r17=PT(R9),r1; /* initialize second base pointer */ \
129 .mem.offset 0,0; st8.spill [r16]=r20,16; /* save original r1 */ \
131 adds r12=-16,r1; /* switch to kernel memory stack (with 16 bytes of scratch) */ \
143 adds r2=IA64_PT_REGS_R16_OFFSET,r1; \
146 movl r1=__gp; /* establish kernel global pointer */ \
/linux-4.1.27/tools/perf/arch/s390/util/
H A Ddwarf-regs.c15 "%r0", "%r1", "%r2", "%r3", "%r4", "%r5", "%r6", "%r7",
/linux-4.1.27/arch/powerpc/math-emu/
H A Dudivmodti4.c13 _FP_W_TYPE q0, q1, r0, r1; _fp_udivmodti4() local
42 r1 = 0; _fp_udivmodti4()
112 r1 = 0; _fp_udivmodti4()
126 r1 = n1; _fp_udivmodti4()
154 r1 = n1; _fp_udivmodti4()
184 r1 = n1 >> bm; _fp_udivmodti4()
190 r[0] = r0, r[1] = r1; _fp_udivmodti4()
/linux-4.1.27/arch/m32r/kernel/
H A Dentry.S32 * @(0x14,sp) - r1
134 ld r1, R1(r8)
135 jl r1
242 mv r1, r9 ; arg2 : __u32 thread_info_flags
253 ld r1, R1(sp)
302 mv r1, sp ; arg1(regs)
336 mv r0, r1 ; arg0(regs)
404 push r1
409 mv r0, r1
416 pop r1
433 srl3 r1, r4, #4
435 and3 r1, r1, #0x0000ffff
438 beqz r1, inst
441 srli r1, #1
444 and3 r1, r4, #2
445 srli r1, #1
446 or3 r1, r1, #8
454 or r1, r3
458 * r1 : unsigned long error-code
479 ldi r1, #0x30 ; error_code
489 ldi r1, #0x20 ; error_code
498 ldi r1, #0 ; error_code ; FIXME
510 ldi r1, #0 ; error_code
519 ldi r1, #0 ; error_code ; FIXME
529 push r1
545 pop r1
/linux-4.1.27/arch/arm/mach-sa1100/
H A Dsleep.S34 mcr p15, 0, r1, c15, c2, 2
57 mov r1, #0
58 str r1, [r5]
66 * r1 = &MSC1
82 ldr r1, =MSC1
89 ldr r4, [r1]
121 str r4, [r1]

Completed in 2153 milliseconds

1234