/linux-4.4.14/arch/arm/crypto/ |
D | aes-armv4.S | 162 orr r0,r0,r4,lsl#8 164 orr r0,r0,r5,lsl#16 166 orr r0,r0,r6,lsl#24 169 orr r1,r1,r4,lsl#8 171 orr r1,r1,r5,lsl#16 173 orr r1,r1,r6,lsl#24 176 orr r2,r2,r4,lsl#8 178 orr r2,r2,r5,lsl#16 180 orr r2,r2,r6,lsl#24 183 orr r3,r3,r4,lsl#8 [all …]
|
D | sha1-armv4-large.S | 61 add r2,r1,r2,lsl#6 @ r2 to point at the end of r1 77 orr r9,r9,r10,lsl#8 79 orr r9,r9,r11,lsl#16 81 orr r9,r9,r12,lsl#24 102 orr r9,r9,r10,lsl#8 104 orr r9,r9,r11,lsl#16 106 orr r9,r9,r12,lsl#24 127 orr r9,r9,r10,lsl#8 129 orr r9,r9,r11,lsl#16 131 orr r9,r9,r12,lsl#24 [all …]
|
D | sha512-core.S_shipped | 148 add r2,r1,r2,lsl#7 @ len to point at the end of inp 191 orr r3,r3,r9,lsl#8 193 orr r3,r3,r10,lsl#16 195 orr r3,r3,r11,lsl#24 196 orr r4,r4,r12,lsl#8 197 orr r4,r4,r9,lsl#16 198 orr r4,r4,r10,lsl#24 214 eor r9,r9,r8,lsl#18 216 eor r10,r10,r7,lsl#18 220 eor r9,r9,r8,lsl#14 [all …]
|
D | sha256-core.S_shipped | 102 add r2,r1,r2,lsl#6 @ len to point at the end of inp 131 orr r2,r2,r12,lsl#8 133 orr r2,r2,r0,lsl#16 138 orr r2,r2,r12,lsl#24 189 orr r2,r2,r3,lsl#8 191 orr r2,r2,r0,lsl#16 196 orr r2,r2,r3,lsl#24 247 orr r2,r2,r12,lsl#8 249 orr r2,r2,r0,lsl#16 254 orr r2,r2,r12,lsl#24 [all …]
|
D | aes-ce-core.S | 154 add ip, \rk, \rounds, lsl #4
|
D | aesbs-core.S_shipped | 1084 sub r12, sp, r10, lsl#7 @ 128 bytes per inner round key 1340 sub r12, sp, r10, lsl#7 @ 128 bytes per inner round key 1573 sub r12, sp, r1, lsl#7 @ 128 bytes per inner round key 2063 sub r12, sp, r1, lsl#7 @ 128 bytes per inner round key
|
/linux-4.4.14/arch/arm/lib/ |
D | io-writesw-armv3.S | 28 orr r3, r3, r3, lsl #16 46 mov ip, r3, lsl #16 51 orr ip, ip, ip, lsl #16 54 mov ip, r4, lsl #16 59 orr ip, ip, ip, lsl #16 62 mov ip, r5, lsl #16 67 orr ip, ip, ip, lsl #16 70 mov ip, r6, lsl #16 75 orr ip, ip, ip, lsl #16 89 mov ip, r3, lsl #16 [all …]
|
D | bitops.h | 13 add r1, r1, r0, lsl #2 @ Get word offset 19 mov r3, r2, lsl r3 38 add r1, r1, r0, lsl #2 @ Get word offset 39 mov r3, r2, lsl r3 @ create mask 68 mov r3, r3, lsl r2 70 ldr r2, [r1, r0, lsl #2] 72 str r2, [r1, r0, lsl #2] 95 ldr r2, [r1, r0, lsl #2]! 97 tst r2, r0, lsl r3 98 \instr r2, r2, r0, lsl r3
|
D | io-readsw-armv3.S | 39 orr ip, ip, ip, lsl #8 48 orr r3, r3, r4, lsl #16 53 orr r4, r4, r5, lsl #16 58 orr r5, r5, r6, lsl #16 63 orr r6, r6, lr, lsl #16 79 orr r3, r3, r4, lsl #16 84 orr r4, r4, r5, lsl #16 94 orr r3, r3, r4, lsl #16
|
D | io-readsw-armv4.S | 15 orr \rd, \hw1, \hw2, lsl #16 17 orr \rd, \hw2, \hw1, lsl #16 21 .Linsw_align: movs ip, r1, lsl #31 71 .Lno_insw_4: movs r2, r2, lsl #31 88 #define pull_hbyte1 lsl #24 92 #define push_hbyte0 lsl #24 109 _BE_ONLY_( mov ip, ip, lsl #24 ) 114 orr ip, ip, r3, lsl #8
|
D | div64.S | 72 mov ip, ip, lsl yl 73 mov yl, r4, lsl yl 81 movcc yl, yl, lsl #1 82 movcc ip, ip, lsl #1 107 4: movs xl, xl, lsl #1 130 mov xl, xl, lsl xh 135 7: movs xl, xl, lsl #1 183 ARM( orr yl, yl, xh, lsl ip ) 184 THUMB( lsl xh, xh, ip ) 186 mov xh, xl, lsl ip
|
D | muldi3.S | 35 bic xl, xl, ip, lsl #16 36 bic yl, yl, yh, lsl #16 41 adds xl, xl, yh, lsl #16 43 adds xl, xl, ip, lsl #16
|
D | ashldi3.S | 45 movmi ah, ah, lsl r2 46 movpl ah, al, lsl r3 50 mov al, al, lsl r2
|
D | io-writesw-armv4.S | 25 .Loutsw_align: movs ip, r1, lsl #31 58 .Lno_outsw_4: movs r2, r2, lsl #31 70 #define pull_hbyte0 lsl #8 74 #define push_hbyte1 lsl #8
|
D | lib1funcs.S | 48 mov \divisor, \divisor, lsl \result 49 mov \curbit, \curbit, lsl \result 59 moveq \divisor, \divisor, lsl #3 69 movlo \divisor, \divisor, lsl #4 70 movlo \curbit, \curbit, lsl #4 77 movlo \divisor, \divisor, lsl #1 78 movlo \curbit, \curbit, lsl #1 144 mov \divisor, \divisor, lsl \order 156 movlo \divisor, \divisor, lsl #4 164 movlo \divisor, \divisor, lsl #1
|
D | csumpartial.S | 61 orr td0, td0, td3, lsl #8 63 orr td0, td3, td0, lsl #8 94 orrne td0, td0, ip, lsl #8 96 orrne td0, ip, td0, lsl #8
|
D | getuser.S | 54 orr r2, r2, rb, lsl #8 56 orr r2, rb, r2, lsl #8 113 orr r3, rb, r3, lsl #8
|
D | memset.S | 28 1: orr r1, r1, r1, lsl #8 29 orr r1, r1, r1, lsl #16 89 movs r8, r8, lsl #(32 - 4)
|
D | memzero.S | 100 movs ip, ip, lsl #(32 - 4) 103 movs ip, ip, lsl #2
|
D | ashrdi3.S | 47 ARM( orrmi al, al, ah, lsl ip )
|
D | lshrdi3.S | 47 ARM( orrmi al, al, ah, lsl ip )
|
D | backtrace.S | 125 ARM( tst instr, r3, lsl reg ) 126 THUMB( lsl r3, reg )
|
D | copy_template.S | 126 lsl ip, ip, #LDR1W_SHIFT 143 lsl ip, ip, #STR1W_SHIFT - LDR1W_SHIFT 167 8: movs r2, r2, lsl #31
|
D | io-writesl.S | 30 2: movs r2, r2, lsl #31
|
D | findbit.S | 185 movne r3, r3, lsl #4 188 movne r3, r3, lsl #2
|
D | io-readsl.S | 30 2: movs r2, r2, lsl #31
|
D | memmove.S | 116 8: movs r2, r2, lsl #31
|
/linux-4.4.14/drivers/scsi/arm/ |
D | acornscsi-io.S | 34 orr r3, r3, r4, lsl #16 36 orr r4, r4, r6, lsl #16 39 orr r5, r5, r6, lsl #16 41 orr r6, r6, ip, lsl #16 50 orr r3, r3, r4, lsl #16 52 orr r4, r4, r6, lsl #16 61 orr r3, r3, r4, lsl #16 85 mov r3, r4, lsl #16 88 orr r4, r4, r4, lsl #16 89 mov r5, r6, lsl #16 [all …]
|
/linux-4.4.14/sound/oss/ |
D | vidc_fill.S | 25 and r4, ip, r4, lsl #8 26 orr r4, r4, r4, lsl #16 38 and r4, ip, r4, lsl #8 39 orr r4, r4, r5, lsl #16 51 and r4, ip, r4, lsl #8 52 orr r4, r4, r4, lsl #16 64 and r4, ip, r4, lsl #8 65 orr r4, r4, r5, lsl #16 79 orr r4, r4, r4, lsl #16 83 andlt r4, r5, ip, lsl #16 [all …]
|
/linux-4.4.14/arch/arm/boot/compressed/ |
D | ll_char_wr.S | 43 mov r1, r1, lsl #3 59 add r0, r0, r5, lsl #3 @ Move to bottom of character 68 ldr r7, [lr, r7, lsl #2] 73 ldr r7, [lr, r7, lsl #2] 87 ldr ip, [lr, ip, lsl #2] 90 ldr ip, [lr, ip, lsl #2] @ avoid r4 97 ldr ip, [lr, ip, lsl #2] 100 ldr ip, [lr, ip, lsl #2] @ avoid r4
|
D | head.S | 239 orr r9, r9, lr, lsl #8 242 orr r9, r9, lr, lsl #16 243 orr r9, r9, r10, lsl #24 689 mov r9, r9, lsl #18 @ start of RAM 713 orr r1, r1, r2, lsl #20 714 add r0, r3, r2, lsl #2 1179 ARM( orr r11, r10, r9, lsl r5 ) @ factor way and cache number into r11 1180 ARM( orr r11, r11, r7, lsl r2 ) @ factor index number into r11 1181 THUMB( lsl r6, r9, r5 ) 1183 THUMB( lsl r6, r7, r2 ) [all …]
|
/linux-4.4.14/arch/arm/include/asm/ |
D | assembler.h | 37 #define lspush lsl 38 #define get_byte_0 lsl #0 42 #define put_byte_0 lsl #0 43 #define put_byte_1 lsl #8 44 #define put_byte_2 lsl #16 45 #define put_byte_3 lsl #24 47 #define lspull lsl 52 #define get_byte_3 lsl #0 53 #define put_byte_0 lsl #24 54 #define put_byte_1 lsl #16 [all …]
|
/linux-4.4.14/arch/avr32/lib/ |
D | __avr32_lsl64.S | 23 lsl r10, r10, r12 24 lsl r11, r11, r12 29 lsl r11, r10, r9
|
D | findbit.S | 37 lsl r8, 2 101 lsl r8, 2 131 lsl r8, 2 161 lsl r8, 2
|
D | csum_partial.S | 39 lsl r9, 16 41 lsl r8, 8
|
D | memcpy.S | 40 lsl r10, 2 55 lsl r9, 2
|
D | csum_partial_copy_generic.S | 62 lsl r5, 8 68 lsl r5, r5, r4
|
D | __avr32_asr64.S | 22 lsl r8, r11, r9
|
D | __avr32_lsr64.S | 22 lsl r8, r11, r9
|
D | memset.S | 65 lsl r9, 1
|
D | clear_user.S | 51 lsl r9, 2
|
/linux-4.4.14/arch/arm64/kernel/vdso/ |
D | gettimeofday.S | 69 lsl x13, x13, x12 143 lsl x14, x14, x12 148 movk x15, #NSEC_PER_SEC_HI16, lsl #16 149 lsl x15, x15, x12 232 movn x15, #0xff00, lsl #48 238 movk x11, #NSEC_PER_SEC_HI16, lsl #16 239 lsl x11, x11, x12
|
/linux-4.4.14/arch/arm/mm/ |
D | abort-lv4t.S | 63 and r9, r8, r7, lsl #1 65 and r9, r8, r7, lsl #2 67 and r9, r8, r7, lsl #3 75 subne r7, r7, r6, lsl #2 @ Undo increment 76 addeq r7, r7, r6, lsl #2 @ Undo decrement 88 ldreq r6, [r2, r9, lsl #2] @ { load Rm value } 102 movs r6, r8, lsl #20 @ Get offset 117 ldr r6, [r2, r7, lsl #2] @ Get register 'Rm' 125 mov r6, r6, lsl r9 @ 0: LSL #!0 204 addeq r7, r7, r6, lsl #2 @ increment SP if PUSH [all …]
|
D | tlb-v7.S | 46 orr r0, r3, r0, lsl #PAGE_SHIFT @ Create initial MVA 47 mov r1, r1, lsl #PAGE_SHIFT 75 mov r0, r0, lsl #PAGE_SHIFT 76 mov r1, r1, lsl #PAGE_SHIFT
|
D | tlb-v6.S | 44 orr r0, r3, r0, lsl #PAGE_SHIFT @ Create initial MVA 45 mov r1, r1, lsl #PAGE_SHIFT 74 mov r0, r0, lsl #PAGE_SHIFT 75 mov r1, r1, lsl #PAGE_SHIFT
|
D | proc-v7-3level.S | 61 orr rpgdh, rpgdh, r2, lsl #(48 - 32) @ upper 32-bits of pgd 146 mov \ttbr1, \ttbr1, lsl #12
|
D | cache-v7.S | 55 mov r5, r3, lsl r1 56 mov r6, r2, lsl r0 151 ARM( orr r11, r10, r4, lsl r5 ) @ factor way and cache number into r11 152 THUMB( lsl r6, r4, r5 ) 154 ARM( orr r11, r11, r9, lsl r2 ) @ factor index number into r11 155 THUMB( lsl r6, r9, r2 )
|
D | proc-arm740.S | 83 orr r0, r0, r4, lsl #1 @ the area register value 96 orr r0, r0, r4, lsl #1 @ the area register value
|
D | proc-macros.S | 77 mov \reg, \reg, lsl \tmp @ actual cache line size 88 mov \reg, \reg, lsl \tmp @ actual cache line size 359 orr \dest, \addr, \dest, lsl #1 @ mask in the region size
|
D | proc-feroceon.S | 68 mov r2, r2, lsl r0 @ actual cache size
|
/linux-4.4.14/arch/m68k/math-emu/ |
D | fp_movem.S | 157 3: lsl.b #1,%d1 164 lsl.w #1,%d2 165 lsl.l #7,%d2 166 lsl.l #8,%d2 176 4: lsl.b #1,%d1 307 lsl.l #5,%d1 312 3: lsl.b #1,%d1 320 lsl.l #5,%d1 331 3: lsl.b #1,%d1 341 4: lsl.b #1,%d1
|
D | fp_util.S | 131 lsl.l #8,%d0 | shift mantissa 168 lsl.l #8,%d0 | shift high mantissa 169 lsl.l #3,%d0 182 lsl.l #8,%d0 183 lsl.l #3,%d0
|
D | fp_move.S | 213 lsl.w #1,%d0 214 lsl.l #7,%d0 215 lsl.l #8,%d0
|
D | fp_decode.h | 164 lsl.l %d1,%d0
|
D | fp_scan.S | 354 lsl.l %d1,%d0
|
/linux-4.4.14/arch/arm64/lib/ |
D | memcmp.S | 93 lsl limit, limit, #3 /* bytes-> bits. */ 96 CPU_LE( lsl mask, mask, limit ) 124 lsl tmp1, tmp1, #3/* Bytes beyond alignment -> bits.*/ 128 CPU_BE( lsl tmp2, tmp2, tmp1 )/*Big-endian.Early bytes are at MSB*/ 231 lsl data1, data1, pos 232 lsl data2, data2, pos
|
D | strncmp.S | 108 lsl limit, limit, #3 /* Bits -> bytes. */ 111 CPU_LE( lsl mask, mask, limit ) 133 neg tmp3, tmp1, lsl #3 /* 64 - bits(bytes beyond align). */ 138 CPU_BE( lsl tmp2, tmp2, tmp3 ) /* Shift (tmp1 & 63). */ 282 lsl data1, data1, pos 283 lsl data2, data2, pos
|
D | memset.S | 62 orr A_lw, A_lw, A_lw, lsl #8 63 orr A_lw, A_lw, A_lw, lsl #16 64 orr A_l, A_l, A_l, lsl #32 166 lsl zva_len, tmp3w, zva_len
|
D | strcmp.S | 96 lsl tmp1, tmp1, #3 /* Bytes beyond alignment -> bits. */ 102 CPU_BE( lsl tmp2, tmp2, tmp1 ) /* Shift (tmp1 & 63). */ 225 lsl data1, data1, pos 226 lsl data2, data2, pos
|
D | bitops.S | 35 lsl x3, x2, x3 // Create mask 53 lsl x4, x2, x3 // Create mask
|
D | strlen.S | 114 lsl tmp1, tmp1, #3 /* Bytes beyond alignment -> bits. */ 117 CPU_BE( lsl tmp2, tmp2, tmp1 ) /* Shift (tmp1 & 63). */
|
D | strnlen.S | 151 lsl tmp4, tmp4, #3 /* Bytes beyond alignment -> bits. */ 155 CPU_BE( lsl tmp2, tmp2, tmp4 ) /* Shift (tmp1 & 63). */
|
D | clear_page.S | 32 lsl x1, x2, x1
|
/linux-4.4.14/arch/arm/mach-tegra/ |
D | sleep.h | 65 movne \rd, \rd, lsl #3 74 movne \rd, \rd, lsl #3 109 moveq \tmp1, \tmp1, lsl #2 111 moveq \tmp2, \tmp2, lsl \tmp1
|
D | reset-handler.S | 157 mov r11, r11, lsl r10 @ R11 = CPU mask 237 mov r1, r0, lsl r10 247 movne r1, r10, lsl #3 269 mov r0, r11, lsl #8
|
D | sleep-tegra30.S | 184 ARM( orr r12, r12, r4, lsl r3 ) 185 THUMB( lsl r4, r4, r3 )
|
D | sleep-tegra20.S | 113 mov r1, r1, lsl r0
|
/linux-4.4.14/arch/arm/mach-shmobile/ |
D | headsmp.S | 58 ldr r8, [r5, r1, lsl #2] 62 ldr r9, [r6, r1, lsl #2] 74 ldr r0, [r7, r1, lsl #2]
|
D | headsmp-scu.S | 32 lsl r1, r1, #3 @ we will shift by cpu_id * 8 bits 35 lsl r3, r3, r1
|
/linux-4.4.14/arch/blackfin/mm/ |
D | sram-alloc.c | 742 struct sram_list_struct *lsl, **tmp; in sram_free_with_lsl() local 748 lsl = *tmp; in sram_free_with_lsl() 750 *tmp = lsl->next; in sram_free_with_lsl() 751 kfree(lsl); in sram_free_with_lsl() 766 struct sram_list_struct *lsl = NULL; in sram_alloc_with_lsl() local 769 lsl = kzalloc(sizeof(struct sram_list_struct), GFP_KERNEL); in sram_alloc_with_lsl() 770 if (!lsl) in sram_alloc_with_lsl() 786 kfree(lsl); in sram_alloc_with_lsl() 789 lsl->addr = addr; in sram_alloc_with_lsl() 790 lsl->length = size; in sram_alloc_with_lsl() [all …]
|
/linux-4.4.14/arch/arm64/mm/ |
D | proc-macros.S | 43 lsl \reg, \reg, \tmp // actual cache line size 53 lsl \reg, \reg, \tmp // actual cache line size
|
/linux-4.4.14/arch/arm64/include/asm/ |
D | fpsimdmacros.h | 85 add \state, \state, x\numnr, lsl #4 86 sub x\tmpnr1, x\tmpnr1, x\numnr, lsl #1 113 add \state, \state, x\tmpnr2, lsl #4 114 sub x\tmpnr1, x\tmpnr1, x\tmpnr2, lsl #1
|
D | assembler.h | 145 orr \rd, \lbits, \hbits, lsl #32
|
/linux-4.4.14/drivers/staging/lustre/lustre/lov/ |
D | lov_lock.c | 107 struct lovsub_lock *lsl; in lov_sublock_adopt() local 114 lsl = cl2sub_lock(sublock); in lov_sublock_adopt() 118 LASSERT(lov_lock_link_find(env, lck, lsl) == NULL); in lov_sublock_adopt() 121 lck->lls_sub[idx].sub_lock = lsl; in lov_sublock_adopt() 124 list_add_tail(&link->lll_list, &lsl->lss_parents); in lov_sublock_adopt() 132 rc = lov_sublock_modify(env, lck, lsl, &sublock->cll_descr, idx); in lov_sublock_adopt() 183 struct lovsub_lock *lsl, in lov_sublock_unlock() argument 188 lsl->lss_active = NULL; in lov_sublock_unlock() 1075 struct lovsub_lock *lsl = lls->sub_lock; in lov_lock_delete() local 1077 if (lsl == NULL) /* already removed */ in lov_lock_delete() [all …]
|
D | lovsub_lock.c | 58 struct lovsub_lock *lsl; in lovsub_lock_fini() local 60 lsl = cl2lovsub_lock(slice); in lovsub_lock_fini() 61 LASSERT(list_empty(&lsl->lss_parents)); in lovsub_lock_fini() 62 kmem_cache_free(lovsub_lock_kmem, lsl); in lovsub_lock_fini()
|
/linux-4.4.14/arch/arm64/kernel/ |
D | head.S | 262 str \tmp2, [\tbl, \tmp1, lsl #3] 294 orr \phys, \flags, \phys, lsl #SWAPPER_BLOCK_SHIFT // table entry 297 9999: str \phys, [\tbl, \start, lsl #3] // store the entry 502 CPU_BE( movk x0, #0x33d0, lsl #16 ) // Set EE and E0E on BE systems 503 CPU_LE( movk x0, #0x30d0, lsl #16 ) // Clear EE and E0E on LE systems
|
D | sleep.S | 95 add x1, x1, x8, lsl #3 170 ldr x0, [x0, x7, lsl #3]
|
D | entry.S | 682 ldr x16, [stbl, scno, lsl #3] // address in the syscall table 713 ldr x16, [stbl, scno, lsl #3] // address in the syscall table
|
/linux-4.4.14/arch/arm/kernel/ |
D | head.S | 238 1: orr r3, r7, r5, lsl #SECTION_SHIFT @ flags + kernel base 239 str r3, [r4, r5, lsl #PMD_ORDER] @ identity mapping 263 orr r3, r7, r3, lsl #SECTION_SHIFT 280 movs r0, r0, lsl #SECTION_SHIFT 304 mov r3, r3, lsl #PMD_ORDER 309 orr r3, r7, r3, lsl #SECTION_SHIFT
|
D | head-common.S | 214 orr r3, r3, r3, lsl #8 215 orr r3, r3, r3, lsl #16
|
D | entry-common.S | 228 ldrcc pc, [tbl, scno, lsl #2] @ call sys_* routine 268 ldrcc pc, [tbl, scno, lsl #2] @ call sys_* routine 325 ldrlo pc, [tbl, scno, lsl #2]
|
D | sleep.S | 79 add r3, r3, r0, lsl #2 154 ldr r0, [r0, r1, lsl #2]
|
D | entry-armv.S | 281 orr r0, r9, r0, lsl #16 1055 THUMB( ldr lr, [r0, lr, lsl #2] ) 1057 ARM( ldr lr, [pc, lr, lsl #2] )
|
D | head-nommu.S | 76 lsl r6, r6, #MPU_RSR_SZ @ Put size in right field
|
/linux-4.4.14/drivers/input/keyboard/ |
D | mpr121_touchkey.c | 130 unsigned char usl, lsl, tl, eleconf; in mpr121_phys_init() local 161 lsl = (usl * 65) / 100; in mpr121_phys_init() 164 ret |= i2c_smbus_write_byte_data(client, AUTO_CONFIG_LSL_ADDR, lsl); in mpr121_phys_init()
|
/linux-4.4.14/arch/m68k/68360/ |
D | entry.S | 50 lsl #2,%d1 81 lsl #2,%d0
|
/linux-4.4.14/arch/arm/mach-davinci/ |
D | sleep.S | 189 add r6, r6, r2, lsl #2 209 add r6, r6, r2, lsl #2
|
/linux-4.4.14/arch/m68k/68000/ |
D | entry.S | 54 lsl #2,%d1 85 lsl #2,%d0
|
/linux-4.4.14/arch/arm/mach-socfpga/ |
D | self-refresh.S | 123 mov r3, r3, lsl #16
|
/linux-4.4.14/arch/arm/include/debug/ |
D | sa1100.S | 47 lsl \rp, \rp, #8
|
D | omap2plus.S | 175 orr \rd, \rd, \rx, lsl #24 @ preserve LSR reg offset
|
/linux-4.4.14/arch/arm/common/ |
D | mcpm_head.S | 84 add r0, r0, r4, lsl #3 196 ldr r5, [r6, r4, lsl #2] @ r5 = CPU entry vector
|
/linux-4.4.14/arch/arm/mach-mvebu/ |
D | coherency_ll.S | 72 lsl r3, r2, r3
|
/linux-4.4.14/arch/arm/boot/bootp/ |
D | init.S | 56 addne r9, r9, r10, lsl #2
|
/linux-4.4.14/arch/arm/kvm/ |
D | interrupts.S | 456 lsl r2, r2, #4 457 orr r2, r2, r1, lsl #24
|
/linux-4.4.14/arch/arm64/kvm/ |
D | hyp.S | 239 add x22, x22, x5, lsl #2 260 add x22, x22, x5, lsl #2 329 add x22, x22, x5, lsl #2 350 add x22, x22, x5, lsl #2 1059 lsl x3, x3, #4 // and present it like HPFAR
|
/linux-4.4.14/arch/arm/mach-omap1/ |
D | ams-delta-fiq-handler.S | 117 orr r8, r11, r8, lsl r10 @ mask spurious interrupt 180 mov r10, r10, lsl #1 @ shift mask left
|
/linux-4.4.14/drivers/staging/lustre/lustre/llite/ |
D | file.c | 1940 struct lustre_swap_layouts *lsl) in ll_swap_layouts() argument 1973 if (lsl->sl_flags & SWAP_LAYOUTS_CHECK_DV1) in ll_swap_layouts() 1976 if (lsl->sl_flags & SWAP_LAYOUTS_CHECK_DV2) in ll_swap_layouts() 1980 llss->dv1 = lsl->sl_dv1; in ll_swap_layouts() 1981 llss->dv2 = lsl->sl_dv2; in ll_swap_layouts() 1996 gid = lsl->sl_gid; in ll_swap_layouts() 2011 if (lsl->sl_flags & in ll_swap_layouts() 2072 if (!(lsl->sl_flags & SWAP_LAYOUTS_KEEP_MTIME)) { in ll_swap_layouts() 2077 if (!(lsl->sl_flags & SWAP_LAYOUTS_KEEP_ATIME)) { in ll_swap_layouts() 2243 struct lustre_swap_layouts lsl; in ll_file_ioctl() local [all …]
|
/linux-4.4.14/arch/m68k/ifpsp060/src/ |
D | pfpsp.S | 4333 lsl.b &0x1,%d1 # should FP1 be moved? 4341 lsl.b &0x1,%d1 # should FP2 be moved? 4348 lsl.b &0x1,%d1 # should FP3 be moved? 4355 lsl.b &0x1,%d1 # should FP4 be moved? 4362 lsl.b &0x1,%d1 # should FP5 be moved? 4369 lsl.b &0x1,%d1 # should FP6 be moved? 4376 lsl.b &0x1,%d1 # should FP7 be moved? 4428 lsl.b &0x1,%d1 # should FP1 be moved? 4436 lsl.b &0x1,%d1 # should FP2 be moved? 4442 lsl.b &0x1,%d1 # should FP3 be moved? [all …]
|
D | itest.S | 411 lsl.w &0x8,%d1 479 lsl.w &0x8,%d1 545 lsl.l &0x8,%d1 547 lsl.l &0x8,%d1 549 lsl.l &0x8,%d1 652 lsl.w &0x8,%d1 716 lsl.w &0x8,%d1 749 lsl.w &0x8,%d1 815 lsl.l &0x8,%d1 817 lsl.l &0x8,%d1 [all …]
|
D | fpsp.S | 4087 lsl.w &0x3,%d1 # shift right 3 bits 8510 lsl.l %d6,%d4 8527 lsl.l %d6,%d4 8529 lsl.l %d6,%d5 10081 lsl.l %d0,%d3 13760 lsl.b &0x3,%d1 13964 lsl.b &0x3,%d1 14305 lsl.b &0x3,%d1 14652 lsl.b &0x3,%d1 15105 lsl.b &0x3,%d1 [all …]
|
D | ilsp.S | 334 lsl.l &0x1, %d7 # shift the divisor 335 lsl.l &0x1, %d6 # shift u4,u3 with overflow to u2
|
D | fplsp.S | 7133 lsl.l &4,%d1 7229 lsl.l &4,%d1 7297 lsl.l &4,%d1 9455 lsl.l %d6,%d4 9464 lsl.l %d6,%d4 9466 lsl.l %d6,%d5 9500 lsl.l %d6,%d1 9509 lsl.l %d6,%d1 9511 lsl.l %d6,%d2 9599 lsl.l %d6,%d1 [all …]
|
D | isp.S | 1360 lsl.l %d2,%d1 # shift index by scale
|
/linux-4.4.14/arch/arm/mach-omap2/ |
D | sram242x.S | 166 movne r5, r5, lsl #1 @ mult by 2 if to full 167 mov r5, r5, lsl #8 @ put rfr field back into place
|
D | sram243x.S | 166 movne r5, r5, lsl #1 @ mult by 2 if to full 167 mov r5, r5, lsl #8 @ put rfr field back into place
|
/linux-4.4.14/arch/arm/vfp/ |
D | vfphw.S | 97 ldr r4, [r3, r11, lsl #2] @ vfp_current_hw_state pointer
|
/linux-4.4.14/arch/avr32/kernel/ |
D | entry-avr32b.S | 494 lsl r1, (32 - PGDIR_SHIFT)
|
/linux-4.4.14/drivers/net/wireless/ |
D | atmel.c | 4350 mov r1, r1, lsl #3
|