rm 465 arch/arm/net/bpf_jit_32.c static inline void emit_udivmod(u8 rd, u8 rm, u8 rn, struct jit_ctx *ctx, u8 op) rm 472 arch/arm/net/bpf_jit_32.c emit(ARM_UDIV(rd, rm, rn), ctx); rm 474 arch/arm/net/bpf_jit_32.c emit(ARM_UDIV(ARM_IP, rm, rn), ctx); rm 475 arch/arm/net/bpf_jit_32.c emit(ARM_MLS(rd, rn, ARM_IP, rm), ctx); rm 493 arch/arm/net/bpf_jit_32.c if (rm != ARM_R0) { rm 495 arch/arm/net/bpf_jit_32.c emit(ARM_MOV_R(ARM_R0, rm), ctx); rm 510 arch/arm/net/bpf_jit_32.c if (rm != ARM_R0) rm 1063 arch/arm/net/bpf_jit_32.c s8 rm = src; rm 1068 arch/arm/net/bpf_jit_32.c rm = tmp[0]; rm 1070 arch/arm/net/bpf_jit_32.c } else if (rd[1] == rm) { rm 1071 arch/arm/net/bpf_jit_32.c emit(ARM_MOV_R(tmp[0], rm), ctx); rm 1072 arch/arm/net/bpf_jit_32.c rm = tmp[0]; rm 1077 arch/arm/net/bpf_jit_32.c emit(ARM_LDRB_I(rd[1], rm, off), ctx); rm 1083 arch/arm/net/bpf_jit_32.c emit(ARM_LDRH_I(rd[1], rm, off), ctx); rm 1089 arch/arm/net/bpf_jit_32.c emit(ARM_LDR_I(rd[1], rm, off), ctx); rm 1095 arch/arm/net/bpf_jit_32.c emit(ARM_LDR_I(rd[1], rm, off), ctx); rm 1096 arch/arm/net/bpf_jit_32.c emit(ARM_LDR_I(rd[0], rm, off + 4), ctx); rm 1103 arch/arm/net/bpf_jit_32.c static inline void emit_ar_r(const u8 rd, const u8 rt, const u8 rm, rm 1110 arch/arm/net/bpf_jit_32.c emit(ARM_AND_R(ARM_LR, rd, rm), ctx); rm 1123 arch/arm/net/bpf_jit_32.c emit(ARM_CMP_R(rd, rm), ctx); rm 1134 arch/arm/net/bpf_jit_32.c emit(ARM_SBCS_R(ARM_IP, rm, rd), ctx); rm 1140 arch/arm/net/bpf_jit_32.c emit(ARM_SBCS_R(ARM_IP, rd, rm), ctx); rm 1357 arch/arm/net/bpf_jit_32.c s8 rd_lo, rt, rm, rn; rm 1670 arch/arm/net/bpf_jit_32.c rm = arm_bpf_get_reg32(src_hi, tmp2[0], ctx); rm 1708 arch/arm/net/bpf_jit_32.c rm = tmp2[0]; rm 1717 arch/arm/net/bpf_jit_32.c emit_ar_r(rd[0], rd[1], rm, rn, ctx, BPF_OP(code), rm 159 arch/arm/net/bpf_jit_32.h #define _AL3_R(op, rd, rn, rm) ((op ## _R) | (rd) << 12 | (rn) << 16 | (rm)) rm 165 arch/arm/net/bpf_jit_32.h #define ARM_ADD_R(rd, rn, rm) _AL3_R(ARM_INST_ADD, rd, rn, rm) rm 166 arch/arm/net/bpf_jit_32.h #define ARM_ADDS_R(rd, rn, rm) _AL3_R(ARM_INST_ADDS, rd, rn, rm) rm 169 arch/arm/net/bpf_jit_32.h #define ARM_ADC_R(rd, rn, rm) _AL3_R(ARM_INST_ADC, rd, rn, rm) rm 172 arch/arm/net/bpf_jit_32.h #define ARM_AND_R(rd, rn, rm) _AL3_R(ARM_INST_AND, rd, rn, rm) rm 173 arch/arm/net/bpf_jit_32.h #define ARM_ANDS_R(rd, rn, rm) _AL3_R(ARM_INST_ANDS, rd, rn, rm) rm 176 arch/arm/net/bpf_jit_32.h #define ARM_BIC_R(rd, rn, rm) _AL3_R(ARM_INST_BIC, rd, rn, rm) rm 180 arch/arm/net/bpf_jit_32.h #define ARM_BX(rm) (ARM_INST_BX | (rm)) rm 181 arch/arm/net/bpf_jit_32.h #define ARM_BLX_R(rm) (ARM_INST_BLX_R | (rm)) rm 183 arch/arm/net/bpf_jit_32.h #define ARM_CMP_R(rn, rm) _AL3_R(ARM_INST_CMP, 0, rn, rm) rm 186 arch/arm/net/bpf_jit_32.h #define ARM_EOR_R(rd, rn, rm) _AL3_R(ARM_INST_EOR, rd, rn, rm) rm 189 arch/arm/net/bpf_jit_32.h #define ARM_LDR_R(rt, rn, rm) (ARM_INST_LDR_R | ARM_INST_LDST__U \ rm 191 arch/arm/net/bpf_jit_32.h | (rm)) rm 192 arch/arm/net/bpf_jit_32.h #define ARM_LDR_R_SI(rt, rn, rm, type, imm) \ rm 195 arch/arm/net/bpf_jit_32.h | (imm) << 7 | (type) << 5 | (rm)) rm 196 arch/arm/net/bpf_jit_32.h #define ARM_LDRB_R(rt, rn, rm) (ARM_INST_LDRB_R | ARM_INST_LDST__U \ rm 198 arch/arm/net/bpf_jit_32.h | (rm)) rm 199 arch/arm/net/bpf_jit_32.h #define ARM_LDRH_R(rt, rn, rm) (ARM_INST_LDRH_R | ARM_INST_LDST__U \ rm 201 arch/arm/net/bpf_jit_32.h | (rm)) rm 206 arch/arm/net/bpf_jit_32.h #define ARM_LSL_R(rd, rn, rm) (_AL3_R(ARM_INST_LSL, rd, 0, rn) | (rm) << 8) rm 209 arch/arm/net/bpf_jit_32.h #define ARM_LSR_R(rd, rn, rm) (_AL3_R(ARM_INST_LSR, rd, 0, rn) | (rm) << 8) rm 211 arch/arm/net/bpf_jit_32.h #define ARM_ASR_R(rd, rn, rm) (_AL3_R(ARM_INST_ASR, rd, 0, rn) | (rm) << 8) rm 214 arch/arm/net/bpf_jit_32.h #define ARM_MOV_R(rd, rm) _AL3_R(ARM_INST_MOV, rd, 0, rm) rm 215 arch/arm/net/bpf_jit_32.h #define ARM_MOVS_R(rd, rm) _AL3_R(ARM_INST_MOVS, rd, 0, rm) rm 217 arch/arm/net/bpf_jit_32.h #define ARM_MOV_SR(rd, rm, type, rs) \ rm 218 arch/arm/net/bpf_jit_32.h (_AL3_SR(ARM_MOV_R(rd, rm)) | (type) << 5 | (rs) << 8) rm 219 arch/arm/net/bpf_jit_32.h #define ARM_MOV_SI(rd, rm, type, imm6) \ rm 220 arch/arm/net/bpf_jit_32.h (ARM_MOV_R(rd, rm) | (type) << 5 | (imm6) << 7) rm 228 arch/arm/net/bpf_jit_32.h #define ARM_MUL(rd, rm, rn) (ARM_INST_MUL | (rd) << 16 | (rm) << 8 | (rn)) rm 233 arch/arm/net/bpf_jit_32.h #define ARM_ORR_R(rd, rn, rm) _AL3_R(ARM_INST_ORR, rd, rn, rm) rm 235 arch/arm/net/bpf_jit_32.h #define ARM_ORR_SR(rd, rn, rm, type, rs) \ rm 236 arch/arm/net/bpf_jit_32.h (_AL3_SR(ARM_ORR_R(rd, rn, rm)) | (type) << 5 | (rs) << 8) rm 237 arch/arm/net/bpf_jit_32.h #define ARM_ORRS_R(rd, rn, rm) _AL3_R(ARM_INST_ORRS, rd, rn, rm) rm 238 arch/arm/net/bpf_jit_32.h #define ARM_ORRS_SR(rd, rn, rm, type, rs) \ rm 239 arch/arm/net/bpf_jit_32.h (_AL3_SR(ARM_ORRS_R(rd, rn, rm)) | (type) << 5 | (rs) << 8) rm 240 arch/arm/net/bpf_jit_32.h #define ARM_ORR_SI(rd, rn, rm, type, imm6) \ rm 241 arch/arm/net/bpf_jit_32.h (ARM_ORR_R(rd, rn, rm) | (type) << 5 | (imm6) << 7) rm 242 arch/arm/net/bpf_jit_32.h #define ARM_ORRS_SI(rd, rn, rm, type, imm6) \ rm 243 arch/arm/net/bpf_jit_32.h (ARM_ORRS_R(rd, rn, rm) | (type) << 5 | (imm6) << 7) rm 245 arch/arm/net/bpf_jit_32.h #define ARM_REV(rd, rm) (ARM_INST_REV | (rd) << 12 | (rm)) rm 246 arch/arm/net/bpf_jit_32.h #define ARM_REV16(rd, rm) (ARM_INST_REV16 | (rd) << 12 | (rm)) rm 252 arch/arm/net/bpf_jit_32.h #define ARM_SUB_R(rd, rn, rm) _AL3_R(ARM_INST_SUB, rd, rn, rm) rm 253 arch/arm/net/bpf_jit_32.h #define ARM_SUBS_R(rd, rn, rm) _AL3_R(ARM_INST_SUBS, rd, rn, rm) rm 254 arch/arm/net/bpf_jit_32.h #define ARM_RSB_R(rd, rn, rm) _AL3_R(ARM_INST_RSB, rd, rn, rm) rm 255 arch/arm/net/bpf_jit_32.h #define ARM_SBC_R(rd, rn, rm) _AL3_R(ARM_INST_SBC, rd, rn, rm) rm 256 arch/arm/net/bpf_jit_32.h #define ARM_SBCS_R(rd, rn, rm) _AL3_R(ARM_INST_SBCS, rd, rn, rm) rm 261 arch/arm/net/bpf_jit_32.h #define ARM_TST_R(rn, rm) _AL3_R(ARM_INST_TST, 0, rn, rm) rm 264 arch/arm/net/bpf_jit_32.h #define ARM_UDIV(rd, rn, rm) (ARM_INST_UDIV | (rd) << 16 | (rn) | (rm) << 8) rm 266 arch/arm/net/bpf_jit_32.h #define ARM_UMULL(rd_lo, rd_hi, rn, rm) (ARM_INST_UMULL | (rd_hi) << 16 \ rm 267 arch/arm/net/bpf_jit_32.h | (rd_lo) << 12 | (rm) << 8 | rn) rm 269 arch/arm/net/bpf_jit_32.h #define ARM_MLS(rd, rn, rm, ra) (ARM_INST_MLS | (rd) << 16 | (rn) | (rm) << 8 \ rm 271 arch/arm/net/bpf_jit_32.h #define ARM_UXTH(rd, rm) (ARM_INST_UXTH | (rd) << 12 | (rm)) rm 79 arch/arm/probes/decode-arm.c int rm = insn & 0xf; rm 80 arch/arm/probes/decode-arm.c long rmv = regs->uregs[rm]; rm 75 arch/arm/probes/kprobes/actions-arm.c int rm = insn & 0xf; rm 81 arch/arm/probes/kprobes/actions-arm.c register unsigned long rmv asm("r3") = regs->uregs[rm]; rm 104 arch/arm/probes/kprobes/actions-arm.c int rm = insn & 0xf; rm 109 arch/arm/probes/kprobes/actions-arm.c register unsigned long rmv asm("r3") = regs->uregs[rm]; rm 135 arch/arm/probes/kprobes/actions-arm.c int rm = insn & 0xf; rm 141 arch/arm/probes/kprobes/actions-arm.c register unsigned long rmv asm("r3") = regs->uregs[rm]; rm 161 arch/arm/probes/kprobes/actions-arm.c int rm = insn & 0xf; rm 167 arch/arm/probes/kprobes/actions-arm.c register unsigned long rmv asm("r3") = (rm == 15) ? pc rm 168 arch/arm/probes/kprobes/actions-arm.c : regs->uregs[rm]; rm 195 arch/arm/probes/kprobes/actions-arm.c int rm = insn & 0xf; rm 199 arch/arm/probes/kprobes/actions-arm.c register unsigned long rmv asm("r3") = regs->uregs[rm]; rm 223 arch/arm/probes/kprobes/actions-arm.c int rm = insn & 0xf; rm 228 arch/arm/probes/kprobes/actions-arm.c register unsigned long rmv asm("r3") = regs->uregs[rm]; rm 251 arch/arm/probes/kprobes/actions-arm.c int rm = insn & 0xf; rm 254 arch/arm/probes/kprobes/actions-arm.c register unsigned long rmv asm("r3") = regs->uregs[rm]; rm 274 arch/arm/probes/kprobes/actions-arm.c int rm = (insn >> 8) & 0xf; rm 279 arch/arm/probes/kprobes/actions-arm.c register unsigned long rmv asm("r1") = regs->uregs[rm]; rm 29 arch/arm/probes/kprobes/actions-thumb.c int rm = insn & 0xf; rm 32 arch/arm/probes/kprobes/actions-thumb.c unsigned long rmv = regs->uregs[rm]; rm 190 arch/arm/probes/kprobes/actions-thumb.c int rm = insn & 0xf; rm 194 arch/arm/probes/kprobes/actions-thumb.c register unsigned long rmv asm("r3") = regs->uregs[rm]; rm 216 arch/arm/probes/kprobes/actions-thumb.c int rm = insn & 0xf; rm 220 arch/arm/probes/kprobes/actions-thumb.c register unsigned long rmv asm("r3") = regs->uregs[rm]; rm 285 arch/arm/probes/kprobes/actions-thumb.c int rm = insn & 0xf; rm 290 arch/arm/probes/kprobes/actions-thumb.c register unsigned long rmv asm("r3") = regs->uregs[rm]; rm 310 arch/arm/probes/kprobes/actions-thumb.c int rm = (insn >> 3) & 0xf; rm 311 arch/arm/probes/kprobes/actions-thumb.c unsigned long rmv = (rm == 15) ? pc : regs->uregs[rm]; rm 486 arch/arm/probes/kprobes/actions-thumb.c int rm = (insn >> 3) & 0xf; rm 493 arch/arm/probes/kprobes/actions-thumb.c rmv = (rm == 15) ? pc : regs->uregs[rm]; rm 1974 arch/mips/math-emu/cp1emu.c oldrm = ieee754_csr.rm; rm 1976 arch/mips/math-emu/cp1emu.c ieee754_csr.rm = MIPSInst_FUNC(ir); rm 1978 arch/mips/math-emu/cp1emu.c ieee754_csr.rm = oldrm; rm 2020 arch/mips/math-emu/cp1emu.c oldrm = ieee754_csr.rm; rm 2022 arch/mips/math-emu/cp1emu.c ieee754_csr.rm = MIPSInst_FUNC(ir); rm 2024 arch/mips/math-emu/cp1emu.c ieee754_csr.rm = oldrm; rm 2325 arch/mips/math-emu/cp1emu.c oldrm = ieee754_csr.rm; rm 2327 arch/mips/math-emu/cp1emu.c ieee754_csr.rm = MIPSInst_FUNC(ir); rm 2329 arch/mips/math-emu/cp1emu.c ieee754_csr.rm = oldrm; rm 2371 arch/mips/math-emu/cp1emu.c oldrm = ieee754_csr.rm; rm 2373 arch/mips/math-emu/cp1emu.c ieee754_csr.rm = MIPSInst_FUNC(ir); rm 2375 arch/mips/math-emu/cp1emu.c ieee754_csr.rm = oldrm; rm 83 arch/mips/math-emu/dp_add.c return ieee754dp_zero(ieee754_csr.rm == FPU_CSR_RD); rm 154 arch/mips/math-emu/dp_add.c return ieee754dp_zero(ieee754_csr.rm == FPU_CSR_RD); rm 14 arch/mips/math-emu/dp_div.c u64 rm; rm 116 arch/mips/math-emu/dp_div.c rm = 0; rm 122 arch/mips/math-emu/dp_div.c rm |= bm; rm 129 arch/mips/math-emu/dp_div.c rm <<= 1; rm 131 arch/mips/math-emu/dp_div.c rm |= 1; /* have remainder, set sticky */ rm 133 arch/mips/math-emu/dp_div.c assert(rm); rm 138 arch/mips/math-emu/dp_div.c while ((rm >> (DP_FBITS + 3)) == 0) { rm 139 arch/mips/math-emu/dp_div.c rm <<= 1; rm 143 arch/mips/math-emu/dp_div.c return ieee754dp_format(xs == ys ? 0 : 1, re, rm); rm 150 arch/mips/math-emu/dp_maddf.c return ieee754dp_zero(ieee754_csr.rm == FPU_CSR_RD); rm 291 arch/mips/math-emu/dp_maddf.c return ieee754dp_zero(ieee754_csr.rm == FPU_CSR_RD); rm 16 arch/mips/math-emu/dp_mul.c u64 rm; rm 144 arch/mips/math-emu/dp_mul.c rm = hrm | (lrm != 0); rm 149 arch/mips/math-emu/dp_mul.c if ((s64) rm < 0) { rm 150 arch/mips/math-emu/dp_mul.c rm = (rm >> (64 - (DP_FBITS + 1 + 3))) | rm 151 arch/mips/math-emu/dp_mul.c ((rm << (DP_FBITS + 1 + 3)) != 0); rm 154 arch/mips/math-emu/dp_mul.c rm = (rm >> (64 - (DP_FBITS + 1 + 3 + 1))) | rm 155 arch/mips/math-emu/dp_mul.c ((rm << (DP_FBITS + 1 + 3 + 1)) != 0); rm 157 arch/mips/math-emu/dp_mul.c assert(rm & (DP_HIDDEN_BIT << 3)); rm 159 arch/mips/math-emu/dp_mul.c return ieee754dp_format(rs, re, rm); rm 54 arch/mips/math-emu/dp_rint.c switch (ieee754_csr.rm) { rm 22 arch/mips/math-emu/dp_simple.c oldrm = ieee754_csr.rm; rm 23 arch/mips/math-emu/dp_simple.c ieee754_csr.rm = FPU_CSR_RD; rm 25 arch/mips/math-emu/dp_simple.c ieee754_csr.rm = oldrm; rm 40 arch/mips/math-emu/dp_simple.c oldrm = ieee754_csr.rm; rm 41 arch/mips/math-emu/dp_simple.c ieee754_csr.rm = FPU_CSR_RD; rm 46 arch/mips/math-emu/dp_simple.c ieee754_csr.rm = oldrm; rm 70 arch/mips/math-emu/dp_sqrt.c ieee754_csr.rm = FPU_CSR_RN; rm 114 arch/mips/math-emu/dp_sqrt.c ieee754_csr.rm = FPU_CSR_RZ; rm 130 arch/mips/math-emu/dp_sqrt.c switch (oldcsr.rm) { rm 83 arch/mips/math-emu/dp_sub.c return ieee754dp_zero(ieee754_csr.rm == FPU_CSR_RD); rm 158 arch/mips/math-emu/dp_sub.c if (ieee754_csr.rm == FPU_CSR_RD) rm 67 arch/mips/math-emu/dp_tint.c switch (ieee754_csr.rm) { rm 72 arch/mips/math-emu/dp_tlong.c switch (ieee754_csr.rm) { rm 151 arch/mips/math-emu/ieee754.h __BITFIELD_FIELD(unsigned rm:2, /* current rounding mode */ rm 158 arch/mips/math-emu/ieee754.h return (ieee754_csr.rm); rm 161 arch/mips/math-emu/ieee754.h static inline unsigned int ieee754_setrm(unsigned int rm) rm 163 arch/mips/math-emu/ieee754.h return (ieee754_csr.rm = rm); rm 61 arch/mips/math-emu/ieee754dp.c switch (ieee754_csr.rm) { rm 102 arch/mips/math-emu/ieee754dp.c switch(ieee754_csr.rm) { rm 166 arch/mips/math-emu/ieee754dp.c switch (ieee754_csr.rm) { rm 61 arch/mips/math-emu/ieee754sp.c switch (ieee754_csr.rm) { rm 102 arch/mips/math-emu/ieee754sp.c switch(ieee754_csr.rm) { rm 165 arch/mips/math-emu/ieee754sp.c switch (ieee754_csr.rm) { rm 83 arch/mips/math-emu/sp_add.c return ieee754sp_zero(ieee754_csr.rm == FPU_CSR_RD); rm 153 arch/mips/math-emu/sp_add.c return ieee754sp_zero(ieee754_csr.rm == FPU_CSR_RD); rm 14 arch/mips/math-emu/sp_div.c unsigned int rm; rm 116 arch/mips/math-emu/sp_div.c rm = 0; rm 122 arch/mips/math-emu/sp_div.c rm |= bm; rm 129 arch/mips/math-emu/sp_div.c rm <<= 1; rm 131 arch/mips/math-emu/sp_div.c rm |= 1; /* have remainder, set sticky */ rm 133 arch/mips/math-emu/sp_div.c assert(rm); rm 137 arch/mips/math-emu/sp_div.c while ((rm >> (SP_FBITS + 3)) == 0) { rm 138 arch/mips/math-emu/sp_div.c rm <<= 1; rm 142 arch/mips/math-emu/sp_div.c return ieee754sp_format(xs == ys ? 0 : 1, re, rm); rm 22 arch/mips/math-emu/sp_fdp.c u32 rm; rm 58 arch/mips/math-emu/sp_fdp.c if ((ieee754_csr.rm == FPU_CSR_RU && !xs) || rm 59 arch/mips/math-emu/sp_fdp.c (ieee754_csr.rm == FPU_CSR_RD && xs)) rm 70 arch/mips/math-emu/sp_fdp.c rm = (xm >> (DP_FBITS - (SP_FBITS + 3))) | rm 73 arch/mips/math-emu/sp_fdp.c return ieee754sp_format(xs, xe, rm); rm 20 arch/mips/math-emu/sp_maddf.c unsigned int rm; rm 119 arch/mips/math-emu/sp_maddf.c return ieee754sp_zero(ieee754_csr.rm == FPU_CSR_RD); rm 185 arch/mips/math-emu/sp_maddf.c rm = XSPSRS64(rm64, (62 - 26)); rm 186 arch/mips/math-emu/sp_maddf.c return ieee754sp_format(rs, re, rm); rm 231 arch/mips/math-emu/sp_maddf.c return ieee754sp_zero(ieee754_csr.rm == FPU_CSR_RD); rm 16 arch/mips/math-emu/sp_mul.c unsigned int rm; rm 139 arch/mips/math-emu/sp_mul.c rm = hrm | (lrm != 0); rm 144 arch/mips/math-emu/sp_mul.c if ((int) rm < 0) { rm 145 arch/mips/math-emu/sp_mul.c rm = (rm >> (32 - (SP_FBITS + 1 + 3))) | rm 146 arch/mips/math-emu/sp_mul.c ((rm << (SP_FBITS + 1 + 3)) != 0); rm 149 arch/mips/math-emu/sp_mul.c rm = (rm >> (32 - (SP_FBITS + 1 + 3 + 1))) | rm 150 arch/mips/math-emu/sp_mul.c ((rm << (SP_FBITS + 1 + 3 + 1)) != 0); rm 152 arch/mips/math-emu/sp_mul.c assert(rm & (SP_HIDDEN_BIT << 3)); rm 154 arch/mips/math-emu/sp_mul.c return ieee754sp_format(rs, re, rm); rm 55 arch/mips/math-emu/sp_rint.c switch (ieee754_csr.rm) { rm 22 arch/mips/math-emu/sp_simple.c oldrm = ieee754_csr.rm; rm 23 arch/mips/math-emu/sp_simple.c ieee754_csr.rm = FPU_CSR_RD; rm 25 arch/mips/math-emu/sp_simple.c ieee754_csr.rm = oldrm; rm 40 arch/mips/math-emu/sp_simple.c oldrm = ieee754_csr.rm; rm 41 arch/mips/math-emu/sp_simple.c ieee754_csr.rm = FPU_CSR_RD; rm 46 arch/mips/math-emu/sp_simple.c ieee754_csr.rm = oldrm; rm 90 arch/mips/math-emu/sp_sqrt.c switch (ieee754_csr.rm) { rm 83 arch/mips/math-emu/sp_sub.c return ieee754sp_zero(ieee754_csr.rm == FPU_CSR_RD); rm 155 arch/mips/math-emu/sp_sub.c if (ieee754_csr.rm == FPU_CSR_RD) rm 72 arch/mips/math-emu/sp_tint.c switch (ieee754_csr.rm) { rm 68 arch/mips/math-emu/sp_tlong.c switch (ieee754_csr.rm) { rm 1902 arch/powerpc/platforms/powernv/pci-ioda.c unsigned long index, unsigned long npages, bool rm) rm 1909 arch/powerpc/platforms/powernv/pci-ioda.c __be64 __iomem *invalidate = pnv_ioda_get_inval_reg(pe->phb, rm); rm 1924 arch/powerpc/platforms/powernv/pci-ioda.c if (rm) rm 1985 arch/powerpc/platforms/powernv/pci-ioda.c static void pnv_pci_phb3_tce_invalidate_entire(struct pnv_phb *phb, bool rm) rm 1987 arch/powerpc/platforms/powernv/pci-ioda.c __be64 __iomem *invalidate = pnv_ioda_get_inval_reg(phb, rm); rm 1991 arch/powerpc/platforms/powernv/pci-ioda.c if (rm) rm 2007 arch/powerpc/platforms/powernv/pci-ioda.c static void pnv_pci_phb3_tce_invalidate(struct pnv_ioda_pe *pe, bool rm, rm 2011 arch/powerpc/platforms/powernv/pci-ioda.c __be64 __iomem *invalidate = pnv_ioda_get_inval_reg(pe->phb, rm); rm 2026 arch/powerpc/platforms/powernv/pci-ioda.c if (rm) rm 2046 arch/powerpc/platforms/powernv/pci-ioda.c unsigned long index, unsigned long npages, bool rm) rm 2067 arch/powerpc/platforms/powernv/pci-ioda.c pnv_pci_phb3_tce_invalidate_entire(phb, rm); rm 2071 arch/powerpc/platforms/powernv/pci-ioda.c pnv_pci_phb3_tce_invalidate(pe, rm, shift, rm 2081 arch/powerpc/platforms/powernv/pci-ioda.c void pnv_pci_ioda2_tce_invalidate_entire(struct pnv_phb *phb, bool rm) rm 2084 arch/powerpc/platforms/powernv/pci-ioda.c pnv_pci_phb3_tce_invalidate_entire(phb, rm); rm 214 arch/powerpc/platforms/powernv/pci.h extern void pnv_pci_ioda2_tce_invalidate_entire(struct pnv_phb *phb, bool rm); rm 101 arch/powerpc/platforms/ps3/mm.c struct mem_region rm; rm 110 arch/powerpc/platforms/ps3/mm.c DBG("%s:%d: map.rm.size = %llxh\n", func, line, m->rm.size); rm 128 arch/powerpc/platforms/ps3/mm.c return (phys_addr < map.rm.size || phys_addr >= map.total) rm 230 arch/powerpc/platforms/ps3/mm.c r->offset = r->base - map.rm.size; rm 282 arch/powerpc/platforms/ps3/mm.c if (result || r->base < map.rm.size) { rm 289 arch/powerpc/platforms/ps3/mm.c r->offset = r->base - map.rm.size; rm 318 arch/powerpc/platforms/ps3/mm.c map.total = map.rm.size; rm 336 arch/powerpc/platforms/ps3/mm.c if (lpar_addr >= map.rm.size) rm 984 arch/powerpc/platforms/ps3/mm.c if (r->offset < map.rm.size) { rm 986 arch/powerpc/platforms/ps3/mm.c virt_addr = map.rm.base + r->offset; rm 987 arch/powerpc/platforms/ps3/mm.c len = map.rm.size - r->offset; rm 996 arch/powerpc/platforms/ps3/mm.c if (r->offset + r->len > map.rm.size) { rm 998 arch/powerpc/platforms/ps3/mm.c virt_addr = map.rm.size; rm 1000 arch/powerpc/platforms/ps3/mm.c if (r->offset >= map.rm.size) rm 1001 arch/powerpc/platforms/ps3/mm.c virt_addr += r->offset - map.rm.size; rm 1003 arch/powerpc/platforms/ps3/mm.c len -= map.rm.size - r->offset; rm 1026 arch/powerpc/platforms/ps3/mm.c if (r->offset < map.rm.size) { rm 1028 arch/powerpc/platforms/ps3/mm.c lpar_addr = map.rm.base + r->offset; rm 1029 arch/powerpc/platforms/ps3/mm.c len = map.rm.size - r->offset; rm 1037 arch/powerpc/platforms/ps3/mm.c if (r->offset + r->len > map.rm.size) { rm 1041 arch/powerpc/platforms/ps3/mm.c if (r->offset >= map.rm.size) rm 1042 arch/powerpc/platforms/ps3/mm.c lpar_addr += r->offset - map.rm.size; rm 1044 arch/powerpc/platforms/ps3/mm.c len -= map.rm.size - r->offset; rm 1126 arch/powerpc/platforms/ps3/mm.c if (r->offset >= map.rm.size) rm 1192 arch/powerpc/platforms/ps3/mm.c result = ps3_repository_read_mm_info(&map.rm.base, &map.rm.size, rm 1198 arch/powerpc/platforms/ps3/mm.c map.rm.offset = map.rm.base; rm 1203 arch/powerpc/platforms/ps3/mm.c BUG_ON(map.rm.base); rm 1204 arch/powerpc/platforms/ps3/mm.c BUG_ON(!map.rm.size); rm 1209 arch/powerpc/platforms/ps3/mm.c result = ps3_mm_region_create(&map.r1, map.total - map.rm.size); rm 1216 arch/powerpc/platforms/ps3/mm.c map.total = map.rm.size + map.r1.size; rm 1222 arch/powerpc/platforms/ps3/mm.c __func__, __LINE__, map.rm.size, rm 1223 arch/powerpc/platforms/ps3/mm.c map.total - map.rm.size); rm 1224 arch/powerpc/platforms/ps3/mm.c memblock_add(map.rm.size, map.total - map.rm.size); rm 302 arch/sh/kernel/disassemble.c int rm = 0; rm 364 arch/sh/kernel/disassemble.c rm = nibs[n]; rm 368 arch/sh/kernel/disassemble.c rm = (nibs[n] & 0x3); rm 408 arch/sh/kernel/disassemble.c printk("r%d", rm); rm 411 arch/sh/kernel/disassemble.c printk("@r%d+", rm); rm 414 arch/sh/kernel/disassemble.c printk("@-r%d", rm); rm 417 arch/sh/kernel/disassemble.c printk("@r%d", rm); rm 420 arch/sh/kernel/disassemble.c printk("@(%d,r%d)", imm, rm); rm 435 arch/sh/kernel/disassemble.c printk("@(r0,r%d)", rm); rm 482 arch/sh/kernel/disassemble.c printk("fr%d", rm); rm 494 arch/sh/kernel/disassemble.c if (rm & 1) { rm 495 arch/sh/kernel/disassemble.c printk("xd%d", rm & ~1); rm 500 arch/sh/kernel/disassemble.c printk("dr%d", rm); rm 517 arch/sh/kernel/disassemble.c printk("fv%d", rm*4); rm 89 arch/sh/kernel/traps_32.c unsigned long *rm, *rn; rm 97 arch/sh/kernel/traps_32.c rm = ®s->regs[index]; rm 113 arch/sh/kernel/traps_32.c srcu = (unsigned char __user *)*rm; rm 127 arch/sh/kernel/traps_32.c src = (unsigned char *)rm; rm 141 arch/sh/kernel/traps_32.c src = (unsigned char*) rm; rm 153 arch/sh/kernel/traps_32.c src = (unsigned char*) rm; rm 164 arch/sh/kernel/traps_32.c srcu = (unsigned char __user *)*rm; rm 175 arch/sh/kernel/traps_32.c srcu = (unsigned char __user *)*rm; rm 177 arch/sh/kernel/traps_32.c *rm += count; rm 197 arch/sh/kernel/traps_32.c dstu = (unsigned char __user *)*rm; /* called Rn in the spec */ rm 206 arch/sh/kernel/traps_32.c srcu = (unsigned char __user *)*rm; rm 306 arch/sh/kernel/traps_32.c u_int rm; rm 316 arch/sh/kernel/traps_32.c rm = regs->regs[index]; rm 344 arch/sh/kernel/traps_32.c regs->pc += rm + 4; rm 351 arch/sh/kernel/traps_32.c regs->pc += rm + 4; rm 371 arch/sh/kernel/traps_32.c regs->pc = rm; rm 378 arch/sh/kernel/traps_32.c regs->pc = rm; rm 752 arch/x86/events/intel/rapl.c struct rapl_model *rm; rm 759 arch/x86/events/intel/rapl.c rm = (struct rapl_model *) id->driver_data; rm 761 arch/x86/events/intel/rapl.c false, (void *) &rm->events); rm 763 arch/x86/events/intel/rapl.c ret = rapl_check_hw_unit(rm->apply_quirk); rm 93 arch/x86/include/asm/fpu/types.h u8 rm; rm 346 arch/x86/lib/insn.c insn_byte_t mod, rm, base; rm 371 arch/x86/lib/insn.c rm = X86_MODRM_RM(insn->modrm.value); rm 379 arch/x86/lib/insn.c if ((mod == 0 && rm == 6) || mod == 2) { rm 385 arch/x86/lib/insn.c if ((mod == 0 && rm == 5) || mod == 2 || rm 111 arch/x86/math-emu/fpu_proto.h extern int FPU_sub(int flags, int rm, int control_w); rm 92 arch/x86/math-emu/fpu_system.h #define FPU_rm (*(u_char *)&(I387->soft.rm)) rm 214 arch/x86/math-emu/get_address.c unsigned rm = FPU_modrm & 7; rm 229 arch/x86/math-emu/get_address.c if (rm == 4 && mod != 3) { rm 232 arch/x86/math-emu/get_address.c cpu_reg_ptr = ®_(rm); rm 235 arch/x86/math-emu/get_address.c if (rm == 5) { rm 299 arch/x86/math-emu/get_address.c unsigned rm = FPU_modrm & 7; rm 315 arch/x86/math-emu/get_address.c if (rm == 6) { rm 347 arch/x86/math-emu/get_address.c switch (rm) { rm 129 arch/x86/math-emu/reg_add_sub.c int FPU_sub(int flags, int rm, int control_w) rm 141 arch/x86/math-emu/reg_add_sub.c b = (FPU_REG *) rm; rm 144 arch/x86/math-emu/reg_add_sub.c b = &st(rm); rm 145 arch/x86/math-emu/reg_add_sub.c tagb = FPU_gettagi(rm); rm 148 arch/x86/math-emu/reg_add_sub.c deststnr = rm; rm 28 arch/x86/math-emu/reg_divide.c int FPU_div(int flags, int rm, int control_w) rm 37 arch/x86/math-emu/reg_divide.c deststnr = rm; rm 46 arch/x86/math-emu/reg_divide.c a = (FPU_REG *) rm; rm 49 arch/x86/math-emu/reg_divide.c a = &st(rm); rm 51 arch/x86/math-emu/reg_divide.c taga = FPU_gettagi(rm); rm 58 arch/x86/math-emu/reg_divide.c b = (FPU_REG *) rm; rm 61 arch/x86/math-emu/reg_divide.c b = &st(rm); rm 63 arch/x86/math-emu/reg_divide.c tagb = FPU_gettagi(rm); rm 133 arch/x86/math-emu/reg_divide.c return real_2op_NaN((FPU_REG *) rm, flags & 0x0f, 0, rm 141 arch/x86/math-emu/reg_divide.c return real_2op_NaN(st0_ptr, tag, rm, rm 142 arch/x86/math-emu/reg_divide.c (flags & REV) ? st0_ptr : &st(rm)); rm 145 arch/x86/math-emu/reg_divide.c tag = FPU_gettagi(rm); rm 147 arch/x86/math-emu/reg_divide.c tag = FPU_Special(&st(rm)); rm 148 arch/x86/math-emu/reg_divide.c return real_2op_NaN(&st(rm), tag, 0, rm 149 arch/x86/math-emu/reg_divide.c (flags & REV) ? st0_ptr : &st(rm)); rm 335 drivers/gpu/drm/armada/armada_crtc.c uint32_t lm, rm, tm, bm, val, sclk; rm 341 drivers/gpu/drm/armada/armada_crtc.c rm = adj->crtc_hsync_start - adj->crtc_hdisplay; rm 348 drivers/gpu/drm/armada/armada_crtc.c DRM_DEBUG_KMS("lm %d rm %d tm %d bm %d\n", lm, rm, tm, bm); rm 379 drivers/gpu/drm/armada/armada_crtc.c armada_reg_queue_set(regs, i, (lm << 16) | rm, LCD_SPU_H_PORCH); rm 1810 drivers/gpu/drm/i915/gvt/handlers.c #define MMIO_F(reg, s, f, am, rm, d, r, w) do { \ rm 1812 drivers/gpu/drm/i915/gvt/handlers.c f, s, am, rm, d, r, w); \ rm 1832 drivers/gpu/drm/i915/gvt/handlers.c #define MMIO_RO(reg, d, f, rm, r, w) \ rm 1833 drivers/gpu/drm/i915/gvt/handlers.c MMIO_F(reg, 4, F_RO | f, 0, rm, d, r, w) rm 1835 drivers/gpu/drm/i915/gvt/handlers.c #define MMIO_RING_F(prefix, s, f, am, rm, d, r, w) do { \ rm 1836 drivers/gpu/drm/i915/gvt/handlers.c MMIO_F(prefix(RENDER_RING_BASE), s, f, am, rm, d, r, w); \ rm 1837 drivers/gpu/drm/i915/gvt/handlers.c MMIO_F(prefix(BLT_RING_BASE), s, f, am, rm, d, r, w); \ rm 1838 drivers/gpu/drm/i915/gvt/handlers.c MMIO_F(prefix(GEN6_BSD_RING_BASE), s, f, am, rm, d, r, w); \ rm 1839 drivers/gpu/drm/i915/gvt/handlers.c MMIO_F(prefix(VEBOX_RING_BASE), s, f, am, rm, d, r, w); \ rm 1841 drivers/gpu/drm/i915/gvt/handlers.c MMIO_F(prefix(GEN8_BSD2_RING_BASE), s, f, am, rm, d, r, w); \ rm 1856 drivers/gpu/drm/i915/gvt/handlers.c #define MMIO_RING_RO(prefix, d, f, rm, r, w) \ rm 1857 drivers/gpu/drm/i915/gvt/handlers.c MMIO_RING_F(prefix, 4, F_RO | f, 0, rm, d, r, w) rm 610 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c ret = dpu_rm_reserve(&dpu_kms->rm, drm_enc, crtc_state, rm 995 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c ret = dpu_rm_reserve(&dpu_kms->rm, drm_enc, drm_crtc->state, rm 1006 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c if (!dpu_rm_get_hw(&dpu_kms->rm, &hw_iter)) rm 1013 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c if (!dpu_rm_get_hw(&dpu_kms->rm, &hw_iter)) rm 1021 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c if (!dpu_rm_get_hw(&dpu_kms->rm, &hw_iter)) rm 1062 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c if (!dpu_rm_get_hw(&dpu_kms->rm, &hw_iter)) rm 1086 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c dpu_rm_release(&dpu_kms->rm, drm_enc); rm 1241 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c dpu_rm_release(&dpu_kms->rm, drm_enc); rm 608 drivers/gpu/drm/msm/disp/dpu1/dpu_kms.c dpu_rm_destroy(&dpu_kms->rm); rm 880 drivers/gpu/drm/msm/disp/dpu1/dpu_kms.c rc = dpu_rm_init(&dpu_kms->rm, dpu_kms->catalog, dpu_kms->mmio); rm 114 drivers/gpu/drm/msm/disp/dpu1/dpu_kms.h struct dpu_rm rm; rm 53 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c static bool _dpu_rm_get_hw_locked(struct dpu_rm *rm, struct dpu_rm_hw_iter *i) rm 57 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c if (!rm || !i || i->type >= DPU_HW_BLK_MAX) { rm 63 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c blk_list = &rm->hw_blks[i->type]; rm 86 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c bool dpu_rm_get_hw(struct dpu_rm *rm, struct dpu_rm_hw_iter *i) rm 90 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c mutex_lock(&rm->rm_lock); rm 91 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c ret = _dpu_rm_get_hw_locked(rm, i); rm 92 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c mutex_unlock(&rm->rm_lock); rm 123 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c int dpu_rm_destroy(struct dpu_rm *rm) rm 129 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c list_for_each_entry_safe(hw_cur, hw_nxt, &rm->hw_blks[type], rm 137 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c mutex_destroy(&rm->rm_lock); rm 143 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c struct dpu_rm *rm, rm 191 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c list_add_tail(&blk->list, &rm->hw_blks[type]); rm 196 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c int dpu_rm_init(struct dpu_rm *rm, rm 203 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c if (!rm || !cat || !mmio) { rm 209 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c memset(rm, 0, sizeof(*rm)); rm 211 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c mutex_init(&rm->rm_lock); rm 214 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c INIT_LIST_HEAD(&rm->hw_blks[type]); rm 225 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c rc = _dpu_rm_hw_blk_create(rm, cat, mmio, DPU_HW_BLK_LM, rm 232 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c if (!rm->lm_max_width) { rm 233 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c rm->lm_max_width = lm->sblk->maxwidth; rm 234 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c } else if (rm->lm_max_width != lm->sblk->maxwidth) { rm 240 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c if (rm->lm_max_width > lm->sblk->maxwidth) rm 241 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c rm->lm_max_width = lm->sblk->maxwidth; rm 246 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c rc = _dpu_rm_hw_blk_create(rm, cat, mmio, DPU_HW_BLK_PINGPONG, rm 260 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c rc = _dpu_rm_hw_blk_create(rm, cat, mmio, DPU_HW_BLK_INTF, rm 269 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c rc = _dpu_rm_hw_blk_create(rm, cat, mmio, DPU_HW_BLK_CTL, rm 280 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c dpu_rm_destroy(rm); rm 306 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c struct dpu_rm *rm, rm 340 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c while (_dpu_rm_get_hw_locked(rm, &iter)) { rm 361 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c static int _dpu_rm_reserve_lms(struct dpu_rm *rm, uint32_t enc_id, rm 379 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c _dpu_rm_get_hw_locked(rm, &iter_i)) { rm 387 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c rm, enc_id, reqs, lm[lm_count], rm 397 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c _dpu_rm_get_hw_locked(rm, &iter_j)) { rm 402 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c rm, enc_id, reqs, iter_j.blk, rm 430 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c struct dpu_rm *rm, rm 447 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c while (_dpu_rm_get_hw_locked(rm, &iter)) { rm 481 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c struct dpu_rm *rm, rm 491 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c while (_dpu_rm_get_hw_locked(rm, &iter)) { rm 515 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c struct dpu_rm *rm, rm 526 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c ret = _dpu_rm_reserve_intf(rm, enc_id, id, rm 536 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c struct dpu_rm *rm, rm 543 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c ret = _dpu_rm_reserve_lms(rm, enc->base.id, reqs); rm 549 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c ret = _dpu_rm_reserve_ctls(rm, enc->base.id, &reqs->topology); rm 555 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c ret = _dpu_rm_reserve_intf_related_hw(rm, enc->base.id, &reqs->hw_res); rm 563 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c struct dpu_rm *rm, rm 580 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c static void _dpu_rm_release_reservation(struct dpu_rm *rm, uint32_t enc_id) rm 586 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c list_for_each_entry(blk, &rm->hw_blks[type], list) { rm 596 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c void dpu_rm_release(struct dpu_rm *rm, struct drm_encoder *enc) rm 598 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c mutex_lock(&rm->rm_lock); rm 600 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c _dpu_rm_release_reservation(rm, enc->base.id); rm 602 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c mutex_unlock(&rm->rm_lock); rm 606 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c struct dpu_rm *rm, rm 622 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c mutex_lock(&rm->rm_lock); rm 624 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c ret = _dpu_rm_populate_requirements(rm, enc, crtc_state, &reqs, rm 631 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c ret = _dpu_rm_make_reservation(rm, enc, crtc_state, &reqs); rm 634 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c _dpu_rm_release_reservation(rm, enc->base.id); rm 639 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c _dpu_rm_release_reservation(rm, enc->base.id); rm 643 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c mutex_unlock(&rm->rm_lock); rm 55 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.h int dpu_rm_init(struct dpu_rm *rm, rm 64 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.h int dpu_rm_destroy(struct dpu_rm *rm); rm 80 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.h int dpu_rm_reserve(struct dpu_rm *rm, rm 93 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.h void dpu_rm_release(struct dpu_rm *rm, struct drm_encoder *enc); rm 119 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.h bool dpu_rm_get_hw(struct dpu_rm *rm, struct dpu_rm_hw_iter *iter); rm 100 drivers/gpu/drm/nouveau/nvkm/engine/fifo/dmanv04.c u32 rm = ((1ULL << c->bits) - 1) << c->regs; rm 102 drivers/gpu/drm/nouveau/nvkm/engine/fifo/dmanv04.c u32 rv = (nvkm_rd32(device, c->regp) & rm) >> c->regs; rm 799 drivers/hwmon/asc7621.c #define PREAD(name, n, pri, rm, rl, m, s, r) \ rm 801 drivers/hwmon/asc7621.c .priority = pri, .msb[0] = rm, .lsb[0] = rl, .mask[0] = m, \ rm 804 drivers/hwmon/asc7621.c #define PWRITE(name, n, pri, rm, rl, m, s, r) \ rm 806 drivers/hwmon/asc7621.c .priority = pri, .msb[0] = rm, .lsb[0] = rl, .mask[0] = m, \ rm 813 drivers/hwmon/asc7621.c #define PWRITEM(name, n, pri, rm, rl, m, s, r) \ rm 815 drivers/hwmon/asc7621.c .priority = pri, .msb = rm, .lsb = rl, .mask = m, .shift = s,} rm 912 drivers/isdn/hardware/mISDN/hfcmulti.c hfcmulti_resync(struct hfc_multi *locked, struct hfc_multi *newmaster, int rm) rm 1002 drivers/isdn/hardware/mISDN/hfcmulti.c if (!rm) rm 1014 drivers/isdn/hardware/mISDN/hfcmulti.c plxsd_checksync(struct hfc_multi *hc, int rm) rm 1022 drivers/isdn/hardware/mISDN/hfcmulti.c hfcmulti_resync(hc, hc, rm); rm 1030 drivers/isdn/hardware/mISDN/hfcmulti.c hfcmulti_resync(hc, NULL, rm); rm 52 drivers/macintosh/rack-meter.c struct rackmeter *rm; rm 95 drivers/macintosh/rack-meter.c static void rackmeter_setup_i2s(struct rackmeter *rm) rm 97 drivers/macintosh/rack-meter.c struct macio_chip *macio = rm->mdev->bus->chip; rm 106 drivers/macintosh/rack-meter.c pmac_call_feature(PMAC_FTR_SOUND_CHIP_ENABLE, rm->i2s, 0, 1); rm 121 drivers/macintosh/rack-meter.c out_le32(rm->i2s_regs + 0x10, 0x01fa0000); rm 122 drivers/macintosh/rack-meter.c (void)in_le32(rm->i2s_regs + 0x10); rm 132 drivers/macintosh/rack-meter.c static void rackmeter_set_default_pattern(struct rackmeter *rm) rm 138 drivers/macintosh/rack-meter.c rm->ubuf[i] = (i & 1) * 255; rm 140 drivers/macintosh/rack-meter.c rm->ubuf[i] = ((~i) & 1) * 255; rm 144 drivers/macintosh/rack-meter.c static void rackmeter_do_pause(struct rackmeter *rm, int pause) rm 146 drivers/macintosh/rack-meter.c struct rackmeter_dma *rdma = rm->dma_buf_v; rm 150 drivers/macintosh/rack-meter.c rm->paused = pause; rm 152 drivers/macintosh/rack-meter.c DBDMA_DO_STOP(rm->dma_regs); rm 158 drivers/macintosh/rack-meter.c rm->dma_buf_v->mark = 0; rm 161 drivers/macintosh/rack-meter.c out_le32(&rm->dma_regs->cmdptr_hi, 0); rm 162 drivers/macintosh/rack-meter.c out_le32(&rm->dma_regs->cmdptr, rm->dma_buf_p); rm 163 drivers/macintosh/rack-meter.c out_le32(&rm->dma_regs->control, (RUN << 16) | RUN); rm 166 drivers/macintosh/rack-meter.c static void rackmeter_setup_dbdma(struct rackmeter *rm) rm 168 drivers/macintosh/rack-meter.c struct rackmeter_dma *db = rm->dma_buf_v; rm 172 drivers/macintosh/rack-meter.c DBDMA_DO_RESET(rm->dma_regs); rm 185 drivers/macintosh/rack-meter.c cmd->phy_addr = cpu_to_le32(rm->dma_buf_p + rm 192 drivers/macintosh/rack-meter.c cmd->phy_addr = cpu_to_le32(rm->dma_buf_p + rm 198 drivers/macintosh/rack-meter.c cmd->phy_addr = cpu_to_le32(rm->dma_buf_p + rm 205 drivers/macintosh/rack-meter.c cmd->phy_addr = cpu_to_le32(rm->dma_buf_p + rm 207 drivers/macintosh/rack-meter.c cmd->cmd_dep = cpu_to_le32(rm->dma_buf_p); rm 209 drivers/macintosh/rack-meter.c rackmeter_do_pause(rm, 0); rm 216 drivers/macintosh/rack-meter.c struct rackmeter *rm = rcpu->rm; rm 240 drivers/macintosh/rack-meter.c rm->ubuf[i + offset] = ub; rm 246 drivers/macintosh/rack-meter.c pause = (rm->cpu[0].zero && rm->cpu[1].zero); rm 247 drivers/macintosh/rack-meter.c if (pause != rm->paused) { rm 248 drivers/macintosh/rack-meter.c mutex_lock(&rm->sem); rm 249 drivers/macintosh/rack-meter.c pause = (rm->cpu[0].zero && rm->cpu[1].zero); rm 250 drivers/macintosh/rack-meter.c rackmeter_do_pause(rm, pause); rm 251 drivers/macintosh/rack-meter.c mutex_unlock(&rm->sem); rm 257 drivers/macintosh/rack-meter.c static void rackmeter_init_cpu_sniffer(struct rackmeter *rm) rm 267 drivers/macintosh/rack-meter.c rm->cpu[0].rm = rm; rm 268 drivers/macintosh/rack-meter.c INIT_DELAYED_WORK(&rm->cpu[0].sniffer, rackmeter_do_timer); rm 269 drivers/macintosh/rack-meter.c rm->cpu[1].rm = rm; rm 270 drivers/macintosh/rack-meter.c INIT_DELAYED_WORK(&rm->cpu[1].sniffer, rackmeter_do_timer); rm 277 drivers/macintosh/rack-meter.c rcpu = &rm->cpu[cpu]; rm 280 drivers/macintosh/rack-meter.c schedule_delayed_work_on(cpu, &rm->cpu[cpu].sniffer, rm 285 drivers/macintosh/rack-meter.c static void rackmeter_stop_cpu_sniffer(struct rackmeter *rm) rm 287 drivers/macintosh/rack-meter.c cancel_delayed_work_sync(&rm->cpu[0].sniffer); rm 288 drivers/macintosh/rack-meter.c cancel_delayed_work_sync(&rm->cpu[1].sniffer); rm 291 drivers/macintosh/rack-meter.c static int rackmeter_setup(struct rackmeter *rm) rm 294 drivers/macintosh/rack-meter.c rackmeter_setup_i2s(rm); rm 297 drivers/macintosh/rack-meter.c rackmeter_set_default_pattern(rm); rm 300 drivers/macintosh/rack-meter.c rackmeter_setup_dbdma(rm); rm 303 drivers/macintosh/rack-meter.c rackmeter_init_cpu_sniffer(rm); rm 311 drivers/macintosh/rack-meter.c static u32 rackmeter_calc_sample(struct rackmeter *rm, unsigned int index) rm 318 drivers/macintosh/rack-meter.c sample |= ((rm->ubuf[led] >= 0x80) << 15); rm 325 drivers/macintosh/rack-meter.c struct rackmeter *rm = arg; rm 326 drivers/macintosh/rack-meter.c struct rackmeter_dma *db = rm->dma_buf_v; rm 334 drivers/macintosh/rack-meter.c (void)in_le32(&rm->dma_regs->status); rm 345 drivers/macintosh/rack-meter.c if (++rm->stale_irq > 3) { rm 348 drivers/macintosh/rack-meter.c DBDMA_DO_RESET(rm->dma_regs); rm 360 drivers/macintosh/rack-meter.c buf[i] = rackmeter_calc_sample(rm, i); rm 370 drivers/macintosh/rack-meter.c struct rackmeter *rm = NULL; rm 399 drivers/macintosh/rack-meter.c rm = kzalloc(sizeof(*rm), GFP_KERNEL); rm 400 drivers/macintosh/rack-meter.c if (rm == NULL) { rm 405 drivers/macintosh/rack-meter.c rm->mdev = mdev; rm 406 drivers/macintosh/rack-meter.c rm->i2s = i2s; rm 407 drivers/macintosh/rack-meter.c mutex_init(&rm->sem); rm 408 drivers/macintosh/rack-meter.c dev_set_drvdata(&mdev->ofdev.dev, rm); rm 426 drivers/macintosh/rack-meter.c rm->irq = macio_irq(mdev, 1); rm 428 drivers/macintosh/rack-meter.c rm->irq = irq_of_parse_and_map(i2s, 1); rm 429 drivers/macintosh/rack-meter.c if (!rm->irq || rm 442 drivers/macintosh/rack-meter.c pr_debug(" irq %d\n", rm->irq); rm 444 drivers/macintosh/rack-meter.c rm->ubuf = (u8 *)__get_free_page(GFP_KERNEL); rm 445 drivers/macintosh/rack-meter.c if (rm->ubuf == NULL) { rm 452 drivers/macintosh/rack-meter.c rm->dma_buf_v = dma_alloc_coherent(&macio_get_pci_dev(mdev)->dev, rm 454 drivers/macintosh/rack-meter.c &rm->dma_buf_p, GFP_KERNEL); rm 455 drivers/macintosh/rack-meter.c if (rm->dma_buf_v == NULL) { rm 462 drivers/macintosh/rack-meter.c rm->i2s_regs = ioremap(macio_resource_start(mdev, 0), 0x1000); rm 464 drivers/macintosh/rack-meter.c rm->i2s_regs = ioremap(ri2s.start, 0x1000); rm 466 drivers/macintosh/rack-meter.c if (rm->i2s_regs == NULL) { rm 473 drivers/macintosh/rack-meter.c rm->dma_regs = ioremap(macio_resource_start(mdev, 1), 0x100); rm 475 drivers/macintosh/rack-meter.c rm->dma_regs = ioremap(rdma.start, 0x100); rm 477 drivers/macintosh/rack-meter.c if (rm->dma_regs == NULL) { rm 484 drivers/macintosh/rack-meter.c rc = rackmeter_setup(rm); rm 492 drivers/macintosh/rack-meter.c rc = request_irq(rm->irq, rackmeter_irq, 0, "rackmeter", rm); rm 502 drivers/macintosh/rack-meter.c DBDMA_DO_RESET(rm->dma_regs); rm 504 drivers/macintosh/rack-meter.c iounmap(rm->dma_regs); rm 506 drivers/macintosh/rack-meter.c iounmap(rm->i2s_regs); rm 510 drivers/macintosh/rack-meter.c rm->dma_buf_v, rm->dma_buf_p); rm 512 drivers/macintosh/rack-meter.c free_page((unsigned long)rm->ubuf); rm 518 drivers/macintosh/rack-meter.c kfree(rm); rm 528 drivers/macintosh/rack-meter.c struct rackmeter *rm = dev_get_drvdata(&mdev->ofdev.dev); rm 531 drivers/macintosh/rack-meter.c rackmeter_stop_cpu_sniffer(rm); rm 537 drivers/macintosh/rack-meter.c DBDMA_DO_RESET(rm->dma_regs); rm 540 drivers/macintosh/rack-meter.c free_irq(rm->irq, rm); rm 543 drivers/macintosh/rack-meter.c iounmap(rm->dma_regs); rm 544 drivers/macintosh/rack-meter.c iounmap(rm->i2s_regs); rm 549 drivers/macintosh/rack-meter.c rm->dma_buf_v, rm->dma_buf_p); rm 552 drivers/macintosh/rack-meter.c free_page((unsigned long)rm->ubuf); rm 560 drivers/macintosh/rack-meter.c kfree(rm); rm 567 drivers/macintosh/rack-meter.c struct rackmeter *rm = dev_get_drvdata(&mdev->ofdev.dev); rm 569 drivers/macintosh/rack-meter.c if (rm == NULL) rm 573 drivers/macintosh/rack-meter.c rackmeter_stop_cpu_sniffer(rm); rm 576 drivers/macintosh/rack-meter.c DBDMA_DO_RESET(rm->dma_regs); rm 2252 drivers/media/pci/ddbridge/ddbridge-core.c const struct ddb_regmap *rm = io_regmap(io, 0); rm 2262 drivers/media/pci/ddbridge/ddbridge-core.c dma->regs = rm->odma->base + rm->odma->size * nr; rm 2263 drivers/media/pci/ddbridge/ddbridge-core.c dma->bufregs = rm->odma_buf->base + rm->odma_buf->size * nr; rm 2269 drivers/media/pci/ddbridge/ddbridge-core.c dma->regs = rm->idma->base + rm->idma->size * nr; rm 2270 drivers/media/pci/ddbridge/ddbridge-core.c dma->bufregs = rm->idma_buf->base + rm->idma_buf->size * nr; rm 2284 drivers/media/pci/ddbridge/ddbridge-core.c const struct ddb_regmap *rm; rm 2289 drivers/media/pci/ddbridge/ddbridge-core.c rm = io_regmap(input, 1); rm 2291 drivers/media/pci/ddbridge/ddbridge-core.c (rm->input->base + rm->input->size * nr); rm 2315 drivers/media/pci/ddbridge/ddbridge-core.c const struct ddb_regmap *rm; rm 2320 drivers/media/pci/ddbridge/ddbridge-core.c rm = io_regmap(output, 1); rm 2322 drivers/media/pci/ddbridge/ddbridge-core.c (rm->output->base + rm->output->size * nr); rm 2370 drivers/media/pci/ddbridge/ddbridge-core.c const struct ddb_regmap *rm; rm 2376 drivers/media/pci/ddbridge/ddbridge-core.c rm = info->regmap; rm 2377 drivers/media/pci/ddbridge/ddbridge-core.c if (!rm) rm 518 drivers/mtd/nand/raw/mpc5121_nfc.c struct mpc512x_reset_module *rm; rm 534 drivers/mtd/nand/raw/mpc5121_nfc.c rm = of_iomap(rmnode, 0); rm 535 drivers/mtd/nand/raw/mpc5121_nfc.c if (!rm) { rm 541 drivers/mtd/nand/raw/mpc5121_nfc.c rcwh = in_be32(&rm->rcwhr); rm 586 drivers/mtd/nand/raw/mpc5121_nfc.c iounmap(rm); rm 470 drivers/net/can/usb/peak_usb/pcan_usb_fd.c struct pucan_rx_msg *rm = (struct pucan_rx_msg *)rx_msg; rm 471 drivers/net/can/usb/peak_usb/pcan_usb_fd.c struct peak_usb_device *dev = usb_if->dev[pucan_msg_get_channel(rm)]; rm 475 drivers/net/can/usb/peak_usb/pcan_usb_fd.c const u16 rx_msg_flags = le16_to_cpu(rm->flags); rm 489 drivers/net/can/usb/peak_usb/pcan_usb_fd.c cfd->len = can_dlc2len(get_canfd_dlc(pucan_msg_get_dlc(rm))); rm 496 drivers/net/can/usb/peak_usb/pcan_usb_fd.c cfd->len = get_can_dlc(pucan_msg_get_dlc(rm)); rm 499 drivers/net/can/usb/peak_usb/pcan_usb_fd.c cfd->can_id = le32_to_cpu(rm->can_id); rm 507 drivers/net/can/usb/peak_usb/pcan_usb_fd.c memcpy(cfd->data, rm->d, cfd->len); rm 509 drivers/net/can/usb/peak_usb/pcan_usb_fd.c peak_usb_netif_rx(skb, &usb_if->time_ref, le32_to_cpu(rm->ts_low)); rm 1891 drivers/net/ethernet/apm/xgene/xgene_enet_main.c pdata->rm = RM3; rm 1899 drivers/net/ethernet/apm/xgene/xgene_enet_main.c pdata->rm = RM1; rm 1908 drivers/net/ethernet/apm/xgene/xgene_enet_main.c pdata->rm = RM0; rm 1966 drivers/net/ethernet/apm/xgene/xgene_enet_main.c pdata->rm = RM0; rm 213 drivers/net/ethernet/apm/xgene/xgene_enet_main.h enum xgene_enet_rm rm; rm 259 drivers/net/ethernet/apm/xgene/xgene_enet_main.h return ((u16)pdata->rm << 10) | ring->num; rm 75 drivers/net/ethernet/chelsio/cxgb/common.h #define t1_rx_mode_promisc(rm) (rm->dev->flags & IFF_PROMISC) rm 76 drivers/net/ethernet/chelsio/cxgb/common.h #define t1_rx_mode_allmulti(rm) (rm->dev->flags & IFF_ALLMULTI) rm 77 drivers/net/ethernet/chelsio/cxgb/common.h #define t1_rx_mode_mc_cnt(rm) (netdev_mc_count(rm->dev)) rm 78 drivers/net/ethernet/chelsio/cxgb/common.h #define t1_get_netdev(rm) (rm->dev) rm 119 drivers/net/ethernet/chelsio/cxgb/cxgb2.c struct t1_rx_mode rm; rm 121 drivers/net/ethernet/chelsio/cxgb/cxgb2.c rm.dev = dev; rm 122 drivers/net/ethernet/chelsio/cxgb/cxgb2.c mac->ops->set_rx_mode(mac, &rm); rm 111 drivers/net/ethernet/chelsio/cxgb/gmac.h int (*set_rx_mode)(struct cmac *, struct t1_rx_mode *rm); rm 345 drivers/net/ethernet/chelsio/cxgb/pm3393.c static int pm3393_set_rx_mode(struct cmac *cmac, struct t1_rx_mode *rm) rm 360 drivers/net/ethernet/chelsio/cxgb/pm3393.c if (t1_rx_mode_promisc(rm)) { rm 364 drivers/net/ethernet/chelsio/cxgb/pm3393.c if (t1_rx_mode_allmulti(rm)) { rm 371 drivers/net/ethernet/chelsio/cxgb/pm3393.c } else if (t1_rx_mode_mc_cnt(rm)) { rm 377 drivers/net/ethernet/chelsio/cxgb/pm3393.c netdev_for_each_mc_addr(ha, t1_get_netdev(rm)) { rm 433 drivers/net/ethernet/chelsio/cxgb/vsc7326.c static int mac_set_rx_mode(struct cmac *mac, struct t1_rx_mode *rm) rm 441 drivers/net/ethernet/chelsio/cxgb/vsc7326.c if (t1_rx_mode_promisc(rm)) rm 139 drivers/parport/parport_gsc.h const unsigned char rm = (PARPORT_CONTROL_STROBE | rm 144 drivers/parport/parport_gsc.h return priv->ctr & rm; /* Use soft copy */ rm 936 drivers/parport/parport_ip32.c const unsigned int rm = rm 938 drivers/parport/parport_ip32.c return __parport_ip32_read_control(p) & rm; rm 1050 drivers/pinctrl/pinctrl-st.c struct regmap *rm = info->regmap; rm 1065 drivers/pinctrl/pinctrl-st.c rt_p->clk1notclk0 = devm_regmap_field_alloc(dev, rm, clk1notclk0); rm 1066 drivers/pinctrl/pinctrl-st.c rt_p->delay_0 = devm_regmap_field_alloc(dev, rm, delay_0); rm 1067 drivers/pinctrl/pinctrl-st.c rt_p->delay_1 = devm_regmap_field_alloc(dev, rm, delay_1); rm 1068 drivers/pinctrl/pinctrl-st.c rt_p->invertclk = devm_regmap_field_alloc(dev, rm, invertclk); rm 1069 drivers/pinctrl/pinctrl-st.c rt_p->retime = devm_regmap_field_alloc(dev, rm, retime); rm 1070 drivers/pinctrl/pinctrl-st.c rt_p->clknotdata = devm_regmap_field_alloc(dev, rm, clknotdata); rm 1071 drivers/pinctrl/pinctrl-st.c rt_p->double_edge = devm_regmap_field_alloc(dev, rm, double_edge); rm 1086 drivers/pinctrl/pinctrl-st.c struct regmap *rm = info->regmap; rm 1097 drivers/pinctrl/pinctrl-st.c rt_d->rt[j] = devm_regmap_field_alloc(dev, rm, reg); rm 1258 drivers/pinctrl/stm32/pinctrl-stm32.c struct regmap *rm; rm 1266 drivers/pinctrl/stm32/pinctrl-stm32.c rm = pctl->regmap; rm 1288 drivers/pinctrl/stm32/pinctrl-stm32.c pctl->irqmux[i] = devm_regmap_field_alloc(dev, rm, mux); rm 2130 drivers/pinctrl/tegra/pinctrl-tegra20.c MUX_PG(rm, I2C1, RSVD2, RSVD3, RSVD4, 0x14, 25, 0x80, 14, 0xa4, 0), rm 131 drivers/s390/net/ctcm_mpc.c __u32 ct, sw, rm, dup; rm 141 drivers/s390/net/ctcm_mpc.c rm = 16; rm 169 drivers/s390/net/ctcm_mpc.c rm--; rm 187 drivers/s390/net/ctcm_mpc.c rm = 16; rm 191 drivers/s390/net/ctcm_mpc.c for ( ; rm > 0; rm--, sw++) { rm 847 drivers/video/console/sticore.c u32 *rm; rm 865 drivers/video/console/sticore.c rm = (u32*) &sti->rm_entry; rm 866 drivers/video/console/sticore.c *rm++ = gsc_readl(address+rm_offset+0x00); rm 867 drivers/video/console/sticore.c *rm++ = gsc_readl(address+rm_offset+0x04); rm 868 drivers/video/console/sticore.c *rm++ = gsc_readl(address+rm_offset+0x08); rm 869 drivers/video/console/sticore.c *rm++ = gsc_readl(address+rm_offset+0x0c); rm 288 drivers/video/fbdev/pm3fb.c unsigned long rm = 1 | (3 << 7); rm 294 drivers/video/fbdev/pm3fb.c rm |= 3 << 15; rm 301 drivers/video/fbdev/pm3fb.c rm |= 2 << 15; rm 313 drivers/video/fbdev/pm3fb.c PM3_WRITE_REG(par, PM3RasterizerMode, rm); rm 295 fs/gfs2/log.c int a, b, rm; rm 302 fs/gfs2/log.c rm = (wrap) ? (a || b) : (a && b); rm 303 fs/gfs2/log.c if (!rm) rm 175 fs/ocfs2/journal.c struct ocfs2_recovery_map *rm; rm 182 fs/ocfs2/journal.c rm = kzalloc(sizeof(struct ocfs2_recovery_map) + rm 185 fs/ocfs2/journal.c if (!rm) { rm 190 fs/ocfs2/journal.c rm->rm_entries = (unsigned int *)((char *)rm + rm 192 fs/ocfs2/journal.c osb->recovery_map = rm; rm 208 fs/ocfs2/journal.c struct ocfs2_recovery_map *rm; rm 227 fs/ocfs2/journal.c rm = osb->recovery_map; rm 230 fs/ocfs2/journal.c kfree(rm); rm 237 fs/ocfs2/journal.c struct ocfs2_recovery_map *rm = osb->recovery_map; rm 241 fs/ocfs2/journal.c for (i = 0; i < rm->rm_used; i++) { rm 242 fs/ocfs2/journal.c if (rm->rm_entries[i] == node_num) rm 253 fs/ocfs2/journal.c struct ocfs2_recovery_map *rm = osb->recovery_map; rm 262 fs/ocfs2/journal.c BUG_ON(rm->rm_used >= osb->max_slots); rm 264 fs/ocfs2/journal.c rm->rm_entries[rm->rm_used] = node_num; rm 265 fs/ocfs2/journal.c rm->rm_used++; rm 275 fs/ocfs2/journal.c struct ocfs2_recovery_map *rm = osb->recovery_map; rm 279 fs/ocfs2/journal.c for (i = 0; i < rm->rm_used; i++) { rm 280 fs/ocfs2/journal.c if (rm->rm_entries[i] == node_num) rm 284 fs/ocfs2/journal.c if (i < rm->rm_used) { rm 286 fs/ocfs2/journal.c memmove(&(rm->rm_entries[i]), &(rm->rm_entries[i + 1]), rm 287 fs/ocfs2/journal.c (rm->rm_used - i - 1) * sizeof(unsigned int)); rm 288 fs/ocfs2/journal.c rm->rm_used--; rm 1127 fs/ocfs2/journal.c struct ocfs2_recovery_map *rm = osb->recovery_map; rm 1130 fs/ocfs2/journal.c empty = (rm->rm_used == 0); rm 1369 fs/ocfs2/journal.c struct ocfs2_recovery_map *rm = osb->recovery_map; rm 1408 fs/ocfs2/journal.c while (rm->rm_used) { rm 1411 fs/ocfs2/journal.c node_num = rm->rm_entries[0]; rm 218 fs/ocfs2/super.c struct ocfs2_recovery_map *rm = osb->recovery_map; rm 271 fs/ocfs2/super.c if (rm->rm_used == 0) rm 274 fs/ocfs2/super.c for (i = 0; i < rm->rm_used; i++) rm 276 fs/ocfs2/super.c rm->rm_entries[i]); rm 68 fs/xfs/scrub/refcount.c struct xfs_rmap_irec rm; rm 132 fs/xfs/scrub/refcount.c memcpy(&frag->rm, rec, sizeof(frag->rm)); rm 178 fs/xfs/scrub/refcount.c if (frag->rm.rm_startblock < bno) rm 180 fs/xfs/scrub/refcount.c bno = frag->rm.rm_startblock; rm 188 fs/xfs/scrub/refcount.c if (frag->rm.rm_startblock > refchk->bno) rm 190 fs/xfs/scrub/refcount.c bno = frag->rm.rm_startblock + frag->rm.rm_blockcount; rm 211 fs/xfs/scrub/refcount.c bno = frag->rm.rm_startblock + frag->rm.rm_blockcount; rm 224 fs/xfs/scrub/refcount.c bno = frag->rm.rm_startblock + frag->rm.rm_blockcount; rm 225 fs/xfs/scrub/refcount.c if (frag->rm.rm_startblock != rbno) rm 221 include/linux/math64.h } rl, rm, rn, rh, a0, b0; rm 228 include/linux/math64.h rm.ll = mul_u32_u32(a0.l.low, b0.l.high); rm 237 include/linux/math64.h rl.l.high = c = (u64)rl.l.high + rm.l.low + rn.l.low; rm 238 include/linux/math64.h rh.l.low = c = (c >> 32) + rm.l.high + rn.l.high + rh.l.low; rm 176 include/linux/parport_pc.h const unsigned char rm = (PARPORT_CONTROL_STROBE | rm 181 include/linux/parport_pc.h return priv->ctr & rm; /* Use soft copy */ rm 27 include/linux/transport_class.h #define DECLARE_TRANSPORT_CLASS(cls, nm, su, rm, cfg) \ rm 33 include/linux/transport_class.h .remove = rm, \ rm 388 net/ipv4/ipconfig.c struct rtentry rm; rm 391 net/ipv4/ipconfig.c memset(&rm, 0, sizeof(rm)); rm 396 net/ipv4/ipconfig.c set_sockaddr((struct sockaddr_in *) &rm.rt_dst, 0, 0); rm 397 net/ipv4/ipconfig.c set_sockaddr((struct sockaddr_in *) &rm.rt_genmask, 0, 0); rm 398 net/ipv4/ipconfig.c set_sockaddr((struct sockaddr_in *) &rm.rt_gateway, ic_gateway, 0); rm 399 net/ipv4/ipconfig.c rm.rt_flags = RTF_UP | RTF_GATEWAY; rm 400 net/ipv4/ipconfig.c if ((err = ip_rt_ioctl(&init_net, SIOCADDRT, &rm)) < 0) { rm 421 net/rds/cong.c struct rds_message *rm; rm 423 net/rds/cong.c rm = rds_message_map_pages(map->m_page_addrs, RDS_CONG_MAP_BYTES); rm 424 net/rds/cong.c if (!IS_ERR(rm)) rm 425 net/rds/cong.c rm->m_inc.i_hdr.h_flags = RDS_FLAG_CONG_BITMAP; rm 427 net/rds/cong.c return rm; rm 431 net/rds/connection.c struct rds_message *rm, *rtmp; rm 444 net/rds/connection.c list_for_each_entry_safe(rm, rtmp, rm 447 net/rds/connection.c list_del_init(&rm->m_conn_item); rm 448 net/rds/connection.c BUG_ON(!list_empty(&rm->m_sock_item)); rm 449 net/rds/connection.c rds_message_put(rm); rm 530 net/rds/connection.c struct rds_message *rm; rm 565 net/rds/connection.c list_for_each_entry(rm, list, m_conn_item) { rm 568 net/rds/connection.c __rds_inc_msg_cp(&rm->m_inc, rm 431 net/rds/ib.h int rds_ib_xmit(struct rds_connection *conn, struct rds_message *rm, rm 1088 net/rds/ib_cm.c struct rds_message *rm; rm 1090 net/rds/ib_cm.c rm = container_of(ic->i_data_op, struct rds_message, data); rm 1091 net/rds/ib_cm.c rds_message_put(rm); rm 47 net/rds/ib_send.c static void rds_ib_send_complete(struct rds_message *rm, rm 49 net/rds/ib_send.c void (*complete)(struct rds_message *rm, int status)) rm 69 net/rds/ib_send.c complete(rm, notify_status); rm 153 net/rds/ib_send.c struct rds_message *rm = NULL; rm 159 net/rds/ib_send.c rm = container_of(send->s_op, struct rds_message, data); rm 166 net/rds/ib_send.c rm = container_of(send->s_op, struct rds_message, rdma); rm 173 net/rds/ib_send.c rm = container_of(send->s_op, struct rds_message, atomic); rm 186 net/rds/ib_send.c return rm; rm 243 net/rds/ib_send.c struct rds_message *rm = NULL; rm 274 net/rds/ib_send.c rm = rds_ib_send_unmap_op(ic, send, wc->status); rm 280 net/rds/ib_send.c if (send->s_op == rm->m_final_op) { rm 284 net/rds/ib_send.c rds_message_unmapped(rm); rm 286 net/rds/ib_send.c rds_message_put(rm); rm 483 net/rds/ib_send.c int rds_ib_xmit(struct rds_connection *conn, struct rds_message *rm, rm 510 net/rds/ib_send.c && rm->m_inc.i_hdr.h_flags & RDS_FLAG_CONG_BITMAP) { rm 512 net/rds/ib_send.c scat = &rm->data.op_sg[sg]; rm 518 net/rds/ib_send.c if (be32_to_cpu(rm->m_inc.i_hdr.h_len) == 0) rm 521 net/rds/ib_send.c i = DIV_ROUND_UP(be32_to_cpu(rm->m_inc.i_hdr.h_len), RDS_FRAG_SIZE); rm 549 net/rds/ib_send.c if (rm->data.op_nents) { rm 550 net/rds/ib_send.c rm->data.op_count = ib_dma_map_sg(dev, rm 551 net/rds/ib_send.c rm->data.op_sg, rm 552 net/rds/ib_send.c rm->data.op_nents, rm 554 net/rds/ib_send.c rdsdebug("ic %p mapping rm %p: %d\n", ic, rm, rm->data.op_count); rm 555 net/rds/ib_send.c if (rm->data.op_count == 0) { rm 562 net/rds/ib_send.c rm->data.op_count = 0; rm 565 net/rds/ib_send.c rds_message_addref(rm); rm 566 net/rds/ib_send.c rm->data.op_dmasg = 0; rm 567 net/rds/ib_send.c rm->data.op_dmaoff = 0; rm 568 net/rds/ib_send.c ic->i_data_op = &rm->data; rm 571 net/rds/ib_send.c if (test_bit(RDS_MSG_ACK_REQUIRED, &rm->m_flags)) rm 572 net/rds/ib_send.c rm->m_inc.i_hdr.h_flags |= RDS_FLAG_ACK_REQUIRED; rm 573 net/rds/ib_send.c if (test_bit(RDS_MSG_RETRANSMITTED, &rm->m_flags)) rm 574 net/rds/ib_send.c rm->m_inc.i_hdr.h_flags |= RDS_FLAG_RETRANSMITTED; rm 578 net/rds/ib_send.c if (rm->rdma.op_active) { rm 581 net/rds/ib_send.c ext_hdr.h_rdma_rkey = cpu_to_be32(rm->rdma.op_rkey); rm 582 net/rds/ib_send.c rds_message_add_extension(&rm->m_inc.i_hdr, rm 585 net/rds/ib_send.c if (rm->m_rdma_cookie) { rm 586 net/rds/ib_send.c rds_message_add_rdma_dest_extension(&rm->m_inc.i_hdr, rm 587 net/rds/ib_send.c rds_rdma_cookie_key(rm->m_rdma_cookie), rm 588 net/rds/ib_send.c rds_rdma_cookie_offset(rm->m_rdma_cookie)); rm 595 net/rds/ib_send.c rm->m_inc.i_hdr.h_ack = cpu_to_be64(rds_ib_piggyb_ack(ic)); rm 596 net/rds/ib_send.c rds_message_make_checksum(&rm->m_inc.i_hdr); rm 614 net/rds/ib_send.c if (rm->rdma.op_active && rm->rdma.op_fence) rm 621 net/rds/ib_send.c scat = &ic->i_data_op->op_sg[rm->data.op_dmasg]; rm 638 net/rds/ib_send.c memcpy(&ic->i_send_hdrs[pos], &rm->m_inc.i_hdr, sizeof(struct rds_header)); rm 642 net/rds/ib_send.c && scat != &rm->data.op_sg[rm->data.op_count]) { rm 644 net/rds/ib_send.c sg_dma_len(scat) - rm->data.op_dmaoff); rm 648 net/rds/ib_send.c send->s_sge[1].addr += rm->data.op_dmaoff; rm 652 net/rds/ib_send.c rm->data.op_dmaoff += len; rm 653 net/rds/ib_send.c if (rm->data.op_dmaoff == sg_dma_len(scat)) { rm 655 net/rds/ib_send.c rm->data.op_dmasg++; rm 656 net/rds/ib_send.c rm->data.op_dmaoff = 0; rm 695 net/rds/ib_send.c && scat != &rm->data.op_sg[rm->data.op_count]); rm 703 net/rds/ib_send.c if (scat == &rm->data.op_sg[rm->data.op_count]) { rm 75 net/rds/loop.c static int rds_loop_xmit(struct rds_connection *conn, struct rds_message *rm, rm 79 net/rds/loop.c struct scatterlist *sgp = &rm->data.op_sg[sg]; rm 81 net/rds/loop.c be32_to_cpu(rm->m_inc.i_hdr.h_len); rm 84 net/rds/loop.c if (rm->m_inc.i_hdr.h_flags & RDS_FLAG_CONG_BITMAP) { rm 92 net/rds/loop.c rds_inc_init(&rm->m_inc, conn, &conn->c_laddr); rm 94 net/rds/loop.c rds_message_addref(rm); rm 96 net/rds/loop.c rds_recv_incoming(conn, &conn->c_laddr, &conn->c_faddr, &rm->m_inc, rm 99 net/rds/loop.c rds_send_drop_acked(conn, be64_to_cpu(rm->m_inc.i_hdr.h_sequence), rm 102 net/rds/loop.c rds_inc_put(&rm->m_inc); rm 113 net/rds/loop.c struct rds_message *rm = container_of(inc, struct rds_message, m_inc); rm 115 net/rds/loop.c rds_message_put(rm); rm 51 net/rds/message.c void rds_message_addref(struct rds_message *rm) rm 53 net/rds/message.c rdsdebug("addref rm %p ref %d\n", rm, refcount_read(&rm->m_refcount)); rm 54 net/rds/message.c refcount_inc(&rm->m_refcount); rm 130 net/rds/message.c static void rds_message_purge(struct rds_message *rm) rm 135 net/rds/message.c if (unlikely(test_bit(RDS_MSG_PAGEVEC, &rm->m_flags))) rm 138 net/rds/message.c spin_lock_irqsave(&rm->m_rs_lock, flags); rm 139 net/rds/message.c if (rm->m_rs) { rm 140 net/rds/message.c struct rds_sock *rs = rm->m_rs; rm 142 net/rds/message.c if (rm->data.op_mmp_znotifier) { rm 144 net/rds/message.c rds_rm_zerocopy_callback(rs, rm->data.op_mmp_znotifier); rm 146 net/rds/message.c rm->data.op_mmp_znotifier = NULL; rm 149 net/rds/message.c rm->m_rs = NULL; rm 151 net/rds/message.c spin_unlock_irqrestore(&rm->m_rs_lock, flags); rm 153 net/rds/message.c for (i = 0; i < rm->data.op_nents; i++) { rm 156 net/rds/message.c __free_page(sg_page(&rm->data.op_sg[i])); rm 158 net/rds/message.c put_page(sg_page(&rm->data.op_sg[i])); rm 160 net/rds/message.c rm->data.op_nents = 0; rm 162 net/rds/message.c if (rm->rdma.op_active) rm 163 net/rds/message.c rds_rdma_free_op(&rm->rdma); rm 164 net/rds/message.c if (rm->rdma.op_rdma_mr) rm 165 net/rds/message.c rds_mr_put(rm->rdma.op_rdma_mr); rm 167 net/rds/message.c if (rm->atomic.op_active) rm 168 net/rds/message.c rds_atomic_free_op(&rm->atomic); rm 169 net/rds/message.c if (rm->atomic.op_rdma_mr) rm 170 net/rds/message.c rds_mr_put(rm->atomic.op_rdma_mr); rm 173 net/rds/message.c void rds_message_put(struct rds_message *rm) rm 175 net/rds/message.c rdsdebug("put rm %p ref %d\n", rm, refcount_read(&rm->m_refcount)); rm 176 net/rds/message.c WARN(!refcount_read(&rm->m_refcount), "danger refcount zero on %p\n", rm); rm 177 net/rds/message.c if (refcount_dec_and_test(&rm->m_refcount)) { rm 178 net/rds/message.c BUG_ON(!list_empty(&rm->m_sock_item)); rm 179 net/rds/message.c BUG_ON(!list_empty(&rm->m_conn_item)); rm 180 net/rds/message.c rds_message_purge(rm); rm 182 net/rds/message.c kfree(rm); rm 286 net/rds/message.c struct rds_message *rm; rm 291 net/rds/message.c rm = kzalloc(sizeof(struct rds_message) + extra_len, gfp); rm 292 net/rds/message.c if (!rm) rm 295 net/rds/message.c rm->m_used_sgs = 0; rm 296 net/rds/message.c rm->m_total_sgs = extra_len / sizeof(struct scatterlist); rm 298 net/rds/message.c refcount_set(&rm->m_refcount, 1); rm 299 net/rds/message.c INIT_LIST_HEAD(&rm->m_sock_item); rm 300 net/rds/message.c INIT_LIST_HEAD(&rm->m_conn_item); rm 301 net/rds/message.c spin_lock_init(&rm->m_rs_lock); rm 302 net/rds/message.c init_waitqueue_head(&rm->m_flush_wait); rm 305 net/rds/message.c return rm; rm 311 net/rds/message.c struct scatterlist *rds_message_alloc_sgs(struct rds_message *rm, int nents) rm 313 net/rds/message.c struct scatterlist *sg_first = (struct scatterlist *) &rm[1]; rm 321 net/rds/message.c if (rm->m_used_sgs + nents > rm->m_total_sgs) { rm 323 net/rds/message.c rm->m_total_sgs, rm->m_used_sgs, nents); rm 327 net/rds/message.c sg_ret = &sg_first[rm->m_used_sgs]; rm 329 net/rds/message.c rm->m_used_sgs += nents; rm 336 net/rds/message.c struct rds_message *rm; rm 341 net/rds/message.c rm = rds_message_alloc(extra_bytes, GFP_NOWAIT); rm 342 net/rds/message.c if (!rm) rm 345 net/rds/message.c set_bit(RDS_MSG_PAGEVEC, &rm->m_flags); rm 346 net/rds/message.c rm->m_inc.i_hdr.h_len = cpu_to_be32(total_len); rm 347 net/rds/message.c rm->data.op_nents = DIV_ROUND_UP(total_len, PAGE_SIZE); rm 348 net/rds/message.c rm->data.op_sg = rds_message_alloc_sgs(rm, num_sgs); rm 349 net/rds/message.c if (IS_ERR(rm->data.op_sg)) { rm 350 net/rds/message.c rds_message_put(rm); rm 351 net/rds/message.c return ERR_CAST(rm->data.op_sg); rm 354 net/rds/message.c for (i = 0; i < rm->data.op_nents; ++i) { rm 355 net/rds/message.c sg_set_page(&rm->data.op_sg[i], rm 360 net/rds/message.c return rm; rm 363 net/rds/message.c static int rds_message_zcopy_from_user(struct rds_message *rm, struct iov_iter *from) rm 371 net/rds/message.c rm->m_inc.i_hdr.h_len = cpu_to_be32(iov_iter_count(from)); rm 376 net/rds/message.c sg = rm->data.op_sg; rm 382 net/rds/message.c rm->data.op_mmp_znotifier = &info->znotif; rm 383 net/rds/message.c if (mm_account_pinned_pages(&rm->data.op_mmp_znotifier->z_mmp, rm 399 net/rds/message.c for (i = 0; i < rm->data.op_nents; i++) rm 400 net/rds/message.c put_page(sg_page(&rm->data.op_sg[i])); rm 401 net/rds/message.c mmp = &rm->data.op_mmp_znotifier->z_mmp; rm 410 net/rds/message.c rm->data.op_nents++; rm 417 net/rds/message.c rm->data.op_mmp_znotifier = NULL; rm 421 net/rds/message.c int rds_message_copy_from_user(struct rds_message *rm, struct iov_iter *from, rm 429 net/rds/message.c rm->m_inc.i_hdr.h_len = cpu_to_be32(iov_iter_count(from)); rm 432 net/rds/message.c sg = rm->data.op_sg; rm 436 net/rds/message.c return rds_message_zcopy_from_user(rm, from); rm 444 net/rds/message.c rm->data.op_nents++; rm 468 net/rds/message.c struct rds_message *rm; rm 476 net/rds/message.c rm = container_of(inc, struct rds_message, m_inc); rm 477 net/rds/message.c len = be32_to_cpu(rm->m_inc.i_hdr.h_len); rm 479 net/rds/message.c sg = rm->data.op_sg; rm 510 net/rds/message.c void rds_message_wait(struct rds_message *rm) rm 512 net/rds/message.c wait_event_interruptible(rm->m_flush_wait, rm 513 net/rds/message.c !test_bit(RDS_MSG_MAPPED, &rm->m_flags)); rm 516 net/rds/message.c void rds_message_unmapped(struct rds_message *rm) rm 518 net/rds/message.c clear_bit(RDS_MSG_MAPPED, &rm->m_flags); rm 519 net/rds/message.c wake_up_interruptible(&rm->m_flush_wait); rm 572 net/rds/rdma.c int rds_cmsg_rdma_args(struct rds_sock *rs, struct rds_message *rm, rm 577 net/rds/rdma.c struct rm_rdma_op *op = &rm->rdma; rm 586 net/rds/rdma.c || rm->rdma.op_active) rm 627 net/rds/rdma.c op->op_sg = rds_message_alloc_sgs(rm, nr_pages); rm 730 net/rds/rdma.c int rds_cmsg_rdma_dest(struct rds_sock *rs, struct rds_message *rm, rm 739 net/rds/rdma.c rm->m_rdma_cookie != 0) rm 742 net/rds/rdma.c memcpy(&rm->m_rdma_cookie, CMSG_DATA(cmsg), sizeof(rm->m_rdma_cookie)); rm 749 net/rds/rdma.c r_key = rds_rdma_cookie_key(rm->m_rdma_cookie); rm 761 net/rds/rdma.c rm->rdma.op_rdma_mr = mr; rm 772 net/rds/rdma.c int rds_cmsg_rdma_map(struct rds_sock *rs, struct rds_message *rm, rm 776 net/rds/rdma.c rm->m_rdma_cookie != 0) rm 779 net/rds/rdma.c return __rds_rdma_map(rs, CMSG_DATA(cmsg), &rm->m_rdma_cookie, rm 780 net/rds/rdma.c &rm->rdma.op_rdma_mr, rm->m_conn_path); rm 786 net/rds/rdma.c int rds_cmsg_atomic(struct rds_sock *rs, struct rds_message *rm, rm 794 net/rds/rdma.c || rm->atomic.op_active) rm 802 net/rds/rdma.c rm->atomic.op_type = RDS_ATOMIC_TYPE_FADD; rm 803 net/rds/rdma.c rm->atomic.op_m_fadd.add = args->fadd.add; rm 804 net/rds/rdma.c rm->atomic.op_m_fadd.nocarry_mask = 0; rm 807 net/rds/rdma.c rm->atomic.op_type = RDS_ATOMIC_TYPE_FADD; rm 808 net/rds/rdma.c rm->atomic.op_m_fadd.add = args->m_fadd.add; rm 809 net/rds/rdma.c rm->atomic.op_m_fadd.nocarry_mask = args->m_fadd.nocarry_mask; rm 812 net/rds/rdma.c rm->atomic.op_type = RDS_ATOMIC_TYPE_CSWP; rm 813 net/rds/rdma.c rm->atomic.op_m_cswp.compare = args->cswp.compare; rm 814 net/rds/rdma.c rm->atomic.op_m_cswp.swap = args->cswp.swap; rm 815 net/rds/rdma.c rm->atomic.op_m_cswp.compare_mask = ~0; rm 816 net/rds/rdma.c rm->atomic.op_m_cswp.swap_mask = ~0; rm 819 net/rds/rdma.c rm->atomic.op_type = RDS_ATOMIC_TYPE_CSWP; rm 820 net/rds/rdma.c rm->atomic.op_m_cswp.compare = args->m_cswp.compare; rm 821 net/rds/rdma.c rm->atomic.op_m_cswp.swap = args->m_cswp.swap; rm 822 net/rds/rdma.c rm->atomic.op_m_cswp.compare_mask = args->m_cswp.compare_mask; rm 823 net/rds/rdma.c rm->atomic.op_m_cswp.swap_mask = args->m_cswp.swap_mask; rm 829 net/rds/rdma.c rm->atomic.op_notify = !!(args->flags & RDS_RDMA_NOTIFY_ME); rm 830 net/rds/rdma.c rm->atomic.op_silent = !!(args->flags & RDS_RDMA_SILENT); rm 831 net/rds/rdma.c rm->atomic.op_active = 1; rm 832 net/rds/rdma.c rm->atomic.op_recverr = rs->rs_recverr; rm 833 net/rds/rdma.c rm->atomic.op_sg = rds_message_alloc_sgs(rm, 1); rm 834 net/rds/rdma.c if (IS_ERR(rm->atomic.op_sg)) { rm 835 net/rds/rdma.c ret = PTR_ERR(rm->atomic.op_sg); rm 850 net/rds/rdma.c sg_set_page(rm->atomic.op_sg, page, 8, offset_in_page(args->local_addr)); rm 852 net/rds/rdma.c if (rm->atomic.op_notify || rm->atomic.op_recverr) { rm 858 net/rds/rdma.c rm->atomic.op_notifier = kmalloc(sizeof(*rm->atomic.op_notifier), GFP_KERNEL); rm 859 net/rds/rdma.c if (!rm->atomic.op_notifier) { rm 864 net/rds/rdma.c rm->atomic.op_notifier->n_user_token = args->user_token; rm 865 net/rds/rdma.c rm->atomic.op_notifier->n_status = RDS_RDMA_SUCCESS; rm 868 net/rds/rdma.c rm->atomic.op_rkey = rds_rdma_cookie_key(args->cookie); rm 869 net/rds/rdma.c rm->atomic.op_remote_addr = args->remote_addr + rds_rdma_cookie_offset(args->cookie); rm 875 net/rds/rdma.c rm->atomic.op_active = 0; rm 876 net/rds/rdma.c kfree(rm->atomic.op_notifier); rm 557 net/rds/rds.h int (*xmit)(struct rds_connection *conn, struct rds_message *rm, rm 852 net/rds/rds.h struct scatterlist *rds_message_alloc_sgs(struct rds_message *rm, int nents); rm 853 net/rds/rds.h int rds_message_copy_from_user(struct rds_message *rm, struct iov_iter *from, rm 865 net/rds/rds.h void rds_message_addref(struct rds_message *rm); rm 866 net/rds/rds.h void rds_message_put(struct rds_message *rm); rm 867 net/rds/rds.h void rds_message_wait(struct rds_message *rm); rm 868 net/rds/rds.h void rds_message_unmapped(struct rds_message *rm); rm 915 net/rds/rds.h typedef int (*is_acked_func)(struct rds_message *rm, uint64_t ack); rm 931 net/rds/rds.h int rds_cmsg_rdma_dest(struct rds_sock *rs, struct rds_message *rm, rm 933 net/rds/rds.h int rds_cmsg_rdma_args(struct rds_sock *rs, struct rds_message *rm, rm 936 net/rds/rds.h int rds_cmsg_rdma_map(struct rds_sock *rs, struct rds_message *rm, rm 940 net/rds/rds.h void rds_rdma_send_complete(struct rds_message *rm, int wc_status); rm 941 net/rds/rds.h void rds_atomic_send_complete(struct rds_message *rm, int wc_status); rm 942 net/rds/rds.h int rds_cmsg_atomic(struct rds_sock *rs, struct rds_message *rm, rm 136 net/rds/recv.c struct rds_message *rm, *tmp; rm 150 net/rds/recv.c list_for_each_entry_safe(rm, tmp, rm 153 net/rds/recv.c set_bit(RDS_MSG_FLUSH, &rm->m_flags); rm 67 net/rds/send.c struct rds_message *rm, *tmp; rm 71 net/rds/send.c rm = cp->cp_xmit_rm; rm 77 net/rds/send.c rds_message_unmapped(rm); rm 78 net/rds/send.c rds_message_put(rm); rm 95 net/rds/send.c list_for_each_entry_safe(rm, tmp, &cp->cp_retrans, m_conn_item) { rm 96 net/rds/send.c set_bit(RDS_MSG_ACK_REQUIRED, &rm->m_flags); rm 97 net/rds/send.c set_bit(RDS_MSG_RETRANSMITTED, &rm->m_flags); rm 140 net/rds/send.c struct rds_message *rm; rm 202 net/rds/send.c rm = cp->cp_xmit_rm; rm 204 net/rds/send.c if (!rm) { rm 219 net/rds/send.c if (!rm && test_and_clear_bit(0, &conn->c_map_queued)) { rm 220 net/rds/send.c rm = rds_cong_update_alloc(conn); rm 221 net/rds/send.c if (IS_ERR(rm)) { rm 222 net/rds/send.c ret = PTR_ERR(rm); rm 225 net/rds/send.c rm->data.op_active = 1; rm 226 net/rds/send.c rm->m_inc.i_conn_path = cp; rm 227 net/rds/send.c rm->m_inc.i_conn = cp->cp_conn; rm 229 net/rds/send.c cp->cp_xmit_rm = rm; rm 239 net/rds/send.c if (!rm) { rm 255 net/rds/send.c rm = list_entry(cp->cp_send_queue.next, rm 258 net/rds/send.c rds_message_addref(rm); rm 264 net/rds/send.c list_move_tail(&rm->m_conn_item, rm 270 net/rds/send.c if (!rm) rm 280 net/rds/send.c if (test_bit(RDS_MSG_FLUSH, &rm->m_flags) || rm 281 net/rds/send.c (rm->rdma.op_active && rm 282 net/rds/send.c test_bit(RDS_MSG_RETRANSMITTED, &rm->m_flags))) { rm 284 net/rds/send.c if (test_and_clear_bit(RDS_MSG_ON_CONN, &rm->m_flags)) rm 285 net/rds/send.c list_move(&rm->m_conn_item, &to_be_dropped); rm 291 net/rds/send.c len = ntohl(rm->m_inc.i_hdr.h_len); rm 294 net/rds/send.c set_bit(RDS_MSG_ACK_REQUIRED, &rm->m_flags); rm 306 net/rds/send.c cp->cp_xmit_rm = rm; rm 310 net/rds/send.c if (rm->rdma.op_active && !cp->cp_xmit_rdma_sent) { rm 311 net/rds/send.c rm->m_final_op = &rm->rdma; rm 315 net/rds/send.c set_bit(RDS_MSG_MAPPED, &rm->m_flags); rm 316 net/rds/send.c ret = conn->c_trans->xmit_rdma(conn, &rm->rdma); rm 318 net/rds/send.c clear_bit(RDS_MSG_MAPPED, &rm->m_flags); rm 319 net/rds/send.c wake_up_interruptible(&rm->m_flush_wait); rm 326 net/rds/send.c if (rm->atomic.op_active && !cp->cp_xmit_atomic_sent) { rm 327 net/rds/send.c rm->m_final_op = &rm->atomic; rm 331 net/rds/send.c set_bit(RDS_MSG_MAPPED, &rm->m_flags); rm 332 net/rds/send.c ret = conn->c_trans->xmit_atomic(conn, &rm->atomic); rm 334 net/rds/send.c clear_bit(RDS_MSG_MAPPED, &rm->m_flags); rm 335 net/rds/send.c wake_up_interruptible(&rm->m_flush_wait); rm 349 net/rds/send.c if (rm->data.op_nents == 0) { rm 353 net/rds/send.c ops_present = (rm->atomic.op_active || rm->rdma.op_active); rm 354 net/rds/send.c if (rm->atomic.op_active && !rm->atomic.op_silent) rm 356 net/rds/send.c if (rm->rdma.op_active && !rm->rdma.op_silent) rm 360 net/rds/send.c && !rm->m_rdma_cookie) rm 361 net/rds/send.c rm->data.op_active = 0; rm 364 net/rds/send.c if (rm->data.op_active && !cp->cp_xmit_data_sent) { rm 365 net/rds/send.c rm->m_final_op = &rm->data; rm 367 net/rds/send.c ret = conn->c_trans->xmit(conn, rm, rm 382 net/rds/send.c sg = &rm->data.op_sg[cp->cp_xmit_sg]; rm 393 net/rds/send.c rm->data.op_nents); rm 398 net/rds/send.c (cp->cp_xmit_sg == rm->data.op_nents)) rm 407 net/rds/send.c if (!rm->data.op_active || cp->cp_xmit_data_sent) { rm 416 net/rds/send.c rds_message_put(rm); rm 428 net/rds/send.c list_for_each_entry(rm, &to_be_dropped, m_conn_item) rm 429 net/rds/send.c rds_message_put(rm); rm 473 net/rds/send.c static void rds_send_sndbuf_remove(struct rds_sock *rs, struct rds_message *rm) rm 475 net/rds/send.c u32 len = be32_to_cpu(rm->m_inc.i_hdr.h_len); rm 486 net/rds/send.c static inline int rds_send_is_acked(struct rds_message *rm, u64 ack, rm 490 net/rds/send.c return is_acked(rm, ack); rm 491 net/rds/send.c return be64_to_cpu(rm->m_inc.i_hdr.h_sequence) <= ack; rm 500 net/rds/send.c void rds_rdma_send_complete(struct rds_message *rm, int status) rm 507 net/rds/send.c spin_lock_irqsave(&rm->m_rs_lock, flags); rm 509 net/rds/send.c ro = &rm->rdma; rm 510 net/rds/send.c if (test_bit(RDS_MSG_ON_SOCK, &rm->m_flags) && rm 513 net/rds/send.c rs = rm->m_rs; rm 524 net/rds/send.c spin_unlock_irqrestore(&rm->m_rs_lock, flags); rm 536 net/rds/send.c void rds_atomic_send_complete(struct rds_message *rm, int status) rm 543 net/rds/send.c spin_lock_irqsave(&rm->m_rs_lock, flags); rm 545 net/rds/send.c ao = &rm->atomic; rm 546 net/rds/send.c if (test_bit(RDS_MSG_ON_SOCK, &rm->m_flags) rm 549 net/rds/send.c rs = rm->m_rs; rm 560 net/rds/send.c spin_unlock_irqrestore(&rm->m_rs_lock, flags); rm 575 net/rds/send.c __rds_send_complete(struct rds_sock *rs, struct rds_message *rm, int status) rm 580 net/rds/send.c ro = &rm->rdma; rm 587 net/rds/send.c ao = &rm->atomic; rm 609 net/rds/send.c struct rds_message *rm; rm 614 net/rds/send.c rm = list_entry(messages->next, struct rds_message, rm 616 net/rds/send.c list_del_init(&rm->m_conn_item); rm 628 net/rds/send.c spin_lock_irqsave(&rm->m_rs_lock, flags); rm 629 net/rds/send.c if (!test_bit(RDS_MSG_ON_SOCK, &rm->m_flags)) rm 632 net/rds/send.c if (rs != rm->m_rs) { rm 637 net/rds/send.c rs = rm->m_rs; rm 645 net/rds/send.c if (test_and_clear_bit(RDS_MSG_ON_SOCK, &rm->m_flags)) { rm 646 net/rds/send.c struct rm_rdma_op *ro = &rm->rdma; rm 649 net/rds/send.c list_del_init(&rm->m_sock_item); rm 650 net/rds/send.c rds_send_sndbuf_remove(rs, rm); rm 659 net/rds/send.c rm->rdma.op_notifier = NULL; rm 666 net/rds/send.c spin_unlock_irqrestore(&rm->m_rs_lock, flags); rm 667 net/rds/send.c rds_message_put(rm); rm 669 net/rds/send.c rds_message_put(rm); rm 689 net/rds/send.c struct rds_message *rm, *tmp; rm 695 net/rds/send.c list_for_each_entry_safe(rm, tmp, &cp->cp_retrans, m_conn_item) { rm 696 net/rds/send.c if (!rds_send_is_acked(rm, ack, is_acked)) rm 699 net/rds/send.c list_move(&rm->m_conn_item, &list); rm 700 net/rds/send.c clear_bit(RDS_MSG_ON_CONN, &rm->m_flags); rm 724 net/rds/send.c struct rds_message *rm, *tmp; rm 733 net/rds/send.c list_for_each_entry_safe(rm, tmp, &rs->rs_send_queue, m_sock_item) { rm 735 net/rds/send.c (!ipv6_addr_equal(&dest->sin6_addr, &rm->m_daddr) || rm 736 net/rds/send.c dest->sin6_port != rm->m_inc.i_hdr.h_dport)) rm 739 net/rds/send.c list_move(&rm->m_sock_item, &list); rm 740 net/rds/send.c rds_send_sndbuf_remove(rs, rm); rm 741 net/rds/send.c clear_bit(RDS_MSG_ON_SOCK, &rm->m_flags); rm 753 net/rds/send.c list_for_each_entry(rm, &list, m_sock_item) { rm 755 net/rds/send.c conn = rm->m_inc.i_conn; rm 757 net/rds/send.c cp = rm->m_inc.i_conn_path; rm 767 net/rds/send.c if (!test_and_clear_bit(RDS_MSG_ON_CONN, &rm->m_flags)) { rm 771 net/rds/send.c list_del_init(&rm->m_conn_item); rm 778 net/rds/send.c spin_lock_irqsave(&rm->m_rs_lock, flags); rm 781 net/rds/send.c __rds_send_complete(rs, rm, RDS_RDMA_CANCELED); rm 784 net/rds/send.c spin_unlock_irqrestore(&rm->m_rs_lock, flags); rm 786 net/rds/send.c rds_message_put(rm); rm 792 net/rds/send.c rm = list_entry(list.next, struct rds_message, m_sock_item); rm 793 net/rds/send.c list_del_init(&rm->m_sock_item); rm 794 net/rds/send.c rds_message_wait(rm); rm 801 net/rds/send.c spin_lock_irqsave(&rm->m_rs_lock, flags); rm 804 net/rds/send.c __rds_send_complete(rs, rm, RDS_RDMA_CANCELED); rm 807 net/rds/send.c spin_unlock_irqrestore(&rm->m_rs_lock, flags); rm 809 net/rds/send.c rds_message_put(rm); rm 820 net/rds/send.c struct rds_message *rm, __be16 sport, rm 829 net/rds/send.c len = be32_to_cpu(rm->m_inc.i_hdr.h_len); rm 852 net/rds/send.c set_bit(RDS_MSG_ACK_REQUIRED, &rm->m_flags); rm 854 net/rds/send.c list_add_tail(&rm->m_sock_item, &rs->rs_send_queue); rm 855 net/rds/send.c set_bit(RDS_MSG_ON_SOCK, &rm->m_flags); rm 856 net/rds/send.c rds_message_addref(rm); rm 858 net/rds/send.c rm->m_rs = rs; rm 862 net/rds/send.c rds_message_populate_header(&rm->m_inc.i_hdr, sport, dport, 0); rm 863 net/rds/send.c rm->m_inc.i_conn = conn; rm 864 net/rds/send.c rm->m_inc.i_conn_path = cp; rm 865 net/rds/send.c rds_message_addref(rm); rm 868 net/rds/send.c rm->m_inc.i_hdr.h_sequence = cpu_to_be64(cp->cp_next_tx_seq++); rm 869 net/rds/send.c list_add_tail(&rm->m_conn_item, &cp->cp_send_queue); rm 870 net/rds/send.c set_bit(RDS_MSG_ON_CONN, &rm->m_flags); rm 874 net/rds/send.c rm, len, rs, rs->rs_snd_bytes, rm 875 net/rds/send.c (unsigned long long)be64_to_cpu(rm->m_inc.i_hdr.h_sequence)); rm 971 net/rds/send.c static int rds_cmsg_zcopy(struct rds_sock *rs, struct rds_message *rm, rm 977 net/rds/send.c !rm->data.op_mmp_znotifier) rm 980 net/rds/send.c rm->data.op_mmp_znotifier->z_cookie = *cookie; rm 984 net/rds/send.c static int rds_cmsg_send(struct rds_sock *rs, struct rds_message *rm, rm 1005 net/rds/send.c ret = rds_cmsg_rdma_args(rs, rm, cmsg, &vct->vec[ind]); rm 1010 net/rds/send.c ret = rds_cmsg_rdma_dest(rs, rm, cmsg); rm 1014 net/rds/send.c ret = rds_cmsg_rdma_map(rs, rm, cmsg); rm 1027 net/rds/send.c ret = rds_cmsg_atomic(rs, rm, cmsg); rm 1031 net/rds/send.c ret = rds_cmsg_zcopy(rs, rm, cmsg); rm 1108 net/rds/send.c struct rds_message *rm = NULL; rm 1269 net/rds/send.c rm = rds_message_alloc(ret, GFP_KERNEL); rm 1270 net/rds/send.c if (!rm) { rm 1277 net/rds/send.c rm->data.op_sg = rds_message_alloc_sgs(rm, num_sgs); rm 1278 net/rds/send.c if (IS_ERR(rm->data.op_sg)) { rm 1279 net/rds/send.c ret = PTR_ERR(rm->data.op_sg); rm 1282 net/rds/send.c ret = rds_message_copy_from_user(rm, &msg->msg_iter, zcopy); rm 1286 net/rds/send.c rm->data.op_active = 1; rm 1288 net/rds/send.c rm->m_daddr = daddr; rm 1313 net/rds/send.c rm->m_conn_path = cpath; rm 1316 net/rds/send.c ret = rds_cmsg_send(rs, rm, msg, &allocated_mr, &vct); rm 1324 net/rds/send.c if (rm->rdma.op_active && !conn->c_trans->xmit_rdma) { rm 1326 net/rds/send.c &rm->rdma, conn->c_trans->xmit_rdma); rm 1331 net/rds/send.c if (rm->atomic.op_active && !conn->c_trans->xmit_atomic) { rm 1333 net/rds/send.c &rm->atomic, conn->c_trans->xmit_atomic); rm 1350 net/rds/send.c while (!rds_send_queue_rm(rs, conn, cpath, rm, rs->rs_bound_port, rm 1360 net/rds/send.c rds_send_queue_rm(rs, conn, cpath, rm, rm 1393 net/rds/send.c rds_message_put(rm); rm 1410 net/rds/send.c rds_rdma_unuse(rs, rds_rdma_cookie_key(rm->m_rdma_cookie), 1); rm 1412 net/rds/send.c if (rm) rm 1413 net/rds/send.c rds_message_put(rm); rm 1429 net/rds/send.c struct rds_message *rm; rm 1433 net/rds/send.c rm = rds_message_alloc(0, GFP_ATOMIC); rm 1434 net/rds/send.c if (!rm) { rm 1439 net/rds/send.c rm->m_daddr = cp->cp_conn->c_faddr; rm 1440 net/rds/send.c rm->data.op_active = 1; rm 1449 net/rds/send.c list_add_tail(&rm->m_conn_item, &cp->cp_send_queue); rm 1450 net/rds/send.c set_bit(RDS_MSG_ON_CONN, &rm->m_flags); rm 1451 net/rds/send.c rds_message_addref(rm); rm 1452 net/rds/send.c rm->m_inc.i_conn = cp->cp_conn; rm 1453 net/rds/send.c rm->m_inc.i_conn_path = cp; rm 1455 net/rds/send.c rds_message_populate_header(&rm->m_inc.i_hdr, sport, dport, rm 1457 net/rds/send.c rm->m_inc.i_hdr.h_flags |= h_flags; rm 1465 net/rds/send.c rds_message_add_extension(&rm->m_inc.i_hdr, rm 1468 net/rds/send.c rds_message_add_extension(&rm->m_inc.i_hdr, rm 1484 net/rds/send.c rds_message_put(rm); rm 1488 net/rds/send.c if (rm) rm 1489 net/rds/send.c rds_message_put(rm); rm 89 net/rds/tcp.h int rds_tcp_xmit(struct rds_connection *conn, struct rds_message *rm, rm 75 net/rds/tcp_send.c int rds_tcp_xmit(struct rds_connection *conn, struct rds_message *rm, rm 78 net/rds/tcp_send.c struct rds_conn_path *cp = rm->m_inc.i_conn_path; rm 90 net/rds/tcp_send.c rm->m_ack_seq = tc->t_last_sent_nxt + rm 92 net/rds/tcp_send.c be32_to_cpu(rm->m_inc.i_hdr.h_len) - 1; rm 94 net/rds/tcp_send.c set_bit(RDS_MSG_HAS_ACK_SEQ, &rm->m_flags); rm 95 net/rds/tcp_send.c tc->t_last_expected_una = rm->m_ack_seq + 1; rm 97 net/rds/tcp_send.c if (test_bit(RDS_MSG_RETRANSMITTED, &rm->m_flags)) rm 98 net/rds/tcp_send.c rm->m_inc.i_hdr.h_flags |= RDS_FLAG_RETRANSMITTED; rm 101 net/rds/tcp_send.c rm, rds_tcp_write_seq(tc), rm 102 net/rds/tcp_send.c (unsigned long long)rm->m_ack_seq); rm 110 net/rds/tcp_send.c (void *)&rm->m_inc.i_hdr + hdr_off, rm 111 net/rds/tcp_send.c sizeof(rm->m_inc.i_hdr) - hdr_off); rm 119 net/rds/tcp_send.c more = rm->data.op_nents > 1 ? (MSG_MORE | MSG_SENDPAGE_NOTLAST) : 0; rm 120 net/rds/tcp_send.c while (sg < rm->data.op_nents) { rm 124 net/rds/tcp_send.c sg_page(&rm->data.op_sg[sg]), rm 125 net/rds/tcp_send.c rm->data.op_sg[sg].offset + off, rm 126 net/rds/tcp_send.c rm->data.op_sg[sg].length - off, rm 128 net/rds/tcp_send.c rdsdebug("tcp sendpage %p:%u:%u ret %d\n", (void *)sg_page(&rm->data.op_sg[sg]), rm 129 net/rds/tcp_send.c rm->data.op_sg[sg].offset + off, rm->data.op_sg[sg].length - off, rm 136 net/rds/tcp_send.c if (off == rm->data.op_sg[sg].length) { rm 140 net/rds/tcp_send.c if (sg == rm->data.op_nents - 1) rm 176 net/rds/tcp_send.c static int rds_tcp_is_acked(struct rds_message *rm, uint64_t ack) rm 178 net/rds/tcp_send.c if (!test_bit(RDS_MSG_HAS_ACK_SEQ, &rm->m_flags)) rm 180 net/rds/tcp_send.c return (__s32)((u32)rm->m_ack_seq - (u32)ack) < 0; rm 813 net/sunrpc/xprtsock.c static int xs_sendpages(struct socket *sock, struct sockaddr *addr, int addrlen, struct xdr_buf *xdr, unsigned int base, rpc_fraghdr rm, int *sent_p) rm 820 net/sunrpc/xprtsock.c unsigned int rmsize = rm ? sizeof(rm) : 0; rm 835 net/sunrpc/xprtsock.c err = xs_send_rm_and_kvec(sock, &msg, rm, rm 960 net/sunrpc/xprtsock.c rpc_fraghdr rm = xs_stream_record_marker(xdr); rm 961 net/sunrpc/xprtsock.c unsigned int msglen = rm ? req->rq_slen + sizeof(rm) : req->rq_slen; rm 976 net/sunrpc/xprtsock.c transport->xmit.offset, rm, &sent); rm 1107 net/sunrpc/xprtsock.c rpc_fraghdr rm = xs_stream_record_marker(xdr); rm 1108 net/sunrpc/xprtsock.c unsigned int msglen = rm ? req->rq_slen + sizeof(rm) : req->rq_slen; rm 1134 net/sunrpc/xprtsock.c transport->xmit.offset, rm, &sent); rm 588 net/tipc/group.c struct tipc_member *m, *rm, *pm; rm 614 net/tipc/group.c rm = list_first_entry(active, struct tipc_member, list); rm 615 net/tipc/group.c rm->state = MBR_RECLAIMING; rm 616 net/tipc/group.c list_del_init(&rm->list); rm 617 net/tipc/group.c tipc_group_proto_xmit(grp, rm, GRP_RECLAIM_MSG, xmitq); rm 255 net/wireless/core.h struct cfg80211_roam_info rm; rm 987 net/wireless/sme.c ev->rm.req_ie = next; rm 988 net/wireless/sme.c ev->rm.req_ie_len = info->req_ie_len; rm 989 net/wireless/sme.c memcpy((void *)ev->rm.req_ie, info->req_ie, info->req_ie_len); rm 993 net/wireless/sme.c ev->rm.resp_ie = next; rm 994 net/wireless/sme.c ev->rm.resp_ie_len = info->resp_ie_len; rm 995 net/wireless/sme.c memcpy((void *)ev->rm.resp_ie, info->resp_ie, rm 1000 net/wireless/sme.c ev->rm.fils.kek = next; rm 1001 net/wireless/sme.c ev->rm.fils.kek_len = info->fils.kek_len; rm 1002 net/wireless/sme.c memcpy((void *)ev->rm.fils.kek, info->fils.kek, rm 1007 net/wireless/sme.c ev->rm.fils.pmk = next; rm 1008 net/wireless/sme.c ev->rm.fils.pmk_len = info->fils.pmk_len; rm 1009 net/wireless/sme.c memcpy((void *)ev->rm.fils.pmk, info->fils.pmk, rm 1014 net/wireless/sme.c ev->rm.fils.pmkid = next; rm 1015 net/wireless/sme.c memcpy((void *)ev->rm.fils.pmkid, info->fils.pmkid, rm 1019 net/wireless/sme.c ev->rm.fils.update_erp_next_seq_num = info->fils.update_erp_next_seq_num; rm 1021 net/wireless/sme.c ev->rm.fils.erp_next_seq_num = info->fils.erp_next_seq_num; rm 1022 net/wireless/sme.c ev->rm.bss = info->bss; rm 876 net/wireless/util.c __cfg80211_roamed(wdev, &ev->rm); rm 1051 net/xfrm/xfrm_policy.c static void xfrm_policy_inexact_gc_tree(struct rb_root *r, bool rm) rm 1059 net/xfrm/xfrm_policy.c xfrm_policy_inexact_gc_tree(&node->root, rm); rm 1063 net/xfrm/xfrm_policy.c WARN_ON_ONCE(rm); rm 298 sound/mips/ad1843.c int lg, rg, lm, rm; rm 308 sound/mips/ad1843.c ad1843_read_multi(ad1843, 2, gp->lmute, &lm, gp->rmute, &rm); rm 311 sound/mips/ad1843.c if (rm) rm 331 sound/mips/ad1843.c int rm = (rg == 0) ? 1 : 0; rm 338 sound/mips/ad1843.c ad1843_write_multi(ad1843, 2, gp->lmute, lm, gp->rmute, rm); rm 5083 sound/soc/codecs/wcd9335.c struct regmap *rm = wcd->regmap; rm 5086 sound/soc/codecs/wcd9335.c regmap_read(rm, WCD9335_CHIP_TIER_CTRL_EFUSE_VAL_OUT0, &val); rm 5087 sound/soc/codecs/wcd9335.c regmap_read(rm, WCD9335_CHIP_TIER_CTRL_CHIP_ID_BYTE0, &byte0); rm 5097 sound/soc/codecs/wcd9335.c regmap_write(rm, WCD9335_CODEC_RPM_RST_CTL, 0x01); rm 5098 sound/soc/codecs/wcd9335.c regmap_write(rm, WCD9335_SIDO_SIDO_TEST_2, 0x00); rm 5099 sound/soc/codecs/wcd9335.c regmap_write(rm, WCD9335_SIDO_SIDO_CCL_8, 0x6F); rm 5100 sound/soc/codecs/wcd9335.c regmap_write(rm, WCD9335_BIAS_VBG_FINE_ADJ, 0x65); rm 5101 sound/soc/codecs/wcd9335.c regmap_write(rm, WCD9335_CODEC_RPM_PWR_CDC_DIG_HM_CTL, 0x5); rm 5102 sound/soc/codecs/wcd9335.c regmap_write(rm, WCD9335_CODEC_RPM_PWR_CDC_DIG_HM_CTL, 0x7); rm 5103 sound/soc/codecs/wcd9335.c regmap_write(rm, WCD9335_CODEC_RPM_PWR_CDC_DIG_HM_CTL, 0x3); rm 5104 sound/soc/codecs/wcd9335.c regmap_write(rm, WCD9335_CODEC_RPM_RST_CTL, 0x3); rm 346 tools/arch/x86/lib/insn.c insn_byte_t mod, rm, base; rm 371 tools/arch/x86/lib/insn.c rm = X86_MODRM_RM(insn->modrm.value); rm 379 tools/arch/x86/lib/insn.c if ((mod == 0 && rm == 6) || mod == 2) { rm 385 tools/arch/x86/lib/insn.c if ((mod == 0 && rm == 5) || mod == 2 || rm 116 tools/perf/arch/powerpc/util/perf_regs.c regmatch_t rm[5]; rm 129 tools/perf/arch/powerpc/util/perf_regs.c if (!regexec(&sdt_op_regex1, old_op, 3, rm, 0)) { rm 133 tools/perf/arch/powerpc/util/perf_regs.c new_len += (int)(rm[2].rm_eo - rm[2].rm_so); rm 140 tools/perf/arch/powerpc/util/perf_regs.c (int)(rm[2].rm_eo - rm[2].rm_so), old_op + rm[2].rm_so); rm 141 tools/perf/arch/powerpc/util/perf_regs.c } else if (!regexec(&sdt_op_regex2, old_op, 5, rm, 0)) { rm 146 tools/perf/arch/powerpc/util/perf_regs.c prefix = (rm[1].rm_so == -1) ? '+' : '-'; rm 149 tools/perf/arch/powerpc/util/perf_regs.c new_len += (int)(rm[2].rm_eo - rm[2].rm_so); rm 150 tools/perf/arch/powerpc/util/perf_regs.c new_len += (int)(rm[4].rm_eo - rm[4].rm_so); rm 157 tools/perf/arch/powerpc/util/perf_regs.c (int)(rm[2].rm_eo - rm[2].rm_so), old_op + rm[2].rm_so, rm 158 tools/perf/arch/powerpc/util/perf_regs.c (int)(rm[4].rm_eo - rm[4].rm_so), old_op + rm[4].rm_so); rm 204 tools/perf/arch/x86/util/perf_regs.c regmatch_t rm[6]; rm 223 tools/perf/arch/x86/util/perf_regs.c regexec(&sdt_op_regex, old_op, 6, rm, 0) || rm 224 tools/perf/arch/x86/util/perf_regs.c rm[4].rm_eo - rm[4].rm_so > SDT_REG_NAME_SIZE) { rm 241 tools/perf/arch/x86/util/perf_regs.c if (rm[3].rm_so != rm[3].rm_eo) { rm 242 tools/perf/arch/x86/util/perf_regs.c if (rm[1].rm_so != rm[1].rm_eo) rm 243 tools/perf/arch/x86/util/perf_regs.c prefix[0] = *(old_op + rm[1].rm_so); rm 244 tools/perf/arch/x86/util/perf_regs.c else if (rm[2].rm_so != rm[2].rm_eo) rm 251 tools/perf/arch/x86/util/perf_regs.c sdt_rename_register(old_op + rm[4].rm_so, rm[4].rm_eo - rm[4].rm_so, rm 256 tools/perf/arch/x86/util/perf_regs.c (rm[2].rm_eo - rm[2].rm_so) + rm 257 tools/perf/arch/x86/util/perf_regs.c (rm[3].rm_eo - rm[3].rm_so) + rm 259 tools/perf/arch/x86/util/perf_regs.c (rm[5].rm_eo - rm[5].rm_so) + rm 268 tools/perf/arch/x86/util/perf_regs.c (int)(rm[2].rm_eo - rm[2].rm_so), old_op + rm[2].rm_so, rm 269 tools/perf/arch/x86/util/perf_regs.c (int)(rm[3].rm_eo - rm[3].rm_so), old_op + rm[3].rm_so, rm 271 tools/perf/arch/x86/util/perf_regs.c (int)(rm[5].rm_eo - rm[5].rm_so), old_op + rm[5].rm_so);