rs 50 arch/arm/mach-omap1/reset.c u16 rs; rs 52 arch/arm/mach-omap1/reset.c rs = __raw_readw(OMAP1_IO_ADDRESS(ARM_SYSST)); rs 54 arch/arm/mach-omap1/reset.c if (rs & (1 << ARM_SYSST_POR_SHIFT)) rs 56 arch/arm/mach-omap1/reset.c if (rs & (1 << ARM_SYSST_EXT_RST_SHIFT)) rs 58 arch/arm/mach-omap1/reset.c if (rs & (1 << ARM_SYSST_ARM_WDRST_SHIFT)) rs 60 arch/arm/mach-omap1/reset.c if (rs & (1 << ARM_SYSST_GLOB_SWRST_SHIFT)) rs 722 arch/arm/net/bpf_jit_32.c const s8 *rs; rs 724 arch/arm/net/bpf_jit_32.c rs = arm_bpf_get_reg64(src, tmp2, ctx); rs 727 arch/arm/net/bpf_jit_32.c emit_alu_r(rd[1], rs[1], true, false, op, ctx); rs 728 arch/arm/net/bpf_jit_32.c emit_alu_r(rd[0], rs[0], true, true, op, ctx); rs 730 arch/arm/net/bpf_jit_32.c s8 rs; rs 732 arch/arm/net/bpf_jit_32.c rs = arm_bpf_get_reg32(src_lo, tmp2[1], ctx); rs 735 arch/arm/net/bpf_jit_32.c emit_alu_r(rd[1], rs, true, false, op, ctx); rs 1356 arch/arm/net/bpf_jit_32.c const s8 *rd, *rs; rs 1633 arch/arm/net/bpf_jit_32.c rs = arm_bpf_get_reg64(src, tmp2, ctx); rs 1634 arch/arm/net/bpf_jit_32.c emit_str_r(dst_lo, rs, off, ctx, BPF_SIZE(code)); rs 217 arch/arm/net/bpf_jit_32.h #define ARM_MOV_SR(rd, rm, type, rs) \ rs 218 arch/arm/net/bpf_jit_32.h (_AL3_SR(ARM_MOV_R(rd, rm)) | (type) << 5 | (rs) << 8) rs 235 arch/arm/net/bpf_jit_32.h #define ARM_ORR_SR(rd, rn, rm, type, rs) \ rs 236 arch/arm/net/bpf_jit_32.h (_AL3_SR(ARM_ORR_R(rd, rn, rm)) | (type) << 5 | (rs) << 8) rs 238 arch/arm/net/bpf_jit_32.h #define ARM_ORRS_SR(rd, rn, rm, type, rs) \ rs 239 arch/arm/net/bpf_jit_32.h (_AL3_SR(ARM_ORRS_R(rd, rn, rm)) | (type) << 5 | (rs) << 8) rs 162 arch/arm/probes/kprobes/actions-arm.c int rs = (insn >> 8) & 0xf; rs 169 arch/arm/probes/kprobes/actions-arm.c register unsigned long rsv asm("r1") = regs->uregs[rs]; rs 224 arch/arm/probes/kprobes/actions-arm.c int rs = (insn >> 8) & 0xf; rs 229 arch/arm/probes/kprobes/actions-arm.c register unsigned long rsv asm("r1") = regs->uregs[rs]; rs 210 arch/arm64/kernel/traps.c static DEFINE_RATELIMIT_STATE(rs, DEFAULT_RATELIMIT_INTERVAL, rs 219 arch/arm64/kernel/traps.c !__ratelimit(&rs)) rs 442 arch/ia64/include/asm/pal.h rs : 1, /* RSE valid */ rs 698 arch/ia64/include/asm/pal.h #define pmci_proc_rse_valid pme_processor.rs rs 134 arch/ia64/include/asm/processor.h __u64 rs : 1; rs 618 arch/ia64/kernel/unwind.c struct unw_reg_state *rs; rs 620 arch/ia64/kernel/unwind.c rs = alloc_reg_state(); rs 621 arch/ia64/kernel/unwind.c if (!rs) { rs 625 arch/ia64/kernel/unwind.c memcpy(rs, &sr->curr, sizeof(*rs)); rs 626 arch/ia64/kernel/unwind.c sr->curr.next = rs; rs 632 arch/ia64/kernel/unwind.c struct unw_reg_state *rs = sr->curr.next; rs 634 arch/ia64/kernel/unwind.c if (!rs) { rs 638 arch/ia64/kernel/unwind.c memcpy(&sr->curr, rs, sizeof(*rs)); rs 639 arch/ia64/kernel/unwind.c free_reg_state(rs); rs 644 arch/ia64/kernel/unwind.c dup_state_stack (struct unw_reg_state *rs) rs 648 arch/ia64/kernel/unwind.c while (rs) { rs 654 arch/ia64/kernel/unwind.c memcpy(copy, rs, sizeof(*copy)); rs 659 arch/ia64/kernel/unwind.c rs = rs->next; rs 667 arch/ia64/kernel/unwind.c free_state_stack (struct unw_reg_state *rs) rs 671 arch/ia64/kernel/unwind.c for (p = rs->next; p != NULL; p = next) { rs 675 arch/ia64/kernel/unwind.c rs->next = NULL; rs 557 arch/ia64/mm/discontig.c unsigned long rs, re, end = start + len; rs 576 arch/ia64/mm/discontig.c rs = max(start, node_memblk[i].start_paddr); rs 580 arch/ia64/mm/discontig.c if (rs < re) rs 581 arch/ia64/mm/discontig.c (*func)(rs, re - rs, node_memblk[i].nid); rs 186 arch/microblaze/include/asm/page.h #define tophys(rd, rs) addik rd, rs, 0 rs 187 arch/microblaze/include/asm/page.h #define tovirt(rd, rs) addik rd, rs, 0 rs 193 arch/microblaze/include/asm/page.h #define tophys(rd, rs) \ rs 194 arch/microblaze/include/asm/page.h addik rd, rs, (CONFIG_KERNEL_BASE_ADDR - CONFIG_KERNEL_START) rs 195 arch/microblaze/include/asm/page.h #define tovirt(rd, rs) \ rs 196 arch/microblaze/include/asm/page.h addik rd, rs, (CONFIG_KERNEL_START - CONFIG_KERNEL_BASE_ADDR) rs 32 arch/mips/crypto/crc32-mips.c _ASM_MACRO_3R(OP, rt, rs, rt2, \ rs 101 arch/mips/dec/kn01-berr.c vaddr = regs->regs[insn.i_format.rs] + rs 204 arch/mips/include/asm/asmmacro.h .macro _EXT rd, rs, p, s rs 208 arch/mips/include/asm/asmmacro.h .macro _EXT rd, rs, p, s rs 251 arch/mips/include/asm/asmmacro.h .macro _ctcmsa cd, rs rs 382 arch/mips/include/asm/asmmacro.h .macro _ctcmsa cd, rs rs 16 arch/mips/include/asm/ginvt.h _ASM_MACRO_1R1I(ginvt, rs, type, rs 1544 arch/mips/include/asm/mipsregs.h _ASM_MACRO_2R_1S(mfhc0, rt, rs, sel, rs 1979 arch/mips/include/asm/mipsregs.h _ASM_MACRO_2R_1S(mfgc0, rt, rs, sel, rs 1982 arch/mips/include/asm/mipsregs.h _ASM_MACRO_2R_1S(dmfgc0, rt, rs, sel, rs 168 arch/mips/include/asm/msa.h _ASM_MACRO_2R(ctcmsa, cd, rs, rs 130 arch/mips/include/asm/octeon/cvmx-asm.h asm ("pop %[rd],%[rs]" : [rd] "=d" (result) : [rs] "d" (input)) rs 132 arch/mips/include/asm/octeon/cvmx-asm.h asm ("dpop %[rd],%[rs]" : [rd] "=d" (result) : [rs] "d" (input)) rs 274 arch/mips/include/asm/octeon/cvmx-pip.h uint64_t rs:1; rs 351 arch/mips/include/asm/octeon/cvmx-sriox-defs.h uint64_t rs:1; rs 357 arch/mips/include/asm/octeon/cvmx-sriox-defs.h uint64_t rs:1; rs 200 arch/mips/include/asm/uasm.h void UASM_i_LA_mostly(u32 **buf, unsigned int rs, long addr); rs 201 arch/mips/include/asm/uasm.h void UASM_i_LA(u32 **buf, unsigned int rs, long addr); rs 211 arch/mips/include/asm/uasm.h # define UASM_i_ADDIU(buf, rs, rt, val) uasm_i_daddiu(buf, rs, rt, val) rs 212 arch/mips/include/asm/uasm.h # define UASM_i_ADDU(buf, rs, rt, rd) uasm_i_daddu(buf, rs, rt, rd) rs 213 arch/mips/include/asm/uasm.h # define UASM_i_LL(buf, rs, rt, off) uasm_i_lld(buf, rs, rt, off) rs 214 arch/mips/include/asm/uasm.h # define UASM_i_LW(buf, rs, rt, off) uasm_i_ld(buf, rs, rt, off) rs 215 arch/mips/include/asm/uasm.h # define UASM_i_LWX(buf, rs, rt, rd) uasm_i_ldx(buf, rs, rt, rd) rs 218 arch/mips/include/asm/uasm.h # define UASM_i_ROTR(buf, rs, rt, sh) uasm_i_drotr(buf, rs, rt, sh) rs 219 arch/mips/include/asm/uasm.h # define UASM_i_SC(buf, rs, rt, off) uasm_i_scd(buf, rs, rt, off) rs 220 arch/mips/include/asm/uasm.h # define UASM_i_SLL(buf, rs, rt, sh) uasm_i_dsll(buf, rs, rt, sh) rs 221 arch/mips/include/asm/uasm.h # define UASM_i_SRA(buf, rs, rt, sh) uasm_i_dsra(buf, rs, rt, sh) rs 222 arch/mips/include/asm/uasm.h # define UASM_i_SRL(buf, rs, rt, sh) uasm_i_dsrl(buf, rs, rt, sh) rs 223 arch/mips/include/asm/uasm.h # define UASM_i_SRL_SAFE(buf, rs, rt, sh) uasm_i_dsrl_safe(buf, rs, rt, sh) rs 224 arch/mips/include/asm/uasm.h # define UASM_i_SUBU(buf, rs, rt, rd) uasm_i_dsubu(buf, rs, rt, rd) rs 225 arch/mips/include/asm/uasm.h # define UASM_i_SW(buf, rs, rt, off) uasm_i_sd(buf, rs, rt, off) rs 227 arch/mips/include/asm/uasm.h # define UASM_i_ADDIU(buf, rs, rt, val) uasm_i_addiu(buf, rs, rt, val) rs 228 arch/mips/include/asm/uasm.h # define UASM_i_ADDU(buf, rs, rt, rd) uasm_i_addu(buf, rs, rt, rd) rs 229 arch/mips/include/asm/uasm.h # define UASM_i_LL(buf, rs, rt, off) uasm_i_ll(buf, rs, rt, off) rs 230 arch/mips/include/asm/uasm.h # define UASM_i_LW(buf, rs, rt, off) uasm_i_lw(buf, rs, rt, off) rs 231 arch/mips/include/asm/uasm.h # define UASM_i_LWX(buf, rs, rt, rd) uasm_i_lwx(buf, rs, rt, rd) rs 234 arch/mips/include/asm/uasm.h # define UASM_i_ROTR(buf, rs, rt, sh) uasm_i_rotr(buf, rs, rt, sh) rs 235 arch/mips/include/asm/uasm.h # define UASM_i_SC(buf, rs, rt, off) uasm_i_sc(buf, rs, rt, off) rs 236 arch/mips/include/asm/uasm.h # define UASM_i_SLL(buf, rs, rt, sh) uasm_i_sll(buf, rs, rt, sh) rs 237 arch/mips/include/asm/uasm.h # define UASM_i_SRA(buf, rs, rt, sh) uasm_i_sra(buf, rs, rt, sh) rs 238 arch/mips/include/asm/uasm.h # define UASM_i_SRL(buf, rs, rt, sh) uasm_i_srl(buf, rs, rt, sh) rs 239 arch/mips/include/asm/uasm.h # define UASM_i_SRL_SAFE(buf, rs, rt, sh) uasm_i_srl(buf, rs, rt, sh) rs 240 arch/mips/include/asm/uasm.h # define UASM_i_SUBU(buf, rs, rt, rd) uasm_i_subu(buf, rs, rt, rd) rs 241 arch/mips/include/asm/uasm.h # define UASM_i_SW(buf, rs, rt, off) uasm_i_sw(buf, rs, rt, off) rs 245 arch/mips/include/asm/uasm.h #define uasm_i_beqz(buf, rs, off) uasm_i_beq(buf, rs, 0, off) rs 246 arch/mips/include/asm/uasm.h #define uasm_i_beqzl(buf, rs, off) uasm_i_beql(buf, rs, 0, off) rs 247 arch/mips/include/asm/uasm.h #define uasm_i_bnez(buf, rs, off) uasm_i_bne(buf, rs, 0, off) rs 248 arch/mips/include/asm/uasm.h #define uasm_i_bnezl(buf, rs, off) uasm_i_bnel(buf, rs, 0, off) rs 634 arch/mips/include/uapi/asm/inst.h __BITFIELD_FIELD(unsigned int rs : 5, rs 642 arch/mips/include/uapi/asm/inst.h __BITFIELD_FIELD(unsigned int rs : 5, rs 650 arch/mips/include/uapi/asm/inst.h __BITFIELD_FIELD(unsigned int rs : 5, rs 659 arch/mips/include/uapi/asm/inst.h __BITFIELD_FIELD(unsigned int rs : 5, rs 669 arch/mips/include/uapi/asm/inst.h __BITFIELD_FIELD(unsigned int rs : 5, rs 679 arch/mips/include/uapi/asm/inst.h __BITFIELD_FIELD(unsigned int rs : 5, rs 699 arch/mips/include/uapi/asm/inst.h __BITFIELD_FIELD(unsigned int rs : 5, rs 738 arch/mips/include/uapi/asm/inst.h __BITFIELD_FIELD(unsigned int rs : 5, rs 760 arch/mips/include/uapi/asm/inst.h __BITFIELD_FIELD(unsigned int rs : 5, rs 779 arch/mips/include/uapi/asm/inst.h __BITFIELD_FIELD(unsigned int rs:5, rs 909 arch/mips/include/uapi/asm/inst.h __BITFIELD_FIELD(unsigned int rs : 5, rs 934 arch/mips/include/uapi/asm/inst.h __BITFIELD_FIELD(unsigned int rs : 3, rs 951 arch/mips/include/uapi/asm/inst.h __BITFIELD_FIELD(unsigned int rs : 3, rs 81 arch/mips/kernel/branch.c *contpc = regs->regs[insn.mm_i_format.rs]; rs 95 arch/mips/kernel/branch.c if ((long)regs->regs[insn.mm_i_format.rs] < 0) rs 111 arch/mips/kernel/branch.c if ((long)regs->regs[insn.mm_i_format.rs] >= 0) rs 121 arch/mips/kernel/branch.c if ((long)regs->regs[insn.mm_i_format.rs] <= 0) rs 131 arch/mips/kernel/branch.c if ((long)regs->regs[insn.mm_i_format.rs] <= 0) rs 160 arch/mips/kernel/branch.c bit = (insn.mm_i_format.rs >> 2); rs 183 arch/mips/kernel/branch.c *contpc = regs->regs[insn.mm_i_format.rs]; rs 188 arch/mips/kernel/branch.c if ((long)regs->regs[reg16to32map[insn.mm_b1_format.rs]] == 0) rs 197 arch/mips/kernel/branch.c if ((long)regs->regs[reg16to32map[insn.mm_b1_format.rs]] != 0) rs 210 arch/mips/kernel/branch.c if (regs->regs[insn.mm_i_format.rs] == rs 221 arch/mips/kernel/branch.c if (regs->regs[insn.mm_i_format.rs] != rs 439 arch/mips/kernel/branch.c regs->cp0_epc = regs->regs[insn.r_format.rs]; rs 456 arch/mips/kernel/branch.c if ((long)regs->regs[insn.i_format.rs] < 0) { rs 470 arch/mips/kernel/branch.c if ((long)regs->regs[insn.i_format.rs] >= 0) { rs 481 arch/mips/kernel/branch.c if (NO_R6EMU && (insn.i_format.rs || rs 491 arch/mips/kernel/branch.c if (!insn.i_format.rs) { rs 502 arch/mips/kernel/branch.c if ((long)regs->regs[insn.i_format.rs] < 0) { rs 513 arch/mips/kernel/branch.c if (NO_R6EMU && (insn.i_format.rs || rs 523 arch/mips/kernel/branch.c if (!insn.i_format.rs) { rs 534 arch/mips/kernel/branch.c if ((long)regs->regs[insn.i_format.rs] >= 0) { rs 583 arch/mips/kernel/branch.c if (regs->regs[insn.i_format.rs] == rs 598 arch/mips/kernel/branch.c if (regs->regs[insn.i_format.rs] != rs 628 arch/mips/kernel/branch.c ((!insn.i_format.rs && insn.i_format.rt) || rs 629 arch/mips/kernel/branch.c (insn.i_format.rs == insn.i_format.rt))) rs 635 arch/mips/kernel/branch.c if ((long)regs->regs[insn.i_format.rs] <= 0) { rs 664 arch/mips/kernel/branch.c ((!insn.i_format.rs && insn.i_format.rt) || rs 665 arch/mips/kernel/branch.c (insn.i_format.rs == insn.i_format.rt))) rs 672 arch/mips/kernel/branch.c if ((long)regs->regs[insn.i_format.rs] > 0) { rs 689 arch/mips/kernel/branch.c ((insn.i_format.rs == bc1eqz_op) || rs 690 arch/mips/kernel/branch.c (insn.i_format.rs == bc1nez_op))) { rs 695 arch/mips/kernel/branch.c if (insn.i_format.rs == bc1eqz_op) rs 750 arch/mips/kernel/branch.c if ((regs->regs[insn.i_format.rs] & (1ull<<insn.i_format.rt)) rs 758 arch/mips/kernel/branch.c if ((regs->regs[insn.i_format.rs] & rs 766 arch/mips/kernel/branch.c if (regs->regs[insn.i_format.rs] & (1ull<<insn.i_format.rt)) rs 773 arch/mips/kernel/branch.c if (regs->regs[insn.i_format.rs] & rs 805 arch/mips/kernel/branch.c if (!insn.i_format.rs) { rs 821 arch/mips/kernel/branch.c if (insn.i_format.rt && !insn.i_format.rs) rs 405 arch/mips/kernel/mips-r2-to-r6-emul.c s32 rt, rs; rs 408 arch/mips/kernel/mips-r2-to-r6-emul.c rs = regs->regs[MIPSInst_RS(ir)]; rs 409 arch/mips/kernel/mips-r2-to-r6-emul.c res = (s64)rt * (s64)rs; rs 411 arch/mips/kernel/mips-r2-to-r6-emul.c rs = res; rs 412 arch/mips/kernel/mips-r2-to-r6-emul.c regs->lo = (s64)rs; rs 432 arch/mips/kernel/mips-r2-to-r6-emul.c u32 rt, rs; rs 435 arch/mips/kernel/mips-r2-to-r6-emul.c rs = regs->regs[MIPSInst_RS(ir)]; rs 436 arch/mips/kernel/mips-r2-to-r6-emul.c res = (u64)rt * (u64)rs; rs 455 arch/mips/kernel/mips-r2-to-r6-emul.c s32 rt, rs; rs 458 arch/mips/kernel/mips-r2-to-r6-emul.c rs = regs->regs[MIPSInst_RS(ir)]; rs 460 arch/mips/kernel/mips-r2-to-r6-emul.c regs->lo = (s64)(rs / rt); rs 461 arch/mips/kernel/mips-r2-to-r6-emul.c regs->hi = (s64)(rs % rt); rs 477 arch/mips/kernel/mips-r2-to-r6-emul.c u32 rt, rs; rs 480 arch/mips/kernel/mips-r2-to-r6-emul.c rs = regs->regs[MIPSInst_RS(ir)]; rs 482 arch/mips/kernel/mips-r2-to-r6-emul.c regs->lo = (s64)(rs / rt); rs 483 arch/mips/kernel/mips-r2-to-r6-emul.c regs->hi = (s64)(rs % rt); rs 500 arch/mips/kernel/mips-r2-to-r6-emul.c s64 rt, rs; rs 506 arch/mips/kernel/mips-r2-to-r6-emul.c rs = regs->regs[MIPSInst_RS(ir)]; rs 507 arch/mips/kernel/mips-r2-to-r6-emul.c res = rt * rs; rs 513 arch/mips/kernel/mips-r2-to-r6-emul.c : "r"(rt), "r"(rs)); rs 532 arch/mips/kernel/mips-r2-to-r6-emul.c u64 rt, rs; rs 538 arch/mips/kernel/mips-r2-to-r6-emul.c rs = regs->regs[MIPSInst_RS(ir)]; rs 539 arch/mips/kernel/mips-r2-to-r6-emul.c res = rt * rs; rs 545 arch/mips/kernel/mips-r2-to-r6-emul.c : "r"(rt), "r"(rs)); rs 563 arch/mips/kernel/mips-r2-to-r6-emul.c s64 rt, rs; rs 569 arch/mips/kernel/mips-r2-to-r6-emul.c rs = regs->regs[MIPSInst_RS(ir)]; rs 571 arch/mips/kernel/mips-r2-to-r6-emul.c regs->lo = rs / rt; rs 572 arch/mips/kernel/mips-r2-to-r6-emul.c regs->hi = rs % rt; rs 588 arch/mips/kernel/mips-r2-to-r6-emul.c u64 rt, rs; rs 594 arch/mips/kernel/mips-r2-to-r6-emul.c rs = regs->regs[MIPSInst_RS(ir)]; rs 596 arch/mips/kernel/mips-r2-to-r6-emul.c regs->lo = rs / rt; rs 597 arch/mips/kernel/mips-r2-to-r6-emul.c regs->hi = rs % rt; rs 636 arch/mips/kernel/mips-r2-to-r6-emul.c s32 rt, rs; rs 639 arch/mips/kernel/mips-r2-to-r6-emul.c rs = regs->regs[MIPSInst_RS(ir)]; rs 640 arch/mips/kernel/mips-r2-to-r6-emul.c res = (s64)rt * (s64)rs; rs 642 arch/mips/kernel/mips-r2-to-r6-emul.c rs = regs->lo; rs 643 arch/mips/kernel/mips-r2-to-r6-emul.c res += ((((s64)rt) << 32) | (u32)rs); rs 647 arch/mips/kernel/mips-r2-to-r6-emul.c rs = res >> 32; rs 648 arch/mips/kernel/mips-r2-to-r6-emul.c regs->hi = (s64)rs; rs 665 arch/mips/kernel/mips-r2-to-r6-emul.c u32 rt, rs; rs 668 arch/mips/kernel/mips-r2-to-r6-emul.c rs = regs->regs[MIPSInst_RS(ir)]; rs 669 arch/mips/kernel/mips-r2-to-r6-emul.c res = (u64)rt * (u64)rs; rs 671 arch/mips/kernel/mips-r2-to-r6-emul.c rs = regs->lo; rs 672 arch/mips/kernel/mips-r2-to-r6-emul.c res += ((((s64)rt) << 32) | (u32)rs); rs 676 arch/mips/kernel/mips-r2-to-r6-emul.c rs = res >> 32; rs 677 arch/mips/kernel/mips-r2-to-r6-emul.c regs->hi = (s64)(s32)rs; rs 694 arch/mips/kernel/mips-r2-to-r6-emul.c s32 rt, rs; rs 697 arch/mips/kernel/mips-r2-to-r6-emul.c rs = regs->regs[MIPSInst_RS(ir)]; rs 698 arch/mips/kernel/mips-r2-to-r6-emul.c res = (s64)rt * (s64)rs; rs 700 arch/mips/kernel/mips-r2-to-r6-emul.c rs = regs->lo; rs 701 arch/mips/kernel/mips-r2-to-r6-emul.c res = ((((s64)rt) << 32) | (u32)rs) - res; rs 705 arch/mips/kernel/mips-r2-to-r6-emul.c rs = res >> 32; rs 706 arch/mips/kernel/mips-r2-to-r6-emul.c regs->hi = (s64)rs; rs 723 arch/mips/kernel/mips-r2-to-r6-emul.c u32 rt, rs; rs 726 arch/mips/kernel/mips-r2-to-r6-emul.c rs = regs->regs[MIPSInst_RS(ir)]; rs 727 arch/mips/kernel/mips-r2-to-r6-emul.c res = (u64)rt * (u64)rs; rs 729 arch/mips/kernel/mips-r2-to-r6-emul.c rs = regs->lo; rs 730 arch/mips/kernel/mips-r2-to-r6-emul.c res = ((((s64)rt) << 32) | (u32)rs) - res; rs 734 arch/mips/kernel/mips-r2-to-r6-emul.c rs = res >> 32; rs 735 arch/mips/kernel/mips-r2-to-r6-emul.c regs->hi = (s64)(s32)rs; rs 752 arch/mips/kernel/mips-r2-to-r6-emul.c s32 rt, rs; rs 757 arch/mips/kernel/mips-r2-to-r6-emul.c rs = regs->regs[MIPSInst_RS(ir)]; rs 758 arch/mips/kernel/mips-r2-to-r6-emul.c res = (s64)rt * (s64)rs; rs 760 arch/mips/kernel/mips-r2-to-r6-emul.c rs = res; rs 761 arch/mips/kernel/mips-r2-to-r6-emul.c regs->regs[MIPSInst_RD(ir)] = (s64)rs; rs 778 arch/mips/kernel/mips-r2-to-r6-emul.c u32 rs; rs 783 arch/mips/kernel/mips-r2-to-r6-emul.c rs = regs->regs[MIPSInst_RS(ir)]; rs 784 arch/mips/kernel/mips-r2-to-r6-emul.c __asm__ __volatile__("clz %0, %1" : "=r"(res) : "r"(rs)); rs 803 arch/mips/kernel/mips-r2-to-r6-emul.c u32 rs; rs 808 arch/mips/kernel/mips-r2-to-r6-emul.c rs = regs->regs[MIPSInst_RS(ir)]; rs 809 arch/mips/kernel/mips-r2-to-r6-emul.c __asm__ __volatile__("clo %0, %1" : "=r"(res) : "r"(rs)); rs 827 arch/mips/kernel/mips-r2-to-r6-emul.c u64 rs; rs 835 arch/mips/kernel/mips-r2-to-r6-emul.c rs = regs->regs[MIPSInst_RS(ir)]; rs 836 arch/mips/kernel/mips-r2-to-r6-emul.c __asm__ __volatile__("dclz %0, %1" : "=r"(res) : "r"(rs)); rs 854 arch/mips/kernel/mips-r2-to-r6-emul.c u64 rs; rs 862 arch/mips/kernel/mips-r2-to-r6-emul.c rs = regs->regs[MIPSInst_RS(ir)]; rs 863 arch/mips/kernel/mips-r2-to-r6-emul.c __asm__ __volatile__("dclo %0, %1" : "=r"(res) : "r"(rs)); rs 911 arch/mips/kernel/mips-r2-to-r6-emul.c unsigned long cpc, epc, nepc, r31, res, rs, rt; rs 941 arch/mips/kernel/mips-r2-to-r6-emul.c rs = MIPSInst_RS(inst); rs 944 arch/mips/kernel/mips-r2-to-r6-emul.c if ((long)regs->regs[rs] >= MIPSInst_SIMM(inst)) rs 951 arch/mips/kernel/mips-r2-to-r6-emul.c if (regs->regs[rs] >= MIPSInst_UIMM(inst)) rs 958 arch/mips/kernel/mips-r2-to-r6-emul.c if ((long)regs->regs[rs] < MIPSInst_SIMM(inst)) rs 965 arch/mips/kernel/mips-r2-to-r6-emul.c if (regs->regs[rs] < MIPSInst_UIMM(inst)) rs 972 arch/mips/kernel/mips-r2-to-r6-emul.c if (regs->regs[rs] == MIPSInst_SIMM(inst)) rs 979 arch/mips/kernel/mips-r2-to-r6-emul.c if (regs->regs[rs] != MIPSInst_SIMM(inst)) rs 1267 arch/mips/kernel/mips-r2-to-r6-emul.c : "+&r"(rt), "=&r"(rs), rs 1342 arch/mips/kernel/mips-r2-to-r6-emul.c : "+&r"(rt), "=&r"(rs), rs 1413 arch/mips/kernel/mips-r2-to-r6-emul.c : "+&r"(rt), "=&r"(rs), rs 1483 arch/mips/kernel/mips-r2-to-r6-emul.c : "+&r"(rt), "=&r"(rs), rs 1602 arch/mips/kernel/mips-r2-to-r6-emul.c : "+&r"(rt), "=&r"(rs), rs 1721 arch/mips/kernel/mips-r2-to-r6-emul.c : "+&r"(rt), "=&r"(rs), rs 1840 arch/mips/kernel/mips-r2-to-r6-emul.c : "+&r"(rt), "=&r"(rs), rs 1958 arch/mips/kernel/mips-r2-to-r6-emul.c : "+&r"(rt), "=&r"(rs), rs 244 arch/mips/kernel/process.c if (ip->i_format.rs != 29) rs 274 arch/mips/kernel/process.c ip->i_format.rs == 29 && ip->i_format.rt == 31) { rs 353 arch/mips/kernel/process.c ip->mm_i_format.rt == 29 && ip->mm_i_format.rs == 29) { rs 359 arch/mips/kernel/process.c if (ip->i_format.rs != 29 || ip->i_format.rt != 29) rs 87 arch/mips/kvm/dyntrans.c synci_inst.i_format.rs = inst.i_format.rs; rs 69 arch/mips/kvm/emulate.c nextpc = arch->gprs[insn.r_format.rs]; rs 85 arch/mips/kvm/emulate.c if ((long)arch->gprs[insn.i_format.rs] < 0) rs 94 arch/mips/kvm/emulate.c if ((long)arch->gprs[insn.i_format.rs] >= 0) rs 104 arch/mips/kvm/emulate.c if ((long)arch->gprs[insn.i_format.rs] < 0) rs 114 arch/mips/kvm/emulate.c if ((long)arch->gprs[insn.i_format.rs] >= 0) rs 155 arch/mips/kvm/emulate.c if (arch->gprs[insn.i_format.rs] == rs 165 arch/mips/kvm/emulate.c if (arch->gprs[insn.i_format.rs] != rs 179 arch/mips/kvm/emulate.c if ((long)arch->gprs[insn.i_format.rs] <= 0) rs 192 arch/mips/kvm/emulate.c if ((long)arch->gprs[insn.i_format.rs] > 0) rs 215 arch/mips/kvm/emulate.c if (insn.i_format.rs != 0 || insn.i_format.rt != 0) rs 221 arch/mips/kvm/emulate.c if (insn.i_format.rs != 0) rs 1314 arch/mips/kvm/emulate.c switch (inst.c0r_format.rs) { rs 1577 arch/mips/kvm/emulate.c vcpu->arch.pc, inst.c0r_format.rs); rs 1818 arch/mips/kvm/emulate.c base = inst.i_format.rs; rs 2515 arch/mips/kvm/emulate.c inst.r_format.rs == 0 && rs 933 arch/mips/kvm/vz.c switch (inst.c0r_format.rs) { rs 984 arch/mips/kvm/vz.c if (inst.c0r_format.rs == mfc_op) rs 989 arch/mips/kvm/vz.c trace_kvm_hwr(vcpu, (inst.c0r_format.rs == mfc_op) ? rs 1000 arch/mips/kvm/vz.c trace_kvm_hwr(vcpu, (inst.c0r_format.rs == mtc_op) ? rs 1026 arch/mips/kvm/vz.c val = mips_process_maar(inst.c0r_format.rs, rs 1083 arch/mips/kvm/vz.c base = inst.i_format.rs; rs 1170 arch/mips/kvm/vz.c if (inst.r_format.rs || (inst.r_format.re >> 3)) rs 1229 arch/mips/kvm/vz.c inst.c0r_format.rs == mtc_op && rs 67 arch/mips/lasat/picvue.c data &= ~picvue->rs; rs 71 arch/mips/lasat/picvue.c data |= picvue->rs; rs 97 arch/mips/lasat/picvue.c data |= picvue->rs; rs 99 arch/mips/lasat/picvue.c data &= ~picvue->rs; rs 104 arch/mips/lasat/picvue.c data &= ~picvue->rs; rs 106 arch/mips/lasat/picvue.c data |= picvue->rs; rs 14 arch/mips/lasat/picvue.h u32 rs; rs 91 arch/mips/math-emu/cp1emu.c mips32_insn.mm_i_format.rt = insn.mm_i_format.rs; rs 92 arch/mips/math-emu/cp1emu.c mips32_insn.mm_i_format.rs = insn.mm_i_format.rt; rs 96 arch/mips/math-emu/cp1emu.c mips32_insn.mm_i_format.rt = insn.mm_i_format.rs; rs 97 arch/mips/math-emu/cp1emu.c mips32_insn.mm_i_format.rs = insn.mm_i_format.rt; rs 101 arch/mips/math-emu/cp1emu.c mips32_insn.mm_i_format.rt = insn.mm_i_format.rs; rs 102 arch/mips/math-emu/cp1emu.c mips32_insn.mm_i_format.rs = insn.mm_i_format.rt; rs 106 arch/mips/math-emu/cp1emu.c mips32_insn.mm_i_format.rt = insn.mm_i_format.rs; rs 107 arch/mips/math-emu/cp1emu.c mips32_insn.mm_i_format.rs = insn.mm_i_format.rt; rs 168 arch/mips/math-emu/cp1emu.c mips32_insn.r_format.rs = rs 254 arch/mips/math-emu/cp1emu.c mips32_insn.r_format.rs = insn.mm_fp4_format.fs; rs 447 arch/mips/math-emu/cp1emu.c *contpc = regs->regs[insn.r_format.rs]; rs 455 arch/mips/math-emu/cp1emu.c if (NO_R6EMU && (insn.i_format.rs || rs 468 arch/mips/math-emu/cp1emu.c if ((long)regs->regs[insn.i_format.rs] < 0) rs 479 arch/mips/math-emu/cp1emu.c if (NO_R6EMU && (insn.i_format.rs || rs 492 arch/mips/math-emu/cp1emu.c if ((long)regs->regs[insn.i_format.rs] >= 0) rs 524 arch/mips/math-emu/cp1emu.c if (regs->regs[insn.i_format.rs] == rs 539 arch/mips/math-emu/cp1emu.c if (regs->regs[insn.i_format.rs] != rs 569 arch/mips/math-emu/cp1emu.c ((!insn.i_format.rs && insn.i_format.rt) || rs 570 arch/mips/math-emu/cp1emu.c (insn.i_format.rs == insn.i_format.rt))) rs 578 arch/mips/math-emu/cp1emu.c if ((long)regs->regs[insn.i_format.rs] <= 0) rs 607 arch/mips/math-emu/cp1emu.c ((!insn.i_format.rs && insn.i_format.rt) || rs 608 arch/mips/math-emu/cp1emu.c (insn.i_format.rs == insn.i_format.rt))) rs 617 arch/mips/math-emu/cp1emu.c if ((long)regs->regs[insn.i_format.rs] > 0) rs 630 arch/mips/math-emu/cp1emu.c if (insn.i_format.rt && !insn.i_format.rs) rs 638 arch/mips/math-emu/cp1emu.c if ((regs->regs[insn.i_format.rs] & (1ull<<insn.i_format.rt)) == 0) rs 644 arch/mips/math-emu/cp1emu.c if ((regs->regs[insn.i_format.rs] & (1ull<<(insn.i_format.rt + 32))) == 0) rs 650 arch/mips/math-emu/cp1emu.c if (regs->regs[insn.i_format.rs] & (1ull<<insn.i_format.rt)) rs 656 arch/mips/math-emu/cp1emu.c if (regs->regs[insn.i_format.rs] & (1ull<<(insn.i_format.rt + 32))) rs 692 arch/mips/math-emu/cp1emu.c if (!insn.i_format.rs) rs 703 arch/mips/math-emu/cp1emu.c ((insn.i_format.rs == bc1eqz_op) || rs 704 arch/mips/math-emu/cp1emu.c (insn.i_format.rs == bc1nez_op))) { rs 708 arch/mips/math-emu/cp1emu.c switch (insn.i_format.rs) { rs 731 arch/mips/math-emu/cp1emu.c if (insn.i_format.rs == bc_op) { rs 44 arch/mips/math-emu/dp_maddf.c int rs; rs 190 arch/mips/math-emu/dp_maddf.c rs = xs ^ ys; rs 192 arch/mips/math-emu/dp_maddf.c rs ^= 1; rs 242 arch/mips/math-emu/dp_maddf.c return ieee754dp_format(rs, re, lrm); rs 270 arch/mips/math-emu/dp_maddf.c if (zs == rs) { rs 288 arch/mips/math-emu/dp_maddf.c zs = rs; rs 15 arch/mips/math-emu/dp_mul.c int rs; rs 110 arch/mips/math-emu/dp_mul.c rs = xs ^ ys; rs 159 arch/mips/math-emu/dp_mul.c return ieee754dp_format(rs, re, rm); rs 235 arch/mips/math-emu/dsemul.c unsigned int rs; rs 238 arch/mips/math-emu/dsemul.c rs = (((insn.mm_a_format.rs + 0xe) & 0xf) + 2); rs 241 arch/mips/math-emu/dsemul.c regs->regs[rs] = (long)v; rs 37 arch/mips/math-emu/ieee754dp.h #define XDPSRS(v,rs) \ rs 38 arch/mips/math-emu/ieee754dp.h ((rs > (DP_FBITS+3))?1:((v) >> (rs)) | ((v) << (64-(rs)) != 0)) rs 37 arch/mips/math-emu/ieee754sp.h #define XSPSRS64(v, rs) \ rs 38 arch/mips/math-emu/ieee754sp.h (((rs) >= 64) ? ((v) != 0) : ((v) >> (rs)) | ((v) << (64-(rs)) != 0)) rs 41 arch/mips/math-emu/ieee754sp.h #define XSPSRS(v, rs) \ rs 42 arch/mips/math-emu/ieee754sp.h ((rs > (SP_FBITS+3))?1:((v) >> (rs)) | ((v) << (32-(rs)) != 0)) rs 19 arch/mips/math-emu/sp_maddf.c int rs; rs 161 arch/mips/math-emu/sp_maddf.c rs = xs ^ ys; rs 163 arch/mips/math-emu/sp_maddf.c rs ^= 1; rs 186 arch/mips/math-emu/sp_maddf.c return ieee754sp_format(rs, re, rm); rs 213 arch/mips/math-emu/sp_maddf.c if (zs == rs) { rs 228 arch/mips/math-emu/sp_maddf.c zs = rs; rs 15 arch/mips/math-emu/sp_mul.c int rs; rs 110 arch/mips/math-emu/sp_mul.c rs = xs ^ ys; rs 154 arch/mips/math-emu/sp_mul.c return ieee754sp_format(rs, re, rm); rs 457 arch/mips/mm/uasm.c void UASM_i_LA_mostly(u32 **buf, unsigned int rs, long addr) rs 460 arch/mips/mm/uasm.c uasm_i_lui(buf, rs, uasm_rel_highest(addr)); rs 462 arch/mips/mm/uasm.c uasm_i_daddiu(buf, rs, rs, uasm_rel_higher(addr)); rs 464 arch/mips/mm/uasm.c uasm_i_dsll(buf, rs, rs, 16); rs 465 arch/mips/mm/uasm.c uasm_i_daddiu(buf, rs, rs, rs 467 arch/mips/mm/uasm.c uasm_i_dsll(buf, rs, rs, 16); rs 469 arch/mips/mm/uasm.c uasm_i_dsll32(buf, rs, rs, 0); rs 471 arch/mips/mm/uasm.c uasm_i_lui(buf, rs, uasm_rel_hi(addr)); rs 475 arch/mips/mm/uasm.c void UASM_i_LA(u32 **buf, unsigned int rs, long addr) rs 477 arch/mips/mm/uasm.c UASM_i_LA_mostly(buf, rs, addr); rs 480 arch/mips/mm/uasm.c uasm_i_daddiu(buf, rs, rs, rs 483 arch/mips/mm/uasm.c uasm_i_addiu(buf, rs, rs, rs 36 arch/mips/oprofile/backtrace.c && ip->i_format.rs == 29 && ip->i_format.rt == 31; rs 42 arch/mips/oprofile/backtrace.c if (ip->i_format.rs != 29 || ip->i_format.rt != 29) rs 56 arch/mips/oprofile/backtrace.c if (ip->r_format.func == jr_op && ip->r_format.rs == 31) rs 44 arch/mips/txx9/generic/mem_tx4927.c unsigned int rs = 0; rs 59 arch/mips/txx9/generic/mem_tx4927.c rs = 2048 << sdccr_rs; rs 64 arch/mips/txx9/generic/mem_tx4927.c return rs * cs * mw * bs; rs 393 arch/powerpc/include/asm/kvm_host.h void (*slbmte)(struct kvm_vcpu *vcpu, u64 rb, u64 rs); rs 78 arch/powerpc/include/asm/kvm_ppc.h unsigned int rs, unsigned int bytes, int is_default_endian); rs 83 arch/powerpc/include/asm/kvm_ppc.h int rs, unsigned int bytes, rs 492 arch/powerpc/include/asm/ppc-opcode.h #define PPC_TLBIE_5(rb,rs,ric,prs,r) \ rs 494 arch/powerpc/include/asm/ppc-opcode.h ___PPC_RB(rb) | ___PPC_RS(rs) | \ rs 497 arch/powerpc/include/asm/ppc-opcode.h #define PPC_TLBIEL(rb,rs,ric,prs,r) \ rs 499 arch/powerpc/include/asm/ppc-opcode.h ___PPC_RB(rb) | ___PPC_RS(rs) | \ rs 483 arch/powerpc/include/asm/ppc_asm.h #define tophys(rd,rs) \ rs 484 arch/powerpc/include/asm/ppc_asm.h addis rd,rs,0 rs 486 arch/powerpc/include/asm/ppc_asm.h #define tovirt(rd,rs) \ rs 487 arch/powerpc/include/asm/ppc_asm.h addis rd,rs,0 rs 493 arch/powerpc/include/asm/ppc_asm.h #define tophys(rd,rs) \ rs 494 arch/powerpc/include/asm/ppc_asm.h clrldi rd,rs,2 rs 496 arch/powerpc/include/asm/ppc_asm.h #define tovirt(rd,rs) \ rs 497 arch/powerpc/include/asm/ppc_asm.h rotldi rd,rs,16; \ rs 504 arch/powerpc/include/asm/ppc_asm.h #define tophys(rd, rs) addis rd, rs, -PAGE_OFFSET@h rs 505 arch/powerpc/include/asm/ppc_asm.h #define tovirt(rd, rs) addis rd, rs, PAGE_OFFSET@h rs 191 arch/powerpc/include/asm/trace.h unsigned long rs, unsigned long ric, unsigned long prs, rs 193 arch/powerpc/include/asm/trace.h TP_ARGS(lpid, local, rb, rs, ric, prs, r), rs 198 arch/powerpc/include/asm/trace.h __field(unsigned long, rs) rs 208 arch/powerpc/include/asm/trace.h __entry->rs = rs; rs 216 arch/powerpc/include/asm/trace.h __entry->rb, __entry->rs, __entry->ric, __entry->prs, rs 175 arch/powerpc/kernel/isa-bridge.c int rlen, i, rs; rs 197 arch/powerpc/kernel/isa-bridge.c rs = na + ns + pna; rs 201 arch/powerpc/kernel/isa-bridge.c if (ranges == NULL || rlen < rs) { rs 208 arch/powerpc/kernel/isa-bridge.c for (i = 0; (i + rs - 1) < rlen; i += rs) { rs 78 arch/powerpc/kernel/legacy_serial.c const __be32 *clk, *spd, *rs; rs 92 arch/powerpc/kernel/legacy_serial.c rs = of_get_property(np, "reg-shift", NULL); rs 93 arch/powerpc/kernel/legacy_serial.c if (rs && *rs) rs 94 arch/powerpc/kernel/legacy_serial.c shift = be32_to_cpup(rs); rs 313 arch/powerpc/kernel/traps.c static DEFINE_RATELIMIT_STATE(rs, DEFAULT_RATELIMIT_INTERVAL, rs 322 arch/powerpc/kernel/traps.c if (!__ratelimit(&rs)) rs 1275 arch/powerpc/kernel/traps.c u32 ra,rs; rs 1279 arch/powerpc/kernel/traps.c rs = (instword >> 21) & 0x1f; rs 1281 arch/powerpc/kernel/traps.c tmp = regs->gpr[rs]; rs 378 arch/powerpc/kvm/book3s_64_mmu.c static void kvmppc_mmu_book3s_64_slbmte(struct kvm_vcpu *vcpu, u64 rs, u64 rb) rs 384 arch/powerpc/kvm/book3s_64_mmu.c dprintk("KVM MMU: slbmte(0x%llx, 0x%llx)\n", rs, rb); rs 395 arch/powerpc/kvm/book3s_64_mmu.c slbe->large = (rs & SLB_VSID_L) ? 1 : 0; rs 396 arch/powerpc/kvm/book3s_64_mmu.c slbe->tb = (rs & SLB_VSID_B_1T) ? 1 : 0; rs 398 arch/powerpc/kvm/book3s_64_mmu.c slbe->vsid = (rs & ~SLB_VSID_B) >> (kvmppc_slb_sid_shift(slbe) - 16); rs 400 arch/powerpc/kvm/book3s_64_mmu.c slbe->Ks = (rs & SLB_VSID_KS) ? 1 : 0; rs 401 arch/powerpc/kvm/book3s_64_mmu.c slbe->Kp = (rs & SLB_VSID_KP) ? 1 : 0; rs 402 arch/powerpc/kvm/book3s_64_mmu.c slbe->nx = (rs & SLB_VSID_N) ? 1 : 0; rs 403 arch/powerpc/kvm/book3s_64_mmu.c slbe->class = (rs & SLB_VSID_C) ? 1 : 0; rs 408 arch/powerpc/kvm/book3s_64_mmu.c switch (rs & SLB_VSID_LP) { rs 421 arch/powerpc/kvm/book3s_64_mmu.c slbe->origv = rs; rs 507 arch/powerpc/kvm/book3s_64_mmu.c u64 rb = 0, rs = 0; rs 537 arch/powerpc/kvm/book3s_64_mmu.c rs |= (value & 0xfffffff) << 12; rs 539 arch/powerpc/kvm/book3s_64_mmu.c rs |= ((value >> 28) & 0x7) << 9; rs 541 arch/powerpc/kvm/book3s_64_mmu.c kvmppc_mmu_book3s_64_slbmte(vcpu, rs, rb); rs 243 arch/powerpc/kvm/book3s_emulate.c int rs = get_rs(inst); rs 303 arch/powerpc/kvm/book3s_emulate.c ulong rs_val = kvmppc_get_gpr(vcpu, rs); rs 314 arch/powerpc/kvm/book3s_emulate.c kvmppc_set_msr(vcpu, kvmppc_get_gpr(vcpu, rs)); rs 343 arch/powerpc/kvm/book3s_emulate.c kvmppc_get_gpr(vcpu, rs)); rs 348 arch/powerpc/kvm/book3s_emulate.c kvmppc_get_gpr(vcpu, rs)); rs 393 arch/powerpc/kvm/book3s_emulate.c kvmppc_get_gpr(vcpu, rs), rs 57 arch/powerpc/kvm/book3s_hv_ras.c unsigned long rs = be64_to_cpu(slb->save_area[i].vsid); rs 60 arch/powerpc/kvm/book3s_hv_ras.c asm volatile("slbmte %0,%1" : : "r" (rs), "r" (rb)); rs 442 arch/powerpc/kvm/book3s_hv_rm_mmu.c unsigned long rb,rs,prs,r,ric; rs 445 arch/powerpc/kvm/book3s_hv_rm_mmu.c rs = 0; /* lpid = 0 */ rs 457 arch/powerpc/kvm/book3s_hv_rm_mmu.c "i"(ric), "r"(rs) : "memory"); rs 45 arch/powerpc/kvm/book3s_hv_tm.c int ra, rs; rs 96 arch/powerpc/kvm/book3s_hv_tm.c rs = (instr >> 21) & 0x1f; rs 97 arch/powerpc/kvm/book3s_hv_tm.c newmsr = kvmppc_get_gpr(vcpu, rs); rs 24 arch/powerpc/kvm/book3s_hv_tm_builtin.c int rs; rs 64 arch/powerpc/kvm/book3s_hv_tm_builtin.c rs = (instr >> 21) & 0x1f; rs 65 arch/powerpc/kvm/book3s_hv_tm_builtin.c newmsr = kvmppc_get_gpr(vcpu, rs); rs 173 arch/powerpc/kvm/book3s_paired_singles.c int rs, ulong addr, int ls_type) rs 191 arch/powerpc/kvm/book3s_paired_singles.c emulated = kvmppc_handle_load(run, vcpu, KVM_MMIO_REG_FPR | rs, rs 201 arch/powerpc/kvm/book3s_paired_singles.c kvm_cvt_fd((u32*)tmp, &VCPU_FPR(vcpu, rs)); rs 202 arch/powerpc/kvm/book3s_paired_singles.c vcpu->arch.qpr[rs] = *((u32*)tmp); rs 205 arch/powerpc/kvm/book3s_paired_singles.c VCPU_FPR(vcpu, rs) = *((u64*)tmp); rs 217 arch/powerpc/kvm/book3s_paired_singles.c int rs, ulong addr, int ls_type) rs 227 arch/powerpc/kvm/book3s_paired_singles.c kvm_cvt_df(&VCPU_FPR(vcpu, rs), (u32*)tmp); rs 232 arch/powerpc/kvm/book3s_paired_singles.c *((u32*)tmp) = VCPU_FPR(vcpu, rs); rs 233 arch/powerpc/kvm/book3s_paired_singles.c val = VCPU_FPR(vcpu, rs) & 0xffffffff; rs 237 arch/powerpc/kvm/book3s_paired_singles.c *((u64*)tmp) = VCPU_FPR(vcpu, rs); rs 238 arch/powerpc/kvm/book3s_paired_singles.c val = VCPU_FPR(vcpu, rs); rs 263 arch/powerpc/kvm/book3s_paired_singles.c int rs, ulong addr, bool w, int i) rs 282 arch/powerpc/kvm/book3s_paired_singles.c emulated = kvmppc_handle_load(run, vcpu, KVM_MMIO_REG_FPR | rs, rs 284 arch/powerpc/kvm/book3s_paired_singles.c vcpu->arch.qpr[rs] = tmp[1]; rs 287 arch/powerpc/kvm/book3s_paired_singles.c emulated = kvmppc_handle_load(run, vcpu, KVM_MMIO_REG_FQPR | rs, rs 295 arch/powerpc/kvm/book3s_paired_singles.c kvm_cvt_fd(&tmp[0], &VCPU_FPR(vcpu, rs)); rs 296 arch/powerpc/kvm/book3s_paired_singles.c vcpu->arch.qpr[rs] = tmp[1]; rs 306 arch/powerpc/kvm/book3s_paired_singles.c int rs, ulong addr, bool w, int i) rs 313 arch/powerpc/kvm/book3s_paired_singles.c kvm_cvt_df(&VCPU_FPR(vcpu, rs), &tmp[0]); rs 314 arch/powerpc/kvm/book3s_paired_singles.c tmp[1] = vcpu->arch.qpr[rs]; rs 1483 arch/powerpc/kvm/book3s_pr.c u64 rs = sregs->u.s.ppc64.slb[i].slbv; rs 1486 arch/powerpc/kvm/book3s_pr.c vcpu->arch.mmu.slbmte(vcpu, rs, rb); rs 46 arch/powerpc/kvm/booke_emulate.c int rs = get_rs(inst); rs 86 arch/powerpc/kvm/booke_emulate.c kvmppc_set_msr(vcpu, kvmppc_get_gpr(vcpu, rs)); rs 91 arch/powerpc/kvm/booke_emulate.c | (kvmppc_get_gpr(vcpu, rs) & MSR_EE); rs 75 arch/powerpc/kvm/emulate.c static int kvmppc_emulate_mtspr(struct kvm_vcpu *vcpu, int sprn, int rs) rs 78 arch/powerpc/kvm/emulate.c ulong spr_val = kvmppc_get_gpr(vcpu, rs); rs 197 arch/powerpc/kvm/emulate.c int rs, rt, sprn; rs 210 arch/powerpc/kvm/emulate.c rs = get_rs(inst); rs 251 arch/powerpc/kvm/emulate.c emulated = kvmppc_emulate_mtspr(vcpu, sprn, rs); rs 1372 arch/powerpc/kvm/powerpc.c static inline int kvmppc_get_vsr_data(struct kvm_vcpu *vcpu, int rs, u64 *val) rs 1390 arch/powerpc/kvm/powerpc.c if (rs < 32) { rs 1391 arch/powerpc/kvm/powerpc.c *val = VCPU_VSX_FPR(vcpu, rs, vsx_offset); rs 1393 arch/powerpc/kvm/powerpc.c reg.vval = VCPU_VSX_VR(vcpu, rs - 32); rs 1407 arch/powerpc/kvm/powerpc.c if (rs < 32) { rs 1410 arch/powerpc/kvm/powerpc.c reg.vsxval[0] = VCPU_VSX_FPR(vcpu, rs, dword_offset); rs 1413 arch/powerpc/kvm/powerpc.c reg.vval = VCPU_VSX_VR(vcpu, rs - 32); rs 1427 arch/powerpc/kvm/powerpc.c int rs, unsigned int bytes, int is_default_endian) rs 1432 arch/powerpc/kvm/powerpc.c vcpu->arch.io_gpr = rs; rs 1439 arch/powerpc/kvm/powerpc.c if (kvmppc_get_vsr_data(vcpu, rs, &val) == -1) rs 1589 arch/powerpc/kvm/powerpc.c unsigned int rs, unsigned int bytes, int is_default_endian) rs 1592 arch/powerpc/kvm/powerpc.c unsigned int index = rs & KVM_MMIO_REG_MASK; rs 1598 arch/powerpc/kvm/powerpc.c vcpu->arch.io_gpr = rs; rs 64 arch/powerpc/mm/book3s64/hash_native.c unsigned long rs; rs 68 arch/powerpc/mm/book3s64/hash_native.c rs = ((unsigned long)pid << PPC_BITLSHIFT(31)); rs 71 arch/powerpc/mm/book3s64/hash_native.c : : "r"(rb), "r"(rs), "i"(ric), "i"(prs), "r"(r) rs 206 arch/powerpc/mm/book3s64/hash_native.c unsigned long rb,rs,prs,r,ric; rs 209 arch/powerpc/mm/book3s64/hash_native.c rs = 0; /* lpid = 0 */ rs 221 arch/powerpc/mm/book3s64/hash_native.c "i"(ric), "r"(rs) : "memory"); rs 33 arch/powerpc/mm/book3s64/radix_tlb.c unsigned long rs; rs 36 arch/powerpc/mm/book3s64/radix_tlb.c rs = ((unsigned long)pid << PPC_BITLSHIFT(31)); rs 39 arch/powerpc/mm/book3s64/radix_tlb.c : : "r"(rb), "r"(rs), "i"(ric), "i"(prs) rs 96 arch/powerpc/mm/book3s64/radix_tlb.c unsigned long rb,rs,prs,r; rs 100 arch/powerpc/mm/book3s64/radix_tlb.c rs = ((unsigned long)pid) << PPC_BITLSHIFT(31); rs 105 arch/powerpc/mm/book3s64/radix_tlb.c : : "r"(rb), "i"(r), "i"(prs), "i"(ric), "r"(rs) : "memory"); rs 106 arch/powerpc/mm/book3s64/radix_tlb.c trace_tlbie(0, 1, rb, rs, ric, prs, r); rs 111 arch/powerpc/mm/book3s64/radix_tlb.c unsigned long rb,rs,prs,r; rs 114 arch/powerpc/mm/book3s64/radix_tlb.c rs = pid << PPC_BITLSHIFT(31); rs 119 arch/powerpc/mm/book3s64/radix_tlb.c : : "r"(rb), "i"(r), "i"(prs), "i"(ric), "r"(rs) : "memory"); rs 120 arch/powerpc/mm/book3s64/radix_tlb.c trace_tlbie(0, 0, rb, rs, ric, prs, r); rs 125 arch/powerpc/mm/book3s64/radix_tlb.c unsigned long rb,rs,prs,r; rs 128 arch/powerpc/mm/book3s64/radix_tlb.c rs = lpid; rs 133 arch/powerpc/mm/book3s64/radix_tlb.c : : "r"(rb), "i"(r), "i"(prs), "i"(ric), "r"(rs) : "memory"); rs 134 arch/powerpc/mm/book3s64/radix_tlb.c trace_tlbie(lpid, 0, rb, rs, ric, prs, r); rs 139 arch/powerpc/mm/book3s64/radix_tlb.c unsigned long rb,rs,prs,r; rs 142 arch/powerpc/mm/book3s64/radix_tlb.c rs = lpid; rs 147 arch/powerpc/mm/book3s64/radix_tlb.c : : "r"(rb), "i"(r), "i"(prs), "i"(ric), "r"(rs) : "memory"); rs 148 arch/powerpc/mm/book3s64/radix_tlb.c trace_tlbie(lpid, 0, rb, rs, ric, prs, r); rs 154 arch/powerpc/mm/book3s64/radix_tlb.c unsigned long rb,rs,prs,r; rs 158 arch/powerpc/mm/book3s64/radix_tlb.c rs = pid << PPC_BITLSHIFT(31); rs 163 arch/powerpc/mm/book3s64/radix_tlb.c : : "r"(rb), "i"(r), "i"(prs), "i"(ric), "r"(rs) : "memory"); rs 164 arch/powerpc/mm/book3s64/radix_tlb.c trace_tlbie(0, 1, rb, rs, ric, prs, r); rs 170 arch/powerpc/mm/book3s64/radix_tlb.c unsigned long rb,rs,prs,r; rs 174 arch/powerpc/mm/book3s64/radix_tlb.c rs = pid << PPC_BITLSHIFT(31); rs 179 arch/powerpc/mm/book3s64/radix_tlb.c : : "r"(rb), "i"(r), "i"(prs), "i"(ric), "r"(rs) : "memory"); rs 180 arch/powerpc/mm/book3s64/radix_tlb.c trace_tlbie(0, 0, rb, rs, ric, prs, r); rs 186 arch/powerpc/mm/book3s64/radix_tlb.c unsigned long rb,rs,prs,r; rs 190 arch/powerpc/mm/book3s64/radix_tlb.c rs = lpid; rs 195 arch/powerpc/mm/book3s64/radix_tlb.c : : "r"(rb), "i"(r), "i"(prs), "i"(ric), "r"(rs) : "memory"); rs 196 arch/powerpc/mm/book3s64/radix_tlb.c trace_tlbie(lpid, 0, rb, rs, ric, prs, r); rs 1194 arch/powerpc/mm/book3s64/radix_tlb.c unsigned long rb,prs,r,rs; rs 1200 arch/powerpc/mm/book3s64/radix_tlb.c rs = 1 & ((1UL << 32) - 1); /* any LPID value to flush guest mappings */ rs 1207 arch/powerpc/mm/book3s64/radix_tlb.c : : "r"(rb), "i"(r), "i"(1), "i"(ric), "r"(rs) : "memory"); rs 675 arch/riscv/net/bpf_jit_comp.c static void init_regs(u8 *rd, u8 *rs, const struct bpf_insn *insn, rs 693 arch/riscv/net/bpf_jit_comp.c *rs = bpf_to_rv_reg(insn->src_reg, ctx); rs 703 arch/riscv/net/bpf_jit_comp.c static void emit_zext_32_rd_rs(u8 *rd, u8 *rs, struct rv_jit_context *ctx) rs 707 arch/riscv/net/bpf_jit_comp.c emit(rv_addi(RV_REG_T1, *rs, 0), ctx); rs 710 arch/riscv/net/bpf_jit_comp.c *rs = RV_REG_T1; rs 713 arch/riscv/net/bpf_jit_comp.c static void emit_sext_32_rd_rs(u8 *rd, u8 *rs, struct rv_jit_context *ctx) rs 716 arch/riscv/net/bpf_jit_comp.c emit(rv_addiw(RV_REG_T1, *rs, 0), ctx); rs 718 arch/riscv/net/bpf_jit_comp.c *rs = RV_REG_T1; rs 742 arch/riscv/net/bpf_jit_comp.c u8 rd = -1, rs = -1, code = insn->code; rs 746 arch/riscv/net/bpf_jit_comp.c init_regs(&rd, &rs, insn, ctx); rs 757 arch/riscv/net/bpf_jit_comp.c emit(is64 ? rv_addi(rd, rs, 0) : rv_addiw(rd, rs, 0), ctx); rs 765 arch/riscv/net/bpf_jit_comp.c emit(is64 ? rv_add(rd, rd, rs) : rv_addw(rd, rd, rs), ctx); rs 771 arch/riscv/net/bpf_jit_comp.c emit(is64 ? rv_sub(rd, rd, rs) : rv_subw(rd, rd, rs), ctx); rs 777 arch/riscv/net/bpf_jit_comp.c emit(rv_and(rd, rd, rs), ctx); rs 783 arch/riscv/net/bpf_jit_comp.c emit(rv_or(rd, rd, rs), ctx); rs 789 arch/riscv/net/bpf_jit_comp.c emit(rv_xor(rd, rd, rs), ctx); rs 795 arch/riscv/net/bpf_jit_comp.c emit(is64 ? rv_mul(rd, rd, rs) : rv_mulw(rd, rd, rs), ctx); rs 801 arch/riscv/net/bpf_jit_comp.c emit(is64 ? rv_divu(rd, rd, rs) : rv_divuw(rd, rd, rs), ctx); rs 807 arch/riscv/net/bpf_jit_comp.c emit(is64 ? rv_remu(rd, rd, rs) : rv_remuw(rd, rd, rs), ctx); rs 813 arch/riscv/net/bpf_jit_comp.c emit(is64 ? rv_sll(rd, rd, rs) : rv_sllw(rd, rd, rs), ctx); rs 819 arch/riscv/net/bpf_jit_comp.c emit(is64 ? rv_srl(rd, rd, rs) : rv_srlw(rd, rd, rs), ctx); rs 825 arch/riscv/net/bpf_jit_comp.c emit(is64 ? rv_sra(rd, rd, rs) : rv_sraw(rd, rd, rs), ctx); rs 1025 arch/riscv/net/bpf_jit_comp.c emit_zext_32_rd_rs(&rd, &rs, ctx); rs 1026 arch/riscv/net/bpf_jit_comp.c emit(rv_beq(rd, rs, rvoff >> 1), ctx); rs 1033 arch/riscv/net/bpf_jit_comp.c emit_zext_32_rd_rs(&rd, &rs, ctx); rs 1034 arch/riscv/net/bpf_jit_comp.c emit(rv_bltu(rs, rd, rvoff >> 1), ctx); rs 1041 arch/riscv/net/bpf_jit_comp.c emit_zext_32_rd_rs(&rd, &rs, ctx); rs 1042 arch/riscv/net/bpf_jit_comp.c emit(rv_bltu(rd, rs, rvoff >> 1), ctx); rs 1049 arch/riscv/net/bpf_jit_comp.c emit_zext_32_rd_rs(&rd, &rs, ctx); rs 1050 arch/riscv/net/bpf_jit_comp.c emit(rv_bgeu(rd, rs, rvoff >> 1), ctx); rs 1057 arch/riscv/net/bpf_jit_comp.c emit_zext_32_rd_rs(&rd, &rs, ctx); rs 1058 arch/riscv/net/bpf_jit_comp.c emit(rv_bgeu(rs, rd, rvoff >> 1), ctx); rs 1065 arch/riscv/net/bpf_jit_comp.c emit_zext_32_rd_rs(&rd, &rs, ctx); rs 1066 arch/riscv/net/bpf_jit_comp.c emit(rv_bne(rd, rs, rvoff >> 1), ctx); rs 1073 arch/riscv/net/bpf_jit_comp.c emit_sext_32_rd_rs(&rd, &rs, ctx); rs 1074 arch/riscv/net/bpf_jit_comp.c emit(rv_blt(rs, rd, rvoff >> 1), ctx); rs 1081 arch/riscv/net/bpf_jit_comp.c emit_sext_32_rd_rs(&rd, &rs, ctx); rs 1082 arch/riscv/net/bpf_jit_comp.c emit(rv_blt(rd, rs, rvoff >> 1), ctx); rs 1089 arch/riscv/net/bpf_jit_comp.c emit_sext_32_rd_rs(&rd, &rs, ctx); rs 1090 arch/riscv/net/bpf_jit_comp.c emit(rv_bge(rd, rs, rvoff >> 1), ctx); rs 1097 arch/riscv/net/bpf_jit_comp.c emit_sext_32_rd_rs(&rd, &rs, ctx); rs 1098 arch/riscv/net/bpf_jit_comp.c emit(rv_bge(rs, rd, rvoff >> 1), ctx); rs 1105 arch/riscv/net/bpf_jit_comp.c emit_zext_32_rd_rs(&rd, &rs, ctx); rs 1106 arch/riscv/net/bpf_jit_comp.c emit(rv_and(RV_REG_T1, rd, rs), ctx); rs 1271 arch/riscv/net/bpf_jit_comp.c emit(rv_lbu(rd, off, rs), ctx); rs 1276 arch/riscv/net/bpf_jit_comp.c emit(rv_add(RV_REG_T1, RV_REG_T1, rs), ctx); rs 1283 arch/riscv/net/bpf_jit_comp.c emit(rv_lhu(rd, off, rs), ctx); rs 1288 arch/riscv/net/bpf_jit_comp.c emit(rv_add(RV_REG_T1, RV_REG_T1, rs), ctx); rs 1295 arch/riscv/net/bpf_jit_comp.c emit(rv_lwu(rd, off, rs), ctx); rs 1300 arch/riscv/net/bpf_jit_comp.c emit(rv_add(RV_REG_T1, RV_REG_T1, rs), ctx); rs 1307 arch/riscv/net/bpf_jit_comp.c emit(rv_ld(rd, off, rs), ctx); rs 1312 arch/riscv/net/bpf_jit_comp.c emit(rv_add(RV_REG_T1, RV_REG_T1, rs), ctx); rs 1366 arch/riscv/net/bpf_jit_comp.c emit(rv_sb(rd, off, rs), ctx); rs 1372 arch/riscv/net/bpf_jit_comp.c emit(rv_sb(RV_REG_T1, 0, rs), ctx); rs 1376 arch/riscv/net/bpf_jit_comp.c emit(rv_sh(rd, off, rs), ctx); rs 1382 arch/riscv/net/bpf_jit_comp.c emit(rv_sh(RV_REG_T1, 0, rs), ctx); rs 1386 arch/riscv/net/bpf_jit_comp.c emit(rv_sw(rd, off, rs), ctx); rs 1392 arch/riscv/net/bpf_jit_comp.c emit(rv_sw(RV_REG_T1, 0, rs), ctx); rs 1396 arch/riscv/net/bpf_jit_comp.c emit(rv_sd(rd, off, rs), ctx); rs 1402 arch/riscv/net/bpf_jit_comp.c emit(rv_sd(RV_REG_T1, 0, rs), ctx); rs 1420 arch/riscv/net/bpf_jit_comp.c rv_amoadd_w(RV_REG_ZERO, rs, rd, 0, 0) : rs 1421 arch/riscv/net/bpf_jit_comp.c rv_amoadd_d(RV_REG_ZERO, rs, rd, 0, 0), ctx); rs 321 arch/s390/net/bpf_jit_comp.c static void save_regs(struct bpf_jit *jit, u32 rs, u32 re) rs 323 arch/s390/net/bpf_jit_comp.c u32 off = STK_OFF_R6 + (rs - 6) * 8; rs 325 arch/s390/net/bpf_jit_comp.c if (rs == re) rs 327 arch/s390/net/bpf_jit_comp.c _EMIT6(0xe300f000 | rs << 20 | off, 0x0024); rs 330 arch/s390/net/bpf_jit_comp.c _EMIT6_DISP(0xeb00f000 | rs << 20 | re << 16, 0x0024, off); rs 336 arch/s390/net/bpf_jit_comp.c static void restore_regs(struct bpf_jit *jit, u32 rs, u32 re, u32 stack_depth) rs 338 arch/s390/net/bpf_jit_comp.c u32 off = STK_OFF_R6 + (rs - 6) * 8; rs 343 arch/s390/net/bpf_jit_comp.c if (rs == re) rs 345 arch/s390/net/bpf_jit_comp.c _EMIT6(0xe300f000 | rs << 20 | off, 0x0004); rs 348 arch/s390/net/bpf_jit_comp.c _EMIT6_DISP(0xeb00f000 | rs << 20 | re << 16, 0x0004, off); rs 388 arch/s390/net/bpf_jit_comp.c int re = 6, rs; rs 391 arch/s390/net/bpf_jit_comp.c rs = get_start(jit, re); rs 392 arch/s390/net/bpf_jit_comp.c if (!rs) rs 394 arch/s390/net/bpf_jit_comp.c re = get_end(jit, rs + 1); rs 396 arch/s390/net/bpf_jit_comp.c save_regs(jit, rs, re); rs 398 arch/s390/net/bpf_jit_comp.c restore_regs(jit, rs, re, stack_depth); rs 72 arch/sh/include/uapi/asm/ptrace_32.h unsigned long rs; rs 858 arch/x86/kernel/ptrace.c #define SEG32(rs) \ rs 859 arch/x86/kernel/ptrace.c case offsetof(struct user32, regs.rs): \ rs 861 arch/x86/kernel/ptrace.c offsetof(struct user_regs_struct, rs), \ rs 932 arch/x86/kernel/ptrace.c #define SEG32(rs) \ rs 933 arch/x86/kernel/ptrace.c case offsetof(struct user32, regs.rs): \ rs 935 arch/x86/kernel/ptrace.c offsetof(struct user_regs_struct, rs)); \ rs 415 arch/xtensa/include/asm/pgtable.h #define _PGD_INDEX(rt,rs) extui rt, rs, PGDIR_SHIFT, 32-PGDIR_SHIFT rs 416 arch/xtensa/include/asm/pgtable.h #define _PTE_INDEX(rt,rs) extui rt, rs, PAGE_SHIFT, PTRS_PER_PTE_SHIFT rs 37 block/partitions/atari.c struct rootsector *rs; rs 53 block/partitions/atari.c rs = read_part_sector(state, 0, §); rs 54 block/partitions/atari.c if (!rs) rs 59 block/partitions/atari.c if (!VALID_PARTITION(&rs->part[0], hd_size) && rs 60 block/partitions/atari.c !VALID_PARTITION(&rs->part[1], hd_size) && rs 61 block/partitions/atari.c !VALID_PARTITION(&rs->part[2], hd_size) && rs 62 block/partitions/atari.c !VALID_PARTITION(&rs->part[3], hd_size)) { rs 72 block/partitions/atari.c pi = &rs->part[0]; rs 74 block/partitions/atari.c for (slot = 1; pi < &rs->part[4] && slot < state->limit; slot++, pi++) { rs 135 block/partitions/atari.c pi = &rs->icdpart[0]; rs 139 block/partitions/atari.c for (; pi < &rs->icdpart[8] && slot < state->limit; slot++, pi++) { rs 594 crypto/serpent_generic.c u32 rs[4], rd[4]; rs 596 crypto/serpent_generic.c rs[0] = swab32(s[3]); rs 597 crypto/serpent_generic.c rs[1] = swab32(s[2]); rs 598 crypto/serpent_generic.c rs[2] = swab32(s[1]); rs 599 crypto/serpent_generic.c rs[3] = swab32(s[0]); rs 601 crypto/serpent_generic.c serpent_encrypt(tfm, (u8 *)rd, (u8 *)rs); rs 614 crypto/serpent_generic.c u32 rs[4], rd[4]; rs 616 crypto/serpent_generic.c rs[0] = swab32(s[3]); rs 617 crypto/serpent_generic.c rs[1] = swab32(s[2]); rs 618 crypto/serpent_generic.c rs[2] = swab32(s[1]); rs 619 crypto/serpent_generic.c rs[3] = swab32(s[0]); rs 621 crypto/serpent_generic.c serpent_decrypt(tfm, (u8 *)rd, (u8 *)rs); rs 62 drivers/auxdisplay/hd44780.c static void hd44780_write_gpio8(struct hd44780 *hd, u8 val, unsigned int rs) rs 68 drivers/auxdisplay/hd44780.c __assign_bit(8, values, rs); rs 78 drivers/auxdisplay/hd44780.c static void hd44780_write_gpio4(struct hd44780 *hd, u8 val, unsigned int rs) rs 85 drivers/auxdisplay/hd44780.c __assign_bit(4, values, rs); rs 211 drivers/auxdisplay/panel.c int rs; rs 920 drivers/auxdisplay/panel.c lcd.pins.rs = PIN_AUTOLF; rs 945 drivers/auxdisplay/panel.c lcd.pins.rs = PIN_SELECP; rs 965 drivers/auxdisplay/panel.c lcd.pins.rs = PIN_SELECP; rs 990 drivers/auxdisplay/panel.c lcd.pins.rs = lcd_rs_pin; rs 1023 drivers/auxdisplay/panel.c if (lcd.pins.rs == PIN_NOT_SET) rs 1024 drivers/auxdisplay/panel.c lcd.pins.rs = DEFAULT_LCD_PIN_RS; rs 1036 drivers/auxdisplay/panel.c if (lcd.pins.rs == PIN_NOT_SET) rs 1037 drivers/auxdisplay/panel.c lcd.pins.rs = PIN_NONE; rs 1057 drivers/auxdisplay/panel.c pin_to_bits(lcd.pins.rs, lcd_bits[LCD_PORT_D][LCD_BIT_RS], rs 1740 drivers/auxdisplay/panel.c lcd.pins.rs = lcd_rs_pin; rs 928 drivers/block/drbd/drbd_nl.c drbd_determine_dev_size(struct drbd_device *device, enum dds_flags flags, struct resize_parms *rs) __must_hold(local) rs 971 drivers/block/drbd/drbd_nl.c if (rs) { rs 973 drivers/block/drbd/drbd_nl.c md->al_stripes = rs->al_stripes; rs 974 drivers/block/drbd/drbd_nl.c md->al_stripe_size_4k = rs->al_stripe_size / 4; rs 975 drivers/block/drbd/drbd_nl.c md->al_size_4k = (u64)rs->al_stripes * rs->al_stripe_size / 4; rs 986 drivers/block/drbd/drbd_nl.c if (rs && u_size == 0) { rs 1028 drivers/block/drbd/drbd_nl.c if (la_size_changed || md_moved || rs) { rs 1061 drivers/block/drbd/drbd_nl.c if (rs) rs 2841 drivers/block/drbd/drbd_nl.c struct resize_parms rs; rs 2863 drivers/block/drbd/drbd_nl.c memset(&rs, 0, sizeof(struct resize_parms)); rs 2864 drivers/block/drbd/drbd_nl.c rs.al_stripes = device->ldev->md.al_stripes; rs 2865 drivers/block/drbd/drbd_nl.c rs.al_stripe_size = device->ldev->md.al_stripe_size_4k * 4; rs 2867 drivers/block/drbd/drbd_nl.c err = resize_parms_from_attrs(&rs, info); rs 2886 drivers/block/drbd/drbd_nl.c if (rs.no_resync && first_peer_device(device)->connection->agreed_pro_version < 93) { rs 2894 drivers/block/drbd/drbd_nl.c if (u_size != (sector_t)rs.resize_size) { rs 2902 drivers/block/drbd/drbd_nl.c if (device->ldev->md.al_stripes != rs.al_stripes || rs 2903 drivers/block/drbd/drbd_nl.c device->ldev->md.al_stripe_size_4k != rs.al_stripe_size / 4) { rs 2904 drivers/block/drbd/drbd_nl.c u32 al_size_k = rs.al_stripes * rs.al_stripe_size; rs 2916 drivers/block/drbd/drbd_nl.c if (device->state.conn != C_CONNECTED && !rs.resize_force) { rs 2931 drivers/block/drbd/drbd_nl.c new_disk_conf->disk_size = (sector_t)rs.resize_size; rs 2939 drivers/block/drbd/drbd_nl.c ddsf = (rs.resize_force ? DDSF_FORCED : 0) | (rs.no_resync ? DDSF_NO_RESYNC : 0); rs 2940 drivers/block/drbd/drbd_nl.c dd = drbd_determine_dev_size(device, ddsf, change_al_layout ? &rs : NULL); rs 37 drivers/clk/bcm/clk-cygnus.c #define RESET_VAL(o, rs, prs) { .offset = o, .reset_shift = rs, \ rs 29 drivers/clk/bcm/clk-ns2.c #define RESET_VAL(o, rs, prs) { .offset = o, .reset_shift = rs, \ rs 29 drivers/clk/bcm/clk-nsp.c #define RESET_VAL(o, rs, prs) { .offset = o, .reset_shift = rs, \ rs 21 drivers/clk/bcm/clk-sr.c #define RESET_VAL(o, rs, prs) { .offset = o, .reset_shift = rs, \ rs 24 drivers/gpu/drm/msm/disp/mdp4/mdp4_irq.c static DEFINE_RATELIMIT_STATE(rs, 5*HZ, 1); rs 29 drivers/gpu/drm/msm/disp/mdp4/mdp4_irq.c if (dumpstate && __ratelimit(&rs)) { rs 26 drivers/gpu/drm/msm/disp/mdp5/mdp5_irq.c static DEFINE_RATELIMIT_STATE(rs, 5*HZ, 1); rs 31 drivers/gpu/drm/msm/disp/mdp5/mdp5_irq.c if (dumpstate && __ratelimit(&rs)) { rs 325 drivers/gpu/drm/nouveau/dispnv04/tvmodesnv17.c uint64_t rs[] = {mode->hdisplay * id3, rs 328 drivers/gpu/drm/nouveau/dispnv04/tvmodesnv17.c do_div(rs[0], overscan * tv_norm->tv_enc_mode.hdisplay); rs 329 drivers/gpu/drm/nouveau/dispnv04/tvmodesnv17.c do_div(rs[1], overscan * tv_norm->tv_enc_mode.vdisplay); rs 332 drivers/gpu/drm/nouveau/dispnv04/tvmodesnv17.c rs[k] = max((int64_t)rs[k], id2); rs 341 drivers/gpu/drm/nouveau/dispnv04/tvmodesnv17.c p->ki3r*i*i*i) * rs[k] rs 345 drivers/gpu/drm/nouveau/dispnv04/tvmodesnv17.c p->ki3rf*i*i*i) * flicker * rs[k]; rs 384 drivers/gpu/drm/tiny/ili9225.c struct gpio_desc *rs; rs 408 drivers/gpu/drm/tiny/ili9225.c rs = devm_gpiod_get(dev, "rs", GPIOD_OUT_LOW); rs 409 drivers/gpu/drm/tiny/ili9225.c if (IS_ERR(rs)) { rs 411 drivers/gpu/drm/tiny/ili9225.c return PTR_ERR(rs); rs 416 drivers/gpu/drm/tiny/ili9225.c ret = mipi_dbi_spi_init(spi, dbi, rs); rs 333 drivers/hid/intel-ish-hid/ishtp/hbm.c struct hbm_client_connect_response *rs) rs 340 drivers/hid/intel-ish-hid/ishtp/hbm.c if (!rs->status && ishtp_hbm_cl_addr_equal(cl, rs)) { rs 378 drivers/hid/intel-ish-hid/ishtp/hbm.c struct hbm_client_connect_response *rs) rs 385 drivers/hid/intel-ish-hid/ishtp/hbm.c if (ishtp_hbm_cl_addr_equal(cl, rs)) { rs 386 drivers/hid/intel-ish-hid/ishtp/hbm.c if (!rs->status) { rs 206 drivers/iommu/amd_iommu.c ratelimit_default_init(&dev_data->rs); rs 545 drivers/iommu/amd_iommu.c if (dev_data && __ratelimit(&dev_data->rs)) { rs 656 drivers/iommu/amd_iommu_types.h struct ratelimit_state rs; /* Ratelimit IOPF messages */ rs 1714 drivers/iommu/dmar.c static DEFINE_RATELIMIT_STATE(rs, rs 1720 drivers/iommu/dmar.c if (fault_status && __ratelimit(&rs)) rs 1731 drivers/iommu/dmar.c bool ratelimited = !__ratelimit(&rs); rs 296 drivers/md/dm-clone-target.c unsigned long *rs, unsigned long *nr_regions) rs 300 drivers/md/dm-clone-target.c *rs = dm_sector_div_up(bio->bi_iter.bi_sector, clone->region_size); rs 303 drivers/md/dm-clone-target.c if (*rs >= end) rs 306 drivers/md/dm-clone-target.c *nr_regions = end - *rs; rs 464 drivers/md/dm-clone-target.c unsigned long rs, nr_regions; rs 473 drivers/md/dm-clone-target.c bio_region_range(clone, bio, &rs, &nr_regions); rs 474 drivers/md/dm-clone-target.c trim_bio(bio, region_to_sector(clone, rs), rs 483 drivers/md/dm-clone-target.c unsigned long rs, nr_regions; rs 485 drivers/md/dm-clone-target.c bio_region_range(clone, bio, &rs, &nr_regions); rs 491 drivers/md/dm-clone-target.c if (WARN_ON(rs >= clone->nr_regions || (rs + nr_regions) < rs || rs 492 drivers/md/dm-clone-target.c (rs + nr_regions) > clone->nr_regions)) { rs 494 drivers/md/dm-clone-target.c clone_device_name(clone), rs, nr_regions, rs 506 drivers/md/dm-clone-target.c if (dm_clone_is_range_hydrated(clone->cmd, rs, nr_regions)) { rs 1187 drivers/md/dm-clone-target.c unsigned long rs, nr_regions; rs 1203 drivers/md/dm-clone-target.c bio_region_range(clone, bio, &rs, &nr_regions); rs 1209 drivers/md/dm-clone-target.c r = dm_clone_cond_set_range(clone->cmd, rs, nr_regions); rs 254 drivers/md/dm-raid.c static void rs_config_backup(struct raid_set *rs, struct rs_layout *l) rs 256 drivers/md/dm-raid.c struct mddev *mddev = &rs->md; rs 263 drivers/md/dm-raid.c static void rs_config_restore(struct raid_set *rs, struct rs_layout *l) rs 265 drivers/md/dm-raid.c struct mddev *mddev = &rs->md; rs 394 drivers/md/dm-raid.c static bool rs_is_raid0(struct raid_set *rs) rs 396 drivers/md/dm-raid.c return !rs->md.level; rs 400 drivers/md/dm-raid.c static bool rs_is_raid1(struct raid_set *rs) rs 402 drivers/md/dm-raid.c return rs->md.level == 1; rs 406 drivers/md/dm-raid.c static bool rs_is_raid10(struct raid_set *rs) rs 408 drivers/md/dm-raid.c return rs->md.level == 10; rs 412 drivers/md/dm-raid.c static bool rs_is_raid6(struct raid_set *rs) rs 414 drivers/md/dm-raid.c return rs->md.level == 6; rs 418 drivers/md/dm-raid.c static bool rs_is_raid456(struct raid_set *rs) rs 420 drivers/md/dm-raid.c return __within_range(rs->md.level, 4, 6); rs 425 drivers/md/dm-raid.c static bool rs_is_reshapable(struct raid_set *rs) rs 427 drivers/md/dm-raid.c return rs_is_raid456(rs) || rs 428 drivers/md/dm-raid.c (rs_is_raid10(rs) && !__is_raid10_far(rs->md.new_layout)); rs 432 drivers/md/dm-raid.c static bool rs_is_recovering(struct raid_set *rs) rs 434 drivers/md/dm-raid.c return rs->md.recovery_cp < rs->md.dev_sectors; rs 438 drivers/md/dm-raid.c static bool rs_is_reshaping(struct raid_set *rs) rs 440 drivers/md/dm-raid.c return rs->md.reshape_position != MaxSector; rs 485 drivers/md/dm-raid.c static unsigned long __valid_flags(struct raid_set *rs) rs 487 drivers/md/dm-raid.c if (rt_is_raid0(rs->raid_type)) rs 489 drivers/md/dm-raid.c else if (rt_is_raid1(rs->raid_type)) rs 491 drivers/md/dm-raid.c else if (rt_is_raid10(rs->raid_type)) rs 493 drivers/md/dm-raid.c else if (rt_is_raid45(rs->raid_type)) rs 495 drivers/md/dm-raid.c else if (rt_is_raid6(rs->raid_type)) rs 506 drivers/md/dm-raid.c static int rs_check_for_valid_flags(struct raid_set *rs) rs 508 drivers/md/dm-raid.c if (rs->ctr_flags & ~__valid_flags(rs)) { rs 509 drivers/md/dm-raid.c rs->ti->error = "Invalid flags combination"; rs 593 drivers/md/dm-raid.c static int raid10_format_to_md_layout(struct raid_set *rs, rs 614 drivers/md/dm-raid.c if (!test_bit(__CTR_FLAG_RAID10_USE_NEAR_SETS, &rs->ctr_flags)) rs 620 drivers/md/dm-raid.c if (!test_bit(__CTR_FLAG_RAID10_USE_NEAR_SETS, &rs->ctr_flags)) rs 678 drivers/md/dm-raid.c static void rs_set_rdev_sectors(struct raid_set *rs) rs 680 drivers/md/dm-raid.c struct mddev *mddev = &rs->md; rs 695 drivers/md/dm-raid.c static void rs_set_capacity(struct raid_set *rs) rs 697 drivers/md/dm-raid.c struct gendisk *gendisk = dm_disk(dm_table_get_md(rs->ti->table)); rs 699 drivers/md/dm-raid.c set_capacity(gendisk, rs->md.array_sectors); rs 707 drivers/md/dm-raid.c static void rs_set_cur(struct raid_set *rs) rs 709 drivers/md/dm-raid.c struct mddev *mddev = &rs->md; rs 720 drivers/md/dm-raid.c static void rs_set_new(struct raid_set *rs) rs 722 drivers/md/dm-raid.c struct mddev *mddev = &rs->md; rs 727 drivers/md/dm-raid.c mddev->raid_disks = rs->raid_disks; rs 735 drivers/md/dm-raid.c struct raid_set *rs; rs 742 drivers/md/dm-raid.c rs = kzalloc(struct_size(rs, dev, raid_devs), GFP_KERNEL); rs 743 drivers/md/dm-raid.c if (!rs) { rs 748 drivers/md/dm-raid.c mddev_init(&rs->md); rs 750 drivers/md/dm-raid.c rs->raid_disks = raid_devs; rs 751 drivers/md/dm-raid.c rs->delta_disks = 0; rs 753 drivers/md/dm-raid.c rs->ti = ti; rs 754 drivers/md/dm-raid.c rs->raid_type = raid_type; rs 755 drivers/md/dm-raid.c rs->stripe_cache_entries = 256; rs 756 drivers/md/dm-raid.c rs->md.raid_disks = raid_devs; rs 757 drivers/md/dm-raid.c rs->md.level = raid_type->level; rs 758 drivers/md/dm-raid.c rs->md.new_level = rs->md.level; rs 759 drivers/md/dm-raid.c rs->md.layout = raid_type->algorithm; rs 760 drivers/md/dm-raid.c rs->md.new_layout = rs->md.layout; rs 761 drivers/md/dm-raid.c rs->md.delta_disks = 0; rs 762 drivers/md/dm-raid.c rs->md.recovery_cp = MaxSector; rs 765 drivers/md/dm-raid.c md_rdev_init(&rs->dev[i].rdev); rs 776 drivers/md/dm-raid.c return rs; rs 780 drivers/md/dm-raid.c static void raid_set_free(struct raid_set *rs) rs 784 drivers/md/dm-raid.c if (rs->journal_dev.dev) { rs 785 drivers/md/dm-raid.c md_rdev_clear(&rs->journal_dev.rdev); rs 786 drivers/md/dm-raid.c dm_put_device(rs->ti, rs->journal_dev.dev); rs 789 drivers/md/dm-raid.c for (i = 0; i < rs->raid_disks; i++) { rs 790 drivers/md/dm-raid.c if (rs->dev[i].meta_dev) rs 791 drivers/md/dm-raid.c dm_put_device(rs->ti, rs->dev[i].meta_dev); rs 792 drivers/md/dm-raid.c md_rdev_clear(&rs->dev[i].rdev); rs 793 drivers/md/dm-raid.c if (rs->dev[i].data_dev) rs 794 drivers/md/dm-raid.c dm_put_device(rs->ti, rs->dev[i].data_dev); rs 797 drivers/md/dm-raid.c kfree(rs); rs 816 drivers/md/dm-raid.c static int parse_dev_params(struct raid_set *rs, struct dm_arg_set *as) rs 829 drivers/md/dm-raid.c for (i = 0; i < rs->raid_disks; i++) { rs 830 drivers/md/dm-raid.c rs->dev[i].rdev.raid_disk = i; rs 832 drivers/md/dm-raid.c rs->dev[i].meta_dev = NULL; rs 833 drivers/md/dm-raid.c rs->dev[i].data_dev = NULL; rs 839 drivers/md/dm-raid.c rs->dev[i].rdev.data_offset = 0; rs 840 drivers/md/dm-raid.c rs->dev[i].rdev.new_data_offset = 0; rs 841 drivers/md/dm-raid.c rs->dev[i].rdev.mddev = &rs->md; rs 848 drivers/md/dm-raid.c r = dm_get_device(rs->ti, arg, dm_table_get_mode(rs->ti->table), rs 849 drivers/md/dm-raid.c &rs->dev[i].meta_dev); rs 851 drivers/md/dm-raid.c rs->ti->error = "RAID metadata device lookup failure"; rs 855 drivers/md/dm-raid.c rs->dev[i].rdev.sb_page = alloc_page(GFP_KERNEL); rs 856 drivers/md/dm-raid.c if (!rs->dev[i].rdev.sb_page) { rs 857 drivers/md/dm-raid.c rs->ti->error = "Failed to allocate superblock page"; rs 867 drivers/md/dm-raid.c if (!test_bit(In_sync, &rs->dev[i].rdev.flags) && rs 868 drivers/md/dm-raid.c (!rs->dev[i].rdev.recovery_offset)) { rs 869 drivers/md/dm-raid.c rs->ti->error = "Drive designated for rebuild not specified"; rs 873 drivers/md/dm-raid.c if (rs->dev[i].meta_dev) { rs 874 drivers/md/dm-raid.c rs->ti->error = "No data device supplied with metadata device"; rs 881 drivers/md/dm-raid.c r = dm_get_device(rs->ti, arg, dm_table_get_mode(rs->ti->table), rs 882 drivers/md/dm-raid.c &rs->dev[i].data_dev); rs 884 drivers/md/dm-raid.c rs->ti->error = "RAID device lookup failure"; rs 888 drivers/md/dm-raid.c if (rs->dev[i].meta_dev) { rs 890 drivers/md/dm-raid.c rs->dev[i].rdev.meta_bdev = rs->dev[i].meta_dev->bdev; rs 892 drivers/md/dm-raid.c rs->dev[i].rdev.bdev = rs->dev[i].data_dev->bdev; rs 893 drivers/md/dm-raid.c list_add_tail(&rs->dev[i].rdev.same_set, &rs->md.disks); rs 894 drivers/md/dm-raid.c if (!test_bit(In_sync, &rs->dev[i].rdev.flags)) rs 898 drivers/md/dm-raid.c if (rs->journal_dev.dev) rs 899 drivers/md/dm-raid.c list_add_tail(&rs->journal_dev.rdev.same_set, &rs->md.disks); rs 902 drivers/md/dm-raid.c rs->md.external = 0; rs 903 drivers/md/dm-raid.c rs->md.persistent = 1; rs 904 drivers/md/dm-raid.c rs->md.major_version = 2; rs 905 drivers/md/dm-raid.c } else if (rebuild && !rs->md.recovery_cp) { rs 917 drivers/md/dm-raid.c rs->ti->error = "Unable to rebuild drive while array is not in-sync"; rs 934 drivers/md/dm-raid.c static int validate_region_size(struct raid_set *rs, unsigned long region_size) rs 936 drivers/md/dm-raid.c unsigned long min_region_size = rs->ti->len / (1 << 21); rs 938 drivers/md/dm-raid.c if (rs_is_raid0(rs)) rs 958 drivers/md/dm-raid.c if (region_size > rs->ti->len) { rs 959 drivers/md/dm-raid.c rs->ti->error = "Supplied region size is too large"; rs 966 drivers/md/dm-raid.c rs->ti->error = "Supplied region size is too small"; rs 971 drivers/md/dm-raid.c rs->ti->error = "Region size is not a power of 2"; rs 975 drivers/md/dm-raid.c if (region_size < rs->md.chunk_sectors) { rs 976 drivers/md/dm-raid.c rs->ti->error = "Region size is smaller than the chunk size"; rs 984 drivers/md/dm-raid.c rs->md.bitmap_info.chunksize = to_bytes(region_size); rs 998 drivers/md/dm-raid.c static int validate_raid_redundancy(struct raid_set *rs) rs 1004 drivers/md/dm-raid.c for (i = 0; i < rs->md.raid_disks; i++) rs 1005 drivers/md/dm-raid.c if (!test_bit(In_sync, &rs->dev[i].rdev.flags) || rs 1006 drivers/md/dm-raid.c !rs->dev[i].rdev.sb_page) rs 1009 drivers/md/dm-raid.c switch (rs->md.level) { rs 1013 drivers/md/dm-raid.c if (rebuild_cnt >= rs->md.raid_disks) rs 1019 drivers/md/dm-raid.c if (rebuild_cnt > rs->raid_type->parity_devs) rs 1023 drivers/md/dm-raid.c copies = raid10_md_layout_to_copies(rs->md.new_layout); rs 1046 drivers/md/dm-raid.c if (__is_raid10_near(rs->md.new_layout)) { rs 1047 drivers/md/dm-raid.c for (i = 0; i < rs->md.raid_disks; i++) { rs 1050 drivers/md/dm-raid.c if ((!rs->dev[i].rdev.sb_page || rs 1051 drivers/md/dm-raid.c !test_bit(In_sync, &rs->dev[i].rdev.flags)) && rs 1070 drivers/md/dm-raid.c group_size = (rs->md.raid_disks / copies); rs 1071 drivers/md/dm-raid.c last_group_start = (rs->md.raid_disks / group_size) - 1; rs 1073 drivers/md/dm-raid.c for (i = 0; i < rs->md.raid_disks; i++) { rs 1076 drivers/md/dm-raid.c if ((!rs->dev[i].rdev.sb_page || rs 1077 drivers/md/dm-raid.c !test_bit(In_sync, &rs->dev[i].rdev.flags)) && rs 1118 drivers/md/dm-raid.c static int parse_raid_params(struct raid_set *rs, struct dm_arg_set *as, rs 1128 drivers/md/dm-raid.c struct raid_type *rt = rs->raid_type; rs 1134 drivers/md/dm-raid.c rs->ti->error = "Bad numerical argument given for chunk_size"; rs 1147 drivers/md/dm-raid.c rs->ti->error = "Chunk size must be a power of 2"; rs 1150 drivers/md/dm-raid.c rs->ti->error = "Chunk size value is too small"; rs 1154 drivers/md/dm-raid.c rs->md.new_chunk_sectors = rs->md.chunk_sectors = value; rs 1173 drivers/md/dm-raid.c for (i = 0; i < rs->raid_disks; i++) { rs 1174 drivers/md/dm-raid.c set_bit(In_sync, &rs->dev[i].rdev.flags); rs 1175 drivers/md/dm-raid.c rs->dev[i].rdev.recovery_offset = MaxSector; rs 1184 drivers/md/dm-raid.c rs->ti->error = "Not enough raid parameters given"; rs 1189 drivers/md/dm-raid.c if (test_and_set_bit(__CTR_FLAG_NOSYNC, &rs->ctr_flags)) { rs 1190 drivers/md/dm-raid.c rs->ti->error = "Only one 'nosync' argument allowed"; rs 1196 drivers/md/dm-raid.c if (test_and_set_bit(__CTR_FLAG_SYNC, &rs->ctr_flags)) { rs 1197 drivers/md/dm-raid.c rs->ti->error = "Only one 'sync' argument allowed"; rs 1203 drivers/md/dm-raid.c if (test_and_set_bit(__CTR_FLAG_RAID10_USE_NEAR_SETS, &rs->ctr_flags)) { rs 1204 drivers/md/dm-raid.c rs->ti->error = "Only one 'raid10_use_new_sets' argument allowed"; rs 1213 drivers/md/dm-raid.c rs->ti->error = "Wrong number of raid parameters given"; rs 1222 drivers/md/dm-raid.c if (test_and_set_bit(__CTR_FLAG_RAID10_FORMAT, &rs->ctr_flags)) { rs 1223 drivers/md/dm-raid.c rs->ti->error = "Only one 'raid10_format' argument pair allowed"; rs 1227 drivers/md/dm-raid.c rs->ti->error = "'raid10_format' is an invalid parameter for this RAID type"; rs 1232 drivers/md/dm-raid.c rs->ti->error = "Invalid 'raid10_format' value given"; rs 1243 drivers/md/dm-raid.c if (test_and_set_bit(__CTR_FLAG_JOURNAL_DEV, &rs->ctr_flags)) { rs 1244 drivers/md/dm-raid.c rs->ti->error = "Only one raid4/5/6 set journaling device allowed"; rs 1248 drivers/md/dm-raid.c rs->ti->error = "'journal_dev' is an invalid parameter for this RAID type"; rs 1251 drivers/md/dm-raid.c r = dm_get_device(rs->ti, arg, dm_table_get_mode(rs->ti->table), rs 1252 drivers/md/dm-raid.c &rs->journal_dev.dev); rs 1254 drivers/md/dm-raid.c rs->ti->error = "raid4/5/6 journal device lookup failure"; rs 1257 drivers/md/dm-raid.c jdev = &rs->journal_dev.rdev; rs 1259 drivers/md/dm-raid.c jdev->mddev = &rs->md; rs 1260 drivers/md/dm-raid.c jdev->bdev = rs->journal_dev.dev->bdev; rs 1263 drivers/md/dm-raid.c rs->ti->error = "No space for raid4/5/6 journal"; rs 1266 drivers/md/dm-raid.c rs->journal_dev.mode = R5C_JOURNAL_MODE_WRITE_THROUGH; rs 1275 drivers/md/dm-raid.c if (!test_bit(__CTR_FLAG_JOURNAL_DEV, &rs->ctr_flags)) { rs 1276 drivers/md/dm-raid.c rs->ti->error = "raid4/5/6 'journal_mode' is invalid without 'journal_dev'"; rs 1279 drivers/md/dm-raid.c if (test_and_set_bit(__CTR_FLAG_JOURNAL_MODE, &rs->ctr_flags)) { rs 1280 drivers/md/dm-raid.c rs->ti->error = "Only one raid4/5/6 'journal_mode' argument allowed"; rs 1285 drivers/md/dm-raid.c rs->ti->error = "Invalid 'journal_mode' argument"; rs 1288 drivers/md/dm-raid.c rs->journal_dev.mode = r; rs 1296 drivers/md/dm-raid.c rs->ti->error = "Bad numerical argument given in raid params"; rs 1306 drivers/md/dm-raid.c if (!__within_range(value, 0, rs->raid_disks - 1)) { rs 1307 drivers/md/dm-raid.c rs->ti->error = "Invalid rebuild index given"; rs 1311 drivers/md/dm-raid.c if (test_and_set_bit(value, (void *) rs->rebuild_disks)) { rs 1312 drivers/md/dm-raid.c rs->ti->error = "rebuild for this index already given"; rs 1316 drivers/md/dm-raid.c rd = rs->dev + value; rs 1320 drivers/md/dm-raid.c set_bit(__CTR_FLAG_REBUILD, &rs->ctr_flags); rs 1323 drivers/md/dm-raid.c rs->ti->error = "write_mostly option is only valid for RAID1"; rs 1327 drivers/md/dm-raid.c if (!__within_range(value, 0, rs->md.raid_disks - 1)) { rs 1328 drivers/md/dm-raid.c rs->ti->error = "Invalid write_mostly index given"; rs 1333 drivers/md/dm-raid.c set_bit(WriteMostly, &rs->dev[value].rdev.flags); rs 1334 drivers/md/dm-raid.c set_bit(__CTR_FLAG_WRITE_MOSTLY, &rs->ctr_flags); rs 1337 drivers/md/dm-raid.c rs->ti->error = "max_write_behind option is only valid for RAID1"; rs 1341 drivers/md/dm-raid.c if (test_and_set_bit(__CTR_FLAG_MAX_WRITE_BEHIND, &rs->ctr_flags)) { rs 1342 drivers/md/dm-raid.c rs->ti->error = "Only one max_write_behind argument pair allowed"; rs 1351 drivers/md/dm-raid.c rs->ti->error = "Max write-behind limit out of range"; rs 1355 drivers/md/dm-raid.c rs->md.bitmap_info.max_write_behind = value / 2; rs 1357 drivers/md/dm-raid.c if (test_and_set_bit(__CTR_FLAG_DAEMON_SLEEP, &rs->ctr_flags)) { rs 1358 drivers/md/dm-raid.c rs->ti->error = "Only one daemon_sleep argument pair allowed"; rs 1362 drivers/md/dm-raid.c rs->ti->error = "daemon sleep period out of range"; rs 1365 drivers/md/dm-raid.c rs->md.bitmap_info.daemon_sleep = value; rs 1368 drivers/md/dm-raid.c if (test_and_set_bit(__CTR_FLAG_DATA_OFFSET, &rs->ctr_flags)) { rs 1369 drivers/md/dm-raid.c rs->ti->error = "Only one data_offset argument pair allowed"; rs 1375 drivers/md/dm-raid.c rs->ti->error = "Bogus data_offset value"; rs 1378 drivers/md/dm-raid.c rs->data_offset = value; rs 1381 drivers/md/dm-raid.c if (test_and_set_bit(__CTR_FLAG_DELTA_DISKS, &rs->ctr_flags)) { rs 1382 drivers/md/dm-raid.c rs->ti->error = "Only one delta_disks argument pair allowed"; rs 1387 drivers/md/dm-raid.c rs->ti->error = "Too many delta_disk requested"; rs 1391 drivers/md/dm-raid.c rs->delta_disks = value; rs 1393 drivers/md/dm-raid.c if (test_and_set_bit(__CTR_FLAG_STRIPE_CACHE, &rs->ctr_flags)) { rs 1394 drivers/md/dm-raid.c rs->ti->error = "Only one stripe_cache argument pair allowed"; rs 1399 drivers/md/dm-raid.c rs->ti->error = "Inappropriate argument: stripe_cache"; rs 1404 drivers/md/dm-raid.c rs->ti->error = "Bogus stripe cache entries value"; rs 1407 drivers/md/dm-raid.c rs->stripe_cache_entries = value; rs 1409 drivers/md/dm-raid.c if (test_and_set_bit(__CTR_FLAG_MIN_RECOVERY_RATE, &rs->ctr_flags)) { rs 1410 drivers/md/dm-raid.c rs->ti->error = "Only one min_recovery_rate argument pair allowed"; rs 1415 drivers/md/dm-raid.c rs->ti->error = "min_recovery_rate out of range"; rs 1418 drivers/md/dm-raid.c rs->md.sync_speed_min = value; rs 1420 drivers/md/dm-raid.c if (test_and_set_bit(__CTR_FLAG_MAX_RECOVERY_RATE, &rs->ctr_flags)) { rs 1421 drivers/md/dm-raid.c rs->ti->error = "Only one max_recovery_rate argument pair allowed"; rs 1426 drivers/md/dm-raid.c rs->ti->error = "max_recovery_rate out of range"; rs 1429 drivers/md/dm-raid.c rs->md.sync_speed_max = value; rs 1431 drivers/md/dm-raid.c if (test_and_set_bit(__CTR_FLAG_REGION_SIZE, &rs->ctr_flags)) { rs 1432 drivers/md/dm-raid.c rs->ti->error = "Only one region_size argument pair allowed"; rs 1437 drivers/md/dm-raid.c rs->requested_bitmap_chunk_sectors = value; rs 1439 drivers/md/dm-raid.c if (test_and_set_bit(__CTR_FLAG_RAID10_COPIES, &rs->ctr_flags)) { rs 1440 drivers/md/dm-raid.c rs->ti->error = "Only one raid10_copies argument pair allowed"; rs 1444 drivers/md/dm-raid.c if (!__within_range(value, 2, rs->md.raid_disks)) { rs 1445 drivers/md/dm-raid.c rs->ti->error = "Bad value for 'raid10_copies'"; rs 1452 drivers/md/dm-raid.c rs->ti->error = "Unable to parse RAID parameter"; rs 1457 drivers/md/dm-raid.c if (test_bit(__CTR_FLAG_SYNC, &rs->ctr_flags) && rs 1458 drivers/md/dm-raid.c test_bit(__CTR_FLAG_NOSYNC, &rs->ctr_flags)) { rs 1459 drivers/md/dm-raid.c rs->ti->error = "sync and nosync are mutually exclusive"; rs 1463 drivers/md/dm-raid.c if (test_bit(__CTR_FLAG_REBUILD, &rs->ctr_flags) && rs 1464 drivers/md/dm-raid.c (test_bit(__CTR_FLAG_SYNC, &rs->ctr_flags) || rs 1465 drivers/md/dm-raid.c test_bit(__CTR_FLAG_NOSYNC, &rs->ctr_flags))) { rs 1466 drivers/md/dm-raid.c rs->ti->error = "sync/nosync and rebuild are mutually exclusive"; rs 1470 drivers/md/dm-raid.c if (write_mostly >= rs->md.raid_disks) { rs 1471 drivers/md/dm-raid.c rs->ti->error = "Can't set all raid1 devices to write_mostly"; rs 1475 drivers/md/dm-raid.c if (rs->md.sync_speed_max && rs 1476 drivers/md/dm-raid.c rs->md.sync_speed_min > rs->md.sync_speed_max) { rs 1477 drivers/md/dm-raid.c rs->ti->error = "Bogus recovery rates"; rs 1481 drivers/md/dm-raid.c if (validate_region_size(rs, region_size)) rs 1484 drivers/md/dm-raid.c if (rs->md.chunk_sectors) rs 1485 drivers/md/dm-raid.c max_io_len = rs->md.chunk_sectors; rs 1489 drivers/md/dm-raid.c if (dm_set_target_max_io_len(rs->ti, max_io_len)) rs 1493 drivers/md/dm-raid.c if (raid10_copies > rs->md.raid_disks) { rs 1494 drivers/md/dm-raid.c rs->ti->error = "Not enough devices to satisfy specification"; rs 1498 drivers/md/dm-raid.c rs->md.new_layout = raid10_format_to_md_layout(rs, raid10_format, raid10_copies); rs 1499 drivers/md/dm-raid.c if (rs->md.new_layout < 0) { rs 1500 drivers/md/dm-raid.c rs->ti->error = "Error getting raid10 format"; rs 1501 drivers/md/dm-raid.c return rs->md.new_layout; rs 1504 drivers/md/dm-raid.c rt = get_raid_type_by_ll(10, rs->md.new_layout); rs 1506 drivers/md/dm-raid.c rs->ti->error = "Failed to recognize new raid10 layout"; rs 1512 drivers/md/dm-raid.c test_bit(__CTR_FLAG_RAID10_USE_NEAR_SETS, &rs->ctr_flags)) { rs 1513 drivers/md/dm-raid.c rs->ti->error = "RAID10 format 'near' and 'raid10_use_near_sets' are incompatible"; rs 1518 drivers/md/dm-raid.c rs->raid10_copies = raid10_copies; rs 1521 drivers/md/dm-raid.c rs->md.persistent = 0; rs 1522 drivers/md/dm-raid.c rs->md.external = 1; rs 1525 drivers/md/dm-raid.c return rs_check_for_valid_flags(rs); rs 1529 drivers/md/dm-raid.c static int rs_set_raid456_stripe_cache(struct raid_set *rs) rs 1533 drivers/md/dm-raid.c struct mddev *mddev = &rs->md; rs 1535 drivers/md/dm-raid.c uint32_t nr_stripes = rs->stripe_cache_entries; rs 1537 drivers/md/dm-raid.c if (!rt_is_raid456(rs->raid_type)) { rs 1538 drivers/md/dm-raid.c rs->ti->error = "Inappropriate raid level; cannot change stripe_cache size"; rs 1550 drivers/md/dm-raid.c rs->ti->error = "Cannot change stripe_cache size on inactive RAID set"; rs 1558 drivers/md/dm-raid.c rs->ti->error = "Failed to set raid4/5/6 stripe cache size"; rs 1569 drivers/md/dm-raid.c static unsigned int mddev_data_stripes(struct raid_set *rs) rs 1571 drivers/md/dm-raid.c return rs->md.raid_disks - rs->raid_type->parity_devs; rs 1575 drivers/md/dm-raid.c static unsigned int rs_data_stripes(struct raid_set *rs) rs 1577 drivers/md/dm-raid.c return rs->raid_disks - rs->raid_type->parity_devs; rs 1584 drivers/md/dm-raid.c static sector_t __rdev_sectors(struct raid_set *rs) rs 1588 drivers/md/dm-raid.c for (i = 0; i < rs->md.raid_disks; i++) { rs 1589 drivers/md/dm-raid.c struct md_rdev *rdev = &rs->dev[i].rdev; rs 1600 drivers/md/dm-raid.c static int _check_data_dev_sectors(struct raid_set *rs) rs 1605 drivers/md/dm-raid.c rdev_for_each(rdev, &rs->md) rs 1608 drivers/md/dm-raid.c if (ds < rs->md.dev_sectors) { rs 1609 drivers/md/dm-raid.c rs->ti->error = "Component device(s) too small"; rs 1618 drivers/md/dm-raid.c static int rs_set_dev_and_array_sectors(struct raid_set *rs, bool use_mddev) rs 1622 drivers/md/dm-raid.c struct mddev *mddev = &rs->md; rs 1624 drivers/md/dm-raid.c sector_t array_sectors = rs->ti->len, dev_sectors = rs->ti->len; rs 1628 drivers/md/dm-raid.c data_stripes = mddev_data_stripes(rs); rs 1630 drivers/md/dm-raid.c delta_disks = rs->delta_disks; rs 1631 drivers/md/dm-raid.c data_stripes = rs_data_stripes(rs); rs 1635 drivers/md/dm-raid.c if (rt_is_raid1(rs->raid_type)) rs 1637 drivers/md/dm-raid.c else if (rt_is_raid10(rs->raid_type)) { rs 1638 drivers/md/dm-raid.c if (rs->raid10_copies < 2 || rs 1640 drivers/md/dm-raid.c rs->ti->error = "Bogus raid10 data copies or delta disks"; rs 1644 drivers/md/dm-raid.c dev_sectors *= rs->raid10_copies; rs 1649 drivers/md/dm-raid.c if (sector_div(array_sectors, rs->raid10_copies)) rs 1666 drivers/md/dm-raid.c return _check_data_dev_sectors(rs); rs 1668 drivers/md/dm-raid.c rs->ti->error = "Target length not divisible by number of data devices"; rs 1673 drivers/md/dm-raid.c static void __rs_setup_recovery(struct raid_set *rs, sector_t dev_sectors) rs 1676 drivers/md/dm-raid.c if (rs_is_raid0(rs)) rs 1677 drivers/md/dm-raid.c rs->md.recovery_cp = MaxSector; rs 1683 drivers/md/dm-raid.c else if (rs_is_raid6(rs)) rs 1684 drivers/md/dm-raid.c rs->md.recovery_cp = dev_sectors; rs 1690 drivers/md/dm-raid.c rs->md.recovery_cp = test_bit(__CTR_FLAG_NOSYNC, &rs->ctr_flags) rs 1695 drivers/md/dm-raid.c static void rs_setup_recovery(struct raid_set *rs, sector_t dev_sectors) rs 1699 drivers/md/dm-raid.c __rs_setup_recovery(rs, 0); rs 1702 drivers/md/dm-raid.c __rs_setup_recovery(rs, MaxSector); rs 1703 drivers/md/dm-raid.c else if (__rdev_sectors(rs) < dev_sectors) rs 1705 drivers/md/dm-raid.c __rs_setup_recovery(rs, __rdev_sectors(rs)); rs 1707 drivers/md/dm-raid.c __rs_setup_recovery(rs, MaxSector); rs 1712 drivers/md/dm-raid.c struct raid_set *rs = container_of(ws, struct raid_set, md.event_work); rs 1715 drivers/md/dm-raid.c if (!rs_is_reshaping(rs)) { rs 1716 drivers/md/dm-raid.c if (rs_is_raid10(rs)) rs 1717 drivers/md/dm-raid.c rs_set_rdev_sectors(rs); rs 1718 drivers/md/dm-raid.c rs_set_capacity(rs); rs 1720 drivers/md/dm-raid.c dm_table_event(rs->ti->table); rs 1725 drivers/md/dm-raid.c struct raid_set *rs = container_of(cb, struct raid_set, callbacks); rs 1727 drivers/md/dm-raid.c return mddev_congested(&rs->md, bits); rs 1736 drivers/md/dm-raid.c static int rs_check_takeover(struct raid_set *rs) rs 1738 drivers/md/dm-raid.c struct mddev *mddev = &rs->md; rs 1741 drivers/md/dm-raid.c if (rs->md.degraded) { rs 1742 drivers/md/dm-raid.c rs->ti->error = "Can't takeover degraded raid set"; rs 1746 drivers/md/dm-raid.c if (rs_is_reshaping(rs)) { rs 1747 drivers/md/dm-raid.c rs->ti->error = "Can't takeover reshaping raid set"; rs 1760 drivers/md/dm-raid.c !(rs->raid_disks % mddev->raid_disks)) rs 1885 drivers/md/dm-raid.c rs->ti->error = "takeover not possible"; rs 1890 drivers/md/dm-raid.c static bool rs_takeover_requested(struct raid_set *rs) rs 1892 drivers/md/dm-raid.c return rs->md.new_level != rs->md.level; rs 1896 drivers/md/dm-raid.c static bool rs_reshape_requested(struct raid_set *rs) rs 1899 drivers/md/dm-raid.c struct mddev *mddev = &rs->md; rs 1901 drivers/md/dm-raid.c if (rs_takeover_requested(rs)) rs 1904 drivers/md/dm-raid.c if (rs_is_raid0(rs)) rs 1909 drivers/md/dm-raid.c rs->delta_disks; rs 1912 drivers/md/dm-raid.c if (rs_is_raid1(rs)) { rs 1913 drivers/md/dm-raid.c if (rs->delta_disks) rs 1914 drivers/md/dm-raid.c return !!rs->delta_disks; rs 1917 drivers/md/dm-raid.c mddev->raid_disks != rs->raid_disks; rs 1920 drivers/md/dm-raid.c if (rs_is_raid10(rs)) rs 1923 drivers/md/dm-raid.c rs->delta_disks >= 0; rs 2027 drivers/md/dm-raid.c static int rs_check_reshape(struct raid_set *rs) rs 2029 drivers/md/dm-raid.c struct mddev *mddev = &rs->md; rs 2032 drivers/md/dm-raid.c rs->ti->error = "Reshape not supported"; rs 2034 drivers/md/dm-raid.c rs->ti->error = "Can't reshape degraded raid set"; rs 2035 drivers/md/dm-raid.c else if (rs_is_recovering(rs)) rs 2036 drivers/md/dm-raid.c rs->ti->error = "Convert request on recovering raid set prohibited"; rs 2037 drivers/md/dm-raid.c else if (rs_is_reshaping(rs)) rs 2038 drivers/md/dm-raid.c rs->ti->error = "raid set already reshaping!"; rs 2039 drivers/md/dm-raid.c else if (!(rs_is_raid1(rs) || rs_is_raid10(rs) || rs_is_raid456(rs))) rs 2040 drivers/md/dm-raid.c rs->ti->error = "Reshaping only supported for raid1/4/5/6/10"; rs 2102 drivers/md/dm-raid.c struct raid_set *rs = container_of(mddev, struct raid_set, md); rs 2114 drivers/md/dm-raid.c for (i = 0; i < rs->raid_disks; i++) rs 2115 drivers/md/dm-raid.c if (!rs->dev[i].data_dev || test_bit(Faulty, &rs->dev[i].rdev.flags)) { rs 2223 drivers/md/dm-raid.c static int super_init_validation(struct raid_set *rs, struct md_rdev *rdev) rs 2227 drivers/md/dm-raid.c struct mddev *mddev = &rs->md; rs 2264 drivers/md/dm-raid.c if (test_bit(__CTR_FLAG_DELTA_DISKS, &rs->ctr_flags)) { rs 2276 drivers/md/dm-raid.c rs->raid_type = get_raid_type_by_ll(mddev->level, mddev->layout); rs 2286 drivers/md/dm-raid.c if (rs_takeover_requested(rs)) { rs 2293 drivers/md/dm-raid.c } else if (rs_reshape_requested(rs)) { rs 2306 drivers/md/dm-raid.c if (rs->delta_disks) rs 2308 drivers/md/dm-raid.c mddev->raid_disks, mddev->raid_disks + rs->delta_disks); rs 2309 drivers/md/dm-raid.c if (rs_is_raid10(rs)) { rs 2323 drivers/md/dm-raid.c if (!test_bit(__CTR_FLAG_NOSYNC, &rs->ctr_flags)) rs 2361 drivers/md/dm-raid.c if (new_devs == rs->raid_disks || !rebuilds) { rs 2363 drivers/md/dm-raid.c if (new_devs == 1 && !rs->delta_disks) rs 2365 drivers/md/dm-raid.c if (new_devs == rs->raid_disks) { rs 2369 drivers/md/dm-raid.c new_devs != rs->delta_disks) { rs 2384 drivers/md/dm-raid.c } else if (!test_bit(__CTR_FLAG_REBUILD, &rs->ctr_flags) && rs_is_recovering(rs)) { rs 2388 drivers/md/dm-raid.c } else if (rs_is_reshaping(rs)) { rs 2417 drivers/md/dm-raid.c if (rs_is_raid10(rs) && __is_raid10_near(mddev->layout)) { rs 2419 drivers/md/dm-raid.c rs->raid_disks % rs->raid10_copies) { rs 2420 drivers/md/dm-raid.c rs->ti->error = rs 2427 drivers/md/dm-raid.c } else if (!(rs_is_raid10(rs) && rt_is_raid0(rs->raid_type)) && rs 2428 drivers/md/dm-raid.c !(rs_is_raid0(rs) && rt_is_raid10(rs->raid_type)) && rs 2429 drivers/md/dm-raid.c !rt_is_raid1(rs->raid_type)) { rs 2430 drivers/md/dm-raid.c rs->ti->error = "Cannot change device positions in raid set"; rs 2449 drivers/md/dm-raid.c static int super_validate(struct raid_set *rs, struct md_rdev *rdev) rs 2451 drivers/md/dm-raid.c struct mddev *mddev = &rs->md; rs 2454 drivers/md/dm-raid.c if (rs_is_raid0(rs) || !rdev->sb_page || rdev->raid_disk < 0) rs 2463 drivers/md/dm-raid.c if (!mddev->events && super_init_validation(rs, rdev)) rs 2468 drivers/md/dm-raid.c rs->ti->error = "Unable to assemble array: Unknown flag(s) in compatible feature flags"; rs 2473 drivers/md/dm-raid.c rs->ti->error = "Unable to assemble array: No incompatible feature flags supported yet"; rs 2478 drivers/md/dm-raid.c mddev->bitmap_info.offset = (rt_is_raid0(rs->raid_type) || rs->journal_dev.dev) ? 0 : to_sector(4096); rs 2497 drivers/md/dm-raid.c else if (!rs_is_reshaping(rs)) rs 2520 drivers/md/dm-raid.c static int analyse_superblocks(struct dm_target *ti, struct raid_set *rs) rs 2524 drivers/md/dm-raid.c struct mddev *mddev = &rs->md; rs 2551 drivers/md/dm-raid.c if (test_bit(__CTR_FLAG_SYNC, &rs->ctr_flags)) rs 2568 drivers/md/dm-raid.c if (rs_is_raid0(rs)) rs 2591 drivers/md/dm-raid.c rs->ti->error = "Unable to assemble array: Invalid superblocks"; rs 2592 drivers/md/dm-raid.c if (super_validate(rs, freshest)) rs 2595 drivers/md/dm-raid.c if (validate_raid_redundancy(rs)) { rs 2596 drivers/md/dm-raid.c rs->ti->error = "Insufficient redundancy to activate array"; rs 2603 drivers/md/dm-raid.c super_validate(rs, rdev)) rs 2616 drivers/md/dm-raid.c static int rs_adjust_data_offsets(struct raid_set *rs) rs 2622 drivers/md/dm-raid.c if (!test_bit(__CTR_FLAG_DATA_OFFSET, &rs->ctr_flags)) { rs 2623 drivers/md/dm-raid.c if (!rs_is_reshapable(rs)) rs 2630 drivers/md/dm-raid.c rdev = &rs->dev[0].rdev; rs 2632 drivers/md/dm-raid.c if (rs->delta_disks < 0) { rs 2642 drivers/md/dm-raid.c new_data_offset = rs->data_offset; rs 2644 drivers/md/dm-raid.c } else if (rs->delta_disks > 0) { rs 2653 drivers/md/dm-raid.c data_offset = rs->data_offset; rs 2675 drivers/md/dm-raid.c data_offset = rs->data_offset ? rdev->data_offset : 0; rs 2676 drivers/md/dm-raid.c new_data_offset = data_offset ? 0 : rs->data_offset; rs 2677 drivers/md/dm-raid.c set_bit(RT_FLAG_UPDATE_SBS, &rs->runtime_flags); rs 2683 drivers/md/dm-raid.c if (rs->data_offset && rs 2684 drivers/md/dm-raid.c to_sector(i_size_read(rdev->bdev->bd_inode)) - rs->md.dev_sectors < MIN_FREE_RESHAPE_SPACE) { rs 2685 drivers/md/dm-raid.c rs->ti->error = data_offset ? "No space for forward reshape" : rs 2694 drivers/md/dm-raid.c if (rs->md.recovery_cp < rs->md.dev_sectors) rs 2695 drivers/md/dm-raid.c rs->md.recovery_cp += rs->dev[0].rdev.data_offset; rs 2698 drivers/md/dm-raid.c rdev_for_each(rdev, &rs->md) { rs 2709 drivers/md/dm-raid.c static void __reorder_raid_disk_indexes(struct raid_set *rs) rs 2714 drivers/md/dm-raid.c rdev_for_each(rdev, &rs->md) { rs 2725 drivers/md/dm-raid.c static int rs_setup_takeover(struct raid_set *rs) rs 2727 drivers/md/dm-raid.c struct mddev *mddev = &rs->md; rs 2729 drivers/md/dm-raid.c unsigned int d = mddev->raid_disks = rs->raid_disks; rs 2730 drivers/md/dm-raid.c sector_t new_data_offset = rs->dev[0].rdev.data_offset ? 0 : rs->data_offset; rs 2732 drivers/md/dm-raid.c if (rt_is_raid10(rs->raid_type)) { rs 2733 drivers/md/dm-raid.c if (rs_is_raid0(rs)) { rs 2735 drivers/md/dm-raid.c __reorder_raid_disk_indexes(rs); rs 2738 drivers/md/dm-raid.c mddev->layout = raid10_format_to_md_layout(rs, ALGORITHM_RAID10_FAR, rs 2739 drivers/md/dm-raid.c rs->raid10_copies); rs 2740 drivers/md/dm-raid.c } else if (rs_is_raid1(rs)) rs 2742 drivers/md/dm-raid.c mddev->layout = raid10_format_to_md_layout(rs, ALGORITHM_RAID10_NEAR, rs 2743 drivers/md/dm-raid.c rs->raid_disks); rs 2753 drivers/md/dm-raid.c rdev = &rs->dev[d].rdev; rs 2755 drivers/md/dm-raid.c if (test_bit(d, (void *) rs->rebuild_disks)) { rs 2770 drivers/md/dm-raid.c static int rs_prepare_reshape(struct raid_set *rs) rs 2773 drivers/md/dm-raid.c struct mddev *mddev = &rs->md; rs 2775 drivers/md/dm-raid.c if (rs_is_raid10(rs)) { rs 2776 drivers/md/dm-raid.c if (rs->raid_disks != mddev->raid_disks && rs 2778 drivers/md/dm-raid.c rs->raid10_copies && rs 2779 drivers/md/dm-raid.c rs->raid10_copies != __raid10_near_copies(mddev->layout)) { rs 2786 drivers/md/dm-raid.c if (rs->raid_disks % rs->raid10_copies) { rs 2787 drivers/md/dm-raid.c rs->ti->error = "Can't reshape raid10 mirror groups"; rs 2792 drivers/md/dm-raid.c __reorder_raid_disk_indexes(rs); rs 2793 drivers/md/dm-raid.c mddev->layout = raid10_format_to_md_layout(rs, ALGORITHM_RAID10_NEAR, rs 2794 drivers/md/dm-raid.c rs->raid10_copies); rs 2800 drivers/md/dm-raid.c } else if (rs_is_raid456(rs)) rs 2803 drivers/md/dm-raid.c else if (rs_is_raid1(rs)) { rs 2804 drivers/md/dm-raid.c if (rs->delta_disks) { rs 2806 drivers/md/dm-raid.c mddev->degraded = rs->delta_disks < 0 ? -rs->delta_disks : rs->delta_disks; rs 2810 drivers/md/dm-raid.c mddev->raid_disks = rs->raid_disks; rs 2814 drivers/md/dm-raid.c rs->ti->error = "Called with bogus raid type"; rs 2819 drivers/md/dm-raid.c set_bit(RT_FLAG_RESHAPE_RS, &rs->runtime_flags); rs 2820 drivers/md/dm-raid.c set_bit(RT_FLAG_UPDATE_SBS, &rs->runtime_flags); rs 2821 drivers/md/dm-raid.c } else if (mddev->raid_disks < rs->raid_disks) rs 2823 drivers/md/dm-raid.c set_bit(RT_FLAG_UPDATE_SBS, &rs->runtime_flags); rs 2829 drivers/md/dm-raid.c static sector_t _get_reshape_sectors(struct raid_set *rs) rs 2834 drivers/md/dm-raid.c rdev_for_each(rdev, &rs->md) rs 2842 drivers/md/dm-raid.c return max(reshape_sectors, (sector_t) rs->data_offset); rs 2852 drivers/md/dm-raid.c static int rs_setup_reshape(struct raid_set *rs) rs 2856 drivers/md/dm-raid.c sector_t reshape_sectors = _get_reshape_sectors(rs); rs 2857 drivers/md/dm-raid.c struct mddev *mddev = &rs->md; rs 2860 drivers/md/dm-raid.c mddev->delta_disks = rs->delta_disks; rs 2866 drivers/md/dm-raid.c DMINFO("Ignoring invalid layout change with delta_disks=%d", rs->delta_disks); rs 2893 drivers/md/dm-raid.c if (rs->delta_disks > 0) { rs 2895 drivers/md/dm-raid.c for (d = cur_raid_devs; d < rs->raid_disks; d++) { rs 2896 drivers/md/dm-raid.c rdev = &rs->dev[d].rdev; rs 2907 drivers/md/dm-raid.c rdev->recovery_offset = rs_is_raid1(rs) ? 0 : MaxSector; rs 2913 drivers/md/dm-raid.c } else if (rs->delta_disks < 0) { rs 2914 drivers/md/dm-raid.c r = rs_set_dev_and_array_sectors(rs, true); rs 2940 drivers/md/dm-raid.c mddev->reshape_backwards = rs->dev[0].rdev.data_offset ? 0 : 1; rs 2948 drivers/md/dm-raid.c rdev_for_each(rdev, &rs->md) rs 2959 drivers/md/dm-raid.c static void configure_discard_support(struct raid_set *rs) rs 2963 drivers/md/dm-raid.c struct dm_target *ti = rs->ti; rs 2968 drivers/md/dm-raid.c raid456 = rs_is_raid456(rs); rs 2970 drivers/md/dm-raid.c for (i = 0; i < rs->raid_disks; i++) { rs 2973 drivers/md/dm-raid.c if (!rs->dev[i].rdev.bdev) rs 2976 drivers/md/dm-raid.c q = bdev_get_queue(rs->dev[i].rdev.bdev); rs 3012 drivers/md/dm-raid.c struct raid_set *rs = NULL; rs 3050 drivers/md/dm-raid.c rs = raid_set_alloc(ti, rt, num_raid_devs); rs 3051 drivers/md/dm-raid.c if (IS_ERR(rs)) rs 3052 drivers/md/dm-raid.c return PTR_ERR(rs); rs 3054 drivers/md/dm-raid.c r = parse_raid_params(rs, &as, num_raid_params); rs 3058 drivers/md/dm-raid.c r = parse_dev_params(rs, &as); rs 3062 drivers/md/dm-raid.c rs->md.sync_super = super_sync; rs 3070 drivers/md/dm-raid.c r = rs_set_dev_and_array_sectors(rs, false); rs 3074 drivers/md/dm-raid.c calculated_dev_sectors = rs->md.dev_sectors; rs 3081 drivers/md/dm-raid.c rs_config_backup(rs, &rs_layout); rs 3083 drivers/md/dm-raid.c r = analyse_superblocks(ti, rs); rs 3087 drivers/md/dm-raid.c rdev_sectors = __rdev_sectors(rs); rs 3095 drivers/md/dm-raid.c reshape_sectors = _get_reshape_sectors(rs); rs 3099 drivers/md/dm-raid.c INIT_WORK(&rs->md.event_work, do_table_event); rs 3100 drivers/md/dm-raid.c ti->private = rs; rs 3104 drivers/md/dm-raid.c rs_config_restore(rs, &rs_layout); rs 3112 drivers/md/dm-raid.c if (test_bit(MD_ARRAY_FIRST_USE, &rs->md.flags)) { rs 3114 drivers/md/dm-raid.c if (rs_is_raid6(rs) && rs 3115 drivers/md/dm-raid.c test_bit(__CTR_FLAG_NOSYNC, &rs->ctr_flags)) { rs 3120 drivers/md/dm-raid.c rs_setup_recovery(rs, 0); rs 3121 drivers/md/dm-raid.c set_bit(RT_FLAG_UPDATE_SBS, &rs->runtime_flags); rs 3122 drivers/md/dm-raid.c rs_set_new(rs); rs 3123 drivers/md/dm-raid.c } else if (rs_is_recovering(rs)) { rs 3125 drivers/md/dm-raid.c if (test_bit(__CTR_FLAG_REBUILD, &rs->ctr_flags)) { rs 3126 drivers/md/dm-raid.c set_bit(RT_FLAG_UPDATE_SBS, &rs->runtime_flags); rs 3127 drivers/md/dm-raid.c rs_setup_recovery(rs, MaxSector); rs 3131 drivers/md/dm-raid.c } else if (rs_is_reshaping(rs)) { rs 3139 drivers/md/dm-raid.c } else if (rs_takeover_requested(rs)) { rs 3140 drivers/md/dm-raid.c if (rs_is_reshaping(rs)) { rs 3147 drivers/md/dm-raid.c if (test_bit(__CTR_FLAG_JOURNAL_DEV, &rs->ctr_flags)) { rs 3161 drivers/md/dm-raid.c r = rs_check_takeover(rs); rs 3165 drivers/md/dm-raid.c r = rs_setup_takeover(rs); rs 3169 drivers/md/dm-raid.c set_bit(RT_FLAG_UPDATE_SBS, &rs->runtime_flags); rs 3171 drivers/md/dm-raid.c rs_setup_recovery(rs, MaxSector); rs 3172 drivers/md/dm-raid.c rs_set_new(rs); rs 3173 drivers/md/dm-raid.c } else if (rs_reshape_requested(rs)) { rs 3180 drivers/md/dm-raid.c if (test_bit(__CTR_FLAG_JOURNAL_DEV, &rs->ctr_flags)) { rs 3187 drivers/md/dm-raid.c if (reshape_sectors || rs_is_raid1(rs)) { rs 3195 drivers/md/dm-raid.c r = rs_prepare_reshape(rs); rs 3200 drivers/md/dm-raid.c rs_setup_recovery(rs, MaxSector); rs 3202 drivers/md/dm-raid.c rs_set_cur(rs); rs 3205 drivers/md/dm-raid.c if (test_bit(__CTR_FLAG_REBUILD, &rs->ctr_flags)) { rs 3206 drivers/md/dm-raid.c rs_setup_recovery(rs, MaxSector); rs 3207 drivers/md/dm-raid.c set_bit(RT_FLAG_UPDATE_SBS, &rs->runtime_flags); rs 3209 drivers/md/dm-raid.c rs_setup_recovery(rs, test_bit(__CTR_FLAG_SYNC, &rs->ctr_flags) ? rs 3211 drivers/md/dm-raid.c rs_set_cur(rs); rs 3215 drivers/md/dm-raid.c r = rs_adjust_data_offsets(rs); rs 3220 drivers/md/dm-raid.c rs->md.ro = 1; rs 3221 drivers/md/dm-raid.c rs->md.in_sync = 1; rs 3224 drivers/md/dm-raid.c set_bit(MD_RECOVERY_FROZEN, &rs->md.recovery); rs 3227 drivers/md/dm-raid.c mddev_lock_nointr(&rs->md); rs 3228 drivers/md/dm-raid.c r = md_run(&rs->md); rs 3229 drivers/md/dm-raid.c rs->md.in_sync = 0; /* Assume already marked dirty */ rs 3232 drivers/md/dm-raid.c mddev_unlock(&rs->md); rs 3236 drivers/md/dm-raid.c r = md_start(&rs->md); rs 3240 drivers/md/dm-raid.c mddev_unlock(&rs->md); rs 3244 drivers/md/dm-raid.c rs->callbacks.congested_fn = raid_is_congested; rs 3245 drivers/md/dm-raid.c dm_table_add_target_callbacks(ti->table, &rs->callbacks); rs 3248 drivers/md/dm-raid.c if (test_bit(__CTR_FLAG_JOURNAL_MODE, &rs->ctr_flags)) { rs 3249 drivers/md/dm-raid.c r = r5c_journal_mode_set(&rs->md, rs->journal_dev.mode); rs 3252 drivers/md/dm-raid.c mddev_unlock(&rs->md); rs 3257 drivers/md/dm-raid.c mddev_suspend(&rs->md); rs 3258 drivers/md/dm-raid.c set_bit(RT_FLAG_RS_SUSPENDED, &rs->runtime_flags); rs 3261 drivers/md/dm-raid.c if (rs_is_raid456(rs)) { rs 3262 drivers/md/dm-raid.c r = rs_set_raid456_stripe_cache(rs); rs 3268 drivers/md/dm-raid.c if (test_bit(RT_FLAG_RESHAPE_RS, &rs->runtime_flags)) { rs 3269 drivers/md/dm-raid.c r = rs_check_reshape(rs); rs 3274 drivers/md/dm-raid.c rs_config_restore(rs, &rs_layout); rs 3276 drivers/md/dm-raid.c if (rs->md.pers->start_reshape) { rs 3277 drivers/md/dm-raid.c r = rs->md.pers->check_reshape(&rs->md); rs 3286 drivers/md/dm-raid.c configure_discard_support(rs); rs 3288 drivers/md/dm-raid.c mddev_unlock(&rs->md); rs 3295 drivers/md/dm-raid.c md_stop(&rs->md); rs 3297 drivers/md/dm-raid.c raid_set_free(rs); rs 3304 drivers/md/dm-raid.c struct raid_set *rs = ti->private; rs 3306 drivers/md/dm-raid.c list_del_init(&rs->callbacks.list); rs 3307 drivers/md/dm-raid.c md_stop(&rs->md); rs 3308 drivers/md/dm-raid.c raid_set_free(rs); rs 3313 drivers/md/dm-raid.c struct raid_set *rs = ti->private; rs 3314 drivers/md/dm-raid.c struct mddev *mddev = &rs->md; rs 3391 drivers/md/dm-raid.c static const char *__raid_dev_status(struct raid_set *rs, struct md_rdev *rdev) rs 3398 drivers/md/dm-raid.c return (rs->journal_dev.mode == R5C_JOURNAL_MODE_WRITE_THROUGH) ? "A" : "a"; rs 3399 drivers/md/dm-raid.c else if (test_bit(RT_FLAG_RS_RESYNCING, &rs->runtime_flags) || rs 3400 drivers/md/dm-raid.c (!test_bit(RT_FLAG_RS_IN_SYNC, &rs->runtime_flags) && rs 3408 drivers/md/dm-raid.c static sector_t rs_get_progress(struct raid_set *rs, unsigned long recovery, rs 3413 drivers/md/dm-raid.c struct mddev *mddev = &rs->md; rs 3415 drivers/md/dm-raid.c clear_bit(RT_FLAG_RS_IN_SYNC, &rs->runtime_flags); rs 3416 drivers/md/dm-raid.c clear_bit(RT_FLAG_RS_RESYNCING, &rs->runtime_flags); rs 3418 drivers/md/dm-raid.c if (rs_is_raid0(rs)) { rs 3420 drivers/md/dm-raid.c set_bit(RT_FLAG_RS_IN_SYNC, &rs->runtime_flags); rs 3436 drivers/md/dm-raid.c set_bit(RT_FLAG_RS_IN_SYNC, &rs->runtime_flags); rs 3450 drivers/md/dm-raid.c set_bit(RT_FLAG_RS_RESYNCING, &rs->runtime_flags); rs 3457 drivers/md/dm-raid.c set_bit(RT_FLAG_RS_RESYNCING, &rs->runtime_flags); rs 3465 drivers/md/dm-raid.c set_bit(RT_FLAG_RS_IN_SYNC, &rs->runtime_flags); rs 3475 drivers/md/dm-raid.c set_bit(RT_FLAG_RS_RESYNCING, &rs->runtime_flags); rs 3483 drivers/md/dm-raid.c set_bit(RT_FLAG_RS_IN_SYNC, &rs->runtime_flags); rs 3487 drivers/md/dm-raid.c clear_bit(RT_FLAG_RS_IN_SYNC, &rs->runtime_flags); rs 3505 drivers/md/dm-raid.c struct raid_set *rs = ti->private; rs 3506 drivers/md/dm-raid.c struct mddev *mddev = &rs->md; rs 3529 drivers/md/dm-raid.c recovery = rs->md.recovery; rs 3531 drivers/md/dm-raid.c resync_max_sectors = test_bit(RT_FLAG_RS_PRERESUMED, &rs->runtime_flags) ? rs 3533 drivers/md/dm-raid.c progress = rs_get_progress(rs, recovery, resync_max_sectors); rs 3536 drivers/md/dm-raid.c sync_action = sync_str(decipher_sync_action(&rs->md, recovery)); rs 3539 drivers/md/dm-raid.c for (i = 0; i < rs->raid_disks; i++) rs 3540 drivers/md/dm-raid.c DMEMIT(__raid_dev_status(rs, &rs->dev[i].rdev)); rs 3585 drivers/md/dm-raid.c DMEMIT(" %llu", (unsigned long long) rs->dev[0].rdev.data_offset); rs 3590 drivers/md/dm-raid.c DMEMIT(" %s", test_bit(__CTR_FLAG_JOURNAL_DEV, &rs->ctr_flags) ? rs 3591 drivers/md/dm-raid.c __raid_dev_status(rs, &rs->journal_dev.rdev) : "-"); rs 3598 drivers/md/dm-raid.c for (i = 0; i < rs->raid_disks; i++) rs 3599 drivers/md/dm-raid.c if (test_bit(WriteMostly, &rs->dev[i].rdev.flags)) rs 3601 drivers/md/dm-raid.c rebuild_disks = memweight(rs->rebuild_disks, DISKS_ARRAY_ELEMS * sizeof(*rs->rebuild_disks)); rs 3604 drivers/md/dm-raid.c hweight32(rs->ctr_flags & CTR_FLAG_OPTIONS_NO_ARGS) + rs 3605 drivers/md/dm-raid.c hweight32(rs->ctr_flags & CTR_FLAG_OPTIONS_ONE_ARG) * 2 + rs 3606 drivers/md/dm-raid.c (test_bit(__CTR_FLAG_JOURNAL_DEV, &rs->ctr_flags) ? 2 : 0) + rs 3607 drivers/md/dm-raid.c (test_bit(__CTR_FLAG_JOURNAL_MODE, &rs->ctr_flags) ? 2 : 0); rs 3611 drivers/md/dm-raid.c DMEMIT("%s %u %u", rs->raid_type->name, raid_param_cnt, mddev->new_chunk_sectors); rs 3612 drivers/md/dm-raid.c if (test_bit(__CTR_FLAG_SYNC, &rs->ctr_flags)) rs 3614 drivers/md/dm-raid.c if (test_bit(__CTR_FLAG_NOSYNC, &rs->ctr_flags)) rs 3617 drivers/md/dm-raid.c for (i = 0; i < rs->raid_disks; i++) rs 3618 drivers/md/dm-raid.c if (test_bit(rs->dev[i].rdev.raid_disk, (void *) rs->rebuild_disks)) rs 3620 drivers/md/dm-raid.c rs->dev[i].rdev.raid_disk); rs 3621 drivers/md/dm-raid.c if (test_bit(__CTR_FLAG_DAEMON_SLEEP, &rs->ctr_flags)) rs 3624 drivers/md/dm-raid.c if (test_bit(__CTR_FLAG_MIN_RECOVERY_RATE, &rs->ctr_flags)) rs 3627 drivers/md/dm-raid.c if (test_bit(__CTR_FLAG_MAX_RECOVERY_RATE, &rs->ctr_flags)) rs 3631 drivers/md/dm-raid.c for (i = 0; i < rs->raid_disks; i++) rs 3632 drivers/md/dm-raid.c if (test_bit(WriteMostly, &rs->dev[i].rdev.flags)) rs 3634 drivers/md/dm-raid.c rs->dev[i].rdev.raid_disk); rs 3635 drivers/md/dm-raid.c if (test_bit(__CTR_FLAG_MAX_WRITE_BEHIND, &rs->ctr_flags)) rs 3638 drivers/md/dm-raid.c if (test_bit(__CTR_FLAG_STRIPE_CACHE, &rs->ctr_flags)) rs 3641 drivers/md/dm-raid.c if (test_bit(__CTR_FLAG_REGION_SIZE, &rs->ctr_flags)) rs 3644 drivers/md/dm-raid.c if (test_bit(__CTR_FLAG_RAID10_COPIES, &rs->ctr_flags)) rs 3647 drivers/md/dm-raid.c if (test_bit(__CTR_FLAG_RAID10_FORMAT, &rs->ctr_flags)) rs 3650 drivers/md/dm-raid.c if (test_bit(__CTR_FLAG_DELTA_DISKS, &rs->ctr_flags)) rs 3652 drivers/md/dm-raid.c max(rs->delta_disks, mddev->delta_disks)); rs 3653 drivers/md/dm-raid.c if (test_bit(__CTR_FLAG_DATA_OFFSET, &rs->ctr_flags)) rs 3655 drivers/md/dm-raid.c (unsigned long long) rs->data_offset); rs 3656 drivers/md/dm-raid.c if (test_bit(__CTR_FLAG_JOURNAL_DEV, &rs->ctr_flags)) rs 3658 drivers/md/dm-raid.c __get_dev_name(rs->journal_dev.dev)); rs 3659 drivers/md/dm-raid.c if (test_bit(__CTR_FLAG_JOURNAL_MODE, &rs->ctr_flags)) rs 3661 drivers/md/dm-raid.c md_journal_mode_to_dm_raid(rs->journal_dev.mode)); rs 3662 drivers/md/dm-raid.c DMEMIT(" %d", rs->raid_disks); rs 3663 drivers/md/dm-raid.c for (i = 0; i < rs->raid_disks; i++) rs 3664 drivers/md/dm-raid.c DMEMIT(" %s %s", __get_dev_name(rs->dev[i].meta_dev), rs 3665 drivers/md/dm-raid.c __get_dev_name(rs->dev[i].data_dev)); rs 3672 drivers/md/dm-raid.c struct raid_set *rs = ti->private; rs 3673 drivers/md/dm-raid.c struct mddev *mddev = &rs->md; rs 3723 drivers/md/dm-raid.c struct raid_set *rs = ti->private; rs 3727 drivers/md/dm-raid.c for (i = 0; !r && i < rs->md.raid_disks; i++) rs 3728 drivers/md/dm-raid.c if (rs->dev[i].data_dev) rs 3730 drivers/md/dm-raid.c rs->dev[i].data_dev, rs 3732 drivers/md/dm-raid.c rs->md.dev_sectors, rs 3740 drivers/md/dm-raid.c struct raid_set *rs = ti->private; rs 3741 drivers/md/dm-raid.c unsigned int chunk_size_bytes = to_bytes(rs->md.chunk_sectors); rs 3744 drivers/md/dm-raid.c blk_limits_io_opt(limits, chunk_size_bytes * mddev_data_stripes(rs)); rs 3750 drivers/md/dm-raid.c if (rs_is_raid1(rs) || rs_is_raid10(rs)) { rs 3752 drivers/md/dm-raid.c limits->max_discard_sectors = rs->md.chunk_sectors; rs 3758 drivers/md/dm-raid.c struct raid_set *rs = ti->private; rs 3760 drivers/md/dm-raid.c if (!test_and_set_bit(RT_FLAG_RS_SUSPENDED, &rs->runtime_flags)) { rs 3762 drivers/md/dm-raid.c if (!test_bit(MD_RECOVERY_FROZEN, &rs->md.recovery)) rs 3763 drivers/md/dm-raid.c md_stop_writes(&rs->md); rs 3765 drivers/md/dm-raid.c mddev_lock_nointr(&rs->md); rs 3766 drivers/md/dm-raid.c mddev_suspend(&rs->md); rs 3767 drivers/md/dm-raid.c mddev_unlock(&rs->md); rs 3771 drivers/md/dm-raid.c static void attempt_restore_of_faulty_devices(struct raid_set *rs) rs 3778 drivers/md/dm-raid.c struct mddev *mddev = &rs->md; rs 3788 drivers/md/dm-raid.c r = &rs->dev[i].rdev; rs 3797 drivers/md/dm-raid.c rs->raid_type->name, i); rs 3839 drivers/md/dm-raid.c rdev_for_each(r, &rs->md) { rs 3854 drivers/md/dm-raid.c static int __load_dirty_region_bitmap(struct raid_set *rs) rs 3859 drivers/md/dm-raid.c if (!rs_is_raid0(rs) && rs 3860 drivers/md/dm-raid.c !test_and_set_bit(RT_FLAG_RS_BITMAP_LOADED, &rs->runtime_flags)) { rs 3861 drivers/md/dm-raid.c r = md_bitmap_load(&rs->md); rs 3870 drivers/md/dm-raid.c static void rs_update_sbs(struct raid_set *rs) rs 3872 drivers/md/dm-raid.c struct mddev *mddev = &rs->md; rs 3888 drivers/md/dm-raid.c static int rs_start_reshape(struct raid_set *rs) rs 3891 drivers/md/dm-raid.c struct mddev *mddev = &rs->md; rs 3897 drivers/md/dm-raid.c r = rs_setup_reshape(rs); rs 3908 drivers/md/dm-raid.c rs->ti->error = "pers->check_reshape() failed"; rs 3919 drivers/md/dm-raid.c rs->ti->error = "pers->start_reshape() failed"; rs 3929 drivers/md/dm-raid.c rs_update_sbs(rs); rs 3937 drivers/md/dm-raid.c struct raid_set *rs = ti->private; rs 3938 drivers/md/dm-raid.c struct mddev *mddev = &rs->md; rs 3941 drivers/md/dm-raid.c if (test_and_set_bit(RT_FLAG_RS_PRERESUMED, &rs->runtime_flags)) rs 3950 drivers/md/dm-raid.c if (test_bit(RT_FLAG_UPDATE_SBS, &rs->runtime_flags)) rs 3951 drivers/md/dm-raid.c rs_update_sbs(rs); rs 3954 drivers/md/dm-raid.c r = __load_dirty_region_bitmap(rs); rs 3959 drivers/md/dm-raid.c if (test_bit(RT_FLAG_RS_BITMAP_LOADED, &rs->runtime_flags) && mddev->bitmap && rs 3960 drivers/md/dm-raid.c mddev->bitmap_info.chunksize != to_bytes(rs->requested_bitmap_chunk_sectors)) { rs 3962 drivers/md/dm-raid.c to_bytes(rs->requested_bitmap_chunk_sectors), 0); rs 3976 drivers/md/dm-raid.c if (test_bit(RT_FLAG_RESHAPE_RS, &rs->runtime_flags)) { rs 3978 drivers/md/dm-raid.c rs_set_rdev_sectors(rs); rs 3980 drivers/md/dm-raid.c r = rs_start_reshape(rs); rs 3992 drivers/md/dm-raid.c struct raid_set *rs = ti->private; rs 3993 drivers/md/dm-raid.c struct mddev *mddev = &rs->md; rs 3995 drivers/md/dm-raid.c if (test_and_set_bit(RT_FLAG_RS_RESUMED, &rs->runtime_flags)) { rs 4001 drivers/md/dm-raid.c attempt_restore_of_faulty_devices(rs); rs 4004 drivers/md/dm-raid.c if (test_and_clear_bit(RT_FLAG_RS_SUSPENDED, &rs->runtime_flags)) { rs 4007 drivers/md/dm-raid.c rs_set_capacity(rs); rs 53 drivers/md/dm-verity-fec.c return decode_rs8(fio->rs, data, par, v->fec->rsn, NULL, neras, rs 310 drivers/md/dm-verity-fec.c if (!fio->rs) rs 311 drivers/md/dm-verity-fec.c fio->rs = mempool_alloc(&v->fec->rs_pool, GFP_NOIO); rs 495 drivers/md/dm-verity-fec.c mempool_free(fio->rs, &f->rs_pool); rs 516 drivers/md/dm-verity-fec.c fio->rs = NULL; rs 578 drivers/md/dm-verity-fec.c struct rs_control *rs = (struct rs_control *)element; rs 580 drivers/md/dm-verity-fec.c if (rs) rs 581 drivers/md/dm-verity-fec.c free_rs(rs); rs 54 drivers/md/dm-verity-fec.h struct rs_control *rs; /* Reed-Solomon state */ rs 74 drivers/media/usb/gspca/stv06xx/stv06xx_hdcs.c u16 rs; rs 185 drivers/media/usb/gspca/stv06xx/stv06xx_hdcs.c rp = hdcs->exp.rs + cp; rs 383 drivers/media/usb/gspca/stv06xx/stv06xx_hdcs.c hdcs->exp.rs = 186; rs 447 drivers/media/usb/gspca/stv06xx/stv06xx_hdcs.c hdcs->exp.rs = 155; rs 497 drivers/misc/mei/hbm.c struct hbm_notification_response *rs = rs 500 drivers/misc/mei/hbm.c return mei_cl_notify_req2fop(rs->start); rs 515 drivers/misc/mei/hbm.c struct hbm_notification_response *rs = rs 518 drivers/misc/mei/hbm.c cl_dbg(dev, cl, "hbm: notify start response status=%d\n", rs->status); rs 520 drivers/misc/mei/hbm.c if (rs->status == MEI_HBMS_SUCCESS || rs 521 drivers/misc/mei/hbm.c rs->status == MEI_HBMS_ALREADY_STARTED) { rs 541 drivers/misc/mei/hbm.c struct hbm_notification_response *rs = rs 544 drivers/misc/mei/hbm.c cl_dbg(dev, cl, "hbm: notify stop response status=%d\n", rs->status); rs 546 drivers/misc/mei/hbm.c if (rs->status == MEI_HBMS_SUCCESS || rs 547 drivers/misc/mei/hbm.c rs->status == MEI_HBMS_NOT_STARTED) { rs 792 drivers/misc/mei/hbm.c struct hbm_client_connect_response *rs = rs 795 drivers/misc/mei/hbm.c cl_dbg(dev, cl, "hbm: disconnect response status=%d\n", rs->status); rs 797 drivers/misc/mei/hbm.c if (rs->status == MEI_CL_DISCONN_SUCCESS) rs 829 drivers/misc/mei/hbm.c struct hbm_client_connect_response *rs = rs 833 drivers/misc/mei/hbm.c mei_cl_conn_status_str(rs->status)); rs 835 drivers/misc/mei/hbm.c if (rs->status == MEI_CL_CONN_SUCCESS) rs 839 drivers/misc/mei/hbm.c if (rs->status == MEI_CL_CONN_NOT_FOUND) { rs 845 drivers/misc/mei/hbm.c cl->status = mei_cl_conn_status_to_errno(rs->status); rs 857 drivers/misc/mei/hbm.c struct mei_hbm_cl_cmd *rs, rs 871 drivers/misc/mei/hbm.c if (mei_hbm_cl_addr_equal(cl, rs)) { rs 882 drivers/misc/mei/hbm.c mei_hbm_cl_connect_res(dev, cl, rs); rs 885 drivers/misc/mei/hbm.c mei_hbm_cl_disconnect_res(dev, cl, rs); rs 888 drivers/misc/mei/hbm.c mei_hbm_cl_notify_start_res(dev, cl, rs); rs 891 drivers/misc/mei/hbm.c mei_hbm_cl_notify_stop_res(dev, cl, rs); rs 89 drivers/misc/phantom.c struct phm_regs rs; rs 129 drivers/misc/phantom.c if (copy_from_user(&rs, argp, sizeof(rs))) rs 132 drivers/misc/phantom.c pr_debug("phantom: SRS %u regs %x\n", rs.count, rs.mask); rs 135 drivers/misc/phantom.c memcpy(&dev->oregs, &rs, sizeof(rs)); rs 137 drivers/misc/phantom.c u32 m = min(rs.count, 8U); rs 139 drivers/misc/phantom.c if (rs.mask & BIT(i)) rs 140 drivers/misc/phantom.c iowrite32(rs.values[i], dev->oaddr + i); rs 162 drivers/misc/phantom.c if (copy_from_user(&rs, argp, sizeof(rs))) rs 165 drivers/misc/phantom.c m = min(rs.count, 8U); rs 167 drivers/misc/phantom.c pr_debug("phantom: GRS %u regs %x\n", rs.count, rs.mask); rs 170 drivers/misc/phantom.c if (rs.mask & BIT(i)) rs 171 drivers/misc/phantom.c rs.values[i] = ioread32(dev->iaddr + i); rs 175 drivers/misc/phantom.c if (copy_to_user(argp, &rs, sizeof(rs))) rs 64 drivers/mtd/nand/raw/cafe_nand.c struct rs_control *rs; rs 393 drivers/mtd/nand/raw/cafe_nand.c syn[i] = cafe->rs->codec->index_of[tmp & 0xfff]; rs 394 drivers/mtd/nand/raw/cafe_nand.c syn[i+1] = cafe->rs->codec->index_of[(tmp >> 16) & 0xfff]; rs 397 drivers/mtd/nand/raw/cafe_nand.c n = decode_rs16(cafe->rs, NULL, NULL, 1367, syn, 0, pos, 0, rs 701 drivers/mtd/nand/raw/cafe_nand.c cafe->rs = init_rs_non_canonical(12, &cafe_mul, 0, 1, 8); rs 702 drivers/mtd/nand/raw/cafe_nand.c if (!cafe->rs) { rs 822 drivers/mtd/nand/raw/cafe_nand.c free_rs(cafe->rs); rs 137 drivers/mtd/nand/raw/diskonchip.c static int doc_ecc_decode(struct rs_control *rs, uint8_t *data, uint8_t *ecc) rs 142 drivers/mtd/nand/raw/diskonchip.c struct rs_codec *cd = rs->codec; rs 174 drivers/mtd/nand/raw/diskonchip.c nerr = decode_rs16(rs, NULL, NULL, 1019, syn, 0, errpos, 0, errval); rs 509 drivers/net/ethernet/cavium/liquidio/liquidio_common.h u64 rs:1; rs 524 drivers/net/ethernet/cavium/liquidio/liquidio_common.h u64 rs:1; rs 662 drivers/net/ethernet/cavium/liquidio/request_manager.c ih2->rs = 1; rs 4345 drivers/net/ethernet/intel/ixgbe/ixgbe_common.c u8 rs, eeprom_data; rs 4350 drivers/net/ethernet/intel/ixgbe/ixgbe_common.c rs = IXGBE_SFF_SOFT_RS_SELECT_10G; rs 4353 drivers/net/ethernet/intel/ixgbe/ixgbe_common.c rs = IXGBE_SFF_SOFT_RS_SELECT_1G; rs 4369 drivers/net/ethernet/intel/ixgbe/ixgbe_common.c eeprom_data = (eeprom_data & ~IXGBE_SFF_SOFT_RS_SELECT_MASK) | rs; rs 4388 drivers/net/ethernet/intel/ixgbe/ixgbe_common.c eeprom_data = (eeprom_data & ~IXGBE_SFF_SOFT_RS_SELECT_MASK) | rs; rs 397 drivers/net/ethernet/sfc/mcdi_port.c bool rs = caps & (1 << MC_CMD_PHY_CAP_RS_FEC_LBN), rs 404 drivers/net/ethernet/sfc/mcdi_port.c if (!baser && !rs) rs 408 drivers/net/ethernet/sfc/mcdi_port.c (baser == baser_req && rs == rs_req ? 0 : ETHTOOL_FEC_AUTO); rs 280 drivers/net/ethernet/stmicro/stmmac/dwmac-sti.c const char *rs; rs 284 drivers/net/ethernet/stmicro/stmmac/dwmac-sti.c err = of_property_read_string(np, "st,tx-retime-src", &rs); rs 288 drivers/net/ethernet/stmicro/stmmac/dwmac-sti.c if (!strcasecmp(rs, "clk_125")) rs 290 drivers/net/ethernet/stmicro/stmmac/dwmac-sti.c else if (!strcasecmp(rs, "txclk")) rs 1183 drivers/net/wireless/ath/ar5523/ar5523.c struct ar5523_cmd_rateset *rs, rs 1207 drivers/net/wireless/ath/ar5523/ar5523.c rs->set[i] = band->bitrates[bit].hw_value; rs 1209 drivers/net/wireless/ath/ar5523/ar5523.c rs->set[i] |= 0x80; rs 1216 drivers/net/wireless/ath/ar5523/ar5523.c rs->length = i; rs 3958 drivers/net/wireless/ath/ath10k/wmi.c struct radar_detector_specs rs; rs 4026 drivers/net/wireless/ath/ath10k/wmi.c if (!ar->dfs_detector->add_pulse(ar->dfs_detector, &pe, &rs)) { rs 4046 drivers/net/wireless/ath/ath10k/wmi.c radar_info->pri_min = rs.pri_min; rs 4047 drivers/net/wireless/ath/ath10k/wmi.c radar_info->pri_max = rs.pri_max; rs 4048 drivers/net/wireless/ath/ath10k/wmi.c radar_info->width_min = rs.width_min; rs 4049 drivers/net/wireless/ath/ath10k/wmi.c radar_info->width_max = rs.width_max; rs 1212 drivers/net/wireless/ath/ath5k/base.c struct ath5k_rx_status *rs) rs 1218 drivers/net/wireless/ath/ath5k/base.c if (!(rs->rs_status & AR5K_RXERR_DECRYPT) && rs 1219 drivers/net/wireless/ath/ath5k/base.c rs->rs_keyix != AR5K_RXKEYIX_INVALID) rs 1227 drivers/net/wireless/ath/ath5k/base.c !(rs->rs_status & AR5K_RXERR_DECRYPT) && rs 1372 drivers/net/wireless/ath/ath5k/base.c struct ath5k_rx_status *rs) rs 1382 drivers/net/wireless/ath/ath5k/base.c if (unlikely(rs->rs_status & AR5K_RXERR_MIC)) rs 1384 drivers/net/wireless/ath/ath5k/base.c if (unlikely(rs->rs_status & AR5K_RXERR_CRC)) rs 1397 drivers/net/wireless/ath/ath5k/base.c rxs->mactime = ath5k_extend_tsf(ah, rs->rs_tstamp); rs 1403 drivers/net/wireless/ath/ath5k/base.c rxs->signal = ah->ah_noise_floor + rs->rs_rssi; rs 1405 drivers/net/wireless/ath/ath5k/base.c rxs->antenna = rs->rs_antenna; rs 1407 drivers/net/wireless/ath/ath5k/base.c if (rs->rs_antenna > 0 && rs->rs_antenna < 5) rs 1408 drivers/net/wireless/ath/ath5k/base.c ah->stats.antenna_rx[rs->rs_antenna]++; rs 1412 drivers/net/wireless/ath/ath5k/base.c rxs->rate_idx = ath5k_hw_to_driver_rix(ah, rs->rs_rate); rs 1413 drivers/net/wireless/ath/ath5k/base.c rxs->flag |= ath5k_rx_decrypted(ah, skb, rs); rs 1425 drivers/net/wireless/ath/ath5k/base.c if (rs->rs_rate == rs 1432 drivers/net/wireless/ath/ath5k/base.c ewma_beacon_rssi_add(&ah->ah_beacon_rssi_avg, rs->rs_rssi); rs 1448 drivers/net/wireless/ath/ath5k/base.c ath5k_receive_frame_ok(struct ath5k_hw *ah, struct ath5k_rx_status *rs) rs 1451 drivers/net/wireless/ath/ath5k/base.c ah->stats.rx_bytes_count += rs->rs_datalen; rs 1453 drivers/net/wireless/ath/ath5k/base.c if (unlikely(rs->rs_status)) { rs 1456 drivers/net/wireless/ath/ath5k/base.c if (rs->rs_status & AR5K_RXERR_CRC) rs 1458 drivers/net/wireless/ath/ath5k/base.c if (rs->rs_status & AR5K_RXERR_FIFO) rs 1460 drivers/net/wireless/ath/ath5k/base.c if (rs->rs_status & AR5K_RXERR_PHY) { rs 1462 drivers/net/wireless/ath/ath5k/base.c if (rs->rs_phyerr > 0 && rs->rs_phyerr < 32) rs 1463 drivers/net/wireless/ath/ath5k/base.c ah->stats.rxerr_phy_code[rs->rs_phyerr]++; rs 1471 drivers/net/wireless/ath/ath5k/base.c if (rs->rs_phyerr == AR5K_RX_PHY_ERROR_OFDM_RESTART || rs 1472 drivers/net/wireless/ath/ath5k/base.c rs->rs_phyerr == AR5K_RX_PHY_ERROR_CCK_RESTART) { rs 1473 drivers/net/wireless/ath/ath5k/base.c rs->rs_status |= AR5K_RXERR_CRC; rs 1474 drivers/net/wireless/ath/ath5k/base.c rs->rs_status &= ~AR5K_RXERR_PHY; rs 1479 drivers/net/wireless/ath/ath5k/base.c if (rs->rs_status & AR5K_RXERR_DECRYPT) { rs 1491 drivers/net/wireless/ath/ath5k/base.c if (rs->rs_keyix == AR5K_RXKEYIX_INVALID && rs 1492 drivers/net/wireless/ath/ath5k/base.c !(rs->rs_status & AR5K_RXERR_CRC)) rs 1495 drivers/net/wireless/ath/ath5k/base.c if (rs->rs_status & AR5K_RXERR_MIC) { rs 1508 drivers/net/wireless/ath/ath5k/base.c if (rs->rs_status & ~filters) rs 1512 drivers/net/wireless/ath/ath5k/base.c if (unlikely(rs->rs_more)) { rs 1541 drivers/net/wireless/ath/ath5k/base.c struct ath5k_rx_status rs = {}; rs 1565 drivers/net/wireless/ath/ath5k/base.c ret = ah->ah_proc_rx_desc(ah, ds, &rs); rs 1574 drivers/net/wireless/ath/ath5k/base.c if (ath5k_receive_frame_ok(ah, &rs)) { rs 1588 drivers/net/wireless/ath/ath5k/base.c skb_put(skb, rs.rs_datalen); rs 1590 drivers/net/wireless/ath/ath5k/base.c ath5k_receive_frame(ah, skb, &rs); rs 1068 drivers/net/wireless/ath/ath5k/debug.c struct ath5k_rx_status *rs) rs 1078 drivers/net/wireless/ath/ath5k/debug.c !done ? ' ' : (rs->rs_status == 0) ? '*' : '!'); rs 1086 drivers/net/wireless/ath/ath5k/debug.c struct ath5k_rx_status rs = {}; rs 1098 drivers/net/wireless/ath/ath5k/debug.c status = ah->ah_proc_rx_desc(ah, ds, &rs); rs 1100 drivers/net/wireless/ath/ath5k/debug.c ath5k_debug_printrxbuf(bf, status == 0, &rs); rs 603 drivers/net/wireless/ath/ath5k/desc.c struct ath5k_rx_status *rs) rs 614 drivers/net/wireless/ath/ath5k/desc.c memset(rs, 0, sizeof(struct ath5k_rx_status)); rs 619 drivers/net/wireless/ath/ath5k/desc.c rs->rs_datalen = rx_status->rx_status_0 & rs 621 drivers/net/wireless/ath/ath5k/desc.c rs->rs_rssi = AR5K_REG_MS(rx_status->rx_status_0, rs 623 drivers/net/wireless/ath/ath5k/desc.c rs->rs_rate = AR5K_REG_MS(rx_status->rx_status_0, rs 625 drivers/net/wireless/ath/ath5k/desc.c rs->rs_more = !!(rx_status->rx_status_0 & rs 632 drivers/net/wireless/ath/ath5k/desc.c rs->rs_tstamp = AR5K_REG_MS(rx_status->rx_status_1, rs 636 drivers/net/wireless/ath/ath5k/desc.c rs->rs_antenna = AR5K_REG_MS(rx_status->rx_status_0, rs 639 drivers/net/wireless/ath/ath5k/desc.c rs->rs_antenna = (rx_status->rx_status_0 & rs 647 drivers/net/wireless/ath/ath5k/desc.c rs->rs_keyix = AR5K_REG_MS(rx_status->rx_status_1, rs 650 drivers/net/wireless/ath/ath5k/desc.c rs->rs_keyix = AR5K_RXKEYIX_INVALID; rs 659 drivers/net/wireless/ath/ath5k/desc.c rs->rs_status |= AR5K_RXERR_CRC; rs 665 drivers/net/wireless/ath/ath5k/desc.c rs->rs_status |= AR5K_RXERR_FIFO; rs 669 drivers/net/wireless/ath/ath5k/desc.c rs->rs_status |= AR5K_RXERR_PHY; rs 670 drivers/net/wireless/ath/ath5k/desc.c rs->rs_phyerr = AR5K_REG_MS(rx_status->rx_status_1, rs 676 drivers/net/wireless/ath/ath5k/desc.c rs->rs_status |= AR5K_RXERR_DECRYPT; rs 697 drivers/net/wireless/ath/ath5k/desc.c struct ath5k_rx_status *rs) rs 709 drivers/net/wireless/ath/ath5k/desc.c memset(rs, 0, sizeof(struct ath5k_rx_status)); rs 715 drivers/net/wireless/ath/ath5k/desc.c rs->rs_datalen = rxstat0 & AR5K_5212_RX_DESC_STATUS0_DATA_LEN; rs 716 drivers/net/wireless/ath/ath5k/desc.c rs->rs_rssi = AR5K_REG_MS(rxstat0, rs 718 drivers/net/wireless/ath/ath5k/desc.c rs->rs_rate = AR5K_REG_MS(rxstat0, rs 720 drivers/net/wireless/ath/ath5k/desc.c rs->rs_antenna = AR5K_REG_MS(rxstat0, rs 722 drivers/net/wireless/ath/ath5k/desc.c rs->rs_more = !!(rxstat0 & AR5K_5212_RX_DESC_STATUS0_MORE); rs 723 drivers/net/wireless/ath/ath5k/desc.c rs->rs_tstamp = AR5K_REG_MS(rxstat1, rs 730 drivers/net/wireless/ath/ath5k/desc.c rs->rs_keyix = AR5K_REG_MS(rxstat1, rs 733 drivers/net/wireless/ath/ath5k/desc.c rs->rs_keyix = AR5K_RXKEYIX_INVALID; rs 740 drivers/net/wireless/ath/ath5k/desc.c rs->rs_status |= AR5K_RXERR_CRC; rs 743 drivers/net/wireless/ath/ath5k/desc.c rs->rs_status |= AR5K_RXERR_PHY; rs 744 drivers/net/wireless/ath/ath5k/desc.c rs->rs_phyerr = AR5K_REG_MS(rxstat1, rs 747 drivers/net/wireless/ath/ath5k/desc.c ath5k_ani_phy_error_report(ah, rs->rs_phyerr); rs 751 drivers/net/wireless/ath/ath5k/desc.c rs->rs_status |= AR5K_RXERR_DECRYPT; rs 754 drivers/net/wireless/ath/ath5k/desc.c rs->rs_status |= AR5K_RXERR_MIC; rs 721 drivers/net/wireless/ath/ath9k/antenna.c void ath_ant_comb_scan(struct ath_softc *sc, struct ath_rx_status *rs) rs 727 drivers/net/wireless/ath/ath9k/antenna.c int main_rssi = rs->rs_rssi_ctl[0]; rs 728 drivers/net/wireless/ath/ath9k/antenna.c int alt_rssi = rs->rs_rssi_ctl[1]; rs 732 drivers/net/wireless/ath/ath9k/antenna.c rx_ant_conf = (rs->rs_rssi_ctl[2] >> ATH_ANT_RX_CURRENT_SHIFT) & rs 734 drivers/net/wireless/ath/ath9k/antenna.c main_ant_conf = (rs->rs_rssi_ctl[2] >> ATH_ANT_RX_MAIN_SHIFT) & rs 769 drivers/net/wireless/ath/ath9k/antenna.c rs->rs_moreaggr) && !short_scan) rs 944 drivers/net/wireless/ath/ath9k/ath9k.h void ath_ant_comb_scan(struct ath_softc *sc, struct ath_rx_status *rs); rs 91 drivers/net/wireless/ath/ath9k/common-debug.c struct ath_rx_status *rs) rs 97 drivers/net/wireless/ath/ath9k/common-debug.c rxstats->rx_bytes_all += rs->rs_datalen; rs 99 drivers/net/wireless/ath/ath9k/common-debug.c if (rs->rs_status & ATH9K_RXERR_CRC) rs 101 drivers/net/wireless/ath/ath9k/common-debug.c if (rs->rs_status & ATH9K_RXERR_DECRYPT) rs 103 drivers/net/wireless/ath/ath9k/common-debug.c if (rs->rs_status & ATH9K_RXERR_MIC) rs 105 drivers/net/wireless/ath/ath9k/common-debug.c if (rs->rs_status & ATH9K_RX_DELIM_CRC_PRE) rs 107 drivers/net/wireless/ath/ath9k/common-debug.c if (rs->rs_status & ATH9K_RX_DELIM_CRC_POST) rs 109 drivers/net/wireless/ath/ath9k/common-debug.c if (rs->rs_status & ATH9K_RX_DECRYPT_BUSY) rs 112 drivers/net/wireless/ath/ath9k/common-debug.c if (rs->rs_status & ATH9K_RXERR_PHY) { rs 114 drivers/net/wireless/ath/ath9k/common-debug.c if (rs->rs_phyerr < ATH9K_PHYERR_MAX) rs 115 drivers/net/wireless/ath/ath9k/common-debug.c RX_PHY_ERR_INC(rs->rs_phyerr); rs 73 drivers/net/wireless/ath/ath9k/common-debug.h struct ath_rx_status *rs); rs 90 drivers/net/wireless/ath/ath9k/common-debug.h struct ath_rx_status *rs) rs 126 drivers/net/wireless/ath/ath9k/common-spectral.c typedef int (ath_cmn_fft_sample_handler) (struct ath_rx_status *rs, rs 131 drivers/net/wireless/ath/ath9k/common-spectral.c ath_cmn_process_ht20_fft(struct ath_rx_status *rs, rs 151 drivers/net/wireless/ath/ath9k/common-spectral.c fft_sample_20.rssi = fix_rssi_inv_only(rs->rs_rssi_ctl[0]); rs 235 drivers/net/wireless/ath/ath9k/common-spectral.c ath_cmn_process_ht20_40_fft(struct ath_rx_status *rs, rs 268 drivers/net/wireless/ath/ath9k/common-spectral.c lower_rssi = fix_rssi_inv_only(rs->rs_rssi_ctl[0]); rs 269 drivers/net/wireless/ath/ath9k/common-spectral.c upper_rssi = fix_rssi_inv_only(rs->rs_rssi_ext[0]); rs 274 drivers/net/wireless/ath/ath9k/common-spectral.c lower_rssi = fix_rssi_inv_only(rs->rs_rssi_ext[0]); rs 275 drivers/net/wireless/ath/ath9k/common-spectral.c upper_rssi = fix_rssi_inv_only(rs->rs_rssi_ctl[0]); rs 469 drivers/net/wireless/ath/ath9k/common-spectral.c struct ath_rx_status *rs, u64 tsf) rs 477 drivers/net/wireless/ath/ath9k/common-spectral.c int len = rs->rs_datalen; rs 492 drivers/net/wireless/ath/ath9k/common-spectral.c if (rs->rs_phyerr != ATH9K_PHYERR_RADAR && rs 493 drivers/net/wireless/ath/ath9k/common-spectral.c rs->rs_phyerr != ATH9K_PHYERR_FALSE_RADAR_EXT && rs 494 drivers/net/wireless/ath/ath9k/common-spectral.c rs->rs_phyerr != ATH9K_PHYERR_SPECTRAL) rs 623 drivers/net/wireless/ath/ath9k/common-spectral.c ret = fft_handler(rs, spec_priv, sample_buf, rs 641 drivers/net/wireless/ath/ath9k/common-spectral.c ret = fft_handler(rs, spec_priv, sample_start, rs 181 drivers/net/wireless/ath/ath9k/common-spectral.h struct ath_rx_status *rs, u64 tsf); rs 199 drivers/net/wireless/ath/ath9k/common-spectral.h struct ath_rx_status *rs, u64 tsf) rs 819 drivers/net/wireless/ath/ath9k/debug.c void ath_debug_stat_rx(struct ath_softc *sc, struct ath_rx_status *rs) rs 821 drivers/net/wireless/ath/ath9k/debug.c ath9k_cmn_debug_stat_rx(&sc->debug.stats.rxstats, rs); rs 261 drivers/net/wireless/ath/ath9k/debug.h void ath_debug_stat_rx(struct ath_softc *sc, struct ath_rx_status *rs); rs 301 drivers/net/wireless/ath/ath9k/debug.h struct ath_rx_status *rs) rs 320 drivers/net/wireless/ath/ath9k/debug.h struct ath_rx_status *rs, rs 324 drivers/net/wireless/ath/ath9k/debug.h struct ath_rx_status *rs, rs 92 drivers/net/wireless/ath/ath9k/debug_sta.c struct ath_rx_status *rs, rs 115 drivers/net/wireless/ath/ath9k/debug_sta.c if (IS_HT_RATE(rs->rs_rate)) { rs 132 drivers/net/wireless/ath/ath9k/debug_sta.c if (IS_CCK_RATE(rs->rs_rate)) { rs 141 drivers/net/wireless/ath/ath9k/debug_sta.c if (IS_OFDM_RATE(rs->rs_rate)) { rs 290 drivers/net/wireless/ath/ath9k/dfs.c struct ath_rx_status *rs, u64 mactime) rs 300 drivers/net/wireless/ath/ath9k/dfs.c if ((rs->rs_phyerr != ATH9K_PHYERR_RADAR) && rs 301 drivers/net/wireless/ath/ath9k/dfs.c (rs->rs_phyerr != ATH9K_PHYERR_FALSE_RADAR_EXT)) { rs 304 drivers/net/wireless/ath/ath9k/dfs.c rs->rs_phyerr); rs 309 drivers/net/wireless/ath/ath9k/dfs.c datalen = rs->rs_datalen; rs 315 drivers/net/wireless/ath/ath9k/dfs.c ard.rssi = rs->rs_rssi_ctl[0]; rs 316 drivers/net/wireless/ath/ath9k/dfs.c ard.ext_rssi = rs->rs_rssi_ext[0]; rs 38 drivers/net/wireless/ath/ath9k/dfs.h struct ath_rx_status *rs, u64 mactime); rs 42 drivers/net/wireless/ath/ath9k/dfs.h struct ath_rx_status *rs, u64 mactime) { } rs 338 drivers/net/wireless/ath/ath9k/htc.h struct ath_rx_status *rs); rs 384 drivers/net/wireless/ath/ath9k/htc.h struct ath_rx_status *rs) rs 246 drivers/net/wireless/ath/ath9k/htc_drv_debug.c struct ath_rx_status *rs) rs 248 drivers/net/wireless/ath/ath9k/htc_drv_debug.c ath9k_cmn_debug_stat_rx(&priv->debug.rx_stats, rs); rs 525 drivers/net/wireless/ath/ath9k/mac.c struct ath_rx_status *rs) rs 536 drivers/net/wireless/ath/ath9k/mac.c rs->rs_status = 0; rs 537 drivers/net/wireless/ath/ath9k/mac.c rs->rs_flags = 0; rs 538 drivers/net/wireless/ath/ath9k/mac.c rs->enc_flags = 0; rs 539 drivers/net/wireless/ath/ath9k/mac.c rs->bw = RATE_INFO_BW_20; rs 541 drivers/net/wireless/ath/ath9k/mac.c rs->rs_datalen = ads.ds_rxstatus1 & AR_DataLen; rs 542 drivers/net/wireless/ath/ath9k/mac.c rs->rs_tstamp = ads.AR_RcvTimestamp; rs 545 drivers/net/wireless/ath/ath9k/mac.c rs->rs_rssi = ATH9K_RSSI_BAD; rs 546 drivers/net/wireless/ath/ath9k/mac.c rs->rs_rssi_ctl[0] = ATH9K_RSSI_BAD; rs 547 drivers/net/wireless/ath/ath9k/mac.c rs->rs_rssi_ctl[1] = ATH9K_RSSI_BAD; rs 548 drivers/net/wireless/ath/ath9k/mac.c rs->rs_rssi_ctl[2] = ATH9K_RSSI_BAD; rs 549 drivers/net/wireless/ath/ath9k/mac.c rs->rs_rssi_ext[0] = ATH9K_RSSI_BAD; rs 550 drivers/net/wireless/ath/ath9k/mac.c rs->rs_rssi_ext[1] = ATH9K_RSSI_BAD; rs 551 drivers/net/wireless/ath/ath9k/mac.c rs->rs_rssi_ext[2] = ATH9K_RSSI_BAD; rs 553 drivers/net/wireless/ath/ath9k/mac.c rs->rs_rssi = MS(ads.ds_rxstatus4, AR_RxRSSICombined); rs 554 drivers/net/wireless/ath/ath9k/mac.c rs->rs_rssi_ctl[0] = MS(ads.ds_rxstatus0, rs 556 drivers/net/wireless/ath/ath9k/mac.c rs->rs_rssi_ctl[1] = MS(ads.ds_rxstatus0, rs 558 drivers/net/wireless/ath/ath9k/mac.c rs->rs_rssi_ctl[2] = MS(ads.ds_rxstatus0, rs 560 drivers/net/wireless/ath/ath9k/mac.c rs->rs_rssi_ext[0] = MS(ads.ds_rxstatus4, rs 562 drivers/net/wireless/ath/ath9k/mac.c rs->rs_rssi_ext[1] = MS(ads.ds_rxstatus4, rs 564 drivers/net/wireless/ath/ath9k/mac.c rs->rs_rssi_ext[2] = MS(ads.ds_rxstatus4, rs 568 drivers/net/wireless/ath/ath9k/mac.c rs->rs_keyix = MS(ads.ds_rxstatus8, AR_KeyIdx); rs 570 drivers/net/wireless/ath/ath9k/mac.c rs->rs_keyix = ATH9K_RXKEYIX_INVALID; rs 572 drivers/net/wireless/ath/ath9k/mac.c rs->rs_rate = MS(ads.ds_rxstatus0, AR_RxRate); rs 573 drivers/net/wireless/ath/ath9k/mac.c rs->rs_more = (ads.ds_rxstatus1 & AR_RxMore) ? 1 : 0; rs 575 drivers/net/wireless/ath/ath9k/mac.c rs->rs_firstaggr = (ads.ds_rxstatus8 & AR_RxFirstAggr) ? 1 : 0; rs 576 drivers/net/wireless/ath/ath9k/mac.c rs->rs_isaggr = (ads.ds_rxstatus8 & AR_RxAggr) ? 1 : 0; rs 577 drivers/net/wireless/ath/ath9k/mac.c rs->rs_moreaggr = (ads.ds_rxstatus8 & AR_RxMoreAggr) ? 1 : 0; rs 578 drivers/net/wireless/ath/ath9k/mac.c rs->rs_antenna = MS(ads.ds_rxstatus3, AR_RxAntenna); rs 581 drivers/net/wireless/ath/ath9k/mac.c rs->enc_flags |= rs 583 drivers/net/wireless/ath/ath9k/mac.c rs->bw = (ads.ds_rxstatus3 & AR_2040) ? RATE_INFO_BW_40 : rs 586 drivers/net/wireless/ath/ath9k/mac.c rs->enc_flags |= rs 592 drivers/net/wireless/ath/ath9k/mac.c rs->rs_flags |= ATH9K_RX_DELIM_CRC_PRE; rs 594 drivers/net/wireless/ath/ath9k/mac.c rs->rs_flags |= ATH9K_RX_DELIM_CRC_POST; rs 596 drivers/net/wireless/ath/ath9k/mac.c rs->rs_flags |= ATH9K_RX_DECRYPT_BUSY; rs 606 drivers/net/wireless/ath/ath9k/mac.c rs->rs_status |= ATH9K_RXERR_PHY; rs 608 drivers/net/wireless/ath/ath9k/mac.c rs->rs_phyerr = phyerr; rs 610 drivers/net/wireless/ath/ath9k/mac.c rs->rs_status |= ATH9K_RXERR_CRC; rs 612 drivers/net/wireless/ath/ath9k/mac.c rs->rs_status |= ATH9K_RXERR_DECRYPT; rs 614 drivers/net/wireless/ath/ath9k/mac.c rs->rs_status |= ATH9K_RXERR_MIC; rs 618 drivers/net/wireless/ath/ath9k/mac.c rs->rs_status |= ATH9K_RXERR_CORRUPT_DESC; rs 621 drivers/net/wireless/ath/ath9k/mac.c if (rs->rs_rate >= 0x90) rs 622 drivers/net/wireless/ath/ath9k/mac.c rs->rs_status |= ATH9K_RXERR_CORRUPT_DESC; rs 626 drivers/net/wireless/ath/ath9k/mac.c rs->rs_status |= ATH9K_RXERR_KEYMISS; rs 732 drivers/net/wireless/ath/ath9k/mac.h struct ath_rx_status *rs); rs 628 drivers/net/wireless/ath/ath9k/recv.c struct ath_rx_status *rs, rs 648 drivers/net/wireless/ath/ath9k/recv.c ret = ath9k_hw_process_rxdesc_edma(ah, rs, skb->data); rs 680 drivers/net/wireless/ath/ath9k/recv.c struct ath_rx_status *rs, rs 685 drivers/net/wireless/ath/ath9k/recv.c while (ath_edma_get_buffers(sc, qtype, rs, &bf)) { rs 695 drivers/net/wireless/ath/ath9k/recv.c struct ath_rx_status *rs) rs 725 drivers/net/wireless/ath/ath9k/recv.c ret = ath9k_hw_rxprocdesc(ah, ds, rs); rs 759 drivers/net/wireless/ath/ath9k/recv.c ret = ath9k_hw_rxprocdesc(ah, ds, rs); rs 765 drivers/net/wireless/ath/ath9k/recv.c rs->rs_datalen = 0; rs 766 drivers/net/wireless/ath/ath9k/recv.c rs->rs_more = true; rs 786 drivers/net/wireless/ath/ath9k/recv.c static void ath9k_process_tsf(struct ath_rx_status *rs, rs 792 drivers/net/wireless/ath/ath9k/recv.c rxs->mactime = (tsf & ~0xffffffffULL) | rs->rs_tstamp; rs 793 drivers/net/wireless/ath/ath9k/recv.c if (rs->rs_tstamp > tsf_lower && rs 794 drivers/net/wireless/ath/ath9k/recv.c unlikely(rs->rs_tstamp - tsf_lower > 0x10000000)) rs 797 drivers/net/wireless/ath/ath9k/recv.c if (rs->rs_tstamp < tsf_lower && rs 798 drivers/net/wireless/ath/ath9k/recv.c unlikely(tsf_lower - rs->rs_tstamp > 0x10000000)) rs 963 drivers/net/wireless/ath/ath9k/recv.c struct ath_rx_status *rs) rs 976 drivers/net/wireless/ath/ath9k/recv.c if (sc->rx.defant != rs->rs_antenna) { rs 978 drivers/net/wireless/ath/ath9k/recv.c ath_setdefantenna(sc, rs->rs_antenna); rs 985 drivers/net/wireless/ath/ath9k/recv.c ath_ant_comb_scan(sc, rs); rs 987 drivers/net/wireless/ath/ath9k/recv.c ath_ant_comb_scan(sc, rs); rs 992 drivers/net/wireless/ath/ath9k/recv.c struct ath_rx_status *rs, struct ieee80211_rx_status *rxs) rs 994 drivers/net/wireless/ath/ath9k/recv.c if (rs->rs_isaggr) { rs 999 drivers/net/wireless/ath/ath9k/recv.c if (!rs->rs_moreaggr) { rs 1004 drivers/net/wireless/ath/ath9k/recv.c if (rs->rs_flags & ATH9K_RX_DELIM_CRC_PRE) rs 1010 drivers/net/wireless/ath/ath9k/recv.c struct ath_rx_status *rs, rs 1021 drivers/net/wireless/ath/ath9k/recv.c u16 len = rs->rs_datalen; rs 1048 drivers/net/wireless/ath/ath9k/recv.c phy = IS_CCK_RATE(rs->rs_rate) ? WLAN_RC_PHY_CCK : WLAN_RC_PHY_OFDM; rs 1068 drivers/net/wireless/ath/ath9k/recv.c struct ath_rx_status rs; rs 1090 drivers/net/wireless/ath/ath9k/recv.c memset(&rs, 0, sizeof(rs)); rs 1092 drivers/net/wireless/ath/ath9k/recv.c bf = ath_edma_get_next_rx_buf(sc, &rs, qtype); rs 1094 drivers/net/wireless/ath/ath9k/recv.c bf = ath_get_next_rx_buf(sc, &rs); rs 1115 drivers/net/wireless/ath/ath9k/recv.c retval = ath9k_rx_skb_preprocess(sc, hdr_skb, &rs, rxs, rs 1148 drivers/net/wireless/ath/ath9k/recv.c skb_put(skb, rs.rs_datalen + ah->caps.rx_status_len); rs 1152 drivers/net/wireless/ath/ath9k/recv.c if (!rs.rs_more) rs 1153 drivers/net/wireless/ath/ath9k/recv.c ath9k_cmn_rx_skb_postprocess(common, hdr_skb, &rs, rs 1156 drivers/net/wireless/ath/ath9k/recv.c if (rs.rs_more) { rs 1199 drivers/net/wireless/ath/ath9k/recv.c ath_rx_ps(sc, skb, rs.is_mybeacon); rs 1202 drivers/net/wireless/ath/ath9k/recv.c ath9k_antenna_check(sc, &rs); rs 1203 drivers/net/wireless/ath/ath9k/recv.c ath9k_apply_ampdu_details(sc, &rs, rxs); rs 1204 drivers/net/wireless/ath/ath9k/recv.c ath_debug_rate_stats(sc, &rs, skb); rs 1205 drivers/net/wireless/ath/ath9k/recv.c ath_rx_count_airtime(sc, &rs, skb); rs 1209 drivers/net/wireless/ath/ath9k/recv.c ath_dynack_sample_ack_ts(sc->sc_ah, skb, rs.rs_tstamp); rs 212 drivers/net/wireless/ath/dfs_pattern_detector.c const struct radar_detector_specs *rs = &dpd->radar_spec[i]; rs 213 drivers/net/wireless/ath/dfs_pattern_detector.c struct pri_detector *de = pri_detector_init(rs); rs 272 drivers/net/wireless/ath/dfs_pattern_detector.c struct radar_detector_specs *rs) rs 298 drivers/net/wireless/ath/dfs_pattern_detector.c if (rs != NULL) rs 299 drivers/net/wireless/ath/dfs_pattern_detector.c memcpy(rs, pd->rs, sizeof(*rs)); rs 303 drivers/net/wireless/ath/dfs_pattern_detector.c event->freq, pd->rs->type_id, rs 101 drivers/net/wireless/ath/dfs_pattern_detector.h struct radar_detector_specs *rs); rs 236 drivers/net/wireless/ath/dfs_pri_detector.c if (delta_ts < pde->rs->pri_min) rs 240 drivers/net/wireless/ath/dfs_pri_detector.c if (delta_ts > pde->rs->pri_max) rs 249 drivers/net/wireless/ath/dfs_pri_detector.c ps.pri = GET_PRI_TO_USE(pde->rs->pri_min, rs 250 drivers/net/wireless/ath/dfs_pri_detector.c pde->rs->pri_max, ts - p->ts); rs 251 drivers/net/wireless/ath/dfs_pri_detector.c ps.dur = ps.pri * (pde->rs->ppb - 1) rs 252 drivers/net/wireless/ath/dfs_pri_detector.c + 2 * pde->rs->max_pri_tolerance; rs 265 drivers/net/wireless/ath/dfs_pri_detector.c pde->rs->max_pri_tolerance); rs 322 drivers/net/wireless/ath/dfs_pri_detector.c pde->rs->max_pri_tolerance); rs 350 drivers/net/wireless/ath/dfs_pri_detector.c if ((ps->count >= pde->rs->ppb_thresh) && rs 351 drivers/net/wireless/ath/dfs_pri_detector.c (ps->count * pde->rs->num_pri >= ps->count_falses)) rs 388 drivers/net/wireless/ath/dfs_pri_detector.c const struct radar_detector_specs *rs = de->rs; rs 391 drivers/net/wireless/ath/dfs_pri_detector.c if ((rs->width_min > event->width) || (rs->width_max < event->width)) rs 394 drivers/net/wireless/ath/dfs_pri_detector.c if ((ts - de->last_ts) < rs->max_pri_tolerance) rs 398 drivers/net/wireless/ath/dfs_pri_detector.c if (rs->chirp && rs->chirp != event->chirp) rs 418 drivers/net/wireless/ath/dfs_pri_detector.c struct pri_detector *pri_detector_init(const struct radar_detector_specs *rs) rs 431 drivers/net/wireless/ath/dfs_pri_detector.c de->window_size = rs->pri_max * rs->ppb * rs->num_pri; rs 432 drivers/net/wireless/ath/dfs_pri_detector.c de->max_count = rs->ppb * 2; rs 433 drivers/net/wireless/ath/dfs_pri_detector.c de->rs = rs; rs 65 drivers/net/wireless/ath/dfs_pri_detector.h const struct radar_detector_specs *rs; rs 76 drivers/net/wireless/ath/dfs_pri_detector.h struct pri_detector *pri_detector_init(const struct radar_detector_specs *rs); rs 262 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c static void brcms_set_basic_rate(struct brcm_rateset *rs, u16 rate, bool is_br) rs 266 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c for (i = 0; i < rs->count; i++) { rs 267 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c if (rate != (rs->rates[i] & 0x7f)) rs 271 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c rs->rates[i] |= BRCMS_RATE_FLAG; rs 273 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c rs->rates[i] &= BRCMS_RATE_MASK; rs 632 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c struct brcm_rateset rs; rs 637 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c brcms_c_get_current_rateset(wl->wlc, &rs); rs 647 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c brcms_set_basic_rate(&rs, rate, br_mask & 1); rs 653 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c error = brcms_c_set_rateset(wl->wlc, &rs); rs 3416 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c brcms_default_rateset(struct brcms_c_info *wlc, struct brcms_c_rateset *rs) rs 3418 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c brcms_c_rateset_default(rs, NULL, wlc->band->phytype, rs 3677 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c struct brcms_c_rateset rs; rs 3683 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c brcms_c_rateset_copy(rs_dflt, &rs); rs 3684 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c brcms_c_rateset_mcs_upd(&rs, wlc->stf->txstreams); rs 3687 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c for (i = 0; i < rs.count; i++) { rs 3688 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c rate = rs.rates[i] & BRCMS_RATE_MASK; rs 3698 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c basic_rate = rs.rates[0] & BRCMS_RATE_MASK; rs 5247 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c struct brcms_c_rateset rs; rs 5274 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c memset(&rs, 0, sizeof(rs)); rs 5279 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c brcms_c_rateset_copy(&gphy_legacy_rates, &rs); rs 5311 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c if (!rs.count) rs 5312 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c brcms_c_rateset_copy(&cck_ofdm_rates, &rs); rs 5315 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c for (i = 0; i < rs.count; i++) { rs 5316 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c if (rs.rates[i] == BRCM_RATE_6M rs 5317 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c || rs.rates[i] == BRCM_RATE_12M rs 5318 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c || rs.rates[i] == BRCM_RATE_24M) rs 5319 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c rs.rates[i] |= BRCMS_RATE_FLAG; rs 5324 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c wlc->default_bss->rateset.count = rs.count; rs 5325 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c memcpy(wlc->default_bss->rateset.rates, rs.rates, rs 5362 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c struct brcms_c_rateset rs, new; rs 5365 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c memcpy(&rs, rs_arg, sizeof(struct brcms_c_rateset)); rs 5368 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c if ((rs.count == 0) || (rs.count > BRCMS_NUMRATES)) rs 5373 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c memcpy(&new, &rs, sizeof(struct brcms_c_rateset)); rs 5382 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c memcpy(&new, &rs, sizeof(struct brcms_c_rateset)); rs 5472 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c struct brcms_c_rateset *rs; rs 5475 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c rs = &wlc->bsscfg->current_bss->rateset; rs 5477 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c rs = &wlc->default_bss->rateset; rs 5480 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c currs->count = rs->count; rs 5481 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c memcpy(&currs->rates, &rs->rates, rs->count); rs 5484 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c int brcms_c_set_rateset(struct brcms_c_info *wlc, struct brcm_rateset *rs) rs 5489 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c if (rs->count > BRCMS_NUMRATES) rs 5495 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c internal_rs.count = rs->count; rs 5496 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c memcpy(&internal_rs.rates, &rs->rates, internal_rs.count); rs 7243 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c struct brcms_c_rateset rs; rs 7254 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c brcms_c_rateset_copy(rs_dflt, &rs); rs 7255 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c brcms_c_rateset_mcs_upd(&rs, wlc->stf->txstreams); rs 7261 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c for (i = 0; i < rs.count; i++) { rs 7262 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c rate = rs.rates[i] & BRCMS_RATE_MASK; rs 316 drivers/net/wireless/broadcom/brcm80211/brcmsmac/pub.h int brcms_c_set_rateset(struct brcms_c_info *wlc, struct brcm_rateset *rs); rs 272 drivers/net/wireless/broadcom/brcm80211/brcmsmac/rate.c static bool brcms_c_rateset_valid(struct brcms_c_rateset *rs, bool check_brate) rs 276 drivers/net/wireless/broadcom/brcm80211/brcmsmac/rate.c if (!rs->count) rs 283 drivers/net/wireless/broadcom/brcm80211/brcmsmac/rate.c for (idx = 0; idx < rs->count; idx++) { rs 284 drivers/net/wireless/broadcom/brcm80211/brcmsmac/rate.c if (rs->rates[idx] & BRCMS_RATE_FLAG) rs 290 drivers/net/wireless/broadcom/brcm80211/brcmsmac/rate.c void brcms_c_rateset_mcs_upd(struct brcms_c_rateset *rs, u8 txstreams) rs 294 drivers/net/wireless/broadcom/brcm80211/brcmsmac/rate.c rs->mcs[i] = 0; rs 302 drivers/net/wireless/broadcom/brcm80211/brcmsmac/rate.c brcms_c_rate_hwrs_filter_sort_validate(struct brcms_c_rateset *rs, rs 312 drivers/net/wireless/broadcom/brcm80211/brcmsmac/rate.c count = rs->count; rs 316 drivers/net/wireless/broadcom/brcm80211/brcmsmac/rate.c r = (int)rs->rates[i] & BRCMS_RATE_MASK; rs 319 drivers/net/wireless/broadcom/brcm80211/brcmsmac/rate.c rateset[r] = rs->rates[i]; /* preserve basic bit! */ rs 327 drivers/net/wireless/broadcom/brcm80211/brcmsmac/rate.c rs->rates[count++] = rateset[r]; rs 330 drivers/net/wireless/broadcom/brcm80211/brcmsmac/rate.c rs->count = count; rs 334 drivers/net/wireless/broadcom/brcm80211/brcmsmac/rate.c rs->mcs[i] = (rs->mcs[i] & hw_rs->mcs[i]); rs 336 drivers/net/wireless/broadcom/brcm80211/brcmsmac/rate.c if (brcms_c_rateset_valid(rs, check_brate)) rs 219 drivers/net/wireless/broadcom/brcm80211/brcmsmac/rate.h bool brcms_c_rate_hwrs_filter_sort_validate(struct brcms_c_rateset *rs, rs 240 drivers/net/wireless/broadcom/brcm80211/brcmsmac/rate.h void brcms_c_rateset_mcs_upd(struct brcms_c_rateset *rs, u8 txstreams); rs 951 drivers/net/wireless/mediatek/mt76/mt7603/mac.c struct mt7603_rate_set *rs; rs 1007 drivers/net/wireless/mediatek/mt76/mt7603/mac.c rs = &sta->rateset[rs_idx]; rs 1009 drivers/net/wireless/mediatek/mt76/mt7603/mac.c if (!first_idx && rs->probe_rate.idx >= 0) { rs 1010 drivers/net/wireless/mediatek/mt76/mt7603/mac.c info->status.rates[0] = rs->probe_rate; rs 1020 drivers/net/wireless/mediatek/mt76/mt7603/mac.c info->status.rates[0] = rs->rates[first_idx / 2]; rs 1028 drivers/net/wireless/mediatek/mt76/mt7603/mac.c cur_rate = &rs->rates[idx / 2]; rs 839 drivers/net/wireless/mediatek/mt76/mt7615/mac.c struct mt7615_rate_set *rs; rs 890 drivers/net/wireless/mediatek/mt76/mt7615/mac.c rs = &sta->rateset[rs_idx]; rs 892 drivers/net/wireless/mediatek/mt76/mt7615/mac.c if (!first_idx && rs->probe_rate.idx >= 0) { rs 893 drivers/net/wireless/mediatek/mt76/mt7615/mac.c info->status.rates[0] = rs->probe_rate; rs 902 drivers/net/wireless/mediatek/mt76/mt7615/mac.c info->status.rates[0] = rs->rates[first_idx / 2]; rs 910 drivers/net/wireless/mediatek/mt76/mt7615/mac.c cur_rate = &rs->rates[idx / 2]; rs 1262 drivers/net/wireless/realtek/rtw88/phy.c u8 bw, u8 rs, u8 ch, s8 pwr_limit) rs 1274 drivers/net/wireless/realtek/rtw88/phy.c rs >= RTW_RATE_SECTION_MAX || ch_idx < 0) { rs 1277 drivers/net/wireless/realtek/rtw88/phy.c regd, band, bw, rs, ch_idx, pwr_limit); rs 1282 drivers/net/wireless/realtek/rtw88/phy.c hal->tx_pwr_limit_2g[regd][bw][rs][ch_idx] = pwr_limit; rs 1283 drivers/net/wireless/realtek/rtw88/phy.c ww = hal->tx_pwr_limit_2g[RTW_REGD_WW][bw][rs][ch_idx]; rs 1285 drivers/net/wireless/realtek/rtw88/phy.c hal->tx_pwr_limit_2g[RTW_REGD_WW][bw][rs][ch_idx] = ww; rs 1287 drivers/net/wireless/realtek/rtw88/phy.c hal->tx_pwr_limit_5g[regd][bw][rs][ch_idx] = pwr_limit; rs 1288 drivers/net/wireless/realtek/rtw88/phy.c ww = hal->tx_pwr_limit_5g[RTW_REGD_WW][bw][rs][ch_idx]; rs 1290 drivers/net/wireless/realtek/rtw88/phy.c hal->tx_pwr_limit_5g[RTW_REGD_WW][bw][rs][ch_idx] = ww; rs 1367 drivers/net/wireless/realtek/rtw88/phy.c p->bw, p->rs, p->ch, p->txpwr_lmt); rs 1682 drivers/net/wireless/realtek/rtw88/phy.c u8 rs; rs 1691 drivers/net/wireless/realtek/rtw88/phy.c rs = RTW_RATE_SECTION_CCK; rs 1693 drivers/net/wireless/realtek/rtw88/phy.c rs = RTW_RATE_SECTION_OFDM; rs 1695 drivers/net/wireless/realtek/rtw88/phy.c rs = RTW_RATE_SECTION_HT_1S; rs 1697 drivers/net/wireless/realtek/rtw88/phy.c rs = RTW_RATE_SECTION_HT_2S; rs 1699 drivers/net/wireless/realtek/rtw88/phy.c rs = RTW_RATE_SECTION_VHT_1S; rs 1701 drivers/net/wireless/realtek/rtw88/phy.c rs = RTW_RATE_SECTION_VHT_2S; rs 1706 drivers/net/wireless/realtek/rtw88/phy.c if (rs == RTW_RATE_SECTION_CCK || rs == RTW_RATE_SECTION_OFDM) rs 1710 drivers/net/wireless/realtek/rtw88/phy.c if (rs == RTW_RATE_SECTION_HT_1S || rs == RTW_RATE_SECTION_HT_2S) rs 1722 drivers/net/wireless/realtek/rtw88/phy.c hal->tx_pwr_limit_2g[regd][cur_bw][rs][ch_idx] : rs 1723 drivers/net/wireless/realtek/rtw88/phy.c hal->tx_pwr_limit_5g[regd][cur_bw][rs][ch_idx]; rs 1794 drivers/net/wireless/realtek/rtw88/phy.c u8 ch, u8 path, u8 rs) rs 1805 drivers/net/wireless/realtek/rtw88/phy.c if (rs >= RTW_RATE_SECTION_MAX) rs 1808 drivers/net/wireless/realtek/rtw88/phy.c rates = rtw_rate_section[rs]; rs 1809 drivers/net/wireless/realtek/rtw88/phy.c size = rtw_rate_size[rs]; rs 1828 drivers/net/wireless/realtek/rtw88/phy.c u8 rs; rs 1832 drivers/net/wireless/realtek/rtw88/phy.c rs = RTW_RATE_SECTION_CCK; rs 1834 drivers/net/wireless/realtek/rtw88/phy.c rs = RTW_RATE_SECTION_OFDM; rs 1836 drivers/net/wireless/realtek/rtw88/phy.c for (; rs < RTW_RATE_SECTION_MAX; rs++) rs 1837 drivers/net/wireless/realtek/rtw88/phy.c rtw_phy_set_tx_power_index_by_rs(rtwdev, ch, path, rs); rs 1857 drivers/net/wireless/realtek/rtw88/phy.c u8 rs, u8 size, u8 *rates) rs 1863 drivers/net/wireless/realtek/rtw88/phy.c if (rs >= RTW_RATE_SECTION_VHT_1S) rs 1869 drivers/net/wireless/realtek/rtw88/phy.c hal->tx_pwr_by_rate_base_2g[path][rs] = base_2g; rs 1870 drivers/net/wireless/realtek/rtw88/phy.c hal->tx_pwr_by_rate_base_5g[path][rs] = base_5g; rs 1905 drivers/net/wireless/realtek/rtw88/phy.c __rtw_phy_tx_power_limit_config(struct rtw_hal *hal, u8 regd, u8 bw, u8 rs) rs 1911 drivers/net/wireless/realtek/rtw88/phy.c base = hal->tx_pwr_by_rate_base_2g[0][rs]; rs 1912 drivers/net/wireless/realtek/rtw88/phy.c hal->tx_pwr_limit_2g[regd][bw][rs][ch] -= base; rs 1916 drivers/net/wireless/realtek/rtw88/phy.c base = hal->tx_pwr_by_rate_base_5g[0][rs]; rs 1917 drivers/net/wireless/realtek/rtw88/phy.c hal->tx_pwr_limit_5g[regd][bw][rs][ch] -= base; rs 1923 drivers/net/wireless/realtek/rtw88/phy.c u8 regd, bw, rs; rs 1930 drivers/net/wireless/realtek/rtw88/phy.c for (rs = 0; rs < RTW_RATE_SECTION_MAX; rs++) rs 1931 drivers/net/wireless/realtek/rtw88/phy.c __rtw_phy_tx_power_limit_config(hal, regd, bw, rs); rs 1935 drivers/net/wireless/realtek/rtw88/phy.c u8 regd, u8 bw, u8 rs) rs 1943 drivers/net/wireless/realtek/rtw88/phy.c hal->tx_pwr_limit_2g[regd][bw][rs][ch] = max_power_index; rs 1947 drivers/net/wireless/realtek/rtw88/phy.c hal->tx_pwr_limit_5g[regd][bw][rs][ch] = max_power_index; rs 1953 drivers/net/wireless/realtek/rtw88/phy.c u8 regd, path, rate, rs, bw; rs 1966 drivers/net/wireless/realtek/rtw88/phy.c for (rs = 0; rs < RTW_RATE_SECTION_MAX; rs++) rs 1968 drivers/net/wireless/realtek/rtw88/phy.c rs); rs 52 drivers/net/wireless/realtek/rtw88/phy.h u8 rs; rs 867 drivers/net/wireless/realtek/rtw88/rtw8822b.c rtw8822b_set_tx_power_index_by_rate(struct rtw_dev *rtwdev, u8 path, u8 rs) rs 875 drivers/net/wireless/realtek/rtw88/rtw8822b.c for (j = 0; j < rtw_rate_size[rs]; j++) { rs 876 drivers/net/wireless/realtek/rtw88/rtw8822b.c rate = rtw_rate_section[rs][j]; rs 892 drivers/net/wireless/realtek/rtw88/rtw8822b.c int rs, path; rs 895 drivers/net/wireless/realtek/rtw88/rtw8822b.c for (rs = 0; rs < RTW_RATE_SECTION_MAX; rs++) rs 896 drivers/net/wireless/realtek/rtw88/rtw8822b.c rtw8822b_set_tx_power_index_by_rate(rtwdev, path, rs); rs 1780 drivers/net/wireless/realtek/rtw88/rtw8822c.c u8 rs, rate, j; rs 1790 drivers/net/wireless/realtek/rtw88/rtw8822c.c for (rs = 0; rs < RTW_RATE_SECTION_MAX; rs++) { rs 1791 drivers/net/wireless/realtek/rtw88/rtw8822c.c for (j = 0; j < rtw_rate_size[rs]; j++) { rs 1792 drivers/net/wireless/realtek/rtw88/rtw8822c.c rate = rtw_rate_section[rs][j]; rs 1795 drivers/net/wireless/realtek/rtw88/rtw8822c.c if (rs == 0) { rs 58 drivers/pci/hotplug/ibmphp_res.c struct resource_node *rs; rs 65 drivers/pci/hotplug/ibmphp_res.c rs = kzalloc(sizeof(struct resource_node), GFP_KERNEL); rs 66 drivers/pci/hotplug/ibmphp_res.c if (!rs) rs 69 drivers/pci/hotplug/ibmphp_res.c rs->busno = curr->bus_num; rs 70 drivers/pci/hotplug/ibmphp_res.c rs->devfunc = curr->dev_fun; rs 71 drivers/pci/hotplug/ibmphp_res.c rs->start = curr->start_addr; rs 72 drivers/pci/hotplug/ibmphp_res.c rs->end = curr->end_addr; rs 73 drivers/pci/hotplug/ibmphp_res.c rs->len = curr->end_addr - curr->start_addr + 1; rs 74 drivers/pci/hotplug/ibmphp_res.c return rs; rs 179 drivers/platform/chrome/wilco_ec/debugfs.c struct hi_gpio_status_response rs; rs 191 drivers/platform/chrome/wilco_ec/debugfs.c msg.response_data = &rs; rs 192 drivers/platform/chrome/wilco_ec/debugfs.c msg.response_size = sizeof(rs); rs 196 drivers/platform/chrome/wilco_ec/debugfs.c if (rs.status) rs 199 drivers/platform/chrome/wilco_ec/debugfs.c *val = rs.val; rs 119 drivers/platform/chrome/wilco_ec/mailbox.c struct wilco_ec_response *rs; rs 151 drivers/platform/chrome/wilco_ec/mailbox.c rs = ec->data_buffer; rs 153 drivers/platform/chrome/wilco_ec/mailbox.c sizeof(*rs) + EC_MAILBOX_DATA_SIZE, rs 154 drivers/platform/chrome/wilco_ec/mailbox.c (u8 *)rs); rs 156 drivers/platform/chrome/wilco_ec/mailbox.c dev_dbg(ec->dev, "bad packet checksum 0x%02x\n", rs->checksum); rs 160 drivers/platform/chrome/wilco_ec/mailbox.c if (rs->result) { rs 161 drivers/platform/chrome/wilco_ec/mailbox.c dev_dbg(ec->dev, "EC reported failure: 0x%02x\n", rs->result); rs 165 drivers/platform/chrome/wilco_ec/mailbox.c if (rs->data_size != EC_MAILBOX_DATA_SIZE) { rs 167 drivers/platform/chrome/wilco_ec/mailbox.c rs->data_size, EC_MAILBOX_DATA_SIZE); rs 171 drivers/platform/chrome/wilco_ec/mailbox.c if (rs->data_size < msg->response_size) { rs 173 drivers/platform/chrome/wilco_ec/mailbox.c rs->data_size, msg->response_size); rs 177 drivers/platform/chrome/wilco_ec/mailbox.c memcpy(msg->response_data, rs->data, msg->response_size); rs 179 drivers/platform/chrome/wilco_ec/mailbox.c return rs->data_size; rs 33 drivers/platform/chrome/wilco_ec/properties.c struct ec_property_response *rs) rs 42 drivers/platform/chrome/wilco_ec/properties.c ec_msg.response_data = rs; rs 43 drivers/platform/chrome/wilco_ec/properties.c ec_msg.response_size = sizeof(*rs); rs 48 drivers/platform/chrome/wilco_ec/properties.c if (rs->op != rq->op) rs 50 drivers/platform/chrome/wilco_ec/properties.c if (memcmp(rq->property_id, rs->property_id, sizeof(rs->property_id))) rs 60 drivers/platform/chrome/wilco_ec/properties.c struct ec_property_response rs; rs 67 drivers/platform/chrome/wilco_ec/properties.c ret = send_property_msg(ec, &rq, &rs); rs 71 drivers/platform/chrome/wilco_ec/properties.c prop_msg->length = rs.length; rs 72 drivers/platform/chrome/wilco_ec/properties.c memcpy(prop_msg->data, rs.data, rs.length); rs 82 drivers/platform/chrome/wilco_ec/properties.c struct ec_property_response rs; rs 91 drivers/platform/chrome/wilco_ec/properties.c ret = send_property_msg(ec, &rq, &rs); rs 94 drivers/platform/chrome/wilco_ec/properties.c if (rs.length != prop_msg->length) rs 289 drivers/s390/cio/chsc.c u8 rs; /* reporting source */ rs 393 drivers/s390/cio/chsc.c sei_area->rs, sei_area->rsid, sei_area->ccdf[0]); rs 410 drivers/s390/cio/chsc.c sei_area->rs, sei_area->rsid, lir->ic, iuparams, rs 416 drivers/s390/cio/chsc.c sei_area->rs, sei_area->rsid, lir->ic, iuparams, rs 432 drivers/s390/cio/chsc.c "rs_id=%04x)\n", sei_area->rs, sei_area->rsid); rs 433 drivers/s390/cio/chsc.c if (sei_area->rs != 4) rs 472 drivers/s390/cio/chsc.c if (sei_area->rs != 0) rs 508 drivers/s390/cio/chsc.c if (sei_area->rs != 0) rs 537 drivers/s390/cio/chsc.c if (sei_area->rs != 7) rs 551 drivers/s390/cio/chsc.c if (sei_area->rs != 7) rs 563 drivers/s390/cio/chsc.c if (sei_area->rs != 5) rs 67 drivers/spi/spi-mt7621.c static inline u32 mt7621_spi_read(struct mt7621_spi *rs, u32 reg) rs 69 drivers/spi/spi-mt7621.c return ioread32(rs->base + reg); rs 72 drivers/spi/spi-mt7621.c static inline void mt7621_spi_write(struct mt7621_spi *rs, u32 reg, u32 val) rs 74 drivers/spi/spi-mt7621.c iowrite32(val, rs->base + reg); rs 79 drivers/spi/spi-mt7621.c struct mt7621_spi *rs = spidev_to_mt7621_spi(spi); rs 89 drivers/spi/spi-mt7621.c master = mt7621_spi_read(rs, MT7621_SPI_MASTER); rs 92 drivers/spi/spi-mt7621.c mt7621_spi_write(rs, MT7621_SPI_MASTER, master); rs 94 drivers/spi/spi-mt7621.c rs->pending_write = 0; rs 98 drivers/spi/spi-mt7621.c mt7621_spi_write(rs, MT7621_SPI_POLAR, polar); rs 103 drivers/spi/spi-mt7621.c struct mt7621_spi *rs = spidev_to_mt7621_spi(spi); rs 109 drivers/spi/spi-mt7621.c rate = DIV_ROUND_UP(rs->sys_freq, speed); rs 118 drivers/spi/spi-mt7621.c reg = mt7621_spi_read(rs, MT7621_SPI_MASTER); rs 121 drivers/spi/spi-mt7621.c rs->speed = speed; rs 134 drivers/spi/spi-mt7621.c mt7621_spi_write(rs, MT7621_SPI_MASTER, reg); rs 139 drivers/spi/spi-mt7621.c static inline int mt7621_spi_wait_till_ready(struct mt7621_spi *rs) rs 146 drivers/spi/spi-mt7621.c status = mt7621_spi_read(rs, MT7621_SPI_TRANS); rs 156 drivers/spi/spi-mt7621.c static void mt7621_spi_read_half_duplex(struct mt7621_spi *rs, rs 166 drivers/spi/spi-mt7621.c tx_len = rs->pending_write; rs 167 drivers/spi/spi-mt7621.c rs->pending_write = 0; rs 177 drivers/spi/spi-mt7621.c mt7621_spi_write(rs, MT7621_SPI_MOREBUF, val); rs 181 drivers/spi/spi-mt7621.c val = mt7621_spi_read(rs, MT7621_SPI_TRANS); rs 183 drivers/spi/spi-mt7621.c mt7621_spi_write(rs, MT7621_SPI_TRANS, val); rs 185 drivers/spi/spi-mt7621.c mt7621_spi_wait_till_ready(rs); rs 189 drivers/spi/spi-mt7621.c val = mt7621_spi_read(rs, MT7621_SPI_DATA0 + i); rs 198 drivers/spi/spi-mt7621.c static inline void mt7621_spi_flush(struct mt7621_spi *rs) rs 200 drivers/spi/spi-mt7621.c mt7621_spi_read_half_duplex(rs, 0, NULL); rs 203 drivers/spi/spi-mt7621.c static void mt7621_spi_write_half_duplex(struct mt7621_spi *rs, rs 206 drivers/spi/spi-mt7621.c int len = rs->pending_write; rs 210 drivers/spi/spi-mt7621.c val = mt7621_spi_read(rs, MT7621_SPI_OPCODE + (len & ~3)); rs 219 drivers/spi/spi-mt7621.c rs->pending_write = len; rs 220 drivers/spi/spi-mt7621.c mt7621_spi_flush(rs); rs 230 drivers/spi/spi-mt7621.c mt7621_spi_write(rs, MT7621_SPI_OPCODE + len - 4, val); rs 241 drivers/spi/spi-mt7621.c mt7621_spi_write(rs, MT7621_SPI_OPCODE + (len & ~3), val); rs 244 drivers/spi/spi-mt7621.c rs->pending_write = len; rs 250 drivers/spi/spi-mt7621.c struct mt7621_spi *rs = spi_controller_get_devdata(master); rs 256 drivers/spi/spi-mt7621.c mt7621_spi_wait_till_ready(rs); rs 283 drivers/spi/spi-mt7621.c mt7621_spi_read_half_duplex(rs, t->len, t->rx_buf); rs 285 drivers/spi/spi-mt7621.c mt7621_spi_write_half_duplex(rs, t->len, t->tx_buf); rs 291 drivers/spi/spi-mt7621.c mt7621_spi_flush(rs); rs 303 drivers/spi/spi-mt7621.c struct mt7621_spi *rs = spidev_to_mt7621_spi(spi); rs 306 drivers/spi/spi-mt7621.c (spi->max_speed_hz > (rs->sys_freq / 2))) rs 307 drivers/spi/spi-mt7621.c spi->max_speed_hz = rs->sys_freq / 2; rs 309 drivers/spi/spi-mt7621.c if (spi->max_speed_hz < (rs->sys_freq / 4097)) { rs 328 drivers/spi/spi-mt7621.c struct mt7621_spi *rs; rs 353 drivers/spi/spi-mt7621.c master = spi_alloc_master(&pdev->dev, sizeof(*rs)); rs 369 drivers/spi/spi-mt7621.c rs = spi_controller_get_devdata(master); rs 370 drivers/spi/spi-mt7621.c rs->base = base; rs 371 drivers/spi/spi-mt7621.c rs->clk = clk; rs 372 drivers/spi/spi-mt7621.c rs->master = master; rs 373 drivers/spi/spi-mt7621.c rs->sys_freq = clk_get_rate(rs->clk); rs 374 drivers/spi/spi-mt7621.c rs->pending_write = 0; rs 375 drivers/spi/spi-mt7621.c dev_info(&pdev->dev, "sys_freq: %u\n", rs->sys_freq); rs 389 drivers/spi/spi-mt7621.c struct mt7621_spi *rs; rs 392 drivers/spi/spi-mt7621.c rs = spi_controller_get_devdata(master); rs 394 drivers/spi/spi-mt7621.c clk_disable_unprepare(rs->clk); rs 188 drivers/spi/spi-rockchip.c static inline void spi_enable_chip(struct rockchip_spi *rs, bool enable) rs 190 drivers/spi/spi-rockchip.c writel_relaxed((enable ? 1U : 0U), rs->regs + ROCKCHIP_SPI_SSIENR); rs 193 drivers/spi/spi-rockchip.c static inline void wait_for_idle(struct rockchip_spi *rs) rs 198 drivers/spi/spi-rockchip.c if (!(readl_relaxed(rs->regs + ROCKCHIP_SPI_SR) & SR_BUSY)) rs 202 drivers/spi/spi-rockchip.c dev_warn(rs->dev, "spi controller is in busy state!\n"); rs 205 drivers/spi/spi-rockchip.c static u32 get_fifo_len(struct rockchip_spi *rs) rs 210 drivers/spi/spi-rockchip.c writel_relaxed(fifo, rs->regs + ROCKCHIP_SPI_TXFTLR); rs 211 drivers/spi/spi-rockchip.c if (fifo != readl_relaxed(rs->regs + ROCKCHIP_SPI_TXFTLR)) rs 215 drivers/spi/spi-rockchip.c writel_relaxed(0, rs->regs + ROCKCHIP_SPI_TXFTLR); rs 223 drivers/spi/spi-rockchip.c struct rockchip_spi *rs = spi_master_get_devdata(master); rs 227 drivers/spi/spi-rockchip.c if (cs_asserted == rs->cs_asserted[spi->chip_select]) rs 232 drivers/spi/spi-rockchip.c pm_runtime_get_sync(rs->dev); rs 234 drivers/spi/spi-rockchip.c ROCKCHIP_SPI_SET_BITS(rs->regs + ROCKCHIP_SPI_SER, rs 237 drivers/spi/spi-rockchip.c ROCKCHIP_SPI_CLR_BITS(rs->regs + ROCKCHIP_SPI_SER, rs 241 drivers/spi/spi-rockchip.c pm_runtime_put(rs->dev); rs 244 drivers/spi/spi-rockchip.c rs->cs_asserted[spi->chip_select] = cs_asserted; rs 250 drivers/spi/spi-rockchip.c struct rockchip_spi *rs = spi_master_get_devdata(master); rs 255 drivers/spi/spi-rockchip.c spi_enable_chip(rs, false); rs 258 drivers/spi/spi-rockchip.c writel_relaxed(0, rs->regs + ROCKCHIP_SPI_IMR); rs 260 drivers/spi/spi-rockchip.c if (atomic_read(&rs->state) & TXDMA) rs 263 drivers/spi/spi-rockchip.c if (atomic_read(&rs->state) & RXDMA) rs 267 drivers/spi/spi-rockchip.c static void rockchip_spi_pio_writer(struct rockchip_spi *rs) rs 269 drivers/spi/spi-rockchip.c u32 tx_free = rs->fifo_len - readl_relaxed(rs->regs + ROCKCHIP_SPI_TXFLR); rs 270 drivers/spi/spi-rockchip.c u32 words = min(rs->tx_left, tx_free); rs 272 drivers/spi/spi-rockchip.c rs->tx_left -= words; rs 276 drivers/spi/spi-rockchip.c if (rs->n_bytes == 1) rs 277 drivers/spi/spi-rockchip.c txw = *(u8 *)rs->tx; rs 279 drivers/spi/spi-rockchip.c txw = *(u16 *)rs->tx; rs 281 drivers/spi/spi-rockchip.c writel_relaxed(txw, rs->regs + ROCKCHIP_SPI_TXDR); rs 282 drivers/spi/spi-rockchip.c rs->tx += rs->n_bytes; rs 286 drivers/spi/spi-rockchip.c static void rockchip_spi_pio_reader(struct rockchip_spi *rs) rs 288 drivers/spi/spi-rockchip.c u32 words = readl_relaxed(rs->regs + ROCKCHIP_SPI_RXFLR); rs 289 drivers/spi/spi-rockchip.c u32 rx_left = rs->rx_left - words; rs 297 drivers/spi/spi-rockchip.c u32 ftl = readl_relaxed(rs->regs + ROCKCHIP_SPI_RXFTLR) + 1; rs 301 drivers/spi/spi-rockchip.c words = rs->rx_left - rx_left; rs 305 drivers/spi/spi-rockchip.c rs->rx_left = rx_left; rs 307 drivers/spi/spi-rockchip.c u32 rxw = readl_relaxed(rs->regs + ROCKCHIP_SPI_RXDR); rs 309 drivers/spi/spi-rockchip.c if (!rs->rx) rs 312 drivers/spi/spi-rockchip.c if (rs->n_bytes == 1) rs 313 drivers/spi/spi-rockchip.c *(u8 *)rs->rx = (u8)rxw; rs 315 drivers/spi/spi-rockchip.c *(u16 *)rs->rx = (u16)rxw; rs 316 drivers/spi/spi-rockchip.c rs->rx += rs->n_bytes; rs 323 drivers/spi/spi-rockchip.c struct rockchip_spi *rs = spi_master_get_devdata(master); rs 325 drivers/spi/spi-rockchip.c if (rs->tx_left) rs 326 drivers/spi/spi-rockchip.c rockchip_spi_pio_writer(rs); rs 328 drivers/spi/spi-rockchip.c rockchip_spi_pio_reader(rs); rs 329 drivers/spi/spi-rockchip.c if (!rs->rx_left) { rs 330 drivers/spi/spi-rockchip.c spi_enable_chip(rs, false); rs 331 drivers/spi/spi-rockchip.c writel_relaxed(0, rs->regs + ROCKCHIP_SPI_IMR); rs 338 drivers/spi/spi-rockchip.c static int rockchip_spi_prepare_irq(struct rockchip_spi *rs, rs 341 drivers/spi/spi-rockchip.c rs->tx = xfer->tx_buf; rs 342 drivers/spi/spi-rockchip.c rs->rx = xfer->rx_buf; rs 343 drivers/spi/spi-rockchip.c rs->tx_left = rs->tx ? xfer->len / rs->n_bytes : 0; rs 344 drivers/spi/spi-rockchip.c rs->rx_left = xfer->len / rs->n_bytes; rs 346 drivers/spi/spi-rockchip.c writel_relaxed(INT_RF_FULL, rs->regs + ROCKCHIP_SPI_IMR); rs 347 drivers/spi/spi-rockchip.c spi_enable_chip(rs, true); rs 349 drivers/spi/spi-rockchip.c if (rs->tx_left) rs 350 drivers/spi/spi-rockchip.c rockchip_spi_pio_writer(rs); rs 359 drivers/spi/spi-rockchip.c struct rockchip_spi *rs = spi_master_get_devdata(master); rs 360 drivers/spi/spi-rockchip.c int state = atomic_fetch_andnot(RXDMA, &rs->state); rs 365 drivers/spi/spi-rockchip.c spi_enable_chip(rs, false); rs 372 drivers/spi/spi-rockchip.c struct rockchip_spi *rs = spi_master_get_devdata(master); rs 373 drivers/spi/spi-rockchip.c int state = atomic_fetch_andnot(TXDMA, &rs->state); rs 379 drivers/spi/spi-rockchip.c wait_for_idle(rs); rs 381 drivers/spi/spi-rockchip.c spi_enable_chip(rs, false); rs 385 drivers/spi/spi-rockchip.c static int rockchip_spi_prepare_dma(struct rockchip_spi *rs, rs 390 drivers/spi/spi-rockchip.c atomic_set(&rs->state, 0); rs 396 drivers/spi/spi-rockchip.c .src_addr = rs->dma_addr_rx, rs 397 drivers/spi/spi-rockchip.c .src_addr_width = rs->n_bytes, rs 418 drivers/spi/spi-rockchip.c .dst_addr = rs->dma_addr_tx, rs 419 drivers/spi/spi-rockchip.c .dst_addr_width = rs->n_bytes, rs 420 drivers/spi/spi-rockchip.c .dst_maxburst = rs->fifo_len / 4, rs 441 drivers/spi/spi-rockchip.c atomic_or(RXDMA, &rs->state); rs 446 drivers/spi/spi-rockchip.c spi_enable_chip(rs, true); rs 449 drivers/spi/spi-rockchip.c atomic_or(TXDMA, &rs->state); rs 458 drivers/spi/spi-rockchip.c static void rockchip_spi_config(struct rockchip_spi *rs, rs 469 drivers/spi/spi-rockchip.c cr0 |= rs->rsd << CR0_RSD_OFFSET; rs 509 drivers/spi/spi-rockchip.c writel_relaxed(cr0, rs->regs + ROCKCHIP_SPI_CTRLR0); rs 510 drivers/spi/spi-rockchip.c writel_relaxed(cr1, rs->regs + ROCKCHIP_SPI_CTRLR1); rs 516 drivers/spi/spi-rockchip.c if (xfer->len < rs->fifo_len) rs 517 drivers/spi/spi-rockchip.c writel_relaxed(xfer->len - 1, rs->regs + ROCKCHIP_SPI_RXFTLR); rs 519 drivers/spi/spi-rockchip.c writel_relaxed(rs->fifo_len / 2 - 1, rs->regs + ROCKCHIP_SPI_RXFTLR); rs 521 drivers/spi/spi-rockchip.c writel_relaxed(rs->fifo_len / 2, rs->regs + ROCKCHIP_SPI_DMATDLR); rs 522 drivers/spi/spi-rockchip.c writel_relaxed(0, rs->regs + ROCKCHIP_SPI_DMARDLR); rs 523 drivers/spi/spi-rockchip.c writel_relaxed(dmacr, rs->regs + ROCKCHIP_SPI_DMACR); rs 529 drivers/spi/spi-rockchip.c writel_relaxed(2 * DIV_ROUND_UP(rs->freq, 2 * xfer->speed_hz), rs 530 drivers/spi/spi-rockchip.c rs->regs + ROCKCHIP_SPI_BAUDR); rs 543 drivers/spi/spi-rockchip.c struct rockchip_spi *rs = spi_master_get_devdata(master); rs 546 drivers/spi/spi-rockchip.c WARN_ON(readl_relaxed(rs->regs + ROCKCHIP_SPI_SSIENR) && rs 547 drivers/spi/spi-rockchip.c (readl_relaxed(rs->regs + ROCKCHIP_SPI_SR) & SR_BUSY)); rs 550 drivers/spi/spi-rockchip.c dev_err(rs->dev, "No buffer for transfer\n"); rs 555 drivers/spi/spi-rockchip.c dev_err(rs->dev, "Transfer is too long (%d)\n", xfer->len); rs 559 drivers/spi/spi-rockchip.c rs->n_bytes = xfer->bits_per_word <= 8 ? 1 : 2; rs 563 drivers/spi/spi-rockchip.c rockchip_spi_config(rs, spi, xfer, use_dma); rs 566 drivers/spi/spi-rockchip.c return rockchip_spi_prepare_dma(rs, master, xfer); rs 568 drivers/spi/spi-rockchip.c return rockchip_spi_prepare_irq(rs, xfer); rs 575 drivers/spi/spi-rockchip.c struct rockchip_spi *rs = spi_master_get_devdata(master); rs 582 drivers/spi/spi-rockchip.c return xfer->len / bytes_per_word >= rs->fifo_len; rs 588 drivers/spi/spi-rockchip.c struct rockchip_spi *rs; rs 599 drivers/spi/spi-rockchip.c rs = spi_master_get_devdata(master); rs 603 drivers/spi/spi-rockchip.c rs->regs = devm_ioremap_resource(&pdev->dev, mem); rs 604 drivers/spi/spi-rockchip.c if (IS_ERR(rs->regs)) { rs 605 drivers/spi/spi-rockchip.c ret = PTR_ERR(rs->regs); rs 609 drivers/spi/spi-rockchip.c rs->apb_pclk = devm_clk_get(&pdev->dev, "apb_pclk"); rs 610 drivers/spi/spi-rockchip.c if (IS_ERR(rs->apb_pclk)) { rs 612 drivers/spi/spi-rockchip.c ret = PTR_ERR(rs->apb_pclk); rs 616 drivers/spi/spi-rockchip.c rs->spiclk = devm_clk_get(&pdev->dev, "spiclk"); rs 617 drivers/spi/spi-rockchip.c if (IS_ERR(rs->spiclk)) { rs 619 drivers/spi/spi-rockchip.c ret = PTR_ERR(rs->spiclk); rs 623 drivers/spi/spi-rockchip.c ret = clk_prepare_enable(rs->apb_pclk); rs 629 drivers/spi/spi-rockchip.c ret = clk_prepare_enable(rs->spiclk); rs 635 drivers/spi/spi-rockchip.c spi_enable_chip(rs, false); rs 646 drivers/spi/spi-rockchip.c rs->dev = &pdev->dev; rs 647 drivers/spi/spi-rockchip.c rs->freq = clk_get_rate(rs->spiclk); rs 652 drivers/spi/spi-rockchip.c u32 rsd = DIV_ROUND_CLOSEST(rsd_nsecs * (rs->freq >> 8), rs 655 drivers/spi/spi-rockchip.c dev_warn(rs->dev, "%u Hz are too slow to express %u ns delay\n", rs 656 drivers/spi/spi-rockchip.c rs->freq, rsd_nsecs); rs 659 drivers/spi/spi-rockchip.c dev_warn(rs->dev, "%u Hz are too fast to express %u ns delay, clamping at %u ns\n", rs 660 drivers/spi/spi-rockchip.c rs->freq, rsd_nsecs, rs 661 drivers/spi/spi-rockchip.c CR0_RSD_MAX * 1000000000U / rs->freq); rs 663 drivers/spi/spi-rockchip.c rs->rsd = rsd; rs 666 drivers/spi/spi-rockchip.c rs->fifo_len = get_fifo_len(rs); rs 667 drivers/spi/spi-rockchip.c if (!rs->fifo_len) { rs 682 drivers/spi/spi-rockchip.c master->min_speed_hz = rs->freq / BAUDR_SCKDV_MAX; rs 683 drivers/spi/spi-rockchip.c master->max_speed_hz = min(rs->freq / BAUDR_SCKDV_MIN, MAX_SCLK_OUT); rs 691 drivers/spi/spi-rockchip.c master->dma_tx = dma_request_chan(rs->dev, "tx"); rs 698 drivers/spi/spi-rockchip.c dev_warn(rs->dev, "Failed to request TX DMA channel\n"); rs 702 drivers/spi/spi-rockchip.c master->dma_rx = dma_request_chan(rs->dev, "rx"); rs 708 drivers/spi/spi-rockchip.c dev_warn(rs->dev, "Failed to request RX DMA channel\n"); rs 713 drivers/spi/spi-rockchip.c rs->dma_addr_tx = mem->start + ROCKCHIP_SPI_TXDR; rs 714 drivers/spi/spi-rockchip.c rs->dma_addr_rx = mem->start + ROCKCHIP_SPI_RXDR; rs 735 drivers/spi/spi-rockchip.c clk_disable_unprepare(rs->spiclk); rs 737 drivers/spi/spi-rockchip.c clk_disable_unprepare(rs->apb_pclk); rs 747 drivers/spi/spi-rockchip.c struct rockchip_spi *rs = spi_master_get_devdata(master); rs 751 drivers/spi/spi-rockchip.c clk_disable_unprepare(rs->spiclk); rs 752 drivers/spi/spi-rockchip.c clk_disable_unprepare(rs->apb_pclk); rs 791 drivers/spi/spi-rockchip.c struct rockchip_spi *rs = spi_master_get_devdata(master); rs 801 drivers/spi/spi-rockchip.c clk_disable_unprepare(rs->spiclk); rs 802 drivers/spi/spi-rockchip.c clk_disable_unprepare(rs->apb_pclk); rs 813 drivers/spi/spi-rockchip.c struct rockchip_spi *rs = spi_master_get_devdata(master); rs 815 drivers/spi/spi-rockchip.c clk_disable_unprepare(rs->spiclk); rs 816 drivers/spi/spi-rockchip.c clk_disable_unprepare(rs->apb_pclk); rs 825 drivers/spi/spi-rockchip.c struct rockchip_spi *rs = spi_master_get_devdata(master); rs 827 drivers/spi/spi-rockchip.c ret = clk_prepare_enable(rs->apb_pclk); rs 831 drivers/spi/spi-rockchip.c ret = clk_prepare_enable(rs->spiclk); rs 833 drivers/spi/spi-rockchip.c clk_disable_unprepare(rs->apb_pclk); rs 285 drivers/staging/comedi/drivers/ni_routes.c struct ni_route_set *rs = &valid_routes->routes[n]; rs 288 drivers/staging/comedi/drivers/ni_routes.c rs->n_src = 0; rs 289 drivers/staging/comedi/drivers/ni_routes.c while (rs->src[rs->n_src]) rs 290 drivers/staging/comedi/drivers/ni_routes.c ++rs->n_src; rs 2441 drivers/tty/cyclades.c u32 rs; rs 2447 drivers/tty/cyclades.c rs = readl(&ch_ctrl->rs_control); rs 2449 drivers/tty/cyclades.c rs |= C_RS_RTS; rs 2451 drivers/tty/cyclades.c rs &= ~C_RS_RTS; rs 2453 drivers/tty/cyclades.c rs |= C_RS_DTR; rs 2459 drivers/tty/cyclades.c rs &= ~C_RS_DTR; rs 2465 drivers/tty/cyclades.c cy_writel(&ch_ctrl->rs_control, rs); rs 3012 drivers/tty/cyclades.c u32 rs; rs 3014 drivers/tty/cyclades.c rs = readl(&ch_ctrl->rs_control); rs 3016 drivers/tty/cyclades.c rs |= C_RS_RTS | C_RS_DTR; rs 3018 drivers/tty/cyclades.c rs &= ~(C_RS_RTS | C_RS_DTR); rs 3019 drivers/tty/cyclades.c cy_writel(&ch_ctrl->rs_control, rs); rs 100 drivers/video/backlight/ili922x.c #define START_BYTE(id, rs, rw) \ rs 101 drivers/video/backlight/ili922x.c (0x70 | (((id) & 0x01) << 2) | (((rs) & 0x01) << 1) | ((rw) & 0x01)) rs 147 drivers/video/backlight/ili922x.c static int ili922x_read_status(struct spi_device *spi, u16 *rs) rs 180 drivers/video/backlight/ili922x.c *rs = (rbuf[2] << 8) + rbuf[3]; rs 212 drivers/video/fbdev/core/bitblit.c unsigned int rs = info->var.xres - rw; rs 220 drivers/video/fbdev/core/bitblit.c region.dx = info->var.xoffset + rs; rs 230 drivers/video/fbdev/core/bitblit.c region.width = rs; rs 181 drivers/video/fbdev/core/fbcon_cw.c unsigned int rs = info->var.yres - rw; rs 189 drivers/video/fbdev/core/fbcon_cw.c region.dy = info->var.yoffset + rs; rs 266 drivers/watchdog/omap_wdt.c u32 rs = pdata->read_reset_sources(); rs 267 drivers/watchdog/omap_wdt.c if (rs & (1 << OMAP_MPU_WD_RST_SRC_ID_SHIFT)) rs 241 fs/9p/vfs_super.c struct p9_rstatfs rs; rs 252 fs/9p/vfs_super.c res = p9_client_statfs(fid, &rs); rs 254 fs/9p/vfs_super.c buf->f_type = rs.type; rs 255 fs/9p/vfs_super.c buf->f_bsize = rs.bsize; rs 256 fs/9p/vfs_super.c buf->f_blocks = rs.blocks; rs 257 fs/9p/vfs_super.c buf->f_bfree = rs.bfree; rs 258 fs/9p/vfs_super.c buf->f_bavail = rs.bavail; rs 259 fs/9p/vfs_super.c buf->f_files = rs.files; rs 260 fs/9p/vfs_super.c buf->f_ffree = rs.ffree; rs 261 fs/9p/vfs_super.c buf->f_fsid.val[0] = rs.fsid & 0xFFFFFFFFUL; rs 262 fs/9p/vfs_super.c buf->f_fsid.val[1] = (rs.fsid >> 32) & 0xFFFFFFFFUL; rs 263 fs/9p/vfs_super.c buf->f_namelen = rs.namelen; rs 305 fs/ceph/inode.c struct ceph_frag_tree_split *rs = (struct ceph_frag_tree_split*)r; rs 307 fs/ceph/inode.c le32_to_cpu(rs->frag)); rs 457 fs/crypto/crypto.c static DEFINE_RATELIMIT_STATE(rs, DEFAULT_RATELIMIT_INTERVAL, rs 462 fs/crypto/crypto.c if (!__ratelimit(&rs)) rs 72 fs/dlm/rcom.c static void set_rcom_status(struct dlm_ls *ls, struct rcom_status *rs, rs 75 fs/dlm/rcom.c rs->rs_flags = cpu_to_le32(flags); rs 196 fs/dlm/rcom.c struct rcom_status *rs; rs 208 fs/dlm/rcom.c rs = (struct rcom_status *)rc_in->rc_buf; rs 210 fs/dlm/rcom.c if (!(le32_to_cpu(rs->rs_flags) & DLM_RSF_NEED_SLOTS)) { rs 353 fs/f2fs/segment.h struct f2fs_sit_entry *rs) rs 355 fs/f2fs/segment.h se->valid_blocks = GET_SIT_VBLOCKS(rs); rs 356 fs/f2fs/segment.h se->ckpt_valid_blocks = GET_SIT_VBLOCKS(rs); rs 357 fs/f2fs/segment.h memcpy(se->cur_valid_map, rs->valid_map, SIT_VBLOCK_MAP_SIZE); rs 358 fs/f2fs/segment.h memcpy(se->ckpt_valid_map, rs->valid_map, SIT_VBLOCK_MAP_SIZE); rs 360 fs/f2fs/segment.h memcpy(se->cur_valid_map_mir, rs->valid_map, SIT_VBLOCK_MAP_SIZE); rs 362 fs/f2fs/segment.h se->type = GET_SIT_TYPE(rs); rs 363 fs/f2fs/segment.h se->mtime = le64_to_cpu(rs->mtime); rs 367 fs/f2fs/segment.h struct f2fs_sit_entry *rs) rs 371 fs/f2fs/segment.h rs->vblocks = cpu_to_le16(raw_vblocks); rs 372 fs/f2fs/segment.h memcpy(rs->valid_map, se->cur_valid_map, SIT_VBLOCK_MAP_SIZE); rs 373 fs/f2fs/segment.h rs->mtime = cpu_to_le64(se->mtime); rs 381 fs/f2fs/segment.h struct f2fs_sit_entry *rs; rs 389 fs/f2fs/segment.h rs = &raw_sit->entries[i]; rs 391 fs/f2fs/segment.h __seg_info_to_raw_sit(se, rs); rs 396 fs/f2fs/segment.h struct f2fs_sit_entry *rs) rs 398 fs/f2fs/segment.h __seg_info_to_raw_sit(se, rs); rs 400 fs/f2fs/segment.h memcpy(se->ckpt_valid_map, rs->valid_map, SIT_VBLOCK_MAP_SIZE); rs 196 fs/gfs2/rgrp.c static inline int rs_cmp(u64 blk, u32 len, struct gfs2_blkreserv *rs) rs 198 fs/gfs2/rgrp.c u64 startblk = gfs2_rbm_to_block(&rs->rs_rbm); rs 200 fs/gfs2/rgrp.c if (blk >= startblk + rs->rs_free) rs 613 fs/gfs2/rgrp.c static void dump_rs(struct seq_file *seq, const struct gfs2_blkreserv *rs, rs 616 fs/gfs2/rgrp.c struct gfs2_inode *ip = container_of(rs, struct gfs2_inode, i_res); rs 620 fs/gfs2/rgrp.c (unsigned long long)gfs2_rbm_to_block(&rs->rs_rbm), rs 621 fs/gfs2/rgrp.c rs->rs_rbm.offset, rs->rs_free); rs 629 fs/gfs2/rgrp.c static void __rs_deltree(struct gfs2_blkreserv *rs) rs 633 fs/gfs2/rgrp.c if (!gfs2_rs_active(rs)) rs 636 fs/gfs2/rgrp.c rgd = rs->rs_rbm.rgd; rs 637 fs/gfs2/rgrp.c trace_gfs2_rs(rs, TRACE_RS_TREEDEL); rs 638 fs/gfs2/rgrp.c rb_erase(&rs->rs_node, &rgd->rd_rstree); rs 639 fs/gfs2/rgrp.c RB_CLEAR_NODE(&rs->rs_node); rs 641 fs/gfs2/rgrp.c if (rs->rs_free) { rs 642 fs/gfs2/rgrp.c u64 last_block = gfs2_rbm_to_block(&rs->rs_rbm) + rs 643 fs/gfs2/rgrp.c rs->rs_free - 1; rs 644 fs/gfs2/rgrp.c struct gfs2_rbm last_rbm = { .rgd = rs->rs_rbm.rgd, }; rs 648 fs/gfs2/rgrp.c BUG_ON(rs->rs_rbm.rgd->rd_reserved < rs->rs_free); rs 649 fs/gfs2/rgrp.c rs->rs_rbm.rgd->rd_reserved -= rs->rs_free; rs 654 fs/gfs2/rgrp.c rgd->rd_extfail_pt += rs->rs_free; rs 655 fs/gfs2/rgrp.c rs->rs_free = 0; rs 658 fs/gfs2/rgrp.c start = rbm_bi(&rs->rs_rbm); rs 671 fs/gfs2/rgrp.c void gfs2_rs_deltree(struct gfs2_blkreserv *rs) rs 675 fs/gfs2/rgrp.c rgd = rs->rs_rbm.rgd; rs 678 fs/gfs2/rgrp.c __rs_deltree(rs); rs 679 fs/gfs2/rgrp.c BUG_ON(rs->rs_free); rs 710 fs/gfs2/rgrp.c struct gfs2_blkreserv *rs; rs 714 fs/gfs2/rgrp.c rs = rb_entry(n, struct gfs2_blkreserv, rs_node); rs 715 fs/gfs2/rgrp.c __rs_deltree(rs); rs 1493 fs/gfs2/rgrp.c struct gfs2_blkreserv *rs = &ip->i_res; rs 1494 fs/gfs2/rgrp.c struct gfs2_rgrpd *rgd = rs->rs_rbm.rgd; rs 1495 fs/gfs2/rgrp.c u64 fsblock = gfs2_rbm_to_block(&rs->rs_rbm); rs 1497 fs/gfs2/rgrp.c BUG_ON(gfs2_rs_active(rs)); rs 1506 fs/gfs2/rgrp.c rc = rs_cmp(fsblock, rs->rs_free, cur); rs 1518 fs/gfs2/rgrp.c rb_link_node(&rs->rs_node, parent, newn); rs 1519 fs/gfs2/rgrp.c rb_insert_color(&rs->rs_node, &rgd->rd_rstree); rs 1522 fs/gfs2/rgrp.c rgd->rd_reserved += rs->rs_free; /* blocks reserved */ rs 1524 fs/gfs2/rgrp.c trace_gfs2_rs(rs, TRACE_RS_INSERT); rs 1539 fs/gfs2/rgrp.c static inline u32 rgd_free(struct gfs2_rgrpd *rgd, struct gfs2_blkreserv *rs) rs 1543 fs/gfs2/rgrp.c if (WARN_ON_ONCE(rgd->rd_reserved < rs->rs_free)) rs 1545 fs/gfs2/rgrp.c tot_reserved = rgd->rd_reserved - rs->rs_free; rs 1568 fs/gfs2/rgrp.c struct gfs2_blkreserv *rs = &ip->i_res; rs 1570 fs/gfs2/rgrp.c u32 free_blocks = rgd_free(rgd, rs); rs 1594 fs/gfs2/rgrp.c rs->rs_rbm = rbm; rs 1595 fs/gfs2/rgrp.c rs->rs_free = extlen; rs 1620 fs/gfs2/rgrp.c struct gfs2_blkreserv *rs; rs 1627 fs/gfs2/rgrp.c rs = rb_entry(n, struct gfs2_blkreserv, rs_node); rs 1628 fs/gfs2/rgrp.c rc = rs_cmp(block, length, rs); rs 1638 fs/gfs2/rgrp.c while ((rs_cmp(block, length, rs) == 0) && (&ip->i_res != rs)) { rs 1639 fs/gfs2/rgrp.c block = gfs2_rbm_to_block(&rs->rs_rbm) + rs->rs_free; rs 1643 fs/gfs2/rgrp.c rs = rb_entry(n, struct gfs2_blkreserv, rs_node); rs 1970 fs/gfs2/rgrp.c static bool gfs2_rgrp_used_recently(const struct gfs2_blkreserv *rs, rs 1976 fs/gfs2/rgrp.c rs->rs_rbm.rgd->rd_gl->gl_dstamp)); rs 2043 fs/gfs2/rgrp.c struct gfs2_blkreserv *rs = &ip->i_res; rs 2053 fs/gfs2/rgrp.c if (gfs2_rs_active(rs)) { rs 2054 fs/gfs2/rgrp.c begin = rs->rs_rbm.rgd; rs 2055 fs/gfs2/rgrp.c } else if (rs->rs_rbm.rgd && rs 2056 fs/gfs2/rgrp.c rgrp_contains_block(rs->rs_rbm.rgd, ip->i_goal)) { rs 2057 fs/gfs2/rgrp.c begin = rs->rs_rbm.rgd; rs 2060 fs/gfs2/rgrp.c rs->rs_rbm.rgd = begin = gfs2_blk2rgrpd(sdp, ip->i_goal, 1); rs 2064 fs/gfs2/rgrp.c if (rs->rs_rbm.rgd == NULL) rs 2070 fs/gfs2/rgrp.c if (!gfs2_glock_is_locked_by_me(rs->rs_rbm.rgd->rd_gl)) { rs 2074 fs/gfs2/rgrp.c if (!gfs2_rs_active(rs)) { rs 2076 fs/gfs2/rgrp.c !fast_to_acquire(rs->rs_rbm.rgd)) rs 2079 fs/gfs2/rgrp.c gfs2_rgrp_used_recently(rs, 1000) && rs 2080 fs/gfs2/rgrp.c gfs2_rgrp_congested(rs->rs_rbm.rgd, loops)) rs 2083 fs/gfs2/rgrp.c error = gfs2_glock_nq_init(rs->rs_rbm.rgd->rd_gl, rs 2088 fs/gfs2/rgrp.c if (!gfs2_rs_active(rs) && (loops < 2) && rs 2089 fs/gfs2/rgrp.c gfs2_rgrp_congested(rs->rs_rbm.rgd, loops)) rs 2092 fs/gfs2/rgrp.c error = update_rgrp_lvb(rs->rs_rbm.rgd); rs 2101 fs/gfs2/rgrp.c if ((rs->rs_rbm.rgd->rd_flags & (GFS2_RGF_NOALLOC | rs 2103 fs/gfs2/rgrp.c (loops == 0 && ap->target > rs->rs_rbm.rgd->rd_extfail_pt)) rs 2107 fs/gfs2/rgrp.c gfs2_rgrp_bh_get(rs->rs_rbm.rgd); rs 2110 fs/gfs2/rgrp.c if (!gfs2_rs_active(rs)) rs 2111 fs/gfs2/rgrp.c rg_mblk_search(rs->rs_rbm.rgd, ip, ap); rs 2114 fs/gfs2/rgrp.c if (!gfs2_rs_active(rs) && (loops < 1)) rs 2118 fs/gfs2/rgrp.c free_blocks = rgd_free(rs->rs_rbm.rgd, rs); rs 2127 fs/gfs2/rgrp.c if (rs->rs_rbm.rgd->rd_flags & GFS2_RDF_CHECK) rs 2128 fs/gfs2/rgrp.c try_rgrp_unlink(rs->rs_rbm.rgd, &last_unlinked, rs 2132 fs/gfs2/rgrp.c if (gfs2_rs_active(rs)) rs 2133 fs/gfs2/rgrp.c gfs2_rs_deltree(rs); rs 2140 fs/gfs2/rgrp.c if (gfs2_select_rgrp(&rs->rs_rbm.rgd, begin)) rs 2312 fs/gfs2/rgrp.c struct gfs2_blkreserv *rs = &ip->i_res; rs 2319 fs/gfs2/rgrp.c if (gfs2_rs_active(rs)) { rs 2320 fs/gfs2/rgrp.c if (gfs2_rbm_eq(&rs->rs_rbm, rbm)) { rs 2322 fs/gfs2/rgrp.c ret = gfs2_rbm_from_block(&rs->rs_rbm, block + len); rs 2323 fs/gfs2/rgrp.c rlen = min(rs->rs_free, len); rs 2324 fs/gfs2/rgrp.c rs->rs_free -= rlen; rs 2326 fs/gfs2/rgrp.c trace_gfs2_rs(rs, TRACE_RS_CLAIM); rs 2327 fs/gfs2/rgrp.c if (rs->rs_free && !ret) rs 2333 fs/gfs2/rgrp.c __rs_deltree(rs); rs 49 fs/gfs2/rgrp.h extern void gfs2_rs_deltree(struct gfs2_blkreserv *rs); rs 80 fs/gfs2/rgrp.h static inline bool gfs2_rs_active(const struct gfs2_blkreserv *rs) rs 82 fs/gfs2/rgrp.h return rs && !RB_EMPTY_NODE(&rs->rs_node); rs 589 fs/gfs2/trace_gfs2.h TP_PROTO(const struct gfs2_blkreserv *rs, u8 func), rs 591 fs/gfs2/trace_gfs2.h TP_ARGS(rs, func), rs 605 fs/gfs2/trace_gfs2.h __entry->dev = rs->rs_rbm.rgd->rd_sbd->sd_vfs->s_dev; rs 606 fs/gfs2/trace_gfs2.h __entry->rd_addr = rs->rs_rbm.rgd->rd_addr; rs 607 fs/gfs2/trace_gfs2.h __entry->rd_free_clone = rs->rs_rbm.rgd->rd_free_clone; rs 608 fs/gfs2/trace_gfs2.h __entry->rd_reserved = rs->rs_rbm.rgd->rd_reserved; rs 609 fs/gfs2/trace_gfs2.h __entry->inum = container_of(rs, struct gfs2_inode, rs 611 fs/gfs2/trace_gfs2.h __entry->start = gfs2_rbm_to_block(&rs->rs_rbm); rs 612 fs/gfs2/trace_gfs2.h __entry->free = rs->rs_free; rs 54 fs/isofs/rock.c struct inode *inode, struct rock_state *rs) rs 56 fs/isofs/rock.c rs->len = sizeof(struct iso_directory_record) + de->name_len[0]; rs 57 fs/isofs/rock.c if (rs->len & 1) rs 58 fs/isofs/rock.c (rs->len)++; rs 59 fs/isofs/rock.c rs->chr = (unsigned char *)de + rs->len; rs 60 fs/isofs/rock.c rs->len = *((unsigned char *)de) - rs->len; rs 61 fs/isofs/rock.c if (rs->len < 0) rs 62 fs/isofs/rock.c rs->len = 0; rs 65 fs/isofs/rock.c rs->len -= ISOFS_SB(inode->i_sb)->s_rock_offset; rs 66 fs/isofs/rock.c rs->chr += ISOFS_SB(inode->i_sb)->s_rock_offset; rs 67 fs/isofs/rock.c if (rs->len < 0) rs 68 fs/isofs/rock.c rs->len = 0; rs 72 fs/isofs/rock.c static void init_rock_state(struct rock_state *rs, struct inode *inode) rs 74 fs/isofs/rock.c memset(rs, 0, sizeof(*rs)); rs 75 fs/isofs/rock.c rs->inode = inode; rs 85 fs/isofs/rock.c static int rock_continue(struct rock_state *rs) rs 88 fs/isofs/rock.c int blocksize = 1 << rs->inode->i_blkbits; rs 91 fs/isofs/rock.c kfree(rs->buffer); rs 92 fs/isofs/rock.c rs->buffer = NULL; rs 94 fs/isofs/rock.c if ((unsigned)rs->cont_offset > blocksize - min_de_size || rs 95 fs/isofs/rock.c (unsigned)rs->cont_size > blocksize || rs 96 fs/isofs/rock.c (unsigned)(rs->cont_offset + rs->cont_size) > blocksize) { rs 99 fs/isofs/rock.c rs->cont_extent, rs->cont_offset, rs->cont_size); rs 104 fs/isofs/rock.c if (rs->cont_extent) { rs 107 fs/isofs/rock.c rs->buffer = kmalloc(rs->cont_size, GFP_KERNEL); rs 108 fs/isofs/rock.c if (!rs->buffer) { rs 113 fs/isofs/rock.c if (++rs->cont_loops >= RR_MAX_CE_ENTRIES) rs 115 fs/isofs/rock.c bh = sb_bread(rs->inode->i_sb, rs->cont_extent); rs 117 fs/isofs/rock.c memcpy(rs->buffer, bh->b_data + rs->cont_offset, rs 118 fs/isofs/rock.c rs->cont_size); rs 120 fs/isofs/rock.c rs->chr = rs->buffer; rs 121 fs/isofs/rock.c rs->len = rs->cont_size; rs 122 fs/isofs/rock.c rs->cont_extent = 0; rs 123 fs/isofs/rock.c rs->cont_size = 0; rs 124 fs/isofs/rock.c rs->cont_offset = 0; rs 130 fs/isofs/rock.c kfree(rs->buffer); rs 131 fs/isofs/rock.c rs->buffer = NULL; rs 139 fs/isofs/rock.c static int rock_check_overflow(struct rock_state *rs, int sig) rs 185 fs/isofs/rock.c if (len > rs->len) { rs 189 fs/isofs/rock.c sig, len, rs->len); rs 201 fs/isofs/rock.c struct rock_state rs; rs 214 fs/isofs/rock.c init_rock_state(&rs, inode); rs 215 fs/isofs/rock.c setup_rock_ridge(de, inode, &rs); rs 218 fs/isofs/rock.c while (rs.len > 2) { /* There may be one byte for padding somewhere */ rs 219 fs/isofs/rock.c rr = (struct rock_ridge *)rs.chr; rs 227 fs/isofs/rock.c sig = isonum_721(rs.chr); rs 228 fs/isofs/rock.c if (rock_check_overflow(&rs, sig)) rs 230 fs/isofs/rock.c rs.chr += rr->len; rs 231 fs/isofs/rock.c rs.len -= rr->len; rs 236 fs/isofs/rock.c if (rs.len < 0) rs 249 fs/isofs/rock.c rs.cont_extent = isonum_733(rr->u.CE.extent); rs 250 fs/isofs/rock.c rs.cont_offset = isonum_733(rr->u.CE.offset); rs 251 fs/isofs/rock.c rs.cont_size = isonum_733(rr->u.CE.size); rs 286 fs/isofs/rock.c kfree(rs.buffer); rs 292 fs/isofs/rock.c ret = rock_continue(&rs); rs 298 fs/isofs/rock.c kfree(rs.buffer); rs 318 fs/isofs/rock.c struct rock_state rs; rs 324 fs/isofs/rock.c init_rock_state(&rs, inode); rs 325 fs/isofs/rock.c setup_rock_ridge(de, inode, &rs); rs 327 fs/isofs/rock.c rs.chr += 14; rs 328 fs/isofs/rock.c rs.len -= 14; rs 329 fs/isofs/rock.c if (rs.len < 0) rs 330 fs/isofs/rock.c rs.len = 0; rs 334 fs/isofs/rock.c while (rs.len > 2) { /* There may be one byte for padding somewhere */ rs 335 fs/isofs/rock.c rr = (struct rock_ridge *)rs.chr; rs 343 fs/isofs/rock.c sig = isonum_721(rs.chr); rs 344 fs/isofs/rock.c if (rock_check_overflow(&rs, sig)) rs 346 fs/isofs/rock.c rs.chr += rr->len; rs 347 fs/isofs/rock.c rs.len -= rr->len; rs 352 fs/isofs/rock.c if (rs.len < 0) rs 368 fs/isofs/rock.c rs.cont_extent = isonum_733(rr->u.CE.extent); rs 369 fs/isofs/rock.c rs.cont_offset = isonum_733(rr->u.CE.offset); rs 370 fs/isofs/rock.c rs.cont_size = isonum_733(rr->u.CE.size); rs 590 fs/isofs/rock.c ret = rock_continue(&rs); rs 596 fs/isofs/rock.c kfree(rs.buffer); rs 707 fs/isofs/rock.c struct rock_state rs; rs 713 fs/isofs/rock.c init_rock_state(&rs, inode); rs 735 fs/isofs/rock.c setup_rock_ridge(raw_de, inode, &rs); rs 738 fs/isofs/rock.c while (rs.len > 2) { /* There may be one byte for padding somewhere */ rs 739 fs/isofs/rock.c rr = (struct rock_ridge *)rs.chr; rs 742 fs/isofs/rock.c sig = isonum_721(rs.chr); rs 743 fs/isofs/rock.c if (rock_check_overflow(&rs, sig)) rs 745 fs/isofs/rock.c rs.chr += rr->len; rs 746 fs/isofs/rock.c rs.len -= rr->len; rs 747 fs/isofs/rock.c if (rs.len < 0) rs 767 fs/isofs/rock.c rs.cont_extent = isonum_733(rr->u.CE.extent); rs 768 fs/isofs/rock.c rs.cont_offset = isonum_733(rr->u.CE.offset); rs 769 fs/isofs/rock.c rs.cont_size = isonum_733(rr->u.CE.size); rs 774 fs/isofs/rock.c ret = rock_continue(&rs); rs 790 fs/isofs/rock.c kfree(rs.buffer); rs 88 fs/jffs2/compr_rubin.c static void init_rubin(struct rubin_state *rs, int div, int *bits) rs 92 fs/jffs2/compr_rubin.c rs->q = 0; rs 93 fs/jffs2/compr_rubin.c rs->p = (long) (2 * UPPER_BIT_RUBIN); rs 94 fs/jffs2/compr_rubin.c rs->bit_number = (long) 0; rs 95 fs/jffs2/compr_rubin.c rs->bit_divider = div; rs 98 fs/jffs2/compr_rubin.c rs->bits[c] = bits[c]; rs 102 fs/jffs2/compr_rubin.c static int encode(struct rubin_state *rs, long A, long B, int symbol) rs 108 fs/jffs2/compr_rubin.c while ((rs->q >= UPPER_BIT_RUBIN) || rs 109 fs/jffs2/compr_rubin.c ((rs->p + rs->q) <= UPPER_BIT_RUBIN)) { rs 110 fs/jffs2/compr_rubin.c rs->bit_number++; rs 112 fs/jffs2/compr_rubin.c ret = pushbit(&rs->pp, (rs->q & UPPER_BIT_RUBIN) ? 1 : 0, 0); rs 115 fs/jffs2/compr_rubin.c rs->q &= LOWER_BITS_RUBIN; rs 116 fs/jffs2/compr_rubin.c rs->q <<= 1; rs 117 fs/jffs2/compr_rubin.c rs->p <<= 1; rs 119 fs/jffs2/compr_rubin.c i0 = A * rs->p / (A + B); rs 123 fs/jffs2/compr_rubin.c if (i0 >= rs->p) rs 124 fs/jffs2/compr_rubin.c i0 = rs->p - 1; rs 126 fs/jffs2/compr_rubin.c i1 = rs->p - i0; rs 129 fs/jffs2/compr_rubin.c rs->p = i0; rs 131 fs/jffs2/compr_rubin.c rs->p = i1; rs 132 fs/jffs2/compr_rubin.c rs->q += i0; rs 138 fs/jffs2/compr_rubin.c static void end_rubin(struct rubin_state *rs) rs 144 fs/jffs2/compr_rubin.c pushbit(&rs->pp, (UPPER_BIT_RUBIN & rs->q) ? 1 : 0, 1); rs 145 fs/jffs2/compr_rubin.c rs->q &= LOWER_BITS_RUBIN; rs 146 fs/jffs2/compr_rubin.c rs->q <<= 1; rs 151 fs/jffs2/compr_rubin.c static void init_decode(struct rubin_state *rs, int div, int *bits) rs 153 fs/jffs2/compr_rubin.c init_rubin(rs, div, bits); rs 156 fs/jffs2/compr_rubin.c rs->rec_q = 0; rs 158 fs/jffs2/compr_rubin.c for (rs->bit_number = 0; rs->bit_number++ < RUBIN_REG_SIZE; rs 159 fs/jffs2/compr_rubin.c rs->rec_q = rs->rec_q * 2 + (long) (pullbit(&rs->pp))) rs 163 fs/jffs2/compr_rubin.c static void __do_decode(struct rubin_state *rs, unsigned long p, rs 182 fs/jffs2/compr_rubin.c rs->p = p; rs 183 fs/jffs2/compr_rubin.c rs->q = q; rs 185 fs/jffs2/compr_rubin.c rs->bit_number += bits; rs 190 fs/jffs2/compr_rubin.c rec_q = rs->rec_q; rs 192 fs/jffs2/compr_rubin.c c = pullbit(&rs->pp); rs 197 fs/jffs2/compr_rubin.c rs->rec_q = rec_q; rs 200 fs/jffs2/compr_rubin.c static int decode(struct rubin_state *rs, long A, long B) rs 202 fs/jffs2/compr_rubin.c unsigned long p = rs->p, q = rs->q; rs 207 fs/jffs2/compr_rubin.c __do_decode(rs, p, q); rs 209 fs/jffs2/compr_rubin.c i0 = A * rs->p / (A + B); rs 213 fs/jffs2/compr_rubin.c if (i0 >= rs->p) rs 214 fs/jffs2/compr_rubin.c i0 = rs->p - 1; rs 216 fs/jffs2/compr_rubin.c threshold = rs->q + i0; rs 217 fs/jffs2/compr_rubin.c symbol = rs->rec_q >= threshold; rs 218 fs/jffs2/compr_rubin.c if (rs->rec_q >= threshold) { rs 219 fs/jffs2/compr_rubin.c rs->q += i0; rs 220 fs/jffs2/compr_rubin.c i0 = rs->p - i0; rs 223 fs/jffs2/compr_rubin.c rs->p = i0; rs 230 fs/jffs2/compr_rubin.c static int out_byte(struct rubin_state *rs, unsigned char byte) rs 234 fs/jffs2/compr_rubin.c rs_copy = *rs; rs 237 fs/jffs2/compr_rubin.c ret = encode(rs, rs->bit_divider-rs->bits[i], rs 238 fs/jffs2/compr_rubin.c rs->bits[i], byte & 1); rs 241 fs/jffs2/compr_rubin.c *rs = rs_copy; rs 249 fs/jffs2/compr_rubin.c static int in_byte(struct rubin_state *rs) rs 251 fs/jffs2/compr_rubin.c int i, result = 0, bit_divider = rs->bit_divider; rs 254 fs/jffs2/compr_rubin.c result |= decode(rs, bit_divider - rs->bits[i], rs 255 fs/jffs2/compr_rubin.c rs->bits[i]) << i; rs 268 fs/jffs2/compr_rubin.c struct rubin_state rs; rs 270 fs/jffs2/compr_rubin.c init_pushpull(&rs.pp, cpage_out, *dstlen * 8, 0, 32); rs 272 fs/jffs2/compr_rubin.c init_rubin(&rs, bit_divider, bits); rs 274 fs/jffs2/compr_rubin.c while (pos < (*sourcelen) && !out_byte(&rs, data_in[pos])) rs 277 fs/jffs2/compr_rubin.c end_rubin(&rs); rs 287 fs/jffs2/compr_rubin.c outpos = (pushedbits(&rs.pp)+7)/8; rs 374 fs/jffs2/compr_rubin.c struct rubin_state rs; rs 376 fs/jffs2/compr_rubin.c init_pushpull(&rs.pp, cdata_in, srclen, 0, 0); rs 377 fs/jffs2/compr_rubin.c init_decode(&rs, bit_divider, bits); rs 380 fs/jffs2/compr_rubin.c page_out[outpos++] = in_byte(&rs); rs 421 fs/reiserfs/bitmap.c struct reiserfs_super_block *rs; rs 428 fs/reiserfs/bitmap.c rs = SB_DISK_SUPER_BLOCK(s); rs 457 fs/reiserfs/bitmap.c set_sb_free_blocks(rs, sb_free_blocks(rs) + 1); rs 2717 fs/reiserfs/journal.c struct reiserfs_super_block *rs; rs 2772 fs/reiserfs/journal.c rs = SB_DISK_SUPER_BLOCK(sb); rs 2786 fs/reiserfs/journal.c if (is_reiserfs_jr(rs) rs 2788 fs/reiserfs/journal.c sb_jp_journal_magic(rs))) { rs 2794 fs/reiserfs/journal.c sb_jp_journal_magic(rs)); rs 11 fs/reiserfs/objectid.c #define objectid_map(s,rs) (old_format_only (s) ? \ rs 12 fs/reiserfs/objectid.c (__le32 *)((struct reiserfs_super_block_v1 *)(rs) + 1) :\ rs 13 fs/reiserfs/objectid.c (__le32 *)((rs) + 1)) rs 52 fs/reiserfs/objectid.c struct reiserfs_super_block *rs = SB_DISK_SUPER_BLOCK(s); rs 53 fs/reiserfs/objectid.c __le32 *map = objectid_map(s, rs); rs 86 fs/reiserfs/objectid.c if (sb_oid_cursize(rs) > 2 && map[1] == map[2]) { rs 88 fs/reiserfs/objectid.c (sb_oid_cursize(rs) - 3) * sizeof(__u32)); rs 89 fs/reiserfs/objectid.c set_sb_oid_cursize(rs, sb_oid_cursize(rs) - 2); rs 101 fs/reiserfs/objectid.c struct reiserfs_super_block *rs = SB_DISK_SUPER_BLOCK(s); rs 102 fs/reiserfs/objectid.c __le32 *map = objectid_map(s, rs); rs 119 fs/reiserfs/objectid.c while (i < sb_oid_cursize(rs)) { rs 131 fs/reiserfs/objectid.c (sb_oid_cursize(rs) - i - rs 133 fs/reiserfs/objectid.c set_sb_oid_cursize(rs, sb_oid_cursize(rs) - 2); rs 135 fs/reiserfs/objectid.c RFALSE(sb_oid_cursize(rs) < 2 || rs 136 fs/reiserfs/objectid.c sb_oid_cursize(rs) > sb_oid_maxsize(rs), rs 138 fs/reiserfs/objectid.c sb_oid_cursize(rs), sb_oid_maxsize(rs)); rs 159 fs/reiserfs/objectid.c if (sb_oid_cursize(rs) == sb_oid_maxsize(rs)) { rs 166 fs/reiserfs/objectid.c (sb_oid_cursize(rs) - i - 1) * sizeof(__u32)); rs 169 fs/reiserfs/objectid.c set_sb_oid_cursize(rs, sb_oid_cursize(rs) + 2); rs 548 fs/reiserfs/prints.c struct reiserfs_super_block *rs = rs 553 fs/reiserfs/prints.c if (is_reiserfs_3_5(rs)) { rs 555 fs/reiserfs/prints.c } else if (is_reiserfs_3_6(rs)) { rs 557 fs/reiserfs/prints.c } else if (is_reiserfs_jr(rs)) { rs 558 fs/reiserfs/prints.c version = ((sb_version(rs) == REISERFS_VERSION_2) ? rs 567 fs/reiserfs/prints.c printk("Block count %u\n", sb_block_count(rs)); rs 568 fs/reiserfs/prints.c printk("Blocksize %d\n", sb_blocksize(rs)); rs 569 fs/reiserfs/prints.c printk("Free blocks %u\n", sb_free_blocks(rs)); rs 576 fs/reiserfs/prints.c data_blocks = sb_block_count(rs) - skipped - 1 - sb_bmap_nr(rs) - rs 577 fs/reiserfs/prints.c (!is_reiserfs_jr(rs) ? sb_jp_journal_size(rs) + rs 578 fs/reiserfs/prints.c 1 : sb_reserved_for_journal(rs)) - sb_free_blocks(rs); rs 581 fs/reiserfs/prints.c "1 super block, %d data blocks\n", skipped, sb_bmap_nr(rs), rs 582 fs/reiserfs/prints.c (!is_reiserfs_jr(rs) ? (sb_jp_journal_size(rs) + 1) : rs 583 fs/reiserfs/prints.c sb_reserved_for_journal(rs)), data_blocks); rs 584 fs/reiserfs/prints.c printk("Root block %u\n", sb_root_block(rs)); rs 585 fs/reiserfs/prints.c printk("Journal block (first) %d\n", sb_jp_journal_1st_block(rs)); rs 586 fs/reiserfs/prints.c printk("Journal dev %d\n", sb_jp_journal_dev(rs)); rs 587 fs/reiserfs/prints.c printk("Journal orig size %d\n", sb_jp_journal_size(rs)); rs 588 fs/reiserfs/prints.c printk("FS state %d\n", sb_fs_state(rs)); rs 590 fs/reiserfs/prints.c reiserfs_hashname(sb_hash_function_code(rs))); rs 592 fs/reiserfs/prints.c printk("Tree height %d\n", sb_tree_height(rs)); rs 57 fs/reiserfs/procfs.c #define DF( x ) D2C( rs -> s_v1.x ) rs 58 fs/reiserfs/procfs.c #define DFL( x ) D4C( rs -> s_v1.x ) rs 60 fs/reiserfs/procfs.c #define objectid_map( s, rs ) (old_format_only (s) ? \ rs 61 fs/reiserfs/procfs.c (__le32 *)((struct reiserfs_super_block_v1 *)rs + 1) : \ rs 62 fs/reiserfs/procfs.c (__le32 *)(rs + 1)) rs 63 fs/reiserfs/procfs.c #define MAP( i ) D4C( objectid_map( sb, rs )[ i ] ) rs 65 fs/reiserfs/procfs.c #define DJF( x ) le32_to_cpu( rs -> x ) rs 229 fs/reiserfs/procfs.c struct reiserfs_super_block *rs = sb_info->s_rs; rs 255 fs/reiserfs/procfs.c rs->s_v1.s_magic, rs 273 fs/reiserfs/procfs.c struct reiserfs_super_block *rs = sb_info->s_rs; rs 274 fs/reiserfs/procfs.c unsigned int mapsize = le16_to_cpu(rs->s_v1.s_oid_cursize); rs 296 fs/reiserfs/procfs.c mapsize, le16_to_cpu(rs->s_v1.s_oid_maxsize), total_used); rs 311 fs/reiserfs/procfs.c struct reiserfs_super_block *rs = r->s_rs; rs 312 fs/reiserfs/procfs.c struct journal_params *jp = &rs->s_v1.s_journal; rs 1104 fs/reiserfs/reiserfs.h int is_reiserfs_3_5(struct reiserfs_super_block *rs); rs 1105 fs/reiserfs/reiserfs.h int is_reiserfs_3_6(struct reiserfs_super_block *rs); rs 1106 fs/reiserfs/reiserfs.h int is_reiserfs_jr(struct reiserfs_super_block *rs); rs 40 fs/reiserfs/super.c int is_reiserfs_3_5(struct reiserfs_super_block *rs) rs 42 fs/reiserfs/super.c return !strncmp(rs->s_v1.s_magic, reiserfs_3_5_magic_string, rs 46 fs/reiserfs/super.c int is_reiserfs_3_6(struct reiserfs_super_block *rs) rs 48 fs/reiserfs/super.c return !strncmp(rs->s_v1.s_magic, reiserfs_3_6_magic_string, rs 52 fs/reiserfs/super.c int is_reiserfs_jr(struct reiserfs_super_block *rs) rs 54 fs/reiserfs/super.c return !strncmp(rs->s_v1.s_magic, reiserfs_jr_magic_string, rs 58 fs/reiserfs/super.c static int is_any_reiserfs_magic_string(struct reiserfs_super_block *rs) rs 60 fs/reiserfs/super.c return (is_reiserfs_3_5(rs) || is_reiserfs_3_6(rs) || rs 61 fs/reiserfs/super.c is_reiserfs_jr(rs)); rs 1397 fs/reiserfs/super.c struct reiserfs_super_block *rs = SB_DISK_SUPER_BLOCK(s); rs 1406 fs/reiserfs/super.c if (!(le32_to_cpu(rs->s_flags) & reiserfs_attrs_cleared)) { rs 1433 fs/reiserfs/super.c struct reiserfs_super_block *rs; rs 1459 fs/reiserfs/super.c rs = SB_DISK_SUPER_BLOCK(s); rs 1532 fs/reiserfs/super.c if (sb_umount_state(rs) == REISERFS_VALID_FS rs 1545 fs/reiserfs/super.c set_sb_umount_state(rs, REISERFS_SB(s)->s_mount_state); rs 1562 fs/reiserfs/super.c REISERFS_SB(s)->s_mount_state = sb_umount_state(rs); rs 1572 fs/reiserfs/super.c REISERFS_SB(s)->s_mount_state = sb_umount_state(rs); rs 1574 fs/reiserfs/super.c set_sb_umount_state(rs, REISERFS_ERROR_FS); rs 1576 fs/reiserfs/super.c set_sb_mnt_count(rs, sb_mnt_count(rs) + 1); rs 1609 fs/reiserfs/super.c struct reiserfs_super_block *rs; rs 1621 fs/reiserfs/super.c rs = (struct reiserfs_super_block *)bh->b_data; rs 1622 fs/reiserfs/super.c if (!is_any_reiserfs_magic_string(rs)) { rs 1629 fs/reiserfs/super.c fs_blocksize = sb_blocksize(rs); rs 1642 fs/reiserfs/super.c rs = (struct reiserfs_super_block *)bh->b_data; rs 1643 fs/reiserfs/super.c if (sb_blocksize(rs) != s->s_blocksize) { rs 1653 fs/reiserfs/super.c if (rs->s_v1.s_root_block == cpu_to_le32(-1)) { rs 1664 fs/reiserfs/super.c SB_DISK_SUPER_BLOCK(s) = rs; rs 1670 fs/reiserfs/super.c if (is_reiserfs_jr(rs)) { rs 1671 fs/reiserfs/super.c if (sb_version(rs) == REISERFS_VERSION_2) rs 1674 fs/reiserfs/super.c else if (sb_version(rs) == REISERFS_VERSION_1) rs 1680 fs/reiserfs/super.c "non-standard magic", sb_version(rs)); rs 1690 fs/reiserfs/super.c is_reiserfs_3_5(rs) ? "3.5" : "3.6"); rs 1905 fs/reiserfs/super.c struct reiserfs_super_block *rs; rs 1983 fs/reiserfs/super.c rs = SB_DISK_SUPER_BLOCK(s); rs 1991 fs/reiserfs/super.c sb_block_count(rs) * sb_blocksize(rs)) { rs 2096 fs/reiserfs/super.c if (is_reiserfs_3_5(rs) rs 2097 fs/reiserfs/super.c || (is_reiserfs_jr(rs) && SB_VERSION(s) == REISERFS_VERSION_1)) rs 2114 fs/reiserfs/super.c set_sb_umount_state(rs, REISERFS_ERROR_FS); rs 2115 fs/reiserfs/super.c set_sb_fs_state(rs, 0); rs 2124 fs/reiserfs/super.c sb_bmap_nr(rs) != 0) { rs 2128 fs/reiserfs/super.c "Clearing to zero.", sb_bmap_nr(rs), rs 2131 fs/reiserfs/super.c set_sb_bmap_nr(rs, 0); rs 2145 fs/reiserfs/super.c if (is_reiserfs_3_5(rs)) rs 2151 fs/reiserfs/super.c memcpy(rs->s_v1.s_magic, rs 2156 fs/reiserfs/super.c set_sb_version(rs, REISERFS_VERSION_2); rs 2164 fs/reiserfs/super.c set_sb_mnt_count(rs, sb_mnt_count(rs) + 1); rs 2205 fs/reiserfs/super.c set_sb_hash_function_code(rs, function2code(sbi->s_hash_function)); rs 2253 fs/reiserfs/super.c struct reiserfs_super_block *rs = SB_DISK_SUPER_BLOCK(dentry->d_sb); rs 2256 fs/reiserfs/super.c buf->f_bfree = sb_free_blocks(rs); rs 2258 fs/reiserfs/super.c buf->f_blocks = sb_block_count(rs) - sb_bmap_nr(rs) - 1; rs 2262 fs/reiserfs/super.c buf->f_fsid.val[0] = (u32)crc32_le(0, rs->s_uuid, sizeof(rs->s_uuid)/2); rs 2263 fs/reiserfs/super.c buf->f_fsid.val[1] = (u32)crc32_le(0, rs->s_uuid + sizeof(rs->s_uuid)/2, rs 2264 fs/reiserfs/super.c sizeof(rs->s_uuid)/2); rs 15 fs/verity/init.c static DEFINE_RATELIMIT_STATE(rs, DEFAULT_RATELIMIT_INTERVAL, rs 20 fs/verity/init.c if (!__ratelimit(&rs)) rs 40 include/linux/ratelimit.h static inline void ratelimit_state_init(struct ratelimit_state *rs, rs 43 include/linux/ratelimit.h memset(rs, 0, sizeof(*rs)); rs 45 include/linux/ratelimit.h raw_spin_lock_init(&rs->lock); rs 46 include/linux/ratelimit.h rs->interval = interval; rs 47 include/linux/ratelimit.h rs->burst = burst; rs 50 include/linux/ratelimit.h static inline void ratelimit_default_init(struct ratelimit_state *rs) rs 52 include/linux/ratelimit.h return ratelimit_state_init(rs, DEFAULT_RATELIMIT_INTERVAL, rs 56 include/linux/ratelimit.h static inline void ratelimit_state_exit(struct ratelimit_state *rs) rs 58 include/linux/ratelimit.h if (!(rs->flags & RATELIMIT_MSG_ON_RELEASE)) rs 61 include/linux/ratelimit.h if (rs->missed) { rs 63 include/linux/ratelimit.h current->comm, rs->missed); rs 64 include/linux/ratelimit.h rs->missed = 0; rs 69 include/linux/ratelimit.h ratelimit_set_flags(struct ratelimit_state *rs, unsigned long flags) rs 71 include/linux/ratelimit.h rs->flags = flags; rs 76 include/linux/ratelimit.h extern int ___ratelimit(struct ratelimit_state *rs, const char *func); rs 62 include/linux/rslib.h int encode_rs8(struct rs_control *rs, uint8_t *data, int len, uint16_t *par, rs 66 include/linux/rslib.h int decode_rs8(struct rs_control *rs, uint8_t *data, uint16_t *par, int len, rs 73 include/linux/rslib.h int encode_rs16(struct rs_control *rs, uint16_t *data, int len, uint16_t *par, rs 77 include/linux/rslib.h int decode_rs16(struct rs_control *rs, uint16_t *data, uint16_t *par, int len, rs 108 include/linux/rslib.h void free_rs(struct rs_control *rs); rs 122 include/linux/rslib.h static inline int rs_modnn(struct rs_codec *rs, int x) rs 124 include/linux/rslib.h while (x >= rs->nn) { rs 125 include/linux/rslib.h x -= rs->nn; rs 126 include/linux/rslib.h x = (x >> rs->mm) + (x & rs->nn); rs 122 include/math-emu/double.h #define FP_FROM_INT_D(X,r,rs,rt) _FP_FROM_INT(D,2,X,r,rs,rt) rs 197 include/math-emu/double.h #define FP_FROM_INT_D(X,r,rs,rt) _FP_FROM_INT(D,1,X,r,rs,rt) rs 128 include/math-emu/quad.h #define FP_FROM_INT_Q(X,r,rs,rt) _FP_FROM_INT(Q,4,X,r,rs,rt) rs 201 include/math-emu/quad.h #define FP_FROM_INT_Q(X,r,rs,rt) _FP_FROM_INT(Q,2,X,r,rs,rt) rs 111 include/math-emu/single.h #define FP_FROM_INT_S(X,r,rs,rt) _FP_FROM_INT(S,1,X,r,rs,rt) rs 1068 include/net/tcp.h void (*cong_control)(struct sock *sk, const struct rate_sample *rs); rs 1136 include/net/tcp.h struct rate_sample *rs); rs 1138 include/net/tcp.h bool is_sack_reneg, struct rate_sample *rs); rs 2035 include/net/tcp.h extern void tcp_rack_update_reo_wnd(struct sock *sk, struct rate_sample *rs); rs 799 kernel/printk/printk.c struct ratelimit_state rs; rs 836 kernel/printk/printk.c if (!___ratelimit(&user->rs, current->comm)) rs 1027 kernel/printk/printk.c ratelimit_default_init(&user->rs); rs 1028 kernel/printk/printk.c ratelimit_set_flags(&user->rs, RATELIMIT_MSG_ON_RELEASE); rs 1048 kernel/printk/printk.c ratelimit_state_exit(&user->rs); rs 183 kernel/rcu/srcutiny.c struct rcu_synchronize rs; rs 185 kernel/rcu/srcutiny.c init_rcu_head_on_stack(&rs.head); rs 186 kernel/rcu/srcutiny.c init_completion(&rs.completion); rs 187 kernel/rcu/srcutiny.c call_srcu(ssp, &rs.head, wakeme_after_rcu); rs 188 kernel/rcu/srcutiny.c wait_for_completion(&rs.completion); rs 189 kernel/rcu/srcutiny.c destroy_rcu_head_on_stack(&rs.head); rs 65 lib/crypto/des.c static const u8 rs[256] = { rs 634 lib/crypto/des.c pe[15 * 2 + 0] = DES_PC2(a, b, c, d); d = rs[d]; rs 635 lib/crypto/des.c pe[14 * 2 + 0] = DES_PC2(d, a, b, c); c = rs[c]; b = rs[b]; rs 636 lib/crypto/des.c pe[13 * 2 + 0] = DES_PC2(b, c, d, a); a = rs[a]; d = rs[d]; rs 637 lib/crypto/des.c pe[12 * 2 + 0] = DES_PC2(d, a, b, c); c = rs[c]; b = rs[b]; rs 638 lib/crypto/des.c pe[11 * 2 + 0] = DES_PC2(b, c, d, a); a = rs[a]; d = rs[d]; rs 639 lib/crypto/des.c pe[10 * 2 + 0] = DES_PC2(d, a, b, c); c = rs[c]; b = rs[b]; rs 640 lib/crypto/des.c pe[ 9 * 2 + 0] = DES_PC2(b, c, d, a); a = rs[a]; d = rs[d]; rs 641 lib/crypto/des.c pe[ 8 * 2 + 0] = DES_PC2(d, a, b, c); c = rs[c]; rs 642 lib/crypto/des.c pe[ 7 * 2 + 0] = DES_PC2(c, d, a, b); b = rs[b]; a = rs[a]; rs 643 lib/crypto/des.c pe[ 6 * 2 + 0] = DES_PC2(a, b, c, d); d = rs[d]; c = rs[c]; rs 644 lib/crypto/des.c pe[ 5 * 2 + 0] = DES_PC2(c, d, a, b); b = rs[b]; a = rs[a]; rs 645 lib/crypto/des.c pe[ 4 * 2 + 0] = DES_PC2(a, b, c, d); d = rs[d]; c = rs[c]; rs 646 lib/crypto/des.c pe[ 3 * 2 + 0] = DES_PC2(c, d, a, b); b = rs[b]; a = rs[a]; rs 647 lib/crypto/des.c pe[ 2 * 2 + 0] = DES_PC2(a, b, c, d); d = rs[d]; c = rs[c]; rs 648 lib/crypto/des.c pe[ 1 * 2 + 0] = DES_PC2(c, d, a, b); b = rs[b]; rs 652 lib/crypto/des.c w = (a ^ c) | (b ^ d) | (rs[a] ^ c) | (b ^ rs[d]); rs 663 lib/crypto/des.c w |= (a ^ c) | (b ^ d) | (rs[a] ^ c) | (b ^ rs[d]); rs 665 lib/crypto/des.c pe[15 * 2 + 1] = DES_PC2(a, b, c, d); d = rs[d]; rs 666 lib/crypto/des.c pe[14 * 2 + 1] = DES_PC2(d, a, b, c); c = rs[c]; b = rs[b]; rs 667 lib/crypto/des.c pe[13 * 2 + 1] = DES_PC2(b, c, d, a); a = rs[a]; d = rs[d]; rs 668 lib/crypto/des.c pe[12 * 2 + 1] = DES_PC2(d, a, b, c); c = rs[c]; b = rs[b]; rs 669 lib/crypto/des.c pe[11 * 2 + 1] = DES_PC2(b, c, d, a); a = rs[a]; d = rs[d]; rs 670 lib/crypto/des.c pe[10 * 2 + 1] = DES_PC2(d, a, b, c); c = rs[c]; b = rs[b]; rs 671 lib/crypto/des.c pe[ 9 * 2 + 1] = DES_PC2(b, c, d, a); a = rs[a]; d = rs[d]; rs 672 lib/crypto/des.c pe[ 8 * 2 + 1] = DES_PC2(d, a, b, c); c = rs[c]; rs 673 lib/crypto/des.c pe[ 7 * 2 + 1] = DES_PC2(c, d, a, b); b = rs[b]; a = rs[a]; rs 674 lib/crypto/des.c pe[ 6 * 2 + 1] = DES_PC2(a, b, c, d); d = rs[d]; c = rs[c]; rs 675 lib/crypto/des.c pe[ 5 * 2 + 1] = DES_PC2(c, d, a, b); b = rs[b]; a = rs[a]; rs 676 lib/crypto/des.c pe[ 4 * 2 + 1] = DES_PC2(a, b, c, d); d = rs[d]; c = rs[c]; rs 677 lib/crypto/des.c pe[ 3 * 2 + 1] = DES_PC2(c, d, a, b); b = rs[b]; a = rs[a]; rs 678 lib/crypto/des.c pe[ 2 * 2 + 1] = DES_PC2(a, b, c, d); d = rs[d]; c = rs[c]; rs 679 lib/crypto/des.c pe[ 1 * 2 + 1] = DES_PC2(c, d, a, b); b = rs[b]; rs 725 lib/crypto/des.c pe[ 0 * 2] = DES_PC2(a, b, c, d); d = rs[d]; rs 726 lib/crypto/des.c pe[ 1 * 2] = DES_PC2(d, a, b, c); c = rs[c]; b = rs[b]; rs 727 lib/crypto/des.c pe[ 2 * 2] = DES_PC2(b, c, d, a); a = rs[a]; d = rs[d]; rs 728 lib/crypto/des.c pe[ 3 * 2] = DES_PC2(d, a, b, c); c = rs[c]; b = rs[b]; rs 729 lib/crypto/des.c pe[ 4 * 2] = DES_PC2(b, c, d, a); a = rs[a]; d = rs[d]; rs 730 lib/crypto/des.c pe[ 5 * 2] = DES_PC2(d, a, b, c); c = rs[c]; b = rs[b]; rs 731 lib/crypto/des.c pe[ 6 * 2] = DES_PC2(b, c, d, a); a = rs[a]; d = rs[d]; rs 732 lib/crypto/des.c pe[ 7 * 2] = DES_PC2(d, a, b, c); c = rs[c]; rs 733 lib/crypto/des.c pe[ 8 * 2] = DES_PC2(c, d, a, b); b = rs[b]; a = rs[a]; rs 734 lib/crypto/des.c pe[ 9 * 2] = DES_PC2(a, b, c, d); d = rs[d]; c = rs[c]; rs 735 lib/crypto/des.c pe[10 * 2] = DES_PC2(c, d, a, b); b = rs[b]; a = rs[a]; rs 736 lib/crypto/des.c pe[11 * 2] = DES_PC2(a, b, c, d); d = rs[d]; c = rs[c]; rs 737 lib/crypto/des.c pe[12 * 2] = DES_PC2(c, d, a, b); b = rs[b]; a = rs[a]; rs 738 lib/crypto/des.c pe[13 * 2] = DES_PC2(a, b, c, d); d = rs[d]; c = rs[c]; rs 739 lib/crypto/des.c pe[14 * 2] = DES_PC2(c, d, a, b); b = rs[b]; rs 750 lib/crypto/des.c pe[ 0 * 2 + 1] = DES_PC2(a, b, c, d); d = rs[d]; rs 751 lib/crypto/des.c pe[ 1 * 2 + 1] = DES_PC2(d, a, b, c); c = rs[c]; b = rs[b]; rs 752 lib/crypto/des.c pe[ 2 * 2 + 1] = DES_PC2(b, c, d, a); a = rs[a]; d = rs[d]; rs 753 lib/crypto/des.c pe[ 3 * 2 + 1] = DES_PC2(d, a, b, c); c = rs[c]; b = rs[b]; rs 754 lib/crypto/des.c pe[ 4 * 2 + 1] = DES_PC2(b, c, d, a); a = rs[a]; d = rs[d]; rs 755 lib/crypto/des.c pe[ 5 * 2 + 1] = DES_PC2(d, a, b, c); c = rs[c]; b = rs[b]; rs 756 lib/crypto/des.c pe[ 6 * 2 + 1] = DES_PC2(b, c, d, a); a = rs[a]; d = rs[d]; rs 757 lib/crypto/des.c pe[ 7 * 2 + 1] = DES_PC2(d, a, b, c); c = rs[c]; rs 758 lib/crypto/des.c pe[ 8 * 2 + 1] = DES_PC2(c, d, a, b); b = rs[b]; a = rs[a]; rs 759 lib/crypto/des.c pe[ 9 * 2 + 1] = DES_PC2(a, b, c, d); d = rs[d]; c = rs[c]; rs 760 lib/crypto/des.c pe[10 * 2 + 1] = DES_PC2(c, d, a, b); b = rs[b]; a = rs[a]; rs 761 lib/crypto/des.c pe[11 * 2 + 1] = DES_PC2(a, b, c, d); d = rs[d]; c = rs[c]; rs 762 lib/crypto/des.c pe[12 * 2 + 1] = DES_PC2(c, d, a, b); b = rs[b]; a = rs[a]; rs 763 lib/crypto/des.c pe[13 * 2 + 1] = DES_PC2(a, b, c, d); d = rs[d]; c = rs[c]; rs 764 lib/crypto/des.c pe[14 * 2 + 1] = DES_PC2(c, d, a, b); b = rs[b]; rs 27 lib/ratelimit.c int ___ratelimit(struct ratelimit_state *rs, const char *func) rs 32 lib/ratelimit.c if (!rs->interval) rs 41 lib/ratelimit.c if (!raw_spin_trylock_irqsave(&rs->lock, flags)) rs 44 lib/ratelimit.c if (!rs->begin) rs 45 lib/ratelimit.c rs->begin = jiffies; rs 47 lib/ratelimit.c if (time_is_before_jiffies(rs->begin + rs->interval)) { rs 48 lib/ratelimit.c if (rs->missed) { rs 49 lib/ratelimit.c if (!(rs->flags & RATELIMIT_MSG_ON_RELEASE)) { rs 52 lib/ratelimit.c func, rs->missed); rs 53 lib/ratelimit.c rs->missed = 0; rs 56 lib/ratelimit.c rs->begin = jiffies; rs 57 lib/ratelimit.c rs->printed = 0; rs 59 lib/ratelimit.c if (rs->burst && rs->burst > rs->printed) { rs 60 lib/ratelimit.c rs->printed++; rs 63 lib/ratelimit.c rs->missed++; rs 66 lib/ratelimit.c raw_spin_unlock_irqrestore(&rs->lock, flags); rs 13 lib/reed_solomon/decode_rs.c struct rs_codec *rs = rsc->codec; rs 16 lib/reed_solomon/decode_rs.c int nn = rs->nn; rs 17 lib/reed_solomon/decode_rs.c int nroots = rs->nroots; rs 18 lib/reed_solomon/decode_rs.c int fcr = rs->fcr; rs 19 lib/reed_solomon/decode_rs.c int prim = rs->prim; rs 20 lib/reed_solomon/decode_rs.c int iprim = rs->iprim; rs 21 lib/reed_solomon/decode_rs.c uint16_t *alpha_to = rs->alpha_to; rs 22 lib/reed_solomon/decode_rs.c uint16_t *index_of = rs->index_of; rs 26 lib/reed_solomon/decode_rs.c uint16_t msk = (uint16_t) rs->nn; rs 72 lib/reed_solomon/decode_rs.c alpha_to[rs_modnn(rs, index_of[syn[i]] + rs 84 lib/reed_solomon/decode_rs.c alpha_to[rs_modnn(rs, index_of[syn[i]] + rs 111 lib/reed_solomon/decode_rs.c lambda[1] = alpha_to[rs_modnn(rs, rs 114 lib/reed_solomon/decode_rs.c u = rs_modnn(rs, prim * (nn - 1 - (eras_pos[i] + pad))); rs 119 lib/reed_solomon/decode_rs.c alpha_to[rs_modnn(rs, u + tmp)]; rs 140 lib/reed_solomon/decode_rs.c alpha_to[rs_modnn(rs, rs 156 lib/reed_solomon/decode_rs.c alpha_to[rs_modnn(rs, discr_r + rs 169 lib/reed_solomon/decode_rs.c rs_modnn(rs, index_of[lambda[i]] rs 200 lib/reed_solomon/decode_rs.c for (i = 1, k = iprim - 1; i <= nn; i++, k = rs_modnn(rs, k + iprim)) { rs 204 lib/reed_solomon/decode_rs.c reg[j] = rs_modnn(rs, reg[j] + j); rs 242 lib/reed_solomon/decode_rs.c alpha_to[rs_modnn(rs, s[i - j] + lambda[j])]; rs 257 lib/reed_solomon/decode_rs.c num1 ^= alpha_to[rs_modnn(rs, omega[i] + rs 267 lib/reed_solomon/decode_rs.c num2 = alpha_to[rs_modnn(rs, root[j] * (fcr - 1) + nn)]; rs 274 lib/reed_solomon/decode_rs.c den ^= alpha_to[rs_modnn(rs, lambda[i + 1] + rs 279 lib/reed_solomon/decode_rs.c b[j] = alpha_to[rs_modnn(rs, index_of[num1] + rs 296 lib/reed_solomon/decode_rs.c tmp ^= alpha_to[rs_modnn(rs, index_of[b[j]] + k)]; rs 13 lib/reed_solomon/encode_rs.c struct rs_codec *rs = rsc->codec; rs 15 lib/reed_solomon/encode_rs.c int nn = rs->nn; rs 16 lib/reed_solomon/encode_rs.c int nroots = rs->nroots; rs 17 lib/reed_solomon/encode_rs.c uint16_t *alpha_to = rs->alpha_to; rs 18 lib/reed_solomon/encode_rs.c uint16_t *index_of = rs->index_of; rs 19 lib/reed_solomon/encode_rs.c uint16_t *genpoly = rs->genpoly; rs 21 lib/reed_solomon/encode_rs.c uint16_t msk = (uint16_t) rs->nn; rs 33 lib/reed_solomon/encode_rs.c par[j] ^= alpha_to[rs_modnn(rs, fb + rs 40 lib/reed_solomon/encode_rs.c par[nroots - 1] = alpha_to[rs_modnn(rs, rs 74 lib/reed_solomon/reed_solomon.c struct rs_codec *rs; rs 76 lib/reed_solomon/reed_solomon.c rs = kzalloc(sizeof(*rs), gfp); rs 77 lib/reed_solomon/reed_solomon.c if (!rs) rs 80 lib/reed_solomon/reed_solomon.c INIT_LIST_HEAD(&rs->list); rs 82 lib/reed_solomon/reed_solomon.c rs->mm = symsize; rs 83 lib/reed_solomon/reed_solomon.c rs->nn = (1 << symsize) - 1; rs 84 lib/reed_solomon/reed_solomon.c rs->fcr = fcr; rs 85 lib/reed_solomon/reed_solomon.c rs->prim = prim; rs 86 lib/reed_solomon/reed_solomon.c rs->nroots = nroots; rs 87 lib/reed_solomon/reed_solomon.c rs->gfpoly = gfpoly; rs 88 lib/reed_solomon/reed_solomon.c rs->gffunc = gffunc; rs 91 lib/reed_solomon/reed_solomon.c rs->alpha_to = kmalloc_array(rs->nn + 1, sizeof(uint16_t), gfp); rs 92 lib/reed_solomon/reed_solomon.c if (rs->alpha_to == NULL) rs 95 lib/reed_solomon/reed_solomon.c rs->index_of = kmalloc_array(rs->nn + 1, sizeof(uint16_t), gfp); rs 96 lib/reed_solomon/reed_solomon.c if (rs->index_of == NULL) rs 99 lib/reed_solomon/reed_solomon.c rs->genpoly = kmalloc_array(rs->nroots + 1, sizeof(uint16_t), gfp); rs 100 lib/reed_solomon/reed_solomon.c if(rs->genpoly == NULL) rs 104 lib/reed_solomon/reed_solomon.c rs->index_of[0] = rs->nn; /* log(zero) = -inf */ rs 105 lib/reed_solomon/reed_solomon.c rs->alpha_to[rs->nn] = 0; /* alpha**-inf = 0 */ rs 108 lib/reed_solomon/reed_solomon.c for (i = 0; i < rs->nn; i++) { rs 109 lib/reed_solomon/reed_solomon.c rs->index_of[sr] = i; rs 110 lib/reed_solomon/reed_solomon.c rs->alpha_to[i] = sr; rs 114 lib/reed_solomon/reed_solomon.c sr &= rs->nn; rs 118 lib/reed_solomon/reed_solomon.c for (i = 0; i < rs->nn; i++) { rs 119 lib/reed_solomon/reed_solomon.c rs->index_of[sr] = i; rs 120 lib/reed_solomon/reed_solomon.c rs->alpha_to[i] = sr; rs 125 lib/reed_solomon/reed_solomon.c if(sr != rs->alpha_to[0]) rs 129 lib/reed_solomon/reed_solomon.c for(iprim = 1; (iprim % prim) != 0; iprim += rs->nn); rs 131 lib/reed_solomon/reed_solomon.c rs->iprim = iprim / prim; rs 134 lib/reed_solomon/reed_solomon.c rs->genpoly[0] = 1; rs 136 lib/reed_solomon/reed_solomon.c rs->genpoly[i + 1] = 1; rs 139 lib/reed_solomon/reed_solomon.c if (rs->genpoly[j] != 0) { rs 140 lib/reed_solomon/reed_solomon.c rs->genpoly[j] = rs->genpoly[j -1] ^ rs 141 lib/reed_solomon/reed_solomon.c rs->alpha_to[rs_modnn(rs, rs 142 lib/reed_solomon/reed_solomon.c rs->index_of[rs->genpoly[j]] + root)]; rs 144 lib/reed_solomon/reed_solomon.c rs->genpoly[j] = rs->genpoly[j - 1]; rs 147 lib/reed_solomon/reed_solomon.c rs->genpoly[0] = rs 148 lib/reed_solomon/reed_solomon.c rs->alpha_to[rs_modnn(rs, rs 149 lib/reed_solomon/reed_solomon.c rs->index_of[rs->genpoly[0]] + root)]; rs 153 lib/reed_solomon/reed_solomon.c rs->genpoly[i] = rs->index_of[rs->genpoly[i]]; rs 155 lib/reed_solomon/reed_solomon.c rs->users = 1; rs 156 lib/reed_solomon/reed_solomon.c list_add(&rs->list, &codec_list); rs 157 lib/reed_solomon/reed_solomon.c return rs; rs 160 lib/reed_solomon/reed_solomon.c kfree(rs->genpoly); rs 161 lib/reed_solomon/reed_solomon.c kfree(rs->index_of); rs 162 lib/reed_solomon/reed_solomon.c kfree(rs->alpha_to); rs 163 lib/reed_solomon/reed_solomon.c kfree(rs); rs 176 lib/reed_solomon/reed_solomon.c void free_rs(struct rs_control *rs) rs 180 lib/reed_solomon/reed_solomon.c if (!rs) rs 183 lib/reed_solomon/reed_solomon.c cd = rs->codec; rs 194 lib/reed_solomon/reed_solomon.c kfree(rs); rs 218 lib/reed_solomon/reed_solomon.c struct rs_control *rs; rs 237 lib/reed_solomon/reed_solomon.c rs = kzalloc(sizeof(*rs) + bsize, gfp); rs 238 lib/reed_solomon/reed_solomon.c if (!rs) rs 261 lib/reed_solomon/reed_solomon.c rs->codec = cd; rs 266 lib/reed_solomon/reed_solomon.c rs->codec = codec_init(symsize, gfpoly, gffunc, fcr, prim, nroots, gfp); rs 267 lib/reed_solomon/reed_solomon.c if (!rs->codec) { rs 268 lib/reed_solomon/reed_solomon.c kfree(rs); rs 269 lib/reed_solomon/reed_solomon.c rs = NULL; rs 273 lib/reed_solomon/reed_solomon.c return rs; rs 108 lib/reed_solomon/test_rslib.c static struct wspace *alloc_ws(struct rs_codec *rs) rs 110 lib/reed_solomon/test_rslib.c int nroots = rs->nroots; rs 112 lib/reed_solomon/test_rslib.c int nn = rs->nn; rs 151 lib/reed_solomon/test_rslib.c static int get_rcw_we(struct rs_control *rs, struct wspace *ws, rs 154 lib/reed_solomon/test_rslib.c int nroots = rs->codec->nroots; rs 158 lib/reed_solomon/test_rslib.c int nn = rs->codec->nn; rs 170 lib/reed_solomon/test_rslib.c encode_rs16(rs, c, dlen, c + dlen, 0); rs 232 lib/reed_solomon/test_rslib.c struct rs_codec *rs = rsc->codec; rs 233 lib/reed_solomon/test_rslib.c uint16_t *alpha_to = rs->alpha_to; rs 234 lib/reed_solomon/test_rslib.c uint16_t *index_of = rs->index_of; rs 235 lib/reed_solomon/test_rslib.c int nroots = rs->nroots; rs 236 lib/reed_solomon/test_rslib.c int prim = rs->prim; rs 237 lib/reed_solomon/test_rslib.c int fcr = rs->fcr; rs 248 lib/reed_solomon/test_rslib.c alpha_to[rs_modnn(rs, index_of[syn[i]] rs 256 lib/reed_solomon/test_rslib.c syn[i] = rs->index_of[syn[i]]; rs 260 lib/reed_solomon/test_rslib.c static void test_uc(struct rs_control *rs, int len, int errs, rs 264 lib/reed_solomon/test_rslib.c int dlen = len - rs->codec->nroots; rs 275 lib/reed_solomon/test_rslib.c nerrs = get_rcw_we(rs, ws, len, errs, eras); rs 279 lib/reed_solomon/test_rslib.c derrs = decode_rs16(rs, r, r + dlen, dlen, rs 284 lib/reed_solomon/test_rslib.c compute_syndrome(rs, r, len, s); rs 285 lib/reed_solomon/test_rslib.c derrs = decode_rs16(rs, NULL, NULL, dlen, rs 290 lib/reed_solomon/test_rslib.c derrs = decode_rs16(rs, r, r + dlen, dlen, rs 313 lib/reed_solomon/test_rslib.c static int ex_rs_helper(struct rs_control *rs, struct wspace *ws, rs 323 lib/reed_solomon/test_rslib.c int nroots = rs->codec->nroots; rs 331 lib/reed_solomon/test_rslib.c test_uc(rs, len, errs, eras, trials, &stat, ws, method); rs 349 lib/reed_solomon/test_rslib.c static int exercise_rs(struct rs_control *rs, struct wspace *ws, rs 360 lib/reed_solomon/test_rslib.c retval |= ex_rs_helper(rs, ws, len, trials, i); rs 366 lib/reed_solomon/test_rslib.c static void test_bc(struct rs_control *rs, int len, int errs, rs 370 lib/reed_solomon/test_rslib.c int nroots = rs->codec->nroots; rs 378 lib/reed_solomon/test_rslib.c get_rcw_we(rs, ws, len, errs, eras); rs 379 lib/reed_solomon/test_rslib.c derrs = decode_rs16(rs, r, r + dlen, dlen, rs 396 lib/reed_solomon/test_rslib.c encode_rs16(rs, r, dlen, corr, 0); rs 407 lib/reed_solomon/test_rslib.c static int exercise_rs_bc(struct rs_control *rs, struct wspace *ws, rs 411 lib/reed_solomon/test_rslib.c int nroots = rs->codec->nroots; rs 424 lib/reed_solomon/test_rslib.c test_bc(rs, len, errs, eras, trials, &stat, ws); rs 74 lib/test_hexdump.c int gs = groupsize, rs = rowsize; rs 78 lib/test_hexdump.c if (rs != 16 && rs != 32) rs 79 lib/test_hexdump.c rs = 16; rs 81 lib/test_hexdump.c if (l > rs) rs 82 lib/test_hexdump.c l = rs; rs 114 lib/test_hexdump.c } while (p < test + rs * 2 + rs / gs + 1); rs 166 lib/test_hexdump.c int rs = rowsize, gs = groupsize; rs 174 lib/test_hexdump.c r = hex_dump_to_buffer(data_b, len, rs, gs, buf, buflen, ascii); rs 180 lib/test_hexdump.c ae = rs * 2 /* hex */ + rs / gs /* spaces */ + 1 /* space */ + len /* ascii */; rs 190 lib/test_hexdump.c test_hexdump_prepare_test(len, rs, gs, test, sizeof(test), ascii); rs 211 lib/test_hexdump.c int rs = (get_random_int() % 2 + 1) * 16; rs 215 lib/test_hexdump.c size_t len = get_random_int() % rs + gs; rs 217 lib/test_hexdump.c test_hexdump_overflow(buflen, rounddown(len, gs), rs, gs, ascii); rs 1515 mm/memory-failure.c #define unpoison_pr_info(fmt, pfn, rs) \ rs 1517 mm/memory-failure.c if (__ratelimit(rs)) \ rs 273 mm/percpu.c static void pcpu_next_unpop(unsigned long *bitmap, int *rs, int *re, int end) rs 275 mm/percpu.c *rs = find_next_zero_bit(bitmap, end, *rs); rs 276 mm/percpu.c *re = find_next_bit(bitmap, end, *rs + 1); rs 279 mm/percpu.c static void pcpu_next_pop(unsigned long *bitmap, int *rs, int *re, int end) rs 281 mm/percpu.c *rs = find_next_bit(bitmap, end, *rs); rs 282 mm/percpu.c *re = find_next_zero_bit(bitmap, end, *rs + 1); rs 290 mm/percpu.c #define pcpu_for_each_unpop_region(bitmap, rs, re, start, end) \ rs 291 mm/percpu.c for ((rs) = (start), pcpu_next_unpop((bitmap), &(rs), &(re), (end)); \ rs 292 mm/percpu.c (rs) < (re); \ rs 293 mm/percpu.c (rs) = (re) + 1, pcpu_next_unpop((bitmap), &(rs), &(re), (end))) rs 295 mm/percpu.c #define pcpu_for_each_pop_region(bitmap, rs, re, start, end) \ rs 296 mm/percpu.c for ((rs) = (start), pcpu_next_pop((bitmap), &(rs), &(re), (end)); \ rs 297 mm/percpu.c (rs) < (re); \ rs 298 mm/percpu.c (rs) = (re) + 1, pcpu_next_pop((bitmap), &(rs), &(re), (end))) rs 752 mm/percpu.c int rs, re, start; /* region start, region end */ rs 768 mm/percpu.c pcpu_for_each_unpop_region(alloc_map, rs, re, start, rs 770 mm/percpu.c pcpu_block_update(block, rs, re); rs 1044 mm/percpu.c int page_start, page_end, rs, re; rs 1049 mm/percpu.c rs = page_start; rs 1050 mm/percpu.c pcpu_next_unpop(chunk->populated, &rs, &re, page_end); rs 1051 mm/percpu.c if (rs >= page_end) rs 1705 mm/percpu.c int page_start, page_end, rs, re; rs 1710 mm/percpu.c pcpu_for_each_unpop_region(chunk->populated, rs, re, rs 1714 mm/percpu.c ret = pcpu_populate_chunk(chunk, rs, re, pcpu_gfp); rs 1722 mm/percpu.c pcpu_chunk_populated(chunk, rs, re); rs 1861 mm/percpu.c int rs, re; rs 1863 mm/percpu.c pcpu_for_each_pop_region(chunk->populated, rs, re, 0, rs 1865 mm/percpu.c pcpu_depopulate_chunk(chunk, rs, re); rs 1867 mm/percpu.c pcpu_chunk_depopulated(chunk, rs, re); rs 1896 mm/percpu.c int nr_unpop = 0, rs, re; rs 1913 mm/percpu.c pcpu_for_each_unpop_region(chunk->populated, rs, re, 0, rs 1915 mm/percpu.c int nr = min(re - rs, nr_to_pop); rs 1917 mm/percpu.c ret = pcpu_populate_chunk(chunk, rs, rs + nr, gfp); rs 1921 mm/percpu.c pcpu_chunk_populated(chunk, rs, rs + nr); rs 175 mm/vmscan.c struct reclaim_state *rs) rs 178 mm/vmscan.c WARN_ON_ONCE(rs && task->reclaim_state); rs 181 mm/vmscan.c WARN_ON_ONCE(!rs && !task->reclaim_state); rs 183 mm/vmscan.c task->reclaim_state = rs; rs 502 net/decnet/dn_neigh.c unsigned char *rs; rs 519 net/decnet/dn_neigh.c s->rs = dn_find_slot(s->ptr, s->n, dn->priority); rs 522 net/decnet/dn_neigh.c if (s->rs == NULL) rs 525 net/decnet/dn_neigh.c dn_dn2eth(s->rs, dn->addr); rs 526 net/decnet/dn_neigh.c s->rs += 6; rs 527 net/decnet/dn_neigh.c *(s->rs) = neigh->nud_state & NUD_CONNECTED ? 0x80 : 0x0; rs 528 net/decnet/dn_neigh.c *(s->rs) |= dn->priority; rs 529 net/decnet/dn_neigh.c s->rs++; rs 540 net/decnet/dn_neigh.c state.rs = ptr; rs 478 net/ipv4/tcp_bbr.c struct sock *sk, const struct rate_sample *rs, u32 acked, u32 *new_cwnd) rs 489 net/ipv4/tcp_bbr.c if (rs->losses > 0) rs 490 net/ipv4/tcp_bbr.c cwnd = max_t(s32, cwnd - rs->losses, 1); rs 516 net/ipv4/tcp_bbr.c static void bbr_set_cwnd(struct sock *sk, const struct rate_sample *rs, rs 526 net/ipv4/tcp_bbr.c if (bbr_set_cwnd_to_recover_or_restore(sk, rs, acked, &cwnd)) rs 552 net/ipv4/tcp_bbr.c const struct rate_sample *rs) rs 567 net/ipv4/tcp_bbr.c inflight = bbr_packets_in_net_at_edt(sk, rs->prior_in_flight); rs 577 net/ipv4/tcp_bbr.c (rs->losses || /* perhaps pacing_gain*BDP won't fit */ rs 599 net/ipv4/tcp_bbr.c const struct rate_sample *rs) rs 603 net/ipv4/tcp_bbr.c if (bbr->mode == BBR_PROBE_BW && bbr_is_next_cycle_phase(sk, rs)) rs 685 net/ipv4/tcp_bbr.c static void bbr_lt_bw_sampling(struct sock *sk, const struct rate_sample *rs) rs 707 net/ipv4/tcp_bbr.c if (!rs->losses) rs 714 net/ipv4/tcp_bbr.c if (rs->is_app_limited) { rs 732 net/ipv4/tcp_bbr.c if (!rs->losses) rs 758 net/ipv4/tcp_bbr.c static void bbr_update_bw(struct sock *sk, const struct rate_sample *rs) rs 765 net/ipv4/tcp_bbr.c if (rs->delivered < 0 || rs->interval_us <= 0) rs 769 net/ipv4/tcp_bbr.c if (!before(rs->prior_delivered, bbr->next_rtt_delivered)) { rs 776 net/ipv4/tcp_bbr.c bbr_lt_bw_sampling(sk, rs); rs 782 net/ipv4/tcp_bbr.c bw = div64_long((u64)rs->delivered * BW_UNIT, rs->interval_us); rs 795 net/ipv4/tcp_bbr.c if (!rs->is_app_limited || bw >= bbr_max_bw(sk)) { rs 815 net/ipv4/tcp_bbr.c const struct rate_sample *rs) rs 821 net/ipv4/tcp_bbr.c if (!bbr_extra_acked_gain || rs->acked_sacked <= 0 || rs 822 net/ipv4/tcp_bbr.c rs->delivered < 0 || rs->interval_us <= 0) rs 846 net/ipv4/tcp_bbr.c (bbr->ack_epoch_acked + rs->acked_sacked >= rs 855 net/ipv4/tcp_bbr.c bbr->ack_epoch_acked + rs->acked_sacked); rs 871 net/ipv4/tcp_bbr.c const struct rate_sample *rs) rs 876 net/ipv4/tcp_bbr.c if (bbr_full_bw_reached(sk) || !bbr->round_start || rs->is_app_limited) rs 890 net/ipv4/tcp_bbr.c static void bbr_check_drain(struct sock *sk, const struct rate_sample *rs) rs 938 net/ipv4/tcp_bbr.c static void bbr_update_min_rtt(struct sock *sk, const struct rate_sample *rs) rs 947 net/ipv4/tcp_bbr.c if (rs->rtt_us >= 0 && rs 948 net/ipv4/tcp_bbr.c (rs->rtt_us <= bbr->min_rtt_us || rs 949 net/ipv4/tcp_bbr.c (filter_expired && !rs->is_ack_delayed))) { rs 950 net/ipv4/tcp_bbr.c bbr->min_rtt_us = rs->rtt_us; rs 980 net/ipv4/tcp_bbr.c if (rs->delivered > 0) rs 1013 net/ipv4/tcp_bbr.c static void bbr_update_model(struct sock *sk, const struct rate_sample *rs) rs 1015 net/ipv4/tcp_bbr.c bbr_update_bw(sk, rs); rs 1016 net/ipv4/tcp_bbr.c bbr_update_ack_aggregation(sk, rs); rs 1017 net/ipv4/tcp_bbr.c bbr_update_cycle_phase(sk, rs); rs 1018 net/ipv4/tcp_bbr.c bbr_check_full_bw_reached(sk, rs); rs 1019 net/ipv4/tcp_bbr.c bbr_check_drain(sk, rs); rs 1020 net/ipv4/tcp_bbr.c bbr_update_min_rtt(sk, rs); rs 1024 net/ipv4/tcp_bbr.c static void bbr_main(struct sock *sk, const struct rate_sample *rs) rs 1029 net/ipv4/tcp_bbr.c bbr_update_model(sk, rs); rs 1033 net/ipv4/tcp_bbr.c bbr_set_cwnd(sk, rs, rs->acked_sacked, bw, bbr->cwnd_gain); rs 1131 net/ipv4/tcp_bbr.c struct rate_sample rs = { .losses = 1 }; rs 1136 net/ipv4/tcp_bbr.c bbr_lt_bw_sampling(sk, &rs); rs 2923 net/ipv4/tcp_input.c long ca_rtt_us, struct rate_sample *rs) rs 2950 net/ipv4/tcp_input.c rs->rtt_us = ca_rtt_us; /* RTT of last (S)ACKed packet (or -1) */ rs 2970 net/ipv4/tcp_input.c struct rate_sample rs; rs 2976 net/ipv4/tcp_input.c tcp_ack_update_rtt(sk, FLAG_SYN_ACKED, rtt_us, -1L, rtt_us, &rs); rs 3324 net/ipv4/tcp_input.c int flag, const struct rate_sample *rs) rs 3329 net/ipv4/tcp_input.c icsk->icsk_ca_ops->cong_control(sk, rs); rs 3590 net/ipv4/tcp_input.c struct rate_sample rs = { .prior_delivered = 0 }; rs 3603 net/ipv4/tcp_input.c sack_state.rate = &rs; rs 3639 net/ipv4/tcp_input.c rs.prior_in_flight = tcp_packets_in_flight(tp); rs 3697 net/ipv4/tcp_input.c tcp_rack_update_reo_wnd(sk, &rs); rs 3721 net/ipv4/tcp_input.c rs.is_ack_delayed = !!(flag & FLAG_ACK_MAYBE_DELAYED); rs 79 net/ipv4/tcp_rate.c struct rate_sample *rs) rs 87 net/ipv4/tcp_rate.c if (!rs->prior_delivered || rs 88 net/ipv4/tcp_rate.c after(scb->tx.delivered, rs->prior_delivered)) { rs 89 net/ipv4/tcp_rate.c rs->prior_delivered = scb->tx.delivered; rs 90 net/ipv4/tcp_rate.c rs->prior_mstamp = scb->tx.delivered_mstamp; rs 91 net/ipv4/tcp_rate.c rs->is_app_limited = scb->tx.is_app_limited; rs 92 net/ipv4/tcp_rate.c rs->is_retrans = scb->sacked & TCPCB_RETRANS; rs 97 net/ipv4/tcp_rate.c rs->interval_us = tcp_stamp_us_delta(tp->first_tx_mstamp, rs 111 net/ipv4/tcp_rate.c bool is_sack_reneg, struct rate_sample *rs) rs 127 net/ipv4/tcp_rate.c rs->acked_sacked = delivered; /* freshly ACKed or SACKed */ rs 128 net/ipv4/tcp_rate.c rs->losses = lost; /* freshly marked lost */ rs 134 net/ipv4/tcp_rate.c if (!rs->prior_mstamp || is_sack_reneg) { rs 135 net/ipv4/tcp_rate.c rs->delivered = -1; rs 136 net/ipv4/tcp_rate.c rs->interval_us = -1; rs 139 net/ipv4/tcp_rate.c rs->delivered = tp->delivered - rs->prior_delivered; rs 146 net/ipv4/tcp_rate.c snd_us = rs->interval_us; /* send phase */ rs 148 net/ipv4/tcp_rate.c rs->prior_mstamp); /* ack phase */ rs 149 net/ipv4/tcp_rate.c rs->interval_us = max(snd_us, ack_us); rs 152 net/ipv4/tcp_rate.c rs->snd_interval_us = snd_us; rs 153 net/ipv4/tcp_rate.c rs->rcv_interval_us = ack_us; rs 162 net/ipv4/tcp_rate.c if (unlikely(rs->interval_us < tcp_min_rtt(tp))) { rs 163 net/ipv4/tcp_rate.c if (!rs->is_retrans) rs 165 net/ipv4/tcp_rate.c rs->interval_us, rs->delivered, rs 168 net/ipv4/tcp_rate.c rs->interval_us = -1; rs 173 net/ipv4/tcp_rate.c if (!rs->is_app_limited || rs 174 net/ipv4/tcp_rate.c ((u64)rs->delivered * tp->rate_interval_us >= rs 175 net/ipv4/tcp_rate.c (u64)tp->rate_delivered * rs->interval_us)) { rs 176 net/ipv4/tcp_rate.c tp->rate_delivered = rs->delivered; rs 177 net/ipv4/tcp_rate.c tp->rate_interval_us = rs->interval_us; rs 178 net/ipv4/tcp_rate.c tp->rate_app_limited = rs->is_app_limited; rs 202 net/ipv4/tcp_recovery.c void tcp_rack_update_reo_wnd(struct sock *sk, struct rate_sample *rs) rs 207 net/ipv4/tcp_recovery.c !rs->prior_delivered) rs 211 net/ipv4/tcp_recovery.c if (before(rs->prior_delivered, tp->rack.last_delivered)) rs 62 net/rds/af_rds.c struct rds_sock *rs; rs 67 net/rds/af_rds.c rs = rds_sk_to_rs(sk); rs 73 net/rds/af_rds.c rds_clear_recv_queue(rs); rs 74 net/rds/af_rds.c rds_cong_remove_socket(rs); rs 76 net/rds/af_rds.c rds_remove_bound(rs); rs 78 net/rds/af_rds.c rds_send_drop_to(rs, NULL); rs 79 net/rds/af_rds.c rds_rdma_drop_keys(rs); rs 80 net/rds/af_rds.c rds_notify_queue_get(rs, NULL); rs 81 net/rds/af_rds.c rds_notify_msg_zcopy_purge(&rs->rs_zcookie_queue); rs 84 net/rds/af_rds.c list_del_init(&rs->rs_item); rs 88 net/rds/af_rds.c rds_trans_put(rs->rs_transport); rs 105 net/rds/af_rds.c void rds_wake_sk_sleep(struct rds_sock *rs) rs 109 net/rds/af_rds.c read_lock_irqsave(&rs->rs_recv_lock, flags); rs 110 net/rds/af_rds.c __rds_wake_sk_sleep(rds_rs_to_sk(rs)); rs 111 net/rds/af_rds.c read_unlock_irqrestore(&rs->rs_recv_lock, flags); rs 117 net/rds/af_rds.c struct rds_sock *rs = rds_sk_to_rs(sock->sk); rs 124 net/rds/af_rds.c if (ipv6_addr_any(&rs->rs_conn_addr)) rs 127 net/rds/af_rds.c if (ipv6_addr_v4mapped(&rs->rs_conn_addr)) { rs 131 net/rds/af_rds.c sin->sin_port = rs->rs_conn_port; rs 132 net/rds/af_rds.c sin->sin_addr.s_addr = rs->rs_conn_addr_v4; rs 137 net/rds/af_rds.c sin6->sin6_port = rs->rs_conn_port; rs 138 net/rds/af_rds.c sin6->sin6_addr = rs->rs_conn_addr; rs 141 net/rds/af_rds.c sin6->sin6_scope_id = rs->rs_bound_scope_id; rs 151 net/rds/af_rds.c if (ipv6_addr_any(&rs->rs_bound_addr)) { rs 152 net/rds/af_rds.c if (ipv6_addr_any(&rs->rs_conn_addr)) { rs 160 net/rds/af_rds.c if (!(ipv6_addr_type(&rs->rs_conn_addr) & rs 174 net/rds/af_rds.c if (ipv6_addr_v4mapped(&rs->rs_bound_addr)) { rs 178 net/rds/af_rds.c sin->sin_port = rs->rs_bound_port; rs 179 net/rds/af_rds.c sin->sin_addr.s_addr = rs->rs_bound_addr_v4; rs 184 net/rds/af_rds.c sin6->sin6_port = rs->rs_bound_port; rs 185 net/rds/af_rds.c sin6->sin6_addr = rs->rs_bound_addr; rs 187 net/rds/af_rds.c sin6->sin6_scope_id = rs->rs_bound_scope_id; rs 216 net/rds/af_rds.c struct rds_sock *rs = rds_sk_to_rs(sk); rs 222 net/rds/af_rds.c if (rs->rs_seen_congestion) rs 225 net/rds/af_rds.c read_lock_irqsave(&rs->rs_recv_lock, flags); rs 226 net/rds/af_rds.c if (!rs->rs_cong_monitor) { rs 230 net/rds/af_rds.c if (rds_cong_updated_since(&rs->rs_cong_track)) rs 233 net/rds/af_rds.c spin_lock(&rs->rs_lock); rs 234 net/rds/af_rds.c if (rs->rs_cong_notify) rs 236 net/rds/af_rds.c spin_unlock(&rs->rs_lock); rs 238 net/rds/af_rds.c if (!list_empty(&rs->rs_recv_queue) || rs 239 net/rds/af_rds.c !list_empty(&rs->rs_notify_queue) || rs 240 net/rds/af_rds.c !list_empty(&rs->rs_zcookie_queue.zcookie_head)) rs 242 net/rds/af_rds.c if (rs->rs_snd_bytes < rds_sk_sndbuf(rs)) rs 246 net/rds/af_rds.c read_unlock_irqrestore(&rs->rs_recv_lock, flags); rs 250 net/rds/af_rds.c rs->rs_seen_congestion = 0; rs 257 net/rds/af_rds.c struct rds_sock *rs = rds_sk_to_rs(sock->sk); rs 265 net/rds/af_rds.c if (rs->rs_transport && rs 266 net/rds/af_rds.c rs->rs_transport->get_tos_map) rs 267 net/rds/af_rds.c tos = rs->rs_transport->get_tos_map(utos); rs 272 net/rds/af_rds.c if (rs->rs_tos || rs->rs_conn) { rs 276 net/rds/af_rds.c rs->rs_tos = tos; rs 281 net/rds/af_rds.c tos = rs->rs_tos; rs 293 net/rds/af_rds.c static int rds_cancel_sent_to(struct rds_sock *rs, char __user *optval, rs 301 net/rds/af_rds.c if (ipv6_addr_any(&rs->rs_bound_addr)) { rs 325 net/rds/af_rds.c rds_send_drop_to(rs, &sin6); rs 343 net/rds/af_rds.c static int rds_cong_monitor(struct rds_sock *rs, char __user *optval, rs 348 net/rds/af_rds.c ret = rds_set_bool_option(&rs->rs_cong_monitor, optval, optlen); rs 350 net/rds/af_rds.c if (rs->rs_cong_monitor) { rs 351 net/rds/af_rds.c rds_cong_add_socket(rs); rs 353 net/rds/af_rds.c rds_cong_remove_socket(rs); rs 354 net/rds/af_rds.c rs->rs_cong_mask = 0; rs 355 net/rds/af_rds.c rs->rs_cong_notify = 0; rs 361 net/rds/af_rds.c static int rds_set_transport(struct rds_sock *rs, char __user *optval, rs 366 net/rds/af_rds.c if (rs->rs_transport) rs 378 net/rds/af_rds.c rs->rs_transport = rds_trans_get(t_type); rs 380 net/rds/af_rds.c return rs->rs_transport ? 0 : -ENOPROTOOPT; rs 407 net/rds/af_rds.c static int rds_recv_track_latency(struct rds_sock *rs, char __user *optval, rs 422 net/rds/af_rds.c rs->rs_rx_traces = trace.rx_traces; rs 423 net/rds/af_rds.c for (i = 0; i < rs->rs_rx_traces; i++) { rs 425 net/rds/af_rds.c rs->rs_rx_traces = 0; rs 428 net/rds/af_rds.c rs->rs_rx_trace[i] = trace.rx_trace_pos[i]; rs 437 net/rds/af_rds.c struct rds_sock *rs = rds_sk_to_rs(sock->sk); rs 447 net/rds/af_rds.c ret = rds_cancel_sent_to(rs, optval, optlen); rs 450 net/rds/af_rds.c ret = rds_get_mr(rs, optval, optlen); rs 453 net/rds/af_rds.c ret = rds_get_mr_for_dest(rs, optval, optlen); rs 456 net/rds/af_rds.c ret = rds_free_mr(rs, optval, optlen); rs 459 net/rds/af_rds.c ret = rds_set_bool_option(&rs->rs_recverr, optval, optlen); rs 462 net/rds/af_rds.c ret = rds_cong_monitor(rs, optval, optlen); rs 466 net/rds/af_rds.c ret = rds_set_transport(rs, optval, optlen); rs 476 net/rds/af_rds.c ret = rds_recv_track_latency(rs, optval, optlen); rs 488 net/rds/af_rds.c struct rds_sock *rs = rds_sk_to_rs(sock->sk); rs 510 net/rds/af_rds.c if (put_user(rs->rs_recverr, (int __user *) optval) || rs 521 net/rds/af_rds.c trans = (rs->rs_transport ? rs->rs_transport->t_type : rs 543 net/rds/af_rds.c struct rds_sock *rs = rds_sk_to_rs(sk); rs 567 net/rds/af_rds.c ipv6_addr_set_v4mapped(sin->sin_addr.s_addr, &rs->rs_conn_addr); rs 568 net/rds/af_rds.c rs->rs_conn_port = sin->sin_port; rs 607 net/rds/af_rds.c (!ipv6_addr_any(&rs->rs_bound_addr) && rs 608 net/rds/af_rds.c rs->rs_bound_scope_id && rs 609 net/rds/af_rds.c sin6->sin6_scope_id != rs->rs_bound_scope_id)) { rs 617 net/rds/af_rds.c rs->rs_bound_scope_id = sin6->sin6_scope_id; rs 619 net/rds/af_rds.c rs->rs_conn_addr = sin6->sin6_addr; rs 620 net/rds/af_rds.c rs->rs_conn_port = sin6->sin6_port; rs 663 net/rds/af_rds.c struct rds_sock *rs = rds_sk_to_rs(sk); rs 665 net/rds/af_rds.c WARN_ON((&rs->rs_item != rs->rs_item.next || rs 666 net/rds/af_rds.c &rs->rs_item != rs->rs_item.prev)); rs 671 net/rds/af_rds.c struct rds_sock *rs; rs 678 net/rds/af_rds.c rs = rds_sk_to_rs(sk); rs 679 net/rds/af_rds.c spin_lock_init(&rs->rs_lock); rs 680 net/rds/af_rds.c rwlock_init(&rs->rs_recv_lock); rs 681 net/rds/af_rds.c INIT_LIST_HEAD(&rs->rs_send_queue); rs 682 net/rds/af_rds.c INIT_LIST_HEAD(&rs->rs_recv_queue); rs 683 net/rds/af_rds.c INIT_LIST_HEAD(&rs->rs_notify_queue); rs 684 net/rds/af_rds.c INIT_LIST_HEAD(&rs->rs_cong_list); rs 685 net/rds/af_rds.c rds_message_zcopy_queue_init(&rs->rs_zcookie_queue); rs 686 net/rds/af_rds.c spin_lock_init(&rs->rs_rdma_lock); rs 687 net/rds/af_rds.c rs->rs_rdma_keys = RB_ROOT; rs 688 net/rds/af_rds.c rs->rs_rx_traces = 0; rs 689 net/rds/af_rds.c rs->rs_tos = 0; rs 690 net/rds/af_rds.c rs->rs_conn = NULL; rs 693 net/rds/af_rds.c list_add_tail(&rs->rs_item, &rds_sock_list); rs 715 net/rds/af_rds.c void rds_sock_addref(struct rds_sock *rs) rs 717 net/rds/af_rds.c sock_hold(rds_rs_to_sk(rs)); rs 720 net/rds/af_rds.c void rds_sock_put(struct rds_sock *rs) rs 722 net/rds/af_rds.c sock_put(rds_rs_to_sk(rs)); rs 735 net/rds/af_rds.c struct rds_sock *rs; rs 743 net/rds/af_rds.c list_for_each_entry(rs, &rds_sock_list, rs_item) { rs 745 net/rds/af_rds.c if (!ipv6_addr_v4mapped(&rs->rs_bound_addr)) rs 748 net/rds/af_rds.c read_lock(&rs->rs_recv_lock); rs 751 net/rds/af_rds.c list_for_each_entry(inc, &rs->rs_recv_queue, i_item) { rs 756 net/rds/af_rds.c rs->rs_bound_addr_v4, rs 760 net/rds/af_rds.c read_unlock(&rs->rs_recv_lock); rs 776 net/rds/af_rds.c struct rds_sock *rs; rs 782 net/rds/af_rds.c list_for_each_entry(rs, &rds_sock_list, rs_item) { rs 783 net/rds/af_rds.c read_lock(&rs->rs_recv_lock); rs 785 net/rds/af_rds.c list_for_each_entry(inc, &rs->rs_recv_queue, i_item) { rs 789 net/rds/af_rds.c &rs->rs_bound_addr, 1); rs 792 net/rds/af_rds.c read_unlock(&rs->rs_recv_lock); rs 808 net/rds/af_rds.c struct rds_sock *rs; rs 819 net/rds/af_rds.c list_for_each_entry(rs, &rds_sock_list, rs_item) { rs 821 net/rds/af_rds.c if (!ipv6_addr_v4mapped(&rs->rs_bound_addr)) rs 823 net/rds/af_rds.c sinfo.sndbuf = rds_sk_sndbuf(rs); rs 824 net/rds/af_rds.c sinfo.rcvbuf = rds_sk_rcvbuf(rs); rs 825 net/rds/af_rds.c sinfo.bound_addr = rs->rs_bound_addr_v4; rs 826 net/rds/af_rds.c sinfo.connected_addr = rs->rs_conn_addr_v4; rs 827 net/rds/af_rds.c sinfo.bound_port = rs->rs_bound_port; rs 828 net/rds/af_rds.c sinfo.connected_port = rs->rs_conn_port; rs 829 net/rds/af_rds.c sinfo.inum = sock_i_ino(rds_rs_to_sk(rs)); rs 848 net/rds/af_rds.c struct rds_sock *rs; rs 857 net/rds/af_rds.c list_for_each_entry(rs, &rds_sock_list, rs_item) { rs 858 net/rds/af_rds.c sinfo6.sndbuf = rds_sk_sndbuf(rs); rs 859 net/rds/af_rds.c sinfo6.rcvbuf = rds_sk_rcvbuf(rs); rs 860 net/rds/af_rds.c sinfo6.bound_addr = rs->rs_bound_addr; rs 861 net/rds/af_rds.c sinfo6.connected_addr = rs->rs_conn_addr; rs 862 net/rds/af_rds.c sinfo6.bound_port = rs->rs_bound_port; rs 863 net/rds/af_rds.c sinfo6.connected_port = rs->rs_conn_port; rs 864 net/rds/af_rds.c sinfo6.inum = sock_i_ino(rds_rs_to_sk(rs)); rs 76 net/rds/bind.c struct rds_sock *rs; rs 80 net/rds/bind.c rs = rhashtable_lookup(&bind_hash_table, key, ht_parms); rs 81 net/rds/bind.c if (rs && (sock_flag(rds_rs_to_sk(rs), SOCK_DEAD) || rs 82 net/rds/bind.c !refcount_inc_not_zero(&rds_rs_to_sk(rs)->sk_refcnt))) rs 83 net/rds/bind.c rs = NULL; rs 87 net/rds/bind.c rdsdebug("returning rs %p for %pI6c:%u\n", rs, addr, rs 90 net/rds/bind.c return rs; rs 94 net/rds/bind.c static int rds_add_bound(struct rds_sock *rs, const struct in6_addr *addr, rs 122 net/rds/bind.c memcpy(rs->rs_bound_key, key, sizeof(rs->rs_bound_key)); rs 123 net/rds/bind.c rs->rs_bound_addr = *addr; rs 124 net/rds/bind.c net_get_random_once(&rs->rs_hash_initval, rs 125 net/rds/bind.c sizeof(rs->rs_hash_initval)); rs 126 net/rds/bind.c rs->rs_bound_port = cpu_to_be16(rover); rs 127 net/rds/bind.c rs->rs_bound_node.next = NULL; rs 128 net/rds/bind.c rds_sock_addref(rs); rs 130 net/rds/bind.c &rs->rs_bound_node, ht_parms)) { rs 131 net/rds/bind.c *port = rs->rs_bound_port; rs 132 net/rds/bind.c rs->rs_bound_scope_id = scope_id; rs 135 net/rds/bind.c rs, addr, (int)ntohs(*port)); rs 138 net/rds/bind.c rs->rs_bound_addr = in6addr_any; rs 139 net/rds/bind.c rds_sock_put(rs); rs 148 net/rds/bind.c void rds_remove_bound(struct rds_sock *rs) rs 151 net/rds/bind.c if (ipv6_addr_any(&rs->rs_bound_addr)) rs 155 net/rds/bind.c rs, &rs->rs_bound_addr, rs 156 net/rds/bind.c ntohs(rs->rs_bound_port)); rs 158 net/rds/bind.c rhashtable_remove_fast(&bind_hash_table, &rs->rs_bound_node, ht_parms); rs 159 net/rds/bind.c rds_sock_put(rs); rs 160 net/rds/bind.c rs->rs_bound_addr = in6addr_any; rs 166 net/rds/bind.c struct rds_sock *rs = rds_sk_to_rs(sk); rs 227 net/rds/bind.c if (!ipv6_addr_any(&rs->rs_bound_addr)) { rs 235 net/rds/bind.c if (!ipv6_addr_any(&rs->rs_conn_addr) && scope_id && rs 236 net/rds/bind.c rs->rs_bound_scope_id && rs 237 net/rds/bind.c scope_id != rs->rs_bound_scope_id) { rs 245 net/rds/bind.c if (rs->rs_transport) { rs 246 net/rds/bind.c trans = rs->rs_transport; rs 262 net/rds/bind.c rs->rs_transport = trans; rs 266 net/rds/bind.c ret = rds_add_bound(rs, binding_addr, &port, scope_id); rs 268 net/rds/bind.c rs->rs_transport = NULL; rs 266 net/rds/cong.c struct rds_sock *rs; rs 269 net/rds/cong.c list_for_each_entry(rs, &rds_cong_monitor, rs_cong_list) { rs 270 net/rds/cong.c spin_lock(&rs->rs_lock); rs 271 net/rds/cong.c rs->rs_cong_notify |= (rs->rs_cong_mask & portmask); rs 272 net/rds/cong.c rs->rs_cong_mask &= ~portmask; rs 273 net/rds/cong.c spin_unlock(&rs->rs_lock); rs 274 net/rds/cong.c if (rs->rs_cong_notify) rs 275 net/rds/cong.c rds_wake_sk_sleep(rs); rs 338 net/rds/cong.c void rds_cong_add_socket(struct rds_sock *rs) rs 343 net/rds/cong.c if (list_empty(&rs->rs_cong_list)) rs 344 net/rds/cong.c list_add(&rs->rs_cong_list, &rds_cong_monitor); rs 348 net/rds/cong.c void rds_cong_remove_socket(struct rds_sock *rs) rs 354 net/rds/cong.c list_del_init(&rs->rs_cong_list); rs 359 net/rds/cong.c map = rds_cong_tree_walk(&rs->rs_bound_addr, NULL); rs 362 net/rds/cong.c if (map && rds_cong_test_bit(map, rs->rs_bound_port)) { rs 363 net/rds/cong.c rds_cong_clear_bit(map, rs->rs_bound_port); rs 369 net/rds/cong.c struct rds_sock *rs) rs 374 net/rds/cong.c if (rs && rs->rs_cong_monitor) { rs 379 net/rds/cong.c spin_lock_irqsave(&rs->rs_lock, flags); rs 380 net/rds/cong.c rs->rs_cong_mask |= RDS_CONG_MONITOR_MASK(ntohs(port)); rs 381 net/rds/cong.c spin_unlock_irqrestore(&rs->rs_lock, flags); rs 124 net/rds/ib_mr.h struct rds_sock *rs, u32 *key_ret, rs 530 net/rds/ib_rdma.c struct rds_sock *rs, u32 *key_ret, rs 538 net/rds/ib_rdma.c rds_ibdev = rds_ib_get_device(rs->rs_bound_addr.s6_addr32[3]); rs 92 net/rds/message.c static void rds_rm_zerocopy_callback(struct rds_sock *rs, rs 103 net/rds/message.c q = &rs->rs_zcookie_queue; rs 140 net/rds/message.c struct rds_sock *rs = rm->m_rs; rs 144 net/rds/message.c rds_rm_zerocopy_callback(rs, rm->data.op_mmp_znotifier); rs 145 net/rds/message.c rds_wake_sk_sleep(rs); rs 148 net/rds/message.c sock_put(rds_rs_to_sk(rs)); rs 97 net/rds/rdma.c struct rds_sock *rs = mr->r_sock; rs 107 net/rds/rdma.c spin_lock_irqsave(&rs->rs_rdma_lock, flags); rs 109 net/rds/rdma.c rb_erase(&mr->r_rb_node, &rs->rs_rdma_keys); rs 112 net/rds/rdma.c spin_unlock_irqrestore(&rs->rs_rdma_lock, flags); rs 128 net/rds/rdma.c void rds_rdma_drop_keys(struct rds_sock *rs) rs 135 net/rds/rdma.c spin_lock_irqsave(&rs->rs_rdma_lock, flags); rs 136 net/rds/rdma.c while ((node = rb_first(&rs->rs_rdma_keys))) { rs 138 net/rds/rdma.c if (mr->r_trans == rs->rs_transport) rs 140 net/rds/rdma.c rb_erase(&mr->r_rb_node, &rs->rs_rdma_keys); rs 142 net/rds/rdma.c spin_unlock_irqrestore(&rs->rs_rdma_lock, flags); rs 145 net/rds/rdma.c spin_lock_irqsave(&rs->rs_rdma_lock, flags); rs 147 net/rds/rdma.c spin_unlock_irqrestore(&rs->rs_rdma_lock, flags); rs 149 net/rds/rdma.c if (rs->rs_transport && rs->rs_transport->flush_mrs) rs 150 net/rds/rdma.c rs->rs_transport->flush_mrs(); rs 173 net/rds/rdma.c static int __rds_rdma_map(struct rds_sock *rs, struct rds_get_mr_args *args, rs 188 net/rds/rdma.c if (ipv6_addr_any(&rs->rs_bound_addr) || !rs->rs_transport) { rs 193 net/rds/rdma.c if (!rs->rs_transport->get_mr) { rs 230 net/rds/rdma.c mr->r_trans = rs->rs_transport; rs 231 net/rds/rdma.c mr->r_sock = rs; rs 273 net/rds/rdma.c trans_private = rs->rs_transport->get_mr(sg, nents, rs, rs 305 net/rds/rdma.c spin_lock_irqsave(&rs->rs_rdma_lock, flags); rs 306 net/rds/rdma.c found = rds_mr_tree_walk(&rs->rs_rdma_keys, mr->r_key, mr); rs 307 net/rds/rdma.c spin_unlock_irqrestore(&rs->rs_rdma_lock, flags); rs 325 net/rds/rdma.c int rds_get_mr(struct rds_sock *rs, char __user *optval, int optlen) rs 336 net/rds/rdma.c return __rds_rdma_map(rs, &args, NULL, NULL, NULL); rs 339 net/rds/rdma.c int rds_get_mr_for_dest(struct rds_sock *rs, char __user *optval, int optlen) rs 360 net/rds/rdma.c return __rds_rdma_map(rs, &new_args, NULL, NULL, NULL); rs 366 net/rds/rdma.c int rds_free_mr(struct rds_sock *rs, char __user *optval, int optlen) rs 381 net/rds/rdma.c if (!rs->rs_transport || !rs->rs_transport->flush_mrs) rs 383 net/rds/rdma.c rs->rs_transport->flush_mrs(); rs 391 net/rds/rdma.c spin_lock_irqsave(&rs->rs_rdma_lock, flags); rs 392 net/rds/rdma.c mr = rds_mr_tree_walk(&rs->rs_rdma_keys, rds_rdma_cookie_key(args.cookie), NULL); rs 394 net/rds/rdma.c rb_erase(&mr->r_rb_node, &rs->rs_rdma_keys); rs 399 net/rds/rdma.c spin_unlock_irqrestore(&rs->rs_rdma_lock, flags); rs 419 net/rds/rdma.c void rds_rdma_unuse(struct rds_sock *rs, u32 r_key, int force) rs 425 net/rds/rdma.c spin_lock_irqsave(&rs->rs_rdma_lock, flags); rs 426 net/rds/rdma.c mr = rds_mr_tree_walk(&rs->rs_rdma_keys, r_key, NULL); rs 430 net/rds/rdma.c spin_unlock_irqrestore(&rs->rs_rdma_lock, flags); rs 435 net/rds/rdma.c rb_erase(&mr->r_rb_node, &rs->rs_rdma_keys); rs 439 net/rds/rdma.c spin_unlock_irqrestore(&rs->rs_rdma_lock, flags); rs 572 net/rds/rdma.c int rds_cmsg_rdma_args(struct rds_sock *rs, struct rds_message *rm, rs 591 net/rds/rdma.c if (ipv6_addr_any(&rs->rs_bound_addr)) { rs 625 net/rds/rdma.c op->op_recverr = rs->rs_recverr; rs 670 net/rds/rdma.c rs->rs_user_addr = iov->addr; rs 671 net/rds/rdma.c rs->rs_user_bytes = iov->bytes; rs 730 net/rds/rdma.c int rds_cmsg_rdma_dest(struct rds_sock *rs, struct rds_message *rm, rs 751 net/rds/rdma.c spin_lock_irqsave(&rs->rs_rdma_lock, flags); rs 752 net/rds/rdma.c mr = rds_mr_tree_walk(&rs->rs_rdma_keys, r_key, NULL); rs 757 net/rds/rdma.c spin_unlock_irqrestore(&rs->rs_rdma_lock, flags); rs 772 net/rds/rdma.c int rds_cmsg_rdma_map(struct rds_sock *rs, struct rds_message *rm, rs 779 net/rds/rdma.c return __rds_rdma_map(rs, CMSG_DATA(cmsg), &rm->m_rdma_cookie, rs 786 net/rds/rdma.c int rds_cmsg_atomic(struct rds_sock *rs, struct rds_message *rm, rs 832 net/rds/rdma.c rm->atomic.op_recverr = rs->rs_recverr; rs 97 net/rds/rds.h #define RDS_MPATH_HASH(rs, n) (jhash_1word((rs)->rs_bound_port, \ rs 98 net/rds/rds.h (rs)->rs_hash_initval) & ((n) - 1)) rs 575 net/rds/rds.h struct rds_sock *rs, u32 *key_ret, rs 670 net/rds/rds.h static inline struct sock *rds_rs_to_sk(struct rds_sock *rs) rs 672 net/rds/rds.h return &rs->rs_sk; rs 680 net/rds/rds.h static inline int rds_sk_sndbuf(struct rds_sock *rs) rs 682 net/rds/rds.h return rds_rs_to_sk(rs)->sk_sndbuf / 2; rs 684 net/rds/rds.h static inline int rds_sk_rcvbuf(struct rds_sock *rs) rs 686 net/rds/rds.h return rds_rs_to_sk(rs)->sk_rcvbuf / 2; rs 729 net/rds/rds.h void rds_sock_addref(struct rds_sock *rs); rs 730 net/rds/rds.h void rds_sock_put(struct rds_sock *rs); rs 731 net/rds/rds.h void rds_wake_sk_sleep(struct rds_sock *rs); rs 744 net/rds/rds.h void rds_remove_bound(struct rds_sock *rs); rs 756 net/rds/rds.h int rds_cong_wait(struct rds_cong_map *map, __be16 port, int nonblock, struct rds_sock *rs); rs 899 net/rds/rds.h void rds_clear_recv_queue(struct rds_sock *rs); rs 900 net/rds/rds.h int rds_notify_queue_get(struct rds_sock *rs, struct msghdr *msg); rs 914 net/rds/rds.h void rds_send_drop_to(struct rds_sock *rs, struct sockaddr_in6 *dest); rs 924 net/rds/rds.h void rds_rdma_unuse(struct rds_sock *rs, u32 r_key, int force); rs 925 net/rds/rds.h int rds_get_mr(struct rds_sock *rs, char __user *optval, int optlen); rs 926 net/rds/rds.h int rds_get_mr_for_dest(struct rds_sock *rs, char __user *optval, int optlen); rs 927 net/rds/rds.h int rds_free_mr(struct rds_sock *rs, char __user *optval, int optlen); rs 928 net/rds/rds.h void rds_rdma_drop_keys(struct rds_sock *rs); rs 931 net/rds/rds.h int rds_cmsg_rdma_dest(struct rds_sock *rs, struct rds_message *rm, rs 933 net/rds/rds.h int rds_cmsg_rdma_args(struct rds_sock *rs, struct rds_message *rm, rs 936 net/rds/rds.h int rds_cmsg_rdma_map(struct rds_sock *rs, struct rds_message *rm, rs 942 net/rds/rds.h int rds_cmsg_atomic(struct rds_sock *rs, struct rds_message *rm, rs 87 net/rds/recv.c static void rds_recv_rcvbuf_delta(struct rds_sock *rs, struct sock *sk, rs 96 net/rds/recv.c rs->rs_rcv_bytes += delta; rs 103 net/rds/recv.c if (rs->rs_transport->t_type == RDS_TRANS_LOOP) rs 106 net/rds/recv.c now_congested = rs->rs_rcv_bytes > rds_sk_rcvbuf(rs); rs 110 net/rds/recv.c rs, &rs->rs_bound_addr, rs 111 net/rds/recv.c ntohs(rs->rs_bound_port), rs->rs_rcv_bytes, rs 112 net/rds/recv.c rds_sk_rcvbuf(rs), now_congested, delta); rs 115 net/rds/recv.c if (!rs->rs_congested && now_congested) { rs 116 net/rds/recv.c rs->rs_congested = 1; rs 123 net/rds/recv.c else if (rs->rs_congested && (rs->rs_rcv_bytes < (rds_sk_rcvbuf(rs)/2))) { rs 124 net/rds/recv.c rs->rs_congested = 0; rs 165 net/rds/recv.c static void rds_recv_incoming_exthdrs(struct rds_incoming *inc, struct rds_sock *rs) rs 183 net/rds/recv.c rds_rdma_unuse(rs, be32_to_cpu(buffer.rdma.h_rdma_rkey), 0); rs 286 net/rds/recv.c struct rds_sock *rs = NULL; rs 362 net/rds/recv.c rs = rds_find_bound(daddr, inc->i_hdr.h_dport, conn->c_bound_if); rs 363 net/rds/recv.c if (!rs) { rs 369 net/rds/recv.c rds_recv_incoming_exthdrs(inc, rs); rs 372 net/rds/recv.c sk = rds_rs_to_sk(rs); rs 375 net/rds/recv.c write_lock_irqsave(&rs->rs_recv_lock, flags); rs 377 net/rds/recv.c rdsdebug("adding inc %p to rs %p's recv queue\n", inc, rs); rs 379 net/rds/recv.c rds_recv_rcvbuf_delta(rs, sk, inc->i_conn->c_lcong, rs 386 net/rds/recv.c list_add_tail(&inc->i_item, &rs->rs_recv_queue); rs 391 net/rds/recv.c write_unlock_irqrestore(&rs->rs_recv_lock, flags); rs 394 net/rds/recv.c if (rs) rs 395 net/rds/recv.c rds_sock_put(rs); rs 403 net/rds/recv.c static int rds_next_incoming(struct rds_sock *rs, struct rds_incoming **inc) rs 408 net/rds/recv.c read_lock_irqsave(&rs->rs_recv_lock, flags); rs 409 net/rds/recv.c if (!list_empty(&rs->rs_recv_queue)) { rs 410 net/rds/recv.c *inc = list_entry(rs->rs_recv_queue.next, rs 415 net/rds/recv.c read_unlock_irqrestore(&rs->rs_recv_lock, flags); rs 421 net/rds/recv.c static int rds_still_queued(struct rds_sock *rs, struct rds_incoming *inc, rs 424 net/rds/recv.c struct sock *sk = rds_rs_to_sk(rs); rs 428 net/rds/recv.c write_lock_irqsave(&rs->rs_recv_lock, flags); rs 433 net/rds/recv.c rds_recv_rcvbuf_delta(rs, sk, inc->i_conn->c_lcong, rs 440 net/rds/recv.c write_unlock_irqrestore(&rs->rs_recv_lock, flags); rs 442 net/rds/recv.c rdsdebug("inc %p rs %p still %d dropped %d\n", inc, rs, ret, drop); rs 450 net/rds/recv.c int rds_notify_queue_get(struct rds_sock *rs, struct msghdr *msghdr) rs 473 net/rds/recv.c spin_lock_irqsave(&rs->rs_lock, flags); rs 474 net/rds/recv.c while (!list_empty(&rs->rs_notify_queue) && count < max_messages) { rs 475 net/rds/recv.c notifier = list_entry(rs->rs_notify_queue.next, rs 480 net/rds/recv.c spin_unlock_irqrestore(&rs->rs_lock, flags); rs 506 net/rds/recv.c spin_lock_irqsave(&rs->rs_lock, flags); rs 507 net/rds/recv.c list_splice(©, &rs->rs_notify_queue); rs 508 net/rds/recv.c spin_unlock_irqrestore(&rs->rs_lock, flags); rs 517 net/rds/recv.c static int rds_notify_cong(struct rds_sock *rs, struct msghdr *msghdr) rs 519 net/rds/recv.c uint64_t notify = rs->rs_cong_notify; rs 528 net/rds/recv.c spin_lock_irqsave(&rs->rs_lock, flags); rs 529 net/rds/recv.c rs->rs_cong_notify &= ~notify; rs 530 net/rds/recv.c spin_unlock_irqrestore(&rs->rs_lock, flags); rs 539 net/rds/recv.c struct rds_sock *rs) rs 552 net/rds/recv.c sock_flag(rds_rs_to_sk(rs), SOCK_RCVTSTAMP)) { rs 556 net/rds/recv.c if (!sock_flag(rds_rs_to_sk(rs), SOCK_TSTAMP_NEW)) { rs 573 net/rds/recv.c if (rs->rs_rx_traces) { rs 579 net/rds/recv.c t.rx_traces = rs->rs_rx_traces; rs 580 net/rds/recv.c for (i = 0; i < rs->rs_rx_traces; i++) { rs 581 net/rds/recv.c j = rs->rs_rx_trace[i]; rs 597 net/rds/recv.c static bool rds_recvmsg_zcookie(struct rds_sock *rs, struct msghdr *msg) rs 599 net/rds/recv.c struct rds_msg_zcopy_queue *q = &rs->rs_zcookie_queue; rs 607 net/rds/recv.c if (!sock_flag(rds_rs_to_sk(rs), SOCK_ZEROCOPY) || rs 636 net/rds/recv.c struct rds_sock *rs = rds_sk_to_rs(sk); rs 655 net/rds/recv.c if (!list_empty(&rs->rs_notify_queue)) { rs 656 net/rds/recv.c ret = rds_notify_queue_get(rs, msg); rs 660 net/rds/recv.c if (rs->rs_cong_notify) { rs 661 net/rds/recv.c ret = rds_notify_cong(rs, msg); rs 665 net/rds/recv.c if (!rds_next_incoming(rs, &inc)) { rs 667 net/rds/recv.c bool reaped = rds_recvmsg_zcookie(rs, msg); rs 674 net/rds/recv.c (!list_empty(&rs->rs_notify_queue) || rs 675 net/rds/recv.c rs->rs_cong_notify || rs 676 net/rds/recv.c rds_next_incoming(rs, &inc)), timeo); rs 700 net/rds/recv.c if (!rds_still_queued(rs, inc, !(msg_flags & MSG_PEEK))) { rs 714 net/rds/recv.c if (rds_cmsg_recv(inc, msg, rs)) { rs 718 net/rds/recv.c rds_recvmsg_zcookie(rs, msg); rs 739 net/rds/recv.c sin6->sin6_scope_id = rs->rs_bound_scope_id; rs 758 net/rds/recv.c void rds_clear_recv_queue(struct rds_sock *rs) rs 760 net/rds/recv.c struct sock *sk = rds_rs_to_sk(rs); rs 764 net/rds/recv.c write_lock_irqsave(&rs->rs_recv_lock, flags); rs 765 net/rds/recv.c list_for_each_entry_safe(inc, tmp, &rs->rs_recv_queue, i_item) { rs 766 net/rds/recv.c rds_recv_rcvbuf_delta(rs, sk, inc->i_conn->c_lcong, rs 772 net/rds/recv.c write_unlock_irqrestore(&rs->rs_recv_lock, flags); rs 473 net/rds/send.c static void rds_send_sndbuf_remove(struct rds_sock *rs, struct rds_message *rm) rs 477 net/rds/send.c assert_spin_locked(&rs->rs_lock); rs 479 net/rds/send.c BUG_ON(rs->rs_snd_bytes < len); rs 480 net/rds/send.c rs->rs_snd_bytes -= len; rs 482 net/rds/send.c if (rs->rs_snd_bytes == 0) rs 502 net/rds/send.c struct rds_sock *rs = NULL; rs 513 net/rds/send.c rs = rm->m_rs; rs 514 net/rds/send.c sock_hold(rds_rs_to_sk(rs)); rs 517 net/rds/send.c spin_lock(&rs->rs_lock); rs 518 net/rds/send.c list_add_tail(¬ifier->n_list, &rs->rs_notify_queue); rs 519 net/rds/send.c spin_unlock(&rs->rs_lock); rs 526 net/rds/send.c if (rs) { rs 527 net/rds/send.c rds_wake_sk_sleep(rs); rs 528 net/rds/send.c sock_put(rds_rs_to_sk(rs)); rs 538 net/rds/send.c struct rds_sock *rs = NULL; rs 549 net/rds/send.c rs = rm->m_rs; rs 550 net/rds/send.c sock_hold(rds_rs_to_sk(rs)); rs 553 net/rds/send.c spin_lock(&rs->rs_lock); rs 554 net/rds/send.c list_add_tail(¬ifier->n_list, &rs->rs_notify_queue); rs 555 net/rds/send.c spin_unlock(&rs->rs_lock); rs 562 net/rds/send.c if (rs) { rs 563 net/rds/send.c rds_wake_sk_sleep(rs); rs 564 net/rds/send.c sock_put(rds_rs_to_sk(rs)); rs 575 net/rds/send.c __rds_send_complete(struct rds_sock *rs, struct rds_message *rm, int status) rs 583 net/rds/send.c list_add_tail(&ro->op_notifier->n_list, &rs->rs_notify_queue); rs 590 net/rds/send.c list_add_tail(&ao->op_notifier->n_list, &rs->rs_notify_queue); rs 608 net/rds/send.c struct rds_sock *rs = NULL; rs 632 net/rds/send.c if (rs != rm->m_rs) { rs 633 net/rds/send.c if (rs) { rs 634 net/rds/send.c rds_wake_sk_sleep(rs); rs 635 net/rds/send.c sock_put(rds_rs_to_sk(rs)); rs 637 net/rds/send.c rs = rm->m_rs; rs 638 net/rds/send.c if (rs) rs 639 net/rds/send.c sock_hold(rds_rs_to_sk(rs)); rs 641 net/rds/send.c if (!rs) rs 643 net/rds/send.c spin_lock(&rs->rs_lock); rs 650 net/rds/send.c rds_send_sndbuf_remove(rs, rm); rs 656 net/rds/send.c &rs->rs_notify_queue); rs 663 net/rds/send.c spin_unlock(&rs->rs_lock); rs 672 net/rds/send.c if (rs) { rs 673 net/rds/send.c rds_wake_sk_sleep(rs); rs 674 net/rds/send.c sock_put(rds_rs_to_sk(rs)); rs 722 net/rds/send.c void rds_send_drop_to(struct rds_sock *rs, struct sockaddr_in6 *dest) rs 731 net/rds/send.c spin_lock_irqsave(&rs->rs_lock, flags); rs 733 net/rds/send.c list_for_each_entry_safe(rm, tmp, &rs->rs_send_queue, m_sock_item) { rs 740 net/rds/send.c rds_send_sndbuf_remove(rs, rm); rs 747 net/rds/send.c spin_unlock_irqrestore(&rs->rs_lock, flags); rs 780 net/rds/send.c spin_lock(&rs->rs_lock); rs 781 net/rds/send.c __rds_send_complete(rs, rm, RDS_RDMA_CANCELED); rs 782 net/rds/send.c spin_unlock(&rs->rs_lock); rs 789 net/rds/send.c rds_wake_sk_sleep(rs); rs 803 net/rds/send.c spin_lock(&rs->rs_lock); rs 804 net/rds/send.c __rds_send_complete(rs, rm, RDS_RDMA_CANCELED); rs 805 net/rds/send.c spin_unlock(&rs->rs_lock); rs 818 net/rds/send.c static int rds_send_queue_rm(struct rds_sock *rs, struct rds_connection *conn, rs 833 net/rds/send.c spin_lock_irqsave(&rs->rs_lock, flags); rs 843 net/rds/send.c if (rs->rs_snd_bytes < rds_sk_sndbuf(rs)) { rs 844 net/rds/send.c rs->rs_snd_bytes += len; rs 851 net/rds/send.c if (rs->rs_snd_bytes >= rds_sk_sndbuf(rs) / 2) rs 854 net/rds/send.c list_add_tail(&rm->m_sock_item, &rs->rs_send_queue); rs 857 net/rds/send.c sock_hold(rds_rs_to_sk(rs)); rs 858 net/rds/send.c rm->m_rs = rs; rs 874 net/rds/send.c rm, len, rs, rs->rs_snd_bytes, rs 880 net/rds/send.c spin_unlock_irqrestore(&rs->rs_lock, flags); rs 971 net/rds/send.c static int rds_cmsg_zcopy(struct rds_sock *rs, struct rds_message *rm, rs 984 net/rds/send.c static int rds_cmsg_send(struct rds_sock *rs, struct rds_message *rm, rs 1005 net/rds/send.c ret = rds_cmsg_rdma_args(rs, rm, cmsg, &vct->vec[ind]); rs 1010 net/rds/send.c ret = rds_cmsg_rdma_dest(rs, rm, cmsg); rs 1014 net/rds/send.c ret = rds_cmsg_rdma_map(rs, rm, cmsg); rs 1027 net/rds/send.c ret = rds_cmsg_atomic(rs, rm, cmsg); rs 1031 net/rds/send.c ret = rds_cmsg_zcopy(rs, rm, cmsg); rs 1045 net/rds/send.c static int rds_send_mprds_hash(struct rds_sock *rs, rs 1051 net/rds/send.c hash = RDS_MPATH_HASH(rs, RDS_MPATH_WORKERS); rs 1053 net/rds/send.c hash = RDS_MPATH_HASH(rs, conn->c_npaths); rs 1104 net/rds/send.c struct rds_sock *rs = rds_sk_to_rs(sk); rs 1119 net/rds/send.c sock_flag(rds_rs_to_sk(rs), SOCK_ZEROCOPY)); rs 1204 net/rds/send.c daddr = rs->rs_conn_addr; rs 1205 net/rds/send.c dport = rs->rs_conn_port; rs 1206 net/rds/send.c scope_id = rs->rs_bound_scope_id; rs 1211 net/rds/send.c if (ipv6_addr_any(&rs->rs_bound_addr) || ipv6_addr_any(&daddr)) { rs 1221 net/rds/send.c ipv6_addr_v4mapped(&rs->rs_bound_addr)) { rs 1230 net/rds/send.c if (scope_id != rs->rs_bound_scope_id) { rs 1232 net/rds/send.c scope_id = rs->rs_bound_scope_id; rs 1233 net/rds/send.c } else if (rs->rs_bound_scope_id) { rs 1252 net/rds/send.c if (payload_len > rds_sk_sndbuf(rs)) { rs 1258 net/rds/send.c if (rs->rs_transport->t_type != RDS_TRANS_TCP) { rs 1292 net/rds/send.c if (rs->rs_conn && ipv6_addr_equal(&rs->rs_conn->c_faddr, &daddr) && rs 1293 net/rds/send.c rs->rs_tos == rs->rs_conn->c_tos) { rs 1294 net/rds/send.c conn = rs->rs_conn; rs 1297 net/rds/send.c &rs->rs_bound_addr, &daddr, rs 1298 net/rds/send.c rs->rs_transport, rs->rs_tos, rs 1305 net/rds/send.c rs->rs_conn = conn; rs 1309 net/rds/send.c cpath = &conn->c_path[rds_send_mprds_hash(rs, conn, nonblock)]; rs 1316 net/rds/send.c ret = rds_cmsg_send(rs, rm, msg, &allocated_mr, &vct); rs 1345 net/rds/send.c ret = rds_cong_wait(conn->c_fcong, dport, nonblock, rs); rs 1347 net/rds/send.c rs->rs_seen_congestion = 1; rs 1350 net/rds/send.c while (!rds_send_queue_rm(rs, conn, cpath, rm, rs->rs_bound_port, rs 1360 net/rds/send.c rds_send_queue_rm(rs, conn, cpath, rm, rs 1361 net/rds/send.c rs->rs_bound_port, rs 1410 net/rds/send.c rds_rdma_unuse(rs, rds_rdma_cookie_key(rm->m_rdma_cookie), 1); rs 124 net/xfrm/xfrm_user.c struct xfrm_replay_state_esn *rs; rs 129 net/xfrm/xfrm_user.c rs = nla_data(rt); rs 131 net/xfrm/xfrm_user.c if (rs->bmp_len > XFRMA_REPLAY_ESN_MAX / sizeof(rs->bmp[0]) / 8) rs 134 net/xfrm/xfrm_user.c if (nla_len(rt) < (int)xfrm_replay_state_esn_len(rs) && rs 135 net/xfrm/xfrm_user.c nla_len(rt) != sizeof(*rs)) rs 251 tools/perf/builtin-stat.c int thread, struct timespec *rs) rs 254 tools/perf/builtin-stat.c u64 val = rs->tv_nsec + rs->tv_sec*1000000000ULL; rs 268 tools/perf/builtin-stat.c static int read_counter(struct evsel *counter, struct timespec *rs) rs 295 tools/perf/builtin-stat.c read_single_counter(counter, cpu, thread, rs)) { rs 324 tools/perf/builtin-stat.c static void read_counters(struct timespec *rs) rs 330 tools/perf/builtin-stat.c ret = read_counter(counter, rs); rs 341 tools/perf/builtin-stat.c struct timespec ts, rs; rs 344 tools/perf/builtin-stat.c diff_timespec(&rs, &ts, &ref_time); rs 346 tools/perf/builtin-stat.c read_counters(&rs); rs 349 tools/perf/builtin-stat.c if (WRITE_STAT_ROUND_EVENT(rs.tv_sec * NSEC_PER_SEC + rs.tv_nsec, INTERVAL)) rs 355 tools/perf/builtin-stat.c print_counters(&rs, 0, NULL); rs 483 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h #define tophys(rd,rs) \ rs 484 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h addis rd,rs,0 rs 486 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h #define tovirt(rd,rs) \ rs 487 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h addis rd,rs,0 rs 493 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h #define tophys(rd,rs) \ rs 494 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h clrldi rd,rs,2 rs 496 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h #define tovirt(rd,rs) \ rs 497 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h rotldi rd,rs,16; \ rs 504 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h #define tophys(rd, rs) addis rd, rs, -PAGE_OFFSET@h rs 505 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h #define tovirt(rd, rs) addis rd, rs, PAGE_OFFSET@h rs 346 tools/testing/selftests/x86/protection_keys.c int r, rs; rs 358 tools/testing/selftests/x86/protection_keys.c rs = sigprocmask(SIG_SETMASK, 0, &newact.sa_mask); rs 359 tools/testing/selftests/x86/protection_keys.c pkey_assert(rs == 0);