rw 79 arch/arc/include/asm/spinlock.h static inline void arch_read_lock(arch_rwlock_t *rw) rw 101 arch/arc/include/asm/spinlock.h : [rwlock] "r" (&(rw->counter)), rw 109 arch/arc/include/asm/spinlock.h static inline int arch_read_trylock(arch_rwlock_t *rw) rw 125 arch/arc/include/asm/spinlock.h : [rwlock] "r" (&(rw->counter)), rw 134 arch/arc/include/asm/spinlock.h static inline void arch_write_lock(arch_rwlock_t *rw) rw 158 arch/arc/include/asm/spinlock.h : [rwlock] "r" (&(rw->counter)), rw 167 arch/arc/include/asm/spinlock.h static inline int arch_write_trylock(arch_rwlock_t *rw) rw 183 arch/arc/include/asm/spinlock.h : [rwlock] "r" (&(rw->counter)), rw 193 arch/arc/include/asm/spinlock.h static inline void arch_read_unlock(arch_rwlock_t *rw) rw 209 arch/arc/include/asm/spinlock.h : [rwlock] "r" (&(rw->counter)) rw 213 arch/arc/include/asm/spinlock.h static inline void arch_write_unlock(arch_rwlock_t *rw) rw 217 arch/arc/include/asm/spinlock.h WRITE_ONCE(rw->counter, __ARCH_RW_LOCK_UNLOCKED__); rw 304 arch/arc/include/asm/spinlock.h static inline int arch_read_trylock(arch_rwlock_t *rw) rw 310 arch/arc/include/asm/spinlock.h arch_spin_lock(&(rw->lock_mutex)); rw 316 arch/arc/include/asm/spinlock.h if (rw->counter > 0) { rw 317 arch/arc/include/asm/spinlock.h rw->counter--; rw 321 arch/arc/include/asm/spinlock.h arch_spin_unlock(&(rw->lock_mutex)); rw 328 arch/arc/include/asm/spinlock.h static inline int arch_write_trylock(arch_rwlock_t *rw) rw 334 arch/arc/include/asm/spinlock.h arch_spin_lock(&(rw->lock_mutex)); rw 342 arch/arc/include/asm/spinlock.h if (rw->counter == __ARCH_RW_LOCK_UNLOCKED__) { rw 343 arch/arc/include/asm/spinlock.h rw->counter = 0; rw 346 arch/arc/include/asm/spinlock.h arch_spin_unlock(&(rw->lock_mutex)); rw 352 arch/arc/include/asm/spinlock.h static inline void arch_read_lock(arch_rwlock_t *rw) rw 354 arch/arc/include/asm/spinlock.h while (!arch_read_trylock(rw)) rw 358 arch/arc/include/asm/spinlock.h static inline void arch_write_lock(arch_rwlock_t *rw) rw 360 arch/arc/include/asm/spinlock.h while (!arch_write_trylock(rw)) rw 364 arch/arc/include/asm/spinlock.h static inline void arch_read_unlock(arch_rwlock_t *rw) rw 369 arch/arc/include/asm/spinlock.h arch_spin_lock(&(rw->lock_mutex)); rw 370 arch/arc/include/asm/spinlock.h rw->counter++; rw 371 arch/arc/include/asm/spinlock.h arch_spin_unlock(&(rw->lock_mutex)); rw 375 arch/arc/include/asm/spinlock.h static inline void arch_write_unlock(arch_rwlock_t *rw) rw 380 arch/arc/include/asm/spinlock.h arch_spin_lock(&(rw->lock_mutex)); rw 381 arch/arc/include/asm/spinlock.h rw->counter = __ARCH_RW_LOCK_UNLOCKED__; rw 382 arch/arc/include/asm/spinlock.h arch_spin_unlock(&(rw->lock_mutex)); rw 139 arch/arm/include/asm/spinlock.h static inline void arch_write_lock(arch_rwlock_t *rw) rw 143 arch/arm/include/asm/spinlock.h prefetchw(&rw->lock); rw 152 arch/arm/include/asm/spinlock.h : "r" (&rw->lock), "r" (0x80000000) rw 158 arch/arm/include/asm/spinlock.h static inline int arch_write_trylock(arch_rwlock_t *rw) rw 162 arch/arm/include/asm/spinlock.h prefetchw(&rw->lock); rw 170 arch/arm/include/asm/spinlock.h : "r" (&rw->lock), "r" (0x80000000) rw 182 arch/arm/include/asm/spinlock.h static inline void arch_write_unlock(arch_rwlock_t *rw) rw 189 arch/arm/include/asm/spinlock.h : "r" (&rw->lock), "r" (0) rw 207 arch/arm/include/asm/spinlock.h static inline void arch_read_lock(arch_rwlock_t *rw) rw 211 arch/arm/include/asm/spinlock.h prefetchw(&rw->lock); rw 221 arch/arm/include/asm/spinlock.h : "r" (&rw->lock) rw 227 arch/arm/include/asm/spinlock.h static inline void arch_read_unlock(arch_rwlock_t *rw) rw 233 arch/arm/include/asm/spinlock.h prefetchw(&rw->lock); rw 241 arch/arm/include/asm/spinlock.h : "r" (&rw->lock) rw 248 arch/arm/include/asm/spinlock.h static inline int arch_read_trylock(arch_rwlock_t *rw) rw 252 arch/arm/include/asm/spinlock.h prefetchw(&rw->lock); rw 260 arch/arm/include/asm/spinlock.h : "r" (&rw->lock) rw 165 arch/ia64/include/asm/spinlock.h #define arch_read_lock_flags(rw, flags) arch_read_lock(rw) rw 167 arch/ia64/include/asm/spinlock.h #define arch_read_lock(rw) \ rw 169 arch/ia64/include/asm/spinlock.h arch_rwlock_t *__read_lock_ptr = (rw); \ rw 180 arch/ia64/include/asm/spinlock.h #define arch_read_unlock(rw) \ rw 182 arch/ia64/include/asm/spinlock.h arch_rwlock_t *__read_lock_ptr = (rw); \ rw 214 arch/ia64/include/asm/spinlock.h #define arch_write_lock(rw) arch_write_lock_flags(rw, 0) rw 216 arch/ia64/include/asm/spinlock.h #define arch_write_trylock(rw) \ rw 224 arch/ia64/include/asm/spinlock.h : "=r"(result) : "r"(rw) : "ar.ccv", "r29", "memory"); \ rw 248 arch/ia64/include/asm/spinlock.h #define arch_write_trylock(rw) \ rw 252 arch/ia64/include/asm/spinlock.h ia64_val = ia64_cmpxchg4_acq((__u32 *)(rw), ia64_set_val, 0); \ rw 59 arch/ia64/include/asm/syscall.h struct pt_regs *regs, unsigned long *args, int rw); rw 2143 arch/ia64/kernel/ptrace.c int rw; rw 2165 arch/ia64/kernel/ptrace.c if (args->rw) rw 2173 arch/ia64/kernel/ptrace.c if (!args->rw) { rw 2182 arch/ia64/kernel/ptrace.c struct pt_regs *regs, unsigned long *args, int rw) rw 2189 arch/ia64/kernel/ptrace.c .rw = rw, rw 1060 arch/microblaze/pci/pci-common.c #define NULL_PCI_OP(rw, size, type) \ rw 1062 arch/microblaze/pci/pci-common.c null_##rw##_config_##size(struct pci_dev *dev, int offset, type val) \ rw 1104 arch/microblaze/pci/pci-common.c #define EARLY_PCI_OP(rw, size, type) \ rw 1105 arch/microblaze/pci/pci-common.c int early_##rw##_config_##size(struct pci_controller *hose, int bus, \ rw 1108 arch/microblaze/pci/pci-common.c return pci_bus_##rw##_config_##size(fake_pci_bus(hose, bus), \ rw 322 arch/mips/include/asm/octeon/cvmx-mio-defs.h uint64_t rw:1; rw 340 arch/mips/include/asm/octeon/cvmx-mio-defs.h uint64_t rw:1; rw 883 arch/mips/include/asm/octeon/cvmx-mio-defs.h uint64_t rw:1; rw 891 arch/mips/include/asm/octeon/cvmx-mio-defs.h uint64_t rw:1; rw 2501 arch/mips/include/asm/octeon/cvmx-mio-defs.h uint64_t rw:1; rw 2519 arch/mips/include/asm/octeon/cvmx-mio-defs.h uint64_t rw:1; rw 66 arch/mips/lasat/picvue.c data |= picvue->rw; rw 95 arch/mips/lasat/picvue.c data &= ~picvue->rw; rw 13 arch/mips/lasat/picvue.h u32 rw; rw 73 arch/parisc/include/asm/spinlock.h static __inline__ void arch_read_lock(arch_rwlock_t *rw) rw 77 arch/parisc/include/asm/spinlock.h arch_spin_lock_flags(&rw->lock, flags); rw 78 arch/parisc/include/asm/spinlock.h rw->counter++; rw 79 arch/parisc/include/asm/spinlock.h arch_spin_unlock(&rw->lock); rw 85 arch/parisc/include/asm/spinlock.h static __inline__ void arch_read_unlock(arch_rwlock_t *rw) rw 89 arch/parisc/include/asm/spinlock.h arch_spin_lock_flags(&rw->lock, flags); rw 90 arch/parisc/include/asm/spinlock.h rw->counter--; rw 91 arch/parisc/include/asm/spinlock.h arch_spin_unlock(&rw->lock); rw 97 arch/parisc/include/asm/spinlock.h static __inline__ int arch_read_trylock(arch_rwlock_t *rw) rw 102 arch/parisc/include/asm/spinlock.h if (arch_spin_trylock(&rw->lock)) { rw 103 arch/parisc/include/asm/spinlock.h rw->counter++; rw 104 arch/parisc/include/asm/spinlock.h arch_spin_unlock(&rw->lock); rw 111 arch/parisc/include/asm/spinlock.h if (rw->counter < 0) rw 115 arch/parisc/include/asm/spinlock.h while (arch_spin_is_locked(&rw->lock) && rw->counter >= 0) rw 123 arch/parisc/include/asm/spinlock.h static __inline__ void arch_write_lock(arch_rwlock_t *rw) rw 128 arch/parisc/include/asm/spinlock.h arch_spin_lock_flags(&rw->lock, flags); rw 130 arch/parisc/include/asm/spinlock.h if (rw->counter != 0) { rw 131 arch/parisc/include/asm/spinlock.h arch_spin_unlock(&rw->lock); rw 134 arch/parisc/include/asm/spinlock.h while (rw->counter != 0) rw 140 arch/parisc/include/asm/spinlock.h rw->counter = -1; /* mark as write-locked */ rw 145 arch/parisc/include/asm/spinlock.h static __inline__ void arch_write_unlock(arch_rwlock_t *rw) rw 147 arch/parisc/include/asm/spinlock.h rw->counter = 0; rw 148 arch/parisc/include/asm/spinlock.h arch_spin_unlock(&rw->lock); rw 153 arch/parisc/include/asm/spinlock.h static __inline__ int arch_write_trylock(arch_rwlock_t *rw) rw 159 arch/parisc/include/asm/spinlock.h if (arch_spin_trylock(&rw->lock)) { rw 160 arch/parisc/include/asm/spinlock.h if (rw->counter == 0) { rw 161 arch/parisc/include/asm/spinlock.h rw->counter = -1; rw 165 arch/parisc/include/asm/spinlock.h arch_spin_unlock(&rw->lock); rw 209 arch/powerpc/include/asm/spinlock.h static inline long __arch_read_trylock(arch_rwlock_t *rw) rw 223 arch/powerpc/include/asm/spinlock.h : "r" (&rw->lock) rw 233 arch/powerpc/include/asm/spinlock.h static inline long __arch_write_trylock(arch_rwlock_t *rw) rw 247 arch/powerpc/include/asm/spinlock.h : "r" (token), "r" (&rw->lock) rw 253 arch/powerpc/include/asm/spinlock.h static inline void arch_read_lock(arch_rwlock_t *rw) rw 256 arch/powerpc/include/asm/spinlock.h if (likely(__arch_read_trylock(rw) > 0)) rw 261 arch/powerpc/include/asm/spinlock.h splpar_rw_yield(rw); rw 262 arch/powerpc/include/asm/spinlock.h } while (unlikely(rw->lock < 0)); rw 267 arch/powerpc/include/asm/spinlock.h static inline void arch_write_lock(arch_rwlock_t *rw) rw 270 arch/powerpc/include/asm/spinlock.h if (likely(__arch_write_trylock(rw) == 0)) rw 275 arch/powerpc/include/asm/spinlock.h splpar_rw_yield(rw); rw 276 arch/powerpc/include/asm/spinlock.h } while (unlikely(rw->lock != 0)); rw 281 arch/powerpc/include/asm/spinlock.h static inline int arch_read_trylock(arch_rwlock_t *rw) rw 283 arch/powerpc/include/asm/spinlock.h return __arch_read_trylock(rw) > 0; rw 286 arch/powerpc/include/asm/spinlock.h static inline int arch_write_trylock(arch_rwlock_t *rw) rw 288 arch/powerpc/include/asm/spinlock.h return __arch_write_trylock(rw) == 0; rw 291 arch/powerpc/include/asm/spinlock.h static inline void arch_read_unlock(arch_rwlock_t *rw) rw 304 arch/powerpc/include/asm/spinlock.h : "r"(&rw->lock) rw 308 arch/powerpc/include/asm/spinlock.h static inline void arch_write_unlock(arch_rwlock_t *rw) rw 312 arch/powerpc/include/asm/spinlock.h rw->lock = 0; rw 1515 arch/powerpc/kernel/pci-common.c #define NULL_PCI_OP(rw, size, type) \ rw 1517 arch/powerpc/kernel/pci-common.c null_##rw##_config_##size(struct pci_dev *dev, int offset, type val) \ rw 1560 arch/powerpc/kernel/pci-common.c #define EARLY_PCI_OP(rw, size, type) \ rw 1561 arch/powerpc/kernel/pci-common.c int early_##rw##_config_##size(struct pci_controller *hose, int bus, \ rw 1564 arch/powerpc/kernel/pci-common.c return pci_bus_##rw##_config_##size(fake_pci_bus(hose, bus), \ rw 46 arch/powerpc/lib/locks.c void splpar_rw_yield(arch_rwlock_t *rw) rw 51 arch/powerpc/lib/locks.c lock_value = rw->lock; rw 60 arch/powerpc/lib/locks.c if (rw->lock != lock_value) rw 115 arch/powerpc/platforms/powermac/low_i2c.c int rw; rw 284 arch/powerpc/platforms/powermac/low_i2c.c else if (host->rw) { rw 444 arch/powerpc/platforms/powermac/low_i2c.c host->rw = (addrdir & 1); rw 26 arch/riscv/include/asm/barrier.h #define __smp_mb() RISCV_FENCE(rw,rw) rw 33 arch/riscv/include/asm/barrier.h RISCV_FENCE(rw,w); \ rw 41 arch/riscv/include/asm/barrier.h RISCV_FENCE(r,rw); \ rw 62 arch/riscv/include/asm/barrier.h #define smp_mb__after_spinlock() RISCV_FENCE(rw,rw) rw 106 arch/s390/include/asm/spinlock.h #define arch_read_relax(rw) barrier() rw 107 arch/s390/include/asm/spinlock.h #define arch_write_relax(rw) barrier() rw 112 arch/s390/include/asm/spinlock.h static inline void arch_read_lock(arch_rwlock_t *rw) rw 116 arch/s390/include/asm/spinlock.h old = __atomic_add(1, &rw->cnts); rw 118 arch/s390/include/asm/spinlock.h arch_read_lock_wait(rw); rw 121 arch/s390/include/asm/spinlock.h static inline void arch_read_unlock(arch_rwlock_t *rw) rw 123 arch/s390/include/asm/spinlock.h __atomic_add_const_barrier(-1, &rw->cnts); rw 126 arch/s390/include/asm/spinlock.h static inline void arch_write_lock(arch_rwlock_t *rw) rw 128 arch/s390/include/asm/spinlock.h if (!__atomic_cmpxchg_bool(&rw->cnts, 0, 0x30000)) rw 129 arch/s390/include/asm/spinlock.h arch_write_lock_wait(rw); rw 132 arch/s390/include/asm/spinlock.h static inline void arch_write_unlock(arch_rwlock_t *rw) rw 134 arch/s390/include/asm/spinlock.h __atomic_add_barrier(-0x30000, &rw->cnts); rw 138 arch/s390/include/asm/spinlock.h static inline int arch_read_trylock(arch_rwlock_t *rw) rw 142 arch/s390/include/asm/spinlock.h old = READ_ONCE(rw->cnts); rw 144 arch/s390/include/asm/spinlock.h __atomic_cmpxchg_bool(&rw->cnts, old, old + 1)); rw 147 arch/s390/include/asm/spinlock.h static inline int arch_write_trylock(arch_rwlock_t *rw) rw 151 arch/s390/include/asm/spinlock.h old = READ_ONCE(rw->cnts); rw 152 arch/s390/include/asm/spinlock.h return !old && __atomic_cmpxchg_bool(&rw->cnts, 0, 0x30000); rw 270 arch/s390/lib/spinlock.c void arch_read_lock_wait(arch_rwlock_t *rw) rw 273 arch/s390/lib/spinlock.c while (READ_ONCE(rw->cnts) & 0x10000) rw 279 arch/s390/lib/spinlock.c __atomic_add_const(-1, &rw->cnts); rw 281 arch/s390/lib/spinlock.c arch_spin_lock(&rw->wait); rw 283 arch/s390/lib/spinlock.c __atomic_add_const(1, &rw->cnts); rw 285 arch/s390/lib/spinlock.c while (READ_ONCE(rw->cnts) & 0x10000) rw 287 arch/s390/lib/spinlock.c arch_spin_unlock(&rw->wait); rw 291 arch/s390/lib/spinlock.c void arch_write_lock_wait(arch_rwlock_t *rw) rw 296 arch/s390/lib/spinlock.c __atomic_add(0x20000, &rw->cnts); rw 299 arch/s390/lib/spinlock.c arch_spin_lock(&rw->wait); rw 302 arch/s390/lib/spinlock.c old = READ_ONCE(rw->cnts); rw 304 arch/s390/lib/spinlock.c __atomic_cmpxchg_bool(&rw->cnts, old, old | 0x10000)) rw 310 arch/s390/lib/spinlock.c arch_spin_unlock(&rw->wait); rw 27 arch/s390/pci/pci_event.c u32 rw : 1; /* read/write */ rw 33 arch/sh/drivers/pci/common.c #define EARLY_PCI_OP(rw, size, type) \ rw 34 arch/sh/drivers/pci/common.c int __init early_##rw##_config_##size(struct pci_channel *hose, \ rw 37 arch/sh/drivers/pci/common.c return pci_##rw##_config_##size( \ rw 52 arch/sh/include/asm/spinlock-cas.h static inline void arch_read_lock(arch_rwlock_t *rw) rw 55 arch/sh/include/asm/spinlock-cas.h do old = rw->lock; rw 56 arch/sh/include/asm/spinlock-cas.h while (!old || __sl_cas(&rw->lock, old, old-1) != old); rw 59 arch/sh/include/asm/spinlock-cas.h static inline void arch_read_unlock(arch_rwlock_t *rw) rw 62 arch/sh/include/asm/spinlock-cas.h do old = rw->lock; rw 63 arch/sh/include/asm/spinlock-cas.h while (__sl_cas(&rw->lock, old, old+1) != old); rw 66 arch/sh/include/asm/spinlock-cas.h static inline void arch_write_lock(arch_rwlock_t *rw) rw 68 arch/sh/include/asm/spinlock-cas.h while (__sl_cas(&rw->lock, RW_LOCK_BIAS, 0) != RW_LOCK_BIAS); rw 71 arch/sh/include/asm/spinlock-cas.h static inline void arch_write_unlock(arch_rwlock_t *rw) rw 73 arch/sh/include/asm/spinlock-cas.h __sl_cas(&rw->lock, 0, RW_LOCK_BIAS); rw 76 arch/sh/include/asm/spinlock-cas.h static inline int arch_read_trylock(arch_rwlock_t *rw) rw 79 arch/sh/include/asm/spinlock-cas.h do old = rw->lock; rw 80 arch/sh/include/asm/spinlock-cas.h while (old && __sl_cas(&rw->lock, old, old-1) != old); rw 84 arch/sh/include/asm/spinlock-cas.h static inline int arch_write_trylock(arch_rwlock_t *rw) rw 86 arch/sh/include/asm/spinlock-cas.h return __sl_cas(&rw->lock, RW_LOCK_BIAS, 0) == RW_LOCK_BIAS; rw 90 arch/sh/include/asm/spinlock-llsc.h static inline void arch_read_lock(arch_rwlock_t *rw) rw 103 arch/sh/include/asm/spinlock-llsc.h : "r" (&rw->lock) rw 108 arch/sh/include/asm/spinlock-llsc.h static inline void arch_read_unlock(arch_rwlock_t *rw) rw 119 arch/sh/include/asm/spinlock-llsc.h : "r" (&rw->lock) rw 124 arch/sh/include/asm/spinlock-llsc.h static inline void arch_write_lock(arch_rwlock_t *rw) rw 137 arch/sh/include/asm/spinlock-llsc.h : "r" (&rw->lock), "r" (RW_LOCK_BIAS) rw 142 arch/sh/include/asm/spinlock-llsc.h static inline void arch_write_unlock(arch_rwlock_t *rw) rw 147 arch/sh/include/asm/spinlock-llsc.h : "r" (&rw->lock), "r" (RW_LOCK_BIAS) rw 152 arch/sh/include/asm/spinlock-llsc.h static inline int arch_read_trylock(arch_rwlock_t *rw) rw 168 arch/sh/include/asm/spinlock-llsc.h : "r" (&rw->lock) rw 175 arch/sh/include/asm/spinlock-llsc.h static inline int arch_write_trylock(arch_rwlock_t *rw) rw 191 arch/sh/include/asm/spinlock-llsc.h : "r" (&rw->lock), "r" (RW_LOCK_BIAS) rw 79 arch/sparc/include/asm/spinlock_32.h static inline void __arch_read_lock(arch_rwlock_t *rw) rw 82 arch/sparc/include/asm/spinlock_32.h lp = rw; rw 99 arch/sparc/include/asm/spinlock_32.h static inline void __arch_read_unlock(arch_rwlock_t *rw) rw 102 arch/sparc/include/asm/spinlock_32.h lp = rw; rw 119 arch/sparc/include/asm/spinlock_32.h static inline void arch_write_lock(arch_rwlock_t *rw) rw 122 arch/sparc/include/asm/spinlock_32.h lp = rw; rw 142 arch/sparc/include/asm/spinlock_32.h static inline int arch_write_trylock(arch_rwlock_t *rw) rw 148 arch/sparc/include/asm/spinlock_32.h : "r" (&rw->lock) rw 152 arch/sparc/include/asm/spinlock_32.h val = rw->lock & ~0xff; rw 154 arch/sparc/include/asm/spinlock_32.h ((volatile u8*)&rw->lock)[3] = 0; rw 156 arch/sparc/include/asm/spinlock_32.h *(volatile u32*)&rw->lock = ~0U; rw 162 arch/sparc/include/asm/spinlock_32.h static inline int __arch_read_trylock(arch_rwlock_t *rw) rw 166 arch/sparc/include/asm/spinlock_32.h lp = rw; rw 124 arch/sparc/kernel/process_32.c struct reg_window32 *rw = (struct reg_window32 *) r->u_regs[14]; rw 140 arch/sparc/kernel/process_32.c rw->locals[0], rw->locals[1], rw->locals[2], rw->locals[3], rw 141 arch/sparc/kernel/process_32.c rw->locals[4], rw->locals[5], rw->locals[6], rw->locals[7]); rw 143 arch/sparc/kernel/process_32.c rw->ins[0], rw->ins[1], rw->ins[2], rw->ins[3], rw 144 arch/sparc/kernel/process_32.c rw->ins[4], rw->ins[5], rw->ins[6], rw->ins[7]); rw 155 arch/sparc/kernel/process_32.c struct reg_window32 *rw; rw 171 arch/sparc/kernel/process_32.c rw = (struct reg_window32 *) fp; rw 172 arch/sparc/kernel/process_32.c pc = rw->ins[7]; rw 175 arch/sparc/kernel/process_32.c fp = rw->ins[6]; rw 466 arch/sparc/kernel/process_32.c struct reg_window32 *rw; rw 479 arch/sparc/kernel/process_32.c rw = (struct reg_window32 *) fp; rw 480 arch/sparc/kernel/process_32.c pc = rw->ins[7]; rw 485 arch/sparc/kernel/process_32.c fp = rw->ins[6] + bias; rw 111 arch/sparc/kernel/process_64.c struct reg_window32 __user *rw; rw 116 arch/sparc/kernel/process_64.c rw = compat_ptr((unsigned int)regs->u_regs[14]); rw 119 arch/sparc/kernel/process_64.c if (copy_from_user (&r_w, rw, sizeof(r_w))) { rw 140 arch/sparc/kernel/process_64.c struct reg_window __user *rw; rw 147 arch/sparc/kernel/process_64.c rw = (struct reg_window __user *) rw 154 arch/sparc/kernel/process_64.c if (copy_from_user (&r_w, rw, sizeof(r_w))) { rw 219 arch/sparc/kernel/process_64.c struct reg_window *rw; rw 221 arch/sparc/kernel/process_64.c rw = (struct reg_window *) rw 223 arch/sparc/kernel/process_64.c if (kstack_valid(tp, (unsigned long) rw)) { rw 224 arch/sparc/kernel/process_64.c rp->i7 = rw->ins[7]; rw 225 arch/sparc/kernel/process_64.c rw = (struct reg_window *) rw 226 arch/sparc/kernel/process_64.c (rw->ins[6] + STACK_BIAS); rw 227 arch/sparc/kernel/process_64.c if (kstack_valid(tp, (unsigned long) rw)) rw 228 arch/sparc/kernel/process_64.c rp->rpc = rw->ins[7]; rw 774 arch/sparc/kernel/process_64.c struct reg_window *rw; rw 789 arch/sparc/kernel/process_64.c rw = (struct reg_window *) fp; rw 790 arch/sparc/kernel/process_64.c pc = rw->ins[7]; rw 795 arch/sparc/kernel/process_64.c fp = rw->ins[6] + bias; rw 72 arch/sparc/kernel/traps_32.c struct reg_window32 *rw = (struct reg_window32 *)regs->u_regs[UREG_FP]; rw 78 arch/sparc/kernel/traps_32.c while(rw && rw 80 arch/sparc/kernel/traps_32.c (((unsigned long) rw) >= PAGE_OFFSET) && rw 81 arch/sparc/kernel/traps_32.c !(((unsigned long) rw) & 0x7)) { rw 82 arch/sparc/kernel/traps_32.c printk("Caller[%08lx]: %pS\n", rw->ins[7], rw 83 arch/sparc/kernel/traps_32.c (void *) rw->ins[7]); rw 84 arch/sparc/kernel/traps_32.c rw = (struct reg_window32 *)rw->ins[6]; rw 2515 arch/sparc/kernel/traps_64.c static inline struct reg_window *kernel_stack_up(struct reg_window *rw) rw 2517 arch/sparc/kernel/traps_64.c unsigned long fp = rw->ins[6]; rw 2544 arch/sparc/kernel/traps_64.c struct reg_window *rw = (struct reg_window *) rw 2550 arch/sparc/kernel/traps_64.c while (rw && rw 2552 arch/sparc/kernel/traps_64.c kstack_valid(tp, (unsigned long) rw)) { rw 2553 arch/sparc/kernel/traps_64.c printk("Caller[%016lx]: %pS\n", rw->ins[7], rw 2554 arch/sparc/kernel/traps_64.c (void *) rw->ins[7]); rw 2556 arch/sparc/kernel/traps_64.c rw = kernel_stack_up(rw); rw 101 arch/x86/kvm/trace.h TP_PROTO(unsigned int rw, unsigned int port, unsigned int size, rw 103 arch/x86/kvm/trace.h TP_ARGS(rw, port, size, count, data), rw 106 arch/x86/kvm/trace.h __field( unsigned int, rw ) rw 114 arch/x86/kvm/trace.h __entry->rw = rw; rw 127 arch/x86/kvm/trace.h __entry->rw ? "write" : "read", rw 195 arch/x86/kvm/trace.h TP_PROTO(unsigned int rw, unsigned int reg, unsigned int val), rw 196 arch/x86/kvm/trace.h TP_ARGS(rw, reg, val), rw 199 arch/x86/kvm/trace.h __field( unsigned int, rw ) rw 205 arch/x86/kvm/trace.h __entry->rw = rw; rw 211 arch/x86/kvm/trace.h __entry->rw ? "write" : "read", rw 362 arch/x86/kvm/trace.h TP_PROTO(unsigned int rw, unsigned int cr, unsigned long val), rw 363 arch/x86/kvm/trace.h TP_ARGS(rw, cr, val), rw 366 arch/x86/kvm/trace.h __field( unsigned int, rw ) rw 372 arch/x86/kvm/trace.h __entry->rw = rw; rw 378 arch/x86/kvm/trace.h __entry->rw ? "write" : "read", rw 1323 arch/x86/kvm/trace.h TP_PROTO(u32 vcpu, u32 offset, bool ft, bool rw, u32 vec), rw 1324 arch/x86/kvm/trace.h TP_ARGS(vcpu, offset, ft, rw, vec), rw 1330 arch/x86/kvm/trace.h __field(bool, rw) rw 1338 arch/x86/kvm/trace.h __entry->rw = rw; rw 1347 arch/x86/kvm/trace.h __entry->rw ? "write" : "read", rw 1350 block/blk-core.c int rw = rq_data_dir(rq); rw 1359 block/blk-core.c part_stat_inc(part, merges[rw]); rw 1374 block/blk-core.c part_inc_in_flight(rq->q, part, rw); rw 1282 block/blk-iocost.c int cpu, rw; rw 1288 block/blk-iocost.c for (rw = READ; rw <= WRITE; rw++) { rw 1289 block/blk-iocost.c u32 this_met = READ_ONCE(stat->missed[rw].nr_met); rw 1290 block/blk-iocost.c u32 this_missed = READ_ONCE(stat->missed[rw].nr_missed); rw 1292 block/blk-iocost.c nr_met[rw] += this_met - stat->missed[rw].last_met; rw 1293 block/blk-iocost.c nr_missed[rw] += this_missed - stat->missed[rw].last_missed; rw 1294 block/blk-iocost.c stat->missed[rw].last_met = this_met; rw 1295 block/blk-iocost.c stat->missed[rw].last_missed = this_missed; rw 1303 block/blk-iocost.c for (rw = READ; rw <= WRITE; rw++) { rw 1304 block/blk-iocost.c if (nr_met[rw] + nr_missed[rw]) rw 1305 block/blk-iocost.c missed_ppm_ar[rw] = rw 1306 block/blk-iocost.c DIV64_U64_ROUND_UP((u64)nr_missed[rw] * MILLION, rw 1307 block/blk-iocost.c nr_met[rw] + nr_missed[rw]); rw 1309 block/blk-iocost.c missed_ppm_ar[rw] = 0; rw 1875 block/blk-iocost.c int pidx, rw; rw 1883 block/blk-iocost.c rw = READ; rw 1887 block/blk-iocost.c rw = WRITE; rw 1897 block/blk-iocost.c this_cpu_inc(ioc->pcpu_stat->missed[rw].nr_met); rw 1899 block/blk-iocost.c this_cpu_inc(ioc->pcpu_stat->missed[rw].nr_missed); rw 292 block/blk-throttle.c static uint64_t tg_bps_limit(struct throtl_grp *tg, int rw) rw 302 block/blk-throttle.c ret = tg->bps[rw][td->limit_index]; rw 306 block/blk-throttle.c tg->iops[rw][td->limit_index]) rw 312 block/blk-throttle.c if (td->limit_index == LIMIT_MAX && tg->bps[rw][LIMIT_LOW] && rw 313 block/blk-throttle.c tg->bps[rw][LIMIT_LOW] != tg->bps[rw][LIMIT_MAX]) { rw 316 block/blk-throttle.c adjusted = throtl_adjusted_limit(tg->bps[rw][LIMIT_LOW], td); rw 317 block/blk-throttle.c ret = min(tg->bps[rw][LIMIT_MAX], adjusted); rw 322 block/blk-throttle.c static unsigned int tg_iops_limit(struct throtl_grp *tg, int rw) rw 332 block/blk-throttle.c ret = tg->iops[rw][td->limit_index]; rw 336 block/blk-throttle.c tg->bps[rw][td->limit_index]) rw 342 block/blk-throttle.c if (td->limit_index == LIMIT_MAX && tg->iops[rw][LIMIT_LOW] && rw 343 block/blk-throttle.c tg->iops[rw][LIMIT_LOW] != tg->iops[rw][LIMIT_MAX]) { rw 346 block/blk-throttle.c adjusted = throtl_adjusted_limit(tg->iops[rw][LIMIT_LOW], td); rw 349 block/blk-throttle.c ret = min_t(unsigned int, tg->iops[rw][LIMIT_MAX], adjusted); rw 486 block/blk-throttle.c int rw; rw 494 block/blk-throttle.c for (rw = READ; rw <= WRITE; rw++) { rw 495 block/blk-throttle.c throtl_qnode_init(&tg->qnode_on_self[rw], tg); rw 496 block/blk-throttle.c throtl_qnode_init(&tg->qnode_on_parent[rw], tg); rw 553 block/blk-throttle.c int rw; rw 555 block/blk-throttle.c for (rw = READ; rw <= WRITE; rw++) rw 556 block/blk-throttle.c tg->has_rules[rw] = (parent_tg && parent_tg->has_rules[rw]) || rw 558 block/blk-throttle.c (tg_bps_limit(tg, rw) != U64_MAX || rw 559 block/blk-throttle.c tg_iops_limit(tg, rw) != UINT_MAX)); rw 760 block/blk-throttle.c bool rw, unsigned long start) rw 762 block/blk-throttle.c tg->bytes_disp[rw] = 0; rw 763 block/blk-throttle.c tg->io_disp[rw] = 0; rw 771 block/blk-throttle.c if (time_after_eq(start, tg->slice_start[rw])) rw 772 block/blk-throttle.c tg->slice_start[rw] = start; rw 774 block/blk-throttle.c tg->slice_end[rw] = jiffies + tg->td->throtl_slice; rw 777 block/blk-throttle.c rw == READ ? 'R' : 'W', tg->slice_start[rw], rw 778 block/blk-throttle.c tg->slice_end[rw], jiffies); rw 781 block/blk-throttle.c static inline void throtl_start_new_slice(struct throtl_grp *tg, bool rw) rw 783 block/blk-throttle.c tg->bytes_disp[rw] = 0; rw 784 block/blk-throttle.c tg->io_disp[rw] = 0; rw 785 block/blk-throttle.c tg->slice_start[rw] = jiffies; rw 786 block/blk-throttle.c tg->slice_end[rw] = jiffies + tg->td->throtl_slice; rw 789 block/blk-throttle.c rw == READ ? 'R' : 'W', tg->slice_start[rw], rw 790 block/blk-throttle.c tg->slice_end[rw], jiffies); rw 793 block/blk-throttle.c static inline void throtl_set_slice_end(struct throtl_grp *tg, bool rw, rw 796 block/blk-throttle.c tg->slice_end[rw] = roundup(jiffy_end, tg->td->throtl_slice); rw 799 block/blk-throttle.c static inline void throtl_extend_slice(struct throtl_grp *tg, bool rw, rw 802 block/blk-throttle.c tg->slice_end[rw] = roundup(jiffy_end, tg->td->throtl_slice); rw 805 block/blk-throttle.c rw == READ ? 'R' : 'W', tg->slice_start[rw], rw 806 block/blk-throttle.c tg->slice_end[rw], jiffies); rw 810 block/blk-throttle.c static bool throtl_slice_used(struct throtl_grp *tg, bool rw) rw 812 block/blk-throttle.c if (time_in_range(jiffies, tg->slice_start[rw], tg->slice_end[rw])) rw 819 block/blk-throttle.c static inline void throtl_trim_slice(struct throtl_grp *tg, bool rw) rw 824 block/blk-throttle.c BUG_ON(time_before(tg->slice_end[rw], tg->slice_start[rw])); rw 831 block/blk-throttle.c if (throtl_slice_used(tg, rw)) rw 842 block/blk-throttle.c throtl_set_slice_end(tg, rw, jiffies + tg->td->throtl_slice); rw 844 block/blk-throttle.c time_elapsed = jiffies - tg->slice_start[rw]; rw 850 block/blk-throttle.c tmp = tg_bps_limit(tg, rw) * tg->td->throtl_slice * nr_slices; rw 854 block/blk-throttle.c io_trim = (tg_iops_limit(tg, rw) * tg->td->throtl_slice * nr_slices) / rw 860 block/blk-throttle.c if (tg->bytes_disp[rw] >= bytes_trim) rw 861 block/blk-throttle.c tg->bytes_disp[rw] -= bytes_trim; rw 863 block/blk-throttle.c tg->bytes_disp[rw] = 0; rw 865 block/blk-throttle.c if (tg->io_disp[rw] >= io_trim) rw 866 block/blk-throttle.c tg->io_disp[rw] -= io_trim; rw 868 block/blk-throttle.c tg->io_disp[rw] = 0; rw 870 block/blk-throttle.c tg->slice_start[rw] += nr_slices * tg->td->throtl_slice; rw 874 block/blk-throttle.c rw == READ ? 'R' : 'W', nr_slices, bytes_trim, io_trim, rw 875 block/blk-throttle.c tg->slice_start[rw], tg->slice_end[rw], jiffies); rw 881 block/blk-throttle.c bool rw = bio_data_dir(bio); rw 886 block/blk-throttle.c jiffy_elapsed = jiffies - tg->slice_start[rw]; rw 898 block/blk-throttle.c tmp = (u64)tg_iops_limit(tg, rw) * jiffy_elapsed_rnd; rw 906 block/blk-throttle.c if (tg->io_disp[rw] + 1 <= io_allowed) { rw 923 block/blk-throttle.c bool rw = bio_data_dir(bio); rw 928 block/blk-throttle.c jiffy_elapsed = jiffy_elapsed_rnd = jiffies - tg->slice_start[rw]; rw 936 block/blk-throttle.c tmp = tg_bps_limit(tg, rw) * jiffy_elapsed_rnd; rw 940 block/blk-throttle.c if (tg->bytes_disp[rw] + bio_size <= bytes_allowed) { rw 947 block/blk-throttle.c extra_bytes = tg->bytes_disp[rw] + bio_size - bytes_allowed; rw 948 block/blk-throttle.c jiffy_wait = div64_u64(extra_bytes * HZ, tg_bps_limit(tg, rw)); rw 970 block/blk-throttle.c bool rw = bio_data_dir(bio); rw 979 block/blk-throttle.c BUG_ON(tg->service_queue.nr_queued[rw] && rw 980 block/blk-throttle.c bio != throtl_peek_queued(&tg->service_queue.queued[rw])); rw 983 block/blk-throttle.c if (tg_bps_limit(tg, rw) == U64_MAX && rw 984 block/blk-throttle.c tg_iops_limit(tg, rw) == UINT_MAX) { rw 997 block/blk-throttle.c if (throtl_slice_used(tg, rw) && !(tg->service_queue.nr_queued[rw])) rw 998 block/blk-throttle.c throtl_start_new_slice(tg, rw); rw 1000 block/blk-throttle.c if (time_before(tg->slice_end[rw], rw 1002 block/blk-throttle.c throtl_extend_slice(tg, rw, rw 1018 block/blk-throttle.c if (time_before(tg->slice_end[rw], jiffies + max_wait)) rw 1019 block/blk-throttle.c throtl_extend_slice(tg, rw, jiffies + max_wait); rw 1026 block/blk-throttle.c bool rw = bio_data_dir(bio); rw 1030 block/blk-throttle.c tg->bytes_disp[rw] += bio_size; rw 1031 block/blk-throttle.c tg->io_disp[rw]++; rw 1032 block/blk-throttle.c tg->last_bytes_disp[rw] += bio_size; rw 1033 block/blk-throttle.c tg->last_io_disp[rw]++; rw 1058 block/blk-throttle.c bool rw = bio_data_dir(bio); rw 1061 block/blk-throttle.c qn = &tg->qnode_on_self[rw]; rw 1069 block/blk-throttle.c if (!sq->nr_queued[rw]) rw 1072 block/blk-throttle.c throtl_qnode_add_bio(bio, qn, &sq->queued[rw]); rw 1074 block/blk-throttle.c sq->nr_queued[rw]++; rw 1105 block/blk-throttle.c struct throtl_grp *parent_tg, bool rw) rw 1107 block/blk-throttle.c if (throtl_slice_used(parent_tg, rw)) { rw 1108 block/blk-throttle.c throtl_start_new_slice_with_credit(parent_tg, rw, rw 1109 block/blk-throttle.c child_tg->slice_start[rw]); rw 1114 block/blk-throttle.c static void tg_dispatch_one_bio(struct throtl_grp *tg, bool rw) rw 1128 block/blk-throttle.c bio = throtl_pop_queued(&sq->queued[rw], &tg_to_put); rw 1129 block/blk-throttle.c sq->nr_queued[rw]--; rw 1141 block/blk-throttle.c throtl_add_bio_tg(bio, &tg->qnode_on_parent[rw], parent_tg); rw 1142 block/blk-throttle.c start_parent_slice_with_credit(tg, parent_tg, rw); rw 1144 block/blk-throttle.c throtl_qnode_add_bio(bio, &tg->qnode_on_parent[rw], rw 1145 block/blk-throttle.c &parent_sq->queued[rw]); rw 1146 block/blk-throttle.c BUG_ON(tg->td->nr_queued[rw] <= 0); rw 1147 block/blk-throttle.c tg->td->nr_queued[rw]--; rw 1150 block/blk-throttle.c throtl_trim_slice(tg, rw); rw 1312 block/blk-throttle.c int rw; rw 1317 block/blk-throttle.c for (rw = READ; rw <= WRITE; rw++) rw 1318 block/blk-throttle.c while ((bio = throtl_pop_queued(&td_sq->queued[rw], NULL))) rw 2038 block/blk-throttle.c int i, cpu, rw; rw 2049 block/blk-throttle.c for (rw = READ; rw <= WRITE; rw++) { rw 2051 block/blk-throttle.c struct latency_bucket *tmp = &td->tmp_buckets[rw][i]; rw 2057 block/blk-throttle.c bucket = per_cpu_ptr(td->latency_buckets[rw], rw 2068 block/blk-throttle.c latency[rw] = tmp->total_latency; rw 2072 block/blk-throttle.c latency[rw] /= samples; rw 2073 block/blk-throttle.c if (latency[rw] == 0) rw 2075 block/blk-throttle.c avg_latency[rw][i].latency = latency[rw]; rw 2080 block/blk-throttle.c for (rw = READ; rw <= WRITE; rw++) { rw 2082 block/blk-throttle.c if (!avg_latency[rw][i].latency) { rw 2083 block/blk-throttle.c if (td->avg_buckets[rw][i].latency < last_latency[rw]) rw 2084 block/blk-throttle.c td->avg_buckets[rw][i].latency = rw 2085 block/blk-throttle.c last_latency[rw]; rw 2089 block/blk-throttle.c if (!td->avg_buckets[rw][i].valid) rw 2090 block/blk-throttle.c latency[rw] = avg_latency[rw][i].latency; rw 2092 block/blk-throttle.c latency[rw] = (td->avg_buckets[rw][i].latency * 7 + rw 2093 block/blk-throttle.c avg_latency[rw][i].latency) >> 3; rw 2095 block/blk-throttle.c td->avg_buckets[rw][i].latency = max(latency[rw], rw 2096 block/blk-throttle.c last_latency[rw]); rw 2097 block/blk-throttle.c td->avg_buckets[rw][i].valid = true; rw 2098 block/blk-throttle.c last_latency[rw] = td->avg_buckets[rw][i].latency; rw 2123 block/blk-throttle.c bool rw = bio_data_dir(bio); rw 2130 block/blk-throttle.c if (bio_flagged(bio, BIO_THROTTLED) || !tg->has_rules[rw]) rw 2143 block/blk-throttle.c if (tg->last_low_overflow_time[rw] == 0) rw 2144 block/blk-throttle.c tg->last_low_overflow_time[rw] = jiffies; rw 2148 block/blk-throttle.c if (sq->nr_queued[rw]) rw 2153 block/blk-throttle.c tg->last_low_overflow_time[rw] = jiffies; rw 2175 block/blk-throttle.c throtl_trim_slice(tg, rw); rw 2182 block/blk-throttle.c qn = &tg->qnode_on_parent[rw]; rw 2191 block/blk-throttle.c rw == READ ? 'R' : 'W', rw 2192 block/blk-throttle.c tg->bytes_disp[rw], bio->bi_iter.bi_size, rw 2193 block/blk-throttle.c tg_bps_limit(tg, rw), rw 2194 block/blk-throttle.c tg->io_disp[rw], tg_iops_limit(tg, rw), rw 2197 block/blk-throttle.c tg->last_low_overflow_time[rw] = jiffies; rw 2199 block/blk-throttle.c td->nr_queued[rw]++; rw 2263 block/blk-throttle.c int rw = bio_data_dir(bio); rw 2289 block/blk-throttle.c threshold = tg->td->avg_buckets[rw][bucket].latency + rw 2343 block/blk-throttle.c int rw; rw 2363 block/blk-throttle.c for (rw = READ; rw <= WRITE; rw++) rw 2364 block/blk-throttle.c while ((bio = throtl_pop_queued(&td->service_queue.queued[rw], rw 458 block/blk-wbt.c static inline unsigned int get_limit(struct rq_wb *rwb, unsigned long rw) rw 469 block/blk-wbt.c if ((rw & REQ_OP_MASK) == REQ_OP_DISCARD) rw 480 block/blk-wbt.c if ((rw & REQ_HIPRIO) || wb_recent_wait(rwb) || current_is_kswapd()) rw 482 block/blk-wbt.c else if ((rw & REQ_BACKGROUND) || close_io(rwb)) { rw 497 block/blk-wbt.c unsigned long rw; rw 503 block/blk-wbt.c return rq_wait_inc_below(rqw, get_limit(data->rwb, data->rw)); rw 517 block/blk-wbt.c unsigned long rw) rw 523 block/blk-wbt.c .rw = rw, rw 290 block/bounce.c int rw = bio_data_dir(*bio_orig); rw 330 block/bounce.c if (rw == WRITE) { rw 348 block/bounce.c if (rw == READ) rw 352 block/bounce.c if (rw == READ) rw 49 block/genhd.c void part_inc_in_flight(struct request_queue *q, struct hd_struct *part, int rw) rw 54 block/genhd.c part_stat_local_inc(part, in_flight[rw]); rw 56 block/genhd.c part_stat_local_inc(&part_to_disk(part)->part0, in_flight[rw]); rw 59 block/genhd.c void part_dec_in_flight(struct request_queue *q, struct hd_struct *part, int rw) rw 64 block/genhd.c part_stat_local_dec(part, in_flight[rw]); rw 66 block/genhd.c part_stat_local_dec(&part_to_disk(part)->part0, in_flight[rw]); rw 2437 drivers/acpi/nfit/core.c resource_size_t dpa, void *iobuf, size_t len, int rw, rw 2447 drivers/acpi/nfit/core.c write_blk_ctl(nfit_blk, lane, dpa, len, rw); rw 2464 drivers/acpi/nfit/core.c if (rw) rw 2478 drivers/acpi/nfit/core.c if (rw) rw 2486 drivers/acpi/nfit/core.c resource_size_t dpa, void *iobuf, u64 len, int rw) rw 2499 drivers/acpi/nfit/core.c iobuf + copied, c, rw, lane); rw 245 drivers/acpi/nfit/nfit.h void *iobuf, u64 len, int rw); rw 543 drivers/ata/libata-sff.c unsigned int buflen, int rw) rw 550 drivers/ata/libata-sff.c if (rw == READ) rw 566 drivers/ata/libata-sff.c if (rw == READ) { rw 598 drivers/ata/libata-sff.c unsigned int buflen, int rw) rw 607 drivers/ata/libata-sff.c return ata_sff_data_xfer(qc, buf, buflen, rw); rw 610 drivers/ata/libata-sff.c if (rw == READ) rw 626 drivers/ata/libata-sff.c if (rw == READ) { rw 777 drivers/ata/libata-sff.c int rw = (qc->tf.flags & ATA_TFLAG_WRITE) ? WRITE : READ; rw 812 drivers/ata/libata-sff.c consumed = ap->ops->sff_data_xfer(qc, buf + offset, count, rw); rw 2947 drivers/ata/libata-sff.c unsigned int rw = (qc->tf.flags & ATA_TFLAG_WRITE); rw 2957 drivers/ata/libata-sff.c if (!rw) rw 381 drivers/ata/pata_arasan_cf.c u32 rw = acdev->qc->tf.flags & ATA_TFLAG_WRITE; rw 383 drivers/ata/pata_arasan_cf.c dev_err(acdev->host->dev, "%s TimeOut", rw ? "write" : "read"); rw 67 drivers/ata/pata_buddha.c unsigned int buflen, int rw) rw 75 drivers/ata/pata_buddha.c if (rw == READ) rw 87 drivers/ata/pata_buddha.c if (rw == READ) { rw 480 drivers/ata/pata_ep93xx.c unsigned int buflen, int rw) rw 489 drivers/ata/pata_ep93xx.c if (rw == READ) rw 503 drivers/ata/pata_ep93xx.c if (rw == READ) { rw 45 drivers/ata/pata_falcon.c unsigned int buflen, int rw) rw 59 drivers/ata/pata_falcon.c if (rw == READ) { rw 78 drivers/ata/pata_falcon.c if (rw == READ) { rw 45 drivers/ata/pata_gayle.c unsigned int buflen, int rw) rw 53 drivers/ata/pata_gayle.c if (rw == READ) rw 65 drivers/ata/pata_gayle.c if (rw == READ) { rw 40 drivers/ata/pata_ixp4xx_cf.c unsigned char *buf, unsigned int buflen, int rw) rw 56 drivers/ata/pata_ixp4xx_cf.c if (rw == READ) rw 68 drivers/ata/pata_ixp4xx_cf.c if (rw == READ) { rw 293 drivers/ata/pata_legacy.c unsigned char *buf, unsigned int buflen, int rw) rw 312 drivers/ata/pata_legacy.c if (rw == READ) rw 319 drivers/ata/pata_legacy.c if (rw == READ) { rw 330 drivers/ata/pata_legacy.c buflen = ata_sff_data_xfer32(qc, buf, buflen, rw); rw 694 drivers/ata/pata_legacy.c unsigned int buflen, int rw) rw 702 drivers/ata/pata_legacy.c if (rw == WRITE) rw 709 drivers/ata/pata_legacy.c if (rw == WRITE) { rw 719 drivers/ata/pata_legacy.c return ata_sff_data_xfer(qc, buf, buflen, rw); rw 124 drivers/ata/pata_ns87415.c unsigned int rw = (qc->tf.flags & ATA_TFLAG_WRITE); rw 137 drivers/ata/pata_ns87415.c if (!rw) rw 296 drivers/ata/pata_octeon_cf.c int rw) rw 304 drivers/ata/pata_octeon_cf.c if (rw) { rw 335 drivers/ata/pata_octeon_cf.c int rw) rw 343 drivers/ata/pata_octeon_cf.c if (rw) { rw 367 drivers/ata/pata_octeon_cf.c if (rw == READ) { rw 588 drivers/ata/pata_octeon_cf.c mio_boot_dma_cfg.s.rw = ((qc->tf.flags & ATA_TFLAG_WRITE) != 0); rw 609 drivers/ata/pata_octeon_cf.c (mio_boot_dma_cfg.s.rw) ? "write" : "read", sg->length, rw 92 drivers/ata/pata_pcmcia.c unsigned char *buf, unsigned int buflen, int rw) rw 96 drivers/ata/pata_pcmcia.c if (rw == READ) rw 265 drivers/ata/pata_samsung_cf.c unsigned char *buf, unsigned int buflen, int rw) rw 274 drivers/ata/pata_samsung_cf.c if (rw == READ) rw 1913 drivers/ata/sata_mv.c unsigned int rw = (qc->tf.flags & ATA_TFLAG_WRITE); rw 1914 drivers/ata/sata_mv.c u32 cmd = (rw ? 0 : ATA_DMA_WR) | ATA_DMA_START; rw 2154 drivers/ata/sata_nv.c unsigned int rw; rw 2168 drivers/ata/sata_nv.c rw = qc->tf.flags & ATA_TFLAG_WRITE; rw 2177 drivers/ata/sata_nv.c if (!rw) rw 451 drivers/ata/sata_rcar.c unsigned int buflen, int rw) rw 458 drivers/ata/sata_rcar.c if (rw == READ) rw 474 drivers/ata/sata_rcar.c if (rw == READ) { rw 564 drivers/ata/sata_rcar.c unsigned int rw = qc->tf.flags & ATA_TFLAG_WRITE; rw 580 drivers/ata/sata_rcar.c if (!rw) rw 278 drivers/ata/sata_sil.c unsigned int rw = (qc->tf.flags & ATA_TFLAG_WRITE); rw 287 drivers/ata/sata_sil.c if (!rw) rw 233 drivers/ata/sata_svw.c unsigned int rw = (qc->tf.flags & ATA_TFLAG_WRITE); rw 244 drivers/ata/sata_svw.c if (!rw) rw 627 drivers/ata/sata_sx4.c unsigned int rw = (qc->tf.flags & ATA_TFLAG_WRITE); rw 641 drivers/ata/sata_sx4.c if (rw && qc->tf.protocol == ATA_PROT_DMA) { rw 212 drivers/auxdisplay/panel.c int rw; rw 946 drivers/auxdisplay/panel.c lcd.pins.rw = PIN_INITP; rw 992 drivers/auxdisplay/panel.c lcd.pins.rw = lcd_rw_pin; rw 1025 drivers/auxdisplay/panel.c if (lcd.pins.rw == PIN_NOT_SET) rw 1026 drivers/auxdisplay/panel.c lcd.pins.rw = DEFAULT_LCD_PIN_RW; rw 1038 drivers/auxdisplay/panel.c if (lcd.pins.rw == PIN_NOT_SET) rw 1039 drivers/auxdisplay/panel.c lcd.pins.rw = PIN_NONE; rw 1059 drivers/auxdisplay/panel.c pin_to_bits(lcd.pins.rw, lcd_bits[LCD_PORT_D][LCD_BIT_RW], rw 1741 drivers/auxdisplay/panel.c lcd.pins.rw = lcd_rw_pin; rw 1211 drivers/block/drbd/drbd_req.c const int rw = bio_data_dir(bio); rw 1240 drivers/block/drbd/drbd_req.c if (rw == WRITE && req->private_bio && req->i.size rw 1325 drivers/block/drbd/drbd_req.c const int rw = bio_data_dir(req->master_bio); rw 1331 drivers/block/drbd/drbd_req.c if (rw == WRITE) { rw 1358 drivers/block/drbd/drbd_req.c if (rw != WRITE) { rw 1369 drivers/block/drbd/drbd_req.c if (rw == WRITE) rw 1375 drivers/block/drbd/drbd_req.c if (rw == WRITE) { rw 1404 drivers/block/drbd/drbd_req.c &device->pending_master_completion[rw == WRITE]); rw 1409 drivers/block/drbd/drbd_req.c &device->pending_completion[rw == WRITE]); rw 1453 drivers/block/drbd/drbd_req.c const int rw = bio_data_dir(req->master_bio); rw 1455 drivers/block/drbd/drbd_req.c if (rw == WRITE /* rw != WRITE should not even end up here! */ rw 514 drivers/block/loop.c loff_t pos, bool rw) rw 561 drivers/block/loop.c iov_iter_bvec(&iter, rw, bvec, nr_bvec, blk_rq_bytes(rq)); rw 572 drivers/block/loop.c if (rw == WRITE) rw 1277 drivers/block/xen-blkback/blkback.c req->u.indirect.nr_segments : req->u.rw.nr_segments; rw 1292 drivers/block/xen-blkback/blkback.c pending_req->id = req->u.rw.id; rw 1298 drivers/block/xen-blkback/blkback.c preq.dev = req->u.rw.handle; rw 1299 drivers/block/xen-blkback/blkback.c preq.sector_number = req->u.rw.sector_number; rw 1301 drivers/block/xen-blkback/blkback.c pages[i]->gref = req->u.rw.seg[i].gref; rw 1302 drivers/block/xen-blkback/blkback.c seg[i].nsec = req->u.rw.seg[i].last_sect - rw 1303 drivers/block/xen-blkback/blkback.c req->u.rw.seg[i].first_sect + 1; rw 1304 drivers/block/xen-blkback/blkback.c seg[i].offset = (req->u.rw.seg[i].first_sect << 9); rw 1305 drivers/block/xen-blkback/blkback.c if ((req->u.rw.seg[i].last_sect >= (XEN_PAGE_SIZE >> 9)) || rw 1306 drivers/block/xen-blkback/blkback.c (req->u.rw.seg[i].last_sect < rw 1307 drivers/block/xen-blkback/blkback.c req->u.rw.seg[i].first_sect)) rw 1421 drivers/block/xen-blkback/blkback.c make_response(ring, req->u.rw.id, req_operation, BLKIF_RSP_ERROR); rw 124 drivers/block/xen-blkback/common.h struct blkif_x86_32_request_rw rw; rw 180 drivers/block/xen-blkback/common.h struct blkif_x86_64_request_rw rw; rw 405 drivers/block/xen-blkback/common.h dst->u.rw.nr_segments = src->u.rw.nr_segments; rw 406 drivers/block/xen-blkback/common.h dst->u.rw.handle = src->u.rw.handle; rw 407 drivers/block/xen-blkback/common.h dst->u.rw.id = src->u.rw.id; rw 408 drivers/block/xen-blkback/common.h dst->u.rw.sector_number = src->u.rw.sector_number; rw 410 drivers/block/xen-blkback/common.h if (n > dst->u.rw.nr_segments) rw 411 drivers/block/xen-blkback/common.h n = dst->u.rw.nr_segments; rw 413 drivers/block/xen-blkback/common.h dst->u.rw.seg[i] = src->u.rw.seg[i]; rw 453 drivers/block/xen-blkback/common.h dst->u.rw.nr_segments = src->u.rw.nr_segments; rw 454 drivers/block/xen-blkback/common.h dst->u.rw.handle = src->u.rw.handle; rw 455 drivers/block/xen-blkback/common.h dst->u.rw.id = src->u.rw.id; rw 456 drivers/block/xen-blkback/common.h dst->u.rw.sector_number = src->u.rw.sector_number; rw 458 drivers/block/xen-blkback/common.h if (n > dst->u.rw.nr_segments) rw 459 drivers/block/xen-blkback/common.h n = dst->u.rw.nr_segments; rw 461 drivers/block/xen-blkback/common.h dst->u.rw.seg[i] = src->u.rw.seg[i]; rw 271 drivers/block/xen-blkfront.c rinfo->shadow_free = rinfo->shadow[free].req.u.rw.id; rw 272 drivers/block/xen-blkfront.c rinfo->shadow[free].req.u.rw.id = 0x0fffffee; /* debug */ rw 279 drivers/block/xen-blkfront.c if (rinfo->shadow[id].req.u.rw.id != id) rw 283 drivers/block/xen-blkfront.c rinfo->shadow[id].req.u.rw.id = rinfo->shadow_free; rw 539 drivers/block/xen-blkfront.c (*ring_req)->u.rw.id = id; rw 657 drivers/block/xen-blkfront.c ring_req->u.rw.seg[grant_idx] = rw 676 drivers/block/xen-blkfront.c uint16_t nr_segments = first->u.rw.nr_segments; rw 682 drivers/block/xen-blkfront.c first->u.rw.nr_segments = BLKIF_MAX_SEGMENTS_PER_REQUEST; rw 684 drivers/block/xen-blkfront.c second->u.rw.nr_segments = nr_segments - BLKIF_MAX_SEGMENTS_PER_REQUEST; rw 685 drivers/block/xen-blkfront.c second->u.rw.sector_number = first->u.rw.sector_number + rw 688 drivers/block/xen-blkfront.c second->u.rw.handle = first->u.rw.handle; rw 767 drivers/block/xen-blkfront.c ring_req->u.rw.sector_number = (blkif_sector_t)blk_rq_pos(req); rw 768 drivers/block/xen-blkfront.c ring_req->u.rw.handle = info->handle; rw 788 drivers/block/xen-blkfront.c ring_req->u.rw.nr_segments = num_grant; rw 1290 drivers/block/xen-blkfront.c rinfo->shadow[i].req.u.rw.nr_segments; rw 1433 drivers/block/xen-blkfront.c s->req.u.indirect.nr_segments : s->req.u.rw.nr_segments; rw 1453 drivers/block/xen-blkfront.c num_grant += s2->req.u.rw.nr_segments; rw 1625 drivers/block/xen-blkfront.c rinfo->shadow[id].req.u.rw.nr_segments == 0)) { rw 1880 drivers/block/xen-blkfront.c rinfo->shadow[j].req.u.rw.id = j + 1; rw 1881 drivers/block/xen-blkfront.c rinfo->shadow[BLK_RING_SIZE(info)-1].req.u.rw.id = 0x0fffffff; rw 2799 drivers/edac/amd64_edac.c u8 rw; rw 2804 drivers/edac/amd64_edac.c rw = dram_rw(pvt, range); rw 2805 drivers/edac/amd64_edac.c if (!rw) rw 2815 drivers/edac/amd64_edac.c (rw & 0x1) ? "R" : "-", rw 2816 drivers/edac/amd64_edac.c (rw & 0x2) ? "W" : "-", rw 743 drivers/gpu/drm/amd/powerplay/inc/amdgpu_smu.h #define smu_thermal_temperature_range_update(smu, range, rw) \ rw 744 drivers/gpu/drm/amd/powerplay/inc/amdgpu_smu.h ((smu)->ppt_funcs->thermal_temperature_range_update? (smu)->ppt_funcs->thermal_temperature_range_update((smu), (range), (rw)) : 0) rw 231 drivers/gpu/drm/amd/powerplay/inc/hwmgr.h int (*smc_table_manager)(struct pp_hwmgr *hwmgr, uint8_t *table, uint16_t table_id, bool rw); /*rw: true for read, false for write */ rw 115 drivers/gpu/drm/amd/powerplay/inc/smumgr.h extern int smum_smc_table_manager(struct pp_hwmgr *hwmgr, uint8_t *table, uint16_t table_id, bool rw); rw 286 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c static int smu10_smc_table_manager(struct pp_hwmgr *hwmgr, uint8_t *table, uint16_t table_id, bool rw) rw 290 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c if (rw) rw 213 drivers/gpu/drm/amd/powerplay/smumgr/smumgr.c int smum_smc_table_manager(struct pp_hwmgr *hwmgr, uint8_t *table, uint16_t table_id, bool rw) rw 216 drivers/gpu/drm/amd/powerplay/smumgr/smumgr.c return hwmgr->smumgr_funcs->smc_table_manager(hwmgr, table, table_id, rw); rw 342 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c uint16_t table_id, bool rw) rw 346 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c if (rw) rw 380 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c uint16_t table_id, bool rw) rw 384 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c if (rw) rw 590 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c uint16_t table_id, bool rw) rw 594 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c if (rw) rw 66 drivers/gpu/drm/i915/gvt/kvmgt.c size_t (*rw)(struct intel_vgpu *vgpu, char *buf, rw 433 drivers/gpu/drm/i915/gvt/kvmgt.c .rw = intel_vgpu_reg_rw_opregion, rw 538 drivers/gpu/drm/i915/gvt/kvmgt.c .rw = intel_vgpu_reg_rw_edid, rw 974 drivers/gpu/drm/i915/gvt/kvmgt.c return vgpu->vdev.region[index].ops->rw(vgpu, buf, count, rw 13 drivers/gpu/drm/nouveau/nvkm/subdev/bios/priv.h bool rw; rw 90 drivers/gpu/drm/nouveau/nvkm/subdev/bios/shadow.c if (mthd->func->rw) rw 163 drivers/gpu/drm/nouveau/nvkm/subdev/bios/shadow.c .rw = false, rw 101 drivers/gpu/drm/nouveau/nvkm/subdev/bios/shadowacpi.c .rw = false, rw 110 drivers/gpu/drm/nouveau/nvkm/subdev/bios/shadowacpi.c .rw = false, rw 76 drivers/gpu/drm/nouveau/nvkm/subdev/bios/shadowof.c .rw = false, rw 88 drivers/gpu/drm/nouveau/nvkm/subdev/bios/shadowpci.c .rw = true, rw 120 drivers/gpu/drm/nouveau/nvkm/subdev/bios/shadowpci.c .rw = true, rw 119 drivers/gpu/drm/nouveau/nvkm/subdev/bios/shadowramin.c .rw = true, rw 63 drivers/gpu/drm/nouveau/nvkm/subdev/bios/shadowrom.c .rw = false, rw 49 drivers/gpu/drm/vmwgfx/ttm_lock.c lock->rw = 0; rw 56 drivers/gpu/drm/vmwgfx/ttm_lock.c if (--lock->rw == 0) rw 66 drivers/gpu/drm/vmwgfx/ttm_lock.c if (lock->rw >= 0 && lock->flags == 0) { rw 67 drivers/gpu/drm/vmwgfx/ttm_lock.c ++lock->rw; rw 93 drivers/gpu/drm/vmwgfx/ttm_lock.c if (lock->rw >= 0 && lock->flags == 0) { rw 94 drivers/gpu/drm/vmwgfx/ttm_lock.c ++lock->rw; rw 127 drivers/gpu/drm/vmwgfx/ttm_lock.c lock->rw = 0; rw 137 drivers/gpu/drm/vmwgfx/ttm_lock.c if (lock->rw == 0 && ((lock->flags & ~TTM_WRITE_LOCK_PENDING) == 0)) { rw 138 drivers/gpu/drm/vmwgfx/ttm_lock.c lock->rw = -1; rw 180 drivers/gpu/drm/vmwgfx/ttm_lock.c if (lock->rw == 0) { rw 72 drivers/gpu/drm/vmwgfx/ttm_lock.h int32_t rw; rw 311 drivers/hid/wacom_wac.c int rw = 0; rw 370 drivers/hid/wacom_wac.c rw = (data[7] & 0x04) - (data[7] & 0x03); rw 373 drivers/hid/wacom_wac.c rw = 44 - (data[6] >> 2); rw 374 drivers/hid/wacom_wac.c rw = clamp_val(rw, 0, 31); rw 375 drivers/hid/wacom_wac.c input_report_abs(input, ABS_DISTANCE, rw); rw 380 drivers/hid/wacom_wac.c rw = (data[6] & 0x01) ? -1 : rw 383 drivers/hid/wacom_wac.c rw = 0; rw 387 drivers/hid/wacom_wac.c rw = -(signed char)data[6]; rw 389 drivers/hid/wacom_wac.c input_report_rel(input, REL_WHEEL, rw); rw 407 drivers/hid/wacom_wac.c rw = ((data[7] & 0x18) >> 3) - ((data[7] & 0x20) >> 3); rw 408 drivers/hid/wacom_wac.c input_report_rel(pad_input, REL_WHEEL, rw); rw 447 drivers/hid/wacom_wac.c rw = (data[7] >> 2 & 0x07); rw 448 drivers/hid/wacom_wac.c battery_capacity = batcap_gr[rw]; rw 449 drivers/hid/wacom_wac.c ps_connected = rw == 7; rw 910 drivers/hwmon/ibmaem.c const struct aem_rw_sensor_template *rw) rw 934 drivers/hwmon/ibmaem.c while (rw->label) { rw 936 drivers/hwmon/ibmaem.c sensors->dev_attr.attr.name = rw->label; rw 938 drivers/hwmon/ibmaem.c sensors->dev_attr.show = rw->show; rw 939 drivers/hwmon/ibmaem.c sensors->dev_attr.store = rw->set; rw 940 drivers/hwmon/ibmaem.c sensors->index = rw->index; rw 948 drivers/hwmon/ibmaem.c rw++; rw 187 drivers/i2c/busses/i2c-ali1563.c union i2c_smbus_data *data, u8 rw) rw 195 drivers/i2c/busses/i2c-ali1563.c if (rw == I2C_SMBUS_WRITE) { rw 209 drivers/i2c/busses/i2c-ali1563.c if (rw == I2C_SMBUS_WRITE) { rw 234 drivers/i2c/busses/i2c-ali1563.c unsigned short flags, char rw, u8 cmd, rw 273 drivers/i2c/busses/i2c-ali1563.c outb_p(((addr & 0x7f) << 1) | (rw & 0x01), SMB_HST_ADD); rw 281 drivers/i2c/busses/i2c-ali1563.c if (rw == I2C_SMBUS_WRITE) rw 287 drivers/i2c/busses/i2c-ali1563.c if (rw == I2C_SMBUS_WRITE) rw 292 drivers/i2c/busses/i2c-ali1563.c if (rw == I2C_SMBUS_WRITE) { rw 299 drivers/i2c/busses/i2c-ali1563.c error = ali1563_block(a, data, rw); rw 307 drivers/i2c/busses/i2c-ali1563.c if ((rw == I2C_SMBUS_WRITE) || (size == HST_CNTL2_QUICK)) rw 114 drivers/i2c/busses/i2c-cbus-gpio.c static int cbus_transfer(struct cbus_host *host, char rw, unsigned dev, rw 133 drivers/i2c/busses/i2c-cbus-gpio.c cbus_send_bit(host, rw == I2C_SMBUS_READ); rw 138 drivers/i2c/busses/i2c-cbus-gpio.c if (rw == I2C_SMBUS_WRITE) { rw 104 drivers/i2c/busses/i2c-ismt.c #define ISMT_DESC_ADDR_RW(addr, rw) (((addr) << 1) | (rw)) rw 203 drivers/i2c/busses/i2c-sprd.c static void sprd_i2c_opt_mode(struct sprd_i2c *i2c_dev, int rw) rw 207 drivers/i2c/busses/i2c-sprd.c writel(cmd | rw << 3, i2c_dev->base + I2C_CTL); rw 270 drivers/i2c/busses/scx200_acb.c char rw, u8 command, int size, rw 287 drivers/i2c/busses/scx200_acb.c buffer = rw ? &data->byte : &command; rw 314 drivers/i2c/busses/scx200_acb.c size, address, command, len, rw); rw 316 drivers/i2c/busses/scx200_acb.c if (!len && rw == I2C_SMBUS_READ) { rw 323 drivers/i2c/busses/scx200_acb.c iface->address_byte = (address << 1) | rw; rw 347 drivers/i2c/busses/scx200_acb.c if (rc == 0 && size == I2C_SMBUS_WORD_DATA && rw == I2C_SMBUS_READ) rw 371 drivers/ide/ide-atapi.c int ireason, int rw) rw 375 drivers/ide/ide-atapi.c debug_log("ireason: 0x%x, rw: 0x%x\n", ireason, rw); rw 377 drivers/ide/ide-atapi.c if (ireason == (!rw << 1)) rw 379 drivers/ide/ide-atapi.c else if (ireason == (rw << 1)) { rw 384 drivers/ide/ide-atapi.c ide_pad_transfer(drive, rw, len); rw 385 drivers/ide/ide-atapi.c } else if (!rw && ireason == ATAPI_COD) { rw 192 drivers/ide/ide-dma-sff.c u8 rw = (cmd->tf_flags & IDE_TFLAG_WRITE) ? 0 : ATA_DMA_WR; rw 210 drivers/ide/ide-dma-sff.c writeb(rw, (void __iomem *)(hwif->dma_base + ATA_DMA_CMD)); rw 212 drivers/ide/ide-dma-sff.c outb(rw, hwif->dma_base + ATA_DMA_CMD); rw 195 drivers/ide/trm290.c unsigned int count, rw = (cmd->tf_flags & IDE_TFLAG_WRITE) ? 1 : 2; rw 202 drivers/ide/trm290.c outl(hwif->dmatable_dma | rw, hwif->dma_base); rw 291 drivers/ide/tx4939ide.c u8 rw = (cmd->tf_flags & IDE_TFLAG_WRITE) ? 0 : ATA_DMA_WR; rw 301 drivers/ide/tx4939ide.c tx4939ide_writeb(rw, base, TX4939IDE_DMA_Cmd); rw 706 drivers/infiniband/hw/mlx5/mr.c MLX5_SET(mkc, mkc, rw, !!(acc & IB_ACCESS_REMOTE_WRITE)); rw 1093 drivers/infiniband/hw/mlx5/mr.c MLX5_SET(mkc, mkc, rw, !!(access_flags & IB_ACCESS_REMOTE_WRITE)); rw 1173 drivers/infiniband/hw/mlx5/mr.c MLX5_SET(mkc, mkc, rw, !!(acc & IB_ACCESS_REMOTE_WRITE)); rw 2419 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c struct ocrdma_sge *rw; rw 2432 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c rw = (struct ocrdma_sge *)(hdr + 1); rw 2434 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c ibwc->byte_len = rw->len; rw 1491 drivers/infiniband/ulp/isert/ib_isert.c if (!cmd->rw.nr_ops) rw 1495 drivers/infiniband/ulp/isert/ib_isert.c rdma_rw_ctx_destroy_signature(&cmd->rw, conn->qp, rw 1500 drivers/infiniband/ulp/isert/ib_isert.c rdma_rw_ctx_destroy(&cmd->rw, conn->qp, conn->cm_id->port_num, rw 1504 drivers/infiniband/ulp/isert/ib_isert.c cmd->rw.nr_ops = 0; rw 1672 drivers/infiniband/ulp/isert/ib_isert.c ret = isert_check_pi_status(cmd, isert_cmd->rw.reg->mr); rw 1718 drivers/infiniband/ulp/isert/ib_isert.c ret = isert_check_pi_status(se_cmd, isert_cmd->rw.reg->mr); rw 2149 drivers/infiniband/ulp/isert/ib_isert.c ret = rdma_rw_ctx_signature_init(&cmd->rw, conn->qp, port_num, rw 2154 drivers/infiniband/ulp/isert/ib_isert.c ret = rdma_rw_ctx_init(&cmd->rw, conn->qp, port_num, rw 2167 drivers/infiniband/ulp/isert/ib_isert.c ret = rdma_rw_ctx_post(&cmd->rw, conn->qp, port_num, cqe, chain_wr); rw 126 drivers/infiniband/ulp/isert/ib_isert.h struct rdma_rw_ctx rw; rw 918 drivers/infiniband/ulp/srpt/ib_srpt.c ret = rdma_rw_ctx_init(&ctx->rw, ch->qp, ch->sport->port, rw 947 drivers/infiniband/ulp/srpt/ib_srpt.c rdma_rw_ctx_destroy(&ctx->rw, ch->qp, ch->sport->port, rw 965 drivers/infiniband/ulp/srpt/ib_srpt.c rdma_rw_ctx_destroy(&ctx->rw, ch->qp, ch->sport->port, rw 2754 drivers/infiniband/ulp/srpt/ib_srpt.c first_wr = rdma_rw_ctx_wrs(&ctx->rw, ch->qp, ch->sport->port, rw 2830 drivers/infiniband/ulp/srpt/ib_srpt.c first_wr = rdma_rw_ctx_wrs(&ctx->rw, ch->qp, rw 195 drivers/infiniband/ulp/srpt/ib_srpt.h struct rdma_rw_ctx rw; rw 32 drivers/input/mouse/cyapa.h #define SMBUS_ENCODE_RW(cmd, rw) ((cmd) | ((rw) & 0x01)) rw 1182 drivers/md/bcache/request.c int rw = bio_data_dir(bio); rw 1229 drivers/md/bcache/request.c if (rw) rw 554 drivers/md/dm-bufio.c static void use_dmio(struct dm_buffer *b, int rw, sector_t sector, rw 559 drivers/md/dm-bufio.c .bi_op = rw, rw 592 drivers/md/dm-bufio.c static void use_bio(struct dm_buffer *b, int rw, sector_t sector, rw 606 drivers/md/dm-bufio.c use_dmio(b, rw, sector, n_sectors, offset); rw 612 drivers/md/dm-bufio.c bio_set_op_attrs(bio, rw, 0); rw 634 drivers/md/dm-bufio.c static void submit_io(struct dm_buffer *b, int rw, void (*end_io)(struct dm_buffer *, blk_status_t)) rw 648 drivers/md/dm-bufio.c if (rw != REQ_OP_WRITE) { rw 667 drivers/md/dm-bufio.c use_bio(b, rw, sector, n_sectors, offset); rw 669 drivers/md/dm-bufio.c use_dmio(b, rw, sector, n_sectors, offset); rw 1421 drivers/md/dm-crypt.c unsigned rw = bio_data_dir(clone); rw 1427 drivers/md/dm-crypt.c if (rw == WRITE) rw 1433 drivers/md/dm-crypt.c if (rw == READ && !error) { rw 355 drivers/md/dm-kcopyd.c int rw; rw 421 drivers/md/dm-kcopyd.c if (job->rw == READ || !test_bit(DM_KCOPYD_WRITE_SEQ, &job->flags)) { rw 523 drivers/md/dm-kcopyd.c if (op_is_write(job->rw)) rw 535 drivers/md/dm-kcopyd.c if (op_is_write(job->rw)) rw 539 drivers/md/dm-kcopyd.c job->rw = WRITE; rw 554 drivers/md/dm-kcopyd.c .bi_op = job->rw, rw 576 drivers/md/dm-kcopyd.c if (job->rw == READ) rw 619 drivers/md/dm-kcopyd.c if (op_is_write(job->rw)) rw 823 drivers/md/dm-kcopyd.c job->rw = READ; rw 832 drivers/md/dm-kcopyd.c job->rw = REQ_OP_WRITE_ZEROES; rw 835 drivers/md/dm-kcopyd.c job->rw = WRITE; rw 121 drivers/md/dm-raid1.c static void queue_bio(struct mirror_set *ms, struct bio *bio, int rw) rw 127 drivers/md/dm-raid1.c bl = (rw == WRITE) ? &ms->writes : &ms->reads; rw 1186 drivers/md/dm-raid1.c int r, rw = bio_data_dir(bio); rw 1195 drivers/md/dm-raid1.c if (rw == WRITE) { rw 1198 drivers/md/dm-raid1.c queue_bio(ms, bio, rw); rw 1213 drivers/md/dm-raid1.c queue_bio(ms, bio, rw); rw 1236 drivers/md/dm-raid1.c int rw = bio_data_dir(bio); rw 1246 drivers/md/dm-raid1.c if (rw == WRITE) { rw 1288 drivers/md/dm-raid1.c queue_bio(ms, bio, rw); rw 2021 drivers/md/dm-thin.c int rw = bio_data_dir(bio); rw 2028 drivers/md/dm-thin.c if (lookup_result.shared && (rw == WRITE) && bio->bi_iter.bi_size) { rw 2043 drivers/md/dm-thin.c if (rw != READ) { rw 1071 drivers/md/dm-writecache.c int rw = bio_data_dir(bio); rw 1081 drivers/md/dm-writecache.c if (rw == READ) { rw 374 drivers/md/md.c const int rw = bio_data_dir(bio); rw 379 drivers/md/md.c if (unlikely(test_bit(MD_BROKEN, &mddev->flags)) && (rw == WRITE)) { rw 390 drivers/md/md.c if (mddev->ro == 1 && unlikely(rw == WRITE)) { rw 1954 drivers/md/raid1.c int sectors, struct page *page, int rw) rw 1956 drivers/md/raid1.c if (sync_page_io(rdev, sector, sectors << 9, page, rw, 0, false)) rw 1959 drivers/md/raid1.c if (rw == WRITE) { rw 2299 drivers/md/raid10.c int sectors, struct page *page, int rw) rw 2305 drivers/md/raid10.c && (rw == READ || test_bit(WriteErrorSeen, &rdev->flags))) rw 2307 drivers/md/raid10.c if (sync_page_io(rdev, sector, sectors << 9, page, rw, 0, false)) rw 2310 drivers/md/raid10.c if (rw == WRITE) { rw 5584 drivers/md/raid5.c const int rw = bio_data_dir(bi); rw 5613 drivers/md/raid5.c if (rw == READ && mddev->degraded == 0 && rw 5712 drivers/md/raid5.c !add_stripe_bio(sh, bi, dd_idx, rw, previous)) { rw 5745 drivers/md/raid5.c if (rw == WRITE) rw 214 drivers/media/dvb-frontends/dib8000.c u16 rw[2]; rw 221 drivers/media/dvb-frontends/dib8000.c rw[0] = __dib8000_read_word(state, reg + 0); rw 222 drivers/media/dvb-frontends/dib8000.c rw[1] = __dib8000_read_word(state, reg + 1); rw 226 drivers/media/dvb-frontends/dib8000.c return ((rw[0] << 16) | (rw[1])); rw 103 drivers/media/pci/cx23885/altera-ci.c int (*fpga_rw) (void *dev, int flag, int data, int rw); rw 27 drivers/media/pci/cx23885/altera-ci.h int (*fpga_rw) (void *dev, int ad_rg, int val, int rw); rw 293 drivers/media/platform/marvell-ccic/cafe-driver.c unsigned short flags, char rw, u8 command, rw 308 drivers/media/platform/marvell-ccic/cafe-driver.c if (rw == I2C_SMBUS_WRITE) rw 310 drivers/media/platform/marvell-ccic/cafe-driver.c else if (rw == I2C_SMBUS_READ) rw 156 drivers/media/usb/uvc/uvc_v4l2.c u16 rw, rh; rw 190 drivers/media/usb/uvc/uvc_v4l2.c rw = fmt->fmt.pix.width; rw 198 drivers/media/usb/uvc/uvc_v4l2.c d = min(w, rw) * min(h, rh); rw 199 drivers/media/usb/uvc/uvc_v4l2.c d = w*h + rw*rh - 2*d; rw 155 drivers/media/v4l2-core/videobuf-dma-sg.c int err, rw = 0; rw 161 drivers/media/v4l2-core/videobuf-dma-sg.c rw = READ; rw 164 drivers/media/v4l2-core/videobuf-dma-sg.c rw = WRITE; rw 180 drivers/media/v4l2-core/videobuf-dma-sg.c if (rw == READ) rw 2121 drivers/mmc/core/mmc_test.c struct mmc_test_multiple_rw *rw) rw 2128 drivers/mmc/core/mmc_test.c if (rw->do_nonblock_req && rw 2134 drivers/mmc/core/mmc_test.c for (i = 0 ; i < rw->len && ret == 0; i++) { rw 2135 drivers/mmc/core/mmc_test.c ret = mmc_test_rw_multiple(test, rw, rw->bs[i], rw->size, 0); rw 2143 drivers/mmc/core/mmc_test.c struct mmc_test_multiple_rw *rw) rw 2148 drivers/mmc/core/mmc_test.c for (i = 0 ; i < rw->len && ret == 0; i++) { rw 2149 drivers/mmc/core/mmc_test.c ret = mmc_test_rw_multiple(test, rw, 512 * 1024, rw->size, rw 2150 drivers/mmc/core/mmc_test.c rw->sg_len[i]); rw 520 drivers/mmc/host/cavium.c int count, rw; rw 527 drivers/mmc/host/cavium.c rw = (data->flags & MMC_DATA_WRITE) ? 1 : 0; rw 529 drivers/mmc/host/cavium.c FIELD_PREP(MIO_EMM_DMA_CFG_RW, rw); rw 542 drivers/mmc/host/cavium.c (rw) ? "W" : "R", sg_dma_len(&data->sg[0]), count); rw 557 drivers/mmc/host/cavium.c int count, i, rw; rw 581 drivers/mmc/host/cavium.c rw = (data->flags & MMC_DATA_WRITE) ? 1 : 0; rw 582 drivers/mmc/host/cavium.c fifo_cmd = FIELD_PREP(MIO_EMM_DMA_FIFO_CMD_RW, rw); rw 599 drivers/mmc/host/cavium.c (rw) ? "W" : "R", sg_dma_len(sg), i, count); rw 1050 drivers/mmc/host/s3cmci.c int rw = (data->flags & MMC_DATA_WRITE) ? 1 : 0; rw 1057 drivers/mmc/host/s3cmci.c host->pio_active = rw ? XFER_WRITE : XFER_READ; rw 1059 drivers/mmc/host/s3cmci.c if (rw) { rw 1072 drivers/mmc/host/s3cmci.c int rw = data->flags & MMC_DATA_WRITE; rw 1086 drivers/mmc/host/s3cmci.c if (!rw) rw 170 drivers/mtd/nand/raw/meson_nand.c struct nand_rw_cmd rw; rw 580 drivers/mtd/nand/raw/meson_nand.c u32 *addrs = nfc->cmdfifo.rw.addrs; rw 588 drivers/mtd/nand/raw/meson_nand.c nfc->cmdfifo.rw.cmd0 = cs | NFC_CMD_CLE | cmd0; rw 616 drivers/mtd/nand/raw/meson_nand.c nfc->cmdfifo.rw.cmd1 = cs | NFC_CMD_CLE | NAND_CMD_READSTART; rw 617 drivers/mtd/nand/raw/meson_nand.c writel(nfc->cmdfifo.rw.cmd1, nfc->reg_base + NFC_REG_CMD); rw 131 drivers/net/dsa/sja1105/sja1105.h sja1105_spi_rw_mode_t rw, u64 reg_addr, rw 134 drivers/net/dsa/sja1105/sja1105.h sja1105_spi_rw_mode_t rw, u64 reg_addr, rw 137 drivers/net/dsa/sja1105/sja1105.h sja1105_spi_rw_mode_t rw, u64 base_addr, rw 70 drivers/net/dsa/sja1105/sja1105_spi.c sja1105_spi_rw_mode_t rw, u64 reg_addr, rw 82 drivers/net/dsa/sja1105/sja1105_spi.c msg.access = rw; rw 84 drivers/net/dsa/sja1105/sja1105_spi.c if (rw == SPI_READ) rw 89 drivers/net/dsa/sja1105/sja1105_spi.c if (rw == SPI_WRITE) rw 97 drivers/net/dsa/sja1105/sja1105_spi.c if (rw == SPI_READ) rw 116 drivers/net/dsa/sja1105/sja1105_spi.c sja1105_spi_rw_mode_t rw, u64 reg_addr, rw 125 drivers/net/dsa/sja1105/sja1105_spi.c if (rw == SPI_WRITE) rw 129 drivers/net/dsa/sja1105/sja1105_spi.c rc = sja1105_spi_send_packed_buf(priv, rw, reg_addr, packed_buf, rw 132 drivers/net/dsa/sja1105/sja1105_spi.c if (rw == SPI_READ) rw 144 drivers/net/dsa/sja1105/sja1105_spi.c sja1105_spi_rw_mode_t rw, u64 base_addr, rw 161 drivers/net/dsa/sja1105/sja1105_spi.c rc = sja1105_spi_send_packed_buf(priv, rw, chunk.spi_address, rw 1774 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h u32 *val, int rw, bool sleep_ok); rw 5311 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c unsigned int rw, bool sleep_ok) rw 5321 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c (rw ? FW_CMD_READ_F : rw 5327 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c c.u.addrval.val = rw ? 0 : cpu_to_be32(vals[i]); rw 5333 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c if (rw) rw 5354 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c u32 *buff, u32 nregs, u32 start_index, int rw, rw 5375 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c rc = t4_tp_fw_ldst_rw(adap, cmd, buff, nregs, start_index, rw, rw 5381 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c if (rw) rw 7458 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c u32 *val, int rw, bool sleep_ok) rw 7476 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c if (rw) rw 767 drivers/net/ethernet/cortina/gemini.c union dma_rwptr rw; rw 775 drivers/net/ethernet/cortina/gemini.c rw.bits32 = readl(ptr_reg); rw 776 drivers/net/ethernet/cortina/gemini.c r = rw.bits.rptr; rw 777 drivers/net/ethernet/cortina/gemini.c w = rw.bits.wptr; rw 880 drivers/net/ethernet/cortina/gemini.c union dma_rwptr rw; rw 888 drivers/net/ethernet/cortina/gemini.c rw.bits32 = readl(geth->base + GLOBAL_SWFQ_RWPTR_REG); rw 889 drivers/net/ethernet/cortina/gemini.c pn = (refill ? rw.bits.wptr : rw.bits.rptr) >> fpp_order; rw 890 drivers/net/ethernet/cortina/gemini.c epn = (rw.bits.rptr >> fpp_order) - 1; rw 1237 drivers/net/ethernet/cortina/gemini.c union dma_rwptr rw; rw 1248 drivers/net/ethernet/cortina/gemini.c rw.bits32 = readl(ptr_reg); rw 1249 drivers/net/ethernet/cortina/gemini.c r = rw.bits.rptr; rw 1250 drivers/net/ethernet/cortina/gemini.c w = rw.bits.wptr; rw 1411 drivers/net/ethernet/cortina/gemini.c union dma_rwptr rw; rw 1415 drivers/net/ethernet/cortina/gemini.c rw.bits32 = readl(ptr_reg); rw 1419 drivers/net/ethernet/cortina/gemini.c r = rw.bits.rptr; rw 1420 drivers/net/ethernet/cortina/gemini.c w = rw.bits.wptr; rw 974 drivers/net/ethernet/dec/tulip/de4x5.c static void mii_ta(u_long rw, u_long ioaddr); rw 4871 drivers/net/ethernet/dec/tulip/de4x5.c mii_ta(u_long rw, u_long ioaddr) rw 4873 drivers/net/ethernet/dec/tulip/de4x5.c if (rw == MII_STWR) { rw 298 drivers/net/ethernet/freescale/fman/fman_keygen.c u32 rw = (u32)(write ? FM_KG_KGAR_WRITE : FM_KG_KGAR_READ); rw 301 drivers/net/ethernet/freescale/fman/fman_keygen.c rw | rw 319 drivers/net/ethernet/freescale/fman/fman_keygen.c u32 rw = write ? (u32)FM_KG_KGAR_WRITE : (u32)FM_KG_KGAR_READ; rw 322 drivers/net/ethernet/freescale/fman/fman_keygen.c rw | rw 363 drivers/net/ethernet/freescale/fman/fman_keygen.c u32 rw = write ? (u32)FM_KG_KGAR_WRITE : (u32)FM_KG_KGAR_READ; rw 366 drivers/net/ethernet/freescale/fman/fman_keygen.c rw | rw 85 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_icm_pool.c MLX5_SET(mkc, mkc, rw, 1); rw 806 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_send.c MLX5_SET(mkc, mkc, rw, 1); rw 158 drivers/net/ethernet/seeq/sgiseeq.c sregs->rw.eth_addr[i] = dev->dev_addr[i]; rw 301 drivers/net/ethernet/seeq/sgiseeq.c sregs->rw.wregs.control = sp->control; rw 302 drivers/net/ethernet/seeq/sgiseeq.c sregs->rw.wregs.frame_gap = 0; rw 419 drivers/net/ethernet/seeq/sgiseeq.c sregs->rw.wregs.control = sp->control & ~(SEEQ_CTRL_XCNT); rw 420 drivers/net/ethernet/seeq/sgiseeq.c sregs->rw.wregs.control = sp->control; rw 783 drivers/net/ethernet/seeq/sgiseeq.c sp->is_edlc = !(sp->sregs->rw.rregs.collision_tx[0] & 0xff); rw 29 drivers/net/ethernet/seeq/sgiseeq.h } rw; rw 109 drivers/net/wireless/ath/ath6kl/sdio.c static inline void ath6kl_sdio_set_cmd53_arg(u32 *arg, u8 rw, u8 func, rw 113 drivers/net/wireless/ath/ath6kl/sdio.c *arg = (((rw & 1) << 31) | rw 267 drivers/net/wireless/ath/ath6kl/sdio.c u8 opcode, rw; rw 292 drivers/net/wireless/ath/ath6kl/sdio.c rw = (scat_req->req & HIF_WRITE) ? CMD53_ARG_WRITE : CMD53_ARG_READ; rw 304 drivers/net/wireless/ath/ath6kl/sdio.c ath6kl_sdio_set_cmd53_arg(&cmd.arg, rw, ar_sdio->func->num, rw 42 drivers/net/wireless/rsi/rsi_91x_sdio.c static u32 rsi_sdio_set_cmd52_arg(bool rw, rw 48 drivers/net/wireless/rsi/rsi_91x_sdio.c return ((rw & 1) << 31) | ((func & 0x7) << 28) | rw 32 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c static int bits(u32 rw, int from, int to) rw 34 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c rw &= ~(0xffffffffU << (to+1)); rw 35 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c rw >>= from; rw 36 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c return rw; rw 39 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c static int bit(u32 rw, int bit) rw 41 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c return bits(rw, bit, bit); rw 44 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c static void dump_regwrite(u32 rw) rw 46 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c int reg = bits(rw, 18, 22); rw 47 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c int rw_flag = bits(rw, 23, 23); rw 48 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c PDEBUG("rf2959 %#010x reg %d rw %d", rw, reg, rw_flag); rw 54 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c bits(rw, 14, 15), bit(rw, 3), bit(rw, 2), bit(rw, 1), rw 55 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c bit(rw, 0)); rw 61 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c bit(rw, 17), bit(rw, 16), bit(rw, 15), bit(rw, 14), rw 62 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c bit(rw, 13), bit(rw, 12), bit(rw, 11), bit(rw, 10), rw 63 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c bits(rw, 7, 9), bits(rw, 4, 6), bits(rw, 0, 3)); rw 67 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c bits(rw, 6, 17), bits(rw, 0, 5)); rw 70 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c PDEBUG("reg3 IFPLL3 num %d", bits(rw, 0, 17)); rw 74 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c bits(rw, 8, 16), bits(rw, 4, 7), bits(rw, 0, 3)); rw 80 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c bit(rw, 17), bit(rw, 16), bit(rw, 15), bit(rw, 14), rw 81 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c bit(rw, 13), bit(rw, 12), bit(rw, 11), bit(rw, 10), rw 82 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c bits(rw, 7, 9), bits(rw, 4, 6), bits(rw, 0,3)); rw 86 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c bits(rw, 6, 17), bits(rw, 0, 5)); rw 89 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c PDEBUG("reg7 RFPLL3 num2 %d", bits(rw, 0, 17)); rw 93 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c bits(rw, 8, 16), bits(rw, 4, 7), bits(rw, 0, 3)); rw 97 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c bits(rw, 13, 17), bits(rw, 8, 12), bits(rw, 3, 7), rw 98 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c bits(rw, 0, 2)); rw 104 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c bit(rw, 17), bits(rw, 15, 16), bits(rw, 10, 14), rw 105 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c bits(rw, 7, 9), bits(rw, 4, 6), bit(rw, 3), bit(rw, 2), rw 106 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c bit(rw, 1), bit(rw, 0)); rw 111 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c bits(rw, 15, 17), bits(rw, 9, 14), bits(rw, 3, 8), rw 112 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c bits(rw, 0, 2)); rw 116 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c bits(rw, 12, 17), bits(rw, 6, 11), bits(rw, 0, 5)); rw 122 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c bit(rw, 17), bit(rw, 16), bit(rw, 15), rw 123 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c bits(rw, 8, 9), bits(rw, 5, 7), bits(rw, 3, 4), rw 124 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c bits(rw, 0, 2)); rw 129 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c bits(rw, 13, 17), bits(rw, 9, 12), bits(rw, 4, 8), rw 130 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c bits(rw, 0, 3)); rw 1912 drivers/net/wireless/zydas/zd1211rw/zd_usb.c struct reg_data *rw = &req->reg_writes[i]; rw 1913 drivers/net/wireless/zydas/zd1211rw/zd_usb.c rw->addr = cpu_to_le16((u16)ioreqs[i].addr); rw 1914 drivers/net/wireless/zydas/zd1211rw/zd_usb.c rw->value = cpu_to_le16(ioreqs[i].value); rw 64 drivers/nvdimm/blk.c struct bio_integrity_payload *bip, u64 lba, int rw) rw 94 drivers/nvdimm/blk.c cur_len, rw); rw 110 drivers/nvdimm/blk.c struct bio_integrity_payload *bip, u64 lba, int rw) rw 118 drivers/nvdimm/blk.c unsigned int len, unsigned int off, int rw, sector_t sector) rw 147 drivers/nvdimm/blk.c err = ndbr->do_io(ndbr, dev_offset, iobuf + off, cur_len, rw); rw 153 drivers/nvdimm/blk.c err = nd_blk_rw_integrity(nsblk, bip, lba, rw); rw 172 drivers/nvdimm/blk.c int err = 0, rw; rw 180 drivers/nvdimm/blk.c rw = bio_data_dir(bio); rw 187 drivers/nvdimm/blk.c bvec.bv_offset, rw, iter.bi_sector); rw 191 drivers/nvdimm/blk.c (rw == READ) ? "READ" : "WRITE", rw 205 drivers/nvdimm/blk.c resource_size_t offset, void *iobuf, size_t n, int rw, rw 222 drivers/nvdimm/blk.c return ndbr->do_io(ndbr, dev_offset, iobuf, n, rw); rw 1142 drivers/nvdimm/btt.c struct arena_info *arena, u32 postmap, int rw) rw 1167 drivers/nvdimm/btt.c if (rw) rw 1191 drivers/nvdimm/btt.c struct arena_info *arena, u32 postmap, int rw) rw 252 drivers/nvdimm/claim.c resource_size_t offset, void *buf, size_t size, int rw, rw 268 drivers/nvdimm/claim.c if (rw == READ) { rw 165 drivers/nvdimm/nd.h void *iobuf, u64 len, int rw); rw 699 drivers/nvme/host/core.c cmnd->rw.opcode = (rq_data_dir(req) ? nvme_cmd_write : nvme_cmd_read); rw 700 drivers/nvme/host/core.c cmnd->rw.nsid = cpu_to_le32(ns->head->ns_id); rw 701 drivers/nvme/host/core.c cmnd->rw.slba = cpu_to_le64(nvme_block_nr(ns, blk_rq_pos(req))); rw 702 drivers/nvme/host/core.c cmnd->rw.length = cpu_to_le16((blk_rq_bytes(req) >> ns->lba_shift) - 1); rw 728 drivers/nvme/host/core.c cmnd->rw.reftag = cpu_to_le32(t10_pi_ref_tag(req)); rw 733 drivers/nvme/host/core.c cmnd->rw.control = cpu_to_le16(control); rw 734 drivers/nvme/host/core.c cmnd->rw.dsmgmt = cpu_to_le32(dsmgmt); rw 1303 drivers/nvme/host/core.c c.rw.opcode = io.opcode; rw 1304 drivers/nvme/host/core.c c.rw.flags = io.flags; rw 1305 drivers/nvme/host/core.c c.rw.nsid = cpu_to_le32(ns->head->ns_id); rw 1306 drivers/nvme/host/core.c c.rw.slba = cpu_to_le64(io.slba); rw 1307 drivers/nvme/host/core.c c.rw.length = cpu_to_le16(io.nblocks); rw 1308 drivers/nvme/host/core.c c.rw.control = cpu_to_le16(io.control); rw 1309 drivers/nvme/host/core.c c.rw.dsmgmt = cpu_to_le32(io.dsmgmt); rw 1310 drivers/nvme/host/core.c c.rw.reftag = cpu_to_le32(io.reftag); rw 1311 drivers/nvme/host/core.c c.rw.apptag = cpu_to_le16(io.apptag); rw 1312 drivers/nvme/host/core.c c.rw.appmask = cpu_to_le16(io.appmask); rw 2260 drivers/nvme/host/fc.c sqe->rw.dptr.sgl.type = (NVME_TRANSPORT_SGL_DATA_DESC << 4) | rw 2262 drivers/nvme/host/fc.c sqe->rw.dptr.sgl.length = cpu_to_le32(data_len); rw 2263 drivers/nvme/host/fc.c sqe->rw.dptr.sgl.addr = 0; rw 808 drivers/nvme/host/pci.c &cmnd->rw, &bv); rw 813 drivers/nvme/host/pci.c &cmnd->rw, &bv); rw 837 drivers/nvme/host/pci.c ret = nvme_pci_setup_sgls(dev, req, &cmnd->rw, nr_mapped); rw 839 drivers/nvme/host/pci.c ret = nvme_pci_setup_prps(dev, req, &cmnd->rw); rw 855 drivers/nvme/host/pci.c cmnd->rw.metadata = cpu_to_le64(iod->meta_dma); rw 601 drivers/nvme/target/admin-cmd.c req->ns = nvmet_find_namespace(req->sq->ctrl, req->cmd->rw.nsid); rw 835 drivers/nvme/target/core.c req->ns = nvmet_find_namespace(req->sq->ctrl, cmd->rw.nsid); rw 124 drivers/nvme/target/io-cmd-bdev.c req->error_slba = le64_to_cpu(req->cmd->rw.slba); rw 158 drivers/nvme/target/io-cmd-bdev.c if (req->cmd->rw.opcode == nvme_cmd_write) { rw 161 drivers/nvme/target/io-cmd-bdev.c if (req->cmd->rw.control & cpu_to_le16(NVME_RW_FUA)) rw 170 drivers/nvme/target/io-cmd-bdev.c sector = le64_to_cpu(req->cmd->rw.slba); rw 96 drivers/nvme/target/io-cmd-file.c int rw; rw 98 drivers/nvme/target/io-cmd-file.c if (req->cmd->rw.opcode == nvme_cmd_write) { rw 99 drivers/nvme/target/io-cmd-file.c if (req->cmd->rw.control & cpu_to_le16(NVME_RW_FUA)) rw 102 drivers/nvme/target/io-cmd-file.c rw = WRITE; rw 105 drivers/nvme/target/io-cmd-file.c rw = READ; rw 108 drivers/nvme/target/io-cmd-file.c iov_iter_bvec(&iter, rw, req->f.bvec, nr_segs, count); rw 148 drivers/nvme/target/io-cmd-file.c pos = le64_to_cpu(req->cmd->rw.slba) << req->ns->blksize_shift; rw 494 drivers/nvme/target/nvmet.h return ((u32)le16_to_cpu(req->cmd->rw.length) + 1) << rw 57 drivers/nvme/target/rdma.c struct rdma_rw_ctx rw; rw 506 drivers/nvme/target/rdma.c rdma_rw_ctx_destroy(&rsp->rw, queue->cm_id->qp, rw 565 drivers/nvme/target/rdma.c first_wr = rdma_rw_ctx_wrs(&rsp->rw, cm_id->qp, rw 590 drivers/nvme/target/rdma.c rdma_rw_ctx_destroy(&rsp->rw, queue->cm_id->qp, rw 678 drivers/nvme/target/rdma.c ret = rdma_rw_ctx_init(&rsp->rw, cm_id->qp, cm_id->port_num, rw 745 drivers/nvme/target/rdma.c if (rdma_rw_ctx_post(&rsp->rw, queue->cm_id->qp, rw 56 drivers/nvme/target/trace.h ns = nvmet_find_namespace(ctrl, req->cmd->rw.nsid); rw 32 drivers/pci/pci-bridge-emul.c u32 rw; rw 44 drivers/pci/pci-bridge-emul.c .rw = (PCI_COMMAND_IO | PCI_COMMAND_MEMORY | rw 89 drivers/pci/pci-bridge-emul.c .rw = GENMASK(24, 0), rw 96 drivers/pci/pci-bridge-emul.c .rw = (GENMASK(15, 12) | GENMASK(7, 4)), rw 115 drivers/pci/pci-bridge-emul.c .rw = GENMASK(31, 20) | GENMASK(15, 4), rw 123 drivers/pci/pci-bridge-emul.c .rw = GENMASK(31, 20) | GENMASK(15, 4), rw 130 drivers/pci/pci-bridge-emul.c .rw = ~0, rw 134 drivers/pci/pci-bridge-emul.c .rw = ~0, rw 138 drivers/pci/pci-bridge-emul.c .rw = ~0, rw 147 drivers/pci/pci-bridge-emul.c .rw = GENMASK(31, 11) | BIT(0), rw 158 drivers/pci/pci-bridge-emul.c .rw = (GENMASK(7, 0) | rw 191 drivers/pci/pci-bridge-emul.c .rw = GENMASK(15, 0), rw 215 drivers/pci/pci-bridge-emul.c .rw = GENMASK(11, 3) | GENMASK(1, 0), rw 233 drivers/pci/pci-bridge-emul.c .rw = GENMASK(12, 0), rw 249 drivers/pci/pci-bridge-emul.c .rw = (PCI_EXP_RTCTL_SECEE | PCI_EXP_RTCTL_SENFEE | rw 301 drivers/pci/pci-bridge-emul.c bridge->pci_regs_behavior[PCI_PREF_MEMORY_BASE / 4].rw = 0; rw 426 drivers/pci/pci-bridge-emul.c new = old & (~mask | ~behavior[reg / 4].rw); rw 429 drivers/pci/pci-bridge-emul.c new |= (value << shift) & (behavior[reg / 4].rw & mask); rw 733 drivers/pinctrl/sh-pfc/core.c unsigned int i, n, rw, fw; rw 740 drivers/pinctrl/sh-pfc/core.c for (i = 0, n = 0, rw = 0; (fw = cfg_reg->var_field_width[i]); i++) { rw 743 drivers/pinctrl/sh-pfc/core.c drvname, cfg_reg->reg, rw, rw + fw - 1); rw 747 drivers/pinctrl/sh-pfc/core.c rw += fw; rw 750 drivers/pinctrl/sh-pfc/core.c if (rw != cfg_reg->reg_width) { rw 752 drivers/pinctrl/sh-pfc/core.c drvname, cfg_reg->reg, rw, cfg_reg->reg_width); rw 160 drivers/regulator/tps6586x-regulator.c TPS6586X_REGULATOR(_id, rw, _pname, vdata, vreg, shift, nbits, \ rw 88 drivers/s390/block/dasd_fba.c define_extent(struct ccw1 * ccw, struct DE_fba_data *data, int rw, rw 96 drivers/s390/block/dasd_fba.c if (rw == WRITE) rw 98 drivers/s390/block/dasd_fba.c else if (rw == READ) rw 108 drivers/s390/block/dasd_fba.c locate_record(struct ccw1 * ccw, struct LO_fba_data *data, int rw, rw 116 drivers/s390/block/dasd_fba.c if (rw == WRITE) rw 118 drivers/s390/block/dasd_fba.c else if (rw == READ) rw 1716 drivers/scsi/sg.c int rw = hp->dxfer_direction == SG_DXFER_TO_DEV ? WRITE : READ; rw 1804 drivers/scsi/sg.c res = import_iovec(rw, hp->dxferp, iov_count, 0, &iov, &i); rw 845 drivers/scsi/st.c if (STps->rw == ST_WRITING) /* Writing */ rw 941 drivers/scsi/st.c STps->rw = ST_IDLE; rw 1077 drivers/scsi/st.c STps->rw = ST_IDLE; rw 1306 drivers/scsi/st.c STps->rw = ST_IDLE; rw 1352 drivers/scsi/st.c if (STps->rw == ST_WRITING && !STp->pos_unknown) { rw 1372 drivers/scsi/st.c if (STps->rw == ST_WRITING && !STp->pos_unknown) { rw 1421 drivers/scsi/st.c if (!STm->sysv || STps->rw != ST_READING) { rw 1659 drivers/scsi/st.c if (STps->rw == ST_READING) { rw 1663 drivers/scsi/st.c STps->rw = ST_WRITING; rw 1664 drivers/scsi/st.c } else if (STps->rw != ST_WRITING && rw 1720 drivers/scsi/st.c STps->rw = ST_WRITING; rw 2124 drivers/scsi/st.c if (STps->rw == ST_WRITING) { rw 2128 drivers/scsi/st.c STps->rw = ST_READING; rw 2165 drivers/scsi/st.c STps->rw = ST_READING; rw 3001 drivers/scsi/st.c STps->rw = ST_IDLE; /* prevent automatic WEOF at close */ rw 3248 drivers/scsi/st.c STps->rw = ST_IDLE; rw 3600 drivers/scsi/st.c if (STps->rw == ST_WRITING && rw 3611 drivers/scsi/st.c STps->rw = ST_IDLE; rw 3637 drivers/scsi/st.c STps->rw = ST_IDLE; /* Prevent automatic WEOF and fsf */ rw 3684 drivers/scsi/st.c STp->ps[i].rw = ST_IDLE; rw 3698 drivers/scsi/st.c STp->ps[0].rw = ST_IDLE; rw 3760 drivers/scsi/st.c if (STps->rw == ST_WRITING) rw 3763 drivers/scsi/st.c else if (STps->rw == ST_READING) rw 4360 drivers/scsi/st.c STps->rw = ST_IDLE; rw 4899 drivers/scsi/st.c size_t count, int rw) rw 4926 drivers/scsi/st.c res = get_user_pages_fast(uaddr, nr_pages, rw == READ ? FOLL_WRITE : 0, rw 87 drivers/scsi/st.h unsigned char rw; rw 2156 drivers/scsi/ufs/ufshcd.c u32 rw; rw 2157 drivers/scsi/ufs/ufshcd.c rw = set & INTERRUPT_MASK_RW_VER_10; rw 2158 drivers/scsi/ufs/ufshcd.c set = rw | ((set ^ intrs) & intrs); rw 2176 drivers/scsi/ufs/ufshcd.c u32 rw; rw 2177 drivers/scsi/ufs/ufshcd.c rw = (set & INTERRUPT_MASK_RW_VER_10) & rw 2179 drivers/scsi/ufs/ufshcd.c set = rw | ((set & intrs) & ~INTERRUPT_MASK_RW_VER_10); rw 204 drivers/spi/spi-omap2-mcspi.c u32 l, rw; rw 209 drivers/spi/spi-omap2-mcspi.c rw = OMAP2_MCSPI_CHCONF_DMAR; rw 211 drivers/spi/spi-omap2-mcspi.c rw = OMAP2_MCSPI_CHCONF_DMAW; rw 214 drivers/spi/spi-omap2-mcspi.c l |= rw; rw 216 drivers/spi/spi-omap2-mcspi.c l &= ~rw; rw 3238 drivers/staging/exfat/exfat_super.c int rw; rw 3240 drivers/staging/exfat/exfat_super.c rw = iov_iter_rw(iter); rw 3242 drivers/staging/exfat/exfat_super.c if (rw == WRITE) { rw 3248 drivers/staging/exfat/exfat_super.c if ((ret < 0) && (rw & WRITE)) rw 857 drivers/target/target_core_pscsi.c int rw = (cmd->data_direction == DMA_TO_DEVICE); rw 893 drivers/target/target_core_pscsi.c if (rw) rw 898 drivers/target/target_core_pscsi.c (rw) ? "rw" : "r", nr_vecs); rw 425 drivers/tty/serial/sunsab.c writeb(up->cached_mode, &up->regs->rw.mode); rw 426 drivers/tty/serial/sunsab.c writeb(up->cached_pvr, &up->regs->rw.pvr); rw 430 drivers/tty/serial/sunsab.c tmp = readb(&up->regs->rw.ccr2); rw 433 drivers/tty/serial/sunsab.c writeb(tmp, &up->regs->rw.ccr2); rw 572 drivers/tty/serial/sunsab.c tmp = readb(&up->regs->rw.ccr0); rw 574 drivers/tty/serial/sunsab.c writeb(tmp, &up->regs->rw.ccr0); rw 611 drivers/tty/serial/sunsab.c up->cached_dafo = readb(&up->regs->rw.dafo); rw 613 drivers/tty/serial/sunsab.c writeb(up->cached_dafo, &up->regs->rw.dafo); rw 617 drivers/tty/serial/sunsab.c writeb(up->cached_mode, &up->regs->rw.mode); rw 631 drivers/tty/serial/sunsab.c tmp = readb(&up->regs->rw.ccr0); rw 633 drivers/tty/serial/sunsab.c writeb(tmp, &up->regs->rw.ccr0); rw 1009 drivers/tty/serial/sunsab.c up->cached_mode = readb(&up->regs->rw.mode); rw 1011 drivers/tty/serial/sunsab.c writeb(up->cached_mode, &up->regs->rw.mode); rw 1013 drivers/tty/serial/sunsab.c writeb(up->cached_mode, &up->regs->rw.mode); rw 116 drivers/tty/serial/sunsab.h __volatile__ struct sab82532_async_rw_regs rw; rw 49 drivers/usb/host/isp1362.h #define ISP1362_REG(name, addr, width, rw) \ rw 50 drivers/usb/host/isp1362.h static isp1362_reg_t ISP1362_REG_##name = ((addr) | (width) | (rw)) rw 58 drivers/usb/host/isp1362.h #define ISP1362_REG(name, addr, width, rw) \ rw 1170 drivers/vfio/pci/vfio_pci.c return vdev->region[index].ops->rw(vdev, buf, rw 51 drivers/vfio/pci/vfio_pci_igd.c .rw = vfio_pci_igd_rw, rw 212 drivers/vfio/pci/vfio_pci_igd.c .rw = vfio_pci_igd_cfg_rw, rw 188 drivers/vfio/pci/vfio_pci_nvlink2.c .rw = vfio_pci_nvgpu_rw, rw 385 drivers/vfio/pci/vfio_pci_nvlink2.c .rw = vfio_pci_npu2_rw, rw 55 drivers/vfio/pci/vfio_pci_private.h size_t (*rw)(struct vfio_pci_device *vdev, char __user *buf, rw 100 drivers/video/backlight/ili922x.c #define START_BYTE(id, rs, rw) \ rw 101 drivers/video/backlight/ili922x.c (0x70 | (((id) & 0x01) << 2) | (((rs) & 0x01) << 1) | ((rw) & 0x01)) rw 210 drivers/video/fbdev/core/bitblit.c unsigned int rw = info->var.xres - (vc->vc_cols*cw); rw 212 drivers/video/fbdev/core/bitblit.c unsigned int rs = info->var.xres - rw; rw 219 drivers/video/fbdev/core/bitblit.c if (rw && !bottom_only) { rw 222 drivers/video/fbdev/core/bitblit.c region.width = rw; rw 196 drivers/video/fbdev/core/fbcon_ccw.c unsigned int rw = info->var.yres - (vc->vc_cols*cw); rw 204 drivers/video/fbdev/core/fbcon_ccw.c if (rw && !bottom_only) { rw 207 drivers/video/fbdev/core/fbcon_ccw.c region.height = rw; rw 179 drivers/video/fbdev/core/fbcon_cw.c unsigned int rw = info->var.yres - (vc->vc_cols*cw); rw 181 drivers/video/fbdev/core/fbcon_cw.c unsigned int rs = info->var.yres - rw; rw 187 drivers/video/fbdev/core/fbcon_cw.c if (rw && !bottom_only) { rw 190 drivers/video/fbdev/core/fbcon_cw.c region.height = rw; rw 227 drivers/video/fbdev/core/fbcon_ud.c unsigned int rw = info->var.xres - (vc->vc_cols*cw); rw 234 drivers/video/fbdev/core/fbcon_ud.c if (rw && !bottom_only) { rw 237 drivers/video/fbdev/core/fbcon_ud.c region.width = rw; rw 147 drivers/w1/w1.c static BIN_ATTR_RW(rw, PAGE_SIZE); rw 201 fs/aio.c struct kiocb rw; rw 564 fs/aio.c struct aio_kiocb *req = container_of(iocb, struct aio_kiocb, rw); rw 624 fs/aio.c req->ki_cancel(&req->rw); rw 1424 fs/aio.c struct aio_kiocb *iocb = container_of(kiocb, struct aio_kiocb, rw); rw 1481 fs/aio.c static ssize_t aio_setup_rw(int rw, const struct iocb *iocb, rw 1489 fs/aio.c ssize_t ret = import_single_range(rw, buf, len, *iovec, iter); rw 1495 fs/aio.c return compat_import_iovec(rw, buf, len, UIO_FASTIOV, iovec, rw 1498 fs/aio.c return import_iovec(rw, buf, len, UIO_FASTIOV, iovec, iter); rw 1664 fs/aio.c struct aio_kiocb *aiocb = container_of(iocb, struct aio_kiocb, rw); rw 1837 fs/aio.c return aio_read(&req->rw, iocb, false, compat); rw 1839 fs/aio.c return aio_write(&req->rw, iocb, false, compat); rw 1841 fs/aio.c return aio_read(&req->rw, iocb, true, compat); rw 1843 fs/aio.c return aio_write(&req->rw, iocb, true, compat); rw 2031 fs/aio.c ret = kiocb->ki_cancel(&kiocb->rw); rw 28 fs/btrfs/locking.h static inline void btrfs_tree_unlock_rw(struct extent_buffer *eb, int rw) rw 30 fs/btrfs/locking.h if (rw == BTRFS_WRITE_LOCK || rw == BTRFS_WRITE_LOCK_BLOCKING) rw 32 fs/btrfs/locking.h else if (rw == BTRFS_READ_LOCK_BLOCKING) rw 34 fs/btrfs/locking.h else if (rw == BTRFS_READ_LOCK) rw 4358 fs/btrfs/relocation.c int rw = 0; rw 4384 fs/btrfs/relocation.c rw = 1; rw 4460 fs/btrfs/relocation.c if (err && rw) rw 579 fs/cifs/cifsproto.h int setup_aio_ctx_iter(struct cifs_aio_ctx *ctx, struct iov_iter *iter, int rw); rw 839 fs/cifs/misc.c setup_aio_ctx_iter(struct cifs_aio_ctx *ctx, struct iov_iter *iter, int rw) rw 924 fs/cifs/misc.c iov_iter_bvec(&ctx->iter, rw, ctx->bv, npages, ctx->len); rw 152 fs/crypto/crypto.c int fscrypt_crypt_block(const struct inode *inode, fscrypt_direction_t rw, rw 185 fs/crypto/crypto.c if (rw == FS_DECRYPT) rw 192 fs/crypto/crypto.c (rw == FS_DECRYPT ? "De" : "En"), lblk_num, res); rw 236 fs/crypto/fscrypt_private.h fscrypt_direction_t rw, u64 lblk_num, rw 2819 fs/f2fs/data.c int rw = iov_iter_rw(iter); rw 2834 fs/f2fs/data.c trace_f2fs_direct_IO_enter(inode, offset, count, rw); rw 2836 fs/f2fs/data.c if (rw == WRITE && whint_mode == WHINT_MODE_OFF) rw 2840 fs/f2fs/data.c if (!down_read_trylock(&fi->i_gc_rwsem[rw])) { rw 2846 fs/f2fs/data.c up_read(&fi->i_gc_rwsem[rw]); rw 2852 fs/f2fs/data.c down_read(&fi->i_gc_rwsem[rw]); rw 2858 fs/f2fs/data.c iter, rw == WRITE ? get_data_block_dio_write : rw 2865 fs/f2fs/data.c up_read(&fi->i_gc_rwsem[rw]); rw 2867 fs/f2fs/data.c if (rw == WRITE) { rw 2881 fs/f2fs/data.c trace_f2fs_direct_IO_exit(inode, offset, count, rw, err); rw 1072 fs/f2fs/f2fs.h #define is_read_io(rw) ((rw) == READ) rw 3708 fs/f2fs/f2fs.h int rw = iov_iter_rw(iter); rw 3710 fs/f2fs/f2fs.h return (test_opt(sbi, LFS) && (rw == WRITE) && rw 3718 fs/f2fs/f2fs.h int rw = iov_iter_rw(iter); rw 3730 fs/f2fs/f2fs.h if (test_opt(sbi, LFS) && (rw == WRITE)) { rw 308 fs/io_uring.c struct kiocb rw; rw 495 fs/io_uring.c int rw = 0; rw 501 fs/io_uring.c rw = !(req->rw.ki_flags & IOCB_DIRECT); rw 506 fs/io_uring.c queue_work(ctx->sqo_wq[rw], &req->work); rw 817 fs/io_uring.c struct kiocb *kiocb = &req->rw; rw 957 fs/io_uring.c struct io_kiocb *req = container_of(kiocb, struct io_kiocb, rw); rw 970 fs/io_uring.c struct io_kiocb *req = container_of(kiocb, struct io_kiocb, rw); rw 1004 fs/io_uring.c if (list_req->rw.ki_filp != req->rw.ki_filp) rw 1080 fs/io_uring.c struct kiocb *kiocb = &req->rw; rw 1165 fs/io_uring.c static int io_import_fixed(struct io_ring_ctx *ctx, int rw, rw 1199 fs/io_uring.c iov_iter_bvec(iter, rw, imu->bvec, imu->nr_bvecs, offset + len); rw 1239 fs/io_uring.c static ssize_t io_import_iovec(struct io_ring_ctx *ctx, int rw, rw 1259 fs/io_uring.c ssize_t ret = io_import_fixed(ctx, rw, sqe, iter); rw 1269 fs/io_uring.c return compat_import_iovec(rw, buf, sqe_len, UIO_FASTIOV, rw 1273 fs/io_uring.c return import_iovec(rw, buf, sqe_len, UIO_FASTIOV, iovec, iter); rw 1304 fs/io_uring.c static void io_async_list_note(int rw, struct io_kiocb *req, size_t len) rw 1306 fs/io_uring.c struct async_list *async_list = &req->ctx->pending_async[rw]; rw 1307 fs/io_uring.c struct kiocb *kiocb = &req->rw; rw 1339 fs/io_uring.c static ssize_t loop_rw_iter(int rw, struct file *file, struct kiocb *kiocb, rw 1368 fs/io_uring.c if (rw == READ) { rw 1397 fs/io_uring.c struct kiocb *kiocb = &req->rw; rw 1462 fs/io_uring.c struct kiocb *kiocb = &req->rw; rw 1592 fs/io_uring.c ret = vfs_fsync_range(req->rw.ki_filp, sqe_off, rw 1640 fs/io_uring.c ret = sync_file_range(req->rw.ki_filp, sqe_off, sqe_len, flags); rw 2211 fs/io_uring.c req->rw.ki_flags &= ~IOCB_NOWAIT; rw 146 fs/nfs/blocklayout/blocklayout.c do_add_page_to_bio(struct bio *bio, int npg, int rw, sector_t isect, rw 156 fs/nfs/blocklayout/blocklayout.c npg, rw, (unsigned long long)isect, offset, *len); rw 183 fs/nfs/blocklayout/blocklayout.c bio_set_op_attrs(bio, rw, 0); rw 192 fs/nfs/blocklayout/blocklayout.c static void bl_mark_devices_unavailable(struct nfs_pgio_header *header, bool rw) rw 203 fs/nfs/blocklayout/blocklayout.c if (!ext_tree_lookup(bl, isect, &be, rw)) rw 182 fs/nfs/blocklayout/blocklayout.h int ext_tree_remove(struct pnfs_block_layout *bl, bool rw, sector_t start, rw 187 fs/nfs/blocklayout/blocklayout.h struct pnfs_block_extent *ret, bool rw); rw 346 fs/nfs/blocklayout/extent_tree.c struct pnfs_block_extent *ret, bool rw) rw 351 fs/nfs/blocklayout/extent_tree.c if (!rw) rw 360 fs/nfs/blocklayout/extent_tree.c int ext_tree_remove(struct pnfs_block_layout *bl, bool rw, rw 368 fs/nfs/blocklayout/extent_tree.c if (rw) { rw 416 fs/ocfs2/refcounttree.c struct ocfs2_refcount_tree *tree, int rw) rw 420 fs/ocfs2/refcounttree.c ret = ocfs2_refcount_lock(tree, rw); rw 426 fs/ocfs2/refcounttree.c if (rw) rw 444 fs/ocfs2/refcounttree.c u64 ref_blkno, int rw, rw 462 fs/ocfs2/refcounttree.c ret = __ocfs2_lock_refcount_tree(osb, tree, rw); rw 473 fs/ocfs2/refcounttree.c ocfs2_unlock_refcount_tree(osb, tree, rw); rw 494 fs/ocfs2/refcounttree.c ocfs2_unlock_refcount_tree(osb, tree, rw); rw 517 fs/ocfs2/refcounttree.c struct ocfs2_refcount_tree *tree, int rw) rw 519 fs/ocfs2/refcounttree.c if (rw) rw 524 fs/ocfs2/refcounttree.c ocfs2_refcount_unlock(tree, rw); rw 29 fs/ocfs2/refcounttree.h int ocfs2_lock_refcount_tree(struct ocfs2_super *osb, u64 ref_blkno, int rw, rw 34 fs/ocfs2/refcounttree.h int rw); rw 405 include/linux/genhd.h int rw); rw 407 include/linux/genhd.h int rw); rw 960 include/linux/libata.h unsigned char *buf, unsigned int buflen, int rw); rw 1844 include/linux/libata.h unsigned char *buf, unsigned int buflen, int rw); rw 1846 include/linux/libata.h unsigned char *buf, unsigned int buflen, int rw); rw 152 include/linux/libnvdimm.h void *iobuf, u64 len, int rw); rw 3247 include/linux/mlx5/mlx5_ifc.h u8 rw[0x1]; rw 108 include/linux/mmiotrace.h extern void mmio_trace_rw(struct mmiotrace_rw *rw); rw 54 include/linux/nd.h void *buf, size_t size, int rw, unsigned long flags); rw 1215 include/linux/nvme.h struct nvme_rw_command rw; rw 55 include/net/llc_conn.h u8 rw; /* rx window size; max = 127 */ rw 354 include/net/llc_pdu.h u8 rw; /* sender receive window */ rw 376 include/net/llc_pdu.h xid_info->rw = rx_window << 1; /* size of receive window */ rw 401 include/net/llc_pdu.h xid_info->rw = rx_window << 1; rw 387 include/trace/events/block.h TP_PROTO(struct request_queue *q, struct bio *bio, int rw), rw 389 include/trace/events/block.h TP_ARGS(q, bio, rw), rw 425 include/trace/events/block.h TP_PROTO(struct request_queue *q, struct bio *bio, int rw), rw 427 include/trace/events/block.h TP_ARGS(q, bio, rw) rw 443 include/trace/events/block.h TP_PROTO(struct request_queue *q, struct bio *bio, int rw), rw 445 include/trace/events/block.h TP_ARGS(q, bio, rw) rw 1310 include/trace/events/ext4.h TP_PROTO(struct inode *inode, loff_t offset, unsigned long len, int rw), rw 1312 include/trace/events/ext4.h TP_ARGS(inode, offset, len, rw), rw 1319 include/trace/events/ext4.h __field( int, rw ) rw 1327 include/trace/events/ext4.h __entry->rw = rw; rw 1333 include/trace/events/ext4.h __entry->pos, __entry->len, __entry->rw) rw 1338 include/trace/events/ext4.h int rw, int ret), rw 1340 include/trace/events/ext4.h TP_ARGS(inode, offset, len, rw, ret), rw 1347 include/trace/events/ext4.h __field( int, rw ) rw 1356 include/trace/events/ext4.h __entry->rw = rw; rw 1364 include/trace/events/ext4.h __entry->rw, __entry->ret) rw 914 include/trace/events/f2fs.h TP_PROTO(struct inode *inode, loff_t offset, unsigned long len, int rw), rw 916 include/trace/events/f2fs.h TP_ARGS(inode, offset, len, rw), rw 923 include/trace/events/f2fs.h __field(int, rw) rw 931 include/trace/events/f2fs.h __entry->rw = rw; rw 938 include/trace/events/f2fs.h __entry->rw) rw 944 include/trace/events/f2fs.h int rw, int ret), rw 946 include/trace/events/f2fs.h TP_ARGS(inode, offset, len, rw, ret), rw 953 include/trace/events/f2fs.h __field(int, rw) rw 962 include/trace/events/f2fs.h __entry->rw = rw; rw 971 include/trace/events/f2fs.h __entry->rw, rw 248 include/xen/interface/io/blkif.h struct blkif_request_rw rw; rw 1020 include/xen/interface/io/sndif.h struct xensnd_rw_req rw; rw 572 kernel/futex.c get_futex_key(u32 __user *uaddr, int fshared, union futex_key *key, enum futex_access rw) rw 618 kernel/futex.c if (err == -EFAULT && rw == FUTEX_READ) { rw 206 kernel/trace/blktrace.c #define MASK_TC_BIT(rw, __name) ((rw & REQ_ ## __name) << \ rw 922 kernel/trace/blktrace.c struct bio *bio, int rw) rw 932 kernel/trace/blktrace.c __blk_add_trace(bt, 0, 0, rw, 0, BLK_TA_GETRQ, 0, 0, rw 941 kernel/trace/blktrace.c struct bio *bio, int rw) rw 951 kernel/trace/blktrace.c __blk_add_trace(bt, 0, 0, rw, 0, BLK_TA_SLEEPRQ, rw 282 kernel/trace/trace_entries.h __field_struct( struct mmiotrace_rw, rw ) rw 283 kernel/trace/trace_entries.h __field_desc( resource_size_t, rw, phys ) rw 284 kernel/trace/trace_entries.h __field_desc( unsigned long, rw, value ) rw 285 kernel/trace/trace_entries.h __field_desc( unsigned long, rw, pc ) rw 286 kernel/trace/trace_entries.h __field_desc( int, rw, map_id ) rw 287 kernel/trace/trace_entries.h __field_desc( unsigned char, rw, opcode ) rw 288 kernel/trace/trace_entries.h __field_desc( unsigned char, rw, width ) rw 171 kernel/trace/trace_mmiotrace.c struct mmiotrace_rw *rw; rw 178 kernel/trace/trace_mmiotrace.c rw = &field->rw; rw 180 kernel/trace/trace_mmiotrace.c switch (rw->opcode) { rw 184 kernel/trace/trace_mmiotrace.c rw->width, secs, usec_rem, rw->map_id, rw 185 kernel/trace/trace_mmiotrace.c (unsigned long long)rw->phys, rw 186 kernel/trace/trace_mmiotrace.c rw->value, rw->pc, 0); rw 191 kernel/trace/trace_mmiotrace.c rw->width, secs, usec_rem, rw->map_id, rw 192 kernel/trace/trace_mmiotrace.c (unsigned long long)rw->phys, rw 193 kernel/trace/trace_mmiotrace.c rw->value, rw->pc, 0); rw 199 kernel/trace/trace_mmiotrace.c secs, usec_rem, rw->map_id, rw 200 kernel/trace/trace_mmiotrace.c (unsigned long long)rw->phys, rw 201 kernel/trace/trace_mmiotrace.c (rw->value >> 16) & 0xff, (rw->value >> 8) & 0xff, rw 202 kernel/trace/trace_mmiotrace.c (rw->value >> 0) & 0xff, rw->pc, 0); rw 297 kernel/trace/trace_mmiotrace.c struct mmiotrace_rw *rw) rw 312 kernel/trace/trace_mmiotrace.c entry->rw = *rw; rw 318 kernel/trace/trace_mmiotrace.c void mmio_trace_rw(struct mmiotrace_rw *rw) rw 322 kernel/trace/trace_mmiotrace.c __trace_mmiotrace_rw(tr, data, rw); rw 1683 lib/iov_iter.c int import_single_range(int rw, void __user *buf, size_t len, rw 1693 lib/iov_iter.c iov_iter_init(i, rw, iov, 1, len); rw 1106 net/llc/af_llc.c llc->rw = opt; rw 1166 net/llc/af_llc.c val = llc->rw; break; rw 59 net/llc/llc_c_ev.c static u16 llc_util_ns_inside_rx_window(u8 ns, u8 vr, u8 rw) rw 62 net/llc/llc_c_ev.c (vr + rw - 1) % LLC_2_SEQ_NBR_MODULO); rw 205 net/llc/llc_c_ev.c !llc_util_ns_inside_rx_window(ns, vr, llc_sk(sk)->rw) ? 0 : 1; rw 217 net/llc/llc_c_ev.c !llc_util_ns_inside_rx_window(ns, vr, llc_sk(sk)->rw) ? 0 : 1; rw 228 net/llc/llc_c_ev.c llc_util_ns_inside_rx_window(ns, vr, llc_sk(sk)->rw) ? 0 : 1; rw 272 net/llc/llc_c_ev.c !llc_util_ns_inside_rx_window(ns, vr, llc_sk(sk)->rw) ? 0 : 1; rw 284 net/llc/llc_c_ev.c !llc_util_ns_inside_rx_window(ns, vr, llc_sk(sk)->rw) ? 0 : 1; rw 295 net/llc/llc_c_ev.c !llc_util_ns_inside_rx_window(ns, vr, llc_sk(sk)->rw) ? 0 : 1; rw 306 net/llc/llc_c_ev.c llc_util_ns_inside_rx_window(ns, vr, llc_sk(sk)->rw) ? 0 : 1; rw 901 net/llc/llc_conn.c llc->rw = 128; /* rx win size (opt and equal to rw 192 net/llc/llc_proc.c llc->retry_count, llc->k, llc->rw, llc->p_flag, llc->f_flag, rw 70 net/nfc/llcp.h u8 rw; rw 115 net/nfc/llcp.h u8 rw; rw 394 net/nfc/llcp_commands.c u8 *rw_tlv = NULL, rw_tlv_length, rw; rw 420 net/nfc/llcp_commands.c rw = sock->rw > LLCP_MAX_RW ? local->rw : sock->rw; rw 430 net/nfc/llcp_commands.c rw_tlv = nfc_llcp_build_tlv(LLCP_TLV_RW, &rw, 0, &rw_tlv_length); rw 469 net/nfc/llcp_commands.c u8 *rw_tlv = NULL, rw_tlv_length, rw; rw 483 net/nfc/llcp_commands.c rw = sock->rw > LLCP_MAX_RW ? local->rw : sock->rw; rw 493 net/nfc/llcp_commands.c rw_tlv = nfc_llcp_build_tlv(LLCP_TLV_RW, &rw, 0, &rw_tlv_length); rw 951 net/nfc/llcp_core.c new_sock->rw = sock->rw; rw 1595 net/nfc/llcp_core.c local->rw = LLCP_MAX_RW; rw 254 net/nfc/llcp_sock.c llcp_sock->rw = (u8) opt; rw 288 net/nfc/llcp_sock.c llcp_sock->rw, llcp_sock->miux); rw 301 net/nfc/llcp_sock.c u8 rw; rw 321 net/nfc/llcp_sock.c rw = llcp_sock->rw > LLCP_MAX_RW ? local->rw : llcp_sock->rw; rw 322 net/nfc/llcp_sock.c if (put_user(rw, (u32 __user *) optval)) rw 969 net/nfc/llcp_sock.c llcp_sock->rw = LLCP_MAX_RW + 1; rw 1006 net/nfc/netlink.c nla_put_u8(msg, NFC_ATTR_LLC_PARAM_RW, local->rw) || rw 1071 net/nfc/netlink.c u8 rw = 0; rw 1083 net/nfc/netlink.c rw = nla_get_u8(info->attrs[NFC_ATTR_LLC_PARAM_RW]); rw 1085 net/nfc/netlink.c if (rw > LLCP_MAX_RW) rw 1120 net/nfc/netlink.c local->rw = rw; rw 145 sound/xen/xen_snd_front.c req->op.rw.length = count; rw 146 sound/xen/xen_snd_front.c req->op.rw.offset = pos; rw 168 sound/xen/xen_snd_front.c req->op.rw.length = count; rw 169 sound/xen/xen_snd_front.c req->op.rw.offset = pos; rw 2645 tools/testing/nvdimm/test/nfit.c void *iobuf, u64 len, int rw) rw 2653 tools/testing/nvdimm/test/nfit.c if (rw)