b 443 arch/alpha/include/asm/bitops.h sched_find_first_bit(const unsigned long b[2]) b 447 arch/alpha/include/asm/bitops.h b0 = b[0]; b 448 arch/alpha/include/asm/bitops.h b1 = b[1]; b 405 arch/alpha/include/asm/core_apecs.h __EXTERN_INLINE void apecs_iowrite8(u8 b, void __iomem *xaddr) b 419 arch/alpha/include/asm/core_apecs.h w = __kernel_insbl(b, addr & 3); b 441 arch/alpha/include/asm/core_apecs.h __EXTERN_INLINE void apecs_iowrite16(u16 b, void __iomem *xaddr) b 455 arch/alpha/include/asm/core_apecs.h w = __kernel_inswl(b, addr & 3); b 467 arch/alpha/include/asm/core_apecs.h __EXTERN_INLINE void apecs_iowrite32(u32 b, void __iomem *xaddr) b 472 arch/alpha/include/asm/core_apecs.h *(vuip)addr = b; b 362 arch/alpha/include/asm/core_cia.h __EXTERN_INLINE void cia_iowrite8(u8 b, void __iomem *xaddr) b 373 arch/alpha/include/asm/core_cia.h w = __kernel_insbl(b, addr & 3); b 392 arch/alpha/include/asm/core_cia.h __EXTERN_INLINE void cia_iowrite16(u16 b, void __iomem *xaddr) b 403 arch/alpha/include/asm/core_cia.h w = __kernel_inswl(b, addr & 3); b 415 arch/alpha/include/asm/core_cia.h __EXTERN_INLINE void cia_iowrite32(u32 b, void __iomem *xaddr) b 420 arch/alpha/include/asm/core_cia.h *(vuip)addr = b; b 251 arch/alpha/include/asm/core_lca.h __EXTERN_INLINE void lca_iowrite8(u8 b, void __iomem *xaddr) b 265 arch/alpha/include/asm/core_lca.h w = __kernel_insbl(b, addr & 3); b 287 arch/alpha/include/asm/core_lca.h __EXTERN_INLINE void lca_iowrite16(u16 b, void __iomem *xaddr) b 301 arch/alpha/include/asm/core_lca.h w = __kernel_inswl(b, addr & 3); b 313 arch/alpha/include/asm/core_lca.h __EXTERN_INLINE void lca_iowrite32(u32 b, void __iomem *xaddr) b 318 arch/alpha/include/asm/core_lca.h *(vuip)addr = b; b 336 arch/alpha/include/asm/core_marvel.h extern void marvel_iowrite8(u8 b, void __iomem *); b 343 arch/alpha/include/asm/core_marvel.h __EXTERN_INLINE void marvel_iowrite16(u16 b, void __iomem *addr) b 345 arch/alpha/include/asm/core_marvel.h __kernel_stw(b, *(vusp)addr); b 282 arch/alpha/include/asm/core_mcpcia.h __EXTERN_INLINE void mcpcia_iowrite8(u8 b, void __iomem *xaddr) b 290 arch/alpha/include/asm/core_mcpcia.h w = __kernel_insbl(b, addr & 3); b 306 arch/alpha/include/asm/core_mcpcia.h __EXTERN_INLINE void mcpcia_iowrite16(u16 b, void __iomem *xaddr) b 314 arch/alpha/include/asm/core_mcpcia.h w = __kernel_inswl(b, addr & 3); b 328 arch/alpha/include/asm/core_mcpcia.h __EXTERN_INLINE void mcpcia_iowrite32(u32 b, void __iomem *xaddr) b 335 arch/alpha/include/asm/core_mcpcia.h *(vuip)addr = b; b 370 arch/alpha/include/asm/core_t2.h extern inline void t2_outb(u8 b, unsigned long addr) b 374 arch/alpha/include/asm/core_t2.h w = __kernel_insbl(b, addr & 3); b 385 arch/alpha/include/asm/core_t2.h extern inline void t2_outw(u16 b, unsigned long addr) b 389 arch/alpha/include/asm/core_t2.h w = __kernel_inswl(b, addr & 3); b 399 arch/alpha/include/asm/core_t2.h extern inline void t2_outl(u32 b, unsigned long addr) b 401 arch/alpha/include/asm/core_t2.h *(vuip) ((addr << 5) + T2_IO + 0x18) = b; b 503 arch/alpha/include/asm/core_t2.h __EXTERN_INLINE void t2_writeb(u8 b, volatile void __iomem *xaddr) b 510 arch/alpha/include/asm/core_t2.h w = __kernel_insbl(b, addr & 3); b 514 arch/alpha/include/asm/core_t2.h __EXTERN_INLINE void t2_writew(u16 b, volatile void __iomem *xaddr) b 521 arch/alpha/include/asm/core_t2.h w = __kernel_inswl(b, addr & 3); b 529 arch/alpha/include/asm/core_t2.h __EXTERN_INLINE void t2_writel(u32 b, volatile void __iomem *xaddr) b 535 arch/alpha/include/asm/core_t2.h *(vuip) ((addr << 5) + T2_SPARSE_MEM + 0x18) = b; b 538 arch/alpha/include/asm/core_t2.h __EXTERN_INLINE void t2_writeq(u64 b, volatile void __iomem *xaddr) b 546 arch/alpha/include/asm/core_t2.h *(vuip)work = b; b 547 arch/alpha/include/asm/core_t2.h *(vuip)(work + (4 << 5)) = b >> 32; b 582 arch/alpha/include/asm/core_t2.h __EXTERN_INLINE void t2_iowrite##NS(u##NS b, void __iomem *xaddr) \ b 585 arch/alpha/include/asm/core_t2.h t2_write##OS(b, xaddr); \ b 587 arch/alpha/include/asm/core_t2.h t2_out##OS(b, (unsigned long)xaddr - T2_IO); \ b 590 arch/alpha/include/asm/core_t2.h IOPORT(b, 8) b 77 arch/alpha/include/asm/err_common.h } b; b 50 arch/alpha/include/asm/extable.h #define swap_ex_entry_fixup(a, b, tmp, delta) \ b 52 arch/alpha/include/asm/extable.h (a)->fixup.unit = (b)->fixup.unit; \ b 53 arch/alpha/include/asm/extable.h (b)->fixup.unit = (tmp).fixup.unit; \ b 135 arch/alpha/include/asm/io.h #define IO_CONCAT(a,b) _IO_CONCAT(a,b) b 136 arch/alpha/include/asm/io.h #define _IO_CONCAT(a,b) a ## _ ## b b 149 arch/alpha/include/asm/io.h static inline void generic_##NAME(TYPE b, QUAL void __iomem *addr) \ b 151 arch/alpha/include/asm/io.h alpha_mv.mv_##NAME(b, addr); \ b 243 arch/alpha/include/asm/io.h extern void outb(u8 b, unsigned long port); b 244 arch/alpha/include/asm/io.h extern void outw(u16 b, unsigned long port); b 245 arch/alpha/include/asm/io.h extern void outl(u32 b, unsigned long port); b 251 arch/alpha/include/asm/io.h extern void writeb(u8 b, volatile void __iomem *addr); b 252 arch/alpha/include/asm/io.h extern void writew(u16 b, volatile void __iomem *addr); b 253 arch/alpha/include/asm/io.h extern void writel(u32 b, volatile void __iomem *addr); b 254 arch/alpha/include/asm/io.h extern void writeq(u64 b, volatile void __iomem *addr); b 260 arch/alpha/include/asm/io.h extern void __raw_writeb(u8 b, volatile void __iomem *addr); b 261 arch/alpha/include/asm/io.h extern void __raw_writew(u16 b, volatile void __iomem *addr); b 262 arch/alpha/include/asm/io.h extern void __raw_writel(u32 b, volatile void __iomem *addr); b 263 arch/alpha/include/asm/io.h extern void __raw_writeq(u64 b, volatile void __iomem *addr); b 337 arch/alpha/include/asm/io.h extern inline void iowrite8(u8 b, void __iomem *addr) b 340 arch/alpha/include/asm/io.h IO_CONCAT(__IO_PREFIX, iowrite8)(b, addr); b 343 arch/alpha/include/asm/io.h extern inline void iowrite16(u16 b, void __iomem *addr) b 346 arch/alpha/include/asm/io.h IO_CONCAT(__IO_PREFIX, iowrite16)(b, addr); b 359 arch/alpha/include/asm/io.h extern inline void outb(u8 b, unsigned long port) b 361 arch/alpha/include/asm/io.h iowrite8(b, ioport_map(port, 1)); b 364 arch/alpha/include/asm/io.h extern inline void outw(u16 b, unsigned long port) b 366 arch/alpha/include/asm/io.h iowrite16(b, ioport_map(port, 2)); b 378 arch/alpha/include/asm/io.h extern inline void iowrite32(u32 b, void __iomem *addr) b 381 arch/alpha/include/asm/io.h IO_CONCAT(__IO_PREFIX, iowrite32)(b, addr); b 389 arch/alpha/include/asm/io.h extern inline void outl(u32 b, unsigned long port) b 391 arch/alpha/include/asm/io.h iowrite32(b, ioport_map(port, 4)); b 406 arch/alpha/include/asm/io.h extern inline void __raw_writeb(u8 b, volatile void __iomem *addr) b 408 arch/alpha/include/asm/io.h IO_CONCAT(__IO_PREFIX,writeb)(b, addr); b 411 arch/alpha/include/asm/io.h extern inline void __raw_writew(u16 b, volatile void __iomem *addr) b 413 arch/alpha/include/asm/io.h IO_CONCAT(__IO_PREFIX,writew)(b, addr); b 430 arch/alpha/include/asm/io.h extern inline void writeb(u8 b, volatile void __iomem *addr) b 433 arch/alpha/include/asm/io.h __raw_writeb(b, addr); b 436 arch/alpha/include/asm/io.h extern inline void writew(u16 b, volatile void __iomem *addr) b 439 arch/alpha/include/asm/io.h __raw_writew(b, addr); b 454 arch/alpha/include/asm/io.h extern inline void __raw_writel(u32 b, volatile void __iomem *addr) b 456 arch/alpha/include/asm/io.h IO_CONCAT(__IO_PREFIX,writel)(b, addr); b 459 arch/alpha/include/asm/io.h extern inline void __raw_writeq(u64 b, volatile void __iomem *addr) b 461 arch/alpha/include/asm/io.h IO_CONCAT(__IO_PREFIX,writeq)(b, addr); b 478 arch/alpha/include/asm/io.h extern inline void writel(u32 b, volatile void __iomem *addr) b 481 arch/alpha/include/asm/io.h __raw_writel(b, addr); b 484 arch/alpha/include/asm/io.h extern inline void writeq(u64 b, volatile void __iomem *addr) b 487 arch/alpha/include/asm/io.h __raw_writeq(b, addr); b 506 arch/alpha/include/asm/io.h #define writeb_relaxed(b, addr) __raw_writeb(b, addr) b 507 arch/alpha/include/asm/io.h #define writew_relaxed(b, addr) __raw_writew(b, addr) b 508 arch/alpha/include/asm/io.h #define writel_relaxed(b, addr) __raw_writel(b, addr) b 509 arch/alpha/include/asm/io.h #define writeq_relaxed(b, addr) __raw_writeq(b, addr) b 22 arch/alpha/include/asm/io_trivial.h IO_CONCAT(__IO_PREFIX,iowrite8)(u8 b, void __iomem *a) b 24 arch/alpha/include/asm/io_trivial.h __kernel_stb(b, *(volatile u8 __force *)a); b 28 arch/alpha/include/asm/io_trivial.h IO_CONCAT(__IO_PREFIX,iowrite16)(u16 b, void __iomem *a) b 30 arch/alpha/include/asm/io_trivial.h __kernel_stw(b, *(volatile u16 __force *)a); b 42 arch/alpha/include/asm/io_trivial.h IO_CONCAT(__IO_PREFIX,iowrite32)(u32 b, void __iomem *a) b 44 arch/alpha/include/asm/io_trivial.h *(volatile u32 __force *)a = b; b 62 arch/alpha/include/asm/io_trivial.h IO_CONCAT(__IO_PREFIX,writeb)(u8 b, volatile void __iomem *a) b 64 arch/alpha/include/asm/io_trivial.h __kernel_stb(b, *(volatile u8 __force *)a); b 68 arch/alpha/include/asm/io_trivial.h IO_CONCAT(__IO_PREFIX,writew)(u16 b, volatile void __iomem *a) b 70 arch/alpha/include/asm/io_trivial.h __kernel_stw(b, *(volatile u16 __force *)a); b 88 arch/alpha/include/asm/io_trivial.h IO_CONCAT(__IO_PREFIX,writeb)(u8 b, volatile void __iomem *a) b 91 arch/alpha/include/asm/io_trivial.h IO_CONCAT(__IO_PREFIX,iowrite8)(b, addr); b 95 arch/alpha/include/asm/io_trivial.h IO_CONCAT(__IO_PREFIX,writew)(u16 b, volatile void __iomem *a) b 98 arch/alpha/include/asm/io_trivial.h IO_CONCAT(__IO_PREFIX,iowrite16)(b, addr); b 116 arch/alpha/include/asm/io_trivial.h IO_CONCAT(__IO_PREFIX,writel)(u32 b, volatile void __iomem *a) b 118 arch/alpha/include/asm/io_trivial.h *(volatile u32 __force *)a = b; b 122 arch/alpha/include/asm/io_trivial.h IO_CONCAT(__IO_PREFIX,writeq)(u64 b, volatile void __iomem *a) b 124 arch/alpha/include/asm/io_trivial.h *(volatile u64 __force *)a = b; b 119 arch/alpha/include/asm/jensen.h static inline void jensen_local_outb(u8 b, unsigned long addr) b 121 arch/alpha/include/asm/jensen.h *(vuip)((addr << 9) + EISA_VL82C106) = b; b 134 arch/alpha/include/asm/jensen.h static inline void jensen_bus_outb(u8 b, unsigned long addr) b 137 arch/alpha/include/asm/jensen.h *(vuip)((addr << 7) + EISA_IO + 0x00) = b * 0x01010101; b 162 arch/alpha/include/asm/jensen.h __EXTERN_INLINE void jensen_outb(u8 b, unsigned long addr) b 165 arch/alpha/include/asm/jensen.h jensen_local_outb(b, addr); b 167 arch/alpha/include/asm/jensen.h jensen_bus_outb(b, addr); b 186 arch/alpha/include/asm/jensen.h __EXTERN_INLINE void jensen_outw(u16 b, unsigned long addr) b 189 arch/alpha/include/asm/jensen.h *(vuip) ((addr << 7) + EISA_IO + 0x20) = b * 0x00010001; b 193 arch/alpha/include/asm/jensen.h __EXTERN_INLINE void jensen_outl(u32 b, unsigned long addr) b 196 arch/alpha/include/asm/jensen.h *(vuip) ((addr << 7) + EISA_IO + 0x60) = b; b 249 arch/alpha/include/asm/jensen.h __EXTERN_INLINE void jensen_writeb(u8 b, volatile void __iomem *xaddr) b 254 arch/alpha/include/asm/jensen.h *(vuip) ((addr << 7) + EISA_MEM + 0x00) = b * 0x01010101; b 257 arch/alpha/include/asm/jensen.h __EXTERN_INLINE void jensen_writew(u16 b, volatile void __iomem *xaddr) b 262 arch/alpha/include/asm/jensen.h *(vuip) ((addr << 7) + EISA_MEM + 0x20) = b * 0x00010001; b 265 arch/alpha/include/asm/jensen.h __EXTERN_INLINE void jensen_writel(u32 b, volatile void __iomem *xaddr) b 270 arch/alpha/include/asm/jensen.h *(vuip) ((addr << 7) + EISA_MEM + 0x60) = b; b 273 arch/alpha/include/asm/jensen.h __EXTERN_INLINE void jensen_writeq(u64 b, volatile void __iomem *xaddr) b 279 arch/alpha/include/asm/jensen.h *(vuip) (addr) = b; b 280 arch/alpha/include/asm/jensen.h *(vuip) (addr + (4 << 7)) = b >> 32; b 315 arch/alpha/include/asm/jensen.h __EXTERN_INLINE void jensen_iowrite##NS(u##NS b, void __iomem *xaddr) \ b 318 arch/alpha/include/asm/jensen.h jensen_write##OS(b, xaddr - 0x100000000ul); \ b 320 arch/alpha/include/asm/jensen.h jensen_out##OS(b, (unsigned long)xaddr); \ b 323 arch/alpha/include/asm/jensen.h IOPORT(b, 8) b 220 arch/alpha/include/asm/mmu_context.h # define switch_mm(a,b,c) alpha_mv.mv_switch_mm((a),(b),(c)) b 224 arch/alpha/include/asm/mmu_context.h # define switch_mm(a,b,c) ev4_switch_mm((a),(b),(c)) b 227 arch/alpha/include/asm/mmu_context.h # define switch_mm(a,b,c) ev5_switch_mm((a),(b),(c)) b 23 arch/alpha/include/asm/uaccess.h #define segment_eq(a, b) ((a).seg == (b).seg) b 20 arch/alpha/include/uapi/asm/compiler.h # define __kernel_cmpbge(a, b) __builtin_alpha_cmpbge(a, b) b 46 arch/alpha/include/uapi/asm/compiler.h # define __kernel_cmpbge(a, b) \ b 48 arch/alpha/include/uapi/asm/compiler.h __asm__("cmpbge %r2,%1,%0" : "=r"(__kir) : "rI"(b), "rJ"(a)); \ b 651 arch/alpha/kernel/core_marvel.c __marvel_rtc_io(u8 b, unsigned long addr, int write) b 660 arch/alpha/kernel/core_marvel.c if (write) index = b; b 666 arch/alpha/kernel/core_marvel.c rtc_access.data = bcd2bin(b); b 828 arch/alpha/kernel/core_marvel.c marvel_iowrite8(u8 b, void __iomem *xaddr) b 834 arch/alpha/kernel/core_marvel.c __marvel_rtc_io(b, addr, 1); b 836 arch/alpha/kernel/core_marvel.c __kernel_stb(b, *(vucp)addr); b 174 arch/alpha/kernel/err_common.c timestamp->b.month, timestamp->b.day, b 175 arch/alpha/kernel/err_common.c timestamp->b.year, timestamp->b.hour, b 176 arch/alpha/kernel/err_common.c timestamp->b.minute, timestamp->b.second); b 38 arch/alpha/kernel/io.c void iowrite8(u8 b, void __iomem *addr) b 41 arch/alpha/kernel/io.c IO_CONCAT(__IO_PREFIX,iowrite8)(b, addr); b 44 arch/alpha/kernel/io.c void iowrite16(u16 b, void __iomem *addr) b 47 arch/alpha/kernel/io.c IO_CONCAT(__IO_PREFIX,iowrite16)(b, addr); b 50 arch/alpha/kernel/io.c void iowrite32(u32 b, void __iomem *addr) b 53 arch/alpha/kernel/io.c IO_CONCAT(__IO_PREFIX,iowrite32)(b, addr); b 78 arch/alpha/kernel/io.c void outb(u8 b, unsigned long port) b 80 arch/alpha/kernel/io.c iowrite8(b, ioport_map(port, 1)); b 83 arch/alpha/kernel/io.c void outw(u16 b, unsigned long port) b 85 arch/alpha/kernel/io.c iowrite16(b, ioport_map(port, 2)); b 88 arch/alpha/kernel/io.c void outl(u32 b, unsigned long port) b 90 arch/alpha/kernel/io.c iowrite32(b, ioport_map(port, 4)); b 120 arch/alpha/kernel/io.c void __raw_writeb(u8 b, volatile void __iomem *addr) b 122 arch/alpha/kernel/io.c IO_CONCAT(__IO_PREFIX,writeb)(b, addr); b 125 arch/alpha/kernel/io.c void __raw_writew(u16 b, volatile void __iomem *addr) b 127 arch/alpha/kernel/io.c IO_CONCAT(__IO_PREFIX,writew)(b, addr); b 130 arch/alpha/kernel/io.c void __raw_writel(u32 b, volatile void __iomem *addr) b 132 arch/alpha/kernel/io.c IO_CONCAT(__IO_PREFIX,writel)(b, addr); b 135 arch/alpha/kernel/io.c void __raw_writeq(u64 b, volatile void __iomem *addr) b 137 arch/alpha/kernel/io.c IO_CONCAT(__IO_PREFIX,writeq)(b, addr); b 177 arch/alpha/kernel/io.c void writeb(u8 b, volatile void __iomem *addr) b 180 arch/alpha/kernel/io.c __raw_writeb(b, addr); b 183 arch/alpha/kernel/io.c void writew(u16 b, volatile void __iomem *addr) b 186 arch/alpha/kernel/io.c __raw_writew(b, addr); b 189 arch/alpha/kernel/io.c void writel(u32 b, volatile void __iomem *addr) b 192 arch/alpha/kernel/io.c __raw_writel(b, addr); b 195 arch/alpha/kernel/io.c void writeq(u64 b, volatile void __iomem *addr) b 198 arch/alpha/kernel/io.c __raw_writeq(b, addr); b 285 arch/alpha/kernel/pci.c pcibios_claim_one_bus(struct pci_bus *b) b 290 arch/alpha/kernel/pci.c list_for_each_entry(dev, &b->devices, bus_list) { b 308 arch/alpha/kernel/pci.c list_for_each_entry(child_bus, &b->children, node) b 315 arch/alpha/kernel/pci.c struct pci_bus *b; b 317 arch/alpha/kernel/pci.c list_for_each_entry(b, &pci_root_buses, node) b 318 arch/alpha/kernel/pci.c pcibios_claim_one_bus(b); b 88 arch/arc/include/asm/entry-compact.h b.d 66f b 56 arch/arc/include/asm/io.h u8 b; b 60 arch/arc/include/asm/io.h : "=r" (b) b 64 arch/arc/include/asm/io.h return b; b 125 arch/arc/include/asm/io.h __raw_readsx(8, b) b 132 arch/arc/include/asm/io.h static inline void __raw_writeb(u8 b, volatile void __iomem *addr) b 137 arch/arc/include/asm/io.h : "r" (b), "m" (*(volatile u8 __force *)addr) b 186 arch/arc/include/asm/io.h __raw_writesx(8, b) b 82 arch/arc/include/asm/mmu.h #define tlb_paranoid_check(a, b) b 18 arch/arc/include/asm/segment.h #define segment_eq(a, b) ((a) == (b)) b 68 arch/arc/include/asm/tlb-mmu1.h b.d 89f b 151 arch/arc/include/asm/unwind.h #define unwind_add_table(a, b, c) b 152 arch/arc/include/asm/unwind.h #define unwind_remove_table(a, b) b 168 arch/arm/common/dmabounce.c struct safe_buffer *b, *rb = NULL; b 173 arch/arm/common/dmabounce.c list_for_each_entry(b, &device_info->safe_buffers, node) b 174 arch/arm/common/dmabounce.c if (b->safe_dma_addr <= safe_dma_addr && b 175 arch/arm/common/dmabounce.c b->safe_dma_addr + b->size > safe_dma_addr) { b 176 arch/arm/common/dmabounce.c rb = b; b 30 arch/arm/common/sharpsl_param.c #define MAGIC_CHG(a,b,c,d) ( ( d << 24 ) | ( c << 16 ) | ( b << 8 ) | a ) b 54 arch/arm/crypto/aes-ce-glue.c u8 b[AES_BLOCK_SIZE]; b 93 arch/arm/crypto/ghash-ce-glue.c dg[0] = be64_to_cpu(dst.b); b 153 arch/arm/crypto/ghash-ce-glue.c h[0] = (be64_to_cpu(k->b) << 1) | carry; b 154 arch/arm/crypto/ghash-ce-glue.c h[1] = (be64_to_cpu(k->a) << 1) | (be64_to_cpu(k->b) >> 63); b 275 arch/arm/include/asm/assembler.h W(b) . + up_b_offset ;\ b 280 arch/arm/include/asm/assembler.h #define ALT_UP_B(label) b label b 376 arch/arm/include/asm/assembler.h \instr\()b\t\cond\().w \reg, [\ptr, #\off] b 417 arch/arm/include/asm/assembler.h \instr\()b\t\cond \reg, [\ptr], #\inc b 135 arch/arm/include/asm/glue-cache.h unsigned long b, unsigned int c) { } b 137 arch/arm/include/asm/glue-cache.h static inline void nop_coherent_kern_range(unsigned long a, unsigned long b) { } b 139 arch/arm/include/asm/glue-cache.h unsigned long b) { return 0; } b 142 arch/arm/include/asm/glue-cache.h static inline void nop_dma_flush_range(const void *a, const void *b) { } b 29 arch/arm/include/asm/jump_label.h WASM(b) " %l[l_yes]\n\t" b 25 arch/arm/include/asm/pgtable-nommu.h #define pmd_offset(a, b) ((void *)0) b 87 arch/arm/include/asm/pgtable.h #define _MOD_PROT(p, b) __pgprot(pgprot_val(p) | (b)) b 79 arch/arm/include/asm/uaccess.h #define segment_eq(a, b) ((a) == (b)) b 266 arch/arm/include/asm/uaccess.h #define segment_eq(a, b) (1) b 16 arch/arm/kernel/atags_proc.c struct buffer *b = PDE_DATA(file_inode(file)); b 17 arch/arm/kernel/atags_proc.c return simple_read_from_buffer(buf, count, ppos, b->data, b->size); b 41 arch/arm/kernel/atags_proc.c struct buffer *b; b 57 arch/arm/kernel/atags_proc.c b = kmalloc(sizeof(*b) + size, GFP_KERNEL); b 58 arch/arm/kernel/atags_proc.c if (!b) b 61 arch/arm/kernel/atags_proc.c b->size = size; b 62 arch/arm/kernel/atags_proc.c memcpy(b->data, atags_copy, size); b 64 arch/arm/kernel/atags_proc.c tags_entry = proc_create_data("atags", 0400, NULL, &atags_fops, b); b 71 arch/arm/kernel/atags_proc.c kfree(b); b 76 arch/arm/kernel/module-plts.c #define cmp_3way(a,b) ((a) < (b) ? -1 : (a) > (b)) b 78 arch/arm/kernel/module-plts.c static int cmp_rel(const void *a, const void *b) b 80 arch/arm/kernel/module-plts.c const Elf32_Rel *x = a, *y = b; b 621 arch/arm/mach-davinci/board-dm365-evm.c static void cpld_led_set(struct led_classdev *cdev, enum led_brightness b) b 626 arch/arm/mach-davinci/board-dm365-evm.c if (b != LED_OFF) b 51 arch/arm/mach-ebsa110/include/mach/io.h #define readb(b) __readb(b) b 52 arch/arm/mach-ebsa110/include/mach/io.h #define readw(b) __readw(b) b 53 arch/arm/mach-ebsa110/include/mach/io.h #define readl(b) __readl(b) b 58 arch/arm/mach-ebsa110/include/mach/io.h #define writeb(v,b) __writeb(v,b) b 59 arch/arm/mach-ebsa110/include/mach/io.h #define writew(v,b) __writew(v,b) b 60 arch/arm/mach-ebsa110/include/mach/io.h #define writel(v,b) __writel(v,b) b 20 arch/arm/mach-ebsa110/leds.c enum led_brightness b) b 24 arch/arm/mach-ebsa110/leds.c if (b != LED_OFF) b 49 arch/arm/mach-footbridge/ebsa285.c enum led_brightness b) b 54 arch/arm/mach-footbridge/ebsa285.c if (b == LED_OFF) b 25 arch/arm/mach-footbridge/include/mach/hardware.h #define MMU_IO(a, b) (a) b 27 arch/arm/mach-footbridge/include/mach/hardware.h #define MMU_IO(a, b) (b) b 689 arch/arm/mach-footbridge/netwinder-hw.c enum led_brightness b) b 698 arch/arm/mach-footbridge/netwinder-hw.c if (b != LED_OFF) b 193 arch/arm/mach-orion5x/dns323-setup.c static int __init dns323_parse_hex_byte(const char *b) b 198 arch/arm/mach-orion5x/dns323-setup.c hi = dns323_parse_hex_nibble(b[0]); b 199 arch/arm/mach-orion5x/dns323-setup.c lo = dns323_parse_hex_nibble(b[1]); b 66 arch/arm/mach-orion5x/tsx09-common.c static int __init qnap_tsx09_parse_hex_byte(const char *b) b 71 arch/arm/mach-orion5x/tsx09-common.c hi = qnap_tsx09_parse_hex_nibble(b[0]); b 72 arch/arm/mach-orion5x/tsx09-common.c lo = qnap_tsx09_parse_hex_nibble(b[1]); b 216 arch/arm/mach-pxa/idp.c enum led_brightness b) b 222 arch/arm/mach-pxa/idp.c if (b != LED_OFF) b 567 arch/arm/mach-pxa/lubbock.c enum led_brightness b) b 573 arch/arm/mach-pxa/lubbock.c if (b != LED_OFF) b 655 arch/arm/mach-pxa/mainstone.c enum led_brightness b) b 661 arch/arm/mach-pxa/mainstone.c if (b != LED_OFF) b 75 arch/arm/mach-s3c24xx/include/mach/io.h DECLARE_IO(char,b,"b") b 1029 arch/arm/nwfpe/softfloat.c static float32 addFloat32Sigs( struct roundingData *roundData, float32 a, float32 b, flag zSign ) b 1037 arch/arm/nwfpe/softfloat.c bSig = extractFloat32Frac( b ); b 1038 arch/arm/nwfpe/softfloat.c bExp = extractFloat32Exp( b ); b 1044 arch/arm/nwfpe/softfloat.c if ( aSig ) return propagateFloat32NaN( a, b ); b 1058 arch/arm/nwfpe/softfloat.c if ( bSig ) return propagateFloat32NaN( a, b ); b 1072 arch/arm/nwfpe/softfloat.c if ( aSig | bSig ) return propagateFloat32NaN( a, b ); b 1101 arch/arm/nwfpe/softfloat.c static float32 subFloat32Sigs( struct roundingData *roundData, float32 a, float32 b, flag zSign ) b 1109 arch/arm/nwfpe/softfloat.c bSig = extractFloat32Frac( b ); b 1110 arch/arm/nwfpe/softfloat.c bExp = extractFloat32Exp( b ); b 1117 arch/arm/nwfpe/softfloat.c if ( aSig | bSig ) return propagateFloat32NaN( a, b ); b 1130 arch/arm/nwfpe/softfloat.c if ( bSig ) return propagateFloat32NaN( a, b ); b 1148 arch/arm/nwfpe/softfloat.c if ( aSig ) return propagateFloat32NaN( a, b ); b 1175 arch/arm/nwfpe/softfloat.c float32 float32_add( struct roundingData *roundData, float32 a, float32 b ) b 1180 arch/arm/nwfpe/softfloat.c bSign = extractFloat32Sign( b ); b 1182 arch/arm/nwfpe/softfloat.c return addFloat32Sigs( roundData, a, b, aSign ); b 1185 arch/arm/nwfpe/softfloat.c return subFloat32Sigs( roundData, a, b, aSign ); b 1197 arch/arm/nwfpe/softfloat.c float32 float32_sub( struct roundingData *roundData, float32 a, float32 b ) b 1202 arch/arm/nwfpe/softfloat.c bSign = extractFloat32Sign( b ); b 1204 arch/arm/nwfpe/softfloat.c return subFloat32Sigs( roundData, a, b, aSign ); b 1207 arch/arm/nwfpe/softfloat.c return addFloat32Sigs( roundData, a, b, aSign ); b 1219 arch/arm/nwfpe/softfloat.c float32 float32_mul( struct roundingData *roundData, float32 a, float32 b ) b 1230 arch/arm/nwfpe/softfloat.c bSig = extractFloat32Frac( b ); b 1231 arch/arm/nwfpe/softfloat.c bExp = extractFloat32Exp( b ); b 1232 arch/arm/nwfpe/softfloat.c bSign = extractFloat32Sign( b ); b 1236 arch/arm/nwfpe/softfloat.c return propagateFloat32NaN( a, b ); b 1245 arch/arm/nwfpe/softfloat.c if ( bSig ) return propagateFloat32NaN( a, b ); b 1280 arch/arm/nwfpe/softfloat.c float32 float32_div( struct roundingData *roundData, float32 a, float32 b ) b 1289 arch/arm/nwfpe/softfloat.c bSig = extractFloat32Frac( b ); b 1290 arch/arm/nwfpe/softfloat.c bExp = extractFloat32Exp( b ); b 1291 arch/arm/nwfpe/softfloat.c bSign = extractFloat32Sign( b ); b 1294 arch/arm/nwfpe/softfloat.c if ( aSig ) return propagateFloat32NaN( a, b ); b 1296 arch/arm/nwfpe/softfloat.c if ( bSig ) return propagateFloat32NaN( a, b ); b 1303 arch/arm/nwfpe/softfloat.c if ( bSig ) return propagateFloat32NaN( a, b ); b 1347 arch/arm/nwfpe/softfloat.c float32 float32_rem( struct roundingData *roundData, float32 a, float32 b ) b 1360 arch/arm/nwfpe/softfloat.c bSig = extractFloat32Frac( b ); b 1361 arch/arm/nwfpe/softfloat.c bExp = extractFloat32Exp( b ); b 1362 arch/arm/nwfpe/softfloat.c bSign = extractFloat32Sign( b ); b 1365 arch/arm/nwfpe/softfloat.c return propagateFloat32NaN( a, b ); b 1371 arch/arm/nwfpe/softfloat.c if ( bSig ) return propagateFloat32NaN( a, b ); b 1505 arch/arm/nwfpe/softfloat.c flag float32_eq( float32 a, float32 b ) b 1509 arch/arm/nwfpe/softfloat.c || ( ( extractFloat32Exp( b ) == 0xFF ) && extractFloat32Frac( b ) ) b 1511 arch/arm/nwfpe/softfloat.c if ( float32_is_signaling_nan( a ) || float32_is_signaling_nan( b ) ) { b 1516 arch/arm/nwfpe/softfloat.c return ( a == b ) || ( (bits32) ( ( a | b )<<1 ) == 0 ); b 1528 arch/arm/nwfpe/softfloat.c flag float32_le( float32 a, float32 b ) b 1533 arch/arm/nwfpe/softfloat.c || ( ( extractFloat32Exp( b ) == 0xFF ) && extractFloat32Frac( b ) ) b 1539 arch/arm/nwfpe/softfloat.c bSign = extractFloat32Sign( b ); b 1540 arch/arm/nwfpe/softfloat.c if ( aSign != bSign ) return aSign || ( (bits32) ( ( a | b )<<1 ) == 0 ); b 1541 arch/arm/nwfpe/softfloat.c return ( a == b ) || ( aSign ^ ( a < b ) ); b 1552 arch/arm/nwfpe/softfloat.c flag float32_lt( float32 a, float32 b ) b 1557 arch/arm/nwfpe/softfloat.c || ( ( extractFloat32Exp( b ) == 0xFF ) && extractFloat32Frac( b ) ) b 1563 arch/arm/nwfpe/softfloat.c bSign = extractFloat32Sign( b ); b 1564 arch/arm/nwfpe/softfloat.c if ( aSign != bSign ) return aSign && ( (bits32) ( ( a | b )<<1 ) != 0 ); b 1565 arch/arm/nwfpe/softfloat.c return ( a != b ) && ( aSign ^ ( a < b ) ); b 1577 arch/arm/nwfpe/softfloat.c flag float32_eq_signaling( float32 a, float32 b ) b 1581 arch/arm/nwfpe/softfloat.c || ( ( extractFloat32Exp( b ) == 0xFF ) && extractFloat32Frac( b ) ) b 1586 arch/arm/nwfpe/softfloat.c return ( a == b ) || ( (bits32) ( ( a | b )<<1 ) == 0 ); b 1598 arch/arm/nwfpe/softfloat.c flag float32_le_quiet( float32 a, float32 b ) b 1604 arch/arm/nwfpe/softfloat.c || ( ( extractFloat32Exp( b ) == 0xFF ) && extractFloat32Frac( b ) ) b 1610 arch/arm/nwfpe/softfloat.c bSign = extractFloat32Sign( b ); b 1611 arch/arm/nwfpe/softfloat.c if ( aSign != bSign ) return aSign || ( (bits32) ( ( a | b )<<1 ) == 0 ); b 1612 arch/arm/nwfpe/softfloat.c return ( a == b ) || ( aSign ^ ( a < b ) ); b 1624 arch/arm/nwfpe/softfloat.c flag float32_lt_quiet( float32 a, float32 b ) b 1629 arch/arm/nwfpe/softfloat.c || ( ( extractFloat32Exp( b ) == 0xFF ) && extractFloat32Frac( b ) ) b 1635 arch/arm/nwfpe/softfloat.c bSign = extractFloat32Sign( b ); b 1636 arch/arm/nwfpe/softfloat.c if ( aSign != bSign ) return aSign && ( (bits32) ( ( a | b )<<1 ) != 0 ); b 1637 arch/arm/nwfpe/softfloat.c return ( a != b ) && ( aSign ^ ( a < b ) ); b 1925 arch/arm/nwfpe/softfloat.c static float64 addFloat64Sigs( struct roundingData *roundData, float64 a, float64 b, flag zSign ) b 1933 arch/arm/nwfpe/softfloat.c bSig = extractFloat64Frac( b ); b 1934 arch/arm/nwfpe/softfloat.c bExp = extractFloat64Exp( b ); b 1940 arch/arm/nwfpe/softfloat.c if ( aSig ) return propagateFloat64NaN( a, b ); b 1954 arch/arm/nwfpe/softfloat.c if ( bSig ) return propagateFloat64NaN( a, b ); b 1968 arch/arm/nwfpe/softfloat.c if ( aSig | bSig ) return propagateFloat64NaN( a, b ); b 1997 arch/arm/nwfpe/softfloat.c static float64 subFloat64Sigs( struct roundingData *roundData, float64 a, float64 b, flag zSign ) b 2005 arch/arm/nwfpe/softfloat.c bSig = extractFloat64Frac( b ); b 2006 arch/arm/nwfpe/softfloat.c bExp = extractFloat64Exp( b ); b 2013 arch/arm/nwfpe/softfloat.c if ( aSig | bSig ) return propagateFloat64NaN( a, b ); b 2026 arch/arm/nwfpe/softfloat.c if ( bSig ) return propagateFloat64NaN( a, b ); b 2044 arch/arm/nwfpe/softfloat.c if ( aSig ) return propagateFloat64NaN( a, b ); b 2071 arch/arm/nwfpe/softfloat.c float64 float64_add( struct roundingData *roundData, float64 a, float64 b ) b 2076 arch/arm/nwfpe/softfloat.c bSign = extractFloat64Sign( b ); b 2078 arch/arm/nwfpe/softfloat.c return addFloat64Sigs( roundData, a, b, aSign ); b 2081 arch/arm/nwfpe/softfloat.c return subFloat64Sigs( roundData, a, b, aSign ); b 2093 arch/arm/nwfpe/softfloat.c float64 float64_sub( struct roundingData *roundData, float64 a, float64 b ) b 2098 arch/arm/nwfpe/softfloat.c bSign = extractFloat64Sign( b ); b 2100 arch/arm/nwfpe/softfloat.c return subFloat64Sigs( roundData, a, b, aSign ); b 2103 arch/arm/nwfpe/softfloat.c return addFloat64Sigs( roundData, a, b, aSign ); b 2115 arch/arm/nwfpe/softfloat.c float64 float64_mul( struct roundingData *roundData, float64 a, float64 b ) b 2124 arch/arm/nwfpe/softfloat.c bSig = extractFloat64Frac( b ); b 2125 arch/arm/nwfpe/softfloat.c bExp = extractFloat64Exp( b ); b 2126 arch/arm/nwfpe/softfloat.c bSign = extractFloat64Sign( b ); b 2130 arch/arm/nwfpe/softfloat.c return propagateFloat64NaN( a, b ); b 2139 arch/arm/nwfpe/softfloat.c if ( bSig ) return propagateFloat64NaN( a, b ); b 2174 arch/arm/nwfpe/softfloat.c float64 float64_div( struct roundingData *roundData, float64 a, float64 b ) b 2185 arch/arm/nwfpe/softfloat.c bSig = extractFloat64Frac( b ); b 2186 arch/arm/nwfpe/softfloat.c bExp = extractFloat64Exp( b ); b 2187 arch/arm/nwfpe/softfloat.c bSign = extractFloat64Sign( b ); b 2190 arch/arm/nwfpe/softfloat.c if ( aSig ) return propagateFloat64NaN( a, b ); b 2192 arch/arm/nwfpe/softfloat.c if ( bSig ) return propagateFloat64NaN( a, b ); b 2199 arch/arm/nwfpe/softfloat.c if ( bSig ) return propagateFloat64NaN( a, b ); b 2245 arch/arm/nwfpe/softfloat.c float64 float64_rem( struct roundingData *roundData, float64 a, float64 b ) b 2256 arch/arm/nwfpe/softfloat.c bSig = extractFloat64Frac( b ); b 2257 arch/arm/nwfpe/softfloat.c bExp = extractFloat64Exp( b ); b 2258 arch/arm/nwfpe/softfloat.c bSign = extractFloat64Sign( b ); b 2261 arch/arm/nwfpe/softfloat.c return propagateFloat64NaN( a, b ); b 2267 arch/arm/nwfpe/softfloat.c if ( bSig ) return propagateFloat64NaN( a, b ); b 2392 arch/arm/nwfpe/softfloat.c flag float64_eq( float64 a, float64 b ) b 2396 arch/arm/nwfpe/softfloat.c || ( ( extractFloat64Exp( b ) == 0x7FF ) && extractFloat64Frac( b ) ) b 2398 arch/arm/nwfpe/softfloat.c if ( float64_is_signaling_nan( a ) || float64_is_signaling_nan( b ) ) { b 2403 arch/arm/nwfpe/softfloat.c return ( a == b ) || ( (bits64) ( ( a | b )<<1 ) == 0 ); b 2415 arch/arm/nwfpe/softfloat.c flag float64_le( float64 a, float64 b ) b 2420 arch/arm/nwfpe/softfloat.c || ( ( extractFloat64Exp( b ) == 0x7FF ) && extractFloat64Frac( b ) ) b 2426 arch/arm/nwfpe/softfloat.c bSign = extractFloat64Sign( b ); b 2427 arch/arm/nwfpe/softfloat.c if ( aSign != bSign ) return aSign || ( (bits64) ( ( a | b )<<1 ) == 0 ); b 2428 arch/arm/nwfpe/softfloat.c return ( a == b ) || ( aSign ^ ( a < b ) ); b 2439 arch/arm/nwfpe/softfloat.c flag float64_lt( float64 a, float64 b ) b 2444 arch/arm/nwfpe/softfloat.c || ( ( extractFloat64Exp( b ) == 0x7FF ) && extractFloat64Frac( b ) ) b 2450 arch/arm/nwfpe/softfloat.c bSign = extractFloat64Sign( b ); b 2451 arch/arm/nwfpe/softfloat.c if ( aSign != bSign ) return aSign && ( (bits64) ( ( a | b )<<1 ) != 0 ); b 2452 arch/arm/nwfpe/softfloat.c return ( a != b ) && ( aSign ^ ( a < b ) ); b 2464 arch/arm/nwfpe/softfloat.c flag float64_eq_signaling( float64 a, float64 b ) b 2468 arch/arm/nwfpe/softfloat.c || ( ( extractFloat64Exp( b ) == 0x7FF ) && extractFloat64Frac( b ) ) b 2473 arch/arm/nwfpe/softfloat.c return ( a == b ) || ( (bits64) ( ( a | b )<<1 ) == 0 ); b 2485 arch/arm/nwfpe/softfloat.c flag float64_le_quiet( float64 a, float64 b ) b 2491 arch/arm/nwfpe/softfloat.c || ( ( extractFloat64Exp( b ) == 0x7FF ) && extractFloat64Frac( b ) ) b 2497 arch/arm/nwfpe/softfloat.c bSign = extractFloat64Sign( b ); b 2498 arch/arm/nwfpe/softfloat.c if ( aSign != bSign ) return aSign || ( (bits64) ( ( a | b )<<1 ) == 0 ); b 2499 arch/arm/nwfpe/softfloat.c return ( a == b ) || ( aSign ^ ( a < b ) ); b 2511 arch/arm/nwfpe/softfloat.c flag float64_lt_quiet( float64 a, float64 b ) b 2516 arch/arm/nwfpe/softfloat.c || ( ( extractFloat64Exp( b ) == 0x7FF ) && extractFloat64Frac( b ) ) b 2522 arch/arm/nwfpe/softfloat.c bSign = extractFloat64Sign( b ); b 2523 arch/arm/nwfpe/softfloat.c if ( aSign != bSign ) return aSign && ( (bits64) ( ( a | b )<<1 ) != 0 ); b 2524 arch/arm/nwfpe/softfloat.c return ( a != b ) && ( aSign ^ ( a < b ) ); b 2745 arch/arm/nwfpe/softfloat.c static floatx80 addFloatx80Sigs( struct roundingData *roundData, floatx80 a, floatx80 b, flag zSign ) b 2753 arch/arm/nwfpe/softfloat.c bSig = extractFloatx80Frac( b ); b 2754 arch/arm/nwfpe/softfloat.c bExp = extractFloatx80Exp( b ); b 2758 arch/arm/nwfpe/softfloat.c if ( (bits64) ( aSig<<1 ) ) return propagateFloatx80NaN( a, b ); b 2767 arch/arm/nwfpe/softfloat.c if ( (bits64) ( bSig<<1 ) ) return propagateFloatx80NaN( a, b ); b 2777 arch/arm/nwfpe/softfloat.c return propagateFloatx80NaN( a, b ); b 2814 arch/arm/nwfpe/softfloat.c static floatx80 subFloatx80Sigs( struct roundingData *roundData, floatx80 a, floatx80 b, flag zSign ) b 2823 arch/arm/nwfpe/softfloat.c bSig = extractFloatx80Frac( b ); b 2824 arch/arm/nwfpe/softfloat.c bExp = extractFloatx80Exp( b ); b 2830 arch/arm/nwfpe/softfloat.c return propagateFloatx80NaN( a, b ); b 2848 arch/arm/nwfpe/softfloat.c if ( (bits64) ( bSig<<1 ) ) return propagateFloatx80NaN( a, b ); b 2860 arch/arm/nwfpe/softfloat.c if ( (bits64) ( aSig<<1 ) ) return propagateFloatx80NaN( a, b ); b 2882 arch/arm/nwfpe/softfloat.c floatx80 floatx80_add( struct roundingData *roundData, floatx80 a, floatx80 b ) b 2887 arch/arm/nwfpe/softfloat.c bSign = extractFloatx80Sign( b ); b 2889 arch/arm/nwfpe/softfloat.c return addFloatx80Sigs( roundData, a, b, aSign ); b 2892 arch/arm/nwfpe/softfloat.c return subFloatx80Sigs( roundData, a, b, aSign ); b 2904 arch/arm/nwfpe/softfloat.c floatx80 floatx80_sub( struct roundingData *roundData, floatx80 a, floatx80 b ) b 2909 arch/arm/nwfpe/softfloat.c bSign = extractFloatx80Sign( b ); b 2911 arch/arm/nwfpe/softfloat.c return subFloatx80Sigs( roundData, a, b, aSign ); b 2914 arch/arm/nwfpe/softfloat.c return addFloatx80Sigs( roundData, a, b, aSign ); b 2926 arch/arm/nwfpe/softfloat.c floatx80 floatx80_mul( struct roundingData *roundData, floatx80 a, floatx80 b ) b 2936 arch/arm/nwfpe/softfloat.c bSig = extractFloatx80Frac( b ); b 2937 arch/arm/nwfpe/softfloat.c bExp = extractFloatx80Exp( b ); b 2938 arch/arm/nwfpe/softfloat.c bSign = extractFloatx80Sign( b ); b 2943 arch/arm/nwfpe/softfloat.c return propagateFloatx80NaN( a, b ); b 2949 arch/arm/nwfpe/softfloat.c if ( (bits64) ( bSig<<1 ) ) return propagateFloatx80NaN( a, b ); b 2987 arch/arm/nwfpe/softfloat.c floatx80 floatx80_div( struct roundingData *roundData, floatx80 a, floatx80 b ) b 2998 arch/arm/nwfpe/softfloat.c bSig = extractFloatx80Frac( b ); b 2999 arch/arm/nwfpe/softfloat.c bExp = extractFloatx80Exp( b ); b 3000 arch/arm/nwfpe/softfloat.c bSign = extractFloatx80Sign( b ); b 3003 arch/arm/nwfpe/softfloat.c if ( (bits64) ( aSig<<1 ) ) return propagateFloatx80NaN( a, b ); b 3005 arch/arm/nwfpe/softfloat.c if ( (bits64) ( bSig<<1 ) ) return propagateFloatx80NaN( a, b ); b 3011 arch/arm/nwfpe/softfloat.c if ( (bits64) ( bSig<<1 ) ) return propagateFloatx80NaN( a, b ); b 3069 arch/arm/nwfpe/softfloat.c floatx80 floatx80_rem( struct roundingData *roundData, floatx80 a, floatx80 b ) b 3080 arch/arm/nwfpe/softfloat.c bSig = extractFloatx80Frac( b ); b 3081 arch/arm/nwfpe/softfloat.c bExp = extractFloatx80Exp( b ); b 3082 arch/arm/nwfpe/softfloat.c bSign = extractFloatx80Sign( b ); b 3086 arch/arm/nwfpe/softfloat.c return propagateFloatx80NaN( a, b ); b 3091 arch/arm/nwfpe/softfloat.c if ( (bits64) ( bSig<<1 ) ) return propagateFloatx80NaN( a, b ); b 3247 arch/arm/nwfpe/softfloat.c flag floatx80_eq( floatx80 a, floatx80 b ) b 3252 arch/arm/nwfpe/softfloat.c || ( ( extractFloatx80Exp( b ) == 0x7FFF ) b 3253 arch/arm/nwfpe/softfloat.c && (bits64) ( extractFloatx80Frac( b )<<1 ) ) b 3256 arch/arm/nwfpe/softfloat.c || floatx80_is_signaling_nan( b ) ) { b 3262 arch/arm/nwfpe/softfloat.c ( a.low == b.low ) b 3263 arch/arm/nwfpe/softfloat.c && ( ( a.high == b.high ) b 3265 arch/arm/nwfpe/softfloat.c && ( (bits16) ( ( a.high | b.high )<<1 ) == 0 ) ) b 3278 arch/arm/nwfpe/softfloat.c flag floatx80_le( floatx80 a, floatx80 b ) b 3284 arch/arm/nwfpe/softfloat.c || ( ( extractFloatx80Exp( b ) == 0x7FFF ) b 3285 arch/arm/nwfpe/softfloat.c && (bits64) ( extractFloatx80Frac( b )<<1 ) ) b 3291 arch/arm/nwfpe/softfloat.c bSign = extractFloatx80Sign( b ); b 3295 arch/arm/nwfpe/softfloat.c || ( ( ( (bits16) ( ( a.high | b.high )<<1 ) ) | a.low | b.low ) b 3299 arch/arm/nwfpe/softfloat.c aSign ? le128( b.high, b.low, a.high, a.low ) b 3300 arch/arm/nwfpe/softfloat.c : le128( a.high, a.low, b.high, b.low ); b 3312 arch/arm/nwfpe/softfloat.c flag floatx80_lt( floatx80 a, floatx80 b ) b 3318 arch/arm/nwfpe/softfloat.c || ( ( extractFloatx80Exp( b ) == 0x7FFF ) b 3319 arch/arm/nwfpe/softfloat.c && (bits64) ( extractFloatx80Frac( b )<<1 ) ) b 3325 arch/arm/nwfpe/softfloat.c bSign = extractFloatx80Sign( b ); b 3329 arch/arm/nwfpe/softfloat.c && ( ( ( (bits16) ( ( a.high | b.high )<<1 ) ) | a.low | b.low ) b 3333 arch/arm/nwfpe/softfloat.c aSign ? lt128( b.high, b.low, a.high, a.low ) b 3334 arch/arm/nwfpe/softfloat.c : lt128( a.high, a.low, b.high, b.low ); b 3346 arch/arm/nwfpe/softfloat.c flag floatx80_eq_signaling( floatx80 a, floatx80 b ) b 3351 arch/arm/nwfpe/softfloat.c || ( ( extractFloatx80Exp( b ) == 0x7FFF ) b 3352 arch/arm/nwfpe/softfloat.c && (bits64) ( extractFloatx80Frac( b )<<1 ) ) b 3358 arch/arm/nwfpe/softfloat.c ( a.low == b.low ) b 3359 arch/arm/nwfpe/softfloat.c && ( ( a.high == b.high ) b 3361 arch/arm/nwfpe/softfloat.c && ( (bits16) ( ( a.high | b.high )<<1 ) == 0 ) ) b 3374 arch/arm/nwfpe/softfloat.c flag floatx80_le_quiet( floatx80 a, floatx80 b ) b 3380 arch/arm/nwfpe/softfloat.c || ( ( extractFloatx80Exp( b ) == 0x7FFF ) b 3381 arch/arm/nwfpe/softfloat.c && (bits64) ( extractFloatx80Frac( b )<<1 ) ) b 3387 arch/arm/nwfpe/softfloat.c bSign = extractFloatx80Sign( b ); b 3391 arch/arm/nwfpe/softfloat.c || ( ( ( (bits16) ( ( a.high | b.high )<<1 ) ) | a.low | b.low ) b 3395 arch/arm/nwfpe/softfloat.c aSign ? le128( b.high, b.low, a.high, a.low ) b 3396 arch/arm/nwfpe/softfloat.c : le128( a.high, a.low, b.high, b.low ); b 3408 arch/arm/nwfpe/softfloat.c flag floatx80_lt_quiet( floatx80 a, floatx80 b ) b 3414 arch/arm/nwfpe/softfloat.c || ( ( extractFloatx80Exp( b ) == 0x7FFF ) b 3415 arch/arm/nwfpe/softfloat.c && (bits64) ( extractFloatx80Frac( b )<<1 ) ) b 3421 arch/arm/nwfpe/softfloat.c bSign = extractFloatx80Sign( b ); b 3425 arch/arm/nwfpe/softfloat.c && ( ( ( (bits16) ( ( a.high | b.high )<<1 ) ) | a.low | b.low ) b 3429 arch/arm/nwfpe/softfloat.c aSign ? lt128( b.high, b.low, a.high, a.low ) b 3430 arch/arm/nwfpe/softfloat.c : lt128( a.high, a.low, b.high, b.low ); b 238 arch/arm/nwfpe/softfloat.h static inline flag float32_eq_nocheck(float32 a, float32 b) b 240 arch/arm/nwfpe/softfloat.h return (a == b) || ((bits32) ((a | b) << 1) == 0); b 243 arch/arm/nwfpe/softfloat.h static inline flag float32_lt_nocheck(float32 a, float32 b) b 248 arch/arm/nwfpe/softfloat.h bSign = extractFloat32Sign(b); b 250 arch/arm/nwfpe/softfloat.h return aSign && ((bits32) ((a | b) << 1) != 0); b 251 arch/arm/nwfpe/softfloat.h return (a != b) && (aSign ^ (a < b)); b 259 arch/arm/nwfpe/softfloat.h static inline flag float64_eq_nocheck(float64 a, float64 b) b 261 arch/arm/nwfpe/softfloat.h return (a == b) || ((bits64) ((a | b) << 1) == 0); b 264 arch/arm/nwfpe/softfloat.h static inline flag float64_lt_nocheck(float64 a, float64 b) b 269 arch/arm/nwfpe/softfloat.h bSign = extractFloat64Sign(b); b 271 arch/arm/nwfpe/softfloat.h return aSign && ((bits64) ((a | b) << 1) != 0); b 272 arch/arm/nwfpe/softfloat.h return (a != b) && (aSign ^ (a < b)); b 80 arch/arm/plat-omap/debug-leds.c enum led_brightness b) b 86 arch/arm/plat-omap/debug-leds.c if (b != LED_OFF) b 38 arch/arm/plat-orion/pcie.c #define PCIE_CONF_BUS(b) (((b) & 0xff) << 16) b 568 arch/arm/probes/kprobes/test-core.c struct benchmarks *b; b 569 arch/arm/probes/kprobes/test-core.c for (b = list; b->fn; ++b) { b 570 arch/arm/probes/kprobes/test-core.c ret = kprobe_benchmark(b->fn, b->offset); b 573 arch/arm/probes/kprobes/test-core.c pr_info(" %dns for kprobe %s\n", ret, b->title); b 24 arch/arm64/crypto/aes-ce-glue.c u8 b[AES_BLOCK_SIZE]; b 806 arch/arm64/crypto/aes-glue.c u64 b = be64_to_cpu(x->b); b 808 arch/arm64/crypto/aes-glue.c y->a = cpu_to_be64((a << 1) | (b >> 63)); b 809 arch/arm64/crypto/aes-glue.c y->b = cpu_to_be64((b << 1) ^ ((a >> 63) ? 0x87 : 0)); b 110 arch/arm64/crypto/ghash-ce-glue.c dg[0] = be64_to_cpu(dst.b); b 218 arch/arm64/crypto/ghash-ce-glue.c h[0] = (be64_to_cpu(k->b) << 1) | carry; b 219 arch/arm64/crypto/ghash-ce-glue.c h[1] = (be64_to_cpu(k->a) << 1) | (be64_to_cpu(k->b) >> 63); b 408 arch/arm64/crypto/ghash-ce-glue.c lengths.b = cpu_to_be64(cryptlen * 8); b 171 arch/arm64/include/asm/alternative.h altinstruction_entry 661f, \cb, ARM64_CB_PATCH, 662f-661f, 0 b 22 arch/arm64/include/asm/asm-prototypes.h long long __ashlti3(long long a, int b); b 23 arch/arm64/include/asm/asm-prototypes.h long long __ashrti3(long long a, int b); b 24 arch/arm64/include/asm/asm-prototypes.h long long __lshrti3(long long a, int b); b 159 arch/arm64/include/asm/assembler.h b \label b 409 arch/arm64/include/asm/assembler.h b.lo 9998b b 429 arch/arm64/include/asm/assembler.h b.lo 9997b b 441 arch/arm64/include/asm/assembler.h b.lt 9000f b 461 arch/arm64/include/asm/assembler.h b.ne 9998b b 608 arch/arm64/include/asm/assembler.h b.ne 10f b 751 arch/arm64/include/asm/assembler.h b \lbl b 753 arch/arm64/include/asm/assembler.h b .Lyield_out_\@ b 299 arch/arm64/include/asm/atomic_ll_sc.h __CMPXCHG_CASE(w, b, , 8, , , , , K) b 303 arch/arm64/include/asm/atomic_ll_sc.h __CMPXCHG_CASE(w, b, acq_, 8, , a, , "memory", K) b 307 arch/arm64/include/asm/atomic_ll_sc.h __CMPXCHG_CASE(w, b, rel_, 8, , , l, "memory", K) b 311 arch/arm64/include/asm/atomic_ll_sc.h __CMPXCHG_CASE(w, b, mb_, 8, dmb ish, , l, "memory", K) b 364 arch/arm64/include/asm/atomic_lse.h __CMPXCHG_CASE(w, b, , 8, ) b 368 arch/arm64/include/asm/atomic_lse.h __CMPXCHG_CASE(w, b, acq_, 8, a, "memory") b 372 arch/arm64/include/asm/atomic_lse.h __CMPXCHG_CASE(w, b, rel_, 8, l, "memory") b 376 arch/arm64/include/asm/atomic_lse.h __CMPXCHG_CASE(w, b, mb_, 8, al, "memory") b 45 arch/arm64/include/asm/cmpxchg.h __XCHG_CASE(w, b, , 8, , , , , , ) b 49 arch/arm64/include/asm/cmpxchg.h __XCHG_CASE(w, b, acq_, 8, , , a, a, , "memory") b 53 arch/arm64/include/asm/cmpxchg.h __XCHG_CASE(w, b, rel_, 8, , , , , l, "memory") b 57 arch/arm64/include/asm/cmpxchg.h __XCHG_CASE(w, b, mb_, 8, dmb ish, nop, , a, l, "memory") b 250 arch/arm64/include/asm/cmpxchg.h __CMPWAIT_CASE(w, b, 8); b 39 arch/arm64/include/asm/fpsimdmacros.h b.eq 9999f b 204 arch/arm64/include/asm/fpsimdmacros.h b.eq 921f b 324 arch/arm64/include/asm/insn.h __AARCH64_INSN_FUNCS(b, 0xFC000000, 0x14000000) b 71 arch/arm64/include/asm/kvm_ptrauth.h b 1000f b 74 arch/arm64/include/asm/kvm_ptrauth.h b 1001f b 87 arch/arm64/include/asm/kvm_ptrauth.h b 2000f b 90 arch/arm64/include/asm/kvm_ptrauth.h b 2001f b 63 arch/arm64/include/asm/module.h bool plt_entries_equal(const struct plt_entry *a, const struct plt_entry *b); b 12 arch/arm64/include/asm/numa.h #define node_distance(a, b) __node_distance(a, b) b 97 arch/arm64/include/asm/percpu.h __PERCPU_OP_CASE(w, b, name, 8, op_llsc, op_lse) \ b 103 arch/arm64/include/asm/percpu.h __PERCPU_RET_OP_CASE(w, b, name, 8, op_llsc, op_lse) \ b 52 arch/arm64/include/asm/uaccess.h #define segment_eq(a, b) ((a) == (b)) b 398 arch/arm64/kernel/fpsimd.c u64 b = swab64(x >> 64); b 400 arch/arm64/kernel/fpsimd.c return ((__uint128_t)a << 64) | b; b 737 arch/arm64/kernel/fpsimd.c unsigned long b; b 763 arch/arm64/kernel/fpsimd.c b = find_last_bit(tmp_map, SVE_VQ_MAX); b 764 arch/arm64/kernel/fpsimd.c if (b >= SVE_VQ_MAX) b 771 arch/arm64/kernel/fpsimd.c if (sve_vl_from_vq(__bit_to_vq(b)) <= sve_max_virtualisable_vl) { b 845 arch/arm64/kernel/fpsimd.c unsigned long b; b 877 arch/arm64/kernel/fpsimd.c b = find_last_bit(tmp_map, SVE_VQ_MAX); b 878 arch/arm64/kernel/fpsimd.c if (b >= SVE_VQ_MAX) b 881 arch/arm64/kernel/fpsimd.c else if (WARN_ON(b == SVE_VQ_MAX - 1)) b 885 arch/arm64/kernel/fpsimd.c sve_max_virtualisable_vl = sve_vl_from_vq(__bit_to_vq(b + 1)); b 39 arch/arm64/kernel/module-plts.c bool plt_entries_equal(const struct plt_entry *a, const struct plt_entry *b) b 49 arch/arm64/kernel/module-plts.c if (a->add != b->add || a->br != b->br) b 53 arch/arm64/kernel/module-plts.c q = ALIGN_DOWN((u64)b, SZ_4K); b 59 arch/arm64/kernel/module-plts.c if (a->adrp == b->adrp && p == q) b 63 arch/arm64/kernel/module-plts.c (q + aarch64_insn_adrp_get_offset(le32_to_cpu(b->adrp))); b 133 arch/arm64/kernel/module-plts.c #define cmp_3way(a,b) ((a) < (b) ? -1 : (a) > (b)) b 135 arch/arm64/kernel/module-plts.c static int cmp_rela(const void *a, const void *b) b 137 arch/arm64/kernel/module-plts.c const Elf64_Rela *x = a, *y = b; b 40 arch/arm64/kernel/pci.c struct pci_bus *b = pci_find_bus(domain, bus); b 42 arch/arm64/kernel/pci.c if (!b) b 44 arch/arm64/kernel/pci.c return b->ops->read(b, devfn, reg, len, val); b 50 arch/arm64/kernel/pci.c struct pci_bus *b = pci_find_bus(domain, bus); b 52 arch/arm64/kernel/pci.c if (!b) b 54 arch/arm64/kernel/pci.c return b->ops->write(b, devfn, reg, len, val); b 29 arch/c6x/include/asm/pgtable.h #define pmd_offset(a, b) ((void *)0) b 33 arch/c6x/include/asm/soc.h #define soc_writel(b, addr) __raw_writel((b), (addr)) b 137 arch/csky/abiv2/fpu.c #define STW_FPU_REGS(a, b, c, d) \ b 139 arch/csky/abiv2/fpu.c "stw %1, (%4, "#b")\n" \ b 143 arch/csky/abiv2/fpu.c #define LDW_FPU_REGS(a, b, c, d) \ b 145 arch/csky/abiv2/fpu.c "ldw %1, (%4, "#b")\n" \ b 16 arch/csky/include/asm/segment.h #define segment_eq(a, b) ((a).seg == (b).seg) b 30 arch/h8300/include/asm/io.h static inline void __raw_writeb(u8 b, const volatile void __iomem *addr) b 32 arch/h8300/include/asm/io.h *(volatile u8 *)addr = b; b 36 arch/h8300/include/asm/io.h static inline void __raw_writew(u16 b, const volatile void __iomem *addr) b 38 arch/h8300/include/asm/io.h *(volatile u16 *)addr = b; b 42 arch/h8300/include/asm/io.h static inline void __raw_writel(u32 b, const volatile void __iomem *addr) b 44 arch/h8300/include/asm/io.h *(volatile u32 *)addr = b; b 47 arch/h8300/include/asm/io.h static inline void ctrl_bclr(int b, void __iomem *addr) b 49 arch/h8300/include/asm/io.h if (__builtin_constant_p(b)) b 50 arch/h8300/include/asm/io.h __asm__("bclr %1,%0" : "+WU"(*(u8 *)addr): "i"(b)); b 52 arch/h8300/include/asm/io.h __asm__("bclr %w1,%0" : "+WU"(*(u8 *)addr): "r"(b)); b 55 arch/h8300/include/asm/io.h static inline void ctrl_bset(int b, void __iomem *addr) b 57 arch/h8300/include/asm/io.h if (__builtin_constant_p(b)) b 58 arch/h8300/include/asm/io.h __asm__("bset %1,%0" : "+WU"(*(u8 *)addr): "i"(b)); b 60 arch/h8300/include/asm/io.h __asm__("bset %w1,%0" : "+WU"(*(u8 *)addr): "r"(b)); b 36 arch/h8300/include/asm/segment.h #define segment_eq(a, b) ((a).seg == (b).seg) b 5 arch/h8300/lib/ashldi3.c __ashldi3(DWtype u, word_type b) b 8 arch/h8300/lib/ashldi3.c const word_type bm = (sizeof (Wtype) * BITS_PER_UNIT) - b; b 11 arch/h8300/lib/ashldi3.c if (b == 0) b 20 arch/h8300/lib/ashldi3.c w.s.low = (UWtype) uu.s.low << b; b 21 arch/h8300/lib/ashldi3.c w.s.high = ((UWtype) uu.s.high << b) | carries; b 4 arch/h8300/lib/ashrdi3.c DWtype __ashrdi3(DWtype u, word_type b) b 7 arch/h8300/lib/ashrdi3.c const word_type bm = (sizeof (Wtype) * BITS_PER_UNIT) - b; b 10 arch/h8300/lib/ashrdi3.c if (b == 0) b 20 arch/h8300/lib/ashrdi3.c w.s.high = uu.s.high >> b; b 21 arch/h8300/lib/ashrdi3.c w.s.low = ((UWtype) uu.s.low >> b) | carries; b 4 arch/h8300/lib/lshrdi3.c DWtype __lshrdi3(DWtype u, word_type b) b 7 arch/h8300/lib/lshrdi3.c const word_type bm = (sizeof (Wtype) * BITS_PER_UNIT) - b; b 10 arch/h8300/lib/lshrdi3.c if (b == 0) b 19 arch/h8300/lib/lshrdi3.c w.s.high = (UWtype) uu.s.high >> b; b 20 arch/h8300/lib/lshrdi3.c w.s.low = ((UWtype) uu.s.low >> b) | carries; b 4 arch/h8300/lib/ucmpdi2.c word_type __ucmpdi2(DWtype a, DWtype b) b 7 arch/h8300/lib/ucmpdi2.c const DWunion bu = {.ll = b}; b 25 arch/hexagon/lib/checksum.c #define VR_NEGATE(a, b, c, d) (SIGN(a, 48) + SIGN(b, 32) + SIGN(c, 16) \ b 27 arch/hexagon/lib/checksum.c #define VR_CARRY(a, b, c, d) (CARRY(a, 48) + CARRY(b, 32) + CARRY(c, 16) \ b 29 arch/hexagon/lib/checksum.c #define VR_SELECT(a, b, c, d) (SELECT(a, 48) + SELECT(b, 32) + SELECT(c, 16) \ b 2099 arch/ia64/hp/common/sba_iommu.c struct pci_bus *b = NULL; b 2100 arch/ia64/hp/common/sba_iommu.c while ((b = pci_find_next_bus(b)) != NULL) b 2101 arch/ia64/hp/common/sba_iommu.c sba_connect_bus(b); b 454 arch/ia64/include/asm/pgtable.h pte_same (pte_t a, pte_t b) b 456 arch/ia64/include/asm/pgtable.h return pte_val(a) == pte_val(b); b 54 arch/ia64/include/asm/uaccess.h #define segment_eq(a, b) ((a).seg == (b).seg) b 162 arch/ia64/include/uapi/asm/gcc_intrin.h #define ia64_shrp(a, b, count) \ b 165 arch/ia64/include/uapi/asm/gcc_intrin.h asm ("shrp %0=%1,%2,%3" : "=r"(ia64_intri_res) : "r"(a), "r"(b), "i"(count)); \ b 247 arch/ia64/kernel/module.c uint64_t b0, b1, *b = (uint64_t *) plt->bundle[1]; b 250 arch/ia64/kernel/module.c b0 = b[0]; b1 = b[1]; b 296 arch/ia64/kernel/module.c uint64_t b0, b1, *b = (uint64_t *) plt->bundle[0]; b 298 arch/ia64/kernel/module.c b0 = b[0]; b1 = b[1]; b 47 arch/ia64/kernel/patch.c u64 m0, m1, v0, v1, b0, b1, *b = (u64 *) (insn_addr & -16); b 51 arch/ia64/kernel/patch.c b0 = b[0]; b1 = b[1]; b 59 arch/ia64/kernel/patch.c b[0] = (b0 & ~m0) | (v0 & m0); b 61 arch/ia64/kernel/patch.c b[1] = (b1 & ~m1) | (v1 & m1); b 127 arch/ia64/kernel/patch.c u64 ip, *b; b 132 arch/ia64/kernel/patch.c b = (u64 *)(ip & -16); b 133 arch/ia64/kernel/patch.c b[1] &= ~0xf800000L; b 580 arch/ia64/kernel/perfmon.c #define pfm_get_cpu_data(a,b) per_cpu(a, b) b 1145 arch/ia64/kernel/perfmon.c pfm_uuid_cmp(pfm_uuid_t a, pfm_uuid_t b) b 1147 arch/ia64/kernel/perfmon.c return memcmp(a, b, sizeof(pfm_uuid_t)); b 1156 arch/ia64/kernel/unwind.c #define UNW_DEC_PROLOGUE(fmt,b,r,arg) desc_prologue(b,r,0,32,arg) b 1161 arch/ia64/kernel/unwind.c #define UNW_DEC_BR_GR(fmt,b,g,arg) desc_br_gr(b,g,arg) b 1162 arch/ia64/kernel/unwind.c #define UNW_DEC_BR_MEM(fmt,b,arg) desc_br_mem(b,arg) b 364 arch/ia64/pci/pci.c void pcibios_fixup_bus(struct pci_bus *b) b 368 arch/ia64/pci/pci.c if (b->self) { b 369 arch/ia64/pci/pci.c pci_read_bridge_bases(b); b 370 arch/ia64/pci/pci.c pcibios_fixup_bridge_resources(b->self); b 372 arch/ia64/pci/pci.c list_for_each_entry(dev, &b->devices, bus_list) b 102 arch/m68k/amiga/config.c static void amiga_mem_console_write(struct console *co, const char *b, b 29 arch/m68k/amiga/pcmcia.c unsigned char b; b 33 arch/m68k/amiga/pcmcia.c b = gayle_reset; b 985 arch/m68k/include/asm/MC68328.h } b; b 992 arch/m68k/include/asm/MC68328.h } b; b 807 arch/m68k/include/asm/MC68EZ328.h } b; b 814 arch/m68k/include/asm/MC68EZ328.h } b; b 900 arch/m68k/include/asm/MC68VZ328.h } b; b 907 arch/m68k/include/asm/MC68VZ328.h } b; b 506 arch/m68k/include/asm/atarihw.h u_char b[4]; b 72 arch/m68k/include/asm/bootstd.h #define _bsc2(type,name,atype,a,btype,b) \ b 73 arch/m68k/include/asm/bootstd.h type name(atype a, btype b) \ b 77 arch/m68k/include/asm/bootstd.h register long __b __asm__ ("%d2") = (long)b; \ b 85 arch/m68k/include/asm/bootstd.h #define _bsc3(type,name,atype,a,btype,b,ctype,c) \ b 86 arch/m68k/include/asm/bootstd.h type name(atype a, btype b, ctype c) \ b 90 arch/m68k/include/asm/bootstd.h register long __b __asm__ ("%d2") = (long)b; \ b 100 arch/m68k/include/asm/bootstd.h #define _bsc4(type,name,atype,a,btype,b,ctype,c,dtype,d) \ b 101 arch/m68k/include/asm/bootstd.h type name(atype a, btype b, ctype c, dtype d) \ b 105 arch/m68k/include/asm/bootstd.h register long __b __asm__ ("%d2") = (long)b; \ b 116 arch/m68k/include/asm/bootstd.h #define _bsc5(type,name,atype,a,btype,b,ctype,c,dtype,d,etype,e) \ b 117 arch/m68k/include/asm/bootstd.h type name(atype a, btype b, ctype c, dtype d, etype e) \ b 121 arch/m68k/include/asm/bootstd.h register long __b __asm__ ("%d2") = (long)b; \ b 43 arch/m68k/include/asm/hash.h u32 a, b; b 54 arch/m68k/include/asm/hash.h : "=&d,d" (a), "=&r,r" (b) b 57 arch/m68k/include/asm/hash.h return ((u16)(x*0x61c8) << 16) + a + b; b 409 arch/m68k/include/asm/io_mm.h #define writeb_relaxed(b, addr) writeb(b, addr) b 410 arch/m68k/include/asm/io_mm.h #define writew_relaxed(b, addr) writew(b, addr) b 411 arch/m68k/include/asm/io_mm.h #define writel_relaxed(b, addr) writel(b, addr) b 23 arch/m68k/include/asm/io_no.h #define __raw_writeb(b, addr) (void)((*(volatile unsigned char *) (addr)) = (b)) b 24 arch/m68k/include/asm/io_no.h #define __raw_writew(b, addr) (void)((*(volatile unsigned short *) (addr)) = (b)) b 25 arch/m68k/include/asm/io_no.h #define __raw_writel(b, addr) (void)((*(volatile unsigned int *) (addr)) = (b)) b 92 arch/m68k/include/asm/m525xsim.h #define MCFINTC2_INTPRI_BITS(b, i) ((b) << (((i) % 8) * 4)) b 134 arch/m68k/include/asm/mac_iop.h } b; b 250 arch/m68k/include/asm/mac_via.h #define IER_SET_BIT(b) (0x80 | (1<<(b)) ) b 251 arch/m68k/include/asm/mac_via.h #define IER_CLR_BIT(b) (0x7F & (1<<(b)) ) b 236 arch/m68k/include/asm/math-emu.h .irp m b,w,l b 16 arch/m68k/include/asm/nubus.h #define nubus_memset_io(a,b,c) memset((void *)(a),(b),(c)) b 17 arch/m68k/include/asm/nubus.h #define nubus_memcpy_fromio(a,b,c) memcpy((a),(void *)(b),(c)) b 18 arch/m68k/include/asm/nubus.h #define nubus_memcpy_toio(a,b,c) memcpy((void *)(a),(b),(c)) b 24 arch/m68k/include/asm/pgtable_no.h #define pmd_offset(a, b) ((void *)0) b 30 arch/m68k/include/asm/raw_io.h #define out_8(addr,b) (void)((*(__force volatile u8 *) (addr)) = (b)) b 82 arch/m68k/include/asm/raw_io.h #define rom_out_8(addr, b) \ b 83 arch/m68k/include/asm/raw_io.h ({u8 __w, __v = (b); u32 _addr = ((u32) (addr)); \ b 55 arch/m68k/include/asm/segment.h #define segment_eq(a, b) ((a).seg == (b).seg) b 68 arch/m68k/include/asm/uaccess_mm.h __put_user_asm(__pu_err, __pu_val, ptr, b, d, -EFAULT); \ b 136 arch/m68k/include/asm/uaccess_mm.h __get_user_asm(__gu_err, x, ptr, u8, b, d, -EFAULT); \ b 186 arch/m68k/include/asm/uaccess_mm.h #define __suffix1 b b 322 arch/m68k/include/asm/uaccess_mm.h __put_user_asm(res, *(u8 *)from, (u8 __user *)to, b, d, 1); b 328 arch/m68k/include/asm/uaccess_mm.h __constant_copy_to_user_asm(res, to, from, tmp, 3, w, b,); b 334 arch/m68k/include/asm/uaccess_mm.h __constant_copy_to_user_asm(res, to, from, tmp, 5, l, b,); b 340 arch/m68k/include/asm/uaccess_mm.h __constant_copy_to_user_asm(res, to, from, tmp, 7, l, w, b); b 346 arch/m68k/include/asm/uaccess_mm.h __constant_copy_to_user_asm(res, to, from, tmp, 9, l, l, b); b 37 arch/m68k/include/asm/uaccess_no.h __put_user_asm(__pu_err, __pu_val, ptr, b); \ b 77 arch/m68k/include/asm/uaccess_no.h __get_user_asm(__gu_err, __gu_val, ptr, b, "=d"); \ b 16 arch/m68k/include/asm/zorro.h #define z_memset_io(a,b,c) memset((void *)(a),(b),(c)) b 17 arch/m68k/include/asm/zorro.h #define z_memcpy_fromio(a,b,c) memcpy((a),(void *)(b),(c)) b 18 arch/m68k/include/asm/zorro.h #define z_memcpy_toio(a,b,c) memcpy((void *)(a),(b),(c)) b 35 arch/m68k/lib/ashldi3.c __ashldi3 (DItype u, word_type b) b 41 arch/m68k/lib/ashldi3.c if (b == 0) b 46 arch/m68k/lib/ashldi3.c bm = (sizeof (SItype) * BITS_PER_UNIT) - b; b 55 arch/m68k/lib/ashldi3.c w.s.low = (USItype)uu.s.low << b; b 56 arch/m68k/lib/ashldi3.c w.s.high = ((USItype)uu.s.high << b) | carries; b 35 arch/m68k/lib/ashrdi3.c __ashrdi3 (DItype u, word_type b) b 41 arch/m68k/lib/ashrdi3.c if (b == 0) b 46 arch/m68k/lib/ashrdi3.c bm = (sizeof (SItype) * BITS_PER_UNIT) - b; b 56 arch/m68k/lib/ashrdi3.c w.s.high = uu.s.high >> b; b 57 arch/m68k/lib/ashrdi3.c w.s.low = ((USItype)uu.s.low >> b) | carries; b 35 arch/m68k/lib/lshrdi3.c __lshrdi3 (DItype u, word_type b) b 41 arch/m68k/lib/lshrdi3.c if (b == 0) b 46 arch/m68k/lib/lshrdi3.c bm = (sizeof (SItype) * BITS_PER_UNIT) - b; b 55 arch/m68k/lib/lshrdi3.c w.s.high = (USItype)uu.s.high >> b; b 56 arch/m68k/lib/lshrdi3.c w.s.low = ((USItype)uu.s.low >> b) | carries; b 580 arch/m68k/mac/misc.c #define DIV(a, b) ((a) / (b) - ((a) % (b) < 0)) b 135 arch/m68k/math-emu/fp_decode.h move.b %d2,%d0 b 46 arch/m68k/mvme147/config.c static int bcd2int (unsigned char b); b 168 arch/m68k/mvme147/config.c static int bcd2int (unsigned char b) b 170 arch/m68k/mvme147/config.c return ((b>>4)*10 + (b&15)); b 51 arch/m68k/mvme16x/config.c int bcd2int (unsigned char b); b 433 arch/m68k/mvme16x/config.c int bcd2int (unsigned char b) b 435 arch/m68k/mvme16x/config.c return ((b>>4)*10 + (b&15)); b 50 arch/m68k/q40/config.c static void q40_mem_console_write(struct console *co, const char *b, b 27 arch/m68k/sun3/sun3dvma.c static inline void dvma_unmap_iommu(unsigned long a, int b) b 40 arch/microblaze/include/asm/cacheflush.h void (*iflr)(unsigned long a, unsigned long b); b 42 arch/microblaze/include/asm/cacheflush.h void (*iinr)(unsigned long a, unsigned long b); b 47 arch/microblaze/include/asm/cacheflush.h void (*dflr)(unsigned long a, unsigned long b); b 49 arch/microblaze/include/asm/cacheflush.h void (*dinr)(unsigned long a, unsigned long b); b 88 arch/microblaze/include/asm/delay.h #define muldiv(a, b, c) (((a)*(b))/(c)) b 35 arch/microblaze/include/asm/hash.h unsigned int b, c; b 38 arch/microblaze/include/asm/hash.h b = a << 23; b 41 arch/microblaze/include/asm/hash.h b += a; b 45 arch/microblaze/include/asm/hash.h a += b; /* (a << 5) + b */ b 49 arch/microblaze/include/asm/hash.h return a - b; /* (a << 11) + (b << 6) + (c << 3) - b */ b 62 arch/microblaze/include/asm/hash.h unsigned int b, c, d; b 64 arch/microblaze/include/asm/hash.h b = a << 4; /* 4 */ b 65 arch/microblaze/include/asm/hash.h c = b << 1; /* 1 5 */ b 66 arch/microblaze/include/asm/hash.h b += a; /* 1 6 */ b 67 arch/microblaze/include/asm/hash.h c += b; /* 1 7 */ b 71 arch/microblaze/include/asm/hash.h d += b; /* 1 19 */ b 75 arch/microblaze/include/asm/hash.h d += b; /* 1 30 */ b 31 arch/microblaze/include/asm/pci.h #define pcibios_scan_all_fns(a, b) 0 b 27 arch/microblaze/include/asm/pgtable.h #define pmd_offset(a, b) ((void *) 0) b 48 arch/microblaze/include/asm/uaccess.h # define segment_eq(a, b) ((a).seg == (b).seg) b 6 arch/microblaze/lib/ashldi3.c long long __ashldi3(long long u, word_type b) b 11 arch/microblaze/lib/ashldi3.c if (b == 0) b 15 arch/microblaze/lib/ashldi3.c bm = 32 - b; b 23 arch/microblaze/lib/ashldi3.c w.s.low = (unsigned int) uu.s.low << b; b 24 arch/microblaze/lib/ashldi3.c w.s.high = ((unsigned int) uu.s.high << b) | carries; b 6 arch/microblaze/lib/ashrdi3.c long long __ashrdi3(long long u, word_type b) b 11 arch/microblaze/lib/ashrdi3.c if (b == 0) b 15 arch/microblaze/lib/ashrdi3.c bm = 32 - b; b 25 arch/microblaze/lib/ashrdi3.c w.s.high = uu.s.high >> b; b 26 arch/microblaze/lib/ashrdi3.c w.s.low = ((unsigned int) uu.s.low >> b) | carries; b 6 arch/microblaze/lib/cmpdi2.c word_type __cmpdi2(long long a, long long b) b 12 arch/microblaze/lib/cmpdi2.c .ll = b b 26 arch/microblaze/lib/libgcc.h extern long long __ashldi3(long long u, word_type b); b 27 arch/microblaze/lib/libgcc.h extern long long __ashrdi3(long long u, word_type b); b 28 arch/microblaze/lib/libgcc.h extern word_type __cmpdi2(long long a, long long b); b 29 arch/microblaze/lib/libgcc.h extern long long __lshrdi3(long long u, word_type b); b 31 arch/microblaze/lib/libgcc.h extern word_type __ucmpdi2(unsigned long long a, unsigned long long b); b 6 arch/microblaze/lib/lshrdi3.c long long __lshrdi3(long long u, word_type b) b 11 arch/microblaze/lib/lshrdi3.c if (b == 0) b 15 arch/microblaze/lib/lshrdi3.c bm = 32 - b; b 23 arch/microblaze/lib/lshrdi3.c w.s.high = (unsigned int) uu.s.high >> b; b 24 arch/microblaze/lib/lshrdi3.c w.s.low = ((unsigned int) uu.s.low >> b) | carries; b 6 arch/microblaze/lib/ucmpdi2.c word_type __ucmpdi2(unsigned long long a, unsigned long long b) b 9 arch/microblaze/lib/ucmpdi2.c const DWunion bu = {.ll = b}; b 676 arch/microblaze/pci/pci-common.c struct pci_bus *b; b 741 arch/microblaze/pci/pci-common.c list_for_each_entry(b, &bus->children, node) b 742 arch/microblaze/pci/pci-common.c pcibios_allocate_bus_resources(b); b 878 arch/microblaze/pci/pci-common.c struct pci_bus *b; b 883 arch/microblaze/pci/pci-common.c list_for_each_entry(b, &pci_root_buses, node) b 884 arch/microblaze/pci/pci-common.c pcibios_allocate_bus_resources(b); b 893 arch/microblaze/pci/pci-common.c list_for_each_entry(b, &pci_root_buses, node) b 894 arch/microblaze/pci/pci-common.c pcibios_reserve_legacy_regions(b); b 242 arch/mips/alchemy/devboards/db1000.c static void db1100_mmcled_set(struct led_classdev *led, enum led_brightness b) b 244 arch/mips/alchemy/devboards/db1000.c if (b != LED_OFF) b 280 arch/mips/alchemy/devboards/db1000.c static void db1100_mmc1led_set(struct led_classdev *led, enum led_brightness b) b 282 arch/mips/alchemy/devboards/db1000.c if (b != LED_OFF) b 2771 arch/mips/cavium-octeon/octeon-irq.c u64 b = ciu3_info->ciu3_addr; b 2791 arch/mips/cavium-octeon/octeon-irq.c cvmx_write_csr(b + CIU3_IDT_CTL(idt_ip2), 0); b 2792 arch/mips/cavium-octeon/octeon-irq.c cvmx_write_csr(b + CIU3_IDT_PP(idt_ip2, 0), 1ull << core); b 2793 arch/mips/cavium-octeon/octeon-irq.c cvmx_write_csr(b + CIU3_IDT_IO(idt_ip2), 0); b 2796 arch/mips/cavium-octeon/octeon-irq.c cvmx_write_csr(b + CIU3_IDT_CTL(idt_ip3), 1); b 2797 arch/mips/cavium-octeon/octeon-irq.c cvmx_write_csr(b + CIU3_IDT_PP(idt_ip3, 0), 1ull << core); b 2798 arch/mips/cavium-octeon/octeon-irq.c cvmx_write_csr(b + CIU3_IDT_IO(idt_ip3), 0); b 2801 arch/mips/cavium-octeon/octeon-irq.c cvmx_write_csr(b + CIU3_IDT_CTL(idt_ip4), 2); b 2802 arch/mips/cavium-octeon/octeon-irq.c cvmx_write_csr(b + CIU3_IDT_PP(idt_ip4, 0), 0); b 2803 arch/mips/cavium-octeon/octeon-irq.c cvmx_write_csr(b + CIU3_IDT_IO(idt_ip4), 0); b 2805 arch/mips/cavium-octeon/octeon-irq.c cvmx_write_csr(b + CIU3_IDT_CTL(unused_idt2), 0); b 2806 arch/mips/cavium-octeon/octeon-irq.c cvmx_write_csr(b + CIU3_IDT_PP(unused_idt2, 0), 0); b 2807 arch/mips/cavium-octeon/octeon-irq.c cvmx_write_csr(b + CIU3_IDT_IO(unused_idt2), 0); b 2812 arch/mips/cavium-octeon/octeon-irq.c cvmx_write_csr(b + CIU3_ISC_W1C(intsn), 2); b 2813 arch/mips/cavium-octeon/octeon-irq.c cvmx_write_csr(b + CIU3_ISC_CTL(intsn), 0); b 104 arch/mips/crypto/crc32-mips.c CRC32(crc, value, b); b 144 arch/mips/crypto/crc32-mips.c CRC32C(crc, value, b); b 82 arch/mips/include/asm/asm.h 9: b 9b; \ b 265 arch/mips/include/asm/asmmacro.h ld.b $w\wd, \off(\base) b 301 arch/mips/include/asm/asmmacro.h st.b $w\wd, \off(\base) b 449 arch/mips/include/asm/io.h BUILDIO_MEM(b, u8) b 467 arch/mips/include/asm/io.h BUILDIO_IOPORT(b, u8) b 573 arch/mips/include/asm/io.h BUILDSTRING(b, u8) b 74 arch/mips/include/asm/ip32/mace.h volatile unsigned char b[4]; b 603 arch/mips/include/asm/mach-au1x00/au1000.h void __iomem *b = (void __iomem *)KSEG1ADDR(AU1000_SYS_PHYS_ADDR); b 605 arch/mips/include/asm/mach-au1x00/au1000.h return __raw_readl(b + regofs); b 610 arch/mips/include/asm/mach-au1x00/au1000.h void __iomem *b = (void __iomem *)KSEG1ADDR(AU1000_SYS_PHYS_ADDR); b 612 arch/mips/include/asm/mach-au1x00/au1000.h __raw_writel(v, b + regofs); b 619 arch/mips/include/asm/mach-au1x00/au1000.h void __iomem *b = (void __iomem *)KSEG1ADDR(AU1000_STATIC_MEM_PHYS_ADDR); b 621 arch/mips/include/asm/mach-au1x00/au1000.h return __raw_readl(b + regofs); b 626 arch/mips/include/asm/mach-au1x00/au1000.h void __iomem *b = (void __iomem *)KSEG1ADDR(AU1000_STATIC_MEM_PHYS_ADDR); b 628 arch/mips/include/asm/mach-au1x00/au1000.h __raw_writel(v, b + regofs); b 146 arch/mips/include/asm/mach-cavium-octeon/kernel-entry-init.h b octeon_wait_forever b 110 arch/mips/include/asm/mach-malta/kernel-entry-init.h b 0f b 124 arch/mips/include/asm/mach-malta/kernel-entry-init.h 1: b 1b b 37 arch/mips/include/asm/mach-paravirt/kernel-entry-init.h b 2b # Unknown CPU, loop forever. b 178 arch/mips/include/asm/netlogic/xlp-hal/bridge.h #define nlm_read_bridge_reg(b, r) nlm_read_reg(b, r) b 179 arch/mips/include/asm/netlogic/xlp-hal/bridge.h #define nlm_write_bridge_reg(b, r, v) nlm_write_reg(b, r, v) b 51 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define XLP_IO_PCI_OFFSET(b, d, f) (((b) << 20) | ((d) << 15) | ((f) << 12)) b 200 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define nlm_read_pci_reg(b, r) nlm_read_reg(b, r) b 201 arch/mips/include/asm/netlogic/xlp-hal/iomap.h #define nlm_write_pci_reg(b, r, v) nlm_write_reg(b, r, v) b 99 arch/mips/include/asm/netlogic/xlp-hal/pcibus.h #define nlm_read_pcie_reg(b, r) nlm_read_reg(b, r) b 100 arch/mips/include/asm/netlogic/xlp-hal/pcibus.h #define nlm_write_pcie_reg(b, r, v) nlm_write_reg(b, r, v) b 217 arch/mips/include/asm/netlogic/xlp-hal/pic.h #define nlm_read_pic_reg(b, r) nlm_read_reg64(b, r) b 218 arch/mips/include/asm/netlogic/xlp-hal/pic.h #define nlm_write_pic_reg(b, r, v) nlm_write_reg64(b, r, v) b 194 arch/mips/include/asm/netlogic/xlp-hal/sys.h #define nlm_read_sys_reg(b, r) nlm_read_reg(b, r) b 195 arch/mips/include/asm/netlogic/xlp-hal/sys.h #define nlm_write_sys_reg(b, r, v) nlm_write_reg(b, r, v) b 94 arch/mips/include/asm/netlogic/xlp-hal/uart.h #define nlm_read_uart_reg(b, r) nlm_read_reg(b, r) b 95 arch/mips/include/asm/netlogic/xlp-hal/uart.h #define nlm_write_uart_reg(b, r, v) nlm_write_reg(b, r, v) b 185 arch/mips/include/asm/netlogic/xlr/fmn.h #define nlm_read_c2_bucksize(b) __read_32bit_c2_register($4, b) b 186 arch/mips/include/asm/netlogic/xlr/fmn.h #define nlm_write_c2_bucksize(b, v) __write_32bit_c2_register($4, b, v) b 154 arch/mips/include/asm/sibyte/bcm1480_regs.h #define A_BCM1480_L2_BANK_ADDRESS(b) (A_BCM1480_L2_BANK_BASE | (((b)&0x7)<<17)) b 148 arch/mips/include/asm/sibyte/sb1250_scd.h b 992f ; \ b 75 arch/mips/include/asm/uaccess.h #define segment_eq(a, b) ((a).seg == (b).seg) b 25 arch/mips/include/asm/uasm.h void uasm_i##op(u32 **buf, unsigned int a, unsigned int b, unsigned int c) b 28 arch/mips/include/asm/uasm.h void uasm_i##op(u32 **buf, unsigned int a, unsigned int b, unsigned int c) b 31 arch/mips/include/asm/uasm.h void uasm_i##op(u32 **buf, unsigned int a, unsigned int b, unsigned int c) b 34 arch/mips/include/asm/uasm.h void uasm_i##op(u32 **buf, unsigned int a, unsigned int b, unsigned int c) b 37 arch/mips/include/asm/uasm.h void uasm_i##op(u32 **buf, unsigned int a, unsigned int b, signed int c) b 40 arch/mips/include/asm/uasm.h void uasm_i##op(u32 **buf, unsigned int a, signed int b, unsigned int c) b 43 arch/mips/include/asm/uasm.h void uasm_i##op(u32 **buf, int a, int b, int c) b 46 arch/mips/include/asm/uasm.h void uasm_i##op(u32 **buf, unsigned int a, unsigned int b, signed int c) b 49 arch/mips/include/asm/uasm.h void uasm_i##op(u32 **buf, unsigned int a, unsigned int b, unsigned int c, \ b 53 arch/mips/include/asm/uasm.h void uasm_i##op(u32 **buf, unsigned int a, unsigned int b) b 56 arch/mips/include/asm/uasm.h void uasm_i##op(u32 **buf, unsigned int a, unsigned int b) b 59 arch/mips/include/asm/uasm.h void uasm_i##op(u32 **buf, unsigned int a, signed int b) b 250 arch/mips/include/asm/uasm.h #define uasm_i_move(buf, a, b) UASM_i_ADDU(buf, a, 0, b) b 1442 arch/mips/kernel/perf_event_mipsxx.c #define IS_BOTH_COUNTERS_24K_EVENT(b) \ b 1443 arch/mips/kernel/perf_event_mipsxx.c ((b) == 0 || (b) == 1 || (b) == 11) b 1446 arch/mips/kernel/perf_event_mipsxx.c #define IS_BOTH_COUNTERS_34K_EVENT(b) \ b 1447 arch/mips/kernel/perf_event_mipsxx.c ((b) == 0 || (b) == 1 || (b) == 11) b 1449 arch/mips/kernel/perf_event_mipsxx.c #define IS_RANGE_P_34K_EVENT(r, b) \ b 1450 arch/mips/kernel/perf_event_mipsxx.c ((b) == 0 || (r) == 18 || (b) == 21 || (b) == 22 || \ b 1451 arch/mips/kernel/perf_event_mipsxx.c (b) == 25 || (b) == 39 || (r) == 44 || (r) == 174 || \ b 1452 arch/mips/kernel/perf_event_mipsxx.c (r) == 176 || ((b) >= 50 && (b) <= 55) || \ b 1453 arch/mips/kernel/perf_event_mipsxx.c ((b) >= 64 && (b) <= 67)) b 1458 arch/mips/kernel/perf_event_mipsxx.c #define IS_BOTH_COUNTERS_74K_EVENT(b) \ b 1459 arch/mips/kernel/perf_event_mipsxx.c ((b) == 0 || (b) == 1) b 1462 arch/mips/kernel/perf_event_mipsxx.c #define IS_BOTH_COUNTERS_PROAPTIV_EVENT(b) \ b 1463 arch/mips/kernel/perf_event_mipsxx.c ((b) == 0 || (b) == 1) b 1465 arch/mips/kernel/perf_event_mipsxx.c #define IS_BOTH_COUNTERS_P5600_EVENT(b) \ b 1466 arch/mips/kernel/perf_event_mipsxx.c ((b) == 0 || (b) == 1) b 1469 arch/mips/kernel/perf_event_mipsxx.c #define IS_BOTH_COUNTERS_1004K_EVENT(b) \ b 1470 arch/mips/kernel/perf_event_mipsxx.c ((b) == 0 || (b) == 1 || (b) == 11) b 1472 arch/mips/kernel/perf_event_mipsxx.c #define IS_RANGE_P_1004K_EVENT(r, b) \ b 1473 arch/mips/kernel/perf_event_mipsxx.c ((b) == 0 || (r) == 18 || (b) == 21 || (b) == 22 || \ b 1474 arch/mips/kernel/perf_event_mipsxx.c (b) == 25 || (b) == 36 || (b) == 39 || (r) == 44 || \ b 1475 arch/mips/kernel/perf_event_mipsxx.c (r) == 174 || (r) == 176 || ((b) >= 50 && (b) <= 59) || \ b 1476 arch/mips/kernel/perf_event_mipsxx.c (r) == 188 || (b) == 61 || (b) == 62 || \ b 1477 arch/mips/kernel/perf_event_mipsxx.c ((b) >= 64 && (b) <= 67)) b 1482 arch/mips/kernel/perf_event_mipsxx.c #define IS_BOTH_COUNTERS_INTERAPTIV_EVENT(b) \ b 1483 arch/mips/kernel/perf_event_mipsxx.c ((b) == 0 || (b) == 1 || (b) == 11) b 1486 arch/mips/kernel/perf_event_mipsxx.c #define IS_RANGE_P_INTERAPTIV_EVENT(r, b) \ b 1487 arch/mips/kernel/perf_event_mipsxx.c ((b) == 0 || (r) == 18 || (b) == 21 || (b) == 22 || \ b 1488 arch/mips/kernel/perf_event_mipsxx.c (b) == 25 || (b) == 36 || (b) == 38 || (b) == 39 || \ b 1489 arch/mips/kernel/perf_event_mipsxx.c (r) == 44 || (r) == 174 || (r) == 176 || ((b) >= 50 && \ b 1490 arch/mips/kernel/perf_event_mipsxx.c (b) <= 59) || (r) == 188 || (b) == 61 || (b) == 62 || \ b 1491 arch/mips/kernel/perf_event_mipsxx.c ((b) >= 64 && (b) <= 67)) b 1496 arch/mips/kernel/perf_event_mipsxx.c #define IS_BOTH_COUNTERS_BMIPS5000_EVENT(b) \ b 1497 arch/mips/kernel/perf_event_mipsxx.c ((b) == 0 || (b) == 1) b 921 arch/mips/kernel/traps.c char b[40]; b 942 arch/mips/kernel/traps.c scnprintf(b, sizeof(b), "%s instruction in kernel code", str); b 943 arch/mips/kernel/traps.c die_if_kernel(b, regs); b 968 arch/mips/kernel/traps.c scnprintf(b, sizeof(b), "%s instruction in kernel code", str); b 969 arch/mips/kernel/traps.c die_if_kernel(b, regs); b 1981 arch/mips/kernel/traps.c unsigned char *b; b 1992 arch/mips/kernel/traps.c b = (unsigned char *)(ebase + 0x200 + n*VECTORSPACING); b 2033 arch/mips/kernel/traps.c set_handler(((unsigned long)b - ebase), vec_start, b 2039 arch/mips/kernel/traps.c h = (u16 *)(b + lui_offset); b 2041 arch/mips/kernel/traps.c h = (u16 *)(b + ori_offset); b 2043 arch/mips/kernel/traps.c local_flush_icache_range((unsigned long)b, b 2044 arch/mips/kernel/traps.c (unsigned long)(b+handler_len)); b 2054 arch/mips/kernel/traps.c h = (u16 *)b; b 2065 arch/mips/kernel/traps.c local_flush_icache_range((unsigned long)b, b 2066 arch/mips/kernel/traps.c (unsigned long)(b+8)); b 45 arch/mips/lasat/interrupt.c int b = 31, s; b 47 arch/mips/lasat/interrupt.c s = 16; if (x << 16 == 0) s = 0; b -= s; x <<= s; b 48 arch/mips/lasat/interrupt.c s = 8; if (x << 8 == 0) s = 0; b -= s; x <<= s; b 49 arch/mips/lasat/interrupt.c s = 4; if (x << 4 == 0) s = 0; b -= s; x <<= s; b 50 arch/mips/lasat/interrupt.c s = 2; if (x << 2 == 0) s = 0; b -= s; x <<= s; b 51 arch/mips/lasat/interrupt.c s = 1; if (x << 1 == 0) s = 0; b -= s; b 53 arch/mips/lasat/interrupt.c return b; b 15 arch/mips/lib/multi3.c static inline long long notrace dmulu(long long a, long long b) b 19 arch/mips/lib/multi3.c asm ("dmulu %0,%1,%2" : "=r" (res) : "r" (a), "r" (b)); b 24 arch/mips/lib/multi3.c static inline long long notrace dmuhu(long long a, long long b) b 28 arch/mips/lib/multi3.c asm ("dmuhu %0,%1,%2" : "=r" (res) : "r" (a), "r" (b)); b 33 arch/mips/lib/multi3.c ti_type notrace __multi3(ti_type a, ti_type b) b 38 arch/mips/lib/multi3.c bb.ti = b; b 1678 arch/mips/math-emu/cp1emu.c union ieee754sp(*b) (union ieee754sp, union ieee754sp); b 1687 arch/mips/math-emu/cp1emu.c handler.b = ieee754sp_add; b 1691 arch/mips/math-emu/cp1emu.c handler.b = ieee754sp_sub; b 1695 arch/mips/math-emu/cp1emu.c handler.b = ieee754sp_mul; b 1699 arch/mips/math-emu/cp1emu.c handler.b = ieee754sp_div; b 1911 arch/mips/math-emu/cp1emu.c rv.s = (*handler.b) (fs, ft); b 2055 arch/mips/math-emu/cp1emu.c union ieee754dp(*b) (union ieee754dp, union ieee754dp); b 2063 arch/mips/math-emu/cp1emu.c handler.b = ieee754dp_add; b 2067 arch/mips/math-emu/cp1emu.c handler.b = ieee754dp_sub; b 2071 arch/mips/math-emu/cp1emu.c handler.b = ieee754dp_mul; b 2075 arch/mips/math-emu/cp1emu.c handler.b = ieee754dp_div; b 2282 arch/mips/math-emu/cp1emu.c rv.d = (*handler.b) (fs, ft); b 28 arch/mips/math-emu/ieee754.c #define xPCNST(s, b, m, ebias) \ b 32 arch/mips/math-emu/ieee754.c .bexp = (b) + ebias, \ b 37 arch/mips/math-emu/ieee754.c #define DPCNST(s, b, m) \ b 38 arch/mips/math-emu/ieee754.c xPCNST(s, b, m, DP_EBIAS) b 61 arch/mips/math-emu/ieee754.c #define SPCNST(s, b, m) \ b 62 arch/mips/math-emu/ieee754.c xPCNST(s, b, m, SP_EBIAS) b 16 arch/mips/mm/context.c static bool asid_versions_eq(int cpu, u64 a, u64 b) b 18 arch/mips/mm/context.c return ((a ^ b) & asid_version_mask(cpu)) == 0; b 78 arch/mips/mm/page.c #define _uasm_i_pref(a, b, c, d) \ b 82 arch/mips/mm/page.c uasm_i_pref(a, b, c, d);\ b 84 arch/mips/mm/page.c uasm_i_pref(a, b, c, d); \ b 69 arch/mips/mm/tlbex.c unsigned long b; b 383 arch/mips/mm/tlbex.c UASM_i_SW(p, 2, offsetof(struct tlb_reg_save, b), K0); b 400 arch/mips/mm/tlbex.c UASM_i_LW(p, 2, offsetof(struct tlb_reg_save, b), K0); b 32 arch/mips/mm/uasm-micromips.c #define M(a, b, c, d, e, f) \ b 34 arch/mips/mm/uasm-micromips.c | (b) << RT_SH \ b 32 arch/mips/mm/uasm-mips.c #define M(a, b, c, d, e, f) \ b 34 arch/mips/mm/uasm-mips.c | (b) << RS_SH \ b 41 arch/mips/mm/uasm-mips.c #define M6(a, b, c, d, e) \ b 43 arch/mips/mm/uasm-mips.c | (b) << RS_SH \ b 155 arch/mips/mm/uasm.c build_insn(buf, insn##op, a, b, c); \ b 162 arch/mips/mm/uasm.c build_insn(buf, insn##op, b, c, a); \ b 169 arch/mips/mm/uasm.c build_insn(buf, insn##op, b, a, c); \ b 176 arch/mips/mm/uasm.c build_insn(buf, insn##op, c, b, a); \ b 183 arch/mips/mm/uasm.c build_insn(buf, insn##op, b, c, a); \ b 190 arch/mips/mm/uasm.c build_insn(buf, insn##op, a, b, c); \ b 197 arch/mips/mm/uasm.c build_insn(buf, insn##op, c, a, b); \ b 204 arch/mips/mm/uasm.c build_insn(buf, insn##op, b, a, c); \ b 211 arch/mips/mm/uasm.c build_insn(buf, insn##op, b, a, c+d-1, c); \ b 218 arch/mips/mm/uasm.c build_insn(buf, insn##op, b, a, c+d-33, c); \ b 225 arch/mips/mm/uasm.c build_insn(buf, insn##op, b, a, c+d-33, c-32); \ b 232 arch/mips/mm/uasm.c build_insn(buf, insn##op, b, a, d-1, c); \ b 239 arch/mips/mm/uasm.c build_insn(buf, insn##op, a, b); \ b 246 arch/mips/mm/uasm.c build_insn(buf, insn##op, b, a); \ b 253 arch/mips/mm/uasm.c build_insn(buf, insn##op, a, b); \ b 394 arch/mips/mm/uasm.c void uasm_i_pref(u32 **buf, unsigned int a, signed int b, b 402 arch/mips/mm/uasm.c build_insn(buf, insn_pref, c, 28, b); b 404 arch/mips/mm/uasm.c build_insn(buf, insn_pref, c, a, b); b 1294 arch/mips/net/ebpf_jit.c emit_instr(ctx, b, b_off); b 143 arch/mips/netlogic/xlp/ahci-init-xlp2.c #define nlm_read_sata_reg(b, r) nlm_read_reg(b, r) b 144 arch/mips/netlogic/xlp/ahci-init-xlp2.c #define nlm_write_sata_reg(b, r, v) nlm_write_reg(b, r, v) b 87 arch/mips/netlogic/xlp/ahci-init.c #define nlm_read_sata_reg(b, r) nlm_read_reg(b, r) b 88 arch/mips/netlogic/xlp/ahci-init.c #define nlm_write_sata_reg(b, r, v) nlm_write_reg(b, r, v) b 84 arch/mips/netlogic/xlp/usb-init-xlp2.c #define nlm_read_usb_reg(b, r) nlm_read_reg(b, r) b 85 arch/mips/netlogic/xlp/usb-init-xlp2.c #define nlm_write_usb_reg(b, r, v) nlm_write_reg(b, r, v) b 63 arch/mips/netlogic/xlp/usb-init.c #define nlm_read_usb_reg(b, r) nlm_read_reg(b, r) b 64 arch/mips/netlogic/xlp/usb-init.c #define nlm_write_usb_reg(b, r, v) nlm_write_reg(b, r, v) b 120 arch/mips/netlogic/xlp/wakeup.c int b = xlp9xx_get_socbus(n); b 121 arch/mips/netlogic/xlp/wakeup.c pr_info("Node %d SoC PCI bus %d.\n", n, b); b 122 arch/mips/netlogic/xlp/wakeup.c if (b == 0) b 50 arch/mips/pci/ops-mace.c *val = mace->pci.config_data.b[(reg & 3) ^ 3]; b 82 arch/mips/pci/ops-mace.c mace->pci.config_data.b[(reg & 3) ^ 3] = val; b 110 arch/mips/sibyte/swarm/rtc_m41t81.c static int m41t81_write(uint8_t addr, int b) b 116 arch/mips/sibyte/swarm/rtc_m41t81.c __raw_writeq(b & 0xff, SMB_CSR(R_SMB_DATA)); b 86 arch/mips/sibyte/swarm/rtc_xicor1241.c static int xicor_write(uint8_t addr, int b) b 92 arch/mips/sibyte/swarm/rtc_xicor1241.c __raw_writeq((addr & 0xff) | ((b & 0xff) << 8), SMB_CSR(R_SMB_DATA)); b 48 arch/nds32/include/asm/uaccess.h #define segment_eq(a, b) ((a) == (b)) b 64 arch/nds32/math-emu/fpuemu.c void (*b)(void *ft, void *fa); b 113 arch/nds32/math-emu/fpuemu.c func.b = fs2d; b 117 arch/nds32/math-emu/fpuemu.c func.b = fs2si; b 121 arch/nds32/math-emu/fpuemu.c func.b = fs2si_z; b 125 arch/nds32/math-emu/fpuemu.c func.b = fs2ui; b 129 arch/nds32/math-emu/fpuemu.c func.b = fs2ui_z; b 133 arch/nds32/math-emu/fpuemu.c func.b = fsi2s; b 137 arch/nds32/math-emu/fpuemu.c func.b = fui2s; b 141 arch/nds32/math-emu/fpuemu.c func.b = fsqrts; b 206 arch/nds32/math-emu/fpuemu.c func.b = fd2s; b 210 arch/nds32/math-emu/fpuemu.c func.b = fd2si; b 214 arch/nds32/math-emu/fpuemu.c func.b = fd2si_z; b 218 arch/nds32/math-emu/fpuemu.c func.b = fd2ui; b 222 arch/nds32/math-emu/fpuemu.c func.b = fd2ui_z; b 226 arch/nds32/math-emu/fpuemu.c func.b = fsi2d; b 230 arch/nds32/math-emu/fpuemu.c func.b = fui2d; b 234 arch/nds32/math-emu/fpuemu.c func.b = fsqrtd; b 275 arch/nds32/math-emu/fpuemu.c func.b(ft, fa); b 292 arch/nds32/math-emu/fpuemu.c func.b(ft, fa); b 314 arch/nds32/math-emu/fpuemu.c func.b(ft, fa); b 331 arch/nds32/math-emu/fpuemu.c func.b(ft, fa); b 33 arch/nios2/include/asm/uaccess.h #define segment_eq(a, b) ((a).seg == (b).seg) b 70 arch/nios2/kernel/misaligned.c u8 a, b, d0, d1, d2, d3; b 91 arch/nios2/kernel/misaligned.c b = (isn >> 22) & 0x1f; b 101 arch/nios2/kernel/misaligned.c put_reg_val(fp, b, val); b 104 arch/nios2/kernel/misaligned.c val = get_reg_val(fp, b); b 119 arch/nios2/kernel/misaligned.c put_reg_val(fp, b, val); b 122 arch/nios2/kernel/misaligned.c val = get_reg_val(fp, b); b 145 arch/nios2/kernel/misaligned.c put_reg_val(fp, b, val); b 46 arch/openrisc/include/asm/uaccess.h #define segment_eq(a, b) ((a) == (b)) b 49 arch/parisc/include/asm/assembly.h #define BL b,l b 33 arch/parisc/include/asm/hash.h u32 a, b, c; b 40 arch/parisc/include/asm/hash.h b = x << 9; a += x; b 41 arch/parisc/include/asm/hash.h c = x << 23; b += a; b 42 arch/parisc/include/asm/hash.h c += b; b 44 arch/parisc/include/asm/hash.h b <<= 11; b 45 arch/parisc/include/asm/hash.h a += c << 3; b -= c; b 46 arch/parisc/include/asm/hash.h return (a << 3) + b; b 120 arch/parisc/include/asm/hash.h u64 b, c, d; b 131 arch/parisc/include/asm/hash.h _ASSIGN(b, a*5); c = a << 13; b 132 arch/parisc/include/asm/hash.h b = (b << 2) + a; _ASSIGN(d, a << 17); b 133 arch/parisc/include/asm/hash.h a = b + (a << 1); c += d; b 136 arch/parisc/include/asm/hash.h c += b; a += b; b 138 arch/parisc/include/asm/hash.h a += c << 3; _ASSIGN(b, b << (7+31), "X" (c), "X" (d)); b 139 arch/parisc/include/asm/hash.h a <<= 31; b += d; b 140 arch/parisc/include/asm/hash.h a += b; b 163 arch/parisc/include/asm/io.h static inline void __raw_writeb(unsigned char b, volatile void __iomem *addr) b 165 arch/parisc/include/asm/io.h *(volatile unsigned char __force *) addr = b; b 167 arch/parisc/include/asm/io.h static inline void __raw_writew(unsigned short b, volatile void __iomem *addr) b 169 arch/parisc/include/asm/io.h *(volatile unsigned short __force *) addr = b; b 171 arch/parisc/include/asm/io.h static inline void __raw_writel(unsigned int b, volatile void __iomem *addr) b 173 arch/parisc/include/asm/io.h *(volatile unsigned int __force *) addr = b; b 175 arch/parisc/include/asm/io.h static inline void __raw_writeq(unsigned long long b, volatile void __iomem *addr) b 177 arch/parisc/include/asm/io.h *(volatile unsigned long long __force *) addr = b; b 197 arch/parisc/include/asm/io.h static inline void writeb(unsigned char b, volatile void __iomem *addr) b 199 arch/parisc/include/asm/io.h __raw_writeb(b, addr); b 227 arch/parisc/include/asm/io.h #define writeb_relaxed(b, addr) writeb(b, addr) b 257 arch/parisc/include/asm/io.h extern void outb(unsigned char b, int addr); b 258 arch/parisc/include/asm/io.h extern void outw(unsigned short b, int addr); b 259 arch/parisc/include/asm/io.h extern void outl(unsigned int b, int addr); b 79 arch/parisc/include/asm/psw.h unsigned int b:1; b 17 arch/parisc/include/asm/uaccess.h #define segment_eq(a, b) ((a).seg == (b).seg) b 69 arch/parisc/kernel/firmware.c # define spin_lock_irqsave(a, b) { b = 1; } b 71 arch/parisc/kernel/firmware.c # define spin_unlock_irqrestore(a, b) b 61 arch/parisc/kernel/pci.c #define EISA_IN(size) if (EISA_bus && (b == 0)) return eisa_in##size(addr) b 62 arch/parisc/kernel/pci.c #define EISA_OUT(size) if (EISA_bus && (b == 0)) return eisa_out##size(d, addr) b 71 arch/parisc/kernel/pci.c int b = PCI_PORT_HBA(addr); \ b 73 arch/parisc/kernel/pci.c if (!parisc_pci_hba[b]) return (u##size) -1; \ b 74 arch/parisc/kernel/pci.c return pci_port->in##type(parisc_pci_hba[b], PCI_PORT_ADDR(addr)); \ b 78 arch/parisc/kernel/pci.c PCI_PORT_IN(b, 8) b 86 arch/parisc/kernel/pci.c int b = PCI_PORT_HBA(addr); \ b 88 arch/parisc/kernel/pci.c if (!parisc_pci_hba[b]) return; \ b 89 arch/parisc/kernel/pci.c pci_port->out##type(parisc_pci_hba[b], PCI_PORT_ADDR(addr), d); \ b 93 arch/parisc/kernel/pci.c PCI_PORT_OUT(b, 8) b 87 arch/parisc/kernel/ptrace.c pa_psw(task)->b = 0; b 42 arch/parisc/kernel/unaligned.c #define OPCODE1(a,b,c) ((a)<<26|(b)<<12|(c)<<6) b 43 arch/parisc/kernel/unaligned.c #define OPCODE2(a,b) ((a)<<26|(b)<<1) b 44 arch/parisc/kernel/unaligned.c #define OPCODE3(a,b) ((a)<<26|(b)<<2) b 128 arch/parisc/kernel/unwind.c static int cmp_unwind_table_entry(const void *a, const void *b) b 131 arch/parisc/kernel/unwind.c - ((const struct unwind_table_entry *)b)->region_start; b 12 arch/parisc/lib/ucmpdi2.c int __ucmpdi2(unsigned long long a, unsigned long long b) b 15 arch/parisc/lib/ucmpdi2.c union ull_union bu = {.ull = b}; b 67 arch/powerpc/boot/ppc_asm.h b $+44; /* Skip trampoline if endian is good */ \ b 40 arch/powerpc/boot/types.h #define min_t(type, a, b) min(((type) a), ((type) b)) b 41 arch/powerpc/boot/types.h #define max_t(type, a, b) max(((type) a), ((type) b)) b 90 arch/powerpc/boot/ugecon.c char *b = (char *)buf; b 93 arch/powerpc/boot/ugecon.c if (*b == '\n') b 95 arch/powerpc/boot/ugecon.c ug_putc(*b++); b 47 arch/powerpc/boot/xz_config.h #define memeq(a, b, size) (memcmp(a, b, size) == 0) b 20 arch/powerpc/include/asm/asm-compat.h #define PPC_LLARX(t, a, b, eh) PPC_LDARX(t, a, b, eh) b 53 arch/powerpc/include/asm/asm-compat.h #define PPC_LLARX(t, a, b, eh) PPC_LWARX(t, a, b, eh) b 121 arch/powerpc/include/asm/elf.h #define VDSO_AUX_ENT(a,b) NEW_AUX_ENT(a,b) b 183 arch/powerpc/include/asm/exception-64e.h b 62f; \ b 89 arch/powerpc/include/asm/exception-64s.h b rfi_flush_fallback b 95 arch/powerpc/include/asm/exception-64s.h b rfi_flush_fallback b 101 arch/powerpc/include/asm/exception-64s.h b rfi_flush_fallback b 110 arch/powerpc/include/asm/exception-64s.h b hrfi_flush_fallback b 116 arch/powerpc/include/asm/exception-64s.h b hrfi_flush_fallback b 122 arch/powerpc/include/asm/exception-64s.h b hrfi_flush_fallback b 128 arch/powerpc/include/asm/exception-64s.h b hrfi_flush_fallback b 30 arch/powerpc/include/asm/io-defs.h DEF_PCI_AC_NORET(readsb, (const PCI_IO_ADDR a, void *b, unsigned long c), b 31 arch/powerpc/include/asm/io-defs.h (a, b, c), mem, a) b 32 arch/powerpc/include/asm/io-defs.h DEF_PCI_AC_NORET(readsw, (const PCI_IO_ADDR a, void *b, unsigned long c), b 33 arch/powerpc/include/asm/io-defs.h (a, b, c), mem, a) b 34 arch/powerpc/include/asm/io-defs.h DEF_PCI_AC_NORET(readsl, (const PCI_IO_ADDR a, void *b, unsigned long c), b 35 arch/powerpc/include/asm/io-defs.h (a, b, c), mem, a) b 36 arch/powerpc/include/asm/io-defs.h DEF_PCI_AC_NORET(writesb, (PCI_IO_ADDR a, const void *b, unsigned long c), b 37 arch/powerpc/include/asm/io-defs.h (a, b, c), mem, a) b 38 arch/powerpc/include/asm/io-defs.h DEF_PCI_AC_NORET(writesw, (PCI_IO_ADDR a, const void *b, unsigned long c), b 39 arch/powerpc/include/asm/io-defs.h (a, b, c), mem, a) b 40 arch/powerpc/include/asm/io-defs.h DEF_PCI_AC_NORET(writesl, (PCI_IO_ADDR a, const void *b, unsigned long c), b 41 arch/powerpc/include/asm/io-defs.h (a, b, c), mem, a) b 43 arch/powerpc/include/asm/io-defs.h DEF_PCI_AC_NORET(insb, (unsigned long p, void *b, unsigned long c), b 44 arch/powerpc/include/asm/io-defs.h (p, b, c), pio, p) b 45 arch/powerpc/include/asm/io-defs.h DEF_PCI_AC_NORET(insw, (unsigned long p, void *b, unsigned long c), b 46 arch/powerpc/include/asm/io-defs.h (p, b, c), pio, p) b 47 arch/powerpc/include/asm/io-defs.h DEF_PCI_AC_NORET(insl, (unsigned long p, void *b, unsigned long c), b 48 arch/powerpc/include/asm/io-defs.h (p, b, c), pio, p) b 49 arch/powerpc/include/asm/io-defs.h DEF_PCI_AC_NORET(outsb, (unsigned long p, const void *b, unsigned long c), b 50 arch/powerpc/include/asm/io-defs.h (p, b, c), pio, p) b 51 arch/powerpc/include/asm/io-defs.h DEF_PCI_AC_NORET(outsw, (unsigned long p, const void *b, unsigned long c), b 52 arch/powerpc/include/asm/io-defs.h (p, b, c), pio, p) b 53 arch/powerpc/include/asm/io-defs.h DEF_PCI_AC_NORET(outsl, (unsigned long p, const void *b, unsigned long c), b 54 arch/powerpc/include/asm/io-defs.h (p, b, c), pio, p) b 530 arch/powerpc/include/asm/io.h #define __do_readsb(a, b, n) eeh_readsb(PCI_FIX_ADDR(a), (b), (n)) b 531 arch/powerpc/include/asm/io.h #define __do_readsw(a, b, n) eeh_readsw(PCI_FIX_ADDR(a), (b), (n)) b 532 arch/powerpc/include/asm/io.h #define __do_readsl(a, b, n) eeh_readsl(PCI_FIX_ADDR(a), (b), (n)) b 534 arch/powerpc/include/asm/io.h #define __do_readsb(a, b, n) _insb(PCI_FIX_ADDR(a), (b), (n)) b 535 arch/powerpc/include/asm/io.h #define __do_readsw(a, b, n) _insw(PCI_FIX_ADDR(a), (b), (n)) b 536 arch/powerpc/include/asm/io.h #define __do_readsl(a, b, n) _insl(PCI_FIX_ADDR(a), (b), (n)) b 538 arch/powerpc/include/asm/io.h #define __do_writesb(a, b, n) _outsb(PCI_FIX_ADDR(a),(b),(n)) b 539 arch/powerpc/include/asm/io.h #define __do_writesw(a, b, n) _outsw(PCI_FIX_ADDR(a),(b),(n)) b 540 arch/powerpc/include/asm/io.h #define __do_writesl(a, b, n) _outsl(PCI_FIX_ADDR(a),(b),(n)) b 542 arch/powerpc/include/asm/io.h #define __do_insb(p, b, n) readsb((PCI_IO_ADDR)_IO_BASE+(p), (b), (n)) b 543 arch/powerpc/include/asm/io.h #define __do_insw(p, b, n) readsw((PCI_IO_ADDR)_IO_BASE+(p), (b), (n)) b 544 arch/powerpc/include/asm/io.h #define __do_insl(p, b, n) readsl((PCI_IO_ADDR)_IO_BASE+(p), (b), (n)) b 545 arch/powerpc/include/asm/io.h #define __do_outsb(p, b, n) writesb((PCI_IO_ADDR)_IO_BASE+(p),(b),(n)) b 546 arch/powerpc/include/asm/io.h #define __do_outsw(p, b, n) writesw((PCI_IO_ADDR)_IO_BASE+(p),(b),(n)) b 547 arch/powerpc/include/asm/io.h #define __do_outsl(p, b, n) writesl((PCI_IO_ADDR)_IO_BASE+(p),(b),(n)) b 51 arch/powerpc/include/asm/kvm_book3s_asm.h b kvmppc_trampoline_\intno b 61 arch/powerpc/include/asm/kvm_booke_hv_asm.h b kvmppc_handler_\intno\()_\srr1 b 90 arch/powerpc/include/asm/pci.h extern void pcibios_claim_one_bus(struct pci_bus *b); b 382 arch/powerpc/include/asm/ppc-opcode.h #define ___PPC_RB(b) (((b) & 0x1f) << 11) b 391 arch/powerpc/include/asm/ppc-opcode.h #define __PPC_RB(b) ___PPC_RB(__REG_##b) b 395 arch/powerpc/include/asm/ppc-opcode.h #define __PPC_XB(b) ((((b) & 0x1f) << 11) | (((b) & 0x20) >> 4)) b 433 arch/powerpc/include/asm/ppc-opcode.h #define PPC_COPY(a, b) stringify_in_c(.long PPC_INST_COPY | \ b 434 arch/powerpc/include/asm/ppc-opcode.h ___PPC_RA(a) | ___PPC_RB(b)) b 438 arch/powerpc/include/asm/ppc-opcode.h #define PPC_DCBAL(a, b) stringify_in_c(.long PPC_INST_DCBAL | \ b 439 arch/powerpc/include/asm/ppc-opcode.h __PPC_RA(a) | __PPC_RB(b)) b 440 arch/powerpc/include/asm/ppc-opcode.h #define PPC_DCBZL(a, b) stringify_in_c(.long PPC_INST_DCBZL | \ b 441 arch/powerpc/include/asm/ppc-opcode.h __PPC_RA(a) | __PPC_RB(b)) b 442 arch/powerpc/include/asm/ppc-opcode.h #define PPC_LQARX(t, a, b, eh) stringify_in_c(.long PPC_INST_LQARX | \ b 444 arch/powerpc/include/asm/ppc-opcode.h ___PPC_RB(b) | __PPC_EH(eh)) b 445 arch/powerpc/include/asm/ppc-opcode.h #define PPC_LDARX(t, a, b, eh) stringify_in_c(.long PPC_INST_LDARX | \ b 447 arch/powerpc/include/asm/ppc-opcode.h ___PPC_RB(b) | __PPC_EH(eh)) b 448 arch/powerpc/include/asm/ppc-opcode.h #define PPC_LWARX(t, a, b, eh) stringify_in_c(.long PPC_INST_LWARX | \ b 450 arch/powerpc/include/asm/ppc-opcode.h ___PPC_RB(b) | __PPC_EH(eh)) b 451 arch/powerpc/include/asm/ppc-opcode.h #define PPC_STQCX(t, a, b) stringify_in_c(.long PPC_INST_STQCX | \ b 453 arch/powerpc/include/asm/ppc-opcode.h ___PPC_RB(b)) b 454 arch/powerpc/include/asm/ppc-opcode.h #define PPC_MADDHD(t, a, b, c) stringify_in_c(.long PPC_INST_MADDHD | \ b 456 arch/powerpc/include/asm/ppc-opcode.h ___PPC_RB(b) | ___PPC_RC(c)) b 457 arch/powerpc/include/asm/ppc-opcode.h #define PPC_MADDHDU(t, a, b, c) stringify_in_c(.long PPC_INST_MADDHDU | \ b 459 arch/powerpc/include/asm/ppc-opcode.h ___PPC_RB(b) | ___PPC_RC(c)) b 460 arch/powerpc/include/asm/ppc-opcode.h #define PPC_MADDLD(t, a, b, c) stringify_in_c(.long PPC_INST_MADDLD | \ b 462 arch/powerpc/include/asm/ppc-opcode.h ___PPC_RB(b) | ___PPC_RC(c)) b 463 arch/powerpc/include/asm/ppc-opcode.h #define PPC_MSGSND(b) stringify_in_c(.long PPC_INST_MSGSND | \ b 464 arch/powerpc/include/asm/ppc-opcode.h ___PPC_RB(b)) b 466 arch/powerpc/include/asm/ppc-opcode.h #define PPC_MSGCLR(b) stringify_in_c(.long PPC_INST_MSGCLR | \ b 467 arch/powerpc/include/asm/ppc-opcode.h ___PPC_RB(b)) b 468 arch/powerpc/include/asm/ppc-opcode.h #define PPC_MSGSNDP(b) stringify_in_c(.long PPC_INST_MSGSNDP | \ b 469 arch/powerpc/include/asm/ppc-opcode.h ___PPC_RB(b)) b 470 arch/powerpc/include/asm/ppc-opcode.h #define PPC_MSGCLRP(b) stringify_in_c(.long PPC_INST_MSGCLRP | \ b 471 arch/powerpc/include/asm/ppc-opcode.h ___PPC_RB(b)) b 472 arch/powerpc/include/asm/ppc-opcode.h #define PPC_PASTE(a, b) stringify_in_c(.long PPC_INST_PASTE | \ b 473 arch/powerpc/include/asm/ppc-opcode.h ___PPC_RA(a) | ___PPC_RB(b)) b 483 arch/powerpc/include/asm/ppc-opcode.h #define PPC_TLBILX(t, a, b) stringify_in_c(.long PPC_INST_TLBILX | \ b 484 arch/powerpc/include/asm/ppc-opcode.h __PPC_T_TLB(t) | __PPC_RA0(a) | __PPC_RB(b)) b 485 arch/powerpc/include/asm/ppc-opcode.h #define PPC_TLBILX_ALL(a, b) PPC_TLBILX(0, a, b) b 486 arch/powerpc/include/asm/ppc-opcode.h #define PPC_TLBILX_PID(a, b) PPC_TLBILX(1, a, b) b 487 arch/powerpc/include/asm/ppc-opcode.h #define PPC_TLBILX_VA(a, b) PPC_TLBILX(3, a, b) b 502 arch/powerpc/include/asm/ppc-opcode.h #define PPC_TLBSRX_DOT(a,b) stringify_in_c(.long PPC_INST_TLBSRX_DOT | \ b 503 arch/powerpc/include/asm/ppc-opcode.h __PPC_RA0(a) | __PPC_RB(b)) b 504 arch/powerpc/include/asm/ppc-opcode.h #define PPC_TLBIVAX(a,b) stringify_in_c(.long PPC_INST_TLBIVAX | \ b 505 arch/powerpc/include/asm/ppc-opcode.h __PPC_RA0(a) | __PPC_RB(b)) b 511 arch/powerpc/include/asm/ppc-opcode.h #define PPC_ERATILX(t, a, b) stringify_in_c(.long PPC_INST_ERATILX | \ b 513 arch/powerpc/include/asm/ppc-opcode.h __PPC_RB(b)) b 514 arch/powerpc/include/asm/ppc-opcode.h #define PPC_ERATIVAX(s, a, b) stringify_in_c(.long PPC_INST_ERATIVAX | \ b 515 arch/powerpc/include/asm/ppc-opcode.h __PPC_RS(s) | __PPC_RA0(a) | __PPC_RB(b)) b 517 arch/powerpc/include/asm/ppc-opcode.h __PPC_RS(t) | __PPC_RA0(a) | __PPC_RB(b)) b 519 arch/powerpc/include/asm/ppc-opcode.h __PPC_RS(t) | __PPC_RA0(a) | __PPC_RB(b)) b 520 arch/powerpc/include/asm/ppc-opcode.h #define PPC_SLBFEE_DOT(t, b) stringify_in_c(.long PPC_INST_SLBFEE | \ b 521 arch/powerpc/include/asm/ppc-opcode.h __PPC_RT(t) | __PPC_RB(b)) b 522 arch/powerpc/include/asm/ppc-opcode.h #define __PPC_SLBFEE_DOT(t, b) stringify_in_c(.long PPC_INST_SLBFEE | \ b 523 arch/powerpc/include/asm/ppc-opcode.h ___PPC_RT(t) | ___PPC_RB(b)) b 524 arch/powerpc/include/asm/ppc-opcode.h #define PPC_ICBT(c,a,b) stringify_in_c(.long PPC_INST_ICBT | \ b 525 arch/powerpc/include/asm/ppc-opcode.h __PPC_CT(c) | __PPC_RA0(a) | __PPC_RB(b)) b 527 arch/powerpc/include/asm/ppc-opcode.h #define LBZCIX(t,a,b) stringify_in_c(.long PPC_INST_LBZCIX | \ b 528 arch/powerpc/include/asm/ppc-opcode.h __PPC_RT(t) | __PPC_RA(a) | __PPC_RB(b)) b 529 arch/powerpc/include/asm/ppc-opcode.h #define STBCIX(s,a,b) stringify_in_c(.long PPC_INST_STBCIX | \ b 530 arch/powerpc/include/asm/ppc-opcode.h __PPC_RS(s) | __PPC_RA(a) | __PPC_RB(b)) b 536 arch/powerpc/include/asm/ppc-opcode.h #define VSX_XX1(s, a, b) (__PPC_XS(s) | __PPC_RA(a) | __PPC_RB(b)) b 537 arch/powerpc/include/asm/ppc-opcode.h #define VSX_XX3(t, a, b) (__PPC_XT(t) | __PPC_XA(a) | __PPC_XB(b)) b 538 arch/powerpc/include/asm/ppc-opcode.h #define STXVD2X(s, a, b) stringify_in_c(.long PPC_INST_STXVD2X | \ b 539 arch/powerpc/include/asm/ppc-opcode.h VSX_XX1((s), a, b)) b 540 arch/powerpc/include/asm/ppc-opcode.h #define LXVD2X(s, a, b) stringify_in_c(.long PPC_INST_LXVD2X | \ b 541 arch/powerpc/include/asm/ppc-opcode.h VSX_XX1((s), a, b)) b 546 arch/powerpc/include/asm/ppc-opcode.h #define VPMSUMW(t, a, b) stringify_in_c(.long PPC_INST_VPMSUMW | \ b 547 arch/powerpc/include/asm/ppc-opcode.h VSX_XX3((t), a, b)) b 548 arch/powerpc/include/asm/ppc-opcode.h #define VPMSUMD(t, a, b) stringify_in_c(.long PPC_INST_VPMSUMD | \ b 549 arch/powerpc/include/asm/ppc-opcode.h VSX_XX3((t), a, b)) b 550 arch/powerpc/include/asm/ppc-opcode.h #define XXLOR(t, a, b) stringify_in_c(.long PPC_INST_XXLOR | \ b 551 arch/powerpc/include/asm/ppc-opcode.h VSX_XX3((t), a, b)) b 554 arch/powerpc/include/asm/ppc-opcode.h #define XVCPSGNDP(t, a, b) stringify_in_c(.long (PPC_INST_XVCPSGNDP | \ b 555 arch/powerpc/include/asm/ppc-opcode.h VSX_XX3((t), (a), (b)))) b 589 arch/powerpc/include/asm/ppc-opcode.h #define PPC_ICSWX(s, a, b) stringify_in_c(.long PPC_INST_ICSWX | \ b 592 arch/powerpc/include/asm/ppc-opcode.h ___PPC_RB(b)) b 593 arch/powerpc/include/asm/ppc-opcode.h #define PPC_ICSWEPX(s, a, b) stringify_in_c(.long PPC_INST_ICSWEPX | \ b 596 arch/powerpc/include/asm/ppc-opcode.h ___PPC_RB(b)) b 110 arch/powerpc/include/asm/ppc_asm.h #define SAVE_VR(n,b,base) li b,16*(n); stvx n,base,b b 111 arch/powerpc/include/asm/ppc_asm.h #define SAVE_2VRS(n,b,base) SAVE_VR(n,b,base); SAVE_VR(n+1,b,base) b 112 arch/powerpc/include/asm/ppc_asm.h #define SAVE_4VRS(n,b,base) SAVE_2VRS(n,b,base); SAVE_2VRS(n+2,b,base) b 113 arch/powerpc/include/asm/ppc_asm.h #define SAVE_8VRS(n,b,base) SAVE_4VRS(n,b,base); SAVE_4VRS(n+4,b,base) b 114 arch/powerpc/include/asm/ppc_asm.h #define SAVE_16VRS(n,b,base) SAVE_8VRS(n,b,base); SAVE_8VRS(n+8,b,base) b 115 arch/powerpc/include/asm/ppc_asm.h #define SAVE_32VRS(n,b,base) SAVE_16VRS(n,b,base); SAVE_16VRS(n+16,b,base) b 116 arch/powerpc/include/asm/ppc_asm.h #define REST_VR(n,b,base) li b,16*(n); lvx n,base,b b 117 arch/powerpc/include/asm/ppc_asm.h #define REST_2VRS(n,b,base) REST_VR(n,b,base); REST_VR(n+1,b,base) b 118 arch/powerpc/include/asm/ppc_asm.h #define REST_4VRS(n,b,base) REST_2VRS(n,b,base); REST_2VRS(n+2,b,base) b 119 arch/powerpc/include/asm/ppc_asm.h #define REST_8VRS(n,b,base) REST_4VRS(n,b,base); REST_4VRS(n+4,b,base) b 120 arch/powerpc/include/asm/ppc_asm.h #define REST_16VRS(n,b,base) REST_8VRS(n,b,base); REST_8VRS(n+8,b,base) b 121 arch/powerpc/include/asm/ppc_asm.h #define REST_32VRS(n,b,base) REST_16VRS(n,b,base); REST_16VRS(n+16,b,base) b 124 arch/powerpc/include/asm/ppc_asm.h #define STXVD2X_ROT(n,b,base) STXVD2X(n,b,base) b 125 arch/powerpc/include/asm/ppc_asm.h #define LXVD2X_ROT(n,b,base) LXVD2X(n,b,base) b 127 arch/powerpc/include/asm/ppc_asm.h #define STXVD2X_ROT(n,b,base) XXSWAPD(n,n); \ b 128 arch/powerpc/include/asm/ppc_asm.h STXVD2X(n,b,base); \ b 131 arch/powerpc/include/asm/ppc_asm.h #define LXVD2X_ROT(n,b,base) LXVD2X(n,b,base); \ b 135 arch/powerpc/include/asm/ppc_asm.h #define SAVE_VSR(n,b,base) li b,16*(n); STXVD2X_ROT(n,R##base,R##b) b 136 arch/powerpc/include/asm/ppc_asm.h #define SAVE_2VSRS(n,b,base) SAVE_VSR(n,b,base); SAVE_VSR(n+1,b,base) b 137 arch/powerpc/include/asm/ppc_asm.h #define SAVE_4VSRS(n,b,base) SAVE_2VSRS(n,b,base); SAVE_2VSRS(n+2,b,base) b 138 arch/powerpc/include/asm/ppc_asm.h #define SAVE_8VSRS(n,b,base) SAVE_4VSRS(n,b,base); SAVE_4VSRS(n+4,b,base) b 139 arch/powerpc/include/asm/ppc_asm.h #define SAVE_16VSRS(n,b,base) SAVE_8VSRS(n,b,base); SAVE_8VSRS(n+8,b,base) b 140 arch/powerpc/include/asm/ppc_asm.h #define SAVE_32VSRS(n,b,base) SAVE_16VSRS(n,b,base); SAVE_16VSRS(n+16,b,base) b 141 arch/powerpc/include/asm/ppc_asm.h #define REST_VSR(n,b,base) li b,16*(n); LXVD2X_ROT(n,R##base,R##b) b 142 arch/powerpc/include/asm/ppc_asm.h #define REST_2VSRS(n,b,base) REST_VSR(n,b,base); REST_VSR(n+1,b,base) b 143 arch/powerpc/include/asm/ppc_asm.h #define REST_4VSRS(n,b,base) REST_2VSRS(n,b,base); REST_2VSRS(n+2,b,base) b 144 arch/powerpc/include/asm/ppc_asm.h #define REST_8VSRS(n,b,base) REST_4VSRS(n,b,base); REST_4VSRS(n+4,b,base) b 145 arch/powerpc/include/asm/ppc_asm.h #define REST_16VSRS(n,b,base) REST_8VSRS(n,b,base); REST_8VSRS(n+8,b,base) b 146 arch/powerpc/include/asm/ppc_asm.h #define REST_32VSRS(n,b,base) REST_16VSRS(n,b,base); REST_16VSRS(n+16,b,base) b 152 arch/powerpc/include/asm/ppc_asm.h #define SAVE_EVR(n,s,b,o) evmergehi s,s,n; stw s,o+4*(n)(b) b 153 arch/powerpc/include/asm/ppc_asm.h #define SAVE_2EVRS(n,s,b,o) SAVE_EVR(n,s,b,o); SAVE_EVR(n+1,s,b,o) b 154 arch/powerpc/include/asm/ppc_asm.h #define SAVE_4EVRS(n,s,b,o) SAVE_2EVRS(n,s,b,o); SAVE_2EVRS(n+2,s,b,o) b 155 arch/powerpc/include/asm/ppc_asm.h #define SAVE_8EVRS(n,s,b,o) SAVE_4EVRS(n,s,b,o); SAVE_4EVRS(n+4,s,b,o) b 156 arch/powerpc/include/asm/ppc_asm.h #define SAVE_16EVRS(n,s,b,o) SAVE_8EVRS(n,s,b,o); SAVE_8EVRS(n+8,s,b,o) b 157 arch/powerpc/include/asm/ppc_asm.h #define SAVE_32EVRS(n,s,b,o) SAVE_16EVRS(n,s,b,o); SAVE_16EVRS(n+16,s,b,o) b 158 arch/powerpc/include/asm/ppc_asm.h #define REST_EVR(n,s,b,o) lwz s,o+4*(n)(b); evmergelo n,s,n b 159 arch/powerpc/include/asm/ppc_asm.h #define REST_2EVRS(n,s,b,o) REST_EVR(n,s,b,o); REST_EVR(n+1,s,b,o) b 160 arch/powerpc/include/asm/ppc_asm.h #define REST_4EVRS(n,s,b,o) REST_2EVRS(n,s,b,o); REST_2EVRS(n+2,s,b,o) b 161 arch/powerpc/include/asm/ppc_asm.h #define REST_8EVRS(n,s,b,o) REST_4EVRS(n,s,b,o); REST_4EVRS(n+4,s,b,o) b 162 arch/powerpc/include/asm/ppc_asm.h #define REST_16EVRS(n,s,b,o) REST_8EVRS(n,s,b,o); REST_8EVRS(n+8,s,b,o) b 163 arch/powerpc/include/asm/ppc_asm.h #define REST_32EVRS(n,s,b,o) REST_16EVRS(n,s,b,o); REST_16EVRS(n+16,s,b,o) b 219 arch/powerpc/include/asm/ppc_asm.h #define XGLUE(a,b) a##b b 220 arch/powerpc/include/asm/ppc_asm.h #define GLUE(a,b) XGLUE(a,b) b 516 arch/powerpc/include/asm/ppc_asm.h #define RFI rfi; b . /* Prevent prefetch past rfi */ b 782 arch/powerpc/include/asm/ppc_asm.h b 191f; /* Skip trampoline if endian is good */ \ b 802 arch/powerpc/include/asm/ppc_asm.h b 191f; /* Skip trampoline if endian is good */ \ b 116 arch/powerpc/include/asm/sstep.h u8 b[16]; b 40 arch/powerpc/include/asm/topology.h #define node_distance(a, b) __node_distance(a, b) b 41 arch/powerpc/include/asm/uaccess.h #define segment_eq(a, b) ((a).seg == (b).seg) b 97 arch/powerpc/kernel/head_booke.h b kvmppc_handler_\intno\()_\srr1 b 253 arch/powerpc/kernel/head_booke.h b 2f; \ b 404 arch/powerpc/kernel/head_booke.h b .; \ b 457 arch/powerpc/kernel/head_booke.h b .; \ b 510 arch/powerpc/kernel/head_booke.h b fast_exception_return; \ b 1161 arch/powerpc/kernel/pci-common.c struct pci_bus *b; b 1226 arch/powerpc/kernel/pci-common.c list_for_each_entry(b, &bus->children, node) b 1227 arch/powerpc/kernel/pci-common.c pcibios_allocate_bus_resources(b); b 1356 arch/powerpc/kernel/pci-common.c struct pci_bus *b; b 1359 arch/powerpc/kernel/pci-common.c list_for_each_entry(b, &pci_root_buses, node) b 1360 arch/powerpc/kernel/pci-common.c pcibios_allocate_bus_resources(b); b 1371 arch/powerpc/kernel/pci-common.c list_for_each_entry(b, &pci_root_buses, node) b 1372 arch/powerpc/kernel/pci-common.c pcibios_reserve_legacy_regions(b); b 811 arch/powerpc/kernel/process.c struct arch_hw_breakpoint *b) b 813 arch/powerpc/kernel/process.c if (a->address != b->address) b 815 arch/powerpc/kernel/process.c if (a->type != b->type) b 817 arch/powerpc/kernel/process.c if (a->len != b->len) b 2305 arch/powerpc/kernel/prom_init.c static int __init prom_set_color(ihandle ih, int i, int r, int g, int b) b 2307 arch/powerpc/kernel/prom_init.c return call_prom("call-method", 6, 1, ADDR("color!"), ih, i, b, g, r); b 1154 arch/powerpc/kernel/time.c unsigned long a, b, c, d; b 1159 arch/powerpc/kernel/time.c b = dividend_high & 0xffffffff; b 1164 arch/powerpc/kernel/time.c ra = ((u64)(a - (w * divisor)) << 32) + b; b 15 arch/powerpc/kernel/vecemu.c extern void vaddfp(vector128 *dst, vector128 *a, vector128 *b); b 16 arch/powerpc/kernel/vecemu.c extern void vsubfp(vector128 *dst, vector128 *a, vector128 *b); b 17 arch/powerpc/kernel/vecemu.c extern void vmaddfp(vector128 *dst, vector128 *a, vector128 *b, vector128 *c); b 18 arch/powerpc/kernel/vecemu.c extern void vnmsubfp(vector128 *dst, vector128 *a, vector128 *b, vector128 *c); b 413 arch/powerpc/kvm/book3s_emulate.c ulong b, t; b 416 arch/powerpc/kvm/book3s_emulate.c b = kvmppc_get_gpr(vcpu, rb); b 417 arch/powerpc/kvm/book3s_emulate.c if (!vcpu->arch.mmu.slbfee(vcpu, b, &t)) b 8 arch/powerpc/kvm/book3s_xive_template.c #define XGLUE(a,b) a##b b 9 arch/powerpc/kvm/book3s_xive_template.c #define GLUE(a,b) XGLUE(a,b) b 2018 arch/powerpc/kvm/booke.c int n, b = 0, w = 0; b 2076 arch/powerpc/kvm/booke.c if (kvmppc_booke_add_breakpoint(dbg_reg, addr, b++)) b 324 arch/powerpc/lib/sstep.c u8 b[sizeof(unsigned long)]; b 331 arch/powerpc/lib/sstep.c err = copy_mem_in(&u.b[i], ea, nb, regs); b 429 arch/powerpc/lib/sstep.c u8 b[sizeof(unsigned long)]; b 435 arch/powerpc/lib/sstep.c return copy_mem_out(&u.b[i], ea, nb, regs); b 468 arch/powerpc/lib/sstep.c u8 b[2 * sizeof(double)]; b 475 arch/powerpc/lib/sstep.c err = copy_mem_in(u.b, ea, nb, regs); b 479 arch/powerpc/lib/sstep.c do_byte_reverse(u.b, min(nb, 8)); b 481 arch/powerpc/lib/sstep.c do_byte_reverse(&u.b[8], 8); b 518 arch/powerpc/lib/sstep.c u8 b[2 * sizeof(double)]; b 545 arch/powerpc/lib/sstep.c do_byte_reverse(u.b, min(nb, 8)); b 547 arch/powerpc/lib/sstep.c do_byte_reverse(&u.b[8], 8); b 549 arch/powerpc/lib/sstep.c return copy_mem_out(u.b, ea, nb, regs); b 563 arch/powerpc/lib/sstep.c u8 b[sizeof(__vector128)]; b 570 arch/powerpc/lib/sstep.c err = copy_mem_in(&u.b[ea & 0xf], ea, size, regs); b 574 arch/powerpc/lib/sstep.c do_byte_reverse(&u.b[ea & 0xf], size); b 590 arch/powerpc/lib/sstep.c u8 b[sizeof(__vector128)]; b 605 arch/powerpc/lib/sstep.c do_byte_reverse(&u.b[ea & 0xf], size); b 606 arch/powerpc/lib/sstep.c return copy_mem_out(&u.b[ea & 0xf], ea, size, regs); b 683 arch/powerpc/lib/sstep.c memcpy(®->b[i], mem, read_size); b 685 arch/powerpc/lib/sstep.c do_byte_reverse(®->b[i], 8); b 732 arch/powerpc/lib/sstep.c reg->b[i] = *bp++; b 778 arch/powerpc/lib/sstep.c memcpy(mem, ®->b[i], write_size); b 808 arch/powerpc/lib/sstep.c *bp++ = reg->b[i]; b 25 arch/powerpc/lib/test_emulate_step.c #define TEST_LWZX(t, a, b) (PPC_INST_LWZX | ___PPC_RT(t) | \ b 26 arch/powerpc/lib/test_emulate_step.c ___PPC_RA(a) | ___PPC_RB(b)) b 29 arch/powerpc/lib/test_emulate_step.c #define TEST_LDARX(t, a, b, eh) (PPC_INST_LDARX | ___PPC_RT(t) | \ b 30 arch/powerpc/lib/test_emulate_step.c ___PPC_RA(a) | ___PPC_RB(b) | \ b 32 arch/powerpc/lib/test_emulate_step.c #define TEST_STDCX(s, a, b) (PPC_INST_STDCX | ___PPC_RS(s) | \ b 33 arch/powerpc/lib/test_emulate_step.c ___PPC_RA(a) | ___PPC_RB(b)) b 34 arch/powerpc/lib/test_emulate_step.c #define TEST_LFSX(t, a, b) (PPC_INST_LFSX | ___PPC_RT(t) | \ b 35 arch/powerpc/lib/test_emulate_step.c ___PPC_RA(a) | ___PPC_RB(b)) b 36 arch/powerpc/lib/test_emulate_step.c #define TEST_STFSX(s, a, b) (PPC_INST_STFSX | ___PPC_RS(s) | \ b 37 arch/powerpc/lib/test_emulate_step.c ___PPC_RA(a) | ___PPC_RB(b)) b 38 arch/powerpc/lib/test_emulate_step.c #define TEST_LFDX(t, a, b) (PPC_INST_LFDX | ___PPC_RT(t) | \ b 39 arch/powerpc/lib/test_emulate_step.c ___PPC_RA(a) | ___PPC_RB(b)) b 40 arch/powerpc/lib/test_emulate_step.c #define TEST_STFDX(s, a, b) (PPC_INST_STFDX | ___PPC_RS(s) | \ b 41 arch/powerpc/lib/test_emulate_step.c ___PPC_RA(a) | ___PPC_RB(b)) b 42 arch/powerpc/lib/test_emulate_step.c #define TEST_LVX(t, a, b) (PPC_INST_LVX | ___PPC_RT(t) | \ b 43 arch/powerpc/lib/test_emulate_step.c ___PPC_RA(a) | ___PPC_RB(b)) b 44 arch/powerpc/lib/test_emulate_step.c #define TEST_STVX(s, a, b) (PPC_INST_STVX | ___PPC_RS(s) | \ b 45 arch/powerpc/lib/test_emulate_step.c ___PPC_RA(a) | ___PPC_RB(b)) b 46 arch/powerpc/lib/test_emulate_step.c #define TEST_LXVD2X(s, a, b) (PPC_INST_LXVD2X | VSX_XX1((s), R##a, R##b)) b 47 arch/powerpc/lib/test_emulate_step.c #define TEST_STXVD2X(s, a, b) (PPC_INST_STXVD2X | VSX_XX1((s), R##a, R##b)) b 48 arch/powerpc/lib/test_emulate_step.c #define TEST_ADD(t, a, b) (PPC_INST_ADD | ___PPC_RT(t) | \ b 49 arch/powerpc/lib/test_emulate_step.c ___PPC_RA(a) | ___PPC_RB(b)) b 50 arch/powerpc/lib/test_emulate_step.c #define TEST_ADD_DOT(t, a, b) (PPC_INST_ADD | ___PPC_RT(t) | \ b 51 arch/powerpc/lib/test_emulate_step.c ___PPC_RA(a) | ___PPC_RB(b) | 0x1) b 52 arch/powerpc/lib/test_emulate_step.c #define TEST_ADDC(t, a, b) (PPC_INST_ADDC | ___PPC_RT(t) | \ b 53 arch/powerpc/lib/test_emulate_step.c ___PPC_RA(a) | ___PPC_RB(b)) b 54 arch/powerpc/lib/test_emulate_step.c #define TEST_ADDC_DOT(t, a, b) (PPC_INST_ADDC | ___PPC_RT(t) | \ b 55 arch/powerpc/lib/test_emulate_step.c ___PPC_RA(a) | ___PPC_RB(b) | 0x1) b 228 arch/powerpc/lib/test_emulate_step.c int b; b 239 arch/powerpc/lib/test_emulate_step.c cached_b = c.b; b 260 arch/powerpc/lib/test_emulate_step.c if (stepped == 1 && c.b == cached_b) b 271 arch/powerpc/lib/test_emulate_step.c long b; b 282 arch/powerpc/lib/test_emulate_step.c cached_b = c.b; b 303 arch/powerpc/lib/test_emulate_step.c if (stepped == 1 && c.b == cached_b) b 328 arch/powerpc/lib/test_emulate_step.c u32 b[4]; b 338 arch/powerpc/lib/test_emulate_step.c cached_b[0] = c.b[0] = 923745; b 339 arch/powerpc/lib/test_emulate_step.c cached_b[1] = c.b[1] = 2139478; b 340 arch/powerpc/lib/test_emulate_step.c cached_b[2] = c.b[2] = 9012; b 341 arch/powerpc/lib/test_emulate_step.c cached_b[3] = c.b[3] = 982134; b 357 arch/powerpc/lib/test_emulate_step.c c.b[0] = 4987513; b 358 arch/powerpc/lib/test_emulate_step.c c.b[1] = 84313948; b 359 arch/powerpc/lib/test_emulate_step.c c.b[2] = 71; b 360 arch/powerpc/lib/test_emulate_step.c c.b[3] = 498532; b 365 arch/powerpc/lib/test_emulate_step.c if (stepped == 1 && cached_b[0] == c.b[0] && cached_b[1] == c.b[1] && b 366 arch/powerpc/lib/test_emulate_step.c cached_b[2] == c.b[2] && cached_b[3] == c.b[3]) b 385 arch/powerpc/lib/test_emulate_step.c u32 b[4]; b 395 arch/powerpc/lib/test_emulate_step.c cached_b[0] = c.b[0] = 18233; b 396 arch/powerpc/lib/test_emulate_step.c cached_b[1] = c.b[1] = 34863571; b 397 arch/powerpc/lib/test_emulate_step.c cached_b[2] = c.b[2] = 834; b 398 arch/powerpc/lib/test_emulate_step.c cached_b[3] = c.b[3] = 6138911; b 418 arch/powerpc/lib/test_emulate_step.c c.b[0] = 21379463; b 419 arch/powerpc/lib/test_emulate_step.c c.b[1] = 87; b 420 arch/powerpc/lib/test_emulate_step.c c.b[2] = 374234; b 421 arch/powerpc/lib/test_emulate_step.c c.b[3] = 4; b 426 arch/powerpc/lib/test_emulate_step.c if (stepped == 1 && cached_b[0] == c.b[0] && cached_b[1] == c.b[1] && b 427 arch/powerpc/lib/test_emulate_step.c cached_b[2] == c.b[2] && cached_b[3] == c.b[3] && b 16 arch/powerpc/lib/xor_vmx.c #define vec_xor(a, b) a ^ b b 15 arch/powerpc/math-emu/udivmodti4.c _FP_I_TYPE b, bm; b 95 arch/powerpc/math-emu/udivmodti4.c b = _FP_W_TYPE_SIZE - bm; b 98 arch/powerpc/math-emu/udivmodti4.c n2 = n1 >> b; b 99 arch/powerpc/math-emu/udivmodti4.c n1 = (n1 << bm) | (n0 >> b); b 163 arch/powerpc/math-emu/udivmodti4.c b = _FP_W_TYPE_SIZE - bm; b 165 arch/powerpc/math-emu/udivmodti4.c d1 = (d1 << bm) | (d0 >> b); b 167 arch/powerpc/math-emu/udivmodti4.c n2 = n1 >> b; b 168 arch/powerpc/math-emu/udivmodti4.c n1 = (n1 << bm) | (n0 >> b); b 184 arch/powerpc/math-emu/udivmodti4.c r0 = (n1 << b) | (n0 >> bm); b 52 arch/powerpc/mm/book3s32/mmu.c int b; b 53 arch/powerpc/mm/book3s32/mmu.c for (b = 0; b < ARRAY_SIZE(bat_addrs); ++b) b 54 arch/powerpc/mm/book3s32/mmu.c if (va >= bat_addrs[b].start && va < bat_addrs[b].limit) b 55 arch/powerpc/mm/book3s32/mmu.c return bat_addrs[b].phys + (va - bat_addrs[b].start); b 64 arch/powerpc/mm/book3s32/mmu.c int b; b 65 arch/powerpc/mm/book3s32/mmu.c for (b = 0; b < ARRAY_SIZE(bat_addrs); ++b) b 66 arch/powerpc/mm/book3s32/mmu.c if (pa >= bat_addrs[b].phys b 67 arch/powerpc/mm/book3s32/mmu.c && pa < (bat_addrs[b].limit-bat_addrs[b].start) b 68 arch/powerpc/mm/book3s32/mmu.c +bat_addrs[b].phys) b 69 arch/powerpc/mm/book3s32/mmu.c return bat_addrs[b].start+(pa-bat_addrs[b].phys); b 75 arch/powerpc/mm/book3s32/mmu.c int b; b 78 arch/powerpc/mm/book3s32/mmu.c for (b = 0; b < 4; b++) { b 79 arch/powerpc/mm/book3s32/mmu.c struct ppc_bat *bat = BATS[b]; b 82 arch/powerpc/mm/book3s32/mmu.c return b; b 87 arch/powerpc/mm/book3s32/mmu.c for (b = 0; b < n; b++) { b 88 arch/powerpc/mm/book3s32/mmu.c struct ppc_bat *bat = BATS[b]; b 91 arch/powerpc/mm/book3s32/mmu.c return b; b 76 arch/powerpc/mm/nohash/fsl_booke.c int b; b 77 arch/powerpc/mm/nohash/fsl_booke.c for (b = 0; b < tlbcam_index; ++b) b 78 arch/powerpc/mm/nohash/fsl_booke.c if (va >= tlbcam_addrs[b].start && va < tlbcam_addrs[b].limit) b 79 arch/powerpc/mm/nohash/fsl_booke.c return tlbcam_addrs[b].phys + (va - tlbcam_addrs[b].start); b 88 arch/powerpc/mm/nohash/fsl_booke.c int b; b 89 arch/powerpc/mm/nohash/fsl_booke.c for (b = 0; b < tlbcam_index; ++b) b 90 arch/powerpc/mm/nohash/fsl_booke.c if (pa >= tlbcam_addrs[b].phys b 91 arch/powerpc/mm/nohash/fsl_booke.c && pa < (tlbcam_addrs[b].limit-tlbcam_addrs[b].start) b 92 arch/powerpc/mm/nohash/fsl_booke.c +tlbcam_addrs[b].phys) b 93 arch/powerpc/mm/nohash/fsl_booke.c return tlbcam_addrs[b].start+(pa-tlbcam_addrs[b].phys); b 188 arch/powerpc/mm/numa.c int __node_distance(int a, int b) b 194 arch/powerpc/mm/numa.c return ((a == b) ? LOCAL_DISTANCE : REMOTE_DISTANCE); b 197 arch/powerpc/mm/numa.c if (distance_lookup_table[a][i] == distance_lookup_table[b][i]) b 50 arch/powerpc/net/bpf_jit.h #define PPC_STDX(r, base, b) EMIT(PPC_INST_STDX | ___PPC_RS(r) | \ b 51 arch/powerpc/net/bpf_jit.h ___PPC_RA(base) | ___PPC_RB(b)) b 67 arch/powerpc/net/bpf_jit.h #define PPC_LDX(r, base, b) EMIT(PPC_INST_LDX | ___PPC_RT(r) | \ b 68 arch/powerpc/net/bpf_jit.h ___PPC_RA(base) | ___PPC_RB(b)) b 73 arch/powerpc/net/bpf_jit.h #define PPC_LHBRX(r, base, b) EMIT(PPC_INST_LHBRX | ___PPC_RT(r) | \ b 74 arch/powerpc/net/bpf_jit.h ___PPC_RA(base) | ___PPC_RB(b)) b 75 arch/powerpc/net/bpf_jit.h #define PPC_LDBRX(r, base, b) EMIT(PPC_INST_LDBRX | ___PPC_RT(r) | \ b 76 arch/powerpc/net/bpf_jit.h ___PPC_RA(base) | ___PPC_RB(b)) b 78 arch/powerpc/net/bpf_jit.h #define PPC_BPF_LDARX(t, a, b, eh) EMIT(PPC_INST_LDARX | ___PPC_RT(t) | \ b 79 arch/powerpc/net/bpf_jit.h ___PPC_RA(a) | ___PPC_RB(b) | \ b 81 arch/powerpc/net/bpf_jit.h #define PPC_BPF_LWARX(t, a, b, eh) EMIT(PPC_INST_LWARX | ___PPC_RT(t) | \ b 82 arch/powerpc/net/bpf_jit.h ___PPC_RA(a) | ___PPC_RB(b) | \ b 84 arch/powerpc/net/bpf_jit.h #define PPC_BPF_STWCX(s, a, b) EMIT(PPC_INST_STWCX | ___PPC_RS(s) | \ b 85 arch/powerpc/net/bpf_jit.h ___PPC_RA(a) | ___PPC_RB(b)) b 86 arch/powerpc/net/bpf_jit.h #define PPC_BPF_STDCX(s, a, b) EMIT(PPC_INST_STDCX | ___PPC_RS(s) | \ b 87 arch/powerpc/net/bpf_jit.h ___PPC_RA(a) | ___PPC_RB(b)) b 90 arch/powerpc/net/bpf_jit.h #define PPC_CMPW(a, b) EMIT(PPC_INST_CMPW | ___PPC_RA(a) | \ b 91 arch/powerpc/net/bpf_jit.h ___PPC_RB(b)) b 92 arch/powerpc/net/bpf_jit.h #define PPC_CMPD(a, b) EMIT(PPC_INST_CMPD | ___PPC_RA(a) | \ b 93 arch/powerpc/net/bpf_jit.h ___PPC_RB(b)) b 96 arch/powerpc/net/bpf_jit.h #define PPC_CMPLW(a, b) EMIT(PPC_INST_CMPLW | ___PPC_RA(a) | \ b 97 arch/powerpc/net/bpf_jit.h ___PPC_RB(b)) b 98 arch/powerpc/net/bpf_jit.h #define PPC_CMPLD(a, b) EMIT(PPC_INST_CMPLD | ___PPC_RA(a) | \ b 99 arch/powerpc/net/bpf_jit.h ___PPC_RB(b)) b 101 arch/powerpc/net/bpf_jit.h #define PPC_SUB(d, a, b) EMIT(PPC_INST_SUB | ___PPC_RT(d) | \ b 102 arch/powerpc/net/bpf_jit.h ___PPC_RB(a) | ___PPC_RA(b)) b 103 arch/powerpc/net/bpf_jit.h #define PPC_ADD(d, a, b) EMIT(PPC_INST_ADD | ___PPC_RT(d) | \ b 104 arch/powerpc/net/bpf_jit.h ___PPC_RA(a) | ___PPC_RB(b)) b 105 arch/powerpc/net/bpf_jit.h #define PPC_MULD(d, a, b) EMIT(PPC_INST_MULLD | ___PPC_RT(d) | \ b 106 arch/powerpc/net/bpf_jit.h ___PPC_RA(a) | ___PPC_RB(b)) b 107 arch/powerpc/net/bpf_jit.h #define PPC_MULW(d, a, b) EMIT(PPC_INST_MULLW | ___PPC_RT(d) | \ b 108 arch/powerpc/net/bpf_jit.h ___PPC_RA(a) | ___PPC_RB(b)) b 109 arch/powerpc/net/bpf_jit.h #define PPC_MULHWU(d, a, b) EMIT(PPC_INST_MULHWU | ___PPC_RT(d) | \ b 110 arch/powerpc/net/bpf_jit.h ___PPC_RA(a) | ___PPC_RB(b)) b 113 arch/powerpc/net/bpf_jit.h #define PPC_DIVWU(d, a, b) EMIT(PPC_INST_DIVWU | ___PPC_RT(d) | \ b 114 arch/powerpc/net/bpf_jit.h ___PPC_RA(a) | ___PPC_RB(b)) b 115 arch/powerpc/net/bpf_jit.h #define PPC_DIVDU(d, a, b) EMIT(PPC_INST_DIVDU | ___PPC_RT(d) | \ b 116 arch/powerpc/net/bpf_jit.h ___PPC_RA(a) | ___PPC_RB(b)) b 117 arch/powerpc/net/bpf_jit.h #define PPC_AND(d, a, b) EMIT(PPC_INST_AND | ___PPC_RA(d) | \ b 118 arch/powerpc/net/bpf_jit.h ___PPC_RS(a) | ___PPC_RB(b)) b 121 arch/powerpc/net/bpf_jit.h #define PPC_AND_DOT(d, a, b) EMIT(PPC_INST_ANDDOT | ___PPC_RA(d) | \ b 122 arch/powerpc/net/bpf_jit.h ___PPC_RS(a) | ___PPC_RB(b)) b 123 arch/powerpc/net/bpf_jit.h #define PPC_OR(d, a, b) EMIT(PPC_INST_OR | ___PPC_RA(d) | \ b 124 arch/powerpc/net/bpf_jit.h ___PPC_RS(a) | ___PPC_RB(b)) b 130 arch/powerpc/net/bpf_jit.h #define PPC_XOR(d, a, b) EMIT(PPC_INST_XOR | ___PPC_RA(d) | \ b 131 arch/powerpc/net/bpf_jit.h ___PPC_RS(a) | ___PPC_RB(b)) b 214 arch/powerpc/perf/isa207-common.h #define P(a, b) PERF_MEM_S(a, b) b 215 arch/powerpc/perf/isa207-common.h #define PH(a, b) (P(LVL, HIT) | P(a, b)) b 216 arch/powerpc/perf/isa207-common.h #define PM(a, b) (P(LVL, MISS) | P(a, b)) b 179 arch/powerpc/perf/power6-pmu.c unsigned int pmc, ev, b, u, s, psel; b 211 arch/powerpc/perf/power6-pmu.c b = (ev >> PM_BYTE_SH) & PM_BYTE_MSK; b 214 arch/powerpc/perf/power6-pmu.c if ((ttmset & (1 << b)) && MMCR1_TTMSEL(mmcr1, b) != u) b 216 arch/powerpc/perf/power6-pmu.c mmcr1 |= (unsigned long)u << MMCR1_TTMSEL_SH(b); b 217 arch/powerpc/perf/power6-pmu.c ttmset |= 1 << b; b 229 arch/powerpc/perf/power6-pmu.c if (b >= 2) b 28 arch/powerpc/perf/req-gen/perf.h #define CAT2_(a, b) a ## b b 29 arch/powerpc/perf/req-gen/perf.h #define CAT2(a, b) CAT2_(a, b) b 30 arch/powerpc/perf/req-gen/perf.h #define CAT3_(a, b, c) a ## b ## c b 31 arch/powerpc/perf/req-gen/perf.h #define CAT3(a, b, c) CAT3_(a, b, c) b 177 arch/powerpc/platforms/85xx/p1022_ds.c u8 b; b 314 arch/powerpc/platforms/85xx/p1022_ds.c b = in_8(lbc_lcs1_ba); b 315 arch/powerpc/platforms/85xx/p1022_ds.c b |= PX_BRDCFG0_ELBC_DIU; b 316 arch/powerpc/platforms/85xx/p1022_ds.c out_8(lbc_lcs1_ba, b); b 329 arch/powerpc/platforms/85xx/p1022_ds.c b = in_8(lbc_lcs1_ba); b 330 arch/powerpc/platforms/85xx/p1022_ds.c b &= ~(PX_BRDCFG1_DFPEN | PX_BRDCFG1_BACKLIGHT); b 331 arch/powerpc/platforms/85xx/p1022_ds.c b |= PX_BRDCFG1_DVIEN; b 332 arch/powerpc/platforms/85xx/p1022_ds.c out_8(lbc_lcs1_ba, b); b 341 arch/powerpc/platforms/85xx/p1022_ds.c b = in_8(lbc_lcs1_ba); b 342 arch/powerpc/platforms/85xx/p1022_ds.c b &= ~PX_BRDCFG1_DVIEN; b 343 arch/powerpc/platforms/85xx/p1022_ds.c b |= PX_BRDCFG1_DFPEN | PX_BRDCFG1_BACKLIGHT; b 344 arch/powerpc/platforms/85xx/p1022_ds.c out_8(lbc_lcs1_ba, b); b 1063 arch/powerpc/platforms/cell/spufs/sched.c int a, b, c; b 1066 arch/powerpc/platforms/cell/spufs/sched.c b = spu_avenrun[1] + (FIXED_1/200); b 1076 arch/powerpc/platforms/cell/spufs/sched.c LOAD_INT(b), LOAD_FRAC(b), b 22 arch/powerpc/platforms/powermac/pfunc_core.c #define LOG_BLOB(t,b,c) b 116 arch/powerpc/platforms/powernv/opal-lpc.c static void opal_lpc_insb(unsigned long p, void *b, unsigned long c) b 118 arch/powerpc/platforms/powernv/opal-lpc.c u8 *ptr = b; b 124 arch/powerpc/platforms/powernv/opal-lpc.c static void opal_lpc_insw(unsigned long p, void *b, unsigned long c) b 126 arch/powerpc/platforms/powernv/opal-lpc.c __le16 *ptr = b; b 132 arch/powerpc/platforms/powernv/opal-lpc.c static void opal_lpc_insl(unsigned long p, void *b, unsigned long c) b 134 arch/powerpc/platforms/powernv/opal-lpc.c __le32 *ptr = b; b 140 arch/powerpc/platforms/powernv/opal-lpc.c static void opal_lpc_outsb(unsigned long p, const void *b, unsigned long c) b 142 arch/powerpc/platforms/powernv/opal-lpc.c const u8 *ptr = b; b 148 arch/powerpc/platforms/powernv/opal-lpc.c static void opal_lpc_outsw(unsigned long p, const void *b, unsigned long c) b 150 arch/powerpc/platforms/powernv/opal-lpc.c const __le16 *ptr = b; b 156 arch/powerpc/platforms/powernv/opal-lpc.c static void opal_lpc_outsl(unsigned long p, const void *b, unsigned long c) b 158 arch/powerpc/platforms/powernv/opal-lpc.c const __le32 *ptr = b; b 43 arch/powerpc/platforms/ps3/mm.c static unsigned long make_page_sizes(unsigned long a, unsigned long b) b 45 arch/powerpc/platforms/ps3/mm.c return (a << 56) | (b << 48); b 68 arch/powerpc/platforms/pseries/pci_dlpar.c struct pci_bus *b = phb->bus; b 73 arch/powerpc/platforms/pseries/pci_dlpar.c pci_domain_nr(b), b->number); b 76 arch/powerpc/platforms/pseries/pci_dlpar.c if (!(list_empty(&b->children) && list_empty(&b->devices))) b 84 arch/powerpc/platforms/pseries/pci_dlpar.c rc = pcibios_unmap_io_space(b); b 87 arch/powerpc/platforms/pseries/pci_dlpar.c __func__, b->name); b 94 arch/powerpc/platforms/pseries/pci_dlpar.c pci_remove_bus(b); b 95 arch/powerpc/platforms/pseries/pci_dlpar.c device_unregister(b->bridge); b 18 arch/powerpc/sysdev/grackle.c #define GRACKLE_CFA(b, d, o) (0x80 | ((b) << 8) | ((d) << 16) \ b 297 arch/powerpc/sysdev/mpic.c #define mpic_read(b,r) _mpic_read(mpic->reg_type,&(b),(r)) b 298 arch/powerpc/sysdev/mpic.c #define mpic_write(b,r,v) _mpic_write(mpic->reg_type,&(b),(r),(v)) b 341 arch/powerpc/sysdev/mpic.c #define mpic_map(m,p,b,o,s) _mpic_map_mmio(m,p,b,o,s) b 117 arch/powerpc/xmon/spu-insns.h #define _A2(a,b) {2,a,b} b 118 arch/powerpc/xmon/spu-insns.h #define _A3(a,b,c) {3,a,b,c} b 119 arch/powerpc/xmon/spu-insns.h #define _A4(a,b,c,d) {4,a,b,c,d} b 2147 arch/powerpc/xmon/xmon.c #define SWAP(a, b, t) ((t) = (a), (a) = (b), (b) = (t)) b 16 arch/riscv/include/asm/asm.h #define __REG_SEL(a, b) __ASM_STR(a) b 18 arch/riscv/include/asm/asm.h #define __REG_SEL(a, b) __ASM_STR(b) b 245 arch/riscv/include/asm/io.h __io_reads_ins(reads, u8, b, __io_br(), __io_ar(addr)) b 252 arch/riscv/include/asm/io.h __io_reads_ins(ins, u8, b, __io_pbr(), __io_par(addr)) b 259 arch/riscv/include/asm/io.h __io_writes_outs(writes, u8, b, __io_bw(), __io_aw()) b 266 arch/riscv/include/asm/io.h __io_writes_outs(outs, u8, b, __io_pbw(), __io_paw()) b 46 arch/riscv/include/asm/uaccess.h #define segment_eq(a, b) ((a).seg == (b).seg) b 28 arch/s390/include/asm/chpid.h static inline int chp_id_is_equal(struct chp_id *a, struct chp_id *b) b 30 arch/s390/include/asm/chpid.h return (a->id == b->id) && (a->cssid == b->cssid); b 112 arch/s390/include/asm/cio.h __u32 b : 1; b 51 arch/s390/include/asm/nmi.h u64 b : 1; /* 14 backed up */ b 138 arch/s390/include/asm/nospec-insn.h 556: b \disp(\reg) b 186 arch/s390/include/asm/nospec-insn.h b \disp(\reg) b 20 arch/s390/include/asm/numa.h int __node_distance(int a, int b); b 806 arch/s390/include/asm/pgtable.h static inline int pte_same(pte_t a, pte_t b) b 808 arch/s390/include/asm/pgtable.h return pte_val(a) == pte_val(b); b 235 arch/s390/include/asm/timex.h static inline int tod_after(unsigned long long a, unsigned long long b) b 238 arch/s390/include/asm/timex.h return (long long) a > (long long) b; b 239 arch/s390/include/asm/timex.h return a > b; b 249 arch/s390/include/asm/timex.h static inline int tod_after_eq(unsigned long long a, unsigned long long b) b 252 arch/s390/include/asm/timex.h return (long long) a >= (long long) b; b 253 arch/s390/include/asm/timex.h return a >= b; b 86 arch/s390/include/asm/topology.h #define node_distance(a, b) __node_distance(a, b) b 35 arch/s390/include/asm/uaccess.h #define segment_eq(a,b) (((a) & 2) == ((b) & 2)) b 21 arch/s390/include/uapi/asm/guarded_storage.h __u8 b : 1; b 40 arch/s390/include/uapi/asm/runtime_instr.h __u32 b : 1; b 26 arch/s390/include/uapi/asm/vtoc.h __u8 b; b 177 arch/s390/include/uapi/asm/vtoc.h __u32 b; /* ending RTA value + 1 */ b 362 arch/s390/kernel/nmi.c if (mci.b) { b 545 arch/s390/kernel/perf_cpum_cf_events.c struct attribute **b, b 553 arch/s390/kernel/perf_cpum_cf_events.c for (i = 0; b[i]; i++) b 565 arch/s390/kernel/perf_cpum_cf_events.c for (i = 0; b[i]; i++) b 566 arch/s390/kernel/perf_cpum_cf_events.c new[j++] = b[i]; b 252 arch/s390/kvm/gaccess.c unsigned long b : 1; /* Base-Space Bit */ b 302 arch/s390/kvm/guestdbg.c static inline int in_addr_range(u64 addr, u64 a, u64 b) b 304 arch/s390/kvm/guestdbg.c if (a <= b) b 305 arch/s390/kvm/guestdbg.c return (addr >= a) && (addr <= b); b 308 arch/s390/kvm/guestdbg.c return (addr >= a) || (addr <= b); b 22 arch/s390/mm/hugetlbpage.c #define move_set_bit(x, a, b) (((x) & (a)) >> ilog2(a) << ilog2(b)) b 254 arch/s390/net/bpf_jit_comp.c #define EMIT6_PCREL_RILB(op, b, target) \ b 257 arch/s390/net/bpf_jit_comp.c _EMIT6(op | reg_high(b) << 16 | rel >> 16, rel & 0xffff); \ b 258 arch/s390/net/bpf_jit_comp.c REG_SET_SEEN(b); \ b 52 arch/s390/numa/numa.c int __node_distance(int a, int b) b 54 arch/s390/numa/numa.c return mode->distance ? mode->distance(a, b) : 0; b 19 arch/s390/numa/numa_mode.h int (*distance)(int a, int b); /* Distance between two nodes */ b 175 arch/s390/tools/gen_opcode_table.c static int cmpformat(const void *a, const void *b) b 177 arch/s390/tools/gen_opcode_table.c return strcmp(((struct insn *)a)->format, ((struct insn *)b)->format); b 199 arch/s390/tools/gen_opcode_table.c static int cmp_long_insn(const void *a, const void *b) b 201 arch/s390/tools/gen_opcode_table.c return strcmp(((struct insn *)a)->name, ((struct insn *)b)->name); b 266 arch/s390/tools/gen_opcode_table.c static int cmpopcode(const void *a, const void *b) b 268 arch/s390/tools/gen_opcode_table.c return strcmp(((struct insn *)a)->opcode, ((struct insn *)b)->opcode); b 85 arch/sh/include/asm/io.h __BUILD_UNCACHED_IO(b, u8) b 115 arch/sh/include/asm/io.h __BUILD_MEMORY_STRING(__raw_, b, u8) b 189 arch/sh/include/asm/io.h BUILDIO_IOPORT(b, u8) b 218 arch/sh/include/asm/io.h __BUILD_IOPORT_STRING(b, u8) b 8 arch/sh/include/asm/io_generic.h #define IO_CONCAT(a,b) _IO_CONCAT(a,b) b 9 arch/sh/include/asm/io_generic.h #define _IO_CONCAT(a,b) a ## _ ## b b 56 arch/sh/include/asm/io_noioport.h #define insb(a, b, c) BUG() b 57 arch/sh/include/asm/io_noioport.h #define insw(a, b, c) BUG() b 58 arch/sh/include/asm/io_noioport.h #define insl(a, b, c) BUG() b 60 arch/sh/include/asm/io_noioport.h #define outsb(a, b, c) BUG() b 61 arch/sh/include/asm/io_noioport.h #define outsw(a, b, c) BUG() b 62 arch/sh/include/asm/io_noioport.h #define outsl(a, b, c) BUG() b 39 arch/sh/include/asm/romimage-macros.h mov.b r0, @r1 b 27 arch/sh/include/asm/segment.h #define segment_eq(a, b) ((a).seg == (b).seg) b 50 arch/sh/include/uapi/asm/swab.h struct { __u32 a,b; } s; b 54 arch/sh/include/uapi/asm/swab.h w.s.b = __arch_swab32(v.s.a); b 55 arch/sh/include/uapi/asm/swab.h w.s.a = __arch_swab32(v.s.b); b 24 arch/sh/kernel/cpu/sh4/fpu.c unsigned long long b); b 25 arch/sh/kernel/cpu/sh4/fpu.c extern unsigned long int float32_div(unsigned long int a, unsigned long int b); b 27 arch/sh/kernel/cpu/sh4/fpu.c unsigned long long b); b 28 arch/sh/kernel/cpu/sh4/fpu.c extern unsigned long int float32_mul(unsigned long int a, unsigned long int b); b 30 arch/sh/kernel/cpu/sh4/fpu.c unsigned long long b); b 31 arch/sh/kernel/cpu/sh4/fpu.c extern unsigned long int float32_add(unsigned long int a, unsigned long int b); b 33 arch/sh/kernel/cpu/sh4/fpu.c unsigned long long b); b 34 arch/sh/kernel/cpu/sh4/fpu.c extern unsigned long int float32_sub(unsigned long int a, unsigned long int b); b 81 arch/sh/kernel/cpu/sh4/softfloat.c float64 float64_sub(float64 a, float64 b); b 82 arch/sh/kernel/cpu/sh4/softfloat.c float32 float32_sub(float32 a, float32 b); b 83 arch/sh/kernel/cpu/sh4/softfloat.c float32 float32_add(float32 a, float32 b); b 84 arch/sh/kernel/cpu/sh4/softfloat.c float64 float64_add(float64 a, float64 b); b 85 arch/sh/kernel/cpu/sh4/softfloat.c float64 float64_div(float64 a, float64 b); b 86 arch/sh/kernel/cpu/sh4/softfloat.c float32 float32_div(float32 a, float32 b); b 87 arch/sh/kernel/cpu/sh4/softfloat.c float32 float32_mul(float32 a, float32 b); b 88 arch/sh/kernel/cpu/sh4/softfloat.c float64 float64_mul(float64 a, float64 b); b 94 arch/sh/kernel/cpu/sh4/softfloat.c void mul64To128(bits64 a, bits64 b, bits64 * z0Ptr, bits64 * z1Ptr); b 100 arch/sh/kernel/cpu/sh4/softfloat.c static float64 subFloat64Sigs(float64 a, float64 b, flag zSign); b 101 arch/sh/kernel/cpu/sh4/softfloat.c static float64 addFloat64Sigs(float64 a, float64 b, flag zSign); b 106 arch/sh/kernel/cpu/sh4/softfloat.c static float32 subFloat32Sigs(float32 a, float32 b, flag zSign); b 107 arch/sh/kernel/cpu/sh4/softfloat.c static float32 addFloat32Sigs(float32 a, float32 b, flag zSign); b 110 arch/sh/kernel/cpu/sh4/softfloat.c static bits64 estimateDiv128To64(bits64 a0, bits64 a1, bits64 b); b 224 arch/sh/kernel/cpu/sh4/softfloat.c static float64 subFloat64Sigs(float64 a, float64 b, flag zSign) b 232 arch/sh/kernel/cpu/sh4/softfloat.c bSig = extractFloat64Frac(b); b 233 arch/sh/kernel/cpu/sh4/softfloat.c bExp = extractFloat64Exp(b); b 285 arch/sh/kernel/cpu/sh4/softfloat.c static float64 addFloat64Sigs(float64 a, float64 b, flag zSign) b 293 arch/sh/kernel/cpu/sh4/softfloat.c bSig = extractFloat64Frac(b); b 294 arch/sh/kernel/cpu/sh4/softfloat.c bExp = extractFloat64Exp(b); b 454 arch/sh/kernel/cpu/sh4/softfloat.c static float32 subFloat32Sigs(float32 a, float32 b, flag zSign) b 462 arch/sh/kernel/cpu/sh4/softfloat.c bSig = extractFloat32Frac(b); b 463 arch/sh/kernel/cpu/sh4/softfloat.c bExp = extractFloat32Exp(b); b 516 arch/sh/kernel/cpu/sh4/softfloat.c static float32 addFloat32Sigs(float32 a, float32 b, flag zSign) b 524 arch/sh/kernel/cpu/sh4/softfloat.c bSig = extractFloat32Frac(b); b 525 arch/sh/kernel/cpu/sh4/softfloat.c bExp = extractFloat32Exp(b); b 573 arch/sh/kernel/cpu/sh4/softfloat.c float64 float64_sub(float64 a, float64 b) b 578 arch/sh/kernel/cpu/sh4/softfloat.c bSign = extractFloat64Sign(b); b 580 arch/sh/kernel/cpu/sh4/softfloat.c return subFloat64Sigs(a, b, aSign); b 582 arch/sh/kernel/cpu/sh4/softfloat.c return addFloat64Sigs(a, b, aSign); b 587 arch/sh/kernel/cpu/sh4/softfloat.c float32 float32_sub(float32 a, float32 b) b 592 arch/sh/kernel/cpu/sh4/softfloat.c bSign = extractFloat32Sign(b); b 594 arch/sh/kernel/cpu/sh4/softfloat.c return subFloat32Sigs(a, b, aSign); b 596 arch/sh/kernel/cpu/sh4/softfloat.c return addFloat32Sigs(a, b, aSign); b 601 arch/sh/kernel/cpu/sh4/softfloat.c float32 float32_add(float32 a, float32 b) b 606 arch/sh/kernel/cpu/sh4/softfloat.c bSign = extractFloat32Sign(b); b 608 arch/sh/kernel/cpu/sh4/softfloat.c return addFloat32Sigs(a, b, aSign); b 610 arch/sh/kernel/cpu/sh4/softfloat.c return subFloat32Sigs(a, b, aSign); b 615 arch/sh/kernel/cpu/sh4/softfloat.c float64 float64_add(float64 a, float64 b) b 620 arch/sh/kernel/cpu/sh4/softfloat.c bSign = extractFloat64Sign(b); b 622 arch/sh/kernel/cpu/sh4/softfloat.c return addFloat64Sigs(a, b, aSign); b 624 arch/sh/kernel/cpu/sh4/softfloat.c return subFloat64Sigs(a, b, aSign); b 656 arch/sh/kernel/cpu/sh4/softfloat.c static bits64 estimateDiv128To64(bits64 a0, bits64 a1, bits64 b) b 661 arch/sh/kernel/cpu/sh4/softfloat.c if (b <= a0) b 663 arch/sh/kernel/cpu/sh4/softfloat.c b0 = b >> 32; b 668 arch/sh/kernel/cpu/sh4/softfloat.c mul64To128(b, z, &term0, &term1); b 672 arch/sh/kernel/cpu/sh4/softfloat.c b1 = b << 32; b 682 arch/sh/kernel/cpu/sh4/softfloat.c void mul64To128(bits64 a, bits64 b, bits64 * z0Ptr, bits64 * z1Ptr) b 689 arch/sh/kernel/cpu/sh4/softfloat.c bLow = b; b 690 arch/sh/kernel/cpu/sh4/softfloat.c bHigh = b >> 32; b 716 arch/sh/kernel/cpu/sh4/softfloat.c float64 float64_div(float64 a, float64 b) b 727 arch/sh/kernel/cpu/sh4/softfloat.c bSig = extractFloat64Frac(b); b 728 arch/sh/kernel/cpu/sh4/softfloat.c bExp = extractFloat64Exp(b); b 729 arch/sh/kernel/cpu/sh4/softfloat.c bSign = extractFloat64Sign(b); b 774 arch/sh/kernel/cpu/sh4/softfloat.c float32 float32_div(float32 a, float32 b) b 784 arch/sh/kernel/cpu/sh4/softfloat.c bSig = extractFloat32Frac(b); b 785 arch/sh/kernel/cpu/sh4/softfloat.c bExp = extractFloat32Exp(b); b 786 arch/sh/kernel/cpu/sh4/softfloat.c bSign = extractFloat32Sign(b); b 824 arch/sh/kernel/cpu/sh4/softfloat.c float32 float32_mul(float32 a, float32 b) b 835 arch/sh/kernel/cpu/sh4/softfloat.c bSig = extractFloat32Frac(b); b 836 arch/sh/kernel/cpu/sh4/softfloat.c bExp = extractFloat32Exp(b); b 837 arch/sh/kernel/cpu/sh4/softfloat.c bSign = extractFloat32Sign(b); b 865 arch/sh/kernel/cpu/sh4/softfloat.c float64 float64_mul(float64 a, float64 b) b 874 arch/sh/kernel/cpu/sh4/softfloat.c bSig = extractFloat64Frac(b); b 875 arch/sh/kernel/cpu/sh4/softfloat.c bExp = extractFloat64Exp(b); b 876 arch/sh/kernel/cpu/sh4/softfloat.c bSign = extractFloat64Sign(b); b 71 arch/sh/kernel/traps_64.c __u64 b; b 73 arch/sh/kernel/traps_64.c b = (__u64)(__s64)(__s32)(a & 0xffffffffUL); b 74 arch/sh/kernel/traps_64.c return (b == a) ? 1 : 0; b 6 arch/sh/lib/ashldi3.c long long __ashldi3(long long u, word_type b) b 11 arch/sh/lib/ashldi3.c if (b == 0) b 15 arch/sh/lib/ashldi3.c bm = 32 - b; b 23 arch/sh/lib/ashldi3.c w.s.low = (unsigned int) uu.s.low << b; b 24 arch/sh/lib/ashldi3.c w.s.high = ((unsigned int) uu.s.high << b) | carries; b 6 arch/sh/lib/ashrdi3.c long long __ashrdi3(long long u, word_type b) b 11 arch/sh/lib/ashrdi3.c if (b == 0) b 15 arch/sh/lib/ashrdi3.c bm = 32 - b; b 25 arch/sh/lib/ashrdi3.c w.s.high = uu.s.high >> b; b 26 arch/sh/lib/ashrdi3.c w.s.low = ((unsigned int) uu.s.low >> b) | carries; b 6 arch/sh/lib/lshrdi3.c long long __lshrdi3(long long u, word_type b) b 11 arch/sh/lib/lshrdi3.c if (b == 0) b 15 arch/sh/lib/lshrdi3.c bm = 32 - b; b 23 arch/sh/lib/lshrdi3.c w.s.high = (unsigned int) uu.s.high >> b; b 24 arch/sh/lib/lshrdi3.c w.s.low = ((unsigned int) uu.s.low >> b) | carries; b 130 arch/sh/mm/pmb.c static inline bool pmb_can_merge(struct pmb_entry *a, struct pmb_entry *b) b 132 arch/sh/mm/pmb.c return (b->vpn == (a->vpn + a->size)) && b 133 arch/sh/mm/pmb.c (b->ppn == (a->ppn + a->size)) && b 134 arch/sh/mm/pmb.c (b->flags == a->flags); b 18 arch/sparc/crypto/opcodes.h #define CRC32C(a,b,c) \ b 19 arch/sparc/crypto/opcodes.h .word (F3F(2,0x36,0x147)|RS1(a)|RS2(b)|RD(c)); b 30 arch/sparc/crypto/opcodes.h #define AES_EROUND01(a,b,c,d) \ b 31 arch/sparc/crypto/opcodes.h .word (F3F(2, 0x19, 0)|RS1(a)|RS2(b)|RS3(c)|RD(d)); b 32 arch/sparc/crypto/opcodes.h #define AES_EROUND23(a,b,c,d) \ b 33 arch/sparc/crypto/opcodes.h .word (F3F(2, 0x19, 1)|RS1(a)|RS2(b)|RS3(c)|RD(d)); b 34 arch/sparc/crypto/opcodes.h #define AES_DROUND01(a,b,c,d) \ b 35 arch/sparc/crypto/opcodes.h .word (F3F(2, 0x19, 2)|RS1(a)|RS2(b)|RS3(c)|RD(d)); b 36 arch/sparc/crypto/opcodes.h #define AES_DROUND23(a,b,c,d) \ b 37 arch/sparc/crypto/opcodes.h .word (F3F(2, 0x19, 3)|RS1(a)|RS2(b)|RS3(c)|RD(d)); b 38 arch/sparc/crypto/opcodes.h #define AES_EROUND01_L(a,b,c,d) \ b 39 arch/sparc/crypto/opcodes.h .word (F3F(2, 0x19, 4)|RS1(a)|RS2(b)|RS3(c)|RD(d)); b 40 arch/sparc/crypto/opcodes.h #define AES_EROUND23_L(a,b,c,d) \ b 41 arch/sparc/crypto/opcodes.h .word (F3F(2, 0x19, 5)|RS1(a)|RS2(b)|RS3(c)|RD(d)); b 42 arch/sparc/crypto/opcodes.h #define AES_DROUND01_L(a,b,c,d) \ b 43 arch/sparc/crypto/opcodes.h .word (F3F(2, 0x19, 6)|RS1(a)|RS2(b)|RS3(c)|RD(d)); b 44 arch/sparc/crypto/opcodes.h #define AES_DROUND23_L(a,b,c,d) \ b 45 arch/sparc/crypto/opcodes.h .word (F3F(2, 0x19, 7)|RS1(a)|RS2(b)|RS3(c)|RD(d)); b 46 arch/sparc/crypto/opcodes.h #define AES_KEXPAND1(a,b,c,d) \ b 47 arch/sparc/crypto/opcodes.h .word (F3F(2, 0x19, 8)|RS1(a)|RS2(b)|IMM5_9(c)|RD(d)); b 48 arch/sparc/crypto/opcodes.h #define AES_KEXPAND0(a,b,c) \ b 49 arch/sparc/crypto/opcodes.h .word (F3F(2, 0x36, 0x130)|RS1(a)|RS2(b)|RD(c)); b 50 arch/sparc/crypto/opcodes.h #define AES_KEXPAND2(a,b,c) \ b 51 arch/sparc/crypto/opcodes.h .word (F3F(2, 0x36, 0x131)|RS1(a)|RS2(b)|RD(c)); b 53 arch/sparc/crypto/opcodes.h #define DES_IP(a,b) \ b 54 arch/sparc/crypto/opcodes.h .word (F3F(2, 0x36, 0x134)|RS1(a)|RD(b)); b 55 arch/sparc/crypto/opcodes.h #define DES_IIP(a,b) \ b 56 arch/sparc/crypto/opcodes.h .word (F3F(2, 0x36, 0x135)|RS1(a)|RD(b)); b 57 arch/sparc/crypto/opcodes.h #define DES_KEXPAND(a,b,c) \ b 58 arch/sparc/crypto/opcodes.h .word (F3F(2, 0x36, 0x136)|RS1(a)|IMM5_0(b)|RD(c)); b 59 arch/sparc/crypto/opcodes.h #define DES_ROUND(a,b,c,d) \ b 60 arch/sparc/crypto/opcodes.h .word (F3F(2, 0x19, 0x009)|RS1(a)|RS2(b)|RS3(c)|RD(d)); b 62 arch/sparc/crypto/opcodes.h #define CAMELLIA_F(a,b,c,d) \ b 63 arch/sparc/crypto/opcodes.h .word (F3F(2, 0x19, 0x00c)|RS1(a)|RS2(b)|RS3(c)|RD(d)); b 64 arch/sparc/crypto/opcodes.h #define CAMELLIA_FL(a,b,c) \ b 65 arch/sparc/crypto/opcodes.h .word (F3F(2, 0x36, 0x13c)|RS1(a)|RS2(b)|RD(c)); b 66 arch/sparc/crypto/opcodes.h #define CAMELLIA_FLI(a,b,c) \ b 67 arch/sparc/crypto/opcodes.h .word (F3F(2, 0x36, 0x13d)|RS1(a)|RS2(b)|RD(c)); b 25 arch/sparc/include/asm/asm-prototypes.h TItype __multi3(TItype a, TItype b); b 22 arch/sparc/include/asm/asmmacro.h #define RESTORE_ALL b ret_trap_entry; clr %l6; b 13 arch/sparc/include/asm/head_32.h rd %psr, %l0; b label; rd %wim, %l3; nop; b 16 arch/sparc/include/asm/head_32.h #define SRMMU_TFAULT rd %psr, %l0; rd %wim, %l3; b srmmu_fault; mov 1, %l7; b 17 arch/sparc/include/asm/head_32.h #define SRMMU_DFAULT rd %psr, %l0; rd %wim, %l3; b srmmu_fault; mov 0, %l7; b 21 arch/sparc/include/asm/head_32.h rd %psr, %l0; mov num, %l7; b bad_trap_handler; rd %wim, %l3; b 37 arch/sparc/include/asm/head_32.h b linux_sparc_syscall; \ b 41 arch/sparc/include/asm/head_32.h b breakpoint_trap; \ b 49 arch/sparc/include/asm/head_32.h b kgdb_trap_low; \ b 59 arch/sparc/include/asm/head_32.h b getcc_trap_handler; rd %psr, %l0; nop; nop; b 63 arch/sparc/include/asm/head_32.h b setcc_trap_handler; rd %psr, %l0; nop; nop; b 73 arch/sparc/include/asm/head_32.h mov int_level, %l7; rd %psr, %l0; b real_irq_entry; rd %wim, %l3; b 79 arch/sparc/include/asm/head_32.h rd %psr, %l0; rd %wim, %l3; b spill_window_entry; andcc %l0, PSR_PS, %g0; b 82 arch/sparc/include/asm/head_32.h rd %psr, %l0; rd %wim, %l3; b fill_window_entry; andcc %l0, PSR_PS, %g0; b 1907 arch/sparc/include/asm/hypervisor.h #define HV_PCI_DEVICE_BUILD(b,d,f) \ b 1908 arch/sparc/include/asm/hypervisor.h ((((b) & 0xff) << 16) | \ b 73 arch/sparc/include/asm/io_32.h static inline void sbus_writeb(u8 b, volatile void __iomem *addr) b 75 arch/sparc/include/asm/io_32.h *(__force volatile u8 *)addr = b; b 69 arch/sparc/include/asm/io_64.h static inline void __raw_writeb(u8 b, const volatile void __iomem *addr) b 73 arch/sparc/include/asm/io_64.h : "Jr" (b), "r" (addr), "i" (ASI_PHYS_BYPASS_EC_E)); b 157 arch/sparc/include/asm/io_64.h static inline void writeb(u8 b, volatile void __iomem *addr) b 161 arch/sparc/include/asm/io_64.h : "Jr" (b), "r" (addr), "i" (ASI_PHYS_BYPASS_EC_E_L) b 214 arch/sparc/include/asm/io_64.h static inline void outb(u8 b, unsigned long addr) b 216 arch/sparc/include/asm/io_64.h writeb(b, (volatile void __iomem *)addr); b 310 arch/sparc/include/asm/io_64.h static inline void sbus_writeb(u8 b, volatile void __iomem *addr) b 312 arch/sparc/include/asm/io_64.h __raw_writeb(b, addr); b 34 arch/sparc/include/asm/topology_64.h #define node_distance(a, b) __node_distance(a, b) b 12 arch/sparc/include/asm/ttable.h #define BOOT_KERNEL b sparc64_boot; nop; nop; nop; nop; nop; nop; nop; b 325 arch/sparc/include/asm/ttable.h b,a,pt %xcc, spill_fixup_dax; \ b 326 arch/sparc/include/asm/ttable.h b,a,pt %xcc, spill_fixup_mna; \ b 327 arch/sparc/include/asm/ttable.h b,a,pt %xcc, spill_fixup; b 421 arch/sparc/include/asm/ttable.h b,a,pt %xcc, spill_fixup_dax; \ b 422 arch/sparc/include/asm/ttable.h b,a,pt %xcc, spill_fixup_mna; \ b 423 arch/sparc/include/asm/ttable.h b,a,pt %xcc, spill_fixup; b 583 arch/sparc/include/asm/ttable.h b,a,pt %xcc, fill_fixup_dax; \ b 584 arch/sparc/include/asm/ttable.h b,a,pt %xcc, fill_fixup_mna; \ b 585 arch/sparc/include/asm/ttable.h b,a,pt %xcc, fill_fixup; b 643 arch/sparc/include/asm/ttable.h b,a,pt %xcc, fill_fixup_dax; \ b 644 arch/sparc/include/asm/ttable.h b,a,pt %xcc, fill_fixup_mna; \ b 645 arch/sparc/include/asm/ttable.h b,a,pt %xcc, fill_fixup; b 31 arch/sparc/include/asm/uaccess_32.h #define segment_eq(a, b) ((a).seg == (b).seg) b 112 arch/sparc/include/asm/uaccess_32.h __put_user_asm(x, b, addr, __pu_ret); \ b 136 arch/sparc/include/asm/uaccess_32.h case 1: __put_user_asm(x, b, addr, __pu_ret); break; \ b 35 arch/sparc/include/asm/uaccess_64.h #define segment_eq(a, b) ((a).seg == (b).seg) b 107 arch/sparc/include/asm/uaccess_64.h case 1: __put_user_asm(data, b, addr, __pu_ret); break; \ b 72 arch/sparc/include/asm/upa.h static inline void _upa_writeb(unsigned char b, unsigned long addr) b 76 arch/sparc/include/asm/upa.h : "r" (b), "r" (addr), "i" (ASI_PHYS_BYPASS_EC_E)); b 422 arch/sparc/kernel/mdesc.c struct mdesc_handle *b, b 460 arch/sparc/kernel/mdesc.c mdesc_for_each_node_by_name(b, b_node, name) { b 461 arch/sparc/kernel/mdesc.c rv = get_info_func(b, b_node, &b_node_info); b 100 arch/sparc/kernel/of_device_common.c u64 b = of_read_addr(base, na); b 102 arch/sparc/kernel/of_device_common.c if (a < b) b 105 arch/sparc/kernel/of_device_common.c b += of_read_addr(size, ns); b 106 arch/sparc/kernel/of_device_common.c if (a >= b) b 719 arch/sparc/kernel/visemul.c s16 b = (rs2 >> (i * 16)) & 0xffff; b 721 arch/sparc/kernel/visemul.c if (a > b) b 729 arch/sparc/kernel/visemul.c s32 b = (rs2 >> (i * 32)) & 0xffffffff; b 731 arch/sparc/kernel/visemul.c if (a > b) b 739 arch/sparc/kernel/visemul.c s16 b = (rs2 >> (i * 16)) & 0xffff; b 741 arch/sparc/kernel/visemul.c if (a <= b) b 749 arch/sparc/kernel/visemul.c s32 b = (rs2 >> (i * 32)) & 0xffffffff; b 751 arch/sparc/kernel/visemul.c if (a <= b) b 759 arch/sparc/kernel/visemul.c s16 b = (rs2 >> (i * 16)) & 0xffff; b 761 arch/sparc/kernel/visemul.c if (a != b) b 769 arch/sparc/kernel/visemul.c s32 b = (rs2 >> (i * 32)) & 0xffffffff; b 771 arch/sparc/kernel/visemul.c if (a != b) b 779 arch/sparc/kernel/visemul.c s16 b = (rs2 >> (i * 16)) & 0xffff; b 781 arch/sparc/kernel/visemul.c if (a == b) b 789 arch/sparc/kernel/visemul.c s32 b = (rs2 >> (i * 32)) & 0xffffffff; b 791 arch/sparc/kernel/visemul.c if (a == b) b 6 arch/sparc/lib/cmpdi2.c word_type __cmpdi2(long long a, long long b) b 12 arch/sparc/lib/cmpdi2.c .ll = b b 5 arch/sparc/lib/ucmpdi2.c word_type __ucmpdi2(unsigned long long a, unsigned long long b) b 8 arch/sparc/lib/ucmpdi2.c const DWunion bu = {.ll = b}; b 282 arch/sparc/math-emu/math_32.c #define TYPE(dummy, r, ru, b, bu, a, au) type = (au << 2) | (a << 0) | (bu << 5) | (b << 3) | (ru << 8) | (r << 6) b 176 arch/sparc/math-emu/math_64.c #define TYPE(ftt, r, ru, b, bu, a, au) type = (au << 2) | (a << 0) | (bu << 5) | (b << 3) | (ru << 8) | (r << 6) | (ftt << 9) b 98 arch/sparc/mm/init_64.c static int cmp_p64(const void *a, const void *b) b 100 arch/sparc/mm/init_64.c const struct linux_prom64_registers *x = a, *y = b; b 600 arch/sparc/mm/init_64.c static int cmp_ptrans(const void *a, const void *b) b 602 arch/sparc/mm/init_64.c const struct linux_prom_translation *x = a, *y = b; b 293 arch/sparc/net/bpf_jit_comp_64.c static void emit_alu3(u32 opcode, u32 a, u32 b, u32 c, struct jit_ctx *ctx) b 295 arch/sparc/net/bpf_jit_comp_64.c emit(opcode | RS1(a) | RS2(b) | RD(c), ctx); b 50 arch/sparc/prom/memory.c static int sp_banks_cmp(const void *a, const void *b) b 52 arch/sparc/prom/memory.c const struct sparc_phys_banks *x = a, *y = b; b 74 arch/um/drivers/mconsole_user.c #define MIN(a,b) ((a)<(b) ? (a):(b)) b 102 arch/unicore32/include/asm/assembler.h \instr\()b.u \reg, [\ptr], #\inc b 118 arch/unicore32/mm/init.c const struct membank *a = _a, *b = _b; b 119 arch/unicore32/mm/init.c long cmp = bank_pfn_start(a) - bank_pfn_start(b); b 56 arch/x86/boot/cpucheck.c #define A32(a, b, c, d) (((d) << 24)+((c) << 16)+((b) << 8)+(a)) b 75 arch/x86/boot/cpuflags.c u32 *a, u32 *b, u32 *c, u32 *d) b 80 arch/x86/boot/cpuflags.c : "=a" (*a), "=c" (*c), "=d" (*d), EBX_REG (*b) b 85 arch/x86/boot/cpuflags.c #define cpuid(id, a, b, c, d) cpuid_count(id, 0, a, b, c, d) b 58 arch/x86/crypto/ghash-clmulni-intel_glue.c u64 a, b; b 67 arch/x86/crypto/ghash-clmulni-intel_glue.c b = be64_to_cpu(x->b); b 69 arch/x86/crypto/ghash-clmulni-intel_glue.c ctx->shash.a = (b << 1) | (a >> 63); b 70 arch/x86/crypto/ghash-clmulni-intel_glue.c ctx->shash.b = (a << 1) | (b >> 63); b 73 arch/x86/crypto/ghash-clmulni-intel_glue.c ctx->shash.b ^= ((u64)0xc2) << 56; b 314 arch/x86/crypto/glue_helper.c le128 b[2]; b 321 arch/x86/crypto/glue_helper.c next_tweak = memcpy(b, req->iv, XTS_BLOCK_SIZE); b 322 arch/x86/crypto/glue_helper.c gf128mul_x_ble(b, b); b 336 arch/x86/crypto/glue_helper.c scatterwalk_map_and_copy(b, dst, 0, XTS_BLOCK_SIZE, 0); b 337 arch/x86/crypto/glue_helper.c memcpy(b + 1, b, tail - XTS_BLOCK_SIZE); b 338 arch/x86/crypto/glue_helper.c scatterwalk_map_and_copy(b, src, XTS_BLOCK_SIZE, b 340 arch/x86/crypto/glue_helper.c scatterwalk_map_and_copy(b, dst, 0, tail, 1); b 52 arch/x86/crypto/poly1305_glue.c static void poly1305_simd_mult(u32 *a, const u32 *b) b 60 arch/x86/crypto/poly1305_glue.c poly1305_block_sse2(a, m, b, 1); b 460 arch/x86/events/core.c u64 b = 0; b 469 arch/x86/events/core.c b |= PERF_SAMPLE_BRANCH_USER; b 472 arch/x86/events/core.c b |= PERF_SAMPLE_BRANCH_KERNEL; b 478 arch/x86/events/core.c return m == b; b 54 arch/x86/events/intel/ds.c #define P(a, b) PERF_MEM_S(a, b) b 82 arch/x86/include/asm/alternative-asm.h #define alt_max_short(a, b) ((a) ^ (((a) ^ (b)) & -(-((a) < (b))))) b 121 arch/x86/include/asm/alternative.h #define alt_max_short(a, b) "((" a ") ^ (((" a ") ^ (" b ")) & -(-((" a ") < (" b ")))))" b 17 arch/x86/include/asm/asm.h # define __ASM_SEL(a,b) __ASM_FORM(a) b 18 arch/x86/include/asm/asm.h # define __ASM_SEL_RAW(a,b) __ASM_FORM_RAW(a) b 21 arch/x86/include/asm/asm.h # define __ASM_SEL(a,b) __ASM_FORM(b) b 22 arch/x86/include/asm/asm.h # define __ASM_SEL_RAW(a,b) __ASM_FORM_RAW(b) b 183 arch/x86/include/asm/checksum_64.h static inline unsigned add32_with_carry(unsigned a, unsigned b) b 188 arch/x86/include/asm/checksum_64.h : "0" (a), "rm" (b)); b 80 arch/x86/include/asm/crypto/glue_helper.h dst->b = cpu_to_be64(le64_to_cpu(src->b)); b 86 arch/x86/include/asm/crypto/glue_helper.h dst->b = cpu_to_le64(be64_to_cpu(src->b)); b 92 arch/x86/include/asm/crypto/glue_helper.h u64 b = le64_to_cpu(i->b); b 94 arch/x86/include/asm/crypto/glue_helper.h b++; b 95 arch/x86/include/asm/crypto/glue_helper.h if (!b) b 99 arch/x86/include/asm/crypto/glue_helper.h i->b = cpu_to_le64(b); b 63 arch/x86/include/asm/div64.h static inline u64 mul_u32_u32(u32 a, u32 b) b 68 arch/x86/include/asm/div64.h : [a] "a" (a), [b] "rm" (b) ); b 64 arch/x86/include/asm/dwarf2.h .macro cfi_ignore a=0, b=0, c=0, d=0 b 24 arch/x86/include/asm/extable.h #define swap_ex_entry_fixup(a, b, tmp, delta) \ b 26 arch/x86/include/asm/extable.h (a)->fixup = (b)->fixup + (delta); \ b 27 arch/x86/include/asm/extable.h (b)->fixup = (tmp).fixup - (delta); \ b 28 arch/x86/include/asm/extable.h (a)->handler = (b)->handler + (delta); \ b 29 arch/x86/include/asm/extable.h (b)->handler = (tmp).handler - (delta); \ b 339 arch/x86/include/asm/io.h BUILDIO(b, b, char) b 316 arch/x86/include/asm/kvm_emulate.h u8 b; b 229 arch/x86/include/asm/mce.h bool machine_check_poll(enum mcp_flags flags, mce_banks_t *b); b 83 arch/x86/include/asm/microcode.h #define QCHAR(a, b, c, d) ((a) + ((b) << 8) + ((c) << 16) + ((d) << 24)) b 91 arch/x86/include/asm/microcode.h #define CPUID_IS(a, b, c, ebx, ecx, edx) \ b 92 arch/x86/include/asm/microcode.h (!((ebx ^ (a))|(edx ^ (b))|(ecx ^ (c)))) b 203 arch/x86/include/asm/paravirt.h #define wrmsr_safe(msr, a, b) paravirt_write_msr_safe(msr, a, b) b 206 arch/x86/include/asm/paravirt.h #define rdmsr_safe(msr, a, b) \ b 211 arch/x86/include/asm/paravirt.h (*b) = _l >> 32; \ b 366 arch/x86/include/asm/paravirt_types.h #define NATIVE_LABEL(a,x,b) "\n\t.globl " a #x "_" #b "\n" a #x "_" #b ":\n\t" b 729 arch/x86/include/asm/pgtable.h static inline int pte_same(pte_t a, pte_t b) b 731 arch/x86/include/asm/pgtable.h return a.pte == b.pte; b 9 arch/x86/include/asm/rmwcc.h #define __RMWcc_CONCAT(a, b) a ## b b 10 arch/x86/include/asm/rmwcc.h #define RMWcc_CONCAT(a, b) __RMWcc_CONCAT(a, b) b 81 arch/x86/include/asm/topology.h #define node_distance(a, b) __node_distance(a, b) b 36 arch/x86/include/asm/uaccess.h #define segment_eq(a, b) ((a).seg == (b).seg) b 719 arch/x86/include/asm/uaccess.h #define user_access_begin(a,b) user_access_begin(a,b) b 78 arch/x86/include/asm/vm86.h #define handle_vm86_fault(a, b) b 81 arch/x86/include/asm/vm86.h static inline int handle_vm86_trap(struct kernel_vm86_regs *a, long b, int c) b 86 arch/x86/include/asm/vm86.h static inline void save_v86_state(struct kernel_vm86_regs *a, int b) { } b 21 arch/x86/include/uapi/asm/swab.h __u32 b; b 27 arch/x86/include/uapi/asm/swab.h : "=r" (v.s.a), "=r" (v.s.b) b 28 arch/x86/include/uapi/asm/swab.h : "0" (v.s.a), "1" (v.s.b)); b 121 arch/x86/kernel/cpu/mce/amd.c struct smca_bank *b; b 126 arch/x86/kernel/cpu/mce/amd.c b = &smca_banks[bank]; b 127 arch/x86/kernel/cpu/mce/amd.c if (!b->hwid) b 130 arch/x86/kernel/cpu/mce/amd.c return b->hwid->bank_type; b 292 arch/x86/kernel/cpu/mce/amd.c struct threshold_block *b; b 312 arch/x86/kernel/cpu/mce/amd.c static const char *bank4_names(const struct threshold_block *b) b 314 arch/x86/kernel/cpu/mce/amd.c switch (b->address) { b 326 arch/x86/kernel/cpu/mce/amd.c WARN(1, "Funny MSR: 0x%08x\n", b->address); b 347 arch/x86/kernel/cpu/mce/amd.c static int lvt_off_valid(struct threshold_block *b, int apic, u32 lo, u32 hi) b 353 arch/x86/kernel/cpu/mce/amd.c "for bank %d, block %d (MSR%08X=0x%x%08x)\n", b->cpu, b 354 arch/x86/kernel/cpu/mce/amd.c b->bank, b->block, b->address, hi, lo); b 369 arch/x86/kernel/cpu/mce/amd.c b->cpu, apic, b->bank, b->block, b->address, hi, lo); b 382 arch/x86/kernel/cpu/mce/amd.c rdmsr(tr->b->address, lo, hi); b 384 arch/x86/kernel/cpu/mce/amd.c if (tr->b->threshold_limit < (hi & THRESHOLD_MAX)) b 390 arch/x86/kernel/cpu/mce/amd.c (THRESHOLD_MAX - tr->b->threshold_limit); b 393 arch/x86/kernel/cpu/mce/amd.c (tr->old_limit - tr->b->threshold_limit); b 402 arch/x86/kernel/cpu/mce/amd.c if (!tr->b->interrupt_capable) b 406 arch/x86/kernel/cpu/mce/amd.c if (lvt_off_valid(tr->b, tr->lvt_off, lo, hi)) { b 413 arch/x86/kernel/cpu/mce/amd.c if (tr->b->interrupt_enable) b 419 arch/x86/kernel/cpu/mce/amd.c wrmsr(tr->b->address, lo, hi); b 422 arch/x86/kernel/cpu/mce/amd.c static void mce_threshold_block_init(struct threshold_block *b, int offset) b 425 arch/x86/kernel/cpu/mce/amd.c .b = b, b 430 arch/x86/kernel/cpu/mce/amd.c b->threshold_limit = THRESHOLD_MAX; b 524 arch/x86/kernel/cpu/mce/amd.c struct threshold_block b; b 530 arch/x86/kernel/cpu/mce/amd.c memset(&b, 0, sizeof(b)); b 531 arch/x86/kernel/cpu/mce/amd.c b.cpu = cpu; b 532 arch/x86/kernel/cpu/mce/amd.c b.bank = bank; b 533 arch/x86/kernel/cpu/mce/amd.c b.block = block; b 534 arch/x86/kernel/cpu/mce/amd.c b.address = addr; b 535 arch/x86/kernel/cpu/mce/amd.c b.interrupt_capable = lvt_interrupt_supported(bank, misc_high); b 537 arch/x86/kernel/cpu/mce/amd.c if (!b.interrupt_capable) b 540 arch/x86/kernel/cpu/mce/amd.c b.interrupt_enable = 1; b 559 arch/x86/kernel/cpu/mce/amd.c mce_threshold_block_init(&b, offset); b 1006 arch/x86/kernel/cpu/mce/amd.c tr.b = block; b 1048 arch/x86/kernel/cpu/mce/amd.c static ssize_t show_ ## name(struct threshold_block *b, char *buf) \ b 1050 arch/x86/kernel/cpu/mce/amd.c return sprintf(buf, "%lu\n", (unsigned long) b->name); \ b 1056 arch/x86/kernel/cpu/mce/amd.c store_interrupt_enable(struct threshold_block *b, const char *buf, size_t size) b 1061 arch/x86/kernel/cpu/mce/amd.c if (!b->interrupt_capable) b 1067 arch/x86/kernel/cpu/mce/amd.c b->interrupt_enable = !!new; b 1070 arch/x86/kernel/cpu/mce/amd.c tr.b = b; b 1072 arch/x86/kernel/cpu/mce/amd.c smp_call_function_single(b->cpu, threshold_restart_bank, &tr, 1); b 1078 arch/x86/kernel/cpu/mce/amd.c store_threshold_limit(struct threshold_block *b, const char *buf, size_t size) b 1092 arch/x86/kernel/cpu/mce/amd.c tr.old_limit = b->threshold_limit; b 1093 arch/x86/kernel/cpu/mce/amd.c b->threshold_limit = new; b 1094 arch/x86/kernel/cpu/mce/amd.c tr.b = b; b 1096 arch/x86/kernel/cpu/mce/amd.c smp_call_function_single(b->cpu, threshold_restart_bank, &tr, 1); b 1101 arch/x86/kernel/cpu/mce/amd.c static ssize_t show_error_count(struct threshold_block *b, char *buf) b 1105 arch/x86/kernel/cpu/mce/amd.c rdmsr_on_cpu(b->cpu, b->address, &lo, &hi); b 1108 arch/x86/kernel/cpu/mce/amd.c (THRESHOLD_MAX - b->threshold_limit))); b 1138 arch/x86/kernel/cpu/mce/amd.c struct threshold_block *b = to_block(kobj); b 1142 arch/x86/kernel/cpu/mce/amd.c ret = a->show ? a->show(b, buf) : -EIO; b 1150 arch/x86/kernel/cpu/mce/amd.c struct threshold_block *b = to_block(kobj); b 1154 arch/x86/kernel/cpu/mce/amd.c ret = a->store ? a->store(b, buf, count) : -EIO; b 1172 arch/x86/kernel/cpu/mce/amd.c static const char *get_name(unsigned int bank, struct threshold_block *b) b 1177 arch/x86/kernel/cpu/mce/amd.c if (b && bank == 4) b 1178 arch/x86/kernel/cpu/mce/amd.c return bank4_names(b); b 1187 arch/x86/kernel/cpu/mce/amd.c if (b && bank_type == SMCA_UMC) { b 1188 arch/x86/kernel/cpu/mce/amd.c if (b->block < ARRAY_SIZE(smca_umc_block_names)) b 1189 arch/x86/kernel/cpu/mce/amd.c return smca_umc_block_names[b->block]; b 1206 arch/x86/kernel/cpu/mce/amd.c struct threshold_block *b = NULL; b 1227 arch/x86/kernel/cpu/mce/amd.c b = kzalloc(sizeof(struct threshold_block), GFP_KERNEL); b 1228 arch/x86/kernel/cpu/mce/amd.c if (!b) b 1231 arch/x86/kernel/cpu/mce/amd.c b->block = block; b 1232 arch/x86/kernel/cpu/mce/amd.c b->bank = bank; b 1233 arch/x86/kernel/cpu/mce/amd.c b->cpu = cpu; b 1234 arch/x86/kernel/cpu/mce/amd.c b->address = address; b 1235 arch/x86/kernel/cpu/mce/amd.c b->interrupt_enable = 0; b 1236 arch/x86/kernel/cpu/mce/amd.c b->interrupt_capable = lvt_interrupt_supported(bank, high); b 1237 arch/x86/kernel/cpu/mce/amd.c b->threshold_limit = THRESHOLD_MAX; b 1239 arch/x86/kernel/cpu/mce/amd.c if (b->interrupt_capable) { b 1241 arch/x86/kernel/cpu/mce/amd.c b->interrupt_enable = 1; b 1246 arch/x86/kernel/cpu/mce/amd.c INIT_LIST_HEAD(&b->miscj); b 1249 arch/x86/kernel/cpu/mce/amd.c list_add(&b->miscj, &tb->blocks->miscj); b 1251 arch/x86/kernel/cpu/mce/amd.c tb->blocks = b; b 1253 arch/x86/kernel/cpu/mce/amd.c err = kobject_init_and_add(&b->kobj, &threshold_ktype, tb->kobj, get_name(bank, b)); b 1265 arch/x86/kernel/cpu/mce/amd.c if (b) b 1266 arch/x86/kernel/cpu/mce/amd.c kobject_uevent(&b->kobj, KOBJ_ADD); b 1271 arch/x86/kernel/cpu/mce/amd.c if (b) { b 1272 arch/x86/kernel/cpu/mce/amd.c kobject_put(&b->kobj); b 1273 arch/x86/kernel/cpu/mce/amd.c list_del(&b->miscj); b 1274 arch/x86/kernel/cpu/mce/amd.c kfree(b); b 1279 arch/x86/kernel/cpu/mce/amd.c static int __threshold_add_blocks(struct threshold_bank *b) b 1281 arch/x86/kernel/cpu/mce/amd.c struct list_head *head = &b->blocks->miscj; b 1286 arch/x86/kernel/cpu/mce/amd.c err = kobject_add(&b->blocks->kobj, b->kobj, b->blocks->kobj.name); b 1292 arch/x86/kernel/cpu/mce/amd.c err = kobject_add(&pos->kobj, b->kobj, pos->kobj.name); b 1307 arch/x86/kernel/cpu/mce/amd.c struct threshold_bank *b = NULL; b 1320 arch/x86/kernel/cpu/mce/amd.c b = nb->bank4; b 1321 arch/x86/kernel/cpu/mce/amd.c err = kobject_add(b->kobj, &dev->kobj, name); b 1325 arch/x86/kernel/cpu/mce/amd.c per_cpu(threshold_banks, cpu)[bank] = b; b 1326 arch/x86/kernel/cpu/mce/amd.c refcount_inc(&b->cpus); b 1328 arch/x86/kernel/cpu/mce/amd.c err = __threshold_add_blocks(b); b 1334 arch/x86/kernel/cpu/mce/amd.c b = kzalloc(sizeof(struct threshold_bank), GFP_KERNEL); b 1335 arch/x86/kernel/cpu/mce/amd.c if (!b) { b 1340 arch/x86/kernel/cpu/mce/amd.c b->kobj = kobject_create_and_add(name, &dev->kobj); b 1341 arch/x86/kernel/cpu/mce/amd.c if (!b->kobj) { b 1347 arch/x86/kernel/cpu/mce/amd.c refcount_set(&b->cpus, 1); b 1352 arch/x86/kernel/cpu/mce/amd.c nb->bank4 = b; b 1356 arch/x86/kernel/cpu/mce/amd.c err = allocate_threshold_blocks(cpu, b, bank, 0, msr_ops.misc(bank)); b 1360 arch/x86/kernel/cpu/mce/amd.c per_cpu(threshold_banks, cpu)[bank] = b; b 1365 arch/x86/kernel/cpu/mce/amd.c kfree(b); b 1393 arch/x86/kernel/cpu/mce/amd.c static void __threshold_remove_blocks(struct threshold_bank *b) b 1398 arch/x86/kernel/cpu/mce/amd.c kobject_del(b->kobj); b 1400 arch/x86/kernel/cpu/mce/amd.c list_for_each_entry_safe(pos, tmp, &b->blocks->miscj, miscj) b 1407 arch/x86/kernel/cpu/mce/amd.c struct threshold_bank *b; b 1409 arch/x86/kernel/cpu/mce/amd.c b = per_cpu(threshold_banks, cpu)[bank]; b 1410 arch/x86/kernel/cpu/mce/amd.c if (!b) b 1413 arch/x86/kernel/cpu/mce/amd.c if (!b->blocks) b 1417 arch/x86/kernel/cpu/mce/amd.c if (!refcount_dec_and_test(&b->cpus)) { b 1418 arch/x86/kernel/cpu/mce/amd.c __threshold_remove_blocks(b); b 1434 arch/x86/kernel/cpu/mce/amd.c kobject_del(b->kobj); b 1435 arch/x86/kernel/cpu/mce/amd.c kobject_put(b->kobj); b 1436 arch/x86/kernel/cpu/mce/amd.c kfree(b); b 699 arch/x86/kernel/cpu/mce/core.c bool machine_check_poll(enum mcp_flags flags, mce_banks_t *b) b 714 arch/x86/kernel/cpu/mce/core.c if (!mce_banks[i].ctl || !test_bit(i, *b)) b 1498 arch/x86/kernel/cpu/mce/core.c struct mce_bank *b = &mce_banks[i]; b 1505 arch/x86/kernel/cpu/mce/core.c b->ctl = -1ULL; b 1506 arch/x86/kernel/cpu/mce/core.c b->init = 1; b 1516 arch/x86/kernel/cpu/mce/core.c u8 b; b 1520 arch/x86/kernel/cpu/mce/core.c b = cap & MCG_BANKCNT_MASK; b 1522 arch/x86/kernel/cpu/mce/core.c if (b > MAX_NR_BANKS) { b 1524 arch/x86/kernel/cpu/mce/core.c smp_processor_id(), MAX_NR_BANKS, b); b 1525 arch/x86/kernel/cpu/mce/core.c b = MAX_NR_BANKS; b 1528 arch/x86/kernel/cpu/mce/core.c this_cpu_write(mce_num_banks, b); b 1568 arch/x86/kernel/cpu/mce/core.c struct mce_bank *b = &mce_banks[i]; b 1570 arch/x86/kernel/cpu/mce/core.c if (!b->init) b 1572 arch/x86/kernel/cpu/mce/core.c wrmsrl(msr_ops.ctl(i), b->ctl); b 1594 arch/x86/kernel/cpu/mce/core.c struct mce_bank *b = &mce_banks[i]; b 1596 arch/x86/kernel/cpu/mce/core.c if (!b->init) b 1600 arch/x86/kernel/cpu/mce/core.c b->init = !!msrval; b 2013 arch/x86/kernel/cpu/mce/core.c struct mce_bank *b = &mce_banks[i]; b 2015 arch/x86/kernel/cpu/mce/core.c if (b->init) b 2121 arch/x86/kernel/cpu/mce/core.c struct mce_bank *b; b 2126 arch/x86/kernel/cpu/mce/core.c b = &per_cpu(mce_banks_array, s->id)[bank]; b 2128 arch/x86/kernel/cpu/mce/core.c if (!b->init) b 2131 arch/x86/kernel/cpu/mce/core.c return sprintf(buf, "%llx\n", b->ctl); b 2138 arch/x86/kernel/cpu/mce/core.c struct mce_bank *b; b 2147 arch/x86/kernel/cpu/mce/core.c b = &per_cpu(mce_banks_array, s->id)[bank]; b 2149 arch/x86/kernel/cpu/mce/core.c if (!b->init) b 2152 arch/x86/kernel/cpu/mce/core.c b->ctl = new; b 2362 arch/x86/kernel/cpu/mce/core.c struct mce_bank *b = &mce_banks[i]; b 2364 arch/x86/kernel/cpu/mce/core.c if (b->init) b 2365 arch/x86/kernel/cpu/mce/core.c wrmsrl(msr_ops.ctl(i), b->ctl); b 2412 arch/x86/kernel/cpu/mce/core.c struct mce_bank_dev *b = &mce_bank_devs[i]; b 2413 arch/x86/kernel/cpu/mce/core.c struct device_attribute *a = &b->attr; b 2415 arch/x86/kernel/cpu/mce/core.c b->bank = i; b 2418 arch/x86/kernel/cpu/mce/core.c a->attr.name = b->attrname; b 2419 arch/x86/kernel/cpu/mce/core.c snprintf(b->attrname, ATTR_LEN, "bank%d", i); b 129 arch/x86/kernel/cpu/mce/inject.c mce_banks_t b; b 131 arch/x86/kernel/cpu/mce/inject.c memset(&b, 0xff, sizeof(mce_banks_t)); b 133 arch/x86/kernel/cpu/mce/inject.c machine_check_poll(0, &b); b 460 arch/x86/kernel/cpu/mce/inject.c u8 b = m.bank; b 466 arch/x86/kernel/cpu/mce/inject.c wrmsrl(MSR_AMD64_SMCA_MCx_DESTAT(b), m.status); b 467 arch/x86/kernel/cpu/mce/inject.c wrmsrl(MSR_AMD64_SMCA_MCx_DEADDR(b), m.addr); b 469 arch/x86/kernel/cpu/mce/inject.c wrmsrl(MSR_AMD64_SMCA_MCx_STATUS(b), m.status); b 470 arch/x86/kernel/cpu/mce/inject.c wrmsrl(MSR_AMD64_SMCA_MCx_ADDR(b), m.addr); b 473 arch/x86/kernel/cpu/mce/inject.c wrmsrl(MSR_AMD64_SMCA_MCx_MISC(b), m.misc); b 474 arch/x86/kernel/cpu/mce/inject.c wrmsrl(MSR_AMD64_SMCA_MCx_SYND(b), m.synd); b 476 arch/x86/kernel/cpu/mce/inject.c wrmsrl(MSR_IA32_MCx_STATUS(b), m.status); b 477 arch/x86/kernel/cpu/mce/inject.c wrmsrl(MSR_IA32_MCx_ADDR(b), m.addr); b 478 arch/x86/kernel/cpu/mce/inject.c wrmsrl(MSR_IA32_MCx_MISC(b), m.misc); b 486 arch/x86/kernel/cpu/mce/inject.c u8 b = i_mce.bank; b 523 arch/x86/kernel/cpu/mce/inject.c b == 4 && b 524 arch/x86/kernel/cpu/mtrr/generic.c void mtrr_wrmsr(unsigned msr, unsigned a, unsigned b) b 526 arch/x86/kernel/cpu/mtrr/generic.c if (wrmsr_safe(msr, a, b) < 0) { b 528 arch/x86/kernel/cpu/mtrr/generic.c smp_processor_id(), msr, a, b); b 290 arch/x86/kernel/e820.c static int __init cpcompare(const void *a, const void *b) b 292 arch/x86/kernel/e820.c struct change_member * const *app = a, * const *bpp = b; b 119 arch/x86/kernel/early-quirks.c u8 b; b 121 arch/x86/kernel/early-quirks.c b = read_pci_config_byte(num, slot, func, 0xac); b 122 arch/x86/kernel/early-quirks.c b &= ~(1<<5); b 123 arch/x86/kernel/early-quirks.c write_pci_config_byte(num, slot, func, 0xac, b); b 137 arch/x86/kernel/early-quirks.c u8 b; b 147 arch/x86/kernel/early-quirks.c outb(0x72, 0xcd6); b = inb(0xcd7); b 148 arch/x86/kernel/early-quirks.c if (!(b & 0x2)) b 83 arch/x86/kernel/kvm.c static struct kvm_task_sleep_node *_find_apf_task(struct kvm_task_sleep_head *b, b 88 arch/x86/kernel/kvm.c hlist_for_each(p, &b->list) { b 105 arch/x86/kernel/kvm.c struct kvm_task_sleep_head *b = &async_pf_sleepers[key]; b 111 arch/x86/kernel/kvm.c raw_spin_lock(&b->lock); b 112 arch/x86/kernel/kvm.c e = _find_apf_task(b, token); b 117 arch/x86/kernel/kvm.c raw_spin_unlock(&b->lock); b 130 arch/x86/kernel/kvm.c hlist_add_head(&n.link, &b->list); b 131 arch/x86/kernel/kvm.c raw_spin_unlock(&b->lock); b 178 arch/x86/kernel/kvm.c struct kvm_task_sleep_head *b = &async_pf_sleepers[i]; b 179 arch/x86/kernel/kvm.c raw_spin_lock(&b->lock); b 180 arch/x86/kernel/kvm.c hlist_for_each_safe(p, next, &b->list) { b 186 arch/x86/kernel/kvm.c raw_spin_unlock(&b->lock); b 193 arch/x86/kernel/kvm.c struct kvm_task_sleep_head *b = &async_pf_sleepers[key]; b 202 arch/x86/kernel/kvm.c raw_spin_lock(&b->lock); b 203 arch/x86/kernel/kvm.c n = _find_apf_task(b, token); b 215 arch/x86/kernel/kvm.c raw_spin_unlock(&b->lock); b 222 arch/x86/kernel/kvm.c hlist_add_head(&n->link, &b->list); b 225 arch/x86/kernel/kvm.c raw_spin_unlock(&b->lock); b 52 arch/x86/kernel/mmconf-fam10h_64.c #define BASE_VALID(b) ((b) + MMCONF_SIZE <= (0xfdULL<<32) || (b) >= (1ULL<<40)) b 65 arch/x86/kernel/paravirt.c struct branch *b = insn_buff; b 74 arch/x86/kernel/paravirt.c b->opcode = 0xe8; /* call */ b 75 arch/x86/kernel/paravirt.c b->delta = delta; b 76 arch/x86/kernel/paravirt.c BUILD_BUG_ON(sizeof(*b) != call_len); b 91 arch/x86/kernel/paravirt.c struct branch *b = insn_buff; b 101 arch/x86/kernel/paravirt.c b->opcode = 0xe9; /* jmp */ b 102 arch/x86/kernel/paravirt.c b->delta = delta; b 360 arch/x86/kernel/quirks.c u8 b = 0; b 362 arch/x86/kernel/quirks.c err = pci_read_config_byte(dev, 0xac, &b); b 363 arch/x86/kernel/quirks.c b &= ~(1<<5); b 364 arch/x86/kernel/quirks.c err |= pci_write_config_byte(dev, 0xac, b); b 380 arch/x86/kernel/quirks.c u8 b; b 399 arch/x86/kernel/quirks.c outb(0x72, 0xcd6); b = inb(0xcd7); b 400 arch/x86/kernel/quirks.c b |= 0x1; b 401 arch/x86/kernel/quirks.c outb(0x72, 0xcd6); outb(b, 0xcd7); b 402 arch/x86/kernel/quirks.c outb(0x72, 0xcd6); b = inb(0xcd7); b 403 arch/x86/kernel/quirks.c if (!(b & 0x1)) b 191 arch/x86/kernel/unwind_orc.c int *a = _a, *b = _b, tmp; b 196 arch/x86/kernel/unwind_orc.c *a = *b + delta; b 197 arch/x86/kernel/unwind_orc.c *b = tmp - delta; b 201 arch/x86/kernel/unwind_orc.c orc_b = cur_orc_table + (b - cur_orc_ip_table); b 210 arch/x86/kernel/unwind_orc.c const int *a = _a, *b = _b; b 212 arch/x86/kernel/unwind_orc.c unsigned long b_val = orc_ip(b); b 358 arch/x86/kvm/emulate.c FOP1E(op##b, al) \ b 389 arch/x86/kvm/emulate.c FOP2E(op##b, al, dl) \ b 407 arch/x86/kvm/emulate.c FOP2E(op##b, al, cl) \ b 416 arch/x86/kvm/emulate.c FOP2E(op##b, dl, al) \ b 1231 arch/x86/kvm/emulate.c reg = (ctxt->b & 7) | ((ctxt->rex_prefix & 1) << 3); b 3038 arch/x86/kvm/emulate.c switch (ctxt->b) { b 3938 arch/x86/kvm/emulate.c (ctxt->b == 0xe2 || test_cc(ctxt->b ^ 0x5, ctxt->eflags))) b 5238 arch/x86/kvm/emulate.c switch (ctxt->b = insn_fetch(u8, ctxt)) { b 5279 arch/x86/kvm/emulate.c ctxt->rex_prefix = ctxt->b; b 5286 arch/x86/kvm/emulate.c ctxt->rep_prefix = ctxt->b; b 5304 arch/x86/kvm/emulate.c opcode = opcode_table[ctxt->b]; b 5306 arch/x86/kvm/emulate.c if (ctxt->b == 0x0f) { b 5308 arch/x86/kvm/emulate.c ctxt->b = insn_fetch(u8, ctxt); b 5309 arch/x86/kvm/emulate.c opcode = twobyte_table[ctxt->b]; b 5312 arch/x86/kvm/emulate.c if (ctxt->b == 0x38) { b 5314 arch/x86/kvm/emulate.c ctxt->b = insn_fetch(u8, ctxt); b 5315 arch/x86/kvm/emulate.c opcode = opcode_map_0f_38[ctxt->b]; b 5324 arch/x86/kvm/emulate.c if (ctxt->opcode_len == 1 && (ctxt->b == 0xc5 || ctxt->b == 0xc4) && b 5493 arch/x86/kvm/emulate.c if (((ctxt->b == 0xa6) || (ctxt->b == 0xa7) || b 5494 arch/x86/kvm/emulate.c (ctxt->b == 0xae) || (ctxt->b == 0xaf)) b 5721 arch/x86/kvm/emulate.c switch (ctxt->b) { b 5723 arch/x86/kvm/emulate.c if (test_cc(ctxt->b, ctxt->eflags)) b 5854 arch/x86/kvm/emulate.c switch (ctxt->b) { b 5870 arch/x86/kvm/emulate.c if (test_cc(ctxt->b, ctxt->eflags)) b 5876 arch/x86/kvm/emulate.c if (test_cc(ctxt->b, ctxt->eflags)) b 5880 arch/x86/kvm/emulate.c ctxt->dst.val = test_cc(ctxt->b, ctxt->eflags); b 5127 arch/x86/kvm/vmx/nested.c u8 b; b 5130 arch/x86/kvm/vmx/nested.c b = -1; b 5142 arch/x86/kvm/vmx/nested.c if (kvm_vcpu_read_guest(vcpu, bitmap, &b, 1)) b 5144 arch/x86/kvm/vmx/nested.c if (b & (1 << (port & 7))) b 5203 arch/x86/kvm/vmx/nested.c unsigned char b; b 5204 arch/x86/kvm/vmx/nested.c if (kvm_vcpu_read_guest(vcpu, bitmap + msr_index/8, &b, 1)) b 5206 arch/x86/kvm/vmx/nested.c return 1 & (b >> (msr_index & 7)); b 5300 arch/x86/kvm/vmx/nested.c u8 b; b 5313 arch/x86/kvm/vmx/nested.c if (kvm_vcpu_read_guest(vcpu, bitmap + field/8, &b, 1)) b 5316 arch/x86/kvm/vmx/nested.c return 1 & (b >> (field & 7)); b 6667 arch/x86/kvm/x86.c switch (ctxt->b) { b 6684 arch/x86/kvm/x86.c switch (ctxt->b) { b 15 arch/x86/lib/csum-partial_64.c unsigned short b = a >> 16; b 18 arch/x86/lib/csum-partial_64.c : "=r" (b) b 19 arch/x86/lib/csum-partial_64.c : "0" (b), "r" (a)); b 20 arch/x86/lib/csum-partial_64.c return b; b 73 arch/x86/lib/insn.c insn_byte_t b, lb; b 81 arch/x86/lib/insn.c b = peek_next(insn_byte_t, insn); b 82 arch/x86/lib/insn.c attr = inat_get_opcode_attribute(b); b 86 arch/x86/lib/insn.c if (prefixes->bytes[i] == b) b 91 arch/x86/lib/insn.c prefixes->bytes[nb++] = b; b 105 arch/x86/lib/insn.c lb = b; b 106 arch/x86/lib/insn.c b = peek_next(insn_byte_t, insn); b 107 arch/x86/lib/insn.c attr = inat_get_opcode_attribute(b); b 113 arch/x86/lib/insn.c b = insn->prefixes.bytes[3]; b 116 arch/x86/lib/insn.c prefixes->bytes[i] = b; b 123 arch/x86/lib/insn.c b = peek_next(insn_byte_t, insn); b 124 arch/x86/lib/insn.c attr = inat_get_opcode_attribute(b); b 126 arch/x86/lib/insn.c insn->rex_prefix.value = b; b 129 arch/x86/lib/insn.c if (X86_REX_W(b)) b 137 arch/x86/lib/insn.c b = peek_next(insn_byte_t, insn); b 138 arch/x86/lib/insn.c attr = inat_get_opcode_attribute(b); b 150 arch/x86/lib/insn.c insn->vex_prefix.bytes[0] = b; b 61 arch/x86/lib/iomem.c void memset_io(volatile void __iomem *a, int b, size_t c) b 67 arch/x86/lib/iomem.c memset((void *)a, b, c); b 413 arch/x86/math-emu/errors.c int real_2op_NaN(FPU_REG const *b, u_char tagb, b 425 arch/x86/math-emu/errors.c tagb = FPU_Special(b); b 432 arch/x86/math-emu/errors.c && !((exponent(b) == EXP_OVER) && (b->sigh & 0x80000000))); b 446 arch/x86/math-emu/errors.c signalling = !(a->sigh & b->sigh & 0x40000000); b 447 arch/x86/math-emu/errors.c if (significand(b) > significand(a)) b 448 arch/x86/math-emu/errors.c x = b; b 449 arch/x86/math-emu/errors.c else if (significand(b) == significand(a)) { b 463 arch/x86/math-emu/errors.c signalling = !(b->sigh & 0x40000000); b 464 arch/x86/math-emu/errors.c x = b; b 476 arch/x86/math-emu/errors.c x = b; b 161 arch/x86/math-emu/fpu_emu.h #define setsign(a,b) { if ((b) != 0) signbyte(a) |= 0x80; else signbyte(a) &= 0x7f; } b 162 arch/x86/math-emu/fpu_emu.h #define copysign(a,b) { if (getsign(a)) signbyte(b) |= 0x80; \ b 163 arch/x86/math-emu/fpu_emu.h else signbyte(b) &= 0x7f; } b 10 arch/x86/math-emu/fpu_proto.h extern int real_2op_NaN(FPU_REG const *b, u_char tagb, int deststnr, b 110 arch/x86/math-emu/fpu_proto.h extern int FPU_add(FPU_REG const *b, u_char tagb, int destrnr, int control_w); b 152 arch/x86/math-emu/fpu_proto.h extern int FPU_mul(FPU_REG const *b, u_char tagb, int deststnr, int control_w); b 1315 arch/x86/math-emu/fpu_trig.c FPU_REG *st1_ptr = &st(1), a, b; b 1326 arch/x86/math-emu/fpu_trig.c FPU_to_exp16(st1_ptr, &b); b 1328 arch/x86/math-emu/fpu_trig.c if (poly_l2p1(sign, sign1, &a, &b, st1_ptr)) b 30 arch/x86/math-emu/poly.h asmlinkage void mul64(unsigned long long const *a, unsigned long long const *b, b 48 arch/x86/math-emu/poly.h #define MK_XSIG(a,b,c) { c, b, a } b 30 arch/x86/math-emu/reg_add_sub.c FPU_REG const *b, u_char tagb, u_char signb, b 37 arch/x86/math-emu/reg_add_sub.c int FPU_add(FPU_REG const *b, u_char tagb, int deststnr, int control_w) b 41 arch/x86/math-emu/reg_add_sub.c u_char signb = getsign(b); b 49 arch/x86/math-emu/reg_add_sub.c expb = exponent(b); b 56 arch/x86/math-emu/reg_add_sub.c FPU_u_add(a, b, dest, control_w, signa, expa, expb); b 61 arch/x86/math-emu/reg_add_sub.c diff = a->sigh - b->sigh; /* This works only if the ms bits b 64 arch/x86/math-emu/reg_add_sub.c diff = a->sigl > b->sigl; b 66 arch/x86/math-emu/reg_add_sub.c diff = -(a->sigl < b->sigl); b 72 arch/x86/math-emu/reg_add_sub.c FPU_u_sub(a, b, dest, control_w, signa, b 76 arch/x86/math-emu/reg_add_sub.c FPU_u_sub(b, a, dest, control_w, signb, b 98 arch/x86/math-emu/reg_add_sub.c tagb = FPU_Special(b); b 109 arch/x86/math-emu/reg_add_sub.c FPU_to_exp16(b, &y); b 111 arch/x86/math-emu/reg_add_sub.c b = &y; b 113 arch/x86/math-emu/reg_add_sub.c expb = exponent16(b); b 119 arch/x86/math-emu/reg_add_sub.c return real_2op_NaN(b, tagb, deststnr, a); b 124 arch/x86/math-emu/reg_add_sub.c return add_sub_specials(a, taga, signa, b, tagb, signb, b 131 arch/x86/math-emu/reg_add_sub.c FPU_REG const *a, *b; b 141 arch/x86/math-emu/reg_add_sub.c b = (FPU_REG *) rm; b 144 arch/x86/math-emu/reg_add_sub.c b = &st(rm); b 152 arch/x86/math-emu/reg_add_sub.c signb = getsign(b); b 164 arch/x86/math-emu/reg_add_sub.c expb = exponent(b); b 172 arch/x86/math-emu/reg_add_sub.c diff = a->sigh - b->sigh; /* Works only if ms bits are identical */ b 174 arch/x86/math-emu/reg_add_sub.c diff = a->sigl > b->sigl; b 176 arch/x86/math-emu/reg_add_sub.c diff = -(a->sigl < b->sigl); b 186 arch/x86/math-emu/reg_add_sub.c FPU_u_sub(a, b, dest, control_w, signa, b 198 arch/x86/math-emu/reg_add_sub.c FPU_u_sub(b, a, dest, control_w, sign, expb, b 204 arch/x86/math-emu/reg_add_sub.c FPU_u_add(a, b, dest, control_w, SIGN_POS, expa, b 209 arch/x86/math-emu/reg_add_sub.c FPU_u_add(a, b, dest, control_w, SIGN_NEG, expa, b 229 arch/x86/math-emu/reg_add_sub.c tagb = FPU_Special(b); b 240 arch/x86/math-emu/reg_add_sub.c FPU_to_exp16(b, &y); b 242 arch/x86/math-emu/reg_add_sub.c b = &y; b 244 arch/x86/math-emu/reg_add_sub.c expb = exponent16(b); b 252 arch/x86/math-emu/reg_add_sub.c d1 = b; b 256 arch/x86/math-emu/reg_add_sub.c d2 = b; b 259 arch/x86/math-emu/reg_add_sub.c return real_2op_NaN(b, tagb, deststnr, d1); b 263 arch/x86/math-emu/reg_add_sub.c return real_2op_NaN(b, tagb, deststnr, d2); b 266 arch/x86/math-emu/reg_add_sub.c return add_sub_specials(a, taga, signa, b, tagb, signb ^ SIGN_NEG, b 272 arch/x86/math-emu/reg_add_sub.c FPU_REG const *b, u_char tagb, u_char signb, b 294 arch/x86/math-emu/reg_add_sub.c reg_copy(b, dest); b 295 arch/x86/math-emu/reg_add_sub.c if ((tagb == TW_Denormal) && (b->sigh & 0x80000000)) { b 325 arch/x86/math-emu/reg_add_sub.c FPU_copy_to_regi(b, TAG_Special, deststnr); b 24 arch/x86/math-emu/reg_compare.c static int compare(FPU_REG const *b, int tagb) b 30 arch/x86/math-emu/reg_compare.c u_char st0_sign, signb = getsign(b); b 37 arch/x86/math-emu/reg_compare.c tagb = FPU_Special(b); b 102 arch/x86/math-emu/reg_compare.c (b->sigh & 0xc0000000) == 0x80000000; b 103 arch/x86/math-emu/reg_compare.c unsupported |= !((exponent(b) == EXP_OVER) b 104 arch/x86/math-emu/reg_compare.c && (b->sigh & 0x80000000)); b 124 arch/x86/math-emu/reg_compare.c FPU_to_exp16(b, &y); b 126 arch/x86/math-emu/reg_compare.c b = &y; b 128 arch/x86/math-emu/reg_compare.c expb = exponent16(b); b 131 arch/x86/math-emu/reg_compare.c expb = exponent(b); b 137 arch/x86/math-emu/reg_compare.c if (!(b->sigh & 0x80000000)) b 143 arch/x86/math-emu/reg_compare.c diff = st0_ptr->sigh - b->sigh; /* Works only if ms bits are b 146 arch/x86/math-emu/reg_compare.c diff = st0_ptr->sigl > b->sigl; b 148 arch/x86/math-emu/reg_compare.c diff = -(st0_ptr->sigl < b->sigl); b 31 arch/x86/math-emu/reg_divide.c FPU_REG const *a, *b, *st0_ptr, *st_ptr; b 42 arch/x86/math-emu/reg_divide.c b = &st(0); b 43 arch/x86/math-emu/reg_divide.c st0_ptr = b; b 58 arch/x86/math-emu/reg_divide.c b = (FPU_REG *) rm; b 61 arch/x86/math-emu/reg_divide.c b = &st(rm); b 62 arch/x86/math-emu/reg_divide.c st_ptr = b; b 68 arch/x86/math-emu/reg_divide.c signb = getsign(b); b 78 arch/x86/math-emu/reg_divide.c reg_copy(b, &y); b 93 arch/x86/math-emu/reg_divide.c tagb = FPU_Special(b); b 102 arch/x86/math-emu/reg_divide.c FPU_to_exp16(b, &y); b 899 arch/x86/math-emu/reg_ld_str.c u_char b; b 947 arch/x86/math-emu/reg_ld_str.c b = FPU_div_small(&ll, 10); b 948 arch/x86/math-emu/reg_ld_str.c b |= (FPU_div_small(&ll, 10)) << 4; b 950 arch/x86/math-emu/reg_ld_str.c FPU_put_user(b, d + i); b 30 arch/x86/math-emu/reg_mul.c int FPU_mul(FPU_REG const *b, u_char tagb, int deststnr, int control_w) b 36 arch/x86/math-emu/reg_mul.c u_char sign = (getsign(a) ^ getsign(b)); b 43 arch/x86/math-emu/reg_mul.c FPU_u_mul(a, b, dest, control_w, sign, b 44 arch/x86/math-emu/reg_mul.c exponent(a) + exponent(b)); b 56 arch/x86/math-emu/reg_mul.c tagb = FPU_Special(b); b 66 arch/x86/math-emu/reg_mul.c FPU_to_exp16(b, &y); b 92 arch/x86/math-emu/reg_mul.c return real_2op_NaN(b, tagb, deststnr, &st(0)); b 104 arch/x86/math-emu/reg_mul.c FPU_copy_to_regi(b, TAG_Special, deststnr); b 163 arch/x86/pci/common.c void pcibios_fixup_bus(struct pci_bus *b) b 167 arch/x86/pci/common.c pci_read_bridge_bases(b); b 168 arch/x86/pci/common.c list_for_each_entry(dev, &b->devices, bus_list) b 203 arch/x86/pci/fixup.c #define GET_INDEX(a, b) ((((a) - PCI_DEVICE_ID_INTEL_MCH_PA) << 3) + ((b) & 7)) b 176 arch/x86/platform/uv/tlb_uv.c int node, b; b 179 arch/x86/platform/uv/tlb_uv.c b = uv_node_to_blade_id(node); b 180 arch/x86/platform/uv/tlb_uv.c if (uvhub == b) b 124 arch/x86/tools/insn_decoder_test.c unsigned int b; b 145 arch/x86/tools/insn_decoder_test.c if (sscanf(s, "%x", &b) == 1) { b 146 arch/x86/tools/insn_decoder_test.c insn_buff[nb++] = (unsigned char) b; b 977 arch/x86/tools/relocs.c const uint32_t *a, *b; b 978 arch/x86/tools/relocs.c a = va; b = vb; b 979 arch/x86/tools/relocs.c return (*a == *b)? 0 : (*a > *b)? 1 : -1; b 8 arch/x86/um/asm/checksum_64.h static inline unsigned add32_with_carry(unsigned a, unsigned b) b 13 arch/x86/um/asm/checksum_64.h : "0" (a), "r" (b)); b 25 arch/x86/um/asm/mm_context.h __u32 b; b 191 arch/x86/um/ldt.c ldt_p->b = 0; b 197 arch/x86/um/ldt.c ldt_p->b = LDT_entry_b(&ldt_info); b 271 arch/x86/um/ldt.c if (ldt[i].a != 0 || ldt[i].b != 0) b 289 arch/x86/um/ldt.c if (ldt[i].a != 0 || ldt[i].b != 0) b 58 arch/x86/xen/multicalls.c struct mc_buffer *b = this_cpu_ptr(&mc_buffer); b 70 arch/x86/xen/multicalls.c trace_xen_mc_flush(b->mcidx, b->argidx, b->cbidx); b 73 arch/x86/xen/multicalls.c memcpy(b->debug, b->entries, b 74 arch/x86/xen/multicalls.c b->mcidx * sizeof(struct multicall_entry)); b 77 arch/x86/xen/multicalls.c switch (b->mcidx) { b 80 arch/x86/xen/multicalls.c BUG_ON(b->argidx != 0); b 86 arch/x86/xen/multicalls.c mc = &b->entries[0]; b 95 arch/x86/xen/multicalls.c if (HYPERVISOR_multicall(b->entries, b->mcidx) != 0) b 97 arch/x86/xen/multicalls.c for (i = 0; i < b->mcidx; i++) b 98 arch/x86/xen/multicalls.c if (b->entries[i].result < 0) b 104 arch/x86/xen/multicalls.c ret, b->mcidx, smp_processor_id()); b 105 arch/x86/xen/multicalls.c for (i = 0; i < b->mcidx; i++) { b 106 arch/x86/xen/multicalls.c if (b->entries[i].result < 0) { b 110 arch/x86/xen/multicalls.c b->debug[i].op, b 111 arch/x86/xen/multicalls.c b->debug[i].args[0], b 112 arch/x86/xen/multicalls.c b->entries[i].result, b 113 arch/x86/xen/multicalls.c b->caller[i]); b 117 arch/x86/xen/multicalls.c b->entries[i].op, b 118 arch/x86/xen/multicalls.c b->entries[i].args[0], b 119 arch/x86/xen/multicalls.c b->entries[i].result); b 125 arch/x86/xen/multicalls.c b->mcidx = 0; b 126 arch/x86/xen/multicalls.c b->argidx = 0; b 128 arch/x86/xen/multicalls.c for (i = 0; i < b->cbidx; i++) { b 129 arch/x86/xen/multicalls.c struct callback *cb = &b->callbacks[i]; b 133 arch/x86/xen/multicalls.c b->cbidx = 0; b 140 arch/x86/xen/multicalls.c struct mc_buffer *b = this_cpu_ptr(&mc_buffer); b 142 arch/x86/xen/multicalls.c unsigned argidx = roundup(b->argidx, sizeof(u64)); b 147 arch/x86/xen/multicalls.c BUG_ON(b->argidx >= MC_ARGS); b 149 arch/x86/xen/multicalls.c if (unlikely(b->mcidx == MC_BATCH || b 151 arch/x86/xen/multicalls.c trace_xen_mc_flush_reason((b->mcidx == MC_BATCH) ? b 154 arch/x86/xen/multicalls.c argidx = roundup(b->argidx, sizeof(u64)); b 157 arch/x86/xen/multicalls.c ret.mc = &b->entries[b->mcidx]; b 159 arch/x86/xen/multicalls.c b->caller[b->mcidx] = __builtin_return_address(0); b 161 arch/x86/xen/multicalls.c b->mcidx++; b 162 arch/x86/xen/multicalls.c ret.args = &b->args[argidx]; b 163 arch/x86/xen/multicalls.c b->argidx = argidx + args; b 165 arch/x86/xen/multicalls.c BUG_ON(b->argidx >= MC_ARGS); b 171 arch/x86/xen/multicalls.c struct mc_buffer *b = this_cpu_ptr(&mc_buffer); b 175 arch/x86/xen/multicalls.c BUG_ON(b->argidx >= MC_ARGS); b 177 arch/x86/xen/multicalls.c if (unlikely(b->mcidx == 0 || b 178 arch/x86/xen/multicalls.c b->entries[b->mcidx - 1].op != op)) { b 183 arch/x86/xen/multicalls.c if (unlikely((b->argidx + size) >= MC_ARGS)) { b 188 arch/x86/xen/multicalls.c ret.mc = &b->entries[b->mcidx - 1]; b 189 arch/x86/xen/multicalls.c ret.args = &b->args[b->argidx]; b 190 arch/x86/xen/multicalls.c b->argidx += size; b 192 arch/x86/xen/multicalls.c BUG_ON(b->argidx >= MC_ARGS); b 201 arch/x86/xen/multicalls.c struct mc_buffer *b = this_cpu_ptr(&mc_buffer); b 204 arch/x86/xen/multicalls.c if (b->cbidx == MC_BATCH) { b 211 arch/x86/xen/multicalls.c cb = &b->callbacks[b->cbidx++]; b 86 arch/xtensa/include/asm/asmmacro.h b\ncond \at, 99f b 22 arch/xtensa/include/asm/coprocessor.h .macro xchal_sa_start a b b 70 arch/xtensa/include/asm/page.h # define DCACHE_ALIAS_EQ(a,b) ((((a) ^ (b)) & DCACHE_ALIAS_MASK) == 0) b 81 arch/xtensa/include/asm/page.h # define ICACHE_ALIAS_EQ(a,b) ((((a) ^ (b)) & ICACHE_ALIAS_MASK) == 0) b 295 arch/xtensa/include/asm/pgtable.h #define pte_same(a,b) (pte_val(a) == pte_val(b)) b 38 arch/xtensa/include/asm/uaccess.h #define segment_eq(a, b) ((a).seg == (b).seg) b 22 arch/xtensa/kernel/platform.c #define _F(r,f,a,b) \ b 23 arch/xtensa/kernel/platform.c r __platform_##f a b; \ b 67 arch/xtensa/platforms/iss/include/platform/simcall.h static inline int __simc(int a, int b, int c, int d) b 71 arch/xtensa/platforms/iss/include/platform/simcall.h register int b1 asm("a3") = b; b 292 arch/xtensa/variants/test_kc705_be/include/variant/tie-asm.h .macro xchal_cp0_store p a b c d continue=0 ofs=-1 select=-1 ; .endm b 293 arch/xtensa/variants/test_kc705_be/include/variant/tie-asm.h .macro xchal_cp0_load p a b c d continue=0 ofs=-1 select=-1 ; .endm b 294 arch/xtensa/variants/test_kc705_be/include/variant/tie-asm.h .macro xchal_cp2_store p a b c d continue=0 ofs=-1 select=-1 ; .endm b 295 arch/xtensa/variants/test_kc705_be/include/variant/tie-asm.h .macro xchal_cp2_load p a b c d continue=0 ofs=-1 select=-1 ; .endm b 296 arch/xtensa/variants/test_kc705_be/include/variant/tie-asm.h .macro xchal_cp3_store p a b c d continue=0 ofs=-1 select=-1 ; .endm b 297 arch/xtensa/variants/test_kc705_be/include/variant/tie-asm.h .macro xchal_cp3_load p a b c d continue=0 ofs=-1 select=-1 ; .endm b 298 arch/xtensa/variants/test_kc705_be/include/variant/tie-asm.h .macro xchal_cp4_store p a b c d continue=0 ofs=-1 select=-1 ; .endm b 299 arch/xtensa/variants/test_kc705_be/include/variant/tie-asm.h .macro xchal_cp4_load p a b c d continue=0 ofs=-1 select=-1 ; .endm b 300 arch/xtensa/variants/test_kc705_be/include/variant/tie-asm.h .macro xchal_cp5_store p a b c d continue=0 ofs=-1 select=-1 ; .endm b 301 arch/xtensa/variants/test_kc705_be/include/variant/tie-asm.h .macro xchal_cp5_load p a b c d continue=0 ofs=-1 select=-1 ; .endm b 302 arch/xtensa/variants/test_kc705_be/include/variant/tie-asm.h .macro xchal_cp6_store p a b c d continue=0 ofs=-1 select=-1 ; .endm b 303 arch/xtensa/variants/test_kc705_be/include/variant/tie-asm.h .macro xchal_cp6_load p a b c d continue=0 ofs=-1 select=-1 ; .endm b 304 arch/xtensa/variants/test_kc705_be/include/variant/tie-asm.h .macro xchal_cp7_store p a b c d continue=0 ofs=-1 select=-1 ; .endm b 305 arch/xtensa/variants/test_kc705_be/include/variant/tie-asm.h .macro xchal_cp7_load p a b c d continue=0 ofs=-1 select=-1 ; .endm b 313 arch/xtensa/variants/test_kc705_hifi/include/variant/tie-asm.h .macro xchal_cp0_store p a b c d continue=0 ofs=-1 select=-1 ; .endm b 314 arch/xtensa/variants/test_kc705_hifi/include/variant/tie-asm.h .macro xchal_cp0_load p a b c d continue=0 ofs=-1 select=-1 ; .endm b 315 arch/xtensa/variants/test_kc705_hifi/include/variant/tie-asm.h .macro xchal_cp2_store p a b c d continue=0 ofs=-1 select=-1 ; .endm b 316 arch/xtensa/variants/test_kc705_hifi/include/variant/tie-asm.h .macro xchal_cp2_load p a b c d continue=0 ofs=-1 select=-1 ; .endm b 317 arch/xtensa/variants/test_kc705_hifi/include/variant/tie-asm.h .macro xchal_cp3_store p a b c d continue=0 ofs=-1 select=-1 ; .endm b 318 arch/xtensa/variants/test_kc705_hifi/include/variant/tie-asm.h .macro xchal_cp3_load p a b c d continue=0 ofs=-1 select=-1 ; .endm b 319 arch/xtensa/variants/test_kc705_hifi/include/variant/tie-asm.h .macro xchal_cp4_store p a b c d continue=0 ofs=-1 select=-1 ; .endm b 320 arch/xtensa/variants/test_kc705_hifi/include/variant/tie-asm.h .macro xchal_cp4_load p a b c d continue=0 ofs=-1 select=-1 ; .endm b 321 arch/xtensa/variants/test_kc705_hifi/include/variant/tie-asm.h .macro xchal_cp5_store p a b c d continue=0 ofs=-1 select=-1 ; .endm b 322 arch/xtensa/variants/test_kc705_hifi/include/variant/tie-asm.h .macro xchal_cp5_load p a b c d continue=0 ofs=-1 select=-1 ; .endm b 323 arch/xtensa/variants/test_kc705_hifi/include/variant/tie-asm.h .macro xchal_cp6_store p a b c d continue=0 ofs=-1 select=-1 ; .endm b 324 arch/xtensa/variants/test_kc705_hifi/include/variant/tie-asm.h .macro xchal_cp6_load p a b c d continue=0 ofs=-1 select=-1 ; .endm b 325 arch/xtensa/variants/test_kc705_hifi/include/variant/tie-asm.h .macro xchal_cp7_store p a b c d continue=0 ofs=-1 select=-1 ; .endm b 326 arch/xtensa/variants/test_kc705_hifi/include/variant/tie-asm.h .macro xchal_cp7_load p a b c d continue=0 ofs=-1 select=-1 ; .endm b 167 arch/xtensa/variants/test_mmuhifi_c3/include/variant/tie-asm.h .macro xchal_cp0_store p a b c d continue=0 ofs=-1 select=-1 ; .endm b 168 arch/xtensa/variants/test_mmuhifi_c3/include/variant/tie-asm.h .macro xchal_cp0_load p a b c d continue=0 ofs=-1 select=-1 ; .endm b 169 arch/xtensa/variants/test_mmuhifi_c3/include/variant/tie-asm.h .macro xchal_cp2_store p a b c d continue=0 ofs=-1 select=-1 ; .endm b 170 arch/xtensa/variants/test_mmuhifi_c3/include/variant/tie-asm.h .macro xchal_cp2_load p a b c d continue=0 ofs=-1 select=-1 ; .endm b 171 arch/xtensa/variants/test_mmuhifi_c3/include/variant/tie-asm.h .macro xchal_cp3_store p a b c d continue=0 ofs=-1 select=-1 ; .endm b 172 arch/xtensa/variants/test_mmuhifi_c3/include/variant/tie-asm.h .macro xchal_cp3_load p a b c d continue=0 ofs=-1 select=-1 ; .endm b 173 arch/xtensa/variants/test_mmuhifi_c3/include/variant/tie-asm.h .macro xchal_cp4_store p a b c d continue=0 ofs=-1 select=-1 ; .endm b 174 arch/xtensa/variants/test_mmuhifi_c3/include/variant/tie-asm.h .macro xchal_cp4_load p a b c d continue=0 ofs=-1 select=-1 ; .endm b 175 arch/xtensa/variants/test_mmuhifi_c3/include/variant/tie-asm.h .macro xchal_cp5_store p a b c d continue=0 ofs=-1 select=-1 ; .endm b 176 arch/xtensa/variants/test_mmuhifi_c3/include/variant/tie-asm.h .macro xchal_cp5_load p a b c d continue=0 ofs=-1 select=-1 ; .endm b 177 arch/xtensa/variants/test_mmuhifi_c3/include/variant/tie-asm.h .macro xchal_cp6_store p a b c d continue=0 ofs=-1 select=-1 ; .endm b 178 arch/xtensa/variants/test_mmuhifi_c3/include/variant/tie-asm.h .macro xchal_cp6_load p a b c d continue=0 ofs=-1 select=-1 ; .endm b 179 arch/xtensa/variants/test_mmuhifi_c3/include/variant/tie-asm.h .macro xchal_cp7_store p a b c d continue=0 ofs=-1 select=-1 ; .endm b 180 arch/xtensa/variants/test_mmuhifi_c3/include/variant/tie-asm.h .macro xchal_cp7_load p a b c d continue=0 ofs=-1 select=-1 ; .endm b 18 block/bfq-wf2q.c static int bfq_gt(u64 a, u64 b) b 20 block/bfq-wf2q.c return (s64)(a - b) > 0; b 659 block/bio.c struct bio *b; b 661 block/bio.c b = bio_alloc_bioset(gfp_mask, 0, bs); b 662 block/bio.c if (!b) b 665 block/bio.c __bio_clone_fast(b, bio); b 670 block/bio.c ret = bio_integrity_clone(b, bio, gfp_mask); b 673 block/bio.c bio_put(b); b 678 block/bio.c return b; b 743 block/blk-core.c char b[BDEVNAME_SIZE]; b 747 block/blk-core.c bio_devname(bio, b), bio->bi_opf, b 792 block/blk-core.c char b[BDEVNAME_SIZE]; b 800 block/blk-core.c bio_devname(bio, b), part->partno); b 875 block/blk-core.c char b[BDEVNAME_SIZE]; b 884 block/blk-core.c bio_devname(bio, b), (long long)bio->bi_iter.bi_sector); b 1172 block/blk-core.c char b[BDEVNAME_SIZE]; b 1177 block/blk-core.c bio_devname(bio, b), count); b 1711 block/blk-mq.c static int plug_rq_cmp(void *priv, struct list_head *a, struct list_head *b) b 1714 block/blk-mq.c struct request *rqb = container_of(b, struct request, queuelist); b 470 block/blk-settings.c void blk_queue_stack_limits(struct request_queue *t, struct request_queue *b) b 472 block/blk-settings.c blk_stack_limits(&t->limits, &b->limits, 0); b 497 block/blk-settings.c int blk_stack_limits(struct queue_limits *t, struct queue_limits *b, b 502 block/blk-settings.c t->max_sectors = min_not_zero(t->max_sectors, b->max_sectors); b 503 block/blk-settings.c t->max_hw_sectors = min_not_zero(t->max_hw_sectors, b->max_hw_sectors); b 504 block/blk-settings.c t->max_dev_sectors = min_not_zero(t->max_dev_sectors, b->max_dev_sectors); b 506 block/blk-settings.c b->max_write_same_sectors); b 508 block/blk-settings.c b->max_write_zeroes_sectors); b 509 block/blk-settings.c t->bounce_pfn = min_not_zero(t->bounce_pfn, b->bounce_pfn); b 512 block/blk-settings.c b->seg_boundary_mask); b 514 block/blk-settings.c b->virt_boundary_mask); b 516 block/blk-settings.c t->max_segments = min_not_zero(t->max_segments, b->max_segments); b 518 block/blk-settings.c b->max_discard_segments); b 520 block/blk-settings.c b->max_integrity_segments); b 523 block/blk-settings.c b->max_segment_size); b 525 block/blk-settings.c t->misaligned |= b->misaligned; b 527 block/blk-settings.c alignment = queue_limit_alignment_offset(b, start); b 536 block/blk-settings.c bottom = max(b->physical_block_size, b->io_min) + alignment; b 546 block/blk-settings.c b->logical_block_size); b 549 block/blk-settings.c b->physical_block_size); b 551 block/blk-settings.c t->io_min = max(t->io_min, b->io_min); b 552 block/blk-settings.c t->io_opt = lcm_not_zero(t->io_opt, b->io_opt); b 577 block/blk-settings.c b->raid_partial_stripes_expensive); b 590 block/blk-settings.c if (b->discard_granularity) { b 591 block/blk-settings.c alignment = queue_limit_discard_alignment(b, start); b 596 block/blk-settings.c bottom = b->discard_granularity + alignment; b 604 block/blk-settings.c b->max_discard_sectors); b 606 block/blk-settings.c b->max_hw_discard_sectors); b 608 block/blk-settings.c b->discard_granularity); b 613 block/blk-settings.c if (b->chunk_sectors) b 615 block/blk-settings.c b->chunk_sectors); b 38 block/partitions/amiga.c char b[BDEVNAME_SIZE]; b 47 block/partitions/amiga.c bdevname(state->bdev, b), blk); b 69 block/partitions/amiga.c bdevname(state->bdev, b), blk); b 90 block/partitions/amiga.c bdevname(state->bdev, b), blk); b 61 block/partitions/ibm.c ptr->b; b 41 block/partitions/sgi.c char b[BDEVNAME_SIZE]; b 61 block/partitions/sgi.c bdevname(state->bdev, b)); b 61 block/partitions/sun.c char b[BDEVNAME_SIZE]; b 82 block/partitions/sun.c bdevname(state->bdev, b)); b 56 block/partitions/sysv68.c struct dkblk0 *b; b 64 block/partitions/sysv68.c b = (struct dkblk0 *)data; b 65 block/partitions/sysv68.c if (memcmp(b->dk_vid.vid_mac, "MOTOROLA", sizeof(b->dk_vid.vid_mac))) { b 69 block/partitions/sysv68.c slices = be16_to_cpu(b->dk_ios.ios_slccnt); b 70 block/partitions/sysv68.c i = be32_to_cpu(b->dk_ios.ios_slcblk); b 784 block/sed-opal.c ssize_t i, b = 0; b 792 block/sed-opal.c u_integer |= ((u64)pos[i] << (8 * b)); b 793 block/sed-opal.c b++; b 196 crypto/adiantum.c u64 x = le64_to_cpu(v1->b); b 197 crypto/adiantum.c u64 y = le64_to_cpu(v2->b); b 199 crypto/adiantum.c r->b = cpu_to_le64(x + y); b 207 crypto/adiantum.c u64 x = le64_to_cpu(v1->b); b 208 crypto/adiantum.c u64 y = le64_to_cpu(v2->b); b 210 crypto/adiantum.c r->b = cpu_to_le64(x - y); b 948 crypto/algapi.c u8 *b = (a + size); b 952 crypto/algapi.c c = *--b + 1; b 953 crypto/algapi.c *b = c; b 961 crypto/algapi.c __be32 *b = (__be32 *)(a + size); b 965 crypto/algapi.c IS_ALIGNED((unsigned long)b, __alignof__(*b))) b 967 crypto/algapi.c c = be32_to_cpu(*--b) + 1; b 968 crypto/algapi.c *b = cpu_to_be32(c); b 204 crypto/asymmetric_keys/verify_pefile.c static int pefile_compare_shdrs(const void *a, const void *b) b 207 crypto/asymmetric_keys/verify_pefile.c const struct section_header *shdrb = b; b 117 crypto/asymmetric_keys/x509_public_key.c bool b = asymmetric_key_id_same(cert->id, cert->sig->auth_ids[0]); b 119 crypto/asymmetric_keys/x509_public_key.c if (!a && !b) b 123 crypto/asymmetric_keys/x509_public_key.c if (((a && !b) || (b && !a)) && b 26 crypto/async_tx/async_pq.c #define P(b, d) (b[d-2]) b 27 crypto/async_tx/async_pq.c #define Q(b, d) (b[d-1]) b 27 crypto/async_tx/async_raid6_recov.c u8 *a, *b, *c; b 70 crypto/async_tx/async_raid6_recov.c b = page_address(srcs[1]); b 75 crypto/async_tx/async_raid6_recov.c bx = bmul[*b++]; b 150 crypto/async_tx/async_raid6_recov.c struct page *p, *q, *a, *b; b 162 crypto/async_tx/async_raid6_recov.c b = blocks[failb]; b 171 crypto/async_tx/async_raid6_recov.c tx = async_sum_product(b, srcs, coef, bytes, submit); b 175 crypto/async_tx/async_raid6_recov.c srcs[1] = b; b 304 crypto/blowfish_common.c #define ROUND(a, b, n) ({ b ^= P[n]; a ^= bf_F(b); }) b 34 crypto/blowfish_generic.c #define ROUND(a, b, n) ({ b ^= P[n]; a ^= bf_F(b); }) b 367 crypto/ecc.c static uint128_t add_128_128(uint128_t a, uint128_t b) b 371 crypto/ecc.c result.m_low = a.m_low + b.m_low; b 372 crypto/ecc.c result.m_high = a.m_high + b.m_high + (result.m_low < a.m_low); b 861 crypto/ecc.c u64 a[ECC_MAX_DIGITS], b[ECC_MAX_DIGITS]; b 872 crypto/ecc.c vli_set(b, mod, ndigits); b 877 crypto/ecc.c while ((cmp_result = vli_cmp(a, b, ndigits)) != 0) { b 889 crypto/ecc.c } else if (EVEN(b)) { b 890 crypto/ecc.c vli_rshift1(b, ndigits); b 899 crypto/ecc.c vli_sub(a, a, b, ndigits); b 913 crypto/ecc.c vli_sub(b, b, a, ndigits); b 914 crypto/ecc.c vli_rshift1(b, ndigits); b 1446 crypto/ecc.c vli_mod_add(w, w, curve->b, curve->p, pk->ndigits); /* a·x + b */ b 70 crypto/ecc.h u64 *b; b 28 crypto/ecc_curve_defs.h .b = nist_p192_b b 54 crypto/ecc_curve_defs.h .b = nist_p256_b b 58 crypto/ecrdsa_defs.h .b = cp256a_b b 91 crypto/ecrdsa_defs.h .b = cp256b_b b 128 crypto/ecrdsa_defs.h .b = cp256c_b b 177 crypto/ecrdsa_defs.h .b = tc512a_b b 222 crypto/ecrdsa_defs.h .b = tc512b_b b 237 crypto/gcm.c lengths.b = cpu_to_be64(gctx->cryptlen * 8); b 142 crypto/gf128mul.c u64 b = be64_to_cpu(x->b); b 143 crypto/gf128mul.c u64 _tt = gf128mul_table_le[b & 0xff]; b 145 crypto/gf128mul.c x->b = cpu_to_be64((b >> 8) | (a << 56)); b 152 crypto/gf128mul.c u64 b = be64_to_cpu(x->b); b 155 crypto/gf128mul.c x->a = cpu_to_be64((a << 8) | (b >> 56)); b 156 crypto/gf128mul.c x->b = cpu_to_be64((b << 8) ^ _tt); b 162 crypto/gf128mul.c u64 b = le64_to_cpu(x->b); b 165 crypto/gf128mul.c r->a = cpu_to_le64((a << 8) | (b >> 56)); b 166 crypto/gf128mul.c r->b = cpu_to_le64((b << 8) ^ _tt); b 170 crypto/gf128mul.c void gf128mul_lle(be128 *r, const be128 *b) b 181 crypto/gf128mul.c u8 ch = ((u8 *)b)[15 - i]; b 208 crypto/gf128mul.c void gf128mul_bbe(be128 *r, const be128 *b) b 219 crypto/gf128mul.c u8 ch = ((u8 *)b)[i]; b 57 crypto/lrw.c static inline void setbit128_bbe(void *b, int bit) b 65 crypto/lrw.c ), b); b 63 crypto/md4.c #define ROUND1(a,b,c,d,k,s) (a = lshift(a + F(b,c,d) + k, s)) b 64 crypto/md4.c #define ROUND2(a,b,c,d,k,s) (a = lshift(a + G(b,c,d) + k + (u32)0x5A827999,s)) b 65 crypto/md4.c #define ROUND3(a,b,c,d,k,s) (a = lshift(a + H(b,c,d) + k + (u32)0x6ED9EBA1,s)) b 69 crypto/md4.c u32 a, b, c, d; b 72 crypto/md4.c b = hash[1]; b 76 crypto/md4.c ROUND1(a, b, c, d, in[0], 3); b 77 crypto/md4.c ROUND1(d, a, b, c, in[1], 7); b 78 crypto/md4.c ROUND1(c, d, a, b, in[2], 11); b 79 crypto/md4.c ROUND1(b, c, d, a, in[3], 19); b 80 crypto/md4.c ROUND1(a, b, c, d, in[4], 3); b 81 crypto/md4.c ROUND1(d, a, b, c, in[5], 7); b 82 crypto/md4.c ROUND1(c, d, a, b, in[6], 11); b 83 crypto/md4.c ROUND1(b, c, d, a, in[7], 19); b 84 crypto/md4.c ROUND1(a, b, c, d, in[8], 3); b 85 crypto/md4.c ROUND1(d, a, b, c, in[9], 7); b 86 crypto/md4.c ROUND1(c, d, a, b, in[10], 11); b 87 crypto/md4.c ROUND1(b, c, d, a, in[11], 19); b 88 crypto/md4.c ROUND1(a, b, c, d, in[12], 3); b 89 crypto/md4.c ROUND1(d, a, b, c, in[13], 7); b 90 crypto/md4.c ROUND1(c, d, a, b, in[14], 11); b 91 crypto/md4.c ROUND1(b, c, d, a, in[15], 19); b 93 crypto/md4.c ROUND2(a, b, c, d,in[ 0], 3); b 94 crypto/md4.c ROUND2(d, a, b, c, in[4], 5); b 95 crypto/md4.c ROUND2(c, d, a, b, in[8], 9); b 96 crypto/md4.c ROUND2(b, c, d, a, in[12], 13); b 97 crypto/md4.c ROUND2(a, b, c, d, in[1], 3); b 98 crypto/md4.c ROUND2(d, a, b, c, in[5], 5); b 99 crypto/md4.c ROUND2(c, d, a, b, in[9], 9); b 100 crypto/md4.c ROUND2(b, c, d, a, in[13], 13); b 101 crypto/md4.c ROUND2(a, b, c, d, in[2], 3); b 102 crypto/md4.c ROUND2(d, a, b, c, in[6], 5); b 103 crypto/md4.c ROUND2(c, d, a, b, in[10], 9); b 104 crypto/md4.c ROUND2(b, c, d, a, in[14], 13); b 105 crypto/md4.c ROUND2(a, b, c, d, in[3], 3); b 106 crypto/md4.c ROUND2(d, a, b, c, in[7], 5); b 107 crypto/md4.c ROUND2(c, d, a, b, in[11], 9); b 108 crypto/md4.c ROUND2(b, c, d, a, in[15], 13); b 110 crypto/md4.c ROUND3(a, b, c, d,in[ 0], 3); b 111 crypto/md4.c ROUND3(d, a, b, c, in[8], 9); b 112 crypto/md4.c ROUND3(c, d, a, b, in[4], 11); b 113 crypto/md4.c ROUND3(b, c, d, a, in[12], 15); b 114 crypto/md4.c ROUND3(a, b, c, d, in[2], 3); b 115 crypto/md4.c ROUND3(d, a, b, c, in[10], 9); b 116 crypto/md4.c ROUND3(c, d, a, b, in[6], 11); b 117 crypto/md4.c ROUND3(b, c, d, a, in[14], 15); b 118 crypto/md4.c ROUND3(a, b, c, d, in[1], 3); b 119 crypto/md4.c ROUND3(d, a, b, c, in[9], 9); b 120 crypto/md4.c ROUND3(c, d, a, b, in[5], 11); b 121 crypto/md4.c ROUND3(b, c, d, a, in[13], 15); b 122 crypto/md4.c ROUND3(a, b, c, d, in[3], 3); b 123 crypto/md4.c ROUND3(d, a, b, c, in[11], 9); b 124 crypto/md4.c ROUND3(c, d, a, b, in[7], 11); b 125 crypto/md4.c ROUND3(b, c, d, a, in[15], 15); b 128 crypto/md4.c hash[1] += b; b 45 crypto/md5.c u32 a, b, c, d; b 48 crypto/md5.c b = hash[1]; b 52 crypto/md5.c MD5STEP(F1, a, b, c, d, in[0] + 0xd76aa478, 7); b 53 crypto/md5.c MD5STEP(F1, d, a, b, c, in[1] + 0xe8c7b756, 12); b 54 crypto/md5.c MD5STEP(F1, c, d, a, b, in[2] + 0x242070db, 17); b 55 crypto/md5.c MD5STEP(F1, b, c, d, a, in[3] + 0xc1bdceee, 22); b 56 crypto/md5.c MD5STEP(F1, a, b, c, d, in[4] + 0xf57c0faf, 7); b 57 crypto/md5.c MD5STEP(F1, d, a, b, c, in[5] + 0x4787c62a, 12); b 58 crypto/md5.c MD5STEP(F1, c, d, a, b, in[6] + 0xa8304613, 17); b 59 crypto/md5.c MD5STEP(F1, b, c, d, a, in[7] + 0xfd469501, 22); b 60 crypto/md5.c MD5STEP(F1, a, b, c, d, in[8] + 0x698098d8, 7); b 61 crypto/md5.c MD5STEP(F1, d, a, b, c, in[9] + 0x8b44f7af, 12); b 62 crypto/md5.c MD5STEP(F1, c, d, a, b, in[10] + 0xffff5bb1, 17); b 63 crypto/md5.c MD5STEP(F1, b, c, d, a, in[11] + 0x895cd7be, 22); b 64 crypto/md5.c MD5STEP(F1, a, b, c, d, in[12] + 0x6b901122, 7); b 65 crypto/md5.c MD5STEP(F1, d, a, b, c, in[13] + 0xfd987193, 12); b 66 crypto/md5.c MD5STEP(F1, c, d, a, b, in[14] + 0xa679438e, 17); b 67 crypto/md5.c MD5STEP(F1, b, c, d, a, in[15] + 0x49b40821, 22); b 69 crypto/md5.c MD5STEP(F2, a, b, c, d, in[1] + 0xf61e2562, 5); b 70 crypto/md5.c MD5STEP(F2, d, a, b, c, in[6] + 0xc040b340, 9); b 71 crypto/md5.c MD5STEP(F2, c, d, a, b, in[11] + 0x265e5a51, 14); b 72 crypto/md5.c MD5STEP(F2, b, c, d, a, in[0] + 0xe9b6c7aa, 20); b 73 crypto/md5.c MD5STEP(F2, a, b, c, d, in[5] + 0xd62f105d, 5); b 74 crypto/md5.c MD5STEP(F2, d, a, b, c, in[10] + 0x02441453, 9); b 75 crypto/md5.c MD5STEP(F2, c, d, a, b, in[15] + 0xd8a1e681, 14); b 76 crypto/md5.c MD5STEP(F2, b, c, d, a, in[4] + 0xe7d3fbc8, 20); b 77 crypto/md5.c MD5STEP(F2, a, b, c, d, in[9] + 0x21e1cde6, 5); b 78 crypto/md5.c MD5STEP(F2, d, a, b, c, in[14] + 0xc33707d6, 9); b 79 crypto/md5.c MD5STEP(F2, c, d, a, b, in[3] + 0xf4d50d87, 14); b 80 crypto/md5.c MD5STEP(F2, b, c, d, a, in[8] + 0x455a14ed, 20); b 81 crypto/md5.c MD5STEP(F2, a, b, c, d, in[13] + 0xa9e3e905, 5); b 82 crypto/md5.c MD5STEP(F2, d, a, b, c, in[2] + 0xfcefa3f8, 9); b 83 crypto/md5.c MD5STEP(F2, c, d, a, b, in[7] + 0x676f02d9, 14); b 84 crypto/md5.c MD5STEP(F2, b, c, d, a, in[12] + 0x8d2a4c8a, 20); b 86 crypto/md5.c MD5STEP(F3, a, b, c, d, in[5] + 0xfffa3942, 4); b 87 crypto/md5.c MD5STEP(F3, d, a, b, c, in[8] + 0x8771f681, 11); b 88 crypto/md5.c MD5STEP(F3, c, d, a, b, in[11] + 0x6d9d6122, 16); b 89 crypto/md5.c MD5STEP(F3, b, c, d, a, in[14] + 0xfde5380c, 23); b 90 crypto/md5.c MD5STEP(F3, a, b, c, d, in[1] + 0xa4beea44, 4); b 91 crypto/md5.c MD5STEP(F3, d, a, b, c, in[4] + 0x4bdecfa9, 11); b 92 crypto/md5.c MD5STEP(F3, c, d, a, b, in[7] + 0xf6bb4b60, 16); b 93 crypto/md5.c MD5STEP(F3, b, c, d, a, in[10] + 0xbebfbc70, 23); b 94 crypto/md5.c MD5STEP(F3, a, b, c, d, in[13] + 0x289b7ec6, 4); b 95 crypto/md5.c MD5STEP(F3, d, a, b, c, in[0] + 0xeaa127fa, 11); b 96 crypto/md5.c MD5STEP(F3, c, d, a, b, in[3] + 0xd4ef3085, 16); b 97 crypto/md5.c MD5STEP(F3, b, c, d, a, in[6] + 0x04881d05, 23); b 98 crypto/md5.c MD5STEP(F3, a, b, c, d, in[9] + 0xd9d4d039, 4); b 99 crypto/md5.c MD5STEP(F3, d, a, b, c, in[12] + 0xe6db99e5, 11); b 100 crypto/md5.c MD5STEP(F3, c, d, a, b, in[15] + 0x1fa27cf8, 16); b 101 crypto/md5.c MD5STEP(F3, b, c, d, a, in[2] + 0xc4ac5665, 23); b 103 crypto/md5.c MD5STEP(F4, a, b, c, d, in[0] + 0xf4292244, 6); b 104 crypto/md5.c MD5STEP(F4, d, a, b, c, in[7] + 0x432aff97, 10); b 105 crypto/md5.c MD5STEP(F4, c, d, a, b, in[14] + 0xab9423a7, 15); b 106 crypto/md5.c MD5STEP(F4, b, c, d, a, in[5] + 0xfc93a039, 21); b 107 crypto/md5.c MD5STEP(F4, a, b, c, d, in[12] + 0x655b59c3, 6); b 108 crypto/md5.c MD5STEP(F4, d, a, b, c, in[3] + 0x8f0ccc92, 10); b 109 crypto/md5.c MD5STEP(F4, c, d, a, b, in[10] + 0xffeff47d, 15); b 110 crypto/md5.c MD5STEP(F4, b, c, d, a, in[1] + 0x85845dd1, 21); b 111 crypto/md5.c MD5STEP(F4, a, b, c, d, in[8] + 0x6fa87e4f, 6); b 112 crypto/md5.c MD5STEP(F4, d, a, b, c, in[15] + 0xfe2ce6e0, 10); b 113 crypto/md5.c MD5STEP(F4, c, d, a, b, in[6] + 0xa3014314, 15); b 114 crypto/md5.c MD5STEP(F4, b, c, d, a, in[13] + 0x4e0811a1, 21); b 115 crypto/md5.c MD5STEP(F4, a, b, c, d, in[4] + 0xf7537e82, 6); b 116 crypto/md5.c MD5STEP(F4, d, a, b, c, in[11] + 0xbd3af235, 10); b 117 crypto/md5.c MD5STEP(F4, c, d, a, b, in[2] + 0x2ad7d2bb, 15); b 118 crypto/md5.c MD5STEP(F4, b, c, d, a, in[9] + 0xeb86d391, 21); b 121 crypto/md5.c hash[1] += b; b 68 crypto/memneq.c __crypto_memneq_generic(const void *a, const void *b, size_t size) b 74 crypto/memneq.c neq |= *(unsigned long *)a ^ *(unsigned long *)b; b 77 crypto/memneq.c b += sizeof(unsigned long); b 82 crypto/memneq.c neq |= *(unsigned char *)a ^ *(unsigned char *)b; b 85 crypto/memneq.c b += 1; b 92 crypto/memneq.c static inline unsigned long __crypto_memneq_16(const void *a, const void *b) b 98 crypto/memneq.c neq |= *(unsigned long *)(a) ^ *(unsigned long *)(b); b 100 crypto/memneq.c neq |= *(unsigned long *)(a+8) ^ *(unsigned long *)(b+8); b 103 crypto/memneq.c neq |= *(unsigned int *)(a) ^ *(unsigned int *)(b); b 105 crypto/memneq.c neq |= *(unsigned int *)(a+4) ^ *(unsigned int *)(b+4); b 107 crypto/memneq.c neq |= *(unsigned int *)(a+8) ^ *(unsigned int *)(b+8); b 109 crypto/memneq.c neq |= *(unsigned int *)(a+12) ^ *(unsigned int *)(b+12); b 114 crypto/memneq.c neq |= *(unsigned char *)(a) ^ *(unsigned char *)(b); b 116 crypto/memneq.c neq |= *(unsigned char *)(a+1) ^ *(unsigned char *)(b+1); b 118 crypto/memneq.c neq |= *(unsigned char *)(a+2) ^ *(unsigned char *)(b+2); b 120 crypto/memneq.c neq |= *(unsigned char *)(a+3) ^ *(unsigned char *)(b+3); b 122 crypto/memneq.c neq |= *(unsigned char *)(a+4) ^ *(unsigned char *)(b+4); b 124 crypto/memneq.c neq |= *(unsigned char *)(a+5) ^ *(unsigned char *)(b+5); b 126 crypto/memneq.c neq |= *(unsigned char *)(a+6) ^ *(unsigned char *)(b+6); b 128 crypto/memneq.c neq |= *(unsigned char *)(a+7) ^ *(unsigned char *)(b+7); b 130 crypto/memneq.c neq |= *(unsigned char *)(a+8) ^ *(unsigned char *)(b+8); b 132 crypto/memneq.c neq |= *(unsigned char *)(a+9) ^ *(unsigned char *)(b+9); b 134 crypto/memneq.c neq |= *(unsigned char *)(a+10) ^ *(unsigned char *)(b+10); b 136 crypto/memneq.c neq |= *(unsigned char *)(a+11) ^ *(unsigned char *)(b+11); b 138 crypto/memneq.c neq |= *(unsigned char *)(a+12) ^ *(unsigned char *)(b+12); b 140 crypto/memneq.c neq |= *(unsigned char *)(a+13) ^ *(unsigned char *)(b+13); b 142 crypto/memneq.c neq |= *(unsigned char *)(a+14) ^ *(unsigned char *)(b+14); b 144 crypto/memneq.c neq |= *(unsigned char *)(a+15) ^ *(unsigned char *)(b+15); b 156 crypto/memneq.c noinline unsigned long __crypto_memneq(const void *a, const void *b, b 161 crypto/memneq.c return __crypto_memneq_16(a, b); b 163 crypto/memneq.c return __crypto_memneq_generic(a, b, size); b 22 crypto/poly1305_generic.c static inline u64 mlt(u64 a, u64 b) b 24 crypto/poly1305_generic.c return a * b; b 40 crypto/rmd128.c #define ROUND(a, b, c, d, f, k, x, s) { \ b 41 crypto/rmd128.c (a) += f((b), (c), (d)) + le32_to_cpup(&(x)) + (k); \ b 43 crypto/rmd160.c #define ROUND(a, b, c, d, e, f, k, x, s) { \ b 44 crypto/rmd160.c (a) += f((b), (c), (d)) + le32_to_cpup(&(x)) + (k); \ b 40 crypto/rmd256.c #define ROUND(a, b, c, d, f, k, x, s) { \ b 41 crypto/rmd256.c (a) += f((b), (c), (d)) + le32_to_cpup(&(x)) + (k); \ b 43 crypto/rmd320.c #define ROUND(a, b, c, d, e, f, k, x, s) { \ b 44 crypto/rmd320.c (a) += f((b), (c), (d)) + le32_to_cpup(&(x)) + (k); \ b 29 crypto/serpent_generic.c #define keyiter(a, b, c, d, i, j) \ b 30 crypto/serpent_generic.c ({ b ^= d; b ^= c; b ^= a; b ^= PHI ^ i; b = rol32(b, 11); k[j] = b; }) b 101 crypto/sha512_generic.c u64 a, b, c, d, e, f, g, h, t1, t2; b 107 crypto/sha512_generic.c a=state[0]; b=state[1]; c=state[2]; d=state[3]; b 127 crypto/sha512_generic.c t2 = e0(a) + Maj(a,b,c); d+=t1; h=t1+t2; b 129 crypto/sha512_generic.c t2 = e0(h) + Maj(h,a,b); c+=t1; g=t1+t2; b 131 crypto/sha512_generic.c t2 = e0(g) + Maj(g,h,a); b+=t1; f=t1+t2; b 132 crypto/sha512_generic.c t1 = e + e1(b) + Ch(b,c,d) + sha512_K[i+3] + W[(i & 15) + 3]; b 134 crypto/sha512_generic.c t1 = d + e1(a) + Ch(a,b,c) + sha512_K[i+4] + W[(i & 15) + 4]; b 136 crypto/sha512_generic.c t1 = c + e1(h) + Ch(h,a,b) + sha512_K[i+5] + W[(i & 15) + 5]; b 138 crypto/sha512_generic.c t1 = b + e1(g) + Ch(g,h,a) + sha512_K[i+6] + W[(i & 15) + 6]; b 139 crypto/sha512_generic.c t2 = e0(c) + Maj(c,d,e); f+=t1; b=t1+t2; b 141 crypto/sha512_generic.c t2 = e0(b) + Maj(b,c,d); e+=t1; a=t1+t2; b 144 crypto/sha512_generic.c state[0] += a; state[1] += b; state[2] += c; state[3] += d; b 148 crypto/sha512_generic.c a = b = c = d = e = f = g = h = t1 = t2 = 0; b 39 crypto/sm3_generic.c static inline u32 ff(unsigned int n, u32 a, u32 b, u32 c) b 41 crypto/sm3_generic.c return (n < 16) ? (a ^ b ^ c) : ((a & b) | (a & c) | (b & c)); b 78 crypto/sm3_generic.c u32 a, b, c, d, e, f, g, h; b 82 crypto/sm3_generic.c b = m[1]; b 96 crypto/sm3_generic.c tt1 = ff(i, a, b, c) + d + ss2 + *wt; b 103 crypto/sm3_generic.c c = rol32(b, 9); b 104 crypto/sm3_generic.c b = a; b 113 crypto/sm3_generic.c m[1] = b ^ m[1]; b 121 crypto/sm3_generic.c a = b = c = d = e = f = g = h = ss1 = ss2 = tt1 = tt2 = 0; b 72 crypto/sm4_generic.c u8 *b = (u8 *)&x; b 75 crypto/sm4_generic.c b[i] = sbox[b[i]]; b 789 crypto/testmgr.c u8 b; b 802 crypto/testmgr.c b = 0x00; b 805 crypto/testmgr.c b = 0xff; b 808 crypto/testmgr.c b = (u8)prandom_u32(); b 811 crypto/testmgr.c memset(buf, b, count); b 817 crypto/testmgr.c b = (u8)prandom_u32(); b 818 crypto/testmgr.c for (i = 0; i < count; i++, b += increment) b 819 crypto/testmgr.c buf[i] = b; b 34 crypto/tgr192.c u64 a, b, c; b 400 crypto/tgr192.c u64 b = *rb; b 406 crypto/tgr192.c b += sbox4[(c >> 8) & 0xff] ^ sbox3[(c >> 24) & 0xff] b 408 crypto/tgr192.c b *= mul; b 411 crypto/tgr192.c *rb = b; b 419 crypto/tgr192.c u64 b = *rb; b 422 crypto/tgr192.c tgr192_round(&a, &b, &c, x[0], mul); b 423 crypto/tgr192.c tgr192_round(&b, &c, &a, x[1], mul); b 424 crypto/tgr192.c tgr192_round(&c, &a, &b, x[2], mul); b 425 crypto/tgr192.c tgr192_round(&a, &b, &c, x[3], mul); b 426 crypto/tgr192.c tgr192_round(&b, &c, &a, x[4], mul); b 427 crypto/tgr192.c tgr192_round(&c, &a, &b, x[5], mul); b 428 crypto/tgr192.c tgr192_round(&a, &b, &c, x[6], mul); b 429 crypto/tgr192.c tgr192_round(&b, &c, &a, x[7], mul); b 432 crypto/tgr192.c *rb = b; b 464 crypto/tgr192.c u64 a, b, c, aa, bb, cc; b 473 crypto/tgr192.c b = bb = tctx->b; b 476 crypto/tgr192.c tgr192_pass(&a, &b, &c, x, 5); b 478 crypto/tgr192.c tgr192_pass(&c, &a, &b, x, 7); b 480 crypto/tgr192.c tgr192_pass(&b, &c, &a, x, 9); b 485 crypto/tgr192.c b -= bb; b 489 crypto/tgr192.c tctx->b = b; b 498 crypto/tgr192.c tctx->b = 0xfedcba9876543210ULL; b 598 crypto/tgr192.c dst[1] = be64p[1] = cpu_to_be64(tctx->b); b 468 crypto/twofish_common.c #define CALC_S(a, b, c, d, i, w, x, y, z) \ b 472 crypto/twofish_common.c (b) ^= exp_to_poly[tmp + (x)]; \ b 483 crypto/twofish_common.c #define CALC_SB_2(i, a, b) \ b 485 crypto/twofish_common.c ctx->s[1][i] = mds[1][q0[(b) ^ sb] ^ sf]; \ b 487 crypto/twofish_common.c ctx->s[3][i] = mds[3][q1[(b) ^ sd] ^ sh] b 491 crypto/twofish_common.c #define CALC_SB192_2(i, a, b) \ b 492 crypto/twofish_common.c ctx->s[0][i] = mds[0][q0[q0[(b) ^ sa] ^ se] ^ si]; \ b 493 crypto/twofish_common.c ctx->s[1][i] = mds[1][q0[q1[(b) ^ sb] ^ sf] ^ sj]; \ b 499 crypto/twofish_common.c #define CALC_SB256_2(i, a, b) \ b 500 crypto/twofish_common.c ctx->s[0][i] = mds[0][q0[q0[q1[(b) ^ sa] ^ se] ^ si] ^ sm]; \ b 503 crypto/twofish_common.c ctx->s[3][i] = mds[3][q1[q1[q0[(b) ^ sd] ^ sh] ^ sl] ^ sp]; b 529 crypto/twofish_common.c #define CALC_K_2(a, b, c, d, j) \ b 531 crypto/twofish_common.c ^ mds[1][q0[b ^ key[(j) + 9]] ^ key[(j) + 1]] \ b 542 crypto/twofish_common.c #define CALC_K192_2(a, b, c, d, j) \ b 544 crypto/twofish_common.c q1[b ^ key[(j) + 17]], \ b 555 crypto/twofish_common.c #define CALC_K256_2(a, b, j) \ b 556 crypto/twofish_common.c CALC_K192_2 (q1[b ^ key[(j) + 24]], \ b 559 crypto/twofish_common.c q0[b ^ key[(j) + 27]], j) b 44 crypto/twofish_generic.c #define G2(b) \ b 45 crypto/twofish_generic.c (ctx->s[1][(b) & 0xFF]) ^ (ctx->s[2][((b) >> 8) & 0xFF]) \ b 46 crypto/twofish_generic.c ^ (ctx->s[3][((b) >> 16) & 0xFF]) ^ (ctx->s[0][(b) >> 24]) b 53 crypto/twofish_generic.c #define ENCROUND(n, a, b, c, d) \ b 54 crypto/twofish_generic.c x = G1 (a); y = G2 (b); \ b 60 crypto/twofish_generic.c #define DECROUND(n, a, b, c, d) \ b 61 crypto/twofish_generic.c x = G1 (a); y = G2 (b); \ b 72 crypto/twofish_generic.c ENCROUND (2 * (n), a, b, c, d); \ b 73 crypto/twofish_generic.c ENCROUND (2 * (n) + 1, c, d, a, b) b 76 crypto/twofish_generic.c DECROUND (2 * (n) + 1, c, d, a, b); \ b 77 crypto/twofish_generic.c DECROUND (2 * (n), a, b, c, d) b 102 crypto/twofish_generic.c u32 a, b, c, d; b 109 crypto/twofish_generic.c INPACK (1, b, 1); b 127 crypto/twofish_generic.c OUTUNPACK (3, b, 7); b 139 crypto/twofish_generic.c u32 a, b, c, d; b 148 crypto/twofish_generic.c INPACK (3, b, 7); b 162 crypto/twofish_generic.c OUTUNPACK (1, b, 1); b 1012 crypto/wp512.c u32 b, carry; b 1026 crypto/wp512.c b = ((source[sourcePos] << sourceGap) & 0xff) | b 1028 crypto/wp512.c buffer[bufferPos++] |= (u8)(b >> bufferRem); b 1034 crypto/wp512.c buffer[bufferPos] = b << (8 - bufferRem); b 1040 crypto/wp512.c b = (source[sourcePos] << sourceGap) & 0xff; b 1041 crypto/wp512.c buffer[bufferPos] |= b >> bufferRem; b 1043 crypto/wp512.c b = 0; b 1055 crypto/wp512.c buffer[bufferPos] = b << (8 - bufferRem); b 150 crypto/xts.c le128 b; b 155 crypto/xts.c scatterwalk_map_and_copy(&b, rctx->tail, 0, XTS_BLOCK_SIZE, 0); b 156 crypto/xts.c le128_xor(&b, &rctx->t, &b); b 157 crypto/xts.c scatterwalk_map_and_copy(&b, rctx->tail, 0, XTS_BLOCK_SIZE, 1); b 171 crypto/xts.c le128 b[2]; b 177 crypto/xts.c scatterwalk_map_and_copy(b, rctx->tail, 0, XTS_BLOCK_SIZE, 0); b 178 crypto/xts.c memcpy(b + 1, b, tail); b 179 crypto/xts.c scatterwalk_map_and_copy(b, req->src, offset, tail, 0); b 181 crypto/xts.c le128_xor(b, &rctx->t, b); b 183 crypto/xts.c scatterwalk_map_and_copy(b, rctx->tail, 0, XTS_BLOCK_SIZE + tail, 1); b 194 crypto/xts.c scatterwalk_map_and_copy(b, rctx->tail, 0, XTS_BLOCK_SIZE, 0); b 195 crypto/xts.c le128_xor(b, &rctx->t, b); b 196 crypto/xts.c scatterwalk_map_and_copy(b, rctx->tail, 0, XTS_BLOCK_SIZE, 1); b 733 drivers/acpi/acpi_video.c acpi_video_cmp_level(const void *a, const void *b) b 735 drivers/acpi/acpi_video.c return *(int *)a - *(int *)b; b 399 drivers/acpi/acpica/acmacros.h #define ARGI_LIST2(a, b) (ARG_1(b)|ARG_2(a)) b 400 drivers/acpi/acpica/acmacros.h #define ARGI_LIST3(a, b, c) (ARG_1(c)|ARG_2(b)|ARG_3(a)) b 401 drivers/acpi/acpica/acmacros.h #define ARGI_LIST4(a, b, c, d) (ARG_1(d)|ARG_2(c)|ARG_3(b)|ARG_4(a)) b 402 drivers/acpi/acpica/acmacros.h #define ARGI_LIST5(a, b, c, d, e) (ARG_1(e)|ARG_2(d)|ARG_3(c)|ARG_4(b)|ARG_5(a)) b 403 drivers/acpi/acpica/acmacros.h #define ARGI_LIST6(a, b, c, d, e, f) (ARG_1(f)|ARG_2(e)|ARG_3(d)|ARG_4(c)|ARG_5(b)|ARG_6(a)) b 406 drivers/acpi/acpica/acmacros.h #define ARGP_LIST2(a, b) (ARG_1(a)|ARG_2(b)) b 407 drivers/acpi/acpica/acmacros.h #define ARGP_LIST3(a, b, c) (ARG_1(a)|ARG_2(b)|ARG_3(c)) b 408 drivers/acpi/acpica/acmacros.h #define ARGP_LIST4(a, b, c, d) (ARG_1(a)|ARG_2(b)|ARG_3(c)|ARG_4(d)) b 409 drivers/acpi/acpica/acmacros.h #define ARGP_LIST5(a, b, c, d, e) (ARG_1(a)|ARG_2(b)|ARG_3(c)|ARG_4(d)|ARG_5(e)) b 410 drivers/acpi/acpica/acmacros.h #define ARGP_LIST6(a, b, c, d, e, f) (ARG_1(a)|ARG_2(b)|ARG_3(c)|ARG_4(d)|ARG_5(e)|ARG_6(f)) b 456 drivers/acpi/acpica/acmacros.h #define ACPI_INIT_UUID(a, b, c, d0, d1, d2, d3, d4, d5, d6, d7) \ b 458 drivers/acpi/acpica/acmacros.h (b) & 0xFF, ((b) >> 8) & 0xFF, \ b 473 drivers/acpi/acpica/acmacros.h #define ASL_CV_CLOSE_PAREN(a,b) cv_close_paren_write_comment(a,b); b 474 drivers/acpi/acpica/acmacros.h #define ASL_CV_CLOSE_BRACE(a,b) cv_close_brace_write_comment(a,b); b 475 drivers/acpi/acpica/acmacros.h #define ASL_CV_SWITCH_FILES(a,b) cv_switch_files(a,b); b 477 drivers/acpi/acpica/acmacros.h #define ASL_CV_PRINT_ONE_COMMENT(a,b,c,d) cv_print_one_comment_type (a,b,c,d); b 478 drivers/acpi/acpica/acmacros.h #define ASL_CV_PRINT_ONE_COMMENT_LIST(a,b) cv_print_one_comment_list (a,b); b 480 drivers/acpi/acpica/acmacros.h #define ASL_CV_INIT_FILETREE(a,b,c) cv_init_file_tree(a,b,c); b 488 drivers/acpi/acpica/acmacros.h #define ASL_CV_CLOSE_PAREN(a,b) acpi_os_printf (")"); b 489 drivers/acpi/acpica/acmacros.h #define ASL_CV_CLOSE_BRACE(a,b) acpi_os_printf ("}"); b 490 drivers/acpi/acpica/acmacros.h #define ASL_CV_SWITCH_FILES(a,b) b 492 drivers/acpi/acpica/acmacros.h #define ASL_CV_PRINT_ONE_COMMENT(a,b,c,d) b 493 drivers/acpi/acpica/acmacros.h #define ASL_CV_PRINT_ONE_COMMENT_LIST(a,b) b 495 drivers/acpi/acpica/acmacros.h #define ASL_CV_INIT_FILETREE(a,b,c) b 124 drivers/acpi/acpica/acpredef.h #define PACKAGE_INFO(a,b,c,d,e,f) {{{(a),(b),(c),(d)}, ((((u16)(f)) << 8) | (e)), 0}} b 261 drivers/acpi/fan.c static int acpi_fan_speed_cmp(const void *a, const void *b) b 264 drivers/acpi/fan.c const struct acpi_fan_fps *fps2 = b; b 526 drivers/acpi/hmat/hmat.c static int initiator_cmp(void *priv, struct list_head *a, struct list_head *b) b 533 drivers/acpi/hmat/hmat.c ib = list_entry(b, struct memory_initiator, node); b 1190 drivers/acpi/nfit/core.c struct nfit_mem *b = container_of(_b, typeof(*b), list); b 1194 drivers/acpi/nfit/core.c handleB = __to_nfit_memdev(b)->device_handle; b 30 drivers/acpi/nvs.c static inline int suspend_nvs_register(unsigned long a, unsigned long b) b 2096 drivers/android/binder.c struct binder_buffer *b, b 2112 drivers/android/binder.c b, buffer_offset, b 2115 drivers/android/binder.c object_size = binder_get_object(proc, b, object_offset, object); b 2164 drivers/android/binder.c struct binder_buffer *b, b 2180 drivers/android/binder.c size_t object_size = binder_get_object(proc, b, last_obj_offset, b 2197 drivers/android/binder.c b, buffer_offset, b 2689 drivers/android/binder.c struct binder_buffer *b = t->buffer; b 2699 drivers/android/binder.c parent = binder_validate_ptr(target_proc, b, &object, bp->parent, b 2708 drivers/android/binder.c if (!binder_validate_fixup(target_proc, b, off_start_offset, b 2725 drivers/android/binder.c (uintptr_t)parent->buffer - (uintptr_t)b->user_data; b 2726 drivers/android/binder.c if (binder_alloc_copy_to_buffer(&target_proc->alloc, b, buffer_offset, b 3238 drivers/ata/libata-core.c void ata_timing_merge(const struct ata_timing *a, const struct ata_timing *b, b 3241 drivers/ata/libata-core.c if (what & ATA_TIMING_SETUP ) m->setup = max(a->setup, b->setup); b 3242 drivers/ata/libata-core.c if (what & ATA_TIMING_ACT8B ) m->act8b = max(a->act8b, b->act8b); b 3243 drivers/ata/libata-core.c if (what & ATA_TIMING_REC8B ) m->rec8b = max(a->rec8b, b->rec8b); b 3244 drivers/ata/libata-core.c if (what & ATA_TIMING_CYC8B ) m->cyc8b = max(a->cyc8b, b->cyc8b); b 3245 drivers/ata/libata-core.c if (what & ATA_TIMING_ACTIVE ) m->active = max(a->active, b->active); b 3246 drivers/ata/libata-core.c if (what & ATA_TIMING_RECOVER) m->recover = max(a->recover, b->recover); b 3247 drivers/ata/libata-core.c if (what & ATA_TIMING_DMACK_HOLD) m->dmack_hold = max(a->dmack_hold, b->dmack_hold); b 3248 drivers/ata/libata-core.c if (what & ATA_TIMING_CYCLE ) m->cycle = max(a->cycle, b->cycle); b 3249 drivers/ata/libata-core.c if (what & ATA_TIMING_UDMA ) m->udma = max(a->udma, b->udma); b 112 drivers/ata/pata_atp867x.c u8 b; b 127 drivers/ata/pata_atp867x.c b = ioread8(dp->dma_mode); b 129 drivers/ata/pata_atp867x.c b = (b & ~ATP867X_IO_DMAMODE_SLAVE_MASK) | b 132 drivers/ata/pata_atp867x.c b = (b & ~ATP867X_IO_DMAMODE_MSTR_MASK) | b 135 drivers/ata/pata_atp867x.c iowrite8(b, dp->dma_mode); b 209 drivers/ata/pata_atp867x.c u8 b; b 220 drivers/ata/pata_atp867x.c b = ioread8(dp->dma_mode); b 222 drivers/ata/pata_atp867x.c b = (b & ~ATP867X_IO_DMAMODE_SLAVE_MASK); b 224 drivers/ata/pata_atp867x.c b = (b & ~ATP867X_IO_DMAMODE_MSTR_MASK); b 225 drivers/ata/pata_atp867x.c iowrite8(b, dp->dma_mode); b 227 drivers/ata/pata_atp867x.c b = atp867x_get_active_clocks_shifted(ap, t.active) | b 231 drivers/ata/pata_atp867x.c iowrite8(b, dp->slave_piospd); b 233 drivers/ata/pata_atp867x.c iowrite8(b, dp->mstr_piospd); b 235 drivers/ata/pata_atp867x.c b = atp867x_get_active_clocks_shifted(ap, t.act8b) | b 238 drivers/ata/pata_atp867x.c iowrite8(b, dp->eightb_piospd); b 1251 drivers/ata/sata_mv.c int b, w; b 1252 drivers/ata/sata_mv.c for (b = 0; b < bytes; ) { b 1253 drivers/ata/sata_mv.c DPRINTK("%p: ", start + b); b 1254 drivers/ata/sata_mv.c for (w = 0; b < bytes && w < 4; w++) { b 1255 drivers/ata/sata_mv.c printk("%08x ", readl(start + b)); b 1256 drivers/ata/sata_mv.c b += sizeof(u32); b 1266 drivers/ata/sata_mv.c int b, w; b 1268 drivers/ata/sata_mv.c for (b = 0; b < bytes; ) { b 1269 drivers/ata/sata_mv.c DPRINTK("%02x: ", b); b 1270 drivers/ata/sata_mv.c for (w = 0; b < bytes && w < 4; w++) { b 1271 drivers/ata/sata_mv.c (void) pci_read_config_dword(pdev, b, &dw); b 1273 drivers/ata/sata_mv.c b += sizeof(u32); b 83 drivers/atm/eni.c #define EVENT(s,a,b) b 114 drivers/atm/eni.c static void EVENT(const char *s,unsigned long a,unsigned long b) b 118 drivers/atm/eni.c ev_b[ec] = b; b 145 drivers/atm/eni.c #define NEPJOK(a0,a1,b) \ b 146 drivers/atm/eni.c ((a0) < (a1) ? (b) <= (a0) || (b) > (a1) : (b) <= (a0) && (b) > (a1)) b 147 drivers/atm/eni.c #define EEPJOK(a0,a1,b) \ b 148 drivers/atm/eni.c ((a0) < (a1) ? (b) < (a0) || (b) >= (a1) : (b) < (a0) && (b) >= (a1)) b 149 drivers/atm/eni.c #define NEPMOK(a0,d,b,c) NEPJOK(a0,(a0+d) & (c-1),b) b 150 drivers/atm/eni.c #define EEPMOK(a0,d,b,c) EEPJOK(a0,(a0+d) & (c-1),b) b 69 drivers/atm/firestream.h #define Q_SA(b) (b + QSA ) b 70 drivers/atm/firestream.h #define Q_EA(b) (b + QEA ) b 71 drivers/atm/firestream.h #define Q_RP(b) (b + QRP ) b 72 drivers/atm/firestream.h #define Q_WP(b) (b + QWP ) b 73 drivers/atm/firestream.h #define Q_CNF(b) (b + QCNF) b 75 drivers/atm/firestream.h #define FP_CNF(b) (b + FPCNF) b 76 drivers/atm/firestream.h #define FP_SA(b) (b + FPSA) b 77 drivers/atm/firestream.h #define FP_EA(b) (b + FPEA) b 78 drivers/atm/firestream.h #define FP_CNT(b) (b + FPCNT) b 79 drivers/atm/firestream.h #define FP_CTU(b) (b + FPCTU) b 2000 drivers/atm/horizon.c u16 b = 0; b 2012 drivers/atm/horizon.c b = read_bia (dev, i/2 + 2); b 2014 drivers/atm/horizon.c b = b >> 8; b 2015 drivers/atm/horizon.c esi[i] = b & 0xFF; b 109 drivers/atm/nicstar.c #define PTR_DIFF(a, b) ((u32)((unsigned long)(a) - (unsigned long)(b))) b 73 drivers/atm/solos-pci.c #define FPGA_VERSION(a,b) (((a) << 8) + (b)) b 64 drivers/atm/zatm.c #define EVENT(s,a,b) b 94 drivers/atm/zatm.c static void EVENT(const char *s,unsigned long a,unsigned long b) b 98 drivers/atm/zatm.c ev_b[ec] = b; b 260 drivers/auxdisplay/cfag12864b.c unsigned short i, j, k, b; b 272 drivers/auxdisplay/cfag12864b.c for (c = 0, b = 0; b < 8; b++) b 275 drivers/auxdisplay/cfag12864b.c + k / 8 + (j * 8 + b) * b 278 drivers/auxdisplay/cfag12864b.c c |= bit(b); b 933 drivers/base/bus.c const struct device *b)) b 937 drivers/base/bus.c struct device *b; b 941 drivers/base/bus.c b = dev_prv->device; b 942 drivers/base/bus.c if (compare(a, b) <= 0) { b 944 drivers/base/bus.c &b->p->knode_bus.n_node); b 953 drivers/base/bus.c const struct device *b)) b 635 drivers/base/regmap/regcache.c static int regcache_default_cmp(const void *a, const void *b) b 638 drivers/base/regmap/regcache.c const struct reg_default *_b = b; b 330 drivers/base/regmap/regmap-irq.c unsigned int b) b 340 drivers/base/regmap/regmap-irq.c (b * map->reg_stride * data->irq_reg_stride), b 341 drivers/base/regmap/regmap-irq.c &data->status_buf[b]); b 343 drivers/base/regmap/regmap-irq.c subreg = &chip->sub_reg_offsets[b]; b 414 drivers/base/regmap/regmap-irq.c unsigned int b; b 417 drivers/base/regmap/regmap-irq.c for_each_set_bit(b, &mreg, map->format.val_bytes * 8) { b 418 drivers/base/regmap/regmap-irq.c if (i * map->format.val_bytes * 8 + b > b 421 drivers/base/regmap/regmap-irq.c ret = read_sub_irq_data(data, b); b 245 drivers/base/regmap/regmap.c u8 *b = buf; b 247 drivers/base/regmap/regmap.c b[0] = val << shift; b 252 drivers/base/regmap/regmap.c __be16 *b = buf; b 254 drivers/base/regmap/regmap.c b[0] = cpu_to_be16(val << shift); b 259 drivers/base/regmap/regmap.c __le16 *b = buf; b 261 drivers/base/regmap/regmap.c b[0] = cpu_to_le16(val << shift); b 272 drivers/base/regmap/regmap.c u8 *b = buf; b 276 drivers/base/regmap/regmap.c b[0] = val >> 16; b 277 drivers/base/regmap/regmap.c b[1] = val >> 8; b 278 drivers/base/regmap/regmap.c b[2] = val; b 283 drivers/base/regmap/regmap.c __be32 *b = buf; b 285 drivers/base/regmap/regmap.c b[0] = cpu_to_be32(val << shift); b 290 drivers/base/regmap/regmap.c __le32 *b = buf; b 292 drivers/base/regmap/regmap.c b[0] = cpu_to_le32(val << shift); b 304 drivers/base/regmap/regmap.c __be64 *b = buf; b 306 drivers/base/regmap/regmap.c b[0] = cpu_to_be64((u64)val << shift); b 311 drivers/base/regmap/regmap.c __le64 *b = buf; b 313 drivers/base/regmap/regmap.c b[0] = cpu_to_le64((u64)val << shift); b 329 drivers/base/regmap/regmap.c const u8 *b = buf; b 331 drivers/base/regmap/regmap.c return b[0]; b 336 drivers/base/regmap/regmap.c const __be16 *b = buf; b 338 drivers/base/regmap/regmap.c return be16_to_cpu(b[0]); b 343 drivers/base/regmap/regmap.c const __le16 *b = buf; b 345 drivers/base/regmap/regmap.c return le16_to_cpu(b[0]); b 350 drivers/base/regmap/regmap.c __be16 *b = buf; b 352 drivers/base/regmap/regmap.c b[0] = be16_to_cpu(b[0]); b 357 drivers/base/regmap/regmap.c __le16 *b = buf; b 359 drivers/base/regmap/regmap.c b[0] = le16_to_cpu(b[0]); b 369 drivers/base/regmap/regmap.c const u8 *b = buf; b 370 drivers/base/regmap/regmap.c unsigned int ret = b[2]; b 371 drivers/base/regmap/regmap.c ret |= ((unsigned int)b[1]) << 8; b 372 drivers/base/regmap/regmap.c ret |= ((unsigned int)b[0]) << 16; b 379 drivers/base/regmap/regmap.c const __be32 *b = buf; b 381 drivers/base/regmap/regmap.c return be32_to_cpu(b[0]); b 386 drivers/base/regmap/regmap.c const __le32 *b = buf; b 388 drivers/base/regmap/regmap.c return le32_to_cpu(b[0]); b 393 drivers/base/regmap/regmap.c __be32 *b = buf; b 395 drivers/base/regmap/regmap.c b[0] = be32_to_cpu(b[0]); b 400 drivers/base/regmap/regmap.c __le32 *b = buf; b 402 drivers/base/regmap/regmap.c b[0] = le32_to_cpu(b[0]); b 413 drivers/base/regmap/regmap.c const __be64 *b = buf; b 415 drivers/base/regmap/regmap.c return be64_to_cpu(b[0]); b 420 drivers/base/regmap/regmap.c const __le64 *b = buf; b 422 drivers/base/regmap/regmap.c return le64_to_cpu(b[0]); b 427 drivers/base/regmap/regmap.c __be64 *b = buf; b 429 drivers/base/regmap/regmap.c b[0] = be64_to_cpu(b[0]); b 434 drivers/base/regmap/regmap.c __le64 *b = buf; b 436 drivers/base/regmap/regmap.c b[0] = le64_to_cpu(b[0]); b 115 drivers/block/drbd/drbd_bitmap.c struct drbd_bitmap *b = device->bitmap; b 120 drivers/block/drbd/drbd_bitmap.c func, b->bm_why ?: "?", b 121 drivers/block/drbd/drbd_bitmap.c b->bm_task->comm, task_pid_nr(b->bm_task)); b 126 drivers/block/drbd/drbd_bitmap.c struct drbd_bitmap *b = device->bitmap; b 129 drivers/block/drbd/drbd_bitmap.c if (!b) { b 134 drivers/block/drbd/drbd_bitmap.c trylock_failed = !mutex_trylock(&b->bm_change); b 139 drivers/block/drbd/drbd_bitmap.c why, b->bm_why ?: "?", b 140 drivers/block/drbd/drbd_bitmap.c b->bm_task->comm, task_pid_nr(b->bm_task)); b 141 drivers/block/drbd/drbd_bitmap.c mutex_lock(&b->bm_change); b 143 drivers/block/drbd/drbd_bitmap.c if (BM_LOCKED_MASK & b->bm_flags) b 145 drivers/block/drbd/drbd_bitmap.c b->bm_flags |= flags & BM_LOCKED_MASK; b 147 drivers/block/drbd/drbd_bitmap.c b->bm_why = why; b 148 drivers/block/drbd/drbd_bitmap.c b->bm_task = current; b 153 drivers/block/drbd/drbd_bitmap.c struct drbd_bitmap *b = device->bitmap; b 154 drivers/block/drbd/drbd_bitmap.c if (!b) { b 162 drivers/block/drbd/drbd_bitmap.c b->bm_flags &= ~BM_LOCKED_MASK; b 163 drivers/block/drbd/drbd_bitmap.c b->bm_why = NULL; b 164 drivers/block/drbd/drbd_bitmap.c b->bm_task = NULL; b 165 drivers/block/drbd/drbd_bitmap.c mutex_unlock(&b->bm_change); b 213 drivers/block/drbd/drbd_bitmap.c struct drbd_bitmap *b = device->bitmap; b 214 drivers/block/drbd/drbd_bitmap.c void *addr = &page_private(b->bm_pages[page_nr]); b 215 drivers/block/drbd/drbd_bitmap.c wait_event(b->bm_io_wait, !test_and_set_bit(BM_PAGE_IO_LOCK, addr)); b 220 drivers/block/drbd/drbd_bitmap.c struct drbd_bitmap *b = device->bitmap; b 221 drivers/block/drbd/drbd_bitmap.c void *addr = &page_private(b->bm_pages[page_nr]); b 256 drivers/block/drbd/drbd_bitmap.c struct drbd_bitmap *b = device->bitmap; b 264 drivers/block/drbd/drbd_bitmap.c BUG_ON(b->n_bitmap_hints >= ARRAY_SIZE(b->al_bitmap_hints)); b 266 drivers/block/drbd/drbd_bitmap.c b->al_bitmap_hints[b->n_bitmap_hints++] = page_nr; b 296 drivers/block/drbd/drbd_bitmap.c static unsigned int bm_word_to_page_idx(struct drbd_bitmap *b, unsigned long long_nr) b 300 drivers/block/drbd/drbd_bitmap.c BUG_ON(page_nr >= b->bm_number_of_pages); b 304 drivers/block/drbd/drbd_bitmap.c static unsigned int bm_bit_to_page_idx(struct drbd_bitmap *b, u64 bitnr) b 308 drivers/block/drbd/drbd_bitmap.c BUG_ON(page_nr >= b->bm_number_of_pages); b 312 drivers/block/drbd/drbd_bitmap.c static unsigned long *__bm_map_pidx(struct drbd_bitmap *b, unsigned int idx) b 314 drivers/block/drbd/drbd_bitmap.c struct page *page = b->bm_pages[idx]; b 318 drivers/block/drbd/drbd_bitmap.c static unsigned long *bm_map_pidx(struct drbd_bitmap *b, unsigned int idx) b 320 drivers/block/drbd/drbd_bitmap.c return __bm_map_pidx(b, idx); b 378 drivers/block/drbd/drbd_bitmap.c static struct page **bm_realloc_pages(struct drbd_bitmap *b, unsigned long want) b 380 drivers/block/drbd/drbd_bitmap.c struct page **old_pages = b->bm_pages; b 383 drivers/block/drbd/drbd_bitmap.c unsigned long have = b->bm_number_of_pages; b 437 drivers/block/drbd/drbd_bitmap.c struct drbd_bitmap *b = device->bitmap; b 438 drivers/block/drbd/drbd_bitmap.c WARN_ON(b != NULL); b 439 drivers/block/drbd/drbd_bitmap.c b = kzalloc(sizeof(struct drbd_bitmap), GFP_KERNEL); b 440 drivers/block/drbd/drbd_bitmap.c if (!b) b 442 drivers/block/drbd/drbd_bitmap.c spin_lock_init(&b->bm_lock); b 443 drivers/block/drbd/drbd_bitmap.c mutex_init(&b->bm_change); b 444 drivers/block/drbd/drbd_bitmap.c init_waitqueue_head(&b->bm_io_wait); b 446 drivers/block/drbd/drbd_bitmap.c device->bitmap = b; b 484 drivers/block/drbd/drbd_bitmap.c static int bm_clear_surplus(struct drbd_bitmap *b) b 492 drivers/block/drbd/drbd_bitmap.c tmp = (b->bm_bits & BITS_PER_PAGE_MASK); b 499 drivers/block/drbd/drbd_bitmap.c p_addr = bm_map_pidx(b, b->bm_number_of_pages - 1); b 521 drivers/block/drbd/drbd_bitmap.c static void bm_set_surplus(struct drbd_bitmap *b) b 528 drivers/block/drbd/drbd_bitmap.c tmp = (b->bm_bits & BITS_PER_PAGE_MASK); b 535 drivers/block/drbd/drbd_bitmap.c p_addr = bm_map_pidx(b, b->bm_number_of_pages - 1); b 556 drivers/block/drbd/drbd_bitmap.c static unsigned long bm_count_bits(struct drbd_bitmap *b) b 560 drivers/block/drbd/drbd_bitmap.c unsigned long mask = (1UL << (b->bm_bits & BITS_PER_LONG_MASK)) -1; b 564 drivers/block/drbd/drbd_bitmap.c for (idx = 0; idx < b->bm_number_of_pages - 1; idx++) { b 565 drivers/block/drbd/drbd_bitmap.c p_addr = __bm_map_pidx(b, idx); b 571 drivers/block/drbd/drbd_bitmap.c last_word = ((b->bm_bits - 1) & BITS_PER_PAGE_MASK) >> LN2_BPL; b 572 drivers/block/drbd/drbd_bitmap.c p_addr = __bm_map_pidx(b, idx); b 584 drivers/block/drbd/drbd_bitmap.c static void bm_memset(struct drbd_bitmap *b, size_t offset, int c, size_t len) b 592 drivers/block/drbd/drbd_bitmap.c if (end > b->bm_words) { b 599 drivers/block/drbd/drbd_bitmap.c idx = bm_word_to_page_idx(b, offset); b 600 drivers/block/drbd/drbd_bitmap.c p_addr = bm_map_pidx(b, idx); b 608 drivers/block/drbd/drbd_bitmap.c bm_set_page_need_writeout(b->bm_pages[idx]); b 634 drivers/block/drbd/drbd_bitmap.c struct drbd_bitmap *b = device->bitmap; b 641 drivers/block/drbd/drbd_bitmap.c if (!expect(b)) b 649 drivers/block/drbd/drbd_bitmap.c if (capacity == b->bm_dev_capacity) b 653 drivers/block/drbd/drbd_bitmap.c spin_lock_irq(&b->bm_lock); b 654 drivers/block/drbd/drbd_bitmap.c opages = b->bm_pages; b 655 drivers/block/drbd/drbd_bitmap.c onpages = b->bm_number_of_pages; b 656 drivers/block/drbd/drbd_bitmap.c owords = b->bm_words; b 657 drivers/block/drbd/drbd_bitmap.c b->bm_pages = NULL; b 658 drivers/block/drbd/drbd_bitmap.c b->bm_number_of_pages = b 659 drivers/block/drbd/drbd_bitmap.c b->bm_set = b 660 drivers/block/drbd/drbd_bitmap.c b->bm_bits = b 661 drivers/block/drbd/drbd_bitmap.c b->bm_words = b 662 drivers/block/drbd/drbd_bitmap.c b->bm_dev_capacity = 0; b 663 drivers/block/drbd/drbd_bitmap.c spin_unlock_irq(&b->bm_lock); b 689 drivers/block/drbd/drbd_bitmap.c have = b->bm_number_of_pages; b 691 drivers/block/drbd/drbd_bitmap.c D_ASSERT(device, b->bm_pages != NULL); b 692 drivers/block/drbd/drbd_bitmap.c npages = b->bm_pages; b 697 drivers/block/drbd/drbd_bitmap.c npages = bm_realloc_pages(b, want); b 705 drivers/block/drbd/drbd_bitmap.c spin_lock_irq(&b->bm_lock); b 706 drivers/block/drbd/drbd_bitmap.c opages = b->bm_pages; b 707 drivers/block/drbd/drbd_bitmap.c owords = b->bm_words; b 708 drivers/block/drbd/drbd_bitmap.c obits = b->bm_bits; b 712 drivers/block/drbd/drbd_bitmap.c bm_set_surplus(b); b 714 drivers/block/drbd/drbd_bitmap.c b->bm_pages = npages; b 715 drivers/block/drbd/drbd_bitmap.c b->bm_number_of_pages = want; b 716 drivers/block/drbd/drbd_bitmap.c b->bm_bits = bits; b 717 drivers/block/drbd/drbd_bitmap.c b->bm_words = words; b 718 drivers/block/drbd/drbd_bitmap.c b->bm_dev_capacity = capacity; b 722 drivers/block/drbd/drbd_bitmap.c bm_memset(b, owords, 0xff, words-owords); b 723 drivers/block/drbd/drbd_bitmap.c b->bm_set += bits - obits; b 725 drivers/block/drbd/drbd_bitmap.c bm_memset(b, owords, 0x00, words-owords); b 734 drivers/block/drbd/drbd_bitmap.c (void)bm_clear_surplus(b); b 736 drivers/block/drbd/drbd_bitmap.c spin_unlock_irq(&b->bm_lock); b 740 drivers/block/drbd/drbd_bitmap.c b->bm_set = bm_count_bits(b); b 758 drivers/block/drbd/drbd_bitmap.c struct drbd_bitmap *b = device->bitmap; b 762 drivers/block/drbd/drbd_bitmap.c if (!expect(b)) b 764 drivers/block/drbd/drbd_bitmap.c if (!expect(b->bm_pages)) b 767 drivers/block/drbd/drbd_bitmap.c spin_lock_irqsave(&b->bm_lock, flags); b 768 drivers/block/drbd/drbd_bitmap.c s = b->bm_set; b 769 drivers/block/drbd/drbd_bitmap.c spin_unlock_irqrestore(&b->bm_lock, flags); b 787 drivers/block/drbd/drbd_bitmap.c struct drbd_bitmap *b = device->bitmap; b 788 drivers/block/drbd/drbd_bitmap.c if (!expect(b)) b 790 drivers/block/drbd/drbd_bitmap.c if (!expect(b->bm_pages)) b 793 drivers/block/drbd/drbd_bitmap.c return b->bm_words; b 798 drivers/block/drbd/drbd_bitmap.c struct drbd_bitmap *b = device->bitmap; b 799 drivers/block/drbd/drbd_bitmap.c if (!expect(b)) b 802 drivers/block/drbd/drbd_bitmap.c return b->bm_bits; b 813 drivers/block/drbd/drbd_bitmap.c struct drbd_bitmap *b = device->bitmap; b 821 drivers/block/drbd/drbd_bitmap.c if (!expect(b)) b 823 drivers/block/drbd/drbd_bitmap.c if (!expect(b->bm_pages)) b 827 drivers/block/drbd/drbd_bitmap.c WARN_ON(offset >= b->bm_words); b 828 drivers/block/drbd/drbd_bitmap.c WARN_ON(end > b->bm_words); b 830 drivers/block/drbd/drbd_bitmap.c spin_lock_irq(&b->bm_lock); b 833 drivers/block/drbd/drbd_bitmap.c idx = bm_word_to_page_idx(b, offset); b 834 drivers/block/drbd/drbd_bitmap.c p_addr = bm_map_pidx(b, idx); b 841 drivers/block/drbd/drbd_bitmap.c b->bm_set += hweight_long(word) - bits; b 844 drivers/block/drbd/drbd_bitmap.c bm_set_page_need_writeout(b->bm_pages[idx]); b 851 drivers/block/drbd/drbd_bitmap.c if (end == b->bm_words) b 852 drivers/block/drbd/drbd_bitmap.c b->bm_set -= bm_clear_surplus(b); b 853 drivers/block/drbd/drbd_bitmap.c spin_unlock_irq(&b->bm_lock); b 862 drivers/block/drbd/drbd_bitmap.c struct drbd_bitmap *b = device->bitmap; b 868 drivers/block/drbd/drbd_bitmap.c if (!expect(b)) b 870 drivers/block/drbd/drbd_bitmap.c if (!expect(b->bm_pages)) b 873 drivers/block/drbd/drbd_bitmap.c spin_lock_irq(&b->bm_lock); b 874 drivers/block/drbd/drbd_bitmap.c if ((offset >= b->bm_words) || b 875 drivers/block/drbd/drbd_bitmap.c (end > b->bm_words) || b 880 drivers/block/drbd/drbd_bitmap.c (unsigned long) b->bm_words); b 884 drivers/block/drbd/drbd_bitmap.c p_addr = bm_map_pidx(b, bm_word_to_page_idx(b, offset)); b 892 drivers/block/drbd/drbd_bitmap.c spin_unlock_irq(&b->bm_lock); b 898 drivers/block/drbd/drbd_bitmap.c struct drbd_bitmap *b = device->bitmap; b 899 drivers/block/drbd/drbd_bitmap.c if (!expect(b)) b 901 drivers/block/drbd/drbd_bitmap.c if (!expect(b->bm_pages)) b 904 drivers/block/drbd/drbd_bitmap.c spin_lock_irq(&b->bm_lock); b 905 drivers/block/drbd/drbd_bitmap.c bm_memset(b, 0, 0xff, b->bm_words); b 906 drivers/block/drbd/drbd_bitmap.c (void)bm_clear_surplus(b); b 907 drivers/block/drbd/drbd_bitmap.c b->bm_set = b->bm_bits; b 908 drivers/block/drbd/drbd_bitmap.c spin_unlock_irq(&b->bm_lock); b 914 drivers/block/drbd/drbd_bitmap.c struct drbd_bitmap *b = device->bitmap; b 915 drivers/block/drbd/drbd_bitmap.c if (!expect(b)) b 917 drivers/block/drbd/drbd_bitmap.c if (!expect(b->bm_pages)) b 920 drivers/block/drbd/drbd_bitmap.c spin_lock_irq(&b->bm_lock); b 921 drivers/block/drbd/drbd_bitmap.c bm_memset(b, 0, 0, b->bm_words); b 922 drivers/block/drbd/drbd_bitmap.c b->bm_set = 0; b 923 drivers/block/drbd/drbd_bitmap.c spin_unlock_irq(&b->bm_lock); b 943 drivers/block/drbd/drbd_bitmap.c struct drbd_bitmap *b = device->bitmap; b 947 drivers/block/drbd/drbd_bitmap.c !bm_test_page_unchanged(b->bm_pages[idx])) b 954 drivers/block/drbd/drbd_bitmap.c bm_set_page_io_err(b->bm_pages[idx]); b 961 drivers/block/drbd/drbd_bitmap.c bm_clear_page_io_err(b->bm_pages[idx]); b 983 drivers/block/drbd/drbd_bitmap.c struct drbd_bitmap *b = device->bitmap; b 1002 drivers/block/drbd/drbd_bitmap.c bm_set_page_unchanged(b->bm_pages[page_nr]); b 1007 drivers/block/drbd/drbd_bitmap.c copy_highpage(page, b->bm_pages[page_nr]); b 1010 drivers/block/drbd/drbd_bitmap.c page = b->bm_pages[page_nr]; b 1036 drivers/block/drbd/drbd_bitmap.c struct drbd_bitmap *b = device->bitmap; b 1074 drivers/block/drbd/drbd_bitmap.c WARN_ON(!(BM_LOCKED_MASK & b->bm_flags)); b 1080 drivers/block/drbd/drbd_bitmap.c num_pages = b->bm_number_of_pages; b 1096 drivers/block/drbd/drbd_bitmap.c for (hint = 0; hint < b->n_bitmap_hints; hint++) { b 1097 drivers/block/drbd/drbd_bitmap.c i = b->al_bitmap_hints[hint]; b 1102 drivers/block/drbd/drbd_bitmap.c &page_private(b->bm_pages[i]))) b 1105 drivers/block/drbd/drbd_bitmap.c if (bm_test_page_unchanged(b->bm_pages[i])) b 1117 drivers/block/drbd/drbd_bitmap.c bm_test_page_unchanged(b->bm_pages[i])) { b 1124 drivers/block/drbd/drbd_bitmap.c !bm_test_page_lazy_writeout(b->bm_pages[i])) { b 1169 drivers/block/drbd/drbd_bitmap.c b->bm_set = bm_count_bits(b); b 1173 drivers/block/drbd/drbd_bitmap.c now = b->bm_set; b 1260 drivers/block/drbd/drbd_bitmap.c struct drbd_bitmap *b = device->bitmap; b 1266 drivers/block/drbd/drbd_bitmap.c if (bm_fo > b->bm_bits) { b 1267 drivers/block/drbd/drbd_bitmap.c drbd_err(device, "bm_fo=%lu bm_bits=%lu\n", bm_fo, b->bm_bits); b 1270 drivers/block/drbd/drbd_bitmap.c while (bm_fo < b->bm_bits) { b 1273 drivers/block/drbd/drbd_bitmap.c p_addr = __bm_map_pidx(b, bm_bit_to_page_idx(b, bm_fo)); b 1285 drivers/block/drbd/drbd_bitmap.c if (bm_fo >= b->bm_bits) b 1300 drivers/block/drbd/drbd_bitmap.c struct drbd_bitmap *b = device->bitmap; b 1303 drivers/block/drbd/drbd_bitmap.c if (!expect(b)) b 1305 drivers/block/drbd/drbd_bitmap.c if (!expect(b->bm_pages)) b 1308 drivers/block/drbd/drbd_bitmap.c spin_lock_irq(&b->bm_lock); b 1309 drivers/block/drbd/drbd_bitmap.c if (BM_DONT_TEST & b->bm_flags) b 1314 drivers/block/drbd/drbd_bitmap.c spin_unlock_irq(&b->bm_lock); b 1354 drivers/block/drbd/drbd_bitmap.c struct drbd_bitmap *b = device->bitmap; b 1361 drivers/block/drbd/drbd_bitmap.c if (e >= b->bm_bits) { b 1363 drivers/block/drbd/drbd_bitmap.c s, e, b->bm_bits); b 1364 drivers/block/drbd/drbd_bitmap.c e = b->bm_bits ? b->bm_bits -1 : 0; b 1367 drivers/block/drbd/drbd_bitmap.c unsigned int page_nr = bm_bit_to_page_idx(b, bitnr); b 1372 drivers/block/drbd/drbd_bitmap.c bm_set_page_lazy_writeout(b->bm_pages[last_page_nr]); b 1374 drivers/block/drbd/drbd_bitmap.c bm_set_page_need_writeout(b->bm_pages[last_page_nr]); b 1377 drivers/block/drbd/drbd_bitmap.c p_addr = __bm_map_pidx(b, page_nr); b 1388 drivers/block/drbd/drbd_bitmap.c bm_set_page_lazy_writeout(b->bm_pages[last_page_nr]); b 1390 drivers/block/drbd/drbd_bitmap.c bm_set_page_need_writeout(b->bm_pages[last_page_nr]); b 1392 drivers/block/drbd/drbd_bitmap.c b->bm_set += changed_total; b 1404 drivers/block/drbd/drbd_bitmap.c struct drbd_bitmap *b = device->bitmap; b 1407 drivers/block/drbd/drbd_bitmap.c if (!expect(b)) b 1409 drivers/block/drbd/drbd_bitmap.c if (!expect(b->bm_pages)) b 1412 drivers/block/drbd/drbd_bitmap.c spin_lock_irqsave(&b->bm_lock, flags); b 1413 drivers/block/drbd/drbd_bitmap.c if ((val ? BM_DONT_SET : BM_DONT_CLEAR) & b->bm_flags) b 1418 drivers/block/drbd/drbd_bitmap.c spin_unlock_irqrestore(&b->bm_lock, flags); b 1436 drivers/block/drbd/drbd_bitmap.c static inline void bm_set_full_words_within_one_page(struct drbd_bitmap *b, b 1442 drivers/block/drbd/drbd_bitmap.c unsigned long *paddr = kmap_atomic(b->bm_pages[page_nr]); b 1456 drivers/block/drbd/drbd_bitmap.c bm_set_page_lazy_writeout(b->bm_pages[page_nr]); b 1457 drivers/block/drbd/drbd_bitmap.c b->bm_set += changed; b 1476 drivers/block/drbd/drbd_bitmap.c struct drbd_bitmap *b = device->bitmap; b 1487 drivers/block/drbd/drbd_bitmap.c spin_lock_irq(&b->bm_lock); b 1489 drivers/block/drbd/drbd_bitmap.c spin_unlock_irq(&b->bm_lock); b 1495 drivers/block/drbd/drbd_bitmap.c spin_lock_irq(&b->bm_lock); b 1512 drivers/block/drbd/drbd_bitmap.c spin_unlock_irq(&b->bm_lock); b 1515 drivers/block/drbd/drbd_bitmap.c spin_lock_irq(&b->bm_lock); b 1536 drivers/block/drbd/drbd_bitmap.c spin_unlock_irq(&b->bm_lock); b 1549 drivers/block/drbd/drbd_bitmap.c struct drbd_bitmap *b = device->bitmap; b 1553 drivers/block/drbd/drbd_bitmap.c if (!expect(b)) b 1555 drivers/block/drbd/drbd_bitmap.c if (!expect(b->bm_pages)) b 1558 drivers/block/drbd/drbd_bitmap.c spin_lock_irqsave(&b->bm_lock, flags); b 1559 drivers/block/drbd/drbd_bitmap.c if (BM_DONT_TEST & b->bm_flags) b 1561 drivers/block/drbd/drbd_bitmap.c if (bitnr < b->bm_bits) { b 1562 drivers/block/drbd/drbd_bitmap.c p_addr = bm_map_pidx(b, bm_bit_to_page_idx(b, bitnr)); b 1565 drivers/block/drbd/drbd_bitmap.c } else if (bitnr == b->bm_bits) { b 1568 drivers/block/drbd/drbd_bitmap.c drbd_err(device, "bitnr=%lu > bm_bits=%lu\n", bitnr, b->bm_bits); b 1572 drivers/block/drbd/drbd_bitmap.c spin_unlock_irqrestore(&b->bm_lock, flags); b 1580 drivers/block/drbd/drbd_bitmap.c struct drbd_bitmap *b = device->bitmap; b 1590 drivers/block/drbd/drbd_bitmap.c if (!expect(b)) b 1592 drivers/block/drbd/drbd_bitmap.c if (!expect(b->bm_pages)) b 1595 drivers/block/drbd/drbd_bitmap.c spin_lock_irqsave(&b->bm_lock, flags); b 1596 drivers/block/drbd/drbd_bitmap.c if (BM_DONT_TEST & b->bm_flags) b 1599 drivers/block/drbd/drbd_bitmap.c unsigned int idx = bm_bit_to_page_idx(b, bitnr); b 1604 drivers/block/drbd/drbd_bitmap.c p_addr = bm_map_pidx(b, idx); b 1606 drivers/block/drbd/drbd_bitmap.c if (expect(bitnr < b->bm_bits)) b 1609 drivers/block/drbd/drbd_bitmap.c drbd_err(device, "bitnr=%lu bm_bits=%lu\n", bitnr, b->bm_bits); b 1613 drivers/block/drbd/drbd_bitmap.c spin_unlock_irqrestore(&b->bm_lock, flags); b 1634 drivers/block/drbd/drbd_bitmap.c struct drbd_bitmap *b = device->bitmap; b 1639 drivers/block/drbd/drbd_bitmap.c if (!expect(b)) b 1641 drivers/block/drbd/drbd_bitmap.c if (!expect(b->bm_pages)) b 1644 drivers/block/drbd/drbd_bitmap.c spin_lock_irqsave(&b->bm_lock, flags); b 1645 drivers/block/drbd/drbd_bitmap.c if (BM_DONT_TEST & b->bm_flags) b 1649 drivers/block/drbd/drbd_bitmap.c e = min((size_t)S2W(enr+1), b->bm_words); b 1651 drivers/block/drbd/drbd_bitmap.c if (s < b->bm_words) { b 1653 drivers/block/drbd/drbd_bitmap.c p_addr = bm_map_pidx(b, bm_word_to_page_idx(b, s)); b 1660 drivers/block/drbd/drbd_bitmap.c spin_unlock_irqrestore(&b->bm_lock, flags); b 1184 drivers/block/drbd/drbd_main.c len = bs.cur.b - p->code + !!bs.cur.bit; b 3099 drivers/block/drbd/drbd_main.c void drbd_md_write(struct drbd_device *device, void *b) b 3101 drivers/block/drbd/drbd_main.c struct meta_data_on_disk *buffer = b; b 1207 drivers/block/drbd/drbd_nl.c struct request_queue *b, b 1215 drivers/block/drbd/drbd_nl.c bool can_do = b ? blk_queue_discard(b) : true; b 1269 drivers/block/drbd/drbd_nl.c struct request_queue *b, struct o_qlim *o, b 1274 drivers/block/drbd/drbd_nl.c bool can_do = b ? b->limits.max_write_same_sectors : true; b 1289 drivers/block/drbd/drbd_nl.c unsigned int me_lbs_b = queue_logical_block_size(b); b 1338 drivers/block/drbd/drbd_nl.c struct request_queue *b = NULL; b 1344 drivers/block/drbd/drbd_nl.c b = bdev->backing_bdev->bd_disk->queue; b 1346 drivers/block/drbd/drbd_nl.c max_hw_sectors = min(queue_max_hw_sectors(b), max_bio_size >> 9); b 1361 drivers/block/drbd/drbd_nl.c decide_on_discard_support(device, q, b, discard_zeroes_if_aligned); b 1362 drivers/block/drbd/drbd_nl.c decide_on_write_same_support(device, q, b, o, disable_write_same); b 1364 drivers/block/drbd/drbd_nl.c if (b) { b 1365 drivers/block/drbd/drbd_nl.c blk_queue_stack_limits(q, b); b 1368 drivers/block/drbd/drbd_nl.c b->backing_dev_info->ra_pages) { b 1371 drivers/block/drbd/drbd_nl.c b->backing_dev_info->ra_pages); b 1373 drivers/block/drbd/drbd_nl.c b->backing_dev_info->ra_pages; b 1505 drivers/block/drbd/drbd_nl.c static bool write_ordering_changed(struct disk_conf *a, struct disk_conf *b) b 1507 drivers/block/drbd/drbd_nl.c return a->disk_barrier != b->disk_barrier || b 1508 drivers/block/drbd/drbd_nl.c a->disk_flushes != b->disk_flushes || b 1509 drivers/block/drbd/drbd_nl.c a->disk_drain != b->disk_drain; b 2305 drivers/block/drbd/drbd_receiver.c static bool seq_greater(u32 a, u32 b) b 2312 drivers/block/drbd/drbd_receiver.c return (s32)a - (s32)b > 0; b 2315 drivers/block/drbd/drbd_receiver.c static u32 seq_max(u32 a, u32 b) b 2317 drivers/block/drbd/drbd_receiver.c return seq_greater(a, b) ? a : b; b 4092 drivers/block/drbd/drbd_receiver.c const char *s, sector_t a, sector_t b) b 4095 drivers/block/drbd/drbd_receiver.c if (a == 0 || b == 0) b 4097 drivers/block/drbd/drbd_receiver.c d = (a > b) ? (a - b) : (b - a); b 4098 drivers/block/drbd/drbd_receiver.c if (d > (a>>3) || d > (b>>3)) b 4100 drivers/block/drbd/drbd_receiver.c (unsigned long long)a, (unsigned long long)b); b 4769 drivers/block/drbd/drbd_receiver.c (unsigned int)(bs.cur.b - p->code), b 528 drivers/block/drbd/drbd_req.c char b[BDEVNAME_SIZE]; b 537 drivers/block/drbd/drbd_req.c bdevname(device->ldev->backing_bdev, b)); b 138 drivers/block/drbd/drbd_vli.h #define LEVEL(t,b,v) \ b 140 drivers/block/drbd/drbd_vli.h if ((in & ((1 << b) -1)) == v) { \ b 141 drivers/block/drbd/drbd_vli.h *out = ((in & ((~0ULL) >> (64-t))) >> b) + adj; \ b 144 drivers/block/drbd/drbd_vli.h adj += 1ULL << (t - b); \ b 164 drivers/block/drbd/drbd_vli.h #define LEVEL(t,b,v) do { \ b 165 drivers/block/drbd/drbd_vli.h max += 1ULL << (t - b); \ b 168 drivers/block/drbd/drbd_vli.h *out = ((in - adj) << b) | v; \ b 193 drivers/block/drbd/drbd_vli.h u8 *b; b 201 drivers/block/drbd/drbd_vli.h cur->b = s; b 210 drivers/block/drbd/drbd_vli.h cur->b = cur->b + (bits >> 3); b 250 drivers/block/drbd/drbd_vli.h unsigned char *b = bs->cur.b; b 256 drivers/block/drbd/drbd_vli.h if ((bs->cur.b + ((bs->cur.bit + bits -1) >> 3)) - bs->buf >= bs->buf_len) b 263 drivers/block/drbd/drbd_vli.h *b++ |= (val & 0xff) << bs->cur.bit; b 266 drivers/block/drbd/drbd_vli.h *b++ |= (val >> tmp) & 0xff; b 289 drivers/block/drbd/drbd_vli.h if (bs->cur.b + ((bs->cur.bit + bs->pad_bits + bits -1) >> 3) - bs->buf >= bs->buf_len) b 290 drivers/block/drbd/drbd_vli.h bits = ((bs->buf_len - (bs->cur.b - bs->buf)) << 3) b 304 drivers/block/drbd/drbd_vli.h memcpy(&val, bs->cur.b+1, n - 1); b 309 drivers/block/drbd/drbd_vli.h val |= bs->cur.b[0] >> bs->cur.bit; b 703 drivers/block/floppy.c #define INFBOUND(a, b) (a) = max_t(int, a, b) b 704 drivers/block/floppy.c #define SUPBOUND(a, b) (a) = min_t(int, a, b) b 314 drivers/block/loop.c struct bio_vec bvec, b; b 329 drivers/block/loop.c b.bv_page = page; b 330 drivers/block/loop.c b.bv_offset = 0; b 331 drivers/block/loop.c b.bv_len = bvec.bv_len; b 332 drivers/block/loop.c ret = lo_write_bvec(lo->lo_backing_file, &b, &pos); b 373 drivers/block/loop.c struct bio_vec bvec, b; b 387 drivers/block/loop.c b.bv_page = page; b 388 drivers/block/loop.c b.bv_offset = 0; b 389 drivers/block/loop.c b.bv_len = bvec.bv_len; b 391 drivers/block/loop.c iov_iter_bvec(&i, READ, &b, 1, b.bv_len); b 766 drivers/block/loop.c struct device_attribute *attr, char *b) \ b 768 drivers/block/loop.c return loop_attr_show(d, b, loop_attr_##_name##_show); \ b 30 drivers/block/paride/aten.c #define j44(a,b) ((((a>>4)&0x0f)|(b&0xf0))^0x88) b 49 drivers/block/paride/aten.c { int a, b, r; b 57 drivers/block/paride/aten.c a = r1(); w0(0x10); b = r1(); w2(0xc); b 58 drivers/block/paride/aten.c return j44(a,b); b 72 drivers/block/paride/aten.c { int k, a, b, c, d; b 79 drivers/block/paride/aten.c a = r1(); w0(0x58); b = r1(); b 82 drivers/block/paride/aten.c buf[2*k+1] = j44(a,b); b 90 drivers/block/paride/aten.c a = r0(); w2(0x20); b = r0(); b 91 drivers/block/paride/aten.c buf[2*k] = b; buf[2*k+1] = a; b 34 drivers/block/paride/comm.c #define j44(a,b) (((a>>3)&0x0f)|((b<<1)&0xf0)) b 35 drivers/block/paride/dstr.c #define j44(a,b) (((a>>3)&0x07)|((~a>>4)&0x08)|((b<<1)&0x70)|((~b)&0x80)) b 49 drivers/block/paride/dstr.c { int a, b, r; b 59 drivers/block/paride/dstr.c case 0: w2(6); a = r1(); w2(4); w2(6); b = r1(); w2(4); b 60 drivers/block/paride/dstr.c return j44(a,b); b 117 drivers/block/paride/dstr.c { int k, a, b; b 127 drivers/block/paride/dstr.c w2(6); b = r1(); w2(4); b 128 drivers/block/paride/dstr.c buf[k] = j44(a,b); b 31 drivers/block/paride/epat.c #define j44(a,b) (((a>>4)&0x0f)+(b&0xf0)) b 32 drivers/block/paride/epat.c #define j53(a,b) (((a>>3)&0x1f)+((b<<4)&0xe0)) b 70 drivers/block/paride/epat.c { int a, b, r; b 77 drivers/block/paride/epat.c a = r1(); w2(4); b = r1(); b 78 drivers/block/paride/epat.c return j44(a,b); b 81 drivers/block/paride/epat.c a = r1(); b = r2(); w0(0xff); b 82 drivers/block/paride/epat.c return j53(a,b); b 99 drivers/block/paride/epat.c { int k, ph, a, b; b 108 drivers/block/paride/epat.c if (a & 8) b = a; b 109 drivers/block/paride/epat.c else { w2(4+ph); b = r1(); } b 110 drivers/block/paride/epat.c buf[k] = j44(a,b); b 121 drivers/block/paride/epat.c a = r1(); b = r2(); b 122 drivers/block/paride/epat.c buf[k] = j53(a,b); b 40 drivers/block/paride/epia.c #define j44(a,b) (((a>>4)&0x0f)+(b&0xf0)) b 41 drivers/block/paride/epia.c #define j53(a,b) (((a>>3)&0x1f)+((b<<4)&0xe0)) b 51 drivers/block/paride/epia.c { int a, b, r; b 59 drivers/block/paride/epia.c a = r1(); w2(1); b = r1(); w2(4); b 60 drivers/block/paride/epia.c return j44(a,b); b 65 drivers/block/paride/epia.c a = r1(); b = r2(); w2(4); b 66 drivers/block/paride/epia.c return j53(a,b); b 138 drivers/block/paride/epia.c { int k, ph, a, b; b 146 drivers/block/paride/epia.c w2(4+ph); b = r1(); b 147 drivers/block/paride/epia.c buf[k] = j44(a,b); b 158 drivers/block/paride/epia.c a = r1(); b = r2(); b 159 drivers/block/paride/epia.c buf[k] = j53(a,b); b 28 drivers/block/paride/fit2.c #define j44(a,b) (((a>>4)&0x0f)|(b&0xf0)) b 48 drivers/block/paride/fit2.c { int a, b, r; b 57 drivers/block/paride/fit2.c w0(1); b = r1(); b 60 drivers/block/paride/fit2.c return j44(a,b); b 66 drivers/block/paride/fit2.c { int k, a, b, c, d; b 73 drivers/block/paride/fit2.c w0(0); a = r1(); w0(1); b = r1(); b 75 drivers/block/paride/fit2.c buf[4*k+0] = j44(a,b); b 79 drivers/block/paride/fit2.c a = r1(); w0(3); b = r1(); b 82 drivers/block/paride/fit2.c buf[4*k+3] = j44(a,b); b 32 drivers/block/paride/fit3.c #define j44(a,b) (((a>>3)&0x0f)|((b<<1)&0xf0)) b 64 drivers/block/paride/fit3.c { int a, b; b 75 drivers/block/paride/fit3.c w2(0xf); b = r1(); b 77 drivers/block/paride/fit3.c return j44(a,b); b 86 drivers/block/paride/fit3.c a = r4(); b = r4(); b 97 drivers/block/paride/fit3.c { int k, a, b, c, d; b 104 drivers/block/paride/fit3.c w2(0xf); b = r1(); b 107 drivers/block/paride/fit3.c buf[2*k ] = j44(a,b); b 117 drivers/block/paride/fit3.c w2(0xee); b = r0(); b 119 drivers/block/paride/fit3.c buf[2*k+1] = b; b 195 drivers/block/paride/frpw.c { int olddelay, a, b; b 209 drivers/block/paride/frpw.c a = r1() & 0xff; w2(4); b = r1() & 0xff; b 216 drivers/block/paride/frpw.c return ((~a&0x40) && (b&0x40)); b 35 drivers/block/paride/kbic.c #define j44(a,b) ((((a>>4)&0x0f)|(b&0xf0))^0x88) b 47 drivers/block/paride/kbic.c { int a, b, s; b 54 drivers/block/paride/kbic.c a = r1(); w0(0x28); b = r1(); w2(4); b 55 drivers/block/paride/kbic.c return j44(a,b); b 68 drivers/block/paride/kbic.c a = r4(); b = r4(); w2(4); w2(0); w2(4); b 136 drivers/block/paride/kbic.c { int k, a, b; b 143 drivers/block/paride/kbic.c w0(0x28); b = r1(); b 144 drivers/block/paride/kbic.c buf[2*k] = j44(a,b); b 145 drivers/block/paride/kbic.c w2(5); b = r1(); b 147 drivers/block/paride/kbic.c buf[2*k+1] = j44(a,b); b 24 drivers/block/paride/ktti.c #define j44(a,b) (((a>>4)&0x0f)|(b&0xf0)) b 44 drivers/block/paride/ktti.c { int a, b, r; b 49 drivers/block/paride/ktti.c a = r1(); w2(0xc); b = r1(); w2(9); w2(0xc); w2(9); b 50 drivers/block/paride/ktti.c return j44(a,b); b 56 drivers/block/paride/ktti.c { int k, a, b; b 60 drivers/block/paride/ktti.c a = r1(); w2(0xc); b = r1(); w2(9); b 61 drivers/block/paride/ktti.c buf[2*k] = j44(a,b); b 62 drivers/block/paride/ktti.c a = r1(); w2(0xc); b = r1(); w2(9); b 63 drivers/block/paride/ktti.c buf[2*k+1] = j44(a,b); b 30 drivers/block/paride/on20.c #define j44(a,b) (((a>>4)&0x0f)|(b&0xf0)) b 38 drivers/block/paride/on26.c #define j44(a,b) (((a>>4)&0x0f)|(b&0xf0)) b 49 drivers/block/paride/on26.c { int a, b, r; b 57 drivers/block/paride/on26.c w2(6); b = r1(); w2(4); b 59 drivers/block/paride/on26.c return j44(a,b); b 188 drivers/block/paride/on26.c { int k, a, b; b 196 drivers/block/paride/on26.c w2(4); b = r1(); b 197 drivers/block/paride/on26.c buf[k] = j44(a,b); b 118 drivers/block/paride/paride.h static inline u16 pi_swab16( char *b, int k) b 122 drivers/block/paride/paride.h r.t[0]=b[2*k+1]; r.t[1]=b[2*k]; b 126 drivers/block/paride/paride.h static inline u32 pi_swab32( char *b, int k) b 130 drivers/block/paride/paride.h r.f[0]=b[4*k+1]; r.f[1]=b[4*k]; b 131 drivers/block/paride/paride.h r.f[2]=b[4*k+3]; r.f[3]=b[4*k+2]; b 864 drivers/block/paride/pcd.c int b, i; b 868 drivers/block/paride/pcd.c b = pcd_bufblk; b 870 drivers/block/paride/pcd.c rd_cmd[5 - i] = b & 0xff; b 871 drivers/block/paride/pcd.c b = b >> 8; b 779 drivers/block/paride/pf.c static int pf_start(struct pf_unit *pf, int cmd, int b, int c) b 785 drivers/block/paride/pf.c io_cmd[5 - i] = b & 0xff; b 786 drivers/block/paride/pf.c b = b >> 8; b 770 drivers/block/paride/pt.c int k, n, r, p, s, t, b; b 792 drivers/block/paride/pt.c b = (n - 1 + tape->bs) / tape->bs; b 793 drivers/block/paride/pt.c n = b * tape->bs; /* rounded up to even block */ b 795 drivers/block/paride/pt.c rd_cmd[4] = b; b 841 drivers/block/paride/pt.c b = k; b 842 drivers/block/paride/pt.c if (b > count) b 843 drivers/block/paride/pt.c b = count; b 844 drivers/block/paride/pt.c if (copy_to_user(buf + t, tape->bufptr, b)) { b 848 drivers/block/paride/pt.c t += b; b 849 drivers/block/paride/pt.c count -= b; b 867 drivers/block/paride/pt.c int k, n, r, p, s, t, b; b 893 drivers/block/paride/pt.c b = (n - 1 + tape->bs) / tape->bs; b 894 drivers/block/paride/pt.c n = b * tape->bs; /* rounded up to even block */ b 896 drivers/block/paride/pt.c wr_cmd[4] = b; b 940 drivers/block/paride/pt.c b = k; b 941 drivers/block/paride/pt.c if (b > count) b 942 drivers/block/paride/pt.c b = count; b 943 drivers/block/paride/pt.c if (copy_from_user(tape->bufptr, buf + t, b)) { b 948 drivers/block/paride/pt.c t += b; b 949 drivers/block/paride/pt.c count -= b; b 2434 drivers/block/pktcdvd.c char b[BDEVNAME_SIZE]; b 2441 drivers/block/pktcdvd.c pr_err("%s incorrect request queue\n", bio_devname(bio, b)); b 2568 drivers/block/pktcdvd.c char b[BDEVNAME_SIZE]; b 2581 drivers/block/pktcdvd.c bdevname(pd2->bdev, b)); b 2618 drivers/block/pktcdvd.c pkt_dbg(1, pd, "writer mapped to %s\n", bdevname(bdev, b)); b 450 drivers/bluetooth/bt3c_cs.c char b[9]; b 475 drivers/bluetooth/bt3c_cs.c memset(b, 0, sizeof(b)); b 476 drivers/bluetooth/bt3c_cs.c memcpy(b, ptr + 2, 2); b 477 drivers/bluetooth/bt3c_cs.c if (kstrtoul(b, 16, &size) < 0) b 480 drivers/bluetooth/bt3c_cs.c memset(b, 0, sizeof(b)); b 481 drivers/bluetooth/bt3c_cs.c memcpy(b, ptr + 4, 8); b 482 drivers/bluetooth/bt3c_cs.c if (kstrtoul(b, 16, &addr) < 0) b 485 drivers/bluetooth/bt3c_cs.c memset(b, 0, sizeof(b)); b 486 drivers/bluetooth/bt3c_cs.c memcpy(b, ptr + (size * 2) + 2, 2); b 487 drivers/bluetooth/bt3c_cs.c if (kstrtoul(b, 16, &fcs) < 0) b 490 drivers/bluetooth/bt3c_cs.c memset(b, 0, sizeof(b)); b 492 drivers/bluetooth/bt3c_cs.c memcpy(b, ptr + (i * 2) + 2, 2); b 493 drivers/bluetooth/bt3c_cs.c if (kstrtouint(b, 16, &tn)) b 507 drivers/bluetooth/bt3c_cs.c memset(b, 0, sizeof(b)); b 509 drivers/bluetooth/bt3c_cs.c memcpy(b, ptr + (i * 4) + 12, 4); b 510 drivers/bluetooth/bt3c_cs.c if (kstrtouint(b, 16, &tmp)) b 172 drivers/char/agp/agp.h #define PGE_EMPTY(b, p) (!(p) || (p) == (unsigned long) (b)->scratch_page) b 156 drivers/char/dsp56k.c dsp56k_host_interface.data.b[1] = fw->data[i]; b 157 drivers/char/dsp56k.c dsp56k_host_interface.data.b[2] = fw->data[i + 1]; b 158 drivers/char/dsp56k.c dsp56k_host_interface.data.b[3] = fw->data[i + 2]; b 163 drivers/char/dsp56k.c dsp56k_host_interface.data.b[1] = 0; b 164 drivers/char/dsp56k.c dsp56k_host_interface.data.b[2] = 0; b 165 drivers/char/dsp56k.c dsp56k_host_interface.data.b[3] = 0; b 170 drivers/char/dsp56k.c get_user(dsp56k_host_interface.data.b[1], bin++); b 171 drivers/char/dsp56k.c get_user(dsp56k_host_interface.data.b[2], bin++); b 172 drivers/char/dsp56k.c get_user(dsp56k_host_interface.data.b[3], bin++); b 202 drivers/char/dsp56k.c put_user(dsp56k_host_interface.data.b[3], buf+n++)); b 219 drivers/char/dsp56k.c put_user(dsp56k_host_interface.data.b[1], buf+n++); b 220 drivers/char/dsp56k.c put_user(dsp56k_host_interface.data.b[2], buf+n++); b 221 drivers/char/dsp56k.c put_user(dsp56k_host_interface.data.b[3], buf+n++)); b 264 drivers/char/dsp56k.c get_user(dsp56k_host_interface.data.b[3], buf+n++)); b 281 drivers/char/dsp56k.c get_user(dsp56k_host_interface.data.b[1], buf+n++); b 282 drivers/char/dsp56k.c get_user(dsp56k_host_interface.data.b[2], buf+n++); b 283 drivers/char/dsp56k.c get_user(dsp56k_host_interface.data.b[3], buf+n++)); b 440 drivers/char/dtlk.c buffer[b++] = inb_p(dtlk_port_lpc); \ b 444 drivers/char/dtlk.c int b = 0, i, j; b 448 drivers/char/dtlk.c buffer[b++] = 0; b 451 drivers/char/dtlk.c buffer[b++] = 0; b 454 drivers/char/dtlk.c buffer[b++] = 0; b 458 drivers/char/dtlk.c for (j = 0; j < b; j++) b 470 drivers/char/dtlk.c buffer[b++] = inb_p(dtlk_port_tts); \ b 474 drivers/char/dtlk.c int b = 0, i, j; b 479 drivers/char/dtlk.c buffer[b++] = 0; b 484 drivers/char/dtlk.c for (j = 0; j < b; j++) b 201 drivers/char/hw_random/optee-rng.c pvt_data.optee_rng.quality = param[0].u.value.b; b 229 drivers/char/hw_random/optee-rng.c memcpy(sess_arg.uuid, rng_device->id.uuid.b, TEE_IOCTL_UUID_LEN); b 38 drivers/char/ipmi/ipmi_si.h unsigned char b); b 13 drivers/char/ipmi/ipmi_si_mem_io.c unsigned char b) b 15 drivers/char/ipmi/ipmi_si_mem_io.c writeb(b, (io->addr)+(offset * io->regspacing)); b 26 drivers/char/ipmi/ipmi_si_mem_io.c unsigned char b) b 28 drivers/char/ipmi/ipmi_si_mem_io.c writeb(b << io->regshift, (io->addr)+(offset * io->regspacing)); b 39 drivers/char/ipmi/ipmi_si_mem_io.c unsigned char b) b 41 drivers/char/ipmi/ipmi_si_mem_io.c writel(b << io->regshift, (io->addr)+(offset * io->regspacing)); b 52 drivers/char/ipmi/ipmi_si_mem_io.c unsigned char b) b 54 drivers/char/ipmi/ipmi_si_mem_io.c writeq((u64)b << io->regshift, (io->addr)+(offset * io->regspacing)); b 14 drivers/char/ipmi/ipmi_si_port_io.c unsigned char b) b 18 drivers/char/ipmi/ipmi_si_port_io.c outb(b, addr + (offset * io->regspacing)); b 29 drivers/char/ipmi/ipmi_si_port_io.c unsigned char b) b 33 drivers/char/ipmi/ipmi_si_port_io.c outw(b << io->regshift, addr + (offset * io->regspacing)); b 44 drivers/char/ipmi/ipmi_si_port_io.c unsigned char b) b 48 drivers/char/ipmi/ipmi_si_port_io.c outl(b << io->regshift, addr+(offset * io->regspacing)); b 78 drivers/char/ipmi/kcs_bmc.h void (*io_outputb)(struct kcs_bmc *kcs_bmc, u32 reg, u8 b); b 196 drivers/char/pcmcia/cm4000_cs.c static void str_invert_revert(unsigned char *b, int len) b 201 drivers/char/pcmcia/cm4000_cs.c b[i] = invert_revert(b[i]); b 667 drivers/char/random.c __u32 a = f->pool[0], b = f->pool[1]; b 670 drivers/char/random.c a += b; c += d; b 671 drivers/char/random.c b = rol32(b, 6); d = rol32(d, 27); b 672 drivers/char/random.c d ^= a; b ^= c; b 674 drivers/char/random.c a += b; c += d; b 675 drivers/char/random.c b = rol32(b, 16); d = rol32(d, 14); b 676 drivers/char/random.c d ^= a; b ^= c; b 678 drivers/char/random.c a += b; c += d; b 679 drivers/char/random.c b = rol32(b, 6); d = rol32(d, 27); b 680 drivers/char/random.c d ^= a; b ^= c; b 682 drivers/char/random.c a += b; c += d; b 683 drivers/char/random.c b = rol32(b, 16); d = rol32(d, 14); b 684 drivers/char/random.c d ^= a; b ^= c; b 686 drivers/char/random.c f->pool[0] = a; f->pool[1] = b; b 2069 drivers/char/random.c int b, i = 0; b 2075 drivers/char/random.c for (b = bytes ; b > 0 ; b -= sizeof(__u32), i++) { b 206 drivers/char/tpm/tpm.h __be32 b; b 370 drivers/char/tpm/tpm1-cmd.c timeout_chip[1] = be32_to_cpu(cap.timeout.b); b 244 drivers/char/tpm/tpm_ftpm_tee.c memcpy(sess_arg.uuid, ftpm_ta_uuid.b, TEE_IOCTL_UUID_LEN); b 450 drivers/clk/clk-si5351.c unsigned long rfrac, denom, a, b, c; b 473 drivers/clk/clk-si5351.c b = 0; b 477 drivers/clk/clk-si5351.c SI5351_PLL_B_MAX, SI5351_PLL_C_MAX, &b, &c); b 481 drivers/clk/clk-si5351.c hwdata->params.p2 = (128 * b) % c; b 483 drivers/clk/clk-si5351.c hwdata->params.p1 += (128 * b / c); b 488 drivers/clk/clk-si5351.c lltmp *= b; b 496 drivers/clk/clk-si5351.c __func__, clk_hw_get_name(hw), a, b, c, b 649 drivers/clk/clk-si5351.c unsigned long a, b, c; b 679 drivers/clk/clk-si5351.c b = 0; b 691 drivers/clk/clk-si5351.c b = 0; b 716 drivers/clk/clk-si5351.c b = 0; b 721 drivers/clk/clk-si5351.c &b, &c); b 727 drivers/clk/clk-si5351.c do_div(lltmp, a * c + b); b 741 drivers/clk/clk-si5351.c hwdata->params.p2 = (128 * b) % c; b 743 drivers/clk/clk-si5351.c hwdata->params.p1 += (128 * b / c); b 749 drivers/clk/clk-si5351.c __func__, clk_hw_get_name(hw), a, b, c, divby4, b 347 drivers/clk/keystone/sci-clk.c static int _cmp_sci_clk(const void *a, const void *b) b 350 drivers/clk/keystone/sci-clk.c const struct sci_clk *cb = *(struct sci_clk **)b; b 492 drivers/clk/keystone/sci-clk.c struct list_head *b) b 495 drivers/clk/keystone/sci-clk.c struct sci_clk *cb = container_of(b, struct sci_clk, node); b 192 drivers/clk/meson/axg-audio.c static AUD_MST_SCLK_PRE_EN(b, AUDIO_MST_B_SCLK_CTRL0); b 228 drivers/clk/meson/axg-audio.c static AUD_MST_SCLK_DIV(b, AUDIO_MST_B_SCLK_CTRL0); b 239 drivers/clk/meson/axg-audio.c static AUD_MST_SCLK_POST_EN(b, AUDIO_MST_B_SCLK_CTRL0); b 279 drivers/clk/meson/axg-audio.c static AUD_MST_SCLK(b, AUDIO_MST_B_SCLK_CTRL1); b 290 drivers/clk/meson/axg-audio.c static AUD_MST_LRCLK_DIV(b, AUDIO_MST_B_SCLK_CTRL0); b 301 drivers/clk/meson/axg-audio.c static AUD_MST_LRCLK(b, AUDIO_MST_B_SCLK_CTRL1); b 384 drivers/clk/pxa/clk-pxa27x.c unsigned int b, osc_forced; b 389 drivers/clk/pxa/clk-pxa27x.c b = clkcfg & (1 << 3); b 393 drivers/clk/pxa/clk-pxa27x.c if (b) b 26 drivers/clk/renesas/r7s9210-cpg-mssr.c unsigned int b; b 146 drivers/clk/renesas/r7s9210-cpg-mssr.c r7s9210_core_clks[i].div = ratio_tab[index].b; b 691 drivers/clk/rockchip/clk.h #define GATE(_id, cname, pname, f, o, b, gf) \ b 700 drivers/clk/rockchip/clk.h .gate_shift = b, \ b 48 drivers/clk/samsung/clk-s3c64xx.c #define GATE_BUS(_id, cname, pname, o, b) \ b 49 drivers/clk/samsung/clk-s3c64xx.c GATE(_id, cname, pname, o, b, 0, 0) b 50 drivers/clk/samsung/clk-s3c64xx.c #define GATE_SCLK(_id, cname, pname, o, b) \ b 51 drivers/clk/samsung/clk-s3c64xx.c GATE(_id, cname, pname, o, b, CLK_SET_RATE_PARENT, 0) b 52 drivers/clk/samsung/clk-s3c64xx.c #define GATE_ON(_id, cname, pname, o, b) \ b 53 drivers/clk/samsung/clk-s3c64xx.c GATE(_id, cname, pname, o, b, CLK_IGNORE_UNUSED, 0) b 211 drivers/clk/samsung/clk.h #define __GATE(_id, cname, pname, o, b, f, gf) \ b 218 drivers/clk/samsung/clk.h .bit_idx = b, \ b 222 drivers/clk/samsung/clk.h #define GATE(_id, cname, pname, o, b, f, gf) \ b 223 drivers/clk/samsung/clk.h __GATE(_id, cname, pname, o, b, f, gf) b 14 drivers/clk/socfpga/clk-gate-a10.c #define streq(a, b) (strcmp((a), (b)) == 0) b 22 drivers/clk/socfpga/clk.h #define streq(a, b) (strcmp((a), (b)) == 0) b 426 drivers/clk/tegra/clk-emc.c const struct emc_timing *b = _b; b 428 drivers/clk/tegra/clk-emc.c if (a->rate < b->rate) b 430 drivers/clk/tegra/clk-emc.c else if (a->rate == b->rate) b 23 drivers/clocksource/h8300_timer16.c #define bset(b, a) iowrite8(ioread8(a) | (1 << (b)), (a)) b 24 drivers/clocksource/h8300_timer16.c #define bclr(b, a) iowrite8(ioread8(a) & ~(1 << (b)), (a)) b 35 drivers/clocksource/h8300_timer8.c #define bset(b, a) iowrite8(ioread8(a) | (1 << (b)), (a)) b 36 drivers/clocksource/h8300_timer8.c #define bclr(b, a) iowrite8(ioread8(a) & ~(1 << (b)), (a)) b 957 drivers/cpufreq/intel_pstate.c static ssize_t store_status(struct kobject *a, struct kobj_attribute *b, b 1040 drivers/cpufreq/intel_pstate.c static ssize_t store_no_turbo(struct kobject *a, struct kobj_attribute *b, b 1128 drivers/cpufreq/intel_pstate.c static ssize_t store_max_perf_pct(struct kobject *a, struct kobj_attribute *b, b 1161 drivers/cpufreq/intel_pstate.c static ssize_t store_min_perf_pct(struct kobject *a, struct kobj_attribute *b, b 1202 drivers/cpufreq/intel_pstate.c struct kobj_attribute *b, b 39 drivers/cpufreq/s3c2440-cpufreq.c static inline int within_khz(unsigned long a, unsigned long b) b 41 drivers/cpufreq/s3c2440-cpufreq.c long diff = a - b; b 517 drivers/cpufreq/s3c24xx-cpufreq.c struct s3c_freq *a, struct s3c_freq *b) b 519 drivers/cpufreq/s3c24xx-cpufreq.c dst->fclk = do_min(a->fclk, b->fclk); b 520 drivers/cpufreq/s3c24xx-cpufreq.c dst->hclk = do_min(a->hclk, b->hclk); b 521 drivers/cpufreq/s3c24xx-cpufreq.c dst->pclk = do_min(a->pclk, b->pclk); b 522 drivers/cpufreq/s3c24xx-cpufreq.c dst->armclk = do_min(a->armclk, b->armclk); b 738 drivers/cpuidle/cpuidle.c static int cpuidle_latency_notify(struct notifier_block *b, b 431 drivers/crypto/axis/artpec6_crypto.c struct artpec6_crypto_bounce_buffer *b; b 434 drivers/crypto/axis/artpec6_crypto.c list_for_each_entry_safe(b, next, &dma->bounce_buffers, list) { b 436 drivers/crypto/axis/artpec6_crypto.c b, b->length, b->offset, b->buf); b 437 drivers/crypto/axis/artpec6_crypto.c sg_pcopy_from_buffer(b->sg, b 439 drivers/crypto/axis/artpec6_crypto.c b->buf, b 440 drivers/crypto/axis/artpec6_crypto.c b->length, b 441 drivers/crypto/axis/artpec6_crypto.c b->offset); b 443 drivers/crypto/axis/artpec6_crypto.c list_del(&b->list); b 444 drivers/crypto/axis/artpec6_crypto.c kfree(b); b 1062 drivers/crypto/axis/artpec6_crypto.c struct artpec6_crypto_bounce_buffer *b; b 1065 drivers/crypto/axis/artpec6_crypto.c list_for_each_entry_safe(b, next, &dma->bounce_buffers, list) { b 1066 drivers/crypto/axis/artpec6_crypto.c kfree(b); b 49 drivers/crypto/cavium/cpt/cpt_common.h #define CPTX_PF_MBOX_INTX(a, b) \ b 50 drivers/crypto/cavium/cpt/cpt_common.h (0x400ll + ((u64)(a) << 36) + ((b) << 3)) b 51 drivers/crypto/cavium/cpt/cpt_common.h #define CPTX_PF_MBOX_INT_W1SX(a, b) \ b 52 drivers/crypto/cavium/cpt/cpt_common.h (0x420ll + ((u64)(a) << 36) + ((b) << 3)) b 53 drivers/crypto/cavium/cpt/cpt_common.h #define CPTX_PF_MBOX_ENA_W1CX(a, b) \ b 54 drivers/crypto/cavium/cpt/cpt_common.h (0x440ll + ((u64)(a) << 36) + ((b) << 3)) b 55 drivers/crypto/cavium/cpt/cpt_common.h #define CPTX_PF_MBOX_ENA_W1SX(a, b) \ b 56 drivers/crypto/cavium/cpt/cpt_common.h (0x460ll + ((u64)(a) << 36) + ((b) << 3)) b 61 drivers/crypto/cavium/cpt/cpt_common.h #define CPTX_PF_GX_EN(a, b) \ b 62 drivers/crypto/cavium/cpt/cpt_common.h (0x600ll + ((u64)(a) << 36) + ((b) << 3)) b 83 drivers/crypto/cavium/cpt/cpt_common.h #define CPTX_PF_EXE_DBG_CNTX(a, b) \ b 84 drivers/crypto/cavium/cpt/cpt_common.h (0x4001100ll + ((u64)(a) << 36) + ((b) << 3)) b 86 drivers/crypto/cavium/cpt/cpt_common.h #define CPTX_PF_EXE_EPCI_INBX_CNT(a, b) \ b 87 drivers/crypto/cavium/cpt/cpt_common.h (0x4001200ll + ((u64)(a) << 36) + ((b) << 3)) b 88 drivers/crypto/cavium/cpt/cpt_common.h #define CPTX_PF_EXE_EPCI_OUTBX_CNT(a, b) \ b 89 drivers/crypto/cavium/cpt/cpt_common.h (0x4001240ll + ((u64)(a) << 36) + ((b) << 3)) b 90 drivers/crypto/cavium/cpt/cpt_common.h #define CPTX_PF_ENGX_UCODE_BASE(a, b) \ b 91 drivers/crypto/cavium/cpt/cpt_common.h (0x4002000ll + ((u64)(a) << 36) + ((b) << 3)) b 92 drivers/crypto/cavium/cpt/cpt_common.h #define CPTX_PF_QX_CTL(a, b) \ b 93 drivers/crypto/cavium/cpt/cpt_common.h (0x8000000ll + ((u64)(a) << 36) + ((b) << 20)) b 94 drivers/crypto/cavium/cpt/cpt_common.h #define CPTX_PF_QX_GMCTL(a, b) \ b 95 drivers/crypto/cavium/cpt/cpt_common.h (0x8000020ll + ((u64)(a) << 36) + ((b) << 20)) b 96 drivers/crypto/cavium/cpt/cpt_common.h #define CPTX_PF_QX_CTL2(a, b) \ b 97 drivers/crypto/cavium/cpt/cpt_common.h (0x8000100ll + ((u64)(a) << 36) + ((b) << 20)) b 98 drivers/crypto/cavium/cpt/cpt_common.h #define CPTX_PF_VFX_MBOXX(a, b, c) \ b 99 drivers/crypto/cavium/cpt/cpt_common.h (0x8001000ll + ((u64)(a) << 36) + ((b) << 20) + ((c) << 8)) b 102 drivers/crypto/cavium/cpt/cpt_common.h #define CPTX_VQX_CTL(a, b) (0x100ll + ((u64)(a) << 36) + ((b) << 20)) b 103 drivers/crypto/cavium/cpt/cpt_common.h #define CPTX_VQX_SADDR(a, b) (0x200ll + ((u64)(a) << 36) + ((b) << 20)) b 104 drivers/crypto/cavium/cpt/cpt_common.h #define CPTX_VQX_DONE_WAIT(a, b) (0x400ll + ((u64)(a) << 36) + ((b) << 20)) b 105 drivers/crypto/cavium/cpt/cpt_common.h #define CPTX_VQX_INPROG(a, b) (0x410ll + ((u64)(a) << 36) + ((b) << 20)) b 106 drivers/crypto/cavium/cpt/cpt_common.h #define CPTX_VQX_DONE(a, b) (0x420ll + ((u64)(a) << 36) + ((b) << 20)) b 107 drivers/crypto/cavium/cpt/cpt_common.h #define CPTX_VQX_DONE_ACK(a, b) (0x440ll + ((u64)(a) << 36) + ((b) << 20)) b 108 drivers/crypto/cavium/cpt/cpt_common.h #define CPTX_VQX_DONE_INT_W1S(a, b) (0x460ll + ((u64)(a) << 36) + ((b) << 20)) b 109 drivers/crypto/cavium/cpt/cpt_common.h #define CPTX_VQX_DONE_INT_W1C(a, b) (0x468ll + ((u64)(a) << 36) + ((b) << 20)) b 110 drivers/crypto/cavium/cpt/cpt_common.h #define CPTX_VQX_DONE_ENA_W1S(a, b) (0x470ll + ((u64)(a) << 36) + ((b) << 20)) b 111 drivers/crypto/cavium/cpt/cpt_common.h #define CPTX_VQX_DONE_ENA_W1C(a, b) (0x478ll + ((u64)(a) << 36) + ((b) << 20)) b 112 drivers/crypto/cavium/cpt/cpt_common.h #define CPTX_VQX_MISC_INT(a, b) (0x500ll + ((u64)(a) << 36) + ((b) << 20)) b 113 drivers/crypto/cavium/cpt/cpt_common.h #define CPTX_VQX_MISC_INT_W1S(a, b) (0x508ll + ((u64)(a) << 36) + ((b) << 20)) b 114 drivers/crypto/cavium/cpt/cpt_common.h #define CPTX_VQX_MISC_ENA_W1S(a, b) (0x510ll + ((u64)(a) << 36) + ((b) << 20)) b 115 drivers/crypto/cavium/cpt/cpt_common.h #define CPTX_VQX_MISC_ENA_W1C(a, b) (0x518ll + ((u64)(a) << 36) + ((b) << 20)) b 116 drivers/crypto/cavium/cpt/cpt_common.h #define CPTX_VQX_DOORBELL(a, b) (0x600ll + ((u64)(a) << 36) + ((b) << 20)) b 117 drivers/crypto/cavium/cpt/cpt_common.h #define CPTX_VFX_PF_MBOXX(a, b, c) \ b 118 drivers/crypto/cavium/cpt/cpt_common.h (0x1000ll + ((u64)(a) << 36) + ((b) << 20) + ((c) << 3)) b 992 drivers/crypto/chelsio/chcr_algo.c __be32 *b = (__be32 *)(dstiv + size); b 997 drivers/crypto/chelsio/chcr_algo.c prev = be32_to_cpu(*--b); b 999 drivers/crypto/chelsio/chcr_algo.c *b = cpu_to_be32(c); b 1009 drivers/crypto/chelsio/chcr_algo.c __be32 *b = (__be32 *)(iv + AES_BLOCK_SIZE); b 1011 drivers/crypto/chelsio/chcr_algo.c u32 temp = be32_to_cpu(*--b); b 1829 drivers/crypto/n2_core.c struct ino_blob *b = &ip->ino_table[i]; b 1830 drivers/crypto/n2_core.c b->intr = i + 1; b 1831 drivers/crypto/n2_core.c b->ino = ino[i]; b 128 drivers/crypto/qat/qat_common/qat_asym_algs.c dma_addr_t b; b 198 drivers/crypto/qat/qat_common/qat_asym_algs.c req->src_align, req->in.dh.in.b); b 200 drivers/crypto/qat/qat_common/qat_asym_algs.c dma_unmap_single(dev, req->in.dh.in.b, b 306 drivers/crypto/qat/qat_common/qat_asym_algs.c qat_req->in.dh.in.b = ctx->dma_g; b 324 drivers/crypto/qat/qat_common/qat_asym_algs.c qat_req->in.dh.in.b = dma_map_single(dev, b 329 drivers/crypto/qat/qat_common/qat_asym_algs.c qat_req->in.dh.in.b))) b 337 drivers/crypto/qat/qat_common/qat_asym_algs.c &qat_req->in.dh.in.b, b 373 drivers/crypto/qat/qat_common/qat_asym_algs.c qat_req->phy_in = dma_map_single(dev, &qat_req->in.dh.in.b, b 419 drivers/crypto/qat/qat_common/qat_asym_algs.c qat_req->in.dh.in.b); b 421 drivers/crypto/qat/qat_common/qat_asym_algs.c if (!dma_mapping_error(dev, qat_req->in.dh.in.b)) b 422 drivers/crypto/qat/qat_common/qat_asym_algs.c dma_unmap_single(dev, qat_req->in.dh.in.b, b 126 drivers/crypto/ux500/cryp/cryp_core.c static inline u8 swap_bits_in_byte(u8 b) b 144 drivers/crypto/ux500/cryp/cryp_core.c n1 = ((b & R_SHIFT_4_MASK) >> 4) | (b & ~(R_SHIFT_4_MASK >> 4)); b 152 drivers/crypto/ux500/cryp/cryp_core.c n2 = ((b & L_SHIFT_4_MASK) << 4) | (b & ~(L_SHIFT_4_MASK << 4)); b 23 drivers/devfreq/governor_simpleondemand.c unsigned long long a, b; b 79 drivers/devfreq/governor_simpleondemand.c b = div_u64(a, stat->total_time); b 80 drivers/devfreq/governor_simpleondemand.c b *= 100; b 81 drivers/devfreq/governor_simpleondemand.c b = div_u64(b, (dfso_upthreshold - dfso_downdifferential / 2)); b 82 drivers/devfreq/governor_simpleondemand.c *freq = (unsigned long) b; b 211 drivers/dma-buf/sync_file.c struct sync_file *b) b 222 drivers/dma-buf/sync_file.c b_fences = get_fences(b, &b_num_fences); b 260 drivers/dma/ioat/hw.h struct ioat_pq16b_descriptor b; b 267 drivers/dma/ioat/hw.h uint64_t b[8]; b 239 drivers/dma/pl330.c #define BYTE_TO_BURST(b, ccr) ((b) / BRST_SIZE(ccr) / BRST_LEN(ccr)) b 1285 drivers/edac/i7core_edac.c struct pci_bus *b = NULL; b 1287 drivers/edac/i7core_edac.c while ((b = pci_find_next_bus(b)) != NULL) { b 1288 drivers/edac/i7core_edac.c bus = b->number; b 333 drivers/edac/skx_base.c #define SKX_TAD_BASE(b) ((u64)GET_BITFIELD((b), 12, 31) << 26) b 334 drivers/edac/skx_base.c #define SKX_TAD_SKT_GRAN(b) GET_BITFIELD((b), 4, 5) b 335 drivers/edac/skx_base.c #define SKX_TAD_CHN_GRAN(b) GET_BITFIELD((b), 6, 7) b 336 drivers/edac/skx_base.c #define SKX_TAD_LIMIT(b) (((u64)GET_BITFIELD((b), 12, 31) << 26) | MASK26) b 337 drivers/edac/skx_base.c #define SKX_TAD_OFFSET(b) ((u64)GET_BITFIELD((b), 4, 23) << 26) b 338 drivers/edac/skx_base.c #define SKX_TAD_SKTWAYS(b) (1 << GET_BITFIELD((b), 10, 11)) b 339 drivers/edac/skx_base.c #define SKX_TAD_CHNWAYS(b) (GET_BITFIELD((b), 8, 9) + 1) b 408 drivers/edac/skx_base.c #define SKX_RIR_VALID(b) GET_BITFIELD((b), 31, 31) b 409 drivers/edac/skx_base.c #define SKX_RIR_LIMIT(b) (((u64)GET_BITFIELD((b), 1, 11) << 29) | MASK29) b 410 drivers/edac/skx_base.c #define SKX_RIR_WAYS(b) (1 << GET_BITFIELD((b), 28, 29)) b 411 drivers/edac/skx_base.c #define SKX_RIR_CHAN_RANK(b) GET_BITFIELD((b), 16, 19) b 412 drivers/edac/skx_base.c #define SKX_RIR_OFFSET(b) ((u64)(GET_BITFIELD((b), 2, 15) << 26)) b 37 drivers/edac/skx_common.h #define MAX(a, b) ((a) > (b) ? (a) : (b)) b 29 drivers/firmware/qcom_scm-64.c #define QCOM_SCM_ARGS_IMPL(num, a, b, c, d, e, f, g, h, i, j, ...) (\ b 31 drivers/firmware/qcom_scm-64.c (((b) & 0x3) << 6) | \ b 453 drivers/firmware/qcom_scm.c int ret, i, b; b 470 drivers/firmware/qcom_scm.c for_each_set_bit(b, &srcvm_bits, BITS_PER_LONG) b 471 drivers/firmware/qcom_scm.c src[i++] = cpu_to_le32(b); b 150 drivers/gpio/gpio-pxa.c #define for_each_gpio_bank(i, b, pc) \ b 151 drivers/gpio/gpio-pxa.c for (i = 0, b = pc->banks; i <= pxa_last_gpio; i += 32, b++) b 500 drivers/gpio/gpio-pxa.c struct pxa_gpio_bank *b = gpio_to_pxabank(&pchip->chip, gpio); b 504 drivers/gpio/gpio-pxa.c b->irq_mask &= ~GPIO_bit(gpio); b 418 drivers/gpio/gpio-tegra.c unsigned int b, p; b 422 drivers/gpio/gpio-tegra.c for (b = 0; b < tgi->bank_count; b++) { b 423 drivers/gpio/gpio-tegra.c struct tegra_gpio_bank *bank = &tgi->bank_info[b]; b 426 drivers/gpio/gpio-tegra.c unsigned int gpio = (b << 5) | (p << 3); b 457 drivers/gpio/gpio-tegra.c unsigned int b, p; b 460 drivers/gpio/gpio-tegra.c for (b = 0; b < tgi->bank_count; b++) { b 461 drivers/gpio/gpio-tegra.c struct tegra_gpio_bank *bank = &tgi->bank_info[b]; b 464 drivers/gpio/gpio-tegra.c unsigned int gpio = (b << 5) | (p << 3); b 1145 drivers/gpu/drm/amd/amdgpu/amdgpu.h #define amdgpu_asic_read_bios_from_rom(adev, b, l) (adev)->asic_funcs->read_bios_from_rom((adev), (b), (l)) b 743 drivers/gpu/drm/amd/amdgpu/amdgpu_display.c fixed20_12 a, b; b 745 drivers/gpu/drm/amd/amdgpu/amdgpu_display.c b.full = dfixed_const(dst_v); b 746 drivers/gpu/drm/amd/amdgpu/amdgpu_display.c amdgpu_crtc->vsc.full = dfixed_div(a, b); b 748 drivers/gpu/drm/amd/amdgpu/amdgpu_display.c b.full = dfixed_const(dst_h); b 749 drivers/gpu/drm/amd/amdgpu/amdgpu_display.c amdgpu_crtc->hsc.full = dfixed_div(a, b); b 250 drivers/gpu/drm/amd/amdgpu/amdgpu_ring.h #define amdgpu_ring_emit_tmz(r, b) (r)->funcs->emit_tmz((r), (b)) b 100 drivers/gpu/drm/amd/amdgpu/amdgpu_sdma.h #define amdgpu_emit_copy_buffer(adev, ib, s, d, b) (adev)->mman.buffer_funcs->emit_copy_buffer((ib), (s), (d), (b)) b 101 drivers/gpu/drm/amd/amdgpu/amdgpu_sdma.h #define amdgpu_emit_fill_buffer(adev, ib, s, d, b) (adev)->mman.buffer_funcs->emit_fill_buffer((ib), (s), (d), (b)) b 806 drivers/gpu/drm/amd/amdgpu/dce_v10_0.c fixed20_12 a, b; b 812 drivers/gpu/drm/amd/amdgpu/dce_v10_0.c b.full = dfixed_mul(a, disp_clk); b 818 drivers/gpu/drm/amd/amdgpu/dce_v10_0.c bandwidth.full = dfixed_mul(b, disp_clk_request_efficiency); b 897 drivers/gpu/drm/amd/amdgpu/dce_v10_0.c fixed20_12 a, b, c; b 903 drivers/gpu/drm/amd/amdgpu/dce_v10_0.c b.full = dfixed_const(1); b 905 drivers/gpu/drm/amd/amdgpu/dce_v10_0.c ((wm->vsc.full > b.full) && (wm->vtaps >= 3)) || b 913 drivers/gpu/drm/amd/amdgpu/dce_v10_0.c b.full = dfixed_const(wm->num_heads); b 914 drivers/gpu/drm/amd/amdgpu/dce_v10_0.c a.full = dfixed_div(a, b); b 921 drivers/gpu/drm/amd/amdgpu/dce_v10_0.c b.full = dfixed_const(1000); b 923 drivers/gpu/drm/amd/amdgpu/dce_v10_0.c b.full = dfixed_div(c, b); b 924 drivers/gpu/drm/amd/amdgpu/dce_v10_0.c a.full = dfixed_div(a, b); b 2097 drivers/gpu/drm/amd/amdgpu/dce_v10_0.c u16 *r, *g, *b; b 2137 drivers/gpu/drm/amd/amdgpu/dce_v10_0.c b = g + crtc->gamma_size; b 2142 drivers/gpu/drm/amd/amdgpu/dce_v10_0.c (*b++ >> 6)); b 832 drivers/gpu/drm/amd/amdgpu/dce_v11_0.c fixed20_12 a, b; b 838 drivers/gpu/drm/amd/amdgpu/dce_v11_0.c b.full = dfixed_mul(a, disp_clk); b 844 drivers/gpu/drm/amd/amdgpu/dce_v11_0.c bandwidth.full = dfixed_mul(b, disp_clk_request_efficiency); b 923 drivers/gpu/drm/amd/amdgpu/dce_v11_0.c fixed20_12 a, b, c; b 929 drivers/gpu/drm/amd/amdgpu/dce_v11_0.c b.full = dfixed_const(1); b 931 drivers/gpu/drm/amd/amdgpu/dce_v11_0.c ((wm->vsc.full > b.full) && (wm->vtaps >= 3)) || b 939 drivers/gpu/drm/amd/amdgpu/dce_v11_0.c b.full = dfixed_const(wm->num_heads); b 940 drivers/gpu/drm/amd/amdgpu/dce_v11_0.c a.full = dfixed_div(a, b); b 947 drivers/gpu/drm/amd/amdgpu/dce_v11_0.c b.full = dfixed_const(1000); b 949 drivers/gpu/drm/amd/amdgpu/dce_v11_0.c b.full = dfixed_div(c, b); b 950 drivers/gpu/drm/amd/amdgpu/dce_v11_0.c a.full = dfixed_div(a, b); b 2139 drivers/gpu/drm/amd/amdgpu/dce_v11_0.c u16 *r, *g, *b; b 2173 drivers/gpu/drm/amd/amdgpu/dce_v11_0.c b = g + crtc->gamma_size; b 2178 drivers/gpu/drm/amd/amdgpu/dce_v11_0.c (*b++ >> 6)); b 605 drivers/gpu/drm/amd/amdgpu/dce_v6_0.c fixed20_12 a, b; b 611 drivers/gpu/drm/amd/amdgpu/dce_v6_0.c b.full = dfixed_mul(a, disp_clk); b 617 drivers/gpu/drm/amd/amdgpu/dce_v6_0.c bandwidth.full = dfixed_mul(b, disp_clk_request_efficiency); b 696 drivers/gpu/drm/amd/amdgpu/dce_v6_0.c fixed20_12 a, b, c; b 702 drivers/gpu/drm/amd/amdgpu/dce_v6_0.c b.full = dfixed_const(1); b 704 drivers/gpu/drm/amd/amdgpu/dce_v6_0.c ((wm->vsc.full > b.full) && (wm->vtaps >= 3)) || b 712 drivers/gpu/drm/amd/amdgpu/dce_v6_0.c b.full = dfixed_const(wm->num_heads); b 713 drivers/gpu/drm/amd/amdgpu/dce_v6_0.c a.full = dfixed_div(a, b); b 720 drivers/gpu/drm/amd/amdgpu/dce_v6_0.c b.full = dfixed_const(1000); b 722 drivers/gpu/drm/amd/amdgpu/dce_v6_0.c b.full = dfixed_div(c, b); b 723 drivers/gpu/drm/amd/amdgpu/dce_v6_0.c a.full = dfixed_div(a, b); b 833 drivers/gpu/drm/amd/amdgpu/dce_v6_0.c fixed20_12 a, b, c; b 925 drivers/gpu/drm/amd/amdgpu/dce_v6_0.c b.full = dfixed_const(mode->clock); b 926 drivers/gpu/drm/amd/amdgpu/dce_v6_0.c b.full = dfixed_div(b, a); b 928 drivers/gpu/drm/amd/amdgpu/dce_v6_0.c c.full = dfixed_mul(c, b); b 937 drivers/gpu/drm/amd/amdgpu/dce_v6_0.c b.full = dfixed_const(mode->clock); b 938 drivers/gpu/drm/amd/amdgpu/dce_v6_0.c b.full = dfixed_div(b, a); b 940 drivers/gpu/drm/amd/amdgpu/dce_v6_0.c c.full = dfixed_mul(c, b); b 2032 drivers/gpu/drm/amd/amdgpu/dce_v6_0.c u16 *r, *g, *b; b 2064 drivers/gpu/drm/amd/amdgpu/dce_v6_0.c b = g + crtc->gamma_size; b 2069 drivers/gpu/drm/amd/amdgpu/dce_v6_0.c (*b++ >> 6)); b 741 drivers/gpu/drm/amd/amdgpu/dce_v8_0.c fixed20_12 a, b; b 747 drivers/gpu/drm/amd/amdgpu/dce_v8_0.c b.full = dfixed_mul(a, disp_clk); b 753 drivers/gpu/drm/amd/amdgpu/dce_v8_0.c bandwidth.full = dfixed_mul(b, disp_clk_request_efficiency); b 832 drivers/gpu/drm/amd/amdgpu/dce_v8_0.c fixed20_12 a, b, c; b 838 drivers/gpu/drm/amd/amdgpu/dce_v8_0.c b.full = dfixed_const(1); b 840 drivers/gpu/drm/amd/amdgpu/dce_v8_0.c ((wm->vsc.full > b.full) && (wm->vtaps >= 3)) || b 848 drivers/gpu/drm/amd/amdgpu/dce_v8_0.c b.full = dfixed_const(wm->num_heads); b 849 drivers/gpu/drm/amd/amdgpu/dce_v8_0.c a.full = dfixed_div(a, b); b 856 drivers/gpu/drm/amd/amdgpu/dce_v8_0.c b.full = dfixed_const(1000); b 858 drivers/gpu/drm/amd/amdgpu/dce_v8_0.c b.full = dfixed_div(c, b); b 859 drivers/gpu/drm/amd/amdgpu/dce_v8_0.c a.full = dfixed_div(a, b); b 2004 drivers/gpu/drm/amd/amdgpu/dce_v8_0.c u16 *r, *g, *b; b 2036 drivers/gpu/drm/amd/amdgpu/dce_v8_0.c b = g + crtc->gamma_size; b 2041 drivers/gpu/drm/amd/amdgpu/dce_v8_0.c (*b++ >> 6)); b 343 drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c #define AMDGPU_RAS_SUB_BLOCK(subblock, a, b, c, d, e, f, g, h) \ b 347 drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c ((a) | ((b) << 1) | ((c) << 2) | ((d) << 3)), \ b 871 drivers/gpu/drm/amd/amdgpu/kv_dpm.c static u32 kv_get_clock_difference(u32 a, u32 b) b 873 drivers/gpu/drm/amd/amdgpu/kv_dpm.c return (a >= b) ? a - b : b - a; b 1917 drivers/gpu/drm/amd/amdgpu/sid.h #define DMA_PACKET(cmd, b, t, s, n) ((((cmd) & 0xF) << 28) | \ b 1918 drivers/gpu/drm/amd/amdgpu/sid.h (((b) & 0x1) << 26) | \ b 129 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_color.c uint32_t r, g, b; b 136 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_color.c b = drm_color_lut_extract(lut[i].blue, 16); b 140 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_color.c gamma->entries.blue[i] = dc_fixpt_from_int(b); b 149 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_color.c b = drm_color_lut_extract(lut[i].blue, 16); b 153 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_color.c gamma->entries.blue[i] = dc_fixpt_from_fraction(b, MAX_DRM_LUT_VALUE); b 132 drivers/gpu/drm/amd/display/dc/calcs/dcn_calc_math.c float dcn_bw_log(float a, float b) b 143 drivers/gpu/drm/amd/display/dc/calcs/dcn_calc_math.c if (b > 2.00001 || b < 1.99999) b 144 drivers/gpu/drm/amd/display/dc/calcs/dcn_calc_math.c return (a + log_2) / dcn_bw_log(b, 2); b 553 drivers/gpu/drm/amd/display/dc/calcs/dcn_calcs.c context->bw_ctx.bw.dcn.watermarks.b.cstate_pstate.cstate_exit_ns = b 555 drivers/gpu/drm/amd/display/dc/calcs/dcn_calcs.c context->bw_ctx.bw.dcn.watermarks.b.cstate_pstate.cstate_enter_plus_exit_ns = b 557 drivers/gpu/drm/amd/display/dc/calcs/dcn_calcs.c context->bw_ctx.bw.dcn.watermarks.b.cstate_pstate.pstate_change_ns = b 559 drivers/gpu/drm/amd/display/dc/calcs/dcn_calcs.c context->bw_ctx.bw.dcn.watermarks.b.pte_meta_urgent_ns = v->ptemeta_urgent_watermark * 1000; b 560 drivers/gpu/drm/amd/display/dc/calcs/dcn_calcs.c context->bw_ctx.bw.dcn.watermarks.b.urgent_ns = v->urgent_watermark * 1000; b 617 drivers/gpu/drm/amd/display/dc/calcs/dcn_calcs.c context->bw_ctx.bw.dcn.watermarks.b = context->bw_ctx.bw.dcn.watermarks.a; b 1125 drivers/gpu/drm/amd/display/dc/calcs/dcn_calcs.c context->bw_ctx.bw.dcn.watermarks.b = context->bw_ctx.bw.dcn.watermarks.a; b 167 drivers/gpu/drm/amd/display/dc/core/dc_link_ddc.c #define DDC_MIN(a, b) (((a) < (b)) ? (a) : (b)) b 337 drivers/gpu/drm/amd/display/dc/dcn10/dcn10_hubbub.c if (safe_to_lower || watermarks->b.urgent_ns > hubbub1->watermarks.b.urgent_ns) { b 338 drivers/gpu/drm/amd/display/dc/dcn10/dcn10_hubbub.c hubbub1->watermarks.b.urgent_ns = watermarks->b.urgent_ns; b 339 drivers/gpu/drm/amd/display/dc/dcn10/dcn10_hubbub.c prog_wm_value = convert_and_clamp(watermarks->b.urgent_ns, b 346 drivers/gpu/drm/amd/display/dc/dcn10/dcn10_hubbub.c watermarks->b.urgent_ns, prog_wm_value); b 349 drivers/gpu/drm/amd/display/dc/dcn10/dcn10_hubbub.c if (safe_to_lower || watermarks->b.pte_meta_urgent_ns > hubbub1->watermarks.b.pte_meta_urgent_ns) { b 350 drivers/gpu/drm/amd/display/dc/dcn10/dcn10_hubbub.c hubbub1->watermarks.b.pte_meta_urgent_ns = watermarks->b.pte_meta_urgent_ns; b 351 drivers/gpu/drm/amd/display/dc/dcn10/dcn10_hubbub.c prog_wm_value = convert_and_clamp(watermarks->b.pte_meta_urgent_ns, b 356 drivers/gpu/drm/amd/display/dc/dcn10/dcn10_hubbub.c watermarks->b.pte_meta_urgent_ns, prog_wm_value); b 445 drivers/gpu/drm/amd/display/dc/dcn10/dcn10_hubbub.c if (safe_to_lower || watermarks->b.cstate_pstate.cstate_enter_plus_exit_ns b 446 drivers/gpu/drm/amd/display/dc/dcn10/dcn10_hubbub.c > hubbub1->watermarks.b.cstate_pstate.cstate_enter_plus_exit_ns) { b 447 drivers/gpu/drm/amd/display/dc/dcn10/dcn10_hubbub.c hubbub1->watermarks.b.cstate_pstate.cstate_enter_plus_exit_ns = b 448 drivers/gpu/drm/amd/display/dc/dcn10/dcn10_hubbub.c watermarks->b.cstate_pstate.cstate_enter_plus_exit_ns; b 450 drivers/gpu/drm/amd/display/dc/dcn10/dcn10_hubbub.c watermarks->b.cstate_pstate.cstate_enter_plus_exit_ns, b 456 drivers/gpu/drm/amd/display/dc/dcn10/dcn10_hubbub.c watermarks->b.cstate_pstate.cstate_enter_plus_exit_ns, prog_wm_value); b 459 drivers/gpu/drm/amd/display/dc/dcn10/dcn10_hubbub.c if (safe_to_lower || watermarks->b.cstate_pstate.cstate_exit_ns b 460 drivers/gpu/drm/amd/display/dc/dcn10/dcn10_hubbub.c > hubbub1->watermarks.b.cstate_pstate.cstate_exit_ns) { b 461 drivers/gpu/drm/amd/display/dc/dcn10/dcn10_hubbub.c hubbub1->watermarks.b.cstate_pstate.cstate_exit_ns = b 462 drivers/gpu/drm/amd/display/dc/dcn10/dcn10_hubbub.c watermarks->b.cstate_pstate.cstate_exit_ns; b 464 drivers/gpu/drm/amd/display/dc/dcn10/dcn10_hubbub.c watermarks->b.cstate_pstate.cstate_exit_ns, b 470 drivers/gpu/drm/amd/display/dc/dcn10/dcn10_hubbub.c watermarks->b.cstate_pstate.cstate_exit_ns, prog_wm_value); b 558 drivers/gpu/drm/amd/display/dc/dcn10/dcn10_hubbub.c if (safe_to_lower || watermarks->b.cstate_pstate.pstate_change_ns b 559 drivers/gpu/drm/amd/display/dc/dcn10/dcn10_hubbub.c > hubbub1->watermarks.b.cstate_pstate.pstate_change_ns) { b 560 drivers/gpu/drm/amd/display/dc/dcn10/dcn10_hubbub.c hubbub1->watermarks.b.cstate_pstate.pstate_change_ns = b 561 drivers/gpu/drm/amd/display/dc/dcn10/dcn10_hubbub.c watermarks->b.cstate_pstate.pstate_change_ns; b 563 drivers/gpu/drm/amd/display/dc/dcn10/dcn10_hubbub.c watermarks->b.cstate_pstate.pstate_change_ns, b 569 drivers/gpu/drm/amd/display/dc/dcn10/dcn10_hubbub.c watermarks->b.cstate_pstate.pstate_change_ns, prog_wm_value); b 2696 drivers/gpu/drm/amd/display/dc/dcn20/dcn20_resource.c context->bw_ctx.bw.dcn.watermarks.b.urgent_ns = get_wm_urgent(&context->bw_ctx.dml, pipes, pipe_cnt) * 1000; b 2697 drivers/gpu/drm/amd/display/dc/dcn20/dcn20_resource.c context->bw_ctx.bw.dcn.watermarks.b.cstate_pstate.cstate_enter_plus_exit_ns = get_wm_stutter_enter_exit(&context->bw_ctx.dml, pipes, pipe_cnt) * 1000; b 2698 drivers/gpu/drm/amd/display/dc/dcn20/dcn20_resource.c context->bw_ctx.bw.dcn.watermarks.b.cstate_pstate.cstate_exit_ns = get_wm_stutter_exit(&context->bw_ctx.dml, pipes, pipe_cnt) * 1000; b 2699 drivers/gpu/drm/amd/display/dc/dcn20/dcn20_resource.c context->bw_ctx.bw.dcn.watermarks.b.cstate_pstate.pstate_change_ns = get_wm_dram_clock_change(&context->bw_ctx.dml, pipes, pipe_cnt) * 1000; b 2700 drivers/gpu/drm/amd/display/dc/dcn20/dcn20_resource.c context->bw_ctx.bw.dcn.watermarks.b.pte_meta_urgent_ns = get_wm_memory_trip(&context->bw_ctx.dml, pipes, pipe_cnt) * 1000; b 175 drivers/gpu/drm/amd/display/dc/dcn21/dcn21_hubbub.c if (safe_to_lower || watermarks->b.urgent_ns > hubbub1->watermarks.b.urgent_ns) { b 176 drivers/gpu/drm/amd/display/dc/dcn21/dcn21_hubbub.c hubbub1->watermarks.b.urgent_ns = watermarks->b.urgent_ns; b 177 drivers/gpu/drm/amd/display/dc/dcn21/dcn21_hubbub.c prog_wm_value = convert_and_clamp(watermarks->b.urgent_ns, b 185 drivers/gpu/drm/amd/display/dc/dcn21/dcn21_hubbub.c watermarks->b.urgent_ns, prog_wm_value); b 309 drivers/gpu/drm/amd/display/dc/dcn21/dcn21_hubbub.c if (safe_to_lower || watermarks->b.cstate_pstate.cstate_enter_plus_exit_ns b 310 drivers/gpu/drm/amd/display/dc/dcn21/dcn21_hubbub.c > hubbub1->watermarks.b.cstate_pstate.cstate_enter_plus_exit_ns) { b 311 drivers/gpu/drm/amd/display/dc/dcn21/dcn21_hubbub.c hubbub1->watermarks.b.cstate_pstate.cstate_enter_plus_exit_ns = b 312 drivers/gpu/drm/amd/display/dc/dcn21/dcn21_hubbub.c watermarks->b.cstate_pstate.cstate_enter_plus_exit_ns; b 314 drivers/gpu/drm/amd/display/dc/dcn21/dcn21_hubbub.c watermarks->b.cstate_pstate.cstate_enter_plus_exit_ns, b 321 drivers/gpu/drm/amd/display/dc/dcn21/dcn21_hubbub.c watermarks->b.cstate_pstate.cstate_enter_plus_exit_ns, prog_wm_value); b 324 drivers/gpu/drm/amd/display/dc/dcn21/dcn21_hubbub.c if (safe_to_lower || watermarks->b.cstate_pstate.cstate_exit_ns b 325 drivers/gpu/drm/amd/display/dc/dcn21/dcn21_hubbub.c > hubbub1->watermarks.b.cstate_pstate.cstate_exit_ns) { b 326 drivers/gpu/drm/amd/display/dc/dcn21/dcn21_hubbub.c hubbub1->watermarks.b.cstate_pstate.cstate_exit_ns = b 327 drivers/gpu/drm/amd/display/dc/dcn21/dcn21_hubbub.c watermarks->b.cstate_pstate.cstate_exit_ns; b 329 drivers/gpu/drm/amd/display/dc/dcn21/dcn21_hubbub.c watermarks->b.cstate_pstate.cstate_exit_ns, b 336 drivers/gpu/drm/amd/display/dc/dcn21/dcn21_hubbub.c watermarks->b.cstate_pstate.cstate_exit_ns, prog_wm_value); b 428 drivers/gpu/drm/amd/display/dc/dcn21/dcn21_hubbub.c if (safe_to_lower || watermarks->b.cstate_pstate.pstate_change_ns b 429 drivers/gpu/drm/amd/display/dc/dcn21/dcn21_hubbub.c > hubbub1->watermarks.b.cstate_pstate.pstate_change_ns) { b 430 drivers/gpu/drm/amd/display/dc/dcn21/dcn21_hubbub.c hubbub1->watermarks.b.cstate_pstate.pstate_change_ns = b 431 drivers/gpu/drm/amd/display/dc/dcn21/dcn21_hubbub.c watermarks->b.cstate_pstate.pstate_change_ns; b 433 drivers/gpu/drm/amd/display/dc/dcn21/dcn21_hubbub.c watermarks->b.cstate_pstate.pstate_change_ns, b 440 drivers/gpu/drm/amd/display/dc/dcn21/dcn21_hubbub.c watermarks->b.cstate_pstate.pstate_change_ns, prog_wm_value); b 1051 drivers/gpu/drm/amd/display/dc/dcn21/dcn21_resource.c calculate_wm_set_for_vlevel(vlevel, table_entry, &context->bw_ctx.bw.dcn.watermarks.b, b 33 drivers/gpu/drm/amd/display/dc/dml/dml_inline_defs.h static inline double dml_min(double a, double b) b 35 drivers/gpu/drm/amd/display/dc/dml/dml_inline_defs.h return (double) dcn_bw_min2(a, b); b 38 drivers/gpu/drm/amd/display/dc/dml/dml_inline_defs.h static inline double dml_min3(double a, double b, double c) b 40 drivers/gpu/drm/amd/display/dc/dml/dml_inline_defs.h return dml_min(dml_min(a, b), c); b 43 drivers/gpu/drm/amd/display/dc/dml/dml_inline_defs.h static inline double dml_min4(double a, double b, double c, double d) b 45 drivers/gpu/drm/amd/display/dc/dml/dml_inline_defs.h return dml_min(dml_min(a, b), dml_min(c, d)); b 48 drivers/gpu/drm/amd/display/dc/dml/dml_inline_defs.h static inline double dml_max(double a, double b) b 50 drivers/gpu/drm/amd/display/dc/dml/dml_inline_defs.h return (double) dcn_bw_max2(a, b); b 53 drivers/gpu/drm/amd/display/dc/dml/dml_inline_defs.h static inline double dml_max3(double a, double b, double c) b 55 drivers/gpu/drm/amd/display/dc/dml/dml_inline_defs.h return dml_max(dml_max(a, b), c); b 58 drivers/gpu/drm/amd/display/dc/dml/dml_inline_defs.h static inline double dml_max4(double a, double b, double c, double d) b 60 drivers/gpu/drm/amd/display/dc/dml/dml_inline_defs.h return dml_max(dml_max(a, b), dml_max(c, d)); b 63 drivers/gpu/drm/amd/display/dc/dml/dml_inline_defs.h static inline double dml_max5(double a, double b, double c, double d, double e) b 65 drivers/gpu/drm/amd/display/dc/dml/dml_inline_defs.h return dml_max(dml_max4(a, b, c, d), e); b 164 drivers/gpu/drm/amd/display/dc/dsc/rc_calc.c int median3(int a, int b, int c) b 166 drivers/gpu/drm/amd/display/dc/dsc/rc_calc.c if (a > b) b 167 drivers/gpu/drm/amd/display/dc/dsc/rc_calc.c swap(a, b); b 168 drivers/gpu/drm/amd/display/dc/dsc/rc_calc.c if (b > c) b 169 drivers/gpu/drm/amd/display/dc/dsc/rc_calc.c swap(b, c); b 170 drivers/gpu/drm/amd/display/dc/dsc/rc_calc.c if (a > b) b 171 drivers/gpu/drm/amd/display/dc/dsc/rc_calc.c swap(b, c); b 173 drivers/gpu/drm/amd/display/dc/dsc/rc_calc.c return b; b 40 drivers/gpu/drm/amd/display/dc/inc/dcn_calc_math.h float dcn_bw_log(float a, float b); b 52 drivers/gpu/drm/amd/display/dc/inc/hw/mem_input.h struct dcn_watermarks b; b 120 drivers/gpu/drm/amd/display/dc/inc/hw/opp.h struct fixed31_32 b; b 154 drivers/gpu/drm/amd/display/dc/inc/hw/opp.h struct fixed31_32 b; b 179 drivers/gpu/drm/amd/display/dc/inc/hw/opp.h struct gamma_point b; b 195 drivers/gpu/drm/amd/display/dc/inc/hw/opp.h struct fixed31_32 b; b 183 drivers/gpu/drm/amd/display/modules/color/color_gamma.c struct fixed31_32 b; b 191 drivers/gpu/drm/amd/display/modules/color/color_gamma.c b = dc_fixpt_from_fraction(28466892, 100000000); b 202 drivers/gpu/drm/amd/display/modules/color/color_gamma.c x = dc_fixpt_add(x, b); b 214 drivers/gpu/drm/amd/display/modules/color/color_gamma.c struct fixed31_32 b; b 222 drivers/gpu/drm/amd/display/modules/color/color_gamma.c b = dc_fixpt_from_fraction(28466892, 100000000); b 233 drivers/gpu/drm/amd/display/modules/color/color_gamma.c x = dc_fixpt_sub(x, b); b 486 drivers/gpu/drm/amd/display/modules/color/color_gamma.c left = axis_x[i].b; b 489 drivers/gpu/drm/amd/display/modules/color/color_gamma.c right = axis_x[i + 1].b; b 491 drivers/gpu/drm/amd/display/modules/color/color_gamma.c right = axis_x[max_number - 1].b; b 592 drivers/gpu/drm/amd/display/modules/color/color_gamma.c point = &coeff[i].b; b 594 drivers/gpu/drm/amd/display/modules/color/color_gamma.c left_pos = axis_x[index_left].b; b 595 drivers/gpu/drm/amd/display/modules/color/color_gamma.c right_pos = axis_x[index_right].b; b 640 drivers/gpu/drm/amd/display/modules/color/color_gamma.c point = &coeff->b; b 674 drivers/gpu/drm/amd/display/modules/color/color_gamma.c rgb[point->right_index].b, b 675 drivers/gpu/drm/amd/display/modules/color/color_gamma.c rgb[point->left_index].b)), b 676 drivers/gpu/drm/amd/display/modules/color/color_gamma.c rgb[point->left_index].b); b 733 drivers/gpu/drm/amd/display/modules/color/color_gamma.c rgb->b = output; b 764 drivers/gpu/drm/amd/display/modules/color/color_gamma.c de_pq[i].b = output; b 794 drivers/gpu/drm/amd/display/modules/color/color_gamma.c rgb->b = rgb->r; b 827 drivers/gpu/drm/amd/display/modules/color/color_gamma.c struct fixed31_32 b = dc_fixpt_from_fraction(5, 10); b 839 drivers/gpu/drm/amd/display/modules/color/color_gamma.c a = dc_fixpt_div(dc_fixpt_add(dc_fixpt_one, b), max_content_pq); // (1+b)/maxContent b 840 drivers/gpu/drm/amd/display/modules/color/color_gamma.c ks = dc_fixpt_sub(dc_fixpt_mul(a, max_lum_pq), b); // a * max_lum_pq - b b 957 drivers/gpu/drm/amd/display/modules/color/color_gamma.c rgb->b = output; b 962 drivers/gpu/drm/amd/display/modules/color/color_gamma.c rgb->b = clip; b 967 drivers/gpu/drm/amd/display/modules/color/color_gamma.c rgb->b = clip; b 1001 drivers/gpu/drm/amd/display/modules/color/color_gamma.c curve[i].b = dc_fixpt_zero; b 1009 drivers/gpu/drm/amd/display/modules/color/color_gamma.c curve[i].b = curve[i].r; b 1015 drivers/gpu/drm/amd/display/modules/color/color_gamma.c curve[i].b = dc_fixpt_one; b 1042 drivers/gpu/drm/amd/display/modules/color/color_gamma.c rgb->b = rgb->r; b 1066 drivers/gpu/drm/amd/display/modules/color/color_gamma.c rgb->b = rgb->r; b 1103 drivers/gpu/drm/amd/display/modules/color/color_gamma.c rgb->b = dc_fixpt_div( b 1114 drivers/gpu/drm/amd/display/modules/color/color_gamma.c rgb->b = dc_fixpt_mul(rgb_last->b, b 1123 drivers/gpu/drm/amd/display/modules/color/color_gamma.c rgb->b = dc_fixpt_mul(rgb_last->b, b 1132 drivers/gpu/drm/amd/display/modules/color/color_gamma.c rgb->b = dc_fixpt_mul(rgb_last->b, b 1179 drivers/gpu/drm/amd/display/modules/color/color_gamma.c pwl_rgb[i].b = dc_fixpt_div( b 1189 drivers/gpu/drm/amd/display/modules/color/color_gamma.c pwl_rgb[i].b = dc_fixpt_sub(dc_fixpt_mul_int( b 1190 drivers/gpu/drm/amd/display/modules/color/color_gamma.c pwl_rgb[i-1].b, 2), pwl_rgb[i-2].b); b 1196 drivers/gpu/drm/amd/display/modules/color/color_gamma.c pwl_rgb[i].b = dc_fixpt_sub(dc_fixpt_mul_int( b 1197 drivers/gpu/drm/amd/display/modules/color/color_gamma.c pwl_rgb[i-1].b, 2), pwl_rgb[i-2].b); b 1232 drivers/gpu/drm/amd/display/modules/color/color_gamma.c rgb->b = dc_fixpt_from_fraction( b 1243 drivers/gpu/drm/amd/display/modules/color/color_gamma.c rgb->b = dc_fixpt_mul(rgb_last->b, b 1252 drivers/gpu/drm/amd/display/modules/color/color_gamma.c rgb->b = dc_fixpt_mul(rgb_last->b, b 1261 drivers/gpu/drm/amd/display/modules/color/color_gamma.c rgb->b = dc_fixpt_mul(rgb_last->b, b 1366 drivers/gpu/drm/amd/display/modules/color/color_gamma.c p->b = value; b 1374 drivers/gpu/drm/amd/display/modules/color/color_gamma.c p->b = dc_fixpt_div(p_last->b, dividers.divider1); b 1380 drivers/gpu/drm/amd/display/modules/color/color_gamma.c p->b = dc_fixpt_div(p_last->b, dividers.divider2); b 1386 drivers/gpu/drm/amd/display/modules/color/color_gamma.c p->b = dc_fixpt_div(p_last->b, dividers.divider3); b 1401 drivers/gpu/drm/amd/display/modules/color/color_gamma.c coords->regamma_y_blue = rgb_regamma->b; b 1532 drivers/gpu/drm/amd/display/modules/color/color_gamma.c lut1 = rgb_user[index].b; b 1533 drivers/gpu/drm/amd/display/modules/color/color_gamma.c lut2 = rgb_user[index_next].b; b 1586 drivers/gpu/drm/amd/display/modules/color/color_gamma.c rgb->b = rgb->r; b 1623 drivers/gpu/drm/amd/display/modules/color/color_gamma.c tf_pts->blue[i] = regamma->b; b 1957 drivers/gpu/drm/amd/display/modules/color/color_gamma.c curve[i].b = curve[i].r; b 2031 drivers/gpu/drm/amd/display/modules/color/color_gamma.c points->blue[i] = rgb_regamma[i].b; b 2058 drivers/gpu/drm/amd/display/modules/color/color_gamma.c points->blue[i] = rgb_regamma[i].b; b 2081 drivers/gpu/drm/amd/display/modules/color/color_gamma.c points->blue[i] = rgb_regamma[i].b; b 2121 drivers/gpu/drm/amd/display/modules/color/color_gamma.c points->blue[i] = rgb_degamma[i].b; b 2144 drivers/gpu/drm/amd/display/modules/color/color_gamma.c points->blue[i] = rgb_degamma[i].b; b 2163 drivers/gpu/drm/amd/display/modules/color/color_gamma.c points->blue[i] = rgb_degamma[i].b; b 28 drivers/gpu/drm/amd/display/modules/power/power_helpers.c #define DIV_ROUNDUP(a, b) (((a)+((b)/2))/(b)) b 1526 drivers/gpu/drm/amd/powerplay/arcturus_ppt.c pptable->DpmDescriptor[PPCLK_GFXCLK].ConversionToAvfsClk.b, b 1528 drivers/gpu/drm/amd/powerplay/arcturus_ppt.c pptable->DpmDescriptor[PPCLK_GFXCLK].SsCurve.b, b 1547 drivers/gpu/drm/amd/powerplay/arcturus_ppt.c pptable->DpmDescriptor[PPCLK_VCLK].ConversionToAvfsClk.b, b 1549 drivers/gpu/drm/amd/powerplay/arcturus_ppt.c pptable->DpmDescriptor[PPCLK_VCLK].SsCurve.b, b 1568 drivers/gpu/drm/amd/powerplay/arcturus_ppt.c pptable->DpmDescriptor[PPCLK_DCLK].ConversionToAvfsClk.b, b 1570 drivers/gpu/drm/amd/powerplay/arcturus_ppt.c pptable->DpmDescriptor[PPCLK_DCLK].SsCurve.b, b 1589 drivers/gpu/drm/amd/powerplay/arcturus_ppt.c pptable->DpmDescriptor[PPCLK_SOCCLK].ConversionToAvfsClk.b, b 1591 drivers/gpu/drm/amd/powerplay/arcturus_ppt.c pptable->DpmDescriptor[PPCLK_SOCCLK].SsCurve.b, b 1610 drivers/gpu/drm/amd/powerplay/arcturus_ppt.c pptable->DpmDescriptor[PPCLK_UCLK].ConversionToAvfsClk.b, b 1612 drivers/gpu/drm/amd/powerplay/arcturus_ppt.c pptable->DpmDescriptor[PPCLK_UCLK].SsCurve.b, b 1631 drivers/gpu/drm/amd/powerplay/arcturus_ppt.c pptable->DpmDescriptor[PPCLK_FCLK].ConversionToAvfsClk.b, b 1633 drivers/gpu/drm/amd/powerplay/arcturus_ppt.c pptable->DpmDescriptor[PPCLK_FCLK].SsCurve.b, b 1718 drivers/gpu/drm/amd/powerplay/arcturus_ppt.c pptable->dBtcGbGfxPll.b, b 1722 drivers/gpu/drm/amd/powerplay/arcturus_ppt.c pptable->dBtcGbGfxAfll.b, b 1726 drivers/gpu/drm/amd/powerplay/arcturus_ppt.c pptable->dBtcGbSoc.b, b 1731 drivers/gpu/drm/amd/powerplay/arcturus_ppt.c pptable->qAgingGb[AVFS_VOLTAGE_GFX].b); b 1734 drivers/gpu/drm/amd/powerplay/arcturus_ppt.c pptable->qAgingGb[AVFS_VOLTAGE_SOC].b); b 1738 drivers/gpu/drm/amd/powerplay/arcturus_ppt.c pptable->qStaticVoltageOffset[AVFS_VOLTAGE_GFX].b, b 1742 drivers/gpu/drm/amd/powerplay/arcturus_ppt.c pptable->qStaticVoltageOffset[AVFS_VOLTAGE_SOC].b, b 1779 drivers/gpu/drm/amd/powerplay/arcturus_ppt.c pptable->ReservedEquation0.b, b 1783 drivers/gpu/drm/amd/powerplay/arcturus_ppt.c pptable->ReservedEquation1.b, b 1787 drivers/gpu/drm/amd/powerplay/arcturus_ppt.c pptable->ReservedEquation2.b, b 1791 drivers/gpu/drm/amd/powerplay/arcturus_ppt.c pptable->ReservedEquation3.b, b 2157 drivers/gpu/drm/amd/powerplay/hwmgr/vega10_hwmgr.c pp_table->AvfsGbCksOn.b = b 2169 drivers/gpu/drm/amd/powerplay/hwmgr/vega10_hwmgr.c pp_table->AvfsGbCksOff.b = b 2187 drivers/gpu/drm/amd/powerplay/hwmgr/vega10_hwmgr.c pp_table->DisplayClock2Gfxclk[DSPCLK_DISPCLK].b = b 2194 drivers/gpu/drm/amd/powerplay/hwmgr/vega10_hwmgr.c pp_table->DisplayClock2Gfxclk[DSPCLK_DISPCLK].b = b 2210 drivers/gpu/drm/amd/powerplay/hwmgr/vega10_hwmgr.c pp_table->DisplayClock2Gfxclk[DSPCLK_DCEFCLK].b = b 2217 drivers/gpu/drm/amd/powerplay/hwmgr/vega10_hwmgr.c pp_table->DisplayClock2Gfxclk[DSPCLK_DCEFCLK].b = b 2233 drivers/gpu/drm/amd/powerplay/hwmgr/vega10_hwmgr.c pp_table->DisplayClock2Gfxclk[DSPCLK_PIXCLK].b = b 2240 drivers/gpu/drm/amd/powerplay/hwmgr/vega10_hwmgr.c pp_table->DisplayClock2Gfxclk[DSPCLK_PIXCLK].b = b 2255 drivers/gpu/drm/amd/powerplay/hwmgr/vega10_hwmgr.c pp_table->DisplayClock2Gfxclk[DSPCLK_PHYCLK].b = b 2262 drivers/gpu/drm/amd/powerplay/hwmgr/vega10_hwmgr.c pp_table->DisplayClock2Gfxclk[DSPCLK_PHYCLK].b = b 2279 drivers/gpu/drm/amd/powerplay/hwmgr/vega10_hwmgr.c pp_table->AcgAvfsGb.b = avfs_params.ulAcgGbFuseTableB; b 136 drivers/gpu/drm/amd/powerplay/hwmgr/vega20_processpptables.c pptable->DpmDescriptor[PPCLK_GFXCLK].ConversionToAvfsClk.b, b 138 drivers/gpu/drm/amd/powerplay/hwmgr/vega20_processpptables.c pptable->DpmDescriptor[PPCLK_GFXCLK].SsCurve.b, b 153 drivers/gpu/drm/amd/powerplay/hwmgr/vega20_processpptables.c pptable->DpmDescriptor[PPCLK_VCLK].ConversionToAvfsClk.b, b 155 drivers/gpu/drm/amd/powerplay/hwmgr/vega20_processpptables.c pptable->DpmDescriptor[PPCLK_VCLK].SsCurve.b, b 170 drivers/gpu/drm/amd/powerplay/hwmgr/vega20_processpptables.c pptable->DpmDescriptor[PPCLK_DCLK].ConversionToAvfsClk.b, b 172 drivers/gpu/drm/amd/powerplay/hwmgr/vega20_processpptables.c pptable->DpmDescriptor[PPCLK_DCLK].SsCurve.b, b 187 drivers/gpu/drm/amd/powerplay/hwmgr/vega20_processpptables.c pptable->DpmDescriptor[PPCLK_ECLK].ConversionToAvfsClk.b, b 189 drivers/gpu/drm/amd/powerplay/hwmgr/vega20_processpptables.c pptable->DpmDescriptor[PPCLK_ECLK].SsCurve.b, b 204 drivers/gpu/drm/amd/powerplay/hwmgr/vega20_processpptables.c pptable->DpmDescriptor[PPCLK_SOCCLK].ConversionToAvfsClk.b, b 206 drivers/gpu/drm/amd/powerplay/hwmgr/vega20_processpptables.c pptable->DpmDescriptor[PPCLK_SOCCLK].SsCurve.b, b 221 drivers/gpu/drm/amd/powerplay/hwmgr/vega20_processpptables.c pptable->DpmDescriptor[PPCLK_UCLK].ConversionToAvfsClk.b, b 223 drivers/gpu/drm/amd/powerplay/hwmgr/vega20_processpptables.c pptable->DpmDescriptor[PPCLK_UCLK].SsCurve.b, b 238 drivers/gpu/drm/amd/powerplay/hwmgr/vega20_processpptables.c pptable->DpmDescriptor[PPCLK_DCEFCLK].ConversionToAvfsClk.b, b 240 drivers/gpu/drm/amd/powerplay/hwmgr/vega20_processpptables.c pptable->DpmDescriptor[PPCLK_DCEFCLK].SsCurve.b, b 255 drivers/gpu/drm/amd/powerplay/hwmgr/vega20_processpptables.c pptable->DpmDescriptor[PPCLK_DISPCLK].ConversionToAvfsClk.b, b 257 drivers/gpu/drm/amd/powerplay/hwmgr/vega20_processpptables.c pptable->DpmDescriptor[PPCLK_DISPCLK].SsCurve.b, b 272 drivers/gpu/drm/amd/powerplay/hwmgr/vega20_processpptables.c pptable->DpmDescriptor[PPCLK_PIXCLK].ConversionToAvfsClk.b, b 274 drivers/gpu/drm/amd/powerplay/hwmgr/vega20_processpptables.c pptable->DpmDescriptor[PPCLK_PIXCLK].SsCurve.b, b 289 drivers/gpu/drm/amd/powerplay/hwmgr/vega20_processpptables.c pptable->DpmDescriptor[PPCLK_PHYCLK].ConversionToAvfsClk.b, b 291 drivers/gpu/drm/amd/powerplay/hwmgr/vega20_processpptables.c pptable->DpmDescriptor[PPCLK_PHYCLK].SsCurve.b, b 306 drivers/gpu/drm/amd/powerplay/hwmgr/vega20_processpptables.c pptable->DpmDescriptor[PPCLK_FCLK].ConversionToAvfsClk.b, b 308 drivers/gpu/drm/amd/powerplay/hwmgr/vega20_processpptables.c pptable->DpmDescriptor[PPCLK_FCLK].SsCurve.b, b 383 drivers/gpu/drm/amd/powerplay/hwmgr/vega20_processpptables.c pptable->CksVoltageOffset.b, b 446 drivers/gpu/drm/amd/powerplay/hwmgr/vega20_processpptables.c pptable->qAvfsGb[AVFS_VOLTAGE_GFX].b, b 450 drivers/gpu/drm/amd/powerplay/hwmgr/vega20_processpptables.c pptable->qAvfsGb[AVFS_VOLTAGE_SOC].b, b 454 drivers/gpu/drm/amd/powerplay/hwmgr/vega20_processpptables.c pptable->dBtcGbGfxCksOn.b, b 458 drivers/gpu/drm/amd/powerplay/hwmgr/vega20_processpptables.c pptable->dBtcGbGfxCksOff.b, b 462 drivers/gpu/drm/amd/powerplay/hwmgr/vega20_processpptables.c pptable->dBtcGbGfxAfll.b, b 466 drivers/gpu/drm/amd/powerplay/hwmgr/vega20_processpptables.c pptable->dBtcGbSoc.b, b 470 drivers/gpu/drm/amd/powerplay/hwmgr/vega20_processpptables.c pptable->qAgingGb[AVFS_VOLTAGE_GFX].b); b 473 drivers/gpu/drm/amd/powerplay/hwmgr/vega20_processpptables.c pptable->qAgingGb[AVFS_VOLTAGE_SOC].b); b 477 drivers/gpu/drm/amd/powerplay/hwmgr/vega20_processpptables.c pptable->qStaticVoltageOffset[AVFS_VOLTAGE_GFX].b, b 481 drivers/gpu/drm/amd/powerplay/hwmgr/vega20_processpptables.c pptable->qStaticVoltageOffset[AVFS_VOLTAGE_SOC].b, b 519 drivers/gpu/drm/amd/powerplay/hwmgr/vega20_processpptables.c pptable->ReservedEquation0.b, b 523 drivers/gpu/drm/amd/powerplay/hwmgr/vega20_processpptables.c pptable->ReservedEquation1.b, b 527 drivers/gpu/drm/amd/powerplay/hwmgr/vega20_processpptables.c pptable->ReservedEquation2.b, b 531 drivers/gpu/drm/amd/powerplay/hwmgr/vega20_processpptables.c pptable->ReservedEquation3.b, b 301 drivers/gpu/drm/amd/powerplay/inc/smu11_driver_if.h uint32_t b; b 307 drivers/gpu/drm/amd/powerplay/inc/smu11_driver_if.h uint32_t b; b 312 drivers/gpu/drm/amd/powerplay/inc/smu11_driver_if.h uint32_t b; b 340 drivers/gpu/drm/amd/powerplay/inc/smu11_driver_if_arcturus.h uint32_t b; // store in IEEE float format in this variable b 346 drivers/gpu/drm/amd/powerplay/inc/smu11_driver_if_arcturus.h uint32_t b; // store in IEEE float format in this variable b 351 drivers/gpu/drm/amd/powerplay/inc/smu11_driver_if_arcturus.h uint32_t b; // store in IEEE float format in this variable b 347 drivers/gpu/drm/amd/powerplay/inc/smu11_driver_if_navi10.h uint32_t b; // store in IEEE float format in this variable b 353 drivers/gpu/drm/amd/powerplay/inc/smu11_driver_if_navi10.h uint32_t b; // store in IEEE float format in this variable b 358 drivers/gpu/drm/amd/powerplay/inc/smu11_driver_if_navi10.h uint32_t b; // store in IEEE float format in this variable b 51 drivers/gpu/drm/amd/powerplay/inc/smu72.h int32_t b; b 43 drivers/gpu/drm/amd/powerplay/inc/smu73.h int32_t b; b 582 drivers/gpu/drm/amd/powerplay/inc/smu73_discrete.h uint32_t b; b 76 drivers/gpu/drm/amd/powerplay/inc/smu74.h int32_t b; b 228 drivers/gpu/drm/amd/powerplay/inc/smu74_discrete.h uint32_t b; b 706 drivers/gpu/drm/amd/powerplay/inc/smu75.h int32_t b; b 102 drivers/gpu/drm/amd/powerplay/inc/smu9_driver_if.h int32_t b; b 203 drivers/gpu/drm/amd/powerplay/inc/vega12/smu9_driver_if.h uint32_t b; b 209 drivers/gpu/drm/amd/powerplay/inc/vega12/smu9_driver_if.h uint32_t b; b 214 drivers/gpu/drm/amd/powerplay/inc/vega12/smu9_driver_if.h uint32_t b; b 1726 drivers/gpu/drm/amd/powerplay/smumgr/polaris10_smumgr.c table->AVFSGB_VDROOP_TABLE[0].b = PP_HOST_TO_SMC_UL(avfs_params.ulAVFSGB_FUSE_TABLE_CKSON_b); b 1731 drivers/gpu/drm/amd/powerplay/smumgr/polaris10_smumgr.c table->AVFSGB_VDROOP_TABLE[1].b = PP_HOST_TO_SMC_UL(avfs_params.ulAVFSGB_FUSE_TABLE_CKSOFF_b); b 1607 drivers/gpu/drm/amd/powerplay/smumgr/vegam_smumgr.c table->AVFSGB_FUSE_TABLE[0].b = b 1615 drivers/gpu/drm/amd/powerplay/smumgr/vegam_smumgr.c table->AVFSGB_FUSE_TABLE[1].b = b 126 drivers/gpu/drm/arm/display/komeda/komeda_kms.c struct drm_plane *b = new->base.plane; b 132 drivers/gpu/drm/arm/display/komeda/komeda_kms.c a->name, b->name, node->base.zpos); b 122 drivers/gpu/drm/arm/malidp_crtc.c #define DE_COEFTAB_DATA(a, b) ((((a) & 0xfff) << 16) | (((b) & 0xfff))) b 131 drivers/gpu/drm/arm/malidp_crtc.c u32 a, b, delta_in, out_start, out_end; b 139 drivers/gpu/drm/arm/malidp_crtc.c b = out_start; b 140 drivers/gpu/drm/arm/malidp_crtc.c coeffs[i] = DE_COEFTAB_DATA(a, b); b 418 drivers/gpu/drm/armada/armada_overlay.c #define R2BGR(r,g,b,s) (C2K(r,s) << 0 | C2K(g,s) << 8 | C2K(b,s) << 16) b 71 drivers/gpu/drm/ast/ast_mode.c u16 *r, *g, *b; b 79 drivers/gpu/drm/ast/ast_mode.c b = g + crtc->gamma_size; b 82 drivers/gpu/drm/ast/ast_mode.c ast_load_palette_index(ast, i, *r++ >> 8, *g++ >> 8, *b++ >> 8); b 1100 drivers/gpu/drm/ast/ast_mode.c u8 b[4]; b 1104 drivers/gpu/drm/ast/ast_mode.c u8 b[2]; b 1124 drivers/gpu/drm/ast/ast_mode.c data32.b[0] = srcdata32[0].b[1] | (srcdata32[0].b[0] >> 4); b 1125 drivers/gpu/drm/ast/ast_mode.c data32.b[1] = srcdata32[0].b[3] | (srcdata32[0].b[2] >> 4); b 1126 drivers/gpu/drm/ast/ast_mode.c data32.b[2] = srcdata32[1].b[1] | (srcdata32[1].b[0] >> 4); b 1127 drivers/gpu/drm/ast/ast_mode.c data32.b[3] = srcdata32[1].b[3] | (srcdata32[1].b[2] >> 4); b 1139 drivers/gpu/drm/ast/ast_mode.c data16.b[0] = srcdata32[0].b[1] | (srcdata32[0].b[0] >> 4); b 1140 drivers/gpu/drm/ast/ast_mode.c data16.b[1] = srcdata32[0].b[3] | (srcdata32[0].b[2] >> 4); b 394 drivers/gpu/drm/bridge/cdns-dsi.c #define DAT_REMAP_CFG(b, l) ((l) << ((b) * 8)) b 657 drivers/gpu/drm/bridge/sil-sii8620.c u8 b[2] = { min(ctx->burst.rx_ack, 255), 0 }; b 661 drivers/gpu/drm/bridge/sil-sii8620.c ctx->burst.rx_ack -= b[0]; b 662 drivers/gpu/drm/bridge/sil-sii8620.c sii8620_write_buf(ctx, REG_EMSC_XMIT_WRITE_PORT, b, 2); b 161 drivers/gpu/drm/bridge/synopsys/dw-hdmi-ahb-audio.c u32 b, sample = *src++; b 163 drivers/gpu/drm/bridge/synopsys/dw-hdmi-ahb-audio.c b = (sample & 8) << (28 - 3); b 167 drivers/gpu/drm/bridge/synopsys/dw-hdmi-ahb-audio.c *dst++ = sample | b; b 103 drivers/gpu/drm/bridge/synopsys/dw-mipi-dsi.c #define VID_NULL_SIZE(b) ((b) & 0x1fff) b 273 drivers/gpu/drm/bridge/tc358767.c static inline struct tc_data *bridge_to_tc(struct drm_bridge *b) b 275 drivers/gpu/drm/bridge/tc358767.c return container_of(b, struct tc_data, bridge); b 411 drivers/gpu/drm/drm_blend.c static int drm_atomic_state_zpos_cmp(const void *a, const void *b) b 414 drivers/gpu/drm/drm_blend.c const struct drm_plane_state *sb = *(struct drm_plane_state **)b; b 2106 drivers/gpu/drm/drm_edid.c bad_std_timing(u8 a, u8 b) b 2108 drivers/gpu/drm/drm_edid.c return (a == 0x00 && b == 0x00) || b 2109 drivers/gpu/drm/drm_edid.c (a == 0x01 && b == 0x01) || b 2110 drivers/gpu/drm/drm_edid.c (a == 0x20 && b == 0x20); b 958 drivers/gpu/drm/drm_fb_helper.c u16 *r, *g, *b; b 972 drivers/gpu/drm/drm_fb_helper.c b = g + crtc->gamma_size; b 976 drivers/gpu/drm/drm_fb_helper.c memcpy(b + cmap->start, cmap->blue, cmap->len * sizeof(*b)); b 978 drivers/gpu/drm/drm_fb_helper.c ret = crtc->funcs->gamma_set(crtc, r, g, b, b 1008 drivers/gpu/drm/drm_fb_helper.c u16 *b = g + crtc->gamma_size; b 1013 drivers/gpu/drm/drm_fb_helper.c lut[i].blue = b[i]; b 1018 drivers/gpu/drm/drm_fb_helper.c lut[i].blue = b[i]; b 1041 drivers/gpu/drm/drm_fb_helper.c u16 *r, *g, *b; b 1089 drivers/gpu/drm/drm_fb_helper.c b = g + crtc->gamma_size; b 1093 drivers/gpu/drm/drm_fb_helper.c memcpy(b + cmap->start, cmap->blue, cmap->len * sizeof(*b)); b 344 drivers/gpu/drm/drm_file.c struct pci_bus *b = list_entry(pci_root_buses.next, b 346 drivers/gpu/drm/drm_file.c if (b) b 347 drivers/gpu/drm/drm_file.c dev->hose = b->sysdata; b 313 drivers/gpu/drm/drm_format_helper.c u8 b = *src & 0x000000ff; b 316 drivers/gpu/drm/drm_format_helper.c *dst++ = (3 * r + 6 * g + b) / 10; b 1325 drivers/gpu/drm/drm_modes.c struct drm_display_mode *b = list_entry(lh_b, struct drm_display_mode, head); b 1328 drivers/gpu/drm/drm_modes.c diff = ((b->type & DRM_MODE_TYPE_PREFERRED) != 0) - b 1332 drivers/gpu/drm/drm_modes.c diff = b->hdisplay * b->vdisplay - a->hdisplay * a->vdisplay; b 1336 drivers/gpu/drm/drm_modes.c diff = b->vrefresh - a->vrefresh; b 1340 drivers/gpu/drm/drm_modes.c diff = b->clock - a->clock; b 1070 drivers/gpu/drm/etnaviv/etnaviv_gpu.c static inline bool fence_after(u32 a, u32 b) b 1072 drivers/gpu/drm/etnaviv/etnaviv_gpu.c return (s32)(a - b) > 0; b 37 drivers/gpu/drm/exynos/exynos_drm_dsi.c #define NEQV(a, b) (!(a) ^ !(b)) b 158 drivers/gpu/drm/gma500/cdv_intel_dp.c int b; b 168 drivers/gpu/drm/gma500/cdv_intel_dp.c for (b = 0; b < len; b++) { b 169 drivers/gpu/drm/gma500/cdv_intel_dp.c ret = i2c_algo_dp_aux_get_byte(adapter, &buf[b]); b 174 drivers/gpu/drm/gma500/cdv_intel_dp.c for (b = 0; b < len; b++) { b 175 drivers/gpu/drm/gma500/cdv_intel_dp.c ret = i2c_algo_dp_aux_put_byte(adapter, buf[b]); b 143 drivers/gpu/drm/gma500/gma_display.c u16 *r, *g, *b; b 152 drivers/gpu/drm/gma500/gma_display.c b = g + crtc->gamma_size; b 159 drivers/gpu/drm/gma500/gma_display.c ((*b++ >> 8) + gma_crtc->lut_adj[i])); b 168 drivers/gpu/drm/gma500/gma_display.c ((*b++ >> 8) + gma_crtc->lut_adj[i]); b 1283 drivers/gpu/drm/gma500/psb_intel_reg.h #define _PIPE(pipe, a, b) ((a) + (pipe)*((b)-(a))) b 109 drivers/gpu/drm/i915/display/intel_atomic.c const struct drm_property_blob *b) b 111 drivers/gpu/drm/i915/display/intel_atomic.c if (a && b) b 112 drivers/gpu/drm/i915/display/intel_atomic.c return a->length == b->length && b 113 drivers/gpu/drm/i915/display/intel_atomic.c !memcmp(a->data, b->data, a->length); b 115 drivers/gpu/drm/i915/display/intel_atomic.c return !a == !b; b 2060 drivers/gpu/drm/i915/display/intel_cdclk.c const struct intel_cdclk_state *b) b 2062 drivers/gpu/drm/i915/display/intel_cdclk.c return a->cdclk != b->cdclk || b 2063 drivers/gpu/drm/i915/display/intel_cdclk.c a->vco != b->vco || b 2064 drivers/gpu/drm/i915/display/intel_cdclk.c a->ref != b->ref; b 2078 drivers/gpu/drm/i915/display/intel_cdclk.c const struct intel_cdclk_state *b) b 2084 drivers/gpu/drm/i915/display/intel_cdclk.c return a->cdclk != b->cdclk && b 2085 drivers/gpu/drm/i915/display/intel_cdclk.c a->vco == b->vco && b 2086 drivers/gpu/drm/i915/display/intel_cdclk.c a->ref == b->ref; b 2098 drivers/gpu/drm/i915/display/intel_cdclk.c const struct intel_cdclk_state *b) b 2100 drivers/gpu/drm/i915/display/intel_cdclk.c return intel_cdclk_needs_modeset(a, b) || b 2101 drivers/gpu/drm/i915/display/intel_cdclk.c a->voltage_level != b->voltage_level; b 27 drivers/gpu/drm/i915/display/intel_cdclk.h const struct intel_cdclk_state *b); b 29 drivers/gpu/drm/i915/display/intel_cdclk.h const struct intel_cdclk_state *b); b 31 drivers/gpu/drm/i915/display/intel_cdclk.h const struct intel_cdclk_state *b); b 11632 drivers/gpu/drm/i915/display/intel_display.c const struct intel_encoder *b) b 11635 drivers/gpu/drm/i915/display/intel_display.c return a == b || (a->cloneable & (1 << b->type) && b 11636 drivers/gpu/drm/i915/display/intel_display.c b->cloneable & (1 << a->type)); b 12473 drivers/gpu/drm/i915/display/intel_display.c const union hdmi_infoframe *b) b 12475 drivers/gpu/drm/i915/display/intel_display.c return memcmp(a, b, sizeof(*a)) == 0; b 12482 drivers/gpu/drm/i915/display/intel_display.c const union hdmi_infoframe *b) b 12492 drivers/gpu/drm/i915/display/intel_display.c hdmi_infoframe_log(KERN_DEBUG, dev_priv->drm.dev, b); b 12498 drivers/gpu/drm/i915/display/intel_display.c hdmi_infoframe_log(KERN_ERR, dev_priv->drm.dev, b); b 692 drivers/gpu/drm/i915/display/intel_dpll_mgr.c u64 a, b, c, d, diff, diff_best; b 717 drivers/gpu/drm/i915/display/intel_dpll_mgr.c b = freq2k * budget * best->p * best->r2; b 724 drivers/gpu/drm/i915/display/intel_dpll_mgr.c if (a < c && b < d) { b 731 drivers/gpu/drm/i915/display/intel_dpll_mgr.c } else if (a >= c && b < d) { b 736 drivers/gpu/drm/i915/display/intel_dpll_mgr.c } else if (a >= c && b >= d) { b 34 drivers/gpu/drm/i915/display/intel_dpll_mgr.h #define abs_diff(a, b) ({ \ b 36 drivers/gpu/drm/i915/display/intel_dpll_mgr.h typeof(b) __b = (b); \ b 54 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c static void __intel_breadcrumbs_disarm_irq(struct intel_breadcrumbs *b) b 56 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c lockdep_assert_held(&b->irq_lock); b 58 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c GEM_BUG_ON(!b->irq_enabled); b 59 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c if (!--b->irq_enabled) b 60 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c irq_disable(container_of(b, b 64 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c b->irq_armed = false; b 69 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c struct intel_breadcrumbs *b = &engine->breadcrumbs; b 72 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c if (!b->irq_armed) b 75 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c spin_lock_irqsave(&b->irq_lock, flags); b 76 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c if (b->irq_armed) b 77 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c __intel_breadcrumbs_disarm_irq(b); b 78 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c spin_unlock_irqrestore(&b->irq_lock, flags); b 133 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c struct intel_breadcrumbs *b = &engine->breadcrumbs; b 139 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c spin_lock(&b->irq_lock); b 141 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c if (b->irq_armed && list_empty(&b->signalers)) b 142 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c __intel_breadcrumbs_disarm_irq(b); b 144 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c list_for_each_entry_safe(ce, cn, &b->signalers, signal_link) { b 185 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c spin_unlock(&b->irq_lock); b 217 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c static void __intel_breadcrumbs_arm_irq(struct intel_breadcrumbs *b) b 220 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c container_of(b, struct intel_engine_cs, breadcrumbs); b 222 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c lockdep_assert_held(&b->irq_lock); b 223 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c if (b->irq_armed) b 232 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c b->irq_armed = true; b 242 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c if (!b->irq_enabled++) b 248 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c struct intel_breadcrumbs *b = &engine->breadcrumbs; b 250 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c spin_lock_init(&b->irq_lock); b 251 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c INIT_LIST_HEAD(&b->signalers); b 253 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c init_irq_work(&b->irq_work, signal_irq_work); b 258 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c struct intel_breadcrumbs *b = &engine->breadcrumbs; b 261 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c spin_lock_irqsave(&b->irq_lock, flags); b 263 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c if (b->irq_enabled) b 268 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c spin_unlock_irqrestore(&b->irq_lock, flags); b 281 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c struct intel_breadcrumbs *b = &rq->engine->breadcrumbs; b 285 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c spin_lock(&b->irq_lock); b 288 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c __intel_breadcrumbs_arm_irq(b); b 313 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c list_move_tail(&ce->signal_link, &b->signalers); b 317 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c spin_unlock(&b->irq_lock); b 325 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c struct intel_breadcrumbs *b = &rq->engine->breadcrumbs; b 336 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c spin_lock(&b->irq_lock); b 346 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c spin_unlock(&b->irq_lock); b 352 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c struct intel_breadcrumbs *b = &engine->breadcrumbs; b 356 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c if (list_empty(&b->signalers)) b 361 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c spin_lock_irq(&b->irq_lock); b 362 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c list_for_each_entry(ce, &b->signalers, signal_link) { b 372 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c spin_unlock_irq(&b->irq_lock); b 55 drivers/gpu/drm/i915/gt/intel_engine_user.c const struct intel_engine_cs *b = b 56 drivers/gpu/drm/i915/gt/intel_engine_user.c container_of((struct rb_node *)B, typeof(*b), uabi_node); b 58 drivers/gpu/drm/i915/gt/intel_engine_user.c if (uabi_classes[a->class] < uabi_classes[b->class]) b 60 drivers/gpu/drm/i915/gt/intel_engine_user.c if (uabi_classes[a->class] > uabi_classes[b->class]) b 63 drivers/gpu/drm/i915/gt/intel_engine_user.c if (a->instance < b->instance) b 65 drivers/gpu/drm/i915/gt/intel_engine_user.c if (a->instance > b->instance) b 753 drivers/gpu/drm/i915/gt/selftest_lrc.c struct preempt_client a, b; b 771 drivers/gpu/drm/i915/gt/selftest_lrc.c if (preempt_client_init(i915, &b)) b 773 drivers/gpu/drm/i915/gt/selftest_lrc.c b.ctx->sched.priority = I915_USER_PRIORITY(I915_PRIORITY_MAX); b 800 drivers/gpu/drm/i915/gt/selftest_lrc.c rq_b = spinner_create_request(&b.spin, b 801 drivers/gpu/drm/i915/gt/selftest_lrc.c b.ctx, engine, b 814 drivers/gpu/drm/i915/gt/selftest_lrc.c if (igt_wait_for_spinner(&b.spin, rq_b)) { b 821 drivers/gpu/drm/i915/gt/selftest_lrc.c if (!igt_wait_for_spinner(&b.spin, rq_b)) { b 826 drivers/gpu/drm/i915/gt/selftest_lrc.c igt_spinner_end(&b.spin); b 841 drivers/gpu/drm/i915/gt/selftest_lrc.c preempt_client_fini(&b); b 850 drivers/gpu/drm/i915/gt/selftest_lrc.c igt_spinner_end(&b.spin); b 864 drivers/gpu/drm/i915/gt/selftest_lrc.c struct preempt_client a, b; b 890 drivers/gpu/drm/i915/gt/selftest_lrc.c if (preempt_client_init(i915, &b)) b 919 drivers/gpu/drm/i915/gt/selftest_lrc.c rq_b = spinner_create_request(&b.spin, b 920 drivers/gpu/drm/i915/gt/selftest_lrc.c b.ctx, engine, b 932 drivers/gpu/drm/i915/gt/selftest_lrc.c if (!igt_wait_for_spinner(&b.spin, rq_b)) { b 937 drivers/gpu/drm/i915/gt/selftest_lrc.c swap(a, b); b 957 drivers/gpu/drm/i915/gt/selftest_lrc.c preempt_client_fini(&b); b 966 drivers/gpu/drm/i915/gt/selftest_lrc.c igt_spinner_end(&b.spin); b 902 drivers/gpu/drm/i915/gt/selftest_workarounds.c u32 a, u32 b, i915_reg_t reg) b 904 drivers/gpu/drm/i915/gt/selftest_workarounds.c if (a != b && !pardon_reg(engine->i915, reg)) { b 906 drivers/gpu/drm/i915/gt/selftest_workarounds.c i915_mmio_reg_offset(reg), a, b); b 924 drivers/gpu/drm/i915/gt/selftest_workarounds.c u32 a, u32 b, i915_reg_t reg) b 926 drivers/gpu/drm/i915/gt/selftest_workarounds.c if (a == b && !writeonly_reg(engine->i915, reg)) { b 940 drivers/gpu/drm/i915/gt/selftest_workarounds.c u32 a, u32 b, b 943 drivers/gpu/drm/i915/gt/selftest_workarounds.c u32 *a, *b; b 950 drivers/gpu/drm/i915/gt/selftest_workarounds.c b = i915_gem_object_pin_map(B->obj, I915_MAP_WB); b 951 drivers/gpu/drm/i915/gt/selftest_workarounds.c if (IS_ERR(b)) { b 952 drivers/gpu/drm/i915/gt/selftest_workarounds.c err = PTR_ERR(b); b 964 drivers/gpu/drm/i915/gt/selftest_workarounds.c if (!fn(engine, a[i], b[i], wa->reg)) b 44 drivers/gpu/drm/i915/gvt/debugfs.c struct list_head *a, struct list_head *b) b 50 drivers/gpu/drm/i915/gvt/debugfs.c mb = container_of(b, struct diff_mmio, node); b 46 drivers/gpu/drm/i915/gvt/execlist.c #define same_context(a, b) (((a)->context_id == (b)->context_id) && \ b 47 drivers/gpu/drm/i915/gvt/execlist.c ((a)->lrca == (b)->lrca)) b 169 drivers/gpu/drm/i915/gvt/firmware.c #define VERIFY(s, a, b) do { \ b 170 drivers/gpu/drm/i915/gvt/firmware.c item = (s); file = (u64)(a); request = (u64)(b); \ b 171 drivers/gpu/drm/i915/gvt/firmware.c if ((a) != (b)) \ b 373 drivers/gpu/drm/i915/gvt/gvt.h #define BYTES_TO_MB(b) ((b) >> 20ULL) b 496 drivers/gpu/drm/i915/gvt/interrupt.c #define SET_BIT_INFO(s, b, e, i) \ b 498 drivers/gpu/drm/i915/gvt/interrupt.c s->events[e].bit = b; \ b 500 drivers/gpu/drm/i915/gvt/interrupt.c s->info[i]->bit_to_event[b] = e;\ b 1459 drivers/gpu/drm/i915/gvt/scheduler.c #define same_context(a, b) (((a)->context_id == (b)->context_id) && \ b 1460 drivers/gpu/drm/i915/gvt/scheduler.c ((a)->lrca == (b)->lrca)) b 4424 drivers/gpu/drm/i915/i915_debugfs.c const struct dpcd_block *b = &i915_dpcd_debug[i]; b 4425 drivers/gpu/drm/i915/i915_debugfs.c size_t size = b->end ? b->end - b->offset + 1 : (b->size ?: 1); b 4427 drivers/gpu/drm/i915/i915_debugfs.c if (b->edp && b 4435 drivers/gpu/drm/i915/i915_debugfs.c err = drm_dp_dpcd_read(&intel_dp->aux, b->offset, buf, size); b 4437 drivers/gpu/drm/i915/i915_debugfs.c seq_printf(m, "%04x: ERROR %d\n", b->offset, (int)err); b 4439 drivers/gpu/drm/i915/i915_debugfs.c seq_printf(m, "%04x: %*ph\n", b->offset, (int)err, buf); b 194 drivers/gpu/drm/i915/i915_reg.h static inline bool i915_mmio_reg_equal(i915_reg_t a, i915_reg_t b) b 196 drivers/gpu/drm/i915/i915_reg.h return i915_mmio_reg_offset(a) == i915_mmio_reg_offset(b); b 228 drivers/gpu/drm/i915/i915_reg.h #define _PIPE(pipe, a, b) _PICK_EVEN(pipe, a, b) b 229 drivers/gpu/drm/i915/i915_reg.h #define _PLANE(plane, a, b) _PICK_EVEN(plane, a, b) b 230 drivers/gpu/drm/i915/i915_reg.h #define _TRANS(tran, a, b) _PICK_EVEN(tran, a, b) b 231 drivers/gpu/drm/i915/i915_reg.h #define _PORT(port, a, b) _PICK_EVEN(port, a, b) b 232 drivers/gpu/drm/i915/i915_reg.h #define _PLL(pll, a, b) _PICK_EVEN(pll, a, b) b 234 drivers/gpu/drm/i915/i915_reg.h #define _MMIO_PIPE(pipe, a, b) _MMIO(_PIPE(pipe, a, b)) b 235 drivers/gpu/drm/i915/i915_reg.h #define _MMIO_PLANE(plane, a, b) _MMIO(_PLANE(plane, a, b)) b 236 drivers/gpu/drm/i915/i915_reg.h #define _MMIO_TRANS(tran, a, b) _MMIO(_TRANS(tran, a, b)) b 237 drivers/gpu/drm/i915/i915_reg.h #define _MMIO_PORT(port, a, b) _MMIO(_PORT(port, a, b)) b 238 drivers/gpu/drm/i915/i915_reg.h #define _MMIO_PLL(pll, a, b) _MMIO(_PLL(pll, a, b)) b 242 drivers/gpu/drm/i915/i915_reg.h #define _MMIO_PIPE3(pipe, a, b, c) _MMIO(_PICK(pipe, a, b, c)) b 243 drivers/gpu/drm/i915/i915_reg.h #define _MMIO_PORT3(pipe, a, b, c) _MMIO(_PICK(pipe, a, b, c)) b 244 drivers/gpu/drm/i915/i915_reg.h #define _MMIO_PHY3(phy, a, b, c) _MMIO(_PHY3(phy, a, b, c)) b 245 drivers/gpu/drm/i915/i915_reg.h #define _MMIO_PLL3(pll, a, b, c) _MMIO(_PICK(pll, a, b, c)) b 7154 drivers/gpu/drm/i915/i915_reg.h #define _ID(id, a, b) _PICK_EVEN(id, a, b) b 135 drivers/gpu/drm/i915/i915_syncmap.c static inline bool seqno_later(u32 a, u32 b) b 137 drivers/gpu/drm/i915/i915_syncmap.c return (s32)(a - b) >= 0; b 88 drivers/gpu/drm/i915/i915_utils.h typeof(B) b = (B); \ b 89 drivers/gpu/drm/i915/i915_utils.h (T)(a + b) < a; \ b 247 drivers/gpu/drm/i915/i915_vma.h static __always_inline ptrdiff_t ptrdiff(const void *a, const void *b) b 249 drivers/gpu/drm/i915/i915_vma.h return a - b; b 3205 drivers/gpu/drm/i915/intel_pm.c const struct intel_pipe_wm *b = &oldstate->wm.ilk.optimal; b 3218 drivers/gpu/drm/i915/intel_pm.c a->pipe_enabled |= b->pipe_enabled; b 3219 drivers/gpu/drm/i915/intel_pm.c a->sprites_enabled |= b->sprites_enabled; b 3220 drivers/gpu/drm/i915/intel_pm.c a->sprites_scaled |= b->sprites_scaled; b 3224 drivers/gpu/drm/i915/intel_pm.c const struct intel_wm_level *b_wm = &b->wm[level]; b 5246 drivers/gpu/drm/i915/intel_pm.c const struct skl_ddb_entry *b) b 5248 drivers/gpu/drm/i915/intel_pm.c return a->start < b->end && b->start < a->end; b 8178 drivers/gpu/drm/i915/intel_pm.c unsigned long m, x, b; b 8186 drivers/gpu/drm/i915/intel_pm.c b = tsfs & TSFS_INTR_MASK; b 8188 drivers/gpu/drm/i915/intel_pm.c return ((m * x) / 127) - b; b 162 drivers/gpu/drm/i915/intel_runtime_pm.c const depot_stack_handle_t * const a = _a, * const b = _b; b 164 drivers/gpu/drm/i915/intel_runtime_pm.c if (*a < *b) b 166 drivers/gpu/drm/i915/intel_runtime_pm.c else if (*a > *b) b 273 drivers/gpu/drm/i915/intel_uncore.h __raw_read(8, b) b 278 drivers/gpu/drm/i915/intel_uncore.h __raw_write(8, b) b 300 drivers/gpu/drm/i915/intel_uncore.h __uncore_read(read8, 8, b, true) b 306 drivers/gpu/drm/i915/intel_uncore.h __uncore_write(write8, 8, b, true) b 1068 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c struct drm_mm_node *b = list_entry(B, typeof(*b), hole_stack); b 1070 drivers/gpu/drm/i915/selftests/i915_gem_gtt.c if (a->start < b->start) b 468 drivers/gpu/drm/i915/selftests/i915_vma.c const struct intel_remapped_plane_info *b) b 470 drivers/gpu/drm/i915/selftests/i915_vma.c return a->width * a->height + b->width * b->height; b 495 drivers/gpu/drm/i915/selftests/i915_vma.c }, *a, *b; b 514 drivers/gpu/drm/i915/selftests/i915_vma.c for (b = planes + ARRAY_SIZE(planes); b-- != planes; ) { b 519 drivers/gpu/drm/i915/selftests/i915_vma.c b->stride * b->height); b 525 drivers/gpu/drm/i915/selftests/i915_vma.c view.rotated.plane[1] = *b; b 545 drivers/gpu/drm/i915/selftests/i915_vma.c vma->size != rotated_size(a, b) * PAGE_SIZE) { b 547 drivers/gpu/drm/i915/selftests/i915_vma.c PAGE_SIZE * rotated_size(a, b), vma->size); b 553 drivers/gpu/drm/i915/selftests/i915_vma.c vma->size > rotated_size(a, b) * PAGE_SIZE) { b 555 drivers/gpu/drm/i915/selftests/i915_vma.c PAGE_SIZE * rotated_size(a, b), vma->size); b 560 drivers/gpu/drm/i915/selftests/i915_vma.c if (vma->pages->nents > rotated_size(a, b)) { b 562 drivers/gpu/drm/i915/selftests/i915_vma.c rotated_size(a, b), vma->pages->nents); b 173 drivers/gpu/drm/mediatek/mtk_hdmi.c static inline struct mtk_hdmi *hdmi_ctx_from_bridge(struct drm_bridge *b) b 175 drivers/gpu/drm/mediatek/mtk_hdmi.c return container_of(b, struct mtk_hdmi, bridge); b 854 drivers/gpu/drm/meson/meson_venc.c static unsigned long modulo(unsigned long a, unsigned long b) b 856 drivers/gpu/drm/meson/meson_venc.c if (a >= b) b 857 drivers/gpu/drm/meson/meson_venc.c return a - b; b 46 drivers/gpu/drm/mgag200/mgag200_mode.c u8 r, b; b 50 drivers/gpu/drm/mgag200/mgag200_mode.c r = b = 0; b 53 drivers/gpu/drm/mgag200/mgag200_mode.c b = *b_ptr++ >> 8; b 59 drivers/gpu/drm/mgag200/mgag200_mode.c b = *b_ptr++ >> 8; b 64 drivers/gpu/drm/mgag200/mgag200_mode.c WREG8(DAC_INDEX + MGA1064_COL_PAL, b); b 33 drivers/gpu/drm/msm/disp/dpu1/dpu_formats.c #define INTERLEAVED_RGB_FMT(fmt, a, r, g, b, e0, e1, e2, e3, uc, alpha, \ b 40 drivers/gpu/drm/msm/disp/dpu1/dpu_formats.c .bits = { g, b, r, a }, \ b 52 drivers/gpu/drm/msm/disp/dpu1/dpu_formats.c #define INTERLEAVED_RGB_FMT_TILED(fmt, a, r, g, b, e0, e1, e2, e3, uc, \ b 59 drivers/gpu/drm/msm/disp/dpu1/dpu_formats.c .bits = { g, b, r, a }, \ b 72 drivers/gpu/drm/msm/disp/dpu1/dpu_formats.c #define INTERLEAVED_YUV_FMT(fmt, a, r, g, b, e0, e1, e2, e3, \ b 79 drivers/gpu/drm/msm/disp/dpu1/dpu_formats.c .bits = { g, b, r, a }, \ b 91 drivers/gpu/drm/msm/disp/dpu1/dpu_formats.c #define PSEUDO_YUV_FMT(fmt, a, r, g, b, e0, e1, chroma, flg, fm, np) \ b 97 drivers/gpu/drm/msm/disp/dpu1/dpu_formats.c .bits = { g, b, r, a }, \ b 109 drivers/gpu/drm/msm/disp/dpu1/dpu_formats.c #define PSEUDO_YUV_FMT_TILED(fmt, a, r, g, b, e0, e1, chroma, \ b 116 drivers/gpu/drm/msm/disp/dpu1/dpu_formats.c .bits = { g, b, r, a }, \ b 128 drivers/gpu/drm/msm/disp/dpu1/dpu_formats.c #define PSEUDO_YUV_FMT_LOOSE(fmt, a, r, g, b, e0, e1, chroma, flg, fm, np)\ b 134 drivers/gpu/drm/msm/disp/dpu1/dpu_formats.c .bits = { g, b, r, a }, \ b 146 drivers/gpu/drm/msm/disp/dpu1/dpu_formats.c #define PSEUDO_YUV_FMT_LOOSE_TILED(fmt, a, r, g, b, e0, e1, chroma, \ b 153 drivers/gpu/drm/msm/disp/dpu1/dpu_formats.c .bits = { g, b, r, a }, \ b 166 drivers/gpu/drm/msm/disp/dpu1/dpu_formats.c #define PLANAR_YUV_FMT(fmt, a, r, g, b, e0, e1, e2, alpha, chroma, bp, \ b 173 drivers/gpu/drm/msm/disp/dpu1/dpu_formats.c .bits = { g, b, r, a }, \ b 34 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c struct dpu_hw_blk_reg_map *b) b 40 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c b->base_off = addr; b 41 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c b->blk_off = m->ctl[i].base; b 42 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c b->length = m->ctl[i].len; b 43 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c b->hwversion = m->hwversion; b 44 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c b->log_mask = DPU_DBG_MASK_CTL; b 62 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_intf.c struct dpu_hw_blk_reg_map *b) b 69 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_intf.c b->base_off = addr; b 70 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_intf.c b->blk_off = m->intf[i].base; b 71 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_intf.c b->length = m->intf[i].len; b 72 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_intf.c b->hwversion = m->hwversion; b 73 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_intf.c b->log_mask = DPU_DBG_MASK_INTF; b 30 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_lm.c struct dpu_hw_blk_reg_map *b) b 36 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_lm.c b->base_off = addr; b 37 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_lm.c b->blk_off = m->mixer[i].base; b 38 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_lm.c b->length = m->mixer[i].len; b 39 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_lm.c b->hwversion = m->hwversion; b 40 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_lm.c b->log_mask = DPU_DBG_MASK_LM; b 34 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_pingpong.c struct dpu_hw_blk_reg_map *b) b 40 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_pingpong.c b->base_off = addr; b 41 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_pingpong.c b->blk_off = m->pingpong[i].base; b 42 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_pingpong.c b->length = m->pingpong[i].len; b 43 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_pingpong.c b->hwversion = m->hwversion; b 44 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_pingpong.c b->log_mask = DPU_DBG_MASK_PINGPONG; b 672 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c struct dpu_hw_blk_reg_map *b) b 676 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c if ((sspp < SSPP_MAX) && catalog && addr && b) { b 679 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c b->base_off = addr; b 680 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c b->blk_off = catalog->sspp[i].base; b 681 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c b->length = catalog->sspp[i].len; b 682 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c b->hwversion = catalog->hwversion; b 683 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c b->log_mask = DPU_DBG_MASK_SSPP; b 295 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_top.c struct dpu_hw_blk_reg_map *b) b 299 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_top.c if (!m || !addr || !b) b 304 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_top.c b->base_off = addr; b 305 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_top.c b->blk_off = m->mdp[i].base; b 306 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_top.c b->length = m->mdp[i].len; b 307 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_top.c b->hwversion = m->hwversion; b 308 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_top.c b->log_mask = DPU_DBG_MASK_TOP; b 216 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_vbif.c struct dpu_hw_blk_reg_map *b) b 222 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_vbif.c b->base_off = addr; b 223 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_vbif.c b->blk_off = m->vbif[i].base; b 224 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_vbif.c b->length = m->vbif[i].len; b 225 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_vbif.c b->hwversion = m->hwversion; b 226 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_vbif.c b->log_mask = DPU_DBG_MASK_VBIF; b 571 drivers/gpu/drm/msm/disp/mdp5/mdp5_crtc.c static int pstate_cmp(const void *a, const void *b) b 574 drivers/gpu/drm/msm/disp/mdp5/mdp5_crtc.c struct plane_state *pb = (struct plane_state *)b; b 63 drivers/gpu/drm/msm/disp/mdp_format.c #define FMT(name, a, r, g, b, e0, e1, e2, e3, alpha, tight, c, cnt, fp, cs, yuv) { \ b 68 drivers/gpu/drm/msm/disp/mdp_format.c .bpc_b = BPC ## b, \ b 766 drivers/gpu/drm/nouveau/dispnv04/crtc.c struct rgb { uint8_t r, g, b; } __attribute__((packed)) *rgbs; b 767 drivers/gpu/drm/nouveau/dispnv04/crtc.c u16 *r, *g, *b; b 773 drivers/gpu/drm/nouveau/dispnv04/crtc.c b = g + crtc->gamma_size; b 778 drivers/gpu/drm/nouveau/dispnv04/crtc.c rgbs[i].b = *b++ >> 8; b 795 drivers/gpu/drm/nouveau/dispnv04/crtc.c nv_crtc_gamma_set(struct drm_crtc *crtc, u16 *r, u16 *g, u16 *b, b 246 drivers/gpu/drm/nouveau/dispnv04/dac.c #define RGB_TEST_DATA(r, g, b) (r << 0 | g << 10 | b << 20) b 56 drivers/gpu/drm/nouveau/dispnv04/tvnv17.c #define RGB_TEST_DATA(r, g, b) (r << 0 | g << 10 | b << 20) b 110 drivers/gpu/drm/nouveau/dispnv50/headc57d.c u16 b = drm_color_lut_extract(in-> blue + 0, 16); b 116 drivers/gpu/drm/nouveau/dispnv50/headc57d.c bi = (drm_color_lut_extract(in-> blue, 16) - b) / 4; b 122 drivers/gpu/drm/nouveau/dispnv50/headc57d.c writew(b + bi * i, mem + 4); b 145 drivers/gpu/drm/nouveau/dispnv50/wndwc57e.c u16 b = fixedU0_16_FP16(drm_color_lut_extract(in-> blue, 16)); b 148 drivers/gpu/drm/nouveau/dispnv50/wndwc57e.c writew(b, mem + 4); b 42 drivers/gpu/drm/nouveau/include/nvif/object.h #define nvif_rd(a,f,b,c) ({ \ b 48 drivers/gpu/drm/nouveau/include/nvif/object.h _data = nvif_object_rd(_object, (b), (c)); \ b 51 drivers/gpu/drm/nouveau/include/nvif/object.h #define nvif_wr(a,f,b,c,d) ({ \ b 56 drivers/gpu/drm/nouveau/include/nvif/object.h nvif_object_wr(_object, (b), (c), (d)); \ b 58 drivers/gpu/drm/nouveau/include/nvif/object.h #define nvif_rd08(a,b) ({ ((u8)nvif_rd((a), ioread8, 1, (b))); }) b 59 drivers/gpu/drm/nouveau/include/nvif/object.h #define nvif_rd16(a,b) ({ ((u16)nvif_rd((a), ioread16_native, 2, (b))); }) b 60 drivers/gpu/drm/nouveau/include/nvif/object.h #define nvif_rd32(a,b) ({ ((u32)nvif_rd((a), ioread32_native, 4, (b))); }) b 61 drivers/gpu/drm/nouveau/include/nvif/object.h #define nvif_wr08(a,b,c) nvif_wr((a), iowrite8, 1, (b), (u8)(c)) b 62 drivers/gpu/drm/nouveau/include/nvif/object.h #define nvif_wr16(a,b,c) nvif_wr((a), iowrite16_native, 2, (b), (u16)(c)) b 63 drivers/gpu/drm/nouveau/include/nvif/object.h #define nvif_wr32(a,b,c) nvif_wr((a), iowrite32_native, 4, (b), (u32)(c)) b 64 drivers/gpu/drm/nouveau/include/nvif/object.h #define nvif_mask(a,b,c,d) ({ \ b 66 drivers/gpu/drm/nouveau/include/nvif/object.h u32 _addr = (b), _data = nvif_rd32(__object, _addr); \ b 71 drivers/gpu/drm/nouveau/include/nvif/object.h #define nvif_mthd(a,b,c,d) nvif_object_mthd((a), (b), (c), (d)) b 335 drivers/gpu/drm/nouveau/nouveau_gem.c struct drm_nouveau_gem_pushbuf_bo *b; b 339 drivers/gpu/drm/nouveau/nouveau_gem.c b = &pbbo[nvbo->pbbo_index]; b 342 drivers/gpu/drm/nouveau/nouveau_gem.c nouveau_bo_fence(nvbo, fence, !!b->write_domains); b 346 drivers/gpu/drm/nouveau/nouveau_gem.c (void *)(unsigned long)b->user_priv; b 395 drivers/gpu/drm/nouveau/nouveau_gem.c struct drm_nouveau_gem_pushbuf_bo *b = &pbbo[i]; b 399 drivers/gpu/drm/nouveau/nouveau_gem.c gem = drm_gem_object_lookup(file_priv, b->handle); b 401 drivers/gpu/drm/nouveau/nouveau_gem.c NV_PRINTK(err, cli, "Unknown handle 0x%08x\n", b->handle); b 414 drivers/gpu/drm/nouveau/nouveau_gem.c "validation list\n", b->handle); b 448 drivers/gpu/drm/nouveau/nouveau_gem.c b->user_priv = (uint64_t)(unsigned long)vma; b 450 drivers/gpu/drm/nouveau/nouveau_gem.c b->user_priv = (uint64_t)(unsigned long)nvbo; b 455 drivers/gpu/drm/nouveau/nouveau_gem.c if ((b->valid_domains & NOUVEAU_GEM_DOMAIN_VRAM) && b 456 drivers/gpu/drm/nouveau/nouveau_gem.c (b->valid_domains & NOUVEAU_GEM_DOMAIN_GART)) b 459 drivers/gpu/drm/nouveau/nouveau_gem.c if (b->valid_domains & NOUVEAU_GEM_DOMAIN_VRAM) b 462 drivers/gpu/drm/nouveau/nouveau_gem.c if (b->valid_domains & NOUVEAU_GEM_DOMAIN_GART) b 466 drivers/gpu/drm/nouveau/nouveau_gem.c b->valid_domains); b 497 drivers/gpu/drm/nouveau/nouveau_gem.c struct drm_nouveau_gem_pushbuf_bo *b = &pbbo[nvbo->pbbo_index]; b 499 drivers/gpu/drm/nouveau/nouveau_gem.c ret = nouveau_gem_set_domain(&nvbo->bo.base, b->read_domains, b 500 drivers/gpu/drm/nouveau/nouveau_gem.c b->write_domains, b 501 drivers/gpu/drm/nouveau/nouveau_gem.c b->valid_domains); b 514 drivers/gpu/drm/nouveau/nouveau_gem.c ret = nouveau_fence_sync(nvbo, chan, !!b->write_domains, true); b 522 drivers/gpu/drm/nouveau/nouveau_gem.c if (nvbo->bo.offset == b->presumed.offset && b 524 drivers/gpu/drm/nouveau/nouveau_gem.c b->presumed.domain & NOUVEAU_GEM_DOMAIN_VRAM) || b 526 drivers/gpu/drm/nouveau/nouveau_gem.c b->presumed.domain & NOUVEAU_GEM_DOMAIN_GART))) b 530 drivers/gpu/drm/nouveau/nouveau_gem.c b->presumed.domain = NOUVEAU_GEM_DOMAIN_GART; b 532 drivers/gpu/drm/nouveau/nouveau_gem.c b->presumed.domain = NOUVEAU_GEM_DOMAIN_VRAM; b 533 drivers/gpu/drm/nouveau/nouveau_gem.c b->presumed.offset = nvbo->bo.offset; b 534 drivers/gpu/drm/nouveau/nouveau_gem.c b->presumed.valid = 0; b 538 drivers/gpu/drm/nouveau/nouveau_gem.c &b->presumed, sizeof(b->presumed))) b 620 drivers/gpu/drm/nouveau/nouveau_gem.c struct drm_nouveau_gem_pushbuf_bo *b; b 630 drivers/gpu/drm/nouveau/nouveau_gem.c b = &bo[r->bo_index]; b 631 drivers/gpu/drm/nouveau/nouveau_gem.c if (b->presumed.valid) b 659 drivers/gpu/drm/nouveau/nouveau_gem.c data = b->presumed.offset + r->data; b 662 drivers/gpu/drm/nouveau/nouveau_gem.c data = (b->presumed.offset + r->data) >> 32; b 667 drivers/gpu/drm/nouveau/nouveau_gem.c if (b->presumed.domain == NOUVEAU_GEM_DOMAIN_GART) b 421 drivers/gpu/drm/nouveau/nouveau_svm.c nouveau_svm_fault_cmp(const void *a, const void *b) b 424 drivers/gpu/drm/nouveau/nouveau_svm.c const struct nouveau_svm_fault *fb = *(struct nouveau_svm_fault **)b; b 88 drivers/gpu/drm/nouveau/nvkm/core/mm.c struct nvkm_mm_node *b; b 93 drivers/gpu/drm/nouveau/nvkm/core/mm.c b = kmalloc(sizeof(*b), GFP_KERNEL); b 94 drivers/gpu/drm/nouveau/nvkm/core/mm.c if (unlikely(b == NULL)) b 97 drivers/gpu/drm/nouveau/nvkm/core/mm.c b->offset = a->offset; b 98 drivers/gpu/drm/nouveau/nvkm/core/mm.c b->length = size; b 99 drivers/gpu/drm/nouveau/nvkm/core/mm.c b->heap = a->heap; b 100 drivers/gpu/drm/nouveau/nvkm/core/mm.c b->type = a->type; b 103 drivers/gpu/drm/nouveau/nvkm/core/mm.c list_add_tail(&b->nl_entry, &a->nl_entry); b 104 drivers/gpu/drm/nouveau/nvkm/core/mm.c if (b->type == NVKM_MM_TYPE_NONE) b 105 drivers/gpu/drm/nouveau/nvkm/core/mm.c list_add_tail(&b->fl_entry, &a->fl_entry); b 107 drivers/gpu/drm/nouveau/nvkm/core/mm.c return b; b 163 drivers/gpu/drm/nouveau/nvkm/core/mm.c struct nvkm_mm_node *b; b 168 drivers/gpu/drm/nouveau/nvkm/core/mm.c b = kmalloc(sizeof(*b), GFP_KERNEL); b 169 drivers/gpu/drm/nouveau/nvkm/core/mm.c if (unlikely(b == NULL)) b 173 drivers/gpu/drm/nouveau/nvkm/core/mm.c b->offset = a->offset + a->length; b 174 drivers/gpu/drm/nouveau/nvkm/core/mm.c b->length = size; b 175 drivers/gpu/drm/nouveau/nvkm/core/mm.c b->heap = a->heap; b 176 drivers/gpu/drm/nouveau/nvkm/core/mm.c b->type = a->type; b 178 drivers/gpu/drm/nouveau/nvkm/core/mm.c list_add(&b->nl_entry, &a->nl_entry); b 179 drivers/gpu/drm/nouveau/nvkm/core/mm.c if (b->type == NVKM_MM_TYPE_NONE) b 180 drivers/gpu/drm/nouveau/nvkm/core/mm.c list_add(&b->fl_entry, &a->fl_entry); b 182 drivers/gpu/drm/nouveau/nvkm/core/mm.c return b; b 1044 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgf100.c const int b = mmio_vram(info, grctx->bundle_size, (1 << s), true); b 1045 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgf100.c mmio_refn(info, 0x408004, 0x00000000, s, b); b 1047 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgf100.c mmio_refn(info, 0x418808, 0x00000000, s, b); b 1056 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgf100.c const int b = mmio_vram(info, grctx->pagepool_size, (1 << s), true); b 1057 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgf100.c mmio_refn(info, 0x40800c, 0x00000000, s, b); b 1059 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgf100.c mmio_refn(info, 0x419004, 0x00000000, s, b); b 1071 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgf100.c const int b = mmio_vram(info, size * gr->tpc_total, (1 << s), false); b 1075 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgf100.c mmio_refn(info, 0x418810, 0x80000000, s, b); b 1076 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgf100.c mmio_refn(info, 0x419848, 0x10000000, s, b); b 18 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgf100.h #define mmio_vram(a,b,c,d) gf100_grctx_mmio_data((a), (b), (c), (d)) b 19 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgf100.h #define mmio_refn(a,b,c,d,e) gf100_grctx_mmio_item((a), (b), (c), (d), (e)) b 20 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgf100.h #define mmio_skip(a,b,c) mmio_refn((a), (b), (c), -1, -1) b 21 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgf100.h #define mmio_wr32(a,b,c) mmio_refn((a), (b), (c), 0, -1) b 744 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgf108.c const int b = mmio_vram(info, size * gr->tpc_total, (1 << s), false); b 751 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgf108.c mmio_refn(info, 0x418810, 0x80000000, s, b); b 752 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgf108.c mmio_refn(info, 0x419848, 0x10000000, s, b); b 759 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgf108.c const u32 b = beta; b 762 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgf108.c mmio_skip(info, o + 0x20, (t << 28) | (b << 16) | ++bo); b 763 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgf108.c mmio_wr32(info, o + 0x20, (t << 28) | (b << 16) | --bo); b 252 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgf117.c const int b = mmio_vram(info, size * gr->tpc_total, (1 << s), false); b 259 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgf117.c mmio_refn(info, 0x418810, 0x80000000, s, b); b 260 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgf117.c mmio_refn(info, 0x419848, 0x10000000, s, b); b 267 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgf117.c const u32 b = beta * gr->ppc_tpc_nr[gpc][ppc]; b 272 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgf117.c mmio_skip(info, o + 0xc0, (t << 28) | (b << 16) | ++bo); b 273 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgf117.c mmio_wr32(info, o + 0xc0, (t << 28) | (b << 16) | --bo); b 882 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgk104.c const int b = mmio_vram(info, grctx->bundle_size, (1 << s), true); b 883 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgk104.c mmio_refn(info, 0x408004, 0x00000000, s, b); b 885 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgk104.c mmio_refn(info, 0x418808, 0x00000000, s, b); b 895 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgk104.c const int b = mmio_vram(info, grctx->pagepool_size, (1 << s), true); b 896 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgk104.c mmio_refn(info, 0x40800c, 0x00000000, s, b); b 898 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgk104.c mmio_refn(info, 0x419004, 0x00000000, s, b); b 886 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgm107.c const int b = mmio_vram(info, grctx->bundle_size, (1 << s), true); b 887 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgm107.c mmio_refn(info, 0x408004, 0x00000000, s, b); b 889 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgm107.c mmio_refn(info, 0x418e24, 0x00000000, s, b); b 899 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgm107.c const int b = mmio_vram(info, grctx->pagepool_size, (1 << s), true); b 900 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgm107.c mmio_refn(info, 0x40800c, 0x00000000, s, b); b 902 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgm107.c mmio_refn(info, 0x419004, 0x00000000, s, b); b 917 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgm107.c const int b = mmio_vram(info, size * gr->tpc_total, (1 << s), false); b 923 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgm107.c mmio_refn(info, 0x418810, 0x80000000, s, b); b 924 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgm107.c mmio_refn(info, 0x419848, 0x10000000, s, b); b 925 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgm107.c mmio_refn(info, 0x419c2c, 0x10000000, s, b); b 37 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgp100.c const int b = mmio_vram(info, grctx->pagepool_size, (1 << s), true); b 38 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgp100.c mmio_refn(info, 0x40800c, 0x00000000, s, b); b 40 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgp100.c mmio_refn(info, 0x419004, 0x00000000, s, b); b 56 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgp100.c int gpc, ppc, b, n = 0; b 61 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgp100.c b = mmio_vram(info, size, (1 << s), false); b 63 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgp100.c mmio_refn(info, 0x418810, 0x80000000, s, b); b 64 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgp100.c mmio_refn(info, 0x419848, 0x10000000, s, b); b 65 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgp100.c mmio_refn(info, 0x419c2c, 0x10000000, s, b); b 66 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgp100.c mmio_refn(info, 0x419b00, 0x00000000, s, b); b 52 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgp102.c int gpc, ppc, b, n = 0; b 57 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgp102.c b = mmio_vram(info, size, (1 << s), false); b 59 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgp102.c mmio_refn(info, 0x418810, 0x80000000, s, b); b 60 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgp102.c mmio_refn(info, 0x419848, 0x10000000, s, b); b 61 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgp102.c mmio_refn(info, 0x419c2c, 0x10000000, s, b); b 62 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgp102.c mmio_refn(info, 0x419b00, 0x00000000, s, b); b 74 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgv100.c int gpc, ppc, b, n = 0; b 78 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgv100.c b = mmio_vram(info, size, (1 << s), false); b 80 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgv100.c mmio_refn(info, 0x418810, 0x80000000, s, b); b 81 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgv100.c mmio_refn(info, 0x419848, 0x10000000, s, b); b 82 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgv100.c mmio_refn(info, 0x419c2c, 0x10000000, s, b); b 83 drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxgv100.c mmio_refn(info, 0x419e00, 0x00000000, s, b); b 399 drivers/gpu/drm/nouveau/nvkm/subdev/bios/init.c #define init_script_table(b) init_table_((b), 0x00, "script table") b 400 drivers/gpu/drm/nouveau/nvkm/subdev/bios/init.c #define init_macro_index_table(b) init_table_((b), 0x02, "macro index table") b 401 drivers/gpu/drm/nouveau/nvkm/subdev/bios/init.c #define init_macro_table(b) init_table_((b), 0x04, "macro table") b 402 drivers/gpu/drm/nouveau/nvkm/subdev/bios/init.c #define init_condition_table(b) init_table_((b), 0x06, "condition table") b 403 drivers/gpu/drm/nouveau/nvkm/subdev/bios/init.c #define init_io_condition_table(b) init_table_((b), 0x08, "io condition table") b 404 drivers/gpu/drm/nouveau/nvkm/subdev/bios/init.c #define init_io_flag_condition_table(b) init_table_((b), 0x0a, "io flag conditon table") b 405 drivers/gpu/drm/nouveau/nvkm/subdev/bios/init.c #define init_function_table(b) init_table_((b), 0x0c, "function table") b 406 drivers/gpu/drm/nouveau/nvkm/subdev/bios/init.c #define init_xlat_table(b) init_table_((b), 0x10, "xlat table"); b 29 drivers/gpu/drm/nouveau/nvkm/subdev/bios/vpstate.c nvbios_vpstate_offset(struct nvkm_bios *b) b 33 drivers/gpu/drm/nouveau/nvkm/subdev/bios/vpstate.c if (!bit_entry(b, 'P', &bit_P)) { b 35 drivers/gpu/drm/nouveau/nvkm/subdev/bios/vpstate.c return nvbios_rd32(b, bit_P.offset + 0x38); b 42 drivers/gpu/drm/nouveau/nvkm/subdev/bios/vpstate.c nvbios_vpstate_parse(struct nvkm_bios *b, struct nvbios_vpstate_header *h) b 47 drivers/gpu/drm/nouveau/nvkm/subdev/bios/vpstate.c h->offset = nvbios_vpstate_offset(b); b 51 drivers/gpu/drm/nouveau/nvkm/subdev/bios/vpstate.c h->version = nvbios_rd08(b, h->offset); b 54 drivers/gpu/drm/nouveau/nvkm/subdev/bios/vpstate.c h->hlen = nvbios_rd08(b, h->offset + 0x1); b 55 drivers/gpu/drm/nouveau/nvkm/subdev/bios/vpstate.c h->elen = nvbios_rd08(b, h->offset + 0x2); b 56 drivers/gpu/drm/nouveau/nvkm/subdev/bios/vpstate.c h->slen = nvbios_rd08(b, h->offset + 0x3); b 57 drivers/gpu/drm/nouveau/nvkm/subdev/bios/vpstate.c h->scount = nvbios_rd08(b, h->offset + 0x4); b 58 drivers/gpu/drm/nouveau/nvkm/subdev/bios/vpstate.c h->ecount = nvbios_rd08(b, h->offset + 0x5); b 60 drivers/gpu/drm/nouveau/nvkm/subdev/bios/vpstate.c h->base_id = nvbios_rd08(b, h->offset + 0x0f); b 62 drivers/gpu/drm/nouveau/nvkm/subdev/bios/vpstate.c h->boost_id = nvbios_rd08(b, h->offset + 0x10); b 66 drivers/gpu/drm/nouveau/nvkm/subdev/bios/vpstate.c h->tdp_id = nvbios_rd08(b, h->offset + 0x11); b 76 drivers/gpu/drm/nouveau/nvkm/subdev/bios/vpstate.c nvbios_vpstate_entry(struct nvkm_bios *b, struct nvbios_vpstate_header *h, b 85 drivers/gpu/drm/nouveau/nvkm/subdev/bios/vpstate.c e->pstate = nvbios_rd08(b, offset); b 86 drivers/gpu/drm/nouveau/nvkm/subdev/bios/vpstate.c e->clock_mhz = nvbios_rd16(b, offset + 0x5); b 362 drivers/gpu/drm/nouveau/nvkm/subdev/clk/nv50.c clk_same(u32 a, u32 b) b 364 drivers/gpu/drm/nouveau/nvkm/subdev/clk/nv50.c return ((a / 1000) == (b / 1000)); b 69 drivers/gpu/drm/nouveau/nvkm/subdev/fb/nv30.c int b = (device->chipset > 0x30 ? b 74 drivers/gpu/drm/nouveau/nvkm/subdev/fb/nv30.c return 2 * (b & 0x8 ? b - 0x10 : b); b 32 drivers/gpu/drm/nouveau/nvkm/subdev/i2c/aux.h #define AUX_MSG(b,l,f,a...) do { \ b 33 drivers/gpu/drm/nouveau/nvkm/subdev/i2c/aux.h struct nvkm_i2c_aux *_aux = (b); \ b 36 drivers/gpu/drm/nouveau/nvkm/subdev/i2c/aux.h #define AUX_ERR(b,f,a...) AUX_MSG((b), error, f, ##a) b 37 drivers/gpu/drm/nouveau/nvkm/subdev/i2c/aux.h #define AUX_DBG(b,f,a...) AUX_MSG((b), debug, f, ##a) b 38 drivers/gpu/drm/nouveau/nvkm/subdev/i2c/aux.h #define AUX_TRACE(b,f,a...) AUX_MSG((b), trace, f, ##a) b 32 drivers/gpu/drm/nouveau/nvkm/subdev/i2c/bus.h #define BUS_MSG(b,l,f,a...) do { \ b 33 drivers/gpu/drm/nouveau/nvkm/subdev/i2c/bus.h struct nvkm_i2c_bus *_bus = (b); \ b 36 drivers/gpu/drm/nouveau/nvkm/subdev/i2c/bus.h #define BUS_ERR(b,f,a...) BUS_MSG((b), error, f, ##a) b 37 drivers/gpu/drm/nouveau/nvkm/subdev/i2c/bus.h #define BUS_DBG(b,f,a...) BUS_MSG((b), debug, f, ##a) b 38 drivers/gpu/drm/nouveau/nvkm/subdev/i2c/bus.h #define BUS_TRACE(b,f,a...) BUS_MSG((b), trace, f, ##a) b 324 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.h #define VMM_FO(m,o,d,c,b) nvkm_fo##b((m)->memory, (o), (d), (c)) b 325 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.h #define VMM_WO(m,o,d,c,b) nvkm_wo##b((m)->memory, (o), (d)) b 326 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.h #define VMM_XO(m,v,o,d,c,b,fn,f,a...) do { \ b 327 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.h const u32 _pteo = (o); u##b _data = (d); \ b 329 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.h VMM_##fn((m), (m)->base + _pteo, _data, (c), b); \ b 3867 drivers/gpu/drm/omapdrm/dss/dispc.c u16 r, g, b; b 3876 drivers/gpu/drm/omapdrm/dss/dispc.c b = (lut[i].blue * (w - j) + lut[i+1].blue * j) / w; b 3880 drivers/gpu/drm/omapdrm/dss/dispc.c b >>= 16 - gdesc->bits; b 3883 drivers/gpu/drm/omapdrm/dss/dispc.c (g << gdesc->bits) | b; b 2938 drivers/gpu/drm/omapdrm/dss/dsi.c int b; b 2948 drivers/gpu/drm/omapdrm/dss/dsi.c for (b = 0; b < 4; ++b) { b 2950 drivers/gpu/drm/omapdrm/dss/dsi.c buf[w] = (val >> (b * 8)) & 0xff; b 407 drivers/gpu/drm/omapdrm/dss/dss.c int b; b 418 drivers/gpu/drm/omapdrm/dss/dss.c b = 0; b 421 drivers/gpu/drm/omapdrm/dss/dss.c b = 1; b 424 drivers/gpu/drm/omapdrm/dss/dss.c b = 2; b 431 drivers/gpu/drm/omapdrm/dss/dss.c REG_FLD_MOD(dss, DSS_CONTROL, b, /* DISPC_CLK_SWITCH */ b 441 drivers/gpu/drm/omapdrm/dss/dss.c int b, pos; b 445 drivers/gpu/drm/omapdrm/dss/dss.c b = 0; b 449 drivers/gpu/drm/omapdrm/dss/dss.c b = 1; b 453 drivers/gpu/drm/omapdrm/dss/dss.c b = 1; b 461 drivers/gpu/drm/omapdrm/dss/dss.c REG_FLD_MOD(dss, DSS_CONTROL, b, pos, pos); /* DSIx_CLK_SWITCH */ b 432 drivers/gpu/drm/omapdrm/dss/dss.h int b; b 433 drivers/gpu/drm/omapdrm/dss/dss.h for (b = 0; b < 32; ++b) { b 434 drivers/gpu/drm/omapdrm/dss/dss.h if (irqstatus & (1 << b)) b 435 drivers/gpu/drm/omapdrm/dss/dss.h irq_arr[b]++; b 422 drivers/gpu/drm/omapdrm/dss/hdmi5_core.c unsigned int y, a, b, s; b 437 drivers/gpu/drm/omapdrm/dss/hdmi5_core.c b = (ptr[0] >> 2) & 0x3; b 456 drivers/gpu/drm/omapdrm/dss/hdmi5_core.c (a << 6) | (s << 4) | (b << 2) | (y << 0)); b 181 drivers/gpu/drm/omapdrm/omap_drv.c static int omap_compare_pipelines(const void *a, const void *b) b 184 drivers/gpu/drm/omapdrm/omap_drv.c const struct omap_drm_pipeline *pipe2 = b; b 282 drivers/gpu/drm/r128/r128_state.c int x, int y, int w, int h, int r, int g, int b) b 292 drivers/gpu/drm/r128/r128_state.c ((g & 0xfc) << 3) | ((b & 0xf8) >> 3)); b 296 drivers/gpu/drm/r128/r128_state.c color = ((r << 16) | (g << 8) | b); b 300 drivers/gpu/drm/r128/r128_state.c color = (((0xff) << 24) | (r << 16) | (g << 8) | b); b 9036 drivers/gpu/drm/radeon/cik.c fixed20_12 a, b; b 9042 drivers/gpu/drm/radeon/cik.c b.full = dfixed_mul(a, disp_clk); b 9048 drivers/gpu/drm/radeon/cik.c bandwidth.full = dfixed_mul(b, disp_clk_request_efficiency); b 9127 drivers/gpu/drm/radeon/cik.c fixed20_12 a, b, c; b 9133 drivers/gpu/drm/radeon/cik.c b.full = dfixed_const(1); b 9135 drivers/gpu/drm/radeon/cik.c ((wm->vsc.full > b.full) && (wm->vtaps >= 3)) || b 9143 drivers/gpu/drm/radeon/cik.c b.full = dfixed_const(wm->num_heads); b 9144 drivers/gpu/drm/radeon/cik.c a.full = dfixed_div(a, b); b 9151 drivers/gpu/drm/radeon/cik.c b.full = dfixed_const(1000); b 9153 drivers/gpu/drm/radeon/cik.c b.full = dfixed_div(c, b); b 9154 drivers/gpu/drm/radeon/cik.c a.full = dfixed_div(a, b); b 2073 drivers/gpu/drm/radeon/evergreen.c fixed20_12 a, b, c; b 2079 drivers/gpu/drm/radeon/evergreen.c b.full = dfixed_const(1); b 2081 drivers/gpu/drm/radeon/evergreen.c ((wm->vsc.full > b.full) && (wm->vtaps >= 3)) || b 2089 drivers/gpu/drm/radeon/evergreen.c b.full = dfixed_const(wm->num_heads); b 2090 drivers/gpu/drm/radeon/evergreen.c a.full = dfixed_div(a, b); b 2095 drivers/gpu/drm/radeon/evergreen.c b.full = dfixed_const(1000); b 2097 drivers/gpu/drm/radeon/evergreen.c b.full = dfixed_div(c, b); b 2098 drivers/gpu/drm/radeon/evergreen.c a.full = dfixed_div(a, b); b 2167 drivers/gpu/drm/radeon/evergreen.c fixed20_12 a, b, c; b 2256 drivers/gpu/drm/radeon/evergreen.c b.full = dfixed_const(mode->clock); b 2257 drivers/gpu/drm/radeon/evergreen.c b.full = dfixed_div(b, a); b 2259 drivers/gpu/drm/radeon/evergreen.c c.full = dfixed_mul(c, b); b 2268 drivers/gpu/drm/radeon/evergreen.c b.full = dfixed_const(mode->clock); b 2269 drivers/gpu/drm/radeon/evergreen.c b.full = dfixed_div(b, a); b 2271 drivers/gpu/drm/radeon/evergreen.c c.full = dfixed_mul(c, b); b 35 drivers/gpu/drm/radeon/evergreen_cs.c #define MAX(a,b) (((a)>(b))?(a):(b)) b 36 drivers/gpu/drm/radeon/evergreen_cs.c #define MIN(a,b) (((a)<(b))?(a):(b)) b 789 drivers/gpu/drm/radeon/kv_dpm.c static u32 kv_get_clock_difference(u32 a, u32 b) b 791 drivers/gpu/drm/radeon/kv_dpm.c return (a >= b) ? a - b : b - a; b 403 drivers/gpu/drm/radeon/radeon.h struct radeon_fence *b) b 406 drivers/gpu/drm/radeon/radeon.h return b; b 409 drivers/gpu/drm/radeon/radeon.h if (!b) { b 413 drivers/gpu/drm/radeon/radeon.h BUG_ON(a->ring != b->ring); b 415 drivers/gpu/drm/radeon/radeon.h if (a->seq > b->seq) { b 418 drivers/gpu/drm/radeon/radeon.h return b; b 423 drivers/gpu/drm/radeon/radeon.h struct radeon_fence *b) b 429 drivers/gpu/drm/radeon/radeon.h if (!b) { b 433 drivers/gpu/drm/radeon/radeon.h BUG_ON(a->ring != b->ring); b 435 drivers/gpu/drm/radeon/radeon.h return a->seq < b->seq; b 2730 drivers/gpu/drm/radeon/radeon.h #define radeon_hdmi_enable(rdev, e, b) (rdev)->asic->display.hdmi_enable((e), (b)) b 51 drivers/gpu/drm/radeon/radeon_cs.c static void radeon_cs_buckets_init(struct radeon_cs_buckets *b) b 56 drivers/gpu/drm/radeon/radeon_cs.c INIT_LIST_HEAD(&b->bucket[i]); b 59 drivers/gpu/drm/radeon/radeon_cs.c static void radeon_cs_buckets_add(struct radeon_cs_buckets *b, b 67 drivers/gpu/drm/radeon/radeon_cs.c list_add_tail(item, &b->bucket[min(priority, RADEON_CS_MAX_PRIORITY)]); b 70 drivers/gpu/drm/radeon/radeon_cs.c static void radeon_cs_buckets_get_list(struct radeon_cs_buckets *b, b 77 drivers/gpu/drm/radeon/radeon_cs.c list_splice(&b->bucket[i], out_list); b 398 drivers/gpu/drm/radeon/radeon_cs.c struct list_head *b) b 401 drivers/gpu/drm/radeon/radeon_cs.c struct radeon_bo_list *lb = list_entry(b, struct radeon_bo_list, tv.head); b 53 drivers/gpu/drm/radeon/radeon_display.c u16 *r, *g, *b; b 74 drivers/gpu/drm/radeon/radeon_display.c b = g + crtc->gamma_size; b 79 drivers/gpu/drm/radeon/radeon_display.c (*b++ >> 6)); b 91 drivers/gpu/drm/radeon/radeon_display.c u16 *r, *g, *b; b 111 drivers/gpu/drm/radeon/radeon_display.c b = g + crtc->gamma_size; b 116 drivers/gpu/drm/radeon/radeon_display.c (*b++ >> 6)); b 125 drivers/gpu/drm/radeon/radeon_display.c u16 *r, *g, *b; b 159 drivers/gpu/drm/radeon/radeon_display.c b = g + crtc->gamma_size; b 164 drivers/gpu/drm/radeon/radeon_display.c (*b++ >> 6)); b 197 drivers/gpu/drm/radeon/radeon_display.c u16 *r, *g, *b; b 211 drivers/gpu/drm/radeon/radeon_display.c b = g + crtc->gamma_size; b 216 drivers/gpu/drm/radeon/radeon_display.c (*b++ >> 6)); b 1761 drivers/gpu/drm/radeon/radeon_display.c fixed20_12 a, b; b 1763 drivers/gpu/drm/radeon/radeon_display.c b.full = dfixed_const(dst_v); b 1764 drivers/gpu/drm/radeon/radeon_display.c radeon_crtc->vsc.full = dfixed_div(a, b); b 1766 drivers/gpu/drm/radeon/radeon_display.c b.full = dfixed_const(dst_h); b 1767 drivers/gpu/drm/radeon/radeon_display.c radeon_crtc->hsc.full = dfixed_div(a, b); b 278 drivers/gpu/drm/radeon/rs690.c fixed20_12 a, b, c; b 310 drivers/gpu/drm/radeon/rs690.c b.full = dfixed_const(mode->crtc_hdisplay); b 312 drivers/gpu/drm/radeon/rs690.c a.full = dfixed_div(b, c); b 328 drivers/gpu/drm/radeon/rs690.c b.full = dfixed_const(1000); b 329 drivers/gpu/drm/radeon/rs690.c a.full = dfixed_div(a, b); b 330 drivers/gpu/drm/radeon/rs690.c pclk.full = dfixed_div(b, a); b 332 drivers/gpu/drm/radeon/rs690.c b.full = dfixed_const(2); b 333 drivers/gpu/drm/radeon/rs690.c if (crtc->vsc.full > b.full) b 334 drivers/gpu/drm/radeon/rs690.c b.full = crtc->vsc.full; b 335 drivers/gpu/drm/radeon/rs690.c b.full = dfixed_mul(b, crtc->hsc); b 337 drivers/gpu/drm/radeon/rs690.c b.full = dfixed_div(b, c); b 338 drivers/gpu/drm/radeon/rs690.c consumption_time.full = dfixed_div(pclk, b); b 360 drivers/gpu/drm/radeon/rs690.c b.full = dfixed_const(crtc->base.mode.crtc_hdisplay); b 361 drivers/gpu/drm/radeon/rs690.c wm->active_time.full = dfixed_mul(line_time, b); b 372 drivers/gpu/drm/radeon/rs690.c b.full = dfixed_div(rdev->pm.igp_sideport_mclk, a); b 373 drivers/gpu/drm/radeon/rs690.c read_delay_latency.full = dfixed_div(read_delay_latency, b); b 469 drivers/gpu/drm/radeon/rs690.c fixed20_12 a, b; b 480 drivers/gpu/drm/radeon/rs690.c b.full = dfixed_mul(wm1->dbpp, wm1->num_line_pair); b 482 drivers/gpu/drm/radeon/rs690.c b.full = wm1->num_line_pair.full; b 483 drivers/gpu/drm/radeon/rs690.c a.full += b.full; b 486 drivers/gpu/drm/radeon/rs690.c b.full = wm0->consumption_rate.full - fill_rate.full; b 487 drivers/gpu/drm/radeon/rs690.c b.full = dfixed_mul(b, wm0->active_time); b 490 drivers/gpu/drm/radeon/rs690.c a.full = a.full + b.full; b 491 drivers/gpu/drm/radeon/rs690.c b.full = dfixed_const(16 * 1000); b 492 drivers/gpu/drm/radeon/rs690.c priority_mark02.full = dfixed_div(a, b); b 496 drivers/gpu/drm/radeon/rs690.c b.full = dfixed_const(16 * 1000); b 497 drivers/gpu/drm/radeon/rs690.c priority_mark02.full = dfixed_div(a, b); b 500 drivers/gpu/drm/radeon/rs690.c b.full = wm1->consumption_rate.full - fill_rate.full; b 501 drivers/gpu/drm/radeon/rs690.c b.full = dfixed_mul(b, wm1->active_time); b 504 drivers/gpu/drm/radeon/rs690.c a.full = a.full + b.full; b 505 drivers/gpu/drm/radeon/rs690.c b.full = dfixed_const(16 * 1000); b 506 drivers/gpu/drm/radeon/rs690.c priority_mark12.full = dfixed_div(a, b); b 510 drivers/gpu/drm/radeon/rs690.c b.full = dfixed_const(16 * 1000); b 511 drivers/gpu/drm/radeon/rs690.c priority_mark12.full = dfixed_div(a, b); b 534 drivers/gpu/drm/radeon/rs690.c b.full = wm0->consumption_rate.full - fill_rate.full; b 535 drivers/gpu/drm/radeon/rs690.c b.full = dfixed_mul(b, wm0->active_time); b 538 drivers/gpu/drm/radeon/rs690.c a.full = a.full + b.full; b 539 drivers/gpu/drm/radeon/rs690.c b.full = dfixed_const(16 * 1000); b 540 drivers/gpu/drm/radeon/rs690.c priority_mark02.full = dfixed_div(a, b); b 544 drivers/gpu/drm/radeon/rs690.c b.full = dfixed_const(16 * 1000); b 545 drivers/gpu/drm/radeon/rs690.c priority_mark02.full = dfixed_div(a, b); b 561 drivers/gpu/drm/radeon/rs690.c b.full = wm1->consumption_rate.full - fill_rate.full; b 562 drivers/gpu/drm/radeon/rs690.c b.full = dfixed_mul(b, wm1->active_time); b 565 drivers/gpu/drm/radeon/rs690.c a.full = a.full + b.full; b 566 drivers/gpu/drm/radeon/rs690.c b.full = dfixed_const(16 * 1000); b 567 drivers/gpu/drm/radeon/rs690.c priority_mark12.full = dfixed_div(a, b); b 571 drivers/gpu/drm/radeon/rs690.c b.full = dfixed_const(16 * 1000); b 572 drivers/gpu/drm/radeon/rs690.c priority_mark12.full = dfixed_div(a, b); b 961 drivers/gpu/drm/radeon/rv515.c fixed20_12 a, b, c; b 990 drivers/gpu/drm/radeon/rv515.c b.full = dfixed_const(mode->crtc_hdisplay); b 992 drivers/gpu/drm/radeon/rv515.c a.full = dfixed_div(b, c); b 1008 drivers/gpu/drm/radeon/rv515.c b.full = dfixed_const(1000); b 1009 drivers/gpu/drm/radeon/rv515.c a.full = dfixed_div(a, b); b 1010 drivers/gpu/drm/radeon/rv515.c pclk.full = dfixed_div(b, a); b 1012 drivers/gpu/drm/radeon/rv515.c b.full = dfixed_const(2); b 1013 drivers/gpu/drm/radeon/rv515.c if (crtc->vsc.full > b.full) b 1014 drivers/gpu/drm/radeon/rv515.c b.full = crtc->vsc.full; b 1015 drivers/gpu/drm/radeon/rv515.c b.full = dfixed_mul(b, crtc->hsc); b 1017 drivers/gpu/drm/radeon/rv515.c b.full = dfixed_div(b, c); b 1018 drivers/gpu/drm/radeon/rv515.c consumption_time.full = dfixed_div(pclk, b); b 1040 drivers/gpu/drm/radeon/rv515.c b.full = dfixed_const(crtc->base.mode.crtc_hdisplay); b 1041 drivers/gpu/drm/radeon/rv515.c wm->active_time.full = dfixed_mul(line_time, b); b 1120 drivers/gpu/drm/radeon/rv515.c fixed20_12 a, b; b 1131 drivers/gpu/drm/radeon/rv515.c b.full = dfixed_div(wm1->dbpp, wm1->num_line_pair); b 1133 drivers/gpu/drm/radeon/rv515.c b.full = wm1->num_line_pair.full; b 1134 drivers/gpu/drm/radeon/rv515.c a.full += b.full; b 1137 drivers/gpu/drm/radeon/rv515.c b.full = wm0->consumption_rate.full - fill_rate.full; b 1138 drivers/gpu/drm/radeon/rv515.c b.full = dfixed_mul(b, wm0->active_time); b 1140 drivers/gpu/drm/radeon/rv515.c b.full = dfixed_div(b, a); b 1143 drivers/gpu/drm/radeon/rv515.c priority_mark02.full = a.full + b.full; b 1147 drivers/gpu/drm/radeon/rv515.c b.full = dfixed_const(16 * 1000); b 1148 drivers/gpu/drm/radeon/rv515.c priority_mark02.full = dfixed_div(a, b); b 1151 drivers/gpu/drm/radeon/rv515.c b.full = wm1->consumption_rate.full - fill_rate.full; b 1152 drivers/gpu/drm/radeon/rv515.c b.full = dfixed_mul(b, wm1->active_time); b 1154 drivers/gpu/drm/radeon/rv515.c b.full = dfixed_div(b, a); b 1157 drivers/gpu/drm/radeon/rv515.c priority_mark12.full = a.full + b.full; b 1161 drivers/gpu/drm/radeon/rv515.c b.full = dfixed_const(16 * 1000); b 1162 drivers/gpu/drm/radeon/rv515.c priority_mark12.full = dfixed_div(a, b); b 1185 drivers/gpu/drm/radeon/rv515.c b.full = wm0->consumption_rate.full - fill_rate.full; b 1186 drivers/gpu/drm/radeon/rv515.c b.full = dfixed_mul(b, wm0->active_time); b 1188 drivers/gpu/drm/radeon/rv515.c b.full = dfixed_div(b, a); b 1191 drivers/gpu/drm/radeon/rv515.c priority_mark02.full = a.full + b.full; b 1195 drivers/gpu/drm/radeon/rv515.c b.full = dfixed_const(16); b 1196 drivers/gpu/drm/radeon/rv515.c priority_mark02.full = dfixed_div(a, b); b 1212 drivers/gpu/drm/radeon/rv515.c b.full = wm1->consumption_rate.full - fill_rate.full; b 1213 drivers/gpu/drm/radeon/rv515.c b.full = dfixed_mul(b, wm1->active_time); b 1215 drivers/gpu/drm/radeon/rv515.c b.full = dfixed_div(b, a); b 1218 drivers/gpu/drm/radeon/rv515.c priority_mark12.full = a.full + b.full; b 1222 drivers/gpu/drm/radeon/rv515.c b.full = dfixed_const(16 * 1000); b 1223 drivers/gpu/drm/radeon/rv515.c priority_mark12.full = dfixed_div(a, b); b 2217 drivers/gpu/drm/radeon/si.c fixed20_12 a, b, c; b 2223 drivers/gpu/drm/radeon/si.c b.full = dfixed_const(1); b 2225 drivers/gpu/drm/radeon/si.c ((wm->vsc.full > b.full) && (wm->vtaps >= 3)) || b 2233 drivers/gpu/drm/radeon/si.c b.full = dfixed_const(wm->num_heads); b 2234 drivers/gpu/drm/radeon/si.c a.full = dfixed_div(a, b); b 2241 drivers/gpu/drm/radeon/si.c b.full = dfixed_const(1000); b 2243 drivers/gpu/drm/radeon/si.c b.full = dfixed_div(c, b); b 2244 drivers/gpu/drm/radeon/si.c a.full = dfixed_div(a, b); b 2312 drivers/gpu/drm/radeon/si.c fixed20_12 a, b, c; b 2407 drivers/gpu/drm/radeon/si.c b.full = dfixed_const(mode->clock); b 2408 drivers/gpu/drm/radeon/si.c b.full = dfixed_div(b, a); b 2410 drivers/gpu/drm/radeon/si.c c.full = dfixed_mul(c, b); b 2419 drivers/gpu/drm/radeon/si.c b.full = dfixed_const(mode->clock); b 2420 drivers/gpu/drm/radeon/si.c b.full = dfixed_div(b, a); b 2422 drivers/gpu/drm/radeon/si.c c.full = dfixed_mul(c, b); b 1853 drivers/gpu/drm/radeon/sid.h #define DMA_PACKET(cmd, b, t, s, n) ((((cmd) & 0xF) << 28) | \ b 1854 drivers/gpu/drm/radeon/sid.h (((b) & 0x1) << 26) | \ b 329 drivers/gpu/drm/rcar-du/rcar_du_regs.h #define DOOR_RGB(r, g, b) (((r) << 18) | ((g) << 10) | ((b) << 2)) b 331 drivers/gpu/drm/rcar-du/rcar_du_regs.h #define CDER_RGB(r, g, b) (((r) << 18) | ((g) << 10) | ((b) << 2)) b 333 drivers/gpu/drm/rcar-du/rcar_du_regs.h #define BPOR_RGB(r, g, b) (((r) << 18) | ((g) << 10) | ((b) << 2)) b 75 drivers/gpu/drm/rcar-du/rcar_lvds.c #define bridge_to_rcar_lvds(b) \ b 76 drivers/gpu/drm/rcar-du/rcar_lvds.c container_of(b, struct rcar_lvds, bridge) b 447 drivers/gpu/drm/savage/savage_drv.h #define BCI_CLIP_BR(b, r) ((((b) << 16) | (r)) & 0x0FFF0FFF) b 151 drivers/gpu/drm/sti/sti_cursor.c u32 a, r, g, b; b 159 drivers/gpu/drm/sti/sti_cursor.c b = (*src >> 6) & 3; b 160 drivers/gpu/drm/sti/sti_cursor.c *dst = a << 6 | r << 4 | g << 2 | b; b 170 drivers/gpu/drm/sti/sti_cursor.c unsigned int a, r, g, b; b 176 drivers/gpu/drm/sti/sti_cursor.c for (b = 0; b < 4; b++) b 180 drivers/gpu/drm/sti/sti_cursor.c (b * 5); b 284 drivers/gpu/drm/sti/sti_gdp.c unsigned int b; b 286 drivers/gpu/drm/sti/sti_gdp.c for (b = 0; b < GDP_NODE_NB_BANK; b++) { b 287 drivers/gpu/drm/sti/sti_gdp.c seq_printf(s, "\n%s[%d].top", sti_plane_to_str(&gdp->plane), b); b 288 drivers/gpu/drm/sti/sti_gdp.c gdp_node_dump_node(s, gdp->node_list[b].top_field); b 289 drivers/gpu/drm/sti/sti_gdp.c seq_printf(s, "\n%s[%d].btm", sti_plane_to_str(&gdp->plane), b); b 290 drivers/gpu/drm/sti/sti_gdp.c gdp_node_dump_node(s, gdp->node_list[b].btm_field); b 454 drivers/gpu/drm/sti/sti_tvout.c int r, g, b, tmp, mask; b 472 drivers/gpu/drm/sti/sti_tvout.c b = (val & mask) >> TVO_VIP_REORDER_B_SHIFT; b 476 drivers/gpu/drm/sti/sti_tvout.c reorder[b], reorder[TVO_VIP_REORDER_CB_B_SEL]); b 37 drivers/gpu/drm/sun4i/sun4i_backend.h #define SUN4I_BACKEND_BACKCOLOR(r, g, b) (((r) << 16) | ((g) << 8) | (b)) b 1641 drivers/gpu/drm/tegra/sor.c unsigned int hfp, hsw, hbp, a = 0, b; b 1649 drivers/gpu/drm/tegra/sor.c b = hfp - 1; b 1651 drivers/gpu/drm/tegra/sor.c pr_info("a: %u, b: %u\n", a, b); b 1659 drivers/gpu/drm/tegra/sor.c if (a > b) b 1669 drivers/gpu/drm/tegra/sor.c if (b > a && a % 2) b 203 drivers/gpu/drm/tiny/repaper.c unsigned int b; b 205 drivers/gpu/drm/tiny/repaper.c for (b = 0; b < (epd->width / 8); b++) { b 207 drivers/gpu/drm/tiny/repaper.c u8 pixels = data[b] & 0xaa; b 212 drivers/gpu/drm/tiny/repaper.c pixel_mask = (mask[b] ^ pixels) & 0xaa; b 249 drivers/gpu/drm/tiny/repaper.c unsigned int b; b 251 drivers/gpu/drm/tiny/repaper.c for (b = epd->width / 8; b > 0; b--) { b 253 drivers/gpu/drm/tiny/repaper.c u8 pixels = data[b - 1] & 0x55; b 257 drivers/gpu/drm/tiny/repaper.c pixel_mask = (mask[b - 1] ^ pixels) & 0x55; b 299 drivers/gpu/drm/tiny/repaper.c unsigned int b; b 301 drivers/gpu/drm/tiny/repaper.c for (b = epd->width / 8; b > 0; b--) { b 303 drivers/gpu/drm/tiny/repaper.c u16 pixels = repaper_interleave_bits(data[b - 1]); b 307 drivers/gpu/drm/tiny/repaper.c pixel_mask = repaper_interleave_bits(mask[b - 1]); b 344 drivers/gpu/drm/tiny/repaper.c unsigned int b; b 356 drivers/gpu/drm/tiny/repaper.c for (b = epd->bytes_per_scan; b > 0; b--) { b 357 drivers/gpu/drm/tiny/repaper.c if (line / 4 == b - 1) b 370 drivers/gpu/drm/tiny/repaper.c for (b = 0; b < epd->bytes_per_scan; b++) { b 371 drivers/gpu/drm/tiny/repaper.c if (0 != (line & 0x01) && line / 8 == b) b 384 drivers/gpu/drm/tiny/repaper.c for (b = epd->bytes_per_scan; b > 0; b--) { b 385 drivers/gpu/drm/tiny/repaper.c if (0 == (line & 0x01) && line / 8 == b - 1) b 475 drivers/gpu/drm/vmwgfx/device_include/svga3d_devcaps.h Bool b; b 141 drivers/gpu/drm/vmwgfx/device_include/svga3d_dx.h float b; b 1108 drivers/gpu/drm/vmwgfx/device_include/svga3d_surfacedefs.h static inline u32 clamped_umul32(u32 a, u32 b) b 1110 drivers/gpu/drm/vmwgfx/device_include/svga3d_surfacedefs.h uint64_t tmp = (uint64_t) a*b; b 684 drivers/gpu/drm/vmwgfx/device_include/svga_reg.h uint32 b : 8; b 265 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c const struct vmw_binding_info *b = &vmw_binding_infos[bt]; b 266 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c size_t offset = b->offsets[shader_slot] + b->size*slot; b 303 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c const struct vmw_binding_info *b = &vmw_binding_infos[bi->bt]; b 308 drivers/gpu/drm/vmwgfx/vmwgfx_binding.c memcpy(loc, bi, b->size); b 157 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c static size_t vmw_ptr_diff(void *a, void *b) b 159 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c return (unsigned long) b - (unsigned long) a; b 2096 drivers/gpu/drm/vmwgfx/vmwgfx_kms.c u16 *r, u16 *g, u16 *b, b 2105 drivers/gpu/drm/vmwgfx/vmwgfx_kms.c r[i], g[i], b[i]); b 2108 drivers/gpu/drm/vmwgfx/vmwgfx_kms.c vmw_write(dev_priv, SVGA_PALETTE_BASE + i * 3 + 2, b[i] >> 8); b 393 drivers/gpu/drm/vmwgfx/vmwgfx_kms.h u16 *r, u16 *g, u16 *b, b 717 drivers/hid/hid-core.c u8 b; b 722 drivers/hid/hid-core.c b = *start++; b 724 drivers/hid/hid-core.c item->type = (b >> 2) & 3; b 725 drivers/hid/hid-core.c item->tag = (b >> 4) & 15; b 746 drivers/hid/hid-core.c item->size = b & 3; b 1179 drivers/hid/hid-input.c int b = field->logical_maximum; b 1183 drivers/hid/hid-input.c b = field->logical_maximum = 255; b 1187 drivers/hid/hid-input.c input_set_abs_params(input, usage->code, a, b, (b - a) >> 8, (b - a) >> 4); b 1188 drivers/hid/hid-input.c else input_set_abs_params(input, usage->code, a, b, 0, 0); b 1308 drivers/hid/hid-input.c int b = field->logical_maximum; b 1309 drivers/hid/hid-input.c input_event(input, EV_KEY, BTN_TOUCH, value > a + ((b - a) >> 3)); b 43 drivers/hid/hid-led.c #define RISO_KAGAKU_IX(r, g, b) riso_kagaku_tbl[((r)?1:0)+((g)?2:0)+((b)?4:0)] b 177 drivers/hid/hid-led.c enum led_brightness r, g, b; b 181 drivers/hid/hid-led.c b = rgb->blue.cdev.brightness; b 184 drivers/hid/hid-led.c return RISO_KAGAKU_IX(r, b, g); b 186 drivers/hid/hid-led.c return RISO_KAGAKU_IX(r, g, b); b 96 drivers/hid/hid-ntrig.c __u8 b = (raw[0] & 0x3c) >> 2; b 106 drivers/hid/hid-ntrig.c return sprintf(buf, "%u.%u.%u.%u.%u", a, b, c, d, e); b 135 drivers/hid/hid-picolcd.h #define hid_hw_request(a, b, c) \ b 137 drivers/hid/hid-picolcd.h picolcd_debug_out_report(hid_get_drvdata(a), a, b); \ b 138 drivers/hid/hid-picolcd.h hid_hw_request(a, b, c); \ b 144 drivers/hid/hid-picolcd_fb.c int i, b, changed = 0; b 149 drivers/hid/hid-picolcd_fb.c for (b = 7; b >= 0; b--) { b 150 drivers/hid/hid-picolcd_fb.c const u8 *bdata = bitmap + tile * 256 + chip * 8 + b * 32; b 157 drivers/hid/hid-picolcd_fb.c for (b = 7; b >= 0; b--) { b 158 drivers/hid/hid-picolcd_fb.c const u8 *bdata = bitmap + (tile * 256 + chip * 8 + b * 32) * 8; b 394 drivers/hid/hid-picolcd_fb.c int i, b; b 397 drivers/hid/hid-picolcd_fb.c for (b = 0; b < 8; b++) { b 399 drivers/hid/hid-picolcd_fb.c p |= o_fb[i*8+b] ? 0x01 : 0x00; b 472 drivers/hid/hid-sony.c u8 r, g, b; b 2204 drivers/hid/hid-sony.c report->b = sc->led_state[2]; b 245 drivers/hid/hid-uclogic-params.c const __u8 *b = p; b 246 drivers/hid/hid-uclogic-params.c return b[0] | (b[1] << 8UL) | (b[2] << 16UL); b 1978 drivers/hid/hid-wiimote-modules.c __u8 o, r, y, g, b, bass, bm, bp; b 2021 drivers/hid/hid-wiimote-modules.c b = !(ext[5] & 0x08); b 2044 drivers/hid/hid-wiimote-modules.c b, &wdata->state.pressure_drums[4], b 26 drivers/hid/intel-ish-hid/ishtp/hbm.c int b; b 29 drivers/hid/intel-ish-hid/ishtp/hbm.c for_each_set_bit(b, dev->fw_clients_map, ISHTP_CLIENTS_MAX) b 265 drivers/hid/wacom_sys.c struct hid_field *b = field->report->field[field->index + 2]; b 270 drivers/hid/wacom_sys.c b->maxusage > 0 && b 271 drivers/hid/wacom_sys.c b->usage[0].hid == 0xFF000000 && b 272 drivers/hid/wacom_sys.c b->report_size == 8) { b 279 drivers/hid/wacom_sys.c b->usage[0].hid = WACOM_HID_WD_SERIALHI; b 280 drivers/hid/wacom_sys.c b->logical_minimum = 0; b 281 drivers/hid/wacom_sys.c b->logical_maximum = U8_MAX; b 48 drivers/hv/hv_trace.h &offer->offer.if_type.b, 16); b 50 drivers/hv/hv_trace.h &offer->offer.if_instance.b, 16); b 289 drivers/hv/hv_trace.h memcpy(__entry->guest_id, &msg->guest_endpoint_id.b, 16); b 290 drivers/hv/hv_trace.h memcpy(__entry->host_id, &msg->host_service_id.b, 16); b 125 drivers/hv/vmbus_drv.c sprintf(&alias_name[i], "%02x", hv_dev->dev_type.b[i/2]); b 204 drivers/hv/vmbus_drv.c hv_dev->channel->offermsg.offer.if_type.b); b 216 drivers/hv/vmbus_drv.c hv_dev->channel->offermsg.offer.if_instance.b); b 1832 drivers/hv/vmbus_drv.c child_device_obj->channel->offermsg.offer.if_instance.b); b 105 drivers/hwmon/pmbus/adm1275.c s16 b; b 750 drivers/hwmon/pmbus/adm1275.c info->b[PSC_VOLTAGE_IN] = coefficients[vindex].b; b 755 drivers/hwmon/pmbus/adm1275.c info->b[PSC_VOLTAGE_OUT] = coefficients[voindex].b; b 762 drivers/hwmon/pmbus/adm1275.c info->b[PSC_CURRENT_OUT] = coefficients[cindex].b; b 768 drivers/hwmon/pmbus/adm1275.c info->b[PSC_POWER] = coefficients[pindex].b; b 773 drivers/hwmon/pmbus/adm1275.c info->b[PSC_TEMPERATURE] = coefficients[tindex].b; b 29 drivers/hwmon/pmbus/ir38064.c .b[PSC_VOLTAGE_OUT] = 0, b 115 drivers/hwmon/pmbus/isl68137.c .b[PSC_VOLTAGE_IN] = 0, b 118 drivers/hwmon/pmbus/isl68137.c .b[PSC_VOLTAGE_OUT] = 0, b 121 drivers/hwmon/pmbus/isl68137.c .b[PSC_CURRENT_IN] = 0, b 124 drivers/hwmon/pmbus/isl68137.c .b[PSC_CURRENT_OUT] = 0, b 127 drivers/hwmon/pmbus/isl68137.c .b[PSC_POWER] = 0, b 130 drivers/hwmon/pmbus/isl68137.c .b[PSC_TEMPERATURE] = 0, b 48 drivers/hwmon/pmbus/lm25066.c short m, b, R; b 78 drivers/hwmon/pmbus/lm25066.c .b = -14500, b 172 drivers/hwmon/pmbus/lm25066.c .b = -140, b 177 drivers/hwmon/pmbus/lm25066.c .b = 500, b 182 drivers/hwmon/pmbus/lm25066.c .b = -504, b 187 drivers/hwmon/pmbus/lm25066.c .b = 100, b 192 drivers/hwmon/pmbus/lm25066.c .b = -4000, b 197 drivers/hwmon/pmbus/lm25066.c .b = -965, b 459 drivers/hwmon/pmbus/lm25066.c info->b[PSC_TEMPERATURE] = coeff[PSC_TEMPERATURE].b; b 462 drivers/hwmon/pmbus/lm25066.c info->b[PSC_VOLTAGE_IN] = coeff[PSC_VOLTAGE_IN].b; b 465 drivers/hwmon/pmbus/lm25066.c info->b[PSC_VOLTAGE_OUT] = coeff[PSC_VOLTAGE_OUT].b; b 471 drivers/hwmon/pmbus/lm25066.c info->b[PSC_CURRENT_IN] = coeff[PSC_CURRENT_IN_L].b; b 473 drivers/hwmon/pmbus/lm25066.c info->b[PSC_POWER] = coeff[PSC_POWER_L].b; b 476 drivers/hwmon/pmbus/lm25066.c info->b[PSC_CURRENT_IN] = coeff[PSC_CURRENT_IN].b; b 478 drivers/hwmon/pmbus/lm25066.c info->b[PSC_POWER] = coeff[PSC_POWER].b; b 153 drivers/hwmon/pmbus/ltc3815.c .b[PSC_VOLTAGE_IN] = 0, b 156 drivers/hwmon/pmbus/ltc3815.c .b[PSC_VOLTAGE_OUT] = 0, b 159 drivers/hwmon/pmbus/ltc3815.c .b[PSC_CURRENT_IN] = 0, b 162 drivers/hwmon/pmbus/ltc3815.c .b[PSC_CURRENT_OUT] = 0, b 165 drivers/hwmon/pmbus/ltc3815.c .b[PSC_TEMPERATURE] = 0, b 70 drivers/hwmon/pmbus/max16064.c .b[PSC_VOLTAGE_IN] = 0, b 73 drivers/hwmon/pmbus/max16064.c .b[PSC_VOLTAGE_OUT] = 0, b 76 drivers/hwmon/pmbus/max16064.c .b[PSC_TEMPERATURE] = 335, b 258 drivers/hwmon/pmbus/max31785.c .b[PSC_FAN] = 0, b 263 drivers/hwmon/pmbus/max31785.c .b[PSC_PWM] = 0, b 274 drivers/hwmon/pmbus/max31785.c .b[PSC_TEMPERATURE] = 0, b 290 drivers/hwmon/pmbus/max31785.c .b[PSC_VOLTAGE_OUT] = 0, b 255 drivers/hwmon/pmbus/max34440.c .b[PSC_VOLTAGE_IN] = 0, b 258 drivers/hwmon/pmbus/max34440.c .b[PSC_VOLTAGE_OUT] = 0, b 261 drivers/hwmon/pmbus/max34440.c .b[PSC_CURRENT_OUT] = 0, b 264 drivers/hwmon/pmbus/max34440.c .b[PSC_TEMPERATURE] = 0, b 298 drivers/hwmon/pmbus/max34440.c .b[PSC_VOLTAGE_IN] = 0, b 301 drivers/hwmon/pmbus/max34440.c .b[PSC_VOLTAGE_OUT] = 0, b 304 drivers/hwmon/pmbus/max34440.c .b[PSC_CURRENT_OUT] = 0, b 307 drivers/hwmon/pmbus/max34440.c .b[PSC_TEMPERATURE] = 0, b 310 drivers/hwmon/pmbus/max34440.c .b[PSC_FAN] = 0, b 341 drivers/hwmon/pmbus/max34440.c .b[PSC_VOLTAGE_IN] = 0, b 344 drivers/hwmon/pmbus/max34440.c .b[PSC_VOLTAGE_OUT] = 0, b 347 drivers/hwmon/pmbus/max34440.c .b[PSC_CURRENT_OUT] = 0, b 350 drivers/hwmon/pmbus/max34440.c .b[PSC_POWER] = 0, b 353 drivers/hwmon/pmbus/max34440.c .b[PSC_TEMPERATURE] = 0, b 376 drivers/hwmon/pmbus/max34440.c .b[PSC_VOLTAGE_OUT] = 0, b 379 drivers/hwmon/pmbus/max34440.c .b[PSC_CURRENT_OUT] = 0, b 382 drivers/hwmon/pmbus/max34440.c .b[PSC_TEMPERATURE] = 0, b 399 drivers/hwmon/pmbus/max34440.c .b[PSC_VOLTAGE_OUT] = 0, b 402 drivers/hwmon/pmbus/max34440.c .b[PSC_TEMPERATURE] = 0, b 430 drivers/hwmon/pmbus/max34440.c .b[PSC_VOLTAGE_OUT] = 0, b 433 drivers/hwmon/pmbus/max34440.c .b[PSC_TEMPERATURE] = 0, b 146 drivers/hwmon/pmbus/max8688.c .b[PSC_VOLTAGE_IN] = 0, b 149 drivers/hwmon/pmbus/max8688.c .b[PSC_VOLTAGE_OUT] = 0, b 152 drivers/hwmon/pmbus/max8688.c .b[PSC_CURRENT_OUT] = 0, b 155 drivers/hwmon/pmbus/max8688.c .b[PSC_TEMPERATURE] = 335, b 391 drivers/hwmon/pmbus/pmbus.h int b[PSC_NUM_CLASSES]; /* offset */ b 652 drivers/hwmon/pmbus/pmbus_core.c s64 b, val = (s16)sensor->data; b 656 drivers/hwmon/pmbus/pmbus_core.c b = data->info->b[sensor->class]; b 667 drivers/hwmon/pmbus/pmbus_core.c b *= 1000; b 673 drivers/hwmon/pmbus/pmbus_core.c b *= 1000; b 685 drivers/hwmon/pmbus/pmbus_core.c val = div_s64(val - b, m); b 813 drivers/hwmon/pmbus/pmbus_core.c s64 b, val64 = val; b 817 drivers/hwmon/pmbus/pmbus_core.c b = data->info->b[sensor->class]; b 823 drivers/hwmon/pmbus/pmbus_core.c b *= 1000; b 829 drivers/hwmon/pmbus/pmbus_core.c b *= 1000; b 831 drivers/hwmon/pmbus/pmbus_core.c val64 = val64 * m + b; b 899 drivers/hwmon/pmbus/pmbus_core.c static int pmbus_get_boolean(struct pmbus_data *data, struct pmbus_boolean *b, b 902 drivers/hwmon/pmbus/pmbus_core.c struct pmbus_sensor *s1 = b->s1; b 903 drivers/hwmon/pmbus/pmbus_core.c struct pmbus_sensor *s2 = b->s2; b 917 drivers/hwmon/pmbus/pmbus_core.c WARN(1, "Bad boolean descriptor %p: s1=%p, s2=%p\n", b, s1, s2); b 95 drivers/hwtracing/stm/p_sys-t.c generate_random_uuid(pn->uuid.b); b 325 drivers/hwtracing/stm/p_sys-t.c sz = stm_data_write(data, m, c, false, op->node.uuid.b, UUID_SIZE); b 125 drivers/i2c/algos/i2c-algo-pca.c __u8 b) b 128 drivers/i2c/algos/i2c-algo-pca.c DEB2("=== WRITE %#04x\n", b); b 129 drivers/i2c/algos/i2c-algo-pca.c pca_outw(adap, I2C_PCA_DAT, b); b 143 drivers/i2c/algos/i2c-algo-pca.c __u8 *b, int ack) b 145 drivers/i2c/algos/i2c-algo-pca.c *b = pca_inw(adap, I2C_PCA_DAT); b 146 drivers/i2c/algos/i2c-algo-pca.c DEB2("=== READ %#04x %s\n", *b, ack ? "ACK" : "NACK"); b 674 drivers/i2c/busses/i2c-cadence.c unsigned int *a, unsigned int *b) b 715 drivers/i2c/busses/i2c-cadence.c *b = calc_div_b; b 72 drivers/i2c/busses/i2c-designware-common.c void dw_writel(struct dw_i2c_dev *dev, u32 b, int offset) b 75 drivers/i2c/busses/i2c-designware-common.c b = swab32(b); b 78 drivers/i2c/busses/i2c-designware-common.c writew_relaxed((u16)b, dev->base + offset); b 79 drivers/i2c/busses/i2c-designware-common.c writew_relaxed((u16)(b >> 16), dev->base + offset + 2); b 81 drivers/i2c/busses/i2c-designware-common.c writel_relaxed(b, dev->base + offset); b 289 drivers/i2c/busses/i2c-designware-core.h void dw_writel(struct dw_i2c_dev *dev, u32 b, int offset); b 406 drivers/i2c/busses/i2c-sis630.c unsigned char b; b 433 drivers/i2c/busses/i2c-sis630.c if (pci_read_config_byte(sis630_dev, SIS630_BIOS_CTL_REG, &b)) { b 439 drivers/i2c/busses/i2c-sis630.c if (!(b & 0x80) && b 440 drivers/i2c/busses/i2c-sis630.c pci_write_config_byte(sis630_dev, SIS630_BIOS_CTL_REG, b | 0x80)) { b 94 drivers/i2c/i2c-stub.c struct smbus_block_data *b, *rb = NULL; b 96 drivers/i2c/i2c-stub.c list_for_each_entry(b, &chip->smbus_blocks, node) { b 97 drivers/i2c/i2c-stub.c if (b->command == command) { b 98 drivers/i2c/i2c-stub.c rb = b; b 130 drivers/i2c/i2c-stub.c struct smbus_block_data *b; b 247 drivers/i2c/i2c-stub.c b = stub_find_block(&adap->dev, chip, command, false); b 254 drivers/i2c/i2c-stub.c if (b == NULL) { b 255 drivers/i2c/i2c-stub.c b = stub_find_block(&adap->dev, chip, command, b 257 drivers/i2c/i2c-stub.c if (b == NULL) { b 263 drivers/i2c/i2c-stub.c if (len > b->len) b 264 drivers/i2c/i2c-stub.c b->len = len; b 266 drivers/i2c/i2c-stub.c b->block[i] = data->block[i + 1]; b 268 drivers/i2c/i2c-stub.c chip->words[command] = (b->block[0] << 8) | b->len; b 273 drivers/i2c/i2c-stub.c if (b == NULL) { b 279 drivers/i2c/i2c-stub.c len = b->len; b 282 drivers/i2c/i2c-stub.c data->block[i + 1] = b->block[i]; b 228 drivers/ide/cmd640.c u8 b; b 231 drivers/ide/cmd640.c b = inb_p(cmd640_key + reg); b 233 drivers/ide/cmd640.c return b; b 253 drivers/ide/cmd640.c u8 b; b 256 drivers/ide/cmd640.c b = __get_cmd640_reg(reg); b 258 drivers/ide/cmd640.c return b; b 322 drivers/ide/cmd640.c u8 b; b 327 drivers/ide/cmd640.c b = get_cmd640_reg(CFR); b 328 drivers/ide/cmd640.c if (b == 0xff || b == 0x00 || (b & CFR_AT_VESA_078h)) { b 330 drivers/ide/cmd640.c b = get_cmd640_reg(CFR); b 331 drivers/ide/cmd640.c if (b == 0xff || b == 0x00 || !(b & CFR_AT_VESA_078h)) b 402 drivers/ide/cmd640.c u8 b = get_cmd640_reg(prefetch_regs[index]); b 404 drivers/ide/cmd640.c __set_prefetch_mode(drive, (b & prefetch_masks[index]) ? 0 : 1); b 415 drivers/ide/cmd640.c u8 b; b 418 drivers/ide/cmd640.c b = __get_cmd640_reg(reg); b 421 drivers/ide/cmd640.c b &= ~prefetch_masks[index]; /* enable prefetch */ b 423 drivers/ide/cmd640.c b |= prefetch_masks[index]; /* disable prefetch */ b 424 drivers/ide/cmd640.c __put_cmd640_reg(reg, b); b 581 drivers/ide/cmd640.c u8 b; b 586 drivers/ide/cmd640.c b = get_cmd640_reg(CNTRL) & ~0x27; b 588 drivers/ide/cmd640.c b |= 0x27; b 589 drivers/ide/cmd640.c put_cmd640_reg(CNTRL, b); b 724 drivers/ide/cmd640.c u8 b, cfr; b 801 drivers/ide/cmd640.c b = get_cmd640_reg(CNTRL); b 807 drivers/ide/cmd640.c if ((b & CNTRL_ENA_2ND)) { b 816 drivers/ide/cmd640.c put_cmd640_reg(CNTRL, b ^ CNTRL_ENA_2ND); /* toggle the bit */ b 821 drivers/ide/cmd640.c put_cmd640_reg(CNTRL, b); /* restore original setting */ b 52 drivers/ide/dtc2278.c static void sub22 (char b, char c) b 58 drivers/ide/dtc2278.c outb_p(b,0xb0); b 99 drivers/ide/ide-floppy_ioctl.c u8 *buf, int b, int l, b 114 drivers/ide/ide-floppy_ioctl.c put_unaligned(cpu_to_be32(b), (unsigned int *)(&buf[4])); b 533 drivers/ide/ide-taskfile.c if (req_task->out_flags.b.data) b 536 drivers/ide/ide-taskfile.c if (req_task->out_flags.b.nsector_hob) b 538 drivers/ide/ide-taskfile.c if (req_task->out_flags.b.sector_hob) b 540 drivers/ide/ide-taskfile.c if (req_task->out_flags.b.lcyl_hob) b 542 drivers/ide/ide-taskfile.c if (req_task->out_flags.b.hcyl_hob) b 545 drivers/ide/ide-taskfile.c if (req_task->out_flags.b.error_feature) b 547 drivers/ide/ide-taskfile.c if (req_task->out_flags.b.nsector) b 549 drivers/ide/ide-taskfile.c if (req_task->out_flags.b.sector) b 551 drivers/ide/ide-taskfile.c if (req_task->out_flags.b.lcyl) b 553 drivers/ide/ide-taskfile.c if (req_task->out_flags.b.hcyl) b 561 drivers/ide/ide-taskfile.c if (req_task->in_flags.b.data) b 109 drivers/ide/ide-timings.c void ide_timing_merge(struct ide_timing *a, struct ide_timing *b, b 113 drivers/ide/ide-timings.c m->setup = max(a->setup, b->setup); b 115 drivers/ide/ide-timings.c m->act8b = max(a->act8b, b->act8b); b 117 drivers/ide/ide-timings.c m->rec8b = max(a->rec8b, b->rec8b); b 119 drivers/ide/ide-timings.c m->cyc8b = max(a->cyc8b, b->cyc8b); b 121 drivers/ide/ide-timings.c m->active = max(a->active, b->active); b 123 drivers/ide/ide-timings.c m->recover = max(a->recover, b->recover); b 125 drivers/ide/ide-timings.c m->cycle = max(a->cycle, b->cycle); b 127 drivers/ide/ide-timings.c m->udma = max(a->udma, b->udma); b 182 drivers/ide/ide.c unsigned int a, b, i, j = 1; b 186 drivers/ide/ide.c if (sscanf(s, "%u.%u:%u", &a, &b, &j) != 3 && b 187 drivers/ide/ide.c sscanf(s, "%u.%u", &a, &b) != 2) b 190 drivers/ide/ide.c i = a * MAX_DRIVES + b; b 250 drivers/ide/ide.c unsigned int a, b, c = 0, h = 0, s = 0, i, j = 1; b 254 drivers/ide/ide.c if (sscanf(str, "%u.%u:%u,%u,%u", &a, &b, &c, &h, &s) != 5 && b 255 drivers/ide/ide.c sscanf(str, "%u.%u:%u", &a, &b, &j) != 3) b 258 drivers/ide/ide.c i = a * MAX_DRIVES + b; b 12 drivers/ide/qd65xx.h #define IDE_IN(a,b,c) ( ((a)<(b)) ? (b) : ( (a)>(c) ? (c) : (a)) ) b 14 drivers/ide/qd65xx.h #define IDE_IMPLY(a,b) ((!(a)) || (b)) b 563 drivers/iio/adc/twl6030-gpadc.c int b, k, gain, x1, x2, i; b 579 drivers/iio/adc/twl6030-gpadc.c b = (d1 * 1000) - (k - 1000) * x1; b 583 drivers/iio/adc/twl6030-gpadc.c gpadc->twl6030_cal_tbl[i].offset_error = b; b 591 drivers/iio/adc/twl6030-gpadc.c dev_dbg(gpadc->dev, "GPADC b for Chn: %d = %d\n", channel, b); b 29 drivers/iio/common/ssp_sensors/ssp_dev.c __le32 b; b 225 drivers/iio/common/ssp_sensors/ssp_dev.c to_send.b = cpu_to_le32(data->batch_latency_buf[type]); b 287 drivers/iio/common/ssp_sensors/ssp_dev.c to_send.b = cpu_to_le32(data->batch_latency_buf[type]); b 241 drivers/iio/dac/ad5758.c static int cmpfunc(const void *a, const void *b) b 243 drivers/iio/dac/ad5758.c return *(int *)a - *(int *)b; b 132 drivers/iio/frequency/ad9523.c #define AD9523_PLL2_FB_NDIV(a, b) (4 * (b) + (a)) b 41 drivers/iio/light/tsl2563.c #define CALIB_FRAC(n, b) (((n) << CALIB_FRAC_BITS) / (b)) b 641 drivers/infiniband/core/cm.c static int be32_lt(__be32 a, __be32 b) b 643 drivers/infiniband/core/cm.c return (__force u32) a < (__force u32) b; b 646 drivers/infiniband/core/cm.c static int be32_gt(__be32 a, __be32 b) b 648 drivers/infiniband/core/cm.c return (__force u32) a > (__force u32) b; b 651 drivers/infiniband/core/cm.c static int be64_lt(__be64 a, __be64 b) b 653 drivers/infiniband/core/cm.c return (__force u64) a < (__force u64) b; b 656 drivers/infiniband/core/cm.c static int be64_gt(__be64 a, __be64 b) b 658 drivers/infiniband/core/cm.c return (__force u64) a > (__force u64) b; b 243 drivers/infiniband/hw/hfi1/eprom.c #define MAGIC4(a, b, c, d) ((d) << 24 | (c) << 16 | (b) << 8 | (a)) b 1424 drivers/infiniband/hw/hfi1/hfi.h #define dc8051_ver(a, b, c) ((a) << 16 | (b) << 8 | (c)) b 1691 drivers/infiniband/hw/hfi1/pio.c #define sent_before(a, b) time_before(a, b) /* a < b */ b 355 drivers/infiniband/hw/hfi1/verbs.h static inline int cmp_psn(u32 a, u32 b) b 357 drivers/infiniband/hw/hfi1/verbs.h return (((int)a) - ((int)b)) << PSN_SHIFT; b 371 drivers/infiniband/hw/hfi1/verbs.h static inline u32 delta_psn(u32 a, u32 b) b 373 drivers/infiniband/hw/hfi1/verbs.h return (((int)a - (int)b) << PSN_SHIFT) >> PSN_SHIFT; b 48 drivers/infiniband/hw/hns/hns_roce_device.h #define HNS_ROCE_ALOGN_UP(a, b) ((((a) + (b) - 1) / (b)) * (b)) b 1509 drivers/infiniband/hw/qib/qib.h #define QLOGIC_IB_HWE_MSG(a, b) { .mask = a, .msg = b } b 258 drivers/infiniband/hw/qib/qib_verbs.h static inline int qib_cmp24(u32 a, u32 b) b 260 drivers/infiniband/hw/qib/qib_verbs.h return (((int) a) - ((int) b)) << 8; b 86 drivers/infiniband/hw/usnic/usnic_uiom_interval_tree.c static int interval_cmp(void *priv, struct list_head *a, struct list_head *b) b 91 drivers/infiniband/hw/usnic/usnic_uiom_interval_tree.c node_b = list_entry(b, struct usnic_uiom_interval_node, link); b 115 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.h #define OPA_VNIC_ENCAP_RC_EXT(w, b) (((w) >> OPA_VNIC_ENCAP_RC_ ## b) & 0x7) b 234 drivers/input/ff-memless.c static inline s16 fixp_mult(s16 a, s16 b) b 237 drivers/input/ff-memless.c return ((s32)(a * b)) >> FRAC_N; b 50 drivers/input/gameport/ns558.c int i, j, b; b 109 drivers/input/gameport/ns558.c for (j = b = 0; j < 1000; j++) b 110 drivers/input/gameport/ns558.c if (inb(io & (-1 << i)) != inb((io & (-1 << i)) + (1 << i) - 1)) b++; b 113 drivers/input/gameport/ns558.c if (b > 300) { /* We allow 30% difference */ b 550 drivers/input/joystick/gamecon.c static void gc_psx_command(struct gc *gc, int b, unsigned char *data) b 557 drivers/input/joystick/gamecon.c for (i = 0; i < GC_PSX_LENGTH; i++, b >>= 1) { b 558 drivers/input/joystick/gamecon.c cmd = (b & 1) ? GC_PSX_COMMAND : 0; b 267 drivers/input/joystick/tmdc.c int i, j, b = 0; b 328 drivers/input/joystick/tmdc.c set_bit(port->btn[j + b], input_dev->keybit); b 329 drivers/input/joystick/tmdc.c b += port->btnc[i]; b 256 drivers/input/keyboard/atkbd.c struct device_attribute *attr, char *b) \ b 258 drivers/input/keyboard/atkbd.c return atkbd_attr_show_helper(d, b, atkbd_show_##_name); \ b 261 drivers/input/keyboard/atkbd.c struct device_attribute *attr, const char *b, size_t s) \ b 263 drivers/input/keyboard/atkbd.c return atkbd_attr_set_helper(d, b, s, atkbd_set_##_name); \ b 278 drivers/input/keyboard/atkbd.c struct device_attribute *attr, char *b) \ b 280 drivers/input/keyboard/atkbd.c return atkbd_attr_show_helper(d, b, atkbd_show_##_name); \ b 48 drivers/input/misc/pcf8574_keypad.c unsigned char x, y, a, b; b 58 drivers/input/misc/pcf8574_keypad.c for (b = 0; y > 0; b++) b 61 drivers/input/misc/pcf8574_keypad.c return ((a - 1) * 4) + b; b 61 drivers/input/misc/yealink.c #define _SEG(t, a, am, b, bm, c, cm, d, dm, e, em, f, fm, g, gm) \ b 63 drivers/input/misc/yealink.c .u = { .s = { _LOC(a, am), _LOC(b, bm), _LOC(c, cm), \ b 111 drivers/input/misc/yealink.c u8 b[sizeof(struct yld_status)]; b 145 drivers/input/misc/yealink.c yld->master.b[a] |= m; b 147 drivers/input/misc/yealink.c yld->master.b[a] &= ~m; b 160 drivers/input/misc/yealink.c yld->master.b[a] |= m; b 162 drivers/input/misc/yealink.c yld->master.b[a] &= ~m; b 334 drivers/input/misc/yealink.c val = yld->master.b[ix]; b 335 drivers/input/misc/yealink.c if (val != yld->copy.b[ix]) b 349 drivers/input/misc/yealink.c yld->copy.b[ix] = val; b 386 drivers/input/misc/yealink.c val = yld->master.b[ix]; b 387 drivers/input/misc/yealink.c yld->copy.b[ix] = val; b 510 drivers/input/misc/yealink.c yld->copy.b[i] = ~yld->master.b[i]; b 581 drivers/input/misc/yealink.c static ssize_t show_line(struct device *dev, char *buf, int a, int b) b 593 drivers/input/misc/yealink.c for (i = a; i < b; i++) b 596 drivers/input/misc/yealink.c for (i = a; i < b; i++) b 602 drivers/input/misc/yealink.c return 3 + ((b - a) << 1); b 116 drivers/input/mouse/inport.c unsigned char a, b, c; b 125 drivers/input/mouse/inport.c b = inb(INPORT_SIGNATURE_PORT); b 127 drivers/input/mouse/inport.c if (a == b || a != c) { b 38 drivers/input/mouse/rpcmouse.c short x, y, dx, dy, b; b 42 drivers/input/mouse/rpcmouse.c b = (short) (__raw_readl(IOMEM(0xe0310000)) ^ 0x70); b 53 drivers/input/mouse/rpcmouse.c input_report_key(dev, BTN_LEFT, b & 0x40); b 54 drivers/input/mouse/rpcmouse.c input_report_key(dev, BTN_MIDDLE, b & 0x20); b 55 drivers/input/mouse/rpcmouse.c input_report_key(dev, BTN_RIGHT, b & 0x10); b 900 drivers/input/mouse/synaptics.c const struct synaptics_hw_state *b, b 904 drivers/input/mouse/synaptics.c synaptics_report_semi_mt_slot(dev, 0, true, min(a->x, b->x), b 905 drivers/input/mouse/synaptics.c min(a->y, b->y)); b 906 drivers/input/mouse/synaptics.c synaptics_report_semi_mt_slot(dev, 1, true, max(a->x, b->x), b 907 drivers/input/mouse/synaptics.c max(a->y, b->y)); b 572 drivers/input/rmi4/rmi_driver.c int b; b 606 drivers/input/rmi4/rmi_driver.c for (b = 0; b < 8; b++) { b 607 drivers/input/rmi4/rmi_driver.c if (buf[i] & (0x1 << b)) b 668 drivers/input/rmi4/rmi_driver.c for (b = 0; b < 7; b++) { b 669 drivers/input/rmi4/rmi_driver.c if (struct_buf[offset] & (0x1 << b)) b 212 drivers/input/touchscreen/ti_am335x_tsc.c static int titsc_cmp_coord(const void *a, const void *b) b 214 drivers/input/touchscreen/ti_am335x_tsc.c return *(int *)a - *(int *)b; b 610 drivers/input/touchscreen/wdt87xx_i2c.c u32 a, b; b 615 drivers/input/touchscreen/wdt87xx_i2c.c b = new_value; b 616 drivers/input/touchscreen/wdt87xx_i2c.c bit0 = a ^ (b & 1); b 624 drivers/input/touchscreen/wdt87xx_i2c.c y = (a << 1) ^ b; b 20 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_GLOBAL_REG_N(b, n, r, v) SET_GLOBAL_REG(b, ((r) + (n << 2)), (v)) b 21 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_GLOBAL_REG_N(b, n, r) GET_GLOBAL_REG(b, ((r) + (n << 2))) b 24 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_GLOBAL_FIELD(b, r, F) GET_FIELD(((b) + (r)), F##_MASK, F##_SHIFT) b 25 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_CONTEXT_FIELD(b, c, r, F) \ b 26 drivers/iommu/msm_iommu_hw-8xxx.h GET_FIELD(((b) + (r) + ((c) << CTX_SHIFT)), F##_MASK, F##_SHIFT) b 28 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_GLOBAL_FIELD(b, r, F, v) \ b 29 drivers/iommu/msm_iommu_hw-8xxx.h SET_FIELD(((b) + (r)), F##_MASK, F##_SHIFT, (v)) b 30 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_CONTEXT_FIELD(b, c, r, F, v) \ b 31 drivers/iommu/msm_iommu_hw-8xxx.h SET_FIELD(((b) + (r) + ((c) << CTX_SHIFT)), F##_MASK, F##_SHIFT, (v)) b 84 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_M2VCBR_N(b, N, v) SET_GLOBAL_REG_N(M2VCBR_N, N, (b), (v)) b 85 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_CBACR_N(b, N, v) SET_GLOBAL_REG_N(CBACR_N, N, (b), (v)) b 86 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_TLBRSW(b, v) SET_GLOBAL_REG(TLBRSW, (b), (v)) b 87 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_TLBTR0(b, v) SET_GLOBAL_REG(TLBTR0, (b), (v)) b 88 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_TLBTR1(b, v) SET_GLOBAL_REG(TLBTR1, (b), (v)) b 89 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_TLBTR2(b, v) SET_GLOBAL_REG(TLBTR2, (b), (v)) b 90 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_TESTBUSCR(b, v) SET_GLOBAL_REG(TESTBUSCR, (b), (v)) b 91 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_GLOBAL_TLBIALL(b, v) SET_GLOBAL_REG(GLOBAL_TLBIALL, (b), (v)) b 92 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_TLBIVMID(b, v) SET_GLOBAL_REG(TLBIVMID, (b), (v)) b 93 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_CR(b, v) SET_GLOBAL_REG(CR, (b), (v)) b 94 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_EAR(b, v) SET_GLOBAL_REG(EAR, (b), (v)) b 95 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_ESR(b, v) SET_GLOBAL_REG(ESR, (b), (v)) b 96 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_ESRRESTORE(b, v) SET_GLOBAL_REG(ESRRESTORE, (b), (v)) b 97 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_ESYNR0(b, v) SET_GLOBAL_REG(ESYNR0, (b), (v)) b 98 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_ESYNR1(b, v) SET_GLOBAL_REG(ESYNR1, (b), (v)) b 99 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_RPU_ACR(b, v) SET_GLOBAL_REG(RPU_ACR, (b), (v)) b 101 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_M2VCBR_N(b, N) GET_GLOBAL_REG_N(M2VCBR_N, N, (b)) b 102 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_CBACR_N(b, N) GET_GLOBAL_REG_N(CBACR_N, N, (b)) b 103 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_TLBTR0(b) GET_GLOBAL_REG(TLBTR0, (b)) b 104 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_TLBTR1(b) GET_GLOBAL_REG(TLBTR1, (b)) b 105 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_TLBTR2(b) GET_GLOBAL_REG(TLBTR2, (b)) b 106 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_TESTBUSCR(b) GET_GLOBAL_REG(TESTBUSCR, (b)) b 107 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_GLOBAL_TLBIALL(b) GET_GLOBAL_REG(GLOBAL_TLBIALL, (b)) b 108 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_TLBIVMID(b) GET_GLOBAL_REG(TLBIVMID, (b)) b 109 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_CR(b) GET_GLOBAL_REG(CR, (b)) b 110 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_EAR(b) GET_GLOBAL_REG(EAR, (b)) b 111 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_ESR(b) GET_GLOBAL_REG(ESR, (b)) b 112 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_ESRRESTORE(b) GET_GLOBAL_REG(ESRRESTORE, (b)) b 113 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_ESYNR0(b) GET_GLOBAL_REG(ESYNR0, (b)) b 114 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_ESYNR1(b) GET_GLOBAL_REG(ESYNR1, (b)) b 115 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_REV(b) GET_GLOBAL_REG(REV, (b)) b 116 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_IDR(b) GET_GLOBAL_REG(IDR, (b)) b 117 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_RPU_ACR(b) GET_GLOBAL_REG(RPU_ACR, (b)) b 121 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_SCTLR(b, c, v) SET_CTX_REG(SCTLR, (b), (c), (v)) b 122 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_ACTLR(b, c, v) SET_CTX_REG(ACTLR, (b), (c), (v)) b 123 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_CONTEXTIDR(b, c, v) SET_CTX_REG(CONTEXTIDR, (b), (c), (v)) b 124 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_TTBR0(b, c, v) SET_CTX_REG(TTBR0, (b), (c), (v)) b 125 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_TTBR1(b, c, v) SET_CTX_REG(TTBR1, (b), (c), (v)) b 126 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_TTBCR(b, c, v) SET_CTX_REG(TTBCR, (b), (c), (v)) b 127 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_PAR(b, c, v) SET_CTX_REG(PAR, (b), (c), (v)) b 128 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_FSR(b, c, v) SET_CTX_REG(FSR, (b), (c), (v)) b 129 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_FSRRESTORE(b, c, v) SET_CTX_REG(FSRRESTORE, (b), (c), (v)) b 130 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_FAR(b, c, v) SET_CTX_REG(FAR, (b), (c), (v)) b 131 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_FSYNR0(b, c, v) SET_CTX_REG(FSYNR0, (b), (c), (v)) b 132 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_FSYNR1(b, c, v) SET_CTX_REG(FSYNR1, (b), (c), (v)) b 133 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_PRRR(b, c, v) SET_CTX_REG(PRRR, (b), (c), (v)) b 134 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_NMRR(b, c, v) SET_CTX_REG(NMRR, (b), (c), (v)) b 135 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_TLBLKCR(b, c, v) SET_CTX_REG(TLBLCKR, (b), (c), (v)) b 136 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_V2PSR(b, c, v) SET_CTX_REG(V2PSR, (b), (c), (v)) b 137 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_TLBFLPTER(b, c, v) SET_CTX_REG(TLBFLPTER, (b), (c), (v)) b 138 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_TLBSLPTER(b, c, v) SET_CTX_REG(TLBSLPTER, (b), (c), (v)) b 139 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_BFBCR(b, c, v) SET_CTX_REG(BFBCR, (b), (c), (v)) b 140 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_CTX_TLBIALL(b, c, v) SET_CTX_REG(CTX_TLBIALL, (b), (c), (v)) b 141 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_TLBIASID(b, c, v) SET_CTX_REG(TLBIASID, (b), (c), (v)) b 142 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_TLBIVA(b, c, v) SET_CTX_REG(TLBIVA, (b), (c), (v)) b 143 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_TLBIVAA(b, c, v) SET_CTX_REG(TLBIVAA, (b), (c), (v)) b 144 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_V2PPR(b, c, v) SET_CTX_REG(V2PPR, (b), (c), (v)) b 145 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_V2PPW(b, c, v) SET_CTX_REG(V2PPW, (b), (c), (v)) b 146 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_V2PUR(b, c, v) SET_CTX_REG(V2PUR, (b), (c), (v)) b 147 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_V2PUW(b, c, v) SET_CTX_REG(V2PUW, (b), (c), (v)) b 148 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_RESUME(b, c, v) SET_CTX_REG(RESUME, (b), (c), (v)) b 150 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_SCTLR(b, c) GET_CTX_REG(SCTLR, (b), (c)) b 151 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_ACTLR(b, c) GET_CTX_REG(ACTLR, (b), (c)) b 152 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_CONTEXTIDR(b, c) GET_CTX_REG(CONTEXTIDR, (b), (c)) b 153 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_TTBR0(b, c) GET_CTX_REG(TTBR0, (b), (c)) b 154 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_TTBR1(b, c) GET_CTX_REG(TTBR1, (b), (c)) b 155 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_TTBCR(b, c) GET_CTX_REG(TTBCR, (b), (c)) b 156 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_PAR(b, c) GET_CTX_REG(PAR, (b), (c)) b 157 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_FSR(b, c) GET_CTX_REG(FSR, (b), (c)) b 158 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_FSRRESTORE(b, c) GET_CTX_REG(FSRRESTORE, (b), (c)) b 159 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_FAR(b, c) GET_CTX_REG(FAR, (b), (c)) b 160 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_FSYNR0(b, c) GET_CTX_REG(FSYNR0, (b), (c)) b 161 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_FSYNR1(b, c) GET_CTX_REG(FSYNR1, (b), (c)) b 162 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_PRRR(b, c) GET_CTX_REG(PRRR, (b), (c)) b 163 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_NMRR(b, c) GET_CTX_REG(NMRR, (b), (c)) b 164 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_TLBLCKR(b, c) GET_CTX_REG(TLBLCKR, (b), (c)) b 165 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_V2PSR(b, c) GET_CTX_REG(V2PSR, (b), (c)) b 166 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_TLBFLPTER(b, c) GET_CTX_REG(TLBFLPTER, (b), (c)) b 167 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_TLBSLPTER(b, c) GET_CTX_REG(TLBSLPTER, (b), (c)) b 168 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_BFBCR(b, c) GET_CTX_REG(BFBCR, (b), (c)) b 169 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_CTX_TLBIALL(b, c) GET_CTX_REG(CTX_TLBIALL, (b), (c)) b 170 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_TLBIASID(b, c) GET_CTX_REG(TLBIASID, (b), (c)) b 171 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_TLBIVA(b, c) GET_CTX_REG(TLBIVA, (b), (c)) b 172 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_TLBIVAA(b, c) GET_CTX_REG(TLBIVAA, (b), (c)) b 173 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_V2PPR(b, c) GET_CTX_REG(V2PPR, (b), (c)) b 174 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_V2PPW(b, c) GET_CTX_REG(V2PPW, (b), (c)) b 175 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_V2PUR(b, c) GET_CTX_REG(V2PUR, (b), (c)) b 176 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_V2PUW(b, c) GET_CTX_REG(V2PUW, (b), (c)) b 177 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_RESUME(b, c) GET_CTX_REG(RESUME, (b), (c)) b 183 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_RWVMID(b, n, v) SET_GLOBAL_FIELD(b, (n<<2)|(CBACR_N), RWVMID, v) b 184 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_RWE(b, n, v) SET_GLOBAL_FIELD(b, (n<<2)|(CBACR_N), RWE, v) b 185 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_RWGE(b, n, v) SET_GLOBAL_FIELD(b, (n<<2)|(CBACR_N), RWGE, v) b 186 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_CBVMID(b, n, v) SET_GLOBAL_FIELD(b, (n<<2)|(CBACR_N), CBVMID, v) b 187 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_IRPTNDX(b, n, v) SET_GLOBAL_FIELD(b, (n<<2)|(CBACR_N), IRPTNDX, v) b 191 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_VMID(b, n, v) SET_GLOBAL_FIELD(b, (n<<2)|(M2VCBR_N), VMID, v) b 192 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_CBNDX(b, n, v) SET_GLOBAL_FIELD(b, (n<<2)|(M2VCBR_N), CBNDX, v) b 193 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_BYPASSD(b, n, v) SET_GLOBAL_FIELD(b, (n<<2)|(M2VCBR_N), BYPASSD, v) b 194 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_BPRCOSH(b, n, v) SET_GLOBAL_FIELD(b, (n<<2)|(M2VCBR_N), BPRCOSH, v) b 195 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_BPRCISH(b, n, v) SET_GLOBAL_FIELD(b, (n<<2)|(M2VCBR_N), BPRCISH, v) b 196 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_BPRCNSH(b, n, v) SET_GLOBAL_FIELD(b, (n<<2)|(M2VCBR_N), BPRCNSH, v) b 197 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_BPSHCFG(b, n, v) SET_GLOBAL_FIELD(b, (n<<2)|(M2VCBR_N), BPSHCFG, v) b 198 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_NSCFG(b, n, v) SET_GLOBAL_FIELD(b, (n<<2)|(M2VCBR_N), NSCFG, v) b 199 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_BPMTCFG(b, n, v) SET_GLOBAL_FIELD(b, (n<<2)|(M2VCBR_N), BPMTCFG, v) b 200 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_BPMEMTYPE(b, n, v) \ b 201 drivers/iommu/msm_iommu_hw-8xxx.h SET_GLOBAL_FIELD(b, (n<<2)|(M2VCBR_N), BPMEMTYPE, v) b 205 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_RPUE(b, v) SET_GLOBAL_FIELD(b, CR, RPUE, v) b 206 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_RPUERE(b, v) SET_GLOBAL_FIELD(b, CR, RPUERE, v) b 207 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_RPUEIE(b, v) SET_GLOBAL_FIELD(b, CR, RPUEIE, v) b 208 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_DCDEE(b, v) SET_GLOBAL_FIELD(b, CR, DCDEE, v) b 209 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_CLIENTPD(b, v) SET_GLOBAL_FIELD(b, CR, CLIENTPD, v) b 210 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_STALLD(b, v) SET_GLOBAL_FIELD(b, CR, STALLD, v) b 211 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_TLBLKCRWE(b, v) SET_GLOBAL_FIELD(b, CR, TLBLKCRWE, v) b 212 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_CR_TLBIALLCFG(b, v) SET_GLOBAL_FIELD(b, CR, CR_TLBIALLCFG, v) b 213 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_TLBIVMIDCFG(b, v) SET_GLOBAL_FIELD(b, CR, TLBIVMIDCFG, v) b 214 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_CR_HUME(b, v) SET_GLOBAL_FIELD(b, CR, CR_HUME, v) b 218 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_CFG(b, v) SET_GLOBAL_FIELD(b, ESR, CFG, v) b 219 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_BYPASS(b, v) SET_GLOBAL_FIELD(b, ESR, BYPASS, v) b 220 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_ESR_MULTI(b, v) SET_GLOBAL_FIELD(b, ESR, ESR_MULTI, v) b 224 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_ESYNR0_AMID(b, v) SET_GLOBAL_FIELD(b, ESYNR0, ESYNR0_AMID, v) b 225 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_ESYNR0_APID(b, v) SET_GLOBAL_FIELD(b, ESYNR0, ESYNR0_APID, v) b 226 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_ESYNR0_ABID(b, v) SET_GLOBAL_FIELD(b, ESYNR0, ESYNR0_ABID, v) b 227 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_ESYNR0_AVMID(b, v) SET_GLOBAL_FIELD(b, ESYNR0, ESYNR0_AVMID, v) b 228 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_ESYNR0_ATID(b, v) SET_GLOBAL_FIELD(b, ESYNR0, ESYNR0_ATID, v) b 232 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_ESYNR1_AMEMTYPE(b, v) \ b 233 drivers/iommu/msm_iommu_hw-8xxx.h SET_GLOBAL_FIELD(b, ESYNR1, ESYNR1_AMEMTYPE, v) b 234 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_ESYNR1_ASHARED(b, v) SET_GLOBAL_FIELD(b, ESYNR1, ESYNR1_ASHARED, v) b 235 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_ESYNR1_AINNERSHARED(b, v) \ b 236 drivers/iommu/msm_iommu_hw-8xxx.h SET_GLOBAL_FIELD(b, ESYNR1, ESYNR1_AINNERSHARED, v) b 237 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_ESYNR1_APRIV(b, v) SET_GLOBAL_FIELD(b, ESYNR1, ESYNR1_APRIV, v) b 238 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_ESYNR1_APROTNS(b, v) SET_GLOBAL_FIELD(b, ESYNR1, ESYNR1_APROTNS, v) b 239 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_ESYNR1_AINST(b, v) SET_GLOBAL_FIELD(b, ESYNR1, ESYNR1_AINST, v) b 240 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_ESYNR1_AWRITE(b, v) SET_GLOBAL_FIELD(b, ESYNR1, ESYNR1_AWRITE, v) b 241 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_ESYNR1_ABURST(b, v) SET_GLOBAL_FIELD(b, ESYNR1, ESYNR1_ABURST, v) b 242 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_ESYNR1_ALEN(b, v) SET_GLOBAL_FIELD(b, ESYNR1, ESYNR1_ALEN, v) b 243 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_ESYNR1_ASIZE(b, v) SET_GLOBAL_FIELD(b, ESYNR1, ESYNR1_ASIZE, v) b 244 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_ESYNR1_ALOCK(b, v) SET_GLOBAL_FIELD(b, ESYNR1, ESYNR1_ALOCK, v) b 245 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_ESYNR1_AOOO(b, v) SET_GLOBAL_FIELD(b, ESYNR1, ESYNR1_AOOO, v) b 246 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_ESYNR1_AFULL(b, v) SET_GLOBAL_FIELD(b, ESYNR1, ESYNR1_AFULL, v) b 247 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_ESYNR1_AC(b, v) SET_GLOBAL_FIELD(b, ESYNR1, ESYNR1_AC, v) b 248 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_ESYNR1_DCD(b, v) SET_GLOBAL_FIELD(b, ESYNR1, ESYNR1_DCD, v) b 252 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_TBE(b, v) SET_GLOBAL_FIELD(b, TESTBUSCR, TBE, v) b 253 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_SPDMBE(b, v) SET_GLOBAL_FIELD(b, TESTBUSCR, SPDMBE, v) b 254 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_WGSEL(b, v) SET_GLOBAL_FIELD(b, TESTBUSCR, WGSEL, v) b 255 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_TBLSEL(b, v) SET_GLOBAL_FIELD(b, TESTBUSCR, TBLSEL, v) b 256 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_TBHSEL(b, v) SET_GLOBAL_FIELD(b, TESTBUSCR, TBHSEL, v) b 257 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_SPDM0SEL(b, v) SET_GLOBAL_FIELD(b, TESTBUSCR, SPDM0SEL, v) b 258 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_SPDM1SEL(b, v) SET_GLOBAL_FIELD(b, TESTBUSCR, SPDM1SEL, v) b 259 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_SPDM2SEL(b, v) SET_GLOBAL_FIELD(b, TESTBUSCR, SPDM2SEL, v) b 260 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_SPDM3SEL(b, v) SET_GLOBAL_FIELD(b, TESTBUSCR, SPDM3SEL, v) b 264 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_TLBIVMID_VMID(b, v) SET_GLOBAL_FIELD(b, TLBIVMID, TLBIVMID_VMID, v) b 268 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_TLBRSW_INDEX(b, v) SET_GLOBAL_FIELD(b, TLBRSW, TLBRSW_INDEX, v) b 269 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_TLBBFBS(b, v) SET_GLOBAL_FIELD(b, TLBRSW, TLBBFBS, v) b 273 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_PR(b, v) SET_GLOBAL_FIELD(b, TLBTR0, PR, v) b 274 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_PW(b, v) SET_GLOBAL_FIELD(b, TLBTR0, PW, v) b 275 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_UR(b, v) SET_GLOBAL_FIELD(b, TLBTR0, UR, v) b 276 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_UW(b, v) SET_GLOBAL_FIELD(b, TLBTR0, UW, v) b 277 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_XN(b, v) SET_GLOBAL_FIELD(b, TLBTR0, XN, v) b 278 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_NSDESC(b, v) SET_GLOBAL_FIELD(b, TLBTR0, NSDESC, v) b 279 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_ISH(b, v) SET_GLOBAL_FIELD(b, TLBTR0, ISH, v) b 280 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_SH(b, v) SET_GLOBAL_FIELD(b, TLBTR0, SH, v) b 281 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_MT(b, v) SET_GLOBAL_FIELD(b, TLBTR0, MT, v) b 282 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_DPSIZR(b, v) SET_GLOBAL_FIELD(b, TLBTR0, DPSIZR, v) b 283 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_DPSIZC(b, v) SET_GLOBAL_FIELD(b, TLBTR0, DPSIZC, v) b 287 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_TLBTR1_VMID(b, v) SET_GLOBAL_FIELD(b, TLBTR1, TLBTR1_VMID, v) b 288 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_TLBTR1_PA(b, v) SET_GLOBAL_FIELD(b, TLBTR1, TLBTR1_PA, v) b 292 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_TLBTR2_ASID(b, v) SET_GLOBAL_FIELD(b, TLBTR2, TLBTR2_ASID, v) b 293 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_TLBTR2_V(b, v) SET_GLOBAL_FIELD(b, TLBTR2, TLBTR2_V, v) b 294 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_TLBTR2_NSTID(b, v) SET_GLOBAL_FIELD(b, TLBTR2, TLBTR2_NSTID, v) b 295 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_TLBTR2_NV(b, v) SET_GLOBAL_FIELD(b, TLBTR2, TLBTR2_NV, v) b 296 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_TLBTR2_VA(b, v) SET_GLOBAL_FIELD(b, TLBTR2, TLBTR2_VA, v) b 301 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_RWVMID(b, n) GET_GLOBAL_FIELD(b, (n<<2)|(CBACR_N), RWVMID) b 302 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_RWE(b, n) GET_GLOBAL_FIELD(b, (n<<2)|(CBACR_N), RWE) b 303 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_RWGE(b, n) GET_GLOBAL_FIELD(b, (n<<2)|(CBACR_N), RWGE) b 304 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_CBVMID(b, n) GET_GLOBAL_FIELD(b, (n<<2)|(CBACR_N), CBVMID) b 305 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_IRPTNDX(b, n) GET_GLOBAL_FIELD(b, (n<<2)|(CBACR_N), IRPTNDX) b 309 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_VMID(b, n) GET_GLOBAL_FIELD(b, (n<<2)|(M2VCBR_N), VMID) b 310 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_CBNDX(b, n) GET_GLOBAL_FIELD(b, (n<<2)|(M2VCBR_N), CBNDX) b 311 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_BYPASSD(b, n) GET_GLOBAL_FIELD(b, (n<<2)|(M2VCBR_N), BYPASSD) b 312 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_BPRCOSH(b, n) GET_GLOBAL_FIELD(b, (n<<2)|(M2VCBR_N), BPRCOSH) b 313 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_BPRCISH(b, n) GET_GLOBAL_FIELD(b, (n<<2)|(M2VCBR_N), BPRCISH) b 314 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_BPRCNSH(b, n) GET_GLOBAL_FIELD(b, (n<<2)|(M2VCBR_N), BPRCNSH) b 315 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_BPSHCFG(b, n) GET_GLOBAL_FIELD(b, (n<<2)|(M2VCBR_N), BPSHCFG) b 316 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_NSCFG(b, n) GET_GLOBAL_FIELD(b, (n<<2)|(M2VCBR_N), NSCFG) b 317 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_BPMTCFG(b, n) GET_GLOBAL_FIELD(b, (n<<2)|(M2VCBR_N), BPMTCFG) b 318 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_BPMEMTYPE(b, n) GET_GLOBAL_FIELD(b, (n<<2)|(M2VCBR_N), BPMEMTYPE) b 322 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_RPUE(b) GET_GLOBAL_FIELD(b, CR, RPUE) b 323 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_RPUERE(b) GET_GLOBAL_FIELD(b, CR, RPUERE) b 324 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_RPUEIE(b) GET_GLOBAL_FIELD(b, CR, RPUEIE) b 325 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_DCDEE(b) GET_GLOBAL_FIELD(b, CR, DCDEE) b 326 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_CLIENTPD(b) GET_GLOBAL_FIELD(b, CR, CLIENTPD) b 327 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_STALLD(b) GET_GLOBAL_FIELD(b, CR, STALLD) b 328 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_TLBLKCRWE(b) GET_GLOBAL_FIELD(b, CR, TLBLKCRWE) b 329 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_CR_TLBIALLCFG(b) GET_GLOBAL_FIELD(b, CR, CR_TLBIALLCFG) b 330 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_TLBIVMIDCFG(b) GET_GLOBAL_FIELD(b, CR, TLBIVMIDCFG) b 331 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_CR_HUME(b) GET_GLOBAL_FIELD(b, CR, CR_HUME) b 335 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_CFG(b) GET_GLOBAL_FIELD(b, ESR, CFG) b 336 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_BYPASS(b) GET_GLOBAL_FIELD(b, ESR, BYPASS) b 337 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_ESR_MULTI(b) GET_GLOBAL_FIELD(b, ESR, ESR_MULTI) b 341 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_ESYNR0_AMID(b) GET_GLOBAL_FIELD(b, ESYNR0, ESYNR0_AMID) b 342 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_ESYNR0_APID(b) GET_GLOBAL_FIELD(b, ESYNR0, ESYNR0_APID) b 343 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_ESYNR0_ABID(b) GET_GLOBAL_FIELD(b, ESYNR0, ESYNR0_ABID) b 344 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_ESYNR0_AVMID(b) GET_GLOBAL_FIELD(b, ESYNR0, ESYNR0_AVMID) b 345 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_ESYNR0_ATID(b) GET_GLOBAL_FIELD(b, ESYNR0, ESYNR0_ATID) b 349 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_ESYNR1_AMEMTYPE(b) GET_GLOBAL_FIELD(b, ESYNR1, ESYNR1_AMEMTYPE) b 350 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_ESYNR1_ASHARED(b) GET_GLOBAL_FIELD(b, ESYNR1, ESYNR1_ASHARED) b 351 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_ESYNR1_AINNERSHARED(b) \ b 352 drivers/iommu/msm_iommu_hw-8xxx.h GET_GLOBAL_FIELD(b, ESYNR1, ESYNR1_AINNERSHARED) b 353 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_ESYNR1_APRIV(b) GET_GLOBAL_FIELD(b, ESYNR1, ESYNR1_APRIV) b 354 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_ESYNR1_APROTNS(b) GET_GLOBAL_FIELD(b, ESYNR1, ESYNR1_APROTNS) b 355 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_ESYNR1_AINST(b) GET_GLOBAL_FIELD(b, ESYNR1, ESYNR1_AINST) b 356 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_ESYNR1_AWRITE(b) GET_GLOBAL_FIELD(b, ESYNR1, ESYNR1_AWRITE) b 357 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_ESYNR1_ABURST(b) GET_GLOBAL_FIELD(b, ESYNR1, ESYNR1_ABURST) b 358 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_ESYNR1_ALEN(b) GET_GLOBAL_FIELD(b, ESYNR1, ESYNR1_ALEN) b 359 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_ESYNR1_ASIZE(b) GET_GLOBAL_FIELD(b, ESYNR1, ESYNR1_ASIZE) b 360 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_ESYNR1_ALOCK(b) GET_GLOBAL_FIELD(b, ESYNR1, ESYNR1_ALOCK) b 361 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_ESYNR1_AOOO(b) GET_GLOBAL_FIELD(b, ESYNR1, ESYNR1_AOOO) b 362 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_ESYNR1_AFULL(b) GET_GLOBAL_FIELD(b, ESYNR1, ESYNR1_AFULL) b 363 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_ESYNR1_AC(b) GET_GLOBAL_FIELD(b, ESYNR1, ESYNR1_AC) b 364 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_ESYNR1_DCD(b) GET_GLOBAL_FIELD(b, ESYNR1, ESYNR1_DCD) b 368 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_NM2VCBMT(b) GET_GLOBAL_FIELD(b, IDR, NM2VCBMT) b 369 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_HTW(b) GET_GLOBAL_FIELD(b, IDR, HTW) b 370 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_HUM(b) GET_GLOBAL_FIELD(b, IDR, HUM) b 371 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_TLBSIZE(b) GET_GLOBAL_FIELD(b, IDR, TLBSIZE) b 372 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_NCB(b) GET_GLOBAL_FIELD(b, IDR, NCB) b 373 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_NIRPT(b) GET_GLOBAL_FIELD(b, IDR, NIRPT) b 377 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_MAJOR(b) GET_GLOBAL_FIELD(b, REV, MAJOR) b 378 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_MINOR(b) GET_GLOBAL_FIELD(b, REV, MINOR) b 382 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_TBE(b) GET_GLOBAL_FIELD(b, TESTBUSCR, TBE) b 383 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_SPDMBE(b) GET_GLOBAL_FIELD(b, TESTBUSCR, SPDMBE) b 384 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_WGSEL(b) GET_GLOBAL_FIELD(b, TESTBUSCR, WGSEL) b 385 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_TBLSEL(b) GET_GLOBAL_FIELD(b, TESTBUSCR, TBLSEL) b 386 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_TBHSEL(b) GET_GLOBAL_FIELD(b, TESTBUSCR, TBHSEL) b 387 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_SPDM0SEL(b) GET_GLOBAL_FIELD(b, TESTBUSCR, SPDM0SEL) b 388 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_SPDM1SEL(b) GET_GLOBAL_FIELD(b, TESTBUSCR, SPDM1SEL) b 389 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_SPDM2SEL(b) GET_GLOBAL_FIELD(b, TESTBUSCR, SPDM2SEL) b 390 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_SPDM3SEL(b) GET_GLOBAL_FIELD(b, TESTBUSCR, SPDM3SEL) b 394 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_TLBIVMID_VMID(b) GET_GLOBAL_FIELD(b, TLBIVMID, TLBIVMID_VMID) b 398 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_PR(b) GET_GLOBAL_FIELD(b, TLBTR0, PR) b 399 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_PW(b) GET_GLOBAL_FIELD(b, TLBTR0, PW) b 400 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_UR(b) GET_GLOBAL_FIELD(b, TLBTR0, UR) b 401 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_UW(b) GET_GLOBAL_FIELD(b, TLBTR0, UW) b 402 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_XN(b) GET_GLOBAL_FIELD(b, TLBTR0, XN) b 403 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_NSDESC(b) GET_GLOBAL_FIELD(b, TLBTR0, NSDESC) b 404 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_ISH(b) GET_GLOBAL_FIELD(b, TLBTR0, ISH) b 405 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_SH(b) GET_GLOBAL_FIELD(b, TLBTR0, SH) b 406 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_MT(b) GET_GLOBAL_FIELD(b, TLBTR0, MT) b 407 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_DPSIZR(b) GET_GLOBAL_FIELD(b, TLBTR0, DPSIZR) b 408 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_DPSIZC(b) GET_GLOBAL_FIELD(b, TLBTR0, DPSIZC) b 412 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_TLBTR1_VMID(b) GET_GLOBAL_FIELD(b, TLBTR1, TLBTR1_VMID) b 413 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_TLBTR1_PA(b) GET_GLOBAL_FIELD(b, TLBTR1, TLBTR1_PA) b 417 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_TLBTR2_ASID(b) GET_GLOBAL_FIELD(b, TLBTR2, TLBTR2_ASID) b 418 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_TLBTR2_V(b) GET_GLOBAL_FIELD(b, TLBTR2, TLBTR2_V) b 419 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_TLBTR2_NSTID(b) GET_GLOBAL_FIELD(b, TLBTR2, TLBTR2_NSTID) b 420 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_TLBTR2_NV(b) GET_GLOBAL_FIELD(b, TLBTR2, TLBTR2_NV) b 421 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_TLBTR2_VA(b) GET_GLOBAL_FIELD(b, TLBTR2, TLBTR2_VA) b 427 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_CFERE(b, c, v) SET_CONTEXT_FIELD(b, c, ACTLR, CFERE, v) b 428 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_CFEIE(b, c, v) SET_CONTEXT_FIELD(b, c, ACTLR, CFEIE, v) b 429 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_PTSHCFG(b, c, v) SET_CONTEXT_FIELD(b, c, ACTLR, PTSHCFG, v) b 430 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_RCOSH(b, c, v) SET_CONTEXT_FIELD(b, c, ACTLR, RCOSH, v) b 431 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_RCISH(b, c, v) SET_CONTEXT_FIELD(b, c, ACTLR, RCISH, v) b 432 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_RCNSH(b, c, v) SET_CONTEXT_FIELD(b, c, ACTLR, RCNSH, v) b 433 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_PRIVCFG(b, c, v) SET_CONTEXT_FIELD(b, c, ACTLR, PRIVCFG, v) b 434 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_DNA(b, c, v) SET_CONTEXT_FIELD(b, c, ACTLR, DNA, v) b 435 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_DNLV2PA(b, c, v) SET_CONTEXT_FIELD(b, c, ACTLR, DNLV2PA, v) b 436 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_TLBMCFG(b, c, v) SET_CONTEXT_FIELD(b, c, ACTLR, TLBMCFG, v) b 437 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_CFCFG(b, c, v) SET_CONTEXT_FIELD(b, c, ACTLR, CFCFG, v) b 438 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_TIPCF(b, c, v) SET_CONTEXT_FIELD(b, c, ACTLR, TIPCF, v) b 439 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_V2PCFG(b, c, v) SET_CONTEXT_FIELD(b, c, ACTLR, V2PCFG, v) b 440 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_HUME(b, c, v) SET_CONTEXT_FIELD(b, c, ACTLR, HUME, v) b 441 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_PTMTCFG(b, c, v) SET_CONTEXT_FIELD(b, c, ACTLR, PTMTCFG, v) b 442 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_PTMEMTYPE(b, c, v) SET_CONTEXT_FIELD(b, c, ACTLR, PTMEMTYPE, v) b 446 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_BFBDFE(b, c, v) SET_CONTEXT_FIELD(b, c, BFBCR, BFBDFE, v) b 447 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_BFBSFE(b, c, v) SET_CONTEXT_FIELD(b, c, BFBCR, BFBSFE, v) b 448 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_SFVS(b, c, v) SET_CONTEXT_FIELD(b, c, BFBCR, SFVS, v) b 449 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_FLVIC(b, c, v) SET_CONTEXT_FIELD(b, c, BFBCR, FLVIC, v) b 450 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_SLVIC(b, c, v) SET_CONTEXT_FIELD(b, c, BFBCR, SLVIC, v) b 454 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_CONTEXTIDR_ASID(b, c, v) \ b 455 drivers/iommu/msm_iommu_hw-8xxx.h SET_CONTEXT_FIELD(b, c, CONTEXTIDR, CONTEXTIDR_ASID, v) b 456 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_CONTEXTIDR_PROCID(b, c, v) \ b 457 drivers/iommu/msm_iommu_hw-8xxx.h SET_CONTEXT_FIELD(b, c, CONTEXTIDR, PROCID, v) b 461 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_TF(b, c, v) SET_CONTEXT_FIELD(b, c, FSR, TF, v) b 462 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_AFF(b, c, v) SET_CONTEXT_FIELD(b, c, FSR, AFF, v) b 463 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_APF(b, c, v) SET_CONTEXT_FIELD(b, c, FSR, APF, v) b 464 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_TLBMF(b, c, v) SET_CONTEXT_FIELD(b, c, FSR, TLBMF, v) b 465 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_HTWDEEF(b, c, v) SET_CONTEXT_FIELD(b, c, FSR, HTWDEEF, v) b 466 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_HTWSEEF(b, c, v) SET_CONTEXT_FIELD(b, c, FSR, HTWSEEF, v) b 467 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_MHF(b, c, v) SET_CONTEXT_FIELD(b, c, FSR, MHF, v) b 468 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_SL(b, c, v) SET_CONTEXT_FIELD(b, c, FSR, SL, v) b 469 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_SS(b, c, v) SET_CONTEXT_FIELD(b, c, FSR, SS, v) b 470 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_MULTI(b, c, v) SET_CONTEXT_FIELD(b, c, FSR, MULTI, v) b 474 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_AMID(b, c, v) SET_CONTEXT_FIELD(b, c, FSYNR0, AMID, v) b 475 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_APID(b, c, v) SET_CONTEXT_FIELD(b, c, FSYNR0, APID, v) b 476 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_ABID(b, c, v) SET_CONTEXT_FIELD(b, c, FSYNR0, ABID, v) b 477 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_ATID(b, c, v) SET_CONTEXT_FIELD(b, c, FSYNR0, ATID, v) b 481 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_AMEMTYPE(b, c, v) SET_CONTEXT_FIELD(b, c, FSYNR1, AMEMTYPE, v) b 482 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_ASHARED(b, c, v) SET_CONTEXT_FIELD(b, c, FSYNR1, ASHARED, v) b 483 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_AINNERSHARED(b, c, v) \ b 484 drivers/iommu/msm_iommu_hw-8xxx.h SET_CONTEXT_FIELD(b, c, FSYNR1, AINNERSHARED, v) b 485 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_APRIV(b, c, v) SET_CONTEXT_FIELD(b, c, FSYNR1, APRIV, v) b 486 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_APROTNS(b, c, v) SET_CONTEXT_FIELD(b, c, FSYNR1, APROTNS, v) b 487 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_AINST(b, c, v) SET_CONTEXT_FIELD(b, c, FSYNR1, AINST, v) b 488 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_AWRITE(b, c, v) SET_CONTEXT_FIELD(b, c, FSYNR1, AWRITE, v) b 489 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_ABURST(b, c, v) SET_CONTEXT_FIELD(b, c, FSYNR1, ABURST, v) b 490 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_ALEN(b, c, v) SET_CONTEXT_FIELD(b, c, FSYNR1, ALEN, v) b 491 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_FSYNR1_ASIZE(b, c, v) \ b 492 drivers/iommu/msm_iommu_hw-8xxx.h SET_CONTEXT_FIELD(b, c, FSYNR1, FSYNR1_ASIZE, v) b 493 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_ALOCK(b, c, v) SET_CONTEXT_FIELD(b, c, FSYNR1, ALOCK, v) b 494 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_AFULL(b, c, v) SET_CONTEXT_FIELD(b, c, FSYNR1, AFULL, v) b 498 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_ICPC0(b, c, v) SET_CONTEXT_FIELD(b, c, NMRR, ICPC0, v) b 499 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_ICPC1(b, c, v) SET_CONTEXT_FIELD(b, c, NMRR, ICPC1, v) b 500 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_ICPC2(b, c, v) SET_CONTEXT_FIELD(b, c, NMRR, ICPC2, v) b 501 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_ICPC3(b, c, v) SET_CONTEXT_FIELD(b, c, NMRR, ICPC3, v) b 502 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_ICPC4(b, c, v) SET_CONTEXT_FIELD(b, c, NMRR, ICPC4, v) b 503 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_ICPC5(b, c, v) SET_CONTEXT_FIELD(b, c, NMRR, ICPC5, v) b 504 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_ICPC6(b, c, v) SET_CONTEXT_FIELD(b, c, NMRR, ICPC6, v) b 505 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_ICPC7(b, c, v) SET_CONTEXT_FIELD(b, c, NMRR, ICPC7, v) b 506 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_OCPC0(b, c, v) SET_CONTEXT_FIELD(b, c, NMRR, OCPC0, v) b 507 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_OCPC1(b, c, v) SET_CONTEXT_FIELD(b, c, NMRR, OCPC1, v) b 508 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_OCPC2(b, c, v) SET_CONTEXT_FIELD(b, c, NMRR, OCPC2, v) b 509 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_OCPC3(b, c, v) SET_CONTEXT_FIELD(b, c, NMRR, OCPC3, v) b 510 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_OCPC4(b, c, v) SET_CONTEXT_FIELD(b, c, NMRR, OCPC4, v) b 511 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_OCPC5(b, c, v) SET_CONTEXT_FIELD(b, c, NMRR, OCPC5, v) b 512 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_OCPC6(b, c, v) SET_CONTEXT_FIELD(b, c, NMRR, OCPC6, v) b 513 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_OCPC7(b, c, v) SET_CONTEXT_FIELD(b, c, NMRR, OCPC7, v) b 517 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_FAULT(b, c, v) SET_CONTEXT_FIELD(b, c, PAR, FAULT, v) b 519 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_FAULT_TF(b, c, v) SET_CONTEXT_FIELD(b, c, PAR, FAULT_TF, v) b 520 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_FAULT_AFF(b, c, v) SET_CONTEXT_FIELD(b, c, PAR, FAULT_AFF, v) b 521 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_FAULT_APF(b, c, v) SET_CONTEXT_FIELD(b, c, PAR, FAULT_APF, v) b 522 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_FAULT_TLBMF(b, c, v) SET_CONTEXT_FIELD(b, c, PAR, FAULT_TLBMF, v) b 523 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_FAULT_HTWDEEF(b, c, v) \ b 524 drivers/iommu/msm_iommu_hw-8xxx.h SET_CONTEXT_FIELD(b, c, PAR, FAULT_HTWDEEF, v) b 525 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_FAULT_HTWSEEF(b, c, v) \ b 526 drivers/iommu/msm_iommu_hw-8xxx.h SET_CONTEXT_FIELD(b, c, PAR, FAULT_HTWSEEF, v) b 527 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_FAULT_MHF(b, c, v) SET_CONTEXT_FIELD(b, c, PAR, FAULT_MHF, v) b 528 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_FAULT_SL(b, c, v) SET_CONTEXT_FIELD(b, c, PAR, FAULT_SL, v) b 529 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_FAULT_SS(b, c, v) SET_CONTEXT_FIELD(b, c, PAR, FAULT_SS, v) b 531 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_NOFAULT_SS(b, c, v) SET_CONTEXT_FIELD(b, c, PAR, NOFAULT_SS, v) b 532 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_NOFAULT_MT(b, c, v) SET_CONTEXT_FIELD(b, c, PAR, NOFAULT_MT, v) b 533 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_NOFAULT_SH(b, c, v) SET_CONTEXT_FIELD(b, c, PAR, NOFAULT_SH, v) b 534 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_NOFAULT_NS(b, c, v) SET_CONTEXT_FIELD(b, c, PAR, NOFAULT_NS, v) b 535 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_NOFAULT_NOS(b, c, v) SET_CONTEXT_FIELD(b, c, PAR, NOFAULT_NOS, v) b 536 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_NPFAULT_PA(b, c, v) SET_CONTEXT_FIELD(b, c, PAR, NPFAULT_PA, v) b 540 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_MTC0(b, c, v) SET_CONTEXT_FIELD(b, c, PRRR, MTC0, v) b 541 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_MTC1(b, c, v) SET_CONTEXT_FIELD(b, c, PRRR, MTC1, v) b 542 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_MTC2(b, c, v) SET_CONTEXT_FIELD(b, c, PRRR, MTC2, v) b 543 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_MTC3(b, c, v) SET_CONTEXT_FIELD(b, c, PRRR, MTC3, v) b 544 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_MTC4(b, c, v) SET_CONTEXT_FIELD(b, c, PRRR, MTC4, v) b 545 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_MTC5(b, c, v) SET_CONTEXT_FIELD(b, c, PRRR, MTC5, v) b 546 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_MTC6(b, c, v) SET_CONTEXT_FIELD(b, c, PRRR, MTC6, v) b 547 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_MTC7(b, c, v) SET_CONTEXT_FIELD(b, c, PRRR, MTC7, v) b 548 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_SHDSH0(b, c, v) SET_CONTEXT_FIELD(b, c, PRRR, SHDSH0, v) b 549 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_SHDSH1(b, c, v) SET_CONTEXT_FIELD(b, c, PRRR, SHDSH1, v) b 550 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_SHNMSH0(b, c, v) SET_CONTEXT_FIELD(b, c, PRRR, SHNMSH0, v) b 551 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_SHNMSH1(b, c, v) SET_CONTEXT_FIELD(b, c, PRRR, SHNMSH1, v) b 552 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_NOS0(b, c, v) SET_CONTEXT_FIELD(b, c, PRRR, NOS0, v) b 553 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_NOS1(b, c, v) SET_CONTEXT_FIELD(b, c, PRRR, NOS1, v) b 554 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_NOS2(b, c, v) SET_CONTEXT_FIELD(b, c, PRRR, NOS2, v) b 555 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_NOS3(b, c, v) SET_CONTEXT_FIELD(b, c, PRRR, NOS3, v) b 556 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_NOS4(b, c, v) SET_CONTEXT_FIELD(b, c, PRRR, NOS4, v) b 557 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_NOS5(b, c, v) SET_CONTEXT_FIELD(b, c, PRRR, NOS5, v) b 558 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_NOS6(b, c, v) SET_CONTEXT_FIELD(b, c, PRRR, NOS6, v) b 559 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_NOS7(b, c, v) SET_CONTEXT_FIELD(b, c, PRRR, NOS7, v) b 563 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_TNR(b, c, v) SET_CONTEXT_FIELD(b, c, RESUME, TNR, v) b 567 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_M(b, c, v) SET_CONTEXT_FIELD(b, c, SCTLR, M, v) b 568 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_TRE(b, c, v) SET_CONTEXT_FIELD(b, c, SCTLR, TRE, v) b 569 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_AFE(b, c, v) SET_CONTEXT_FIELD(b, c, SCTLR, AFE, v) b 570 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_HAF(b, c, v) SET_CONTEXT_FIELD(b, c, SCTLR, HAF, v) b 571 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_BE(b, c, v) SET_CONTEXT_FIELD(b, c, SCTLR, BE, v) b 572 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_AFFD(b, c, v) SET_CONTEXT_FIELD(b, c, SCTLR, AFFD, v) b 576 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_LKE(b, c, v) SET_CONTEXT_FIELD(b, c, TLBLKCR, LKE, v) b 577 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_TLBLKCR_TLBIALLCFG(b, c, v) \ b 578 drivers/iommu/msm_iommu_hw-8xxx.h SET_CONTEXT_FIELD(b, c, TLBLKCR, TLBLCKR_TLBIALLCFG, v) b 579 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_TLBIASIDCFG(b, c, v) \ b 580 drivers/iommu/msm_iommu_hw-8xxx.h SET_CONTEXT_FIELD(b, c, TLBLKCR, TLBIASIDCFG, v) b 581 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_TLBIVAACFG(b, c, v) SET_CONTEXT_FIELD(b, c, TLBLKCR, TLBIVAACFG, v) b 582 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_FLOOR(b, c, v) SET_CONTEXT_FIELD(b, c, TLBLKCR, FLOOR, v) b 583 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_VICTIM(b, c, v) SET_CONTEXT_FIELD(b, c, TLBLKCR, VICTIM, v) b 587 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_N(b, c, v) SET_CONTEXT_FIELD(b, c, TTBCR, N, v) b 588 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_PD0(b, c, v) SET_CONTEXT_FIELD(b, c, TTBCR, PD0, v) b 589 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_PD1(b, c, v) SET_CONTEXT_FIELD(b, c, TTBCR, PD1, v) b 593 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_TTBR0_IRGNH(b, c, v) SET_CONTEXT_FIELD(b, c, TTBR0, TTBR0_IRGNH, v) b 594 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_TTBR0_SH(b, c, v) SET_CONTEXT_FIELD(b, c, TTBR0, TTBR0_SH, v) b 595 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_TTBR0_ORGN(b, c, v) SET_CONTEXT_FIELD(b, c, TTBR0, TTBR0_ORGN, v) b 596 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_TTBR0_NOS(b, c, v) SET_CONTEXT_FIELD(b, c, TTBR0, TTBR0_NOS, v) b 597 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_TTBR0_IRGNL(b, c, v) SET_CONTEXT_FIELD(b, c, TTBR0, TTBR0_IRGNL, v) b 598 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_TTBR0_PA(b, c, v) SET_CONTEXT_FIELD(b, c, TTBR0, TTBR0_PA, v) b 602 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_TTBR1_IRGNH(b, c, v) SET_CONTEXT_FIELD(b, c, TTBR1, TTBR1_IRGNH, v) b 603 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_TTBR1_SH(b, c, v) SET_CONTEXT_FIELD(b, c, TTBR1, TTBR1_SH, v) b 604 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_TTBR1_ORGN(b, c, v) SET_CONTEXT_FIELD(b, c, TTBR1, TTBR1_ORGN, v) b 605 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_TTBR1_NOS(b, c, v) SET_CONTEXT_FIELD(b, c, TTBR1, TTBR1_NOS, v) b 606 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_TTBR1_IRGNL(b, c, v) SET_CONTEXT_FIELD(b, c, TTBR1, TTBR1_IRGNL, v) b 607 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_TTBR1_PA(b, c, v) SET_CONTEXT_FIELD(b, c, TTBR1, TTBR1_PA, v) b 611 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_HIT(b, c, v) SET_CONTEXT_FIELD(b, c, V2PSR, HIT, v) b 612 drivers/iommu/msm_iommu_hw-8xxx.h #define SET_INDEX(b, c, v) SET_CONTEXT_FIELD(b, c, V2PSR, INDEX, v) b 617 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_CFERE(b, c) GET_CONTEXT_FIELD(b, c, ACTLR, CFERE) b 618 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_CFEIE(b, c) GET_CONTEXT_FIELD(b, c, ACTLR, CFEIE) b 619 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_PTSHCFG(b, c) GET_CONTEXT_FIELD(b, c, ACTLR, PTSHCFG) b 620 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_RCOSH(b, c) GET_CONTEXT_FIELD(b, c, ACTLR, RCOSH) b 621 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_RCISH(b, c) GET_CONTEXT_FIELD(b, c, ACTLR, RCISH) b 622 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_RCNSH(b, c) GET_CONTEXT_FIELD(b, c, ACTLR, RCNSH) b 623 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_PRIVCFG(b, c) GET_CONTEXT_FIELD(b, c, ACTLR, PRIVCFG) b 624 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_DNA(b, c) GET_CONTEXT_FIELD(b, c, ACTLR, DNA) b 625 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_DNLV2PA(b, c) GET_CONTEXT_FIELD(b, c, ACTLR, DNLV2PA) b 626 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_TLBMCFG(b, c) GET_CONTEXT_FIELD(b, c, ACTLR, TLBMCFG) b 627 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_CFCFG(b, c) GET_CONTEXT_FIELD(b, c, ACTLR, CFCFG) b 628 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_TIPCF(b, c) GET_CONTEXT_FIELD(b, c, ACTLR, TIPCF) b 629 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_V2PCFG(b, c) GET_CONTEXT_FIELD(b, c, ACTLR, V2PCFG) b 630 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_HUME(b, c) GET_CONTEXT_FIELD(b, c, ACTLR, HUME) b 631 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_PTMTCFG(b, c) GET_CONTEXT_FIELD(b, c, ACTLR, PTMTCFG) b 632 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_PTMEMTYPE(b, c) GET_CONTEXT_FIELD(b, c, ACTLR, PTMEMTYPE) b 635 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_BFBDFE(b, c) GET_CONTEXT_FIELD(b, c, BFBCR, BFBDFE) b 636 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_BFBSFE(b, c) GET_CONTEXT_FIELD(b, c, BFBCR, BFBSFE) b 637 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_SFVS(b, c) GET_CONTEXT_FIELD(b, c, BFBCR, SFVS) b 638 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_FLVIC(b, c) GET_CONTEXT_FIELD(b, c, BFBCR, FLVIC) b 639 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_SLVIC(b, c) GET_CONTEXT_FIELD(b, c, BFBCR, SLVIC) b 643 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_CONTEXTIDR_ASID(b, c) \ b 644 drivers/iommu/msm_iommu_hw-8xxx.h GET_CONTEXT_FIELD(b, c, CONTEXTIDR, CONTEXTIDR_ASID) b 645 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_CONTEXTIDR_PROCID(b, c) GET_CONTEXT_FIELD(b, c, CONTEXTIDR, PROCID) b 649 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_TF(b, c) GET_CONTEXT_FIELD(b, c, FSR, TF) b 650 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_AFF(b, c) GET_CONTEXT_FIELD(b, c, FSR, AFF) b 651 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_APF(b, c) GET_CONTEXT_FIELD(b, c, FSR, APF) b 652 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_TLBMF(b, c) GET_CONTEXT_FIELD(b, c, FSR, TLBMF) b 653 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_HTWDEEF(b, c) GET_CONTEXT_FIELD(b, c, FSR, HTWDEEF) b 654 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_HTWSEEF(b, c) GET_CONTEXT_FIELD(b, c, FSR, HTWSEEF) b 655 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_MHF(b, c) GET_CONTEXT_FIELD(b, c, FSR, MHF) b 656 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_SL(b, c) GET_CONTEXT_FIELD(b, c, FSR, SL) b 657 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_SS(b, c) GET_CONTEXT_FIELD(b, c, FSR, SS) b 658 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_MULTI(b, c) GET_CONTEXT_FIELD(b, c, FSR, MULTI) b 662 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_AMID(b, c) GET_CONTEXT_FIELD(b, c, FSYNR0, AMID) b 663 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_APID(b, c) GET_CONTEXT_FIELD(b, c, FSYNR0, APID) b 664 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_ABID(b, c) GET_CONTEXT_FIELD(b, c, FSYNR0, ABID) b 665 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_ATID(b, c) GET_CONTEXT_FIELD(b, c, FSYNR0, ATID) b 669 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_AMEMTYPE(b, c) GET_CONTEXT_FIELD(b, c, FSYNR1, AMEMTYPE) b 670 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_ASHARED(b, c) GET_CONTEXT_FIELD(b, c, FSYNR1, ASHARED) b 671 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_AINNERSHARED(b, c) GET_CONTEXT_FIELD(b, c, FSYNR1, AINNERSHARED) b 672 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_APRIV(b, c) GET_CONTEXT_FIELD(b, c, FSYNR1, APRIV) b 673 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_APROTNS(b, c) GET_CONTEXT_FIELD(b, c, FSYNR1, APROTNS) b 674 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_AINST(b, c) GET_CONTEXT_FIELD(b, c, FSYNR1, AINST) b 675 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_AWRITE(b, c) GET_CONTEXT_FIELD(b, c, FSYNR1, AWRITE) b 676 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_ABURST(b, c) GET_CONTEXT_FIELD(b, c, FSYNR1, ABURST) b 677 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_ALEN(b, c) GET_CONTEXT_FIELD(b, c, FSYNR1, ALEN) b 678 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_FSYNR1_ASIZE(b, c) GET_CONTEXT_FIELD(b, c, FSYNR1, FSYNR1_ASIZE) b 679 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_ALOCK(b, c) GET_CONTEXT_FIELD(b, c, FSYNR1, ALOCK) b 680 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_AFULL(b, c) GET_CONTEXT_FIELD(b, c, FSYNR1, AFULL) b 684 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_ICPC0(b, c) GET_CONTEXT_FIELD(b, c, NMRR, ICPC0) b 685 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_ICPC1(b, c) GET_CONTEXT_FIELD(b, c, NMRR, ICPC1) b 686 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_ICPC2(b, c) GET_CONTEXT_FIELD(b, c, NMRR, ICPC2) b 687 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_ICPC3(b, c) GET_CONTEXT_FIELD(b, c, NMRR, ICPC3) b 688 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_ICPC4(b, c) GET_CONTEXT_FIELD(b, c, NMRR, ICPC4) b 689 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_ICPC5(b, c) GET_CONTEXT_FIELD(b, c, NMRR, ICPC5) b 690 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_ICPC6(b, c) GET_CONTEXT_FIELD(b, c, NMRR, ICPC6) b 691 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_ICPC7(b, c) GET_CONTEXT_FIELD(b, c, NMRR, ICPC7) b 692 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_OCPC0(b, c) GET_CONTEXT_FIELD(b, c, NMRR, OCPC0) b 693 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_OCPC1(b, c) GET_CONTEXT_FIELD(b, c, NMRR, OCPC1) b 694 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_OCPC2(b, c) GET_CONTEXT_FIELD(b, c, NMRR, OCPC2) b 695 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_OCPC3(b, c) GET_CONTEXT_FIELD(b, c, NMRR, OCPC3) b 696 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_OCPC4(b, c) GET_CONTEXT_FIELD(b, c, NMRR, OCPC4) b 697 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_OCPC5(b, c) GET_CONTEXT_FIELD(b, c, NMRR, OCPC5) b 698 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_OCPC6(b, c) GET_CONTEXT_FIELD(b, c, NMRR, OCPC6) b 699 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_OCPC7(b, c) GET_CONTEXT_FIELD(b, c, NMRR, OCPC7) b 705 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_FAULT(b, c) GET_CONTEXT_FIELD(b, c, PAR, FAULT) b 707 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_FAULT_TF(b, c) GET_CONTEXT_FIELD(b, c, PAR, FAULT_TF) b 708 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_FAULT_AFF(b, c) GET_CONTEXT_FIELD(b, c, PAR, FAULT_AFF) b 709 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_FAULT_APF(b, c) GET_CONTEXT_FIELD(b, c, PAR, FAULT_APF) b 710 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_FAULT_TLBMF(b, c) GET_CONTEXT_FIELD(b, c, PAR, FAULT_TLBMF) b 711 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_FAULT_HTWDEEF(b, c) GET_CONTEXT_FIELD(b, c, PAR, FAULT_HTWDEEF) b 712 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_FAULT_HTWSEEF(b, c) GET_CONTEXT_FIELD(b, c, PAR, FAULT_HTWSEEF) b 713 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_FAULT_MHF(b, c) GET_CONTEXT_FIELD(b, c, PAR, FAULT_MHF) b 714 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_FAULT_SL(b, c) GET_CONTEXT_FIELD(b, c, PAR, FAULT_SL) b 715 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_FAULT_SS(b, c) GET_CONTEXT_FIELD(b, c, PAR, FAULT_SS) b 717 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_NOFAULT_SS(b, c) GET_CONTEXT_FIELD(b, c, PAR, PAR_NOFAULT_SS) b 718 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_NOFAULT_MT(b, c) GET_CONTEXT_FIELD(b, c, PAR, PAR_NOFAULT_MT) b 719 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_NOFAULT_SH(b, c) GET_CONTEXT_FIELD(b, c, PAR, PAR_NOFAULT_SH) b 720 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_NOFAULT_NS(b, c) GET_CONTEXT_FIELD(b, c, PAR, PAR_NOFAULT_NS) b 721 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_NOFAULT_NOS(b, c) GET_CONTEXT_FIELD(b, c, PAR, PAR_NOFAULT_NOS) b 722 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_NPFAULT_PA(b, c) GET_CONTEXT_FIELD(b, c, PAR, PAR_NPFAULT_PA) b 726 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_MTC0(b, c) GET_CONTEXT_FIELD(b, c, PRRR, MTC0) b 727 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_MTC1(b, c) GET_CONTEXT_FIELD(b, c, PRRR, MTC1) b 728 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_MTC2(b, c) GET_CONTEXT_FIELD(b, c, PRRR, MTC2) b 729 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_MTC3(b, c) GET_CONTEXT_FIELD(b, c, PRRR, MTC3) b 730 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_MTC4(b, c) GET_CONTEXT_FIELD(b, c, PRRR, MTC4) b 731 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_MTC5(b, c) GET_CONTEXT_FIELD(b, c, PRRR, MTC5) b 732 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_MTC6(b, c) GET_CONTEXT_FIELD(b, c, PRRR, MTC6) b 733 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_MTC7(b, c) GET_CONTEXT_FIELD(b, c, PRRR, MTC7) b 734 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_SHDSH0(b, c) GET_CONTEXT_FIELD(b, c, PRRR, SHDSH0) b 735 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_SHDSH1(b, c) GET_CONTEXT_FIELD(b, c, PRRR, SHDSH1) b 736 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_SHNMSH0(b, c) GET_CONTEXT_FIELD(b, c, PRRR, SHNMSH0) b 737 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_SHNMSH1(b, c) GET_CONTEXT_FIELD(b, c, PRRR, SHNMSH1) b 738 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_NOS0(b, c) GET_CONTEXT_FIELD(b, c, PRRR, NOS0) b 739 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_NOS1(b, c) GET_CONTEXT_FIELD(b, c, PRRR, NOS1) b 740 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_NOS2(b, c) GET_CONTEXT_FIELD(b, c, PRRR, NOS2) b 741 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_NOS3(b, c) GET_CONTEXT_FIELD(b, c, PRRR, NOS3) b 742 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_NOS4(b, c) GET_CONTEXT_FIELD(b, c, PRRR, NOS4) b 743 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_NOS5(b, c) GET_CONTEXT_FIELD(b, c, PRRR, NOS5) b 744 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_NOS6(b, c) GET_CONTEXT_FIELD(b, c, PRRR, NOS6) b 745 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_NOS7(b, c) GET_CONTEXT_FIELD(b, c, PRRR, NOS7) b 751 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_TNR(b, c) GET_CONTEXT_FIELD(b, c, RESUME, TNR) b 755 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_M(b, c) GET_CONTEXT_FIELD(b, c, SCTLR, M) b 756 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_TRE(b, c) GET_CONTEXT_FIELD(b, c, SCTLR, TRE) b 757 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_AFE(b, c) GET_CONTEXT_FIELD(b, c, SCTLR, AFE) b 758 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_HAF(b, c) GET_CONTEXT_FIELD(b, c, SCTLR, HAF) b 759 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_BE(b, c) GET_CONTEXT_FIELD(b, c, SCTLR, BE) b 760 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_AFFD(b, c) GET_CONTEXT_FIELD(b, c, SCTLR, AFFD) b 764 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_LKE(b, c) GET_CONTEXT_FIELD(b, c, TLBLKCR, LKE) b 765 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_TLBLCKR_TLBIALLCFG(b, c) \ b 766 drivers/iommu/msm_iommu_hw-8xxx.h GET_CONTEXT_FIELD(b, c, TLBLKCR, TLBLCKR_TLBIALLCFG) b 767 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_TLBIASIDCFG(b, c) GET_CONTEXT_FIELD(b, c, TLBLKCR, TLBIASIDCFG) b 768 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_TLBIVAACFG(b, c) GET_CONTEXT_FIELD(b, c, TLBLKCR, TLBIVAACFG) b 769 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_FLOOR(b, c) GET_CONTEXT_FIELD(b, c, TLBLKCR, FLOOR) b 770 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_VICTIM(b, c) GET_CONTEXT_FIELD(b, c, TLBLKCR, VICTIM) b 774 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_N(b, c) GET_CONTEXT_FIELD(b, c, TTBCR, N) b 775 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_PD0(b, c) GET_CONTEXT_FIELD(b, c, TTBCR, PD0) b 776 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_PD1(b, c) GET_CONTEXT_FIELD(b, c, TTBCR, PD1) b 780 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_TTBR0_IRGNH(b, c) GET_CONTEXT_FIELD(b, c, TTBR0, TTBR0_IRGNH) b 781 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_TTBR0_SH(b, c) GET_CONTEXT_FIELD(b, c, TTBR0, TTBR0_SH) b 782 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_TTBR0_ORGN(b, c) GET_CONTEXT_FIELD(b, c, TTBR0, TTBR0_ORGN) b 783 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_TTBR0_NOS(b, c) GET_CONTEXT_FIELD(b, c, TTBR0, TTBR0_NOS) b 784 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_TTBR0_IRGNL(b, c) GET_CONTEXT_FIELD(b, c, TTBR0, TTBR0_IRGNL) b 785 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_TTBR0_PA(b, c) GET_CONTEXT_FIELD(b, c, TTBR0, TTBR0_PA) b 789 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_TTBR1_IRGNH(b, c) GET_CONTEXT_FIELD(b, c, TTBR1, TTBR1_IRGNH) b 790 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_TTBR1_SH(b, c) GET_CONTEXT_FIELD(b, c, TTBR1, TTBR1_SH) b 791 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_TTBR1_ORGN(b, c) GET_CONTEXT_FIELD(b, c, TTBR1, TTBR1_ORGN) b 792 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_TTBR1_NOS(b, c) GET_CONTEXT_FIELD(b, c, TTBR1, TTBR1_NOS) b 793 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_TTBR1_IRGNL(b, c) GET_CONTEXT_FIELD(b, c, TTBR1, TTBR1_IRGNL) b 794 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_TTBR1_PA(b, c) GET_CONTEXT_FIELD(b, c, TTBR1, TTBR1_PA) b 798 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_HIT(b, c) GET_CONTEXT_FIELD(b, c, V2PSR, HIT) b 799 drivers/iommu/msm_iommu_hw-8xxx.h #define GET_INDEX(b, c) GET_CONTEXT_FIELD(b, c, V2PSR, INDEX) b 503 drivers/irqchip/irq-armada-370-xp.c static void armada_370_xp_handle_msi_irq(struct pt_regs *r, bool b) {} b 48 drivers/irqchip/irq-bcm2835.c #define MAKE_HWIRQ(b, n) ((b << 5) | (n)) b 137 drivers/irqchip/irq-bcm2835.c int irq, b, i; b 148 drivers/irqchip/irq-bcm2835.c for (b = 0; b < NR_BANKS; b++) { b 149 drivers/irqchip/irq-bcm2835.c intc.pending[b] = base + reg_pending[b]; b 150 drivers/irqchip/irq-bcm2835.c intc.enable[b] = base + reg_enable[b]; b 151 drivers/irqchip/irq-bcm2835.c intc.disable[b] = base + reg_disable[b]; b 153 drivers/irqchip/irq-bcm2835.c for (i = 0; i < bank_irqs[b]; i++) { b 154 drivers/irqchip/irq-bcm2835.c irq = irq_create_mapping(intc.domain, MAKE_HWIRQ(b, i)); b 37 drivers/irqchip/irq-bcm7120-l2.c struct bcm7120_l2_intc_data *b; b 58 drivers/irqchip/irq-bcm7120-l2.c struct bcm7120_l2_intc_data *b = data->b; b 64 drivers/irqchip/irq-bcm7120-l2.c for (idx = 0; idx < b->n_words; idx++) { b 67 drivers/irqchip/irq-bcm7120-l2.c irq_get_domain_generic_chip(b->domain, base); b 72 drivers/irqchip/irq-bcm7120-l2.c pending = irq_reg_readl(gc, b->stat_offset[idx]) & b 78 drivers/irqchip/irq-bcm7120-l2.c generic_handle_irq(irq_find_mapping(b->domain, b 88 drivers/irqchip/irq-bcm7120-l2.c struct bcm7120_l2_intc_data *b = gc->private; b 92 drivers/irqchip/irq-bcm7120-l2.c if (b->can_wake) b 142 drivers/irqchip/irq-bcm7120-l2.c l1_data->b = data; b 93 drivers/irqchip/irq-brcmstb-l2.c struct brcmstb_l2_intc_data *b = irq_desc_get_handler_data(desc); b 100 drivers/irqchip/irq-brcmstb-l2.c status = irq_reg_readl(b->gc, b->status_offset) & b 101 drivers/irqchip/irq-brcmstb-l2.c ~(irq_reg_readl(b->gc, b->mask_offset)); b 113 drivers/irqchip/irq-brcmstb-l2.c generic_handle_irq(irq_linear_revmap(b->domain, irq)); b 123 drivers/irqchip/irq-brcmstb-l2.c struct brcmstb_l2_intc_data *b = gc->private; b 128 drivers/irqchip/irq-brcmstb-l2.c b->saved_mask = irq_reg_readl(gc, ct->regs.mask); b 130 drivers/irqchip/irq-brcmstb-l2.c if (b->can_wake) { b 142 drivers/irqchip/irq-brcmstb-l2.c struct brcmstb_l2_intc_data *b = gc->private; b 148 drivers/irqchip/irq-brcmstb-l2.c irq_reg_writel(gc, ~b->saved_mask & ~gc->wake_active, b 153 drivers/irqchip/irq-brcmstb-l2.c irq_reg_writel(gc, b->saved_mask, ct->regs.disable); b 154 drivers/irqchip/irq-brcmstb-l2.c irq_reg_writel(gc, ~b->saved_mask, ct->regs.enable); b 1522 drivers/irqchip/irq-gic-v3-its.c static void merge_lpi_ranges(struct lpi_range *a, struct lpi_range *b) b 1524 drivers/irqchip/irq-gic-v3-its.c if (&a->entry == &lpi_range_list || &b->entry == &lpi_range_list) b 1526 drivers/irqchip/irq-gic-v3-its.c if (a->base_id + a->span != b->base_id) b 1528 drivers/irqchip/irq-gic-v3-its.c b->base_id = a->base_id; b 1529 drivers/irqchip/irq-gic-v3-its.c b->span += a->span; b 213 drivers/isdn/hardware/mISDN/hfc_pci.h #define Write_hfc(a, b, c) (writeb(c, (a->hw.pci_io) + b)) b 214 drivers/isdn/hardware/mISDN/hfc_pci.h #define Read_hfc(a, b) (readb((a->hw.pci_io) + b)) b 1697 drivers/isdn/hardware/mISDN/hfcsusb.c u_char b; b 1708 drivers/isdn/hardware/mISDN/hfcsusb.c memcpy(&b, dmabuf, sizeof(u_char)); b 1717 drivers/isdn/hardware/mISDN/hfcsusb.c if (b != HFCUSB_CHIPID) { b 1719 drivers/isdn/hardware/mISDN/hfcsusb.c hw->name, __func__, b); b 107 drivers/isdn/hardware/mISDN/hfcsusb.h #define write_reg_atomic(a, b, c) \ b 108 drivers/isdn/hardware/mISDN/hfcsusb.h usb_control_msg((a)->dev, (a)->ctrl_out_pipe, 0, 0x40, (c), (b), \ b 110 drivers/isdn/hardware/mISDN/hfcsusb.h #define read_reg_atomic(a, b, c) \ b 111 drivers/isdn/hardware/mISDN/hfcsusb.h usb_control_msg((a)->dev, (a)->ctrl_in_pipe, 1, 0xC0, 0, (b), (c), \ b 360 drivers/isdn/mISDN/dsp_blowfish.c #define EROUND(a, b, n) do { b ^= P[n]; a ^= bf_F(b); } while (0) b 361 drivers/isdn/mISDN/dsp_blowfish.c #define DROUND(a, b, n) do { a ^= bf_F(b); b ^= P[n]; } while (0) b 294 drivers/leds/leds-aat1290.c int i, b = 0, e = AAT1290_MM_CURRENT_SCALE_SIZE; b 296 drivers/leds/leds-aat1290.c while (e - b > 1) { b 297 drivers/leds/leds-aat1290.c i = b + (e - b) / 2; b 301 drivers/leds/leds-aat1290.c b = i; b 304 drivers/leds/leds-aat1290.c cfg->max_mm_current = led->mm_current_scale[b]; b 305 drivers/leds/leds-aat1290.c cfg->max_brightness = b + 1; b 64 drivers/leds/leds-bd2802.c unsigned b:2; b 116 drivers/leds/leds-bd2802.c return !led->led[id].b; b 125 drivers/leds/leds-bd2802.c if (led->led[id].r || led->led[id].g || led->led[id].b) b 186 drivers/leds/leds-bd2802.c led->led[i].b = led_bit; b 743 drivers/leds/leds-bd2802.c if (led->led[i].b) b 744 drivers/leds/leds-bd2802.c bd2802_turn_on(led, i, BLUE, led->led[i].b); b 147 drivers/leds/leds-lp3952.c u8 cmd_index, u8 r, u8 g, u8 b, b 156 drivers/leds/leds-lp3952.c .b = b, b 397 drivers/leds/leds-lp5562.c ptn->b, ptn->size_b); b 117 drivers/leds/leds-pca9532.c int a = 0, b = 0, i = 0; b 123 drivers/leds/leds-pca9532.c b += data->leds[i].ldev.brightness; b 129 drivers/leds/leds-pca9532.c b, a, value); b 132 drivers/leds/leds-pca9532.c b = b/a; b 133 drivers/leds/leds-pca9532.c if (b > 0xFF) b 135 drivers/leds/leds-pca9532.c data->pwm[pwm] = b; b 381 drivers/leds/leds-tca6507.c struct bank *b = tca->bank + led->bank; b 383 drivers/leds/leds-tca6507.c b->time_use--; b 384 drivers/leds/leds-tca6507.c b->level_use--; b 398 drivers/leds/leds-tca6507.c struct bank *b; b 497 drivers/leds/leds-tca6507.c b = &tca->bank[i]; b 498 drivers/leds/leds-tca6507.c if (b->level_use == 0) b 500 drivers/leds/leds-tca6507.c b->level_use++; b 503 drivers/leds/leds-tca6507.c if (b->on_dflt || b 505 drivers/leds/leds-tca6507.c b->time_use == 0) { b 506 drivers/leds/leds-tca6507.c b->ontime = led->ontime; b 507 drivers/leds/leds-tca6507.c b->on_dflt = led->on_dflt; b 511 drivers/leds/leds-tca6507.c if (b->off_dflt || b 513 drivers/leds/leds-tca6507.c b->time_use == 0) { b 514 drivers/leds/leds-tca6507.c b->offtime = led->offtime; b 515 drivers/leds/leds-tca6507.c b->off_dflt = led->off_dflt; b 522 drivers/leds/leds-tca6507.c led->ontime = b->ontime; b 523 drivers/leds/leds-tca6507.c led->offtime = b->offtime; b 525 drivers/leds/leds-tca6507.c b->time_use++; b 527 drivers/leds/leds-tca6507.c led->led_cdev.brightness = TO_BRIGHT(b->level); b 76 drivers/md/bcache/alloc.c uint8_t bch_inc_gen(struct cache *ca, struct bucket *b) b 78 drivers/md/bcache/alloc.c uint8_t ret = ++b->gen; b 80 drivers/md/bcache/alloc.c ca->set->need_gc = max(ca->set->need_gc, bucket_gc_gen(b)); b 89 drivers/md/bcache/alloc.c struct bucket *b; b 108 drivers/md/bcache/alloc.c for_each_bucket(b, ca) b 109 drivers/md/bcache/alloc.c if (b->prio && b 110 drivers/md/bcache/alloc.c b->prio != BTREE_PRIO && b 111 drivers/md/bcache/alloc.c !atomic_read(&b->pin)) { b 112 drivers/md/bcache/alloc.c b->prio--; b 113 drivers/md/bcache/alloc.c c->min_prio = min(c->min_prio, b->prio); b 126 drivers/md/bcache/alloc.c static inline bool can_inc_bucket_gen(struct bucket *b) b 128 drivers/md/bcache/alloc.c return bucket_gc_gen(b) < BUCKET_GC_GEN_MAX; b 131 drivers/md/bcache/alloc.c bool bch_can_invalidate_bucket(struct cache *ca, struct bucket *b) b 135 drivers/md/bcache/alloc.c return (!GC_MARK(b) || b 136 drivers/md/bcache/alloc.c GC_MARK(b) == GC_MARK_RECLAIMABLE) && b 137 drivers/md/bcache/alloc.c !atomic_read(&b->pin) && b 138 drivers/md/bcache/alloc.c can_inc_bucket_gen(b); b 141 drivers/md/bcache/alloc.c void __bch_invalidate_one_bucket(struct cache *ca, struct bucket *b) b 144 drivers/md/bcache/alloc.c BUG_ON(GC_MARK(b) && GC_MARK(b) != GC_MARK_RECLAIMABLE); b 146 drivers/md/bcache/alloc.c if (GC_SECTORS_USED(b)) b 147 drivers/md/bcache/alloc.c trace_bcache_invalidate(ca, b - ca->buckets); b 149 drivers/md/bcache/alloc.c bch_inc_gen(ca, b); b 150 drivers/md/bcache/alloc.c b->prio = INITIAL_PRIO; b 151 drivers/md/bcache/alloc.c atomic_inc(&b->pin); b 154 drivers/md/bcache/alloc.c static void bch_invalidate_one_bucket(struct cache *ca, struct bucket *b) b 156 drivers/md/bcache/alloc.c __bch_invalidate_one_bucket(ca, b); b 158 drivers/md/bcache/alloc.c fifo_push(&ca->free_inc, b - ca->buckets); b 170 drivers/md/bcache/alloc.c #define bucket_prio(b) \ b 174 drivers/md/bcache/alloc.c (b->prio - ca->set->min_prio + min_prio) * GC_SECTORS_USED(b); \ b 182 drivers/md/bcache/alloc.c struct bucket *b; b 187 drivers/md/bcache/alloc.c for_each_bucket(b, ca) { b 188 drivers/md/bcache/alloc.c if (!bch_can_invalidate_bucket(ca, b)) b 192 drivers/md/bcache/alloc.c heap_add(&ca->heap, b, bucket_max_cmp); b 193 drivers/md/bcache/alloc.c else if (bucket_max_cmp(b, heap_peek(&ca->heap))) { b 194 drivers/md/bcache/alloc.c ca->heap.data[0] = b; b 203 drivers/md/bcache/alloc.c if (!heap_pop(&ca->heap, b, bucket_min_cmp)) { b 213 drivers/md/bcache/alloc.c bch_invalidate_one_bucket(ca, b); b 219 drivers/md/bcache/alloc.c struct bucket *b; b 227 drivers/md/bcache/alloc.c b = ca->buckets + ca->fifo_last_bucket++; b 229 drivers/md/bcache/alloc.c if (bch_can_invalidate_bucket(ca, b)) b 230 drivers/md/bcache/alloc.c bch_invalidate_one_bucket(ca, b); b 242 drivers/md/bcache/alloc.c struct bucket *b; b 253 drivers/md/bcache/alloc.c b = ca->buckets + n; b 255 drivers/md/bcache/alloc.c if (bch_can_invalidate_bucket(ca, b)) b 256 drivers/md/bcache/alloc.c bch_invalidate_one_bucket(ca, b); b 396 drivers/md/bcache/alloc.c struct bucket *b; b 446 drivers/md/bcache/alloc.c b = ca->buckets + r; b 448 drivers/md/bcache/alloc.c BUG_ON(atomic_read(&b->pin) != 1); b 450 drivers/md/bcache/alloc.c SET_GC_SECTORS_USED(b, ca->sb.bucket_size); b 453 drivers/md/bcache/alloc.c SET_GC_MARK(b, GC_MARK_METADATA); b 454 drivers/md/bcache/alloc.c SET_GC_MOVE(b, 0); b 455 drivers/md/bcache/alloc.c b->prio = BTREE_PRIO; b 457 drivers/md/bcache/alloc.c SET_GC_MARK(b, GC_MARK_RECLAIMABLE); b 458 drivers/md/bcache/alloc.c SET_GC_MOVE(b, 0); b 459 drivers/md/bcache/alloc.c b->prio = INITIAL_PRIO; b 470 drivers/md/bcache/alloc.c void __bch_bucket_free(struct cache *ca, struct bucket *b) b 472 drivers/md/bcache/alloc.c SET_GC_MARK(b, 0); b 473 drivers/md/bcache/alloc.c SET_GC_SECTORS_USED(b, 0); b 508 drivers/md/bcache/alloc.c long b = bch_bucket_alloc(ca, reserve, wait); b 510 drivers/md/bcache/alloc.c if (b == -1) b 513 drivers/md/bcache/alloc.c k->ptr[i] = MAKE_PTR(ca->buckets[b].gen, b 514 drivers/md/bcache/alloc.c bucket_to_sector(c, b), b 620 drivers/md/bcache/alloc.c struct open_bucket *b; b 634 drivers/md/bcache/alloc.c while (!(b = pick_data_bucket(c, k, write_point, &alloc.key))) { b 655 drivers/md/bcache/alloc.c for (i = 0; i < KEY_PTRS(&b->key); i++) b 656 drivers/md/bcache/alloc.c EBUG_ON(ptr_stale(c, &b->key, i)); b 660 drivers/md/bcache/alloc.c for (i = 0; i < KEY_PTRS(&b->key); i++) b 661 drivers/md/bcache/alloc.c k->ptr[i] = b->key.ptr[i]; b 663 drivers/md/bcache/alloc.c sectors = min(sectors, b->sectors_free); b 667 drivers/md/bcache/alloc.c SET_KEY_PTRS(k, KEY_PTRS(&b->key)); b 673 drivers/md/bcache/alloc.c list_move_tail(&b->list, &c->data_buckets); b 674 drivers/md/bcache/alloc.c bkey_copy_key(&b->key, k); b 675 drivers/md/bcache/alloc.c b->last_write_point = write_point; b 677 drivers/md/bcache/alloc.c b->sectors_free -= sectors; b 679 drivers/md/bcache/alloc.c for (i = 0; i < KEY_PTRS(&b->key); i++) { b 680 drivers/md/bcache/alloc.c SET_PTR_OFFSET(&b->key, i, PTR_OFFSET(&b->key, i) + sectors); b 683 drivers/md/bcache/alloc.c &PTR_CACHE(c, &b->key, i)->sectors_written); b 686 drivers/md/bcache/alloc.c if (b->sectors_free < c->sb.block_size) b 687 drivers/md/bcache/alloc.c b->sectors_free = 0; b 694 drivers/md/bcache/alloc.c if (b->sectors_free) b 695 drivers/md/bcache/alloc.c for (i = 0; i < KEY_PTRS(&b->key); i++) b 696 drivers/md/bcache/alloc.c atomic_inc(&PTR_BUCKET(c, &b->key, i)->pin); b 706 drivers/md/bcache/alloc.c struct open_bucket *b; b 709 drivers/md/bcache/alloc.c b = list_first_entry(&c->data_buckets, b 711 drivers/md/bcache/alloc.c list_del(&b->list); b 712 drivers/md/bcache/alloc.c kfree(b); b 723 drivers/md/bcache/alloc.c struct open_bucket *b = kzalloc(sizeof(*b), GFP_KERNEL); b 725 drivers/md/bcache/alloc.c if (!b) b 728 drivers/md/bcache/alloc.c list_add(&b->list, &c->data_buckets); b 276 drivers/md/bcache/bcache.h int (*cache_miss)(struct btree *b, struct search *s, b 751 drivers/md/bcache/bcache.h #define btree_blocks(b) \ b 752 drivers/md/bcache/bcache.h ((unsigned int) (KEY_SIZE(&b->key) >> (b)->c->block_bits)) b 772 drivers/md/bcache/bcache.h static inline sector_t bucket_to_sector(struct cache_set *c, size_t b) b 774 drivers/md/bcache/bcache.h return ((sector_t) b) << c->bucket_bits; b 803 drivers/md/bcache/bcache.h static inline uint8_t gen_after(uint8_t a, uint8_t b) b 805 drivers/md/bcache/bcache.h uint8_t r = a - b; b 835 drivers/md/bcache/bcache.h #define btree_bug(b, ...) \ b 837 drivers/md/bcache/bcache.h if (bch_cache_set_error((b)->c, __VA_ARGS__)) \ b 847 drivers/md/bcache/bcache.h #define btree_bug_on(cond, b, ...) \ b 850 drivers/md/bcache/bcache.h btree_bug(b, __VA_ARGS__); \ b 870 drivers/md/bcache/bcache.h #define for_each_bucket(b, ca) \ b 871 drivers/md/bcache/bcache.h for (b = (ca)->buckets + (ca)->sb.first_bucket; \ b 872 drivers/md/bcache/bcache.h b < (ca)->buckets + (ca)->sb.nbuckets; b++) b 895 drivers/md/bcache/bcache.h static inline uint8_t bucket_gc_gen(struct bucket *b) b 897 drivers/md/bcache/bcache.h return b->gen - b->last_gc; b 961 drivers/md/bcache/bcache.h uint8_t bch_inc_gen(struct cache *ca, struct bucket *b); b 964 drivers/md/bcache/bcache.h bool bch_can_invalidate_bucket(struct cache *ca, struct bucket *b); b 965 drivers/md/bcache/bcache.h void __bch_invalidate_one_bucket(struct cache *ca, struct bucket *b); b 967 drivers/md/bcache/bcache.h void __bch_bucket_free(struct cache *ca, struct bucket *b); b 21 drivers/md/bcache/bset.c void bch_dump_bset(struct btree_keys *b, struct bset *i, unsigned int set) b 31 drivers/md/bcache/bset.c if (b->ops->key_dump) b 32 drivers/md/bcache/bset.c b->ops->key_dump(b, k); b 37 drivers/md/bcache/bset.c bkey_cmp(k, b->ops->is_extents ? b 43 drivers/md/bcache/bset.c void bch_dump_bucket(struct btree_keys *b) b 48 drivers/md/bcache/bset.c for (i = 0; i <= b->nsets; i++) b 49 drivers/md/bcache/bset.c bch_dump_bset(b, b->set[i].data, b 50 drivers/md/bcache/bset.c bset_sector_offset(b, b->set[i].data)); b 54 drivers/md/bcache/bset.c int __bch_count_data(struct btree_keys *b) b 60 drivers/md/bcache/bset.c if (b->ops->is_extents) b 61 drivers/md/bcache/bset.c for_each_key(b, k, &iter) b 66 drivers/md/bcache/bset.c void __bch_check_keys(struct btree_keys *b, const char *fmt, ...) b 73 drivers/md/bcache/bset.c for_each_key(b, k, &iter) { b 74 drivers/md/bcache/bset.c if (b->ops->is_extents) { b 79 drivers/md/bcache/bset.c if (bch_ptr_invalid(b, k)) b 86 drivers/md/bcache/bset.c if (bch_ptr_bad(b, k)) b 97 drivers/md/bcache/bset.c if (p && bkey_cmp(p, &b->key) > 0) b 102 drivers/md/bcache/bset.c bch_dump_bucket(b); b 116 drivers/md/bcache/bset.c bkey_cmp(k, iter->b->ops->is_extents ? b 118 drivers/md/bcache/bset.c bch_dump_bucket(iter->b); b 266 drivers/md/bcache/bset.c static inline size_t btree_keys_bytes(struct btree_keys *b) b 268 drivers/md/bcache/bset.c return PAGE_SIZE << b->page_order; b 271 drivers/md/bcache/bset.c static inline size_t btree_keys_cachelines(struct btree_keys *b) b 273 drivers/md/bcache/bset.c return btree_keys_bytes(b) / BSET_CACHELINE; b 277 drivers/md/bcache/bset.c static inline size_t bset_tree_bytes(struct btree_keys *b) b 279 drivers/md/bcache/bset.c return btree_keys_cachelines(b) * sizeof(struct bkey_float); b 283 drivers/md/bcache/bset.c static inline size_t bset_prev_bytes(struct btree_keys *b) b 285 drivers/md/bcache/bset.c return btree_keys_cachelines(b) * sizeof(uint8_t); b 290 drivers/md/bcache/bset.c void bch_btree_keys_free(struct btree_keys *b) b 292 drivers/md/bcache/bset.c struct bset_tree *t = b->set; b 294 drivers/md/bcache/bset.c if (bset_prev_bytes(b) < PAGE_SIZE) b 298 drivers/md/bcache/bset.c get_order(bset_prev_bytes(b))); b 300 drivers/md/bcache/bset.c if (bset_tree_bytes(b) < PAGE_SIZE) b 304 drivers/md/bcache/bset.c get_order(bset_tree_bytes(b))); b 306 drivers/md/bcache/bset.c free_pages((unsigned long) t->data, b->page_order); b 314 drivers/md/bcache/bset.c int bch_btree_keys_alloc(struct btree_keys *b, b 318 drivers/md/bcache/bset.c struct bset_tree *t = b->set; b 322 drivers/md/bcache/bset.c b->page_order = page_order; b 324 drivers/md/bcache/bset.c t->data = (void *) __get_free_pages(gfp, b->page_order); b 328 drivers/md/bcache/bset.c t->tree = bset_tree_bytes(b) < PAGE_SIZE b 329 drivers/md/bcache/bset.c ? kmalloc(bset_tree_bytes(b), gfp) b 330 drivers/md/bcache/bset.c : (void *) __get_free_pages(gfp, get_order(bset_tree_bytes(b))); b 334 drivers/md/bcache/bset.c t->prev = bset_prev_bytes(b) < PAGE_SIZE b 335 drivers/md/bcache/bset.c ? kmalloc(bset_prev_bytes(b), gfp) b 336 drivers/md/bcache/bset.c : (void *) __get_free_pages(gfp, get_order(bset_prev_bytes(b))); b 342 drivers/md/bcache/bset.c bch_btree_keys_free(b); b 347 drivers/md/bcache/bset.c void bch_btree_keys_init(struct btree_keys *b, const struct btree_keys_ops *ops, b 350 drivers/md/bcache/bset.c b->ops = ops; b 351 drivers/md/bcache/bset.c b->expensive_debug_checks = expensive_debug_checks; b 352 drivers/md/bcache/bset.c b->nsets = 0; b 353 drivers/md/bcache/bset.c b->last_set_unwritten = 0; b 420 drivers/md/bcache/bset.c unsigned int b = fls(j); b 421 drivers/md/bcache/bset.c unsigned int shift = fls(size - 1) - b; b 423 drivers/md/bcache/bset.c j ^= 1U << (b - 1); b 638 drivers/md/bcache/bset.c static void bset_alloc_tree(struct btree_keys *b, struct bset_tree *t) b 640 drivers/md/bcache/bset.c if (t != b->set) { b 648 drivers/md/bcache/bset.c while (t < b->set + MAX_BSETS) b 652 drivers/md/bcache/bset.c static void bch_bset_build_unwritten_tree(struct btree_keys *b) b 654 drivers/md/bcache/bset.c struct bset_tree *t = bset_tree_last(b); b 656 drivers/md/bcache/bset.c BUG_ON(b->last_set_unwritten); b 657 drivers/md/bcache/bset.c b->last_set_unwritten = 1; b 659 drivers/md/bcache/bset.c bset_alloc_tree(b, t); b 661 drivers/md/bcache/bset.c if (t->tree != b->set->tree + btree_keys_cachelines(b)) { b 667 drivers/md/bcache/bset.c void bch_bset_init_next(struct btree_keys *b, struct bset *i, uint64_t magic) b 669 drivers/md/bcache/bset.c if (i != b->set->data) { b 670 drivers/md/bcache/bset.c b->set[++b->nsets].data = i; b 671 drivers/md/bcache/bset.c i->seq = b->set->data->seq; b 679 drivers/md/bcache/bset.c bch_bset_build_unwritten_tree(b); b 692 drivers/md/bcache/bset.c void bch_bset_build_written_tree(struct btree_keys *b) b 694 drivers/md/bcache/bset.c struct bset_tree *t = bset_tree_last(b); b 698 drivers/md/bcache/bset.c b->last_set_unwritten = 0; b 700 drivers/md/bcache/bset.c bset_alloc_tree(b, t); b 704 drivers/md/bcache/bset.c b->set->tree + btree_keys_cachelines(b) - t->tree); b 739 drivers/md/bcache/bset.c void bch_bset_fix_invalidated_key(struct btree_keys *b, struct bkey *k) b 744 drivers/md/bcache/bset.c for (t = b->set; t <= bset_tree_last(b); t++) b 750 drivers/md/bcache/bset.c if (!t->size || !bset_written(b, t)) b 785 drivers/md/bcache/bset.c static void bch_bset_fix_lookup_table(struct btree_keys *b, b 822 drivers/md/bcache/bset.c if (t->size == b->set->tree + btree_keys_cachelines(b) - t->tree) b 842 drivers/md/bcache/bset.c bool bch_bkey_try_merge(struct btree_keys *b, struct bkey *l, struct bkey *r) b 844 drivers/md/bcache/bset.c if (!b->ops->key_merge) b 856 drivers/md/bcache/bset.c return b->ops->key_merge(b, l, r); b 860 drivers/md/bcache/bset.c void bch_bset_insert(struct btree_keys *b, struct bkey *where, b 863 drivers/md/bcache/bset.c struct bset_tree *t = bset_tree_last(b); b 865 drivers/md/bcache/bset.c BUG_ON(!b->last_set_unwritten); b 866 drivers/md/bcache/bset.c BUG_ON(bset_byte_offset(b, t->data) + b 868 drivers/md/bcache/bset.c PAGE_SIZE << b->page_order); b 876 drivers/md/bcache/bset.c bch_bset_fix_lookup_table(b, t, where); b 880 drivers/md/bcache/bset.c unsigned int bch_btree_insert_key(struct btree_keys *b, struct bkey *k, b 884 drivers/md/bcache/bset.c struct bset *i = bset_tree_last(b)->data; b 890 drivers/md/bcache/bset.c BUG_ON(b->ops->is_extents && !KEY_SIZE(k)); b 897 drivers/md/bcache/bset.c if (b->ops->is_extents) b 902 drivers/md/bcache/bset.c m = bch_btree_iter_init(b, &iter, preceding_key_p); b 904 drivers/md/bcache/bset.c if (b->ops->insert_fixup(b, k, &iter, replace_key)) b 910 drivers/md/bcache/bset.c bkey_cmp(k, b->ops->is_extents ? &START_KEY(m) : m) > 0) b 916 drivers/md/bcache/bset.c bch_bkey_try_merge(b, prev, k)) b 926 drivers/md/bcache/bset.c bch_bkey_try_merge(b, k, m)) b 929 drivers/md/bcache/bset.c bch_bset_insert(b, m, k); b 1018 drivers/md/bcache/bset.c struct bkey *__bch_bset_search(struct btree_keys *b, struct bset_tree *t, b 1041 drivers/md/bcache/bset.c } else if (bset_written(b, t)) { b 1057 drivers/md/bcache/bset.c BUG_ON(!b->nsets && b 1063 drivers/md/bcache/bset.c if (btree_keys_expensive_checks(b)) { b 1064 drivers/md/bcache/bset.c BUG_ON(bset_written(b, t) && b 1107 drivers/md/bcache/bset.c static struct bkey *__bch_btree_iter_init(struct btree_keys *b, b 1118 drivers/md/bcache/bset.c iter->b = b; b 1121 drivers/md/bcache/bset.c for (; start <= bset_tree_last(b); start++) { b 1122 drivers/md/bcache/bset.c ret = bch_bset_search(b, start, search); b 1129 drivers/md/bcache/bset.c struct bkey *bch_btree_iter_init(struct btree_keys *b, b 1133 drivers/md/bcache/bset.c return __bch_btree_iter_init(b, iter, search, b->set); b 1140 drivers/md/bcache/bset.c struct btree_iter_set b __maybe_unused; b 1155 drivers/md/bcache/bset.c heap_pop(iter, b, cmp); b 1171 drivers/md/bcache/bset.c struct btree_keys *b, ptr_filter_fn fn) b 1177 drivers/md/bcache/bset.c } while (ret && fn(b, ret)); b 1201 drivers/md/bcache/bset.c static void btree_mergesort(struct btree_keys *b, struct bset *out, b 1214 drivers/md/bcache/bset.c heap_sift(iter, i, b->ops->sort_cmp); b 1217 drivers/md/bcache/bset.c if (b->ops->sort_fixup && fixup) b 1218 drivers/md/bcache/bset.c k = b->ops->sort_fixup(iter, &tmp.k); b 1223 drivers/md/bcache/bset.c k = __bch_btree_iter_next(iter, b->ops->sort_cmp); b 1225 drivers/md/bcache/bset.c if (bad(b, k)) b 1231 drivers/md/bcache/bset.c } else if (!bch_bkey_try_merge(b, last, k)) { b 1242 drivers/md/bcache/bset.c static void __btree_sort(struct btree_keys *b, struct btree_iter *iter, b 1263 drivers/md/bcache/bset.c btree_mergesort(b, out, iter, fixup, false); b 1264 drivers/md/bcache/bset.c b->nsets = start; b 1266 drivers/md/bcache/bset.c if (!start && order == b->page_order) { b 1273 drivers/md/bcache/bset.c out->magic = b->set->data->magic; b 1274 drivers/md/bcache/bset.c out->seq = b->set->data->seq; b 1275 drivers/md/bcache/bset.c out->version = b->set->data->version; b 1276 drivers/md/bcache/bset.c swap(out, b->set->data); b 1278 drivers/md/bcache/bset.c b->set[start].data->keys = out->keys; b 1279 drivers/md/bcache/bset.c memcpy(b->set[start].data->start, out->start, b 1288 drivers/md/bcache/bset.c bch_bset_build_written_tree(b); b 1294 drivers/md/bcache/bset.c void bch_btree_sort_partial(struct btree_keys *b, unsigned int start, b 1297 drivers/md/bcache/bset.c size_t order = b->page_order, keys = 0; b 1299 drivers/md/bcache/bset.c int oldsize = bch_count_data(b); b 1301 drivers/md/bcache/bset.c __bch_btree_iter_init(b, &iter, NULL, &b->set[start]); b 1306 drivers/md/bcache/bset.c for (i = start; i <= b->nsets; i++) b 1307 drivers/md/bcache/bset.c keys += b->set[i].data->keys; b 1309 drivers/md/bcache/bset.c order = get_order(__set_bytes(b->set->data, keys)); b 1312 drivers/md/bcache/bset.c __btree_sort(b, &iter, start, order, false, state); b 1314 drivers/md/bcache/bset.c EBUG_ON(oldsize >= 0 && bch_count_data(b) != oldsize); b 1318 drivers/md/bcache/bset.c void bch_btree_sort_and_fix_extents(struct btree_keys *b, b 1322 drivers/md/bcache/bset.c __btree_sort(b, iter, 0, b->page_order, true, state); b 1325 drivers/md/bcache/bset.c void bch_btree_sort_into(struct btree_keys *b, struct btree_keys *new, b 1331 drivers/md/bcache/bset.c bch_btree_iter_init(b, &iter, NULL); b 1333 drivers/md/bcache/bset.c btree_mergesort(b, new->set->data, &iter, false, true); b 1342 drivers/md/bcache/bset.c void bch_btree_sort_lazy(struct btree_keys *b, struct bset_sort_state *state) b 1348 drivers/md/bcache/bset.c if (!b->nsets) b 1351 drivers/md/bcache/bset.c for (i = b->nsets - 1; i >= 0; --i) { b 1354 drivers/md/bcache/bset.c if (b->set[i].data->keys < crit) { b 1355 drivers/md/bcache/bset.c bch_btree_sort_partial(b, i, state); b 1361 drivers/md/bcache/bset.c if (b->nsets + 1 == MAX_BSETS) { b 1362 drivers/md/bcache/bset.c bch_btree_sort(b, state); b 1367 drivers/md/bcache/bset.c bch_bset_build_written_tree(b); b 1371 drivers/md/bcache/bset.c void bch_btree_keys_stats(struct btree_keys *b, struct bset_stats *stats) b 1375 drivers/md/bcache/bset.c for (i = 0; i <= b->nsets; i++) { b 1376 drivers/md/bcache/bset.c struct bset_tree *t = &b->set[i]; b 1380 drivers/md/bcache/bset.c if (bset_written(b, t)) { b 194 drivers/md/bcache/bset.h bool (*insert_fixup)(struct btree_keys *b, b 234 drivers/md/bcache/bset.h static inline struct bset_tree *bset_tree_last(struct btree_keys *b) b 236 drivers/md/bcache/bset.h return b->set + b->nsets; b 239 drivers/md/bcache/bset.h static inline bool bset_written(struct btree_keys *b, struct bset_tree *t) b 241 drivers/md/bcache/bset.h return t <= b->set + b->nsets - b->last_set_unwritten; b 244 drivers/md/bcache/bset.h static inline bool bkey_written(struct btree_keys *b, struct bkey *k) b 246 drivers/md/bcache/bset.h return !b->last_set_unwritten || k < b->set[b->nsets].data->start; b 249 drivers/md/bcache/bset.h static inline unsigned int bset_byte_offset(struct btree_keys *b, b 252 drivers/md/bcache/bset.h return ((size_t) i) - ((size_t) b->set->data); b 255 drivers/md/bcache/bset.h static inline unsigned int bset_sector_offset(struct btree_keys *b, b 258 drivers/md/bcache/bset.h return bset_byte_offset(b, i) >> 9; b 269 drivers/md/bcache/bset.h static inline size_t bch_btree_keys_u64s_remaining(struct btree_keys *b) b 271 drivers/md/bcache/bset.h struct bset_tree *t = bset_tree_last(b); b 273 drivers/md/bcache/bset.h BUG_ON((PAGE_SIZE << b->page_order) < b 274 drivers/md/bcache/bset.h (bset_byte_offset(b, t->data) + set_bytes(t->data))); b 276 drivers/md/bcache/bset.h if (!b->last_set_unwritten) b 279 drivers/md/bcache/bset.h return ((PAGE_SIZE << b->page_order) - b 280 drivers/md/bcache/bset.h (bset_byte_offset(b, t->data) + set_bytes(t->data))) / b 284 drivers/md/bcache/bset.h static inline struct bset *bset_next_set(struct btree_keys *b, b 287 drivers/md/bcache/bset.h struct bset *i = bset_tree_last(b)->data; b 292 drivers/md/bcache/bset.h void bch_btree_keys_free(struct btree_keys *b); b 293 drivers/md/bcache/bset.h int bch_btree_keys_alloc(struct btree_keys *b, unsigned int page_order, b 295 drivers/md/bcache/bset.h void bch_btree_keys_init(struct btree_keys *b, const struct btree_keys_ops *ops, b 298 drivers/md/bcache/bset.h void bch_bset_init_next(struct btree_keys *b, struct bset *i, uint64_t magic); b 299 drivers/md/bcache/bset.h void bch_bset_build_written_tree(struct btree_keys *b); b 300 drivers/md/bcache/bset.h void bch_bset_fix_invalidated_key(struct btree_keys *b, struct bkey *k); b 301 drivers/md/bcache/bset.h bool bch_bkey_try_merge(struct btree_keys *b, struct bkey *l, struct bkey *r); b 302 drivers/md/bcache/bset.h void bch_bset_insert(struct btree_keys *b, struct bkey *where, b 304 drivers/md/bcache/bset.h unsigned int bch_btree_insert_key(struct btree_keys *b, struct bkey *k, b 320 drivers/md/bcache/bset.h struct btree_keys *b; b 327 drivers/md/bcache/bset.h typedef bool (*ptr_filter_fn)(struct btree_keys *b, const struct bkey *k); b 331 drivers/md/bcache/bset.h struct btree_keys *b, b 336 drivers/md/bcache/bset.h struct bkey *bch_btree_iter_init(struct btree_keys *b, b 340 drivers/md/bcache/bset.h struct bkey *__bch_bset_search(struct btree_keys *b, struct bset_tree *t, b 346 drivers/md/bcache/bset.h static inline struct bkey *bch_bset_search(struct btree_keys *b, b 350 drivers/md/bcache/bset.h return search ? __bch_bset_search(b, t, search) : t->data->start; b 353 drivers/md/bcache/bset.h #define for_each_key_filter(b, k, iter, filter) \ b 354 drivers/md/bcache/bset.h for (bch_btree_iter_init((b), (iter), NULL); \ b 355 drivers/md/bcache/bset.h ((k) = bch_btree_iter_next_filter((iter), (b), filter));) b 357 drivers/md/bcache/bset.h #define for_each_key(b, k, iter) \ b 358 drivers/md/bcache/bset.h for (bch_btree_iter_init((b), (iter), NULL); \ b 375 drivers/md/bcache/bset.h void bch_btree_sort_lazy(struct btree_keys *b, struct bset_sort_state *state); b 376 drivers/md/bcache/bset.h void bch_btree_sort_into(struct btree_keys *b, struct btree_keys *new, b 378 drivers/md/bcache/bset.h void bch_btree_sort_and_fix_extents(struct btree_keys *b, b 381 drivers/md/bcache/bset.h void bch_btree_sort_partial(struct btree_keys *b, unsigned int start, b 384 drivers/md/bcache/bset.h static inline void bch_btree_sort(struct btree_keys *b, b 387 drivers/md/bcache/bset.h bch_btree_sort_partial(b, 0, state); b 396 drivers/md/bcache/bset.h void bch_btree_keys_stats(struct btree_keys *b, struct bset_stats *state); b 459 drivers/md/bcache/bset.h static inline bool bch_ptr_invalid(struct btree_keys *b, const struct bkey *k) b 461 drivers/md/bcache/bset.h return b->ops->key_invalid(b, k); b 464 drivers/md/bcache/bset.h static inline bool bch_ptr_bad(struct btree_keys *b, const struct bkey *k) b 466 drivers/md/bcache/bset.h return b->ops->key_bad(b, k); b 469 drivers/md/bcache/bset.h static inline void bch_bkey_to_text(struct btree_keys *b, char *buf, b 472 drivers/md/bcache/bset.h return b->ops->key_to_text(buf, size, k); b 556 drivers/md/bcache/bset.h int __bch_count_data(struct btree_keys *b); b 557 drivers/md/bcache/bset.h void __printf(2, 3) __bch_check_keys(struct btree_keys *b, b 560 drivers/md/bcache/bset.h void bch_dump_bset(struct btree_keys *b, struct bset *i, unsigned int set); b 561 drivers/md/bcache/bset.h void bch_dump_bucket(struct btree_keys *b); b 565 drivers/md/bcache/bset.h static inline int __bch_count_data(struct btree_keys *b) { return -1; } b 567 drivers/md/bcache/bset.h __bch_check_keys(struct btree_keys *b, const char *fmt, ...) {} b 568 drivers/md/bcache/bset.h static inline void bch_dump_bucket(struct btree_keys *b) {} b 569 drivers/md/bcache/bset.h void bch_dump_bset(struct btree_keys *b, struct bset *i, unsigned int set); b 573 drivers/md/bcache/bset.h static inline bool btree_keys_expensive_checks(struct btree_keys *b) b 576 drivers/md/bcache/bset.h return *b->expensive_debug_checks; b 582 drivers/md/bcache/bset.h static inline int bch_count_data(struct btree_keys *b) b 584 drivers/md/bcache/bset.h return btree_keys_expensive_checks(b) ? __bch_count_data(b) : -1; b 587 drivers/md/bcache/bset.h #define bch_check_keys(b, ...) \ b 589 drivers/md/bcache/bset.h if (btree_keys_expensive_checks(b)) \ b 590 drivers/md/bcache/bset.h __bch_check_keys(b, __VA_ARGS__); \ b 102 drivers/md/bcache/btree.c #define insert_lock(s, b) ((b)->level <= (s)->lock) b 122 drivers/md/bcache/btree.c #define btree(fn, key, b, op, ...) \ b 124 drivers/md/bcache/btree.c int _r, l = (b)->level - 1; \ b 126 drivers/md/bcache/btree.c struct btree *_child = bch_btree_node_get((b)->c, op, key, l, \ b 127 drivers/md/bcache/btree.c _w, b); \ b 163 drivers/md/bcache/btree.c static inline struct bset *write_block(struct btree *b) b 165 drivers/md/bcache/btree.c return ((void *) btree_bset_first(b)) + b->written * block_bytes(b->c); b 168 drivers/md/bcache/btree.c static void bch_btree_init_next(struct btree *b) b 171 drivers/md/bcache/btree.c if (b->level && b->keys.nsets) b 172 drivers/md/bcache/btree.c bch_btree_sort(&b->keys, &b->c->sort); b 174 drivers/md/bcache/btree.c bch_btree_sort_lazy(&b->keys, &b->c->sort); b 176 drivers/md/bcache/btree.c if (b->written < btree_blocks(b)) b 177 drivers/md/bcache/btree.c bch_bset_init_next(&b->keys, write_block(b), b 178 drivers/md/bcache/btree.c bset_magic(&b->c->sb)); b 195 drivers/md/bcache/btree.c static uint64_t btree_csum_set(struct btree *b, struct bset *i) b 197 drivers/md/bcache/btree.c uint64_t crc = b->key.ptr[0]; b 204 drivers/md/bcache/btree.c void bch_btree_node_read_done(struct btree *b) b 207 drivers/md/bcache/btree.c struct bset *i = btree_bset_first(b); b 215 drivers/md/bcache/btree.c iter = mempool_alloc(&b->c->fill_iter, GFP_NOIO); b 216 drivers/md/bcache/btree.c iter->size = b->c->sb.bucket_size / b->c->sb.block_size; b 220 drivers/md/bcache/btree.c iter->b = &b->keys; b 227 drivers/md/bcache/btree.c b->written < btree_blocks(b) && i->seq == b->keys.set[0].data->seq; b 228 drivers/md/bcache/btree.c i = write_block(b)) { b 234 drivers/md/bcache/btree.c if (b->written + set_blocks(i, block_bytes(b->c)) > b 235 drivers/md/bcache/btree.c btree_blocks(b)) b 239 drivers/md/bcache/btree.c if (i->magic != bset_magic(&b->c->sb)) b 249 drivers/md/bcache/btree.c if (i->csum != btree_csum_set(b, i)) b 255 drivers/md/bcache/btree.c if (i != b->keys.set[0].data && !i->keys) b 260 drivers/md/bcache/btree.c b->written += set_blocks(i, block_bytes(b->c)); b 264 drivers/md/bcache/btree.c for (i = write_block(b); b 265 drivers/md/bcache/btree.c bset_sector_offset(&b->keys, i) < KEY_SIZE(&b->key); b 266 drivers/md/bcache/btree.c i = ((void *) i) + block_bytes(b->c)) b 267 drivers/md/bcache/btree.c if (i->seq == b->keys.set[0].data->seq) b 270 drivers/md/bcache/btree.c bch_btree_sort_and_fix_extents(&b->keys, iter, &b->c->sort); b 272 drivers/md/bcache/btree.c i = b->keys.set[0].data; b 274 drivers/md/bcache/btree.c if (b->keys.set[0].size && b 275 drivers/md/bcache/btree.c bkey_cmp(&b->key, &b->keys.set[0].end) < 0) b 278 drivers/md/bcache/btree.c if (b->written < btree_blocks(b)) b 279 drivers/md/bcache/btree.c bch_bset_init_next(&b->keys, write_block(b), b 280 drivers/md/bcache/btree.c bset_magic(&b->c->sb)); b 282 drivers/md/bcache/btree.c mempool_free(iter, &b->c->fill_iter); b 285 drivers/md/bcache/btree.c set_btree_node_io_error(b); b 286 drivers/md/bcache/btree.c bch_cache_set_error(b->c, "%s at bucket %zu, block %u, %u keys", b 287 drivers/md/bcache/btree.c err, PTR_BUCKET_NR(b->c, &b->key, 0), b 288 drivers/md/bcache/btree.c bset_block_offset(b, i), i->keys); b 299 drivers/md/bcache/btree.c static void bch_btree_node_read(struct btree *b) b 305 drivers/md/bcache/btree.c trace_bcache_btree_read(b); b 309 drivers/md/bcache/btree.c bio = bch_bbio_alloc(b->c); b 310 drivers/md/bcache/btree.c bio->bi_iter.bi_size = KEY_SIZE(&b->key) << 9; b 315 drivers/md/bcache/btree.c bch_bio_map(bio, b->keys.set[0].data); b 317 drivers/md/bcache/btree.c bch_submit_bbio(bio, b->c, &b->key, 0); b 321 drivers/md/bcache/btree.c set_btree_node_io_error(b); b 323 drivers/md/bcache/btree.c bch_bbio_free(bio, b->c); b 325 drivers/md/bcache/btree.c if (btree_node_io_error(b)) b 328 drivers/md/bcache/btree.c bch_btree_node_read_done(b); b 329 drivers/md/bcache/btree.c bch_time_stats_update(&b->c->btree_read_time, start_time); b 333 drivers/md/bcache/btree.c bch_cache_set_error(b->c, "io error reading bucket %zu", b 334 drivers/md/bcache/btree.c PTR_BUCKET_NR(b->c, &b->key, 0)); b 337 drivers/md/bcache/btree.c static void btree_complete_write(struct btree *b, struct btree_write *w) b 340 drivers/md/bcache/btree.c !atomic_sub_return(w->prio_blocked, &b->c->prio_blocked)) b 341 drivers/md/bcache/btree.c wake_up_allocators(b->c); b 345 drivers/md/bcache/btree.c __closure_wake_up(&b->c->journal.wait); b 354 drivers/md/bcache/btree.c struct btree *b = container_of(cl, struct btree, io); b 356 drivers/md/bcache/btree.c up(&b->io_mutex); b 361 drivers/md/bcache/btree.c struct btree *b = container_of(cl, struct btree, io); b 362 drivers/md/bcache/btree.c struct btree_write *w = btree_prev_write(b); b 364 drivers/md/bcache/btree.c bch_bbio_free(b->bio, b->c); b 365 drivers/md/bcache/btree.c b->bio = NULL; b 366 drivers/md/bcache/btree.c btree_complete_write(b, w); b 368 drivers/md/bcache/btree.c if (btree_node_dirty(b)) b 369 drivers/md/bcache/btree.c schedule_delayed_work(&b->work, 30 * HZ); b 376 drivers/md/bcache/btree.c struct btree *b = container_of(cl, struct btree, io); b 378 drivers/md/bcache/btree.c bio_free_pages(b->bio); b 385 drivers/md/bcache/btree.c struct btree *b = container_of(cl, struct btree, io); b 388 drivers/md/bcache/btree.c set_btree_node_io_error(b); b 390 drivers/md/bcache/btree.c bch_bbio_count_io_errors(b->c, bio, bio->bi_status, "writing btree"); b 394 drivers/md/bcache/btree.c static void do_btree_node_write(struct btree *b) b 396 drivers/md/bcache/btree.c struct closure *cl = &b->io; b 397 drivers/md/bcache/btree.c struct bset *i = btree_bset_last(b); b 401 drivers/md/bcache/btree.c i->csum = btree_csum_set(b, i); b 403 drivers/md/bcache/btree.c BUG_ON(b->bio); b 404 drivers/md/bcache/btree.c b->bio = bch_bbio_alloc(b->c); b 406 drivers/md/bcache/btree.c b->bio->bi_end_io = btree_node_write_endio; b 407 drivers/md/bcache/btree.c b->bio->bi_private = cl; b 408 drivers/md/bcache/btree.c b->bio->bi_iter.bi_size = roundup(set_bytes(i), block_bytes(b->c)); b 409 drivers/md/bcache/btree.c b->bio->bi_opf = REQ_OP_WRITE | REQ_META | REQ_FUA; b 410 drivers/md/bcache/btree.c bch_bio_map(b->bio, i); b 427 drivers/md/bcache/btree.c bkey_copy(&k.key, &b->key); b 429 drivers/md/bcache/btree.c bset_sector_offset(&b->keys, i)); b 431 drivers/md/bcache/btree.c if (!bch_bio_alloc_pages(b->bio, __GFP_NOWARN|GFP_NOWAIT)) { b 436 drivers/md/bcache/btree.c bio_for_each_segment_all(bv, b->bio, iter_all) { b 441 drivers/md/bcache/btree.c bch_submit_bbio(b->bio, b->c, &k.key, 0); b 449 drivers/md/bcache/btree.c b->bio->bi_vcnt = 0; b 450 drivers/md/bcache/btree.c bch_bio_map(b->bio, i); b 452 drivers/md/bcache/btree.c bch_submit_bbio(b->bio, b->c, &k.key, 0); b 459 drivers/md/bcache/btree.c void __bch_btree_node_write(struct btree *b, struct closure *parent) b 461 drivers/md/bcache/btree.c struct bset *i = btree_bset_last(b); b 463 drivers/md/bcache/btree.c lockdep_assert_held(&b->write_lock); b 465 drivers/md/bcache/btree.c trace_bcache_btree_write(b); b 468 drivers/md/bcache/btree.c BUG_ON(b->written >= btree_blocks(b)); b 469 drivers/md/bcache/btree.c BUG_ON(b->written && !i->keys); b 470 drivers/md/bcache/btree.c BUG_ON(btree_bset_first(b)->seq != i->seq); b 471 drivers/md/bcache/btree.c bch_check_keys(&b->keys, "writing"); b 473 drivers/md/bcache/btree.c cancel_delayed_work(&b->work); b 476 drivers/md/bcache/btree.c down(&b->io_mutex); b 477 drivers/md/bcache/btree.c closure_init(&b->io, parent ?: &b->c->cl); b 479 drivers/md/bcache/btree.c clear_bit(BTREE_NODE_dirty, &b->flags); b 480 drivers/md/bcache/btree.c change_bit(BTREE_NODE_write_idx, &b->flags); b 482 drivers/md/bcache/btree.c do_btree_node_write(b); b 484 drivers/md/bcache/btree.c atomic_long_add(set_blocks(i, block_bytes(b->c)) * b->c->sb.block_size, b 485 drivers/md/bcache/btree.c &PTR_CACHE(b->c, &b->key, 0)->btree_sectors_written); b 487 drivers/md/bcache/btree.c b->written += set_blocks(i, block_bytes(b->c)); b 490 drivers/md/bcache/btree.c void bch_btree_node_write(struct btree *b, struct closure *parent) b 492 drivers/md/bcache/btree.c unsigned int nsets = b->keys.nsets; b 494 drivers/md/bcache/btree.c lockdep_assert_held(&b->lock); b 496 drivers/md/bcache/btree.c __bch_btree_node_write(b, parent); b 502 drivers/md/bcache/btree.c if (nsets && !b->keys.nsets) b 503 drivers/md/bcache/btree.c bch_btree_verify(b); b 505 drivers/md/bcache/btree.c bch_btree_init_next(b); b 508 drivers/md/bcache/btree.c static void bch_btree_node_write_sync(struct btree *b) b 514 drivers/md/bcache/btree.c mutex_lock(&b->write_lock); b 515 drivers/md/bcache/btree.c bch_btree_node_write(b, &cl); b 516 drivers/md/bcache/btree.c mutex_unlock(&b->write_lock); b 523 drivers/md/bcache/btree.c struct btree *b = container_of(to_delayed_work(w), struct btree, work); b 525 drivers/md/bcache/btree.c mutex_lock(&b->write_lock); b 526 drivers/md/bcache/btree.c if (btree_node_dirty(b)) b 527 drivers/md/bcache/btree.c __bch_btree_node_write(b, NULL); b 528 drivers/md/bcache/btree.c mutex_unlock(&b->write_lock); b 531 drivers/md/bcache/btree.c static void bch_btree_leaf_dirty(struct btree *b, atomic_t *journal_ref) b 533 drivers/md/bcache/btree.c struct bset *i = btree_bset_last(b); b 534 drivers/md/bcache/btree.c struct btree_write *w = btree_current_write(b); b 536 drivers/md/bcache/btree.c lockdep_assert_held(&b->write_lock); b 538 drivers/md/bcache/btree.c BUG_ON(!b->written); b 541 drivers/md/bcache/btree.c if (!btree_node_dirty(b)) b 542 drivers/md/bcache/btree.c schedule_delayed_work(&b->work, 30 * HZ); b 544 drivers/md/bcache/btree.c set_btree_node_dirty(b); b 548 drivers/md/bcache/btree.c journal_pin_cmp(b->c, w->journal, journal_ref)) { b 562 drivers/md/bcache/btree.c bch_btree_node_write(b, NULL); b 575 drivers/md/bcache/btree.c static void mca_data_free(struct btree *b) b 577 drivers/md/bcache/btree.c BUG_ON(b->io_mutex.count != 1); b 579 drivers/md/bcache/btree.c bch_btree_keys_free(&b->keys); b 581 drivers/md/bcache/btree.c b->c->btree_cache_used--; b 582 drivers/md/bcache/btree.c list_move(&b->list, &b->c->btree_cache_freed); b 585 drivers/md/bcache/btree.c static void mca_bucket_free(struct btree *b) b 587 drivers/md/bcache/btree.c BUG_ON(btree_node_dirty(b)); b 589 drivers/md/bcache/btree.c b->key.ptr[0] = 0; b 590 drivers/md/bcache/btree.c hlist_del_init_rcu(&b->hash); b 591 drivers/md/bcache/btree.c list_move(&b->list, &b->c->btree_cache_freeable); b 599 drivers/md/bcache/btree.c static void mca_data_alloc(struct btree *b, struct bkey *k, gfp_t gfp) b 601 drivers/md/bcache/btree.c if (!bch_btree_keys_alloc(&b->keys, b 603 drivers/md/bcache/btree.c ilog2(b->c->btree_pages), b 606 drivers/md/bcache/btree.c b->c->btree_cache_used++; b 607 drivers/md/bcache/btree.c list_move(&b->list, &b->c->btree_cache); b 609 drivers/md/bcache/btree.c list_move(&b->list, &b->c->btree_cache_freed); b 620 drivers/md/bcache/btree.c struct btree *b = kzalloc(sizeof(struct btree), gfp); b 622 drivers/md/bcache/btree.c if (!b) b 625 drivers/md/bcache/btree.c init_rwsem(&b->lock); b 626 drivers/md/bcache/btree.c lockdep_set_novalidate_class(&b->lock); b 627 drivers/md/bcache/btree.c mutex_init(&b->write_lock); b 628 drivers/md/bcache/btree.c lockdep_set_novalidate_class(&b->write_lock); b 629 drivers/md/bcache/btree.c INIT_LIST_HEAD(&b->list); b 630 drivers/md/bcache/btree.c INIT_DELAYED_WORK(&b->work, btree_node_write_work); b 631 drivers/md/bcache/btree.c b->c = c; b 632 drivers/md/bcache/btree.c sema_init(&b->io_mutex, 1); b 634 drivers/md/bcache/btree.c mca_data_alloc(b, k, gfp); b 635 drivers/md/bcache/btree.c return b; b 638 drivers/md/bcache/btree.c static int mca_reap(struct btree *b, unsigned int min_order, bool flush) b 643 drivers/md/bcache/btree.c lockdep_assert_held(&b->c->bucket_lock); b 645 drivers/md/bcache/btree.c if (!down_write_trylock(&b->lock)) b 648 drivers/md/bcache/btree.c BUG_ON(btree_node_dirty(b) && !b->keys.set[0].data); b 650 drivers/md/bcache/btree.c if (b->keys.page_order < min_order) b 654 drivers/md/bcache/btree.c if (btree_node_dirty(b)) b 657 drivers/md/bcache/btree.c if (down_trylock(&b->io_mutex)) b 659 drivers/md/bcache/btree.c up(&b->io_mutex); b 668 drivers/md/bcache/btree.c mutex_lock(&b->write_lock); b 674 drivers/md/bcache/btree.c if (btree_node_journal_flush(b)) { b 675 drivers/md/bcache/btree.c pr_debug("bnode %p is flushing by journal, retry", b); b 676 drivers/md/bcache/btree.c mutex_unlock(&b->write_lock); b 681 drivers/md/bcache/btree.c if (btree_node_dirty(b)) b 682 drivers/md/bcache/btree.c __bch_btree_node_write(b, &cl); b 683 drivers/md/bcache/btree.c mutex_unlock(&b->write_lock); b 688 drivers/md/bcache/btree.c down(&b->io_mutex); b 689 drivers/md/bcache/btree.c up(&b->io_mutex); b 693 drivers/md/bcache/btree.c rw_unlock(true, b); b 701 drivers/md/bcache/btree.c struct btree *b, *t; b 732 drivers/md/bcache/btree.c list_for_each_entry_safe(b, t, &c->btree_cache_freeable, list) { b 737 drivers/md/bcache/btree.c !mca_reap(b, 0, false)) { b 738 drivers/md/bcache/btree.c mca_data_free(b); b 739 drivers/md/bcache/btree.c rw_unlock(true, b); b 749 drivers/md/bcache/btree.c b = list_first_entry(&c->btree_cache, struct btree, list); b 752 drivers/md/bcache/btree.c if (!b->accessed && b 753 drivers/md/bcache/btree.c !mca_reap(b, 0, false)) { b 754 drivers/md/bcache/btree.c mca_bucket_free(b); b 755 drivers/md/bcache/btree.c mca_data_free(b); b 756 drivers/md/bcache/btree.c rw_unlock(true, b); b 759 drivers/md/bcache/btree.c b->accessed = 0; b 782 drivers/md/bcache/btree.c struct btree *b; b 803 drivers/md/bcache/btree.c b = list_first_entry(&c->btree_cache, struct btree, list); b 810 drivers/md/bcache/btree.c if (btree_node_dirty(b)) { b 811 drivers/md/bcache/btree.c btree_complete_write(b, btree_current_write(b)); b 812 drivers/md/bcache/btree.c clear_bit(BTREE_NODE_dirty, &b->flags); b 814 drivers/md/bcache/btree.c mca_data_free(b); b 818 drivers/md/bcache/btree.c b = list_first_entry(&c->btree_cache_freed, b 820 drivers/md/bcache/btree.c list_del(&b->list); b 821 drivers/md/bcache/btree.c cancel_delayed_work_sync(&b->work); b 822 drivers/md/bcache/btree.c kfree(b); b 875 drivers/md/bcache/btree.c struct btree *b; b 878 drivers/md/bcache/btree.c hlist_for_each_entry_rcu(b, mca_hash(c, k), hash) b 879 drivers/md/bcache/btree.c if (PTR_HASH(c, &b->key) == PTR_HASH(c, k)) b 881 drivers/md/bcache/btree.c b = NULL; b 884 drivers/md/bcache/btree.c return b; b 905 drivers/md/bcache/btree.c struct btree *b; b 912 drivers/md/bcache/btree.c list_for_each_entry_reverse(b, &c->btree_cache, list) b 913 drivers/md/bcache/btree.c if (!mca_reap(b, btree_order(k), false)) b 914 drivers/md/bcache/btree.c return b; b 916 drivers/md/bcache/btree.c list_for_each_entry_reverse(b, &c->btree_cache, list) b 917 drivers/md/bcache/btree.c if (!mca_reap(b, btree_order(k), true)) b 918 drivers/md/bcache/btree.c return b; b 941 drivers/md/bcache/btree.c struct btree *b; b 953 drivers/md/bcache/btree.c list_for_each_entry(b, &c->btree_cache_freeable, list) b 954 drivers/md/bcache/btree.c if (!mca_reap(b, btree_order(k), false)) b 960 drivers/md/bcache/btree.c list_for_each_entry(b, &c->btree_cache_freed, list) b 961 drivers/md/bcache/btree.c if (!mca_reap(b, 0, false)) { b 962 drivers/md/bcache/btree.c mca_data_alloc(b, k, __GFP_NOWARN|GFP_NOIO); b 963 drivers/md/bcache/btree.c if (!b->keys.set[0].data) b 969 drivers/md/bcache/btree.c b = mca_bucket_alloc(c, k, __GFP_NOWARN|GFP_NOIO); b 970 drivers/md/bcache/btree.c if (!b) b 973 drivers/md/bcache/btree.c BUG_ON(!down_write_trylock(&b->lock)); b 974 drivers/md/bcache/btree.c if (!b->keys.set->data) b 977 drivers/md/bcache/btree.c BUG_ON(b->io_mutex.count != 1); b 979 drivers/md/bcache/btree.c bkey_copy(&b->key, k); b 980 drivers/md/bcache/btree.c list_move(&b->list, &c->btree_cache); b 981 drivers/md/bcache/btree.c hlist_del_init_rcu(&b->hash); b 982 drivers/md/bcache/btree.c hlist_add_head_rcu(&b->hash, mca_hash(c, k)); b 984 drivers/md/bcache/btree.c lock_set_subclass(&b->lock.dep_map, level + 1, _THIS_IP_); b 985 drivers/md/bcache/btree.c b->parent = (void *) ~0UL; b 986 drivers/md/bcache/btree.c b->flags = 0; b 987 drivers/md/bcache/btree.c b->written = 0; b 988 drivers/md/bcache/btree.c b->level = level; b 990 drivers/md/bcache/btree.c if (!b->level) b 991 drivers/md/bcache/btree.c bch_btree_keys_init(&b->keys, &bch_extent_keys_ops, b 992 drivers/md/bcache/btree.c &b->c->expensive_debug_checks); b 994 drivers/md/bcache/btree.c bch_btree_keys_init(&b->keys, &bch_btree_keys_ops, b 995 drivers/md/bcache/btree.c &b->c->expensive_debug_checks); b 997 drivers/md/bcache/btree.c return b; b 999 drivers/md/bcache/btree.c if (b) b 1000 drivers/md/bcache/btree.c rw_unlock(true, b); b 1002 drivers/md/bcache/btree.c b = mca_cannibalize(c, op, k); b 1003 drivers/md/bcache/btree.c if (!IS_ERR(b)) b 1006 drivers/md/bcache/btree.c return b; b 1023 drivers/md/bcache/btree.c struct btree *b; b 1027 drivers/md/bcache/btree.c b = mca_find(c, k); b 1029 drivers/md/bcache/btree.c if (!b) { b 1034 drivers/md/bcache/btree.c b = mca_alloc(c, op, k, level); b 1037 drivers/md/bcache/btree.c if (!b) b 1039 drivers/md/bcache/btree.c if (IS_ERR(b)) b 1040 drivers/md/bcache/btree.c return b; b 1042 drivers/md/bcache/btree.c bch_btree_node_read(b); b 1045 drivers/md/bcache/btree.c downgrade_write(&b->lock); b 1047 drivers/md/bcache/btree.c rw_lock(write, b, level); b 1048 drivers/md/bcache/btree.c if (PTR_HASH(c, &b->key) != PTR_HASH(c, k)) { b 1049 drivers/md/bcache/btree.c rw_unlock(write, b); b 1052 drivers/md/bcache/btree.c BUG_ON(b->level != level); b 1055 drivers/md/bcache/btree.c if (btree_node_io_error(b)) { b 1056 drivers/md/bcache/btree.c rw_unlock(write, b); b 1060 drivers/md/bcache/btree.c BUG_ON(!b->written); b 1062 drivers/md/bcache/btree.c b->parent = parent; b 1063 drivers/md/bcache/btree.c b->accessed = 1; b 1065 drivers/md/bcache/btree.c for (; i <= b->keys.nsets && b->keys.set[i].size; i++) { b 1066 drivers/md/bcache/btree.c prefetch(b->keys.set[i].tree); b 1067 drivers/md/bcache/btree.c prefetch(b->keys.set[i].data); b 1070 drivers/md/bcache/btree.c for (; i <= b->keys.nsets; i++) b 1071 drivers/md/bcache/btree.c prefetch(b->keys.set[i].data); b 1073 drivers/md/bcache/btree.c return b; b 1078 drivers/md/bcache/btree.c struct btree *b; b 1081 drivers/md/bcache/btree.c b = mca_alloc(parent->c, NULL, k, parent->level - 1); b 1084 drivers/md/bcache/btree.c if (!IS_ERR_OR_NULL(b)) { b 1085 drivers/md/bcache/btree.c b->parent = parent; b 1086 drivers/md/bcache/btree.c bch_btree_node_read(b); b 1087 drivers/md/bcache/btree.c rw_unlock(true, b); b 1093 drivers/md/bcache/btree.c static void btree_node_free(struct btree *b) b 1095 drivers/md/bcache/btree.c trace_bcache_btree_node_free(b); b 1097 drivers/md/bcache/btree.c BUG_ON(b == b->c->root); b 1100 drivers/md/bcache/btree.c mutex_lock(&b->write_lock); b 1107 drivers/md/bcache/btree.c if (btree_node_journal_flush(b)) { b 1108 drivers/md/bcache/btree.c mutex_unlock(&b->write_lock); b 1109 drivers/md/bcache/btree.c pr_debug("bnode %p journal_flush set, retry", b); b 1114 drivers/md/bcache/btree.c if (btree_node_dirty(b)) { b 1115 drivers/md/bcache/btree.c btree_complete_write(b, btree_current_write(b)); b 1116 drivers/md/bcache/btree.c clear_bit(BTREE_NODE_dirty, &b->flags); b 1119 drivers/md/bcache/btree.c mutex_unlock(&b->write_lock); b 1121 drivers/md/bcache/btree.c cancel_delayed_work(&b->work); b 1123 drivers/md/bcache/btree.c mutex_lock(&b->c->bucket_lock); b 1124 drivers/md/bcache/btree.c bch_bucket_free(b->c, &b->key); b 1125 drivers/md/bcache/btree.c mca_bucket_free(b); b 1126 drivers/md/bcache/btree.c mutex_unlock(&b->c->bucket_lock); b 1134 drivers/md/bcache/btree.c struct btree *b = ERR_PTR(-EAGAIN); b 1144 drivers/md/bcache/btree.c b = mca_alloc(c, op, &k.key, level); b 1145 drivers/md/bcache/btree.c if (IS_ERR(b)) b 1148 drivers/md/bcache/btree.c if (!b) { b 1154 drivers/md/bcache/btree.c b->accessed = 1; b 1155 drivers/md/bcache/btree.c b->parent = parent; b 1156 drivers/md/bcache/btree.c bch_bset_init_next(&b->keys, b->keys.set->data, bset_magic(&b->c->sb)); b 1160 drivers/md/bcache/btree.c trace_bcache_btree_node_alloc(b); b 1161 drivers/md/bcache/btree.c return b; b 1168 drivers/md/bcache/btree.c return b; b 1178 drivers/md/bcache/btree.c static struct btree *btree_node_alloc_replacement(struct btree *b, b 1181 drivers/md/bcache/btree.c struct btree *n = bch_btree_node_alloc(b->c, op, b->level, b->parent); b 1185 drivers/md/bcache/btree.c bch_btree_sort_into(&b->keys, &n->keys, &b->c->sort); b 1186 drivers/md/bcache/btree.c bkey_copy_key(&n->key, &b->key); b 1193 drivers/md/bcache/btree.c static void make_btree_freeing_key(struct btree *b, struct bkey *k) b 1197 drivers/md/bcache/btree.c mutex_lock(&b->c->bucket_lock); b 1199 drivers/md/bcache/btree.c atomic_inc(&b->c->prio_blocked); b 1201 drivers/md/bcache/btree.c bkey_copy(k, &b->key); b 1206 drivers/md/bcache/btree.c bch_inc_gen(PTR_CACHE(b->c, &b->key, i), b 1207 drivers/md/bcache/btree.c PTR_BUCKET(b->c, &b->key, i))); b 1209 drivers/md/bcache/btree.c mutex_unlock(&b->c->bucket_lock); b 1212 drivers/md/bcache/btree.c static int btree_check_reserve(struct btree *b, struct btree_op *op) b 1214 drivers/md/bcache/btree.c struct cache_set *c = b->c; b 1216 drivers/md/bcache/btree.c unsigned int i, reserve = (c->root->level - b->level) * 2 + 1; b 1231 drivers/md/bcache/btree.c return mca_cannibalize_lock(b->c, op); b 1288 drivers/md/bcache/btree.c #define btree_mark_key(b, k) __bch_btree_mark_key(b->c, b->level, k) b 1297 drivers/md/bcache/btree.c struct bucket *b = PTR_BUCKET(c, k, i); b 1299 drivers/md/bcache/btree.c b->gen = PTR_GEN(k, i); b 1302 drivers/md/bcache/btree.c b->prio = BTREE_PRIO; b 1303 drivers/md/bcache/btree.c else if (!level && b->prio == BTREE_PRIO) b 1304 drivers/md/bcache/btree.c b->prio = INITIAL_PRIO; b 1315 drivers/md/bcache/btree.c static bool btree_gc_mark_node(struct btree *b, struct gc_stat *gc) b 1325 drivers/md/bcache/btree.c for_each_key_filter(&b->keys, k, &iter, bch_ptr_invalid) { b 1326 drivers/md/bcache/btree.c stale = max(stale, btree_mark_key(b, k)); b 1329 drivers/md/bcache/btree.c if (bch_ptr_bad(&b->keys, k)) b 1339 drivers/md/bcache/btree.c for (t = b->keys.set; t <= &b->keys.set[b->keys.nsets]; t++) b 1341 drivers/md/bcache/btree.c bset_written(&b->keys, t) && b 1342 drivers/md/bcache/btree.c bkey_cmp(&b->key, &t->end) < 0, b 1343 drivers/md/bcache/btree.c b, "found short btree key in gc"); b 1345 drivers/md/bcache/btree.c if (b->c->gc_always_rewrite) b 1360 drivers/md/bcache/btree.c struct btree *b; b 1364 drivers/md/bcache/btree.c static int bch_btree_insert_node(struct btree *b, struct btree_op *op, b 1369 drivers/md/bcache/btree.c static int btree_gc_coalesce(struct btree *b, struct btree_op *op, b 1380 drivers/md/bcache/btree.c if (btree_check_reserve(b, NULL)) b 1386 drivers/md/bcache/btree.c while (nodes < GC_MERGE_NODES && !IS_ERR_OR_NULL(r[nodes].b)) b 1389 drivers/md/bcache/btree.c blocks = btree_default_blocks(b->c) * 2 / 3; b 1392 drivers/md/bcache/btree.c __set_blocks(b->keys.set[0].data, keys, b 1393 drivers/md/bcache/btree.c block_bytes(b->c)) > blocks * (nodes - 1)) b 1397 drivers/md/bcache/btree.c new_nodes[i] = btree_node_alloc_replacement(r[i].b, NULL); b 1408 drivers/md/bcache/btree.c if (btree_check_reserve(b, NULL)) b 1427 drivers/md/bcache/btree.c block_bytes(b->c)) > blocks) b 1443 drivers/md/bcache/btree.c block_bytes(b->c)) > b 1449 drivers/md/bcache/btree.c last = &r->b->key; b 1452 drivers/md/bcache/btree.c BUG_ON(__set_blocks(n1, n1->keys + keys, block_bytes(b->c)) > b 1492 drivers/md/bcache/btree.c if (__bch_keylist_realloc(&keylist, bkey_u64s(&r[i].b->key))) b 1495 drivers/md/bcache/btree.c make_btree_freeing_key(r[i].b, keylist.top); b 1499 drivers/md/bcache/btree.c bch_btree_insert_node(b, op, &keylist, NULL, NULL); b 1503 drivers/md/bcache/btree.c btree_node_free(r[i].b); b 1504 drivers/md/bcache/btree.c rw_unlock(true, r[i].b); b 1506 drivers/md/bcache/btree.c r[i].b = new_nodes[i]; b 1510 drivers/md/bcache/btree.c r[nodes - 1].b = ERR_PTR(-EINTR); b 1525 drivers/md/bcache/btree.c atomic_dec(&b->c->prio_blocked); b 1536 drivers/md/bcache/btree.c static int btree_gc_rewrite_node(struct btree *b, struct btree_op *op, b 1542 drivers/md/bcache/btree.c if (btree_check_reserve(b, NULL)) b 1548 drivers/md/bcache/btree.c if (btree_check_reserve(b, NULL)) { b 1562 drivers/md/bcache/btree.c bch_btree_insert_node(b, op, &keys, NULL, NULL); b 1572 drivers/md/bcache/btree.c static unsigned int btree_gc_count_keys(struct btree *b) b 1578 drivers/md/bcache/btree.c for_each_key_filter(&b->keys, k, &iter, bch_ptr_bad) b 1610 drivers/md/bcache/btree.c static int btree_gc_recurse(struct btree *b, struct btree_op *op, b 1620 drivers/md/bcache/btree.c bch_btree_iter_init(&b->keys, &iter, &b->c->gc_done); b 1623 drivers/md/bcache/btree.c i->b = ERR_PTR(-EINTR); b 1626 drivers/md/bcache/btree.c k = bch_btree_iter_next_filter(&iter, &b->keys, bch_ptr_bad); b 1628 drivers/md/bcache/btree.c r->b = bch_btree_node_get(b->c, op, k, b->level - 1, b 1629 drivers/md/bcache/btree.c true, b); b 1630 drivers/md/bcache/btree.c if (IS_ERR(r->b)) { b 1631 drivers/md/bcache/btree.c ret = PTR_ERR(r->b); b 1635 drivers/md/bcache/btree.c r->keys = btree_gc_count_keys(r->b); b 1637 drivers/md/bcache/btree.c ret = btree_gc_coalesce(b, op, gc, r); b 1642 drivers/md/bcache/btree.c if (!last->b) b 1645 drivers/md/bcache/btree.c if (!IS_ERR(last->b)) { b 1646 drivers/md/bcache/btree.c should_rewrite = btree_gc_mark_node(last->b, gc); b 1648 drivers/md/bcache/btree.c ret = btree_gc_rewrite_node(b, op, last->b); b 1653 drivers/md/bcache/btree.c if (last->b->level) { b 1654 drivers/md/bcache/btree.c ret = btree_gc_recurse(last->b, op, writes, gc); b 1659 drivers/md/bcache/btree.c bkey_copy_key(&b->c->gc_done, &last->b->key); b 1665 drivers/md/bcache/btree.c mutex_lock(&last->b->write_lock); b 1666 drivers/md/bcache/btree.c if (btree_node_dirty(last->b)) b 1667 drivers/md/bcache/btree.c bch_btree_node_write(last->b, writes); b 1668 drivers/md/bcache/btree.c mutex_unlock(&last->b->write_lock); b 1669 drivers/md/bcache/btree.c rw_unlock(true, last->b); b 1673 drivers/md/bcache/btree.c r->b = NULL; b 1675 drivers/md/bcache/btree.c if (atomic_read(&b->c->search_inflight) && b 1676 drivers/md/bcache/btree.c gc->nodes >= gc->nodes_pre + btree_gc_min_nodes(b->c)) { b 1689 drivers/md/bcache/btree.c if (!IS_ERR_OR_NULL(i->b)) { b 1690 drivers/md/bcache/btree.c mutex_lock(&i->b->write_lock); b 1691 drivers/md/bcache/btree.c if (btree_node_dirty(i->b)) b 1692 drivers/md/bcache/btree.c bch_btree_node_write(i->b, writes); b 1693 drivers/md/bcache/btree.c mutex_unlock(&i->b->write_lock); b 1694 drivers/md/bcache/btree.c rw_unlock(true, i->b); b 1700 drivers/md/bcache/btree.c static int bch_btree_gc_root(struct btree *b, struct btree_op *op, b 1707 drivers/md/bcache/btree.c should_rewrite = btree_gc_mark_node(b, gc); b 1709 drivers/md/bcache/btree.c n = btree_node_alloc_replacement(b, NULL); b 1715 drivers/md/bcache/btree.c btree_node_free(b); b 1722 drivers/md/bcache/btree.c __bch_btree_mark_key(b->c, b->level + 1, &b->key); b 1724 drivers/md/bcache/btree.c if (b->level) { b 1725 drivers/md/bcache/btree.c ret = btree_gc_recurse(b, op, writes, gc); b 1730 drivers/md/bcache/btree.c bkey_copy_key(&b->c->gc_done, &b->key); b 1738 drivers/md/bcache/btree.c struct bucket *b; b 1750 drivers/md/bcache/btree.c for_each_bucket(b, ca) { b 1751 drivers/md/bcache/btree.c b->last_gc = b->gen; b 1752 drivers/md/bcache/btree.c if (!atomic_read(&b->pin)) { b 1753 drivers/md/bcache/btree.c SET_GC_MARK(b, 0); b 1754 drivers/md/bcache/btree.c SET_GC_SECTORS_USED(b, 0); b 1763 drivers/md/bcache/btree.c struct bucket *b; b 1812 drivers/md/bcache/btree.c for_each_bucket(b, ca) { b 1813 drivers/md/bcache/btree.c c->need_gc = max(c->need_gc, bucket_gc_gen(b)); b 1815 drivers/md/bcache/btree.c if (atomic_read(&b->pin)) b 1818 drivers/md/bcache/btree.c BUG_ON(!GC_MARK(b) && GC_SECTORS_USED(b)); b 1820 drivers/md/bcache/btree.c if (!GC_MARK(b) || GC_MARK(b) == GC_MARK_RECLAIMABLE) b 1917 drivers/md/bcache/btree.c static int bch_btree_check_recurse(struct btree *b, struct btree_op *op) b 1923 drivers/md/bcache/btree.c for_each_key_filter(&b->keys, k, &iter, bch_ptr_invalid) b 1924 drivers/md/bcache/btree.c bch_initial_mark_key(b->c, b->level, k); b 1926 drivers/md/bcache/btree.c bch_initial_mark_key(b->c, b->level + 1, &b->key); b 1928 drivers/md/bcache/btree.c if (b->level) { b 1929 drivers/md/bcache/btree.c bch_btree_iter_init(&b->keys, &iter, NULL); b 1932 drivers/md/bcache/btree.c k = bch_btree_iter_next_filter(&iter, &b->keys, b 1935 drivers/md/bcache/btree.c btree_node_prefetch(b, k); b 1940 drivers/md/bcache/btree.c b->c->gc_stats.nodes++; b 1944 drivers/md/bcache/btree.c ret = btree(check_recurse, p, b, op); b 1965 drivers/md/bcache/btree.c struct bucket *b; b 1982 drivers/md/bcache/btree.c for_each_bucket(b, ca) { b 1987 drivers/md/bcache/btree.c if (bch_can_invalidate_bucket(ca, b) && b 1988 drivers/md/bcache/btree.c !GC_MARK(b)) { b 1989 drivers/md/bcache/btree.c __bch_invalidate_one_bucket(ca, b); b 1991 drivers/md/bcache/btree.c b - ca->buckets)) b 1993 drivers/md/bcache/btree.c b - ca->buckets); b 2003 drivers/md/bcache/btree.c static bool btree_insert_key(struct btree *b, struct bkey *k, b 2008 drivers/md/bcache/btree.c BUG_ON(bkey_cmp(k, &b->key) > 0); b 2010 drivers/md/bcache/btree.c status = bch_btree_insert_key(&b->keys, k, replace_key); b 2012 drivers/md/bcache/btree.c bch_check_keys(&b->keys, "%u for %s", status, b 2015 drivers/md/bcache/btree.c trace_bcache_btree_insert_key(b, k, replace_key != NULL, b 2022 drivers/md/bcache/btree.c static size_t insert_u64s_remaining(struct btree *b) b 2024 drivers/md/bcache/btree.c long ret = bch_btree_keys_u64s_remaining(&b->keys); b 2029 drivers/md/bcache/btree.c if (b->keys.ops->is_extents) b 2035 drivers/md/bcache/btree.c static bool bch_btree_insert_keys(struct btree *b, struct btree_op *op, b 2040 drivers/md/bcache/btree.c int oldsize = bch_count_data(&b->keys); b 2045 drivers/md/bcache/btree.c if (bkey_u64s(k) > insert_u64s_remaining(b)) b 2048 drivers/md/bcache/btree.c if (bkey_cmp(k, &b->key) <= 0) { b 2049 drivers/md/bcache/btree.c if (!b->level) b 2050 drivers/md/bcache/btree.c bkey_put(b->c, k); b 2052 drivers/md/bcache/btree.c ret |= btree_insert_key(b, k, replace_key); b 2054 drivers/md/bcache/btree.c } else if (bkey_cmp(&START_KEY(k), &b->key) < 0) { b 2058 drivers/md/bcache/btree.c bch_cut_back(&b->key, &temp.key); b 2059 drivers/md/bcache/btree.c bch_cut_front(&b->key, insert_keys->keys); b 2061 drivers/md/bcache/btree.c ret |= btree_insert_key(b, &temp.key, replace_key); b 2071 drivers/md/bcache/btree.c BUG_ON(!bch_keylist_empty(insert_keys) && b->level); b 2073 drivers/md/bcache/btree.c BUG_ON(bch_count_data(&b->keys) < oldsize); b 2077 drivers/md/bcache/btree.c static int btree_split(struct btree *b, struct btree_op *op, b 2090 drivers/md/bcache/btree.c if (btree_check_reserve(b, op)) { b 2091 drivers/md/bcache/btree.c if (!b->level) b 2097 drivers/md/bcache/btree.c n1 = btree_node_alloc_replacement(b, op); b 2102 drivers/md/bcache/btree.c block_bytes(n1->c)) > (btree_blocks(b) * 4) / 5; b 2107 drivers/md/bcache/btree.c trace_bcache_btree_node_split(b, btree_bset_first(n1)->keys); b 2109 drivers/md/bcache/btree.c n2 = bch_btree_node_alloc(b->c, op, b->level, b->parent); b 2113 drivers/md/bcache/btree.c if (!b->parent) { b 2114 drivers/md/bcache/btree.c n3 = bch_btree_node_alloc(b->c, op, b->level + 1, NULL); b 2144 drivers/md/bcache/btree.c bkey_copy_key(&n2->key, &b->key); b 2151 drivers/md/bcache/btree.c trace_bcache_btree_node_compact(b, btree_bset_first(n1)->keys); b 2172 drivers/md/bcache/btree.c } else if (!b->parent) { b 2179 drivers/md/bcache/btree.c make_btree_freeing_key(b, parent_keys.top); b 2182 drivers/md/bcache/btree.c bch_btree_insert_node(b->parent, op, &parent_keys, NULL, NULL); b 2186 drivers/md/bcache/btree.c btree_node_free(b); b 2189 drivers/md/bcache/btree.c bch_time_stats_update(&b->c->btree_split_time, start_time); b 2193 drivers/md/bcache/btree.c bkey_put(b->c, &n2->key); b 2197 drivers/md/bcache/btree.c bkey_put(b->c, &n1->key); b 2201 drivers/md/bcache/btree.c WARN(1, "bcache: btree split failed (level %u)", b->level); b 2211 drivers/md/bcache/btree.c static int bch_btree_insert_node(struct btree *b, struct btree_op *op, b 2218 drivers/md/bcache/btree.c BUG_ON(b->level && replace_key); b 2222 drivers/md/bcache/btree.c mutex_lock(&b->write_lock); b 2224 drivers/md/bcache/btree.c if (write_block(b) != btree_bset_last(b) && b 2225 drivers/md/bcache/btree.c b->keys.last_set_unwritten) b 2226 drivers/md/bcache/btree.c bch_btree_init_next(b); /* just wrote a set */ b 2228 drivers/md/bcache/btree.c if (bch_keylist_nkeys(insert_keys) > insert_u64s_remaining(b)) { b 2229 drivers/md/bcache/btree.c mutex_unlock(&b->write_lock); b 2233 drivers/md/bcache/btree.c BUG_ON(write_block(b) != btree_bset_last(b)); b 2235 drivers/md/bcache/btree.c if (bch_btree_insert_keys(b, op, insert_keys, replace_key)) { b 2236 drivers/md/bcache/btree.c if (!b->level) b 2237 drivers/md/bcache/btree.c bch_btree_leaf_dirty(b, journal_ref); b 2239 drivers/md/bcache/btree.c bch_btree_node_write(b, &cl); b 2242 drivers/md/bcache/btree.c mutex_unlock(&b->write_lock); b 2250 drivers/md/bcache/btree.c op->lock = b->c->root->level + 1; b 2252 drivers/md/bcache/btree.c } else if (op->lock <= b->c->root->level) { b 2253 drivers/md/bcache/btree.c op->lock = b->c->root->level + 1; b 2257 drivers/md/bcache/btree.c int ret = btree_split(b, op, insert_keys, replace_key); b 2267 drivers/md/bcache/btree.c int bch_btree_insert_check_key(struct btree *b, struct btree_op *op, b 2271 drivers/md/bcache/btree.c uint64_t btree_ptr = b->key.ptr[0]; b 2272 drivers/md/bcache/btree.c unsigned long seq = b->seq; b 2279 drivers/md/bcache/btree.c rw_unlock(false, b); b 2280 drivers/md/bcache/btree.c rw_lock(true, b, b->level); b 2282 drivers/md/bcache/btree.c if (b->key.ptr[0] != btree_ptr || b 2283 drivers/md/bcache/btree.c b->seq != seq + 1) { b 2284 drivers/md/bcache/btree.c op->lock = b->level; b 2296 drivers/md/bcache/btree.c ret = bch_btree_insert_node(b, op, &insert, NULL, NULL); b 2301 drivers/md/bcache/btree.c downgrade_write(&b->lock); b 2312 drivers/md/bcache/btree.c static int btree_insert_fn(struct btree_op *b_op, struct btree *b) b 2317 drivers/md/bcache/btree.c int ret = bch_btree_insert_node(b, &op->op, op->keys, b 2359 drivers/md/bcache/btree.c void bch_btree_set_root(struct btree *b) b 2366 drivers/md/bcache/btree.c trace_bcache_btree_set_root(b); b 2368 drivers/md/bcache/btree.c BUG_ON(!b->written); b 2370 drivers/md/bcache/btree.c for (i = 0; i < KEY_PTRS(&b->key); i++) b 2371 drivers/md/bcache/btree.c BUG_ON(PTR_BUCKET(b->c, &b->key, i)->prio != BTREE_PRIO); b 2373 drivers/md/bcache/btree.c mutex_lock(&b->c->bucket_lock); b 2374 drivers/md/bcache/btree.c list_del_init(&b->list); b 2375 drivers/md/bcache/btree.c mutex_unlock(&b->c->bucket_lock); b 2377 drivers/md/bcache/btree.c b->c->root = b; b 2379 drivers/md/bcache/btree.c bch_journal_meta(b->c, &cl); b 2385 drivers/md/bcache/btree.c static int bch_btree_map_nodes_recurse(struct btree *b, struct btree_op *op, b 2391 drivers/md/bcache/btree.c if (b->level) { b 2395 drivers/md/bcache/btree.c bch_btree_iter_init(&b->keys, &iter, from); b 2397 drivers/md/bcache/btree.c while ((k = bch_btree_iter_next_filter(&iter, &b->keys, b 2399 drivers/md/bcache/btree.c ret = btree(map_nodes_recurse, k, b, b 2408 drivers/md/bcache/btree.c if (!b->level || flags == MAP_ALL_NODES) b 2409 drivers/md/bcache/btree.c ret = fn(op, b); b 2420 drivers/md/bcache/btree.c static int bch_btree_map_keys_recurse(struct btree *b, struct btree_op *op, b 2428 drivers/md/bcache/btree.c bch_btree_iter_init(&b->keys, &iter, from); b 2430 drivers/md/bcache/btree.c while ((k = bch_btree_iter_next_filter(&iter, &b->keys, bch_ptr_bad))) { b 2431 drivers/md/bcache/btree.c ret = !b->level b 2432 drivers/md/bcache/btree.c ? fn(op, b, k) b 2433 drivers/md/bcache/btree.c : btree(map_keys_recurse, k, b, op, from, fn, flags); b 2440 drivers/md/bcache/btree.c if (!b->level && (flags & MAP_END_KEY)) b 2441 drivers/md/bcache/btree.c ret = fn(op, b, &KEY(KEY_INODE(&b->key), b 2442 drivers/md/bcache/btree.c KEY_OFFSET(&b->key), 0)); b 2479 drivers/md/bcache/btree.c static int refill_keybuf_fn(struct btree_op *op, struct btree *b, b 151 drivers/md/bcache/btree.h static inline bool btree_node_ ## flag(struct btree *b) \ b 152 drivers/md/bcache/btree.h { return test_bit(BTREE_NODE_ ## flag, &b->flags); } \ b 154 drivers/md/bcache/btree.h static inline void set_btree_node_ ## flag(struct btree *b) \ b 155 drivers/md/bcache/btree.h { set_bit(BTREE_NODE_ ## flag, &b->flags); } b 169 drivers/md/bcache/btree.h static inline struct btree_write *btree_current_write(struct btree *b) b 171 drivers/md/bcache/btree.h return b->writes + btree_node_write_idx(b); b 174 drivers/md/bcache/btree.h static inline struct btree_write *btree_prev_write(struct btree *b) b 176 drivers/md/bcache/btree.h return b->writes + (btree_node_write_idx(b) ^ 1); b 179 drivers/md/bcache/btree.h static inline struct bset *btree_bset_first(struct btree *b) b 181 drivers/md/bcache/btree.h return b->keys.set->data; b 184 drivers/md/bcache/btree.h static inline struct bset *btree_bset_last(struct btree *b) b 186 drivers/md/bcache/btree.h return bset_tree_last(&b->keys)->data; b 189 drivers/md/bcache/btree.h static inline unsigned int bset_block_offset(struct btree *b, struct bset *i) b 191 drivers/md/bcache/btree.h return bset_sector_offset(&b->keys, i) >> b->c->block_bits; b 203 drivers/md/bcache/btree.h #define for_each_cached_btree(b, c, iter) \ b 207 drivers/md/bcache/btree.h hlist_for_each_entry_rcu((b), (c)->bucket_hash + iter, hash) b 228 drivers/md/bcache/btree.h static inline void rw_lock(bool w, struct btree *b, int level) b 230 drivers/md/bcache/btree.h w ? down_write_nested(&b->lock, level + 1) b 231 drivers/md/bcache/btree.h : down_read_nested(&b->lock, level + 1); b 233 drivers/md/bcache/btree.h b->seq++; b 236 drivers/md/bcache/btree.h static inline void rw_unlock(bool w, struct btree *b) b 239 drivers/md/bcache/btree.h b->seq++; b 240 drivers/md/bcache/btree.h (w ? up_write : up_read)(&b->lock); b 243 drivers/md/bcache/btree.h void bch_btree_node_read_done(struct btree *b); b 244 drivers/md/bcache/btree.h void __bch_btree_node_write(struct btree *b, struct closure *parent); b 245 drivers/md/bcache/btree.h void bch_btree_node_write(struct btree *b, struct closure *parent); b 247 drivers/md/bcache/btree.h void bch_btree_set_root(struct btree *b); b 255 drivers/md/bcache/btree.h int bch_btree_insert_check_key(struct btree *b, struct btree_op *op, b 297 drivers/md/bcache/btree.h typedef int (btree_map_nodes_fn)(struct btree_op *b_op, struct btree *b); b 315 drivers/md/bcache/btree.h typedef int (btree_map_keys_fn)(struct btree_op *op, struct btree *b, b 24 drivers/md/bcache/debug.c #define for_each_written_bset(b, start, i) \ b 26 drivers/md/bcache/debug.c (void *) i < (void *) (start) + (KEY_SIZE(&b->key) << 9) &&\ b 28 drivers/md/bcache/debug.c i = (void *) i + set_blocks(i, block_bytes(b->c)) * \ b 29 drivers/md/bcache/debug.c block_bytes(b->c)) b 31 drivers/md/bcache/debug.c void bch_btree_verify(struct btree *b) b 33 drivers/md/bcache/debug.c struct btree *v = b->c->verify_data; b 37 drivers/md/bcache/debug.c if (!b->c->verify || !b->c->verify_ondisk) b 40 drivers/md/bcache/debug.c down(&b->io_mutex); b 41 drivers/md/bcache/debug.c mutex_lock(&b->c->verify_lock); b 43 drivers/md/bcache/debug.c ondisk = b->c->verify_ondisk; b 44 drivers/md/bcache/debug.c sorted = b->c->verify_data->keys.set->data; b 45 drivers/md/bcache/debug.c inmemory = b->keys.set->data; b 47 drivers/md/bcache/debug.c bkey_copy(&v->key, &b->key); b 49 drivers/md/bcache/debug.c v->level = b->level; b 50 drivers/md/bcache/debug.c v->keys.ops = b->keys.ops; b 52 drivers/md/bcache/debug.c bio = bch_bbio_alloc(b->c); b 53 drivers/md/bcache/debug.c bio_set_dev(bio, PTR_CACHE(b->c, &b->key, 0)->bdev); b 54 drivers/md/bcache/debug.c bio->bi_iter.bi_sector = PTR_OFFSET(&b->key, 0); b 60 drivers/md/bcache/debug.c bch_bbio_free(bio, b->c); b 78 drivers/md/bcache/debug.c bch_dump_bset(&b->keys, inmemory, 0); b 83 drivers/md/bcache/debug.c for_each_written_bset(b, ondisk, i) { b 85 drivers/md/bcache/debug.c block_bytes(b->c); b 88 drivers/md/bcache/debug.c bch_dump_bset(&b->keys, i, block); b 92 drivers/md/bcache/debug.c ((void *) i - (void *) ondisk) / block_bytes(b->c)); b 98 drivers/md/bcache/debug.c pr_err("b->written %u\n", b->written); b 104 drivers/md/bcache/debug.c mutex_unlock(&b->c->verify_lock); b 105 drivers/md/bcache/debug.c up(&b->io_mutex); b 11 drivers/md/bcache/debug.h void bch_btree_verify(struct btree *b); b 20 drivers/md/bcache/debug.h static inline void bch_btree_verify(struct btree *b) {} b 128 drivers/md/bcache/extents.c struct btree *b = container_of(keys, struct btree, keys); b 136 drivers/md/bcache/extents.c size_t n = PTR_BUCKET_NR(b->c, k, j); b 139 drivers/md/bcache/extents.c if (n >= b->c->sb.first_bucket && n < b->c->sb.nbuckets) b 141 drivers/md/bcache/extents.c PTR_BUCKET(b->c, k, j)->prio); b 144 drivers/md/bcache/extents.c pr_err(" %s\n", bch_ptr_status(b->c, k)); b 168 drivers/md/bcache/extents.c struct btree *b = container_of(bk, struct btree, keys); b 170 drivers/md/bcache/extents.c return __bch_btree_ptr_invalid(b->c, k); b 173 drivers/md/bcache/extents.c static bool btree_ptr_bad_expensive(struct btree *b, const struct bkey *k) b 179 drivers/md/bcache/extents.c if (mutex_trylock(&b->c->bucket_lock)) { b 181 drivers/md/bcache/extents.c if (ptr_available(b->c, k, i)) { b 182 drivers/md/bcache/extents.c g = PTR_BUCKET(b->c, k, i); b 186 drivers/md/bcache/extents.c (b->c->gc_mark_valid && b 191 drivers/md/bcache/extents.c mutex_unlock(&b->c->bucket_lock); b 196 drivers/md/bcache/extents.c mutex_unlock(&b->c->bucket_lock); b 198 drivers/md/bcache/extents.c btree_bug(b, b 200 drivers/md/bcache/extents.c buf, PTR_BUCKET_NR(b->c, k, i), atomic_read(&g->pin), b 207 drivers/md/bcache/extents.c struct btree *b = container_of(bk, struct btree, keys); b 216 drivers/md/bcache/extents.c if (!ptr_available(b->c, k, i) || b 217 drivers/md/bcache/extents.c ptr_stale(b->c, k, i)) b 220 drivers/md/bcache/extents.c if (expensive_debug_checks(b->c) && b 221 drivers/md/bcache/extents.c btree_ptr_bad_expensive(b, k)) b 232 drivers/md/bcache/extents.c struct btree *b = container_of(bk, struct btree, keys); b 235 drivers/md/bcache/extents.c btree_current_write(b)->prio_blocked++; b 323 drivers/md/bcache/extents.c static bool bch_extent_insert_fixup(struct btree_keys *b, b 328 drivers/md/bcache/extents.c struct cache_set *c = container_of(b, struct btree, keys)->c; b 410 drivers/md/bcache/extents.c if (bkey_written(b, k)) { b 423 drivers/md/bcache/extents.c top = bch_bset_search(b, bset_tree_last(b), b 425 drivers/md/bcache/extents.c bch_bset_insert(b, top, k); b 429 drivers/md/bcache/extents.c bch_bset_insert(b, k, &temp.key); b 435 drivers/md/bcache/extents.c bch_bset_fix_invalidated_key(b, k); b 445 drivers/md/bcache/extents.c if (bkey_written(b, k) && b 454 drivers/md/bcache/extents.c bch_bset_fix_invalidated_key(b, k); b 502 drivers/md/bcache/extents.c struct btree *b = container_of(bk, struct btree, keys); b 504 drivers/md/bcache/extents.c return __bch_extent_invalid(b->c, k); b 507 drivers/md/bcache/extents.c static bool bch_extent_bad_expensive(struct btree *b, const struct bkey *k, b 510 drivers/md/bcache/extents.c struct bucket *g = PTR_BUCKET(b->c, k, ptr); b 513 drivers/md/bcache/extents.c if (mutex_trylock(&b->c->bucket_lock)) { b 514 drivers/md/bcache/extents.c if (b->c->gc_mark_valid && b 523 drivers/md/bcache/extents.c mutex_unlock(&b->c->bucket_lock); b 528 drivers/md/bcache/extents.c mutex_unlock(&b->c->bucket_lock); b 530 drivers/md/bcache/extents.c btree_bug(b, b 532 drivers/md/bcache/extents.c buf, PTR_BUCKET_NR(b->c, k, ptr), atomic_read(&g->pin), b 539 drivers/md/bcache/extents.c struct btree *b = container_of(bk, struct btree, keys); b 548 drivers/md/bcache/extents.c if (!ptr_available(b->c, k, i)) b 552 drivers/md/bcache/extents.c stale = ptr_stale(b->c, k, i); b 560 drivers/md/bcache/extents.c btree_bug_on(stale > BUCKET_GC_GEN_MAX, b, b 562 drivers/md/bcache/extents.c stale, b->c->need_gc); b 567 drivers/md/bcache/extents.c if (expensive_debug_checks(b->c) && b 568 drivers/md/bcache/extents.c bch_extent_bad_expensive(b, k, i)) b 585 drivers/md/bcache/extents.c struct btree *b = container_of(bk, struct btree, keys); b 588 drivers/md/bcache/extents.c if (key_merging_disabled(b->c)) b 593 drivers/md/bcache/extents.c PTR_BUCKET_NR(b->c, l, i) != PTR_BUCKET_NR(b->c, r, i)) b 19 drivers/md/bcache/io.c struct bbio *b = container_of(bio, struct bbio, bio); b 21 drivers/md/bcache/io.c mempool_free(b, &c->bio_meta); b 26 drivers/md/bcache/io.c struct bbio *b = mempool_alloc(&c->bio_meta, GFP_NOIO); b 27 drivers/md/bcache/io.c struct bio *bio = &b->bio; b 36 drivers/md/bcache/io.c struct bbio *b = container_of(bio, struct bbio, bio); b 38 drivers/md/bcache/io.c bio->bi_iter.bi_sector = PTR_OFFSET(&b->key, 0); b 39 drivers/md/bcache/io.c bio_set_dev(bio, PTR_CACHE(c, &b->key, 0)->bdev); b 41 drivers/md/bcache/io.c b->submit_time_us = local_clock_us(); b 48 drivers/md/bcache/io.c struct bbio *b = container_of(bio, struct bbio, bio); b 50 drivers/md/bcache/io.c bch_bkey_copy_single_ptr(&b->key, k, ptr); b 139 drivers/md/bcache/io.c struct bbio *b = container_of(bio, struct bbio, bio); b 140 drivers/md/bcache/io.c struct cache *ca = PTR_CACHE(c, &b->key, 0); b 149 drivers/md/bcache/io.c int us = t - b->submit_time_us; b 173 drivers/md/bcache/journal.c #define read_bucket(b) \ b 175 drivers/md/bcache/journal.c ret = journal_read_bucket(ca, list, b); \ b 176 drivers/md/bcache/journal.c __set_bit(b, bitmap); \ b 424 drivers/md/bcache/journal.c struct btree *b, *t, *btree_nodes[BTREE_FLUSH_NR]; b 461 drivers/md/bcache/journal.c list_for_each_entry_safe_reverse(b, t, &c->btree_cache, list) { b 485 drivers/md/bcache/journal.c if (btree_node_journal_flush(b)) b 488 drivers/md/bcache/journal.c mutex_lock(&b->write_lock); b 490 drivers/md/bcache/journal.c if (!btree_node_dirty(b)) { b 491 drivers/md/bcache/journal.c mutex_unlock(&b->write_lock); b 495 drivers/md/bcache/journal.c if (!btree_current_write(b)->journal) { b 496 drivers/md/bcache/journal.c mutex_unlock(&b->write_lock); b 513 drivers/md/bcache/journal.c if (nr_to_fifo_front(btree_current_write(b)->journal, b 516 drivers/md/bcache/journal.c mutex_unlock(&b->write_lock); b 520 drivers/md/bcache/journal.c set_btree_node_journal_flush(b); b 522 drivers/md/bcache/journal.c mutex_unlock(&b->write_lock); b 524 drivers/md/bcache/journal.c btree_nodes[nr++] = b; b 538 drivers/md/bcache/journal.c b = btree_nodes[i]; b 539 drivers/md/bcache/journal.c if (!b) { b 545 drivers/md/bcache/journal.c if (!btree_node_journal_flush(b)) { b 546 drivers/md/bcache/journal.c pr_err("BUG: bnode %p: journal_flush bit cleaned", b); b 550 drivers/md/bcache/journal.c mutex_lock(&b->write_lock); b 551 drivers/md/bcache/journal.c if (!btree_current_write(b)->journal) { b 552 drivers/md/bcache/journal.c clear_bit(BTREE_NODE_journal_flush, &b->flags); b 553 drivers/md/bcache/journal.c mutex_unlock(&b->write_lock); b 554 drivers/md/bcache/journal.c pr_debug("bnode %p: written by others", b); b 558 drivers/md/bcache/journal.c if (!btree_node_dirty(b)) { b 559 drivers/md/bcache/journal.c clear_bit(BTREE_NODE_journal_flush, &b->flags); b 560 drivers/md/bcache/journal.c mutex_unlock(&b->write_lock); b 561 drivers/md/bcache/journal.c pr_debug("bnode %p: dirty bit cleaned by others", b); b 565 drivers/md/bcache/journal.c __bch_btree_node_write(b, NULL); b 566 drivers/md/bcache/journal.c clear_bit(BTREE_NODE_journal_flush, &b->flags); b 567 drivers/md/bcache/journal.c mutex_unlock(&b->write_lock); b 64 drivers/md/bcache/movinggc.c struct bbio *b = container_of(bio, struct bbio, bio); b 70 drivers/md/bcache/movinggc.c else if (!KEY_DIRTY(&b->key) && b 71 drivers/md/bcache/movinggc.c ptr_stale(io->op.c, &b->key, 0)) { b 192 drivers/md/bcache/movinggc.c struct bucket *b; b 194 drivers/md/bcache/movinggc.c return (b = heap_peek(&ca->heap)) ? GC_SECTORS_USED(b) : 0; b 200 drivers/md/bcache/movinggc.c struct bucket *b; b 215 drivers/md/bcache/movinggc.c for_each_bucket(b, ca) { b 216 drivers/md/bcache/movinggc.c if (GC_MARK(b) == GC_MARK_METADATA || b 217 drivers/md/bcache/movinggc.c !GC_SECTORS_USED(b) || b 218 drivers/md/bcache/movinggc.c GC_SECTORS_USED(b) == ca->sb.bucket_size || b 219 drivers/md/bcache/movinggc.c atomic_read(&b->pin)) b 223 drivers/md/bcache/movinggc.c sectors_to_move += GC_SECTORS_USED(b); b 224 drivers/md/bcache/movinggc.c heap_add(&ca->heap, b, bucket_cmp); b 225 drivers/md/bcache/movinggc.c } else if (bucket_cmp(b, heap_peek(&ca->heap))) { b 227 drivers/md/bcache/movinggc.c sectors_to_move += GC_SECTORS_USED(b); b 229 drivers/md/bcache/movinggc.c ca->heap.data[0] = b; b 235 drivers/md/bcache/movinggc.c heap_pop(&ca->heap, b, bucket_cmp); b 236 drivers/md/bcache/movinggc.c sectors_to_move -= GC_SECTORS_USED(b); b 239 drivers/md/bcache/movinggc.c while (heap_pop(&ca->heap, b, bucket_cmp)) b 240 drivers/md/bcache/movinggc.c SET_GC_MOVE(b, 1); b 498 drivers/md/bcache/request.c struct bbio *b = container_of(bio, struct bbio, bio); b 511 drivers/md/bcache/request.c else if (!KEY_DIRTY(&b->key) && b 512 drivers/md/bcache/request.c ptr_stale(s->iop.c, &b->key, 0)) { b 524 drivers/md/bcache/request.c static int cache_lookup_fn(struct btree_op *op, struct btree *b, struct bkey *k) b 541 drivers/md/bcache/request.c int ret = s->d->cache_miss(b, s, bio, sectors); b 556 drivers/md/bcache/request.c PTR_BUCKET(b->c, k, ptr)->prio = INITIAL_PRIO; b 585 drivers/md/bcache/request.c __bch_submit_bbio(n, b->c); b 888 drivers/md/bcache/request.c static int cached_dev_cache_miss(struct btree *b, struct search *s, b 916 drivers/md/bcache/request.c ret = bch_btree_insert_check_key(b, &s->op, &s->iop.replace_key); b 1289 drivers/md/bcache/request.c static int flash_dev_cache_miss(struct btree *b, struct search *s, b 535 drivers/md/bcache/super.c struct bucket *b; b 570 drivers/md/bcache/super.c for (b = ca->buckets + i * prios_per_bucket(ca); b 571 drivers/md/bcache/super.c b < ca->buckets + ca->sb.nbuckets && d < end; b 572 drivers/md/bcache/super.c b++, d++) { b 573 drivers/md/bcache/super.c d->prio = cpu_to_le16(b->prio); b 574 drivers/md/bcache/super.c d->gen = b->gen; b 617 drivers/md/bcache/super.c struct bucket *b; b 620 drivers/md/bcache/super.c for (b = ca->buckets; b 621 drivers/md/bcache/super.c b < ca->buckets + ca->sb.nbuckets; b 622 drivers/md/bcache/super.c b++, d++) { b 641 drivers/md/bcache/super.c b->prio = le16_to_cpu(d->prio); b 642 drivers/md/bcache/super.c b->gen = b->last_gc = d->gen; b 648 drivers/md/bcache/super.c static int open_dev(struct block_device *b, fmode_t mode) b 650 drivers/md/bcache/super.c struct bcache_device *d = b->bd_disk->private_data; b 659 drivers/md/bcache/super.c static void release_dev(struct gendisk *b, fmode_t mode) b 661 drivers/md/bcache/super.c struct bcache_device *d = b->private_data; b 666 drivers/md/bcache/super.c static int ioctl_dev(struct block_device *b, fmode_t mode, b 669 drivers/md/bcache/super.c struct bcache_device *d = b->bd_disk->private_data; b 1615 drivers/md/bcache/super.c struct btree *b; b 1634 drivers/md/bcache/super.c list_for_each_entry(b, &c->btree_cache, list) { b 1635 drivers/md/bcache/super.c mutex_lock(&b->write_lock); b 1636 drivers/md/bcache/super.c if (btree_node_dirty(b)) b 1637 drivers/md/bcache/super.c __bch_btree_node_write(b, NULL); b 1638 drivers/md/bcache/super.c mutex_unlock(&b->write_lock); b 2153 drivers/md/bcache/super.c struct bucket *b; b 2234 drivers/md/bcache/super.c for_each_bucket(b, ca) b 2235 drivers/md/bcache/super.c atomic_set(&b->pin, 0); b 596 drivers/md/bcache/sysfs.c static int bch_btree_bset_stats(struct btree_op *b_op, struct btree *b) b 601 drivers/md/bcache/sysfs.c bch_btree_keys_stats(&b->keys, &op->stats); b 636 drivers/md/bcache/sysfs.c struct btree *b; b 642 drivers/md/bcache/sysfs.c rw_unlock(false, b); b 644 drivers/md/bcache/sysfs.c b = c->root; b 645 drivers/md/bcache/sysfs.c rw_lock(false, b, b->level); b 646 drivers/md/bcache/sysfs.c } while (b != c->root); b 648 drivers/md/bcache/sysfs.c for_each_key_filter(&b->keys, k, &iter, bch_ptr_bad) b 651 drivers/md/bcache/sysfs.c rw_unlock(false, b); b 659 drivers/md/bcache/sysfs.c struct btree *b; b 662 drivers/md/bcache/sysfs.c list_for_each_entry(b, &c->btree_cache, list) b 663 drivers/md/bcache/sysfs.c ret += 1 << (b->keys.page_order + PAGE_SHIFT); b 1017 drivers/md/bcache/sysfs.c struct bucket *b; b 1031 drivers/md/bcache/sysfs.c for_each_bucket(b, ca) { b 1032 drivers/md/bcache/sysfs.c if (!GC_SECTORS_USED(b)) b 1034 drivers/md/bcache/sysfs.c if (GC_MARK(b) == GC_MARK_RECLAIMABLE) b 1036 drivers/md/bcache/sysfs.c if (GC_MARK(b) == GC_MARK_DIRTY) b 1038 drivers/md/bcache/sysfs.c if (GC_MARK(b) == GC_MARK_METADATA) b 762 drivers/md/bcache/writeback.c static int sectors_dirty_init_fn(struct btree_op *_op, struct btree *b, b 771 drivers/md/bcache/writeback.c bcache_dev_sectors_dirty_add(b->c, KEY_INODE(k), b 775 drivers/md/bcache/writeback.c if (atomic_read(&b->c->search_inflight) && b 239 drivers/md/dm-bufio.c static void buffer_record_stack(struct dm_buffer *b) b 241 drivers/md/dm-bufio.c b->stack_len = stack_trace_save(b->stack_entries, MAX_STACK, 2); b 251 drivers/md/dm-bufio.c struct dm_buffer *b; b 254 drivers/md/dm-bufio.c b = container_of(n, struct dm_buffer, node); b 256 drivers/md/dm-bufio.c if (b->block == block) b 257 drivers/md/dm-bufio.c return b; b 259 drivers/md/dm-bufio.c n = (b->block < block) ? n->rb_left : n->rb_right; b 265 drivers/md/dm-bufio.c static void __insert(struct dm_bufio_client *c, struct dm_buffer *b) b 273 drivers/md/dm-bufio.c if (found->block == b->block) { b 274 drivers/md/dm-bufio.c BUG_ON(found != b); b 279 drivers/md/dm-bufio.c new = (found->block < b->block) ? b 283 drivers/md/dm-bufio.c rb_link_node(&b->node, parent, new); b 284 drivers/md/dm-bufio.c rb_insert_color(&b->node, &c->buffer_tree); b 287 drivers/md/dm-bufio.c static void __remove(struct dm_bufio_client *c, struct dm_buffer *b) b 289 drivers/md/dm-bufio.c rb_erase(&b->node, &c->buffer_tree); b 294 drivers/md/dm-bufio.c static void adjust_total_allocated(struct dm_buffer *b, bool unlink) b 305 drivers/md/dm-bufio.c data_mode = b->data_mode; b 306 drivers/md/dm-bufio.c diff = (long)b->c->block_size; b 319 drivers/md/dm-bufio.c b->accessed = 1; b 322 drivers/md/dm-bufio.c list_add(&b->global_list, &global_queue); b 327 drivers/md/dm-bufio.c list_del(&b->global_list); b 444 drivers/md/dm-bufio.c struct dm_buffer *b = kmem_cache_alloc(c->slab_buffer, gfp_mask); b 446 drivers/md/dm-bufio.c if (!b) b 449 drivers/md/dm-bufio.c b->c = c; b 451 drivers/md/dm-bufio.c b->data = alloc_buffer_data(c, gfp_mask, &b->data_mode); b 452 drivers/md/dm-bufio.c if (!b->data) { b 453 drivers/md/dm-bufio.c kmem_cache_free(c->slab_buffer, b); b 458 drivers/md/dm-bufio.c b->stack_len = 0; b 460 drivers/md/dm-bufio.c return b; b 466 drivers/md/dm-bufio.c static void free_buffer(struct dm_buffer *b) b 468 drivers/md/dm-bufio.c struct dm_bufio_client *c = b->c; b 470 drivers/md/dm-bufio.c free_buffer_data(c, b->data, b->data_mode); b 471 drivers/md/dm-bufio.c kmem_cache_free(c->slab_buffer, b); b 477 drivers/md/dm-bufio.c static void __link_buffer(struct dm_buffer *b, sector_t block, int dirty) b 479 drivers/md/dm-bufio.c struct dm_bufio_client *c = b->c; b 482 drivers/md/dm-bufio.c b->block = block; b 483 drivers/md/dm-bufio.c b->list_mode = dirty; b 484 drivers/md/dm-bufio.c list_add(&b->lru_list, &c->lru[dirty]); b 485 drivers/md/dm-bufio.c __insert(b->c, b); b 486 drivers/md/dm-bufio.c b->last_accessed = jiffies; b 488 drivers/md/dm-bufio.c adjust_total_allocated(b, false); b 494 drivers/md/dm-bufio.c static void __unlink_buffer(struct dm_buffer *b) b 496 drivers/md/dm-bufio.c struct dm_bufio_client *c = b->c; b 498 drivers/md/dm-bufio.c BUG_ON(!c->n_buffers[b->list_mode]); b 500 drivers/md/dm-bufio.c c->n_buffers[b->list_mode]--; b 501 drivers/md/dm-bufio.c __remove(b->c, b); b 502 drivers/md/dm-bufio.c list_del(&b->lru_list); b 504 drivers/md/dm-bufio.c adjust_total_allocated(b, true); b 510 drivers/md/dm-bufio.c static void __relink_lru(struct dm_buffer *b, int dirty) b 512 drivers/md/dm-bufio.c struct dm_bufio_client *c = b->c; b 514 drivers/md/dm-bufio.c b->accessed = 1; b 516 drivers/md/dm-bufio.c BUG_ON(!c->n_buffers[b->list_mode]); b 518 drivers/md/dm-bufio.c c->n_buffers[b->list_mode]--; b 520 drivers/md/dm-bufio.c b->list_mode = dirty; b 521 drivers/md/dm-bufio.c list_move(&b->lru_list, &c->lru[dirty]); b 522 drivers/md/dm-bufio.c b->last_accessed = jiffies; b 549 drivers/md/dm-bufio.c struct dm_buffer *b = context; b 551 drivers/md/dm-bufio.c b->end_io(b, unlikely(error != 0) ? BLK_STS_IOERR : 0); b 554 drivers/md/dm-bufio.c static void use_dmio(struct dm_buffer *b, int rw, sector_t sector, b 562 drivers/md/dm-bufio.c .notify.context = b, b 563 drivers/md/dm-bufio.c .client = b->c->dm_io, b 566 drivers/md/dm-bufio.c .bdev = b->c->bdev, b 571 drivers/md/dm-bufio.c if (b->data_mode != DATA_MODE_VMALLOC) { b 573 drivers/md/dm-bufio.c io_req.mem.ptr.addr = (char *)b->data + offset; b 576 drivers/md/dm-bufio.c io_req.mem.ptr.vma = (char *)b->data + offset; b 581 drivers/md/dm-bufio.c b->end_io(b, errno_to_blk_status(r)); b 586 drivers/md/dm-bufio.c struct dm_buffer *b = bio->bi_private; b 589 drivers/md/dm-bufio.c b->end_io(b, status); b 592 drivers/md/dm-bufio.c static void use_bio(struct dm_buffer *b, int rw, sector_t sector, b 599 drivers/md/dm-bufio.c vec_size = b->c->block_size >> PAGE_SHIFT; b 600 drivers/md/dm-bufio.c if (unlikely(b->c->sectors_per_block_bits < PAGE_SHIFT - SECTOR_SHIFT)) b 606 drivers/md/dm-bufio.c use_dmio(b, rw, sector, n_sectors, offset); b 611 drivers/md/dm-bufio.c bio_set_dev(bio, b->c->bdev); b 614 drivers/md/dm-bufio.c bio->bi_private = b; b 616 drivers/md/dm-bufio.c ptr = (char *)b->data + offset; b 634 drivers/md/dm-bufio.c static void submit_io(struct dm_buffer *b, int rw, void (*end_io)(struct dm_buffer *, blk_status_t)) b 640 drivers/md/dm-bufio.c b->end_io = end_io; b 642 drivers/md/dm-bufio.c if (likely(b->c->sectors_per_block_bits >= 0)) b 643 drivers/md/dm-bufio.c sector = b->block << b->c->sectors_per_block_bits; b 645 drivers/md/dm-bufio.c sector = b->block * (b->c->block_size >> SECTOR_SHIFT); b 646 drivers/md/dm-bufio.c sector += b->c->start; b 649 drivers/md/dm-bufio.c n_sectors = b->c->block_size >> SECTOR_SHIFT; b 652 drivers/md/dm-bufio.c if (b->c->write_callback) b 653 drivers/md/dm-bufio.c b->c->write_callback(b); b 654 drivers/md/dm-bufio.c offset = b->write_start; b 655 drivers/md/dm-bufio.c end = b->write_end; b 659 drivers/md/dm-bufio.c if (unlikely(end > b->c->block_size)) b 660 drivers/md/dm-bufio.c end = b->c->block_size; b 666 drivers/md/dm-bufio.c if (b->data_mode != DATA_MODE_VMALLOC) b 667 drivers/md/dm-bufio.c use_bio(b, rw, sector, n_sectors, offset); b 669 drivers/md/dm-bufio.c use_dmio(b, rw, sector, n_sectors, offset); b 682 drivers/md/dm-bufio.c static void write_endio(struct dm_buffer *b, blk_status_t status) b 684 drivers/md/dm-bufio.c b->write_error = status; b 686 drivers/md/dm-bufio.c struct dm_bufio_client *c = b->c; b 692 drivers/md/dm-bufio.c BUG_ON(!test_bit(B_WRITING, &b->state)); b 695 drivers/md/dm-bufio.c clear_bit(B_WRITING, &b->state); b 698 drivers/md/dm-bufio.c wake_up_bit(&b->state, B_WRITING); b 710 drivers/md/dm-bufio.c static void __write_dirty_buffer(struct dm_buffer *b, b 713 drivers/md/dm-bufio.c if (!test_bit(B_DIRTY, &b->state)) b 716 drivers/md/dm-bufio.c clear_bit(B_DIRTY, &b->state); b 717 drivers/md/dm-bufio.c wait_on_bit_lock_io(&b->state, B_WRITING, TASK_UNINTERRUPTIBLE); b 719 drivers/md/dm-bufio.c b->write_start = b->dirty_start; b 720 drivers/md/dm-bufio.c b->write_end = b->dirty_end; b 723 drivers/md/dm-bufio.c submit_io(b, REQ_OP_WRITE, write_endio); b 725 drivers/md/dm-bufio.c list_add_tail(&b->write_list, write_list); b 733 drivers/md/dm-bufio.c struct dm_buffer *b = b 735 drivers/md/dm-bufio.c list_del(&b->write_list); b 736 drivers/md/dm-bufio.c submit_io(b, REQ_OP_WRITE, write_endio); b 747 drivers/md/dm-bufio.c static void __make_buffer_clean(struct dm_buffer *b) b 749 drivers/md/dm-bufio.c BUG_ON(b->hold_count); b 751 drivers/md/dm-bufio.c if (!b->state) /* fast case */ b 754 drivers/md/dm-bufio.c wait_on_bit_io(&b->state, B_READING, TASK_UNINTERRUPTIBLE); b 755 drivers/md/dm-bufio.c __write_dirty_buffer(b, NULL); b 756 drivers/md/dm-bufio.c wait_on_bit_io(&b->state, B_WRITING, TASK_UNINTERRUPTIBLE); b 765 drivers/md/dm-bufio.c struct dm_buffer *b; b 767 drivers/md/dm-bufio.c list_for_each_entry_reverse(b, &c->lru[LIST_CLEAN], lru_list) { b 768 drivers/md/dm-bufio.c BUG_ON(test_bit(B_WRITING, &b->state)); b 769 drivers/md/dm-bufio.c BUG_ON(test_bit(B_DIRTY, &b->state)); b 771 drivers/md/dm-bufio.c if (!b->hold_count) { b 772 drivers/md/dm-bufio.c __make_buffer_clean(b); b 773 drivers/md/dm-bufio.c __unlink_buffer(b); b 774 drivers/md/dm-bufio.c return b; b 779 drivers/md/dm-bufio.c list_for_each_entry_reverse(b, &c->lru[LIST_DIRTY], lru_list) { b 780 drivers/md/dm-bufio.c BUG_ON(test_bit(B_READING, &b->state)); b 782 drivers/md/dm-bufio.c if (!b->hold_count) { b 783 drivers/md/dm-bufio.c __make_buffer_clean(b); b 784 drivers/md/dm-bufio.c __unlink_buffer(b); b 785 drivers/md/dm-bufio.c return b; b 830 drivers/md/dm-bufio.c struct dm_buffer *b; b 848 drivers/md/dm-bufio.c b = alloc_buffer(c, GFP_NOWAIT | __GFP_NORETRY | __GFP_NOMEMALLOC | __GFP_NOWARN); b 849 drivers/md/dm-bufio.c if (b) b 850 drivers/md/dm-bufio.c return b; b 858 drivers/md/dm-bufio.c b = alloc_buffer(c, GFP_NOIO | __GFP_NORETRY | __GFP_NOMEMALLOC | __GFP_NOWARN); b 860 drivers/md/dm-bufio.c if (b) b 861 drivers/md/dm-bufio.c return b; b 866 drivers/md/dm-bufio.c b = list_entry(c->reserved_buffers.next, b 868 drivers/md/dm-bufio.c list_del(&b->lru_list); b 871 drivers/md/dm-bufio.c return b; b 874 drivers/md/dm-bufio.c b = __get_unclaimed_buffer(c); b 875 drivers/md/dm-bufio.c if (b) b 876 drivers/md/dm-bufio.c return b; b 884 drivers/md/dm-bufio.c struct dm_buffer *b = __alloc_buffer_wait_no_callback(c, nf); b 886 drivers/md/dm-bufio.c if (!b) b 890 drivers/md/dm-bufio.c c->alloc_callback(b); b 892 drivers/md/dm-bufio.c return b; b 898 drivers/md/dm-bufio.c static void __free_buffer_wake(struct dm_buffer *b) b 900 drivers/md/dm-bufio.c struct dm_bufio_client *c = b->c; b 903 drivers/md/dm-bufio.c free_buffer(b); b 905 drivers/md/dm-bufio.c list_add(&b->lru_list, &c->reserved_buffers); b 915 drivers/md/dm-bufio.c struct dm_buffer *b, *tmp; b 917 drivers/md/dm-bufio.c list_for_each_entry_safe_reverse(b, tmp, &c->lru[LIST_DIRTY], lru_list) { b 918 drivers/md/dm-bufio.c BUG_ON(test_bit(B_READING, &b->state)); b 920 drivers/md/dm-bufio.c if (!test_bit(B_DIRTY, &b->state) && b 921 drivers/md/dm-bufio.c !test_bit(B_WRITING, &b->state)) { b 922 drivers/md/dm-bufio.c __relink_lru(b, LIST_CLEAN); b 926 drivers/md/dm-bufio.c if (no_wait && test_bit(B_WRITING, &b->state)) b 929 drivers/md/dm-bufio.c __write_dirty_buffer(b, write_list); b 954 drivers/md/dm-bufio.c struct dm_buffer *b, *new_b = NULL; b 958 drivers/md/dm-bufio.c b = __find(c, block); b 959 drivers/md/dm-bufio.c if (b) b 973 drivers/md/dm-bufio.c b = __find(c, block); b 974 drivers/md/dm-bufio.c if (b) { b 981 drivers/md/dm-bufio.c b = new_b; b 982 drivers/md/dm-bufio.c b->hold_count = 1; b 983 drivers/md/dm-bufio.c b->read_error = 0; b 984 drivers/md/dm-bufio.c b->write_error = 0; b 985 drivers/md/dm-bufio.c __link_buffer(b, block, LIST_CLEAN); b 988 drivers/md/dm-bufio.c b->state = 0; b 989 drivers/md/dm-bufio.c return b; b 992 drivers/md/dm-bufio.c b->state = 1 << B_READING; b 995 drivers/md/dm-bufio.c return b; b 1007 drivers/md/dm-bufio.c if (nf == NF_GET && unlikely(test_bit(B_READING, &b->state))) b 1010 drivers/md/dm-bufio.c b->hold_count++; b 1011 drivers/md/dm-bufio.c __relink_lru(b, test_bit(B_DIRTY, &b->state) || b 1012 drivers/md/dm-bufio.c test_bit(B_WRITING, &b->state)); b 1013 drivers/md/dm-bufio.c return b; b 1020 drivers/md/dm-bufio.c static void read_endio(struct dm_buffer *b, blk_status_t status) b 1022 drivers/md/dm-bufio.c b->read_error = status; b 1024 drivers/md/dm-bufio.c BUG_ON(!test_bit(B_READING, &b->state)); b 1027 drivers/md/dm-bufio.c clear_bit(B_READING, &b->state); b 1030 drivers/md/dm-bufio.c wake_up_bit(&b->state, B_READING); b 1043 drivers/md/dm-bufio.c struct dm_buffer *b; b 1048 drivers/md/dm-bufio.c b = __bufio_new(c, block, nf, &need_submit, &write_list); b 1050 drivers/md/dm-bufio.c if (b && b->hold_count == 1) b 1051 drivers/md/dm-bufio.c buffer_record_stack(b); b 1057 drivers/md/dm-bufio.c if (!b) b 1061 drivers/md/dm-bufio.c submit_io(b, REQ_OP_READ, read_endio); b 1063 drivers/md/dm-bufio.c wait_on_bit_io(&b->state, B_READING, TASK_UNINTERRUPTIBLE); b 1065 drivers/md/dm-bufio.c if (b->read_error) { b 1066 drivers/md/dm-bufio.c int error = blk_status_to_errno(b->read_error); b 1068 drivers/md/dm-bufio.c dm_bufio_release(b); b 1073 drivers/md/dm-bufio.c *bp = b; b 1075 drivers/md/dm-bufio.c return b->data; b 1117 drivers/md/dm-bufio.c struct dm_buffer *b; b 1118 drivers/md/dm-bufio.c b = __bufio_new(c, block, NF_PREFETCH, &need_submit, b 1127 drivers/md/dm-bufio.c if (unlikely(b != NULL)) { b 1131 drivers/md/dm-bufio.c submit_io(b, REQ_OP_READ, read_endio); b 1132 drivers/md/dm-bufio.c dm_bufio_release(b); b 1149 drivers/md/dm-bufio.c void dm_bufio_release(struct dm_buffer *b) b 1151 drivers/md/dm-bufio.c struct dm_bufio_client *c = b->c; b 1155 drivers/md/dm-bufio.c BUG_ON(!b->hold_count); b 1157 drivers/md/dm-bufio.c b->hold_count--; b 1158 drivers/md/dm-bufio.c if (!b->hold_count) { b 1166 drivers/md/dm-bufio.c if ((b->read_error || b->write_error) && b 1167 drivers/md/dm-bufio.c !test_bit(B_READING, &b->state) && b 1168 drivers/md/dm-bufio.c !test_bit(B_WRITING, &b->state) && b 1169 drivers/md/dm-bufio.c !test_bit(B_DIRTY, &b->state)) { b 1170 drivers/md/dm-bufio.c __unlink_buffer(b); b 1171 drivers/md/dm-bufio.c __free_buffer_wake(b); b 1179 drivers/md/dm-bufio.c void dm_bufio_mark_partial_buffer_dirty(struct dm_buffer *b, b 1182 drivers/md/dm-bufio.c struct dm_bufio_client *c = b->c; b 1185 drivers/md/dm-bufio.c BUG_ON(end > b->c->block_size); b 1189 drivers/md/dm-bufio.c BUG_ON(test_bit(B_READING, &b->state)); b 1191 drivers/md/dm-bufio.c if (!test_and_set_bit(B_DIRTY, &b->state)) { b 1192 drivers/md/dm-bufio.c b->dirty_start = start; b 1193 drivers/md/dm-bufio.c b->dirty_end = end; b 1194 drivers/md/dm-bufio.c __relink_lru(b, LIST_DIRTY); b 1196 drivers/md/dm-bufio.c if (start < b->dirty_start) b 1197 drivers/md/dm-bufio.c b->dirty_start = start; b 1198 drivers/md/dm-bufio.c if (end > b->dirty_end) b 1199 drivers/md/dm-bufio.c b->dirty_end = end; b 1206 drivers/md/dm-bufio.c void dm_bufio_mark_buffer_dirty(struct dm_buffer *b) b 1208 drivers/md/dm-bufio.c dm_bufio_mark_partial_buffer_dirty(b, 0, b->c->block_size); b 1236 drivers/md/dm-bufio.c struct dm_buffer *b, *tmp; b 1247 drivers/md/dm-bufio.c list_for_each_entry_safe_reverse(b, tmp, &c->lru[LIST_DIRTY], lru_list) { b 1253 drivers/md/dm-bufio.c BUG_ON(test_bit(B_READING, &b->state)); b 1255 drivers/md/dm-bufio.c if (test_bit(B_WRITING, &b->state)) { b 1258 drivers/md/dm-bufio.c b->hold_count++; b 1260 drivers/md/dm-bufio.c wait_on_bit_io(&b->state, B_WRITING, b 1263 drivers/md/dm-bufio.c b->hold_count--; b 1265 drivers/md/dm-bufio.c wait_on_bit_io(&b->state, B_WRITING, b 1269 drivers/md/dm-bufio.c if (!test_bit(B_DIRTY, &b->state) && b 1270 drivers/md/dm-bufio.c !test_bit(B_WRITING, &b->state)) b 1271 drivers/md/dm-bufio.c __relink_lru(b, LIST_CLEAN); b 1340 drivers/md/dm-bufio.c void dm_bufio_release_move(struct dm_buffer *b, sector_t new_block) b 1342 drivers/md/dm-bufio.c struct dm_bufio_client *c = b->c; b 1366 drivers/md/dm-bufio.c BUG_ON(!b->hold_count); b 1367 drivers/md/dm-bufio.c BUG_ON(test_bit(B_READING, &b->state)); b 1369 drivers/md/dm-bufio.c __write_dirty_buffer(b, NULL); b 1370 drivers/md/dm-bufio.c if (b->hold_count == 1) { b 1371 drivers/md/dm-bufio.c wait_on_bit_io(&b->state, B_WRITING, b 1373 drivers/md/dm-bufio.c set_bit(B_DIRTY, &b->state); b 1374 drivers/md/dm-bufio.c b->dirty_start = 0; b 1375 drivers/md/dm-bufio.c b->dirty_end = c->block_size; b 1376 drivers/md/dm-bufio.c __unlink_buffer(b); b 1377 drivers/md/dm-bufio.c __link_buffer(b, new_block, LIST_DIRTY); b 1380 drivers/md/dm-bufio.c wait_on_bit_lock_io(&b->state, B_WRITING, b 1389 drivers/md/dm-bufio.c old_block = b->block; b 1390 drivers/md/dm-bufio.c __unlink_buffer(b); b 1391 drivers/md/dm-bufio.c __link_buffer(b, new_block, b->list_mode); b 1392 drivers/md/dm-bufio.c submit_io(b, REQ_OP_WRITE, write_endio); b 1393 drivers/md/dm-bufio.c wait_on_bit_io(&b->state, B_WRITING, b 1395 drivers/md/dm-bufio.c __unlink_buffer(b); b 1396 drivers/md/dm-bufio.c __link_buffer(b, old_block, b->list_mode); b 1400 drivers/md/dm-bufio.c dm_bufio_release(b); b 1412 drivers/md/dm-bufio.c struct dm_buffer *b; b 1416 drivers/md/dm-bufio.c b = __find(c, block); b 1417 drivers/md/dm-bufio.c if (b && likely(!b->hold_count) && likely(!b->state)) { b 1418 drivers/md/dm-bufio.c __unlink_buffer(b); b 1419 drivers/md/dm-bufio.c __free_buffer_wake(b); b 1449 drivers/md/dm-bufio.c sector_t dm_bufio_get_block_number(struct dm_buffer *b) b 1451 drivers/md/dm-bufio.c return b->block; b 1455 drivers/md/dm-bufio.c void *dm_bufio_get_block_data(struct dm_buffer *b) b 1457 drivers/md/dm-bufio.c return b->data; b 1461 drivers/md/dm-bufio.c void *dm_bufio_get_aux_data(struct dm_buffer *b) b 1463 drivers/md/dm-bufio.c return b + 1; b 1467 drivers/md/dm-bufio.c struct dm_bufio_client *dm_bufio_get_client(struct dm_buffer *b) b 1469 drivers/md/dm-bufio.c return b->c; b 1475 drivers/md/dm-bufio.c struct dm_buffer *b; b 1488 drivers/md/dm-bufio.c while ((b = __get_unclaimed_buffer(c))) b 1489 drivers/md/dm-bufio.c __free_buffer_wake(b); b 1492 drivers/md/dm-bufio.c list_for_each_entry(b, &c->lru[i], lru_list) { b 1496 drivers/md/dm-bufio.c (unsigned long long)b->block, b->hold_count, i); b 1498 drivers/md/dm-bufio.c stack_trace_print(b->stack_entries, b->stack_len, 1); b 1500 drivers/md/dm-bufio.c b->hold_count = 0; b 1505 drivers/md/dm-bufio.c while ((b = __get_unclaimed_buffer(c))) b 1506 drivers/md/dm-bufio.c __free_buffer_wake(b); b 1523 drivers/md/dm-bufio.c static bool __try_evict_buffer(struct dm_buffer *b, gfp_t gfp) b 1526 drivers/md/dm-bufio.c if (test_bit(B_READING, &b->state) || b 1527 drivers/md/dm-bufio.c test_bit(B_WRITING, &b->state) || b 1528 drivers/md/dm-bufio.c test_bit(B_DIRTY, &b->state)) b 1532 drivers/md/dm-bufio.c if (b->hold_count) b 1535 drivers/md/dm-bufio.c __make_buffer_clean(b); b 1536 drivers/md/dm-bufio.c __unlink_buffer(b); b 1537 drivers/md/dm-bufio.c __free_buffer_wake(b); b 1556 drivers/md/dm-bufio.c struct dm_buffer *b, *tmp; b 1563 drivers/md/dm-bufio.c list_for_each_entry_safe_reverse(b, tmp, &c->lru[l], lru_list) { b 1564 drivers/md/dm-bufio.c if (__try_evict_buffer(b, gfp_mask)) b 1681 drivers/md/dm-bufio.c struct dm_buffer *b = alloc_buffer(c, GFP_KERNEL); b 1683 drivers/md/dm-bufio.c if (!b) { b 1687 drivers/md/dm-bufio.c __free_buffer_wake(b); b 1708 drivers/md/dm-bufio.c struct dm_buffer *b = list_entry(c->reserved_buffers.next, b 1710 drivers/md/dm-bufio.c list_del(&b->lru_list); b 1711 drivers/md/dm-bufio.c free_buffer(b); b 1748 drivers/md/dm-bufio.c struct dm_buffer *b = list_entry(c->reserved_buffers.next, b 1750 drivers/md/dm-bufio.c list_del(&b->lru_list); b 1751 drivers/md/dm-bufio.c free_buffer(b); b 1785 drivers/md/dm-bufio.c static bool older_than(struct dm_buffer *b, unsigned long age_hz) b 1787 drivers/md/dm-bufio.c return time_after_eq(jiffies, b->last_accessed + age_hz); b 1792 drivers/md/dm-bufio.c struct dm_buffer *b, *tmp; b 1807 drivers/md/dm-bufio.c list_for_each_entry_safe_reverse(b, tmp, &c->lru[LIST_CLEAN], lru_list) { b 1811 drivers/md/dm-bufio.c if (!older_than(b, age_hz)) b 1814 drivers/md/dm-bufio.c if (__try_evict_buffer(b, 0)) b 1827 drivers/md/dm-bufio.c struct dm_buffer *b; b 1848 drivers/md/dm-bufio.c b = list_entry(global_queue.prev, struct dm_buffer, global_list); b 1850 drivers/md/dm-bufio.c if (b->accessed) { b 1851 drivers/md/dm-bufio.c b->accessed = 0; b 1852 drivers/md/dm-bufio.c list_move(&b->global_list, &global_queue); b 1859 drivers/md/dm-bufio.c current_client = b->c; b 1876 drivers/md/dm-bufio.c if (unlikely(!__try_evict_buffer(b, GFP_KERNEL))) { b 1878 drivers/md/dm-bufio.c list_move(&b->global_list, &global_queue); b 34 drivers/md/dm-cache-background-tracker.c struct background_tracker *b = kmalloc(sizeof(*b), GFP_KERNEL); b 36 drivers/md/dm-cache-background-tracker.c if (!b) { b 41 drivers/md/dm-cache-background-tracker.c b->max_work = max_work; b 42 drivers/md/dm-cache-background-tracker.c atomic_set(&b->pending_promotes, 0); b 43 drivers/md/dm-cache-background-tracker.c atomic_set(&b->pending_writebacks, 0); b 44 drivers/md/dm-cache-background-tracker.c atomic_set(&b->pending_demotes, 0); b 46 drivers/md/dm-cache-background-tracker.c INIT_LIST_HEAD(&b->issued); b 47 drivers/md/dm-cache-background-tracker.c INIT_LIST_HEAD(&b->queued); b 49 drivers/md/dm-cache-background-tracker.c b->pending = RB_ROOT; b 50 drivers/md/dm-cache-background-tracker.c b->work_cache = KMEM_CACHE(bt_work, 0); b 51 drivers/md/dm-cache-background-tracker.c if (!b->work_cache) { b 53 drivers/md/dm-cache-background-tracker.c kfree(b); b 54 drivers/md/dm-cache-background-tracker.c b = NULL; b 57 drivers/md/dm-cache-background-tracker.c return b; b 61 drivers/md/dm-cache-background-tracker.c void btracker_destroy(struct background_tracker *b) b 63 drivers/md/dm-cache-background-tracker.c kmem_cache_destroy(b->work_cache); b 64 drivers/md/dm-cache-background-tracker.c kfree(b); b 79 drivers/md/dm-cache-background-tracker.c static bool __insert_pending(struct background_tracker *b, b 84 drivers/md/dm-cache-background-tracker.c struct rb_node **new = &b->pending.rb_node, *parent = NULL; b 103 drivers/md/dm-cache-background-tracker.c rb_insert_color(&nw->node, &b->pending); b 108 drivers/md/dm-cache-background-tracker.c static struct bt_work *__find_pending(struct background_tracker *b, b 113 drivers/md/dm-cache-background-tracker.c struct rb_node **new = &b->pending.rb_node; b 133 drivers/md/dm-cache-background-tracker.c static void update_stats(struct background_tracker *b, struct policy_work *w, int delta) b 137 drivers/md/dm-cache-background-tracker.c atomic_add(delta, &b->pending_promotes); b 141 drivers/md/dm-cache-background-tracker.c atomic_add(delta, &b->pending_demotes); b 145 drivers/md/dm-cache-background-tracker.c atomic_add(delta, &b->pending_writebacks); b 150 drivers/md/dm-cache-background-tracker.c unsigned btracker_nr_writebacks_queued(struct background_tracker *b) b 152 drivers/md/dm-cache-background-tracker.c return atomic_read(&b->pending_writebacks); b 156 drivers/md/dm-cache-background-tracker.c unsigned btracker_nr_demotions_queued(struct background_tracker *b) b 158 drivers/md/dm-cache-background-tracker.c return atomic_read(&b->pending_demotes); b 162 drivers/md/dm-cache-background-tracker.c static bool max_work_reached(struct background_tracker *b) b 164 drivers/md/dm-cache-background-tracker.c return atomic_read(&b->pending_promotes) + b 165 drivers/md/dm-cache-background-tracker.c atomic_read(&b->pending_writebacks) + b 166 drivers/md/dm-cache-background-tracker.c atomic_read(&b->pending_demotes) >= b->max_work; b 169 drivers/md/dm-cache-background-tracker.c static struct bt_work *alloc_work(struct background_tracker *b) b 171 drivers/md/dm-cache-background-tracker.c if (max_work_reached(b)) b 174 drivers/md/dm-cache-background-tracker.c return kmem_cache_alloc(b->work_cache, GFP_NOWAIT); b 177 drivers/md/dm-cache-background-tracker.c int btracker_queue(struct background_tracker *b, b 186 drivers/md/dm-cache-background-tracker.c w = alloc_work(b); b 192 drivers/md/dm-cache-background-tracker.c if (!__insert_pending(b, w)) { b 197 drivers/md/dm-cache-background-tracker.c kmem_cache_free(b->work_cache, w); b 203 drivers/md/dm-cache-background-tracker.c list_add(&w->list, &b->issued); b 205 drivers/md/dm-cache-background-tracker.c list_add(&w->list, &b->queued); b 206 drivers/md/dm-cache-background-tracker.c update_stats(b, &w->work, 1); b 215 drivers/md/dm-cache-background-tracker.c int btracker_issue(struct background_tracker *b, struct policy_work **work) b 219 drivers/md/dm-cache-background-tracker.c if (list_empty(&b->queued)) b 222 drivers/md/dm-cache-background-tracker.c w = list_first_entry(&b->queued, struct bt_work, list); b 223 drivers/md/dm-cache-background-tracker.c list_move(&w->list, &b->issued); b 230 drivers/md/dm-cache-background-tracker.c void btracker_complete(struct background_tracker *b, b 235 drivers/md/dm-cache-background-tracker.c update_stats(b, &w->work, -1); b 236 drivers/md/dm-cache-background-tracker.c rb_erase(&w->node, &b->pending); b 238 drivers/md/dm-cache-background-tracker.c kmem_cache_free(b->work_cache, w); b 242 drivers/md/dm-cache-background-tracker.c bool btracker_promotion_already_present(struct background_tracker *b, b 245 drivers/md/dm-cache-background-tracker.c return __find_pending(b, oblock) != NULL; b 22 drivers/md/dm-cache-background-tracker.h void btracker_destroy(struct background_tracker *b); b 24 drivers/md/dm-cache-background-tracker.h unsigned btracker_nr_writebacks_queued(struct background_tracker *b); b 25 drivers/md/dm-cache-background-tracker.h unsigned btracker_nr_demotions_queued(struct background_tracker *b); b 31 drivers/md/dm-cache-background-tracker.h int btracker_queue(struct background_tracker *b, b 38 drivers/md/dm-cache-background-tracker.h int btracker_issue(struct background_tracker *b, struct policy_work **work); b 39 drivers/md/dm-cache-background-tracker.h void btracker_complete(struct background_tracker *b, b 41 drivers/md/dm-cache-background-tracker.h bool btracker_promotion_already_present(struct background_tracker *b, b 24 drivers/md/dm-cache-block-types.h static inline dm_oblock_t to_oblock(dm_block_t b) b 26 drivers/md/dm-cache-block-types.h return (__force dm_oblock_t) b; b 29 drivers/md/dm-cache-block-types.h static inline dm_block_t from_oblock(dm_oblock_t b) b 31 drivers/md/dm-cache-block-types.h return (__force dm_block_t) b; b 34 drivers/md/dm-cache-block-types.h static inline dm_cblock_t to_cblock(uint32_t b) b 36 drivers/md/dm-cache-block-types.h return (__force dm_cblock_t) b; b 39 drivers/md/dm-cache-block-types.h static inline uint32_t from_cblock(dm_cblock_t b) b 41 drivers/md/dm-cache-block-types.h return (__force uint32_t) b; b 44 drivers/md/dm-cache-block-types.h static inline dm_dblock_t to_dblock(dm_block_t b) b 46 drivers/md/dm-cache-block-types.h return (__force dm_dblock_t) b; b 49 drivers/md/dm-cache-block-types.h static inline dm_block_t from_dblock(dm_dblock_t b) b 51 drivers/md/dm-cache-block-types.h return (__force dm_block_t) b; b 172 drivers/md/dm-cache-metadata.c struct dm_block *b, b 175 drivers/md/dm-cache-metadata.c struct cache_disk_superblock *disk_super = dm_block_data(b); b 177 drivers/md/dm-cache-metadata.c disk_super->blocknr = cpu_to_le64(dm_block_location(b)); b 197 drivers/md/dm-cache-metadata.c struct dm_block *b, b 200 drivers/md/dm-cache-metadata.c struct cache_disk_superblock *disk_super = dm_block_data(b); b 203 drivers/md/dm-cache-metadata.c if (dm_block_location(b) != le64_to_cpu(disk_super->blocknr)) { b 206 drivers/md/dm-cache-metadata.c (unsigned long long)dm_block_location(b)); b 264 drivers/md/dm-cache-metadata.c struct dm_block *b; b 271 drivers/md/dm-cache-metadata.c r = dm_bm_read_lock(bm, CACHE_SUPERBLOCK_LOCATION, NULL, &b); b 275 drivers/md/dm-cache-metadata.c data_le = dm_block_data(b); b 284 drivers/md/dm-cache-metadata.c dm_bm_unlock(b); b 735 drivers/md/dm-cache-metadata.c uint64_t b = value >> 16; b 736 drivers/md/dm-cache-metadata.c *block = to_oblock(b); b 881 drivers/md/dm-cache-metadata.c static int block_clean_combined_dirty(struct dm_cache_metadata *cmd, dm_cblock_t b, b 889 drivers/md/dm-cache-metadata.c r = dm_array_get_value(&cmd->info, cmd->root, from_cblock(b), &value); b 1112 drivers/md/dm-cache-metadata.c static int __set_discard(struct dm_cache_metadata *cmd, dm_dblock_t b) b 1115 drivers/md/dm-cache-metadata.c from_dblock(b), &cmd->discard_root); b 1118 drivers/md/dm-cache-metadata.c static int __clear_discard(struct dm_cache_metadata *cmd, dm_dblock_t b) b 1121 drivers/md/dm-cache-metadata.c from_dblock(b), &cmd->discard_root); b 1153 drivers/md/dm-cache-metadata.c uint32_t b; b 1170 drivers/md/dm-cache-metadata.c for (b = 0; ; b++) { b 1171 drivers/md/dm-cache-metadata.c r = fn(context, cmd->discard_block_size, to_dblock(b), b 1176 drivers/md/dm-cache-metadata.c if (b >= (from_dblock(cmd->discard_nr_blocks) - 1)) b 1187 drivers/md/dm-cache-metadata.c for (b = 0; b < from_dblock(cmd->discard_nr_blocks); b++) { b 1188 drivers/md/dm-cache-metadata.c r = fn(context, cmd->discard_block_size, to_dblock(b), false); b 1292 drivers/md/dm-cache-policy-smq.c static dm_oblock_t to_hblock(struct smq_policy *mq, dm_oblock_t b) b 1294 drivers/md/dm-cache-policy-smq.c sector_t r = from_oblock(b); b 1299 drivers/md/dm-cache-policy-smq.c static struct entry *update_hotspot_queue(struct smq_policy *mq, dm_oblock_t b) b 1302 drivers/md/dm-cache-policy-smq.c dm_oblock_t hb = to_hblock(mq, b); b 173 drivers/md/dm-cache-target.c struct batcher *b = container_of(_ws, struct batcher, commit_work); b 189 drivers/md/dm-cache-target.c spin_lock_irqsave(&b->lock, flags); b 190 drivers/md/dm-cache-target.c list_splice_init(&b->work_items, &work_items); b 191 drivers/md/dm-cache-target.c bio_list_merge(&bios, &b->bios); b 192 drivers/md/dm-cache-target.c bio_list_init(&b->bios); b 193 drivers/md/dm-cache-target.c b->commit_scheduled = false; b 194 drivers/md/dm-cache-target.c spin_unlock_irqrestore(&b->lock, flags); b 196 drivers/md/dm-cache-target.c r = b->commit_op(b->commit_context); b 202 drivers/md/dm-cache-target.c queue_work(b->wq, ws); b 210 drivers/md/dm-cache-target.c b->issue_op(bio, b->issue_context); b 214 drivers/md/dm-cache-target.c static void batcher_init(struct batcher *b, b 221 drivers/md/dm-cache-target.c b->commit_op = commit_op; b 222 drivers/md/dm-cache-target.c b->commit_context = commit_context; b 223 drivers/md/dm-cache-target.c b->issue_op = issue_op; b 224 drivers/md/dm-cache-target.c b->issue_context = issue_context; b 225 drivers/md/dm-cache-target.c b->wq = wq; b 227 drivers/md/dm-cache-target.c spin_lock_init(&b->lock); b 228 drivers/md/dm-cache-target.c INIT_LIST_HEAD(&b->work_items); b 229 drivers/md/dm-cache-target.c bio_list_init(&b->bios); b 230 drivers/md/dm-cache-target.c INIT_WORK(&b->commit_work, __commit); b 231 drivers/md/dm-cache-target.c b->commit_scheduled = false; b 234 drivers/md/dm-cache-target.c static void async_commit(struct batcher *b) b 236 drivers/md/dm-cache-target.c queue_work(b->wq, &b->commit_work); b 239 drivers/md/dm-cache-target.c static void continue_after_commit(struct batcher *b, struct continuation *k) b 244 drivers/md/dm-cache-target.c spin_lock_irqsave(&b->lock, flags); b 245 drivers/md/dm-cache-target.c commit_scheduled = b->commit_scheduled; b 246 drivers/md/dm-cache-target.c list_add_tail(&k->ws.entry, &b->work_items); b 247 drivers/md/dm-cache-target.c spin_unlock_irqrestore(&b->lock, flags); b 250 drivers/md/dm-cache-target.c async_commit(b); b 256 drivers/md/dm-cache-target.c static void issue_after_commit(struct batcher *b, struct bio *bio) b 261 drivers/md/dm-cache-target.c spin_lock_irqsave(&b->lock, flags); b 262 drivers/md/dm-cache-target.c commit_scheduled = b->commit_scheduled; b 263 drivers/md/dm-cache-target.c bio_list_add(&b->bios, bio); b 264 drivers/md/dm-cache-target.c spin_unlock_irqrestore(&b->lock, flags); b 267 drivers/md/dm-cache-target.c async_commit(b); b 273 drivers/md/dm-cache-target.c static void schedule_commit(struct batcher *b) b 278 drivers/md/dm-cache-target.c spin_lock_irqsave(&b->lock, flags); b 279 drivers/md/dm-cache-target.c immediate = !list_empty(&b->work_items) || !bio_list_empty(&b->bios); b 280 drivers/md/dm-cache-target.c b->commit_scheduled = true; b 281 drivers/md/dm-cache-target.c spin_unlock_irqrestore(&b->lock, flags); b 284 drivers/md/dm-cache-target.c async_commit(b); b 579 drivers/md/dm-cache-target.c static inline dm_oblock_t oblock_succ(dm_oblock_t b) b 581 drivers/md/dm-cache-target.c return to_oblock(from_oblock(b) + 1ull); b 687 drivers/md/dm-cache-target.c static bool is_dirty(struct cache *cache, dm_cblock_t b) b 689 drivers/md/dm-cache-target.c return test_bit(from_cblock(b), cache->dirty_bitset); b 732 drivers/md/dm-cache-target.c static dm_block_t block_div(dm_block_t b, uint32_t n) b 734 drivers/md/dm-cache-target.c do_div(b, n); b 736 drivers/md/dm-cache-target.c return b; b 757 drivers/md/dm-cache-target.c static void set_discard(struct cache *cache, dm_dblock_t b) b 761 drivers/md/dm-cache-target.c BUG_ON(from_dblock(b) >= from_dblock(cache->discard_nr_blocks)); b 765 drivers/md/dm-cache-target.c set_bit(from_dblock(b), cache->discard_bitset); b 769 drivers/md/dm-cache-target.c static void clear_discard(struct cache *cache, dm_dblock_t b) b 774 drivers/md/dm-cache-target.c clear_bit(from_dblock(b), cache->discard_bitset); b 778 drivers/md/dm-cache-target.c static bool is_discarded(struct cache *cache, dm_dblock_t b) b 784 drivers/md/dm-cache-target.c r = test_bit(from_dblock(b), cache->discard_bitset); b 790 drivers/md/dm-cache-target.c static bool is_discarded_oblock(struct cache *cache, dm_oblock_t b) b 796 drivers/md/dm-cache-target.c r = test_bit(from_dblock(oblock_to_dblock(cache, b)), b 1102 drivers/md/dm-cache-target.c dm_dblock_t *b, dm_dblock_t *e) b 1107 drivers/md/dm-cache-target.c *b = to_dblock(dm_sector_div_up(sb, cache->discard_block_size)); b 1110 drivers/md/dm-cache-target.c *e = *b; b 1868 drivers/md/dm-cache-target.c dm_dblock_t b, e; b 1873 drivers/md/dm-cache-target.c calc_discard_block_range(cache, bio, &b, &e); b 1874 drivers/md/dm-cache-target.c while (b != e) { b 1875 drivers/md/dm-cache-target.c set_discard(cache, b); b 1876 drivers/md/dm-cache-target.c b = to_dblock(from_dblock(b) + 1); b 1957 drivers/md/dm-cache-target.c enum busy b; b 1960 drivers/md/dm-cache-target.c b = spare_migration_bandwidth(cache); b 1962 drivers/md/dm-cache-target.c r = policy_get_background_work(cache->policy, b == IDLE, &op); b 2123 drivers/md/dm-cache-target.c char b[BDEVNAME_SIZE]; b 2138 drivers/md/dm-cache-target.c bdevname(ca->metadata_dev->bdev, b), THIN_METADATA_MAX_SECTORS); b 2929 drivers/md/dm-cache-target.c sector_t b, e; b 2937 drivers/md/dm-cache-target.c b = li->discard_begin * li->block_size; b 2943 drivers/md/dm-cache-target.c b = dm_sector_div_up(b, li->cache->discard_block_size); b 2953 drivers/md/dm-cache-target.c for (; b < e; b++) b 2954 drivers/md/dm-cache-target.c set_discard(li->cache, to_dblock(b)); b 3271 drivers/md/dm-cache-target.c uint64_t b, e; b 3277 drivers/md/dm-cache-target.c r = sscanf(str, "%llu-%llu%c", &b, &e, &dummy); b 3282 drivers/md/dm-cache-target.c result->begin = to_cblock(b); b 3290 drivers/md/dm-cache-target.c r = sscanf(str, "%llu%c", &b, &dummy); b 3295 drivers/md/dm-cache-target.c result->begin = to_cblock(b); b 3306 drivers/md/dm-cache-target.c uint64_t b = from_cblock(range->begin); b 3310 drivers/md/dm-cache-target.c if (b >= n) { b 3312 drivers/md/dm-cache-target.c cache_device_name(cache), b, n); b 3322 drivers/md/dm-cache-target.c if (b >= e) { b 3324 drivers/md/dm-cache-target.c cache_device_name(cache), b, e); b 3331 drivers/md/dm-cache-target.c static inline dm_cblock_t cblock_succ(dm_cblock_t b) b 3333 drivers/md/dm-cache-target.c return to_cblock(from_cblock(b) + 1); b 167 drivers/md/dm-clone-metadata.c struct dm_block *b, size_t sb_block_size) b 172 drivers/md/dm-clone-metadata.c sb = dm_block_data(b); b 173 drivers/md/dm-clone-metadata.c sb->blocknr = cpu_to_le64(dm_block_location(b)); b 180 drivers/md/dm-clone-metadata.c static int sb_check(struct dm_block_validator *v, struct dm_block *b, b 186 drivers/md/dm-clone-metadata.c sb = dm_block_data(b); b 188 drivers/md/dm-clone-metadata.c if (dm_block_location(b) != le64_to_cpu(sb->blocknr)) { b 191 drivers/md/dm-clone-metadata.c (unsigned long long)dm_block_location(b)); b 1699 drivers/md/dm-clone-target.c char b[BDEVNAME_SIZE]; b 1711 drivers/md/dm-clone-target.c bdevname(clone->metadata_dev->bdev, b), DM_CLONE_METADATA_MAX_SECTORS); b 1048 drivers/md/dm-crypt.c char b[BDEVNAME_SIZE]; b 1049 drivers/md/dm-crypt.c DMERR_LIMIT("%s: INTEGRITY AEAD ERROR, sector %llu", bio_devname(ctx->bio_in, b), b 1694 drivers/md/dm-crypt.c char b[BDEVNAME_SIZE]; b 1695 drivers/md/dm-crypt.c DMERR_LIMIT("%s: INTEGRITY AEAD ERROR, sector %llu", bio_devname(ctx->bio_in, b), b 192 drivers/md/dm-era-target.c struct dm_block *b, b 195 drivers/md/dm-era-target.c struct superblock_disk *disk = dm_block_data(b); b 197 drivers/md/dm-era-target.c disk->blocknr = cpu_to_le64(dm_block_location(b)); b 216 drivers/md/dm-era-target.c struct dm_block *b, b 219 drivers/md/dm-era-target.c struct superblock_disk *disk = dm_block_data(b); b 222 drivers/md/dm-era-target.c if (dm_block_location(b) != le64_to_cpu(disk->blocknr)) { b 225 drivers/md/dm-era-target.c (unsigned long long)dm_block_location(b)); b 326 drivers/md/dm-era-target.c struct dm_block *b; b 333 drivers/md/dm-era-target.c r = dm_bm_read_lock(bm, SUPERBLOCK_LOCATION, NULL, &b); b 337 drivers/md/dm-era-target.c data_le = dm_block_data(b); b 346 drivers/md/dm-era-target.c dm_bm_unlock(b); b 369 drivers/md/dm-era-target.c dm_block_t b; b 372 drivers/md/dm-era-target.c b = le64_to_cpu(ws_d.root); b 374 drivers/md/dm-era-target.c dm_tm_inc(md->tm, b); b 381 drivers/md/dm-era-target.c dm_block_t b; b 384 drivers/md/dm-era-target.c b = le64_to_cpu(ws_d.root); b 386 drivers/md/dm-era-target.c dm_bitset_del(&md->bitset_info, b); b 691 drivers/md/dm-era-target.c unsigned b, e = min(d->current_bit + INSERTS_PER_STEP, d->nr_bits); b 693 drivers/md/dm-era-target.c for (b = d->current_bit; b < e; b++) { b 694 drivers/md/dm-era-target.c r = writeset_marked_on_disk(&d->info, &d->writeset, b, &marked); b 705 drivers/md/dm-era-target.c b, &d->value, &md->era_array_root); b 712 drivers/md/dm-era-target.c if (b == d->nr_bits) b 715 drivers/md/dm-era-target.c d->current_bit = b; b 1299 drivers/md/dm-integrity.c struct dm_buffer *b; b 1307 drivers/md/dm-integrity.c data = dm_bufio_read(ic->bufio, *metadata_block, &b); b 1317 drivers/md/dm-integrity.c dm_bufio_mark_partial_buffer_dirty(b, *metadata_offset, *metadata_offset + to_copy); b 1328 drivers/md/dm-integrity.c dm_bufio_release(b); b 1332 drivers/md/dm-integrity.c dm_bufio_release(b); b 813 drivers/md/dm-mpath.c char b[BDEVNAME_SIZE]; b 816 drivers/md/dm-mpath.c bdevname(bdev, b)); b 985 drivers/md/dm-snap.c struct bio *b = NULL; b 1001 drivers/md/dm-snap.c b = __release_queued_bios_after_merge(s); b 1005 drivers/md/dm-snap.c if (b) b 1006 drivers/md/dm-snap.c flush_bios(b); b 1127 drivers/md/dm-snap.c struct bio *b = NULL; b 1153 drivers/md/dm-snap.c b = __release_queued_bios_after_merge(s); b 1155 drivers/md/dm-snap.c error_bios(b); b 289 drivers/md/dm-table.c char b[BDEVNAME_SIZE]; b 300 drivers/md/dm-table.c dm_device_name(ti->table->md), bdevname(bdev, b), b 313 drivers/md/dm-table.c dm_device_name(ti->table->md), bdevname(bdev, b), b 331 drivers/md/dm-table.c zone_sectors, bdevname(bdev, b)); b 348 drivers/md/dm-table.c zone_sectors, bdevname(bdev, b)); b 361 drivers/md/dm-table.c limits->logical_block_size, bdevname(bdev, b)); b 370 drivers/md/dm-table.c limits->logical_block_size, bdevname(bdev, b)); b 473 drivers/md/dm-table.c char b[BDEVNAME_SIZE]; b 477 drivers/md/dm-table.c dm_device_name(ti->table->md), bdevname(bdev, b)); b 485 drivers/md/dm-table.c dm_device_name(ti->table->md), bdevname(bdev, b), b 1735 drivers/md/dm-table.c char b[BDEVNAME_SIZE]; b 1738 drivers/md/dm-table.c return (strncmp(bdevname(dev->bdev, b), "nvme", 4) == 0); b 2096 drivers/md/dm-table.c char b[BDEVNAME_SIZE]; b 2103 drivers/md/dm-table.c bdevname(dd->dm_dev->bdev, b)); b 249 drivers/md/dm-thin-metadata.c struct dm_block *b, b 252 drivers/md/dm-thin-metadata.c struct thin_disk_superblock *disk_super = dm_block_data(b); b 254 drivers/md/dm-thin-metadata.c disk_super->blocknr = cpu_to_le64(dm_block_location(b)); b 261 drivers/md/dm-thin-metadata.c struct dm_block *b, b 264 drivers/md/dm-thin-metadata.c struct thin_disk_superblock *disk_super = dm_block_data(b); b 267 drivers/md/dm-thin-metadata.c if (dm_block_location(b) != le64_to_cpu(disk_super->blocknr)) { b 270 drivers/md/dm-thin-metadata.c (unsigned long long)dm_block_location(b)); b 303 drivers/md/dm-thin-metadata.c static uint64_t pack_block_time(dm_block_t b, uint32_t t) b 305 drivers/md/dm-thin-metadata.c return (b << 24) | t; b 308 drivers/md/dm-thin-metadata.c static void unpack_block_time(uint64_t v, dm_block_t *b, uint32_t *t) b 310 drivers/md/dm-thin-metadata.c *b = v >> 24; b 318 drivers/md/dm-thin-metadata.c uint64_t b; b 322 drivers/md/dm-thin-metadata.c unpack_block_time(le64_to_cpu(v_le), &b, &t); b 323 drivers/md/dm-thin-metadata.c dm_sm_inc_block(sm, b); b 330 drivers/md/dm-thin-metadata.c uint64_t b; b 334 drivers/md/dm-thin-metadata.c unpack_block_time(le64_to_cpu(v_le), &b, &t); b 335 drivers/md/dm-thin-metadata.c dm_sm_dec_block(sm, b); b 429 drivers/md/dm-thin-metadata.c struct dm_block *b; b 436 drivers/md/dm-thin-metadata.c r = dm_bm_read_lock(bm, THIN_SUPERBLOCK_LOCATION, NULL, &b); b 440 drivers/md/dm-thin-metadata.c data_le = dm_block_data(b); b 449 drivers/md/dm-thin-metadata.c dm_bm_unlock(b); b 1743 drivers/md/dm-thin-metadata.c int dm_pool_block_is_shared(struct dm_pool_metadata *pmd, dm_block_t b, bool *result) b 1749 drivers/md/dm-thin-metadata.c r = dm_sm_get_count(pmd->data_sm, b, &ref_count); b 1757 drivers/md/dm-thin-metadata.c int dm_pool_inc_data_range(struct dm_pool_metadata *pmd, dm_block_t b, dm_block_t e) b 1762 drivers/md/dm-thin-metadata.c for (; b != e; b++) { b 1763 drivers/md/dm-thin-metadata.c r = dm_sm_inc_block(pmd->data_sm, b); b 1772 drivers/md/dm-thin-metadata.c int dm_pool_dec_data_range(struct dm_pool_metadata *pmd, dm_block_t b, dm_block_t e) b 1777 drivers/md/dm-thin-metadata.c for (; b != e; b++) { b 1778 drivers/md/dm-thin-metadata.c r = dm_sm_dec_block(pmd->data_sm, b); b 198 drivers/md/dm-thin-metadata.h int dm_pool_block_is_shared(struct dm_pool_metadata *pmd, dm_block_t b, bool *result); b 200 drivers/md/dm-thin-metadata.h int dm_pool_inc_data_range(struct dm_pool_metadata *pmd, dm_block_t b, dm_block_t e); b 201 drivers/md/dm-thin-metadata.h int dm_pool_dec_data_range(struct dm_pool_metadata *pmd, dm_block_t b, dm_block_t e); b 121 drivers/md/dm-thin.c dm_block_t b, dm_block_t e, struct dm_cell_key *key) b 125 drivers/md/dm-thin.c key->block_begin = b; b 129 drivers/md/dm-thin.c static void build_data_key(struct dm_thin_device *td, dm_block_t b, b 132 drivers/md/dm-thin.c build_key(td, PHYSICAL, b, b + 1llu, key); b 135 drivers/md/dm-thin.c static void build_virtual_key(struct dm_thin_device *td, dm_block_t b, b 138 drivers/md/dm-thin.c build_key(td, VIRTUAL, b, b + 1llu, key); b 371 drivers/md/dm-thin.c static sector_t block_to_sectors(struct pool *pool, dm_block_t b) b 374 drivers/md/dm-thin.c (b << pool->sectors_per_block_shift) : b 375 drivers/md/dm-thin.c (b * pool->sectors_per_block); b 700 drivers/md/dm-thin.c sector_t b = bio->bi_iter.bi_sector; b 701 drivers/md/dm-thin.c sector_t e = b + (bio->bi_iter.bi_size >> SECTOR_SHIFT); b 703 drivers/md/dm-thin.c b += pool->sectors_per_block - 1ull; /* so we round up */ b 706 drivers/md/dm-thin.c b >>= pool->sectors_per_block_shift; b 709 drivers/md/dm-thin.c (void) sector_div(b, pool->sectors_per_block); b 713 drivers/md/dm-thin.c if (e < b) b 715 drivers/md/dm-thin.c e = b; b 717 drivers/md/dm-thin.c *begin = b; b 1090 drivers/md/dm-thin.c dm_block_t b = m->data_block, e, end = m->data_block + m->virt_end - m->virt_begin; b 1094 drivers/md/dm-thin.c while (b != end) { b 1096 drivers/md/dm-thin.c for (; b < end; b++) { b 1097 drivers/md/dm-thin.c r = dm_pool_block_is_shared(pool->pmd, b, &shared); b 1105 drivers/md/dm-thin.c if (b == end) b 1109 drivers/md/dm-thin.c for (e = b + 1; e != end; e++) { b 1118 drivers/md/dm-thin.c r = issue_discard(&op, b, e); b 1122 drivers/md/dm-thin.c b = e; b 472 drivers/md/dm-verity-target.c unsigned b; b 475 drivers/md/dm-verity-target.c for (b = 0; b < io->n_blocks; b++) { b 477 drivers/md/dm-verity-target.c sector_t cur_block = io->block + b; b 864 drivers/md/dm-writecache.c size_t b; b 871 drivers/md/dm-writecache.c for (b = 0; b < wc->n_blocks; b++) { b 872 drivers/md/dm-writecache.c struct wc_entry *e = &wc->entries[b]; b 873 drivers/md/dm-writecache.c e->index = b; b 902 drivers/md/dm-writecache.c size_t b; b 940 drivers/md/dm-writecache.c for (b = 0; b < wc->n_blocks; b++) { b 941 drivers/md/dm-writecache.c struct wc_entry *e = &wc->entries[b]; b 951 drivers/md/dm-writecache.c (unsigned long)b, r); b 961 drivers/md/dm-writecache.c for (b = 0; b < wc->n_blocks; b++) { b 962 drivers/md/dm-writecache.c struct wc_entry *e = &wc->entries[b]; b 980 drivers/md/dm-writecache.c (unsigned long long)b, (unsigned long long)read_original_sector(wc, e), b 1785 drivers/md/dm-writecache.c size_t b; b 1796 drivers/md/dm-writecache.c for (b = 0; b < ARRAY_SIZE(sb(wc)->padding); b++) b 1797 drivers/md/dm-writecache.c pmem_assign(sb(wc)->padding[b], cpu_to_le64(0)); b 1803 drivers/md/dm-writecache.c for (b = 0; b < wc->n_blocks; b++) { b 1804 drivers/md/dm-writecache.c write_original_sector_seq_count(wc, &wc->entries[b], -1, -1); b 330 drivers/md/dm-zoned-metadata.c struct dmz_mblock *b; b 334 drivers/md/dm-zoned-metadata.c b = container_of(*new, struct dmz_mblock, node); b 336 drivers/md/dm-zoned-metadata.c new = (b->no < mblk->no) ? &((*new)->rb_left) : &((*new)->rb_right); b 42 drivers/md/dm-zoned.h #define dmz_blk2sect(b) ((sector_t)(b) << DMZ_BLOCK_SECTORS_SHIFT) b 71 drivers/md/dm-zoned.h #define dmz_chunk_block(dev, b) ((b) & ((dev)->zone_nr_blocks - 1)) b 1648 drivers/md/dm.c struct bio *b = bio_split(bio, bio_sectors(bio) - ci.sector_count, b 1650 drivers/md/dm.c ci.io->orig_bio = b; b 1664 drivers/md/dm.c bio_chain(b, bio); b 1665 drivers/md/dm.c trace_block_split(md->queue, b, bio->bi_iter.bi_sector); b 1110 drivers/md/md-bitmap.c int b; b 1155 drivers/md/md-bitmap.c b = test_bit(bit, paddr); b 1157 drivers/md/md-bitmap.c b = test_bit_le(bit, paddr); b 1159 drivers/md/md-bitmap.c if (b) { b 66 drivers/md/md-faulty.c struct bio *b = bio->bi_private; b 68 drivers/md/md-faulty.c b->bi_iter.bi_size = bio->bi_iter.bi_size; b 69 drivers/md/md-faulty.c b->bi_iter.bi_sector = bio->bi_iter.bi_sector; b 73 drivers/md/md-faulty.c bio_io_error(b); b 208 drivers/md/md-faulty.c struct bio *b = bio_clone_fast(bio, GFP_NOIO, &mddev->bio_set); b 210 drivers/md/md-faulty.c bio_set_dev(b, conf->rdev->bdev); b 211 drivers/md/md-faulty.c b->bi_private = bio; b 212 drivers/md/md-faulty.c b->bi_end_io = faulty_fail; b 213 drivers/md/md-faulty.c bio = b; b 242 drivers/md/md-linear.c char b[BDEVNAME_SIZE]; b 297 drivers/md/md-linear.c bdevname(tmp_dev->rdev->bdev, b), b 90 drivers/md/md-multipath.c char b[BDEVNAME_SIZE]; b 93 drivers/md/md-multipath.c bdevname(rdev->bdev,b), b 182 drivers/md/md-multipath.c char b[BDEVNAME_SIZE]; b 207 drivers/md/md-multipath.c bdevname(rdev->bdev, b), b 225 drivers/md/md-multipath.c char b[BDEVNAME_SIZE]; b 230 drivers/md/md-multipath.c bdevname(tmp->rdev->bdev,b)); b 324 drivers/md/md-multipath.c char b[BDEVNAME_SIZE]; b 337 drivers/md/md-multipath.c bio_devname(bio, b), b 342 drivers/md/md-multipath.c bio_devname(bio, b), b 947 drivers/md/md.c char b[BDEVNAME_SIZE]; b 959 drivers/md/md.c bdevname(rdev->bdev,b)); b 1105 drivers/md/md.c char b[BDEVNAME_SIZE], b2[BDEVNAME_SIZE]; b 1124 drivers/md/md.c bdevname(rdev->bdev, b); b 1128 drivers/md/md.c pr_warn("md: invalid raid superblock magic on %s\n", b); b 1136 drivers/md/md.c sb->major_version, sb->minor_version, b); b 1144 drivers/md/md.c pr_warn("md: invalid superblock checksum on %s\n", b); b 1177 drivers/md/md.c b, bdevname(refdev->bdev,b2)); b 1182 drivers/md/md.c b, bdevname(refdev->bdev, b2)); b 1546 drivers/md/md.c char b[BDEVNAME_SIZE], b2[BDEVNAME_SIZE]; b 1592 drivers/md/md.c bdevname(rdev->bdev,b)); b 1597 drivers/md/md.c bdevname(rdev->bdev,b)); b 1704 drivers/md/md.c bdevname(rdev->bdev,b), b 2281 drivers/md/md.c char b[BDEVNAME_SIZE]; b 2333 drivers/md/md.c bdevname(rdev->bdev,b); b 2334 drivers/md/md.c strreplace(b, '/', '!'); b 2337 drivers/md/md.c pr_debug("md: bind<%s>\n", b); b 2342 drivers/md/md.c if ((err = kobject_add(&rdev->kobj, &mddev->kobj, "dev-%s", b))) b 2360 drivers/md/md.c b, mdname(mddev)); b 2373 drivers/md/md.c char b[BDEVNAME_SIZE]; b 2377 drivers/md/md.c pr_debug("md: unbind<%s>\n", bdevname(rdev->bdev,b)); b 2403 drivers/md/md.c char b[BDEVNAME_SIZE]; b 2408 drivers/md/md.c pr_warn("md: could not open %s.\n", __bdevname(dev, b)); b 2426 drivers/md/md.c char b[BDEVNAME_SIZE]; b 2428 drivers/md/md.c pr_debug("md: export_rdev(%s)\n", bdevname(rdev->bdev,b)); b 2684 drivers/md/md.c char b[BDEVNAME_SIZE]; b 2694 drivers/md/md.c bdevname(rdev->bdev, b), b 2707 drivers/md/md.c bdevname(rdev->bdev, b)); b 3572 drivers/md/md.c char b[BDEVNAME_SIZE]; b 3597 drivers/md/md.c bdevname(rdev->bdev,b)); b 3607 drivers/md/md.c bdevname(rdev->bdev,b), b 3613 drivers/md/md.c bdevname(rdev->bdev,b)); b 3636 drivers/md/md.c char b[BDEVNAME_SIZE]; b 3649 drivers/md/md.c bdevname(rdev->bdev,b)); b 3668 drivers/md/md.c mdname(mddev), bdevname(rdev->bdev, b), b 3677 drivers/md/md.c bdevname(rdev->bdev,b)); b 5713 drivers/md/md.c char b[BDEVNAME_SIZE], b2[BDEVNAME_SIZE]; b 5724 drivers/md/md.c bdevname(rdev->bdev,b), b 6245 drivers/md/md.c char b[BDEVNAME_SIZE]; b 6246 drivers/md/md.c pr_cont("<%s>", bdevname(rdev->bdev,b)); b 6273 drivers/md/md.c char b[BDEVNAME_SIZE]; b 6283 drivers/md/md.c pr_debug("md: considering %s ...\n", bdevname(rdev0->bdev,b)); b 6288 drivers/md/md.c bdevname(rdev->bdev,b)); b 6306 drivers/md/md.c bdevname(rdev0->bdev, b), rdev0->preferred_minor); b 6322 drivers/md/md.c mdname(mddev), bdevname(rdev0->bdev,b)); b 6496 drivers/md/md.c char b[BDEVNAME_SIZE], b2[BDEVNAME_SIZE]; b 6527 drivers/md/md.c bdevname(rdev->bdev,b), b 6703 drivers/md/md.c char b[BDEVNAME_SIZE]; b 6737 drivers/md/md.c bdevname(rdev->bdev,b), mdname(mddev)); b 6743 drivers/md/md.c char b[BDEVNAME_SIZE]; b 6777 drivers/md/md.c bdevname(rdev->bdev,b), mdname(mddev)); b 7800 drivers/md/md.c char b[BDEVNAME_SIZE]; b 7803 drivers/md/md.c bdevname(rdev->bdev,b)); b 8039 drivers/md/md.c char b[BDEVNAME_SIZE]; b 8041 drivers/md/md.c bdevname(rdev->bdev,b), rdev->desc_nr); b 9361 drivers/md/md.c char b[BDEVNAME_SIZE]; b 9385 drivers/md/md.c pr_info("md: Removing Candidate device %s because add failed\n", bdevname(rdev2->bdev,b)); b 9403 drivers/md/md.c bdevname(rdev2->bdev,b)); b 41 drivers/md/persistent-data/dm-array.c struct dm_block *b, b 44 drivers/md/persistent-data/dm-array.c struct array_block *bh_le = dm_block_data(b); b 46 drivers/md/persistent-data/dm-array.c bh_le->blocknr = cpu_to_le64(dm_block_location(b)); b 53 drivers/md/persistent-data/dm-array.c struct dm_block *b, b 56 drivers/md/persistent-data/dm-array.c struct array_block *bh_le = dm_block_data(b); b 59 drivers/md/persistent-data/dm-array.c if (dm_block_location(b) != le64_to_cpu(bh_le->blocknr)) { b 62 drivers/md/persistent-data/dm-array.c (unsigned long long) dm_block_location(b)); b 220 drivers/md/persistent-data/dm-array.c static int get_ablock(struct dm_array_info *info, dm_block_t b, b 225 drivers/md/persistent-data/dm-array.c r = dm_tm_read_lock(info->btree_info.tm, b, &array_validator, block); b 282 drivers/md/persistent-data/dm-array.c static int __shadow_ablock(struct dm_array_info *info, dm_block_t b, b 286 drivers/md/persistent-data/dm-array.c int r = dm_tm_shadow_block(info->btree_info.tm, b, b 303 drivers/md/persistent-data/dm-array.c struct dm_block *block, dm_block_t b, b 308 drivers/md/persistent-data/dm-array.c if (dm_block_location(block) != b) { b 315 drivers/md/persistent-data/dm-array.c dm_tm_inc(info->btree_info.tm, b); b 333 drivers/md/persistent-data/dm-array.c dm_block_t b; b 339 drivers/md/persistent-data/dm-array.c b = le64_to_cpu(block_le); b 341 drivers/md/persistent-data/dm-array.c r = __shadow_ablock(info, b, block, ab); b 345 drivers/md/persistent-data/dm-array.c return __reinsert_ablock(info, index, *block, b, root); b 588 drivers/md/persistent-data/dm-array.c uint64_t b; b 596 drivers/md/persistent-data/dm-array.c b = le64_to_cpu(block_le); b 598 drivers/md/persistent-data/dm-array.c r = dm_tm_ref(info->btree_info.tm, b, &ref_count); b 601 drivers/md/persistent-data/dm-array.c (unsigned long long) b); b 610 drivers/md/persistent-data/dm-array.c r = get_ablock(info, b, &block, &ab); b 613 drivers/md/persistent-data/dm-array.c (unsigned long long) b); b 621 drivers/md/persistent-data/dm-array.c dm_tm_dec(info->btree_info.tm, b); b 168 drivers/md/persistent-data/dm-bitset.c unsigned b = index % BITS_PER_ARRAY_ENTRY; b 174 drivers/md/persistent-data/dm-bitset.c set_bit(b, (unsigned long *) &info->current_bits); b 185 drivers/md/persistent-data/dm-bitset.c unsigned b = index % BITS_PER_ARRAY_ENTRY; b 191 drivers/md/persistent-data/dm-bitset.c clear_bit(b, (unsigned long *) &info->current_bits); b 202 drivers/md/persistent-data/dm-bitset.c unsigned b = index % BITS_PER_ARRAY_ENTRY; b 208 drivers/md/persistent-data/dm-bitset.c *result = test_bit(b, (unsigned long *) &info->current_bits); b 300 drivers/md/persistent-data/dm-block-manager.c static void report_recursive_bug(dm_block_t b, int r) b 304 drivers/md/persistent-data/dm-block-manager.c (unsigned long long) b); b 328 drivers/md/persistent-data/dm-block-manager.c static struct dm_buffer *to_buffer(struct dm_block *b) b 330 drivers/md/persistent-data/dm-block-manager.c return (struct dm_buffer *) b; b 333 drivers/md/persistent-data/dm-block-manager.c dm_block_t dm_block_location(struct dm_block *b) b 335 drivers/md/persistent-data/dm-block-manager.c return dm_bufio_get_block_number(to_buffer(b)); b 339 drivers/md/persistent-data/dm-block-manager.c void *dm_block_data(struct dm_block *b) b 341 drivers/md/persistent-data/dm-block-manager.c return dm_bufio_get_block_data(to_buffer(b)); b 455 drivers/md/persistent-data/dm-block-manager.c int dm_bm_read_lock(struct dm_block_manager *bm, dm_block_t b, b 463 drivers/md/persistent-data/dm-block-manager.c p = dm_bufio_read(bm->bufio, b, (struct dm_buffer **) result); b 471 drivers/md/persistent-data/dm-block-manager.c report_recursive_bug(b, r); b 489 drivers/md/persistent-data/dm-block-manager.c dm_block_t b, struct dm_block_validator *v, b 499 drivers/md/persistent-data/dm-block-manager.c p = dm_bufio_read(bm->bufio, b, (struct dm_buffer **) result); b 507 drivers/md/persistent-data/dm-block-manager.c report_recursive_bug(b, r); b 525 drivers/md/persistent-data/dm-block-manager.c dm_block_t b, struct dm_block_validator *v, b 532 drivers/md/persistent-data/dm-block-manager.c p = dm_bufio_get(bm->bufio, b, (struct dm_buffer **) result); b 542 drivers/md/persistent-data/dm-block-manager.c report_recursive_bug(b, r); b 558 drivers/md/persistent-data/dm-block-manager.c dm_block_t b, struct dm_block_validator *v, b 568 drivers/md/persistent-data/dm-block-manager.c p = dm_bufio_new(bm->bufio, b, (struct dm_buffer **) result); b 588 drivers/md/persistent-data/dm-block-manager.c void dm_bm_unlock(struct dm_block *b) b 591 drivers/md/persistent-data/dm-block-manager.c aux = dm_bufio_get_aux_data(to_buffer(b)); b 594 drivers/md/persistent-data/dm-block-manager.c dm_bufio_mark_buffer_dirty(to_buffer(b)); b 599 drivers/md/persistent-data/dm-block-manager.c dm_bufio_release(to_buffer(b)); b 612 drivers/md/persistent-data/dm-block-manager.c void dm_bm_prefetch(struct dm_block_manager *bm, dm_block_t b) b 614 drivers/md/persistent-data/dm-block-manager.c dm_bufio_prefetch(bm->bufio, b, 1); b 21 drivers/md/persistent-data/dm-block-manager.h dm_block_t dm_block_location(struct dm_block *b); b 22 drivers/md/persistent-data/dm-block-manager.h void *dm_block_data(struct dm_block *b); b 52 drivers/md/persistent-data/dm-block-manager.h void (*prepare_for_write)(struct dm_block_validator *v, struct dm_block *b, size_t block_size); b 57 drivers/md/persistent-data/dm-block-manager.h int (*check)(struct dm_block_validator *v, struct dm_block *b, size_t block_size); b 73 drivers/md/persistent-data/dm-block-manager.h int dm_bm_read_lock(struct dm_block_manager *bm, dm_block_t b, b 77 drivers/md/persistent-data/dm-block-manager.h int dm_bm_write_lock(struct dm_block_manager *bm, dm_block_t b, b 85 drivers/md/persistent-data/dm-block-manager.h int dm_bm_read_try_lock(struct dm_block_manager *bm, dm_block_t b, b 93 drivers/md/persistent-data/dm-block-manager.h int dm_bm_write_lock_zero(struct dm_block_manager *bm, dm_block_t b, b 97 drivers/md/persistent-data/dm-block-manager.h void dm_bm_unlock(struct dm_block *b); b 113 drivers/md/persistent-data/dm-block-manager.h void dm_bm_prefetch(struct dm_block_manager *bm, dm_block_t b); b 48 drivers/md/persistent-data/dm-btree-internal.h int bn_read_lock(struct dm_btree_info *info, dm_block_t b, b 55 drivers/md/persistent-data/dm-btree-internal.h void unlock_block(struct dm_btree_info *info, struct dm_block *b); b 88 drivers/md/persistent-data/dm-btree-internal.h int shadow_step(struct shadow_spine *s, dm_block_t b, b 418 drivers/md/persistent-data/dm-btree-remove.c dm_block_t b = value64(n, 0); b 420 drivers/md/persistent-data/dm-btree-remove.c r = dm_tm_read_lock(info->tm, b, &btree_node_validator, &child); b 19 drivers/md/persistent-data/dm-btree-spine.c struct dm_block *b, b 23 drivers/md/persistent-data/dm-btree-spine.c struct dm_block *b, b 26 drivers/md/persistent-data/dm-btree-spine.c struct btree_node *n = dm_block_data(b); b 29 drivers/md/persistent-data/dm-btree-spine.c h->blocknr = cpu_to_le64(dm_block_location(b)); b 34 drivers/md/persistent-data/dm-btree-spine.c BUG_ON(node_check(v, b, 4096)); b 38 drivers/md/persistent-data/dm-btree-spine.c struct dm_block *b, b 41 drivers/md/persistent-data/dm-btree-spine.c struct btree_node *n = dm_block_data(b); b 47 drivers/md/persistent-data/dm-btree-spine.c if (dm_block_location(b) != le64_to_cpu(h->blocknr)) { b 49 drivers/md/persistent-data/dm-btree-spine.c le64_to_cpu(h->blocknr), dm_block_location(b)); b 95 drivers/md/persistent-data/dm-btree-spine.c int bn_read_lock(struct dm_btree_info *info, dm_block_t b, b 98 drivers/md/persistent-data/dm-btree-spine.c return dm_tm_read_lock(info->tm, b, &btree_node_validator, result); b 120 drivers/md/persistent-data/dm-btree-spine.c void unlock_block(struct dm_btree_info *info, struct dm_block *b) b 122 drivers/md/persistent-data/dm-btree-spine.c dm_tm_unlock(info->tm, b); b 199 drivers/md/persistent-data/dm-btree-spine.c int shadow_step(struct shadow_spine *s, dm_block_t b, b 210 drivers/md/persistent-data/dm-btree-spine.c r = bn_shadow(s->info, b, vt, s->nodes + s->count); b 129 drivers/md/persistent-data/dm-btree.c struct dm_block *b; b 134 drivers/md/persistent-data/dm-btree.c r = new_block(info, &b); b 141 drivers/md/persistent-data/dm-btree.c n = dm_block_data(b); b 148 drivers/md/persistent-data/dm-btree.c *root = dm_block_location(b); b 149 drivers/md/persistent-data/dm-btree.c unlock_block(info, b); b 163 drivers/md/persistent-data/dm-btree.c struct dm_block *b; b 208 drivers/md/persistent-data/dm-btree.c static int push_frame(struct del_stack *s, dm_block_t b, unsigned level) b 218 drivers/md/persistent-data/dm-btree.c r = dm_tm_ref(s->tm, b, &ref_count); b 227 drivers/md/persistent-data/dm-btree.c dm_tm_dec(s->tm, b); b 233 drivers/md/persistent-data/dm-btree.c r = dm_tm_read_lock(s->tm, b, &btree_node_validator, &f->b); b 239 drivers/md/persistent-data/dm-btree.c f->n = dm_block_data(f->b); b 256 drivers/md/persistent-data/dm-btree.c dm_tm_dec(s->tm, dm_block_location(f->b)); b 257 drivers/md/persistent-data/dm-btree.c dm_tm_unlock(s->tm, f->b); b 266 drivers/md/persistent-data/dm-btree.c dm_tm_unlock(s->tm, f->b); b 294 drivers/md/persistent-data/dm-btree.c dm_block_t b; b 307 drivers/md/persistent-data/dm-btree.c b = value64(f->n, f->current_child); b 309 drivers/md/persistent-data/dm-btree.c r = push_frame(s, b, f->level); b 314 drivers/md/persistent-data/dm-btree.c b = value64(f->n, f->current_child); b 316 drivers/md/persistent-data/dm-btree.c r = push_frame(s, b, f->level + 1); b 1000 drivers/md/persistent-data/dm-btree.c struct btree_node *bn = dm_block_data(n->b); b 1015 drivers/md/persistent-data/dm-btree.c struct btree_node *bn = dm_block_data(n->b); b 1020 drivers/md/persistent-data/dm-btree.c static int push_node(struct dm_btree_cursor *c, dm_block_t b) b 1030 drivers/md/persistent-data/dm-btree.c r = bn_read_lock(c->info, b, &n->b); b 1046 drivers/md/persistent-data/dm-btree.c unlock_block(c->info, c->nodes[c->depth].b); b 1059 drivers/md/persistent-data/dm-btree.c bn = dm_block_data(n->b); b 1080 drivers/md/persistent-data/dm-btree.c bn = dm_block_data(n->b); b 1152 drivers/md/persistent-data/dm-btree.c struct btree_node *bn = dm_block_data(n->b); b 190 drivers/md/persistent-data/dm-btree.h struct dm_block *b; b 12 drivers/md/persistent-data/dm-persistent-data-internal.h static inline unsigned dm_hash_block(dm_block_t b, unsigned hash_mask) b 16 drivers/md/persistent-data/dm-persistent-data-internal.h return (((unsigned) b) * BIG_PRIME) & hash_mask; b 23 drivers/md/persistent-data/dm-space-map-common.c struct dm_block *b, b 26 drivers/md/persistent-data/dm-space-map-common.c struct disk_metadata_index *mi_le = dm_block_data(b); b 28 drivers/md/persistent-data/dm-space-map-common.c mi_le->blocknr = cpu_to_le64(dm_block_location(b)); b 35 drivers/md/persistent-data/dm-space-map-common.c struct dm_block *b, b 38 drivers/md/persistent-data/dm-space-map-common.c struct disk_metadata_index *mi_le = dm_block_data(b); b 41 drivers/md/persistent-data/dm-space-map-common.c if (dm_block_location(b) != le64_to_cpu(mi_le->blocknr)) { b 43 drivers/md/persistent-data/dm-space-map-common.c le64_to_cpu(mi_le->blocknr), dm_block_location(b)); b 73 drivers/md/persistent-data/dm-space-map-common.c struct dm_block *b, b 76 drivers/md/persistent-data/dm-space-map-common.c struct disk_bitmap_header *disk_header = dm_block_data(b); b 78 drivers/md/persistent-data/dm-space-map-common.c disk_header->blocknr = cpu_to_le64(dm_block_location(b)); b 85 drivers/md/persistent-data/dm-space-map-common.c struct dm_block *b, b 88 drivers/md/persistent-data/dm-space-map-common.c struct disk_bitmap_header *disk_header = dm_block_data(b); b 91 drivers/md/persistent-data/dm-space-map-common.c if (dm_block_location(b) != le64_to_cpu(disk_header->blocknr)) { b 93 drivers/md/persistent-data/dm-space-map-common.c le64_to_cpu(disk_header->blocknr), dm_block_location(b)); b 120 drivers/md/persistent-data/dm-space-map-common.c static void *dm_bitmap_data(struct dm_block *b) b 122 drivers/md/persistent-data/dm-space-map-common.c return dm_block_data(b) + sizeof(struct disk_bitmap_header); b 127 drivers/md/persistent-data/dm-space-map-common.c static unsigned dm_bitmap_word_used(void *addr, unsigned b) b 130 drivers/md/persistent-data/dm-space-map-common.c __le64 *w_le = words_le + (b >> ENTRIES_SHIFT); b 138 drivers/md/persistent-data/dm-space-map-common.c static unsigned sm_lookup_bitmap(void *addr, unsigned b) b 141 drivers/md/persistent-data/dm-space-map-common.c __le64 *w_le = words_le + (b >> ENTRIES_SHIFT); b 144 drivers/md/persistent-data/dm-space-map-common.c b = (b & (ENTRIES_PER_WORD - 1)) << 1; b 145 drivers/md/persistent-data/dm-space-map-common.c hi = !!test_bit_le(b, (void *) w_le); b 146 drivers/md/persistent-data/dm-space-map-common.c lo = !!test_bit_le(b + 1, (void *) w_le); b 150 drivers/md/persistent-data/dm-space-map-common.c static void sm_set_bitmap(void *addr, unsigned b, unsigned val) b 153 drivers/md/persistent-data/dm-space-map-common.c __le64 *w_le = words_le + (b >> ENTRIES_SHIFT); b 155 drivers/md/persistent-data/dm-space-map-common.c b = (b & (ENTRIES_PER_WORD - 1)) << 1; b 158 drivers/md/persistent-data/dm-space-map-common.c __set_bit_le(b, (void *) w_le); b 160 drivers/md/persistent-data/dm-space-map-common.c __clear_bit_le(b, (void *) w_le); b 163 drivers/md/persistent-data/dm-space-map-common.c __set_bit_le(b + 1, (void *) w_le); b 165 drivers/md/persistent-data/dm-space-map-common.c __clear_bit_le(b + 1, (void *) w_le); b 255 drivers/md/persistent-data/dm-space-map-common.c struct dm_block *b; b 258 drivers/md/persistent-data/dm-space-map-common.c r = dm_tm_new_block(ll->tm, &dm_sm_bitmap_validator, &b); b 262 drivers/md/persistent-data/dm-space-map-common.c idx.blocknr = cpu_to_le64(dm_block_location(b)); b 264 drivers/md/persistent-data/dm-space-map-common.c dm_tm_unlock(ll->tm, b); b 277 drivers/md/persistent-data/dm-space-map-common.c int sm_ll_lookup_bitmap(struct ll_disk *ll, dm_block_t b, uint32_t *result) b 280 drivers/md/persistent-data/dm-space-map-common.c dm_block_t index = b; b 284 drivers/md/persistent-data/dm-space-map-common.c b = do_div(index, ll->entries_per_block); b 294 drivers/md/persistent-data/dm-space-map-common.c *result = sm_lookup_bitmap(dm_bitmap_data(blk), b); b 301 drivers/md/persistent-data/dm-space-map-common.c static int sm_ll_lookup_big_ref_count(struct ll_disk *ll, dm_block_t b, b 307 drivers/md/persistent-data/dm-space-map-common.c r = dm_btree_lookup(&ll->ref_count_info, ll->ref_count_root, &b, &le_rc); b 316 drivers/md/persistent-data/dm-space-map-common.c int sm_ll_lookup(struct ll_disk *ll, dm_block_t b, uint32_t *result) b 318 drivers/md/persistent-data/dm-space-map-common.c int r = sm_ll_lookup_bitmap(ll, b, result); b 326 drivers/md/persistent-data/dm-space-map-common.c return sm_ll_lookup_big_ref_count(ll, b, result); b 384 drivers/md/persistent-data/dm-space-map-common.c dm_block_t begin, dm_block_t end, dm_block_t *b) b 390 drivers/md/persistent-data/dm-space-map-common.c r = sm_ll_find_free_block(new_ll, begin, new_ll->nr_blocks, b); b 395 drivers/md/persistent-data/dm-space-map-common.c if (*b >= old_ll->nr_blocks) b 398 drivers/md/persistent-data/dm-space-map-common.c r = sm_ll_lookup(old_ll, *b, &count); b 403 drivers/md/persistent-data/dm-space-map-common.c begin = *b + 1; b 410 drivers/md/persistent-data/dm-space-map-common.c static int sm_ll_mutate(struct ll_disk *ll, dm_block_t b, b 417 drivers/md/persistent-data/dm-space-map-common.c dm_block_t index = b; b 439 drivers/md/persistent-data/dm-space-map-common.c r = sm_ll_lookup_big_ref_count(ll, b, &old); b 460 drivers/md/persistent-data/dm-space-map-common.c &b, &ll->ref_count_root); b 473 drivers/md/persistent-data/dm-space-map-common.c &b, &le_rc, &ll->ref_count_root); b 504 drivers/md/persistent-data/dm-space-map-common.c int sm_ll_insert(struct ll_disk *ll, dm_block_t b, b 507 drivers/md/persistent-data/dm-space-map-common.c return sm_ll_mutate(ll, b, set_ref_count, &ref_count, ev); b 516 drivers/md/persistent-data/dm-space-map-common.c int sm_ll_inc(struct ll_disk *ll, dm_block_t b, enum allocation_event *ev) b 518 drivers/md/persistent-data/dm-space-map-common.c return sm_ll_mutate(ll, b, inc_ref_count, NULL, ev); b 532 drivers/md/persistent-data/dm-space-map-common.c int sm_ll_dec(struct ll_disk *ll, dm_block_t b, enum allocation_event *ev) b 534 drivers/md/persistent-data/dm-space-map-common.c return sm_ll_mutate(ll, b, dec_ref_count, NULL, ev); b 570 drivers/md/persistent-data/dm-space-map-common.c struct dm_block *b; b 572 drivers/md/persistent-data/dm-space-map-common.c r = dm_tm_new_block(ll->tm, &index_validator, &b); b 576 drivers/md/persistent-data/dm-space-map-common.c ll->bitmap_root = dm_block_location(b); b 578 drivers/md/persistent-data/dm-space-map-common.c dm_tm_unlock(ll->tm, b); b 607 drivers/md/persistent-data/dm-space-map-common.c struct dm_block *b; b 609 drivers/md/persistent-data/dm-space-map-common.c r = dm_tm_shadow_block(ll->tm, ll->bitmap_root, &index_validator, &b, &inc); b 613 drivers/md/persistent-data/dm-space-map-common.c memcpy(dm_block_data(b), &ll->mi_le, sizeof(ll->mi_le)); b 614 drivers/md/persistent-data/dm-space-map-common.c ll->bitmap_root = dm_block_location(b); b 616 drivers/md/persistent-data/dm-space-map-common.c dm_tm_unlock(ll->tm, b); b 108 drivers/md/persistent-data/dm-space-map-common.h int sm_ll_lookup_bitmap(struct ll_disk *ll, dm_block_t b, uint32_t *result); b 109 drivers/md/persistent-data/dm-space-map-common.h int sm_ll_lookup(struct ll_disk *ll, dm_block_t b, uint32_t *result); b 114 drivers/md/persistent-data/dm-space-map-common.h int sm_ll_insert(struct ll_disk *ll, dm_block_t b, uint32_t ref_count, enum allocation_event *ev); b 115 drivers/md/persistent-data/dm-space-map-common.h int sm_ll_inc(struct ll_disk *ll, dm_block_t b, enum allocation_event *ev); b 116 drivers/md/persistent-data/dm-space-map-common.h int sm_ll_dec(struct ll_disk *ll, dm_block_t b, enum allocation_event *ev); b 64 drivers/md/persistent-data/dm-space-map-disk.c static int sm_disk_get_count(struct dm_space_map *sm, dm_block_t b, b 68 drivers/md/persistent-data/dm-space-map-disk.c return sm_ll_lookup(&smd->ll, b, result); b 71 drivers/md/persistent-data/dm-space-map-disk.c static int sm_disk_count_is_more_than_one(struct dm_space_map *sm, dm_block_t b, b 77 drivers/md/persistent-data/dm-space-map-disk.c r = sm_disk_get_count(sm, b, &count); b 86 drivers/md/persistent-data/dm-space-map-disk.c static int sm_disk_set_count(struct dm_space_map *sm, dm_block_t b, b 94 drivers/md/persistent-data/dm-space-map-disk.c r = sm_ll_insert(&smd->ll, b, count, &ev); b 113 drivers/md/persistent-data/dm-space-map-disk.c r = sm_ll_lookup(&smd->old_ll, b, &old_count); b 126 drivers/md/persistent-data/dm-space-map-disk.c static int sm_disk_inc_block(struct dm_space_map *sm, dm_block_t b) b 132 drivers/md/persistent-data/dm-space-map-disk.c r = sm_ll_inc(&smd->ll, b, &ev); b 143 drivers/md/persistent-data/dm-space-map-disk.c static int sm_disk_dec_block(struct dm_space_map *sm, dm_block_t b) b 150 drivers/md/persistent-data/dm-space-map-disk.c r = sm_ll_dec(&smd->ll, b, &ev); b 156 drivers/md/persistent-data/dm-space-map-disk.c r = sm_ll_lookup(&smd->old_ll, b, &old_count); b 164 drivers/md/persistent-data/dm-space-map-disk.c static int sm_disk_new_block(struct dm_space_map *sm, dm_block_t *b) b 173 drivers/md/persistent-data/dm-space-map-disk.c r = sm_ll_find_common_free_block(&smd->old_ll, &smd->ll, smd->begin, smd->ll.nr_blocks, b); b 177 drivers/md/persistent-data/dm-space-map-disk.c smd->begin = *b + 1; b 178 drivers/md/persistent-data/dm-space-map-disk.c r = sm_ll_inc(&smd->ll, *b, &ev); b 119 drivers/md/persistent-data/dm-space-map-metadata.c enum block_op_type type, dm_block_t b) b 133 drivers/md/persistent-data/dm-space-map-metadata.c bop->block = b; b 181 drivers/md/persistent-data/dm-space-map-metadata.c static int add_bop(struct sm_metadata *smm, enum block_op_type type, dm_block_t b) b 183 drivers/md/persistent-data/dm-space-map-metadata.c int r = brb_push(&smm->uncommitted, type, b); b 300 drivers/md/persistent-data/dm-space-map-metadata.c static int sm_metadata_get_count(struct dm_space_map *sm, dm_block_t b, b 317 drivers/md/persistent-data/dm-space-map-metadata.c if (op->block != b) b 331 drivers/md/persistent-data/dm-space-map-metadata.c r = sm_ll_lookup(&smm->ll, b, result); b 341 drivers/md/persistent-data/dm-space-map-metadata.c dm_block_t b, int *result) b 358 drivers/md/persistent-data/dm-space-map-metadata.c if (op->block != b) b 377 drivers/md/persistent-data/dm-space-map-metadata.c r = sm_ll_lookup_bitmap(&smm->ll, b, &rc); b 392 drivers/md/persistent-data/dm-space-map-metadata.c static int sm_metadata_set_count(struct dm_space_map *sm, dm_block_t b, b 405 drivers/md/persistent-data/dm-space-map-metadata.c r = sm_ll_insert(&smm->ll, b, count, &ev); b 411 drivers/md/persistent-data/dm-space-map-metadata.c static int sm_metadata_inc_block(struct dm_space_map *sm, dm_block_t b) b 418 drivers/md/persistent-data/dm-space-map-metadata.c r = add_bop(smm, BOP_INC, b); b 421 drivers/md/persistent-data/dm-space-map-metadata.c r = sm_ll_inc(&smm->ll, b, &ev); b 428 drivers/md/persistent-data/dm-space-map-metadata.c static int sm_metadata_dec_block(struct dm_space_map *sm, dm_block_t b) b 435 drivers/md/persistent-data/dm-space-map-metadata.c r = add_bop(smm, BOP_DEC, b); b 438 drivers/md/persistent-data/dm-space-map-metadata.c r = sm_ll_dec(&smm->ll, b, &ev); b 445 drivers/md/persistent-data/dm-space-map-metadata.c static int sm_metadata_new_block_(struct dm_space_map *sm, dm_block_t *b) b 454 drivers/md/persistent-data/dm-space-map-metadata.c r = sm_ll_find_common_free_block(&smm->old_ll, &smm->ll, smm->begin, smm->ll.nr_blocks, b); b 458 drivers/md/persistent-data/dm-space-map-metadata.c smm->begin = *b + 1; b 461 drivers/md/persistent-data/dm-space-map-metadata.c r = add_bop(smm, BOP_INC, *b); b 464 drivers/md/persistent-data/dm-space-map-metadata.c r = sm_ll_inc(&smm->ll, *b, &ev); b 474 drivers/md/persistent-data/dm-space-map-metadata.c static int sm_metadata_new_block(struct dm_space_map *sm, dm_block_t *b) b 479 drivers/md/persistent-data/dm-space-map-metadata.c int r = sm_metadata_new_block_(sm, b); b 603 drivers/md/persistent-data/dm-space-map-metadata.c static int sm_bootstrap_get_count(struct dm_space_map *sm, dm_block_t b, b 608 drivers/md/persistent-data/dm-space-map-metadata.c *result = (b < smm->begin) ? 1 : 0; b 614 drivers/md/persistent-data/dm-space-map-metadata.c dm_block_t b, int *result) b 621 drivers/md/persistent-data/dm-space-map-metadata.c static int sm_bootstrap_set_count(struct dm_space_map *sm, dm_block_t b, b 629 drivers/md/persistent-data/dm-space-map-metadata.c static int sm_bootstrap_new_block(struct dm_space_map *sm, dm_block_t *b) b 639 drivers/md/persistent-data/dm-space-map-metadata.c *b = smm->begin++; b 644 drivers/md/persistent-data/dm-space-map-metadata.c static int sm_bootstrap_inc_block(struct dm_space_map *sm, dm_block_t b) b 648 drivers/md/persistent-data/dm-space-map-metadata.c return add_bop(smm, BOP_INC, b); b 651 drivers/md/persistent-data/dm-space-map-metadata.c static int sm_bootstrap_dec_block(struct dm_space_map *sm, dm_block_t b) b 655 drivers/md/persistent-data/dm-space-map-metadata.c return add_bop(smm, BOP_DEC, b); b 42 drivers/md/persistent-data/dm-space-map.h int (*get_count)(struct dm_space_map *sm, dm_block_t b, uint32_t *result); b 43 drivers/md/persistent-data/dm-space-map.h int (*count_is_more_than_one)(struct dm_space_map *sm, dm_block_t b, b 45 drivers/md/persistent-data/dm-space-map.h int (*set_count)(struct dm_space_map *sm, dm_block_t b, uint32_t count); b 49 drivers/md/persistent-data/dm-space-map.h int (*inc_block)(struct dm_space_map *sm, dm_block_t b); b 50 drivers/md/persistent-data/dm-space-map.h int (*dec_block)(struct dm_space_map *sm, dm_block_t b); b 55 drivers/md/persistent-data/dm-space-map.h int (*new_block)(struct dm_space_map *sm, dm_block_t *b); b 97 drivers/md/persistent-data/dm-space-map.h static inline int dm_sm_get_count(struct dm_space_map *sm, dm_block_t b, b 100 drivers/md/persistent-data/dm-space-map.h return sm->get_count(sm, b, result); b 104 drivers/md/persistent-data/dm-space-map.h dm_block_t b, int *result) b 106 drivers/md/persistent-data/dm-space-map.h return sm->count_is_more_than_one(sm, b, result); b 109 drivers/md/persistent-data/dm-space-map.h static inline int dm_sm_set_count(struct dm_space_map *sm, dm_block_t b, b 112 drivers/md/persistent-data/dm-space-map.h return sm->set_count(sm, b, count); b 120 drivers/md/persistent-data/dm-space-map.h static inline int dm_sm_inc_block(struct dm_space_map *sm, dm_block_t b) b 122 drivers/md/persistent-data/dm-space-map.h return sm->inc_block(sm, b); b 125 drivers/md/persistent-data/dm-space-map.h static inline int dm_sm_dec_block(struct dm_space_map *sm, dm_block_t b) b 127 drivers/md/persistent-data/dm-space-map.h return sm->dec_block(sm, b); b 130 drivers/md/persistent-data/dm-space-map.h static inline int dm_sm_new_block(struct dm_space_map *sm, dm_block_t *b) b 132 drivers/md/persistent-data/dm-space-map.h return sm->new_block(sm, b); b 31 drivers/md/persistent-data/dm-transaction-manager.c static unsigned prefetch_hash(dm_block_t b) b 33 drivers/md/persistent-data/dm-transaction-manager.c return hash_64(b, PREFETCH_BITS); b 49 drivers/md/persistent-data/dm-transaction-manager.c static void prefetch_add(struct prefetch_set *p, dm_block_t b) b 51 drivers/md/persistent-data/dm-transaction-manager.c unsigned h = prefetch_hash(b); b 55 drivers/md/persistent-data/dm-transaction-manager.c p->blocks[h] = b; b 103 drivers/md/persistent-data/dm-transaction-manager.c static int is_shadow(struct dm_transaction_manager *tm, dm_block_t b) b 106 drivers/md/persistent-data/dm-transaction-manager.c unsigned bucket = dm_hash_block(b, DM_HASH_MASK); b 111 drivers/md/persistent-data/dm-transaction-manager.c if (si->where == b) { b 124 drivers/md/persistent-data/dm-transaction-manager.c static void insert_shadow(struct dm_transaction_manager *tm, dm_block_t b) b 131 drivers/md/persistent-data/dm-transaction-manager.c si->where = b; b 132 drivers/md/persistent-data/dm-transaction-manager.c bucket = dm_hash_block(b, DM_HASH_MASK); b 328 drivers/md/persistent-data/dm-transaction-manager.c int dm_tm_read_lock(struct dm_transaction_manager *tm, dm_block_t b, b 333 drivers/md/persistent-data/dm-transaction-manager.c int r = dm_bm_read_try_lock(tm->real->bm, b, v, blk); b 336 drivers/md/persistent-data/dm-transaction-manager.c prefetch_add(&tm->real->prefetches, b); b 341 drivers/md/persistent-data/dm-transaction-manager.c return dm_bm_read_lock(tm->bm, b, v, blk); b 345 drivers/md/persistent-data/dm-transaction-manager.c void dm_tm_unlock(struct dm_transaction_manager *tm, struct dm_block *b) b 347 drivers/md/persistent-data/dm-transaction-manager.c dm_bm_unlock(b); b 351 drivers/md/persistent-data/dm-transaction-manager.c void dm_tm_inc(struct dm_transaction_manager *tm, dm_block_t b) b 358 drivers/md/persistent-data/dm-transaction-manager.c dm_sm_inc_block(tm->sm, b); b 362 drivers/md/persistent-data/dm-transaction-manager.c void dm_tm_dec(struct dm_transaction_manager *tm, dm_block_t b) b 369 drivers/md/persistent-data/dm-transaction-manager.c dm_sm_dec_block(tm->sm, b); b 373 drivers/md/persistent-data/dm-transaction-manager.c int dm_tm_ref(struct dm_transaction_manager *tm, dm_block_t b, b 379 drivers/md/persistent-data/dm-transaction-manager.c return dm_sm_get_count(tm->sm, b, result); b 93 drivers/md/persistent-data/dm-transaction-manager.h int dm_tm_read_lock(struct dm_transaction_manager *tm, dm_block_t b, b 97 drivers/md/persistent-data/dm-transaction-manager.h void dm_tm_unlock(struct dm_transaction_manager *tm, struct dm_block *b); b 102 drivers/md/persistent-data/dm-transaction-manager.h void dm_tm_inc(struct dm_transaction_manager *tm, dm_block_t b); b 104 drivers/md/persistent-data/dm-transaction-manager.h void dm_tm_dec(struct dm_transaction_manager *tm, dm_block_t b); b 106 drivers/md/persistent-data/dm-transaction-manager.h int dm_tm_ref(struct dm_transaction_manager *tm, dm_block_t b, b 55 drivers/md/raid0.c char b[BDEVNAME_SIZE]; b 68 drivers/md/raid0.c + k]->bdev, b)); b 87 drivers/md/raid0.c char b[BDEVNAME_SIZE]; b 98 drivers/md/raid0.c bdevname(rdev1->bdev, b)); b 113 drivers/md/raid0.c bdevname(rdev1->bdev,b), b 260 drivers/md/raid0.c bdevname(rdev->bdev, b)); b 266 drivers/md/raid0.c bdevname(rdev->bdev, b), c); b 381 drivers/md/raid1.c char b[BDEVNAME_SIZE]; b 384 drivers/md/raid1.c bdevname(rdev->bdev, b), b 1224 drivers/md/raid1.c char b[BDEVNAME_SIZE]; b 1239 drivers/md/raid1.c bdevname(rdev->bdev, b); b 1241 drivers/md/raid1.c strcpy(b, "???"); b 1268 drivers/md/raid1.c b, b 1280 drivers/md/raid1.c bdevname(mirror->rdev->bdev, b)); b 1613 drivers/md/raid1.c char b[BDEVNAME_SIZE]; b 1649 drivers/md/raid1.c mdname(mddev), bdevname(rdev->bdev, b), b 1667 drivers/md/raid1.c char b[BDEVNAME_SIZE]; b 1673 drivers/md/raid1.c bdevname(rdev->bdev,b)); b 2034 drivers/md/raid1.c char b[BDEVNAME_SIZE]; b 2042 drivers/md/raid1.c mdname(mddev), bio_devname(bio, b), b 2123 drivers/md/raid1.c struct bio *b = r1_bio->bios[i]; b 2124 drivers/md/raid1.c struct resync_pages *rp = get_resync_pages(b); b 2125 drivers/md/raid1.c if (b->bi_end_io != end_sync_read) b 2128 drivers/md/raid1.c status = b->bi_status; b 2129 drivers/md/raid1.c bio_reset(b); b 2130 drivers/md/raid1.c b->bi_status = status; b 2131 drivers/md/raid1.c b->bi_iter.bi_sector = r1_bio->sector + b 2133 drivers/md/raid1.c bio_set_dev(b, conf->mirrors[i].rdev->bdev); b 2134 drivers/md/raid1.c b->bi_end_io = end_sync_read; b 2136 drivers/md/raid1.c b->bi_private = rp; b 2139 drivers/md/raid1.c md_bio_reset_resync_pages(b, rp, r1_bio->sectors << 9); b 2312 drivers/md/raid1.c char b[BDEVNAME_SIZE]; b 2329 drivers/md/raid1.c bdevname(rdev->bdev, b)); b 392 drivers/md/raid10.c char b[BDEVNAME_SIZE]; b 395 drivers/md/raid10.c bdevname(rdev->bdev, b), b 1143 drivers/md/raid10.c char b[BDEVNAME_SIZE]; b 1167 drivers/md/raid10.c bdevname(err_rdev->bdev, b); b 1169 drivers/md/raid10.c strcpy(b, "???"); b 1181 drivers/md/raid10.c mdname(mddev), b, b 1190 drivers/md/raid10.c bdevname(rdev->bdev, b), b 1636 drivers/md/raid10.c char b[BDEVNAME_SIZE]; b 1668 drivers/md/raid10.c mdname(mddev), bdevname(rdev->bdev, b), b 1688 drivers/md/raid10.c char b[BDEVNAME_SIZE]; b 1694 drivers/md/raid10.c bdevname(rdev->bdev,b)); b 2351 drivers/md/raid10.c char b[BDEVNAME_SIZE]; b 2352 drivers/md/raid10.c bdevname(rdev->bdev, b); b 2355 drivers/md/raid10.c mdname(mddev), b, b 2358 drivers/md/raid10.c mdname(mddev), b); b 2428 drivers/md/raid10.c char b[BDEVNAME_SIZE]; b 2454 drivers/md/raid10.c bdevname(rdev->bdev, b)); b 2457 drivers/md/raid10.c bdevname(rdev->bdev, b)); b 2464 drivers/md/raid10.c char b[BDEVNAME_SIZE]; b 2490 drivers/md/raid10.c bdevname(rdev->bdev, b)); b 2493 drivers/md/raid10.c bdevname(rdev->bdev, b)); b 2501 drivers/md/raid10.c bdevname(rdev->bdev, b)); b 4609 drivers/md/raid10.c struct bio *b; b 4614 drivers/md/raid10.c b = r10_bio->devs[s/2].repl_bio; b 4617 drivers/md/raid10.c b = r10_bio->devs[s/2].bio; b 4622 drivers/md/raid10.c bio_set_dev(b, rdev2->bdev); b 4623 drivers/md/raid10.c b->bi_iter.bi_sector = r10_bio->devs[s/2].addr + b 4625 drivers/md/raid10.c b->bi_end_io = end_reshape_write; b 4626 drivers/md/raid10.c bio_set_op_attrs(b, REQ_OP_WRITE, 0); b 4627 drivers/md/raid10.c b->bi_next = blist; b 4628 drivers/md/raid10.c blist = b; b 4700 drivers/md/raid10.c struct bio *b; b 4706 drivers/md/raid10.c b = r10_bio->devs[s/2].repl_bio; b 4709 drivers/md/raid10.c b = r10_bio->devs[s/2].bio; b 4717 drivers/md/raid10.c md_sync_acct_bio(b, r10_bio->sectors); b 4719 drivers/md/raid10.c b->bi_next = NULL; b 4720 drivers/md/raid10.c generic_make_request(b); b 3068 drivers/md/raid5-cache.c char b[BDEVNAME_SIZE]; b 3072 drivers/md/raid5-cache.c mdname(conf->mddev), bdevname(rdev->bdev, b)); b 419 drivers/md/raid5-ppl.c char b[BDEVNAME_SIZE]; b 424 drivers/md/raid5-ppl.c bio_devname(bio, b)); b 593 drivers/md/raid5-ppl.c char b[BDEVNAME_SIZE]; b 595 drivers/md/raid5-ppl.c pr_debug("%s: dev: %s\n", __func__, bio_devname(bio, b)); b 638 drivers/md/raid5-ppl.c char b[BDEVNAME_SIZE]; b 647 drivers/md/raid5-ppl.c bio_devname(bio, b)); b 810 drivers/md/raid5-ppl.c char b[BDEVNAME_SIZE]; b 905 drivers/md/raid5-ppl.c __func__, indent, "", bdevname(rdev->bdev, b), b 952 drivers/md/raid5-ppl.c bdevname(parity_rdev->bdev, b)); b 1264 drivers/md/raid5-ppl.c char b[BDEVNAME_SIZE]; b 1281 drivers/md/raid5-ppl.c mdname(rdev->mddev), bdevname(rdev->bdev, b)); b 1292 drivers/md/raid5-ppl.c mdname(rdev->mddev), bdevname(rdev->bdev, b)); b 1300 drivers/md/raid5-ppl.c mdname(rdev->mddev), bdevname(rdev->bdev, b)); b 1470 drivers/md/raid5-ppl.c char b[BDEVNAME_SIZE]; b 1477 drivers/md/raid5-ppl.c bdevname(rdev->bdev, b)); b 879 drivers/md/raid5.c static int cmp_stripe(void *priv, struct list_head *a, struct list_head *b) b 883 drivers/md/raid5.c const struct r5pending_data *db = list_entry(b, b 2466 drivers/md/raid5.c char b[BDEVNAME_SIZE]; b 2507 drivers/md/raid5.c bdevname(rdev->bdev, b)); b 2524 drivers/md/raid5.c const char *bdn = bdevname(rdev->bdev, b); b 2673 drivers/md/raid5.c char b[BDEVNAME_SIZE]; b 2703 drivers/md/raid5.c bdevname(rdev->bdev, b), b 7041 drivers/md/raid5.c char b[BDEVNAME_SIZE]; b 7043 drivers/md/raid5.c mdname(mddev), bdevname(rdev->bdev, b), raid_disk); b 7543 drivers/md/raid5.c char b[BDEVNAME_SIZE]; b 7548 drivers/md/raid5.c bdevname(tmp->rdev->bdev, b)); b 276 drivers/media/common/b2c2/flexcop.c u8 *b = fc->dvb_adapter.proposed_mac; b 277 drivers/media/common/b2c2/flexcop.c info("MAC address = %pM", b); b 278 drivers/media/common/b2c2/flexcop.c flexcop_set_mac_filter(fc,b); b 428 drivers/media/common/cx2341x.c int b = ctrl->value + 1; b 431 drivers/media/common/cx2341x.c params->video_gop_size = b * ((gop + b - 1) / b); b 434 drivers/media/common/cx2341x.c params->video_gop_size -= b; b 438 drivers/media/common/cx2341x.c int b = params->video_b_frames + 1; b 440 drivers/media/common/cx2341x.c params->video_gop_size = b * ((gop + b - 1) / b); b 443 drivers/media/common/cx2341x.c params->video_gop_size -= b; b 1329 drivers/media/common/cx2341x.c int b = val + 1; b 1332 drivers/media/common/cx2341x.c gop = b * ((gop + b - 1) / b); b 1336 drivers/media/common/cx2341x.c gop -= b; b 44 drivers/media/common/cypress_firmware.c u8 *b = (u8 *) &fw->data[*pos]; b 51 drivers/media/common/cypress_firmware.c hx->len = b[0]; b 56 drivers/media/common/cypress_firmware.c hx->addr = b[1] | (b[2] << 8); b 57 drivers/media/common/cypress_firmware.c hx->type = b[3]; b 62 drivers/media/common/cypress_firmware.c hx->addr |= (b[4] << 24) | (b[5] << 16); b 65 drivers/media/common/cypress_firmware.c memcpy(hx->data, &b[data_offs], hx->len); b 66 drivers/media/common/cypress_firmware.c hx->chk = b[hx->len + data_offs]; b 364 drivers/media/common/saa7146/saa7146_hlp.c int l = 0, r = 0, t = 0, b = 0; b 390 drivers/media/common/saa7146/saa7146_hlp.c b = y[i]+h[i]; b 397 drivers/media/common/saa7146/saa7146_hlp.c line_list[(2*i)+1] = min_t(int, b, height); b 874 drivers/media/common/saa7146/saa7146_video.c static int vidioc_reqbufs(struct file *file, void *__fh, struct v4l2_requestbuffers *b) b 878 drivers/media/common/saa7146/saa7146_video.c if (b->type == V4L2_BUF_TYPE_VIDEO_CAPTURE) b 879 drivers/media/common/saa7146/saa7146_video.c return videobuf_reqbufs(&fh->video_q, b); b 880 drivers/media/common/saa7146/saa7146_video.c if (b->type == V4L2_BUF_TYPE_VBI_CAPTURE) b 881 drivers/media/common/saa7146/saa7146_video.c return videobuf_reqbufs(&fh->vbi_q, b); b 1153 drivers/media/common/v4l2-tpg/v4l2-tpg-colors.c static void mult_matrix(double *r, double *g, double *b, const double m[3][3]) b 1157 drivers/media/common/v4l2-tpg/v4l2-tpg-colors.c ir = m[0][0] * (*r) + m[0][1] * (*g) + m[0][2] * (*b); b 1158 drivers/media/common/v4l2-tpg/v4l2-tpg-colors.c ig = m[1][0] * (*r) + m[1][1] * (*g) + m[1][2] * (*b); b 1159 drivers/media/common/v4l2-tpg/v4l2-tpg-colors.c ib = m[2][0] * (*r) + m[2][1] * (*g) + m[2][2] * (*b); b 1162 drivers/media/common/v4l2-tpg/v4l2-tpg-colors.c *b = ib; b 1232 drivers/media/common/v4l2-tpg/v4l2-tpg-colors.c double *r, double *g, double *b) b 1238 drivers/media/common/v4l2-tpg/v4l2-tpg-colors.c *b = transfer_srgb_to_rgb(*b); b 1243 drivers/media/common/v4l2-tpg/v4l2-tpg-colors.c mult_matrix(r, g, b, rec709_to_240m); b 1246 drivers/media/common/v4l2-tpg/v4l2-tpg-colors.c mult_matrix(r, g, b, rec709_to_170m); b 1249 drivers/media/common/v4l2-tpg/v4l2-tpg-colors.c mult_matrix(r, g, b, rec709_to_ebu); b 1252 drivers/media/common/v4l2-tpg/v4l2-tpg-colors.c mult_matrix(r, g, b, rec709_to_ntsc1953); b 1255 drivers/media/common/v4l2-tpg/v4l2-tpg-colors.c mult_matrix(r, g, b, rec709_to_oprgb); b 1258 drivers/media/common/v4l2-tpg/v4l2-tpg-colors.c mult_matrix(r, g, b, rec709_to_bt2020); b 1261 drivers/media/common/v4l2-tpg/v4l2-tpg-colors.c mult_matrix(r, g, b, rec709_to_dcip3); b 1273 drivers/media/common/v4l2-tpg/v4l2-tpg-colors.c *b = ((*b) < 0) ? 0 : (((*b) > 1) ? 1 : (*b)); b 1280 drivers/media/common/v4l2-tpg/v4l2-tpg-colors.c *b = transfer_rgb_to_rec709(*b); b 1285 drivers/media/common/v4l2-tpg/v4l2-tpg-colors.c *b = transfer_rgb_to_srgb(*b); b 1290 drivers/media/common/v4l2-tpg/v4l2-tpg-colors.c *b = transfer_rgb_to_oprgb(*b); b 1295 drivers/media/common/v4l2-tpg/v4l2-tpg-colors.c *b = transfer_rgb_to_dcip3(*b); b 1300 drivers/media/common/v4l2-tpg/v4l2-tpg-colors.c *b = transfer_rgb_to_smpte2084(*b); b 1305 drivers/media/common/v4l2-tpg/v4l2-tpg-colors.c *b = transfer_rgb_to_smpte240m(*b); b 1387 drivers/media/common/v4l2-tpg/v4l2-tpg-colors.c double r, g, b; b 1394 drivers/media/common/v4l2-tpg/v4l2-tpg-colors.c b = tpg_colors[i].b / 255.0; b 1396 drivers/media/common/v4l2-tpg/v4l2-tpg-colors.c csc(c, x, &r, &g, &b); b 1401 drivers/media/common/v4l2-tpg/v4l2-tpg-colors.c (int)(r * 4080), (int)(g * 4080), (int)(b * 4080)); b 549 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c static void color_to_hsv(struct tpg_data *tpg, int r, int g, int b, b 559 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c b >>= 4; b 562 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c max_rgb = max3(r, g, b); b 571 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c min_rgb = min3(r, g, b); b 586 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c aux = g - b; b 589 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c aux = b - r; b 614 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c static void rgb2ycbcr(const int m[3][3], int r, int g, int b, b 617 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c *y = ((m[0][0] * r + m[0][1] * g + m[0][2] * b) >> 16) + (y_offset << 4); b 618 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c *cb = ((m[1][0] * r + m[1][1] * g + m[1][2] * b) >> 16) + (128 << 4); b 619 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c *cr = ((m[2][0] * r + m[2][1] * g + m[2][2] * b) >> 16) + (128 << 4); b 622 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c static void color_to_ycbcr(struct tpg_data *tpg, int r, int g, int b, b 682 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c rgb2ycbcr(full ? bt601_full : bt601, r, g, b, y_offset, y, cb, cr); b 687 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c rgb2ycbcr(bt601, r, g, b, 16, y, cb, cr); b 692 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c rgb2ycbcr(rec709, r, g, b, 16, y, cb, cr); b 695 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c rgb2ycbcr(full ? bt2020_full : bt2020, r, g, b, y_offset, y, cb, cr); b 700 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c COEFF(0.0593, 255) * rec709_to_linear(b)) >> 16; b 703 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c if (b <= yc) b 704 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c *cb = (((b - yc) * (full ? bt2020c_full[0] : bt2020c[0])) >> 16) + (128 << 4); b 706 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c *cb = (((b - yc) * (full ? bt2020c_full[1] : bt2020c[1])) >> 16) + (128 << 4); b 713 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c rgb2ycbcr(full ? smpte240m_full : smpte240m, r, g, b, y_offset, y, cb, cr); b 717 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c rgb2ycbcr(full ? rec709_full : rec709, r, g, b, y_offset, y, cb, cr); b 723 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c int y_offset, int *r, int *g, int *b) b 730 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c *b = m[2][0] * y + m[2][1] * cb + m[2][2] * cr; b 733 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c *b = clamp(*b >> 12, 0, 0xff0); b 737 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c int *r, int *g, int *b) b 797 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c ycbcr2rgb(full ? bt601_full : bt601, y, cb, cr, y_offset, r, g, b); b 802 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c ycbcr2rgb(bt601, y, cb, cr, 16, r, g, b); b 807 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c ycbcr2rgb(rec709, y, cb, cr, 16, r, g, b); b 810 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c ycbcr2rgb(full ? bt2020_full : bt2020, y, cb, cr, y_offset, r, g, b); b 818 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c *b = y_fac * y + (full ? bt2020c_full[0] : bt2020c[0]) * cb; b 820 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c *b = y_fac * y + (full ? bt2020c_full[1] : bt2020c[1]) * cb; b 821 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c *b = *b >> 12; b 828 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c lin_b = rec709_to_linear(*b); b 837 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c ycbcr2rgb(full ? smpte240m_full : smpte240m, y, cb, cr, y_offset, r, g, b); b 841 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c ycbcr2rgb(full ? rec709_full : rec709, y, cb, cr, y_offset, r, g, b); b 852 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c int b = tpg_colors[col].b; b 861 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c b = tpg_colors[col].b; b 867 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c b = tpg_colors[col].b; b 869 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c r = g = b = prandom_u32_max(256); b 871 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c r = g = b = tpg->qual_offset + prandom_u32_max(196); b 873 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c r = g = b = k - TPG_COLOR_RAMP; b 879 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c b = tpg_csc_colors[tpg->colorspace][tpg->real_xfer_func][col].b; b 883 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c b <<= 4; b 890 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c r = g = b = (13879 * r + 46688 * g + 4713 * b) >> 16; b 910 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c b = (b * 219) / 255 + (16 << 4); b 921 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c b = clamp(b, 16 << 4, 235 << 4); b 924 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c b = (b - (16 << 4)) * 255 / 219; b 935 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c color_to_ycbcr(tpg, r, g, b, &y, &cb, &cr); b 950 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c ycbcr_to_color(tpg, y, cb, cr, &r, &g, &b); b 962 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c color_to_hsv(tpg, r, g, b, &h, &s, &v); b 972 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c color_to_ycbcr(tpg, r, g, b, &y, &cb, &cr); b 1025 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c b = (b * 219) / 255 + (16 << 4); b 1031 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c b >>= 10; b 1037 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c b >>= 7; b 1050 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c b >>= 8; b 1066 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c b >>= 7; b 1071 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c b >>= 6; b 1076 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c b >>= 4; b 1082 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c tpg->colors[k][2] = b; b 2290 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c const struct v4l2_rect *b = &tpg->border; b 2304 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c if (tpg->show_border && frame_line >= b->top && b 2305 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c frame_line < b->top + b->height) { b 2306 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c unsigned bottom = b->top + b->height - 1; b 2310 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c if (frame_line == b->top || frame_line == b->top + 1 || b 2315 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c if (b->left >= c->left && b 2316 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c b->left < c->left + c->width) b 2319 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c if (b->left + b->width > c->left && b 2320 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c b->left + b->width <= c->left + c->width) b 2325 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c if (tpg->qual != TPG_QUAL_NOISE && frame_line >= b->top && b 2326 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c frame_line < b->top + b->height) { b 2638 drivers/media/common/videobuf2/videobuf2-core.c struct vb2_buffer *b; b 2651 drivers/media/common/videobuf2/videobuf2-core.c b = q->bufs[index]; b 2662 drivers/media/common/videobuf2/videobuf2-core.c b->planes[0].data_offset < buf->size) { b 2663 drivers/media/common/videobuf2/videobuf2-core.c buf->pos = b->planes[0].data_offset; b 2702 drivers/media/common/videobuf2/videobuf2-core.c struct vb2_buffer *b = q->bufs[index]; b 2715 drivers/media/common/videobuf2/videobuf2-core.c b->planes[0].bytesused = buf->pos; b 2718 drivers/media/common/videobuf2/videobuf2-core.c b->timestamp = ktime_get_ns(); b 59 drivers/media/common/videobuf2/videobuf2-v4l2.c static int __verify_planes_array(struct vb2_buffer *vb, const struct v4l2_buffer *b) b 61 drivers/media/common/videobuf2/videobuf2-v4l2.c if (!V4L2_TYPE_IS_MULTIPLANAR(b->type)) b 65 drivers/media/common/videobuf2/videobuf2-v4l2.c if (b->m.planes == NULL) { b 70 drivers/media/common/videobuf2/videobuf2-v4l2.c if (b->length < vb->num_planes || b->length > VB2_MAX_PLANES) { b 72 drivers/media/common/videobuf2/videobuf2-v4l2.c vb->num_planes, b->length); b 88 drivers/media/common/videobuf2/videobuf2-v4l2.c static int __verify_length(struct vb2_buffer *vb, const struct v4l2_buffer *b) b 94 drivers/media/common/videobuf2/videobuf2-v4l2.c if (!V4L2_TYPE_IS_OUTPUT(b->type)) b 97 drivers/media/common/videobuf2/videobuf2-v4l2.c if (V4L2_TYPE_IS_MULTIPLANAR(b->type)) { b 99 drivers/media/common/videobuf2/videobuf2-v4l2.c length = (b->memory == VB2_MEMORY_USERPTR || b 100 drivers/media/common/videobuf2/videobuf2-v4l2.c b->memory == VB2_MEMORY_DMABUF) b 101 drivers/media/common/videobuf2/videobuf2-v4l2.c ? b->m.planes[plane].length b 103 drivers/media/common/videobuf2/videobuf2-v4l2.c bytesused = b->m.planes[plane].bytesused b 104 drivers/media/common/videobuf2/videobuf2-v4l2.c ? b->m.planes[plane].bytesused : length; b 106 drivers/media/common/videobuf2/videobuf2-v4l2.c if (b->m.planes[plane].bytesused > length) b 109 drivers/media/common/videobuf2/videobuf2-v4l2.c if (b->m.planes[plane].data_offset > 0 && b 110 drivers/media/common/videobuf2/videobuf2-v4l2.c b->m.planes[plane].data_offset >= bytesused) b 114 drivers/media/common/videobuf2/videobuf2-v4l2.c length = (b->memory == VB2_MEMORY_USERPTR) b 115 drivers/media/common/videobuf2/videobuf2-v4l2.c ? b->length : vb->planes[0].length; b 117 drivers/media/common/videobuf2/videobuf2-v4l2.c if (b->bytesused > length) b 136 drivers/media/common/videobuf2/videobuf2-v4l2.c const struct v4l2_buffer *b = pb; b 146 drivers/media/common/videobuf2/videobuf2-v4l2.c vb->timestamp = v4l2_timeval_to_ns(&b->timestamp); b 147 drivers/media/common/videobuf2/videobuf2-v4l2.c vbuf->flags |= b->flags & V4L2_BUF_FLAG_TIMECODE; b 148 drivers/media/common/videobuf2/videobuf2-v4l2.c if (b->flags & V4L2_BUF_FLAG_TIMECODE) b 149 drivers/media/common/videobuf2/videobuf2-v4l2.c vbuf->timecode = b->timecode; b 169 drivers/media/common/videobuf2/videobuf2-v4l2.c static int vb2_fill_vb2_v4l2_buffer(struct vb2_buffer *vb, struct v4l2_buffer *b) b 177 drivers/media/common/videobuf2/videobuf2-v4l2.c ret = __verify_length(vb, b); b 182 drivers/media/common/videobuf2/videobuf2-v4l2.c if (b->field == V4L2_FIELD_ALTERNATE && q->is_output) { b 198 drivers/media/common/videobuf2/videobuf2-v4l2.c if (V4L2_TYPE_IS_MULTIPLANAR(b->type)) { b 199 drivers/media/common/videobuf2/videobuf2-v4l2.c switch (b->memory) { b 203 drivers/media/common/videobuf2/videobuf2-v4l2.c b->m.planes[plane].m.userptr; b 205 drivers/media/common/videobuf2/videobuf2-v4l2.c b->m.planes[plane].length; b 211 drivers/media/common/videobuf2/videobuf2-v4l2.c b->m.planes[plane].m.fd; b 213 drivers/media/common/videobuf2/videobuf2-v4l2.c b->m.planes[plane].length; b 227 drivers/media/common/videobuf2/videobuf2-v4l2.c if (V4L2_TYPE_IS_OUTPUT(b->type)) { b 246 drivers/media/common/videobuf2/videobuf2-v4l2.c struct v4l2_plane *psrc = &b->m.planes[plane]; b 274 drivers/media/common/videobuf2/videobuf2-v4l2.c switch (b->memory) { b 276 drivers/media/common/videobuf2/videobuf2-v4l2.c planes[0].m.userptr = b->m.userptr; b 277 drivers/media/common/videobuf2/videobuf2-v4l2.c planes[0].length = b->length; b 280 drivers/media/common/videobuf2/videobuf2-v4l2.c planes[0].m.fd = b->m.fd; b 281 drivers/media/common/videobuf2/videobuf2-v4l2.c planes[0].length = b->length; b 290 drivers/media/common/videobuf2/videobuf2-v4l2.c if (V4L2_TYPE_IS_OUTPUT(b->type)) { b 291 drivers/media/common/videobuf2/videobuf2-v4l2.c if (b->bytesused == 0) b 295 drivers/media/common/videobuf2/videobuf2-v4l2.c planes[0].bytesused = b->bytesused; b 297 drivers/media/common/videobuf2/videobuf2-v4l2.c planes[0].bytesused = b->bytesused ? b 298 drivers/media/common/videobuf2/videobuf2-v4l2.c b->bytesused : planes[0].length; b 305 drivers/media/common/videobuf2/videobuf2-v4l2.c vbuf->flags = b->flags & ~V4L2_BUFFER_MASK_FLAGS; b 306 drivers/media/common/videobuf2/videobuf2-v4l2.c if (!vb->vb2_queue->copy_timestamp || !V4L2_TYPE_IS_OUTPUT(b->type)) { b 315 drivers/media/common/videobuf2/videobuf2-v4l2.c if (V4L2_TYPE_IS_OUTPUT(b->type)) { b 323 drivers/media/common/videobuf2/videobuf2-v4l2.c vbuf->field = b->field; b 335 drivers/media/common/videobuf2/videobuf2-v4l2.c struct v4l2_buffer *b, bool is_prepare, b 344 drivers/media/common/videobuf2/videobuf2-v4l2.c if (b->type != q->type) { b 349 drivers/media/common/videobuf2/videobuf2-v4l2.c if (b->index >= q->num_buffers) { b 354 drivers/media/common/videobuf2/videobuf2-v4l2.c if (q->bufs[b->index] == NULL) { b 360 drivers/media/common/videobuf2/videobuf2-v4l2.c if (b->memory != q->memory) { b 365 drivers/media/common/videobuf2/videobuf2-v4l2.c vb = q->bufs[b->index]; b 367 drivers/media/common/videobuf2/videobuf2-v4l2.c ret = __verify_planes_array(vb, b); b 371 drivers/media/common/videobuf2/videobuf2-v4l2.c if (!is_prepare && (b->flags & V4L2_BUF_FLAG_REQUEST_FD) && b 381 drivers/media/common/videobuf2/videobuf2-v4l2.c ret = vb2_fill_vb2_v4l2_buffer(vb, b); b 389 drivers/media/common/videobuf2/videobuf2-v4l2.c if (!(b->flags & V4L2_BUF_FLAG_REQUEST_FD)) { b 432 drivers/media/common/videobuf2/videobuf2-v4l2.c if (b->request_fd < 0) { b 437 drivers/media/common/videobuf2/videobuf2-v4l2.c req = media_request_get_by_fd(mdev, b->request_fd); b 455 drivers/media/common/videobuf2/videobuf2-v4l2.c vbuf->request_fd = b->request_fd; b 466 drivers/media/common/videobuf2/videobuf2-v4l2.c struct v4l2_buffer *b = pb; b 472 drivers/media/common/videobuf2/videobuf2-v4l2.c b->index = vb->index; b 473 drivers/media/common/videobuf2/videobuf2-v4l2.c b->type = vb->type; b 474 drivers/media/common/videobuf2/videobuf2-v4l2.c b->memory = vb->memory; b 475 drivers/media/common/videobuf2/videobuf2-v4l2.c b->bytesused = 0; b 477 drivers/media/common/videobuf2/videobuf2-v4l2.c b->flags = vbuf->flags; b 478 drivers/media/common/videobuf2/videobuf2-v4l2.c b->field = vbuf->field; b 479 drivers/media/common/videobuf2/videobuf2-v4l2.c b->timestamp = ns_to_timeval(vb->timestamp); b 480 drivers/media/common/videobuf2/videobuf2-v4l2.c b->timecode = vbuf->timecode; b 481 drivers/media/common/videobuf2/videobuf2-v4l2.c b->sequence = vbuf->sequence; b 482 drivers/media/common/videobuf2/videobuf2-v4l2.c b->reserved2 = 0; b 483 drivers/media/common/videobuf2/videobuf2-v4l2.c b->request_fd = 0; b 490 drivers/media/common/videobuf2/videobuf2-v4l2.c b->length = vb->num_planes; b 492 drivers/media/common/videobuf2/videobuf2-v4l2.c struct v4l2_plane *pdst = &b->m.planes[plane]; b 511 drivers/media/common/videobuf2/videobuf2-v4l2.c b->length = vb->planes[0].length; b 512 drivers/media/common/videobuf2/videobuf2-v4l2.c b->bytesused = vb->planes[0].bytesused; b 514 drivers/media/common/videobuf2/videobuf2-v4l2.c b->m.offset = vb->planes[0].m.offset; b 516 drivers/media/common/videobuf2/videobuf2-v4l2.c b->m.userptr = vb->planes[0].m.userptr; b 518 drivers/media/common/videobuf2/videobuf2-v4l2.c b->m.fd = vb->planes[0].m.fd; b 524 drivers/media/common/videobuf2/videobuf2-v4l2.c b->flags &= ~V4L2_BUFFER_MASK_FLAGS; b 525 drivers/media/common/videobuf2/videobuf2-v4l2.c b->flags |= q->timestamp_flags & V4L2_BUF_FLAG_TIMESTAMP_MASK; b 531 drivers/media/common/videobuf2/videobuf2-v4l2.c b->flags &= ~V4L2_BUF_FLAG_TSTAMP_SRC_MASK; b 532 drivers/media/common/videobuf2/videobuf2-v4l2.c b->flags |= q->timestamp_flags & V4L2_BUF_FLAG_TSTAMP_SRC_MASK; b 538 drivers/media/common/videobuf2/videobuf2-v4l2.c b->flags |= V4L2_BUF_FLAG_QUEUED; b 541 drivers/media/common/videobuf2/videobuf2-v4l2.c b->flags |= V4L2_BUF_FLAG_IN_REQUEST; b 544 drivers/media/common/videobuf2/videobuf2-v4l2.c b->flags |= V4L2_BUF_FLAG_ERROR; b 547 drivers/media/common/videobuf2/videobuf2-v4l2.c b->flags |= V4L2_BUF_FLAG_DONE; b 558 drivers/media/common/videobuf2/videobuf2-v4l2.c b->flags |= V4L2_BUF_FLAG_PREPARED; b 561 drivers/media/common/videobuf2/videobuf2-v4l2.c b->flags |= V4L2_BUF_FLAG_MAPPED; b 563 drivers/media/common/videobuf2/videobuf2-v4l2.c b->flags |= V4L2_BUF_FLAG_REQUEST_FD; b 564 drivers/media/common/videobuf2/videobuf2-v4l2.c b->request_fd = vbuf->request_fd; b 626 drivers/media/common/videobuf2/videobuf2-v4l2.c int vb2_querybuf(struct vb2_queue *q, struct v4l2_buffer *b) b 631 drivers/media/common/videobuf2/videobuf2-v4l2.c if (b->type != q->type) { b 636 drivers/media/common/videobuf2/videobuf2-v4l2.c if (b->index >= q->num_buffers) { b 640 drivers/media/common/videobuf2/videobuf2-v4l2.c vb = q->bufs[b->index]; b 641 drivers/media/common/videobuf2/videobuf2-v4l2.c ret = __verify_planes_array(vb, b); b 643 drivers/media/common/videobuf2/videobuf2-v4l2.c vb2_core_querybuf(q, b->index, b); b 673 drivers/media/common/videobuf2/videobuf2-v4l2.c struct v4l2_buffer *b) b 682 drivers/media/common/videobuf2/videobuf2-v4l2.c if (b->flags & V4L2_BUF_FLAG_REQUEST_FD) b 685 drivers/media/common/videobuf2/videobuf2-v4l2.c ret = vb2_queue_or_prepare_buf(q, mdev, b, true, NULL); b 687 drivers/media/common/videobuf2/videobuf2-v4l2.c return ret ? ret : vb2_core_prepare_buf(q, b->index, b); b 748 drivers/media/common/videobuf2/videobuf2-v4l2.c struct v4l2_buffer *b) b 758 drivers/media/common/videobuf2/videobuf2-v4l2.c ret = vb2_queue_or_prepare_buf(q, mdev, b, false, &req); b 761 drivers/media/common/videobuf2/videobuf2-v4l2.c ret = vb2_core_qbuf(q, b->index, b, req); b 768 drivers/media/common/videobuf2/videobuf2-v4l2.c int vb2_dqbuf(struct vb2_queue *q, struct v4l2_buffer *b, bool nonblocking) b 777 drivers/media/common/videobuf2/videobuf2-v4l2.c if (b->type != q->type) { b 782 drivers/media/common/videobuf2/videobuf2-v4l2.c ret = vb2_core_dqbuf(q, NULL, b, nonblocking); b 785 drivers/media/common/videobuf2/videobuf2-v4l2.c b->flags & V4L2_BUF_FLAG_DONE && b 786 drivers/media/common/videobuf2/videobuf2-v4l2.c b->flags & V4L2_BUF_FLAG_LAST) b 793 drivers/media/common/videobuf2/videobuf2-v4l2.c b->flags &= ~V4L2_BUF_FLAG_DONE; b 1061 drivers/media/dvb-core/dvb_frontend.c #define _DTV_CMD(n, s, b) \ b 1066 drivers/media/dvb-core/dvb_frontend.c .buffer = b \ b 140 drivers/media/dvb-core/dvb_vb2.c struct dmx_buffer *b = pb; b 142 drivers/media/dvb-core/dvb_vb2.c b->index = vb->index; b 143 drivers/media/dvb-core/dvb_vb2.c b->length = vb->planes[0].length; b 144 drivers/media/dvb-core/dvb_vb2.c b->bytesused = vb->planes[0].bytesused; b 145 drivers/media/dvb-core/dvb_vb2.c b->offset = vb->planes[0].m.offset; b 359 drivers/media/dvb-core/dvb_vb2.c int dvb_vb2_querybuf(struct dvb_vb2_ctx *ctx, struct dmx_buffer *b) b 361 drivers/media/dvb-core/dvb_vb2.c vb2_core_querybuf(&ctx->vb_q, b->index, b); b 362 drivers/media/dvb-core/dvb_vb2.c dprintk(3, "[%s] index=%d\n", ctx->name, b->index); b 383 drivers/media/dvb-core/dvb_vb2.c int dvb_vb2_qbuf(struct dvb_vb2_ctx *ctx, struct dmx_buffer *b) b 387 drivers/media/dvb-core/dvb_vb2.c ret = vb2_core_qbuf(&ctx->vb_q, b->index, b, NULL); b 390 drivers/media/dvb-core/dvb_vb2.c b->index, ret); b 393 drivers/media/dvb-core/dvb_vb2.c dprintk(5, "[%s] index=%d\n", ctx->name, b->index); b 398 drivers/media/dvb-core/dvb_vb2.c int dvb_vb2_dqbuf(struct dvb_vb2_ctx *ctx, struct dmx_buffer *b) b 403 drivers/media/dvb-core/dvb_vb2.c ret = vb2_core_dqbuf(&ctx->vb_q, &b->index, b, ctx->nonblocking); b 410 drivers/media/dvb-core/dvb_vb2.c b->count = ctx->count++; b 411 drivers/media/dvb-core/dvb_vb2.c b->flags = ctx->flags; b 416 drivers/media/dvb-core/dvb_vb2.c ctx->name, b->index, ctx->count, b->flags); b 71 drivers/media/dvb-frontends/bcm3510.c #define dbufout(b,l,m) {\ b 74 drivers/media/dvb-frontends/bcm3510.c m("%02x ",b[i]); \ b 83 drivers/media/dvb-frontends/bcm3510.c u8 b[256]; b 85 drivers/media/dvb-frontends/bcm3510.c struct i2c_msg msg = { .addr = state->config->demod_address, .flags = 0, .buf = b, .len = len + 1 }; b 87 drivers/media/dvb-frontends/bcm3510.c b[0] = reg; b 88 drivers/media/dvb-frontends/bcm3510.c memcpy(&b[1],buf,len); b 267 drivers/media/dvb-frontends/bcm3510.c bcm3510_register_value b; b 270 drivers/media/dvb-frontends/bcm3510.c if ((ret = bcm3510_readB(st,0xfa,&b)) < 0) b 273 drivers/media/dvb-frontends/bcm3510.c b.BERCTL_fa.RESYNC = 0; bcm3510_writeB(st,0xfa,b); b 274 drivers/media/dvb-frontends/bcm3510.c b.BERCTL_fa.RESYNC = 1; bcm3510_writeB(st,0xfa,b); b 275 drivers/media/dvb-frontends/bcm3510.c b.BERCTL_fa.RESYNC = 0; bcm3510_writeB(st,0xfa,b); b 276 drivers/media/dvb-frontends/bcm3510.c b.BERCTL_fa.CNTCTL = 1; b.BERCTL_fa.BITCNT = 1; bcm3510_writeB(st,0xfa,b); b 610 drivers/media/dvb-frontends/bcm3510.c static int bcm3510_write_ram(struct bcm3510_state *st, u16 addr, const u8 *b, b 622 drivers/media/dvb-frontends/bcm3510.c vD.MDATA_ab = b[i]; b 635 drivers/media/dvb-frontends/bcm3510.c const u8 *b; b 645 drivers/media/dvb-frontends/bcm3510.c b = fw->data; b 647 drivers/media/dvb-frontends/bcm3510.c addr = le16_to_cpu(*((__le16 *)&b[i])); b 648 drivers/media/dvb-frontends/bcm3510.c len = le16_to_cpu(*((__le16 *)&b[i+2])); b 650 drivers/media/dvb-frontends/bcm3510.c if ((ret = bcm3510_write_ram(st,addr,&b[i+4],len)) < 0) { b 108 drivers/media/dvb-frontends/cx24113.c u8 b; b 113 drivers/media/dvb-frontends/cx24113.c .flags = I2C_M_RD, .buf = &b, .len = 1 } b 124 drivers/media/dvb-frontends/cx24113.c return b; b 256 drivers/media/dvb-frontends/cx24123.c u8 b = 0; b 259 drivers/media/dvb-frontends/cx24123.c { .addr = i2c_addr, .flags = I2C_M_RD, .buf = &b, .len = 1 } b 271 drivers/media/dvb-frontends/cx24123.c return b; b 428 drivers/media/dvb-frontends/cx24123.c static u32 cx24123_int_log2(u32 a, u32 b) b 431 drivers/media/dvb-frontends/cx24123.c u32 div = a / b; b 432 drivers/media/dvb-frontends/cx24123.c if (a % b >= b / 2) b 320 drivers/media/dvb-frontends/dib0090.c static void dib0090_write_regs(struct dib0090_state *state, u8 r, const u16 * b, u8 c) b 323 drivers/media/dvb-frontends/dib0090.c dib0090_write_reg(state, r++, *b++); b 71 drivers/media/dvb-frontends/dib3000mb.c u8 b[] = { b 76 drivers/media/dvb-frontends/dib3000mb.c { .addr = state->config.demod_address, .flags = 0, .buf = b, .len = 4 } b 60 drivers/media/dvb-frontends/dib3000mc.c u8 *b; b 62 drivers/media/dvb-frontends/dib3000mc.c b = kmalloc(4, GFP_KERNEL); b 63 drivers/media/dvb-frontends/dib3000mc.c if (!b) b 66 drivers/media/dvb-frontends/dib3000mc.c b[0] = (reg >> 8) | 0x80; b 67 drivers/media/dvb-frontends/dib3000mc.c b[1] = reg; b 68 drivers/media/dvb-frontends/dib3000mc.c b[2] = 0; b 69 drivers/media/dvb-frontends/dib3000mc.c b[3] = 0; b 71 drivers/media/dvb-frontends/dib3000mc.c msg[0].buf = b; b 72 drivers/media/dvb-frontends/dib3000mc.c msg[1].buf = b + 2; b 77 drivers/media/dvb-frontends/dib3000mc.c word = (b[2] << 8) | b[3]; b 78 drivers/media/dvb-frontends/dib3000mc.c kfree(b); b 89 drivers/media/dvb-frontends/dib3000mc.c u8 *b; b 91 drivers/media/dvb-frontends/dib3000mc.c b = kmalloc(4, GFP_KERNEL); b 92 drivers/media/dvb-frontends/dib3000mc.c if (!b) b 95 drivers/media/dvb-frontends/dib3000mc.c b[0] = reg >> 8; b 96 drivers/media/dvb-frontends/dib3000mc.c b[1] = reg; b 97 drivers/media/dvb-frontends/dib3000mc.c b[2] = val >> 8; b 98 drivers/media/dvb-frontends/dib3000mc.c b[3] = val; b 100 drivers/media/dvb-frontends/dib3000mc.c msg.buf = b; b 103 drivers/media/dvb-frontends/dib3000mc.c kfree(b); b 736 drivers/media/dvb-frontends/dib7000m.c u16 b[9] = { 676, 696, 717, 737, 758, 778, 799, 819, 840 }; b 738 drivers/media/dvb-frontends/dib7000m.c dib7000m_write_word(state, 88 + i, b[i]); b 208 drivers/media/dvb-frontends/dib9000.c static int dib9000_risc_apb_access_read(struct dib9000_state *state, u32 address, u16 attribute, const u8 * tx, u32 txlen, u8 * b, u32 len); b 209 drivers/media/dvb-frontends/dib9000.c static int dib9000_risc_apb_access_write(struct dib9000_state *state, u32 address, u16 attribute, const u8 * b, u32 len); b 233 drivers/media/dvb-frontends/dib9000.c static int dib9000_read16_attr(struct dib9000_state *state, u16 reg, u8 *b, u32 len, u16 attribute) b 240 drivers/media/dvb-frontends/dib9000.c return dib9000_risc_apb_access_read(state, reg, attribute, NULL, 0, b, len); b 249 drivers/media/dvb-frontends/dib9000.c state->msg[1].buf = b; b 263 drivers/media/dvb-frontends/dib9000.c state->msg[1].buf = b; b 270 drivers/media/dvb-frontends/dib9000.c b += l; b 315 drivers/media/dvb-frontends/dib9000.c #define dib9000_read16_noinc_attr(state, reg, b, len, attribute) dib9000_read16_attr(state, reg, b, len, (attribute) | DATA_BUS_ACCESS_MODE_NO_ADDRESS_INCREMENT) b 378 drivers/media/dvb-frontends/dib9000.c u8 b[2] = { val >> 8, val & 0xff }; b 379 drivers/media/dvb-frontends/dib9000.c return dib9000_write16_attr(state, reg, b, 2, 0); b 384 drivers/media/dvb-frontends/dib9000.c u8 b[2] = { val >> 8, val & 0xff }; b 385 drivers/media/dvb-frontends/dib9000.c return dib9000_write16_attr(state, reg, b, 2, attribute); b 398 drivers/media/dvb-frontends/dib9000.c #define dib9000_risc_mem_read_chunks(state, b, len) dib9000_read16_attr(state, 1063, b, len, DATA_BUS_ACCESS_MODE_8BIT | DATA_BUS_ACCESS_MODE_NO_ADDRESS_INCREMENT) b 403 drivers/media/dvb-frontends/dib9000.c u8 b[14] = { 0 }; b 407 drivers/media/dvb-frontends/dib9000.c b[1] = 1; b 411 drivers/media/dvb-frontends/dib9000.c b[4] = (u8) (addr >> 8); b 412 drivers/media/dvb-frontends/dib9000.c b[5] = (u8) (addr & 0xff); b 416 drivers/media/dvb-frontends/dib9000.c b[12] = (u8) (addr >> 8); b 417 drivers/media/dvb-frontends/dib9000.c b[13] = (u8) (addr & 0xff); b 422 drivers/media/dvb-frontends/dib9000.c b[8] = (u8) (addr >> 8); b 423 drivers/media/dvb-frontends/dib9000.c b[9] = (u8) (addr & 0xff); b 425 drivers/media/dvb-frontends/dib9000.c dib9000_write(state, 1056, b, 14); b 442 drivers/media/dvb-frontends/dib9000.c static int dib9000_risc_mem_read(struct dib9000_state *state, u8 cmd, u8 * b, u16 len) b 452 drivers/media/dvb-frontends/dib9000.c dib9000_risc_mem_read_chunks(state, b, len); b 457 drivers/media/dvb-frontends/dib9000.c static int dib9000_risc_mem_write(struct dib9000_state *state, u8 cmd, const u8 * b) b 468 drivers/media/dvb-frontends/dib9000.c dib9000_risc_mem_write_chunks(state, b, m->size); b 529 drivers/media/dvb-frontends/dib9000.c u8 *d, b[2]; b 571 drivers/media/dvb-frontends/dib9000.c b[0] = id; b 572 drivers/media/dvb-frontends/dib9000.c b[1] = len + 1; b 573 drivers/media/dvb-frontends/dib9000.c if (dib9000_write16_noinc_attr(state, 1045, b, 2, attr) != 0 || dib9000_write16_noinc_attr(state, 1045, (u8 *) data, len * 2, attr) != 0) { b 649 drivers/media/dvb-frontends/dib9000.c char *b = (char *)&data[2]; b 651 drivers/media/dvb-frontends/dib9000.c b[2 * (size - 2) - 1] = '\0'; /* Bullet proof the buffer */ b 652 drivers/media/dvb-frontends/dib9000.c if (*b == '~') { b 653 drivers/media/dvb-frontends/dib9000.c b++; b 654 drivers/media/dvb-frontends/dib9000.c dprintk("%s\n", b); b 658 drivers/media/dvb-frontends/dib9000.c ts / 10000, ts % 10000, *b ? b : "<empty>"); b 1012 drivers/media/dvb-frontends/dib9000.c static int dib9000_risc_apb_access_read(struct dib9000_state *state, u32 address, u16 attribute, const u8 * tx, u32 txlen, u8 * b, u32 len) b 1029 drivers/media/dvb-frontends/dib9000.c b[i * 2] = (mb[i + 1] >> 8) & 0xff; b 1030 drivers/media/dvb-frontends/dib9000.c b[i * 2 + 1] = (mb[i + 1]) & 0xff; b 1039 drivers/media/dvb-frontends/dib9000.c static int dib9000_risc_apb_access_write(struct dib9000_state *state, u32 address, u16 attribute, const u8 * b, u32 len) b 1054 drivers/media/dvb-frontends/dib9000.c mb[1 + i / 2] = b[i] << 8 | b[i + 1]; b 1056 drivers/media/dvb-frontends/dib9000.c mb[1 + len / 2] = b[len - 1] << 8; b 1081 drivers/media/dvb-frontends/dib9000.c u16 b[40] = { 0 }; b 1094 drivers/media/dvb-frontends/dib9000.c b[0] = (u16) f->mask; b 1095 drivers/media/dvb-frontends/dib9000.c b[1] = (u16) f->direction; b 1096 drivers/media/dvb-frontends/dib9000.c b[2] = (u16) f->value; b 1099 drivers/media/dvb-frontends/dib9000.c b[3] = (u16) f->mask; b 1100 drivers/media/dvb-frontends/dib9000.c b[4] = (u16) f->direction; b 1101 drivers/media/dvb-frontends/dib9000.c b[5] = (u16) f->value; b 1106 drivers/media/dvb-frontends/dib9000.c if (dib9000_mbx_send(state, OUT_MSG_CONF_GPIO, b, 15) != 0) b 1110 drivers/media/dvb-frontends/dib9000.c b[0] = state->chip.d9.cfg.subband.size; /* type == 0 -> GPIO - PWM not yet supported */ b 1112 drivers/media/dvb-frontends/dib9000.c b[1 + i * 4] = state->chip.d9.cfg.subband.subband[i].f_mhz; b 1113 drivers/media/dvb-frontends/dib9000.c b[2 + i * 4] = (u16) state->chip.d9.cfg.subband.subband[i].gpio.mask; b 1114 drivers/media/dvb-frontends/dib9000.c b[3 + i * 4] = (u16) state->chip.d9.cfg.subband.subband[i].gpio.direction; b 1115 drivers/media/dvb-frontends/dib9000.c b[4 + i * 4] = (u16) state->chip.d9.cfg.subband.subband[i].gpio.value; b 1117 drivers/media/dvb-frontends/dib9000.c b[1 + i * 4] = 0; /* fe_id */ b 1118 drivers/media/dvb-frontends/dib9000.c if (dib9000_mbx_send(state, OUT_MSG_SUBBAND_SEL, b, 2 + 4 * i) != 0) b 1122 drivers/media/dvb-frontends/dib9000.c b[0] = (0 << 8) | 1; b 1124 drivers/media/dvb-frontends/dib9000.c b[1] = (0 << 8) | (0); b 1125 drivers/media/dvb-frontends/dib9000.c b[2] = (u16) (((state->chip.d9.cfg.xtal_clock_khz * 1000) >> 16) & 0xffff); b 1126 drivers/media/dvb-frontends/dib9000.c b[3] = (u16) (((state->chip.d9.cfg.xtal_clock_khz * 1000)) & 0xffff); b 1127 drivers/media/dvb-frontends/dib9000.c b[4] = (u16) ((state->chip.d9.cfg.vcxo_timer >> 16) & 0xffff); b 1128 drivers/media/dvb-frontends/dib9000.c b[5] = (u16) ((state->chip.d9.cfg.vcxo_timer) & 0xffff); b 1129 drivers/media/dvb-frontends/dib9000.c b[6] = (u16) ((state->chip.d9.cfg.timing_frequency >> 16) & 0xffff); b 1130 drivers/media/dvb-frontends/dib9000.c b[7] = (u16) ((state->chip.d9.cfg.timing_frequency) & 0xffff); b 1131 drivers/media/dvb-frontends/dib9000.c b[29] = state->chip.d9.cfg.if_drives; b 1132 drivers/media/dvb-frontends/dib9000.c if (dib9000_mbx_send(state, OUT_MSG_INIT_DEMOD, b, ARRAY_SIZE(b)) != 0) b 1138 drivers/media/dvb-frontends/dib9000.c if (dib9000_mbx_get_message(state, IN_MSG_FE_FW_DL_DONE, b, &size) < 0) b 1141 drivers/media/dvb-frontends/dib9000.c if (size > ARRAY_SIZE(b)) { b 1143 drivers/media/dvb-frontends/dib9000.c (int)ARRAY_SIZE(b)); b 1148 drivers/media/dvb-frontends/dib9000.c state->platform.risc.fe_mm[i / 2].addr = b[i + 0]; b 1149 drivers/media/dvb-frontends/dib9000.c state->platform.risc.fe_mm[i / 2].size = b[i + 1]; b 1157 drivers/media/dvb-frontends/dib9000.c u8 b[9]; b 1162 drivers/media/dvb-frontends/dib9000.c b[0] = (u8) ((freq >> 0) & 0xff); b 1163 drivers/media/dvb-frontends/dib9000.c b[1] = (u8) ((freq >> 8) & 0xff); b 1164 drivers/media/dvb-frontends/dib9000.c b[2] = (u8) ((freq >> 16) & 0xff); b 1165 drivers/media/dvb-frontends/dib9000.c b[3] = (u8) ((freq >> 24) & 0xff); b 1166 drivers/media/dvb-frontends/dib9000.c b[4] = (u8) ((state->fe[0]->dtv_property_cache.bandwidth_hz / 1000 >> 0) & 0xff); b 1167 drivers/media/dvb-frontends/dib9000.c b[5] = (u8) ((state->fe[0]->dtv_property_cache.bandwidth_hz / 1000 >> 8) & 0xff); b 1168 drivers/media/dvb-frontends/dib9000.c b[6] = (u8) ((state->fe[0]->dtv_property_cache.bandwidth_hz / 1000 >> 16) & 0xff); b 1169 drivers/media/dvb-frontends/dib9000.c b[7] = (u8) ((state->fe[0]->dtv_property_cache.bandwidth_hz / 1000 >> 24) & 0xff); b 1170 drivers/media/dvb-frontends/dib9000.c b[8] = 0x80; /* do not wait for CELL ID when doing autosearch */ b 1172 drivers/media/dvb-frontends/dib9000.c b[8] |= 1; b 1173 drivers/media/dvb-frontends/dib9000.c dib9000_risc_mem_write(state, FE_MM_W_CHANNEL_HEAD, b); b 102 drivers/media/dvb-frontends/dibx000_common.c const u8 *b = msg->buf; b 109 drivers/media/dvb-frontends/dibx000_common.c data = *b++ << 8; b 111 drivers/media/dvb-frontends/dibx000_common.c data |= *b++; b 142 drivers/media/dvb-frontends/dibx000_common.c u8 *b = msg->buf; b 170 drivers/media/dvb-frontends/dibx000_common.c *b++ = (da >> 8) & 0xff; b 173 drivers/media/dvb-frontends/dibx000_common.c *b++ = da & 0xff; b 1588 drivers/media/dvb-frontends/drx39xyj/drx_driver.h struct drx_aud_carrier b; b 215 drivers/media/dvb-frontends/drxd_hard.c static inline u32 MulDiv32(u32 a, u32 b, u32 c) b 219 drivers/media/dvb-frontends/drxd_hard.c tmp64 = (u64)a * (u64)b; b 159 drivers/media/dvb-frontends/drxk_hard.c static inline u32 MulDiv32(u32 a, u32 b, u32 c) b 163 drivers/media/dvb-frontends/drxk_hard.c tmp64 = (u64) a * (u64) b; b 2490 drivers/media/dvb-frontends/drxk_hard.c u32 b = 0; b 2567 drivers/media/dvb-frontends/drxk_hard.c b = log10times100(eq_reg_td_req_smb_cnt * tps_cnt); b 2571 drivers/media/dvb-frontends/drxk_hard.c i_mer = a + b - c; b 65 drivers/media/dvb-frontends/gp8psk-fe.h int (*in)(void *priv, u8 req, u16 value, u16 index, u8 *b, int blen); b 66 drivers/media/dvb-frontends/gp8psk-fe.h int (*out)(void *priv, u8 req, u16 value, u16 index, u8 *b, int blen); b 153 drivers/media/dvb-frontends/m88rs2000.c u8 b[3]; b 167 drivers/media/dvb-frontends/m88rs2000.c b[0] = (u8) (temp >> 16) & 0xff; b 168 drivers/media/dvb-frontends/m88rs2000.c b[1] = (u8) (temp >> 8) & 0xff; b 169 drivers/media/dvb-frontends/m88rs2000.c b[2] = (u8) temp & 0xff; b 171 drivers/media/dvb-frontends/m88rs2000.c ret = m88rs2000_writereg(state, 0x93, b[2]); b 172 drivers/media/dvb-frontends/m88rs2000.c ret |= m88rs2000_writereg(state, 0x94, b[1]); b 173 drivers/media/dvb-frontends/m88rs2000.c ret |= m88rs2000_writereg(state, 0x95, b[0]); b 138 drivers/media/dvb-frontends/mt312.c static inline u32 mt312_div(u32 a, u32 b) b 140 drivers/media/dvb-frontends/mt312.c return (a + (b / 2)) / b; b 789 drivers/media/dvb-frontends/nxt200x.c u8 b[3]; b 791 drivers/media/dvb-frontends/nxt200x.c nxt200x_readreg_multibyte(state, 0xE6, b, 3); b 793 drivers/media/dvb-frontends/nxt200x.c *ber = ((b[0] << 8) + b[1]) * 8; b 801 drivers/media/dvb-frontends/nxt200x.c u8 b[2]; b 805 drivers/media/dvb-frontends/nxt200x.c b[0] = 0x00; b 806 drivers/media/dvb-frontends/nxt200x.c nxt200x_writebytes(state, 0xA1, b, 1); b 809 drivers/media/dvb-frontends/nxt200x.c nxt200x_readreg_multibyte(state, 0xA6, b, 2); b 811 drivers/media/dvb-frontends/nxt200x.c temp = (b[0] << 8) | b[1]; b 821 drivers/media/dvb-frontends/nxt200x.c u8 b[2]; b 826 drivers/media/dvb-frontends/nxt200x.c b[0] = 0x00; b 827 drivers/media/dvb-frontends/nxt200x.c nxt200x_writebytes(state, 0xA1, b, 1); b 830 drivers/media/dvb-frontends/nxt200x.c nxt200x_readreg_multibyte(state, 0xA6, b, 2); b 832 drivers/media/dvb-frontends/nxt200x.c temp = (b[0] << 8) | b[1]; b 854 drivers/media/dvb-frontends/nxt200x.c u8 b[3]; b 856 drivers/media/dvb-frontends/nxt200x.c nxt200x_readreg_multibyte(state, 0xE6, b, 3); b 857 drivers/media/dvb-frontends/nxt200x.c *ucblocks = b[2]; b 11 drivers/media/dvb-frontends/rtl2832.c #define REG_MASK(b) (BIT(b + 1) - 1) b 68 drivers/media/dvb-frontends/s5h1420.c u8 b[2]; b 70 drivers/media/dvb-frontends/s5h1420.c { .addr = state->config->demod_address, .flags = 0, .buf = b, .len = 2 }, b 72 drivers/media/dvb-frontends/s5h1420.c { .addr = state->config->demod_address, .flags = I2C_M_RD, .buf = b, .len = 1 }, b 75 drivers/media/dvb-frontends/s5h1420.c b[0] = (reg - 1) & 0xff; b 76 drivers/media/dvb-frontends/s5h1420.c b[1] = state->shadow[(reg - 1) & 0xff]; b 93 drivers/media/dvb-frontends/s5h1420.c return b[0]; b 307 drivers/media/dvb-frontends/si21xx.c static int si21_readregs(struct si21xx_state *state, u8 reg1, u8 *b, u8 len) b 319 drivers/media/dvb-frontends/si21xx.c .buf = b, b 47 drivers/media/dvb-frontends/stb0899_priv.h #define MAKEWORD32(a, b, c, d) (((a) << 24) | ((b) << 16) | ((c) << 8) | (d)) b 48 drivers/media/dvb-frontends/stb0899_priv.h #define MAKEWORD16(a, b) (((a) << 8) | (b)) b 113 drivers/media/dvb-frontends/stv0288.c unsigned char b[3]; b 128 drivers/media/dvb-frontends/stv0288.c b[0] = (unsigned char)((temp >> 12) & 0xff); b 129 drivers/media/dvb-frontends/stv0288.c b[1] = (unsigned char)((temp >> 4) & 0xff); b 130 drivers/media/dvb-frontends/stv0288.c b[2] = (unsigned char)((temp << 4) & 0xf0); b 135 drivers/media/dvb-frontends/stv0288.c stv0288_writeregI(state, 0x28, b[0]); b 136 drivers/media/dvb-frontends/stv0288.c stv0288_writeregI(state, 0x29, b[1]); b 137 drivers/media/dvb-frontends/stv0288.c stv0288_writeregI(state, 0x2a, b[2]); b 95 drivers/media/dvb-frontends/stv0297.c static int stv0297_readregs(struct stv0297_state *state, u8 reg1, u8 * b, u8 len) b 100 drivers/media/dvb-frontends/stv0297.c {.addr = state->config->demod_address,.flags = I2C_M_RD,.buf = b,.len = len} b 112 drivers/media/dvb-frontends/stv0299.c static int stv0299_readregs (struct stv0299_state* state, u8 reg1, u8 *b, u8 len) b 116 drivers/media/dvb-frontends/stv0299.c { .addr = state->config->demod_address, .flags = I2C_M_RD, .buf = b, .len = len } }; b 67 drivers/media/dvb-frontends/stv0910.c static inline u32 muldiv32(u32 a, u32 b, u32 c) b 71 drivers/media/dvb-frontends/stv0910.c tmp64 = (u64)a * (u64)b; b 40 drivers/media/dvb-frontends/stv6110.c static s32 abssub(s32 a, s32 b) b 42 drivers/media/dvb-frontends/stv6110.c if (a > b) b 43 drivers/media/dvb-frontends/stv6110.c return a - b; b 45 drivers/media/dvb-frontends/stv6110.c return b - a; b 46 drivers/media/dvb-frontends/stv6110x_priv.h #define MAKEWORD16(a, b) (((a) << 8) | (b)) b 294 drivers/media/dvb-frontends/stv6111.c static inline u32 muldiv32(u32 a, u32 b, u32 c) b 298 drivers/media/dvb-frontends/stv6111.c tmp64 = (u64)a * (u64)b; b 394 drivers/media/dvb-frontends/tda10023.c u8 a,b,c; b 396 drivers/media/dvb-frontends/tda10023.c b=tda10023_readreg(state, 0x15); b 400 drivers/media/dvb-frontends/tda10023.c *ber = a | (b<<8)| (c<<16); b 432 drivers/media/dvb-frontends/tda10023.c u8 a,b,c,d; b 434 drivers/media/dvb-frontends/tda10023.c b= tda10023_readreg (state, 0x75); b 437 drivers/media/dvb-frontends/tda10023.c *ucblocks = a | (b<<8)|(c<<16)|(d<<24); b 64 drivers/media/dvb-frontends/tda8083.c static int tda8083_readregs (struct tda8083_state* state, u8 reg1, u8 *b, u8 len) b 68 drivers/media/dvb-frontends/tda8083.c { .addr = state->config->demod_address, .flags = I2C_M_RD, .buf = b, .len = len } }; b 795 drivers/media/i2c/ad9389b.c u8 b[128]; b 796 drivers/media/i2c/ad9389b.c u8 *bp = b; b 805 drivers/media/i2c/ad9389b.c v4l2_dbg(lvl, debug, sd, "%s\n", b); b 127 drivers/media/i2c/adv748x/adv748x-hdmi.c int a, b; b 130 drivers/media/i2c/adv748x/adv748x-hdmi.c b = hdmi_read(state, ADV748X_HDMI_TMDS_2); b 131 drivers/media/i2c/adv748x/adv748x-hdmi.c if (a < 0 || b < 0) b 139 drivers/media/i2c/adv748x/adv748x-hdmi.c return ((a << 1) | (b >> 7)) * 1000000 + (b & 0x7f) * 1000000 / 128; b 1460 drivers/media/i2c/adv7511-v4l2.c u8 b[128]; b 1461 drivers/media/i2c/adv7511-v4l2.c u8 *bp = b; b 1469 drivers/media/i2c/adv7511-v4l2.c v4l2_dbg(lvl, debug, sd, "%s\n", b); b 1507 drivers/media/i2c/adv7604.c int a, b; b 1510 drivers/media/i2c/adv7604.c b = hdmi_read(sd, 0x3b); b 1511 drivers/media/i2c/adv7604.c if (a < 0 || b < 0) b 1513 drivers/media/i2c/adv7604.c freq = a * 1000000 + ((b & 0x30) >> 4) * 250000; b 1527 drivers/media/i2c/adv7604.c int a, b; b 1530 drivers/media/i2c/adv7604.c b = hdmi_read(sd, 0x52); b 1531 drivers/media/i2c/adv7604.c if (a < 0 || b < 0) b 1533 drivers/media/i2c/adv7604.c return ((a << 1) | (b >> 7)) * 1000000 + (b & 0x7f) * 1000000 / 128; b 1860 drivers/media/i2c/adv7604.c #define _SEL(a,b,c,d,e,f) { \ b 1861 drivers/media/i2c/adv7604.c ADV76XX_OP_CH_SEL_##a, ADV76XX_OP_CH_SEL_##b, ADV76XX_OP_CH_SEL_##c, \ b 2042 drivers/media/i2c/adv7842.c #define _SEL(a, b, c, d, e, f) { \ b 2043 drivers/media/i2c/adv7842.c ADV7842_OP_CH_SEL_##a, ADV7842_OP_CH_SEL_##b, ADV7842_OP_CH_SEL_##c, \ b 482 drivers/media/i2c/et8ek8/et8ek8_driver.c static int et8ek8_reglist_cmp(const void *a, const void *b) b 485 drivers/media/i2c/et8ek8/et8ek8_driver.c **list2 = (const struct et8ek8_reglist **)b; b 162 drivers/media/i2c/ir-kbd-i2c.c unsigned char b; b 165 drivers/media/i2c/ir-kbd-i2c.c rc = i2c_master_recv(ir->c, &b, 1); b 174 drivers/media/i2c/ir-kbd-i2c.c *scancode = b; b 211 drivers/media/i2c/ir-kbd-i2c.c unsigned char b; b 214 drivers/media/i2c/ir-kbd-i2c.c rc = i2c_master_recv(ir->c, &b, 1); b 226 drivers/media/i2c/ir-kbd-i2c.c dev_dbg(&ir->rc->dev, "key %02x\n", b); b 228 drivers/media/i2c/ir-kbd-i2c.c if (b == 0xff) b 231 drivers/media/i2c/ir-kbd-i2c.c if (b == 0xfe) b 236 drivers/media/i2c/ir-kbd-i2c.c *scancode = b; b 470 drivers/media/i2c/ir-kbd-i2c.c static int cmp_no_trail(u8 *a, u8 *b, unsigned int count) b 473 drivers/media/i2c/ir-kbd-i2c.c if (*a++ != *b++) b 477 drivers/media/i2c/ir-kbd-i2c.c return (*a & 0xf0) - (*b & 0xf0); b 60 drivers/media/i2c/mt9t112.c #define mt9t112_reg_write(ret, client, a, b) \ b 61 drivers/media/i2c/mt9t112.c ECHECKER(ret, __mt9t112_reg_write(client, a, b)) b 62 drivers/media/i2c/mt9t112.c #define mt9t112_mcu_write(ret, client, a, b) \ b 63 drivers/media/i2c/mt9t112.c ECHECKER(ret, __mt9t112_mcu_write(client, a, b)) b 65 drivers/media/i2c/mt9t112.c #define mt9t112_reg_mask_set(ret, client, a, b, c) \ b 66 drivers/media/i2c/mt9t112.c ECHECKER(ret, __mt9t112_reg_mask_set(client, a, b, c)) b 67 drivers/media/i2c/mt9t112.c #define mt9t112_mcu_mask_set(ret, client, a, b, c) \ b 68 drivers/media/i2c/mt9t112.c ECHECKER(ret, __mt9t112_mcu_mask_set(client, a, b, c)) b 272 drivers/media/i2c/mt9t112.c #define CLOCK_INFO(a, b) b 274 drivers/media/i2c/mt9t112.c #define CLOCK_INFO(a, b) mt9t112_clock_info(a, b) b 1946 drivers/media/i2c/smiapp/smiapp-core.c u32 min, max, a, b, max_m; b 1961 drivers/media/i2c/smiapp/smiapp-core.c b = crops[SMIAPP_PAD_SINK]->height b 1969 drivers/media/i2c/smiapp/smiapp-core.c b = clamp(b, sensor->limits[SMIAPP_LIMIT_SCALER_M_MIN], b 1974 drivers/media/i2c/smiapp/smiapp-core.c dev_dbg(&client->dev, "scaling: a %d b %d max_m %d\n", a, b, max_m); b 1976 drivers/media/i2c/smiapp/smiapp-core.c min = min(max_m, min(a, b)); b 1977 drivers/media/i2c/smiapp/smiapp-core.c max = min(max_m, max(a, b)); b 4903 drivers/media/pci/bt8xx/bttv-cards.c unsigned char b; b 4904 drivers/media/pci/bt8xx/bttv-cards.c pci_read_config_byte(dev, 0x53, &b); b 4907 drivers/media/pci/bt8xx/bttv-cards.c b); b 1899 drivers/media/pci/bt8xx/bttv-driver.c const struct v4l2_rect * b, b 1915 drivers/media/pci/bt8xx/bttv-driver.c max_left = b->left + b->width - width; b 1929 drivers/media/pci/bt8xx/bttv-driver.c max_top = b->top + b->height - c->rect.height; b 1955 drivers/media/pci/bt8xx/bttv-driver.c const struct v4l2_rect *b; b 1969 drivers/media/pci/bt8xx/bttv-driver.c b = &bttv_tvnorms[btv->tvnorm].cropcap.bounds; b 1984 drivers/media/pci/bt8xx/bttv-driver.c max_width = min_t(unsigned int, b->width, MAX_HACTIVE); b 1985 drivers/media/pci/bt8xx/bttv-driver.c max_height = b->height; b 1990 drivers/media/pci/bt8xx/bttv-driver.c if (btv->vbi_end > b->top) { b 1991 drivers/media/pci/bt8xx/bttv-driver.c max_height -= btv->vbi_end - b->top; b 2024 drivers/media/pci/bt8xx/bttv-driver.c bttv_crop_adjust(c, b, *width, *height, field); b 2650 drivers/media/pci/bt8xx/bttv-driver.c struct v4l2_buffer *b) b 2653 drivers/media/pci/bt8xx/bttv-driver.c return videobuf_querybuf(bttv_queue(fh), b); b 2656 drivers/media/pci/bt8xx/bttv-driver.c static int bttv_qbuf(struct file *file, void *priv, struct v4l2_buffer *b) b 2665 drivers/media/pci/bt8xx/bttv-driver.c return videobuf_qbuf(bttv_queue(fh), b); b 2668 drivers/media/pci/bt8xx/bttv-driver.c static int bttv_dqbuf(struct file *file, void *priv, struct v4l2_buffer *b) b 2671 drivers/media/pci/bt8xx/bttv-driver.c return videobuf_dqbuf(bttv_queue(fh), b, b 2791 drivers/media/pci/bt8xx/bttv-driver.c const struct v4l2_rect *b; b 2815 drivers/media/pci/bt8xx/bttv-driver.c b = &bttv_tvnorms[btv->tvnorm].cropcap.bounds; b 2817 drivers/media/pci/bt8xx/bttv-driver.c b_left = b->left; b 2818 drivers/media/pci/bt8xx/bttv-driver.c b_right = b_left + b->width; b 2819 drivers/media/pci/bt8xx/bttv-driver.c b_bottom = b->top + b->height; b 2821 drivers/media/pci/bt8xx/bttv-driver.c b_top = max(b->top, btv->vbi_end); b 327 drivers/media/pci/bt8xx/bttv-input.c unsigned char b; b 330 drivers/media/pci/bt8xx/bttv-input.c rc = i2c_master_recv(ir->c, &b, 1); b 339 drivers/media/pci/bt8xx/bttv-input.c if (b==0xaa) b 341 drivers/media/pci/bt8xx/bttv-input.c dprintk("key %02x\n", b); b 356 drivers/media/pci/bt8xx/bttv-input.c *scancode = b; b 269 drivers/media/pci/bt8xx/bttv.h #define _MUXSELe(a, b...) 0##a << 28 | _MUXSELf(b) b 270 drivers/media/pci/bt8xx/bttv.h #define _MUXSELd(a, b...) 0##a << 26 | _MUXSELe(b) b 271 drivers/media/pci/bt8xx/bttv.h #define _MUXSELc(a, b...) 0##a << 24 | _MUXSELd(b) b 272 drivers/media/pci/bt8xx/bttv.h #define _MUXSELb(a, b...) 0##a << 22 | _MUXSELc(b) b 273 drivers/media/pci/bt8xx/bttv.h #define _MUXSELa(a, b...) 0##a << 20 | _MUXSELb(b) b 274 drivers/media/pci/bt8xx/bttv.h #define _MUXSEL9(a, b...) 0##a << 18 | _MUXSELa(b) b 275 drivers/media/pci/bt8xx/bttv.h #define _MUXSEL8(a, b...) 0##a << 16 | _MUXSEL9(b) b 276 drivers/media/pci/bt8xx/bttv.h #define _MUXSEL7(a, b...) 0##a << 14 | _MUXSEL8(b) b 277 drivers/media/pci/bt8xx/bttv.h #define _MUXSEL6(a, b...) 0##a << 12 | _MUXSEL7(b) b 278 drivers/media/pci/bt8xx/bttv.h #define _MUXSEL5(a, b...) 0##a << 10 | _MUXSEL6(b) b 279 drivers/media/pci/bt8xx/bttv.h #define _MUXSEL4(a, b...) 0##a << 8 | _MUXSEL5(b) b 280 drivers/media/pci/bt8xx/bttv.h #define _MUXSEL3(a, b...) 0##a << 6 | _MUXSEL4(b) b 281 drivers/media/pci/bt8xx/bttv.h #define _MUXSEL2(a, b...) 0##a << 4 | _MUXSEL3(b) b 282 drivers/media/pci/bt8xx/bttv.h #define _MUXSEL1(a, b...) 0##a << 2 | _MUXSEL2(b) b 283 drivers/media/pci/bt8xx/bttv.h #define MUXSEL(a, b...) (a | _MUXSEL1(b)) b 870 drivers/media/pci/cx18/cx18-ioctl.c struct v4l2_buffer *b) b 880 drivers/media/pci/cx18/cx18-ioctl.c return videobuf_querybuf(cx18_vb_queue(id), b); b 883 drivers/media/pci/cx18/cx18-ioctl.c static int cx18_qbuf(struct file *file, void *priv, struct v4l2_buffer *b) b 893 drivers/media/pci/cx18/cx18-ioctl.c return videobuf_qbuf(cx18_vb_queue(id), b); b 896 drivers/media/pci/cx18/cx18-ioctl.c static int cx18_dqbuf(struct file *file, void *priv, struct v4l2_buffer *b) b 906 drivers/media/pci/cx18/cx18-ioctl.c return videobuf_dqbuf(cx18_vb_queue(id), b, file->f_flags & O_NONBLOCK); b 652 drivers/media/pci/cx88/cx88-alsa.c int v, b; b 657 drivers/media/pci/cx88/cx88-alsa.c b = left ? (0x8000 * right) / left : 0x8000; b 660 drivers/media/pci/cx88/cx88-alsa.c b = right ? 0xffff - (0x8000 * left) / right : 0x8000; b 663 drivers/media/pci/cx88/cx88-alsa.c wm8775_s_ctrl(core, V4L2_CID_AUDIO_BALANCE, b); b 672 drivers/media/pci/cx88/cx88-alsa.c int left, right, v, b; b 681 drivers/media/pci/cx88/cx88-alsa.c b = right - left; b 682 drivers/media/pci/cx88/cx88-alsa.c if (b < 0) { b 684 drivers/media/pci/cx88/cx88-alsa.c b = (-b) | 0x40; b 695 drivers/media/pci/cx88/cx88-alsa.c if ((cx_read(AUD_BAL_CTL) & 0x7f) != b) { b 696 drivers/media/pci/cx88/cx88-alsa.c cx_write(AUD_BAL_CTL, b); b 19 drivers/media/pci/cx88/cx88-dsp.c #define compat_remainder(a, b) \ b 20 drivers/media/pci/cx88/cx88-dsp.c ((float)(((s32)((a) * 100)) % ((s32)((b) * 100))) / 100.0) b 175 drivers/media/pci/intel/ipu3/ipu3-cio2.c struct cio2_buffer *b, b 179 drivers/media/pci/intel/ipu3/ipu3-cio2.c struct vb2_buffer *vb = &b->vbb.vb2_buf; b 183 drivers/media/pci/intel/ipu3/ipu3-cio2.c entry[0].first_entry.first_page_offset = b->offset; b 203 drivers/media/pci/intel/ipu3/ipu3-cio2.c entry->lop_page_addr = b->lop_bus_addr[i] >> PAGE_SHIFT; b 271 drivers/media/pci/intel/ipu3/ipu3-cio2.c static s32 cio2_rx_timing(s32 a, s32 b, s64 freq, int def) b 285 drivers/media/pci/intel/ipu3/ipu3-cio2.c r = accinv * b * (uiinv >> LIMIT_SHIFT); b 560 drivers/media/pci/intel/ipu3/ipu3-cio2.c struct cio2_buffer *b; b 565 drivers/media/pci/intel/ipu3/ipu3-cio2.c b = q->bufs[q->bufs_first]; b 566 drivers/media/pci/intel/ipu3/ipu3-cio2.c if (b) { b 572 drivers/media/pci/intel/ipu3/ipu3-cio2.c "buffer %i done\n", b->vbb.vb2_buf.index); b 574 drivers/media/pci/intel/ipu3/ipu3-cio2.c b->vbb.vb2_buf.timestamp = ns; b 575 drivers/media/pci/intel/ipu3/ipu3-cio2.c b->vbb.field = V4L2_FIELD_NONE; b 576 drivers/media/pci/intel/ipu3/ipu3-cio2.c b->vbb.sequence = atomic_read(&q->frame_sequence); b 577 drivers/media/pci/intel/ipu3/ipu3-cio2.c if (b->vbb.vb2_buf.planes[0].length != bytes) b 579 drivers/media/pci/intel/ipu3/ipu3-cio2.c b->vbb.vb2_buf.planes[0].length, b 581 drivers/media/pci/intel/ipu3/ipu3-cio2.c vb2_buffer_done(&b->vbb.vb2_buf, VB2_BUF_STATE_DONE); b 842 drivers/media/pci/intel/ipu3/ipu3-cio2.c struct cio2_buffer *b = b 858 drivers/media/pci/intel/ipu3/ipu3-cio2.c memset(b->lop, 0, sizeof(b->lop)); b 861 drivers/media/pci/intel/ipu3/ipu3-cio2.c b->lop[i] = dma_alloc_coherent(dev, CIO2_PAGE_SIZE, b 862 drivers/media/pci/intel/ipu3/ipu3-cio2.c &b->lop_bus_addr[i], GFP_KERNEL); b 863 drivers/media/pci/intel/ipu3/ipu3-cio2.c if (!b->lop[i]) b 873 drivers/media/pci/intel/ipu3/ipu3-cio2.c b->offset = sg->sgl->offset; b 879 drivers/media/pci/intel/ipu3/ipu3-cio2.c b->lop[i][j] = sg_page_iter_dma_address(&sg_iter) >> PAGE_SHIFT; b 887 drivers/media/pci/intel/ipu3/ipu3-cio2.c b->lop[i][j] = cio2->dummy_page_bus_addr >> PAGE_SHIFT; b 892 drivers/media/pci/intel/ipu3/ipu3-cio2.c b->lop[i], b->lop_bus_addr[i]); b 902 drivers/media/pci/intel/ipu3/ipu3-cio2.c struct cio2_buffer *b = b 950 drivers/media/pci/intel/ipu3/ipu3-cio2.c q->bufs[next] = b; b 952 drivers/media/pci/intel/ipu3/ipu3-cio2.c cio2_fbpt_entry_init_buf(cio2, b, entry); b 975 drivers/media/pci/intel/ipu3/ipu3-cio2.c struct cio2_buffer *b = b 981 drivers/media/pci/intel/ipu3/ipu3-cio2.c if (b->lop[i]) b 983 drivers/media/pci/intel/ipu3/ipu3-cio2.c b->lop[i], b->lop_bus_addr[i]); b 42 drivers/media/pci/mantis/mantis_vp2033.c u8 b = 0xff; b 45 drivers/media/pci/mantis/mantis_vp2033.c {.addr = 0x50, .flags = 0, .buf = &b, .len = 1}, b 98 drivers/media/pci/mantis/mantis_vp2040.c u8 b = 0xff; b 101 drivers/media/pci/mantis/mantis_vp2040.c {.addr = 0x50, .flags = 0, .buf = &b, .len = 1}, b 239 drivers/media/pci/ngene/ngene-core.c u8 buf[8], *b; b 247 drivers/media/pci/ngene/ngene-core.c b = dev->hosttongene; b 248 drivers/media/pci/ngene/ngene-core.c dev_err(pdev, "dev->hosttongene (%p): %*ph\n", b, 8, b); b 250 drivers/media/pci/ngene/ngene-core.c b = dev->ngenetohost; b 251 drivers/media/pci/ngene/ngene-core.c dev_err(pdev, "dev->ngenetohost (%p): %*ph\n", b, 8, b); b 110 drivers/media/pci/saa7134/saa7134-input.c unsigned char b; b 131 drivers/media/pci/saa7134/saa7134-input.c b = 0; b 133 drivers/media/pci/saa7134/saa7134-input.c while (1 != i2c_master_send(ir->c, &b, 1)) { b 146 drivers/media/pci/saa7134/saa7134-input.c rc = i2c_master_recv(ir->c, &b, 1); b 155 drivers/media/pci/saa7134/saa7134-input.c *scancode = b; b 164 drivers/media/pci/saa7134/saa7134-input.c unsigned char b; b 189 drivers/media/pci/saa7134/saa7134-input.c rc = i2c_master_recv(ir->c, &b, 1); b 199 drivers/media/pci/saa7134/saa7134-input.c if (b == 0xff) b 204 drivers/media/pci/saa7134/saa7134-input.c input_dbg("get_key_msi_tvanywhere_plus: Key = 0x%02X\n", b); b 206 drivers/media/pci/saa7134/saa7134-input.c *scancode = b; b 215 drivers/media/pci/saa7134/saa7134-input.c unsigned char b; b 241 drivers/media/pci/saa7134/saa7134-input.c rc = i2c_master_recv(ir->c, &b, 1); b 251 drivers/media/pci/saa7134/saa7134-input.c if (b == 0xff) b 256 drivers/media/pci/saa7134/saa7134-input.c input_dbg("get_key_kworld_pc150u: Key = 0x%02X\n", b); b 258 drivers/media/pci/saa7134/saa7134-input.c *scancode = b; b 267 drivers/media/pci/saa7134/saa7134-input.c unsigned char b; b 270 drivers/media/pci/saa7134/saa7134-input.c rc = i2c_master_recv(ir->c, &b, 1); b 279 drivers/media/pci/saa7134/saa7134-input.c if (b==0) b 283 drivers/media/pci/saa7134/saa7134-input.c if (b & 0x80) b 287 drivers/media/pci/saa7134/saa7134-input.c *scancode = b; b 337 drivers/media/pci/saa7134/saa7134-input.c unsigned char b[4]; b 341 drivers/media/pci/saa7134/saa7134-input.c rc = i2c_master_recv(ir->c, b, 4); b 349 drivers/media/pci/saa7134/saa7134-input.c for (start = 0; start < ARRAY_SIZE(b); start++) { b 350 drivers/media/pci/saa7134/saa7134-input.c if (b[start] == marker) { b 351 drivers/media/pci/saa7134/saa7134-input.c code=b[(start+parity_offset + 1) % 4]; b 352 drivers/media/pci/saa7134/saa7134-input.c parity=b[(start+parity_offset) % 4]; b 614 drivers/media/pci/saa7134/saa7134-video.c static int cliplist_cmp(const void *a, const void *b) b 617 drivers/media/pci/saa7134/saa7134-video.c const struct cliplist *clb = b; b 1647 drivers/media/pci/saa7134/saa7134-video.c struct v4l2_rect *b = &dev->crop_bounds; b 1663 drivers/media/pci/saa7134/saa7134-video.c if (c->top < b->top) b 1664 drivers/media/pci/saa7134/saa7134-video.c c->top = b->top; b 1665 drivers/media/pci/saa7134/saa7134-video.c if (c->top > b->top + b->height) b 1666 drivers/media/pci/saa7134/saa7134-video.c c->top = b->top + b->height; b 1667 drivers/media/pci/saa7134/saa7134-video.c if (c->height > b->top - c->top + b->height) b 1668 drivers/media/pci/saa7134/saa7134-video.c c->height = b->top - c->top + b->height; b 1670 drivers/media/pci/saa7134/saa7134-video.c if (c->left < b->left) b 1671 drivers/media/pci/saa7134/saa7134-video.c c->left = b->left; b 1672 drivers/media/pci/saa7134/saa7134-video.c if (c->left > b->left + b->width) b 1673 drivers/media/pci/saa7134/saa7134-video.c c->left = b->left + b->width; b 1674 drivers/media/pci/saa7134/saa7134-video.c if (c->width > b->left - c->left + b->width) b 1675 drivers/media/pci/saa7134/saa7134-video.c c->width = b->left - c->left + b->width; b 15 drivers/media/pci/saa7164/saa7164-bus.c struct tmComResBusInfo *b = &dev->bus; b 17 drivers/media/pci/saa7164/saa7164-bus.c mutex_init(&b->lock); b 19 drivers/media/pci/saa7164/saa7164-bus.c b->Type = TYPE_BUS_PCIe; b 20 drivers/media/pci/saa7164/saa7164-bus.c b->m_wMaxReqSize = SAA_DEVICE_MAXREQUESTSIZE; b 22 drivers/media/pci/saa7164/saa7164-bus.c b->m_pdwSetRing = (u8 __iomem *)(dev->bmmio + b 25 drivers/media/pci/saa7164/saa7164-bus.c b->m_dwSizeSetRing = SAA_DEVICE_BUFFERBLOCKSIZE; b 27 drivers/media/pci/saa7164/saa7164-bus.c b->m_pdwGetRing = (u8 __iomem *)(dev->bmmio + b 30 drivers/media/pci/saa7164/saa7164-bus.c b->m_dwSizeGetRing = SAA_DEVICE_BUFFERBLOCKSIZE; b 32 drivers/media/pci/saa7164/saa7164-bus.c b->m_dwSetWritePos = ((u32)dev->intfdesc.BARLocation) + b 34 drivers/media/pci/saa7164/saa7164-bus.c b->m_dwSetReadPos = b->m_dwSetWritePos + (1 * sizeof(u32)); b 36 drivers/media/pci/saa7164/saa7164-bus.c b->m_dwGetWritePos = b->m_dwSetWritePos + (2 * sizeof(u32)); b 37 drivers/media/pci/saa7164/saa7164-bus.c b->m_dwGetReadPos = b->m_dwSetWritePos + (3 * sizeof(u32)); b 44 drivers/media/pci/saa7164/saa7164-bus.c struct tmComResBusInfo *b = &dev->bus; b 47 drivers/media/pci/saa7164/saa7164-bus.c dprintk(DBGLVL_BUS, " .type = %d\n", b->Type); b 49 drivers/media/pci/saa7164/saa7164-bus.c dprintk(DBGLVL_BUS, " .m_wMaxReqSize = 0x%x\n", b->m_wMaxReqSize); b 50 drivers/media/pci/saa7164/saa7164-bus.c dprintk(DBGLVL_BUS, " .m_pdwSetRing = 0x%p\n", b->m_pdwSetRing); b 51 drivers/media/pci/saa7164/saa7164-bus.c dprintk(DBGLVL_BUS, " .m_dwSizeSetRing = 0x%x\n", b->m_dwSizeSetRing); b 52 drivers/media/pci/saa7164/saa7164-bus.c dprintk(DBGLVL_BUS, " .m_pdwGetRing = 0x%p\n", b->m_pdwGetRing); b 53 drivers/media/pci/saa7164/saa7164-bus.c dprintk(DBGLVL_BUS, " .m_dwSizeGetRing = 0x%x\n", b->m_dwSizeGetRing); b 56 drivers/media/pci/saa7164/saa7164-bus.c b->m_dwSetReadPos, saa7164_readl(b->m_dwSetReadPos)); b 59 drivers/media/pci/saa7164/saa7164-bus.c b->m_dwSetWritePos, saa7164_readl(b->m_dwSetWritePos)); b 62 drivers/media/pci/saa7164/saa7164-bus.c b->m_dwGetReadPos, saa7164_readl(b->m_dwGetReadPos)); b 65 drivers/media/pci/saa7164/saa7164-bus.c b->m_dwGetWritePos, saa7164_readl(b->m_dwGetWritePos)); b 72 drivers/media/pci/saa7164/saa7164-bus.c struct tmComResBusInfo *b = &dev->bus; b 75 drivers/media/pci/saa7164/saa7164-bus.c if (saa7164_readl(b->m_dwSetReadPos) > b->m_dwSizeSetRing) b 78 drivers/media/pci/saa7164/saa7164-bus.c if (saa7164_readl(b->m_dwSetWritePos) > b->m_dwSizeSetRing) b 81 drivers/media/pci/saa7164/saa7164-bus.c if (saa7164_readl(b->m_dwGetReadPos) > b->m_dwSizeGetRing) b 84 drivers/media/pci/saa7164/saa7164-bus.c if (saa7164_readl(b->m_dwGetWritePos) > b->m_dwSizeGetRing) b 1052 drivers/media/pci/saa7164/saa7164-core.c struct tmComResBusInfo *b; b 1064 drivers/media/pci/saa7164/saa7164-core.c b = &dev->bus; b 1065 drivers/media/pci/saa7164/saa7164-core.c mutex_lock(&b->lock); b 1068 drivers/media/pci/saa7164/saa7164-core.c b->m_dwSetReadPos, saa7164_readl(b->m_dwSetReadPos)); b 1071 drivers/media/pci/saa7164/saa7164-core.c b->m_dwSetWritePos, saa7164_readl(b->m_dwSetWritePos)); b 1074 drivers/media/pci/saa7164/saa7164-core.c b->m_dwGetReadPos, saa7164_readl(b->m_dwGetReadPos)); b 1077 drivers/media/pci/saa7164/saa7164-core.c b->m_dwGetWritePos, saa7164_readl(b->m_dwGetWritePos)); b 1081 drivers/media/pci/saa7164/saa7164-core.c for (i = 0; i < b->m_dwSizeSetRing; i++) { b 1085 drivers/media/pci/saa7164/saa7164-core.c seq_printf(m, " %02x", readb(b->m_pdwSetRing + i)); b 1096 drivers/media/pci/saa7164/saa7164-core.c for (i = 0; i < b->m_dwSizeGetRing; i++) { b 1100 drivers/media/pci/saa7164/saa7164-core.c seq_printf(m, " %02x", readb(b->m_pdwGetRing + i)); b 1108 drivers/media/pci/saa7164/saa7164-core.c mutex_unlock(&b->lock); b 476 drivers/media/pci/saa7164/saa7164-dvb.c struct saa7164_buffer *b; b 488 drivers/media/pci/saa7164/saa7164-dvb.c b = list_entry(c, struct saa7164_buffer, list); b 490 drivers/media/pci/saa7164/saa7164-dvb.c saa7164_buffer_dealloc(b); b 2078 drivers/media/pci/ttpci/av7110.c u8 b = 0xff; b 2080 drivers/media/pci/ttpci/av7110.c struct i2c_msg msg[] = { { .addr = 0x50,.flags = 0,.buf = &b,.len = 1 }, b 945 drivers/media/pci/ttpci/av7110_hw.c static int OSDSetColor(struct av7110 *av7110, u8 color, u8 r, u8 g, u8 b, u8 blend) b 952 drivers/media/pci/ttpci/av7110_hw.c yuv = blend ? RGB2YUV(r,g,b) : 0; b 1088 drivers/media/pci/ttpci/av7110_hw.c u8 r, g = 0, b = 0, blend = 0; b 1093 drivers/media/pci/ttpci/av7110_hw.c get_user(b, colors + i * 4 + 2) || b 1098 drivers/media/pci/ttpci/av7110_hw.c ret = OSDSetColor(av7110, dc->color + i, r, g, b, blend); b 1153 drivers/media/pci/ttpci/budget-av.c u8 b = 0xff; b 1155 drivers/media/pci/ttpci/budget-av.c struct i2c_msg msg[] = { {.addr = 0x50,.flags = 0,.buf = &b,.len = 1}, b 427 drivers/media/pci/ttpci/budget.c u8 b = 0xff; b 429 drivers/media/pci/ttpci/budget.c struct i2c_msg msg[] = { { .addr = 0x50,.flags = 0,.buf = &b,.len = 1 }, b 30 drivers/media/pci/ttpci/dvb_filter.c u8 *b = mbuf+c; b 32 drivers/media/pci/ttpci/dvb_filter.c if ( b[0] == 0x0b && b[1] == 0x77 ) b 629 drivers/media/platform/am437x/am437x-vpfe.c val = ((bcomp->b & VPFE_BLK_COMP_MASK) | b 2115 drivers/media/platform/am437x/am437x-vpfe.c static int enclosed_rectangle(struct v4l2_rect *a, struct v4l2_rect *b) b 2117 drivers/media/platform/am437x/am437x-vpfe.c if (a->left < b->left || a->top < b->top) b 2120 drivers/media/platform/am437x/am437x-vpfe.c if (a->left + a->width > b->left + b->width) b 2123 drivers/media/platform/am437x/am437x-vpfe.c if (a->top + a->height > b->top + b->height) b 345 drivers/media/platform/davinci/dm355_ccdc.c val = (bcomp->b & CCDC_BLK_COMP_MASK) | b 358 drivers/media/platform/davinci/dm644x_ccdc.c val = ((bcomp->b & CCDC_BLK_COMP_MASK) | b 369 drivers/media/platform/davinci/vpif.h unsigned long btm_vbi, unsigned long a, unsigned long b) b 376 drivers/media/platform/davinci/vpif.h unsigned long btm_vbi, unsigned long a, unsigned long b) b 383 drivers/media/platform/davinci/vpif.h unsigned long btm_vbi, unsigned long a, unsigned long b) b 390 drivers/media/platform/davinci/vpif.h unsigned long btm_vbi, unsigned long a, unsigned long b) b 435 drivers/media/platform/exynos-gsc/gsc-m2m.c static int is_rectangle_enclosed(struct v4l2_rect *a, struct v4l2_rect *b) b 437 drivers/media/platform/exynos-gsc/gsc-m2m.c if (a->left < b->left || a->top < b->top) b 440 drivers/media/platform/exynos-gsc/gsc-m2m.c if (a->left + a->width > b->left + b->width) b 443 drivers/media/platform/exynos-gsc/gsc-m2m.c if (a->top + a->height > b->top + b->height) b 1303 drivers/media/platform/exynos4-is/fimc-capture.c static int enclosed_rectangle(struct v4l2_rect *a, struct v4l2_rect *b) b 1305 drivers/media/platform/exynos4-is/fimc-capture.c if (a->left < b->left || a->top < b->top) b 1307 drivers/media/platform/exynos4-is/fimc-capture.c if (a->left + a->width > b->left + b->width) b 1309 drivers/media/platform/exynos4-is/fimc-capture.c if (a->top + a->height > b->top + b->height) b 872 drivers/media/platform/exynos4-is/fimc-lite.c static int enclosed_rectangle(struct v4l2_rect *a, struct v4l2_rect *b) b 874 drivers/media/platform/exynos4-is/fimc-lite.c if (a->left < b->left || a->top < b->top) b 876 drivers/media/platform/exynos4-is/fimc-lite.c if (a->left + a->width > b->left + b->width) b 878 drivers/media/platform/exynos4-is/fimc-lite.c if (a->top + a->height > b->top + b->height) b 64 drivers/media/platform/mtk-jpeg/mtk_jpeg_core.c struct vb2_v4l2_buffer b; b 81 drivers/media/platform/mtk-jpeg/mtk_jpeg_core.c return container_of(to_vb2_v4l2_buffer(vb), struct mtk_jpeg_src_buf, b); b 915 drivers/media/platform/omap3isp/ispvideo.c isp_video_querybuf(struct file *file, void *fh, struct v4l2_buffer *b) b 922 drivers/media/platform/omap3isp/ispvideo.c ret = vb2_querybuf(&vfh->queue, b); b 929 drivers/media/platform/omap3isp/ispvideo.c isp_video_qbuf(struct file *file, void *fh, struct v4l2_buffer *b) b 936 drivers/media/platform/omap3isp/ispvideo.c ret = vb2_qbuf(&vfh->queue, video->video.v4l2_dev->mdev, b); b 943 drivers/media/platform/omap3isp/ispvideo.c isp_video_dqbuf(struct file *file, void *fh, struct v4l2_buffer *b) b 950 drivers/media/platform/omap3isp/ispvideo.c ret = vb2_dqbuf(&vfh->queue, b, file->f_flags & O_NONBLOCK); b 27 drivers/media/platform/qcom/venus/venc_ctrls.c u32 b, p, ratio; b 40 drivers/media/platform/qcom/venus/venc_ctrls.c b = p = half; b 42 drivers/media/platform/qcom/venus/venc_ctrls.c for (; b <= gop_size - 1; b++, p--) { b 43 drivers/media/platform/qcom/venus/venc_ctrls.c if (b % p) b 46 drivers/media/platform/qcom/venus/venc_ctrls.c ratio = b / p; b 60 drivers/media/platform/qcom/venus/venc_ctrls.c if (b + p + 1 != gop_size) b 63 drivers/media/platform/qcom/venus/venc_ctrls.c *bf = b; b 469 drivers/media/platform/rcar_jpu.c struct v4l2_m2m_buffer *b = b 472 drivers/media/platform/rcar_jpu.c return container_of(b, struct jpu_buffer, buf); b 971 drivers/media/platform/s3c-camif/camif-capture.c struct v4l2_buffer *b) b 974 drivers/media/platform/s3c-camif/camif-capture.c return vb2_prepare_buf(&vp->vb_queue, vp->vdev.v4l2_dev->mdev, b); b 1740 drivers/media/platform/s5p-jpeg/jpeg-core.c static int enclosed_rectangle(struct v4l2_rect *a, struct v4l2_rect *b) b 1742 drivers/media/platform/s5p-jpeg/jpeg-core.c if (a->left < b->left || a->top < b->top) b 1744 drivers/media/platform/s5p-jpeg/jpeg-core.c if (a->left + a->width > b->left + b->width) b 1746 drivers/media/platform/s5p-jpeg/jpeg-core.c if (a->top + a->height > b->top + b->height) b 132 drivers/media/platform/s5p-mfc/s5p_mfc.c struct s5p_mfc_buf *b; b 136 drivers/media/platform/s5p-mfc/s5p_mfc.c b = list_entry(lh->next, struct s5p_mfc_buf, list); b 137 drivers/media/platform/s5p-mfc/s5p_mfc.c for (i = 0; i < b->b->vb2_buf.num_planes; i++) b 138 drivers/media/platform/s5p-mfc/s5p_mfc.c vb2_set_plane_payload(&b->b->vb2_buf, i, 0); b 139 drivers/media/platform/s5p-mfc/s5p_mfc.c vb2_buffer_done(&b->b->vb2_buf, VB2_BUF_STATE_ERROR); b 140 drivers/media/platform/s5p-mfc/s5p_mfc.c list_del(&b->list); b 230 drivers/media/platform/s5p-mfc/s5p_mfc.c dst_buf->b->vb2_buf.index); b 231 drivers/media/platform/s5p-mfc/s5p_mfc.c vb2_set_plane_payload(&dst_buf->b->vb2_buf, 0, 0); b 232 drivers/media/platform/s5p-mfc/s5p_mfc.c vb2_set_plane_payload(&dst_buf->b->vb2_buf, 1, 0); b 236 drivers/media/platform/s5p-mfc/s5p_mfc.c dst_buf->b->sequence = (ctx->sequence++); b 240 drivers/media/platform/s5p-mfc/s5p_mfc.c dst_buf->b->field = V4L2_FIELD_NONE; b 242 drivers/media/platform/s5p-mfc/s5p_mfc.c dst_buf->b->field = V4L2_FIELD_INTERLACED; b 243 drivers/media/platform/s5p-mfc/s5p_mfc.c dst_buf->b->flags |= V4L2_BUF_FLAG_LAST; b 245 drivers/media/platform/s5p-mfc/s5p_mfc.c ctx->dec_dst_flag &= ~(1 << dst_buf->b->vb2_buf.index); b 246 drivers/media/platform/s5p-mfc/s5p_mfc.c vb2_buffer_done(&dst_buf->b->vb2_buf, VB2_BUF_STATE_DONE); b 267 drivers/media/platform/s5p-mfc/s5p_mfc.c u32 addr = (u32)vb2_dma_contig_plane_dma_addr(&dst_buf->b->vb2_buf, 0); b 270 drivers/media/platform/s5p-mfc/s5p_mfc.c dst_buf->b->timecode = src_buf->b->timecode; b 271 drivers/media/platform/s5p-mfc/s5p_mfc.c dst_buf->b->vb2_buf.timestamp = b 272 drivers/media/platform/s5p-mfc/s5p_mfc.c src_buf->b->vb2_buf.timestamp; b 273 drivers/media/platform/s5p-mfc/s5p_mfc.c dst_buf->b->flags &= b 275 drivers/media/platform/s5p-mfc/s5p_mfc.c dst_buf->b->flags |= b 276 drivers/media/platform/s5p-mfc/s5p_mfc.c src_buf->b->flags b 280 drivers/media/platform/s5p-mfc/s5p_mfc.c dst_buf->b->flags |= b 284 drivers/media/platform/s5p-mfc/s5p_mfc.c dst_buf->b->flags |= b 288 drivers/media/platform/s5p-mfc/s5p_mfc.c dst_buf->b->flags |= b 328 drivers/media/platform/s5p-mfc/s5p_mfc.c u32 addr = (u32)vb2_dma_contig_plane_dma_addr(&dst_buf->b->vb2_buf, 0); b 334 drivers/media/platform/s5p-mfc/s5p_mfc.c dst_buf->b->sequence = ctx->sequence; b 339 drivers/media/platform/s5p-mfc/s5p_mfc.c dst_buf->b->field = V4L2_FIELD_NONE; b 341 drivers/media/platform/s5p-mfc/s5p_mfc.c dst_buf->b->field = b 343 drivers/media/platform/s5p-mfc/s5p_mfc.c vb2_set_plane_payload(&dst_buf->b->vb2_buf, 0, b 345 drivers/media/platform/s5p-mfc/s5p_mfc.c vb2_set_plane_payload(&dst_buf->b->vb2_buf, 1, b 347 drivers/media/platform/s5p-mfc/s5p_mfc.c clear_bit(dst_buf->b->vb2_buf.index, b 350 drivers/media/platform/s5p-mfc/s5p_mfc.c vb2_buffer_done(&dst_buf->b->vb2_buf, err ? b 430 drivers/media/platform/s5p-mfc/s5p_mfc.c src_buf->b->vb2_buf.planes[0].bytesused) { b 442 drivers/media/platform/s5p-mfc/s5p_mfc.c vb2_buffer_done(&src_buf->b->vb2_buf, b 445 drivers/media/platform/s5p-mfc/s5p_mfc.c vb2_buffer_done(&src_buf->b->vb2_buf, b 542 drivers/media/platform/s5p-mfc/s5p_mfc.c src_buf->b->vb2_buf.planes[0].bytesused) b 581 drivers/media/platform/s5p-mfc/s5p_mfc.c vb2_buffer_done(&src_buf->b->vb2_buf, b 616 drivers/media/platform/s5p-mfc/s5p_mfc.c vb2_set_plane_payload(&mb_entry->b->vb2_buf, 0, 0); b 617 drivers/media/platform/s5p-mfc/s5p_mfc.c vb2_buffer_done(&mb_entry->b->vb2_buf, VB2_BUF_STATE_DONE); b 178 drivers/media/platform/s5p-mfc/s5p_mfc_common.h struct vb2_v4l2_buffer *b; b 965 drivers/media/platform/s5p-mfc/s5p_mfc_dec.c ctx->dst_bufs[i].b = vbuf; b 983 drivers/media/platform/s5p-mfc/s5p_mfc_dec.c ctx->src_bufs[i].b = vbuf; b 1135 drivers/media/platform/s5p-mfc/s5p_mfc_enc.c dst_addr = vb2_dma_contig_plane_dma_addr(&dst_mb->b->vb2_buf, 0); b 1136 drivers/media/platform/s5p-mfc/s5p_mfc_enc.c dst_size = vb2_plane_size(&dst_mb->b->vb2_buf, 0); b 1155 drivers/media/platform/s5p-mfc/s5p_mfc_enc.c vb2_set_plane_payload(&dst_mb->b->vb2_buf, 0, b 1158 drivers/media/platform/s5p-mfc/s5p_mfc_enc.c vb2_buffer_done(&dst_mb->b->vb2_buf, b 1193 drivers/media/platform/s5p-mfc/s5p_mfc_enc.c src_y_addr = vb2_dma_contig_plane_dma_addr(&src_mb->b->vb2_buf, 0); b 1194 drivers/media/platform/s5p-mfc/s5p_mfc_enc.c src_c_addr = vb2_dma_contig_plane_dma_addr(&src_mb->b->vb2_buf, 1); b 1199 drivers/media/platform/s5p-mfc/s5p_mfc_enc.c dst_addr = vb2_dma_contig_plane_dma_addr(&dst_mb->b->vb2_buf, 0); b 1200 drivers/media/platform/s5p-mfc/s5p_mfc_enc.c dst_size = vb2_plane_size(&dst_mb->b->vb2_buf, 0); b 1227 drivers/media/platform/s5p-mfc/s5p_mfc_enc.c &mb_entry->b->vb2_buf, 0); b 1229 drivers/media/platform/s5p-mfc/s5p_mfc_enc.c &mb_entry->b->vb2_buf, 1); b 1234 drivers/media/platform/s5p-mfc/s5p_mfc_enc.c vb2_buffer_done(&mb_entry->b->vb2_buf, b 1241 drivers/media/platform/s5p-mfc/s5p_mfc_enc.c &mb_entry->b->vb2_buf, 0); b 1243 drivers/media/platform/s5p-mfc/s5p_mfc_enc.c &mb_entry->b->vb2_buf, 1); b 1248 drivers/media/platform/s5p-mfc/s5p_mfc_enc.c vb2_buffer_done(&mb_entry->b->vb2_buf, b 1273 drivers/media/platform/s5p-mfc/s5p_mfc_enc.c mb_entry->b->flags |= V4L2_BUF_FLAG_KEYFRAME; b 1276 drivers/media/platform/s5p-mfc/s5p_mfc_enc.c mb_entry->b->flags |= V4L2_BUF_FLAG_PFRAME; b 1279 drivers/media/platform/s5p-mfc/s5p_mfc_enc.c mb_entry->b->flags |= V4L2_BUF_FLAG_BFRAME; b 1282 drivers/media/platform/s5p-mfc/s5p_mfc_enc.c vb2_set_plane_payload(&mb_entry->b->vb2_buf, 0, strm_size); b 1283 drivers/media/platform/s5p-mfc/s5p_mfc_enc.c vb2_buffer_done(&mb_entry->b->vb2_buf, VB2_BUF_STATE_DONE); b 2431 drivers/media/platform/s5p-mfc/s5p_mfc_enc.c ctx->dst_bufs[i].b = vbuf; b 2440 drivers/media/platform/s5p-mfc/s5p_mfc_enc.c ctx->src_bufs[i].b = vbuf; b 38 drivers/media/platform/s5p-mfc/s5p_mfc_opr.c struct s5p_mfc_priv_buf *b) b 41 drivers/media/platform/s5p-mfc/s5p_mfc_opr.c unsigned int count = b->size >> PAGE_SHIFT; b 45 drivers/media/platform/s5p-mfc/s5p_mfc_opr.c mfc_debug(3, "Allocating priv: %zu\n", b->size); b 54 drivers/media/platform/s5p-mfc/s5p_mfc_opr.c b->virt = dev->mem_virt + offset; b 55 drivers/media/platform/s5p-mfc/s5p_mfc_opr.c b->dma = dev->mem_base + offset; b 60 drivers/media/platform/s5p-mfc/s5p_mfc_opr.c b->ctx = mem_ctx; b 61 drivers/media/platform/s5p-mfc/s5p_mfc_opr.c b->virt = dma_alloc_coherent(mem_dev, b->size, &b->dma, GFP_KERNEL); b 62 drivers/media/platform/s5p-mfc/s5p_mfc_opr.c if (!b->virt) b 64 drivers/media/platform/s5p-mfc/s5p_mfc_opr.c if (b->dma < base) { b 66 drivers/media/platform/s5p-mfc/s5p_mfc_opr.c &b->dma, &base); b 67 drivers/media/platform/s5p-mfc/s5p_mfc_opr.c dma_free_coherent(mem_dev, b->size, b->virt, b->dma); b 72 drivers/media/platform/s5p-mfc/s5p_mfc_opr.c mfc_debug(3, "Allocated addr %p %pad\n", b->virt, &b->dma); b 75 drivers/media/platform/s5p-mfc/s5p_mfc_opr.c mfc_err("Allocating private buffer of size %zu failed\n", b->size); b 80 drivers/media/platform/s5p-mfc/s5p_mfc_opr.c struct s5p_mfc_priv_buf *b) b 84 drivers/media/platform/s5p-mfc/s5p_mfc_opr.c mfc_debug(3, "Allocating generic buf: %zu\n", b->size); b 86 drivers/media/platform/s5p-mfc/s5p_mfc_opr.c b->ctx = mem_ctx; b 87 drivers/media/platform/s5p-mfc/s5p_mfc_opr.c b->virt = dma_alloc_coherent(mem_dev, b->size, &b->dma, GFP_KERNEL); b 88 drivers/media/platform/s5p-mfc/s5p_mfc_opr.c if (!b->virt) b 91 drivers/media/platform/s5p-mfc/s5p_mfc_opr.c mfc_debug(3, "Allocated addr %p %pad\n", b->virt, &b->dma); b 94 drivers/media/platform/s5p-mfc/s5p_mfc_opr.c mfc_err("Allocating generic buffer of size %zu failed\n", b->size); b 99 drivers/media/platform/s5p-mfc/s5p_mfc_opr.c struct s5p_mfc_priv_buf *b) b 102 drivers/media/platform/s5p-mfc/s5p_mfc_opr.c unsigned int start = (b->dma - dev->mem_base) >> PAGE_SHIFT; b 103 drivers/media/platform/s5p-mfc/s5p_mfc_opr.c unsigned int count = b->size >> PAGE_SHIFT; b 107 drivers/media/platform/s5p-mfc/s5p_mfc_opr.c struct device *mem_dev = dev->mem_dev[b->ctx]; b 109 drivers/media/platform/s5p-mfc/s5p_mfc_opr.c dma_free_coherent(mem_dev, b->size, b->virt, b->dma); b 111 drivers/media/platform/s5p-mfc/s5p_mfc_opr.c b->virt = NULL; b 112 drivers/media/platform/s5p-mfc/s5p_mfc_opr.c b->dma = 0; b 113 drivers/media/platform/s5p-mfc/s5p_mfc_opr.c b->size = 0; b 117 drivers/media/platform/s5p-mfc/s5p_mfc_opr.c struct s5p_mfc_priv_buf *b) b 119 drivers/media/platform/s5p-mfc/s5p_mfc_opr.c struct device *mem_dev = dev->mem_dev[b->ctx]; b 120 drivers/media/platform/s5p-mfc/s5p_mfc_opr.c dma_free_coherent(mem_dev, b->size, b->virt, b->dma); b 121 drivers/media/platform/s5p-mfc/s5p_mfc_opr.c b->virt = NULL; b 122 drivers/media/platform/s5p-mfc/s5p_mfc_opr.c b->dma = 0; b 123 drivers/media/platform/s5p-mfc/s5p_mfc_opr.c b->size = 0; b 330 drivers/media/platform/s5p-mfc/s5p_mfc_opr.h struct s5p_mfc_priv_buf *b); b 332 drivers/media/platform/s5p-mfc/s5p_mfc_opr.h struct s5p_mfc_priv_buf *b); b 334 drivers/media/platform/s5p-mfc/s5p_mfc_opr.h struct s5p_mfc_priv_buf *b); b 336 drivers/media/platform/s5p-mfc/s5p_mfc_opr.h struct s5p_mfc_priv_buf *b); b 1182 drivers/media/platform/s5p-mfc/s5p_mfc_opr_v5.c vb2_dma_contig_plane_dma_addr(&temp_vb->b->vb2_buf, 0), b 1183 drivers/media/platform/s5p-mfc/s5p_mfc_opr_v5.c ctx->consumed_stream, temp_vb->b->vb2_buf.planes[0].bytesused); b 1185 drivers/media/platform/s5p-mfc/s5p_mfc_opr_v5.c if (temp_vb->b->vb2_buf.planes[0].bytesused == 0) { b 1219 drivers/media/platform/s5p-mfc/s5p_mfc_opr_v5.c if (src_mb->b->vb2_buf.planes[0].bytesused == 0) { b 1227 drivers/media/platform/s5p-mfc/s5p_mfc_opr_v5.c &src_mb->b->vb2_buf, 0); b 1229 drivers/media/platform/s5p-mfc/s5p_mfc_opr_v5.c &src_mb->b->vb2_buf, 1); b 1238 drivers/media/platform/s5p-mfc/s5p_mfc_opr_v5.c dst_addr = vb2_dma_contig_plane_dma_addr(&dst_mb->b->vb2_buf, 0); b 1239 drivers/media/platform/s5p-mfc/s5p_mfc_opr_v5.c dst_size = vb2_plane_size(&dst_mb->b->vb2_buf, 0); b 1243 drivers/media/platform/s5p-mfc/s5p_mfc_opr_v5.c src_mb ? src_mb->b->vb2_buf.index : -1, ctx->state); b 1258 drivers/media/platform/s5p-mfc/s5p_mfc_opr_v5.c temp_vb->b->vb2_buf.planes[0].bytesused); b 1260 drivers/media/platform/s5p-mfc/s5p_mfc_opr_v5.c vb2_dma_contig_plane_dma_addr(&temp_vb->b->vb2_buf, 0), b 1261 drivers/media/platform/s5p-mfc/s5p_mfc_opr_v5.c 0, temp_vb->b->vb2_buf.planes[0].bytesused); b 1275 drivers/media/platform/s5p-mfc/s5p_mfc_opr_v5.c dst_addr = vb2_dma_contig_plane_dma_addr(&dst_mb->b->vb2_buf, 0); b 1276 drivers/media/platform/s5p-mfc/s5p_mfc_opr_v5.c dst_size = vb2_plane_size(&dst_mb->b->vb2_buf, 0); b 1302 drivers/media/platform/s5p-mfc/s5p_mfc_opr_v5.c temp_vb->b->vb2_buf.planes[0].bytesused); b 1304 drivers/media/platform/s5p-mfc/s5p_mfc_opr_v5.c vb2_dma_contig_plane_dma_addr(&temp_vb->b->vb2_buf, 0), b 1305 drivers/media/platform/s5p-mfc/s5p_mfc_opr_v5.c 0, temp_vb->b->vb2_buf.planes[0].bytesused); b 1815 drivers/media/platform/s5p-mfc/s5p_mfc_opr_v6.c vb2_dma_contig_plane_dma_addr(&temp_vb->b->vb2_buf, 0), b 1817 drivers/media/platform/s5p-mfc/s5p_mfc_opr_v6.c temp_vb->b->vb2_buf.planes[0].bytesused); b 1820 drivers/media/platform/s5p-mfc/s5p_mfc_opr_v6.c if (temp_vb->b->vb2_buf.planes[0].bytesused == 0) { b 1858 drivers/media/platform/s5p-mfc/s5p_mfc_opr_v6.c if (src_mb->b->vb2_buf.planes[0].bytesused == 0) { b 1862 drivers/media/platform/s5p-mfc/s5p_mfc_opr_v6.c src_y_addr = vb2_dma_contig_plane_dma_addr(&src_mb->b->vb2_buf, 0); b 1863 drivers/media/platform/s5p-mfc/s5p_mfc_opr_v6.c src_c_addr = vb2_dma_contig_plane_dma_addr(&src_mb->b->vb2_buf, 1); b 1876 drivers/media/platform/s5p-mfc/s5p_mfc_opr_v6.c dst_addr = vb2_dma_contig_plane_dma_addr(&dst_mb->b->vb2_buf, 0); b 1877 drivers/media/platform/s5p-mfc/s5p_mfc_opr_v6.c dst_size = vb2_plane_size(&dst_mb->b->vb2_buf, 0); b 1895 drivers/media/platform/s5p-mfc/s5p_mfc_opr_v6.c mfc_debug(2, "Header size: %d\n", temp_vb->b->vb2_buf.planes[0].bytesused); b 1897 drivers/media/platform/s5p-mfc/s5p_mfc_opr_v6.c vb2_dma_contig_plane_dma_addr(&temp_vb->b->vb2_buf, 0), 0, b 1898 drivers/media/platform/s5p-mfc/s5p_mfc_opr_v6.c temp_vb->b->vb2_buf.planes[0].bytesused); b 1911 drivers/media/platform/s5p-mfc/s5p_mfc_opr_v6.c dst_addr = vb2_dma_contig_plane_dma_addr(&dst_mb->b->vb2_buf, 0); b 1912 drivers/media/platform/s5p-mfc/s5p_mfc_opr_v6.c dst_size = vb2_plane_size(&dst_mb->b->vb2_buf, 0); b 905 drivers/media/platform/sti/bdisp/bdisp-v4l2.c static int is_rect_enclosed(struct v4l2_rect *a, struct v4l2_rect *b) b 909 drivers/media/platform/sti/bdisp/bdisp-v4l2.c if (a->left < b->left || a->top < b->top) b 912 drivers/media/platform/sti/bdisp/bdisp-v4l2.c if (a->left + a->width > b->left + b->width) b 915 drivers/media/platform/sti/bdisp/bdisp-v4l2.c if (a->top + a->height > b->top + b->height) b 36 drivers/media/platform/sti/delta/delta-mjpeg-dec.c char *b = str; b 41 drivers/media/platform/sti/delta/delta-mjpeg-dec.c b += snprintf(b, len, b 54 drivers/media/platform/sti/delta/delta-mjpeg-dec.c char *b = str; b 59 drivers/media/platform/sti/delta/delta-mjpeg-dec.c b += snprintf(b, len, b 15 drivers/media/platform/sti/hva/hva-mem.c struct hva_buffer *b; b 19 drivers/media/platform/sti/hva/hva-mem.c b = devm_kzalloc(dev, sizeof(*b), GFP_KERNEL); b 20 drivers/media/platform/sti/hva/hva-mem.c if (!b) { b 31 drivers/media/platform/sti/hva/hva-mem.c devm_kfree(dev, b); b 35 drivers/media/platform/sti/hva/hva-mem.c b->size = size; b 36 drivers/media/platform/sti/hva/hva-mem.c b->paddr = paddr; b 37 drivers/media/platform/sti/hva/hva-mem.c b->vaddr = base; b 38 drivers/media/platform/sti/hva/hva-mem.c b->name = name; b 42 drivers/media/platform/sti/hva/hva-mem.c ctx->name, size, b->vaddr, &b->paddr, b->name); b 45 drivers/media/platform/sti/hva/hva-mem.c *buf = b; b 293 drivers/media/platform/vim2m.c u8 _r[2], _g[2], _b[2], *r, *g, *b; b 300 drivers/media/platform/vim2m.c b = _b; b 309 drivers/media/platform/vim2m.c *b++ = (u8)((pix & 0x1f) << 3) | 0x07; b 318 drivers/media/platform/vim2m.c *b++ = (u8)((pix & 0x1f) << 3) | 0x07; b 326 drivers/media/platform/vim2m.c *b++ = src[i][2]; b 331 drivers/media/platform/vim2m.c *b++ = src[i][0]; b 342 drivers/media/platform/vim2m.c b = _b; b 351 drivers/media/platform/vim2m.c (*b >> 3); b 364 drivers/media/platform/vim2m.c (*b >> 3); b 375 drivers/media/platform/vim2m.c *(*dst)++ = *b++; b 380 drivers/media/platform/vim2m.c *(*dst)++ = *b++; b 390 drivers/media/platform/vim2m.c y = ((8453 * (*r) + 16594 * (*g) + 3223 * (*b) b 392 drivers/media/platform/vim2m.c u = ((-4878 * (*r) - 9578 * (*g) + 14456 * (*b) b 394 drivers/media/platform/vim2m.c v = ((14456 * (*r++) - 12105 * (*g++) - 2351 * (*b++) b 396 drivers/media/platform/vim2m.c y1 = ((8453 * (*r) + 16594 * (*g) + 3223 * (*b) b 408 drivers/media/platform/vim2m.c *(*dst)++ = *b; b 418 drivers/media/platform/vim2m.c *(*dst)++ = *++b; b 429 drivers/media/platform/vim2m.c *(*dst)++ = *b; b 439 drivers/media/platform/vim2m.c *(*dst)++ = *++b; b 167 drivers/media/radio/radio-zoltrix.c int a, b; b 175 drivers/media/radio/radio-zoltrix.c b = inb(isa->io); b 177 drivers/media/radio/radio-zoltrix.c return (a == b && a == 0xcf) ? b 184 drivers/media/radio/radio-zoltrix.c int a, b; b 192 drivers/media/radio/radio-zoltrix.c b = inb(isa->io); b 194 drivers/media/radio/radio-zoltrix.c if (a != b) b 75 drivers/media/radio/si4713/si4713.c #define get_status_bit(p, b, m) (((p) & (m)) >> (b)) b 76 drivers/media/radio/si4713/si4713.c #define set_bits(p, v, b, m) (((p) & ~(m)) | ((v) << (b))) b 1147 drivers/media/rc/imon.c static int stabilize(int a, int b, u16 timeout, u16 threshold) b 1167 drivers/media/rc/imon.c y += b; b 112 drivers/media/rc/ttusbir.c unsigned i, v, b; b 133 drivers/media/rc/ttusbir.c b = ffz(v | 1); b 136 drivers/media/rc/ttusbir.c b = ffs(v) - 1; b 140 drivers/media/rc/ttusbir.c rawir.duration = NS_PER_BIT * (8 - b); b 145 drivers/media/rc/ttusbir.c rawir.duration = NS_PER_BIT * b; b 627 drivers/media/rc/winbond-cir.c wbcir_tx(struct rc_dev *dev, unsigned *b, unsigned count) b 634 drivers/media/rc/winbond-cir.c buf = kmalloc_array(count, sizeof(*b), GFP_KERNEL); b 640 drivers/media/rc/winbond-cir.c buf[i] = DIV_ROUND_CLOSEST(b[i], 10); b 35 drivers/media/tuners/mt2060.c u8 *b; b 37 drivers/media/tuners/mt2060.c b = kmalloc(2, GFP_KERNEL); b 38 drivers/media/tuners/mt2060.c if (!b) b 41 drivers/media/tuners/mt2060.c b[0] = reg; b 42 drivers/media/tuners/mt2060.c b[1] = 0; b 44 drivers/media/tuners/mt2060.c msg[0].buf = b; b 45 drivers/media/tuners/mt2060.c msg[1].buf = b + 1; b 51 drivers/media/tuners/mt2060.c *val = b[1]; b 52 drivers/media/tuners/mt2060.c kfree(b); b 197 drivers/media/tuners/mt2060.c u8 b[8]; b 203 drivers/media/tuners/mt2060.c b[0] = REG_LO1B1; b 204 drivers/media/tuners/mt2060.c b[1] = 0xFF; b 209 drivers/media/tuners/mt2060.c mt2060_writeregs(priv,b,2); b 247 drivers/media/tuners/mt2060.c b[0] = REG_LO1C1; b 248 drivers/media/tuners/mt2060.c b[1] = lnaband | ((num1 >>2) & 0x0F); b 249 drivers/media/tuners/mt2060.c b[2] = div1; b 250 drivers/media/tuners/mt2060.c b[3] = (num2 & 0x0F) | ((num1 & 3) << 4); b 251 drivers/media/tuners/mt2060.c b[4] = num2 >> 4; b 252 drivers/media/tuners/mt2060.c b[5] = ((num2 >>12) & 1) | (div2 << 1); b 257 drivers/media/tuners/mt2060.c dprintk("PLL [1..5]: %2x %2x %2x %2x %2x",(int)b[1],(int)b[2],(int)b[3],(int)b[4],(int)b[5]); b 259 drivers/media/tuners/mt2060.c mt2060_writeregs(priv,b,6); b 264 drivers/media/tuners/mt2060.c mt2060_readreg(priv,REG_LO_STATUS,b); b 265 drivers/media/tuners/mt2060.c if ((b[0] & 0x88)==0x88) b 279 drivers/media/tuners/mt2060.c u8 b = 0; b 291 drivers/media/tuners/mt2060.c b |= (1 << 6); // FM1SS; b 292 drivers/media/tuners/mt2060.c mt2060_writereg(priv, REG_LO2C1,b); b 296 drivers/media/tuners/mt2060.c b |= (1 << 7); // FM1CA; b 297 drivers/media/tuners/mt2060.c mt2060_writereg(priv, REG_LO2C1,b); b 298 drivers/media/tuners/mt2060.c b &= ~(1 << 7); // FM1CA; b 302 drivers/media/tuners/mt2060.c b &= ~(1 << 6); // FM1SS b 303 drivers/media/tuners/mt2060.c mt2060_writereg(priv, REG_LO2C1,b); b 310 drivers/media/tuners/mt2060.c while (i++ < 10 && mt2060_readreg(priv, REG_MISC_STAT, &b) == 0 && (b & (1 << 6)) == 0) b 90 drivers/media/tuners/mt2131.c u8 b[8]; b 134 drivers/media/tuners/mt2131.c b[0] = 1; b 135 drivers/media/tuners/mt2131.c b[1] = (num1 >> 5) & 0xFF; b 136 drivers/media/tuners/mt2131.c b[2] = (num1 & 0x1F); b 137 drivers/media/tuners/mt2131.c b[3] = div1; b 138 drivers/media/tuners/mt2131.c b[4] = (num2 >> 5) & 0xFF; b 139 drivers/media/tuners/mt2131.c b[5] = num2 & 0x1F; b 140 drivers/media/tuners/mt2131.c b[6] = div2; b 148 drivers/media/tuners/mt2131.c (int)b[1], (int)b[2], (int)b[3], (int)b[4], (int)b[5], b 149 drivers/media/tuners/mt2131.c (int)b[6]); b 151 drivers/media/tuners/mt2131.c ret = mt2131_writeregs(priv,b,7); b 124 drivers/media/tuners/mt2266.c u8 b[10]; b 199 drivers/media/tuners/mt2266.c b[0] = REG_TUNE; b 200 drivers/media/tuners/mt2266.c b[1] = (tune >> 8) & 0x1F; b 201 drivers/media/tuners/mt2266.c b[2] = tune & 0xFF; b 202 drivers/media/tuners/mt2266.c b[3] = tune >> 13; b 203 drivers/media/tuners/mt2266.c mt2266_writeregs(priv,b,4); b 209 drivers/media/tuners/mt2266.c (int) b[1], (int) b[2], (int)b[3]); b 212 drivers/media/tuners/mt2266.c b[0] = 0x05; b 213 drivers/media/tuners/mt2266.c b[1] = (priv->band == MT2266_VHF) ? 0x52 : 0x62; b 214 drivers/media/tuners/mt2266.c b[2] = lnaband; b 215 drivers/media/tuners/mt2266.c mt2266_writeregs(priv, b, 3); b 221 drivers/media/tuners/mt2266.c mt2266_readreg(priv,REG_LOCK,b); b 222 drivers/media/tuners/mt2266.c if (b[0] & 0x40) b 186 drivers/media/tuners/qm1d1c0042.c s32 b; b 239 drivers/media/tuners/qm1d1c0042.c b = (s32)div64_s64(((s64) freq) << 20, state->cfg.xtal_freq) b 242 drivers/media/tuners/qm1d1c0042.c if (b >= 0) b 243 drivers/media/tuners/qm1d1c0042.c sd = b; b 245 drivers/media/tuners/qm1d1c0042.c sd = (1 << 22) + b; b 51 drivers/media/tuners/tda9887.c unsigned char b; b 141 drivers/media/tuners/tda9887.c .b = ( cNegativeFmTV | b 152 drivers/media/tuners/tda9887.c .b = ( cNegativeFmTV | b 163 drivers/media/tuners/tda9887.c .b = ( cNegativeFmTV | b 174 drivers/media/tuners/tda9887.c .b = ( cNegativeFmTV | b 185 drivers/media/tuners/tda9887.c .b = ( cNegativeFmTV | b 193 drivers/media/tuners/tda9887.c .b = ( cPositiveAmTV | b 202 drivers/media/tuners/tda9887.c .b = ( cOutputPort2Inactive | b 212 drivers/media/tuners/tda9887.c .b = ( cNegativeFmTV | b 223 drivers/media/tuners/tda9887.c .b = ( cNegativeFmTV | b 234 drivers/media/tuners/tda9887.c .b = ( cNegativeFmTV | b 247 drivers/media/tuners/tda9887.c .b = ( cFmRadio | b 259 drivers/media/tuners/tda9887.c .b = ( cFmRadio | b 427 drivers/media/tuners/tda9887.c buf[1] = norm->b; b 1066 drivers/media/tuners/tuner-simple.c u8 b[1]; b 1069 drivers/media/tuners/tuner-simple.c .buf = b, .len = 1, b 728 drivers/media/tuners/xc5000.c int b; b 790 drivers/media/tuners/xc5000.c b = 6; b 794 drivers/media/tuners/xc5000.c b = 7; b 798 drivers/media/tuners/xc5000.c b = 8; b 801 drivers/media/tuners/xc5000.c b, bw); b 21 drivers/media/usb/b2c2/flexcop-usb.c #define debug_dump(b, l, method) do {\ b 24 drivers/media/usb/b2c2/flexcop-usb.c method("%02x ", b[i]); \ b 31 drivers/media/usb/b2c2/flexcop-usb.c #define debug_dump(b, l, method) b 332 drivers/media/usb/b2c2/flexcop-usb.c u8 *b; b 342 drivers/media/usb/b2c2/flexcop-usb.c b = fc_usb->tmp_buffer; b 345 drivers/media/usb/b2c2/flexcop-usb.c b=buffer; b 350 drivers/media/usb/b2c2/flexcop-usb.c if (*b == 0xff) { b 351 drivers/media/usb/b2c2/flexcop-usb.c switch (*(b+1) & 0x03) { b 353 drivers/media/usb/b2c2/flexcop-usb.c if (*(b+2) == 0x47) b 355 drivers/media/usb/b2c2/flexcop-usb.c fc_usb->fc_dev, b+2, 1); b 357 drivers/media/usb/b2c2/flexcop-usb.c deb_ts("not ts packet %*ph\n", 4, b+2); b 358 drivers/media/usb/b2c2/flexcop-usb.c b += 190; b 373 drivers/media/usb/b2c2/flexcop-usb.c memcpy(fc_usb->tmp_buffer, b, l); b 1663 drivers/media/usb/cx231xx/cx231xx-417.c static int vidioc_dqbuf(struct file *file, void *priv, struct v4l2_buffer *b) b 1667 drivers/media/usb/cx231xx/cx231xx-417.c return videobuf_dqbuf(&fh->vidq, b, file->f_flags & O_NONBLOCK); b 1662 drivers/media/usb/cx231xx/cx231xx-video.c static int vidioc_querybuf(struct file *file, void *priv, struct v4l2_buffer *b) b 1672 drivers/media/usb/cx231xx/cx231xx-video.c return videobuf_querybuf(&fh->vb_vidq, b); b 1675 drivers/media/usb/cx231xx/cx231xx-video.c static int vidioc_qbuf(struct file *file, void *priv, struct v4l2_buffer *b) b 1685 drivers/media/usb/cx231xx/cx231xx-video.c return videobuf_qbuf(&fh->vb_vidq, b); b 1688 drivers/media/usb/cx231xx/cx231xx-video.c static int vidioc_dqbuf(struct file *file, void *priv, struct v4l2_buffer *b) b 1698 drivers/media/usb/cx231xx/cx231xx-video.c return videobuf_dqbuf(&fh->vb_vidq, b, file->f_flags & O_NONBLOCK); b 101 drivers/media/usb/dvb-usb-v2/az6007.c u16 index, u8 *b, int blen) b 109 drivers/media/usb/dvb-usb-v2/az6007.c value, index, b, blen, 5000); b 119 drivers/media/usb/dvb-usb-v2/az6007.c DUMP_PREFIX_NONE, b, blen); b 126 drivers/media/usb/dvb-usb-v2/az6007.c u16 index, u8 *b, int blen) b 134 drivers/media/usb/dvb-usb-v2/az6007.c ret = __az6007_read(d->udev, req, value, index, b, blen); b 142 drivers/media/usb/dvb-usb-v2/az6007.c u16 index, u8 *b, int blen) b 150 drivers/media/usb/dvb-usb-v2/az6007.c DUMP_PREFIX_NONE, b, blen); b 163 drivers/media/usb/dvb-usb-v2/az6007.c value, index, b, blen, 5000); b 173 drivers/media/usb/dvb-usb-v2/az6007.c u16 index, u8 *b, int blen) b 181 drivers/media/usb/dvb-usb-v2/az6007.c ret = __az6007_write(d->udev, req, value, index, b, blen); b 259 drivers/media/usb/dvb-usb-v2/az6007.c u8 *b; b 264 drivers/media/usb/dvb-usb-v2/az6007.c b = kmalloc(12, GFP_KERNEL); b 265 drivers/media/usb/dvb-usb-v2/az6007.c if (!b) b 275 drivers/media/usb/dvb-usb-v2/az6007.c ret = az6007_read(d, req, value, index, b, blen); b 280 drivers/media/usb/dvb-usb-v2/az6007.c ret = b[0]; b 284 drivers/media/usb/dvb-usb-v2/az6007.c kfree(b); b 332 drivers/media/usb/dvb-usb-v2/az6007.c u8 *b; b 337 drivers/media/usb/dvb-usb-v2/az6007.c b = kmalloc(12, GFP_KERNEL); b 338 drivers/media/usb/dvb-usb-v2/az6007.c if (!b) b 348 drivers/media/usb/dvb-usb-v2/az6007.c ret = az6007_read(d, req, value, index, b, blen); b 353 drivers/media/usb/dvb-usb-v2/az6007.c if (b[0] == 0) b 356 drivers/media/usb/dvb-usb-v2/az6007.c ret = b[1]; b 357 drivers/media/usb/dvb-usb-v2/az6007.c pr_debug("read cam data = %x from 0x%x\n", b[1], value); b 361 drivers/media/usb/dvb-usb-v2/az6007.c kfree(b); b 408 drivers/media/usb/dvb-usb-v2/az6007.c u8 *b; b 410 drivers/media/usb/dvb-usb-v2/az6007.c b = kmalloc(12, GFP_KERNEL); b 411 drivers/media/usb/dvb-usb-v2/az6007.c if (!b) b 419 drivers/media/usb/dvb-usb-v2/az6007.c ret = az6007_read(d, req, value, index, b, blen); b 424 drivers/media/usb/dvb-usb-v2/az6007.c ret = b[0]; b 426 drivers/media/usb/dvb-usb-v2/az6007.c kfree(b); b 524 drivers/media/usb/dvb-usb-v2/az6007.c u8 *b; b 526 drivers/media/usb/dvb-usb-v2/az6007.c b = kmalloc(12, GFP_KERNEL); b 527 drivers/media/usb/dvb-usb-v2/az6007.c if (!b) b 536 drivers/media/usb/dvb-usb-v2/az6007.c ret = az6007_read(d, req, value, index, b, blen); b 543 drivers/media/usb/dvb-usb-v2/az6007.c if (!ret && b[0] == 1) { b 549 drivers/media/usb/dvb-usb-v2/az6007.c kfree(b); b 54 drivers/media/usb/dvb-usb-v2/dvb_usb.h #define dvb_usb_dbg_usb_control_msg(udev, r, t, v, i, b, l) { \ b 62 drivers/media/usb/dvb-usb-v2/dvb_usb.h i & 0xff, i >> 8, l & 0xff, l >> 8, direction, l, b); \ b 178 drivers/media/usb/dvb-usb-v2/mxl111sf-i2c.c u8 b = 0; b 182 drivers/media/usb/dvb-usb-v2/mxl111sf-i2c.c ret = mxl111sf_read_reg(state, SW_I2C_BUSY_ADDR, &b); b 24 drivers/media/usb/dvb-usb-v2/usb_urb.c u8 *b; b 48 drivers/media/usb/dvb-usb-v2/usb_urb.c b = (u8 *) urb->transfer_buffer; b 59 drivers/media/usb/dvb-usb-v2/usb_urb.c b + urb->iso_frame_desc[i].offset, b 68 drivers/media/usb/dvb-usb-v2/usb_urb.c stream->complete(stream, b, urb->actual_length); b 299 drivers/media/usb/dvb-usb/az6027.c u16 value, u16 index, u8 *b, int blen) b 311 drivers/media/usb/dvb-usb/az6027.c b, b 322 drivers/media/usb/dvb-usb/az6027.c debug_dump(b, blen, deb_xfer); b 332 drivers/media/usb/dvb-usb/az6027.c u8 *b, b 338 drivers/media/usb/dvb-usb/az6027.c debug_dump(b, blen, deb_xfer); b 349 drivers/media/usb/dvb-usb/az6027.c b, b 417 drivers/media/usb/dvb-usb/az6027.c u8 *b; b 422 drivers/media/usb/dvb-usb/az6027.c b = kmalloc(12, GFP_KERNEL); b 423 drivers/media/usb/dvb-usb/az6027.c if (!b) b 433 drivers/media/usb/dvb-usb/az6027.c ret = az6027_usb_in_op(d, req, value, index, b, blen); b 438 drivers/media/usb/dvb-usb/az6027.c ret = b[0]; b 442 drivers/media/usb/dvb-usb/az6027.c kfree(b); b 490 drivers/media/usb/dvb-usb/az6027.c u8 *b; b 495 drivers/media/usb/dvb-usb/az6027.c b = kmalloc(12, GFP_KERNEL); b 496 drivers/media/usb/dvb-usb/az6027.c if (!b) b 506 drivers/media/usb/dvb-usb/az6027.c ret = az6027_usb_in_op(d, req, value, index, b, blen); b 511 drivers/media/usb/dvb-usb/az6027.c if (b[0] == 0) b 514 drivers/media/usb/dvb-usb/az6027.c ret = b[1]; b 515 drivers/media/usb/dvb-usb/az6027.c deb_info("read cam data = %x from 0x%x", b[1], value); b 519 drivers/media/usb/dvb-usb/az6027.c kfree(b); b 566 drivers/media/usb/dvb-usb/az6027.c u8 *b; b 568 drivers/media/usb/dvb-usb/az6027.c b = kmalloc(12, GFP_KERNEL); b 569 drivers/media/usb/dvb-usb/az6027.c if (!b) b 577 drivers/media/usb/dvb-usb/az6027.c ret = az6027_usb_in_op(d, req, value, index, b, blen); b 582 drivers/media/usb/dvb-usb/az6027.c ret = b[0]; b 584 drivers/media/usb/dvb-usb/az6027.c kfree(b); b 682 drivers/media/usb/dvb-usb/az6027.c u8 *b; b 684 drivers/media/usb/dvb-usb/az6027.c b = kmalloc(12, GFP_KERNEL); b 685 drivers/media/usb/dvb-usb/az6027.c if (!b) b 694 drivers/media/usb/dvb-usb/az6027.c ret = az6027_usb_in_op(d, req, value, index, b, blen); b 701 drivers/media/usb/dvb-usb/az6027.c if (!ret && b[0] == 1) { b 707 drivers/media/usb/dvb-usb/az6027.c kfree(b); b 1058 drivers/media/usb/dvb-usb/az6027.c u8 *b; b 1061 drivers/media/usb/dvb-usb/az6027.c b = kmalloc(16, GFP_KERNEL); b 1062 drivers/media/usb/dvb-usb/az6027.c if (!b) b 1071 drivers/media/usb/dvb-usb/az6027.c b, b 1076 drivers/media/usb/dvb-usb/az6027.c kfree(b); b 300 drivers/media/usb/dvb-usb/cxusb.c u8 b = 0; b 305 drivers/media/usb/dvb-usb/cxusb.c return cxusb_ctrl_msg(d, CMD_POWER_ON, &b, 1, NULL, 0); b 307 drivers/media/usb/dvb-usb/cxusb.c return cxusb_ctrl_msg(d, CMD_POWER_OFF, &b, 1, NULL, 0); b 388 drivers/media/usb/dvb-usb/cxusb.c u8 b = 0; b 391 drivers/media/usb/dvb-usb/cxusb.c return cxusb_ctrl_msg(d, CMD_POWER_ON, &b, 1, NULL, 0); b 410 drivers/media/usb/dvb-usb/cxusb.c u8 b; b 417 drivers/media/usb/dvb-usb/cxusb.c cxusb_ctrl_msg(d, CMD_DIGITAL, NULL, 0, &b, 1); b 841 drivers/media/usb/dvb-usb/cxusb.c u8 b; b 882 drivers/media/usb/dvb-usb/cxusb.c NULL, 0, &b, 1); b 379 drivers/media/usb/dvb-usb/dib0700_core.c u8 *b; b 381 drivers/media/usb/dvb-usb/dib0700_core.c b = kmalloc(16, GFP_KERNEL); b 382 drivers/media/usb/dvb-usb/dib0700_core.c if (!b) b 387 drivers/media/usb/dvb-usb/dib0700_core.c REQUEST_GET_VERSION, USB_TYPE_VENDOR | USB_DIR_IN, 0, 0, b, 16, USB_CTRL_GET_TIMEOUT); b 394 drivers/media/usb/dvb-usb/dib0700_core.c kfree(b); b 62 drivers/media/usb/dvb-usb/dibusb-common.c u8 *b; b 65 drivers/media/usb/dvb-usb/dibusb-common.c b = kmalloc(3, GFP_KERNEL); b 66 drivers/media/usb/dvb-usb/dibusb-common.c if (!b) b 69 drivers/media/usb/dvb-usb/dibusb-common.c b[0] = DIBUSB_REQ_SET_IOCTL; b 70 drivers/media/usb/dvb-usb/dibusb-common.c b[1] = DIBUSB_IOCTL_CMD_POWER_MODE; b 71 drivers/media/usb/dvb-usb/dibusb-common.c b[2] = onoff ? DIBUSB_IOCTL_POWER_WAKEUP : DIBUSB_IOCTL_POWER_SLEEP; b 73 drivers/media/usb/dvb-usb/dibusb-common.c ret = dvb_usb_generic_write(d, b, 3); b 75 drivers/media/usb/dvb-usb/dibusb-common.c kfree(b); b 86 drivers/media/usb/dvb-usb/dibusb-common.c u8 *b; b 88 drivers/media/usb/dvb-usb/dibusb-common.c b = kmalloc(3, GFP_KERNEL); b 89 drivers/media/usb/dvb-usb/dibusb-common.c if (!b) b 96 drivers/media/usb/dvb-usb/dibusb-common.c b[0] = DIBUSB_REQ_SET_STREAMING_MODE; b 97 drivers/media/usb/dvb-usb/dibusb-common.c b[1] = 0x00; b 98 drivers/media/usb/dvb-usb/dibusb-common.c ret = dvb_usb_generic_write(adap->dev, b, 2); b 103 drivers/media/usb/dvb-usb/dibusb-common.c b[0] = DIBUSB_REQ_SET_IOCTL; b 104 drivers/media/usb/dvb-usb/dibusb-common.c b[1] = onoff ? DIBUSB_IOCTL_CMD_ENABLE_STREAM : DIBUSB_IOCTL_CMD_DISABLE_STREAM; b 105 drivers/media/usb/dvb-usb/dibusb-common.c ret = dvb_usb_generic_write(adap->dev, b, 3); b 108 drivers/media/usb/dvb-usb/dibusb-common.c kfree(b); b 115 drivers/media/usb/dvb-usb/dibusb-common.c u8 *b; b 121 drivers/media/usb/dvb-usb/dibusb-common.c b = kmalloc(3, GFP_KERNEL); b 122 drivers/media/usb/dvb-usb/dibusb-common.c if (!b) b 125 drivers/media/usb/dvb-usb/dibusb-common.c b[0] = DIBUSB_REQ_SET_IOCTL; b 126 drivers/media/usb/dvb-usb/dibusb-common.c b[1] = DIBUSB_IOCTL_CMD_POWER_MODE; b 127 drivers/media/usb/dvb-usb/dibusb-common.c b[2] = DIBUSB_IOCTL_POWER_WAKEUP; b 129 drivers/media/usb/dvb-usb/dibusb-common.c ret = dvb_usb_generic_write(d, b, 3); b 131 drivers/media/usb/dvb-usb/dibusb-common.c kfree(b); b 68 drivers/media/usb/dvb-usb/dibusb-mb.c u8 b[2] = { 0,0 }, b2[1]; b 71 drivers/media/usb/dvb-usb/dibusb-mb.c { .flags = 0, .buf = b, .len = 2 }, b 119 drivers/media/usb/dvb-usb/dibusb-mc-common.c u8 a,b; b 128 drivers/media/usb/dvb-usb/dibusb-mc-common.c dibusb_read_eeprom_byte(adap->dev,0x7F,&b); b 131 drivers/media/usb/dvb-usb/dibusb-mc-common.c if1 += b; b 133 drivers/media/usb/dvb-usb/dibusb-mc-common.c if1 -= b; b 135 drivers/media/usb/dvb-usb/dibusb-mc-common.c warn("LITE-ON DVB-T: Strange IF1 calibration :%2X %2X\n", a, b); b 128 drivers/media/usb/dvb-usb/digitv.c u8 b[5]; b 130 drivers/media/usb/dvb-usb/digitv.c fe->ops.tuner_ops.calc_regs(fe, b, sizeof(b)); b 133 drivers/media/usb/dvb-usb/digitv.c return digitv_ctrl_msg(adap->dev, USB_WRITE_TUNER, 0, &b[1], 4, NULL, 0); b 235 drivers/media/usb/dvb-usb/digitv.c u8 b[4] = { 0 }; b 246 drivers/media/usb/dvb-usb/digitv.c ret = digitv_ctrl_msg(d, USB_WRITE_REMOTE, 0, b, 4, NULL, 0); b 278 drivers/media/usb/dvb-usb/digitv.c u8 b[4] = { 0 }; b 281 drivers/media/usb/dvb-usb/digitv.c b[0] = 1; b 282 drivers/media/usb/dvb-usb/digitv.c digitv_ctrl_msg(d,USB_WRITE_REMOTE_TYPE,0,b,4,NULL,0); b 284 drivers/media/usb/dvb-usb/digitv.c b[0] = 0; b 285 drivers/media/usb/dvb-usb/digitv.c digitv_ctrl_msg(d,USB_WRITE_REMOTE,0,b,4,NULL,0); b 126 drivers/media/usb/dvb-usb/dvb-usb-firmware.c u8 *b = (u8 *) &fw->data[*pos]; b 133 drivers/media/usb/dvb-usb/dvb-usb-firmware.c hx->len = b[0]; b 138 drivers/media/usb/dvb-usb/dvb-usb-firmware.c hx->addr = b[1] | (b[2] << 8); b 139 drivers/media/usb/dvb-usb/dvb-usb-firmware.c hx->type = b[3]; b 143 drivers/media/usb/dvb-usb/dvb-usb-firmware.c hx->addr |= (b[4] << 24) | (b[5] << 16); b 147 drivers/media/usb/dvb-usb/dvb-usb-firmware.c memcpy(hx->data,&b[data_offs],hx->len); b 148 drivers/media/usb/dvb-usb/dvb-usb-firmware.c hx->chk = b[hx->len + data_offs]; b 34 drivers/media/usb/dvb-usb/dvb-usb.h #define debug_dump(b,l,func) {\ b 36 drivers/media/usb/dvb-usb/dvb-usb.h for (loop_ = 0; loop_ < l; loop_++) func("%02x ", b[loop_]); \ b 42 drivers/media/usb/dvb-usb/dvb-usb.h #define debug_dump(b,l,func) b 1787 drivers/media/usb/dvb-usb/dw2102.c u8 *b, *p; b 1815 drivers/media/usb/dvb-usb/dw2102.c b = (u8 *) p + i; b 1816 drivers/media/usb/dvb-usb/dw2102.c if (dw210x_op_rw(dev, 0xa0, i, 0, b , 0x40, b 30 drivers/media/usb/dvb-usb/gp8psk.c u16 index, u8 *b, int blen) b 57 drivers/media/usb/dvb-usb/gp8psk.c memcpy(b, st->data, blen); b 61 drivers/media/usb/dvb-usb/gp8psk.c debug_dump(b,blen,deb_xfer); b 69 drivers/media/usb/dvb-usb/gp8psk.c u16 index, u8 *b, int blen) b 75 drivers/media/usb/dvb-usb/gp8psk.c debug_dump(b,blen,deb_xfer); b 83 drivers/media/usb/dvb-usb/gp8psk.c memcpy(st->data, b, blen); b 256 drivers/media/usb/dvb-usb/gp8psk.c u16 index, u8 *b, int blen) b 260 drivers/media/usb/dvb-usb/gp8psk.c return gp8psk_usb_in_op(d, req, value, index, b, blen); b 264 drivers/media/usb/dvb-usb/gp8psk.c u16 index, u8 *b, int blen) b 268 drivers/media/usb/dvb-usb/gp8psk.c return gp8psk_usb_out_op(d, req, value, index, b, blen); b 134 drivers/media/usb/dvb-usb/nova-t-usb2.c u8 b; b 142 drivers/media/usb/dvb-usb/nova-t-usb2.c dibusb_read_eeprom_byte(d,i, &b); b 144 drivers/media/usb/dvb-usb/nova-t-usb2.c mac[5 - (i - 136)] = b; b 447 drivers/media/usb/dvb-usb/opera1.c u8 *b, *p; b 470 drivers/media/usb/dvb-usb/opera1.c b = (u8 *) p + i; b 472 drivers/media/usb/dvb-usb/opera1.c (dev, OPERA_WRITE_FX2, 0x0, b , fpgasize, b 561 drivers/media/usb/dvb-usb/pctv452e.c u8 *b, *rx; b 565 drivers/media/usb/dvb-usb/pctv452e.c b = kmalloc(CMD_BUFFER_SIZE + PCTV_ANSWER_LEN, GFP_KERNEL); b 566 drivers/media/usb/dvb-usb/pctv452e.c if (!b) b 569 drivers/media/usb/dvb-usb/pctv452e.c rx = b + CMD_BUFFER_SIZE; b 574 drivers/media/usb/dvb-usb/pctv452e.c b[0] = SYNC_BYTE_OUT; b 575 drivers/media/usb/dvb-usb/pctv452e.c b[1] = id; b 576 drivers/media/usb/dvb-usb/pctv452e.c b[2] = PCTV_CMD_IR; b 577 drivers/media/usb/dvb-usb/pctv452e.c b[3] = 0; b 580 drivers/media/usb/dvb-usb/pctv452e.c ret = dvb_usb_generic_rw(d, b, 4, rx, PCTV_ANSWER_LEN, 0); b 605 drivers/media/usb/dvb-usb/pctv452e.c kfree(b); b 87 drivers/media/usb/dvb-usb/technisat-usb2.c u8 *b; b 90 drivers/media/usb/dvb-usb/technisat-usb2.c b = kmalloc(64, GFP_KERNEL); b 91 drivers/media/usb/dvb-usb/technisat-usb2.c if (!b) b 109 drivers/media/usb/dvb-usb/technisat-usb2.c b[0] = I2C_SPEED_100KHZ_BIT; b 110 drivers/media/usb/dvb-usb/technisat-usb2.c b[1] = device_addr << 1; b 113 drivers/media/usb/dvb-usb/technisat-usb2.c b[0] |= rxlen; b 114 drivers/media/usb/dvb-usb/technisat-usb2.c b[1] |= 1; b 117 drivers/media/usb/dvb-usb/technisat-usb2.c memcpy(&b[2], tx, txlen); b 120 drivers/media/usb/dvb-usb/technisat-usb2.c b, 2 + txlen, b 130 drivers/media/usb/dvb-usb/technisat-usb2.c b, 64, &actual_length, 1000); b 136 drivers/media/usb/dvb-usb/technisat-usb2.c if (b[0] != I2C_STATUS_OK) { b 137 drivers/media/usb/dvb-usb/technisat-usb2.c err("i2c-error: %02x = %d", device_addr, b[0]); b 139 drivers/media/usb/dvb-usb/technisat-usb2.c if (!(b[0] == I2C_STATUS_NAK && b 145 drivers/media/usb/dvb-usb/technisat-usb2.c deb_i2c("status: %d, ", b[0]); b 148 drivers/media/usb/dvb-usb/technisat-usb2.c memcpy(rx, &b[2], rxlen); b 157 drivers/media/usb/dvb-usb/technisat-usb2.c kfree(b); b 278 drivers/media/usb/dvb-usb/technisat-usb2.c u8 *b = state->buf; b 281 drivers/media/usb/dvb-usb/technisat-usb2.c b[0] = 0; b 290 drivers/media/usb/dvb-usb/technisat-usb2.c b, 1, 500); b 391 drivers/media/usb/dvb-usb/technisat-usb2.c u8 b[16]; b 397 drivers/media/usb/dvb-usb/technisat-usb2.c if (technisat_usb2_i2c_access(d->udev, 0x50 + j / 256, ®, 1, b, 16) != 0) b 402 drivers/media/usb/dvb-usb/technisat-usb2.c deb_eeprom("%02x ", b[i]); b 408 drivers/media/usb/dvb-usb/technisat-usb2.c static u8 technisat_usb2_calc_lrc(const u8 *b, u16 length) b 412 drivers/media/usb/dvb-usb/technisat-usb2.c lrc ^= *b++; b 417 drivers/media/usb/dvb-usb/technisat-usb2.c u16 offset, u8 *b, u16 length, u8 tries) b 428 drivers/media/usb/dvb-usb/technisat-usb2.c .buf = b, b 440 drivers/media/usb/dvb-usb/technisat-usb2.c technisat_usb2_calc_lrc(b, length - 1) == b[length - 1]; b 480 drivers/media/usb/dvb-usb/ttusb2.c u8 b = onoff; b 481 drivers/media/usb/dvb-usb/ttusb2.c ttusb2_msg(d, CMD_POWER, &b, 0, NULL, 0); b 482 drivers/media/usb/dvb-usb/ttusb2.c return ttusb2_msg(d, CMD_POWER, &b, 1, NULL, 0); b 20 drivers/media/usb/dvb-usb/usb-urb.c u8 *b; b 40 drivers/media/usb/dvb-usb/usb-urb.c b = (u8 *) urb->transfer_buffer; b 48 drivers/media/usb/dvb-usb/usb-urb.c stream->complete(stream, b + urb->iso_frame_desc[i].offset, urb->iso_frame_desc[i].actual_length); b 53 drivers/media/usb/dvb-usb/usb-urb.c debug_dump(b,20,deb_uxfer); b 57 drivers/media/usb/dvb-usb/usb-urb.c stream->complete(stream, b, urb->actual_length); b 31 drivers/media/usb/dvb-usb/vp702x.c u16 value, u16 index, u8 *b, int blen) b 39 drivers/media/usb/dvb-usb/vp702x.c value, index, b, blen, b 50 drivers/media/usb/dvb-usb/vp702x.c debug_dump(b,blen,deb_xfer); b 56 drivers/media/usb/dvb-usb/vp702x.c u16 index, u8 *b, int blen) b 61 drivers/media/usb/dvb-usb/vp702x.c ret = vp702x_usb_in_op_unlocked(d, req, value, index, b, blen); b 68 drivers/media/usb/dvb-usb/vp702x.c u16 value, u16 index, u8 *b, int blen) b 72 drivers/media/usb/dvb-usb/vp702x.c debug_dump(b,blen,deb_xfer); b 78 drivers/media/usb/dvb-usb/vp702x.c value,index,b,blen, b 87 drivers/media/usb/dvb-usb/vp702x.c u16 index, u8 *b, int blen) b 92 drivers/media/usb/dvb-usb/vp702x.c ret = vp702x_usb_out_op_unlocked(d, req, value, index, b, blen); b 222 drivers/media/usb/dvb-usb/vp702x.c u8 *b; b 234 drivers/media/usb/dvb-usb/vp702x.c b = dst->buf; b 235 drivers/media/usb/dvb-usb/vp702x.c memset(b, 0, 10); b 236 drivers/media/usb/dvb-usb/vp702x.c vp702x_usb_in_op(adap->dev, 0xb5, 3, 0, b, 10); b 237 drivers/media/usb/dvb-usb/vp702x.c vp702x_usb_in_op(adap->dev, 0xb5, 0, 0, b, 10); b 238 drivers/media/usb/dvb-usb/vp702x.c vp702x_usb_in_op(adap->dev, 0xb5, 1, 0, b, 10); b 112 drivers/media/usb/dvb-usb/vp702x.h extern int vp702x_usb_in_op(struct dvb_usb_device *d, u8 req, u16 value, u16 index, u8 *b, int blen); b 86 drivers/media/usb/em28xx/em28xx-input.c unsigned char b; b 89 drivers/media/usb/em28xx/em28xx-input.c rc = i2c_master_recv(i2c_dev, &b, 1); b 101 drivers/media/usb/em28xx/em28xx-input.c if (b == 0xff) b 104 drivers/media/usb/em28xx/em28xx-input.c if (b == 0xfe) b 109 drivers/media/usb/em28xx/em28xx-input.c *scancode = b; b 54 drivers/media/usb/go7007/go7007-fw.c int b; /* bit position of most recently-written bit */ b 61 drivers/media/usb/go7007/go7007-fw.c name.b -= (length); \ b 62 drivers/media/usb/go7007/go7007-fw.c name.a |= (val) << name.b; \ b 63 drivers/media/usb/go7007/go7007-fw.c while (name.b <= 24) { \ b 67 drivers/media/usb/go7007/go7007-fw.c name.b += 8; \ b 72 drivers/media/usb/go7007/go7007-fw.c #define CODE_LENGTH(name) (name.len + (32 - name.b)) b 438 drivers/media/usb/gspca/cpia1.c u8 a, u8 b, u8 c, u8 d) b 466 drivers/media/usb/gspca/cpia1.c cmd[3] = b; b 560 drivers/media/usb/gspca/cpia1.c u8 a, u8 b, u8 c, u8 d, b 569 drivers/media/usb/gspca/cpia1.c cmd[3] = b; b 629 drivers/media/usb/gspca/etoms.c __u8 r, g, b; b 641 drivers/media/usb/gspca/etoms.c b = gspca_dev->usb_buf[2]; b 643 drivers/media/usb/gspca/etoms.c b = ((b << 7) >> 10); b 645 drivers/media/usb/gspca/etoms.c luma = LIMIT(r + g + b); b 401 drivers/media/usb/gspca/pac7302.c static const int b[9] = b 409 drivers/media/usb/gspca/pac7302.c v += b[i]; b 1700 drivers/media/usb/gspca/sonixj.c int a, b; b 1706 drivers/media/usb/gspca/sonixj.c b = a * a * 4; /* heuristic */ b 1707 drivers/media/usb/gspca/sonixj.c i2c_w1(gspca_dev, 0x03, b >> 8); b 1708 drivers/media/usb/gspca/sonixj.c i2c_w1(gspca_dev, 0x04, b); b 155 drivers/media/usb/gspca/w996Xcf.c u16 a, b; b 162 drivers/media/usb/gspca/w996Xcf.c b = UV_QUANTABLE[j] | ((unsigned)(UV_QUANTABLE[j + 1]) << 8); b 164 drivers/media/usb/gspca/w996Xcf.c reg_w(sd, 0x60 + i, b); b 3139 drivers/media/usb/pvrusb2/pvrusb2-hdw.c u32 b; b 3141 drivers/media/usb/pvrusb2/pvrusb2-hdw.c pvr2_hdw_gpio_get_out(hdw,&b); b 68 drivers/media/usb/tm6000/tm6000-i2c.c u8 b[2]; b 92 drivers/media/usb/tm6000/tm6000-i2c.c REQ_16_SET_GET_I2C_WR1_RDN, addr | reg << 8, 0, b, len); b 94 drivers/media/usb/tm6000/tm6000-i2c.c *buf = b[1]; b 359 drivers/media/usb/tm6000/tm6000.h struct v4l2_buffer *b); b 360 drivers/media/usb/tm6000/tm6000.h int tm6000_vidioc_qbuf(struct file *file, void *priv, struct v4l2_buffer *b); b 361 drivers/media/usb/tm6000/tm6000.h int tm6000_vidioc_dqbuf(struct file *file, void *priv, struct v4l2_buffer *b); b 198 drivers/media/usb/ttusb-budget/dvb-ttusb-budget.c u8 b[0x28]; b 205 drivers/media/usb/ttusb-budget/dvb-ttusb-budget.c b[0] = 0xaa; b 206 drivers/media/usb/ttusb-budget/dvb-ttusb-budget.c b[1] = id; b 207 drivers/media/usb/ttusb-budget/dvb-ttusb-budget.c b[2] = 0x31; b 208 drivers/media/usb/ttusb-budget/dvb-ttusb-budget.c b[3] = snd_len + 3; b 209 drivers/media/usb/ttusb-budget/dvb-ttusb-budget.c b[4] = addr << 1; b 210 drivers/media/usb/ttusb-budget/dvb-ttusb-budget.c b[5] = snd_len; b 211 drivers/media/usb/ttusb-budget/dvb-ttusb-budget.c b[6] = rcv_len; b 214 drivers/media/usb/ttusb-budget/dvb-ttusb-budget.c b[7 + i] = snd_buf[i]; b 216 drivers/media/usb/ttusb-budget/dvb-ttusb-budget.c err = ttusb_cmd(ttusb, b, snd_len + 7, 1); b 221 drivers/media/usb/ttusb-budget/dvb-ttusb-budget.c err = ttusb_result(ttusb, b, 0x20); b 224 drivers/media/usb/ttusb-budget/dvb-ttusb-budget.c if ((snd_len != b[5]) || (rcv_len != b[6])) return -EREMOTEIO; b 228 drivers/media/usb/ttusb-budget/dvb-ttusb-budget.c if (err || b[0] != 0x55 || b[1] != id) { b 236 drivers/media/usb/ttusb-budget/dvb-ttusb-budget.c rcv_buf[i] = b[7 + i]; b 290 drivers/media/usb/ttusb-budget/dvb-ttusb-budget.c u8 b[40]; b 300 drivers/media/usb/ttusb-budget/dvb-ttusb-budget.c b[0] = 0xaa; b 301 drivers/media/usb/ttusb-budget/dvb-ttusb-budget.c b[2] = 0x13; b 302 drivers/media/usb/ttusb-budget/dvb-ttusb-budget.c b[3] = 28; b 307 drivers/media/usb/ttusb-budget/dvb-ttusb-budget.c memcpy(&b[4], &fw->data[i], 28); b 309 drivers/media/usb/ttusb-budget/dvb-ttusb-budget.c b[1] = ++ttusb->c; b 311 drivers/media/usb/ttusb-budget/dvb-ttusb-budget.c err = ttusb_cmd(ttusb, b, 32, 0); b 317 drivers/media/usb/ttusb-budget/dvb-ttusb-budget.c b[1] = ++ttusb->c; b 318 drivers/media/usb/ttusb-budget/dvb-ttusb-budget.c b[2] = 0x13; b 319 drivers/media/usb/ttusb-budget/dvb-ttusb-budget.c b[3] = 0; b 321 drivers/media/usb/ttusb-budget/dvb-ttusb-budget.c err = ttusb_cmd(ttusb, b, 4, 0); b 326 drivers/media/usb/ttusb-budget/dvb-ttusb-budget.c b[1] = ++ttusb->c; b 327 drivers/media/usb/ttusb-budget/dvb-ttusb-budget.c b[2] = 0x14; b 328 drivers/media/usb/ttusb-budget/dvb-ttusb-budget.c b[3] = 0; b 330 drivers/media/usb/ttusb-budget/dvb-ttusb-budget.c err = ttusb_cmd(ttusb, b, 4, 0); b 347 drivers/media/usb/ttusb-budget/dvb-ttusb-budget.c u8 b[] = { 0xaa, ++ttusb->c, 0x22, 4, chan_id, filter_type, b 351 drivers/media/usb/ttusb-budget/dvb-ttusb-budget.c err = ttusb_cmd(ttusb, b, sizeof(b), 0); b 359 drivers/media/usb/ttusb-budget/dvb-ttusb-budget.c u8 b[] = { 0xaa, ++ttusb->c, 0x23, 1, channel_id }; b 361 drivers/media/usb/ttusb-budget/dvb-ttusb-budget.c err = ttusb_cmd(ttusb, b, sizeof(b), 0); b 371 drivers/media/usb/ttusb-budget/dvb-ttusb-budget.c u8 b[] = { 0xaa, 0, 0x24, 0x1a, filter_id, associated_chan, b 380 drivers/media/usb/ttusb-budget/dvb-ttusb-budget.c err = ttusb_cmd(ttusb, b, sizeof(b), 0); b 388 drivers/media/usb/ttusb-budget/dvb-ttusb-budget.c u8 b[] = { 0xaa, ++ttusb->c, 0x25, 1, filter_id }; b 390 drivers/media/usb/ttusb-budget/dvb-ttusb-budget.c err = ttusb_cmd(ttusb, b, sizeof(b), 0); b 472 drivers/media/usb/ttusb-budget/dvb-ttusb-budget.c u8 b[12] = { 0xaa, ++ttusb->c, 0x18 }; b 476 drivers/media/usb/ttusb-budget/dvb-ttusb-budget.c b[3] = 4 + 2 + cmd->msg_len; b 477 drivers/media/usb/ttusb-budget/dvb-ttusb-budget.c b[4] = 0xFF; /* send diseqc master, not burst */ b 478 drivers/media/usb/ttusb-budget/dvb-ttusb-budget.c b[5] = cmd->msg_len; b 480 drivers/media/usb/ttusb-budget/dvb-ttusb-budget.c memcpy(b + 5, cmd->msg, cmd->msg_len); b 483 drivers/media/usb/ttusb-budget/dvb-ttusb-budget.c if ((err = ttusb_cmd(ttusb, b, 4 + b[3], 0))) { b 494 drivers/media/usb/ttusb-budget/dvb-ttusb-budget.c u8 b[] = { 0xaa, ++ttusb->c, 0x16, 5, /*power: */ 1, b 501 drivers/media/usb/ttusb-budget/dvb-ttusb-budget.c if ((err = ttusb_cmd(ttusb, b, sizeof(b), 0))) { b 532 drivers/media/usb/ttusb-budget/dvb-ttusb-budget.c u8 b[] = { 0xaa, ++ttusb->c, 0x19, 1, freq }; b 535 drivers/media/usb/ttusb-budget/dvb-ttusb-budget.c err = ttusb_cmd(ttusb, b, sizeof(b), 0); b 1374 drivers/media/usb/ttusb-budget/dvb-ttusb-budget.c u8 b = 0xff; b 1376 drivers/media/usb/ttusb-budget/dvb-ttusb-budget.c struct i2c_msg msg[] = { { .addr = 0x50,.flags = 0,.buf = &b,.len = 1 }, b 318 drivers/media/usb/ttusb-dec/ttusb_dec.c u8 *b; b 322 drivers/media/usb/ttusb-dec/ttusb_dec.c b = kzalloc(COMMAND_PACKET_SIZE + 4, GFP_KERNEL); b 323 drivers/media/usb/ttusb-dec/ttusb_dec.c if (!b) b 327 drivers/media/usb/ttusb-dec/ttusb_dec.c kfree(b); b 332 drivers/media/usb/ttusb-dec/ttusb_dec.c b[0] = 0xaa; b 333 drivers/media/usb/ttusb-dec/ttusb_dec.c b[1] = ++dec->trans_count; b 334 drivers/media/usb/ttusb-dec/ttusb_dec.c b[2] = command; b 335 drivers/media/usb/ttusb-dec/ttusb_dec.c b[3] = param_length; b 338 drivers/media/usb/ttusb-dec/ttusb_dec.c memcpy(&b[4], params, param_length); b 342 drivers/media/usb/ttusb-dec/ttusb_dec.c __func__, param_length, b); b 345 drivers/media/usb/ttusb-dec/ttusb_dec.c result = usb_bulk_msg(dec->udev, dec->command_pipe, b, b 352 drivers/media/usb/ttusb-dec/ttusb_dec.c kfree(b); b 356 drivers/media/usb/ttusb-dec/ttusb_dec.c result = usb_bulk_msg(dec->udev, dec->result_pipe, b, b 363 drivers/media/usb/ttusb-dec/ttusb_dec.c kfree(b); b 368 drivers/media/usb/ttusb-dec/ttusb_dec.c __func__, actual_len, b); b 372 drivers/media/usb/ttusb-dec/ttusb_dec.c *result_length = b[3]; b 373 drivers/media/usb/ttusb-dec/ttusb_dec.c if (cmd_result && b[3] > 0) b 374 drivers/media/usb/ttusb-dec/ttusb_dec.c memcpy(cmd_result, &b[4], b[3]); b 378 drivers/media/usb/ttusb-dec/ttusb_dec.c kfree(b); b 438 drivers/media/usb/ttusb-dec/ttusb_dec.c u8 b[] = { 0x00, 0x00, 0x00, 0x00, b 448 drivers/media/usb/ttusb-dec/ttusb_dec.c memcpy(&b[0], &pcr, 2); b 449 drivers/media/usb/ttusb-dec/ttusb_dec.c memcpy(&b[2], &audio, 2); b 450 drivers/media/usb/ttusb-dec/ttusb_dec.c memcpy(&b[4], &video, 2); b 452 drivers/media/usb/ttusb-dec/ttusb_dec.c ttusb_dec_send_command(dec, 0x50, sizeof(b), b, NULL, NULL); b 635 drivers/media/usb/ttusb-dec/ttusb_dec.c static void swap_bytes(u8 *b, int length) b 638 drivers/media/usb/ttusb-dec/ttusb_dec.c for (; length; b += 2, length -= 2) b 639 drivers/media/usb/ttusb-dec/ttusb_dec.c swap(*b, *(b + 1)); b 642 drivers/media/usb/ttusb-dec/ttusb_dec.c static void ttusb_dec_process_urb_frame(struct ttusb_dec *dec, u8 *b, b 645 drivers/media/usb/ttusb-dec/ttusb_dec.c swap_bytes(b, length); b 653 drivers/media/usb/ttusb-dec/ttusb_dec.c if (*b++ == 0xaa) b 662 drivers/media/usb/ttusb-dec/ttusb_dec.c if (*b == 0x00) { b 665 drivers/media/usb/ttusb-dec/ttusb_dec.c } else if (*b != 0xaa) { b 669 drivers/media/usb/ttusb-dec/ttusb_dec.c b++; b 674 drivers/media/usb/ttusb-dec/ttusb_dec.c dec->packet[dec->packet_length++] = *b++; b 703 drivers/media/usb/ttusb-dec/ttusb_dec.c dec->packet[dec->packet_length++] = *b++; b 729 drivers/media/usb/ttusb-dec/ttusb_dec.c b, remainder); b 731 drivers/media/usb/ttusb-dec/ttusb_dec.c b += remainder; b 736 drivers/media/usb/ttusb-dec/ttusb_dec.c b, length); b 747 drivers/media/usb/ttusb-dec/ttusb_dec.c dec->packet[dec->packet_length++] = *b++; b 805 drivers/media/usb/ttusb-dec/ttusb_dec.c u8 *b; b 810 drivers/media/usb/ttusb-dec/ttusb_dec.c b = urb->transfer_buffer + d->offset; b 817 drivers/media/usb/ttusb-dec/ttusb_dec.c memcpy(frame->data, b, length); b 898 drivers/media/usb/ttusb-dec/ttusb_dec.c u8 b[] = { 0x05 }; b 906 drivers/media/usb/ttusb-dec/ttusb_dec.c result = ttusb_dec_send_command(dec, 0x80, sizeof(b), b 907 drivers/media/usb/ttusb-dec/ttusb_dec.c b, NULL, NULL); b 1218 drivers/media/usb/ttusb-dec/ttusb_dec.c u8 b[] = { 0x00, 0x01 }; b 1249 drivers/media/usb/ttusb-dec/ttusb_dec.c ttusb_dec_send_command(dec,0xb0,sizeof(b),b,NULL,NULL); b 1313 drivers/media/usb/ttusb-dec/ttusb_dec.c u8 *b; b 1375 drivers/media/usb/ttusb-dec/ttusb_dec.c b = kmalloc(ARM_PACKET_SIZE, GFP_KERNEL); b 1376 drivers/media/usb/ttusb-dec/ttusb_dec.c if (b == NULL) { b 1386 drivers/media/usb/ttusb-dec/ttusb_dec.c b[j + 0] = 0xaa; b 1387 drivers/media/usb/ttusb-dec/ttusb_dec.c b[j + 1] = trans_count++; b 1388 drivers/media/usb/ttusb-dec/ttusb_dec.c b[j + 2] = 0xf0; b 1389 drivers/media/usb/ttusb-dec/ttusb_dec.c b[j + 3] = size; b 1390 drivers/media/usb/ttusb-dec/ttusb_dec.c memcpy(&b[j + 4], &firmware[i], size); b 1395 drivers/media/usb/ttusb-dec/ttusb_dec.c result = usb_bulk_msg(dec->udev, dec->command_pipe, b, b 1400 drivers/media/usb/ttusb-dec/ttusb_dec.c result = usb_bulk_msg(dec->udev, dec->command_pipe, b, b 1409 drivers/media/usb/ttusb-dec/ttusb_dec.c kfree(b); b 40 drivers/media/usb/ttusb-dec/ttusbdecfe.c u8 b[] = { 0x00, 0x00, 0x00, 0x00, b 47 drivers/media/usb/ttusb-dec/ttusbdecfe.c ret=state->config->send_command(fe, 0x73, sizeof(b), b, &len, result); b 80 drivers/media/usb/ttusb-dec/ttusbdecfe.c u8 b[] = { 0x00, 0x00, 0x00, 0x03, b 87 drivers/media/usb/ttusb-dec/ttusbdecfe.c memcpy(&b[4], &freq, sizeof (u32)); b 88 drivers/media/usb/ttusb-dec/ttusbdecfe.c state->config->send_command(fe, 0x71, sizeof(b), b, NULL, NULL); b 108 drivers/media/usb/ttusb-dec/ttusbdecfe.c u8 b[] = { 0x00, 0x00, 0x00, 0x01, b 125 drivers/media/usb/ttusb-dec/ttusbdecfe.c memcpy(&b[4], &freq, sizeof(u32)); b 127 drivers/media/usb/ttusb-dec/ttusbdecfe.c memcpy(&b[12], &sym_rate, sizeof(u32)); b 129 drivers/media/usb/ttusb-dec/ttusbdecfe.c memcpy(&b[24], &band, sizeof(u32)); b 131 drivers/media/usb/ttusb-dec/ttusbdecfe.c memcpy(&b[28], &lnb_voltage, sizeof(u32)); b 133 drivers/media/usb/ttusb-dec/ttusbdecfe.c state->config->send_command(fe, 0x71, sizeof(b), b, NULL, NULL); b 141 drivers/media/usb/ttusb-dec/ttusbdecfe.c u8 b[] = { 0x00, 0xff, 0x00, 0x00, b 145 drivers/media/usb/ttusb-dec/ttusbdecfe.c if (cmd->msg_len > sizeof(b) - 4) b 148 drivers/media/usb/ttusb-dec/ttusbdecfe.c memcpy(&b[4], cmd->msg, cmd->msg_len); b 151 drivers/media/usb/ttusb-dec/ttusbdecfe.c sizeof(b) - (6 - cmd->msg_len), b, b 2022 drivers/media/v4l2-core/v4l2-ioctl.c struct v4l2_buffer *b = arg; b 2023 drivers/media/v4l2-core/v4l2-ioctl.c int ret = check_fmt(file, b->type); b 2025 drivers/media/v4l2-core/v4l2-ioctl.c return ret ? ret : ops->vidioc_prepare_buf(file, fh, b); b 132 drivers/media/v4l2-core/v4l2-mem2mem.c struct v4l2_m2m_buffer *b; b 142 drivers/media/v4l2-core/v4l2-mem2mem.c b = list_first_entry(&q_ctx->rdy_queue, struct v4l2_m2m_buffer, list); b 144 drivers/media/v4l2-core/v4l2-mem2mem.c return &b->vb; b 150 drivers/media/v4l2-core/v4l2-mem2mem.c struct v4l2_m2m_buffer *b; b 160 drivers/media/v4l2-core/v4l2-mem2mem.c b = list_last_entry(&q_ctx->rdy_queue, struct v4l2_m2m_buffer, list); b 162 drivers/media/v4l2-core/v4l2-mem2mem.c return &b->vb; b 168 drivers/media/v4l2-core/v4l2-mem2mem.c struct v4l2_m2m_buffer *b; b 176 drivers/media/v4l2-core/v4l2-mem2mem.c b = list_first_entry(&q_ctx->rdy_queue, struct v4l2_m2m_buffer, list); b 177 drivers/media/v4l2-core/v4l2-mem2mem.c list_del(&b->list); b 181 drivers/media/v4l2-core/v4l2-mem2mem.c return &b->vb; b 188 drivers/media/v4l2-core/v4l2-mem2mem.c struct v4l2_m2m_buffer *b; b 192 drivers/media/v4l2-core/v4l2-mem2mem.c b = container_of(vbuf, struct v4l2_m2m_buffer, vb); b 193 drivers/media/v4l2-core/v4l2-mem2mem.c list_del(&b->list); b 203 drivers/media/v4l2-core/v4l2-mem2mem.c struct v4l2_m2m_buffer *b, *tmp; b 208 drivers/media/v4l2-core/v4l2-mem2mem.c list_for_each_entry_safe(b, tmp, &q_ctx->rdy_queue, list) { b 209 drivers/media/v4l2-core/v4l2-mem2mem.c if (b->vb.vb2_buf.index == idx) { b 210 drivers/media/v4l2-core/v4l2-mem2mem.c list_del(&b->list); b 212 drivers/media/v4l2-core/v4l2-mem2mem.c ret = &b->vb; b 962 drivers/media/v4l2-core/v4l2-mem2mem.c struct v4l2_m2m_buffer *b = container_of(vbuf, b 972 drivers/media/v4l2-core/v4l2-mem2mem.c list_add_tail(&b->list, &q_ctx->rdy_queue); b 317 drivers/media/v4l2-core/videobuf-core.c static void videobuf_status(struct videobuf_queue *q, struct v4l2_buffer *b, b 323 drivers/media/v4l2-core/videobuf-core.c b->index = vb->i; b 324 drivers/media/v4l2-core/videobuf-core.c b->type = type; b 326 drivers/media/v4l2-core/videobuf-core.c b->memory = vb->memory; b 327 drivers/media/v4l2-core/videobuf-core.c switch (b->memory) { b 329 drivers/media/v4l2-core/videobuf-core.c b->m.offset = vb->boff; b 330 drivers/media/v4l2-core/videobuf-core.c b->length = vb->bsize; b 333 drivers/media/v4l2-core/videobuf-core.c b->m.userptr = vb->baddr; b 334 drivers/media/v4l2-core/videobuf-core.c b->length = vb->bsize; b 337 drivers/media/v4l2-core/videobuf-core.c b->m.offset = vb->boff; b 344 drivers/media/v4l2-core/videobuf-core.c b->flags = V4L2_BUF_FLAG_TIMESTAMP_MONOTONIC; b 346 drivers/media/v4l2-core/videobuf-core.c b->flags |= V4L2_BUF_FLAG_MAPPED; b 352 drivers/media/v4l2-core/videobuf-core.c b->flags |= V4L2_BUF_FLAG_QUEUED; b 355 drivers/media/v4l2-core/videobuf-core.c b->flags |= V4L2_BUF_FLAG_ERROR; b 358 drivers/media/v4l2-core/videobuf-core.c b->flags |= V4L2_BUF_FLAG_DONE; b 366 drivers/media/v4l2-core/videobuf-core.c b->field = vb->field; b 367 drivers/media/v4l2-core/videobuf-core.c b->timestamp = ns_to_timeval(vb->ts); b 368 drivers/media/v4l2-core/videobuf-core.c b->bytesused = vb->size; b 369 drivers/media/v4l2-core/videobuf-core.c b->sequence = vb->field_count >> 1; b 500 drivers/media/v4l2-core/videobuf-core.c int videobuf_querybuf(struct videobuf_queue *q, struct v4l2_buffer *b) b 505 drivers/media/v4l2-core/videobuf-core.c if (unlikely(b->type != q->type)) { b 509 drivers/media/v4l2-core/videobuf-core.c if (unlikely(b->index >= VIDEO_MAX_FRAME)) { b 513 drivers/media/v4l2-core/videobuf-core.c if (unlikely(NULL == q->bufs[b->index])) { b 518 drivers/media/v4l2-core/videobuf-core.c videobuf_status(q, b, q->bufs[b->index], q->type); b 527 drivers/media/v4l2-core/videobuf-core.c int videobuf_qbuf(struct videobuf_queue *q, struct v4l2_buffer *b) b 536 drivers/media/v4l2-core/videobuf-core.c if (b->memory == V4L2_MEMORY_MMAP) b 546 drivers/media/v4l2-core/videobuf-core.c if (b->type != q->type) { b 550 drivers/media/v4l2-core/videobuf-core.c if (b->index >= VIDEO_MAX_FRAME) { b 554 drivers/media/v4l2-core/videobuf-core.c buf = q->bufs[b->index]; b 560 drivers/media/v4l2-core/videobuf-core.c if (buf->memory != b->memory) { b 569 drivers/media/v4l2-core/videobuf-core.c switch (b->memory) { b 579 drivers/media/v4l2-core/videobuf-core.c buf->size = b->bytesused; b 580 drivers/media/v4l2-core/videobuf-core.c buf->field = b->field; b 581 drivers/media/v4l2-core/videobuf-core.c buf->ts = v4l2_timeval_to_ns(&b->timestamp); b 585 drivers/media/v4l2-core/videobuf-core.c if (b->length < buf->bsize) { b 590 drivers/media/v4l2-core/videobuf-core.c buf->baddr != b->m.userptr) b 592 drivers/media/v4l2-core/videobuf-core.c buf->baddr = b->m.userptr; b 595 drivers/media/v4l2-core/videobuf-core.c buf->boff = b->m.offset; b 623 drivers/media/v4l2-core/videobuf-core.c if (b->memory == V4L2_MEMORY_MMAP) b 695 drivers/media/v4l2-core/videobuf-core.c struct v4l2_buffer *b, int nonblocking) b 702 drivers/media/v4l2-core/videobuf-core.c memset(b, 0, sizeof(*b)); b 724 drivers/media/v4l2-core/videobuf-core.c videobuf_status(q, b, buf, q->type); b 727 drivers/media/v4l2-core/videobuf-core.c b->flags &= ~V4L2_BUF_FLAG_DONE; b 935 drivers/memory/tegra/tegra124-emc.c const struct emc_timing *b = _b; b 937 drivers/memory/tegra/tegra124-emc.c if (a->rate < b->rate) b 939 drivers/memory/tegra/tegra124-emc.c else if (a->rate == b->rate) b 322 drivers/memory/tegra/tegra20-emc.c const struct emc_timing *b = _b; b 324 drivers/memory/tegra/tegra20-emc.c if (a->rate < b->rate) b 327 drivers/memory/tegra/tegra20-emc.c if (a->rate > b->rate) b 265 drivers/message/fusion/mptfc.c mptfc_FcDevPage0_cmp_func(const void *a, const void *b) b 268 drivers/message/fusion/mptfc.c FCDevicePage0_t **bb = (FCDevicePage0_t **)b; b 2283 drivers/mfd/ab8500-debugfs.c static int strval_len(char *b) b 2285 drivers/mfd/ab8500-debugfs.c char *s = b; b 2301 drivers/mfd/ab8500-debugfs.c return (int) (s-b); b 2308 drivers/mfd/ab8500-debugfs.c static ssize_t hwreg_common_write(char *b, struct hwreg_cfg *cfg, b 2323 drivers/mfd/ab8500-debugfs.c if (!strncmp(b, "read ", 5)) { b 2325 drivers/mfd/ab8500-debugfs.c b += 5; b 2326 drivers/mfd/ab8500-debugfs.c } else if (!strncmp(b, "write ", 6)) { b 2328 drivers/mfd/ab8500-debugfs.c b += 6; b 2333 drivers/mfd/ab8500-debugfs.c while ((*b == ' ') || (*b == '-')) { b 2334 drivers/mfd/ab8500-debugfs.c if (*(b-1) != ' ') { b 2335 drivers/mfd/ab8500-debugfs.c b++; b 2338 drivers/mfd/ab8500-debugfs.c if ((!strncmp(b, "-d ", 3)) || b 2339 drivers/mfd/ab8500-debugfs.c (!strncmp(b, "-dec ", 5))) { b 2340 drivers/mfd/ab8500-debugfs.c b += (*(b+2) == ' ') ? 3 : 5; b 2342 drivers/mfd/ab8500-debugfs.c } else if ((!strncmp(b, "-h ", 3)) || b 2343 drivers/mfd/ab8500-debugfs.c (!strncmp(b, "-hex ", 5))) { b 2344 drivers/mfd/ab8500-debugfs.c b += (*(b+2) == ' ') ? 3 : 5; b 2346 drivers/mfd/ab8500-debugfs.c } else if ((!strncmp(b, "-m ", 3)) || b 2347 drivers/mfd/ab8500-debugfs.c (!strncmp(b, "-mask ", 6))) { b 2348 drivers/mfd/ab8500-debugfs.c b += (*(b+2) == ' ') ? 3 : 6; b 2349 drivers/mfd/ab8500-debugfs.c if (strval_len(b) == 0) b 2351 drivers/mfd/ab8500-debugfs.c ret = kstrtoul(b, 0, &loc.mask); b 2354 drivers/mfd/ab8500-debugfs.c } else if ((!strncmp(b, "-s ", 3)) || b 2355 drivers/mfd/ab8500-debugfs.c (!strncmp(b, "-shift ", 7))) { b 2356 drivers/mfd/ab8500-debugfs.c b += (*(b+2) == ' ') ? 3 : 7; b 2357 drivers/mfd/ab8500-debugfs.c if (strval_len(b) == 0) b 2359 drivers/mfd/ab8500-debugfs.c ret = kstrtol(b, 0, &loc.shift); b 2367 drivers/mfd/ab8500-debugfs.c if (strval_len(b) == 0) b 2369 drivers/mfd/ab8500-debugfs.c ret = kstrtouint(b, 0, &loc.bank); b 2372 drivers/mfd/ab8500-debugfs.c while (*b == ' ') b 2373 drivers/mfd/ab8500-debugfs.c b++; b 2374 drivers/mfd/ab8500-debugfs.c if (strval_len(b) == 0) b 2376 drivers/mfd/ab8500-debugfs.c ret = kstrtoul(b, 0, &loc.addr); b 2381 drivers/mfd/ab8500-debugfs.c while (*b == ' ') b 2382 drivers/mfd/ab8500-debugfs.c b++; b 2383 drivers/mfd/ab8500-debugfs.c if (strval_len(b) == 0) b 2385 drivers/mfd/ab8500-debugfs.c ret = kstrtouint(b, 0, &val); b 83 drivers/mfd/lpc_ich.c #define wdt_res(b, i) (&wdt_ich_res[(b) + (i)]) b 376 drivers/mfd/menelaus.c int b; b 382 drivers/mfd/menelaus.c b = menelaus_read_reg(MENELAUS_MCT_CTRL2); b 383 drivers/mfd/menelaus.c b &= ~(MCT_CTRL2_VS2_SEL_D0 | MCT_CTRL2_VS2_SEL_D1); b 384 drivers/mfd/menelaus.c b |= power; b 385 drivers/mfd/menelaus.c ret = menelaus_write_reg(MENELAUS_MCT_CTRL2, b); b 37 drivers/mfd/qcom-pm8xxx.c #define PM8821_SSBI_REG(m, b, offset) \ b 39 drivers/mfd/qcom-pm8xxx.c (PM8821_SSBI_REG_ADDR_IRQ_MASTER0 + b + offset) : \ b 40 drivers/mfd/qcom-pm8xxx.c (PM8821_SSBI_REG_ADDR_IRQ_MASTER1 + b + offset)) b 41 drivers/mfd/qcom-pm8xxx.c #define PM8821_SSBI_ADDR_IRQ_ROOT(m, b) PM8821_SSBI_REG(m, b, 0x0) b 42 drivers/mfd/qcom-pm8xxx.c #define PM8821_SSBI_ADDR_IRQ_CLEAR(m, b) PM8821_SSBI_REG(m, b, 0x01) b 43 drivers/mfd/qcom-pm8xxx.c #define PM8821_SSBI_ADDR_IRQ_MASK(m, b) PM8821_SSBI_REG(m, b, 0x08) b 44 drivers/mfd/qcom-pm8xxx.c #define PM8821_SSBI_ADDR_IRQ_RT_STATUS(m, b) PM8821_SSBI_REG(m, b, 0x0f) b 584 drivers/mfd/twl4030-power.c struct twl4030_resconfig *b = board; b 586 drivers/mfd/twl4030-power.c while (b->resource) { b 587 drivers/mfd/twl4030-power.c if (b->resource == common->resource) { b 588 drivers/mfd/twl4030-power.c *common = *b; b 591 drivers/mfd/twl4030-power.c b++; b 24 drivers/mfd/wl1273-core.c u8 b[2]; b 27 drivers/mfd/wl1273-core.c r = i2c_smbus_read_i2c_block_data(client, reg, sizeof(b), b); b 33 drivers/mfd/wl1273-core.c *value = (u16)b[0] << 8 | b[1]; b 19 drivers/misc/altera-stapl/altera-jtag.c #define alt_jtag_io(a, b, c)\ b 20 drivers/misc/altera-stapl/altera-jtag.c astate->config->jtag_io(astate->config->dev, a, b, c); b 678 drivers/misc/altera-stapl/altera.c s32 b = stack[--stack_ptr]; b 688 drivers/misc/altera-stapl/altera.c ((a & long_tmp) == (b & long_tmp)) b 2009 drivers/misc/altera-stapl/altera.c s32 a, b; b 2039 drivers/misc/altera-stapl/altera.c b = 1 + long_count - mask_index; b 2040 drivers/misc/altera-stapl/altera.c a = (a < b) ? a : b; b 2042 drivers/misc/altera-stapl/altera.c b = 1 + mask_left - mask_right; b 2043 drivers/misc/altera-stapl/altera.c a = (a < b) ? a : b; b 2064 drivers/misc/altera-stapl/altera.c b = source2[index2 >> 3] & b 2068 drivers/misc/altera-stapl/altera.c if (a != b) /* failure */ b 361 drivers/misc/fastrpc.c static int olaps_cmp(const void *a, const void *b) b 364 drivers/misc/fastrpc.c struct fastrpc_buf_overlap *pb = (struct fastrpc_buf_overlap *)b; b 66 drivers/misc/ibmasm/event.c static inline int event_available(struct event_buffer *b, struct event_reader *r) b 68 drivers/misc/ibmasm/event.c return (r->next_serial_number < b->next_serial_number); b 87 drivers/misc/sgi-gru/grukservices.c #define ASYNC_BID_TO_HAN(b) ((b) + 1) b 746 drivers/misc/sgi-gru/grumain.c #define next_gru(b, g) (((g) < &(b)->bs_grus[GRU_CHIPLETS_PER_BLADE - 1]) ? \ b 747 drivers/misc/sgi-gru/grumain.c ((g)+1) : &(b)->bs_grus[0]) b 486 drivers/misc/sgi-gru/grutables.h #define get_gru(b, c) (&gru_base[b]->bs_grus[c]) b 148 drivers/misc/sram.c struct list_head *b) b 151 drivers/misc/sram.c struct sram_reserve *rb = list_entry(b, struct sram_reserve, list); b 421 drivers/misc/vmw_balloon.c static inline void vmballoon_stats_op_inc(struct vmballoon *b, unsigned int op, b 425 drivers/misc/vmw_balloon.c atomic64_inc(&b->stats->ops[op][type]); b 428 drivers/misc/vmw_balloon.c static inline void vmballoon_stats_gen_inc(struct vmballoon *b, b 432 drivers/misc/vmw_balloon.c atomic64_inc(&b->stats->general_stat[stat]); b 435 drivers/misc/vmw_balloon.c static inline void vmballoon_stats_gen_add(struct vmballoon *b, b 440 drivers/misc/vmw_balloon.c atomic64_add(val, &b->stats->general_stat[stat]); b 443 drivers/misc/vmw_balloon.c static inline void vmballoon_stats_page_inc(struct vmballoon *b, b 448 drivers/misc/vmw_balloon.c atomic64_inc(&b->stats->page_stat[stat][size]); b 451 drivers/misc/vmw_balloon.c static inline void vmballoon_stats_page_add(struct vmballoon *b, b 457 drivers/misc/vmw_balloon.c atomic64_add(val, &b->stats->page_stat[stat][size]); b 461 drivers/misc/vmw_balloon.c __vmballoon_cmd(struct vmballoon *b, unsigned long cmd, unsigned long arg1, b 466 drivers/misc/vmw_balloon.c vmballoon_stats_op_inc(b, cmd, VMW_BALLOON_OP_STAT); b 489 drivers/misc/vmw_balloon.c WRITE_ONCE(b->target, local_result); b 493 drivers/misc/vmw_balloon.c vmballoon_stats_op_inc(b, cmd, VMW_BALLOON_OP_FAIL_STAT); b 501 drivers/misc/vmw_balloon.c b->reset_required = true; b 507 drivers/misc/vmw_balloon.c vmballoon_cmd(struct vmballoon *b, unsigned long cmd, unsigned long arg1, b 512 drivers/misc/vmw_balloon.c return __vmballoon_cmd(b, cmd, arg1, arg2, &dummy); b 519 drivers/misc/vmw_balloon.c static int vmballoon_send_start(struct vmballoon *b, unsigned long req_caps) b 523 drivers/misc/vmw_balloon.c status = __vmballoon_cmd(b, VMW_BALLOON_CMD_START, req_caps, 0, b 528 drivers/misc/vmw_balloon.c b->capabilities = capabilities; b 531 drivers/misc/vmw_balloon.c b->capabilities = VMW_BALLOON_BASIC_CMDS; b 542 drivers/misc/vmw_balloon.c b->max_page_size = VMW_BALLOON_4K_PAGE; b 543 drivers/misc/vmw_balloon.c if ((b->capabilities & VMW_BALLOON_BATCHED_2M_CMDS) && b 544 drivers/misc/vmw_balloon.c (b->capabilities & VMW_BALLOON_BATCHED_CMDS)) b 545 drivers/misc/vmw_balloon.c b->max_page_size = VMW_BALLOON_2M_PAGE; b 563 drivers/misc/vmw_balloon.c static int vmballoon_send_guest_id(struct vmballoon *b) b 567 drivers/misc/vmw_balloon.c status = vmballoon_cmd(b, VMW_BALLOON_CMD_GUEST_ID, b 636 drivers/misc/vmw_balloon.c static int vmballoon_send_get_target(struct vmballoon *b) b 644 drivers/misc/vmw_balloon.c if (!(b->capabilities & VMW_BALLOON_64_BIT_TARGET) && b 648 drivers/misc/vmw_balloon.c status = vmballoon_cmd(b, VMW_BALLOON_CMD_GET_TARGET, limit, 0); b 665 drivers/misc/vmw_balloon.c static int vmballoon_alloc_page_list(struct vmballoon *b, b 689 drivers/misc/vmw_balloon.c vmballoon_stats_page_inc(b, VMW_BALLOON_PAGE_STAT_ALLOC, b 700 drivers/misc/vmw_balloon.c vmballoon_stats_page_inc(b, VMW_BALLOON_PAGE_STAT_ALLOC_FAIL, b 718 drivers/misc/vmw_balloon.c static int vmballoon_handle_one_result(struct vmballoon *b, struct page *page, b 731 drivers/misc/vmw_balloon.c vmballoon_stats_page_inc(b, VMW_BALLOON_PAGE_STAT_REFUSED_ALLOC, b 750 drivers/misc/vmw_balloon.c static unsigned long vmballoon_status_page(struct vmballoon *b, int idx, b 755 drivers/misc/vmw_balloon.c *p = pfn_to_page(b->batch_page[idx].pfn); b 756 drivers/misc/vmw_balloon.c return b->batch_page[idx].status; b 760 drivers/misc/vmw_balloon.c *p = b->page; b 786 drivers/misc/vmw_balloon.c static unsigned long vmballoon_lock_op(struct vmballoon *b, b 793 drivers/misc/vmw_balloon.c lockdep_assert_held(&b->comm_lock); b 805 drivers/misc/vmw_balloon.c pfn = PHYS_PFN(virt_to_phys(b->batch_page)); b 809 drivers/misc/vmw_balloon.c pfn = page_to_pfn(b->page); b 816 drivers/misc/vmw_balloon.c return vmballoon_cmd(b, cmd, pfn, num_pages); b 828 drivers/misc/vmw_balloon.c static void vmballoon_add_page(struct vmballoon *b, unsigned int idx, b 831 drivers/misc/vmw_balloon.c lockdep_assert_held(&b->comm_lock); b 834 drivers/misc/vmw_balloon.c b->batch_page[idx] = (struct vmballoon_batch_entry) b 837 drivers/misc/vmw_balloon.c b->page = p; b 859 drivers/misc/vmw_balloon.c static int vmballoon_lock(struct vmballoon *b, struct vmballoon_ctl *ctl) b 870 drivers/misc/vmw_balloon.c spin_lock(&b->comm_lock); b 874 drivers/misc/vmw_balloon.c vmballoon_add_page(b, i++, page); b 876 drivers/misc/vmw_balloon.c batch_status = vmballoon_lock_op(b, ctl->n_pages, ctl->page_size, b 887 drivers/misc/vmw_balloon.c status = vmballoon_status_page(b, i, &page); b 897 drivers/misc/vmw_balloon.c if (!vmballoon_handle_one_result(b, page, ctl->page_size, b 910 drivers/misc/vmw_balloon.c spin_unlock(&b->comm_lock); b 944 drivers/misc/vmw_balloon.c static void vmballoon_release_refused_pages(struct vmballoon *b, b 947 drivers/misc/vmw_balloon.c vmballoon_stats_page_inc(b, VMW_BALLOON_PAGE_STAT_REFUSED_FREE, b 962 drivers/misc/vmw_balloon.c static int64_t vmballoon_change(struct vmballoon *b) b 966 drivers/misc/vmw_balloon.c size = atomic64_read(&b->size); b 967 drivers/misc/vmw_balloon.c target = READ_ONCE(b->target); b 974 drivers/misc/vmw_balloon.c if (b->reset_required) b 983 drivers/misc/vmw_balloon.c if (target > size && time_before(jiffies, READ_ONCE(b->shrink_timeout))) b 1000 drivers/misc/vmw_balloon.c static void vmballoon_enqueue_page_list(struct vmballoon *b, b 1009 drivers/misc/vmw_balloon.c balloon_page_list_enqueue(&b->b_dev_info, pages); b 1015 drivers/misc/vmw_balloon.c spin_lock_irqsave(&b->b_dev_info.pages_lock, flags); b 1021 drivers/misc/vmw_balloon.c list_splice_init(pages, &b->huge_pages); b 1024 drivers/misc/vmw_balloon.c spin_unlock_irqrestore(&b->b_dev_info.pages_lock, flags); b 1043 drivers/misc/vmw_balloon.c static void vmballoon_dequeue_page_list(struct vmballoon *b, b 1055 drivers/misc/vmw_balloon.c *n_pages = balloon_page_list_dequeue(&b->b_dev_info, pages, b 1061 drivers/misc/vmw_balloon.c spin_lock_irqsave(&b->b_dev_info.pages_lock, flags); b 1062 drivers/misc/vmw_balloon.c list_for_each_entry_safe(page, tmp, &b->huge_pages, lru) { b 1072 drivers/misc/vmw_balloon.c spin_unlock_irqrestore(&b->b_dev_info.pages_lock, flags); b 1107 drivers/misc/vmw_balloon.c static void vmballoon_inflate(struct vmballoon *b) b 1114 drivers/misc/vmw_balloon.c .page_size = b->max_page_size, b 1118 drivers/misc/vmw_balloon.c while ((to_inflate_frames = vmballoon_change(b)) > 0) { b 1127 drivers/misc/vmw_balloon.c to_inflate_pages = min_t(unsigned long, b->batch_max_pages, b 1132 drivers/misc/vmw_balloon.c alloc_error = vmballoon_alloc_page_list(b, &ctl, b 1136 drivers/misc/vmw_balloon.c lock_error = vmballoon_lock(b, &ctl); b 1146 drivers/misc/vmw_balloon.c atomic64_add(ctl.n_pages * page_in_frames, &b->size); b 1148 drivers/misc/vmw_balloon.c vmballoon_enqueue_page_list(b, &ctl.pages, &ctl.n_pages, b 1177 drivers/misc/vmw_balloon.c vmballoon_release_refused_pages(b, &ctl); b 1194 drivers/misc/vmw_balloon.c static unsigned long vmballoon_deflate(struct vmballoon *b, uint64_t n_frames, b 1226 drivers/misc/vmw_balloon.c -vmballoon_change(b); b 1236 drivers/misc/vmw_balloon.c to_deflate_pages = min_t(unsigned long, b->batch_max_pages, b 1241 drivers/misc/vmw_balloon.c vmballoon_dequeue_page_list(b, &ctl.pages, &ctl.n_pages, b 1258 drivers/misc/vmw_balloon.c vmballoon_lock(b, &ctl); b 1270 drivers/misc/vmw_balloon.c atomic64_sub(n_unlocked_frames, &b->size); b 1273 drivers/misc/vmw_balloon.c vmballoon_stats_page_add(b, VMW_BALLOON_PAGE_STAT_FREE, b 1281 drivers/misc/vmw_balloon.c vmballoon_enqueue_page_list(b, &ctl.refused_pages, b 1287 drivers/misc/vmw_balloon.c if (ctl.page_size == b->max_page_size) b 1306 drivers/misc/vmw_balloon.c static void vmballoon_deinit_batching(struct vmballoon *b) b 1308 drivers/misc/vmw_balloon.c free_page((unsigned long)b->batch_page); b 1309 drivers/misc/vmw_balloon.c b->batch_page = NULL; b 1311 drivers/misc/vmw_balloon.c b->batch_max_pages = 1; b 1324 drivers/misc/vmw_balloon.c static int vmballoon_init_batching(struct vmballoon *b) b 1332 drivers/misc/vmw_balloon.c b->batch_page = page_address(page); b 1333 drivers/misc/vmw_balloon.c b->batch_max_pages = PAGE_SIZE / sizeof(struct vmballoon_batch_entry); b 1345 drivers/misc/vmw_balloon.c struct vmballoon *b = client_data; b 1347 drivers/misc/vmw_balloon.c vmballoon_stats_gen_inc(b, VMW_BALLOON_STAT_DOORBELL); b 1349 drivers/misc/vmw_balloon.c mod_delayed_work(system_freezable_wq, &b->dwork, 0); b 1355 drivers/misc/vmw_balloon.c static void vmballoon_vmci_cleanup(struct vmballoon *b) b 1357 drivers/misc/vmw_balloon.c vmballoon_cmd(b, VMW_BALLOON_CMD_VMCI_DOORBELL_SET, b 1360 drivers/misc/vmw_balloon.c if (!vmci_handle_is_invalid(b->vmci_doorbell)) { b 1361 drivers/misc/vmw_balloon.c vmci_doorbell_destroy(b->vmci_doorbell); b 1362 drivers/misc/vmw_balloon.c b->vmci_doorbell = VMCI_INVALID_HANDLE; b 1376 drivers/misc/vmw_balloon.c static int vmballoon_vmci_init(struct vmballoon *b) b 1380 drivers/misc/vmw_balloon.c if ((b->capabilities & VMW_BALLOON_SIGNALLED_WAKEUP_CMD) == 0) b 1383 drivers/misc/vmw_balloon.c error = vmci_doorbell_create(&b->vmci_doorbell, VMCI_FLAG_DELAYED_CB, b 1385 drivers/misc/vmw_balloon.c vmballoon_doorbell, b); b 1390 drivers/misc/vmw_balloon.c error = __vmballoon_cmd(b, VMW_BALLOON_CMD_VMCI_DOORBELL_SET, b 1391 drivers/misc/vmw_balloon.c b->vmci_doorbell.context, b 1392 drivers/misc/vmw_balloon.c b->vmci_doorbell.resource, NULL); b 1399 drivers/misc/vmw_balloon.c vmballoon_vmci_cleanup(b); b 1412 drivers/misc/vmw_balloon.c static void vmballoon_pop(struct vmballoon *b) b 1416 drivers/misc/vmw_balloon.c while ((size = atomic64_read(&b->size))) b 1417 drivers/misc/vmw_balloon.c vmballoon_deflate(b, size, false); b 1425 drivers/misc/vmw_balloon.c static void vmballoon_reset(struct vmballoon *b) b 1429 drivers/misc/vmw_balloon.c down_write(&b->conf_sem); b 1431 drivers/misc/vmw_balloon.c vmballoon_vmci_cleanup(b); b 1434 drivers/misc/vmw_balloon.c vmballoon_pop(b); b 1436 drivers/misc/vmw_balloon.c if (vmballoon_send_start(b, VMW_BALLOON_CAPABILITIES)) b 1439 drivers/misc/vmw_balloon.c if ((b->capabilities & VMW_BALLOON_BATCHED_CMDS) != 0) { b 1440 drivers/misc/vmw_balloon.c if (vmballoon_init_batching(b)) { b 1447 drivers/misc/vmw_balloon.c vmballoon_send_start(b, 0); b 1450 drivers/misc/vmw_balloon.c } else if ((b->capabilities & VMW_BALLOON_BASIC_CMDS) != 0) { b 1451 drivers/misc/vmw_balloon.c vmballoon_deinit_batching(b); b 1454 drivers/misc/vmw_balloon.c vmballoon_stats_gen_inc(b, VMW_BALLOON_STAT_RESET); b 1455 drivers/misc/vmw_balloon.c b->reset_required = false; b 1457 drivers/misc/vmw_balloon.c error = vmballoon_vmci_init(b); b 1461 drivers/misc/vmw_balloon.c if (vmballoon_send_guest_id(b)) b 1465 drivers/misc/vmw_balloon.c up_write(&b->conf_sem); b 1479 drivers/misc/vmw_balloon.c struct vmballoon *b = container_of(dwork, struct vmballoon, dwork); b 1482 drivers/misc/vmw_balloon.c if (b->reset_required) b 1483 drivers/misc/vmw_balloon.c vmballoon_reset(b); b 1485 drivers/misc/vmw_balloon.c down_read(&b->conf_sem); b 1492 drivers/misc/vmw_balloon.c vmballoon_stats_gen_inc(b, VMW_BALLOON_STAT_TIMER); b 1494 drivers/misc/vmw_balloon.c if (!vmballoon_send_get_target(b)) b 1495 drivers/misc/vmw_balloon.c change = vmballoon_change(b); b 1499 drivers/misc/vmw_balloon.c atomic64_read(&b->size), READ_ONCE(b->target)); b 1502 drivers/misc/vmw_balloon.c vmballoon_inflate(b); b 1504 drivers/misc/vmw_balloon.c vmballoon_deflate(b, 0, true); b 1507 drivers/misc/vmw_balloon.c up_read(&b->conf_sem); b 1528 drivers/misc/vmw_balloon.c struct vmballoon *b = &balloon; b 1531 drivers/misc/vmw_balloon.c pr_debug("%s - size: %llu", __func__, atomic64_read(&b->size)); b 1533 drivers/misc/vmw_balloon.c vmballoon_stats_gen_inc(b, VMW_BALLOON_STAT_SHRINK); b 1539 drivers/misc/vmw_balloon.c if (!down_read_trylock(&b->conf_sem)) b 1542 drivers/misc/vmw_balloon.c deflated_frames = vmballoon_deflate(b, sc->nr_to_scan, true); b 1544 drivers/misc/vmw_balloon.c vmballoon_stats_gen_add(b, VMW_BALLOON_STAT_SHRINK_FREE, b 1552 drivers/misc/vmw_balloon.c WRITE_ONCE(b->shrink_timeout, jiffies + HZ * VMBALLOON_SHRINK_DELAY); b 1554 drivers/misc/vmw_balloon.c up_read(&b->conf_sem); b 1570 drivers/misc/vmw_balloon.c struct vmballoon *b = &balloon; b 1572 drivers/misc/vmw_balloon.c return atomic64_read(&b->size); b 1575 drivers/misc/vmw_balloon.c static void vmballoon_unregister_shrinker(struct vmballoon *b) b 1577 drivers/misc/vmw_balloon.c if (b->shrinker_registered) b 1578 drivers/misc/vmw_balloon.c unregister_shrinker(&b->shrinker); b 1579 drivers/misc/vmw_balloon.c b->shrinker_registered = false; b 1582 drivers/misc/vmw_balloon.c static int vmballoon_register_shrinker(struct vmballoon *b) b 1590 drivers/misc/vmw_balloon.c b->shrinker.scan_objects = vmballoon_shrinker_scan; b 1591 drivers/misc/vmw_balloon.c b->shrinker.count_objects = vmballoon_shrinker_count; b 1592 drivers/misc/vmw_balloon.c b->shrinker.seeks = DEFAULT_SEEKS; b 1594 drivers/misc/vmw_balloon.c r = register_shrinker(&b->shrinker); b 1597 drivers/misc/vmw_balloon.c b->shrinker_registered = true; b 1623 drivers/misc/vmw_balloon.c static int vmballoon_enable_stats(struct vmballoon *b) b 1627 drivers/misc/vmw_balloon.c down_write(&b->conf_sem); b 1630 drivers/misc/vmw_balloon.c if (b->stats) b 1633 drivers/misc/vmw_balloon.c b->stats = kzalloc(sizeof(*b->stats), GFP_KERNEL); b 1635 drivers/misc/vmw_balloon.c if (!b->stats) { b 1642 drivers/misc/vmw_balloon.c up_write(&b->conf_sem); b 1659 drivers/misc/vmw_balloon.c struct vmballoon *b = f->private; b 1663 drivers/misc/vmw_balloon.c if (!b->stats) { b 1664 drivers/misc/vmw_balloon.c int r = vmballoon_enable_stats(b); b 1673 drivers/misc/vmw_balloon.c seq_printf(f, "%-22s: %#16lx\n", "used capabilities", b->capabilities); b 1675 drivers/misc/vmw_balloon.c b->reset_required ? "y" : "n"); b 1678 drivers/misc/vmw_balloon.c seq_printf(f, "%-22s: %16lu\n", "target", READ_ONCE(b->target)); b 1679 drivers/misc/vmw_balloon.c seq_printf(f, "%-22s: %16llu\n", "current", atomic64_read(&b->size)); b 1687 drivers/misc/vmw_balloon.c atomic64_read(&b->stats->ops[i][VMW_BALLOON_OP_STAT]), b 1688 drivers/misc/vmw_balloon.c atomic64_read(&b->stats->ops[i][VMW_BALLOON_OP_FAIL_STAT])); b 1694 drivers/misc/vmw_balloon.c atomic64_read(&b->stats->general_stat[i])); b 1701 drivers/misc/vmw_balloon.c atomic64_read(&b->stats->page_stat[i][j])); b 1709 drivers/misc/vmw_balloon.c static void __init vmballoon_debugfs_init(struct vmballoon *b) b 1711 drivers/misc/vmw_balloon.c b->dbg_entry = debugfs_create_file("vmmemctl", S_IRUGO, NULL, b, b 1715 drivers/misc/vmw_balloon.c static void __exit vmballoon_debugfs_exit(struct vmballoon *b) b 1718 drivers/misc/vmw_balloon.c debugfs_remove(b->dbg_entry); b 1719 drivers/misc/vmw_balloon.c kfree(b->stats); b 1720 drivers/misc/vmw_balloon.c b->stats = NULL; b 1725 drivers/misc/vmw_balloon.c static inline void vmballoon_debugfs_init(struct vmballoon *b) b 1729 drivers/misc/vmw_balloon.c static inline void vmballoon_debugfs_exit(struct vmballoon *b) b 1770 drivers/misc/vmw_balloon.c struct vmballoon *b; b 1773 drivers/misc/vmw_balloon.c b = container_of(b_dev_info, struct vmballoon, b_dev_info); b 1779 drivers/misc/vmw_balloon.c if (!down_read_trylock(&b->conf_sem)) b 1782 drivers/misc/vmw_balloon.c spin_lock(&b->comm_lock); b 1790 drivers/misc/vmw_balloon.c vmballoon_add_page(b, 0, page); b 1791 drivers/misc/vmw_balloon.c status = vmballoon_lock_op(b, 1, VMW_BALLOON_4K_PAGE, b 1795 drivers/misc/vmw_balloon.c status = vmballoon_status_page(b, 0, &page); b 1802 drivers/misc/vmw_balloon.c spin_unlock(&b->comm_lock); b 1817 drivers/misc/vmw_balloon.c vmballoon_add_page(b, 0, newpage); b 1818 drivers/misc/vmw_balloon.c status = vmballoon_lock_op(b, 1, VMW_BALLOON_4K_PAGE, b 1822 drivers/misc/vmw_balloon.c status = vmballoon_status_page(b, 0, &newpage); b 1824 drivers/misc/vmw_balloon.c spin_unlock(&b->comm_lock); b 1833 drivers/misc/vmw_balloon.c atomic64_dec(&b->size); b 1845 drivers/misc/vmw_balloon.c spin_lock_irqsave(&b->b_dev_info.pages_lock, flags); b 1853 drivers/misc/vmw_balloon.c balloon_page_insert(&b->b_dev_info, newpage); b 1861 drivers/misc/vmw_balloon.c b->b_dev_info.isolated_pages--; b 1862 drivers/misc/vmw_balloon.c spin_unlock_irqrestore(&b->b_dev_info.pages_lock, flags); b 1865 drivers/misc/vmw_balloon.c up_read(&b->conf_sem); b 1874 drivers/misc/vmw_balloon.c static void vmballoon_compaction_deinit(struct vmballoon *b) b 1876 drivers/misc/vmw_balloon.c if (!IS_ERR(b->b_dev_info.inode)) b 1877 drivers/misc/vmw_balloon.c iput(b->b_dev_info.inode); b 1879 drivers/misc/vmw_balloon.c b->b_dev_info.inode = NULL; b 1895 drivers/misc/vmw_balloon.c static __init int vmballoon_compaction_init(struct vmballoon *b) b 1901 drivers/misc/vmw_balloon.c b->b_dev_info.migratepage = vmballoon_migratepage; b 1902 drivers/misc/vmw_balloon.c b->b_dev_info.inode = alloc_anon_inode(vmballoon_mnt->mnt_sb); b 1904 drivers/misc/vmw_balloon.c if (IS_ERR(b->b_dev_info.inode)) b 1905 drivers/misc/vmw_balloon.c return PTR_ERR(b->b_dev_info.inode); b 1907 drivers/misc/vmw_balloon.c b->b_dev_info.inode->i_mapping->a_ops = &balloon_aops; b 1913 drivers/misc/vmw_balloon.c static void vmballoon_compaction_deinit(struct vmballoon *b) b 1917 drivers/misc/vmw_balloon.c static int vmballoon_compaction_init(struct vmballoon *b) b 136 drivers/mmc/core/mmc.c unsigned int e, m, a, b; b 177 drivers/mmc/core/mmc.c b = UNSTUFF_BITS(resp, 37, 5); b 178 drivers/mmc/core/mmc.c csd->erase_size = (a + 1) * (b + 1); b 412 drivers/mmc/core/sdio_io.c void sdio_writeb(struct sdio_func *func, u8 b, unsigned int addr, int *err_ret) b 422 drivers/mmc/core/sdio_io.c ret = mmc_io_rw_direct(func->card, 1, func->num, addr, b, NULL); b 561 drivers/mmc/core/sdio_io.c void sdio_writew(struct sdio_func *func, u16 b, unsigned int addr, int *err_ret) b 565 drivers/mmc/core/sdio_io.c *(__le16 *)func->tmpbuf = cpu_to_le16(b); b 609 drivers/mmc/core/sdio_io.c void sdio_writel(struct sdio_func *func, u32 b, unsigned int addr, int *err_ret) b 613 drivers/mmc/core/sdio_io.c *(__le32 *)func->tmpbuf = cpu_to_le32(b); b 667 drivers/mmc/core/sdio_io.c void sdio_f0_writeb(struct sdio_func *func, unsigned char b, unsigned int addr, b 684 drivers/mmc/core/sdio_io.c ret = mmc_io_rw_direct(func->card, 1, 0, addr, b, NULL); b 497 drivers/mmc/host/mxcmmc.c u32 a, b, c; b 514 drivers/mmc/host/mxcmmc.c b = mxcmci_readw(host, MMC_REG_RES_FIFO); b 515 drivers/mmc/host/mxcmmc.c cmd->resp[i] = a << 16 | b; b 519 drivers/mmc/host/mxcmmc.c b = mxcmci_readw(host, MMC_REG_RES_FIFO); b 521 drivers/mmc/host/mxcmmc.c cmd->resp[0] = a << 24 | b << 8 | c >> 8; b 561 drivers/mmc/host/mxcmmc.c u8 *b = (u8 *)buf; b 569 drivers/mmc/host/mxcmmc.c memcpy(b, &tmp, bytes); b 589 drivers/mmc/host/mxcmmc.c u8 *b = (u8 *)buf; b 596 drivers/mmc/host/mxcmmc.c memcpy(&tmp, b, bytes); b 197 drivers/mmc/host/sdhci-pci-gli.c static void gli_set_9750_rx_inv(struct sdhci_host *host, bool b) b 205 drivers/mmc/host/sdhci-pci-gli.c if (b) { b 2353 drivers/mmc/host/sdhci.c u32 b = host->sdma_boundary; b 2369 drivers/mmc/host/sdhci.c sdhci_writew(host, SDHCI_MAKE_BLKSZ(b, 128), SDHCI_BLOCK_SIZE); b 2371 drivers/mmc/host/sdhci.c sdhci_writew(host, SDHCI_MAKE_BLKSZ(b, 64), SDHCI_BLOCK_SIZE); b 1753 drivers/mtd/devices/st_spi_fsm.c uint8_t *b = (uint8_t *)buf; b 1767 drivers/mtd/devices/st_spi_fsm.c ret = stfsm_write(fsm, b, bytes, to); b 1771 drivers/mtd/devices/st_spi_fsm.c b += bytes; b 277 drivers/mtd/inftlcore.c status = oob.b.Status | oob.b.Status1; b 348 drivers/mtd/inftlcore.c oob.b.Status = oob.b.Status1 = SECTOR_USED; b 599 drivers/mtd/inftlcore.c oob.u.b.virtualUnitNo = cpu_to_le16(thisVUC); b 600 drivers/mtd/inftlcore.c oob.u.b.prevUnitNo = cpu_to_le16(prev_block); b 601 drivers/mtd/inftlcore.c oob.u.b.ANAC = anac; b 602 drivers/mtd/inftlcore.c oob.u.b.NACs = nacs; b 603 drivers/mtd/inftlcore.c oob.u.b.parityPerField = parity; b 604 drivers/mtd/inftlcore.c oob.u.b.discarded = 0xaa; b 836 drivers/mtd/inftlcore.c oob.b.Status = oob.b.Status1 = SECTOR_USED; b 146 drivers/mtd/maps/nettel.c unsigned long b; b 149 drivers/mtd/maps/nettel.c for (b = 0; (b < nettel_intel_partitions[3].size); b += 0x100000) { b 150 drivers/mtd/maps/nettel.c cfi_send_gen_cmd(0xff, 0x55, b, &nettel_intel_map, cfi, b 598 drivers/mtd/nand/raw/atmel/pmecc.c s16 a, b, c; b 604 drivers/mtd/nand/raw/atmel/pmecc.c b = index_of[dmu[ro]]; b 606 drivers/mtd/nand/raw/atmel/pmecc.c tmp = a + (cw_len - b) + c; b 628 drivers/mtd/nand/raw/atmel/pmecc.c s16 a, b, c; b 631 drivers/mtd/nand/raw/atmel/pmecc.c b = si[2 * (i - 1) + 3 - k]; b 632 drivers/mtd/nand/raw/atmel/pmecc.c c = index_of[b]; b 555 drivers/mtd/nand/raw/cafe_nand.c static unsigned short gf64_mul(u8 a, u8 b) b 563 drivers/mtd/nand/raw/cafe_nand.c c ^= b; b 565 drivers/mtd/nand/raw/cafe_nand.c b <<= 1; b 566 drivers/mtd/nand/raw/cafe_nand.c if ((b & 0x40) != 0) b 567 drivers/mtd/nand/raw/cafe_nand.c b ^= 0x43; b 574 drivers/mtd/nand/raw/cafe_nand.c static u16 gf4096_mul(u16 a, u16 b) b 580 drivers/mtd/nand/raw/cafe_nand.c bh = b >> 6; b 581 drivers/mtd/nand/raw/cafe_nand.c bl = b & 0x3f; b 188 drivers/mtd/nand/raw/meson_nand.c #define MESON_ECC_DATA(b, s) { .bch = (b), .strength = (s)} b 624 drivers/mtd/nand/raw/mtk_nand.c static void mtk_nfc_no_bad_mark_swap(struct mtd_info *a, u8 *b, int c) b 2132 drivers/mtd/nand/raw/nand_base.c const struct nand_op_parser_ctx *b) b 2134 drivers/mtd/nand/raw/nand_base.c if (a->subop.ninstrs < b->subop.ninstrs) b 2136 drivers/mtd/nand/raw/nand_base.c else if (a->subop.ninstrs > b->subop.ninstrs) b 2139 drivers/mtd/nand/raw/nand_base.c if (a->subop.last_instr_end_off < b->subop.last_instr_end_off) b 2141 drivers/mtd/nand/raw/nand_base.c else if (a->subop.last_instr_end_off > b->subop.last_instr_end_off) b 565 drivers/mtd/nand/raw/stm32_fmc2_nand.c u32 byte_addr = 0, b; b 572 drivers/mtd/nand/raw/stm32_fmc2_nand.c b = b0 | (b1 << 8) | (b2 << 16); b 575 drivers/mtd/nand/raw/stm32_fmc2_nand.c if (likely(!b)) b 580 drivers/mtd/nand/raw/stm32_fmc2_nand.c switch (b % 4) { b 589 drivers/mtd/nand/raw/stm32_fmc2_nand.c b >>= 2; b 595 drivers/mtd/nand/raw/stm32_fmc2_nand.c switch (b % 4) { b 604 drivers/mtd/nand/raw/stm32_fmc2_nand.c b >>= 2; b 152 drivers/mtd/nand/raw/sunxi_nand.c #define NFC_ECC_ERR_CNT(b, x) (((x) >> (((b) % 4) * 8)) & 0xff) b 281 drivers/mtd/nftlcore.c status = oob.b.Status | oob.b.Status1; b 421 drivers/mtd/nftlcore.c oob.b.Status = oob.b.Status1 = SECTOR_USED; b 693 drivers/mtd/nftlcore.c oob.b.Status = oob.b.Status1 = SECTOR_USED; b 25 drivers/mtd/ubi/debug.h #define ubi_dbg_print_hex_dump(l, ps, pt, r, g, b, len, a) \ b 26 drivers/mtd/ubi/debug.h print_hex_dump(l, ps, pt, r, g, b, len, a) b 362 drivers/net/can/grcan.c static inline u32 grcan_ring_add(u32 a, u32 b, u32 size) b 364 drivers/net/can/grcan.c u32 sum = a + b; b 373 drivers/net/can/grcan.c static inline u32 grcan_ring_sub(u32 a, u32 b, u32 size) b 375 drivers/net/can/grcan.c return grcan_ring_add(a, size - b, size); b 428 drivers/net/can/rcar/rcar_canfd.c #define RCANFD_F_RMDF(q, b) (0x200c + (0x04 * (b)) + (0x20 * (q))) b 456 drivers/net/can/rcar/rcar_canfd.c #define RCANFD_F_TMDF(p, b) (0x400c + (0x20 * (p)) + (0x04 * (b))) b 21 drivers/net/can/rx-offload.c static inline bool can_rx_offload_le(struct can_rx_offload *offload, unsigned int a, unsigned int b) b 24 drivers/net/can/rx-offload.c return a <= b; b 26 drivers/net/can/rx-offload.c return a >= b; b 69 drivers/net/can/rx-offload.c int (*compare)(struct sk_buff *a, struct sk_buff *b)) b 97 drivers/net/can/rx-offload.c static int can_rx_offload_compare(struct sk_buff *a, struct sk_buff *b) b 102 drivers/net/can/rx-offload.c cb_b = can_rx_offload_get_cb(b); b 550 drivers/net/ethernet/8390/ne2k-pci.c __le16 *b = (__le16 *)buf; b 552 drivers/net/ethernet/8390/ne2k-pci.c *b++ = cpu_to_le16(inw(NE_BASE + NE_DATAPORT)); b 553 drivers/net/ethernet/8390/ne2k-pci.c buf = (char *)b; b 615 drivers/net/ethernet/8390/ne2k-pci.c __le16 *b = (__le16 *)buf; b 617 drivers/net/ethernet/8390/ne2k-pci.c outw(le16_to_cpu(*b++), NE_BASE + NE_DATAPORT); b 618 drivers/net/ethernet/8390/ne2k-pci.c buf = (char *)b; b 262 drivers/net/ethernet/apple/bmac.c int b; b 265 drivers/net/ethernet/apple/bmac.c b = (val & (1 << nb))? 6: 4; b 266 drivers/net/ethernet/apple/bmac.c bmwrite(dev, MIFCSR, b); b 268 drivers/net/ethernet/apple/bmac.c bmwrite(dev, MIFCSR, b|1); b 747 drivers/net/ethernet/broadcom/bnx2x/bnx2x.h #define SUB_S16(a, b) (s16)((s16)(a) - (s16)(b)) b 748 drivers/net/ethernet/broadcom/bnx2x/bnx2x.h #define SUB_S32(a, b) (s32)((s32)(a) - (s32)(b)) b 107 drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.h #define MAXVAL(a, b) (((a) > (b)) ? (a) : (b)) b 4038 drivers/net/ethernet/cavium/liquidio/lio_main.c u8 *b = (u8 *)&data[1]; b 4043 drivers/net/ethernet/cavium/liquidio/lio_main.c vf_idx, b + 2); b 2844 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c static void init_cong_ctrl(unsigned short *a, unsigned short *b) b 2871 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c b[0] = b[1] = b[2] = b[3] = b[4] = b[5] = b[6] = b[7] = b[8] = 0; b 2872 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c b[9] = b[10] = 1; b 2873 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c b[11] = b[12] = 2; b 2874 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c b[13] = b[14] = b[15] = b[16] = 3; b 2875 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c b[17] = b[18] = b[19] = b[20] = b[21] = 4; b 2876 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c b[22] = b[23] = b[24] = b[25] = b[26] = b[27] = 5; b 2877 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c b[28] = b[29] = 6; b 2878 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c b[30] = b[31] = 7; b 123 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c static int cudbg_mem_desc_cmp(const void *a, const void *b) b 126 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c ((const struct cudbg_mem_desc *)b)->base; b 1537 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h u32 b = ((u32)addr[3] << 16) | ((u32)addr[4] << 8) | addr[5]; b 1539 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h a ^= b; b 2764 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c u16 a, b; b 2780 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c b = (hw_addr[1] << 8) | hw_addr[2]; b 2781 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c a ^= b; b 5811 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c static void init_cong_ctrl(unsigned short *a, unsigned short *b) b 5838 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c b[0] = b[1] = b[2] = b[3] = b[4] = b[5] = b[6] = b[7] = b[8] = 0; b 5839 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c b[9] = b[10] = 1; b 5840 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c b[11] = b[12] = 2; b 5841 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c b[13] = b[14] = b[15] = b[16] = 3; b 5842 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c b[17] = b[18] = b[19] = b[20] = b[21] = 4; b 5843 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c b[22] = b[23] = b[24] = b[25] = b[26] = b[27] = 5; b 5844 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c b[28] = b[29] = 6; b 5845 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c b[30] = b[31] = 7; b 357 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_common.h u32 b = ((u32)addr[3] << 16) | ((u32)addr[4] << 8) | addr[5]; b 359 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_common.h a ^= b; b 2226 drivers/net/ethernet/cisco/enic/enic_main.c rss_key_buf_va->key[kidx].b[bidx] = enic->rss_key[i]; b 2260 drivers/net/ethernet/cisco/enic/enic_main.c (*rss_cpu_buf_va).cpu[i/4].b[i%4] = i % enic->rq_count; b 30 drivers/net/ethernet/cisco/enic/vnic_rss.h u8 b[ENIC_RSS_BYTES_PER_KEY]; b 39 drivers/net/ethernet/cisco/enic/vnic_rss.h u8 b[4] ; b 3995 drivers/net/ethernet/dec/tulip/de4x5.c u32 b; b 4004 drivers/net/ethernet/dec/tulip/de4x5.c dev.llsig.b = ETH_PROM_SIG; b 1735 drivers/net/ethernet/hisilicon/hns/hns_dsaf_main.c const u16 *b = (const u16 *)src; b 1737 drivers/net/ethernet/hisilicon/hns/hns_dsaf_main.c a[0] &= b[0]; b 1738 drivers/net/ethernet/hisilicon/hns/hns_dsaf_main.c a[1] &= b[1]; b 1739 drivers/net/ethernet/hisilicon/hns/hns_dsaf_main.c a[2] &= b[2]; b 135 drivers/net/ethernet/i825xx/lasi_82596.c u16 a, b; b 139 drivers/net/ethernet/i825xx/lasi_82596.c b = v & 0xffff; b 142 drivers/net/ethernet/i825xx/lasi_82596.c b = v >> 16; b 148 drivers/net/ethernet/i825xx/lasi_82596.c gsc_writel(b, dev->base_addr + PA_CPU_PORT_L_ACCESS); b 426 drivers/net/ethernet/intel/e100.c #define X(a,b) b,a b 428 drivers/net/ethernet/intel/e100.c #define X(a,b) a,b b 3404 drivers/net/ethernet/intel/e1000/e1000_main.c struct my_u { __le64 a; __le64 b; }; b 3418 drivers/net/ethernet/intel/e1000/e1000_main.c ((le64_to_cpu(u->b) & (1<<20)) ? 'd' : 'c'), i, b 3419 drivers/net/ethernet/intel/e1000/e1000_main.c le64_to_cpu(u->a), le64_to_cpu(u->b), b 3446 drivers/net/ethernet/intel/e1000/e1000_main.c struct my_u { __le64 a; __le64 b; }; b 3458 drivers/net/ethernet/intel/e1000/e1000_main.c i, le64_to_cpu(u->a), le64_to_cpu(u->b), b 205 drivers/net/ethernet/intel/e1000e/netdev.c __le64 b; b 213 drivers/net/ethernet/intel/e1000e/netdev.c __le64 b; b 303 drivers/net/ethernet/intel/e1000e/netdev.c (!(le64_to_cpu(u0->b) & BIT(29)) ? 'l' : b 304 drivers/net/ethernet/intel/e1000e/netdev.c ((le64_to_cpu(u0->b) & BIT(20)) ? 'd' : 'c')), b 307 drivers/net/ethernet/intel/e1000e/netdev.c (unsigned long long)le64_to_cpu(u0->b), b 380 drivers/net/ethernet/intel/e1000e/netdev.c (unsigned long long)le64_to_cpu(u1->b), b 388 drivers/net/ethernet/intel/e1000e/netdev.c (unsigned long long)le64_to_cpu(u1->b), b 446 drivers/net/ethernet/intel/e1000e/netdev.c (unsigned long long)le64_to_cpu(u1->b), b 452 drivers/net/ethernet/intel/e1000e/netdev.c (unsigned long long)le64_to_cpu(u1->b), b 17 drivers/net/ethernet/intel/i40e/i40e_ddp.c struct i40e_profile_info *b) b 19 drivers/net/ethernet/intel/i40e/i40e_ddp.c return a->track_id == b->track_id && b 20 drivers/net/ethernet/intel/i40e/i40e_ddp.c !memcmp(&a->version, &b->version, sizeof(a->version)) && b 21 drivers/net/ethernet/intel/i40e/i40e_ddp.c !memcmp(&a->name, &b->name, I40E_DDP_NAME_SIZE); b 4361 drivers/net/ethernet/intel/i40e/i40e_ethtool.c struct i40e_fdir_filter *b) b 4364 drivers/net/ethernet/intel/i40e/i40e_ethtool.c if (a->dst_ip != b->dst_ip || b 4365 drivers/net/ethernet/intel/i40e/i40e_ethtool.c a->src_ip != b->src_ip || b 4366 drivers/net/ethernet/intel/i40e/i40e_ethtool.c a->dst_port != b->dst_port || b 4367 drivers/net/ethernet/intel/i40e/i40e_ethtool.c a->src_port != b->src_port || b 4368 drivers/net/ethernet/intel/i40e/i40e_ethtool.c a->flow_type != b->flow_type || b 4369 drivers/net/ethernet/intel/i40e/i40e_ethtool.c a->ip4_proto != b->ip4_proto) b 365 drivers/net/ethernet/intel/igb/igb_main.c struct my_u0 { u64 a; u64 b; } *u0; b 450 drivers/net/ethernet/intel/igb/igb_main.c le64_to_cpu(u0->b), b 531 drivers/net/ethernet/intel/igb/igb_main.c le64_to_cpu(u0->b), b 537 drivers/net/ethernet/intel/igb/igb_main.c le64_to_cpu(u0->b), b 386 drivers/net/ethernet/intel/igbvf/ethtool.c char *b = (char *)adapter + b 389 drivers/net/ethernet/intel/igbvf/ethtool.c sizeof(u64)) ? (*(u64 *)p - *(u64 *)b) : b 390 drivers/net/ethernet/intel/igbvf/ethtool.c (*(u32 *)p - *(u32 *)b)); b 1519 drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c bool b = false; b 1523 drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c b = reg_pattern_test(adapter, data, b 1529 drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c b = reg_set_and_check(adapter, data, b 1540 drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c b = reg_pattern_test(adapter, data, b 1546 drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c b = reg_pattern_test(adapter, data, b 1552 drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c b = reg_pattern_test(adapter, data, b 1558 drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c if (b) b 575 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c struct my_u0 { u64 a; u64 b; } *u0; b 691 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c le64_to_cpu((__force __le64)u0->b), b 803 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c le64_to_cpu((__force __le64)u0->b), b 810 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c le64_to_cpu((__force __le64)u0->b), b 677 drivers/net/ethernet/intel/ixgbevf/ethtool.c bool b = false; b 681 drivers/net/ethernet/intel/ixgbevf/ethtool.c b = reg_pattern_test(adapter, data, b 687 drivers/net/ethernet/intel/ixgbevf/ethtool.c b = reg_set_and_check(adapter, data, b 698 drivers/net/ethernet/intel/ixgbevf/ethtool.c b = reg_pattern_test(adapter, data, b 704 drivers/net/ethernet/intel/ixgbevf/ethtool.c b = reg_pattern_test(adapter, data, b 710 drivers/net/ethernet/intel/ixgbevf/ethtool.c b = reg_pattern_test(adapter, data, b 716 drivers/net/ethernet/intel/ixgbevf/ethtool.c if (b) b 179 drivers/net/ethernet/marvell/octeontx2/af/common.h #define NIX_LINK_CGX_LMAC(a, b) (0 + 4 * (a) + (b)) b 181 drivers/net/ethernet/marvell/octeontx2/af/common.h #define NIX_CHAN_CGX_LMAC_CHX(a, b, c) (0x800 + 0x100 * (a) + 0x10 * (b) + (c)) b 182 drivers/net/ethernet/marvell/octeontx2/af/common.h #define NIX_CHAN_LBK_CHX(a, b) (0 + 0x100 * (a) + (b)) b 29 drivers/net/ethernet/marvell/octeontx2/af/rvu_reg.h #define RVU_AF_AFPFX_MBOXX(a, b) (0x2000 | (a) << 4 | (b) << 3) b 75 drivers/net/ethernet/marvell/octeontx2/af/rvu_reg.h #define RVU_PF_VFX_PFVF_MBOXX(a, b) (0x0 | (a) << 12 | (b) << 3) b 359 drivers/net/ethernet/marvell/octeontx2/af/rvu_reg.h #define NIX_AF_TL3_TL2X_LINKX_CFG(a, b) (0x1700 | (a) << 16 | (b) << 3) b 360 drivers/net/ethernet/marvell/octeontx2/af/rvu_reg.h #define NIX_AF_RX_FLOW_KEY_ALGX_FIELDX(a, b) (0x1800 | (a) << 18 | (b) << 3) b 367 drivers/net/ethernet/marvell/octeontx2/af/rvu_reg.h #define NIX_AF_LSO_FORMATX_FIELDX(a, b) (0x1B00 | (a) << 16 | (b) << 3) b 391 drivers/net/ethernet/marvell/octeontx2/af/rvu_reg.h #define NIX_AF_LFX_RX_VTAG_TYPEX(a, b) (0x4200 | (a) << 17 | (b) << 3) b 392 drivers/net/ethernet/marvell/octeontx2/af/rvu_reg.h #define NIX_AF_LFX_LOCKX(a, b) (0x4300 | (a) << 17 | (b) << 3) b 393 drivers/net/ethernet/marvell/octeontx2/af/rvu_reg.h #define NIX_AF_LFX_TX_STATX(a, b) (0x4400 | (a) << 17 | (b) << 3) b 394 drivers/net/ethernet/marvell/octeontx2/af/rvu_reg.h #define NIX_AF_LFX_RX_STATX(a, b) (0x4500 | (a) << 17 | (b) << 3) b 395 drivers/net/ethernet/marvell/octeontx2/af/rvu_reg.h #define NIX_AF_LFX_RSS_GRPX(a, b) (0x4600 | (a) << 17 | (b) << 3) b 459 drivers/net/ethernet/marvell/octeontx2/af/rvu_reg.h #define NPC_AF_PKINDX_CPI_DEFX(a, b) (0x80020ull | (a) << 6 | (b) << 3) b 460 drivers/net/ethernet/marvell/octeontx2/af/rvu_reg.h #define NPC_AF_KPUX_ENTRYX_CAMX(a, b, c) \ b 461 drivers/net/ethernet/marvell/octeontx2/af/rvu_reg.h (0x100000 | (a) << 14 | (b) << 6 | (c) << 3) b 462 drivers/net/ethernet/marvell/octeontx2/af/rvu_reg.h #define NPC_AF_KPUX_ENTRYX_ACTION0(a, b) \ b 463 drivers/net/ethernet/marvell/octeontx2/af/rvu_reg.h (0x100020 | (a) << 14 | (b) << 6) b 464 drivers/net/ethernet/marvell/octeontx2/af/rvu_reg.h #define NPC_AF_KPUX_ENTRYX_ACTION1(a, b) \ b 465 drivers/net/ethernet/marvell/octeontx2/af/rvu_reg.h (0x100028 | (a) << 14 | (b) << 6) b 466 drivers/net/ethernet/marvell/octeontx2/af/rvu_reg.h #define NPC_AF_KPUX_ENTRY_DISX(a, b) (0x180000 | (a) << 6 | (b) << 3) b 468 drivers/net/ethernet/marvell/octeontx2/af/rvu_reg.h #define NPC_AF_INTFX_LIDX_LTX_LDX_CFG(a, b, c, d) \ b 469 drivers/net/ethernet/marvell/octeontx2/af/rvu_reg.h (0x900000 | (a) << 16 | (b) << 12 | (c) << 5 | (d) << 3) b 470 drivers/net/ethernet/marvell/octeontx2/af/rvu_reg.h #define NPC_AF_INTFX_LDATAX_FLAGSX_CFG(a, b, c) \ b 471 drivers/net/ethernet/marvell/octeontx2/af/rvu_reg.h (0x980000 | (a) << 16 | (b) << 12 | (c) << 3) b 472 drivers/net/ethernet/marvell/octeontx2/af/rvu_reg.h #define NPC_AF_MCAMEX_BANKX_CAMX_INTF(a, b, c) \ b 473 drivers/net/ethernet/marvell/octeontx2/af/rvu_reg.h (0x1000000ull | (a) << 10 | (b) << 6 | (c) << 3) b 474 drivers/net/ethernet/marvell/octeontx2/af/rvu_reg.h #define NPC_AF_MCAMEX_BANKX_CAMX_W0(a, b, c) \ b 475 drivers/net/ethernet/marvell/octeontx2/af/rvu_reg.h (0x1000010ull | (a) << 10 | (b) << 6 | (c) << 3) b 476 drivers/net/ethernet/marvell/octeontx2/af/rvu_reg.h #define NPC_AF_MCAMEX_BANKX_CAMX_W1(a, b, c) \ b 477 drivers/net/ethernet/marvell/octeontx2/af/rvu_reg.h (0x1000020ull | (a) << 10 | (b) << 6 | (c) << 3) b 478 drivers/net/ethernet/marvell/octeontx2/af/rvu_reg.h #define NPC_AF_MCAMEX_BANKX_CFG(a, b) (0x1800000ull | (a) << 8 | (b) << 4) b 479 drivers/net/ethernet/marvell/octeontx2/af/rvu_reg.h #define NPC_AF_MCAMEX_BANKX_STAT_ACT(a, b) \ b 480 drivers/net/ethernet/marvell/octeontx2/af/rvu_reg.h (0x1880000 | (a) << 8 | (b) << 4) b 483 drivers/net/ethernet/marvell/octeontx2/af/rvu_reg.h #define NPC_AF_MCAMEX_BANKX_ACTION(a, b) (0x1900000ull | (a) << 8 | (b) << 4) b 484 drivers/net/ethernet/marvell/octeontx2/af/rvu_reg.h #define NPC_AF_MCAMEX_BANKX_TAG_ACT(a, b) \ b 485 drivers/net/ethernet/marvell/octeontx2/af/rvu_reg.h (0x1900008 | (a) << 8 | (b) << 4) b 488 drivers/net/ethernet/marvell/octeontx2/af/rvu_reg.h #define NPC_AF_MCAM_BANKX_HITX(a, b) (0x1c80000 | (a) << 8 | (b) << 4) b 4182 drivers/net/ethernet/marvell/sky2.c static int sky2_reg_access_ok(struct sky2_hw *hw, unsigned int b) b 4188 drivers/net/ethernet/marvell/sky2.c switch (b) { b 4235 drivers/net/ethernet/marvell/sky2.c unsigned int b; b 4239 drivers/net/ethernet/marvell/sky2.c for (b = 0; b < 128; b++) { b 4241 drivers/net/ethernet/marvell/sky2.c if (b == 3) b 4243 drivers/net/ethernet/marvell/sky2.c else if (sky2_reg_access_ok(sky2->hw, b)) b 291 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c struct mod_hdr_key *b) b 293 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c if (a->num_actions != b->num_actions) b 296 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c return memcmp(a->actions, b->actions, a->num_actions * MLX5_MH_ACT_SZ); b 2928 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c struct encap_key *b) b 2930 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c return memcmp(a->ip_tun_key, b->ip_tun_key, sizeof(*a->ip_tun_key)) || b 2931 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c a->tc_tunnel->tunnel_type != b->tc_tunnel->tunnel_type; b 685 drivers/net/ethernet/mellanox/mlx5/core/port.c MLX5_SET(qetc_reg, in, tc_configuration[i].b, 1); b 48 drivers/net/ethernet/mellanox/mlxsw/item.h u8 *b = (u8 *) buf; b 51 drivers/net/ethernet/mellanox/mlxsw/item.h tmp = b[offset]; b 64 drivers/net/ethernet/mellanox/mlxsw/item.h u8 *b = (u8 *) buf; b 71 drivers/net/ethernet/mellanox/mlxsw/item.h tmp = b[offset]; b 74 drivers/net/ethernet/mellanox/mlxsw/item.h b[offset] = tmp; b 82 drivers/net/ethernet/mellanox/mlxsw/item.h __be16 *b = (__be16 *) buf; b 85 drivers/net/ethernet/mellanox/mlxsw/item.h tmp = be16_to_cpu(b[offset]); b 98 drivers/net/ethernet/mellanox/mlxsw/item.h __be16 *b = (__be16 *) buf; b 105 drivers/net/ethernet/mellanox/mlxsw/item.h tmp = be16_to_cpu(b[offset]); b 108 drivers/net/ethernet/mellanox/mlxsw/item.h b[offset] = cpu_to_be16(tmp); b 116 drivers/net/ethernet/mellanox/mlxsw/item.h __be32 *b = (__be32 *) buf; b 119 drivers/net/ethernet/mellanox/mlxsw/item.h tmp = be32_to_cpu(b[offset]); b 132 drivers/net/ethernet/mellanox/mlxsw/item.h __be32 *b = (__be32 *) buf; b 139 drivers/net/ethernet/mellanox/mlxsw/item.h tmp = be32_to_cpu(b[offset]); b 142 drivers/net/ethernet/mellanox/mlxsw/item.h b[offset] = cpu_to_be32(tmp); b 150 drivers/net/ethernet/mellanox/mlxsw/item.h __be64 *b = (__be64 *) buf; b 153 drivers/net/ethernet/mellanox/mlxsw/item.h tmp = be64_to_cpu(b[offset]); b 165 drivers/net/ethernet/mellanox/mlxsw/item.h __be64 *b = (__be64 *) buf; b 172 drivers/net/ethernet/mellanox/mlxsw/item.h tmp = be64_to_cpu(b[offset]); b 175 drivers/net/ethernet/mellanox/mlxsw/item.h b[offset] = cpu_to_be64(tmp); b 325 drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c DECLARE_BITMAP(b, MLXSW_SP_PREFIX_COUNT); b 329 drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c for_each_set_bit(prefix, (prefix_usage)->b, MLXSW_SP_PREFIX_COUNT) b 349 drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c set_bit(prefix_len, prefix_usage->b); b 356 drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c clear_bit(prefix_len, prefix_usage->b); b 197 drivers/net/ethernet/microchip/enc28j60.c u8 b = (addr & BANK_MASK) >> 5; b 206 drivers/net/ethernet/microchip/enc28j60.c if ((b & ECON1_BSEL0) != (priv->bank & ECON1_BSEL0)) { b 207 drivers/net/ethernet/microchip/enc28j60.c if (b & ECON1_BSEL0) b 214 drivers/net/ethernet/microchip/enc28j60.c if ((b & ECON1_BSEL1) != (priv->bank & ECON1_BSEL1)) { b 215 drivers/net/ethernet/microchip/enc28j60.c if (b & ECON1_BSEL1) b 222 drivers/net/ethernet/microchip/enc28j60.c priv->bank = b; b 2516 drivers/net/ethernet/natsemi/natsemi.c int b = (ether_crc(ETH_ALEN, ha->addr) >> 23) & 0x1ff; b 2517 drivers/net/ethernet/natsemi/natsemi.c mc_filter[b/8] |= (1 << (b & 0x07)); b 1859 drivers/net/ethernet/natsemi/ns83820.c unsigned a, b; b 1878 drivers/net/ethernet/natsemi/ns83820.c b = ns83820_mii_read_reg(dev, 1, 0x1d); b 1879 drivers/net/ethernet/natsemi/ns83820.c dprintk("version: 0x%04x 0x%04x\n", a, b); b 111 drivers/net/ethernet/neterion/vxge/vxge-traffic.h #define VXGE_HW_SET_LEVEL(a, b) (((a) > (b)) ? (a) : (b)) b 516 drivers/net/ethernet/netronome/nfp/nfpcore/nfp6000_pcie.c const struct nfp_bar *a = aptr, *b = bptr; b 518 drivers/net/ethernet/netronome/nfp/nfpcore/nfp6000_pcie.c if (a->bitsize == b->bitsize) b 519 drivers/net/ethernet/netronome/nfp/nfpcore/nfp6000_pcie.c return a->index - b->index; b 521 drivers/net/ethernet/netronome/nfp/nfpcore/nfp6000_pcie.c return a->bitsize - b->bitsize; b 3200 drivers/net/ethernet/nvidia/forcedeth.c u32 a, b; b 3203 drivers/net/ethernet/nvidia/forcedeth.c b = le16_to_cpu(*(__le16 *) (&hw_addr[4])); b 3206 drivers/net/ethernet/nvidia/forcedeth.c alwaysOn[1] &= b; b 3207 drivers/net/ethernet/nvidia/forcedeth.c alwaysOff[1] &= ~b; b 41 drivers/net/ethernet/qlogic/netxen/netxen_nic.h #define NETXEN_VERSION_CODE(a, b, c) (((a) << 24) + ((b) << 16) + (c)) b 70 drivers/net/ethernet/qlogic/netxen/netxen_nic.h #define find_diff_among(a,b,range) ((a)<(b)?((b)-(a)):((b)+(range)-(a))) b 6200 drivers/net/ethernet/qlogic/qed/qed_debug.c static u32 qed_cyclic_add(u32 a, u32 b, u32 size) b 6202 drivers/net/ethernet/qlogic/qed/qed_debug.c return (a + b) % size; b 6205 drivers/net/ethernet/qlogic/qed/qed_debug.c static u32 qed_cyclic_sub(u32 a, u32 b, u32 size) b 6207 drivers/net/ethernet/qlogic/qed/qed_debug.c return (size + a - b) % size; b 60 drivers/net/ethernet/qlogic/qede/qede_filter.c bool (*ip_comp)(struct qede_arfs_tuple *a, struct qede_arfs_tuple *b); b 1499 drivers/net/ethernet/qlogic/qede/qede_filter.c struct qede_arfs_tuple *b) b 1502 drivers/net/ethernet/qlogic/qede/qede_filter.c b->eth_proto != htons(ETH_P_IP)) b 1505 drivers/net/ethernet/qlogic/qede/qede_filter.c return (a->src_ipv4 == b->src_ipv4) && b 1506 drivers/net/ethernet/qlogic/qede/qede_filter.c (a->dst_ipv4 == b->dst_ipv4); b 1541 drivers/net/ethernet/qlogic/qede/qede_filter.c struct qede_arfs_tuple *b) b 1544 drivers/net/ethernet/qlogic/qede/qede_filter.c b->eth_proto != htons(ETH_P_IPV6)) b 1547 drivers/net/ethernet/qlogic/qede/qede_filter.c if (memcmp(&a->src_ipv6, &b->src_ipv6, sizeof(struct in6_addr))) b 1550 drivers/net/ethernet/qlogic/qede/qede_filter.c if (memcmp(&a->dst_ipv6, &b->dst_ipv6, sizeof(struct in6_addr))) b 46 drivers/net/ethernet/qlogic/qlcnic/qlcnic.h #define QLCNIC_VERSION_CODE(a, b, c) (((a) << 24) + ((b) << 16) + (c)) b 2378 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c #define IS_QLC_83XX_USED(a, b, c) (((1 << a->portnum) & b) || ((c >> 6) & 0x1)) b 538 drivers/net/ethernet/sfc/bitfield.h #define EFX_OWORD32(a, b, c, d) \ b 539 drivers/net/ethernet/sfc/bitfield.h { .u32 = { cpu_to_le32(a), cpu_to_le32(b), \ b 535 drivers/net/ethernet/sfc/falcon/bitfield.h #define EF4_OWORD32(a, b, c, d) \ b 536 drivers/net/ethernet/sfc/falcon/bitfield.h { .u32 = { cpu_to_le32(a), cpu_to_le32(b), \ b 91 drivers/net/ethernet/sfc/falcon/farch.c static bool ef4_masked_compare_oword(const ef4_oword_t *a, const ef4_oword_t *b, b 94 drivers/net/ethernet/sfc/falcon/farch.c return ((a->u64[0] ^ b->u64[0]) & mask->u64[0]) || b 95 drivers/net/ethernet/sfc/falcon/farch.c ((a->u64[1] ^ b->u64[1]) & mask->u64[1]); b 93 drivers/net/ethernet/sfc/farch.c static bool efx_masked_compare_oword(const efx_oword_t *a, const efx_oword_t *b, b 96 drivers/net/ethernet/sfc/farch.c return ((a->u64[0] ^ b->u64[0]) & mask->u64[0]) || b 97 drivers/net/ethernet/sfc/farch.c ((a->u64[1] ^ b->u64[1]) & mask->u64[1]); b 263 drivers/net/ethernet/sgi/ioc3-eth.c int a, b, index, disc; b 273 drivers/net/ethernet/sgi/ioc3-eth.c b = nic_read_bit(mcr); b 275 drivers/net/ethernet/sgi/ioc3-eth.c if (a && b) { b 281 drivers/net/ethernet/sgi/ioc3-eth.c if (!a && !b) { b 123 drivers/net/ethernet/smsc/smc911x.c u_char a, b; b 125 drivers/net/ethernet/smsc/smc911x.c b = *buf++; b 126 drivers/net/ethernet/smsc/smc911x.c pr_cont("%02x%02x ", a, b); b 132 drivers/net/ethernet/smsc/smc911x.c u_char a, b; b 134 drivers/net/ethernet/smsc/smc911x.c b = *buf++; b 135 drivers/net/ethernet/smsc/smc911x.c pr_cont("%02x%02x ", a, b); b 172 drivers/net/ethernet/smsc/smc91x.c u_char a, b; b 174 drivers/net/ethernet/smsc/smc91x.c b = *buf++; b 175 drivers/net/ethernet/smsc/smc91x.c pr_cont("%02x%02x ", a, b); b 181 drivers/net/ethernet/smsc/smc91x.c u_char a, b; b 183 drivers/net/ethernet/smsc/smc91x.c b = *buf++; b 184 drivers/net/ethernet/smsc/smc91x.c pr_cont("%02x%02x ", a, b); b 440 drivers/net/ethernet/ti/tlan.h #define CIRC_INC(a, b) if (++a >= b) a = 0 b 689 drivers/net/ethernet/via/via-rhine.c unsigned char b = readb(ioaddr+reg); b 691 drivers/net/ethernet/via/via-rhine.c if (a != b) { b 694 drivers/net/ethernet/via/via-rhine.c reg, a, b); b 25 drivers/net/fddi/skfp/queue.c #define PRINTF(a,b,c) b 1249 drivers/net/geneve.c struct ip_tunnel_info *b) b 1252 drivers/net/geneve.c return a->key.u.ipv4.dst == b->key.u.ipv4.dst; b 1254 drivers/net/geneve.c return ipv6_addr_equal(&a->key.u.ipv6.dst, &b->key.u.ipv6.dst); b 322 drivers/net/hamradio/baycom_epp.c #define PKP(a,b) printk(KERN_INFO "baycomm_epp: channel params: " a "\n", b) b 324 drivers/net/hamradio/baycom_epp.c #define PKP(a,b) b 216 drivers/net/hamradio/hdlcdrv.c #define PKP(a,b) printk(KERN_INFO "hdlcdrv.c: channel params: " a "\n", b) b 218 drivers/net/hamradio/hdlcdrv.c #define PKP(a,b) b 651 drivers/net/hamradio/yam.c unsigned char b, temp; b 685 drivers/net/hamradio/yam.c b = yp->tx_buf[yp->tx_count++]; b 686 drivers/net/hamradio/yam.c outb(b, THR(dev->base_addr)); b 689 drivers/net/hamradio/yam.c yp->tx_crch = chktabh[temp] ^ b; b 69 drivers/net/phy/phy-core.c #define PHY_SETTING(s, d, b) { .speed = SPEED_ ## s, .duplex = DUPLEX_ ## d, \ b 70 drivers/net/phy/phy-core.c .bit = ETHTOOL_LINK_MODE_ ## b ## _BIT} b 1093 drivers/net/phy/phylink.c static void phylink_merge_link_mode(unsigned long *dst, const unsigned long *b) b 1101 drivers/net/phy/phylink.c linkmode_or(dst, dst, b); b 194 drivers/net/plip/plip.c } b; b 602 drivers/net/plip/plip.c &rcv->nibble, &rcv->length.b.lsb)) { b 614 drivers/net/plip/plip.c &rcv->nibble, &rcv->length.b.lsb)) b 622 drivers/net/plip/plip.c &rcv->nibble, &rcv->length.b.msb)) b 814 drivers/net/plip/plip.c &snd->nibble, snd->length.b.lsb)) b 821 drivers/net/plip/plip.c &snd->nibble, snd->length.b.msb)) b 203 drivers/net/ppp/bsd_comp.c #define MAXCODE(b) ((1 << (b)) - 1) b 250 drivers/net/ppp/ppp_generic.c #define seq_before(a, b) ((s32)((a) - (b)) < 0) b 251 drivers/net/ppp/ppp_generic.c #define seq_after(a, b) ((s32)((a) - (b)) > 0) b 127 drivers/net/ppp/pppoe.c static inline int cmp_2_addr(struct pppoe_addr *a, struct pppoe_addr *b) b 129 drivers/net/ppp/pppoe.c return a->sid == b->sid && ether_addr_equal(a->remote, b->remote); b 96 drivers/net/vxlan.c bool vxlan_addr_equal(const union vxlan_addr *a, const union vxlan_addr *b) b 98 drivers/net/vxlan.c if (a->sa.sa_family != b->sa.sa_family) b 101 drivers/net/vxlan.c return ipv6_addr_equal(&a->sin6.sin6_addr, &b->sin6.sin6_addr); b 103 drivers/net/vxlan.c return a->sin.sin_addr.s_addr == b->sin.sin_addr.s_addr; b 133 drivers/net/vxlan.c bool vxlan_addr_equal(const union vxlan_addr *a, const union vxlan_addr *b) b 135 drivers/net/vxlan.c return a->sin.sin_addr.s_addr == b->sin.sin_addr.s_addr; b 1362 drivers/net/wan/cosa.c unsigned long b = (unsigned long)buf; b 1363 drivers/net/wan/cosa.c if (b+len >= MAX_DMA_ADDRESS) b 1365 drivers/net/wan/cosa.c if ((b^ (b+len)) & 0x10000) { b 1185 drivers/net/wan/ixp4xx_hss.c static u32 check_clock(u32 rate, u32 a, u32 b, u32 c, b 1193 drivers/net/wan/ixp4xx_hss.c do_div(new_rate, a * (c + 1) + b + 1); b 1199 drivers/net/wan/ixp4xx_hss.c *reg = (a << 22) | (b << 12) | c; b 1206 drivers/net/wan/ixp4xx_hss.c u32 a, b, diff = 0xFFFFFFFF; b 1224 drivers/net/wan/ixp4xx_hss.c for (b = 0; b < 0x400; b++) { b 1225 drivers/net/wan/ixp4xx_hss.c u64 c = (b + 1) * (u64)rate; b 1229 drivers/net/wan/ixp4xx_hss.c if (b == 0 && /* also try a bit higher rate */ b 1232 drivers/net/wan/ixp4xx_hss.c check_clock(rate, a, b, 0xFFF, best, &diff, reg); b 1235 drivers/net/wan/ixp4xx_hss.c if (!check_clock(rate, a, b, c, best, &diff, reg)) b 1237 drivers/net/wan/ixp4xx_hss.c if (!check_clock(rate, a, b, c + 1, best, &diff, reg)) b 207 drivers/net/wan/sealevel.c struct slvl_board *b; b 218 drivers/net/wan/sealevel.c b = kzalloc(sizeof(struct slvl_board), GFP_KERNEL); b 219 drivers/net/wan/sealevel.c if (!b) b 222 drivers/net/wan/sealevel.c b->dev[0].chan = &b->board.chanA; b 223 drivers/net/wan/sealevel.c b->dev[0].channel = 0; b 225 drivers/net/wan/sealevel.c b->dev[1].chan = &b->board.chanB; b 226 drivers/net/wan/sealevel.c b->dev[1].channel = 1; b 228 drivers/net/wan/sealevel.c dev = &b->board; b 236 drivers/net/wan/sealevel.c b->iobase = iobase; b 257 drivers/net/wan/sealevel.c outb(3 | (1 << 7), b->iobase + 4); b 270 drivers/net/wan/sealevel.c dev->chanA.private = &b->dev[0]; b 271 drivers/net/wan/sealevel.c dev->chanB.private = &b->dev[1]; b 308 drivers/net/wan/sealevel.c if (slvl_setup(&b->dev[0], iobase, irq)) b 310 drivers/net/wan/sealevel.c if (slvl_setup(&b->dev[1], iobase, irq)) b 315 drivers/net/wan/sealevel.c return b; b 318 drivers/net/wan/sealevel.c unregister_hdlc_device(b->dev[0].chan->netdevice); b 319 drivers/net/wan/sealevel.c free_netdev(b->dev[0].chan->netdevice); b 327 drivers/net/wan/sealevel.c kfree(b); b 333 drivers/net/wan/sealevel.c static void __exit slvl_shutdown(struct slvl_board *b) b 337 drivers/net/wan/sealevel.c z8530_shutdown(&b->board); b 340 drivers/net/wan/sealevel.c struct net_device *d = b->dev[u].chan->netdevice; b 345 drivers/net/wan/sealevel.c free_irq(b->board.irq, &b->board); b 346 drivers/net/wan/sealevel.c free_dma(b->board.chanA.rxdma); b 347 drivers/net/wan/sealevel.c free_dma(b->board.chanA.txdma); b 349 drivers/net/wan/sealevel.c outb(0, b->iobase); b 350 drivers/net/wan/sealevel.c release_region(b->iobase, 8); b 351 drivers/net/wan/sealevel.c kfree(b); b 344 drivers/net/wireless/ath/ath5k/base.c ath5k_setup_rate_idx(struct ath5k_hw *ah, struct ieee80211_supported_band *b) b 349 drivers/net/wireless/ath/ath5k/base.c ah->rate_idx[b->band][i] = -1; b 351 drivers/net/wireless/ath/ath5k/base.c for (i = 0; i < b->n_bitrates; i++) { b 352 drivers/net/wireless/ath/ath5k/base.c ah->rate_idx[b->band][b->bitrates[i].hw_value] = i; b 353 drivers/net/wireless/ath/ath5k/base.c if (b->bitrates[i].hw_value_short) b 354 drivers/net/wireless/ath/ath5k/base.c ah->rate_idx[b->band][b->bitrates[i].hw_value_short] = i; b 1026 drivers/net/wireless/ath/ath5k/debug.c unsigned int b, i; b 1031 drivers/net/wireless/ath/ath5k/debug.c for (b = 0; b < NUM_NL80211_BANDS; b++) { b 1032 drivers/net/wireless/ath/ath5k/debug.c struct ieee80211_supported_band *band = &ah->sbands[b]; b 740 drivers/net/wireless/ath/ath5k/pcu.c ath5k_check_timer_win(int a, int b, int window, int intval) b 748 drivers/net/wireless/ath/ath5k/pcu.c if ((b - a == window) || /* 1.) */ b 749 drivers/net/wireless/ath/ath5k/pcu.c (a - b == intval - window) || /* 2.) */ b 750 drivers/net/wireless/ath/ath5k/pcu.c ((a | 0x10000) - b == intval - window) || /* 3.) */ b 751 drivers/net/wireless/ath/ath5k/pcu.c ((b | 0x10000) - a == window)) /* 4.) */ b 1650 drivers/net/wireless/ath/carl9170/main.c int err, b, i; b 1664 drivers/net/wireless/ath/carl9170/main.c for (b = 0; b < NUM_NL80211_BANDS; b++) { b 1665 drivers/net/wireless/ath/carl9170/main.c band = ar->hw->wiphy->bands[b]; b 595 drivers/net/wireless/ath/carl9170/rx.c #define TID_CHECK(a, b) ( \ b 597 drivers/net/wireless/ath/carl9170/rx.c ((b) & cpu_to_le16(IEEE80211_BAR_CTRL_TID_INFO_MASK))) \ b 1834 drivers/net/wireless/ath/wil6210/cfg80211.c static void wil_print_bcon_data(struct cfg80211_beacon_data *b) b 1837 drivers/net/wireless/ath/wil6210/cfg80211.c b->head, b->head_len, true); b 1839 drivers/net/wireless/ath/wil6210/cfg80211.c b->tail, b->tail_len, true); b 1841 drivers/net/wireless/ath/wil6210/cfg80211.c b->beacon_ies, b->beacon_ies_len, true); b 1843 drivers/net/wireless/ath/wil6210/cfg80211.c b->probe_resp, b->probe_resp_len, true); b 1845 drivers/net/wireless/ath/wil6210/cfg80211.c b->proberesp_ies, b->proberesp_ies_len, true); b 1847 drivers/net/wireless/ath/wil6210/cfg80211.c b->assocresp_ies, b->assocresp_ies_len, true); b 4304 drivers/net/wireless/atmel/atmel.c b RESET_HANDLER b 4306 drivers/net/wireless/atmel/atmel.c b HALT1 b 4308 drivers/net/wireless/atmel/atmel.c b HALT1 b 4310 drivers/net/wireless/atmel/atmel.c b HALT1 b 4313 drivers/net/wireless/atmel/atmel.c b HALT1 b 4315 drivers/net/wireless/atmel/atmel.c b HALT1 b 4317 drivers/net/wireless/atmel/atmel.c b HALT1 b 4318 drivers/net/wireless/atmel/atmel.c HALT1: b HALT1 b 4355 drivers/net/wireless/atmel/atmel.c HALT2: b HALT2 b 4386 drivers/net/wireless/atmel/atmel.c b DELAYLOOP b 954 drivers/net/wireless/broadcom/b43/dma.c #define divide(a, b) ({ \ b 956 drivers/net/wireless/broadcom/b43/dma.c do_div(__a, b); \ b 960 drivers/net/wireless/broadcom/b43/dma.c #define modulo(a, b) ({ \ b 962 drivers/net/wireless/broadcom/b43/dma.c do_div(__a, b); \ b 513 drivers/net/wireless/broadcom/b43/phy_common.c unsigned int a, b, c, d; b 519 drivers/net/wireless/broadcom/b43/phy_common.c b = (tmp >> 8) & 0xFF; b 523 drivers/net/wireless/broadcom/b43/phy_common.c b == 0 || b == B43_TSSI_MAX || b 534 drivers/net/wireless/broadcom/b43/phy_common.c b = (b + 32) & 0x3F; b 540 drivers/net/wireless/broadcom/b43/phy_common.c average = (a + b + c + d + 2) / 4; b 702 drivers/net/wireless/broadcom/b43/phy_g.c s32 a, b; b 721 drivers/net/wireless/broadcom/b43/phy_g.c b = 0xA; b 724 drivers/net/wireless/broadcom/b43/phy_g.c b = 0x12; b 727 drivers/net/wireless/broadcom/b43/phy_g.c b = 0x11; b 739 drivers/net/wireless/broadcom/b43/phy_g.c b = b * (gphy->nrssi[1] - gphy->nrssi[0]); b 740 drivers/net/wireless/broadcom/b43/phy_g.c b += (gphy->nrssi[0] << 6); b 741 drivers/net/wireless/broadcom/b43/phy_g.c if (b < 32) b 742 drivers/net/wireless/broadcom/b43/phy_g.c b += 31; b 744 drivers/net/wireless/broadcom/b43/phy_g.c b += 32; b 745 drivers/net/wireless/broadcom/b43/phy_g.c b = b >> 6; b 746 drivers/net/wireless/broadcom/b43/phy_g.c b = clamp_val(b, -31, 31); b 749 drivers/net/wireless/broadcom/b43/phy_g.c tmp_u16 |= ((u32) b & 0x0000003F); b 84 drivers/net/wireless/broadcom/b43/phy_g.h const struct b43_rfatt *b) b 86 drivers/net/wireless/broadcom/b43/phy_g.h return ((a->att == b->att) && b 87 drivers/net/wireless/broadcom/b43/phy_g.h (a->with_padmix == b->with_padmix)); b 105 drivers/net/wireless/broadcom/b43/phy_g.h const struct b43_bbatt *b) b 107 drivers/net/wireless/broadcom/b43/phy_g.h return (a->att == b->att); b 1360 drivers/net/wireless/broadcom/b43/phy_lp.c static void lpphy_set_tx_iqcc(struct b43_wldev *dev, u16 a, u16 b) b 1365 drivers/net/wireless/broadcom/b43/phy_lp.c tmp[1] = b; b 4488 drivers/net/wireless/broadcom/b43/phy_n.c u16 tmp, a, b; b 4540 drivers/net/wireless/broadcom/b43/phy_n.c b = (qq << (31 - qq_nbits)); b 4543 drivers/net/wireless/broadcom/b43/phy_n.c b = (qq << (31 - qq_nbits)); b 4550 drivers/net/wireless/broadcom/b43/phy_n.c b = int_sqrt(b / tmp - a * a) - (1 << 10); b 4555 drivers/net/wireless/broadcom/b43/phy_n.c new.b0 = b & 0x3FF; b 4557 drivers/net/wireless/broadcom/b43/phy_n.c new.a0 = b & 0x3FF; b 4563 drivers/net/wireless/broadcom/b43/phy_n.c new.b1 = b & 0x3FF; b 4565 drivers/net/wireless/broadcom/b43/phy_n.c new.a1 = b & 0x3FF; b 806 drivers/net/wireless/broadcom/b43legacy/radio.c s32 b; b 861 drivers/net/wireless/broadcom/b43legacy/radio.c b = 0xA; b 865 drivers/net/wireless/broadcom/b43legacy/radio.c b = 0x12; b 868 drivers/net/wireless/broadcom/b43legacy/radio.c b = 0x11; b 880 drivers/net/wireless/broadcom/b43legacy/radio.c b = b * (phy->nrssi[1] - phy->nrssi[0]); b 881 drivers/net/wireless/broadcom/b43legacy/radio.c b += (phy->nrssi[0] << 6); b 882 drivers/net/wireless/broadcom/b43legacy/radio.c if (b < 32) b 883 drivers/net/wireless/broadcom/b43legacy/radio.c b += 31; b 885 drivers/net/wireless/broadcom/b43legacy/radio.c b += 32; b 886 drivers/net/wireless/broadcom/b43legacy/radio.c b = b >> 6; b 887 drivers/net/wireless/broadcom/b43legacy/radio.c b = clamp_val(b, -31, 31); b 890 drivers/net/wireless/broadcom/b43legacy/radio.c tmp_u16 |= ((u32)b & 0x0000003F); b 266 drivers/net/wireless/broadcom/brcm80211/brcmfmac/core.c const u16 *a, *b; b 279 drivers/net/wireless/broadcom/brcm80211/brcmfmac/core.c b = (const u16 *)iapp_l2_update_packet; b 281 drivers/net/wireless/broadcom/brcm80211/brcmfmac/core.c return !((a[0] ^ b[0]) | (a[1] ^ b[1]) | (a[2] ^ b[2])); b 321 drivers/net/wireless/broadcom/brcm80211/brcmsmac/aiutils.c #define GOODCOREADDR(x, b) \ b 322 drivers/net/wireless/broadcom/brcm80211/brcmsmac/aiutils.c (((x) >= (b)) && ((x) < ((b) + SI_MAXCORES * SI_CORE_SIZE)) && \ b 1008 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_int.h void wlc_lcnphy_set_tx_iqcc(struct brcms_phy *pi, u16 a, u16 b); b 1010 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_int.h void wlc_lcnphy_get_tx_iqcc(struct brcms_phy *pi, u16 *a, u16 *b); b 159 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c s16 b; b 1168 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c static void wlc_lcnphy_set_rx_iq_comp(struct brcms_phy *pi, u16 a, u16 b) b 1172 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c mod_phy_reg(pi, 0x646, (0x3ff << 0), (b) << 0); b 1176 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c mod_phy_reg(pi, 0x648, (0x3ff << 0), (b) << 0); b 1180 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c mod_phy_reg(pi, 0x64a, (0x3ff << 0), (b) << 0); b 1235 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c s32 a, b, temp; b 1280 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c b = (qq << (31 - qq_nbits)); b 1285 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c b = (qq << (31 - qq_nbits)); b 1290 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c b /= temp; b 1291 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c b -= a * a; b 1292 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c b = (s32) int_sqrt((unsigned long) b); b 1293 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c b -= (1 << 10); b 1295 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c b0_new = (u16) (b & 0x3ff); b 1388 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c iqcomp[iqcomp_sz].b); b 2270 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c s32 a, b, p; b 2273 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c b = (1024 * b0) + (64 * b1 * tssi); b 2274 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c p = ((2 * b) + a) / (2 * a); b 3161 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c void wlc_lcnphy_set_tx_iqcc(struct brcms_phy *pi, u16 a, u16 b) b 3167 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c iqcc[1] = b; b 3192 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c u16 a, b; b 3231 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c b = (u16) (bbmultiqcomp & 0x3ff); b 3232 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c wlc_lcnphy_set_tx_iqcc(pi, a, b); b 3534 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c u16 a, b, didq; b 3541 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c wlc_lcnphy_get_tx_iqcc(pi, &a, &b); b 3543 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c cc.im = b; b 3843 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c void wlc_lcnphy_get_tx_iqcc(struct brcms_phy *pi, u16 *a, u16 *b) b 3856 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c *b = iqcc[1]; b 3901 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c u16 a, b, didq, save_pa_gain = 0; b 3965 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c wlc_lcnphy_get_tx_iqcc(pi, &a, &b); b 3981 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c ((u32) (a & 0x3FF) << 10) | (b & 0x3ff); b 3990 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c pi_lcn->lcnphy_cal_results.txiqlocal_b = b; b 26123 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c s32 a, b, temp; b 26182 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c b = (qq << (31 - qq_nbits)); b 26189 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c b = (qq << (31 - qq_nbits)); b 26196 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c b /= temp; b 26197 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c b -= a * a; b 26198 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c b = (s32) int_sqrt((unsigned long) b); b 26199 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c b -= (1 << 10); b 26204 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c new_comp.b0 = (s16) b & 0x3ff; b 26207 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c new_comp.a0 = (s16) b & 0x3ff; b 26214 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c new_comp.b1 = (s16) b & 0x3ff; b 26217 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c new_comp.a1 = (s16) b & 0x3ff; b 196 drivers/net/wireless/broadcom/brcm80211/include/brcmu_utils.h #define brcmu_prpkt(a, b) b 292 drivers/net/wireless/intel/ipw2x00/ipw2200.c #define ipw_read_reg32(a, b) _ipw_read_reg32(a, b) b 296 drivers/net/wireless/intel/ipw2x00/ipw2200.c #define ipw_read_reg8(a, b) _ipw_read_reg8(a, b) b 300 drivers/net/wireless/intel/ipw2x00/ipw2200.c static inline void ipw_write_reg8(struct ipw_priv *a, u32 b, u8 c) b 303 drivers/net/wireless/intel/ipw2x00/ipw2200.c __LINE__, (u32) (b), (u32) (c)); b 304 drivers/net/wireless/intel/ipw2x00/ipw2200.c _ipw_write_reg8(a, b, c); b 309 drivers/net/wireless/intel/ipw2x00/ipw2200.c static inline void ipw_write_reg16(struct ipw_priv *a, u32 b, u16 c) b 312 drivers/net/wireless/intel/ipw2x00/ipw2200.c __LINE__, (u32) (b), (u32) (c)); b 313 drivers/net/wireless/intel/ipw2x00/ipw2200.c _ipw_write_reg16(a, b, c); b 318 drivers/net/wireless/intel/ipw2x00/ipw2200.c static inline void ipw_write_reg32(struct ipw_priv *a, u32 b, u32 c) b 321 drivers/net/wireless/intel/ipw2x00/ipw2200.c __LINE__, (u32) (b), (u32) (c)); b 322 drivers/net/wireless/intel/ipw2x00/ipw2200.c _ipw_write_reg32(a, b, c); b 408 drivers/net/wireless/intel/ipw2x00/ipw2200.c #define ipw_read_indirect(a, b, c, d) ({ \ b 410 drivers/net/wireless/intel/ipw2x00/ipw2200.c __LINE__, (u32)(b), (u32)(d)); \ b 411 drivers/net/wireless/intel/ipw2x00/ipw2200.c _ipw_read_indirect(a, b, c, d); \ b 417 drivers/net/wireless/intel/ipw2x00/ipw2200.c #define ipw_write_indirect(a, b, c, d) do { \ b 419 drivers/net/wireless/intel/ipw2x00/ipw2200.c __LINE__, (u32)(b), (u32)(d)); \ b 420 drivers/net/wireless/intel/ipw2x00/ipw2200.c _ipw_write_indirect(a, b, c, d); \ b 538 drivers/net/wireless/intel/ipw2x00/ipw2200.h __le32 b[SUP_RATE_11B_MAX_NUM_CHANNELS]; b 543 drivers/net/wireless/intel/ipw2x00/ipw2200.h __le32 b[SUP_RATE_11B_MAX_NUM_CHANNELS]; b 289 drivers/net/wireless/intel/iwlegacy/3945.h s32 a, b, c, d, e; /* coefficients for voltage->power b 781 drivers/net/wireless/intel/iwlegacy/4965-rs.c struct il_scale_tbl_info *b) b 783 drivers/net/wireless/intel/iwlegacy/4965-rs.c return (a->lq_type == b->lq_type && a->ant_type == b->ant_type && b 784 drivers/net/wireless/intel/iwlegacy/4965-rs.c a->is_SGI == b->is_SGI); b 630 drivers/net/wireless/intel/iwlegacy/4965.c s32 b = -1; b 632 drivers/net/wireless/intel/iwlegacy/4965.c for (b = 0; b < EEPROM_TX_POWER_BANDS; b++) { b 633 drivers/net/wireless/intel/iwlegacy/4965.c if (il->calib_info->band_info[b].ch_from == 0) b 636 drivers/net/wireless/intel/iwlegacy/4965.c if (channel >= il->calib_info->band_info[b].ch_from && b 637 drivers/net/wireless/intel/iwlegacy/4965.c channel <= il->calib_info->band_info[b].ch_to) b 641 drivers/net/wireless/intel/iwlegacy/4965.c return b; b 2694 drivers/net/wireless/intel/iwlegacy/commands.h __le32 b[SUP_RATE_11B_MAX_NUM_CHANNELS]; b 2699 drivers/net/wireless/intel/iwlegacy/commands.h __le32 b[SUP_RATE_11B_MAX_NUM_CHANNELS]; b 2497 drivers/net/wireless/intel/iwlwifi/dvm/commands.h __le32 b[SUP_RATE_11B_MAX_NUM_CHANNELS]; b 2502 drivers/net/wireless/intel/iwlwifi/dvm/commands.h __le32 b[SUP_RATE_11B_MAX_NUM_CHANNELS]; b 837 drivers/net/wireless/intel/iwlwifi/dvm/rs.c struct iwl_scale_tbl_info *b) b 839 drivers/net/wireless/intel/iwlwifi/dvm/rs.c return (a->lq_type == b->lq_type) && (a->ant_type == b->ant_type) && b 840 drivers/net/wireless/intel/iwlwifi/dvm/rs.c (a->is_SGI == b->is_SGI); b 1141 drivers/net/wireless/intel/iwlwifi/mvm/rs.c struct rs_rate *b) b 1146 drivers/net/wireless/intel/iwlwifi/mvm/rs.c ant_match = (b->ant == ANT_A || b->ant == ANT_B); b 1148 drivers/net/wireless/intel/iwlwifi/mvm/rs.c ant_match = (a->ant == b->ant); b 1150 drivers/net/wireless/intel/iwlwifi/mvm/rs.c return (a->type == b->type) && (a->bw == b->bw) && (a->sgi == b->sgi) b 553 drivers/net/wireless/intel/iwlwifi/mvm/tt.c static int compare_temps(const void *a, const void *b) b 556 drivers/net/wireless/intel/iwlwifi/mvm/tt.c (s16)le16_to_cpu(*(__le16 *)b)); b 108 drivers/net/wireless/intersil/p54/eeprom.c const struct p54_channel_entry *b = _b; b 110 drivers/net/wireless/intersil/p54/eeprom.c return a->freq - b->freq; b 117 drivers/net/wireless/intersil/p54/eeprom.c const struct p54_rssi_db_entry *b = _b; b 119 drivers/net/wireless/intersil/p54/eeprom.c return a->freq - b->freq; b 616 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c #define TID_CHECK(a, b) ( \ b 618 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c ((b) & cpu_to_le16(IEEE80211_BAR_CTRL_TID_INFO_MASK))) \ b 664 drivers/net/wireless/realtek/rtl8xxxu/rtl8xxxu.h int b:4; b 666 drivers/net/wireless/realtek/rtl8xxxu/rtl8xxxu.h int b:4; b 536 drivers/net/wireless/realtek/rtl8xxxu/rtl8xxxu_8192e.c ofdmbase += priv->ofdm_tx_power_diff[tx_idx].b; b 545 drivers/net/wireless/realtek/rtl8xxxu/rtl8xxxu_8192e.c mcsbase += priv->ht40_tx_power_diff[tx_idx++].b; b 547 drivers/net/wireless/realtek/rtl8xxxu/rtl8xxxu_8192e.c mcsbase += priv->ht20_tx_power_diff[tx_idx++].b; b 580 drivers/net/wireless/realtek/rtl8xxxu/rtl8xxxu_8192e.c efuse->tx_power_index_A.ht20_ofdm_1s_diff.b; b 581 drivers/net/wireless/realtek/rtl8xxxu/rtl8xxxu_8192e.c priv->ht20_tx_power_diff[0].b = b 582 drivers/net/wireless/realtek/rtl8xxxu/rtl8xxxu_8192e.c efuse->tx_power_index_B.ht20_ofdm_1s_diff.b; b 585 drivers/net/wireless/realtek/rtl8xxxu/rtl8xxxu_8192e.c priv->ht40_tx_power_diff[0].b = 0; b 590 drivers/net/wireless/realtek/rtl8xxxu/rtl8xxxu_8192e.c priv->ofdm_tx_power_diff[i].b = b 595 drivers/net/wireless/realtek/rtl8xxxu/rtl8xxxu_8192e.c priv->ht20_tx_power_diff[i].b = b 600 drivers/net/wireless/realtek/rtl8xxxu/rtl8xxxu_8192e.c priv->ht40_tx_power_diff[i].b = b 379 drivers/net/wireless/realtek/rtl8xxxu/rtl8xxxu_8723b.c ofdmbase += priv->ofdm_tx_power_diff[tx_idx].b; b 387 drivers/net/wireless/realtek/rtl8xxxu/rtl8xxxu_8723b.c mcsbase += priv->ht40_tx_power_diff[tx_idx++].b; b 389 drivers/net/wireless/realtek/rtl8xxxu/rtl8xxxu_8723b.c mcsbase += priv->ht20_tx_power_diff[tx_idx++].b; b 420 drivers/net/wireless/realtek/rtl8xxxu/rtl8xxxu_8723b.c priv->ofdm_tx_power_diff[0].b = b 424 drivers/net/wireless/realtek/rtl8xxxu/rtl8xxxu_8723b.c efuse->tx_power_index_A.ht20_ofdm_1s_diff.b; b 425 drivers/net/wireless/realtek/rtl8xxxu/rtl8xxxu_8723b.c priv->ht20_tx_power_diff[0].b = b 426 drivers/net/wireless/realtek/rtl8xxxu/rtl8xxxu_8723b.c efuse->tx_power_index_B.ht20_ofdm_1s_diff.b; b 429 drivers/net/wireless/realtek/rtl8xxxu/rtl8xxxu_8723b.c priv->ht40_tx_power_diff[0].b = 0; b 434 drivers/net/wireless/realtek/rtl8xxxu/rtl8xxxu_8723b.c priv->ofdm_tx_power_diff[i].b = b 439 drivers/net/wireless/realtek/rtl8xxxu/rtl8xxxu_8723b.c priv->ht20_tx_power_diff[i].b = b 444 drivers/net/wireless/realtek/rtl8xxxu/rtl8xxxu_8723b.c priv->ht40_tx_power_diff[i].b = b 1416 drivers/net/wireless/realtek/rtl8xxxu/rtl8xxxu_core.c ofdmbase[1] = ofdm[1] + priv->ofdm_tx_power_index_diff[group].b; b 1422 drivers/net/wireless/realtek/rtl8xxxu/rtl8xxxu_core.c mcsbase[1] += priv->ht20_tx_power_index_diff[group].b; b 1428 drivers/net/wireless/realtek/rtl8xxxu/rtl8xxxu_core.c if (ofdm[1] > priv->ht40_2s_tx_power_index_diff[group].b) b 1429 drivers/net/wireless/realtek/rtl8xxxu/rtl8xxxu_core.c ofdm[1] -= priv->ht40_2s_tx_power_index_diff[group].b; b 163 drivers/net/wireless/realtek/rtlwifi/debug.c RTL_DEBUG_IMPL_BB_SERIES(b, 0x0b00); b 208 drivers/net/wireless/realtek/rtlwifi/debug.c RTL_DEBUG_IMPL_RF_SERIES(b, RF90_PATH_B); b 602 drivers/net/wireless/realtek/rtw88/debug.c rtw_debug_impl_bb(b, 0x0b00); b 1248 drivers/net/wireless/rsi/rsi_91x_mgmt.c static int rsi_compare(const void *a, const void *b) b 1251 drivers/net/wireless/rsi/rsi_91x_mgmt.c u16 _b = *(const u16 *)(b); b 16 drivers/net/wireless/ti/wl12xx/debugfs.c #define WL12XX_DEBUGFS_FWSTATS_FILE(a, b, c) \ b 17 drivers/net/wireless/ti/wl12xx/debugfs.c DEBUGFS_FWSTATS_FILE(a, b, c, wl12xx_acx_statistics) b 21 drivers/net/wireless/ti/wl18xx/debugfs.c #define WL18XX_DEBUGFS_FWSTATS_FILE(a, b, c) \ b 22 drivers/net/wireless/ti/wl18xx/debugfs.c DEBUGFS_FWSTATS_FILE(a, b, c, wl18xx_acx_statistics) b 23 drivers/net/wireless/ti/wl18xx/debugfs.c #define WL18XX_DEBUGFS_FWSTATS_FILE_ARRAY(a, b, c) \ b 24 drivers/net/wireless/ti/wl18xx/debugfs.c DEBUGFS_FWSTATS_FILE_ARRAY(a, b, c, wl18xx_acx_statistics) b 1695 drivers/net/wireless/ti/wlcore/cmd.c int ret = 0, i, b, ch_bit_idx; b 1708 drivers/net/wireless/ti/wlcore/cmd.c for (b = NL80211_BAND_2GHZ; b <= NL80211_BAND_5GHZ; b++) { b 1709 drivers/net/wireless/ti/wlcore/cmd.c band = wiphy->bands[b]; b 1723 drivers/net/wireless/ti/wlcore/cmd.c ch_bit_idx = wlcore_get_reg_conf_ch_idx(b, ch); b 68 drivers/net/wireless/wl3501_cs.c #define wl3501_outb(a, b) { outb(a, b); slow_down_io(); } b 69 drivers/net/wireless/wl3501_cs.c #define wl3501_outb_p(a, b) { outb_p(a, b); slow_down_io(); } b 70 drivers/net/wireless/wl3501_cs.c #define wl3501_outsb(a, b, c) { outsb(a, b, c); slow_down_io(); } b 84 drivers/net/wireless/zydas/zd1211rw/zd_rf_uw2453.c #define RF_CHANPAIR(a,b) [CHAN_TO_PAIRIDX(a)] b 625 drivers/net/xen-netback/xenbus.c unsigned long b, u; b 637 drivers/net/xen-netback/xenbus.c b = simple_strtoul(s, &e, 10); b 646 drivers/net/xen-netback/xenbus.c *bytes = b; b 22 drivers/nvdimm/label.c static u32 best_seq(u32 a, u32 b) b 25 drivers/nvdimm/label.c b &= NSINDEX_SEQ_MASK; b 27 drivers/nvdimm/label.c if (a == 0 || a == b) b 28 drivers/nvdimm/label.c return b; b 29 drivers/nvdimm/label.c else if (b == 0) b 31 drivers/nvdimm/label.c else if (nd_inc_seq(a) == b) b 32 drivers/nvdimm/label.c return b; b 2276 drivers/nvdimm/namespace_devs.c static int cmp_dpa(const void *a, const void *b) b 2279 drivers/nvdimm/namespace_devs.c const struct device *dev_b = *(const struct device **) b; b 482 drivers/nvdimm/security.c #define C(a, b, c) a b 485 drivers/nvdimm/security.c #define C(a, b, c) { b, c } b 1760 drivers/nvme/host/core.c static bool nvme_ns_ids_equal(struct nvme_ns_ids *a, struct nvme_ns_ids *b) b 1762 drivers/nvme/host/core.c return uuid_equal(&a->uuid, &b->uuid) && b 1763 drivers/nvme/host/core.c memcmp(&a->nguid, &b->nguid, sizeof(a->nguid)) == 0 && b 1764 drivers/nvme/host/core.c memcmp(&a->eui64, &b->eui64, sizeof(a->eui64)) == 0; b 3427 drivers/nvme/host/core.c static int ns_cmp(void *priv, struct list_head *a, struct list_head *b) b 3430 drivers/nvme/host/core.c struct nvme_ns *nsb = container_of(b, struct nvme_ns, list); b 141 drivers/nvme/target/io-cmd-bdev.c if (bio != &req->b.inline_bio) b 174 drivers/nvme/target/io-cmd-bdev.c bio = &req->b.inline_bio; b 208 drivers/nvme/target/io-cmd-bdev.c struct bio *bio = &req->b.inline_bio; b 298 drivers/nvme/target/nvmet.h } b; b 909 drivers/nvmem/core.c u8 *p, *b; b 912 drivers/nvmem/core.c p = b = buf; b 915 drivers/nvmem/core.c *b++ >>= bit_offset; b 920 drivers/nvmem/core.c *p |= *b << (BITS_PER_BYTE - bit_offset); b 922 drivers/nvmem/core.c p = b; b 923 drivers/nvmem/core.c *b++ >>= bit_offset; b 998 drivers/nvmem/core.c u8 v, *p, *buf, *b, pbyte, pbits; b 1006 drivers/nvmem/core.c p = b = buf; b 1009 drivers/nvmem/core.c pbyte = *b; b 1010 drivers/nvmem/core.c *b <<= bit_offset; b 1016 drivers/nvmem/core.c *b++ |= GENMASK(bit_offset - 1, 0) & v; b 1022 drivers/nvmem/core.c pbyte = *b; b 1023 drivers/nvmem/core.c p = b; b 1024 drivers/nvmem/core.c *b <<= bit_offset; b 1025 drivers/nvmem/core.c *b++ |= pbits; b 193 drivers/of/of_reserved_mem.c static int __init __rmem_cmp(const void *a, const void *b) b 195 drivers/of/of_reserved_mem.c const struct reserved_mem *ra = a, *rb = b; b 77 drivers/oprofile/cpu_buffer.c struct oprofile_cpu_buffer *b = &per_cpu(op_cpu_buffer, i); b 79 drivers/oprofile/cpu_buffer.c b->last_task = NULL; b 80 drivers/oprofile/cpu_buffer.c b->last_is_kernel = -1; b 81 drivers/oprofile/cpu_buffer.c b->tracing = 0; b 82 drivers/oprofile/cpu_buffer.c b->buffer_size = buffer_size; b 83 drivers/oprofile/cpu_buffer.c b->sample_received = 0; b 84 drivers/oprofile/cpu_buffer.c b->sample_lost_overflow = 0; b 85 drivers/oprofile/cpu_buffer.c b->backtrace_aborted = 0; b 86 drivers/oprofile/cpu_buffer.c b->sample_invalid_eip = 0; b 87 drivers/oprofile/cpu_buffer.c b->cpu = i; b 88 drivers/oprofile/cpu_buffer.c INIT_DELAYED_WORK(&b->work, wq_sync_buffer); b 104 drivers/oprofile/cpu_buffer.c struct oprofile_cpu_buffer *b = &per_cpu(op_cpu_buffer, i); b 110 drivers/oprofile/cpu_buffer.c schedule_delayed_work_on(i, &b->work, DEFAULT_TIMER_EXPIRE + i); b 124 drivers/oprofile/cpu_buffer.c struct oprofile_cpu_buffer *b = &per_cpu(op_cpu_buffer, i); b 127 drivers/oprofile/cpu_buffer.c flush_delayed_work(&b->work); b 454 drivers/oprofile/cpu_buffer.c struct oprofile_cpu_buffer *b = b 456 drivers/oprofile/cpu_buffer.c if (b->cpu != smp_processor_id() && !cpu_online(b->cpu)) { b 457 drivers/oprofile/cpu_buffer.c cancel_delayed_work(&b->work); b 460 drivers/oprofile/cpu_buffer.c sync_buffer(b->cpu); b 464 drivers/oprofile/cpu_buffer.c schedule_delayed_work(&b->work, DEFAULT_TIMER_EXPIRE); b 277 drivers/parisc/dino.c DINO_PORT_IN(b, 8, 3) b 293 drivers/parisc/dino.c DINO_PORT_OUT(b, 8, 3) b 440 drivers/parport/parport_ip32.c #define CHECK_EXTRA_BITS(p, b, m) \ b 442 drivers/parport/parport_ip32.c unsigned int __b = (b), __m = (m); \ b 446 drivers/parport/parport_ip32.c (p)->name, __func__, #b, __b, __m); \ b 57 drivers/parport/share.c static void dead_write_lines(struct parport *p, unsigned char b){} b 59 drivers/parport/share.c static unsigned char dead_frob_lines(struct parport *p, unsigned char b, b 64 drivers/parport/share.c static size_t dead_write(struct parport *p, const void *b, size_t l, int f) b 66 drivers/parport/share.c static size_t dead_read(struct parport *p, void *b, size_t l, int f) b 109 drivers/pci/controller/pci-ftpci100.c #define PCI_CONF_BUS(b) (((b) & 0xFF) << 16) b 1213 drivers/pci/controller/pci-hyperv.c params->device_id = (hbus->hdev->dev_instance.b[5] << 24) | b 1214 drivers/pci/controller/pci-hyperv.c (hbus->hdev->dev_instance.b[4] << 16) | b 1215 drivers/pci/controller/pci-hyperv.c (hbus->hdev->dev_instance.b[7] << 8) | b 1216 drivers/pci/controller/pci-hyperv.c (hbus->hdev->dev_instance.b[6] & 0xf8) | b 2893 drivers/pci/controller/pci-hyperv.c dom_req = hdev->dev_instance.b[5] << 8 | hdev->dev_instance.b[4]; b 47 drivers/pci/controller/pci-mvebu.c #define PCIE_CONF_BUS(b) (((b) & 0xff) << 16) b 122 drivers/pci/controller/pci-xgene.c unsigned int b, d, f; b 125 drivers/pci/controller/pci-xgene.c b = bus->number; b 130 drivers/pci/controller/pci-xgene.c rtdid_val = (b << 8) | (d << 3) | f; b 88 drivers/pci/controller/pcie-cadence-ep.c u32 addr0, addr1, reg, cfg, b, aperture, ctrl; b 131 drivers/pci/controller/pcie-cadence-ep.c b = bar; b 134 drivers/pci/controller/pcie-cadence-ep.c b = bar - BAR_4; b 138 drivers/pci/controller/pcie-cadence-ep.c cfg &= ~(CDNS_PCIE_LM_EP_FUNC_BAR_CFG_BAR_APERTURE_MASK(b) | b 139 drivers/pci/controller/pcie-cadence-ep.c CDNS_PCIE_LM_EP_FUNC_BAR_CFG_BAR_CTRL_MASK(b)); b 140 drivers/pci/controller/pcie-cadence-ep.c cfg |= (CDNS_PCIE_LM_EP_FUNC_BAR_CFG_BAR_APERTURE(b, aperture) | b 141 drivers/pci/controller/pcie-cadence-ep.c CDNS_PCIE_LM_EP_FUNC_BAR_CFG_BAR_CTRL(b, ctrl)); b 153 drivers/pci/controller/pcie-cadence-ep.c u32 reg, cfg, b, ctrl; b 157 drivers/pci/controller/pcie-cadence-ep.c b = bar; b 160 drivers/pci/controller/pcie-cadence-ep.c b = bar - BAR_4; b 165 drivers/pci/controller/pcie-cadence-ep.c cfg &= ~(CDNS_PCIE_LM_EP_FUNC_BAR_CFG_BAR_APERTURE_MASK(b) | b 166 drivers/pci/controller/pcie-cadence-ep.c CDNS_PCIE_LM_EP_FUNC_BAR_CFG_BAR_CTRL_MASK(b)); b 167 drivers/pci/controller/pcie-cadence-ep.c cfg |= CDNS_PCIE_LM_EP_FUNC_BAR_CFG_BAR_CTRL(b, ctrl); b 48 drivers/pci/controller/pcie-cadence.h #define CDNS_PCIE_LM_EP_FUNC_BAR_CFG_BAR_APERTURE_MASK(b) \ b 49 drivers/pci/controller/pcie-cadence.h (GENMASK(4, 0) << ((b) * 8)) b 50 drivers/pci/controller/pcie-cadence.h #define CDNS_PCIE_LM_EP_FUNC_BAR_CFG_BAR_APERTURE(b, a) \ b 51 drivers/pci/controller/pcie-cadence.h (((a) << ((b) * 8)) & CDNS_PCIE_LM_EP_FUNC_BAR_CFG_BAR_APERTURE_MASK(b)) b 52 drivers/pci/controller/pcie-cadence.h #define CDNS_PCIE_LM_EP_FUNC_BAR_CFG_BAR_CTRL_MASK(b) \ b 53 drivers/pci/controller/pcie-cadence.h (GENMASK(7, 5) << ((b) * 8)) b 54 drivers/pci/controller/pcie-cadence.h #define CDNS_PCIE_LM_EP_FUNC_BAR_CFG_BAR_CTRL(b, c) \ b 55 drivers/pci/controller/pcie-cadence.h (((c) << ((b) * 8 + 5)) & CDNS_PCIE_LM_EP_FUNC_BAR_CFG_BAR_CTRL_MASK(b)) b 127 drivers/pci/controller/pcie-rcar.c #define PCIE_CONF_BUS(b) (((b) & 0xff) << 24) b 169 drivers/pci/controller/pcie-rockchip-ep.c u32 addr0, addr1, reg, cfg, b, aperture, ctrl; b 205 drivers/pci/controller/pcie-rockchip-ep.c b = bar; b 208 drivers/pci/controller/pcie-rockchip-ep.c b = bar - BAR_4; b 215 drivers/pci/controller/pcie-rockchip-ep.c cfg &= ~(ROCKCHIP_PCIE_CORE_EP_FUNC_BAR_CFG_BAR_APERTURE_MASK(b) | b 216 drivers/pci/controller/pcie-rockchip-ep.c ROCKCHIP_PCIE_CORE_EP_FUNC_BAR_CFG_BAR_CTRL_MASK(b)); b 217 drivers/pci/controller/pcie-rockchip-ep.c cfg |= (ROCKCHIP_PCIE_CORE_EP_FUNC_BAR_CFG_BAR_APERTURE(b, aperture) | b 218 drivers/pci/controller/pcie-rockchip-ep.c ROCKCHIP_PCIE_CORE_EP_FUNC_BAR_CFG_BAR_CTRL(b, ctrl)); b 234 drivers/pci/controller/pcie-rockchip-ep.c u32 reg, cfg, b, ctrl; b 239 drivers/pci/controller/pcie-rockchip-ep.c b = bar; b 242 drivers/pci/controller/pcie-rockchip-ep.c b = bar - BAR_4; b 247 drivers/pci/controller/pcie-rockchip-ep.c cfg &= ~(ROCKCHIP_PCIE_CORE_EP_FUNC_BAR_CFG_BAR_APERTURE_MASK(b) | b 248 drivers/pci/controller/pcie-rockchip-ep.c ROCKCHIP_PCIE_CORE_EP_FUNC_BAR_CFG_BAR_CTRL_MASK(b)); b 249 drivers/pci/controller/pcie-rockchip-ep.c cfg |= ROCKCHIP_PCIE_CORE_EP_FUNC_BAR_CFG_BAR_CTRL(b, ctrl); b 267 drivers/pci/controller/pcie-rockchip.h #define ROCKCHIP_PCIE_CORE_EP_FUNC_BAR_CFG_BAR_APERTURE_MASK(b) \ b 268 drivers/pci/controller/pcie-rockchip.h (GENMASK(4, 0) << ((b) * 8)) b 269 drivers/pci/controller/pcie-rockchip.h #define ROCKCHIP_PCIE_CORE_EP_FUNC_BAR_CFG_BAR_APERTURE(b, a) \ b 270 drivers/pci/controller/pcie-rockchip.h (((a) << ((b) * 8)) & \ b 271 drivers/pci/controller/pcie-rockchip.h ROCKCHIP_PCIE_CORE_EP_FUNC_BAR_CFG_BAR_APERTURE_MASK(b)) b 272 drivers/pci/controller/pcie-rockchip.h #define ROCKCHIP_PCIE_CORE_EP_FUNC_BAR_CFG_BAR_CTRL_MASK(b) \ b 273 drivers/pci/controller/pcie-rockchip.h (GENMASK(7, 5) << ((b) * 8)) b 274 drivers/pci/controller/pcie-rockchip.h #define ROCKCHIP_PCIE_CORE_EP_FUNC_BAR_CFG_BAR_CTRL(b, c) \ b 275 drivers/pci/controller/pcie-rockchip.h (((c) << ((b) * 8 + 5)) & \ b 276 drivers/pci/controller/pcie-rockchip.h ROCKCHIP_PCIE_CORE_EP_FUNC_BAR_CFG_BAR_CTRL_MASK(b)) b 691 drivers/pci/hotplug/ibmphp_core.c struct pci_bus *bus, *b; b 716 drivers/pci/hotplug/ibmphp_core.c b = pci_scan_bus(busno, ibmphp_pci_bus->ops, NULL); b 717 drivers/pci/hotplug/ibmphp_core.c if (!b) b 720 drivers/pci/hotplug/ibmphp_core.c pci_bus_add_devices(b); b 325 drivers/pci/p2pdma.c static bool host_bridge_whitelist(struct pci_dev *a, struct pci_dev *b) b 328 drivers/pci/p2pdma.c struct pci_host_bridge *host_b = pci_find_host_bridge(b->bus); b 344 drivers/pci/p2pdma.c struct pci_dev *a = provider, *b = client, *bb; b 366 drivers/pci/p2pdma.c bb = b; b 386 drivers/pci/p2pdma.c bb = b; b 421 drivers/pci/pci-sysfs.c struct pci_bus *b = NULL; b 428 drivers/pci/pci-sysfs.c while ((b = pci_find_next_bus(b)) != NULL) b 429 drivers/pci/pci-sysfs.c pci_rescan_bus(b); b 916 drivers/pci/pci-sysfs.c void __weak pci_adjust_legacy_attr(struct pci_bus *b, b 932 drivers/pci/pci-sysfs.c void pci_create_legacy_files(struct pci_bus *b) b 936 drivers/pci/pci-sysfs.c b->legacy_io = kcalloc(2, sizeof(struct bin_attribute), b 938 drivers/pci/pci-sysfs.c if (!b->legacy_io) b 941 drivers/pci/pci-sysfs.c sysfs_bin_attr_init(b->legacy_io); b 942 drivers/pci/pci-sysfs.c b->legacy_io->attr.name = "legacy_io"; b 943 drivers/pci/pci-sysfs.c b->legacy_io->size = 0xffff; b 944 drivers/pci/pci-sysfs.c b->legacy_io->attr.mode = 0600; b 945 drivers/pci/pci-sysfs.c b->legacy_io->read = pci_read_legacy_io; b 946 drivers/pci/pci-sysfs.c b->legacy_io->write = pci_write_legacy_io; b 947 drivers/pci/pci-sysfs.c b->legacy_io->mmap = pci_mmap_legacy_io; b 948 drivers/pci/pci-sysfs.c pci_adjust_legacy_attr(b, pci_mmap_io); b 949 drivers/pci/pci-sysfs.c error = device_create_bin_file(&b->dev, b->legacy_io); b 954 drivers/pci/pci-sysfs.c b->legacy_mem = b->legacy_io + 1; b 955 drivers/pci/pci-sysfs.c sysfs_bin_attr_init(b->legacy_mem); b 956 drivers/pci/pci-sysfs.c b->legacy_mem->attr.name = "legacy_mem"; b 957 drivers/pci/pci-sysfs.c b->legacy_mem->size = 1024*1024; b 958 drivers/pci/pci-sysfs.c b->legacy_mem->attr.mode = 0600; b 959 drivers/pci/pci-sysfs.c b->legacy_mem->mmap = pci_mmap_legacy_mem; b 960 drivers/pci/pci-sysfs.c pci_adjust_legacy_attr(b, pci_mmap_mem); b 961 drivers/pci/pci-sysfs.c error = device_create_bin_file(&b->dev, b->legacy_mem); b 968 drivers/pci/pci-sysfs.c device_remove_bin_file(&b->dev, b->legacy_io); b 970 drivers/pci/pci-sysfs.c kfree(b->legacy_io); b 971 drivers/pci/pci-sysfs.c b->legacy_io = NULL; b 973 drivers/pci/pci-sysfs.c dev_warn(&b->dev, "could not create legacy I/O port and ISA memory resources in sysfs\n"); b 976 drivers/pci/pci-sysfs.c void pci_remove_legacy_files(struct pci_bus *b) b 978 drivers/pci/pci-sysfs.c if (b->legacy_io) { b 979 drivers/pci/pci-sysfs.c device_remove_bin_file(&b->dev, b->legacy_io); b 980 drivers/pci/pci-sysfs.c device_remove_bin_file(&b->dev, b->legacy_mem); b 981 drivers/pci/pci-sysfs.c kfree(b->legacy_io); /* both are allocated here */ b 547 drivers/pci/probe.c struct pci_bus *b; b 549 drivers/pci/probe.c b = kzalloc(sizeof(*b), GFP_KERNEL); b 550 drivers/pci/probe.c if (!b) b 553 drivers/pci/probe.c INIT_LIST_HEAD(&b->node); b 554 drivers/pci/probe.c INIT_LIST_HEAD(&b->children); b 555 drivers/pci/probe.c INIT_LIST_HEAD(&b->devices); b 556 drivers/pci/probe.c INIT_LIST_HEAD(&b->slots); b 557 drivers/pci/probe.c INIT_LIST_HEAD(&b->resources); b 558 drivers/pci/probe.c b->max_bus_speed = PCI_SPEED_UNKNOWN; b 559 drivers/pci/probe.c b->cur_bus_speed = PCI_SPEED_UNKNOWN; b 562 drivers/pci/probe.c b->domain_nr = parent->domain_nr; b 564 drivers/pci/probe.c return b; b 808 drivers/pci/probe.c struct pci_bus *b; b 815 drivers/pci/probe.c for (b = bus, d = NULL; !d && !pci_is_root_bus(b); b = b->parent) { b 816 drivers/pci/probe.c if (b->self) b 817 drivers/pci/probe.c d = dev_get_msi_domain(&b->self->dev); b 821 drivers/pci/probe.c d = pci_host_bridge_msi_domain(b); b 830 drivers/pci/probe.c struct pci_bus *bus, *b; b 854 drivers/pci/probe.c b = pci_find_bus(pci_domain_nr(bus), bridge->busnr); b 855 drivers/pci/probe.c if (b) { b 857 drivers/pci/probe.c dev_dbg(&b->dev, "bus already known\n"); b 2964 drivers/pci/probe.c int pci_bus_insert_busn_res(struct pci_bus *b, int bus, int bus_max) b 2966 drivers/pci/probe.c struct resource *res = &b->busn_res; b 2973 drivers/pci/probe.c if (!pci_is_root_bus(b)) b 2974 drivers/pci/probe.c parent_res = &b->parent->busn_res; b 2976 drivers/pci/probe.c parent_res = get_pci_domain_busn_res(pci_domain_nr(b)); b 2983 drivers/pci/probe.c dev_info(&b->dev, b 2985 drivers/pci/probe.c res, pci_is_root_bus(b) ? "domain " : "", b 2991 drivers/pci/probe.c int pci_bus_update_busn_res_end(struct pci_bus *b, int bus_max) b 2993 drivers/pci/probe.c struct resource *res = &b->busn_res; b 3003 drivers/pci/probe.c dev_info(&b->dev, "busn_res: %pR end %s updated to %02x\n", b 3007 drivers/pci/probe.c pci_bus_insert_busn_res(b, res->start, res->end); b 3012 drivers/pci/probe.c void pci_bus_release_busn_res(struct pci_bus *b) b 3014 drivers/pci/probe.c struct resource *res = &b->busn_res; b 3021 drivers/pci/probe.c dev_info(&b->dev, "busn_res: %pR %s released\n", b 3029 drivers/pci/probe.c struct pci_bus *b; b 3045 drivers/pci/probe.c b = bridge->bus; b 3049 drivers/pci/probe.c dev_info(&b->dev, b 3052 drivers/pci/probe.c pci_bus_insert_busn_res(b, bus, 255); b 3055 drivers/pci/probe.c max = pci_scan_child_bus(b); b 3058 drivers/pci/probe.c pci_bus_update_busn_res_end(b, max); b 3069 drivers/pci/probe.c struct pci_bus *b; b 3078 drivers/pci/probe.c b = pci_create_root_bus(parent, bus, ops, sysdata, resources); b 3079 drivers/pci/probe.c if (!b) b 3083 drivers/pci/probe.c dev_info(&b->dev, b 3086 drivers/pci/probe.c pci_bus_insert_busn_res(b, bus, 255); b 3089 drivers/pci/probe.c max = pci_scan_child_bus(b); b 3092 drivers/pci/probe.c pci_bus_update_busn_res_end(b, max); b 3094 drivers/pci/probe.c return b; b 3102 drivers/pci/probe.c struct pci_bus *b; b 3107 drivers/pci/probe.c b = pci_create_root_bus(NULL, bus, ops, sysdata, &resources); b 3108 drivers/pci/probe.c if (b) { b 3109 drivers/pci/probe.c pci_scan_child_bus(b); b 3113 drivers/pci/probe.c return b; b 3185 drivers/pci/probe.c const struct pci_dev *b = to_pci_dev(d_b); b 3187 drivers/pci/probe.c if (pci_domain_nr(a->bus) < pci_domain_nr(b->bus)) return -1; b 3188 drivers/pci/probe.c else if (pci_domain_nr(a->bus) > pci_domain_nr(b->bus)) return 1; b 3190 drivers/pci/probe.c if (a->bus->number < b->bus->number) return -1; b 3191 drivers/pci/probe.c else if (a->bus->number > b->bus->number) return 1; b 3193 drivers/pci/probe.c if (a->devfn < b->devfn) return -1; b 3194 drivers/pci/probe.c else if (a->devfn > b->devfn) return 1; b 2388 drivers/pci/quirks.c uint8_t b; b 2390 drivers/pci/quirks.c if (pci_read_config_byte(dev, 0xf41, &b) == 0) { b 2391 drivers/pci/quirks.c if (!(b & 0x20)) { b 2392 drivers/pci/quirks.c pci_write_config_byte(dev, 0xf41, b | 0x20); b 2417 drivers/pci/quirks.c uint8_t b; b 2428 drivers/pci/quirks.c if (pci_read_config_byte(dev, 0x76, &b) == 0) { b 2429 drivers/pci/quirks.c if (b & 0x40) { b 2431 drivers/pci/quirks.c pci_write_config_byte(dev, 0x76, b ^ 0x40); b 2437 drivers/pci/quirks.c if (pci_read_config_byte(dev, 0x72, &b) == 0) { b 2438 drivers/pci/quirks.c if (b != 0) { b 163 drivers/pci/search.c struct pci_bus *b = NULL; b 169 drivers/pci/search.c b = list_entry(n, struct pci_bus, node); b 171 drivers/pci/search.c return b; b 1186 drivers/pci/setup-bus.c struct pci_bus *b = dev->subordinate; b 1187 drivers/pci/setup-bus.c if (!b) b 1192 drivers/pci/setup-bus.c pci_bus_size_cardbus(b, realloc_head); b 1197 drivers/pci/setup-bus.c __pci_bus_size_bridges(b, realloc_head); b 1300 drivers/pci/setup-bus.c static void assign_fixed_resource_on_bus(struct pci_bus *b, struct resource *r) b 1307 drivers/pci/setup-bus.c pci_bus_for_each_resource(b, parent_r, i) { b 1326 drivers/pci/setup-bus.c struct pci_bus *b; b 1333 drivers/pci/setup-bus.c b = dev->bus; b 1334 drivers/pci/setup-bus.c while (b && !r->parent) { b 1335 drivers/pci/setup-bus.c assign_fixed_resource_on_bus(b, r); b 1336 drivers/pci/setup-bus.c b = b->parent; b 1345 drivers/pci/setup-bus.c struct pci_bus *b; b 1353 drivers/pci/setup-bus.c b = dev->subordinate; b 1354 drivers/pci/setup-bus.c if (!b) b 1357 drivers/pci/setup-bus.c __pci_bus_assign_resources(b, realloc_head, fail_head); b 1362 drivers/pci/setup-bus.c pci_setup_bridge(b); b 1366 drivers/pci/setup-bus.c pci_setup_cardbus(b); b 1371 drivers/pci/setup-bus.c pci_domain_nr(b), b->number); b 1411 drivers/pci/setup-bus.c static void pci_bus_allocate_dev_resources(struct pci_bus *b) b 1416 drivers/pci/setup-bus.c list_for_each_entry(dev, &b->devices, bus_list) { b 1425 drivers/pci/setup-bus.c static void pci_bus_allocate_resources(struct pci_bus *b) b 1434 drivers/pci/setup-bus.c if (b->self) { b 1435 drivers/pci/setup-bus.c pci_read_bridge_bases(b); b 1436 drivers/pci/setup-bus.c pci_claim_bridge_resources(b->self); b 1439 drivers/pci/setup-bus.c list_for_each_entry(child, &b->children, node) b 1443 drivers/pci/setup-bus.c void pci_bus_claim_resources(struct pci_bus *b) b 1445 drivers/pci/setup-bus.c pci_bus_allocate_resources(b); b 1446 drivers/pci/setup-bus.c pci_bus_allocate_dev_resources(b); b 1454 drivers/pci/setup-bus.c struct pci_bus *b; b 1459 drivers/pci/setup-bus.c b = bridge->subordinate; b 1460 drivers/pci/setup-bus.c if (!b) b 1463 drivers/pci/setup-bus.c __pci_bus_assign_resources(b, add_head, fail_head); b 1467 drivers/pci/setup-bus.c pci_setup_bridge(b); b 1471 drivers/pci/setup-bus.c pci_setup_cardbus(b); b 1476 drivers/pci/setup-bus.c pci_domain_nr(b), b->number); b 1561 drivers/pci/setup-bus.c struct pci_bus *b = dev->subordinate; b 1562 drivers/pci/setup-bus.c if (!b) b 1571 drivers/pci/setup-bus.c pci_bus_release_bridge_resources(b, type, b 1600 drivers/pci/setup-bus.c struct pci_bus *b; b 1607 drivers/pci/setup-bus.c b = dev->subordinate; b 1608 drivers/pci/setup-bus.c if (!b) b 1611 drivers/pci/setup-bus.c pci_bus_dump_resources(b); b 1943 drivers/pci/setup-bus.c struct pci_bus *b; b 1945 drivers/pci/setup-bus.c b = dev->subordinate; b 1946 drivers/pci/setup-bus.c if (!b || !dev->is_hotplug_bridge) b 1969 drivers/pci/setup-bus.c pci_bus_distribute_available_resources(b, add_list, io, mmio, b 418 drivers/pci/xen-pcifront.c struct pci_bus *b) b 428 drivers/pci/xen-pcifront.c d = pci_get_slot(b, devfn); b 435 drivers/pci/xen-pcifront.c d = pci_scan_single_device(b, devfn); b 448 drivers/pci/xen-pcifront.c struct pci_bus *b; b 486 drivers/pci/xen-pcifront.c b = pci_scan_root_bus(&pdev->xdev->dev, bus, b 488 drivers/pci/xen-pcifront.c if (!b) { b 497 drivers/pci/xen-pcifront.c bus_entry->bus = b; b 503 drivers/pci/xen-pcifront.c err = pcifront_scan_bus(pdev, domain, bus, b); b 506 drivers/pci/xen-pcifront.c pci_walk_bus(b, pcifront_claim_resource, pdev); b 509 drivers/pci/xen-pcifront.c pci_bus_add_devices(b); b 525 drivers/pci/xen-pcifront.c struct pci_bus *b; b 540 drivers/pci/xen-pcifront.c b = pci_find_bus(domain, bus); b 541 drivers/pci/xen-pcifront.c if (!b) b 545 drivers/pci/xen-pcifront.c err = pcifront_scan_bus(pdev, domain, bus, b); b 548 drivers/pci/xen-pcifront.c pci_walk_bus(b, pcifront_claim_resource, pdev); b 551 drivers/pci/xen-pcifront.c pci_bus_add_devices(b); b 257 drivers/pcmcia/i82365.c static void i365_bflip(u_short sock, u_short reg, u_char mask, int b) b 260 drivers/pcmcia/i82365.c if (b) b 269 drivers/pcmcia/i82365.c u_short a, b; b 271 drivers/pcmcia/i82365.c b = i365_get(sock, reg+1); b 272 drivers/pcmcia/i82365.c return (a + (b<<8)); b 292 drivers/pcmcia/i82365.c #define flip(v,b,f) (v = ((f)<0) ? v : ((f) ? ((v)|(b)) : ((v)&(~b)))) b 118 drivers/pcmcia/o2micro.h u8 a, b; b 123 drivers/pcmcia/o2micro.h b = config_readb(socket, O2_RESERVED2); b 124 drivers/pcmcia/o2micro.h dev_dbg(&socket->dev->dev, "O2: 0x94/0xD4: %02x/%02x\n", a, b); b 160 drivers/pcmcia/o2micro.h b | O2_RES_READ_PREFETCH | O2_RES_WRITE_BURST); b 167 drivers/pcmcia/o2micro.h b & ~(O2_RES_READ_PREFETCH | O2_RES_WRITE_BURST)); b 576 drivers/pcmcia/pcmcia_resource.c u8 b = c->io[0].start & 0xff; b 577 drivers/pcmcia/pcmcia_resource.c pcmcia_write_cis_mem(s, 1, (base + CISREG_IOBASE_0)>>1, 1, &b); b 578 drivers/pcmcia/pcmcia_resource.c b = (c->io[0].start >> 8) & 0xff; b 579 drivers/pcmcia/pcmcia_resource.c pcmcia_write_cis_mem(s, 1, (base + CISREG_IOBASE_1)>>1, 1, &b); b 582 drivers/pcmcia/pcmcia_resource.c u8 b = resource_size(&c->io[0]) + resource_size(&c->io[1]) - 1; b 583 drivers/pcmcia/pcmcia_resource.c pcmcia_write_cis_mem(s, 1, (base + CISREG_IOSIZE)>>1, 1, &b); b 189 drivers/pcmcia/rsrc_nonstatic.c u_char *b, hole, most; b 194 drivers/pcmcia/rsrc_nonstatic.c b = kzalloc(256, GFP_KERNEL); b 195 drivers/pcmcia/rsrc_nonstatic.c if (!b) { b 209 drivers/pcmcia/rsrc_nonstatic.c if ((j == 8) && (++b[hole] > b[most])) b 211 drivers/pcmcia/rsrc_nonstatic.c if (b[most] == 127) b 214 drivers/pcmcia/rsrc_nonstatic.c kfree(b); b 298 drivers/pcmcia/rsrc_nonstatic.c int i, a = 0, b = -1, d; b 314 drivers/pcmcia/rsrc_nonstatic.c b &= d; b 323 drivers/pcmcia/rsrc_nonstatic.c if (b == -1) b 484 drivers/pcmcia/rsrc_nonstatic.c unsigned long b, i, ok = 0; b 510 drivers/pcmcia/rsrc_nonstatic.c b = order[i] << 12; b 511 drivers/pcmcia/rsrc_nonstatic.c if ((b >= mm.base) && (b+0x10000 <= mm.base+mm.num)) { b 513 drivers/pcmcia/rsrc_nonstatic.c sub_interval(&s_data->mem_db, b, 0x10000); b 515 drivers/pcmcia/rsrc_nonstatic.c ok += do_mem_probe(s, b, 0x10000, b 690 drivers/pcmcia/soc_common.c char *b = *p; b 693 drivers/pcmcia/soc_common.c b += sprintf(b, "%-9s:", prefix); b 696 drivers/pcmcia/soc_common.c b += sprintf(b, " %s", bits[i].name); b 697 drivers/pcmcia/soc_common.c *b++ = '\n'; b 698 drivers/pcmcia/soc_common.c *p = b; b 609 drivers/perf/arm-ccn.c static int arm_ccn_pmu_type_eq(u32 a, u32 b) b 611 drivers/perf/arm-ccn.c if (a == b) b 621 drivers/perf/arm-ccn.c switch (b) { b 700 drivers/perf/arm_pmu.c static int cpu_pm_pmu_notify(struct notifier_block *b, unsigned long cmd, b 703 drivers/perf/arm_pmu.c struct arm_pmu *armpmu = container_of(b, struct arm_pmu, cpu_pm_nb); b 193 drivers/perf/fsl_imx8_ddr_perf.c struct perf_event *b) b 197 drivers/perf/fsl_imx8_ddr_perf.c if (!ddr_perf_is_filtered(b)) b 199 drivers/perf/fsl_imx8_ddr_perf.c return ddr_perf_filter_val(a) == ddr_perf_filter_val(b); b 130 drivers/phy/freescale/phy-fsl-imx8-mipi-dphy.c u32 b = *pdenom; b 137 drivers/phy/freescale/phy-fsl-imx8-mipi-dphy.c while (b) { b 139 drivers/phy/freescale/phy-fsl-imx8-mipi-dphy.c whole = a / b; b 146 drivers/phy/freescale/phy-fsl-imx8-mipi-dphy.c c = a - (b * whole); b 147 drivers/phy/freescale/phy-fsl-imx8-mipi-dphy.c a = b; b 148 drivers/phy/freescale/phy-fsl-imx8-mipi-dphy.c b = c; b 24 drivers/phy/qualcomm/phy-qcom-ipq806x-sata.c #define __set(v, a, b) (((v) << (b)) & GENMASK(a, b)) b 375 drivers/pinctrl/bcm/pinctrl-bcm281xx.c #define BCM281XX_PIN_DESC(a, b, c) \ b 376 drivers/pinctrl/bcm/pinctrl-bcm281xx.c { .number = a, .name = b, .drv_data = &c##_pin } b 170 drivers/pinctrl/bcm/pinctrl-ns2-mux.c #define NS2_PIN_DESC(p, n, b, o, s, i, pu, d) \ b 175 drivers/pinctrl/bcm/pinctrl-ns2-mux.c .base = b, \ b 35 drivers/pinctrl/intel/pinctrl-cannonlake.c #define CNL_COMMUNITY(b, s, e, o, g) \ b 37 drivers/pinctrl/intel/pinctrl-cannonlake.c .barno = (b), \ b 49 drivers/pinctrl/intel/pinctrl-cannonlake.c #define CNLLP_COMMUNITY(b, s, e, g) \ b 50 drivers/pinctrl/intel/pinctrl-cannonlake.c CNL_COMMUNITY(b, s, e, CNL_LP_HOSTSW_OWN, g) b 52 drivers/pinctrl/intel/pinctrl-cannonlake.c #define CNLH_COMMUNITY(b, s, e, g) \ b 53 drivers/pinctrl/intel/pinctrl-cannonlake.c CNL_COMMUNITY(b, s, e, CNL_H_HOSTSW_OWN, g) b 30 drivers/pinctrl/intel/pinctrl-cedarfork.c #define CDF_COMMUNITY(b, s, e, g) \ b 32 drivers/pinctrl/intel/pinctrl-cedarfork.c .barno = (b), \ b 30 drivers/pinctrl/intel/pinctrl-denverton.c #define DNV_COMMUNITY(b, s, e, g) \ b 32 drivers/pinctrl/intel/pinctrl-denverton.c .barno = (b), \ b 34 drivers/pinctrl/intel/pinctrl-icelake.c #define ICL_COMMUNITY(b, s, e, g) \ b 36 drivers/pinctrl/intel/pinctrl-icelake.c .barno = (b), \ b 23 drivers/pinctrl/intel/pinctrl-lewisburg.c #define LBG_COMMUNITY(b, s, e) \ b 25 drivers/pinctrl/intel/pinctrl-lewisburg.c .barno = (b), \ b 72 drivers/pinctrl/intel/pinctrl-merrifield.c #define MRFLD_FAMILY(b, s, e) \ b 74 drivers/pinctrl/intel/pinctrl-merrifield.c .barno = (b), \ b 79 drivers/pinctrl/intel/pinctrl-merrifield.c #define MRFLD_FAMILY_PROTECTED(b, s, e) \ b 81 drivers/pinctrl/intel/pinctrl-merrifield.c .barno = (b), \ b 25 drivers/pinctrl/intel/pinctrl-sunrisepoint.c #define SPT_COMMUNITY(b, s, e) \ b 27 drivers/pinctrl/intel/pinctrl-sunrisepoint.c .barno = (b), \ b 47 drivers/pinctrl/intel/pinctrl-sunrisepoint.c #define SPTH_COMMUNITY(b, s, e, g) \ b 49 drivers/pinctrl/intel/pinctrl-sunrisepoint.c .barno = (b), \ b 15 drivers/pinctrl/meson/pinctrl-meson8-pmx.h #define PMX_DATA(r, b, g) \ b 18 drivers/pinctrl/meson/pinctrl-meson8-pmx.h .bit = b, \ b 22 drivers/pinctrl/meson/pinctrl-meson8-pmx.h #define GROUP(grp, r, b) \ b 28 drivers/pinctrl/meson/pinctrl-meson8-pmx.h PMX_DATA(r, b, false), \ b 227 drivers/pinctrl/nomadik/pinctrl-ab8500.c #define AB8500_PIN_GROUP(a, b) { .name = #a, .pins = a##_pins, \ b 228 drivers/pinctrl/nomadik/pinctrl-ab8500.c .npins = ARRAY_SIZE(a##_pins), .altsetting = b } b 303 drivers/pinctrl/nomadik/pinctrl-ab8500.c #define AB8500_FUNC_GROUPS(a, b...) \ b 304 drivers/pinctrl/nomadik/pinctrl-ab8500.c static const char * const a##_groups[] = { b }; b 149 drivers/pinctrl/nomadik/pinctrl-ab8505.c #define AB8505_PIN_GROUP(a, b) { .name = #a, .pins = a##_pins, \ b 150 drivers/pinctrl/nomadik/pinctrl-ab8505.c .npins = ARRAY_SIZE(a##_pins), .altsetting = b } b 190 drivers/pinctrl/nomadik/pinctrl-ab8505.c #define AB8505_FUNC_GROUPS(a, b...) \ b 191 drivers/pinctrl/nomadik/pinctrl-ab8505.c static const char * const a##_groups[] = { b }; b 97 drivers/pinctrl/nomadik/pinctrl-abx500.h #define GPIO_IRQ_CLUSTER(a, b, c) \ b 100 drivers/pinctrl/nomadik/pinctrl-abx500.h .end = b, \ b 135 drivers/pinctrl/nomadik/pinctrl-abx500.h #define ABX500_PINRANGE(a, b, c) { .offset = a, .npins = b, .altfunc = c } b 659 drivers/pinctrl/nomadik/pinctrl-nomadik-db8500.c #define DB8500_PIN_GROUP(a, b) { .name = #a, .pins = a##_pins, \ b 660 drivers/pinctrl/nomadik/pinctrl-nomadik-db8500.c .npins = ARRAY_SIZE(a##_pins), .altsetting = b } b 812 drivers/pinctrl/nomadik/pinctrl-nomadik-db8500.c #define DB8500_FUNC_GROUPS(a, b...) \ b 813 drivers/pinctrl/nomadik/pinctrl-nomadik-db8500.c static const char * const a##_groups[] = { b }; b 306 drivers/pinctrl/nomadik/pinctrl-nomadik-stn8815.c #define STN8815_PIN_GROUP(a, b) { .name = #a, .pins = a##_pins, \ b 307 drivers/pinctrl/nomadik/pinctrl-nomadik-stn8815.c .npins = ARRAY_SIZE(a##_pins), .altsetting = b } b 326 drivers/pinctrl/nomadik/pinctrl-nomadik-stn8815.c #define STN8815_FUNC_GROUPS(a, b...) \ b 327 drivers/pinctrl/nomadik/pinctrl-nomadik-stn8815.c static const char * const a##_groups[] = { b }; b 653 drivers/pinctrl/nuvoton/pinctrl-npcm7xx.c #define NPCM7XX_FUNC(a, b...) static const char *a ## _grp[] = { b } b 897 drivers/pinctrl/nuvoton/pinctrl-npcm7xx.c #define NPCM7XX_PINCFG(a, b, c, d, e, f, g, h, i, j, k) \ b 898 drivers/pinctrl/nuvoton/pinctrl-npcm7xx.c [a] { .fn0 = fn_ ## b, .reg0 = NPCM7XX_GCR_ ## c, .bit0 = d, \ b 595 drivers/pinctrl/pinctrl-coh901.c #define COH901_PINRANGE(a, b) { .offset = a, .pin_base = b } b 345 drivers/pinctrl/pinctrl-lpc18xx.c LPC_P(b,0, R, CTOUT, LCD, R, GPIO, R, R, R, 0, ND); b 346 drivers/pinctrl/pinctrl-lpc18xx.c LPC_P(b,1, R, USB1, LCD, R, GPIO, CTOUT, R, R, 0, ND); b 347 drivers/pinctrl/pinctrl-lpc18xx.c LPC_P(b,2, R, USB1, LCD, R, GPIO, CTOUT, R, R, 0, ND); b 348 drivers/pinctrl/pinctrl-lpc18xx.c LPC_P(b,3, R, USB1, LCD, R, GPIO, CTOUT, R, R, 0, ND); b 349 drivers/pinctrl/pinctrl-lpc18xx.c LPC_P(b,4, R, USB1, LCD, R, GPIO, CTIN, R, R, 0, ND); b 350 drivers/pinctrl/pinctrl-lpc18xx.c LPC_P(b,5, R, USB1, LCD, R, GPIO, CTIN, LCD_ALT, R, 0, ND); b 351 drivers/pinctrl/pinctrl-lpc18xx.c LPC_P(b,6, R, USB1, LCD, R, GPIO, CTIN, LCD_ALT, R, ADC0|6, ND); b 555 drivers/pinctrl/pinctrl-lpc18xx.c LPC18XX_PIN_P(b,0), b 556 drivers/pinctrl/pinctrl-lpc18xx.c LPC18XX_PIN_P(b,1), b 557 drivers/pinctrl/pinctrl-lpc18xx.c LPC18XX_PIN_P(b,2), b 558 drivers/pinctrl/pinctrl-lpc18xx.c LPC18XX_PIN_P(b,3), b 559 drivers/pinctrl/pinctrl-lpc18xx.c LPC18XX_PIN_P(b,4), b 560 drivers/pinctrl/pinctrl-lpc18xx.c LPC18XX_PIN_P(b,5), b 561 drivers/pinctrl/pinctrl-lpc18xx.c LPC18XX_PIN_P(b,6), b 429 drivers/pinctrl/pinctrl-rockchip.c struct rockchip_pin_bank *b = info->ctrl->pin_banks; b 431 drivers/pinctrl/pinctrl-rockchip.c while (pin >= (b->pin_base + b->nr_pins)) b 432 drivers/pinctrl/pinctrl-rockchip.c b++; b 434 drivers/pinctrl/pinctrl-rockchip.c return b; b 441 drivers/pinctrl/pinctrl-rockchip.c struct rockchip_pin_bank *b = info->ctrl->pin_banks; b 444 drivers/pinctrl/pinctrl-rockchip.c for (i = 0; i < info->ctrl->nr_banks; i++, b++) { b 445 drivers/pinctrl/pinctrl-rockchip.c if (b->bank_num == num) b 446 drivers/pinctrl/pinctrl-rockchip.c return b; b 948 drivers/pinctrl/pinctrl-sx150x.c unsigned int a, b; b 975 drivers/pinctrl/pinctrl-sx150x.c b = val & 0x0000ff00; b 978 drivers/pinctrl/pinctrl-sx150x.c val |= b << 8; b 227 drivers/pinctrl/samsung/pinctrl-exynos.c struct samsung_pin_bank *b = h->host_data; b 229 drivers/pinctrl/samsung/pinctrl-exynos.c irq_set_chip_data(virq, b); b 230 drivers/pinctrl/samsung/pinctrl-exynos.c irq_set_chip_and_handler(virq, &b->irq_chip->chip, b 454 drivers/pinctrl/samsung/pinctrl-exynos.c struct samsung_pin_bank *b = eintd->banks[i]; b 455 drivers/pinctrl/samsung/pinctrl-exynos.c pend = readl(b->eint_base + b->irq_chip->eint_pend b 456 drivers/pinctrl/samsung/pinctrl-exynos.c + b->eint_offset); b 457 drivers/pinctrl/samsung/pinctrl-exynos.c mask = readl(b->eint_base + b->irq_chip->eint_mask b 458 drivers/pinctrl/samsung/pinctrl-exynos.c + b->eint_offset); b 459 drivers/pinctrl/samsung/pinctrl-exynos.c exynos_irq_demux_eint(pend & ~mask, b->irq_domain); b 361 drivers/pinctrl/samsung/pinctrl-samsung.c struct samsung_pin_bank *b; b 363 drivers/pinctrl/samsung/pinctrl-samsung.c b = drvdata->pin_banks; b 365 drivers/pinctrl/samsung/pinctrl-samsung.c while ((pin >= b->pin_base) && b 366 drivers/pinctrl/samsung/pinctrl-samsung.c ((b->pin_base + b->nr_pins - 1) < pin)) b 367 drivers/pinctrl/samsung/pinctrl-samsung.c b++; b 369 drivers/pinctrl/samsung/pinctrl-samsung.c *reg = b->pctl_base + b->pctl_offset; b 370 drivers/pinctrl/samsung/pinctrl-samsung.c *offset = pin - b->pin_base; b 372 drivers/pinctrl/samsung/pinctrl-samsung.c *bank = b; b 245 drivers/pinctrl/sirf/pinctrl-atlas7.c #define MUX(b, pad, f, dr, db, dvr, dvb) \ b 247 drivers/pinctrl/sirf/pinctrl-atlas7.c .bank = b, \ b 324 drivers/pinctrl/sirf/pinctrl-atlas7.c #define ATLAS7_GPIO_BASE(g, b) ((g)->reg + 0x100 * (b)) b 325 drivers/pinctrl/sirf/pinctrl-atlas7.c #define ATLAS7_GPIO_CTRL(b, i) ((b)->base + 4 * (i)) b 326 drivers/pinctrl/sirf/pinctrl-atlas7.c #define ATLAS7_GPIO_INT_STATUS(b) ((b)->base + 0x8C) b 26 drivers/pinctrl/sprd/pinctrl-sprd.h #define SPRD_PINCTRL_PIN_DATA(a, b) \ b 28 drivers/pinctrl/sprd/pinctrl-sprd.h .name = b, \ b 1547 drivers/pinctrl/tegra/pinctrl-tegra114.c #define PINGROUP_BIT_Y(b) (b) b 1548 drivers/pinctrl/tegra/pinctrl-tegra114.c #define PINGROUP_BIT_N(b) (-1) b 1716 drivers/pinctrl/tegra/pinctrl-tegra124.c #define PINGROUP_BIT_Y(b) (b) b 1717 drivers/pinctrl/tegra/pinctrl-tegra124.c #define PINGROUP_BIT_N(b) (-1) b 1780 drivers/pinctrl/tegra/pinctrl-tegra124.c #define MIPI_PAD_CTRL_PINGROUP(pg_name, r, b, f0, f1) \ b 1793 drivers/pinctrl/tegra/pinctrl-tegra124.c .mux_bit = b, \ b 1275 drivers/pinctrl/tegra/pinctrl-tegra210.c #define PINGROUP_BIT_Y(b) (b) b 1276 drivers/pinctrl/tegra/pinctrl-tegra210.c #define PINGROUP_BIT_N(b) (-1) b 2108 drivers/pinctrl/tegra/pinctrl-tegra30.c #define PINGROUP_BIT_Y(b) (b) b 2109 drivers/pinctrl/tegra/pinctrl-tegra30.c #define PINGROUP_BIT_N(b) (-1) b 152 drivers/pinctrl/uniphier/pinctrl-uniphier.h #define UNIPHIER_PINCTRL_PIN(a, b, c, d, e, f, g) \ b 155 drivers/pinctrl/uniphier/pinctrl-uniphier.h .name = b, \ b 316 drivers/platform/x86/compal-laptop.c static int bl_get_brightness(struct backlight_device *b) b 321 drivers/platform/x86/compal-laptop.c static int bl_update_status(struct backlight_device *b) b 323 drivers/platform/x86/compal-laptop.c int ret = set_backlight_level(b->props.brightness); b 327 drivers/platform/x86/compal-laptop.c set_backlight_state((b->props.power == FB_BLANK_UNBLANK) b 328 drivers/platform/x86/compal-laptop.c && !(b->props.state & BL_CORE_SUSPENDED) b 329 drivers/platform/x86/compal-laptop.c && !(b->props.state & BL_CORE_FBBLANK)); b 244 drivers/platform/x86/fujitsu-laptop.c static int bl_get_brightness(struct backlight_device *b) b 246 drivers/platform/x86/fujitsu-laptop.c struct acpi_device *device = bl_get_data(b); b 248 drivers/platform/x86/fujitsu-laptop.c return b->props.power == FB_BLANK_POWERDOWN ? 0 : get_lcd_level(device); b 251 drivers/platform/x86/fujitsu-laptop.c static int bl_update_status(struct backlight_device *b) b 253 drivers/platform/x86/fujitsu-laptop.c struct acpi_device *device = bl_get_data(b); b 256 drivers/platform/x86/fujitsu-laptop.c if (b->props.power == FB_BLANK_POWERDOWN) b 264 drivers/platform/x86/fujitsu-laptop.c return set_lcd_level(device, b->props.brightness); b 206 drivers/platform/x86/intel_oaktrail.c static int get_backlight_brightness(struct backlight_device *b) b 214 drivers/platform/x86/intel_oaktrail.c static int set_backlight_brightness(struct backlight_device *b) b 216 drivers/platform/x86/intel_oaktrail.c u8 percent = (u8) b->props.brightness; b 287 drivers/platform/x86/msi-laptop.c static int bl_get_brightness(struct backlight_device *b) b 293 drivers/platform/x86/msi-laptop.c static int bl_update_status(struct backlight_device *b) b 295 drivers/platform/x86/msi-laptop.c return set_lcd_level(b->props.brightness); b 787 drivers/platform/x86/sony-laptop.c #define MIN(a, b) (a > b ? b : a) b 252 drivers/platform/x86/thinkpad_acpi.c #define strlencmp(a, b) (strncmp((a), (b), strlen(b))) b 6938 drivers/platform/x86/thinkpad_acpi.c unsigned int b; b 6951 drivers/platform/x86/thinkpad_acpi.c b = tpacpi_check_std_acpi_brightness_support(); b 6952 drivers/platform/x86/thinkpad_acpi.c switch (b) { b 6962 drivers/platform/x86/thinkpad_acpi.c bright_maxlvl = b - 1; b 6970 drivers/platform/x86/thinkpad_acpi.c int b; b 7031 drivers/platform/x86/thinkpad_acpi.c if (tpacpi_brightness_get_raw(&b) < 0) b 7037 drivers/platform/x86/thinkpad_acpi.c props.brightness = b & TP_EC_BACKLIGHT_LVLMSK; b 755 drivers/pnp/isapnp/core.c unsigned char checksum = 0x6a, bit, b; b 758 drivers/pnp/isapnp/core.c b = data[i]; b 761 drivers/pnp/isapnp/core.c if (b & (1 << j)) b 2234 drivers/power/supply/ab8500_fg.c const struct abx500_battery_type *b; b 2236 drivers/power/supply/ab8500_fg.c b = &(di->bm->bat_type[di->bm->batt_id]); b 2242 drivers/power/supply/ab8500_fg.c b->charge_full_design; b 2247 drivers/power/supply/ab8500_fg.c di->vbat_nom = b->nominal_voltage; b 32 drivers/power/supply/axp20x_usb_power.c #define AXP20X_VBUS_VHOLD_uV(b) (4000000 + (((b) >> 3) & 7) * 100000) b 91 drivers/power/supply/ipaq_micro_battery.c static int get_capacity(struct power_supply *b) b 93 drivers/power/supply/ipaq_micro_battery.c struct micro_battery *mb = dev_get_drvdata(b->dev.parent); b 111 drivers/power/supply/ipaq_micro_battery.c static int get_status(struct power_supply *b) b 113 drivers/power/supply/ipaq_micro_battery.c struct micro_battery *mb = dev_get_drvdata(b->dev.parent); b 128 drivers/power/supply/ipaq_micro_battery.c static int micro_batt_get_property(struct power_supply *b, b 132 drivers/power/supply/ipaq_micro_battery.c struct micro_battery *mb = dev_get_drvdata(b->dev.parent); b 155 drivers/power/supply/ipaq_micro_battery.c val->intval = get_status(b); b 161 drivers/power/supply/ipaq_micro_battery.c val->intval = get_capacity(b); b 176 drivers/power/supply/ipaq_micro_battery.c static int micro_ac_get_property(struct power_supply *b, b 180 drivers/power/supply/ipaq_micro_battery.c struct micro_battery *mb = dev_get_drvdata(b->dev.parent); b 186 drivers/power/supply/twl4030_madc_battery.c static int twl4030_cmp(const void *a, const void *b) b 188 drivers/power/supply/twl4030_madc_battery.c return ((struct twl4030_madc_bat_calibration *)b)->voltage - b 150 drivers/pps/generators/pps_gen_parport.c struct timespec64 a, b; b 156 drivers/pps/generators/pps_gen_parport.c ktime_get_real_ts64(&b); b 159 drivers/pps/generators/pps_gen_parport.c b = timespec64_sub(b, a); b 160 drivers/pps/generators/pps_gen_parport.c acc += timespec64_to_ns(&b); b 78 drivers/pwm/pwm-fsl-ftm.c const struct fsl_pwm_periodcfg *b) b 80 drivers/pwm/pwm-fsl-ftm.c if (a->clk_select != b->clk_select) b 82 drivers/pwm/pwm-fsl-ftm.c if (a->clk_ps != b->clk_ps) b 84 drivers/pwm/pwm-fsl-ftm.c if (a->mod_period != b->mod_period) b 1600 drivers/rapidio/devices/rio_mport_cdev.c struct rio_pw_filter *b) b 1602 drivers/rapidio/devices/rio_mport_cdev.c if ((a->mask == b->mask) && (a->low == b->low) && (a->high == b->high)) b 310 drivers/regulator/vctrl-regulator.c static int vctrl_cmp_ctrl_uV(const void *a, const void *b) b 313 drivers/regulator/vctrl-regulator.c const struct vctrl_voltage_table *bt = b; b 255 drivers/remoteproc/imx_rproc.c int a, b = 0, err, nph; b 264 drivers/remoteproc/imx_rproc.c if (b >= IMX7D_RPROC_MEM_MAX) b 267 drivers/remoteproc/imx_rproc.c priv->mem[b].cpu_addr = devm_ioremap(&pdev->dev, b 269 drivers/remoteproc/imx_rproc.c if (!priv->mem[b].cpu_addr) { b 273 drivers/remoteproc/imx_rproc.c priv->mem[b].sys_addr = att->sa; b 274 drivers/remoteproc/imx_rproc.c priv->mem[b].size = att->size; b 275 drivers/remoteproc/imx_rproc.c b++; b 295 drivers/remoteproc/imx_rproc.c if (b >= IMX7D_RPROC_MEM_MAX) b 298 drivers/remoteproc/imx_rproc.c priv->mem[b].cpu_addr = devm_ioremap_resource(&pdev->dev, &res); b 299 drivers/remoteproc/imx_rproc.c if (IS_ERR(priv->mem[b].cpu_addr)) { b 301 drivers/remoteproc/imx_rproc.c err = PTR_ERR(priv->mem[b].cpu_addr); b 304 drivers/remoteproc/imx_rproc.c priv->mem[b].sys_addr = res.start; b 305 drivers/remoteproc/imx_rproc.c priv->mem[b].size = resource_size(&res); b 306 drivers/remoteproc/imx_rproc.c b++; b 722 drivers/s390/cio/cio.c orb->tm.b = 1; b 749 drivers/s390/cio/cio.c if (!to_io_private(sch)->orb.tm.b) b 62 drivers/s390/cio/device_fsm.c if (orb->tm.b) { b 47 drivers/s390/cio/orb.h u32 b:1; b 600 drivers/s390/cio/qdio_main.c int j, b = start; b 603 drivers/s390/cio/qdio_main.c get_buf_state(q, b, &state, 0); b 605 drivers/s390/cio/qdio_main.c struct qaob *aob = q->u.out.aobs[b]; b 609 drivers/s390/cio/qdio_main.c q->u.out.sbal_state[b].flags |= b 611 drivers/s390/cio/qdio_main.c q->u.out.aobs[b] = NULL; b 613 drivers/s390/cio/qdio_main.c b = next_buf(b); b 259 drivers/s390/cio/vfio_ccw_fsm.c if (orb->tm.b) { b 905 drivers/s390/crypto/ap_bus.c int i, n, b; b 917 drivers/s390/crypto/ap_bus.c b = hex_to_bin(*str); b 919 drivers/s390/crypto/ap_bus.c if (b & (0x08 >> n)) b 461 drivers/scsi/3w-9xxx.h #define TW_PRINTK(h,a,b,c) { \ b 463 drivers/scsi/3w-9xxx.h printk(KERN_WARNING "3w-9xxx: scsi%d: ERROR: (0x%02X:0x%04X): %s.\n",h->host_no,a,b,c); \ b 465 drivers/scsi/3w-9xxx.h printk(KERN_WARNING "3w-9xxx: ERROR: (0x%02X:0x%04X): %s.\n",a,b,c); \ b 191 drivers/scsi/3w-sas.h #define TW_PRINTK(h,a,b,c) { \ b 193 drivers/scsi/3w-sas.h printk(KERN_WARNING "3w-sas: scsi%d: ERROR: (0x%02X:0x%04X): %s.\n",h->host_no,a,b,c); \ b 195 drivers/scsi/3w-sas.h printk(KERN_WARNING "3w-sas: ERROR: (0x%02X:0x%04X): %s.\n",a,b,c); \ b 845 drivers/scsi/aacraid/aachba.c static void inqstrcpy(char *a, char *b) b 849 drivers/scsi/aacraid/aachba.c *b++ = *a++; b 2551 drivers/scsi/advansys.c static void asc_prt_adv_sgblock(int sgblockno, ADV_SG_BLOCK *b) b 2556 drivers/scsi/advansys.c (ulong)b, sgblockno); b 2558 drivers/scsi/advansys.c b->sg_cnt, (u32)le32_to_cpu(b->sg_ptr)); b 2559 drivers/scsi/advansys.c BUG_ON(b->sg_cnt > NO_OF_SG_PER_BLOCK); b 2560 drivers/scsi/advansys.c if (b->sg_ptr != 0) b 2561 drivers/scsi/advansys.c BUG_ON(b->sg_cnt != NO_OF_SG_PER_BLOCK); b 2562 drivers/scsi/advansys.c for (i = 0; i < b->sg_cnt; i++) { b 2564 drivers/scsi/advansys.c i, (u32)le32_to_cpu(b->sg_list[i].sg_addr), b 2565 drivers/scsi/advansys.c (u32)le32_to_cpu(b->sg_list[i].sg_count)); b 69 drivers/scsi/aic94xx/aic94xx_reg.h ASD_READ_OCM(u8, byte, b); b 82 drivers/scsi/aic94xx/aic94xx_reg.h ASD_WRITE_OCM(u8, byte, b); b 427 drivers/scsi/aic94xx/aic94xx_scb.c struct asd_ascb *a, *b; b 442 drivers/scsi/aic94xx/aic94xx_scb.c list_for_each_entry_safe(a, b, &asd_ha->seq.pend_q, list) { b 468 drivers/scsi/aic94xx/aic94xx_scb.c list_for_each_entry_safe(a, b, &asd_ha->seq.pend_q, list) { b 595 drivers/scsi/aic94xx/aic94xx_task.c struct asd_ascb *b = a; b 597 drivers/scsi/aic94xx/aic94xx_task.c if (a == b) b 3691 drivers/scsi/arcmsr/arcmsr_hba.c } b; b 3707 drivers/scsi/arcmsr/arcmsr_hba.c writel(datetime.b.msg_time[0], ®->message_rwbuffer[0]); b 3708 drivers/scsi/arcmsr/arcmsr_hba.c writel(datetime.b.msg_time[1], ®->message_rwbuffer[1]); b 3716 drivers/scsi/arcmsr/arcmsr_hba.c writel(datetime.b.msg_time[0], rwbuffer++); b 3717 drivers/scsi/arcmsr/arcmsr_hba.c writel(datetime.b.msg_time[1], rwbuffer++); b 3723 drivers/scsi/arcmsr/arcmsr_hba.c writel(datetime.b.msg_time[0], ®->msgcode_rwbuffer[0]); b 3724 drivers/scsi/arcmsr/arcmsr_hba.c writel(datetime.b.msg_time[1], ®->msgcode_rwbuffer[1]); b 3733 drivers/scsi/arcmsr/arcmsr_hba.c writel(datetime.b.msg_time[0], rwbuffer++); b 3734 drivers/scsi/arcmsr/arcmsr_hba.c writel(datetime.b.msg_time[1], rwbuffer++); b 3740 drivers/scsi/arcmsr/arcmsr_hba.c writel(datetime.b.msg_time[0], ®->msgcode_rwbuffer[0]); b 3741 drivers/scsi/arcmsr/arcmsr_hba.c writel(datetime.b.msg_time[1], ®->msgcode_rwbuffer[1]); b 65 drivers/scsi/arm/oak.c unsigned long b; b 87 drivers/scsi/arm/oak.c b = (unsigned long) readw(base + DATA); b 88 drivers/scsi/arm/oak.c *addr ++ = b; b 91 drivers/scsi/arm/oak.c *addr ++ = b>>8; b 763 drivers/scsi/atari_scsi.c unsigned char b; b 765 drivers/scsi/atari_scsi.c ssize_t count = nvram_read(&b, 1, &offset); b 770 drivers/scsi/atari_scsi.c if ((count == 1) && (b & 0x80)) b 771 drivers/scsi/atari_scsi.c atari_scsi_template.this_id = b & 7; b 982 drivers/scsi/bfa/bfa_defs.h u8 b; b 1007 drivers/scsi/bfa/bfa_defs.h u8 b; b 1017 drivers/scsi/bfa/bfa_defs.h u8 b; b 1027 drivers/scsi/bfa/bfa_defs.h u8 b; b 1048 drivers/scsi/bfa/bfa_defs.h u8 b; b 1062 drivers/scsi/bfa/bfa_defs.h u8 b; b 3925 drivers/scsi/bfa/bfa_ioc.c e10g.b = sfpmem->srlid_base.xcvr[0]; b 3926 drivers/scsi/bfa/bfa_ioc.c bfa_trc(sfp, e10g.b); b 3968 drivers/scsi/bfa/bfa_ioc.c bfa_trc(sfp, e10g.b); b 3980 drivers/scsi/bfa/bfa_ioc.c bfa_trc(sfp, fc3.b); b 3981 drivers/scsi/bfa/bfa_ioc.c bfa_trc(sfp, e10g.b); b 1317 drivers/scsi/esas2r/esas2r_main.c struct atto_vda_buzzer_info *b; b 1324 drivers/scsi/esas2r/esas2r_main.c b = (struct atto_vda_buzzer_info *)data; b 1326 drivers/scsi/esas2r/esas2r_main.c b->duration = le32_to_cpu(b->duration); b 128 drivers/scsi/gdth.c static int gdth_fill_raw_cmd(gdth_ha_str *ha, struct scsi_cmnd *scp, u8 b); b 184 drivers/scsi/gdth.c #define GDTOFFSOF(a,b) (size_t)&(((a*)0)->b) b 187 drivers/scsi/gdth.c #define BUS_L2P(a,b) ((b)>(a)->virt_bus ? (b-1):(b)) b 1580 drivers/scsi/gdth.c u8 b, t, l, firsttime; b 1599 drivers/scsi/gdth.c b = nscp->device->channel; b 1603 drivers/scsi/gdth.c if ((b != ha->virt_bus && ha->raw[BUS_L2P(ha,b)].lock) || b 1604 drivers/scsi/gdth.c (b == ha->virt_bus && t < MAX_HDRIVES && ha->hdr[t].lock)) b 1608 drivers/scsi/gdth.c b = t = l = 0; b 1628 drivers/scsi/gdth.c b, t, l)); b 1631 drivers/scsi/gdth.c if (b == 0 && t == 0 && l == 0) { b 1636 drivers/scsi/gdth.c if (b == 0 && ((t == 0 && l == 1) || b 1649 drivers/scsi/gdth.c if (b == ha->bus_cnt && t == ha->tid_cnt-1) { b 1658 drivers/scsi/gdth.c if (b == ha->virt_bus && nscp->cmnd[0] != INQUIRY && b 1673 drivers/scsi/gdth.c if (!(cmd_index=gdth_fill_raw_cmd(ha, nscp, BUS_L2P(ha, b)))) b 1690 drivers/scsi/gdth.c } else if (b != ha->virt_bus) { b 1691 drivers/scsi/gdth.c if (ha->raw[BUS_L2P(ha,b)].io_cnt[t] >= GDTH_MAX_RAW || b 1692 drivers/scsi/gdth.c !(cmd_index=gdth_fill_raw_cmd(ha, nscp, BUS_L2P(ha, b)))) b 1695 drivers/scsi/gdth.c ha->raw[BUS_L2P(ha,b)].io_cnt[t]++; b 1698 drivers/scsi/gdth.c nscp->cmnd[0], b, t, l)); b 2137 drivers/scsi/gdth.c static int gdth_fill_raw_cmd(gdth_ha_str *ha, struct scsi_cmnd *scp, u8 b) b 2150 drivers/scsi/gdth.c scp->cmnd[0],b,t,l)); b 2198 drivers/scsi/gdth.c cmdp->u.raw64.bus = b; b 2214 drivers/scsi/gdth.c cmdp->u.raw.bus = b; b 2647 drivers/scsi/gdth.c u8 b, t; b 2723 drivers/scsi/gdth.c b = scp->device->channel; b 2725 drivers/scsi/gdth.c if (cmndinfo->OpCode == -1 && b != ha->virt_bus) { b 2726 drivers/scsi/gdth.c ha->raw[BUS_L2P(ha,b)].io_cnt[t]--; b 3092 drivers/scsi/gdth.c stack.b[j++] = *(u32*)&dvr->eu.stream[(int)f[i]]; b 3095 drivers/scsi/gdth.c stack.b[j++] = *(u16*)&dvr->eu.stream[(int)f[i]]; b 3098 drivers/scsi/gdth.c stack.b[j++] = *(u8*)&dvr->eu.stream[(int)f[i]]; b 3278 drivers/scsi/gdth.c u8 b, t; b 3283 drivers/scsi/gdth.c b = scp->device->channel; b 3296 drivers/scsi/gdth.c if ((b != ha->virt_bus && ha->raw[BUS_L2P(ha, b)].lock) || b 3297 drivers/scsi/gdth.c (b == ha->virt_bus && t < MAX_HDRIVES && ha->hdr[t].lock)) { b 3313 drivers/scsi/gdth.c u8 b; b 3317 drivers/scsi/gdth.c b = scp->device->channel; b 3323 drivers/scsi/gdth.c if (!SPECIAL_SCP(cmnd) && cmnd->device->channel == b) b 3328 drivers/scsi/gdth.c if (b == ha->virt_bus) { b 3347 drivers/scsi/gdth.c ha->raw[BUS_L2P(ha,b)].io_cnt[i] = 0; b 3352 drivers/scsi/gdth.c BUS_L2P(ha,b), 0, 0); b 3361 drivers/scsi/gdth.c u8 b, t; b 3368 drivers/scsi/gdth.c b = sd->channel; b 3370 drivers/scsi/gdth.c TRACE2(("gdth_bios_param() ha %d bus %d target %d\n", ha->hanum, b, t)); b 3372 drivers/scsi/gdth.c if (b != ha->virt_bus || ha->hdr[t].heads == 0) { b 972 drivers/scsi/gdth.h unsigned long b[10]; /* 32/64 bit compiler ! */ b 566 drivers/scsi/gdth_proc.c u8 b, t; b 574 drivers/scsi/gdth_proc.c b = scp->device->channel; b 577 drivers/scsi/gdth_proc.c b == (u8)busnum) { b 1444 drivers/scsi/hpsa.c #define SCSI3ADDR_EQ(a, b) ( \ b 1445 drivers/scsi/hpsa.c (a)[7] == (b)[7] && \ b 1446 drivers/scsi/hpsa.c (a)[6] == (b)[6] && \ b 1447 drivers/scsi/hpsa.c (a)[5] == (b)[5] && \ b 1448 drivers/scsi/hpsa.c (a)[4] == (b)[4] && \ b 1449 drivers/scsi/hpsa.c (a)[3] == (b)[3] && \ b 1450 drivers/scsi/hpsa.c (a)[2] == (b)[2] && \ b 1451 drivers/scsi/hpsa.c (a)[1] == (b)[1] && \ b 1452 drivers/scsi/hpsa.c (a)[0] == (b)[0]) b 9094 drivers/scsi/hpsa.c int i, j, b, size; b 9100 drivers/scsi/hpsa.c b = num_buckets; /* Assume the biggest bucket */ b 9104 drivers/scsi/hpsa.c b = j; b 9109 drivers/scsi/hpsa.c bucket_map[i] = b; b 452 drivers/scsi/imm.c static int imm_cpp(unsigned short ppb, unsigned char b) b 487 drivers/scsi/imm.c w_dtr(ppb, b); b 279 drivers/scsi/initio.c u8 b; b 288 drivers/scsi/initio.c b = SE2CS | SE2DO; /* -CLK+dataBit */ b 290 drivers/scsi/initio.c b = SE2CS; /* -CLK */ b 291 drivers/scsi/initio.c outb(b, base + TUL_NVRAM); b 293 drivers/scsi/initio.c outb(b | SE2CLK, base + TUL_NVRAM); /* +CLK */ b 729 drivers/scsi/libsas/sas_expander.c u8 a, b; b 731 drivers/scsi/libsas/sas_expander.c b = resp[j + 1]; b 734 drivers/scsi/libsas/sas_expander.c resp[j + 2] = b; b 78 drivers/scsi/libsas/sas_init.c int b; b 80 drivers/scsi/libsas/sas_init.c for (b = (SAS_ADDR_SIZE - 1); b >= 0; b--) { b 82 drivers/scsi/libsas/sas_init.c if ((1 << b) & sas_addr[i]) { b 3734 drivers/scsi/lpfc/lpfc_els.c switch (stat.un.b.lsRjtRsnCode) { b 3749 drivers/scsi/lpfc/lpfc_els.c if (stat.un.b.lsRjtRsnCodeExp == b 3758 drivers/scsi/lpfc/lpfc_els.c if (stat.un.b.lsRjtRsnCodeExp == b 3775 drivers/scsi/lpfc/lpfc_els.c (stat.un.b.lsRjtRsnCodeExp == LSEXP_OUT_OF_RESOURCE)){ b 3806 drivers/scsi/lpfc/lpfc_els.c stat.un.b.lsRjtRsnCodeExp == LSEXP_PORT_LOGIN_REQ) { b 3811 drivers/scsi/lpfc/lpfc_els.c stat.un.b.lsRjtRsnCodeExp == b 3825 drivers/scsi/lpfc/lpfc_els.c ((stat.un.b.lsRjtRsnCodeExp == LSEXP_INVALID_PNAME) || b 3826 drivers/scsi/lpfc/lpfc_els.c (stat.un.b.lsRjtRsnCodeExp == LSEXP_INVALID_NPORT_ID)) b 3837 drivers/scsi/lpfc/lpfc_els.c if ((stat.un.b.vendorUnique == 0x45) && b 3848 drivers/scsi/lpfc/lpfc_els.c if (stat.un.b.lsRjtRsnCodeExp == b 5822 drivers/scsi/lpfc/lpfc_els.c stat->un.b.lsRjtRsnCode = LSRJT_UNABLE_TPC; b 5953 drivers/scsi/lpfc/lpfc_els.c stat.un.b.lsRjtRsnCode = rjt_err; b 5954 drivers/scsi/lpfc/lpfc_els.c stat.un.b.lsRjtRsnCodeExp = rjt_expl; b 6051 drivers/scsi/lpfc/lpfc_els.c stat->un.b.lsRjtRsnCode = LSRJT_UNABLE_TPC; b 6054 drivers/scsi/lpfc/lpfc_els.c stat->un.b.lsRjtRsnCodeExp = LSEXP_CMD_IN_PROGRESS; b 6217 drivers/scsi/lpfc/lpfc_els.c stat.un.b.lsRjtRsnCode = rjt_err; b 6309 drivers/scsi/lpfc/lpfc_els.c switch (rscn_did.un.b.resv & RSCN_ADDRESS_FORMAT_MASK) { b 6311 drivers/scsi/lpfc/lpfc_els.c if ((ns_did.un.b.domain == rscn_did.un.b.domain) b 6312 drivers/scsi/lpfc/lpfc_els.c && (ns_did.un.b.area == rscn_did.un.b.area) b 6313 drivers/scsi/lpfc/lpfc_els.c && (ns_did.un.b.id == rscn_did.un.b.id)) b 6317 drivers/scsi/lpfc/lpfc_els.c if ((ns_did.un.b.domain == rscn_did.un.b.domain) b 6318 drivers/scsi/lpfc/lpfc_els.c && (ns_did.un.b.area == rscn_did.un.b.area)) b 6322 drivers/scsi/lpfc/lpfc_els.c if (ns_did.un.b.domain == rscn_did.un.b.domain) b 6952 drivers/scsi/lpfc/lpfc_els.c stat.un.b.lsRjtRsvd0 = 0; b 6953 drivers/scsi/lpfc/lpfc_els.c stat.un.b.lsRjtRsnCode = LSRJT_UNABLE_TPC; b 6954 drivers/scsi/lpfc/lpfc_els.c stat.un.b.lsRjtRsnCodeExp = LSEXP_CANT_GIVE_DATA; b 6955 drivers/scsi/lpfc/lpfc_els.c stat.un.b.vendorUnique = 0; b 7006 drivers/scsi/lpfc/lpfc_els.c stat.un.b.lsRjtRsvd0 = 0; b 7007 drivers/scsi/lpfc/lpfc_els.c stat.un.b.lsRjtRsnCode = LSRJT_UNABLE_TPC; b 7008 drivers/scsi/lpfc/lpfc_els.c stat.un.b.lsRjtRsnCodeExp = LSEXP_CANT_GIVE_DATA; b 7009 drivers/scsi/lpfc/lpfc_els.c stat.un.b.vendorUnique = 0; b 7274 drivers/scsi/lpfc/lpfc_els.c stat.un.b.lsRjtRsvd0 = 0; b 7275 drivers/scsi/lpfc/lpfc_els.c stat.un.b.lsRjtRsnCode = LSRJT_UNABLE_TPC; b 7276 drivers/scsi/lpfc/lpfc_els.c stat.un.b.lsRjtRsnCodeExp = LSEXP_CANT_GIVE_DATA; b 7277 drivers/scsi/lpfc/lpfc_els.c stat.un.b.vendorUnique = 0; b 7362 drivers/scsi/lpfc/lpfc_els.c stat.un.b.lsRjtRsvd0 = 0; b 7363 drivers/scsi/lpfc/lpfc_els.c stat.un.b.lsRjtRsnCode = LSRJT_UNABLE_TPC; b 7364 drivers/scsi/lpfc/lpfc_els.c stat.un.b.lsRjtRsnCodeExp = LSEXP_CANT_GIVE_DATA; b 7365 drivers/scsi/lpfc/lpfc_els.c stat.un.b.vendorUnique = 0; b 7439 drivers/scsi/lpfc/lpfc_els.c stat.un.b.lsRjtRsvd0 = 0; b 7440 drivers/scsi/lpfc/lpfc_els.c stat.un.b.lsRjtRsnCode = LSRJT_UNABLE_TPC; b 7441 drivers/scsi/lpfc/lpfc_els.c stat.un.b.lsRjtRsnCodeExp = LSEXP_CANT_GIVE_DATA; b 7442 drivers/scsi/lpfc/lpfc_els.c stat.un.b.vendorUnique = 0; b 7641 drivers/scsi/lpfc/lpfc_els.c stat.un.b.lsRjtRsvd0 = 0; b 7642 drivers/scsi/lpfc/lpfc_els.c stat.un.b.lsRjtRsnCode = LSRJT_UNABLE_TPC; b 7643 drivers/scsi/lpfc/lpfc_els.c stat.un.b.lsRjtRsnCodeExp = LSEXP_CANT_GIVE_DATA; b 7644 drivers/scsi/lpfc/lpfc_els.c stat.un.b.vendorUnique = 0; b 8192 drivers/scsi/lpfc/lpfc_els.c lsrjt_event.reason_code = stat.un.b.lsRjtRsnCode; b 8193 drivers/scsi/lpfc/lpfc_els.c lsrjt_event.explanation = stat.un.b.lsRjtRsnCodeExp; b 8718 drivers/scsi/lpfc/lpfc_els.c stat.un.b.lsRjtRsnCode = rjt_err; b 8719 drivers/scsi/lpfc/lpfc_els.c stat.un.b.lsRjtRsnCodeExp = rjt_exp; b 9724 drivers/scsi/lpfc/lpfc_els.c if ((stat.un.b.lsRjtRsnCode == LSRJT_UNABLE_TPC) || b 9725 drivers/scsi/lpfc/lpfc_els.c (stat.un.b.lsRjtRsnCode == LSRJT_LOGICAL_BSY)) b 5281 drivers/scsi/lpfc/lpfc_hbadisc.c if ((mydid.un.b.domain == 0) && (mydid.un.b.area == 0)) { b 5287 drivers/scsi/lpfc/lpfc_hbadisc.c if (matchdid.un.b.id == ndlpdid.un.b.id) { b 5288 drivers/scsi/lpfc/lpfc_hbadisc.c if ((mydid.un.b.domain == matchdid.un.b.domain) && b 5289 drivers/scsi/lpfc/lpfc_hbadisc.c (mydid.un.b.area == matchdid.un.b.area)) { b 5298 drivers/scsi/lpfc/lpfc_hbadisc.c if ((ndlpdid.un.b.domain == 0) && b 5299 drivers/scsi/lpfc/lpfc_hbadisc.c (ndlpdid.un.b.area == 0)) { b 5300 drivers/scsi/lpfc/lpfc_hbadisc.c if (ndlpdid.un.b.id && b 5309 drivers/scsi/lpfc/lpfc_hbadisc.c if ((mydid.un.b.domain == ndlpdid.un.b.domain) && b 5310 drivers/scsi/lpfc/lpfc_hbadisc.c (mydid.un.b.area == ndlpdid.un.b.area)) { b 5311 drivers/scsi/lpfc/lpfc_hbadisc.c if ((matchdid.un.b.domain == 0) && b 5312 drivers/scsi/lpfc/lpfc_hbadisc.c (matchdid.un.b.area == 0)) { b 5313 drivers/scsi/lpfc/lpfc_hbadisc.c if (matchdid.un.b.id) b 701 drivers/scsi/lpfc/lpfc_hw.h } b; b 717 drivers/scsi/lpfc/lpfc_hw.h } b; b 1034 drivers/scsi/lpfc/lpfc_hw.h } b; b 2685 drivers/scsi/lpfc/lpfc_hw.h } b; b 306 drivers/scsi/lpfc/lpfc_nportdisc.c stat.un.b.lsRjtRsnCode = LSRJT_UNABLE_TPC; b 307 drivers/scsi/lpfc/lpfc_nportdisc.c stat.un.b.lsRjtRsnCodeExp = LSEXP_INVALID_PNAME; b 315 drivers/scsi/lpfc/lpfc_nportdisc.c stat.un.b.lsRjtRsnCode = LSRJT_UNABLE_TPC; b 316 drivers/scsi/lpfc/lpfc_nportdisc.c stat.un.b.lsRjtRsnCodeExp = LSEXP_INVALID_NNAME; b 325 drivers/scsi/lpfc/lpfc_nportdisc.c stat.un.b.lsRjtRsnCode = LSRJT_UNABLE_TPC; b 326 drivers/scsi/lpfc/lpfc_nportdisc.c stat.un.b.lsRjtRsnCodeExp = LSEXP_SPARM_OPTIONS; b 506 drivers/scsi/lpfc/lpfc_nportdisc.c stat.un.b.lsRjtRsnCode = LSRJT_INVALID_CMD; b 507 drivers/scsi/lpfc/lpfc_nportdisc.c stat.un.b.lsRjtRsnCodeExp = LSEXP_NOTHING_MORE; b 519 drivers/scsi/lpfc/lpfc_nportdisc.c stat.un.b.lsRjtRsnCode = LSRJT_UNABLE_TPC; b 520 drivers/scsi/lpfc/lpfc_nportdisc.c stat.un.b.lsRjtRsnCodeExp = LSEXP_OUT_OF_RESOURCE; b 627 drivers/scsi/lpfc/lpfc_nportdisc.c stat.un.b.lsRjtRsvd0 = 0; b 628 drivers/scsi/lpfc/lpfc_nportdisc.c stat.un.b.lsRjtRsnCode = LSRJT_UNABLE_TPC; b 629 drivers/scsi/lpfc/lpfc_nportdisc.c stat.un.b.lsRjtRsnCodeExp = LSEXP_SPARM_OPTIONS; b 630 drivers/scsi/lpfc/lpfc_nportdisc.c stat.un.b.vendorUnique = 0; b 766 drivers/scsi/lpfc/lpfc_nportdisc.c stat.un.b.lsRjtRsnCode = LSRJT_CMD_UNSUPPORTED; b 767 drivers/scsi/lpfc/lpfc_nportdisc.c stat.un.b.lsRjtRsnCodeExp = LSEXP_REQ_UNSUPPORTED; b 1058 drivers/scsi/lpfc/lpfc_nportdisc.c stat.un.b.lsRjtRsnCode = LSRJT_UNABLE_TPC; b 1059 drivers/scsi/lpfc/lpfc_nportdisc.c stat.un.b.lsRjtRsnCodeExp = LSEXP_CMD_IN_PROGRESS; b 1092 drivers/scsi/lpfc/lpfc_nportdisc.c stat.un.b.lsRjtRsnCode = LSRJT_LOGICAL_BSY; b 1093 drivers/scsi/lpfc/lpfc_nportdisc.c stat.un.b.lsRjtRsnCodeExp = LSEXP_NOTHING_MORE; b 1648 drivers/scsi/lpfc/lpfc_nportdisc.c stat.un.b.lsRjtRsnCode = LSRJT_LOGICAL_BSY; b 1649 drivers/scsi/lpfc/lpfc_nportdisc.c stat.un.b.lsRjtRsnCodeExp = LSEXP_NOTHING_MORE; b 2211 drivers/scsi/lpfc/lpfc_nportdisc.c stat.un.b.lsRjtRsnCode = LSRJT_UNABLE_TPC; b 2212 drivers/scsi/lpfc/lpfc_nportdisc.c stat.un.b.lsRjtRsnCodeExp = LSEXP_NOTHING_MORE; b 2225 drivers/scsi/lpfc/lpfc_nportdisc.c stat.un.b.lsRjtRsnCode = LSRJT_UNABLE_TPC; b 2226 drivers/scsi/lpfc/lpfc_nportdisc.c stat.un.b.lsRjtRsnCodeExp = LSEXP_NOTHING_MORE; b 2253 drivers/scsi/lpfc/lpfc_nportdisc.c stat.un.b.lsRjtRsnCode = LSRJT_UNABLE_TPC; b 2254 drivers/scsi/lpfc/lpfc_nportdisc.c stat.un.b.lsRjtRsnCodeExp = LSEXP_NOTHING_MORE; b 2267 drivers/scsi/lpfc/lpfc_nportdisc.c stat.un.b.lsRjtRsnCode = LSRJT_UNABLE_TPC; b 2268 drivers/scsi/lpfc/lpfc_nportdisc.c stat.un.b.lsRjtRsnCodeExp = LSEXP_NOTHING_MORE; b 2502 drivers/scsi/lpfc/lpfc_nportdisc.c stat.un.b.lsRjtRsnCode = LSRJT_UNABLE_TPC; b 2503 drivers/scsi/lpfc/lpfc_nportdisc.c stat.un.b.lsRjtRsnCodeExp = LSEXP_NOTHING_MORE; b 122 drivers/scsi/lpfc/lpfc_sli.c #define lpfc_sli4_pcimem_bcopy(a, b, c) lpfc_sli_pcimem_bcopy(a, b, c) b 1912 drivers/scsi/megaraid/megaraid_sas.h u8 b[12]; b 2190 drivers/scsi/megaraid/megaraid_sas.h u8 b[96]; b 2335 drivers/scsi/megaraid/megaraid_sas_base.c memset(dcmd->mbox.b, 0, MFI_MBOX_SIZE); b 2442 drivers/scsi/megaraid/megaraid_sas_base.c memset(dcmd->mbox.b, 0, MFI_MBOX_SIZE); b 2608 drivers/scsi/megaraid/megaraid_sas_base.c memset(dcmd->mbox.b, 0, MFI_MBOX_SIZE); b 3588 drivers/scsi/megaraid/megaraid_sas_base.c && (cmd->frame->dcmd.mbox.b[1] == 1)) { b 3635 drivers/scsi/megaraid/megaraid_sas_base.c (cmd->frame->dcmd.mbox.b[0] == 1)) { b 4423 drivers/scsi/megaraid/megaraid_sas_base.c memset(dcmd->mbox.b, 0, MFI_MBOX_SIZE); b 4515 drivers/scsi/megaraid/megaraid_sas_base.c memset(dcmd->mbox.b, 0, MFI_MBOX_SIZE); b 4517 drivers/scsi/megaraid/megaraid_sas_base.c dcmd->mbox.b[0] = MR_PD_QUERY_TYPE_EXPOSED_TO_HOST; b 4518 drivers/scsi/megaraid/megaraid_sas_base.c dcmd->mbox.b[1] = 0; b 4646 drivers/scsi/megaraid/megaraid_sas_base.c memset(dcmd->mbox.b, 0, MFI_MBOX_SIZE); b 4649 drivers/scsi/megaraid/megaraid_sas_base.c dcmd->mbox.b[0] = 1; b 4764 drivers/scsi/megaraid/megaraid_sas_base.c memset(dcmd->mbox.b, 0, MFI_MBOX_SIZE); b 4766 drivers/scsi/megaraid/megaraid_sas_base.c dcmd->mbox.b[0] = query_type; b 4768 drivers/scsi/megaraid/megaraid_sas_base.c dcmd->mbox.b[2] = 1; b 4886 drivers/scsi/megaraid/megaraid_sas_base.c memset(dcmd->mbox.b, 0, MFI_MBOX_SIZE); b 4888 drivers/scsi/megaraid/megaraid_sas_base.c dcmd->mbox.b[0] = is_probe ? 0 : 1; b 5072 drivers/scsi/megaraid/megaraid_sas_base.c memset(dcmd->mbox.b, 0, MFI_MBOX_SIZE); b 5154 drivers/scsi/megaraid/megaraid_sas_base.c memset(dcmd->mbox.b, 0, MFI_MBOX_SIZE); b 5164 drivers/scsi/megaraid/megaraid_sas_base.c dcmd->mbox.b[0] = 1; b 5303 drivers/scsi/megaraid/megaraid_sas_base.c memset(dcmd->mbox.b, 0, MFI_MBOX_SIZE); b 5304 drivers/scsi/megaraid/megaraid_sas_base.c dcmd->mbox.b[0] = crash_buf_state; b 6437 drivers/scsi/megaraid/megaraid_sas_base.c memset(dcmd->mbox.b, 0, MFI_MBOX_SIZE); b 6573 drivers/scsi/megaraid/megaraid_sas_base.c memset(dcmd->mbox.b, 0, MFI_MBOX_SIZE); b 6649 drivers/scsi/megaraid/megaraid_sas_base.c memset(dcmd->mbox.b, 0, MFI_MBOX_SIZE); b 6650 drivers/scsi/megaraid/megaraid_sas_base.c dcmd->mbox.b[0] = MEGASAS_IS_LOGICAL(sdev); b 7452 drivers/scsi/megaraid/megaraid_sas_base.c memset(dcmd->mbox.b, 0, MFI_MBOX_SIZE); b 7462 drivers/scsi/megaraid/megaraid_sas_base.c dcmd->mbox.b[0] = MR_FLUSH_CTRL_CACHE | MR_FLUSH_DISK_CACHE; b 7505 drivers/scsi/megaraid/megaraid_sas_base.c memset(dcmd->mbox.b, 0, MFI_MBOX_SIZE); b 54 drivers/scsi/megaraid/megaraid_sas_fp.c #define ABS_DIFF(a, b) (((a) > (b)) ? ((a) - (b)) : ((b) - (a))) b 1276 drivers/scsi/megaraid/megaraid_sas_fusion.c memset(dcmd->mbox.b, 0, MFI_MBOX_SIZE); b 1279 drivers/scsi/megaraid/megaraid_sas_fusion.c dcmd->mbox.b[0] = MEGASAS_DCMD_MBOX_PEND_FLAG; b 1379 drivers/scsi/megaraid/megaraid_sas_fusion.c memset(dcmd->mbox.b, 0, MFI_MBOX_SIZE); b 1463 drivers/scsi/megaraid/megaraid_sas_fusion.c memset(dcmd->mbox.b, 0, MFI_MBOX_SIZE); b 1488 drivers/scsi/megaraid/megaraid_sas_fusion.c dcmd->mbox.b[0] = num_lds; b 1489 drivers/scsi/megaraid/megaraid_sas_fusion.c dcmd->mbox.b[1] = MEGASAS_DCMD_MBOX_PEND_FLAG; b 300 drivers/scsi/mesh.c #define MKWORD(a, b, c, d) (((a) << 24) + ((b) << 16) + ((c) << 8) + (d)) b 608 drivers/scsi/mesh.c unsigned char *b = cmd->request_buffer; b 610 drivers/scsi/mesh.c b[0], b[1], b[2], b[3], b[4], b[5], b[6], b[7]); b 840 drivers/scsi/mesh.c int b, n; b 844 drivers/scsi/mesh.c b = ms->msgin[0]; b 845 drivers/scsi/mesh.c if (b == 1) { b 848 drivers/scsi/mesh.c } else if (0x20 <= b && b <= 0x2f) { b 861 drivers/scsi/mesh.c int b, t, prev; b 934 drivers/scsi/mesh.c b = in_8(&mr->fifo); b 935 drivers/scsi/mesh.c dlog(ms, "reseldata %x", b); b 938 drivers/scsi/mesh.c if ((b & (1 << t)) != 0 && t != ms->host->this_id) b 940 drivers/scsi/mesh.c if (b != (1 << t) + (1 << ms->host->this_id)) { b 941 drivers/scsi/mesh.c printk(KERN_ERR "mesh: bad reselection data %x\n", b); b 3663 drivers/scsi/mpt3sas/mpt3sas_base.c _base_mpi_ep_writeq(__u64 b, volatile void __iomem *addr, b 3669 drivers/scsi/mpt3sas/mpt3sas_base.c __raw_writel((u32)(b), addr); b 3670 drivers/scsi/mpt3sas/mpt3sas_base.c __raw_writel((u32)(b >> 32), (addr + 4)); b 3686 drivers/scsi/mpt3sas/mpt3sas_base.c _base_writeq(__u64 b, volatile void __iomem *addr, spinlock_t *writeq_lock) b 3689 drivers/scsi/mpt3sas/mpt3sas_base.c __raw_writeq(b, addr); b 3694 drivers/scsi/mpt3sas/mpt3sas_base.c _base_writeq(__u64 b, volatile void __iomem *addr, spinlock_t *writeq_lock) b 3696 drivers/scsi/mpt3sas/mpt3sas_base.c _base_mpi_ep_writeq(b, addr, writeq_lock); b 4348 drivers/scsi/mpt3sas/mpt3sas_scsih.c _scsih_set_volume_handle_for_tr(u16 handle, u16 *a, u16 *b) b 4350 drivers/scsi/mpt3sas/mpt3sas_scsih.c if (!handle || handle == *a || handle == *b) b 4354 drivers/scsi/mpt3sas/mpt3sas_scsih.c else if (!*b) b 4355 drivers/scsi/mpt3sas/mpt3sas_scsih.c *b = handle; b 4376 drivers/scsi/mpt3sas/mpt3sas_scsih.c u16 handle, volume_handle, a, b; b 4380 drivers/scsi/mpt3sas/mpt3sas_scsih.c b = 0; b 4397 drivers/scsi/mpt3sas/mpt3sas_scsih.c _scsih_set_volume_handle_for_tr(volume_handle, &a, &b); b 4409 drivers/scsi/mpt3sas/mpt3sas_scsih.c _scsih_set_volume_handle_for_tr(volume_handle, &a, &b); b 4415 drivers/scsi/mpt3sas/mpt3sas_scsih.c if (b) b 4416 drivers/scsi/mpt3sas/mpt3sas_scsih.c _scsih_tm_tr_volume_send(ioc, b); b 4428 drivers/scsi/mpt3sas/mpt3sas_scsih.c else if (volume_handle == a || volume_handle == b) { b 338 drivers/scsi/mvumi.h #define HS_SET_STATE(a, b) (a |= (b & 0xFFFF)) b 339 drivers/scsi/mvumi.h #define HS_SET_STATUS(a, b) (a |= ((b & 0xFFFF) << 16)) b 270 drivers/scsi/ncr53c8xx.c m_addr_t a, b; b 294 drivers/scsi/ncr53c8xx.c b = a ^ s; b 296 drivers/scsi/ncr53c8xx.c while (q->next && q->next != (m_link_s *) b) { b 305 drivers/scsi/ncr53c8xx.c a = a & b; b 377 drivers/scsi/ppa.c static inline void ppa_d_pulse(unsigned short ppb, unsigned char b) b 379 drivers/scsi/ppa.c w_dtr(ppb, b); b 397 drivers/scsi/ppa.c static inline void ppa_c_pulse(unsigned short ppb, unsigned char b) b 399 drivers/scsi/ppa.c w_dtr(ppb, b); b 568 drivers/scsi/qla1280.c #define qla1280_dump_buffer(a, b, c) do{}while(0) b 569 drivers/scsi/qla1280.c #define qla1280_print_scsi_cmd(a, b) do{}while(0) b 3926 drivers/scsi/qla1280.c __qla1280_dump_buffer(char *b, int size) b 3937 drivers/scsi/qla1280.c c = *b++; b 820 drivers/scsi/qla2xxx/qla_attr.c did.b.domain = (type & 0x00ff0000) >> 16; b 821 drivers/scsi/qla2xxx/qla_attr.c did.b.area = (type & 0x0000ff00) >> 8; b 822 drivers/scsi/qla2xxx/qla_attr.c did.b.al_pa = (type & 0x000000ff); b 825 drivers/scsi/qla2xxx/qla_attr.c did.b.domain, did.b.area, did.b.al_pa); b 2381 drivers/scsi/qla2xxx/qla_attr.c fc_host_port_id(shost) = vha->d_id.b.domain << 16 | b 2382 drivers/scsi/qla2xxx/qla_attr.c vha->d_id.b.area << 8 | vha->d_id.b.al_pa; b 2509 drivers/scsi/qla2xxx/qla_attr.c port_id = fcport->d_id.b.domain << 16 | b 2510 drivers/scsi/qla2xxx/qla_attr.c fcport->d_id.b.area << 8 | fcport->d_id.b.al_pa; b 2581 drivers/scsi/qla2xxx/qla_attr.c fcport->loop_id, fcport->d_id.b.domain, b 2582 drivers/scsi/qla2xxx/qla_attr.c fcport->d_id.b.area, fcport->d_id.b.al_pa); b 325 drivers/scsi/qla2xxx/qla_bsg.c fcport->d_id.b.al_pa = b 327 drivers/scsi/qla2xxx/qla_bsg.c fcport->d_id.b.area = b 329 drivers/scsi/qla2xxx/qla_bsg.c fcport->d_id.b.domain = b 332 drivers/scsi/qla2xxx/qla_bsg.c (fcport->d_id.b.al_pa == 0xFD) ? b 387 drivers/scsi/qla2xxx/qla_bsg.c fcport->d_id.b.domain, fcport->d_id.b.area, fcport->d_id.b.al_pa); b 509 drivers/scsi/qla2xxx/qla_bsg.c fcport->d_id.b.al_pa = bsg_request->rqst_data.h_ct.port_id[0]; b 510 drivers/scsi/qla2xxx/qla_bsg.c fcport->d_id.b.area = bsg_request->rqst_data.h_ct.port_id[1]; b 511 drivers/scsi/qla2xxx/qla_bsg.c fcport->d_id.b.domain = bsg_request->rqst_data.h_ct.port_id[2]; b 534 drivers/scsi/qla2xxx/qla_bsg.c fcport->loop_id, fcport->d_id.b.domain, fcport->d_id.b.area, b 535 drivers/scsi/qla2xxx/qla_bsg.c fcport->d_id.b.al_pa); b 1835 drivers/scsi/qla2xxx/qla_bsg.c vha->bidir_fcport.d_id.b.al_pa = vha->d_id.b.al_pa; b 1836 drivers/scsi/qla2xxx/qla_bsg.c vha->bidir_fcport.d_id.b.area = vha->d_id.b.area; b 1837 drivers/scsi/qla2xxx/qla_bsg.c vha->bidir_fcport.d_id.b.domain = vha->d_id.b.domain; b 223 drivers/scsi/qla2xxx/qla_dbg.h #define fce_calc_size(b) ((FCE_BYTES_PER_BUFFER) * (b)) b 357 drivers/scsi/qla2xxx/qla_def.h } b; b 387 drivers/scsi/qla2xxx/qla_def.h res.b.domain = id.domain; b 388 drivers/scsi/qla2xxx/qla_def.h res.b.area = id.area; b 389 drivers/scsi/qla2xxx/qla_def.h res.b.al_pa = id.al_pa; b 390 drivers/scsi/qla2xxx/qla_def.h res.b.rsvd_1 = 0; b 399 drivers/scsi/qla2xxx/qla_def.h res.domain = port_id.b.domain; b 400 drivers/scsi/qla2xxx/qla_def.h res.area = port_id.b.area; b 401 drivers/scsi/qla2xxx/qla_def.h res.al_pa = port_id.b.al_pa; b 31 drivers/scsi/qla2xxx/qla_dfs.c sess->d_id.b.domain, sess->d_id.b.area, b 32 drivers/scsi/qla2xxx/qla_dfs.c sess->d_id.b.al_pa, sess->port_name, b 96 drivers/scsi/qla2xxx/qla_dfs.c fc_port.port_name, fc_port.d_id.b.domain, b 97 drivers/scsi/qla2xxx/qla_dfs.c fc_port.d_id.b.area, fc_port.d_id.b.al_pa, b 131 drivers/scsi/qla2xxx/qla_gs.c routine, ms_pkt->entry_status, vha->d_id.b.domain, b 132 drivers/scsi/qla2xxx/qla_gs.c vha->d_id.b.area, vha->d_id.b.al_pa); b 147 drivers/scsi/qla2xxx/qla_gs.c routine, vha->d_id.b.domain, b 148 drivers/scsi/qla2xxx/qla_gs.c vha->d_id.b.area, vha->d_id.b.al_pa, b 183 drivers/scsi/qla2xxx/qla_gs.c vha->d_id.b.domain, vha->d_id.b.area, b 184 drivers/scsi/qla2xxx/qla_gs.c vha->d_id.b.al_pa); b 255 drivers/scsi/qla2xxx/qla_gs.c fcport->d_id.b.domain = 0xf0; b 261 drivers/scsi/qla2xxx/qla_gs.c fcport->d_id.b.domain, fcport->d_id.b.area, b 262 drivers/scsi/qla2xxx/qla_gs.c fcport->d_id.b.al_pa); b 342 drivers/scsi/qla2xxx/qla_gs.c list[i].d_id.b.rsvd_1 = gid_data->control_byte; b 421 drivers/scsi/qla2xxx/qla_gs.c if (list[i].d_id.b.rsvd_1 != 0) b 490 drivers/scsi/qla2xxx/qla_gs.c list[i].d_id.b.domain, list[i].d_id.b.area, b 491 drivers/scsi/qla2xxx/qla_gs.c list[i].d_id.b.al_pa); b 495 drivers/scsi/qla2xxx/qla_gs.c if (list[i].d_id.b.rsvd_1 != 0) b 1015 drivers/scsi/qla2xxx/qla_gs.c sns_cmd->p.cmd.param[0] = fcport->d_id.b.al_pa; b 1016 drivers/scsi/qla2xxx/qla_gs.c sns_cmd->p.cmd.param[1] = fcport->d_id.b.area; b 1017 drivers/scsi/qla2xxx/qla_gs.c sns_cmd->p.cmd.param[2] = fcport->d_id.b.domain; b 1035 drivers/scsi/qla2xxx/qla_gs.c fcport->d_id.b.domain = sns_cmd->p.gan_data[17]; b 1036 drivers/scsi/qla2xxx/qla_gs.c fcport->d_id.b.area = sns_cmd->p.gan_data[18]; b 1037 drivers/scsi/qla2xxx/qla_gs.c fcport->d_id.b.al_pa = sns_cmd->p.gan_data[19]; b 1044 drivers/scsi/qla2xxx/qla_gs.c fcport->d_id.b.domain = 0xf0; b 1050 drivers/scsi/qla2xxx/qla_gs.c fcport->d_id.b.domain, fcport->d_id.b.area, b 1051 drivers/scsi/qla2xxx/qla_gs.c fcport->d_id.b.al_pa); b 1106 drivers/scsi/qla2xxx/qla_gs.c list[i].d_id.b.domain = entry[1]; b 1107 drivers/scsi/qla2xxx/qla_gs.c list[i].d_id.b.area = entry[2]; b 1108 drivers/scsi/qla2xxx/qla_gs.c list[i].d_id.b.al_pa = entry[3]; b 1112 drivers/scsi/qla2xxx/qla_gs.c list[i].d_id.b.rsvd_1 = entry[0]; b 1154 drivers/scsi/qla2xxx/qla_gs.c sns_cmd->p.cmd.param[0] = list[i].d_id.b.al_pa; b 1155 drivers/scsi/qla2xxx/qla_gs.c sns_cmd->p.cmd.param[1] = list[i].d_id.b.area; b 1156 drivers/scsi/qla2xxx/qla_gs.c sns_cmd->p.cmd.param[2] = list[i].d_id.b.domain; b 1179 drivers/scsi/qla2xxx/qla_gs.c if (list[i].d_id.b.rsvd_1 != 0) b 1210 drivers/scsi/qla2xxx/qla_gs.c sns_cmd->p.cmd.param[0] = list[i].d_id.b.al_pa; b 1211 drivers/scsi/qla2xxx/qla_gs.c sns_cmd->p.cmd.param[1] = list[i].d_id.b.area; b 1212 drivers/scsi/qla2xxx/qla_gs.c sns_cmd->p.cmd.param[2] = list[i].d_id.b.domain; b 1237 drivers/scsi/qla2xxx/qla_gs.c list[i].d_id.b.domain, list[i].d_id.b.area, b 1238 drivers/scsi/qla2xxx/qla_gs.c list[i].d_id.b.al_pa); b 1242 drivers/scsi/qla2xxx/qla_gs.c if (list[i].d_id.b.rsvd_1 != 0) b 1270 drivers/scsi/qla2xxx/qla_gs.c sns_cmd->p.cmd.param[0] = vha->d_id.b.al_pa; b 1271 drivers/scsi/qla2xxx/qla_gs.c sns_cmd->p.cmd.param[1] = vha->d_id.b.area; b 1272 drivers/scsi/qla2xxx/qla_gs.c sns_cmd->p.cmd.param[2] = vha->d_id.b.domain; b 1319 drivers/scsi/qla2xxx/qla_gs.c sns_cmd->p.cmd.param[0] = vha->d_id.b.al_pa; b 1320 drivers/scsi/qla2xxx/qla_gs.c sns_cmd->p.cmd.param[1] = vha->d_id.b.area; b 1321 drivers/scsi/qla2xxx/qla_gs.c sns_cmd->p.cmd.param[2] = vha->d_id.b.domain; b 2735 drivers/scsi/qla2xxx/qla_gs.c if (list[i].d_id.b.rsvd_1 != 0) b 2866 drivers/scsi/qla2xxx/qla_gs.c if (list[i].d_id.b.rsvd_1 != 0) b 2947 drivers/scsi/qla2xxx/qla_gs.c if (list[i].d_id.b.rsvd_1 != 0) b 3085 drivers/scsi/qla2xxx/qla_gs.c fcport->loop_id, fcport->d_id.b.domain, b 3086 drivers/scsi/qla2xxx/qla_gs.c fcport->d_id.b.area, fcport->d_id.b.al_pa); b 3502 drivers/scsi/qla2xxx/qla_gs.c ct_req->req.gff_id.port_id[0] = fcport->d_id.b.domain; b 3503 drivers/scsi/qla2xxx/qla_gs.c ct_req->req.gff_id.port_id[1] = fcport->d_id.b.area; b 3504 drivers/scsi/qla2xxx/qla_gs.c ct_req->req.gff_id.port_id[2] = fcport->d_id.b.al_pa; b 3622 drivers/scsi/qla2xxx/qla_gs.c if ((rp->id.b.domain & 0xf0) == 0xf0) b 3783 drivers/scsi/qla2xxx/qla_gs.c id.b.rsvd_1 = 0; b 3784 drivers/scsi/qla2xxx/qla_gs.c id.b.domain = d->port_id[0]; b 3785 drivers/scsi/qla2xxx/qla_gs.c id.b.area = d->port_id[1]; b 3786 drivers/scsi/qla2xxx/qla_gs.c id.b.al_pa = d->port_id[2]; b 355 drivers/scsi/qla2xxx/qla_init.c fcport->d_id.b.domain, fcport->d_id.b.area, fcport->d_id.b.al_pa, b 406 drivers/scsi/qla2xxx/qla_init.c sp->handle, fcport->loop_id, fcport->d_id.b.domain, b 407 drivers/scsi/qla2xxx/qla_init.c fcport->d_id.b.area, fcport->d_id.b.al_pa, b 468 drivers/scsi/qla2xxx/qla_init.c sp->handle, fcport->loop_id, fcport->d_id.b.domain, b 469 drivers/scsi/qla2xxx/qla_init.c fcport->d_id.b.area, fcport->d_id.b.al_pa); b 727 drivers/scsi/qla2xxx/qla_init.c fcport->d_id.b.domain, fcport->d_id.b.area, b 728 drivers/scsi/qla2xxx/qla_init.c fcport->d_id.b.al_pa, fcport->loop_id); b 733 drivers/scsi/qla2xxx/qla_init.c id.b.domain = e->port_id[2]; b 734 drivers/scsi/qla2xxx/qla_init.c id.b.area = e->port_id[1]; b 735 drivers/scsi/qla2xxx/qla_init.c id.b.al_pa = e->port_id[0]; b 736 drivers/scsi/qla2xxx/qla_init.c id.b.rsvd_1 = 0; b 765 drivers/scsi/qla2xxx/qla_init.c fcport->fc4f_nvme, id.b.domain, id.b.area, id.b.al_pa, b 766 drivers/scsi/qla2xxx/qla_init.c fcport->d_id.b.domain, fcport->d_id.b.area, b 767 drivers/scsi/qla2xxx/qla_init.c fcport->d_id.b.al_pa, loop_id, fcport->loop_id); b 895 drivers/scsi/qla2xxx/qla_init.c id.b.domain = e->port_id[0]; b 896 drivers/scsi/qla2xxx/qla_init.c id.b.area = e->port_id[1]; b 897 drivers/scsi/qla2xxx/qla_init.c id.b.al_pa = e->port_id[2]; b 898 drivers/scsi/qla2xxx/qla_init.c id.b.rsvd_1 = 0; b 1047 drivers/scsi/qla2xxx/qla_init.c id.b.domain = e->port_id[2]; b 1048 drivers/scsi/qla2xxx/qla_init.c id.b.area = e->port_id[1]; b 1049 drivers/scsi/qla2xxx/qla_init.c id.b.al_pa = e->port_id[0]; b 1050 drivers/scsi/qla2xxx/qla_init.c id.b.rsvd_1 = 0; b 1758 drivers/scsi/qla2xxx/qla_init.c ea.id.b.rsvd_1 = RSCN_PORT_ADDR; b 1808 drivers/scsi/qla2xxx/qla_init.c sp->handle, fcport->loop_id, fcport->d_id.b.domain, b 1809 drivers/scsi/qla2xxx/qla_init.c fcport->d_id.b.area, fcport->d_id.b.al_pa); b 2026 drivers/scsi/qla2xxx/qla_init.c cid.b.domain = (ea->iop[1] >> 16) & 0xff; b 2027 drivers/scsi/qla2xxx/qla_init.c cid.b.area = (ea->iop[1] >> 8) & 0xff; b 2028 drivers/scsi/qla2xxx/qla_init.c cid.b.al_pa = ea->iop[1] & 0xff; b 2029 drivers/scsi/qla2xxx/qla_init.c cid.b.rsvd_1 = 0; b 4381 drivers/scsi/qla2xxx/qla_init.c id.b.domain = domain; b 4382 drivers/scsi/qla2xxx/qla_init.c id.b.area = area; b 4383 drivers/scsi/qla2xxx/qla_init.c id.b.al_pa = al_pa; b 4384 drivers/scsi/qla2xxx/qla_init.c id.b.rsvd_1 = 0; b 4815 drivers/scsi/qla2xxx/qla_init.c port_state_str[state], fcport->d_id.b.domain, b 4816 drivers/scsi/qla2xxx/qla_init.c fcport->d_id.b.area, fcport->d_id.b.al_pa); b 5150 drivers/scsi/qla2xxx/qla_init.c if (area && domain && ((area != vha->d_id.b.area) || b 5151 drivers/scsi/qla2xxx/qla_init.c (domain != vha->d_id.b.domain)) && b 5163 drivers/scsi/qla2xxx/qla_init.c new_fcport->d_id.b.domain = domain; b 5164 drivers/scsi/qla2xxx/qla_init.c new_fcport->d_id.b.area = area; b 5165 drivers/scsi/qla2xxx/qla_init.c new_fcport->d_id.b.al_pa = al_pa; b 5336 drivers/scsi/qla2xxx/qla_init.c rport_ids.port_id = fcport->d_id.b.domain << 16 | b 5337 drivers/scsi/qla2xxx/qla_init.c fcport->d_id.b.area << 8 | fcport->d_id.b.al_pa; b 5753 drivers/scsi/qla2xxx/qla_init.c if (swl[swl_idx].d_id.b.rsvd_1 != 0) { b 5777 drivers/scsi/qla2xxx/qla_init.c new_fcport->d_id.b.domain, b 5778 drivers/scsi/qla2xxx/qla_init.c new_fcport->d_id.b.area, b 5779 drivers/scsi/qla2xxx/qla_init.c new_fcport->d_id.b.al_pa); b 5798 drivers/scsi/qla2xxx/qla_init.c if ((new_fcport->d_id.b.domain & 0xf0) == 0xf0) b 5856 drivers/scsi/qla2xxx/qla_init.c fcport->d_id.b.domain, b 5857 drivers/scsi/qla2xxx/qla_init.c fcport->d_id.b.area, b 5858 drivers/scsi/qla2xxx/qla_init.c fcport->d_id.b.al_pa, b 5860 drivers/scsi/qla2xxx/qla_init.c new_fcport->d_id.b.domain, b 5861 drivers/scsi/qla2xxx/qla_init.c new_fcport->d_id.b.area, b 5862 drivers/scsi/qla2xxx/qla_init.c new_fcport->d_id.b.al_pa); b 6001 drivers/scsi/qla2xxx/qla_init.c fcport->loop_id, fcport->d_id.b.domain, b 6002 drivers/scsi/qla2xxx/qla_init.c fcport->d_id.b.area, fcport->d_id.b.al_pa); b 6006 drivers/scsi/qla2xxx/qla_init.c fcport->d_id.b.domain, fcport->d_id.b.area, b 6007 drivers/scsi/qla2xxx/qla_init.c fcport->d_id.b.al_pa, mb, BIT_0); b 6026 drivers/scsi/qla2xxx/qla_init.c fcport->loop_id, fcport->d_id.b.domain, b 6027 drivers/scsi/qla2xxx/qla_init.c fcport->d_id.b.area, fcport->d_id.b.al_pa); b 6084 drivers/scsi/qla2xxx/qla_init.c fcport->d_id.b.domain, fcport->d_id.b.area, b 6085 drivers/scsi/qla2xxx/qla_init.c fcport->d_id.b.al_pa); b 6096 drivers/scsi/qla2xxx/qla_init.c "jiffies=%lx.\n", mb[0], fcport->d_id.b.domain, b 6097 drivers/scsi/qla2xxx/qla_init.c fcport->d_id.b.area, fcport->d_id.b.al_pa, b 6102 drivers/scsi/qla2xxx/qla_init.c fcport->d_id.b.domain, fcport->d_id.b.area, b 6103 drivers/scsi/qla2xxx/qla_init.c fcport->d_id.b.al_pa); b 8844 drivers/scsi/qla2xxx/qla_init.c fcport->loop_id, fcport->d_id.b.domain, b 8845 drivers/scsi/qla2xxx/qla_init.c fcport->d_id.b.area, fcport->d_id.b.al_pa); b 8851 drivers/scsi/qla2xxx/qla_init.c fcport->d_id.b.domain, fcport->d_id.b.area, b 8852 drivers/scsi/qla2xxx/qla_init.c fcport->d_id.b.al_pa); b 1668 drivers/scsi/qla2xxx/qla_iocb.c cmd_pkt->port_id[0] = sp->fcport->d_id.b.al_pa; b 1669 drivers/scsi/qla2xxx/qla_iocb.c cmd_pkt->port_id[1] = sp->fcport->d_id.b.area; b 1670 drivers/scsi/qla2xxx/qla_iocb.c cmd_pkt->port_id[2] = sp->fcport->d_id.b.domain; b 1853 drivers/scsi/qla2xxx/qla_iocb.c cmd_pkt->port_id[0] = sp->fcport->d_id.b.al_pa; b 1854 drivers/scsi/qla2xxx/qla_iocb.c cmd_pkt->port_id[1] = sp->fcport->d_id.b.area; b 1855 drivers/scsi/qla2xxx/qla_iocb.c cmd_pkt->port_id[2] = sp->fcport->d_id.b.domain; b 1988 drivers/scsi/qla2xxx/qla_iocb.c cmd_pkt->port_id[0] = sp->fcport->d_id.b.al_pa; b 1989 drivers/scsi/qla2xxx/qla_iocb.c cmd_pkt->port_id[1] = sp->fcport->d_id.b.area; b 1990 drivers/scsi/qla2xxx/qla_iocb.c cmd_pkt->port_id[2] = sp->fcport->d_id.b.domain; b 2188 drivers/scsi/qla2xxx/qla_iocb.c cmd_pkt->port_id[0] = sp->fcport->d_id.b.al_pa; b 2189 drivers/scsi/qla2xxx/qla_iocb.c cmd_pkt->port_id[1] = sp->fcport->d_id.b.area; b 2190 drivers/scsi/qla2xxx/qla_iocb.c cmd_pkt->port_id[2] = sp->fcport->d_id.b.domain; b 2353 drivers/scsi/qla2xxx/qla_iocb.c logio->port_id[0] = sp->fcport->d_id.b.al_pa; b 2354 drivers/scsi/qla2xxx/qla_iocb.c logio->port_id[1] = sp->fcport->d_id.b.area; b 2355 drivers/scsi/qla2xxx/qla_iocb.c logio->port_id[2] = sp->fcport->d_id.b.domain; b 2375 drivers/scsi/qla2xxx/qla_iocb.c logio->port_id[0] = sp->fcport->d_id.b.al_pa; b 2376 drivers/scsi/qla2xxx/qla_iocb.c logio->port_id[1] = sp->fcport->d_id.b.area; b 2377 drivers/scsi/qla2xxx/qla_iocb.c logio->port_id[2] = sp->fcport->d_id.b.domain; b 2399 drivers/scsi/qla2xxx/qla_iocb.c mbx->mb2 = cpu_to_le16(sp->fcport->d_id.b.domain); b 2400 drivers/scsi/qla2xxx/qla_iocb.c mbx->mb3 = cpu_to_le16(sp->fcport->d_id.b.area << 8 | b 2401 drivers/scsi/qla2xxx/qla_iocb.c sp->fcport->d_id.b.al_pa); b 2422 drivers/scsi/qla2xxx/qla_iocb.c logio->port_id[0] = sp->fcport->d_id.b.al_pa; b 2423 drivers/scsi/qla2xxx/qla_iocb.c logio->port_id[1] = sp->fcport->d_id.b.area; b 2424 drivers/scsi/qla2xxx/qla_iocb.c logio->port_id[2] = sp->fcport->d_id.b.domain; b 2439 drivers/scsi/qla2xxx/qla_iocb.c mbx->mb2 = cpu_to_le16(sp->fcport->d_id.b.domain); b 2440 drivers/scsi/qla2xxx/qla_iocb.c mbx->mb3 = cpu_to_le16(sp->fcport->d_id.b.area << 8 | b 2441 drivers/scsi/qla2xxx/qla_iocb.c sp->fcport->d_id.b.al_pa); b 2496 drivers/scsi/qla2xxx/qla_iocb.c tsk->port_id[0] = fcport->d_id.b.al_pa; b 2497 drivers/scsi/qla2xxx/qla_iocb.c tsk->port_id[1] = fcport->d_id.b.area; b 2498 drivers/scsi/qla2xxx/qla_iocb.c tsk->port_id[2] = fcport->d_id.b.domain; b 2543 drivers/scsi/qla2xxx/qla_iocb.c sp->name, sp->handle, fcport->d_id.b.domain, fcport->d_id.b.area, b 2544 drivers/scsi/qla2xxx/qla_iocb.c fcport->d_id.b.al_pa); b 2557 drivers/scsi/qla2xxx/qla_iocb.c sp->name, sp->handle, fcport->d_id.b.domain, b 2558 drivers/scsi/qla2xxx/qla_iocb.c fcport->d_id.b.area, fcport->d_id.b.al_pa); b 2591 drivers/scsi/qla2xxx/qla_iocb.c fcport->d_id.b.domain = remote_did.b.domain; b 2592 drivers/scsi/qla2xxx/qla_iocb.c fcport->d_id.b.area = remote_did.b.area; b 2593 drivers/scsi/qla2xxx/qla_iocb.c fcport->d_id.b.al_pa = remote_did.b.al_pa; b 2596 drivers/scsi/qla2xxx/qla_iocb.c fcport->d_id.b.domain, fcport->d_id.b.area, fcport->d_id.b.al_pa); b 2620 drivers/scsi/qla2xxx/qla_iocb.c logo_pyld.s_id[0] = vha->d_id.b.al_pa; b 2621 drivers/scsi/qla2xxx/qla_iocb.c logo_pyld.s_id[1] = vha->d_id.b.area; b 2622 drivers/scsi/qla2xxx/qla_iocb.c logo_pyld.s_id[2] = vha->d_id.b.domain; b 2637 drivers/scsi/qla2xxx/qla_iocb.c sp->name, sp->handle, fcport->loop_id, fcport->d_id.b.domain, b 2638 drivers/scsi/qla2xxx/qla_iocb.c fcport->d_id.b.area, fcport->d_id.b.al_pa); b 2664 drivers/scsi/qla2xxx/qla_iocb.c els_iocb->port_id[0] = sp->fcport->d_id.b.al_pa; b 2665 drivers/scsi/qla2xxx/qla_iocb.c els_iocb->port_id[1] = sp->fcport->d_id.b.area; b 2666 drivers/scsi/qla2xxx/qla_iocb.c els_iocb->port_id[2] = sp->fcport->d_id.b.domain; b 2668 drivers/scsi/qla2xxx/qla_iocb.c els_iocb->s_id[1] = vha->d_id.b.al_pa; b 2669 drivers/scsi/qla2xxx/qla_iocb.c els_iocb->s_id[2] = vha->d_id.b.area; b 2670 drivers/scsi/qla2xxx/qla_iocb.c els_iocb->s_id[0] = vha->d_id.b.domain; b 2905 drivers/scsi/qla2xxx/qla_iocb.c els_iocb->port_id[0] = sp->fcport->d_id.b.al_pa; b 2906 drivers/scsi/qla2xxx/qla_iocb.c els_iocb->port_id[1] = sp->fcport->d_id.b.area; b 2907 drivers/scsi/qla2xxx/qla_iocb.c els_iocb->port_id[2] = sp->fcport->d_id.b.domain; b 3243 drivers/scsi/qla2xxx/qla_iocb.c cmd_pkt->port_id[0] = sp->fcport->d_id.b.al_pa; b 3244 drivers/scsi/qla2xxx/qla_iocb.c cmd_pkt->port_id[1] = sp->fcport->d_id.b.area; b 3245 drivers/scsi/qla2xxx/qla_iocb.c cmd_pkt->port_id[2] = sp->fcport->d_id.b.domain; b 3314 drivers/scsi/qla2xxx/qla_iocb.c cmd_pkt->port_id[0] = sp->fcport->d_id.b.al_pa; b 3315 drivers/scsi/qla2xxx/qla_iocb.c cmd_pkt->port_id[1] = sp->fcport->d_id.b.area; b 3316 drivers/scsi/qla2xxx/qla_iocb.c cmd_pkt->port_id[2] = sp->fcport->d_id.b.domain; b 3412 drivers/scsi/qla2xxx/qla_iocb.c abt_iocb->port_id[0] = sp->fcport->d_id.b.al_pa; b 3413 drivers/scsi/qla2xxx/qla_iocb.c abt_iocb->port_id[1] = sp->fcport->d_id.b.area; b 3414 drivers/scsi/qla2xxx/qla_iocb.c abt_iocb->port_id[2] = sp->fcport->d_id.b.domain; b 3533 drivers/scsi/qla2xxx/qla_iocb.c logio->port_id[0] = sp->fcport->d_id.b.al_pa; b 3534 drivers/scsi/qla2xxx/qla_iocb.c logio->port_id[1] = sp->fcport->d_id.b.area; b 3535 drivers/scsi/qla2xxx/qla_iocb.c logio->port_id[2] = sp->fcport->d_id.b.domain; b 3789 drivers/scsi/qla2xxx/qla_iocb.c cmd_pkt->port_id[0] = vha->d_id.b.al_pa; b 3790 drivers/scsi/qla2xxx/qla_iocb.c cmd_pkt->port_id[1] = vha->d_id.b.area; b 3791 drivers/scsi/qla2xxx/qla_iocb.c cmd_pkt->port_id[2] = vha->d_id.b.domain; b 1095 drivers/scsi/qla2xxx/qla_isr.c host_pid = (vha->d_id.b.domain << 16) | (vha->d_id.b.area << 8) b 1096 drivers/scsi/qla2xxx/qla_isr.c | vha->d_id.b.al_pa; b 1118 drivers/scsi/qla2xxx/qla_isr.c ea.id.b.rsvd_1 = rscn_entry >> 24; b 1375 drivers/scsi/qla2xxx/qla_isr.c fcport->d_id.b.domain, fcport->d_id.b.area, b 1376 drivers/scsi/qla2xxx/qla_isr.c fcport->d_id.b.al_pa, mbx->entry_status, b 1393 drivers/scsi/qla2xxx/qla_isr.c type, sp->handle, fcport->d_id.b.domain, b 1394 drivers/scsi/qla2xxx/qla_isr.c fcport->d_id.b.area, fcport->d_id.b.al_pa, b 1423 drivers/scsi/qla2xxx/qla_isr.c fcport->d_id.b.domain, fcport->d_id.b.area, fcport->d_id.b.al_pa, b 1705 drivers/scsi/qla2xxx/qla_isr.c type, fcport->port_name, sp->handle, fcport->d_id.b.domain, b 1706 drivers/scsi/qla2xxx/qla_isr.c fcport->d_id.b.area, fcport->d_id.b.al_pa, b 1718 drivers/scsi/qla2xxx/qla_isr.c fcport->d_id.b.domain, b 1719 drivers/scsi/qla2xxx/qla_isr.c fcport->d_id.b.area, fcport->d_id.b.al_pa, b 1801 drivers/scsi/qla2xxx/qla_isr.c sp->handle, fcport->d_id.b.domain, b 1802 drivers/scsi/qla2xxx/qla_isr.c fcport->d_id.b.area, fcport->d_id.b.al_pa, b 2734 drivers/scsi/qla2xxx/qla_isr.c "port state= %s comp_status %x.\n", fcport->d_id.b.domain, b 2735 drivers/scsi/qla2xxx/qla_isr.c fcport->d_id.b.area, fcport->d_id.b.al_pa, b 2789 drivers/scsi/qla2xxx/qla_isr.c cp->device->id, cp->device->lun, fcport->d_id.b.domain, b 2790 drivers/scsi/qla2xxx/qla_isr.c fcport->d_id.b.area, fcport->d_id.b.al_pa, ox_id, b 1971 drivers/scsi/qla2xxx/qla_mbx.c fcport->d_id.b.domain = pd24->port_id[0]; b 1972 drivers/scsi/qla2xxx/qla_mbx.c fcport->d_id.b.area = pd24->port_id[1]; b 1973 drivers/scsi/qla2xxx/qla_mbx.c fcport->d_id.b.al_pa = pd24->port_id[2]; b 1974 drivers/scsi/qla2xxx/qla_mbx.c fcport->d_id.b.rsvd_1 = 0; b 1997 drivers/scsi/qla2xxx/qla_mbx.c pd->slave_state, fcport->d_id.b.domain, b 1998 drivers/scsi/qla2xxx/qla_mbx.c fcport->d_id.b.area, fcport->d_id.b.al_pa); b 2016 drivers/scsi/qla2xxx/qla_mbx.c fcport->d_id.b.domain = pd->port_id[0]; b 2017 drivers/scsi/qla2xxx/qla_mbx.c fcport->d_id.b.area = pd->port_id[3]; b 2018 drivers/scsi/qla2xxx/qla_mbx.c fcport->d_id.b.al_pa = pd->port_id[2]; b 2019 drivers/scsi/qla2xxx/qla_mbx.c fcport->d_id.b.rsvd_1 = 0; b 2589 drivers/scsi/qla2xxx/qla_mbx.c fcport->d_id.b.domain, fcport->d_id.b.area, b 2590 drivers/scsi/qla2xxx/qla_mbx.c fcport->d_id.b.al_pa, mb_ret, opt); b 3151 drivers/scsi/qla2xxx/qla_mbx.c abt->port_id[0] = fcport->d_id.b.al_pa; b 3152 drivers/scsi/qla2xxx/qla_mbx.c abt->port_id[1] = fcport->d_id.b.area; b 3153 drivers/scsi/qla2xxx/qla_mbx.c abt->port_id[2] = fcport->d_id.b.domain; b 3231 drivers/scsi/qla2xxx/qla_mbx.c tsk->p.tsk.port_id[0] = fcport->d_id.b.al_pa; b 3232 drivers/scsi/qla2xxx/qla_mbx.c tsk->p.tsk.port_id[1] = fcport->d_id.b.area; b 3233 drivers/scsi/qla2xxx/qla_mbx.c tsk->p.tsk.port_id[2] = fcport->d_id.b.domain; b 3848 drivers/scsi/qla2xxx/qla_mbx.c id.b.domain = rptid_entry->port_id[2]; b 3849 drivers/scsi/qla2xxx/qla_mbx.c id.b.area = rptid_entry->port_id[1]; b 3850 drivers/scsi/qla2xxx/qla_mbx.c id.b.al_pa = rptid_entry->port_id[0]; b 3851 drivers/scsi/qla2xxx/qla_mbx.c id.b.rsvd_1 = 0; b 3921 drivers/scsi/qla2xxx/qla_mbx.c vha->d_id.b.al_pa = 1; b 3925 drivers/scsi/qla2xxx/qla_mbx.c id.b.al_pa = 2; b 4027 drivers/scsi/qla2xxx/qla_mbx.c vha->d_id.b.domain = rptid_entry->port_id[2]; b 4028 drivers/scsi/qla2xxx/qla_mbx.c vha->d_id.b.area = rptid_entry->port_id[1]; b 4029 drivers/scsi/qla2xxx/qla_mbx.c vha->d_id.b.al_pa = rptid_entry->port_id[0]; b 4050 drivers/scsi/qla2xxx/qla_mbx.c fcport->d_id.b.domain = b 4052 drivers/scsi/qla2xxx/qla_mbx.c fcport->d_id.b.area = b 4054 drivers/scsi/qla2xxx/qla_mbx.c fcport->d_id.b.al_pa = b 6391 drivers/scsi/qla2xxx/qla_mbx.c fcport->d_id.b.domain = pd->port_id[0]; b 6392 drivers/scsi/qla2xxx/qla_mbx.c fcport->d_id.b.area = pd->port_id[1]; b 6393 drivers/scsi/qla2xxx/qla_mbx.c fcport->d_id.b.al_pa = pd->port_id[2]; b 6394 drivers/scsi/qla2xxx/qla_mbx.c fcport->d_id.b.rsvd_1 = 0; b 1967 drivers/scsi/qla2xxx/qla_mr.c vha->d_id.b.domain = pinfo->port_id[0]; b 1968 drivers/scsi/qla2xxx/qla_mr.c vha->d_id.b.area = pinfo->port_id[1]; b 1969 drivers/scsi/qla2xxx/qla_mr.c vha->d_id.b.al_pa = pinfo->port_id[2]; b 451 drivers/scsi/qla2xxx/qla_nvme.c cmd_pkt->port_id[0] = sp->fcport->d_id.b.al_pa; b 452 drivers/scsi/qla2xxx/qla_nvme.c cmd_pkt->port_id[1] = sp->fcport->d_id.b.area; b 453 drivers/scsi/qla2xxx/qla_nvme.c cmd_pkt->port_id[2] = sp->fcport->d_id.b.domain; b 196 drivers/scsi/qla2xxx/qla_target.c if (vha->d_id.b.area == d_id.area && b 197 drivers/scsi/qla2xxx/qla_target.c vha->d_id.b.domain == d_id.domain && b 198 drivers/scsi/qla2xxx/qla_target.c vha->d_id.b.al_pa == d_id.al_pa) b 840 drivers/scsi/qla2xxx/qla_target.c port_id.b.domain = iocb->u.isp24.port_id[2]; b 841 drivers/scsi/qla2xxx/qla_target.c port_id.b.area = iocb->u.isp24.port_id[1]; b 842 drivers/scsi/qla2xxx/qla_target.c port_id.b.al_pa = iocb->u.isp24.port_id[0]; b 843 drivers/scsi/qla2xxx/qla_target.c port_id.b.rsvd_1 = 0; b 943 drivers/scsi/qla2xxx/qla_target.c logo->id.b.domain, logo->id.b.area, logo->id.b.al_pa, b 964 drivers/scsi/qla2xxx/qla_target.c sess->d_id.b.domain, sess->d_id.b.area, sess->d_id.b.al_pa, b 1406 drivers/scsi/qla2xxx/qla_target.c fcport->loop_id, sess->d_id.b.domain, sess->d_id.b.area, b 1407 drivers/scsi/qla2xxx/qla_target.c sess->d_id.b.al_pa, sess->conf_compl_supported ? "" : "not "); b 4547 drivers/scsi/qla2xxx/qla_target.c fcport->d_id.b.domain, fcport->d_id.b.area, b 4548 drivers/scsi/qla2xxx/qla_target.c fcport->d_id.b.al_pa, rc); b 4636 drivers/scsi/qla2xxx/qla_target.c key = (((u32)s_id->b.domain << 16) | b 4637 drivers/scsi/qla2xxx/qla_target.c ((u32)s_id->b.area << 8) | b 4638 drivers/scsi/qla2xxx/qla_target.c ((u32)s_id->b.al_pa)); b 4687 drivers/scsi/qla2xxx/qla_target.c port_id.b.domain = iocb->u.isp24.port_id[2]; b 4688 drivers/scsi/qla2xxx/qla_target.c port_id.b.area = iocb->u.isp24.port_id[1]; b 4689 drivers/scsi/qla2xxx/qla_target.c port_id.b.al_pa = iocb->u.isp24.port_id[0]; b 4690 drivers/scsi/qla2xxx/qla_target.c port_id.b.rsvd_1 = 0; b 4865 drivers/scsi/qla2xxx/qla_target.c port_id.b.domain = iocb->u.isp24.port_id[2]; b 4866 drivers/scsi/qla2xxx/qla_target.c port_id.b.area = iocb->u.isp24.port_id[1]; b 4867 drivers/scsi/qla2xxx/qla_target.c port_id.b.al_pa = iocb->u.isp24.port_id[0]; b 4868 drivers/scsi/qla2xxx/qla_target.c port_id.b.rsvd_1 = 0; b 6456 drivers/scsi/qla2xxx/qla_target.c unsigned char *b) b 6460 drivers/scsi/qla2xxx/qla_target.c put_unaligned_be64(wwpn, b); b 6461 drivers/scsi/qla2xxx/qla_target.c pr_debug("qla2xxx passed configfs WWPN: %8phC\n", b); b 6483 drivers/scsi/qla2xxx/qla_target.c u8 b[WWN_SIZE]; b 6521 drivers/scsi/qla2xxx/qla_target.c qlt_lport_dump(vha, phys_wwpn, b); b 6523 drivers/scsi/qla2xxx/qla_target.c if (memcmp(vha->port_name, b, WWN_SIZE)) { b 986 drivers/scsi/qla2xxx/qla_target.h ((_s_id.b.domain == 0xff) && ((_s_id.b.area & 0xf0) == 0xf0)) b 89 drivers/scsi/qla2xxx/tcm_qla2xxx.c u8 b[8]; b 91 drivers/scsi/qla2xxx/tcm_qla2xxx.c put_unaligned_be64(wwn, b); b 94 drivers/scsi/qla2xxx/tcm_qla2xxx.c b[0], b[1], b[2], b[3], b[4], b[5], b[6], b[7]); b 1487 drivers/scsi/qla2xxx/tcm_qla2xxx.c sess->loop_id, loop_id, sess->d_id.b.domain, b 1488 drivers/scsi/qla2xxx/tcm_qla2xxx.c sess->d_id.b.area, sess->d_id.b.al_pa, s_id.b.domain, b 1489 drivers/scsi/qla2xxx/tcm_qla2xxx.c s_id.b.area, s_id.b.al_pa); b 1509 drivers/scsi/qla2xxx/tcm_qla2xxx.c key = (((u32) sess->d_id.b.domain << 16) | b 1510 drivers/scsi/qla2xxx/tcm_qla2xxx.c ((u32) sess->d_id.b.area << 8) | b 1511 drivers/scsi/qla2xxx/tcm_qla2xxx.c ((u32) sess->d_id.b.al_pa)); b 1516 drivers/scsi/qla2xxx/tcm_qla2xxx.c sess->d_id.b.domain, sess->d_id.b.area, b 1517 drivers/scsi/qla2xxx/tcm_qla2xxx.c sess->d_id.b.al_pa); b 1520 drivers/scsi/qla2xxx/tcm_qla2xxx.c sess->d_id.b.domain, sess->d_id.b.area, b 1521 drivers/scsi/qla2xxx/tcm_qla2xxx.c sess->d_id.b.al_pa); b 1523 drivers/scsi/qla2xxx/tcm_qla2xxx.c key = (((u32) s_id.b.domain << 16) | b 1524 drivers/scsi/qla2xxx/tcm_qla2xxx.c ((u32) s_id.b.area << 8) | b 1525 drivers/scsi/qla2xxx/tcm_qla2xxx.c ((u32) s_id.b.al_pa)); b 1529 drivers/scsi/qla2xxx/tcm_qla2xxx.c s_id.b.domain, s_id.b.area, s_id.b.al_pa); b 13 drivers/scsi/qla4xxx/ql4_dbg.c void qla4xxx_dump_buffer(void *b, uint32_t size) b 16 drivers/scsi/qla4xxx/ql4_dbg.c uint8_t *c = b; b 77 drivers/scsi/qla4xxx/ql4_glbl.h void qla4xxx_dump_buffer(void *b, uint32_t size); b 60 drivers/scsi/qlogicpti.c #define PACKB(a, b) (((a)<<4)|(b)) b 1077 drivers/scsi/scsi_debug.c char b[32]; b 1150 drivers/scsi/scsi_debug.c snprintf(b, sizeof(b), "%08X", target_dev_id); b 1151 drivers/scsi/scsi_debug.c memcpy(arr + num, b, 8); b 2799 drivers/scsi/scsi_debug.c char b[128]; b 2805 drivers/scsi/scsi_debug.c n += scnprintf(b + n, sizeof(b) - n, b 2808 drivers/scsi/scsi_debug.c n += scnprintf(b + n, sizeof(b) - n, b 2811 drivers/scsi/scsi_debug.c pr_err("%04d: %s\n", i, b); b 5612 drivers/scsi/scsi_debug.c char b[120]; b 5616 drivers/scsi/scsi_debug.c sb = (int)sizeof(b); b 5618 drivers/scsi/scsi_debug.c strcpy(b, "too long, over 32 bytes"); b 5621 drivers/scsi/scsi_debug.c n += scnprintf(b + n, sb - n, "%02x ", b 5625 drivers/scsi/scsi_debug.c blk_mq_unique_tag(scp->request), b); b 72 drivers/scsi/scsi_dh.c const struct scsi_dh_blist *b; b 77 drivers/scsi/scsi_dh.c for (b = scsi_dh_blist; b->vendor; b++) { b 78 drivers/scsi/scsi_dh.c if (!strncmp(sdev->vendor, b->vendor, strlen(b->vendor)) && b 79 drivers/scsi/scsi_dh.c !strncmp(sdev->model, b->model, strlen(b->model))) { b 80 drivers/scsi/scsi_dh.c return b->driver; b 1862 drivers/scsi/storvsc_drv.c target = (device->dev_instance.b[5] << 8 | b 1863 drivers/scsi/storvsc_drv.c device->dev_instance.b[4]); b 98 drivers/scsi/sym53c8xx_2/sym_malloc.c unsigned long a, b; b 125 drivers/scsi/sym53c8xx_2/sym_malloc.c b = a ^ s; b 127 drivers/scsi/sym53c8xx_2/sym_malloc.c while (q->next && q->next != (m_link_p) b) { b 136 drivers/scsi/sym53c8xx_2/sym_malloc.c a = a & b; b 351 drivers/scsi/sym53c8xx_2/sym_malloc.c dma_addr_t b; b 362 drivers/scsi/sym53c8xx_2/sym_malloc.c b = vp->baddr + (m - a); b 364 drivers/scsi/sym53c8xx_2/sym_malloc.c return b; b 106 drivers/sh/intc/internals.h static inline int intc_handle_int_cmp(const void *a, const void *b) b 109 drivers/sh/intc/internals.h const struct intc_handle_int *_b = b; b 335 drivers/slimbus/core.c static bool slim_eaddr_equal(struct slim_eaddr *a, struct slim_eaddr *b) b 337 drivers/slimbus/core.c return (a->manf_id == b->manf_id && b 338 drivers/slimbus/core.c a->prod_code == b->prod_code && b 339 drivers/slimbus/core.c a->dev_index == b->dev_index && b 340 drivers/slimbus/core.c a->instance == b->instance); b 38 drivers/slimbus/slimbus.h #define SLIM_HEADER_GET_MT(b) ((b >> SLIM_MSG_MT_SHIFT) & SLIM_MSG_MT_MASK) b 39 drivers/slimbus/slimbus.h #define SLIM_HEADER_GET_RL(b) ((b >> SLIM_MSG_RL_SHIFT) & SLIM_MSG_RL_MASK) b 40 drivers/slimbus/slimbus.h #define SLIM_HEADER_GET_MC(b) ((b >> SLIM_MSG_MC_SHIFT) & SLIM_MSG_MC_MASK) b 41 drivers/slimbus/slimbus.h #define SLIM_HEADER_GET_DT(b) ((b >> SLIM_MSG_DT_SHIFT) & SLIM_MSG_DT_MASK) b 51 drivers/soc/fsl/qbman/bman_test_api.c static inline int bufs_cmp(const struct bm_buffer *a, const struct bm_buffer *b) b 67 drivers/soc/fsl/qbman/bman_test_api.c (bm_buffer_get64(b) & BMAN_TOKEN_MASK)) b 70 drivers/soc/fsl/qbman/bman_test_api.c (bm_buffer_get64(b) & BMAN_TOKEN_MASK)) b 73 drivers/soc/fsl/qbman/bman_test_api.c if (bm_buffer_get64(a) < bm_buffer_get64(b)) b 75 drivers/soc/fsl/qbman/bman_test_api.c if (bm_buffer_get64(a) > bm_buffer_get64(b)) b 130 drivers/soc/fsl/qbman/qman_priv.h const struct qman_cgrs *a, const struct qman_cgrs *b) b 135 drivers/soc/fsl/qbman/qman_priv.h const u32 *_b = b->q.state; b 142 drivers/soc/fsl/qbman/qman_priv.h const struct qman_cgrs *a, const struct qman_cgrs *b) b 147 drivers/soc/fsl/qbman/qman_priv.h const u32 *_b = b->q.state; b 93 drivers/soc/fsl/qbman/qman_test_api.c static bool fd_neq(const struct qm_fd *a, const struct qm_fd *b) b 95 drivers/soc/fsl/qbman/qman_test_api.c bool neq = qm_fd_addr_get64(a) != qm_fd_addr_get64(b); b 97 drivers/soc/fsl/qbman/qman_test_api.c neq |= qm_fd_get_format(a) != qm_fd_get_format(b); b 98 drivers/soc/fsl/qbman/qman_test_api.c neq |= a->cfg != b->cfg; b 99 drivers/soc/fsl/qbman/qman_test_api.c neq |= a->cmd != b->cmd; b 173 drivers/soc/qcom/apr.c struct apr_rx_buf *abuf, *b; b 177 drivers/soc/qcom/apr.c list_for_each_entry_safe(abuf, b, &apr->rx_list, node) { b 404 drivers/spi/spi-loopback-test.c u8 b; b 432 drivers/spi/spi-loopback-test.c b = ((u8 *)xfer->rx_buf)[xfer->len - 1 - i]; b 433 drivers/spi/spi-loopback-test.c if (b != SPI_TEST_PATTERN_UNWRITTEN) b 451 drivers/spi/spi-loopback-test.c static int rx_ranges_cmp(void *priv, struct list_head *a, struct list_head *b) b 454 drivers/spi/spi-loopback-test.c struct rx_ranges *rx_b = list_entry(b, struct rx_ranges, list); b 215 drivers/spi/spi-pic32.c BUILD_SPI_FIFO_RW(byte, u8, b); b 903 drivers/staging/comedi/comedi.h #define __RANGE(a, b) ((((a) & 0xffff) << 16) | ((b) & 0xffff)) b 906 drivers/staging/comedi/comedi.h #define RANGE_LENGTH(b) ((b) & 0xffff) b 213 drivers/staging/comedi/comedi_buf.c void *b = bm->page_list[pg].virt_addr + pgoff; b 216 drivers/staging/comedi/comedi_buf.c memcpy(b, buf, l); b 218 drivers/staging/comedi/comedi_buf.c memcpy(buf, b, l); b 21 drivers/staging/comedi/comedidev.h #define COMEDI_VERSION(a, b, c) (((a) << 16) + ((b) << 8) + (c)) b 604 drivers/staging/comedi/comedidev.h #define RANGE(a, b) {(a) * 1e6, (b) * 1e6, 0} b 605 drivers/staging/comedi/comedidev.h #define RANGE_ext(a, b) {(a) * 1e6, (b) * 1e6, RF_EXTERNAL} b 606 drivers/staging/comedi/comedidev.h #define RANGE_mA(a, b) {(a) * 1e6, (b) * 1e6, UNIT_mA} b 607 drivers/staging/comedi/comedidev.h #define RANGE_unitless(a, b) {(a) * 1e6, (b) * 1e6, 0} b 244 drivers/staging/comedi/drivers/ni_routes.c const struct ni_route_set *b = vb; b 246 drivers/staging/comedi/drivers/ni_routes.c if (a->dest < b->dest) b 248 drivers/staging/comedi/drivers/ni_routes.c else if (a->dest > b->dest) b 752 drivers/staging/exfat/exfat.h int nls_dosname_cmp(struct super_block *sb, u8 *a, u8 *b); b 753 drivers/staging/exfat/exfat.h int nls_uniname_cmp(struct super_block *sb, u16 *a, u16 *b); b 594 drivers/staging/exfat/exfat_core.c int i, b; b 600 drivers/staging/exfat/exfat_core.c b = clu & ((p_bd->sector_size << 3) - 1); b 604 drivers/staging/exfat/exfat_core.c exfat_bitmap_set((u8 *)p_fs->vol_amap[i]->b_data, b); b 611 drivers/staging/exfat/exfat_core.c int i, b; b 622 drivers/staging/exfat/exfat_core.c b = clu & ((p_bd->sector_size << 3) - 1); b 626 drivers/staging/exfat/exfat_core.c exfat_bitmap_clear((u8 *)p_fs->vol_amap[i]->b_data, b); b 99 drivers/staging/exfat/exfat_nls.c int nls_dosname_cmp(struct super_block *sb, u8 *a, u8 *b) b 101 drivers/staging/exfat/exfat_nls.c return strncmp(a, b, DOS_NAME_LENGTH); b 104 drivers/staging/exfat/exfat_nls.c int nls_uniname_cmp(struct super_block *sb, u16 *a, u16 *b) b 108 drivers/staging/exfat/exfat_nls.c for (i = 0; i < MAX_NAME_LENGTH; i++, a++, b++) { b 109 drivers/staging/exfat/exfat_nls.c if (nls_upper(sb, *a) != nls_upper(sb, *b)) b 309 drivers/staging/fbtft/fb_agm1264k-fl.c u16 b = pixel & 0x1f; b 313 drivers/staging/fbtft/fb_agm1264k-fl.c pixel = (299 * r + 587 * g + 114 * b) / 200; b 65 drivers/staging/fbtft/fb_ssd1325.c u16 b = pixel & 0x1f; b 69 drivers/staging/fbtft/fb_ssd1325.c pixel = (299 * r + 587 * g + 114 * b) / 195; b 418 drivers/staging/fieldbus/anybuss/host.c unsigned int a, b, i = 0; b 422 drivers/staging/fieldbus/anybuss/host.c regmap_read(regmap, REG_IND_AB, &b); b 423 drivers/staging/fieldbus/anybuss/host.c if (likely(a == b)) b 95 drivers/staging/fwserial/fwserial.c #define to_device(a, b) (a->b) b 753 drivers/staging/greybus/power_supply.c static int get_property(struct power_supply *b, b 757 drivers/staging/greybus/power_supply.c struct gb_power_supply *gbpsy = to_gb_power_supply(b); b 800 drivers/staging/greybus/power_supply.c static int set_property(struct power_supply *b, b 804 drivers/staging/greybus/power_supply.c struct gb_power_supply *gbpsy = to_gb_power_supply(b); b 809 drivers/staging/greybus/power_supply.c static int property_is_writeable(struct power_supply *b, b 812 drivers/staging/greybus/power_supply.c struct gb_power_supply *gbpsy = to_gb_power_supply(b); b 40 drivers/staging/isdn/hysdn/boardergo.c unsigned char volatile b; b 55 drivers/staging/isdn/hysdn/boardergo.c b = dpr->ToPcInt; /* clear for ergo */ b 56 drivers/staging/isdn/hysdn/boardergo.c b |= dpr->ToPcIntMetro; /* same for metro */ b 57 drivers/staging/isdn/hysdn/boardergo.c b |= dpr->ToHyInt; /* and for champ */ b 269 drivers/staging/media/hantro/hantro_h264.c struct hantro_h264_reflist_builder *b) b 286 drivers/staging/media/hantro/hantro_h264.c memset(b, 0, sizeof(*b)); b 287 drivers/staging/media/hantro/hantro_h264.c b->dpb = dpb; b 288 drivers/staging/media/hantro/hantro_h264.c b->curpoc = get_poc(buf->field, dec_param->top_field_order_cnt, b 310 drivers/staging/media/hantro/hantro_h264.c b->frame_nums[i] = (int)dpb[i].frame_num - max_frame_num; b 312 drivers/staging/media/hantro/hantro_h264.c b->frame_nums[i] = dpb[i].frame_num; b 314 drivers/staging/media/hantro/hantro_h264.c b->pocs[i] = get_poc(buf->field, dpb[i].top_field_order_cnt, b 316 drivers/staging/media/hantro/hantro_h264.c b->unordered_reflist[b->num_valid] = i; b 317 drivers/staging/media/hantro/hantro_h264.c b->num_valid++; b 320 drivers/staging/media/hantro/hantro_h264.c for (i = b->num_valid; i < ARRAY_SIZE(ctx->h264_dec.dpb); i++) b 321 drivers/staging/media/hantro/hantro_h264.c b->unordered_reflist[i] = i; b 327 drivers/staging/media/hantro/hantro_h264.c const struct v4l2_h264_dpb_entry *a, *b; b 333 drivers/staging/media/hantro/hantro_h264.c b = &builder->dpb[idxb]; b 336 drivers/staging/media/hantro/hantro_h264.c (b->flags & V4L2_H264_DPB_ENTRY_FLAG_LONG_TERM)) { b 351 drivers/staging/media/hantro/hantro_h264.c return a->pic_num - b->pic_num; b 357 drivers/staging/media/hantro/hantro_h264.c const struct v4l2_h264_dpb_entry *a, *b; b 364 drivers/staging/media/hantro/hantro_h264.c b = &builder->dpb[idxb]; b 367 drivers/staging/media/hantro/hantro_h264.c (b->flags & V4L2_H264_DPB_ENTRY_FLAG_LONG_TERM)) { b 377 drivers/staging/media/hantro/hantro_h264.c return a->pic_num - b->pic_num; b 398 drivers/staging/media/hantro/hantro_h264.c const struct v4l2_h264_dpb_entry *a, *b; b 405 drivers/staging/media/hantro/hantro_h264.c b = &builder->dpb[idxb]; b 408 drivers/staging/media/hantro/hantro_h264.c (b->flags & V4L2_H264_DPB_ENTRY_FLAG_LONG_TERM)) { b 418 drivers/staging/media/hantro/hantro_h264.c return a->pic_num - b->pic_num; b 466 drivers/staging/media/hantro/hantro_h264.c const struct v4l2_h264_dpb_entry *b) b 468 drivers/staging/media/hantro/hantro_h264.c return a->top_field_order_cnt == b->top_field_order_cnt && b 469 drivers/staging/media/hantro/hantro_h264.c a->bottom_field_order_cnt == b->bottom_field_order_cnt; b 601 drivers/staging/media/ipu3/include/intel-ipu3.h __u16 b; b 620 drivers/staging/media/ipu3/include/intel-ipu3.h __u8 b; b 667 drivers/staging/media/ipu3/include/intel-ipu3.h __u8 b; b 1185 drivers/staging/media/ipu3/include/intel-ipu3.h __u16 b; b 2196 drivers/staging/media/ipu3/include/intel-ipu3.h __s32 b:12; b 2277 drivers/staging/media/ipu3/include/intel-ipu3.h __u16 b; b 2569 drivers/staging/media/ipu3/include/intel-ipu3.h __u16 b[IPU3_UAPI_ISP_VEC_ELEMS]; b 2648 drivers/staging/media/ipu3/include/intel-ipu3.h __u16 b; b 1382 drivers/staging/media/ipu3/ipu3-abi.h struct imgu_abi_frame_sp_plane b; b 1401 drivers/staging/media/ipu3/ipu3-abi.h struct imgu_abi_frame_sp_plane b; b 11 drivers/staging/media/ipu3/ipu3-css-params.c #define DIV_ROUND_CLOSEST_DOWN(a, b) (((a) + ((b) / 2) - 1) / (b)) b 12 drivers/staging/media/ipu3/ipu3-css-params.c #define roundclosest_down(a, b) (DIV_ROUND_CLOSEST_DOWN(a, b) * (b)) b 2265 drivers/staging/media/ipu3/ipu3-css-params.c acc->tcc.macc_table.entries[i].b = 0; b 2792 drivers/staging/media/ipu3/ipu3-css-params.c xnr_vmem->b[i] = imgu_css_xnr3_vmem_defaults.b b 1401 drivers/staging/media/ipu3/ipu3-css.c struct imgu_css_buffer *b, *b0; b 1425 drivers/staging/media/ipu3/ipu3-css.c list_for_each_entry_safe(b, b0, b 1428 drivers/staging/media/ipu3/ipu3-css.c b->state = IPU3_CSS_BUFFER_FAILED; b 1429 drivers/staging/media/ipu3/ipu3-css.c list_del(&b->list); b 1932 drivers/staging/media/ipu3/ipu3-css.c struct imgu_css_buffer *b) b 1943 drivers/staging/media/ipu3/ipu3-css.c if (b->queue >= IPU3_CSS_QUEUES || !imgu_css_queues[b->queue].qid) b 1946 drivers/staging/media/ipu3/ipu3-css.c b->queue_pos = imgu_css_queue_pos(css, imgu_css_queues[b->queue].qid, b 1949 drivers/staging/media/ipu3/ipu3-css.c if (b->queue_pos >= ARRAY_SIZE(css->pipes[pipe].abi_buffers[b->queue])) b 1951 drivers/staging/media/ipu3/ipu3-css.c abi_buf = css->pipes[pipe].abi_buffers[b->queue][b->queue_pos].vaddr; b 1956 drivers/staging/media/ipu3/ipu3-css.c buf_addr = (void *)abi_buf + imgu_css_queues[b->queue].ptr_ofs; b 1957 drivers/staging/media/ipu3/ipu3-css.c *(imgu_addr_t *)buf_addr = b->daddr; b 1959 drivers/staging/media/ipu3/ipu3-css.c if (b->queue == IPU3_CSS_QUEUE_STAT_3A) b 1960 drivers/staging/media/ipu3/ipu3-css.c abi_buf->payload.s3a.data.dmem.s3a_tbl = b->daddr; b 1962 drivers/staging/media/ipu3/ipu3-css.c if (b->queue == IPU3_CSS_QUEUE_OUT) b 1966 drivers/staging/media/ipu3/ipu3-css.c if (b->queue == IPU3_CSS_QUEUE_VF) b 1971 drivers/staging/media/ipu3/ipu3-css.c list_add_tail(&b->list, &css_pipe->queue[b->queue].bufs); b 1973 drivers/staging/media/ipu3/ipu3-css.c b->state = IPU3_CSS_BUFFER_QUEUED; b 1975 drivers/staging/media/ipu3/ipu3-css.c data = css->pipes[pipe].abi_buffers[b->queue][b->queue_pos].daddr; b 1976 drivers/staging/media/ipu3/ipu3-css.c r = imgu_css_queue_data(css, imgu_css_queues[b->queue].qid, b 1982 drivers/staging/media/ipu3/ipu3-css.c imgu_css_queues[b->queue].qid); b 1988 drivers/staging/media/ipu3/ipu3-css.c b, b->queue, pipe); b 1993 drivers/staging/media/ipu3/ipu3-css.c b->state = (r == -EBUSY || r == -EAGAIN) ? b 1995 drivers/staging/media/ipu3/ipu3-css.c list_del(&b->list); b 2013 drivers/staging/media/ipu3/ipu3-css.c struct imgu_css_buffer *b = ERR_PTR(-EAGAIN); b 2074 drivers/staging/media/ipu3/ipu3-css.c b = list_first_entry(&css_pipe->queue[queue].bufs, b 2076 drivers/staging/media/ipu3/ipu3-css.c if (queue != b->queue || b 2078 drivers/staging/media/ipu3/ipu3-css.c [b->queue][b->queue_pos].daddr) { b 2085 drivers/staging/media/ipu3/ipu3-css.c b->pipe = pipe; b 2086 drivers/staging/media/ipu3/ipu3-css.c b->state = IPU3_CSS_BUFFER_DONE; b 2087 drivers/staging/media/ipu3/ipu3-css.c list_del(&b->list); b 2129 drivers/staging/media/ipu3/ipu3-css.c return b; b 181 drivers/staging/media/ipu3/ipu3-css.h struct imgu_css_buffer *b); b 200 drivers/staging/media/ipu3/ipu3-css.h imgu_css_buf_state(struct imgu_css_buffer *b) b 202 drivers/staging/media/ipu3/ipu3-css.h return b->state; b 206 drivers/staging/media/ipu3/ipu3-css.h static inline void imgu_css_buf_init(struct imgu_css_buffer *b, b 209 drivers/staging/media/ipu3/ipu3-css.h b->state = IPU3_CSS_BUFFER_NEW; b 210 drivers/staging/media/ipu3/ipu3-css.h b->queue = queue; b 211 drivers/staging/media/ipu3/ipu3-css.h b->daddr = daddr; b 9304 drivers/staging/media/ipu3/ipu3-tables.c .b = { b 36 drivers/staging/media/ipu3/ipu3-tables.h s16 b[IMGU_XNR3_VMEM_LUT_LEN]; b 447 drivers/staging/media/ipu3/ipu3-v4l2.c struct imgu_vb2_buffer *b, *b0; b 451 drivers/staging/media/ipu3/ipu3-v4l2.c list_for_each_entry_safe(b, b0, &node->buffers, list) { b 452 drivers/staging/media/ipu3/ipu3-v4l2.c list_del(&b->list); b 453 drivers/staging/media/ipu3/ipu3-v4l2.c vb2_buffer_done(&b->vbb.vb2_buf, state); b 1417 drivers/staging/media/ipu3/ipu3-v4l2.c struct imgu_vb2_buffer *b = b 1420 drivers/staging/media/ipu3/ipu3-v4l2.c list_del(&b->list); b 1421 drivers/staging/media/ipu3/ipu3-v4l2.c vb2_buffer_done(&b->vbb.vb2_buf, state); b 510 drivers/staging/media/ipu3/ipu3.c struct imgu_css_buffer *b; b 517 drivers/staging/media/ipu3/ipu3.c b = imgu_css_buf_dequeue(&imgu->css); b 519 drivers/staging/media/ipu3/ipu3.c } while (PTR_ERR(b) == -EAGAIN); b 521 drivers/staging/media/ipu3/ipu3.c if (IS_ERR(b)) { b 522 drivers/staging/media/ipu3/ipu3.c if (PTR_ERR(b) != -EBUSY) /* All done */ b 525 drivers/staging/media/ipu3/ipu3.c PTR_ERR(b)); b 529 drivers/staging/media/ipu3/ipu3.c node = imgu_map_node(imgu, b->queue); b 530 drivers/staging/media/ipu3/ipu3.c pipe = b->pipe; b 531 drivers/staging/media/ipu3/ipu3.c dummy = imgu_dummybufs_check(imgu, b, pipe); b 533 drivers/staging/media/ipu3/ipu3.c buf = container_of(b, struct imgu_buffer, css_buf); b 539 drivers/staging/media/ipu3/ipu3.c (u32)b->daddr); b 787 drivers/staging/media/omap4iss/iss_video.c iss_video_querybuf(struct file *file, void *fh, struct v4l2_buffer *b) b 791 drivers/staging/media/omap4iss/iss_video.c return vb2_querybuf(&vfh->queue, b); b 795 drivers/staging/media/omap4iss/iss_video.c iss_video_qbuf(struct file *file, void *fh, struct v4l2_buffer *b) b 800 drivers/staging/media/omap4iss/iss_video.c return vb2_qbuf(&vfh->queue, video->video.v4l2_dev->mdev, b); b 812 drivers/staging/media/omap4iss/iss_video.c iss_video_dqbuf(struct file *file, void *fh, struct v4l2_buffer *b) b 816 drivers/staging/media/omap4iss/iss_video.c return vb2_dqbuf(&vfh->queue, b, file->f_flags & O_NONBLOCK); b 425 drivers/staging/media/soc_camera/soc_camera.c struct v4l2_buffer *b) b 429 drivers/staging/media/soc_camera/soc_camera.c return vb2_prepare_buf(&icd->vb2_vidq, NULL, b); b 545 drivers/staging/most/core.c static int split_string(char *buf, char **a, char **b, char **c, char **d) b 551 drivers/staging/most/core.c *b = strsep(&buf, ":\n"); b 552 drivers/staging/most/core.c if (!*b) b 25 drivers/staging/octeon/octeon-stubs.h #define CVMX_GMXX_RXX_ADR_CAM_EN(a, b) ((a)+(b)) b 26 drivers/staging/octeon/octeon-stubs.h #define CVMX_GMXX_RXX_ADR_CTL(a, b) ((a)+(b)) b 27 drivers/staging/octeon/octeon-stubs.h #define CVMX_GMXX_PRTX_CFG(a, b) ((a)+(b)) b 28 drivers/staging/octeon/octeon-stubs.h #define CVMX_GMXX_RXX_FRM_MAX(a, b) ((a)+(b)) b 29 drivers/staging/octeon/octeon-stubs.h #define CVMX_GMXX_RXX_JABBER(a, b) ((a)+(b)) b 1412 drivers/staging/octeon/octeon-stubs.h #define CVMX_ASXX_RX_CLK_SETX(a, b) ((a)+(b)) b 1413 drivers/staging/octeon/octeon-stubs.h #define CVMX_ASXX_TX_CLK_SETX(a, b) ((a)+(b)) b 1415 drivers/staging/octeon/octeon-stubs.h #define CVMX_GMXX_RXX_ADR_CAM0(a, b) ((a)+(b)) b 1416 drivers/staging/octeon/octeon-stubs.h #define CVMX_GMXX_RXX_ADR_CAM1(a, b) ((a)+(b)) b 1417 drivers/staging/octeon/octeon-stubs.h #define CVMX_GMXX_RXX_ADR_CAM2(a, b) ((a)+(b)) b 1418 drivers/staging/octeon/octeon-stubs.h #define CVMX_GMXX_RXX_ADR_CAM3(a, b) ((a)+(b)) b 1419 drivers/staging/octeon/octeon-stubs.h #define CVMX_GMXX_RXX_ADR_CAM4(a, b) ((a)+(b)) b 1420 drivers/staging/octeon/octeon-stubs.h #define CVMX_GMXX_RXX_ADR_CAM5(a, b) ((a)+(b)) b 1421 drivers/staging/octeon/octeon-stubs.h #define CVMX_GMXX_RXX_FRM_CTL(a, b) ((a)+(b)) b 1422 drivers/staging/octeon/octeon-stubs.h #define CVMX_GMXX_RXX_INT_REG(a, b) ((a)+(b)) b 1423 drivers/staging/octeon/octeon-stubs.h #define CVMX_GMXX_SMACX(a, b) ((a)+(b)) b 291 drivers/staging/rtl8188eu/core/rtw_mlme.c static int is_same_ess(struct wlan_bssid_ex *a, struct wlan_bssid_ex *b) b 293 drivers/staging/rtl8188eu/core/rtw_mlme.c return (a->ssid.ssid_length == b->ssid.ssid_length) && b 294 drivers/staging/rtl8188eu/core/rtw_mlme.c !memcmp(a->ssid.ssid, b->ssid.ssid, a->ssid.ssid_length); b 296 drivers/staging/rtl8188eu/core/rtw_security.c void rtw_secmicappendbyte(struct mic_data *pmicdata, u8 b) b 299 drivers/staging/rtl8188eu/core/rtw_security.c pmicdata->M |= ((unsigned long)b) << (8*pmicdata->nBytesInM); b 781 drivers/staging/rtl8188eu/core/rtw_security.c static void xor_128(u8 *a, u8 *b, u8 *out); b 782 drivers/staging/rtl8188eu/core/rtw_security.c static void xor_32(u8 *a, u8 *b, u8 *out); b 795 drivers/staging/rtl8188eu/core/rtw_security.c static void xor_128(u8 *a, u8 *b, u8 *out) b 800 drivers/staging/rtl8188eu/core/rtw_security.c out[i] = a[i] ^ b[i]; b 803 drivers/staging/rtl8188eu/core/rtw_security.c static void xor_32(u8 *a, u8 *b, u8 *out) b 808 drivers/staging/rtl8188eu/core/rtw_security.c out[i] = a[i] ^ b[i]; b 132 drivers/staging/rtl8188eu/include/rtw_recv.h #define SN_LESS(a, b) (((a - b) & 0x800) != 0) b 133 drivers/staging/rtl8188eu/include/rtw_recv.h #define SN_EQUAL(a, b) (a == b) b 293 drivers/staging/rtl8188eu/include/rtw_security.h void rtw_secmicappendbyte(struct mic_data *pmicdata, u8 b); b 705 drivers/staging/rtl8188eu/os_dep/ioctl_linux.c union iwreq_data *wrqu, char *b) b 753 drivers/staging/rtl8188eu/os_dep/ioctl_linux.c union iwreq_data *wrqu, char *b) b 1941 drivers/staging/rtl8188eu/os_dep/ioctl_linux.c union iwreq_data *wrqu, char *b) b 23 drivers/staging/rtl8192e/rtl8192e/rtl_wx.c union iwreq_data *wrqu, char *b) b 27 drivers/staging/rtl8192e/rtl8192e/rtl_wx.c return rtllib_wx_get_freq(priv->rtllib, a, wrqu, b); b 32 drivers/staging/rtl8192e/rtl8192e/rtl_wx.c union iwreq_data *wrqu, char *b) b 36 drivers/staging/rtl8192e/rtl8192e/rtl_wx.c return rtllib_wx_get_mode(priv->rtllib, a, wrqu, b); b 245 drivers/staging/rtl8192e/rtl8192e/rtl_wx.c union iwreq_data *wrqu, char *b) b 277 drivers/staging/rtl8192e/rtl8192e/rtl_wx.c ret = rtllib_wx_set_mode(priv->rtllib, a, wrqu, b); b 378 drivers/staging/rtl8192e/rtl8192e/rtl_wx.c union iwreq_data *wrqu, char *b) b 406 drivers/staging/rtl8192e/rtl8192e/rtl_wx.c struct iw_scan_req *req = (struct iw_scan_req *)b; b 458 drivers/staging/rtl8192e/rtl8192e/rtl_wx.c ret = rtllib_wx_set_scan(priv->rtllib, a, wrqu, b); b 467 drivers/staging/rtl8192e/rtl8192e/rtl_wx.c union iwreq_data *wrqu, char *b) b 480 drivers/staging/rtl8192e/rtl8192e/rtl_wx.c ret = rtllib_wx_get_scan(priv->rtllib, a, wrqu, b); b 489 drivers/staging/rtl8192e/rtl8192e/rtl_wx.c union iwreq_data *wrqu, char *b) b 501 drivers/staging/rtl8192e/rtl8192e/rtl_wx.c ret = rtllib_wx_set_essid(priv->rtllib, a, wrqu, b); b 510 drivers/staging/rtl8192e/rtl8192e/rtl_wx.c union iwreq_data *wrqu, char *b) b 517 drivers/staging/rtl8192e/rtl8192e/rtl_wx.c ret = rtllib_wx_get_essid(priv->rtllib, a, wrqu, b); b 557 drivers/staging/rtl8192e/rtl8192e/rtl_wx.c union iwreq_data *wrqu, char *b) b 567 drivers/staging/rtl8192e/rtl8192e/rtl_wx.c ret = rtllib_wx_set_freq(priv->rtllib, a, wrqu, b); b 417 drivers/staging/rtl8192e/rtllib.h #define SN_LESS(a, b) (((a-b)&0x800) != 0) b 418 drivers/staging/rtl8192e/rtllib.h #define SN_EQUAL(a, b) (a == b) b 2028 drivers/staging/rtl8192e/rtllib.h union iwreq_data *wrqu, char *b); b 2037 drivers/staging/rtl8192e/rtllib.h union iwreq_data *wrqu, char *b); b 2040 drivers/staging/rtl8192e/rtllib.h union iwreq_data *wrqu, char *b); b 2046 drivers/staging/rtl8192e/rtllib.h union iwreq_data *wrqu, char *b); b 2049 drivers/staging/rtl8192e/rtllib.h union iwreq_data *wrqu, char *b); b 2052 drivers/staging/rtl8192e/rtllib.h union iwreq_data *wrqu, char *b); b 2807 drivers/staging/rtl8192e/rtllib_softmac.c struct rtllib_probe_response *b; b 2814 drivers/staging/rtl8192e/rtllib_softmac.c b = (struct rtllib_probe_response *) skb->data; b 2815 drivers/staging/rtl8192e/rtllib_softmac.c b->header.frame_ctl = cpu_to_le16(RTLLIB_STYPE_BEACON); b 2824 drivers/staging/rtl8192e/rtllib_softmac.c struct rtllib_probe_response *b; b 2830 drivers/staging/rtl8192e/rtllib_softmac.c b = (struct rtllib_probe_response *) skb->data; b 2831 drivers/staging/rtl8192e/rtllib_softmac.c b->header.seq_ctl = cpu_to_le16(ieee->seq_ctrl[0] << 4); b 30 drivers/staging/rtl8192e/rtllib_softmac_wx.c union iwreq_data *wrqu, char *b) b 89 drivers/staging/rtl8192e/rtllib_softmac_wx.c union iwreq_data *wrqu, char *b) b 191 drivers/staging/rtl8192e/rtllib_softmac_wx.c union iwreq_data *wrqu, char *b) b 216 drivers/staging/rtl8192e/rtllib_softmac_wx.c strncpy(b, ieee->current_network.ssid, len); b 282 drivers/staging/rtl8192e/rtllib_softmac_wx.c union iwreq_data *wrqu, char *b) b 417 drivers/staging/rtl8192e/rtllib_softmac_wx.c union iwreq_data *wrqu, char *b) b 498 drivers/staging/rtl8192e/rtllib_softmac_wx.c union iwreq_data *wrqu, char *b) b 416 drivers/staging/rtl8192u/ieee80211/ieee80211.h #define SN_LESS(a, b) (((a - b) & 0x800) != 0) b 417 drivers/staging/rtl8192u/ieee80211/ieee80211.h #define SN_EQUAL(a, b) (a == b) b 2285 drivers/staging/rtl8192u/ieee80211/ieee80211.h union iwreq_data *wrqu, char *b); b 2297 drivers/staging/rtl8192u/ieee80211/ieee80211.h union iwreq_data *wrqu, char *b); b 2301 drivers/staging/rtl8192u/ieee80211/ieee80211.h union iwreq_data *wrqu, char *b); b 2309 drivers/staging/rtl8192u/ieee80211/ieee80211.h union iwreq_data *wrqu, char *b); b 2313 drivers/staging/rtl8192u/ieee80211/ieee80211.h union iwreq_data *wrqu, char *b); b 2317 drivers/staging/rtl8192u/ieee80211/ieee80211.h union iwreq_data *wrqu, char *b); b 2413 drivers/staging/rtl8192u/ieee80211/ieee80211_softmac.c struct ieee80211_probe_response *b; b 2420 drivers/staging/rtl8192u/ieee80211/ieee80211_softmac.c b = (struct ieee80211_probe_response *)skb->data; b 2421 drivers/staging/rtl8192u/ieee80211/ieee80211_softmac.c b->header.frame_ctl = cpu_to_le16(IEEE80211_STYPE_BEACON); b 2429 drivers/staging/rtl8192u/ieee80211/ieee80211_softmac.c struct ieee80211_probe_response *b; b 2435 drivers/staging/rtl8192u/ieee80211/ieee80211_softmac.c b = (struct ieee80211_probe_response *)skb->data; b 2436 drivers/staging/rtl8192u/ieee80211/ieee80211_softmac.c b->header.seq_ctl = cpu_to_le16(ieee->seq_ctrl[0] << 4); b 31 drivers/staging/rtl8192u/ieee80211/ieee80211_softmac_wx.c union iwreq_data *wrqu, char *b) b 88 drivers/staging/rtl8192u/ieee80211/ieee80211_softmac_wx.c union iwreq_data *wrqu, char *b) b 180 drivers/staging/rtl8192u/ieee80211/ieee80211_softmac_wx.c int ieee80211_wx_get_essid(struct ieee80211_device *ieee, struct iw_request_info *a, union iwreq_data *wrqu, char *b) b 205 drivers/staging/rtl8192u/ieee80211/ieee80211_softmac_wx.c strncpy(b, ieee->current_network.ssid, len); b 271 drivers/staging/rtl8192u/ieee80211/ieee80211_softmac_wx.c union iwreq_data *wrqu, char *b) b 360 drivers/staging/rtl8192u/ieee80211/ieee80211_softmac_wx.c union iwreq_data *wrqu, char *b) b 438 drivers/staging/rtl8192u/ieee80211/ieee80211_softmac_wx.c union iwreq_data *wrqu, char *b) b 94 drivers/staging/rtl8192u/r8180_93cx6.c static void eprom_send_bits_string(struct net_device *dev, short b[], int len) b 99 drivers/staging/rtl8192u/r8180_93cx6.c eprom_w(dev, b[i]); b 24 drivers/staging/rtl8192u/r8192U_core.c double __adddf3(double a, double b) b 26 drivers/staging/rtl8192u/r8192U_core.c return a + b; b 29 drivers/staging/rtl8192u/r8192U_core.c double __addsf3(float a, float b) b 31 drivers/staging/rtl8192u/r8192U_core.c return a + b; b 34 drivers/staging/rtl8192u/r8192U_core.c double __subdf3(double a, double b) b 36 drivers/staging/rtl8192u/r8192U_core.c return a - b; b 39 drivers/staging/rtl8192u/r8192U_wx.c union iwreq_data *wrqu, char *b) b 43 drivers/staging/rtl8192u/r8192U_wx.c return ieee80211_wx_get_freq(priv->ieee80211, a, wrqu, b); b 47 drivers/staging/rtl8192u/r8192U_wx.c union iwreq_data *wrqu, char *b) b 51 drivers/staging/rtl8192u/r8192U_wx.c return ieee80211_wx_get_mode(priv->ieee80211, a, wrqu, b); b 185 drivers/staging/rtl8192u/r8192U_wx.c union iwreq_data *wrqu, char *b) b 192 drivers/staging/rtl8192u/r8192U_wx.c ret = ieee80211_wx_set_mode(priv->ieee80211, a, wrqu, b); b 321 drivers/staging/rtl8192u/r8192U_wx.c union iwreq_data *wrqu, char *b) b 333 drivers/staging/rtl8192u/r8192U_wx.c struct iw_scan_req *req = (struct iw_scan_req *)b; b 347 drivers/staging/rtl8192u/r8192U_wx.c ret = ieee80211_wx_set_scan(priv->ieee80211, a, wrqu, b); b 355 drivers/staging/rtl8192u/r8192U_wx.c union iwreq_data *wrqu, char *b) b 366 drivers/staging/rtl8192u/r8192U_wx.c ret = ieee80211_wx_get_scan(priv->ieee80211, a, wrqu, b); b 375 drivers/staging/rtl8192u/r8192U_wx.c union iwreq_data *wrqu, char *b) b 382 drivers/staging/rtl8192u/r8192U_wx.c ret = ieee80211_wx_set_essid(priv->ieee80211, a, wrqu, b); b 391 drivers/staging/rtl8192u/r8192U_wx.c union iwreq_data *wrqu, char *b) b 398 drivers/staging/rtl8192u/r8192U_wx.c ret = ieee80211_wx_get_essid(priv->ieee80211, a, wrqu, b); b 406 drivers/staging/rtl8192u/r8192U_wx.c union iwreq_data *wrqu, char *b) b 413 drivers/staging/rtl8192u/r8192U_wx.c ret = ieee80211_wx_set_freq(priv->ieee80211, a, wrqu, b); b 841 drivers/staging/rtl8192u/r8192U_wx.c union iwreq_data *wrqu, char *b) b 36 drivers/staging/rtl8712/rtl8712_recv.h #define SN_LESS(a, b) (((a-b) & 0x800) != 0) b 37 drivers/staging/rtl8712/rtl8712_recv.h #define SN_EQUAL(a, b) (a == b) b 715 drivers/staging/rtl8712/rtl871x_ioctl_linux.c union iwreq_data *wrqu, char *b) b 746 drivers/staging/rtl8712/rtl871x_ioctl_linux.c union iwreq_data *wrqu, char *b) b 1853 drivers/staging/rtl8712/rtl871x_ioctl_linux.c union iwreq_data *wrqu, char *b) b 285 drivers/staging/rtl8712/rtl871x_security.c static void secmicappendbyte(struct mic_data *pmicdata, u8 b) b 288 drivers/staging/rtl8712/rtl871x_security.c pmicdata->M |= ((u32)b) << (8 * pmicdata->nBytesInM); b 740 drivers/staging/rtl8712/rtl871x_security.c static void xor_128(u8 *a, u8 *b, u8 *out) b 745 drivers/staging/rtl8712/rtl871x_security.c out[i] = a[i] ^ b[i]; b 748 drivers/staging/rtl8712/rtl871x_security.c static void xor_32(u8 *a, u8 *b, u8 *out) b 753 drivers/staging/rtl8712/rtl871x_security.c out[i] = a[i] ^ b[i]; b 409 drivers/staging/rtl8723bs/core/rtw_mlme.c inline int is_same_ess(struct wlan_bssid_ex *a, struct wlan_bssid_ex *b) b 413 drivers/staging/rtl8723bs/core/rtw_mlme.c return (a->Ssid.SsidLength == b->Ssid.SsidLength) b 414 drivers/staging/rtl8723bs/core/rtw_mlme.c && !memcmp(a->Ssid.Ssid, b->Ssid.Ssid, a->Ssid.SsidLength); b 359 drivers/staging/rtl8723bs/core/rtw_security.c void rtw_secmicappendbyte(struct mic_data *pmicdata, u8 b) b 362 drivers/staging/rtl8723bs/core/rtw_security.c pmicdata->M |= ((unsigned long)b) << (8*pmicdata->nBytesInM); b 947 drivers/staging/rtl8723bs/core/rtw_security.c static void xor_128(u8 *a, u8 *b, u8 *out); b 948 drivers/staging/rtl8723bs/core/rtw_security.c static void xor_32(u8 *a, u8 *b, u8 *out); b 962 drivers/staging/rtl8723bs/core/rtw_security.c static void xor_128(u8 *a, u8 *b, u8 *out) b 967 drivers/staging/rtl8723bs/core/rtw_security.c out[i] = a[i] ^ b[i]; b 972 drivers/staging/rtl8723bs/core/rtw_security.c static void xor_32(u8 *a, u8 *b, u8 *out) b 977 drivers/staging/rtl8723bs/core/rtw_security.c out[i] = a[i] ^ b[i]; b 99 drivers/staging/rtl8723bs/include/hal_com_h2c.h #define eqMacAddr(a, b) (((a)[0]==(b)[0] && (a)[1]==(b)[1] && (a)[2]==(b)[2] && (a)[3]==(b)[3] && (a)[4]==(b)[4] && (a)[5]==(b)[5]) ? 1:0) b 190 drivers/staging/rtl8723bs/include/rtw_recv.h #define SN_LESS(a, b) (((a-b)&0x800)!= 0) b 191 drivers/staging/rtl8723bs/include/rtw_recv.h #define SN_EQUAL(a, b) (a == b) b 408 drivers/staging/rtl8723bs/include/rtw_security.h void rtw_secmicappendbyte(struct mic_data *pmicdata, u8 b); b 846 drivers/staging/rtl8723bs/os_dep/ioctl_linux.c union iwreq_data *wrqu, char *b) b 912 drivers/staging/rtl8723bs/os_dep/ioctl_linux.c union iwreq_data *wrqu, char *b) b 2361 drivers/staging/rtl8723bs/os_dep/ioctl_linux.c union iwreq_data *wrqu, char *b) b 2367 drivers/staging/rtl8723bs/os_dep/ioctl_linux.c union iwreq_data *wrqu, char *b) b 2395 drivers/staging/rtl8723bs/os_dep/ioctl_linux.c union iwreq_data *wrqu, char *b) b 19 drivers/staging/vc04_services/bcm2835-camera/mmal-common.h #define MMAL_FOURCC(a, b, c, d) ((a) | (b << 8) | (c << 16) | (d << 24)) b 118 drivers/staging/vc04_services/interface/vchiq_arm/vchiq_core.h #define BITSET_SIZE(b) ((b + 31) >> 5) b 119 drivers/staging/vc04_services/interface/vchiq_arm/vchiq_core.h #define BITSET_WORD(b) (b >> 5) b 120 drivers/staging/vc04_services/interface/vchiq_arm/vchiq_core.h #define BITSET_BIT(b) (1 << (b & 31)) b 121 drivers/staging/vc04_services/interface/vchiq_arm/vchiq_core.h #define BITSET_IS_SET(bs, b) (bs[BITSET_WORD(b)] & BITSET_BIT(b)) b 122 drivers/staging/vc04_services/interface/vchiq_arm/vchiq_core.h #define BITSET_SET(bs, b) (bs[BITSET_WORD(b)] |= BITSET_BIT(b)) b 123 drivers/staging/vc04_services/interface/vchiq_arm/vchiq_core.h #define BITSET_CLR(bs, b) (bs[BITSET_WORD(b)] &= ~BITSET_BIT(b)) b 893 drivers/staging/vt6655/rf.c long b = (byCurrRSSI & 0x3F); b 907 drivers/staging/vt6655/rf.c *pldBm = -1 * (a + b * 2); b 505 drivers/staging/vt6655/rxtx.c PK_TYPE_11B, &buf->b); b 539 drivers/staging/vt6655/rxtx.c PK_TYPE_11B, &buf->b); b 655 drivers/staging/vt6655/rxtx.c PK_TYPE_11B, &buf->b); b 690 drivers/staging/vt6655/rxtx.c PK_TYPE_11B, &buf->b); b 847 drivers/staging/vt6655/rxtx.c PK_TYPE_11B, &buf->b); b 885 drivers/staging/vt6655/rxtx.c PK_TYPE_11B, &buf->b); b 72 drivers/staging/vt6655/rxtx.h struct vnt_phy_field b; b 81 drivers/staging/vt6655/rxtx.h struct vnt_phy_field b; b 107 drivers/staging/vt6655/rxtx.h struct vnt_phy_field b; b 117 drivers/staging/vt6655/rxtx.h struct vnt_phy_field b; b 148 drivers/staging/vt6655/rxtx.h struct vnt_phy_field b; b 156 drivers/staging/vt6655/rxtx.h struct vnt_phy_field b; b 794 drivers/staging/vt6656/rf.c long b = rssi & 0x3f; b 811 drivers/staging/vt6656/rf.c *dbm = -1 * (a + b * 2); b 278 drivers/staging/vt6656/rxtx.c PK_TYPE_11B, &buf->b); b 312 drivers/staging/vt6656/rxtx.c PK_TYPE_11B, &buf->b); b 410 drivers/staging/vt6656/rxtx.c PK_TYPE_11B, &buf->b); b 437 drivers/staging/vt6656/rxtx.c PK_TYPE_11B, &buf->b); b 531 drivers/staging/vt6656/rxtx.c PK_TYPE_11B, &buf->b); b 567 drivers/staging/vt6656/rxtx.c PK_TYPE_11B, &buf->b); b 67 drivers/staging/vt6656/rxtx.h struct vnt_phy_field b; b 77 drivers/staging/vt6656/rxtx.h struct vnt_phy_field b; b 106 drivers/staging/vt6656/rxtx.h struct vnt_phy_field b; b 117 drivers/staging/vt6656/rxtx.h struct vnt_phy_field b; b 151 drivers/staging/vt6656/rxtx.h struct vnt_phy_field b; b 160 drivers/staging/vt6656/rxtx.h struct vnt_phy_field b; b 260 drivers/staging/wilc1000/wilc_netdev.c u8 b; b 274 drivers/staging/wilc1000/wilc_netdev.c b = WILC_FW_BSS_TYPE_INFRA; b 275 drivers/staging/wilc1000/wilc_netdev.c if (!wilc_wlan_cfg_set(vif, 0, WID_BSS_TYPE, &b, 1, 0, 0)) b 278 drivers/staging/wilc1000/wilc_netdev.c b = WILC_FW_TX_RATE_AUTO; b 279 drivers/staging/wilc1000/wilc_netdev.c if (!wilc_wlan_cfg_set(vif, 0, WID_CURRENT_TX_RATE, &b, 1, 0, 0)) b 282 drivers/staging/wilc1000/wilc_netdev.c b = WILC_FW_OPER_MODE_G_MIXED_11B_2; b 283 drivers/staging/wilc1000/wilc_netdev.c if (!wilc_wlan_cfg_set(vif, 0, WID_11G_OPERATING_MODE, &b, 1, 0, 0)) b 286 drivers/staging/wilc1000/wilc_netdev.c b = WILC_FW_PREAMBLE_SHORT; b 287 drivers/staging/wilc1000/wilc_netdev.c if (!wilc_wlan_cfg_set(vif, 0, WID_PREAMBLE, &b, 1, 0, 0)) b 290 drivers/staging/wilc1000/wilc_netdev.c b = WILC_FW_11N_PROT_AUTO; b 291 drivers/staging/wilc1000/wilc_netdev.c if (!wilc_wlan_cfg_set(vif, 0, WID_11N_PROT_MECH, &b, 1, 0, 0)) b 294 drivers/staging/wilc1000/wilc_netdev.c b = WILC_FW_ACTIVE_SCAN; b 295 drivers/staging/wilc1000/wilc_netdev.c if (!wilc_wlan_cfg_set(vif, 0, WID_SCAN_TYPE, &b, 1, 0, 0)) b 298 drivers/staging/wilc1000/wilc_netdev.c b = WILC_FW_SITE_SURVEY_OFF; b 299 drivers/staging/wilc1000/wilc_netdev.c if (!wilc_wlan_cfg_set(vif, 0, WID_SITE_SURVEY, &b, 1, 0, 0)) b 312 drivers/staging/wilc1000/wilc_netdev.c b = 0; b 313 drivers/staging/wilc1000/wilc_netdev.c if (!wilc_wlan_cfg_set(vif, 0, WID_BCAST_SSID, &b, 1, 0, 0)) b 316 drivers/staging/wilc1000/wilc_netdev.c b = 1; b 317 drivers/staging/wilc1000/wilc_netdev.c if (!wilc_wlan_cfg_set(vif, 0, WID_QOS_ENABLE, &b, 1, 0, 0)) b 320 drivers/staging/wilc1000/wilc_netdev.c b = WILC_FW_NO_POWERSAVE; b 321 drivers/staging/wilc1000/wilc_netdev.c if (!wilc_wlan_cfg_set(vif, 0, WID_POWER_MANAGEMENT, &b, 1, 0, 0)) b 324 drivers/staging/wilc1000/wilc_netdev.c b = WILC_FW_SEC_NO; b 325 drivers/staging/wilc1000/wilc_netdev.c if (!wilc_wlan_cfg_set(vif, 0, WID_11I_MODE, &b, 1, 0, 0)) b 328 drivers/staging/wilc1000/wilc_netdev.c b = WILC_FW_AUTH_OPEN_SYSTEM; b 329 drivers/staging/wilc1000/wilc_netdev.c if (!wilc_wlan_cfg_set(vif, 0, WID_AUTH_TYPE, &b, 1, 0, 0)) b 332 drivers/staging/wilc1000/wilc_netdev.c b = 3; b 333 drivers/staging/wilc1000/wilc_netdev.c if (!wilc_wlan_cfg_set(vif, 0, WID_LISTEN_INTERVAL, &b, 1, 0, 0)) b 336 drivers/staging/wilc1000/wilc_netdev.c b = 3; b 337 drivers/staging/wilc1000/wilc_netdev.c if (!wilc_wlan_cfg_set(vif, 0, WID_DTIM_PERIOD, &b, 1, 0, 0)) b 340 drivers/staging/wilc1000/wilc_netdev.c b = WILC_FW_ACK_POLICY_NORMAL; b 341 drivers/staging/wilc1000/wilc_netdev.c if (!wilc_wlan_cfg_set(vif, 0, WID_ACK_POLICY, &b, 1, 0, 0)) b 344 drivers/staging/wilc1000/wilc_netdev.c b = 0; b 345 drivers/staging/wilc1000/wilc_netdev.c if (!wilc_wlan_cfg_set(vif, 0, WID_USER_CONTROL_ON_TX_POWER, &b, 1, b 349 drivers/staging/wilc1000/wilc_netdev.c b = 48; b 350 drivers/staging/wilc1000/wilc_netdev.c if (!wilc_wlan_cfg_set(vif, 0, WID_TX_POWER_LEVEL_11A, &b, 1, 0, 0)) b 353 drivers/staging/wilc1000/wilc_netdev.c b = 28; b 354 drivers/staging/wilc1000/wilc_netdev.c if (!wilc_wlan_cfg_set(vif, 0, WID_TX_POWER_LEVEL_11B, &b, 1, 0, 0)) b 362 drivers/staging/wilc1000/wilc_netdev.c b = WILC_FW_REKEY_POLICY_DISABLE; b 363 drivers/staging/wilc1000/wilc_netdev.c if (!wilc_wlan_cfg_set(vif, 0, WID_REKEY_POLICY, &b, 1, 0, 0)) b 377 drivers/staging/wilc1000/wilc_netdev.c b = 1; b 378 drivers/staging/wilc1000/wilc_netdev.c if (!wilc_wlan_cfg_set(vif, 0, WID_SHORT_SLOT_ALLOWED, &b, 1, 0, b 382 drivers/staging/wilc1000/wilc_netdev.c b = WILC_FW_ERP_PROT_SELF_CTS; b 383 drivers/staging/wilc1000/wilc_netdev.c if (!wilc_wlan_cfg_set(vif, 0, WID_11N_ERP_PROT_TYPE, &b, 1, 0, 0)) b 386 drivers/staging/wilc1000/wilc_netdev.c b = 1; b 387 drivers/staging/wilc1000/wilc_netdev.c if (!wilc_wlan_cfg_set(vif, 0, WID_11N_ENABLE, &b, 1, 0, 0)) b 390 drivers/staging/wilc1000/wilc_netdev.c b = WILC_FW_11N_OP_MODE_HT_MIXED; b 391 drivers/staging/wilc1000/wilc_netdev.c if (!wilc_wlan_cfg_set(vif, 0, WID_11N_OPERATING_MODE, &b, 1, 0, 0)) b 394 drivers/staging/wilc1000/wilc_netdev.c b = 1; b 395 drivers/staging/wilc1000/wilc_netdev.c if (!wilc_wlan_cfg_set(vif, 0, WID_11N_TXOP_PROT_DISABLE, &b, 1, 0, 0)) b 398 drivers/staging/wilc1000/wilc_netdev.c b = WILC_FW_OBBS_NONHT_DETECT_PROTECT_REPORT; b 399 drivers/staging/wilc1000/wilc_netdev.c if (!wilc_wlan_cfg_set(vif, 0, WID_11N_OBSS_NONHT_DETECTION, &b, 1, b 403 drivers/staging/wilc1000/wilc_netdev.c b = WILC_FW_HT_PROT_RTS_CTS_NONHT; b 404 drivers/staging/wilc1000/wilc_netdev.c if (!wilc_wlan_cfg_set(vif, 0, WID_11N_HT_PROT_TYPE, &b, 1, 0, 0)) b 407 drivers/staging/wilc1000/wilc_netdev.c b = 0; b 408 drivers/staging/wilc1000/wilc_netdev.c if (!wilc_wlan_cfg_set(vif, 0, WID_11N_RIFS_PROT_ENABLE, &b, 1, 0, b 412 drivers/staging/wilc1000/wilc_netdev.c b = 7; b 413 drivers/staging/wilc1000/wilc_netdev.c if (!wilc_wlan_cfg_set(vif, 0, WID_11N_CURRENT_TX_MCS, &b, 1, 0, 0)) b 416 drivers/staging/wilc1000/wilc_netdev.c b = 1; b 417 drivers/staging/wilc1000/wilc_netdev.c if (!wilc_wlan_cfg_set(vif, 0, WID_11N_IMMEDIATE_BA_ENABLED, &b, 1, b 166 drivers/staging/wilc1000/wilc_spi.c static int wilc_spi_tx(struct wilc *wilc, u8 *b, u32 len) b 172 drivers/staging/wilc1000/wilc_spi.c if (len > 0 && b) { b 174 drivers/staging/wilc1000/wilc_spi.c .tx_buf = b, b 282 drivers/staging/wilc1000/wilc_spi.c static int spi_cmd_complete(struct wilc *wilc, u8 cmd, u32 adr, u8 *b, u32 sz, b 361 drivers/staging/wilc1000/wilc_spi.c wb[3] = b[3]; b 362 drivers/staging/wilc1000/wilc_spi.c wb[4] = b[2]; b 363 drivers/staging/wilc1000/wilc_spi.c wb[5] = b[1]; b 364 drivers/staging/wilc1000/wilc_spi.c wb[6] = b[0]; b 372 drivers/staging/wilc1000/wilc_spi.c wb[4] = b[3]; b 373 drivers/staging/wilc1000/wilc_spi.c wb[5] = b[2]; b 374 drivers/staging/wilc1000/wilc_spi.c wb[6] = b[1]; b 375 drivers/staging/wilc1000/wilc_spi.c wb[7] = b[0]; b 487 drivers/staging/wilc1000/wilc_spi.c b[0] = rb[rix++]; b 488 drivers/staging/wilc1000/wilc_spi.c b[1] = rb[rix++]; b 489 drivers/staging/wilc1000/wilc_spi.c b[2] = rb[rix++]; b 490 drivers/staging/wilc1000/wilc_spi.c b[3] = rb[rix++]; b 515 drivers/staging/wilc1000/wilc_spi.c b[ix++] = rb[rix++]; b 530 drivers/staging/wilc1000/wilc_spi.c if (wilc_spi_rx(wilc, &b[ix], nbytes)) { b 587 drivers/staging/wilc1000/wilc_spi.c if (wilc_spi_rx(wilc, &b[ix], nbytes)) { b 611 drivers/staging/wilc1000/wilc_spi.c static int spi_data_write(struct wilc *wilc, u8 *b, u32 sz) b 651 drivers/staging/wilc1000/wilc_spi.c if (wilc_spi_tx(wilc, &b[ix], nbytes)) { b 102 drivers/staging/wilc1000/wilc_wlan_cfg.c static int wilc_wlan_cfg_set_bin(u8 *frame, u32 offset, u16 id, u8 *b, u32 size) b 113 drivers/staging/wilc1000/wilc_wlan_cfg.c if ((b) && size != 0) { b 114 drivers/staging/wilc1000/wilc_wlan_cfg.c memcpy(&frame[offset + 4], b, size); b 143 drivers/staging/wilc1000/wilc_wlan_cfg.c if (wl->cfg.b[i].id == WID_NIL) b 146 drivers/staging/wilc1000/wilc_wlan_cfg.c if (wl->cfg.b[i].id == wid) { b 147 drivers/staging/wilc1000/wilc_wlan_cfg.c wl->cfg.b[i].val = info[4]; b 222 drivers/staging/wilc1000/wilc_wlan_cfg.c if (wl->cfg.b[i].id == WID_NIL) b 225 drivers/staging/wilc1000/wilc_wlan_cfg.c if (wl->cfg.b[i].id == wid) { b 226 drivers/staging/wilc1000/wilc_wlan_cfg.c wl->cfg.b[i].val = info[3]; b 294 drivers/staging/wilc1000/wilc_wlan_cfg.c if (wl->cfg.b[i].id == WID_NIL) b 297 drivers/staging/wilc1000/wilc_wlan_cfg.c if (wl->cfg.b[i].id == wid) { b 298 drivers/staging/wilc1000/wilc_wlan_cfg.c memcpy(buffer, &wl->cfg.b[i].val, 1); b 404 drivers/staging/wilc1000/wilc_wlan_cfg.c wl->cfg.b = kmemdup(g_cfg_byte, sizeof(g_cfg_byte), GFP_KERNEL); b 405 drivers/staging/wilc1000/wilc_wlan_cfg.c if (!wl->cfg.b) b 446 drivers/staging/wilc1000/wilc_wlan_cfg.c kfree(wl->cfg.b); b 452 drivers/staging/wilc1000/wilc_wlan_cfg.c kfree(wl->cfg.b); b 37 drivers/staging/wilc1000/wilc_wlan_cfg.h struct wilc_cfg_byte *b; b 60 drivers/staging/wlan-ng/hfa384x.h #define HFA384x_FIRMWARE_VERSION(a, b, c) (((a) << 16) + ((b) << 8) + (c)) b 182 drivers/staging/wusbcore/crypto.c const struct aes_ccm_label *a, const void *b, b 216 drivers/staging/wusbcore/crypto.c crypto_shash_finup(desc, b, blen, iv); b 245 drivers/staging/wusbcore/crypto.c const void *b, size_t blen, size_t len) b 271 drivers/staging/wusbcore/crypto.c &n, a, b, blen); b 153 drivers/staging/wusbcore/host/whci/whci-hc.h #define QH_INFO2_BURST(b) ((b) << 0) /* maximum burst length */ b 318 drivers/staging/wusbcore/host/whci/whci-hc.h # define WUSBCMD_BCID(b) ((b) << 16) b 300 drivers/staging/wusbcore/include/wusb.h const void *b, size_t blen, size_t len); b 305 drivers/staging/wusbcore/include/wusb.h const void *b, size_t blen) b 307 drivers/staging/wusbcore/include/wusb.h return wusb_prf(out, out_size, key, n, a, b, blen, 64); b 313 drivers/staging/wusbcore/include/wusb.h const void *b, size_t blen) b 315 drivers/staging/wusbcore/include/wusb.h return wusb_prf(out, out_size, key, n, a, b, blen, 128); b 321 drivers/staging/wusbcore/include/wusb.h const void *b, size_t blen) b 323 drivers/staging/wusbcore/include/wusb.h return wusb_prf(out, out_size, key, n, a, b, blen, 256); b 431 drivers/staging/wusbcore/wa-xfer.c struct wa_xfer_abort_buffer *b = urb->context; b 432 drivers/staging/wusbcore/wa-xfer.c struct wahc *wa = b->wa; b 443 drivers/staging/wusbcore/wa-xfer.c xfer = wa_xfer_get_by_id(wa, le32_to_cpu(b->cmd.dwTransferID)); b 476 drivers/staging/wusbcore/wa-xfer.c __func__, le32_to_cpu(b->cmd.dwTransferID)); b 481 drivers/staging/wusbcore/wa-xfer.c usb_put_urb(&b->urb); b 498 drivers/staging/wusbcore/wa-xfer.c struct wa_xfer_abort_buffer *b; b 501 drivers/staging/wusbcore/wa-xfer.c b = kmalloc(sizeof(*b), GFP_ATOMIC); b 502 drivers/staging/wusbcore/wa-xfer.c if (b == NULL) b 504 drivers/staging/wusbcore/wa-xfer.c b->cmd.bLength = sizeof(b->cmd); b 505 drivers/staging/wusbcore/wa-xfer.c b->cmd.bRequestType = WA_XFER_ABORT; b 506 drivers/staging/wusbcore/wa-xfer.c b->cmd.wRPipe = rpipe->descr.wRPipeIndex; b 507 drivers/staging/wusbcore/wa-xfer.c b->cmd.dwTransferID = wa_xfer_id_le32(xfer); b 508 drivers/staging/wusbcore/wa-xfer.c b->wa = wa_get(xfer->wa); b 510 drivers/staging/wusbcore/wa-xfer.c usb_init_urb(&b->urb); b 511 drivers/staging/wusbcore/wa-xfer.c usb_fill_bulk_urb(&b->urb, xfer->wa->usb_dev, b 514 drivers/staging/wusbcore/wa-xfer.c &b->cmd, sizeof(b->cmd), __wa_xfer_abort_cb, b); b 515 drivers/staging/wusbcore/wa-xfer.c result = usb_submit_urb(&b->urb, GFP_ATOMIC); b 526 drivers/staging/wusbcore/wa-xfer.c kfree(b); b 2301 drivers/target/target_core_configfs.c char *b = page; b 2307 drivers/target/target_core_configfs.c bl += sprintf(b + bl, "%u %u\n", b 2311 drivers/target/target_core_configfs.c bl += sprintf(b + bl, "%llu %llu", b 2332 drivers/target/target_core_configfs.c bl += sprintf(b + bl, " %d:%c", b 2335 drivers/target/target_core_configfs.c bl += sprintf(b + bl, "\n"); b 801 drivers/target/target_core_file.c static ssize_t fd_show_configfs_dev_params(struct se_device *dev, char *b) b 806 drivers/target/target_core_file.c bl = sprintf(b + bl, "TCM FILEIO ID: %u", fd_dev->fd_dev_id); b 807 drivers/target/target_core_file.c bl += sprintf(b + bl, " File: %s Size: %llu Mode: %s Async: %d\n", b 596 drivers/target/target_core_iblock.c static ssize_t iblock_show_configfs_dev_params(struct se_device *dev, char *b) b 604 drivers/target/target_core_iblock.c bl += sprintf(b + bl, "iBlock device: %s", b 607 drivers/target/target_core_iblock.c bl += sprintf(b + bl, " UDEV PATH: %s", b 609 drivers/target/target_core_iblock.c bl += sprintf(b + bl, " readonly: %d\n", ib_dev->ibd_readonly); b 611 drivers/target/target_core_iblock.c bl += sprintf(b + bl, " "); b 613 drivers/target/target_core_iblock.c bl += sprintf(b + bl, "Major: %d Minor: %d %s\n", b 618 drivers/target/target_core_iblock.c bl += sprintf(b + bl, "Major: 0 Minor: 0\n"); b 794 drivers/target/target_core_pscsi.c static ssize_t pscsi_show_configfs_dev_params(struct se_device *dev, char *b) b 807 drivers/target/target_core_pscsi.c bl = sprintf(b, "SCSI Device Bus Location:" b 813 drivers/target/target_core_pscsi.c bl += sprintf(b + bl, " Vendor: %." b 815 drivers/target/target_core_pscsi.c bl += sprintf(b + bl, " Model: %." b 817 drivers/target/target_core_pscsi.c bl += sprintf(b + bl, " Rev: %." b 586 drivers/target/target_core_rd.c static ssize_t rd_show_configfs_dev_params(struct se_device *dev, char *b) b 590 drivers/target/target_core_rd.c ssize_t bl = sprintf(b, "TCM RamDisk ID: %u RamDisk Makeup: rd_mcp\n", b 592 drivers/target/target_core_rd.c bl += sprintf(b + bl, " PAGES/PAGE_SIZE: %u*%lu" b 958 drivers/target/target_core_transport.c char *b, b 961 drivers/target/target_core_transport.c *bl += sprintf(b + *bl, "Status: "); b 963 drivers/target/target_core_transport.c *bl += sprintf(b + *bl, "ACTIVATED"); b 965 drivers/target/target_core_transport.c *bl += sprintf(b + *bl, "DEACTIVATED"); b 967 drivers/target/target_core_transport.c *bl += sprintf(b + *bl, " Max Queue Depth: %d", dev->queue_depth); b 968 drivers/target/target_core_transport.c *bl += sprintf(b + *bl, " SectorSize: %u HwMaxSectors: %u\n", b 971 drivers/target/target_core_transport.c *bl += sprintf(b + *bl, " "); b 2216 drivers/target/target_core_user.c static ssize_t tcmu_show_configfs_dev_params(struct se_device *dev, char *b) b 2221 drivers/target/target_core_user.c bl = sprintf(b + bl, "Config: %s ", b 2223 drivers/target/target_core_user.c bl += sprintf(b + bl, "Size: %llu ", udev->dev_size); b 2224 drivers/target/target_core_user.c bl += sprintf(b + bl, "MaxDataAreaMB: %u\n", b 91 drivers/target/tcm_fc/tfc_conf.c u8 b[8]; b 93 drivers/target/tcm_fc/tfc_conf.c put_unaligned_be64(wwn, b); b 96 drivers/target/tcm_fc/tfc_conf.c b[0], b[1], b[2], b[3], b[4], b[5], b[6], b[7]); b 60 drivers/tee/optee/core.c p->u.value.b = mp->u.value.b; b 186 drivers/tee/optee/core.c mp->u.value.b = p->u.value.b; b 110 drivers/tee/optee/device.c memcpy(sess_arg.uuid, pta_uuid.b, TEE_IOCTL_UUID_LEN); b 139 drivers/tee/optee/optee_msg.h u64 b; b 44 drivers/tee/optee/rpc.c arg->params[0].u.value.b = ts.tv_nsec; b 106 drivers/tee/optee/rpc.c wq_sleep(&optee->wait_queue, arg->params[0].u.value.b); b 109 drivers/tee/optee/rpc.c wq_wakeup(&optee->wait_queue, arg->params[0].u.value.b); b 179 drivers/tee/optee/rpc.c param.u.value.b = sz; b 217 drivers/tee/optee/rpc.c sz = arg->params[0].u.value.b; b 295 drivers/tee/optee/rpc.c param.u.value.b = tee_shm_get_id(shm); b 327 drivers/tee/optee/rpc.c shm = (struct tee_shm *)(unsigned long)arg->params[0].u.value.b; b 275 drivers/tee/optee/supp.c param->u.value.b = 0; b 365 drivers/tee/optee/supp.c p->u.value.b = param[n + num_meta].u.value.b; b 228 drivers/tee/tee_core.c params[n].u.value.b = ip.b; b 251 drivers/tee/tee_core.c if ((ip.a + ip.b) < ip.a || b 252 drivers/tee/tee_core.c (ip.a + ip.b) > shm->size) { b 258 drivers/tee/tee_core.c params[n].u.memref.size = ip.b; b 282 drivers/tee/tee_core.c put_user(p->u.value.b, &up->b) || b 288 drivers/tee/tee_core.c if (put_user((u64)p->u.memref.size, &up->b)) b 475 drivers/tee/tee_core.c ip.b = p->u.value.b; b 481 drivers/tee/tee_core.c ip.b = p->u.memref.size; b 492 drivers/tee/tee_core.c ip.b = 0; b 577 drivers/tee/tee_core.c p->u.value.b = ip.b; b 591 drivers/tee/tee_core.c p->u.memref.size = ip.b; b 376 drivers/thermal/armada_thermal.c s64 sample, b, m; b 387 drivers/thermal/armada_thermal.c b = priv->data->coef_b; b 392 drivers/thermal/armada_thermal.c *temp = div_s64((m * sample) - b, div); b 394 drivers/thermal/armada_thermal.c *temp = div_s64(b - (m * sample), div); b 459 drivers/thermal/armada_thermal.c s64 b = data->coef_b; b 465 drivers/thermal/armada_thermal.c sample = div_s64(((temp_mc * div) + b), m); b 467 drivers/thermal/armada_thermal.c sample = div_s64((b - (temp_mc * div)), m); b 62 drivers/thermal/tegra/soctherm-fuse.c static s64 div64_s64_precise(s64 a, s32 b) b 69 drivers/thermal/tegra/soctherm-fuse.c r = div64_s64(al * 2 + 1, 2 * b); b 290 drivers/thermal/ti-soc-thermal/ti-bandgap.h #define TI_BANDGAP_HAS(b, f) \ b 291 drivers/thermal/ti-soc-thermal/ti-bandgap.h ((b)->conf->features & TI_BANDGAP_FEATURE_ ## f) b 3404 drivers/tty/cyclades.c const struct zfile_block *b, *bs; b 3447 drivers/tty/cyclades.c for (b = bs; b < bs + h->n_blocks; b++) b 3448 drivers/tty/cyclades.c if (b->file_offset + b->size > len) { b 3459 drivers/tty/cyclades.c b = &bs[c->block_list[a]]; b 3460 drivers/tty/cyclades.c if (b->type == ZBLOCK_FPGA) { b 3462 drivers/tty/cyclades.c cyz_fpga_copy(fpga, ptr + b->file_offset, b 3463 drivers/tty/cyclades.c b->size); b 3466 drivers/tty/cyclades.c memcpy_toio(base + b->ram_offset, b 3467 drivers/tty/cyclades.c ptr + b->file_offset, b->size); b 208 drivers/tty/goldfish.c static void goldfish_tty_console_write(struct console *co, const char *b, b 211 drivers/tty/goldfish.c goldfish_tty_do_write(co->index, b, count); b 151 drivers/tty/hvc/hvc_console.c static void hvc_console_print(struct console *co, const char *b, b 168 drivers/tty/hvc/hvc_console.c if (b[n] == '\n' && !donecr) { b 172 drivers/tty/hvc/hvc_console.c c[i++] = b[n++]; b 141 drivers/tty/isicom.c #define isicom_paranoia_check(a, b, c) __isicom_paranoia_check((a), (b), (c)) b 143 drivers/tty/isicom.c #define isicom_paranoia_check(a, b, c) 0 b 2704 drivers/tty/mxser.c unsigned int b, i, m; b 2733 drivers/tty/mxser.c for (m = 0, b = 0; b < MXSER_BOARDS; b++) { b 2734 drivers/tty/mxser.c if (!ioaddr[b]) b 2738 drivers/tty/mxser.c retval = mxser_get_ISA_conf(ioaddr[b], brd); b 2745 drivers/tty/mxser.c brd->info->name, ioaddr[b]); b 191 drivers/tty/n_hdlc.c #define bset(p,b) ((p)[(b) >> 5] |= (1 << ((b) & 0x1f))) b 1963 drivers/tty/n_tty.c unsigned char __user **b, b 1979 drivers/tty/n_tty.c retval = copy_to_user(*b, from, n); b 1989 drivers/tty/n_tty.c *b += n; b 2019 drivers/tty/n_tty.c unsigned char __user **b, b 2062 drivers/tty/n_tty.c ret = tty_copy_to_user(tty, *b, tail, n); b 2065 drivers/tty/n_tty.c *b += n; b 2136 drivers/tty/n_tty.c unsigned char __user *b = buf; b 2182 drivers/tty/n_tty.c if (b != buf) b 2188 drivers/tty/n_tty.c if (put_user(cs, b)) { b 2192 drivers/tty/n_tty.c b++; b 2235 drivers/tty/n_tty.c retval = canon_copy_from_read_buf(tty, &b, &nr); b 2242 drivers/tty/n_tty.c if (packet && b == buf) { b 2243 drivers/tty/n_tty.c if (put_user(TIOCPKT_DATA, b)) { b 2247 drivers/tty/n_tty.c b++; b 2251 drivers/tty/n_tty.c uncopied = copy_from_read_buf(tty, &b, &nr); b 2252 drivers/tty/n_tty.c uncopied += copy_from_read_buf(tty, &b, &nr); b 2261 drivers/tty/n_tty.c if (b - buf >= minimum) b 2273 drivers/tty/n_tty.c if (b - buf) b 2274 drivers/tty/n_tty.c retval = b - buf; b 2304 drivers/tty/n_tty.c const unsigned char *b = buf; b 2333 drivers/tty/n_tty.c ssize_t num = process_output_block(tty, b, nr); b 2340 drivers/tty/n_tty.c b += num; b 2344 drivers/tty/n_tty.c c = *b; b 2347 drivers/tty/n_tty.c b++; nr--; b 2356 drivers/tty/n_tty.c c = tty->ops->write(tty, b, nr); b 2364 drivers/tty/n_tty.c b += c; b 2385 drivers/tty/n_tty.c return (b - buf) ? b - buf : retval; b 1557 drivers/tty/rocket.c const unsigned char *b; b 1581 drivers/tty/rocket.c b = buf; b 1584 drivers/tty/rocket.c sOutStrW(sGetTxRxDataIO(cp), (unsigned short *) b, c / 2); b 1588 drivers/tty/rocket.c sOutB(sGetTxRxDataIO(cp), b[c - 1]); b 1613 drivers/tty/rocket.c b = buf; b 1614 drivers/tty/rocket.c memcpy(info->xmit_buf + info->xmit_head, b, c); b 214 drivers/tty/serial/21285.c unsigned int baud, quot, h_lcr, b; b 232 drivers/tty/serial/21285.c b = port->uartclk / (16 * quot); b 233 drivers/tty/serial/21285.c tty_termios_encode_baud_rate(termios, b, b); b 583 drivers/tty/serial/max3100.c char b[12]; b 598 drivers/tty/serial/max3100.c sprintf(b, "max3100-%d", s->minor); b 599 drivers/tty/serial/max3100.c s->workqueue = create_freezable_workqueue(b); b 258 drivers/tty/serial/sccnxp.c static int sccnxp_update_best_err(int a, int b, int *besterr) b 260 drivers/tty/serial/sccnxp.c int err = abs(a - b); b 516 drivers/tty/synclink.c #define usc_UnlatchRxstatusBits(a,b) usc_OutReg( (a), RCSR, (u16)((b) & RXSTATUS_ALL) ) b 554 drivers/tty/synclink.c #define usc_UnlatchTxstatusBits(a,b) usc_OutReg( (a), TCSR, (u16)((a)->tcsr_value + ((b) & 0x00FF)) ) b 574 drivers/tty/synclink.c #define usc_UnlatchIostatusBits(a,b) usc_OutReg((a),MISR,(u16)((b) & 0xaaa0)) b 575 drivers/tty/synclink.c #define usc_UnlatchMiscstatusBits(a,b) usc_OutReg((a),MISR,(u16)((b) & 0x000f)) b 606 drivers/tty/synclink.c #define usc_EnableInterrupts( a, b ) \ b 607 drivers/tty/synclink.c usc_OutReg( (a), ICR, (u16)((usc_InReg((a),ICR) & 0xff00) + 0xc0 + (b)) ) b 609 drivers/tty/synclink.c #define usc_DisableInterrupts( a, b ) \ b 610 drivers/tty/synclink.c usc_OutReg( (a), ICR, (u16)((usc_InReg((a),ICR) & 0xff00) + 0x80 + (b)) ) b 618 drivers/tty/synclink.c #define usc_ClearIrqPendingBits( a, b ) usc_OutReg( (a), DCCR, 0x40 + (b) ) b 638 drivers/tty/synclink.c #define usc_EnableDmaInterrupts(a,b) \ b 639 drivers/tty/synclink.c usc_OutDmaReg( (a), DICR, (u16)(usc_InDmaReg((a),DICR) | (b)) ) b 641 drivers/tty/synclink.c #define usc_DisableDmaInterrupts(a,b) \ b 642 drivers/tty/synclink.c usc_OutDmaReg( (a), DICR, (u16)(usc_InDmaReg((a),DICR) & ~(b)) ) b 644 drivers/tty/synclink.c #define usc_EnableStatusIrqs(a,b) \ b 645 drivers/tty/synclink.c usc_OutReg( (a), SICR, (u16)(usc_InReg((a),SICR) | (b)) ) b 647 drivers/tty/synclink.c #define usc_DisablestatusIrqs(a,b) \ b 648 drivers/tty/synclink.c usc_OutReg( (a), SICR, (u16)(usc_InReg((a),SICR) & ~(b)) ) b 659 drivers/tty/synclink.c #define usc_EnableTransmitter(a,b) \ b 660 drivers/tty/synclink.c usc_OutReg( (a), TMR, (u16)((usc_InReg((a),TMR) & 0xfffc) | (b)) ) b 661 drivers/tty/synclink.c #define usc_EnableReceiver(a,b) \ b 662 drivers/tty/synclink.c usc_OutReg( (a), RMR, (u16)((usc_InReg((a),RMR) & 0xfffc) | (b)) ) b 674 drivers/tty/synclink.c #define usc_TCmd(a,b) usc_OutReg((a), TCSR, (u16)((a)->tcsr_value + (b))) b 675 drivers/tty/synclink.c #define usc_RCmd(a,b) usc_OutReg((a), RCSR, (b)) b 211 drivers/tty/synclink_gt.c #define set_desc_buffer(a,b) (a).pbuf = cpu_to_le32((unsigned int)(b)) b 212 drivers/tty/synclink_gt.c #define set_desc_next(a,b) (a).next = cpu_to_le32((unsigned int)(b)) b 213 drivers/tty/synclink_gt.c #define set_desc_count(a,b)(a).count = cpu_to_le16((unsigned short)(b)) b 214 drivers/tty/synclink_gt.c #define set_desc_eof(a,b) (a).status = cpu_to_le16((b) ? (le16_to_cpu((a).status) | BIT0) : (le16_to_cpu((a).status) & ~BIT0)) b 215 drivers/tty/synclink_gt.c #define set_desc_status(a, b) (a).status = cpu_to_le16((unsigned short)(b)) b 194 drivers/tty/tty_buffer.c static void tty_buffer_free(struct tty_port *port, struct tty_buffer *b) b 199 drivers/tty/tty_buffer.c WARN_ON(atomic_sub_return(b->size, &buf->mem_used) < 0); b 201 drivers/tty/tty_buffer.c if (b->size > MIN_TTYB_SIZE) b 202 drivers/tty/tty_buffer.c kfree(b); b 203 drivers/tty/tty_buffer.c else if (b->size > 0) b 204 drivers/tty/tty_buffer.c llist_add(&b->free, &buf->free); b 261 drivers/tty/tty_buffer.c struct tty_buffer *b, *n; b 264 drivers/tty/tty_buffer.c b = buf->tail; b 265 drivers/tty/tty_buffer.c if (b->flags & TTYB_NORMAL) b 266 drivers/tty/tty_buffer.c left = 2 * b->size - b->used; b 268 drivers/tty/tty_buffer.c left = b->size - b->used; b 270 drivers/tty/tty_buffer.c change = (b->flags & TTYB_NORMAL) && (~flags & TTYB_NORMAL); b 280 drivers/tty/tty_buffer.c smp_store_release(&b->commit, b->used); b 285 drivers/tty/tty_buffer.c smp_store_release(&b->next, n); b 293 drivers/tty/tty_ioctl.c int tty_termios_hw_change(const struct ktermios *a, const struct ktermios *b) b 295 drivers/tty/tty_ioctl.c if (a->c_ispeed != b->c_ispeed || a->c_ospeed != b->c_ospeed) b 297 drivers/tty/tty_ioctl.c if ((a->c_cflag ^ b->c_cflag) & ~(HUPCL | CREAD | CLOCAL)) b 438 drivers/tty/vt/vt.c static void vc_uniscr_scroll(struct vc_data *vc, unsigned int t, unsigned int b, b 446 drivers/tty/vt/vt.c sz = b - t; b 447 drivers/tty/vt/vt.c clear = b - nr; b 627 drivers/tty/vt/vt.c static void con_scroll(struct vc_data *vc, unsigned int t, unsigned int b, b 632 drivers/tty/vt/vt.c if (t + nr >= b) b 633 drivers/tty/vt/vt.c nr = b - t - 1; b 634 drivers/tty/vt/vt.c if (b > vc->vc_rows || t >= b || nr < 1) b 636 drivers/tty/vt/vt.c vc_uniscr_scroll(vc, t, b, dir, nr); b 637 drivers/tty/vt/vt.c if (con_is_visible(vc) && vc->vc_sw->con_scroll(vc, t, b, dir, nr)) b 644 drivers/tty/vt/vt.c clear = s + (b - t - nr) * vc->vc_cols; b 647 drivers/tty/vt/vt.c scr_memmovew(d, s, (b - t - nr) * vc->vc_size_row); b 1619 drivers/tty/vt/vt.c struct rgb { u8 r; u8 g; u8 b; }; b 1626 drivers/tty/vt/vt.c c->b = i&4 ? 0xaa : 0x00; b 1630 drivers/tty/vt/vt.c c->b = i&4 ? 0xff : 0x55; b 1634 drivers/tty/vt/vt.c c->b = (i - 16) % 6 * 85 / 2; b 1636 drivers/tty/vt/vt.c c->r = c->g = c->b = i * 10 - 2312; b 1641 drivers/tty/vt/vt.c u8 hue = 0, max = max3(c->r, c->g, c->b); b 1647 drivers/tty/vt/vt.c if (c->b > max / 2) b 1665 drivers/tty/vt/vt.c | (c->r&0x80) >> 1 | (c->g&0x80) >> 2 | (c->b&0x80) >> 3; b 1694 drivers/tty/vt/vt.c c.b = vc->vc_par[i + 3]; b 2946 drivers/tty/vt/vt.c static void vt_console_print(struct console *co, const char *b, unsigned count) b 2982 drivers/tty/vt/vt.c c = *b++; b 418 drivers/usb/atm/ueagle-atm.c #define E1_MAKESA(a, b, c, d) \ b 422 drivers/usb/atm/ueagle-atm.c ((b) & 0xff)) b 168 drivers/usb/c67x00/c67x00-sched.c static inline u16 frame_add(u16 a, u16 b) b 170 drivers/usb/c67x00/c67x00-sched.c return (a + b) & HOST_FRAME_MASK; b 176 drivers/usb/c67x00/c67x00-sched.c static inline int frame_after(u16 a, u16 b) b 178 drivers/usb/c67x00/c67x00-sched.c return ((HOST_FRAME_MASK + a - b) & HOST_FRAME_MASK) < b 185 drivers/usb/c67x00/c67x00-sched.c static inline int frame_after_eq(u16 a, u16 b) b 187 drivers/usb/c67x00/c67x00-sched.c return ((HOST_FRAME_MASK + 1 + a - b) & HOST_FRAME_MASK) < b 1118 drivers/usb/dwc2/core.h } b; b 1762 drivers/usb/dwc2/hcd.c hsotg->flags.b.port_connect_status_change = 1; b 1763 drivers/usb/dwc2/hcd.c hsotg->flags.b.port_connect_status = 1; b 1780 drivers/usb/dwc2/hcd.c hsotg->flags.b.port_connect_status_change = 1; b 1781 drivers/usb/dwc2/hcd.c hsotg->flags.b.port_connect_status = 0; b 1844 drivers/usb/dwc2/hcd.c hsotg->flags.b.port_suspend_change = 1; b 1849 drivers/usb/dwc2/hcd.c hsotg->flags.b.port_l1_change = 1; b 1886 drivers/usb/dwc2/hcd.c if (!hsotg->flags.b.port_connect_status) { b 1962 drivers/usb/dwc2/hcd.c if (hsotg->flags.b.port_connect_status) b 3446 drivers/usb/dwc2/hcd.c hsotg->flags.b.port_connect_status_change = 0; b 3453 drivers/usb/dwc2/hcd.c hsotg->flags.b.port_reset_change = 0; b 3463 drivers/usb/dwc2/hcd.c hsotg->flags.b.port_enable_change = 0; b 3474 drivers/usb/dwc2/hcd.c hsotg->flags.b.port_suspend_change = 0; b 3480 drivers/usb/dwc2/hcd.c hsotg->flags.b.port_l1_change = 0; b 3486 drivers/usb/dwc2/hcd.c hsotg->flags.b.port_over_current_change = 0; b 3525 drivers/usb/dwc2/hcd.c if (hsotg->flags.b.port_connect_status_change) b 3527 drivers/usb/dwc2/hcd.c if (hsotg->flags.b.port_enable_change) b 3529 drivers/usb/dwc2/hcd.c if (hsotg->flags.b.port_suspend_change) b 3531 drivers/usb/dwc2/hcd.c if (hsotg->flags.b.port_l1_change) b 3533 drivers/usb/dwc2/hcd.c if (hsotg->flags.b.port_reset_change) b 3535 drivers/usb/dwc2/hcd.c if (hsotg->flags.b.port_over_current_change) { b 3540 drivers/usb/dwc2/hcd.c if (!hsotg->flags.b.port_connect_status) { b 3614 drivers/usb/dwc2/hcd.c if (!hsotg->flags.b.port_connect_status) { b 3730 drivers/usb/dwc2/hcd.c retval = (hsotg->flags.b.port_connect_status_change || b 3731 drivers/usb/dwc2/hcd.c hsotg->flags.b.port_reset_change || b 3732 drivers/usb/dwc2/hcd.c hsotg->flags.b.port_enable_change || b 3733 drivers/usb/dwc2/hcd.c hsotg->flags.b.port_suspend_change || b 3734 drivers/usb/dwc2/hcd.c hsotg->flags.b.port_over_current_change); b 3740 drivers/usb/dwc2/hcd.c hsotg->flags.b.port_connect_status_change); b 3742 drivers/usb/dwc2/hcd.c hsotg->flags.b.port_reset_change); b 3744 drivers/usb/dwc2/hcd.c hsotg->flags.b.port_enable_change); b 3746 drivers/usb/dwc2/hcd.c hsotg->flags.b.port_suspend_change); b 3748 drivers/usb/dwc2/hcd.c hsotg->flags.b.port_over_current_change); b 4197 drivers/usb/dwc2/hcd.c hsotg->flags.b.port_reset_change = 1; b 4434 drivers/usb/dwc2/hcd.c hsotg->flags.b.port_suspend_change = 1; b 4838 drivers/usb/dwc2/hcd.c if (hsotg->flags.b.port_connect_status) b 286 drivers/usb/dwc2/hcd_intr.c hsotg->flags.b.port_reset_change = 1; b 346 drivers/usb/dwc2/hcd_intr.c hsotg->flags.b.port_reset_change = 1; b 403 drivers/usb/dwc2/hcd_intr.c hsotg->flags.b.port_enable_change = 1; b 423 drivers/usb/dwc2/hcd_intr.c hsotg->flags.b.port_over_current_change = 1; b 107 drivers/usb/early/xhci-dbc.c static u32 __init xdbc_find_dbgp(int xdbc_num, u32 *b, u32 *d, u32 *f) b 122 drivers/usb/early/xhci-dbc.c *b = bus; b 1129 drivers/usb/gadget/composite.c struct usb_os_string *b = buf; b 1130 drivers/usb/gadget/composite.c b->bLength = sizeof(*b); b 1131 drivers/usb/gadget/composite.c b->bDescriptorType = USB_DT_STRING; b 1133 drivers/usb/gadget/composite.c sizeof(b->qwSignature) == sizeof(cdev->qw_sign), b 1135 drivers/usb/gadget/composite.c memcpy(&b->qwSignature, cdev->qw_sign, sizeof(b->qwSignature)); b 1136 drivers/usb/gadget/composite.c b->bMS_VendorCode = cdev->b_vendor_code; b 1137 drivers/usb/gadget/composite.c b->bPad = 0; b 1138 drivers/usb/gadget/composite.c return sizeof(*b); b 448 drivers/usb/gadget/function/f_midi.c struct gmidi_in_port *port, uint8_t b) b 453 drivers/usb/gadget/function/f_midi.c switch (b) { b 457 drivers/usb/gadget/function/f_midi.c p[1] = b; b 494 drivers/usb/gadget/function/f_midi.c switch (b) { b 496 drivers/usb/gadget/function/f_midi.c port->data[0] = b; b 502 drivers/usb/gadget/function/f_midi.c port->data[0] = b; b 506 drivers/usb/gadget/function/f_midi.c port->data[0] = b; b 526 drivers/usb/gadget/function/f_midi.c port->data[0] = b; b 529 drivers/usb/gadget/function/f_midi.c if (b >= 0xc0 && b <= 0xdf) b 545 drivers/usb/gadget/function/f_midi.c p[2] = b; b 550 drivers/usb/gadget/function/f_midi.c port->data[1] = b; b 561 drivers/usb/gadget/function/f_midi.c p[3] = b; b 566 drivers/usb/gadget/function/f_midi.c port->data[0] = b; b 570 drivers/usb/gadget/function/f_midi.c port->data[1] = b; b 577 drivers/usb/gadget/function/f_midi.c p[3] = b; b 640 drivers/usb/gadget/function/f_midi.c uint8_t b; b 642 drivers/usb/gadget/function/f_midi.c if (snd_rawmidi_transmit(substream, &b, 1) != 1) { b 646 drivers/usb/gadget/function/f_midi.c f_midi_transmit_byte(req, port, b); b 138 drivers/usb/gadget/function/uvc_v4l2.c uvc_v4l2_reqbufs(struct file *file, void *fh, struct v4l2_requestbuffers *b) b 144 drivers/usb/gadget/function/uvc_v4l2.c if (b->type != video->queue.queue.type) b 147 drivers/usb/gadget/function/uvc_v4l2.c return uvcg_alloc_buffers(&video->queue, b); b 151 drivers/usb/gadget/function/uvc_v4l2.c uvc_v4l2_querybuf(struct file *file, void *fh, struct v4l2_buffer *b) b 157 drivers/usb/gadget/function/uvc_v4l2.c return uvcg_query_buffer(&video->queue, b); b 161 drivers/usb/gadget/function/uvc_v4l2.c uvc_v4l2_qbuf(struct file *file, void *fh, struct v4l2_buffer *b) b 168 drivers/usb/gadget/function/uvc_v4l2.c ret = uvcg_queue_buffer(&video->queue, b); b 176 drivers/usb/gadget/function/uvc_v4l2.c uvc_v4l2_dqbuf(struct file *file, void *fh, struct v4l2_buffer *b) b 182 drivers/usb/gadget/function/uvc_v4l2.c return uvcg_dequeue_buffer(&video->queue, b, file->f_flags & O_NONBLOCK); b 1900 drivers/usb/gadget/udc/net2280.c #define device_create_file(a, b) (0) b 1901 drivers/usb/gadget/udc/net2280.c #define device_remove_file(a, b) do { } while (0) b 206 drivers/usb/gadget/udc/s3c2410_udc.c inline void s3c2410_udc_set_ep0_ss(void __iomem *b) b 208 drivers/usb/gadget/udc/s3c2410_udc.c udc_writeb(b, S3C2410_UDC_INDEX_EP0, S3C2410_UDC_INDEX_REG); b 209 drivers/usb/gadget/udc/s3c2410_udc.c udc_writeb(b, S3C2410_UDC_EP0_CSR_SENDSTL, S3C2410_UDC_EP0_CSR_REG); b 692 drivers/usb/host/ohci-hcd.c &ohci->regs->roothub.b); b 75 drivers/usb/host/ohci-pci.c struct pci_dev *b; b 77 drivers/usb/host/ohci-pci.c b = pci_get_slot (pdev->bus, PCI_DEVFN (PCI_SLOT (pdev->devfn), 1)); b 78 drivers/usb/host/ohci-pci.c if (b && b->device == PCI_DEVICE_ID_NS_87560_LIO b 79 drivers/usb/host/ohci-pci.c && b->vendor == PCI_VENDOR_ID_NS) { b 85 drivers/usb/host/ohci-pci.c pci_dev_put(b); b 31 drivers/usb/host/ohci-ps3.c ohci_writel(ohci, 0x00060000, &ohci->regs->roothub.b); b 238 drivers/usb/host/ohci.h __hc32 b; b 721 drivers/usb/host/ohci.h { return ohci_readl (hc, &hc->regs->roothub.b); } b 647 drivers/usb/host/u132-hcd.c u8 *b = buf; b 651 drivers/usb/host/u132-hcd.c *u++ = *b++; b 798 drivers/usb/host/u132-hcd.c u8 *b = buf; b 802 drivers/usb/host/u132-hcd.c *u++ = *b++; b 937 drivers/usb/host/u132-hcd.c u8 *b = buf; b 941 drivers/usb/host/u132-hcd.c *u++ = *b++; b 1226 drivers/usb/host/u132-hcd.c u8 *b = buf; b 1230 drivers/usb/host/u132-hcd.c *u++ = *b++; b 1757 drivers/usb/host/u132-hcd.c retval = u132_write_pcimem(u132, roothub.b, b 2336 drivers/usb/host/u132-hcd.c u8 *b = urb->setup_packet; b 2346 drivers/usb/host/u132-hcd.c int w = sprintf(d, " %02X", *b++); b 2589 drivers/usb/host/u132-hcd.c retval = u132_read_pcimem(u132, roothub.b, &rh_b); b 164 drivers/usb/host/xhci-dbgcap.h #define dbc_epctx_info2(t, p, b) \ b 165 drivers/usb/host/xhci-dbgcap.h cpu_to_le32(EP_TYPE(t) | MAX_PACKET(p) | MAX_BURST(b)) b 223 drivers/usb/image/mdc800.c static int mdc800_endpoint_equals (struct usb_endpoint_descriptor *a,struct usb_endpoint_descriptor *b) b 226 drivers/usb/image/mdc800.c ( a->bEndpointAddress == b->bEndpointAddress ) b 227 drivers/usb/image/mdc800.c && ( a->bmAttributes == b->bmAttributes ) b 228 drivers/usb/image/mdc800.c && ( a->wMaxPacketSize == b->wMaxPacketSize ) b 272 drivers/usb/image/mdc800.c unsigned char* b=urb->transfer_buffer; b 278 drivers/usb/image/mdc800.c if (mdc800_isBusy (b)) b 288 drivers/usb/image/mdc800.c if (mdc800->camera_busy && mdc800_isReady (b)) b 294 drivers/usb/image/mdc800.c if (!(mdc800_isBusy (b) || mdc800_isReady (b))) b 297 drivers/usb/image/mdc800.c dev_dbg(dev, "%i %i %i %i %i %i %i %i \n",b[0],b[1],b[2],b[3],b[4],b[5],b[6],b[7]); b 299 drivers/usb/image/mdc800.c memcpy (mdc800->camera_response,b,8); b 734 drivers/usb/misc/ftdi-elan.c int b = 0; b 745 drivers/usb/misc/ftdi-elan.c buf[b++] = command->header; b 746 drivers/usb/misc/ftdi-elan.c buf[b++] = (command->length >> 0) & 0x00FF; b 747 drivers/usb/misc/ftdi-elan.c buf[b++] = (command->length >> 8) & 0x00FF; b 748 drivers/usb/misc/ftdi-elan.c buf[b++] = command->address; b 749 drivers/usb/misc/ftdi-elan.c buf[b++] = command->width; b 751 drivers/usb/misc/ftdi-elan.c buf[b++] = *f++; b 849 drivers/usb/misc/ftdi-elan.c char *b) b 870 drivers/usb/misc/ftdi-elan.c return b; b 886 drivers/usb/misc/ftdi-elan.c char *b) b 914 drivers/usb/misc/ftdi-elan.c u8 *b = ftdi->response + ftdi->received; b 979 drivers/usb/misc/ftdi-elan.c *b++ = c; b 1005 drivers/usb/misc/ftdi-elan.c b = ftdi->response; b 1025 drivers/usb/misc/ftdi-elan.c b = ftdi->response; b 1057 drivers/usb/misc/ftdi-elan.c b = have_ed_set_response(ftdi, target, b 1059 drivers/usb/misc/ftdi-elan.c b); b 1062 drivers/usb/misc/ftdi-elan.c b = have_ed_get_response(ftdi, target, b 1064 drivers/usb/misc/ftdi-elan.c b); b 1621 drivers/usb/misc/ftdi-elan.c u8 *b; b 1643 drivers/usb/misc/ftdi-elan.c b = command->buffer; b 1649 drivers/usb/misc/ftdi-elan.c int w = sprintf(d, " %02X", *b++); b 1819 drivers/usb/misc/ftdi-elan.c char *b = ftdi->bulk_in_buffer; b 1823 drivers/usb/misc/ftdi-elan.c char c = *b++; b 1992 drivers/usb/misc/ftdi-elan.c char *b = ftdi->bulk_in_buffer; b 1997 drivers/usb/misc/ftdi-elan.c c = *b++; b 2087 drivers/usb/misc/ftdi-elan.c char *b = ftdi->bulk_in_buffer; b 2091 drivers/usb/misc/ftdi-elan.c char c = *b++; b 2373 drivers/usb/misc/ftdi-elan.c retval = ftdi_write_pcimem(ftdi, roothub.b, b 766 drivers/usb/misc/sisusbvga/sisusb_con.c unsigned int t, unsigned int b, enum con_scroll dir, b 770 drivers/usb/misc/sisusbvga/sisusb_con.c int length = ((b - t) * cols) * 2; b 786 drivers/usb/misc/sisusbvga/sisusb_con.c (b - t - lines) * cols * 2); b 787 drivers/usb/misc/sisusbvga/sisusb_con.c sisusbcon_memsetw(sisusb_vaddr(sisusb, c, 0, b - lines), b 794 drivers/usb/misc/sisusbvga/sisusb_con.c (b - t - lines) * cols * 2); b 810 drivers/usb/misc/sisusbvga/sisusb_con.c sisusbcon_scroll(struct vc_data *c, unsigned int t, unsigned int b, b 841 drivers/usb/misc/sisusbvga/sisusb_con.c if (t || b != c->vc_rows) b 842 drivers/usb/misc/sisusbvga/sisusb_con.c return sisusbcon_scroll_area(c, sisusb, t, b, dir, lines); b 111 drivers/usb/renesas_usbhs/mod_gadget.c #define usbhsg_status_set(gp, b) (gp->status |= b) b 112 drivers/usb/renesas_usbhs/mod_gadget.c #define usbhsg_status_clr(gp, b) (gp->status &= ~b) b 113 drivers/usb/renesas_usbhs/mod_gadget.c #define usbhsg_status_has(gp, b) (gp->status & b) b 71 drivers/usb/serial/belkin_sa.h #define BELKIN_SA_BAUD(b) (230400/b) b 73 drivers/usb/serial/belkin_sa.h #define BELKIN_SA_STOP_BITS(b) (b-1) b 75 drivers/usb/serial/belkin_sa.h #define BELKIN_SA_DATA_BITS(b) (b-5) b 40 drivers/usb/serial/garmin_gps.c #define _DRIVER_VERSION(a, b) "v" _STR(a) "." _STR(b) b 970 drivers/usb/serial/iuu_phoenix.c #define SOUP(a, b, c, d) do { \ b 973 drivers/usb/serial/iuu_phoenix.c b, a, c, d, NULL, 0, 1000); \ b 974 drivers/usb/serial/iuu_phoenix.c dev_dbg(dev, "0x%x:0x%x:0x%x:0x%x %d\n", a, b, c, d, result); } while (0) b 245 drivers/usb/serial/keyspan_usa26msg.h b, b 188 drivers/usb/serial/keyspan_usa28msg.h b; // typically a data byte b 271 drivers/usb/serial/keyspan_usa49msg.h b; // typically a data byte b 239 drivers/usb/serial/keyspan_usa67msg.h b, b 573 drivers/usb/serial/pl2303.c static bool pl2303_termios_change(const struct ktermios *a, const struct ktermios *b) b 577 drivers/usb/serial/pl2303.c ixon_change = ((a->c_iflag ^ b->c_iflag) & (IXON | IXANY)) || b 578 drivers/usb/serial/pl2303.c a->c_cc[VSTART] != b->c_cc[VSTART] || b 579 drivers/usb/serial/pl2303.c a->c_cc[VSTOP] != b->c_cc[VSTOP]; b 581 drivers/usb/serial/pl2303.c return tty_termios_hw_change(a, b) || ixon_change; b 122 drivers/video/backlight/ili922x.c #define set_tx_byte(b) (tx_invert ? ~(b) : b) b 45 drivers/video/backlight/sky81452-backlight.c #define CTZ(b) __builtin_ctz(b) b 504 drivers/video/console/mdacon.c static bool mdacon_scroll(struct vc_data *c, unsigned int t, unsigned int b, b 519 drivers/video/console/mdacon.c (b-t-lines)*mda_num_columns*2); b 520 drivers/video/console/mdacon.c scr_memsetw(mda_addr(0, b - lines), eattr, b 526 drivers/video/console/mdacon.c (b-t-lines)*mda_num_columns*2); b 578 drivers/video/console/newport_con.c static bool newport_scroll(struct vc_data *vc, unsigned int t, unsigned int b, b 587 drivers/video/console/newport_con.c if (t == 0 && b == vc->vc_rows) { b 602 drivers/video/console/newport_con.c count = (b - t - lines) * vc->vc_cols; b 623 drivers/video/console/newport_con.c vc->vc_size_row * (b - lines)); b 625 drivers/video/console/newport_con.c y = b - lines; b 640 drivers/video/console/newport_con.c y = b - 1; b 642 drivers/video/console/newport_con.c vc->vc_size_row * (b - lines) - 2); b 644 drivers/video/console/newport_con.c vc->vc_size_row * b - 2); b 157 drivers/video/console/sticon.c unsigned int b, enum con_scroll dir, unsigned int count) b 168 drivers/video/console/sticon.c sti_bmove(sti, t + count, 0, t, 0, b - t - count, conp->vc_cols); b 169 drivers/video/console/sticon.c sti_clear(sti, b - count, 0, count, conp->vc_cols, conp->vc_video_erase_char); b 173 drivers/video/console/sticon.c sti_bmove(sti, t, 0, t + count, 0, b - t - count, conp->vc_cols); b 1367 drivers/video/console/vgacon.c static bool vgacon_scroll(struct vc_data *c, unsigned int t, unsigned int b, b 1373 drivers/video/console/vgacon.c if (t || b != c->vc_rows || vga_is_gfx || c->vc_mode != KD_TEXT) b 2580 drivers/video/fbdev/amifb.c static inline unsigned long comp(unsigned long a, unsigned long b, b 2583 drivers/video/fbdev/amifb.c return ((a ^ b) & mask) ^ b; b 2587 drivers/video/fbdev/amifb.c static inline unsigned long xor(unsigned long a, unsigned long b, b 2590 drivers/video/fbdev/amifb.c return (a & mask) ^ b; b 1438 drivers/video/fbdev/aty/aty128fb.c s32 x, b, p, ron, roff; b 1458 drivers/video/fbdev/aty/aty128fb.c b = 0; b 1461 drivers/video/fbdev/aty/aty128fb.c b++; b 1463 drivers/video/fbdev/aty/aty128fb.c p = b + 1; b 153 drivers/video/fbdev/aty/mach64_cursor.c u8 m, b; b 163 drivers/video/fbdev/aty/mach64_cursor.c b = *src++; b 168 drivers/video/fbdev/aty/mach64_cursor.c l = cursor_bits_lookup[(b ^ m) >> 4] | b 170 drivers/video/fbdev/aty/mach64_cursor.c (cursor_bits_lookup[(b ^ m) & 0x0f] << 8); b 174 drivers/video/fbdev/aty/mach64_cursor.c l = cursor_bits_lookup[(b & m) >> 4] | b 176 drivers/video/fbdev/aty/mach64_cursor.c (cursor_bits_lookup[(b & m) & 0x0f] << 8); b 150 drivers/video/fbdev/c2p_core.h static inline u32 comp(u32 a, u32 b, u32 mask) b 152 drivers/video/fbdev/c2p_core.h return ((a ^ b) & mask) ^ b; b 348 drivers/video/fbdev/controlfb.c __u8 r, g, b; b 355 drivers/video/fbdev/controlfb.c b = blue >> 8; b 360 drivers/video/fbdev/controlfb.c out_8(&p->cmap_regs->lut, b); b 15 drivers/video/fbdev/core/fb_draw.h comp(unsigned long a, unsigned long b, unsigned long mask) b 17 drivers/video/fbdev/core/fb_draw.h return ((a ^ b) & mask) ^ b; b 168 drivers/video/fbdev/core/fb_draw.h #define fb_shifted_pixels_mask_u32(p, i, b) FB_SHIFT_HIGH((p), ~(u32)0, (i)) b 169 drivers/video/fbdev/core/fb_draw.h #define fb_shifted_pixels_mask_long(p, i, b) FB_SHIFT_HIGH((p), ~0UL, (i)) b 178 drivers/video/fbdev/core/fbcon.c #define divides(a, b) ((!(a) || (b)%(a)) ? 0 : 1) b 1340 drivers/video/fbdev/core/fbcon.c u_int b = y_break - sy; b 1341 drivers/video/fbdev/core/fbcon.c ops->clear(vc, info, real_y(p, sy), sx, b, width); b 1342 drivers/video/fbdev/core/fbcon.c ops->clear(vc, info, real_y(p, sy + b), sx, height - b, b 1872 drivers/video/fbdev/core/fbcon.c static bool fbcon_scroll(struct vc_data *vc, unsigned int t, unsigned int b, b 1900 drivers/video/fbdev/core/fbcon.c fbcon_redraw_blit(vc, info, p, t, b - t - count, b 1902 drivers/video/fbdev/core/fbcon.c fbcon_clear(vc, b - count, 0, count, vc->vc_cols); b 1905 drivers/video/fbdev/core/fbcon.c (b - count)), b 1912 drivers/video/fbdev/core/fbcon.c if (b - t - count > 3 * vc->vc_rows >> 2) { b 1917 drivers/video/fbdev/core/fbcon.c if (vc->vc_rows - b > 0) b 1918 drivers/video/fbdev/core/fbcon.c fbcon_bmove(vc, b - count, 0, b, 0, b 1919 drivers/video/fbdev/core/fbcon.c vc->vc_rows - b, b 1923 drivers/video/fbdev/core/fbcon.c b - t - count, vc->vc_cols); b 1926 drivers/video/fbdev/core/fbcon.c fbcon_clear(vc, b - count, 0, count, vc->vc_cols); b 1932 drivers/video/fbdev/core/fbcon.c && ((!scroll_partial && (b - t == vc->vc_rows)) b 1934 drivers/video/fbdev/core/fbcon.c && (b - t - count > b 1939 drivers/video/fbdev/core/fbcon.c if (vc->vc_rows - b > 0) b 1940 drivers/video/fbdev/core/fbcon.c fbcon_redraw_move(vc, p, b, b 1941 drivers/video/fbdev/core/fbcon.c vc->vc_rows - b, b); b 1943 drivers/video/fbdev/core/fbcon.c fbcon_redraw_move(vc, p, t + count, b - t - count, t); b 1944 drivers/video/fbdev/core/fbcon.c fbcon_clear(vc, b - count, 0, count, vc->vc_cols); b 1950 drivers/video/fbdev/core/fbcon.c && ((!scroll_partial && (b - t == vc->vc_rows)) b 1952 drivers/video/fbdev/core/fbcon.c && (b - t - count > b 1958 drivers/video/fbdev/core/fbcon.c if (vc->vc_rows - b > 0) b 1959 drivers/video/fbdev/core/fbcon.c fbcon_bmove(vc, b - count, 0, b, 0, b 1960 drivers/video/fbdev/core/fbcon.c vc->vc_rows - b, b 1964 drivers/video/fbdev/core/fbcon.c b - t - count, vc->vc_cols); b 1967 drivers/video/fbdev/core/fbcon.c fbcon_clear(vc, b - count, 0, count, vc->vc_cols); b 1972 drivers/video/fbdev/core/fbcon.c fbcon_redraw(vc, p, t, b - t - count, b 1974 drivers/video/fbdev/core/fbcon.c fbcon_clear(vc, b - count, 0, count, vc->vc_cols); b 1977 drivers/video/fbdev/core/fbcon.c (b - count)), b 1991 drivers/video/fbdev/core/fbcon.c fbcon_redraw_blit(vc, info, p, b - 1, b - t - count, b 2003 drivers/video/fbdev/core/fbcon.c if (b - t - count > 3 * vc->vc_rows >> 2) { b 2004 drivers/video/fbdev/core/fbcon.c if (vc->vc_rows - b > 0) b 2005 drivers/video/fbdev/core/fbcon.c fbcon_bmove(vc, b, 0, b - count, 0, b 2006 drivers/video/fbdev/core/fbcon.c vc->vc_rows - b, b 2014 drivers/video/fbdev/core/fbcon.c b - t - count, vc->vc_cols); b 2022 drivers/video/fbdev/core/fbcon.c && ((!scroll_partial && (b - t == vc->vc_rows)) b 2024 drivers/video/fbdev/core/fbcon.c && (b - t - count > b 2026 drivers/video/fbdev/core/fbcon.c if (vc->vc_rows - b > 0) b 2027 drivers/video/fbdev/core/fbcon.c fbcon_bmove(vc, b, 0, b - count, 0, b 2028 drivers/video/fbdev/core/fbcon.c vc->vc_rows - b, b 2036 drivers/video/fbdev/core/fbcon.c b - t - count, vc->vc_cols); b 2044 drivers/video/fbdev/core/fbcon.c && ((!scroll_partial && (b - t == vc->vc_rows)) b 2046 drivers/video/fbdev/core/fbcon.c && (b - t - count > b 2048 drivers/video/fbdev/core/fbcon.c if (vc->vc_rows - b > 0) b 2049 drivers/video/fbdev/core/fbcon.c fbcon_redraw_move(vc, p, b, vc->vc_rows - b, b 2050 drivers/video/fbdev/core/fbcon.c b - count); b 2055 drivers/video/fbdev/core/fbcon.c fbcon_redraw_move(vc, p, t, b - t - count, t + count); b 2061 drivers/video/fbdev/core/fbcon.c fbcon_redraw(vc, p, b - 1, b - t - count, b 2104 drivers/video/fbdev/core/fbcon.c u_int b; b 2107 drivers/video/fbdev/core/fbcon.c b = y_break - sy; b 2109 drivers/video/fbdev/core/fbcon.c fbcon_bmove_rec(vc, p, sy, sx, dy, dx, b, width, b 2111 drivers/video/fbdev/core/fbcon.c fbcon_bmove_rec(vc, p, sy + b, sx, dy + b, dx, b 2112 drivers/video/fbdev/core/fbcon.c height - b, width, y_break); b 2114 drivers/video/fbdev/core/fbcon.c fbcon_bmove_rec(vc, p, sy + b, sx, dy + b, dx, b 2115 drivers/video/fbdev/core/fbcon.c height - b, width, y_break); b 2116 drivers/video/fbdev/core/fbcon.c fbcon_bmove_rec(vc, p, sy, sx, dy, dx, b, width, b 2123 drivers/video/fbdev/core/fbcon.c b = y_break - dy; b 2125 drivers/video/fbdev/core/fbcon.c fbcon_bmove_rec(vc, p, sy, sx, dy, dx, b, width, b 2127 drivers/video/fbdev/core/fbcon.c fbcon_bmove_rec(vc, p, sy + b, sx, dy + b, dx, b 2128 drivers/video/fbdev/core/fbcon.c height - b, width, y_break); b 2130 drivers/video/fbdev/core/fbcon.c fbcon_bmove_rec(vc, p, sy + b, sx, dy + b, dx, b 2131 drivers/video/fbdev/core/fbcon.c height - b, width, y_break); b 2132 drivers/video/fbdev/core/fbcon.c fbcon_bmove_rec(vc, p, sy, sx, dy, dx, b, width, b 148 drivers/video/fbdev/core/fbmon.c unsigned char *b; b 177 drivers/video/fbdev/core/fbmon.c b = edid + EDID_STRUCT_DISPLAY; b 180 drivers/video/fbdev/core/fbmon.c if (b[4] & 0x01 && b[0] & 0x80) b 184 drivers/video/fbdev/core/fbmon.c b = edid + DETAILED_TIMING_DESCRIPTIONS_START; b 188 drivers/video/fbdev/core/fbmon.c if (edid_is_limits_block(b)) { b 193 drivers/video/fbdev/core/fbmon.c b += DETAILED_TIMING_DESCRIPTION_SIZE; b 210 drivers/video/fbdev/core/fbmon.c unsigned char *b, csum = 0; b 219 drivers/video/fbdev/core/fbmon.c b = edid + EDID_STRUCT_DISPLAY; b 220 drivers/video/fbdev/core/fbmon.c b[0] &= ~0x80; b 225 drivers/video/fbdev/core/fbmon.c b = edid + DETAILED_TIMING_DESCRIPTIONS_START; b 227 drivers/video/fbdev/core/fbmon.c if (!(edid_is_serial_block(b) || b 228 drivers/video/fbdev/core/fbmon.c edid_is_ascii_block(b) || b 229 drivers/video/fbdev/core/fbmon.c edid_is_monitor_block(b) || b 230 drivers/video/fbdev/core/fbmon.c edid_is_timing_block(b))) { b 231 drivers/video/fbdev/core/fbmon.c b[0] = 0x00; b 232 drivers/video/fbdev/core/fbmon.c b[1] = 0x00; b 233 drivers/video/fbdev/core/fbmon.c b[2] = 0x00; b 234 drivers/video/fbdev/core/fbmon.c b[3] = 0xfd; b 235 drivers/video/fbdev/core/fbmon.c b[4] = 0x00; b 236 drivers/video/fbdev/core/fbmon.c b[5] = 60; /* vfmin */ b 237 drivers/video/fbdev/core/fbmon.c b[6] = 60; /* vfmax */ b 238 drivers/video/fbdev/core/fbmon.c b[7] = 30; /* hfmin */ b 239 drivers/video/fbdev/core/fbmon.c b[8] = 75; /* hfmax */ b 240 drivers/video/fbdev/core/fbmon.c b[9] = 17; /* pixclock - 170 MHz*/ b 241 drivers/video/fbdev/core/fbmon.c b[10] = 0; /* GTF */ b 245 drivers/video/fbdev/core/fbmon.c b += DETAILED_TIMING_DESCRIPTION_SIZE; b 376 drivers/video/fbdev/core/svgalib.c static inline u32 abs_diff(u32 a, u32 b) b 378 drivers/video/fbdev/core/svgalib.c return (a > b) ? (a - b) : (b - a); b 103 drivers/video/fbdev/efifb.c u8 r, g, b; b 106 drivers/video/fbdev/efifb.c b = *src++; b 111 drivers/video/fbdev/efifb.c (b << si->blue_pos); b 864 drivers/video/fbdev/i740fb.c u32 r, g, b; b 883 drivers/video/fbdev/i740fb.c b = (blue >> (16 - info->var.blue.length)) b 887 drivers/video/fbdev/i740fb.c ((u32 *) info->pseudo_palette)[regno] = r | g | b; b 1312 drivers/video/fbdev/i810/i810_main.c u8 r, g, b; b 1320 drivers/video/fbdev/i810/i810_main.c i810_read_dac((u8) (regno*4), &r, &g, &b, mmio); b 1323 drivers/video/fbdev/i810/i810_main.c b, mmio); b 73 drivers/video/fbdev/matrox/i2c-matroxfb.c struct i2c_bit_adapter* b = data; b 74 drivers/video/fbdev/matrox/i2c-matroxfb.c matroxfb_i2c_set(b->minfo, b->mask.data, state); b 78 drivers/video/fbdev/matrox/i2c-matroxfb.c struct i2c_bit_adapter* b = data; b 79 drivers/video/fbdev/matrox/i2c-matroxfb.c matroxfb_i2c_set(b->minfo, b->mask.clock, state); b 83 drivers/video/fbdev/matrox/i2c-matroxfb.c struct i2c_bit_adapter* b = data; b 84 drivers/video/fbdev/matrox/i2c-matroxfb.c return (matroxfb_read_gpio(b->minfo) & b->mask.data) ? 1 : 0; b 88 drivers/video/fbdev/matrox/i2c-matroxfb.c struct i2c_bit_adapter* b = data; b 89 drivers/video/fbdev/matrox/i2c-matroxfb.c return (matroxfb_read_gpio(b->minfo) & b->mask.clock) ? 1 : 0; b 102 drivers/video/fbdev/matrox/i2c-matroxfb.c static int i2c_bus_reg(struct i2c_bit_adapter* b, struct matrox_fb_info* minfo, b 108 drivers/video/fbdev/matrox/i2c-matroxfb.c b->minfo = minfo; b 109 drivers/video/fbdev/matrox/i2c-matroxfb.c b->mask.data = data; b 110 drivers/video/fbdev/matrox/i2c-matroxfb.c b->mask.clock = clock; b 111 drivers/video/fbdev/matrox/i2c-matroxfb.c b->adapter.owner = THIS_MODULE; b 112 drivers/video/fbdev/matrox/i2c-matroxfb.c snprintf(b->adapter.name, sizeof(b->adapter.name), name, b 114 drivers/video/fbdev/matrox/i2c-matroxfb.c i2c_set_adapdata(&b->adapter, b); b 115 drivers/video/fbdev/matrox/i2c-matroxfb.c b->adapter.class = class; b 116 drivers/video/fbdev/matrox/i2c-matroxfb.c b->adapter.algo_data = &b->bac; b 117 drivers/video/fbdev/matrox/i2c-matroxfb.c b->adapter.dev.parent = &minfo->pcidev->dev; b 118 drivers/video/fbdev/matrox/i2c-matroxfb.c b->bac = matrox_i2c_algo_template; b 119 drivers/video/fbdev/matrox/i2c-matroxfb.c b->bac.data = b; b 120 drivers/video/fbdev/matrox/i2c-matroxfb.c err = i2c_bit_add_bus(&b->adapter); b 121 drivers/video/fbdev/matrox/i2c-matroxfb.c b->initialized = !err; b 125 drivers/video/fbdev/matrox/i2c-matroxfb.c static void i2c_bit_bus_del(struct i2c_bit_adapter* b) { b 126 drivers/video/fbdev/matrox/i2c-matroxfb.c if (b->initialized) { b 127 drivers/video/fbdev/matrox/i2c-matroxfb.c i2c_del_adapter(&b->adapter); b 128 drivers/video/fbdev/matrox/i2c-matroxfb.c b->initialized = 0; b 982 drivers/video/fbdev/matrox/matroxfb_DAC1064.c u_int8_t b; b 992 drivers/video/fbdev/matrox/matroxfb_DAC1064.c pci_read_config_byte(ibm, PCI_SECONDARY_BUS, &b); b 993 drivers/video/fbdev/matrox/matroxfb_DAC1064.c if (b == minfo->pcidev->bus->number) { b 1028 drivers/video/fbdev/matrox/matroxfb_DAC1064.c b = inDAC1064(minfo, M1064_XGENIODATA) & ~1; b 1029 drivers/video/fbdev/matrox/matroxfb_DAC1064.c outDAC1064(minfo, M1064_XGENIODATA, b); b 1030 drivers/video/fbdev/matrox/matroxfb_DAC1064.c b = inDAC1064(minfo, M1064_XGENIOCTRL) | 1; b 1031 drivers/video/fbdev/matrox/matroxfb_DAC1064.c outDAC1064(minfo, M1064_XGENIOCTRL, b); b 1620 drivers/video/fbdev/matrox/matroxfb_base.c static int initMatrox2(struct matrox_fb_info *minfo, struct board *b) b 1637 drivers/video/fbdev/matrox/matroxfb_base.c minfo->hw_switch = b->base->lowlevel; b 1638 drivers/video/fbdev/matrox/matroxfb_base.c minfo->devflags.accelerator = b->base->accelID; b 1639 drivers/video/fbdev/matrox/matroxfb_base.c minfo->max_pixel_clock = b->maxclk; b 1641 drivers/video/fbdev/matrox/matroxfb_base.c printk(KERN_INFO "matroxfb: Matrox %s detected\n", b->name); b 1643 drivers/video/fbdev/matrox/matroxfb_base.c minfo->chip = b->chip; b 1644 drivers/video/fbdev/matrox/matroxfb_base.c minfo->capable.srcorg = b->flags & DEVF_SRCORG; b 1645 drivers/video/fbdev/matrox/matroxfb_base.c minfo->devflags.video64bits = b->flags & DEVF_VIDEO64BIT; b 1646 drivers/video/fbdev/matrox/matroxfb_base.c if (b->flags & DEVF_TEXT4B) { b 1650 drivers/video/fbdev/matrox/matroxfb_base.c } else if (b->flags & DEVF_TEXT16B) { b 1659 drivers/video/fbdev/matrox/matroxfb_base.c minfo->devflags.support32MB = (b->flags & DEVF_SUPPORT32MB) != 0; b 1660 drivers/video/fbdev/matrox/matroxfb_base.c minfo->devflags.precise_width = !(b->flags & DEVF_ANY_VXRES); b 1661 drivers/video/fbdev/matrox/matroxfb_base.c minfo->devflags.crtc2 = (b->flags & DEVF_CRTC2) != 0; b 1662 drivers/video/fbdev/matrox/matroxfb_base.c minfo->devflags.maven_capable = (b->flags & DEVF_MAVEN_CAPABLE) != 0; b 1663 drivers/video/fbdev/matrox/matroxfb_base.c minfo->devflags.dualhead = (b->flags & DEVF_DUALHEAD) != 0; b 1665 drivers/video/fbdev/matrox/matroxfb_base.c minfo->devflags.g450dac = (b->flags & DEVF_G450DAC) != 0; b 1669 drivers/video/fbdev/matrox/matroxfb_base.c if (b->flags & DEVF_PANELLINK_CAPABLE) { b 1678 drivers/video/fbdev/matrox/matroxfb_base.c minfo->capable.cross4MB = b->flags & DEVF_CROSS4MB; b 1679 drivers/video/fbdev/matrox/matroxfb_base.c if (b->flags & DEVF_SWAPS) { b 1697 drivers/video/fbdev/matrox/matroxfb_base.c memsize = b->base->maxvram; b 1768 drivers/video/fbdev/matrox/matroxfb_base.c if (minfo->video.len_usable > b->base->maxdisplayable) b 1769 drivers/video/fbdev/matrox/matroxfb_base.c minfo->video.len_usable = b->base->maxdisplayable; b 2030 drivers/video/fbdev/matrox/matroxfb_base.c struct board* b; b 2040 drivers/video/fbdev/matrox/matroxfb_base.c for (b = dev_list; b->vendor; b++) { b 2041 drivers/video/fbdev/matrox/matroxfb_base.c if ((b->vendor != pdev->vendor) || (b->device != pdev->device) || (b->rev < pdev->revision)) continue; b 2042 drivers/video/fbdev/matrox/matroxfb_base.c if (b->svid) b 2043 drivers/video/fbdev/matrox/matroxfb_base.c if ((b->svid != svid) || (b->sid != sid)) continue; b 2047 drivers/video/fbdev/matrox/matroxfb_base.c if (!b->vendor) b 2103 drivers/video/fbdev/matrox/matroxfb_base.c err = initMatrox2(minfo, b); b 135 drivers/video/fbdev/matrox/matroxfb_g450.c const int b = minfo->altout.tvo_params.brightness + BLMIN; b 138 drivers/video/fbdev/matrox/matroxfb_g450.c *bl = max(b - c, BLMIN); b 139 drivers/video/fbdev/matrox/matroxfb_g450.c *wl = min(b + c, WLMAX); b 360 drivers/video/fbdev/matrox/matroxfb_maven.c const int b = md->primary_head->altout.tvo_params.brightness + BLMIN; b 363 drivers/video/fbdev/matrox/matroxfb_maven.c *bl = max(b - c, BLMIN); b 364 drivers/video/fbdev/matrox/matroxfb_maven.c *wl = min(b + c, WLMAX); b 735 drivers/video/fbdev/matrox/matroxfb_maven.c unsigned int uninitialized_var(a), uninitialized_var(b), b 739 drivers/video/fbdev/matrox/matroxfb_maven.c if (!matroxfb_mavenclock((m->mode == MATROXFB_OUTPUT_MODE_PAL) ? &maven_PAL : &maven_NTSC, h, vt, &a, &b, &c, &h2)) { b 745 drivers/video/fbdev/matrox/matroxfb_maven.c m->regs[0x81] = b - 1; b 892 drivers/video/fbdev/matrox/matroxfb_maven.c u32 b; b 895 drivers/video/fbdev/matrox/matroxfb_maven.c b = (mt->VTotal - 1) * (m->htotal + 2) + m->hcorr + 2; b 898 drivers/video/fbdev/matrox/matroxfb_maven.c do_div(f1, b); b 453 drivers/video/fbdev/matrox/matroxfb_misc.c unsigned char b; b 455 drivers/video/fbdev/matrox/matroxfb_misc.c b = readb(vbios + 0x7FF1); b 456 drivers/video/fbdev/matrox/matroxfb_misc.c if (b == 0xFF) { b 457 drivers/video/fbdev/matrox/matroxfb_misc.c b = 0; b 459 drivers/video/fbdev/matrox/matroxfb_misc.c bd->output.state = b; b 474 drivers/video/fbdev/matrox/matroxfb_misc.c unsigned char b = readb(vbios + i); b 476 drivers/video/fbdev/matrox/matroxfb_misc.c if (b == '(' && readb(vbios + i + 1) == 'V') { b 484 drivers/video/fbdev/matrox/matroxfb_misc.c if (b == 0) b 782 drivers/video/fbdev/matrox/matroxfb_misc.c unsigned char __iomem* b; b 784 drivers/video/fbdev/matrox/matroxfb_misc.c b = ioremap(0x000C0000, 65536); b 785 drivers/video/fbdev/matrox/matroxfb_misc.c if (!b) { b 788 drivers/video/fbdev/matrox/matroxfb_misc.c unsigned int ven = readb(b+0x64+0) | (readb(b+0x64+1) << 8); b 789 drivers/video/fbdev/matrox/matroxfb_misc.c unsigned int dev = readb(b+0x64+2) | (readb(b+0x64+3) << 8); b 795 drivers/video/fbdev/matrox/matroxfb_misc.c parse_bios(b, &minfo->bios); b 797 drivers/video/fbdev/matrox/matroxfb_misc.c iounmap(b); b 118 drivers/video/fbdev/mb862xx/mb862xxfb.h #define pack(a, b) (((a) << 16) | (b)) b 256 drivers/video/fbdev/mmp/hw/mmp_ctrl.h #define CFG_H_BACK_PORCH(b) ((b)<<16) b 259 drivers/video/fbdev/mmp/hw/mmp_ctrl.h #define CFG_V_BACK_PORCH(b) ((b)<<16) b 407 drivers/video/fbdev/nvidia/nv_hw.c int b = fifo_data.graphics_burst_size >> 4; b 409 drivers/video/fbdev/nvidia/nv_hw.c while (b >>= 1) b 647 drivers/video/fbdev/nvidia/nv_hw.c int b = fifo_data.graphics_burst_size >> 4; b 649 drivers/video/fbdev/nvidia/nv_hw.c while (b >>= 1) b 749 drivers/video/fbdev/nvidia/nv_hw.c int b = fifo_data.graphics_burst_size >> 4; b 751 drivers/video/fbdev/nvidia/nv_hw.c while (b >>= 1) b 17 drivers/video/fbdev/nvidia/nv_type.h #define BITMASK(t,b) (((unsigned)(1U << (((t)-(b)+1)))-1) << (b)) b 119 drivers/video/fbdev/nvidia/nvidia.c u32 b, tmp; b 124 drivers/video/fbdev/nvidia/nvidia.c b = *data++; b 125 drivers/video/fbdev/nvidia/nvidia.c reverse_order(&b); b 130 drivers/video/fbdev/nvidia/nvidia.c tmp = (b & (1 << 31)) ? fg << 16 : bg << 16; b 131 drivers/video/fbdev/nvidia/nvidia.c b <<= 1; b 132 drivers/video/fbdev/nvidia/nvidia.c tmp |= (b & (1 << 31)) ? fg : bg; b 133 drivers/video/fbdev/nvidia/nvidia.c b <<= 1; b 135 drivers/video/fbdev/nvidia/nvidia.c tmp = (b & 1) ? fg : bg; b 136 drivers/video/fbdev/nvidia/nvidia.c b >>= 1; b 137 drivers/video/fbdev/nvidia/nvidia.c tmp |= (b & 1) ? fg << 16 : bg << 16; b 138 drivers/video/fbdev/nvidia/nvidia.c b >>= 1; b 732 drivers/video/fbdev/nvidia/nvidia.c u8 r, g, b; b 742 drivers/video/fbdev/nvidia/nvidia.c nvidia_read_clut(par, regno * 4, &r, &g, &b); b 746 drivers/video/fbdev/nvidia/nvidia.c green >> 8, b); b 191 drivers/video/fbdev/omap/hwa742.c u8 b; b 193 drivers/video/fbdev/omap/hwa742.c b = hwa742_read_reg(HWA742_NDP_CTRL); b 194 drivers/video/fbdev/omap/hwa742.c b |= 1 << 2; b 195 drivers/video/fbdev/omap/hwa742.c hwa742_write_reg(HWA742_NDP_CTRL, b); b 218 drivers/video/fbdev/omap/hwa742.c u8 b; b 222 drivers/video/fbdev/omap/hwa742.c b = hwa742_read_reg(HWA742_NDP_CTRL); b 223 drivers/video/fbdev/omap/hwa742.c b &= ~(1 << 2); b 224 drivers/video/fbdev/omap/hwa742.c hwa742_write_reg(HWA742_NDP_CTRL, b); b 783 drivers/video/fbdev/omap/hwa742.c u8 b; b 866 drivers/video/fbdev/omap/hwa742.c b = hwa742_read_reg(HWA742_NDP_CTRL); b 867 drivers/video/fbdev/omap/hwa742.c b &= ~0x3; b 868 drivers/video/fbdev/omap/hwa742.c b |= use_hsvs ? 1 : 0; b 869 drivers/video/fbdev/omap/hwa742.c b |= (use_ndp && use_hsvs) ? 0 : 2; b 870 drivers/video/fbdev/omap/hwa742.c hwa742_write_reg(HWA742_NDP_CTRL, b); b 879 drivers/video/fbdev/omap/hwa742.c hs, vs, (b & 0x3), !use_hsvs); b 132 drivers/video/fbdev/omap2/omapfb/dss/core.c static int omap_dss_pm_notif(struct notifier_block *b, unsigned long v, void *d) b 3007 drivers/video/fbdev/omap2/omapfb/dss/dsi.c int b; b 3017 drivers/video/fbdev/omap2/omapfb/dss/dsi.c for (b = 0; b < 4; ++b) { b 3019 drivers/video/fbdev/omap2/omapfb/dss/dsi.c buf[w] = (val >> (b * 8)) & 0xff; b 397 drivers/video/fbdev/omap2/omapfb/dss/dss.c int b; b 402 drivers/video/fbdev/omap2/omapfb/dss/dss.c b = 0; b 405 drivers/video/fbdev/omap2/omapfb/dss/dss.c b = 1; b 408 drivers/video/fbdev/omap2/omapfb/dss/dss.c b = 2; b 417 drivers/video/fbdev/omap2/omapfb/dss/dss.c REG_FLD_MOD(DSS_CONTROL, b, start, end); /* DISPC_CLK_SWITCH */ b 425 drivers/video/fbdev/omap2/omapfb/dss/dss.c int b, pos; b 429 drivers/video/fbdev/omap2/omapfb/dss/dss.c b = 0; b 433 drivers/video/fbdev/omap2/omapfb/dss/dss.c b = 1; b 437 drivers/video/fbdev/omap2/omapfb/dss/dss.c b = 1; b 445 drivers/video/fbdev/omap2/omapfb/dss/dss.c REG_FLD_MOD(DSS_CONTROL, b, pos, pos); /* DSIx_CLK_SWITCH */ b 453 drivers/video/fbdev/omap2/omapfb/dss/dss.c int b, ix, pos; b 462 drivers/video/fbdev/omap2/omapfb/dss/dss.c b = 0; b 466 drivers/video/fbdev/omap2/omapfb/dss/dss.c b = 1; b 471 drivers/video/fbdev/omap2/omapfb/dss/dss.c b = 1; b 480 drivers/video/fbdev/omap2/omapfb/dss/dss.c REG_FLD_MOD(DSS_CONTROL, b, pos, pos); /* LCDx_CLK_SWITCH */ b 468 drivers/video/fbdev/omap2/omapfb/dss/dss.h int b; b 469 drivers/video/fbdev/omap2/omapfb/dss/dss.h for (b = 0; b < 32; ++b) { b 470 drivers/video/fbdev/omap2/omapfb/dss/dss.h if (irqstatus & (1 << b)) b 471 drivers/video/fbdev/omap2/omapfb/dss/dss.h irq_arr[b]++; b 414 drivers/video/fbdev/omap2/omapfb/dss/hdmi5_core.c unsigned y, a, b, s; b 429 drivers/video/fbdev/omap2/omapfb/dss/hdmi5_core.c b = (ptr[0] >> 2) & 0x3; b 448 drivers/video/fbdev/omap2/omapfb/dss/hdmi5_core.c (a << 6) | (s << 4) | (b << 2) | (y << 0)); b 64 drivers/video/fbdev/omap2/omapfb/omapfb-main.c int b = (color >> 0) & 0xff; b 72 drivers/video/fbdev/omap2/omapfb/omapfb-main.c b = b * 32 / 256; b 74 drivers/video/fbdev/omap2/omapfb/omapfb-main.c __raw_writew((r << 11) | (g << 5) | (b << 0), p); b 79 drivers/video/fbdev/omap2/omapfb/omapfb-main.c __raw_writeb(b, p + 0); b 123 drivers/video/fbdev/omap2/omapfb/omapfb-main.c unsigned r = 0, g = 0, b = 0; b 127 drivers/video/fbdev/omap2/omapfb/omapfb-main.c b = (y % 32) * 256 / 32; b 134 drivers/video/fbdev/omap2/omapfb/omapfb-main.c b = (y % 256); b 140 drivers/video/fbdev/omap2/omapfb/omapfb-main.c c = (r << 16) | (g << 8) | (b << 0); b 54 drivers/video/fbdev/pm2fb.c #define DPRINTK(a, b...) \ b 55 drivers/video/fbdev/pm2fb.c printk(KERN_DEBUG "pm2fb: %s: " a, __func__ , ## b) b 57 drivers/video/fbdev/pm2fb.c #define DPRINTK(a, b...) b 434 drivers/video/fbdev/pm2fb.c unsigned char r, unsigned char g, unsigned char b) b 443 drivers/video/fbdev/pm2fb.c pm2_WR(p, PM2R_RD_PALETTE_DATA, b); b 44 drivers/video/fbdev/pm3fb.c #define DPRINTK(a, b...) \ b 45 drivers/video/fbdev/pm3fb.c printk(KERN_DEBUG "pm3fb: %s: " a, __func__ , ## b) b 47 drivers/video/fbdev/pm3fb.c #define DPRINTK(a, b...) b 121 drivers/video/fbdev/pm3fb.c unsigned char r, unsigned char g, unsigned char b) b 130 drivers/video/fbdev/pm3fb.c PM3_WRITE_REG(par, PM3RD_PaletteData, b); b 135 drivers/video/fbdev/pm3fb.c unsigned char r, unsigned char g, unsigned char b) b 140 drivers/video/fbdev/pm3fb.c pm3fb_set_color(par, i, r, g, b); b 83 drivers/video/fbdev/pxa168fb.h #define CFG_H_BACK_PORCH(b) ((b) << 16) b 86 drivers/video/fbdev/pxa168fb.h #define CFG_V_BACK_PORCH(b) ((b) << 16) b 309 drivers/video/fbdev/pxafb.c #define SET_PIXFMT(v, r, g, b, t) \ b 311 drivers/video/fbdev/pxafb.c (v)->transp.offset = (t) ? (r) + (g) + (b) : 0; \ b 313 drivers/video/fbdev/pxafb.c (v)->blue.length = (b); (v)->blue.offset = 0; \ b 314 drivers/video/fbdev/pxafb.c (v)->green.length = (g); (v)->green.offset = (b); \ b 315 drivers/video/fbdev/pxafb.c (v)->red.length = (r); (v)->red.offset = (b) + (g); \ b 491 drivers/video/fbdev/riva/fbdev.c u32 b, tmp; b 499 drivers/video/fbdev/riva/fbdev.c b = *data++; b 500 drivers/video/fbdev/riva/fbdev.c reverse_order(&b); b 505 drivers/video/fbdev/riva/fbdev.c tmp = (b & (1 << 31)) ? fg << 16 : bg << 16; b 506 drivers/video/fbdev/riva/fbdev.c b <<= 1; b 507 drivers/video/fbdev/riva/fbdev.c tmp |= (b & (1 << 31)) ? fg : bg; b 508 drivers/video/fbdev/riva/fbdev.c b <<= 1; b 510 drivers/video/fbdev/riva/fbdev.c tmp = (b & 1) ? fg : bg; b 511 drivers/video/fbdev/riva/fbdev.c b >>= 1; b 512 drivers/video/fbdev/riva/fbdev.c tmp |= (b & 1) ? fg << 16 : bg << 16; b 513 drivers/video/fbdev/riva/fbdev.c b >>= 1; b 1353 drivers/video/fbdev/riva/fbdev.c u8 r, g, b; b 1362 drivers/video/fbdev/riva/fbdev.c riva_rclut(chip, regno*4, &r, &g, &b); b 1365 drivers/video/fbdev/riva/fbdev.c green >> 8, b); b 30 drivers/video/fbdev/riva/nvreg.h #define BITMASK(t,b) (((unsigned)(1U << (((t)-(b)+1)))-1) << (b)) b 641 drivers/video/fbdev/riva/riva_hw.c int b = fifo_data.graphics_burst_size >> 4; b 643 drivers/video/fbdev/riva/riva_hw.c while (b >>= 1) b 832 drivers/video/fbdev/riva/riva_hw.c int b = fifo_data.graphics_burst_size >> 4; b 834 drivers/video/fbdev/riva/riva_hw.c while (b >>= 1) b 1097 drivers/video/fbdev/riva/riva_hw.c int b = fifo_data.graphics_burst_size >> 4; b 1099 drivers/video/fbdev/riva/riva_hw.c while (b >>= 1) b 1153 drivers/video/fbdev/riva/riva_hw.c int b = fifo_data.graphics_burst_size >> 4; b 1155 drivers/video/fbdev/riva/riva_hw.c while (b >>= 1) b 149 drivers/video/fbdev/s3c-fb.c struct fb_bitfield b; b 771 drivers/video/fbdev/s3c-fb.c val |= chan_to_field(blue, &win->palette.b); b 1222 drivers/video/fbdev/s3c-fb.c win->palette.b.offset = 0; b 1223 drivers/video/fbdev/s3c-fb.c win->palette.b.length = 5; b 1231 drivers/video/fbdev/s3c-fb.c win->palette.b.offset = 0; b 1232 drivers/video/fbdev/s3c-fb.c win->palette.b.length = 8; b 117 drivers/video/fbdev/savage/savagefb.h #define BCI_CLIP_BR(b, r) ((((b) << 16) | (r)) & 0x0FFF0FFF) b 6609 drivers/video/fbdev/sis/init301.c unsigned short a, b, p = 0; b 6612 drivers/video/fbdev/sis/init301.c b = SiS_Pr->SiS_HDE; b 6615 drivers/video/fbdev/sis/init301.c b = SiS_Pr->SiS_VDE; b 6618 drivers/video/fbdev/sis/init301.c if(a < b) { b 6620 drivers/video/fbdev/sis/init301.c } else if(a == b) { b 96 drivers/video/fbdev/sm712.h #define pal_rgb(r, g, b, val) (((r & 0xf800) >> 8) | \ b 99 drivers/video/fbdev/sm712.h ((b & 0xf800) >> 3)) b 106 drivers/video/fbdev/sm712.h #define pal_rgb(r, g, b, val) val b 80 drivers/video/fbdev/tdfxfb.c #define DPRINTK(a, b...) pr_debug("fb: %s: " a, __func__ , ## b) b 340 drivers/video/fbdev/tridentfb.c u32 x, u32 y, u32 w, u32 h, u32 c, u32 b) b 345 drivers/video/fbdev/tridentfb.c writemmr(par, BGCOLOR, b); b 475 drivers/video/fbdev/via/hw.c static void write_dac_reg(u8 index, u8 r, u8 g, u8 b) b 480 drivers/video/fbdev/via/hw.c outb(b, LUT_DATA); b 300 drivers/video/fbdev/via/viafbdev.c u32 r, g, b; b 319 drivers/video/fbdev/via/viafbdev.c b = (blue >> (16 - info->var.blue.length)) b 323 drivers/video/fbdev/via/viafbdev.c ((u32 *) info->pseudo_palette)[regno] = r | g | b; b 152 drivers/virtio/virtio_mmio.c u8 b; b 167 drivers/virtio/virtio_mmio.c b = readb(base + offset); b 168 drivers/virtio/virtio_mmio.c memcpy(buf, &b, sizeof b); b 194 drivers/virtio/virtio_mmio.c u8 b; b 210 drivers/virtio/virtio_mmio.c memcpy(&b, buf, sizeof b); b 211 drivers/virtio/virtio_mmio.c writeb(b, base + offset); b 194 drivers/virtio/virtio_pci_modern.c u8 b; b 202 drivers/virtio/virtio_pci_modern.c b = ioread8(vp_dev->device + offset); b 203 drivers/virtio/virtio_pci_modern.c memcpy(buf, &b, sizeof b); b 230 drivers/virtio/virtio_pci_modern.c u8 b; b 238 drivers/virtio/virtio_pci_modern.c memcpy(&b, buf, sizeof b); b 239 drivers/virtio/virtio_pci_modern.c iowrite8(b, vp_dev->device + offset); b 75 drivers/xen/events/events_fifo.c #define EVTCHN_FIFO_BIT(b, w) \ b 76 drivers/xen/events/events_fifo.c (((unsigned long)w & 0x4UL) ? (EVTCHN_FIFO_ ##b + 32) : EVTCHN_FIFO_ ##b) b 81 drivers/xen/events/events_fifo.c #define EVTCHN_FIFO_BIT(b, w) EVTCHN_FIFO_ ##b b 51 drivers/xen/xen-pciback/conf_space.c if (field->u.b.read) b 52 drivers/xen/xen-pciback/conf_space.c ret = field->u.b.read(dev, offset, (u8 *) value, b 77 drivers/xen/xen-pciback/conf_space.c if (field->u.b.write) b 78 drivers/xen/xen-pciback/conf_space.c ret = field->u.b.write(dev, offset, (u8) value, b 56 drivers/xen/xen-pciback/conf_space.h } b; b 182 drivers/xen/xen-pciback/conf_space_capability.c .u.b.read = xen_pcibk_read_config_byte, b 187 drivers/xen/xen-pciback/conf_space_capability.c .u.b.read = xen_pcibk_read_config_byte, b 330 drivers/xen/xen-pciback/conf_space_header.c .u.b.read = interrupt_read, b 335 drivers/xen/xen-pciback/conf_space_header.c .u.b.read = xen_pcibk_read_config_byte, b 341 drivers/xen/xen-pciback/conf_space_header.c .u.b.read = xen_pcibk_read_config_byte, b 342 drivers/xen/xen-pciback/conf_space_header.c .u.b.write = xen_pcibk_write_config_byte, b 347 drivers/xen/xen-pciback/conf_space_header.c .u.b.read = xen_pcibk_read_config_byte, b 352 drivers/xen/xen-pciback/conf_space_header.c .u.b.read = xen_pcibk_read_config_byte, b 353 drivers/xen/xen-pciback/conf_space_header.c .u.b.write = bist_write, b 71 drivers/xen/xen-pciback/conf_space_quirks.c field->u.b.read = xen_pcibk_read_config_byte; b 72 drivers/xen/xen-pciback/conf_space_quirks.c field->u.b.write = xen_pcibk_write_config_byte; b 304 drivers/xen/xen-pciback/xenbus.c unsigned int d, b; b 326 drivers/xen/xen-pciback/xenbus.c str, "%x:%x", &d, &b); b 334 drivers/xen/xen-pciback/xenbus.c if (d == domain && b == bus) { b 424 fs/afs/cmservice.c __be32 *b; b 454 fs/afs/cmservice.c b = call->buffer; b 456 fs/afs/cmservice.c r->time_low = b[0]; b 457 fs/afs/cmservice.c r->time_mid = htons(ntohl(b[1])); b 458 fs/afs/cmservice.c r->time_hi_and_version = htons(ntohl(b[2])); b 459 fs/afs/cmservice.c r->clock_seq_hi_and_reserved = ntohl(b[3]); b 460 fs/afs/cmservice.c r->clock_seq_low = ntohl(b[4]); b 463 fs/afs/cmservice.c r->node[loop] = ntohl(b[loop + 5]); b 537 fs/afs/cmservice.c __be32 *b; b 565 fs/afs/cmservice.c b = call->buffer; b 567 fs/afs/cmservice.c r->time_low = b[0]; b 568 fs/afs/cmservice.c r->time_mid = htons(ntohl(b[1])); b 569 fs/afs/cmservice.c r->time_hi_and_version = htons(ntohl(b[2])); b 570 fs/afs/cmservice.c r->clock_seq_hi_and_reserved = ntohl(b[3]); b 571 fs/afs/cmservice.c r->clock_seq_low = ntohl(b[4]); b 574 fs/afs/cmservice.c r->node[loop] = ntohl(b[loop + 5]); b 194 fs/afs/dir_edit.c unsigned int need_slots, nr_blocks, b; b 230 fs/afs/dir_edit.c for (b = 0; b < nr_blocks + 1; b++) { b 234 fs/afs/dir_edit.c index = b / AFS_DIR_BLOCKS_PER_PAGE; b 257 fs/afs/dir_edit.c block = &dir_page->blocks[b % AFS_DIR_BLOCKS_PER_PAGE]; b 260 fs/afs/dir_edit.c b, b 261 fs/afs/dir_edit.c (b < AFS_DIR_BLOCKS_WITH_CTR) ? meta->meta.alloc_ctrs[b] : 99, b 266 fs/afs/dir_edit.c if (b == nr_blocks) { b 267 fs/afs/dir_edit.c _debug("init %u", b); b 268 fs/afs/dir_edit.c afs_edit_init_block(meta, block, b); b 269 fs/afs/dir_edit.c i_size_write(&vnode->vfs_inode, (b + 1) * AFS_DIR_BLOCK_SIZE); b 273 fs/afs/dir_edit.c if (b >= AFS_DIR_BLOCKS_WITH_CTR || b 274 fs/afs/dir_edit.c meta->meta.alloc_ctrs[b] >= need_slots) { b 307 fs/afs/dir_edit.c b = 0; b 311 fs/afs/dir_edit.c trace_afs_edit_dir(vnode, why, afs_edit_dir_create, b, slot, b 331 fs/afs/dir_edit.c if (b < AFS_DIR_BLOCKS_WITH_CTR) b 332 fs/afs/dir_edit.c meta->meta.alloc_ctrs[b] -= need_slots; b 336 fs/afs/dir_edit.c _debug("Insert %s in %u[%u]", name->name, b, slot); b 374 fs/afs/dir_edit.c unsigned int need_slots, nr_blocks, b; b 407 fs/afs/dir_edit.c for (b = 0; b < nr_blocks; b++) { b 408 fs/afs/dir_edit.c index = b / AFS_DIR_BLOCKS_PER_PAGE; b 423 fs/afs/dir_edit.c block = &dir_page->blocks[b % AFS_DIR_BLOCKS_PER_PAGE]; b 425 fs/afs/dir_edit.c if (b > AFS_DIR_BLOCKS_WITH_CTR || b 426 fs/afs/dir_edit.c meta->meta.alloc_ctrs[b] <= AFS_DIR_SLOTS_PER_BLOCK - 1 - need_slots) { b 427 fs/afs/dir_edit.c slot = afs_dir_scan_block(block, name, b); b 448 fs/afs/dir_edit.c trace_afs_edit_dir(vnode, why, afs_edit_dir_delete, b, slot, b 463 fs/afs/dir_edit.c if (b < AFS_DIR_BLOCKS_WITH_CTR) b 464 fs/afs/dir_edit.c meta->meta.alloc_ctrs[b] += need_slots; b 468 fs/afs/dir_edit.c _debug("Remove %s from %u[%u]", name->name, b, slot); b 49 fs/afs/server.c const struct sockaddr_in6 *a = &srx->transport.sin6, *b; b 53 fs/afs/server.c b = &alist->addrs[i].transport.sin6; b 55 fs/afs/server.c (u16 __force)b->sin6_port); b 58 fs/afs/server.c &b->sin6_addr, b 65 fs/afs/server.c const struct sockaddr_in *a = &srx->transport.sin, *b; b 69 fs/afs/server.c b = &alist->addrs[i].transport.sin; b 71 fs/afs/server.c (u16 __force)b->sin_port); b 74 fs/afs/server.c (u32 __force)b->sin_addr.s_addr); b 96 fs/afs/vl_list.c const u8 *b = *_b; b 105 fs/afs/vl_list.c for (; nr_addrs > 0 && end - b >= nr_addrs; nr_addrs--) { b 109 fs/afs/vl_list.c hdr.address_type = *b++; b 113 fs/afs/vl_list.c if (end - b < 4) { b 117 fs/afs/vl_list.c memcpy(x, b, 4); b 119 fs/afs/vl_list.c b += 4; b 123 fs/afs/vl_list.c if (end - b < 16) { b 127 fs/afs/vl_list.c memcpy(x, b, 16); b 129 fs/afs/vl_list.c b += 16; b 144 fs/afs/vl_list.c *_b = b; b 148 fs/afs/vl_list.c *_b = b; b 165 fs/afs/vl_list.c const u8 *b = buffer, *end = buffer + buffer_size; b 171 fs/afs/vl_list.c if (end - b < sizeof(*hdr) || b 175 fs/afs/vl_list.c hdr->hdr.content, hdr->hdr.version, end - b); b 197 fs/afs/vl_list.c b += sizeof(*hdr); b 198 fs/afs/vl_list.c while (end - b >= sizeof(bs)) { b 199 fs/afs/vl_list.c bs.name_len = afs_extract_le16(&b); b 200 fs/afs/vl_list.c bs.priority = afs_extract_le16(&b); b 201 fs/afs/vl_list.c bs.weight = afs_extract_le16(&b); b 202 fs/afs/vl_list.c bs.port = afs_extract_le16(&b); b 203 fs/afs/vl_list.c bs.source = *b++; b 204 fs/afs/vl_list.c bs.status = *b++; b 205 fs/afs/vl_list.c bs.protocol = *b++; b 206 fs/afs/vl_list.c bs.nr_addrs = *b++; b 211 fs/afs/vl_list.c bs.name_len, bs.name_len, b); b 213 fs/afs/vl_list.c if (end - b < bs.name_len) b 238 fs/afs/vl_list.c strncasecmp(b, p->name, bs.name_len) == 0) { b 246 fs/afs/vl_list.c server = afs_alloc_vlserver(b, bs.name_len, bs.port); b 251 fs/afs/vl_list.c b += bs.name_len; b 256 fs/afs/vl_list.c addrs = afs_extract_vl_addrs(&b, end, bs.nr_addrs, bs.port); b 314 fs/afs/vl_list.c if (b != end) { b 315 fs/afs/vl_list.c _debug("parse error %zd", b - end); b 330 fs/afs/vl_list.c printk(KERN_DEBUG "DNS: at %zu\n", (const void *)b - buffer); b 251 fs/btrfs/check-integrity.c static void btrfsic_block_init(struct btrfsic_block *b); b 253 fs/btrfs/check-integrity.c static void btrfsic_block_free(struct btrfsic_block *b); b 261 fs/btrfs/check-integrity.c static void btrfsic_block_hashtable_add(struct btrfsic_block *b, b 263 fs/btrfs/check-integrity.c static void btrfsic_block_hashtable_remove(struct btrfsic_block *b); b 385 fs/btrfs/check-integrity.c static void btrfsic_block_init(struct btrfsic_block *b) b 387 fs/btrfs/check-integrity.c b->magic_num = BTRFSIC_BLOCK_MAGIC_NUMBER; b 388 fs/btrfs/check-integrity.c b->dev_state = NULL; b 389 fs/btrfs/check-integrity.c b->dev_bytenr = 0; b 390 fs/btrfs/check-integrity.c b->logical_bytenr = 0; b 391 fs/btrfs/check-integrity.c b->generation = BTRFSIC_GENERATION_UNKNOWN; b 392 fs/btrfs/check-integrity.c b->disk_key.objectid = 0; b 393 fs/btrfs/check-integrity.c b->disk_key.type = 0; b 394 fs/btrfs/check-integrity.c b->disk_key.offset = 0; b 395 fs/btrfs/check-integrity.c b->is_metadata = 0; b 396 fs/btrfs/check-integrity.c b->is_superblock = 0; b 397 fs/btrfs/check-integrity.c b->is_iodone = 0; b 398 fs/btrfs/check-integrity.c b->iodone_w_error = 0; b 399 fs/btrfs/check-integrity.c b->never_written = 0; b 400 fs/btrfs/check-integrity.c b->mirror_num = 0; b 401 fs/btrfs/check-integrity.c b->next_in_same_bio = NULL; b 402 fs/btrfs/check-integrity.c b->orig_bio_bh_private = NULL; b 403 fs/btrfs/check-integrity.c b->orig_bio_bh_end_io.bio = NULL; b 404 fs/btrfs/check-integrity.c INIT_LIST_HEAD(&b->collision_resolving_node); b 405 fs/btrfs/check-integrity.c INIT_LIST_HEAD(&b->all_blocks_node); b 406 fs/btrfs/check-integrity.c INIT_LIST_HEAD(&b->ref_to_list); b 407 fs/btrfs/check-integrity.c INIT_LIST_HEAD(&b->ref_from_list); b 408 fs/btrfs/check-integrity.c b->submit_bio_bh_rw = 0; b 409 fs/btrfs/check-integrity.c b->flush_gen = 0; b 414 fs/btrfs/check-integrity.c struct btrfsic_block *b; b 416 fs/btrfs/check-integrity.c b = kzalloc(sizeof(*b), GFP_NOFS); b 417 fs/btrfs/check-integrity.c if (NULL != b) b 418 fs/btrfs/check-integrity.c btrfsic_block_init(b); b 420 fs/btrfs/check-integrity.c return b; b 423 fs/btrfs/check-integrity.c static void btrfsic_block_free(struct btrfsic_block *b) b 425 fs/btrfs/check-integrity.c BUG_ON(!(NULL == b || BTRFSIC_BLOCK_MAGIC_NUMBER == b->magic_num)); b 426 fs/btrfs/check-integrity.c kfree(b); b 496 fs/btrfs/check-integrity.c static void btrfsic_block_hashtable_add(struct btrfsic_block *b, b 500 fs/btrfs/check-integrity.c (((unsigned int)(b->dev_bytenr >> 16)) ^ b 501 fs/btrfs/check-integrity.c ((unsigned int)((uintptr_t)b->dev_state->bdev))) & b 504 fs/btrfs/check-integrity.c list_add(&b->collision_resolving_node, h->table + hashval); b 507 fs/btrfs/check-integrity.c static void btrfsic_block_hashtable_remove(struct btrfsic_block *b) b 509 fs/btrfs/check-integrity.c list_del(&b->collision_resolving_node); b 521 fs/btrfs/check-integrity.c struct btrfsic_block *b; b 523 fs/btrfs/check-integrity.c list_for_each_entry(b, h->table + hashval, collision_resolving_node) { b 524 fs/btrfs/check-integrity.c if (b->dev_state->bdev == bdev && b->dev_bytenr == dev_bytenr) b 525 fs/btrfs/check-integrity.c return b; b 2419 fs/btrfs/ctree.c struct extent_buffer *b = *eb_ret; b 2425 fs/btrfs/ctree.c blocknr = btrfs_node_blockptr(b, slot); b 2426 fs/btrfs/ctree.c gen = btrfs_node_ptr_generation(b, slot); b 2427 fs/btrfs/ctree.c parent_level = btrfs_header_level(b); b 2428 fs/btrfs/ctree.c btrfs_node_key_to_cpu(b, &first_key, slot); b 2513 fs/btrfs/ctree.c struct extent_buffer *b, int level, int ins_len, b 2519 fs/btrfs/ctree.c if ((p->search_for_split || ins_len > 0) && btrfs_header_nritems(b) >= b 2538 fs/btrfs/ctree.c b = p->nodes[level]; b 2539 fs/btrfs/ctree.c } else if (ins_len < 0 && btrfs_header_nritems(b) < b 2557 fs/btrfs/ctree.c b = p->nodes[level]; b 2558 fs/btrfs/ctree.c if (!b) { b 2562 fs/btrfs/ctree.c BUG_ON(btrfs_header_nritems(b) == 1); b 2572 fs/btrfs/ctree.c static int key_search(struct extent_buffer *b, const struct btrfs_key *key, b 2576 fs/btrfs/ctree.c *prev_cmp = btrfs_bin_search(b, key, level, slot); b 2625 fs/btrfs/ctree.c struct extent_buffer *b; b 2644 fs/btrfs/ctree.c b = btrfs_clone_extent_buffer(root->commit_root); b 2646 fs/btrfs/ctree.c if (!b) b 2650 fs/btrfs/ctree.c b = root->commit_root; b 2651 fs/btrfs/ctree.c extent_buffer_get(b); b 2653 fs/btrfs/ctree.c level = btrfs_header_level(b); b 2664 fs/btrfs/ctree.c b = btrfs_root_node(root); b 2665 fs/btrfs/ctree.c level = btrfs_header_level(b); b 2678 fs/btrfs/ctree.c b = btrfs_read_lock_root_node(root); b 2679 fs/btrfs/ctree.c level = btrfs_header_level(b); b 2684 fs/btrfs/ctree.c btrfs_tree_read_unlock(b); b 2685 fs/btrfs/ctree.c free_extent_buffer(b); b 2688 fs/btrfs/ctree.c b = btrfs_lock_root_node(root); b 2692 fs/btrfs/ctree.c level = btrfs_header_level(b); b 2695 fs/btrfs/ctree.c p->nodes[level] = b; b 2701 fs/btrfs/ctree.c return b; b 2734 fs/btrfs/ctree.c struct extent_buffer *b; b 2777 fs/btrfs/ctree.c b = btrfs_search_slot_get_root(root, p, write_lock_level); b 2778 fs/btrfs/ctree.c if (IS_ERR(b)) { b 2779 fs/btrfs/ctree.c ret = PTR_ERR(b); b 2783 fs/btrfs/ctree.c while (b) { b 2784 fs/btrfs/ctree.c level = btrfs_header_level(b); b 2798 fs/btrfs/ctree.c if (!should_cow_block(trans, root, b)) { b 2818 fs/btrfs/ctree.c err = btrfs_cow_block(trans, root, b, NULL, 0, b 2819 fs/btrfs/ctree.c &b); b 2821 fs/btrfs/ctree.c err = btrfs_cow_block(trans, root, b, b 2823 fs/btrfs/ctree.c p->slots[level + 1], &b); b 2830 fs/btrfs/ctree.c p->nodes[level] = b; b 2856 fs/btrfs/ctree.c ret = key_search(b, key, level, &prev_cmp, &slot); b 2867 fs/btrfs/ctree.c err = setup_nodes_for_search(trans, root, p, b, level, b 2875 fs/btrfs/ctree.c b = p->nodes[level]; b 2900 fs/btrfs/ctree.c err = read_block_for_search(root, p, &b, level, b 2910 fs/btrfs/ctree.c level = btrfs_header_level(b); b 2912 fs/btrfs/ctree.c if (!btrfs_try_tree_write_lock(b)) { b 2914 fs/btrfs/ctree.c btrfs_tree_lock(b); b 2918 fs/btrfs/ctree.c if (!btrfs_tree_read_lock_atomic(b)) { b 2920 fs/btrfs/ctree.c btrfs_tree_read_lock(b); b 2924 fs/btrfs/ctree.c p->nodes[level] = b; b 2929 fs/btrfs/ctree.c btrfs_leaf_free_space(b) < ins_len) { b 2980 fs/btrfs/ctree.c struct extent_buffer *b; b 2998 fs/btrfs/ctree.c b = get_old_root(root, time_seq); b 2999 fs/btrfs/ctree.c if (!b) { b 3003 fs/btrfs/ctree.c level = btrfs_header_level(b); b 3006 fs/btrfs/ctree.c while (b) { b 3007 fs/btrfs/ctree.c level = btrfs_header_level(b); b 3008 fs/btrfs/ctree.c p->nodes[level] = b; b 3023 fs/btrfs/ctree.c ret = key_search(b, key, level, &prev_cmp, &slot); b 3042 fs/btrfs/ctree.c err = read_block_for_search(root, p, &b, level, b 3051 fs/btrfs/ctree.c level = btrfs_header_level(b); b 3052 fs/btrfs/ctree.c if (!btrfs_tree_read_lock_atomic(b)) { b 3054 fs/btrfs/ctree.c btrfs_tree_read_lock(b); b 3056 fs/btrfs/ctree.c b = tree_mod_log_rewind(fs_info, p, b, time_seq); b 3057 fs/btrfs/ctree.c if (!b) { b 3062 fs/btrfs/ctree.c p->nodes[level] = b; b 3413 fs/btrfs/ctree.h #define in_range(b, first, len) ((b) >= (first) && (b) < (first) + (len)) b 1273 fs/btrfs/disk-io.c memcpy(root->root_item.uuid, uuid.b, BTRFS_UUID_SIZE); b 648 fs/btrfs/ioctl.c memcpy(root_item->uuid, new_uuid.b, BTRFS_UUID_SIZE); b 10 fs/btrfs/misc.h #define in_range(b, first, len) ((b) >= (first) && (b) < (first) + (len)) b 1672 fs/btrfs/raid56.c static int plug_cmp(void *priv, struct list_head *a, struct list_head *b) b 1676 fs/btrfs/raid56.c struct btrfs_raid_bio *rb = container_of(b, struct btrfs_raid_bio, b 48 fs/btrfs/root-tree.c memcpy(item->uuid, uuid.b, BTRFS_UUID_SIZE); b 1521 fs/btrfs/transaction.c memcpy(new_root_item->uuid, new_uuid.b, BTRFS_UUID_SIZE); b 1632 fs/btrfs/transaction.c ret = btrfs_uuid_tree_add(trans, new_uuid.b, BTRFS_UUID_KEY_SUBVOL, b 4083 fs/btrfs/tree-log.c static int extent_cmp(void *priv, struct list_head *a, struct list_head *b) b 4088 fs/btrfs/tree-log.c em2 = list_entry(b, struct extent_map, list); b 1399 fs/btrfs/volumes.c static int devid_cmp(void *priv, struct list_head *a, struct list_head *b) b 1404 fs/btrfs/volumes.c dev2 = list_entry(b, struct btrfs_device, dev_list); b 4930 fs/btrfs/volumes.c static int btrfs_cmp_device_info(const void *a, const void *b) b 4933 fs/btrfs/volumes.c const struct btrfs_device_info *di_b = b; b 5562 fs/btrfs/volumes.c static inline int parity_smaller(u64 a, u64 b) b 5564 fs/btrfs/volumes.c return a > b; b 1231 fs/buffer.c struct bh_lru *b; b 1237 fs/buffer.c b = this_cpu_ptr(&bh_lrus); b 1239 fs/buffer.c swap(evictee, b->bhs[i]); b 1382 fs/buffer.c struct bh_lru *b = &get_cpu_var(bh_lrus); b 1386 fs/buffer.c brelse(b->bhs[i]); b 1387 fs/buffer.c b->bhs[i] = NULL; b 1394 fs/buffer.c struct bh_lru *b = per_cpu_ptr(&bh_lrus, cpu); b 1398 fs/buffer.c if (b->bhs[i]) b 3382 fs/buffer.c struct bh_lru *b = &per_cpu(bh_lrus, cpu); b 3385 fs/buffer.c brelse(b->bhs[i]); b 3386 fs/buffer.c b->bhs[i] = NULL; b 8 fs/ceph/ceph_frag.c int ceph_frag_compare(__u32 a, __u32 b) b 11 fs/ceph/ceph_frag.c unsigned vb = ceph_frag_value(b); b 17 fs/ceph/ceph_frag.c vb = ceph_frag_bits(b); b 295 fs/ceph/snap.c static int cmpu64_rev(const void *a, const void *b) b 297 fs/ceph/snap.c if (*(u64 *)a < *(u64 *)b) b 299 fs/ceph/snap.c if (*(u64 *)a > *(u64 *)b) b 1081 fs/compat_ioctl.c unsigned int a, b; b 1083 fs/compat_ioctl.c b = *(unsigned int *)q; b 1084 fs/compat_ioctl.c if (a > b) b 1086 fs/compat_ioctl.c if (a < b) b 196 fs/dcache.c unsigned long a,b,mask; b 200 fs/dcache.c b = load_unaligned_zeropad(ct); b 203 fs/dcache.c if (unlikely(a != b)) b 212 fs/dcache.c return unlikely(!!((a ^ b) & mask)); b 478 fs/dcache.c struct hlist_bl_head *b; b 485 fs/dcache.c b = &dentry->d_sb->s_roots; b 487 fs/dcache.c b = d_hash(dentry->d_name.hash); b 489 fs/dcache.c hlist_bl_lock(b); b 491 fs/dcache.c hlist_bl_unlock(b); b 2218 fs/dcache.c struct hlist_bl_head *b = d_hash(hashlen_hash(hashlen)); b 2242 fs/dcache.c hlist_bl_for_each_entry_rcu(dentry, node, b, d_hash) { b 2340 fs/dcache.c struct hlist_bl_head *b = d_hash(hash); b 2367 fs/dcache.c hlist_bl_for_each_entry_rcu(dentry, node, b, d_hash) { b 2460 fs/dcache.c struct hlist_bl_head *b = d_hash(entry->d_name.hash); b 2462 fs/dcache.c hlist_bl_lock(b); b 2463 fs/dcache.c hlist_bl_add_head_rcu(&entry->d_hash, b); b 2464 fs/dcache.c hlist_bl_unlock(b); b 2517 fs/dcache.c struct hlist_bl_head *b = in_lookup_hash(parent, hash); b 2555 fs/dcache.c hlist_bl_lock(b); b 2557 fs/dcache.c hlist_bl_unlock(b); b 2568 fs/dcache.c hlist_bl_for_each_entry(dentry, node, b, d_u.d_in_lookup_hash) { b 2575 fs/dcache.c hlist_bl_unlock(b); b 2612 fs/dcache.c hlist_bl_add_head_rcu(&new->d_u.d_in_lookup_hash, b); b 2613 fs/dcache.c hlist_bl_unlock(b); b 2624 fs/dcache.c struct hlist_bl_head *b = in_lookup_hash(dentry->d_parent, b 2626 fs/dcache.c hlist_bl_lock(b); b 2631 fs/dcache.c hlist_bl_unlock(b); b 64 fs/dlm/dir.c char *b, *last_name = NULL; b 102 fs/dlm/dir.c b = ls->ls_recover_buf->rc_buf; b 113 fs/dlm/dir.c memcpy(&v, b, sizeof(__be16)); b 115 fs/dlm/dir.c b += sizeof(__be16); b 134 fs/dlm/dir.c b, namelen, b 157 fs/dlm/dir.c b, namelen); b 176 fs/dlm/dir.c memcpy(last_name, b, namelen); b 177 fs/dlm/dir.c b += namelen; b 544 fs/dlm/lock.c uint32_t hash, uint32_t b, b 592 fs/dlm/lock.c spin_lock(&ls->ls_rsbtbl[b].lock); b 594 fs/dlm/lock.c error = dlm_search_rsb_tree(&ls->ls_rsbtbl[b].keep, name, len, &r); b 608 fs/dlm/lock.c error = dlm_search_rsb_tree(&ls->ls_rsbtbl[b].toss, name, len, &r); b 648 fs/dlm/lock.c rb_erase(&r->res_hashnode, &ls->ls_rsbtbl[b].toss); b 649 fs/dlm/lock.c error = rsb_insert(r, &ls->ls_rsbtbl[b].keep); b 663 fs/dlm/lock.c spin_unlock(&ls->ls_rsbtbl[b].lock); b 670 fs/dlm/lock.c r->res_bucket = b; b 710 fs/dlm/lock.c error = rsb_insert(r, &ls->ls_rsbtbl[b].keep); b 712 fs/dlm/lock.c spin_unlock(&ls->ls_rsbtbl[b].lock); b 723 fs/dlm/lock.c uint32_t hash, uint32_t b, b 737 fs/dlm/lock.c spin_lock(&ls->ls_rsbtbl[b].lock); b 739 fs/dlm/lock.c error = dlm_search_rsb_tree(&ls->ls_rsbtbl[b].keep, name, len, &r); b 752 fs/dlm/lock.c error = dlm_search_rsb_tree(&ls->ls_rsbtbl[b].toss, name, len, &r); b 783 fs/dlm/lock.c rb_erase(&r->res_hashnode, &ls->ls_rsbtbl[b].toss); b 784 fs/dlm/lock.c error = rsb_insert(r, &ls->ls_rsbtbl[b].keep); b 795 fs/dlm/lock.c spin_unlock(&ls->ls_rsbtbl[b].lock); b 802 fs/dlm/lock.c r->res_bucket = b; b 808 fs/dlm/lock.c error = rsb_insert(r, &ls->ls_rsbtbl[b].keep); b 810 fs/dlm/lock.c spin_unlock(&ls->ls_rsbtbl[b].lock); b 819 fs/dlm/lock.c uint32_t hash, b; b 826 fs/dlm/lock.c b = hash & (ls->ls_rsbtbl_size - 1); b 831 fs/dlm/lock.c return find_rsb_nodir(ls, name, len, hash, b, dir_nodeid, b 834 fs/dlm/lock.c return find_rsb_dir(ls, name, len, hash, b, dir_nodeid, b 914 fs/dlm/lock.c uint32_t hash, b; b 930 fs/dlm/lock.c b = hash & (ls->ls_rsbtbl_size - 1); b 946 fs/dlm/lock.c spin_lock(&ls->ls_rsbtbl[b].lock); b 947 fs/dlm/lock.c error = dlm_search_rsb_tree(&ls->ls_rsbtbl[b].keep, name, len, &r); b 953 fs/dlm/lock.c spin_unlock(&ls->ls_rsbtbl[b].lock); b 958 fs/dlm/lock.c error = dlm_search_rsb_tree(&ls->ls_rsbtbl[b].toss, name, len, &r); b 1041 fs/dlm/lock.c spin_unlock(&ls->ls_rsbtbl[b].lock); b 1052 fs/dlm/lock.c spin_unlock(&ls->ls_rsbtbl[b].lock); b 1059 fs/dlm/lock.c r->res_bucket = b; b 1066 fs/dlm/lock.c error = rsb_insert(r, &ls->ls_rsbtbl[b].toss); b 1070 fs/dlm/lock.c spin_unlock(&ls->ls_rsbtbl[b].lock); b 1079 fs/dlm/lock.c spin_unlock(&ls->ls_rsbtbl[b].lock); b 1103 fs/dlm/lock.c uint32_t hash, b; b 1107 fs/dlm/lock.c b = hash & (ls->ls_rsbtbl_size - 1); b 1109 fs/dlm/lock.c spin_lock(&ls->ls_rsbtbl[b].lock); b 1110 fs/dlm/lock.c error = dlm_search_rsb_tree(&ls->ls_rsbtbl[b].keep, name, len, &r); b 1114 fs/dlm/lock.c error = dlm_search_rsb_tree(&ls->ls_rsbtbl[b].toss, name, len, &r); b 1120 fs/dlm/lock.c spin_unlock(&ls->ls_rsbtbl[b].lock); b 1648 fs/dlm/lock.c static void shrink_bucket(struct dlm_ls *ls, int b) b 1660 fs/dlm/lock.c spin_lock(&ls->ls_rsbtbl[b].lock); b 1662 fs/dlm/lock.c if (!(ls->ls_rsbtbl[b].flags & DLM_RTF_SHRINK)) { b 1663 fs/dlm/lock.c spin_unlock(&ls->ls_rsbtbl[b].lock); b 1667 fs/dlm/lock.c for (n = rb_first(&ls->ls_rsbtbl[b].toss); n; n = next) { b 1712 fs/dlm/lock.c rb_erase(&r->res_hashnode, &ls->ls_rsbtbl[b].toss); b 1717 fs/dlm/lock.c ls->ls_rsbtbl[b].flags |= DLM_RTF_SHRINK; b 1719 fs/dlm/lock.c ls->ls_rsbtbl[b].flags &= ~DLM_RTF_SHRINK; b 1720 fs/dlm/lock.c spin_unlock(&ls->ls_rsbtbl[b].lock); b 1740 fs/dlm/lock.c spin_lock(&ls->ls_rsbtbl[b].lock); b 1741 fs/dlm/lock.c rv = dlm_search_rsb_tree(&ls->ls_rsbtbl[b].toss, name, len, &r); b 1743 fs/dlm/lock.c spin_unlock(&ls->ls_rsbtbl[b].lock); b 1749 fs/dlm/lock.c spin_unlock(&ls->ls_rsbtbl[b].lock); b 1758 fs/dlm/lock.c spin_unlock(&ls->ls_rsbtbl[b].lock); b 1767 fs/dlm/lock.c spin_unlock(&ls->ls_rsbtbl[b].lock); b 1774 fs/dlm/lock.c spin_unlock(&ls->ls_rsbtbl[b].lock); b 1779 fs/dlm/lock.c rb_erase(&r->res_hashnode, &ls->ls_rsbtbl[b].toss); b 1786 fs/dlm/lock.c spin_unlock(&ls->ls_rsbtbl[b].lock); b 1947 fs/dlm/lock.c int b, len = r->res_ls->ls_lvblen; b 1953 fs/dlm/lock.c b = dlm_lvb_operations[lkb->lkb_grmode + 1][lkb->lkb_rqmode + 1]; b 1955 fs/dlm/lock.c if (b == 1) { b 1968 fs/dlm/lock.c } else if (b == 0) { b 2028 fs/dlm/lock.c int b; b 2036 fs/dlm/lock.c b = dlm_lvb_operations[lkb->lkb_grmode + 1][lkb->lkb_rqmode + 1]; b 2037 fs/dlm/lock.c if (b == 1) { b 4020 fs/dlm/lock.c uint32_t hash, b; b 4027 fs/dlm/lock.c b = hash & (ls->ls_rsbtbl_size - 1); b 4033 fs/dlm/lock.c spin_lock(&ls->ls_rsbtbl[b].lock); b 4034 fs/dlm/lock.c rv = dlm_search_rsb_tree(&ls->ls_rsbtbl[b].keep, name, len, &r); b 4036 fs/dlm/lock.c spin_unlock(&ls->ls_rsbtbl[b].lock); b 4041 fs/dlm/lock.c rv = dlm_search_rsb_tree(&ls->ls_rsbtbl[b].toss, name, len, &r); b 4043 fs/dlm/lock.c spin_unlock(&ls->ls_rsbtbl[b].lock); b 4054 fs/dlm/lock.c spin_unlock(&ls->ls_rsbtbl[b].lock); b 4390 fs/dlm/lock.c uint32_t hash, b; b 4423 fs/dlm/lock.c b = hash & (ls->ls_rsbtbl_size - 1); b 4425 fs/dlm/lock.c spin_lock(&ls->ls_rsbtbl[b].lock); b 4427 fs/dlm/lock.c rv = dlm_search_rsb_tree(&ls->ls_rsbtbl[b].toss, name, len, &r); b 4430 fs/dlm/lock.c rv = dlm_search_rsb_tree(&ls->ls_rsbtbl[b].keep, name, len, &r); b 4435 fs/dlm/lock.c spin_unlock(&ls->ls_rsbtbl[b].lock); b 4443 fs/dlm/lock.c spin_unlock(&ls->ls_rsbtbl[b].lock); b 4450 fs/dlm/lock.c spin_unlock(&ls->ls_rsbtbl[b].lock); b 4458 fs/dlm/lock.c spin_unlock(&ls->ls_rsbtbl[b].lock); b 4463 fs/dlm/lock.c rb_erase(&r->res_hashnode, &ls->ls_rsbtbl[b].toss); b 4464 fs/dlm/lock.c spin_unlock(&ls->ls_rsbtbl[b].lock); b 4470 fs/dlm/lock.c spin_unlock(&ls->ls_rsbtbl[b].lock); b 766 fs/dlm/lowcomms.c unsigned char *b=(unsigned char *)&peeraddr; b 769 fs/dlm/lowcomms.c b, sizeof(struct sockaddr_storage)); b 892 fs/dlm/lowcomms.c unsigned char *b = (unsigned char *)&prim.ssp_addr; b 896 fs/dlm/lowcomms.c b, sizeof(struct sockaddr_storage)); b 39 fs/ext2/balloc.c #define in_range(b, first, len) ((b) >= (first) && (b) <= (first) + (len) - 1) b 1510 fs/ext2/balloc.c static inline int test_root(int a, int b) b 1512 fs/ext2/balloc.c int num = b; b 1515 fs/ext2/balloc.c num *= b; b 748 fs/ext4/balloc.c static inline int test_root(ext4_group_t a, int b) b 751 fs/ext4/balloc.c if (a < b) b 753 fs/ext4/balloc.c if (a == b) b 755 fs/ext4/balloc.c if ((a % b) != 0) b 757 fs/ext4/balloc.c a = a / b; b 3373 fs/ext4/ext4.h #define in_range(b, first, len) ((b) >= (first) && (b) <= (first) + (len) - 1) b 2739 fs/ext4/extents.c ext4_lblk_t a, b; b 2779 fs/ext4/extents.c b = ex_ee_block+ex_ee_len - 1 < end ? b 2782 fs/ext4/extents.c ext_debug(" border %u:%u\n", a, b); b 2802 fs/ext4/extents.c } else if (b != ex_ee_block + ex_ee_len - 1) { b 2838 fs/ext4/extents.c err = ext4_remove_blocks(handle, inode, ex, partial, a, b); b 355 fs/ext4/fsmap.c struct list_head *b) b 361 fs/ext4/fsmap.c fb = container_of(b, struct ext4_fsmap, fmr_list); b 20 fs/ext4/hash.c __u32 a = in[0], b = in[1], c = in[2], d = in[3]; b 25 fs/ext4/hash.c b0 += ((b1 << 4)+a) ^ (b1+sum) ^ ((b1 >> 5)+b); b 44 fs/ext4/hash.c #define ROUND(f, a, b, c, d, x, s) \ b 45 fs/ext4/hash.c (a += f(b, c, d) + x, a = rol32(a, s)) b 55 fs/ext4/hash.c __u32 a = buf[0], b = buf[1], c = buf[2], d = buf[3]; b 58 fs/ext4/hash.c ROUND(F, a, b, c, d, in[0] + K1, 3); b 59 fs/ext4/hash.c ROUND(F, d, a, b, c, in[1] + K1, 7); b 60 fs/ext4/hash.c ROUND(F, c, d, a, b, in[2] + K1, 11); b 61 fs/ext4/hash.c ROUND(F, b, c, d, a, in[3] + K1, 19); b 62 fs/ext4/hash.c ROUND(F, a, b, c, d, in[4] + K1, 3); b 63 fs/ext4/hash.c ROUND(F, d, a, b, c, in[5] + K1, 7); b 64 fs/ext4/hash.c ROUND(F, c, d, a, b, in[6] + K1, 11); b 65 fs/ext4/hash.c ROUND(F, b, c, d, a, in[7] + K1, 19); b 68 fs/ext4/hash.c ROUND(G, a, b, c, d, in[1] + K2, 3); b 69 fs/ext4/hash.c ROUND(G, d, a, b, c, in[3] + K2, 5); b 70 fs/ext4/hash.c ROUND(G, c, d, a, b, in[5] + K2, 9); b 71 fs/ext4/hash.c ROUND(G, b, c, d, a, in[7] + K2, 13); b 72 fs/ext4/hash.c ROUND(G, a, b, c, d, in[0] + K2, 3); b 73 fs/ext4/hash.c ROUND(G, d, a, b, c, in[2] + K2, 5); b 74 fs/ext4/hash.c ROUND(G, c, d, a, b, in[4] + K2, 9); b 75 fs/ext4/hash.c ROUND(G, b, c, d, a, in[6] + K2, 13); b 78 fs/ext4/hash.c ROUND(H, a, b, c, d, in[3] + K3, 3); b 79 fs/ext4/hash.c ROUND(H, d, a, b, c, in[7] + K3, 9); b 80 fs/ext4/hash.c ROUND(H, c, d, a, b, in[2] + K3, 11); b 81 fs/ext4/hash.c ROUND(H, b, c, d, a, in[6] + K3, 15); b 82 fs/ext4/hash.c ROUND(H, a, b, c, d, in[1] + K3, 3); b 83 fs/ext4/hash.c ROUND(H, d, a, b, c, in[5] + K3, 9); b 84 fs/ext4/hash.c ROUND(H, c, d, a, b, in[0] + K3, 11); b 85 fs/ext4/hash.c ROUND(H, b, c, d, a, in[4] + K3, 15); b 88 fs/ext4/hash.c buf[1] += b; b 327 fs/ext4/indirect.c ext4_fsblk_t b, new_blocks[4]; b 362 fs/ext4/indirect.c b = new_blocks[i]; b 367 fs/ext4/indirect.c *p++ = cpu_to_le32(b++); b 4666 fs/ext4/inode.c ext4_fsblk_t b, end, table; b 4672 fs/ext4/inode.c b = block & ~((ext4_fsblk_t) ra_blks - 1); b 4673 fs/ext4/inode.c if (table > b) b 4674 fs/ext4/inode.c b = table; b 4675 fs/ext4/inode.c end = b + ra_blks; b 4682 fs/ext4/inode.c while (b <= end) b 4683 fs/ext4/inode.c sb_breadahead_unmovable(sb, b++); b 37 fs/ext4/ioctl.c static void memswap(void *a, void *b, size_t len) b 42 fs/ext4/ioctl.c bp = (unsigned char *)b; b 293 fs/ext4/page-io.c char b[BDEVNAME_SIZE]; b 296 fs/ext4/page-io.c bio_devname(bio, b), b 112 fs/ext4/resize.c #define outside(b, first, last) ((b) < (first) || (b) >= (last)) b 113 fs/ext4/resize.c #define inside(b, first, last) ((b) >= (first) && (b) < (last)) b 887 fs/ext4/super.c char b[BDEVNAME_SIZE]; b 896 fs/ext4/super.c __bdevname(dev, b), PTR_ERR(bdev)); b 3466 fs/ext4/super.c ext4_fsblk_t first_block, last_block, b; b 3479 fs/ext4/super.c b = ext4_block_bitmap(sb, gdp); b 3480 fs/ext4/super.c if (b >= first_block && b <= last_block) { b 3481 fs/ext4/super.c ext4_set_bit(EXT4_B2C(sbi, b - first_block), buf); b 3484 fs/ext4/super.c b = ext4_inode_bitmap(sb, gdp); b 3485 fs/ext4/super.c if (b >= first_block && b <= last_block) { b 3486 fs/ext4/super.c ext4_set_bit(EXT4_B2C(sbi, b - first_block), buf); b 3489 fs/ext4/super.c b = ext4_inode_table(sb, gdp); b 3490 fs/ext4/super.c if (b >= first_block && b + sbi->s_itb_per_group <= last_block) b 3491 fs/ext4/super.c for (j = 0; j < sbi->s_itb_per_group; j++, b++) { b 3492 fs/ext4/super.c int c = EXT4_B2C(sbi, b - first_block); b 88 fs/ext4/xattr.h #define EXT4_XATTR_MIN_LARGE_EA_SIZE(b) \ b 89 fs/ext4/xattr.h ((b) - EXT4_XATTR_LEN(3) - sizeof(struct ext4_xattr_header) - 4) b 279 fs/f2fs/data.c struct block_device *b = f2fs_target_device(sbi, blk_addr, NULL); b 280 fs/f2fs/data.c return bio->bi_disk == b->bd_disk && bio->bi_partno == b->bd_partno; b 110 fs/f2fs/f2fs.h #define ver_after(a, b) (typecheck(unsigned long long, a) && \ b 111 fs/f2fs/f2fs.h typecheck(unsigned long long, b) && \ b 112 fs/f2fs/f2fs.h ((long long)((a) - (b)) > 0)) b 30 fs/f2fs/hash.c __u32 a = in[0], b = in[1], c = in[2], d = in[3]; b 35 fs/f2fs/hash.c b0 += ((b1 << 4)+a) ^ (b1+sum) ^ ((b1 >> 5)+b); b 66 fs/f2fs/node.h #define nat_set_blkaddr(nat, b) ((nat)->ni.blk_addr = (b)) b 256 fs/fat/dir.c const unsigned char *b, int b_len) b 262 fs/fat/dir.c return !nls_strnicmp(sbi->nls_io, a, b, a_len); b 264 fs/fat/dir.c return !memcmp(a, b, a_len); b 669 fs/fat/inode.c struct fat_boot_sector *b; b 693 fs/fat/inode.c b = (struct fat_boot_sector *) bh->b_data; b 697 fs/fat/inode.c b->fat32.state |= FAT_STATE_DIRTY; b 699 fs/fat/inode.c b->fat32.state &= ~FAT_STATE_DIRTY; b 702 fs/fat/inode.c b->fat16.state |= FAT_STATE_DIRTY; b 704 fs/fat/inode.c b->fat16.state &= ~FAT_STATE_DIRTY; b 1431 fs/fat/inode.c static bool fat_bpb_is_zero(struct fat_boot_sector *b) b 1433 fs/fat/inode.c if (get_unaligned_le16(&b->sector_size)) b 1435 fs/fat/inode.c if (b->sec_per_clus) b 1437 fs/fat/inode.c if (b->reserved) b 1439 fs/fat/inode.c if (b->fats) b 1441 fs/fat/inode.c if (get_unaligned_le16(&b->dir_entries)) b 1443 fs/fat/inode.c if (get_unaligned_le16(&b->sectors)) b 1445 fs/fat/inode.c if (b->media) b 1447 fs/fat/inode.c if (b->fat_length) b 1449 fs/fat/inode.c if (b->secs_track) b 1451 fs/fat/inode.c if (b->heads) b 1456 fs/fat/inode.c static int fat_read_bpb(struct super_block *sb, struct fat_boot_sector *b, b 1463 fs/fat/inode.c bpb->fat_sector_size = get_unaligned_le16(&b->sector_size); b 1464 fs/fat/inode.c bpb->fat_sec_per_clus = b->sec_per_clus; b 1465 fs/fat/inode.c bpb->fat_reserved = le16_to_cpu(b->reserved); b 1466 fs/fat/inode.c bpb->fat_fats = b->fats; b 1467 fs/fat/inode.c bpb->fat_dir_entries = get_unaligned_le16(&b->dir_entries); b 1468 fs/fat/inode.c bpb->fat_sectors = get_unaligned_le16(&b->sectors); b 1469 fs/fat/inode.c bpb->fat_fat_length = le16_to_cpu(b->fat_length); b 1470 fs/fat/inode.c bpb->fat_total_sect = le32_to_cpu(b->total_sect); b 1472 fs/fat/inode.c bpb->fat16_state = b->fat16.state; b 1473 fs/fat/inode.c bpb->fat16_vol_id = get_unaligned_le32(b->fat16.vol_id); b 1475 fs/fat/inode.c bpb->fat32_length = le32_to_cpu(b->fat32.length); b 1476 fs/fat/inode.c bpb->fat32_root_cluster = le32_to_cpu(b->fat32.root_cluster); b 1477 fs/fat/inode.c bpb->fat32_info_sector = le16_to_cpu(b->fat32.info_sector); b 1478 fs/fat/inode.c bpb->fat32_state = b->fat32.state; b 1479 fs/fat/inode.c bpb->fat32_vol_id = get_unaligned_le32(b->fat32.vol_id); b 1499 fs/fat/inode.c if (!fat_valid_media(b->media)) { b 1502 fs/fat/inode.c (unsigned)b->media); b 1535 fs/fat/inode.c struct fat_boot_sector *b, int silent, b 1548 fs/fat/inode.c if (b->ignored[0] != 0xeb || b->ignored[2] != 0x90) { b 1559 fs/fat/inode.c if (!fat_bpb_is_zero(b)) { b 86 fs/fs_parser.c int ret = -ENOPARAM, b; b 164 fs/fs_parser.c b = lookup_constant(bool_names, param->string, -1); b 165 fs/fs_parser.c if (b == -1) b 167 fs/fs_parser.c result->boolean = b; b 109 fs/fscache/cookie.c const struct fscache_cookie *b) b 113 fs/fscache/cookie.c if (a->key_hash != b->key_hash) b 114 fs/fscache/cookie.c return (long)a->key_hash - (long)b->key_hash; b 115 fs/fscache/cookie.c if (a->parent != b->parent) b 116 fs/fscache/cookie.c return (long)a->parent - (long)b->parent; b 117 fs/fscache/cookie.c if (a->key_len != b->key_len) b 118 fs/fscache/cookie.c return (long)a->key_len - (long)b->key_len; b 119 fs/fscache/cookie.c if (a->type != b->type) b 120 fs/fscache/cookie.c return (long)a->type - (long)b->type; b 124 fs/fscache/cookie.c kb = &b->inline_key; b 127 fs/fscache/cookie.c kb = b->key; b 1221 fs/gfs2/dir.c static int compare_dents(const void *a, const void *b) b 1230 fs/gfs2/dir.c dent_b = *(const struct gfs2_dirent **)b; b 1336 fs/gfs2/glock.c const struct lm_lockname *b = &gh_b->gh_gl->gl_name; b 1338 fs/gfs2/glock.c if (a->ln_number > b->ln_number) b 1340 fs/gfs2/glock.c if (a->ln_number < b->ln_number) b 1514 fs/gfs2/glock.c static int glock_cmp(void *priv, struct list_head *a, struct list_head *b) b 1519 fs/gfs2/glock.c glb = list_entry(b, struct gfs2_glock, gl_lru); b 295 fs/gfs2/log.c int a, b, rm; b 301 fs/gfs2/log.c b = (tr->tr_first < new_tail); b 302 fs/gfs2/log.c rm = (wrap) ? (a || b) : (a && b); b 531 fs/gfs2/log.c static int ip_cmp(void *priv, struct list_head *a, struct list_head *b) b 536 fs/gfs2/log.c ipb = list_entry(b, struct gfs2_inode, i_ordered); b 624 fs/gfs2/lops.c static int blocknr_cmp(void *priv, struct list_head *a, struct list_head *b) b 629 fs/gfs2/lops.c bdb = list_entry(b, struct gfs2_bufdata, bd_list); b 633 fs/gfs2/quota.c static int sort_qd(const void *a, const void *b) b 636 fs/gfs2/quota.c const struct gfs2_quota_data *qd_b = *(const struct gfs2_quota_data **)b; b 87 fs/gfs2/recovery.c int wrap, a, b, revoke; b 102 fs/gfs2/recovery.c b = (where < rr->rr_where); b 103 fs/gfs2/recovery.c revoke = (wrap) ? (a || b) : (a && b); b 121 fs/gfs2/sys.c unsigned int b = test_bit(SDF_WITHDRAWN, &sdp->sd_flags); b 122 fs/gfs2/sys.c return snprintf(buf, PAGE_SIZE, "%u\n", b); b 48 fs/hfs/bfind.c int b, e; b 51 fs/hfs/bfind.c b = 0; b 55 fs/hfs/bfind.c rec = (e + b) / 2; b 70 fs/hfs/bfind.c b = rec + 1; b 73 fs/hfs/bfind.c } while (b <= e); b 119 fs/hfsplus/bfind.c int b, e; b 123 fs/hfsplus/bfind.c b = 0; b 127 fs/hfsplus/bfind.c rec = (e + b) / 2; b 135 fs/hfsplus/bfind.c if (rec_found(bnode, fd, &b, &e, &rec)) { b 139 fs/hfsplus/bfind.c } while (b <= e); b 123 fs/hpfs/alloc.c int a, b; b 138 fs/hpfs/alloc.c q = nr + n; b = 0; b 142 fs/hpfs/alloc.c if (!b) { b 144 fs/hpfs/alloc.c b = 1; b 184 fs/hpfs/alloc.c goto b; b 189 fs/hpfs/alloc.c b: b 362 fs/hpfs/alloc.c int b = hpfs_sb(s)->sb_c_bitmap & 0x0fffffff; b 379 fs/hpfs/alloc.c bmp = hpfs_map_bitmap(s, b, &qbh, "chkdn1"); b 383 fs/hpfs/alloc.c if (i == b) i++; b 187 fs/hpfs/hpfs_fn.h static inline unsigned tstbits(__le32 *bmp, unsigned b, unsigned n) b 190 fs/hpfs/hpfs_fn.h if ((b >= 0x4000) || (b + n - 1 >= 0x4000)) return n; b 191 fs/hpfs/hpfs_fn.h if (!((le32_to_cpu(bmp[(b & 0x3fff) >> 5]) >> (b & 0x1f)) & 1)) return 1; b 193 fs/hpfs/hpfs_fn.h if (!((le32_to_cpu(bmp[((b+i) & 0x3fff) >> 5]) >> ((b+i) & 0x1f)) & 1)) b 117 fs/hpfs/map.c __le32 *b; b 118 fs/hpfs/map.c if (!(b = kmalloc_array(n, 512, GFP_KERNEL))) { b 125 fs/hpfs/map.c kfree(b); b 128 fs/hpfs/map.c memcpy((char *)b + 512 * i, d, 512); b 131 fs/hpfs/map.c return b; b 271 fs/hpfs/map.c int b = 0; b 296 fs/hpfs/map.c if (hpfs_sb(s)->sb_chk >= 2) b |= 1 << de->down; b 312 fs/hpfs/map.c if (b == 3) b 495 fs/inode.c struct hlist_head *b = inode_hashtable + hash(inode->i_sb, hashval); b 499 fs/inode.c hlist_add_head(&inode->i_hash, b); b 1243 fs/inode.c struct hlist_head *b = inode_hashtable + hash(sb, ino); b 1247 fs/inode.c hlist_for_each_entry(inode, b, i_hash) { b 47 fs/jbd2/recovery.c static void journal_brelse_array(struct buffer_head *b[], int n) b 50 fs/jbd2/recovery.c brelse (b[n]); b 301 fs/jffs2/nodelist.h #define ref_totlen(a, b, c) __jffs2_ref_totlen((a), (b), (c)) b 87 fs/jffs2/os-linux.h #define jffs2_flash_writev(a,b,c,d,e,f) jffs2_flash_direct_writev(a,b,c,d,e) b 201 fs/jffs2/summary.h #define jffs2_sum_add_kvec(a,b,c,d) (0) b 202 fs/jffs2/summary.h #define jffs2_sum_move_collected(a,b) b 204 fs/jffs2/summary.h #define jffs2_sum_add_padding_mem(a,b) b 205 fs/jffs2/summary.h #define jffs2_sum_add_inode_mem(a,b,c) b 206 fs/jffs2/summary.h #define jffs2_sum_add_dirent_mem(a,b,c) b 207 fs/jffs2/summary.h #define jffs2_sum_add_xattr_mem(a,b,c) b 208 fs/jffs2/summary.h #define jffs2_sum_add_xref_mem(a,b,c) b 209 fs/jffs2/summary.h #define jffs2_sum_scan_sumnode(a,b,c,d,e) (0) b 269 fs/jffs2/wbuf.c #define jffs2_verify_write(c,b,o) (0) b 1717 fs/jfs/jfs_dmap.c s64 b, lblkno; b 1727 fs/jfs/jfs_dmap.c for (lev = level, b = *blkno; lev >= 0; lev--) { b 1731 fs/jfs/jfs_dmap.c lblkno = BLKTOCTL(b, bmp->db_l2nbperpage, lev); b 1771 fs/jfs/jfs_dmap.c b += (((s64) leafidx) << budmin); b 1782 fs/jfs/jfs_dmap.c *blkno = b; b 1836 fs/jfs/jfs_dmap.c s64 b, lblkno, n; b 1869 fs/jfs/jfs_dmap.c for (n = nblocks, b = blkno; n > 0; n -= nb, b += nb) { b 1872 fs/jfs/jfs_dmap.c lblkno = BLKTODMAP(b, bmp->db_l2nbperpage); b 1896 fs/jfs/jfs_dmap.c if ((rc = dbAllocDmap(bmp, dp, b, nb))) { b 1921 fs/jfs/jfs_dmap.c for (n = nblocks - n, b = blkno; n > 0; b 1922 fs/jfs/jfs_dmap.c n -= BPERDMAP, b += BPERDMAP) { b 1925 fs/jfs/jfs_dmap.c lblkno = BLKTODMAP(b, bmp->db_l2nbperpage); b 1939 fs/jfs/jfs_dmap.c if (dbFreeDmap(bmp, dp, b, BPERDMAP)) { b 3736 fs/jfs/jfs_dmap.c int blkno, w, b, r, nw, nb, i; b 3765 fs/jfs/jfs_dmap.c b = blkno & (DBWORD - 1); b 3767 fs/jfs/jfs_dmap.c nb = min(r, DBWORD - b); b 3773 fs/jfs/jfs_dmap.c >> b)); b 3775 fs/jfs/jfs_dmap.c >> b)); b 3803 fs/jfs/jfs_dmap.c b = blkno & (DBWORD - 1); b 3804 fs/jfs/jfs_dmap.c if (b) { b 3806 fs/jfs/jfs_dmap.c dp->wmap[w] = dp->pmap[w] = cpu_to_le32(ONES >> b); b 72 fs/jfs/jfs_dmap.h #define BLKTODMAP(b,s) \ b 73 fs/jfs/jfs_dmap.h ((((b) >> 13) + ((b) >> 23) + ((b) >> 33) + 3 + 1) << (s)) b 87 fs/jfs/jfs_dmap.h #define BLKTOL0(b,s) \ b 88 fs/jfs/jfs_dmap.h (((((b) >> 23) << 10) + ((b) >> 23) + ((b) >> 33) + 2 + 1) << (s)) b 102 fs/jfs/jfs_dmap.h #define BLKTOL1(b,s) \ b 103 fs/jfs/jfs_dmap.h (((((b) >> 33) << 20) + (((b) >> 33) << 10) + ((b) >> 33) + 1 + 1) << (s)) b 109 fs/jfs/jfs_dmap.h #define BLKTOCTL(b,s,l) \ b 110 fs/jfs/jfs_dmap.h (((l) == 2) ? 1 : ((l) == 1) ? BLKTOL1((b),(s)) : BLKTOL0((b),(s))) b 121 fs/jfs/jfs_dmap.h #define BLKTOAG(b,sbi) ((b) >> ((sbi)->bmap->db_agl2size)) b 268 fs/jfs/jfs_dmap.h #define LITOL2BSZ(n,m,b) ((((n) == 0) ? (m) : cnttz((n))) + (b)) b 271 fs/jfs/jfs_dmap.h #define BLKTOCTLLEAF(b,m) \ b 272 fs/jfs/jfs_dmap.h (((b) & (((s64)1 << ((m) + L2LPERCTL)) - 1)) >> (m)) b 3458 fs/jfs/jfs_dtree.c goto b; b 3478 fs/jfs/jfs_dtree.c b: b 139 fs/jfs/jfs_filsys.h #define LBLK2PBLK(sb,b) ((b) << (sb->s_blocksize_bits - L2PBSIZE)) b 140 fs/jfs/jfs_filsys.h #define PBLK2LBLK(sb,b) ((b) >> (sb->s_blocksize_bits - L2PBSIZE)) b 25 fs/jfs/resize.c #define BLKTODMAPN(b)\ b 26 fs/jfs/resize.c (((b) >> 13) + ((b) >> 23) + ((b) >> 33) + 3 + 1) b 63 fs/kernfs/dir.c struct kernfs_node *b) b 66 fs/kernfs/dir.c struct kernfs_root *ra = kernfs_root(a), *rb = kernfs_root(b); b 72 fs/kernfs/dir.c db = kernfs_depth(rb->kn, b); b 79 fs/kernfs/dir.c b = b->parent; b 84 fs/kernfs/dir.c while (b != a) { b 85 fs/kernfs/dir.c b = b->parent; b 92 fs/lockd/svclock.c struct nlm_block *b; b 107 fs/lockd/svclock.c b = list_entry(pos, struct nlm_block, b_list); b 108 fs/lockd/svclock.c if (time_after(b->b_when,when) || b->b_when == NLM_NEVER) b 170 fs/lockd/svclock.c static inline int nlm_cookie_match(struct nlm_cookie *a, struct nlm_cookie *b) b 172 fs/lockd/svclock.c if (a->len != b->len) b 174 fs/lockd/svclock.c if (memcmp(a->data, b->data, a->len)) b 1980 fs/namei.c unsigned long a = 0, b, x = 0, y = (unsigned long)salt; b 1992 fs/namei.c b = a ^ REPEAT_BYTE('/'); b 1993 fs/namei.c } while (!(has_zero(a, &adata, &constants) | has_zero(b, &bdata, &constants))); b 1996 fs/namei.c bdata = prep_zero_mask(b, bdata, &constants); b 193 fs/nfs/blocklayout/blocklayout.h struct pnfs_block_volume *b, gfp_t gfp_mask); b 53 fs/nfs/blocklayout/dev.c nfs4_block_decode_volume(struct xdr_stream *xdr, struct pnfs_block_volume *b) b 61 fs/nfs/blocklayout/dev.c b->type = be32_to_cpup(p++); b 63 fs/nfs/blocklayout/dev.c switch (b->type) { b 68 fs/nfs/blocklayout/dev.c b->simple.nr_sigs = be32_to_cpup(p++); b 69 fs/nfs/blocklayout/dev.c if (!b->simple.nr_sigs || b->simple.nr_sigs > PNFS_BLOCK_MAX_UUIDS) { b 70 fs/nfs/blocklayout/dev.c dprintk("Bad signature count: %d\n", b->simple.nr_sigs); b 74 fs/nfs/blocklayout/dev.c b->simple.len = 4 + 4; b 75 fs/nfs/blocklayout/dev.c for (i = 0; i < b->simple.nr_sigs; i++) { b 79 fs/nfs/blocklayout/dev.c p = xdr_decode_hyper(p, &b->simple.sigs[i].offset); b 80 fs/nfs/blocklayout/dev.c b->simple.sigs[i].sig_len = be32_to_cpup(p++); b 81 fs/nfs/blocklayout/dev.c if (b->simple.sigs[i].sig_len > PNFS_BLOCK_UUID_LEN) { b 83 fs/nfs/blocklayout/dev.c b->simple.sigs[i].sig_len); b 87 fs/nfs/blocklayout/dev.c p = xdr_inline_decode(xdr, b->simple.sigs[i].sig_len); b 90 fs/nfs/blocklayout/dev.c memcpy(&b->simple.sigs[i].sig, p, b 91 fs/nfs/blocklayout/dev.c b->simple.sigs[i].sig_len); b 93 fs/nfs/blocklayout/dev.c b->simple.len += 8 + 4 + \ b 94 fs/nfs/blocklayout/dev.c (XDR_QUADLEN(b->simple.sigs[i].sig_len) << 2); b 101 fs/nfs/blocklayout/dev.c p = xdr_decode_hyper(p, &b->slice.start); b 102 fs/nfs/blocklayout/dev.c p = xdr_decode_hyper(p, &b->slice.len); b 103 fs/nfs/blocklayout/dev.c b->slice.volume = be32_to_cpup(p++); b 110 fs/nfs/blocklayout/dev.c b->concat.volumes_count = be32_to_cpup(p++); b 111 fs/nfs/blocklayout/dev.c if (b->concat.volumes_count > PNFS_BLOCK_MAX_DEVICES) { b 112 fs/nfs/blocklayout/dev.c dprintk("Too many volumes: %d\n", b->concat.volumes_count); b 116 fs/nfs/blocklayout/dev.c p = xdr_inline_decode(xdr, b->concat.volumes_count * 4); b 119 fs/nfs/blocklayout/dev.c for (i = 0; i < b->concat.volumes_count; i++) b 120 fs/nfs/blocklayout/dev.c b->concat.volumes[i] = be32_to_cpup(p++); b 127 fs/nfs/blocklayout/dev.c p = xdr_decode_hyper(p, &b->stripe.chunk_size); b 128 fs/nfs/blocklayout/dev.c b->stripe.volumes_count = be32_to_cpup(p++); b 129 fs/nfs/blocklayout/dev.c if (b->stripe.volumes_count > PNFS_BLOCK_MAX_DEVICES) { b 130 fs/nfs/blocklayout/dev.c dprintk("Too many volumes: %d\n", b->stripe.volumes_count); b 134 fs/nfs/blocklayout/dev.c p = xdr_inline_decode(xdr, b->stripe.volumes_count * 4); b 137 fs/nfs/blocklayout/dev.c for (i = 0; i < b->stripe.volumes_count; i++) b 138 fs/nfs/blocklayout/dev.c b->stripe.volumes[i] = be32_to_cpup(p++); b 144 fs/nfs/blocklayout/dev.c b->scsi.code_set = be32_to_cpup(p++); b 145 fs/nfs/blocklayout/dev.c b->scsi.designator_type = be32_to_cpup(p++); b 146 fs/nfs/blocklayout/dev.c b->scsi.designator_len = be32_to_cpup(p++); b 147 fs/nfs/blocklayout/dev.c p = xdr_inline_decode(xdr, b->scsi.designator_len); b 150 fs/nfs/blocklayout/dev.c if (b->scsi.designator_len > 256) b 152 fs/nfs/blocklayout/dev.c memcpy(&b->scsi.designator, p, b->scsi.designator_len); b 156 fs/nfs/blocklayout/dev.c p = xdr_decode_hyper(p, &b->scsi.pr_key); b 38 fs/nfs/blocklayout/rpc_pipefs.c nfs4_encode_simple(__be32 *p, struct pnfs_block_volume *b) b 43 fs/nfs/blocklayout/rpc_pipefs.c *p++ = cpu_to_be32(b->type); b 44 fs/nfs/blocklayout/rpc_pipefs.c *p++ = cpu_to_be32(b->simple.nr_sigs); b 45 fs/nfs/blocklayout/rpc_pipefs.c for (i = 0; i < b->simple.nr_sigs; i++) { b 46 fs/nfs/blocklayout/rpc_pipefs.c p = xdr_encode_hyper(p, b->simple.sigs[i].offset); b 47 fs/nfs/blocklayout/rpc_pipefs.c p = xdr_encode_opaque(p, b->simple.sigs[i].sig, b 48 fs/nfs/blocklayout/rpc_pipefs.c b->simple.sigs[i].sig_len); b 53 fs/nfs/blocklayout/rpc_pipefs.c bl_resolve_deviceid(struct nfs_server *server, struct pnfs_block_volume *b, b 71 fs/nfs/blocklayout/rpc_pipefs.c b->simple.len += 4; /* single volume */ b 72 fs/nfs/blocklayout/rpc_pipefs.c if (b->simple.len > PAGE_SIZE) b 76 fs/nfs/blocklayout/rpc_pipefs.c msg->len = sizeof(*bl_msg) + b->simple.len; b 83 fs/nfs/blocklayout/rpc_pipefs.c bl_msg->totallen = b->simple.len; b 84 fs/nfs/blocklayout/rpc_pipefs.c nfs4_encode_simple(msg->data + sizeof(*bl_msg), b); b 167 fs/nfs/dns_resolve.c struct nfs_dns_ent *b; b 170 fs/nfs/dns_resolve.c b = container_of(cb, struct nfs_dns_ent, h); b 172 fs/nfs/dns_resolve.c if (a->namelen == 0 || a->namelen != b->namelen) b 174 fs/nfs/dns_resolve.c return memcmp(a->hostname, b->hostname, a->namelen) == 0; b 1092 fs/nfs/filelayout/filelayout.c struct pnfs_commit_bucket *b; b 1097 fs/nfs/filelayout/filelayout.c for (i = 0, b = cinfo->ds->buckets; i < cinfo->ds->nbuckets; i++, b++) { b 1098 fs/nfs/filelayout/filelayout.c list_for_each_entry_safe(freq, t, &b->written, wb_list) { b 1102 fs/nfs/filelayout/filelayout.c list_for_each_entry_safe(freq, t, &b->committing, wb_list) { b 596 fs/nfs/nfs4_fs.h #define nfs4_close_state(a, b) do { } while (0) b 597 fs/nfs/nfs4_fs.h #define nfs4_close_sync(a, b) do { } while (0) b 598 fs/nfs/nfs4_fs.h #define nfs4_state_protect(a, b, c, d) do { } while (0) b 599 fs/nfs/nfs4_fs.h #define nfs4_state_protect_write(a, b, c, d) do { } while (0) b 7844 fs/nfs/nfs4proc.c struct nfs41_server_scope *b) b 7846 fs/nfs/nfs4proc.c if (a->server_scope_sz != b->server_scope_sz) b 7848 fs/nfs/nfs4proc.c return memcmp(a->server_scope, b->server_scope, a->server_scope_sz) == 0; b 137 fs/nfs/pnfs_nfs.c struct pnfs_commit_bucket *b; b 144 fs/nfs/pnfs_nfs.c for (i = 0, b = cinfo->ds->buckets; i < cinfo->ds->nbuckets; i++, b++) { b 145 fs/nfs/pnfs_nfs.c nwritten = nfs_scan_commit_list(&b->written, dst, cinfo, 0); b 149 fs/nfs/pnfs_nfs.c if (list_empty(&b->written)) { b 150 fs/nfs/pnfs_nfs.c freeme = b->wlseg; b 151 fs/nfs/pnfs_nfs.c b->wlseg = NULL; b 346 fs/nfs/pnfs_nfs.c struct sockaddr_in *a, *b; b 355 fs/nfs/pnfs_nfs.c b = (struct sockaddr_in *)addr2; b 357 fs/nfs/pnfs_nfs.c if (a->sin_addr.s_addr == b->sin_addr.s_addr && b 358 fs/nfs/pnfs_nfs.c a->sin_port == b->sin_port) b 2436 fs/nfs/super.c static int nfs_compare_mount_options(const struct super_block *s, const struct nfs_server *b, int flags) b 2440 fs/nfs/super.c const struct rpc_clnt *clnt_b = b->client; b 2444 fs/nfs/super.c if (a->nfs_client != b->nfs_client) b 2446 fs/nfs/super.c if ((a->flags ^ b->flags) & NFS_MOUNT_CMP_FLAGMASK) b 2448 fs/nfs/super.c if (a->wsize != b->wsize) b 2450 fs/nfs/super.c if (a->rsize != b->rsize) b 2452 fs/nfs/super.c if (a->acregmin != b->acregmin) b 2454 fs/nfs/super.c if (a->acregmax != b->acregmax) b 2456 fs/nfs/super.c if (a->acdirmin != b->acdirmin) b 2458 fs/nfs/super.c if (a->acdirmax != b->acdirmax) b 199 fs/nfs_common/nfsacl.c const struct posix_acl_entry *a = x, *b = y; b 201 fs/nfs_common/nfsacl.c if (a->e_tag != b->e_tag) b 202 fs/nfs_common/nfsacl.c return a->e_tag - b->e_tag; b 203 fs/nfs_common/nfsacl.c else if ((a->e_tag == ACL_USER) && uid_gt(a->e_uid, b->e_uid)) b 205 fs/nfs_common/nfsacl.c else if ((a->e_tag == ACL_USER) && uid_lt(a->e_uid, b->e_uid)) b 207 fs/nfs_common/nfsacl.c else if ((a->e_tag == ACL_GROUP) && gid_gt(a->e_gid, b->e_gid)) b 209 fs/nfs_common/nfsacl.c else if ((a->e_tag == ACL_GROUP) && gid_lt(a->e_gid, b->e_gid)) b 150 fs/nfsd/blocklayout.c struct pnfs_block_volume *b; b 159 fs/nfsd/blocklayout.c b = &dev->volumes[0]; b 161 fs/nfsd/blocklayout.c b->type = PNFS_BLOCK_VOLUME_SIMPLE; b 162 fs/nfsd/blocklayout.c b->simple.sig_len = PNFS_BLOCK_UUID_LEN; b 163 fs/nfsd/blocklayout.c return sb->s_export_op->get_uuid(sb, b->simple.sig, &b->simple.sig_len, b 164 fs/nfsd/blocklayout.c &b->simple.offset); b 215 fs/nfsd/blocklayout.c struct pnfs_block_volume *b) b 295 fs/nfsd/blocklayout.c b->scsi.code_set = PS_CODE_SET_BINARY; b 296 fs/nfsd/blocklayout.c b->scsi.designator_type = type == 0x02 ? b 298 fs/nfsd/blocklayout.c b->scsi.designator_len = id_len; b 299 fs/nfsd/blocklayout.c memcpy(b->scsi.designator, d + 4, id_len); b 334 fs/nfsd/blocklayout.c struct pnfs_block_volume *b; b 345 fs/nfsd/blocklayout.c b = &dev->volumes[0]; b 347 fs/nfsd/blocklayout.c b->type = PNFS_BLOCK_VOLUME_SCSI; b 348 fs/nfsd/blocklayout.c b->scsi.pr_key = nfsd4_scsi_pr_key(clp); b 350 fs/nfsd/blocklayout.c error = nfsd4_scsi_identify_device(sb->s_bdev, b); b 20 fs/nfsd/blocklayoutxdr.c struct pnfs_block_extent *b = lgp->lg_content; b 31 fs/nfsd/blocklayoutxdr.c p = xdr_encode_opaque_fixed(p, &b->vol_id, b 33 fs/nfsd/blocklayoutxdr.c p = xdr_encode_hyper(p, b->foff); b 34 fs/nfsd/blocklayoutxdr.c p = xdr_encode_hyper(p, b->len); b 35 fs/nfsd/blocklayoutxdr.c p = xdr_encode_hyper(p, b->soff); b 36 fs/nfsd/blocklayoutxdr.c *p++ = cpu_to_be32(b->es); b 41 fs/nfsd/blocklayoutxdr.c nfsd4_block_encode_volume(struct xdr_stream *xdr, struct pnfs_block_volume *b) b 46 fs/nfsd/blocklayoutxdr.c switch (b->type) { b 48 fs/nfsd/blocklayoutxdr.c len = 4 + 4 + 8 + 4 + (XDR_QUADLEN(b->simple.sig_len) << 2); b 53 fs/nfsd/blocklayoutxdr.c *p++ = cpu_to_be32(b->type); b 55 fs/nfsd/blocklayoutxdr.c p = xdr_encode_hyper(p, b->simple.offset); b 56 fs/nfsd/blocklayoutxdr.c p = xdr_encode_opaque(p, b->simple.sig, b->simple.sig_len); b 59 fs/nfsd/blocklayoutxdr.c len = 4 + 4 + 4 + 4 + (XDR_QUADLEN(b->scsi.designator_len) << 2) + 8; b 64 fs/nfsd/blocklayoutxdr.c *p++ = cpu_to_be32(b->type); b 65 fs/nfsd/blocklayoutxdr.c *p++ = cpu_to_be32(b->scsi.code_set); b 66 fs/nfsd/blocklayoutxdr.c *p++ = cpu_to_be32(b->scsi.designator_type); b 67 fs/nfsd/blocklayoutxdr.c p = xdr_encode_opaque(p, b->scsi.designator, b->scsi.designator_len); b 68 fs/nfsd/blocklayoutxdr.c p = xdr_encode_hyper(p, b->scsi.pr_key); b 192 fs/nfsd/export.c static inline int expkey_match (struct cache_head *a, struct cache_head *b) b 195 fs/nfsd/export.c struct svc_expkey *new = container_of(b, struct svc_expkey, h); b 704 fs/nfsd/export.c static int svc_export_match(struct cache_head *a, struct cache_head *b) b 707 fs/nfsd/export.c struct svc_export *new = container_of(b, struct svc_export, h); b 144 fs/nfsd/nfs4idmap.c struct ent *b = container_of(cb, struct ent, h); b 146 fs/nfsd/nfs4idmap.c return (a->id == b->id && a->type == b->type && b 147 fs/nfsd/nfs4idmap.c strcmp(a->authname, b->authname) == 0); b 315 fs/nfsd/nfs4idmap.c struct ent *b = container_of(cb, struct ent, h); b 317 fs/nfsd/nfs4idmap.c return (a->type == b->type && strcmp(a->name, b->name) == 0 && b 318 fs/nfsd/nfs4idmap.c strcmp(a->authname, b->authname) == 0); b 119 fs/nfsd/nfscache.c nfsd_reply_cache_free_locked(struct nfsd_drc_bucket *b, struct svc_cacherep *rp, b 127 fs/nfsd/nfscache.c rb_erase(&rp->c_node, &b->rb_head); b 136 fs/nfsd/nfscache.c nfsd_reply_cache_free(struct nfsd_drc_bucket *b, struct svc_cacherep *rp, b 139 fs/nfsd/nfscache.c spin_lock(&b->cache_lock); b 140 fs/nfsd/nfscache.c nfsd_reply_cache_free_locked(b, rp, nn); b 141 fs/nfsd/nfscache.c spin_unlock(&b->cache_lock); b 221 fs/nfsd/nfscache.c lru_put_end(struct nfsd_drc_bucket *b, struct svc_cacherep *rp) b 224 fs/nfsd/nfscache.c list_move_tail(&rp->c_lru, &b->lru_head); b 228 fs/nfsd/nfscache.c prune_bucket(struct nfsd_drc_bucket *b, struct nfsd_net *nn) b 233 fs/nfsd/nfscache.c list_for_each_entry_safe(rp, tmp, &b->lru_head, c_lru) { b 243 fs/nfsd/nfscache.c nfsd_reply_cache_free_locked(b, rp, nn); b 260 fs/nfsd/nfscache.c struct nfsd_drc_bucket *b = &nn->drc_hashtbl[i]; b 262 fs/nfsd/nfscache.c if (list_empty(&b->lru_head)) b 264 fs/nfsd/nfscache.c spin_lock(&b->cache_lock); b 265 fs/nfsd/nfscache.c freed += prune_bucket(b, nn); b 266 fs/nfsd/nfscache.c spin_unlock(&b->cache_lock); b 338 fs/nfsd/nfscache.c nfsd_cache_insert(struct nfsd_drc_bucket *b, struct svc_cacherep *key, b 342 fs/nfsd/nfscache.c struct rb_node **p = &b->rb_head.rb_node, b 363 fs/nfsd/nfscache.c rb_insert_color(&key->c_node, &b->rb_head); b 376 fs/nfsd/nfscache.c lru_put_end(b, ret); b 395 fs/nfsd/nfscache.c struct nfsd_drc_bucket *b = &nn->drc_hashtbl[hash]; b 417 fs/nfsd/nfscache.c spin_lock(&b->cache_lock); b 418 fs/nfsd/nfscache.c found = nfsd_cache_insert(b, rp, nn); b 433 fs/nfsd/nfscache.c prune_bucket(b, nn); b 435 fs/nfsd/nfscache.c spin_unlock(&b->cache_lock); b 468 fs/nfsd/nfscache.c nfsd_reply_cache_free_locked(b, rp, nn); b 497 fs/nfsd/nfscache.c struct nfsd_drc_bucket *b; b 505 fs/nfsd/nfscache.c b = &nn->drc_hashtbl[hash]; b 512 fs/nfsd/nfscache.c nfsd_reply_cache_free(b, rp, nn); b 527 fs/nfsd/nfscache.c nfsd_reply_cache_free(b, rp, nn); b 534 fs/nfsd/nfscache.c nfsd_reply_cache_free(b, rp, nn); b 537 fs/nfsd/nfscache.c spin_lock(&b->cache_lock); b 539 fs/nfsd/nfscache.c lru_put_end(b, rp); b 543 fs/nfsd/nfscache.c spin_unlock(&b->cache_lock); b 595 fs/nfsd/state.h static inline bool nfsd4_stateid_generation_after(stateid_t *a, stateid_t *b) b 597 fs/nfsd/state.h return (s32)(a->si_generation - b->si_generation) > 0; b 694 fs/nilfs2/inode.c __u64 b; b 700 fs/nilfs2/inode.c ret = nilfs_bmap_last_key(ii->i_bmap, &b); b 706 fs/nilfs2/inode.c if (b < from) b 709 fs/nilfs2/inode.c b -= min_t(__u64, NILFS_MAX_TRUNCATE_BLOCKS, b - from); b 710 fs/nilfs2/inode.c ret = nilfs_bmap_truncate(ii->i_bmap, b); b 713 fs/nilfs2/inode.c nilfs_bmap_truncate(ii->i_bmap, b) == 0)) b 499 fs/nilfs2/page.c sector_t b; b 520 fs/nilfs2/page.c b = pvec.pages[0]->index << (PAGE_SHIFT - inode->i_blkbits); b 531 fs/nilfs2/page.c if (b < start_blk) b 535 fs/nilfs2/page.c *blkoff = b; b 540 fs/nilfs2/page.c } while (++b, bh = bh->b_this_page, bh != head); b 545 fs/nilfs2/page.c b += nblocks_in_page; b 808 fs/nilfs2/recovery.c sector_t b, end; b 827 fs/nilfs2/recovery.c b = seg_start; b 828 fs/nilfs2/recovery.c while (b <= seg_end) b 829 fs/nilfs2/recovery.c __breadahead(nilfs->ns_bdev, b++, nilfs->ns_blocksize); b 873 fs/nilfs2/recovery.c nilfs_get_segment_range(nilfs, nextnum, &b, &end); b 874 fs/nilfs2/recovery.c while (b <= end) b 875 fs/nilfs2/recovery.c __breadahead(nilfs->ns_bdev, b++, b 137 fs/nilfs2/segment.c #define nilfs_cnt32_gt(a, b) \ b 138 fs/nilfs2/segment.c (typecheck(__u32, a) && typecheck(__u32, b) && \ b 139 fs/nilfs2/segment.c ((__s32)(b) - (__s32)(a) < 0)) b 140 fs/nilfs2/segment.c #define nilfs_cnt32_ge(a, b) \ b 141 fs/nilfs2/segment.c (typecheck(__u32, a) && typecheck(__u32, b) && \ b 142 fs/nilfs2/segment.c ((__s32)(a) - (__s32)(b) >= 0)) b 143 fs/nilfs2/segment.c #define nilfs_cnt32_lt(a, b) nilfs_cnt32_gt(b, a) b 144 fs/nilfs2/segment.c #define nilfs_cnt32_le(a, b) nilfs_cnt32_ge(b, a) b 38 fs/notify/fsnotify.h struct fsnotify_group *b); b 455 fs/notify/mark.c int fsnotify_compare_groups(struct fsnotify_group *a, struct fsnotify_group *b) b 457 fs/notify/mark.c if (a == b) b 461 fs/notify/mark.c if (!b) b 463 fs/notify/mark.c if (a->priority < b->priority) b 465 fs/notify/mark.c if (a->priority > b->priority) b 467 fs/notify/mark.c if (a < b) b 1135 fs/ntfs/mft.c u8 pass, b; b 1205 fs/ntfs/mft.c b = ffz((unsigned long)*byte); b 1206 fs/ntfs/mft.c if (b < 8 && b >= (bit & 7)) { b 1207 fs/ntfs/mft.c ll = data_pos + (bit & ~7ull) + b; b 1212 fs/ntfs/mft.c *byte |= 1 << b; b 1286 fs/ntfs/mft.c u8 *b, tb; b 1332 fs/ntfs/mft.c b = (u8*)page_address(page) + (ll & ~PAGE_MASK); b 1335 fs/ntfs/mft.c if (*b != 0xff && !(*b & tb)) { b 1337 fs/ntfs/mft.c *b |= tb; b 28 fs/ntfs/mst.c int post_read_mst_fixup(NTFS_RECORD *b, const u32 size) b 34 fs/ntfs/mst.c usa_ofs = le16_to_cpu(b->usa_ofs); b 36 fs/ntfs/mst.c usa_count = le16_to_cpu(b->usa_count) - 1; b 44 fs/ntfs/mst.c usa_pos = (u16*)b + usa_ofs/sizeof(u16); b 56 fs/ntfs/mst.c data_pos = (u16*)b + NTFS_BLOCK_SIZE/sizeof(u16) - 1; b 67 fs/ntfs/mst.c b->magic = magic_BAAD; b 73 fs/ntfs/mst.c usa_count = le16_to_cpu(b->usa_count) - 1; b 74 fs/ntfs/mst.c data_pos = (u16*)b + NTFS_BLOCK_SIZE/sizeof(u16) - 1; b 109 fs/ntfs/mst.c int pre_write_mst_fixup(NTFS_RECORD *b, const u32 size) b 116 fs/ntfs/mst.c if (!b || ntfs_is_baad_record(b->magic) || b 117 fs/ntfs/mst.c ntfs_is_hole_record(b->magic)) b 120 fs/ntfs/mst.c usa_ofs = le16_to_cpu(b->usa_ofs); b 122 fs/ntfs/mst.c usa_count = le16_to_cpu(b->usa_count) - 1; b 130 fs/ntfs/mst.c usa_pos = (le16*)((u8*)b + usa_ofs); b 141 fs/ntfs/mst.c data_pos = (le16*)b + NTFS_BLOCK_SIZE/sizeof(le16) - 1; b 165 fs/ntfs/mst.c void post_write_mst_fixup(NTFS_RECORD *b) b 169 fs/ntfs/mst.c u16 usa_ofs = le16_to_cpu(b->usa_ofs); b 170 fs/ntfs/mst.c u16 usa_count = le16_to_cpu(b->usa_count) - 1; b 173 fs/ntfs/mst.c usa_pos = (le16*)b + usa_ofs/sizeof(le16); b 176 fs/ntfs/mst.c data_pos = (le16*)b + NTFS_BLOCK_SIZE/sizeof(le16) - 1; b 89 fs/ntfs/ntfs.h extern int post_read_mst_fixup(NTFS_RECORD *b, const u32 size); b 90 fs/ntfs/ntfs.h extern int pre_write_mst_fixup(NTFS_RECORD *b, const u32 size); b 91 fs/ntfs/ntfs.h extern void post_write_mst_fixup(NTFS_RECORD *b); b 747 fs/ntfs/runlist.c u8 b; /* Current byte offset in buf. */ b 812 fs/ntfs/runlist.c b = *buf & 0xf; b 813 fs/ntfs/runlist.c if (b) { b 814 fs/ntfs/runlist.c if (unlikely(buf + b > attr_end)) b 816 fs/ntfs/runlist.c for (deltaxcn = (s8)buf[b--]; b; b--) b 817 fs/ntfs/runlist.c deltaxcn = (deltaxcn << 8) + buf[b]; b 849 fs/ntfs/runlist.c b = b2 + ((*buf >> 4) & 0xf); b 850 fs/ntfs/runlist.c if (buf + b > attr_end) b 852 fs/ntfs/runlist.c for (deltaxcn = (s8)buf[b--]; b > b2; b--) b 853 fs/ntfs/runlist.c deltaxcn = (deltaxcn << 8) + buf[b]; b 570 fs/ntfs/super.c const NTFS_BOOT_SECTOR *b, const bool silent) b 579 fs/ntfs/super.c if ((void*)b < (void*)&b->checksum && b->checksum && !silent) { b 583 fs/ntfs/super.c for (i = 0, u = (le32*)b; u < (le32*)(&b->checksum); ++u) b 585 fs/ntfs/super.c if (le32_to_cpu(b->checksum) != i) b 589 fs/ntfs/super.c if (b->oem_id != magicNTFS) b 592 fs/ntfs/super.c if (le16_to_cpu(b->bpb.bytes_per_sector) < 0x100 || b 593 fs/ntfs/super.c le16_to_cpu(b->bpb.bytes_per_sector) > 0x1000) b 596 fs/ntfs/super.c switch (b->bpb.sectors_per_cluster) { b 603 fs/ntfs/super.c if ((u32)le16_to_cpu(b->bpb.bytes_per_sector) * b 604 fs/ntfs/super.c b->bpb.sectors_per_cluster > NTFS_MAX_CLUSTER_SIZE) b 607 fs/ntfs/super.c if (le16_to_cpu(b->bpb.reserved_sectors) || b 608 fs/ntfs/super.c le16_to_cpu(b->bpb.root_entries) || b 609 fs/ntfs/super.c le16_to_cpu(b->bpb.sectors) || b 610 fs/ntfs/super.c le16_to_cpu(b->bpb.sectors_per_fat) || b 611 fs/ntfs/super.c le32_to_cpu(b->bpb.large_sectors) || b->bpb.fats) b 614 fs/ntfs/super.c if ((u8)b->clusters_per_mft_record < 0xe1 || b 615 fs/ntfs/super.c (u8)b->clusters_per_mft_record > 0xf7) b 616 fs/ntfs/super.c switch (b->clusters_per_mft_record) { b 623 fs/ntfs/super.c if ((u8)b->clusters_per_index_record < 0xe1 || b 624 fs/ntfs/super.c (u8)b->clusters_per_index_record > 0xf7) b 625 fs/ntfs/super.c switch (b->clusters_per_index_record) { b 636 fs/ntfs/super.c if (!silent && b->end_of_sector_marker != cpu_to_le16(0xaa55)) b 752 fs/ntfs/super.c static bool parse_ntfs_boot_sector(ntfs_volume *vol, const NTFS_BOOT_SECTOR *b) b 758 fs/ntfs/super.c vol->sector_size = le16_to_cpu(b->bpb.bytes_per_sector); b 771 fs/ntfs/super.c ntfs_debug("sectors_per_cluster = 0x%x", b->bpb.sectors_per_cluster); b 772 fs/ntfs/super.c sectors_per_cluster_bits = ffs(b->bpb.sectors_per_cluster) - 1; b 775 fs/ntfs/super.c nr_hidden_sects = le32_to_cpu(b->bpb.hidden_sectors); b 790 fs/ntfs/super.c clusters_per_mft_record = b->clusters_per_mft_record; b 830 fs/ntfs/super.c clusters_per_index_record = b->clusters_per_index_record; b 866 fs/ntfs/super.c ll = sle64_to_cpu(b->number_of_sectors) >> sectors_per_cluster_bits; b 888 fs/ntfs/super.c ll = sle64_to_cpu(b->mft_lcn); b 897 fs/ntfs/super.c ll = sle64_to_cpu(b->mftmirr_lcn); b 922 fs/ntfs/super.c vol->serial_no = le64_to_cpu(b->volume_serial_number); b 58 fs/ocfs2/blockcheck.c unsigned int b, p = 0; b 64 fs/ocfs2/blockcheck.c b = i + 1; b 69 fs/ocfs2/blockcheck.c b += p; b 79 fs/ocfs2/blockcheck.c for (; (1 << p) < (b + 1); p++) b 80 fs/ocfs2/blockcheck.c b++; b 85 fs/ocfs2/blockcheck.c return b; b 101 fs/ocfs2/blockcheck.c unsigned int i, b, p = 0; b 120 fs/ocfs2/blockcheck.c b = calc_code_bit(nr + i, &p); b 138 fs/ocfs2/blockcheck.c parity ^= b; b 161 fs/ocfs2/blockcheck.c unsigned int i, b; b 184 fs/ocfs2/blockcheck.c b = calc_code_bit(nr, NULL); b 186 fs/ocfs2/blockcheck.c if (fix < b) b 189 fs/ocfs2/blockcheck.c for (i = 0; i < d; i++, b++) b 192 fs/ocfs2/blockcheck.c while (hweight32(b) == 1) b 193 fs/ocfs2/blockcheck.c b++; b 203 fs/ocfs2/blockcheck.c if (b == fix) b 202 fs/ocfs2/cluster/tcp.c # define o2net_init_nst(a, b, c, d, e) b 206 fs/ocfs2/cluster/tcp.c # define o2net_set_nst_sock_container(a, b) b 207 fs/ocfs2/cluster/tcp.c # define o2net_set_nst_msg_id(a, b) b 246 fs/ocfs2/cluster/tcp.c # define o2net_update_send_stats(a, b) b 213 fs/ocfs2/dir.c __u32 a = in[0], b = in[1], c = in[2], d = in[3]; b 218 fs/ocfs2/dir.c b0 += ((b1 << 4)+a) ^ (b1+sum) ^ ((b1 >> 5)+b); b 673 fs/ocfs2/dir.c unsigned long start, block, b; b 697 fs/ocfs2/dir.c b = block; b 704 fs/ocfs2/dir.c if (b >= nblocks || (num && block == start)) { b 711 fs/ocfs2/dir.c err = ocfs2_read_dir_block(dir, b++, &bh, b 3482 fs/ocfs2/dir.c static int dx_leaf_sort_cmp(const void *a, const void *b) b 3485 fs/ocfs2/dir.c const struct ocfs2_dx_entry *entry2 = b; b 3506 fs/ocfs2/dir.c static void dx_leaf_sort_swap(void *a, void *b, int size) b 3509 fs/ocfs2/dir.c struct ocfs2_dx_entry *entry2 = b; b 1048 fs/ocfs2/dlm/dlmmaster.c int b; b 1052 fs/ocfs2/dlm/dlmmaster.c b = (mle->type == DLM_MLE_BLOCK); b 1053 fs/ocfs2/dlm/dlmmaster.c if ((*blocked && !b) || (!*blocked && b)) { b 1056 fs/ocfs2/dlm/dlmmaster.c *blocked, b); b 1057 fs/ocfs2/dlm/dlmmaster.c *blocked = b; b 149 fs/ocfs2/dlmglue.h #define ocfs2_inode_lock_nested(i, b, e, s)\ b 150 fs/ocfs2/dlmglue.h ocfs2_inode_lock_full_nested(i, b, e, 0, s) b 153 fs/ocfs2/dlmglue.h #define ocfs2_inode_lock(i, b, e) ocfs2_inode_lock_full_nested(i, b, e, 0, OI_LS_NORMAL) b 154 fs/ocfs2/dlmglue.h #define ocfs2_try_inode_lock(i, b, e)\ b 155 fs/ocfs2/dlmglue.h ocfs2_inode_lock_full_nested(i, b, e, OCFS2_META_LOCK_NOQUEUE,\ b 33 fs/ocfs2/ioctl.c #define o2info_from_user(a, b) \ b 34 fs/ocfs2/ioctl.c copy_from_user(&(a), (b), sizeof(a)) b 35 fs/ocfs2/ioctl.c #define o2info_to_user(a, b) \ b 36 fs/ocfs2/ioctl.c copy_to_user((typeof(a) __user *)b, &(a), sizeof(a)) b 1369 fs/ocfs2/refcounttree.c static int cmp_refcount_rec_by_low_cpos(const void *a, const void *b) b 1371 fs/ocfs2/refcounttree.c const struct ocfs2_refcount_rec *l = a, *r = b; b 1382 fs/ocfs2/refcounttree.c static int cmp_refcount_rec_by_cpos(const void *a, const void *b) b 1384 fs/ocfs2/refcounttree.c const struct ocfs2_refcount_rec *l = a, *r = b; b 1395 fs/ocfs2/refcounttree.c static void swap_refcount_rec(void *a, void *b, int size) b 1397 fs/ocfs2/refcounttree.c struct ocfs2_refcount_rec *l = a, *r = b; b 68 fs/ocfs2/slot_map.c int b, i, slotno; b 72 fs/ocfs2/slot_map.c for (b = 0; b < si->si_blocks; b++) { b 73 fs/ocfs2/slot_map.c se = (struct ocfs2_slot_map_extended *)si->si_bh[b]->b_data; b 4154 fs/ocfs2/xattr.c static int cmp_xe(const void *a, const void *b) b 4156 fs/ocfs2/xattr.c const struct ocfs2_xattr_entry *l = a, *r = b; b 4167 fs/ocfs2/xattr.c static void swap_xe(void *a, void *b, int size) b 4169 fs/ocfs2/xattr.c struct ocfs2_xattr_entry *l = a, *r = b, tmp; b 4366 fs/ocfs2/xattr.c static int cmp_xe_offset(const void *a, const void *b) b 4368 fs/ocfs2/xattr.c const struct ocfs2_xattr_entry *l = a, *r = b; b 36 fs/reiserfs/hashes.c b0 += ((b1 << 4)+a) ^ (b1+sum) ^ ((b1 >> 5)+b); \ b 49 fs/reiserfs/hashes.c u32 a, b, c, d; b 61 fs/reiserfs/hashes.c b = (u32) msg[4] | b 79 fs/reiserfs/hashes.c b = (u32) msg[4] | b 93 fs/reiserfs/hashes.c b = (u32) msg[4] | b 105 fs/reiserfs/hashes.c b = c = d = pad; b 107 fs/reiserfs/hashes.c b <<= 8; b 108 fs/reiserfs/hashes.c b |= msg[i]; b 111 fs/reiserfs/hashes.c a = b = c = d = pad; b 2604 fs/reiserfs/journal.c char b[BDEVNAME_SIZE]; b 2627 fs/reiserfs/journal.c __bdevname(jdev, b), result); b 410 fs/reiserfs/procfs.c char b[BDEVNAME_SIZE]; b 414 fs/reiserfs/procfs.c strlcpy(b, sb->s_id, BDEVNAME_SIZE); b 415 fs/reiserfs/procfs.c s = strchr(b, '/'); b 420 fs/reiserfs/procfs.c REISERFS_SB(sb)->procdir = proc_mkdir_data(b, 0, proc_info_root, sb); b 432 fs/reiserfs/procfs.c proc_info_root_name, b); b 440 fs/reiserfs/procfs.c char b[BDEVNAME_SIZE]; b 444 fs/reiserfs/procfs.c strlcpy(b, sb->s_id, BDEVNAME_SIZE); b 445 fs/reiserfs/procfs.c s = strchr(b, '/'); b 449 fs/reiserfs/procfs.c remove_proc_subtree(b, proc_info_root); b 531 fs/reiserfs/stree.c b_blocknr_t *b, int num) b 537 fs/reiserfs/stree.c bh[i] = sb_getblk(s, b[i]); b 804 fs/reiserfs/xattr.c struct listxattr_buf *b = b 812 fs/reiserfs/xattr.c handler = find_xattr_handler_prefix(b->dentry->d_sb->s_xattr, b 815 fs/reiserfs/xattr.c (handler->list && !handler->list(b->dentry))) b 818 fs/reiserfs/xattr.c if (b->buf) { b 819 fs/reiserfs/xattr.c if (b->pos + size > b->size) { b 820 fs/reiserfs/xattr.c b->pos = -ERANGE; b 823 fs/reiserfs/xattr.c memcpy(b->buf + b->pos, name, namelen); b 824 fs/reiserfs/xattr.c b->buf[b->pos + namelen] = 0; b 826 fs/reiserfs/xattr.c b->pos += size; b 85 fs/squashfs/block.c int bytes, compressed, b = 0, k = 0, avail, i; b 109 fs/squashfs/block.c for (b = 0; bytes < length; b++, cur_index++) { b 110 fs/squashfs/block.c bh[b] = sb_getblk(sb, cur_index); b 111 fs/squashfs/block.c if (bh[b] == NULL) b 115 fs/squashfs/block.c ll_rw_block(REQ_OP_READ, 0, b, bh); b 126 fs/squashfs/block.c b = 1; b 141 fs/squashfs/block.c for (; bytes < length; b++) { b 142 fs/squashfs/block.c bh[b] = sb_getblk(sb, ++cur_index); b 143 fs/squashfs/block.c if (bh[b] == NULL) b 147 fs/squashfs/block.c ll_rw_block(REQ_OP_READ, 0, b - 1, bh + 1); b 150 fs/squashfs/block.c for (i = 0; i < b; i++) { b 159 fs/squashfs/block.c length = squashfs_decompress(msblk, bh, b, offset, length, b 170 fs/squashfs/block.c for (bytes = length; k < b; k++) { b 196 fs/squashfs/block.c for (; k < b; k++) b 184 fs/squashfs/decompressor_multi.c int b, int offset, int length, struct squashfs_page_actor *output) b 190 fs/squashfs/decompressor_multi.c bh, b, offset, length, output); b 76 fs/squashfs/decompressor_multi_percpu.c int b, int offset, int length, struct squashfs_page_actor *output) b 81 fs/squashfs/decompressor_multi_percpu.c int res = msblk->decompressor->decompress(msblk, stream->stream, bh, b, b 63 fs/squashfs/decompressor_single.c int b, int offset, int length, struct squashfs_page_actor *output) b 69 fs/squashfs/decompressor_single.c res = msblk->decompressor->decompress(msblk, stream->stream, bh, b, b 92 fs/squashfs/lz4_wrapper.c struct buffer_head **bh, int b, int offset, int length, b 99 fs/squashfs/lz4_wrapper.c for (i = 0; i < b; i++) { b 66 fs/squashfs/lzo_wrapper.c struct buffer_head **bh, int b, int offset, int length, b 74 fs/squashfs/lzo_wrapper.c for (i = 0; i < b; i++) { b 120 fs/squashfs/xz_wrapper.c struct buffer_head **bh, int b, int offset, int length, b 135 fs/squashfs/xz_wrapper.c if (stream->buf.in_pos == stream->buf.in_size && k < b) { b 154 fs/squashfs/xz_wrapper.c if (stream->buf.in_pos == stream->buf.in_size && k < b) b 160 fs/squashfs/xz_wrapper.c if (xz_err != XZ_STREAM_END || k < b) b 166 fs/squashfs/xz_wrapper.c for (; k < b; k++) b 53 fs/squashfs/zlib_wrapper.c struct buffer_head **bh, int b, int offset, int length, b 64 fs/squashfs/zlib_wrapper.c if (stream->avail_in == 0 && k < b) { b 89 fs/squashfs/zlib_wrapper.c if (stream->avail_in == 0 && k < b) b 102 fs/squashfs/zlib_wrapper.c if (k < b) b 108 fs/squashfs/zlib_wrapper.c for (; k < b; k++) b 62 fs/squashfs/zstd_wrapper.c struct buffer_head **bh, int b, int offset, int length, b 84 fs/squashfs/zstd_wrapper.c if (in_buf.pos == in_buf.size && k < b) { b 111 fs/squashfs/zstd_wrapper.c if (in_buf.pos == in_buf.size && k < b) b 123 fs/squashfs/zstd_wrapper.c if (k < b) b 129 fs/squashfs/zstd_wrapper.c for (; k < b; k++) b 288 fs/stat.c # define choose_32_64(a,b) a b 290 fs/stat.c # define choose_32_64(a,b) b b 730 fs/ubifs/find.c const struct ubifs_lprops **b) b 733 fs/ubifs/find.c const struct ubifs_lprops *lpb = *b; b 109 fs/ubifs/gc.c static int data_nodes_cmp(void *priv, struct list_head *a, struct list_head *b) b 116 fs/ubifs/gc.c if (a == b) b 120 fs/ubifs/gc.c sb = list_entry(b, struct ubifs_scan_node, list); b 153 fs/ubifs/gc.c struct list_head *b) b 160 fs/ubifs/gc.c if (a == b) b 164 fs/ubifs/gc.c sb = list_entry(b, struct ubifs_scan_node, list); b 115 fs/ubifs/log.c struct ubifs_bud *b; b 122 fs/ubifs/log.c b = rb_entry(parent, struct ubifs_bud, rb); b 123 fs/ubifs/log.c ubifs_assert(c, bud->lnum != b->lnum); b 124 fs/ubifs/log.c if (bud->lnum < b->lnum) b 172 fs/ubifs/lprops.c const int b = LPT_HEAP_SZ / 2 - 1; b 177 fs/ubifs/lprops.c cpos = (((size_t)lprops >> 4) & b) + b; b 178 fs/ubifs/lprops.c ubifs_assert(c, cpos >= b); b 225 fs/ubifs/lpt.c int b = *pos; b 232 fs/ubifs/lpt.c if (b) { b 233 fs/ubifs/lpt.c *p |= ((uint8_t)val) << b; b 234 fs/ubifs/lpt.c nrbits += b; b 236 fs/ubifs/lpt.c *++p = (uint8_t)(val >>= (8 - b)); b 257 fs/ubifs/lpt.c b = nrbits & 7; b 258 fs/ubifs/lpt.c if (b == 0) b 261 fs/ubifs/lpt.c *pos = b; b 277 fs/ubifs/lpt.c int b = *pos; b 279 fs/ubifs/lpt.c const int bytes = (nrbits + b + 7) >> 3; b 285 fs/ubifs/lpt.c if (b) { b 302 fs/ubifs/lpt.c val <<= (8 - b); b 303 fs/ubifs/lpt.c val |= *p >> b; b 304 fs/ubifs/lpt.c nrbits += b; b 326 fs/ubifs/lpt.c b = nrbits & 7; b 329 fs/ubifs/lpt.c *pos = b; b 330 fs/ubifs/lpt.c ubifs_assert(c, (val >> nrbits) == 0 || nrbits - b == 32); b 88 fs/ubifs/replay.c static int set_bud_lprops(struct ubifs_info *c, struct bud_entry *b) b 95 fs/ubifs/replay.c lp = ubifs_lpt_lookup_dirty(c, b->bud->lnum); b 102 fs/ubifs/replay.c if (b->bud->start == 0 && (lp->free != c->leb_size || lp->dirty != 0)) { b 122 fs/ubifs/replay.c dbg_mnt("bud LEB %d was GC'd (%d free, %d dirty)", b->bud->lnum, b 124 fs/ubifs/replay.c dbg_gc("bud LEB %d was GC'd (%d free, %d dirty)", b->bud->lnum, b 136 fs/ubifs/replay.c b->bud->lnum, lp->free, lp->dirty, b->free, b 137 fs/ubifs/replay.c b->dirty); b 139 fs/ubifs/replay.c lp = ubifs_change_lp(c, lp, b->free, dirty + b->dirty, b 147 fs/ubifs/replay.c err = ubifs_wbuf_seek_nolock(&c->jheads[b->bud->jhead].wbuf, b 148 fs/ubifs/replay.c b->bud->lnum, c->leb_size - b->free); b 164 fs/ubifs/replay.c struct bud_entry *b; b 167 fs/ubifs/replay.c list_for_each_entry(b, &c->replay_buds, list) { b 168 fs/ubifs/replay.c err = set_bud_lprops(c, b); b 302 fs/ubifs/replay.c struct list_head *b) b 308 fs/ubifs/replay.c if (a == b) b 312 fs/ubifs/replay.c rb = list_entry(b, struct replay_entry, list); b 671 fs/ubifs/replay.c static int replay_bud(struct ubifs_info *c, struct bud_entry *b) b 673 fs/ubifs/replay.c int is_last = is_last_bud(c, b->bud); b 674 fs/ubifs/replay.c int err = 0, used = 0, lnum = b->bud->lnum, offs = b->bud->start; b 680 fs/ubifs/replay.c lnum, b->bud->jhead, offs, is_last); b 689 fs/ubifs/replay.c sleb = ubifs_recover_leb(c, lnum, offs, c->sbuf, b->bud->jhead); b 695 fs/ubifs/replay.c n_nodes = authenticate_sleb(c, sleb, b->bud->log_hash, is_last); b 701 fs/ubifs/replay.c ubifs_shash_copy_state(c, b->bud->log_hash, b 702 fs/ubifs/replay.c c->jheads[b->bud->jhead].log_hash); b 827 fs/ubifs/replay.c b->dirty = sleb->endpt - offs - used; b 828 fs/ubifs/replay.c b->free = c->leb_size - sleb->endpt; b 830 fs/ubifs/replay.c lnum, b->dirty, b->free); b 852 fs/ubifs/replay.c struct bud_entry *b; b 856 fs/ubifs/replay.c list_for_each_entry(b, &c->replay_buds, list) { b 857 fs/ubifs/replay.c err = replay_bud(c, b); b 861 fs/ubifs/replay.c ubifs_assert(c, b->sqnum > prev_sqnum); b 862 fs/ubifs/replay.c prev_sqnum = b->sqnum; b 874 fs/ubifs/replay.c struct bud_entry *b; b 877 fs/ubifs/replay.c b = list_entry(c->replay_buds.next, struct bud_entry, list); b 878 fs/ubifs/replay.c list_del(&b->list); b 879 fs/ubifs/replay.c kfree(b); b 898 fs/ubifs/replay.c struct bud_entry *b; b 907 fs/ubifs/replay.c b = kmalloc(sizeof(struct bud_entry), GFP_KERNEL); b 908 fs/ubifs/replay.c if (!b) { b 926 fs/ubifs/replay.c b->bud = bud; b 927 fs/ubifs/replay.c b->sqnum = sqnum; b 928 fs/ubifs/replay.c list_add_tail(&b->list, &c->replay_buds); b 933 fs/ubifs/replay.c kfree(b); b 155 fs/ufs/ufs.h static inline u64 ufs_dtog(struct ufs_sb_private_info * uspi, u64 b) b 157 fs/ufs/ufs.h do_div(b, uspi->s_fpg); b 158 fs/ufs/ufs.h return b; b 161 fs/ufs/ufs.h static inline u32 ufs_dtogd(struct ufs_sb_private_info * uspi, u64 b) b 163 fs/ufs/ufs.h return do_div(b, uspi->s_fpg); b 208 fs/ufs/ufs_fs.h #define ufs_fsbtodb(uspi, b) ((b) << (uspi)->s_fsbtodb) b 209 fs/ufs/ufs_fs.h #define ufs_dbtofsb(uspi, b) ((b) >> (uspi)->s_fsbtodb) b 18 fs/ufs/util.h #define in_range(b,first,len) ((b)>=(first)&&(b)<(first)+(len)) b 34 fs/xfs/libxfs/xfs_alloc.c #define XFS_ABSDIFF(a,b) (((a) <= (b)) ? ((b) - (a)) : ((a) - (b))) b 102 fs/xfs/libxfs/xfs_bmap.c cur->bc_rec.b = *irec; b 111 fs/xfs/libxfs/xfs_bmap.c cur->bc_rec.b.br_startoff = 0; b 112 fs/xfs/libxfs/xfs_bmap.c cur->bc_rec.b.br_startblock = 0; b 113 fs/xfs/libxfs/xfs_bmap.c cur->bc_rec.b.br_blockcount = 0; b 690 fs/xfs/libxfs/xfs_bmap.c cur->bc_private.b.flags = wasdel ? XFS_BTCUR_BPRV_WASDEL : 0; b 727 fs/xfs/libxfs/xfs_bmap.c cur->bc_private.b.allocated++; b 946 fs/xfs/libxfs/xfs_bmap.c cur->bc_private.b.allocated = 0; b 973 fs/xfs/libxfs/xfs_bmap.c cur->bc_private.b.allocated = 0; b 1558 fs/xfs/libxfs/xfs_bmap.c (bma->cur->bc_private.b.flags & XFS_BTCUR_BPRV_WASDEL)); b 1818 fs/xfs/libxfs/xfs_bmap.c (bma->cur ? bma->cur->bc_private.b.allocated : 0)); b 1895 fs/xfs/libxfs/xfs_bmap.c (bma->cur ? bma->cur->bc_private.b.allocated : 0)); b 2010 fs/xfs/libxfs/xfs_bmap.c da_new += bma->cur->bc_private.b.allocated; b 2011 fs/xfs/libxfs/xfs_bmap.c bma->cur->bc_private.b.allocated = 0; b 2325 fs/xfs/libxfs/xfs_bmap.c cur->bc_rec.b = *new; b 2439 fs/xfs/libxfs/xfs_bmap.c cur->bc_rec.b = PREV; b 2489 fs/xfs/libxfs/xfs_bmap.c cur->bc_private.b.allocated = 0; b 2668 fs/xfs/libxfs/xfs_bmap.c ASSERT(!cur || !(cur->bc_private.b.flags & XFS_BTCUR_BPRV_WASDEL)); b 2850 fs/xfs/libxfs/xfs_bmap.c cur->bc_private.b.allocated = 0; b 4085 fs/xfs/libxfs/xfs_bmap.c bma->cur->bc_private.b.flags = b 4609 fs/xfs/libxfs/xfs_bmap.c cur->bc_private.b.flags = 0; b 5094 fs/xfs/libxfs/xfs_bmap.c cur->bc_rec.b = new; b 5258 fs/xfs/libxfs/xfs_bmap.c cur->bc_private.b.flags = 0; b 5511 fs/xfs/libxfs/xfs_bmap.c cur->bc_private.b.allocated = 0; b 5729 fs/xfs/libxfs/xfs_bmap.c cur->bc_private.b.flags = 0; b 5847 fs/xfs/libxfs/xfs_bmap.c cur->bc_private.b.flags = 0; b 5967 fs/xfs/libxfs/xfs_bmap.c cur->bc_private.b.flags = 0; b 6017 fs/xfs/libxfs/xfs_bmap.c cur->bc_private.b.allocated = 0; b 169 fs/xfs/libxfs/xfs_bmap_btree.c cur->bc_private.b.ip, cur->bc_private.b.whichfork); b 175 fs/xfs/libxfs/xfs_bmap_btree.c new->bc_private.b.flags = cur->bc_private.b.flags; b 186 fs/xfs/libxfs/xfs_bmap_btree.c (dst->bc_private.b.ip->i_d.di_flags & XFS_DIFLAG_REALTIME)); b 188 fs/xfs/libxfs/xfs_bmap_btree.c dst->bc_private.b.allocated += src->bc_private.b.allocated; b 191 fs/xfs/libxfs/xfs_bmap_btree.c src->bc_private.b.allocated = 0; b 208 fs/xfs/libxfs/xfs_bmap_btree.c xfs_rmap_ino_bmbt_owner(&args.oinfo, cur->bc_private.b.ip->i_ino, b 209 fs/xfs/libxfs/xfs_bmap_btree.c cur->bc_private.b.whichfork); b 233 fs/xfs/libxfs/xfs_bmap_btree.c args.wasdel = cur->bc_private.b.flags & XFS_BTCUR_BPRV_WASDEL; b 262 fs/xfs/libxfs/xfs_bmap_btree.c cur->bc_private.b.allocated++; b 263 fs/xfs/libxfs/xfs_bmap_btree.c cur->bc_private.b.ip->i_d.di_nblocks++; b 264 fs/xfs/libxfs/xfs_bmap_btree.c xfs_trans_log_inode(args.tp, cur->bc_private.b.ip, XFS_ILOG_CORE); b 265 fs/xfs/libxfs/xfs_bmap_btree.c xfs_trans_mod_dquot_byino(args.tp, cur->bc_private.b.ip, b 283 fs/xfs/libxfs/xfs_bmap_btree.c struct xfs_inode *ip = cur->bc_private.b.ip; b 288 fs/xfs/libxfs/xfs_bmap_btree.c xfs_rmap_ino_bmbt_owner(&oinfo, ip->i_ino, cur->bc_private.b.whichfork); b 305 fs/xfs/libxfs/xfs_bmap_btree.c ifp = XFS_IFORK_PTR(cur->bc_private.b.ip, b 306 fs/xfs/libxfs/xfs_bmap_btree.c cur->bc_private.b.whichfork); b 323 fs/xfs/libxfs/xfs_bmap_btree.c ifp = XFS_IFORK_PTR(cur->bc_private.b.ip, b 324 fs/xfs/libxfs/xfs_bmap_btree.c cur->bc_private.b.whichfork); b 350 fs/xfs/libxfs/xfs_bmap_btree.c return xfs_bmdr_maxrecs(cur->bc_private.b.forksize, level == 0); b 377 fs/xfs/libxfs/xfs_bmap_btree.c xfs_bmbt_disk_set_all(&rec->bmbt, &cur->bc_rec.b); b 394 fs/xfs/libxfs/xfs_bmap_btree.c cur->bc_rec.b.br_startoff; b 404 fs/xfs/libxfs/xfs_bmap_btree.c uint64_t b = be64_to_cpu(k2->bmbt.br_startoff); b 412 fs/xfs/libxfs/xfs_bmap_btree.c if (a > b) b 414 fs/xfs/libxfs/xfs_bmap_btree.c if (b > a) b 569 fs/xfs/libxfs/xfs_bmap_btree.c cur->bc_private.b.forksize = XFS_IFORK_SIZE(ip, whichfork); b 570 fs/xfs/libxfs/xfs_bmap_btree.c cur->bc_private.b.ip = ip; b 571 fs/xfs/libxfs/xfs_bmap_btree.c cur->bc_private.b.allocated = 0; b 572 fs/xfs/libxfs/xfs_bmap_btree.c cur->bc_private.b.flags = 0; b 573 fs/xfs/libxfs/xfs_bmap_btree.c cur->bc_private.b.whichfork = whichfork; b 647 fs/xfs/libxfs/xfs_bmap_btree.c cur->bc_private.b.flags |= XFS_BTCUR_BPRV_INVALID_OWNER; b 239 fs/xfs/libxfs/xfs_btree.c cur->bc_private.b.ip->i_ino, b 240 fs/xfs/libxfs/xfs_btree.c cur->bc_private.b.whichfork, cur->bc_btnum, b 383 fs/xfs/libxfs/xfs_btree.c cur->bc_private.b.allocated == 0); b 659 fs/xfs/libxfs/xfs_btree.c ifp = XFS_IFORK_PTR(cur->bc_private.b.ip, cur->bc_private.b.whichfork); b 1042 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_block *b; /* btree block */ b 1049 fs/xfs/libxfs/xfs_btree.c b = XFS_BUF_TO_BLOCK(bp); b 1051 fs/xfs/libxfs/xfs_btree.c if (b->bb_u.l.bb_leftsib == cpu_to_be64(NULLFSBLOCK)) b 1053 fs/xfs/libxfs/xfs_btree.c if (b->bb_u.l.bb_rightsib == cpu_to_be64(NULLFSBLOCK)) b 1056 fs/xfs/libxfs/xfs_btree.c if (b->bb_u.s.bb_leftsib == cpu_to_be32(NULLAGBLOCK)) b 1058 fs/xfs/libxfs/xfs_btree.c if (b->bb_u.s.bb_rightsib == cpu_to_be32(NULLAGBLOCK)) b 1204 fs/xfs/libxfs/xfs_btree.c owner = cur->bc_private.b.ip->i_ino; b 1454 fs/xfs/libxfs/xfs_btree.c xfs_trans_log_inode(cur->bc_tp, cur->bc_private.b.ip, b 1455 fs/xfs/libxfs/xfs_btree.c xfs_ilog_fbroot(cur->bc_private.b.whichfork)); b 1497 fs/xfs/libxfs/xfs_btree.c xfs_trans_log_inode(cur->bc_tp, cur->bc_private.b.ip, b 1498 fs/xfs/libxfs/xfs_btree.c xfs_ilog_fbroot(cur->bc_private.b.whichfork)); b 1566 fs/xfs/libxfs/xfs_btree.c xfs_trans_log_inode(cur->bc_tp, cur->bc_private.b.ip, b 1567 fs/xfs/libxfs/xfs_btree.c xfs_ilog_fbroot(cur->bc_private.b.whichfork)); b 1804 fs/xfs/libxfs/xfs_btree.c !(cur->bc_private.b.flags & XFS_BTCUR_BPRV_INVALID_OWNER) && b 1807 fs/xfs/libxfs/xfs_btree.c cur->bc_private.b.ip->i_ino) b 2993 fs/xfs/libxfs/xfs_btree.c xfs_iroot_realloc(cur->bc_private.b.ip, b 2995 fs/xfs/libxfs/xfs_btree.c cur->bc_private.b.whichfork); b 3008 fs/xfs/libxfs/xfs_btree.c XFS_ILOG_CORE | xfs_ilog_fbroot(cur->bc_private.b.whichfork); b 3160 fs/xfs/libxfs/xfs_btree.c struct xfs_inode *ip = cur->bc_private.b.ip; b 3164 fs/xfs/libxfs/xfs_btree.c xfs_iroot_realloc(ip, 1, cur->bc_private.b.whichfork); b 3507 fs/xfs/libxfs/xfs_btree.c int whichfork = cur->bc_private.b.whichfork; b 3508 fs/xfs/libxfs/xfs_btree.c struct xfs_inode *ip = cur->bc_private.b.ip; b 3566 fs/xfs/libxfs/xfs_btree.c xfs_iroot_realloc(cur->bc_private.b.ip, index, b 3567 fs/xfs/libxfs/xfs_btree.c cur->bc_private.b.whichfork); b 3596 fs/xfs/libxfs/xfs_btree.c XFS_ILOG_CORE | xfs_ilog_fbroot(cur->bc_private.b.whichfork)); b 3764 fs/xfs/libxfs/xfs_btree.c xfs_iroot_realloc(cur->bc_private.b.ip, -1, b 3765 fs/xfs/libxfs/xfs_btree.c cur->bc_private.b.whichfork); b 4876 fs/xfs/libxfs/xfs_btree.c const union xfs_btree_ptr *b) b 4879 fs/xfs/libxfs/xfs_btree.c return (int64_t)be64_to_cpu(a->l) - be64_to_cpu(b->l); b 4880 fs/xfs/libxfs/xfs_btree.c return (int64_t)be32_to_cpu(a->s) - be32_to_cpu(b->s); b 174 fs/xfs/libxfs/xfs_btree.h struct xfs_bmbt_irec b; b 222 fs/xfs/libxfs/xfs_btree.h } b; b 447 fs/xfs/libxfs/xfs_btree.h #define XFS_EXTLEN_MIN(a,b) min_t(xfs_extlen_t, (a), (b)) b 448 fs/xfs/libxfs/xfs_btree.h #define XFS_EXTLEN_MAX(a,b) max_t(xfs_extlen_t, (a), (b)) b 449 fs/xfs/libxfs/xfs_btree.h #define XFS_AGBLOCK_MIN(a,b) min_t(xfs_agblock_t, (a), (b)) b 450 fs/xfs/libxfs/xfs_btree.h #define XFS_AGBLOCK_MAX(a,b) max_t(xfs_agblock_t, (a), (b)) b 451 fs/xfs/libxfs/xfs_btree.h #define XFS_FILEOFF_MIN(a,b) min_t(xfs_fileoff_t, (a), (b)) b 452 fs/xfs/libxfs/xfs_btree.h #define XFS_FILEOFF_MAX(a,b) max_t(xfs_fileoff_t, (a), (b)) b 453 fs/xfs/libxfs/xfs_btree.h #define XFS_FILBLKS_MIN(a,b) min_t(xfs_filblks_t, (a), (b)) b 454 fs/xfs/libxfs/xfs_btree.h #define XFS_FILBLKS_MAX(a,b) max_t(xfs_filblks_t, (a), (b)) b 505 fs/xfs/libxfs/xfs_btree.h const union xfs_btree_ptr *b); b 2094 fs/xfs/libxfs/xfs_da_btree.c xfs_fileoff_t b; b 2102 fs/xfs/libxfs/xfs_da_btree.c for (b = *bno, mapi = 0; b < *bno + count; ) { b 2104 fs/xfs/libxfs/xfs_da_btree.c c = (int)(*bno + count - b); b 2105 fs/xfs/libxfs/xfs_da_btree.c error = xfs_bmapi_write(tp, dp, b, c, b 2113 fs/xfs/libxfs/xfs_da_btree.c b = mapp[mapi - 1].br_startoff + b 32 fs/xfs/libxfs/xfs_dir2_block.c static int xfs_dir2_block_sort(const void *a, const void *b); b 879 fs/xfs/libxfs/xfs_dir2_block.c const void *b) /* second leaf entry */ b 885 fs/xfs/libxfs/xfs_dir2_block.c lb = b; b 588 fs/xfs/libxfs/xfs_format.h #define XFS_B_TO_FSB(mp,b) \ b 589 fs/xfs/libxfs/xfs_format.h ((((uint64_t)(b)) + (mp)->m_blockmask) >> (mp)->m_sb.sb_blocklog) b 590 fs/xfs/libxfs/xfs_format.h #define XFS_B_TO_FSBT(mp,b) (((uint64_t)(b)) >> (mp)->m_sb.sb_blocklog) b 591 fs/xfs/libxfs/xfs_format.h #define XFS_B_FSB_OFFSET(mp,b) ((b) & (mp)->m_blockmask) b 1094 fs/xfs/libxfs/xfs_format.h #define XFS_OFFBNO_TO_AGINO(mp,b,o) \ b 1095 fs/xfs/libxfs/xfs_format.h ((xfs_agino_t)(((b) << XFS_INO_OFFSET_BITS(mp)) | (o))) b 1096 fs/xfs/libxfs/xfs_format.h #define XFS_FSB_TO_INO(mp, b) ((xfs_ino_t)((b) << XFS_INO_OFFSET_BITS(mp))) b 1097 fs/xfs/libxfs/xfs_format.h #define XFS_AGB_TO_AGINO(mp, b) ((xfs_agino_t)((b) << XFS_INO_OFFSET_BITS(mp))) b 1131 fs/xfs/libxfs/xfs_format.h #define XFS_RTMIN(a,b) ((a) < (b) ? (a) : (b)) b 1132 fs/xfs/libxfs/xfs_format.h #define XFS_RTMAX(a,b) ((a) > (b) ? (a) : (b)) b 1137 fs/xfs/libxfs/xfs_format.h #define XFS_RTBLOCKLOG(b) xfs_highbit64(b) b 30 fs/xfs/libxfs/xfs_ialloc.h xfs_make_iptr(struct xfs_mount *mp, struct xfs_buf *b, int o) b 32 fs/xfs/libxfs/xfs_ialloc.h return xfs_buf_offset(b, o << (mp)->m_sb.sb_inodelog); b 37 fs/xfs/libxfs/xfs_log_format.h #define XLOG_BTOLSUNIT(log, b) (((b)+(log)->l_mp->m_sb.sb_logsunit-1) / \ b 2401 fs/xfs/libxfs/xfs_rmap.c const struct xfs_rmap_irec *b) b 2407 fs/xfs/libxfs/xfs_rmap.c ob = xfs_rmap_irec_offset_pack(b); b 2409 fs/xfs/libxfs/xfs_rmap.c if (a->rm_startblock < b->rm_startblock) b 2411 fs/xfs/libxfs/xfs_rmap.c else if (a->rm_startblock > b->rm_startblock) b 2413 fs/xfs/libxfs/xfs_rmap.c else if (a->rm_owner < b->rm_owner) b 2415 fs/xfs/libxfs/xfs_rmap.c else if (a->rm_owner > b->rm_owner) b 191 fs/xfs/libxfs/xfs_rmap.h const struct xfs_rmap_irec *b); b 379 fs/xfs/libxfs/xfs_rmap_btree.c uint64_t b; b 388 fs/xfs/libxfs/xfs_rmap_btree.c b = be64_to_cpu(k2->rmap.rm_owner); b 389 fs/xfs/libxfs/xfs_rmap_btree.c if (a < b) b 391 fs/xfs/libxfs/xfs_rmap_btree.c else if (a > b) b 394 fs/xfs/libxfs/xfs_rmap_btree.c b = XFS_RMAP_OFF(be64_to_cpu(k2->rmap.rm_offset)); b 395 fs/xfs/libxfs/xfs_rmap_btree.c if (a <= b) b 409 fs/xfs/libxfs/xfs_rmap_btree.c uint64_t b; b 418 fs/xfs/libxfs/xfs_rmap_btree.c b = be64_to_cpu(r2->rmap.rm_owner); b 419 fs/xfs/libxfs/xfs_rmap_btree.c if (a < b) b 421 fs/xfs/libxfs/xfs_rmap_btree.c else if (a > b) b 424 fs/xfs/libxfs/xfs_rmap_btree.c b = XFS_RMAP_OFF(be64_to_cpu(r2->rmap.rm_offset)); b 425 fs/xfs/libxfs/xfs_rmap_btree.c if (a <= b) b 101 fs/xfs/libxfs/xfs_rtbitmap.c xfs_rtword_t *b; /* current word in buffer */ b 128 fs/xfs/libxfs/xfs_rtbitmap.c b = &bufp[word]; b 135 fs/xfs/libxfs/xfs_rtbitmap.c want = (*b & ((xfs_rtword_t)1 << bit)) ? -1 : 0; b 152 fs/xfs/libxfs/xfs_rtbitmap.c if ((wdiff = (*b ^ want) & mask)) { b 177 fs/xfs/libxfs/xfs_rtbitmap.c b = &bufp[word]; b 182 fs/xfs/libxfs/xfs_rtbitmap.c b--; b 198 fs/xfs/libxfs/xfs_rtbitmap.c if ((wdiff = *b ^ want)) { b 223 fs/xfs/libxfs/xfs_rtbitmap.c b = &bufp[word]; b 228 fs/xfs/libxfs/xfs_rtbitmap.c b--; b 245 fs/xfs/libxfs/xfs_rtbitmap.c if ((wdiff = (*b ^ want) & mask)) { b 276 fs/xfs/libxfs/xfs_rtbitmap.c xfs_rtword_t *b; /* current word in buffer */ b 303 fs/xfs/libxfs/xfs_rtbitmap.c b = &bufp[word]; b 310 fs/xfs/libxfs/xfs_rtbitmap.c want = (*b & ((xfs_rtword_t)1 << bit)) ? -1 : 0; b 326 fs/xfs/libxfs/xfs_rtbitmap.c if ((wdiff = (*b ^ want) & mask)) { b 349 fs/xfs/libxfs/xfs_rtbitmap.c b = bufp = bp->b_addr; b 355 fs/xfs/libxfs/xfs_rtbitmap.c b++; b 371 fs/xfs/libxfs/xfs_rtbitmap.c if ((wdiff = *b ^ want)) { b 394 fs/xfs/libxfs/xfs_rtbitmap.c b = bufp = bp->b_addr; b 400 fs/xfs/libxfs/xfs_rtbitmap.c b++; b 415 fs/xfs/libxfs/xfs_rtbitmap.c if ((wdiff = (*b ^ want) & mask)) { b 539 fs/xfs/libxfs/xfs_rtbitmap.c xfs_rtword_t *b; /* current word in buffer */ b 567 fs/xfs/libxfs/xfs_rtbitmap.c first = b = &bufp[word]; b 587 fs/xfs/libxfs/xfs_rtbitmap.c *b |= mask; b 589 fs/xfs/libxfs/xfs_rtbitmap.c *b &= ~mask; b 602 fs/xfs/libxfs/xfs_rtbitmap.c (uint)((char *)b - (char *)bufp)); b 607 fs/xfs/libxfs/xfs_rtbitmap.c first = b = bufp = bp->b_addr; b 613 fs/xfs/libxfs/xfs_rtbitmap.c b++; b 629 fs/xfs/libxfs/xfs_rtbitmap.c *b = val; b 642 fs/xfs/libxfs/xfs_rtbitmap.c (uint)((char *)b - (char *)bufp)); b 647 fs/xfs/libxfs/xfs_rtbitmap.c first = b = bufp = bp->b_addr; b 653 fs/xfs/libxfs/xfs_rtbitmap.c b++; b 669 fs/xfs/libxfs/xfs_rtbitmap.c *b |= mask; b 671 fs/xfs/libxfs/xfs_rtbitmap.c *b &= ~mask; b 672 fs/xfs/libxfs/xfs_rtbitmap.c b++; b 677 fs/xfs/libxfs/xfs_rtbitmap.c if (b > first) b 679 fs/xfs/libxfs/xfs_rtbitmap.c (uint)((char *)b - (char *)bufp - 1)); b 773 fs/xfs/libxfs/xfs_rtbitmap.c xfs_rtword_t *b; /* current word in buffer */ b 801 fs/xfs/libxfs/xfs_rtbitmap.c b = &bufp[word]; b 823 fs/xfs/libxfs/xfs_rtbitmap.c if ((wdiff = (*b ^ val) & mask)) { b 847 fs/xfs/libxfs/xfs_rtbitmap.c b = bufp = bp->b_addr; b 853 fs/xfs/libxfs/xfs_rtbitmap.c b++; b 869 fs/xfs/libxfs/xfs_rtbitmap.c if ((wdiff = *b ^ val)) { b 893 fs/xfs/libxfs/xfs_rtbitmap.c b = bufp = bp->b_addr; b 899 fs/xfs/libxfs/xfs_rtbitmap.c b++; b 914 fs/xfs/libxfs/xfs_rtbitmap.c if ((wdiff = (*b ^ val) & mask)) { b 957 fs/xfs/libxfs/xfs_rtbitmap.c #define xfs_rtcheck_alloc_range(m,t,b,l) (0) b 21 fs/xfs/libxfs/xfs_trans_space.h #define XFS_NRMAPADD_SPACE_RES(mp, b)\ b 22 fs/xfs/libxfs/xfs_trans_space.h (((b + XFS_MAX_CONTIG_RMAPS_PER_BLOCK(mp) - 1) / \ b 29 fs/xfs/libxfs/xfs_trans_space.h #define XFS_NEXTENTADD_SPACE_RES(mp,b,w)\ b 30 fs/xfs/libxfs/xfs_trans_space.h (((b + XFS_MAX_CONTIG_EXTENTS_PER_BLOCK(mp) - 1) / \ b 35 fs/xfs/libxfs/xfs_trans_space.h #define XFS_SWAP_RMAP_SPACE_RES(mp,b,w)\ b 36 fs/xfs/libxfs/xfs_trans_space.h (XFS_NEXTENTADD_SPACE_RES((mp), (b), (w)) + \ b 37 fs/xfs/libxfs/xfs_trans_space.h XFS_NRMAPADD_SPACE_RES((mp), (b))) b 79 fs/xfs/libxfs/xfs_trans_space.h #define XFS_GROWFSRT_SPACE_RES(mp,b) \ b 80 fs/xfs/libxfs/xfs_trans_space.h ((b) + XFS_EXTENTADD_SPACE_RES(mp, XFS_DATA_FORK)) b 94 fs/xfs/libxfs/xfs_trans_space.h #define XFS_SYMLINK_SPACE_RES(mp,nl,b) \ b 95 fs/xfs/libxfs/xfs_trans_space.h (XFS_IALLOC_SPACE_RES(mp) + XFS_DIRENTER_SPACE_RES(mp,nl) + (b)) b 653 fs/xfs/scrub/agheader.c const xfs_agblock_t *b = pb; b 655 fs/xfs/scrub/agheader.c return (int)*a - (int)*b; b 67 fs/xfs/scrub/bitmap.c struct list_head *b) b 73 fs/xfs/scrub/bitmap.c bp = container_of(b, struct xfs_bitmap_range, list); b 25 fs/xfs/scrub/bitmap.h #define for_each_xfs_bitmap_block(b, bex, n, bitmap) \ b 27 fs/xfs/scrub/bitmap.h for ((b) = bex->start; (b) < bex->start + bex->len; (b)++) b 377 fs/xfs/scrub/bmap.c struct xfs_inode *ip = bs->cur->bc_private.b.ip; b 834 fs/xfs/scrub/repair.c xfs_agblock_t b; b 843 fs/xfs/scrub/repair.c for (b = 0; b < rec->rm_blockcount; b++) { b 849 fs/xfs/scrub/repair.c rec->rm_owner, rec->rm_startblock + b, b 27 fs/xfs/scrub/trace.c return XFS_INO_TO_FSB(cur->bc_mp, cur->bc_private.b.ip->i_ino); b 382 fs/xfs/scrub/trace.h __entry->whichfork = cur->bc_private.b.whichfork; b 462 fs/xfs/scrub/trace.h __entry->whichfork = cur->bc_private.b.whichfork; b 355 fs/xfs/xfs_aops.c struct list_head *b) b 361 fs/xfs/xfs_aops.c ib = container_of(b, struct xfs_ioend, io_list); b 26 fs/xfs/xfs_attr_list.c xfs_attr_shortform_compare(const void *a, const void *b) b 31 fs/xfs/xfs_attr_list.c sb = (xfs_attr_sf_sort_t *)b; b 271 fs/xfs/xfs_bmap_item.c struct list_head *b) b 277 fs/xfs/xfs_bmap_item.c bb = container_of(b, struct xfs_bmap_intent, bi_list); b 242 fs/xfs/xfs_bmap_util.c int b; b 245 fs/xfs/xfs_bmap_util.c for (b = 1; b <= numrecs; b++) { b 246 fs/xfs/xfs_bmap_util.c frp = XFS_BMBT_REC_ADDR(mp, block, b); b 1867 fs/xfs/xfs_buf.c struct list_head *b) b 1870 fs/xfs/xfs_buf.c struct xfs_buf *bp = container_of(b, struct xfs_buf, b_list); b 61 fs/xfs/xfs_extent_busy.h xfs_extent_busy_ag_cmp(void *priv, struct list_head *a, struct list_head *b); b 402 fs/xfs/xfs_extfree_item.c struct list_head *b) b 409 fs/xfs/xfs_extfree_item.c rb = container_of(b, struct xfs_extent_free_item, xefi_list); b 75 fs/xfs/xfs_inode.c xfs_extlen_t a, b; b 80 fs/xfs/xfs_inode.c b = xfs_get_extsz_hint(ip); b 82 fs/xfs/xfs_inode.c a = max(a, b); b 30 fs/xfs/xfs_itable.h char __user *b = breq->ubuffer; b 32 fs/xfs/xfs_itable.h breq->ubuffer = b + bytes; b 106 fs/xfs/xfs_log.c #define xlog_verify_dest_ptr(a,b) b 108 fs/xfs/xfs_log.c #define xlog_verify_iclog(a,b,c) b 109 fs/xfs/xfs_log.c #define xlog_verify_tail_lsn(a,b,c) b 775 fs/xfs/xfs_mount.c (get_unaligned_be16(&sbp->sb_uuid.b[8]) << 16) | b 776 fs/xfs/xfs_mount.c get_unaligned_be16(&sbp->sb_uuid.b[4]); b 777 fs/xfs/xfs_mount.c mp->m_fixedfsid[1] = get_unaligned_be32(&sbp->sb_uuid.b[0]); b 147 fs/xfs/xfs_quota.h #define xfs_qm_newmount(mp, a, b) (0) b 275 fs/xfs/xfs_refcount_item.c struct list_head *b) b 282 fs/xfs/xfs_refcount_item.c rb = container_of(b, struct xfs_refcount_intent, ri_list); b 341 fs/xfs/xfs_rmap_item.c struct list_head *b) b 348 fs/xfs/xfs_rmap_item.c rb = container_of(b, struct xfs_rmap_intent, ri_list); b 1278 fs/xfs/xfs_rtalloc.c xfs_rtblock_t b; /* result block */ b 1293 fs/xfs/xfs_rtalloc.c b = 0; b 1296 fs/xfs/xfs_rtalloc.c b = (mp->m_sb.sb_rextents * ((resid << 1) + 1ULL)) >> b 1298 fs/xfs/xfs_rtalloc.c if (b >= mp->m_sb.sb_rextents) b 1299 fs/xfs/xfs_rtalloc.c div64_u64_rem(b, mp->m_sb.sb_rextents, &b); b 1300 fs/xfs/xfs_rtalloc.c if (b + len > mp->m_sb.sb_rextents) b 1301 fs/xfs/xfs_rtalloc.c b = mp->m_sb.sb_rextents - len; b 1305 fs/xfs/xfs_rtalloc.c *pick = b; b 139 fs/xfs/xfs_rtalloc.h # define xfs_rtallocate_extent(t,b,min,max,l,f,p,rb) (ENOSYS) b 140 fs/xfs/xfs_rtalloc.h # define xfs_rtfree_extent(t,b,l) (ENOSYS) b 145 fs/xfs/xfs_rtalloc.h # define xfs_rtbuf_get(m,t,b,i,p) (ENOSYS) b 424 include/acpi/acoutput.h #define ACPI_DUMP_OPERANDS(a, b ,c) acpi_ex_dump_operands(a, b, c) b 425 include/acpi/acoutput.h #define ACPI_DUMP_ENTRY(a, b) acpi_ns_dump_entry (a, b) b 426 include/acpi/acoutput.h #define ACPI_DUMP_PATHNAME(a, b, c, d) acpi_ns_dump_pathname(a, b, c, d) b 427 include/acpi/acoutput.h #define ACPI_DUMP_BUFFER(a, b) acpi_ut_debug_dump_buffer((u8 *) a, b, DB_BYTE_DISPLAY, _COMPONENT) b 429 include/acpi/acoutput.h #define ACPI_TRACE_POINT(a, b, c, d) acpi_trace_point (a, b, c, d) b 442 include/acpi/acoutput.h #define ACPI_FUNCTION_TRACE_PTR(a, b) b 443 include/acpi/acoutput.h #define ACPI_FUNCTION_TRACE_U32(a, b) b 444 include/acpi/acoutput.h #define ACPI_FUNCTION_TRACE_STR(a, b) b 447 include/acpi/acoutput.h #define ACPI_DUMP_OPERANDS(a, b, c) b 448 include/acpi/acoutput.h #define ACPI_DUMP_ENTRY(a, b) b 449 include/acpi/acoutput.h #define ACPI_DUMP_PATHNAME(a, b, c, d) b 450 include/acpi/acoutput.h #define ACPI_DUMP_BUFFER(a, b) b 452 include/acpi/acoutput.h #define ACPI_TRACE_POINT(a, b, c, d) b 441 include/acpi/actypes.h #define ACPI_TIME_AFTER(a, b) ((s64)((b) - (a)) < 0) b 492 include/acpi/actypes.h #define ACPI_MIN(a,b) (((a)<(b))?(a):(b)) b 493 include/acpi/actypes.h #define ACPI_MAX(a,b) (((a)>(b))?(a):(b)) b 503 include/acpi/actypes.h #define ACPI_ADD_PTR(t, a, b) ACPI_CAST_PTR (t, (ACPI_CAST_PTR (u8, (a)) + (acpi_size)(b))) b 504 include/acpi/actypes.h #define ACPI_SUB_PTR(t, a, b) ACPI_CAST_PTR (t, (ACPI_CAST_PTR (u8, (a)) - (acpi_size)(b))) b 505 include/acpi/actypes.h #define ACPI_PTR_DIFF(a, b) ((acpi_size) (ACPI_CAST_PTR (u8, (a)) - ACPI_CAST_PTR (u8, (b)))) b 518 include/acpi/actypes.h #define ACPI_COMPARE_NAMESEG(a,b) (*ACPI_CAST_PTR (u32, (a)) == *ACPI_CAST_PTR (u32, (b))) b 521 include/acpi/actypes.h #define ACPI_COMPARE_NAMESEG(a,b) (!strncmp (ACPI_CAST_PTR (char, (a)), ACPI_CAST_PTR (char, (b)), ACPI_NAMESEG_SIZE)) b 79 include/acpi/platform/aclinux.h #define ACPI_INIT_GLOBAL(t,a,b) b 13 include/asm-generic/bitops/sched.h static inline int sched_find_first_bit(const unsigned long *b) b 16 include/asm-generic/bitops/sched.h if (b[0]) b 17 include/asm-generic/bitops/sched.h return __ffs(b[0]); b 18 include/asm-generic/bitops/sched.h return __ffs(b[1]) + 64; b 20 include/asm-generic/bitops/sched.h if (b[0]) b 21 include/asm-generic/bitops/sched.h return __ffs(b[0]); b 22 include/asm-generic/bitops/sched.h if (b[1]) b 23 include/asm-generic/bitops/sched.h return __ffs(b[1]) + 32; b 24 include/asm-generic/bitops/sched.h if (b[2]) b 25 include/asm-generic/bitops/sched.h return __ffs(b[2]) + 64; b 26 include/asm-generic/bitops/sched.h return __ffs(b[3]) + 96; b 90 include/asm-generic/uaccess.h #define segment_eq(a, b) ((a).seg == (b).seg) b 393 include/crypto/algapi.h noinline unsigned long __crypto_memneq(const void *a, const void *b, size_t size); b 405 include/crypto/algapi.h static inline int crypto_memneq(const void *a, const void *b, size_t size) b 407 include/crypto/algapi.h return __crypto_memneq(a, b, size) != 0UL ? 1 : 0; b 53 include/crypto/b128ops.h u64 a, b; b 57 include/crypto/b128ops.h __be64 a, b; b 61 include/crypto/b128ops.h __le64 b, a; b 67 include/crypto/b128ops.h r->b = p->b ^ q->b; b 163 include/crypto/gf128mul.h void gf128mul_lle(be128 *a, const be128 *b); b 165 include/crypto/gf128mul.h void gf128mul_bbe(be128 *a, const be128 *b); b 185 include/crypto/gf128mul.h u64 b = be64_to_cpu(x->b); b 189 include/crypto/gf128mul.h u64 _tt = gf128mul_mask_from_bit(b, 0) & ((u64)0xe1 << 56); b 191 include/crypto/gf128mul.h r->b = cpu_to_be64((b >> 1) | (a << 63)); b 198 include/crypto/gf128mul.h u64 b = be64_to_cpu(x->b); b 203 include/crypto/gf128mul.h r->a = cpu_to_be64((a << 1) | (b >> 63)); b 204 include/crypto/gf128mul.h r->b = cpu_to_be64((b << 1) ^ _tt); b 211 include/crypto/gf128mul.h u64 b = le64_to_cpu(x->b); b 216 include/crypto/gf128mul.h r->a = cpu_to_le64((a << 1) | (b >> 63)); b 217 include/crypto/gf128mul.h r->b = cpu_to_le64((b << 1) ^ _tt); b 481 include/drm/drm_crtc.h int (*gamma_set)(struct drm_crtc *crtc, u16 *r, u16 *g, u16 *b, b 108 include/drm/drm_fixed.h static inline s64 drm_fixp_mul(s64 a, s64 b) b 110 include/drm/drm_fixed.h unsigned shift = drm_fixp_msbset(a) + drm_fixp_msbset(b); b 116 include/drm/drm_fixed.h b >>= shift >> 1; b 120 include/drm/drm_fixed.h result = a * b; b 131 include/drm/drm_fixed.h static inline s64 drm_fixp_div(s64 a, s64 b) b 139 include/drm/drm_fixed.h b >>= (DRM_FIXED_POINT - shift); b 141 include/drm/drm_fixed.h result = div64_s64(a, b); b 149 include/drm/drm_fixed.h static inline s64 drm_fixp_from_fraction(s64 a, s64 b) b 153 include/drm/drm_fixed.h bool b_neg = b < 0; b 155 include/drm/drm_fixed.h u64 b_abs = b_neg ? -b : b; b 15 include/dt-bindings/pinctrl/r7s72100-pinctrl.h #define RZA1_PINMUX(b, p, f) ((b) * RZA1_PINS_PER_PORT + (p) | (f << 16)) b 40 include/dt-bindings/pinctrl/r7s9210-pinctrl.h #define RZA2_PINMUX(b, p, f) ((b) * RZA2_PINS_PER_PORT + (p) | (f << 16)) b 735 include/linux/blk-cgroup.h char b[BDEVNAME_SIZE]; b 739 include/linux/blk-cgroup.h bio_devname(bio, b)); b 763 include/linux/blkdev.h static inline bool blk_write_same_mergeable(struct bio *a, struct bio *b) b 765 include/linux/blkdev.h if (bio_page(a) == bio_page(b) && b 766 include/linux/blkdev.h bio_offset(a) == bio_offset(b)) b 1094 include/linux/blkdev.h extern int blk_stack_limits(struct queue_limits *t, struct queue_limits *b, b 1100 include/linux/blkdev.h extern void blk_queue_stack_limits(struct request_queue *t, struct request_queue *b); b 1627 include/linux/blkdev.h struct bio *b) b 1632 include/linux/blkdev.h struct bio *b, b 1637 include/linux/blkdev.h static inline struct blk_integrity *bdev_get_integrity(struct block_device *b) b 1645 include/linux/blkdev.h static inline int blk_integrity_compare(struct gendisk *a, struct gendisk *b) b 1650 include/linux/blkdev.h struct blk_integrity *b) b 1672 include/linux/blkdev.h struct bio *b) b 66 include/linux/can/dev/peak_canfd.h #define PUCAN_TSLOW_BRP(b) ((b) & PUCAN_TSLOW_BRP_MASK) b 93 include/linux/can/dev/peak_canfd.h #define PUCAN_TFAST_BRP(b) ((b) & PUCAN_TFAST_BRP_MASK) b 102 include/linux/capability.h #define CAP_BOP_ALL(c, a, b, OP) \ b 106 include/linux/capability.h c.cap[__capi] = a.cap[__capi] OP b.cap[__capi]; \ b 119 include/linux/capability.h const kernel_cap_t b) b 122 include/linux/capability.h CAP_BOP_ALL(dest, a, b, |); b 127 include/linux/capability.h const kernel_cap_t b) b 130 include/linux/capability.h CAP_BOP_ALL(dest, a, b, &); b 25 include/linux/ceph/buffer.h static inline struct ceph_buffer *ceph_buffer_get(struct ceph_buffer *b) b 27 include/linux/ceph/buffer.h kref_get(&b->kref); b 28 include/linux/ceph/buffer.h return b; b 31 include/linux/ceph/buffer.h static inline void ceph_buffer_put(struct ceph_buffer *b) b 33 include/linux/ceph/buffer.h if (b) b 34 include/linux/ceph/buffer.h kref_put(&b->kref, ceph_buffer_release); b 37 include/linux/ceph/buffer.h extern int ceph_decode_buffer(struct ceph_buffer **b, void **p, void *end); b 22 include/linux/ceph/ceph_frag.h static inline __u32 ceph_frag_make(__u32 b, __u32 v) b 24 include/linux/ceph/ceph_frag.h return (b << 24) | b 25 include/linux/ceph/ceph_frag.h (v & (0xffffffu << (24-b)) & 0xffffffu); b 73 include/linux/ceph/ceph_frag.h int ceph_frag_compare(__u32 a, __u32 b); b 188 include/linux/ceph/libceph.h #define RB_CMP3WAY(a, b) ((a) < (b) ? -1 : (a) > (b)) b 35 include/linux/ceph/msgr.h static inline __s32 ceph_seq_cmp(__u32 a, __u32 b) b 37 include/linux/ceph/msgr.h return (__s32)a - (__s32)b; b 20 include/linux/ceph/rados.h const struct ceph_fsid *b) b 22 include/linux/ceph/rados.h return memcmp(a, b, sizeof(*a)); b 98 include/linux/ceph/rados.h static inline int ceph_stable_mod(int x, int b, int bmask) b 100 include/linux/ceph/rados.h if ((x & bmask) < b) b 158 include/linux/compiler_attributes.h #define __printf(a, b) __attribute__((__format__(printf, a, b))) b 159 include/linux/compiler_attributes.h #define __scanf(a, b) __attribute__((__format__(scanf, a, b))) b 53 include/linux/compiler_types.h #define ___PASTE(a,b) a##b b 54 include/linux/compiler_types.h #define __PASTE(a,b) ___PASTE(a,b) b 129 include/linux/compiler_types.h #define __compiler_offsetof(a, b) __builtin_offsetof(a, b) b 220 include/linux/compiler_types.h #define __same_type(a, b) __builtin_types_compatible_p(typeof(a), typeof(b)) b 311 include/linux/crush/crush.h extern int crush_get_bucket_item_weight(const struct crush_bucket *b, int pos); b 312 include/linux/crush/crush.h extern void crush_destroy_bucket_uniform(struct crush_bucket_uniform *b); b 313 include/linux/crush/crush.h extern void crush_destroy_bucket_list(struct crush_bucket_list *b); b 314 include/linux/crush/crush.h extern void crush_destroy_bucket_tree(struct crush_bucket_tree *b); b 315 include/linux/crush/crush.h extern void crush_destroy_bucket_straw(struct crush_bucket_straw *b); b 316 include/linux/crush/crush.h extern void crush_destroy_bucket_straw2(struct crush_bucket_straw2 *b); b 317 include/linux/crush/crush.h extern void crush_destroy_bucket(struct crush_bucket *b); b 18 include/linux/crush/hash.h extern __u32 crush_hash32_2(int type, __u32 a, __u32 b); b 19 include/linux/crush/hash.h extern __u32 crush_hash32_3(int type, __u32 a, __u32 b, __u32 c); b 20 include/linux/crush/hash.h extern __u32 crush_hash32_4(int type, __u32 a, __u32 b, __u32 c, __u32 d); b 21 include/linux/crush/hash.h extern __u32 crush_hash32_5(int type, __u32 a, __u32 b, __u32 c, __u32 d, b 269 include/linux/device.h const struct device *b)); b 84 include/linux/dm-bufio.h void dm_bufio_release(struct dm_buffer *b); b 94 include/linux/dm-bufio.h void dm_bufio_mark_buffer_dirty(struct dm_buffer *b); b 102 include/linux/dm-bufio.h void dm_bufio_mark_partial_buffer_dirty(struct dm_buffer *b, b 125 include/linux/dm-bufio.h void dm_bufio_release_move(struct dm_buffer *b, sector_t new_block); b 141 include/linux/dm-bufio.h sector_t dm_bufio_get_block_number(struct dm_buffer *b); b 142 include/linux/dm-bufio.h void *dm_bufio_get_block_data(struct dm_buffer *b); b 143 include/linux/dm-bufio.h void *dm_bufio_get_aux_data(struct dm_buffer *b); b 144 include/linux/dm-bufio.h struct dm_bufio_client *dm_bufio_get_client(struct dm_buffer *b); b 66 include/linux/efi.h #define EFI_GUID(a,b,c,d0,d1,d2,d3,d4,d5,d6,d7) \ b 67 include/linux/efi.h GUID_INIT(a, b, c, d0, d1, d2, d3, d4, d5, d6, d7) b 1033 include/linux/efi.h sprintf(out, "%pUl", guid->b); b 71 include/linux/elfnote.h #define _ELFNOTE_PASTE(a,b) a##b b 79 include/linux/etherdevice.h static const __be16 *b = (const __be16 *)eth_reserved_addr_base; b 83 include/linux/etherdevice.h return (((*(const u32 *)addr) ^ (*(const u32 *)b)) | b 84 include/linux/etherdevice.h (__force int)((a[2] ^ b[2]) & m)) == 0; b 86 include/linux/etherdevice.h return ((a[0] ^ b[0]) | (a[1] ^ b[1]) | ((a[2] ^ b[2]) & m)) == 0; b 283 include/linux/etherdevice.h const u16 *b = (const u16 *)src; b 285 include/linux/etherdevice.h a[0] = b[0]; b 286 include/linux/etherdevice.h a[1] = b[1]; b 287 include/linux/etherdevice.h a[2] = b[2]; b 324 include/linux/etherdevice.h const u16 *b = (const u16 *)addr2; b 326 include/linux/etherdevice.h return ((a[0] ^ b[0]) | (a[1] ^ b[1]) | (a[2] ^ b[2])) == 0; b 498 include/linux/etherdevice.h static inline unsigned long compare_ether_header(const void *a, const void *b) b 511 include/linux/etherdevice.h fold = *(unsigned long *)a ^ *(unsigned long *)b; b 512 include/linux/etherdevice.h fold |= *(unsigned long *)(a + 6) ^ *(unsigned long *)(b + 6); b 516 include/linux/etherdevice.h u32 *b32 = (u32 *)((u8 *)b + 2); b 518 include/linux/etherdevice.h return (*(u16 *)a ^ *(u16 *)b) | (a32[0] ^ b32[0]) | b 569 include/linux/fb.h #define fb_writeb(b,addr) (*(volatile u8 *) (addr) = (b)) b 570 include/linux/fb.h #define fb_writew(b,addr) (*(volatile u16 *) (addr) = (b)) b 571 include/linux/fb.h #define fb_writel(b,addr) (*(volatile u32 *) (addr) = (b)) b 572 include/linux/fb.h #define fb_writeq(b,addr) (*(volatile u64 *) (addr) = (b)) b 7 include/linux/gcd.h unsigned long gcd(unsigned long a, unsigned long b) __attribute_const__; b 95 include/linux/genl_magic_func.h #define DPRINT_TLA(a, op, b) pr_info("%s %s %s\n", a, op, b); b 121 include/linux/genl_magic_func.h #define DPRINT_TLA(a, op, b) do {} while (0) b 20 include/linux/genl_magic_struct.h #define CONCAT__(a,b) a ## b b 21 include/linux/genl_magic_struct.h #define CONCAT_(a,b) CONCAT__(a,b) b 40 include/linux/hid-debug.h #define hid_dump_input(a,b,c) do { } while (0) b 41 include/linux/hid-debug.h #define hid_dump_report(a,b,c,d) do { } while (0) b 42 include/linux/hid-debug.h #define hid_dump_device(a,b) do { } while (0) b 43 include/linux/hid-debug.h #define hid_dump_field(a,b,c) do { } while (0) b 44 include/linux/hid-debug.h #define hid_resolv_usage(a,b) do { } while (0) b 45 include/linux/hid-debug.h #define hid_debug_register(a, b) do { } while (0) b 49 include/linux/hid-debug.h #define hid_debug_event(a,b) do { } while (0) b 670 include/linux/hid.h #define HID_DEVICE(b, g, ven, prod) \ b 671 include/linux/hid.h .bus = (b), .group = (g), .vendor = (ven), .product = (prod) b 167 include/linux/hugetlb.h #define follow_hugetlb_page(m,v,p,vs,a,b,i,w,n) ({ BUG(); 0; }) b 179 include/linux/inetdevice.h int inet_addr_onlink(struct in_device *in_dev, __be32 a, __be32 b); b 17 include/linux/isapnp.h #define ISAPNP_VENDOR(a,b,c) (((((a)-'A'+1)&0x3f)<<2)|\ b 18 include/linux/isapnp.h ((((b)-'A'+1)&0x18)>>3)|((((b)-'A'+1)&7)<<13)|\ b 35 include/linux/jhash.h #define __jhash_mix(a, b, c) \ b 37 include/linux/jhash.h a -= c; a ^= rol32(c, 4); c += b; \ b 38 include/linux/jhash.h b -= a; b ^= rol32(a, 6); a += c; \ b 39 include/linux/jhash.h c -= b; c ^= rol32(b, 8); b += a; \ b 40 include/linux/jhash.h a -= c; a ^= rol32(c, 16); c += b; \ b 41 include/linux/jhash.h b -= a; b ^= rol32(a, 19); a += c; \ b 42 include/linux/jhash.h c -= b; c ^= rol32(b, 4); b += a; \ b 46 include/linux/jhash.h #define __jhash_final(a, b, c) \ b 48 include/linux/jhash.h c ^= b; c -= rol32(b, 14); \ b 50 include/linux/jhash.h b ^= a; b -= rol32(a, 25); \ b 51 include/linux/jhash.h c ^= b; c -= rol32(b, 16); \ b 53 include/linux/jhash.h b ^= a; b -= rol32(a, 14); \ b 54 include/linux/jhash.h c ^= b; c -= rol32(b, 24); \ b 72 include/linux/jhash.h u32 a, b, c; b 76 include/linux/jhash.h a = b = c = JHASH_INITVAL + length + initval; b 81 include/linux/jhash.h b += __get_unaligned_cpu32(k + 4); b 83 include/linux/jhash.h __jhash_mix(a, b, c); b 93 include/linux/jhash.h case 8: b += (u32)k[7]<<24; /* fall through */ b 94 include/linux/jhash.h case 7: b += (u32)k[6]<<16; /* fall through */ b 95 include/linux/jhash.h case 6: b += (u32)k[5]<<8; /* fall through */ b 96 include/linux/jhash.h case 5: b += k[4]; /* fall through */ b 101 include/linux/jhash.h __jhash_final(a, b, c); b 118 include/linux/jhash.h u32 a, b, c; b 121 include/linux/jhash.h a = b = c = JHASH_INITVAL + (length<<2) + initval; b 126 include/linux/jhash.h b += k[1]; b 128 include/linux/jhash.h __jhash_mix(a, b, c); b 136 include/linux/jhash.h case 2: b += k[1]; /* fall through */ b 138 include/linux/jhash.h __jhash_final(a, b, c); b 148 include/linux/jhash.h static inline u32 __jhash_nwords(u32 a, u32 b, u32 c, u32 initval) b 151 include/linux/jhash.h b += initval; b 154 include/linux/jhash.h __jhash_final(a, b, c); b 159 include/linux/jhash.h static inline u32 jhash_3words(u32 a, u32 b, u32 c, u32 initval) b 161 include/linux/jhash.h return __jhash_nwords(a, b, c, initval + JHASH_INITVAL + (3 << 2)); b 164 include/linux/jhash.h static inline u32 jhash_2words(u32 a, u32 b, u32 initval) b 166 include/linux/jhash.h return __jhash_nwords(a, b, 0, initval + JHASH_INITVAL + (2 << 2)); b 105 include/linux/jiffies.h #define time_after(a,b) \ b 107 include/linux/jiffies.h typecheck(unsigned long, b) && \ b 108 include/linux/jiffies.h ((long)((b) - (a)) < 0)) b 109 include/linux/jiffies.h #define time_before(a,b) time_after(b,a) b 111 include/linux/jiffies.h #define time_after_eq(a,b) \ b 113 include/linux/jiffies.h typecheck(unsigned long, b) && \ b 114 include/linux/jiffies.h ((long)((a) - (b)) >= 0)) b 115 include/linux/jiffies.h #define time_before_eq(a,b) time_after_eq(b,a) b 120 include/linux/jiffies.h #define time_in_range(a,b,c) \ b 121 include/linux/jiffies.h (time_after_eq(a,b) && \ b 127 include/linux/jiffies.h #define time_in_range_open(a,b,c) \ b 128 include/linux/jiffies.h (time_after_eq(a,b) && \ b 134 include/linux/jiffies.h #define time_after64(a,b) \ b 136 include/linux/jiffies.h typecheck(__u64, b) && \ b 137 include/linux/jiffies.h ((__s64)((b) - (a)) < 0)) b 138 include/linux/jiffies.h #define time_before64(a,b) time_after64(b,a) b 140 include/linux/jiffies.h #define time_after_eq64(a,b) \ b 142 include/linux/jiffies.h typecheck(__u64, b) && \ b 143 include/linux/jiffies.h ((__s64)((a) - (b)) >= 0)) b 144 include/linux/jiffies.h #define time_before_eq64(a,b) time_after_eq64(b,a) b 146 include/linux/jiffies.h #define time_in_range64(a, b, c) \ b 147 include/linux/jiffies.h (time_after_eq64(a, b) && \ b 182 include/linux/kernel.h #define sector_div(a, b) do_div(a, b) b 975 include/linux/kernel.h #define swap(a, b) \ b 976 include/linux/kernel.h do { typeof(a) __tmp = (a); (a) = (b); (b) = __tmp; } while (0) b 982 include/linux/kernel.h #define __CONCAT(a, b) a ## b b 983 include/linux/kernel.h #define CONCATENATE(a, b) __CONCAT(a, b) b 7 include/linux/lcm.h unsigned long lcm(unsigned long a, unsigned long b) __attribute_const__; b 8 include/linux/lcm.h unsigned long lcm_not_zero(unsigned long a, unsigned long b) __attribute_const__; b 102 include/linux/leds-lp3952.h u16 b:3; b 19 include/linux/linkmode.h const unsigned long *b) b 21 include/linux/linkmode.h bitmap_and(dst, a, b, __ETHTOOL_LINK_MODE_MASK_NBITS); b 25 include/linux/linkmode.h const unsigned long *b) b 27 include/linux/linkmode.h bitmap_or(dst, a, b, __ETHTOOL_LINK_MODE_MASK_NBITS); b 146 include/linux/list_bl.h static inline void hlist_bl_lock(struct hlist_bl_head *b) b 148 include/linux/list_bl.h bit_spin_lock(0, (unsigned long *)b); b 151 include/linux/list_bl.h static inline void hlist_bl_unlock(struct hlist_bl_head *b) b 153 include/linux/list_bl.h __bit_spin_unlock(0, (unsigned long *)b); b 156 include/linux/list_bl.h static inline bool hlist_bl_is_locked(struct hlist_bl_head *b) b 158 include/linux/list_bl.h return bit_spin_is_locked(0, (unsigned long *)b); b 12 include/linux/list_sort.h struct list_head *b)); b 168 include/linux/math64.h static inline u64 mul_u32_u32(u32 a, u32 b) b 170 include/linux/math64.h return (u64)a * b; b 210 include/linux/math64.h static inline u64 mul_u64_u64_shr(u64 a, u64 b, unsigned int shift) b 225 include/linux/math64.h b0.ll = b; b 103 include/linux/mempolicy.h extern bool __mpol_equal(struct mempolicy *a, struct mempolicy *b); b 104 include/linux/mempolicy.h static inline bool mpol_equal(struct mempolicy *a, struct mempolicy *b) b 106 include/linux/mempolicy.h if (a == b) b 108 include/linux/mempolicy.h return __mpol_equal(a, b); b 212 include/linux/mempolicy.h static inline bool mpol_equal(struct mempolicy *a, struct mempolicy *b) b 45 include/linux/mfd/palmas.h #define PALMAS_PMIC_HAS(b, f) \ b 46 include/linux/mfd/palmas.h ((b)->features & PALMAS_PMIC_FEATURE_ ## f) b 15 include/linux/mfd/tmio.h #define tmio_ioread16_rep(r, b, l) readsw(r, b, l) b 21 include/linux/mfd/tmio.h #define tmio_iowrite16_rep(r, b, l) writesw(r, b, l) b 9579 include/linux/mlx5/mlx5_ifc.h u8 b[0x1]; b 147 include/linux/mmc/sdio_func.h extern void sdio_writeb(struct sdio_func *func, u8 b, b 149 include/linux/mmc/sdio_func.h extern void sdio_writew(struct sdio_func *func, u16 b, b 151 include/linux/mmc/sdio_func.h extern void sdio_writel(struct sdio_func *func, u32 b, b 164 include/linux/mmc/sdio_func.h extern void sdio_f0_writeb(struct sdio_func *func, unsigned char b, b 576 include/linux/mod_devicetable.h #define DMI_MATCH(a, b) { .slot = a, .substr = b } b 577 include/linux/mod_devicetable.h #define DMI_EXACT_MATCH(a, b) { .slot = a, .substr = b, .exact_match = 1 } b 639 include/linux/module.h int ref_module(struct module *a, struct module *b); b 63 include/linux/mtd/ftl.h #define BLOCK_FREE(b) ((b) == 0xffffffff) b 64 include/linux/mtd/ftl.h #define BLOCK_DELETED(b) (((b) == 0) || ((b) == 0xfffffffe)) b 66 include/linux/mtd/ftl.h #define BLOCK_TYPE(b) ((b) & 0x7f) b 67 include/linux/mtd/ftl.h #define BLOCK_ADDRESS(b) ((b) & ~0x7f) b 68 include/linux/mtd/ftl.h #define BLOCK_NUMBER(b) ((b) >> 9) b 481 include/linux/mtd/nand.h const struct nand_pos *b) b 483 include/linux/mtd/nand.h if (a->target != b->target) b 484 include/linux/mtd/nand.h return a->target < b->target ? -1 : 1; b 486 include/linux/mtd/nand.h if (a->lun != b->lun) b 487 include/linux/mtd/nand.h return a->lun < b->lun ? -1 : 1; b 489 include/linux/mtd/nand.h if (a->eraseblock != b->eraseblock) b 490 include/linux/mtd/nand.h return a->eraseblock < b->eraseblock ? -1 : 1; b 492 include/linux/mtd/nand.h if (a->page != b->page) b 493 include/linux/mtd/nand.h return a->page < b->page ? -1 : 1; b 644 include/linux/mtd/rawnand.h #define NAND_OP_DATA_IN(l, b, ns) \ b 649 include/linux/mtd/rawnand.h .buf.in = b, \ b 655 include/linux/mtd/rawnand.h #define NAND_OP_DATA_OUT(l, b, ns) \ b 660 include/linux/mtd/rawnand.h .buf.out = b, \ b 666 include/linux/mtd/rawnand.h #define NAND_OP_8BIT_DATA_IN(l, b, ns) \ b 671 include/linux/mtd/rawnand.h .buf.in = b, \ b 677 include/linux/mtd/rawnand.h #define NAND_OP_8BIT_DATA_OUT(l, b, ns) \ b 682 include/linux/mtd/rawnand.h .buf.out = b, \ b 828 include/linux/netdevice.h struct netdev_phys_item_id *b) b 830 include/linux/netdevice.h return a->id_len == b->id_len && b 831 include/linux/netdevice.h memcmp(a->id, b->id, a->id_len) == 0; b 3689 include/linux/netdevice.h bool netdev_port_same_parent_id(struct net_device *a, struct net_device *b); b 20 include/linux/netfilter/ipset/ip_set.h #define _IP_SET_MODULE_DESC(a, b, c) \ b 21 include/linux/netfilter/ipset/ip_set.h MODULE_DESCRIPTION(a " type of IP sets, revisions " b "-" c) b 22 include/linux/netfilter/ipset/ip_set.h #define IP_SET_MODULE_DESC(a, b, c) \ b 23 include/linux/netfilter/ipset/ip_set.h _IP_SET_MODULE_DESC(a, __stringify(b), __stringify(c)) b 190 include/linux/netfilter/ipset/ip_set.h bool (*same_set)(const struct ip_set *a, const struct ip_set *b); b 700 include/linux/netfilter/ipset/ip_set.h #define IPSET_CONCAT(a, b) a##b b 701 include/linux/netfilter/ipset/ip_set.h #define IPSET_TOKEN(a, b) IPSET_CONCAT(a, b) b 407 include/linux/netfilter/x_tables.h const unsigned long *b = (const unsigned long *)_b; b 411 include/linux/netfilter/x_tables.h ret = (a[0] ^ b[0]) & mask[0]; b 413 include/linux/netfilter/x_tables.h ret |= (a[1] ^ b[1]) & mask[1]; b 415 include/linux/netfilter/x_tables.h ret |= (a[2] ^ b[2]) & mask[2]; b 417 include/linux/netfilter/x_tables.h ret |= (a[3] ^ b[3]) & mask[3]; b 28 include/linux/nfs.h static inline int nfs_compare_fh(const struct nfs_fh *a, const struct nfs_fh *b) b 30 include/linux/nfs.h return a->size != b->size || memcmp(a->data, b->data, a->size) != 0; b 31 include/linux/nfs_xdr.h static inline int nfs_fsid_equal(const struct nfs_fsid *a, const struct nfs_fsid *b) b 33 include/linux/nfs_xdr.h return a->major == b->major && a->minor == b->minor; b 129 include/linux/nubus.h #define for_each_board_func_rsrc(b, f) \ b 130 include/linux/nubus.h for_each_func_rsrc(f) if (f->board != b) {} else b 55 include/linux/overflow.h #define check_add_overflow(a, b, d) ({ \ b 57 include/linux/overflow.h typeof(b) __b = (b); \ b 64 include/linux/overflow.h #define check_sub_overflow(a, b, d) ({ \ b 66 include/linux/overflow.h typeof(b) __b = (b); \ b 73 include/linux/overflow.h #define check_mul_overflow(a, b, d) ({ \ b 75 include/linux/overflow.h typeof(b) __b = (b); \ b 86 include/linux/overflow.h #define __unsigned_add_overflow(a, b, d) ({ \ b 88 include/linux/overflow.h typeof(b) __b = (b); \ b 95 include/linux/overflow.h #define __unsigned_sub_overflow(a, b, d) ({ \ b 97 include/linux/overflow.h typeof(b) __b = (b); \ b 107 include/linux/overflow.h #define __unsigned_mul_overflow(a, b, d) ({ \ b 109 include/linux/overflow.h typeof(b) __b = (b); \ b 135 include/linux/overflow.h #define __signed_add_overflow(a, b, d) ({ \ b 137 include/linux/overflow.h typeof(b) __b = (b); \ b 151 include/linux/overflow.h #define __signed_sub_overflow(a, b, d) ({ \ b 153 include/linux/overflow.h typeof(b) __b = (b); \ b 178 include/linux/overflow.h #define __signed_mul_overflow(a, b, d) ({ \ b 180 include/linux/overflow.h typeof(b) __b = (b); \ b 193 include/linux/overflow.h #define check_add_overflow(a, b, d) \ b 195 include/linux/overflow.h __signed_add_overflow(a, b, d), \ b 196 include/linux/overflow.h __unsigned_add_overflow(a, b, d)) b 198 include/linux/overflow.h #define check_sub_overflow(a, b, d) \ b 200 include/linux/overflow.h __signed_sub_overflow(a, b, d), \ b 201 include/linux/overflow.h __unsigned_sub_overflow(a, b, d)) b 203 include/linux/overflow.h #define check_mul_overflow(a, b, d) \ b 205 include/linux/overflow.h __signed_mul_overflow(a, b, d), \ b 206 include/linux/overflow.h __unsigned_mul_overflow(a, b, d)) b 253 include/linux/overflow.h static inline __must_check size_t array_size(size_t a, size_t b) b 257 include/linux/overflow.h if (check_mul_overflow(a, b, &bytes)) b 275 include/linux/overflow.h static inline __must_check size_t array3_size(size_t a, size_t b, size_t c) b 279 include/linux/overflow.h if (check_mul_overflow(a, b, &bytes)) b 291 include/linux/overflow.h static inline __must_check size_t __ab_c_size(size_t a, size_t b, size_t c) b 295 include/linux/overflow.h if (check_mul_overflow(a, b, &bytes)) b 991 include/linux/pci.h int pci_bus_insert_busn_res(struct pci_bus *b, int bus, int busmax); b 992 include/linux/pci.h int pci_bus_update_busn_res_end(struct pci_bus *b, int busmax); b 993 include/linux/pci.h void pci_bus_release_busn_res(struct pci_bus *b); b 1023 include/linux/pci.h void pci_remove_bus(struct pci_bus *b); b 167 include/linux/pinctrl/pinconf-generic.h #define PCONFDUMP(a, b, c, d) { \ b 168 include/linux/pinctrl/pinconf-generic.h .param = a, .display = b, .format = c, .has_arg = d \ b 43 include/linux/pinctrl/pinctrl.h #define PINCTRL_PIN(a, b) { .number = a, .name = b } b 1777 include/linux/platform_data/cros_ec_commands.h uint8_t r, g, b; b 31 include/linux/platform_data/leds-lp55xx.h const u8 *b; b 774 include/linux/pm.h static inline int device_pm_wait_for_dev(struct device *a, struct device *b) b 145 include/linux/posix-timers.h #define INIT_CPU_TIMERBASE(b) { \ b 149 include/linux/posix-timers.h #define INIT_CPU_TIMERBASES(b) { \ b 150 include/linux/posix-timers.h INIT_CPU_TIMERBASE(b[0]), \ b 151 include/linux/posix-timers.h INIT_CPU_TIMERBASE(b[1]), \ b 152 include/linux/posix-timers.h INIT_CPU_TIMERBASE(b[2]), \ b 65 include/linux/rbtree_latch.h bool (*less)(struct latch_tree_node *a, struct latch_tree_node *b); b 66 include/linux/rbtree_latch.h int (*comp)(void *key, struct latch_tree_node *b); b 77 include/linux/rbtree_latch.h bool (*less)(struct latch_tree_node *a, struct latch_tree_node *b)) b 33 include/linux/rcupdate.h #define ULONG_CMP_GE(a, b) (ULONG_MAX / 2 >= (a) - (b)) b 34 include/linux/rcupdate.h #define ULONG_CMP_LT(a, b) (ULONG_MAX / 2 < (a) - (b)) b 609 include/linux/sched.h } b; /* Bits. */ b 23 include/linux/sched/deadline.h static inline bool dl_time_before(u64 a, u64 b) b 25 include/linux/sched/deadline.h return (s64)(a - b) < 0; b 122 include/linux/signal.h static inline void name(sigset_t *r, const sigset_t *a, const sigset_t *b) \ b 129 include/linux/signal.h b3 = b->sig[3]; b2 = b->sig[2]; \ b 134 include/linux/signal.h a1 = a->sig[1]; b1 = b->sig[1]; \ b 138 include/linux/signal.h a0 = a->sig[0]; b0 = b->sig[0]; \ b 35 include/linux/siphash.h u64 siphash_2u64(const u64 a, const u64 b, const siphash_key_t *key); b 36 include/linux/siphash.h u64 siphash_3u64(const u64 a, const u64 b, const u64 c, b 38 include/linux/siphash.h u64 siphash_4u64(const u64 a, const u64 b, const u64 c, const u64 d, b 41 include/linux/siphash.h u64 siphash_3u32(const u32 a, const u32 b, const u32 c, b 44 include/linux/siphash.h static inline u64 siphash_2u32(const u32 a, const u32 b, b 47 include/linux/siphash.h return siphash_1u64((u64)b << 32 | a, key); b 49 include/linux/siphash.h static inline u64 siphash_4u32(const u32 a, const u32 b, const u32 c, b 52 include/linux/siphash.h return siphash_2u64((u64)b << 32 | a, (u64)d << 32 | c, key); b 105 include/linux/siphash.h u32 hsiphash_2u32(const u32 a, const u32 b, const hsiphash_key_t *key); b 106 include/linux/siphash.h u32 hsiphash_3u32(const u32 a, const u32 b, const u32 c, b 108 include/linux/siphash.h u32 hsiphash_4u32(const u32 a, const u32 b, const u32 c, const u32 d, b 3708 include/linux/skbuff.h const void *b = skb_metadata_end(skb_b); b 3715 include/linux/skbuff.h #define __it_diff(a, b, op) (*(u##op *)__it(a, op)) ^ (*(u##op *)__it(b, op)) b 3716 include/linux/skbuff.h case 32: diffs |= __it_diff(a, b, 64); b 3718 include/linux/skbuff.h case 24: diffs |= __it_diff(a, b, 64); b 3720 include/linux/skbuff.h case 16: diffs |= __it_diff(a, b, 64); b 3722 include/linux/skbuff.h case 8: diffs |= __it_diff(a, b, 64); b 3724 include/linux/skbuff.h case 28: diffs |= __it_diff(a, b, 64); b 3726 include/linux/skbuff.h case 20: diffs |= __it_diff(a, b, 64); b 3728 include/linux/skbuff.h case 12: diffs |= __it_diff(a, b, 64); b 3730 include/linux/skbuff.h case 4: diffs |= __it_diff(a, b, 32); b 3735 include/linux/skbuff.h return memcmp(a - meta_len, b - meta_len, meta_len); b 138 include/linux/string.h extern void **__memcat_p(void **a, void **b); b 139 include/linux/string.h #define memcat_p(a, b) ({ \ b 140 include/linux/string.h BUILD_BUG_ON_MSG(!__same_type(*(a), *(b)), \ b 142 include/linux/string.h (typeof(*a) *)__memcat_p((void **)(a), (void **)(b)); \ b 431 include/linux/suspend.h extern void __register_nosave_region(unsigned long b, unsigned long e, int km); b 432 include/linux/suspend.h static inline void __init register_nosave_region(unsigned long b, unsigned long e) b 434 include/linux/suspend.h __register_nosave_region(b, e, 0); b 436 include/linux/suspend.h static inline void __init register_nosave_region_late(unsigned long b, unsigned long e) b 438 include/linux/suspend.h __register_nosave_region(b, e, 1); b 455 include/linux/suspend.h static inline void register_nosave_region(unsigned long b, unsigned long e) {} b 456 include/linux/suspend.h static inline void register_nosave_region_late(unsigned long b, unsigned long e) {} b 67 include/linux/tee_drv.h u64 b; b 199 include/linux/ti_wilink_st.h #define MAKEWORD(a, b) ((unsigned short)(((unsigned char)(a)) \ b 200 include/linux/ti_wilink_st.h | ((unsigned short)((unsigned char)(b))) << 8)) b 97 include/linux/time.h #define time_after32(a, b) ((s32)((u32)(b) - (u32)(a)) < 0) b 98 include/linux/time.h #define time_before32(b, a) time_after32(a, b) b 110 include/linux/time32.h const struct timespec *b) b 112 include/linux/time32.h return (a->tv_sec == b->tv_sec) && (a->tv_nsec == b->tv_nsec); b 50 include/linux/time64.h const struct timespec64 *b) b 52 include/linux/time64.h return (a->tv_sec == b->tv_sec) && (a->tv_nsec == b->tv_nsec); b 35 include/linux/tnum.h struct tnum tnum_add(struct tnum a, struct tnum b); b 37 include/linux/tnum.h struct tnum tnum_sub(struct tnum a, struct tnum b); b 39 include/linux/tnum.h struct tnum tnum_and(struct tnum a, struct tnum b); b 41 include/linux/tnum.h struct tnum tnum_or(struct tnum a, struct tnum b); b 43 include/linux/tnum.h struct tnum tnum_xor(struct tnum a, struct tnum b); b 45 include/linux/tnum.h struct tnum tnum_mul(struct tnum a, struct tnum b); b 48 include/linux/tnum.h struct tnum tnum_intersect(struct tnum a, struct tnum b); b 60 include/linux/tnum.h static inline bool tnum_equals_const(struct tnum a, u64 b) b 62 include/linux/tnum.h return tnum_is_const(a) && a.value == b; b 77 include/linux/tnum.h bool tnum_in(struct tnum a, struct tnum b); b 75 include/linux/tty.h static inline unsigned char *char_buf_ptr(struct tty_buffer *b, int ofs) b 77 include/linux/tty.h return ((unsigned char *)b->data) + ofs; b 80 include/linux/tty.h static inline char *flag_buf_ptr(struct tty_buffer *b, int ofs) b 82 include/linux/tty.h return (char *)char_buf_ptr(b, ofs) + b->size; b 540 include/linux/tty.h extern int tty_termios_hw_change(const struct ktermios *a, const struct ktermios *b); b 13 include/linux/ucs2_string.h int ucs2_strncmp(const ucs2_char_t *a, const ucs2_char_t *b, size_t len); b 17 include/linux/uuid.h __u8 b[UUID_SIZE]; b 20 include/linux/uuid.h #define UUID_INIT(a, b, c, d0, d1, d2, d3, d4, d5, d6, d7) \ b 23 include/linux/uuid.h ((b) >> 8) & 0xff, (b) & 0xff, \ b 214 include/linux/xz.h XZ_EXTERN enum xz_ret xz_dec_run(struct xz_dec *s, struct xz_buf *b); b 113 include/media/davinci/dm355_ccdc.h unsigned char b; b 92 include/media/davinci/dm644x_ccdc.h char b; b 232 include/media/dvb_vb2.h int dvb_vb2_querybuf(struct dvb_vb2_ctx *ctx, struct dmx_buffer *b); b 255 include/media/dvb_vb2.h int dvb_vb2_qbuf(struct dvb_vb2_ctx *ctx, struct dmx_buffer *b); b 267 include/media/dvb_vb2.h int dvb_vb2_dqbuf(struct dvb_vb2_ctx *ctx, struct dmx_buffer *b); b 308 include/media/dvbdev.h #define dvb_register_media_controller(a, b) {} b 19 include/media/tpg/v4l2-tpg.h unsigned char r, g, b; b 23 include/media/tpg/v4l2-tpg.h __u16 r, g, b; b 451 include/media/v4l2-common.h #define V4L2_FRACT_COMPARE(a, OP, b) \ b 452 include/media/v4l2-common.h ((u64)(a).numerator * (b).denominator OP \ b 453 include/media/v4l2-common.h (u64)(b).numerator * (a).denominator) b 411 include/media/v4l2-ioctl.h struct v4l2_requestbuffers *b); b 413 include/media/v4l2-ioctl.h struct v4l2_buffer *b); b 415 include/media/v4l2-ioctl.h struct v4l2_buffer *b); b 419 include/media/v4l2-ioctl.h struct v4l2_buffer *b); b 422 include/media/v4l2-ioctl.h struct v4l2_create_buffers *b); b 424 include/media/v4l2-ioctl.h struct v4l2_buffer *b); b 488 include/media/v4l2-mem2mem.h #define v4l2_m2m_for_each_dst_buf(m2m_ctx, b) \ b 489 include/media/v4l2-mem2mem.h list_for_each_entry(b, &m2m_ctx->cap_q_ctx.rdy_queue, list) b 497 include/media/v4l2-mem2mem.h #define v4l2_m2m_for_each_src_buf(m2m_ctx, b) \ b 498 include/media/v4l2-mem2mem.h list_for_each_entry(b, &m2m_ctx->out_q_ctx.rdy_queue, list) b 508 include/media/v4l2-mem2mem.h #define v4l2_m2m_for_each_dst_buf_safe(m2m_ctx, b, n) \ b 509 include/media/v4l2-mem2mem.h list_for_each_entry_safe(b, n, &m2m_ctx->cap_q_ctx.rdy_queue, list) b 519 include/media/v4l2-mem2mem.h #define v4l2_m2m_for_each_src_buf_safe(m2m_ctx, b, n) \ b 520 include/media/v4l2-mem2mem.h list_for_each_entry_safe(b, n, &m2m_ctx->out_q_ctx.rdy_queue, list) b 201 include/media/videobuf-core.h int videobuf_querybuf(struct videobuf_queue *q, struct v4l2_buffer *b); b 203 include/media/videobuf-core.h struct v4l2_buffer *b); b 205 include/media/videobuf-core.h struct v4l2_buffer *b, int nonblocking); b 74 include/media/videobuf2-v4l2.h int vb2_querybuf(struct vb2_queue *q, struct v4l2_buffer *b); b 119 include/media/videobuf2-v4l2.h struct v4l2_buffer *b); b 145 include/media/videobuf2-v4l2.h struct v4l2_buffer *b); b 182 include/media/videobuf2-v4l2.h int vb2_dqbuf(struct vb2_queue *q, struct v4l2_buffer *b, bool nonblocking); b 195 include/net/bluetooth/bluetooth.h __u8 b[6]; b 1400 include/net/bluetooth/hci_core.h if ((bdaddr->b[5] & 0xc0) == 0x40) b 1412 include/net/bluetooth/hci_core.h if ((addr->b[5] & 0xc0) == 0xc0) b 118 include/net/cfg802154.h wpan_phy_supported_bool(bool b, enum nl802154_supported_bool_states st) b 122 include/net/cfg802154.h return b; b 124 include/net/cfg802154.h return !b; b 151 include/net/cfg802154.h wpan_phy_cca_cmp(const struct wpan_phy_cca *a, const struct wpan_phy_cca *b) b 153 include/net/cfg802154.h if (a->mode != b->mode) b 157 include/net/cfg802154.h return a->opt == b->opt; b 78 include/net/codel.h #define codel_time_after(a, b) \ b 80 include/net/codel.h typecheck(codel_time_t, b) && \ b 81 include/net/codel.h ((s32)((a) - (b)) > 0)) b 82 include/net/codel.h #define codel_time_before(a, b) codel_time_after(b, a) b 84 include/net/codel.h #define codel_time_after_eq(a, b) \ b 86 include/net/codel.h typecheck(codel_time_t, b) && \ b 87 include/net/codel.h ((s32)((a) - (b)) >= 0)) b 88 include/net/codel.h #define codel_time_before_eq(a, b) codel_time_after_eq(b, a) b 64 include/net/dst_metadata.h const struct metadata_dst *a, *b; b 70 include/net/dst_metadata.h b = (const struct metadata_dst *) skb_dst(skb_b); b 72 include/net/dst_metadata.h if (!a != !b || a->type != b->type) b 77 include/net/dst_metadata.h return memcmp(&a->u.port_info, &b->u.port_info, b 80 include/net/dst_metadata.h return memcmp(&a->u.tun_info, &b->u.tun_info, b 176 include/net/fib_rules.h struct fib_rule_port_range *b) b 178 include/net/fib_rules.h return a->start == b->start && b 179 include/net/fib_rules.h a->end == b->end; b 42 include/net/gen_stats.h struct gnet_stats_basic_packed *b); b 46 include/net/gen_stats.h struct gnet_stats_basic_packed *b); b 50 include/net/gen_stats.h struct gnet_stats_basic_packed *b); b 126 include/net/inetpeer.h const struct inetpeer_addr *b) b 136 include/net/inetpeer.h if (a->key[i] == b->key[i]) b 138 include/net/inetpeer.h if (a->key[i] < b->key[i]) b 296 include/net/ip6_route.h static inline bool rt6_duplicate_nexthop(struct fib6_info *a, struct fib6_info *b) b 300 include/net/ip6_route.h if (a->nh || b->nh) b 301 include/net/ip6_route.h return nexthop_cmp(a->nh, b->nh); b 304 include/net/ip6_route.h nhb = b->fib6_nh; b 177 include/net/ip_vs.h const union nf_inet_addr *b) b 181 include/net/ip_vs.h return ipv6_addr_equal(&a->in6, &b->in6); b 183 include/net/ip_vs.h return a->ip == b->ip; b 521 include/net/ipv6.h b = plen & 0x7; b 525 include/net/ipv6.h if (b != 0) b 526 include/net/ipv6.h pfx->s6_addr[o] = addr->s6_addr[o] & (0xff00 >> b); b 535 include/net/ipv6.h b = plen & 0x7; b 538 include/net/ipv6.h if (b != 0) { b 539 include/net/ipv6.h addr->s6_addr[o] &= ~(0xff00 >> b); b 540 include/net/ipv6.h addr->s6_addr[o] |= (pfx->s6_addr[o] & (0xff00 >> b)); b 166 include/net/llc_c_ac.h u8 llc_circular_between(u8 a, u8 b, u8 c); b 47 include/net/lwtunnel.h int (*cmp_encap)(struct lwtunnel_state *a, struct lwtunnel_state *b); b 125 include/net/lwtunnel.h int lwtunnel_cmp_encap(struct lwtunnel_state *a, struct lwtunnel_state *b); b 239 include/net/lwtunnel.h struct lwtunnel_state *b) b 12 include/net/netfilter/br_netfilter.h struct nf_bridge_info *b = skb_ext_add(skb, SKB_EXT_BRIDGE_NF); b 14 include/net/netfilter/br_netfilter.h if (b) b 15 include/net/netfilter/br_netfilter.h memset(b, 0, sizeof(*b)); b 17 include/net/netfilter/br_netfilter.h return b; b 68 include/net/netfilter/nf_conntrack_zones.h const struct nf_conntrack_zone *b, b 73 include/net/netfilter/nf_conntrack_zones.h nf_ct_zone_id(b, dir); b 80 include/net/netfilter/nf_conntrack_zones.h const struct nf_conntrack_zone *b) b 83 include/net/netfilter/nf_conntrack_zones.h return nf_ct_zone(a)->id == b->id; b 59 include/net/netfilter/nf_queue.h u32 a, b, c; b 64 include/net/netfilter/nf_queue.h b = (__force u32) ip6h->daddr.s6_addr32[3]; b 66 include/net/netfilter/nf_queue.h b = (__force u32) ip6h->saddr.s6_addr32[3]; b 76 include/net/netfilter/nf_queue.h return jhash_3words(a, b, c, initval); b 341 include/net/sctp/sm.h #define TSN_lt(a,b) \ b 343 include/net/sctp/sm.h typecheck(__u32, b) && \ b 344 include/net/sctp/sm.h ((__s32)((a) - (b)) < 0)) b 346 include/net/sctp/sm.h #define TSN_lte(a,b) \ b 348 include/net/sctp/sm.h typecheck(__u32, b) && \ b 349 include/net/sctp/sm.h ((__s32)((a) - (b)) <= 0)) b 352 include/net/sctp/sm.h #define MID_lt(a, b) \ b 354 include/net/sctp/sm.h typecheck(__u32, b) && \ b 355 include/net/sctp/sm.h ((__s32)((a) - (b)) < 0)) b 358 include/net/sctp/sm.h #define SSN_lt(a,b) \ b 360 include/net/sctp/sm.h typecheck(__u16, b) && \ b 361 include/net/sctp/sm.h ((__s16)((a) - (b)) < 0)) b 364 include/net/sctp/sm.h #define ADDIP_SERIAL_gte(a,b) \ b 366 include/net/sctp/sm.h typecheck(__u32, b) && \ b 367 include/net/sctp/sm.h ((__s32)((b) - (a)) <= 0)) b 1695 include/net/xfrm.h const xfrm_address_t *b) b 1698 include/net/xfrm.h (const struct in6_addr *)b); b 1702 include/net/xfrm.h const xfrm_address_t *b, b 1708 include/net/xfrm.h return ((__force u32)a->a4 ^ (__force u32)b->a4) == 0; b 1710 include/net/xfrm.h return xfrm6_addr_equal(a, b); b 384 include/ras/ras_event.h #define EM(a, b) TRACE_DEFINE_ENUM(a); b 385 include/ras/ras_event.h #define EMe(a, b) TRACE_DEFINE_ENUM(a); b 396 include/ras/ras_event.h #define EM(a, b) { a, b }, b 397 include/ras/ras_event.h #define EMe(a, b) { a, b } b 638 include/rdma/rdmavt_qp.h static inline int rvt_cmp_msn(u32 a, u32 b) b 640 include/rdma/rdmavt_qp.h return (((int)a) - ((int)b)) << 8; b 448 include/sound/gus.h static inline void snd_gf1_uart_cmd(struct snd_gus_card * gus, unsigned char b) b 450 include/sound/gus.h outb(gus->gf1.uart_cmd = b, GUSP(gus, MIDICTRL)); b 458 include/sound/gus.h static inline void snd_gf1_uart_put(struct snd_gus_card * gus, unsigned char b) b 460 include/sound/gus.h outb(b, GUSP(gus, MIDIDATA)); b 83 include/trace/events/9p.h #define EM(a, b) TRACE_DEFINE_ENUM(a); b 84 include/trace/events/9p.h #define EMe(a, b) TRACE_DEFINE_ENUM(a); b 94 include/trace/events/9p.h #define EM(a, b) { a, b }, b 95 include/trace/events/9p.h #define EMe(a, b) { a, b } b 438 include/trace/events/afs.h #define EM(a, b) TRACE_DEFINE_ENUM(a); b 439 include/trace/events/afs.h #define E_(a, b) TRACE_DEFINE_ENUM(a); b 460 include/trace/events/afs.h #define EM(a, b) { a, b }, b 461 include/trace/events/afs.h #define E_(a, b) { a, b } b 64 include/trace/events/bcache.h TP_PROTO(struct btree *b), b 65 include/trace/events/bcache.h TP_ARGS(b), b 72 include/trace/events/bcache.h __entry->bucket = PTR_BUCKET_NR(b->c, &b->key, 0); b 258 include/trace/events/bcache.h TP_PROTO(struct btree *b), b 259 include/trace/events/bcache.h TP_ARGS(b) b 263 include/trace/events/bcache.h TP_PROTO(struct btree *b), b 264 include/trace/events/bcache.h TP_ARGS(b), b 273 include/trace/events/bcache.h __entry->bucket = PTR_BUCKET_NR(b->c, &b->key, 0); b 274 include/trace/events/bcache.h __entry->block = b->written; b 275 include/trace/events/bcache.h __entry->keys = b->keys.set[b->keys.nsets].data->keys; b 282 include/trace/events/bcache.h TP_PROTO(struct btree *b), b 283 include/trace/events/bcache.h TP_ARGS(b) b 292 include/trace/events/bcache.h TP_PROTO(struct btree *b), b 293 include/trace/events/bcache.h TP_ARGS(b) b 332 include/trace/events/bcache.h TP_PROTO(struct btree *b, struct bkey *k, unsigned op, unsigned status), b 333 include/trace/events/bcache.h TP_ARGS(b, k, op, status), b 347 include/trace/events/bcache.h __entry->btree_node = PTR_BUCKET_NR(b->c, &b->key, 0); b 348 include/trace/events/bcache.h __entry->btree_level = b->level; b 365 include/trace/events/bcache.h TP_PROTO(struct btree *b, unsigned keys), b 366 include/trace/events/bcache.h TP_ARGS(b, keys), b 374 include/trace/events/bcache.h __entry->bucket = PTR_BUCKET_NR(b->c, &b->key, 0); b 382 include/trace/events/bcache.h TP_PROTO(struct btree *b, unsigned keys), b 383 include/trace/events/bcache.h TP_ARGS(b, keys) b 387 include/trace/events/bcache.h TP_PROTO(struct btree *b, unsigned keys), b 388 include/trace/events/bcache.h TP_ARGS(b, keys) b 392 include/trace/events/bcache.h TP_PROTO(struct btree *b), b 393 include/trace/events/bcache.h TP_ARGS(b) b 55 include/trace/events/cachefiles.h #define EM(a, b) TRACE_DEFINE_ENUM(a); b 56 include/trace/events/cachefiles.h #define E_(a, b) TRACE_DEFINE_ENUM(a); b 67 include/trace/events/cachefiles.h #define EM(a, b) { a, b }, b 68 include/trace/events/cachefiles.h #define E_(a, b) { a, b } b 147 include/trace/events/fscache.h #define EM(a, b) TRACE_DEFINE_ENUM(a); b 148 include/trace/events/fscache.h #define E_(a, b) TRACE_DEFINE_ENUM(a); b 158 include/trace/events/fscache.h #define EM(a, b) { a, b }, b 159 include/trace/events/fscache.h #define E_(a, b) { a, b } b 39 include/trace/events/huge_memory.h #define EM(a, b) TRACE_DEFINE_ENUM(a); b 40 include/trace/events/huge_memory.h #define EMe(a, b) TRACE_DEFINE_ENUM(a); b 46 include/trace/events/huge_memory.h #define EM(a, b) {a, b}, b 47 include/trace/events/huge_memory.h #define EMe(a, b) {a, b} b 31 include/trace/events/migrate.h #define EM(a, b) TRACE_DEFINE_ENUM(a); b 32 include/trace/events/migrate.h #define EMe(a, b) TRACE_DEFINE_ENUM(a); b 43 include/trace/events/migrate.h #define EM(a, b) {a, b}, b 44 include/trace/events/migrate.h #define EMe(a, b) {a, b} b 248 include/trace/events/mmflags.h #define EM(a, b) TRACE_DEFINE_ENUM(a); b 249 include/trace/events/mmflags.h #define EMe(a, b) TRACE_DEFINE_ENUM(a); b 263 include/trace/events/mmflags.h #define EM(a, b) {a, b}, b 264 include/trace/events/mmflags.h #define EMe(a, b) {a, b} b 465 include/trace/events/rxrpc.h #define EM(a, b) TRACE_DEFINE_ENUM(a); b 466 include/trace/events/rxrpc.h #define E_(a, b) TRACE_DEFINE_ENUM(a); b 491 include/trace/events/rxrpc.h #define EM(a, b) { a, b }, b 492 include/trace/events/rxrpc.h #define E_(a, b) { a, b } b 507 include/trace/events/sunrpc.h #define EM(a, b) TRACE_DEFINE_ENUM(a); b 508 include/trace/events/sunrpc.h #define EMe(a, b) TRACE_DEFINE_ENUM(a); b 546 include/trace/events/sunrpc.h #define EM(a, b) {a, b}, b 547 include/trace/events/sunrpc.h #define EMe(a, b) {a, b} b 24 include/trace/events/tlb.h #define EM(a,b) TRACE_DEFINE_ENUM(a); b 25 include/trace/events/tlb.h #define EMe(a,b) TRACE_DEFINE_ENUM(a); b 35 include/trace/events/tlb.h #define EM(a,b) { a, b }, b 36 include/trace/events/tlb.h #define EMe(a,b) { a, b } b 14 include/trace/events/v4l2.h #define EM(a, b) TRACE_DEFINE_ENUM(a); b 15 include/trace/events/v4l2.h #define EMe(a, b) TRACE_DEFINE_ENUM(a); b 61 include/trace/events/v4l2.h #define EM(a, b) {a, b}, b 62 include/trace/events/v4l2.h #define EMe(a, b) {a, b} b 30 include/trace/events/writeback.h #define EM(a,b) TRACE_DEFINE_ENUM(a); b 31 include/trace/events/writeback.h #define EMe(a,b) TRACE_DEFINE_ENUM(a); b 51 include/trace/events/writeback.h #define EM(a,b) { a, b }, b 52 include/trace/events/writeback.h #define EMe(a,b) { a, b } b 69 include/uapi/drm/drm_fourcc.h #define fourcc_code(a, b, c, d) ((__u32)(a) | ((__u32)(b) << 8) | \ b 97 include/uapi/linux/am437x-vpfe.h char b; b 57 include/uapi/linux/hdreg.h } b; b 98 include/uapi/linux/map_to_7segment.h #define _SEG7(l,a,b,c,d,e,f,g) \ b 99 include/uapi/linux/map_to_7segment.h ( a<<BIT_SEG7_A | b<<BIT_SEG7_B | c<<BIT_SEG7_C | d<<BIT_SEG7_D | \ b 107 include/uapi/linux/netfilter/x_tables.h #define SET_COUNTER(c,b,p) do { (c).bcnt = (b); (c).pcnt = (p); } while(0) b 108 include/uapi/linux/netfilter/x_tables.h #define ADD_COUNTER(c,b,p) do { (c).bcnt += (b); (c).pcnt += (p); } while(0) b 19 include/uapi/linux/romfs_fs.h #define __mk4(a,b,c,d) cpu_to_be32(__mkl(__mkw(a,b),__mkw(c,d))) b 197 include/uapi/linux/tee.h __u64 b; b 24 include/uapi/linux/uuid.h __u8 b[16]; b 27 include/uapi/linux/uuid.h #define GUID_INIT(a, b, c, d0, d1, d2, d3, d4, d5, d6, d7) \ b 30 include/uapi/linux/uuid.h (b) & 0xff, ((b) >> 8) & 0xff, \ b 36 include/uapi/linux/uuid.h #define UUID_LE(a, b, c, d0, d1, d2, d3, d4, d5, d6, d7) \ b 37 include/uapi/linux/uuid.h GUID_INIT(a, b, c, d0, d1, d2, d3, d4, d5, d6, d7) b 81 include/uapi/linux/videodev2.h #define v4l2_fourcc(a, b, c, d)\ b 82 include/uapi/linux/videodev2.h ((__u32)(a) | ((__u32)(b) << 8) | ((__u32)(c) << 16) | ((__u32)(d) << 24)) b 83 include/uapi/linux/videodev2.h #define v4l2_fourcc_be(a, b, c, d) (v4l2_fourcc(a, b, c, d) | (1U << 31)) b 51 include/uapi/mtd/inftl-user.h struct inftl_unithead2 b; b 56 include/uapi/mtd/inftl-user.h struct inftl_bci b; b 57 include/uapi/mtd/nftl-user.h struct nftl_uci1 b; b 62 include/uapi/mtd/nftl-user.h struct nftl_bci b; b 504 include/video/pm3fb.h #define PM3ConstantColorDDA_B(b) (((b) & 0xff) << 16) b 410 init/do_mounts.c char b[BDEVNAME_SIZE]; b 412 init/do_mounts.c const char *b = name; b 432 init/do_mounts.c __bdevname(ROOT_DEV, b); b 435 init/do_mounts.c root_device_name, b, err); b 443 init/do_mounts.c panic("VFS: Unable to mount root fs on %s", b); b 457 init/do_mounts.c __bdevname(ROOT_DEV, b); b 459 init/do_mounts.c panic("VFS: Unable to mount root fs on %s", b); b 696 kernel/auditfilter.c static int audit_compare_rule(struct audit_krule *a, struct audit_krule *b) b 700 kernel/auditfilter.c if (a->flags != b->flags || b 701 kernel/auditfilter.c a->pflags != b->pflags || b 702 kernel/auditfilter.c a->listnr != b->listnr || b 703 kernel/auditfilter.c a->action != b->action || b 704 kernel/auditfilter.c a->field_count != b->field_count) b 708 kernel/auditfilter.c if (a->fields[i].type != b->fields[i].type || b 709 kernel/auditfilter.c a->fields[i].op != b->fields[i].op) b 723 kernel/auditfilter.c if (strcmp(a->fields[i].lsm_str, b->fields[i].lsm_str)) b 728 kernel/auditfilter.c audit_watch_path(b->watch))) b 733 kernel/auditfilter.c audit_tree_path(b->tree))) b 738 kernel/auditfilter.c if (strcmp(a->filterkey, b->filterkey)) b 744 kernel/auditfilter.c audit_mark_path(b->exe))) b 753 kernel/auditfilter.c if (!uid_eq(a->fields[i].uid, b->fields[i].uid)) b 761 kernel/auditfilter.c if (!gid_eq(a->fields[i].gid, b->fields[i].gid)) b 765 kernel/auditfilter.c if (a->fields[i].val != b->fields[i].val) b 771 kernel/auditfilter.c if (a->mask[i] != b->mask[i]) b 3155 kernel/bpf/btf.c static int btf_sec_info_cmp(const void *a, const void *b) b 3158 kernel/bpf/btf.c const struct btf_sec_info *y = b; b 588 kernel/bpf/core.c struct latch_tree_node *b) b 590 kernel/bpf/core.c return bpf_get_prog_addr_start(a) < bpf_get_prog_addr_start(b); b 577 kernel/bpf/hashtab.c struct bucket *b; b 580 kernel/bpf/hashtab.c b = __select_bucket(htab, tgt_l->hash); b 581 kernel/bpf/hashtab.c head = &b->head; b 583 kernel/bpf/hashtab.c raw_spin_lock_irqsave(&b->lock, flags); b 591 kernel/bpf/hashtab.c raw_spin_unlock_irqrestore(&b->lock, flags); b 826 kernel/bpf/hashtab.c struct bucket *b; b 840 kernel/bpf/hashtab.c b = __select_bucket(htab, hash); b 841 kernel/bpf/hashtab.c head = &b->head; b 866 kernel/bpf/hashtab.c raw_spin_lock_irqsave(&b->lock, flags); b 907 kernel/bpf/hashtab.c raw_spin_unlock_irqrestore(&b->lock, flags); b 918 kernel/bpf/hashtab.c struct bucket *b; b 932 kernel/bpf/hashtab.c b = __select_bucket(htab, hash); b 933 kernel/bpf/hashtab.c head = &b->head; b 946 kernel/bpf/hashtab.c raw_spin_lock_irqsave(&b->lock, flags); b 965 kernel/bpf/hashtab.c raw_spin_unlock_irqrestore(&b->lock, flags); b 983 kernel/bpf/hashtab.c struct bucket *b; b 997 kernel/bpf/hashtab.c b = __select_bucket(htab, hash); b 998 kernel/bpf/hashtab.c head = &b->head; b 1001 kernel/bpf/hashtab.c raw_spin_lock_irqsave(&b->lock, flags); b 1024 kernel/bpf/hashtab.c raw_spin_unlock_irqrestore(&b->lock, flags); b 1036 kernel/bpf/hashtab.c struct bucket *b; b 1050 kernel/bpf/hashtab.c b = __select_bucket(htab, hash); b 1051 kernel/bpf/hashtab.c head = &b->head; b 1065 kernel/bpf/hashtab.c raw_spin_lock_irqsave(&b->lock, flags); b 1087 kernel/bpf/hashtab.c raw_spin_unlock_irqrestore(&b->lock, flags); b 1111 kernel/bpf/hashtab.c struct bucket *b; b 1122 kernel/bpf/hashtab.c b = __select_bucket(htab, hash); b 1123 kernel/bpf/hashtab.c head = &b->head; b 1125 kernel/bpf/hashtab.c raw_spin_lock_irqsave(&b->lock, flags); b 1135 kernel/bpf/hashtab.c raw_spin_unlock_irqrestore(&b->lock, flags); b 1143 kernel/bpf/hashtab.c struct bucket *b; b 1154 kernel/bpf/hashtab.c b = __select_bucket(htab, hash); b 1155 kernel/bpf/hashtab.c head = &b->head; b 1157 kernel/bpf/hashtab.c raw_spin_lock_irqsave(&b->lock, flags); b 1166 kernel/bpf/hashtab.c raw_spin_unlock_irqrestore(&b->lock, flags); b 62 kernel/bpf/tnum.c struct tnum tnum_add(struct tnum a, struct tnum b) b 66 kernel/bpf/tnum.c sm = a.mask + b.mask; b 67 kernel/bpf/tnum.c sv = a.value + b.value; b 70 kernel/bpf/tnum.c mu = chi | a.mask | b.mask; b 74 kernel/bpf/tnum.c struct tnum tnum_sub(struct tnum a, struct tnum b) b 78 kernel/bpf/tnum.c dv = a.value - b.value; b 80 kernel/bpf/tnum.c beta = dv - b.mask; b 82 kernel/bpf/tnum.c mu = chi | a.mask | b.mask; b 86 kernel/bpf/tnum.c struct tnum tnum_and(struct tnum a, struct tnum b) b 91 kernel/bpf/tnum.c beta = b.value | b.mask; b 92 kernel/bpf/tnum.c v = a.value & b.value; b 96 kernel/bpf/tnum.c struct tnum tnum_or(struct tnum a, struct tnum b) b 100 kernel/bpf/tnum.c v = a.value | b.value; b 101 kernel/bpf/tnum.c mu = a.mask | b.mask; b 105 kernel/bpf/tnum.c struct tnum tnum_xor(struct tnum a, struct tnum b) b 109 kernel/bpf/tnum.c v = a.value ^ b.value; b 110 kernel/bpf/tnum.c mu = a.mask | b.mask; b 128 kernel/bpf/tnum.c struct tnum tnum_mul(struct tnum a, struct tnum b) b 133 kernel/bpf/tnum.c pi = a.value * b.value; b 134 kernel/bpf/tnum.c acc = hma(TNUM(pi, 0), a.mask, b.mask | b.value); b 135 kernel/bpf/tnum.c return hma(acc, b.mask, a.value); b 141 kernel/bpf/tnum.c struct tnum tnum_intersect(struct tnum a, struct tnum b) b 145 kernel/bpf/tnum.c v = a.value | b.value; b 146 kernel/bpf/tnum.c mu = a.mask & b.mask; b 164 kernel/bpf/tnum.c bool tnum_in(struct tnum a, struct tnum b) b 166 kernel/bpf/tnum.c if (b.mask & ~a.mask) b 168 kernel/bpf/tnum.c b.value &= ~a.mask; b 169 kernel/bpf/tnum.c return a.value == b.value; b 1081 kernel/bpf/verifier.c static int cmp_subprogs(const void *a, const void *b) b 1084 kernel/bpf/verifier.c ((struct bpf_subprog_info *)b)->start; b 4177 kernel/bpf/verifier.c static bool signed_add_overflows(s64 a, s64 b) b 4180 kernel/bpf/verifier.c s64 res = (s64)((u64)a + (u64)b); b 4182 kernel/bpf/verifier.c if (b < 0) b 4187 kernel/bpf/verifier.c static bool signed_sub_overflows(s64 a, s64 b) b 4190 kernel/bpf/verifier.c s64 res = (s64)((u64)a - (u64)b); b 4192 kernel/bpf/verifier.c if (b < 0) b 276 kernel/cgroup/cgroup-v1.c static int cmppid(const void *a, const void *b) b 278 kernel/cgroup/cgroup-v1.c return *(pid_t *)a - *(pid_t *)b; b 640 kernel/cgroup/cpuset.c static int cpusets_overlap(struct cpuset *a, struct cpuset *b) b 642 kernel/cgroup/cpuset.c return cpumask_intersects(a->effective_cpus, b->effective_cpus); b 822 kernel/cgroup/cpuset.c struct cpuset *b = csa[j]; b 823 kernel/cgroup/cpuset.c int bpn = b->pn; b 825 kernel/cgroup/cpuset.c if (apn != bpn && cpusets_overlap(a, b)) { b 879 kernel/cgroup/cpuset.c struct cpuset *b = csa[j]; b 881 kernel/cgroup/cpuset.c if (apn == b->pn) { b 882 kernel/cgroup/cpuset.c cpumask_or(dp, dp, b->effective_cpus); b 885 kernel/cgroup/cpuset.c update_domain_attr_tree(dattr + nslot, b); b 888 kernel/cgroup/cpuset.c b->pn = -1; b 613 kernel/cred.c int cred_fscmp(const struct cred *a, const struct cred *b) b 618 kernel/cred.c if (a == b) b 620 kernel/cred.c if (uid_lt(a->fsuid, b->fsuid)) b 622 kernel/cred.c if (uid_gt(a->fsuid, b->fsuid)) b 625 kernel/cred.c if (gid_lt(a->fsgid, b->fsgid)) b 627 kernel/cred.c if (gid_gt(a->fsgid, b->fsgid)) b 631 kernel/cred.c gb = b->group_info; b 269 kernel/dma/debug.c static bool exact_match(struct dma_debug_entry *a, struct dma_debug_entry *b) b 271 kernel/dma/debug.c return ((a->dev_addr == b->dev_addr) && b 272 kernel/dma/debug.c (a->dev == b->dev)) ? true : false; b 276 kernel/dma/debug.c struct dma_debug_entry *b) b 278 kernel/dma/debug.c if (a->dev != b->dev) b 281 kernel/dma/debug.c if ((b->dev_addr <= a->dev_addr) && b 282 kernel/dma/debug.c ((b->dev_addr + b->size) >= (a->dev_addr + a->size))) b 3621 kernel/events/core.c #define REDUCE_FLS(a, b) \ b 3623 kernel/events/core.c if (a##_fls > b##_fls) { \ b 3627 kernel/events/core.c b >>= 1; \ b 3628 kernel/events/core.c b##_fls--; \ b 10814 kernel/events/core.c static void mutex_lock_double(struct mutex *a, struct mutex *b) b 10816 kernel/events/core.c if (b < a) b 10817 kernel/events/core.c swap(a, b); b 10820 kernel/events/core.c mutex_lock_nested(b, SINGLE_DEPTH_NESTING); b 84 kernel/groups.c kgid_t b = *(kgid_t *)_b; b 86 kernel/groups.c return gid_gt(a, b) - gid_lt(a, b); b 35 kernel/jump_label.c static int jump_label_cmp(const void *a, const void *b) b 38 kernel/jump_label.c const struct jump_entry *jeb = b; b 63 kernel/jump_label.c static void jump_label_swap(void *a, void *b, int size) b 65 kernel/jump_label.c long delta = (unsigned long)a - (unsigned long)b; b 67 kernel/jump_label.c struct jump_entry *jeb = b; b 352 kernel/locking/mutex.c __ww_ctx_stamp_after(struct ww_acquire_ctx *a, struct ww_acquire_ctx *b) b 355 kernel/locking/mutex.c return (signed long)(a->stamp - b->stamp) > 0; b 120 kernel/module.c mod_tree_less(struct latch_tree_node *a, struct latch_tree_node *b) b 122 kernel/module.c return __mod_tree_val(a) < __mod_tree_val(b); b 831 kernel/module.c static int already_uses(struct module *a, struct module *b) b 835 kernel/module.c list_for_each_entry(use, &b->source_list, source_list) { b 837 kernel/module.c pr_debug("%s uses %s!\n", a->name, b->name); b 841 kernel/module.c pr_debug("%s does not use %s!\n", a->name, b->name); b 852 kernel/module.c static int add_module_usage(struct module *a, struct module *b) b 862 kernel/module.c use->target = b; b 863 kernel/module.c list_add(&use->source_list, &b->source_list); b 869 kernel/module.c int ref_module(struct module *a, struct module *b) b 873 kernel/module.c if (b == NULL || already_uses(a, b)) b 877 kernel/module.c err = strong_try_module_get(b); b 881 kernel/module.c err = add_module_usage(a, b); b 883 kernel/module.c module_put(b); b 1169 kernel/module.c int ref_module(struct module *a, struct module *b) b 1171 kernel/module.c return strong_try_module_get(b); b 84 kernel/params.c bool parameqn(const char *a, const char *b, size_t n) b 89 kernel/params.c if (dash2underscore(a[i]) != dash2underscore(b[i])) b 95 kernel/params.c bool parameq(const char *a, const char *b) b 97 kernel/params.c return parameqn(a, b, strlen(a)+1); b 2844 kernel/printk/printk.c struct console *a, *b; b 2861 kernel/printk/printk.c for (a=console_drivers->next, b=console_drivers ; b 2862 kernel/printk/printk.c a; b=a, a=b->next) { b 2864 kernel/printk/printk.c b->next = a->next; b 1401 kernel/rcu/tree.c rdp->cpu_no_qs.b.norm = need_gp; b 1958 kernel/rcu/tree.c if (rdp->cpu_no_qs.b.norm || rdp->gp_seq != rnp->gp_seq || b 1967 kernel/rcu/tree.c rdp->cpu_no_qs.b.norm = true; /* need qs for new gp. */ b 2013 kernel/rcu/tree.c if (rdp->cpu_no_qs.b.norm) b 2807 kernel/rcu/tree.c if (rdp->core_needs_qs && !rdp->cpu_no_qs.b.norm) b 3046 kernel/rcu/tree.c rdp->cpu_no_qs.b.norm = true; b 144 kernel/rcu/tree.h } b; /* Bits. */ b 639 kernel/rcu/tree_exp.h t->rcu_read_unlock_special.b.exp_hint = true; b 702 kernel/rcu/tree_exp.h __this_cpu_write(rcu_data.cpu_no_qs.b.exp, true); b 718 kernel/rcu/tree_exp.h __this_cpu_read(rcu_data.cpu_no_qs.b.exp)) b 741 kernel/rcu/tree_exp.h __this_cpu_read(rcu_data.cpu_no_qs.b.exp)) { b 266 kernel/rcu/tree_plugin.h __this_cpu_write(rcu_data.cpu_no_qs.b.norm, false); b 268 kernel/rcu/tree_plugin.h WRITE_ONCE(current->rcu_read_unlock_special.b.need_qs, false); b 295 kernel/rcu/tree_plugin.h !t->rcu_read_unlock_special.b.blocked) { b 300 kernel/rcu/tree_plugin.h t->rcu_read_unlock_special.b.blocked = true; b 447 kernel/rcu/tree_plugin.h t->rcu_read_unlock_special.b.deferred_qs = false; b 448 kernel/rcu/tree_plugin.h if (special.b.need_qs) { b 450 kernel/rcu/tree_plugin.h t->rcu_read_unlock_special.b.need_qs = false; b 472 kernel/rcu/tree_plugin.h if (special.b.blocked) { b 473 kernel/rcu/tree_plugin.h t->rcu_read_unlock_special.b.blocked = false; b 613 kernel/rcu/tree_plugin.h t->rcu_read_unlock_special.b.exp_hint = false; b 620 kernel/rcu/tree_plugin.h (exp && !t->rcu_read_unlock_special.b.deferred_qs))) { b 639 kernel/rcu/tree_plugin.h t->rcu_read_unlock_special.b.deferred_qs = true; b 643 kernel/rcu/tree_plugin.h WRITE_ONCE(t->rcu_read_unlock_special.b.exp_hint, false); b 707 kernel/rcu/tree_plugin.h __this_cpu_read(rcu_data.cpu_no_qs.b.norm) && b 708 kernel/rcu/tree_plugin.h !t->rcu_read_unlock_special.b.need_qs && b 710 kernel/rcu/tree_plugin.h t->rcu_read_unlock_special.b.need_qs = true; b 728 kernel/rcu/tree_plugin.h WRITE_ONCE(t->rcu_read_unlock_special.b.blocked, true); b 804 kernel/rcu/tree_plugin.h __this_cpu_write(rcu_data.cpu_no_qs.b.norm, false); b 805 kernel/rcu/tree_plugin.h if (!__this_cpu_read(rcu_data.cpu_no_qs.b.exp)) b 807 kernel/rcu/tree_plugin.h __this_cpu_write(rcu_data.cpu_no_qs.b.exp, false); b 525 kernel/sched/fair.c struct sched_entity *b) b 527 kernel/sched/fair.c return (s64)(a->vruntime - b->vruntime) < 0; b 2083 kernel/sched/fair.c int a, b; b 2095 kernel/sched/fair.c for_each_node_mask(b, nodes) { b 2096 kernel/sched/fair.c if (node_distance(a, b) < dist) { b 2097 kernel/sched/fair.c faults += group_faults(p, b); b 2098 kernel/sched/fair.c node_set(b, this_group); b 2099 kernel/sched/fair.c node_clear(b, nodes); b 225 kernel/sched/sched.h dl_entity_preempt(struct sched_dl_entity *a, struct sched_dl_entity *b) b 228 kernel/sched/sched.h dl_time_before(a->deadline, b->deadline); b 715 kernel/sched/sched.h static inline bool sched_asym_prefer(int a, int b) b 717 kernel/sched/sched.h return arch_asym_cpu_priority(a) > arch_asym_cpu_priority(b); b 1524 kernel/sched/topology.c int a, b, c, n; b 1534 kernel/sched/topology.c for_each_online_node(b) { b 1536 kernel/sched/topology.c if (node_distance(a, b) < n) b 1542 kernel/sched/topology.c node_distance(b, c) < n) { b 152 kernel/signal.c #define PENDING(p,b) has_pending_signals(&(p)->signal, (b)) b 77 kernel/sys.c # define SET_UNALIGN_CTL(a, b) (-EINVAL) b 80 kernel/sys.c # define GET_UNALIGN_CTL(a, b) (-EINVAL) b 83 kernel/sys.c # define SET_FPEMU_CTL(a, b) (-EINVAL) b 86 kernel/sys.c # define GET_FPEMU_CTL(a, b) (-EINVAL) b 89 kernel/sys.c # define SET_FPEXC_CTL(a, b) (-EINVAL) b 92 kernel/sys.c # define GET_FPEXC_CTL(a, b) (-EINVAL) b 95 kernel/sys.c # define GET_ENDIAN(a, b) (-EINVAL) b 98 kernel/sys.c # define SET_ENDIAN(a, b) (-EINVAL) b 110 kernel/sys.c # define SET_FP_MODE(a,b) (-EINVAL) b 119 kernel/sys.c # define PAC_RESET_KEYS(a, b) (-EINVAL) b 287 kernel/time/hrtimer.c # define switch_hrtimer_base(t, b, p) (b) b 44 kernel/time/timeconv.c static long math_div(long a, long b) b 46 kernel/time/timeconv.c return a / b - (a % b < 0); b 1961 kernel/time/timer.c int b; b 1963 kernel/time/timer.c for (b = 0; b < NR_BASES; b++) { b 1964 kernel/time/timer.c base = per_cpu_ptr(&timer_bases[b], cpu); b 1977 kernel/time/timer.c int b, i; b 1981 kernel/time/timer.c for (b = 0; b < NR_BASES; b++) { b 1982 kernel/time/timer.c old_base = per_cpu_ptr(&timer_bases[b], cpu); b 1983 kernel/time/timer.c new_base = get_cpu_ptr(&timer_bases[b]); b 702 kernel/trace/blktrace.c char b[BDEVNAME_SIZE]; b 712 kernel/trace/blktrace.c bdevname(bdev, b); b 713 kernel/trace/blktrace.c ret = __blk_trace_setup(q, b, bdev->bd_dev, bdev, arg); b 717 kernel/trace/blktrace.c bdevname(bdev, b); b 718 kernel/trace/blktrace.c ret = compat_blk_trace_setup(q, b, bdev->bd_dev, bdev, arg); b 469 kernel/trace/ftrace.c struct ftrace_profile *b = p2; b 471 kernel/trace/ftrace.c if (a->time < b->time) b 473 kernel/trace/ftrace.c if (a->time > b->time) b 483 kernel/trace/ftrace.c struct ftrace_profile *b = p2; b 485 kernel/trace/ftrace.c if (a->counter < b->counter) b 487 kernel/trace/ftrace.c if (a->counter > b->counter) b 1525 kernel/trace/ftrace.c static int ftrace_cmp_recs(const void *a, const void *b) b 1528 kernel/trace/ftrace.c const struct dyn_ftrace *rec = b; b 5562 kernel/trace/ftrace.c static int ftrace_cmp_ips(const void *a, const void *b) b 5565 kernel/trace/ftrace.c const unsigned long *ipb = b; b 727 kernel/trace/ring_buffer.c #define RB_WARN_ON(b, cond) \ b 731 kernel/trace/ring_buffer.c if (__same_type(*(b), struct ring_buffer_per_cpu)) { \ b 733 kernel/trace/ring_buffer.c (void *)b; \ b 736 kernel/trace/ring_buffer.c atomic_inc(&b->record_disabled); \ b 1338 kernel/trace/trace.c #define C(a, b) b b 436 kernel/trace/trace.h #define TRACER_OPT(s, b) .name = #s, .bit = b b 1284 kernel/trace/trace.h #define C(a, b) TRACE_ITER_##a##_BIT b 1297 kernel/trace/trace.h #define C(a, b) TRACE_ITER_##a = (1 << TRACE_ITER_##a##_BIT) b 338 kernel/trace/trace_branch.c struct ftrace_branch_data *b = p2; b 343 kernel/trace/trace_branch.c percent_b = get_incorrect_percent(b); b 350 kernel/trace/trace_branch.c if (a->incorrect < b->incorrect) b 352 kernel/trace/trace_branch.c if (a->incorrect > b->incorrect) b 360 kernel/trace/trace_branch.c if (a->correct > b->correct) b 362 kernel/trace/trace_branch.c if (a->correct < b->correct) b 36 kernel/trace/trace_events_filter.c #define C(a, b) a b 41 kernel/trace/trace_events_filter.c #define C(a, b) b b 73 kernel/trace/trace_events_filter.c #define C(a, b) FILT_ERR_##a b 78 kernel/trace/trace_events_filter.c #define C(a, b) b b 2107 kernel/trace/trace_events_filter.c .rec = { .a = va, .b = vb, .c = vc, .d = vd, \ b 12 kernel/trace/trace_events_filter_test.h TP_PROTO(int a, int b, int c, int d, int e, int f, int g, int h), b 14 kernel/trace/trace_events_filter_test.h TP_ARGS(a, b, c, d, e, f, g, h), b 18 kernel/trace/trace_events_filter_test.h __field(int, b) b 29 kernel/trace/trace_events_filter_test.h __entry->b = b; b 39 kernel/trace/trace_events_filter_test.h __entry->a, __entry->b, __entry->c, __entry->d, b 72 kernel/trace/trace_events_hist.c #define C(a, b) HIST_ERR_##a b 77 kernel/trace/trace_events_hist.c #define C(a, b) b b 136 kernel/trace/trace_hwlat.c #define time_sub(a, b) ((a) - (b)) b 137 kernel/trace/trace_hwlat.c #define init_time(a, b) (a = b) b 17 kernel/trace/trace_probe.c #define C(a, b) b b 1099 kernel/trace/trace_probe.c int trace_probe_compare_arg_type(struct trace_probe *a, struct trace_probe *b) b 1104 kernel/trace/trace_probe.c if (a->nr_args < b->nr_args) b 1106 kernel/trace/trace_probe.c if (a->nr_args > b->nr_args) b 1107 kernel/trace/trace_probe.c return b->nr_args + 1; b 1110 kernel/trace/trace_probe.c if ((b->nr_args <= i) || b 1111 kernel/trace/trace_probe.c ((a->args[i].type != b->args[i].type) || b 1112 kernel/trace/trace_probe.c (a->args[i].count != b->args[i].count) || b 1113 kernel/trace/trace_probe.c strcmp(a->args[i].name, b->args[i].name))) b 342 kernel/trace/trace_probe.h int trace_probe_compare_arg_type(struct trace_probe *a, struct trace_probe *b); b 450 kernel/trace/trace_probe.h #define C(a, b) TP_ERR_##a b 130 kernel/trace/tracing_map.c char *b = val_b; b 132 kernel/trace/tracing_map.c return strcmp(a, b); b 143 kernel/trace/tracing_map.c u64 b = atomic64_read((atomic64_t *)val_b); b 145 kernel/trace/tracing_map.c return (a > b) ? 1 : ((a < b) ? -1 : 0); b 152 kernel/trace/tracing_map.c type b = (type)(*(u64 *)val_b); \ b 154 kernel/trace/tracing_map.c return (a > b) ? 1 : ((a < b) ? -1 : 0); \ b 838 kernel/trace/tracing_map.c const struct tracing_map_sort_entry **b) b 842 kernel/trace/tracing_map.c if (memcmp((*a)->key, (*b)->key, (*a)->elt->map->key_size)) b 849 kernel/trace/tracing_map.c const struct tracing_map_sort_entry **b) b 859 kernel/trace/tracing_map.c elt_b = (*b)->elt; b 877 kernel/trace/tracing_map.c const struct tracing_map_sort_entry **b) b 887 kernel/trace/tracing_map.c elt_b = (*b)->elt; b 1010 kernel/trace/tracing_map.c const struct tracing_map_sort_entry **b = &entries[i + 1]; b 1012 kernel/trace/tracing_map.c if (primary_fn(a, b) == 0) { b 788 kernel/user_namespace.c static int cmp_extents_forward(const void *a, const void *b) b 791 kernel/user_namespace.c const struct uid_gid_extent *e2 = b; b 803 kernel/user_namespace.c static int cmp_extents_reverse(const void *a, const void *b) b 806 kernel/user_namespace.c const struct uid_gid_extent *e2 = b; b 3400 kernel/workqueue.c const struct workqueue_attrs *b) b 3402 kernel/workqueue.c if (a->nice != b->nice) b 3404 kernel/workqueue.c if (!cpumask_equal(a->cpumask, b->cpumask)) b 109 lib/842/842_compress.c #define get_input_data(p, o, b) \ b 110 lib/842/842_compress.c be##b##_to_cpu(get_unaligned((__be##b *)((p)->in + (o)))) b 112 lib/842/842_compress.c #define init_hashtable_nodes(p, b) do { \ b 114 lib/842/842_compress.c hash_init((p)->htable##b); \ b 115 lib/842/842_compress.c for (_i = 0; _i < ARRAY_SIZE((p)->node##b); _i++) { \ b 116 lib/842/842_compress.c (p)->node##b[_i].index = _i; \ b 117 lib/842/842_compress.c (p)->node##b[_i].data = 0; \ b 118 lib/842/842_compress.c INIT_HLIST_NODE(&(p)->node##b[_i].node); \ b 122 lib/842/842_compress.c #define find_index(p, b, n) ({ \ b 123 lib/842/842_compress.c struct sw842_hlist_node##b *_n; \ b 124 lib/842/842_compress.c p->index##b[n] = INDEX_NOT_FOUND; \ b 125 lib/842/842_compress.c hash_for_each_possible(p->htable##b, _n, node, p->data##b[n]) { \ b 126 lib/842/842_compress.c if (p->data##b[n] == _n->data) { \ b 127 lib/842/842_compress.c p->index##b[n] = _n->index; \ b 131 lib/842/842_compress.c p->index##b[n] >= 0; \ b 134 lib/842/842_compress.c #define check_index(p, b, n) \ b 135 lib/842/842_compress.c ((p)->index##b[n] == INDEX_NOT_CHECKED \ b 136 lib/842/842_compress.c ? find_index(p, b, n) \ b 137 lib/842/842_compress.c : (p)->index##b[n] >= 0) b 139 lib/842/842_compress.c #define replace_hash(p, b, i, d) do { \ b 140 lib/842/842_compress.c struct sw842_hlist_node##b *_n = &(p)->node##b[(i)+(d)]; \ b 142 lib/842/842_compress.c _n->data = (p)->data##b[d]; \ b 143 lib/842/842_compress.c pr_debug("add hash index%x %x pos %x data %lx\n", b, \ b 147 lib/842/842_compress.c hash_add((p)->htable##b, &_n->node, _n->data); \ b 169 lib/842/842_compress.c int b = p->bit, bits = b + n, s = round_up(bits, 8) - bits; b 191 lib/842/842_compress.c o = *out & bmask[b]; b 224 lib/842/842_compress.c int ret, i, b = 0; b 242 lib/842/842_compress.c if (b) b 252 lib/842/842_compress.c if (b == 2 && t[i] & OP_ACTION_DATA) b 254 lib/842/842_compress.c else if (b != 0 && b != 4) b 257 lib/842/842_compress.c ret = add_bits(p, p->index4[b >> 2], I4_BITS); b 259 lib/842/842_compress.c ret = add_bits(p, p->data4[b >> 2], 32); b 264 lib/842/842_compress.c if (b != 0 && b != 2 && b != 4 && b != 6) b 267 lib/842/842_compress.c ret = add_bits(p, p->index2[b >> 1], I2_BITS); b 269 lib/842/842_compress.c ret = add_bits(p, p->data2[b >> 1], 16); b 274 lib/842/842_compress.c inv = (b != 8) || !(t[i] & OP_ACTION_NOOP); b 290 lib/842/842_compress.c b += t[i] & OP_AMOUNT; b 293 lib/842/842_compress.c if (b != 8) { b 295 lib/842/842_compress.c c, b, t[0], t[1], t[2], t[3]); b 327 lib/842/842_compress.c static int add_short_data_template(struct sw842_param *p, u8 b) b 331 lib/842/842_compress.c if (!b || b > SHORT_DATA_BITS_MAX) b 338 lib/842/842_compress.c ret = add_bits(p, b, SHORT_DATA_BITS); b 342 lib/842/842_compress.c for (i = 0; i < b; i++) { b 383 lib/842/842_compress.c int i, match, b = 0; b 391 lib/842/842_compress.c match = check_index(p, 2, b >> 1); b 393 lib/842/842_compress.c match = check_index(p, 4, b >> 2); b 402 lib/842/842_compress.c b += t[i] & OP_AMOUNT; b 89 lib/842/842_decompress.c u8 *in = p->in, b = p->bit, bits = b + n; b 9 lib/ashldi3.c long long notrace __ashldi3(long long u, word_type b) b 14 lib/ashldi3.c if (b == 0) b 18 lib/ashldi3.c bm = 32 - b; b 26 lib/ashldi3.c w.s.low = (unsigned int) uu.s.low << b; b 27 lib/ashldi3.c w.s.high = ((unsigned int) uu.s.high << b) | carries; b 9 lib/ashrdi3.c long long notrace __ashrdi3(long long u, word_type b) b 14 lib/ashrdi3.c if (b == 0) b 18 lib/ashrdi3.c bm = 32 - b; b 28 lib/ashrdi3.c w.s.high = uu.s.high >> b; b 29 lib/ashrdi3.c w.s.low = ((unsigned int) uu.s.low >> b) | carries; b 306 lib/bch.c unsigned int b) b 308 lib/bch.c return (a && b) ? bch->a_pow_tab[mod_s(bch, bch->a_log_tab[a]+ b 309 lib/bch.c bch->a_log_tab[b])] : 0; b 318 lib/bch.c unsigned int b) b 321 lib/bch.c GF_N(bch)-bch->a_log_tab[b])] : 0; b 519 lib/bch.c unsigned int b, unsigned int c, b 526 lib/bch.c j = a_log(bch, b); b 534 lib/bch.c (b ? bch->a_pow_tab[mod_s(bch, j)] : 0); b 616 lib/bch.c unsigned int a, b, c, a2, b2, c2, e3, tmp[4]; b 627 lib/bch.c b = gf_mul(bch, a2, b2)^c2; /* b = a2b2 + c2 */ b 631 lib/bch.c if (find_affine4_roots(bch, a, b, c, tmp) == 4) { b 649 lib/bch.c unsigned int a, b, c, d, e = 0, f, a2, b2, c2, e4; b 658 lib/bch.c b = gf_div(bch, poly->c[2], e4); b 677 lib/bch.c d = a_pow(bch, 2*l)^gf_mul(bch, b, f)^d; b 678 lib/bch.c b = gf_mul(bch, a, e)^b; b 687 lib/bch.c a2 = gf_div(bch, b, d); b 692 lib/bch.c a2 = b; b 723 lib/bch.c const struct gf_poly *b, int *rep) b 727 lib/bch.c const unsigned int d = b->deg; b 735 lib/bch.c gf_poly_logrep(bch, b, rep); b 759 lib/bch.c const struct gf_poly *b, struct gf_poly *q) b 761 lib/bch.c if (a->deg >= b->deg) { b 762 lib/bch.c q->deg = a->deg-b->deg; b 764 lib/bch.c gf_poly_mod(bch, a, b, NULL); b 766 lib/bch.c memcpy(q->c, &a->c[b->deg], (1+q->deg)*sizeof(unsigned int)); b 777 lib/bch.c struct gf_poly *b) b 781 lib/bch.c dbg("gcd(%s,%s)=", gf_poly_str(a), gf_poly_str(b)); b 783 lib/bch.c if (a->deg < b->deg) { b 784 lib/bch.c tmp = b; b 785 lib/bch.c b = a; b 789 lib/bch.c while (b->deg > 0) { b 790 lib/bch.c gf_poly_mod(bch, a, b, NULL); b 791 lib/bch.c tmp = b; b 792 lib/bch.c b = a; b 1091 lib/bch.c int i, j, b, d; b 1101 lib/bch.c for (b = 0; b < 4; b++) { b 1103 lib/bch.c tab = bch->mod8_tab + (b*256+i)*l; b 1104 lib/bch.c data = i << (8*b); b 46 lib/btree.c #define MAX(a, b) ((a) > (b) ? (a) : (b)) b 9 lib/cmpdi2.c word_type notrace __cmpdi2(long long a, long long b) b 15 lib/cmpdi2.c .ll = b b 73 lib/crc32.c const u32 *b; b 99 lib/crc32.c b = (const u32 *)buf; b 101 lib/crc32.c --b; b 104 lib/crc32.c for (--b; len; --len) { b 106 lib/crc32.c q = crc ^ *++b; /* use pre increment for speed */ b 111 lib/crc32.c q = *++b; b 118 lib/crc32.c u8 *p = (u8 *)(b + 1) - 1; b 40 lib/crypto/arc4.c u32 x, y, a, b; b 51 lib/crypto/arc4.c b = S[y]; b 55 lib/crypto/arc4.c a = (a + b) & 0xff; b 56 lib/crypto/arc4.c S[x] = b; b 66 lib/crypto/arc4.c b = tb; b 609 lib/crypto/des.c #define DES_PC2(a, b, c, d) (T4(d) | T3(c) | T2(b) | T1(a)) b 626 lib/crypto/des.c unsigned long a, b, c, d, w; b 631 lib/crypto/des.c b = k[6]; b &= 0x0e; b <<= 4; b |= k[2] & 0x1e; b = pc1[b]; b 634 lib/crypto/des.c pe[15 * 2 + 0] = DES_PC2(a, b, c, d); d = rs[d]; b 635 lib/crypto/des.c pe[14 * 2 + 0] = DES_PC2(d, a, b, c); c = rs[c]; b = rs[b]; b 636 lib/crypto/des.c pe[13 * 2 + 0] = DES_PC2(b, c, d, a); a = rs[a]; d = rs[d]; b 637 lib/crypto/des.c pe[12 * 2 + 0] = DES_PC2(d, a, b, c); c = rs[c]; b = rs[b]; b 638 lib/crypto/des.c pe[11 * 2 + 0] = DES_PC2(b, c, d, a); a = rs[a]; d = rs[d]; b 639 lib/crypto/des.c pe[10 * 2 + 0] = DES_PC2(d, a, b, c); c = rs[c]; b = rs[b]; b 640 lib/crypto/des.c pe[ 9 * 2 + 0] = DES_PC2(b, c, d, a); a = rs[a]; d = rs[d]; b 641 lib/crypto/des.c pe[ 8 * 2 + 0] = DES_PC2(d, a, b, c); c = rs[c]; b 642 lib/crypto/des.c pe[ 7 * 2 + 0] = DES_PC2(c, d, a, b); b = rs[b]; a = rs[a]; b 643 lib/crypto/des.c pe[ 6 * 2 + 0] = DES_PC2(a, b, c, d); d = rs[d]; c = rs[c]; b 644 lib/crypto/des.c pe[ 5 * 2 + 0] = DES_PC2(c, d, a, b); b = rs[b]; a = rs[a]; b 645 lib/crypto/des.c pe[ 4 * 2 + 0] = DES_PC2(a, b, c, d); d = rs[d]; c = rs[c]; b 646 lib/crypto/des.c pe[ 3 * 2 + 0] = DES_PC2(c, d, a, b); b = rs[b]; a = rs[a]; b 647 lib/crypto/des.c pe[ 2 * 2 + 0] = DES_PC2(a, b, c, d); d = rs[d]; c = rs[c]; b 648 lib/crypto/des.c pe[ 1 * 2 + 0] = DES_PC2(c, d, a, b); b = rs[b]; b 649 lib/crypto/des.c pe[ 0 * 2 + 0] = DES_PC2(b, c, d, a); b 652 lib/crypto/des.c w = (a ^ c) | (b ^ d) | (rs[a] ^ c) | (b ^ rs[d]); b 659 lib/crypto/des.c b = k[2]; b &= 0xe0; b >>= 4; b |= k[6] & 0xf0; b = pc1[b + 1]; b 663 lib/crypto/des.c w |= (a ^ c) | (b ^ d) | (rs[a] ^ c) | (b ^ rs[d]); b 665 lib/crypto/des.c pe[15 * 2 + 1] = DES_PC2(a, b, c, d); d = rs[d]; b 666 lib/crypto/des.c pe[14 * 2 + 1] = DES_PC2(d, a, b, c); c = rs[c]; b = rs[b]; b 667 lib/crypto/des.c pe[13 * 2 + 1] = DES_PC2(b, c, d, a); a = rs[a]; d = rs[d]; b 668 lib/crypto/des.c pe[12 * 2 + 1] = DES_PC2(d, a, b, c); c = rs[c]; b = rs[b]; b 669 lib/crypto/des.c pe[11 * 2 + 1] = DES_PC2(b, c, d, a); a = rs[a]; d = rs[d]; b 670 lib/crypto/des.c pe[10 * 2 + 1] = DES_PC2(d, a, b, c); c = rs[c]; b = rs[b]; b 671 lib/crypto/des.c pe[ 9 * 2 + 1] = DES_PC2(b, c, d, a); a = rs[a]; d = rs[d]; b 672 lib/crypto/des.c pe[ 8 * 2 + 1] = DES_PC2(d, a, b, c); c = rs[c]; b 673 lib/crypto/des.c pe[ 7 * 2 + 1] = DES_PC2(c, d, a, b); b = rs[b]; a = rs[a]; b 674 lib/crypto/des.c pe[ 6 * 2 + 1] = DES_PC2(a, b, c, d); d = rs[d]; c = rs[c]; b 675 lib/crypto/des.c pe[ 5 * 2 + 1] = DES_PC2(c, d, a, b); b = rs[b]; a = rs[a]; b 676 lib/crypto/des.c pe[ 4 * 2 + 1] = DES_PC2(a, b, c, d); d = rs[d]; c = rs[c]; b 677 lib/crypto/des.c pe[ 3 * 2 + 1] = DES_PC2(c, d, a, b); b = rs[b]; a = rs[a]; b 678 lib/crypto/des.c pe[ 2 * 2 + 1] = DES_PC2(a, b, c, d); d = rs[d]; c = rs[c]; b 679 lib/crypto/des.c pe[ 1 * 2 + 1] = DES_PC2(c, d, a, b); b = rs[b]; b 680 lib/crypto/des.c pe[ 0 * 2 + 1] = DES_PC2(b, c, d, a); b 685 lib/crypto/des.c b = pe[2 * d + 1]; b 686 lib/crypto/des.c c = a ^ b; b 689 lib/crypto/des.c b ^= c; b 690 lib/crypto/des.c ROL(b, 18); b 692 lib/crypto/des.c pe[2 * d + 1] = b; b 717 lib/crypto/des.c unsigned long a, b, c, d; b 722 lib/crypto/des.c b = k[6]; b &= 0x0e; b <<= 4; b |= k[2] & 0x1e; b = pc1[b]; b 725 lib/crypto/des.c pe[ 0 * 2] = DES_PC2(a, b, c, d); d = rs[d]; b 726 lib/crypto/des.c pe[ 1 * 2] = DES_PC2(d, a, b, c); c = rs[c]; b = rs[b]; b 727 lib/crypto/des.c pe[ 2 * 2] = DES_PC2(b, c, d, a); a = rs[a]; d = rs[d]; b 728 lib/crypto/des.c pe[ 3 * 2] = DES_PC2(d, a, b, c); c = rs[c]; b = rs[b]; b 729 lib/crypto/des.c pe[ 4 * 2] = DES_PC2(b, c, d, a); a = rs[a]; d = rs[d]; b 730 lib/crypto/des.c pe[ 5 * 2] = DES_PC2(d, a, b, c); c = rs[c]; b = rs[b]; b 731 lib/crypto/des.c pe[ 6 * 2] = DES_PC2(b, c, d, a); a = rs[a]; d = rs[d]; b 732 lib/crypto/des.c pe[ 7 * 2] = DES_PC2(d, a, b, c); c = rs[c]; b 733 lib/crypto/des.c pe[ 8 * 2] = DES_PC2(c, d, a, b); b = rs[b]; a = rs[a]; b 734 lib/crypto/des.c pe[ 9 * 2] = DES_PC2(a, b, c, d); d = rs[d]; c = rs[c]; b 735 lib/crypto/des.c pe[10 * 2] = DES_PC2(c, d, a, b); b = rs[b]; a = rs[a]; b 736 lib/crypto/des.c pe[11 * 2] = DES_PC2(a, b, c, d); d = rs[d]; c = rs[c]; b 737 lib/crypto/des.c pe[12 * 2] = DES_PC2(c, d, a, b); b = rs[b]; a = rs[a]; b 738 lib/crypto/des.c pe[13 * 2] = DES_PC2(a, b, c, d); d = rs[d]; c = rs[c]; b 739 lib/crypto/des.c pe[14 * 2] = DES_PC2(c, d, a, b); b = rs[b]; b 740 lib/crypto/des.c pe[15 * 2] = DES_PC2(b, c, d, a); b 747 lib/crypto/des.c b = k[2]; b &= 0xe0; b >>= 4; b |= k[6] & 0xf0; b = pc1[b + 1]; b 750 lib/crypto/des.c pe[ 0 * 2 + 1] = DES_PC2(a, b, c, d); d = rs[d]; b 751 lib/crypto/des.c pe[ 1 * 2 + 1] = DES_PC2(d, a, b, c); c = rs[c]; b = rs[b]; b 752 lib/crypto/des.c pe[ 2 * 2 + 1] = DES_PC2(b, c, d, a); a = rs[a]; d = rs[d]; b 753 lib/crypto/des.c pe[ 3 * 2 + 1] = DES_PC2(d, a, b, c); c = rs[c]; b = rs[b]; b 754 lib/crypto/des.c pe[ 4 * 2 + 1] = DES_PC2(b, c, d, a); a = rs[a]; d = rs[d]; b 755 lib/crypto/des.c pe[ 5 * 2 + 1] = DES_PC2(d, a, b, c); c = rs[c]; b = rs[b]; b 756 lib/crypto/des.c pe[ 6 * 2 + 1] = DES_PC2(b, c, d, a); a = rs[a]; d = rs[d]; b 757 lib/crypto/des.c pe[ 7 * 2 + 1] = DES_PC2(d, a, b, c); c = rs[c]; b 758 lib/crypto/des.c pe[ 8 * 2 + 1] = DES_PC2(c, d, a, b); b = rs[b]; a = rs[a]; b 759 lib/crypto/des.c pe[ 9 * 2 + 1] = DES_PC2(a, b, c, d); d = rs[d]; c = rs[c]; b 760 lib/crypto/des.c pe[10 * 2 + 1] = DES_PC2(c, d, a, b); b = rs[b]; a = rs[a]; b 761 lib/crypto/des.c pe[11 * 2 + 1] = DES_PC2(a, b, c, d); d = rs[d]; c = rs[c]; b 762 lib/crypto/des.c pe[12 * 2 + 1] = DES_PC2(c, d, a, b); b = rs[b]; a = rs[a]; b 763 lib/crypto/des.c pe[13 * 2 + 1] = DES_PC2(a, b, c, d); d = rs[d]; c = rs[c]; b 764 lib/crypto/des.c pe[14 * 2 + 1] = DES_PC2(c, d, a, b); b = rs[b]; b 765 lib/crypto/des.c pe[15 * 2 + 1] = DES_PC2(b, c, d, a); b 770 lib/crypto/des.c b = pe[2 * d + 1]; b 771 lib/crypto/des.c c = a ^ b; b 774 lib/crypto/des.c b ^= c; b 775 lib/crypto/des.c ROL(b, 18); b 777 lib/crypto/des.c pe[2 * d + 1] = b; b 48 lib/crypto/sha256.c u32 a, b, c, d, e, f, g, h, t1, t2; b 61 lib/crypto/sha256.c a = state[0]; b = state[1]; c = state[2]; d = state[3]; b 66 lib/crypto/sha256.c t2 = e0(a) + Maj(a, b, c); d += t1; h = t1 + t2; b 68 lib/crypto/sha256.c t2 = e0(h) + Maj(h, a, b); c += t1; g = t1 + t2; b 70 lib/crypto/sha256.c t2 = e0(g) + Maj(g, h, a); b += t1; f = t1 + t2; b 71 lib/crypto/sha256.c t1 = e + e1(b) + Ch(b, c, d) + 0xe9b5dba5 + W[3]; b 73 lib/crypto/sha256.c t1 = d + e1(a) + Ch(a, b, c) + 0x3956c25b + W[4]; b 75 lib/crypto/sha256.c t1 = c + e1(h) + Ch(h, a, b) + 0x59f111f1 + W[5]; b 77 lib/crypto/sha256.c t1 = b + e1(g) + Ch(g, h, a) + 0x923f82a4 + W[6]; b 78 lib/crypto/sha256.c t2 = e0(c) + Maj(c, d, e); f += t1; b = t1 + t2; b 80 lib/crypto/sha256.c t2 = e0(b) + Maj(b, c, d); e += t1; a = t1 + t2; b 83 lib/crypto/sha256.c t2 = e0(a) + Maj(a, b, c); d += t1; h = t1 + t2; b 85 lib/crypto/sha256.c t2 = e0(h) + Maj(h, a, b); c += t1; g = t1 + t2; b 87 lib/crypto/sha256.c t2 = e0(g) + Maj(g, h, a); b += t1; f = t1 + t2; b 88 lib/crypto/sha256.c t1 = e + e1(b) + Ch(b, c, d) + 0x550c7dc3 + W[11]; b 90 lib/crypto/sha256.c t1 = d + e1(a) + Ch(a, b, c) + 0x72be5d74 + W[12]; b 92 lib/crypto/sha256.c t1 = c + e1(h) + Ch(h, a, b) + 0x80deb1fe + W[13]; b 94 lib/crypto/sha256.c t1 = b + e1(g) + Ch(g, h, a) + 0x9bdc06a7 + W[14]; b 95 lib/crypto/sha256.c t2 = e0(c) + Maj(c, d, e); f += t1; b = t1 + t2; b 97 lib/crypto/sha256.c t2 = e0(b) + Maj(b, c, d); e += t1; a = t1 + t2; b 100 lib/crypto/sha256.c t2 = e0(a) + Maj(a, b, c); d += t1; h = t1 + t2; b 102 lib/crypto/sha256.c t2 = e0(h) + Maj(h, a, b); c += t1; g = t1 + t2; b 104 lib/crypto/sha256.c t2 = e0(g) + Maj(g, h, a); b += t1; f = t1 + t2; b 105 lib/crypto/sha256.c t1 = e + e1(b) + Ch(b, c, d) + 0x240ca1cc + W[19]; b 107 lib/crypto/sha256.c t1 = d + e1(a) + Ch(a, b, c) + 0x2de92c6f + W[20]; b 109 lib/crypto/sha256.c t1 = c + e1(h) + Ch(h, a, b) + 0x4a7484aa + W[21]; b 111 lib/crypto/sha256.c t1 = b + e1(g) + Ch(g, h, a) + 0x5cb0a9dc + W[22]; b 112 lib/crypto/sha256.c t2 = e0(c) + Maj(c, d, e); f += t1; b = t1 + t2; b 114 lib/crypto/sha256.c t2 = e0(b) + Maj(b, c, d); e += t1; a = t1 + t2; b 117 lib/crypto/sha256.c t2 = e0(a) + Maj(a, b, c); d += t1; h = t1 + t2; b 119 lib/crypto/sha256.c t2 = e0(h) + Maj(h, a, b); c += t1; g = t1 + t2; b 121 lib/crypto/sha256.c t2 = e0(g) + Maj(g, h, a); b += t1; f = t1 + t2; b 122 lib/crypto/sha256.c t1 = e + e1(b) + Ch(b, c, d) + 0xbf597fc7 + W[27]; b 124 lib/crypto/sha256.c t1 = d + e1(a) + Ch(a, b, c) + 0xc6e00bf3 + W[28]; b 126 lib/crypto/sha256.c t1 = c + e1(h) + Ch(h, a, b) + 0xd5a79147 + W[29]; b 128 lib/crypto/sha256.c t1 = b + e1(g) + Ch(g, h, a) + 0x06ca6351 + W[30]; b 129 lib/crypto/sha256.c t2 = e0(c) + Maj(c, d, e); f += t1; b = t1 + t2; b 131 lib/crypto/sha256.c t2 = e0(b) + Maj(b, c, d); e += t1; a = t1 + t2; b 134 lib/crypto/sha256.c t2 = e0(a) + Maj(a, b, c); d += t1; h = t1 + t2; b 136 lib/crypto/sha256.c t2 = e0(h) + Maj(h, a, b); c += t1; g = t1 + t2; b 138 lib/crypto/sha256.c t2 = e0(g) + Maj(g, h, a); b += t1; f = t1 + t2; b 139 lib/crypto/sha256.c t1 = e + e1(b) + Ch(b, c, d) + 0x53380d13 + W[35]; b 141 lib/crypto/sha256.c t1 = d + e1(a) + Ch(a, b, c) + 0x650a7354 + W[36]; b 143 lib/crypto/sha256.c t1 = c + e1(h) + Ch(h, a, b) + 0x766a0abb + W[37]; b 145 lib/crypto/sha256.c t1 = b + e1(g) + Ch(g, h, a) + 0x81c2c92e + W[38]; b 146 lib/crypto/sha256.c t2 = e0(c) + Maj(c, d, e); f += t1; b = t1 + t2; b 148 lib/crypto/sha256.c t2 = e0(b) + Maj(b, c, d); e += t1; a = t1 + t2; b 151 lib/crypto/sha256.c t2 = e0(a) + Maj(a, b, c); d += t1; h = t1 + t2; b 153 lib/crypto/sha256.c t2 = e0(h) + Maj(h, a, b); c += t1; g = t1 + t2; b 155 lib/crypto/sha256.c t2 = e0(g) + Maj(g, h, a); b += t1; f = t1 + t2; b 156 lib/crypto/sha256.c t1 = e + e1(b) + Ch(b, c, d) + 0xc76c51a3 + W[43]; b 158 lib/crypto/sha256.c t1 = d + e1(a) + Ch(a, b, c) + 0xd192e819 + W[44]; b 160 lib/crypto/sha256.c t1 = c + e1(h) + Ch(h, a, b) + 0xd6990624 + W[45]; b 162 lib/crypto/sha256.c t1 = b + e1(g) + Ch(g, h, a) + 0xf40e3585 + W[46]; b 163 lib/crypto/sha256.c t2 = e0(c) + Maj(c, d, e); f += t1; b = t1 + t2; b 165 lib/crypto/sha256.c t2 = e0(b) + Maj(b, c, d); e += t1; a = t1 + t2; b 168 lib/crypto/sha256.c t2 = e0(a) + Maj(a, b, c); d += t1; h = t1 + t2; b 170 lib/crypto/sha256.c t2 = e0(h) + Maj(h, a, b); c += t1; g = t1 + t2; b 172 lib/crypto/sha256.c t2 = e0(g) + Maj(g, h, a); b += t1; f = t1 + t2; b 173 lib/crypto/sha256.c t1 = e + e1(b) + Ch(b, c, d) + 0x34b0bcb5 + W[51]; b 175 lib/crypto/sha256.c t1 = d + e1(a) + Ch(a, b, c) + 0x391c0cb3 + W[52]; b 177 lib/crypto/sha256.c t1 = c + e1(h) + Ch(h, a, b) + 0x4ed8aa4a + W[53]; b 179 lib/crypto/sha256.c t1 = b + e1(g) + Ch(g, h, a) + 0x5b9cca4f + W[54]; b 180 lib/crypto/sha256.c t2 = e0(c) + Maj(c, d, e); f += t1; b = t1 + t2; b 182 lib/crypto/sha256.c t2 = e0(b) + Maj(b, c, d); e += t1; a = t1 + t2; b 185 lib/crypto/sha256.c t2 = e0(a) + Maj(a, b, c); d += t1; h = t1 + t2; b 187 lib/crypto/sha256.c t2 = e0(h) + Maj(h, a, b); c += t1; g = t1 + t2; b 189 lib/crypto/sha256.c t2 = e0(g) + Maj(g, h, a); b += t1; f = t1 + t2; b 190 lib/crypto/sha256.c t1 = e + e1(b) + Ch(b, c, d) + 0x8cc70208 + W[59]; b 192 lib/crypto/sha256.c t1 = d + e1(a) + Ch(a, b, c) + 0x90befffa + W[60]; b 194 lib/crypto/sha256.c t1 = c + e1(h) + Ch(h, a, b) + 0xa4506ceb + W[61]; b 196 lib/crypto/sha256.c t1 = b + e1(g) + Ch(g, h, a) + 0xbef9a3f7 + W[62]; b 197 lib/crypto/sha256.c t2 = e0(c) + Maj(c, d, e); f += t1; b = t1 + t2; b 199 lib/crypto/sha256.c t2 = e0(b) + Maj(b, c, d); e += t1; a = t1 + t2; b 201 lib/crypto/sha256.c state[0] += a; state[1] += b; state[2] += c; state[3] += d; b 205 lib/crypto/sha256.c a = b = c = d = e = f = g = h = t1 = t2 = 0; b 190 lib/debugobjects.c static struct debug_obj *lookup_object(void *addr, struct debug_bucket *b) b 195 lib/debugobjects.c hlist_for_each_entry(obj, &b->list, node) { b 226 lib/debugobjects.c alloc_object(void *addr, struct debug_bucket *b, struct debug_obj_descr *descr) b 280 lib/debugobjects.c hlist_add_head(&obj->node, &b->list); b 40 lib/decompress_unlzma.c #define MIN(a, b) (((a) < (b)) ? (a) : (b)) b 177 lib/decompress_unxz.c static bool memeq(const void *a, const void *b, size_t size) b 180 lib/decompress_unxz.c const uint8_t *y = b; b 194 lib/decompress_unxz.c uint8_t *b = buf; b 195 lib/decompress_unxz.c uint8_t *e = b + size; b 197 lib/decompress_unxz.c while (b != e) b 198 lib/decompress_unxz.c *b++ = '\0'; b 257 lib/decompress_unxz.c struct xz_buf b; b 278 lib/decompress_unxz.c b.out = out; b 279 lib/decompress_unxz.c b.out_size = (size_t)-1; b 281 lib/decompress_unxz.c b.out_size = XZ_IOBUF_SIZE; b 282 lib/decompress_unxz.c b.out = malloc(XZ_IOBUF_SIZE); b 283 lib/decompress_unxz.c if (b.out == NULL) b 294 lib/decompress_unxz.c b.in = in; b 295 lib/decompress_unxz.c b.in_pos = 0; b 296 lib/decompress_unxz.c b.in_size = in_size; b 297 lib/decompress_unxz.c b.out_pos = 0; b 300 lib/decompress_unxz.c ret = xz_dec_run(s, &b); b 303 lib/decompress_unxz.c if (b.in_pos == b.in_size && fill != NULL) { b 305 lib/decompress_unxz.c *in_used += b.in_pos; b 307 lib/decompress_unxz.c b.in_pos = 0; b 320 lib/decompress_unxz.c b.in_size = in_size; b 323 lib/decompress_unxz.c ret = xz_dec_run(s, &b); b 325 lib/decompress_unxz.c if (flush != NULL && (b.out_pos == b.out_size b 326 lib/decompress_unxz.c || (ret != XZ_OK && b.out_pos > 0))) { b 332 lib/decompress_unxz.c if (flush(b.out, b.out_pos) != (long)b.out_pos) b 335 lib/decompress_unxz.c b.out_pos = 0; b 343 lib/decompress_unxz.c free(b.out); b 347 lib/decompress_unxz.c *in_used += b.in_pos; b 383 lib/decompress_unxz.c free(b.out); b 28 lib/extable.c static void swap_ex(void *a, void *b, int size) b 30 lib/extable.c struct exception_table_entry *x = a, *y = b, tmp; b 31 lib/extable.c int delta = b - a; b 52 lib/extable.c static int cmp_ex_sort(const void *a, const void *b) b 54 lib/extable.c const struct exception_table_entry *x = a, *y = b; b 81 lib/glob.c unsigned char b = a; b 87 lib/glob.c b = class[1]; b 89 lib/glob.c if (b == '\0') b 95 lib/glob.c match |= (a <= c && c <= b); b 141 lib/inflate.c uch b; /* number of bits in this code or subcode */ b 234 lib/inflate.c #define NEEDBITS(n) {while(k<(n)){b|=((ulg)NEXTBYTE())<<k;k+=8;}} b 235 lib/inflate.c #define DUMPBITS(n) {b>>=(n);k-=(n);} b 323 lib/inflate.c unsigned *b, /* code lengths in bits (all assumed <= BMAX) */ b 375 lib/inflate.c p = b; i = n; b 434 lib/inflate.c p = b; i = 0; b 507 lib/inflate.c r.b = (uch)l; /* bits to dump before this table */ b 518 lib/inflate.c r.b = (uch)(k - w); b 603 lib/inflate.c register ulg b; /* bit buffer */ b 608 lib/inflate.c b = bb; /* initialize bit buffer */ b 618 lib/inflate.c if ((e = (t = tl + ((unsigned)b & ml))->e) > 16) b 622 lib/inflate.c DUMPBITS(t->b) b 625 lib/inflate.c } while ((e = (t = t->v.t + ((unsigned)b & mask_bits[e]))->e) > 16); b 626 lib/inflate.c DUMPBITS(t->b) b 645 lib/inflate.c n = t->v.n + ((unsigned)b & mask_bits[e]); b 650 lib/inflate.c if ((e = (t = td + ((unsigned)b & md))->e) > 16) b 654 lib/inflate.c DUMPBITS(t->b) b 657 lib/inflate.c } while ((e = (t = t->v.t + ((unsigned)b & mask_bits[e]))->e) > 16); b 658 lib/inflate.c DUMPBITS(t->b) b 660 lib/inflate.c d = w - t->v.n - ((unsigned)b & mask_bits[e]); b 692 lib/inflate.c bb = b; /* restore global bit buffer */ b 709 lib/inflate.c register ulg b; /* bit buffer */ b 715 lib/inflate.c b = bb; /* initialize bit buffer */ b 727 lib/inflate.c n = ((unsigned)b & 0xffff); b 730 lib/inflate.c if (n != (unsigned)((~b) & 0xffff)) b 739 lib/inflate.c slide[w++] = (uch)b; b 751 lib/inflate.c bb = b; /* restore global bit buffer */ b 845 lib/inflate.c register ulg b; /* bit buffer */ b 861 lib/inflate.c b = bb; b 867 lib/inflate.c nl = 257 + ((unsigned)b & 0x1f); /* number of literal/length codes */ b 870 lib/inflate.c nd = 1 + ((unsigned)b & 0x1f); /* number of distance codes */ b 873 lib/inflate.c nb = 4 + ((unsigned)b & 0xf); /* number of bit length codes */ b 891 lib/inflate.c ll[border[j]] = (unsigned)b & 7; b 918 lib/inflate.c j = (td = tl + ((unsigned)b & m))->b; b 926 lib/inflate.c j = 3 + ((unsigned)b & 3); b 938 lib/inflate.c j = 3 + ((unsigned)b & 7); b 951 lib/inflate.c j = 11 + ((unsigned)b & 0x7f); b 971 lib/inflate.c bb = b; b 1040 lib/inflate.c register ulg b; /* bit buffer */ b 1046 lib/inflate.c b = bb; b 1052 lib/inflate.c *e = (int)b & 1; b 1058 lib/inflate.c t = (unsigned)b & 3; b 1063 lib/inflate.c bb = b; b 46 lib/interval_tree_test.c u32 b = (prandom_u32_state(&rnd) >> 4) % max_endpoint; b 47 lib/interval_tree_test.c u32 a = (prandom_u32_state(&rnd) >> 4) % b; b 50 lib/interval_tree_test.c nodes[i].last = b; b 20 lib/list_sort.c struct list_head *a, struct list_head *b) b 26 lib/list_sort.c if (cmp(priv, a, b) <= 0) { b 31 lib/list_sort.c *tail = b; b 35 lib/list_sort.c *tail = b; b 36 lib/list_sort.c tail = &b->next; b 37 lib/list_sort.c b = b->next; b 38 lib/list_sort.c if (!b) { b 56 lib/list_sort.c struct list_head *a, struct list_head *b) b 63 lib/list_sort.c if (cmp(priv, a, b) <= 0) { b 71 lib/list_sort.c tail->next = b; b 72 lib/list_sort.c b->prev = tail; b 73 lib/list_sort.c tail = b; b 74 lib/list_sort.c b = b->next; b 75 lib/list_sort.c if (!b) { b 76 lib/list_sort.c b = a; b 83 lib/list_sort.c tail->next = b; b 93 lib/list_sort.c cmp(priv, b, b); b 94 lib/list_sort.c b->prev = tail; b 95 lib/list_sort.c tail = b; b 96 lib/list_sort.c b = b->next; b 97 lib/list_sort.c } while (b); b 190 lib/list_sort.c struct list_head *b)) b 228 lib/list_sort.c struct list_head *a = *tail, *b = a->prev; b 230 lib/list_sort.c a = merge(priv, (cmp_func)cmp, b, a); b 232 lib/list_sort.c a->prev = b->prev; b 23 lib/logic_pio.c #define in_range(b, first, len) ((b) >= (first) && (b) < (first) + (len)) b 297 lib/logic_pio.c BUILD_LOGIC_IO(b, u8) b 9 lib/lshrdi3.c long long notrace __lshrdi3(long long u, word_type b) b 14 lib/lshrdi3.c if (b == 0) b 18 lib/lshrdi3.c bm = 32 - b; b 26 lib/lshrdi3.c w.s.high = (unsigned int) uu.s.high >> b; b 27 lib/lshrdi3.c w.s.low = ((unsigned int) uu.s.low >> b) | carries; b 148 lib/lz4/lz4defs.h U32 b = get_unaligned((const U32 *)src + 1); b 151 lib/lz4/lz4defs.h put_unaligned(b, (U32 *)dst + 1); b 32 lib/math/div64.c uint64_t b = base; b 44 lib/math/div64.c while ((int64_t)b > 0 && b < rem) { b 45 lib/math/div64.c b = b+b; b 50 lib/math/div64.c if (rem >= b) { b 51 lib/math/div64.c rem -= b; b 54 lib/math/div64.c b >>= 1; b 23 lib/math/gcd.c unsigned long gcd(unsigned long a, unsigned long b) b 25 lib/math/gcd.c unsigned long r = a | b; b 27 lib/math/gcd.c if (!a || !b) b 30 lib/math/gcd.c b >>= __ffs(b); b 31 lib/math/gcd.c if (b == 1) b 38 lib/math/gcd.c if (a == b) b 41 lib/math/gcd.c if (a < b) b 42 lib/math/gcd.c swap(a, b); b 43 lib/math/gcd.c a -= b; b 50 lib/math/gcd.c unsigned long gcd(unsigned long a, unsigned long b) b 52 lib/math/gcd.c unsigned long r = a | b; b 54 lib/math/gcd.c if (!a || !b) b 60 lib/math/gcd.c while (!(b & r)) b 61 lib/math/gcd.c b >>= 1; b 62 lib/math/gcd.c if (b == r) b 70 lib/math/gcd.c if (a == b) b 73 lib/math/gcd.c if (a < b) b 74 lib/math/gcd.c swap(a, b); b 75 lib/math/gcd.c a -= b; b 78 lib/math/gcd.c a += b; b 21 lib/math/int_sqrt.c unsigned long b, m, y = 0; b 28 lib/math/int_sqrt.c b = y + m; b 31 lib/math/int_sqrt.c if (x >= b) { b 32 lib/math/int_sqrt.c x -= b; b 50 lib/math/int_sqrt.c u64 b, m, y = 0; b 57 lib/math/int_sqrt.c b = y + m; b 60 lib/math/int_sqrt.c if (x >= b) { b 61 lib/math/int_sqrt.c x -= b; b 8 lib/math/lcm.c unsigned long lcm(unsigned long a, unsigned long b) b 10 lib/math/lcm.c if (a && b) b 11 lib/math/lcm.c return (a / gcd(a, b)) * b; b 17 lib/math/lcm.c unsigned long lcm_not_zero(unsigned long a, unsigned long b) b 19 lib/math/lcm.c unsigned long l = lcm(a, b); b 24 lib/math/lcm.c return (b ? : a); b 10 lib/memcat_p.c void **__memcat_p(void **a, void **b) b 18 lib/memcat_p.c for (p = b; *p; nr++, p++) b 28 lib/memcat_p.c for (nr--; nr >= 0; nr--, p = p == b ? &a[nr] : p - 1) b 195 lib/mpi/longlong.h #define umul_ppmm(xh, xl, a, b) \ b 212 lib/mpi/longlong.h "r" ((USItype)(b)) \ b 215 lib/mpi/longlong.h #define umul_ppmm(xh, xl, a, b) \ b 221 lib/mpi/longlong.h "r" ((USItype)(b)) \ b 557 lib/mpi/longlong.h #define umul_ppmm(xh, xl, a, b) \ b 583 lib/mpi/longlong.h : "%2" ((USItype)(a)), "d" ((USItype)(b))); \ b 573 lib/objagg.c static int objagg_stats_info_sort_cmp_func(const void *a, const void *b) b 576 lib/objagg.c const struct objagg_obj_stats_info *stats_info2 = b; b 76 lib/oid_registry.c unsigned char b = octets[len]; b 77 lib/oid_registry.c if (a > b) { b 81 lib/oid_registry.c if (a < b) { b 21 lib/raid6/mktables.c static uint8_t gfmul(uint8_t a, uint8_t b) b 25 lib/raid6/mktables.c while (b) { b 26 lib/raid6/mktables.c if (b & 1) b 29 lib/raid6/mktables.c b >>= 1; b 35 lib/raid6/mktables.c static uint8_t gfpow(uint8_t a, int b) b 39 lib/raid6/mktables.c b %= 255; b 40 lib/raid6/mktables.c if (b < 0) b 41 lib/raid6/mktables.c b += 255; b 43 lib/raid6/mktables.c while (b) { b 44 lib/raid6/mktables.c if (b & 1) b 47 lib/raid6/mktables.c b >>= 1; b 15 lib/raid6/recov_neon_inner.c static uint8x16_t vqtbl1q_u8(uint8x16_t a, uint8x16_t b) b 22 lib/raid6/recov_neon_inner.c return vcombine_u8(vtbl2_u8(__a.pair, vget_low_u8(b)), b 23 lib/raid6/recov_neon_inner.c vtbl2_u8(__a.pair, vget_high_u8(b))); b 62 lib/random32.c #define TAUSWORTHE(s, a, b, c, d) ((s & c) << d) ^ (((s << a) ^ s) >> b) b 34 lib/reed_solomon/decode_rs.c uint16_t *b = rsc->buffers + RS_DECODE_B * (nroots + 1); b 126 lib/reed_solomon/decode_rs.c b[i] = index_of[lambda[i]]; b 148 lib/reed_solomon/decode_rs.c memmove (&b[1], b, nroots * sizeof (b[0])); b 149 lib/reed_solomon/decode_rs.c b[0] = nn; b 154 lib/reed_solomon/decode_rs.c if (b[i] != nn) { b 157 lib/reed_solomon/decode_rs.c b[i])]; b 168 lib/reed_solomon/decode_rs.c b[i] = (lambda[i] == 0) ? nn : b 174 lib/reed_solomon/decode_rs.c memmove(&b[1], b, nroots * sizeof(b[0])); b 175 lib/reed_solomon/decode_rs.c b[0] = nn; b 263 lib/reed_solomon/decode_rs.c b[j] = 0; b 279 lib/reed_solomon/decode_rs.c b[j] = alpha_to[rs_modnn(rs, index_of[num1] + b 292 lib/reed_solomon/decode_rs.c if (b[j] == 0) b 296 lib/reed_solomon/decode_rs.c tmp ^= alpha_to[rs_modnn(rs, index_of[b[j]] + k)]; b 310 lib/reed_solomon/decode_rs.c if (b[i]) { b 311 lib/reed_solomon/decode_rs.c corr[j] = b[i]; b 319 lib/reed_solomon/decode_rs.c data[loc[i] - pad] ^= b[i]; b 321 lib/reed_solomon/decode_rs.c par[loc[i] - pad - len] ^= b[i]; b 34 lib/siphash.c u64 b = ((u64)(len)) << 56; \ b 41 lib/siphash.c v3 ^= b; \ b 44 lib/siphash.c v0 ^= b; \ b 67 lib/siphash.c b |= le64_to_cpu((__force __le64)(load_unaligned_zeropad(data) & b 71 lib/siphash.c case 7: b |= ((u64)end[6]) << 48; /* fall through */ b 72 lib/siphash.c case 6: b |= ((u64)end[5]) << 40; /* fall through */ b 73 lib/siphash.c case 5: b |= ((u64)end[4]) << 32; /* fall through */ b 74 lib/siphash.c case 4: b |= le32_to_cpup(data); break; b 75 lib/siphash.c case 3: b |= ((u64)end[2]) << 16; /* fall through */ b 76 lib/siphash.c case 2: b |= le16_to_cpup(data); break; b 77 lib/siphash.c case 1: b |= end[0]; b 100 lib/siphash.c b |= le64_to_cpu((__force __le64)(load_unaligned_zeropad(data) & b 104 lib/siphash.c case 7: b |= ((u64)end[6]) << 48; /* fall through */ b 105 lib/siphash.c case 6: b |= ((u64)end[5]) << 40; /* fall through */ b 106 lib/siphash.c case 5: b |= ((u64)end[4]) << 32; /* fall through */ b 107 lib/siphash.c case 4: b |= get_unaligned_le32(end); break; b 108 lib/siphash.c case 3: b |= ((u64)end[2]) << 16; /* fall through */ b 109 lib/siphash.c case 2: b |= get_unaligned_le16(end); break; b 110 lib/siphash.c case 1: b |= end[0]; b 217 lib/siphash.c b |= first; b 231 lib/siphash.c b |= third; b 244 lib/siphash.c v3 ^= b; \ b 246 lib/siphash.c v0 ^= b; \ b 267 lib/siphash.c b |= le64_to_cpu((__force __le64)(load_unaligned_zeropad(data) & b 271 lib/siphash.c case 7: b |= ((u64)end[6]) << 48; /* fall through */ b 272 lib/siphash.c case 6: b |= ((u64)end[5]) << 40; /* fall through */ b 273 lib/siphash.c case 5: b |= ((u64)end[4]) << 32; /* fall through */ b 274 lib/siphash.c case 4: b |= le32_to_cpup(data); break; b 275 lib/siphash.c case 3: b |= ((u64)end[2]) << 16; /* fall through */ b 276 lib/siphash.c case 2: b |= le16_to_cpup(data); break; b 277 lib/siphash.c case 1: b |= end[0]; b 300 lib/siphash.c b |= le64_to_cpu((__force __le64)(load_unaligned_zeropad(data) & b 304 lib/siphash.c case 7: b |= ((u64)end[6]) << 48; /* fall through */ b 305 lib/siphash.c case 6: b |= ((u64)end[5]) << 40; /* fall through */ b 306 lib/siphash.c case 5: b |= ((u64)end[4]) << 32; /* fall through */ b 307 lib/siphash.c case 4: b |= get_unaligned_le32(end); break; b 308 lib/siphash.c case 3: b |= ((u64)end[2]) << 16; /* fall through */ b 309 lib/siphash.c case 2: b |= get_unaligned_le16(end); break; b 310 lib/siphash.c case 1: b |= end[0]; b 326 lib/siphash.c b |= first; b 363 lib/siphash.c b |= third; b 405 lib/siphash.c u32 b = ((u32)(len)) << 24; \ b 412 lib/siphash.c v3 ^= b; \ b 414 lib/siphash.c v0 ^= b; \ b 434 lib/siphash.c case 3: b |= ((u32)end[2]) << 16; /* fall through */ b 435 lib/siphash.c case 2: b |= le16_to_cpup(data); break; b 436 lib/siphash.c case 1: b |= end[0]; b 457 lib/siphash.c case 3: b |= ((u32)end[2]) << 16; /* fall through */ b 458 lib/siphash.c case 2: b |= get_unaligned_le16(end); break; b 459 lib/siphash.c case 1: b |= end[0]; b 58 lib/sort.c static void swap_words_32(void *a, void *b, size_t n) b 62 lib/sort.c *(u32 *)(a + n) = *(u32 *)(b + n); b 63 lib/sort.c *(u32 *)(b + n) = t; b 83 lib/sort.c static void swap_words_64(void *a, void *b, size_t n) b 88 lib/sort.c *(u64 *)(a + n) = *(u64 *)(b + n); b 89 lib/sort.c *(u64 *)(b + n) = t; b 93 lib/sort.c *(u32 *)(a + n) = *(u32 *)(b + n); b 94 lib/sort.c *(u32 *)(b + n) = t; b 97 lib/sort.c *(u32 *)(a + n) = *(u32 *)(b + n); b 98 lib/sort.c *(u32 *)(b + n) = t; b 111 lib/sort.c static void swap_bytes(void *a, void *b, size_t n) b 115 lib/sort.c ((char *)a)[n] = ((char *)b)[n]; b 116 lib/sort.c ((char *)b)[n] = t; b 120 lib/sort.c typedef void (*swap_func_t)(void *a, void *b, int size); b 135 lib/sort.c static void do_swap(void *a, void *b, size_t size, swap_func_t swap_func) b 138 lib/sort.c swap_words_64(a, b, size); b 140 lib/sort.c swap_words_32(a, b, size); b 142 lib/sort.c swap_bytes(a, b, size); b 144 lib/sort.c swap_func(a, b, (int)size); b 151 lib/sort.c static int do_cmp(const void *a, const void *b, b 155 lib/sort.c return ((cmp_func_t)(priv))(a, b); b 156 lib/sort.c return cmp(a, b, priv); b 233 lib/sort.c size_t b, c, d; b 254 lib/sort.c for (b = a; c = 2*b + size, (d = c + size) < n;) b 255 lib/sort.c b = do_cmp(base + c, base + d, cmp_func, priv) >= 0 ? c : d; b 257 lib/sort.c b = c; b 260 lib/sort.c while (b != a && do_cmp(base + a, base + b, cmp_func, priv) >= 0) b 261 lib/sort.c b = parent(b, lsbit, size); b 262 lib/sort.c c = b; /* Where "a" belongs */ b 263 lib/sort.c while (b != a) { /* Shift it into place */ b 264 lib/sort.c b = parent(b, lsbit, size); b 265 lib/sort.c do_swap(base + b, base + c, size, swap_func); b 905 lib/string.c int bcmp(const void *a, const void *b, size_t len) b 907 lib/string.c return memcmp(a, b, len); b 59 lib/test_list_sort.c static int __init cmp(void *priv, struct list_head *a, struct list_head *b) b 64 lib/test_list_sort.c elb = container_of(b, struct debug_el, list); b 19 lib/test_overflow.c t a, b; \ b 218 lib/test_overflow.c #define check_one_op(t, fmt, op, sym, a, b, r, of) do { \ b 222 lib/test_overflow.c _of = check_ ## op ## _overflow(a, b, &_r); \ b 226 lib/test_overflow.c a, b, of ? "" : " not", #t); \ b 232 lib/test_overflow.c a, b, r, _r, #t); \ b 242 lib/test_overflow.c check_one_op(t, fmt, add, "+", p->a, p->b, p->sum, p->s_of); \ b 243 lib/test_overflow.c check_one_op(t, fmt, add, "+", p->b, p->a, p->sum, p->s_of); \ b 244 lib/test_overflow.c check_one_op(t, fmt, sub, "-", p->a, p->b, p->diff, p->d_of); \ b 245 lib/test_overflow.c check_one_op(t, fmt, mul, "*", p->a, p->b, p->prod, p->p_of); \ b 246 lib/test_overflow.c check_one_op(t, fmt, mul, "*", p->b, p->a, p->prod, p->p_of); \ b 505 lib/test_overflow.c volatile size_t b = (SIZE_MAX / TEST_SIZE) + 1; \ b 518 lib/test_overflow.c a * b); \ b 527 lib/test_overflow.c array_size(a, b)); \ b 10 lib/test_sort.c static int __init cmpint(const void *a, const void *b) b 12 lib/test_sort.c return *(int *)a - *(int *)b; b 8 lib/ucmpdi2.c word_type notrace __ucmpdi2(unsigned long long a, unsigned long long b) b 11 lib/ucmpdi2.c const DWunion bu = {.ll = b}; b 36 lib/ucs2_string.c ucs2_strncmp(const ucs2_char_t *a, const ucs2_char_t *b, size_t len) b 41 lib/ucs2_string.c if (*a < *b) b 43 lib/ucs2_string.c if (*a > *b) b 48 lib/ucs2_string.c b++; b 43 lib/uuid.c static void __uuid_gen_common(__u8 b[16]) b 45 lib/uuid.c prandom_bytes(b, 16); b 47 lib/uuid.c b[8] = (b[8] & 0x3F) | 0x80; b 52 lib/uuid.c __uuid_gen_common(lu->b); b 54 lib/uuid.c lu->b[7] = (lu->b[7] & 0x0F) | 0x40; b 60 lib/uuid.c __uuid_gen_common(bu->b); b 62 lib/uuid.c bu->b[6] = (bu->b[6] & 0x0F) | 0x40; b 95 lib/uuid.c static int __uuid_parse(const char *uuid, __u8 b[16], const u8 ei[16]) b 107 lib/uuid.c b[ei[i]] = (hi << 4) | lo; b 115 lib/uuid.c return __uuid_parse(uuid, u->b, guid_index); b 121 lib/uuid.c return __uuid_parse(uuid, u->b, uuid_index); b 83 lib/xz/xz_dec_bcj.c static inline int bcj_x86_test_msbyte(uint8_t b) b 85 lib/xz/xz_dec_bcj.c return b == 0x00 || b == 0xFF; b 101 lib/xz/xz_dec_bcj.c uint8_t b; b 117 lib/xz/xz_dec_bcj.c b = buf[i + 4 - mask_to_bit_num[prev_mask]]; b 119 lib/xz/xz_dec_bcj.c || bcj_x86_test_msbyte(b)) { b 137 lib/xz/xz_dec_bcj.c b = (uint8_t)(dest >> (24 - j)); b 138 lib/xz/xz_dec_bcj.c if (!bcj_x86_test_msbyte(b)) b 399 lib/xz/xz_dec_bcj.c static void bcj_flush(struct xz_dec_bcj *s, struct xz_buf *b) b 403 lib/xz/xz_dec_bcj.c copy_size = min_t(size_t, s->temp.filtered, b->out_size - b->out_pos); b 404 lib/xz/xz_dec_bcj.c memcpy(b->out + b->out_pos, s->temp.buf, copy_size); b 405 lib/xz/xz_dec_bcj.c b->out_pos += copy_size; b 419 lib/xz/xz_dec_bcj.c struct xz_buf *b) b 429 lib/xz/xz_dec_bcj.c bcj_flush(s, b); b 449 lib/xz/xz_dec_bcj.c if (s->temp.size < b->out_size - b->out_pos || s->temp.size == 0) { b 450 lib/xz/xz_dec_bcj.c out_start = b->out_pos; b 451 lib/xz/xz_dec_bcj.c memcpy(b->out + b->out_pos, s->temp.buf, s->temp.size); b 452 lib/xz/xz_dec_bcj.c b->out_pos += s->temp.size; b 454 lib/xz/xz_dec_bcj.c s->ret = xz_dec_lzma2_run(lzma2, b); b 459 lib/xz/xz_dec_bcj.c bcj_apply(s, b->out, &out_start, b->out_pos); b 469 lib/xz/xz_dec_bcj.c s->temp.size = b->out_pos - out_start; b 470 lib/xz/xz_dec_bcj.c b->out_pos -= s->temp.size; b 471 lib/xz/xz_dec_bcj.c memcpy(s->temp.buf, b->out + b->out_pos, s->temp.size); b 478 lib/xz/xz_dec_bcj.c if (b->out_pos + s->temp.size < b->out_size) b 490 lib/xz/xz_dec_bcj.c if (b->out_pos < b->out_size) { b 492 lib/xz/xz_dec_bcj.c s->out = b->out; b 493 lib/xz/xz_dec_bcj.c s->out_pos = b->out_pos; b 494 lib/xz/xz_dec_bcj.c s->out_size = b->out_size; b 495 lib/xz/xz_dec_bcj.c b->out = s->temp.buf; b 496 lib/xz/xz_dec_bcj.c b->out_pos = s->temp.size; b 497 lib/xz/xz_dec_bcj.c b->out_size = sizeof(s->temp.buf); b 499 lib/xz/xz_dec_bcj.c s->ret = xz_dec_lzma2_run(lzma2, b); b 501 lib/xz/xz_dec_bcj.c s->temp.size = b->out_pos; b 502 lib/xz/xz_dec_bcj.c b->out = s->out; b 503 lib/xz/xz_dec_bcj.c b->out_pos = s->out_pos; b 504 lib/xz/xz_dec_bcj.c b->out_size = s->out_size; b 519 lib/xz/xz_dec_bcj.c bcj_flush(s, b); b 286 lib/xz/xz_dec_lzma2.c static void dict_reset(struct dictionary *dict, struct xz_buf *b) b 289 lib/xz/xz_dec_lzma2.c dict->buf = b->out + b->out_pos; b 290 lib/xz/xz_dec_lzma2.c dict->end = b->out_size - b->out_pos; b 374 lib/xz/xz_dec_lzma2.c static void dict_uncompressed(struct dictionary *dict, struct xz_buf *b, b 379 lib/xz/xz_dec_lzma2.c while (*left > 0 && b->in_pos < b->in_size b 380 lib/xz/xz_dec_lzma2.c && b->out_pos < b->out_size) { b 381 lib/xz/xz_dec_lzma2.c copy_size = min(b->in_size - b->in_pos, b 382 lib/xz/xz_dec_lzma2.c b->out_size - b->out_pos); b 390 lib/xz/xz_dec_lzma2.c memcpy(dict->buf + dict->pos, b->in + b->in_pos, copy_size); b 400 lib/xz/xz_dec_lzma2.c memcpy(b->out + b->out_pos, b->in + b->in_pos, b 406 lib/xz/xz_dec_lzma2.c b->out_pos += copy_size; b 407 lib/xz/xz_dec_lzma2.c b->in_pos += copy_size; b 416 lib/xz/xz_dec_lzma2.c static uint32_t dict_flush(struct dictionary *dict, struct xz_buf *b) b 424 lib/xz/xz_dec_lzma2.c memcpy(b->out + b->out_pos, dict->buf + dict->start, b 429 lib/xz/xz_dec_lzma2.c b->out_pos += copy_size; b 449 lib/xz/xz_dec_lzma2.c static bool rc_read_init(struct rc_dec *rc, struct xz_buf *b) b 452 lib/xz/xz_dec_lzma2.c if (b->in_pos == b->in_size) b 455 lib/xz/xz_dec_lzma2.c rc->code = (rc->code << 8) + b->in[b->in_pos++]; b 846 lib/xz/xz_dec_lzma2.c static bool lzma2_lzma(struct xz_dec_lzma2 *s, struct xz_buf *b) b 851 lib/xz/xz_dec_lzma2.c in_avail = b->in_size - b->in_pos; b 859 lib/xz/xz_dec_lzma2.c memcpy(s->temp.buf + s->temp.size, b->in + b->in_pos, tmp); b 868 lib/xz/xz_dec_lzma2.c b->in_pos += tmp; b 889 lib/xz/xz_dec_lzma2.c b->in_pos += s->rc.in_pos - s->temp.size; b 893 lib/xz/xz_dec_lzma2.c in_avail = b->in_size - b->in_pos; b 895 lib/xz/xz_dec_lzma2.c s->rc.in = b->in; b 896 lib/xz/xz_dec_lzma2.c s->rc.in_pos = b->in_pos; b 899 lib/xz/xz_dec_lzma2.c s->rc.in_limit = b->in_pos + s->lzma2.compressed; b 901 lib/xz/xz_dec_lzma2.c s->rc.in_limit = b->in_size - LZMA_IN_REQUIRED; b 906 lib/xz/xz_dec_lzma2.c in_avail = s->rc.in_pos - b->in_pos; b 911 lib/xz/xz_dec_lzma2.c b->in_pos = s->rc.in_pos; b 914 lib/xz/xz_dec_lzma2.c in_avail = b->in_size - b->in_pos; b 919 lib/xz/xz_dec_lzma2.c memcpy(s->temp.buf, b->in + b->in_pos, in_avail); b 921 lib/xz/xz_dec_lzma2.c b->in_pos += in_avail; b 932 lib/xz/xz_dec_lzma2.c struct xz_buf *b) b 936 lib/xz/xz_dec_lzma2.c while (b->in_pos < b->in_size || s->lzma2.sequence == SEQ_LZMA_RUN) { b 970 lib/xz/xz_dec_lzma2.c tmp = b->in[b->in_pos++]; b 978 lib/xz/xz_dec_lzma2.c dict_reset(&s->dict, b); b 1018 lib/xz/xz_dec_lzma2.c += (uint32_t)b->in[b->in_pos++] << 8; b 1024 lib/xz/xz_dec_lzma2.c += (uint32_t)b->in[b->in_pos++] + 1; b 1030 lib/xz/xz_dec_lzma2.c = (uint32_t)b->in[b->in_pos++] << 8; b 1036 lib/xz/xz_dec_lzma2.c += (uint32_t)b->in[b->in_pos++] + 1; b 1041 lib/xz/xz_dec_lzma2.c if (!lzma_props(s, b->in[b->in_pos++])) b 1052 lib/xz/xz_dec_lzma2.c if (!rc_read_init(&s->rc, b)) b 1071 lib/xz/xz_dec_lzma2.c b->out_size - b->out_pos, b 1073 lib/xz/xz_dec_lzma2.c if (!lzma2_lzma(s, b)) b 1076 lib/xz/xz_dec_lzma2.c s->lzma2.uncompressed -= dict_flush(&s->dict, b); b 1086 lib/xz/xz_dec_lzma2.c } else if (b->out_pos == b->out_size b 1087 lib/xz/xz_dec_lzma2.c || (b->in_pos == b->in_size b 1096 lib/xz/xz_dec_lzma2.c dict_uncompressed(&s->dict, b, &s->lzma2.compressed); b 157 lib/xz/xz_dec_stream.c static bool fill_temp(struct xz_dec *s, struct xz_buf *b) b 160 lib/xz/xz_dec_stream.c b->in_size - b->in_pos, s->temp.size - s->temp.pos); b 162 lib/xz/xz_dec_stream.c memcpy(s->temp.buf + s->temp.pos, b->in + b->in_pos, copy_size); b 163 lib/xz/xz_dec_stream.c b->in_pos += copy_size; b 218 lib/xz/xz_dec_stream.c static enum xz_ret dec_block(struct xz_dec *s, struct xz_buf *b) b 222 lib/xz/xz_dec_stream.c s->in_start = b->in_pos; b 223 lib/xz/xz_dec_stream.c s->out_start = b->out_pos; b 227 lib/xz/xz_dec_stream.c ret = xz_dec_bcj_run(s->bcj, s->lzma2, b); b 230 lib/xz/xz_dec_stream.c ret = xz_dec_lzma2_run(s->lzma2, b); b 232 lib/xz/xz_dec_stream.c s->block.compressed += b->in_pos - s->in_start; b 233 lib/xz/xz_dec_stream.c s->block.uncompressed += b->out_pos - s->out_start; b 245 lib/xz/xz_dec_stream.c s->crc32 = xz_crc32(b->out + s->out_start, b 246 lib/xz/xz_dec_stream.c b->out_pos - s->out_start, s->crc32); b 281 lib/xz/xz_dec_stream.c static void index_update(struct xz_dec *s, const struct xz_buf *b) b 283 lib/xz/xz_dec_stream.c size_t in_used = b->in_pos - s->in_start; b 285 lib/xz/xz_dec_stream.c s->crc32 = xz_crc32(b->in + s->in_start, in_used, s->crc32); b 296 lib/xz/xz_dec_stream.c static enum xz_ret dec_index(struct xz_dec *s, struct xz_buf *b) b 301 lib/xz/xz_dec_stream.c ret = dec_vli(s, b->in, &b->in_pos, b->in_size); b 303 lib/xz/xz_dec_stream.c index_update(s, b); b 346 lib/xz/xz_dec_stream.c static enum xz_ret crc32_validate(struct xz_dec *s, struct xz_buf *b) b 349 lib/xz/xz_dec_stream.c if (b->in_pos == b->in_size) b 352 lib/xz/xz_dec_stream.c if (((s->crc32 >> s->pos) & 0xFF) != b->in[b->in_pos++]) b 370 lib/xz/xz_dec_stream.c static bool check_skip(struct xz_dec *s, struct xz_buf *b) b 373 lib/xz/xz_dec_stream.c if (b->in_pos == b->in_size) b 376 lib/xz/xz_dec_stream.c ++b->in_pos; b 549 lib/xz/xz_dec_stream.c static enum xz_ret dec_main(struct xz_dec *s, struct xz_buf *b) b 557 lib/xz/xz_dec_stream.c s->in_start = b->in_pos; b 570 lib/xz/xz_dec_stream.c if (!fill_temp(s, b)) b 590 lib/xz/xz_dec_stream.c if (b->in_pos == b->in_size) b 594 lib/xz/xz_dec_stream.c if (b->in[b->in_pos] == 0) { b 595 lib/xz/xz_dec_stream.c s->in_start = b->in_pos++; b 605 lib/xz/xz_dec_stream.c = ((uint32_t)b->in[b->in_pos] + 1) * 4; b 614 lib/xz/xz_dec_stream.c if (!fill_temp(s, b)) b 626 lib/xz/xz_dec_stream.c ret = dec_block(s, b); b 643 lib/xz/xz_dec_stream.c if (b->in_pos == b->in_size) b 646 lib/xz/xz_dec_stream.c if (b->in[b->in_pos++] != 0) b 658 lib/xz/xz_dec_stream.c ret = crc32_validate(s, b); b 663 lib/xz/xz_dec_stream.c else if (!check_skip(s, b)) { b 672 lib/xz/xz_dec_stream.c ret = dec_index(s, b); b 681 lib/xz/xz_dec_stream.c while ((s->index.size + (b->in_pos - s->in_start)) b 683 lib/xz/xz_dec_stream.c if (b->in_pos == b->in_size) { b 684 lib/xz/xz_dec_stream.c index_update(s, b); b 688 lib/xz/xz_dec_stream.c if (b->in[b->in_pos++] != 0) b 693 lib/xz/xz_dec_stream.c index_update(s, b); b 705 lib/xz/xz_dec_stream.c ret = crc32_validate(s, b); b 715 lib/xz/xz_dec_stream.c if (!fill_temp(s, b)) b 750 lib/xz/xz_dec_stream.c XZ_EXTERN enum xz_ret xz_dec_run(struct xz_dec *s, struct xz_buf *b) b 759 lib/xz/xz_dec_stream.c in_start = b->in_pos; b 760 lib/xz/xz_dec_stream.c out_start = b->out_pos; b 761 lib/xz/xz_dec_stream.c ret = dec_main(s, b); b 765 lib/xz/xz_dec_stream.c ret = b->in_pos == b->in_size b 769 lib/xz/xz_dec_stream.c b->in_pos = in_start; b 770 lib/xz/xz_dec_stream.c b->out_pos = out_start; b 773 lib/xz/xz_dec_stream.c } else if (ret == XZ_OK && in_start == b->in_pos b 774 lib/xz/xz_dec_stream.c && out_start == b->out_pos) { b 40 lib/xz/xz_private.h # define memeq(a, b, size) (memcmp(a, b, size) == 0) b 127 lib/xz/xz_private.h struct xz_buf *b); b 154 lib/xz/xz_private.h struct xz_buf *b); b 74 lib/zlib_deflate/deflate.c static void putShortMSB (deflate_state *s, uInt b); b 290 lib/zlib_deflate/deflate.c uInt b b 293 lib/zlib_deflate/deflate.c put_byte(s, (Byte)(b >> 8)); b 294 lib/zlib_deflate/deflate.c put_byte(s, (Byte)(b & 0xff)); b 26 lib/zlib_inflate/inffast.c unsigned char b[2]; b 34 lib/zlib_inflate/inffast.c unsigned char *b = (unsigned char *)p; b 36 lib/zlib_inflate/inffast.c mm.b[0] = b[0]; b 37 lib/zlib_inflate/inffast.c mm.b[1] = b[1]; b 294 lib/zlib_inflate/inffast.c mm.b[0] = mm.b[1]; b 39 lib/zstd/zstd_internal.h #define MIN(a, b) ((a) < (b) ? (a) : (b)) b 40 lib/zstd/zstd_internal.h #define MAX(a, b) ((a) > (b) ? (a) : (b)) b 116 mm/frontswap.c DECLARE_BITMAP(b, MAX_SWAPFILES); b 121 mm/frontswap.c bitmap_zero(b, MAX_SWAPFILES); b 148 mm/frontswap.c set_bit(si->type, b); b 158 mm/frontswap.c if (unlikely(!bitmap_equal(a, b, MAX_SWAPFILES))) { b 160 mm/frontswap.c if (!test_bit(i, a) && test_bit(i, b)) b 162 mm/frontswap.c else if (test_bit(i, a) && !test_bit(i, b)) b 4084 mm/memcontrol.c static int compare_thresholds(const void *a, const void *b) b 4087 mm/memcontrol.c const struct mem_cgroup_threshold *_b = b; b 2246 mm/mempolicy.c bool __mpol_equal(struct mempolicy *a, struct mempolicy *b) b 2248 mm/mempolicy.c if (!a || !b) b 2250 mm/mempolicy.c if (a->mode != b->mode) b 2252 mm/mempolicy.c if (a->flags != b->flags) b 2255 mm/mempolicy.c if (!nodes_equal(a->w.user_nodemask, b->w.user_nodemask)) b 2262 mm/mempolicy.c return !!nodes_equal(a->v.nodes, b->v.nodes); b 2267 mm/mempolicy.c return a->v.preferred_node == b->v.preferred_node; b 1241 mm/mmap.c static int anon_vma_compatible(struct vm_area_struct *a, struct vm_area_struct *b) b 1243 mm/mmap.c return a->vm_end == b->vm_start && b 1244 mm/mmap.c mpol_equal(vma_policy(a), vma_policy(b)) && b 1245 mm/mmap.c a->vm_file == b->vm_file && b 1246 mm/mmap.c !((a->vm_flags ^ b->vm_flags) & ~(VM_READ|VM_WRITE|VM_EXEC|VM_SOFTDIRTY)) && b 1247 mm/mmap.c b->vm_pgoff == a->vm_pgoff + ((b->vm_start - a->vm_start) >> PAGE_SHIFT); b 1272 mm/mmap.c static struct anon_vma *reusable_anon_vma(struct vm_area_struct *old, struct vm_area_struct *a, struct vm_area_struct *b) b 1274 mm/mmap.c if (anon_vma_compatible(a, b)) { b 59 mm/page_poison.c static bool single_bit_flip(unsigned char a, unsigned char b) b 61 mm/page_poison.c unsigned char error = a ^ b; b 25 mm/percpu-stats.c static int cmpint(const void *a, const void *b) b 27 mm/percpu-stats.c return *(int *)a - *(int *)b; b 588 mm/percpu.c static inline bool pcpu_region_overlap(int a, int b, int x, int y) b 590 mm/percpu.c return (a < y) && (x < b); b 2421 mm/slab.c static void swap_free_obj(struct page *page, unsigned int a, unsigned int b) b 2424 mm/slab.c ((freelist_idx_t *)page->freelist)[b]); b 3034 mm/slab.c #define cache_alloc_debugcheck_after(a,b,objp,d) (objp) b 210 mm/slob.c static void slob_free_pages(void *b, int order) b 212 mm/slob.c struct page *sp = virt_to_page(b); b 306 mm/slob.c slob_t *b = NULL; b 333 mm/slob.c b = slob_page_alloc(sp, size, align, align_offset, &page_removed_from_list); b 334 mm/slob.c if (!b) b 357 mm/slob.c if (!b) { b 358 mm/slob.c b = slob_new_pages(gfp & ~__GFP_ZERO, 0, node); b 359 mm/slob.c if (!b) b 361 mm/slob.c sp = virt_to_page(b); b 366 mm/slob.c sp->freelist = b; b 368 mm/slob.c set_slob(b, SLOB_UNITS(PAGE_SIZE), b + SLOB_UNITS(PAGE_SIZE)); b 370 mm/slob.c b = slob_page_alloc(sp, size, align, align_offset, &_unused); b 371 mm/slob.c BUG_ON(!b); b 375 mm/slob.c memset(b, 0, size); b 376 mm/slob.c return b; b 385 mm/slob.c slob_t *prev, *next, *b = (slob_t *)block; b 406 mm/slob.c slob_free_pages(b, 0); b 413 mm/slob.c sp->freelist = b; b 414 mm/slob.c set_slob(b, units, b 415 mm/slob.c (void *)((unsigned long)(b + b 433 mm/slob.c if (b < (slob_t *)sp->freelist) { b 434 mm/slob.c if (b + units == sp->freelist) { b 438 mm/slob.c set_slob(b, units, sp->freelist); b 439 mm/slob.c sp->freelist = b; b 443 mm/slob.c while (b > next) { b 448 mm/slob.c if (!slob_last(prev) && b + units == next) { b 450 mm/slob.c set_slob(b, units, slob_next(next)); b 452 mm/slob.c set_slob(b, units, next); b 454 mm/slob.c if (prev + slob_units(prev) == b) { b 455 mm/slob.c units = slob_units(b) + slob_units(prev); b 456 mm/slob.c set_slob(prev, units, slob_next(b)); b 458 mm/slob.c set_slob(prev, slob_units(prev), b); b 594 mm/slob.c void *b; b 602 mm/slob.c b = slob_alloc(c->size, flags, c->align, node, 0); b 603 mm/slob.c trace_kmem_cache_alloc_node(_RET_IP_, b, c->object_size, b 607 mm/slob.c b = slob_new_pages(flags, get_order(c->size), node); b 608 mm/slob.c trace_kmem_cache_alloc_node(_RET_IP_, b, c->object_size, b 613 mm/slob.c if (b && c->ctor) { b 615 mm/slob.c c->ctor(b); b 618 mm/slob.c kmemleak_alloc_recursive(b, c->size, 1, c->flags, flags); b 619 mm/slob.c return b; b 642 mm/slob.c static void __kmem_cache_free(void *b, int size) b 645 mm/slob.c slob_free(b, size); b 647 mm/slob.c slob_free_pages(b, get_order(size)); b 653 mm/slob.c void *b = (void *)slob_rcu - (slob_rcu->size - sizeof(struct slob_rcu)); b 655 mm/slob.c __kmem_cache_free(b, slob_rcu->size); b 658 mm/slob.c void kmem_cache_free(struct kmem_cache *c, void *b) b 660 mm/slob.c kmemleak_free_recursive(b, c->flags); b 663 mm/slob.c slob_rcu = b + (c->size - sizeof(struct slob_rcu)); b 667 mm/slob.c __kmem_cache_free(b, c->size); b 670 mm/slob.c trace_kmem_cache_free(_RET_IP_, b); b 129 mm/zsmalloc.c #define MAX(a, b) ((a) >= (b) ? (a) : (b)) b 28 net/atm/addr.c static int identical(const struct sockaddr_atmsvc *a, const struct sockaddr_atmsvc *b) b 31 net/atm/addr.c if (memcmp(a->sas_addr.prv, b->sas_addr.prv, ATM_ESA_LEN)) b 34 net/atm/addr.c return !*b->sas_addr.pub; b 35 net/atm/addr.c if (!*b->sas_addr.pub) b 37 net/atm/addr.c return !strcmp(a->sas_addr.pub, b->sas_addr.pub); b 725 net/atm/br2684.c atm_backend_t b; b 731 net/atm/br2684.c err = get_user(b, (atm_backend_t __user *) argp); b 734 net/atm/br2684.c if (b != ATM_BACKEND_BR2684) b 452 net/atm/pppoatm.c atm_backend_t b; b 453 net/atm/pppoatm.c if (get_user(b, (atm_backend_t __user *) argp)) b 455 net/atm/pppoatm.c if (b != ATM_BACKEND_PPP) b 114 net/ax25/ax25_addr.c int ax25cmp(const ax25_address *a, const ax25_address *b) b 119 net/ax25/ax25_addr.c if ((a->ax25_call[ct] & 0xFE) != (b->ax25_call[ct] & 0xFE)) /* Clean off repeater bits */ b 124 net/ax25/ax25_addr.c if ((a->ax25_call[ct] & 0x1E) == (b->ax25_call[ct] & 0x1E)) /* SSID without control bit */ b 370 net/batman-adv/main.h #define batadv_inc_counter(b, i) batadv_add_counter(b, i, 1) b 989 net/bluetooth/6lowpan.c &addr->b[5], &addr->b[4], &addr->b[3], b 990 net/bluetooth/6lowpan.c &addr->b[2], &addr->b[1], &addr->b[0], b 2428 net/bluetooth/hci_core.c if (addr_type == ADDR_LE_DEV_RANDOM && (bdaddr->b[5] & 0xc0) != 0xc0) b 2898 net/bluetooth/hci_core.c struct bdaddr_list *b; b 2900 net/bluetooth/hci_core.c list_for_each_entry(b, bdaddr_list, list) { b 2901 net/bluetooth/hci_core.c if (!bacmp(&b->bdaddr, bdaddr) && b->bdaddr_type == type) b 2902 net/bluetooth/hci_core.c return b; b 2912 net/bluetooth/hci_core.c struct bdaddr_list_with_irk *b; b 2914 net/bluetooth/hci_core.c list_for_each_entry(b, bdaddr_list, list) { b 2915 net/bluetooth/hci_core.c if (!bacmp(&b->bdaddr, bdaddr) && b->bdaddr_type == type) b 2916 net/bluetooth/hci_core.c return b; b 2924 net/bluetooth/hci_core.c struct bdaddr_list *b, *n; b 2926 net/bluetooth/hci_core.c list_for_each_entry_safe(b, n, bdaddr_list, list) { b 2927 net/bluetooth/hci_core.c list_del(&b->list); b 2928 net/bluetooth/hci_core.c kfree(b); b 124 net/bluetooth/hci_debugfs.c struct bdaddr_list *b; b 127 net/bluetooth/hci_debugfs.c list_for_each_entry(b, &hdev->whitelist, list) b 128 net/bluetooth/hci_debugfs.c seq_printf(f, "%pMR (type %u)\n", &b->bdaddr, b->bdaddr_type); b 143 net/bluetooth/hci_debugfs.c struct bdaddr_list *b; b 146 net/bluetooth/hci_debugfs.c list_for_each_entry(b, &hdev->blacklist, list) b 147 net/bluetooth/hci_debugfs.c seq_printf(f, "%pMR (type %u)\n", &b->bdaddr, b->bdaddr_type); b 716 net/bluetooth/hci_debugfs.c struct bdaddr_list *b; b 719 net/bluetooth/hci_debugfs.c list_for_each_entry(b, &hdev->le_white_list, list) b 720 net/bluetooth/hci_debugfs.c seq_printf(f, "%pMR (type %u)\n", &b->bdaddr, b->bdaddr_type); b 731 net/bluetooth/hci_debugfs.c struct bdaddr_list *b; b 734 net/bluetooth/hci_debugfs.c list_for_each_entry(b, &hdev->le_resolv_list, list) b 735 net/bluetooth/hci_debugfs.c seq_printf(f, "%pMR (type %u)\n", &b->bdaddr, b->bdaddr_type); b 688 net/bluetooth/hci_request.c struct bdaddr_list *b; b 697 net/bluetooth/hci_request.c list_for_each_entry(b, &hdev->le_white_list, list) { b 702 net/bluetooth/hci_request.c &b->bdaddr, b->bdaddr_type) && b 704 net/bluetooth/hci_request.c &b->bdaddr, b->bdaddr_type)) { b 707 net/bluetooth/hci_request.c cp.bdaddr_type = b->bdaddr_type; b 708 net/bluetooth/hci_request.c bacpy(&cp.bdaddr, &b->bdaddr); b 715 net/bluetooth/hci_request.c if (hci_find_irk_by_addr(hdev, &b->bdaddr, b->bdaddr_type)) { b 1532 net/bluetooth/hci_request.c nrpa.b[5] &= 0x3f; b 1944 net/bluetooth/hci_request.c nrpa.b[5] &= 0x3f; b 1988 net/bluetooth/hci_request.c struct bdaddr_list *b; b 1990 net/bluetooth/hci_request.c list_for_each_entry(b, &hdev->whitelist, list) { b 1993 net/bluetooth/hci_request.c conn = hci_conn_hash_lookup_ba(hdev, ACL_LINK, &b->bdaddr); b 4431 net/bluetooth/mgmt.c if ((cp->bdaddr.b[5] & 0xc0) != 0xc0) b 4997 net/bluetooth/mgmt.c if ((irk->addr.bdaddr.b[5] & 0xc0) != 0xc0) b 5079 net/bluetooth/mgmt.c if ((key->addr.bdaddr.b[5] & 0xc0) != 0xc0) b 5778 net/bluetooth/mgmt.c struct bdaddr_list *b, *btmp; b 5788 net/bluetooth/mgmt.c list_for_each_entry_safe(b, btmp, &hdev->whitelist, list) { b 5789 net/bluetooth/mgmt.c device_removed(sk, hdev, &b->bdaddr, b->bdaddr_type); b 5790 net/bluetooth/mgmt.c list_del(&b->list); b 5791 net/bluetooth/mgmt.c kfree(b); b 7141 net/bluetooth/mgmt.c (key->bdaddr.b[5] & 0xc0) != 0xc0) b 7199 net/bluetooth/mgmt.c (csrk->bdaddr.b[5] & 0xc0) != 0xc0) b 75 net/bluetooth/rfcomm/core.c #define __get_dlci(b) ((b & 0xfc) >> 2) b 76 net/bluetooth/rfcomm/core.c #define __get_channel(b) ((b & 0xf8) >> 3) b 77 net/bluetooth/rfcomm/core.c #define __get_dir(b) ((b & 0x04) >> 2) b 78 net/bluetooth/rfcomm/core.c #define __get_type(b) ((b & 0xef)) b 80 net/bluetooth/rfcomm/core.c #define __test_ea(b) ((b & 0x01)) b 81 net/bluetooth/rfcomm/core.c #define __test_cr(b) (!!(b & 0x02)) b 82 net/bluetooth/rfcomm/core.c #define __test_pf(b) (!!(b & 0x10)) b 97 net/bluetooth/rfcomm/core.c #define __get_mcc_type(b) ((b & 0xfc) >> 2) b 98 net/bluetooth/rfcomm/core.c #define __get_mcc_len(b) ((b & 0xfe) >> 1) b 516 net/bluetooth/smp.c err = smp_ah(irk, &bdaddr->b[3], hash); b 520 net/bluetooth/smp.c return !crypto_memneq(bdaddr->b, hash, 3); b 534 net/bluetooth/smp.c get_random_bytes(&rpa->b[3], 3); b 536 net/bluetooth/smp.c rpa->b[5] &= 0x3f; /* Clear two most significant bits */ b 537 net/bluetooth/smp.c rpa->b[5] |= 0x40; /* Set second most significant bit */ b 539 net/bluetooth/smp.c err = smp_ah(irk, &rpa->b[3], rpa->b); b 1416 net/bluetooth/smp.c u8 *na, *nb, a[7], b[7]; b 1427 net/bluetooth/smp.c memcpy(b, &hcon->resp_addr, 6); b 1429 net/bluetooth/smp.c b[6] = hcon->resp_addr_type; b 1431 net/bluetooth/smp.c return smp_f5(smp->tfm_cmac, smp->dhkey, na, nb, a, b, mackey, ltk); b 1438 net/bluetooth/smp.c u8 a[7], b[7], *local_addr, *remote_addr; b 1442 net/bluetooth/smp.c memcpy(b, &hcon->resp_addr, 6); b 1444 net/bluetooth/smp.c b[6] = hcon->resp_addr_type; b 1448 net/bluetooth/smp.c remote_addr = b; b 1451 net/bluetooth/smp.c local_addr = b; b 2809 net/bluetooth/smp.c u8 a[7], b[7], *local_addr, *remote_addr; b 2819 net/bluetooth/smp.c memcpy(b, &hcon->resp_addr, 6); b 2821 net/bluetooth/smp.c b[6] = hcon->resp_addr_type; b 2825 net/bluetooth/smp.c remote_addr = b; b 2828 net/bluetooth/smp.c local_addr = b; b 126 net/bridge/br_ioctl.c struct __bridge_info b; b 128 net/bridge/br_ioctl.c memset(&b, 0, sizeof(struct __bridge_info)); b 130 net/bridge/br_ioctl.c memcpy(&b.designated_root, &br->designated_root, 8); b 131 net/bridge/br_ioctl.c memcpy(&b.bridge_id, &br->bridge_id, 8); b 132 net/bridge/br_ioctl.c b.root_path_cost = br->root_path_cost; b 133 net/bridge/br_ioctl.c b.max_age = jiffies_to_clock_t(br->max_age); b 134 net/bridge/br_ioctl.c b.hello_time = jiffies_to_clock_t(br->hello_time); b 135 net/bridge/br_ioctl.c b.forward_delay = br->forward_delay; b 136 net/bridge/br_ioctl.c b.bridge_max_age = br->bridge_max_age; b 137 net/bridge/br_ioctl.c b.bridge_hello_time = br->bridge_hello_time; b 138 net/bridge/br_ioctl.c b.bridge_forward_delay = jiffies_to_clock_t(br->bridge_forward_delay); b 139 net/bridge/br_ioctl.c b.topology_change = br->topology_change; b 140 net/bridge/br_ioctl.c b.topology_change_detected = br->topology_change_detected; b 141 net/bridge/br_ioctl.c b.root_port = br->root_port; b 143 net/bridge/br_ioctl.c b.stp_enabled = (br->stp_enabled != BR_NO_STP); b 144 net/bridge/br_ioctl.c b.ageing_time = jiffies_to_clock_t(br->ageing_time); b 145 net/bridge/br_ioctl.c b.hello_timer_value = br_timer_value(&br->hello_timer); b 146 net/bridge/br_ioctl.c b.tcn_timer_value = br_timer_value(&br->tcn_timer); b 147 net/bridge/br_ioctl.c b.topology_change_timer_value = br_timer_value(&br->topology_change_timer); b 148 net/bridge/br_ioctl.c b.gc_timer_value = br_timer_value(&br->gc_work.timer); b 151 net/bridge/br_ioctl.c if (copy_to_user((void __user *)args[1], &b, sizeof(b))) b 195 net/bridge/netfilter/ebt_among.c static bool wormhash_sizes_valid(const struct ebt_mac_wormhash *wh, int a, int b) b 200 net/bridge/netfilter/ebt_among.c return ebt_mac_wormhash_size(wh) + a == b; b 1887 net/bridge/netfilter/ebtables.c char *b = state->buf_kern_start; b 1889 net/bridge/netfilter/ebtables.c if (WARN_ON(b && state->buf_kern_offset > state->buf_kern_len)) b 1892 net/bridge/netfilter/ebtables.c if (b != NULL && sz > 0) b 1893 net/bridge/netfilter/ebtables.c memset(b + state->buf_kern_offset, 0, sz); b 43 net/ceph/armor.c unsigned char a, b, c; b 48 net/ceph/armor.c b = *src++; b 49 net/ceph/armor.c *dst++ = encode_bits(((a & 3) << 4) | (b >> 4)); b 52 net/ceph/armor.c *dst++ = encode_bits(((b & 15) << 2) | b 56 net/ceph/armor.c *dst++ = encode_bits((b & 15) << 2); b 80 net/ceph/armor.c int a, b, c, d; b 89 net/ceph/armor.c b = decode_bits(src[1]); b 92 net/ceph/armor.c if (a < 0 || b < 0 || c < 0 || d < 0) b 95 net/ceph/armor.c *dst++ = (a << 2) | (b >> 4); b 98 net/ceph/armor.c *dst++ = ((b & 15) << 4) | (c >> 2); b 844 net/ceph/auth_x.c __le64 a, b, c, d; b 864 net/ceph/auth_x.c *psig = penc->a ^ penc->b ^ penc->c ^ penc->d; b 14 net/ceph/buffer.c struct ceph_buffer *b; b 16 net/ceph/buffer.c b = kmalloc(sizeof(*b), gfp); b 17 net/ceph/buffer.c if (!b) b 20 net/ceph/buffer.c b->vec.iov_base = ceph_kvmalloc(len, gfp); b 21 net/ceph/buffer.c if (!b->vec.iov_base) { b 22 net/ceph/buffer.c kfree(b); b 26 net/ceph/buffer.c kref_init(&b->kref); b 27 net/ceph/buffer.c b->alloc_len = len; b 28 net/ceph/buffer.c b->vec.iov_len = len; b 29 net/ceph/buffer.c dout("buffer_new %p\n", b); b 30 net/ceph/buffer.c return b; b 36 net/ceph/buffer.c struct ceph_buffer *b = container_of(kref, struct ceph_buffer, kref); b 38 net/ceph/buffer.c dout("buffer_release %p\n", b); b 39 net/ceph/buffer.c kvfree(b->vec.iov_base); b 40 net/ceph/buffer.c kfree(b); b 44 net/ceph/buffer.c int ceph_decode_buffer(struct ceph_buffer **b, void **p, void *end) b 52 net/ceph/buffer.c *b = ceph_buffer_new(len, GFP_NOFS); b 53 net/ceph/buffer.c if (!*b) b 55 net/ceph/buffer.c ceph_decode_copy(p, (*b)->vec.iov_base, len); b 10 net/ceph/ceph_hash.c #define mix(a, b, c) \ b 12 net/ceph/ceph_hash.c a = a - b; a = a - c; a = a ^ (c >> 13); \ b 13 net/ceph/ceph_hash.c b = b - c; b = b - a; b = b ^ (a << 8); \ b 14 net/ceph/ceph_hash.c c = c - a; c = c - b; c = c ^ (b >> 13); \ b 15 net/ceph/ceph_hash.c a = a - b; a = a - c; a = a ^ (c >> 12); \ b 16 net/ceph/ceph_hash.c b = b - c; b = b - a; b = b ^ (a << 16); \ b 17 net/ceph/ceph_hash.c c = c - a; c = c - b; c = c ^ (b >> 5); \ b 18 net/ceph/ceph_hash.c a = a - b; a = a - c; a = a ^ (c >> 3); \ b 19 net/ceph/ceph_hash.c b = b - c; b = b - a; b = b ^ (a << 10); \ b 20 net/ceph/ceph_hash.c c = c - a; c = c - b; c = c ^ (b >> 15); \ b 26 net/ceph/ceph_hash.c __u32 a, b, c; /* the internal state */ b 32 net/ceph/ceph_hash.c b = a; b 39 net/ceph/ceph_hash.c b = b + (k[4] + ((__u32)k[5] << 8) + ((__u32)k[6] << 16) + b 43 net/ceph/ceph_hash.c mix(a, b, c); b 62 net/ceph/ceph_hash.c b = b + ((__u32)k[7] << 24); b 65 net/ceph/ceph_hash.c b = b + ((__u32)k[6] << 16); b 68 net/ceph/ceph_hash.c b = b + ((__u32)k[5] << 8); b 71 net/ceph/ceph_hash.c b = b + k[4]; b 86 net/ceph/ceph_hash.c mix(a, b, c); b 28 net/ceph/crush/crush.c int crush_get_bucket_item_weight(const struct crush_bucket *b, int p) b 30 net/ceph/crush/crush.c if ((__u32)p >= b->size) b 33 net/ceph/crush/crush.c switch (b->alg) { b 35 net/ceph/crush/crush.c return ((struct crush_bucket_uniform *)b)->item_weight; b 37 net/ceph/crush/crush.c return ((struct crush_bucket_list *)b)->item_weights[p]; b 39 net/ceph/crush/crush.c return ((struct crush_bucket_tree *)b)->node_weights[crush_calc_tree_node(p)]; b 41 net/ceph/crush/crush.c return ((struct crush_bucket_straw *)b)->item_weights[p]; b 43 net/ceph/crush/crush.c return ((struct crush_bucket_straw2 *)b)->item_weights[p]; b 48 net/ceph/crush/crush.c void crush_destroy_bucket_uniform(struct crush_bucket_uniform *b) b 50 net/ceph/crush/crush.c kfree(b->h.items); b 51 net/ceph/crush/crush.c kfree(b); b 54 net/ceph/crush/crush.c void crush_destroy_bucket_list(struct crush_bucket_list *b) b 56 net/ceph/crush/crush.c kfree(b->item_weights); b 57 net/ceph/crush/crush.c kfree(b->sum_weights); b 58 net/ceph/crush/crush.c kfree(b->h.items); b 59 net/ceph/crush/crush.c kfree(b); b 62 net/ceph/crush/crush.c void crush_destroy_bucket_tree(struct crush_bucket_tree *b) b 64 net/ceph/crush/crush.c kfree(b->h.items); b 65 net/ceph/crush/crush.c kfree(b->node_weights); b 66 net/ceph/crush/crush.c kfree(b); b 69 net/ceph/crush/crush.c void crush_destroy_bucket_straw(struct crush_bucket_straw *b) b 71 net/ceph/crush/crush.c kfree(b->straws); b 72 net/ceph/crush/crush.c kfree(b->item_weights); b 73 net/ceph/crush/crush.c kfree(b->h.items); b 74 net/ceph/crush/crush.c kfree(b); b 77 net/ceph/crush/crush.c void crush_destroy_bucket_straw2(struct crush_bucket_straw2 *b) b 79 net/ceph/crush/crush.c kfree(b->item_weights); b 80 net/ceph/crush/crush.c kfree(b->h.items); b 81 net/ceph/crush/crush.c kfree(b); b 84 net/ceph/crush/crush.c void crush_destroy_bucket(struct crush_bucket *b) b 86 net/ceph/crush/crush.c switch (b->alg) { b 88 net/ceph/crush/crush.c crush_destroy_bucket_uniform((struct crush_bucket_uniform *)b); b 91 net/ceph/crush/crush.c crush_destroy_bucket_list((struct crush_bucket_list *)b); b 94 net/ceph/crush/crush.c crush_destroy_bucket_tree((struct crush_bucket_tree *)b); b 97 net/ceph/crush/crush.c crush_destroy_bucket_straw((struct crush_bucket_straw *)b); b 100 net/ceph/crush/crush.c crush_destroy_bucket_straw2((struct crush_bucket_straw2 *)b); b 113 net/ceph/crush/crush.c __s32 b; b 114 net/ceph/crush/crush.c for (b = 0; b < map->max_buckets; b++) { b 115 net/ceph/crush/crush.c if (map->buckets[b] == NULL) b 117 net/ceph/crush/crush.c crush_destroy_bucket(map->buckets[b]); b 124 net/ceph/crush/crush.c __u32 b; b 125 net/ceph/crush/crush.c for (b = 0; b < map->max_rules; b++) b 126 net/ceph/crush/crush.c crush_destroy_rule(map->rules[b]); b 13 net/ceph/crush/hash.c #define crush_hashmix(a, b, c) do { \ b 14 net/ceph/crush/hash.c a = a-b; a = a-c; a = a^(c>>13); \ b 15 net/ceph/crush/hash.c b = b-c; b = b-a; b = b^(a<<8); \ b 16 net/ceph/crush/hash.c c = c-a; c = c-b; c = c^(b>>13); \ b 17 net/ceph/crush/hash.c a = a-b; a = a-c; a = a^(c>>12); \ b 18 net/ceph/crush/hash.c b = b-c; b = b-a; b = b^(a<<16); \ b 19 net/ceph/crush/hash.c c = c-a; c = c-b; c = c^(b>>5); \ b 20 net/ceph/crush/hash.c a = a-b; a = a-c; a = a^(c>>3); \ b 21 net/ceph/crush/hash.c b = b-c; b = b-a; b = b^(a<<10); \ b 22 net/ceph/crush/hash.c c = c-a; c = c-b; c = c^(b>>15); \ b 30 net/ceph/crush/hash.c __u32 b = a; b 33 net/ceph/crush/hash.c crush_hashmix(b, x, hash); b 38 net/ceph/crush/hash.c static __u32 crush_hash32_rjenkins1_2(__u32 a, __u32 b) b 40 net/ceph/crush/hash.c __u32 hash = crush_hash_seed ^ a ^ b; b 43 net/ceph/crush/hash.c crush_hashmix(a, b, hash); b 45 net/ceph/crush/hash.c crush_hashmix(b, y, hash); b 49 net/ceph/crush/hash.c static __u32 crush_hash32_rjenkins1_3(__u32 a, __u32 b, __u32 c) b 51 net/ceph/crush/hash.c __u32 hash = crush_hash_seed ^ a ^ b ^ c; b 54 net/ceph/crush/hash.c crush_hashmix(a, b, hash); b 57 net/ceph/crush/hash.c crush_hashmix(b, x, hash); b 62 net/ceph/crush/hash.c static __u32 crush_hash32_rjenkins1_4(__u32 a, __u32 b, __u32 c, __u32 d) b 64 net/ceph/crush/hash.c __u32 hash = crush_hash_seed ^ a ^ b ^ c ^ d; b 67 net/ceph/crush/hash.c crush_hashmix(a, b, hash); b 70 net/ceph/crush/hash.c crush_hashmix(y, b, hash); b 76 net/ceph/crush/hash.c static __u32 crush_hash32_rjenkins1_5(__u32 a, __u32 b, __u32 c, __u32 d, b 79 net/ceph/crush/hash.c __u32 hash = crush_hash_seed ^ a ^ b ^ c ^ d ^ e; b 82 net/ceph/crush/hash.c crush_hashmix(a, b, hash); b 86 net/ceph/crush/hash.c crush_hashmix(b, x, hash); b 104 net/ceph/crush/hash.c __u32 crush_hash32_2(int type, __u32 a, __u32 b) b 108 net/ceph/crush/hash.c return crush_hash32_rjenkins1_2(a, b); b 114 net/ceph/crush/hash.c __u32 crush_hash32_3(int type, __u32 a, __u32 b, __u32 c) b 118 net/ceph/crush/hash.c return crush_hash32_rjenkins1_3(a, b, c); b 124 net/ceph/crush/hash.c __u32 crush_hash32_4(int type, __u32 a, __u32 b, __u32 c, __u32 d) b 128 net/ceph/crush/hash.c return crush_hash32_rjenkins1_4(a, b, c, d); b 134 net/ceph/crush/hash.c __u32 crush_hash32_5(int type, __u32 a, __u32 b, __u32 c, __u32 d, __u32 e) b 138 net/ceph/crush/hash.c return crush_hash32_rjenkins1_5(a, b, c, d, e); b 851 net/ceph/crush/mapper.c __s32 b; b 864 net/ceph/crush/mapper.c for (b = 0; b < map->max_buckets; ++b) { b 865 net/ceph/crush/mapper.c if (!map->buckets[b]) b 868 net/ceph/crush/mapper.c w->work[b] = v; b 869 net/ceph/crush/mapper.c switch (map->buckets[b]->alg) { b 874 net/ceph/crush/mapper.c w->work[b]->perm_x = 0; b 875 net/ceph/crush/mapper.c w->work[b]->perm_n = 0; b 876 net/ceph/crush/mapper.c w->work[b]->perm = v; b 877 net/ceph/crush/mapper.c v += map->buckets[b]->size * sizeof(__u32); b 902 net/ceph/crush/mapper.c int *b = a + result_max; b 903 net/ceph/crush/mapper.c int *c = b + result_max; b 905 net/ceph/crush/mapper.c int *o = b; b 35 net/ceph/osdmap.c int b = 0; b 38 net/ceph/osdmap.c b++; b 40 net/ceph/osdmap.c return b; b 56 net/ceph/osdmap.c struct crush_bucket_uniform *b) b 59 net/ceph/osdmap.c ceph_decode_need(p, end, (1+b->h.size) * sizeof(u32), bad); b 60 net/ceph/osdmap.c b->item_weight = ceph_decode_32(p); b 67 net/ceph/osdmap.c struct crush_bucket_list *b) b 71 net/ceph/osdmap.c b->item_weights = kcalloc(b->h.size, sizeof(u32), GFP_NOFS); b 72 net/ceph/osdmap.c if (b->item_weights == NULL) b 74 net/ceph/osdmap.c b->sum_weights = kcalloc(b->h.size, sizeof(u32), GFP_NOFS); b 75 net/ceph/osdmap.c if (b->sum_weights == NULL) b 77 net/ceph/osdmap.c ceph_decode_need(p, end, 2 * b->h.size * sizeof(u32), bad); b 78 net/ceph/osdmap.c for (j = 0; j < b->h.size; j++) { b 79 net/ceph/osdmap.c b->item_weights[j] = ceph_decode_32(p); b 80 net/ceph/osdmap.c b->sum_weights[j] = ceph_decode_32(p); b 88 net/ceph/osdmap.c struct crush_bucket_tree *b) b 92 net/ceph/osdmap.c ceph_decode_8_safe(p, end, b->num_nodes, bad); b 93 net/ceph/osdmap.c b->node_weights = kcalloc(b->num_nodes, sizeof(u32), GFP_NOFS); b 94 net/ceph/osdmap.c if (b->node_weights == NULL) b 96 net/ceph/osdmap.c ceph_decode_need(p, end, b->num_nodes * sizeof(u32), bad); b 97 net/ceph/osdmap.c for (j = 0; j < b->num_nodes; j++) b 98 net/ceph/osdmap.c b->node_weights[j] = ceph_decode_32(p); b 105 net/ceph/osdmap.c struct crush_bucket_straw *b) b 109 net/ceph/osdmap.c b->item_weights = kcalloc(b->h.size, sizeof(u32), GFP_NOFS); b 110 net/ceph/osdmap.c if (b->item_weights == NULL) b 112 net/ceph/osdmap.c b->straws = kcalloc(b->h.size, sizeof(u32), GFP_NOFS); b 113 net/ceph/osdmap.c if (b->straws == NULL) b 115 net/ceph/osdmap.c ceph_decode_need(p, end, 2 * b->h.size * sizeof(u32), bad); b 116 net/ceph/osdmap.c for (j = 0; j < b->h.size; j++) { b 117 net/ceph/osdmap.c b->item_weights[j] = ceph_decode_32(p); b 118 net/ceph/osdmap.c b->straws[j] = ceph_decode_32(p); b 126 net/ceph/osdmap.c struct crush_bucket_straw2 *b) b 130 net/ceph/osdmap.c b->item_weights = kcalloc(b->h.size, sizeof(u32), GFP_NOFS); b 131 net/ceph/osdmap.c if (b->item_weights == NULL) b 133 net/ceph/osdmap.c ceph_decode_need(p, end, b->h.size * sizeof(u32), bad); b 134 net/ceph/osdmap.c for (j = 0; j < b->h.size; j++) b 135 net/ceph/osdmap.c b->item_weights[j] = ceph_decode_32(p); b 318 net/ceph/osdmap.c __s32 b; b 324 net/ceph/osdmap.c for (b = 0; b < c->max_buckets; b++) { b 325 net/ceph/osdmap.c if (!c->buckets[b]) b 328 net/ceph/osdmap.c switch (c->buckets[b]->alg) { b 338 net/ceph/osdmap.c c->working_size += c->buckets[b]->size * sizeof(__u32); b 387 net/ceph/osdmap.c struct crush_bucket *b; b 417 net/ceph/osdmap.c b = c->buckets[i] = kzalloc(size, GFP_NOFS); b 418 net/ceph/osdmap.c if (b == NULL) b 422 net/ceph/osdmap.c b->id = ceph_decode_32(p); b 423 net/ceph/osdmap.c b->type = ceph_decode_16(p); b 424 net/ceph/osdmap.c b->alg = ceph_decode_8(p); b 425 net/ceph/osdmap.c b->hash = ceph_decode_8(p); b 426 net/ceph/osdmap.c b->weight = ceph_decode_32(p); b 427 net/ceph/osdmap.c b->size = ceph_decode_32(p); b 430 net/ceph/osdmap.c b->size, (int)(*p-start), *p, end); b 432 net/ceph/osdmap.c b->items = kcalloc(b->size, sizeof(__s32), GFP_NOFS); b 433 net/ceph/osdmap.c if (b->items == NULL) b 436 net/ceph/osdmap.c ceph_decode_need(p, end, b->size*sizeof(u32), bad); b 437 net/ceph/osdmap.c for (j = 0; j < b->size; j++) b 438 net/ceph/osdmap.c b->items[j] = ceph_decode_32(p); b 440 net/ceph/osdmap.c switch (b->alg) { b 443 net/ceph/osdmap.c (struct crush_bucket_uniform *)b); b 449 net/ceph/osdmap.c (struct crush_bucket_list *)b); b 455 net/ceph/osdmap.c (struct crush_bucket_tree *)b); b 461 net/ceph/osdmap.c (struct crush_bucket_straw *)b); b 467 net/ceph/osdmap.c (struct crush_bucket_straw2 *)b); b 225 net/core/bpf_sk_storage.c struct bucket *b; b 232 net/core/bpf_sk_storage.c b = select_bucket(smap, selem); b 233 net/core/bpf_sk_storage.c raw_spin_lock_bh(&b->lock); b 236 net/core/bpf_sk_storage.c raw_spin_unlock_bh(&b->lock); b 242 net/core/bpf_sk_storage.c struct bucket *b = select_bucket(smap, selem); b 244 net/core/bpf_sk_storage.c raw_spin_lock_bh(&b->lock); b 246 net/core/bpf_sk_storage.c hlist_add_head_rcu(&selem->map_node, &b->list); b 247 net/core/bpf_sk_storage.c raw_spin_unlock_bh(&b->lock); b 557 net/core/bpf_sk_storage.c struct bucket *b; b 578 net/core/bpf_sk_storage.c b = &smap->buckets[i]; b 582 net/core/bpf_sk_storage.c while ((selem = hlist_entry_safe(rcu_dereference_raw(hlist_first_rcu(&b->list)), b 8257 net/core/dev.c bool netdev_port_same_parent_id(struct net_device *a, struct net_device *b) b 8263 net/core/dev.c dev_get_port_parent_id(b, &b_id, true)) b 2559 net/core/filter.c int b = sge->length - pop - a; b 2571 net/core/filter.c b, sge->offset + pop + a); b 2578 net/core/filter.c get_order(a + b)); b 2587 net/core/filter.c memcpy(to + a, from + a + pop, b); b 2588 net/core/filter.c sg_set_page(sge, page, a + b, 0); b 63 net/core/gen_estimator.c struct gnet_stats_basic_packed *b) b 65 net/core/gen_estimator.c memset(b, 0, sizeof(*b)); b 69 net/core/gen_estimator.c __gnet_stats_copy_basic(e->running, b, e->cpu_bstats, e->bstats); b 79 net/core/gen_estimator.c struct gnet_stats_basic_packed b; b 82 net/core/gen_estimator.c est_fetch_counters(est, &b); b 83 net/core/gen_estimator.c brate = (b.bytes - est->last_bytes) << (10 - est->ewma_log - est->intvl_log); b 86 net/core/gen_estimator.c rate = (u64)(b.packets - est->last_packets) << (10 - est->ewma_log - est->intvl_log); b 94 net/core/gen_estimator.c est->last_bytes = b.bytes; b 95 net/core/gen_estimator.c est->last_packets = b.packets; b 133 net/core/gen_estimator.c struct gnet_stats_basic_packed b; b 161 net/core/gen_estimator.c est_fetch_counters(est, &b); b 164 net/core/gen_estimator.c est->last_bytes = b.bytes; b 165 net/core/gen_estimator.c est->last_packets = b.packets; b 144 net/core/gen_stats.c struct gnet_stats_basic_packed *b) b 155 net/core/gen_stats.c bstats->bytes = b->bytes; b 156 net/core/gen_stats.c bstats->packets = b->packets; b 165 net/core/gen_stats.c struct gnet_stats_basic_packed *b, b 170 net/core/gen_stats.c __gnet_stats_copy_basic(running, &bstats, cpu, b); b 206 net/core/gen_stats.c struct gnet_stats_basic_packed *b) b 208 net/core/gen_stats.c return ___gnet_stats_copy_basic(running, d, cpu, b, b 230 net/core/gen_stats.c struct gnet_stats_basic_packed *b) b 232 net/core/gen_stats.c return ___gnet_stats_copy_basic(running, d, cpu, b, b 487 net/core/lwt_bpf.c static int bpf_lwt_prog_cmp(struct bpf_lwt_prog *a, struct bpf_lwt_prog *b) b 493 net/core/lwt_bpf.c if (!a->name && !b->name) b 496 net/core/lwt_bpf.c if (!a->name || !b->name) b 499 net/core/lwt_bpf.c return strcmp(a->name, b->name); b 502 net/core/lwt_bpf.c static int bpf_encap_cmp(struct lwtunnel_state *a, struct lwtunnel_state *b) b 505 net/core/lwt_bpf.c struct bpf_lwt *b_bpf = bpf_lwt_lwtunnel(b); b 284 net/core/lwtunnel.c int lwtunnel_cmp_encap(struct lwtunnel_state *a, struct lwtunnel_state *b) b 289 net/core/lwtunnel.c if (!a && !b) b 292 net/core/lwtunnel.c if (!a || !b) b 295 net/core/lwtunnel.c if (a->type != b->type) b 305 net/core/lwtunnel.c ret = ops->cmp_encap(a, b); b 11 net/core/net-procfs.c #define set_bucket_offset(b, o) ((b) << BUCKET_SPACE | (o)) b 193 net/core/page_pool.c #define _distance(a, b) (s32)((a) - (b)) b 1971 net/core/pktgen.c char b[IFNAMSIZ+5]; b 1978 net/core/pktgen.c b[i] = ifname[i]; b 1980 net/core/pktgen.c b[i] = 0; b 1982 net/core/pktgen.c return dev_get_by_name(pn->net, b); b 872 net/core/rtnetlink.c const struct rtnl_link_stats64 *b) b 874 net/core/rtnetlink.c a->rx_packets = b->rx_packets; b 875 net/core/rtnetlink.c a->tx_packets = b->tx_packets; b 876 net/core/rtnetlink.c a->rx_bytes = b->rx_bytes; b 877 net/core/rtnetlink.c a->tx_bytes = b->tx_bytes; b 878 net/core/rtnetlink.c a->rx_errors = b->rx_errors; b 879 net/core/rtnetlink.c a->tx_errors = b->tx_errors; b 880 net/core/rtnetlink.c a->rx_dropped = b->rx_dropped; b 881 net/core/rtnetlink.c a->tx_dropped = b->tx_dropped; b 883 net/core/rtnetlink.c a->multicast = b->multicast; b 884 net/core/rtnetlink.c a->collisions = b->collisions; b 886 net/core/rtnetlink.c a->rx_length_errors = b->rx_length_errors; b 887 net/core/rtnetlink.c a->rx_over_errors = b->rx_over_errors; b 888 net/core/rtnetlink.c a->rx_crc_errors = b->rx_crc_errors; b 889 net/core/rtnetlink.c a->rx_frame_errors = b->rx_frame_errors; b 890 net/core/rtnetlink.c a->rx_fifo_errors = b->rx_fifo_errors; b 891 net/core/rtnetlink.c a->rx_missed_errors = b->rx_missed_errors; b 893 net/core/rtnetlink.c a->tx_aborted_errors = b->tx_aborted_errors; b 894 net/core/rtnetlink.c a->tx_carrier_errors = b->tx_carrier_errors; b 895 net/core/rtnetlink.c a->tx_fifo_errors = b->tx_fifo_errors; b 896 net/core/rtnetlink.c a->tx_heartbeat_errors = b->tx_heartbeat_errors; b 897 net/core/rtnetlink.c a->tx_window_errors = b->tx_window_errors; b 899 net/core/rtnetlink.c a->rx_compressed = b->rx_compressed; b 900 net/core/rtnetlink.c a->tx_compressed = b->tx_compressed; b 902 net/core/rtnetlink.c a->rx_nohandler = b->rx_nohandler; b 107 net/dccp/ackvec.c static inline u16 __ackvec_idx_add(const u16 a, const u16 b) b 109 net/dccp/ackvec.c return (a + b) % DCCPAV_MAX_ACKVEC_LEN; b 112 net/dccp/ackvec.c static inline u16 __ackvec_idx_sub(const u16 a, const u16 b) b 114 net/dccp/ackvec.c return __ackvec_idx_add(a, DCCPAV_MAX_ACKVEC_LEN - b); b 135 net/dccp/ccids/lib/packet_history.c static void tfrc_rx_hist_swap(struct tfrc_rx_hist *h, const u8 a, const u8 b) b 138 net/dccp/ccids/lib/packet_history.c idx_b = tfrc_rx_hist_index(h, b); b 54 net/dccp/ccids/lib/packet_history.h #define SUB16(a, b) (((a) + 16 - (b)) & 0xF) b 27 net/dccp/ccids/lib/tfrc.h static inline u64 scaled_div(u64 a, u64 b) b 29 net/dccp/ccids/lib/tfrc.h BUG_ON(b == 0); b 30 net/dccp/ccids/lib/tfrc.h return div64_u64(a * 1000000, b); b 33 net/dccp/ccids/lib/tfrc.h static inline u32 scaled_div32(u64 a, u64 b) b 35 net/dccp/ccids/lib/tfrc.h u64 result = scaled_div(a, b); b 39 net/dccp/ccids/lib/tfrc.h (unsigned long long)a, (unsigned long long)b); b 108 net/dccp/dccp.h #define ADD48(a, b) (((a) + (b)) & UINT48_MAX) b 109 net/dccp/dccp.h #define SUB48(a, b) ADD48((a), COMPLEMENT48(b)) b 113 net/decnet/dn_table.c static inline int dn_key_eq(dn_fib_key_t a, dn_fib_key_t b) b 115 net/decnet/dn_table.c return a.datum == b.datum; b 118 net/decnet/dn_table.c static inline int dn_key_leq(dn_fib_key_t a, dn_fib_key_t b) b 120 net/decnet/dn_table.c return a.datum <= b.datum; b 26 net/hsr/hsr_framereg.c static bool seq_nr_after(u16 a, u16 b) b 31 net/hsr/hsr_framereg.c if ((int)b - a == 32768) b 34 net/hsr/hsr_framereg.c return (((s16)(b - a)) < 0); b 37 net/hsr/hsr_framereg.c #define seq_nr_before(a, b) seq_nr_after((b), (a)) b 38 net/hsr/hsr_framereg.c #define seq_nr_after_or_eq(a, b) (!seq_nr_before((a), (b))) b 39 net/hsr/hsr_framereg.c #define seq_nr_before_or_eq(a, b) (!seq_nr_after((a), (b))) b 331 net/ipv4/devinet.c int inet_addr_onlink(struct in_device *in_dev, __be32 a, __be32 b) b 338 net/ipv4/devinet.c if (!b || inet_ifa_match(b, ifa)) { b 315 net/ipv4/ip_tunnel_core.c static int ip_tun_cmp_encap(struct lwtunnel_state *a, struct lwtunnel_state *b) b 317 net/ipv4/ip_tunnel_core.c return memcmp(lwt_tun_info(a), lwt_tun_info(b), b 800 net/ipv4/ipconfig.c struct bootp_pkt *b; b 811 net/ipv4/ipconfig.c b = skb_put_zero(skb, sizeof(struct bootp_pkt)); b 826 net/ipv4/ipconfig.c b->udph.source = htons(68); b 827 net/ipv4/ipconfig.c b->udph.dest = htons(67); b 828 net/ipv4/ipconfig.c b->udph.len = htons(sizeof(struct bootp_pkt) - sizeof(struct iphdr)); b 832 net/ipv4/ipconfig.c b->op = BOOTP_REQUEST; b 834 net/ipv4/ipconfig.c b->htype = dev->type; b 836 net/ipv4/ipconfig.c b->htype = ARPHRD_ETHER; b 840 net/ipv4/ipconfig.c b->htype = dev->type; /* can cause undefined behavior */ b 844 net/ipv4/ipconfig.c b->hlen = dev->addr_len; b 845 net/ipv4/ipconfig.c memcpy(b->hw_addr, dev->dev_addr, dev->addr_len); b 846 net/ipv4/ipconfig.c b->secs = htons(jiffies_diff / HZ); b 847 net/ipv4/ipconfig.c b->xid = d->xid; b 852 net/ipv4/ipconfig.c ic_dhcp_init_options(b->exten, d); b 855 net/ipv4/ipconfig.c ic_bootp_init_ext(b->exten); b 960 net/ipv4/ipconfig.c struct bootp_pkt *b; b 981 net/ipv4/ipconfig.c b = (struct bootp_pkt *)skb_network_header(skb); b 982 net/ipv4/ipconfig.c h = &b->iph; b 999 net/ipv4/ipconfig.c if (b->udph.source != htons(67) || b->udph.dest != htons(68)) b 1002 net/ipv4/ipconfig.c if (ntohs(h->tot_len) < ntohs(b->udph.len) + sizeof(struct iphdr)) b 1005 net/ipv4/ipconfig.c len = ntohs(b->udph.len) - sizeof(struct udphdr); b 1006 net/ipv4/ipconfig.c ext_len = len - (sizeof(*b) - b 1009 net/ipv4/ipconfig.c sizeof(b->exten)); b 1017 net/ipv4/ipconfig.c b = (struct bootp_pkt *)skb_network_header(skb); b 1018 net/ipv4/ipconfig.c h = &b->iph; b 1035 net/ipv4/ipconfig.c if (b->op != BOOTP_REPLY || b 1036 net/ipv4/ipconfig.c b->xid != d->xid) { b 1038 net/ipv4/ipconfig.c d->dev->name, b->op, b->xid); b 1044 net/ipv4/ipconfig.c !memcmp(b->exten, ic_bootp_cookie, 4)) { /* Check magic cookie */ b 1045 net/ipv4/ipconfig.c u8 *end = (u8 *) b + ntohs(b->iph.tot_len); b 1053 net/ipv4/ipconfig.c ext = &b->exten[4]; b 1084 net/ipv4/ipconfig.c ic_myaddr = b->your_ip; b 1087 net/ipv4/ipconfig.c &ic_myaddr, &b->iph.saddr); b 1093 net/ipv4/ipconfig.c (b->server_ip != server_id)) b 1094 net/ipv4/ipconfig.c b->server_ip = ic_servaddr; b 1098 net/ipv4/ipconfig.c if (memcmp(dev->dev_addr, b->hw_addr, dev->addr_len) != 0) b 1116 net/ipv4/ipconfig.c ext = &b->exten[4]; b 1129 net/ipv4/ipconfig.c ic_myaddr = b->your_ip; b 1130 net/ipv4/ipconfig.c ic_servaddr = b->server_ip; b 1131 net/ipv4/ipconfig.c ic_addrservaddr = b->iph.saddr; b 1132 net/ipv4/ipconfig.c if (ic_gateway == NONE && b->relay_ip) b 1133 net/ipv4/ipconfig.c ic_gateway = b->relay_ip; b 71 net/ipv4/netfilter/arp_tables.c const u16 *b = (const u16 *)_b; b 76 net/ipv4/netfilter/arp_tables.c ret |= (a[i] ^ b[i]) & mask[i]; b 180 net/ipv4/tcp_cubic.c u32 x, b, shift; b 200 net/ipv4/tcp_cubic.c b = fls64(a); b 201 net/ipv4/tcp_cubic.c if (b < 7) { b 206 net/ipv4/tcp_cubic.c b = ((b * 84) >> 8) - 1; b 207 net/ipv4/tcp_cubic.c shift = (a >> (b * 3)); b 209 net/ipv4/tcp_cubic.c x = ((u32)(((u32)v[shift] + 10) << b)) >> 6; b 79 net/ipv4/tcp_metrics.c const struct inetpeer_addr *b) b 81 net/ipv4/tcp_metrics.c return inetpeer_addr_cmp(a, b) == 0; b 81 net/ipv4/tcp_westwood.c static inline u32 westwood_do_filter(u32 a, u32 b) b 83 net/ipv4/tcp_westwood.c return ((7 * a) + b) >> 3; b 298 net/ipv6/ila/ila_lwt.c static int ila_encap_cmp(struct lwtunnel_state *a, struct lwtunnel_state *b) b 301 net/ipv6/ila/ila_lwt.c struct ila_params *b_p = ila_params_lwtunnel(b); b 29 net/ipv6/proc.c #define MAX4(a, b, c, d) \ b 30 net/ipv6/proc.c max_t(u32, max_t(u32, a, b), max_t(u32, c, d)) b 482 net/ipv6/seg6_iptunnel.c static int seg6_encap_cmp(struct lwtunnel_state *a, struct lwtunnel_state *b) b 485 net/ipv6/seg6_iptunnel.c struct seg6_iptunnel_encap *b_hdr = seg6_encap_lwtunnel(b); b 689 net/ipv6/seg6_local.c static int cmp_nla_srh(struct seg6_local_lwt *a, struct seg6_local_lwt *b) b 693 net/ipv6/seg6_local.c if (len != ((b->srh->hdrlen + 1) << 3)) b 696 net/ipv6/seg6_local.c return memcmp(a->srh, b->srh, len); b 714 net/ipv6/seg6_local.c static int cmp_nla_table(struct seg6_local_lwt *a, struct seg6_local_lwt *b) b 716 net/ipv6/seg6_local.c if (a->table != b->table) b 743 net/ipv6/seg6_local.c static int cmp_nla_nh4(struct seg6_local_lwt *a, struct seg6_local_lwt *b) b 745 net/ipv6/seg6_local.c return memcmp(&a->nh4, &b->nh4, sizeof(struct in_addr)); b 769 net/ipv6/seg6_local.c static int cmp_nla_nh6(struct seg6_local_lwt *a, struct seg6_local_lwt *b) b 771 net/ipv6/seg6_local.c return memcmp(&a->nh6, &b->nh6, sizeof(struct in6_addr)); b 789 net/ipv6/seg6_local.c static int cmp_nla_iif(struct seg6_local_lwt *a, struct seg6_local_lwt *b) b 791 net/ipv6/seg6_local.c if (a->iif != b->iif) b 812 net/ipv6/seg6_local.c static int cmp_nla_oif(struct seg6_local_lwt *a, struct seg6_local_lwt *b) b 814 net/ipv6/seg6_local.c if (a->oif != b->oif) b 879 net/ipv6/seg6_local.c static int cmp_nla_bpf(struct seg6_local_lwt *a, struct seg6_local_lwt *b) b 881 net/ipv6/seg6_local.c if (!a->bpf.name && !b->bpf.name) b 884 net/ipv6/seg6_local.c if (!a->bpf.name || !b->bpf.name) b 887 net/ipv6/seg6_local.c return strcmp(a->bpf.name, b->bpf.name); b 893 net/ipv6/seg6_local.c int (*cmp)(struct seg6_local_lwt *a, struct seg6_local_lwt *b); b 1078 net/ipv6/seg6_local.c struct lwtunnel_state *b) b 1085 net/ipv6/seg6_local.c slwt_b = seg6_local_lwtunnel(b); b 1419 net/llc/llc_c_ac.c u8 llc_circular_between(u8 a, u8 b, u8 c) b 1421 net/llc/llc_c_ac.c b = b - a; b 1423 net/llc/llc_c_ac.c return b <= c; b 463 net/mac80211/debugfs_sta.c #define PFLAG(a, b) \ b 467 net/mac80211/debugfs_sta.c "\t\t%s\n", b); \ b 588 net/mac80211/debugfs_sta.c #define PFLAG(t, n, a, b) \ b 591 net/mac80211/debugfs_sta.c PRINT("%s", b); \ b 601 net/mac80211/debugfs_sta.c #define PFLAG_RANGE_DEFAULT(t, i, n, s, m, off, fmt, a, b) \ b 604 net/mac80211/debugfs_sta.c PRINT("%s", b); \ b 19 net/mac80211/fils_aead.c u64 b = get_unaligned_be64(pad + 8); b 21 net/mac80211/fils_aead.c put_unaligned_be64((a << 1) | (b >> 63), pad); b 22 net/mac80211/fils_aead.c put_unaligned_be64((b << 1) ^ ((a >> 63) ? 0x87 : 0), pad + 8); b 53 net/mac80211/rc80211_minstrel_ht.c #define _MAX(a, b) (((a)>(b))?(a):(b)) b 23 net/mac802154/llsec.c const struct ieee802154_llsec_key_id *b); b 189 net/mac802154/llsec.c const struct ieee802154_llsec_key_id *b) b 191 net/mac802154/llsec.c if (a->mode != b->mode) b 195 net/mac802154/llsec.c return ieee802154_addr_equal(&a->device_addr, &b->device_addr); b 197 net/mac802154/llsec.c if (a->id != b->id) b 204 net/mac802154/llsec.c return a->short_source == b->short_source; b 206 net/mac802154/llsec.c return a->extended_source == b->extended_source; b 264 net/mpls/mpls_iptunnel.c static int mpls_encap_cmp(struct lwtunnel_state *a, struct lwtunnel_state *b) b 267 net/mpls/mpls_iptunnel.c struct mpls_iptunnel_encap *b_hdr = mpls_lwtunnel_encap(b); b 197 net/netfilter/ipset/ip_set_bitmap_ip.c bitmap_ip_same_set(const struct ip_set *a, const struct ip_set *b) b 200 net/netfilter/ipset/ip_set_bitmap_ip.c const struct bitmap_ip *y = b->data; b 205 net/netfilter/ipset/ip_set_bitmap_ip.c a->timeout == b->timeout && b 206 net/netfilter/ipset/ip_set_bitmap_ip.c a->extensions == b->extensions; b 281 net/netfilter/ipset/ip_set_bitmap_ipmac.c bitmap_ipmac_same_set(const struct ip_set *a, const struct ip_set *b) b 284 net/netfilter/ipset/ip_set_bitmap_ipmac.c const struct bitmap_ipmac *y = b->data; b 288 net/netfilter/ipset/ip_set_bitmap_ipmac.c a->timeout == b->timeout && b 289 net/netfilter/ipset/ip_set_bitmap_ipmac.c a->extensions == b->extensions; b 183 net/netfilter/ipset/ip_set_bitmap_port.c bitmap_port_same_set(const struct ip_set *a, const struct ip_set *b) b 186 net/netfilter/ipset/ip_set_bitmap_port.c const struct bitmap_port *y = b->data; b 190 net/netfilter/ipset/ip_set_bitmap_port.c a->timeout == b->timeout && b 191 net/netfilter/ipset/ip_set_bitmap_port.c a->extensions == b->extensions; b 44 net/netfilter/ipset/ip_set_core.c #define STRNCMP(a, b) (strncmp(a, b, IPSET_MAXNAMELEN) == 0) b 495 net/netfilter/ipset/ip_set_hash_gen.h mtype_same_set(const struct ip_set *a, const struct ip_set *b) b 498 net/netfilter/ipset/ip_set_hash_gen.h const struct htype *y = b->data; b 502 net/netfilter/ipset/ip_set_hash_gen.h a->timeout == b->timeout && b 509 net/netfilter/ipset/ip_set_hash_gen.h a->extensions == b->extensions; b 41 net/netfilter/ipset/ip_set_hash_netiface.c #define STRLCPY(a, b) strlcpy(a, b, IFNAMSIZ) b 538 net/netfilter/ipset/ip_set_list_set.c list_set_same_set(const struct ip_set *a, const struct ip_set *b) b 541 net/netfilter/ipset/ip_set_list_set.c const struct list_set *y = b->data; b 544 net/netfilter/ipset/ip_set_list_set.c a->timeout == b->timeout && b 545 net/netfilter/ipset/ip_set_list_set.c a->extensions == b->extensions; b 142 net/netfilter/ipset/pfxlen.c #define E(a, b, c, d) \ b 144 net/netfilter/ipset/pfxlen.c htonl(a), htonl(b), \ b 157 net/netfilter/ipset/pfxlen.c #define E(a, b, c, d) \ b 158 net/netfilter/ipset/pfxlen.c {.ip6 = { (__force __be32)a, (__force __be32)b, \ b 101 net/netfilter/ipvs/ip_vs_dh.c struct ip_vs_dh_bucket *b; b 106 net/netfilter/ipvs/ip_vs_dh.c b = &s->buckets[0]; b 110 net/netfilter/ipvs/ip_vs_dh.c dest = rcu_dereference_protected(b->dest, 1); b 114 net/netfilter/ipvs/ip_vs_dh.c RCU_INIT_POINTER(b->dest, NULL); b 121 net/netfilter/ipvs/ip_vs_dh.c RCU_INIT_POINTER(b->dest, dest); b 125 net/netfilter/ipvs/ip_vs_dh.c b++; b 137 net/netfilter/ipvs/ip_vs_dh.c struct ip_vs_dh_bucket *b; b 140 net/netfilter/ipvs/ip_vs_dh.c b = &s->buckets[0]; b 142 net/netfilter/ipvs/ip_vs_dh.c dest = rcu_dereference_protected(b->dest, 1); b 145 net/netfilter/ipvs/ip_vs_dh.c RCU_INIT_POINTER(b->dest, NULL); b 147 net/netfilter/ipvs/ip_vs_dh.c b++; b 166 net/netfilter/ipvs/ip_vs_sh.c struct ip_vs_sh_bucket *b; b 172 net/netfilter/ipvs/ip_vs_sh.c b = &s->buckets[0]; b 177 net/netfilter/ipvs/ip_vs_sh.c dest = rcu_dereference_protected(b->dest, 1); b 181 net/netfilter/ipvs/ip_vs_sh.c RCU_INIT_POINTER(b->dest, NULL); b 188 net/netfilter/ipvs/ip_vs_sh.c RCU_INIT_POINTER(b->dest, dest); b 201 net/netfilter/ipvs/ip_vs_sh.c b++; b 213 net/netfilter/ipvs/ip_vs_sh.c struct ip_vs_sh_bucket *b; b 216 net/netfilter/ipvs/ip_vs_sh.c b = &s->buckets[0]; b 218 net/netfilter/ipvs/ip_vs_sh.c dest = rcu_dereference_protected(b->dest, 1); b 221 net/netfilter/ipvs/ip_vs_sh.c RCU_INIT_POINTER(b->dest, NULL); b 223 net/netfilter/ipvs/ip_vs_sh.c b++; b 80 net/netfilter/nf_conncount.c static int key_diff(const u32 *a, const u32 *b, unsigned int klen) b 82 net/netfilter/nf_conncount.c return memcmp(a, b, klen * sizeof(u32)); b 101 net/netfilter/nf_conncount.c unsigned long a, b; b 108 net/netfilter/nf_conncount.c b = conn->jiffies32; b 116 net/netfilter/nf_conncount.c age = a - b; b 465 net/netfilter/nf_conntrack_core.c unsigned long a, b, c, d; b 470 net/netfilter/nf_conntrack_core.c b = (unsigned long)ct->master; b 476 net/netfilter/nf_conntrack_core.c return siphash_4u64((u64)a, (u64)b, (u64)c, (u64)d, &ct_id_seed); b 478 net/netfilter/nf_conntrack_core.c return siphash_4u32((u32)a, (u32)b, (u32)c, (u32)d, &ct_id_seed); b 232 net/netfilter/nf_conntrack_expect.c const struct nf_conntrack_expect *b) b 239 net/netfilter/nf_conntrack_expect.c intersect_mask.src.u.all = a->mask.src.u.all & b->mask.src.u.all; b 243 net/netfilter/nf_conntrack_expect.c a->mask.src.u3.all[count] & b->mask.src.u3.all[count]; b 246 net/netfilter/nf_conntrack_expect.c return nf_ct_tuple_mask_cmp(&a->tuple, &b->tuple, &intersect_mask) && b 247 net/netfilter/nf_conntrack_expect.c net_eq(nf_ct_net(a->master), nf_ct_net(b->master)) && b 248 net/netfilter/nf_conntrack_expect.c nf_ct_zone_equal_any(a->master, nf_ct_zone(b->master)); b 252 net/netfilter/nf_conntrack_expect.c const struct nf_conntrack_expect *b) b 254 net/netfilter/nf_conntrack_expect.c return nf_ct_tuple_equal(&a->tuple, &b->tuple) && b 255 net/netfilter/nf_conntrack_expect.c nf_ct_tuple_mask_equal(&a->mask, &b->mask) && b 256 net/netfilter/nf_conntrack_expect.c net_eq(nf_ct_net(a->master), nf_ct_net(b->master)) && b 257 net/netfilter/nf_conntrack_expect.c nf_ct_zone_equal_any(a->master, nf_ct_zone(b->master)); b 261 net/netfilter/nf_conntrack_expect.c const struct nf_conntrack_expect *b, b 267 net/netfilter/nf_conntrack_expect.c return a->master == b->master; b 101 net/netfilter/nf_conntrack_h323_asn1.c #define INC_BITS(bs,b) if(((bs)->bit+=(b))>7){(bs)->cur+=(bs)->bit>>3;(bs)->bit&=7;} b 105 net/netfilter/nf_conntrack_h323_asn1.c static unsigned int get_bits(struct bitstr *bs, unsigned int b); b 106 net/netfilter/nf_conntrack_h323_asn1.c static unsigned int get_bitmap(struct bitstr *bs, unsigned int b); b 107 net/netfilter/nf_conntrack_h323_asn1.c static unsigned int get_uint(struct bitstr *bs, int b); b 180 net/netfilter/nf_conntrack_h323_asn1.c unsigned int b = (*bs->cur) & (0x80 >> bs->bit); b 184 net/netfilter/nf_conntrack_h323_asn1.c return b; b 188 net/netfilter/nf_conntrack_h323_asn1.c static unsigned int get_bits(struct bitstr *bs, unsigned int b) b 193 net/netfilter/nf_conntrack_h323_asn1.c l = b + bs->bit; b 213 net/netfilter/nf_conntrack_h323_asn1.c static unsigned int get_bitmap(struct bitstr *bs, unsigned int b) b 217 net/netfilter/nf_conntrack_h323_asn1.c if (!b) b 220 net/netfilter/nf_conntrack_h323_asn1.c l = bs->bit + b; b 244 net/netfilter/nf_conntrack_h323_asn1.c v &= 0xffffffff << (32 - b); b 252 net/netfilter/nf_conntrack_h323_asn1.c static unsigned int get_uint(struct bitstr *bs, int b) b 256 net/netfilter/nf_conntrack_h323_asn1.c switch (b) { b 325 net/netfilter/nf_conntrack_netlink.c #define ctnetlink_dump_mark(a, b) (0) b 354 net/netfilter/nf_conntrack_netlink.c #define ctnetlink_dump_secctx(a, b) (0) b 387 net/netfilter/nf_conntrack_netlink.c #define ctnetlink_dump_labels(a, b) (0) b 2711 net/netfilter/nf_conntrack_netlink.c unsigned long a, b, c, d; b 2716 net/netfilter/nf_conntrack_netlink.c b = (unsigned long)exp->helper; b 2721 net/netfilter/nf_conntrack_netlink.c return (__force __be32)siphash_4u64((u64)a, (u64)b, (u64)c, (u64)d, &exp_id_seed); b 2723 net/netfilter/nf_conntrack_netlink.c return (__force __be32)siphash_4u32((u32)a, (u32)b, (u32)c, (u32)d, &exp_id_seed); b 4011 net/netfilter/nf_tables_api.c unsigned char *b = skb_tail_pointer(skb); b 4076 net/netfilter/nf_tables_api.c nlmsg_trim(skb, b); b 178 net/netfilter/xt_hashlimit.c const struct dsthash_dst *b) b 180 net/netfilter/xt_hashlimit.c return !memcmp(&ent->dst, b, sizeof(ent->dst)); b 53 net/netfilter/xt_iprange.c iprange_ipv6_lt(const struct in6_addr *a, const struct in6_addr *b) b 58 net/netfilter/xt_iprange.c if (a->s6_addr32[i] != b->s6_addr32[i]) b 59 net/netfilter/xt_iprange.c return ntohl(a->s6_addr32[i]) < ntohl(b->s6_addr32[i]); b 38 net/netfilter/xt_set.c #define ADT_OPT(n, f, d, fs, cfs, t, p, b, po, bo) \ b 46 net/netfilter/xt_set.c .ext.bytes = b, \ b 596 net/openvswitch/flow_table.c const struct sw_flow_mask *b) b 599 net/openvswitch/flow_table.c const u8 *b_ = (const u8 *)&b->key + b->range.start; b 601 net/openvswitch/flow_table.c return (a->range.end == b->range.end) b 602 net/openvswitch/flow_table.c && (a->range.start == b->range.start) b 293 net/rds/threads.c u32 a, b; b 299 net/rds/threads.c b = ntohl(addr2->s6_addr32[i]); b 300 net/rds/threads.c if (a < b) b 302 net/rds/threads.c else if (a > b) b 30 net/rxrpc/call_accept.c struct rxrpc_backlog *b, b 55 net/rxrpc/call_accept.c call_head = b->call_backlog_head; b 56 net/rxrpc/call_accept.c call_tail = READ_ONCE(b->call_backlog_tail); b 64 net/rxrpc/call_accept.c head = b->peer_backlog_head; b 65 net/rxrpc/call_accept.c tail = READ_ONCE(b->peer_backlog_tail); b 70 net/rxrpc/call_accept.c b->peer_backlog[head] = peer; b 71 net/rxrpc/call_accept.c smp_store_release(&b->peer_backlog_head, b 75 net/rxrpc/call_accept.c head = b->conn_backlog_head; b 76 net/rxrpc/call_accept.c tail = READ_ONCE(b->conn_backlog_tail); b 83 net/rxrpc/call_accept.c b->conn_backlog[head] = conn; b 84 net/rxrpc/call_accept.c smp_store_release(&b->conn_backlog_head, b 142 net/rxrpc/call_accept.c b->call_backlog[call_head] = call; b 143 net/rxrpc/call_accept.c smp_store_release(&b->call_backlog_head, (call_head + 1) & (size - 1)); b 163 net/rxrpc/call_accept.c struct rxrpc_backlog *b = rx->backlog; b 165 net/rxrpc/call_accept.c if (!b) { b 166 net/rxrpc/call_accept.c b = kzalloc(sizeof(struct rxrpc_backlog), gfp); b 167 net/rxrpc/call_accept.c if (!b) b 169 net/rxrpc/call_accept.c rx->backlog = b; b 175 net/rxrpc/call_accept.c while (rxrpc_service_prealloc_one(rx, b, NULL, NULL, 0, gfp, b 187 net/rxrpc/call_accept.c struct rxrpc_backlog *b = rx->backlog; b 191 net/rxrpc/call_accept.c if (!b) b 201 net/rxrpc/call_accept.c head = b->peer_backlog_head; b 202 net/rxrpc/call_accept.c tail = b->peer_backlog_tail; b 204 net/rxrpc/call_accept.c struct rxrpc_peer *peer = b->peer_backlog[tail]; b 209 net/rxrpc/call_accept.c head = b->conn_backlog_head; b 210 net/rxrpc/call_accept.c tail = b->conn_backlog_tail; b 212 net/rxrpc/call_accept.c struct rxrpc_connection *conn = b->conn_backlog[tail]; b 223 net/rxrpc/call_accept.c head = b->call_backlog_head; b 224 net/rxrpc/call_accept.c tail = b->call_backlog_tail; b 226 net/rxrpc/call_accept.c struct rxrpc_call *call = b->call_backlog[tail]; b 239 net/rxrpc/call_accept.c kfree(b); b 270 net/rxrpc/call_accept.c struct rxrpc_backlog *b = rx->backlog; b 277 net/rxrpc/call_accept.c call_head = smp_load_acquire(&b->call_backlog_head); b 278 net/rxrpc/call_accept.c call_tail = b->call_backlog_tail; b 280 net/rxrpc/call_accept.c conn_head = smp_load_acquire(&b->conn_backlog_head); b 281 net/rxrpc/call_accept.c conn_tail = b->conn_backlog_tail; b 284 net/rxrpc/call_accept.c peer_head = smp_load_acquire(&b->peer_backlog_head); b 285 net/rxrpc/call_accept.c peer_tail = b->peer_backlog_tail; b 296 net/rxrpc/call_accept.c peer = b->peer_backlog[peer_tail]; b 299 net/rxrpc/call_accept.c b->peer_backlog[peer_tail] = NULL; b 300 net/rxrpc/call_accept.c smp_store_release(&b->peer_backlog_tail, b 308 net/rxrpc/call_accept.c conn = b->conn_backlog[conn_tail]; b 309 net/rxrpc/call_accept.c b->conn_backlog[conn_tail] = NULL; b 310 net/rxrpc/call_accept.c smp_store_release(&b->conn_backlog_tail, b 321 net/rxrpc/call_accept.c call = b->call_backlog[call_tail]; b 322 net/rxrpc/call_accept.c b->call_backlog[call_tail] = NULL; b 323 net/rxrpc/call_accept.c smp_store_release(&b->call_backlog_tail, b 661 net/rxrpc/call_accept.c struct rxrpc_backlog *b = rx->backlog; b 666 net/rxrpc/call_accept.c return rxrpc_service_prealloc_one(rx, b, notify_rx, b 757 net/sched/act_api.c unsigned char *b = skb_tail_pointer(skb); b 786 net/sched/act_api.c nlmsg_trim(skb, b); b 1044 net/sched/act_api.c unsigned char *b = skb_tail_pointer(skb); b 1064 net/sched/act_api.c nlh->nlmsg_len = skb_tail_pointer(skb) - b; b 1068 net/sched/act_api.c nlmsg_trim(skb, b); b 1141 net/sched/act_api.c unsigned char *b; b 1155 net/sched/act_api.c b = skb_tail_pointer(skb); b 1195 net/sched/act_api.c nlh->nlmsg_len = skb_tail_pointer(skb) - b; b 1459 net/sched/act_api.c unsigned char *b = skb_tail_pointer(skb); b 1529 net/sched/act_api.c nlmsg_trim(skb, b); b 1531 net/sched/act_api.c nlh->nlmsg_len = skb_tail_pointer(skb) - b; b 1539 net/sched/act_api.c nlmsg_trim(skb, b); b 172 net/sched/act_connmark.c unsigned char *b = skb_tail_pointer(skb); b 197 net/sched/act_connmark.c nlmsg_trim(skb, b); b 635 net/sched/act_csum.c unsigned char *b = skb_tail_pointer(skb); b 663 net/sched/act_csum.c nlmsg_trim(skb, b); b 832 net/sched/act_ct.c unsigned char *b = skb_tail_pointer(skb); b 893 net/sched/act_ct.c nlmsg_trim(skb, b); b 291 net/sched/act_ctinfo.c unsigned char *b = skb_tail_pointer(skb); b 342 net/sched/act_ctinfo.c nlmsg_trim(skb, b); b 195 net/sched/act_gact.c unsigned char *b = skb_tail_pointer(skb); b 229 net/sched/act_gact.c nlmsg_trim(skb, b); b 378 net/sched/act_ife.c unsigned char *b = skb_tail_pointer(skb); b 402 net/sched/act_ife.c nlmsg_trim(skb, b); b 621 net/sched/act_ife.c unsigned char *b = skb_tail_pointer(skb); b 667 net/sched/act_ife.c nlmsg_trim(skb, b); b 278 net/sched/act_ipt.c unsigned char *b = skb_tail_pointer(skb); b 315 net/sched/act_ipt.c nlmsg_trim(skb, b); b 331 net/sched/act_mirred.c unsigned char *b = skb_tail_pointer(skb); b 360 net/sched/act_mirred.c nlmsg_trim(skb, b); b 300 net/sched/act_mpls.c unsigned char *b = skb_tail_pointer(skb); b 347 net/sched/act_mpls.c nlmsg_trim(skb, b); b 265 net/sched/act_nat.c unsigned char *b = skb_tail_pointer(skb); b 293 net/sched/act_nat.c nlmsg_trim(skb, b); b 414 net/sched/act_pedit.c unsigned char *b = skb_tail_pointer(skb); b 460 net/sched/act_pedit.c nlmsg_trim(skb, b); b 307 net/sched/act_police.c unsigned char *b = skb_tail_pointer(skb); b 360 net/sched/act_police.c nlmsg_trim(skb, b); b 202 net/sched/act_sample.c unsigned char *b = skb_tail_pointer(skb); b 235 net/sched/act_sample.c nlmsg_trim(skb, b); b 173 net/sched/act_simple.c unsigned char *b = skb_tail_pointer(skb); b 197 net/sched/act_simple.c nlmsg_trim(skb, b); b 231 net/sched/act_skbedit.c unsigned char *b = skb_tail_pointer(skb); b 279 net/sched/act_skbedit.c nlmsg_trim(skb, b); b 218 net/sched/act_skbmod.c unsigned char *b = skb_tail_pointer(skb); b 252 net/sched/act_skbmod.c nlmsg_trim(skb, b); b 515 net/sched/act_tunnel_key.c unsigned char *b = skb_tail_pointer(skb); b 569 net/sched/act_tunnel_key.c nlmsg_trim(skb, b); b 255 net/sched/act_vlan.c unsigned char *b = skb_tail_pointer(skb); b 290 net/sched/act_vlan.c nlmsg_trim(skb, b); b 1789 net/sched/cls_api.c unsigned char *b = skb_tail_pointer(skb); b 1817 net/sched/cls_api.c nlh->nlmsg_len = skb_tail_pointer(skb) - b; b 1822 net/sched/cls_api.c nlmsg_trim(skb, b); b 2611 net/sched/cls_api.c unsigned char *b = skb_tail_pointer(skb); b 2646 net/sched/cls_api.c nlh->nlmsg_len = skb_tail_pointer(skb) - b; b 2651 net/sched/cls_api.c nlmsg_trim(skb, b); b 129 net/sched/cls_route.c struct route4_bucket *b; b 162 net/sched/cls_route.c b = rcu_dereference_bh(head->table[h]); b 163 net/sched/cls_route.c if (b) { b 164 net/sched/cls_route.c for (f = rcu_dereference_bh(b->ht[route4_hash_from(id)]); b 170 net/sched/cls_route.c for (f = rcu_dereference_bh(b->ht[route4_hash_iif(iif)]); b 176 net/sched/cls_route.c for (f = rcu_dereference_bh(b->ht[route4_hash_wild()]); b 218 net/sched/cls_route.c struct route4_bucket *b; b 230 net/sched/cls_route.c b = rtnl_dereference(head->table[h1]); b 231 net/sched/cls_route.c if (b) { b 232 net/sched/cls_route.c for (f = rtnl_dereference(b->ht[h2]); b 285 net/sched/cls_route.c struct route4_bucket *b; b 287 net/sched/cls_route.c b = rtnl_dereference(head->table[h1]); b 288 net/sched/cls_route.c if (b) { b 292 net/sched/cls_route.c while ((f = rtnl_dereference(b->ht[h2])) != NULL) { b 296 net/sched/cls_route.c RCU_INIT_POINTER(b->ht[h2], next); b 305 net/sched/cls_route.c kfree_rcu(b, rcu); b 318 net/sched/cls_route.c struct route4_bucket *b; b 326 net/sched/cls_route.c b = f->bkt; b 328 net/sched/cls_route.c fp = &b->ht[from_hash(h >> 16)]; b 350 net/sched/cls_route.c rt = rtnl_dereference(b->ht[i]); b 357 net/sched/cls_route.c kfree_rcu(b, rcu); b 390 net/sched/cls_route.c struct route4_bucket *b; b 428 net/sched/cls_route.c b = rtnl_dereference(head->table[h1]); b 429 net/sched/cls_route.c if (!b) { b 430 net/sched/cls_route.c b = kzalloc(sizeof(struct route4_bucket), GFP_KERNEL); b 431 net/sched/cls_route.c if (b == NULL) b 434 net/sched/cls_route.c rcu_assign_pointer(head->table[h1], b); b 438 net/sched/cls_route.c for (fp = rtnl_dereference(b->ht[h2]); b 454 net/sched/cls_route.c f->bkt = b; b 473 net/sched/cls_route.c struct route4_bucket *b; b 532 net/sched/cls_route.c b = rtnl_dereference(head->table[th]); b 533 net/sched/cls_route.c if (b) { b 534 net/sched/cls_route.c fp = &b->ht[h]; b 571 net/sched/cls_route.c struct route4_bucket *b = rtnl_dereference(head->table[h]); b 573 net/sched/cls_route.c if (b) { b 577 net/sched/cls_route.c for (f = rtnl_dereference(b->ht[h1]); b 411 net/sched/cls_rsvp.h u32 b = 1 << (data->tgenerator & 0x1F); b 413 net/sched/cls_rsvp.h if (data->tmap[n] & b) b 415 net/sched/cls_rsvp.h data->tmap[n] |= b; b 682 net/sched/em_meta.c static int meta_var_compare(struct meta_obj *a, struct meta_obj *b) b 684 net/sched/em_meta.c int r = a->len - b->len; b 687 net/sched/em_meta.c r = memcmp((void *) a->value, (void *) b->value, a->len); b 732 net/sched/em_meta.c static int meta_int_compare(struct meta_obj *a, struct meta_obj *b) b 737 net/sched/em_meta.c if (unlikely(a->value == b->value)) b 739 net/sched/em_meta.c else if (a->value < b->value) b 874 net/sched/sch_api.c unsigned char *b = skb_tail_pointer(skb); b 937 net/sched/sch_api.c nlh->nlmsg_len = skb_tail_pointer(skb) - b; b 942 net/sched/sch_api.c nlmsg_trim(skb, b); b 1680 net/sched/sch_api.c int b; b 1706 net/sched/sch_api.c hash_for_each(qdisc_dev(root)->qdisc_hash, b, q, hash) { b 1791 net/sched/sch_api.c unsigned char *b = skb_tail_pointer(skb); b 1822 net/sched/sch_api.c nlh->nlmsg_len = skb_tail_pointer(skb) - b; b 1827 net/sched/sch_api.c nlmsg_trim(skb, b); b 2162 net/sched/sch_api.c int b; b 2180 net/sched/sch_api.c hash_for_each(qdisc_dev(root)->qdisc_hash, b, q, hash) { b 146 net/sched/sch_cake.c u16 t:3, b:10; b 1364 net/sched/sch_cake.c q->tins[ii.t].overflow_idx[ii.b] = j; b 1365 net/sched/sch_cake.c q->tins[jj.t].overflow_idx[jj.b] = i; b 1372 net/sched/sch_cake.c return q->tins[ii.t].backlogs[ii.b]; b 1429 net/sched/sch_cake.c struct cake_tin_data *b, b 1439 net/sched/sch_cake.c u64 tin_dur = (len * b->tin_rate_ns) >> b->tin_rate_shft; b 1443 net/sched/sch_cake.c if (ktime_before(b->time_next_packet, now)) b 1444 net/sched/sch_cake.c b->time_next_packet = ktime_add_ns(b->time_next_packet, b 1447 net/sched/sch_cake.c else if (ktime_before(b->time_next_packet, b 1449 net/sched/sch_cake.c b->time_next_packet = ktime_add_ns(now, tin_dur); b 1467 net/sched/sch_cake.c struct cake_tin_data *b; b 1482 net/sched/sch_cake.c idx = qq.b; b 1484 net/sched/sch_cake.c b = &q->tins[tin]; b 1485 net/sched/sch_cake.c flow = &b->flows[idx]; b 1493 net/sched/sch_cake.c if (cobalt_queue_full(&flow->cvars, &b->cparams, now)) b 1494 net/sched/sch_cake.c b->unresponsive_flow_count++; b 1498 net/sched/sch_cake.c b->backlogs[idx] -= len; b 1499 net/sched/sch_cake.c b->tin_backlog -= len; b 1504 net/sched/sch_cake.c b->tin_dropped++; b 1508 net/sched/sch_cake.c cake_advance_shaper(q, b, skb, now, true); b 1638 net/sched/sch_cake.c struct cake_tin_data *b; b 1643 net/sched/sch_cake.c idx = cake_classify(sch, &b, skb, q->flow_mode, &ret); b 1651 net/sched/sch_cake.c flow = &b->flows[idx]; b 1654 net/sched/sch_cake.c if (!b->tin_backlog) { b 1655 net/sched/sch_cake.c if (ktime_before(b->time_next_packet, now)) b 1656 net/sched/sch_cake.c b->time_next_packet = now; b 1674 net/sched/sch_cake.c if (unlikely(len > b->max_skblen)) b 1675 net/sched/sch_cake.c b->max_skblen = len; b 1699 net/sched/sch_cake.c b->packets++; b 1704 net/sched/sch_cake.c b->bytes += slen; b 1705 net/sched/sch_cake.c b->backlogs[idx] += slen; b 1706 net/sched/sch_cake.c b->tin_backlog += slen; b 1722 net/sched/sch_cake.c b->ack_drops++; b 1724 net/sched/sch_cake.c b->bytes += qdisc_pkt_len(ack); b 1728 net/sched/sch_cake.c cake_advance_shaper(q, b, ack, now, true); b 1738 net/sched/sch_cake.c b->packets++; b 1739 net/sched/sch_cake.c b->bytes += len; b 1740 net/sched/sch_cake.c b->backlogs[idx] += len; b 1741 net/sched/sch_cake.c b->tin_backlog += len; b 1747 net/sched/sch_cake.c cake_heapify_up(q, b->overflow_idx[idx]); b 1770 net/sched/sch_cake.c u64 b = q->avg_window_bytes * (u64)NSEC_PER_SEC; b 1772 net/sched/sch_cake.c b = div64_u64(b, window_interval); b 1774 net/sched/sch_cake.c cake_ewma(q->avg_peak_bandwidth, b, b 1775 net/sched/sch_cake.c b > q->avg_peak_bandwidth ? 2 : 8); b 1793 net/sched/sch_cake.c struct cake_host *srchost = &b->hosts[flow->srchost]; b 1794 net/sched/sch_cake.c struct cake_host *dsthost = &b->hosts[flow->dsthost]; b 1798 net/sched/sch_cake.c list_add_tail(&flow->flowchain, &b->new_flows); b 1800 net/sched/sch_cake.c b->decaying_flow_count--; b 1801 net/sched/sch_cake.c list_move_tail(&flow->flowchain, &b->new_flows); b 1804 net/sched/sch_cake.c b->sparse_flow_count++; b 1812 net/sched/sch_cake.c flow->deficit = (b->flow_quantum * b 1815 net/sched/sch_cake.c struct cake_host *srchost = &b->hosts[flow->srchost]; b 1816 net/sched/sch_cake.c struct cake_host *dsthost = &b->hosts[flow->dsthost]; b 1822 net/sched/sch_cake.c b->sparse_flow_count--; b 1823 net/sched/sch_cake.c b->bulk_flow_count++; b 1843 net/sched/sch_cake.c b->drop_overlimit += dropped; b 1851 net/sched/sch_cake.c struct cake_tin_data *b = &q->tins[q->cur_tin]; b 1852 net/sched/sch_cake.c struct cake_flow *flow = &b->flows[q->cur_flow]; b 1859 net/sched/sch_cake.c b->backlogs[q->cur_flow] -= len; b 1860 net/sched/sch_cake.c b->tin_backlog -= len; b 1866 net/sched/sch_cake.c cake_heapify(q, b->overflow_idx[q->cur_flow]); b 1886 net/sched/sch_cake.c struct cake_tin_data *b = &q->tins[q->cur_tin]; b 1919 net/sched/sch_cake.c while (b->tin_deficit < 0 || b 1920 net/sched/sch_cake.c !(b->sparse_flow_count + b->bulk_flow_count)) { b 1921 net/sched/sch_cake.c if (b->tin_deficit <= 0) b 1922 net/sched/sch_cake.c b->tin_deficit += b->tin_quantum_band; b 1923 net/sched/sch_cake.c if (b->sparse_flow_count + b->bulk_flow_count) b 1927 net/sched/sch_cake.c b++; b 1930 net/sched/sch_cake.c b = q->tins; b 1953 net/sched/sch_cake.c b = q->tins + tin; b 1954 net/sched/sch_cake.c if ((b->sparse_flow_count + b->bulk_flow_count) > 0) { b 1956 net/sched/sch_cake.c ktime_sub(b->time_next_packet, now); b 1968 net/sched/sch_cake.c b = q->tins + best_tin; b 1971 net/sched/sch_cake.c if (unlikely(!(b->sparse_flow_count + b->bulk_flow_count))) b 1977 net/sched/sch_cake.c head = &b->decaying_flows; b 1979 net/sched/sch_cake.c head = &b->new_flows; b 1981 net/sched/sch_cake.c head = &b->old_flows; b 1983 net/sched/sch_cake.c head = &b->decaying_flows; b 1990 net/sched/sch_cake.c q->cur_flow = flow - b->flows; b 1994 net/sched/sch_cake.c srchost = &b->hosts[flow->srchost]; b 1995 net/sched/sch_cake.c dsthost = &b->hosts[flow->dsthost]; b 2006 net/sched/sch_cake.c b->sparse_flow_count--; b 2007 net/sched/sch_cake.c b->bulk_flow_count++; b 2036 net/sched/sch_cake.c flow->deficit += (b->flow_quantum * quantum_div[host_load] + b 2038 net/sched/sch_cake.c list_move_tail(&flow->flowchain, &b->old_flows); b 2048 net/sched/sch_cake.c if (cobalt_queue_empty(&flow->cvars, &b->cparams, now)) b 2049 net/sched/sch_cake.c b->unresponsive_flow_count--; b 2057 net/sched/sch_cake.c &b->decaying_flows); b 2059 net/sched/sch_cake.c b->bulk_flow_count--; b 2067 net/sched/sch_cake.c b->decaying_flow_count++; b 2070 net/sched/sch_cake.c b->sparse_flow_count--; b 2071 net/sched/sch_cake.c b->decaying_flow_count++; b 2079 net/sched/sch_cake.c b->sparse_flow_count--; b 2081 net/sched/sch_cake.c b->bulk_flow_count--; b 2090 net/sched/sch_cake.c b->decaying_flow_count--; b 2098 net/sched/sch_cake.c if (!cobalt_should_drop(&flow->cvars, &b->cparams, now, skb, b 2099 net/sched/sch_cake.c (b->bulk_flow_count * b 2107 net/sched/sch_cake.c len = cake_advance_shaper(q, b, skb, b 2110 net/sched/sch_cake.c b->tin_deficit -= len; b 2113 net/sched/sch_cake.c b->tin_dropped++; b 2121 net/sched/sch_cake.c b->tin_ecn_mark += !!flow->cvars.ecn_marked; b 2126 net/sched/sch_cake.c b->avge_delay = cake_ewma(b->avge_delay, delay, 8); b 2127 net/sched/sch_cake.c b->peak_delay = cake_ewma(b->peak_delay, delay, b 2128 net/sched/sch_cake.c delay > b->peak_delay ? 2 : 8); b 2129 net/sched/sch_cake.c b->base_delay = cake_ewma(b->base_delay, delay, b 2130 net/sched/sch_cake.c delay < b->base_delay ? 2 : 8); b 2132 net/sched/sch_cake.c len = cake_advance_shaper(q, b, skb, now, false); b 2134 net/sched/sch_cake.c b->tin_deficit -= len; b 2191 net/sched/sch_cake.c static void cake_set_rate(struct cake_tin_data *b, u64 rate, u32 mtu, b 2203 net/sched/sch_cake.c b->flow_quantum = 1514; b 2205 net/sched/sch_cake.c b->flow_quantum = max(min(rate >> 12, 1514ULL), 300ULL); b 2215 net/sched/sch_cake.c b->tin_rate_bps = rate; b 2216 net/sched/sch_cake.c b->tin_rate_ns = rate_ns; b 2217 net/sched/sch_cake.c b->tin_rate_shft = rate_shft; b 2221 net/sched/sch_cake.c b->cparams.target = max((byte_target_ns * 3) / 2, target_ns); b 2222 net/sched/sch_cake.c b->cparams.interval = max(rtt_est_ns + b 2223 net/sched/sch_cake.c b->cparams.target - target_ns, b 2224 net/sched/sch_cake.c b->cparams.target * 2); b 2225 net/sched/sch_cake.c b->cparams.mtu_time = byte_target_ns; b 2226 net/sched/sch_cake.c b->cparams.p_inc = 1 << 24; /* 1/256 */ b 2227 net/sched/sch_cake.c b->cparams.p_dec = 1 << 20; /* 1/4096 */ b 2233 net/sched/sch_cake.c struct cake_tin_data *b = &q->tins[0]; b 2242 net/sched/sch_cake.c cake_set_rate(b, rate, mtu, b 2244 net/sched/sch_cake.c b->tin_quantum_band = 65535; b 2245 net/sched/sch_cake.c b->tin_quantum_prio = 65535; b 2265 net/sched/sch_cake.c struct cake_tin_data *b = &q->tins[i]; b 2267 net/sched/sch_cake.c cake_set_rate(b, rate, mtu, us_to_ns(q->target), b 2270 net/sched/sch_cake.c b->tin_quantum_prio = max_t(u16, 1U, quantum1); b 2271 net/sched/sch_cake.c b->tin_quantum_band = max_t(u16, 1U, quantum2); b 2362 net/sched/sch_cake.c struct cake_tin_data *b = &q->tins[i]; b 2364 net/sched/sch_cake.c cake_set_rate(b, rate, mtu, us_to_ns(q->target), b 2367 net/sched/sch_cake.c b->tin_quantum_prio = max_t(u16, 1U, quantum1); b 2368 net/sched/sch_cake.c b->tin_quantum_band = max_t(u16, 1U, quantum2); b 2702 net/sched/sch_cake.c struct cake_tin_data *b = q->tins + i; b 2704 net/sched/sch_cake.c INIT_LIST_HEAD(&b->new_flows); b 2705 net/sched/sch_cake.c INIT_LIST_HEAD(&b->old_flows); b 2706 net/sched/sch_cake.c INIT_LIST_HEAD(&b->decaying_flows); b 2707 net/sched/sch_cake.c b->sparse_flow_count = 0; b 2708 net/sched/sch_cake.c b->bulk_flow_count = 0; b 2709 net/sched/sch_cake.c b->decaying_flow_count = 0; b 2712 net/sched/sch_cake.c struct cake_flow *flow = b->flows + j; b 2719 net/sched/sch_cake.c q->overflow_heap[k].b = j; b 2720 net/sched/sch_cake.c b->overflow_idx[j] = k; b 2856 net/sched/sch_cake.c struct cake_tin_data *b = &q->tins[q->tin_order[i]]; b 2862 net/sched/sch_cake.c PUT_TSTAT_U64(THRESHOLD_RATE64, b->tin_rate_bps); b 2863 net/sched/sch_cake.c PUT_TSTAT_U64(SENT_BYTES64, b->bytes); b 2864 net/sched/sch_cake.c PUT_TSTAT_U32(BACKLOG_BYTES, b->tin_backlog); b 2867 net/sched/sch_cake.c ktime_to_us(ns_to_ktime(b->cparams.target))); b 2869 net/sched/sch_cake.c ktime_to_us(ns_to_ktime(b->cparams.interval))); b 2871 net/sched/sch_cake.c PUT_TSTAT_U32(SENT_PACKETS, b->packets); b 2872 net/sched/sch_cake.c PUT_TSTAT_U32(DROPPED_PACKETS, b->tin_dropped); b 2873 net/sched/sch_cake.c PUT_TSTAT_U32(ECN_MARKED_PACKETS, b->tin_ecn_mark); b 2874 net/sched/sch_cake.c PUT_TSTAT_U32(ACKS_DROPPED_PACKETS, b->ack_drops); b 2877 net/sched/sch_cake.c ktime_to_us(ns_to_ktime(b->peak_delay))); b 2879 net/sched/sch_cake.c ktime_to_us(ns_to_ktime(b->avge_delay))); b 2881 net/sched/sch_cake.c ktime_to_us(ns_to_ktime(b->base_delay))); b 2883 net/sched/sch_cake.c PUT_TSTAT_U32(WAY_INDIRECT_HITS, b->way_hits); b 2884 net/sched/sch_cake.c PUT_TSTAT_U32(WAY_MISSES, b->way_misses); b 2885 net/sched/sch_cake.c PUT_TSTAT_U32(WAY_COLLISIONS, b->way_collisions); b 2887 net/sched/sch_cake.c PUT_TSTAT_U32(SPARSE_FLOWS, b->sparse_flow_count + b 2888 net/sched/sch_cake.c b->decaying_flow_count); b 2889 net/sched/sch_cake.c PUT_TSTAT_U32(BULK_FLOWS, b->bulk_flow_count); b 2890 net/sched/sch_cake.c PUT_TSTAT_U32(UNRESPONSIVE_FLOWS, b->unresponsive_flow_count); b 2891 net/sched/sch_cake.c PUT_TSTAT_U32(MAX_SKBLEN, b->max_skblen); b 2893 net/sched/sch_cake.c PUT_TSTAT_U32(FLOW_QUANTUM, b->flow_quantum); b 2955 net/sched/sch_cake.c const struct cake_tin_data *b = \ b 2959 net/sched/sch_cake.c flow = &b->flows[idx % CAKE_QUEUES]; b 2970 net/sched/sch_cake.c qs.backlog = b->backlogs[idx % CAKE_QUEUES]; b 3028 net/sched/sch_cake.c struct cake_tin_data *b = &q->tins[q->tin_order[i]]; b 3031 net/sched/sch_cake.c if (list_empty(&b->flows[j].flowchain) || b 428 net/sched/sch_cbq.c struct cbq_class *b; b 431 net/sched/sch_cbq.c for (b = cl->borrow; b; b = b->borrow) { b 432 net/sched/sch_cbq.c delay = b->undertime - q->now; b 1234 net/sched/sch_cbq.c unsigned char *b = skb_tail_pointer(skb); b 1241 net/sched/sch_cbq.c nlmsg_trim(skb, b); b 1247 net/sched/sch_cbq.c unsigned char *b = skb_tail_pointer(skb); b 1267 net/sched/sch_cbq.c nlmsg_trim(skb, b); b 1273 net/sched/sch_cbq.c unsigned char *b = skb_tail_pointer(skb); b 1287 net/sched/sch_cbq.c nlmsg_trim(skb, b); b 1293 net/sched/sch_cbq.c unsigned char *b = skb_tail_pointer(skb); b 1306 net/sched/sch_cbq.c nlmsg_trim(skb, b); b 1518 net/sched/sch_hfsc.c unsigned char *b = skb_tail_pointer(skb); b 1527 net/sched/sch_hfsc.c nlmsg_trim(skb, b); b 109 net/sched/sch_hhf.c #define hhf_time_before(a, b) \ b 110 net/sched/sch_hhf.c (typecheck(u32, a) && typecheck(u32, b) && ((s32)((a) - (b)) < 0)) b 265 net/sched/sch_multiq.c unsigned char *b = skb_tail_pointer(skb); b 277 net/sched/sch_multiq.c nlmsg_trim(skb, b); b 267 net/sched/sch_prio.c unsigned char *b = skb_tail_pointer(skb); b 284 net/sched/sch_prio.c nlmsg_trim(skb, b); b 717 net/sched/sch_qfq.c static inline int qfq_gt(u64 a, u64 b) b 719 net/sched/sch_qfq.c return (s64)(a - b) > 0; b 126 net/sched/sch_sfb.c struct sfb_bucket *b = &q->bins[slot].bins[0][0]; b 132 net/sched/sch_sfb.c if (b[hash].qlen < 0xFFFF) b 133 net/sched/sch_sfb.c b[hash].qlen++; b 134 net/sched/sch_sfb.c b += SFB_NUMBUCKETS; /* next level */ b 155 net/sched/sch_sfb.c struct sfb_bucket *b = &q->bins[slot].bins[0][0]; b 161 net/sched/sch_sfb.c if (b[hash].qlen > 0) b 162 net/sched/sch_sfb.c b[hash].qlen--; b 163 net/sched/sch_sfb.c b += SFB_NUMBUCKETS; /* next level */ b 180 net/sched/sch_sfb.c static void decrement_prob(struct sfb_bucket *b, struct sfb_sched_data *q) b 182 net/sched/sch_sfb.c b->p_mark = prob_minus(b->p_mark, q->decrement); b 185 net/sched/sch_sfb.c static void increment_prob(struct sfb_bucket *b, struct sfb_sched_data *q) b 187 net/sched/sch_sfb.c b->p_mark = prob_plus(b->p_mark, q->increment); b 202 net/sched/sch_sfb.c const struct sfb_bucket *b = &q->bins[q->slot].bins[0][0]; b 205 net/sched/sch_sfb.c if (qlen < b->qlen) b 206 net/sched/sch_sfb.c qlen = b->qlen; b 207 net/sched/sch_sfb.c totalpm += b->p_mark; b 208 net/sched/sch_sfb.c if (prob < b->p_mark) b 209 net/sched/sch_sfb.c prob = b->p_mark; b 210 net/sched/sch_sfb.c b++; b 330 net/sched/sch_sfb.c struct sfb_bucket *b = &q->bins[slot].bins[i][hash]; b 333 net/sched/sch_sfb.c if (b->qlen == 0) b 334 net/sched/sch_sfb.c decrement_prob(b, q); b 335 net/sched/sch_sfb.c else if (b->qlen >= q->bin_size) b 336 net/sched/sch_sfb.c increment_prob(b, q); b 337 net/sched/sch_sfb.c if (minqlen > b->qlen) b 338 net/sched/sch_sfb.c minqlen = b->qlen; b 339 net/sched/sch_sfb.c if (p_min > b->p_mark) b 340 net/sched/sch_sfb.c p_min = b->p_mark; b 363 net/sched/sch_sfb.c struct sfb_bucket *b = &q->bins[slot].bins[i][hash]; b 366 net/sched/sch_sfb.c if (b->qlen == 0) b 367 net/sched/sch_sfb.c decrement_prob(b, q); b 368 net/sched/sch_sfb.c else if (b->qlen >= q->bin_size) b 369 net/sched/sch_sfb.c increment_prob(b, q); b 791 net/sched/sch_sfq.c unsigned char *b = skb_tail_pointer(skb); b 821 net/sched/sch_sfq.c nlmsg_trim(skb, b); b 67 net/sunrpc/auth_gss/svcauth_gss.c static int netobj_equal(struct xdr_netobj *a, struct xdr_netobj *b) b 69 net/sunrpc/auth_gss/svcauth_gss.c return a->len == b->len && 0 == memcmp(a->data, b->data, a->len); b 115 net/sunrpc/auth_gss/svcauth_gss.c static int rsi_match(struct cache_head *a, struct cache_head *b) b 118 net/sunrpc/auth_gss/svcauth_gss.c struct rsi *tmp = container_of(b, struct rsi, h); b 382 net/sunrpc/auth_gss/svcauth_gss.c rsc_match(struct cache_head *a, struct cache_head *b) b 385 net/sunrpc/auth_gss/svcauth_gss.c struct rsc *tmp = container_of(b, struct rsc, h); b 214 net/sunrpc/stats.c static void _add_rpc_iostats(struct rpc_iostats *a, struct rpc_iostats *b) b 216 net/sunrpc/stats.c a->om_ops += b->om_ops; b 217 net/sunrpc/stats.c a->om_ntrans += b->om_ntrans; b 218 net/sunrpc/stats.c a->om_timeouts += b->om_timeouts; b 219 net/sunrpc/stats.c a->om_bytes_sent += b->om_bytes_sent; b 220 net/sunrpc/stats.c a->om_bytes_recv += b->om_bytes_recv; b 221 net/sunrpc/stats.c a->om_queue = ktime_add(a->om_queue, b->om_queue); b 222 net/sunrpc/stats.c a->om_rtt = ktime_add(a->om_rtt, b->om_rtt); b 223 net/sunrpc/stats.c a->om_execute = ktime_add(a->om_execute, b->om_execute); b 224 net/sunrpc/stats.c a->om_error_status += b->om_error_status; b 68 net/tipc/bearer.c static void bearer_disable(struct net *net, struct tipc_bearer *b); b 175 net/tipc/bearer.c struct tipc_bearer *b; b 179 net/tipc/bearer.c b = rtnl_dereference(tn->bearer_list[i]); b 180 net/tipc/bearer.c if (b && (!strcmp(b->name, name))) b 181 net/tipc/bearer.c return b; b 194 net/tipc/bearer.c struct tipc_bearer *b; b 199 net/tipc/bearer.c b = rtnl_dereference(tn->bearer_list[bearer_id]); b 200 net/tipc/bearer.c if (!b) b 203 net/tipc/bearer.c strcpy(name, b->name); b 210 net/tipc/bearer.c struct tipc_bearer *b; b 213 net/tipc/bearer.c b = rcu_dereference(tn->bearer_list[bearer_id]); b 214 net/tipc/bearer.c if (b) b 215 net/tipc/bearer.c tipc_disc_add_dest(b->disc); b 222 net/tipc/bearer.c struct tipc_bearer *b; b 225 net/tipc/bearer.c b = rcu_dereference(tn->bearer_list[bearer_id]); b 226 net/tipc/bearer.c if (b) b 227 net/tipc/bearer.c tipc_disc_remove_dest(b->disc); b 241 net/tipc/bearer.c struct tipc_bearer *b; b 269 net/tipc/bearer.c b = rtnl_dereference(tn->bearer_list[bearer_id]); b 270 net/tipc/bearer.c if (!b) b 272 net/tipc/bearer.c if (!strcmp(name, b->name)) { b 277 net/tipc/bearer.c if (b->priority != prio) b 298 net/tipc/bearer.c b = kzalloc(sizeof(*b), GFP_ATOMIC); b 299 net/tipc/bearer.c if (!b) b 302 net/tipc/bearer.c strcpy(b->name, name); b 303 net/tipc/bearer.c b->media = m; b 304 net/tipc/bearer.c res = m->enable_media(net, b, attr); b 306 net/tipc/bearer.c kfree(b); b 311 net/tipc/bearer.c b->identity = bearer_id; b 312 net/tipc/bearer.c b->tolerance = m->tolerance; b 313 net/tipc/bearer.c b->window = m->window; b 314 net/tipc/bearer.c b->domain = disc_domain; b 315 net/tipc/bearer.c b->net_plane = bearer_id + 'A'; b 316 net/tipc/bearer.c b->priority = prio; b 317 net/tipc/bearer.c test_and_set_bit_lock(0, &b->up); b 319 net/tipc/bearer.c res = tipc_disc_create(net, b, &b->bcast_addr, &skb); b 321 net/tipc/bearer.c bearer_disable(net, b); b 326 net/tipc/bearer.c rcu_assign_pointer(tn->bearer_list[bearer_id], b); b 328 net/tipc/bearer.c tipc_bearer_xmit_skb(net, bearer_id, skb, &b->bcast_addr); b 331 net/tipc/bearer.c bearer_disable(net, b); b 346 net/tipc/bearer.c static int tipc_reset_bearer(struct net *net, struct tipc_bearer *b) b 348 net/tipc/bearer.c pr_info("Resetting bearer <%s>\n", b->name); b 349 net/tipc/bearer.c tipc_node_delete_links(net, b->identity); b 350 net/tipc/bearer.c tipc_disc_reset(net, b); b 359 net/tipc/bearer.c static void bearer_disable(struct net *net, struct tipc_bearer *b) b 362 net/tipc/bearer.c int bearer_id = b->identity; b 364 net/tipc/bearer.c pr_info("Disabling bearer <%s>\n", b->name); b 365 net/tipc/bearer.c clear_bit_unlock(0, &b->up); b 367 net/tipc/bearer.c b->media->disable_media(b); b 368 net/tipc/bearer.c RCU_INIT_POINTER(b->media_ptr, NULL); b 369 net/tipc/bearer.c if (b->disc) b 370 net/tipc/bearer.c tipc_disc_delete(b->disc); b 372 net/tipc/bearer.c kfree_rcu(b, rcu); b 376 net/tipc/bearer.c int tipc_enable_l2_media(struct net *net, struct tipc_bearer *b, b 379 net/tipc/bearer.c char *dev_name = strchr((const char *)b->name, ':') + 1; b 380 net/tipc/bearer.c int hwaddr_len = b->media->hwaddr_len; b 394 net/tipc/bearer.c pr_info("Enabling <%s> not permitted\n", b->name); b 410 net/tipc/bearer.c rcu_assign_pointer(b->media_ptr, dev); b 411 net/tipc/bearer.c b->pt.dev = dev; b 412 net/tipc/bearer.c b->pt.type = htons(ETH_P_TIPC); b 413 net/tipc/bearer.c b->pt.func = tipc_l2_rcv_msg; b 414 net/tipc/bearer.c dev_add_pack(&b->pt); b 415 net/tipc/bearer.c memset(&b->bcast_addr, 0, sizeof(b->bcast_addr)); b 416 net/tipc/bearer.c memcpy(b->bcast_addr.value, dev->broadcast, hwaddr_len); b 417 net/tipc/bearer.c b->bcast_addr.media_id = b->media->type_id; b 418 net/tipc/bearer.c b->bcast_addr.broadcast = TIPC_BROADCAST_SUPPORT; b 419 net/tipc/bearer.c b->mtu = dev->mtu; b 420 net/tipc/bearer.c b->media->raw2addr(b, &b->addr, (char *)dev->dev_addr); b 421 net/tipc/bearer.c rcu_assign_pointer(dev->tipc_ptr, b); b 429 net/tipc/bearer.c void tipc_disable_l2_media(struct tipc_bearer *b) b 433 net/tipc/bearer.c dev = (struct net_device *)rtnl_dereference(b->media_ptr); b 434 net/tipc/bearer.c dev_remove_pack(&b->pt); b 447 net/tipc/bearer.c struct tipc_bearer *b, struct tipc_media_addr *dest) b 452 net/tipc/bearer.c dev = (struct net_device *)rcu_dereference(b->media_ptr); b 473 net/tipc/bearer.c struct tipc_bearer *b; b 476 net/tipc/bearer.c b = bearer_get(net, bearer_id); b 477 net/tipc/bearer.c if (b) b 478 net/tipc/bearer.c supp = (b->bcast_addr.broadcast == TIPC_BROADCAST_SUPPORT); b 486 net/tipc/bearer.c struct tipc_bearer *b; b 489 net/tipc/bearer.c b = rcu_dereference(tipc_net(net)->bearer_list[bearer_id]); b 490 net/tipc/bearer.c if (b) b 491 net/tipc/bearer.c mtu = b->mtu; b 503 net/tipc/bearer.c struct tipc_bearer *b; b 506 net/tipc/bearer.c b = bearer_get(net, bearer_id); b 507 net/tipc/bearer.c if (likely(b && (test_bit(0, &b->up) || msg_is_reset(hdr)))) b 508 net/tipc/bearer.c b->media->send_msg(net, skb, b, dest); b 520 net/tipc/bearer.c struct tipc_bearer *b; b 527 net/tipc/bearer.c b = bearer_get(net, bearer_id); b 528 net/tipc/bearer.c if (unlikely(!b)) b 532 net/tipc/bearer.c if (likely(test_bit(0, &b->up) || msg_is_reset(buf_msg(skb)))) b 533 net/tipc/bearer.c b->media->send_msg(net, skb, b, dst); b 547 net/tipc/bearer.c struct tipc_bearer *b; b 552 net/tipc/bearer.c b = bearer_get(net, bearer_id); b 553 net/tipc/bearer.c if (unlikely(!b || !test_bit(0, &b->up))) b 560 net/tipc/bearer.c b->media->send_msg(net, skb, b, &b->bcast_addr); b 579 net/tipc/bearer.c struct tipc_bearer *b; b 582 net/tipc/bearer.c b = rcu_dereference(dev->tipc_ptr) ?: b 584 net/tipc/bearer.c if (likely(b && test_bit(0, &b->up) && b 587 net/tipc/bearer.c tipc_rcv(dev_net(b->pt.dev), skb, b); b 610 net/tipc/bearer.c struct tipc_bearer *b; b 612 net/tipc/bearer.c b = rtnl_dereference(dev->tipc_ptr); b 613 net/tipc/bearer.c if (!b) b 616 net/tipc/bearer.c trace_tipc_l2_device_event(dev, b, evt); b 620 net/tipc/bearer.c test_and_set_bit_lock(0, &b->up); b 625 net/tipc/bearer.c clear_bit_unlock(0, &b->up); b 626 net/tipc/bearer.c tipc_reset_bearer(net, b); b 629 net/tipc/bearer.c test_and_set_bit_lock(0, &b->up); b 633 net/tipc/bearer.c bearer_disable(net, b); b 636 net/tipc/bearer.c b->mtu = dev->mtu; b 637 net/tipc/bearer.c tipc_reset_bearer(net, b); b 640 net/tipc/bearer.c b->media->raw2addr(b, &b->addr, b 642 net/tipc/bearer.c tipc_reset_bearer(net, b); b 646 net/tipc/bearer.c bearer_disable(net, b); b 670 net/tipc/bearer.c struct tipc_bearer *b; b 674 net/tipc/bearer.c b = rtnl_dereference(tn->bearer_list[i]); b 675 net/tipc/bearer.c if (b) { b 676 net/tipc/bearer.c bearer_disable(net, b); b 981 net/tipc/bearer.c struct tipc_bearer *b; b 999 net/tipc/bearer.c b = tipc_bearer_find(net, name); b 1000 net/tipc/bearer.c if (!b) { b 1007 net/tipc/bearer.c err = tipc_udp_nl_bearer_add(b, b 1022 net/tipc/bearer.c struct tipc_bearer *b; b 1041 net/tipc/bearer.c b = tipc_bearer_find(net, name); b 1042 net/tipc/bearer.c if (!b) b 1054 net/tipc/bearer.c b->tolerance = nla_get_u32(props[TIPC_NLA_PROP_TOL]); b 1055 net/tipc/bearer.c tipc_node_apply_property(net, b, TIPC_NLA_PROP_TOL); b 1058 net/tipc/bearer.c b->priority = nla_get_u32(props[TIPC_NLA_PROP_PRIO]); b 1060 net/tipc/bearer.c b->window = nla_get_u32(props[TIPC_NLA_PROP_WIN]); b 1062 net/tipc/bearer.c if (b->media->type_id != TIPC_MEDIA_TYPE_UDP) b 1068 net/tipc/bearer.c b->mtu = nla_get_u32(props[TIPC_NLA_PROP_MTU]); b 1069 net/tipc/bearer.c tipc_node_apply_property(net, b, TIPC_NLA_PROP_MTU); b 105 net/tipc/bearer.h struct tipc_bearer *b, b 107 net/tipc/bearer.h int (*enable_media)(struct net *net, struct tipc_bearer *b, b 109 net/tipc/bearer.h void (*disable_media)(struct tipc_bearer *b); b 114 net/tipc/bearer.h int (*msg2addr)(struct tipc_bearer *b, b 117 net/tipc/bearer.h int (*raw2addr)(struct tipc_bearer *b, b 179 net/tipc/bearer.h void tipc_rcv(struct net *net, struct sk_buff *skb, struct tipc_bearer *b); b 211 net/tipc/bearer.h int tipc_enable_l2_media(struct net *net, struct tipc_bearer *b, b 213 net/tipc/bearer.h void tipc_disable_l2_media(struct tipc_bearer *b); b 215 net/tipc/bearer.h struct tipc_bearer *b, struct tipc_media_addr *dest); b 81 net/tipc/discover.c u32 mtyp, struct tipc_bearer *b) b 84 net/tipc/discover.c u32 dest_domain = b->domain; b 96 net/tipc/discover.c b->media->addr2msg(msg_media_addr(hdr), &b->addr); b 103 net/tipc/discover.c struct tipc_bearer *b) b 112 net/tipc/discover.c tipc_disc_init_msg(net, skb, mtyp, b); b 115 net/tipc/discover.c tipc_bearer_xmit_skb(net, b->identity, skb, maddr); b 124 net/tipc/discover.c static void disc_dupl_alert(struct tipc_bearer *b, u32 node_addr, b 132 net/tipc/discover.c media_addr_str, b->name); b 141 net/tipc/discover.c struct tipc_bearer *b, b 181 net/tipc/discover.c self, sugg_addr, maddr, b); b 192 net/tipc/discover.c struct tipc_bearer *b) b 219 net/tipc/discover.c err = b->media->msg2addr(b, &maddr, msg_media_addr(hdr)); b 226 net/tipc/discover.c if (!memcmp(&maddr, &b->addr, sizeof(maddr))) b 230 net/tipc/discover.c if (tipc_disc_addr_trial_msg(b->disc, &maddr, b, dst, b 237 net/tipc/discover.c disc_dupl_alert(b, self, &maddr); b 242 net/tipc/discover.c if (!tipc_in_scope(legacy, b->domain, src)) b 244 net/tipc/discover.c tipc_node_check_dest(net, src, peer_id, b, caps, signature, b 247 net/tipc/discover.c disc_dupl_alert(b, src, &maddr); b 252 net/tipc/discover.c tipc_disc_msg_xmit(net, DSC_RESP_MSG, src, self, 0, &maddr, b); b 344 net/tipc/discover.c int tipc_disc_create(struct net *net, struct tipc_bearer *b, b 358 net/tipc/discover.c tipc_disc_init_msg(net, d->skb, DSC_REQ_MSG, b); b 367 net/tipc/discover.c d->bearer_id = b->identity; b 368 net/tipc/discover.c d->domain = b->domain; b 374 net/tipc/discover.c b->disc = d; b 396 net/tipc/discover.c void tipc_disc_reset(struct net *net, struct tipc_bearer *b) b 398 net/tipc/discover.c struct tipc_discoverer *d = b->disc; b 403 net/tipc/discover.c tipc_disc_init_msg(net, d->skb, DSC_REQ_MSG, b); b 405 net/tipc/discover.c d->bearer_id = b->identity; b 406 net/tipc/discover.c d->domain = b->domain; b 414 net/tipc/discover.c tipc_bearer_xmit_skb(net, b->identity, skb, &maddr); b 61 net/tipc/eth_media.c static int tipc_eth_raw2addr(struct tipc_bearer *b, b 75 net/tipc/eth_media.c static int tipc_eth_msg2addr(struct tipc_bearer *b, b 81 net/tipc/eth_media.c return tipc_eth_raw2addr(b, addr, msg); b 66 net/tipc/ib_media.c static int tipc_ib_raw2addr(struct tipc_bearer *b, b 73 net/tipc/ib_media.c addr->broadcast = !memcmp(msg, b->bcast_addr.value, b 79 net/tipc/ib_media.c static int tipc_ib_msg2addr(struct tipc_bearer *b, b 83 net/tipc/ib_media.c return tipc_ib_raw2addr(b, addr, msg); b 182 net/tipc/msg.h static inline void msg_swap_words(struct tipc_msg *msg, u32 a, u32 b) b 186 net/tipc/msg.h msg->hdr[a] = msg->hdr[b]; b 187 net/tipc/msg.h msg->hdr[b] = temp; b 808 net/tipc/name_table.c struct nlattr *b; b 836 net/tipc/name_table.c b = nla_nest_start_noflag(msg->skb, TIPC_NLA_NAME_TABLE_PUBL); b 837 net/tipc/name_table.c if (!b) b 855 net/tipc/name_table.c nla_nest_end(msg->skb, b); b 864 net/tipc/name_table.c nla_nest_cancel(msg->skb, b); b 400 net/tipc/netlink_compat.c struct tipc_bearer_config *b; b 403 net/tipc/netlink_compat.c b = (struct tipc_bearer_config *)TLV_DATA(msg->req); b 415 net/tipc/netlink_compat.c if (!string_is_valid(b->name, len)) b 418 net/tipc/netlink_compat.c if (nla_put_string(skb, TIPC_NLA_BEARER_NAME, b->name)) b 421 net/tipc/netlink_compat.c if (nla_put_u32(skb, TIPC_NLA_BEARER_DOMAIN, ntohl(b->disc_domain))) b 424 net/tipc/netlink_compat.c if (ntohl(b->priority) <= TIPC_MAX_LINK_PRI) { b 428 net/tipc/netlink_compat.c if (nla_put_u32(skb, TIPC_NLA_PROP_PRIO, ntohl(b->priority))) b 981 net/tipc/node.c u8 *peer_id, struct tipc_bearer *b, b 1007 net/tipc/node.c le = &n->links[b->identity]; b 1082 net/tipc/node.c if_name = strchr(b->name, ':') + 1; b 1084 net/tipc/node.c if (!tipc_link_create(net, if_name, b->identity, b->tolerance, b 1085 net/tipc/node.c b->net_plane, b->mtu, b->priority, b 1086 net/tipc/node.c b->window, session, b 1113 net/tipc/node.c tipc_node_link_down(n, b->identity, false); b 1800 net/tipc/node.c void tipc_rcv(struct net *net, struct sk_buff *skb, struct tipc_bearer *b) b 1805 net/tipc/node.c int bearer_id = b->identity; b 1824 net/tipc/node.c return tipc_disc_rcv(net, skb, b); b 1894 net/tipc/node.c void tipc_node_apply_property(struct net *net, struct tipc_bearer *b, b 1898 net/tipc/node.c int bearer_id = b->identity; b 1912 net/tipc/node.c tipc_link_set_tolerance(e->link, b->tolerance, b 1915 net/tipc/node.c tipc_link_set_mtu(e->link, b->mtu); b 82 net/tipc/node.h void tipc_node_apply_property(struct net *net, struct tipc_bearer *b, int prop); b 394 net/tipc/trace.h TP_PROTO(struct net_device *dev, struct tipc_bearer *b, b 397 net/tipc/trace.h TP_ARGS(dev, b, evt), b 401 net/tipc/trace.h __string(b_name, b->name) b 410 net/tipc/trace.h __assign_str(b_name, b->name); b 412 net/tipc/trace.h __entry->b_up = test_bit(0, &b->up); b 137 net/tipc/udp_media.c static int tipc_udp_msg2addr(struct tipc_bearer *b, struct tipc_media_addr *a, b 225 net/tipc/udp_media.c struct tipc_bearer *b, b 228 net/tipc/udp_media.c struct udp_media_addr *src = (struct udp_media_addr *)&b->addr.value; b 241 net/tipc/udp_media.c ub = rcu_dereference(b->media_ptr); b 272 net/tipc/udp_media.c static bool tipc_udp_is_known_peer(struct tipc_bearer *b, b 278 net/tipc/udp_media.c ub = rcu_dereference_rtnl(b->media_ptr); b 292 net/tipc/udp_media.c static int tipc_udp_rcast_add(struct tipc_bearer *b, b 298 net/tipc/udp_media.c ub = rcu_dereference_rtnl(b->media_ptr); b 319 net/tipc/udp_media.c b->bcast_addr.broadcast = TIPC_REPLICAST_SUPPORT; b 324 net/tipc/udp_media.c static int tipc_udp_rcast_disc(struct tipc_bearer *b, struct sk_buff *skb) b 329 net/tipc/udp_media.c dst = (struct udp_media_addr *)&b->bcast_addr.value; b 355 net/tipc/udp_media.c if (likely(tipc_udp_is_known_peer(b, &src))) b 358 net/tipc/udp_media.c return tipc_udp_rcast_add(b, &src); b 365 net/tipc/udp_media.c struct tipc_bearer *b; b 377 net/tipc/udp_media.c b = rcu_dereference(ub->bearer); b 378 net/tipc/udp_media.c if (!b) b 381 net/tipc/udp_media.c if (b && test_bit(0, &b->up)) { b 382 net/tipc/udp_media.c tipc_rcv(sock_net(sk), skb, b); b 387 net/tipc/udp_media.c err = tipc_udp_rcast_disc(b, skb); b 451 net/tipc/udp_media.c struct tipc_bearer *b; b 482 net/tipc/udp_media.c b = tipc_bearer_find(net, bname); b 483 net/tipc/udp_media.c if (!b) { b 487 net/tipc/udp_media.c bid = b->identity; b 493 net/tipc/udp_media.c b = rtnl_dereference(tn->bearer_list[bid]); b 494 net/tipc/udp_media.c if (!b) { b 500 net/tipc/udp_media.c ub = rtnl_dereference(b->media_ptr); b 535 net/tipc/udp_media.c int tipc_udp_nl_add_bearer_data(struct tipc_nl_msg *msg, struct tipc_bearer *b) b 537 net/tipc/udp_media.c struct udp_media_addr *src = (struct udp_media_addr *)&b->addr.value; b 542 net/tipc/udp_media.c ub = rtnl_dereference(b->media_ptr); b 553 net/tipc/udp_media.c dst = (struct udp_media_addr *)&b->bcast_addr.value; b 617 net/tipc/udp_media.c int tipc_udp_nl_bearer_add(struct tipc_bearer *b, struct nlattr *attr) b 634 net/tipc/udp_media.c dst = (struct udp_media_addr *)&b->bcast_addr.value; b 640 net/tipc/udp_media.c if (tipc_udp_is_known_peer(b, &addr)) b 643 net/tipc/udp_media.c return tipc_udp_rcast_add(b, &addr); b 655 net/tipc/udp_media.c static int tipc_udp_enable(struct net *net, struct tipc_bearer *b, b 714 net/tipc/udp_media.c b->bcast_addr.media_id = TIPC_MEDIA_TYPE_UDP; b 715 net/tipc/udp_media.c b->bcast_addr.broadcast = TIPC_BROADCAST_SUPPORT; b 716 net/tipc/udp_media.c rcu_assign_pointer(b->media_ptr, ub); b 717 net/tipc/udp_media.c rcu_assign_pointer(ub->bearer, b); b 718 net/tipc/udp_media.c tipc_udp_media_addr_set(&b->addr, &local); b 741 net/tipc/udp_media.c b->mtu = b->media->mtu; b 751 net/tipc/udp_media.c b->mtu = 1280; b 775 net/tipc/udp_media.c memcpy(&b->bcast_addr.value, &remote, sizeof(remote)); b 779 net/tipc/udp_media.c err = tipc_udp_rcast_add(b, &remote); b 812 net/tipc/udp_media.c static void tipc_udp_disable(struct tipc_bearer *b) b 816 net/tipc/udp_media.c ub = rtnl_dereference(b->media_ptr); b 44 net/tipc/udp_media.h int tipc_udp_nl_bearer_add(struct tipc_bearer *b, struct nlattr *attr); b 45 net/tipc/udp_media.h int tipc_udp_nl_add_bearer_data(struct tipc_nl_msg *msg, struct tipc_bearer *b); b 2703 net/unix/af_unix.c #define set_bucket_offset(b, o) ((b) << BUCKET_SPACE | (o)) b 84 net/vmw_vsock/hyperv_transport.c unsigned char b[sizeof(guid_t) - sizeof(unsigned int)]; b 155 net/vmw_vsock/hyperv_transport.c return !memcmp(&id->b[4], &srv_id_template.b[4], sizeof(guid_t) - 4); b 7295 net/wireless/nl80211.c static bool is_band_valid(struct wiphy *wiphy, enum nl80211_band b) b 7297 net/wireless/nl80211.c return b < NUM_NL80211_BANDS && wiphy->bands[b]; b 767 net/wireless/scan.c struct cfg80211_bss *b, b 775 net/wireless/scan.c if (a->channel != b->channel) b 776 net/wireless/scan.c return b->channel->center_freq - a->channel->center_freq; b 781 net/wireless/scan.c b_ies = rcu_access_pointer(b->ies); b 788 net/wireless/scan.c if (WLAN_CAPABILITY_IS_STA_BSS(b->capability)) b 812 net/wireless/scan.c r = memcmp(a->bssid, b->bssid, sizeof(a->bssid)); b 184 net/xfrm/xfrm_policy.c struct xfrm_pol_inexact_bin *b, b 790 net/xfrm/xfrm_policy.c const xfrm_address_t *b, b 799 net/xfrm/xfrm_policy.c return ntohl(a->a4) - ntohl(b->a4); b 801 net/xfrm/xfrm_policy.c (ntohl(b->a4) & ((~0UL << (32 - prefixlen)))); b 807 net/xfrm/xfrm_policy.c delta = memcmp(a->a6, b->a6, pdw << 2); b 815 net/xfrm/xfrm_policy.c (ntohl(b->a6[pdw]) & mask); b 1072 net/xfrm/xfrm_policy.c static void __xfrm_policy_inexact_prune_bin(struct xfrm_pol_inexact_bin *b, bool net_exit) b 1074 net/xfrm/xfrm_policy.c write_seqcount_begin(&b->count); b 1075 net/xfrm/xfrm_policy.c xfrm_policy_inexact_gc_tree(&b->root_d, net_exit); b 1076 net/xfrm/xfrm_policy.c xfrm_policy_inexact_gc_tree(&b->root_s, net_exit); b 1077 net/xfrm/xfrm_policy.c write_seqcount_end(&b->count); b 1079 net/xfrm/xfrm_policy.c if (!RB_EMPTY_ROOT(&b->root_d) || !RB_EMPTY_ROOT(&b->root_s) || b 1080 net/xfrm/xfrm_policy.c !hlist_empty(&b->hhead)) { b 1085 net/xfrm/xfrm_policy.c if (rhashtable_remove_fast(&xfrm_policy_inexact_table, &b->head, b 1087 net/xfrm/xfrm_policy.c list_del(&b->inexact_bins); b 1088 net/xfrm/xfrm_policy.c kfree_rcu(b, rcu); b 1092 net/xfrm/xfrm_policy.c static void xfrm_policy_inexact_prune_bin(struct xfrm_pol_inexact_bin *b) b 1094 net/xfrm/xfrm_policy.c struct net *net = read_pnet(&b->k.net); b 1097 net/xfrm/xfrm_policy.c __xfrm_policy_inexact_prune_bin(b, false); b 1454 net/xfrm/xfrm_policy.c const struct xfrm_pol_inexact_bin *b = data; b 1456 net/xfrm/xfrm_policy.c return xfrm_pol_bin_key(&b->k, 0, seed); b 1463 net/xfrm/xfrm_policy.c const struct xfrm_pol_inexact_bin *b = ptr; b 1466 net/xfrm/xfrm_policy.c if (!net_eq(read_pnet(&b->k.net), read_pnet(&key->net))) b 1469 net/xfrm/xfrm_policy.c ret = b->k.dir ^ key->dir; b 1473 net/xfrm/xfrm_policy.c ret = b->k.type ^ key->type; b 1477 net/xfrm/xfrm_policy.c ret = b->k.family ^ key->family; b 1481 net/xfrm/xfrm_policy.c return b->k.if_id ^ key->if_id; b 1943 net/xfrm/xfrm_policy.c struct xfrm_pol_inexact_bin *b, b 1950 net/xfrm/xfrm_policy.c if (!b) b 1953 net/xfrm/xfrm_policy.c family = b->k.family; b 1955 net/xfrm/xfrm_policy.c cand->res[XFRM_POL_CAND_ANY] = &b->hhead; b 1957 net/xfrm/xfrm_policy.c n = xfrm_policy_lookup_inexact_addr(&b->root_d, &b->count, daddr, b 1961 net/xfrm/xfrm_policy.c n = xfrm_policy_lookup_inexact_addr(&n->root, &b->count, saddr, b 1967 net/xfrm/xfrm_policy.c n = xfrm_policy_lookup_inexact_addr(&b->root_s, &b->count, saddr, b 4070 net/xfrm/xfrm_policy.c struct xfrm_pol_inexact_bin *b, *t; b 4098 net/xfrm/xfrm_policy.c list_for_each_entry_safe(b, t, &net->xfrm.inexact_bins, inexact_bins) b 4099 net/xfrm/xfrm_policy.c __xfrm_policy_inexact_prune_bin(b, true); b 24 samples/bpf/test_lru_dist.c #define min(a, b) ((a) < (b) ? (a) : (b)) b 177 samples/bpf/test_lru_dist.c char *b, *l; b 184 samples/bpf/test_lru_dist.c b = malloc(fst.st_size); b 185 samples/bpf/test_lru_dist.c assert(b); b 187 samples/bpf/test_lru_dist.c assert(read(dist_fd, b, fst.st_size) == fst.st_size); b 190 samples/bpf/test_lru_dist.c if (b[i] == '\n') b 199 samples/bpf/test_lru_dist.c for (l = strtok(b, "\n"); l; l = strtok(NULL, "\n")) b 201 samples/bpf/test_lru_dist.c free(b); b 576 samples/bpf/xdp_monitor_user.c static inline void swap(struct stats_record **a, struct stats_record **b) b 581 samples/bpf/xdp_monitor_user.c *a = *b; b 582 samples/bpf/xdp_monitor_user.c *b = tmp; b 490 samples/bpf/xdp_redirect_cpu_user.c static inline void swap(struct stats_record **a, struct stats_record **b) b 495 samples/bpf/xdp_redirect_cpu_user.c *a = *b; b 496 samples/bpf/xdp_redirect_cpu_user.c *b = tmp; b 428 samples/bpf/xdp_rxq_info_user.c static inline void swap(struct stats_record **a, struct stats_record **b) b 433 samples/bpf/xdp_rxq_info_user.c *a = *b; b 434 samples/bpf/xdp_rxq_info_user.c *b = tmp; b 721 scripts/asn1_compiler.c const struct type *const *a = _a, *const *b = _b; b 723 scripts/asn1_compiler.c if ((*a)->name->size != (*b)->name->size) b 724 scripts/asn1_compiler.c return (*a)->name->size - (*b)->name->size; b 726 scripts/asn1_compiler.c return memcmp((*a)->name->content, (*b)->name->content, b 55 scripts/dtc/dtc.h #define streq(a, b) (strcmp((a), (b)) == 0) b 57 scripts/dtc/dtc.h #define strprefixeq(a, n, b) (strlen(b) == (n) && (memcmp(a, b, n) == 0)) b 15 scripts/dtc/include-prefixes/dt-bindings/pinctrl/r7s72100-pinctrl.h #define RZA1_PINMUX(b, p, f) ((b) * RZA1_PINS_PER_PORT + (p) | (f << 16)) b 40 scripts/dtc/include-prefixes/dt-bindings/pinctrl/r7s9210-pinctrl.h #define RZA2_PINMUX(b, p, f) ((b) * RZA2_PINS_PER_PORT + (p) | (f << 16)) b 649 scripts/dtc/livetree.c const struct reserve_info *a, *b; b 652 scripts/dtc/livetree.c b = *((const struct reserve_info * const *)bx); b 654 scripts/dtc/livetree.c if (a->address < b->address) b 656 scripts/dtc/livetree.c else if (a->address > b->address) b 658 scripts/dtc/livetree.c else if (a->size < b->size) b 660 scripts/dtc/livetree.c else if (a->size > b->size) b 698 scripts/dtc/livetree.c const struct property *a, *b; b 701 scripts/dtc/livetree.c b = *((const struct property * const *)bx); b 703 scripts/dtc/livetree.c return strcmp(a->name, b->name); b 734 scripts/dtc/livetree.c const struct node *a, *b; b 737 scripts/dtc/livetree.c b = *((const struct node * const *)bx); b 739 scripts/dtc/livetree.c return strcmp(a->name, b->name); b 138 scripts/extract-cert.c BIO *b; b 141 scripts/extract-cert.c b = BIO_new_file(cert_src, "rb"); b 142 scripts/extract-cert.c ERR(!b, "%s", cert_src); b 145 scripts/extract-cert.c x509 = PEM_read_bio_X509(b, NULL, NULL, NULL); b 142 scripts/gcc-plugins/randomize_layout_plugin.c typedef struct ranctx { u64 a; u64 b; u64 c; u64 d; } ranctx; b 146 scripts/gcc-plugins/randomize_layout_plugin.c u64 e = x->a - rot(x->b, 7); b 147 scripts/gcc-plugins/randomize_layout_plugin.c x->a = x->b ^ rot(x->c, 13); b 148 scripts/gcc-plugins/randomize_layout_plugin.c x->b = x->c + rot(x->d, 37); b 158 scripts/gcc-plugins/randomize_layout_plugin.c x->b = seed[1]; b 56 scripts/genksyms/genksyms.c static int equal_list(struct string_list *a, struct string_list *b); b 394 scripts/genksyms/genksyms.c static int equal_list(struct string_list *a, struct string_list *b) b 396 scripts/genksyms/genksyms.c while (a && b) { b 397 scripts/genksyms/genksyms.c if (a->tag != b->tag || strcmp(a->string, b->string)) b 400 scripts/genksyms/genksyms.c b = b->next; b 403 scripts/genksyms/genksyms.c return !a && !b; b 503 scripts/genksyms/genksyms.c struct string_list **e, **b; b 516 scripts/genksyms/genksyms.c b = alloca(elem * sizeof(*e)); b 517 scripts/genksyms/genksyms.c e = b + elem; b 524 scripts/genksyms/genksyms.c while (b != e) { b 525 scripts/genksyms/genksyms.c print_node(f, *b++); b 533 scripts/genksyms/genksyms.c struct string_list **e, **b; b 544 scripts/genksyms/genksyms.c b = alloca(elem * sizeof(*e)); b 545 scripts/genksyms/genksyms.c e = b + elem; b 552 scripts/genksyms/genksyms.c while (b != e) { b 556 scripts/genksyms/genksyms.c cur = *(b++); b 37 scripts/genksyms/parse.y struct string_list *b = *pb, *e = *pe; b 39 scripts/genksyms/parse.y free_list(b, e); b 48 scripts/genksyms/parse.y struct string_list *b = *body, *i = *ident, *r; b 58 scripts/genksyms/parse.y add_symbol(i->string, type, b, is_extern); b 673 scripts/kallsyms.c static int compare_symbols(const void *a, const void *b) b 680 scripts/kallsyms.c sb = b; b 51 scripts/kconfig/lxdialog/util.c #define DLG_COLOR(dialog, f, b, h) \ b 54 scripts/kconfig/lxdialog/util.c dlg.dialog.bg = (b); \ b 25 scripts/kconfig/nconf.h #define max(a, b) ({\ b 27 scripts/kconfig/nconf.h typeof(b) _b = b;\ b 30 scripts/kconfig/nconf.h #define min(a, b) ({\ b 32 scripts/kconfig/nconf.h typeof(b) _b = b;\ b 954 scripts/kconfig/qconf.cc void ConfigView::setShowName(bool b) b 956 scripts/kconfig/qconf.cc if (list->showName != b) { b 957 scripts/kconfig/qconf.cc list->showName = b; b 959 scripts/kconfig/qconf.cc emit showNameChanged(b); b 963 scripts/kconfig/qconf.cc void ConfigView::setShowRange(bool b) b 965 scripts/kconfig/qconf.cc if (list->showRange != b) { b 966 scripts/kconfig/qconf.cc list->showRange = b; b 968 scripts/kconfig/qconf.cc emit showRangeChanged(b); b 972 scripts/kconfig/qconf.cc void ConfigView::setShowData(bool b) b 974 scripts/kconfig/qconf.cc if (list->showData != b) { b 975 scripts/kconfig/qconf.cc list->showData = b; b 977 scripts/kconfig/qconf.cc emit showDataChanged(b); b 1031 scripts/kconfig/qconf.cc void ConfigInfoView::setShowDebug(bool b) b 1033 scripts/kconfig/qconf.cc if (_showDebug != b) { b 1034 scripts/kconfig/qconf.cc _showDebug = b; b 1039 scripts/kconfig/qconf.cc emit showDebugChanged(b); b 38 scripts/mod/file2alias.c __u8 b[16]; b 43 scripts/mod/file2alias.c __u8 b[16]; b 46 scripts/mod/file2alias.c __u8 b[16]; b 112 scripts/mod/file2alias.c uuid.b[3], uuid.b[2], uuid.b[1], uuid.b[0], b 113 scripts/mod/file2alias.c uuid.b[5], uuid.b[4], uuid.b[7], uuid.b[6], b 114 scripts/mod/file2alias.c uuid.b[8], uuid.b[9], uuid.b[10], uuid.b[11], b 115 scripts/mod/file2alias.c uuid.b[12], uuid.b[13], uuid.b[14], uuid.b[15]); b 894 scripts/mod/file2alias.c sprintf(&guid_name[i], "%02x", TO_NATIVE((guid->b)[i/2])); b 1308 scripts/mod/file2alias.c uuid.b[0], uuid.b[1], uuid.b[2], uuid.b[3], uuid.b[4], b 1309 scripts/mod/file2alias.c uuid.b[5], uuid.b[6], uuid.b[7], uuid.b[8], uuid.b[9], b 1310 scripts/mod/file2alias.c uuid.b[10], uuid.b[11], uuid.b[12], uuid.b[13], uuid.b[14], b 1311 scripts/mod/file2alias.c uuid.b[15]); b 2259 scripts/mod/modpost.c static void add_header(struct buffer *b, struct module *mod) b 2261 scripts/mod/modpost.c buf_printf(b, "#include <linux/build-salt.h>\n"); b 2262 scripts/mod/modpost.c buf_printf(b, "#include <linux/module.h>\n"); b 2263 scripts/mod/modpost.c buf_printf(b, "#include <linux/vermagic.h>\n"); b 2264 scripts/mod/modpost.c buf_printf(b, "#include <linux/compiler.h>\n"); b 2265 scripts/mod/modpost.c buf_printf(b, "\n"); b 2266 scripts/mod/modpost.c buf_printf(b, "BUILD_SALT;\n"); b 2267 scripts/mod/modpost.c buf_printf(b, "\n"); b 2268 scripts/mod/modpost.c buf_printf(b, "MODULE_INFO(vermagic, VERMAGIC_STRING);\n"); b 2269 scripts/mod/modpost.c buf_printf(b, "MODULE_INFO(name, KBUILD_MODNAME);\n"); b 2270 scripts/mod/modpost.c buf_printf(b, "\n"); b 2271 scripts/mod/modpost.c buf_printf(b, "__visible struct module __this_module\n"); b 2272 scripts/mod/modpost.c buf_printf(b, "__section(.gnu.linkonce.this_module) = {\n"); b 2273 scripts/mod/modpost.c buf_printf(b, "\t.name = KBUILD_MODNAME,\n"); b 2275 scripts/mod/modpost.c buf_printf(b, "\t.init = init_module,\n"); b 2277 scripts/mod/modpost.c buf_printf(b, "#ifdef CONFIG_MODULE_UNLOAD\n" b 2280 scripts/mod/modpost.c buf_printf(b, "\t.arch = MODULE_ARCH_INIT,\n"); b 2281 scripts/mod/modpost.c buf_printf(b, "};\n"); b 2284 scripts/mod/modpost.c static void add_intree_flag(struct buffer *b, int is_intree) b 2287 scripts/mod/modpost.c buf_printf(b, "\nMODULE_INFO(intree, \"Y\");\n"); b 2291 scripts/mod/modpost.c static void add_retpoline(struct buffer *b) b 2293 scripts/mod/modpost.c buf_printf(b, "\n#ifdef CONFIG_RETPOLINE\n"); b 2294 scripts/mod/modpost.c buf_printf(b, "MODULE_INFO(retpoline, \"Y\");\n"); b 2295 scripts/mod/modpost.c buf_printf(b, "#endif\n"); b 2298 scripts/mod/modpost.c static void add_staging_flag(struct buffer *b, const char *name) b 2301 scripts/mod/modpost.c buf_printf(b, "\nMODULE_INFO(staging, \"Y\");\n"); b 2307 scripts/mod/modpost.c static int add_versions(struct buffer *b, struct module *mod) b 2324 scripts/mod/modpost.c buf_printf(b, "\n"); b 2325 scripts/mod/modpost.c buf_printf(b, "static const struct modversion_info ____versions[]\n"); b 2326 scripts/mod/modpost.c buf_printf(b, "__used __section(__versions) = {\n"); b 2342 scripts/mod/modpost.c buf_printf(b, "\t{ %#8x, \"%s\" },\n", b 2346 scripts/mod/modpost.c buf_printf(b, "};\n"); b 2351 scripts/mod/modpost.c static void add_depends(struct buffer *b, struct module *mod) b 2361 scripts/mod/modpost.c buf_printf(b, "\n"); b 2362 scripts/mod/modpost.c buf_printf(b, "MODULE_INFO(depends, \""); b 2377 scripts/mod/modpost.c buf_printf(b, "%s%s", first ? "" : ",", p); b 2380 scripts/mod/modpost.c buf_printf(b, "\");\n"); b 2383 scripts/mod/modpost.c static void add_srcversion(struct buffer *b, struct module *mod) b 2386 scripts/mod/modpost.c buf_printf(b, "\n"); b 2387 scripts/mod/modpost.c buf_printf(b, "MODULE_INFO(srcversion, \"%s\");\n", b 2392 scripts/mod/modpost.c static void write_if_changed(struct buffer *b, const char *fname) b 2405 scripts/mod/modpost.c if (st.st_size != b->pos) b 2408 scripts/mod/modpost.c tmp = NOFAIL(malloc(b->pos)); b 2409 scripts/mod/modpost.c if (fread(tmp, 1, b->pos, file) != b->pos) b 2412 scripts/mod/modpost.c if (memcmp(tmp, b->p, b->pos) != 0) b 2429 scripts/mod/modpost.c if (fwrite(b->p, 1, b->pos, file) != b->pos) { b 67 scripts/mod/sumversion.c #define ROUND1(a,b,c,d,k,s) (a = lshift(a + F(b,c,d) + k, s)) b 68 scripts/mod/sumversion.c #define ROUND2(a,b,c,d,k,s) (a = lshift(a + G(b,c,d) + k + (uint32_t)0x5A827999,s)) b 69 scripts/mod/sumversion.c #define ROUND3(a,b,c,d,k,s) (a = lshift(a + H(b,c,d) + k + (uint32_t)0x6ED9EBA1,s)) b 90 scripts/mod/sumversion.c uint32_t a, b, c, d; b 93 scripts/mod/sumversion.c b = hash[1]; b 97 scripts/mod/sumversion.c ROUND1(a, b, c, d, in[0], 3); b 98 scripts/mod/sumversion.c ROUND1(d, a, b, c, in[1], 7); b 99 scripts/mod/sumversion.c ROUND1(c, d, a, b, in[2], 11); b 100 scripts/mod/sumversion.c ROUND1(b, c, d, a, in[3], 19); b 101 scripts/mod/sumversion.c ROUND1(a, b, c, d, in[4], 3); b 102 scripts/mod/sumversion.c ROUND1(d, a, b, c, in[5], 7); b 103 scripts/mod/sumversion.c ROUND1(c, d, a, b, in[6], 11); b 104 scripts/mod/sumversion.c ROUND1(b, c, d, a, in[7], 19); b 105 scripts/mod/sumversion.c ROUND1(a, b, c, d, in[8], 3); b 106 scripts/mod/sumversion.c ROUND1(d, a, b, c, in[9], 7); b 107 scripts/mod/sumversion.c ROUND1(c, d, a, b, in[10], 11); b 108 scripts/mod/sumversion.c ROUND1(b, c, d, a, in[11], 19); b 109 scripts/mod/sumversion.c ROUND1(a, b, c, d, in[12], 3); b 110 scripts/mod/sumversion.c ROUND1(d, a, b, c, in[13], 7); b 111 scripts/mod/sumversion.c ROUND1(c, d, a, b, in[14], 11); b 112 scripts/mod/sumversion.c ROUND1(b, c, d, a, in[15], 19); b 114 scripts/mod/sumversion.c ROUND2(a, b, c, d,in[ 0], 3); b 115 scripts/mod/sumversion.c ROUND2(d, a, b, c, in[4], 5); b 116 scripts/mod/sumversion.c ROUND2(c, d, a, b, in[8], 9); b 117 scripts/mod/sumversion.c ROUND2(b, c, d, a, in[12], 13); b 118 scripts/mod/sumversion.c ROUND2(a, b, c, d, in[1], 3); b 119 scripts/mod/sumversion.c ROUND2(d, a, b, c, in[5], 5); b 120 scripts/mod/sumversion.c ROUND2(c, d, a, b, in[9], 9); b 121 scripts/mod/sumversion.c ROUND2(b, c, d, a, in[13], 13); b 122 scripts/mod/sumversion.c ROUND2(a, b, c, d, in[2], 3); b 123 scripts/mod/sumversion.c ROUND2(d, a, b, c, in[6], 5); b 124 scripts/mod/sumversion.c ROUND2(c, d, a, b, in[10], 9); b 125 scripts/mod/sumversion.c ROUND2(b, c, d, a, in[14], 13); b 126 scripts/mod/sumversion.c ROUND2(a, b, c, d, in[3], 3); b 127 scripts/mod/sumversion.c ROUND2(d, a, b, c, in[7], 5); b 128 scripts/mod/sumversion.c ROUND2(c, d, a, b, in[11], 9); b 129 scripts/mod/sumversion.c ROUND2(b, c, d, a, in[15], 13); b 131 scripts/mod/sumversion.c ROUND3(a, b, c, d,in[ 0], 3); b 132 scripts/mod/sumversion.c ROUND3(d, a, b, c, in[8], 9); b 133 scripts/mod/sumversion.c ROUND3(c, d, a, b, in[4], 11); b 134 scripts/mod/sumversion.c ROUND3(b, c, d, a, in[12], 15); b 135 scripts/mod/sumversion.c ROUND3(a, b, c, d, in[2], 3); b 136 scripts/mod/sumversion.c ROUND3(d, a, b, c, in[10], 9); b 137 scripts/mod/sumversion.c ROUND3(c, d, a, b, in[6], 11); b 138 scripts/mod/sumversion.c ROUND3(b, c, d, a, in[14], 15); b 139 scripts/mod/sumversion.c ROUND3(a, b, c, d, in[1], 3); b 140 scripts/mod/sumversion.c ROUND3(d, a, b, c, in[9], 9); b 141 scripts/mod/sumversion.c ROUND3(c, d, a, b, in[5], 11); b 142 scripts/mod/sumversion.c ROUND3(b, c, d, a, in[13], 15); b 143 scripts/mod/sumversion.c ROUND3(a, b, c, d, in[3], 3); b 144 scripts/mod/sumversion.c ROUND3(d, a, b, c, in[11], 9); b 145 scripts/mod/sumversion.c ROUND3(c, d, a, b, in[7], 11); b 146 scripts/mod/sumversion.c ROUND3(b, c, d, a, in[15], 15); b 149 scripts/mod/sumversion.c hash[1] += b; b 157 scripts/sign-file.c BIO *b; b 159 scripts/sign-file.c b = BIO_new_file(private_key_name, "rb"); b 160 scripts/sign-file.c ERR(!b, "%s", private_key_name); b 161 scripts/sign-file.c private_key = PEM_read_bio_PrivateKey(b, NULL, pem_pw_cb, b 164 scripts/sign-file.c BIO_free(b); b 174 scripts/sign-file.c BIO *b; b 177 scripts/sign-file.c b = BIO_new_file(x509_name, "rb"); b 178 scripts/sign-file.c ERR(!b, "%s", x509_name); b 181 scripts/sign-file.c n = BIO_read(b, buf, 2); b 183 scripts/sign-file.c if (BIO_should_retry(b)) { b 194 scripts/sign-file.c ERR(BIO_reset(b) != 0, "%s", x509_name); b 198 scripts/sign-file.c x509 = d2i_X509_bio(b, NULL); b 201 scripts/sign-file.c x509 = PEM_read_bio_X509(b, NULL, NULL, NULL); b 203 scripts/sign-file.c BIO_free(b); b 330 scripts/sign-file.c BIO *b; b 334 scripts/sign-file.c b = BIO_new_file(sig_file_name, "wb"); b 335 scripts/sign-file.c ERR(!b, "%s", sig_file_name); b 337 scripts/sign-file.c ERR(i2d_CMS_bio_stream(b, cms, NULL, 0) < 0, b 340 scripts/sign-file.c ERR(i2d_PKCS7_bio(b, pkcs7) < 0, b 343 scripts/sign-file.c BIO_free(b); b 375 scripts/sign-file.c BIO *b; b 380 scripts/sign-file.c b = BIO_new_file(raw_sig_name, "rb"); b 381 scripts/sign-file.c ERR(!b, "%s", raw_sig_name); b 382 scripts/sign-file.c while ((n = BIO_read(b, buf, sizeof(buf))), n > 0) b 384 scripts/sign-file.c BIO_free(b); b 200 scripts/sortextable.c static int compare_relative_table(const void *a, const void *b) b 203 scripts/sortextable.c int32_t bv = (int32_t)r(b); b 78 scripts/sortextable.h static int compare_extable(const void *a, const void *b) b 81 scripts/sortextable.h Elf_Addr bv = _r(b); b 750 scripts/unifdef.c static Linetype op_lt(int *p, Linetype at, int a, Linetype bt, int b) { b 751 scripts/unifdef.c return op_strict(p, a < b, at, bt); b 753 scripts/unifdef.c static Linetype op_gt(int *p, Linetype at, int a, Linetype bt, int b) { b 754 scripts/unifdef.c return op_strict(p, a > b, at, bt); b 756 scripts/unifdef.c static Linetype op_le(int *p, Linetype at, int a, Linetype bt, int b) { b 757 scripts/unifdef.c return op_strict(p, a <= b, at, bt); b 759 scripts/unifdef.c static Linetype op_ge(int *p, Linetype at, int a, Linetype bt, int b) { b 760 scripts/unifdef.c return op_strict(p, a >= b, at, bt); b 762 scripts/unifdef.c static Linetype op_eq(int *p, Linetype at, int a, Linetype bt, int b) { b 763 scripts/unifdef.c return op_strict(p, a == b, at, bt); b 765 scripts/unifdef.c static Linetype op_ne(int *p, Linetype at, int a, Linetype bt, int b) { b 766 scripts/unifdef.c return op_strict(p, a != b, at, bt); b 768 scripts/unifdef.c static Linetype op_or(int *p, Linetype at, int a, Linetype bt, int b) { b 771 scripts/unifdef.c return op_strict(p, a || b, at, bt); b 773 scripts/unifdef.c static Linetype op_and(int *p, Linetype at, int a, Linetype bt, int b) { b 776 scripts/unifdef.c return op_strict(p, a && b, at, bt); b 297 security/apparmor/include/label.h struct aa_label *b); b 298 security/apparmor/include/label.h struct aa_label *aa_label_find_merge(struct aa_label *a, struct aa_label *b); b 299 security/apparmor/include/label.h struct aa_label *aa_label_merge(struct aa_label *a, struct aa_label *b, b 107 security/apparmor/label.c static int ns_cmp(struct aa_ns *a, struct aa_ns *b) b 112 security/apparmor/label.c AA_BUG(!b); b 114 security/apparmor/label.c AA_BUG(!b->base.hname); b 116 security/apparmor/label.c if (a == b) b 119 security/apparmor/label.c res = a->level - b->level; b 123 security/apparmor/label.c return strcmp(a->base.hname, b->base.hname); b 135 security/apparmor/label.c static int profile_cmp(struct aa_profile *a, struct aa_profile *b) b 140 security/apparmor/label.c AA_BUG(!b); b 142 security/apparmor/label.c AA_BUG(!b->ns); b 144 security/apparmor/label.c AA_BUG(!b->base.hname); b 146 security/apparmor/label.c if (a == b || a->base.hname == b->base.hname) b 148 security/apparmor/label.c res = ns_cmp(a->ns, b->ns); b 152 security/apparmor/label.c return strcmp(a->base.hname, b->base.hname); b 165 security/apparmor/label.c static int vec_cmp(struct aa_profile **a, int an, struct aa_profile **b, int bn) b 171 security/apparmor/label.c AA_BUG(!b); b 172 security/apparmor/label.c AA_BUG(!*b); b 177 security/apparmor/label.c int res = profile_cmp(a[i], b[i]); b 214 security/apparmor/label.c static int sort_cmp(const void *a, const void *b) b 216 security/apparmor/label.c return profile_cmp(*(struct aa_profile **)a, *(struct aa_profile **)b); b 469 security/apparmor/label.c static int label_cmp(struct aa_label *a, struct aa_label *b) b 471 security/apparmor/label.c AA_BUG(!b); b 473 security/apparmor/label.c if (a == b) b 476 security/apparmor/label.c return vec_cmp(a->vec, a->size, b->vec, b->size); b 934 security/apparmor/label.c struct aa_label *b) b 937 security/apparmor/label.c AA_BUG(!b); b 942 security/apparmor/label.c AA_BUG(I->j > b->size); b 945 security/apparmor/label.c if (I->j < b->size) { b 946 security/apparmor/label.c int res = profile_cmp(a->vec[I->i], b->vec[I->j]); b 949 security/apparmor/label.c return b->vec[(I->j)++]; b 957 security/apparmor/label.c if (I->j < b->size) b 958 security/apparmor/label.c return b->vec[(I->j)++]; b 975 security/apparmor/label.c static int label_merge_cmp(struct aa_label *a, struct aa_label *b, b 983 security/apparmor/label.c AA_BUG(!b); b 987 security/apparmor/label.c k < z->size && (p = aa_label_next_in_merge(&i, a, b)); b 1021 security/apparmor/label.c struct aa_label *b) b 1033 security/apparmor/label.c AA_BUG(!b); b 1034 security/apparmor/label.c AA_BUG(b->size < 0); b 1036 security/apparmor/label.c AA_BUG(new->size < a->size + b->size); b 1038 security/apparmor/label.c label_for_each_in_merge(i, a, b, next) { b 1070 security/apparmor/label.c else if (k == b->size) b 1071 security/apparmor/label.c return aa_get_label(b); b 1091 security/apparmor/label.c struct aa_label *b) b 1094 security/apparmor/label.c struct aa_ns *nsb = labels_ns(b); b 1114 security/apparmor/label.c struct aa_label *b) b 1120 security/apparmor/label.c AA_BUG(!b); b 1122 security/apparmor/label.c if (a == b) b 1129 security/apparmor/label.c int result = label_merge_cmp(a, b, this); b 1153 security/apparmor/label.c struct aa_label *aa_label_find_merge(struct aa_label *a, struct aa_label *b) b 1160 security/apparmor/label.c AA_BUG(!b); b 1164 security/apparmor/label.c if (label_is_stale(b)) b 1165 security/apparmor/label.c b = br = aa_get_newest_label(b); b 1166 security/apparmor/label.c ls = labelset_of_merge(a, b); b 1168 security/apparmor/label.c label = __label_find_merge(ls, a, b); b 1190 security/apparmor/label.c struct aa_label *aa_label_merge(struct aa_label *a, struct aa_label *b, b 1196 security/apparmor/label.c AA_BUG(!b); b 1198 security/apparmor/label.c if (a == b) b 1211 security/apparmor/label.c b = aa_get_newest_label(b); b 1216 security/apparmor/label.c new = aa_label_alloc(a->size + b->size, NULL, gfp); b 1220 security/apparmor/label.c label = label_merge_insert(new, a, b); b 1224 security/apparmor/label.c aa_put_label(b); b 372 security/apparmor/match.c u32 b = (base)[(state)]; \ b 373 security/apparmor/match.c unsigned int pos = base_idx(b) + (C); \ b 376 security/apparmor/match.c if (b & MATCH_FLAG_DIFF_ENCODE) \ b 598 security/device_cgroup.c const char *b; b 608 security/device_cgroup.c b = buffer; b 610 security/device_cgroup.c switch (*b) { b 649 security/device_cgroup.c b++; b 650 security/device_cgroup.c if (!isspace(*b)) b 652 security/device_cgroup.c b++; b 653 security/device_cgroup.c if (*b == '*') { b 655 security/device_cgroup.c b++; b 656 security/device_cgroup.c } else if (isdigit(*b)) { b 659 security/device_cgroup.c temp[count] = *b; b 660 security/device_cgroup.c b++; b 661 security/device_cgroup.c if (!isdigit(*b)) b 670 security/device_cgroup.c if (*b != ':') b 672 security/device_cgroup.c b++; b 675 security/device_cgroup.c if (*b == '*') { b 677 security/device_cgroup.c b++; b 678 security/device_cgroup.c } else if (isdigit(*b)) { b 681 security/device_cgroup.c temp[count] = *b; b 682 security/device_cgroup.c b++; b 683 security/device_cgroup.c if (!isdigit(*b)) b 692 security/device_cgroup.c if (!isspace(*b)) b 694 security/device_cgroup.c for (b++, count = 0; count < 3; count++, b++) { b 695 security/device_cgroup.c switch (*b) { b 58 security/integrity/platform_certs/efi_parser.c list.signature_type.b, list.signature_list_size, b 329 security/keys/keyring.c const struct keyring_index_key *b = data; b 335 security/keys/keyring.c seg_b = b->hash; b 345 security/keys/keyring.c seg_b = b->x; b 352 security/keys/keyring.c seg_b = (unsigned long)b->type; b 358 security/keys/keyring.c seg_b = (unsigned long)b->domain_tag; b 369 security/keys/keyring.c seg_b = *(unsigned char *)(b->description + i); b 209 security/selinux/ss/conditional.c static int bool_isvalid(struct cond_bool_datum *b) b 211 security/selinux/ss/conditional.c if (!(b->state == 0 || b->state == 1)) b 845 security/tomoyo/common.c const struct tomoyo_acl_head *b) b 848 security/tomoyo/common.c container_of(b, struct tomoyo_manager, head)->manager; b 1044 security/tomoyo/common.c const struct tomoyo_acl_info *b) b 1047 security/tomoyo/common.c const struct tomoyo_task_acl *p2 = container_of(b, typeof(*p2), head); b 1166 security/tomoyo/common.h const struct tomoyo_path_info *b) b 1168 security/tomoyo/common.h return a->hash != b->hash || strcmp(a->name, b->name); b 1234 security/tomoyo/common.h (const struct tomoyo_name_union *a, const struct tomoyo_name_union *b) b 1236 security/tomoyo/common.h return a->filename == b->filename && a->group == b->group; b 1248 security/tomoyo/common.h (const struct tomoyo_number_union *a, const struct tomoyo_number_union *b) b 1250 security/tomoyo/common.h return a->values[0] == b->values[0] && a->values[1] == b->values[1] && b 1251 security/tomoyo/common.h a->group == b->group && a->value_type[0] == b->value_type[0] && b 1252 security/tomoyo/common.h a->value_type[1] == b->value_type[1]; b 1264 security/tomoyo/common.h (const struct tomoyo_ipaddr_union *a, const struct tomoyo_ipaddr_union *b) b 1266 security/tomoyo/common.h return !memcmp(a->ip, b->ip, sizeof(a->ip)) && a->group == b->group && b 1267 security/tomoyo/common.h a->is_ipv6 == b->is_ipv6; b 358 security/tomoyo/condition.c const struct tomoyo_condition *b) b 360 security/tomoyo/condition.c return a->size == b->size && a->condc == b->condc && b 361 security/tomoyo/condition.c a->numbers_count == b->numbers_count && b 362 security/tomoyo/condition.c a->names_count == b->names_count && b 363 security/tomoyo/condition.c a->argc == b->argc && a->envc == b->envc && b 364 security/tomoyo/condition.c a->grant_log == b->grant_log && a->transit == b->transit && b 365 security/tomoyo/condition.c !memcmp(a + 1, b + 1, a->size - sizeof(*a)); b 74 security/tomoyo/domain.c const struct tomoyo_acl_info *b) b 76 security/tomoyo/domain.c return a->type == b->type && a->cond == b->cond; b 220 security/tomoyo/domain.c const struct tomoyo_acl_head *b) b 225 security/tomoyo/domain.c const struct tomoyo_transition_control *p2 = container_of(b, b 383 security/tomoyo/domain.c const struct tomoyo_acl_head *b) b 387 security/tomoyo/domain.c const struct tomoyo_aggregator *p2 = container_of(b, typeof(*p2), b 77 security/tomoyo/environ.c const struct tomoyo_acl_info *b) b 80 security/tomoyo/environ.c const struct tomoyo_env_acl *p2 = container_of(b, typeof(*p2), head); b 342 security/tomoyo/file.c const struct tomoyo_acl_info *b) b 345 security/tomoyo/file.c const struct tomoyo_path_acl *p2 = container_of(b, typeof(*p2), head); b 360 security/tomoyo/file.c struct tomoyo_acl_info *b, b 366 security/tomoyo/file.c const u16 b_perm = container_of(b, struct tomoyo_path_acl, head)->perm; b 414 security/tomoyo/file.c const struct tomoyo_acl_info *b) b 417 security/tomoyo/file.c const struct tomoyo_mkdev_acl *p2 = container_of(b, typeof(*p2), head); b 435 security/tomoyo/file.c struct tomoyo_acl_info *b, b 441 security/tomoyo/file.c const u8 b_perm = container_of(b, struct tomoyo_mkdev_acl, head) b 496 security/tomoyo/file.c const struct tomoyo_acl_info *b) b 499 security/tomoyo/file.c const struct tomoyo_path2_acl *p2 = container_of(b, typeof(*p2), head); b 515 security/tomoyo/file.c struct tomoyo_acl_info *b, b 521 security/tomoyo/file.c const u8 b_perm = container_of(b, struct tomoyo_path2_acl, head)->perm; b 632 security/tomoyo/file.c const struct tomoyo_acl_info *b) b 636 security/tomoyo/file.c const struct tomoyo_path_number_acl *p2 = container_of(b, typeof(*p2), b 653 security/tomoyo/file.c struct tomoyo_acl_info *b, b 659 security/tomoyo/file.c const u8 b_perm = container_of(b, struct tomoyo_path_number_acl, head) b 963 security/tomoyo/file.c const struct tomoyo_acl_info *b) b 966 security/tomoyo/file.c const struct tomoyo_mount_acl *p2 = container_of(b, typeof(*p2), head); b 22 security/tomoyo/group.c const struct tomoyo_acl_head *b) b 25 security/tomoyo/group.c container_of(b, struct tomoyo_path_group, head)->member_name; b 37 security/tomoyo/group.c const struct tomoyo_acl_head *b) b 41 security/tomoyo/group.c &container_of(b, struct tomoyo_number_group, head) b 56 security/tomoyo/group.c const struct tomoyo_acl_head *b) b 60 security/tomoyo/group.c const struct tomoyo_address_group *p2 = container_of(b, typeof(*p2), b 193 security/tomoyo/network.c const struct tomoyo_acl_info *b) b 196 security/tomoyo/network.c const struct tomoyo_inet_acl *p2 = container_of(b, typeof(*p2), head); b 212 security/tomoyo/network.c const struct tomoyo_acl_info *b) b 215 security/tomoyo/network.c const struct tomoyo_unix_acl *p2 = container_of(b, typeof(*p2), head); b 231 security/tomoyo/network.c struct tomoyo_acl_info *b, b 237 security/tomoyo/network.c const u8 b_perm = container_of(b, struct tomoyo_inet_acl, head)->perm; b 257 security/tomoyo/network.c struct tomoyo_acl_info *b, b 263 security/tomoyo/network.c const u8 b_perm = container_of(b, struct tomoyo_unix_acl, head)->perm; b 318 sound/core/oss/pcm_oss.c int b, int bdir, b 323 sound/core/oss/pcm_oss.c *c = a - b; b 334 sound/core/oss/pcm_oss.c unsigned int b, int bdir) b 342 sound/core/oss/pcm_oss.c b--; b 346 sound/core/oss/pcm_oss.c return a < b || (a == b && adir < bdir); b 503 sound/core/pcm_lib.c static inline unsigned int div32(unsigned int a, unsigned int b, b 506 sound/core/pcm_lib.c if (b == 0) { b 510 sound/core/pcm_lib.c *r = a % b; b 511 sound/core/pcm_lib.c return a / b; b 514 sound/core/pcm_lib.c static inline unsigned int div_down(unsigned int a, unsigned int b) b 516 sound/core/pcm_lib.c if (b == 0) b 518 sound/core/pcm_lib.c return a / b; b 521 sound/core/pcm_lib.c static inline unsigned int div_up(unsigned int a, unsigned int b) b 525 sound/core/pcm_lib.c if (b == 0) b 527 sound/core/pcm_lib.c q = div32(a, b, &r); b 533 sound/core/pcm_lib.c static inline unsigned int mul(unsigned int a, unsigned int b) b 537 sound/core/pcm_lib.c if (div_down(UINT_MAX, a) < b) b 539 sound/core/pcm_lib.c return a * b; b 542 sound/core/pcm_lib.c static inline unsigned int muldiv32(unsigned int a, unsigned int b, b 545 sound/core/pcm_lib.c u_int64_t n = (u_int64_t) a * b; b 646 sound/core/pcm_lib.c void snd_interval_mul(const struct snd_interval *a, const struct snd_interval *b, struct snd_interval *c) b 648 sound/core/pcm_lib.c if (a->empty || b->empty) { b 653 sound/core/pcm_lib.c c->min = mul(a->min, b->min); b 654 sound/core/pcm_lib.c c->openmin = (a->openmin || b->openmin); b 655 sound/core/pcm_lib.c c->max = mul(a->max, b->max); b 656 sound/core/pcm_lib.c c->openmax = (a->openmax || b->openmax); b 657 sound/core/pcm_lib.c c->integer = (a->integer && b->integer); b 670 sound/core/pcm_lib.c void snd_interval_div(const struct snd_interval *a, const struct snd_interval *b, struct snd_interval *c) b 673 sound/core/pcm_lib.c if (a->empty || b->empty) { b 678 sound/core/pcm_lib.c c->min = div32(a->min, b->max, &r); b 679 sound/core/pcm_lib.c c->openmin = (r || a->openmin || b->openmax); b 680 sound/core/pcm_lib.c if (b->min > 0) { b 681 sound/core/pcm_lib.c c->max = div32(a->max, b->min, &r); b 686 sound/core/pcm_lib.c c->openmax = (a->openmax || b->openmin); b 705 sound/core/pcm_lib.c void snd_interval_muldivk(const struct snd_interval *a, const struct snd_interval *b, b 709 sound/core/pcm_lib.c if (a->empty || b->empty) { b 714 sound/core/pcm_lib.c c->min = muldiv32(a->min, b->min, k, &r); b 715 sound/core/pcm_lib.c c->openmin = (r || a->openmin || b->openmin); b 716 sound/core/pcm_lib.c c->max = muldiv32(a->max, b->max, k, &r); b 721 sound/core/pcm_lib.c c->openmax = (a->openmax || b->openmax); b 737 sound/core/pcm_lib.c const struct snd_interval *b, struct snd_interval *c) b 740 sound/core/pcm_lib.c if (a->empty || b->empty) { b 745 sound/core/pcm_lib.c c->min = muldiv32(a->min, k, b->max, &r); b 746 sound/core/pcm_lib.c c->openmin = (r || a->openmin || b->openmax); b 747 sound/core/pcm_lib.c if (b->min > 0) { b 748 sound/core/pcm_lib.c c->max = muldiv32(a->max, k, b->min, &r); b 753 sound/core/pcm_lib.c c->openmax = (a->openmax || b->openmin); b 14 sound/core/pcm_local.h const struct snd_interval *b, struct snd_interval *c); b 16 sound/core/pcm_local.h const struct snd_interval *b, struct snd_interval *c); b 18 sound/core/pcm_local.h const struct snd_interval *b, b 21 sound/core/pcm_local.h const struct snd_interval *b, struct snd_interval *c); b 87 sound/core/seq/seq_prioq.c struct snd_seq_event *b) b 91 sound/core/seq/seq_prioq.c return (snd_seq_compare_tick_time(&a->time.tick, &b->time.tick)); b 94 sound/core/seq/seq_prioq.c return (snd_seq_compare_real_time(&a->time.time, &b->time.time)); b 104 sound/core/seq/seq_prioq.c struct snd_seq_event *b) b 108 sound/core/seq/seq_prioq.c if (a->time.tick > b->time.tick) b 110 sound/core/seq/seq_prioq.c else if (a->time.tick == b->time.tick) b 116 sound/core/seq/seq_prioq.c if (a->time.time.tv_sec > b->time.time.tv_sec) b 118 sound/core/seq/seq_prioq.c else if (a->time.time.tv_sec == b->time.time.tv_sec) { b 119 sound/core/seq/seq_prioq.c if (a->time.time.tv_nsec > b->time.time.tv_nsec) b 121 sound/core/seq/seq_prioq.c else if (a->time.time.tv_nsec == b->time.time.tv_nsec) b 66 sound/core/seq/seq_timer.h static inline int snd_seq_compare_tick_time(snd_seq_tick_time_t *a, snd_seq_tick_time_t *b) b 69 sound/core/seq/seq_timer.h return (*a >= *b); b 72 sound/core/seq/seq_timer.h static inline int snd_seq_compare_real_time(snd_seq_real_time_t *a, snd_seq_real_time_t *b) b 75 sound/core/seq/seq_timer.h if (a->tv_sec > b->tv_sec) b 77 sound/core/seq/seq_timer.h if ((a->tv_sec == b->tv_sec) && (a->tv_nsec >= b->tv_nsec)) b 48 sound/drivers/vx/vx_mixer.c } b; b 59 sound/drivers/vx/vx_mixer.c } b; b 63 sound/drivers/vx/vx_mixer.c #define SET_CDC_DATA_SEL(di,s) ((di).b.mh = (u8) (s)) b 64 sound/drivers/vx/vx_mixer.c #define SET_CDC_DATA_REG(di,r) ((di).b.ml = (u8) (r)) b 65 sound/drivers/vx/vx_mixer.c #define SET_CDC_DATA_VAL(di,d) ((di).b.ll = (u8) (d)) b 301 sound/firewire/amdtp-am824.c u8 *b; b 304 sound/firewire/amdtp-am824.c b = (u8 *)&buffer[p->midi_position]; b 310 sound/firewire/amdtp-am824.c snd_rawmidi_transmit(p->midi[port], &b[1], 1) == 1) { b 312 sound/firewire/amdtp-am824.c b[0] = 0x81; b 314 sound/firewire/amdtp-am824.c b[0] = 0x80; b 315 sound/firewire/amdtp-am824.c b[1] = 0; b 317 sound/firewire/amdtp-am824.c b[2] = 0; b 318 sound/firewire/amdtp-am824.c b[3] = 0; b 329 sound/firewire/amdtp-am824.c u8 *b; b 338 sound/firewire/amdtp-am824.c b = (u8 *)&buffer[p->midi_position]; b 340 sound/firewire/amdtp-am824.c len = b[0] - 0x80; b 342 sound/firewire/amdtp-am824.c snd_rawmidi_receive(p->midi[port], b + 1, len); b 160 sound/firewire/bebob/bebob_proc.c void (*op)(struct snd_info_entry *e, struct snd_info_buffer *b)) b 254 sound/firewire/digi00x/amdtp-dot.c u8 *b; b 258 sound/firewire/digi00x/amdtp-dot.c b = (u8 *)&buffer[0]; b 264 sound/firewire/digi00x/amdtp-dot.c len = snd_rawmidi_transmit(p->midi[port], b + 1, 2); b 274 sound/firewire/digi00x/amdtp-dot.c b[3] = 0xe0; b 276 sound/firewire/digi00x/amdtp-dot.c b[3] = 0x20; b 278 sound/firewire/digi00x/amdtp-dot.c b[3] = 0x00; b 279 sound/firewire/digi00x/amdtp-dot.c b[3] |= len; b 282 sound/firewire/digi00x/amdtp-dot.c b[1] = 0; b 283 sound/firewire/digi00x/amdtp-dot.c b[2] = 0; b 284 sound/firewire/digi00x/amdtp-dot.c b[3] = 0; b 286 sound/firewire/digi00x/amdtp-dot.c b[0] = 0x80; b 297 sound/firewire/digi00x/amdtp-dot.c u8 *b; b 300 sound/firewire/digi00x/amdtp-dot.c b = (u8 *)&buffer[0]; b 302 sound/firewire/digi00x/amdtp-dot.c len = b[3] & 0x0f; b 309 sound/firewire/digi00x/amdtp-dot.c if (b[3] >> 4 > 0) b 315 sound/firewire/digi00x/amdtp-dot.c snd_rawmidi_receive(p->midi[port], b + 1, len); b 38 sound/firewire/fireface/ff-proc.c struct snd_info_buffer *b)) b 196 sound/firewire/fireworks/fireworks_proc.c void (*op)(struct snd_info_entry *e, struct snd_info_buffer *b)) b 243 sound/firewire/motu/amdtp-motu.c u8 *b; b 247 sound/firewire/motu/amdtp-motu.c b = (u8 *)buffer; b 250 sound/firewire/motu/amdtp-motu.c snd_rawmidi_transmit(midi, b + p->midi_byte_offset, 1) == 1) { b 251 sound/firewire/motu/amdtp-motu.c b[p->midi_flag_offset] = 0x01; b 253 sound/firewire/motu/amdtp-motu.c b[p->midi_byte_offset] = 0x00; b 254 sound/firewire/motu/amdtp-motu.c b[p->midi_flag_offset] = 0x00; b 269 sound/firewire/motu/amdtp-motu.c u8 *b; b 273 sound/firewire/motu/amdtp-motu.c b = (u8 *)buffer; b 276 sound/firewire/motu/amdtp-motu.c if (midi && (b[p->midi_flag_offset] & 0x01)) b 277 sound/firewire/motu/amdtp-motu.c snd_rawmidi_receive(midi, b + p->midi_byte_offset, 1); b 84 sound/firewire/motu/motu-proc.c struct snd_info_buffer *b)) b 80 sound/firewire/oxfw/oxfw-proc.c struct snd_info_buffer *b)) b 21 sound/firewire/packets-buffer.c int iso_packets_buffer_init(struct iso_packets_buffer *b, struct fw_unit *unit, b 30 sound/firewire/packets-buffer.c b->packets = kmalloc_array(count, sizeof(*b->packets), GFP_KERNEL); b 31 sound/firewire/packets-buffer.c if (!b->packets) { b 44 sound/firewire/packets-buffer.c err = fw_iso_buffer_init(&b->iso_buffer, fw_parent_device(unit)->card, b 51 sound/firewire/packets-buffer.c p = page_address(b->iso_buffer.pages[page_index]); b 53 sound/firewire/packets-buffer.c b->packets[i].buffer = p + offset_in_page; b 54 sound/firewire/packets-buffer.c b->packets[i].offset = page_index * PAGE_SIZE + offset_in_page; b 60 sound/firewire/packets-buffer.c kfree(b->packets); b 71 sound/firewire/packets-buffer.c void iso_packets_buffer_destroy(struct iso_packets_buffer *b, b 74 sound/firewire/packets-buffer.c fw_iso_buffer_destroy(&b->iso_buffer, fw_parent_device(unit)->card); b 75 sound/firewire/packets-buffer.c kfree(b->packets); b 21 sound/firewire/packets-buffer.h int iso_packets_buffer_init(struct iso_packets_buffer *b, struct fw_unit *unit, b 24 sound/firewire/packets-buffer.h void iso_packets_buffer_destroy(struct iso_packets_buffer *b, b 55 sound/firewire/tascam/tascam-proc.c struct snd_info_buffer *b)) b 256 sound/firewire/tascam/tascam-transaction.c u8 *b; b 264 sound/firewire/tascam/tascam-transaction.c b = (u8 *)(buf + i * 2); b 266 sound/firewire/tascam/tascam-transaction.c port = b[0] >> 4; b 272 sound/firewire/tascam/tascam-transaction.c bytes = calculate_message_bytes(b[1]); b 277 sound/firewire/tascam/tascam-transaction.c if (b[bytes] == 0xf7) b 286 sound/firewire/tascam/tascam-transaction.c snd_rawmidi_receive(substream, b + 1, bytes); b 150 sound/isa/msnd/msnd_pinnacle_mixer.c #define update_volm(a, b) \ b 154 sound/isa/msnd/msnd_pinnacle_mixer.c dev->SMA + SMA_##b##Left); \ b 157 sound/isa/msnd/msnd_pinnacle_mixer.c dev->SMA + SMA_##b##Right); \ b 45 sound/isa/sb/emu8000_callback.c #define LIMITVALUE(x, a, b) do { if ((x) < (a)) (x) = (a); else if ((x) > (b)) (x) = (b); } while (0) b 32 sound/isa/sb/sb16_csp.c #define CSP_HDR_VALUE(a,b,c,d) ((a) | ((b)<<8) | ((c)<<16) | ((d)<<24)) b 34 sound/isa/sb/sb16_csp.c #define CSP_HDR_VALUE(a,b,c,d) ((d) | ((c)<<8) | ((b)<<16) | ((a)<<24)) b 725 sound/pci/asihpi/asihpi.c static inline unsigned int modulo_min(unsigned int a, unsigned int b, b 729 sound/pci/asihpi/asihpi.c if (((a-b) % modulus) < (modulus/2)) b 730 sound/pci/asihpi/asihpi.c result = b; b 54 sound/pci/asihpi/hpi6205.h struct hpi_fifo_buffer b; b 47 sound/pci/au88x0/au88x0_a3d.c a3dsrc_SetAtmosTarget(a3dsrc_t * a, short aa, short b, short c, short d, b 56 sound/pci/au88x0/au88x0_a3d.c (b << 0x10) | aa); b 62 sound/pci/au88x0/au88x0_a3d.c a3dsrc_SetAtmosCurrent(a3dsrc_t * a, short aa, short b, short c, short d, b 71 sound/pci/au88x0/au88x0_a3d.c (b << 0x10) | aa); b 88 sound/pci/au88x0/au88x0_a3d.c a3dsrc_GetAtmosTarget(a3dsrc_t * a, short *aa, short *b, short *c, b 124 sound/pci/au88x0/au88x0_a3d.c a3dsrc_SetHrtfTarget(a3dsrc_t * a, a3d_Hrtf_t const aa, a3d_Hrtf_t const b) b 133 sound/pci/au88x0/au88x0_a3d.c (b[i] << 0x10) | aa[i]); b 137 sound/pci/au88x0/au88x0_a3d.c a3dsrc_SetHrtfCurrent(a3dsrc_t * a, a3d_Hrtf_t const aa, a3d_Hrtf_t const b) b 146 sound/pci/au88x0/au88x0_a3d.c (b[i] << 0x10) | aa[i]); b 150 sound/pci/au88x0/au88x0_a3d.c a3dsrc_SetHrtfState(a3dsrc_t * a, a3d_Hrtf_t const aa, a3d_Hrtf_t const b) b 159 sound/pci/au88x0/au88x0_a3d.c (b[i] << 0x10) | aa[i]); b 172 sound/pci/au88x0/au88x0_a3d.c static void a3dsrc_GetHrtfTarget(a3dsrc_t * a, a3d_Hrtf_t aa, a3d_Hrtf_t b) b 183 sound/pci/au88x0/au88x0_a3d.c b[i] = b 189 sound/pci/au88x0/au88x0_a3d.c static void a3dsrc_GetHrtfCurrent(a3dsrc_t * a, a3d_Hrtf_t aa, a3d_Hrtf_t b) b 200 sound/pci/au88x0/au88x0_a3d.c b[i] = b 206 sound/pci/au88x0/au88x0_a3d.c static void a3dsrc_GetHrtfState(a3dsrc_t * a, a3d_Hrtf_t aa, a3d_Hrtf_t b) b 217 sound/pci/au88x0/au88x0_a3d.c b[i] = b 367 sound/pci/au88x0/au88x0_a3d.c static void CA3dIO_WriteReg(a3dsrc_t * a, unsigned long addr, short aa, short b) b 370 sound/pci/au88x0/au88x0_a3d.c hwwrite(vortex->mmio, addr, (aa << 0x10) | b); b 560 sound/pci/au88x0/au88x0_core.c unsigned int cr, unsigned int b, int sweep, int d, b 620 sound/pci/au88x0/au88x0_core.c hwwrite(card->mmio, VORTEX_SRC_U1 + (src << 2), b & 0xffff); b 99 sound/pci/au88x0/au88x0_eq.c static void vortex_EqHw_SetLeftStates(vortex_t * vortex, u16 a[], u16 b[]) b 108 sound/pci/au88x0/au88x0_eq.c hwwrite(vortex->mmio, 0x2b014 + (i * 0xc), b[i]); b 109 sound/pci/au88x0/au88x0_eq.c hwwrite(vortex->mmio, 0x2b018 + (i * 0xc), b[1 + i]); b 110 sound/pci/au88x0/au88x0_eq.c hwwrite(vortex->mmio, 0x2b01c + (i * 0xc), b[2 + i]); b 111 sound/pci/au88x0/au88x0_eq.c hwwrite(vortex->mmio, 0x2b020 + (i * 0xc), b[3 + i]); b 116 sound/pci/au88x0/au88x0_eq.c static void vortex_EqHw_SetRightStates(vortex_t * vortex, u16 a[], u16 b[]) b 125 sound/pci/au88x0/au88x0_eq.c hwwrite(vortex->mmio, 0x2b1f4 + (i * 0xc), b[i]); b 126 sound/pci/au88x0/au88x0_eq.c hwwrite(vortex->mmio, 0x2b1f8 + (i * 0xc), b[1 + i]); b 127 sound/pci/au88x0/au88x0_eq.c hwwrite(vortex->mmio, 0x2b1fc + (i * 0xc), b[2 + i]); b 128 sound/pci/au88x0/au88x0_eq.c hwwrite(vortex->mmio, 0x2b200 + (i * 0xc), b[3 + i]); b 134 sound/pci/au88x0/au88x0_eq.c static void vortex_EqHw_GetTimeConsts(vortex_t * vortex, u16 * a, u16 * b) b 137 sound/pci/au88x0/au88x0_eq.c *b = hwread(vortex->mmio, 0x2b3c8); b 150 sound/pci/au88x0/au88x0_eq.c static void vortex_EqHw_GetLeftStates(vortex_t * vortex, u16 * a, u16 b[]) b 155 sound/pci/au88x0/au88x0_eq.c static void vortex_EqHw_GetRightStates(vortex_t * vortex, u16 * a, u16 b[]) b 162 sound/pci/au88x0/au88x0_eq.c static void vortex_EqHw_SetBypassGain(vortex_t * vortex, u16 a, u16 b) b 167 sound/pci/au88x0/au88x0_eq.c hwwrite(vortex->mmio, 0x2b3ec, b); b 170 sound/pci/au88x0/au88x0_eq.c hwwrite(vortex->mmio, 0x2b3ec, sign_invert(b)); b 174 sound/pci/au88x0/au88x0_eq.c static void vortex_EqHw_SetA3DBypassGain(vortex_t * vortex, u16 a, u16 b) b 178 sound/pci/au88x0/au88x0_eq.c hwwrite(vortex->mmio, 0x2b3f8, b); b 182 sound/pci/au88x0/au88x0_eq.c static void vortex_EqHw_SetCurrBypassGain(vortex_t * vortex, u16 a, u16 b) b 186 sound/pci/au88x0/au88x0_eq.c hwwrite(vortex->mmio, 0x2b3e8, b); b 189 sound/pci/au88x0/au88x0_eq.c static void vortex_EqHw_SetCurrA3DBypassGain(vortex_t * vortex, u16 a, u16 b) b 193 sound/pci/au88x0/au88x0_eq.c hwwrite(vortex->mmio, 0x2b3f4, b); b 198 sound/pci/au88x0/au88x0_eq.c vortex_EqHw_SetLeftGainsSingleTarget(vortex_t * vortex, u16 index, u16 b) b 200 sound/pci/au88x0/au88x0_eq.c hwwrite(vortex->mmio, 0x2b02c + (index * 0x30), b); b 204 sound/pci/au88x0/au88x0_eq.c vortex_EqHw_SetRightGainsSingleTarget(vortex_t * vortex, u16 index, u16 b) b 206 sound/pci/au88x0/au88x0_eq.c hwwrite(vortex->mmio, 0x2b20c + (index * 0x30), b); b 595 sound/pci/au88x0/au88x0_eq.c vortex_Eqlzr_SetA3dBypassGain(vortex_t * vortex, u32 a, u32 b) b 601 sound/pci/au88x0/au88x0_eq.c eq->this5c = b; b 19 sound/pci/cs46xx/cs46xx_dsp_scb_types.h #define ___DSP_DUAL_16BIT_ALLOC(a,b) u16 a; u16 b; b 21 sound/pci/cs46xx/cs46xx_dsp_scb_types.h #define ___DSP_DUAL_16BIT_ALLOC(a,b) u16 b; u16 a; b 196 sound/pci/ctxfi/ctatc.c int b; b 203 sound/pci/ctxfi/ctatc.c for (b = 31; ((b >= 0) && !(input_rate >> b)); ) b 204 sound/pci/ctxfi/ctatc.c b--; b 206 sound/pci/ctxfi/ctatc.c if (b >= 0) { b 207 sound/pci/ctxfi/ctatc.c input_rate <<= (31 - b); b 209 sound/pci/ctxfi/ctatc.c b = 24 - (31 - b); b 210 sound/pci/ctxfi/ctatc.c if (b >= 0) b 211 sound/pci/ctxfi/ctatc.c input_rate <<= b; b 213 sound/pci/ctxfi/ctatc.c input_rate >>= -b; b 1784 sound/pci/ctxfi/cthw20k1.c #define CTLBITS(a, b, c, d) (((a) << 24) | ((b) << 16) | ((c) << 8) | (d)) b 44 sound/pci/emu10k1/emu10k1_callback.c #define LIMITVALUE(x, a, b) do { if ((x) < (a)) (x) = (a); else if ((x) > (b)) (x) = (b); } while (0) b 37 sound/pci/hda/hda_auto_parser.c const struct auto_out_pin *b = bp; b 38 sound/pci/hda/hda_auto_parser.c return (int)(a->seq - b->seq); b 71 sound/pci/hda/hda_auto_parser.c const struct auto_pin_cfg_item *b = bp; b 72 sound/pci/hda/hda_auto_parser.c if (a->type != b->type) b 73 sound/pci/hda/hda_auto_parser.c return (int)(a->type - b->type); b 77 sound/pci/hda/hda_auto_parser.c return (int)(b->has_boost_on_pin - a->has_boost_on_pin); b 4858 sound/pci/hda/hda_generic.c const struct automic_entry *b = bp; b 4859 sound/pci/hda/hda_generic.c return (int)(a->attr - b->attr); b 532 sound/pci/hda/hda_sysfs.c static inline int strmatch(const char *a, const char *b) b 534 sound/pci/hda/hda_sysfs.c return strncasecmp(a, b, strlen(b)) == 0; b 990 sound/pci/oxygen/xonar_wm87x6.c #define WM8776_FIELD_CTL_VOLUME(a, b, c, d, e, f, g, h, tlv_p) { \ b 991 sound/pci/oxygen/xonar_wm87x6.c _WM8776_FIELD_CTL(a " Capture Volume", b, c, d, e, f, g, h), \ b 229 sound/pci/riptide/riptide.c #define SEND_GETV(p,b) sendcmd(p,RESP,GETV,0,RET(b)) /* get version */ b 230 sound/pci/riptide/riptide.c #define SEND_GETC(p,b,c) sendcmd(p,PARM|RESP,GETC,c,RET(b)) b 231 sound/pci/riptide/riptide.c #define SEND_GUNS(p,b) sendcmd(p,RESP,GUNS,0,RET(b)) b 232 sound/pci/riptide/riptide.c #define SEND_SCID(p,b) sendcmd(p,RESP,SCID,0,RET(b)) b 233 sound/pci/riptide/riptide.c #define SEND_RMEM(p,b,c,d) sendcmd(p,PARM|RESP,RMEM|BYTE1(b),LONG0(c),RET(d)) /* memory access for firmware write */ b 234 sound/pci/riptide/riptide.c #define SEND_SMEM(p,b,c) sendcmd(p,PARM,SMEM|BYTE1(b),LONG0(c),RET(0)) /* memory access for firmware write */ b 235 sound/pci/riptide/riptide.c #define SEND_WMEM(p,b,c) sendcmd(p,PARM,WMEM|BYTE1(b),LONG0(c),RET(0)) /* memory access for firmware write */ b 236 sound/pci/riptide/riptide.c #define SEND_SDTM(p,b,c) sendcmd(p,PARM|RESP,SDTM|TRINIB1(b),0,RET(c)) /* memory access for firmware write */ b 237 sound/pci/riptide/riptide.c #define SEND_GOTO(p,b) sendcmd(p,PARM,GOTO,LONG0(b),RET(0)) /* memory access for firmware write */ b 239 sound/pci/riptide/riptide.c #define SEND_SSTR(p,b,c) sendcmd(p,PARM,SSTR|BYTE3(b),LONG0(c),RET(0)) /* start stream */ b 240 sound/pci/riptide/riptide.c #define SEND_PSTR(p,b) sendcmd(p,PARM,PSTR,BYTE3(b),RET(0)) /* pause stream */ b 241 sound/pci/riptide/riptide.c #define SEND_KSTR(p,b) sendcmd(p,PARM,KSTR,BYTE3(b),RET(0)) /* stop stream */ b 243 sound/pci/riptide/riptide.c #define SEND_GPOS(p,b,c,d) sendcmd(p,PARM|RESP,GPOS,BYTE3(c)|BYTE2(b),RET(d)) /* get position in dma */ b 244 sound/pci/riptide/riptide.c #define SEND_SETF(p,b,c,d,e,f,g) sendcmd(p,PARM,SETF|WORD1(b)|BYTE3(c),d|BYTE1(e)|BYTE2(f)|BYTE3(g),RET(0)) /* set sample format at mixer */ b 245 sound/pci/riptide/riptide.c #define SEND_GSTS(p,b,c,d) sendcmd(p,PARM|RESP,GSTS,BYTE3(c)|BYTE2(b),RET(d)) b 246 sound/pci/riptide/riptide.c #define SEND_NGPOS(p,b,c,d) sendcmd(p,PARM|RESP,NGPOS,BYTE3(c)|BYTE2(b),RET(d)) b 247 sound/pci/riptide/riptide.c #define SEND_PSEL(p,b,c) sendcmd(p,PARM,PSEL,BYTE2(b)|BYTE3(c),RET(0)) /* activate lbus path */ b 248 sound/pci/riptide/riptide.c #define SEND_PCLR(p,b,c) sendcmd(p,PARM,PCLR,BYTE2(b)|BYTE3(c),RET(0)) /* deactivate lbus path */ b 249 sound/pci/riptide/riptide.c #define SEND_PLST(p,b) sendcmd(p,PARM,PLST,BYTE3(b),RET(0)) b 250 sound/pci/riptide/riptide.c #define SEND_RSSV(p,b,c,d) sendcmd(p,PARM|RESP,RSSV,BYTE2(b)|BYTE3(c),RET(d)) b 251 sound/pci/riptide/riptide.c #define SEND_LSEL(p,b,c,d,e,f,g,h) sendcmd(p,PARM,LSEL|BYTE1(b)|BYTE2(c)|BYTE3(d),BYTE0(e)|BYTE1(f)|BYTE2(g)|BYTE3(h),RET(0)) /* select paths for internal connections */ b 252 sound/pci/riptide/riptide.c #define SEND_SSRC(p,b,c,d,e) sendcmd(p,PARM,SSRC|BYTE1(b)|WORD2(c),WORD0(d)|WORD2(e),RET(0)) /* configure source */ b 253 sound/pci/riptide/riptide.c #define SEND_SLST(p,b) sendcmd(p,PARM,SLST,BYTE3(b),RET(0)) b 254 sound/pci/riptide/riptide.c #define SEND_RSRC(p,b,c) sendcmd(p,RESP,RSRC|BYTE1(b),0,RET(c)) /* read source config */ b 255 sound/pci/riptide/riptide.c #define SEND_SSRB(p,b,c) sendcmd(p,PARM,SSRB|BYTE1(b),WORD2(c),RET(0)) b 256 sound/pci/riptide/riptide.c #define SEND_SDGV(p,b,c,d,e) sendcmd(p,PARM,SDGV|BYTE2(b)|BYTE3(c),WORD0(d)|WORD2(e),RET(0)) /* set digital mixer */ b 257 sound/pci/riptide/riptide.c #define SEND_RDGV(p,b,c,d) sendcmd(p,PARM|RESP,RDGV|BYTE2(b)|BYTE3(c),0,RET(d)) /* read digital mixer */ b 258 sound/pci/riptide/riptide.c #define SEND_DLST(p,b) sendcmd(p,PARM,DLST,BYTE3(b),RET(0)) b 259 sound/pci/riptide/riptide.c #define SEND_SACR(p,b,c) sendcmd(p,PARM,SACR,WORD0(b)|WORD2(c),RET(0)) /* set AC97 register */ b 260 sound/pci/riptide/riptide.c #define SEND_RACR(p,b,c) sendcmd(p,PARM|RESP,RACR,WORD2(b),RET(c)) /* get AC97 register */ b 261 sound/pci/riptide/riptide.c #define SEND_ALST(p,b) sendcmd(p,PARM,ALST,BYTE3(b),RET(0)) b 262 sound/pci/riptide/riptide.c #define SEND_TXAC(p,b,c,d,e,f) sendcmd(p,PARM,TXAC|BYTE1(b)|WORD2(c),WORD0(d)|BYTE2(e)|BYTE3(f),RET(0)) b 263 sound/pci/riptide/riptide.c #define SEND_RXAC(p,b,c,d) sendcmd(p,PARM|RESP,RXAC,BYTE2(b)|BYTE3(c),RET(d)) b 264 sound/pci/riptide/riptide.c #define SEND_SI2S(p,b) sendcmd(p,PARM,SI2S,WORD2(b),RET(0)) b 490 sound/pci/riptide/riptide.c unsigned char *b); b 959 sound/pci/riptide/riptide.c unsigned char *a, unsigned char *b) b 967 sound/pci/riptide/riptide.c *b = rptr.retbytes[1]; b 968 sound/pci/riptide/riptide.c snd_printdd("getsourcesink 0x%x 0x%x\n", *a, *b); b 141 sound/soc/codecs/ak4613.c #define AUDIO_IFACE(b, fmt) { b, SND_SOC_DAIFMT_##fmt } b 2711 sound/soc/codecs/arizona.c s16 b = be16_to_cpu(_b); b 2716 sound/soc/codecs/arizona.c if (abs(b) >= 4096) b 2719 sound/soc/codecs/arizona.c return (abs((a << 16) / (4096 - b)) >= 4096 << 4); b 27 sound/soc/codecs/cs42l51.h #define CS42L51_MK_CHIP_REV(a, b) ((a)<<3|(b)) b 4608 sound/soc/codecs/madera.c s16 b = be16_to_cpu(_b); b 4613 sound/soc/codecs/madera.c if (abs(b) >= 4096) b 4616 sound/soc/codecs/madera.c return (abs((a << 16) / (4096 - b)) >= 4096 << 4); b 114 sound/soc/codecs/tscs42xx.h #define RM(m, b) ((m)<<(b)) b 115 sound/soc/codecs/tscs42xx.h #define RV(v, b) ((v)<<(b)) b 249 sound/soc/codecs/wm8580.c struct pll_state b; b 484 sound/soc/codecs/wm8580.c state = &wm8580->b; b 122 sound/soc/fsl/fsl_asrc.h #define ASRCNCR_ANCi_SHIFT(i, b) (b * i) b 123 sound/soc/fsl/fsl_asrc.h #define ASRCNCR_ANCi_MASK(i, b) (((1 << b) - 1) << ASRCNCR_ANCi_SHIFT(i, b)) b 124 sound/soc/fsl/fsl_asrc.h #define ASRCNCR_ANCi(i, v, b) ((v << ASRCNCR_ANCi_SHIFT(i, b)) & ASRCNCR_ANCi_MASK(i, b)) b 42 sound/soc/samsung/s3c-i2s-v2.c #define bit_set(v, b) (((v) & (b)) ? 1 : 0) b 26 sound/soc/samsung/s3c24xx-i2s.h #define S3C24XX_PRESCALE(a,b) \ b 27 sound/soc/samsung/s3c24xx-i2s.h (((a - 1) << S3C2410_IISPSR_INTSHIFT) | ((b - 1) << S3C2410_IISPSR_EXTSHFIT)) b 1472 sound/soc/soc-dapm.c struct snd_soc_dapm_widget *b, b 1486 sound/soc/soc-dapm.c WARN_ONCE(sort[b->id] == 0, "offset b->id %d not initialized\n", b->id); b 1488 sound/soc/soc-dapm.c if (sort[a->id] != sort[b->id]) b 1489 sound/soc/soc-dapm.c return sort[a->id] - sort[b->id]; b 1490 sound/soc/soc-dapm.c if (a->subseq != b->subseq) { b 1492 sound/soc/soc-dapm.c return a->subseq - b->subseq; b 1494 sound/soc/soc-dapm.c return b->subseq - a->subseq; b 1496 sound/soc/soc-dapm.c if (a->reg != b->reg) b 1497 sound/soc/soc-dapm.c return a->reg - b->reg; b 1498 sound/soc/soc-dapm.c if (a->dapm != b->dapm) b 1499 sound/soc/soc-dapm.c return (unsigned long)a->dapm - (unsigned long)b->dapm; b 578 sound/sparc/dbri.c static __u32 reverse_bytes(__u32 b, int len) b 582 sound/sparc/dbri.c b = ((b & 0xffff0000) >> 16) | ((b & 0x0000ffff) << 16); b 585 sound/sparc/dbri.c b = ((b & 0xff00ff00) >> 8) | ((b & 0x00ff00ff) << 8); b 588 sound/sparc/dbri.c b = ((b & 0xf0f0f0f0) >> 4) | ((b & 0x0f0f0f0f) << 4); b 591 sound/sparc/dbri.c b = ((b & 0xcccccccc) >> 2) | ((b & 0x33333333) << 2); b 594 sound/sparc/dbri.c b = ((b & 0xaaaaaaaa) >> 1) | ((b & 0x55555555) << 1); b 602 sound/sparc/dbri.c return b; b 23 sound/synth/emux/emux_synth.c #define LIMITVALUE(x, a, b) do { if ((x) < (a)) (x) = (a); else if ((x) > (b)) (x) = (b); } while (0) b 131 sound/usb/caiaq/input.c static unsigned int decode_erp(unsigned char a, unsigned char b) b 151 sound/usb/caiaq/input.c pos_b = b - LOW_PEAK + DEG270; b 156 sound/usb/caiaq/input.c pos_b = HIGH_PEAK - b + DEG90; b 159 sound/usb/caiaq/input.c if (b > mid_value) b 558 sound/usb/midi.c uint8_t b, struct urb *urb) b 564 sound/usb/midi.c if (b >= 0xf8) { b 565 sound/usb/midi.c output_packet(urb, p0 | 0x0f, b, 0, 0); b 566 sound/usb/midi.c } else if (b >= 0xf0) { b 567 sound/usb/midi.c switch (b) { b 569 sound/usb/midi.c port->data[0] = b; b 574 sound/usb/midi.c port->data[0] = b; b 578 sound/usb/midi.c port->data[0] = b; b 606 sound/usb/midi.c } else if (b >= 0x80) { b 607 sound/usb/midi.c port->data[0] = b; b 608 sound/usb/midi.c if (b >= 0xc0 && b <= 0xdf) b 621 sound/usb/midi.c output_packet(urb, p0, port->data[0], b, 0); b 624 sound/usb/midi.c port->data[1] = b; b 635 sound/usb/midi.c output_packet(urb, p0, port->data[0], port->data[1], b); b 638 sound/usb/midi.c port->data[0] = b; b 642 sound/usb/midi.c port->data[1] = b; b 647 sound/usb/midi.c port->data[1], b); b 665 sound/usb/midi.c uint8_t b; b 666 sound/usb/midi.c if (snd_rawmidi_transmit(port->substream, &b, 1) != 1) { b 670 sound/usb/midi.c snd_usbmidi_transmit_byte(port, b, urb); b 1095 sound/usb/misc/ua101.c unsigned int b, u = 0; b 1097 sound/usb/misc/ua101.c for (b = 0; b < ARRAY_SIZE(stream->buffers); ++b) { b 1098 sound/usb/misc/ua101.c unsigned int size = stream->buffers[b].size; b 1099 sound/usb/misc/ua101.c u8 *addr = stream->buffers[b].addr; b 1100 sound/usb/misc/ua101.c dma_addr_t dma = stream->buffers[b].dma; b 933 sound/usb/mixer.c #define PTYPE(a, b) ((a) << 8 | (b)) b 73 tools/arch/x86/lib/insn.c insn_byte_t b, lb; b 81 tools/arch/x86/lib/insn.c b = peek_next(insn_byte_t, insn); b 82 tools/arch/x86/lib/insn.c attr = inat_get_opcode_attribute(b); b 86 tools/arch/x86/lib/insn.c if (prefixes->bytes[i] == b) b 91 tools/arch/x86/lib/insn.c prefixes->bytes[nb++] = b; b 105 tools/arch/x86/lib/insn.c lb = b; b 106 tools/arch/x86/lib/insn.c b = peek_next(insn_byte_t, insn); b 107 tools/arch/x86/lib/insn.c attr = inat_get_opcode_attribute(b); b 113 tools/arch/x86/lib/insn.c b = insn->prefixes.bytes[3]; b 116 tools/arch/x86/lib/insn.c prefixes->bytes[i] = b; b 123 tools/arch/x86/lib/insn.c b = peek_next(insn_byte_t, insn); b 124 tools/arch/x86/lib/insn.c attr = inat_get_opcode_attribute(b); b 126 tools/arch/x86/lib/insn.c insn->rex_prefix.value = b; b 129 tools/arch/x86/lib/insn.c if (X86_REX_W(b)) b 137 tools/arch/x86/lib/insn.c b = peek_next(insn_byte_t, insn); b 138 tools/arch/x86/lib/insn.c attr = inat_get_opcode_attribute(b); b 150 tools/arch/x86/lib/insn.c insn->vex_prefix.bytes[0] = b; b 672 tools/bpf/bpftool/prog.c const struct map_replace *a = p1, *b = p2; b 674 tools/bpf/bpftool/prog.c return a->idx - b->idx; b 4 tools/build/tests/ex/ex.c int b(void); b 14 tools/build/tests/ex/ex.c b(); b 39 tools/include/linux/compiler-gcc.h #define __printf(a, b) __attribute__((format(printf, a, b))) b 40 tools/include/linux/compiler-gcc.h #define __scanf(a, b) __attribute__((format(scanf, a, b))) b 27 tools/include/linux/compiler.h # define __same_type(a, b) __builtin_types_compatible_p(typeof(a), typeof(b)) b 37 tools/include/linux/irqflags.h #define trace_lock_acquire(a, b, c, d, e, f, g) b 35 tools/include/linux/jhash.h #define __jhash_mix(a, b, c) \ b 37 tools/include/linux/jhash.h a -= c; a ^= rol32(c, 4); c += b; \ b 38 tools/include/linux/jhash.h b -= a; b ^= rol32(a, 6); a += c; \ b 39 tools/include/linux/jhash.h c -= b; c ^= rol32(b, 8); b += a; \ b 40 tools/include/linux/jhash.h a -= c; a ^= rol32(c, 16); c += b; \ b 41 tools/include/linux/jhash.h b -= a; b ^= rol32(a, 19); a += c; \ b 42 tools/include/linux/jhash.h c -= b; c ^= rol32(b, 4); b += a; \ b 46 tools/include/linux/jhash.h #define __jhash_final(a, b, c) \ b 48 tools/include/linux/jhash.h c ^= b; c -= rol32(b, 14); \ b 50 tools/include/linux/jhash.h b ^= a; b -= rol32(a, 25); \ b 51 tools/include/linux/jhash.h c ^= b; c -= rol32(b, 16); \ b 53 tools/include/linux/jhash.h b ^= a; b -= rol32(a, 14); \ b 54 tools/include/linux/jhash.h c ^= b; c -= rol32(b, 24); \ b 72 tools/include/linux/jhash.h u32 a, b, c; b 76 tools/include/linux/jhash.h a = b = c = JHASH_INITVAL + length + initval; b 81 tools/include/linux/jhash.h b += __get_unaligned_cpu32(k + 4); b 83 tools/include/linux/jhash.h __jhash_mix(a, b, c); b 94 tools/include/linux/jhash.h case 8: b += (u32)k[7]<<24; b 95 tools/include/linux/jhash.h case 7: b += (u32)k[6]<<16; b 96 tools/include/linux/jhash.h case 6: b += (u32)k[5]<<8; b 97 tools/include/linux/jhash.h case 5: b += k[4]; b 102 tools/include/linux/jhash.h __jhash_final(a, b, c); b 119 tools/include/linux/jhash.h u32 a, b, c; b 122 tools/include/linux/jhash.h a = b = c = JHASH_INITVAL + (length<<2) + initval; b 127 tools/include/linux/jhash.h b += k[1]; b 129 tools/include/linux/jhash.h __jhash_mix(a, b, c); b 137 tools/include/linux/jhash.h case 2: b += k[1]; b 139 tools/include/linux/jhash.h __jhash_final(a, b, c); b 149 tools/include/linux/jhash.h static inline u32 __jhash_nwords(u32 a, u32 b, u32 c, u32 initval) b 152 tools/include/linux/jhash.h b += initval; b 155 tools/include/linux/jhash.h __jhash_final(a, b, c); b 160 tools/include/linux/jhash.h static inline u32 jhash_3words(u32 a, u32 b, u32 c, u32 initval) b 162 tools/include/linux/jhash.h return __jhash_nwords(a, b, c, initval + JHASH_INITVAL + (3 << 2)); b 165 tools/include/linux/jhash.h static inline u32 jhash_2words(u32 a, u32 b, u32 initval) b 167 tools/include/linux/jhash.h return __jhash_nwords(a, b, 0, initval + JHASH_INITVAL + (2 << 2)); b 49 tools/include/linux/overflow.h #define check_add_overflow(a, b, d) ({ \ b 51 tools/include/linux/overflow.h typeof(b) __b = (b); \ b 58 tools/include/linux/overflow.h #define check_sub_overflow(a, b, d) ({ \ b 60 tools/include/linux/overflow.h typeof(b) __b = (b); \ b 67 tools/include/linux/overflow.h #define check_mul_overflow(a, b, d) ({ \ b 69 tools/include/linux/overflow.h typeof(b) __b = (b); \ b 80 tools/include/linux/overflow.h #define __unsigned_add_overflow(a, b, d) ({ \ b 82 tools/include/linux/overflow.h typeof(b) __b = (b); \ b 89 tools/include/linux/overflow.h #define __unsigned_sub_overflow(a, b, d) ({ \ b 91 tools/include/linux/overflow.h typeof(b) __b = (b); \ b 101 tools/include/linux/overflow.h #define __unsigned_mul_overflow(a, b, d) ({ \ b 103 tools/include/linux/overflow.h typeof(b) __b = (b); \ b 129 tools/include/linux/overflow.h #define __signed_add_overflow(a, b, d) ({ \ b 131 tools/include/linux/overflow.h typeof(b) __b = (b); \ b 145 tools/include/linux/overflow.h #define __signed_sub_overflow(a, b, d) ({ \ b 147 tools/include/linux/overflow.h typeof(b) __b = (b); \ b 172 tools/include/linux/overflow.h #define __signed_mul_overflow(a, b, d) ({ \ b 174 tools/include/linux/overflow.h typeof(b) __b = (b); \ b 187 tools/include/linux/overflow.h #define check_add_overflow(a, b, d) \ b 189 tools/include/linux/overflow.h __signed_add_overflow(a, b, d), \ b 190 tools/include/linux/overflow.h __unsigned_add_overflow(a, b, d)) b 192 tools/include/linux/overflow.h #define check_sub_overflow(a, b, d) \ b 194 tools/include/linux/overflow.h __signed_sub_overflow(a, b, d), \ b 195 tools/include/linux/overflow.h __unsigned_sub_overflow(a, b, d)) b 197 tools/include/linux/overflow.h #define check_mul_overflow(a, b, d) \ b 199 tools/include/linux/overflow.h __signed_mul_overflow(a, b, d), \ b 200 tools/include/linux/overflow.h __unsigned_mul_overflow(a, b, d)) b 216 tools/include/linux/overflow.h static inline __must_check size_t array_size(size_t a, size_t b) b 220 tools/include/linux/overflow.h if (check_mul_overflow(a, b, &bytes)) b 238 tools/include/linux/overflow.h static inline __must_check size_t array3_size(size_t a, size_t b, size_t c) b 242 tools/include/linux/overflow.h if (check_mul_overflow(a, b, &bytes)) b 2297 tools/include/nolibc/nolibc.h void *memset(void *dst, int b, size_t len) b 2302 tools/include/nolibc/nolibc.h *(p++) = b; b 35 tools/io_uring/io_uring-bench.c #define min(a, b) ((a < b) ? (a) : (b)) b 500 tools/lib/bpf/btf.c const struct btf_var_secinfo *b = _b; b 502 tools/lib/bpf/btf.c return a->offset - b->offset; b 100 tools/lib/bpf/btf_dump.c static bool str_equal_fn(const void *a, const void *b, void *ctx) b 102 tools/lib/bpf/btf_dump.c return strcmp(a, b) == 0; b 67 tools/lib/bpf/libbpf.c #define __printf(a, b) __attribute__((format(printf, a, b))) b 657 tools/lib/bpf/libbpf.c const struct bpf_map *b = _b; b 659 tools/lib/bpf/libbpf.c if (a->sec_idx != b->sec_idx) b 660 tools/lib/bpf/libbpf.c return a->sec_idx - b->sec_idx; b 661 tools/lib/bpf/libbpf.c return a->sec_offset - b->sec_offset; b 1000 tools/lib/bpf/libbpf.c char *b; b 1001 tools/lib/bpf/libbpf.c for (b = ((char *)def) + sizeof(struct bpf_map_def); b 1002 tools/lib/bpf/libbpf.c b < ((char *)def) + map_def_sz; b++) { b 1003 tools/lib/bpf/libbpf.c if (*b != 0) { b 6 tools/lib/lockdep/tests/ABA.c pthread_mutex_t a, b; b 9 tools/lib/lockdep/tests/ABA.c pthread_mutex_init(&b, NULL); b 12 tools/lib/lockdep/tests/ABA.c pthread_mutex_lock(&b); b 7 tools/lib/lockdep/tests/ABBA.c pthread_mutex_t a, b; b 10 tools/lib/lockdep/tests/ABBA.c pthread_mutex_init(&b, NULL); b 12 tools/lib/lockdep/tests/ABBA.c LOCK_UNLOCK_2(a, b); b 13 tools/lib/lockdep/tests/ABBA.c LOCK_UNLOCK_2(b, a); b 15 tools/lib/lockdep/tests/ABBA.c pthread_mutex_destroy(&b); b 19 tools/lib/lockdep/tests/ABBA.c pthread_mutex_init(&b, NULL); b 21 tools/lib/lockdep/tests/ABBA.c LOCK_UNLOCK_2(a, b); b 22 tools/lib/lockdep/tests/ABBA.c LOCK_UNLOCK_2(b, a); b 24 tools/lib/lockdep/tests/ABBA.c pthread_mutex_destroy(&b); b 6 tools/lib/lockdep/tests/ABBA_2threads.c pthread_mutex_t b = PTHREAD_MUTEX_INITIALIZER; b 13 tools/lib/lockdep/tests/ABBA_2threads.c pthread_mutex_lock(&b); b 21 tools/lib/lockdep/tests/ABBA_2threads.c pthread_mutex_unlock(&b); b 39 tools/lib/lockdep/tests/ABBA_2threads.c pthread_mutex_lock(&b); b 41 tools/lib/lockdep/tests/ABBA_2threads.c pthread_mutex_unlock(&b); b 7 tools/lib/lockdep/tests/ABBCCA.c pthread_mutex_t a, b, c; b 10 tools/lib/lockdep/tests/ABBCCA.c pthread_mutex_init(&b, NULL); b 13 tools/lib/lockdep/tests/ABBCCA.c LOCK_UNLOCK_2(a, b); b 14 tools/lib/lockdep/tests/ABBCCA.c LOCK_UNLOCK_2(b, c); b 18 tools/lib/lockdep/tests/ABBCCA.c pthread_mutex_destroy(&b); b 7 tools/lib/lockdep/tests/ABBCCDDA.c pthread_mutex_t a, b, c, d; b 10 tools/lib/lockdep/tests/ABBCCDDA.c pthread_mutex_init(&b, NULL); b 14 tools/lib/lockdep/tests/ABBCCDDA.c LOCK_UNLOCK_2(a, b); b 15 tools/lib/lockdep/tests/ABBCCDDA.c LOCK_UNLOCK_2(b, c); b 21 tools/lib/lockdep/tests/ABBCCDDA.c pthread_mutex_destroy(&b); b 7 tools/lib/lockdep/tests/ABCABC.c pthread_mutex_t a, b, c; b 10 tools/lib/lockdep/tests/ABCABC.c pthread_mutex_init(&b, NULL); b 13 tools/lib/lockdep/tests/ABCABC.c LOCK_UNLOCK_2(a, b); b 15 tools/lib/lockdep/tests/ABCABC.c LOCK_UNLOCK_2(b, c); b 18 tools/lib/lockdep/tests/ABCABC.c pthread_mutex_destroy(&b); b 7 tools/lib/lockdep/tests/ABCDBCDA.c pthread_mutex_t a, b, c, d; b 10 tools/lib/lockdep/tests/ABCDBCDA.c pthread_mutex_init(&b, NULL); b 14 tools/lib/lockdep/tests/ABCDBCDA.c LOCK_UNLOCK_2(a, b); b 16 tools/lib/lockdep/tests/ABCDBCDA.c LOCK_UNLOCK_2(b, c); b 21 tools/lib/lockdep/tests/ABCDBCDA.c pthread_mutex_destroy(&b); b 7 tools/lib/lockdep/tests/ABCDBDDA.c pthread_mutex_t a, b, c, d; b 10 tools/lib/lockdep/tests/ABCDBDDA.c pthread_mutex_init(&b, NULL); b 14 tools/lib/lockdep/tests/ABCDBDDA.c LOCK_UNLOCK_2(a, b); b 16 tools/lib/lockdep/tests/ABCDBDDA.c LOCK_UNLOCK_2(b, d); b 21 tools/lib/lockdep/tests/ABCDBDDA.c pthread_mutex_destroy(&b); b 6 tools/lib/lockdep/tests/WW.c pthread_rwlock_t a, b; b 9 tools/lib/lockdep/tests/WW.c pthread_rwlock_init(&b, NULL); b 12 tools/lib/lockdep/tests/WW.c pthread_rwlock_rdlock(&b); b 5 tools/lib/lockdep/tests/common.h #define LOCK_UNLOCK_2(a, b) \ b 8 tools/lib/lockdep/tests/common.h pthread_mutex_lock(&(b)); \ b 9 tools/lib/lockdep/tests/common.h pthread_mutex_unlock(&(b)); \ b 42 tools/lib/subcmd/help.c struct cmdname *b = *(struct cmdname **)b_; b 43 tools/lib/subcmd/help.c return strcmp(a->name, b->name); b 28 tools/lib/subcmd/help.h int cmdname_compare(const void *a, const void *b); b 785 tools/lib/subcmd/parse-options.c const struct option *a = va, *b = vb; b 786 tools/lib/subcmd/parse-options.c int sa = tolower(a->short_name), sb = tolower(b->short_name), ret; b 797 tools/lib/subcmd/parse-options.c *lb = b->long_name ?: ""; b 124 tools/lib/subcmd/parse-options.h #define OPT_BIT(s, l, v, h, b) { .type = OPTION_BIT, .short_name = (s), .long_name = (l), .value = check_vtype(v, int *), .help = (h), .defval = (b) } b 132 tools/lib/traceevent/event-parse.c static int cmdline_cmp(const void *a, const void *b) b 135 tools/lib/traceevent/event-parse.c const struct tep_cmdline *cb = b; b 146 tools/lib/traceevent/event-parse.c static int cmdline_slot_cmp(const void *a, const void *b) b 149 tools/lib/traceevent/event-parse.c const struct tep_cmdline *cb = b; b 409 tools/lib/traceevent/event-parse.c static int func_cmp(const void *a, const void *b) b 412 tools/lib/traceevent/event-parse.c const struct func_map *fb = b; b 426 tools/lib/traceevent/event-parse.c static int func_bcmp(const void *a, const void *b) b 429 tools/lib/traceevent/event-parse.c const struct func_map *fb = b; b 684 tools/lib/traceevent/event-parse.c static int printk_cmp(const void *a, const void *b) b 687 tools/lib/traceevent/event-parse.c const struct printk_map *pb = b; b 3529 tools/lib/traceevent/event-parse.c static int events_id_cmp(const void *a, const void *b); b 5700 tools/lib/traceevent/event-parse.c static int events_id_cmp(const void *a, const void *b) b 5703 tools/lib/traceevent/event-parse.c struct tep_event * const * eb = b; b 5714 tools/lib/traceevent/event-parse.c static int events_name_cmp(const void *a, const void *b) b 5717 tools/lib/traceevent/event-parse.c struct tep_event * const * eb = b; b 5728 tools/lib/traceevent/event-parse.c return events_id_cmp(a, b); b 5731 tools/lib/traceevent/event-parse.c static int events_system_cmp(const void *a, const void *b) b 5734 tools/lib/traceevent/event-parse.c struct tep_event * const * eb = b; b 5745 tools/lib/traceevent/event-parse.c return events_id_cmp(a, b); b 5767 tools/lib/traceevent/event-parse.c int (*sort)(const void *a, const void *b); b 96 tools/lib/traceevent/parse-filter.c static int filter_cmp(const void *a, const void *b) b 99 tools/lib/traceevent/parse-filter.c const struct tep_filter_type *eb = b; b 610 tools/lib/traceevent/parse-filter.c rotate_op_right(struct tep_filter_arg *a, struct tep_filter_arg *b) b 615 tools/lib/traceevent/parse-filter.c a->op.right = b; b 14 tools/perf/arch/x86/util/header.c cpuid(unsigned int op, unsigned int *a, unsigned int *b, unsigned int *c, b 20 tools/perf/arch/x86/util/header.c "=S" (*b), b 29 tools/perf/arch/x86/util/header.c unsigned int a, b, c, d, lvl; b 34 tools/perf/arch/x86/util/header.c cpuid(0, &lvl, &b, &c, &d); b 35 tools/perf/arch/x86/util/header.c strncpy(&vendor[0], (char *)(&b), 4); b 41 tools/perf/arch/x86/util/header.c cpuid(1, &a, &b, &c, &d); b 344 tools/perf/arch/x86/util/intel-bts.c int i, a, b; b 346 tools/perf/arch/x86/util/intel-bts.c b = buf_size >> 3; b 347 tools/perf/arch/x86/util/intel-bts.c a = b - 512; b 351 tools/perf/arch/x86/util/intel-bts.c for (i = a; i < b; i++) { b 1010 tools/perf/arch/x86/util/intel-pt.c int i, a, b; b 1012 tools/perf/arch/x86/util/intel-pt.c b = buf_size >> 3; b 1013 tools/perf/arch/x86/util/intel-pt.c a = b - 512; b 1017 tools/perf/arch/x86/util/intel-pt.c for (i = a; i < b; i++) { b 1691 tools/perf/builtin-c2c.c static bool fmt_equal(struct perf_hpp_fmt *a, struct perf_hpp_fmt *b) b 1694 tools/perf/builtin-c2c.c struct c2c_fmt *c2c_b = container_of(b, struct c2c_fmt, fmt); b 1731 tools/perf/builtin-c2c.c struct hist_entry *a, struct hist_entry *b) b 1736 tools/perf/builtin-c2c.c return dim->se->se_cmp(a, b); b 1740 tools/perf/builtin-c2c.c struct hist_entry *a, struct hist_entry *b) b 1747 tools/perf/builtin-c2c.c return collapse_fn(a, b); b 2402 tools/perf/builtin-c2c.c ui_browser__help_window(&browser->b, help); b 2467 tools/perf/builtin-c2c.c ui_browser__help_window(&browser->b, help); b 654 tools/perf/builtin-diff.c static int block_pair_cmp(struct hist_entry *a, struct hist_entry *b) b 657 tools/perf/builtin-diff.c struct block_info *bi_b = b->block_info; b 308 tools/perf/builtin-kmem.c static int funcmp(const void *a, const void *b) b 311 tools/perf/builtin-kmem.c const struct alloc_func *fb = b; b 319 tools/perf/builtin-kmem.c static int callcmp(const void *a, const void *b) b 322 tools/perf/builtin-kmem.c const struct alloc_func *fb = b; b 618 tools/perf/builtin-kmem.c static int gfpcmp(const void *a, const void *b) b 621 tools/perf/builtin-kmem.c const struct gfp_flag *fb = b; b 1414 tools/perf/builtin-kmem.c static int ptr_cmp(void *a, void *b) b 1417 tools/perf/builtin-kmem.c struct alloc_stat *r = b; b 1431 tools/perf/builtin-kmem.c static int slab_callsite_cmp(void *a, void *b) b 1434 tools/perf/builtin-kmem.c struct alloc_stat *r = b; b 1448 tools/perf/builtin-kmem.c static int hit_cmp(void *a, void *b) b 1451 tools/perf/builtin-kmem.c struct alloc_stat *r = b; b 1465 tools/perf/builtin-kmem.c static int bytes_cmp(void *a, void *b) b 1468 tools/perf/builtin-kmem.c struct alloc_stat *r = b; b 1482 tools/perf/builtin-kmem.c static int frag_cmp(void *a, void *b) b 1486 tools/perf/builtin-kmem.c struct alloc_stat *r = b; b 1503 tools/perf/builtin-kmem.c static int pingpong_cmp(void *a, void *b) b 1506 tools/perf/builtin-kmem.c struct alloc_stat *r = b; b 1521 tools/perf/builtin-kmem.c static int page_cmp(void *a, void *b) b 1524 tools/perf/builtin-kmem.c struct page_stat *r = b; b 1538 tools/perf/builtin-kmem.c static int page_callsite_cmp(void *a, void *b) b 1541 tools/perf/builtin-kmem.c struct page_stat *r = b; b 1555 tools/perf/builtin-kmem.c static int page_hit_cmp(void *a, void *b) b 1558 tools/perf/builtin-kmem.c struct page_stat *r = b; b 1572 tools/perf/builtin-kmem.c static int page_bytes_cmp(void *a, void *b) b 1575 tools/perf/builtin-kmem.c struct page_stat *r = b; b 1589 tools/perf/builtin-kmem.c static int page_order_cmp(void *a, void *b) b 1592 tools/perf/builtin-kmem.c struct page_stat *r = b; b 1606 tools/perf/builtin-kmem.c static int migrate_type_cmp(void *a, void *b) b 1609 tools/perf/builtin-kmem.c struct page_stat *r = b; b 1627 tools/perf/builtin-kmem.c static int gfp_flags_cmp(void *a, void *b) b 1630 tools/perf/builtin-kmem.c struct page_stat *r = b; b 192 tools/perf/builtin-stat.c struct timespec *b) b 194 tools/perf/builtin-stat.c r->tv_sec = a->tv_sec - b->tv_sec; b 195 tools/perf/builtin-stat.c if (a->tv_nsec < b->tv_nsec) { b 196 tools/perf/builtin-stat.c r->tv_nsec = a->tv_nsec + NSEC_PER_SEC - b->tv_nsec; b 199 tools/perf/builtin-stat.c r->tv_nsec = a->tv_nsec - b->tv_nsec ; b 1555 tools/perf/builtin-trace.c static int intcmp(const void *a, const void *b) b 1557 tools/perf/builtin-trace.c const int *one = a, *another = b; b 3667 tools/perf/builtin-trace.c DEFINE_RESORT_RB(syscall_stats, a->msecs > b->msecs, b 3757 tools/perf/builtin-trace.c DEFINE_RESORT_RB(threads, (thread__nr_events(a->thread->priv) < thread__nr_events(b->thread->priv)), b 3 tools/perf/lib/core.c #define __printf(a, b) __attribute__((format(printf, a, b))) b 442 tools/perf/tests/code-reading.c static int comp(const void *a, const void *b) b 444 tools/perf/tests/code-reading.c return *(int *)a - *(int *)b; b 251 tools/perf/tests/switch-tracking.c static int compar(const void *a, const void *b) b 254 tools/perf/tests/switch-tracking.c const struct event_node *nodeb = b; b 159 tools/perf/tests/time-utils-test.c u64 b = 1234567123456789ULL; b 162 tools/perf/tests/time-utils-test.c .ptime = { {b, b + 1}, }, b 164 tools/perf/tests/time-utils-test.c .skip = { b - 1, b + 2, }, b 165 tools/perf/tests/time-utils-test.c .noskip = { b, b + 1, }, b 172 tools/perf/tests/time-utils-test.c u64 b = 1234567123456789ULL; b 179 tools/perf/tests/time-utils-test.c .ptime = { {b, b + 1}, {c, c + 123}, {e, e + 5}, }, b 181 tools/perf/tests/time-utils-test.c .skip = { b - 1, b + 2, c - 1, c + 124, e - 1, e + 6 }, b 182 tools/perf/tests/time-utils-test.c .noskip = { b, b + 1, c, c + 123, e, e + 5 }, b 189 tools/perf/tests/time-utils-test.c u64 b = 7654321ULL * NSEC_PER_SEC; b 192 tools/perf/tests/time-utils-test.c .first = b, b 193 tools/perf/tests/time-utils-test.c .last = b + 100, b 194 tools/perf/tests/time-utils-test.c .ptime = { {b, b + 9}, }, b 196 tools/perf/tests/time-utils-test.c .skip = { b - 1, b + 10, }, b 197 tools/perf/tests/time-utils-test.c .noskip = { b, b + 9, }, b 204 tools/perf/tests/time-utils-test.c u64 b = 7654321ULL * NSEC_PER_SEC; b 207 tools/perf/tests/time-utils-test.c .first = b, b 208 tools/perf/tests/time-utils-test.c .last = b + 100, b 209 tools/perf/tests/time-utils-test.c .ptime = { {b + 10, b + 19}, }, b 211 tools/perf/tests/time-utils-test.c .skip = { b + 9, b + 20, }, b 212 tools/perf/tests/time-utils-test.c .noskip = { b + 10, b + 19, }, b 219 tools/perf/tests/time-utils-test.c u64 b = 11223344ULL * NSEC_PER_SEC; b 222 tools/perf/tests/time-utils-test.c .first = b, b 223 tools/perf/tests/time-utils-test.c .last = b + 100, b 224 tools/perf/tests/time-utils-test.c .ptime = { {b, b + 9}, {b + 10, b + 19}, }, b 226 tools/perf/tests/time-utils-test.c .skip = { b - 1, b + 20, }, b 227 tools/perf/tests/time-utils-test.c .noskip = { b, b + 8, b + 9, b + 10, b + 11, b + 12, b + 19, }, b 234 tools/perf/tests/time-utils-test.c u64 b = 11223344ULL * NSEC_PER_SEC; b 237 tools/perf/tests/time-utils-test.c .first = b, b 238 tools/perf/tests/time-utils-test.c .last = b + 100, b 239 tools/perf/tests/time-utils-test.c .ptime = { {b, b + 9}, {b + 20, b + 29}, { b + 90, b + 100}, }, b 241 tools/perf/tests/time-utils-test.c .skip = { b - 1, b + 10, b + 19, b + 30, b + 89, b + 101 }, b 242 tools/perf/tests/time-utils-test.c .noskip = { b, b + 9, b + 20, b + 29, b + 90, b + 100}, b 30 tools/perf/ui/browsers/annotate.c struct ui_browser b; b 98 tools/perf/ui/browsers/annotate.c struct annotate_browser *ab = container_of(browser, struct annotate_browser, b); b 149 tools/perf/ui/browsers/annotate.c struct annotate_browser *ab = container_of(browser, struct annotate_browser, b); b 153 tools/perf/ui/browsers/annotate.c struct map_symbol *ms = ab->b.priv; b 230 tools/perf/ui/browsers/annotate.c static double disasm__cmp(struct annotation_line *a, struct annotation_line *b, b 236 tools/perf/ui/browsers/annotate.c if (a->data[i].percent[percent_type] == b->data[i].percent[percent_type]) b 239 tools/perf/ui/browsers/annotate.c b->data[i].percent[percent_type]; b 268 tools/perf/ui/browsers/annotate.c struct annotation *notes = browser__annotation(&browser->b); b 271 tools/perf/ui/browsers/annotate.c ui_browser__refresh_dimensions(&browser->b); b 272 tools/perf/ui/browsers/annotate.c back = browser->b.height / 2; b 273 tools/perf/ui/browsers/annotate.c browser->b.top_idx = browser->b.index = idx; b 275 tools/perf/ui/browsers/annotate.c while (browser->b.top_idx != 0 && back != 0) { b 281 tools/perf/ui/browsers/annotate.c --browser->b.top_idx; b 285 tools/perf/ui/browsers/annotate.c browser->b.top = pos; b 286 tools/perf/ui/browsers/annotate.c browser->b.navkeypressed = true; b 292 tools/perf/ui/browsers/annotate.c struct annotation *notes = browser__annotation(&browser->b); b 305 tools/perf/ui/browsers/annotate.c struct map_symbol *ms = browser->b.priv; b 348 tools/perf/ui/browsers/annotate.c struct annotation *notes = browser__annotation(&browser->b); b 350 tools/perf/ui/browsers/annotate.c off_t offset = browser->b.index - browser->b.top_idx; b 352 tools/perf/ui/browsers/annotate.c browser->b.seek(&browser->b, offset, SEEK_CUR); b 353 tools/perf/ui/browsers/annotate.c al = list_entry(browser->b.top, struct annotation_line, node); b 359 tools/perf/ui/browsers/annotate.c browser->b.nr_entries = notes->nr_entries; b 361 tools/perf/ui/browsers/annotate.c browser->b.seek(&browser->b, -offset, SEEK_CUR); b 362 tools/perf/ui/browsers/annotate.c browser->b.top_idx = al->idx - offset; b 363 tools/perf/ui/browsers/annotate.c browser->b.index = al->idx; b 367 tools/perf/ui/browsers/annotate.c browser->b.seek(&browser->b, -offset, SEEK_CUR); b 374 tools/perf/ui/browsers/annotate.c browser->b.nr_entries = notes->nr_asm_entries; b 376 tools/perf/ui/browsers/annotate.c browser->b.seek(&browser->b, -offset, SEEK_CUR); b 377 tools/perf/ui/browsers/annotate.c browser->b.top_idx = al->idx_asm - offset; b 378 tools/perf/ui/browsers/annotate.c browser->b.index = al->idx_asm; b 413 tools/perf/ui/browsers/annotate.c struct map_symbol *ms = browser->b.priv; b 436 tools/perf/ui/browsers/annotate.c ui_browser__show_title(&browser->b, title); b 444 tools/perf/ui/browsers/annotate.c struct annotation *notes = browser__annotation(&browser->b); b 490 tools/perf/ui/browsers/annotate.c struct annotation *notes = browser__annotation(&browser->b); b 493 tools/perf/ui/browsers/annotate.c *idx = browser->b.index; b 527 tools/perf/ui/browsers/annotate.c struct annotation *notes = browser__annotation(&browser->b); b 530 tools/perf/ui/browsers/annotate.c *idx = browser->b.index; b 610 tools/perf/ui/browsers/annotate.c struct annotate_browser *ab = container_of(browser, struct annotate_browser, b); b 665 tools/perf/ui/browsers/annotate.c struct map_symbol *ms = browser->b.priv; b 674 tools/perf/ui/browsers/annotate.c if (annotate_browser__show(&browser->b, title, help) < 0) b 681 tools/perf/ui/browsers/annotate.c browser->b.navkeypressed = false; b 687 tools/perf/ui/browsers/annotate.c key = ui_browser__run(&browser->b, delay_secs); b 708 tools/perf/ui/browsers/annotate.c annotate_browser__show(&browser->b, title, help); b 729 tools/perf/ui/browsers/annotate.c ui_browser__help_window(&browser->b, b 803 tools/perf/ui/browsers/annotate.c seq++, browser->b.nr_entries, b 804 tools/perf/ui/browsers/annotate.c browser->b.height, b 805 tools/perf/ui/browsers/annotate.c browser->b.index, b 806 tools/perf/ui/browsers/annotate.c browser->b.top_idx, b 854 tools/perf/ui/browsers/annotate.c annotate_browser__show(&browser->b, title, help); b 869 tools/perf/ui/browsers/annotate.c ui_browser__hide(&browser->b); b 902 tools/perf/ui/browsers/annotate.c .b = { b 931 tools/perf/ui/browsers/annotate.c browser.b.width = notes->max_line_len; b 932 tools/perf/ui/browsers/annotate.c browser.b.nr_entries = notes->nr_entries; b 933 tools/perf/ui/browsers/annotate.c browser.b.entries = ¬es->src->source, b 934 tools/perf/ui/browsers/annotate.c browser.b.width += 18; /* Percentage */ b 937 tools/perf/ui/browsers/annotate.c ui_browser__init_asm_mode(&browser.b); b 78 tools/perf/ui/browsers/hists.c struct ui_browser *browser = &hb->b; b 102 tools/perf/ui/browsers/hists.c struct ui_browser *browser = &hb->b; b 126 tools/perf/ui/browsers/hists.c struct hist_browser *hb = container_of(browser, struct hist_browser, b); b 148 tools/perf/ui/browsers/hists.c browser->b.nr_entries = hist_browser__nr_entries(browser); b 149 tools/perf/ui/browsers/hists.c hist_browser__refresh_dimensions(&browser->b); b 150 tools/perf/ui/browsers/hists.c ui_browser__reset_index(&browser->b); b 412 tools/perf/ui/browsers/hists.c browser->b.nr_entries -= he->nr_rows; b 431 tools/perf/ui/browsers/hists.c browser->b.nr_entries += child_rows - he->nr_rows; b 439 tools/perf/ui/browsers/hists.c browser->b.nr_entries -= child_rows - he->nr_rows; b 447 tools/perf/ui/browsers/hists.c browser->b.nr_entries += he->nr_rows; b 598 tools/perf/ui/browsers/hists.c browser->b.nr_entries = hist_browser__nr_entries(browser); b 600 tools/perf/ui/browsers/hists.c ui_browser__reset_index(&browser->b); b 609 tools/perf/ui/browsers/hists.c browser->b.nr_entries = hist_browser__nr_entries(browser); b 634 tools/perf/ui/browsers/hists.c browser->b.entries = &browser->hists->entries; b 635 tools/perf/ui/browsers/hists.c browser->b.nr_entries = hist_browser__nr_entries(browser); b 639 tools/perf/ui/browsers/hists.c if (ui_browser__show(&browser->b, title, "%s", help) < 0) b 643 tools/perf/ui/browsers/hists.c key = ui_browser__run(&browser->b, delay_secs); b 659 tools/perf/ui/browsers/hists.c ui_browser__update_nr_entries(&browser->b, nr_entries); b 666 tools/perf/ui/browsers/hists.c ui_browser__warn_lost_events(&browser->b); b 670 tools/perf/ui/browsers/hists.c ui_browser__show_title(&browser->b, title); b 675 tools/perf/ui/browsers/hists.c struct hist_entry *h = rb_entry(browser->b.top, b 679 tools/perf/ui/browsers/hists.c seq++, browser->b.nr_entries, b 681 tools/perf/ui/browsers/hists.c browser->b.extra_title_lines, b 682 tools/perf/ui/browsers/hists.c browser->b.rows, b 683 tools/perf/ui/browsers/hists.c browser->b.index, b 684 tools/perf/ui/browsers/hists.c browser->b.top_idx, b 717 tools/perf/ui/browsers/hists.c ui_browser__hide(&browser->b); b 748 tools/perf/ui/browsers/hists.c width = browser->b.width - (offset + 2); b 749 tools/perf/ui/browsers/hists.c if (ui_browser__is_current_entry(&browser->b, row)) { b 755 tools/perf/ui/browsers/hists.c ui_browser__set_color(&browser->b, color); b 756 tools/perf/ui/browsers/hists.c ui_browser__gotorc(&browser->b, row, 0); b 757 tools/perf/ui/browsers/hists.c ui_browser__write_nstring(&browser->b, " ", offset); b 758 tools/perf/ui/browsers/hists.c ui_browser__printf(&browser->b, "%c", folded_sign); b 759 tools/perf/ui/browsers/hists.c ui_browser__write_graph(&browser->b, show_annotated ? SLSMG_RARROW_CHAR : ' '); b 760 tools/perf/ui/browsers/hists.c ui_browser__write_nstring(&browser->b, str, width); b 763 tools/perf/ui/browsers/hists.c static void hist_browser__fprintf_callchain_entry(struct hist_browser *b __maybe_unused, b 781 tools/perf/ui/browsers/hists.c return browser->b.rows == row; b 1152 tools/perf/ui/browsers/hists.c struct ui_browser *b; b 1169 tools/perf/ui/browsers/hists.c ui_browser__set_percent_color(arg->b, percent, arg->current_entry); b 1172 tools/perf/ui/browsers/hists.c ui_browser__printf(arg->b, "%s", hpp->buf); b 1208 tools/perf/ui/browsers/hists.c ui_browser__printf(arg->b, "%s", hpp->buf); \ b 1249 tools/perf/ui/browsers/hists.c int width = browser->b.width; b 1251 tools/perf/ui/browsers/hists.c bool current_entry = ui_browser__is_current_entry(&browser->b, row); b 1269 tools/perf/ui/browsers/hists.c .b = &browser->b, b 1275 tools/perf/ui/browsers/hists.c ui_browser__gotorc(&browser->b, row, 0); b 1286 tools/perf/ui/browsers/hists.c column++ < browser->b.horiz_scroll) b 1289 tools/perf/ui/browsers/hists.c if (current_entry && browser->b.navkeypressed) { b 1290 tools/perf/ui/browsers/hists.c ui_browser__set_color(&browser->b, b 1293 tools/perf/ui/browsers/hists.c ui_browser__set_color(&browser->b, b 1299 tools/perf/ui/browsers/hists.c ui_browser__printf(&browser->b, "%c ", folded_sign); b 1304 tools/perf/ui/browsers/hists.c ui_browser__printf(&browser->b, " "); b 1315 tools/perf/ui/browsers/hists.c ui_browser__printf(&browser->b, "%s", s + ret); b 1318 tools/perf/ui/browsers/hists.c ui_browser__printf(&browser->b, "%s", s); b 1324 tools/perf/ui/browsers/hists.c if (!browser->b.navkeypressed) b 1327 tools/perf/ui/browsers/hists.c ui_browser__write_nstring(&browser->b, "", width); b 1334 tools/perf/ui/browsers/hists.c if (folded_sign == '-' && row != browser->b.rows) { b 1356 tools/perf/ui/browsers/hists.c int width = browser->b.width; b 1358 tools/perf/ui/browsers/hists.c bool current_entry = ui_browser__is_current_entry(&browser->b, row); b 1364 tools/perf/ui/browsers/hists.c .b = &browser->b, b 1384 tools/perf/ui/browsers/hists.c ui_browser__gotorc(&browser->b, row, 0); b 1386 tools/perf/ui/browsers/hists.c if (current_entry && browser->b.navkeypressed) b 1387 tools/perf/ui/browsers/hists.c ui_browser__set_color(&browser->b, HE_COLORSET_SELECTED); b 1389 tools/perf/ui/browsers/hists.c ui_browser__set_color(&browser->b, HE_COLORSET_NORMAL); b 1391 tools/perf/ui/browsers/hists.c ui_browser__write_nstring(&browser->b, "", level * HIERARCHY_INDENT); b 1406 tools/perf/ui/browsers/hists.c column++ < browser->b.horiz_scroll) b 1409 tools/perf/ui/browsers/hists.c if (current_entry && browser->b.navkeypressed) { b 1410 tools/perf/ui/browsers/hists.c ui_browser__set_color(&browser->b, b 1413 tools/perf/ui/browsers/hists.c ui_browser__set_color(&browser->b, b 1418 tools/perf/ui/browsers/hists.c ui_browser__printf(&browser->b, "%c ", folded_sign); b 1422 tools/perf/ui/browsers/hists.c ui_browser__printf(&browser->b, " "); b 1433 tools/perf/ui/browsers/hists.c ui_browser__printf(&browser->b, "%s", s + ret); b 1437 tools/perf/ui/browsers/hists.c ui_browser__printf(&browser->b, "%s", s); b 1443 tools/perf/ui/browsers/hists.c ui_browser__write_nstring(&browser->b, "", hierarchy_indent); b 1447 tools/perf/ui/browsers/hists.c if (column >= browser->b.horiz_scroll) { b 1455 tools/perf/ui/browsers/hists.c if (current_entry && browser->b.navkeypressed) { b 1456 tools/perf/ui/browsers/hists.c ui_browser__set_color(&browser->b, b 1459 tools/perf/ui/browsers/hists.c ui_browser__set_color(&browser->b, b 1465 tools/perf/ui/browsers/hists.c ui_browser__printf(&browser->b, "%c ", folded_sign); b 1468 tools/perf/ui/browsers/hists.c ui_browser__write_nstring(&browser->b, "", 2); b 1484 tools/perf/ui/browsers/hists.c ui_browser__printf(&browser->b, "%s", skip_spaces(s)); b 1493 tools/perf/ui/browsers/hists.c if (!browser->b.navkeypressed) b 1496 tools/perf/ui/browsers/hists.c ui_browser__write_nstring(&browser->b, "", width); b 1502 tools/perf/ui/browsers/hists.c if (entry->leaf && folded_sign == '-' && row != browser->b.rows) { b 1519 tools/perf/ui/browsers/hists.c int width = browser->b.width; b 1520 tools/perf/ui/browsers/hists.c bool current_entry = ui_browser__is_current_entry(&browser->b, row); b 1533 tools/perf/ui/browsers/hists.c ui_browser__gotorc(&browser->b, row, 0); b 1535 tools/perf/ui/browsers/hists.c if (current_entry && browser->b.navkeypressed) b 1536 tools/perf/ui/browsers/hists.c ui_browser__set_color(&browser->b, HE_COLORSET_SELECTED); b 1538 tools/perf/ui/browsers/hists.c ui_browser__set_color(&browser->b, HE_COLORSET_NORMAL); b 1540 tools/perf/ui/browsers/hists.c ui_browser__write_nstring(&browser->b, "", level * HIERARCHY_INDENT); b 1548 tools/perf/ui/browsers/hists.c column++ < browser->b.horiz_scroll) b 1562 tools/perf/ui/browsers/hists.c ui_browser__write_nstring(&browser->b, "", ret); b 1566 tools/perf/ui/browsers/hists.c ui_browser__write_nstring(&browser->b, "", indent * HIERARCHY_INDENT); b 1569 tools/perf/ui/browsers/hists.c if (column >= browser->b.horiz_scroll) { b 1573 tools/perf/ui/browsers/hists.c ui_browser__printf(&browser->b, " %s", buf); b 1578 tools/perf/ui/browsers/hists.c if (!browser->b.navkeypressed) b 1581 tools/perf/ui/browsers/hists.c ui_browser__write_nstring(&browser->b, "", width); b 1612 tools/perf/ui/browsers/hists.c if (perf_hpp__should_skip(fmt, hists) || column++ < browser->b.horiz_scroll) b 1653 tools/perf/ui/browsers/hists.c if (column++ < browser->b.horiz_scroll) b 1721 tools/perf/ui/browsers/hists.c ui_browser__gotorc(&browser->b, 0, 0); b 1722 tools/perf/ui/browsers/hists.c ui_browser__set_color(&browser->b, HE_COLORSET_ROOT); b 1723 tools/perf/ui/browsers/hists.c ui_browser__write_nstring(&browser->b, headers, browser->b.width + 1); b 1739 tools/perf/ui/browsers/hists.c ui_browser__gotorc_title(&browser->b, line, 0); b 1740 tools/perf/ui/browsers/hists.c ui_browser__set_color(&browser->b, HE_COLORSET_ROOT); b 1741 tools/perf/ui/browsers/hists.c ui_browser__write_nstring(&browser->b, headers, browser->b.width + 1); b 1758 tools/perf/ui/browsers/hists.c hb = container_of(browser, struct hist_browser, b); b 1767 tools/perf/ui/browsers/hists.c struct hist_browser *hb = container_of(browser, struct hist_browser, b); b 1858 tools/perf/ui/browsers/hists.c hb = container_of(browser, struct hist_browser, b); b 2097 tools/perf/ui/browsers/hists.c struct rb_node *nd = hists__filter_entries(rb_first(browser->b.entries), b 2159 tools/perf/ui/browsers/hists.c browser->b.refresh = hist_browser__refresh; b 2160 tools/perf/ui/browsers/hists.c browser->b.refresh_dimensions = hist_browser__refresh_dimensions; b 2161 tools/perf/ui/browsers/hists.c browser->b.seek = ui_browser__hists_seek; b 2162 tools/perf/ui/browsers/hists.c browser->b.use_navkeypressed = true; b 2173 tools/perf/ui/browsers/hists.c ++browser->b.columns; b 2176 tools/perf/ui/browsers/hists.c ++browser->b.columns; b 2179 tools/perf/ui/browsers/hists.c ++browser->b.columns; b 2394 tools/perf/ui/browsers/hists.c ui_browser__update_nr_entries(&browser->b, browser->hists->nr_entries); b 2396 tools/perf/ui/browsers/hists.c ui_browser__handle_resize(&browser->b); b 2904 tools/perf/ui/browsers/hists.c browser->b.no_samples_msg = "Collecting samples..."; b 2934 tools/perf/ui/browsers/hists.c ui_browser__warning(&browser->b, delay_secs * 2, b 3019 tools/perf/ui/browsers/hists.c ui_browser__warning(&browser->b, delay_secs * 2, b 3031 tools/perf/ui/browsers/hists.c ui_browser__help_window(&browser->b, b 3051 tools/perf/ui/browsers/hists.c ui_browser__dialog_yesno(&browser->b, b 3212 tools/perf/ui/browsers/hists.c struct ui_browser b; b 3224 tools/perf/ui/browsers/hists.c struct evsel_menu, b); b 3275 tools/perf/ui/browsers/hists.c struct evlist *evlist = menu->b.priv; b 3281 tools/perf/ui/browsers/hists.c if (ui_browser__show(&menu->b, title, b 3286 tools/perf/ui/browsers/hists.c key = ui_browser__run(&menu->b, delay_secs); b 3296 tools/perf/ui/browsers/hists.c ui_browser__warn_lost_events(&menu->b); b 3319 tools/perf/ui/browsers/hists.c ui_browser__show_title(&menu->b, title); b 3344 tools/perf/ui/browsers/hists.c if (!ui_browser__dialog_yesno(&menu->b, b 3357 tools/perf/ui/browsers/hists.c ui_browser__hide(&menu->b); b 3382 tools/perf/ui/browsers/hists.c .b = { b 3402 tools/perf/ui/browsers/hists.c if (menu.b.width < line_len) b 3403 tools/perf/ui/browsers/hists.c menu.b.width = line_len; b 10 tools/perf/ui/browsers/hists.h struct ui_browser b; b 20 tools/perf/ui/browsers/map.c struct ui_browser b; b 28 tools/perf/ui/browsers/map.c struct map_browser *mb = container_of(browser, struct map_browser, b); b 67 tools/perf/ui/browsers/map.c browser->b.top = &sym->rb_node; b 68 tools/perf/ui/browsers/map.c browser->b.index = browser->b.top_idx = *idx; b 79 tools/perf/ui/browsers/map.c if (ui_browser__show(&browser->b, browser->map->dso->long_name, b 85 tools/perf/ui/browsers/map.c key = ui_browser__run(&browser->b, 0); b 101 tools/perf/ui/browsers/map.c ui_browser__hide(&browser->b); b 108 tools/perf/ui/browsers/map.c .b = { b 120 tools/perf/ui/browsers/map.c for (nd = rb_first(mb.b.entries); nd; nd = rb_next(nd)) { b 127 tools/perf/ui/browsers/map.c *idx = mb.b.nr_entries; b 129 tools/perf/ui/browsers/map.c ++mb.b.nr_entries; b 154 tools/perf/ui/hist.c static int __hpp__sort(struct hist_entry *a, struct hist_entry *b, b 163 tools/perf/ui/hist.c ret = field_cmp(get_field(a), get_field(b)); b 183 tools/perf/ui/hist.c list_for_each_entry(pair, &b->pairs.head, pairs.node) { b 201 tools/perf/ui/hist.c static int __hpp__sort_acc(struct hist_entry *a, struct hist_entry *b, b 210 tools/perf/ui/hist.c ret = field_cmp(get_field(a), get_field(b)); b 214 tools/perf/ui/hist.c if (a->thread != b->thread || !hist_entry__has_callchains(a) || !symbol_conf.use_callchain) b 217 tools/perf/ui/hist.c ret = b->callchain->max_depth - a->callchain->max_depth; b 300 tools/perf/ui/hist.c struct hist_entry *a, struct hist_entry *b) \ b 302 tools/perf/ui/hist.c return __hpp__sort(a, b, he_get_##_field); \ b 328 tools/perf/ui/hist.c struct hist_entry *a, struct hist_entry *b) \ b 330 tools/perf/ui/hist.c return __hpp__sort_acc(a, b, he_get_acc_##_field); \ b 348 tools/perf/ui/hist.c struct hist_entry *a, struct hist_entry *b) \ b 350 tools/perf/ui/hist.c return __hpp__sort(a, b, he_get_raw_##_field); \ b 380 tools/perf/ui/hist.c struct hist_entry *b __maybe_unused) b 390 tools/perf/ui/hist.c static bool hpp__equal(struct perf_hpp_fmt *a, struct perf_hpp_fmt *b) b 392 tools/perf/ui/hist.c if (!perf_hpp__is_hpp_entry(a) || !perf_hpp__is_hpp_entry(b)) b 395 tools/perf/ui/hist.c return a->idx == b->idx; b 564 tools/perf/ui/hist.c static bool fmt_equal(struct perf_hpp_fmt *a, struct perf_hpp_fmt *b) b 566 tools/perf/ui/hist.c return a->equal && a->equal(a, b); b 666 tools/perf/util/annotate.c static int ins__cmp(const void *a, const void *b) b 669 tools/perf/util/annotate.c const struct ins *ib = b; b 712 tools/perf/util/annotate.c static int arch__cmp(const void *a, const void *b) b 715 tools/perf/util/annotate.c const struct arch *ab = b; b 2151 tools/perf/util/annotate.c static int cmp_source_line(struct annotation_line *a, struct annotation_line *b) b 2156 tools/perf/util/annotate.c if (a->data[i].percent_sum == b->data[i].percent_sum) b 2158 tools/perf/util/annotate.c return a->data[i].percent_sum > b->data[i].percent_sum; b 280 tools/perf/util/auxtrace.c struct auxtrace_buffer *b; b 284 tools/perf/util/auxtrace.c b = memdup(buffer, sizeof(struct auxtrace_buffer)); b 285 tools/perf/util/auxtrace.c if (!b) b 287 tools/perf/util/auxtrace.c b->size = BUFFER_LIMIT_FOR_32_BIT; b 288 tools/perf/util/auxtrace.c b->consecutive = consecutive; b 289 tools/perf/util/auxtrace.c err = auxtrace_queues__queue_buffer(queues, idx, b); b 291 tools/perf/util/auxtrace.c auxtrace_buffer__free(b); b 300 tools/perf/util/build-id.c struct perf_record_header_build_id b; b 306 tools/perf/util/build-id.c memset(&b, 0, sizeof(b)); b 307 tools/perf/util/build-id.c memcpy(&b.build_id, build_id, BUILD_ID_SIZE); b 308 tools/perf/util/build-id.c b.pid = pid; b 309 tools/perf/util/build-id.c b.header.misc = misc; b 310 tools/perf/util/build-id.c b.header.size = sizeof(b) + len; b 312 tools/perf/util/build-id.c err = do_write(fd, &b, sizeof(b)); b 126 tools/perf/util/cpumap.c static int cmp_ids(const void *a, const void *b) b 128 tools/perf/util/cpumap.c return *(int *)a - *(int *)b; b 263 tools/perf/util/evsel.h #define perf_evsel__cmp(a, b) \ b 265 tools/perf/util/evsel.h (b) && \ b 266 tools/perf/util/evsel.h (a)->core.attr.type == (b)->core.attr.type && \ b 267 tools/perf/util/evsel.h (a)->core.attr.config == (b)->core.attr.config) b 997 tools/perf/util/header.c static int cpu_cache_level__sort(const void *a, const void *b) b 1000 tools/perf/util/header.c struct cpu_cache_level *cache_b = (struct cpu_cache_level *)b; b 1005 tools/perf/util/header.c static bool cpu_cache_level__cmp(struct cpu_cache_level *a, struct cpu_cache_level *b) b 1007 tools/perf/util/header.c if (a->level != b->level) b 1010 tools/perf/util/header.c if (a->line_size != b->line_size) b 1013 tools/perf/util/header.c if (a->sets != b->sets) b 1016 tools/perf/util/header.c if (a->ways != b->ways) b 1019 tools/perf/util/header.c if (strcmp(a->type, b->type)) b 1022 tools/perf/util/header.c if (strcmp(a->size, b->size)) b 1025 tools/perf/util/header.c if (strcmp(a->map, b->map)) b 1251 tools/perf/util/header.c static int memory_node__sort(const void *a, const void *b) b 1254 tools/perf/util/header.c const struct memory_node *nb = b; b 1628 tools/perf/util/hist.c static int64_t hist_entry__sort(struct hist_entry *a, struct hist_entry *b) b 1638 tools/perf/util/hist.c cmp = fmt->sort(fmt, a, b); b 267 tools/perf/util/hist.h struct hist_entry *a, struct hist_entry *b); b 269 tools/perf/util/hist.h struct hist_entry *a, struct hist_entry *b); b 271 tools/perf/util/hist.h struct hist_entry *a, struct hist_entry *b); b 272 tools/perf/util/hist.h bool (*equal)(struct perf_hpp_fmt *a, struct perf_hpp_fmt *b); b 249 tools/perf/util/intel-bts.c struct auxtrace_buffer *b) b 254 tools/perf/util/intel-bts.c if (b->list.prev == &queue->head) b 256 tools/perf/util/intel-bts.c a = list_entry(b->list.prev, struct auxtrace_buffer, list); b 257 tools/perf/util/intel-bts.c start = intel_bts_find_overlap(a->data, a->size, b->data, b->size); b 260 tools/perf/util/intel-bts.c b->use_size = b->data + b->size - start; b 261 tools/perf/util/intel-bts.c b->use_data = start; b 237 tools/perf/util/intel-pt.c struct auxtrace_buffer *b) b 242 tools/perf/util/intel-pt.c start = intel_pt_find_overlap(a->data, a->size, b->data, b->size, b 246 tools/perf/util/intel-pt.c b->use_size = b->data + b->size - start; b 247 tools/perf/util/intel-pt.c b->use_data = start; b 248 tools/perf/util/intel-pt.c if (b->use_size && consecutive) b 249 tools/perf/util/intel-pt.c b->consecutive = true; b 256 tools/perf/util/intel-pt.c struct intel_pt_buffer *b) b 274 tools/perf/util/intel-pt.c b->len = buffer->use_size; b 275 tools/perf/util/intel-pt.c b->buf = buffer->use_data; b 277 tools/perf/util/intel-pt.c b->len = buffer->size; b 278 tools/perf/util/intel-pt.c b->buf = buffer->data; b 280 tools/perf/util/intel-pt.c b->ref_timestamp = buffer->reference; b 283 tools/perf/util/intel-pt.c b->consecutive = false; b 284 tools/perf/util/intel-pt.c b->trace_nr = buffer->buffer_nr + 1; b 286 tools/perf/util/intel-pt.c b->consecutive = true; b 315 tools/perf/util/intel-pt.c struct intel_pt_buffer b = { .len = 0 }; b 321 tools/perf/util/intel-pt.c err = intel_pt_get_buffer(ptq, buffer, old_buffer, &b); b 325 tools/perf/util/intel-pt.c if (b.len) { b 333 tools/perf/util/intel-pt.c err = cb(&b, cb_data); b 349 tools/perf/util/intel-pt.c static int intel_pt_get_trace(struct intel_pt_buffer *b, void *data) b 358 tools/perf/util/intel-pt.c b->len = 0; b 368 tools/perf/util/intel-pt.c b->len = 0; b 374 tools/perf/util/intel-pt.c err = intel_pt_get_buffer(ptq, buffer, old_buffer, b); b 381 tools/perf/util/intel-pt.c if (b->len) { b 387 tools/perf/util/intel-pt.c return intel_pt_get_trace(b, data); b 70 tools/perf/util/jitdump.c #define hmax(a, b) ((a) > (b) ? (a) : (b)) b 333 tools/perf/util/mem-events.c #define P(a, b) PERF_MEM_##a##_##b b 55 tools/perf/util/metricgroup.c const struct metric_event *b = entry; b 57 tools/perf/util/metricgroup.c if (a->evsel == b->evsel) b 59 tools/perf/util/metricgroup.c if ((char *)a->evsel < (char *)b->evsel) b 230 tools/perf/util/metricgroup.c struct mep *b = (struct mep *)entry; b 232 tools/perf/util/metricgroup.c return strcmp(a->name, b->name); b 2181 tools/perf/util/parse-events.c static int cmp_string(const void *a, const void *b) b 2184 tools/perf/util/parse-events.c const char * const *bs = b; b 14 tools/perf/util/perf_regs.h #define SMPL_REG(n, b) { .name = #n, .mask = 1ULL << (b) } b 15 tools/perf/util/perf_regs.h #define SMPL_REG2(n, b) { .name = #n, .mask = 3ULL << (b) } b 1287 tools/perf/util/pmu.c long b; b 1293 tools/perf/util/pmu.c for (b = from; b <= to; b++) b 1294 tools/perf/util/pmu.c set_bit(b, bits); b 1297 tools/perf/util/pmu.c static int sub_non_neg(int a, int b) b 1299 tools/perf/util/pmu.c if (b > a) b 1301 tools/perf/util/pmu.c return a - b; b 1347 tools/perf/util/pmu.c static int cmp_sevent(const void *a, const void *b) b 1350 tools/perf/util/pmu.c const struct sevent *bs = b; b 3121 tools/perf/util/probe-event.c static void *memcat(void *a, size_t sz_a, void *b, size_t sz_b) b 3128 tools/perf/util/probe-event.c memcpy(ret + sz_a, b, sz_b); b 584 tools/perf/util/probe-file.c static bool streql(const char *a, const char *b) b 586 tools/perf/util/probe-file.c if (a == b) b 589 tools/perf/util/probe-file.c if (!a || !b) b 592 tools/perf/util/probe-file.c return !strcmp(a, b); b 65 tools/perf/util/rb_resort.h struct __name##_sorted_entry *a, *b; \ b 67 tools/perf/util/rb_resort.h b = rb_entry(ndb, struct __name##_sorted_entry, rb_node); \ b 688 tools/perf/util/session.c static u8 revbyte(u8 b) b 690 tools/perf/util/session.c int rev = (b >> 4) | ((b & 0xf) << 4); b 1800 tools/perf/util/sort.c struct hist_entry *a, struct hist_entry *b) b 1805 tools/perf/util/sort.c return hse->se->se_cmp(a, b); b 1809 tools/perf/util/sort.c struct hist_entry *a, struct hist_entry *b) b 1816 tools/perf/util/sort.c return collapse_fn(a, b); b 1820 tools/perf/util/sort.c struct hist_entry *a, struct hist_entry *b) b 1827 tools/perf/util/sort.c return sort_fn(a, b); b 1856 tools/perf/util/sort.c static bool __sort__hpp_equal(struct perf_hpp_fmt *a, struct perf_hpp_fmt *b) b 1861 tools/perf/util/sort.c if (!perf_hpp__is_sort_entry(a) || !perf_hpp__is_sort_entry(b)) b 1865 tools/perf/util/sort.c hse_b = container_of(b, struct hpp_sort_entry, hpp); b 2164 tools/perf/util/sort.c struct hist_entry *a, struct hist_entry *b) b 2172 tools/perf/util/sort.c if (b == NULL) { b 2193 tools/perf/util/sort.c return memcmp(a->raw_data + offset, b->raw_data + offset, size); b 2201 tools/perf/util/sort.c static bool __sort__hde_equal(struct perf_hpp_fmt *a, struct perf_hpp_fmt *b) b 2206 tools/perf/util/sort.c if (!perf_hpp__is_dynamic_entry(a) || !perf_hpp__is_dynamic_entry(b)) b 2210 tools/perf/util/sort.c hde_b = container_of(b, struct hpp_dynamic_entry, hpp); b 688 tools/perf/util/stat-display.c static int cmp_val(const void *a, const void *b) b 690 tools/perf/util/stat-display.c return ((struct perf_aggr_thread_value *)b)->val - b 42 tools/perf/util/stat-shadow.c const struct saved_value *b = entry; b 44 tools/perf/util/stat-shadow.c if (a->cpu != b->cpu) b 45 tools/perf/util/stat-shadow.c return a->cpu - b->cpu; b 54 tools/perf/util/stat-shadow.c if (a->type != b->type) b 55 tools/perf/util/stat-shadow.c return a->type - b->type; b 57 tools/perf/util/stat-shadow.c if (a->ctx != b->ctx) b 58 tools/perf/util/stat-shadow.c return a->ctx - b->ctx; b 60 tools/perf/util/stat-shadow.c if (a->evsel == NULL && b->evsel == NULL) { b 61 tools/perf/util/stat-shadow.c if (a->stat == b->stat) b 64 tools/perf/util/stat-shadow.c if ((char *)a->stat < (char *)b->stat) b 70 tools/perf/util/stat-shadow.c if (a->evsel == b->evsel) b 72 tools/perf/util/stat-shadow.c if ((char *)a->evsel < (char *)b->evsel) b 726 tools/perf/util/svghelper.c static int str_to_bitmap(char *s, cpumask_t *b, int nr_cpus) b 744 tools/perf/util/svghelper.c set_bit(c, cpumask_bits(b)); b 139 tools/perf/util/symbol.c s64 b; b 144 tools/perf/util/symbol.c b = symb->end - symb->start; b 145 tools/perf/util/symbol.c if ((b == 0) && (a > 0)) b 147 tools/perf/util/symbol.c else if ((a == 0) && (b > 0)) b 152 tools/perf/util/symbol.c b = symb->binding == STB_WEAK; b 153 tools/perf/util/symbol.c if (b && !a) b 155 tools/perf/util/symbol.c if (a && !b) b 160 tools/perf/util/symbol.c b = symb->binding == STB_GLOBAL; b 161 tools/perf/util/symbol.c if (a && !b) b 163 tools/perf/util/symbol.c if (b && !a) b 168 tools/perf/util/symbol.c b = prefix_underscores_count(symb->name); b 169 tools/perf/util/symbol.c if (b > a) b 171 tools/perf/util/symbol.c else if (a > b) b 54 tools/perf/util/syscalltbl.c const struct syscall *a = va, *b = vb; b 56 tools/perf/util/syscalltbl.c return strcmp(a->name, b->name); b 15 tools/power/acpi/tools/acpidump/acpidump.h #define INIT_GLOBAL(a,b) a=b b 18 tools/power/acpi/tools/acpidump/acpidump.h #define INIT_GLOBAL(a,b) a b 17 tools/power/cpupower/utils/helpers/bitmask.c #define max(a, b) ((a) > (b) ? (a) : (b)) b 201 tools/power/cpupower/utils/helpers/bitmask.c unsigned int b; /* end of range */ b 210 tools/power/cpupower/utils/helpers/bitmask.c b = a; b 215 tools/power/cpupower/utils/helpers/bitmask.c sret = sscanf(c1, "%u%c", &b, &nextc); b 225 tools/power/cpupower/utils/helpers/bitmask.c if (!(a <= b)) b 227 tools/power/cpupower/utils/helpers/bitmask.c if (b >= bmp->size) b 229 tools/power/cpupower/utils/helpers/bitmask.c while (a <= b) { b 148 tools/power/x86/turbostat/turbostat.c #define MAX(a, b) ((a) > (b) ? (a) : (b)) b 1 tools/testing/radix-tree/linux/cpu.h #define cpuhp_setup_state_nocalls(a, b, c, d) (0) b 61 tools/testing/scatterlist/linux/mm.h #define ___PASTE(a,b) a##b b 62 tools/testing/scatterlist/linux/mm.h #define __PASTE(a,b) ___PASTE(a,b) b 119 tools/testing/scatterlist/linux/mm.h #define kmemleak_alloc(a, b, c, d) b 9 tools/testing/selftests/bpf/prog_tests/core_reloc.c .b = 0xc001, \ b 33 tools/testing/selftests/bpf/prog_tests/core_reloc.c .b = { .b = { .b = 0xc001 } }, \ b 56 tools/testing/selftests/bpf/prog_tests/core_reloc.c .b = { [1] = { [2] = { [3] = 2 } } }, \ b 86 tools/testing/selftests/bpf/prog_tests/core_reloc.c .b = 2, \ b 116 tools/testing/selftests/bpf/prog_tests/core_reloc.c .b = 2, \ b 126 tools/testing/selftests/bpf/prog_tests/core_reloc.c .a = 1, .b = 2, .c = 3, .d = 4, \ b 277 tools/testing/selftests/bpf/prog_tests/core_reloc.c .b = 1, b 70 tools/testing/selftests/bpf/prog_tests/global_data.c __u32 b; b 98 tools/testing/selftests/bpf/prog_tests/global_data.c err, val.a, val.b, val.c, tests[i].val.a, tests[i].val.b, tests[i].val.c); b 15 tools/testing/selftests/bpf/prog_tests/tcp_rtt.c char b = 0x55; b 17 tools/testing/selftests/bpf/prog_tests/tcp_rtt.c if (CHECK_FAIL(write(fd, &b, sizeof(b)) != 1)) b 30 tools/testing/selftests/bpf/progs/btf_dump_test_case_bitfields.c long int b: 2; b 59 tools/testing/selftests/bpf/progs/btf_dump_test_case_bitfields.c short b; /* combined with previous bitfield */ b 82 tools/testing/selftests/bpf/progs/btf_dump_test_case_bitfields.c long b: 16; b 62 tools/testing/selftests/bpf/progs/btf_dump_test_case_namespacing.c union B b; b 11 tools/testing/selftests/bpf/progs/btf_dump_test_case_packing.c short b; b 16 tools/testing/selftests/bpf/progs/btf_dump_test_case_packing.c short b; b 21 tools/testing/selftests/bpf/progs/btf_dump_test_case_packing.c int b; b 26 tools/testing/selftests/bpf/progs/btf_dump_test_case_packing.c int b; b 32 tools/testing/selftests/bpf/progs/btf_dump_test_case_packing.c long int b; b 41 tools/testing/selftests/bpf/progs/btf_dump_test_case_packing.c char b; b 48 tools/testing/selftests/bpf/progs/btf_dump_test_case_packing.c int b; b 12 tools/testing/selftests/bpf/progs/btf_dump_test_case_padding.c long int b; b 32 tools/testing/selftests/bpf/progs/btf_dump_test_case_padding.c int b; b 53 tools/testing/selftests/bpf/progs/btf_dump_test_case_padding.c int b; b 76 tools/testing/selftests/bpf/progs/btf_dump_test_case_padding.c int b __attribute__((aligned(32))); b 101 tools/testing/selftests/bpf/progs/btf_dump_test_case_padding.c short b; b 80 tools/testing/selftests/bpf/progs/btf_dump_test_case_syntax.c void (*b)(int, struct { b 101 tools/testing/selftests/bpf/progs/btf_dump_test_case_syntax.c crazy_ptr_t b; b 136 tools/testing/selftests/bpf/progs/btf_dump_test_case_syntax.c char b; b 164 tools/testing/selftests/bpf/progs/btf_dump_test_case_syntax.c int b; b 180 tools/testing/selftests/bpf/progs/btf_dump_test_case_syntax.c int b; b 9 tools/testing/selftests/bpf/progs/core_reloc_types.h int b; b 16 tools/testing/selftests/bpf/progs/core_reloc_types.h int b; b 29 tools/testing/selftests/bpf/progs/core_reloc_types.h int b; b 37 tools/testing/selftests/bpf/progs/core_reloc_types.h union core_reloc_nesting_subunion b; b 38 tools/testing/selftests/bpf/progs/core_reloc_types.h } b; b 51 tools/testing/selftests/bpf/progs/core_reloc_types.h int b; b 52 tools/testing/selftests/bpf/progs/core_reloc_types.h } b; b 53 tools/testing/selftests/bpf/progs/core_reloc_types.h } b; b 71 tools/testing/selftests/bpf/progs/core_reloc_types.h int b; b 72 tools/testing/selftests/bpf/progs/core_reloc_types.h } b; b 73 tools/testing/selftests/bpf/progs/core_reloc_types.h } b; b 95 tools/testing/selftests/bpf/progs/core_reloc_types.h int b; b 97 tools/testing/selftests/bpf/progs/core_reloc_types.h } b; b 99 tools/testing/selftests/bpf/progs/core_reloc_types.h } b; b 117 tools/testing/selftests/bpf/progs/core_reloc_types.h int b; /* offset 16 */ b 118 tools/testing/selftests/bpf/progs/core_reloc_types.h } b; b 119 tools/testing/selftests/bpf/progs/core_reloc_types.h } b; b 136 tools/testing/selftests/bpf/progs/core_reloc_types.h int b; /* offset 16 */ b 137 tools/testing/selftests/bpf/progs/core_reloc_types.h } b; b 140 tools/testing/selftests/bpf/progs/core_reloc_types.h } b; b 154 tools/testing/selftests/bpf/progs/core_reloc_types.h int b; b 155 tools/testing/selftests/bpf/progs/core_reloc_types.h } b; b 156 tools/testing/selftests/bpf/progs/core_reloc_types.h } b __attribute__((aligned(16))); b 169 tools/testing/selftests/bpf/progs/core_reloc_types.h } b; b 170 tools/testing/selftests/bpf/progs/core_reloc_types.h } b; b 182 tools/testing/selftests/bpf/progs/core_reloc_types.h int b[1]; b 183 tools/testing/selftests/bpf/progs/core_reloc_types.h } b; b 184 tools/testing/selftests/bpf/progs/core_reloc_types.h } b; b 196 tools/testing/selftests/bpf/progs/core_reloc_types.h } b; b 208 tools/testing/selftests/bpf/progs/core_reloc_types.h int b; b 209 tools/testing/selftests/bpf/progs/core_reloc_types.h } *b; b 210 tools/testing/selftests/bpf/progs/core_reloc_types.h } b; b 222 tools/testing/selftests/bpf/progs/core_reloc_types.h int b; b 223 tools/testing/selftests/bpf/progs/core_reloc_types.h } b[1]; b 224 tools/testing/selftests/bpf/progs/core_reloc_types.h } b; b 236 tools/testing/selftests/bpf/progs/core_reloc_types.h int b; /* offset 4 */ b 237 tools/testing/selftests/bpf/progs/core_reloc_types.h } b; b 238 tools/testing/selftests/bpf/progs/core_reloc_types.h } b; b 250 tools/testing/selftests/bpf/progs/core_reloc_types.h int b; /* offset 8 (!) */ b 251 tools/testing/selftests/bpf/progs/core_reloc_types.h } b; b 252 tools/testing/selftests/bpf/progs/core_reloc_types.h } b; b 267 tools/testing/selftests/bpf/progs/core_reloc_types.h int b; b 268 tools/testing/selftests/bpf/progs/core_reloc_types.h } b; b 269 tools/testing/selftests/bpf/progs/core_reloc_types.h } b; b 295 tools/testing/selftests/bpf/progs/core_reloc_types.h int b; b 309 tools/testing/selftests/bpf/progs/core_reloc_types.h } b; b 310 tools/testing/selftests/bpf/progs/core_reloc_types.h } b; b 330 tools/testing/selftests/bpf/progs/core_reloc_types.h char b[2][3][4]; b 338 tools/testing/selftests/bpf/progs/core_reloc_types.h char b[3][4][5]; b 346 tools/testing/selftests/bpf/progs/core_reloc_types.h char b[2][3][4]; b 361 tools/testing/selftests/bpf/progs/core_reloc_types.h char b[2][3][4]; b 368 tools/testing/selftests/bpf/progs/core_reloc_types.h char b[2][3]; /* this one lacks one dimension */ b 375 tools/testing/selftests/bpf/progs/core_reloc_types.h char b[2][3][4]; b 382 tools/testing/selftests/bpf/progs/core_reloc_types.h char b[2][3][4]; b 389 tools/testing/selftests/bpf/progs/core_reloc_types.h char b[2][3][4]; b 404 tools/testing/selftests/bpf/progs/core_reloc_types.h int b; b 412 tools/testing/selftests/bpf/progs/core_reloc_types.h int b; b 425 tools/testing/selftests/bpf/progs/core_reloc_types.h int b; b 432 tools/testing/selftests/bpf/progs/core_reloc_types.h int b; b 439 tools/testing/selftests/bpf/progs/core_reloc_types.h int b; b 447 tools/testing/selftests/bpf/progs/core_reloc_types.h int *b; /* ptr instead of int */ b 455 tools/testing/selftests/bpf/progs/core_reloc_types.h int b; b 465 tools/testing/selftests/bpf/progs/core_reloc_types.h int a, b, c, d, e, f, g, h; b 484 tools/testing/selftests/bpf/progs/core_reloc_types.h int_t b; b 495 tools/testing/selftests/bpf/progs/core_reloc_types.h int b; b 529 tools/testing/selftests/bpf/progs/core_reloc_types.h int3_t b; b 648 tools/testing/selftests/bpf/progs/core_reloc_types.h int a, b, c; b 664 tools/testing/selftests/bpf/progs/core_reloc_types.h int b; b 29 tools/testing/selftests/bpf/progs/test_core_reloc_arrays.c char b[2][3][4]; b 44 tools/testing/selftests/bpf/progs/test_core_reloc_arrays.c if (BPF_CORE_READ(&out->b123, &in->b[1][2][3])) b 17 tools/testing/selftests/bpf/progs/test_core_reloc_flavors.c int b; b 24 tools/testing/selftests/bpf/progs/test_core_reloc_flavors.c int b; b 31 tools/testing/selftests/bpf/progs/test_core_reloc_flavors.c int b; b 54 tools/testing/selftests/bpf/progs/test_core_reloc_flavors.c if (BPF_CORE_READ(&out->b, &in_rev->b)) b 16 tools/testing/selftests/bpf/progs/test_core_reloc_misc.c int a, b, c; b 32 tools/testing/selftests/bpf/progs/test_core_reloc_misc.c int b; b 45 tools/testing/selftests/bpf/progs/test_core_reloc_misc.c BPF_CORE_READ(&out->b, &in_b->b1)) /* accessor: 0:0 */ b 16 tools/testing/selftests/bpf/progs/test_core_reloc_mods.c int a, b, c, d, e, f, g, h; b 35 tools/testing/selftests/bpf/progs/test_core_reloc_mods.c int_t b; b 51 tools/testing/selftests/bpf/progs/test_core_reloc_mods.c BPF_CORE_READ(&out->b, &in->b) || b 20 tools/testing/selftests/bpf/progs/test_core_reloc_nesting.c int b; b 29 tools/testing/selftests/bpf/progs/test_core_reloc_nesting.c union core_reloc_nesting_subunion b; b 30 tools/testing/selftests/bpf/progs/test_core_reloc_nesting.c } b; b 41 tools/testing/selftests/bpf/progs/test_core_reloc_nesting.c if (BPF_CORE_READ(&out->b.b.b, &in->b.b.b)) b 22 tools/testing/selftests/bpf/progs/test_core_reloc_primitives.c int b; b 35 tools/testing/selftests/bpf/progs/test_core_reloc_primitives.c BPF_CORE_READ(&out->b, &in->b) || b 26 tools/testing/selftests/bpf/progs/test_global_data.c __u32 b; b 54 tools/testing/selftests/bpf/progs/test_global_data.c .b = 0xfefeefef, b 61 tools/testing/selftests/bpf/progs/test_global_data.c .b = 0xeeeeefef, b 12 tools/testing/selftests/bpf/progs/test_jhash.h #define __jhash_mix(a, b, c) \ b 14 tools/testing/selftests/bpf/progs/test_jhash.h a -= c; a ^= rol32(c, 4); c += b; \ b 15 tools/testing/selftests/bpf/progs/test_jhash.h b -= a; b ^= rol32(a, 6); a += c; \ b 16 tools/testing/selftests/bpf/progs/test_jhash.h c -= b; c ^= rol32(b, 8); b += a; \ b 17 tools/testing/selftests/bpf/progs/test_jhash.h a -= c; a ^= rol32(c, 16); c += b; \ b 18 tools/testing/selftests/bpf/progs/test_jhash.h b -= a; b ^= rol32(a, 19); a += c; \ b 19 tools/testing/selftests/bpf/progs/test_jhash.h c -= b; c ^= rol32(b, 4); b += a; \ b 22 tools/testing/selftests/bpf/progs/test_jhash.h #define __jhash_final(a, b, c) \ b 24 tools/testing/selftests/bpf/progs/test_jhash.h c ^= b; c -= rol32(b, 14); \ b 26 tools/testing/selftests/bpf/progs/test_jhash.h b ^= a; b -= rol32(a, 25); \ b 27 tools/testing/selftests/bpf/progs/test_jhash.h c ^= b; c -= rol32(b, 16); \ b 29 tools/testing/selftests/bpf/progs/test_jhash.h b ^= a; b -= rol32(a, 14); \ b 30 tools/testing/selftests/bpf/progs/test_jhash.h c ^= b; c -= rol32(b, 24); \ b 38 tools/testing/selftests/bpf/progs/test_jhash.h u32 a, b, c; b 41 tools/testing/selftests/bpf/progs/test_jhash.h a = b = c = JHASH_INITVAL + length + initval; b 45 tools/testing/selftests/bpf/progs/test_jhash.h b += *(volatile u32 *)(k + 4); b 47 tools/testing/selftests/bpf/progs/test_jhash.h __jhash_mix(a, b, c); b 56 tools/testing/selftests/bpf/progs/test_jhash.h case 8: b += (u32)k[7]<<24; b 57 tools/testing/selftests/bpf/progs/test_jhash.h case 7: b += (u32)k[6]<<16; b 58 tools/testing/selftests/bpf/progs/test_jhash.h case 6: b += (u32)k[5]<<8; b 59 tools/testing/selftests/bpf/progs/test_jhash.h case 5: b += k[4]; b 65 tools/testing/selftests/bpf/progs/test_jhash.h __jhash_final(a, b, c); b 34 tools/testing/selftests/bpf/progs/test_l4lb.c #define __jhash_mix(a, b, c) \ b 36 tools/testing/selftests/bpf/progs/test_l4lb.c a -= c; a ^= rol32(c, 4); c += b; \ b 37 tools/testing/selftests/bpf/progs/test_l4lb.c b -= a; b ^= rol32(a, 6); a += c; \ b 38 tools/testing/selftests/bpf/progs/test_l4lb.c c -= b; c ^= rol32(b, 8); b += a; \ b 39 tools/testing/selftests/bpf/progs/test_l4lb.c a -= c; a ^= rol32(c, 16); c += b; \ b 40 tools/testing/selftests/bpf/progs/test_l4lb.c b -= a; b ^= rol32(a, 19); a += c; \ b 41 tools/testing/selftests/bpf/progs/test_l4lb.c c -= b; c ^= rol32(b, 4); b += a; \ b 44 tools/testing/selftests/bpf/progs/test_l4lb.c #define __jhash_final(a, b, c) \ b 46 tools/testing/selftests/bpf/progs/test_l4lb.c c ^= b; c -= rol32(b, 14); \ b 48 tools/testing/selftests/bpf/progs/test_l4lb.c b ^= a; b -= rol32(a, 25); \ b 49 tools/testing/selftests/bpf/progs/test_l4lb.c c ^= b; c -= rol32(b, 16); \ b 51 tools/testing/selftests/bpf/progs/test_l4lb.c b ^= a; b -= rol32(a, 14); \ b 52 tools/testing/selftests/bpf/progs/test_l4lb.c c ^= b; c -= rol32(b, 24); \ b 61 tools/testing/selftests/bpf/progs/test_l4lb.c u32 a, b, c; b 64 tools/testing/selftests/bpf/progs/test_l4lb.c a = b = c = JHASH_INITVAL + length + initval; b 68 tools/testing/selftests/bpf/progs/test_l4lb.c b += *(u32 *)(k + 4); b 70 tools/testing/selftests/bpf/progs/test_l4lb.c __jhash_mix(a, b, c); b 79 tools/testing/selftests/bpf/progs/test_l4lb.c case 8: b += (u32)k[7]<<24; b 80 tools/testing/selftests/bpf/progs/test_l4lb.c case 7: b += (u32)k[6]<<16; b 81 tools/testing/selftests/bpf/progs/test_l4lb.c case 6: b += (u32)k[5]<<8; b 82 tools/testing/selftests/bpf/progs/test_l4lb.c case 5: b += k[4]; b 87 tools/testing/selftests/bpf/progs/test_l4lb.c __jhash_final(a, b, c); b 95 tools/testing/selftests/bpf/progs/test_l4lb.c static inline u32 __jhash_nwords(u32 a, u32 b, u32 c, u32 initval) b 98 tools/testing/selftests/bpf/progs/test_l4lb.c b += initval; b 100 tools/testing/selftests/bpf/progs/test_l4lb.c __jhash_final(a, b, c); b 104 tools/testing/selftests/bpf/progs/test_l4lb.c static inline u32 jhash_2words(u32 a, u32 b, u32 initval) b 106 tools/testing/selftests/bpf/progs/test_l4lb.c return __jhash_nwords(a, b, 0, initval + JHASH_INITVAL + (2 << 2)); b 30 tools/testing/selftests/bpf/progs/test_l4lb_noinline.c #define __jhash_mix(a, b, c) \ b 32 tools/testing/selftests/bpf/progs/test_l4lb_noinline.c a -= c; a ^= rol32(c, 4); c += b; \ b 33 tools/testing/selftests/bpf/progs/test_l4lb_noinline.c b -= a; b ^= rol32(a, 6); a += c; \ b 34 tools/testing/selftests/bpf/progs/test_l4lb_noinline.c c -= b; c ^= rol32(b, 8); b += a; \ b 35 tools/testing/selftests/bpf/progs/test_l4lb_noinline.c a -= c; a ^= rol32(c, 16); c += b; \ b 36 tools/testing/selftests/bpf/progs/test_l4lb_noinline.c b -= a; b ^= rol32(a, 19); a += c; \ b 37 tools/testing/selftests/bpf/progs/test_l4lb_noinline.c c -= b; c ^= rol32(b, 4); b += a; \ b 40 tools/testing/selftests/bpf/progs/test_l4lb_noinline.c #define __jhash_final(a, b, c) \ b 42 tools/testing/selftests/bpf/progs/test_l4lb_noinline.c c ^= b; c -= rol32(b, 14); \ b 44 tools/testing/selftests/bpf/progs/test_l4lb_noinline.c b ^= a; b -= rol32(a, 25); \ b 45 tools/testing/selftests/bpf/progs/test_l4lb_noinline.c c ^= b; c -= rol32(b, 16); \ b 47 tools/testing/selftests/bpf/progs/test_l4lb_noinline.c b ^= a; b -= rol32(a, 14); \ b 48 tools/testing/selftests/bpf/progs/test_l4lb_noinline.c c ^= b; c -= rol32(b, 24); \ b 57 tools/testing/selftests/bpf/progs/test_l4lb_noinline.c u32 a, b, c; b 60 tools/testing/selftests/bpf/progs/test_l4lb_noinline.c a = b = c = JHASH_INITVAL + length + initval; b 64 tools/testing/selftests/bpf/progs/test_l4lb_noinline.c b += *(u32 *)(k + 4); b 66 tools/testing/selftests/bpf/progs/test_l4lb_noinline.c __jhash_mix(a, b, c); b 75 tools/testing/selftests/bpf/progs/test_l4lb_noinline.c case 8: b += (u32)k[7]<<24; b 76 tools/testing/selftests/bpf/progs/test_l4lb_noinline.c case 7: b += (u32)k[6]<<16; b 77 tools/testing/selftests/bpf/progs/test_l4lb_noinline.c case 6: b += (u32)k[5]<<8; b 78 tools/testing/selftests/bpf/progs/test_l4lb_noinline.c case 5: b += k[4]; b 83 tools/testing/selftests/bpf/progs/test_l4lb_noinline.c __jhash_final(a, b, c); b 91 tools/testing/selftests/bpf/progs/test_l4lb_noinline.c static u32 __jhash_nwords(u32 a, u32 b, u32 c, u32 initval) b 94 tools/testing/selftests/bpf/progs/test_l4lb_noinline.c b += initval; b 96 tools/testing/selftests/bpf/progs/test_l4lb_noinline.c __jhash_final(a, b, c); b 100 tools/testing/selftests/bpf/progs/test_l4lb_noinline.c static u32 jhash_2words(u32 a, u32 b, u32 initval) b 102 tools/testing/selftests/bpf/progs/test_l4lb_noinline.c return __jhash_nwords(a, b, 0, initval + JHASH_INITVAL + (2 << 2)); b 20 tools/testing/selftests/bpf/progs/test_pkt_md_access.c #define TEST_FIELD_OFFSET(a, b) ((sizeof(a) - sizeof(b)) / sizeof(b)) b 27 tools/testing/selftests/bpf/progs/test_xdp_noinline.c #define __jhash_mix(a, b, c) \ b 29 tools/testing/selftests/bpf/progs/test_xdp_noinline.c a -= c; a ^= rol32(c, 4); c += b; \ b 30 tools/testing/selftests/bpf/progs/test_xdp_noinline.c b -= a; b ^= rol32(a, 6); a += c; \ b 31 tools/testing/selftests/bpf/progs/test_xdp_noinline.c c -= b; c ^= rol32(b, 8); b += a; \ b 32 tools/testing/selftests/bpf/progs/test_xdp_noinline.c a -= c; a ^= rol32(c, 16); c += b; \ b 33 tools/testing/selftests/bpf/progs/test_xdp_noinline.c b -= a; b ^= rol32(a, 19); a += c; \ b 34 tools/testing/selftests/bpf/progs/test_xdp_noinline.c c -= b; c ^= rol32(b, 4); b += a; \ b 37 tools/testing/selftests/bpf/progs/test_xdp_noinline.c #define __jhash_final(a, b, c) \ b 39 tools/testing/selftests/bpf/progs/test_xdp_noinline.c c ^= b; c -= rol32(b, 14); \ b 41 tools/testing/selftests/bpf/progs/test_xdp_noinline.c b ^= a; b -= rol32(a, 25); \ b 42 tools/testing/selftests/bpf/progs/test_xdp_noinline.c c ^= b; c -= rol32(b, 16); \ b 44 tools/testing/selftests/bpf/progs/test_xdp_noinline.c b ^= a; b -= rol32(a, 14); \ b 45 tools/testing/selftests/bpf/progs/test_xdp_noinline.c c ^= b; c -= rol32(b, 24); \ b 55 tools/testing/selftests/bpf/progs/test_xdp_noinline.c u32 a, b, c; b 58 tools/testing/selftests/bpf/progs/test_xdp_noinline.c a = b = c = JHASH_INITVAL + length + initval; b 62 tools/testing/selftests/bpf/progs/test_xdp_noinline.c b += *(u32 *)(k + 4); b 64 tools/testing/selftests/bpf/progs/test_xdp_noinline.c __jhash_mix(a, b, c); b 73 tools/testing/selftests/bpf/progs/test_xdp_noinline.c case 8: b += (u32)k[7]<<24; b 74 tools/testing/selftests/bpf/progs/test_xdp_noinline.c case 7: b += (u32)k[6]<<16; b 75 tools/testing/selftests/bpf/progs/test_xdp_noinline.c case 6: b += (u32)k[5]<<8; b 76 tools/testing/selftests/bpf/progs/test_xdp_noinline.c case 5: b += k[4]; b 81 tools/testing/selftests/bpf/progs/test_xdp_noinline.c __jhash_final(a, b, c); b 90 tools/testing/selftests/bpf/progs/test_xdp_noinline.c u32 __jhash_nwords(u32 a, u32 b, u32 c, u32 initval) b 93 tools/testing/selftests/bpf/progs/test_xdp_noinline.c b += initval; b 95 tools/testing/selftests/bpf/progs/test_xdp_noinline.c __jhash_final(a, b, c); b 100 tools/testing/selftests/bpf/progs/test_xdp_noinline.c u32 jhash_2words(u32 a, u32 b, u32 initval) b 102 tools/testing/selftests/bpf/progs/test_xdp_noinline.c return __jhash_nwords(a, b, 0, initval + JHASH_INITVAL + (2 << 2)); b 27 tools/testing/selftests/bpf/test_hashmap.c bool equal_fn(const void *a, const void *b, void *ctx) b 29 tools/testing/selftests/bpf/test_hashmap.c return (long)a == (long)b; b 1236 tools/testing/selftests/bpf/test_maps.c char b[116]; b 13 tools/testing/selftests/cgroup/cgroup_util.h static inline int values_close(long a, long b, int err) b 15 tools/testing/selftests/cgroup/cgroup_util.h return abs(a - b) <= (a + b) / 100 * err; b 32 tools/testing/selftests/kvm/include/test_util.h #define ASSERT_EQ(a, b) do { \ b 34 tools/testing/selftests/kvm/include/test_util.h typeof(b) __b = (b); \ b 39 tools/testing/selftests/kvm/include/test_util.h #a, #b, #a, (unsigned long) __a, #b, (unsigned long) __b); \ b 76 tools/testing/selftests/mqueue/mq_perf_tests.c #define min(a, b) ((a) < (b) ? (a) : (b)) b 38 tools/testing/selftests/net/nettest.c #define MAX(a, b) ((a) > (b) ? (a) : (b)) b 41 tools/testing/selftests/net/nettest.c #define MIN(a, b) ((a) < (b) ? (a) : (b)) b 39 tools/testing/selftests/powerpc/benchmarks/context_switch.c vector int a, b, c; b 50 tools/testing/selftests/powerpc/benchmarks/context_switch.c c = a + b; b 68 tools/testing/selftests/powerpc/benchmarks/context_switch.c c = a + b; b 20 tools/testing/selftests/powerpc/primitives/asm/asm-compat.h #define PPC_LLARX(t, a, b, eh) PPC_LDARX(t, a, b, eh) b 53 tools/testing/selftests/powerpc/primitives/asm/asm-compat.h #define PPC_LLARX(t, a, b, eh) PPC_LWARX(t, a, b, eh) b 110 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h #define SAVE_VR(n,b,base) li b,16*(n); stvx n,base,b b 111 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h #define SAVE_2VRS(n,b,base) SAVE_VR(n,b,base); SAVE_VR(n+1,b,base) b 112 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h #define SAVE_4VRS(n,b,base) SAVE_2VRS(n,b,base); SAVE_2VRS(n+2,b,base) b 113 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h #define SAVE_8VRS(n,b,base) SAVE_4VRS(n,b,base); SAVE_4VRS(n+4,b,base) b 114 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h #define SAVE_16VRS(n,b,base) SAVE_8VRS(n,b,base); SAVE_8VRS(n+8,b,base) b 115 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h #define SAVE_32VRS(n,b,base) SAVE_16VRS(n,b,base); SAVE_16VRS(n+16,b,base) b 116 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h #define REST_VR(n,b,base) li b,16*(n); lvx n,base,b b 117 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h #define REST_2VRS(n,b,base) REST_VR(n,b,base); REST_VR(n+1,b,base) b 118 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h #define REST_4VRS(n,b,base) REST_2VRS(n,b,base); REST_2VRS(n+2,b,base) b 119 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h #define REST_8VRS(n,b,base) REST_4VRS(n,b,base); REST_4VRS(n+4,b,base) b 120 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h #define REST_16VRS(n,b,base) REST_8VRS(n,b,base); REST_8VRS(n+8,b,base) b 121 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h #define REST_32VRS(n,b,base) REST_16VRS(n,b,base); REST_16VRS(n+16,b,base) b 124 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h #define STXVD2X_ROT(n,b,base) STXVD2X(n,b,base) b 125 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h #define LXVD2X_ROT(n,b,base) LXVD2X(n,b,base) b 127 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h #define STXVD2X_ROT(n,b,base) XXSWAPD(n,n); \ b 128 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h STXVD2X(n,b,base); \ b 131 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h #define LXVD2X_ROT(n,b,base) LXVD2X(n,b,base); \ b 135 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h #define SAVE_VSR(n,b,base) li b,16*(n); STXVD2X_ROT(n,R##base,R##b) b 136 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h #define SAVE_2VSRS(n,b,base) SAVE_VSR(n,b,base); SAVE_VSR(n+1,b,base) b 137 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h #define SAVE_4VSRS(n,b,base) SAVE_2VSRS(n,b,base); SAVE_2VSRS(n+2,b,base) b 138 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h #define SAVE_8VSRS(n,b,base) SAVE_4VSRS(n,b,base); SAVE_4VSRS(n+4,b,base) b 139 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h #define SAVE_16VSRS(n,b,base) SAVE_8VSRS(n,b,base); SAVE_8VSRS(n+8,b,base) b 140 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h #define SAVE_32VSRS(n,b,base) SAVE_16VSRS(n,b,base); SAVE_16VSRS(n+16,b,base) b 141 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h #define REST_VSR(n,b,base) li b,16*(n); LXVD2X_ROT(n,R##base,R##b) b 142 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h #define REST_2VSRS(n,b,base) REST_VSR(n,b,base); REST_VSR(n+1,b,base) b 143 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h #define REST_4VSRS(n,b,base) REST_2VSRS(n,b,base); REST_2VSRS(n+2,b,base) b 144 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h #define REST_8VSRS(n,b,base) REST_4VSRS(n,b,base); REST_4VSRS(n+4,b,base) b 145 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h #define REST_16VSRS(n,b,base) REST_8VSRS(n,b,base); REST_8VSRS(n+8,b,base) b 146 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h #define REST_32VSRS(n,b,base) REST_16VSRS(n,b,base); REST_16VSRS(n+16,b,base) b 152 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h #define SAVE_EVR(n,s,b,o) evmergehi s,s,n; stw s,o+4*(n)(b) b 153 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h #define SAVE_2EVRS(n,s,b,o) SAVE_EVR(n,s,b,o); SAVE_EVR(n+1,s,b,o) b 154 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h #define SAVE_4EVRS(n,s,b,o) SAVE_2EVRS(n,s,b,o); SAVE_2EVRS(n+2,s,b,o) b 155 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h #define SAVE_8EVRS(n,s,b,o) SAVE_4EVRS(n,s,b,o); SAVE_4EVRS(n+4,s,b,o) b 156 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h #define SAVE_16EVRS(n,s,b,o) SAVE_8EVRS(n,s,b,o); SAVE_8EVRS(n+8,s,b,o) b 157 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h #define SAVE_32EVRS(n,s,b,o) SAVE_16EVRS(n,s,b,o); SAVE_16EVRS(n+16,s,b,o) b 158 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h #define REST_EVR(n,s,b,o) lwz s,o+4*(n)(b); evmergelo n,s,n b 159 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h #define REST_2EVRS(n,s,b,o) REST_EVR(n,s,b,o); REST_EVR(n+1,s,b,o) b 160 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h #define REST_4EVRS(n,s,b,o) REST_2EVRS(n,s,b,o); REST_2EVRS(n+2,s,b,o) b 161 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h #define REST_8EVRS(n,s,b,o) REST_4EVRS(n,s,b,o); REST_4EVRS(n+4,s,b,o) b 162 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h #define REST_16EVRS(n,s,b,o) REST_8EVRS(n,s,b,o); REST_8EVRS(n+8,s,b,o) b 163 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h #define REST_32EVRS(n,s,b,o) REST_16EVRS(n,s,b,o); REST_16EVRS(n+16,s,b,o) b 219 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h #define XGLUE(a,b) a##b b 220 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h #define GLUE(a,b) XGLUE(a,b) b 516 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h #define RFI rfi; b . /* Prevent prefetch past rfi */ b 782 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h b 191f; /* Skip trampoline if endian is good */ \ b 802 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h b 191f; /* Skip trampoline if endian is good */ \ b 16 tools/testing/selftests/powerpc/ptrace/ptrace-gpr.c float b = FPR_2; b 16 tools/testing/selftests/powerpc/ptrace/ptrace-tm-gpr.c float b = FPR_2; b 59 tools/testing/selftests/powerpc/ptrace/ptrace-tm-gpr.c [flt_2] "b" (&b), [cptr1] "b" (&cptr[1]) b 16 tools/testing/selftests/powerpc/ptrace/ptrace-tm-spd-gpr.c float b = FPR_2; b 23 tools/testing/selftests/powerpc/stringloops/asm/ppc-opcode.h #define ___PPC_RB(b) (((b) & 0x1f) << 11) b 26 tools/testing/selftests/proc/proc-self-map-files-001.c static void pass(const char *fmt, unsigned long a, unsigned long b) b 31 tools/testing/selftests/proc/proc-self-map-files-001.c snprintf(name, sizeof(name), fmt, a, b); b 36 tools/testing/selftests/proc/proc-self-map-files-001.c static void fail(const char *fmt, unsigned long a, unsigned long b) b 41 tools/testing/selftests/proc/proc-self-map-files-001.c snprintf(name, sizeof(name), fmt, a, b); b 52 tools/testing/selftests/proc/proc-self-map-files-001.c unsigned long a, b; b 63 tools/testing/selftests/proc/proc-self-map-files-001.c b = (unsigned long)p + PAGE_SIZE; b 65 tools/testing/selftests/proc/proc-self-map-files-001.c pass("/proc/self/map_files/%lx-%lx", a, b); b 66 tools/testing/selftests/proc/proc-self-map-files-001.c fail("/proc/self/map_files/ %lx-%lx", a, b); b 67 tools/testing/selftests/proc/proc-self-map-files-001.c fail("/proc/self/map_files/%lx -%lx", a, b); b 68 tools/testing/selftests/proc/proc-self-map-files-001.c fail("/proc/self/map_files/%lx- %lx", a, b); b 69 tools/testing/selftests/proc/proc-self-map-files-001.c fail("/proc/self/map_files/%lx-%lx ", a, b); b 70 tools/testing/selftests/proc/proc-self-map-files-001.c fail("/proc/self/map_files/0%lx-%lx", a, b); b 71 tools/testing/selftests/proc/proc-self-map-files-001.c fail("/proc/self/map_files/%lx-0%lx", a, b); b 73 tools/testing/selftests/proc/proc-self-map-files-001.c fail("/proc/self/map_files/100000000%lx-%lx", a, b); b 74 tools/testing/selftests/proc/proc-self-map-files-001.c fail("/proc/self/map_files/%lx-100000000%lx", a, b); b 76 tools/testing/selftests/proc/proc-self-map-files-001.c fail("/proc/self/map_files/10000000000000000%lx-%lx", a, b); b 77 tools/testing/selftests/proc/proc-self-map-files-001.c fail("/proc/self/map_files/%lx-10000000000000000%lx", a, b); b 26 tools/testing/selftests/proc/proc-self-map-files-002.c static void pass(const char *fmt, unsigned long a, unsigned long b) b 31 tools/testing/selftests/proc/proc-self-map-files-002.c snprintf(name, sizeof(name), fmt, a, b); b 36 tools/testing/selftests/proc/proc-self-map-files-002.c static void fail(const char *fmt, unsigned long a, unsigned long b) b 41 tools/testing/selftests/proc/proc-self-map-files-002.c snprintf(name, sizeof(name), fmt, a, b); b 58 tools/testing/selftests/proc/proc-self-map-files-002.c unsigned long a, b; b 75 tools/testing/selftests/proc/proc-self-map-files-002.c b = (unsigned long)p + PAGE_SIZE; b 77 tools/testing/selftests/proc/proc-self-map-files-002.c pass("/proc/self/map_files/%lx-%lx", a, b); b 78 tools/testing/selftests/proc/proc-self-map-files-002.c fail("/proc/self/map_files/ %lx-%lx", a, b); b 79 tools/testing/selftests/proc/proc-self-map-files-002.c fail("/proc/self/map_files/%lx -%lx", a, b); b 80 tools/testing/selftests/proc/proc-self-map-files-002.c fail("/proc/self/map_files/%lx- %lx", a, b); b 81 tools/testing/selftests/proc/proc-self-map-files-002.c fail("/proc/self/map_files/%lx-%lx ", a, b); b 82 tools/testing/selftests/proc/proc-self-map-files-002.c fail("/proc/self/map_files/0%lx-%lx", a, b); b 83 tools/testing/selftests/proc/proc-self-map-files-002.c fail("/proc/self/map_files/%lx-0%lx", a, b); b 85 tools/testing/selftests/proc/proc-self-map-files-002.c fail("/proc/self/map_files/100000000%lx-%lx", a, b); b 86 tools/testing/selftests/proc/proc-self-map-files-002.c fail("/proc/self/map_files/%lx-100000000%lx", a, b); b 88 tools/testing/selftests/proc/proc-self-map-files-002.c fail("/proc/self/map_files/10000000000000000%lx-%lx", a, b); b 89 tools/testing/selftests/proc/proc-self-map-files-002.c fail("/proc/self/map_files/%lx-10000000000000000%lx", a, b); b 78 tools/testing/selftests/sync/sync_fence.c int a, b, c, d, valid; b 83 tools/testing/selftests/sync/sync_fence.c b = sw_sync_fence_create(timeline, "allocFence", 2); b 87 tools/testing/selftests/sync/sync_fence.c sw_sync_fence_is_valid(b) && b 91 tools/testing/selftests/sync/sync_fence.c d = sync_merge("mergeFence", b, a); b 114 tools/testing/selftests/sync/sync_fence.c ASSERT(sync_fence_count_with_status(b, FENCE_STATUS_SIGNALED) == 1, b 128 tools/testing/selftests/sync/sync_fence.c sw_sync_fence_destroy(b); b 77 tools/testing/selftests/sync/sync_stress_parallelism.c pthread_t a, b; b 96 tools/testing/selftests/sync/sync_stress_parallelism.c pthread_create(&b, NULL, (void *(*)(void *)) b 101 tools/testing/selftests/sync/sync_stress_parallelism.c pthread_join(b, NULL); b 87 tools/testing/selftests/timers/alarmtimer-suspend.c long long timespec_sub(struct timespec a, struct timespec b) b 89 tools/testing/selftests/timers/alarmtimer-suspend.c long long ret = NSEC_PER_SEC * b.tv_sec + b.tv_nsec; b 80 tools/testing/selftests/timers/inconsistency-check.c static inline int in_order(struct timespec a, struct timespec b) b 83 tools/testing/selftests/timers/inconsistency-check.c if ((unsigned long)a.tv_sec < (unsigned long)b.tv_sec) b 85 tools/testing/selftests/timers/inconsistency-check.c if ((unsigned long)a.tv_sec > (unsigned long)b.tv_sec) b 87 tools/testing/selftests/timers/inconsistency-check.c if (a.tv_nsec > b.tv_nsec) b 60 tools/testing/selftests/timers/leap-a-day.c static inline int in_order(struct timespec a, struct timespec b) b 62 tools/testing/selftests/timers/leap-a-day.c if (a.tv_sec < b.tv_sec) b 64 tools/testing/selftests/timers/leap-a-day.c if (a.tv_sec > b.tv_sec) b 66 tools/testing/selftests/timers/leap-a-day.c if (a.tv_nsec > b.tv_nsec) b 40 tools/testing/selftests/timers/mqueue-lat.c long long timespec_sub(struct timespec a, struct timespec b) b 42 tools/testing/selftests/timers/mqueue-lat.c long long ret = NSEC_PER_SEC * b.tv_sec + b.tv_nsec; b 80 tools/testing/selftests/timers/nanosleep.c static inline int in_order(struct timespec a, struct timespec b) b 82 tools/testing/selftests/timers/nanosleep.c if (a.tv_sec < b.tv_sec) b 84 tools/testing/selftests/timers/nanosleep.c if (a.tv_sec > b.tv_sec) b 86 tools/testing/selftests/timers/nanosleep.c if (a.tv_nsec > b.tv_nsec) b 90 tools/testing/selftests/timers/nsleep-lat.c long long timespec_sub(struct timespec a, struct timespec b) b 92 tools/testing/selftests/timers/nsleep-lat.c long long ret = NSEC_PER_SEC * b.tv_sec + b.tv_nsec; b 88 tools/testing/selftests/timers/set-timer-lat.c long long timespec_sub(struct timespec a, struct timespec b) b 90 tools/testing/selftests/timers/set-timer-lat.c long long ret = NSEC_PER_SEC * b.tv_sec + b.tv_nsec; b 44 tools/testing/selftests/timers/threadtest.c struct timespec *a, *b; b 49 tools/testing/selftests/timers/threadtest.c b = &list[i+1]; b 52 tools/testing/selftests/timers/threadtest.c if ((b->tv_sec <= a->tv_sec) && b 53 tools/testing/selftests/timers/threadtest.c (b->tv_nsec < a->tv_nsec)) { b 24 tools/testing/selftests/vDSO/vdso_standalone_test_x86.c int strcmp(const char *a, const char *b) b 27 tools/testing/selftests/vDSO/vdso_standalone_test_x86.c while (*a || *b) { b 28 tools/testing/selftests/vDSO/vdso_standalone_test_x86.c if (*a != *b) b 30 tools/testing/selftests/vDSO/vdso_standalone_test_x86.c if (*a == 0 || *b == 0) b 33 tools/testing/selftests/vDSO/vdso_standalone_test_x86.c b++; b 62 tools/testing/selftests/vm/transhuge-stress.c struct timespec a, b; b 136 tools/testing/selftests/vm/transhuge-stress.c clock_gettime(CLOCK_MONOTONIC, &b); b 137 tools/testing/selftests/vm/transhuge-stress.c s = b.tv_sec - a.tv_sec + (b.tv_nsec - a.tv_nsec) / 1000000000.; b 186 tools/testing/selftests/x86/test_vdso.c static bool ts_leq(const struct timespec *a, const struct timespec *b) b 188 tools/testing/selftests/x86/test_vdso.c if (a->tv_sec != b->tv_sec) b 189 tools/testing/selftests/x86/test_vdso.c return a->tv_sec < b->tv_sec; b 191 tools/testing/selftests/x86/test_vdso.c return a->tv_nsec <= b->tv_nsec; b 194 tools/testing/selftests/x86/test_vdso.c static bool tv_leq(const struct timeval *a, const struct timeval *b) b 196 tools/testing/selftests/x86/test_vdso.c if (a->tv_sec != b->tv_sec) b 197 tools/testing/selftests/x86/test_vdso.c return a->tv_sec < b->tv_sec; b 199 tools/testing/selftests/x86/test_vdso.c return a->tv_usec <= b->tv_usec; b 191 tools/testing/selftests/x86/test_vsyscall.c static double tv_diff(const struct timeval *a, const struct timeval *b) b 193 tools/testing/selftests/x86/test_vsyscall.c return (double)(a->tv_sec - b->tv_sec) + b 194 tools/testing/selftests/x86/test_vsyscall.c (double)((int)a->tv_usec - (int)b->tv_usec) * 1e-6; b 48 tools/virtio/linux/kernel.h #define __printf(a,b) __attribute__((format(printf,a,b))) b 139 tools/virtio/linux/kernel.h #define list_add_tail(a, b) do {} while (0) b 141 tools/virtio/linux/kernel.h #define list_for_each_entry(a, b, c) while (0) b 389 tools/vm/slabinfo.c char b[20]; b 391 tools/vm/slabinfo.c store_size(b, s->numa[node]); b 392 tools/vm/slabinfo.c printf(" %4s", b); b 398 tools/vm/slabinfo.c char b[20]; b 400 tools/vm/slabinfo.c store_size(b, s->numa_partial[node]); b 401 tools/vm/slabinfo.c printf(" %4s", b); b 2186 virt/kvm/arm/vgic/vgic-its.c struct list_head *b) b 2189 virt/kvm/arm/vgic/vgic-its.c struct its_ite *iteb = container_of(b, struct its_ite, ite_list); b 2325 virt/kvm/arm/vgic/vgic-its.c struct list_head *b) b 2328 virt/kvm/arm/vgic/vgic-its.c struct its_device *devb = container_of(b, struct its_device, dev_list); b 258 virt/kvm/arm/vgic/vgic.c static int vgic_irq_cmp(void *priv, struct list_head *a, struct list_head *b) b 261 virt/kvm/arm/vgic/vgic.c struct vgic_irq *irqb = container_of(b, struct vgic_irq, ap_list);