/linux-4.4.14/fs/ |
D | fs-writeback.c | 102 static bool wb_io_lists_populated(struct bdi_writeback *wb) in wb_io_lists_populated() argument 104 if (wb_has_dirty_io(wb)) { in wb_io_lists_populated() 107 set_bit(WB_has_dirty_io, &wb->state); in wb_io_lists_populated() 108 WARN_ON_ONCE(!wb->avg_write_bandwidth); in wb_io_lists_populated() 109 atomic_long_add(wb->avg_write_bandwidth, in wb_io_lists_populated() 110 &wb->bdi->tot_write_bandwidth); in wb_io_lists_populated() 115 static void wb_io_lists_depopulated(struct bdi_writeback *wb) in wb_io_lists_depopulated() argument 117 if (wb_has_dirty_io(wb) && list_empty(&wb->b_dirty) && in wb_io_lists_depopulated() 118 list_empty(&wb->b_io) && list_empty(&wb->b_more_io)) { in wb_io_lists_depopulated() 119 clear_bit(WB_has_dirty_io, &wb->state); in wb_io_lists_depopulated() [all …]
|
/linux-4.4.14/mm/ |
D | backing-dev.c | 49 struct bdi_writeback *wb = &bdi->wb; in bdi_debug_stats_show() local 57 spin_lock(&wb->list_lock); in bdi_debug_stats_show() 58 list_for_each_entry(inode, &wb->b_dirty, i_io_list) in bdi_debug_stats_show() 60 list_for_each_entry(inode, &wb->b_io, i_io_list) in bdi_debug_stats_show() 62 list_for_each_entry(inode, &wb->b_more_io, i_io_list) in bdi_debug_stats_show() 64 list_for_each_entry(inode, &wb->b_dirty_time, i_io_list) in bdi_debug_stats_show() 67 spin_unlock(&wb->list_lock); in bdi_debug_stats_show() 70 wb_thresh = wb_calc_thresh(wb, dirty_thresh); in bdi_debug_stats_show() 88 (unsigned long) K(wb_stat(wb, WB_WRITEBACK)), in bdi_debug_stats_show() 89 (unsigned long) K(wb_stat(wb, WB_RECLAIMABLE)), in bdi_debug_stats_show() [all …]
|
D | page-writeback.c | 133 struct bdi_writeback *wb; member 157 #define GDTC_INIT(__wb) .wb = (__wb), \ 163 #define MDTC_INIT(__wb, __gdtc) .wb = (__wb), \ 183 static struct fprop_local_percpu *wb_memcg_completions(struct bdi_writeback *wb) in wb_memcg_completions() argument 185 return &wb->memcg_completions; in wb_memcg_completions() 188 static void wb_min_max_ratio(struct bdi_writeback *wb, in wb_min_max_ratio() argument 191 unsigned long this_bw = wb->avg_write_bandwidth; in wb_min_max_ratio() 192 unsigned long tot_bw = atomic_long_read(&wb->bdi->tot_write_bandwidth); in wb_min_max_ratio() 193 unsigned long long min = wb->bdi->min_ratio; in wb_min_max_ratio() 194 unsigned long long max = wb->bdi->max_ratio; in wb_min_max_ratio() [all …]
|
D | memcontrol.c | 3725 struct wb_domain *mem_cgroup_wb_domain(struct bdi_writeback *wb) in mem_cgroup_wb_domain() argument 3727 struct mem_cgroup *memcg = mem_cgroup_from_css(wb->memcg_css); in mem_cgroup_wb_domain() 3753 void mem_cgroup_wb_stats(struct bdi_writeback *wb, unsigned long *pfilepages, in mem_cgroup_wb_stats() argument 3757 struct mem_cgroup *memcg = mem_cgroup_from_css(wb->memcg_css); in mem_cgroup_wb_stats()
|
/linux-4.4.14/fs/afs/ |
D | write.c | 19 static int afs_write_back_from_locked_page(struct afs_writeback *wb, 35 static void afs_unlink_writeback(struct afs_writeback *wb) in afs_unlink_writeback() argument 38 struct afs_vnode *vnode = wb->vnode; in afs_unlink_writeback() 40 list_del_init(&wb->link); in afs_unlink_writeback() 57 static void afs_free_writeback(struct afs_writeback *wb) in afs_free_writeback() argument 60 key_put(wb->key); in afs_free_writeback() 61 kfree(wb); in afs_free_writeback() 67 void afs_put_writeback(struct afs_writeback *wb) in afs_put_writeback() argument 69 struct afs_vnode *vnode = wb->vnode; in afs_put_writeback() 71 _enter("{%d}", wb->usage); in afs_put_writeback() [all …]
|
D | file.c | 315 struct afs_writeback *wb = (struct afs_writeback *) page_private(page); in afs_invalidatepage() local 332 if (wb && !PageWriteback(page)) { in afs_invalidatepage() 334 afs_put_writeback(wb); in afs_invalidatepage() 351 struct afs_writeback *wb = (struct afs_writeback *) page_private(page); in afs_releasepage() local 368 if (wb) { in afs_releasepage() 370 afs_put_writeback(wb); in afs_releasepage()
|
D | fsclient.c | 1156 struct afs_writeback *wb, in afs_fs_store_data64() argument 1162 struct afs_vnode *vnode = wb->vnode; in afs_fs_store_data64() 1167 key_serial(wb->key), vnode->fid.vid, vnode->fid.vnode); in afs_fs_store_data64() 1175 call->wb = wb; in afs_fs_store_data64() 1176 call->key = wb->key; in afs_fs_store_data64() 1215 int afs_fs_store_data(struct afs_server *server, struct afs_writeback *wb, in afs_fs_store_data() argument 1220 struct afs_vnode *vnode = wb->vnode; in afs_fs_store_data() 1226 key_serial(wb->key), vnode->fid.vid, vnode->fid.vnode); in afs_fs_store_data() 1243 return afs_fs_store_data64(server, wb, first, last, offset, to, in afs_fs_store_data() 1252 call->wb = wb; in afs_fs_store_data() [all …]
|
D | vnode.c | 769 int afs_vnode_store_data(struct afs_writeback *wb, pgoff_t first, pgoff_t last, in afs_vnode_store_data() argument 773 struct afs_vnode *vnode = wb->vnode; in afs_vnode_store_data() 781 key_serial(wb->key), in afs_vnode_store_data() 797 ret = afs_fs_store_data(server, wb, first, last, offset, to, in afs_vnode_store_data()
|
D | internal.h | 87 struct afs_writeback *wb; /* writeback being performed */ member
|
/linux-4.4.14/include/linux/ |
D | backing-dev.h | 32 void wb_start_writeback(struct bdi_writeback *wb, long nr_pages, 34 void wb_start_background_writeback(struct bdi_writeback *wb); 36 void wb_wakeup_delayed(struct bdi_writeback *wb); 43 static inline bool wb_has_dirty_io(struct bdi_writeback *wb) in wb_has_dirty_io() argument 45 return test_bit(WB_has_dirty_io, &wb->state); in wb_has_dirty_io() 57 static inline void __add_wb_stat(struct bdi_writeback *wb, in __add_wb_stat() argument 60 __percpu_counter_add(&wb->stat[item], amount, WB_STAT_BATCH); in __add_wb_stat() 63 static inline void __inc_wb_stat(struct bdi_writeback *wb, in __inc_wb_stat() argument 66 __add_wb_stat(wb, item, 1); in __inc_wb_stat() 69 static inline void inc_wb_stat(struct bdi_writeback *wb, enum wb_stat_item item) in inc_wb_stat() argument [all …]
|
D | backing-dev-defs.h | 154 struct bdi_writeback wb; /* the root writeback info for this bdi */ member 185 clear_wb_congested(bdi->wb.congested, sync); in clear_bdi_congested() 190 set_wb_congested(bdi->wb.congested, sync); in set_bdi_congested() 199 static inline bool wb_tryget(struct bdi_writeback *wb) in wb_tryget() argument 201 if (wb != &wb->bdi->wb) in wb_tryget() 202 return percpu_ref_tryget(&wb->refcnt); in wb_tryget() 210 static inline void wb_get(struct bdi_writeback *wb) in wb_get() argument 212 if (wb != &wb->bdi->wb) in wb_get() 213 percpu_ref_get(&wb->refcnt); in wb_get() 220 static inline void wb_put(struct bdi_writeback *wb) in wb_put() argument [all …]
|
D | writeback.h | 90 struct bdi_writeback *wb; /* wb this writeback is issued under */ member 266 if (wbc->wb) in wbc_init_bio() 267 bio_associate_blkcg(bio, wbc->wb->blkcg_css); in wbc_init_bio() 363 unsigned long wb_calc_thresh(struct bdi_writeback *wb, unsigned long thresh); 365 void wb_update_bandwidth(struct bdi_writeback *wb, unsigned long start_time); 368 bool wb_over_bg_thresh(struct bdi_writeback *wb);
|
D | memcontrol.h | 683 struct wb_domain *mem_cgroup_wb_domain(struct bdi_writeback *wb); 684 void mem_cgroup_wb_stats(struct bdi_writeback *wb, unsigned long *pfilepages, 690 static inline struct wb_domain *mem_cgroup_wb_domain(struct bdi_writeback *wb) in mem_cgroup_wb_domain() argument 695 static inline void mem_cgroup_wb_stats(struct bdi_writeback *wb, in mem_cgroup_wb_stats() argument
|
D | mm.h | 1273 struct mem_cgroup *memcg, struct bdi_writeback *wb);
|
/linux-4.4.14/drivers/staging/wilc1000/ |
D | wilc_spi.c | 273 u8 wb[32], rb[32]; in spi_cmd_complete() local 280 wb[0] = cmd; in spi_cmd_complete() 283 wb[1] = (u8)(adr >> 16); in spi_cmd_complete() 284 wb[2] = (u8)(adr >> 8); in spi_cmd_complete() 285 wb[3] = (u8)adr; in spi_cmd_complete() 290 wb[1] = (u8)(adr >> 8); in spi_cmd_complete() 292 wb[1] |= BIT(7); in spi_cmd_complete() 293 wb[2] = (u8)adr; in spi_cmd_complete() 294 wb[3] = 0x00; in spi_cmd_complete() 299 wb[1] = 0x00; in spi_cmd_complete() [all …]
|
D | linux_wlan_spi.h | 12 int linux_spi_write_read(u8 *wb, u8 *rb, u32 rlen);
|
D | linux_wlan_spi.c | 366 int linux_spi_write_read(u8 *wb, u8 *rb, u32 rlen) in linux_spi_write_read() argument 375 .tx_buf = wb, in linux_spi_write_read()
|
/linux-4.4.14/include/trace/events/ |
D | writeback.h | 137 static inline size_t __trace_wb_cgroup_size(struct bdi_writeback *wb) in __trace_wb_cgroup_size() argument 139 return kernfs_path_len(wb->memcg_css->cgroup->kn) + 1; in __trace_wb_cgroup_size() 142 static inline void __trace_wb_assign_cgroup(char *buf, struct bdi_writeback *wb) in __trace_wb_assign_cgroup() argument 144 struct cgroup *cgrp = wb->memcg_css->cgroup; in __trace_wb_assign_cgroup() 153 if (wbc->wb) in __trace_wbc_cgroup_size() 154 return __trace_wb_cgroup_size(wbc->wb); in __trace_wbc_cgroup_size() 162 if (wbc->wb) in __trace_wbc_assign_cgroup() 163 __trace_wb_assign_cgroup(buf, wbc->wb); in __trace_wbc_assign_cgroup() 170 static inline size_t __trace_wb_cgroup_size(struct bdi_writeback *wb) in __trace_wb_cgroup_size() argument 175 static inline void __trace_wb_assign_cgroup(char *buf, struct bdi_writeback *wb) in __trace_wb_assign_cgroup() argument [all …]
|
/linux-4.4.14/arch/xtensa/kernel/ |
D | ptrace.c | 56 unsigned long wb = regs->windowbase; in ptrace_getregs() local 73 gregset->a + ((wb * 4 + i) % XCHAL_NUM_AREGS)); in ptrace_getregs() 84 unsigned long wb, ws; in ptrace_setregs() local 95 __get_user(wb, &gregset->windowbase); in ptrace_setregs() 100 if (wb >= XCHAL_NUM_AREGS / 4) in ptrace_setregs() 103 if (wb != regs->windowbase || ws != regs->windowstart) { in ptrace_setregs() 106 rotws = (((ws | (ws << WSBITS)) >> wb) & in ptrace_setregs() 110 regs->windowbase = wb; in ptrace_setregs() 115 if (wb != 0 && __copy_from_user(regs->areg + XCHAL_NUM_AREGS - wb * 4, in ptrace_setregs() 116 gregset->a, wb * 16)) in ptrace_setregs() [all …]
|
D | process.c | 323 unsigned long wb, ws, wm; in xtensa_elf_core_copy_regs() local 326 wb = regs->windowbase; in xtensa_elf_core_copy_regs() 329 ws = ((ws >> wb) | (ws << (WSBITS - wb))) & ((1 << WSBITS) - 1); in xtensa_elf_core_copy_regs()
|
D | signal.c | 58 const unsigned long wb = regs->windowbase; in flush_window_regs_user() local 71 wm = (ws >> wb) | (ws << (XCHAL_NUM_AREGS / 4 - wb)); in flush_window_regs_user() 118 regs->windowstart = 1 << wb; in flush_window_regs_user()
|
/linux-4.4.14/Documentation/devicetree/bindings/interrupt-controller/ |
D | qca,ath79-cpu-intc.txt | 5 qca,ddr-wb-channels and qca,ddr-wb-channel-interrupts properties. 20 - qca,ddr-wb-channel-interrupts: List of the interrupts needing a write 22 - qca,ddr-wb-channels: List of phandles to the write buffer channels for 23 each interrupt. If qca,ddr-wb-channel-interrupts is not present the interrupt 34 qca,ddr-wb-channel-interrupts = <2>, <3>, <4>, <5>; 35 qca,ddr-wb-channels = <&ddr_ctrl 3>, <&ddr_ctrl 2>, 43 #qca,ddr-wb-channel-cells = <1>;
|
/linux-4.4.14/scripts/ |
D | extract-cert.c | 72 static BIO *wb; variable 80 if (!wb) { in write_cert() 81 wb = BIO_new_file(cert_dst, "wb"); in write_cert() 82 ERR(!wb, "%s", cert_dst); in write_cert() 85 ERR(!i2d_X509_bio(wb, x509), "%s", cert_dst); in write_cert() 146 if (wb && !x509) { in main() 159 BIO_free(wb); in main()
|
D | kallsyms.c | 641 int wa, wb; in compare_symbols() local 654 wb = (sb->sym[0] == 'w') || (sb->sym[0] == 'W'); in compare_symbols() 655 if (wa != wb) in compare_symbols() 656 return wa - wb; in compare_symbols() 660 wb = may_be_linker_script_provide_symbol(sb); in compare_symbols() 661 if (wa != wb) in compare_symbols() 662 return wa - wb; in compare_symbols() 666 wb = prefix_underscores_count((const char *)sb->sym + 1); in compare_symbols() 667 if (wa != wb) in compare_symbols() 668 return wa - wb; in compare_symbols()
|
/linux-4.4.14/drivers/gpu/drm/radeon/ |
D | r600_dma.c | 56 if (rdev->wb.enabled) in r600_dma_get_rptr() 57 rptr = rdev->wb.wb[ring->rptr_offs/4]; in r600_dma_get_rptr() 144 upper_32_bits(rdev->wb.gpu_addr + R600_WB_DMA_RPTR_OFFSET) & 0xFF); in r600_dma_resume() 146 ((rdev->wb.gpu_addr + R600_WB_DMA_RPTR_OFFSET) & 0xFFFFFFFC)); in r600_dma_resume() 148 if (rdev->wb.enabled) in r600_dma_resume() 244 gpu_addr = rdev->wb.gpu_addr + index; in r600_dma_ring_test() 247 rdev->wb.wb[index/4] = cpu_to_le32(tmp); in r600_dma_ring_test() 261 tmp = le32_to_cpu(rdev->wb.wb[index/4]); in r600_dma_ring_test() 351 gpu_addr = rdev->wb.gpu_addr + index; in r600_dma_ib_test() 377 tmp = le32_to_cpu(rdev->wb.wb[index/4]); in r600_dma_ib_test() [all …]
|
D | radeon_device.c | 427 rdev->wb.enabled = false; in radeon_wb_disable() 441 if (rdev->wb.wb_obj) { in radeon_wb_fini() 442 if (!radeon_bo_reserve(rdev->wb.wb_obj, false)) { in radeon_wb_fini() 443 radeon_bo_kunmap(rdev->wb.wb_obj); in radeon_wb_fini() 444 radeon_bo_unpin(rdev->wb.wb_obj); in radeon_wb_fini() 445 radeon_bo_unreserve(rdev->wb.wb_obj); in radeon_wb_fini() 447 radeon_bo_unref(&rdev->wb.wb_obj); in radeon_wb_fini() 448 rdev->wb.wb = NULL; in radeon_wb_fini() 449 rdev->wb.wb_obj = NULL; in radeon_wb_fini() 466 if (rdev->wb.wb_obj == NULL) { in radeon_wb_init() [all …]
|
D | cik_sdma.c | 68 if (rdev->wb.enabled) { in cik_sdma_get_rptr() 69 rptr = rdev->wb.wb[ring->rptr_offs/4]; in cik_sdma_get_rptr() 139 if (rdev->wb.enabled) { in cik_sdma_ring_ib_execute() 401 upper_32_bits(rdev->wb.gpu_addr + wb_offset) & 0xFFFFFFFF); in cik_sdma_gfx_resume() 403 ((rdev->wb.gpu_addr + wb_offset) & 0xFFFFFFFC)); in cik_sdma_gfx_resume() 405 if (rdev->wb.enabled) in cik_sdma_gfx_resume() 659 gpu_addr = rdev->wb.gpu_addr + index; in cik_sdma_ring_test() 662 rdev->wb.wb[index/4] = cpu_to_le32(tmp); in cik_sdma_ring_test() 677 tmp = le32_to_cpu(rdev->wb.wb[index/4]); in cik_sdma_ring_test() 716 gpu_addr = rdev->wb.gpu_addr + index; in cik_sdma_ib_test() [all …]
|
D | ni_dma.c | 58 if (rdev->wb.enabled) { in cayman_dma_get_rptr() 59 rptr = rdev->wb.wb[ring->rptr_offs/4]; in cayman_dma_get_rptr() 128 if (rdev->wb.enabled) { in cayman_dma_ring_ib_execute() 223 upper_32_bits(rdev->wb.gpu_addr + wb_offset) & 0xFF); in cayman_dma_resume() 225 ((rdev->wb.gpu_addr + wb_offset) & 0xFFFFFFFC)); in cayman_dma_resume() 227 if (rdev->wb.enabled) in cayman_dma_resume()
|
D | radeon_fence.c | 65 if (likely(rdev->wb.enabled || !drv->scratch_reg)) { in radeon_fence_write() 88 if (likely(rdev->wb.enabled || !drv->scratch_reg)) { in radeon_fence_read() 807 if (rdev->wb.use_event || !radeon_ring_supports_scratch_reg(rdev, &rdev->ring[ring])) { in radeon_fence_driver_start_ring() 811 rdev->fence_drv[ring].cpu_addr = &rdev->wb.wb[index/4]; in radeon_fence_driver_start_ring() 812 rdev->fence_drv[ring].gpu_addr = rdev->wb.gpu_addr + in radeon_fence_driver_start_ring() 831 rdev->fence_drv[ring].cpu_addr = &rdev->wb.wb[index/4]; in radeon_fence_driver_start_ring() 832 rdev->fence_drv[ring].gpu_addr = rdev->wb.gpu_addr + index; in radeon_fence_driver_start_ring()
|
D | radeon_ring.c | 300 else if (rdev->wb.enabled) in radeon_ring_backup() 412 if (rdev->wb.enabled) { in radeon_ring_init() 414 ring->next_rptr_gpu_addr = rdev->wb.gpu_addr + index; in radeon_ring_init() 415 ring->next_rptr_cpu_addr = &rdev->wb.wb[index/4]; in radeon_ring_init()
|
D | evergreen_dma.c | 72 if (rdev->wb.enabled) { in evergreen_dma_ring_ib_execute()
|
D | cik.c | 4161 } else if (rdev->wb.enabled) { in cik_ring_ib_execute() 4479 WREG32(SCRATCH_ADDR, ((rdev->wb.gpu_addr + RADEON_WB_SCRATCH_OFFSET) >> 8) & 0xFFFFFFFF); in cik_cp_gfx_resume() 4497 WREG32(CP_RB0_RPTR_ADDR, (rdev->wb.gpu_addr + RADEON_WB_CP_RPTR_OFFSET) & 0xFFFFFFFC); in cik_cp_gfx_resume() 4498 WREG32(CP_RB0_RPTR_ADDR_HI, upper_32_bits(rdev->wb.gpu_addr + RADEON_WB_CP_RPTR_OFFSET) & 0xFF); in cik_cp_gfx_resume() 4503 if (!rdev->wb.enabled) in cik_cp_gfx_resume() 4533 if (rdev->wb.enabled) in cik_gfx_get_rptr() 4534 rptr = rdev->wb.wb[ring->rptr_offs/4]; in cik_gfx_get_rptr() 4563 if (rdev->wb.enabled) { in cik_compute_get_rptr() 4564 rptr = rdev->wb.wb[ring->rptr_offs/4]; in cik_compute_get_rptr() 4581 if (rdev->wb.enabled) { in cik_compute_get_wptr() [all …]
|
D | r600.c | 2616 if (rdev->wb.enabled) in r600_gfx_get_rptr() 2617 rptr = rdev->wb.wb[ring->rptr_offs/4]; in r600_gfx_get_rptr() 2746 ((rdev->wb.gpu_addr + RADEON_WB_CP_RPTR_OFFSET) & 0xFFFFFFFC)); in r600_cp_resume() 2747 WREG32(CP_RB_RPTR_ADDR_HI, upper_32_bits(rdev->wb.gpu_addr + RADEON_WB_CP_RPTR_OFFSET) & 0xFF); in r600_cp_resume() 2748 WREG32(SCRATCH_ADDR, ((rdev->wb.gpu_addr + RADEON_WB_SCRATCH_OFFSET) >> 8) & 0xFFFFFFFF); in r600_cp_resume() 2750 if (rdev->wb.enabled) in r600_cp_resume() 2874 if (rdev->wb.use_event) { in r600_fence_ring_emit() 3337 } else if (rdev->wb.enabled) { in r600_ring_ib_execute() 3668 if (rdev->wb.enabled) in r600_irq_init() 3672 WREG32(IH_RB_WPTR_ADDR_LO, (rdev->wb.gpu_addr + R600_WB_IH_WPTR_OFFSET) & 0xFFFFFFFC); in r600_irq_init() [all …]
|
D | si.c | 3421 } else if (rdev->wb.enabled) { in si_ring_ib_execute() 3661 WREG32(SCRATCH_ADDR, ((rdev->wb.gpu_addr + RADEON_WB_SCRATCH_OFFSET) >> 8) & 0xFFFFFFFF); in si_cp_resume() 3679 WREG32(CP_RB0_RPTR_ADDR, (rdev->wb.gpu_addr + RADEON_WB_CP_RPTR_OFFSET) & 0xFFFFFFFC); in si_cp_resume() 3680 WREG32(CP_RB0_RPTR_ADDR_HI, upper_32_bits(rdev->wb.gpu_addr + RADEON_WB_CP_RPTR_OFFSET) & 0xFF); in si_cp_resume() 3682 if (rdev->wb.enabled) in si_cp_resume() 3710 WREG32(CP_RB1_RPTR_ADDR, (rdev->wb.gpu_addr + RADEON_WB_CP1_RPTR_OFFSET) & 0xFFFFFFFC); in si_cp_resume() 3711 WREG32(CP_RB1_RPTR_ADDR_HI, upper_32_bits(rdev->wb.gpu_addr + RADEON_WB_CP1_RPTR_OFFSET) & 0xFF); in si_cp_resume() 3734 WREG32(CP_RB2_RPTR_ADDR, (rdev->wb.gpu_addr + RADEON_WB_CP2_RPTR_OFFSET) & 0xFFFFFFFC); in si_cp_resume() 3735 WREG32(CP_RB2_RPTR_ADDR_HI, upper_32_bits(rdev->wb.gpu_addr + RADEON_WB_CP2_RPTR_OFFSET) & 0xFF); in si_cp_resume() 6034 if (rdev->wb.enabled) in si_irq_init() [all …]
|
D | ni.c | 1485 if (rdev->wb.enabled) in cayman_gfx_get_rptr() 1486 rptr = rdev->wb.wb[ring->rptr_offs/4]; in cayman_gfx_get_rptr() 1691 WREG32(SCRATCH_ADDR, ((rdev->wb.gpu_addr + RADEON_WB_SCRATCH_OFFSET) >> 8) & 0xFFFFFFFF); in cayman_cp_resume() 1708 addr = rdev->wb.gpu_addr + RADEON_WB_CP_RPTR_OFFSET; in cayman_cp_resume()
|
D | evergreen.c | 3034 } else if (rdev->wb.enabled) { in evergreen_ring_ib_execute() 3192 ((rdev->wb.gpu_addr + RADEON_WB_CP_RPTR_OFFSET) & 0xFFFFFFFC)); in evergreen_cp_resume() 3193 WREG32(CP_RB_RPTR_ADDR_HI, upper_32_bits(rdev->wb.gpu_addr + RADEON_WB_CP_RPTR_OFFSET) & 0xFF); in evergreen_cp_resume() 3194 WREG32(SCRATCH_ADDR, ((rdev->wb.gpu_addr + RADEON_WB_SCRATCH_OFFSET) >> 8) & 0xFFFFFFFF); in evergreen_cp_resume() 3196 if (rdev->wb.enabled) in evergreen_cp_resume() 5014 if (rdev->wb.enabled) in evergreen_get_ih_wptr() 5015 wptr = le32_to_cpu(rdev->wb.wb[R600_WB_IH_WPTR_OFFSET/4]); in evergreen_get_ih_wptr()
|
D | r100.c | 1061 if (rdev->wb.enabled) in r100_gfx_get_rptr() 1062 rptr = le32_to_cpu(rdev->wb.wb[ring->rptr_offs/4]); in r100_gfx_get_rptr() 1187 S_00070C_RB_RPTR_ADDR((rdev->wb.gpu_addr + RADEON_WB_CP_RPTR_OFFSET) >> 2)); in r100_cp_init() 1188 WREG32(R_000774_SCRATCH_ADDR, rdev->wb.gpu_addr + RADEON_WB_SCRATCH_OFFSET); in r100_cp_init() 1190 if (rdev->wb.enabled) in r100_cp_init()
|
D | radeon.h | 1143 volatile uint32_t *wb; member 2392 struct radeon_wb wb; member
|
/linux-4.4.14/drivers/usb/class/ |
D | cdc-acm.c | 169 struct acm_wb *wb; in acm_wb_alloc() local 174 wb = &acm->wb[wbn]; in acm_wb_alloc() 175 if (!wb->use) { in acm_wb_alloc() 176 wb->use = 1; in acm_wb_alloc() 193 n -= acm->wb[i].use; in acm_wb_is_avail() 201 static void acm_write_done(struct acm *acm, struct acm_wb *wb) in acm_write_done() argument 203 wb->use = 0; in acm_write_done() 214 static int acm_start_wb(struct acm *acm, struct acm_wb *wb) in acm_start_wb() argument 220 wb->urb->transfer_buffer = wb->buf; in acm_start_wb() 221 wb->urb->transfer_dma = wb->dmah; in acm_start_wb() [all …]
|
D | cdc-acm.h | 93 struct acm_wb wb[ACM_NW]; member
|
/linux-4.4.14/drivers/gpu/drm/amd/amdgpu/ |
D | amdgpu_device.c | 426 if (adev->wb.wb_obj) { in amdgpu_wb_fini() 427 if (!amdgpu_bo_reserve(adev->wb.wb_obj, false)) { in amdgpu_wb_fini() 428 amdgpu_bo_kunmap(adev->wb.wb_obj); in amdgpu_wb_fini() 429 amdgpu_bo_unpin(adev->wb.wb_obj); in amdgpu_wb_fini() 430 amdgpu_bo_unreserve(adev->wb.wb_obj); in amdgpu_wb_fini() 432 amdgpu_bo_unref(&adev->wb.wb_obj); in amdgpu_wb_fini() 433 adev->wb.wb = NULL; in amdgpu_wb_fini() 434 adev->wb.wb_obj = NULL; in amdgpu_wb_fini() 451 if (adev->wb.wb_obj == NULL) { in amdgpu_wb_init() 454 &adev->wb.wb_obj); in amdgpu_wb_init() [all …]
|
D | sdma_v3_0.c | 283 rptr = ring->adev->wb.wb[ring->rptr_offs] >> 2; in sdma_v3_0_ring_get_rptr() 302 wptr = ring->adev->wb.wb[ring->wptr_offs] >> 2; in sdma_v3_0_ring_get_wptr() 325 adev->wb.wb[ring->wptr_offs] = ring->wptr << 2; in sdma_v3_0_ring_set_wptr() 618 upper_32_bits(adev->wb.gpu_addr + wb_offset) & 0xFFFFFFFF); in sdma_v3_0_gfx_resume() 620 lower_32_bits(adev->wb.gpu_addr + wb_offset) & 0xFFFFFFFC); in sdma_v3_0_gfx_resume() 785 gpu_addr = adev->wb.gpu_addr + (index * 4); in sdma_v3_0_ring_test_ring() 787 adev->wb.wb[index] = cpu_to_le32(tmp); in sdma_v3_0_ring_test_ring() 805 tmp = le32_to_cpu(adev->wb.wb[index]); in sdma_v3_0_ring_test_ring() 848 gpu_addr = adev->wb.gpu_addr + (index * 4); in sdma_v3_0_ring_test_ib() 850 adev->wb.wb[index] = cpu_to_le32(tmp); in sdma_v3_0_ring_test_ib() [all …]
|
D | cik_sdma.c | 156 rptr = ring->adev->wb.wb[ring->rptr_offs]; in cik_sdma_ring_get_rptr() 438 upper_32_bits(adev->wb.gpu_addr + wb_offset) & 0xFFFFFFFF); in cik_sdma_gfx_resume() 440 ((adev->wb.gpu_addr + wb_offset) & 0xFFFFFFFC)); in cik_sdma_gfx_resume() 583 gpu_addr = adev->wb.gpu_addr + (index * 4); in cik_sdma_ring_test_ring() 585 adev->wb.wb[index] = cpu_to_le32(tmp); in cik_sdma_ring_test_ring() 601 tmp = le32_to_cpu(adev->wb.wb[index]); in cik_sdma_ring_test_ring() 644 gpu_addr = adev->wb.gpu_addr + (index * 4); in cik_sdma_ring_test_ib() 646 adev->wb.wb[index] = cpu_to_le32(tmp); in cik_sdma_ring_test_ib() 672 tmp = le32_to_cpu(adev->wb.wb[index]); in cik_sdma_ring_test_ib()
|
D | sdma_v2_4.c | 187 rptr = ring->adev->wb.wb[ring->rptr_offs] >> 2; in sdma_v2_4_ring_get_rptr() 481 upper_32_bits(adev->wb.gpu_addr + wb_offset) & 0xFFFFFFFF); in sdma_v2_4_gfx_resume() 483 lower_32_bits(adev->wb.gpu_addr + wb_offset) & 0xFFFFFFFC); in sdma_v2_4_gfx_resume() 635 gpu_addr = adev->wb.gpu_addr + (index * 4); in sdma_v2_4_ring_test_ring() 637 adev->wb.wb[index] = cpu_to_le32(tmp); in sdma_v2_4_ring_test_ring() 655 tmp = le32_to_cpu(adev->wb.wb[index]); in sdma_v2_4_ring_test_ring() 698 gpu_addr = adev->wb.gpu_addr + (index * 4); in sdma_v2_4_ring_test_ib() 700 adev->wb.wb[index] = cpu_to_le32(tmp); in sdma_v2_4_ring_test_ib() 731 tmp = le32_to_cpu(adev->wb.wb[index]); in sdma_v2_4_ring_test_ib()
|
D | tonga_ih.c | 143 wptr_off = adev->wb.gpu_addr + (adev->irq.ih.wptr_offs * 4); in tonga_ih_irq_init() 204 wptr = le32_to_cpu(adev->wb.wb[adev->irq.ih.wptr_offs]); in tonga_ih_get_wptr() 266 adev->wb.wb[adev->irq.ih.rptr_offs] = adev->irq.ih.rptr; in tonga_ih_set_rptr()
|
D | cik_ih.c | 135 wptr_off = adev->wb.gpu_addr + (adev->irq.ih.wptr_offs * 4); in cik_ih_irq_init() 191 wptr = le32_to_cpu(adev->wb.wb[adev->irq.ih.wptr_offs]); in cik_ih_get_wptr()
|
D | cz_ih.c | 137 wptr_off = adev->wb.gpu_addr + (adev->irq.ih.wptr_offs * 4); in cz_ih_irq_init() 193 wptr = le32_to_cpu(adev->wb.wb[adev->irq.ih.wptr_offs]); in cz_ih_get_wptr()
|
D | iceland_ih.c | 137 wptr_off = adev->wb.gpu_addr + (adev->irq.ih.wptr_offs * 4); in iceland_ih_irq_init() 193 wptr = le32_to_cpu(adev->wb.wb[adev->irq.ih.wptr_offs]); in iceland_ih_get_wptr()
|
D | amdgpu_ring.c | 346 ring->next_rptr_gpu_addr = adev->wb.gpu_addr + (ring->next_rptr_offs * 4); in amdgpu_ring_init() 347 ring->next_rptr_cpu_addr = &adev->wb.wb[ring->next_rptr_offs]; in amdgpu_ring_init()
|
D | amdgpu_fence.c | 442 ring->fence_drv.cpu_addr = &adev->wb.wb[ring->fence_offs]; in amdgpu_fence_driver_start_ring() 443 ring->fence_drv.gpu_addr = adev->wb.gpu_addr + (ring->fence_offs * 4); in amdgpu_fence_driver_start_ring()
|
D | gfx_v8_0.c | 3328 rptr_addr = adev->wb.gpu_addr + (ring->rptr_offs * 4); in gfx_v8_0_cp_gfx_resume() 3889 wb_gpu_addr = adev->wb.gpu_addr + (ring->rptr_offs * 4); in gfx_v8_0_cp_compute_resume() 3899 wb_gpu_addr = adev->wb.gpu_addr + (ring->wptr_offs * 4); in gfx_v8_0_cp_compute_resume() 4477 rptr = ring->adev->wb.wb[ring->rptr_offs]; in gfx_v8_0_ring_get_rptr_gfx() 4489 wptr = ring->adev->wb.wb[ring->wptr_offs]; in gfx_v8_0_ring_get_wptr_gfx() 4502 adev->wb.wb[ring->wptr_offs] = ring->wptr; in gfx_v8_0_ring_set_wptr_gfx() 4748 return ring->adev->wb.wb[ring->rptr_offs]; in gfx_v8_0_ring_get_rptr_compute() 4753 return ring->adev->wb.wb[ring->wptr_offs]; in gfx_v8_0_ring_get_wptr_compute() 4761 adev->wb.wb[ring->wptr_offs] = ring->wptr; in gfx_v8_0_ring_set_wptr_compute()
|
D | amdgpu.h | 1293 volatile uint32_t *wb; member 1299 int amdgpu_wb_get(struct amdgpu_device *adev, u32 *wb); 1300 void amdgpu_wb_free(struct amdgpu_device *adev, u32 wb); 2037 struct amdgpu_wb wb; member 2256 …amdgpu_ring_emit_gds_switch(r, v, db, ds, wb, ws, ab, as) (r)->funcs->emit_gds_switch((r), (v), (d… argument
|
D | gfx_v7_0.c | 2964 rptr_addr = adev->wb.gpu_addr + (ring->rptr_offs * 4); in gfx_v7_0_cp_gfx_resume() 2994 rptr = ring->adev->wb.wb[ring->rptr_offs]; in gfx_v7_0_ring_get_rptr_gfx() 3021 rptr = ring->adev->wb.wb[ring->rptr_offs]; in gfx_v7_0_ring_get_rptr_compute() 3031 wptr = ring->adev->wb.wb[ring->wptr_offs]; in gfx_v7_0_ring_get_wptr_compute() 3041 adev->wb.wb[ring->wptr_offs] = ring->wptr; in gfx_v7_0_ring_set_wptr_compute() 3490 wb_gpu_addr = adev->wb.gpu_addr + (ring->wptr_offs * 4); in gfx_v7_0_cp_compute_resume() 3498 wb_gpu_addr = adev->wb.gpu_addr + (ring->rptr_offs * 4); in gfx_v7_0_cp_compute_resume()
|
D | amdgpu_test.c | 49 if (adev->wb.wb_obj) in amdgpu_do_test_moves()
|
/linux-4.4.14/Documentation/devicetree/bindings/memory-controllers/ |
D | ath79-ddr-controller.txt | 15 - #qca,ddr-wb-channel-cells: has to be 1, the index of the write buffer 25 #qca,ddr-wb-channel-cells = <1>; 32 qca,ddr-wb-channel-interrupts = <2>, <3>, <4>, <5>; 33 qca,ddr-wb-channels = <&ddr_ctrl 3>, <&ddr_ctrl 2>,
|
/linux-4.4.14/arch/mips/boot/dts/qca/ |
D | ar9132.dtsi | 24 qca,ddr-wb-channel-interrupts = <2>, <3>, <4>, <5>; 25 qca,ddr-wb-channels = <&ddr_ctrl 3>, <&ddr_ctrl 2>, 52 #qca,ddr-wb-channel-cells = <1>;
|
/linux-4.4.14/drivers/staging/media/davinci_vpfe/ |
D | dm365_ipipe_hw.c | 612 void ipipe_set_wb_regs(void __iomem *base_addr, struct vpfe_ipipe_wb *wb) in ipipe_set_wb_regs() argument 618 regw_ip(base_addr, wb->ofst_r & WB_OFFSET_MASK, WB2_OFT_R); in ipipe_set_wb_regs() 619 regw_ip(base_addr, wb->ofst_gr & WB_OFFSET_MASK, WB2_OFT_GR); in ipipe_set_wb_regs() 620 regw_ip(base_addr, wb->ofst_gb & WB_OFFSET_MASK, WB2_OFT_GB); in ipipe_set_wb_regs() 621 regw_ip(base_addr, wb->ofst_b & WB_OFFSET_MASK, WB2_OFT_B); in ipipe_set_wb_regs() 624 val = IPIPE_U13Q9(wb->gain_r.decimal, wb->gain_r.integer); in ipipe_set_wb_regs() 627 val = IPIPE_U13Q9(wb->gain_gr.decimal, wb->gain_gr.integer); in ipipe_set_wb_regs() 630 val = IPIPE_U13Q9(wb->gain_gb.decimal, wb->gain_gb.integer); in ipipe_set_wb_regs() 633 val = IPIPE_U13Q9(wb->gain_b.decimal, wb->gain_b.integer); in ipipe_set_wb_regs()
|
D | dm365_ipipe_hw.h | 556 void ipipe_set_wb_regs(void __iomem *base_addr, struct vpfe_ipipe_wb *wb);
|
/linux-4.4.14/arch/arm/boot/dts/ |
D | exynos4x12.dtsi | 304 samsung,isp-wb; 312 samsung,isp-wb; 320 samsung,isp-wb; 321 samsung,lcd-wb; 330 samsung,isp-wb; 331 samsung,lcd-wb;
|
D | exynos4210.dtsi | 212 samsung,lcd-wb; 219 samsung,lcd-wb;
|
D | s5pv210.dtsi | 627 samsung,lcd-wb;
|
/linux-4.4.14/drivers/crypto/sunxi-ss/ |
D | sun4i-ss-hash.c | 352 u32 wb = 0; in sun4i_hash_final() local 380 wb = *(u32 *)(op->buf + nwait * 4); in sun4i_hash_final() 381 wb &= (0xFFFFFFFF >> (4 - nbw) * 8); in sun4i_hash_final() 386 wb |= ((1 << 7) << (nbw * 8)); in sun4i_hash_final() 387 bf[j++] = wb; in sun4i_hash_final()
|
/linux-4.4.14/drivers/media/i2c/m5mols/ |
D | m5mols_controls.c | 281 static const unsigned short wb[][2] = { in m5mols_set_white_balance() local 296 for (i = 0; i < ARRAY_SIZE(wb); i++) { in m5mols_set_white_balance() 298 if (wb[i][0] != val) in m5mols_set_white_balance() 302 "Setting white balance to: %#x\n", wb[i][0]); in m5mols_set_white_balance() 304 awb = wb[i][0] == V4L2_WHITE_BALANCE_AUTO; in m5mols_set_white_balance() 311 ret = m5mols_write(sd, AWB_MANUAL, wb[i][1]); in m5mols_set_white_balance()
|
/linux-4.4.14/drivers/net/ethernet/intel/igb/ |
D | e1000_82575.h | 107 } wb; /* writeback */ member 126 } wb; member
|
D | igb.h | 311 return rx_desc->wb.upper.status_error & cpu_to_le32(stat_err_bits); in igb_test_staterr()
|
D | igb_main.c | 523 staterr = le32_to_cpu(rx_desc->wb.upper.status_error); in igb_dump() 6428 if (!(eop_desc->wb.status & cpu_to_le32(E1000_TXD_STAT_DD))) in igb_clean_tx_irq() 6533 tx_buffer->next_to_watch->wb.status); in igb_clean_tx_irq() 6652 unsigned int size = le16_to_cpu(rx_desc->wb.upper.length); in igb_add_rx_frag() 6795 le32_to_cpu(rx_desc->wb.upper.status_error)); in igb_rx_checksum() 6804 le32_to_cpu(rx_desc->wb.lower.hi_dword.rss), in igb_rx_hash() 6900 vid = be16_to_cpu(rx_desc->wb.upper.vlan); in igb_process_skb_fields() 6902 vid = le16_to_cpu(rx_desc->wb.upper.vlan); in igb_process_skb_fields() 6930 if (!rx_desc->wb.upper.status_error) in igb_clean_rx_irq() 7063 rx_desc->wb.upper.status_error = 0; in igb_alloc_rx_buffers()
|
/linux-4.4.14/drivers/net/ethernet/intel/ixgbevf/ |
D | defines.h | 212 } wb; member 243 } wb; /* writeback */ member
|
D | ixgbevf.h | 344 return rx_desc->wb.upper.status_error & cpu_to_le32(stat_err_bits); in ixgbevf_test_staterr()
|
D | ixgbevf_main.c | 318 if (!(eop_desc->wb.status & cpu_to_le32(IXGBE_TXD_STAT_DD))) in ixgbevf_clean_tx_irq() 408 eop_desc, (eop_desc ? eop_desc->wb.status : 0), in ixgbevf_clean_tx_irq() 475 rss_type = le16_to_cpu(rx_desc->wb.lower.lo_dword.hs_rss.pkt_info) & in ixgbevf_rx_hash() 481 skb_set_hash(skb, le32_to_cpu(rx_desc->wb.lower.hi_dword.rss), in ixgbevf_rx_hash() 539 u16 vid = le16_to_cpu(rx_desc->wb.upper.vlan); in ixgbevf_process_skb_fields() 776 unsigned int size = le16_to_cpu(rx_desc->wb.upper.length); in ixgbevf_add_rx_frag()
|
/linux-4.4.14/drivers/char/ |
D | ps3flash.c | 270 int wb; in ps3flash_kernel_write() local 277 wb = ps3flash_writeback(ps3flash_dev); in ps3flash_kernel_write() 278 if (wb) in ps3flash_kernel_write() 279 return wb; in ps3flash_kernel_write()
|
/linux-4.4.14/drivers/net/ethernet/intel/igbvf/ |
D | vf.h | 87 } wb; /* writeback */ member 104 } wb; member
|
D | netdev.c | 281 staterr = le32_to_cpu(rx_desc->wb.upper.status_error); in igbvf_clean_rx_irq() 296 hlen = (le16_to_cpu(rx_desc->wb.lower.lo_dword.hs_rss.hdr_info) in igbvf_clean_rx_irq() 302 length = le16_to_cpu(rx_desc->wb.upper.length); in igbvf_clean_rx_irq() 376 rx_desc->wb.upper.vlan); in igbvf_clean_rx_irq() 379 rx_desc->wb.upper.status_error = 0; in igbvf_clean_rx_irq() 391 staterr = le32_to_cpu(rx_desc->wb.upper.status_error); in igbvf_clean_rx_irq() 815 if (!(eop_desc->wb.status & cpu_to_le32(E1000_TXD_STAT_DD))) in igbvf_clean_tx_irq() 839 tx_desc->wb.status = 0; in igbvf_clean_tx_irq()
|
/linux-4.4.14/drivers/media/i2c/s5c73m3/ |
D | s5c73m3-ctrls.c | 152 static const unsigned short wb[][2] = { in s5c73m3_set_white_balance() local 162 for (i = 0; i < ARRAY_SIZE(wb); i++) { in s5c73m3_set_white_balance() 163 if (wb[i][0] != val) in s5c73m3_set_white_balance() 170 return s5c73m3_isp_command(state, COMM_AWB_MODE, wb[i][1]); in s5c73m3_set_white_balance()
|
/linux-4.4.14/arch/alpha/kernel/ |
D | smc37c669.c | 944 #define wb( _x_, _y_ ) outb( _y_, (unsigned int)((unsigned long)_x_) ) macro 1975 wb( &SMC37c669->index_port, SMC37c669_CONFIG_ON_KEY ); in SMC37c669_config_mode() 1976 wb( &SMC37c669->index_port, SMC37c669_CONFIG_ON_KEY ); in SMC37c669_config_mode() 1980 wb( &SMC37c669->index_port, SMC37c669_CONFIG_OFF_KEY ); in SMC37c669_config_mode() 2012 wb( &SMC37c669->index_port, index ); in SMC37c669_read_config() 2047 wb( &SMC37c669->index_port, index ); in SMC37c669_write_config() 2048 wb( &SMC37c669->data_port, data ); in SMC37c669_write_config()
|
/linux-4.4.14/drivers/mtd/nand/ |
D | nandsim.c | 849 struct weak_block *wb; in parse_weakblocks() local 868 wb = kzalloc(sizeof(*wb), GFP_KERNEL); in parse_weakblocks() 869 if (!wb) { in parse_weakblocks() 873 wb->erase_block_no = erase_block_no; in parse_weakblocks() 874 wb->max_erases = max_erases; in parse_weakblocks() 875 list_add(&wb->list, &weak_blocks); in parse_weakblocks() 882 struct weak_block *wb; in erase_error() local 884 list_for_each_entry(wb, &weak_blocks, list) in erase_error() 885 if (wb->erase_block_no == erase_block_no) { in erase_error() 886 if (wb->erases_done >= wb->max_erases) in erase_error() [all …]
|
/linux-4.4.14/drivers/net/ethernet/intel/i40e/ |
D | i40e_txrx.c | 463 qw = le64_to_cpu(rx_desc->wb.qword1.status_error_len); in i40e_fd_handle_status() 468 pf->fd_inv = le32_to_cpu(rx_desc->wb.qword0.hi_dword.fd_id); in i40e_fd_handle_status() 469 if ((rx_desc->wb.qword0.hi_dword.fd_id != 0) || in i40e_fd_handle_status() 487 if ((rx_desc->wb.qword0.hi_dword.fd_id == 0) && in i40e_fd_handle_status() 516 rx_desc->wb.qword0.hi_dword.fd_id); in i40e_fd_handle_status() 930 qw = le64_to_cpu(rx_desc->wb.qword1.status_error_len); in i40e_clean_programming_status() 1431 (rx_desc->wb.qword1.status_error_len & rss_mask) == rss_mask) in i40e_rx_hash() 1432 return le32_to_cpu(rx_desc->wb.qword0.hi_dword.rss); in i40e_rx_hash() 1495 qword = le64_to_cpu(rx_desc->wb.qword1.status_error_len); in i40e_clean_rx_irq_ps() 1627 ? le16_to_cpu(rx_desc->wb.qword0.lo_dword.l2tag1) in i40e_clean_rx_irq_ps() [all …]
|
D | i40e_fcoe.c | 633 xid = le32_to_cpu(rx_desc->wb.qword0.hi_dword.fcoe_param) & in i40e_fcoe_handle_status() 642 qw = le64_to_cpu(rx_desc->wb.qword1.status_error_len); in i40e_fcoe_handle_status() 705 qw = le64_to_cpu(rx_desc->wb.qword1.status_error_len); in i40e_fcoe_handle_offload() 744 xid = le16_to_cpu(rx_desc->wb.qword0.lo_dword.mirr_fcoe.fcoe_ctx_id); in i40e_fcoe_handle_offload() 759 ddp->len = le32_to_cpu(rx_desc->wb.qword0.hi_dword.fcoe_param); in i40e_fcoe_handle_offload()
|
D | i40e_type.h | 594 } wb; /* writeback */ member 643 } wb; /* writeback */ member
|
/linux-4.4.14/drivers/net/ethernet/intel/i40evf/ |
D | i40e_txrx.c | 904 (rx_desc->wb.qword1.status_error_len & rss_mask) == rss_mask) in i40e_rx_hash() 905 return le32_to_cpu(rx_desc->wb.qword0.hi_dword.rss); in i40e_rx_hash() 965 qword = le64_to_cpu(rx_desc->wb.qword1.status_error_len); in i40e_clean_rx_irq_ps() 1085 ? le16_to_cpu(rx_desc->wb.qword0.lo_dword.l2tag1) in i40e_clean_rx_irq_ps() 1096 rx_desc->wb.qword1.status_error_len = 0; in i40e_clean_rx_irq_ps() 1141 qword = le64_to_cpu(rx_desc->wb.qword1.status_error_len); in i40e_clean_rx_irq_1buf() 1203 ? le16_to_cpu(rx_desc->wb.qword0.lo_dword.l2tag1) in i40e_clean_rx_irq_1buf() 1207 rx_desc->wb.qword1.status_error_len = 0; in i40e_clean_rx_irq_1buf()
|
D | i40e_type.h | 581 } wb; /* writeback */ member 630 } wb; /* writeback */ member
|
/linux-4.4.14/drivers/net/ethernet/broadcom/bnx2x/ |
D | bnx2x_init_ops.h | 70 u8 wb) in bnx2x_write_big_buf() argument 76 else if (wb && CHIP_IS_E1(bp)) in bnx2x_write_big_buf() 85 u32 len, u8 wb) in bnx2x_init_fill() argument 96 bnx2x_write_big_buf(bp, addr + i*4, cur_len, wb); in bnx2x_init_fill()
|
/linux-4.4.14/drivers/net/ethernet/intel/e1000e/ |
D | hw.h | 229 } wb; /* writeback */ member 265 } wb; /* writeback */ member
|
D | netdev.c | 383 le32_to_cpu(rx_desc_ps->wb.middle.status_error); in e1000e_dump() 449 staterr = le32_to_cpu(rx_desc->wb.upper.status_error); in e1000e_dump() 941 staterr = le32_to_cpu(rx_desc->wb.upper.status_error); in e1000_clean_rx_irq() 971 length = le16_to_cpu(rx_desc->wb.upper.length); in e1000_clean_rx_irq() 1040 e1000_rx_hash(netdev, rx_desc->wb.lower.hi_dword.rss, skb); in e1000_clean_rx_irq() 1043 rx_desc->wb.upper.vlan); in e1000_clean_rx_irq() 1046 rx_desc->wb.upper.status_error &= cpu_to_le32(~0xFF); in e1000_clean_rx_irq() 1059 staterr = le32_to_cpu(rx_desc->wb.upper.status_error); in e1000_clean_rx_irq() 1327 staterr = le32_to_cpu(rx_desc->wb.middle.status_error); in e1000_clean_rx_irq_ps() 1372 length = le16_to_cpu(rx_desc->wb.middle.length0); in e1000_clean_rx_irq_ps() [all …]
|
/linux-4.4.14/Documentation/filesystems/ |
D | affs.txt | 192 do care about the case. Example (with /wb being an affs mounted fs): 193 rm /wb/WRONGCASE 195 rm /wb/WR*
|
/linux-4.4.14/arch/unicore32/boot/compressed/ |
D | head.S | 92 mov r0, #0x1c @ en icache and wb dcache
|
/linux-4.4.14/Documentation/devicetree/bindings/media/ |
D | samsung-fimc.txt | 87 - samsung,isp-wb: this property must be present if the IP block has the ISP 89 - samsung,lcd-wb: this property must be present if the IP block has the LCD
|
/linux-4.4.14/drivers/media/dvb-frontends/ |
D | dib3000mb.c | 55 u8 wb[] = { ((reg >> 8) | 0x80) & 0xff, reg & 0xff }; in dib3000_read_reg() local 58 { .addr = state->config.demod_address, .flags = 0, .buf = wb, .len = 2 }, in dib3000_read_reg()
|
D | dib3000mc.c | 52 u8 wb[2] = { (reg >> 8) | 0x80, reg & 0xff }; in dib3000mc_read_word() local 55 { .addr = state->i2c_addr >> 1, .flags = 0, .buf = wb, .len = 2 }, in dib3000mc_read_word()
|
/linux-4.4.14/drivers/media/usb/dvb-usb/ |
D | dib0700_devices.c | 2206 u8 wb[4] = { 0xc >> 8, 0xc & 0xff, 0, 0 }; in dib01x0_pmu_update() local 2209 {.addr = 0x1e >> 1, .flags = 0, .buf = wb, .len = 2}, in dib01x0_pmu_update() 2235 wb[2] = (data[index_data + 1] >> 8) & 0xff; in dib01x0_pmu_update() 2236 wb[3] = (data[index_data + 1]) & 0xff; in dib01x0_pmu_update() 2239 wb[0] = (data[index_data] >> 8) & 0xff; in dib01x0_pmu_update() 2240 wb[1] = (data[index_data]) & 0xff; in dib01x0_pmu_update() 2244 wb[2] |= rb[0]; in dib01x0_pmu_update() 2245 wb[3] |= rb[1] & ~(3 << 4); in dib01x0_pmu_update() 2248 wb[0] = (data[index_data] >> 8)&0xff; in dib01x0_pmu_update() 2249 wb[1] = (data[index_data])&0xff; in dib01x0_pmu_update()
|
/linux-4.4.14/drivers/gpu/drm/msm/mdp/mdp5/ |
D | mdp5.xml.h | 1393 case 0: return (mdp5_cfg->wb.base[0]); in __offset_WB() 1394 case 1: return (mdp5_cfg->wb.base[1]); in __offset_WB() 1395 case 2: return (mdp5_cfg->wb.base[2]); in __offset_WB() 1396 case 3: return (mdp5_cfg->wb.base[3]); in __offset_WB() 1397 case 4: return (mdp5_cfg->wb.base[4]); in __offset_WB()
|
/linux-4.4.14/drivers/net/ethernet/intel/ixgbe/ |
D | ixgbe_fcoe.c | 460 ddp->len = le32_to_cpu(rx_desc->wb.lower.hi_dword.rss); in ixgbe_fcoe_ddp() 474 ddp->len = le32_to_cpu(rx_desc->wb.lower.hi_dword.rss); in ixgbe_fcoe_ddp()
|
D | ixgbe_main.c | 785 staterr = le32_to_cpu(rx_desc->wb.upper.status_error); in ixgbe_dump() 1120 if (!(eop_desc->wb.status & cpu_to_le32(IXGBE_TXD_STAT_DD))) in ixgbe_clean_tx_irq() 1400 rss_type = le16_to_cpu(rx_desc->wb.lower.lo_dword.hs_rss.pkt_info) & in ixgbe_rx_hash() 1406 skb_set_hash(skb, le32_to_cpu(rx_desc->wb.lower.hi_dword.rss), in ixgbe_rx_hash() 1422 __le16 pkt_info = rx_desc->wb.lower.lo_dword.hs_rss.pkt_info; in ixgbe_rx_is_fcoe() 1441 __le16 pkt_info = rx_desc->wb.lower.lo_dword.hs_rss.pkt_info; in ixgbe_rx_checksum() 1442 __le16 hdr_info = rx_desc->wb.lower.lo_dword.hs_rss.hdr_info; in ixgbe_rx_checksum() 1573 rx_desc->wb.upper.status_error = 0; in ixgbe_alloc_rx_buffers() 1650 u16 vid = le16_to_cpu(rx_desc->wb.upper.vlan); in ixgbe_process_skb_fields() 1693 __le32 rsc_enabled = rx_desc->wb.lower.lo_dword.data & in ixgbe_is_non_eop() [all …]
|
D | ixgbe.h | 556 return rx_desc->wb.upper.status_error & cpu_to_le32(stat_err_bits); in ixgbe_test_staterr()
|
D | ixgbe_type.h | 2690 } wb; member 2721 } wb; /* writeback */ member
|
/linux-4.4.14/drivers/gpu/drm/exynos/ |
D | exynos_drm_fimc.c | 228 static void fimc_set_type_ctrl(struct fimc_context *ctx, enum fimc_wb wb) in fimc_set_type_ctrl() argument 232 DRM_DEBUG_KMS("wb[%d]\n", wb); in fimc_set_type_ctrl() 242 switch (wb) { in fimc_set_type_ctrl()
|
/linux-4.4.14/fs/fuse/ |
D | file.c | 1455 dec_wb_stat(&bdi->wb, WB_WRITEBACK); in fuse_writepage_finish() 1457 wb_writeout_inc(&bdi->wb); in fuse_writepage_finish() 1645 inc_wb_stat(&inode_to_bdi(inode)->wb, WB_WRITEBACK); in fuse_writepage_locked() 1759 dec_wb_stat(&bdi->wb, WB_WRITEBACK); in fuse_writepage_in_flight() 1761 wb_writeout_inc(&bdi->wb); in fuse_writepage_in_flight() 1858 inc_wb_stat(&inode_to_bdi(inode)->wb, WB_WRITEBACK); in fuse_writepages_fill()
|
/linux-4.4.14/arch/mips/include/asm/sn/sn0/ |
D | hubio.h | 892 wb: 1, /* 15: writeback pending. */ member
|
/linux-4.4.14/Documentation/x86/ |
D | pat.txt | 157 will first make a region uc, wc or wt and switch it back to wb after use.
|
/linux-4.4.14/fs/nfs/ |
D | internal.h | 630 inc_wb_stat(&inode_to_bdi(inode)->wb, WB_RECLAIMABLE); in nfs_mark_page_unstable()
|
D | write.c | 875 dec_wb_stat(&inode_to_bdi(page_file_mapping(page)->host)->wb, in nfs_clear_page_commit()
|
/linux-4.4.14/fs/f2fs/ |
D | segment.h | 699 if (sbi->sb->s_bdi->wb.dirty_exceeded) in nr_pages_to_skip()
|
D | node.c | 56 if (sbi->sb->s_bdi->wb.dirty_exceeded) in available_free_memory() 73 if (sbi->sb->s_bdi->wb.dirty_exceeded) in available_free_memory()
|
/linux-4.4.14/drivers/net/ethernet/intel/e1000/ |
D | e1000_hw.h | 553 } wb; /* writeback */ member 584 } wb; /* writeback */ member
|
/linux-4.4.14/block/ |
D | blk-core.c | 76 clear_wb_congested(rl->q->backing_dev_info.wb.congested, sync); in blk_clear_congested() 87 set_wb_congested(rl->q->backing_dev_info.wb.congested, sync); in blk_set_congested()
|
/linux-4.4.14/fs/gfs2/ |
D | super.c | 751 if (bdi->wb.dirty_exceeded) in gfs2_write_inode()
|
/linux-4.4.14/drivers/md/ |
D | dm.c | 2155 r = md->queue->backing_dev_info.wb.state & in dm_any_congested()
|