Home
last modified time | relevance | path

Searched refs:wb (Results 1 – 106 of 106) sorted by relevance

/linux-4.4.14/fs/
Dfs-writeback.c102 static bool wb_io_lists_populated(struct bdi_writeback *wb) in wb_io_lists_populated() argument
104 if (wb_has_dirty_io(wb)) { in wb_io_lists_populated()
107 set_bit(WB_has_dirty_io, &wb->state); in wb_io_lists_populated()
108 WARN_ON_ONCE(!wb->avg_write_bandwidth); in wb_io_lists_populated()
109 atomic_long_add(wb->avg_write_bandwidth, in wb_io_lists_populated()
110 &wb->bdi->tot_write_bandwidth); in wb_io_lists_populated()
115 static void wb_io_lists_depopulated(struct bdi_writeback *wb) in wb_io_lists_depopulated() argument
117 if (wb_has_dirty_io(wb) && list_empty(&wb->b_dirty) && in wb_io_lists_depopulated()
118 list_empty(&wb->b_io) && list_empty(&wb->b_more_io)) { in wb_io_lists_depopulated()
119 clear_bit(WB_has_dirty_io, &wb->state); in wb_io_lists_depopulated()
[all …]
/linux-4.4.14/mm/
Dbacking-dev.c49 struct bdi_writeback *wb = &bdi->wb; in bdi_debug_stats_show() local
57 spin_lock(&wb->list_lock); in bdi_debug_stats_show()
58 list_for_each_entry(inode, &wb->b_dirty, i_io_list) in bdi_debug_stats_show()
60 list_for_each_entry(inode, &wb->b_io, i_io_list) in bdi_debug_stats_show()
62 list_for_each_entry(inode, &wb->b_more_io, i_io_list) in bdi_debug_stats_show()
64 list_for_each_entry(inode, &wb->b_dirty_time, i_io_list) in bdi_debug_stats_show()
67 spin_unlock(&wb->list_lock); in bdi_debug_stats_show()
70 wb_thresh = wb_calc_thresh(wb, dirty_thresh); in bdi_debug_stats_show()
88 (unsigned long) K(wb_stat(wb, WB_WRITEBACK)), in bdi_debug_stats_show()
89 (unsigned long) K(wb_stat(wb, WB_RECLAIMABLE)), in bdi_debug_stats_show()
[all …]
Dpage-writeback.c133 struct bdi_writeback *wb; member
157 #define GDTC_INIT(__wb) .wb = (__wb), \
163 #define MDTC_INIT(__wb, __gdtc) .wb = (__wb), \
183 static struct fprop_local_percpu *wb_memcg_completions(struct bdi_writeback *wb) in wb_memcg_completions() argument
185 return &wb->memcg_completions; in wb_memcg_completions()
188 static void wb_min_max_ratio(struct bdi_writeback *wb, in wb_min_max_ratio() argument
191 unsigned long this_bw = wb->avg_write_bandwidth; in wb_min_max_ratio()
192 unsigned long tot_bw = atomic_long_read(&wb->bdi->tot_write_bandwidth); in wb_min_max_ratio()
193 unsigned long long min = wb->bdi->min_ratio; in wb_min_max_ratio()
194 unsigned long long max = wb->bdi->max_ratio; in wb_min_max_ratio()
[all …]
Dmemcontrol.c3725 struct wb_domain *mem_cgroup_wb_domain(struct bdi_writeback *wb) in mem_cgroup_wb_domain() argument
3727 struct mem_cgroup *memcg = mem_cgroup_from_css(wb->memcg_css); in mem_cgroup_wb_domain()
3753 void mem_cgroup_wb_stats(struct bdi_writeback *wb, unsigned long *pfilepages, in mem_cgroup_wb_stats() argument
3757 struct mem_cgroup *memcg = mem_cgroup_from_css(wb->memcg_css); in mem_cgroup_wb_stats()
/linux-4.4.14/fs/afs/
Dwrite.c19 static int afs_write_back_from_locked_page(struct afs_writeback *wb,
35 static void afs_unlink_writeback(struct afs_writeback *wb) in afs_unlink_writeback() argument
38 struct afs_vnode *vnode = wb->vnode; in afs_unlink_writeback()
40 list_del_init(&wb->link); in afs_unlink_writeback()
57 static void afs_free_writeback(struct afs_writeback *wb) in afs_free_writeback() argument
60 key_put(wb->key); in afs_free_writeback()
61 kfree(wb); in afs_free_writeback()
67 void afs_put_writeback(struct afs_writeback *wb) in afs_put_writeback() argument
69 struct afs_vnode *vnode = wb->vnode; in afs_put_writeback()
71 _enter("{%d}", wb->usage); in afs_put_writeback()
[all …]
Dfile.c315 struct afs_writeback *wb = (struct afs_writeback *) page_private(page); in afs_invalidatepage() local
332 if (wb && !PageWriteback(page)) { in afs_invalidatepage()
334 afs_put_writeback(wb); in afs_invalidatepage()
351 struct afs_writeback *wb = (struct afs_writeback *) page_private(page); in afs_releasepage() local
368 if (wb) { in afs_releasepage()
370 afs_put_writeback(wb); in afs_releasepage()
Dfsclient.c1156 struct afs_writeback *wb, in afs_fs_store_data64() argument
1162 struct afs_vnode *vnode = wb->vnode; in afs_fs_store_data64()
1167 key_serial(wb->key), vnode->fid.vid, vnode->fid.vnode); in afs_fs_store_data64()
1175 call->wb = wb; in afs_fs_store_data64()
1176 call->key = wb->key; in afs_fs_store_data64()
1215 int afs_fs_store_data(struct afs_server *server, struct afs_writeback *wb, in afs_fs_store_data() argument
1220 struct afs_vnode *vnode = wb->vnode; in afs_fs_store_data()
1226 key_serial(wb->key), vnode->fid.vid, vnode->fid.vnode); in afs_fs_store_data()
1243 return afs_fs_store_data64(server, wb, first, last, offset, to, in afs_fs_store_data()
1252 call->wb = wb; in afs_fs_store_data()
[all …]
Dvnode.c769 int afs_vnode_store_data(struct afs_writeback *wb, pgoff_t first, pgoff_t last, in afs_vnode_store_data() argument
773 struct afs_vnode *vnode = wb->vnode; in afs_vnode_store_data()
781 key_serial(wb->key), in afs_vnode_store_data()
797 ret = afs_fs_store_data(server, wb, first, last, offset, to, in afs_vnode_store_data()
Dinternal.h87 struct afs_writeback *wb; /* writeback being performed */ member
/linux-4.4.14/include/linux/
Dbacking-dev.h32 void wb_start_writeback(struct bdi_writeback *wb, long nr_pages,
34 void wb_start_background_writeback(struct bdi_writeback *wb);
36 void wb_wakeup_delayed(struct bdi_writeback *wb);
43 static inline bool wb_has_dirty_io(struct bdi_writeback *wb) in wb_has_dirty_io() argument
45 return test_bit(WB_has_dirty_io, &wb->state); in wb_has_dirty_io()
57 static inline void __add_wb_stat(struct bdi_writeback *wb, in __add_wb_stat() argument
60 __percpu_counter_add(&wb->stat[item], amount, WB_STAT_BATCH); in __add_wb_stat()
63 static inline void __inc_wb_stat(struct bdi_writeback *wb, in __inc_wb_stat() argument
66 __add_wb_stat(wb, item, 1); in __inc_wb_stat()
69 static inline void inc_wb_stat(struct bdi_writeback *wb, enum wb_stat_item item) in inc_wb_stat() argument
[all …]
Dbacking-dev-defs.h154 struct bdi_writeback wb; /* the root writeback info for this bdi */ member
185 clear_wb_congested(bdi->wb.congested, sync); in clear_bdi_congested()
190 set_wb_congested(bdi->wb.congested, sync); in set_bdi_congested()
199 static inline bool wb_tryget(struct bdi_writeback *wb) in wb_tryget() argument
201 if (wb != &wb->bdi->wb) in wb_tryget()
202 return percpu_ref_tryget(&wb->refcnt); in wb_tryget()
210 static inline void wb_get(struct bdi_writeback *wb) in wb_get() argument
212 if (wb != &wb->bdi->wb) in wb_get()
213 percpu_ref_get(&wb->refcnt); in wb_get()
220 static inline void wb_put(struct bdi_writeback *wb) in wb_put() argument
[all …]
Dwriteback.h90 struct bdi_writeback *wb; /* wb this writeback is issued under */ member
266 if (wbc->wb) in wbc_init_bio()
267 bio_associate_blkcg(bio, wbc->wb->blkcg_css); in wbc_init_bio()
363 unsigned long wb_calc_thresh(struct bdi_writeback *wb, unsigned long thresh);
365 void wb_update_bandwidth(struct bdi_writeback *wb, unsigned long start_time);
368 bool wb_over_bg_thresh(struct bdi_writeback *wb);
Dmemcontrol.h683 struct wb_domain *mem_cgroup_wb_domain(struct bdi_writeback *wb);
684 void mem_cgroup_wb_stats(struct bdi_writeback *wb, unsigned long *pfilepages,
690 static inline struct wb_domain *mem_cgroup_wb_domain(struct bdi_writeback *wb) in mem_cgroup_wb_domain() argument
695 static inline void mem_cgroup_wb_stats(struct bdi_writeback *wb, in mem_cgroup_wb_stats() argument
Dmm.h1273 struct mem_cgroup *memcg, struct bdi_writeback *wb);
/linux-4.4.14/drivers/staging/wilc1000/
Dwilc_spi.c273 u8 wb[32], rb[32]; in spi_cmd_complete() local
280 wb[0] = cmd; in spi_cmd_complete()
283 wb[1] = (u8)(adr >> 16); in spi_cmd_complete()
284 wb[2] = (u8)(adr >> 8); in spi_cmd_complete()
285 wb[3] = (u8)adr; in spi_cmd_complete()
290 wb[1] = (u8)(adr >> 8); in spi_cmd_complete()
292 wb[1] |= BIT(7); in spi_cmd_complete()
293 wb[2] = (u8)adr; in spi_cmd_complete()
294 wb[3] = 0x00; in spi_cmd_complete()
299 wb[1] = 0x00; in spi_cmd_complete()
[all …]
Dlinux_wlan_spi.h12 int linux_spi_write_read(u8 *wb, u8 *rb, u32 rlen);
Dlinux_wlan_spi.c366 int linux_spi_write_read(u8 *wb, u8 *rb, u32 rlen) in linux_spi_write_read() argument
375 .tx_buf = wb, in linux_spi_write_read()
/linux-4.4.14/include/trace/events/
Dwriteback.h137 static inline size_t __trace_wb_cgroup_size(struct bdi_writeback *wb) in __trace_wb_cgroup_size() argument
139 return kernfs_path_len(wb->memcg_css->cgroup->kn) + 1; in __trace_wb_cgroup_size()
142 static inline void __trace_wb_assign_cgroup(char *buf, struct bdi_writeback *wb) in __trace_wb_assign_cgroup() argument
144 struct cgroup *cgrp = wb->memcg_css->cgroup; in __trace_wb_assign_cgroup()
153 if (wbc->wb) in __trace_wbc_cgroup_size()
154 return __trace_wb_cgroup_size(wbc->wb); in __trace_wbc_cgroup_size()
162 if (wbc->wb) in __trace_wbc_assign_cgroup()
163 __trace_wb_assign_cgroup(buf, wbc->wb); in __trace_wbc_assign_cgroup()
170 static inline size_t __trace_wb_cgroup_size(struct bdi_writeback *wb) in __trace_wb_cgroup_size() argument
175 static inline void __trace_wb_assign_cgroup(char *buf, struct bdi_writeback *wb) in __trace_wb_assign_cgroup() argument
[all …]
/linux-4.4.14/arch/xtensa/kernel/
Dptrace.c56 unsigned long wb = regs->windowbase; in ptrace_getregs() local
73 gregset->a + ((wb * 4 + i) % XCHAL_NUM_AREGS)); in ptrace_getregs()
84 unsigned long wb, ws; in ptrace_setregs() local
95 __get_user(wb, &gregset->windowbase); in ptrace_setregs()
100 if (wb >= XCHAL_NUM_AREGS / 4) in ptrace_setregs()
103 if (wb != regs->windowbase || ws != regs->windowstart) { in ptrace_setregs()
106 rotws = (((ws | (ws << WSBITS)) >> wb) & in ptrace_setregs()
110 regs->windowbase = wb; in ptrace_setregs()
115 if (wb != 0 && __copy_from_user(regs->areg + XCHAL_NUM_AREGS - wb * 4, in ptrace_setregs()
116 gregset->a, wb * 16)) in ptrace_setregs()
[all …]
Dprocess.c323 unsigned long wb, ws, wm; in xtensa_elf_core_copy_regs() local
326 wb = regs->windowbase; in xtensa_elf_core_copy_regs()
329 ws = ((ws >> wb) | (ws << (WSBITS - wb))) & ((1 << WSBITS) - 1); in xtensa_elf_core_copy_regs()
Dsignal.c58 const unsigned long wb = regs->windowbase; in flush_window_regs_user() local
71 wm = (ws >> wb) | (ws << (XCHAL_NUM_AREGS / 4 - wb)); in flush_window_regs_user()
118 regs->windowstart = 1 << wb; in flush_window_regs_user()
/linux-4.4.14/Documentation/devicetree/bindings/interrupt-controller/
Dqca,ath79-cpu-intc.txt5 qca,ddr-wb-channels and qca,ddr-wb-channel-interrupts properties.
20 - qca,ddr-wb-channel-interrupts: List of the interrupts needing a write
22 - qca,ddr-wb-channels: List of phandles to the write buffer channels for
23 each interrupt. If qca,ddr-wb-channel-interrupts is not present the interrupt
34 qca,ddr-wb-channel-interrupts = <2>, <3>, <4>, <5>;
35 qca,ddr-wb-channels = <&ddr_ctrl 3>, <&ddr_ctrl 2>,
43 #qca,ddr-wb-channel-cells = <1>;
/linux-4.4.14/scripts/
Dextract-cert.c72 static BIO *wb; variable
80 if (!wb) { in write_cert()
81 wb = BIO_new_file(cert_dst, "wb"); in write_cert()
82 ERR(!wb, "%s", cert_dst); in write_cert()
85 ERR(!i2d_X509_bio(wb, x509), "%s", cert_dst); in write_cert()
146 if (wb && !x509) { in main()
159 BIO_free(wb); in main()
Dkallsyms.c641 int wa, wb; in compare_symbols() local
654 wb = (sb->sym[0] == 'w') || (sb->sym[0] == 'W'); in compare_symbols()
655 if (wa != wb) in compare_symbols()
656 return wa - wb; in compare_symbols()
660 wb = may_be_linker_script_provide_symbol(sb); in compare_symbols()
661 if (wa != wb) in compare_symbols()
662 return wa - wb; in compare_symbols()
666 wb = prefix_underscores_count((const char *)sb->sym + 1); in compare_symbols()
667 if (wa != wb) in compare_symbols()
668 return wa - wb; in compare_symbols()
/linux-4.4.14/drivers/gpu/drm/radeon/
Dr600_dma.c56 if (rdev->wb.enabled) in r600_dma_get_rptr()
57 rptr = rdev->wb.wb[ring->rptr_offs/4]; in r600_dma_get_rptr()
144 upper_32_bits(rdev->wb.gpu_addr + R600_WB_DMA_RPTR_OFFSET) & 0xFF); in r600_dma_resume()
146 ((rdev->wb.gpu_addr + R600_WB_DMA_RPTR_OFFSET) & 0xFFFFFFFC)); in r600_dma_resume()
148 if (rdev->wb.enabled) in r600_dma_resume()
244 gpu_addr = rdev->wb.gpu_addr + index; in r600_dma_ring_test()
247 rdev->wb.wb[index/4] = cpu_to_le32(tmp); in r600_dma_ring_test()
261 tmp = le32_to_cpu(rdev->wb.wb[index/4]); in r600_dma_ring_test()
351 gpu_addr = rdev->wb.gpu_addr + index; in r600_dma_ib_test()
377 tmp = le32_to_cpu(rdev->wb.wb[index/4]); in r600_dma_ib_test()
[all …]
Dradeon_device.c427 rdev->wb.enabled = false; in radeon_wb_disable()
441 if (rdev->wb.wb_obj) { in radeon_wb_fini()
442 if (!radeon_bo_reserve(rdev->wb.wb_obj, false)) { in radeon_wb_fini()
443 radeon_bo_kunmap(rdev->wb.wb_obj); in radeon_wb_fini()
444 radeon_bo_unpin(rdev->wb.wb_obj); in radeon_wb_fini()
445 radeon_bo_unreserve(rdev->wb.wb_obj); in radeon_wb_fini()
447 radeon_bo_unref(&rdev->wb.wb_obj); in radeon_wb_fini()
448 rdev->wb.wb = NULL; in radeon_wb_fini()
449 rdev->wb.wb_obj = NULL; in radeon_wb_fini()
466 if (rdev->wb.wb_obj == NULL) { in radeon_wb_init()
[all …]
Dcik_sdma.c68 if (rdev->wb.enabled) { in cik_sdma_get_rptr()
69 rptr = rdev->wb.wb[ring->rptr_offs/4]; in cik_sdma_get_rptr()
139 if (rdev->wb.enabled) { in cik_sdma_ring_ib_execute()
401 upper_32_bits(rdev->wb.gpu_addr + wb_offset) & 0xFFFFFFFF); in cik_sdma_gfx_resume()
403 ((rdev->wb.gpu_addr + wb_offset) & 0xFFFFFFFC)); in cik_sdma_gfx_resume()
405 if (rdev->wb.enabled) in cik_sdma_gfx_resume()
659 gpu_addr = rdev->wb.gpu_addr + index; in cik_sdma_ring_test()
662 rdev->wb.wb[index/4] = cpu_to_le32(tmp); in cik_sdma_ring_test()
677 tmp = le32_to_cpu(rdev->wb.wb[index/4]); in cik_sdma_ring_test()
716 gpu_addr = rdev->wb.gpu_addr + index; in cik_sdma_ib_test()
[all …]
Dni_dma.c58 if (rdev->wb.enabled) { in cayman_dma_get_rptr()
59 rptr = rdev->wb.wb[ring->rptr_offs/4]; in cayman_dma_get_rptr()
128 if (rdev->wb.enabled) { in cayman_dma_ring_ib_execute()
223 upper_32_bits(rdev->wb.gpu_addr + wb_offset) & 0xFF); in cayman_dma_resume()
225 ((rdev->wb.gpu_addr + wb_offset) & 0xFFFFFFFC)); in cayman_dma_resume()
227 if (rdev->wb.enabled) in cayman_dma_resume()
Dradeon_fence.c65 if (likely(rdev->wb.enabled || !drv->scratch_reg)) { in radeon_fence_write()
88 if (likely(rdev->wb.enabled || !drv->scratch_reg)) { in radeon_fence_read()
807 if (rdev->wb.use_event || !radeon_ring_supports_scratch_reg(rdev, &rdev->ring[ring])) { in radeon_fence_driver_start_ring()
811 rdev->fence_drv[ring].cpu_addr = &rdev->wb.wb[index/4]; in radeon_fence_driver_start_ring()
812 rdev->fence_drv[ring].gpu_addr = rdev->wb.gpu_addr + in radeon_fence_driver_start_ring()
831 rdev->fence_drv[ring].cpu_addr = &rdev->wb.wb[index/4]; in radeon_fence_driver_start_ring()
832 rdev->fence_drv[ring].gpu_addr = rdev->wb.gpu_addr + index; in radeon_fence_driver_start_ring()
Dradeon_ring.c300 else if (rdev->wb.enabled) in radeon_ring_backup()
412 if (rdev->wb.enabled) { in radeon_ring_init()
414 ring->next_rptr_gpu_addr = rdev->wb.gpu_addr + index; in radeon_ring_init()
415 ring->next_rptr_cpu_addr = &rdev->wb.wb[index/4]; in radeon_ring_init()
Devergreen_dma.c72 if (rdev->wb.enabled) { in evergreen_dma_ring_ib_execute()
Dcik.c4161 } else if (rdev->wb.enabled) { in cik_ring_ib_execute()
4479 WREG32(SCRATCH_ADDR, ((rdev->wb.gpu_addr + RADEON_WB_SCRATCH_OFFSET) >> 8) & 0xFFFFFFFF); in cik_cp_gfx_resume()
4497 WREG32(CP_RB0_RPTR_ADDR, (rdev->wb.gpu_addr + RADEON_WB_CP_RPTR_OFFSET) & 0xFFFFFFFC); in cik_cp_gfx_resume()
4498 WREG32(CP_RB0_RPTR_ADDR_HI, upper_32_bits(rdev->wb.gpu_addr + RADEON_WB_CP_RPTR_OFFSET) & 0xFF); in cik_cp_gfx_resume()
4503 if (!rdev->wb.enabled) in cik_cp_gfx_resume()
4533 if (rdev->wb.enabled) in cik_gfx_get_rptr()
4534 rptr = rdev->wb.wb[ring->rptr_offs/4]; in cik_gfx_get_rptr()
4563 if (rdev->wb.enabled) { in cik_compute_get_rptr()
4564 rptr = rdev->wb.wb[ring->rptr_offs/4]; in cik_compute_get_rptr()
4581 if (rdev->wb.enabled) { in cik_compute_get_wptr()
[all …]
Dr600.c2616 if (rdev->wb.enabled) in r600_gfx_get_rptr()
2617 rptr = rdev->wb.wb[ring->rptr_offs/4]; in r600_gfx_get_rptr()
2746 ((rdev->wb.gpu_addr + RADEON_WB_CP_RPTR_OFFSET) & 0xFFFFFFFC)); in r600_cp_resume()
2747 WREG32(CP_RB_RPTR_ADDR_HI, upper_32_bits(rdev->wb.gpu_addr + RADEON_WB_CP_RPTR_OFFSET) & 0xFF); in r600_cp_resume()
2748 WREG32(SCRATCH_ADDR, ((rdev->wb.gpu_addr + RADEON_WB_SCRATCH_OFFSET) >> 8) & 0xFFFFFFFF); in r600_cp_resume()
2750 if (rdev->wb.enabled) in r600_cp_resume()
2874 if (rdev->wb.use_event) { in r600_fence_ring_emit()
3337 } else if (rdev->wb.enabled) { in r600_ring_ib_execute()
3668 if (rdev->wb.enabled) in r600_irq_init()
3672 WREG32(IH_RB_WPTR_ADDR_LO, (rdev->wb.gpu_addr + R600_WB_IH_WPTR_OFFSET) & 0xFFFFFFFC); in r600_irq_init()
[all …]
Dsi.c3421 } else if (rdev->wb.enabled) { in si_ring_ib_execute()
3661 WREG32(SCRATCH_ADDR, ((rdev->wb.gpu_addr + RADEON_WB_SCRATCH_OFFSET) >> 8) & 0xFFFFFFFF); in si_cp_resume()
3679 WREG32(CP_RB0_RPTR_ADDR, (rdev->wb.gpu_addr + RADEON_WB_CP_RPTR_OFFSET) & 0xFFFFFFFC); in si_cp_resume()
3680 WREG32(CP_RB0_RPTR_ADDR_HI, upper_32_bits(rdev->wb.gpu_addr + RADEON_WB_CP_RPTR_OFFSET) & 0xFF); in si_cp_resume()
3682 if (rdev->wb.enabled) in si_cp_resume()
3710 WREG32(CP_RB1_RPTR_ADDR, (rdev->wb.gpu_addr + RADEON_WB_CP1_RPTR_OFFSET) & 0xFFFFFFFC); in si_cp_resume()
3711 WREG32(CP_RB1_RPTR_ADDR_HI, upper_32_bits(rdev->wb.gpu_addr + RADEON_WB_CP1_RPTR_OFFSET) & 0xFF); in si_cp_resume()
3734 WREG32(CP_RB2_RPTR_ADDR, (rdev->wb.gpu_addr + RADEON_WB_CP2_RPTR_OFFSET) & 0xFFFFFFFC); in si_cp_resume()
3735 WREG32(CP_RB2_RPTR_ADDR_HI, upper_32_bits(rdev->wb.gpu_addr + RADEON_WB_CP2_RPTR_OFFSET) & 0xFF); in si_cp_resume()
6034 if (rdev->wb.enabled) in si_irq_init()
[all …]
Dni.c1485 if (rdev->wb.enabled) in cayman_gfx_get_rptr()
1486 rptr = rdev->wb.wb[ring->rptr_offs/4]; in cayman_gfx_get_rptr()
1691 WREG32(SCRATCH_ADDR, ((rdev->wb.gpu_addr + RADEON_WB_SCRATCH_OFFSET) >> 8) & 0xFFFFFFFF); in cayman_cp_resume()
1708 addr = rdev->wb.gpu_addr + RADEON_WB_CP_RPTR_OFFSET; in cayman_cp_resume()
Devergreen.c3034 } else if (rdev->wb.enabled) { in evergreen_ring_ib_execute()
3192 ((rdev->wb.gpu_addr + RADEON_WB_CP_RPTR_OFFSET) & 0xFFFFFFFC)); in evergreen_cp_resume()
3193 WREG32(CP_RB_RPTR_ADDR_HI, upper_32_bits(rdev->wb.gpu_addr + RADEON_WB_CP_RPTR_OFFSET) & 0xFF); in evergreen_cp_resume()
3194 WREG32(SCRATCH_ADDR, ((rdev->wb.gpu_addr + RADEON_WB_SCRATCH_OFFSET) >> 8) & 0xFFFFFFFF); in evergreen_cp_resume()
3196 if (rdev->wb.enabled) in evergreen_cp_resume()
5014 if (rdev->wb.enabled) in evergreen_get_ih_wptr()
5015 wptr = le32_to_cpu(rdev->wb.wb[R600_WB_IH_WPTR_OFFSET/4]); in evergreen_get_ih_wptr()
Dr100.c1061 if (rdev->wb.enabled) in r100_gfx_get_rptr()
1062 rptr = le32_to_cpu(rdev->wb.wb[ring->rptr_offs/4]); in r100_gfx_get_rptr()
1187 S_00070C_RB_RPTR_ADDR((rdev->wb.gpu_addr + RADEON_WB_CP_RPTR_OFFSET) >> 2)); in r100_cp_init()
1188 WREG32(R_000774_SCRATCH_ADDR, rdev->wb.gpu_addr + RADEON_WB_SCRATCH_OFFSET); in r100_cp_init()
1190 if (rdev->wb.enabled) in r100_cp_init()
Dradeon.h1143 volatile uint32_t *wb; member
2392 struct radeon_wb wb; member
/linux-4.4.14/drivers/usb/class/
Dcdc-acm.c169 struct acm_wb *wb; in acm_wb_alloc() local
174 wb = &acm->wb[wbn]; in acm_wb_alloc()
175 if (!wb->use) { in acm_wb_alloc()
176 wb->use = 1; in acm_wb_alloc()
193 n -= acm->wb[i].use; in acm_wb_is_avail()
201 static void acm_write_done(struct acm *acm, struct acm_wb *wb) in acm_write_done() argument
203 wb->use = 0; in acm_write_done()
214 static int acm_start_wb(struct acm *acm, struct acm_wb *wb) in acm_start_wb() argument
220 wb->urb->transfer_buffer = wb->buf; in acm_start_wb()
221 wb->urb->transfer_dma = wb->dmah; in acm_start_wb()
[all …]
Dcdc-acm.h93 struct acm_wb wb[ACM_NW]; member
/linux-4.4.14/drivers/gpu/drm/amd/amdgpu/
Damdgpu_device.c426 if (adev->wb.wb_obj) { in amdgpu_wb_fini()
427 if (!amdgpu_bo_reserve(adev->wb.wb_obj, false)) { in amdgpu_wb_fini()
428 amdgpu_bo_kunmap(adev->wb.wb_obj); in amdgpu_wb_fini()
429 amdgpu_bo_unpin(adev->wb.wb_obj); in amdgpu_wb_fini()
430 amdgpu_bo_unreserve(adev->wb.wb_obj); in amdgpu_wb_fini()
432 amdgpu_bo_unref(&adev->wb.wb_obj); in amdgpu_wb_fini()
433 adev->wb.wb = NULL; in amdgpu_wb_fini()
434 adev->wb.wb_obj = NULL; in amdgpu_wb_fini()
451 if (adev->wb.wb_obj == NULL) { in amdgpu_wb_init()
454 &adev->wb.wb_obj); in amdgpu_wb_init()
[all …]
Dsdma_v3_0.c283 rptr = ring->adev->wb.wb[ring->rptr_offs] >> 2; in sdma_v3_0_ring_get_rptr()
302 wptr = ring->adev->wb.wb[ring->wptr_offs] >> 2; in sdma_v3_0_ring_get_wptr()
325 adev->wb.wb[ring->wptr_offs] = ring->wptr << 2; in sdma_v3_0_ring_set_wptr()
618 upper_32_bits(adev->wb.gpu_addr + wb_offset) & 0xFFFFFFFF); in sdma_v3_0_gfx_resume()
620 lower_32_bits(adev->wb.gpu_addr + wb_offset) & 0xFFFFFFFC); in sdma_v3_0_gfx_resume()
785 gpu_addr = adev->wb.gpu_addr + (index * 4); in sdma_v3_0_ring_test_ring()
787 adev->wb.wb[index] = cpu_to_le32(tmp); in sdma_v3_0_ring_test_ring()
805 tmp = le32_to_cpu(adev->wb.wb[index]); in sdma_v3_0_ring_test_ring()
848 gpu_addr = adev->wb.gpu_addr + (index * 4); in sdma_v3_0_ring_test_ib()
850 adev->wb.wb[index] = cpu_to_le32(tmp); in sdma_v3_0_ring_test_ib()
[all …]
Dcik_sdma.c156 rptr = ring->adev->wb.wb[ring->rptr_offs]; in cik_sdma_ring_get_rptr()
438 upper_32_bits(adev->wb.gpu_addr + wb_offset) & 0xFFFFFFFF); in cik_sdma_gfx_resume()
440 ((adev->wb.gpu_addr + wb_offset) & 0xFFFFFFFC)); in cik_sdma_gfx_resume()
583 gpu_addr = adev->wb.gpu_addr + (index * 4); in cik_sdma_ring_test_ring()
585 adev->wb.wb[index] = cpu_to_le32(tmp); in cik_sdma_ring_test_ring()
601 tmp = le32_to_cpu(adev->wb.wb[index]); in cik_sdma_ring_test_ring()
644 gpu_addr = adev->wb.gpu_addr + (index * 4); in cik_sdma_ring_test_ib()
646 adev->wb.wb[index] = cpu_to_le32(tmp); in cik_sdma_ring_test_ib()
672 tmp = le32_to_cpu(adev->wb.wb[index]); in cik_sdma_ring_test_ib()
Dsdma_v2_4.c187 rptr = ring->adev->wb.wb[ring->rptr_offs] >> 2; in sdma_v2_4_ring_get_rptr()
481 upper_32_bits(adev->wb.gpu_addr + wb_offset) & 0xFFFFFFFF); in sdma_v2_4_gfx_resume()
483 lower_32_bits(adev->wb.gpu_addr + wb_offset) & 0xFFFFFFFC); in sdma_v2_4_gfx_resume()
635 gpu_addr = adev->wb.gpu_addr + (index * 4); in sdma_v2_4_ring_test_ring()
637 adev->wb.wb[index] = cpu_to_le32(tmp); in sdma_v2_4_ring_test_ring()
655 tmp = le32_to_cpu(adev->wb.wb[index]); in sdma_v2_4_ring_test_ring()
698 gpu_addr = adev->wb.gpu_addr + (index * 4); in sdma_v2_4_ring_test_ib()
700 adev->wb.wb[index] = cpu_to_le32(tmp); in sdma_v2_4_ring_test_ib()
731 tmp = le32_to_cpu(adev->wb.wb[index]); in sdma_v2_4_ring_test_ib()
Dtonga_ih.c143 wptr_off = adev->wb.gpu_addr + (adev->irq.ih.wptr_offs * 4); in tonga_ih_irq_init()
204 wptr = le32_to_cpu(adev->wb.wb[adev->irq.ih.wptr_offs]); in tonga_ih_get_wptr()
266 adev->wb.wb[adev->irq.ih.rptr_offs] = adev->irq.ih.rptr; in tonga_ih_set_rptr()
Dcik_ih.c135 wptr_off = adev->wb.gpu_addr + (adev->irq.ih.wptr_offs * 4); in cik_ih_irq_init()
191 wptr = le32_to_cpu(adev->wb.wb[adev->irq.ih.wptr_offs]); in cik_ih_get_wptr()
Dcz_ih.c137 wptr_off = adev->wb.gpu_addr + (adev->irq.ih.wptr_offs * 4); in cz_ih_irq_init()
193 wptr = le32_to_cpu(adev->wb.wb[adev->irq.ih.wptr_offs]); in cz_ih_get_wptr()
Diceland_ih.c137 wptr_off = adev->wb.gpu_addr + (adev->irq.ih.wptr_offs * 4); in iceland_ih_irq_init()
193 wptr = le32_to_cpu(adev->wb.wb[adev->irq.ih.wptr_offs]); in iceland_ih_get_wptr()
Damdgpu_ring.c346 ring->next_rptr_gpu_addr = adev->wb.gpu_addr + (ring->next_rptr_offs * 4); in amdgpu_ring_init()
347 ring->next_rptr_cpu_addr = &adev->wb.wb[ring->next_rptr_offs]; in amdgpu_ring_init()
Damdgpu_fence.c442 ring->fence_drv.cpu_addr = &adev->wb.wb[ring->fence_offs]; in amdgpu_fence_driver_start_ring()
443 ring->fence_drv.gpu_addr = adev->wb.gpu_addr + (ring->fence_offs * 4); in amdgpu_fence_driver_start_ring()
Dgfx_v8_0.c3328 rptr_addr = adev->wb.gpu_addr + (ring->rptr_offs * 4); in gfx_v8_0_cp_gfx_resume()
3889 wb_gpu_addr = adev->wb.gpu_addr + (ring->rptr_offs * 4); in gfx_v8_0_cp_compute_resume()
3899 wb_gpu_addr = adev->wb.gpu_addr + (ring->wptr_offs * 4); in gfx_v8_0_cp_compute_resume()
4477 rptr = ring->adev->wb.wb[ring->rptr_offs]; in gfx_v8_0_ring_get_rptr_gfx()
4489 wptr = ring->adev->wb.wb[ring->wptr_offs]; in gfx_v8_0_ring_get_wptr_gfx()
4502 adev->wb.wb[ring->wptr_offs] = ring->wptr; in gfx_v8_0_ring_set_wptr_gfx()
4748 return ring->adev->wb.wb[ring->rptr_offs]; in gfx_v8_0_ring_get_rptr_compute()
4753 return ring->adev->wb.wb[ring->wptr_offs]; in gfx_v8_0_ring_get_wptr_compute()
4761 adev->wb.wb[ring->wptr_offs] = ring->wptr; in gfx_v8_0_ring_set_wptr_compute()
Damdgpu.h1293 volatile uint32_t *wb; member
1299 int amdgpu_wb_get(struct amdgpu_device *adev, u32 *wb);
1300 void amdgpu_wb_free(struct amdgpu_device *adev, u32 wb);
2037 struct amdgpu_wb wb; member
2256 …amdgpu_ring_emit_gds_switch(r, v, db, ds, wb, ws, ab, as) (r)->funcs->emit_gds_switch((r), (v), (d… argument
Dgfx_v7_0.c2964 rptr_addr = adev->wb.gpu_addr + (ring->rptr_offs * 4); in gfx_v7_0_cp_gfx_resume()
2994 rptr = ring->adev->wb.wb[ring->rptr_offs]; in gfx_v7_0_ring_get_rptr_gfx()
3021 rptr = ring->adev->wb.wb[ring->rptr_offs]; in gfx_v7_0_ring_get_rptr_compute()
3031 wptr = ring->adev->wb.wb[ring->wptr_offs]; in gfx_v7_0_ring_get_wptr_compute()
3041 adev->wb.wb[ring->wptr_offs] = ring->wptr; in gfx_v7_0_ring_set_wptr_compute()
3490 wb_gpu_addr = adev->wb.gpu_addr + (ring->wptr_offs * 4); in gfx_v7_0_cp_compute_resume()
3498 wb_gpu_addr = adev->wb.gpu_addr + (ring->rptr_offs * 4); in gfx_v7_0_cp_compute_resume()
Damdgpu_test.c49 if (adev->wb.wb_obj) in amdgpu_do_test_moves()
/linux-4.4.14/Documentation/devicetree/bindings/memory-controllers/
Dath79-ddr-controller.txt15 - #qca,ddr-wb-channel-cells: has to be 1, the index of the write buffer
25 #qca,ddr-wb-channel-cells = <1>;
32 qca,ddr-wb-channel-interrupts = <2>, <3>, <4>, <5>;
33 qca,ddr-wb-channels = <&ddr_ctrl 3>, <&ddr_ctrl 2>,
/linux-4.4.14/arch/mips/boot/dts/qca/
Dar9132.dtsi24 qca,ddr-wb-channel-interrupts = <2>, <3>, <4>, <5>;
25 qca,ddr-wb-channels = <&ddr_ctrl 3>, <&ddr_ctrl 2>,
52 #qca,ddr-wb-channel-cells = <1>;
/linux-4.4.14/drivers/staging/media/davinci_vpfe/
Ddm365_ipipe_hw.c612 void ipipe_set_wb_regs(void __iomem *base_addr, struct vpfe_ipipe_wb *wb) in ipipe_set_wb_regs() argument
618 regw_ip(base_addr, wb->ofst_r & WB_OFFSET_MASK, WB2_OFT_R); in ipipe_set_wb_regs()
619 regw_ip(base_addr, wb->ofst_gr & WB_OFFSET_MASK, WB2_OFT_GR); in ipipe_set_wb_regs()
620 regw_ip(base_addr, wb->ofst_gb & WB_OFFSET_MASK, WB2_OFT_GB); in ipipe_set_wb_regs()
621 regw_ip(base_addr, wb->ofst_b & WB_OFFSET_MASK, WB2_OFT_B); in ipipe_set_wb_regs()
624 val = IPIPE_U13Q9(wb->gain_r.decimal, wb->gain_r.integer); in ipipe_set_wb_regs()
627 val = IPIPE_U13Q9(wb->gain_gr.decimal, wb->gain_gr.integer); in ipipe_set_wb_regs()
630 val = IPIPE_U13Q9(wb->gain_gb.decimal, wb->gain_gb.integer); in ipipe_set_wb_regs()
633 val = IPIPE_U13Q9(wb->gain_b.decimal, wb->gain_b.integer); in ipipe_set_wb_regs()
Ddm365_ipipe_hw.h556 void ipipe_set_wb_regs(void __iomem *base_addr, struct vpfe_ipipe_wb *wb);
/linux-4.4.14/arch/arm/boot/dts/
Dexynos4x12.dtsi304 samsung,isp-wb;
312 samsung,isp-wb;
320 samsung,isp-wb;
321 samsung,lcd-wb;
330 samsung,isp-wb;
331 samsung,lcd-wb;
Dexynos4210.dtsi212 samsung,lcd-wb;
219 samsung,lcd-wb;
Ds5pv210.dtsi627 samsung,lcd-wb;
/linux-4.4.14/drivers/crypto/sunxi-ss/
Dsun4i-ss-hash.c352 u32 wb = 0; in sun4i_hash_final() local
380 wb = *(u32 *)(op->buf + nwait * 4); in sun4i_hash_final()
381 wb &= (0xFFFFFFFF >> (4 - nbw) * 8); in sun4i_hash_final()
386 wb |= ((1 << 7) << (nbw * 8)); in sun4i_hash_final()
387 bf[j++] = wb; in sun4i_hash_final()
/linux-4.4.14/drivers/media/i2c/m5mols/
Dm5mols_controls.c281 static const unsigned short wb[][2] = { in m5mols_set_white_balance() local
296 for (i = 0; i < ARRAY_SIZE(wb); i++) { in m5mols_set_white_balance()
298 if (wb[i][0] != val) in m5mols_set_white_balance()
302 "Setting white balance to: %#x\n", wb[i][0]); in m5mols_set_white_balance()
304 awb = wb[i][0] == V4L2_WHITE_BALANCE_AUTO; in m5mols_set_white_balance()
311 ret = m5mols_write(sd, AWB_MANUAL, wb[i][1]); in m5mols_set_white_balance()
/linux-4.4.14/drivers/net/ethernet/intel/igb/
De1000_82575.h107 } wb; /* writeback */ member
126 } wb; member
Digb.h311 return rx_desc->wb.upper.status_error & cpu_to_le32(stat_err_bits); in igb_test_staterr()
Digb_main.c523 staterr = le32_to_cpu(rx_desc->wb.upper.status_error); in igb_dump()
6428 if (!(eop_desc->wb.status & cpu_to_le32(E1000_TXD_STAT_DD))) in igb_clean_tx_irq()
6533 tx_buffer->next_to_watch->wb.status); in igb_clean_tx_irq()
6652 unsigned int size = le16_to_cpu(rx_desc->wb.upper.length); in igb_add_rx_frag()
6795 le32_to_cpu(rx_desc->wb.upper.status_error)); in igb_rx_checksum()
6804 le32_to_cpu(rx_desc->wb.lower.hi_dword.rss), in igb_rx_hash()
6900 vid = be16_to_cpu(rx_desc->wb.upper.vlan); in igb_process_skb_fields()
6902 vid = le16_to_cpu(rx_desc->wb.upper.vlan); in igb_process_skb_fields()
6930 if (!rx_desc->wb.upper.status_error) in igb_clean_rx_irq()
7063 rx_desc->wb.upper.status_error = 0; in igb_alloc_rx_buffers()
/linux-4.4.14/drivers/net/ethernet/intel/ixgbevf/
Ddefines.h212 } wb; member
243 } wb; /* writeback */ member
Dixgbevf.h344 return rx_desc->wb.upper.status_error & cpu_to_le32(stat_err_bits); in ixgbevf_test_staterr()
Dixgbevf_main.c318 if (!(eop_desc->wb.status & cpu_to_le32(IXGBE_TXD_STAT_DD))) in ixgbevf_clean_tx_irq()
408 eop_desc, (eop_desc ? eop_desc->wb.status : 0), in ixgbevf_clean_tx_irq()
475 rss_type = le16_to_cpu(rx_desc->wb.lower.lo_dword.hs_rss.pkt_info) & in ixgbevf_rx_hash()
481 skb_set_hash(skb, le32_to_cpu(rx_desc->wb.lower.hi_dword.rss), in ixgbevf_rx_hash()
539 u16 vid = le16_to_cpu(rx_desc->wb.upper.vlan); in ixgbevf_process_skb_fields()
776 unsigned int size = le16_to_cpu(rx_desc->wb.upper.length); in ixgbevf_add_rx_frag()
/linux-4.4.14/drivers/char/
Dps3flash.c270 int wb; in ps3flash_kernel_write() local
277 wb = ps3flash_writeback(ps3flash_dev); in ps3flash_kernel_write()
278 if (wb) in ps3flash_kernel_write()
279 return wb; in ps3flash_kernel_write()
/linux-4.4.14/drivers/net/ethernet/intel/igbvf/
Dvf.h87 } wb; /* writeback */ member
104 } wb; member
Dnetdev.c281 staterr = le32_to_cpu(rx_desc->wb.upper.status_error); in igbvf_clean_rx_irq()
296 hlen = (le16_to_cpu(rx_desc->wb.lower.lo_dword.hs_rss.hdr_info) in igbvf_clean_rx_irq()
302 length = le16_to_cpu(rx_desc->wb.upper.length); in igbvf_clean_rx_irq()
376 rx_desc->wb.upper.vlan); in igbvf_clean_rx_irq()
379 rx_desc->wb.upper.status_error = 0; in igbvf_clean_rx_irq()
391 staterr = le32_to_cpu(rx_desc->wb.upper.status_error); in igbvf_clean_rx_irq()
815 if (!(eop_desc->wb.status & cpu_to_le32(E1000_TXD_STAT_DD))) in igbvf_clean_tx_irq()
839 tx_desc->wb.status = 0; in igbvf_clean_tx_irq()
/linux-4.4.14/drivers/media/i2c/s5c73m3/
Ds5c73m3-ctrls.c152 static const unsigned short wb[][2] = { in s5c73m3_set_white_balance() local
162 for (i = 0; i < ARRAY_SIZE(wb); i++) { in s5c73m3_set_white_balance()
163 if (wb[i][0] != val) in s5c73m3_set_white_balance()
170 return s5c73m3_isp_command(state, COMM_AWB_MODE, wb[i][1]); in s5c73m3_set_white_balance()
/linux-4.4.14/arch/alpha/kernel/
Dsmc37c669.c944 #define wb( _x_, _y_ ) outb( _y_, (unsigned int)((unsigned long)_x_) ) macro
1975 wb( &SMC37c669->index_port, SMC37c669_CONFIG_ON_KEY ); in SMC37c669_config_mode()
1976 wb( &SMC37c669->index_port, SMC37c669_CONFIG_ON_KEY ); in SMC37c669_config_mode()
1980 wb( &SMC37c669->index_port, SMC37c669_CONFIG_OFF_KEY ); in SMC37c669_config_mode()
2012 wb( &SMC37c669->index_port, index ); in SMC37c669_read_config()
2047 wb( &SMC37c669->index_port, index ); in SMC37c669_write_config()
2048 wb( &SMC37c669->data_port, data ); in SMC37c669_write_config()
/linux-4.4.14/drivers/mtd/nand/
Dnandsim.c849 struct weak_block *wb; in parse_weakblocks() local
868 wb = kzalloc(sizeof(*wb), GFP_KERNEL); in parse_weakblocks()
869 if (!wb) { in parse_weakblocks()
873 wb->erase_block_no = erase_block_no; in parse_weakblocks()
874 wb->max_erases = max_erases; in parse_weakblocks()
875 list_add(&wb->list, &weak_blocks); in parse_weakblocks()
882 struct weak_block *wb; in erase_error() local
884 list_for_each_entry(wb, &weak_blocks, list) in erase_error()
885 if (wb->erase_block_no == erase_block_no) { in erase_error()
886 if (wb->erases_done >= wb->max_erases) in erase_error()
[all …]
/linux-4.4.14/drivers/net/ethernet/intel/i40e/
Di40e_txrx.c463 qw = le64_to_cpu(rx_desc->wb.qword1.status_error_len); in i40e_fd_handle_status()
468 pf->fd_inv = le32_to_cpu(rx_desc->wb.qword0.hi_dword.fd_id); in i40e_fd_handle_status()
469 if ((rx_desc->wb.qword0.hi_dword.fd_id != 0) || in i40e_fd_handle_status()
487 if ((rx_desc->wb.qword0.hi_dword.fd_id == 0) && in i40e_fd_handle_status()
516 rx_desc->wb.qword0.hi_dword.fd_id); in i40e_fd_handle_status()
930 qw = le64_to_cpu(rx_desc->wb.qword1.status_error_len); in i40e_clean_programming_status()
1431 (rx_desc->wb.qword1.status_error_len & rss_mask) == rss_mask) in i40e_rx_hash()
1432 return le32_to_cpu(rx_desc->wb.qword0.hi_dword.rss); in i40e_rx_hash()
1495 qword = le64_to_cpu(rx_desc->wb.qword1.status_error_len); in i40e_clean_rx_irq_ps()
1627 ? le16_to_cpu(rx_desc->wb.qword0.lo_dword.l2tag1) in i40e_clean_rx_irq_ps()
[all …]
Di40e_fcoe.c633 xid = le32_to_cpu(rx_desc->wb.qword0.hi_dword.fcoe_param) & in i40e_fcoe_handle_status()
642 qw = le64_to_cpu(rx_desc->wb.qword1.status_error_len); in i40e_fcoe_handle_status()
705 qw = le64_to_cpu(rx_desc->wb.qword1.status_error_len); in i40e_fcoe_handle_offload()
744 xid = le16_to_cpu(rx_desc->wb.qword0.lo_dword.mirr_fcoe.fcoe_ctx_id); in i40e_fcoe_handle_offload()
759 ddp->len = le32_to_cpu(rx_desc->wb.qword0.hi_dword.fcoe_param); in i40e_fcoe_handle_offload()
Di40e_type.h594 } wb; /* writeback */ member
643 } wb; /* writeback */ member
/linux-4.4.14/drivers/net/ethernet/intel/i40evf/
Di40e_txrx.c904 (rx_desc->wb.qword1.status_error_len & rss_mask) == rss_mask) in i40e_rx_hash()
905 return le32_to_cpu(rx_desc->wb.qword0.hi_dword.rss); in i40e_rx_hash()
965 qword = le64_to_cpu(rx_desc->wb.qword1.status_error_len); in i40e_clean_rx_irq_ps()
1085 ? le16_to_cpu(rx_desc->wb.qword0.lo_dword.l2tag1) in i40e_clean_rx_irq_ps()
1096 rx_desc->wb.qword1.status_error_len = 0; in i40e_clean_rx_irq_ps()
1141 qword = le64_to_cpu(rx_desc->wb.qword1.status_error_len); in i40e_clean_rx_irq_1buf()
1203 ? le16_to_cpu(rx_desc->wb.qword0.lo_dword.l2tag1) in i40e_clean_rx_irq_1buf()
1207 rx_desc->wb.qword1.status_error_len = 0; in i40e_clean_rx_irq_1buf()
Di40e_type.h581 } wb; /* writeback */ member
630 } wb; /* writeback */ member
/linux-4.4.14/drivers/net/ethernet/broadcom/bnx2x/
Dbnx2x_init_ops.h70 u8 wb) in bnx2x_write_big_buf() argument
76 else if (wb && CHIP_IS_E1(bp)) in bnx2x_write_big_buf()
85 u32 len, u8 wb) in bnx2x_init_fill() argument
96 bnx2x_write_big_buf(bp, addr + i*4, cur_len, wb); in bnx2x_init_fill()
/linux-4.4.14/drivers/net/ethernet/intel/e1000e/
Dhw.h229 } wb; /* writeback */ member
265 } wb; /* writeback */ member
Dnetdev.c383 le32_to_cpu(rx_desc_ps->wb.middle.status_error); in e1000e_dump()
449 staterr = le32_to_cpu(rx_desc->wb.upper.status_error); in e1000e_dump()
941 staterr = le32_to_cpu(rx_desc->wb.upper.status_error); in e1000_clean_rx_irq()
971 length = le16_to_cpu(rx_desc->wb.upper.length); in e1000_clean_rx_irq()
1040 e1000_rx_hash(netdev, rx_desc->wb.lower.hi_dword.rss, skb); in e1000_clean_rx_irq()
1043 rx_desc->wb.upper.vlan); in e1000_clean_rx_irq()
1046 rx_desc->wb.upper.status_error &= cpu_to_le32(~0xFF); in e1000_clean_rx_irq()
1059 staterr = le32_to_cpu(rx_desc->wb.upper.status_error); in e1000_clean_rx_irq()
1327 staterr = le32_to_cpu(rx_desc->wb.middle.status_error); in e1000_clean_rx_irq_ps()
1372 length = le16_to_cpu(rx_desc->wb.middle.length0); in e1000_clean_rx_irq_ps()
[all …]
/linux-4.4.14/Documentation/filesystems/
Daffs.txt192 do care about the case. Example (with /wb being an affs mounted fs):
193 rm /wb/WRONGCASE
195 rm /wb/WR*
/linux-4.4.14/arch/unicore32/boot/compressed/
Dhead.S92 mov r0, #0x1c @ en icache and wb dcache
/linux-4.4.14/Documentation/devicetree/bindings/media/
Dsamsung-fimc.txt87 - samsung,isp-wb: this property must be present if the IP block has the ISP
89 - samsung,lcd-wb: this property must be present if the IP block has the LCD
/linux-4.4.14/drivers/media/dvb-frontends/
Ddib3000mb.c55 u8 wb[] = { ((reg >> 8) | 0x80) & 0xff, reg & 0xff }; in dib3000_read_reg() local
58 { .addr = state->config.demod_address, .flags = 0, .buf = wb, .len = 2 }, in dib3000_read_reg()
Ddib3000mc.c52 u8 wb[2] = { (reg >> 8) | 0x80, reg & 0xff }; in dib3000mc_read_word() local
55 { .addr = state->i2c_addr >> 1, .flags = 0, .buf = wb, .len = 2 }, in dib3000mc_read_word()
/linux-4.4.14/drivers/media/usb/dvb-usb/
Ddib0700_devices.c2206 u8 wb[4] = { 0xc >> 8, 0xc & 0xff, 0, 0 }; in dib01x0_pmu_update() local
2209 {.addr = 0x1e >> 1, .flags = 0, .buf = wb, .len = 2}, in dib01x0_pmu_update()
2235 wb[2] = (data[index_data + 1] >> 8) & 0xff; in dib01x0_pmu_update()
2236 wb[3] = (data[index_data + 1]) & 0xff; in dib01x0_pmu_update()
2239 wb[0] = (data[index_data] >> 8) & 0xff; in dib01x0_pmu_update()
2240 wb[1] = (data[index_data]) & 0xff; in dib01x0_pmu_update()
2244 wb[2] |= rb[0]; in dib01x0_pmu_update()
2245 wb[3] |= rb[1] & ~(3 << 4); in dib01x0_pmu_update()
2248 wb[0] = (data[index_data] >> 8)&0xff; in dib01x0_pmu_update()
2249 wb[1] = (data[index_data])&0xff; in dib01x0_pmu_update()
/linux-4.4.14/drivers/gpu/drm/msm/mdp/mdp5/
Dmdp5.xml.h1393 case 0: return (mdp5_cfg->wb.base[0]); in __offset_WB()
1394 case 1: return (mdp5_cfg->wb.base[1]); in __offset_WB()
1395 case 2: return (mdp5_cfg->wb.base[2]); in __offset_WB()
1396 case 3: return (mdp5_cfg->wb.base[3]); in __offset_WB()
1397 case 4: return (mdp5_cfg->wb.base[4]); in __offset_WB()
/linux-4.4.14/drivers/net/ethernet/intel/ixgbe/
Dixgbe_fcoe.c460 ddp->len = le32_to_cpu(rx_desc->wb.lower.hi_dword.rss); in ixgbe_fcoe_ddp()
474 ddp->len = le32_to_cpu(rx_desc->wb.lower.hi_dword.rss); in ixgbe_fcoe_ddp()
Dixgbe_main.c785 staterr = le32_to_cpu(rx_desc->wb.upper.status_error); in ixgbe_dump()
1120 if (!(eop_desc->wb.status & cpu_to_le32(IXGBE_TXD_STAT_DD))) in ixgbe_clean_tx_irq()
1400 rss_type = le16_to_cpu(rx_desc->wb.lower.lo_dword.hs_rss.pkt_info) & in ixgbe_rx_hash()
1406 skb_set_hash(skb, le32_to_cpu(rx_desc->wb.lower.hi_dword.rss), in ixgbe_rx_hash()
1422 __le16 pkt_info = rx_desc->wb.lower.lo_dword.hs_rss.pkt_info; in ixgbe_rx_is_fcoe()
1441 __le16 pkt_info = rx_desc->wb.lower.lo_dword.hs_rss.pkt_info; in ixgbe_rx_checksum()
1442 __le16 hdr_info = rx_desc->wb.lower.lo_dword.hs_rss.hdr_info; in ixgbe_rx_checksum()
1573 rx_desc->wb.upper.status_error = 0; in ixgbe_alloc_rx_buffers()
1650 u16 vid = le16_to_cpu(rx_desc->wb.upper.vlan); in ixgbe_process_skb_fields()
1693 __le32 rsc_enabled = rx_desc->wb.lower.lo_dword.data & in ixgbe_is_non_eop()
[all …]
Dixgbe.h556 return rx_desc->wb.upper.status_error & cpu_to_le32(stat_err_bits); in ixgbe_test_staterr()
Dixgbe_type.h2690 } wb; member
2721 } wb; /* writeback */ member
/linux-4.4.14/drivers/gpu/drm/exynos/
Dexynos_drm_fimc.c228 static void fimc_set_type_ctrl(struct fimc_context *ctx, enum fimc_wb wb) in fimc_set_type_ctrl() argument
232 DRM_DEBUG_KMS("wb[%d]\n", wb); in fimc_set_type_ctrl()
242 switch (wb) { in fimc_set_type_ctrl()
/linux-4.4.14/fs/fuse/
Dfile.c1455 dec_wb_stat(&bdi->wb, WB_WRITEBACK); in fuse_writepage_finish()
1457 wb_writeout_inc(&bdi->wb); in fuse_writepage_finish()
1645 inc_wb_stat(&inode_to_bdi(inode)->wb, WB_WRITEBACK); in fuse_writepage_locked()
1759 dec_wb_stat(&bdi->wb, WB_WRITEBACK); in fuse_writepage_in_flight()
1761 wb_writeout_inc(&bdi->wb); in fuse_writepage_in_flight()
1858 inc_wb_stat(&inode_to_bdi(inode)->wb, WB_WRITEBACK); in fuse_writepages_fill()
/linux-4.4.14/arch/mips/include/asm/sn/sn0/
Dhubio.h892 wb: 1, /* 15: writeback pending. */ member
/linux-4.4.14/Documentation/x86/
Dpat.txt157 will first make a region uc, wc or wt and switch it back to wb after use.
/linux-4.4.14/fs/nfs/
Dinternal.h630 inc_wb_stat(&inode_to_bdi(inode)->wb, WB_RECLAIMABLE); in nfs_mark_page_unstable()
Dwrite.c875 dec_wb_stat(&inode_to_bdi(page_file_mapping(page)->host)->wb, in nfs_clear_page_commit()
/linux-4.4.14/fs/f2fs/
Dsegment.h699 if (sbi->sb->s_bdi->wb.dirty_exceeded) in nr_pages_to_skip()
Dnode.c56 if (sbi->sb->s_bdi->wb.dirty_exceeded) in available_free_memory()
73 if (sbi->sb->s_bdi->wb.dirty_exceeded) in available_free_memory()
/linux-4.4.14/drivers/net/ethernet/intel/e1000/
De1000_hw.h553 } wb; /* writeback */ member
584 } wb; /* writeback */ member
/linux-4.4.14/block/
Dblk-core.c76 clear_wb_congested(rl->q->backing_dev_info.wb.congested, sync); in blk_clear_congested()
87 set_wb_congested(rl->q->backing_dev_info.wb.congested, sync); in blk_set_congested()
/linux-4.4.14/fs/gfs2/
Dsuper.c751 if (bdi->wb.dirty_exceeded) in gfs2_write_inode()
/linux-4.4.14/drivers/md/
Ddm.c2155 r = md->queue->backing_dev_info.wb.state & in dm_any_congested()