Home
last modified time | relevance | path

Searched refs:ndw (Results 1 – 15 of 15) sorted by relevance

/linux-4.4.14/drivers/gpu/drm/radeon/
Dni_dma.c320 unsigned ndw; in cayman_dma_vm_copy_pages() local
323 ndw = count * 2; in cayman_dma_vm_copy_pages()
324 if (ndw > 0xFFFFE) in cayman_dma_vm_copy_pages()
325 ndw = 0xFFFFE; in cayman_dma_vm_copy_pages()
328 0, 0, ndw); in cayman_dma_vm_copy_pages()
334 pe += ndw * 4; in cayman_dma_vm_copy_pages()
335 src += ndw * 4; in cayman_dma_vm_copy_pages()
336 count -= ndw / 2; in cayman_dma_vm_copy_pages()
360 unsigned ndw; in cayman_dma_vm_write_pages() local
363 ndw = count * 2; in cayman_dma_vm_write_pages()
[all …]
Dsi_dma.c112 unsigned ndw; in si_dma_vm_write_pages() local
115 ndw = count * 2; in si_dma_vm_write_pages()
116 if (ndw > 0xFFFFE) in si_dma_vm_write_pages()
117 ndw = 0xFFFFE; in si_dma_vm_write_pages()
120 ib->ptr[ib->length_dw++] = DMA_PACKET(DMA_PACKET_WRITE, 0, 0, 0, ndw); in si_dma_vm_write_pages()
123 for (; ndw > 0; ndw -= 2, --count, pe += 8) { in si_dma_vm_write_pages()
159 unsigned ndw; in si_dma_vm_set_pages() local
162 ndw = count * 2; in si_dma_vm_set_pages()
163 if (ndw > 0xFFFFE) in si_dma_vm_set_pages()
164 ndw = 0xFFFFE; in si_dma_vm_set_pages()
[all …]
Dradeon_ring.c104 int radeon_ring_alloc(struct radeon_device *rdev, struct radeon_ring *ring, unsigned ndw) in radeon_ring_alloc() argument
109 if (ndw > (ring->ring_size / 4)) in radeon_ring_alloc()
114 ndw = (ndw + ring->align_mask) & ~ring->align_mask; in radeon_ring_alloc()
115 while (ndw > (ring->ring_free_dw - 1)) { in radeon_ring_alloc()
117 if (ndw < ring->ring_free_dw) { in radeon_ring_alloc()
124 ring->count_dw = ndw; in radeon_ring_alloc()
140 int radeon_ring_lock(struct radeon_device *rdev, struct radeon_ring *ring, unsigned ndw) in radeon_ring_lock() argument
145 r = radeon_ring_alloc(rdev, ring, ndw); in radeon_ring_lock()
Dcik_sdma.c843 unsigned ndw; in cik_sdma_vm_write_pages() local
846 ndw = count * 2; in cik_sdma_vm_write_pages()
847 if (ndw > 0xFFFFE) in cik_sdma_vm_write_pages()
848 ndw = 0xFFFFE; in cik_sdma_vm_write_pages()
855 ib->ptr[ib->length_dw++] = ndw; in cik_sdma_vm_write_pages()
856 for (; ndw > 0; ndw -= 2, --count, pe += 8) { in cik_sdma_vm_write_pages()
892 unsigned ndw; in cik_sdma_vm_set_pages() local
895 ndw = count; in cik_sdma_vm_set_pages()
896 if (ndw > 0x7FFFF) in cik_sdma_vm_set_pages()
897 ndw = 0x7FFFF; in cik_sdma_vm_set_pages()
[all …]
Dradeon_vm.c646 unsigned count = 0, pt_idx, ndw; in radeon_vm_update_page_directory() local
651 ndw = 64; in radeon_vm_update_page_directory()
654 ndw += vm->max_pde_used * 6; in radeon_vm_update_page_directory()
657 if (ndw > 0xfffff) in radeon_vm_update_page_directory()
660 r = radeon_ib_get(rdev, R600_RING_TYPE_DMA_INDEX, &ib, NULL, ndw * 4); in radeon_vm_update_page_directory()
704 WARN_ON(ib.length_dw > ndw); in radeon_vm_update_page_directory()
916 unsigned nptes, ncmds, ndw; in radeon_vm_bo_update() local
972 ndw = 64; in radeon_vm_bo_update()
977 ndw += ncmds * 7; in radeon_vm_bo_update()
981 ndw += ncmds * 4; in radeon_vm_bo_update()
[all …]
Dr100.c894 unsigned ndw; in r100_copy_blit() local
906 ndw = 64 + (10 * num_loops); in r100_copy_blit()
907 r = radeon_ring_lock(rdev, ring, ndw); in r100_copy_blit()
909 DRM_ERROR("radeon: moving bo (%d) asking for %u dw.\n", r, ndw); in r100_copy_blit()
Dradeon.h1032 int radeon_ring_alloc(struct radeon_device *rdev, struct radeon_ring *cp, unsigned ndw);
1033 int radeon_ring_lock(struct radeon_device *rdev, struct radeon_ring *cp, unsigned ndw);
/linux-4.4.14/drivers/gpu/drm/amd/amdgpu/
Damdgpu_ring.c83 int amdgpu_ring_alloc(struct amdgpu_ring *ring, unsigned ndw) in amdgpu_ring_alloc() argument
88 if (ndw > (ring->ring_size / 4)) in amdgpu_ring_alloc()
93 ndw = (ndw + ring->align_mask) & ~ring->align_mask; in amdgpu_ring_alloc()
94 while (ndw > (ring->ring_free_dw - 1)) { in amdgpu_ring_alloc()
96 if (ndw < ring->ring_free_dw) { in amdgpu_ring_alloc()
103 ring->count_dw = ndw; in amdgpu_ring_alloc()
119 int amdgpu_ring_lock(struct amdgpu_ring *ring, unsigned ndw) in amdgpu_ring_lock() argument
124 r = amdgpu_ring_alloc(ring, ndw); in amdgpu_ring_lock()
Damdgpu_vm.c435 unsigned count = 0, pt_idx, ndw; in amdgpu_vm_update_page_directory() local
442 ndw = 64; in amdgpu_vm_update_page_directory()
445 ndw += vm->max_pde_used * 6; in amdgpu_vm_update_page_directory()
448 if (ndw > 0xfffff) in amdgpu_vm_update_page_directory()
455 r = amdgpu_ib_get(ring, NULL, ndw * 4, ib); in amdgpu_vm_update_page_directory()
500 WARN_ON(ib->length_dw > ndw); in amdgpu_vm_update_page_directory()
710 unsigned nptes, ncmds, ndw; in amdgpu_vm_bo_update_mapping() local
735 ndw = 64; in amdgpu_vm_bo_update_mapping()
739 ndw += ncmds * 7; in amdgpu_vm_bo_update_mapping()
743 ndw += ncmds * 4; in amdgpu_vm_bo_update_mapping()
[all …]
Dcik_sdma.c746 unsigned ndw; in cik_sdma_vm_write_pte() local
749 ndw = count * 2; in cik_sdma_vm_write_pte()
750 if (ndw > 0xFFFFE) in cik_sdma_vm_write_pte()
751 ndw = 0xFFFFE; in cik_sdma_vm_write_pte()
758 ib->ptr[ib->length_dw++] = ndw; in cik_sdma_vm_write_pte()
759 for (; ndw > 0; ndw -= 2, --count, pe += 8) { in cik_sdma_vm_write_pte()
794 unsigned ndw; in cik_sdma_vm_set_pte_pde() local
797 ndw = count; in cik_sdma_vm_set_pte_pde()
798 if (ndw > 0x7FFFF) in cik_sdma_vm_set_pte_pde()
799 ndw = 0x7FFFF; in cik_sdma_vm_set_pte_pde()
[all …]
Dsdma_v2_4.c805 unsigned ndw; in sdma_v2_4_vm_write_pte() local
808 ndw = count * 2; in sdma_v2_4_vm_write_pte()
809 if (ndw > 0xFFFFE) in sdma_v2_4_vm_write_pte()
810 ndw = 0xFFFFE; in sdma_v2_4_vm_write_pte()
817 ib->ptr[ib->length_dw++] = ndw; in sdma_v2_4_vm_write_pte()
818 for (; ndw > 0; ndw -= 2, --count, pe += 8) { in sdma_v2_4_vm_write_pte()
853 unsigned ndw; in sdma_v2_4_vm_set_pte_pde() local
856 ndw = count; in sdma_v2_4_vm_set_pte_pde()
857 if (ndw > 0x7FFFF) in sdma_v2_4_vm_set_pte_pde()
858 ndw = 0x7FFFF; in sdma_v2_4_vm_set_pte_pde()
[all …]
Dsdma_v3_0.c954 unsigned ndw; in sdma_v3_0_vm_write_pte() local
957 ndw = count * 2; in sdma_v3_0_vm_write_pte()
958 if (ndw > 0xFFFFE) in sdma_v3_0_vm_write_pte()
959 ndw = 0xFFFFE; in sdma_v3_0_vm_write_pte()
966 ib->ptr[ib->length_dw++] = ndw; in sdma_v3_0_vm_write_pte()
967 for (; ndw > 0; ndw -= 2, --count, pe += 8) { in sdma_v3_0_vm_write_pte()
1002 unsigned ndw; in sdma_v3_0_vm_set_pte_pde() local
1005 ndw = count; in sdma_v3_0_vm_set_pte_pde()
1006 if (ndw > 0x7FFFF) in sdma_v3_0_vm_set_pte_pde()
1007 ndw = 0x7FFFF; in sdma_v3_0_vm_set_pte_pde()
[all …]
Damdgpu.h1218 int amdgpu_ring_alloc(struct amdgpu_ring *ring, unsigned ndw);
1219 int amdgpu_ring_lock(struct amdgpu_ring *ring, unsigned ndw);
/linux-4.4.14/Documentation/RCU/
Dtrace.txt383 0!np=26111 qsp=29 rpq=5386 cbr=1 cng=570 gpc=3674 gps=577 nn=15903 ndw=0
384 1!np=28913 qsp=35 rpq=6097 cbr=1 cng=448 gpc=3700 gps=554 nn=18113 ndw=0
385 2!np=32740 qsp=37 rpq=6202 cbr=0 cng=476 gpc=4627 gps=546 nn=20889 ndw=0
386 3 np=23679 qsp=22 rpq=5044 cbr=1 cng=415 gpc=3403 gps=347 nn=14469 ndw=0
387 4!np=30714 qsp=4 rpq=5574 cbr=0 cng=528 gpc=3931 gps=639 nn=20042 ndw=0
388 5 np=28910 qsp=2 rpq=5246 cbr=0 cng=428 gpc=4105 gps=709 nn=18422 ndw=0
389 6!np=38648 qsp=5 rpq=7076 cbr=0 cng=840 gpc=4072 gps=961 nn=25699 ndw=0
390 7 np=37275 qsp=2 rpq=6873 cbr=0 cng=868 gpc=3416 gps=971 nn=25147 ndw=0
419 o "ndw" is the number of times that a wakeup of an rcuo
/linux-4.4.14/kernel/rcu/
Dtree_plugin.h2298 int ndw; in do_nocb_deferred_wakeup() local
2302 ndw = READ_ONCE(rdp->nocb_defer_wakeup); in do_nocb_deferred_wakeup()
2304 wake_nocb_leader(rdp, ndw == RCU_NOGP_WAKE_FORCE); in do_nocb_deferred_wakeup()