| /linux-4.4.14/drivers/gpu/drm/radeon/ |
| D | si_dma.c | 79 ib->ptr[ib->length_dw++] = DMA_PACKET(DMA_PACKET_COPY, in si_dma_vm_copy_pages() 81 ib->ptr[ib->length_dw++] = lower_32_bits(pe); in si_dma_vm_copy_pages() 82 ib->ptr[ib->length_dw++] = lower_32_bits(src); in si_dma_vm_copy_pages() 83 ib->ptr[ib->length_dw++] = upper_32_bits(pe) & 0xff; in si_dma_vm_copy_pages() 84 ib->ptr[ib->length_dw++] = upper_32_bits(src) & 0xff; in si_dma_vm_copy_pages() 120 ib->ptr[ib->length_dw++] = DMA_PACKET(DMA_PACKET_WRITE, 0, 0, 0, ndw); in si_dma_vm_write_pages() 121 ib->ptr[ib->length_dw++] = pe; in si_dma_vm_write_pages() 122 ib->ptr[ib->length_dw++] = upper_32_bits(pe) & 0xff; in si_dma_vm_write_pages() 133 ib->ptr[ib->length_dw++] = value; in si_dma_vm_write_pages() 134 ib->ptr[ib->length_dw++] = upper_32_bits(value); in si_dma_vm_write_pages() [all …]
|
| D | radeon_vce.c | 363 ib.length_dw = 0; in radeon_vce_get_create_msg() 364 ib.ptr[ib.length_dw++] = cpu_to_le32(0x0000000c); /* len */ in radeon_vce_get_create_msg() 365 ib.ptr[ib.length_dw++] = cpu_to_le32(0x00000001); /* session cmd */ in radeon_vce_get_create_msg() 366 ib.ptr[ib.length_dw++] = cpu_to_le32(handle); in radeon_vce_get_create_msg() 368 ib.ptr[ib.length_dw++] = cpu_to_le32(0x00000030); /* len */ in radeon_vce_get_create_msg() 369 ib.ptr[ib.length_dw++] = cpu_to_le32(0x01000001); /* create cmd */ in radeon_vce_get_create_msg() 370 ib.ptr[ib.length_dw++] = cpu_to_le32(0x00000000); in radeon_vce_get_create_msg() 371 ib.ptr[ib.length_dw++] = cpu_to_le32(0x00000042); in radeon_vce_get_create_msg() 372 ib.ptr[ib.length_dw++] = cpu_to_le32(0x0000000a); in radeon_vce_get_create_msg() 373 ib.ptr[ib.length_dw++] = cpu_to_le32(0x00000001); in radeon_vce_get_create_msg() [all …]
|
| D | ni_dma.c | 146 radeon_ring_write(ring, (ib->length_dw << 12) | (upper_32_bits(ib->gpu_addr) & 0xFF)); in cayman_dma_ring_ib_execute() 327 ib->ptr[ib->length_dw++] = DMA_PACKET(DMA_PACKET_COPY, in cayman_dma_vm_copy_pages() 329 ib->ptr[ib->length_dw++] = lower_32_bits(pe); in cayman_dma_vm_copy_pages() 330 ib->ptr[ib->length_dw++] = lower_32_bits(src); in cayman_dma_vm_copy_pages() 331 ib->ptr[ib->length_dw++] = upper_32_bits(pe) & 0xff; in cayman_dma_vm_copy_pages() 332 ib->ptr[ib->length_dw++] = upper_32_bits(src) & 0xff; in cayman_dma_vm_copy_pages() 368 ib->ptr[ib->length_dw++] = DMA_PACKET(DMA_PACKET_WRITE, in cayman_dma_vm_write_pages() 370 ib->ptr[ib->length_dw++] = pe; in cayman_dma_vm_write_pages() 371 ib->ptr[ib->length_dw++] = upper_32_bits(pe) & 0xff; in cayman_dma_vm_write_pages() 382 ib->ptr[ib->length_dw++] = value; in cayman_dma_vm_write_pages() [all …]
|
| D | radeon_cs.c | 90 p->nrelocs = chunk->length_dw / 4; in radeon_cs_parser_relocs() 298 p->chunks[i].length_dw = user_chunk.length_dw; in radeon_cs_parser_init() 305 if (p->chunks[i].length_dw == 0) in radeon_cs_parser_init() 311 if (p->chunks[i].length_dw == 0) in radeon_cs_parser_init() 317 if (p->chunks[i].length_dw == 0) in radeon_cs_parser_init() 321 size = p->chunks[i].length_dw; in radeon_cs_parser_init() 342 if (p->chunks[i].length_dw > 1) in radeon_cs_parser_init() 344 if (p->chunks[i].length_dw > 2) in radeon_cs_parser_init() 535 if (parser->const_ib.length_dw) { in radeon_cs_ib_vm_chunk() 601 if (ib_chunk->length_dw > RADEON_IB_VM_MAX_SIZE) { in radeon_cs_ib_fill() [all …]
|
| D | cik_sdma.c | 157 radeon_ring_write(ring, ib->length_dw); in cik_sdma_ring_ib_execute() 732 ib.length_dw = 5; in cik_sdma_ib_test() 808 ib->ptr[ib->length_dw++] = SDMA_PACKET(SDMA_OPCODE_COPY, in cik_sdma_vm_copy_pages() 810 ib->ptr[ib->length_dw++] = bytes; in cik_sdma_vm_copy_pages() 811 ib->ptr[ib->length_dw++] = 0; /* src/dst endian swap */ in cik_sdma_vm_copy_pages() 812 ib->ptr[ib->length_dw++] = lower_32_bits(src); in cik_sdma_vm_copy_pages() 813 ib->ptr[ib->length_dw++] = upper_32_bits(src); in cik_sdma_vm_copy_pages() 814 ib->ptr[ib->length_dw++] = lower_32_bits(pe); in cik_sdma_vm_copy_pages() 815 ib->ptr[ib->length_dw++] = upper_32_bits(pe); in cik_sdma_vm_copy_pages() 851 ib->ptr[ib->length_dw++] = SDMA_PACKET(SDMA_OPCODE_WRITE, in cik_sdma_vm_write_pages() [all …]
|
| D | radeon_uvd.c | 541 if (idx >= relocs_chunk->length_dw) { in radeon_uvd_cs_reloc() 543 idx, relocs_chunk->length_dw); in radeon_uvd_cs_reloc() 655 if (p->chunk_ib->length_dw % 16) { in radeon_uvd_cs_parse() 657 p->chunk_ib->length_dw); in radeon_uvd_cs_parse() 685 } while (p->idx < p->chunk_ib->length_dw); in radeon_uvd_cs_parse() 714 ib.length_dw = 16; in radeon_uvd_send_msg()
|
| D | radeon_vm.c | 410 ib.length_dw = 0; in radeon_vm_clear_bo() 414 WARN_ON(ib.length_dw > 64); in radeon_vm_clear_bo() 663 ib.length_dw = 0; in radeon_vm_update_page_directory() 700 if (ib.length_dw != 0) { in radeon_vm_update_page_directory() 704 WARN_ON(ib.length_dw > ndw); in radeon_vm_update_page_directory() 1001 ib.length_dw = 0; in radeon_vm_bo_update() 1019 WARN_ON(ib.length_dw > ndw); in radeon_vm_bo_update()
|
| D | r600_dma.c | 363 ib.length_dw = 4; in r600_dma_ib_test() 422 radeon_ring_write(ring, (ib->length_dw << 16) | (upper_32_bits(ib->gpu_addr) & 0xFF)); in r600_dma_ring_ib_execute()
|
| D | evergreen_dma.c | 90 radeon_ring_write(ring, (ib->length_dw << 12) | (upper_32_bits(ib->gpu_addr) & 0xFF)); in evergreen_dma_ring_ib_execute()
|
| D | r600_cs.c | 2319 } while (p->idx < p->chunk_ib->length_dw); in r600_cs_parse() 2321 for (r = 0; r < p->ib.length_dw; r++) { in r600_cs_parse() 2402 parser.ib.length_dw = ib_chunk->length_dw; in r600_cs_legacy() 2403 *l = parser.ib.length_dw; in r600_cs_legacy() 2404 if (copy_from_user(ib, ib_chunk->user_ptr, ib_chunk->length_dw * 4)) { in r600_cs_legacy() 2484 if (p->idx >= ib_chunk->length_dw) { in r600_dma_cs_parse() 2486 p->idx, ib_chunk->length_dw); in r600_dma_cs_parse() 2622 } while (p->idx < p->chunk_ib->length_dw); in r600_dma_cs_parse() 2624 for (r = 0; r < p->ib->length_dw; r++) { in r600_dma_cs_parse()
|
| D | radeon_ib.c | 128 if (!ib->length_dw || !ring->ready) { in radeon_ib_schedule()
|
| D | radeon_trace.h | 40 __entry->dw = p->chunk_ib->length_dw;
|
| D | evergreen_cs.c | 2724 } while (p->idx < p->chunk_ib->length_dw); in evergreen_cs_parse() 2726 for (r = 0; r < p->ib.length_dw; r++) { in evergreen_cs_parse() 2756 if (p->idx >= ib_chunk->length_dw) { in evergreen_dma_cs_parse() 2758 p->idx, ib_chunk->length_dw); in evergreen_dma_cs_parse() 3163 } while (p->idx < p->chunk_ib->length_dw); in evergreen_dma_cs_parse() 3165 for (r = 0; r < p->ib->length_dw; r++) { in evergreen_dma_cs_parse() 3478 } while (idx < ib->length_dw); in evergreen_ib_parse() 3580 } while (idx < ib->length_dw); in evergreen_dma_ib_parse()
|
| D | uvd_v1_0.c | 488 radeon_ring_write(ring, ib->length_dw); in uvd_v1_0_ib_execute()
|
| D | r100.c | 2074 } while (p->idx < p->chunk_ib->length_dw); in r100_cs_parse() 3698 radeon_ring_write(ring, ib->length_dw); in r100_ring_ib_execute() 3728 ib.length_dw = 8; in r100_ib_test()
|
| D | r300.c | 1317 } while (p->idx < p->chunk_ib->length_dw); in r300_cs_parse()
|
| D | radeon.h | 837 uint32_t length_dw; member 1066 uint32_t length_dw; member
|
| D | si.c | 3440 radeon_ring_write(ring, ib->length_dw | (vm_id << 24)); in si_ring_ib_execute() 4769 for (i = 0; i < ib->length_dw; i++) { in si_ib_parse() 4777 } while (idx < ib->length_dw); in si_ib_parse()
|
| D | ni.c | 1457 radeon_ring_write(ring, ib->length_dw | (vm_id << 24)); in cayman_ring_ib_execute()
|
| D | r600.c | 3353 radeon_ring_write(ring, ib->length_dw); in r600_ring_ib_execute() 3378 ib.length_dw = 3; in r600_ib_test()
|
| D | cik.c | 4173 control |= ib->length_dw | (vm_id << 24); in cik_ring_ib_execute() 4214 ib.length_dw = 3; in cik_ib_test()
|
| D | evergreen.c | 3050 radeon_ring_write(ring, ib->length_dw); in evergreen_ring_ib_execute()
|
| /linux-4.4.14/drivers/gpu/drm/amd/amdgpu/ |
| D | amdgpu_vce.c | 391 ib->length_dw = 0; in amdgpu_vce_get_create_msg() 392 ib->ptr[ib->length_dw++] = 0x0000000c; /* len */ in amdgpu_vce_get_create_msg() 393 ib->ptr[ib->length_dw++] = 0x00000001; /* session cmd */ in amdgpu_vce_get_create_msg() 394 ib->ptr[ib->length_dw++] = handle; in amdgpu_vce_get_create_msg() 397 ib->ptr[ib->length_dw++] = 0x00000040; /* len */ in amdgpu_vce_get_create_msg() 399 ib->ptr[ib->length_dw++] = 0x00000030; /* len */ in amdgpu_vce_get_create_msg() 400 ib->ptr[ib->length_dw++] = 0x01000001; /* create cmd */ in amdgpu_vce_get_create_msg() 401 ib->ptr[ib->length_dw++] = 0x00000000; in amdgpu_vce_get_create_msg() 402 ib->ptr[ib->length_dw++] = 0x00000042; in amdgpu_vce_get_create_msg() 403 ib->ptr[ib->length_dw++] = 0x0000000a; in amdgpu_vce_get_create_msg() [all …]
|
| D | cik_sdma.c | 234 amdgpu_ring_write(ring, ib->length_dw); in cik_sdma_ring_emit_ib() 659 ib.length_dw = 5; in cik_sdma_ring_test_ib() 713 ib->ptr[ib->length_dw++] = SDMA_PACKET(SDMA_OPCODE_COPY, in cik_sdma_vm_copy_pte() 715 ib->ptr[ib->length_dw++] = bytes; in cik_sdma_vm_copy_pte() 716 ib->ptr[ib->length_dw++] = 0; /* src/dst endian swap */ in cik_sdma_vm_copy_pte() 717 ib->ptr[ib->length_dw++] = lower_32_bits(src); in cik_sdma_vm_copy_pte() 718 ib->ptr[ib->length_dw++] = upper_32_bits(src); in cik_sdma_vm_copy_pte() 719 ib->ptr[ib->length_dw++] = lower_32_bits(pe); in cik_sdma_vm_copy_pte() 720 ib->ptr[ib->length_dw++] = upper_32_bits(pe); in cik_sdma_vm_copy_pte() 754 ib->ptr[ib->length_dw++] = SDMA_PACKET(SDMA_OPCODE_WRITE, in cik_sdma_vm_write_pte() [all …]
|
| D | sdma_v2_4.c | 270 amdgpu_ring_write(ring, ib->length_dw); in sdma_v2_4_ring_emit_ib() 717 ib.length_dw = 8; in sdma_v2_4_ring_test_ib() 772 ib->ptr[ib->length_dw++] = SDMA_PKT_HEADER_OP(SDMA_OP_COPY) | in sdma_v2_4_vm_copy_pte() 774 ib->ptr[ib->length_dw++] = bytes; in sdma_v2_4_vm_copy_pte() 775 ib->ptr[ib->length_dw++] = 0; /* src/dst endian swap */ in sdma_v2_4_vm_copy_pte() 776 ib->ptr[ib->length_dw++] = lower_32_bits(src); in sdma_v2_4_vm_copy_pte() 777 ib->ptr[ib->length_dw++] = upper_32_bits(src); in sdma_v2_4_vm_copy_pte() 778 ib->ptr[ib->length_dw++] = lower_32_bits(pe); in sdma_v2_4_vm_copy_pte() 779 ib->ptr[ib->length_dw++] = upper_32_bits(pe); in sdma_v2_4_vm_copy_pte() 813 ib->ptr[ib->length_dw++] = SDMA_PKT_HEADER_OP(SDMA_OP_WRITE) | in sdma_v2_4_vm_write_pte() [all …]
|
| D | sdma_v3_0.c | 380 amdgpu_ring_write(ring, ib->length_dw); in sdma_v3_0_ring_emit_ib() 867 ib.length_dw = 8; in sdma_v3_0_ring_test_ib() 921 ib->ptr[ib->length_dw++] = SDMA_PKT_HEADER_OP(SDMA_OP_COPY) | in sdma_v3_0_vm_copy_pte() 923 ib->ptr[ib->length_dw++] = bytes; in sdma_v3_0_vm_copy_pte() 924 ib->ptr[ib->length_dw++] = 0; /* src/dst endian swap */ in sdma_v3_0_vm_copy_pte() 925 ib->ptr[ib->length_dw++] = lower_32_bits(src); in sdma_v3_0_vm_copy_pte() 926 ib->ptr[ib->length_dw++] = upper_32_bits(src); in sdma_v3_0_vm_copy_pte() 927 ib->ptr[ib->length_dw++] = lower_32_bits(pe); in sdma_v3_0_vm_copy_pte() 928 ib->ptr[ib->length_dw++] = upper_32_bits(pe); in sdma_v3_0_vm_copy_pte() 962 ib->ptr[ib->length_dw++] = SDMA_PKT_HEADER_OP(SDMA_OP_WRITE) | in sdma_v3_0_vm_write_pte() [all …]
|
| D | amdgpu_vm.c | 367 ib->length_dw = 0; in amdgpu_vm_clear_bo() 371 WARN_ON(ib->length_dw > 64); in amdgpu_vm_clear_bo() 460 ib->length_dw = 0; in amdgpu_vm_update_page_directory() 497 if (ib->length_dw != 0) { in amdgpu_vm_update_page_directory() 500 WARN_ON(ib->length_dw > ndw); in amdgpu_vm_update_page_directory() 514 if (!amdgpu_enable_scheduler || ib->length_dw == 0) { in amdgpu_vm_update_page_directory() 770 ib->length_dw = 0; in amdgpu_vm_bo_update_mapping() 783 WARN_ON(ib->length_dw > ndw); in amdgpu_vm_bo_update_mapping()
|
| D | amdgpu_cs.c | 216 p->chunks[i].length_dw = user_chunk.length_dw; in amdgpu_cs_parser_init() 218 size = p->chunks[i].length_dw; in amdgpu_cs_parser_init() 241 if (p->chunks[i].length_dw * sizeof(uint32_t) < size) { in amdgpu_cs_parser_init() 689 ib->length_dw = chunk_ib->ib_bytes / 4; in amdgpu_cs_ib_fill() 756 num_deps = chunk->length_dw * 4 / in amdgpu_cs_dependencies()
|
| D | amdgpu_uvd.c | 717 if (ctx->idx >= ib->length_dw) { in amdgpu_uvd_cs_reg() 759 for (ctx->idx = 0 ; ctx->idx < ib->length_dw; ) { in amdgpu_uvd_cs_packets() 801 if (ib->length_dw % 16) { in amdgpu_uvd_ring_parse_cs() 803 ib->length_dw); in amdgpu_uvd_ring_parse_cs() 888 ib->length_dw = 16; in amdgpu_uvd_send_msg()
|
| D | amdgpu_trace.h | 42 __entry->dw = p->ibs[i].length_dw;
|
| D | amdgpu_ttm.c | 1042 ib->length_dw = 0; in amdgpu_copy_buffer() 1065 WARN_ON(ib->length_dw > num_dw); in amdgpu_copy_buffer()
|
| D | uvd_v6_0.c | 570 amdgpu_ring_write(ring, ib->length_dw); in uvd_v6_0_ring_emit_ib()
|
| D | uvd_v5_0.c | 570 amdgpu_ring_write(ring, ib->length_dw); in uvd_v5_0_ring_emit_ib()
|
| D | uvd_v4_2.c | 524 amdgpu_ring_write(ring, ib->length_dw); in uvd_v4_2_ring_emit_ib()
|
| D | amdgpu.h | 807 uint32_t length_dw; member 1241 uint32_t length_dw; member
|
| D | gfx_v8_0.c | 689 ib.length_dw = 3; in gfx_v8_0_ring_test_ib() 4574 control |= ib->length_dw | in gfx_v8_0_ring_emit_ib_gfx() 4604 control |= ib->length_dw | in gfx_v8_0_ring_emit_ib_compute()
|
| D | gfx_v7_0.c | 2596 control |= ib->length_dw | in gfx_v7_0_ring_emit_ib_gfx() 2625 control |= ib->length_dw | in gfx_v7_0_ring_emit_ib_compute() 2672 ib.length_dw = 3; in gfx_v7_0_ring_test_ib()
|
| /linux-4.4.14/include/uapi/drm/ |
| D | amdgpu_drm.h | 358 uint32_t length_dw; member
|
| D | radeon_drm.h | 965 uint32_t length_dw; member
|