chunk_ib 800 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c struct drm_amdgpu_cs_chunk_ib *chunk_ib; chunk_ib 810 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c chunk_ib = chunk->kdata; chunk_ib 815 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c va_start = chunk_ib->va_start & AMDGPU_GMC_HOLE_MASK; chunk_ib 822 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c if ((va_start + chunk_ib->ib_bytes) > chunk_ib 838 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c memcpy(ib->ptr, kptr, chunk_ib->ib_bytes); chunk_ib 952 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c struct drm_amdgpu_cs_chunk_ib *chunk_ib; chunk_ib 957 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c chunk_ib = (struct drm_amdgpu_cs_chunk_ib *)chunk->kdata; chunk_ib 962 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c if (chunk_ib->ip_type == AMDGPU_HW_IP_GFX && chunk_ib 964 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c if (chunk_ib->flags & AMDGPU_IB_FLAG_PREEMPT) { chunk_ib 965 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c if (chunk_ib->flags & AMDGPU_IB_FLAG_CE) chunk_ib 976 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c r = amdgpu_ctx_get_entity(parser->ctx, chunk_ib->ip_type, chunk_ib 977 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c chunk_ib->ip_instance, chunk_ib->ring, chunk_ib 982 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c if (chunk_ib->flags & AMDGPU_IB_FLAG_PREAMBLE) chunk_ib 993 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c chunk_ib->ib_bytes : 0, ib); chunk_ib 999 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c ib->gpu_addr = chunk_ib->va_start; chunk_ib 1000 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c ib->length_dw = chunk_ib->ib_bytes / 4; chunk_ib 1001 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c ib->flags = chunk_ib->flags; chunk_ib 2776 drivers/gpu/drm/radeon/evergreen_cs.c } while (p->idx < p->chunk_ib->length_dw); chunk_ib 2799 drivers/gpu/drm/radeon/evergreen_cs.c struct radeon_cs_chunk *ib_chunk = p->chunk_ib; chunk_ib 3215 drivers/gpu/drm/radeon/evergreen_cs.c } while (p->idx < p->chunk_ib->length_dw); chunk_ib 2073 drivers/gpu/drm/radeon/r100.c } while (p->idx < p->chunk_ib->length_dw); chunk_ib 1320 drivers/gpu/drm/radeon/r300.c } while (p->idx < p->chunk_ib->length_dw); chunk_ib 2319 drivers/gpu/drm/radeon/r600_cs.c } while (p->idx < p->chunk_ib->length_dw); chunk_ib 2380 drivers/gpu/drm/radeon/r600_cs.c struct radeon_cs_chunk *ib_chunk = p->chunk_ib; chunk_ib 2527 drivers/gpu/drm/radeon/r600_cs.c } while (p->idx < p->chunk_ib->length_dw); chunk_ib 1074 drivers/gpu/drm/radeon/radeon.h struct radeon_cs_chunk *chunk_ib; chunk_ib 1091 drivers/gpu/drm/radeon/radeon.h struct radeon_cs_chunk *ibc = p->chunk_ib; chunk_ib 288 drivers/gpu/drm/radeon/radeon_cs.c p->chunk_ib = NULL; chunk_ib 322 drivers/gpu/drm/radeon/radeon_cs.c p->chunk_ib = &p->chunks[i]; chunk_ib 465 drivers/gpu/drm/radeon/radeon_cs.c if (parser->chunk_ib == NULL) chunk_ib 549 drivers/gpu/drm/radeon/radeon_cs.c if (parser->chunk_ib == NULL) chunk_ib 610 drivers/gpu/drm/radeon/radeon_cs.c if (parser->chunk_ib == NULL) chunk_ib 638 drivers/gpu/drm/radeon/radeon_cs.c ib_chunk = parser->chunk_ib; chunk_ib 644 drivers/gpu/drm/radeon/radeon_cs.c ib_chunk = parser->chunk_ib; chunk_ib 736 drivers/gpu/drm/radeon/radeon_cs.c struct radeon_cs_chunk *ib_chunk = p->chunk_ib; chunk_ib 41 drivers/gpu/drm/radeon/radeon_trace.h __entry->dw = p->chunk_ib->length_dw; chunk_ib 700 drivers/gpu/drm/radeon/radeon_uvd.c if (p->chunk_ib->length_dw % 16) { chunk_ib 702 drivers/gpu/drm/radeon/radeon_uvd.c p->chunk_ib->length_dw); chunk_ib 730 drivers/gpu/drm/radeon/radeon_uvd.c } while (p->idx < p->chunk_ib->length_dw); chunk_ib 565 drivers/gpu/drm/radeon/radeon_vce.c while (p->idx < p->chunk_ib->length_dw) {