chunk_ib          800 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c 			struct drm_amdgpu_cs_chunk_ib *chunk_ib;
chunk_ib          810 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c 			chunk_ib = chunk->kdata;
chunk_ib          815 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c 			va_start = chunk_ib->va_start & AMDGPU_GMC_HOLE_MASK;
chunk_ib          822 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c 			if ((va_start + chunk_ib->ib_bytes) >
chunk_ib          838 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c 				memcpy(ib->ptr, kptr, chunk_ib->ib_bytes);
chunk_ib          952 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c 		struct drm_amdgpu_cs_chunk_ib *chunk_ib;
chunk_ib          957 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c 		chunk_ib = (struct drm_amdgpu_cs_chunk_ib *)chunk->kdata;
chunk_ib          962 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c 		if (chunk_ib->ip_type == AMDGPU_HW_IP_GFX &&
chunk_ib          964 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c 			if (chunk_ib->flags & AMDGPU_IB_FLAG_PREEMPT) {
chunk_ib          965 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c 				if (chunk_ib->flags & AMDGPU_IB_FLAG_CE)
chunk_ib          976 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c 		r = amdgpu_ctx_get_entity(parser->ctx, chunk_ib->ip_type,
chunk_ib          977 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c 					  chunk_ib->ip_instance, chunk_ib->ring,
chunk_ib          982 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c 		if (chunk_ib->flags & AMDGPU_IB_FLAG_PREAMBLE)
chunk_ib          993 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c 				   chunk_ib->ib_bytes : 0, ib);
chunk_ib          999 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c 		ib->gpu_addr = chunk_ib->va_start;
chunk_ib         1000 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c 		ib->length_dw = chunk_ib->ib_bytes / 4;
chunk_ib         1001 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c 		ib->flags = chunk_ib->flags;
chunk_ib         2776 drivers/gpu/drm/radeon/evergreen_cs.c 	} while (p->idx < p->chunk_ib->length_dw);
chunk_ib         2799 drivers/gpu/drm/radeon/evergreen_cs.c 	struct radeon_cs_chunk *ib_chunk = p->chunk_ib;
chunk_ib         3215 drivers/gpu/drm/radeon/evergreen_cs.c 	} while (p->idx < p->chunk_ib->length_dw);
chunk_ib         2073 drivers/gpu/drm/radeon/r100.c 	} while (p->idx < p->chunk_ib->length_dw);
chunk_ib         1320 drivers/gpu/drm/radeon/r300.c 	} while (p->idx < p->chunk_ib->length_dw);
chunk_ib         2319 drivers/gpu/drm/radeon/r600_cs.c 	} while (p->idx < p->chunk_ib->length_dw);
chunk_ib         2380 drivers/gpu/drm/radeon/r600_cs.c 	struct radeon_cs_chunk *ib_chunk = p->chunk_ib;
chunk_ib         2527 drivers/gpu/drm/radeon/r600_cs.c 	} while (p->idx < p->chunk_ib->length_dw);
chunk_ib         1074 drivers/gpu/drm/radeon/radeon.h 	struct radeon_cs_chunk  *chunk_ib;
chunk_ib         1091 drivers/gpu/drm/radeon/radeon.h 	struct radeon_cs_chunk *ibc = p->chunk_ib;
chunk_ib          288 drivers/gpu/drm/radeon/radeon_cs.c 	p->chunk_ib = NULL;
chunk_ib          322 drivers/gpu/drm/radeon/radeon_cs.c 			p->chunk_ib = &p->chunks[i];
chunk_ib          465 drivers/gpu/drm/radeon/radeon_cs.c 	if (parser->chunk_ib == NULL)
chunk_ib          549 drivers/gpu/drm/radeon/radeon_cs.c 	if (parser->chunk_ib == NULL)
chunk_ib          610 drivers/gpu/drm/radeon/radeon_cs.c 	if (parser->chunk_ib == NULL)
chunk_ib          638 drivers/gpu/drm/radeon/radeon_cs.c 		ib_chunk = parser->chunk_ib;
chunk_ib          644 drivers/gpu/drm/radeon/radeon_cs.c 	ib_chunk = parser->chunk_ib;
chunk_ib          736 drivers/gpu/drm/radeon/radeon_cs.c 	struct radeon_cs_chunk *ib_chunk = p->chunk_ib;
chunk_ib           41 drivers/gpu/drm/radeon/radeon_trace.h 			   __entry->dw = p->chunk_ib->length_dw;
chunk_ib          700 drivers/gpu/drm/radeon/radeon_uvd.c 	if (p->chunk_ib->length_dw % 16) {
chunk_ib          702 drivers/gpu/drm/radeon/radeon_uvd.c 			  p->chunk_ib->length_dw);
chunk_ib          730 drivers/gpu/drm/radeon/radeon_uvd.c 	} while (p->idx < p->chunk_ib->length_dw);
chunk_ib          565 drivers/gpu/drm/radeon/radeon_vce.c 	while (p->idx < p->chunk_ib->length_dw) {