vc4                40 drivers/gpu/drm/vc4/vc4_bo.c static void vc4_bo_stats_print(struct drm_printer *p, struct vc4_dev *vc4)
vc4                44 drivers/gpu/drm/vc4/vc4_bo.c 	for (i = 0; i < vc4->num_labels; i++) {
vc4                45 drivers/gpu/drm/vc4/vc4_bo.c 		if (!vc4->bo_labels[i].num_allocated)
vc4                49 drivers/gpu/drm/vc4/vc4_bo.c 			   vc4->bo_labels[i].name,
vc4                50 drivers/gpu/drm/vc4/vc4_bo.c 			   vc4->bo_labels[i].size_allocated / 1024,
vc4                51 drivers/gpu/drm/vc4/vc4_bo.c 			   vc4->bo_labels[i].num_allocated);
vc4                54 drivers/gpu/drm/vc4/vc4_bo.c 	mutex_lock(&vc4->purgeable.lock);
vc4                55 drivers/gpu/drm/vc4/vc4_bo.c 	if (vc4->purgeable.num)
vc4                57 drivers/gpu/drm/vc4/vc4_bo.c 			   vc4->purgeable.size / 1024, vc4->purgeable.num);
vc4                59 drivers/gpu/drm/vc4/vc4_bo.c 	if (vc4->purgeable.purged_num)
vc4                61 drivers/gpu/drm/vc4/vc4_bo.c 			   vc4->purgeable.purged_size / 1024,
vc4                62 drivers/gpu/drm/vc4/vc4_bo.c 			   vc4->purgeable.purged_num);
vc4                63 drivers/gpu/drm/vc4/vc4_bo.c 	mutex_unlock(&vc4->purgeable.lock);
vc4                70 drivers/gpu/drm/vc4/vc4_bo.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4                73 drivers/gpu/drm/vc4/vc4_bo.c 	vc4_bo_stats_print(&p, vc4);
vc4                86 drivers/gpu/drm/vc4/vc4_bo.c static int vc4_get_user_label(struct vc4_dev *vc4, const char *name)
vc4                91 drivers/gpu/drm/vc4/vc4_bo.c 	for (i = 0; i < vc4->num_labels; i++) {
vc4                92 drivers/gpu/drm/vc4/vc4_bo.c 		if (!vc4->bo_labels[i].name) {
vc4                94 drivers/gpu/drm/vc4/vc4_bo.c 		} else if (strcmp(vc4->bo_labels[i].name, name) == 0) {
vc4               101 drivers/gpu/drm/vc4/vc4_bo.c 		WARN_ON(vc4->bo_labels[free_slot].num_allocated != 0);
vc4               102 drivers/gpu/drm/vc4/vc4_bo.c 		vc4->bo_labels[free_slot].name = name;
vc4               105 drivers/gpu/drm/vc4/vc4_bo.c 		u32 new_label_count = vc4->num_labels + 1;
vc4               107 drivers/gpu/drm/vc4/vc4_bo.c 			krealloc(vc4->bo_labels,
vc4               116 drivers/gpu/drm/vc4/vc4_bo.c 		free_slot = vc4->num_labels;
vc4               117 drivers/gpu/drm/vc4/vc4_bo.c 		vc4->bo_labels = new_labels;
vc4               118 drivers/gpu/drm/vc4/vc4_bo.c 		vc4->num_labels = new_label_count;
vc4               120 drivers/gpu/drm/vc4/vc4_bo.c 		vc4->bo_labels[free_slot].name = name;
vc4               121 drivers/gpu/drm/vc4/vc4_bo.c 		vc4->bo_labels[free_slot].num_allocated = 0;
vc4               122 drivers/gpu/drm/vc4/vc4_bo.c 		vc4->bo_labels[free_slot].size_allocated = 0;
vc4               131 drivers/gpu/drm/vc4/vc4_bo.c 	struct vc4_dev *vc4 = to_vc4_dev(gem_obj->dev);
vc4               133 drivers/gpu/drm/vc4/vc4_bo.c 	lockdep_assert_held(&vc4->bo_lock);
vc4               136 drivers/gpu/drm/vc4/vc4_bo.c 		vc4->bo_labels[label].num_allocated++;
vc4               137 drivers/gpu/drm/vc4/vc4_bo.c 		vc4->bo_labels[label].size_allocated += gem_obj->size;
vc4               140 drivers/gpu/drm/vc4/vc4_bo.c 	vc4->bo_labels[bo->label].num_allocated--;
vc4               141 drivers/gpu/drm/vc4/vc4_bo.c 	vc4->bo_labels[bo->label].size_allocated -= gem_obj->size;
vc4               143 drivers/gpu/drm/vc4/vc4_bo.c 	if (vc4->bo_labels[bo->label].num_allocated == 0 &&
vc4               150 drivers/gpu/drm/vc4/vc4_bo.c 		kfree(vc4->bo_labels[bo->label].name);
vc4               151 drivers/gpu/drm/vc4/vc4_bo.c 		vc4->bo_labels[bo->label].name = NULL;
vc4               165 drivers/gpu/drm/vc4/vc4_bo.c 	struct vc4_dev *vc4 = to_vc4_dev(obj->dev);
vc4               167 drivers/gpu/drm/vc4/vc4_bo.c 	lockdep_assert_held(&vc4->bo_lock);
vc4               183 drivers/gpu/drm/vc4/vc4_bo.c 	struct vc4_dev *vc4 = to_vc4_dev(bo->base.base.dev);
vc4               185 drivers/gpu/drm/vc4/vc4_bo.c 	lockdep_assert_held(&vc4->bo_lock);
vc4               193 drivers/gpu/drm/vc4/vc4_bo.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               196 drivers/gpu/drm/vc4/vc4_bo.c 	if (vc4->bo_cache.size_list_size <= page_index) {
vc4               197 drivers/gpu/drm/vc4/vc4_bo.c 		uint32_t new_size = max(vc4->bo_cache.size_list_size * 2,
vc4               210 drivers/gpu/drm/vc4/vc4_bo.c 		for (i = 0; i < vc4->bo_cache.size_list_size; i++) {
vc4               212 drivers/gpu/drm/vc4/vc4_bo.c 				&vc4->bo_cache.size_list[i];
vc4               220 drivers/gpu/drm/vc4/vc4_bo.c 		for (i = vc4->bo_cache.size_list_size; i < new_size; i++)
vc4               223 drivers/gpu/drm/vc4/vc4_bo.c 		kfree(vc4->bo_cache.size_list);
vc4               224 drivers/gpu/drm/vc4/vc4_bo.c 		vc4->bo_cache.size_list = new_list;
vc4               225 drivers/gpu/drm/vc4/vc4_bo.c 		vc4->bo_cache.size_list_size = new_size;
vc4               228 drivers/gpu/drm/vc4/vc4_bo.c 	return &vc4->bo_cache.size_list[page_index];
vc4               233 drivers/gpu/drm/vc4/vc4_bo.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               235 drivers/gpu/drm/vc4/vc4_bo.c 	mutex_lock(&vc4->bo_lock);
vc4               236 drivers/gpu/drm/vc4/vc4_bo.c 	while (!list_empty(&vc4->bo_cache.time_list)) {
vc4               237 drivers/gpu/drm/vc4/vc4_bo.c 		struct vc4_bo *bo = list_last_entry(&vc4->bo_cache.time_list,
vc4               242 drivers/gpu/drm/vc4/vc4_bo.c 	mutex_unlock(&vc4->bo_lock);
vc4               247 drivers/gpu/drm/vc4/vc4_bo.c 	struct vc4_dev *vc4 = to_vc4_dev(bo->base.base.dev);
vc4               249 drivers/gpu/drm/vc4/vc4_bo.c 	mutex_lock(&vc4->purgeable.lock);
vc4               250 drivers/gpu/drm/vc4/vc4_bo.c 	list_add_tail(&bo->size_head, &vc4->purgeable.list);
vc4               251 drivers/gpu/drm/vc4/vc4_bo.c 	vc4->purgeable.num++;
vc4               252 drivers/gpu/drm/vc4/vc4_bo.c 	vc4->purgeable.size += bo->base.base.size;
vc4               253 drivers/gpu/drm/vc4/vc4_bo.c 	mutex_unlock(&vc4->purgeable.lock);
vc4               258 drivers/gpu/drm/vc4/vc4_bo.c 	struct vc4_dev *vc4 = to_vc4_dev(bo->base.base.dev);
vc4               273 drivers/gpu/drm/vc4/vc4_bo.c 	vc4->purgeable.num--;
vc4               274 drivers/gpu/drm/vc4/vc4_bo.c 	vc4->purgeable.size -= bo->base.base.size;
vc4               279 drivers/gpu/drm/vc4/vc4_bo.c 	struct vc4_dev *vc4 = to_vc4_dev(bo->base.base.dev);
vc4               281 drivers/gpu/drm/vc4/vc4_bo.c 	mutex_lock(&vc4->purgeable.lock);
vc4               283 drivers/gpu/drm/vc4/vc4_bo.c 	mutex_unlock(&vc4->purgeable.lock);
vc4               303 drivers/gpu/drm/vc4/vc4_bo.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               305 drivers/gpu/drm/vc4/vc4_bo.c 	mutex_lock(&vc4->purgeable.lock);
vc4               306 drivers/gpu/drm/vc4/vc4_bo.c 	while (!list_empty(&vc4->purgeable.list)) {
vc4               307 drivers/gpu/drm/vc4/vc4_bo.c 		struct vc4_bo *bo = list_first_entry(&vc4->purgeable.list,
vc4               319 drivers/gpu/drm/vc4/vc4_bo.c 		mutex_unlock(&vc4->purgeable.lock);
vc4               338 drivers/gpu/drm/vc4/vc4_bo.c 		mutex_lock(&vc4->purgeable.lock);
vc4               341 drivers/gpu/drm/vc4/vc4_bo.c 			vc4->purgeable.purged_size += purged_size;
vc4               342 drivers/gpu/drm/vc4/vc4_bo.c 			vc4->purgeable.purged_num++;
vc4               345 drivers/gpu/drm/vc4/vc4_bo.c 	mutex_unlock(&vc4->purgeable.lock);
vc4               352 drivers/gpu/drm/vc4/vc4_bo.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               358 drivers/gpu/drm/vc4/vc4_bo.c 	mutex_lock(&vc4->bo_lock);
vc4               359 drivers/gpu/drm/vc4/vc4_bo.c 	if (page_index >= vc4->bo_cache.size_list_size)
vc4               362 drivers/gpu/drm/vc4/vc4_bo.c 	if (list_empty(&vc4->bo_cache.size_list[page_index]))
vc4               365 drivers/gpu/drm/vc4/vc4_bo.c 	bo = list_first_entry(&vc4->bo_cache.size_list[page_index],
vc4               373 drivers/gpu/drm/vc4/vc4_bo.c 	mutex_unlock(&vc4->bo_lock);
vc4               387 drivers/gpu/drm/vc4/vc4_bo.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               397 drivers/gpu/drm/vc4/vc4_bo.c 	mutex_lock(&vc4->bo_lock);
vc4               399 drivers/gpu/drm/vc4/vc4_bo.c 	vc4->bo_labels[VC4_BO_TYPE_KERNEL].num_allocated++;
vc4               400 drivers/gpu/drm/vc4/vc4_bo.c 	vc4->bo_labels[VC4_BO_TYPE_KERNEL].size_allocated += size;
vc4               401 drivers/gpu/drm/vc4/vc4_bo.c 	mutex_unlock(&vc4->bo_lock);
vc4               410 drivers/gpu/drm/vc4/vc4_bo.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               452 drivers/gpu/drm/vc4/vc4_bo.c 		struct drm_printer p = drm_info_printer(vc4->dev->dev);
vc4               454 drivers/gpu/drm/vc4/vc4_bo.c 		vc4_bo_stats_print(&p, vc4);
vc4               465 drivers/gpu/drm/vc4/vc4_bo.c 	mutex_lock(&vc4->bo_lock);
vc4               467 drivers/gpu/drm/vc4/vc4_bo.c 	mutex_unlock(&vc4->bo_lock);
vc4               500 drivers/gpu/drm/vc4/vc4_bo.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               503 drivers/gpu/drm/vc4/vc4_bo.c 	lockdep_assert_held(&vc4->bo_lock);
vc4               505 drivers/gpu/drm/vc4/vc4_bo.c 	while (!list_empty(&vc4->bo_cache.time_list)) {
vc4               506 drivers/gpu/drm/vc4/vc4_bo.c 		struct vc4_bo *bo = list_last_entry(&vc4->bo_cache.time_list,
vc4               509 drivers/gpu/drm/vc4/vc4_bo.c 			mod_timer(&vc4->bo_cache.time_timer,
vc4               526 drivers/gpu/drm/vc4/vc4_bo.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               536 drivers/gpu/drm/vc4/vc4_bo.c 	mutex_lock(&vc4->bo_lock);
vc4               579 drivers/gpu/drm/vc4/vc4_bo.c 	list_add(&bo->unref_head, &vc4->bo_cache.time_list);
vc4               586 drivers/gpu/drm/vc4/vc4_bo.c 	mutex_unlock(&vc4->bo_lock);
vc4               591 drivers/gpu/drm/vc4/vc4_bo.c 	struct vc4_dev *vc4 =
vc4               593 drivers/gpu/drm/vc4/vc4_bo.c 	struct drm_device *dev = vc4->dev;
vc4               595 drivers/gpu/drm/vc4/vc4_bo.c 	mutex_lock(&vc4->bo_lock);
vc4               597 drivers/gpu/drm/vc4/vc4_bo.c 	mutex_unlock(&vc4->bo_lock);
vc4               653 drivers/gpu/drm/vc4/vc4_bo.c 	struct vc4_dev *vc4 = from_timer(vc4, t, bo_cache.time_timer);
vc4               655 drivers/gpu/drm/vc4/vc4_bo.c 	schedule_work(&vc4->bo_cache.time_work);
vc4               796 drivers/gpu/drm/vc4/vc4_bo.c static int vc4_grab_bin_bo(struct vc4_dev *vc4, struct vc4_file *vc4file)
vc4               800 drivers/gpu/drm/vc4/vc4_bo.c 	if (!vc4->v3d)
vc4               806 drivers/gpu/drm/vc4/vc4_bo.c 	ret = vc4_v3d_bin_bo_get(vc4, &vc4file->bin_bo_used);
vc4               818 drivers/gpu/drm/vc4/vc4_bo.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               822 drivers/gpu/drm/vc4/vc4_bo.c 	ret = vc4_grab_bin_bo(vc4, vc4file);
vc4               867 drivers/gpu/drm/vc4/vc4_bo.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               887 drivers/gpu/drm/vc4/vc4_bo.c 	ret = vc4_grab_bin_bo(vc4, vc4file);
vc4              1010 drivers/gpu/drm/vc4/vc4_bo.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4              1017 drivers/gpu/drm/vc4/vc4_bo.c 	vc4->bo_labels = kcalloc(VC4_BO_TYPE_COUNT, sizeof(*vc4->bo_labels),
vc4              1019 drivers/gpu/drm/vc4/vc4_bo.c 	if (!vc4->bo_labels)
vc4              1021 drivers/gpu/drm/vc4/vc4_bo.c 	vc4->num_labels = VC4_BO_TYPE_COUNT;
vc4              1025 drivers/gpu/drm/vc4/vc4_bo.c 		vc4->bo_labels[i].name = bo_type_names[i];
vc4              1027 drivers/gpu/drm/vc4/vc4_bo.c 	mutex_init(&vc4->bo_lock);
vc4              1031 drivers/gpu/drm/vc4/vc4_bo.c 	INIT_LIST_HEAD(&vc4->bo_cache.time_list);
vc4              1033 drivers/gpu/drm/vc4/vc4_bo.c 	INIT_WORK(&vc4->bo_cache.time_work, vc4_bo_cache_time_work);
vc4              1034 drivers/gpu/drm/vc4/vc4_bo.c 	timer_setup(&vc4->bo_cache.time_timer, vc4_bo_cache_time_timer, 0);
vc4              1041 drivers/gpu/drm/vc4/vc4_bo.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4              1044 drivers/gpu/drm/vc4/vc4_bo.c 	del_timer(&vc4->bo_cache.time_timer);
vc4              1045 drivers/gpu/drm/vc4/vc4_bo.c 	cancel_work_sync(&vc4->bo_cache.time_work);
vc4              1049 drivers/gpu/drm/vc4/vc4_bo.c 	for (i = 0; i < vc4->num_labels; i++) {
vc4              1050 drivers/gpu/drm/vc4/vc4_bo.c 		if (vc4->bo_labels[i].num_allocated) {
vc4              1053 drivers/gpu/drm/vc4/vc4_bo.c 				  vc4->bo_labels[i].num_allocated,
vc4              1054 drivers/gpu/drm/vc4/vc4_bo.c 				  vc4->bo_labels[i].name);
vc4              1058 drivers/gpu/drm/vc4/vc4_bo.c 			kfree(vc4->bo_labels[i].name);
vc4              1060 drivers/gpu/drm/vc4/vc4_bo.c 	kfree(vc4->bo_labels);
vc4              1066 drivers/gpu/drm/vc4/vc4_bo.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4              1086 drivers/gpu/drm/vc4/vc4_bo.c 	mutex_lock(&vc4->bo_lock);
vc4              1087 drivers/gpu/drm/vc4/vc4_bo.c 	label = vc4_get_user_label(vc4, name);
vc4              1092 drivers/gpu/drm/vc4/vc4_bo.c 	mutex_unlock(&vc4->bo_lock);
vc4                92 drivers/gpu/drm/vc4/vc4_crtc.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               214 drivers/gpu/drm/vc4/vc4_crtc.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               385 drivers/gpu/drm/vc4/vc4_crtc.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               446 drivers/gpu/drm/vc4/vc4_crtc.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               456 drivers/gpu/drm/vc4/vc4_crtc.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               517 drivers/gpu/drm/vc4/vc4_crtc.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               549 drivers/gpu/drm/vc4/vc4_crtc.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               630 drivers/gpu/drm/vc4/vc4_crtc.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               650 drivers/gpu/drm/vc4/vc4_crtc.c 	spin_lock_irqsave(&vc4->hvs->mm_lock, flags);
vc4               651 drivers/gpu/drm/vc4/vc4_crtc.c 	ret = drm_mm_insert_node(&vc4->hvs->dlist_mm, &vc4_state->mm,
vc4               653 drivers/gpu/drm/vc4/vc4_crtc.c 	spin_unlock_irqrestore(&vc4->hvs->mm_lock, flags);
vc4               686 drivers/gpu/drm/vc4/vc4_crtc.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               693 drivers/gpu/drm/vc4/vc4_crtc.c 	u32 __iomem *dlist_start = vc4->hvs->dlist + vc4_state->mm.start;
vc4               785 drivers/gpu/drm/vc4/vc4_crtc.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               850 drivers/gpu/drm/vc4/vc4_crtc.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               883 drivers/gpu/drm/vc4/vc4_crtc.c 	up(&vc4->async_modeset);
vc4               898 drivers/gpu/drm/vc4/vc4_crtc.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               928 drivers/gpu/drm/vc4/vc4_crtc.c 	ret = down_interruptible(&vc4->async_modeset);
vc4               994 drivers/gpu/drm/vc4/vc4_crtc.c 	struct vc4_dev *vc4 = to_vc4_dev(crtc->dev);
vc4              1000 drivers/gpu/drm/vc4/vc4_crtc.c 		spin_lock_irqsave(&vc4->hvs->mm_lock, flags);
vc4              1002 drivers/gpu/drm/vc4/vc4_crtc.c 		spin_unlock_irqrestore(&vc4->hvs->mm_lock, flags);
vc4              1112 drivers/gpu/drm/vc4/vc4_crtc.c 	struct vc4_dev *vc4 = to_vc4_dev(drm);
vc4                26 drivers/gpu/drm/vc4/vc4_debugfs.c 	struct vc4_dev *vc4 = to_vc4_dev(minor->dev);
vc4                30 drivers/gpu/drm/vc4/vc4_debugfs.c 			    minor->debugfs_root, &vc4->load_tracker_enabled);
vc4                32 drivers/gpu/drm/vc4/vc4_debugfs.c 	list_for_each_entry(entry, &vc4->debugfs_list, link) {
vc4                68 drivers/gpu/drm/vc4/vc4_debugfs.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4                80 drivers/gpu/drm/vc4/vc4_debugfs.c 	list_add(&entry->link, &vc4->debugfs_list);
vc4               261 drivers/gpu/drm/vc4/vc4_dpi.c 	struct vc4_dev *vc4 = to_vc4_dev(drm);
vc4               321 drivers/gpu/drm/vc4/vc4_dpi.c 	vc4->dpi = dpi;
vc4               337 drivers/gpu/drm/vc4/vc4_dpi.c 	struct vc4_dev *vc4 = to_vc4_dev(drm);
vc4               346 drivers/gpu/drm/vc4/vc4_dpi.c 	vc4->dpi = NULL;
vc4                70 drivers/gpu/drm/vc4/vc4_drv.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4                77 drivers/gpu/drm/vc4/vc4_drv.c 	if (!vc4->v3d)
vc4                82 drivers/gpu/drm/vc4/vc4_drv.c 		ret = vc4_v3d_pm_get(vc4);
vc4                86 drivers/gpu/drm/vc4/vc4_drv.c 		vc4_v3d_pm_put(vc4);
vc4                89 drivers/gpu/drm/vc4/vc4_drv.c 		ret = vc4_v3d_pm_get(vc4);
vc4                93 drivers/gpu/drm/vc4/vc4_drv.c 		vc4_v3d_pm_put(vc4);
vc4                96 drivers/gpu/drm/vc4/vc4_drv.c 		ret = vc4_v3d_pm_get(vc4);
vc4               100 drivers/gpu/drm/vc4/vc4_drv.c 		vc4_v3d_pm_put(vc4);
vc4               133 drivers/gpu/drm/vc4/vc4_drv.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               137 drivers/gpu/drm/vc4/vc4_drv.c 		vc4_v3d_bin_bo_put(vc4);
vc4               256 drivers/gpu/drm/vc4/vc4_drv.c 	struct vc4_dev *vc4;
vc4               262 drivers/gpu/drm/vc4/vc4_drv.c 	vc4 = devm_kzalloc(dev, sizeof(*vc4), GFP_KERNEL);
vc4               263 drivers/gpu/drm/vc4/vc4_drv.c 	if (!vc4)
vc4               276 drivers/gpu/drm/vc4/vc4_drv.c 	vc4->dev = drm;
vc4               277 drivers/gpu/drm/vc4/vc4_drv.c 	drm->dev_private = vc4;
vc4               278 drivers/gpu/drm/vc4/vc4_drv.c 	INIT_LIST_HEAD(&vc4->debugfs_list);
vc4               280 drivers/gpu/drm/vc4/vc4_drv.c 	mutex_init(&vc4->bin_bo_lock);
vc4               321 drivers/gpu/drm/vc4/vc4_drv.c 	struct vc4_dev *vc4 = to_vc4_dev(drm);
vc4               329 drivers/gpu/drm/vc4/vc4_drv.c 	drm_atomic_private_obj_fini(&vc4->load_tracker);
vc4               330 drivers/gpu/drm/vc4/vc4_drv.c 	drm_atomic_private_obj_fini(&vc4->ctm_manager);
vc4               309 drivers/gpu/drm/vc4/vc4_drv.h 	struct vc4_dev *vc4;
vc4               478 drivers/gpu/drm/vc4/vc4_drv.h #define V3D_READ(offset) readl(vc4->v3d->regs + offset)
vc4               479 drivers/gpu/drm/vc4/vc4_drv.h #define V3D_WRITE(offset, val) writel(val, vc4->v3d->regs + offset)
vc4               480 drivers/gpu/drm/vc4/vc4_drv.h #define HVS_READ(offset) readl(vc4->hvs->regs + offset)
vc4               481 drivers/gpu/drm/vc4/vc4_drv.h #define HVS_WRITE(offset, val) writel(val, vc4->hvs->regs + offset)
vc4               616 drivers/gpu/drm/vc4/vc4_drv.h vc4_first_bin_job(struct vc4_dev *vc4)
vc4               618 drivers/gpu/drm/vc4/vc4_drv.h 	return list_first_entry_or_null(&vc4->bin_job_list,
vc4               623 drivers/gpu/drm/vc4/vc4_drv.h vc4_first_render_job(struct vc4_dev *vc4)
vc4               625 drivers/gpu/drm/vc4/vc4_drv.h 	return list_first_entry_or_null(&vc4->render_job_list,
vc4               630 drivers/gpu/drm/vc4/vc4_drv.h vc4_last_render_job(struct vc4_dev *vc4)
vc4               632 drivers/gpu/drm/vc4/vc4_drv.h 	if (list_empty(&vc4->render_job_list))
vc4               634 drivers/gpu/drm/vc4/vc4_drv.h 	return list_last_entry(&vc4->render_job_list,
vc4               807 drivers/gpu/drm/vc4/vc4_drv.h void vc4_job_handle_completed(struct vc4_dev *vc4);
vc4               850 drivers/gpu/drm/vc4/vc4_drv.h int vc4_v3d_get_bin_slot(struct vc4_dev *vc4);
vc4               851 drivers/gpu/drm/vc4/vc4_drv.h int vc4_v3d_bin_bo_get(struct vc4_dev *vc4, bool *used);
vc4               852 drivers/gpu/drm/vc4/vc4_drv.h void vc4_v3d_bin_bo_put(struct vc4_dev *vc4);
vc4               853 drivers/gpu/drm/vc4/vc4_drv.h int vc4_v3d_pm_get(struct vc4_dev *vc4);
vc4               854 drivers/gpu/drm/vc4/vc4_drv.h void vc4_v3d_pm_put(struct vc4_dev *vc4);
vc4               883 drivers/gpu/drm/vc4/vc4_drv.h void vc4_perfmon_start(struct vc4_dev *vc4, struct vc4_perfmon *perfmon);
vc4               884 drivers/gpu/drm/vc4/vc4_drv.h void vc4_perfmon_stop(struct vc4_dev *vc4, struct vc4_perfmon *perfmon,
vc4              1444 drivers/gpu/drm/vc4/vc4_dsi.c 	struct vc4_dev *vc4 = to_vc4_dev(drm);
vc4              1596 drivers/gpu/drm/vc4/vc4_dsi.c 		vc4->dsi1 = dsi;
vc4              1628 drivers/gpu/drm/vc4/vc4_dsi.c 	struct vc4_dev *vc4 = to_vc4_dev(drm);
vc4              1637 drivers/gpu/drm/vc4/vc4_dsi.c 		vc4->dsi1 = NULL;
vc4                39 drivers/gpu/drm/vc4/vc4_fence.c 	struct vc4_dev *vc4 = to_vc4_dev(f->dev);
vc4                41 drivers/gpu/drm/vc4/vc4_fence.c 	return vc4->finished_seqno >= f->seqno;
vc4                42 drivers/gpu/drm/vc4/vc4_gem.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4                44 drivers/gpu/drm/vc4/vc4_gem.c 	mod_timer(&vc4->hangcheck.timer,
vc4                74 drivers/gpu/drm/vc4/vc4_gem.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4                79 drivers/gpu/drm/vc4/vc4_gem.c 	if (!vc4->v3d) {
vc4                84 drivers/gpu/drm/vc4/vc4_gem.c 	spin_lock_irqsave(&vc4->job_lock, irqflags);
vc4                85 drivers/gpu/drm/vc4/vc4_gem.c 	kernel_state = vc4->hang_state;
vc4                87 drivers/gpu/drm/vc4/vc4_gem.c 		spin_unlock_irqrestore(&vc4->job_lock, irqflags);
vc4                97 drivers/gpu/drm/vc4/vc4_gem.c 		spin_unlock_irqrestore(&vc4->job_lock, irqflags);
vc4               101 drivers/gpu/drm/vc4/vc4_gem.c 	vc4->hang_state = NULL;
vc4               102 drivers/gpu/drm/vc4/vc4_gem.c 	spin_unlock_irqrestore(&vc4->job_lock, irqflags);
vc4               151 drivers/gpu/drm/vc4/vc4_gem.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               165 drivers/gpu/drm/vc4/vc4_gem.c 	spin_lock_irqsave(&vc4->job_lock, irqflags);
vc4               166 drivers/gpu/drm/vc4/vc4_gem.c 	exec[0] = vc4_first_bin_job(vc4);
vc4               167 drivers/gpu/drm/vc4/vc4_gem.c 	exec[1] = vc4_first_render_job(vc4);
vc4               169 drivers/gpu/drm/vc4/vc4_gem.c 		spin_unlock_irqrestore(&vc4->job_lock, irqflags);
vc4               189 drivers/gpu/drm/vc4/vc4_gem.c 		spin_unlock_irqrestore(&vc4->job_lock, irqflags);
vc4               227 drivers/gpu/drm/vc4/vc4_gem.c 	spin_unlock_irqrestore(&vc4->job_lock, irqflags);
vc4               275 drivers/gpu/drm/vc4/vc4_gem.c 	spin_lock_irqsave(&vc4->job_lock, irqflags);
vc4               276 drivers/gpu/drm/vc4/vc4_gem.c 	if (vc4->hang_state) {
vc4               277 drivers/gpu/drm/vc4/vc4_gem.c 		spin_unlock_irqrestore(&vc4->job_lock, irqflags);
vc4               280 drivers/gpu/drm/vc4/vc4_gem.c 		vc4->hang_state = kernel_state;
vc4               281 drivers/gpu/drm/vc4/vc4_gem.c 		spin_unlock_irqrestore(&vc4->job_lock, irqflags);
vc4               288 drivers/gpu/drm/vc4/vc4_gem.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               292 drivers/gpu/drm/vc4/vc4_gem.c 	mutex_lock(&vc4->power_lock);
vc4               293 drivers/gpu/drm/vc4/vc4_gem.c 	if (vc4->power_refcount) {
vc4               297 drivers/gpu/drm/vc4/vc4_gem.c 		pm_runtime_put_sync_suspend(&vc4->v3d->pdev->dev);
vc4               298 drivers/gpu/drm/vc4/vc4_gem.c 		pm_runtime_get_sync(&vc4->v3d->pdev->dev);
vc4               300 drivers/gpu/drm/vc4/vc4_gem.c 	mutex_unlock(&vc4->power_lock);
vc4               314 drivers/gpu/drm/vc4/vc4_gem.c 	struct vc4_dev *vc4 =
vc4               317 drivers/gpu/drm/vc4/vc4_gem.c 	vc4_save_hang_state(vc4->dev);
vc4               319 drivers/gpu/drm/vc4/vc4_gem.c 	vc4_reset(vc4->dev);
vc4               325 drivers/gpu/drm/vc4/vc4_gem.c 	struct vc4_dev *vc4 = from_timer(vc4, t, hangcheck.timer);
vc4               326 drivers/gpu/drm/vc4/vc4_gem.c 	struct drm_device *dev = vc4->dev;
vc4               331 drivers/gpu/drm/vc4/vc4_gem.c 	spin_lock_irqsave(&vc4->job_lock, irqflags);
vc4               333 drivers/gpu/drm/vc4/vc4_gem.c 	bin_exec = vc4_first_bin_job(vc4);
vc4               334 drivers/gpu/drm/vc4/vc4_gem.c 	render_exec = vc4_first_render_job(vc4);
vc4               338 drivers/gpu/drm/vc4/vc4_gem.c 		spin_unlock_irqrestore(&vc4->job_lock, irqflags);
vc4               354 drivers/gpu/drm/vc4/vc4_gem.c 		spin_unlock_irqrestore(&vc4->job_lock, irqflags);
vc4               359 drivers/gpu/drm/vc4/vc4_gem.c 	spin_unlock_irqrestore(&vc4->job_lock, irqflags);
vc4               365 drivers/gpu/drm/vc4/vc4_gem.c 	schedule_work(&vc4->hangcheck.reset_work);
vc4               371 drivers/gpu/drm/vc4/vc4_gem.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               384 drivers/gpu/drm/vc4/vc4_gem.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               389 drivers/gpu/drm/vc4/vc4_gem.c 	if (vc4->finished_seqno >= seqno)
vc4               399 drivers/gpu/drm/vc4/vc4_gem.c 		prepare_to_wait(&vc4->job_wait_queue, &wait,
vc4               408 drivers/gpu/drm/vc4/vc4_gem.c 		if (vc4->finished_seqno >= seqno)
vc4               422 drivers/gpu/drm/vc4/vc4_gem.c 	finish_wait(&vc4->job_wait_queue, &wait);
vc4               431 drivers/gpu/drm/vc4/vc4_gem.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               450 drivers/gpu/drm/vc4/vc4_gem.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               468 drivers/gpu/drm/vc4/vc4_gem.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               472 drivers/gpu/drm/vc4/vc4_gem.c 	exec = vc4_first_bin_job(vc4);
vc4               481 drivers/gpu/drm/vc4/vc4_gem.c 	if (exec->perfmon && vc4->active_perfmon != exec->perfmon)
vc4               482 drivers/gpu/drm/vc4/vc4_gem.c 		vc4_perfmon_start(vc4, exec->perfmon);
vc4               493 drivers/gpu/drm/vc4/vc4_gem.c 		next = vc4_first_bin_job(vc4);
vc4               508 drivers/gpu/drm/vc4/vc4_gem.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               509 drivers/gpu/drm/vc4/vc4_gem.c 	struct vc4_exec_info *exec = vc4_first_render_job(vc4);
vc4               528 drivers/gpu/drm/vc4/vc4_gem.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               529 drivers/gpu/drm/vc4/vc4_gem.c 	bool was_empty = list_empty(&vc4->render_job_list);
vc4               531 drivers/gpu/drm/vc4/vc4_gem.c 	list_move_tail(&exec->head, &vc4->render_job_list);
vc4               669 drivers/gpu/drm/vc4/vc4_gem.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               680 drivers/gpu/drm/vc4/vc4_gem.c 	spin_lock_irqsave(&vc4->job_lock, irqflags);
vc4               682 drivers/gpu/drm/vc4/vc4_gem.c 	seqno = ++vc4->emit_seqno;
vc4               685 drivers/gpu/drm/vc4/vc4_gem.c 	dma_fence_init(&fence->base, &vc4_fence_ops, &vc4->job_lock,
vc4               686 drivers/gpu/drm/vc4/vc4_gem.c 		       vc4->dma_fence_context, exec->seqno);
vc4               697 drivers/gpu/drm/vc4/vc4_gem.c 	list_add_tail(&exec->head, &vc4->bin_job_list);
vc4               704 drivers/gpu/drm/vc4/vc4_gem.c 	renderjob = vc4_first_render_job(vc4);
vc4               705 drivers/gpu/drm/vc4/vc4_gem.c 	if (vc4_first_bin_job(vc4) == exec &&
vc4               711 drivers/gpu/drm/vc4/vc4_gem.c 	spin_unlock_irqrestore(&vc4->job_lock, irqflags);
vc4               825 drivers/gpu/drm/vc4/vc4_gem.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               925 drivers/gpu/drm/vc4/vc4_gem.c 		ret = vc4_v3d_bin_bo_get(vc4, &exec->bin_bo_used);
vc4               944 drivers/gpu/drm/vc4/vc4_gem.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               974 drivers/gpu/drm/vc4/vc4_gem.c 	spin_lock_irqsave(&vc4->job_lock, irqflags);
vc4               975 drivers/gpu/drm/vc4/vc4_gem.c 	vc4->bin_alloc_used &= ~exec->bin_slots;
vc4               976 drivers/gpu/drm/vc4/vc4_gem.c 	spin_unlock_irqrestore(&vc4->job_lock, irqflags);
vc4               980 drivers/gpu/drm/vc4/vc4_gem.c 		vc4_v3d_bin_bo_put(vc4);
vc4               985 drivers/gpu/drm/vc4/vc4_gem.c 	vc4_v3d_pm_put(vc4);
vc4               991 drivers/gpu/drm/vc4/vc4_gem.c vc4_job_handle_completed(struct vc4_dev *vc4)
vc4               996 drivers/gpu/drm/vc4/vc4_gem.c 	spin_lock_irqsave(&vc4->job_lock, irqflags);
vc4               997 drivers/gpu/drm/vc4/vc4_gem.c 	while (!list_empty(&vc4->job_done_list)) {
vc4               999 drivers/gpu/drm/vc4/vc4_gem.c 			list_first_entry(&vc4->job_done_list,
vc4              1003 drivers/gpu/drm/vc4/vc4_gem.c 		spin_unlock_irqrestore(&vc4->job_lock, irqflags);
vc4              1004 drivers/gpu/drm/vc4/vc4_gem.c 		vc4_complete_exec(vc4->dev, exec);
vc4              1005 drivers/gpu/drm/vc4/vc4_gem.c 		spin_lock_irqsave(&vc4->job_lock, irqflags);
vc4              1008 drivers/gpu/drm/vc4/vc4_gem.c 	list_for_each_entry_safe(cb, cb_temp, &vc4->seqno_cb_list, work.entry) {
vc4              1009 drivers/gpu/drm/vc4/vc4_gem.c 		if (cb->seqno <= vc4->finished_seqno) {
vc4              1015 drivers/gpu/drm/vc4/vc4_gem.c 	spin_unlock_irqrestore(&vc4->job_lock, irqflags);
vc4              1029 drivers/gpu/drm/vc4/vc4_gem.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4              1036 drivers/gpu/drm/vc4/vc4_gem.c 	spin_lock_irqsave(&vc4->job_lock, irqflags);
vc4              1037 drivers/gpu/drm/vc4/vc4_gem.c 	if (seqno > vc4->finished_seqno) {
vc4              1039 drivers/gpu/drm/vc4/vc4_gem.c 		list_add_tail(&cb->work.entry, &vc4->seqno_cb_list);
vc4              1043 drivers/gpu/drm/vc4/vc4_gem.c 	spin_unlock_irqrestore(&vc4->job_lock, irqflags);
vc4              1055 drivers/gpu/drm/vc4/vc4_gem.c 	struct vc4_dev *vc4 =
vc4              1058 drivers/gpu/drm/vc4/vc4_gem.c 	vc4_job_handle_completed(vc4);
vc4              1131 drivers/gpu/drm/vc4/vc4_gem.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4              1140 drivers/gpu/drm/vc4/vc4_gem.c 	if (!vc4->v3d) {
vc4              1164 drivers/gpu/drm/vc4/vc4_gem.c 	ret = vc4_v3d_pm_get(vc4);
vc4              1198 drivers/gpu/drm/vc4/vc4_gem.c 					     vc4->dma_fence_context)) {
vc4              1257 drivers/gpu/drm/vc4/vc4_gem.c 	args->seqno = vc4->emit_seqno;
vc4              1262 drivers/gpu/drm/vc4/vc4_gem.c 	vc4_complete_exec(vc4->dev, exec);
vc4              1270 drivers/gpu/drm/vc4/vc4_gem.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4              1272 drivers/gpu/drm/vc4/vc4_gem.c 	vc4->dma_fence_context = dma_fence_context_alloc(1);
vc4              1274 drivers/gpu/drm/vc4/vc4_gem.c 	INIT_LIST_HEAD(&vc4->bin_job_list);
vc4              1275 drivers/gpu/drm/vc4/vc4_gem.c 	INIT_LIST_HEAD(&vc4->render_job_list);
vc4              1276 drivers/gpu/drm/vc4/vc4_gem.c 	INIT_LIST_HEAD(&vc4->job_done_list);
vc4              1277 drivers/gpu/drm/vc4/vc4_gem.c 	INIT_LIST_HEAD(&vc4->seqno_cb_list);
vc4              1278 drivers/gpu/drm/vc4/vc4_gem.c 	spin_lock_init(&vc4->job_lock);
vc4              1280 drivers/gpu/drm/vc4/vc4_gem.c 	INIT_WORK(&vc4->hangcheck.reset_work, vc4_reset_work);
vc4              1281 drivers/gpu/drm/vc4/vc4_gem.c 	timer_setup(&vc4->hangcheck.timer, vc4_hangcheck_elapsed, 0);
vc4              1283 drivers/gpu/drm/vc4/vc4_gem.c 	INIT_WORK(&vc4->job_done_work, vc4_job_done_work);
vc4              1285 drivers/gpu/drm/vc4/vc4_gem.c 	mutex_init(&vc4->power_lock);
vc4              1287 drivers/gpu/drm/vc4/vc4_gem.c 	INIT_LIST_HEAD(&vc4->purgeable.list);
vc4              1288 drivers/gpu/drm/vc4/vc4_gem.c 	mutex_init(&vc4->purgeable.lock);
vc4              1294 drivers/gpu/drm/vc4/vc4_gem.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4              1299 drivers/gpu/drm/vc4/vc4_gem.c 	WARN_ON(vc4->emit_seqno != vc4->finished_seqno);
vc4              1304 drivers/gpu/drm/vc4/vc4_gem.c 	if (vc4->bin_bo) {
vc4              1305 drivers/gpu/drm/vc4/vc4_gem.c 		drm_gem_object_put_unlocked(&vc4->bin_bo->base.base);
vc4              1306 drivers/gpu/drm/vc4/vc4_gem.c 		vc4->bin_bo = NULL;
vc4              1309 drivers/gpu/drm/vc4/vc4_gem.c 	if (vc4->hang_state)
vc4              1310 drivers/gpu/drm/vc4/vc4_gem.c 		vc4_free_hang_state(dev, vc4->hang_state);
vc4                97 drivers/gpu/drm/vc4/vc4_hdmi.c #define HDMI_READ(offset) readl(vc4->hdmi->hdmicore_regs + offset)
vc4                98 drivers/gpu/drm/vc4/vc4_hdmi.c #define HDMI_WRITE(offset, val) writel(val, vc4->hdmi->hdmicore_regs + offset)
vc4                99 drivers/gpu/drm/vc4/vc4_hdmi.c #define HD_READ(offset) readl(vc4->hdmi->hd_regs + offset)
vc4               100 drivers/gpu/drm/vc4/vc4_hdmi.c #define HD_WRITE(offset, val) writel(val, vc4->hdmi->hd_regs + offset)
vc4               186 drivers/gpu/drm/vc4/vc4_hdmi.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               187 drivers/gpu/drm/vc4/vc4_hdmi.c 	struct vc4_hdmi *hdmi = vc4->hdmi;
vc4               200 drivers/gpu/drm/vc4/vc4_hdmi.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               202 drivers/gpu/drm/vc4/vc4_hdmi.c 	if (vc4->hdmi->hpd_gpio) {
vc4               203 drivers/gpu/drm/vc4/vc4_hdmi.c 		if (gpio_get_value_cansleep(vc4->hdmi->hpd_gpio) ^
vc4               204 drivers/gpu/drm/vc4/vc4_hdmi.c 		    vc4->hdmi->hpd_active_low)
vc4               206 drivers/gpu/drm/vc4/vc4_hdmi.c 		cec_phys_addr_invalidate(vc4->hdmi->cec_adap);
vc4               210 drivers/gpu/drm/vc4/vc4_hdmi.c 	if (drm_probe_ddc(vc4->hdmi->ddc))
vc4               215 drivers/gpu/drm/vc4/vc4_hdmi.c 	cec_phys_addr_invalidate(vc4->hdmi->cec_adap);
vc4               232 drivers/gpu/drm/vc4/vc4_hdmi.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               236 drivers/gpu/drm/vc4/vc4_hdmi.c 	edid = drm_get_edid(connector, vc4->hdmi->ddc);
vc4               237 drivers/gpu/drm/vc4/vc4_hdmi.c 	cec_s_phys_addr_from_edid(vc4->hdmi->cec_adap, edid);
vc4               319 drivers/gpu/drm/vc4/vc4_hdmi.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               333 drivers/gpu/drm/vc4/vc4_hdmi.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               380 drivers/gpu/drm/vc4/vc4_hdmi.c 	struct vc4_dev *vc4 = encoder->dev->dev_private;
vc4               381 drivers/gpu/drm/vc4/vc4_hdmi.c 	struct vc4_hdmi *hdmi = vc4->hdmi;
vc4               428 drivers/gpu/drm/vc4/vc4_hdmi.c 	struct vc4_dev *vc4 = drm->dev_private;
vc4               429 drivers/gpu/drm/vc4/vc4_hdmi.c 	struct vc4_hdmi *hdmi = vc4->hdmi;
vc4               452 drivers/gpu/drm/vc4/vc4_hdmi.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               453 drivers/gpu/drm/vc4/vc4_hdmi.c 	struct vc4_hdmi *hdmi = vc4->hdmi;
vc4               474 drivers/gpu/drm/vc4/vc4_hdmi.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               475 drivers/gpu/drm/vc4/vc4_hdmi.c 	struct vc4_hdmi *hdmi = vc4->hdmi;
vc4               716 drivers/gpu/drm/vc4/vc4_hdmi.c 	struct vc4_dev *vc4 = to_vc4_dev(drm);
vc4               737 drivers/gpu/drm/vc4/vc4_hdmi.c 	struct vc4_dev *vc4 = to_vc4_dev(drm);
vc4               773 drivers/gpu/drm/vc4/vc4_hdmi.c 	struct vc4_dev *vc4 = to_vc4_dev(encoder->dev);
vc4               807 drivers/gpu/drm/vc4/vc4_hdmi.c 	struct vc4_dev *vc4 = to_vc4_dev(drm);
vc4               841 drivers/gpu/drm/vc4/vc4_hdmi.c 	struct vc4_dev *vc4 = to_vc4_dev(drm);
vc4               913 drivers/gpu/drm/vc4/vc4_hdmi.c 	struct vc4_dev *vc4 = to_vc4_dev(drm);
vc4              1148 drivers/gpu/drm/vc4/vc4_hdmi.c 	struct vc4_dev *vc4 = priv;
vc4              1149 drivers/gpu/drm/vc4/vc4_hdmi.c 	struct vc4_hdmi *hdmi = vc4->hdmi;
vc4              1168 drivers/gpu/drm/vc4/vc4_hdmi.c static void vc4_cec_read_msg(struct vc4_dev *vc4, u32 cntrl1)
vc4              1170 drivers/gpu/drm/vc4/vc4_hdmi.c 	struct cec_msg *msg = &vc4->hdmi->cec_rx_msg;
vc4              1187 drivers/gpu/drm/vc4/vc4_hdmi.c 	struct vc4_dev *vc4 = priv;
vc4              1188 drivers/gpu/drm/vc4/vc4_hdmi.c 	struct vc4_hdmi *hdmi = vc4->hdmi;
vc4              1199 drivers/gpu/drm/vc4/vc4_hdmi.c 		vc4_cec_read_msg(vc4, cntrl1);
vc4              1215 drivers/gpu/drm/vc4/vc4_hdmi.c 	struct vc4_dev *vc4 = cec_get_drvdata(adap);
vc4              1258 drivers/gpu/drm/vc4/vc4_hdmi.c 	struct vc4_dev *vc4 = cec_get_drvdata(adap);
vc4              1269 drivers/gpu/drm/vc4/vc4_hdmi.c 	struct vc4_dev *vc4 = cec_get_drvdata(adap);
vc4              1305 drivers/gpu/drm/vc4/vc4_hdmi.c 	struct vc4_dev *vc4 = drm->dev_private;
vc4              1397 drivers/gpu/drm/vc4/vc4_hdmi.c 	vc4->hdmi = hdmi;
vc4              1420 drivers/gpu/drm/vc4/vc4_hdmi.c 					      vc4, "vc4",
vc4              1444 drivers/gpu/drm/vc4/vc4_hdmi.c 					"vc4 hdmi cec", vc4);
vc4              1481 drivers/gpu/drm/vc4/vc4_hdmi.c 	struct vc4_dev *vc4 = drm->dev_private;
vc4              1482 drivers/gpu/drm/vc4/vc4_hdmi.c 	struct vc4_hdmi *hdmi = vc4->hdmi;
vc4              1493 drivers/gpu/drm/vc4/vc4_hdmi.c 	vc4->hdmi = NULL;
vc4                66 drivers/gpu/drm/vc4/vc4_hvs.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4                67 drivers/gpu/drm/vc4/vc4_hvs.c 	struct drm_printer p = drm_info_printer(&vc4->hvs->pdev->dev);
vc4                70 drivers/gpu/drm/vc4/vc4_hvs.c 	drm_print_regset32(&p, &vc4->hvs->regset);
vc4                76 drivers/gpu/drm/vc4/vc4_hvs.c 			 readl((u32 __iomem *)vc4->hvs->dlist + i + 0),
vc4                77 drivers/gpu/drm/vc4/vc4_hvs.c 			 readl((u32 __iomem *)vc4->hvs->dlist + i + 1),
vc4                78 drivers/gpu/drm/vc4/vc4_hvs.c 			 readl((u32 __iomem *)vc4->hvs->dlist + i + 2),
vc4                79 drivers/gpu/drm/vc4/vc4_hvs.c 			 readl((u32 __iomem *)vc4->hvs->dlist + i + 3));
vc4                87 drivers/gpu/drm/vc4/vc4_hvs.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4                90 drivers/gpu/drm/vc4/vc4_hvs.c 	drm_printf(&p, "%d\n", atomic_read(&vc4->underrun));
vc4               159 drivers/gpu/drm/vc4/vc4_hvs.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               169 drivers/gpu/drm/vc4/vc4_hvs.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               181 drivers/gpu/drm/vc4/vc4_hvs.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               183 drivers/gpu/drm/vc4/vc4_hvs.c 	atomic_inc(&vc4->underrun);
vc4               190 drivers/gpu/drm/vc4/vc4_hvs.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               222 drivers/gpu/drm/vc4/vc4_hvs.c 	struct vc4_dev *vc4 = drm->dev_private;
vc4               270 drivers/gpu/drm/vc4/vc4_hvs.c 	vc4->hvs = hvs;
vc4               316 drivers/gpu/drm/vc4/vc4_hvs.c 	struct vc4_dev *vc4 = drm->dev_private;
vc4               318 drivers/gpu/drm/vc4/vc4_hvs.c 	if (vc4->hvs->mitchell_netravali_filter.allocated)
vc4               319 drivers/gpu/drm/vc4/vc4_hvs.c 		drm_mm_remove_node(&vc4->hvs->mitchell_netravali_filter);
vc4               321 drivers/gpu/drm/vc4/vc4_hvs.c 	drm_mm_takedown(&vc4->hvs->dlist_mm);
vc4               322 drivers/gpu/drm/vc4/vc4_hvs.c 	drm_mm_takedown(&vc4->hvs->lbm_mm);
vc4               324 drivers/gpu/drm/vc4/vc4_hvs.c 	vc4->hvs = NULL;
vc4                60 drivers/gpu/drm/vc4/vc4_irq.c 	struct vc4_dev *vc4 =
vc4                67 drivers/gpu/drm/vc4/vc4_irq.c 	mutex_lock(&vc4->bin_bo_lock);
vc4                69 drivers/gpu/drm/vc4/vc4_irq.c 	if (!vc4->bin_bo)
vc4                72 drivers/gpu/drm/vc4/vc4_irq.c 	bo = vc4->bin_bo;
vc4                74 drivers/gpu/drm/vc4/vc4_irq.c 	bin_bo_slot = vc4_v3d_get_bin_slot(vc4);
vc4                80 drivers/gpu/drm/vc4/vc4_irq.c 	spin_lock_irqsave(&vc4->job_lock, irqflags);
vc4                82 drivers/gpu/drm/vc4/vc4_irq.c 	if (vc4->bin_alloc_overflow) {
vc4                89 drivers/gpu/drm/vc4/vc4_irq.c 		exec = vc4_first_bin_job(vc4);
vc4                91 drivers/gpu/drm/vc4/vc4_irq.c 			exec = vc4_last_render_job(vc4);
vc4                93 drivers/gpu/drm/vc4/vc4_irq.c 			exec->bin_slots |= vc4->bin_alloc_overflow;
vc4                98 drivers/gpu/drm/vc4/vc4_irq.c 			vc4->bin_alloc_used &= ~vc4->bin_alloc_overflow;
vc4               101 drivers/gpu/drm/vc4/vc4_irq.c 	vc4->bin_alloc_overflow = BIT(bin_bo_slot);
vc4               103 drivers/gpu/drm/vc4/vc4_irq.c 	V3D_WRITE(V3D_BPOA, bo->base.paddr + bin_bo_slot * vc4->bin_alloc_size);
vc4               107 drivers/gpu/drm/vc4/vc4_irq.c 	spin_unlock_irqrestore(&vc4->job_lock, irqflags);
vc4               110 drivers/gpu/drm/vc4/vc4_irq.c 	mutex_unlock(&vc4->bin_bo_lock);
vc4               116 drivers/gpu/drm/vc4/vc4_irq.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               117 drivers/gpu/drm/vc4/vc4_irq.c 	struct vc4_exec_info *next, *exec = vc4_first_bin_job(vc4);
vc4               123 drivers/gpu/drm/vc4/vc4_irq.c 	next = vc4_first_bin_job(vc4);
vc4               136 drivers/gpu/drm/vc4/vc4_irq.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               137 drivers/gpu/drm/vc4/vc4_irq.c 	struct vc4_exec_info *exec = vc4_first_bin_job(vc4);
vc4               144 drivers/gpu/drm/vc4/vc4_irq.c 		vc4_perfmon_stop(vc4, exec->perfmon, false);
vc4               146 drivers/gpu/drm/vc4/vc4_irq.c 	list_move_tail(&exec->head, &vc4->bin_job_list);
vc4               153 drivers/gpu/drm/vc4/vc4_irq.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               154 drivers/gpu/drm/vc4/vc4_irq.c 	struct vc4_exec_info *exec = vc4_first_render_job(vc4);
vc4               160 drivers/gpu/drm/vc4/vc4_irq.c 	vc4->finished_seqno++;
vc4               161 drivers/gpu/drm/vc4/vc4_irq.c 	list_move_tail(&exec->head, &vc4->job_done_list);
vc4               163 drivers/gpu/drm/vc4/vc4_irq.c 	nextbin = vc4_first_bin_job(vc4);
vc4               164 drivers/gpu/drm/vc4/vc4_irq.c 	nextrender = vc4_first_render_job(vc4);
vc4               171 drivers/gpu/drm/vc4/vc4_irq.c 		vc4_perfmon_stop(vc4, exec->perfmon, true);
vc4               191 drivers/gpu/drm/vc4/vc4_irq.c 	wake_up_all(&vc4->job_wait_queue);
vc4               192 drivers/gpu/drm/vc4/vc4_irq.c 	schedule_work(&vc4->job_done_work);
vc4               199 drivers/gpu/drm/vc4/vc4_irq.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               216 drivers/gpu/drm/vc4/vc4_irq.c 		schedule_work(&vc4->overflow_mem_work);
vc4               221 drivers/gpu/drm/vc4/vc4_irq.c 		spin_lock(&vc4->job_lock);
vc4               223 drivers/gpu/drm/vc4/vc4_irq.c 		spin_unlock(&vc4->job_lock);
vc4               228 drivers/gpu/drm/vc4/vc4_irq.c 		spin_lock(&vc4->job_lock);
vc4               230 drivers/gpu/drm/vc4/vc4_irq.c 		spin_unlock(&vc4->job_lock);
vc4               240 drivers/gpu/drm/vc4/vc4_irq.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               242 drivers/gpu/drm/vc4/vc4_irq.c 	if (!vc4->v3d)
vc4               245 drivers/gpu/drm/vc4/vc4_irq.c 	init_waitqueue_head(&vc4->job_wait_queue);
vc4               246 drivers/gpu/drm/vc4/vc4_irq.c 	INIT_WORK(&vc4->overflow_mem_work, vc4_overflow_mem_work);
vc4               257 drivers/gpu/drm/vc4/vc4_irq.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               259 drivers/gpu/drm/vc4/vc4_irq.c 	if (!vc4->v3d)
vc4               273 drivers/gpu/drm/vc4/vc4_irq.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               275 drivers/gpu/drm/vc4/vc4_irq.c 	if (!vc4->v3d)
vc4               287 drivers/gpu/drm/vc4/vc4_irq.c 	cancel_work_sync(&vc4->overflow_mem_work);
vc4               293 drivers/gpu/drm/vc4/vc4_irq.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               307 drivers/gpu/drm/vc4/vc4_irq.c 	spin_lock_irqsave(&vc4->job_lock, irqflags);
vc4               310 drivers/gpu/drm/vc4/vc4_irq.c 	spin_unlock_irqrestore(&vc4->job_lock, irqflags);
vc4                52 drivers/gpu/drm/vc4/vc4_kms.c 	struct vc4_dev *vc4 = dev->dev_private;
vc4                56 drivers/gpu/drm/vc4/vc4_kms.c 	ret = drm_modeset_lock(&vc4->ctm_state_lock, state->acquire_ctx);
vc4               114 drivers/gpu/drm/vc4/vc4_kms.c vc4_ctm_commit(struct vc4_dev *vc4, struct drm_atomic_state *state)
vc4               116 drivers/gpu/drm/vc4/vc4_kms.c 	struct vc4_ctm_state *ctm_state = to_vc4_ctm_state(vc4->ctm_manager.state);
vc4               151 drivers/gpu/drm/vc4/vc4_kms.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               169 drivers/gpu/drm/vc4/vc4_kms.c 	vc4_ctm_commit(vc4, state);
vc4               187 drivers/gpu/drm/vc4/vc4_kms.c 	up(&vc4->async_modeset);
vc4               215 drivers/gpu/drm/vc4/vc4_kms.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               219 drivers/gpu/drm/vc4/vc4_kms.c 		ret = down_interruptible(&vc4->async_modeset);
vc4               225 drivers/gpu/drm/vc4/vc4_kms.c 			up(&vc4->async_modeset);
vc4               233 drivers/gpu/drm/vc4/vc4_kms.c 		up(&vc4->async_modeset);
vc4               250 drivers/gpu/drm/vc4/vc4_kms.c 	ret = down_interruptible(&vc4->async_modeset);
vc4               256 drivers/gpu/drm/vc4/vc4_kms.c 		up(&vc4->async_modeset);
vc4               264 drivers/gpu/drm/vc4/vc4_kms.c 			up(&vc4->async_modeset);
vc4               348 drivers/gpu/drm/vc4/vc4_kms.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               358 drivers/gpu/drm/vc4/vc4_kms.c 			ctm_state = vc4_get_ctm_state(state, &vc4->ctm_manager);
vc4               370 drivers/gpu/drm/vc4/vc4_kms.c 			ctm_state = vc4_get_ctm_state(state, &vc4->ctm_manager);
vc4               412 drivers/gpu/drm/vc4/vc4_kms.c 	struct vc4_dev *vc4 = to_vc4_dev(state->dev);
vc4               419 drivers/gpu/drm/vc4/vc4_kms.c 						      &vc4->load_tracker);
vc4               442 drivers/gpu/drm/vc4/vc4_kms.c 	if (!vc4->load_tracker_enabled)
vc4               512 drivers/gpu/drm/vc4/vc4_kms.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               520 drivers/gpu/drm/vc4/vc4_kms.c 	vc4->load_tracker_enabled = true;
vc4               522 drivers/gpu/drm/vc4/vc4_kms.c 	sema_init(&vc4->async_modeset, 1);
vc4               541 drivers/gpu/drm/vc4/vc4_kms.c 	drm_modeset_lock_init(&vc4->ctm_state_lock);
vc4               547 drivers/gpu/drm/vc4/vc4_kms.c 	drm_atomic_private_obj_init(dev, &vc4->ctm_manager, &ctm_state->base,
vc4               552 drivers/gpu/drm/vc4/vc4_kms.c 		drm_atomic_private_obj_fini(&vc4->ctm_manager);
vc4               556 drivers/gpu/drm/vc4/vc4_kms.c 	drm_atomic_private_obj_init(dev, &vc4->load_tracker, &load_state->base,
vc4                30 drivers/gpu/drm/vc4/vc4_perfmon.c void vc4_perfmon_start(struct vc4_dev *vc4, struct vc4_perfmon *perfmon)
vc4                35 drivers/gpu/drm/vc4/vc4_perfmon.c 	if (WARN_ON_ONCE(!perfmon || vc4->active_perfmon))
vc4                44 drivers/gpu/drm/vc4/vc4_perfmon.c 	vc4->active_perfmon = perfmon;
vc4                47 drivers/gpu/drm/vc4/vc4_perfmon.c void vc4_perfmon_stop(struct vc4_dev *vc4, struct vc4_perfmon *perfmon,
vc4                52 drivers/gpu/drm/vc4/vc4_perfmon.c 	if (WARN_ON_ONCE(!vc4->active_perfmon ||
vc4                53 drivers/gpu/drm/vc4/vc4_perfmon.c 			 perfmon != vc4->active_perfmon))
vc4                62 drivers/gpu/drm/vc4/vc4_perfmon.c 	vc4->active_perfmon = NULL;
vc4               103 drivers/gpu/drm/vc4/vc4_perfmon.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               110 drivers/gpu/drm/vc4/vc4_perfmon.c 	if (!vc4->v3d) {
vc4               155 drivers/gpu/drm/vc4/vc4_perfmon.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               160 drivers/gpu/drm/vc4/vc4_perfmon.c 	if (!vc4->v3d) {
vc4               179 drivers/gpu/drm/vc4/vc4_perfmon.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               185 drivers/gpu/drm/vc4/vc4_perfmon.c 	if (!vc4->v3d) {
vc4               178 drivers/gpu/drm/vc4/vc4_plane.c 	struct vc4_dev *vc4 = to_vc4_dev(plane->dev);
vc4               184 drivers/gpu/drm/vc4/vc4_plane.c 		spin_lock_irqsave(&vc4->hvs->mm_lock, irqflags);
vc4               186 drivers/gpu/drm/vc4/vc4_plane.c 		spin_unlock_irqrestore(&vc4->hvs->mm_lock, irqflags);
vc4               545 drivers/gpu/drm/vc4/vc4_plane.c 	struct vc4_dev *vc4 = to_vc4_dev(state->plane->dev);
vc4               563 drivers/gpu/drm/vc4/vc4_plane.c 		spin_lock_irqsave(&vc4->hvs->mm_lock, irqflags);
vc4               564 drivers/gpu/drm/vc4/vc4_plane.c 		ret = drm_mm_insert_node_generic(&vc4->hvs->lbm_mm,
vc4               567 drivers/gpu/drm/vc4/vc4_plane.c 		spin_unlock_irqrestore(&vc4->hvs->mm_lock, irqflags);
vc4               586 drivers/gpu/drm/vc4/vc4_plane.c 	struct vc4_dev *vc4 = to_vc4_dev(plane->dev);
vc4               893 drivers/gpu/drm/vc4/vc4_plane.c 			u32 kernel = VC4_SET_FIELD(vc4->hvs->mitchell_netravali_filter.start,
vc4                14 drivers/gpu/drm/vc4/vc4_trace.h #define TRACE_SYSTEM vc4
vc4                59 drivers/gpu/drm/vc4/vc4_trace.h #define TRACE_INCLUDE_PATH ../../drivers/gpu/drm/vc4
vc4               373 drivers/gpu/drm/vc4/vc4_txp.c 	struct vc4_dev *vc4 = to_vc4_dev(drm);
vc4               409 drivers/gpu/drm/vc4/vc4_txp.c 	vc4->txp = txp;
vc4               420 drivers/gpu/drm/vc4/vc4_txp.c 	struct vc4_dev *vc4 = to_vc4_dev(drm);
vc4               425 drivers/gpu/drm/vc4/vc4_txp.c 	vc4->txp = NULL;
vc4               103 drivers/gpu/drm/vc4/vc4_v3d.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               104 drivers/gpu/drm/vc4/vc4_v3d.c 	int ret = vc4_v3d_pm_get(vc4);
vc4               119 drivers/gpu/drm/vc4/vc4_v3d.c 		vc4_v3d_pm_put(vc4);
vc4               130 drivers/gpu/drm/vc4/vc4_v3d.c vc4_v3d_pm_get(struct vc4_dev *vc4)
vc4               132 drivers/gpu/drm/vc4/vc4_v3d.c 	mutex_lock(&vc4->power_lock);
vc4               133 drivers/gpu/drm/vc4/vc4_v3d.c 	if (vc4->power_refcount++ == 0) {
vc4               134 drivers/gpu/drm/vc4/vc4_v3d.c 		int ret = pm_runtime_get_sync(&vc4->v3d->pdev->dev);
vc4               137 drivers/gpu/drm/vc4/vc4_v3d.c 			vc4->power_refcount--;
vc4               138 drivers/gpu/drm/vc4/vc4_v3d.c 			mutex_unlock(&vc4->power_lock);
vc4               142 drivers/gpu/drm/vc4/vc4_v3d.c 	mutex_unlock(&vc4->power_lock);
vc4               148 drivers/gpu/drm/vc4/vc4_v3d.c vc4_v3d_pm_put(struct vc4_dev *vc4)
vc4               150 drivers/gpu/drm/vc4/vc4_v3d.c 	mutex_lock(&vc4->power_lock);
vc4               151 drivers/gpu/drm/vc4/vc4_v3d.c 	if (--vc4->power_refcount == 0) {
vc4               152 drivers/gpu/drm/vc4/vc4_v3d.c 		pm_runtime_mark_last_busy(&vc4->v3d->pdev->dev);
vc4               153 drivers/gpu/drm/vc4/vc4_v3d.c 		pm_runtime_put_autosuspend(&vc4->v3d->pdev->dev);
vc4               155 drivers/gpu/drm/vc4/vc4_v3d.c 	mutex_unlock(&vc4->power_lock);
vc4               160 drivers/gpu/drm/vc4/vc4_v3d.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               169 drivers/gpu/drm/vc4/vc4_v3d.c int vc4_v3d_get_bin_slot(struct vc4_dev *vc4)
vc4               171 drivers/gpu/drm/vc4/vc4_v3d.c 	struct drm_device *dev = vc4->dev;
vc4               178 drivers/gpu/drm/vc4/vc4_v3d.c 	spin_lock_irqsave(&vc4->job_lock, irqflags);
vc4               179 drivers/gpu/drm/vc4/vc4_v3d.c 	slot = ffs(~vc4->bin_alloc_used);
vc4               183 drivers/gpu/drm/vc4/vc4_v3d.c 		vc4->bin_alloc_used |= BIT(slot);
vc4               184 drivers/gpu/drm/vc4/vc4_v3d.c 		spin_unlock_irqrestore(&vc4->job_lock, irqflags);
vc4               191 drivers/gpu/drm/vc4/vc4_v3d.c 	exec = vc4_last_render_job(vc4);
vc4               194 drivers/gpu/drm/vc4/vc4_v3d.c 	spin_unlock_irqrestore(&vc4->job_lock, irqflags);
vc4               230 drivers/gpu/drm/vc4/vc4_v3d.c static int bin_bo_alloc(struct vc4_dev *vc4)
vc4               232 drivers/gpu/drm/vc4/vc4_v3d.c 	struct vc4_v3d *v3d = vc4->v3d;
vc4               249 drivers/gpu/drm/vc4/vc4_v3d.c 		struct vc4_bo *bo = vc4_bo_create(vc4->dev, size, true,
vc4               266 drivers/gpu/drm/vc4/vc4_v3d.c 			vc4->bin_bo = bo;
vc4               284 drivers/gpu/drm/vc4/vc4_v3d.c 			vc4->bin_alloc_size = 512 * 1024;
vc4               285 drivers/gpu/drm/vc4/vc4_v3d.c 			vc4->bin_alloc_used = 0;
vc4               286 drivers/gpu/drm/vc4/vc4_v3d.c 			vc4->bin_alloc_overflow = 0;
vc4               287 drivers/gpu/drm/vc4/vc4_v3d.c 			WARN_ON_ONCE(sizeof(vc4->bin_alloc_used) * 8 !=
vc4               288 drivers/gpu/drm/vc4/vc4_v3d.c 				     bo->base.base.size / vc4->bin_alloc_size);
vc4               290 drivers/gpu/drm/vc4/vc4_v3d.c 			kref_init(&vc4->bin_bo_kref);
vc4               317 drivers/gpu/drm/vc4/vc4_v3d.c int vc4_v3d_bin_bo_get(struct vc4_dev *vc4, bool *used)
vc4               321 drivers/gpu/drm/vc4/vc4_v3d.c 	mutex_lock(&vc4->bin_bo_lock);
vc4               326 drivers/gpu/drm/vc4/vc4_v3d.c 	if (vc4->bin_bo)
vc4               327 drivers/gpu/drm/vc4/vc4_v3d.c 		kref_get(&vc4->bin_bo_kref);
vc4               329 drivers/gpu/drm/vc4/vc4_v3d.c 		ret = bin_bo_alloc(vc4);
vc4               335 drivers/gpu/drm/vc4/vc4_v3d.c 	mutex_unlock(&vc4->bin_bo_lock);
vc4               342 drivers/gpu/drm/vc4/vc4_v3d.c 	struct vc4_dev *vc4 = container_of(ref, struct vc4_dev, bin_bo_kref);
vc4               344 drivers/gpu/drm/vc4/vc4_v3d.c 	if (WARN_ON_ONCE(!vc4->bin_bo))
vc4               347 drivers/gpu/drm/vc4/vc4_v3d.c 	drm_gem_object_put_unlocked(&vc4->bin_bo->base.base);
vc4               348 drivers/gpu/drm/vc4/vc4_v3d.c 	vc4->bin_bo = NULL;
vc4               351 drivers/gpu/drm/vc4/vc4_v3d.c void vc4_v3d_bin_bo_put(struct vc4_dev *vc4)
vc4               353 drivers/gpu/drm/vc4/vc4_v3d.c 	mutex_lock(&vc4->bin_bo_lock);
vc4               354 drivers/gpu/drm/vc4/vc4_v3d.c 	kref_put(&vc4->bin_bo_kref, bin_bo_release);
vc4               355 drivers/gpu/drm/vc4/vc4_v3d.c 	mutex_unlock(&vc4->bin_bo_lock);
vc4               362 drivers/gpu/drm/vc4/vc4_v3d.c 	struct vc4_dev *vc4 = v3d->vc4;
vc4               364 drivers/gpu/drm/vc4/vc4_v3d.c 	vc4_irq_uninstall(vc4->dev);
vc4               374 drivers/gpu/drm/vc4/vc4_v3d.c 	struct vc4_dev *vc4 = v3d->vc4;
vc4               381 drivers/gpu/drm/vc4/vc4_v3d.c 	vc4_v3d_init_hw(vc4->dev);
vc4               384 drivers/gpu/drm/vc4/vc4_v3d.c 	enable_irq(vc4->dev->irq);
vc4               385 drivers/gpu/drm/vc4/vc4_v3d.c 	vc4_irq_postinstall(vc4->dev);
vc4               395 drivers/gpu/drm/vc4/vc4_v3d.c 	struct vc4_dev *vc4 = to_vc4_dev(drm);
vc4               414 drivers/gpu/drm/vc4/vc4_v3d.c 	vc4->v3d = v3d;
vc4               415 drivers/gpu/drm/vc4/vc4_v3d.c 	v3d->vc4 = vc4;
vc4               472 drivers/gpu/drm/vc4/vc4_v3d.c 	struct vc4_dev *vc4 = to_vc4_dev(drm);
vc4               485 drivers/gpu/drm/vc4/vc4_v3d.c 	vc4->v3d = NULL;
vc4               352 drivers/gpu/drm/vc4/vc4_validate.c 	struct vc4_dev *vc4 = to_vc4_dev(dev);
vc4               382 drivers/gpu/drm/vc4/vc4_validate.c 	bin_slot = vc4_v3d_get_bin_slot(vc4);
vc4               395 drivers/gpu/drm/vc4/vc4_validate.c 	bin_addr = vc4->bin_bo->base.paddr + bin_slot * vc4->bin_alloc_size;
vc4               417 drivers/gpu/drm/vc4/vc4_validate.c 	*(uint32_t *)(validated + 4) = (bin_addr + vc4->bin_alloc_size -
vc4               529 drivers/gpu/drm/vc4/vc4_vec.c 	struct vc4_dev *vc4 = to_vc4_dev(drm);
vc4               581 drivers/gpu/drm/vc4/vc4_vec.c 	vc4->vec = vec;
vc4               598 drivers/gpu/drm/vc4/vc4_vec.c 	struct vc4_dev *vc4 = to_vc4_dev(drm);
vc4               605 drivers/gpu/drm/vc4/vc4_vec.c 	vc4->vec = NULL;
vc4               357 drivers/net/dsa/b53/b53_common.c 	u8 mgmt, vc0, vc1, vc4 = 0, vc5;
vc4               364 drivers/net/dsa/b53/b53_common.c 		b53_read8(dev, B53_VLAN_PAGE, B53_VLAN_CTRL4_25, &vc4);
vc4               367 drivers/net/dsa/b53/b53_common.c 		b53_read8(dev, B53_VLAN_PAGE, B53_VLAN_CTRL4_63XX, &vc4);
vc4               370 drivers/net/dsa/b53/b53_common.c 		b53_read8(dev, B53_VLAN_PAGE, B53_VLAN_CTRL4, &vc4);
vc4               379 drivers/net/dsa/b53/b53_common.c 		vc4 &= ~VC4_ING_VID_CHECK_MASK;
vc4               381 drivers/net/dsa/b53/b53_common.c 			vc4 |= VC4_ING_VID_VIO_DROP << VC4_ING_VID_CHECK_S;
vc4               384 drivers/net/dsa/b53/b53_common.c 			vc4 |= VC4_ING_VID_VIO_FWD << VC4_ING_VID_CHECK_S;
vc4               397 drivers/net/dsa/b53/b53_common.c 		vc4 &= ~VC4_ING_VID_CHECK_MASK;
vc4               401 drivers/net/dsa/b53/b53_common.c 			vc4 |= VC4_ING_VID_VIO_FWD << VC4_ING_VID_CHECK_S;
vc4               403 drivers/net/dsa/b53/b53_common.c 			vc4 |= VC4_ING_VID_VIO_TO_IMP << VC4_ING_VID_CHECK_S;
vc4               423 drivers/net/dsa/b53/b53_common.c 		b53_write8(dev, B53_VLAN_PAGE, B53_VLAN_CTRL4_25, vc4);
vc4               427 drivers/net/dsa/b53/b53_common.c 		b53_write8(dev, B53_VLAN_PAGE, B53_VLAN_CTRL4_63XX, vc4);
vc4               431 drivers/net/dsa/b53/b53_common.c 		b53_write8(dev, B53_VLAN_PAGE, B53_VLAN_CTRL4, vc4);
vc4              2312 drivers/net/dsa/b53/b53_common.c 		u8 vc4;
vc4              2314 drivers/net/dsa/b53/b53_common.c 		b53_read8(dev, B53_VLAN_PAGE, B53_VLAN_CTRL4_25, &vc4);
vc4              2317 drivers/net/dsa/b53/b53_common.c 		switch (vc4 & 3) {