vc4 40 drivers/gpu/drm/vc4/vc4_bo.c static void vc4_bo_stats_print(struct drm_printer *p, struct vc4_dev *vc4) vc4 44 drivers/gpu/drm/vc4/vc4_bo.c for (i = 0; i < vc4->num_labels; i++) { vc4 45 drivers/gpu/drm/vc4/vc4_bo.c if (!vc4->bo_labels[i].num_allocated) vc4 49 drivers/gpu/drm/vc4/vc4_bo.c vc4->bo_labels[i].name, vc4 50 drivers/gpu/drm/vc4/vc4_bo.c vc4->bo_labels[i].size_allocated / 1024, vc4 51 drivers/gpu/drm/vc4/vc4_bo.c vc4->bo_labels[i].num_allocated); vc4 54 drivers/gpu/drm/vc4/vc4_bo.c mutex_lock(&vc4->purgeable.lock); vc4 55 drivers/gpu/drm/vc4/vc4_bo.c if (vc4->purgeable.num) vc4 57 drivers/gpu/drm/vc4/vc4_bo.c vc4->purgeable.size / 1024, vc4->purgeable.num); vc4 59 drivers/gpu/drm/vc4/vc4_bo.c if (vc4->purgeable.purged_num) vc4 61 drivers/gpu/drm/vc4/vc4_bo.c vc4->purgeable.purged_size / 1024, vc4 62 drivers/gpu/drm/vc4/vc4_bo.c vc4->purgeable.purged_num); vc4 63 drivers/gpu/drm/vc4/vc4_bo.c mutex_unlock(&vc4->purgeable.lock); vc4 70 drivers/gpu/drm/vc4/vc4_bo.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 73 drivers/gpu/drm/vc4/vc4_bo.c vc4_bo_stats_print(&p, vc4); vc4 86 drivers/gpu/drm/vc4/vc4_bo.c static int vc4_get_user_label(struct vc4_dev *vc4, const char *name) vc4 91 drivers/gpu/drm/vc4/vc4_bo.c for (i = 0; i < vc4->num_labels; i++) { vc4 92 drivers/gpu/drm/vc4/vc4_bo.c if (!vc4->bo_labels[i].name) { vc4 94 drivers/gpu/drm/vc4/vc4_bo.c } else if (strcmp(vc4->bo_labels[i].name, name) == 0) { vc4 101 drivers/gpu/drm/vc4/vc4_bo.c WARN_ON(vc4->bo_labels[free_slot].num_allocated != 0); vc4 102 drivers/gpu/drm/vc4/vc4_bo.c vc4->bo_labels[free_slot].name = name; vc4 105 drivers/gpu/drm/vc4/vc4_bo.c u32 new_label_count = vc4->num_labels + 1; vc4 107 drivers/gpu/drm/vc4/vc4_bo.c krealloc(vc4->bo_labels, vc4 116 drivers/gpu/drm/vc4/vc4_bo.c free_slot = vc4->num_labels; vc4 117 drivers/gpu/drm/vc4/vc4_bo.c vc4->bo_labels = new_labels; vc4 118 drivers/gpu/drm/vc4/vc4_bo.c vc4->num_labels = new_label_count; vc4 120 drivers/gpu/drm/vc4/vc4_bo.c vc4->bo_labels[free_slot].name = name; vc4 121 drivers/gpu/drm/vc4/vc4_bo.c vc4->bo_labels[free_slot].num_allocated = 0; vc4 122 drivers/gpu/drm/vc4/vc4_bo.c vc4->bo_labels[free_slot].size_allocated = 0; vc4 131 drivers/gpu/drm/vc4/vc4_bo.c struct vc4_dev *vc4 = to_vc4_dev(gem_obj->dev); vc4 133 drivers/gpu/drm/vc4/vc4_bo.c lockdep_assert_held(&vc4->bo_lock); vc4 136 drivers/gpu/drm/vc4/vc4_bo.c vc4->bo_labels[label].num_allocated++; vc4 137 drivers/gpu/drm/vc4/vc4_bo.c vc4->bo_labels[label].size_allocated += gem_obj->size; vc4 140 drivers/gpu/drm/vc4/vc4_bo.c vc4->bo_labels[bo->label].num_allocated--; vc4 141 drivers/gpu/drm/vc4/vc4_bo.c vc4->bo_labels[bo->label].size_allocated -= gem_obj->size; vc4 143 drivers/gpu/drm/vc4/vc4_bo.c if (vc4->bo_labels[bo->label].num_allocated == 0 && vc4 150 drivers/gpu/drm/vc4/vc4_bo.c kfree(vc4->bo_labels[bo->label].name); vc4 151 drivers/gpu/drm/vc4/vc4_bo.c vc4->bo_labels[bo->label].name = NULL; vc4 165 drivers/gpu/drm/vc4/vc4_bo.c struct vc4_dev *vc4 = to_vc4_dev(obj->dev); vc4 167 drivers/gpu/drm/vc4/vc4_bo.c lockdep_assert_held(&vc4->bo_lock); vc4 183 drivers/gpu/drm/vc4/vc4_bo.c struct vc4_dev *vc4 = to_vc4_dev(bo->base.base.dev); vc4 185 drivers/gpu/drm/vc4/vc4_bo.c lockdep_assert_held(&vc4->bo_lock); vc4 193 drivers/gpu/drm/vc4/vc4_bo.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 196 drivers/gpu/drm/vc4/vc4_bo.c if (vc4->bo_cache.size_list_size <= page_index) { vc4 197 drivers/gpu/drm/vc4/vc4_bo.c uint32_t new_size = max(vc4->bo_cache.size_list_size * 2, vc4 210 drivers/gpu/drm/vc4/vc4_bo.c for (i = 0; i < vc4->bo_cache.size_list_size; i++) { vc4 212 drivers/gpu/drm/vc4/vc4_bo.c &vc4->bo_cache.size_list[i]; vc4 220 drivers/gpu/drm/vc4/vc4_bo.c for (i = vc4->bo_cache.size_list_size; i < new_size; i++) vc4 223 drivers/gpu/drm/vc4/vc4_bo.c kfree(vc4->bo_cache.size_list); vc4 224 drivers/gpu/drm/vc4/vc4_bo.c vc4->bo_cache.size_list = new_list; vc4 225 drivers/gpu/drm/vc4/vc4_bo.c vc4->bo_cache.size_list_size = new_size; vc4 228 drivers/gpu/drm/vc4/vc4_bo.c return &vc4->bo_cache.size_list[page_index]; vc4 233 drivers/gpu/drm/vc4/vc4_bo.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 235 drivers/gpu/drm/vc4/vc4_bo.c mutex_lock(&vc4->bo_lock); vc4 236 drivers/gpu/drm/vc4/vc4_bo.c while (!list_empty(&vc4->bo_cache.time_list)) { vc4 237 drivers/gpu/drm/vc4/vc4_bo.c struct vc4_bo *bo = list_last_entry(&vc4->bo_cache.time_list, vc4 242 drivers/gpu/drm/vc4/vc4_bo.c mutex_unlock(&vc4->bo_lock); vc4 247 drivers/gpu/drm/vc4/vc4_bo.c struct vc4_dev *vc4 = to_vc4_dev(bo->base.base.dev); vc4 249 drivers/gpu/drm/vc4/vc4_bo.c mutex_lock(&vc4->purgeable.lock); vc4 250 drivers/gpu/drm/vc4/vc4_bo.c list_add_tail(&bo->size_head, &vc4->purgeable.list); vc4 251 drivers/gpu/drm/vc4/vc4_bo.c vc4->purgeable.num++; vc4 252 drivers/gpu/drm/vc4/vc4_bo.c vc4->purgeable.size += bo->base.base.size; vc4 253 drivers/gpu/drm/vc4/vc4_bo.c mutex_unlock(&vc4->purgeable.lock); vc4 258 drivers/gpu/drm/vc4/vc4_bo.c struct vc4_dev *vc4 = to_vc4_dev(bo->base.base.dev); vc4 273 drivers/gpu/drm/vc4/vc4_bo.c vc4->purgeable.num--; vc4 274 drivers/gpu/drm/vc4/vc4_bo.c vc4->purgeable.size -= bo->base.base.size; vc4 279 drivers/gpu/drm/vc4/vc4_bo.c struct vc4_dev *vc4 = to_vc4_dev(bo->base.base.dev); vc4 281 drivers/gpu/drm/vc4/vc4_bo.c mutex_lock(&vc4->purgeable.lock); vc4 283 drivers/gpu/drm/vc4/vc4_bo.c mutex_unlock(&vc4->purgeable.lock); vc4 303 drivers/gpu/drm/vc4/vc4_bo.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 305 drivers/gpu/drm/vc4/vc4_bo.c mutex_lock(&vc4->purgeable.lock); vc4 306 drivers/gpu/drm/vc4/vc4_bo.c while (!list_empty(&vc4->purgeable.list)) { vc4 307 drivers/gpu/drm/vc4/vc4_bo.c struct vc4_bo *bo = list_first_entry(&vc4->purgeable.list, vc4 319 drivers/gpu/drm/vc4/vc4_bo.c mutex_unlock(&vc4->purgeable.lock); vc4 338 drivers/gpu/drm/vc4/vc4_bo.c mutex_lock(&vc4->purgeable.lock); vc4 341 drivers/gpu/drm/vc4/vc4_bo.c vc4->purgeable.purged_size += purged_size; vc4 342 drivers/gpu/drm/vc4/vc4_bo.c vc4->purgeable.purged_num++; vc4 345 drivers/gpu/drm/vc4/vc4_bo.c mutex_unlock(&vc4->purgeable.lock); vc4 352 drivers/gpu/drm/vc4/vc4_bo.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 358 drivers/gpu/drm/vc4/vc4_bo.c mutex_lock(&vc4->bo_lock); vc4 359 drivers/gpu/drm/vc4/vc4_bo.c if (page_index >= vc4->bo_cache.size_list_size) vc4 362 drivers/gpu/drm/vc4/vc4_bo.c if (list_empty(&vc4->bo_cache.size_list[page_index])) vc4 365 drivers/gpu/drm/vc4/vc4_bo.c bo = list_first_entry(&vc4->bo_cache.size_list[page_index], vc4 373 drivers/gpu/drm/vc4/vc4_bo.c mutex_unlock(&vc4->bo_lock); vc4 387 drivers/gpu/drm/vc4/vc4_bo.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 397 drivers/gpu/drm/vc4/vc4_bo.c mutex_lock(&vc4->bo_lock); vc4 399 drivers/gpu/drm/vc4/vc4_bo.c vc4->bo_labels[VC4_BO_TYPE_KERNEL].num_allocated++; vc4 400 drivers/gpu/drm/vc4/vc4_bo.c vc4->bo_labels[VC4_BO_TYPE_KERNEL].size_allocated += size; vc4 401 drivers/gpu/drm/vc4/vc4_bo.c mutex_unlock(&vc4->bo_lock); vc4 410 drivers/gpu/drm/vc4/vc4_bo.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 452 drivers/gpu/drm/vc4/vc4_bo.c struct drm_printer p = drm_info_printer(vc4->dev->dev); vc4 454 drivers/gpu/drm/vc4/vc4_bo.c vc4_bo_stats_print(&p, vc4); vc4 465 drivers/gpu/drm/vc4/vc4_bo.c mutex_lock(&vc4->bo_lock); vc4 467 drivers/gpu/drm/vc4/vc4_bo.c mutex_unlock(&vc4->bo_lock); vc4 500 drivers/gpu/drm/vc4/vc4_bo.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 503 drivers/gpu/drm/vc4/vc4_bo.c lockdep_assert_held(&vc4->bo_lock); vc4 505 drivers/gpu/drm/vc4/vc4_bo.c while (!list_empty(&vc4->bo_cache.time_list)) { vc4 506 drivers/gpu/drm/vc4/vc4_bo.c struct vc4_bo *bo = list_last_entry(&vc4->bo_cache.time_list, vc4 509 drivers/gpu/drm/vc4/vc4_bo.c mod_timer(&vc4->bo_cache.time_timer, vc4 526 drivers/gpu/drm/vc4/vc4_bo.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 536 drivers/gpu/drm/vc4/vc4_bo.c mutex_lock(&vc4->bo_lock); vc4 579 drivers/gpu/drm/vc4/vc4_bo.c list_add(&bo->unref_head, &vc4->bo_cache.time_list); vc4 586 drivers/gpu/drm/vc4/vc4_bo.c mutex_unlock(&vc4->bo_lock); vc4 591 drivers/gpu/drm/vc4/vc4_bo.c struct vc4_dev *vc4 = vc4 593 drivers/gpu/drm/vc4/vc4_bo.c struct drm_device *dev = vc4->dev; vc4 595 drivers/gpu/drm/vc4/vc4_bo.c mutex_lock(&vc4->bo_lock); vc4 597 drivers/gpu/drm/vc4/vc4_bo.c mutex_unlock(&vc4->bo_lock); vc4 653 drivers/gpu/drm/vc4/vc4_bo.c struct vc4_dev *vc4 = from_timer(vc4, t, bo_cache.time_timer); vc4 655 drivers/gpu/drm/vc4/vc4_bo.c schedule_work(&vc4->bo_cache.time_work); vc4 796 drivers/gpu/drm/vc4/vc4_bo.c static int vc4_grab_bin_bo(struct vc4_dev *vc4, struct vc4_file *vc4file) vc4 800 drivers/gpu/drm/vc4/vc4_bo.c if (!vc4->v3d) vc4 806 drivers/gpu/drm/vc4/vc4_bo.c ret = vc4_v3d_bin_bo_get(vc4, &vc4file->bin_bo_used); vc4 818 drivers/gpu/drm/vc4/vc4_bo.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 822 drivers/gpu/drm/vc4/vc4_bo.c ret = vc4_grab_bin_bo(vc4, vc4file); vc4 867 drivers/gpu/drm/vc4/vc4_bo.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 887 drivers/gpu/drm/vc4/vc4_bo.c ret = vc4_grab_bin_bo(vc4, vc4file); vc4 1010 drivers/gpu/drm/vc4/vc4_bo.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 1017 drivers/gpu/drm/vc4/vc4_bo.c vc4->bo_labels = kcalloc(VC4_BO_TYPE_COUNT, sizeof(*vc4->bo_labels), vc4 1019 drivers/gpu/drm/vc4/vc4_bo.c if (!vc4->bo_labels) vc4 1021 drivers/gpu/drm/vc4/vc4_bo.c vc4->num_labels = VC4_BO_TYPE_COUNT; vc4 1025 drivers/gpu/drm/vc4/vc4_bo.c vc4->bo_labels[i].name = bo_type_names[i]; vc4 1027 drivers/gpu/drm/vc4/vc4_bo.c mutex_init(&vc4->bo_lock); vc4 1031 drivers/gpu/drm/vc4/vc4_bo.c INIT_LIST_HEAD(&vc4->bo_cache.time_list); vc4 1033 drivers/gpu/drm/vc4/vc4_bo.c INIT_WORK(&vc4->bo_cache.time_work, vc4_bo_cache_time_work); vc4 1034 drivers/gpu/drm/vc4/vc4_bo.c timer_setup(&vc4->bo_cache.time_timer, vc4_bo_cache_time_timer, 0); vc4 1041 drivers/gpu/drm/vc4/vc4_bo.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 1044 drivers/gpu/drm/vc4/vc4_bo.c del_timer(&vc4->bo_cache.time_timer); vc4 1045 drivers/gpu/drm/vc4/vc4_bo.c cancel_work_sync(&vc4->bo_cache.time_work); vc4 1049 drivers/gpu/drm/vc4/vc4_bo.c for (i = 0; i < vc4->num_labels; i++) { vc4 1050 drivers/gpu/drm/vc4/vc4_bo.c if (vc4->bo_labels[i].num_allocated) { vc4 1053 drivers/gpu/drm/vc4/vc4_bo.c vc4->bo_labels[i].num_allocated, vc4 1054 drivers/gpu/drm/vc4/vc4_bo.c vc4->bo_labels[i].name); vc4 1058 drivers/gpu/drm/vc4/vc4_bo.c kfree(vc4->bo_labels[i].name); vc4 1060 drivers/gpu/drm/vc4/vc4_bo.c kfree(vc4->bo_labels); vc4 1066 drivers/gpu/drm/vc4/vc4_bo.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 1086 drivers/gpu/drm/vc4/vc4_bo.c mutex_lock(&vc4->bo_lock); vc4 1087 drivers/gpu/drm/vc4/vc4_bo.c label = vc4_get_user_label(vc4, name); vc4 1092 drivers/gpu/drm/vc4/vc4_bo.c mutex_unlock(&vc4->bo_lock); vc4 92 drivers/gpu/drm/vc4/vc4_crtc.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 214 drivers/gpu/drm/vc4/vc4_crtc.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 385 drivers/gpu/drm/vc4/vc4_crtc.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 446 drivers/gpu/drm/vc4/vc4_crtc.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 456 drivers/gpu/drm/vc4/vc4_crtc.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 517 drivers/gpu/drm/vc4/vc4_crtc.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 549 drivers/gpu/drm/vc4/vc4_crtc.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 630 drivers/gpu/drm/vc4/vc4_crtc.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 650 drivers/gpu/drm/vc4/vc4_crtc.c spin_lock_irqsave(&vc4->hvs->mm_lock, flags); vc4 651 drivers/gpu/drm/vc4/vc4_crtc.c ret = drm_mm_insert_node(&vc4->hvs->dlist_mm, &vc4_state->mm, vc4 653 drivers/gpu/drm/vc4/vc4_crtc.c spin_unlock_irqrestore(&vc4->hvs->mm_lock, flags); vc4 686 drivers/gpu/drm/vc4/vc4_crtc.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 693 drivers/gpu/drm/vc4/vc4_crtc.c u32 __iomem *dlist_start = vc4->hvs->dlist + vc4_state->mm.start; vc4 785 drivers/gpu/drm/vc4/vc4_crtc.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 850 drivers/gpu/drm/vc4/vc4_crtc.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 883 drivers/gpu/drm/vc4/vc4_crtc.c up(&vc4->async_modeset); vc4 898 drivers/gpu/drm/vc4/vc4_crtc.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 928 drivers/gpu/drm/vc4/vc4_crtc.c ret = down_interruptible(&vc4->async_modeset); vc4 994 drivers/gpu/drm/vc4/vc4_crtc.c struct vc4_dev *vc4 = to_vc4_dev(crtc->dev); vc4 1000 drivers/gpu/drm/vc4/vc4_crtc.c spin_lock_irqsave(&vc4->hvs->mm_lock, flags); vc4 1002 drivers/gpu/drm/vc4/vc4_crtc.c spin_unlock_irqrestore(&vc4->hvs->mm_lock, flags); vc4 1112 drivers/gpu/drm/vc4/vc4_crtc.c struct vc4_dev *vc4 = to_vc4_dev(drm); vc4 26 drivers/gpu/drm/vc4/vc4_debugfs.c struct vc4_dev *vc4 = to_vc4_dev(minor->dev); vc4 30 drivers/gpu/drm/vc4/vc4_debugfs.c minor->debugfs_root, &vc4->load_tracker_enabled); vc4 32 drivers/gpu/drm/vc4/vc4_debugfs.c list_for_each_entry(entry, &vc4->debugfs_list, link) { vc4 68 drivers/gpu/drm/vc4/vc4_debugfs.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 80 drivers/gpu/drm/vc4/vc4_debugfs.c list_add(&entry->link, &vc4->debugfs_list); vc4 261 drivers/gpu/drm/vc4/vc4_dpi.c struct vc4_dev *vc4 = to_vc4_dev(drm); vc4 321 drivers/gpu/drm/vc4/vc4_dpi.c vc4->dpi = dpi; vc4 337 drivers/gpu/drm/vc4/vc4_dpi.c struct vc4_dev *vc4 = to_vc4_dev(drm); vc4 346 drivers/gpu/drm/vc4/vc4_dpi.c vc4->dpi = NULL; vc4 70 drivers/gpu/drm/vc4/vc4_drv.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 77 drivers/gpu/drm/vc4/vc4_drv.c if (!vc4->v3d) vc4 82 drivers/gpu/drm/vc4/vc4_drv.c ret = vc4_v3d_pm_get(vc4); vc4 86 drivers/gpu/drm/vc4/vc4_drv.c vc4_v3d_pm_put(vc4); vc4 89 drivers/gpu/drm/vc4/vc4_drv.c ret = vc4_v3d_pm_get(vc4); vc4 93 drivers/gpu/drm/vc4/vc4_drv.c vc4_v3d_pm_put(vc4); vc4 96 drivers/gpu/drm/vc4/vc4_drv.c ret = vc4_v3d_pm_get(vc4); vc4 100 drivers/gpu/drm/vc4/vc4_drv.c vc4_v3d_pm_put(vc4); vc4 133 drivers/gpu/drm/vc4/vc4_drv.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 137 drivers/gpu/drm/vc4/vc4_drv.c vc4_v3d_bin_bo_put(vc4); vc4 256 drivers/gpu/drm/vc4/vc4_drv.c struct vc4_dev *vc4; vc4 262 drivers/gpu/drm/vc4/vc4_drv.c vc4 = devm_kzalloc(dev, sizeof(*vc4), GFP_KERNEL); vc4 263 drivers/gpu/drm/vc4/vc4_drv.c if (!vc4) vc4 276 drivers/gpu/drm/vc4/vc4_drv.c vc4->dev = drm; vc4 277 drivers/gpu/drm/vc4/vc4_drv.c drm->dev_private = vc4; vc4 278 drivers/gpu/drm/vc4/vc4_drv.c INIT_LIST_HEAD(&vc4->debugfs_list); vc4 280 drivers/gpu/drm/vc4/vc4_drv.c mutex_init(&vc4->bin_bo_lock); vc4 321 drivers/gpu/drm/vc4/vc4_drv.c struct vc4_dev *vc4 = to_vc4_dev(drm); vc4 329 drivers/gpu/drm/vc4/vc4_drv.c drm_atomic_private_obj_fini(&vc4->load_tracker); vc4 330 drivers/gpu/drm/vc4/vc4_drv.c drm_atomic_private_obj_fini(&vc4->ctm_manager); vc4 309 drivers/gpu/drm/vc4/vc4_drv.h struct vc4_dev *vc4; vc4 478 drivers/gpu/drm/vc4/vc4_drv.h #define V3D_READ(offset) readl(vc4->v3d->regs + offset) vc4 479 drivers/gpu/drm/vc4/vc4_drv.h #define V3D_WRITE(offset, val) writel(val, vc4->v3d->regs + offset) vc4 480 drivers/gpu/drm/vc4/vc4_drv.h #define HVS_READ(offset) readl(vc4->hvs->regs + offset) vc4 481 drivers/gpu/drm/vc4/vc4_drv.h #define HVS_WRITE(offset, val) writel(val, vc4->hvs->regs + offset) vc4 616 drivers/gpu/drm/vc4/vc4_drv.h vc4_first_bin_job(struct vc4_dev *vc4) vc4 618 drivers/gpu/drm/vc4/vc4_drv.h return list_first_entry_or_null(&vc4->bin_job_list, vc4 623 drivers/gpu/drm/vc4/vc4_drv.h vc4_first_render_job(struct vc4_dev *vc4) vc4 625 drivers/gpu/drm/vc4/vc4_drv.h return list_first_entry_or_null(&vc4->render_job_list, vc4 630 drivers/gpu/drm/vc4/vc4_drv.h vc4_last_render_job(struct vc4_dev *vc4) vc4 632 drivers/gpu/drm/vc4/vc4_drv.h if (list_empty(&vc4->render_job_list)) vc4 634 drivers/gpu/drm/vc4/vc4_drv.h return list_last_entry(&vc4->render_job_list, vc4 807 drivers/gpu/drm/vc4/vc4_drv.h void vc4_job_handle_completed(struct vc4_dev *vc4); vc4 850 drivers/gpu/drm/vc4/vc4_drv.h int vc4_v3d_get_bin_slot(struct vc4_dev *vc4); vc4 851 drivers/gpu/drm/vc4/vc4_drv.h int vc4_v3d_bin_bo_get(struct vc4_dev *vc4, bool *used); vc4 852 drivers/gpu/drm/vc4/vc4_drv.h void vc4_v3d_bin_bo_put(struct vc4_dev *vc4); vc4 853 drivers/gpu/drm/vc4/vc4_drv.h int vc4_v3d_pm_get(struct vc4_dev *vc4); vc4 854 drivers/gpu/drm/vc4/vc4_drv.h void vc4_v3d_pm_put(struct vc4_dev *vc4); vc4 883 drivers/gpu/drm/vc4/vc4_drv.h void vc4_perfmon_start(struct vc4_dev *vc4, struct vc4_perfmon *perfmon); vc4 884 drivers/gpu/drm/vc4/vc4_drv.h void vc4_perfmon_stop(struct vc4_dev *vc4, struct vc4_perfmon *perfmon, vc4 1444 drivers/gpu/drm/vc4/vc4_dsi.c struct vc4_dev *vc4 = to_vc4_dev(drm); vc4 1596 drivers/gpu/drm/vc4/vc4_dsi.c vc4->dsi1 = dsi; vc4 1628 drivers/gpu/drm/vc4/vc4_dsi.c struct vc4_dev *vc4 = to_vc4_dev(drm); vc4 1637 drivers/gpu/drm/vc4/vc4_dsi.c vc4->dsi1 = NULL; vc4 39 drivers/gpu/drm/vc4/vc4_fence.c struct vc4_dev *vc4 = to_vc4_dev(f->dev); vc4 41 drivers/gpu/drm/vc4/vc4_fence.c return vc4->finished_seqno >= f->seqno; vc4 42 drivers/gpu/drm/vc4/vc4_gem.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 44 drivers/gpu/drm/vc4/vc4_gem.c mod_timer(&vc4->hangcheck.timer, vc4 74 drivers/gpu/drm/vc4/vc4_gem.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 79 drivers/gpu/drm/vc4/vc4_gem.c if (!vc4->v3d) { vc4 84 drivers/gpu/drm/vc4/vc4_gem.c spin_lock_irqsave(&vc4->job_lock, irqflags); vc4 85 drivers/gpu/drm/vc4/vc4_gem.c kernel_state = vc4->hang_state; vc4 87 drivers/gpu/drm/vc4/vc4_gem.c spin_unlock_irqrestore(&vc4->job_lock, irqflags); vc4 97 drivers/gpu/drm/vc4/vc4_gem.c spin_unlock_irqrestore(&vc4->job_lock, irqflags); vc4 101 drivers/gpu/drm/vc4/vc4_gem.c vc4->hang_state = NULL; vc4 102 drivers/gpu/drm/vc4/vc4_gem.c spin_unlock_irqrestore(&vc4->job_lock, irqflags); vc4 151 drivers/gpu/drm/vc4/vc4_gem.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 165 drivers/gpu/drm/vc4/vc4_gem.c spin_lock_irqsave(&vc4->job_lock, irqflags); vc4 166 drivers/gpu/drm/vc4/vc4_gem.c exec[0] = vc4_first_bin_job(vc4); vc4 167 drivers/gpu/drm/vc4/vc4_gem.c exec[1] = vc4_first_render_job(vc4); vc4 169 drivers/gpu/drm/vc4/vc4_gem.c spin_unlock_irqrestore(&vc4->job_lock, irqflags); vc4 189 drivers/gpu/drm/vc4/vc4_gem.c spin_unlock_irqrestore(&vc4->job_lock, irqflags); vc4 227 drivers/gpu/drm/vc4/vc4_gem.c spin_unlock_irqrestore(&vc4->job_lock, irqflags); vc4 275 drivers/gpu/drm/vc4/vc4_gem.c spin_lock_irqsave(&vc4->job_lock, irqflags); vc4 276 drivers/gpu/drm/vc4/vc4_gem.c if (vc4->hang_state) { vc4 277 drivers/gpu/drm/vc4/vc4_gem.c spin_unlock_irqrestore(&vc4->job_lock, irqflags); vc4 280 drivers/gpu/drm/vc4/vc4_gem.c vc4->hang_state = kernel_state; vc4 281 drivers/gpu/drm/vc4/vc4_gem.c spin_unlock_irqrestore(&vc4->job_lock, irqflags); vc4 288 drivers/gpu/drm/vc4/vc4_gem.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 292 drivers/gpu/drm/vc4/vc4_gem.c mutex_lock(&vc4->power_lock); vc4 293 drivers/gpu/drm/vc4/vc4_gem.c if (vc4->power_refcount) { vc4 297 drivers/gpu/drm/vc4/vc4_gem.c pm_runtime_put_sync_suspend(&vc4->v3d->pdev->dev); vc4 298 drivers/gpu/drm/vc4/vc4_gem.c pm_runtime_get_sync(&vc4->v3d->pdev->dev); vc4 300 drivers/gpu/drm/vc4/vc4_gem.c mutex_unlock(&vc4->power_lock); vc4 314 drivers/gpu/drm/vc4/vc4_gem.c struct vc4_dev *vc4 = vc4 317 drivers/gpu/drm/vc4/vc4_gem.c vc4_save_hang_state(vc4->dev); vc4 319 drivers/gpu/drm/vc4/vc4_gem.c vc4_reset(vc4->dev); vc4 325 drivers/gpu/drm/vc4/vc4_gem.c struct vc4_dev *vc4 = from_timer(vc4, t, hangcheck.timer); vc4 326 drivers/gpu/drm/vc4/vc4_gem.c struct drm_device *dev = vc4->dev; vc4 331 drivers/gpu/drm/vc4/vc4_gem.c spin_lock_irqsave(&vc4->job_lock, irqflags); vc4 333 drivers/gpu/drm/vc4/vc4_gem.c bin_exec = vc4_first_bin_job(vc4); vc4 334 drivers/gpu/drm/vc4/vc4_gem.c render_exec = vc4_first_render_job(vc4); vc4 338 drivers/gpu/drm/vc4/vc4_gem.c spin_unlock_irqrestore(&vc4->job_lock, irqflags); vc4 354 drivers/gpu/drm/vc4/vc4_gem.c spin_unlock_irqrestore(&vc4->job_lock, irqflags); vc4 359 drivers/gpu/drm/vc4/vc4_gem.c spin_unlock_irqrestore(&vc4->job_lock, irqflags); vc4 365 drivers/gpu/drm/vc4/vc4_gem.c schedule_work(&vc4->hangcheck.reset_work); vc4 371 drivers/gpu/drm/vc4/vc4_gem.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 384 drivers/gpu/drm/vc4/vc4_gem.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 389 drivers/gpu/drm/vc4/vc4_gem.c if (vc4->finished_seqno >= seqno) vc4 399 drivers/gpu/drm/vc4/vc4_gem.c prepare_to_wait(&vc4->job_wait_queue, &wait, vc4 408 drivers/gpu/drm/vc4/vc4_gem.c if (vc4->finished_seqno >= seqno) vc4 422 drivers/gpu/drm/vc4/vc4_gem.c finish_wait(&vc4->job_wait_queue, &wait); vc4 431 drivers/gpu/drm/vc4/vc4_gem.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 450 drivers/gpu/drm/vc4/vc4_gem.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 468 drivers/gpu/drm/vc4/vc4_gem.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 472 drivers/gpu/drm/vc4/vc4_gem.c exec = vc4_first_bin_job(vc4); vc4 481 drivers/gpu/drm/vc4/vc4_gem.c if (exec->perfmon && vc4->active_perfmon != exec->perfmon) vc4 482 drivers/gpu/drm/vc4/vc4_gem.c vc4_perfmon_start(vc4, exec->perfmon); vc4 493 drivers/gpu/drm/vc4/vc4_gem.c next = vc4_first_bin_job(vc4); vc4 508 drivers/gpu/drm/vc4/vc4_gem.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 509 drivers/gpu/drm/vc4/vc4_gem.c struct vc4_exec_info *exec = vc4_first_render_job(vc4); vc4 528 drivers/gpu/drm/vc4/vc4_gem.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 529 drivers/gpu/drm/vc4/vc4_gem.c bool was_empty = list_empty(&vc4->render_job_list); vc4 531 drivers/gpu/drm/vc4/vc4_gem.c list_move_tail(&exec->head, &vc4->render_job_list); vc4 669 drivers/gpu/drm/vc4/vc4_gem.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 680 drivers/gpu/drm/vc4/vc4_gem.c spin_lock_irqsave(&vc4->job_lock, irqflags); vc4 682 drivers/gpu/drm/vc4/vc4_gem.c seqno = ++vc4->emit_seqno; vc4 685 drivers/gpu/drm/vc4/vc4_gem.c dma_fence_init(&fence->base, &vc4_fence_ops, &vc4->job_lock, vc4 686 drivers/gpu/drm/vc4/vc4_gem.c vc4->dma_fence_context, exec->seqno); vc4 697 drivers/gpu/drm/vc4/vc4_gem.c list_add_tail(&exec->head, &vc4->bin_job_list); vc4 704 drivers/gpu/drm/vc4/vc4_gem.c renderjob = vc4_first_render_job(vc4); vc4 705 drivers/gpu/drm/vc4/vc4_gem.c if (vc4_first_bin_job(vc4) == exec && vc4 711 drivers/gpu/drm/vc4/vc4_gem.c spin_unlock_irqrestore(&vc4->job_lock, irqflags); vc4 825 drivers/gpu/drm/vc4/vc4_gem.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 925 drivers/gpu/drm/vc4/vc4_gem.c ret = vc4_v3d_bin_bo_get(vc4, &exec->bin_bo_used); vc4 944 drivers/gpu/drm/vc4/vc4_gem.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 974 drivers/gpu/drm/vc4/vc4_gem.c spin_lock_irqsave(&vc4->job_lock, irqflags); vc4 975 drivers/gpu/drm/vc4/vc4_gem.c vc4->bin_alloc_used &= ~exec->bin_slots; vc4 976 drivers/gpu/drm/vc4/vc4_gem.c spin_unlock_irqrestore(&vc4->job_lock, irqflags); vc4 980 drivers/gpu/drm/vc4/vc4_gem.c vc4_v3d_bin_bo_put(vc4); vc4 985 drivers/gpu/drm/vc4/vc4_gem.c vc4_v3d_pm_put(vc4); vc4 991 drivers/gpu/drm/vc4/vc4_gem.c vc4_job_handle_completed(struct vc4_dev *vc4) vc4 996 drivers/gpu/drm/vc4/vc4_gem.c spin_lock_irqsave(&vc4->job_lock, irqflags); vc4 997 drivers/gpu/drm/vc4/vc4_gem.c while (!list_empty(&vc4->job_done_list)) { vc4 999 drivers/gpu/drm/vc4/vc4_gem.c list_first_entry(&vc4->job_done_list, vc4 1003 drivers/gpu/drm/vc4/vc4_gem.c spin_unlock_irqrestore(&vc4->job_lock, irqflags); vc4 1004 drivers/gpu/drm/vc4/vc4_gem.c vc4_complete_exec(vc4->dev, exec); vc4 1005 drivers/gpu/drm/vc4/vc4_gem.c spin_lock_irqsave(&vc4->job_lock, irqflags); vc4 1008 drivers/gpu/drm/vc4/vc4_gem.c list_for_each_entry_safe(cb, cb_temp, &vc4->seqno_cb_list, work.entry) { vc4 1009 drivers/gpu/drm/vc4/vc4_gem.c if (cb->seqno <= vc4->finished_seqno) { vc4 1015 drivers/gpu/drm/vc4/vc4_gem.c spin_unlock_irqrestore(&vc4->job_lock, irqflags); vc4 1029 drivers/gpu/drm/vc4/vc4_gem.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 1036 drivers/gpu/drm/vc4/vc4_gem.c spin_lock_irqsave(&vc4->job_lock, irqflags); vc4 1037 drivers/gpu/drm/vc4/vc4_gem.c if (seqno > vc4->finished_seqno) { vc4 1039 drivers/gpu/drm/vc4/vc4_gem.c list_add_tail(&cb->work.entry, &vc4->seqno_cb_list); vc4 1043 drivers/gpu/drm/vc4/vc4_gem.c spin_unlock_irqrestore(&vc4->job_lock, irqflags); vc4 1055 drivers/gpu/drm/vc4/vc4_gem.c struct vc4_dev *vc4 = vc4 1058 drivers/gpu/drm/vc4/vc4_gem.c vc4_job_handle_completed(vc4); vc4 1131 drivers/gpu/drm/vc4/vc4_gem.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 1140 drivers/gpu/drm/vc4/vc4_gem.c if (!vc4->v3d) { vc4 1164 drivers/gpu/drm/vc4/vc4_gem.c ret = vc4_v3d_pm_get(vc4); vc4 1198 drivers/gpu/drm/vc4/vc4_gem.c vc4->dma_fence_context)) { vc4 1257 drivers/gpu/drm/vc4/vc4_gem.c args->seqno = vc4->emit_seqno; vc4 1262 drivers/gpu/drm/vc4/vc4_gem.c vc4_complete_exec(vc4->dev, exec); vc4 1270 drivers/gpu/drm/vc4/vc4_gem.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 1272 drivers/gpu/drm/vc4/vc4_gem.c vc4->dma_fence_context = dma_fence_context_alloc(1); vc4 1274 drivers/gpu/drm/vc4/vc4_gem.c INIT_LIST_HEAD(&vc4->bin_job_list); vc4 1275 drivers/gpu/drm/vc4/vc4_gem.c INIT_LIST_HEAD(&vc4->render_job_list); vc4 1276 drivers/gpu/drm/vc4/vc4_gem.c INIT_LIST_HEAD(&vc4->job_done_list); vc4 1277 drivers/gpu/drm/vc4/vc4_gem.c INIT_LIST_HEAD(&vc4->seqno_cb_list); vc4 1278 drivers/gpu/drm/vc4/vc4_gem.c spin_lock_init(&vc4->job_lock); vc4 1280 drivers/gpu/drm/vc4/vc4_gem.c INIT_WORK(&vc4->hangcheck.reset_work, vc4_reset_work); vc4 1281 drivers/gpu/drm/vc4/vc4_gem.c timer_setup(&vc4->hangcheck.timer, vc4_hangcheck_elapsed, 0); vc4 1283 drivers/gpu/drm/vc4/vc4_gem.c INIT_WORK(&vc4->job_done_work, vc4_job_done_work); vc4 1285 drivers/gpu/drm/vc4/vc4_gem.c mutex_init(&vc4->power_lock); vc4 1287 drivers/gpu/drm/vc4/vc4_gem.c INIT_LIST_HEAD(&vc4->purgeable.list); vc4 1288 drivers/gpu/drm/vc4/vc4_gem.c mutex_init(&vc4->purgeable.lock); vc4 1294 drivers/gpu/drm/vc4/vc4_gem.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 1299 drivers/gpu/drm/vc4/vc4_gem.c WARN_ON(vc4->emit_seqno != vc4->finished_seqno); vc4 1304 drivers/gpu/drm/vc4/vc4_gem.c if (vc4->bin_bo) { vc4 1305 drivers/gpu/drm/vc4/vc4_gem.c drm_gem_object_put_unlocked(&vc4->bin_bo->base.base); vc4 1306 drivers/gpu/drm/vc4/vc4_gem.c vc4->bin_bo = NULL; vc4 1309 drivers/gpu/drm/vc4/vc4_gem.c if (vc4->hang_state) vc4 1310 drivers/gpu/drm/vc4/vc4_gem.c vc4_free_hang_state(dev, vc4->hang_state); vc4 97 drivers/gpu/drm/vc4/vc4_hdmi.c #define HDMI_READ(offset) readl(vc4->hdmi->hdmicore_regs + offset) vc4 98 drivers/gpu/drm/vc4/vc4_hdmi.c #define HDMI_WRITE(offset, val) writel(val, vc4->hdmi->hdmicore_regs + offset) vc4 99 drivers/gpu/drm/vc4/vc4_hdmi.c #define HD_READ(offset) readl(vc4->hdmi->hd_regs + offset) vc4 100 drivers/gpu/drm/vc4/vc4_hdmi.c #define HD_WRITE(offset, val) writel(val, vc4->hdmi->hd_regs + offset) vc4 186 drivers/gpu/drm/vc4/vc4_hdmi.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 187 drivers/gpu/drm/vc4/vc4_hdmi.c struct vc4_hdmi *hdmi = vc4->hdmi; vc4 200 drivers/gpu/drm/vc4/vc4_hdmi.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 202 drivers/gpu/drm/vc4/vc4_hdmi.c if (vc4->hdmi->hpd_gpio) { vc4 203 drivers/gpu/drm/vc4/vc4_hdmi.c if (gpio_get_value_cansleep(vc4->hdmi->hpd_gpio) ^ vc4 204 drivers/gpu/drm/vc4/vc4_hdmi.c vc4->hdmi->hpd_active_low) vc4 206 drivers/gpu/drm/vc4/vc4_hdmi.c cec_phys_addr_invalidate(vc4->hdmi->cec_adap); vc4 210 drivers/gpu/drm/vc4/vc4_hdmi.c if (drm_probe_ddc(vc4->hdmi->ddc)) vc4 215 drivers/gpu/drm/vc4/vc4_hdmi.c cec_phys_addr_invalidate(vc4->hdmi->cec_adap); vc4 232 drivers/gpu/drm/vc4/vc4_hdmi.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 236 drivers/gpu/drm/vc4/vc4_hdmi.c edid = drm_get_edid(connector, vc4->hdmi->ddc); vc4 237 drivers/gpu/drm/vc4/vc4_hdmi.c cec_s_phys_addr_from_edid(vc4->hdmi->cec_adap, edid); vc4 319 drivers/gpu/drm/vc4/vc4_hdmi.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 333 drivers/gpu/drm/vc4/vc4_hdmi.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 380 drivers/gpu/drm/vc4/vc4_hdmi.c struct vc4_dev *vc4 = encoder->dev->dev_private; vc4 381 drivers/gpu/drm/vc4/vc4_hdmi.c struct vc4_hdmi *hdmi = vc4->hdmi; vc4 428 drivers/gpu/drm/vc4/vc4_hdmi.c struct vc4_dev *vc4 = drm->dev_private; vc4 429 drivers/gpu/drm/vc4/vc4_hdmi.c struct vc4_hdmi *hdmi = vc4->hdmi; vc4 452 drivers/gpu/drm/vc4/vc4_hdmi.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 453 drivers/gpu/drm/vc4/vc4_hdmi.c struct vc4_hdmi *hdmi = vc4->hdmi; vc4 474 drivers/gpu/drm/vc4/vc4_hdmi.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 475 drivers/gpu/drm/vc4/vc4_hdmi.c struct vc4_hdmi *hdmi = vc4->hdmi; vc4 716 drivers/gpu/drm/vc4/vc4_hdmi.c struct vc4_dev *vc4 = to_vc4_dev(drm); vc4 737 drivers/gpu/drm/vc4/vc4_hdmi.c struct vc4_dev *vc4 = to_vc4_dev(drm); vc4 773 drivers/gpu/drm/vc4/vc4_hdmi.c struct vc4_dev *vc4 = to_vc4_dev(encoder->dev); vc4 807 drivers/gpu/drm/vc4/vc4_hdmi.c struct vc4_dev *vc4 = to_vc4_dev(drm); vc4 841 drivers/gpu/drm/vc4/vc4_hdmi.c struct vc4_dev *vc4 = to_vc4_dev(drm); vc4 913 drivers/gpu/drm/vc4/vc4_hdmi.c struct vc4_dev *vc4 = to_vc4_dev(drm); vc4 1148 drivers/gpu/drm/vc4/vc4_hdmi.c struct vc4_dev *vc4 = priv; vc4 1149 drivers/gpu/drm/vc4/vc4_hdmi.c struct vc4_hdmi *hdmi = vc4->hdmi; vc4 1168 drivers/gpu/drm/vc4/vc4_hdmi.c static void vc4_cec_read_msg(struct vc4_dev *vc4, u32 cntrl1) vc4 1170 drivers/gpu/drm/vc4/vc4_hdmi.c struct cec_msg *msg = &vc4->hdmi->cec_rx_msg; vc4 1187 drivers/gpu/drm/vc4/vc4_hdmi.c struct vc4_dev *vc4 = priv; vc4 1188 drivers/gpu/drm/vc4/vc4_hdmi.c struct vc4_hdmi *hdmi = vc4->hdmi; vc4 1199 drivers/gpu/drm/vc4/vc4_hdmi.c vc4_cec_read_msg(vc4, cntrl1); vc4 1215 drivers/gpu/drm/vc4/vc4_hdmi.c struct vc4_dev *vc4 = cec_get_drvdata(adap); vc4 1258 drivers/gpu/drm/vc4/vc4_hdmi.c struct vc4_dev *vc4 = cec_get_drvdata(adap); vc4 1269 drivers/gpu/drm/vc4/vc4_hdmi.c struct vc4_dev *vc4 = cec_get_drvdata(adap); vc4 1305 drivers/gpu/drm/vc4/vc4_hdmi.c struct vc4_dev *vc4 = drm->dev_private; vc4 1397 drivers/gpu/drm/vc4/vc4_hdmi.c vc4->hdmi = hdmi; vc4 1420 drivers/gpu/drm/vc4/vc4_hdmi.c vc4, "vc4", vc4 1444 drivers/gpu/drm/vc4/vc4_hdmi.c "vc4 hdmi cec", vc4); vc4 1481 drivers/gpu/drm/vc4/vc4_hdmi.c struct vc4_dev *vc4 = drm->dev_private; vc4 1482 drivers/gpu/drm/vc4/vc4_hdmi.c struct vc4_hdmi *hdmi = vc4->hdmi; vc4 1493 drivers/gpu/drm/vc4/vc4_hdmi.c vc4->hdmi = NULL; vc4 66 drivers/gpu/drm/vc4/vc4_hvs.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 67 drivers/gpu/drm/vc4/vc4_hvs.c struct drm_printer p = drm_info_printer(&vc4->hvs->pdev->dev); vc4 70 drivers/gpu/drm/vc4/vc4_hvs.c drm_print_regset32(&p, &vc4->hvs->regset); vc4 76 drivers/gpu/drm/vc4/vc4_hvs.c readl((u32 __iomem *)vc4->hvs->dlist + i + 0), vc4 77 drivers/gpu/drm/vc4/vc4_hvs.c readl((u32 __iomem *)vc4->hvs->dlist + i + 1), vc4 78 drivers/gpu/drm/vc4/vc4_hvs.c readl((u32 __iomem *)vc4->hvs->dlist + i + 2), vc4 79 drivers/gpu/drm/vc4/vc4_hvs.c readl((u32 __iomem *)vc4->hvs->dlist + i + 3)); vc4 87 drivers/gpu/drm/vc4/vc4_hvs.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 90 drivers/gpu/drm/vc4/vc4_hvs.c drm_printf(&p, "%d\n", atomic_read(&vc4->underrun)); vc4 159 drivers/gpu/drm/vc4/vc4_hvs.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 169 drivers/gpu/drm/vc4/vc4_hvs.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 181 drivers/gpu/drm/vc4/vc4_hvs.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 183 drivers/gpu/drm/vc4/vc4_hvs.c atomic_inc(&vc4->underrun); vc4 190 drivers/gpu/drm/vc4/vc4_hvs.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 222 drivers/gpu/drm/vc4/vc4_hvs.c struct vc4_dev *vc4 = drm->dev_private; vc4 270 drivers/gpu/drm/vc4/vc4_hvs.c vc4->hvs = hvs; vc4 316 drivers/gpu/drm/vc4/vc4_hvs.c struct vc4_dev *vc4 = drm->dev_private; vc4 318 drivers/gpu/drm/vc4/vc4_hvs.c if (vc4->hvs->mitchell_netravali_filter.allocated) vc4 319 drivers/gpu/drm/vc4/vc4_hvs.c drm_mm_remove_node(&vc4->hvs->mitchell_netravali_filter); vc4 321 drivers/gpu/drm/vc4/vc4_hvs.c drm_mm_takedown(&vc4->hvs->dlist_mm); vc4 322 drivers/gpu/drm/vc4/vc4_hvs.c drm_mm_takedown(&vc4->hvs->lbm_mm); vc4 324 drivers/gpu/drm/vc4/vc4_hvs.c vc4->hvs = NULL; vc4 60 drivers/gpu/drm/vc4/vc4_irq.c struct vc4_dev *vc4 = vc4 67 drivers/gpu/drm/vc4/vc4_irq.c mutex_lock(&vc4->bin_bo_lock); vc4 69 drivers/gpu/drm/vc4/vc4_irq.c if (!vc4->bin_bo) vc4 72 drivers/gpu/drm/vc4/vc4_irq.c bo = vc4->bin_bo; vc4 74 drivers/gpu/drm/vc4/vc4_irq.c bin_bo_slot = vc4_v3d_get_bin_slot(vc4); vc4 80 drivers/gpu/drm/vc4/vc4_irq.c spin_lock_irqsave(&vc4->job_lock, irqflags); vc4 82 drivers/gpu/drm/vc4/vc4_irq.c if (vc4->bin_alloc_overflow) { vc4 89 drivers/gpu/drm/vc4/vc4_irq.c exec = vc4_first_bin_job(vc4); vc4 91 drivers/gpu/drm/vc4/vc4_irq.c exec = vc4_last_render_job(vc4); vc4 93 drivers/gpu/drm/vc4/vc4_irq.c exec->bin_slots |= vc4->bin_alloc_overflow; vc4 98 drivers/gpu/drm/vc4/vc4_irq.c vc4->bin_alloc_used &= ~vc4->bin_alloc_overflow; vc4 101 drivers/gpu/drm/vc4/vc4_irq.c vc4->bin_alloc_overflow = BIT(bin_bo_slot); vc4 103 drivers/gpu/drm/vc4/vc4_irq.c V3D_WRITE(V3D_BPOA, bo->base.paddr + bin_bo_slot * vc4->bin_alloc_size); vc4 107 drivers/gpu/drm/vc4/vc4_irq.c spin_unlock_irqrestore(&vc4->job_lock, irqflags); vc4 110 drivers/gpu/drm/vc4/vc4_irq.c mutex_unlock(&vc4->bin_bo_lock); vc4 116 drivers/gpu/drm/vc4/vc4_irq.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 117 drivers/gpu/drm/vc4/vc4_irq.c struct vc4_exec_info *next, *exec = vc4_first_bin_job(vc4); vc4 123 drivers/gpu/drm/vc4/vc4_irq.c next = vc4_first_bin_job(vc4); vc4 136 drivers/gpu/drm/vc4/vc4_irq.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 137 drivers/gpu/drm/vc4/vc4_irq.c struct vc4_exec_info *exec = vc4_first_bin_job(vc4); vc4 144 drivers/gpu/drm/vc4/vc4_irq.c vc4_perfmon_stop(vc4, exec->perfmon, false); vc4 146 drivers/gpu/drm/vc4/vc4_irq.c list_move_tail(&exec->head, &vc4->bin_job_list); vc4 153 drivers/gpu/drm/vc4/vc4_irq.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 154 drivers/gpu/drm/vc4/vc4_irq.c struct vc4_exec_info *exec = vc4_first_render_job(vc4); vc4 160 drivers/gpu/drm/vc4/vc4_irq.c vc4->finished_seqno++; vc4 161 drivers/gpu/drm/vc4/vc4_irq.c list_move_tail(&exec->head, &vc4->job_done_list); vc4 163 drivers/gpu/drm/vc4/vc4_irq.c nextbin = vc4_first_bin_job(vc4); vc4 164 drivers/gpu/drm/vc4/vc4_irq.c nextrender = vc4_first_render_job(vc4); vc4 171 drivers/gpu/drm/vc4/vc4_irq.c vc4_perfmon_stop(vc4, exec->perfmon, true); vc4 191 drivers/gpu/drm/vc4/vc4_irq.c wake_up_all(&vc4->job_wait_queue); vc4 192 drivers/gpu/drm/vc4/vc4_irq.c schedule_work(&vc4->job_done_work); vc4 199 drivers/gpu/drm/vc4/vc4_irq.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 216 drivers/gpu/drm/vc4/vc4_irq.c schedule_work(&vc4->overflow_mem_work); vc4 221 drivers/gpu/drm/vc4/vc4_irq.c spin_lock(&vc4->job_lock); vc4 223 drivers/gpu/drm/vc4/vc4_irq.c spin_unlock(&vc4->job_lock); vc4 228 drivers/gpu/drm/vc4/vc4_irq.c spin_lock(&vc4->job_lock); vc4 230 drivers/gpu/drm/vc4/vc4_irq.c spin_unlock(&vc4->job_lock); vc4 240 drivers/gpu/drm/vc4/vc4_irq.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 242 drivers/gpu/drm/vc4/vc4_irq.c if (!vc4->v3d) vc4 245 drivers/gpu/drm/vc4/vc4_irq.c init_waitqueue_head(&vc4->job_wait_queue); vc4 246 drivers/gpu/drm/vc4/vc4_irq.c INIT_WORK(&vc4->overflow_mem_work, vc4_overflow_mem_work); vc4 257 drivers/gpu/drm/vc4/vc4_irq.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 259 drivers/gpu/drm/vc4/vc4_irq.c if (!vc4->v3d) vc4 273 drivers/gpu/drm/vc4/vc4_irq.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 275 drivers/gpu/drm/vc4/vc4_irq.c if (!vc4->v3d) vc4 287 drivers/gpu/drm/vc4/vc4_irq.c cancel_work_sync(&vc4->overflow_mem_work); vc4 293 drivers/gpu/drm/vc4/vc4_irq.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 307 drivers/gpu/drm/vc4/vc4_irq.c spin_lock_irqsave(&vc4->job_lock, irqflags); vc4 310 drivers/gpu/drm/vc4/vc4_irq.c spin_unlock_irqrestore(&vc4->job_lock, irqflags); vc4 52 drivers/gpu/drm/vc4/vc4_kms.c struct vc4_dev *vc4 = dev->dev_private; vc4 56 drivers/gpu/drm/vc4/vc4_kms.c ret = drm_modeset_lock(&vc4->ctm_state_lock, state->acquire_ctx); vc4 114 drivers/gpu/drm/vc4/vc4_kms.c vc4_ctm_commit(struct vc4_dev *vc4, struct drm_atomic_state *state) vc4 116 drivers/gpu/drm/vc4/vc4_kms.c struct vc4_ctm_state *ctm_state = to_vc4_ctm_state(vc4->ctm_manager.state); vc4 151 drivers/gpu/drm/vc4/vc4_kms.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 169 drivers/gpu/drm/vc4/vc4_kms.c vc4_ctm_commit(vc4, state); vc4 187 drivers/gpu/drm/vc4/vc4_kms.c up(&vc4->async_modeset); vc4 215 drivers/gpu/drm/vc4/vc4_kms.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 219 drivers/gpu/drm/vc4/vc4_kms.c ret = down_interruptible(&vc4->async_modeset); vc4 225 drivers/gpu/drm/vc4/vc4_kms.c up(&vc4->async_modeset); vc4 233 drivers/gpu/drm/vc4/vc4_kms.c up(&vc4->async_modeset); vc4 250 drivers/gpu/drm/vc4/vc4_kms.c ret = down_interruptible(&vc4->async_modeset); vc4 256 drivers/gpu/drm/vc4/vc4_kms.c up(&vc4->async_modeset); vc4 264 drivers/gpu/drm/vc4/vc4_kms.c up(&vc4->async_modeset); vc4 348 drivers/gpu/drm/vc4/vc4_kms.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 358 drivers/gpu/drm/vc4/vc4_kms.c ctm_state = vc4_get_ctm_state(state, &vc4->ctm_manager); vc4 370 drivers/gpu/drm/vc4/vc4_kms.c ctm_state = vc4_get_ctm_state(state, &vc4->ctm_manager); vc4 412 drivers/gpu/drm/vc4/vc4_kms.c struct vc4_dev *vc4 = to_vc4_dev(state->dev); vc4 419 drivers/gpu/drm/vc4/vc4_kms.c &vc4->load_tracker); vc4 442 drivers/gpu/drm/vc4/vc4_kms.c if (!vc4->load_tracker_enabled) vc4 512 drivers/gpu/drm/vc4/vc4_kms.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 520 drivers/gpu/drm/vc4/vc4_kms.c vc4->load_tracker_enabled = true; vc4 522 drivers/gpu/drm/vc4/vc4_kms.c sema_init(&vc4->async_modeset, 1); vc4 541 drivers/gpu/drm/vc4/vc4_kms.c drm_modeset_lock_init(&vc4->ctm_state_lock); vc4 547 drivers/gpu/drm/vc4/vc4_kms.c drm_atomic_private_obj_init(dev, &vc4->ctm_manager, &ctm_state->base, vc4 552 drivers/gpu/drm/vc4/vc4_kms.c drm_atomic_private_obj_fini(&vc4->ctm_manager); vc4 556 drivers/gpu/drm/vc4/vc4_kms.c drm_atomic_private_obj_init(dev, &vc4->load_tracker, &load_state->base, vc4 30 drivers/gpu/drm/vc4/vc4_perfmon.c void vc4_perfmon_start(struct vc4_dev *vc4, struct vc4_perfmon *perfmon) vc4 35 drivers/gpu/drm/vc4/vc4_perfmon.c if (WARN_ON_ONCE(!perfmon || vc4->active_perfmon)) vc4 44 drivers/gpu/drm/vc4/vc4_perfmon.c vc4->active_perfmon = perfmon; vc4 47 drivers/gpu/drm/vc4/vc4_perfmon.c void vc4_perfmon_stop(struct vc4_dev *vc4, struct vc4_perfmon *perfmon, vc4 52 drivers/gpu/drm/vc4/vc4_perfmon.c if (WARN_ON_ONCE(!vc4->active_perfmon || vc4 53 drivers/gpu/drm/vc4/vc4_perfmon.c perfmon != vc4->active_perfmon)) vc4 62 drivers/gpu/drm/vc4/vc4_perfmon.c vc4->active_perfmon = NULL; vc4 103 drivers/gpu/drm/vc4/vc4_perfmon.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 110 drivers/gpu/drm/vc4/vc4_perfmon.c if (!vc4->v3d) { vc4 155 drivers/gpu/drm/vc4/vc4_perfmon.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 160 drivers/gpu/drm/vc4/vc4_perfmon.c if (!vc4->v3d) { vc4 179 drivers/gpu/drm/vc4/vc4_perfmon.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 185 drivers/gpu/drm/vc4/vc4_perfmon.c if (!vc4->v3d) { vc4 178 drivers/gpu/drm/vc4/vc4_plane.c struct vc4_dev *vc4 = to_vc4_dev(plane->dev); vc4 184 drivers/gpu/drm/vc4/vc4_plane.c spin_lock_irqsave(&vc4->hvs->mm_lock, irqflags); vc4 186 drivers/gpu/drm/vc4/vc4_plane.c spin_unlock_irqrestore(&vc4->hvs->mm_lock, irqflags); vc4 545 drivers/gpu/drm/vc4/vc4_plane.c struct vc4_dev *vc4 = to_vc4_dev(state->plane->dev); vc4 563 drivers/gpu/drm/vc4/vc4_plane.c spin_lock_irqsave(&vc4->hvs->mm_lock, irqflags); vc4 564 drivers/gpu/drm/vc4/vc4_plane.c ret = drm_mm_insert_node_generic(&vc4->hvs->lbm_mm, vc4 567 drivers/gpu/drm/vc4/vc4_plane.c spin_unlock_irqrestore(&vc4->hvs->mm_lock, irqflags); vc4 586 drivers/gpu/drm/vc4/vc4_plane.c struct vc4_dev *vc4 = to_vc4_dev(plane->dev); vc4 893 drivers/gpu/drm/vc4/vc4_plane.c u32 kernel = VC4_SET_FIELD(vc4->hvs->mitchell_netravali_filter.start, vc4 14 drivers/gpu/drm/vc4/vc4_trace.h #define TRACE_SYSTEM vc4 vc4 59 drivers/gpu/drm/vc4/vc4_trace.h #define TRACE_INCLUDE_PATH ../../drivers/gpu/drm/vc4 vc4 373 drivers/gpu/drm/vc4/vc4_txp.c struct vc4_dev *vc4 = to_vc4_dev(drm); vc4 409 drivers/gpu/drm/vc4/vc4_txp.c vc4->txp = txp; vc4 420 drivers/gpu/drm/vc4/vc4_txp.c struct vc4_dev *vc4 = to_vc4_dev(drm); vc4 425 drivers/gpu/drm/vc4/vc4_txp.c vc4->txp = NULL; vc4 103 drivers/gpu/drm/vc4/vc4_v3d.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 104 drivers/gpu/drm/vc4/vc4_v3d.c int ret = vc4_v3d_pm_get(vc4); vc4 119 drivers/gpu/drm/vc4/vc4_v3d.c vc4_v3d_pm_put(vc4); vc4 130 drivers/gpu/drm/vc4/vc4_v3d.c vc4_v3d_pm_get(struct vc4_dev *vc4) vc4 132 drivers/gpu/drm/vc4/vc4_v3d.c mutex_lock(&vc4->power_lock); vc4 133 drivers/gpu/drm/vc4/vc4_v3d.c if (vc4->power_refcount++ == 0) { vc4 134 drivers/gpu/drm/vc4/vc4_v3d.c int ret = pm_runtime_get_sync(&vc4->v3d->pdev->dev); vc4 137 drivers/gpu/drm/vc4/vc4_v3d.c vc4->power_refcount--; vc4 138 drivers/gpu/drm/vc4/vc4_v3d.c mutex_unlock(&vc4->power_lock); vc4 142 drivers/gpu/drm/vc4/vc4_v3d.c mutex_unlock(&vc4->power_lock); vc4 148 drivers/gpu/drm/vc4/vc4_v3d.c vc4_v3d_pm_put(struct vc4_dev *vc4) vc4 150 drivers/gpu/drm/vc4/vc4_v3d.c mutex_lock(&vc4->power_lock); vc4 151 drivers/gpu/drm/vc4/vc4_v3d.c if (--vc4->power_refcount == 0) { vc4 152 drivers/gpu/drm/vc4/vc4_v3d.c pm_runtime_mark_last_busy(&vc4->v3d->pdev->dev); vc4 153 drivers/gpu/drm/vc4/vc4_v3d.c pm_runtime_put_autosuspend(&vc4->v3d->pdev->dev); vc4 155 drivers/gpu/drm/vc4/vc4_v3d.c mutex_unlock(&vc4->power_lock); vc4 160 drivers/gpu/drm/vc4/vc4_v3d.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 169 drivers/gpu/drm/vc4/vc4_v3d.c int vc4_v3d_get_bin_slot(struct vc4_dev *vc4) vc4 171 drivers/gpu/drm/vc4/vc4_v3d.c struct drm_device *dev = vc4->dev; vc4 178 drivers/gpu/drm/vc4/vc4_v3d.c spin_lock_irqsave(&vc4->job_lock, irqflags); vc4 179 drivers/gpu/drm/vc4/vc4_v3d.c slot = ffs(~vc4->bin_alloc_used); vc4 183 drivers/gpu/drm/vc4/vc4_v3d.c vc4->bin_alloc_used |= BIT(slot); vc4 184 drivers/gpu/drm/vc4/vc4_v3d.c spin_unlock_irqrestore(&vc4->job_lock, irqflags); vc4 191 drivers/gpu/drm/vc4/vc4_v3d.c exec = vc4_last_render_job(vc4); vc4 194 drivers/gpu/drm/vc4/vc4_v3d.c spin_unlock_irqrestore(&vc4->job_lock, irqflags); vc4 230 drivers/gpu/drm/vc4/vc4_v3d.c static int bin_bo_alloc(struct vc4_dev *vc4) vc4 232 drivers/gpu/drm/vc4/vc4_v3d.c struct vc4_v3d *v3d = vc4->v3d; vc4 249 drivers/gpu/drm/vc4/vc4_v3d.c struct vc4_bo *bo = vc4_bo_create(vc4->dev, size, true, vc4 266 drivers/gpu/drm/vc4/vc4_v3d.c vc4->bin_bo = bo; vc4 284 drivers/gpu/drm/vc4/vc4_v3d.c vc4->bin_alloc_size = 512 * 1024; vc4 285 drivers/gpu/drm/vc4/vc4_v3d.c vc4->bin_alloc_used = 0; vc4 286 drivers/gpu/drm/vc4/vc4_v3d.c vc4->bin_alloc_overflow = 0; vc4 287 drivers/gpu/drm/vc4/vc4_v3d.c WARN_ON_ONCE(sizeof(vc4->bin_alloc_used) * 8 != vc4 288 drivers/gpu/drm/vc4/vc4_v3d.c bo->base.base.size / vc4->bin_alloc_size); vc4 290 drivers/gpu/drm/vc4/vc4_v3d.c kref_init(&vc4->bin_bo_kref); vc4 317 drivers/gpu/drm/vc4/vc4_v3d.c int vc4_v3d_bin_bo_get(struct vc4_dev *vc4, bool *used) vc4 321 drivers/gpu/drm/vc4/vc4_v3d.c mutex_lock(&vc4->bin_bo_lock); vc4 326 drivers/gpu/drm/vc4/vc4_v3d.c if (vc4->bin_bo) vc4 327 drivers/gpu/drm/vc4/vc4_v3d.c kref_get(&vc4->bin_bo_kref); vc4 329 drivers/gpu/drm/vc4/vc4_v3d.c ret = bin_bo_alloc(vc4); vc4 335 drivers/gpu/drm/vc4/vc4_v3d.c mutex_unlock(&vc4->bin_bo_lock); vc4 342 drivers/gpu/drm/vc4/vc4_v3d.c struct vc4_dev *vc4 = container_of(ref, struct vc4_dev, bin_bo_kref); vc4 344 drivers/gpu/drm/vc4/vc4_v3d.c if (WARN_ON_ONCE(!vc4->bin_bo)) vc4 347 drivers/gpu/drm/vc4/vc4_v3d.c drm_gem_object_put_unlocked(&vc4->bin_bo->base.base); vc4 348 drivers/gpu/drm/vc4/vc4_v3d.c vc4->bin_bo = NULL; vc4 351 drivers/gpu/drm/vc4/vc4_v3d.c void vc4_v3d_bin_bo_put(struct vc4_dev *vc4) vc4 353 drivers/gpu/drm/vc4/vc4_v3d.c mutex_lock(&vc4->bin_bo_lock); vc4 354 drivers/gpu/drm/vc4/vc4_v3d.c kref_put(&vc4->bin_bo_kref, bin_bo_release); vc4 355 drivers/gpu/drm/vc4/vc4_v3d.c mutex_unlock(&vc4->bin_bo_lock); vc4 362 drivers/gpu/drm/vc4/vc4_v3d.c struct vc4_dev *vc4 = v3d->vc4; vc4 364 drivers/gpu/drm/vc4/vc4_v3d.c vc4_irq_uninstall(vc4->dev); vc4 374 drivers/gpu/drm/vc4/vc4_v3d.c struct vc4_dev *vc4 = v3d->vc4; vc4 381 drivers/gpu/drm/vc4/vc4_v3d.c vc4_v3d_init_hw(vc4->dev); vc4 384 drivers/gpu/drm/vc4/vc4_v3d.c enable_irq(vc4->dev->irq); vc4 385 drivers/gpu/drm/vc4/vc4_v3d.c vc4_irq_postinstall(vc4->dev); vc4 395 drivers/gpu/drm/vc4/vc4_v3d.c struct vc4_dev *vc4 = to_vc4_dev(drm); vc4 414 drivers/gpu/drm/vc4/vc4_v3d.c vc4->v3d = v3d; vc4 415 drivers/gpu/drm/vc4/vc4_v3d.c v3d->vc4 = vc4; vc4 472 drivers/gpu/drm/vc4/vc4_v3d.c struct vc4_dev *vc4 = to_vc4_dev(drm); vc4 485 drivers/gpu/drm/vc4/vc4_v3d.c vc4->v3d = NULL; vc4 352 drivers/gpu/drm/vc4/vc4_validate.c struct vc4_dev *vc4 = to_vc4_dev(dev); vc4 382 drivers/gpu/drm/vc4/vc4_validate.c bin_slot = vc4_v3d_get_bin_slot(vc4); vc4 395 drivers/gpu/drm/vc4/vc4_validate.c bin_addr = vc4->bin_bo->base.paddr + bin_slot * vc4->bin_alloc_size; vc4 417 drivers/gpu/drm/vc4/vc4_validate.c *(uint32_t *)(validated + 4) = (bin_addr + vc4->bin_alloc_size - vc4 529 drivers/gpu/drm/vc4/vc4_vec.c struct vc4_dev *vc4 = to_vc4_dev(drm); vc4 581 drivers/gpu/drm/vc4/vc4_vec.c vc4->vec = vec; vc4 598 drivers/gpu/drm/vc4/vc4_vec.c struct vc4_dev *vc4 = to_vc4_dev(drm); vc4 605 drivers/gpu/drm/vc4/vc4_vec.c vc4->vec = NULL; vc4 357 drivers/net/dsa/b53/b53_common.c u8 mgmt, vc0, vc1, vc4 = 0, vc5; vc4 364 drivers/net/dsa/b53/b53_common.c b53_read8(dev, B53_VLAN_PAGE, B53_VLAN_CTRL4_25, &vc4); vc4 367 drivers/net/dsa/b53/b53_common.c b53_read8(dev, B53_VLAN_PAGE, B53_VLAN_CTRL4_63XX, &vc4); vc4 370 drivers/net/dsa/b53/b53_common.c b53_read8(dev, B53_VLAN_PAGE, B53_VLAN_CTRL4, &vc4); vc4 379 drivers/net/dsa/b53/b53_common.c vc4 &= ~VC4_ING_VID_CHECK_MASK; vc4 381 drivers/net/dsa/b53/b53_common.c vc4 |= VC4_ING_VID_VIO_DROP << VC4_ING_VID_CHECK_S; vc4 384 drivers/net/dsa/b53/b53_common.c vc4 |= VC4_ING_VID_VIO_FWD << VC4_ING_VID_CHECK_S; vc4 397 drivers/net/dsa/b53/b53_common.c vc4 &= ~VC4_ING_VID_CHECK_MASK; vc4 401 drivers/net/dsa/b53/b53_common.c vc4 |= VC4_ING_VID_VIO_FWD << VC4_ING_VID_CHECK_S; vc4 403 drivers/net/dsa/b53/b53_common.c vc4 |= VC4_ING_VID_VIO_TO_IMP << VC4_ING_VID_CHECK_S; vc4 423 drivers/net/dsa/b53/b53_common.c b53_write8(dev, B53_VLAN_PAGE, B53_VLAN_CTRL4_25, vc4); vc4 427 drivers/net/dsa/b53/b53_common.c b53_write8(dev, B53_VLAN_PAGE, B53_VLAN_CTRL4_63XX, vc4); vc4 431 drivers/net/dsa/b53/b53_common.c b53_write8(dev, B53_VLAN_PAGE, B53_VLAN_CTRL4, vc4); vc4 2312 drivers/net/dsa/b53/b53_common.c u8 vc4; vc4 2314 drivers/net/dsa/b53/b53_common.c b53_read8(dev, B53_VLAN_PAGE, B53_VLAN_CTRL4_25, &vc4); vc4 2317 drivers/net/dsa/b53/b53_common.c switch (vc4 & 3) {