topology_state 3245 drivers/gpu/drm/drm_dp_mst_topology.c struct drm_dp_mst_topology_state *topology_state; topology_state 3249 drivers/gpu/drm/drm_dp_mst_topology.c topology_state = drm_atomic_get_mst_topology_state(state, mgr); topology_state 3250 drivers/gpu/drm/drm_dp_mst_topology.c if (IS_ERR(topology_state)) topology_state 3251 drivers/gpu/drm/drm_dp_mst_topology.c return PTR_ERR(topology_state); topology_state 3254 drivers/gpu/drm/drm_dp_mst_topology.c list_for_each_entry(pos, &topology_state->vcpis, next) { topology_state 3290 drivers/gpu/drm/drm_dp_mst_topology.c list_add(&vcpi->next, &topology_state->vcpis); topology_state 3329 drivers/gpu/drm/drm_dp_mst_topology.c struct drm_dp_mst_topology_state *topology_state; topology_state 3333 drivers/gpu/drm/drm_dp_mst_topology.c topology_state = drm_atomic_get_mst_topology_state(state, mgr); topology_state 3334 drivers/gpu/drm/drm_dp_mst_topology.c if (IS_ERR(topology_state)) topology_state 3335 drivers/gpu/drm/drm_dp_mst_topology.c return PTR_ERR(topology_state); topology_state 3337 drivers/gpu/drm/drm_dp_mst_topology.c list_for_each_entry(pos, &topology_state->vcpis, next) { topology_state 3345 drivers/gpu/drm/drm_dp_mst_topology.c port, &topology_state->base);