topology_state   3245 drivers/gpu/drm/drm_dp_mst_topology.c 	struct drm_dp_mst_topology_state *topology_state;
topology_state   3249 drivers/gpu/drm/drm_dp_mst_topology.c 	topology_state = drm_atomic_get_mst_topology_state(state, mgr);
topology_state   3250 drivers/gpu/drm/drm_dp_mst_topology.c 	if (IS_ERR(topology_state))
topology_state   3251 drivers/gpu/drm/drm_dp_mst_topology.c 		return PTR_ERR(topology_state);
topology_state   3254 drivers/gpu/drm/drm_dp_mst_topology.c 	list_for_each_entry(pos, &topology_state->vcpis, next) {
topology_state   3290 drivers/gpu/drm/drm_dp_mst_topology.c 		list_add(&vcpi->next, &topology_state->vcpis);
topology_state   3329 drivers/gpu/drm/drm_dp_mst_topology.c 	struct drm_dp_mst_topology_state *topology_state;
topology_state   3333 drivers/gpu/drm/drm_dp_mst_topology.c 	topology_state = drm_atomic_get_mst_topology_state(state, mgr);
topology_state   3334 drivers/gpu/drm/drm_dp_mst_topology.c 	if (IS_ERR(topology_state))
topology_state   3335 drivers/gpu/drm/drm_dp_mst_topology.c 		return PTR_ERR(topology_state);
topology_state   3337 drivers/gpu/drm/drm_dp_mst_topology.c 	list_for_each_entry(pos, &topology_state->vcpis, next) {
topology_state   3345 drivers/gpu/drm/drm_dp_mst_topology.c 			  port, &topology_state->base);