mstb 66 drivers/gpu/drm/drm_dp_mst_topology.c struct drm_dp_mst_branch *mstb); mstb 68 drivers/gpu/drm/drm_dp_mst_topology.c struct drm_dp_mst_branch *mstb, mstb 869 drivers/gpu/drm/drm_dp_mst_topology.c static int drm_dp_mst_wait_tx_reply(struct drm_dp_mst_branch *mstb, mstb 872 drivers/gpu/drm/drm_dp_mst_topology.c struct drm_dp_mst_topology_mgr *mgr = mstb->mgr; mstb 878 drivers/gpu/drm/drm_dp_mst_topology.c mutex_lock(&mstb->mgr->qlock); mstb 898 drivers/gpu/drm/drm_dp_mst_topology.c mstb->tx_slots[txmsg->seqno] = NULL; mstb 909 drivers/gpu/drm/drm_dp_mst_topology.c struct drm_dp_mst_branch *mstb; mstb 911 drivers/gpu/drm/drm_dp_mst_topology.c mstb = kzalloc(sizeof(*mstb), GFP_KERNEL); mstb 912 drivers/gpu/drm/drm_dp_mst_topology.c if (!mstb) mstb 915 drivers/gpu/drm/drm_dp_mst_topology.c mstb->lct = lct; mstb 917 drivers/gpu/drm/drm_dp_mst_topology.c memcpy(mstb->rad, rad, lct / 2); mstb 918 drivers/gpu/drm/drm_dp_mst_topology.c INIT_LIST_HEAD(&mstb->ports); mstb 919 drivers/gpu/drm/drm_dp_mst_topology.c kref_init(&mstb->topology_kref); mstb 920 drivers/gpu/drm/drm_dp_mst_topology.c kref_init(&mstb->malloc_kref); mstb 921 drivers/gpu/drm/drm_dp_mst_topology.c return mstb; mstb 926 drivers/gpu/drm/drm_dp_mst_topology.c struct drm_dp_mst_branch *mstb = mstb 929 drivers/gpu/drm/drm_dp_mst_topology.c if (mstb->port_parent) mstb 930 drivers/gpu/drm/drm_dp_mst_topology.c drm_dp_mst_put_port_malloc(mstb->port_parent); mstb 932 drivers/gpu/drm/drm_dp_mst_topology.c kfree(mstb); mstb 1035 drivers/gpu/drm/drm_dp_mst_topology.c drm_dp_mst_get_mstb_malloc(struct drm_dp_mst_branch *mstb) mstb 1037 drivers/gpu/drm/drm_dp_mst_topology.c kref_get(&mstb->malloc_kref); mstb 1038 drivers/gpu/drm/drm_dp_mst_topology.c DRM_DEBUG("mstb %p (%d)\n", mstb, kref_read(&mstb->malloc_kref)); mstb 1053 drivers/gpu/drm/drm_dp_mst_topology.c drm_dp_mst_put_mstb_malloc(struct drm_dp_mst_branch *mstb) mstb 1055 drivers/gpu/drm/drm_dp_mst_topology.c DRM_DEBUG("mstb %p (%d)\n", mstb, kref_read(&mstb->malloc_kref) - 1); mstb 1056 drivers/gpu/drm/drm_dp_mst_topology.c kref_put(&mstb->malloc_kref, drm_dp_free_mst_branch_device); mstb 1113 drivers/gpu/drm/drm_dp_mst_topology.c struct drm_dp_mst_branch *mstb = mstb 1115 drivers/gpu/drm/drm_dp_mst_topology.c struct drm_dp_mst_topology_mgr *mgr = mstb->mgr; mstb 1120 drivers/gpu/drm/drm_dp_mst_topology.c list_for_each_entry_safe(port, tmp, &mstb->ports, next) { mstb 1127 drivers/gpu/drm/drm_dp_mst_topology.c mutex_lock(&mstb->mgr->qlock); mstb 1128 drivers/gpu/drm/drm_dp_mst_topology.c if (mstb->tx_slots[0]) { mstb 1129 drivers/gpu/drm/drm_dp_mst_topology.c mstb->tx_slots[0]->state = DRM_DP_SIDEBAND_TX_TIMEOUT; mstb 1130 drivers/gpu/drm/drm_dp_mst_topology.c mstb->tx_slots[0] = NULL; mstb 1133 drivers/gpu/drm/drm_dp_mst_topology.c if (mstb->tx_slots[1]) { mstb 1134 drivers/gpu/drm/drm_dp_mst_topology.c mstb->tx_slots[1]->state = DRM_DP_SIDEBAND_TX_TIMEOUT; mstb 1135 drivers/gpu/drm/drm_dp_mst_topology.c mstb->tx_slots[1] = NULL; mstb 1138 drivers/gpu/drm/drm_dp_mst_topology.c mutex_unlock(&mstb->mgr->qlock); mstb 1141 drivers/gpu/drm/drm_dp_mst_topology.c wake_up_all(&mstb->mgr->tx_waitq); mstb 1143 drivers/gpu/drm/drm_dp_mst_topology.c drm_dp_mst_put_mstb_malloc(mstb); mstb 1169 drivers/gpu/drm/drm_dp_mst_topology.c drm_dp_mst_topology_try_get_mstb(struct drm_dp_mst_branch *mstb) mstb 1171 drivers/gpu/drm/drm_dp_mst_topology.c int ret = kref_get_unless_zero(&mstb->topology_kref); mstb 1174 drivers/gpu/drm/drm_dp_mst_topology.c DRM_DEBUG("mstb %p (%d)\n", mstb, mstb 1175 drivers/gpu/drm/drm_dp_mst_topology.c kref_read(&mstb->topology_kref)); mstb 1194 drivers/gpu/drm/drm_dp_mst_topology.c static void drm_dp_mst_topology_get_mstb(struct drm_dp_mst_branch *mstb) mstb 1196 drivers/gpu/drm/drm_dp_mst_topology.c WARN_ON(kref_read(&mstb->topology_kref) == 0); mstb 1197 drivers/gpu/drm/drm_dp_mst_topology.c kref_get(&mstb->topology_kref); mstb 1198 drivers/gpu/drm/drm_dp_mst_topology.c DRM_DEBUG("mstb %p (%d)\n", mstb, kref_read(&mstb->topology_kref)); mstb 1214 drivers/gpu/drm/drm_dp_mst_topology.c drm_dp_mst_topology_put_mstb(struct drm_dp_mst_branch *mstb) mstb 1217 drivers/gpu/drm/drm_dp_mst_topology.c mstb, kref_read(&mstb->topology_kref) - 1); mstb 1218 drivers/gpu/drm/drm_dp_mst_topology.c kref_put(&mstb->topology_kref, drm_dp_destroy_mst_branch_device); mstb 1223 drivers/gpu/drm/drm_dp_mst_topology.c struct drm_dp_mst_branch *mstb; mstb 1232 drivers/gpu/drm/drm_dp_mst_topology.c mstb = port->mstb; mstb 1233 drivers/gpu/drm/drm_dp_mst_topology.c port->mstb = NULL; mstb 1234 drivers/gpu/drm/drm_dp_mst_topology.c drm_dp_mst_topology_put_mstb(mstb); mstb 1345 drivers/gpu/drm/drm_dp_mst_topology.c drm_dp_mst_topology_get_mstb_validated_locked(struct drm_dp_mst_branch *mstb, mstb 1351 drivers/gpu/drm/drm_dp_mst_topology.c if (to_find == mstb) mstb 1352 drivers/gpu/drm/drm_dp_mst_topology.c return mstb; mstb 1354 drivers/gpu/drm/drm_dp_mst_topology.c list_for_each_entry(port, &mstb->ports, next) { mstb 1355 drivers/gpu/drm/drm_dp_mst_topology.c if (port->mstb) { mstb 1357 drivers/gpu/drm/drm_dp_mst_topology.c port->mstb, to_find); mstb 1367 drivers/gpu/drm/drm_dp_mst_topology.c struct drm_dp_mst_branch *mstb) mstb 1374 drivers/gpu/drm/drm_dp_mst_topology.c mgr->mst_primary, mstb); mstb 1384 drivers/gpu/drm/drm_dp_mst_topology.c drm_dp_mst_topology_get_port_validated_locked(struct drm_dp_mst_branch *mstb, mstb 1389 drivers/gpu/drm/drm_dp_mst_topology.c list_for_each_entry(port, &mstb->ports, next) { mstb 1393 drivers/gpu/drm/drm_dp_mst_topology.c if (port->mstb) { mstb 1395 drivers/gpu/drm/drm_dp_mst_topology.c port->mstb, to_find); mstb 1421 drivers/gpu/drm/drm_dp_mst_topology.c static struct drm_dp_mst_port *drm_dp_get_port(struct drm_dp_mst_branch *mstb, u8 port_num) mstb 1426 drivers/gpu/drm/drm_dp_mst_topology.c list_for_each_entry(port, &mstb->ports, next) { mstb 1474 drivers/gpu/drm/drm_dp_mst_topology.c port->mstb = drm_dp_add_mst_branch_device(lct, rad); mstb 1475 drivers/gpu/drm/drm_dp_mst_topology.c if (port->mstb) { mstb 1476 drivers/gpu/drm/drm_dp_mst_topology.c port->mstb->mgr = port->mgr; mstb 1477 drivers/gpu/drm/drm_dp_mst_topology.c port->mstb->port_parent = port; mstb 1537 drivers/gpu/drm/drm_dp_mst_topology.c static void drm_dp_check_mstb_guid(struct drm_dp_mst_branch *mstb, u8 *guid) mstb 1541 drivers/gpu/drm/drm_dp_mst_topology.c memcpy(mstb->guid, guid, 16); mstb 1543 drivers/gpu/drm/drm_dp_mst_topology.c if (!drm_dp_validate_guid(mstb->mgr, mstb->guid)) { mstb 1544 drivers/gpu/drm/drm_dp_mst_topology.c if (mstb->port_parent) { mstb 1546 drivers/gpu/drm/drm_dp_mst_topology.c mstb->mgr, mstb 1547 drivers/gpu/drm/drm_dp_mst_topology.c mstb->port_parent, mstb 1550 drivers/gpu/drm/drm_dp_mst_topology.c mstb->guid); mstb 1554 drivers/gpu/drm/drm_dp_mst_topology.c mstb->mgr->aux, mstb 1556 drivers/gpu/drm/drm_dp_mst_topology.c mstb->guid, mstb 1562 drivers/gpu/drm/drm_dp_mst_topology.c static void build_mst_prop_path(const struct drm_dp_mst_branch *mstb, mstb 1569 drivers/gpu/drm/drm_dp_mst_topology.c snprintf(proppath, proppath_size, "mst:%d", mstb->mgr->conn_base_id); mstb 1570 drivers/gpu/drm/drm_dp_mst_topology.c for (i = 0; i < (mstb->lct - 1); i++) { mstb 1572 drivers/gpu/drm/drm_dp_mst_topology.c int port_num = (mstb->rad[i / 2] >> shift) & 0xf; mstb 1620 drivers/gpu/drm/drm_dp_mst_topology.c static void drm_dp_add_port(struct drm_dp_mst_branch *mstb, mstb 1630 drivers/gpu/drm/drm_dp_mst_topology.c port = drm_dp_get_port(mstb, port_msg->port_number); mstb 1637 drivers/gpu/drm/drm_dp_mst_topology.c port->parent = mstb; mstb 1639 drivers/gpu/drm/drm_dp_mst_topology.c port->mgr = mstb->mgr; mstb 1648 drivers/gpu/drm/drm_dp_mst_topology.c drm_dp_mst_get_mstb_malloc(mstb); mstb 1668 drivers/gpu/drm/drm_dp_mst_topology.c mutex_lock(&mstb->mgr->lock); mstb 1670 drivers/gpu/drm/drm_dp_mst_topology.c list_add(&port->next, &mstb->ports); mstb 1671 drivers/gpu/drm/drm_dp_mst_topology.c mutex_unlock(&mstb->mgr->lock); mstb 1677 drivers/gpu/drm/drm_dp_mst_topology.c drm_dp_send_enum_path_resources(mstb->mgr, mstb 1678 drivers/gpu/drm/drm_dp_mst_topology.c mstb, port); mstb 1690 drivers/gpu/drm/drm_dp_mst_topology.c drm_dp_send_link_address(mstb->mgr, port->mstb); mstb 1696 drivers/gpu/drm/drm_dp_mst_topology.c build_mst_prop_path(mstb, port->port_num, proppath, mstb 1698 drivers/gpu/drm/drm_dp_mst_topology.c port->connector = (*mstb->mgr->cbs->add_connector)(mstb->mgr, mstb 1703 drivers/gpu/drm/drm_dp_mst_topology.c mutex_lock(&mstb->mgr->lock); mstb 1705 drivers/gpu/drm/drm_dp_mst_topology.c mutex_unlock(&mstb->mgr->lock); mstb 1717 drivers/gpu/drm/drm_dp_mst_topology.c (*mstb->mgr->cbs->register_connector)(port->connector); mstb 1725 drivers/gpu/drm/drm_dp_mst_topology.c static void drm_dp_update_port(struct drm_dp_mst_branch *mstb, mstb 1732 drivers/gpu/drm/drm_dp_mst_topology.c port = drm_dp_get_port(mstb, conn_stat->port_number); mstb 1759 drivers/gpu/drm/drm_dp_mst_topology.c queue_work(system_long_wq, &mstb->mgr->work); mstb 1766 drivers/gpu/drm/drm_dp_mst_topology.c struct drm_dp_mst_branch *mstb; mstb 1772 drivers/gpu/drm/drm_dp_mst_topology.c mstb = mgr->mst_primary; mstb 1774 drivers/gpu/drm/drm_dp_mst_topology.c if (!mstb) mstb 1781 drivers/gpu/drm/drm_dp_mst_topology.c list_for_each_entry(port, &mstb->ports, next) { mstb 1783 drivers/gpu/drm/drm_dp_mst_topology.c mstb = port->mstb; mstb 1784 drivers/gpu/drm/drm_dp_mst_topology.c if (!mstb) { mstb 1793 drivers/gpu/drm/drm_dp_mst_topology.c ret = drm_dp_mst_topology_try_get_mstb(mstb); mstb 1795 drivers/gpu/drm/drm_dp_mst_topology.c mstb = NULL; mstb 1798 drivers/gpu/drm/drm_dp_mst_topology.c return mstb; mstb 1802 drivers/gpu/drm/drm_dp_mst_topology.c struct drm_dp_mst_branch *mstb, mstb 1808 drivers/gpu/drm/drm_dp_mst_topology.c if (memcmp(mstb->guid, guid, 16) == 0) mstb 1809 drivers/gpu/drm/drm_dp_mst_topology.c return mstb; mstb 1812 drivers/gpu/drm/drm_dp_mst_topology.c list_for_each_entry(port, &mstb->ports, next) { mstb 1813 drivers/gpu/drm/drm_dp_mst_topology.c if (!port->mstb) mstb 1816 drivers/gpu/drm/drm_dp_mst_topology.c found_mstb = get_mst_branch_device_by_guid_helper(port->mstb, guid); mstb 1829 drivers/gpu/drm/drm_dp_mst_topology.c struct drm_dp_mst_branch *mstb; mstb 1835 drivers/gpu/drm/drm_dp_mst_topology.c mstb = get_mst_branch_device_by_guid_helper(mgr->mst_primary, guid); mstb 1836 drivers/gpu/drm/drm_dp_mst_topology.c if (mstb) { mstb 1837 drivers/gpu/drm/drm_dp_mst_topology.c ret = drm_dp_mst_topology_try_get_mstb(mstb); mstb 1839 drivers/gpu/drm/drm_dp_mst_topology.c mstb = NULL; mstb 1843 drivers/gpu/drm/drm_dp_mst_topology.c return mstb; mstb 1847 drivers/gpu/drm/drm_dp_mst_topology.c struct drm_dp_mst_branch *mstb) mstb 1851 drivers/gpu/drm/drm_dp_mst_topology.c if (!mstb->link_address_sent) mstb 1852 drivers/gpu/drm/drm_dp_mst_topology.c drm_dp_send_link_address(mgr, mstb); mstb 1854 drivers/gpu/drm/drm_dp_mst_topology.c list_for_each_entry(port, &mstb->ports, next) { mstb 1862 drivers/gpu/drm/drm_dp_mst_topology.c drm_dp_send_enum_path_resources(mgr, mstb, port); mstb 1864 drivers/gpu/drm/drm_dp_mst_topology.c if (port->mstb) { mstb 1866 drivers/gpu/drm/drm_dp_mst_topology.c mgr, port->mstb); mstb 1878 drivers/gpu/drm/drm_dp_mst_topology.c struct drm_dp_mst_branch *mstb; mstb 1882 drivers/gpu/drm/drm_dp_mst_topology.c mstb = mgr->mst_primary; mstb 1883 drivers/gpu/drm/drm_dp_mst_topology.c if (mstb) { mstb 1884 drivers/gpu/drm/drm_dp_mst_topology.c ret = drm_dp_mst_topology_try_get_mstb(mstb); mstb 1886 drivers/gpu/drm/drm_dp_mst_topology.c mstb = NULL; mstb 1889 drivers/gpu/drm/drm_dp_mst_topology.c if (mstb) { mstb 1890 drivers/gpu/drm/drm_dp_mst_topology.c drm_dp_check_and_send_link_address(mgr, mstb); mstb 1891 drivers/gpu/drm/drm_dp_mst_topology.c drm_dp_mst_topology_put_mstb(mstb); mstb 1959 drivers/gpu/drm/drm_dp_mst_topology.c struct drm_dp_mst_branch *mstb = txmsg->dst; mstb 1964 drivers/gpu/drm/drm_dp_mst_topology.c if (mstb->tx_slots[0] && mstb->tx_slots[1]) { mstb 1968 drivers/gpu/drm/drm_dp_mst_topology.c if (mstb->tx_slots[0] == NULL && mstb->tx_slots[1] == NULL) { mstb 1969 drivers/gpu/drm/drm_dp_mst_topology.c txmsg->seqno = mstb->last_seqno; mstb 1970 drivers/gpu/drm/drm_dp_mst_topology.c mstb->last_seqno ^= 1; mstb 1971 drivers/gpu/drm/drm_dp_mst_topology.c } else if (mstb->tx_slots[0] == NULL) mstb 1975 drivers/gpu/drm/drm_dp_mst_topology.c mstb->tx_slots[txmsg->seqno] = txmsg; mstb 1985 drivers/gpu/drm/drm_dp_mst_topology.c hdr->lct = mstb->lct; mstb 1986 drivers/gpu/drm/drm_dp_mst_topology.c hdr->lcr = mstb->lct - 1; mstb 1987 drivers/gpu/drm/drm_dp_mst_topology.c if (mstb->lct > 1) mstb 1988 drivers/gpu/drm/drm_dp_mst_topology.c memcpy(hdr->rad, mstb->rad, mstb->lct / 2); mstb 2107 drivers/gpu/drm/drm_dp_mst_topology.c struct drm_dp_mst_branch *mstb) mstb 2117 drivers/gpu/drm/drm_dp_mst_topology.c txmsg->dst = mstb; mstb 2120 drivers/gpu/drm/drm_dp_mst_topology.c mstb->link_address_sent = true; mstb 2123 drivers/gpu/drm/drm_dp_mst_topology.c ret = drm_dp_mst_wait_tx_reply(mstb, txmsg); mstb 2144 drivers/gpu/drm/drm_dp_mst_topology.c drm_dp_check_mstb_guid(mstb, txmsg->reply.u.link_addr.guid); mstb 2147 drivers/gpu/drm/drm_dp_mst_topology.c drm_dp_add_port(mstb, mgr->dev, &txmsg->reply.u.link_addr.ports[i]); mstb 2152 drivers/gpu/drm/drm_dp_mst_topology.c mstb->link_address_sent = false; mstb 2160 drivers/gpu/drm/drm_dp_mst_topology.c struct drm_dp_mst_branch *mstb, mstb 2171 drivers/gpu/drm/drm_dp_mst_topology.c txmsg->dst = mstb; mstb 2176 drivers/gpu/drm/drm_dp_mst_topology.c ret = drm_dp_mst_wait_tx_reply(mstb, txmsg); mstb 2193 drivers/gpu/drm/drm_dp_mst_topology.c static struct drm_dp_mst_port *drm_dp_get_last_connected_port_to_mstb(struct drm_dp_mst_branch *mstb) mstb 2195 drivers/gpu/drm/drm_dp_mst_topology.c if (!mstb->port_parent) mstb 2198 drivers/gpu/drm/drm_dp_mst_topology.c if (mstb->port_parent->mstb != mstb) mstb 2199 drivers/gpu/drm/drm_dp_mst_topology.c return mstb->port_parent; mstb 2201 drivers/gpu/drm/drm_dp_mst_topology.c return drm_dp_get_last_connected_port_to_mstb(mstb->port_parent->parent); mstb 2214 drivers/gpu/drm/drm_dp_mst_topology.c struct drm_dp_mst_branch *mstb, mstb 2225 drivers/gpu/drm/drm_dp_mst_topology.c found_port = drm_dp_get_last_connected_port_to_mstb(mstb); mstb 2234 drivers/gpu/drm/drm_dp_mst_topology.c mstb = found_port->parent; mstb 2248 drivers/gpu/drm/drm_dp_mst_topology.c struct drm_dp_mst_branch *mstb; mstb 2254 drivers/gpu/drm/drm_dp_mst_topology.c mstb = drm_dp_mst_topology_get_mstb_validated(mgr, port->parent); mstb 2255 drivers/gpu/drm/drm_dp_mst_topology.c if (!mstb) { mstb 2256 drivers/gpu/drm/drm_dp_mst_topology.c mstb = drm_dp_get_last_connected_port_and_mstb(mgr, mstb 2260 drivers/gpu/drm/drm_dp_mst_topology.c if (!mstb) mstb 2273 drivers/gpu/drm/drm_dp_mst_topology.c txmsg->dst = mstb; mstb 2288 drivers/gpu/drm/drm_dp_mst_topology.c ret = drm_dp_mst_wait_tx_reply(mstb, txmsg); mstb 2297 drivers/gpu/drm/drm_dp_mst_topology.c drm_dp_mst_topology_put_mstb(mstb); mstb 2543 drivers/gpu/drm/drm_dp_mst_topology.c struct drm_dp_mst_branch *mstb; mstb 2545 drivers/gpu/drm/drm_dp_mst_topology.c mstb = drm_dp_mst_topology_get_mstb_validated(mgr, port->parent); mstb 2546 drivers/gpu/drm/drm_dp_mst_topology.c if (!mstb) mstb 2560 drivers/gpu/drm/drm_dp_mst_topology.c ret = drm_dp_mst_wait_tx_reply(mstb, txmsg); mstb 2567 drivers/gpu/drm/drm_dp_mst_topology.c mstb, port->port_num, offset, size); mstb 2584 drivers/gpu/drm/drm_dp_mst_topology.c drm_dp_mst_topology_put_mstb(mstb); mstb 2596 drivers/gpu/drm/drm_dp_mst_topology.c struct drm_dp_mst_branch *mstb; mstb 2598 drivers/gpu/drm/drm_dp_mst_topology.c mstb = drm_dp_mst_topology_get_mstb_validated(mgr, port->parent); mstb 2599 drivers/gpu/drm/drm_dp_mst_topology.c if (!mstb) mstb 2609 drivers/gpu/drm/drm_dp_mst_topology.c txmsg->dst = mstb; mstb 2613 drivers/gpu/drm/drm_dp_mst_topology.c ret = drm_dp_mst_wait_tx_reply(mstb, txmsg); mstb 2622 drivers/gpu/drm/drm_dp_mst_topology.c drm_dp_mst_topology_put_mstb(mstb); mstb 2637 drivers/gpu/drm/drm_dp_mst_topology.c struct drm_dp_mst_branch *mstb, mstb 2646 drivers/gpu/drm/drm_dp_mst_topology.c txmsg->dst = mstb; mstb 2697 drivers/gpu/drm/drm_dp_mst_topology.c struct drm_dp_mst_branch *mstb = NULL; mstb 2724 drivers/gpu/drm/drm_dp_mst_topology.c mstb = drm_dp_add_mst_branch_device(1, NULL); mstb 2725 drivers/gpu/drm/drm_dp_mst_topology.c if (mstb == NULL) { mstb 2729 drivers/gpu/drm/drm_dp_mst_topology.c mstb->mgr = mgr; mstb 2732 drivers/gpu/drm/drm_dp_mst_topology.c mgr->mst_primary = mstb; mstb 2753 drivers/gpu/drm/drm_dp_mst_topology.c mstb = mgr->mst_primary; mstb 2770 drivers/gpu/drm/drm_dp_mst_topology.c if (mstb) mstb 2771 drivers/gpu/drm/drm_dp_mst_topology.c drm_dp_mst_topology_put_mstb(mstb); mstb 2910 drivers/gpu/drm/drm_dp_mst_topology.c struct drm_dp_mst_branch *mstb; mstb 2912 drivers/gpu/drm/drm_dp_mst_topology.c mstb = drm_dp_get_mst_branch_device(mgr, mstb 2916 drivers/gpu/drm/drm_dp_mst_topology.c if (!mstb) { mstb 2925 drivers/gpu/drm/drm_dp_mst_topology.c txmsg = mstb->tx_slots[slot]; mstb 2931 drivers/gpu/drm/drm_dp_mst_topology.c mstb, mstb 2936 drivers/gpu/drm/drm_dp_mst_topology.c drm_dp_mst_topology_put_mstb(mstb); mstb 2952 drivers/gpu/drm/drm_dp_mst_topology.c drm_dp_mst_topology_put_mstb(mstb); mstb 2956 drivers/gpu/drm/drm_dp_mst_topology.c mstb->tx_slots[slot] = NULL; mstb 2976 drivers/gpu/drm/drm_dp_mst_topology.c struct drm_dp_mst_branch *mstb = NULL; mstb 2980 drivers/gpu/drm/drm_dp_mst_topology.c mstb = drm_dp_get_mst_branch_device(mgr, mstb 2983 drivers/gpu/drm/drm_dp_mst_topology.c if (!mstb) { mstb 2996 drivers/gpu/drm/drm_dp_mst_topology.c if (!mstb) mstb 2997 drivers/gpu/drm/drm_dp_mst_topology.c mstb = drm_dp_get_mst_branch_device_by_guid(mgr, msg.u.conn_stat.guid); mstb 2999 drivers/gpu/drm/drm_dp_mst_topology.c if (!mstb) { mstb 3005 drivers/gpu/drm/drm_dp_mst_topology.c drm_dp_update_port(mstb, &msg.u.conn_stat); mstb 3012 drivers/gpu/drm/drm_dp_mst_topology.c if (!mstb) mstb 3013 drivers/gpu/drm/drm_dp_mst_topology.c mstb = drm_dp_get_mst_branch_device_by_guid(mgr, msg.u.resource_stat.guid); mstb 3015 drivers/gpu/drm/drm_dp_mst_topology.c if (!mstb) { mstb 3024 drivers/gpu/drm/drm_dp_mst_topology.c if (mstb) mstb 3025 drivers/gpu/drm/drm_dp_mst_topology.c drm_dp_mst_topology_put_mstb(mstb); mstb 3608 drivers/gpu/drm/drm_dp_mst_topology.c struct drm_dp_mst_branch *mstb) mstb 3611 drivers/gpu/drm/drm_dp_mst_topology.c int tabs = mstb->lct; mstb 3619 drivers/gpu/drm/drm_dp_mst_topology.c seq_printf(m, "%smst: %p, %d\n", prefix, mstb, mstb->num_ports); mstb 3620 drivers/gpu/drm/drm_dp_mst_topology.c list_for_each_entry(port, &mstb->ports, next) { mstb 3622 drivers/gpu/drm/drm_dp_mst_topology.c if (port->mstb) mstb 3623 drivers/gpu/drm/drm_dp_mst_topology.c drm_dp_mst_dump_mstb(m, port->mstb); mstb 4043 drivers/gpu/drm/drm_dp_mst_topology.c struct drm_dp_mst_branch *mstb; mstb 4050 drivers/gpu/drm/drm_dp_mst_topology.c mstb = drm_dp_mst_topology_get_mstb_validated(mgr, port->parent); mstb 4051 drivers/gpu/drm/drm_dp_mst_topology.c if (!mstb) mstb 4079 drivers/gpu/drm/drm_dp_mst_topology.c txmsg->dst = mstb; mstb 4084 drivers/gpu/drm/drm_dp_mst_topology.c ret = drm_dp_mst_wait_tx_reply(mstb, txmsg); mstb 4100 drivers/gpu/drm/drm_dp_mst_topology.c drm_dp_mst_topology_put_mstb(mstb); mstb 93 include/drm/drm_dp_mst_helper.h struct drm_dp_mst_branch *mstb; /* pointer to an mstb if this port has one */