Lines Matching refs:vif
122 #define callback_param(vif, pending_idx) \ argument
123 (vif->pending_tx_info[pending_idx].callback_struct)
184 netif_tx_stop_queue(netdev_get_tx_queue(queue->vif->dev, queue->id)); in xenvif_rx_queue_tail()
209 netif_tx_wake_queue(netdev_get_tx_queue(queue->vif->dev, queue->id)); in xenvif_rx_queue_maybe_wake()
321 copy_gop->dest.domid = queue->vif->domid; in xenvif_gop_frag_copy()
346 if (*head && ((1 << gso_type) & queue->vif->gso_mask)) in xenvif_gop_frag_copy()
370 struct xenvif *vif = netdev_priv(skb->dev); in xenvif_gop_skb() local
391 if ((1 << gso_type) & vif->gso_prefix_mask) { in xenvif_gop_skb()
403 if ((1 << gso_type) & vif->gso_mask) { in xenvif_gop_skb()
446 static int xenvif_check_gop(struct xenvif *vif, int nr_meta_slots, in xenvif_check_gop() argument
456 netdev_dbg(vif->dev, in xenvif_check_gop()
458 copy_op->status, vif->domid); in xenvif_check_gop()
541 queue->vif->gso_prefix_mask) { in xenvif_rx_action()
559 status = xenvif_check_gop(queue->vif, in xenvif_rx_action()
581 queue->vif->gso_mask) { in xenvif_rx_action()
670 static void xenvif_fatal_tx_err(struct xenvif *vif) in xenvif_fatal_tx_err() argument
672 netdev_err(vif->dev, "fatal error; disabling device\n"); in xenvif_fatal_tx_err()
673 vif->disabled = true; in xenvif_fatal_tx_err()
675 if (vif->queues) in xenvif_fatal_tx_err()
676 xenvif_kick_thread(&vif->queues[0]); in xenvif_fatal_tx_err()
696 netdev_err(queue->vif->dev, in xenvif_count_requests()
699 xenvif_fatal_tx_err(queue->vif); in xenvif_count_requests()
707 netdev_err(queue->vif->dev, in xenvif_count_requests()
710 xenvif_fatal_tx_err(queue->vif); in xenvif_count_requests()
723 netdev_dbg(queue->vif->dev, in xenvif_count_requests()
746 netdev_dbg(queue->vif->dev, in xenvif_count_requests()
756 netdev_err(queue->vif->dev, "Cross page boundary, txp->offset: %x, size: %u\n", in xenvif_count_requests()
758 xenvif_fatal_tx_err(queue->vif); in xenvif_count_requests()
792 txp->gref, queue->vif->domid); in xenvif_tx_create_map_op()
852 netdev_err(queue->vif->dev, in xenvif_get_requests()
881 netdev_err(queue->vif->dev, in xenvif_grant_handle_set()
894 netdev_err(queue->vif->dev, in xenvif_grant_handle_reset()
926 netdev_dbg(queue->vif->dev, in xenvif_tx_check_gop()
970 netdev_dbg(queue->vif->dev, in xenvif_tx_check_gop()
1074 netdev_err(queue->vif->dev, "Missing extra info\n"); in xenvif_get_extras()
1075 xenvif_fatal_tx_err(queue->vif); in xenvif_get_extras()
1084 netdev_err(queue->vif->dev, in xenvif_get_extras()
1086 xenvif_fatal_tx_err(queue->vif); in xenvif_get_extras()
1097 static int xenvif_set_skb_gso(struct xenvif *vif, in xenvif_set_skb_gso() argument
1102 netdev_err(vif->dev, "GSO size must not be zero.\n"); in xenvif_set_skb_gso()
1103 xenvif_fatal_tx_err(vif); in xenvif_set_skb_gso()
1115 netdev_err(vif->dev, "Bad GSO type %d.\n", gso->u.gso.type); in xenvif_set_skb_gso()
1116 xenvif_fatal_tx_err(vif); in xenvif_set_skb_gso()
1199 netdev_err(queue->vif->dev, in xenvif_tx_build_gops()
1204 xenvif_fatal_tx_err(queue->vif); in xenvif_tx_build_gops()
1242 netdev_dbg(queue->vif->dev, in xenvif_tx_build_gops()
1250 netdev_err(queue->vif->dev, in xenvif_tx_build_gops()
1254 xenvif_fatal_tx_err(queue->vif); in xenvif_tx_build_gops()
1267 netdev_dbg(queue->vif->dev, in xenvif_tx_build_gops()
1277 if (xenvif_set_skb_gso(queue->vif, skb, gso)) { in xenvif_tx_build_gops()
1288 queue->tx_copy_ops[*copy_ops].source.domid = queue->vif->domid; in xenvif_tx_build_gops()
1462 netdev_err(queue->vif->dev, in xenvif_tx_submit()
1470 skb->dev = queue->vif->dev; in xenvif_tx_submit()
1475 netdev_dbg(queue->vif->dev, in xenvif_tx_submit()
1601 netdev_err(queue->vif->dev, "Unmap fail: nr_ops %tx ret %d\n", in xenvif_tx_dealloc_action()
1605 netdev_err(queue->vif->dev, in xenvif_tx_dealloc_action()
1738 netdev_err(queue->vif->dev, in xenvif_idx_unmap()
1765 xenbus_unmap_ring_vfree(xenvif_to_xenbus_device(queue->vif), in xenvif_unmap_frontend_rings()
1768 xenbus_unmap_ring_vfree(xenvif_to_xenbus_device(queue->vif), in xenvif_unmap_frontend_rings()
1782 err = xenbus_map_ring_valloc(xenvif_to_xenbus_device(queue->vif), in xenvif_map_frontend_rings()
1790 err = xenbus_map_ring_valloc(xenvif_to_xenbus_device(queue->vif), in xenvif_map_frontend_rings()
1807 struct xenvif *vif = queue->vif; in xenvif_queue_carrier_off() local
1812 spin_lock(&vif->lock); in xenvif_queue_carrier_off()
1813 if (vif->stalled_queues++ == 0) { in xenvif_queue_carrier_off()
1814 netdev_info(vif->dev, "Guest Rx stalled"); in xenvif_queue_carrier_off()
1815 netif_carrier_off(vif->dev); in xenvif_queue_carrier_off()
1817 spin_unlock(&vif->lock); in xenvif_queue_carrier_off()
1822 struct xenvif *vif = queue->vif; in xenvif_queue_carrier_on() local
1828 spin_lock(&vif->lock); in xenvif_queue_carrier_on()
1829 if (--vif->stalled_queues == 0) { in xenvif_queue_carrier_on()
1830 netdev_info(vif->dev, "Guest Rx ready"); in xenvif_queue_carrier_on()
1831 netif_carrier_on(vif->dev); in xenvif_queue_carrier_on()
1833 spin_unlock(&vif->lock); in xenvif_queue_carrier_on()
1846 queue->last_rx_time + queue->vif->stall_timeout); in xenvif_rx_queue_stalled()
1864 || (queue->vif->stall_timeout && in xenvif_have_rx_work()
1868 || queue->vif->disabled; in xenvif_have_rx_work()
1917 struct xenvif *vif = queue->vif; in xenvif_kthread_guest_rx() local
1919 if (!vif->stall_timeout) in xenvif_kthread_guest_rx()
1935 if (unlikely(vif->disabled && queue->id == 0)) { in xenvif_kthread_guest_rx()
1936 xenvif_carrier_off(vif); in xenvif_kthread_guest_rx()
1947 if (vif->stall_timeout) { in xenvif_kthread_guest_rx()