Lines Matching refs:dev_queue

2503 static inline void netif_tx_start_queue(struct netdev_queue *dev_queue)  in netif_tx_start_queue()  argument
2505 clear_bit(__QUEUE_STATE_DRV_XOFF, &dev_queue->state); in netif_tx_start_queue()
2529 void netif_tx_wake_queue(struct netdev_queue *dev_queue);
2553 static inline void netif_tx_stop_queue(struct netdev_queue *dev_queue) in netif_tx_stop_queue() argument
2555 if (WARN_ON(!dev_queue)) { in netif_tx_stop_queue()
2559 set_bit(__QUEUE_STATE_DRV_XOFF, &dev_queue->state); in netif_tx_stop_queue()
2584 static inline bool netif_tx_queue_stopped(const struct netdev_queue *dev_queue) in netif_tx_queue_stopped() argument
2586 return test_bit(__QUEUE_STATE_DRV_XOFF, &dev_queue->state); in netif_tx_queue_stopped()
2600 static inline bool netif_xmit_stopped(const struct netdev_queue *dev_queue) in netif_xmit_stopped() argument
2602 return dev_queue->state & QUEUE_STATE_ANY_XOFF; in netif_xmit_stopped()
2606 netif_xmit_frozen_or_stopped(const struct netdev_queue *dev_queue) in netif_xmit_frozen_or_stopped() argument
2608 return dev_queue->state & QUEUE_STATE_ANY_XOFF_OR_FROZEN; in netif_xmit_frozen_or_stopped()
2612 netif_xmit_frozen_or_drv_stopped(const struct netdev_queue *dev_queue) in netif_xmit_frozen_or_drv_stopped() argument
2614 return dev_queue->state & QUEUE_STATE_DRV_XOFF_OR_FROZEN; in netif_xmit_frozen_or_drv_stopped()
2624 static inline void netdev_txq_bql_enqueue_prefetchw(struct netdev_queue *dev_queue) in netdev_txq_bql_enqueue_prefetchw() argument
2627 prefetchw(&dev_queue->dql.num_queued); in netdev_txq_bql_enqueue_prefetchw()
2638 static inline void netdev_txq_bql_complete_prefetchw(struct netdev_queue *dev_queue) in netdev_txq_bql_complete_prefetchw() argument
2641 prefetchw(&dev_queue->dql.limit); in netdev_txq_bql_complete_prefetchw()
2645 static inline void netdev_tx_sent_queue(struct netdev_queue *dev_queue, in netdev_tx_sent_queue() argument
2649 dql_queued(&dev_queue->dql, bytes); in netdev_tx_sent_queue()
2651 if (likely(dql_avail(&dev_queue->dql) >= 0)) in netdev_tx_sent_queue()
2654 set_bit(__QUEUE_STATE_STACK_XOFF, &dev_queue->state); in netdev_tx_sent_queue()
2664 if (unlikely(dql_avail(&dev_queue->dql) >= 0)) in netdev_tx_sent_queue()
2665 clear_bit(__QUEUE_STATE_STACK_XOFF, &dev_queue->state); in netdev_tx_sent_queue()
2683 static inline void netdev_tx_completed_queue(struct netdev_queue *dev_queue, in netdev_tx_completed_queue() argument
2690 dql_completed(&dev_queue->dql, bytes); in netdev_tx_completed_queue()
2699 if (dql_avail(&dev_queue->dql) < 0) in netdev_tx_completed_queue()
2702 if (test_and_clear_bit(__QUEUE_STATE_STACK_XOFF, &dev_queue->state)) in netdev_tx_completed_queue()
2703 netif_schedule_queue(dev_queue); in netdev_tx_completed_queue()
2738 static inline void netdev_reset_queue(struct net_device *dev_queue) in netdev_reset_queue() argument
2740 netdev_tx_reset_queue(netdev_get_tx_queue(dev_queue, 0)); in netdev_reset_queue()