Lines Matching refs:queue

49 						unsigned int queue)  in __carl9170_get_queue()  argument
52 return queue; in __carl9170_get_queue()
78 int queue, i; in carl9170_tx_accounting() local
83 queue = skb_get_queue_mapping(skb); in carl9170_tx_accounting()
92 ar->tx_stats[queue].len++; in carl9170_tx_accounting()
93 ar->tx_stats[queue].count++; in carl9170_tx_accounting()
158 int queue; in carl9170_tx_accounting_free() local
160 queue = skb_get_queue_mapping(skb); in carl9170_tx_accounting_free()
164 ar->tx_stats[queue].len--; in carl9170_tx_accounting_free()
455 int queue = skb_get_queue_mapping(skb); in carl9170_tx_bar_status() local
458 list_for_each_entry_rcu(entry, &ar->bar_list[queue], list) { in carl9170_tx_bar_status()
460 spin_lock_bh(&ar->bar_list_lock[queue]); in carl9170_tx_bar_status()
462 spin_unlock_bh(&ar->bar_list_lock[queue]); in carl9170_tx_bar_status()
469 queue, bar->ra, bar->ta, bar->control, in carl9170_tx_bar_status()
514 struct sk_buff_head *queue) in carl9170_get_queued_skb() argument
518 spin_lock_bh(&queue->lock); in carl9170_get_queued_skb()
519 skb_queue_walk(queue, skb) { in carl9170_get_queued_skb()
525 __skb_unlink(skb, queue); in carl9170_get_queued_skb()
526 spin_unlock_bh(&queue->lock); in carl9170_get_queued_skb()
531 spin_unlock_bh(&queue->lock); in carl9170_get_queued_skb()
617 skb = skb_peek(&iter->queue); in carl9170_tx_ampdu_timeout()
1131 u16 seq, queue, tmpssn; in carl9170_tx_ampdu() local
1158 queue = TID_TO_WME_AC(tid_info->tid); in carl9170_tx_ampdu()
1165 first = skb_peek(&tid_info->queue); in carl9170_tx_ampdu()
1176 while ((skb = skb_peek(&tid_info->queue))) { in carl9170_tx_ampdu()
1196 __skb_unlink(skb, &tid_info->queue); in carl9170_tx_ampdu()
1204 if (skb_queue_empty(&tid_info->queue) || in carl9170_tx_ampdu()
1205 carl9170_get_seq(skb_peek(&tid_info->queue)) != in carl9170_tx_ampdu()
1227 spin_lock_bh(&ar->tx_pending[queue].lock); in carl9170_tx_ampdu()
1228 skb_queue_splice_tail_init(&agg, &ar->tx_pending[queue]); in carl9170_tx_ampdu()
1229 spin_unlock_bh(&ar->tx_pending[queue].lock); in carl9170_tx_ampdu()
1240 struct sk_buff_head *queue) in carl9170_tx_pick_skb() argument
1248 spin_lock_bh(&queue->lock); in carl9170_tx_pick_skb()
1249 skb = skb_peek(queue); in carl9170_tx_pick_skb()
1256 __skb_unlink(skb, queue); in carl9170_tx_pick_skb()
1257 spin_unlock_bh(&queue->lock); in carl9170_tx_pick_skb()
1266 spin_unlock_bh(&queue->lock); in carl9170_tx_pick_skb()
1324 unsigned int queue = skb_get_queue_mapping(skb); in carl9170_bar_check() local
1329 spin_lock_bh(&ar->bar_list_lock[queue]); in carl9170_bar_check()
1330 list_add_tail_rcu(&entry->list, &ar->bar_list[queue]); in carl9170_bar_check()
1331 spin_unlock_bh(&ar->bar_list_lock[queue]); in carl9170_bar_check()
1427 __skb_queue_tail(&agg->queue, skb); in carl9170_tx_ampdu_queue()
1432 skb_queue_reverse_walk(&agg->queue, iter) { in carl9170_tx_ampdu_queue()
1436 __skb_queue_after(&agg->queue, iter, skb); in carl9170_tx_ampdu_queue()
1441 __skb_queue_head(&agg->queue, skb); in carl9170_tx_ampdu_queue()
1445 if (agg->snx == carl9170_get_seq(skb_peek(&agg->queue))) { in carl9170_tx_ampdu_queue()
1510 unsigned int queue = skb_get_queue_mapping(skb); in carl9170_op_tx() local
1514 skb_queue_tail(&ar->tx_pending[queue], skb); in carl9170_op_tx()