txq               613 drivers/atm/ambassador.c   amb_txq * txq = &dev->txq;
txq               621 drivers/atm/ambassador.c   spin_lock_irqsave (&txq->lock, flags);
txq               623 drivers/atm/ambassador.c   if (txq->pending < txq->maximum) {
txq               624 drivers/atm/ambassador.c     PRINTD (DBG_TX, "TX in slot %p", txq->in.ptr);
txq               626 drivers/atm/ambassador.c     *txq->in.ptr = *tx;
txq               627 drivers/atm/ambassador.c     txq->pending++;
txq               628 drivers/atm/ambassador.c     txq->in.ptr = NEXTQ (txq->in.ptr, txq->in.start, txq->in.limit);
txq               630 drivers/atm/ambassador.c     wr_mem (dev, offsetof(amb_mem, mb.adapter.tx_address), virt_to_bus (txq->in.ptr));
txq               633 drivers/atm/ambassador.c     if (txq->pending > txq->high)
txq               634 drivers/atm/ambassador.c       txq->high = txq->pending;
txq               635 drivers/atm/ambassador.c     spin_unlock_irqrestore (&txq->lock, flags);
txq               638 drivers/atm/ambassador.c     txq->filled++;
txq               639 drivers/atm/ambassador.c     spin_unlock_irqrestore (&txq->lock, flags);
txq               645 drivers/atm/ambassador.c   amb_txq * txq = &dev->txq;
txq               650 drivers/atm/ambassador.c   spin_lock_irqsave (&txq->lock, flags);
txq               652 drivers/atm/ambassador.c   if (txq->pending && txq->out.ptr->handle) {
txq               654 drivers/atm/ambassador.c     tx_complete (dev, txq->out.ptr);
txq               656 drivers/atm/ambassador.c     txq->out.ptr->handle = 0;
txq               658 drivers/atm/ambassador.c     txq->pending--;
txq               659 drivers/atm/ambassador.c     txq->out.ptr = NEXTQ (txq->out.ptr, txq->out.start, txq->out.limit);
txq               661 drivers/atm/ambassador.c     spin_unlock_irqrestore (&txq->lock, flags);
txq               665 drivers/atm/ambassador.c     spin_unlock_irqrestore (&txq->lock, flags);
txq              1435 drivers/atm/ambassador.c     amb_txq * t = &dev->txq;
txq              1549 drivers/atm/ambassador.c     amb_txq * txq = &dev->txq;
txq              1551 drivers/atm/ambassador.c     txq->pending = 0;
txq              1552 drivers/atm/ambassador.c     txq->high = 0;
txq              1553 drivers/atm/ambassador.c     txq->filled = 0;
txq              1554 drivers/atm/ambassador.c     txq->maximum = txs - 1;
txq              1556 drivers/atm/ambassador.c     txq->in.start = in;
txq              1557 drivers/atm/ambassador.c     txq->in.ptr = in;
txq              1558 drivers/atm/ambassador.c     txq->in.limit = in + txs;
txq              1560 drivers/atm/ambassador.c     memory = txq->in.limit;
txq              1563 drivers/atm/ambassador.c     txq->out.start = out;
txq              1564 drivers/atm/ambassador.c     txq->out.ptr = out;
txq              1565 drivers/atm/ambassador.c     txq->out.limit = out + txs;
txq              1567 drivers/atm/ambassador.c     memory = txq->out.limit;
txq              1981 drivers/atm/ambassador.c   a.tx_start      = bus_addr (dev->txq.in.start);
txq              1982 drivers/atm/ambassador.c   a.tx_end        = bus_addr (dev->txq.in.limit);
txq              1983 drivers/atm/ambassador.c   a.txcom_start   = bus_addr (dev->txq.out.start);
txq              1984 drivers/atm/ambassador.c   a.txcom_end     = bus_addr (dev->txq.out.limit);
txq              2153 drivers/atm/ambassador.c       spin_lock_init (&dev->txq.lock);
txq               619 drivers/atm/ambassador.h   amb_txq          txq;
txq              1393 drivers/atm/firestream.c static int init_q(struct fs_dev *dev, struct queue *txq, int queue,
txq              1420 drivers/atm/firestream.c 	txq->sa = p;
txq              1421 drivers/atm/firestream.c 	txq->ea = p;
txq              1422 drivers/atm/firestream.c 	txq->offset = queue; 
txq              1523 drivers/atm/firestream.c static void free_queue(struct fs_dev *dev, struct queue *txq)
txq              1527 drivers/atm/firestream.c 	write_fs (dev, Q_SA(txq->offset), 0);
txq              1528 drivers/atm/firestream.c 	write_fs (dev, Q_EA(txq->offset), 0);
txq              1529 drivers/atm/firestream.c 	write_fs (dev, Q_RP(txq->offset), 0);
txq              1530 drivers/atm/firestream.c 	write_fs (dev, Q_WP(txq->offset), 0);
txq              1533 drivers/atm/firestream.c 	fs_dprintk (FS_DEBUG_ALLOC, "Free queue: %p\n", txq->sa);
txq              1534 drivers/atm/firestream.c 	kfree (txq->sa);
txq               772 drivers/atm/fore200e.c     struct host_txq*        txq = &fore200e->host_txq;
txq               782 drivers/atm/fore200e.c 	entry = &txq->host_entry[ txq->tail ];
txq               789 drivers/atm/fore200e.c 		entry, txq->tail, entry->vc_map, entry->skb);
txq               856 drivers/atm/fore200e.c 	FORE200E_NEXT_ENTRY(txq->tail, QUEUE_SIZE_TX);
txq              1472 drivers/atm/fore200e.c     struct host_txq*        txq;
txq              1494 drivers/atm/fore200e.c     txq = &fore200e->host_txq;
txq              1561 drivers/atm/fore200e.c     entry = &txq->host_entry[ txq->head ];
txq              1563 drivers/atm/fore200e.c     if ((*entry->status != STATUS_FREE) || (txq->txing >= QUEUE_SIZE_TX - 2)) {
txq              1613 drivers/atm/fore200e.c     FORE200E_NEXT_ENTRY(txq->head, QUEUE_SIZE_TX);
txq              1614 drivers/atm/fore200e.c     txq->txing++;
txq              2197 drivers/atm/fore200e.c     struct host_txq*     txq =  &fore200e->host_txq;
txq              2205 drivers/atm/fore200e.c 				       &txq->status,
txq              2214 drivers/atm/fore200e.c 				       &txq->tpd,
txq              2219 drivers/atm/fore200e.c 	fore200e_dma_chunk_free(fore200e, &txq->status);
txq              2229 drivers/atm/fore200e.c 	txq->host_entry[ i ].status = 
txq              2230 drivers/atm/fore200e.c 	                     FORE200E_INDEX(txq->status.align_addr, enum status, i);
txq              2231 drivers/atm/fore200e.c 	txq->host_entry[ i ].tpd = 
txq              2232 drivers/atm/fore200e.c 	                     FORE200E_INDEX(txq->tpd.align_addr, struct tpd, i);
txq              2233 drivers/atm/fore200e.c 	txq->host_entry[ i ].tpd_dma  = 
txq              2234 drivers/atm/fore200e.c                              FORE200E_DMA_INDEX(txq->tpd.dma_addr, struct tpd, i);
txq              2235 drivers/atm/fore200e.c 	txq->host_entry[ i ].cp_entry = &cp_entry[ i ];
txq              2237 drivers/atm/fore200e.c 	*txq->host_entry[ i ].status = STATUS_FREE;
txq              2239 drivers/atm/fore200e.c 	fore200e->bus->write(FORE200E_DMA_INDEX(txq->status.dma_addr, enum status, i), 
txq              2249 drivers/atm/fore200e.c     txq->head = 0;
txq              2250 drivers/atm/fore200e.c     txq->tail = 0;
txq                72 drivers/bluetooth/bluecard_cs.c 	struct sk_buff_head txq;
txq               259 drivers/bluetooth/bluecard_cs.c 		skb = skb_dequeue(&(info->txq));
txq               324 drivers/bluetooth/bluecard_cs.c 			skb_queue_head(&(info->txq), skb);
txq               597 drivers/bluetooth/bluecard_cs.c 	skb_queue_tail(&(info->txq), skb);
txq               614 drivers/bluetooth/bluecard_cs.c 	skb_queue_purge(&(info->txq));
txq               670 drivers/bluetooth/bluecard_cs.c 	skb_queue_tail(&(info->txq), skb);
txq               692 drivers/bluetooth/bluecard_cs.c 	skb_queue_head_init(&(info->txq));
txq               775 drivers/bluetooth/bluecard_cs.c 	skb_queue_purge(&(info->txq));
txq                77 drivers/bluetooth/bt3c_cs.c 	struct sk_buff_head txq;
txq               196 drivers/bluetooth/bt3c_cs.c 		skb = skb_dequeue(&(info->txq));
txq               390 drivers/bluetooth/bt3c_cs.c 	skb_queue_purge(&(info->txq));
txq               429 drivers/bluetooth/bt3c_cs.c 	skb_queue_tail(&(info->txq), skb);
txq               546 drivers/bluetooth/bt3c_cs.c 	skb_queue_head_init(&(info->txq));
txq               162 drivers/bluetooth/btmtksdio.c 	struct sk_buff_head txq;
txq               314 drivers/bluetooth/btmtksdio.c 	while ((skb = skb_dequeue(&bdev->txq))) {
txq               318 drivers/bluetooth/btmtksdio.c 			skb_queue_head(&bdev->txq, skb);
txq               651 drivers/bluetooth/btmtksdio.c 	skb_queue_purge(&bdev->txq);
txq               918 drivers/bluetooth/btmtksdio.c 	skb_queue_tail(&bdev->txq, skb);
txq               944 drivers/bluetooth/btmtksdio.c 	skb_queue_head_init(&bdev->txq);
txq               132 drivers/bluetooth/btmtkuart.c 	struct sk_buff_head txq;
txq               379 drivers/bluetooth/btmtkuart.c 			struct sk_buff *skb = skb_dequeue(&bdev->txq);
txq               391 drivers/bluetooth/btmtkuart.c 				skb_queue_head(&bdev->txq, skb);
txq               627 drivers/bluetooth/btmtkuart.c 	skb_queue_purge(&bdev->txq);
txq               892 drivers/bluetooth/btmtkuart.c 	skb_queue_tail(&bdev->txq, skb);
txq               990 drivers/bluetooth/btmtkuart.c 	skb_queue_head_init(&bdev->txq);
txq                49 drivers/bluetooth/btsdio.c 	struct sk_buff_head txq;
txq               101 drivers/bluetooth/btsdio.c 	while ((skb = skb_dequeue(&data->txq))) {
txq               105 drivers/bluetooth/btsdio.c 			skb_queue_head(&data->txq, skb);
txq               231 drivers/bluetooth/btsdio.c 	skb_queue_purge(&data->txq);
txq               259 drivers/bluetooth/btsdio.c 	skb_queue_tail(&data->txq, skb);
txq               302 drivers/bluetooth/btsdio.c 	skb_queue_head_init(&data->txq);
txq                75 drivers/bluetooth/dtl1_cs.c 	struct sk_buff_head txq;
txq               156 drivers/bluetooth/dtl1_cs.c 		skb = skb_dequeue(&(info->txq));
txq               168 drivers/bluetooth/dtl1_cs.c 			skb_queue_head(&(info->txq), skb);
txq               370 drivers/bluetooth/dtl1_cs.c 	skb_queue_purge(&(info->txq));
txq               421 drivers/bluetooth/dtl1_cs.c 	skb_queue_tail(&(info->txq), s);
txq               443 drivers/bluetooth/dtl1_cs.c 	skb_queue_head_init(&(info->txq));
txq                24 drivers/bluetooth/hci_ag6xx.c 	struct sk_buff_head txq;
txq                43 drivers/bluetooth/hci_ag6xx.c 	skb_queue_head_init(&ag6xx->txq);
txq                55 drivers/bluetooth/hci_ag6xx.c 	skb_queue_purge(&ag6xx->txq);
txq                69 drivers/bluetooth/hci_ag6xx.c 	skb_queue_purge(&ag6xx->txq);
txq                78 drivers/bluetooth/hci_ag6xx.c 	skb = skb_dequeue(&ag6xx->txq);
txq                91 drivers/bluetooth/hci_ag6xx.c 	skb_queue_tail(&ag6xx->txq, skb);
txq                35 drivers/bluetooth/hci_ath.c 	struct sk_buff_head txq;
txq               108 drivers/bluetooth/hci_ath.c 	skb_queue_head_init(&ath->txq);
txq               124 drivers/bluetooth/hci_ath.c 	skb_queue_purge(&ath->txq);
txq               142 drivers/bluetooth/hci_ath.c 	skb_queue_purge(&ath->txq);
txq               232 drivers/bluetooth/hci_ath.c 	skb_queue_tail(&ath->txq, skb);
txq               244 drivers/bluetooth/hci_ath.c 	return skb_dequeue(&ath->txq);
txq               120 drivers/bluetooth/hci_bcm.c 	struct sk_buff_head	txq;
txq               395 drivers/bluetooth/hci_bcm.c 	skb_queue_tail(&bcm->txq, skb);
txq               416 drivers/bluetooth/hci_bcm.c 	skb_queue_head_init(&bcm->txq);
txq               504 drivers/bluetooth/hci_bcm.c 	skb_queue_purge(&bcm->txq);
txq               518 drivers/bluetooth/hci_bcm.c 	skb_queue_purge(&bcm->txq);
txq               664 drivers/bluetooth/hci_bcm.c 	skb_queue_tail(&bcm->txq, skb);
txq               683 drivers/bluetooth/hci_bcm.c 	skb = skb_dequeue(&bcm->txq);
txq                37 drivers/bluetooth/hci_h4.c 	struct sk_buff_head txq;
txq                51 drivers/bluetooth/hci_h4.c 	skb_queue_head_init(&h4->txq);
txq                64 drivers/bluetooth/hci_h4.c 	skb_queue_purge(&h4->txq);
txq                78 drivers/bluetooth/hci_h4.c 	skb_queue_purge(&h4->txq);
txq                97 drivers/bluetooth/hci_h4.c 	skb_queue_tail(&h4->txq, skb);
txq               131 drivers/bluetooth/hci_h4.c 	return skb_dequeue(&h4->txq);
txq                69 drivers/bluetooth/hci_intel.c 	struct sk_buff_head txq;
txq               179 drivers/bluetooth/hci_intel.c 	skb_queue_head(&intel->txq, skb);
txq               223 drivers/bluetooth/hci_intel.c 	skb_queue_head(&intel->txq, skb);
txq               262 drivers/bluetooth/hci_intel.c 	skb_queue_head(&intel->txq, skb);
txq               401 drivers/bluetooth/hci_intel.c 	skb_queue_head_init(&intel->txq);
txq               424 drivers/bluetooth/hci_intel.c 	skb_queue_purge(&intel->txq);
txq               438 drivers/bluetooth/hci_intel.c 	skb_queue_purge(&intel->txq);
txq               517 drivers/bluetooth/hci_intel.c 	skb_queue_tail(&intel->txq, skb);
txq              1026 drivers/bluetooth/hci_intel.c 	skb_queue_tail(&intel->txq, skb);
txq              1036 drivers/bluetooth/hci_intel.c 	skb = skb_dequeue(&intel->txq);
txq                75 drivers/bluetooth/hci_ll.c 	struct sk_buff_head txq;
txq               105 drivers/bluetooth/hci_ll.c 	skb_queue_tail(&ll->txq, skb);
txq               121 drivers/bluetooth/hci_ll.c 	skb_queue_head_init(&ll->txq);
txq               147 drivers/bluetooth/hci_ll.c 	skb_queue_purge(&ll->txq);
txq               160 drivers/bluetooth/hci_ll.c 	skb_queue_purge(&ll->txq);
txq               191 drivers/bluetooth/hci_ll.c 		skb_queue_tail(&ll->txq, skb);
txq               327 drivers/bluetooth/hci_ll.c 		skb_queue_tail(&ll->txq, skb);
txq               448 drivers/bluetooth/hci_ll.c 	return skb_dequeue(&ll->txq);
txq                38 drivers/bluetooth/hci_mrvl.c 	struct sk_buff_head txq;
txq                69 drivers/bluetooth/hci_mrvl.c 	skb_queue_head_init(&mrvl->txq);
txq                98 drivers/bluetooth/hci_mrvl.c 	skb_queue_purge(&mrvl->txq);
txq               113 drivers/bluetooth/hci_mrvl.c 	skb_queue_purge(&mrvl->txq);
txq               124 drivers/bluetooth/hci_mrvl.c 	skb = skb_dequeue(&mrvl->txq);
txq               140 drivers/bluetooth/hci_mrvl.c 	skb_queue_tail(&mrvl->txq, skb);
txq               157 drivers/bluetooth/hci_mrvl.c 	skb_queue_tail(&mrvl->txq, skb);
txq               135 drivers/bluetooth/hci_nokia.c 	struct sk_buff_head txq;
txq               482 drivers/bluetooth/hci_nokia.c 	skb_queue_purge(&btdev->txq);
txq               496 drivers/bluetooth/hci_nokia.c 	skb_queue_purge(&btdev->txq);
txq               526 drivers/bluetooth/hci_nokia.c 	skb_queue_tail(&btdev->txq, skb);
txq               648 drivers/bluetooth/hci_nokia.c 	struct sk_buff *result = skb_dequeue(&btdev->txq);
txq               741 drivers/bluetooth/hci_nokia.c 	skb_queue_head_init(&btdev->txq);
txq                85 drivers/bluetooth/hci_qca.c 	struct sk_buff_head txq;
txq               301 drivers/bluetooth/hci_qca.c 	skb_queue_tail(&qca->txq, skb);
txq               485 drivers/bluetooth/hci_qca.c 	skb_queue_head_init(&qca->txq);
txq               598 drivers/bluetooth/hci_qca.c 	skb_queue_purge(&qca->txq);
txq               614 drivers/bluetooth/hci_qca.c 	skb_queue_purge(&qca->txq);
txq               742 drivers/bluetooth/hci_qca.c 			skb_queue_tail(&qca->txq, skb);
txq               786 drivers/bluetooth/hci_qca.c 		skb_queue_tail(&qca->txq, skb);
txq               795 drivers/bluetooth/hci_qca.c 		skb_queue_tail(&qca->txq, skb);
txq               959 drivers/bluetooth/hci_qca.c 	return skb_dequeue(&qca->txq);
txq              1020 drivers/bluetooth/hci_qca.c 	skb_queue_tail(&qca->txq, skb);
txq              1025 drivers/bluetooth/hci_qca.c 	while (!skb_queue_empty(&qca->txq))
txq               585 drivers/crypto/chelsio/chcr_algo.c 	struct sge_uld_txq *txq;
txq               589 drivers/crypto/chelsio/chcr_algo.c 	txq = &txq_info->uldtxq[idx];
txq               590 drivers/crypto/chelsio/chcr_algo.c 	spin_lock(&txq->sendq.lock);
txq               591 drivers/crypto/chelsio/chcr_algo.c 	if (txq->full)
txq               593 drivers/crypto/chelsio/chcr_algo.c 	spin_unlock(&txq->sendq.lock);
txq               569 drivers/crypto/chelsio/chcr_ipsec.c 		netif_tx_stop_queue(q->txq);
txq               658 drivers/crypto/chelsio/chcr_ipsec.c 	netif_tx_stop_queue(q->txq);
txq               190 drivers/crypto/chelsio/chtls/chtls.h 	struct sk_buff_head txq;
txq                65 drivers/crypto/chelsio/chtls/chtls_cm.c 	skb_queue_head_init(&csk->txq);
txq               145 drivers/crypto/chelsio/chtls/chtls_cm.c 	while ((skb = __skb_dequeue(&csk->txq))) {
txq              2133 drivers/crypto/chelsio/chtls/chtls_cm.c 	if (skb_queue_len(&csk->txq) && chtls_push_frames(csk, 0))
txq               313 drivers/crypto/chelsio/chtls/chtls_hw.c 		if (skb_queue_len(&csk->txq))
txq               400 drivers/crypto/chelsio/chtls/chtls_io.c 					 skb_queue_empty(&csk->txq)));
txq               536 drivers/crypto/chelsio/chtls/chtls_io.c 				   skb_queue_empty(&csk->txq)));
txq               615 drivers/crypto/chelsio/chtls/chtls_io.c 	while (csk->wr_credits && (skb = skb_peek(&csk->txq)) &&
txq               617 drivers/crypto/chelsio/chtls/chtls_io.c 		skb_queue_len(&csk->txq) > 1)) {
txq               660 drivers/crypto/chelsio/chtls/chtls_io.c 		__skb_unlink(skb, &csk->txq);
txq               768 drivers/crypto/chelsio/chtls/chtls_io.c 	int qlen = skb_queue_len(&csk->txq);
txq               771 drivers/crypto/chelsio/chtls/chtls_io.c 		struct sk_buff *skb = skb_peek_tail(&csk->txq);
txq               826 drivers/crypto/chelsio/chtls/chtls_io.c 	__skb_queue_tail(&csk->txq, skb);
txq               881 drivers/crypto/chelsio/chtls/chtls_io.c 	if (skb_queue_len(&csk->txq) == 1)
txq              1016 drivers/crypto/chelsio/chtls/chtls_io.c 		skb = skb_peek_tail(&csk->txq);
txq              1186 drivers/crypto/chelsio/chtls/chtls_io.c 		__skb_unlink(skb, &csk->txq);
txq              1224 drivers/crypto/chelsio/chtls/chtls_io.c 		struct sk_buff *skb = skb_peek_tail(&csk->txq);
txq              1435 drivers/crypto/chelsio/chtls/chtls_io.c 		    skb_queue_len(&csk->txq) &&
txq              1741 drivers/crypto/chelsio/chtls/chtls_io.c 		    skb_queue_len(&csk->txq) &&
txq               206 drivers/misc/ti-st/st_core.c 		skb_queue_tail(&st_gdata->txq, waiting_skb);
txq               396 drivers/misc/ti-st/st_core.c 	return skb_dequeue(&st_gdata->txq);
txq               418 drivers/misc/ti-st/st_core.c 		skb_queue_tail(&st_gdata->txq, skb);
txq               770 drivers/misc/ti-st/st_core.c 	skb_queue_purge(&st_gdata->txq);
txq               866 drivers/misc/ti-st/st_core.c 	skb_queue_head_init(&st_gdata->txq);
txq               898 drivers/misc/ti-st/st_core.c 		skb_queue_purge(&st_gdata->txq);
txq              4184 drivers/net/bonding/bond_main.c 	u16 txq = skb_rx_queue_recorded(skb) ? skb_get_rx_queue(skb) : 0;
txq              4189 drivers/net/bonding/bond_main.c 	if (unlikely(txq >= dev->real_num_tx_queues)) {
txq              4191 drivers/net/bonding/bond_main.c 			txq -= dev->real_num_tx_queues;
txq              4192 drivers/net/bonding/bond_main.c 		} while (txq >= dev->real_num_tx_queues);
txq              4194 drivers/net/bonding/bond_main.c 	return txq;
txq               546 drivers/net/ethernet/alacritech/slic.h 	struct slic_tx_queue txq;
txq               114 drivers/net/ethernet/alacritech/slicoss.c static unsigned int slic_get_free_tx_descs(struct slic_tx_queue *txq)
txq               118 drivers/net/ethernet/alacritech/slicoss.c 	return slic_get_free_queue_descs(txq->put_idx, txq->done_idx, txq->len);
txq               345 drivers/net/ethernet/alacritech/slicoss.c 	struct slic_tx_queue *txq = &sdev->txq;
txq               360 drivers/net/ethernet/alacritech/slicoss.c 		txq->done_idx = idx;
txq               361 drivers/net/ethernet/alacritech/slicoss.c 		buff = &txq->txbuffs[idx];
txq               388 drivers/net/ethernet/alacritech/slicoss.c 	    (slic_get_free_tx_descs(txq) >= SLIC_MIN_TX_WAKEUP_DESCS))
txq               839 drivers/net/ethernet/alacritech/slicoss.c 	struct slic_tx_queue *txq = &sdev->txq;
txq               845 drivers/net/ethernet/alacritech/slicoss.c 	txq->len = SLIC_NUM_TX_DESCS;
txq               846 drivers/net/ethernet/alacritech/slicoss.c 	txq->put_idx = 0;
txq               847 drivers/net/ethernet/alacritech/slicoss.c 	txq->done_idx = 0;
txq               849 drivers/net/ethernet/alacritech/slicoss.c 	txq->txbuffs = kcalloc(txq->len, sizeof(*buff), GFP_KERNEL);
txq               850 drivers/net/ethernet/alacritech/slicoss.c 	if (!txq->txbuffs)
txq               853 drivers/net/ethernet/alacritech/slicoss.c 	txq->dma_pool = dma_pool_create("slic_pool", &sdev->pdev->dev,
txq               856 drivers/net/ethernet/alacritech/slicoss.c 	if (!txq->dma_pool) {
txq               862 drivers/net/ethernet/alacritech/slicoss.c 	for (i = 0; i < txq->len; i++) {
txq               863 drivers/net/ethernet/alacritech/slicoss.c 		buff = &txq->txbuffs[i];
txq               864 drivers/net/ethernet/alacritech/slicoss.c 		desc = dma_pool_zalloc(txq->dma_pool, GFP_KERNEL,
txq               884 drivers/net/ethernet/alacritech/slicoss.c 		buff = &txq->txbuffs[i];
txq               885 drivers/net/ethernet/alacritech/slicoss.c 		dma_pool_free(txq->dma_pool, buff->desc, buff->desc_paddr);
txq               887 drivers/net/ethernet/alacritech/slicoss.c 	dma_pool_destroy(txq->dma_pool);
txq               890 drivers/net/ethernet/alacritech/slicoss.c 	kfree(txq->txbuffs);
txq               897 drivers/net/ethernet/alacritech/slicoss.c 	struct slic_tx_queue *txq = &sdev->txq;
txq               901 drivers/net/ethernet/alacritech/slicoss.c 	for (i = 0; i < txq->len; i++) {
txq               902 drivers/net/ethernet/alacritech/slicoss.c 		buff = &txq->txbuffs[i];
txq               903 drivers/net/ethernet/alacritech/slicoss.c 		dma_pool_free(txq->dma_pool, buff->desc, buff->desc_paddr);
txq               912 drivers/net/ethernet/alacritech/slicoss.c 	dma_pool_destroy(txq->dma_pool);
txq               914 drivers/net/ethernet/alacritech/slicoss.c 	kfree(txq->txbuffs);
txq              1411 drivers/net/ethernet/alacritech/slicoss.c 	struct slic_tx_queue *txq = &sdev->txq;
txq              1418 drivers/net/ethernet/alacritech/slicoss.c 	if (unlikely(slic_get_free_tx_descs(txq) < SLIC_MAX_REQ_TX_DESCS)) {
txq              1420 drivers/net/ethernet/alacritech/slicoss.c 			   slic_get_free_tx_descs(txq));
txq              1432 drivers/net/ethernet/alacritech/slicoss.c 	buff = &txq->txbuffs[txq->put_idx];
txq              1443 drivers/net/ethernet/alacritech/slicoss.c 	txq->put_idx = slic_next_queue_idx(txq->put_idx, txq->len);
txq              1451 drivers/net/ethernet/alacritech/slicoss.c 	if (slic_get_free_tx_descs(txq) < SLIC_MAX_REQ_TX_DESCS)
txq               755 drivers/net/ethernet/amazon/ena/ena_netdev.c 	struct netdev_queue *txq;
txq               765 drivers/net/ethernet/amazon/ena/ena_netdev.c 	txq = netdev_get_tx_queue(tx_ring->netdev, tx_ring->qid);
txq               809 drivers/net/ethernet/amazon/ena/ena_netdev.c 	netdev_tx_completed_queue(txq, tx_pkts, tx_bytes);
txq               822 drivers/net/ethernet/amazon/ena/ena_netdev.c 	if (unlikely(netif_tx_queue_stopped(txq) && above_thresh)) {
txq               823 drivers/net/ethernet/amazon/ena/ena_netdev.c 		__netif_tx_lock(txq, smp_processor_id());
txq               827 drivers/net/ethernet/amazon/ena/ena_netdev.c 		if (netif_tx_queue_stopped(txq) && above_thresh &&
txq               829 drivers/net/ethernet/amazon/ena/ena_netdev.c 			netif_tx_wake_queue(txq);
txq               834 drivers/net/ethernet/amazon/ena/ena_netdev.c 		__netif_tx_unlock(txq);
txq              2252 drivers/net/ethernet/amazon/ena/ena_netdev.c 	struct netdev_queue *txq;
txq              2261 drivers/net/ethernet/amazon/ena/ena_netdev.c 	txq = netdev_get_tx_queue(dev, qid);
txq              2320 drivers/net/ethernet/amazon/ena/ena_netdev.c 	netdev_tx_sent_queue(txq, skb->len);
txq              2343 drivers/net/ethernet/amazon/ena/ena_netdev.c 		netif_tx_stop_queue(txq);
txq              2360 drivers/net/ethernet/amazon/ena/ena_netdev.c 			netif_tx_wake_queue(txq);
txq              2367 drivers/net/ethernet/amazon/ena/ena_netdev.c 	if (netif_xmit_stopped(txq) || !netdev_xmit_more()) {
txq              1437 drivers/net/ethernet/amd/xgbe/xgbe-drv.c 	struct netdev_queue *txq;
txq              1468 drivers/net/ethernet/amd/xgbe/xgbe-drv.c 		txq = netdev_get_tx_queue(netdev, channel->queue_index);
txq              1469 drivers/net/ethernet/amd/xgbe/xgbe-drv.c 		netdev_tx_reset_queue(txq);
txq              2019 drivers/net/ethernet/amd/xgbe/xgbe-drv.c 	struct netdev_queue *txq;
txq              2025 drivers/net/ethernet/amd/xgbe/xgbe-drv.c 	txq = netdev_get_tx_queue(netdev, channel->queue_index);
txq              2064 drivers/net/ethernet/amd/xgbe/xgbe-drv.c 	netdev_tx_sent_queue(txq, packet->tx_bytes);
txq              2621 drivers/net/ethernet/amd/xgbe/xgbe-drv.c 	struct netdev_queue *txq;
txq              2637 drivers/net/ethernet/amd/xgbe/xgbe-drv.c 	txq = netdev_get_tx_queue(netdev, channel->queue_index);
txq              2670 drivers/net/ethernet/amd/xgbe/xgbe-drv.c 	netdev_tx_completed_queue(txq, tx_packets, tx_bytes);
txq              2675 drivers/net/ethernet/amd/xgbe/xgbe-drv.c 		netif_tx_wake_queue(txq);
txq               865 drivers/net/ethernet/apm/xgene/xgene_enet_main.c 	struct netdev_queue *txq;
txq               871 drivers/net/ethernet/apm/xgene/xgene_enet_main.c 		txq = netdev_get_tx_queue(ndev, i);
txq               872 drivers/net/ethernet/apm/xgene/xgene_enet_main.c 		txq->trans_start = jiffies;
txq               873 drivers/net/ethernet/apm/xgene/xgene_enet_main.c 		netif_tx_start_queue(txq);
txq                97 drivers/net/ethernet/atheros/alx/alx.h 	struct alx_tx_queue	*txq;
txq               378 drivers/net/ethernet/atheros/alx/hw.c 	u32 rxq, txq, val;
txq               383 drivers/net/ethernet/atheros/alx/hw.c 	txq = alx_read_mem32(hw, ALX_TXQ0);
txq               384 drivers/net/ethernet/atheros/alx/hw.c 	alx_write_mem32(hw, ALX_TXQ0, txq & ~ALX_TXQ0_EN);
txq               609 drivers/net/ethernet/atheros/alx/hw.c 	u32 mac, txq, rxq;
txq               613 drivers/net/ethernet/atheros/alx/hw.c 	txq = alx_read_mem32(hw, ALX_TXQ0);
txq               614 drivers/net/ethernet/atheros/alx/hw.c 	alx_write_mem32(hw, ALX_TXQ0, txq | ALX_TXQ0_EN);
txq                54 drivers/net/ethernet/atheros/alx/main.c static void alx_free_txbuf(struct alx_tx_queue *txq, int entry)
txq                56 drivers/net/ethernet/atheros/alx/main.c 	struct alx_buffer *txb = &txq->bufs[entry];
txq                59 drivers/net/ethernet/atheros/alx/main.c 		dma_unmap_single(txq->dev,
txq               150 drivers/net/ethernet/atheros/alx/main.c 	return alx->qnapi[r_idx]->txq;
txq               153 drivers/net/ethernet/atheros/alx/main.c static struct netdev_queue *alx_get_tx_queue(const struct alx_tx_queue *txq)
txq               155 drivers/net/ethernet/atheros/alx/main.c 	return netdev_get_tx_queue(txq->netdev, txq->queue_idx);
txq               158 drivers/net/ethernet/atheros/alx/main.c static inline int alx_tpd_avail(struct alx_tx_queue *txq)
txq               160 drivers/net/ethernet/atheros/alx/main.c 	if (txq->write_idx >= txq->read_idx)
txq               161 drivers/net/ethernet/atheros/alx/main.c 		return txq->count + txq->read_idx - txq->write_idx - 1;
txq               162 drivers/net/ethernet/atheros/alx/main.c 	return txq->read_idx - txq->write_idx - 1;
txq               165 drivers/net/ethernet/atheros/alx/main.c static bool alx_clean_tx_irq(struct alx_tx_queue *txq)
txq               173 drivers/net/ethernet/atheros/alx/main.c 	alx = netdev_priv(txq->netdev);
txq               174 drivers/net/ethernet/atheros/alx/main.c 	tx_queue = alx_get_tx_queue(txq);
txq               176 drivers/net/ethernet/atheros/alx/main.c 	sw_read_idx = txq->read_idx;
txq               177 drivers/net/ethernet/atheros/alx/main.c 	hw_read_idx = alx_read_mem16(&alx->hw, txq->c_reg);
txq               183 drivers/net/ethernet/atheros/alx/main.c 			skb = txq->bufs[sw_read_idx].skb;
txq               190 drivers/net/ethernet/atheros/alx/main.c 			alx_free_txbuf(txq, sw_read_idx);
txq               192 drivers/net/ethernet/atheros/alx/main.c 			if (++sw_read_idx == txq->count)
txq               195 drivers/net/ethernet/atheros/alx/main.c 		txq->read_idx = sw_read_idx;
txq               201 drivers/net/ethernet/atheros/alx/main.c 	    alx_tpd_avail(txq) > txq->count / 4)
txq               306 drivers/net/ethernet/atheros/alx/main.c 	if (np->txq)
txq               307 drivers/net/ethernet/atheros/alx/main.c 		tx_complete = alx_clean_tx_irq(np->txq);
txq               460 drivers/net/ethernet/atheros/alx/main.c 		if (np->txq) {
txq               461 drivers/net/ethernet/atheros/alx/main.c 			np->txq->read_idx = 0;
txq               462 drivers/net/ethernet/atheros/alx/main.c 			np->txq->write_idx = 0;
txq               464 drivers/net/ethernet/atheros/alx/main.c 					txring_header_reg[np->txq->queue_idx],
txq               465 drivers/net/ethernet/atheros/alx/main.c 					np->txq->tpd_dma);
txq               489 drivers/net/ethernet/atheros/alx/main.c static void alx_free_txring_buf(struct alx_tx_queue *txq)
txq               493 drivers/net/ethernet/atheros/alx/main.c 	if (!txq->bufs)
txq               496 drivers/net/ethernet/atheros/alx/main.c 	for (i = 0; i < txq->count; i++)
txq               497 drivers/net/ethernet/atheros/alx/main.c 		alx_free_txbuf(txq, i);
txq               499 drivers/net/ethernet/atheros/alx/main.c 	memset(txq->bufs, 0, txq->count * sizeof(struct alx_buffer));
txq               500 drivers/net/ethernet/atheros/alx/main.c 	memset(txq->tpd, 0, txq->count * sizeof(struct alx_txd));
txq               501 drivers/net/ethernet/atheros/alx/main.c 	txq->write_idx = 0;
txq               502 drivers/net/ethernet/atheros/alx/main.c 	txq->read_idx = 0;
txq               504 drivers/net/ethernet/atheros/alx/main.c 	netdev_tx_reset_queue(alx_get_tx_queue(txq));
txq               539 drivers/net/ethernet/atheros/alx/main.c 		if (alx->qnapi[i] && alx->qnapi[i]->txq)
txq               540 drivers/net/ethernet/atheros/alx/main.c 			alx_free_txring_buf(alx->qnapi[i]->txq);
txq               617 drivers/net/ethernet/atheros/alx/main.c static int alx_alloc_tx_ring(struct alx_priv *alx, struct alx_tx_queue *txq,
txq               620 drivers/net/ethernet/atheros/alx/main.c 	txq->bufs = kcalloc(txq->count, sizeof(struct alx_buffer), GFP_KERNEL);
txq               621 drivers/net/ethernet/atheros/alx/main.c 	if (!txq->bufs)
txq               624 drivers/net/ethernet/atheros/alx/main.c 	txq->tpd = alx->descmem.virt + offset;
txq               625 drivers/net/ethernet/atheros/alx/main.c 	txq->tpd_dma = alx->descmem.dma + offset;
txq               626 drivers/net/ethernet/atheros/alx/main.c 	offset += sizeof(struct alx_txd) * txq->count;
txq               674 drivers/net/ethernet/atheros/alx/main.c 		offset = alx_alloc_tx_ring(alx, alx->qnapi[i]->txq, offset);
txq               697 drivers/net/ethernet/atheros/alx/main.c 		if (alx->qnapi[i] && alx->qnapi[i]->txq)
txq               698 drivers/net/ethernet/atheros/alx/main.c 			kfree(alx->qnapi[i]->txq->bufs);
txq               721 drivers/net/ethernet/atheros/alx/main.c 		kfree(np->txq);
txq               743 drivers/net/ethernet/atheros/alx/main.c 	struct alx_tx_queue *txq;
txq               762 drivers/net/ethernet/atheros/alx/main.c 		txq = kzalloc(sizeof(*txq), GFP_KERNEL);
txq               763 drivers/net/ethernet/atheros/alx/main.c 		if (!txq)
txq               766 drivers/net/ethernet/atheros/alx/main.c 		np->txq = txq;
txq               767 drivers/net/ethernet/atheros/alx/main.c 		txq->p_reg = tx_pidx_reg[i];
txq               768 drivers/net/ethernet/atheros/alx/main.c 		txq->c_reg = tx_cidx_reg[i];
txq               769 drivers/net/ethernet/atheros/alx/main.c 		txq->queue_idx = i;
txq               770 drivers/net/ethernet/atheros/alx/main.c 		txq->count = alx->tx_ringsz;
txq               771 drivers/net/ethernet/atheros/alx/main.c 		txq->netdev = alx->dev;
txq               772 drivers/net/ethernet/atheros/alx/main.c 		txq->dev = &alx->hw.pdev->dev;
txq               868 drivers/net/ethernet/atheros/alx/main.c 		if (np->txq && np->rxq)
txq               870 drivers/net/ethernet/atheros/alx/main.c 				np->txq->queue_idx);
txq               871 drivers/net/ethernet/atheros/alx/main.c 		else if (np->txq)
txq               873 drivers/net/ethernet/atheros/alx/main.c 				np->txq->queue_idx);
txq              1436 drivers/net/ethernet/atheros/alx/main.c static int alx_map_tx_skb(struct alx_tx_queue *txq, struct sk_buff *skb)
txq              1440 drivers/net/ethernet/atheros/alx/main.c 	int maplen, f, first_idx = txq->write_idx;
txq              1442 drivers/net/ethernet/atheros/alx/main.c 	first_tpd = &txq->tpd[txq->write_idx];
txq              1446 drivers/net/ethernet/atheros/alx/main.c 		if (++txq->write_idx == txq->count)
txq              1447 drivers/net/ethernet/atheros/alx/main.c 			txq->write_idx = 0;
txq              1449 drivers/net/ethernet/atheros/alx/main.c 		tpd = &txq->tpd[txq->write_idx];
txq              1456 drivers/net/ethernet/atheros/alx/main.c 	dma = dma_map_single(txq->dev, skb->data, maplen,
txq              1458 drivers/net/ethernet/atheros/alx/main.c 	if (dma_mapping_error(txq->dev, dma))
txq              1461 drivers/net/ethernet/atheros/alx/main.c 	dma_unmap_len_set(&txq->bufs[txq->write_idx], size, maplen);
txq              1462 drivers/net/ethernet/atheros/alx/main.c 	dma_unmap_addr_set(&txq->bufs[txq->write_idx], dma, dma);
txq              1470 drivers/net/ethernet/atheros/alx/main.c 		if (++txq->write_idx == txq->count)
txq              1471 drivers/net/ethernet/atheros/alx/main.c 			txq->write_idx = 0;
txq              1472 drivers/net/ethernet/atheros/alx/main.c 		tpd = &txq->tpd[txq->write_idx];
txq              1477 drivers/net/ethernet/atheros/alx/main.c 		dma = skb_frag_dma_map(txq->dev, frag, 0,
txq              1479 drivers/net/ethernet/atheros/alx/main.c 		if (dma_mapping_error(txq->dev, dma))
txq              1481 drivers/net/ethernet/atheros/alx/main.c 		dma_unmap_len_set(&txq->bufs[txq->write_idx], size, maplen);
txq              1482 drivers/net/ethernet/atheros/alx/main.c 		dma_unmap_addr_set(&txq->bufs[txq->write_idx], dma, dma);
txq              1490 drivers/net/ethernet/atheros/alx/main.c 	txq->bufs[txq->write_idx].skb = skb;
txq              1492 drivers/net/ethernet/atheros/alx/main.c 	if (++txq->write_idx == txq->count)
txq              1493 drivers/net/ethernet/atheros/alx/main.c 		txq->write_idx = 0;
txq              1499 drivers/net/ethernet/atheros/alx/main.c 	while (f != txq->write_idx) {
txq              1500 drivers/net/ethernet/atheros/alx/main.c 		alx_free_txbuf(txq, f);
txq              1501 drivers/net/ethernet/atheros/alx/main.c 		if (++f == txq->count)
txq              1508 drivers/net/ethernet/atheros/alx/main.c 				       struct alx_tx_queue *txq)
txq              1514 drivers/net/ethernet/atheros/alx/main.c 	alx = netdev_priv(txq->netdev);
txq              1516 drivers/net/ethernet/atheros/alx/main.c 	if (alx_tpd_avail(txq) < alx_tpd_req(skb)) {
txq              1517 drivers/net/ethernet/atheros/alx/main.c 		netif_tx_stop_queue(alx_get_tx_queue(txq));
txq              1521 drivers/net/ethernet/atheros/alx/main.c 	first = &txq->tpd[txq->write_idx];
txq              1530 drivers/net/ethernet/atheros/alx/main.c 	if (alx_map_tx_skb(txq, skb) < 0)
txq              1533 drivers/net/ethernet/atheros/alx/main.c 	netdev_tx_sent_queue(alx_get_tx_queue(txq), skb->len);
txq              1537 drivers/net/ethernet/atheros/alx/main.c 	alx_write_mem16(&alx->hw, txq->p_reg, txq->write_idx);
txq              1539 drivers/net/ethernet/atheros/alx/main.c 	if (alx_tpd_avail(txq) < txq->count / 8)
txq              1540 drivers/net/ethernet/atheros/alx/main.c 		netif_tx_stop_queue(alx_get_tx_queue(txq));
txq              1190 drivers/net/ethernet/atheros/atl1c/atl1c_main.c 	u32 mac, txq, rxq;
txq              1196 drivers/net/ethernet/atheros/atl1c/atl1c_main.c 	AT_READ_REG(hw, REG_TXQ_CTRL, &txq);
txq              1200 drivers/net/ethernet/atheros/atl1c/atl1c_main.c 	txq |= TXQ_CTRL_EN;
txq              1214 drivers/net/ethernet/atheros/atl1c/atl1c_main.c 	AT_WRITE_REG(hw, REG_TXQ_CTRL, txq);
txq               933 drivers/net/ethernet/broadcom/bcmsysport.c 	struct netdev_queue *txq;
txq               937 drivers/net/ethernet/broadcom/bcmsysport.c 	txq = netdev_get_tx_queue(priv->netdev, ring->index);
txq               942 drivers/net/ethernet/broadcom/bcmsysport.c 		netif_tx_wake_queue(txq);
txq              1281 drivers/net/ethernet/broadcom/bcmsysport.c 	struct netdev_queue *txq;
txq              1290 drivers/net/ethernet/broadcom/bcmsysport.c 	txq = netdev_get_tx_queue(dev, queue);
txq              1296 drivers/net/ethernet/broadcom/bcmsysport.c 		netif_tx_stop_queue(txq);
txq              1347 drivers/net/ethernet/broadcom/bcmsysport.c 		netif_tx_stop_queue(txq);
txq              2424 drivers/net/ethernet/broadcom/bcmsysport.c 	u32 txq, rxq;
txq              2436 drivers/net/ethernet/broadcom/bcmsysport.c 	if (of_property_read_u32(dn, "systemport,num-txq", &txq))
txq              2437 drivers/net/ethernet/broadcom/bcmsysport.c 		txq = TDMA_NUM_RINGS;
txq              2442 drivers/net/ethernet/broadcom/bcmsysport.c 	if (!txq || txq > TDMA_NUM_RINGS)
txq              2445 drivers/net/ethernet/broadcom/bcmsysport.c 	dev = alloc_etherdev_mqs(sizeof(*priv), txq, rxq);
txq              2453 drivers/net/ethernet/broadcom/bcmsysport.c 	priv->tx_rings = devm_kcalloc(&pdev->dev, txq,
txq              2559 drivers/net/ethernet/broadcom/bcmsysport.c 		 priv->irq0, priv->irq1, txq, rxq);
txq              2854 drivers/net/ethernet/broadcom/bnx2.c 	struct netdev_queue *txq;
txq              2857 drivers/net/ethernet/broadcom/bnx2.c 	txq = netdev_get_tx_queue(bp->dev, index);
txq              2919 drivers/net/ethernet/broadcom/bnx2.c 	netdev_tx_completed_queue(txq, tx_pkt, tx_bytes);
txq              2930 drivers/net/ethernet/broadcom/bnx2.c 	if (unlikely(netif_tx_queue_stopped(txq)) &&
txq              2932 drivers/net/ethernet/broadcom/bnx2.c 		__netif_tx_lock(txq, smp_processor_id());
txq              2933 drivers/net/ethernet/broadcom/bnx2.c 		if ((netif_tx_queue_stopped(txq)) &&
txq              2935 drivers/net/ethernet/broadcom/bnx2.c 			netif_tx_wake_queue(txq);
txq              2936 drivers/net/ethernet/broadcom/bnx2.c 		__netif_tx_unlock(txq);
txq              6606 drivers/net/ethernet/broadcom/bnx2.c 	struct netdev_queue *txq;
txq              6612 drivers/net/ethernet/broadcom/bnx2.c 	txq = netdev_get_tx_queue(dev, i);
txq              6616 drivers/net/ethernet/broadcom/bnx2.c 		netif_tx_stop_queue(txq);
txq              6716 drivers/net/ethernet/broadcom/bnx2.c 	netdev_tx_sent_queue(txq, skb->len);
txq              6727 drivers/net/ethernet/broadcom/bnx2.c 		netif_tx_stop_queue(txq);
txq              6736 drivers/net/ethernet/broadcom/bnx2.c 			netif_tx_wake_queue(txq);
txq               275 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c 	struct netdev_queue *txq;
txq               284 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c 	txq = netdev_get_tx_queue(bp->dev, txdata->txq_index);
txq               306 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c 	netdev_tx_completed_queue(txq, pkts_compl, bytes_compl);
txq               322 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c 	if (unlikely(netif_tx_queue_stopped(txq))) {
txq               333 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c 		__netif_tx_lock(txq, smp_processor_id());
txq               335 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c 		if ((netif_tx_queue_stopped(txq)) &&
txq               338 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c 			netif_tx_wake_queue(txq);
txq               340 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c 		__netif_tx_unlock(txq);
txq              3737 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c 	struct netdev_queue *txq;
txq              3762 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c 	txq = netdev_get_tx_queue(dev, txq_index);
txq              3790 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c 		netif_tx_stop_queue(txq);
txq              4154 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c 	netdev_tx_sent_queue(txq, skb->len);
txq              4177 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c 		netif_tx_stop_queue(txq);
txq              4186 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c 			netif_tx_wake_queue(txq);
txq              2512 drivers/net/ethernet/broadcom/bnx2x/bnx2x_ethtool.c 	struct netdev_queue *txq = netdev_get_tx_queue(bp->dev,
txq              2579 drivers/net/ethernet/broadcom/bnx2x/bnx2x_ethtool.c 	netdev_tx_sent_queue(txq, skb->len);
txq               654 drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.c 	req->txq.txq_addr = fp->txdata_ptr[FIRST_TX_COS_INDEX]->tx_desc_mapping;
txq               655 drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.c 	req->txq.vf_sb = fp_idx;
txq               656 drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.c 	req->txq.sb_index = HC_INDEX_ETH_TX_CQ_CONS_COS0;
txq               657 drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.c 	req->txq.hc_rate = bp->tx_ticks ? 1000000/bp->tx_ticks : 0;
txq               658 drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.c 	req->txq.flags = flags;
txq               659 drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.c 	req->txq.traffic_type = LLFC_TRAFFIC_TYPE_NW;
txq              1562 drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.c 			q->sb_idx = setup_q->txq.vf_sb;
txq              1565 drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.c 			init_p->tx.hc_rate = setup_q->txq.hc_rate;
txq              1566 drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.c 			init_p->tx.sb_cq_index = setup_q->txq.sb_index;
txq              1568 drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.c 			bnx2x_vf_mbx_set_q_flags(bp, setup_q->txq.flags,
txq              1572 drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.c 			bnx2x_vf_mbx_set_q_flags(bp, setup_q->txq.flags,
txq              1578 drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.c 			txq_params->dscr_map = setup_q->txq.txq_addr;
txq              1579 drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.c 			txq_params->sb_cq_index = setup_q->txq.sb_index;
txq              1580 drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.c 			txq_params->traffic_type = setup_q->txq.traffic_type;
txq               288 drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.h 	} txq;
txq               350 drivers/net/ethernet/broadcom/bnxt/bnxt.c 	struct netdev_queue *txq;
txq               366 drivers/net/ethernet/broadcom/bnxt/bnxt.c 	txq = netdev_get_tx_queue(dev, i);
txq               372 drivers/net/ethernet/broadcom/bnxt/bnxt.c 		netif_tx_stop_queue(txq);
txq               457 drivers/net/ethernet/broadcom/bnxt/bnxt.c 		netdev_tx_sent_queue(txq, skb->len);
txq               565 drivers/net/ethernet/broadcom/bnxt/bnxt.c 	netdev_tx_sent_queue(txq, skb->len);
txq               573 drivers/net/ethernet/broadcom/bnxt/bnxt.c 	if (!netdev_xmit_more() || netif_xmit_stopped(txq))
txq               582 drivers/net/ethernet/broadcom/bnxt/bnxt.c 		netif_tx_stop_queue(txq);
txq               591 drivers/net/ethernet/broadcom/bnxt/bnxt.c 			netif_tx_wake_queue(txq);
txq               622 drivers/net/ethernet/broadcom/bnxt/bnxt.c 	struct netdev_queue *txq = netdev_get_tx_queue(bp->dev, txr->txq_index);
txq               664 drivers/net/ethernet/broadcom/bnxt/bnxt.c 	netdev_tx_completed_queue(txq, nr_pkts, tx_bytes);
txq               674 drivers/net/ethernet/broadcom/bnxt/bnxt.c 	if (unlikely(netif_tx_queue_stopped(txq)) &&
txq               676 drivers/net/ethernet/broadcom/bnxt/bnxt.c 		__netif_tx_lock(txq, smp_processor_id());
txq               677 drivers/net/ethernet/broadcom/bnxt/bnxt.c 		if (netif_tx_queue_stopped(txq) &&
txq               680 drivers/net/ethernet/broadcom/bnxt/bnxt.c 			netif_tx_wake_queue(txq);
txq               681 drivers/net/ethernet/broadcom/bnxt/bnxt.c 		__netif_tx_unlock(txq);
txq              1449 drivers/net/ethernet/broadcom/genet/bcmgenet.c 	struct netdev_queue *txq;
txq              1454 drivers/net/ethernet/broadcom/genet/bcmgenet.c 		txq = netdev_get_tx_queue(ring->priv->dev, ring->queue);
txq              1455 drivers/net/ethernet/broadcom/genet/bcmgenet.c 		netif_tx_wake_queue(txq);
txq              1552 drivers/net/ethernet/broadcom/genet/bcmgenet.c 	struct netdev_queue *txq;
txq              1575 drivers/net/ethernet/broadcom/genet/bcmgenet.c 	txq = netdev_get_tx_queue(dev, ring->queue);
txq              1581 drivers/net/ethernet/broadcom/genet/bcmgenet.c 		if (!netif_tx_queue_stopped(txq)) {
txq              1582 drivers/net/ethernet/broadcom/genet/bcmgenet.c 			netif_tx_stop_queue(txq);
txq              1663 drivers/net/ethernet/broadcom/genet/bcmgenet.c 	netdev_tx_sent_queue(txq, GENET_CB(skb)->bytes_sent);
txq              1666 drivers/net/ethernet/broadcom/genet/bcmgenet.c 		netif_tx_stop_queue(txq);
txq              1668 drivers/net/ethernet/broadcom/genet/bcmgenet.c 	if (!netdev_xmit_more() || netif_xmit_stopped(txq))
txq              2521 drivers/net/ethernet/broadcom/genet/bcmgenet.c 	struct netdev_queue *txq;
txq              2532 drivers/net/ethernet/broadcom/genet/bcmgenet.c 		txq = netdev_get_tx_queue(priv->dev, priv->tx_rings[i].queue);
txq              2533 drivers/net/ethernet/broadcom/genet/bcmgenet.c 		netdev_tx_reset_queue(txq);
txq              2536 drivers/net/ethernet/broadcom/genet/bcmgenet.c 	txq = netdev_get_tx_queue(priv->dev, priv->tx_rings[DESC_INDEX].queue);
txq              2537 drivers/net/ethernet/broadcom/genet/bcmgenet.c 	netdev_tx_reset_queue(txq);
txq              3019 drivers/net/ethernet/broadcom/genet/bcmgenet.c 	struct netdev_queue *txq;
txq              3026 drivers/net/ethernet/broadcom/genet/bcmgenet.c 	txq = netdev_get_tx_queue(priv->dev, ring->queue);
txq              3038 drivers/net/ethernet/broadcom/genet/bcmgenet.c 	txq_stopped = netif_tx_queue_stopped(txq);
txq              6550 drivers/net/ethernet/broadcom/tg3.c 	struct netdev_queue *txq;
txq              6557 drivers/net/ethernet/broadcom/tg3.c 	txq = netdev_get_tx_queue(tp->dev, index);
txq              6624 drivers/net/ethernet/broadcom/tg3.c 	netdev_tx_completed_queue(txq, pkts_compl, bytes_compl);
txq              6635 drivers/net/ethernet/broadcom/tg3.c 	if (unlikely(netif_tx_queue_stopped(txq) &&
txq              6637 drivers/net/ethernet/broadcom/tg3.c 		__netif_tx_lock(txq, smp_processor_id());
txq              6638 drivers/net/ethernet/broadcom/tg3.c 		if (netif_tx_queue_stopped(txq) &&
txq              6640 drivers/net/ethernet/broadcom/tg3.c 			netif_tx_wake_queue(txq);
txq              6641 drivers/net/ethernet/broadcom/tg3.c 		__netif_tx_unlock(txq);
txq              7875 drivers/net/ethernet/broadcom/tg3.c 		       struct netdev_queue *txq, struct sk_buff *skb)
txq              7882 drivers/net/ethernet/broadcom/tg3.c 		netif_tx_stop_queue(txq);
txq              7893 drivers/net/ethernet/broadcom/tg3.c 		netif_tx_wake_queue(txq);
txq              7923 drivers/net/ethernet/broadcom/tg3.c 	struct netdev_queue *txq;
txq              7930 drivers/net/ethernet/broadcom/tg3.c 	txq = netdev_get_tx_queue(dev, skb_get_queue_mapping(skb));
txq              7943 drivers/net/ethernet/broadcom/tg3.c 		if (!netif_tx_queue_stopped(txq)) {
txq              7944 drivers/net/ethernet/broadcom/tg3.c 			netif_tx_stop_queue(txq);
txq              7974 drivers/net/ethernet/broadcom/tg3.c 				return tg3_tso_bug(tp, tnapi, txq, skb);
txq              7982 drivers/net/ethernet/broadcom/tg3.c 					return tg3_tso_bug(tp, tnapi, txq, skb);
txq              8125 drivers/net/ethernet/broadcom/tg3.c 			return tg3_tso_bug(tp, tnapi, txq, skb);
txq              8139 drivers/net/ethernet/broadcom/tg3.c 	netdev_tx_sent_queue(txq, skb->len);
txq              8146 drivers/net/ethernet/broadcom/tg3.c 		netif_tx_stop_queue(txq);
txq              8155 drivers/net/ethernet/broadcom/tg3.c 			netif_tx_wake_queue(txq);
txq              8158 drivers/net/ethernet/broadcom/tg3.c 	if (!netdev_xmit_more() || netif_xmit_stopped(txq)) {
txq              2870 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	struct bna_txq *txq;
txq              2873 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	list_for_each_entry(txq, &tx->txq_q, qe) {
txq              2874 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		txq->tcb->priority = txq->priority;
txq              2876 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		bna_ib_start(tx->bna, &txq->ib, is_regular);
txq              3089 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	struct bna_txq *txq = NULL;
txq              3099 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		txq = txq ? list_next_entry(txq, qe)
txq              3101 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		bfi_enet_datapath_q_init(&cfg_req->q_cfg[i].q.q, &txq->qpt);
txq              3102 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		cfg_req->q_cfg[i].q.priority = txq->priority;
txq              3105 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 			txq->ib.ib_seg_host_addr.lsb;
txq              3107 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 			txq->ib.ib_seg_host_addr.msb;
txq              3109 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 			htons((u16)txq->ib.intr_vector);
txq              3116 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	cfg_req->ib_cfg.msix = (txq->ib.intr_type == BNA_INTR_T_MSIX)
txq              3119 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 			htonl((u32)txq->ib.coalescing_timeo);
txq              3121 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 			htonl((u32)txq->ib.interpkt_timeo);
txq              3122 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	cfg_req->ib_cfg.inter_pkt_count = (u8)txq->ib.interpkt_count;
txq              3151 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	struct bna_txq *txq;
txq              3154 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	list_for_each_entry(txq, &tx->txq_q, qe)
txq              3155 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		bna_ib_stop(tx->bna, &txq->ib);
txq              3161 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_txq_qpt_setup(struct bna_txq *txq, int page_count, int page_size,
txq              3171 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	txq->qpt.hw_qpt_ptr.lsb = qpt_mem->dma.lsb;
txq              3172 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	txq->qpt.hw_qpt_ptr.msb = qpt_mem->dma.msb;
txq              3173 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	txq->qpt.kv_qpt_ptr = qpt_mem->kva;
txq              3174 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	txq->qpt.page_count = page_count;
txq              3175 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	txq->qpt.page_size = page_size;
txq              3177 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	txq->tcb->sw_qpt = (void **) swqpt_mem->kva;
txq              3178 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	txq->tcb->sw_q = page_mem->kva;
txq              3184 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		txq->tcb->sw_qpt[i] = kva;
txq              3188 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		((struct bna_dma_addr *)txq->qpt.kv_qpt_ptr)[i].lsb =
txq              3190 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		((struct bna_dma_addr *)txq->qpt.kv_qpt_ptr)[i].msb =
txq              3217 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	struct bna_txq *txq;
txq              3221 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		txq = list_first_entry(&tx->txq_q, struct bna_txq, qe);
txq              3222 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		txq->tcb = NULL;
txq              3223 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		txq->tx = NULL;
txq              3224 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		list_move_tail(&txq->qe, &tx_mod->txq_free_q);
txq              3273 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	struct bna_txq *txq = NULL;
txq              3281 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	for (i = 0, txq = list_first_entry(&tx->txq_q, struct bna_txq, qe);
txq              3282 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	     i < tx->num_txq; i++, txq = list_next_entry(txq, qe)) {
txq              3284 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		txq->tcb->i_dbell->doorbell_addr =
txq              3287 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		txq->tcb->q_dbell =
txq              3290 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		txq->hw_id = cfg_rsp->q_handles[i].hw_qid;
txq              3293 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		(*txq->tcb->hw_consumer_index) = 0;
txq              3294 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		txq->tcb->producer_index = txq->tcb->consumer_index = 0;
txq              3371 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	struct bna_txq *txq;
txq              3401 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		txq = list_first_entry(&tx_mod->txq_free_q, struct bna_txq, qe);
txq              3402 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		list_move_tail(&txq->qe, &tx->txq_q);
txq              3403 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		txq->tx = tx;
txq              3441 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	list_for_each_entry(txq, &tx->txq_q, qe) {
txq              3442 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		txq->tcb = (struct bna_tcb *)
txq              3444 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		txq->tx_packets = 0;
txq              3445 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		txq->tx_bytes = 0;
txq              3448 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		txq->ib.ib_seg_host_addr.lsb =
txq              3450 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		txq->ib.ib_seg_host_addr.msb =
txq              3452 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		txq->ib.ib_seg_host_addr_kva =
txq              3454 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		txq->ib.intr_type = intr_info->intr_type;
txq              3455 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		txq->ib.intr_vector = (intr_info->num == 1) ?
txq              3459 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 			txq->ib.intr_vector = BIT(txq->ib.intr_vector);
txq              3460 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		txq->ib.coalescing_timeo = tx_cfg->coalescing_timeo;
txq              3461 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		txq->ib.interpkt_timeo = BFI_TX_INTERPKT_TIMEO;
txq              3462 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		txq->ib.interpkt_count = BFI_TX_INTERPKT_COUNT;
txq              3466 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		txq->tcb->q_depth = tx_cfg->txq_depth;
txq              3467 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		txq->tcb->unmap_q = (void *)
txq              3469 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		txq->tcb->hw_consumer_index =
txq              3470 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 			(u32 *)txq->ib.ib_seg_host_addr_kva;
txq              3471 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		txq->tcb->i_dbell = &txq->ib.door_bell;
txq              3472 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		txq->tcb->intr_type = txq->ib.intr_type;
txq              3473 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		txq->tcb->intr_vector = txq->ib.intr_vector;
txq              3474 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		txq->tcb->txq = txq;
txq              3475 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		txq->tcb->bnad = bnad;
txq              3476 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		txq->tcb->id = i;
txq              3479 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		bna_txq_qpt_setup(txq, page_count, PAGE_SIZE,
txq              3487 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 			(tx->tcb_setup_cbfn)(bna->bnad, txq->tcb);
txq              3490 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 			txq->priority = txq->tcb->id;
txq              3492 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 			txq->priority = tx_mod->default_prio;
txq              3513 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	struct bna_txq *txq;
txq              3515 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	list_for_each_entry(txq, &tx->txq_q, qe)
txq              3517 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 			(tx->tcb_destroy_cbfn)(tx->bna->bnad, txq->tcb);
txq              3587 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	tx_mod->txq = (struct bna_txq *)
txq              3598 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		list_add_tail(&tx_mod->txq[i].qe, &tx_mod->txq_free_q);
txq              3663 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	struct bna_txq *txq;
txq              3665 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	list_for_each_entry(txq, &tx->txq_q, qe)
txq              3666 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		bna_ib_coalescing_timeo_set(&txq->ib, coalescing_timeo);
txq               426 drivers/net/ethernet/brocade/bna/bna_types.h 	struct bna_txq *txq;
txq               515 drivers/net/ethernet/brocade/bna/bna_types.h 	struct bna_txq *txq;		/* BFI_MAX_TXQ entries */
txq               197 drivers/net/ethernet/brocade/bna/bnad.c 	tcb->txq->tx_packets += sent_packets;
txq               198 drivers/net/ethernet/brocade/bna/bnad.c 	tcb->txq->tx_bytes += sent_bytes;
txq              1002 drivers/net/ethernet/brocade/bna/bnad.c 			(struct bnad_tx_info *)tcb->txq->tx->priv;
txq              1012 drivers/net/ethernet/brocade/bna/bnad.c 			(struct bnad_tx_info *)tcb->txq->tx->priv;
txq              2414 drivers/net/ethernet/brocade/bna/bnad.c 				bnad->tx_info[i].tcb[j]->txq->tx_packets;
txq              2416 drivers/net/ethernet/brocade/bna/bnad.c 					bnad->tx_info[i].tcb[j]->txq->tx_bytes;
txq               827 drivers/net/ethernet/brocade/bna/bnad_ethtool.c 				bnad->tx_info[i].tcb[j]->txq) {
txq               829 drivers/net/ethernet/brocade/bna/bnad_ethtool.c 				buf[bi++] = tcb->txq->tx_packets;
txq               830 drivers/net/ethernet/brocade/bna/bnad_ethtool.c 				buf[bi++] = tcb->txq->tx_bytes;
txq               182 drivers/net/ethernet/cavium/liquidio/lio_core.c void octeon_report_tx_completion_to_bql(void *txq, unsigned int pkts_compl,
txq               185 drivers/net/ethernet/cavium/liquidio/lio_core.c 	struct netdev_queue *netdev_queue = txq;
txq               224 drivers/net/ethernet/cavium/liquidio/lio_core.c 	struct netdev_queue *txq;
txq               243 drivers/net/ethernet/cavium/liquidio/lio_core.c 	txq = netdev_get_tx_queue(skb->dev, skb_get_queue_mapping(skb));
txq               244 drivers/net/ethernet/cavium/liquidio/lio_core.c 	netdev_tx_sent_queue(txq, skb->len);
txq               246 drivers/net/ethernet/cavium/liquidio/lio_core.c 	return netif_xmit_stopped(txq);
txq              1071 drivers/net/ethernet/cavium/liquidio/lio_ethtool.c 	lio->txq = lio->linfo.txpciq[0].s.q_no;
txq              3655 drivers/net/ethernet/cavium/liquidio/lio_main.c 		lio->txq = lio->linfo.txpciq[0].s.q_no;
txq              3666 drivers/net/ethernet/cavium/liquidio/lio_main.c 		lio->tx_qsize = octeon_get_tx_qsize(octeon_dev, lio->txq);
txq              2153 drivers/net/ethernet/cavium/liquidio/lio_vf_main.c 		lio->txq = lio->linfo.txpciq[0].s.q_no;
txq              2156 drivers/net/ethernet/cavium/liquidio/lio_vf_main.c 		lio->tx_qsize = octeon_get_tx_qsize(octeon_dev, lio->txq);
txq               386 drivers/net/ethernet/cavium/liquidio/lio_vf_rep.c 	if (octnet_iq_is_full(vf_rep->oct, parent_lio->txq)) {
txq               417 drivers/net/ethernet/cavium/liquidio/lio_vf_rep.c 	sc->iq_no = parent_lio->txq;
txq                70 drivers/net/ethernet/cavium/liquidio/octeon_main.h void octeon_report_tx_completion_to_bql(void *txq, unsigned int pkts_compl,
txq               105 drivers/net/ethernet/cavium/liquidio/octeon_network.h 	int txq;
txq               858 drivers/net/ethernet/cavium/thunder/nicvf_main.c 	struct netdev_queue *txq;
txq               939 drivers/net/ethernet/cavium/thunder/nicvf_main.c 		txq = netdev_get_tx_queue(netdev, txq_idx);
txq               941 drivers/net/ethernet/cavium/thunder/nicvf_main.c 			netdev_tx_completed_queue(txq, tx_pkts, tx_bytes);
txq               945 drivers/net/ethernet/cavium/thunder/nicvf_main.c 		if (netif_tx_queue_stopped(txq) && netif_carrier_ok(netdev)) {
txq               946 drivers/net/ethernet/cavium/thunder/nicvf_main.c 			netif_tx_wake_queue(txq);
txq              1258 drivers/net/ethernet/cavium/thunder/nicvf_main.c 	struct netdev_queue *txq = netdev_get_tx_queue(netdev, qid);
txq              1292 drivers/net/ethernet/cavium/thunder/nicvf_main.c 	if (!netif_tx_queue_stopped(txq) &&
txq              1294 drivers/net/ethernet/cavium/thunder/nicvf_main.c 		netif_tx_stop_queue(txq);
txq              1301 drivers/net/ethernet/cavium/thunder/nicvf_main.c 			netif_tx_wake_queue(txq);
txq              1468 drivers/net/ethernet/cavium/thunder/nicvf_queues.c 	struct netdev_queue *txq;
txq              1470 drivers/net/ethernet/cavium/thunder/nicvf_queues.c 	txq = netdev_get_tx_queue(nic->pnicvf->netdev,
txq              1473 drivers/net/ethernet/cavium/thunder/nicvf_queues.c 	netdev_tx_sent_queue(txq, skb->len);
txq               204 drivers/net/ethernet/chelsio/cxgb3/adapter.h 	struct sge_txq txq[SGE_TXQ_PER_SET];
txq               601 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c 				t3_write_reg(adap, A_SG_KDOORBELL, F_SELEGRCNTX | V_EGRCNTX(qs->txq[j].cntxt_id));
txq               179 drivers/net/ethernet/chelsio/cxgb3/sge.c 	return container_of(q, struct sge_qset, txq[qidx]);
txq               658 drivers/net/ethernet/chelsio/cxgb3/sge.c 	memset(q->txq, 0, sizeof(struct sge_txq) * SGE_TXQ_PER_SET);
txq               695 drivers/net/ethernet/chelsio/cxgb3/sge.c 		if (q->txq[i].desc) {
txq               697 drivers/net/ethernet/chelsio/cxgb3/sge.c 			t3_sge_enable_ecntxt(adapter, q->txq[i].cntxt_id, 0);
txq               699 drivers/net/ethernet/chelsio/cxgb3/sge.c 			if (q->txq[i].sdesc) {
txq               700 drivers/net/ethernet/chelsio/cxgb3/sge.c 				free_tx_desc(adapter, &q->txq[i],
txq               701 drivers/net/ethernet/chelsio/cxgb3/sge.c 					     q->txq[i].in_use);
txq               702 drivers/net/ethernet/chelsio/cxgb3/sge.c 				kfree(q->txq[i].sdesc);
txq               705 drivers/net/ethernet/chelsio/cxgb3/sge.c 					  q->txq[i].size *
txq               707 drivers/net/ethernet/chelsio/cxgb3/sge.c 					  q->txq[i].desc, q->txq[i].phys_addr);
txq               708 drivers/net/ethernet/chelsio/cxgb3/sge.c 			__skb_queue_purge(&q->txq[i].sendq);
txq               735 drivers/net/ethernet/chelsio/cxgb3/sge.c 	qs->txq[TXQ_ETH].cntxt_id = FW_TUNNEL_SGEEC_START + id;
txq               736 drivers/net/ethernet/chelsio/cxgb3/sge.c 	qs->txq[TXQ_ETH].token = FW_TUNNEL_TID_START + id;
txq               737 drivers/net/ethernet/chelsio/cxgb3/sge.c 	qs->txq[TXQ_OFLD].cntxt_id = FW_OFLD_SGEEC_START + id;
txq               738 drivers/net/ethernet/chelsio/cxgb3/sge.c 	qs->txq[TXQ_CTRL].cntxt_id = FW_CTRL_SGEEC_START + id;
txq               739 drivers/net/ethernet/chelsio/cxgb3/sge.c 	qs->txq[TXQ_CTRL].token = FW_CTRL_TID_START + id;
txq              1248 drivers/net/ethernet/chelsio/cxgb3/sge.c static inline void t3_stop_tx_queue(struct netdev_queue *txq,
txq              1251 drivers/net/ethernet/chelsio/cxgb3/sge.c 	netif_tx_stop_queue(txq);
txq              1269 drivers/net/ethernet/chelsio/cxgb3/sge.c 	struct netdev_queue *txq;
txq              1285 drivers/net/ethernet/chelsio/cxgb3/sge.c 	q = &qs->txq[TXQ_ETH];
txq              1286 drivers/net/ethernet/chelsio/cxgb3/sge.c 	txq = netdev_get_tx_queue(dev, qidx);
txq              1294 drivers/net/ethernet/chelsio/cxgb3/sge.c 		t3_stop_tx_queue(txq, qs, q);
txq              1311 drivers/net/ethernet/chelsio/cxgb3/sge.c 		t3_stop_tx_queue(txq, qs, q);
txq              1316 drivers/net/ethernet/chelsio/cxgb3/sge.c 			netif_tx_start_queue(txq);
txq              1527 drivers/net/ethernet/chelsio/cxgb3/sge.c 	struct sge_txq *q = &qs->txq[TXQ_CTRL];
txq              1567 drivers/net/ethernet/chelsio/cxgb3/sge.c 	ret = ctrl_xmit(adap, &adap->sge.qs[0].txq[TXQ_CTRL], skb);
txq              1744 drivers/net/ethernet/chelsio/cxgb3/sge.c 	struct sge_txq *q = &qs->txq[TXQ_OFLD];
txq              1838 drivers/net/ethernet/chelsio/cxgb3/sge.c 		return ctrl_xmit(adap, &qs->txq[TXQ_CTRL], skb);
txq              1840 drivers/net/ethernet/chelsio/cxgb3/sge.c 	return ofld_xmit(adap, &qs->txq[TXQ_OFLD], skb);
txq              1987 drivers/net/ethernet/chelsio/cxgb3/sge.c 	    should_restart_tx(&qs->txq[TXQ_ETH]) &&
txq              1989 drivers/net/ethernet/chelsio/cxgb3/sge.c 		qs->txq[TXQ_ETH].restarts++;
txq              1995 drivers/net/ethernet/chelsio/cxgb3/sge.c 	    should_restart_tx(&qs->txq[TXQ_OFLD]) &&
txq              1997 drivers/net/ethernet/chelsio/cxgb3/sge.c 		qs->txq[TXQ_OFLD].restarts++;
txq              1998 drivers/net/ethernet/chelsio/cxgb3/sge.c 		tasklet_schedule(&qs->txq[TXQ_OFLD].qresume_tsk);
txq              2001 drivers/net/ethernet/chelsio/cxgb3/sge.c 	    should_restart_tx(&qs->txq[TXQ_CTRL]) &&
txq              2003 drivers/net/ethernet/chelsio/cxgb3/sge.c 		qs->txq[TXQ_CTRL].restarts++;
txq              2004 drivers/net/ethernet/chelsio/cxgb3/sge.c 		tasklet_schedule(&qs->txq[TXQ_CTRL].qresume_tsk);
txq              2220 drivers/net/ethernet/chelsio/cxgb3/sge.c 		clear_bit(TXQ_RUNNING, &qs->txq[TXQ_ETH].flags);
txq              2225 drivers/net/ethernet/chelsio/cxgb3/sge.c 		qs->txq[TXQ_ETH].processed += credits;
txq              2229 drivers/net/ethernet/chelsio/cxgb3/sge.c 		qs->txq[TXQ_CTRL].processed += credits;
txq              2233 drivers/net/ethernet/chelsio/cxgb3/sge.c 		clear_bit(TXQ_RUNNING, &qs->txq[TXQ_OFLD].flags);
txq              2237 drivers/net/ethernet/chelsio/cxgb3/sge.c 		qs->txq[TXQ_OFLD].processed += credits;
txq              2254 drivers/net/ethernet/chelsio/cxgb3/sge.c 		struct sge_txq *txq = &qs->txq[TXQ_ETH];
txq              2256 drivers/net/ethernet/chelsio/cxgb3/sge.c 		if (txq->cleaned + txq->in_use != txq->processed &&
txq              2257 drivers/net/ethernet/chelsio/cxgb3/sge.c 		    !test_and_set_bit(TXQ_LAST_PKT_DB, &txq->flags)) {
txq              2258 drivers/net/ethernet/chelsio/cxgb3/sge.c 			set_bit(TXQ_RUNNING, &txq->flags);
txq              2260 drivers/net/ethernet/chelsio/cxgb3/sge.c 				     V_EGRCNTX(txq->cntxt_id));
txq              2265 drivers/net/ethernet/chelsio/cxgb3/sge.c 		struct sge_txq *txq = &qs->txq[TXQ_OFLD];
txq              2267 drivers/net/ethernet/chelsio/cxgb3/sge.c 		if (txq->cleaned + txq->in_use != txq->processed &&
txq              2268 drivers/net/ethernet/chelsio/cxgb3/sge.c 		    !test_and_set_bit(TXQ_LAST_PKT_DB, &txq->flags)) {
txq              2269 drivers/net/ethernet/chelsio/cxgb3/sge.c 			set_bit(TXQ_RUNNING, &txq->flags);
txq              2271 drivers/net/ethernet/chelsio/cxgb3/sge.c 				     V_EGRCNTX(txq->cntxt_id));
txq              2930 drivers/net/ethernet/chelsio/cxgb3/sge.c                 tbd[TXQ_ETH] = reclaim_completed_tx(adap, &qs->txq[TXQ_ETH],
txq              2935 drivers/net/ethernet/chelsio/cxgb3/sge.c 	if (spin_trylock(&qs->txq[TXQ_OFLD].lock)) {
txq              2936 drivers/net/ethernet/chelsio/cxgb3/sge.c 		tbd[TXQ_OFLD] = reclaim_completed_tx(adap, &qs->txq[TXQ_OFLD],
txq              2938 drivers/net/ethernet/chelsio/cxgb3/sge.c 		spin_unlock(&qs->txq[TXQ_OFLD].lock);
txq              3074 drivers/net/ethernet/chelsio/cxgb3/sge.c 		q->txq[i].desc = alloc_ring(adapter->pdev, p->txq_size[i],
txq              3076 drivers/net/ethernet/chelsio/cxgb3/sge.c 					    &q->txq[i].phys_addr,
txq              3077 drivers/net/ethernet/chelsio/cxgb3/sge.c 					    &q->txq[i].sdesc);
txq              3078 drivers/net/ethernet/chelsio/cxgb3/sge.c 		if (!q->txq[i].desc)
txq              3081 drivers/net/ethernet/chelsio/cxgb3/sge.c 		q->txq[i].gen = 1;
txq              3082 drivers/net/ethernet/chelsio/cxgb3/sge.c 		q->txq[i].size = p->txq_size[i];
txq              3083 drivers/net/ethernet/chelsio/cxgb3/sge.c 		spin_lock_init(&q->txq[i].lock);
txq              3084 drivers/net/ethernet/chelsio/cxgb3/sge.c 		skb_queue_head_init(&q->txq[i].sendq);
txq              3087 drivers/net/ethernet/chelsio/cxgb3/sge.c 	tasklet_init(&q->txq[TXQ_OFLD].qresume_tsk, restart_offloadq,
txq              3089 drivers/net/ethernet/chelsio/cxgb3/sge.c 	tasklet_init(&q->txq[TXQ_CTRL].qresume_tsk, restart_ctrlq,
txq              3101 drivers/net/ethernet/chelsio/cxgb3/sge.c 	q->txq[TXQ_ETH].stop_thres = nports *
txq              3142 drivers/net/ethernet/chelsio/cxgb3/sge.c 	ret = t3_sge_init_ecntxt(adapter, q->txq[TXQ_ETH].cntxt_id, USE_GTS,
txq              3143 drivers/net/ethernet/chelsio/cxgb3/sge.c 				 SGE_CNTXT_ETH, id, q->txq[TXQ_ETH].phys_addr,
txq              3144 drivers/net/ethernet/chelsio/cxgb3/sge.c 				 q->txq[TXQ_ETH].size, q->txq[TXQ_ETH].token,
txq              3150 drivers/net/ethernet/chelsio/cxgb3/sge.c 		ret = t3_sge_init_ecntxt(adapter, q->txq[TXQ_OFLD].cntxt_id,
txq              3152 drivers/net/ethernet/chelsio/cxgb3/sge.c 					 q->txq[TXQ_OFLD].phys_addr,
txq              3153 drivers/net/ethernet/chelsio/cxgb3/sge.c 					 q->txq[TXQ_OFLD].size, 0, 1, 0);
txq              3159 drivers/net/ethernet/chelsio/cxgb3/sge.c 		ret = t3_sge_init_ecntxt(adapter, q->txq[TXQ_CTRL].cntxt_id, 0,
txq              3161 drivers/net/ethernet/chelsio/cxgb3/sge.c 					 q->txq[TXQ_CTRL].phys_addr,
txq              3162 drivers/net/ethernet/chelsio/cxgb3/sge.c 					 q->txq[TXQ_CTRL].size,
txq              3163 drivers/net/ethernet/chelsio/cxgb3/sge.c 					 q->txq[TXQ_CTRL].token, 1, 0);
txq              3295 drivers/net/ethernet/chelsio/cxgb3/sge.c 			tasklet_kill(&qs->txq[TXQ_OFLD].qresume_tsk);
txq              3296 drivers/net/ethernet/chelsio/cxgb3/sge.c 			tasklet_kill(&qs->txq[TXQ_CTRL].qresume_tsk);
txq              3003 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c 	struct sge_txq *txq = (struct sge_txq *)q; \
txq              3004 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c 	QDESC_GET(txq, txq->desc, type, label); \
txq               237 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h static inline void cudbg_fill_qdesc_txq(const struct sge_txq *txq,
txq               242 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h 	entry->qid = txq->cntxt_id;
txq               244 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h 	entry->num_desc = txq->size;
txq               245 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h 	entry->data_size = txq->size * sizeof(struct tx_desc);
txq               246 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h 	memcpy(entry->data, txq->desc, entry->data_size);
txq               759 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h 	struct netdev_queue *txq;   /* associated netdev TX queue */
txq              1419 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h int t4_sge_alloc_eth_txq(struct adapter *adap, struct sge_eth_txq *txq,
txq              1422 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h int t4_sge_alloc_ctrl_txq(struct adapter *adap, struct sge_ctrl_txq *txq,
txq              1427 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h int t4_sge_alloc_uld_txq(struct adapter *adap, struct sge_uld_txq *txq,
txq              1783 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h 		unsigned int vf, unsigned int txq, unsigned int txq_eth_ctrl,
txq               931 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c 	struct sge_eth_txq *txq;
txq               933 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c 	txq = &adap->sge.ethtxq[pi->first_qset];
txq               939 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c 	return adap->sge.dbqtimer_val[txq->dbqtimerix];
txq               989 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c 	struct sge_eth_txq *txq;
txq              1015 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c 	txq = &s->ethtxq[pi->first_qset];
txq              1016 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c 	if (txq->dbqtimerix == min_timerix)
txq              1019 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c 	for (qix = 0; qix < pi->nqsets; qix++, txq++) {
txq              1024 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c 			  FW_PARAMS_PARAM_YZ_V(txq->q.cntxt_id));
txq              1031 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c 		txq->dbqtimerix = min_timerix;
txq               236 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	struct sge_eth_txq *txq = &adap->sge.ethtxq[pi->first_qset];
txq               242 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	for (i = 0; i < pi->nqsets; i++, txq++) {
txq               249 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 			FW_PARAMS_PARAM_YZ_V(txq->q.cntxt_id));
txq               265 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 			txq->dcb_prio = enable ? value : 0;
txq               576 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 		struct sge_txq *txq;
txq               578 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 		txq = q->adap->sge.egr_map[qid - q->adap->sge.egr_start];
txq               579 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 		txq->restarts++;
txq               580 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 		if (txq->q_type == CXGB4_TXQ_ETH) {
txq               583 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 			eq = container_of(txq, struct sge_eth_txq, q);
txq               588 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 			oq = container_of(txq, struct sge_uld_txq, q);
txq              1064 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	int txq;
txq              1081 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 			txq = 0;
txq              1083 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 			txq = (vlan_tci & VLAN_PRIO_MASK) >> VLAN_PRIO_SHIFT;
txq              1086 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 				txq = skb->priority & 0x7;
txq              1089 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 		return txq;
txq              1094 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 		txq = (skb_rx_queue_recorded(skb)
txq              1098 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 		while (unlikely(txq >= dev->real_num_tx_queues))
txq              1099 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 			txq -= dev->real_num_tx_queues;
txq              1101 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 		return txq;
txq              2064 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 				struct sge_uld_txq *txq = &txq_info->uldtxq[i];
txq              2066 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 				disable_txq_db(&txq->q);
txq              2086 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 				struct sge_uld_txq *txq = &txq_info->uldtxq[i];
txq              2088 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 				enable_txq_db(adap, &txq->q);
txq              2167 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 				struct sge_uld_txq *txq = &txq_info->uldtxq[i];
txq              2169 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 				sync_txq_pidx(adap, &txq->q);
txq               464 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c 		struct sge_uld_txq *txq = &txq_info->uldtxq[i];
txq               466 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c 		if (txq && txq->q.desc) {
txq               467 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c 			tasklet_kill(&txq->qresume_tsk);
txq               469 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c 					txq->q.cntxt_id);
txq               470 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c 			free_tx_desc(adap, &txq->q, txq->q.in_use, false);
txq               471 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c 			kfree(txq->q.sdesc);
txq               472 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c 			__skb_queue_purge(&txq->sendq);
txq               473 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c 			free_txq(adap, &txq->q);
txq               488 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c 		struct sge_uld_txq *txq = &txq_info->uldtxq[i];
txq               490 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c 		txq->q.size = 1024;
txq               491 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c 		err = t4_sge_alloc_uld_txq(adap, txq, adap->port[i / j],
txq               148 drivers/net/ethernet/chelsio/cxgb4/sched.c 	struct sge_eth_txq *txq;
txq               156 drivers/net/ethernet/chelsio/cxgb4/sched.c 	txq = &adap->sge.ethtxq[pi->first_qset + p->queue];
txq               157 drivers/net/ethernet/chelsio/cxgb4/sched.c 	qid = txq->q.cntxt_id;
txq               190 drivers/net/ethernet/chelsio/cxgb4/sched.c 	struct sge_eth_txq *txq;
txq               201 drivers/net/ethernet/chelsio/cxgb4/sched.c 	txq = &adap->sge.ethtxq[pi->first_qset + p->queue];
txq               202 drivers/net/ethernet/chelsio/cxgb4/sched.c 	qid = txq->q.cntxt_id;
txq              1169 drivers/net/ethernet/chelsio/cxgb4/sge.c 	netif_tx_stop_queue(q->txq);
txq              1331 drivers/net/ethernet/chelsio/cxgb4/sge.c 	if (!q->in_use || !__netif_tx_trylock(eq->txq))
txq              1347 drivers/net/ethernet/chelsio/cxgb4/sge.c 	if (netif_tx_queue_stopped(eq->txq) && hw_in_use < (q->size / 2)) {
txq              1348 drivers/net/ethernet/chelsio/cxgb4/sge.c 		netif_tx_wake_queue(eq->txq);
txq              1352 drivers/net/ethernet/chelsio/cxgb4/sge.c 	__netif_tx_unlock(eq->txq);
txq              1714 drivers/net/ethernet/chelsio/cxgb4/sge.c 	struct sge_eth_txq *txq;
txq              1743 drivers/net/ethernet/chelsio/cxgb4/sge.c 	txq = &adapter->sge.ethtxq[pi->first_qset + qidx];
txq              1748 drivers/net/ethernet/chelsio/cxgb4/sge.c 	reclaim_completed_tx(adapter, &txq->q, -1, true);
txq              1756 drivers/net/ethernet/chelsio/cxgb4/sge.c 	credits = txq_avail(&txq->q) - ndesc;
txq              1764 drivers/net/ethernet/chelsio/cxgb4/sge.c 		eth_txq_stop(txq);
txq              1777 drivers/net/ethernet/chelsio/cxgb4/sge.c 		txq->mapping_err++;
txq              1791 drivers/net/ethernet/chelsio/cxgb4/sge.c 		eth_txq_stop(txq);
txq              1801 drivers/net/ethernet/chelsio/cxgb4/sge.c 	wr = (void *)&txq->q.desc[txq->q.pidx];
txq              1853 drivers/net/ethernet/chelsio/cxgb4/sge.c 		txq->tso++;
txq              1854 drivers/net/ethernet/chelsio/cxgb4/sge.c 		txq->tx_cso += ssi->gso_segs;
txq              1872 drivers/net/ethernet/chelsio/cxgb4/sge.c 			txq->tx_cso++;
txq              1882 drivers/net/ethernet/chelsio/cxgb4/sge.c 		txq->vlan_ins++;
txq              1901 drivers/net/ethernet/chelsio/cxgb4/sge.c 		cxgb4_inline_tx_skb(skb, &txq->q, cpl + 1);
txq              1941 drivers/net/ethernet/chelsio/cxgb4/sge.c 		struct sge_txq *tq = &txq->q;
txq              1969 drivers/net/ethernet/chelsio/cxgb4/sge.c 	txq_advance(&txq->q, ndesc);
txq              1971 drivers/net/ethernet/chelsio/cxgb4/sge.c 	cxgb4_ring_tx_db(adapter, &txq->q, ndesc);
txq              2251 drivers/net/ethernet/chelsio/cxgb4/sge.c 	struct sge_txq *txq;
txq              2301 drivers/net/ethernet/chelsio/cxgb4/sge.c 			txq = &q->q;
txq              2306 drivers/net/ethernet/chelsio/cxgb4/sge.c 				left = (u8 *)end - (u8 *)txq->stat;
txq              2307 drivers/net/ethernet/chelsio/cxgb4/sge.c 				end = (void *)txq->desc + left;
txq              2314 drivers/net/ethernet/chelsio/cxgb4/sge.c 			if (pos == (u64 *)txq->stat) {
txq              2315 drivers/net/ethernet/chelsio/cxgb4/sge.c 				left = (u8 *)end - (u8 *)txq->stat;
txq              2316 drivers/net/ethernet/chelsio/cxgb4/sge.c 				end = (void *)txq->desc + left;
txq              2317 drivers/net/ethernet/chelsio/cxgb4/sge.c 				pos = (void *)txq->desc;
txq              2431 drivers/net/ethernet/chelsio/cxgb4/sge.c 	struct sge_uld_txq *txq;
txq              2447 drivers/net/ethernet/chelsio/cxgb4/sge.c 	txq = &txq_info->uldtxq[idx];
txq              2448 drivers/net/ethernet/chelsio/cxgb4/sge.c 	return ofld_xmit(txq, skb);
txq              2560 drivers/net/ethernet/chelsio/cxgb4/sge.c 	struct sge_uld_txq *txq;
txq              2573 drivers/net/ethernet/chelsio/cxgb4/sge.c 	txq = &txq_info->uldtxq[idx];
txq              2575 drivers/net/ethernet/chelsio/cxgb4/sge.c 	ret = ofld_xmit_direct(txq, src, len);
txq              2893 drivers/net/ethernet/chelsio/cxgb4/sge.c 	struct sge_eth_txq *txq;
txq              2914 drivers/net/ethernet/chelsio/cxgb4/sge.c 	txq = &s->ethtxq[pi->first_qset + rspq->idx];
txq              2915 drivers/net/ethernet/chelsio/cxgb4/sge.c 	t4_sge_eth_txq_egress_update(adapter, txq, -1);
txq              3431 drivers/net/ethernet/chelsio/cxgb4/sge.c 			struct sge_uld_txq *txq = s->egr_map[id];
txq              3434 drivers/net/ethernet/chelsio/cxgb4/sge.c 			tasklet_schedule(&txq->qresume_tsk);
txq              3728 drivers/net/ethernet/chelsio/cxgb4/sge.c int t4_sge_alloc_eth_txq(struct adapter *adap, struct sge_eth_txq *txq,
txq              3739 drivers/net/ethernet/chelsio/cxgb4/sge.c 	nentries = txq->q.size + s->stat_len / sizeof(struct tx_desc);
txq              3741 drivers/net/ethernet/chelsio/cxgb4/sge.c 	txq->q.desc = alloc_ring(adap->pdev_dev, txq->q.size,
txq              3743 drivers/net/ethernet/chelsio/cxgb4/sge.c 			&txq->q.phys_addr, &txq->q.sdesc, s->stat_len,
txq              3745 drivers/net/ethernet/chelsio/cxgb4/sge.c 	if (!txq->q.desc)
txq              3779 drivers/net/ethernet/chelsio/cxgb4/sge.c 	c.eqaddr = cpu_to_be64(txq->q.phys_addr);
txq              3790 drivers/net/ethernet/chelsio/cxgb4/sge.c 				    FW_EQ_ETH_CMD_TIMERIX_V(txq->dbqtimerix));
txq              3794 drivers/net/ethernet/chelsio/cxgb4/sge.c 		kfree(txq->q.sdesc);
txq              3795 drivers/net/ethernet/chelsio/cxgb4/sge.c 		txq->q.sdesc = NULL;
txq              3798 drivers/net/ethernet/chelsio/cxgb4/sge.c 				  txq->q.desc, txq->q.phys_addr);
txq              3799 drivers/net/ethernet/chelsio/cxgb4/sge.c 		txq->q.desc = NULL;
txq              3803 drivers/net/ethernet/chelsio/cxgb4/sge.c 	txq->q.q_type = CXGB4_TXQ_ETH;
txq              3804 drivers/net/ethernet/chelsio/cxgb4/sge.c 	init_txq(adap, &txq->q, FW_EQ_ETH_CMD_EQID_G(ntohl(c.eqid_pkd)));
txq              3805 drivers/net/ethernet/chelsio/cxgb4/sge.c 	txq->txq = netdevq;
txq              3806 drivers/net/ethernet/chelsio/cxgb4/sge.c 	txq->tso = txq->tx_cso = txq->vlan_ins = 0;
txq              3807 drivers/net/ethernet/chelsio/cxgb4/sge.c 	txq->mapping_err = 0;
txq              3808 drivers/net/ethernet/chelsio/cxgb4/sge.c 	txq->dbqt = dbqt;
txq              3813 drivers/net/ethernet/chelsio/cxgb4/sge.c int t4_sge_alloc_ctrl_txq(struct adapter *adap, struct sge_ctrl_txq *txq,
txq              3824 drivers/net/ethernet/chelsio/cxgb4/sge.c 	nentries = txq->q.size + s->stat_len / sizeof(struct tx_desc);
txq              3826 drivers/net/ethernet/chelsio/cxgb4/sge.c 	txq->q.desc = alloc_ring(adap->pdev_dev, nentries,
txq              3827 drivers/net/ethernet/chelsio/cxgb4/sge.c 				 sizeof(struct tx_desc), 0, &txq->q.phys_addr,
txq              3829 drivers/net/ethernet/chelsio/cxgb4/sge.c 	if (!txq->q.desc)
txq              3851 drivers/net/ethernet/chelsio/cxgb4/sge.c 	c.eqaddr = cpu_to_be64(txq->q.phys_addr);
txq              3857 drivers/net/ethernet/chelsio/cxgb4/sge.c 				  txq->q.desc, txq->q.phys_addr);
txq              3858 drivers/net/ethernet/chelsio/cxgb4/sge.c 		txq->q.desc = NULL;
txq              3862 drivers/net/ethernet/chelsio/cxgb4/sge.c 	txq->q.q_type = CXGB4_TXQ_CTRL;
txq              3863 drivers/net/ethernet/chelsio/cxgb4/sge.c 	init_txq(adap, &txq->q, FW_EQ_CTRL_CMD_EQID_G(ntohl(c.cmpliqid_eqid)));
txq              3864 drivers/net/ethernet/chelsio/cxgb4/sge.c 	txq->adap = adap;
txq              3865 drivers/net/ethernet/chelsio/cxgb4/sge.c 	skb_queue_head_init(&txq->sendq);
txq              3866 drivers/net/ethernet/chelsio/cxgb4/sge.c 	tasklet_init(&txq->qresume_tsk, restart_ctrlq, (unsigned long)txq);
txq              3867 drivers/net/ethernet/chelsio/cxgb4/sge.c 	txq->full = 0;
txq              3883 drivers/net/ethernet/chelsio/cxgb4/sge.c int t4_sge_alloc_uld_txq(struct adapter *adap, struct sge_uld_txq *txq,
txq              3895 drivers/net/ethernet/chelsio/cxgb4/sge.c 	nentries = txq->q.size + s->stat_len / sizeof(struct tx_desc);
txq              3897 drivers/net/ethernet/chelsio/cxgb4/sge.c 	txq->q.desc = alloc_ring(adap->pdev_dev, txq->q.size,
txq              3899 drivers/net/ethernet/chelsio/cxgb4/sge.c 			&txq->q.phys_addr, &txq->q.sdesc, s->stat_len,
txq              3901 drivers/net/ethernet/chelsio/cxgb4/sge.c 	if (!txq->q.desc)
txq              3924 drivers/net/ethernet/chelsio/cxgb4/sge.c 	c.eqaddr = cpu_to_be64(txq->q.phys_addr);
txq              3928 drivers/net/ethernet/chelsio/cxgb4/sge.c 		kfree(txq->q.sdesc);
txq              3929 drivers/net/ethernet/chelsio/cxgb4/sge.c 		txq->q.sdesc = NULL;
txq              3932 drivers/net/ethernet/chelsio/cxgb4/sge.c 				  txq->q.desc, txq->q.phys_addr);
txq              3933 drivers/net/ethernet/chelsio/cxgb4/sge.c 		txq->q.desc = NULL;
txq              3937 drivers/net/ethernet/chelsio/cxgb4/sge.c 	txq->q.q_type = CXGB4_TXQ_ULD;
txq              3938 drivers/net/ethernet/chelsio/cxgb4/sge.c 	init_txq(adap, &txq->q, FW_EQ_OFLD_CMD_EQID_G(ntohl(c.eqid_pkd)));
txq              3939 drivers/net/ethernet/chelsio/cxgb4/sge.c 	txq->adap = adap;
txq              3940 drivers/net/ethernet/chelsio/cxgb4/sge.c 	skb_queue_head_init(&txq->sendq);
txq              3941 drivers/net/ethernet/chelsio/cxgb4/sge.c 	tasklet_init(&txq->qresume_tsk, restart_ofldq, (unsigned long)txq);
txq              3942 drivers/net/ethernet/chelsio/cxgb4/sge.c 	txq->full = 0;
txq              3943 drivers/net/ethernet/chelsio/cxgb4/sge.c 	txq->mapping_err = 0;
txq              4036 drivers/net/ethernet/chelsio/cxgb4/sge.c 			__netif_tx_lock_bh(etq->txq);
txq              4038 drivers/net/ethernet/chelsio/cxgb4/sge.c 			__netif_tx_unlock_bh(etq->txq);
txq              4114 drivers/net/ethernet/chelsio/cxgb4/sge.c 			struct sge_uld_txq *txq = txq_info->uldtxq;
txq              4117 drivers/net/ethernet/chelsio/cxgb4/sge.c 				if (txq->q.desc)
txq              4118 drivers/net/ethernet/chelsio/cxgb4/sge.c 					tasklet_kill(&txq->qresume_tsk);
txq              4128 drivers/net/ethernet/chelsio/cxgb4/sge.c 			struct sge_uld_txq *txq = txq_info->uldtxq;
txq              4131 drivers/net/ethernet/chelsio/cxgb4/sge.c 				if (txq->q.desc)
txq              4132 drivers/net/ethernet/chelsio/cxgb4/sge.c 					tasklet_kill(&txq->qresume_tsk);
txq              7587 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 		unsigned int vf, unsigned int txq, unsigned int txq_eth_ctrl,
txq              7603 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 				    FW_PFVF_CMD_NEQ_V(txq));
txq               260 drivers/net/ethernet/chelsio/cxgb4vf/adapter.h 	struct netdev_queue *txq;	/* associated netdev TX queue */
txq               538 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 		struct sge_eth_txq *txq;
txq               560 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 		txq = container_of(tq, struct sge_eth_txq, q);
txq               572 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 		txq->q.restarts++;
txq               573 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 		netif_tx_wake_queue(txq->txq);
txq               635 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 		struct sge_eth_txq *txq = &s->ethtxq[pi->first_qset];
txq               638 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 		for (qs = 0; qs < pi->nqsets; qs++, rxq++, txq++) {
txq               645 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 			err = t4vf_sge_alloc_eth_txq(adapter, txq, dev,
txq               666 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 		struct sge_eth_txq *txq = &s->ethtxq[pi->first_qset];
txq               669 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 		for (qs = 0; qs < pi->nqsets; qs++, rxq++, txq++) {
txq               671 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 			EQ_MAP(s, txq->q.abs_id) = &txq->q;
txq              1804 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 	const struct sge_eth_txq *txq = &adapter->sge.ethtxq[pi->first_qset];
txq              1809 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 	for (qs = 0; qs < pi->nqsets; qs++, rxq++, txq++) {
txq              1810 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 		stats->tso += txq->tso;
txq              1811 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 		stats->tx_csum += txq->tx_cso;
txq              1814 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 		stats->vlan_ins += txq->vlan_ins;
txq              2070 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 	#define T(s, v)		S3("u", s, txq[qs].v)
txq              2075 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 		const struct sge_eth_txq *txq = &adapter->sge.ethtxq[r * QPL];
txq              2224 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 	#define T3(fmt, s, v)	S3(fmt, s, txq[qs].v)
txq              2232 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 		const struct sge_eth_txq *txq = &adapter->sge.ethtxq[r * QPL];
txq              2793 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 		struct sge_eth_txq *txq = &s->ethtxq[qs];
txq              2797 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 		txq->q.size = 1024;
txq              1133 drivers/net/ethernet/chelsio/cxgb4vf/sge.c static void txq_stop(struct sge_eth_txq *txq)
txq              1135 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	netif_tx_stop_queue(txq->txq);
txq              1136 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	txq->q.stops++;
txq              1164 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	struct sge_eth_txq *txq;
txq              1198 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	txq = &adapter->sge.ethtxq[pi->first_qset + qidx];
txq              1208 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	reclaim_completed_tx(adapter, &txq->q, true);
txq              1217 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	credits = txq_avail(&txq->q) - ndesc;
txq              1226 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		txq_stop(txq);
txq              1240 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		txq->mapping_err++;
txq              1255 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		txq_stop(txq);
txq              1266 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	wr = (void *)&txq->q.desc[txq->q.pidx];
txq              1322 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		txq->tso++;
txq              1323 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		txq->tx_cso += ssi->gso_segs;
txq              1340 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 			txq->tx_cso++;
txq              1350 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		txq->vlan_ins++;
txq              1365 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	T4_TRACE5(adapter->tb[txq->q.cntxt_id & 7],
txq              1367 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		  ndesc, credits, txq->q.pidx, skb->len, ssi->nr_frags);
txq              1379 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		inline_tx_skb(skb, &txq->q, cpl + 1);
txq              1420 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		struct sge_txq *tq = &txq->q;
txq              1449 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	txq_advance(&txq->q, ndesc);
txq              1451 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	ring_tx_db(adapter, &txq->q, ndesc);
txq              2134 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		struct sge_eth_txq *txq = &s->ethtxq[i];
txq              2136 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		if (reclaimable(&txq->q) && __netif_tx_trylock(txq->txq)) {
txq              2137 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 			int avail = reclaimable(&txq->q);
txq              2142 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 			free_tx_desc(adapter, &txq->q, avail, true);
txq              2143 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 			txq->q.in_use -= avail;
txq              2144 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 			__netif_tx_unlock(txq->txq);
txq              2412 drivers/net/ethernet/chelsio/cxgb4vf/sge.c int t4vf_sge_alloc_eth_txq(struct adapter *adapter, struct sge_eth_txq *txq,
txq              2426 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	nentries = txq->q.size + s->stat_len / sizeof(struct tx_desc);
txq              2432 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	txq->q.desc = alloc_ring(adapter->pdev_dev, txq->q.size,
txq              2435 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 				 &txq->q.phys_addr, &txq->q.sdesc, s->stat_len);
txq              2436 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	if (!txq->q.desc)
txq              2468 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	cmd.eqaddr = cpu_to_be64(txq->q.phys_addr);
txq              2480 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		kfree(txq->q.sdesc);
txq              2481 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		txq->q.sdesc = NULL;
txq              2484 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 				  txq->q.desc, txq->q.phys_addr);
txq              2485 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		txq->q.desc = NULL;
txq              2489 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	txq->q.in_use = 0;
txq              2490 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	txq->q.cidx = 0;
txq              2491 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	txq->q.pidx = 0;
txq              2492 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	txq->q.stat = (void *)&txq->q.desc[txq->q.size];
txq              2493 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	txq->q.cntxt_id = FW_EQ_ETH_CMD_EQID_G(be32_to_cpu(rpl.eqid_pkd));
txq              2494 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	txq->q.bar2_addr = bar2_address(adapter,
txq              2495 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 					txq->q.cntxt_id,
txq              2497 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 					&txq->q.bar2_qid);
txq              2498 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	txq->q.abs_id =
txq              2500 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	txq->txq = devq;
txq              2501 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	txq->tso = 0;
txq              2502 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	txq->tx_cso = 0;
txq              2503 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	txq->vlan_ins = 0;
txq              2504 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	txq->q.stops = 0;
txq              2505 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	txq->q.restarts = 0;
txq              2506 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	txq->mapping_err = 0;
txq              2567 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	struct sge_eth_txq *txq = s->ethtxq;
txq              2572 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	for (qs = 0; qs < adapter->sge.ethqsets; qs++, rxq++, txq++) {
txq              2575 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		if (txq->q.desc) {
txq              2576 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 			t4vf_eth_eq_free(adapter, txq->q.cntxt_id);
txq              2577 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 			free_tx_desc(adapter, &txq->q, txq->q.in_use, true);
txq              2578 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 			kfree(txq->q.sdesc);
txq              2579 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 			free_txq(adapter, &txq->q);
txq               861 drivers/net/ethernet/cisco/enic/enic_main.c 	struct netdev_queue *txq;
txq               870 drivers/net/ethernet/cisco/enic/enic_main.c 	txq = netdev_get_tx_queue(netdev, txq_map);
txq               888 drivers/net/ethernet/cisco/enic/enic_main.c 		netif_tx_stop_queue(txq);
txq               898 drivers/net/ethernet/cisco/enic/enic_main.c 		netif_tx_stop_queue(txq);
txq               900 drivers/net/ethernet/cisco/enic/enic_main.c 	if (!netdev_xmit_more() || netif_xmit_stopped(txq))
txq               125 drivers/net/ethernet/cortina/gemini.c 	struct gmac_txq		txq[TX_QUEUE_NUM];
txq               555 drivers/net/ethernet/cortina/gemini.c 	struct gmac_txq *txq = port->txq;
txq               589 drivers/net/ethernet/cortina/gemini.c 		txq->ring = desc_ring;
txq               590 drivers/net/ethernet/cortina/gemini.c 		txq->skb = skb_tab;
txq               591 drivers/net/ethernet/cortina/gemini.c 		txq->noirq_packets = 0;
txq               597 drivers/net/ethernet/cortina/gemini.c 		txq->cptr = r;
txq               599 drivers/net/ethernet/cortina/gemini.c 		txq++;
txq               607 drivers/net/ethernet/cortina/gemini.c static void gmac_clean_txq(struct net_device *netdev, struct gmac_txq *txq,
txq               613 drivers/net/ethernet/cortina/gemini.c 	unsigned int c = txq->cptr;
txq               629 drivers/net/ethernet/cortina/gemini.c 		txd = txq->ring + c;
txq               639 drivers/net/ethernet/cortina/gemini.c 			dev_kfree_skb(txq->skb[c]);
txq               676 drivers/net/ethernet/cortina/gemini.c 	txq->cptr = c;
txq               695 drivers/net/ethernet/cortina/gemini.c 		gmac_clean_txq(netdev, port->txq + i, r);
txq               699 drivers/net/ethernet/cortina/gemini.c 	kfree(port->txq->skb);
txq               701 drivers/net/ethernet/cortina/gemini.c 			  n_txq * sizeof(*port->txq->ring) << port->txq_order,
txq               702 drivers/net/ethernet/cortina/gemini.c 			  port->txq->ring, port->txq_dma_base);
txq              1111 drivers/net/ethernet/cortina/gemini.c 			       unsigned int txq, int en)
txq              1119 drivers/net/ethernet/cortina/gemini.c 	mask = GMAC0_IRQ0_TXQ0_INTS << (6 * netdev->dev_id + txq);
txq              1138 drivers/net/ethernet/cortina/gemini.c 			    struct gmac_txq *txq, unsigned short *desc)
txq              1193 drivers/net/ethernet/cortina/gemini.c 			txq->skb[w] = skb;
txq              1201 drivers/net/ethernet/cortina/gemini.c 		txd = txq->ring + w;
txq              1221 drivers/net/ethernet/cortina/gemini.c 		dma_unmap_page(geth->dev, txq->ring[w].word2.buf_adr,
txq              1222 drivers/net/ethernet/cortina/gemini.c 			       txq->ring[w].word0.bits.buffer_size,
txq              1235 drivers/net/ethernet/cortina/gemini.c 	struct gmac_txq *txq;
txq              1244 drivers/net/ethernet/cortina/gemini.c 	txq = &port->txq[txq_num];
txq              1252 drivers/net/ethernet/cortina/gemini.c 	d = txq->cptr - w - 1;
txq              1256 drivers/net/ethernet/cortina/gemini.c 		gmac_clean_txq(netdev, txq, r);
txq              1257 drivers/net/ethernet/cortina/gemini.c 		d = txq->cptr - w - 1;
txq              1263 drivers/net/ethernet/cortina/gemini.c 			d = txq->cptr + nfrags + 16;
txq              1265 drivers/net/ethernet/cortina/gemini.c 			txq->ring[d].word3.bits.eofie = 1;
txq              1275 drivers/net/ethernet/cortina/gemini.c 	if (gmac_map_tx_bufs(netdev, skb, txq, &w)) {
txq              1283 drivers/net/ethernet/cortina/gemini.c 		if (gmac_map_tx_bufs(netdev, skb, txq, &w))
txq              1289 drivers/net/ethernet/cortina/gemini.c 	gmac_clean_txq(netdev, txq, r);
txq              1361 drivers/net/ethernet/emulex/benet/be_cmds.c 	struct be_queue_info *txq = &txo->q;
txq              1363 drivers/net/ethernet/emulex/benet/be_cmds.c 	struct be_dma_mem *q_mem = &txq->dma_mem;
txq              1385 drivers/net/ethernet/emulex/benet/be_cmds.c 	req->queue_size = be_encoded_q_len(txq->len);
txq              1393 drivers/net/ethernet/emulex/benet/be_cmds.c 		txq->id = le16_to_cpu(resp->cid);
txq              1398 drivers/net/ethernet/emulex/benet/be_cmds.c 		txq->created = true;
txq               934 drivers/net/ethernet/emulex/benet/be_main.c 	struct be_queue_info *txq = &txo->q;
txq               935 drivers/net/ethernet/emulex/benet/be_main.c 	struct be_eth_hdr_wrb *hdr = queue_index_node(txq, head);
txq               943 drivers/net/ethernet/emulex/benet/be_main.c 	atomic_add(num_frags, &txq->used);
txq               953 drivers/net/ethernet/emulex/benet/be_main.c 	struct be_queue_info *txq = &txo->q;
txq               955 drivers/net/ethernet/emulex/benet/be_main.c 	wrb = queue_head_node(txq);
txq               957 drivers/net/ethernet/emulex/benet/be_main.c 	queue_head_inc(txq);
txq               970 drivers/net/ethernet/emulex/benet/be_main.c 	struct be_queue_info *txq = &txo->q;
txq               973 drivers/net/ethernet/emulex/benet/be_main.c 	txq->head = head;
txq               976 drivers/net/ethernet/emulex/benet/be_main.c 	queue_head_inc(txq);
txq               978 drivers/net/ethernet/emulex/benet/be_main.c 		wrb = queue_head_node(txq);
txq               982 drivers/net/ethernet/emulex/benet/be_main.c 		queue_head_inc(txq);
txq               985 drivers/net/ethernet/emulex/benet/be_main.c 	txq->head = head;
txq              1227 drivers/net/ethernet/emulex/benet/be_main.c 	struct be_queue_info *txq = &txo->q;
txq              1228 drivers/net/ethernet/emulex/benet/be_main.c 	struct be_eth_hdr_wrb *hdr = queue_index_node(txq, txo->last_req_hdr);
txq              1236 drivers/net/ethernet/emulex/benet/be_main.c 		wrb_fill_dummy(queue_head_node(txq));
txq              1237 drivers/net/ethernet/emulex/benet/be_main.c 		queue_head_inc(txq);
txq              1238 drivers/net/ethernet/emulex/benet/be_main.c 		atomic_inc(&txq->used);
txq              2752 drivers/net/ethernet/emulex/benet/be_main.c 	struct be_queue_info *txq = &txo->q;
txq              2760 drivers/net/ethernet/emulex/benet/be_main.c 		if (sent_skbs[txq->tail]) {
txq              2764 drivers/net/ethernet/emulex/benet/be_main.c 			skb = sent_skbs[txq->tail];
txq              2765 drivers/net/ethernet/emulex/benet/be_main.c 			sent_skbs[txq->tail] = NULL;
txq              2766 drivers/net/ethernet/emulex/benet/be_main.c 			queue_tail_inc(txq);  /* skip hdr wrb */
txq              2770 drivers/net/ethernet/emulex/benet/be_main.c 		wrb = queue_tail_node(txq);
txq              2771 drivers/net/ethernet/emulex/benet/be_main.c 		frag_index = txq->tail;
txq              2775 drivers/net/ethernet/emulex/benet/be_main.c 		queue_tail_inc(txq);
txq              2872 drivers/net/ethernet/emulex/benet/be_main.c 	struct be_queue_info *txq;
txq              2884 drivers/net/ethernet/emulex/benet/be_main.c 			txq = &txo->q;
txq              2893 drivers/net/ethernet/emulex/benet/be_main.c 				atomic_sub(num_wrbs, &txq->used);
txq              2909 drivers/net/ethernet/emulex/benet/be_main.c 		txq = &txo->q;
txq              2911 drivers/net/ethernet/emulex/benet/be_main.c 		if (atomic_read(&txq->used)) {
txq              2913 drivers/net/ethernet/emulex/benet/be_main.c 				 i, atomic_read(&txq->used));
txq              2914 drivers/net/ethernet/emulex/benet/be_main.c 			notified_idx = txq->tail;
txq              2915 drivers/net/ethernet/emulex/benet/be_main.c 			end_idx = txq->tail;
txq              2916 drivers/net/ethernet/emulex/benet/be_main.c 			index_adv(&end_idx, atomic_read(&txq->used) - 1,
txq              2917 drivers/net/ethernet/emulex/benet/be_main.c 				  txq->len);
txq              2922 drivers/net/ethernet/emulex/benet/be_main.c 			atomic_sub(num_wrbs, &txq->used);
txq              2923 drivers/net/ethernet/emulex/benet/be_main.c 			BUG_ON(atomic_read(&txq->used));
txq              2928 drivers/net/ethernet/emulex/benet/be_main.c 			txq->head = notified_idx;
txq              2929 drivers/net/ethernet/emulex/benet/be_main.c 			txq->tail = notified_idx;
txq               672 drivers/net/ethernet/faraday/ftgmac100.c 		struct netdev_queue *txq;
txq               674 drivers/net/ethernet/faraday/ftgmac100.c 		txq = netdev_get_tx_queue(netdev, 0);
txq               675 drivers/net/ethernet/faraday/ftgmac100.c 		__netif_tx_lock(txq, smp_processor_id());
txq               679 drivers/net/ethernet/faraday/ftgmac100.c 		__netif_tx_unlock(txq);
txq              2059 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c 	struct netdev_queue *txq;
txq              2109 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c 	txq = netdev_get_tx_queue(net_dev, queue_mapping);
txq              2112 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c 	txq->trans_start = jiffies;
txq               279 drivers/net/ethernet/freescale/fec_main.c #define IS_TSO_HEADER(txq, addr) \
txq               280 drivers/net/ethernet/freescale/fec_main.c 	((addr >= txq->tso_hdrs_dma) && \
txq               281 drivers/net/ethernet/freescale/fec_main.c 	(addr < txq->tso_hdrs_dma + txq->bd.ring_size * TSO_HEADER_SIZE))
txq               305 drivers/net/ethernet/freescale/fec_main.c static int fec_enet_get_free_txdesc_num(struct fec_enet_priv_tx_q *txq)
txq               309 drivers/net/ethernet/freescale/fec_main.c 	entries = (((const char *)txq->dirty_tx -
txq               310 drivers/net/ethernet/freescale/fec_main.c 			(const char *)txq->bd.cur) >> txq->bd.dsize_log2) - 1;
txq               312 drivers/net/ethernet/freescale/fec_main.c 	return entries >= 0 ? entries : entries + txq->bd.ring_size;
txq               338 drivers/net/ethernet/freescale/fec_main.c 	struct fec_enet_priv_tx_q *txq;
txq               344 drivers/net/ethernet/freescale/fec_main.c 	txq = fep->tx_queue[0];
txq               345 drivers/net/ethernet/freescale/fec_main.c 	bdp = txq->bd.base;
txq               350 drivers/net/ethernet/freescale/fec_main.c 			bdp == txq->bd.cur ? 'S' : ' ',
txq               351 drivers/net/ethernet/freescale/fec_main.c 			bdp == txq->dirty_tx ? 'H' : ' ',
txq               355 drivers/net/ethernet/freescale/fec_main.c 			txq->tx_skbuff[index]);
txq               356 drivers/net/ethernet/freescale/fec_main.c 		bdp = fec_enet_get_nextdesc(bdp, &txq->bd);
txq               358 drivers/net/ethernet/freescale/fec_main.c 	} while (bdp != txq->bd.base);
txq               384 drivers/net/ethernet/freescale/fec_main.c fec_enet_txq_submit_frag_skb(struct fec_enet_priv_tx_q *txq,
txq               389 drivers/net/ethernet/freescale/fec_main.c 	struct bufdesc *bdp = txq->bd.cur;
txq               403 drivers/net/ethernet/freescale/fec_main.c 		bdp = fec_enet_get_nextdesc(bdp, &txq->bd);
txq               424 drivers/net/ethernet/freescale/fec_main.c 				estatus |= FEC_TX_BD_FTYPE(txq->bd.qid);
txq               433 drivers/net/ethernet/freescale/fec_main.c 		index = fec_enet_get_bd_index(bdp, &txq->bd);
txq               436 drivers/net/ethernet/freescale/fec_main.c 			memcpy(txq->tx_bounce[index], bufaddr, frag_len);
txq               437 drivers/net/ethernet/freescale/fec_main.c 			bufaddr = txq->tx_bounce[index];
txq               462 drivers/net/ethernet/freescale/fec_main.c 	bdp = txq->bd.cur;
txq               464 drivers/net/ethernet/freescale/fec_main.c 		bdp = fec_enet_get_nextdesc(bdp, &txq->bd);
txq               471 drivers/net/ethernet/freescale/fec_main.c static int fec_enet_txq_submit_skb(struct fec_enet_priv_tx_q *txq,
txq               485 drivers/net/ethernet/freescale/fec_main.c 	entries_free = fec_enet_get_free_txdesc_num(txq);
txq               500 drivers/net/ethernet/freescale/fec_main.c 	bdp = txq->bd.cur;
txq               509 drivers/net/ethernet/freescale/fec_main.c 	index = fec_enet_get_bd_index(bdp, &txq->bd);
txq               512 drivers/net/ethernet/freescale/fec_main.c 		memcpy(txq->tx_bounce[index], skb->data, buflen);
txq               513 drivers/net/ethernet/freescale/fec_main.c 		bufaddr = txq->tx_bounce[index];
txq               529 drivers/net/ethernet/freescale/fec_main.c 		last_bdp = fec_enet_txq_submit_frag_skb(txq, skb, ndev);
txq               557 drivers/net/ethernet/freescale/fec_main.c 			estatus |= FEC_TX_BD_FTYPE(txq->bd.qid);
txq               566 drivers/net/ethernet/freescale/fec_main.c 	index = fec_enet_get_bd_index(last_bdp, &txq->bd);
txq               568 drivers/net/ethernet/freescale/fec_main.c 	txq->tx_skbuff[index] = skb;
txq               582 drivers/net/ethernet/freescale/fec_main.c 	bdp = fec_enet_get_nextdesc(last_bdp, &txq->bd);
txq               590 drivers/net/ethernet/freescale/fec_main.c 	txq->bd.cur = bdp;
txq               593 drivers/net/ethernet/freescale/fec_main.c 	writel(0, txq->bd.reg_desc_active);
txq               599 drivers/net/ethernet/freescale/fec_main.c fec_enet_txq_put_data_tso(struct fec_enet_priv_tx_q *txq, struct sk_buff *skb,
txq               617 drivers/net/ethernet/freescale/fec_main.c 		memcpy(txq->tx_bounce[index], data, size);
txq               618 drivers/net/ethernet/freescale/fec_main.c 		data = txq->tx_bounce[index];
txq               637 drivers/net/ethernet/freescale/fec_main.c 			estatus |= FEC_TX_BD_FTYPE(txq->bd.qid);
txq               659 drivers/net/ethernet/freescale/fec_main.c fec_enet_txq_put_hdr_tso(struct fec_enet_priv_tx_q *txq,
txq               675 drivers/net/ethernet/freescale/fec_main.c 	bufaddr = txq->tso_hdrs + index * TSO_HEADER_SIZE;
txq               676 drivers/net/ethernet/freescale/fec_main.c 	dmabuf = txq->tso_hdrs_dma + index * TSO_HEADER_SIZE;
txq               679 drivers/net/ethernet/freescale/fec_main.c 		memcpy(txq->tx_bounce[index], skb->data, hdr_len);
txq               680 drivers/net/ethernet/freescale/fec_main.c 		bufaddr = txq->tx_bounce[index];
txq               700 drivers/net/ethernet/freescale/fec_main.c 			estatus |= FEC_TX_BD_FTYPE(txq->bd.qid);
txq               712 drivers/net/ethernet/freescale/fec_main.c static int fec_enet_txq_submit_tso(struct fec_enet_priv_tx_q *txq,
txq               719 drivers/net/ethernet/freescale/fec_main.c 	struct bufdesc *bdp = txq->bd.cur;
txq               724 drivers/net/ethernet/freescale/fec_main.c 	if (tso_count_descs(skb) >= fec_enet_get_free_txdesc_num(txq)) {
txq               744 drivers/net/ethernet/freescale/fec_main.c 		index = fec_enet_get_bd_index(bdp, &txq->bd);
txq               749 drivers/net/ethernet/freescale/fec_main.c 		hdr = txq->tso_hdrs + index * TSO_HEADER_SIZE;
txq               751 drivers/net/ethernet/freescale/fec_main.c 		ret = fec_enet_txq_put_hdr_tso(txq, skb, ndev, bdp, index);
txq               759 drivers/net/ethernet/freescale/fec_main.c 			bdp = fec_enet_get_nextdesc(bdp, &txq->bd);
txq               760 drivers/net/ethernet/freescale/fec_main.c 			index = fec_enet_get_bd_index(bdp, &txq->bd);
txq               761 drivers/net/ethernet/freescale/fec_main.c 			ret = fec_enet_txq_put_data_tso(txq, skb, ndev,
txq               773 drivers/net/ethernet/freescale/fec_main.c 		bdp = fec_enet_get_nextdesc(bdp, &txq->bd);
txq               777 drivers/net/ethernet/freescale/fec_main.c 	txq->tx_skbuff[index] = skb;
txq               780 drivers/net/ethernet/freescale/fec_main.c 	txq->bd.cur = bdp;
txq               784 drivers/net/ethernet/freescale/fec_main.c 	    !readl(txq->bd.reg_desc_active) ||
txq               785 drivers/net/ethernet/freescale/fec_main.c 	    !readl(txq->bd.reg_desc_active) ||
txq               786 drivers/net/ethernet/freescale/fec_main.c 	    !readl(txq->bd.reg_desc_active) ||
txq               787 drivers/net/ethernet/freescale/fec_main.c 	    !readl(txq->bd.reg_desc_active))
txq               788 drivers/net/ethernet/freescale/fec_main.c 		writel(0, txq->bd.reg_desc_active);
txq               803 drivers/net/ethernet/freescale/fec_main.c 	struct fec_enet_priv_tx_q *txq;
txq               808 drivers/net/ethernet/freescale/fec_main.c 	txq = fep->tx_queue[queue];
txq               812 drivers/net/ethernet/freescale/fec_main.c 		ret = fec_enet_txq_submit_tso(txq, skb, ndev);
txq               814 drivers/net/ethernet/freescale/fec_main.c 		ret = fec_enet_txq_submit_skb(txq, skb, ndev);
txq               818 drivers/net/ethernet/freescale/fec_main.c 	entries_free = fec_enet_get_free_txdesc_num(txq);
txq               819 drivers/net/ethernet/freescale/fec_main.c 	if (entries_free <= txq->tx_stop_threshold)
txq               830 drivers/net/ethernet/freescale/fec_main.c 	struct fec_enet_priv_tx_q *txq;
txq               860 drivers/net/ethernet/freescale/fec_main.c 		txq = fep->tx_queue[q];
txq               861 drivers/net/ethernet/freescale/fec_main.c 		bdp = txq->bd.base;
txq               862 drivers/net/ethernet/freescale/fec_main.c 		txq->bd.cur = bdp;
txq               864 drivers/net/ethernet/freescale/fec_main.c 		for (i = 0; i < txq->bd.ring_size; i++) {
txq               868 drivers/net/ethernet/freescale/fec_main.c 			    !IS_TSO_HEADER(txq, fec32_to_cpu(bdp->cbd_bufaddr)))
txq               873 drivers/net/ethernet/freescale/fec_main.c 			if (txq->tx_skbuff[i]) {
txq               874 drivers/net/ethernet/freescale/fec_main.c 				dev_kfree_skb_any(txq->tx_skbuff[i]);
txq               875 drivers/net/ethernet/freescale/fec_main.c 				txq->tx_skbuff[i] = NULL;
txq               878 drivers/net/ethernet/freescale/fec_main.c 			bdp = fec_enet_get_nextdesc(bdp, &txq->bd);
txq               882 drivers/net/ethernet/freescale/fec_main.c 		bdp = fec_enet_get_prevdesc(bdp, &txq->bd);
txq               884 drivers/net/ethernet/freescale/fec_main.c 		txq->dirty_tx = bdp;
txq               900 drivers/net/ethernet/freescale/fec_main.c 	struct fec_enet_priv_tx_q *txq;
txq               916 drivers/net/ethernet/freescale/fec_main.c 		txq = fep->tx_queue[i];
txq               917 drivers/net/ethernet/freescale/fec_main.c 		writel(txq->bd.dma, fep->hwp + FEC_X_DES_START(i));
txq               929 drivers/net/ethernet/freescale/fec_main.c 	struct fec_enet_priv_tx_q *txq;
txq               933 drivers/net/ethernet/freescale/fec_main.c 		txq = fep->tx_queue[i];
txq               935 drivers/net/ethernet/freescale/fec_main.c 		for (j = 0; j < txq->bd.ring_size; j++) {
txq               936 drivers/net/ethernet/freescale/fec_main.c 			if (txq->tx_skbuff[j]) {
txq               937 drivers/net/ethernet/freescale/fec_main.c 				dev_kfree_skb_any(txq->tx_skbuff[j]);
txq               938 drivers/net/ethernet/freescale/fec_main.c 				txq->tx_skbuff[j] = NULL;
txq              1248 drivers/net/ethernet/freescale/fec_main.c 	struct fec_enet_priv_tx_q *txq;
txq              1257 drivers/net/ethernet/freescale/fec_main.c 	txq = fep->tx_queue[queue_id];
txq              1260 drivers/net/ethernet/freescale/fec_main.c 	bdp = txq->dirty_tx;
txq              1263 drivers/net/ethernet/freescale/fec_main.c 	bdp = fec_enet_get_nextdesc(bdp, &txq->bd);
txq              1265 drivers/net/ethernet/freescale/fec_main.c 	while (bdp != READ_ONCE(txq->bd.cur)) {
txq              1272 drivers/net/ethernet/freescale/fec_main.c 		index = fec_enet_get_bd_index(bdp, &txq->bd);
txq              1274 drivers/net/ethernet/freescale/fec_main.c 		skb = txq->tx_skbuff[index];
txq              1275 drivers/net/ethernet/freescale/fec_main.c 		txq->tx_skbuff[index] = NULL;
txq              1276 drivers/net/ethernet/freescale/fec_main.c 		if (!IS_TSO_HEADER(txq, fec32_to_cpu(bdp->cbd_bufaddr)))
txq              1327 drivers/net/ethernet/freescale/fec_main.c 		txq->dirty_tx = bdp;
txq              1330 drivers/net/ethernet/freescale/fec_main.c 		bdp = fec_enet_get_nextdesc(bdp, &txq->bd);
txq              1335 drivers/net/ethernet/freescale/fec_main.c 			entries_free = fec_enet_get_free_txdesc_num(txq);
txq              1336 drivers/net/ethernet/freescale/fec_main.c 			if (entries_free >= txq->tx_wake_threshold)
txq              1342 drivers/net/ethernet/freescale/fec_main.c 	if (bdp != txq->bd.cur &&
txq              1343 drivers/net/ethernet/freescale/fec_main.c 	    readl(txq->bd.reg_desc_active) == 0)
txq              1344 drivers/net/ethernet/freescale/fec_main.c 		writel(0, txq->bd.reg_desc_active);
txq              2748 drivers/net/ethernet/freescale/fec_main.c 	struct fec_enet_priv_tx_q *txq;
txq              2770 drivers/net/ethernet/freescale/fec_main.c 		txq = fep->tx_queue[q];
txq              2771 drivers/net/ethernet/freescale/fec_main.c 		bdp = txq->bd.base;
txq              2772 drivers/net/ethernet/freescale/fec_main.c 		for (i = 0; i < txq->bd.ring_size; i++) {
txq              2773 drivers/net/ethernet/freescale/fec_main.c 			kfree(txq->tx_bounce[i]);
txq              2774 drivers/net/ethernet/freescale/fec_main.c 			txq->tx_bounce[i] = NULL;
txq              2775 drivers/net/ethernet/freescale/fec_main.c 			skb = txq->tx_skbuff[i];
txq              2776 drivers/net/ethernet/freescale/fec_main.c 			txq->tx_skbuff[i] = NULL;
txq              2786 drivers/net/ethernet/freescale/fec_main.c 	struct fec_enet_priv_tx_q *txq;
txq              2790 drivers/net/ethernet/freescale/fec_main.c 			txq = fep->tx_queue[i];
txq              2792 drivers/net/ethernet/freescale/fec_main.c 					  txq->bd.ring_size * TSO_HEADER_SIZE,
txq              2793 drivers/net/ethernet/freescale/fec_main.c 					  txq->tso_hdrs,
txq              2794 drivers/net/ethernet/freescale/fec_main.c 					  txq->tso_hdrs_dma);
txq              2808 drivers/net/ethernet/freescale/fec_main.c 	struct fec_enet_priv_tx_q *txq;
txq              2811 drivers/net/ethernet/freescale/fec_main.c 		txq = kzalloc(sizeof(*txq), GFP_KERNEL);
txq              2812 drivers/net/ethernet/freescale/fec_main.c 		if (!txq) {
txq              2817 drivers/net/ethernet/freescale/fec_main.c 		fep->tx_queue[i] = txq;
txq              2818 drivers/net/ethernet/freescale/fec_main.c 		txq->bd.ring_size = TX_RING_SIZE;
txq              2821 drivers/net/ethernet/freescale/fec_main.c 		txq->tx_stop_threshold = FEC_MAX_SKB_DESCS;
txq              2822 drivers/net/ethernet/freescale/fec_main.c 		txq->tx_wake_threshold =
txq              2823 drivers/net/ethernet/freescale/fec_main.c 			(txq->bd.ring_size - txq->tx_stop_threshold) / 2;
txq              2825 drivers/net/ethernet/freescale/fec_main.c 		txq->tso_hdrs = dma_alloc_coherent(&fep->pdev->dev,
txq              2826 drivers/net/ethernet/freescale/fec_main.c 					txq->bd.ring_size * TSO_HEADER_SIZE,
txq              2827 drivers/net/ethernet/freescale/fec_main.c 					&txq->tso_hdrs_dma,
txq              2829 drivers/net/ethernet/freescale/fec_main.c 		if (!txq->tso_hdrs) {
txq              2901 drivers/net/ethernet/freescale/fec_main.c 	struct fec_enet_priv_tx_q *txq;
txq              2903 drivers/net/ethernet/freescale/fec_main.c 	txq = fep->tx_queue[queue];
txq              2904 drivers/net/ethernet/freescale/fec_main.c 	bdp = txq->bd.base;
txq              2905 drivers/net/ethernet/freescale/fec_main.c 	for (i = 0; i < txq->bd.ring_size; i++) {
txq              2906 drivers/net/ethernet/freescale/fec_main.c 		txq->tx_bounce[i] = kmalloc(FEC_ENET_TX_FRSIZE, GFP_KERNEL);
txq              2907 drivers/net/ethernet/freescale/fec_main.c 		if (!txq->tx_bounce[i])
txq              2918 drivers/net/ethernet/freescale/fec_main.c 		bdp = fec_enet_get_nextdesc(bdp, &txq->bd);
txq              2922 drivers/net/ethernet/freescale/fec_main.c 	bdp = fec_enet_get_prevdesc(bdp, &txq->bd);
txq              3290 drivers/net/ethernet/freescale/fec_main.c 		struct fec_enet_priv_tx_q *txq = fep->tx_queue[i];
txq              3291 drivers/net/ethernet/freescale/fec_main.c 		unsigned size = dsize * txq->bd.ring_size;
txq              3293 drivers/net/ethernet/freescale/fec_main.c 		txq->bd.qid = i;
txq              3294 drivers/net/ethernet/freescale/fec_main.c 		txq->bd.base = cbd_base;
txq              3295 drivers/net/ethernet/freescale/fec_main.c 		txq->bd.cur = cbd_base;
txq              3296 drivers/net/ethernet/freescale/fec_main.c 		txq->bd.dma = bd_dma;
txq              3297 drivers/net/ethernet/freescale/fec_main.c 		txq->bd.dsize = dsize;
txq              3298 drivers/net/ethernet/freescale/fec_main.c 		txq->bd.dsize_log2 = dsize_log2;
txq              3299 drivers/net/ethernet/freescale/fec_main.c 		txq->bd.reg_desc_active = fep->hwp + offset_des_active_txq[i];
txq              3302 drivers/net/ethernet/freescale/fec_main.c 		txq->bd.last = (struct bufdesc *)(((void *)cbd_base) - dsize);
txq              1156 drivers/net/ethernet/freescale/gianfar.c 		struct netdev_queue *txq;
txq              1159 drivers/net/ethernet/freescale/gianfar.c 		txq = netdev_get_tx_queue(tx_queue->dev, tx_queue->qindex);
txq              1162 drivers/net/ethernet/freescale/gianfar.c 		netdev_tx_reset_queue(txq);
txq              1797 drivers/net/ethernet/freescale/gianfar.c 	struct netdev_queue *txq;
txq              1810 drivers/net/ethernet/freescale/gianfar.c 	txq = netdev_get_tx_queue(dev, rq);
txq              1855 drivers/net/ethernet/freescale/gianfar.c 		netif_tx_stop_queue(txq);
txq              1979 drivers/net/ethernet/freescale/gianfar.c 	netdev_tx_sent_queue(txq, bytes_sent);
txq              2011 drivers/net/ethernet/freescale/gianfar.c 		netif_tx_stop_queue(txq);
txq              2186 drivers/net/ethernet/freescale/gianfar.c 	struct netdev_queue *txq;
txq              2202 drivers/net/ethernet/freescale/gianfar.c 	txq = netdev_get_tx_queue(dev, tqi);
txq              2282 drivers/net/ethernet/freescale/gianfar.c 	    netif_tx_queue_stopped(txq) &&
txq              2290 drivers/net/ethernet/freescale/gianfar.c 	netdev_tx_completed_queue(txq, howmany, bytes_sent);
txq               120 drivers/net/ethernet/hisilicon/hisi_femac.c 	struct hisi_femac_queue txq;
txq               147 drivers/net/ethernet/hisilicon/hisi_femac.c 	dma_addr = priv->txq.dma_phys[pos];
txq               155 drivers/net/ethernet/hisilicon/hisi_femac.c 	struct hisi_femac_queue *txq = &priv->txq;
txq               163 drivers/net/ethernet/hisilicon/hisi_femac.c 		skb = txq->skb[txq->tail];
txq               169 drivers/net/ethernet/hisilicon/hisi_femac.c 		hisi_femac_tx_dma_unmap(priv, skb, txq->tail);
txq               177 drivers/net/ethernet/hisilicon/hisi_femac.c 		txq->skb[txq->tail] = NULL;
txq               178 drivers/net/ethernet/hisilicon/hisi_femac.c 		txq->tail = (txq->tail + 1) % txq->num;
txq               372 drivers/net/ethernet/hisilicon/hisi_femac.c 	ret = hisi_femac_init_queue(priv->dev, &priv->txq, TXQ_NUM);
txq               387 drivers/net/ethernet/hisilicon/hisi_femac.c 	struct hisi_femac_queue *txq = &priv->txq;
txq               412 drivers/net/ethernet/hisilicon/hisi_femac.c 	pos = txq->tail;
txq               413 drivers/net/ethernet/hisilicon/hisi_femac.c 	while (pos != txq->head) {
txq               414 drivers/net/ethernet/hisilicon/hisi_femac.c 		skb = txq->skb[pos];
txq               417 drivers/net/ethernet/hisilicon/hisi_femac.c 				   pos, txq->head);
txq               422 drivers/net/ethernet/hisilicon/hisi_femac.c 		txq->skb[pos] = NULL;
txq               423 drivers/net/ethernet/hisilicon/hisi_femac.c 		pos = (pos + 1) % txq->num;
txq               425 drivers/net/ethernet/hisilicon/hisi_femac.c 	txq->tail = pos;
txq               503 drivers/net/ethernet/hisilicon/hisi_femac.c 	struct hisi_femac_queue *txq = &priv->txq;
txq               517 drivers/net/ethernet/hisilicon/hisi_femac.c 	if (unlikely(!CIRC_SPACE(txq->head, txq->tail,
txq               518 drivers/net/ethernet/hisilicon/hisi_femac.c 				 txq->num))) {
txq               533 drivers/net/ethernet/hisilicon/hisi_femac.c 	txq->dma_phys[txq->head] = addr;
txq               535 drivers/net/ethernet/hisilicon/hisi_femac.c 	txq->skb[txq->head] = skb;
txq               536 drivers/net/ethernet/hisilicon/hisi_femac.c 	txq->head = (txq->head + 1) % txq->num;
txq                94 drivers/net/ethernet/huawei/hinic/hinic_main.c static void update_tx_stats(struct hinic_dev *nic_dev, struct hinic_txq *txq)
txq               101 drivers/net/ethernet/huawei/hinic/hinic_main.c 	hinic_txq_get_stats(txq, &tx_stats);
txq               112 drivers/net/ethernet/huawei/hinic/hinic_main.c 	hinic_txq_clean_stats(txq);
txq                76 drivers/net/ethernet/huawei/hinic/hinic_tx.c void hinic_txq_clean_stats(struct hinic_txq *txq)
txq                78 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	struct hinic_txq_stats *txq_stats = &txq->txq_stats;
txq                95 drivers/net/ethernet/huawei/hinic/hinic_tx.c void hinic_txq_get_stats(struct hinic_txq *txq, struct hinic_txq_stats *stats)
txq                97 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	struct hinic_txq_stats *txq_stats = &txq->txq_stats;
txq               117 drivers/net/ethernet/huawei/hinic/hinic_tx.c static void txq_stats_init(struct hinic_txq *txq)
txq               119 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	struct hinic_txq_stats *txq_stats = &txq->txq_stats;
txq               122 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	hinic_txq_clean_stats(txq);
txq               470 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	struct hinic_txq *txq;
txq               473 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	txq = &nic_dev->txqs[q_id];
txq               474 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	qp = container_of(txq->sq, struct hinic_qp, sq);
txq               487 drivers/net/ethernet/huawei/hinic/hinic_tx.c 		u64_stats_update_begin(&txq->txq_stats.syncp);
txq               488 drivers/net/ethernet/huawei/hinic/hinic_tx.c 		txq->txq_stats.big_frags_pkts++;
txq               489 drivers/net/ethernet/huawei/hinic/hinic_tx.c 		u64_stats_update_end(&txq->txq_stats.syncp);
txq               492 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	if (nr_sges > txq->max_sges) {
txq               497 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	err = tx_map_skb(nic_dev, skb, txq->sges);
txq               503 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	sq_wqe = hinic_sq_get_wqe(txq->sq, wqe_size, &prod_idx);
txq               510 drivers/net/ethernet/huawei/hinic/hinic_tx.c 		sq_wqe = hinic_sq_get_wqe(txq->sq, wqe_size, &prod_idx);
txq               516 drivers/net/ethernet/huawei/hinic/hinic_tx.c 		tx_unmap_skb(nic_dev, skb, txq->sges);
txq               518 drivers/net/ethernet/huawei/hinic/hinic_tx.c 		u64_stats_update_begin(&txq->txq_stats.syncp);
txq               519 drivers/net/ethernet/huawei/hinic/hinic_tx.c 		txq->txq_stats.tx_busy++;
txq               520 drivers/net/ethernet/huawei/hinic/hinic_tx.c 		u64_stats_update_end(&txq->txq_stats.syncp);
txq               527 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	hinic_sq_prepare_wqe(txq->sq, prod_idx, sq_wqe, txq->sges, nr_sges);
txq               533 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	hinic_sq_write_wqe(txq->sq, prod_idx, sq_wqe, skb, wqe_size);
txq               538 drivers/net/ethernet/huawei/hinic/hinic_tx.c 		hinic_sq_write_db(txq->sq, prod_idx, wqe_size, 0);
txq               543 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	hinic_sq_return_wqe(txq->sq, wqe_size);
txq               544 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	tx_unmap_skb(nic_dev, skb, txq->sges);
txq               550 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	u64_stats_update_begin(&txq->txq_stats.syncp);
txq               551 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	txq->txq_stats.tx_dropped++;
txq               552 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	u64_stats_update_end(&txq->txq_stats.syncp);
txq               575 drivers/net/ethernet/huawei/hinic/hinic_tx.c static void free_all_tx_skbs(struct hinic_txq *txq)
txq               577 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	struct hinic_dev *nic_dev = netdev_priv(txq->netdev);
txq               578 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	struct hinic_sq *sq = txq->sq;
txq               592 drivers/net/ethernet/huawei/hinic/hinic_tx.c 		hinic_sq_get_sges(sq_wqe, txq->free_sges, nr_sges);
txq               596 drivers/net/ethernet/huawei/hinic/hinic_tx.c 		tx_free_skb(nic_dev, skb, txq->free_sges);
txq               609 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	struct hinic_txq *txq = container_of(napi, struct hinic_txq, napi);
txq               610 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	struct hinic_qp *qp = container_of(txq->sq, struct hinic_qp, sq);
txq               611 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	struct hinic_dev *nic_dev = netdev_priv(txq->netdev);
txq               613 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	struct hinic_sq *sq = txq->sq;
txq               647 drivers/net/ethernet/huawei/hinic/hinic_tx.c 		hinic_sq_get_sges(sq_wqe, txq->free_sges, nr_sges);
txq               651 drivers/net/ethernet/huawei/hinic/hinic_tx.c 		tx_free_skb(nic_dev, skb, txq->free_sges);
txq               656 drivers/net/ethernet/huawei/hinic/hinic_tx.c 		netdev_txq = netdev_get_tx_queue(txq->netdev, qp->q_id);
txq               664 drivers/net/ethernet/huawei/hinic/hinic_tx.c 		u64_stats_update_begin(&txq->txq_stats.syncp);
txq               665 drivers/net/ethernet/huawei/hinic/hinic_tx.c 		txq->txq_stats.tx_wake++;
txq               666 drivers/net/ethernet/huawei/hinic/hinic_tx.c 		u64_stats_update_end(&txq->txq_stats.syncp);
txq               669 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	u64_stats_update_begin(&txq->txq_stats.syncp);
txq               670 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	txq->txq_stats.bytes += tx_bytes;
txq               671 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	txq->txq_stats.pkts += pkts;
txq               672 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	u64_stats_update_end(&txq->txq_stats.syncp);
txq               685 drivers/net/ethernet/huawei/hinic/hinic_tx.c static void tx_napi_add(struct hinic_txq *txq, int weight)
txq               687 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	netif_napi_add(txq->netdev, &txq->napi, free_tx_poll, weight);
txq               688 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	napi_enable(&txq->napi);
txq               691 drivers/net/ethernet/huawei/hinic/hinic_tx.c static void tx_napi_del(struct hinic_txq *txq)
txq               693 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	napi_disable(&txq->napi);
txq               694 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	netif_napi_del(&txq->napi);
txq               699 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	struct hinic_txq *txq = data;
txq               702 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	nic_dev = netdev_priv(txq->netdev);
txq               706 drivers/net/ethernet/huawei/hinic/hinic_tx.c 				   txq->sq->msix_entry,
txq               709 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	hinic_hwdev_msix_cnt_set(nic_dev->hwdev, txq->sq->msix_entry);
txq               711 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	napi_schedule(&txq->napi);
txq               715 drivers/net/ethernet/huawei/hinic/hinic_tx.c static int tx_request_irq(struct hinic_txq *txq)
txq               717 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	struct hinic_dev *nic_dev = netdev_priv(txq->netdev);
txq               721 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	struct hinic_sq *sq = txq->sq;
txq               724 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	tx_napi_add(txq, nic_dev->tx_weight);
txq               731 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	err = request_irq(sq->irq, tx_irq, 0, txq->irq_name, txq);
txq               734 drivers/net/ethernet/huawei/hinic/hinic_tx.c 		tx_napi_del(txq);
txq               741 drivers/net/ethernet/huawei/hinic/hinic_tx.c static void tx_free_irq(struct hinic_txq *txq)
txq               743 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	struct hinic_sq *sq = txq->sq;
txq               745 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	free_irq(sq->irq, txq);
txq               746 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	tx_napi_del(txq);
txq               757 drivers/net/ethernet/huawei/hinic/hinic_tx.c int hinic_init_txq(struct hinic_txq *txq, struct hinic_sq *sq,
txq               766 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	txq->netdev = netdev;
txq               767 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	txq->sq = sq;
txq               769 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	txq_stats_init(txq);
txq               771 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	txq->max_sges = HINIC_MAX_SQ_BUFDESCS;
txq               773 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	sges_size = txq->max_sges * sizeof(*txq->sges);
txq               774 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	txq->sges = devm_kzalloc(&netdev->dev, sges_size, GFP_KERNEL);
txq               775 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	if (!txq->sges)
txq               778 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	sges_size = txq->max_sges * sizeof(*txq->free_sges);
txq               779 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	txq->free_sges = devm_kzalloc(&netdev->dev, sges_size, GFP_KERNEL);
txq               780 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	if (!txq->free_sges) {
txq               786 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	txq->irq_name = devm_kzalloc(&netdev->dev, irqname_len, GFP_KERNEL);
txq               787 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	if (!txq->irq_name) {
txq               792 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	sprintf(txq->irq_name, "hinic_txq%d", qp->q_id);
txq               799 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	err = tx_request_irq(txq);
txq               809 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	devm_kfree(&netdev->dev, txq->irq_name);
txq               812 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	devm_kfree(&netdev->dev, txq->free_sges);
txq               815 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	devm_kfree(&netdev->dev, txq->sges);
txq               823 drivers/net/ethernet/huawei/hinic/hinic_tx.c void hinic_clean_txq(struct hinic_txq *txq)
txq               825 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	struct net_device *netdev = txq->netdev;
txq               827 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	tx_free_irq(txq);
txq               829 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	free_all_tx_skbs(txq);
txq               831 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	devm_kfree(&netdev->dev, txq->irq_name);
txq               832 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	devm_kfree(&netdev->dev, txq->free_sges);
txq               833 drivers/net/ethernet/huawei/hinic/hinic_tx.c 	devm_kfree(&netdev->dev, txq->sges);
txq                43 drivers/net/ethernet/huawei/hinic/hinic_tx.h void hinic_txq_clean_stats(struct hinic_txq *txq);
txq                45 drivers/net/ethernet/huawei/hinic/hinic_tx.h void hinic_txq_get_stats(struct hinic_txq *txq, struct hinic_txq_stats *stats);
txq                49 drivers/net/ethernet/huawei/hinic/hinic_tx.h int hinic_init_txq(struct hinic_txq *txq, struct hinic_sq *sq,
txq                52 drivers/net/ethernet/huawei/hinic/hinic_tx.h void hinic_clean_txq(struct hinic_txq *txq);
txq               806 drivers/net/ethernet/ibm/ehea/ehea_main.c 	struct netdev_queue *txq = netdev_get_tx_queue(pr->port->netdev,
txq               857 drivers/net/ethernet/ibm/ehea/ehea_main.c 	if (unlikely(netif_tx_queue_stopped(txq) &&
txq               859 drivers/net/ethernet/ibm/ehea/ehea_main.c 		__netif_tx_lock(txq, smp_processor_id());
txq               860 drivers/net/ethernet/ibm/ehea/ehea_main.c 		if (netif_tx_queue_stopped(txq) &&
txq               862 drivers/net/ethernet/ibm/ehea/ehea_main.c 			netif_tx_wake_queue(txq);
txq               863 drivers/net/ethernet/ibm/ehea/ehea_main.c 		__netif_tx_unlock(txq);
txq              2018 drivers/net/ethernet/ibm/ehea/ehea_main.c 	struct netdev_queue *txq;
txq              2021 drivers/net/ethernet/ibm/ehea/ehea_main.c 	txq = netdev_get_tx_queue(dev, skb_get_queue_mapping(skb));
txq              2071 drivers/net/ethernet/ibm/ehea/ehea_main.c 		netif_tx_stop_queue(txq);
txq              2079 drivers/net/ethernet/ibm/ehea/ehea_main.c 		netif_tx_stop_queue(txq);
txq              1420 drivers/net/ethernet/ibm/ibmvnic.c 	struct netdev_queue *txq;
txq              1454 drivers/net/ethernet/ibm/ibmvnic.c 	txq = netdev_get_tx_queue(netdev, skb_get_queue_mapping(skb));
txq              1608 drivers/net/ethernet/ibm/ibmvnic.c 	txq->trans_start = jiffies;
txq              2130 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c 						      qpi->txq.queue_id)) {
txq              2135 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c 			vsi_queue_id = qpi->txq.queue_id;
txq              2137 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c 			if (qpi->txq.vsi_id != qci->vsi_id ||
txq              2156 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c 					     &qpi->txq)) {
txq               272 drivers/net/ethernet/intel/iavf/iavf_virtchnl.c 		vqpi->txq.vsi_id = vqci->vsi_id;
txq               273 drivers/net/ethernet/intel/iavf/iavf_virtchnl.c 		vqpi->txq.queue_id = i;
txq               274 drivers/net/ethernet/intel/iavf/iavf_virtchnl.c 		vqpi->txq.ring_len = adapter->tx_rings[i].count;
txq               275 drivers/net/ethernet/intel/iavf/iavf_virtchnl.c 		vqpi->txq.dma_ring_addr = adapter->tx_rings[i].dma;
txq              1727 drivers/net/ethernet/intel/ice/ice_lib.c 	struct ice_aqc_add_txqs_perq *txq;
txq              1763 drivers/net/ethernet/intel/ice/ice_lib.c 	txq = &qg_buf->txqs[0];
txq              1764 drivers/net/ethernet/intel/ice/ice_lib.c 	if (pf_q == le16_to_cpu(txq->txq_id))
txq              1765 drivers/net/ethernet/intel/ice/ice_lib.c 		ring->txq_teid = le32_to_cpu(txq->q_teid);
txq              1927 drivers/net/ethernet/intel/ice/ice_lib.c ice_cfg_txq_interrupt(struct ice_vsi *vsi, u16 txq, u16 msix_idx, u16 itr_idx)
txq              1930 drivers/net/ethernet/intel/ice/ice_lib.c ice_cfg_txq_interrupt(struct ice_vsi *vsi, u16 txq, u16 msix_idx, u16 itr_idx)
txq              1942 drivers/net/ethernet/intel/ice/ice_lib.c 	wr32(hw, QINT_TQCTL(vsi->txq_map[txq]), val);
txq              1988 drivers/net/ethernet/intel/ice/ice_lib.c 	u32 txq = 0, rxq = 0;
txq              2012 drivers/net/ethernet/intel/ice/ice_lib.c 			ice_cfg_txq_interrupt(vsi, txq, reg_idx,
txq              2014 drivers/net/ethernet/intel/ice/ice_lib.c 			txq++;
txq              2681 drivers/net/ethernet/intel/ice/ice_lib.c 	u32 txq = 0;
txq              2692 drivers/net/ethernet/intel/ice/ice_lib.c 			wr32(hw, QINT_TQCTL(vsi->txq_map[txq]), 0);
txq              2693 drivers/net/ethernet/intel/ice/ice_lib.c 			txq++;
txq                38 drivers/net/ethernet/intel/ice/ice_lib.h ice_cfg_txq_interrupt(struct ice_vsi *vsi, u16 txq, u16 msix_idx, u16 itr_idx);
txq              2267 drivers/net/ethernet/intel/ice/ice_virtchnl_pf.c 		if (qpi->txq.vsi_id != qci->vsi_id ||
txq              2269 drivers/net/ethernet/intel/ice/ice_virtchnl_pf.c 		    qpi->rxq.queue_id != qpi->txq.queue_id ||
txq              2270 drivers/net/ethernet/intel/ice/ice_virtchnl_pf.c 		    qpi->txq.headwb_enabled ||
txq              2271 drivers/net/ethernet/intel/ice/ice_virtchnl_pf.c 		    !ice_vc_isvalid_ring_len(qpi->txq.ring_len) ||
txq              2273 drivers/net/ethernet/intel/ice/ice_virtchnl_pf.c 		    !ice_vc_isvalid_q_id(vf, qci->vsi_id, qpi->txq.queue_id)) {
txq              2278 drivers/net/ethernet/intel/ice/ice_virtchnl_pf.c 		if (qpi->txq.ring_len > 0) {
txq              2280 drivers/net/ethernet/intel/ice/ice_virtchnl_pf.c 			vsi->tx_rings[i]->dma = qpi->txq.dma_ring_addr;
txq              2281 drivers/net/ethernet/intel/ice/ice_virtchnl_pf.c 			vsi->tx_rings[i]->count = qpi->txq.ring_len;
txq              8507 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 	int txq;
txq              8513 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 		txq = vdev->tc_to_txq[tc].offset;
txq              8514 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 		txq += reciprocal_scale(skb_get_hash(skb),
txq              8517 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 		return txq;
txq              8538 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 	txq = skb_rx_queue_recorded(skb) ? skb_get_rx_queue(skb) :
txq              8541 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 	while (txq >= f->indices)
txq              8542 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 		txq -= f->indices;
txq              8544 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 	return txq + f->offset;
txq               174 drivers/net/ethernet/lantiq_etop.c 	struct netdev_queue *txq =
txq               190 drivers/net/ethernet/lantiq_etop.c 	if (netif_tx_queue_stopped(txq))
txq               191 drivers/net/ethernet/lantiq_etop.c 		netif_tx_start_queue(txq);
txq               458 drivers/net/ethernet/lantiq_etop.c 	struct netdev_queue *txq = netdev_get_tx_queue(dev, queue);
txq               471 drivers/net/ethernet/lantiq_etop.c 		netif_tx_stop_queue(txq);
txq               492 drivers/net/ethernet/lantiq_etop.c 		netif_tx_stop_queue(txq);
txq               178 drivers/net/ethernet/marvell/mv643xx_eth.c #define IS_TSO_HEADER(txq, addr) \
txq               179 drivers/net/ethernet/marvell/mv643xx_eth.c 	((addr >= txq->tso_hdrs_dma) && \
txq               180 drivers/net/ethernet/marvell/mv643xx_eth.c 	 (addr < txq->tso_hdrs_dma + txq->tx_ring_size * TSO_HEADER_SIZE))
txq               407 drivers/net/ethernet/marvell/mv643xx_eth.c 	struct tx_queue txq[8];
txq               445 drivers/net/ethernet/marvell/mv643xx_eth.c static struct mv643xx_eth_private *txq_to_mp(struct tx_queue *txq)
txq               447 drivers/net/ethernet/marvell/mv643xx_eth.c 	return container_of(txq, struct mv643xx_eth_private, txq[txq->index]);
txq               466 drivers/net/ethernet/marvell/mv643xx_eth.c static void txq_reset_hw_ptr(struct tx_queue *txq)
txq               468 drivers/net/ethernet/marvell/mv643xx_eth.c 	struct mv643xx_eth_private *mp = txq_to_mp(txq);
txq               471 drivers/net/ethernet/marvell/mv643xx_eth.c 	addr = (u32)txq->tx_desc_dma;
txq               472 drivers/net/ethernet/marvell/mv643xx_eth.c 	addr += txq->tx_curr_desc * sizeof(struct tx_desc);
txq               473 drivers/net/ethernet/marvell/mv643xx_eth.c 	wrlp(mp, TXQ_CURRENT_DESC_PTR(txq->index), addr);
txq               476 drivers/net/ethernet/marvell/mv643xx_eth.c static void txq_enable(struct tx_queue *txq)
txq               478 drivers/net/ethernet/marvell/mv643xx_eth.c 	struct mv643xx_eth_private *mp = txq_to_mp(txq);
txq               479 drivers/net/ethernet/marvell/mv643xx_eth.c 	wrlp(mp, TXQ_COMMAND, 1 << txq->index);
txq               482 drivers/net/ethernet/marvell/mv643xx_eth.c static void txq_disable(struct tx_queue *txq)
txq               484 drivers/net/ethernet/marvell/mv643xx_eth.c 	struct mv643xx_eth_private *mp = txq_to_mp(txq);
txq               485 drivers/net/ethernet/marvell/mv643xx_eth.c 	u8 mask = 1 << txq->index;
txq               492 drivers/net/ethernet/marvell/mv643xx_eth.c static void txq_maybe_wake(struct tx_queue *txq)
txq               494 drivers/net/ethernet/marvell/mv643xx_eth.c 	struct mv643xx_eth_private *mp = txq_to_mp(txq);
txq               495 drivers/net/ethernet/marvell/mv643xx_eth.c 	struct netdev_queue *nq = netdev_get_tx_queue(mp->dev, txq->index);
txq               499 drivers/net/ethernet/marvell/mv643xx_eth.c 		if (txq->tx_desc_count <= txq->tx_wake_threshold)
txq               730 drivers/net/ethernet/marvell/mv643xx_eth.c txq_put_data_tso(struct net_device *dev, struct tx_queue *txq,
txq               738 drivers/net/ethernet/marvell/mv643xx_eth.c 	tx_index = txq->tx_curr_desc++;
txq               739 drivers/net/ethernet/marvell/mv643xx_eth.c 	if (txq->tx_curr_desc == txq->tx_ring_size)
txq               740 drivers/net/ethernet/marvell/mv643xx_eth.c 		txq->tx_curr_desc = 0;
txq               741 drivers/net/ethernet/marvell/mv643xx_eth.c 	desc = &txq->tx_desc_area[tx_index];
txq               742 drivers/net/ethernet/marvell/mv643xx_eth.c 	txq->tx_desc_mapping[tx_index] = DESC_DMA_MAP_SINGLE;
txq               749 drivers/net/ethernet/marvell/mv643xx_eth.c 		memcpy(txq->tso_hdrs + tx_index * TSO_HEADER_SIZE,
txq               751 drivers/net/ethernet/marvell/mv643xx_eth.c 		desc->buf_ptr = txq->tso_hdrs_dma
txq               755 drivers/net/ethernet/marvell/mv643xx_eth.c 		txq->tx_desc_mapping[tx_index] = DESC_DMA_MAP_SINGLE;
txq               778 drivers/net/ethernet/marvell/mv643xx_eth.c txq_put_hdr_tso(struct sk_buff *skb, struct tx_queue *txq, int length,
txq               781 drivers/net/ethernet/marvell/mv643xx_eth.c 	struct mv643xx_eth_private *mp = txq_to_mp(txq);
txq               790 drivers/net/ethernet/marvell/mv643xx_eth.c 	tx_index = txq->tx_curr_desc;
txq               791 drivers/net/ethernet/marvell/mv643xx_eth.c 	desc = &txq->tx_desc_area[tx_index];
txq               802 drivers/net/ethernet/marvell/mv643xx_eth.c 	desc->buf_ptr = txq->tso_hdrs_dma +
txq               803 drivers/net/ethernet/marvell/mv643xx_eth.c 			txq->tx_curr_desc * TSO_HEADER_SIZE;
txq               815 drivers/net/ethernet/marvell/mv643xx_eth.c 	txq->tx_curr_desc++;
txq               816 drivers/net/ethernet/marvell/mv643xx_eth.c 	if (txq->tx_curr_desc == txq->tx_ring_size)
txq               817 drivers/net/ethernet/marvell/mv643xx_eth.c 		txq->tx_curr_desc = 0;
txq               820 drivers/net/ethernet/marvell/mv643xx_eth.c static int txq_submit_tso(struct tx_queue *txq, struct sk_buff *skb,
txq               823 drivers/net/ethernet/marvell/mv643xx_eth.c 	struct mv643xx_eth_private *mp = txq_to_mp(txq);
txq               832 drivers/net/ethernet/marvell/mv643xx_eth.c 	if ((txq->tx_desc_count + tso_count_descs(skb)) >= txq->tx_ring_size) {
txq               837 drivers/net/ethernet/marvell/mv643xx_eth.c 	first_tx_desc = &txq->tx_desc_area[txq->tx_curr_desc];
txq               852 drivers/net/ethernet/marvell/mv643xx_eth.c 		hdr = txq->tso_hdrs + txq->tx_curr_desc * TSO_HEADER_SIZE;
txq               854 drivers/net/ethernet/marvell/mv643xx_eth.c 		txq_put_hdr_tso(skb, txq, data_left, &first_cmd_sts,
txq               862 drivers/net/ethernet/marvell/mv643xx_eth.c 			ret = txq_put_data_tso(dev, txq, skb, tso.data, size,
txq               872 drivers/net/ethernet/marvell/mv643xx_eth.c 	__skb_queue_tail(&txq->tx_skb, skb);
txq               880 drivers/net/ethernet/marvell/mv643xx_eth.c 	mp->work_tx_end &= ~(1 << txq->index);
txq               884 drivers/net/ethernet/marvell/mv643xx_eth.c 	txq_enable(txq);
txq               885 drivers/net/ethernet/marvell/mv643xx_eth.c 	txq->tx_desc_count += desc_count;
txq               894 drivers/net/ethernet/marvell/mv643xx_eth.c static void txq_submit_frag_skb(struct tx_queue *txq, struct sk_buff *skb)
txq               896 drivers/net/ethernet/marvell/mv643xx_eth.c 	struct mv643xx_eth_private *mp = txq_to_mp(txq);
txq               906 drivers/net/ethernet/marvell/mv643xx_eth.c 		tx_index = txq->tx_curr_desc++;
txq               907 drivers/net/ethernet/marvell/mv643xx_eth.c 		if (txq->tx_curr_desc == txq->tx_ring_size)
txq               908 drivers/net/ethernet/marvell/mv643xx_eth.c 			txq->tx_curr_desc = 0;
txq               909 drivers/net/ethernet/marvell/mv643xx_eth.c 		desc = &txq->tx_desc_area[tx_index];
txq               910 drivers/net/ethernet/marvell/mv643xx_eth.c 		txq->tx_desc_mapping[tx_index] = DESC_DMA_MAP_PAGE;
txq               932 drivers/net/ethernet/marvell/mv643xx_eth.c static int txq_submit_skb(struct tx_queue *txq, struct sk_buff *skb,
txq               935 drivers/net/ethernet/marvell/mv643xx_eth.c 	struct mv643xx_eth_private *mp = txq_to_mp(txq);
txq               946 drivers/net/ethernet/marvell/mv643xx_eth.c 	if (txq->tx_ring_size - txq->tx_desc_count < MAX_SKB_FRAGS + 1) {
txq               957 drivers/net/ethernet/marvell/mv643xx_eth.c 	tx_index = txq->tx_curr_desc++;
txq               958 drivers/net/ethernet/marvell/mv643xx_eth.c 	if (txq->tx_curr_desc == txq->tx_ring_size)
txq               959 drivers/net/ethernet/marvell/mv643xx_eth.c 		txq->tx_curr_desc = 0;
txq               960 drivers/net/ethernet/marvell/mv643xx_eth.c 	desc = &txq->tx_desc_area[tx_index];
txq               961 drivers/net/ethernet/marvell/mv643xx_eth.c 	txq->tx_desc_mapping[tx_index] = DESC_DMA_MAP_SINGLE;
txq               964 drivers/net/ethernet/marvell/mv643xx_eth.c 		txq_submit_frag_skb(txq, skb);
txq               976 drivers/net/ethernet/marvell/mv643xx_eth.c 	__skb_queue_tail(&txq->tx_skb, skb);
txq               985 drivers/net/ethernet/marvell/mv643xx_eth.c 	mp->work_tx_end &= ~(1 << txq->index);
txq               989 drivers/net/ethernet/marvell/mv643xx_eth.c 	txq_enable(txq);
txq               991 drivers/net/ethernet/marvell/mv643xx_eth.c 	txq->tx_desc_count += nr_frags + 1;
txq              1000 drivers/net/ethernet/marvell/mv643xx_eth.c 	struct tx_queue *txq;
txq              1004 drivers/net/ethernet/marvell/mv643xx_eth.c 	txq = mp->txq + queue;
txq              1016 drivers/net/ethernet/marvell/mv643xx_eth.c 		ret = txq_submit_tso(txq, skb, dev);
txq              1018 drivers/net/ethernet/marvell/mv643xx_eth.c 		ret = txq_submit_skb(txq, skb, dev);
txq              1020 drivers/net/ethernet/marvell/mv643xx_eth.c 		txq->tx_bytes += length;
txq              1021 drivers/net/ethernet/marvell/mv643xx_eth.c 		txq->tx_packets++;
txq              1023 drivers/net/ethernet/marvell/mv643xx_eth.c 		if (txq->tx_desc_count >= txq->tx_stop_threshold)
txq              1026 drivers/net/ethernet/marvell/mv643xx_eth.c 		txq->tx_dropped++;
txq              1035 drivers/net/ethernet/marvell/mv643xx_eth.c static void txq_kick(struct tx_queue *txq)
txq              1037 drivers/net/ethernet/marvell/mv643xx_eth.c 	struct mv643xx_eth_private *mp = txq_to_mp(txq);
txq              1038 drivers/net/ethernet/marvell/mv643xx_eth.c 	struct netdev_queue *nq = netdev_get_tx_queue(mp->dev, txq->index);
txq              1044 drivers/net/ethernet/marvell/mv643xx_eth.c 	if (rdlp(mp, TXQ_COMMAND) & (1 << txq->index))
txq              1047 drivers/net/ethernet/marvell/mv643xx_eth.c 	hw_desc_ptr = rdlp(mp, TXQ_CURRENT_DESC_PTR(txq->index));
txq              1048 drivers/net/ethernet/marvell/mv643xx_eth.c 	expected_ptr = (u32)txq->tx_desc_dma +
txq              1049 drivers/net/ethernet/marvell/mv643xx_eth.c 				txq->tx_curr_desc * sizeof(struct tx_desc);
txq              1052 drivers/net/ethernet/marvell/mv643xx_eth.c 		txq_enable(txq);
txq              1057 drivers/net/ethernet/marvell/mv643xx_eth.c 	mp->work_tx_end &= ~(1 << txq->index);
txq              1060 drivers/net/ethernet/marvell/mv643xx_eth.c static int txq_reclaim(struct tx_queue *txq, int budget, int force)
txq              1062 drivers/net/ethernet/marvell/mv643xx_eth.c 	struct mv643xx_eth_private *mp = txq_to_mp(txq);
txq              1063 drivers/net/ethernet/marvell/mv643xx_eth.c 	struct netdev_queue *nq = netdev_get_tx_queue(mp->dev, txq->index);
txq              1069 drivers/net/ethernet/marvell/mv643xx_eth.c 	while (reclaimed < budget && txq->tx_desc_count > 0) {
txq              1075 drivers/net/ethernet/marvell/mv643xx_eth.c 		tx_index = txq->tx_used_desc;
txq              1076 drivers/net/ethernet/marvell/mv643xx_eth.c 		desc = &txq->tx_desc_area[tx_index];
txq              1077 drivers/net/ethernet/marvell/mv643xx_eth.c 		desc_dma_map = txq->tx_desc_mapping[tx_index];
txq              1087 drivers/net/ethernet/marvell/mv643xx_eth.c 		txq->tx_used_desc = tx_index + 1;
txq              1088 drivers/net/ethernet/marvell/mv643xx_eth.c 		if (txq->tx_used_desc == txq->tx_ring_size)
txq              1089 drivers/net/ethernet/marvell/mv643xx_eth.c 			txq->tx_used_desc = 0;
txq              1092 drivers/net/ethernet/marvell/mv643xx_eth.c 		txq->tx_desc_count--;
txq              1094 drivers/net/ethernet/marvell/mv643xx_eth.c 		if (!IS_TSO_HEADER(txq, desc->buf_ptr)) {
txq              1109 drivers/net/ethernet/marvell/mv643xx_eth.c 			struct sk_buff *skb = __skb_dequeue(&txq->tx_skb);
txq              1125 drivers/net/ethernet/marvell/mv643xx_eth.c 		mp->work_tx &= ~(1 << txq->index);
txq              1168 drivers/net/ethernet/marvell/mv643xx_eth.c static void txq_set_rate(struct tx_queue *txq, int rate, int burst)
txq              1170 drivers/net/ethernet/marvell/mv643xx_eth.c 	struct mv643xx_eth_private *mp = txq_to_mp(txq);
txq              1182 drivers/net/ethernet/marvell/mv643xx_eth.c 	wrlp(mp, TXQ_BW_TOKENS(txq->index), token_rate << 14);
txq              1183 drivers/net/ethernet/marvell/mv643xx_eth.c 	wrlp(mp, TXQ_BW_CONF(txq->index), (bucket_size << 10) | token_rate);
txq              1186 drivers/net/ethernet/marvell/mv643xx_eth.c static void txq_set_fixed_prio_mode(struct tx_queue *txq)
txq              1188 drivers/net/ethernet/marvell/mv643xx_eth.c 	struct mv643xx_eth_private *mp = txq_to_mp(txq);
txq              1207 drivers/net/ethernet/marvell/mv643xx_eth.c 		val |= 1 << txq->index;
txq              1265 drivers/net/ethernet/marvell/mv643xx_eth.c 		struct tx_queue *txq = mp->txq + i;
txq              1267 drivers/net/ethernet/marvell/mv643xx_eth.c 		tx_packets += txq->tx_packets;
txq              1268 drivers/net/ethernet/marvell/mv643xx_eth.c 		tx_bytes += txq->tx_bytes;
txq              1269 drivers/net/ethernet/marvell/mv643xx_eth.c 		tx_dropped += txq->tx_dropped;
txq              2030 drivers/net/ethernet/marvell/mv643xx_eth.c 	struct tx_queue *txq = mp->txq + index;
txq              2036 drivers/net/ethernet/marvell/mv643xx_eth.c 	txq->index = index;
txq              2038 drivers/net/ethernet/marvell/mv643xx_eth.c 	txq->tx_ring_size = mp->tx_ring_size;
txq              2044 drivers/net/ethernet/marvell/mv643xx_eth.c 	txq->tx_stop_threshold = txq->tx_ring_size - MV643XX_MAX_SKB_DESCS;
txq              2045 drivers/net/ethernet/marvell/mv643xx_eth.c 	txq->tx_wake_threshold = txq->tx_stop_threshold / 2;
txq              2047 drivers/net/ethernet/marvell/mv643xx_eth.c 	txq->tx_desc_count = 0;
txq              2048 drivers/net/ethernet/marvell/mv643xx_eth.c 	txq->tx_curr_desc = 0;
txq              2049 drivers/net/ethernet/marvell/mv643xx_eth.c 	txq->tx_used_desc = 0;
txq              2051 drivers/net/ethernet/marvell/mv643xx_eth.c 	size = txq->tx_ring_size * sizeof(struct tx_desc);
txq              2054 drivers/net/ethernet/marvell/mv643xx_eth.c 		txq->tx_desc_area = ioremap(mp->tx_desc_sram_addr,
txq              2056 drivers/net/ethernet/marvell/mv643xx_eth.c 		txq->tx_desc_dma = mp->tx_desc_sram_addr;
txq              2058 drivers/net/ethernet/marvell/mv643xx_eth.c 		txq->tx_desc_area = dma_alloc_coherent(mp->dev->dev.parent,
txq              2059 drivers/net/ethernet/marvell/mv643xx_eth.c 						       size, &txq->tx_desc_dma,
txq              2063 drivers/net/ethernet/marvell/mv643xx_eth.c 	if (txq->tx_desc_area == NULL) {
txq              2068 drivers/net/ethernet/marvell/mv643xx_eth.c 	memset(txq->tx_desc_area, 0, size);
txq              2070 drivers/net/ethernet/marvell/mv643xx_eth.c 	txq->tx_desc_area_size = size;
txq              2072 drivers/net/ethernet/marvell/mv643xx_eth.c 	tx_desc = txq->tx_desc_area;
txq              2073 drivers/net/ethernet/marvell/mv643xx_eth.c 	for (i = 0; i < txq->tx_ring_size; i++) {
txq              2078 drivers/net/ethernet/marvell/mv643xx_eth.c 		if (nexti == txq->tx_ring_size)
txq              2082 drivers/net/ethernet/marvell/mv643xx_eth.c 		txd->next_desc_ptr = txq->tx_desc_dma +
txq              2086 drivers/net/ethernet/marvell/mv643xx_eth.c 	txq->tx_desc_mapping = kcalloc(txq->tx_ring_size, sizeof(char),
txq              2088 drivers/net/ethernet/marvell/mv643xx_eth.c 	if (!txq->tx_desc_mapping) {
txq              2094 drivers/net/ethernet/marvell/mv643xx_eth.c 	txq->tso_hdrs = dma_alloc_coherent(mp->dev->dev.parent,
txq              2095 drivers/net/ethernet/marvell/mv643xx_eth.c 					   txq->tx_ring_size * TSO_HEADER_SIZE,
txq              2096 drivers/net/ethernet/marvell/mv643xx_eth.c 					   &txq->tso_hdrs_dma, GFP_KERNEL);
txq              2097 drivers/net/ethernet/marvell/mv643xx_eth.c 	if (txq->tso_hdrs == NULL) {
txq              2101 drivers/net/ethernet/marvell/mv643xx_eth.c 	skb_queue_head_init(&txq->tx_skb);
txq              2106 drivers/net/ethernet/marvell/mv643xx_eth.c 	kfree(txq->tx_desc_mapping);
txq              2109 drivers/net/ethernet/marvell/mv643xx_eth.c 		iounmap(txq->tx_desc_area);
txq              2111 drivers/net/ethernet/marvell/mv643xx_eth.c 		dma_free_coherent(mp->dev->dev.parent, txq->tx_desc_area_size,
txq              2112 drivers/net/ethernet/marvell/mv643xx_eth.c 				  txq->tx_desc_area, txq->tx_desc_dma);
txq              2116 drivers/net/ethernet/marvell/mv643xx_eth.c static void txq_deinit(struct tx_queue *txq)
txq              2118 drivers/net/ethernet/marvell/mv643xx_eth.c 	struct mv643xx_eth_private *mp = txq_to_mp(txq);
txq              2120 drivers/net/ethernet/marvell/mv643xx_eth.c 	txq_disable(txq);
txq              2121 drivers/net/ethernet/marvell/mv643xx_eth.c 	txq_reclaim(txq, txq->tx_ring_size, 1);
txq              2123 drivers/net/ethernet/marvell/mv643xx_eth.c 	BUG_ON(txq->tx_used_desc != txq->tx_curr_desc);
txq              2125 drivers/net/ethernet/marvell/mv643xx_eth.c 	if (txq->index == 0 &&
txq              2126 drivers/net/ethernet/marvell/mv643xx_eth.c 	    txq->tx_desc_area_size <= mp->tx_desc_sram_size)
txq              2127 drivers/net/ethernet/marvell/mv643xx_eth.c 		iounmap(txq->tx_desc_area);
txq              2129 drivers/net/ethernet/marvell/mv643xx_eth.c 		dma_free_coherent(mp->dev->dev.parent, txq->tx_desc_area_size,
txq              2130 drivers/net/ethernet/marvell/mv643xx_eth.c 				  txq->tx_desc_area, txq->tx_desc_dma);
txq              2131 drivers/net/ethernet/marvell/mv643xx_eth.c 	kfree(txq->tx_desc_mapping);
txq              2133 drivers/net/ethernet/marvell/mv643xx_eth.c 	if (txq->tso_hdrs)
txq              2135 drivers/net/ethernet/marvell/mv643xx_eth.c 				  txq->tx_ring_size * TSO_HEADER_SIZE,
txq              2136 drivers/net/ethernet/marvell/mv643xx_eth.c 				  txq->tso_hdrs, txq->tso_hdrs_dma);
txq              2206 drivers/net/ethernet/marvell/mv643xx_eth.c 				struct tx_queue *txq = mp->txq + i;
txq              2208 drivers/net/ethernet/marvell/mv643xx_eth.c 				txq_reclaim(txq, txq->tx_ring_size, 1);
txq              2209 drivers/net/ethernet/marvell/mv643xx_eth.c 				txq_reset_hw_ptr(txq);
txq              2282 drivers/net/ethernet/marvell/mv643xx_eth.c 			txq_kick(mp->txq + queue);
txq              2284 drivers/net/ethernet/marvell/mv643xx_eth.c 			work_done += txq_reclaim(mp->txq + queue, work_tbd, 0);
txq              2285 drivers/net/ethernet/marvell/mv643xx_eth.c 			txq_maybe_wake(mp->txq + queue);
txq              2349 drivers/net/ethernet/marvell/mv643xx_eth.c 		struct tx_queue *txq = mp->txq + i;
txq              2351 drivers/net/ethernet/marvell/mv643xx_eth.c 		txq_reset_hw_ptr(txq);
txq              2352 drivers/net/ethernet/marvell/mv643xx_eth.c 		txq_set_rate(txq, 1000000000, 16777216);
txq              2353 drivers/net/ethernet/marvell/mv643xx_eth.c 		txq_set_fixed_prio_mode(txq);
txq              2460 drivers/net/ethernet/marvell/mv643xx_eth.c 				txq_deinit(mp->txq + i);
txq              2492 drivers/net/ethernet/marvell/mv643xx_eth.c 		txq_disable(mp->txq + i);
txq              2536 drivers/net/ethernet/marvell/mv643xx_eth.c 		txq_deinit(mp->txq + i);
txq               126 drivers/net/ethernet/marvell/mvneta.c #define      MVNETA_CPU_TXQ_ACCESS(txq)		 BIT(txq + 8)
txq               325 drivers/net/ethernet/marvell/mvneta.c #define IS_TSO_HEADER(txq, addr) \
txq               326 drivers/net/ethernet/marvell/mvneta.c 	((addr >= txq->tso_hdrs_phys) && \
txq               327 drivers/net/ethernet/marvell/mvneta.c 	 (addr < txq->tso_hdrs_phys + txq->size * TSO_HEADER_SIZE))
txq               669 drivers/net/ethernet/marvell/mvneta.c static void mvneta_txq_inc_get(struct mvneta_tx_queue *txq)
txq               671 drivers/net/ethernet/marvell/mvneta.c 	txq->txq_get_index++;
txq               672 drivers/net/ethernet/marvell/mvneta.c 	if (txq->txq_get_index == txq->size)
txq               673 drivers/net/ethernet/marvell/mvneta.c 		txq->txq_get_index = 0;
txq               677 drivers/net/ethernet/marvell/mvneta.c static void mvneta_txq_inc_put(struct mvneta_tx_queue *txq)
txq               679 drivers/net/ethernet/marvell/mvneta.c 	txq->txq_put_index++;
txq               680 drivers/net/ethernet/marvell/mvneta.c 	if (txq->txq_put_index == txq->size)
txq               681 drivers/net/ethernet/marvell/mvneta.c 		txq->txq_put_index = 0;
txq               860 drivers/net/ethernet/marvell/mvneta.c 				     struct mvneta_tx_queue *txq,
txq               865 drivers/net/ethernet/marvell/mvneta.c 	pend_desc += txq->pending;
txq               870 drivers/net/ethernet/marvell/mvneta.c 		mvreg_write(pp, MVNETA_TXQ_UPDATE_REG(txq->id), val);
txq               873 drivers/net/ethernet/marvell/mvneta.c 	txq->pending = 0;
txq               878 drivers/net/ethernet/marvell/mvneta.c mvneta_txq_next_desc_get(struct mvneta_tx_queue *txq)
txq               880 drivers/net/ethernet/marvell/mvneta.c 	int tx_desc = txq->next_desc_to_proc;
txq               882 drivers/net/ethernet/marvell/mvneta.c 	txq->next_desc_to_proc = MVNETA_QUEUE_NEXT_DESC(txq, tx_desc);
txq               883 drivers/net/ethernet/marvell/mvneta.c 	return txq->descs + tx_desc;
txq               889 drivers/net/ethernet/marvell/mvneta.c static void mvneta_txq_desc_put(struct mvneta_tx_queue *txq)
txq               891 drivers/net/ethernet/marvell/mvneta.c 	if (txq->next_desc_to_proc == 0)
txq               892 drivers/net/ethernet/marvell/mvneta.c 		txq->next_desc_to_proc = txq->last_desc - 1;
txq               894 drivers/net/ethernet/marvell/mvneta.c 		txq->next_desc_to_proc--;
txq              1155 drivers/net/ethernet/marvell/mvneta.c 		struct mvneta_tx_queue *txq = &pp->txqs[queue];
txq              1156 drivers/net/ethernet/marvell/mvneta.c 		if (txq->descs)
txq              1393 drivers/net/ethernet/marvell/mvneta.c 		int rxq, txq;
txq              1399 drivers/net/ethernet/marvell/mvneta.c 			for (txq = 0; txq < txq_number; txq++)
txq              1400 drivers/net/ethernet/marvell/mvneta.c 				if ((txq % max_cpu) == cpu)
txq              1401 drivers/net/ethernet/marvell/mvneta.c 					txq_map |= MVNETA_CPU_TXQ_ACCESS(txq);
txq              1604 drivers/net/ethernet/marvell/mvneta.c 					 struct mvneta_tx_queue *txq, u32 value)
txq              1608 drivers/net/ethernet/marvell/mvneta.c 	val = mvreg_read(pp, MVNETA_TXQ_SIZE_REG(txq->id));
txq              1613 drivers/net/ethernet/marvell/mvneta.c 	mvreg_write(pp, MVNETA_TXQ_SIZE_REG(txq->id), val);
txq              1630 drivers/net/ethernet/marvell/mvneta.c 				     struct mvneta_tx_queue *txq,
txq              1638 drivers/net/ethernet/marvell/mvneta.c 		mvreg_write(pp, MVNETA_TXQ_UPDATE_REG(txq->id), val);
txq              1643 drivers/net/ethernet/marvell/mvneta.c 	mvreg_write(pp, MVNETA_TXQ_UPDATE_REG(txq->id), val);
txq              1648 drivers/net/ethernet/marvell/mvneta.c 					struct mvneta_tx_queue *txq)
txq              1653 drivers/net/ethernet/marvell/mvneta.c 	val = mvreg_read(pp, MVNETA_TXQ_STATUS_REG(txq->id));
txq              1664 drivers/net/ethernet/marvell/mvneta.c 				     struct mvneta_tx_queue *txq)
txq              1669 drivers/net/ethernet/marvell/mvneta.c 	sent_desc = mvneta_txq_sent_desc_num_get(pp, txq);
txq              1673 drivers/net/ethernet/marvell/mvneta.c 		mvneta_txq_sent_desc_dec(pp, txq, sent_desc);
txq              1768 drivers/net/ethernet/marvell/mvneta.c 				 struct mvneta_tx_queue *txq, int num,
txq              1775 drivers/net/ethernet/marvell/mvneta.c 		struct mvneta_tx_desc *tx_desc = txq->descs +
txq              1776 drivers/net/ethernet/marvell/mvneta.c 			txq->txq_get_index;
txq              1777 drivers/net/ethernet/marvell/mvneta.c 		struct sk_buff *skb = txq->tx_skb[txq->txq_get_index];
txq              1784 drivers/net/ethernet/marvell/mvneta.c 		mvneta_txq_inc_get(txq);
txq              1786 drivers/net/ethernet/marvell/mvneta.c 		if (!IS_TSO_HEADER(txq, tx_desc->buf_phys_addr))
txq              1800 drivers/net/ethernet/marvell/mvneta.c 			   struct mvneta_tx_queue *txq)
txq              1802 drivers/net/ethernet/marvell/mvneta.c 	struct netdev_queue *nq = netdev_get_tx_queue(pp->dev, txq->id);
txq              1805 drivers/net/ethernet/marvell/mvneta.c 	tx_done = mvneta_txq_sent_desc_proc(pp, txq);
txq              1809 drivers/net/ethernet/marvell/mvneta.c 	mvneta_txq_bufs_free(pp, txq, tx_done, nq);
txq              1811 drivers/net/ethernet/marvell/mvneta.c 	txq->count -= tx_done;
txq              1814 drivers/net/ethernet/marvell/mvneta.c 		if (txq->count <= txq->tx_wake_threshold)
txq              2241 drivers/net/ethernet/marvell/mvneta.c 		   struct mvneta_port *pp, struct mvneta_tx_queue *txq)
txq              2246 drivers/net/ethernet/marvell/mvneta.c 	txq->tx_skb[txq->txq_put_index] = NULL;
txq              2247 drivers/net/ethernet/marvell/mvneta.c 	tx_desc = mvneta_txq_next_desc_get(txq);
txq              2251 drivers/net/ethernet/marvell/mvneta.c 	tx_desc->buf_phys_addr = txq->tso_hdrs_phys +
txq              2252 drivers/net/ethernet/marvell/mvneta.c 				 txq->txq_put_index * TSO_HEADER_SIZE;
txq              2253 drivers/net/ethernet/marvell/mvneta.c 	mvneta_txq_inc_put(txq);
txq              2257 drivers/net/ethernet/marvell/mvneta.c mvneta_tso_put_data(struct net_device *dev, struct mvneta_tx_queue *txq,
txq              2263 drivers/net/ethernet/marvell/mvneta.c 	tx_desc = mvneta_txq_next_desc_get(txq);
txq              2269 drivers/net/ethernet/marvell/mvneta.c 		mvneta_txq_desc_put(txq);
txq              2274 drivers/net/ethernet/marvell/mvneta.c 	txq->tx_skb[txq->txq_put_index] = NULL;
txq              2282 drivers/net/ethernet/marvell/mvneta.c 			txq->tx_skb[txq->txq_put_index] = skb;
txq              2284 drivers/net/ethernet/marvell/mvneta.c 	mvneta_txq_inc_put(txq);
txq              2289 drivers/net/ethernet/marvell/mvneta.c 			 struct mvneta_tx_queue *txq)
txq              2299 drivers/net/ethernet/marvell/mvneta.c 	if ((txq->count + tso_count_descs(skb)) >= txq->size)
txq              2319 drivers/net/ethernet/marvell/mvneta.c 		hdr = txq->tso_hdrs + txq->txq_put_index * TSO_HEADER_SIZE;
txq              2322 drivers/net/ethernet/marvell/mvneta.c 		mvneta_tso_put_hdr(skb, pp, txq);
txq              2330 drivers/net/ethernet/marvell/mvneta.c 			if (mvneta_tso_put_data(dev, txq, skb,
txq              2348 drivers/net/ethernet/marvell/mvneta.c 		struct mvneta_tx_desc *tx_desc = txq->descs + i;
txq              2349 drivers/net/ethernet/marvell/mvneta.c 		if (!IS_TSO_HEADER(txq, tx_desc->buf_phys_addr))
txq              2354 drivers/net/ethernet/marvell/mvneta.c 		mvneta_txq_desc_put(txq);
txq              2361 drivers/net/ethernet/marvell/mvneta.c 				  struct mvneta_tx_queue *txq)
txq              2370 drivers/net/ethernet/marvell/mvneta.c 		tx_desc = mvneta_txq_next_desc_get(txq);
txq              2379 drivers/net/ethernet/marvell/mvneta.c 			mvneta_txq_desc_put(txq);
txq              2386 drivers/net/ethernet/marvell/mvneta.c 			txq->tx_skb[txq->txq_put_index] = skb;
txq              2390 drivers/net/ethernet/marvell/mvneta.c 			txq->tx_skb[txq->txq_put_index] = NULL;
txq              2392 drivers/net/ethernet/marvell/mvneta.c 		mvneta_txq_inc_put(txq);
txq              2402 drivers/net/ethernet/marvell/mvneta.c 		tx_desc = txq->descs + i;
txq              2407 drivers/net/ethernet/marvell/mvneta.c 		mvneta_txq_desc_put(txq);
txq              2418 drivers/net/ethernet/marvell/mvneta.c 	struct mvneta_tx_queue *txq = &pp->txqs[txq_id];
txq              2428 drivers/net/ethernet/marvell/mvneta.c 		frags = mvneta_tx_tso(skb, dev, txq);
txq              2435 drivers/net/ethernet/marvell/mvneta.c 	tx_desc = mvneta_txq_next_desc_get(txq);
txq              2446 drivers/net/ethernet/marvell/mvneta.c 		mvneta_txq_desc_put(txq);
txq              2455 drivers/net/ethernet/marvell/mvneta.c 		txq->tx_skb[txq->txq_put_index] = skb;
txq              2456 drivers/net/ethernet/marvell/mvneta.c 		mvneta_txq_inc_put(txq);
txq              2460 drivers/net/ethernet/marvell/mvneta.c 		txq->tx_skb[txq->txq_put_index] = NULL;
txq              2461 drivers/net/ethernet/marvell/mvneta.c 		mvneta_txq_inc_put(txq);
txq              2464 drivers/net/ethernet/marvell/mvneta.c 		if (mvneta_tx_frag_process(pp, skb, txq)) {
txq              2469 drivers/net/ethernet/marvell/mvneta.c 			mvneta_txq_desc_put(txq);
txq              2482 drivers/net/ethernet/marvell/mvneta.c 		txq->count += frags;
txq              2483 drivers/net/ethernet/marvell/mvneta.c 		if (txq->count >= txq->tx_stop_threshold)
txq              2487 drivers/net/ethernet/marvell/mvneta.c 		    txq->pending + frags > MVNETA_TXQ_DEC_SENT_MASK)
txq              2488 drivers/net/ethernet/marvell/mvneta.c 			mvneta_txq_pend_desc_add(pp, txq, frags);
txq              2490 drivers/net/ethernet/marvell/mvneta.c 			txq->pending += frags;
txq              2507 drivers/net/ethernet/marvell/mvneta.c 				  struct mvneta_tx_queue *txq)
txq              2510 drivers/net/ethernet/marvell/mvneta.c 	struct netdev_queue *nq = netdev_get_tx_queue(pp->dev, txq->id);
txq              2511 drivers/net/ethernet/marvell/mvneta.c 	int tx_done = txq->count;
txq              2513 drivers/net/ethernet/marvell/mvneta.c 	mvneta_txq_bufs_free(pp, txq, tx_done, nq);
txq              2516 drivers/net/ethernet/marvell/mvneta.c 	txq->count = 0;
txq              2517 drivers/net/ethernet/marvell/mvneta.c 	txq->txq_put_index = 0;
txq              2518 drivers/net/ethernet/marvell/mvneta.c 	txq->txq_get_index = 0;
txq              2526 drivers/net/ethernet/marvell/mvneta.c 	struct mvneta_tx_queue *txq;
txq              2531 drivers/net/ethernet/marvell/mvneta.c 		txq = mvneta_tx_done_policy(pp, cause_tx_done);
txq              2533 drivers/net/ethernet/marvell/mvneta.c 		nq = netdev_get_tx_queue(pp->dev, txq->id);
txq              2536 drivers/net/ethernet/marvell/mvneta.c 		if (txq->count)
txq              2537 drivers/net/ethernet/marvell/mvneta.c 			mvneta_txq_done(pp, txq);
txq              2540 drivers/net/ethernet/marvell/mvneta.c 		cause_tx_done &= ~((1 << txq->id));
txq              2984 drivers/net/ethernet/marvell/mvneta.c 			      struct mvneta_tx_queue *txq)
txq              2988 drivers/net/ethernet/marvell/mvneta.c 	txq->size = pp->tx_ring_size;
txq              2994 drivers/net/ethernet/marvell/mvneta.c 	txq->tx_stop_threshold = txq->size - MVNETA_MAX_SKB_DESCS;
txq              2995 drivers/net/ethernet/marvell/mvneta.c 	txq->tx_wake_threshold = txq->tx_stop_threshold / 2;
txq              2998 drivers/net/ethernet/marvell/mvneta.c 	txq->descs = dma_alloc_coherent(pp->dev->dev.parent,
txq              2999 drivers/net/ethernet/marvell/mvneta.c 					txq->size * MVNETA_DESC_ALIGNED_SIZE,
txq              3000 drivers/net/ethernet/marvell/mvneta.c 					&txq->descs_phys, GFP_KERNEL);
txq              3001 drivers/net/ethernet/marvell/mvneta.c 	if (!txq->descs)
txq              3004 drivers/net/ethernet/marvell/mvneta.c 	txq->last_desc = txq->size - 1;
txq              3006 drivers/net/ethernet/marvell/mvneta.c 	txq->tx_skb = kmalloc_array(txq->size, sizeof(*txq->tx_skb),
txq              3008 drivers/net/ethernet/marvell/mvneta.c 	if (!txq->tx_skb) {
txq              3010 drivers/net/ethernet/marvell/mvneta.c 				  txq->size * MVNETA_DESC_ALIGNED_SIZE,
txq              3011 drivers/net/ethernet/marvell/mvneta.c 				  txq->descs, txq->descs_phys);
txq              3016 drivers/net/ethernet/marvell/mvneta.c 	txq->tso_hdrs = dma_alloc_coherent(pp->dev->dev.parent,
txq              3017 drivers/net/ethernet/marvell/mvneta.c 					   txq->size * TSO_HEADER_SIZE,
txq              3018 drivers/net/ethernet/marvell/mvneta.c 					   &txq->tso_hdrs_phys, GFP_KERNEL);
txq              3019 drivers/net/ethernet/marvell/mvneta.c 	if (!txq->tso_hdrs) {
txq              3020 drivers/net/ethernet/marvell/mvneta.c 		kfree(txq->tx_skb);
txq              3022 drivers/net/ethernet/marvell/mvneta.c 				  txq->size * MVNETA_DESC_ALIGNED_SIZE,
txq              3023 drivers/net/ethernet/marvell/mvneta.c 				  txq->descs, txq->descs_phys);
txq              3029 drivers/net/ethernet/marvell/mvneta.c 		cpu = txq->id % num_present_cpus();
txq              3032 drivers/net/ethernet/marvell/mvneta.c 	cpumask_set_cpu(cpu, &txq->affinity_mask);
txq              3033 drivers/net/ethernet/marvell/mvneta.c 	netif_set_xps_queue(pp->dev, &txq->affinity_mask, txq->id);
txq              3039 drivers/net/ethernet/marvell/mvneta.c 			       struct mvneta_tx_queue *txq)
txq              3042 drivers/net/ethernet/marvell/mvneta.c 	mvreg_write(pp, MVETH_TXQ_TOKEN_CFG_REG(txq->id), 0x03ffffff);
txq              3043 drivers/net/ethernet/marvell/mvneta.c 	mvreg_write(pp, MVETH_TXQ_TOKEN_COUNT_REG(txq->id), 0x3fffffff);
txq              3046 drivers/net/ethernet/marvell/mvneta.c 	mvreg_write(pp, MVNETA_TXQ_BASE_ADDR_REG(txq->id), txq->descs_phys);
txq              3047 drivers/net/ethernet/marvell/mvneta.c 	mvreg_write(pp, MVNETA_TXQ_SIZE_REG(txq->id), txq->size);
txq              3049 drivers/net/ethernet/marvell/mvneta.c 	mvneta_tx_done_pkts_coal_set(pp, txq, txq->done_pkts_coal);
txq              3054 drivers/net/ethernet/marvell/mvneta.c 			   struct mvneta_tx_queue *txq)
txq              3058 drivers/net/ethernet/marvell/mvneta.c 	ret = mvneta_txq_sw_init(pp, txq);
txq              3062 drivers/net/ethernet/marvell/mvneta.c 	mvneta_txq_hw_init(pp, txq);
txq              3069 drivers/net/ethernet/marvell/mvneta.c 				 struct mvneta_tx_queue *txq)
txq              3071 drivers/net/ethernet/marvell/mvneta.c 	struct netdev_queue *nq = netdev_get_tx_queue(pp->dev, txq->id);
txq              3073 drivers/net/ethernet/marvell/mvneta.c 	kfree(txq->tx_skb);
txq              3075 drivers/net/ethernet/marvell/mvneta.c 	if (txq->tso_hdrs)
txq              3077 drivers/net/ethernet/marvell/mvneta.c 				  txq->size * TSO_HEADER_SIZE,
txq              3078 drivers/net/ethernet/marvell/mvneta.c 				  txq->tso_hdrs, txq->tso_hdrs_phys);
txq              3079 drivers/net/ethernet/marvell/mvneta.c 	if (txq->descs)
txq              3081 drivers/net/ethernet/marvell/mvneta.c 				  txq->size * MVNETA_DESC_ALIGNED_SIZE,
txq              3082 drivers/net/ethernet/marvell/mvneta.c 				  txq->descs, txq->descs_phys);
txq              3086 drivers/net/ethernet/marvell/mvneta.c 	txq->descs             = NULL;
txq              3087 drivers/net/ethernet/marvell/mvneta.c 	txq->last_desc         = 0;
txq              3088 drivers/net/ethernet/marvell/mvneta.c 	txq->next_desc_to_proc = 0;
txq              3089 drivers/net/ethernet/marvell/mvneta.c 	txq->descs_phys        = 0;
txq              3093 drivers/net/ethernet/marvell/mvneta.c 				 struct mvneta_tx_queue *txq)
txq              3096 drivers/net/ethernet/marvell/mvneta.c 	mvreg_write(pp, MVETH_TXQ_TOKEN_CFG_REG(txq->id), 0);
txq              3097 drivers/net/ethernet/marvell/mvneta.c 	mvreg_write(pp, MVETH_TXQ_TOKEN_COUNT_REG(txq->id), 0);
txq              3100 drivers/net/ethernet/marvell/mvneta.c 	mvreg_write(pp, MVNETA_TXQ_BASE_ADDR_REG(txq->id), 0);
txq              3101 drivers/net/ethernet/marvell/mvneta.c 	mvreg_write(pp, MVNETA_TXQ_SIZE_REG(txq->id), 0);
txq              3105 drivers/net/ethernet/marvell/mvneta.c 			      struct mvneta_tx_queue *txq)
txq              3107 drivers/net/ethernet/marvell/mvneta.c 	mvneta_txq_sw_deinit(pp, txq);
txq              3108 drivers/net/ethernet/marvell/mvneta.c 	mvneta_txq_hw_deinit(pp, txq);
txq              3994 drivers/net/ethernet/marvell/mvneta.c 		struct mvneta_tx_queue *txq = &pp->txqs[queue];
txq              3995 drivers/net/ethernet/marvell/mvneta.c 		txq->done_pkts_coal = c->tx_max_coalesced_frames;
txq              3996 drivers/net/ethernet/marvell/mvneta.c 		mvneta_tx_done_pkts_coal_set(pp, txq, txq->done_pkts_coal);
txq              4389 drivers/net/ethernet/marvell/mvneta.c 		struct mvneta_tx_queue *txq = &pp->txqs[queue];
txq              4390 drivers/net/ethernet/marvell/mvneta.c 		txq->id = queue;
txq              4391 drivers/net/ethernet/marvell/mvneta.c 		txq->size = pp->tx_ring_size;
txq              4392 drivers/net/ethernet/marvell/mvneta.c 		txq->done_pkts_coal = MVNETA_TXDONE_COAL_PKTS;
txq              4788 drivers/net/ethernet/marvell/mvneta.c 		struct mvneta_tx_queue *txq = &pp->txqs[queue];
txq              4790 drivers/net/ethernet/marvell/mvneta.c 		mvneta_txq_hw_deinit(pp, txq);
txq              4840 drivers/net/ethernet/marvell/mvneta.c 		struct mvneta_tx_queue *txq = &pp->txqs[queue];
txq              4842 drivers/net/ethernet/marvell/mvneta.c 		txq->next_desc_to_proc = 0;
txq              4843 drivers/net/ethernet/marvell/mvneta.c 		mvneta_txq_hw_init(pp, txq);
txq               189 drivers/net/ethernet/marvell/mvpp2/mvpp2.h #define MVPP2_TXQ_SENT_REG(txq)			(0x3c00 + 4 * (txq))
txq               338 drivers/net/ethernet/marvell/mvpp2/mvpp2.h #define     MVPP22_CTRS_TX_CTR(port, txq)	((txq) | ((port) << 3) | BIT(7))
txq               207 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 				 unsigned int txq)
txq               210 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 		tx_desc->pp21.phys_txq = txq;
txq               212 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 		tx_desc->pp22.phys_txq = txq;
txq               322 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c static inline int mvpp2_txq_phys(int port, int txq)
txq               324 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 	return (MVPP2_MAX_TCONT + port) * MVPP2_MAX_TXQ + txq;
txq              1791 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 		struct mvpp2_tx_queue *txq = port->txqs[queue];
txq              1793 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 		if (txq->descs)
txq              1897 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c mvpp2_txq_next_desc_get(struct mvpp2_tx_queue *txq)
txq              1899 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 	int tx_desc = txq->next_desc_to_proc;
txq              1901 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 	txq->next_desc_to_proc = MVPP2_QUEUE_NEXT_DESC(txq, tx_desc);
txq              1902 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 	return txq->descs + tx_desc;
txq              1949 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 					 struct mvpp2_tx_queue *txq, int num)
txq              1955 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 	val = (txq->id << MVPP2_TXQ_RSVD_REQ_Q_OFFSET) | num;
txq              1967 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 					    struct mvpp2_tx_queue *txq,
txq              1986 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 		txq_pcpu_aux = per_cpu_ptr(txq->pcpu, thread);
txq              1995 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 	   (txq->size - (MVPP2_MAX_THREADS * MVPP2_CPU_DESC_CHUNK)))
txq              1998 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 	txq_pcpu->reserved_num += mvpp2_txq_alloc_reserved_desc(port, txq, req);
txq              2009 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c static void mvpp2_txq_desc_put(struct mvpp2_tx_queue *txq)
txq              2011 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 	if (txq->next_desc_to_proc == 0)
txq              2012 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 		txq->next_desc_to_proc = txq->last_desc - 1;
txq              2014 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 		txq->next_desc_to_proc--;
txq              2059 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 					   struct mvpp2_tx_queue *txq)
txq              2066 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 					MVPP2_TXQ_SENT_REG(txq->id));
txq              2097 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 	int	txq, tx_port_num;
txq              2126 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 	for (txq = 0; txq < port->ntxqs; txq++) {
txq              2128 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 				 MVPP2_TXQ_SCHED_TOKEN_SIZE_REG(txq));
txq              2136 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 				    MVPP2_TXQ_SCHED_TOKEN_SIZE_REG(txq),
txq              2162 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 				   struct mvpp2_tx_queue *txq)
txq              2167 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 	if (txq->done_pkts_coal > MVPP2_TXQ_THRESH_MASK)
txq              2168 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 		txq->done_pkts_coal = MVPP2_TXQ_THRESH_MASK;
txq              2170 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 	val = (txq->done_pkts_coal << MVPP2_TXQ_THRESH_OFFSET);
txq              2171 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 	mvpp2_thread_write(port->priv, thread, MVPP2_TXQ_NUM_REG, txq->id);
txq              2231 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 				struct mvpp2_tx_queue *txq,
txq              2267 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c static void mvpp2_txq_done(struct mvpp2_port *port, struct mvpp2_tx_queue *txq,
txq              2270 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 	struct netdev_queue *nq = netdev_get_tx_queue(port->dev, txq->log_id);
txq              2276 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 	tx_done = mvpp2_txq_sent_desc_proc(port, txq);
txq              2279 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 	mvpp2_txq_bufs_free(port, txq, txq_pcpu, tx_done);
txq              2291 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 	struct mvpp2_tx_queue *txq;
txq              2296 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 		txq = mvpp2_get_tx_queue(port, cause);
txq              2297 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 		if (!txq)
txq              2300 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 		txq_pcpu = per_cpu_ptr(txq->pcpu, thread);
txq              2303 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 			mvpp2_txq_done(port, txq, txq_pcpu);
txq              2307 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 		cause &= ~(1 << txq->log_id);
txq              2454 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 			  struct mvpp2_tx_queue *txq)
txq              2461 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 	txq->size = port->tx_ring_size;
txq              2464 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 	txq->descs = dma_alloc_coherent(port->dev->dev.parent,
txq              2465 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 				txq->size * MVPP2_DESC_ALIGNED_SIZE,
txq              2466 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 				&txq->descs_dma, GFP_KERNEL);
txq              2467 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 	if (!txq->descs)
txq              2470 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 	txq->last_desc = txq->size - 1;
txq              2474 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 	mvpp2_thread_write(port->priv, thread, MVPP2_TXQ_NUM_REG, txq->id);
txq              2476 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 			   txq->descs_dma);
txq              2478 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 			   txq->size & MVPP2_TXQ_DESC_SIZE_MASK);
txq              2481 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 			   txq->id << MVPP2_TXQ_RSVD_CLR_OFFSET);
txq              2493 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 	       (txq->log_id * desc_per_txq);
txq              2504 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 	val = mvpp2_read(port->priv, MVPP2_TXQ_SCHED_REFILL_REG(txq->log_id));
txq              2508 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 	mvpp2_write(port->priv, MVPP2_TXQ_SCHED_REFILL_REG(txq->log_id), val);
txq              2511 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 	mvpp2_write(port->priv, MVPP2_TXQ_SCHED_TOKEN_SIZE_REG(txq->log_id),
txq              2515 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 		txq_pcpu = per_cpu_ptr(txq->pcpu, thread);
txq              2516 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 		txq_pcpu->size = txq->size;
txq              2529 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 		txq_pcpu->stop_threshold = txq->size - MVPP2_MAX_SKB_DESCS;
txq              2546 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 			     struct mvpp2_tx_queue *txq)
txq              2552 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 		txq_pcpu = per_cpu_ptr(txq->pcpu, thread);
txq              2564 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 	if (txq->descs)
txq              2566 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 				  txq->size * MVPP2_DESC_ALIGNED_SIZE,
txq              2567 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 				  txq->descs, txq->descs_dma);
txq              2569 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 	txq->descs             = NULL;
txq              2570 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 	txq->last_desc         = 0;
txq              2571 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 	txq->next_desc_to_proc = 0;
txq              2572 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 	txq->descs_dma         = 0;
txq              2575 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 	mvpp2_write(port->priv, MVPP2_TXQ_SCHED_TOKEN_CNTR_REG(txq->log_id), 0);
txq              2579 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 	mvpp2_thread_write(port->priv, thread, MVPP2_TXQ_NUM_REG, txq->id);
txq              2586 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c static void mvpp2_txq_clean(struct mvpp2_port *port, struct mvpp2_tx_queue *txq)
txq              2593 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 	mvpp2_thread_write(port->priv, thread, MVPP2_TXQ_NUM_REG, txq->id);
txq              2606 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 				    port->id, txq->log_id);
txq              2622 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 		txq_pcpu = per_cpu_ptr(txq->pcpu, thread);
txq              2625 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 		mvpp2_txq_bufs_free(port, txq, txq_pcpu, txq_pcpu->count);
txq              2637 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 	struct mvpp2_tx_queue *txq;
txq              2648 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 		txq = port->txqs[queue];
txq              2649 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 		mvpp2_txq_clean(port, txq);
txq              2650 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 		mvpp2_txq_deinit(port, txq);
txq              2688 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 	struct mvpp2_tx_queue *txq;
txq              2692 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 		txq = port->txqs[queue];
txq              2693 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 		err = mvpp2_txq_init(port, txq);
txq              2705 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 			txq = port->txqs[queue];
txq              2706 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 			mvpp2_tx_pkts_coal_set(port, txq);
txq              3017 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c tx_desc_unmap_put(struct mvpp2_port *port, struct mvpp2_tx_queue *txq,
txq              3021 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 	struct mvpp2_txq_pcpu *txq_pcpu = per_cpu_ptr(txq->pcpu, thread);
txq              3030 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 	mvpp2_txq_desc_put(txq);
txq              3036 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 				 struct mvpp2_tx_queue *txq)
txq              3039 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 	struct mvpp2_txq_pcpu *txq_pcpu = per_cpu_ptr(txq->pcpu, thread);
txq              3049 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 		mvpp2_txdesc_txq_set(port, tx_desc, txq->id);
txq              3056 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 			mvpp2_txq_desc_put(txq);
txq              3080 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 		tx_desc = txq->descs + i;
txq              3081 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 		tx_desc_unmap_put(port, txq, tx_desc);
txq              3089 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 				     struct mvpp2_tx_queue *txq,
txq              3098 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 	mvpp2_txdesc_txq_set(port, tx_desc, txq->id);
txq              3113 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 				     struct mvpp2_tx_queue *txq,
txq              3122 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 	mvpp2_txdesc_txq_set(port, tx_desc, txq->id);
txq              3128 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 		mvpp2_txq_desc_put(txq);
txq              3149 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 			struct mvpp2_tx_queue *txq,
txq              3160 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 	    mvpp2_txq_reserved_desc_num_proc(port, txq, txq_pcpu,
txq              3175 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 		mvpp2_tso_put_hdr(skb, dev, txq, aggr_txq, txq_pcpu, hdr_sz);
txq              3182 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 			if (mvpp2_tso_put_data(skb, dev, &tso, txq, aggr_txq,
txq              3193 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 		struct mvpp2_tx_desc *tx_desc = txq->descs + i;
txq              3194 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 		tx_desc_unmap_put(port, txq, tx_desc);
txq              3203 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 	struct mvpp2_tx_queue *txq, *aggr_txq;
txq              3216 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 	txq = port->txqs[txq_id];
txq              3217 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 	txq_pcpu = per_cpu_ptr(txq->pcpu, thread);
txq              3224 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 		frags = mvpp2_tx_tso(skb, dev, txq, aggr_txq, txq_pcpu);
txq              3231 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 	    mvpp2_txq_reserved_desc_num_proc(port, txq, txq_pcpu, frags)) {
txq              3238 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 	mvpp2_txdesc_txq_set(port, tx_desc, txq->id);
txq              3244 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 		mvpp2_txq_desc_put(txq);
txq              3265 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 		if (mvpp2_tx_frag_process(port, skb, aggr_txq, txq)) {
txq              3266 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 			tx_desc_unmap_put(port, txq, tx_desc);
txq              3297 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 	if (!port->has_tx_irqs && txq_pcpu->count >= txq->done_pkts_coal)
txq              3298 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 		mvpp2_txq_done(port, txq, txq_pcpu);
txq              4060 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 		struct mvpp2_tx_queue *txq = port->txqs[queue];
txq              4062 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 		txq->done_pkts_coal = c->tx_max_coalesced_frames;
txq              4065 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 			mvpp2_tx_pkts_coal_set(port, txq);
txq              4572 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 		struct mvpp2_tx_queue *txq;
txq              4574 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 		txq = devm_kzalloc(dev, sizeof(*txq), GFP_KERNEL);
txq              4575 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 		if (!txq) {
txq              4580 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 		txq->pcpu = alloc_percpu(struct mvpp2_txq_pcpu);
txq              4581 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 		if (!txq->pcpu) {
txq              4586 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 		txq->id = queue_phy_id;
txq              4587 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 		txq->log_id = queue;
txq              4588 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 		txq->done_pkts_coal = MVPP2_TXDONE_COAL_PKTS_THRESH;
txq              4590 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 			txq_pcpu = per_cpu_ptr(txq->pcpu, thread);
txq              4594 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 		port->txqs[queue] = txq;
txq               408 drivers/net/ethernet/mellanox/mlx5/core/en.h 	struct netdev_queue       *txq;
txq              1071 drivers/net/ethernet/mellanox/mlx5/core/en.h void mlx5e_tx_disable_queue(struct netdev_queue *txq);
txq                59 drivers/net/ethernet/mellanox/mlx5/core/en/reporter_tx.c 	mlx5e_tx_disable_queue(sq->txq);
txq               124 drivers/net/ethernet/mellanox/mlx5/core/en/reporter_tx.c 		jiffies_to_usecs(jiffies - sq->txq->trans_start));
txq               152 drivers/net/ethernet/mellanox/mlx5/core/en/reporter_tx.c 	bool stopped = netif_xmit_stopped(sq->txq);
txq              1350 drivers/net/ethernet/mellanox/mlx5/core/en_main.c 	sq->txq = netdev_get_tx_queue(sq->channel->netdev, sq->txq_ix);
txq              1352 drivers/net/ethernet/mellanox/mlx5/core/en_main.c 	netdev_tx_reset_queue(sq->txq);
txq              1353 drivers/net/ethernet/mellanox/mlx5/core/en_main.c 	netif_tx_start_queue(sq->txq);
txq              1356 drivers/net/ethernet/mellanox/mlx5/core/en_main.c void mlx5e_tx_disable_queue(struct netdev_queue *txq)
txq              1358 drivers/net/ethernet/mellanox/mlx5/core/en_main.c 	__netif_tx_lock_bh(txq);
txq              1359 drivers/net/ethernet/mellanox/mlx5/core/en_main.c 	netif_tx_stop_queue(txq);
txq              1360 drivers/net/ethernet/mellanox/mlx5/core/en_main.c 	__netif_tx_unlock_bh(txq);
txq              1372 drivers/net/ethernet/mellanox/mlx5/core/en_main.c 	mlx5e_tx_disable_queue(sq->txq);
txq               259 drivers/net/ethernet/mellanox/mlx5/core/en_tx.c 		netif_tx_stop_queue(sq->txq);
txq               263 drivers/net/ethernet/mellanox/mlx5/core/en_tx.c 	send_doorbell = __netdev_tx_sent_queue(sq->txq, num_bytes,
txq               525 drivers/net/ethernet/mellanox/mlx5/core/en_tx.c 	netdev_tx_completed_queue(sq->txq, npkts, nbytes);
txq               527 drivers/net/ethernet/mellanox/mlx5/core/en_tx.c 	if (netif_tx_queue_stopped(sq->txq) &&
txq               530 drivers/net/ethernet/mellanox/mlx5/core/en_tx.c 		netif_tx_wake_queue(sq->txq);
txq               575 drivers/net/ethernet/mellanox/mlx5/core/en_tx.c 	netdev_tx_completed_queue(sq->txq, npkts, nbytes);
txq               127 drivers/net/ethernet/micrel/ks8851.c 	struct sk_buff_head	txq;
txq               760 drivers/net/ethernet/micrel/ks8851.c 	bool last = skb_queue_empty(&ks->txq);
txq               765 drivers/net/ethernet/micrel/ks8851.c 		txb = skb_dequeue(&ks->txq);
txq               766 drivers/net/ethernet/micrel/ks8851.c 		last = skb_queue_empty(&ks->txq);
txq               906 drivers/net/ethernet/micrel/ks8851.c 	while (!skb_queue_empty(&ks->txq)) {
txq               907 drivers/net/ethernet/micrel/ks8851.c 		struct sk_buff *txb = skb_dequeue(&ks->txq);
txq               950 drivers/net/ethernet/micrel/ks8851.c 		skb_queue_tail(&ks->txq, skb);
txq              1519 drivers/net/ethernet/micrel/ks8851.c 	skb_queue_head_init(&ks->txq);
txq               109 drivers/net/ethernet/neterion/vxge/vxge-main.c 		if (__netif_tx_trylock(fifo->txq)) {
txq               112 drivers/net/ethernet/neterion/vxge/vxge-main.c 			__netif_tx_unlock(fifo->txq);
txq               621 drivers/net/ethernet/neterion/vxge/vxge-main.c 	if (netif_tx_queue_stopped(fifo->txq))
txq               622 drivers/net/ethernet/neterion/vxge/vxge-main.c 		netif_tx_wake_queue(fifo->txq);
txq               863 drivers/net/ethernet/neterion/vxge/vxge-main.c 	if (netif_tx_queue_stopped(fifo->txq))
txq               878 drivers/net/ethernet/neterion/vxge/vxge-main.c 		netif_tx_stop_queue(fifo->txq);
txq               991 drivers/net/ethernet/neterion/vxge/vxge-main.c 	netif_tx_stop_queue(fifo->txq);
txq              1569 drivers/net/ethernet/neterion/vxge/vxge-main.c 	if (netif_tx_queue_stopped(vpath->fifo.txq))
txq              1570 drivers/net/ethernet/neterion/vxge/vxge-main.c 		netif_tx_wake_queue(vpath->fifo.txq);
txq              2077 drivers/net/ethernet/neterion/vxge/vxge-main.c 				vpath->fifo.txq =
txq              2080 drivers/net/ethernet/neterion/vxge/vxge-main.c 				vpath->fifo.txq =
txq              2638 drivers/net/ethernet/neterion/vxge/vxge-main.c 					netif_tx_stop_queue(vpath->fifo.txq);
txq              3600 drivers/net/ethernet/neterion/vxge/vxge-main.c 				netif_tx_stop_queue(vpath->fifo.txq);
txq               240 drivers/net/ethernet/neterion/vxge/vxge-main.h 	struct netdev_queue *txq;
txq               962 drivers/net/ethernet/nxp/lpc_eth.c 	struct netdev_queue *txq = netdev_get_tx_queue(ndev, 0);
txq               964 drivers/net/ethernet/nxp/lpc_eth.c 	__netif_tx_lock(txq, smp_processor_id());
txq               966 drivers/net/ethernet/nxp/lpc_eth.c 	__netif_tx_unlock(txq);
txq               648 drivers/net/ethernet/qlogic/netxen/netxen_nic.h 	struct netdev_queue *txq;
txq               563 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 	__netif_tx_lock_bh(tx_ring->txq);
txq               568 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 		netif_tx_stop_queue(tx_ring->txq);
txq               572 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 				netif_tx_wake_queue(tx_ring->txq);
txq               574 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 			__netif_tx_unlock_bh(tx_ring->txq);
txq               596 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 	__netif_tx_unlock_bh(tx_ring->txq);
txq               197 drivers/net/ethernet/qlogic/netxen/netxen_nic_init.c 	tx_ring->txq = netdev_get_tx_queue(netdev, 0);
txq               415 drivers/net/ethernet/qlogic/qede/qede.h #define QEDE_TXQ_XDP_TO_IDX(edev, txq)	((txq)->index - \
txq               421 drivers/net/ethernet/qlogic/qede/qede.h #define QEDE_TXQ_TO_NDEV_TXQ_ID(edev, txq)	((QEDE_TSS_COUNT(edev) * \
txq               422 drivers/net/ethernet/qlogic/qede/qede.h 						 (txq)->cos) + (txq)->index)
txq               424 drivers/net/ethernet/qlogic/qede/qede.h 	(&((edev)->fp_array[QEDE_NDEV_TXQ_ID_TO_FP_ID(edev, idx)].txq \
txq               426 drivers/net/ethernet/qlogic/qede/qede.h #define QEDE_FP_TC0_TXQ(fp)	(&((fp)->txq[0]))
txq               466 drivers/net/ethernet/qlogic/qede/qede.h 	struct qede_tx_queue	*txq;
txq               527 drivers/net/ethernet/qlogic/qede/qede.h 		     struct qede_tx_queue *txq, int *len);
txq               567 drivers/net/ethernet/qlogic/qede/qede.h int qede_txq_has_work(struct qede_tx_queue *txq);
txq               221 drivers/net/ethernet/qlogic/qede/qede_ethtool.c 				       struct qede_tx_queue *txq, u8 **buf)
txq               226 drivers/net/ethernet/qlogic/qede/qede_ethtool.c 		if (txq->is_xdp)
txq               228 drivers/net/ethernet/qlogic/qede/qede_ethtool.c 				QEDE_TXQ_XDP_TO_IDX(edev, txq),
txq               231 drivers/net/ethernet/qlogic/qede/qede_ethtool.c 			sprintf(*buf, "%d_%d: %s", txq->index, txq->cos,
txq               276 drivers/net/ethernet/qlogic/qede/qede_ethtool.c 							   &fp->txq[cos], &buf);
txq               311 drivers/net/ethernet/qlogic/qede/qede_ethtool.c static void qede_get_ethtool_stats_txq(struct qede_tx_queue *txq, u64 **buf)
txq               316 drivers/net/ethernet/qlogic/qede/qede_ethtool.c 		**buf = *((u64 *)(((void *)txq) + qede_tqstats_arr[i].offset));
txq               356 drivers/net/ethernet/qlogic/qede/qede_ethtool.c 				qede_get_ethtool_stats_txq(&fp->txq[cos], &buf);
txq               840 drivers/net/ethernet/qlogic/qede/qede_ethtool.c 			struct qede_tx_queue *txq;
txq               849 drivers/net/ethernet/qlogic/qede/qede_ethtool.c 				txq = QEDE_FP_TC0_TXQ(fp);
txq               850 drivers/net/ethernet/qlogic/qede/qede_ethtool.c 				tx_handle = txq->handle;
txq               908 drivers/net/ethernet/qlogic/qede/qede_ethtool.c 			struct qede_tx_queue *txq;
txq               914 drivers/net/ethernet/qlogic/qede/qede_ethtool.c 			txq = QEDE_FP_TC0_TXQ(fp);
txq               918 drivers/net/ethernet/qlogic/qede/qede_ethtool.c 							     txq->handle);
txq              1493 drivers/net/ethernet/qlogic/qede/qede_ethtool.c 	struct qede_tx_queue *txq = NULL;
txq              1503 drivers/net/ethernet/qlogic/qede/qede_ethtool.c 			txq = QEDE_FP_TC0_TXQ(fp);
txq              1508 drivers/net/ethernet/qlogic/qede/qede_ethtool.c 	if (!txq) {
txq              1514 drivers/net/ethernet/qlogic/qede/qede_ethtool.c 	idx = txq->sw_tx_prod;
txq              1515 drivers/net/ethernet/qlogic/qede/qede_ethtool.c 	txq->sw_tx_ring.skbs[idx].skb = skb;
txq              1516 drivers/net/ethernet/qlogic/qede/qede_ethtool.c 	first_bd = qed_chain_produce(&txq->tx_pbl);
txq              1535 drivers/net/ethernet/qlogic/qede/qede_ethtool.c 	txq->sw_tx_prod = (txq->sw_tx_prod + 1) % txq->num_tx_buffers;
txq              1537 drivers/net/ethernet/qlogic/qede/qede_ethtool.c 	val = qed_chain_get_prod_idx(&txq->tx_pbl);
txq              1538 drivers/net/ethernet/qlogic/qede/qede_ethtool.c 	txq->tx_db.data.bd_prod = cpu_to_le16(val);
txq              1545 drivers/net/ethernet/qlogic/qede/qede_ethtool.c 	writel(txq->tx_db.raw, txq->doorbell_addr);
txq              1548 drivers/net/ethernet/qlogic/qede/qede_ethtool.c 		if (qede_txq_has_work(txq))
txq              1553 drivers/net/ethernet/qlogic/qede/qede_ethtool.c 	if (!qede_txq_has_work(txq)) {
txq              1558 drivers/net/ethernet/qlogic/qede/qede_ethtool.c 	first_bd = (struct eth_tx_1st_bd *)qed_chain_consume(&txq->tx_pbl);
txq              1561 drivers/net/ethernet/qlogic/qede/qede_ethtool.c 	txq->sw_tx_cons = (txq->sw_tx_cons + 1) % txq->num_tx_buffers;
txq              1562 drivers/net/ethernet/qlogic/qede/qede_ethtool.c 	txq->sw_tx_ring.skbs[idx].skb = NULL;
txq               100 drivers/net/ethernet/qlogic/qede/qede_fp.c int qede_free_tx_pkt(struct qede_dev *edev, struct qede_tx_queue *txq, int *len)
txq               102 drivers/net/ethernet/qlogic/qede/qede_fp.c 	u16 idx = txq->sw_tx_cons;
txq               103 drivers/net/ethernet/qlogic/qede/qede_fp.c 	struct sk_buff *skb = txq->sw_tx_ring.skbs[idx].skb;
txq               108 drivers/net/ethernet/qlogic/qede/qede_fp.c 	bool data_split = txq->sw_tx_ring.skbs[idx].flags & QEDE_TSO_SPLIT_BD;
txq               114 drivers/net/ethernet/qlogic/qede/qede_fp.c 		       idx, txq->sw_tx_cons, txq->sw_tx_prod);
txq               120 drivers/net/ethernet/qlogic/qede/qede_fp.c 	first_bd = (struct eth_tx_1st_bd *)qed_chain_consume(&txq->tx_pbl);
txq               128 drivers/net/ethernet/qlogic/qede/qede_fp.c 			qed_chain_consume(&txq->tx_pbl);
txq               138 drivers/net/ethernet/qlogic/qede/qede_fp.c 			qed_chain_consume(&txq->tx_pbl);
txq               144 drivers/net/ethernet/qlogic/qede/qede_fp.c 		qed_chain_consume(&txq->tx_pbl);
txq               148 drivers/net/ethernet/qlogic/qede/qede_fp.c 	txq->sw_tx_ring.skbs[idx].skb = NULL;
txq               149 drivers/net/ethernet/qlogic/qede/qede_fp.c 	txq->sw_tx_ring.skbs[idx].flags = 0;
txq               155 drivers/net/ethernet/qlogic/qede/qede_fp.c static void qede_free_failed_tx_pkt(struct qede_tx_queue *txq,
txq               159 drivers/net/ethernet/qlogic/qede/qede_fp.c 	u16 idx = txq->sw_tx_prod;
txq               160 drivers/net/ethernet/qlogic/qede/qede_fp.c 	struct sk_buff *skb = txq->sw_tx_ring.skbs[idx].skb;
txq               165 drivers/net/ethernet/qlogic/qede/qede_fp.c 	qed_chain_set_prod(&txq->tx_pbl,
txq               166 drivers/net/ethernet/qlogic/qede/qede_fp.c 			   le16_to_cpu(txq->tx_db.data.bd_prod), first_bd);
txq               168 drivers/net/ethernet/qlogic/qede/qede_fp.c 	first_bd = (struct eth_tx_1st_bd *)qed_chain_produce(&txq->tx_pbl);
txq               172 drivers/net/ethernet/qlogic/qede/qede_fp.c 					  qed_chain_produce(&txq->tx_pbl);
txq               177 drivers/net/ethernet/qlogic/qede/qede_fp.c 	dma_unmap_single(txq->dev, BD_UNMAP_ADDR(first_bd),
txq               183 drivers/net/ethernet/qlogic/qede/qede_fp.c 			qed_chain_produce(&txq->tx_pbl);
txq               185 drivers/net/ethernet/qlogic/qede/qede_fp.c 			dma_unmap_page(txq->dev,
txq               191 drivers/net/ethernet/qlogic/qede/qede_fp.c 	qed_chain_set_prod(&txq->tx_pbl,
txq               192 drivers/net/ethernet/qlogic/qede/qede_fp.c 			   le16_to_cpu(txq->tx_db.data.bd_prod), first_bd);
txq               196 drivers/net/ethernet/qlogic/qede/qede_fp.c 	txq->sw_tx_ring.skbs[idx].skb = NULL;
txq               197 drivers/net/ethernet/qlogic/qede/qede_fp.c 	txq->sw_tx_ring.skbs[idx].flags = 0;
txq               267 drivers/net/ethernet/qlogic/qede/qede_fp.c static int map_frag_to_bd(struct qede_tx_queue *txq,
txq               273 drivers/net/ethernet/qlogic/qede/qede_fp.c 	mapping = skb_frag_dma_map(txq->dev, frag, 0,
txq               275 drivers/net/ethernet/qlogic/qede/qede_fp.c 	if (unlikely(dma_mapping_error(txq->dev, mapping)))
txq               314 drivers/net/ethernet/qlogic/qede/qede_fp.c static inline void qede_update_tx_producer(struct qede_tx_queue *txq)
txq               321 drivers/net/ethernet/qlogic/qede/qede_fp.c 	writel(txq->tx_db.raw, txq->doorbell_addr);
txq               333 drivers/net/ethernet/qlogic/qede/qede_fp.c 	struct qede_tx_queue *txq = fp->xdp_tx;
txq               335 drivers/net/ethernet/qlogic/qede/qede_fp.c 	u16 idx = txq->sw_tx_prod;
txq               338 drivers/net/ethernet/qlogic/qede/qede_fp.c 	if (!qed_chain_get_elem_left(&txq->tx_pbl)) {
txq               339 drivers/net/ethernet/qlogic/qede/qede_fp.c 		txq->stopped_cnt++;
txq               343 drivers/net/ethernet/qlogic/qede/qede_fp.c 	first_bd = (struct eth_tx_1st_bd *)qed_chain_produce(&txq->tx_pbl);
txq               365 drivers/net/ethernet/qlogic/qede/qede_fp.c 	txq->sw_tx_ring.xdp[idx].page = metadata->data;
txq               366 drivers/net/ethernet/qlogic/qede/qede_fp.c 	txq->sw_tx_ring.xdp[idx].mapping = metadata->mapping;
txq               367 drivers/net/ethernet/qlogic/qede/qede_fp.c 	txq->sw_tx_prod = (txq->sw_tx_prod + 1) % txq->num_tx_buffers;
txq               375 drivers/net/ethernet/qlogic/qede/qede_fp.c int qede_txq_has_work(struct qede_tx_queue *txq)
txq               381 drivers/net/ethernet/qlogic/qede/qede_fp.c 	hw_bd_cons = le16_to_cpu(*txq->hw_cons_ptr);
txq               382 drivers/net/ethernet/qlogic/qede/qede_fp.c 	if (qed_chain_get_cons_idx(&txq->tx_pbl) == hw_bd_cons + 1)
txq               385 drivers/net/ethernet/qlogic/qede/qede_fp.c 	return hw_bd_cons != qed_chain_get_cons_idx(&txq->tx_pbl);
txq               388 drivers/net/ethernet/qlogic/qede/qede_fp.c static void qede_xdp_tx_int(struct qede_dev *edev, struct qede_tx_queue *txq)
txq               392 drivers/net/ethernet/qlogic/qede/qede_fp.c 	hw_bd_cons = le16_to_cpu(*txq->hw_cons_ptr);
txq               395 drivers/net/ethernet/qlogic/qede/qede_fp.c 	while (hw_bd_cons != qed_chain_get_cons_idx(&txq->tx_pbl)) {
txq               396 drivers/net/ethernet/qlogic/qede/qede_fp.c 		qed_chain_consume(&txq->tx_pbl);
txq               397 drivers/net/ethernet/qlogic/qede/qede_fp.c 		idx = txq->sw_tx_cons;
txq               400 drivers/net/ethernet/qlogic/qede/qede_fp.c 			       txq->sw_tx_ring.xdp[idx].mapping,
txq               402 drivers/net/ethernet/qlogic/qede/qede_fp.c 		__free_page(txq->sw_tx_ring.xdp[idx].page);
txq               404 drivers/net/ethernet/qlogic/qede/qede_fp.c 		txq->sw_tx_cons = (txq->sw_tx_cons + 1) % txq->num_tx_buffers;
txq               405 drivers/net/ethernet/qlogic/qede/qede_fp.c 		txq->xmit_pkts++;
txq               409 drivers/net/ethernet/qlogic/qede/qede_fp.c static int qede_tx_int(struct qede_dev *edev, struct qede_tx_queue *txq)
txq               416 drivers/net/ethernet/qlogic/qede/qede_fp.c 	netdev_txq = netdev_get_tx_queue(edev->ndev, txq->ndev_txq_id);
txq               418 drivers/net/ethernet/qlogic/qede/qede_fp.c 	hw_bd_cons = le16_to_cpu(*txq->hw_cons_ptr);
txq               421 drivers/net/ethernet/qlogic/qede/qede_fp.c 	while (hw_bd_cons != qed_chain_get_cons_idx(&txq->tx_pbl)) {
txq               424 drivers/net/ethernet/qlogic/qede/qede_fp.c 		rc = qede_free_tx_pkt(edev, txq, &len);
txq               428 drivers/net/ethernet/qlogic/qede/qede_fp.c 				  qed_chain_get_cons_idx(&txq->tx_pbl));
txq               434 drivers/net/ethernet/qlogic/qede/qede_fp.c 		txq->sw_tx_cons = (txq->sw_tx_cons + 1) % txq->num_tx_buffers;
txq               435 drivers/net/ethernet/qlogic/qede/qede_fp.c 		txq->xmit_pkts++;
txq               466 drivers/net/ethernet/qlogic/qede/qede_fp.c 		    (qed_chain_get_elem_left(&txq->tx_pbl)
txq              1363 drivers/net/ethernet/qlogic/qede/qede_fp.c 			if (qede_txq_has_work(&fp->txq[cos]))
txq              1385 drivers/net/ethernet/qlogic/qede/qede_fp.c 			if (qede_txq_has_work(&fp->txq[cos]))
txq              1386 drivers/net/ethernet/qlogic/qede/qede_fp.c 				qede_tx_int(edev, &fp->txq[cos]);
txq              1433 drivers/net/ethernet/qlogic/qede/qede_fp.c 	struct qede_tx_queue *txq;
txq              1450 drivers/net/ethernet/qlogic/qede/qede_fp.c 	txq = QEDE_NDEV_TXQ_ID_TO_TXQ(edev, txq_index);
txq              1453 drivers/net/ethernet/qlogic/qede/qede_fp.c 	WARN_ON(qed_chain_get_elem_left(&txq->tx_pbl) < (MAX_SKB_FRAGS + 1));
txq              1460 drivers/net/ethernet/qlogic/qede/qede_fp.c 			txq->tx_mem_alloc_err++;
txq              1469 drivers/net/ethernet/qlogic/qede/qede_fp.c 	idx = txq->sw_tx_prod;
txq              1470 drivers/net/ethernet/qlogic/qede/qede_fp.c 	txq->sw_tx_ring.skbs[idx].skb = skb;
txq              1472 drivers/net/ethernet/qlogic/qede/qede_fp.c 		   qed_chain_produce(&txq->tx_pbl);
txq              1481 drivers/net/ethernet/qlogic/qede/qede_fp.c 	mapping = dma_map_single(txq->dev, skb->data,
txq              1483 drivers/net/ethernet/qlogic/qede/qede_fp.c 	if (unlikely(dma_mapping_error(txq->dev, mapping))) {
txq              1485 drivers/net/ethernet/qlogic/qede/qede_fp.c 		qede_free_failed_tx_pkt(txq, first_bd, 0, false);
txq              1486 drivers/net/ethernet/qlogic/qede/qede_fp.c 		qede_update_tx_producer(txq);
txq              1497 drivers/net/ethernet/qlogic/qede/qede_fp.c 			qed_chain_produce(&txq->tx_pbl);
txq              1502 drivers/net/ethernet/qlogic/qede/qede_fp.c 			qed_chain_produce(&txq->tx_pbl);
txq              1535 drivers/net/ethernet/qlogic/qede/qede_fp.c 		if (unlikely(txq->is_legacy))
txq              1592 drivers/net/ethernet/qlogic/qede/qede_fp.c 			txq->sw_tx_ring.skbs[idx].flags |= QEDE_TSO_SPLIT_BD;
txq              1609 drivers/net/ethernet/qlogic/qede/qede_fp.c 		rc = map_frag_to_bd(txq,
txq              1613 drivers/net/ethernet/qlogic/qede/qede_fp.c 			qede_free_failed_tx_pkt(txq, first_bd, nbd, data_split);
txq              1614 drivers/net/ethernet/qlogic/qede/qede_fp.c 			qede_update_tx_producer(txq);
txq              1629 drivers/net/ethernet/qlogic/qede/qede_fp.c 			     qed_chain_produce(&txq->tx_pbl);
txq              1633 drivers/net/ethernet/qlogic/qede/qede_fp.c 		rc = map_frag_to_bd(txq,
txq              1637 drivers/net/ethernet/qlogic/qede/qede_fp.c 			qede_free_failed_tx_pkt(txq, first_bd, nbd, data_split);
txq              1638 drivers/net/ethernet/qlogic/qede/qede_fp.c 			qede_update_tx_producer(txq);
txq              1653 drivers/net/ethernet/qlogic/qede/qede_fp.c 	txq->sw_tx_prod = (txq->sw_tx_prod + 1) % txq->num_tx_buffers;
txq              1656 drivers/net/ethernet/qlogic/qede/qede_fp.c 	txq->tx_db.data.bd_prod =
txq              1657 drivers/net/ethernet/qlogic/qede/qede_fp.c 		cpu_to_le16(qed_chain_get_prod_idx(&txq->tx_pbl));
txq              1660 drivers/net/ethernet/qlogic/qede/qede_fp.c 		qede_update_tx_producer(txq);
txq              1662 drivers/net/ethernet/qlogic/qede/qede_fp.c 	if (unlikely(qed_chain_get_elem_left(&txq->tx_pbl)
txq              1665 drivers/net/ethernet/qlogic/qede/qede_fp.c 			qede_update_tx_producer(txq);
txq              1668 drivers/net/ethernet/qlogic/qede/qede_fp.c 		txq->stopped_cnt++;
txq              1677 drivers/net/ethernet/qlogic/qede/qede_fp.c 		if ((qed_chain_get_elem_left(&txq->tx_pbl) >=
txq               856 drivers/net/ethernet/qlogic/qede/qede_main.c 			kfree(fp->txq);
txq               906 drivers/net/ethernet/qlogic/qede/qede_main.c 			fp->txq = kcalloc(edev->dev_info.num_tc,
txq               907 drivers/net/ethernet/qlogic/qede/qede_main.c 					  sizeof(*fp->txq), GFP_KERNEL);
txq               908 drivers/net/ethernet/qlogic/qede/qede_main.c 			if (!fp->txq)
txq              1461 drivers/net/ethernet/qlogic/qede/qede_main.c static void qede_free_mem_txq(struct qede_dev *edev, struct qede_tx_queue *txq)
txq              1464 drivers/net/ethernet/qlogic/qede/qede_main.c 	if (txq->is_xdp)
txq              1465 drivers/net/ethernet/qlogic/qede/qede_main.c 		kfree(txq->sw_tx_ring.xdp);
txq              1467 drivers/net/ethernet/qlogic/qede/qede_main.c 		kfree(txq->sw_tx_ring.skbs);
txq              1470 drivers/net/ethernet/qlogic/qede/qede_main.c 	edev->ops->common->chain_free(edev->cdev, &txq->tx_pbl);
txq              1474 drivers/net/ethernet/qlogic/qede/qede_main.c static int qede_alloc_mem_txq(struct qede_dev *edev, struct qede_tx_queue *txq)
txq              1479 drivers/net/ethernet/qlogic/qede/qede_main.c 	txq->num_tx_buffers = edev->q_num_tx_buffers;
txq              1482 drivers/net/ethernet/qlogic/qede/qede_main.c 	if (txq->is_xdp) {
txq              1483 drivers/net/ethernet/qlogic/qede/qede_main.c 		size = sizeof(*txq->sw_tx_ring.xdp) * txq->num_tx_buffers;
txq              1484 drivers/net/ethernet/qlogic/qede/qede_main.c 		txq->sw_tx_ring.xdp = kzalloc(size, GFP_KERNEL);
txq              1485 drivers/net/ethernet/qlogic/qede/qede_main.c 		if (!txq->sw_tx_ring.xdp)
txq              1488 drivers/net/ethernet/qlogic/qede/qede_main.c 		size = sizeof(*txq->sw_tx_ring.skbs) * txq->num_tx_buffers;
txq              1489 drivers/net/ethernet/qlogic/qede/qede_main.c 		txq->sw_tx_ring.skbs = kzalloc(size, GFP_KERNEL);
txq              1490 drivers/net/ethernet/qlogic/qede/qede_main.c 		if (!txq->sw_tx_ring.skbs)
txq              1498 drivers/net/ethernet/qlogic/qede/qede_main.c 					    txq->num_tx_buffers,
txq              1500 drivers/net/ethernet/qlogic/qede/qede_main.c 					    &txq->tx_pbl, NULL);
txq              1507 drivers/net/ethernet/qlogic/qede/qede_main.c 	qede_free_mem_txq(edev, txq);
txq              1526 drivers/net/ethernet/qlogic/qede/qede_main.c 			qede_free_mem_txq(edev, &fp->txq[cos]);
txq              1557 drivers/net/ethernet/qlogic/qede/qede_main.c 			rc = qede_alloc_mem_txq(edev, &fp->txq[cos]);
txq              1600 drivers/net/ethernet/qlogic/qede/qede_main.c 				struct qede_tx_queue *txq)
txq              1606 drivers/net/ethernet/qlogic/qede/qede_main.c 	netdev_txq = netdev_get_tx_queue(edev->ndev, txq->ndev_txq_id);
txq              1608 drivers/net/ethernet/qlogic/qede/qede_main.c 	while (qed_chain_get_cons_idx(&txq->tx_pbl) !=
txq              1609 drivers/net/ethernet/qlogic/qede/qede_main.c 	       qed_chain_get_prod_idx(&txq->tx_pbl)) {
txq              1612 drivers/net/ethernet/qlogic/qede/qede_main.c 			   txq->index, qed_chain_get_cons_idx(&txq->tx_pbl),
txq              1613 drivers/net/ethernet/qlogic/qede/qede_main.c 			   qed_chain_get_prod_idx(&txq->tx_pbl));
txq              1615 drivers/net/ethernet/qlogic/qede/qede_main.c 		rc = qede_free_tx_pkt(edev, txq, &len);
txq              1619 drivers/net/ethernet/qlogic/qede/qede_main.c 				  txq->index,
txq              1620 drivers/net/ethernet/qlogic/qede/qede_main.c 				  qed_chain_get_cons_idx(&txq->tx_pbl),
txq              1621 drivers/net/ethernet/qlogic/qede/qede_main.c 				  qed_chain_get_prod_idx(&txq->tx_pbl));
txq              1627 drivers/net/ethernet/qlogic/qede/qede_main.c 		txq->sw_tx_cons++;
txq              1646 drivers/net/ethernet/qlogic/qede/qede_main.c 						    &fp->txq[cos]);
txq              1688 drivers/net/ethernet/qlogic/qede/qede_main.c 				struct qede_tx_queue *txq = &fp->txq[cos];
txq              1691 drivers/net/ethernet/qlogic/qede/qede_main.c 				txq->cos = cos;
txq              1692 drivers/net/ethernet/qlogic/qede/qede_main.c 				txq->index = txq_index;
txq              1693 drivers/net/ethernet/qlogic/qede/qede_main.c 				ndev_tx_id = QEDE_TXQ_TO_NDEV_TXQ_ID(edev, txq);
txq              1694 drivers/net/ethernet/qlogic/qede/qede_main.c 				txq->ndev_txq_id = ndev_tx_id;
txq              1697 drivers/net/ethernet/qlogic/qede/qede_main.c 					txq->is_legacy = 1;
txq              1698 drivers/net/ethernet/qlogic/qede/qede_main.c 				txq->dev = &edev->pdev->dev;
txq              1849 drivers/net/ethernet/qlogic/qede/qede_main.c 			  struct qede_tx_queue *txq, bool allow_drain)
txq              1853 drivers/net/ethernet/qlogic/qede/qede_main.c 	while (txq->sw_tx_cons != txq->sw_tx_prod) {
txq              1858 drivers/net/ethernet/qlogic/qede/qede_main.c 					  txq->index);
txq              1862 drivers/net/ethernet/qlogic/qede/qede_main.c 				return qede_drain_txq(edev, txq, false);
txq              1866 drivers/net/ethernet/qlogic/qede/qede_main.c 				  txq->index, txq->sw_tx_prod,
txq              1867 drivers/net/ethernet/qlogic/qede/qede_main.c 				  txq->sw_tx_cons);
txq              1882 drivers/net/ethernet/qlogic/qede/qede_main.c 			 struct qede_tx_queue *txq, int rss_id)
txq              1885 drivers/net/ethernet/qlogic/qede/qede_main.c 	edev->ops->common->db_recovery_del(edev->cdev, txq->doorbell_addr,
txq              1886 drivers/net/ethernet/qlogic/qede/qede_main.c 					   &txq->tx_db);
txq              1888 drivers/net/ethernet/qlogic/qede/qede_main.c 	return edev->ops->q_tx_stop(edev->cdev, rss_id, txq->handle);
txq              1924 drivers/net/ethernet/qlogic/qede/qede_main.c 				rc = qede_drain_txq(edev, &fp->txq[cos], true);
txq              1946 drivers/net/ethernet/qlogic/qede/qede_main.c 				rc = qede_stop_txq(edev, &fp->txq[cos], i);
txq              1981 drivers/net/ethernet/qlogic/qede/qede_main.c 			  struct qede_tx_queue *txq, u8 rss_id, u16 sb_idx)
txq              1983 drivers/net/ethernet/qlogic/qede/qede_main.c 	dma_addr_t phys_table = qed_chain_get_pbl_phys(&txq->tx_pbl);
txq              1984 drivers/net/ethernet/qlogic/qede/qede_main.c 	u32 page_cnt = qed_chain_get_page_cnt(&txq->tx_pbl);
txq              1995 drivers/net/ethernet/qlogic/qede/qede_main.c 	if (txq->is_xdp)
txq              1996 drivers/net/ethernet/qlogic/qede/qede_main.c 		params.queue_id = QEDE_TXQ_XDP_TO_IDX(edev, txq);
txq              1998 drivers/net/ethernet/qlogic/qede/qede_main.c 		params.queue_id = txq->index;
txq              2002 drivers/net/ethernet/qlogic/qede/qede_main.c 	params.tc = txq->cos;
txq              2007 drivers/net/ethernet/qlogic/qede/qede_main.c 		DP_ERR(edev, "Start TXQ #%d failed %d\n", txq->index, rc);
txq              2011 drivers/net/ethernet/qlogic/qede/qede_main.c 	txq->doorbell_addr = ret_params.p_doorbell;
txq              2012 drivers/net/ethernet/qlogic/qede/qede_main.c 	txq->handle = ret_params.p_handle;
txq              2015 drivers/net/ethernet/qlogic/qede/qede_main.c 	txq->hw_cons_ptr = &fp->sb_info->sb_virt->pi_array[sb_idx];
txq              2018 drivers/net/ethernet/qlogic/qede/qede_main.c 	SET_FIELD(txq->tx_db.data.params, ETH_DB_DATA_DEST, DB_DEST_XCM);
txq              2019 drivers/net/ethernet/qlogic/qede/qede_main.c 	SET_FIELD(txq->tx_db.data.params, ETH_DB_DATA_AGG_CMD, DB_AGG_CMD_SET);
txq              2020 drivers/net/ethernet/qlogic/qede/qede_main.c 	SET_FIELD(txq->tx_db.data.params, ETH_DB_DATA_AGG_VAL_SEL,
txq              2022 drivers/net/ethernet/qlogic/qede/qede_main.c 	txq->tx_db.data.agg_flags = DQ_XCM_ETH_DQ_CF_CMD;
txq              2025 drivers/net/ethernet/qlogic/qede/qede_main.c 	rc = edev->ops->common->db_recovery_add(edev->cdev, txq->doorbell_addr,
txq              2026 drivers/net/ethernet/qlogic/qede/qede_main.c 						&txq->tx_db, DB_REC_WIDTH_32B,
txq              2130 drivers/net/ethernet/qlogic/qede/qede_main.c 				rc = qede_start_txq(edev, fp, &fp->txq[cos], i,
txq              2489 drivers/net/ethernet/qlogic/qede/qede_main.c static bool qede_is_txq_full(struct qede_dev *edev, struct qede_tx_queue *txq)
txq              2493 drivers/net/ethernet/qlogic/qede/qede_main.c 	netdev_txq = netdev_get_tx_queue(edev->ndev, txq->ndev_txq_id);
txq              2558 drivers/net/ethernet/qlogic/qede/qede_main.c 			struct qede_tx_queue *txq = QEDE_FP_TC0_TXQ(fp);
txq              2560 drivers/net/ethernet/qlogic/qede/qede_main.c 			if (txq->sw_tx_cons != txq->sw_tx_prod)
txq              2562 drivers/net/ethernet/qlogic/qede/qede_main.c 			if (qede_is_txq_full(edev, txq))
txq               638 drivers/net/ethernet/qlogic/qlcnic/qlcnic.h 	struct netdev_queue *txq;
txq               396 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 	__netif_tx_lock_bh(tx_ring->txq);
txq               401 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 		netif_tx_stop_queue(tx_ring->txq);
txq               405 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 				netif_tx_wake_queue(tx_ring->txq);
txq               408 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 			__netif_tx_unlock_bh(tx_ring->txq);
txq               432 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 	__netif_tx_unlock_bh(tx_ring->txq);
txq               693 drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c 		netif_tx_stop_queue(tx_ring->txq);
txq               695 drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c 			netif_tx_start_queue(tx_ring->txq);
txq               930 drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c 		if (netif_tx_queue_stopped(tx_ring->txq) &&
txq               933 drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c 				netif_tx_wake_queue(tx_ring->txq);
txq              2413 drivers/net/ethernet/qlogic/qlcnic/qlcnic_main.c 		tx_ring->txq = netdev_get_tx_queue(netdev, ring);
txq               454 drivers/net/ethernet/samsung/sxgbe/sxgbe_common.h 	struct sxgbe_tx_queue *txq[SXGBE_TX_QUEUES];
txq               333 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c 			priv->hw->desc->init_tx_desc(&priv->txq[j]->dma_tx[i]);
txq               566 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c 				   priv->txq[queue_num], tx_rsize);
txq               575 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c 		priv->txq[queue_num]->priv_ptr = priv;
txq               599 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c 		free_tx_ring(priv->device, priv->txq[queue_num], tx_rsize);
txq               635 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c 		struct sxgbe_tx_queue *tqueue = priv->txq[queue_num];
txq               651 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c 		free_tx_ring(priv->device, priv->txq[queue_num], tx_rsize);
txq               665 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c 		priv->txq[queue_num] = devm_kmalloc(priv->device,
txq               667 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c 		if (!priv->txq[queue_num])
txq               796 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c 		struct sxgbe_tx_queue *tqueue = priv->txq[queue_num];
txq               815 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c 	struct sxgbe_tx_queue *tx_ring = priv->txq[queue_num];
txq               960 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c 					(priv->txq[queue_num])->dma_tx_phy,
txq              1024 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c 		struct sxgbe_tx_queue *p = priv->txq[queue_num];
txq              1038 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c 		struct sxgbe_tx_queue *p = priv->txq[queue_num];
txq              1122 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c 				       (priv->txq[queue_num])->irq_no,
txq              1124 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c 				       dev->name, priv->txq[queue_num]);
txq              1274 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c 	struct sxgbe_tx_queue *tqueue = priv->txq[txq_index];
txq              1623 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c 	struct sxgbe_tx_queue *txq = (struct sxgbe_tx_queue *)dev_id;
txq              1624 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c 	struct sxgbe_priv_data *priv = txq->priv_ptr;
txq              1627 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c 	status = priv->hw->dma->tx_dma_int_status(priv->ioaddr, txq->queue_no,
txq              1635 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c 		sxgbe_restart_tx_queue(priv, txq->queue_no);
txq              1644 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c 					       txq->queue_no, priv->tx_tc);
txq               127 drivers/net/ethernet/samsung/sxgbe/sxgbe_platform.c 		priv->txq[i]->irq_no = irq_of_parse_and_map(node, chan++);
txq               128 drivers/net/ethernet/samsung/sxgbe/sxgbe_platform.c 		if (priv->txq[i]->irq_no <= 0) {
txq               160 drivers/net/ethernet/samsung/sxgbe/sxgbe_platform.c 		irq_dispose_mapping(priv->txq[i]->irq_no);
txq               284 drivers/net/ethernet/sfc/ptp.c 	struct sk_buff_head txq;
txq              1365 drivers/net/ethernet/sfc/ptp.c 	skb_queue_purge(&efx->ptp_data->txq);
txq              1418 drivers/net/ethernet/sfc/ptp.c 	while ((skb = skb_dequeue(&ptp_data->txq)))
txq              1462 drivers/net/ethernet/sfc/ptp.c 	skb_queue_head_init(&ptp->txq);
txq              1570 drivers/net/ethernet/sfc/ptp.c 	skb_queue_purge(&efx->ptp_data->txq);
txq              1708 drivers/net/ethernet/sfc/ptp.c 	skb_queue_tail(&ptp->txq, skb);
txq              3586 drivers/net/ethernet/sun/niu.c 	struct netdev_queue *txq;
txq              3592 drivers/net/ethernet/sun/niu.c 	txq = netdev_get_tx_queue(np->dev, index);
txq              3616 drivers/net/ethernet/sun/niu.c 	if (unlikely(netif_tx_queue_stopped(txq) &&
txq              3618 drivers/net/ethernet/sun/niu.c 		__netif_tx_lock(txq, smp_processor_id());
txq              3619 drivers/net/ethernet/sun/niu.c 		if (netif_tx_queue_stopped(txq) &&
txq              3621 drivers/net/ethernet/sun/niu.c 			netif_tx_wake_queue(txq);
txq              3622 drivers/net/ethernet/sun/niu.c 		__netif_tx_unlock(txq);
txq              6612 drivers/net/ethernet/sun/niu.c 	struct netdev_queue *txq;
txq              6622 drivers/net/ethernet/sun/niu.c 	txq = netdev_get_tx_queue(dev, i);
txq              6625 drivers/net/ethernet/sun/niu.c 		netif_tx_stop_queue(txq);
txq              6716 drivers/net/ethernet/sun/niu.c 		netif_tx_stop_queue(txq);
txq              6718 drivers/net/ethernet/sun/niu.c 			netif_tx_wake_queue(txq);
txq               703 drivers/net/ethernet/sun/sungem.c 		struct netdev_queue *txq = netdev_get_tx_queue(dev, 0);
txq               705 drivers/net/ethernet/sun/sungem.c 		__netif_tx_lock(txq, smp_processor_id());
txq               709 drivers/net/ethernet/sun/sungem.c 		__netif_tx_unlock(txq);
txq               894 drivers/net/ethernet/sun/sungem.c 			struct netdev_queue *txq = netdev_get_tx_queue(dev, 0);
txq               902 drivers/net/ethernet/sun/sungem.c 			__netif_tx_lock(txq, smp_processor_id());
txq               904 drivers/net/ethernet/sun/sungem.c 			__netif_tx_unlock(txq);
txq              1349 drivers/net/ethernet/sun/sungem.c 	struct netdev_queue *txq = netdev_get_tx_queue(gp->dev, 0);
txq              1379 drivers/net/ethernet/sun/sungem.c 	__netif_tx_lock(txq, smp_processor_id());
txq              1441 drivers/net/ethernet/sun/sungem.c 	__netif_tx_unlock(txq);
txq               692 drivers/net/ethernet/sun/sunvnet_common.c 	struct netdev_queue *txq;
txq               726 drivers/net/ethernet/sun/sunvnet_common.c 	txq = netdev_get_tx_queue(dev, port->q_index);
txq               727 drivers/net/ethernet/sun/sunvnet_common.c 	if (unlikely(netif_tx_queue_stopped(txq) &&
txq               761 drivers/net/ethernet/sun/sunvnet_common.c 	struct netdev_queue *txq;
txq               763 drivers/net/ethernet/sun/sunvnet_common.c 	txq = netdev_get_tx_queue(VNET_PORT_TO_NET_DEVICE(port),
txq               765 drivers/net/ethernet/sun/sunvnet_common.c 	__netif_tx_lock(txq, smp_processor_id());
txq               766 drivers/net/ethernet/sun/sunvnet_common.c 	if (likely(netif_tx_queue_stopped(txq)))
txq               767 drivers/net/ethernet/sun/sunvnet_common.c 		netif_tx_wake_queue(txq);
txq               768 drivers/net/ethernet/sun/sunvnet_common.c 	__netif_tx_unlock(txq);
txq              1257 drivers/net/ethernet/sun/sunvnet_common.c 		struct netdev_queue *txq;
txq              1259 drivers/net/ethernet/sun/sunvnet_common.c 		txq  = netdev_get_tx_queue(dev, port->q_index);
txq              1260 drivers/net/ethernet/sun/sunvnet_common.c 		netif_tx_stop_queue(txq);
txq              1263 drivers/net/ethernet/sun/sunvnet_common.c 		netif_tx_wake_queue(txq);
txq              1337 drivers/net/ethernet/sun/sunvnet_common.c 	struct netdev_queue *txq;
txq              1397 drivers/net/ethernet/sun/sunvnet_common.c 	txq = netdev_get_tx_queue(dev, i);
txq              1399 drivers/net/ethernet/sun/sunvnet_common.c 		if (!netif_tx_queue_stopped(txq)) {
txq              1400 drivers/net/ethernet/sun/sunvnet_common.c 			netif_tx_stop_queue(txq);
txq              1515 drivers/net/ethernet/sun/sunvnet_common.c 		netif_tx_stop_queue(txq);
txq              1518 drivers/net/ethernet/sun/sunvnet_common.c 			netif_tx_wake_queue(txq);
txq               592 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c 	struct netdev_queue *txq;
txq               608 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c 		txq = netdev_get_tx_queue(netdev, channel->queue_index);
txq               609 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c 		netdev_tx_reset_queue(txq);
txq               707 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c 	struct netdev_queue *txq;
txq               717 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c 	txq = netdev_get_tx_queue(netdev, channel->queue_index);
txq               753 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c 	netdev_tx_sent_queue(txq, tx_pkt_info->tx_bytes);
txq              1044 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c 	struct netdev_queue *txq;
txq              1060 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c 	txq = netdev_get_tx_queue(netdev, channel->queue_index);
txq              1094 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c 	netdev_tx_completed_queue(txq, tx_packets, tx_bytes);
txq              1099 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c 		netif_tx_wake_queue(txq);
txq               378 drivers/net/ethernet/ti/cpsw.c 	struct netdev_queue	*txq;
txq               399 drivers/net/ethernet/ti/cpsw.c 	txq = netdev_get_tx_queue(ndev, ch);
txq               400 drivers/net/ethernet/ti/cpsw.c 	if (unlikely(netif_tx_queue_stopped(txq)))
txq               401 drivers/net/ethernet/ti/cpsw.c 		netif_tx_wake_queue(txq);
txq              1801 drivers/net/ethernet/ti/cpsw.c 	struct netdev_queue *txq;
txq              1820 drivers/net/ethernet/ti/cpsw.c 	txq = netdev_get_tx_queue(ndev, q_idx);
txq              1833 drivers/net/ethernet/ti/cpsw.c 		netif_tx_stop_queue(txq);
txq              1839 drivers/net/ethernet/ti/cpsw.c 			netif_tx_wake_queue(txq);
txq              1845 drivers/net/ethernet/ti/cpsw.c 	netif_tx_stop_queue(txq);
txq              1851 drivers/net/ethernet/ti/cpsw.c 		netif_tx_wake_queue(txq);
txq                84 drivers/net/ethernet/ti/cpts.c 	skb_queue_walk_safe(&cpts->txq, skb, tmp) {
txq                87 drivers/net/ethernet/ti/cpts.c 			__skb_unlink(skb, &cpts->txq);
txq               108 drivers/net/ethernet/ti/cpts.c 	skb_queue_walk_safe(&cpts->txq, skb, tmp) {
txq               121 drivers/net/ethernet/ti/cpts.c 			__skb_unlink(skb, &cpts->txq);
txq               131 drivers/net/ethernet/ti/cpts.c 			__skb_unlink(skb, &cpts->txq);
txq               303 drivers/net/ethernet/ti/cpts.c 	if (!skb_queue_empty(&cpts->txq)) {
txq               305 drivers/net/ethernet/ti/cpts.c 		if (!skb_queue_empty(&cpts->txq))
txq               409 drivers/net/ethernet/ti/cpts.c 		__skb_queue_tail(&cpts->txq, skb);
txq               451 drivers/net/ethernet/ti/cpts.c 	skb_queue_head_init(&cpts->txq);
txq               493 drivers/net/ethernet/ti/cpts.c 	skb_queue_purge(&cpts->txq);
txq               116 drivers/net/ethernet/ti/cpts.h 	struct sk_buff_head txq;
txq               710 drivers/net/hyperv/netvsc.c 		struct netdev_queue *txq = netdev_get_tx_queue(ndev, q_idx);
txq               712 drivers/net/hyperv/netvsc.c 		if (netif_tx_queue_stopped(txq) && !net_device->tx_disable &&
txq               715 drivers/net/hyperv/netvsc.c 			netif_tx_wake_queue(txq);
txq               818 drivers/net/hyperv/netvsc.c 	struct netdev_queue *txq = netdev_get_tx_queue(ndev, packet->q_idx);
txq               861 drivers/net/hyperv/netvsc.c 			netif_tx_stop_queue(txq);
txq               865 drivers/net/hyperv/netvsc.c 		netif_tx_stop_queue(txq);
txq               874 drivers/net/hyperv/netvsc.c 	if (netif_tx_queue_stopped(txq) &&
txq               877 drivers/net/hyperv/netvsc.c 		netif_tx_wake_queue(txq);
txq               347 drivers/net/hyperv/netvsc_drv.c 	u16 txq;
txq               355 drivers/net/hyperv/netvsc_drv.c 			txq = vf_ops->ndo_select_queue(vf_netdev, skb, sb_dev);
txq               357 drivers/net/hyperv/netvsc_drv.c 			txq = netdev_pick_tx(vf_netdev, skb, NULL);
txq               363 drivers/net/hyperv/netvsc_drv.c 		qdisc_skb_cb(skb)->slave_dev_queue_mapping = txq;
txq               365 drivers/net/hyperv/netvsc_drv.c 		txq = netvsc_pick_tx(ndev, skb);
txq               369 drivers/net/hyperv/netvsc_drv.c 	while (unlikely(txq >= ndev->real_num_tx_queues))
txq               370 drivers/net/hyperv/netvsc_drv.c 		txq -= ndev->real_num_tx_queues;
txq               372 drivers/net/hyperv/netvsc_drv.c 	return txq;
txq                65 drivers/net/ifb.c 	struct netdev_queue *txq;
txq                68 drivers/net/ifb.c 	txq = netdev_get_tx_queue(txp->dev, txp->txqnum);
txq                71 drivers/net/ifb.c 		if (!__netif_tx_trylock(txq))
txq                74 drivers/net/ifb.c 		__netif_tx_unlock(txq);
txq               107 drivers/net/ifb.c 	if (__netif_tx_trylock(txq)) {
txq               111 drivers/net/ifb.c 			if (netif_tx_queue_stopped(txq))
txq               112 drivers/net/ifb.c 				netif_tx_wake_queue(txq);
txq               114 drivers/net/ifb.c 			__netif_tx_unlock(txq);
txq               117 drivers/net/ifb.c 		__netif_tx_unlock(txq);
txq               123 drivers/net/net_failover.c 	u16 txq;
txq               130 drivers/net/net_failover.c 			txq = ops->ndo_select_queue(primary_dev, skb, sb_dev);
txq               132 drivers/net/net_failover.c 			txq = netdev_pick_tx(primary_dev, skb, NULL);
txq               136 drivers/net/net_failover.c 		return txq;
txq               139 drivers/net/net_failover.c 	txq = skb_rx_queue_recorded(skb) ? skb_get_rx_queue(skb) : 0;
txq               144 drivers/net/net_failover.c 	if (unlikely(txq >= dev->real_num_tx_queues)) {
txq               146 drivers/net/net_failover.c 			txq -= dev->real_num_tx_queues;
txq               147 drivers/net/net_failover.c 		} while (txq >= dev->real_num_tx_queues);
txq               150 drivers/net/net_failover.c 	return txq;
txq              1736 drivers/net/team/team.c 	u16 txq = skb_rx_queue_recorded(skb) ? skb_get_rx_queue(skb) : 0;
txq              1743 drivers/net/team/team.c 	if (unlikely(txq >= dev->real_num_tx_queues)) {
txq              1745 drivers/net/team/team.c 			txq -= dev->real_num_tx_queues;
txq              1746 drivers/net/team/team.c 		} while (txq >= dev->real_num_tx_queues);
txq              1748 drivers/net/team/team.c 	return txq;
txq               568 drivers/net/tun.c 	u32 txq = 0;
txq               573 drivers/net/tun.c 	txq = __skb_get_hash_symmetric(skb);
txq               574 drivers/net/tun.c 	e = tun_flow_find(&tun->flows[tun_hashfn(txq)], txq);
txq               576 drivers/net/tun.c 		tun_flow_save_rps_rxhash(e, txq);
txq               577 drivers/net/tun.c 		txq = e->queue_index;
txq               580 drivers/net/tun.c 		txq = ((u64)txq * numqueues) >> 32;
txq               583 drivers/net/tun.c 	return txq;
txq              1065 drivers/net/tun.c 	int txq = skb->queue_mapping;
txq              1070 drivers/net/tun.c 	tfile = rcu_dereference(tun->tfiles[txq]);
txq               372 drivers/net/usb/lan78xx.c 	struct sk_buff_head	txq;
txq              2675 drivers/net/usb/lan78xx.c 	temp = unlink_urbs(dev, &dev->txq) + unlink_urbs(dev, &dev->rxq);
txq              2679 drivers/net/usb/lan78xx.c 	       !skb_queue_empty(&dev->txq) &&
txq              2829 drivers/net/usb/lan78xx.c 	defer_bh(dev, skb, &dev->txq, tx_done);
txq              3382 drivers/net/usb/lan78xx.c 	spin_lock_irqsave(&dev->txq.lock, flags);
txq              3385 drivers/net/usb/lan78xx.c 		spin_unlock_irqrestore(&dev->txq.lock, flags);
txq              3405 drivers/net/usb/lan78xx.c 		spin_unlock_irqrestore(&dev->txq.lock, flags);
txq              3415 drivers/net/usb/lan78xx.c 		lan78xx_queue_skb(&dev->txq, skb, tx_start);
txq              3416 drivers/net/usb/lan78xx.c 		if (skb_queue_len(&dev->txq) >= dev->tx_qlen)
txq              3431 drivers/net/usb/lan78xx.c 	spin_unlock_irqrestore(&dev->txq.lock, flags);
txq              3463 drivers/net/usb/lan78xx.c 	if (skb_queue_len(&dev->txq) < dev->tx_qlen)
txq              3519 drivers/net/usb/lan78xx.c 		unlink_urbs(dev, &dev->txq);
txq              3668 drivers/net/usb/lan78xx.c 	unlink_urbs(dev, &dev->txq);
txq              3740 drivers/net/usb/lan78xx.c 	skb_queue_head_init(&dev->txq);
txq              4017 drivers/net/usb/lan78xx.c 		spin_lock_irq(&dev->txq.lock);
txq              4019 drivers/net/usb/lan78xx.c 		if ((skb_queue_len(&dev->txq) ||
txq              4022 drivers/net/usb/lan78xx.c 			spin_unlock_irq(&dev->txq.lock);
txq              4027 drivers/net/usb/lan78xx.c 			spin_unlock_irq(&dev->txq.lock);
txq              4121 drivers/net/usb/lan78xx.c 		spin_lock_irq(&dev->txq.lock);
txq              4131 drivers/net/usb/lan78xx.c 				lan78xx_queue_skb(&dev->txq, skb, tx_start);
txq              4136 drivers/net/usb/lan78xx.c 		spin_unlock_irq(&dev->txq.lock);
txq              4139 drivers/net/usb/lan78xx.c 			if (!(skb_queue_len(&dev->txq) >= dev->tx_qlen))
txq               786 drivers/net/usb/usbnet.c 	temp = unlink_urbs(dev, &dev->txq) +
txq               791 drivers/net/usb/usbnet.c 	wait_skb_queue_empty(&dev->txq);
txq              1140 drivers/net/usb/usbnet.c 		unlink_urbs (dev, &dev->txq);
txq              1291 drivers/net/usb/usbnet.c 	(void) defer_bh(dev, skb, &dev->txq, tx_done);
txq              1300 drivers/net/usb/usbnet.c 	unlink_urbs (dev, &dev->txq);
txq              1425 drivers/net/usb/usbnet.c 	spin_lock_irqsave(&dev->txq.lock, flags);
txq              1428 drivers/net/usb/usbnet.c 		spin_unlock_irqrestore(&dev->txq.lock, flags);
txq              1433 drivers/net/usb/usbnet.c 		spin_unlock_irqrestore(&dev->txq.lock, flags);
txq              1445 drivers/net/usb/usbnet.c 		spin_unlock_irqrestore(&dev->txq.lock, flags);
txq              1464 drivers/net/usb/usbnet.c 		__usbnet_queue_skb(&dev->txq, skb, tx_start);
txq              1465 drivers/net/usb/usbnet.c 		if (dev->txq.qlen >= TX_QLEN (dev))
txq              1468 drivers/net/usb/usbnet.c 	spin_unlock_irqrestore (&dev->txq.lock, flags);
txq              1549 drivers/net/usb/usbnet.c 		if (dev->txq.qlen + dev->rxq.qlen + dev->done.qlen == 0)
txq              1571 drivers/net/usb/usbnet.c 		if (dev->txq.qlen < TX_QLEN (dev))
txq              1700 drivers/net/usb/usbnet.c 	skb_queue_head_init (&dev->txq);
txq              1853 drivers/net/usb/usbnet.c 		spin_lock_irq(&dev->txq.lock);
txq              1855 drivers/net/usb/usbnet.c 		if (dev->txq.qlen && PMSG_IS_AUTO(message)) {
txq              1857 drivers/net/usb/usbnet.c 			spin_unlock_irq(&dev->txq.lock);
txq              1861 drivers/net/usb/usbnet.c 			spin_unlock_irq(&dev->txq.lock);
txq              1892 drivers/net/usb/usbnet.c 		spin_lock_irq(&dev->txq.lock);
txq              1904 drivers/net/usb/usbnet.c 				__skb_queue_tail(&dev->txq, skb);
txq              1910 drivers/net/usb/usbnet.c 		spin_unlock_irq(&dev->txq.lock);
txq              1921 drivers/net/usb/usbnet.c 			if (!(dev->txq.qlen >= TX_QLEN(dev)))
txq               267 drivers/net/virtio_net.c static int txq2vq(int txq)
txq               269 drivers/net/virtio_net.c 	return txq * 2 + 1;
txq              1405 drivers/net/virtio_net.c 	struct netdev_queue *txq = netdev_get_tx_queue(vi->dev, index);
txq              1410 drivers/net/virtio_net.c 	if (__netif_tx_trylock(txq)) {
txq              1412 drivers/net/virtio_net.c 		__netif_tx_unlock(txq);
txq              1416 drivers/net/virtio_net.c 		netif_tx_wake_queue(txq);
txq              1485 drivers/net/virtio_net.c 	struct netdev_queue *txq;
txq              1493 drivers/net/virtio_net.c 	txq = netdev_get_tx_queue(vi->dev, index);
txq              1494 drivers/net/virtio_net.c 	__netif_tx_lock(txq, raw_smp_processor_id());
txq              1496 drivers/net/virtio_net.c 	__netif_tx_unlock(txq);
txq              1501 drivers/net/virtio_net.c 		netif_tx_wake_queue(txq);
txq              1559 drivers/net/virtio_net.c 	struct netdev_queue *txq = netdev_get_tx_queue(dev, qnum);
txq              1616 drivers/net/virtio_net.c 	if (kick || netif_xmit_stopped(txq)) {
txq              1946 drivers/net/vmxnet3/vmxnet3_drv.c 			struct vmxnet3_tx_queue *txq = &adapter->tx_queue[i];
txq              1947 drivers/net/vmxnet3/vmxnet3_drv.c 			vmxnet3_tq_tx_complete(txq, adapter);
txq               111 drivers/net/vrf.c 	struct netdev_queue *txq;
txq               117 drivers/net/vrf.c 	txq = netdev_get_tx_queue(dev, 0);
txq               118 drivers/net/vrf.c 	qdisc = rcu_access_pointer(txq->qdisc);
txq               192 drivers/net/wan/dlci.c 		struct netdev_queue *txq = skb_get_tx_queue(dev, skb);
txq               193 drivers/net/wan/dlci.c 		netdev_start_xmit(skb, dlp->slave, txq, false);
txq               439 drivers/net/wan/farsync.c 	struct sk_buff *txq[FST_TXQ_DEPTH];	/* The queue */
txq              1362 drivers/net/wan/farsync.c 				skb = port->txq[port->txqs];
txq              2330 drivers/net/wan/farsync.c 	port->txq[port->txqe] = skb;
txq               121 drivers/net/wireless/ath/ath10k/core.h 	struct ieee80211_txq *txq;
txq              2270 drivers/net/wireless/ath/ath10k/htt.h 			      struct ieee80211_txq *txq);
txq              2272 drivers/net/wireless/ath/ath10k/htt.h 			      struct ieee80211_txq *txq);
txq              3042 drivers/net/wireless/ath/ath10k/htt_rx.c 	struct ieee80211_txq *txq;
txq              3114 drivers/net/wireless/ath/ath10k/htt_rx.c 		txq = ath10k_mac_txq_lookup(ar, peer_id, tid);
txq              3121 drivers/net/wireless/ath/ath10k/htt_rx.c 		if (unlikely(!txq)) {
txq              3130 drivers/net/wireless/ath/ath10k/htt_rx.c 		ieee80211_txq_schedule_start(hw, txq->ac);
txq              3131 drivers/net/wireless/ath/ath10k/htt_rx.c 		may_tx = ieee80211_txq_may_transmit(hw, txq);
txq              3137 drivers/net/wireless/ath/ath10k/htt_rx.c 			ret = ath10k_mac_tx_push_txq(hw, txq);
txq              3144 drivers/net/wireless/ath/ath10k/htt_rx.c 		ieee80211_return_txq(hw, txq, false);
txq              3145 drivers/net/wireless/ath/ath10k/htt_rx.c 		ieee80211_txq_schedule_end(hw, txq->ac);
txq              3150 drivers/net/wireless/ath/ath10k/htt_rx.c 		ath10k_htt_tx_txq_recalc(hw, txq);
txq              3205 drivers/net/wireless/ath/ath10k/htt_rx.c 	struct ieee80211_txq *txq;
txq              3278 drivers/net/wireless/ath/ath10k/htt_rx.c 		txq = ath10k_mac_txq_lookup(ar, peer_id, tid);
txq              3285 drivers/net/wireless/ath/ath10k/htt_rx.c 		if (unlikely(!txq)) {
txq              3292 drivers/net/wireless/ath/ath10k/htt_rx.c 		artxq = (void *)txq->drv_priv;
txq                38 drivers/net/wireless/ath/ath10k/htt_tx.c 				       struct ieee80211_txq *txq)
txq                42 drivers/net/wireless/ath/ath10k/htt_tx.c 	struct ath10k_vif *arvif = (void *)txq->vif->drv_priv;
txq                59 drivers/net/wireless/ath/ath10k/htt_tx.c 	if (txq->sta) {
txq                60 drivers/net/wireless/ath/ath10k/htt_tx.c 		arsta = (void *)txq->sta->drv_priv;
txq                66 drivers/net/wireless/ath/ath10k/htt_tx.c 	tid = txq->tid;
txq                70 drivers/net/wireless/ath/ath10k/htt_tx.c 	ieee80211_txq_get_depth(txq, &frame_cnt, &byte_cnt);
txq               116 drivers/net/wireless/ath/ath10k/htt_tx.c 			      struct ieee80211_txq *txq)
txq               121 drivers/net/wireless/ath/ath10k/htt_tx.c 	__ath10k_htt_tx_txq_recalc(hw, txq);
txq               133 drivers/net/wireless/ath/ath10k/htt_tx.c 			      struct ieee80211_txq *txq)
txq               138 drivers/net/wireless/ath/ath10k/htt_tx.c 	__ath10k_htt_tx_txq_recalc(hw, txq);
txq              3577 drivers/net/wireless/ath/ath10k/mac.c 				    struct ieee80211_txq *txq,
txq              3607 drivers/net/wireless/ath/ath10k/mac.c 	cb->txq = txq;
txq              3929 drivers/net/wireless/ath/ath10k/mac.c static void ath10k_mac_txq_init(struct ieee80211_txq *txq)
txq              3933 drivers/net/wireless/ath/ath10k/mac.c 	if (!txq)
txq              3936 drivers/net/wireless/ath/ath10k/mac.c 	artxq = (void *)txq->drv_priv;
txq              3940 drivers/net/wireless/ath/ath10k/mac.c static void ath10k_mac_txq_unref(struct ath10k *ar, struct ieee80211_txq *txq)
txq              3946 drivers/net/wireless/ath/ath10k/mac.c 	if (!txq)
txq              3952 drivers/net/wireless/ath/ath10k/mac.c 		if (cb->txq == txq)
txq              3953 drivers/net/wireless/ath/ath10k/mac.c 			cb->txq = NULL;
txq              3974 drivers/net/wireless/ath/ath10k/mac.c 		return peer->sta->txq[tid];
txq              3976 drivers/net/wireless/ath/ath10k/mac.c 		return peer->vif->txq;
txq              3982 drivers/net/wireless/ath/ath10k/mac.c 				   struct ieee80211_txq *txq)
txq              3985 drivers/net/wireless/ath/ath10k/mac.c 	struct ath10k_txq *artxq = (void *)txq->drv_priv;
txq              4009 drivers/net/wireless/ath/ath10k/mac.c 				     struct ieee80211_txq *txq,
txq              4016 drivers/net/wireless/ath/ath10k/mac.c 	if (!txq || !txq->sta)
txq              4023 drivers/net/wireless/ath/ath10k/mac.c 	arsta = (struct ath10k_sta *)txq->sta->drv_priv;
txq              4047 drivers/net/wireless/ath/ath10k/mac.c 			   struct ieee80211_txq *txq)
txq              4051 drivers/net/wireless/ath/ath10k/mac.c 	struct ath10k_txq *artxq = (void *)txq->drv_priv;
txq              4052 drivers/net/wireless/ath/ath10k/mac.c 	struct ieee80211_vif *vif = txq->vif;
txq              4053 drivers/net/wireless/ath/ath10k/mac.c 	struct ieee80211_sta *sta = txq->sta;
txq              4070 drivers/net/wireless/ath/ath10k/mac.c 	skb = ieee80211_tx_dequeue(hw, txq);
txq              4079 drivers/net/wireless/ath/ath10k/mac.c 	airtime = ath10k_mac_update_airtime(ar, txq, skb);
txq              4080 drivers/net/wireless/ath/ath10k/mac.c 	ath10k_mac_tx_h_fill_cb(ar, vif, txq, skb, airtime);
txq              4124 drivers/net/wireless/ath/ath10k/mac.c 	struct ieee80211_txq *txq;
txq              4128 drivers/net/wireless/ath/ath10k/mac.c 	while ((txq = ieee80211_next_txq(hw, ac))) {
txq              4129 drivers/net/wireless/ath/ath10k/mac.c 		while (ath10k_mac_tx_can_push(hw, txq)) {
txq              4130 drivers/net/wireless/ath/ath10k/mac.c 			ret = ath10k_mac_tx_push_txq(hw, txq);
txq              4134 drivers/net/wireless/ath/ath10k/mac.c 		ieee80211_return_txq(hw, txq, false);
txq              4135 drivers/net/wireless/ath/ath10k/mac.c 		ath10k_htt_tx_txq_update(hw, txq);
txq              4339 drivers/net/wireless/ath/ath10k/mac.c 	struct ieee80211_txq *txq = NULL;
txq              4349 drivers/net/wireless/ath/ath10k/mac.c 	airtime = ath10k_mac_update_airtime(ar, txq, skb);
txq              4350 drivers/net/wireless/ath/ath10k/mac.c 	ath10k_mac_tx_h_fill_cb(ar, vif, txq, skb, airtime);
txq              4398 drivers/net/wireless/ath/ath10k/mac.c 					struct ieee80211_txq *txq)
txq              4404 drivers/net/wireless/ath/ath10k/mac.c 	ath10k_htt_tx_txq_update(hw, txq);
txq              4408 drivers/net/wireless/ath/ath10k/mac.c 	ac = txq->ac;
txq              4410 drivers/net/wireless/ath/ath10k/mac.c 	txq = ieee80211_next_txq(hw, ac);
txq              4411 drivers/net/wireless/ath/ath10k/mac.c 	if (!txq)
txq              4414 drivers/net/wireless/ath/ath10k/mac.c 	while (ath10k_mac_tx_can_push(hw, txq)) {
txq              4415 drivers/net/wireless/ath/ath10k/mac.c 		ret = ath10k_mac_tx_push_txq(hw, txq);
txq              4419 drivers/net/wireless/ath/ath10k/mac.c 	ieee80211_return_txq(hw, txq, false);
txq              4420 drivers/net/wireless/ath/ath10k/mac.c 	ath10k_htt_tx_txq_update(hw, txq);
txq              5150 drivers/net/wireless/ath/ath10k/mac.c 	ath10k_mac_txq_init(vif->txq);
txq              5584 drivers/net/wireless/ath/ath10k/mac.c 	ath10k_mac_txq_unref(ar, vif->txq);
txq              5601 drivers/net/wireless/ath/ath10k/mac.c 	ath10k_mac_txq_unref(ar, vif->txq);
txq              6471 drivers/net/wireless/ath/ath10k/mac.c 		for (i = 0; i < ARRAY_SIZE(sta->txq); i++)
txq              6472 drivers/net/wireless/ath/ath10k/mac.c 			ath10k_mac_txq_init(sta->txq[i]);
txq              6637 drivers/net/wireless/ath/ath10k/mac.c 		for (i = 0; i < ARRAY_SIZE(sta->txq); i++)
txq              6638 drivers/net/wireless/ath/ath10k/mac.c 			ath10k_mac_txq_unref(ar, sta->txq[i]);
txq                69 drivers/net/wireless/ath/ath10k/mac.h 			   struct ieee80211_txq *txq);
txq                49 drivers/net/wireless/ath/ath10k/txrx.c 	struct ieee80211_txq *txq;
txq                74 drivers/net/wireless/ath/ath10k/txrx.c 	txq = skb_cb->txq;
txq                76 drivers/net/wireless/ath/ath10k/txrx.c 	if (txq) {
txq                77 drivers/net/wireless/ath/ath10k/txrx.c 		artxq = (void *)txq->drv_priv;
txq                87 drivers/net/wireless/ath/ath10k/txrx.c 	if (txq && txq->sta && skb_cb->airtime_est)
txq                88 drivers/net/wireless/ath/ath10k/txrx.c 		ieee80211_sta_register_airtime(txq->sta, txq->tid,
txq               733 drivers/net/wireless/ath/ath5k/base.c 		  struct ath5k_txq *txq, int padsize,
txq               830 drivers/net/wireless/ath/ath5k/base.c 	spin_lock_bh(&txq->lock);
txq               831 drivers/net/wireless/ath/ath5k/base.c 	list_add_tail(&bf->list, &txq->q);
txq               832 drivers/net/wireless/ath/ath5k/base.c 	txq->txq_len++;
txq               833 drivers/net/wireless/ath/ath5k/base.c 	if (txq->link == NULL) /* is this first packet? */
txq               834 drivers/net/wireless/ath/ath5k/base.c 		ath5k_hw_set_txdp(ah, txq->qnum, bf->daddr);
txq               836 drivers/net/wireless/ath/ath5k/base.c 		*txq->link = bf->daddr;
txq               838 drivers/net/wireless/ath/ath5k/base.c 	txq->link = &ds->ds_link;
txq               839 drivers/net/wireless/ath/ath5k/base.c 	ath5k_hw_start_tx_dma(ah, txq->qnum);
txq               840 drivers/net/wireless/ath/ath5k/base.c 	spin_unlock_bh(&txq->lock);
txq               977 drivers/net/wireless/ath/ath5k/base.c 	struct ath5k_txq *txq;
txq              1010 drivers/net/wireless/ath/ath5k/base.c 	txq = &ah->txqs[qnum];
txq              1011 drivers/net/wireless/ath/ath5k/base.c 	if (!txq->setup) {
txq              1012 drivers/net/wireless/ath/ath5k/base.c 		txq->qnum = qnum;
txq              1013 drivers/net/wireless/ath/ath5k/base.c 		txq->link = NULL;
txq              1014 drivers/net/wireless/ath/ath5k/base.c 		INIT_LIST_HEAD(&txq->q);
txq              1015 drivers/net/wireless/ath/ath5k/base.c 		spin_lock_init(&txq->lock);
txq              1016 drivers/net/wireless/ath/ath5k/base.c 		txq->setup = true;
txq              1017 drivers/net/wireless/ath/ath5k/base.c 		txq->txq_len = 0;
txq              1018 drivers/net/wireless/ath/ath5k/base.c 		txq->txq_max = ATH5K_TXQ_LEN_MAX;
txq              1019 drivers/net/wireless/ath/ath5k/base.c 		txq->txq_poll_mark = false;
txq              1020 drivers/net/wireless/ath/ath5k/base.c 		txq->txq_stuck = 0;
txq              1112 drivers/net/wireless/ath/ath5k/base.c 	struct ath5k_txq *txq;
txq              1118 drivers/net/wireless/ath/ath5k/base.c 			txq = &ah->txqs[i];
txq              1119 drivers/net/wireless/ath/ath5k/base.c 			spin_lock_bh(&txq->lock);
txq              1120 drivers/net/wireless/ath/ath5k/base.c 			list_for_each_entry_safe(bf, bf0, &txq->q, list) {
txq              1128 drivers/net/wireless/ath/ath5k/base.c 				txq->txq_len--;
txq              1131 drivers/net/wireless/ath/ath5k/base.c 			txq->link = NULL;
txq              1132 drivers/net/wireless/ath/ath5k/base.c 			txq->txq_poll_mark = false;
txq              1133 drivers/net/wireless/ath/ath5k/base.c 			spin_unlock_bh(&txq->lock);
txq              1141 drivers/net/wireless/ath/ath5k/base.c 	struct ath5k_txq *txq = ah->txqs;
txq              1144 drivers/net/wireless/ath/ath5k/base.c 	for (i = 0; i < ARRAY_SIZE(ah->txqs); i++, txq++)
txq              1145 drivers/net/wireless/ath/ath5k/base.c 		if (txq->setup) {
txq              1146 drivers/net/wireless/ath/ath5k/base.c 			ath5k_hw_release_tx_queue(ah, txq->qnum);
txq              1147 drivers/net/wireless/ath/ath5k/base.c 			txq->setup = false;
txq              1611 drivers/net/wireless/ath/ath5k/base.c 	       struct ath5k_txq *txq, struct ieee80211_tx_control *control)
txq              1618 drivers/net/wireless/ath/ath5k/base.c 	trace_ath5k_tx(ah, skb, txq);
txq              1631 drivers/net/wireless/ath/ath5k/base.c 	if (txq->txq_len >= txq->txq_max &&
txq              1632 drivers/net/wireless/ath/ath5k/base.c 	    txq->qnum <= AR5K_TX_QUEUE_ID_DATA_MAX)
txq              1633 drivers/net/wireless/ath/ath5k/base.c 		ieee80211_stop_queue(hw, txq->qnum);
txq              1651 drivers/net/wireless/ath/ath5k/base.c 	if (ath5k_txbuf_setup(ah, bf, txq, padsize, control)) {
txq              1667 drivers/net/wireless/ath/ath5k/base.c 			 struct ath5k_txq *txq, struct ath5k_tx_status *ts,
txq              1727 drivers/net/wireless/ath/ath5k/base.c 	trace_ath5k_tx_complete(ah, skb, txq, ts);
txq              1732 drivers/net/wireless/ath/ath5k/base.c ath5k_tx_processq(struct ath5k_hw *ah, struct ath5k_txq *txq)
txq              1740 drivers/net/wireless/ath/ath5k/base.c 	spin_lock(&txq->lock);
txq              1741 drivers/net/wireless/ath/ath5k/base.c 	list_for_each_entry_safe(bf, bf0, &txq->q, list) {
txq              1743 drivers/net/wireless/ath/ath5k/base.c 		txq->txq_poll_mark = false;
txq              1755 drivers/net/wireless/ath/ath5k/base.c 					"queue %u\n", ret, txq->qnum);
txq              1764 drivers/net/wireless/ath/ath5k/base.c 			ath5k_tx_frame_completed(ah, skb, txq, &ts, bf);
txq              1773 drivers/net/wireless/ath/ath5k/base.c 		if (ath5k_hw_get_txdp(ah, txq->qnum) != bf->daddr) {
txq              1777 drivers/net/wireless/ath/ath5k/base.c 			txq->txq_len--;
txq              1781 drivers/net/wireless/ath/ath5k/base.c 	spin_unlock(&txq->lock);
txq              1782 drivers/net/wireless/ath/ath5k/base.c 	if (txq->txq_len < ATH5K_TXQ_LEN_LOW && txq->qnum < 4)
txq              1783 drivers/net/wireless/ath/ath5k/base.c 		ieee80211_wake_queue(ah->hw, txq->qnum);
txq              2465 drivers/net/wireless/ath/ath5k/base.c 	struct ath5k_txq *txq;
txq              2476 drivers/net/wireless/ath/ath5k/base.c 			txq = &ah->txqs[i];
txq              2477 drivers/net/wireless/ath/ath5k/base.c 			spin_lock_bh(&txq->lock);
txq              2478 drivers/net/wireless/ath/ath5k/base.c 			if (txq->txq_len > 1) {
txq              2479 drivers/net/wireless/ath/ath5k/base.c 				if (txq->txq_poll_mark) {
txq              2482 drivers/net/wireless/ath/ath5k/base.c 						  txq->qnum);
txq              2484 drivers/net/wireless/ath/ath5k/base.c 					txq->txq_stuck++;
txq              2485 drivers/net/wireless/ath/ath5k/base.c 					spin_unlock_bh(&txq->lock);
txq              2488 drivers/net/wireless/ath/ath5k/base.c 					txq->txq_poll_mark = true;
txq              2491 drivers/net/wireless/ath/ath5k/base.c 			spin_unlock_bh(&txq->lock);
txq              2986 drivers/net/wireless/ath/ath5k/base.c 	struct ath5k_txq *txq;
txq              3036 drivers/net/wireless/ath/ath5k/base.c 		txq = ath5k_txq_setup(ah, AR5K_TX_QUEUE_DATA, AR5K_WME_AC_VO);
txq              3037 drivers/net/wireless/ath/ath5k/base.c 		if (IS_ERR(txq)) {
txq              3039 drivers/net/wireless/ath/ath5k/base.c 			ret = PTR_ERR(txq);
txq              3042 drivers/net/wireless/ath/ath5k/base.c 		txq = ath5k_txq_setup(ah, AR5K_TX_QUEUE_DATA, AR5K_WME_AC_VI);
txq              3043 drivers/net/wireless/ath/ath5k/base.c 		if (IS_ERR(txq)) {
txq              3045 drivers/net/wireless/ath/ath5k/base.c 			ret = PTR_ERR(txq);
txq              3048 drivers/net/wireless/ath/ath5k/base.c 		txq = ath5k_txq_setup(ah, AR5K_TX_QUEUE_DATA, AR5K_WME_AC_BE);
txq              3049 drivers/net/wireless/ath/ath5k/base.c 		if (IS_ERR(txq)) {
txq              3051 drivers/net/wireless/ath/ath5k/base.c 			ret = PTR_ERR(txq);
txq              3054 drivers/net/wireless/ath/ath5k/base.c 		txq = ath5k_txq_setup(ah, AR5K_TX_QUEUE_DATA, AR5K_WME_AC_BK);
txq              3055 drivers/net/wireless/ath/ath5k/base.c 		if (IS_ERR(txq)) {
txq              3057 drivers/net/wireless/ath/ath5k/base.c 			ret = PTR_ERR(txq);
txq              3063 drivers/net/wireless/ath/ath5k/base.c 		txq = ath5k_txq_setup(ah, AR5K_TX_QUEUE_DATA, AR5K_WME_AC_BE);
txq              3064 drivers/net/wireless/ath/ath5k/base.c 		if (IS_ERR(txq)) {
txq              3066 drivers/net/wireless/ath/ath5k/base.c 			ret = PTR_ERR(txq);
txq               108 drivers/net/wireless/ath/ath5k/base.h 		    struct ath5k_txq *txq, struct ieee80211_tx_control *control);
txq               835 drivers/net/wireless/ath/ath5k/debug.c 	struct ath5k_txq *txq;
txq               843 drivers/net/wireless/ath/ath5k/debug.c 		txq = &ah->txqs[i];
txq               846 drivers/net/wireless/ath/ath5k/debug.c 			"%02d: %ssetup\n", i, txq->setup ? "" : "not ");
txq               848 drivers/net/wireless/ath/ath5k/debug.c 		if (!txq->setup)
txq               852 drivers/net/wireless/ath/ath5k/debug.c 		spin_lock_bh(&txq->lock);
txq               853 drivers/net/wireless/ath/ath5k/debug.c 		list_for_each_entry_safe(bf, bf0, &txq->q, list)
txq               855 drivers/net/wireless/ath/ath5k/debug.c 		spin_unlock_bh(&txq->lock);
txq               858 drivers/net/wireless/ath/ath5k/debug.c 				"  len: %d bufs: %d\n", txq->txq_len, n);
txq               860 drivers/net/wireless/ath/ath5k/debug.c 				"  stuck: %d\n", txq->txq_stuck);
txq               272 drivers/net/wireless/ath/ath6kl/debug.c 		   get_queue_depth(&ep_dist->htc_ep->txq));
txq               744 drivers/net/wireless/ath/ath6kl/debug.c 				 get_queue_depth(&ep_list->htc_ep->txq));
txq               511 drivers/net/wireless/ath/ath6kl/htc.h 	struct list_head txq;
txq               426 drivers/net/wireless/ath/ath6kl/htc_mbox.c 	endpoint->cred_dist.txq_depth = get_queue_depth(&endpoint->txq);
txq               439 drivers/net/wireless/ath/ath6kl/htc_mbox.c 			    struct list_head *txq)
txq               441 drivers/net/wireless/ath/ath6kl/htc_mbox.c 	if (list_empty(txq))
txq               446 drivers/net/wireless/ath/ath6kl/htc_mbox.c 		   endpoint->eid, get_queue_depth(txq));
txq               448 drivers/net/wireless/ath/ath6kl/htc_mbox.c 	ath6kl_tx_complete(endpoint->target, txq);
txq               611 drivers/net/wireless/ath/ath6kl/htc_mbox.c 		if (list_empty(&endpoint->txq))
txq               613 drivers/net/wireless/ath/ath6kl/htc_mbox.c 		packet = list_first_entry(&endpoint->txq, struct htc_packet,
txq               618 drivers/net/wireless/ath/ath6kl/htc_mbox.c 			   packet, get_queue_depth(&endpoint->txq));
txq               628 drivers/net/wireless/ath/ath6kl/htc_mbox.c 		packet = list_first_entry(&endpoint->txq, struct htc_packet,
txq               861 drivers/net/wireless/ath/ath6kl/htc_mbox.c 	struct list_head txq;
txq               882 drivers/net/wireless/ath/ath6kl/htc_mbox.c 	INIT_LIST_HEAD(&txq);
txq               889 drivers/net/wireless/ath/ath6kl/htc_mbox.c 		if (list_empty(&endpoint->txq))
txq               892 drivers/net/wireless/ath/ath6kl/htc_mbox.c 		ath6kl_htc_tx_pkts_get(target, endpoint, &txq);
txq               894 drivers/net/wireless/ath/ath6kl/htc_mbox.c 		if (list_empty(&txq))
txq               905 drivers/net/wireless/ath/ath6kl/htc_mbox.c 			    (get_queue_depth(&txq) >=
txq               911 drivers/net/wireless/ath/ath6kl/htc_mbox.c 					ath6kl_htc_tx_bundle(endpoint, &txq,
txq               918 drivers/net/wireless/ath/ath6kl/htc_mbox.c 			if (list_empty(&txq))
txq               921 drivers/net/wireless/ath/ath6kl/htc_mbox.c 			packet = list_first_entry(&txq, struct htc_packet,
txq               976 drivers/net/wireless/ath/ath6kl/htc_mbox.c 	txq_depth = get_queue_depth(&endpoint->txq);
txq               997 drivers/net/wireless/ath/ath6kl/htc_mbox.c 	list_add_tail(&tx_pkt->list, &endpoint->txq);
txq              1020 drivers/net/wireless/ath/ath6kl/htc_mbox.c 		if (!list_empty(&endpoint->txq)) {
txq              1025 drivers/net/wireless/ath/ath6kl/htc_mbox.c 				   get_queue_depth(&endpoint->txq));
txq              1167 drivers/net/wireless/ath/ath6kl/htc_mbox.c 	list_for_each_entry_safe(packet, tmp_pkt, &endpoint->txq, list) {
txq              1233 drivers/net/wireless/ath/ath6kl/htc_mbox.c 			get_queue_depth(&endpoint->txq);
txq              1649 drivers/net/wireless/ath/ath6kl/htc_mbox.c 			get_queue_depth(&endpoint->txq);
txq              2588 drivers/net/wireless/ath/ath6kl/htc_mbox.c 		INIT_LIST_HEAD(&endpoint->txq);
txq               106 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		if (list_empty(&ep->txq))
txq               110 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		packet = list_first_entry(&ep->txq, struct htc_packet, list);
txq               114 drivers/net/wireless/ath/ath6kl/htc_pipe.c 			   __func__, packet, get_queue_depth(&ep->txq));
txq               159 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		packet = list_first_entry(&ep->txq, struct htc_packet, list);
txq               183 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		if (list_empty(&ep->txq))
txq               186 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		packet = list_first_entry(&ep->txq, struct htc_packet, list);
txq               191 drivers/net/wireless/ath/ath6kl/htc_pipe.c 			   __func__, packet, get_queue_depth(&ep->txq));
txq               302 drivers/net/wireless/ath/ath6kl/htc_pipe.c 					       struct list_head *txq)
txq               312 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		   __func__, txq,
txq               313 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		   (txq == NULL) ? 0 : get_queue_depth(txq));
txq               323 drivers/net/wireless/ath/ath6kl/htc_pipe.c 	if (txq != NULL) {
txq               324 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		if (list_empty(txq)) {
txq               330 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		txqueue_depth = get_queue_depth(&ep->txq);
txq               335 drivers/net/wireless/ath/ath6kl/htc_pipe.c 			overflow = get_queue_depth(txq);
txq               339 drivers/net/wireless/ath/ath6kl/htc_pipe.c 			overflow += get_queue_depth(txq);
txq               358 drivers/net/wireless/ath/ath6kl/htc_pipe.c 			list_splice_tail_init(txq, &send_queue);
txq               360 drivers/net/wireless/ath/ath6kl/htc_pipe.c 			good_pkts = get_queue_depth(txq) - overflow;
txq               370 drivers/net/wireless/ath/ath6kl/htc_pipe.c 				packet = list_first_entry(txq,
txq               383 drivers/net/wireless/ath/ath6kl/htc_pipe.c 						 txq, list) {
txq               421 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		list_splice_tail_init(&send_queue, &ep->txq);
txq               451 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		if (get_queue_depth(&ep->txq) == 0)
txq               675 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		if (ep->cred_dist.credits && get_queue_depth(&ep->txq)) {
txq               697 drivers/net/wireless/ath/ath6kl/htc_pipe.c 	while (get_queue_depth(&ep->txq)) {
txq               698 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		packet = list_first_entry(&ep->txq, struct htc_packet, list);
txq              1183 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		INIT_LIST_HEAD(&ep->txq);
txq               177 drivers/net/wireless/ath/ath9k/ath9k.h 	s8 txq;
txq               235 drivers/net/wireless/ath/ath9k/ath9k.h 	struct ath_txq *txq;
txq               273 drivers/net/wireless/ath/ath9k/ath9k.h 	struct ath_txq *txq;
txq               289 drivers/net/wireless/ath/ath9k/ath9k.h 	struct ath_txq txq[ATH9K_NUM_TX_QUEUES];
txq               431 drivers/net/wireless/ath/ath9k/ath9k.h 	struct ieee80211_txq *txq;
txq               435 drivers/net/wireless/ath/ath9k/ath9k.h 		txq = sta->txq[tidno % ARRAY_SIZE(sta->txq)];
txq               437 drivers/net/wireless/ath/ath9k/ath9k.h 		txq = vif->txq;
txq               439 drivers/net/wireless/ath/ath9k/ath9k.h 	return (struct ath_atx_tid *) txq->drv_priv;
txq               565 drivers/net/wireless/ath/ath9k/ath9k.h static inline void ath_txq_lock(struct ath_softc *sc, struct ath_txq *txq)
txq               567 drivers/net/wireless/ath/ath9k/ath9k.h 	spin_lock_bh(&txq->axq_lock);
txq               569 drivers/net/wireless/ath/ath9k/ath9k.h static inline void ath_txq_unlock(struct ath_softc *sc, struct ath_txq *txq)
txq               571 drivers/net/wireless/ath/ath9k/ath9k.h 	spin_unlock_bh(&txq->axq_lock);
txq               581 drivers/net/wireless/ath/ath9k/ath9k.h void ath_txq_unlock_complete(struct ath_softc *sc, struct ath_txq *txq);
txq               582 drivers/net/wireless/ath/ath9k/ath9k.h void ath_tx_cleanupq(struct ath_softc *sc, struct ath_txq *txq);
txq               584 drivers/net/wireless/ath/ath9k/ath9k.h void ath_draintxq(struct ath_softc *sc, struct ath_txq *txq);
txq               587 drivers/net/wireless/ath/ath9k/ath9k.h void ath_txq_schedule(struct ath_softc *sc, struct ath_txq *txq);
txq                38 drivers/net/wireless/ath/ath9k/beacon.c 	struct ath_txq *txq;
txq                50 drivers/net/wireless/ath/ath9k/beacon.c 		txq = sc->tx.txq_map[IEEE80211_AC_BE];
txq                51 drivers/net/wireless/ath/ath9k/beacon.c 		ath9k_hw_get_txq_props(ah, txq->axq_qnum, &qi_be);
txq              1016 drivers/net/wireless/ath/ath9k/channel.c 	txctl.txq = sc->tx.txq_map[IEEE80211_AC_VO];
txq              1137 drivers/net/wireless/ath/ath9k/channel.c 	txctl.txq = sc->tx.txq_map[IEEE80211_AC_VO];
txq               623 drivers/net/wireless/ath/ath9k/debug.c static void print_queue(struct ath_softc *sc, struct ath_txq *txq,
txq               626 drivers/net/wireless/ath/ath9k/debug.c 	ath_txq_lock(sc, txq);
txq               628 drivers/net/wireless/ath/ath9k/debug.c 	seq_printf(file, "%s: %d ", "qnum", txq->axq_qnum);
txq               629 drivers/net/wireless/ath/ath9k/debug.c 	seq_printf(file, "%s: %2d ", "qdepth", txq->axq_depth);
txq               630 drivers/net/wireless/ath/ath9k/debug.c 	seq_printf(file, "%s: %2d ", "ampdu-depth", txq->axq_ampdu_depth);
txq               631 drivers/net/wireless/ath/ath9k/debug.c 	seq_printf(file, "%s: %3d\n", "pending", txq->pending_frames);
txq               633 drivers/net/wireless/ath/ath9k/debug.c 	ath_txq_unlock(sc, txq);
txq               640 drivers/net/wireless/ath/ath9k/debug.c 	struct ath_txq *txq;
txq               647 drivers/net/wireless/ath/ath9k/debug.c 		txq = sc->tx.txq_map[i];
txq               649 drivers/net/wireless/ath/ath9k/debug.c 		print_queue(sc, txq, file);
txq               783 drivers/net/wireless/ath/ath9k/debug.c 		       struct ath_tx_status *ts, struct ath_txq *txq,
txq               786 drivers/net/wireless/ath/ath9k/debug.c 	int qnum = txq->axq_qnum;
txq               259 drivers/net/wireless/ath/ath9k/debug.h 		       struct ath_tx_status *ts, struct ath_txq *txq,
txq               296 drivers/net/wireless/ath/ath9k/debug.h 				     struct ath_txq *txq,
txq                29 drivers/net/wireless/ath/ath9k/debug_sta.c 	struct ath_txq *txq;
txq                57 drivers/net/wireless/ath/ath9k/debug_sta.c 		txq = tid->txq;
txq                58 drivers/net/wireless/ath/ath9k/debug_sta.c 		ath_txq_lock(sc, txq);
txq                71 drivers/net/wireless/ath/ath9k/debug_sta.c 		ath_txq_unlock(sc, txq);
txq               414 drivers/net/wireless/ath/ath9k/gpio.c 	struct ath_txq *txq;
txq               429 drivers/net/wireless/ath/ath9k/gpio.c 		txq = sc->tx.txq_map[IEEE80211_AC_BE];
txq               430 drivers/net/wireless/ath/ath9k/gpio.c 		ath9k_hw_init_btcoex_hw(sc->sc_ah, txq->axq_qnum);
txq               821 drivers/net/wireless/ath/ath9k/hw.h 	struct ath9k_tx_queue_info txq[ATH9K_NUM_TX_QUEUES];
txq               784 drivers/net/wireless/ath/ath9k/init.c 			ath_tx_cleanupq(sc, &sc->tx.txq[i]);
txq              1109 drivers/net/wireless/ath/ath9k/init.c 			ath_tx_cleanupq(sc, &sc->tx.txq[i]);
txq                25 drivers/net/wireless/ath/ath9k/link.c 	struct ath_txq *txq;
txq                32 drivers/net/wireless/ath/ath9k/link.c 		txq = sc->tx.txq_map[i];
txq                34 drivers/net/wireless/ath/ath9k/link.c 		ath_txq_lock(sc, txq);
txq                35 drivers/net/wireless/ath/ath9k/link.c 		if (txq->axq_depth) {
txq                36 drivers/net/wireless/ath/ath9k/link.c 			if (txq->axq_tx_inprogress) {
txq                37 drivers/net/wireless/ath/ath9k/link.c 				ath_txq_unlock(sc, txq);
txq                41 drivers/net/wireless/ath/ath9k/link.c 			txq->axq_tx_inprogress = true;
txq                43 drivers/net/wireless/ath/ath9k/link.c 		ath_txq_unlock(sc, txq);
txq               182 drivers/net/wireless/ath/ath9k/link.c 	txctl.txq = sc->tx.txq_map[IEEE80211_AC_BE];
txq               203 drivers/net/wireless/ath/ath9k/mac.c 	qi = &ah->txq[q];
txq               267 drivers/net/wireless/ath/ath9k/mac.c 	qi = &ah->txq[q];
txq               323 drivers/net/wireless/ath/ath9k/mac.c 	qi = &ah->txq[q];
txq               351 drivers/net/wireless/ath/ath9k/mac.c 	qi = &ah->txq[q];
txq               373 drivers/net/wireless/ath/ath9k/mac.c 	qi = &ah->txq[q];
txq                57 drivers/net/wireless/ath/ath9k/main.c static bool ath9k_has_pending_frames(struct ath_softc *sc, struct ath_txq *txq,
txq                62 drivers/net/wireless/ath/ath9k/main.c 	spin_lock_bh(&txq->axq_lock);
txq                64 drivers/net/wireless/ath/ath9k/main.c 	if (txq->axq_depth) {
txq                72 drivers/net/wireless/ath/ath9k/main.c 	if (txq->mac80211_qnum >= 0) {
txq                75 drivers/net/wireless/ath/ath9k/main.c 		acq = &sc->cur_chan->acq[txq->mac80211_qnum];
txq                80 drivers/net/wireless/ath/ath9k/main.c 	spin_unlock_bh(&txq->axq_lock);
txq               805 drivers/net/wireless/ath/ath9k/main.c 	txctl.txq = sc->tx.txq_map[skb_get_queue_mapping(skb)];
txq               812 drivers/net/wireless/ath/ath9k/main.c 		TX_STAT_INC(sc, txctl.txq->axq_qnum, txfailed);
txq              1636 drivers/net/wireless/ath/ath9k/main.c 	struct ath_txq *txq;
txq              1643 drivers/net/wireless/ath/ath9k/main.c 	txq = sc->tx.txq_map[queue];
txq              1657 drivers/net/wireless/ath/ath9k/main.c 		queue, txq->axq_qnum, params->aifs, params->cw_min,
txq              1661 drivers/net/wireless/ath/ath9k/main.c 	ret = ath_txq_update(sc, txq->axq_qnum, &qi);
txq              2048 drivers/net/wireless/ath/ath9k/main.c 		npend = ath9k_has_pending_frames(sc, &sc->tx.txq[i],
txq               126 drivers/net/wireless/ath/ath9k/tx99.c 	txctl.txq = sc->tx.txq_map[IEEE80211_AC_VO];
txq                50 drivers/net/wireless/ath/ath9k/xmit.c static void ath_tx_send_normal(struct ath_softc *sc, struct ath_txq *txq,
txq                53 drivers/net/wireless/ath/ath9k/xmit.c 			    int tx_flags, struct ath_txq *txq,
txq                56 drivers/net/wireless/ath/ath9k/xmit.c 				struct ath_txq *txq, struct list_head *bf_q,
txq                59 drivers/net/wireless/ath/ath9k/xmit.c static void ath_tx_txqaddbuf(struct ath_softc *sc, struct ath_txq *txq,
txq                67 drivers/net/wireless/ath/ath9k/xmit.c 					   struct ath_txq *txq,
txq               101 drivers/net/wireless/ath/ath9k/xmit.c void ath_txq_unlock_complete(struct ath_softc *sc, struct ath_txq *txq)
txq               102 drivers/net/wireless/ath/ath9k/xmit.c 	__releases(&txq->axq_lock)
txq               109 drivers/net/wireless/ath/ath9k/xmit.c 	skb_queue_splice_init(&txq->complete_q, &q);
txq               110 drivers/net/wireless/ath/ath9k/xmit.c 	spin_unlock_bh(&txq->axq_lock);
txq               129 drivers/net/wireless/ath/ath9k/xmit.c 	struct ath_txq *txq = tid->txq;
txq               135 drivers/net/wireless/ath/ath9k/xmit.c 	ath_txq_lock(sc, txq);
txq               136 drivers/net/wireless/ath/ath9k/xmit.c 	ath_txq_schedule(sc, txq);
txq               137 drivers/net/wireless/ath/ath9k/xmit.c 	ath_txq_unlock(sc, txq);
txq               164 drivers/net/wireless/ath/ath9k/xmit.c static void ath_txq_skb_done(struct ath_softc *sc, struct ath_txq *txq,
txq               168 drivers/net/wireless/ath/ath9k/xmit.c 	int q = fi->txq;
txq               173 drivers/net/wireless/ath/ath9k/xmit.c 	txq = sc->tx.txq_map[q];
txq               174 drivers/net/wireless/ath/ath9k/xmit.c 	if (WARN_ON(--txq->pending_frames < 0))
txq               175 drivers/net/wireless/ath/ath9k/xmit.c 		txq->pending_frames = 0;
txq               189 drivers/net/wireless/ath/ath9k/xmit.c 	struct ieee80211_txq *txq = container_of((void*)tid, struct ieee80211_txq, drv_priv);
txq               193 drivers/net/wireless/ath/ath9k/xmit.c 		.txq = tid->txq,
txq               200 drivers/net/wireless/ath/ath9k/xmit.c 	skb = ieee80211_tx_dequeue(hw, txq);
txq               211 drivers/net/wireless/ath/ath9k/xmit.c 	if (tid->txq == sc->tx.txq_map[q]) {
txq               213 drivers/net/wireless/ath/ath9k/xmit.c 		fi->txq = q;
txq               214 drivers/net/wireless/ath/ath9k/xmit.c 		++tid->txq->pending_frames;
txq               234 drivers/net/wireless/ath/ath9k/xmit.c 	struct ath_txq *txq = tid->txq;
txq               250 drivers/net/wireless/ath/ath9k/xmit.c 			ath_txq_skb_done(sc, txq, skb);
txq               261 drivers/net/wireless/ath/ath9k/xmit.c 		ath_tx_complete_buf(sc, bf, txq, &bf_head, NULL, &ts, 0);
txq               265 drivers/net/wireless/ath/ath9k/xmit.c 		ath_txq_unlock(sc, txq);
txq               267 drivers/net/wireless/ath/ath9k/xmit.c 		ath_txq_lock(sc, txq);
txq               316 drivers/net/wireless/ath/ath9k/xmit.c static void ath_tid_drain(struct ath_softc *sc, struct ath_txq *txq,
txq               335 drivers/net/wireless/ath/ath9k/xmit.c 			ath_tx_complete(sc, skb, ATH_TX_ERROR, txq, NULL);
txq               340 drivers/net/wireless/ath/ath9k/xmit.c 		ath_tx_complete_buf(sc, bf, txq, &bf_head, NULL, &ts, 0);
txq               344 drivers/net/wireless/ath/ath9k/xmit.c static void ath_tx_set_retry(struct ath_softc *sc, struct ath_txq *txq,
txq               352 drivers/net/wireless/ath/ath9k/xmit.c 	TX_STAT_INC(sc, txq->axq_qnum, a_retries);
txq               439 drivers/net/wireless/ath/ath9k/xmit.c static void ath_tx_complete_aggr(struct ath_softc *sc, struct ath_txq *txq,
txq               479 drivers/net/wireless/ath/ath9k/xmit.c 			ath_tx_complete_buf(sc, bf, txq, &bf_head, NULL, ts, 0);
txq               552 drivers/net/wireless/ath/ath9k/xmit.c 				ath_tx_set_retry(sc, txq, bf->bf_mpdu,
txq               588 drivers/net/wireless/ath/ath9k/xmit.c 			ath_tx_complete_buf(sc, bf, txq, &bf_head, sta, ts,
txq               608 drivers/net/wireless/ath/ath9k/xmit.c 					ath_tx_complete_buf(sc, bf, txq,
txq               648 drivers/net/wireless/ath/ath9k/xmit.c 		ath_txq_unlock(sc, txq);
txq               650 drivers/net/wireless/ath/ath9k/xmit.c 		ath_txq_lock(sc, txq);
txq               681 drivers/net/wireless/ath/ath9k/xmit.c static void ath_tx_process_buffer(struct ath_softc *sc, struct ath_txq *txq,
txq               694 drivers/net/wireless/ath/ath9k/xmit.c 	txq->axq_tx_inprogress = false;
txq               696 drivers/net/wireless/ath/ath9k/xmit.c 	txq->axq_depth--;
txq               698 drivers/net/wireless/ath/ath9k/xmit.c 		txq->axq_ampdu_depth--;
txq               722 drivers/net/wireless/ath/ath9k/xmit.c 		ath_tx_complete_buf(sc, bf, txq, bf_head, sta, ts, txok);
txq               724 drivers/net/wireless/ath/ath9k/xmit.c 		ath_tx_complete_aggr(sc, txq, bf, bf_head, sta, tid, ts, txok);
txq               727 drivers/net/wireless/ath/ath9k/xmit.c 		ath_txq_schedule(sc, txq);
txq               760 drivers/net/wireless/ath/ath9k/xmit.c 	int q = tid->txq->mac80211_qnum;
txq               894 drivers/net/wireless/ath/ath9k/xmit.c ath_tx_get_tid_subframe(struct ath_softc *sc, struct ath_txq *txq,
txq               912 drivers/net/wireless/ath/ath9k/xmit.c 			bf = ath_tx_setup_buffer(sc, txq, tid, skb);
txq               917 drivers/net/wireless/ath/ath9k/xmit.c 			ath_txq_skb_done(sc, txq, skb);
txq               968 drivers/net/wireless/ath/ath9k/xmit.c 			ath_tx_complete_buf(sc, bf, txq, &bf_head, NULL, &ts, 0);
txq               983 drivers/net/wireless/ath/ath9k/xmit.c ath_tx_form_aggr(struct ath_softc *sc, struct ath_txq *txq,
txq              1041 drivers/net/wireless/ath/ath9k/xmit.c 		ret = ath_tx_get_tid_subframe(sc, txq, tid, &bf);
txq              1056 drivers/net/wireless/ath/ath9k/xmit.c 		TX_STAT_INC(sc, txq->axq_qnum, a_aggr);
txq              1343 drivers/net/wireless/ath/ath9k/xmit.c 			     struct ath_txq *txq, int len)
txq              1354 drivers/net/wireless/ath/ath9k/xmit.c 	info.qcu = txq->axq_qnum;
txq              1374 drivers/net/wireless/ath/ath9k/xmit.c 			    txq == sc->tx.uapsdq)
txq              1435 drivers/net/wireless/ath/ath9k/xmit.c ath_tx_form_burst(struct ath_softc *sc, struct ath_txq *txq,
txq              1454 drivers/net/wireless/ath/ath9k/xmit.c 		ret = ath_tx_get_tid_subframe(sc, txq, tid, &bf);
txq              1468 drivers/net/wireless/ath/ath9k/xmit.c static int ath_tx_sched_aggr(struct ath_softc *sc, struct ath_txq *txq,
txq              1479 drivers/net/wireless/ath/ath9k/xmit.c 	ret = ath_tx_get_tid_subframe(sc, txq, tid, &bf);
txq              1485 drivers/net/wireless/ath/ath9k/xmit.c 	if ((aggr && txq->axq_ampdu_depth >= ATH_AGGR_MIN_QDEPTH) ||
txq              1486 drivers/net/wireless/ath/ath9k/xmit.c 	    (!aggr && txq->axq_depth >= ATH_NON_AGGR_MIN_QDEPTH)) {
txq              1493 drivers/net/wireless/ath/ath9k/xmit.c 		aggr_len = ath_tx_form_aggr(sc, txq, tid, &bf_q, bf);
txq              1495 drivers/net/wireless/ath/ath9k/xmit.c 		ath_tx_form_burst(sc, txq, tid, &bf_q, bf);
txq              1505 drivers/net/wireless/ath/ath9k/xmit.c 	ath_tx_fill_desc(sc, bf, txq, aggr_len);
txq              1506 drivers/net/wireless/ath/ath9k/xmit.c 	ath_tx_txqaddbuf(sc, txq, &bf_q, false);
txq              1515 drivers/net/wireless/ath/ath9k/xmit.c 	struct ath_txq *txq;
txq              1523 drivers/net/wireless/ath/ath9k/xmit.c 	txq = txtid->txq;
txq              1525 drivers/net/wireless/ath/ath9k/xmit.c 	ath_txq_lock(sc, txq);
txq              1545 drivers/net/wireless/ath/ath9k/xmit.c 	ath_txq_unlock_complete(sc, txq);
txq              1555 drivers/net/wireless/ath/ath9k/xmit.c 	struct ath_txq *txq = txtid->txq;
txq              1559 drivers/net/wireless/ath/ath9k/xmit.c 	ath_txq_lock(sc, txq);
txq              1562 drivers/net/wireless/ath/ath9k/xmit.c 	ath_txq_unlock_complete(sc, txq);
txq              1587 drivers/net/wireless/ath/ath9k/xmit.c 	struct ath_txq *txq;
txq              1594 drivers/net/wireless/ath/ath9k/xmit.c 		txq = tid->txq;
txq              1596 drivers/net/wireless/ath/ath9k/xmit.c 		ath_txq_lock(sc, txq);
txq              1600 drivers/net/wireless/ath/ath9k/xmit.c 			ath_txq_schedule(sc, txq);
txq              1602 drivers/net/wireless/ath/ath9k/xmit.c 		ath_txq_unlock_complete(sc, txq);
txq              1631 drivers/net/wireless/ath/ath9k/xmit.c 	struct ath_txq *txq = sc->tx.uapsdq;
txq              1647 drivers/net/wireless/ath/ath9k/xmit.c 		ath_txq_lock(sc, tid->txq);
txq              1665 drivers/net/wireless/ath/ath9k/xmit.c 			TX_STAT_INC(sc, txq->axq_qnum, a_queued_hw);
txq              1670 drivers/net/wireless/ath/ath9k/xmit.c 		ath_txq_unlock_complete(sc, tid->txq);
txq              1683 drivers/net/wireless/ath/ath9k/xmit.c 	ath_txq_lock(sc, txq);
txq              1684 drivers/net/wireless/ath/ath9k/xmit.c 	ath_tx_fill_desc(sc, bf, txq, 0);
txq              1685 drivers/net/wireless/ath/ath9k/xmit.c 	ath_tx_txqaddbuf(sc, txq, &bf_q, false);
txq              1686 drivers/net/wireless/ath/ath9k/xmit.c 	ath_txq_unlock(sc, txq);
txq              1745 drivers/net/wireless/ath/ath9k/xmit.c 		struct ath_txq *txq = &sc->tx.txq[axq_qnum];
txq              1747 drivers/net/wireless/ath/ath9k/xmit.c 		txq->axq_qnum = axq_qnum;
txq              1748 drivers/net/wireless/ath/ath9k/xmit.c 		txq->mac80211_qnum = -1;
txq              1749 drivers/net/wireless/ath/ath9k/xmit.c 		txq->axq_link = NULL;
txq              1750 drivers/net/wireless/ath/ath9k/xmit.c 		__skb_queue_head_init(&txq->complete_q);
txq              1751 drivers/net/wireless/ath/ath9k/xmit.c 		INIT_LIST_HEAD(&txq->axq_q);
txq              1752 drivers/net/wireless/ath/ath9k/xmit.c 		spin_lock_init(&txq->axq_lock);
txq              1753 drivers/net/wireless/ath/ath9k/xmit.c 		txq->axq_depth = 0;
txq              1754 drivers/net/wireless/ath/ath9k/xmit.c 		txq->axq_ampdu_depth = 0;
txq              1755 drivers/net/wireless/ath/ath9k/xmit.c 		txq->axq_tx_inprogress = false;
txq              1758 drivers/net/wireless/ath/ath9k/xmit.c 		txq->txq_headidx = txq->txq_tailidx = 0;
txq              1760 drivers/net/wireless/ath/ath9k/xmit.c 			INIT_LIST_HEAD(&txq->txq_fifo[i]);
txq              1762 drivers/net/wireless/ath/ath9k/xmit.c 	return &sc->tx.txq[axq_qnum];
txq              1772 drivers/net/wireless/ath/ath9k/xmit.c 	BUG_ON(sc->tx.txq[qnum].axq_qnum != qnum);
txq              1807 drivers/net/wireless/ath/ath9k/xmit.c static void ath_drain_txq_list(struct ath_softc *sc, struct ath_txq *txq,
txq              1830 drivers/net/wireless/ath/ath9k/xmit.c 		ath_tx_process_buffer(sc, txq, &ts, bf, &bf_head);
txq              1840 drivers/net/wireless/ath/ath9k/xmit.c void ath_draintxq(struct ath_softc *sc, struct ath_txq *txq)
txq              1843 drivers/net/wireless/ath/ath9k/xmit.c 	ath_txq_lock(sc, txq);
txq              1846 drivers/net/wireless/ath/ath9k/xmit.c 		int idx = txq->txq_tailidx;
txq              1848 drivers/net/wireless/ath/ath9k/xmit.c 		while (!list_empty(&txq->txq_fifo[idx])) {
txq              1849 drivers/net/wireless/ath/ath9k/xmit.c 			ath_drain_txq_list(sc, txq, &txq->txq_fifo[idx]);
txq              1853 drivers/net/wireless/ath/ath9k/xmit.c 		txq->txq_tailidx = idx;
txq              1856 drivers/net/wireless/ath/ath9k/xmit.c 	txq->axq_link = NULL;
txq              1857 drivers/net/wireless/ath/ath9k/xmit.c 	txq->axq_tx_inprogress = false;
txq              1858 drivers/net/wireless/ath/ath9k/xmit.c 	ath_drain_txq_list(sc, txq, &txq->axq_q);
txq              1860 drivers/net/wireless/ath/ath9k/xmit.c 	ath_txq_unlock_complete(sc, txq);
txq              1868 drivers/net/wireless/ath/ath9k/xmit.c 	struct ath_txq *txq;
txq              1882 drivers/net/wireless/ath/ath9k/xmit.c 		if (!sc->tx.txq[i].axq_depth)
txq              1885 drivers/net/wireless/ath/ath9k/xmit.c 		if (ath9k_hw_numtxpending(ah, sc->tx.txq[i].axq_qnum))
txq              1899 drivers/net/wireless/ath/ath9k/xmit.c 		txq = &sc->tx.txq[i];
txq              1900 drivers/net/wireless/ath/ath9k/xmit.c 		ath_draintxq(sc, txq);
txq              1906 drivers/net/wireless/ath/ath9k/xmit.c void ath_tx_cleanupq(struct ath_softc *sc, struct ath_txq *txq)
txq              1908 drivers/net/wireless/ath/ath9k/xmit.c 	ath9k_hw_releasetxqueue(sc->sc_ah, txq->axq_qnum);
txq              1909 drivers/net/wireless/ath/ath9k/xmit.c 	sc->tx.txqsetup &= ~(1<<txq->axq_qnum);
txq              1915 drivers/net/wireless/ath/ath9k/xmit.c void ath_txq_schedule(struct ath_softc *sc, struct ath_txq *txq)
txq              1923 drivers/net/wireless/ath/ath9k/xmit.c 	if (txq->mac80211_qnum < 0)
txq              1929 drivers/net/wireless/ath/ath9k/xmit.c 	ieee80211_txq_schedule_start(hw, txq->mac80211_qnum);
txq              1936 drivers/net/wireless/ath/ath9k/xmit.c 	while ((queue = ieee80211_next_txq(hw, txq->mac80211_qnum))) {
txq              1941 drivers/net/wireless/ath/ath9k/xmit.c 		ret = ath_tx_sched_aggr(sc, txq, tid);
txq              1951 drivers/net/wireless/ath/ath9k/xmit.c 	ieee80211_txq_schedule_end(hw, txq->mac80211_qnum);
txq              1956 drivers/net/wireless/ath/ath9k/xmit.c 	struct ath_txq *txq;
txq              1960 drivers/net/wireless/ath/ath9k/xmit.c 		txq = sc->tx.txq_map[i];
txq              1962 drivers/net/wireless/ath/ath9k/xmit.c 		spin_lock_bh(&txq->axq_lock);
txq              1963 drivers/net/wireless/ath/ath9k/xmit.c 		ath_txq_schedule(sc, txq);
txq              1964 drivers/net/wireless/ath/ath9k/xmit.c 		spin_unlock_bh(&txq->axq_lock);
txq              1976 drivers/net/wireless/ath/ath9k/xmit.c static void ath_tx_txqaddbuf(struct ath_softc *sc, struct ath_txq *txq,
txq              1998 drivers/net/wireless/ath/ath9k/xmit.c 		txq->axq_qnum, txq->axq_depth);
txq              2000 drivers/net/wireless/ath/ath9k/xmit.c 	if (edma && list_empty(&txq->txq_fifo[txq->txq_headidx])) {
txq              2001 drivers/net/wireless/ath/ath9k/xmit.c 		list_splice_tail_init(head, &txq->txq_fifo[txq->txq_headidx]);
txq              2002 drivers/net/wireless/ath/ath9k/xmit.c 		INCR(txq->txq_headidx, ATH_TXFIFO_DEPTH);
txq              2005 drivers/net/wireless/ath/ath9k/xmit.c 		list_splice_tail_init(head, &txq->axq_q);
txq              2007 drivers/net/wireless/ath/ath9k/xmit.c 		if (txq->axq_link) {
txq              2008 drivers/net/wireless/ath/ath9k/xmit.c 			ath9k_hw_set_desc_link(ah, txq->axq_link, bf->bf_daddr);
txq              2010 drivers/net/wireless/ath/ath9k/xmit.c 				txq->axq_qnum, txq->axq_link,
txq              2015 drivers/net/wireless/ath/ath9k/xmit.c 		txq->axq_link = bf_last->bf_desc;
txq              2019 drivers/net/wireless/ath/ath9k/xmit.c 		TX_STAT_INC(sc, txq->axq_qnum, puttxbuf);
txq              2020 drivers/net/wireless/ath/ath9k/xmit.c 		ath9k_hw_puttxbuf(ah, txq->axq_qnum, bf->bf_daddr);
txq              2022 drivers/net/wireless/ath/ath9k/xmit.c 			txq->axq_qnum, ito64(bf->bf_daddr), bf->bf_desc);
txq              2026 drivers/net/wireless/ath/ath9k/xmit.c 		TX_STAT_INC(sc, txq->axq_qnum, txstart);
txq              2027 drivers/net/wireless/ath/ath9k/xmit.c 		ath9k_hw_txstart(ah, txq->axq_qnum);
txq              2032 drivers/net/wireless/ath/ath9k/xmit.c 			txq->axq_depth++;
txq              2034 drivers/net/wireless/ath/ath9k/xmit.c 				txq->axq_ampdu_depth++;
txq              2043 drivers/net/wireless/ath/ath9k/xmit.c static void ath_tx_send_normal(struct ath_softc *sc, struct ath_txq *txq,
txq              2061 drivers/net/wireless/ath/ath9k/xmit.c 	ath_tx_fill_desc(sc, bf, txq, fi->framelen);
txq              2062 drivers/net/wireless/ath/ath9k/xmit.c 	ath_tx_txqaddbuf(sc, txq, &bf_head, false);
txq              2063 drivers/net/wireless/ath/ath9k/xmit.c 	TX_STAT_INC(sc, txq->axq_qnum, queued);
txq              2107 drivers/net/wireless/ath/ath9k/xmit.c 	fi->txq = -1;
txq              2145 drivers/net/wireless/ath/ath9k/xmit.c 					   struct ath_txq *txq,
txq              2273 drivers/net/wireless/ath/ath9k/xmit.c 	struct ath_txq *txq = txctl->txq;
txq              2294 drivers/net/wireless/ath/ath9k/xmit.c 		txq = sc->tx.uapsdq;
txq              2301 drivers/net/wireless/ath/ath9k/xmit.c 	ath_txq_lock(sc, txq);
txq              2302 drivers/net/wireless/ath/ath9k/xmit.c 	if (txq == sc->tx.txq_map[q]) {
txq              2303 drivers/net/wireless/ath/ath9k/xmit.c 		fi->txq = q;
txq              2304 drivers/net/wireless/ath/ath9k/xmit.c 		++txq->pending_frames;
txq              2307 drivers/net/wireless/ath/ath9k/xmit.c 	bf = ath_tx_setup_buffer(sc, txq, tid, skb);
txq              2309 drivers/net/wireless/ath/ath9k/xmit.c 		ath_txq_skb_done(sc, txq, skb);
txq              2323 drivers/net/wireless/ath/ath9k/xmit.c 	ath_tx_send_normal(sc, txq, tid, skb);
txq              2326 drivers/net/wireless/ath/ath9k/xmit.c 	ath_txq_unlock(sc, txq);
txq              2336 drivers/net/wireless/ath/ath9k/xmit.c 		.txq = sc->beacon.cabq
txq              2355 drivers/net/wireless/ath/ath9k/xmit.c 		bf = ath_tx_setup_buffer(sc, txctl.txq, NULL, skb);
txq              2386 drivers/net/wireless/ath/ath9k/xmit.c 	ath_txq_lock(sc, txctl.txq);
txq              2387 drivers/net/wireless/ath/ath9k/xmit.c 	ath_tx_fill_desc(sc, bf, txctl.txq, 0);
txq              2388 drivers/net/wireless/ath/ath9k/xmit.c 	ath_tx_txqaddbuf(sc, txctl.txq, &bf_q, false);
txq              2389 drivers/net/wireless/ath/ath9k/xmit.c 	TX_STAT_INC(sc, txctl.txq->axq_qnum, queued);
txq              2390 drivers/net/wireless/ath/ath9k/xmit.c 	ath_txq_unlock(sc, txctl.txq);
txq              2398 drivers/net/wireless/ath/ath9k/xmit.c 			    int tx_flags, struct ath_txq *txq,
txq              2433 drivers/net/wireless/ath/ath9k/xmit.c 	if ((sc->ps_flags & PS_WAIT_FOR_TX_ACK) && !txq->axq_depth) {
txq              2444 drivers/net/wireless/ath/ath9k/xmit.c 	ath_txq_skb_done(sc, txq, skb);
txq              2446 drivers/net/wireless/ath/ath9k/xmit.c 	__skb_queue_tail(&txq->complete_q, skb);
txq              2450 drivers/net/wireless/ath/ath9k/xmit.c 				struct ath_txq *txq, struct list_head *bf_q,
txq              2478 drivers/net/wireless/ath/ath9k/xmit.c 		ath_debug_stat_tx(sc, bf, ts, txq, tx_flags);
txq              2479 drivers/net/wireless/ath/ath9k/xmit.c 		ath_tx_complete(sc, skb, tx_flags, txq, sta);
txq              2553 drivers/net/wireless/ath/ath9k/xmit.c static void ath_tx_processq(struct ath_softc *sc, struct ath_txq *txq)
txq              2564 drivers/net/wireless/ath/ath9k/xmit.c 		txq->axq_qnum, ath9k_hw_gettxbuf(sc->sc_ah, txq->axq_qnum),
txq              2565 drivers/net/wireless/ath/ath9k/xmit.c 		txq->axq_link);
txq              2567 drivers/net/wireless/ath/ath9k/xmit.c 	ath_txq_lock(sc, txq);
txq              2572 drivers/net/wireless/ath/ath9k/xmit.c 		if (list_empty(&txq->axq_q)) {
txq              2573 drivers/net/wireless/ath/ath9k/xmit.c 			txq->axq_link = NULL;
txq              2574 drivers/net/wireless/ath/ath9k/xmit.c 			ath_txq_schedule(sc, txq);
txq              2577 drivers/net/wireless/ath/ath9k/xmit.c 		bf = list_first_entry(&txq->axq_q, struct ath_buf, list);
txq              2590 drivers/net/wireless/ath/ath9k/xmit.c 			if (list_is_last(&bf_held->list, &txq->axq_q))
txq              2605 drivers/net/wireless/ath/ath9k/xmit.c 		TX_STAT_INC(sc, txq->axq_qnum, txprocdesc);
txq              2616 drivers/net/wireless/ath/ath9k/xmit.c 				&txq->axq_q, lastbf->list.prev);
txq              2623 drivers/net/wireless/ath/ath9k/xmit.c 		ath_tx_process_buffer(sc, txq, &ts, bf, &bf_head);
txq              2625 drivers/net/wireless/ath/ath9k/xmit.c 	ath_txq_unlock_complete(sc, txq);
txq              2637 drivers/net/wireless/ath/ath9k/xmit.c 			ath_tx_processq(sc, &sc->tx.txq[i]);
txq              2647 drivers/net/wireless/ath/ath9k/xmit.c 	struct ath_txq *txq;
txq              2680 drivers/net/wireless/ath/ath9k/xmit.c 		txq = &sc->tx.txq[ts.qid];
txq              2682 drivers/net/wireless/ath/ath9k/xmit.c 		ath_txq_lock(sc, txq);
txq              2684 drivers/net/wireless/ath/ath9k/xmit.c 		TX_STAT_INC(sc, txq->axq_qnum, txprocdesc);
txq              2686 drivers/net/wireless/ath/ath9k/xmit.c 		fifo_list = &txq->txq_fifo[txq->txq_tailidx];
txq              2688 drivers/net/wireless/ath/ath9k/xmit.c 			ath_txq_unlock(sc, txq);
txq              2704 drivers/net/wireless/ath/ath9k/xmit.c 			INCR(txq->txq_tailidx, ATH_TXFIFO_DEPTH);
txq              2706 drivers/net/wireless/ath/ath9k/xmit.c 			if (!list_empty(&txq->axq_q)) {
txq              2710 drivers/net/wireless/ath/ath9k/xmit.c 				txq->axq_link = NULL;
txq              2711 drivers/net/wireless/ath/ath9k/xmit.c 				list_splice_tail_init(&txq->axq_q, &bf_q);
txq              2712 drivers/net/wireless/ath/ath9k/xmit.c 				ath_tx_txqaddbuf(sc, txq, &bf_q, true);
txq              2721 drivers/net/wireless/ath/ath9k/xmit.c 		ath_tx_process_buffer(sc, txq, &ts, bf, &bf_head);
txq              2722 drivers/net/wireless/ath/ath9k/xmit.c 		ath_txq_unlock_complete(sc, txq);
txq              2804 drivers/net/wireless/ath/ath9k/xmit.c 		tid->txq = sc->tx.txq_map[acno];
txq              2814 drivers/net/wireless/ath/ath9k/xmit.c 	struct ath_txq *txq;
txq              2821 drivers/net/wireless/ath/ath9k/xmit.c 		txq = tid->txq;
txq              2823 drivers/net/wireless/ath/ath9k/xmit.c 		ath_txq_lock(sc, txq);
txq              2828 drivers/net/wireless/ath/ath9k/xmit.c 		ath_tid_drain(sc, txq, tid);
txq              2831 drivers/net/wireless/ath/ath9k/xmit.c 		ath_txq_unlock(sc, txq);
txq              2869 drivers/net/wireless/ath/ath9k/xmit.c 	bf = ath_tx_setup_buffer(sc, txctl->txq, NULL, skb);
txq              2878 drivers/net/wireless/ath/ath9k/xmit.c 	ath9k_hw_tx99_start(sc->sc_ah, txctl->txq->axq_qnum);
txq              2880 drivers/net/wireless/ath/ath9k/xmit.c 	ath_tx_send_normal(sc, txctl->txq, NULL, skb);
txq              1495 drivers/net/wireless/ath/wil6210/debugfs.c 		struct netdev_queue *txq = netdev_get_tx_queue(ndev, i);
txq              1496 drivers/net/wireless/ath/wil6210/debugfs.c 		unsigned long state = txq->state;
txq               829 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	struct pktq *txq;
txq               834 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	txq = brcmf_bus_gettxq(fws->drvr->bus_if);
txq               835 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	if (IS_ERR(txq)) {
txq               840 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 	for (prec = 0; prec < txq->num_prec; prec++) {
txq               841 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 		skb = brcmu_pktq_pdeq_match(txq, prec, fn, &ifidx);
txq               848 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c 			skb = brcmu_pktq_pdeq_match(txq, prec, fn, &ifidx);
txq               444 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c 	struct pktq txq;	/* Queue length used for flow-control */
txq              2315 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c 				brcmu_pktq_mlen(&bus->txq, ~bus->flowcontrol));
txq              2319 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c 			pkt = brcmu_pktq_mdeq(&bus->txq, tx_prec_map,
txq              2351 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c 	    bus->txoff && (pktq_len(&bus->txq) < TXLOW)) {
txq              2486 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c 	brcmu_pktq_flush(&bus->txq, true, NULL, NULL);
txq              2677 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c 	    brcmu_pktq_mlen(&bus->txq, ~bus->flowcontrol) && txlimit &&
txq              2700 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c 		    brcmu_pktq_mlen(&bus->txq, ~bus->flowcontrol) &&
txq              2712 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c 	return &bus->txq;
txq              2777 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c 	brcmf_dbg(TRACE, "deferring pktq len %d\n", pktq_len(&bus->txq));
txq              2784 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c 	if (!brcmf_sdio_prec_enq(&bus->txq, pkt, prec)) {
txq              2792 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c 	if (pktq_len(&bus->txq) >= TXHI) {
txq              2799 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c 	if (pktq_plen(&bus->txq, prec) > qcount[prec])
txq              2800 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c 		qcount[prec] = pktq_plen(&bus->txq, prec);
txq              4014 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c 	brcmu_pktq_init(&bus->txq, (PRIOMASK + 1), TXQLEN);
txq              1229 drivers/net/wireless/cisco/airo.c 	struct sk_buff_head txq;// tx queue used by mpi350 code
txq              1932 drivers/net/wireless/cisco/airo.c 	npacks = skb_queue_len (&ai->txq);
txq              1940 drivers/net/wireless/cisco/airo.c 		skb_queue_tail (&ai->txq, skb);
txq              1945 drivers/net/wireless/cisco/airo.c 	skb_queue_tail (&ai->txq, skb);
txq              1975 drivers/net/wireless/cisco/airo.c 	if ((skb = skb_dequeue(&ai->txq)) == NULL) {
txq              2402 drivers/net/wireless/cisco/airo.c 	if (test_bit(FLAG_MPI, &ai->flags) && !skb_queue_empty(&ai->txq)) {
txq              2404 drivers/net/wireless/cisco/airo.c 		for (;(skb = skb_dequeue(&ai->txq));)
txq              2812 drivers/net/wireless/cisco/airo.c 	skb_queue_head_init (&ai->txq);
txq              3441 drivers/net/wireless/cisco/airo.c 		if (!skb_queue_empty(&ai->txq)) {
txq              2808 drivers/net/wireless/intel/ipw2x00/ipw2100.c 	struct ipw2100_bd_queue *txq = &priv->tx_queue;
txq              2822 drivers/net/wireless/intel/ipw2x00/ipw2100.c 	tbd = &txq->drv[packet->index];
txq              2829 drivers/net/wireless/intel/ipw2x00/ipw2100.c 		e = txq->oldest;
txq              2836 drivers/net/wireless/intel/ipw2x00/ipw2100.c 		e = txq->oldest + frag_num;
txq              2837 drivers/net/wireless/intel/ipw2x00/ipw2100.c 		e %= txq->entries;
txq              2854 drivers/net/wireless/intel/ipw2x00/ipw2100.c 	if (w != txq->next)
txq              2890 drivers/net/wireless/intel/ipw2x00/ipw2100.c 		i = txq->oldest;
txq              2892 drivers/net/wireless/intel/ipw2x00/ipw2100.c 			     &txq->drv[i],
txq              2893 drivers/net/wireless/intel/ipw2x00/ipw2100.c 			     (u32) (txq->nic + i * sizeof(struct ipw2100_bd)),
txq              2894 drivers/net/wireless/intel/ipw2x00/ipw2100.c 			     txq->drv[i].host_addr, txq->drv[i].buf_length);
txq              2897 drivers/net/wireless/intel/ipw2x00/ipw2100.c 			i = (i + 1) % txq->entries;
txq              2900 drivers/net/wireless/intel/ipw2x00/ipw2100.c 				     &txq->drv[i],
txq              2901 drivers/net/wireless/intel/ipw2x00/ipw2100.c 				     (u32) (txq->nic + i *
txq              2903 drivers/net/wireless/intel/ipw2x00/ipw2100.c 				     (u32) txq->drv[i].host_addr,
txq              2904 drivers/net/wireless/intel/ipw2x00/ipw2100.c 				     txq->drv[i].buf_length);
txq              2911 drivers/net/wireless/intel/ipw2x00/ipw2100.c 		if (txq->drv[txq->oldest].status.info.fields.txType != 0)
txq              2915 drivers/net/wireless/intel/ipw2x00/ipw2100.c 			       priv->net_dev->name, txq->oldest, packet->index);
txq              2919 drivers/net/wireless/intel/ipw2x00/ipw2100.c 			tbd = &txq->drv[(packet->index + 1 + i) % txq->entries];
txq              2922 drivers/net/wireless/intel/ipw2x00/ipw2100.c 				     (packet->index + 1 + i) % txq->entries,
txq              2948 drivers/net/wireless/intel/ipw2x00/ipw2100.c 		if (txq->drv[txq->oldest].status.info.fields.txType != 1)
txq              2952 drivers/net/wireless/intel/ipw2x00/ipw2100.c 			       priv->net_dev->name, txq->oldest, packet->index);
txq              2971 drivers/net/wireless/intel/ipw2x00/ipw2100.c 	txq->oldest = (e + 1) % txq->entries;
txq              2973 drivers/net/wireless/intel/ipw2x00/ipw2100.c 	txq->available += descriptors_used;
txq              2974 drivers/net/wireless/intel/ipw2x00/ipw2100.c 	SET_STAT(&priv->txq_stat, txq->available);
txq              3000 drivers/net/wireless/intel/ipw2x00/ipw2100.c 	struct ipw2100_bd_queue *txq = &priv->tx_queue;
txq              3002 drivers/net/wireless/intel/ipw2x00/ipw2100.c 	int next = txq->next;
txq              3011 drivers/net/wireless/intel/ipw2x00/ipw2100.c 		if (txq->available <= 3) {
txq              3023 drivers/net/wireless/intel/ipw2x00/ipw2100.c 			     &txq->drv[txq->next],
txq              3024 drivers/net/wireless/intel/ipw2x00/ipw2100.c 			     (u32) (txq->nic + txq->next *
txq              3027 drivers/net/wireless/intel/ipw2x00/ipw2100.c 		packet->index = txq->next;
txq              3029 drivers/net/wireless/intel/ipw2x00/ipw2100.c 		tbd = &txq->drv[txq->next];
txq              3042 drivers/net/wireless/intel/ipw2x00/ipw2100.c 		txq->next++;
txq              3043 drivers/net/wireless/intel/ipw2x00/ipw2100.c 		txq->next %= txq->entries;
txq              3044 drivers/net/wireless/intel/ipw2x00/ipw2100.c 		txq->available--;
txq              3051 drivers/net/wireless/intel/ipw2x00/ipw2100.c 	if (txq->next != next) {
txq              3057 drivers/net/wireless/intel/ipw2x00/ipw2100.c 			       txq->next);
txq              3069 drivers/net/wireless/intel/ipw2x00/ipw2100.c 	struct ipw2100_bd_queue *txq = &priv->tx_queue;
txq              3071 drivers/net/wireless/intel/ipw2x00/ipw2100.c 	int next = txq->next;
txq              3095 drivers/net/wireless/intel/ipw2x00/ipw2100.c 		if (txq->available <= 3 + packet->info.d_struct.txb->nr_frags) {
txq              3103 drivers/net/wireless/intel/ipw2x00/ipw2100.c 		tbd = &txq->drv[txq->next];
txq              3105 drivers/net/wireless/intel/ipw2x00/ipw2100.c 		packet->index = txq->next;
txq              3142 drivers/net/wireless/intel/ipw2x00/ipw2100.c 		txq->next++;
txq              3143 drivers/net/wireless/intel/ipw2x00/ipw2100.c 		txq->next %= txq->entries;
txq              3154 drivers/net/wireless/intel/ipw2x00/ipw2100.c 			tbd = &txq->drv[txq->next];
txq              3181 drivers/net/wireless/intel/ipw2x00/ipw2100.c 				     txq->next, tbd->host_addr,
txq              3189 drivers/net/wireless/intel/ipw2x00/ipw2100.c 			txq->next++;
txq              3190 drivers/net/wireless/intel/ipw2x00/ipw2100.c 			txq->next %= txq->entries;
txq              3193 drivers/net/wireless/intel/ipw2x00/ipw2100.c 		txq->available -= 1 + packet->info.d_struct.txb->nr_frags;
txq              3194 drivers/net/wireless/intel/ipw2x00/ipw2100.c 		SET_STAT(&priv->txq_stat, txq->available);
txq              3200 drivers/net/wireless/intel/ipw2x00/ipw2100.c 	if (txq->next != next) {
txq              3205 drivers/net/wireless/intel/ipw2x00/ipw2100.c 			       txq->next);
txq               199 drivers/net/wireless/intel/ipw2x00/ipw2200.c 				struct clx2_tx_queue *txq, int qindex);
txq              1991 drivers/net/wireless/intel/ipw2x00/ipw2200.c 		rc = ipw_queue_tx_reclaim(priv, &priv->txq[0], 0);
txq              1997 drivers/net/wireless/intel/ipw2x00/ipw2200.c 		rc = ipw_queue_tx_reclaim(priv, &priv->txq[1], 1);
txq              2003 drivers/net/wireless/intel/ipw2x00/ipw2200.c 		rc = ipw_queue_tx_reclaim(priv, &priv->txq[2], 2);
txq              2009 drivers/net/wireless/intel/ipw2x00/ipw2200.c 		rc = ipw_queue_tx_reclaim(priv, &priv->txq[3], 3);
txq              3800 drivers/net/wireless/intel/ipw2x00/ipw2200.c 				  struct clx2_tx_queue *txq)
txq              3802 drivers/net/wireless/intel/ipw2x00/ipw2200.c 	struct tfd_frame *bd = &txq->bd[txq->q.last_used];
txq              3824 drivers/net/wireless/intel/ipw2x00/ipw2200.c 		if (txq->txb[txq->q.last_used]) {
txq              3825 drivers/net/wireless/intel/ipw2x00/ipw2200.c 			libipw_txb_free(txq->txb[txq->q.last_used]);
txq              3826 drivers/net/wireless/intel/ipw2x00/ipw2200.c 			txq->txb[txq->q.last_used] = NULL;
txq              3840 drivers/net/wireless/intel/ipw2x00/ipw2200.c static void ipw_queue_tx_free(struct ipw_priv *priv, struct clx2_tx_queue *txq)
txq              3842 drivers/net/wireless/intel/ipw2x00/ipw2200.c 	struct clx2_queue *q = &txq->q;
txq              3851 drivers/net/wireless/intel/ipw2x00/ipw2200.c 		ipw_queue_tx_free_tfd(priv, txq);
txq              3855 drivers/net/wireless/intel/ipw2x00/ipw2200.c 	pci_free_consistent(dev, sizeof(txq->bd[0]) * q->n_bd, txq->bd,
txq              3857 drivers/net/wireless/intel/ipw2x00/ipw2200.c 	kfree(txq->txb);
txq              3860 drivers/net/wireless/intel/ipw2x00/ipw2200.c 	memset(txq, 0, sizeof(*txq));
txq              3874 drivers/net/wireless/intel/ipw2x00/ipw2200.c 	ipw_queue_tx_free(priv, &priv->txq[0]);
txq              3875 drivers/net/wireless/intel/ipw2x00/ipw2200.c 	ipw_queue_tx_free(priv, &priv->txq[1]);
txq              3876 drivers/net/wireless/intel/ipw2x00/ipw2200.c 	ipw_queue_tx_free(priv, &priv->txq[2]);
txq              3877 drivers/net/wireless/intel/ipw2x00/ipw2200.c 	ipw_queue_tx_free(priv, &priv->txq[3]);
txq              4951 drivers/net/wireless/intel/ipw2x00/ipw2200.c 	rc = ipw_queue_tx_init(priv, &priv->txq[0], nTx,
txq              4959 drivers/net/wireless/intel/ipw2x00/ipw2200.c 	rc = ipw_queue_tx_init(priv, &priv->txq[1], nTx,
txq              4967 drivers/net/wireless/intel/ipw2x00/ipw2200.c 	rc = ipw_queue_tx_init(priv, &priv->txq[2], nTx,
txq              4975 drivers/net/wireless/intel/ipw2x00/ipw2200.c 	rc = ipw_queue_tx_init(priv, &priv->txq[3], nTx,
txq              5007 drivers/net/wireless/intel/ipw2x00/ipw2200.c 				struct clx2_tx_queue *txq, int qindex)
txq              5011 drivers/net/wireless/intel/ipw2x00/ipw2200.c 	struct clx2_queue *q = &txq->q;
txq              5022 drivers/net/wireless/intel/ipw2x00/ipw2200.c 		ipw_queue_tx_free_tfd(priv, txq);
txq              5039 drivers/net/wireless/intel/ipw2x00/ipw2200.c 	struct clx2_tx_queue *txq = &priv->txq_cmd;
txq              5040 drivers/net/wireless/intel/ipw2x00/ipw2200.c 	struct clx2_queue *q = &txq->q;
txq              5048 drivers/net/wireless/intel/ipw2x00/ipw2200.c 	tfd = &txq->bd[q->first_empty];
txq              5049 drivers/net/wireless/intel/ipw2x00/ipw2200.c 	txq->txb[q->first_empty] = NULL;
txq              10101 drivers/net/wireless/intel/ipw2x00/ipw2200.c 	struct clx2_tx_queue *txq = &priv->txq[tx_id];
txq              10103 drivers/net/wireless/intel/ipw2x00/ipw2200.c 	struct clx2_tx_queue *txq = &priv->txq[0];
txq              10105 drivers/net/wireless/intel/ipw2x00/ipw2200.c 	struct clx2_queue *q = &txq->q;
txq              10135 drivers/net/wireless/intel/ipw2x00/ipw2200.c 	tfd = &txq->bd[q->first_empty];
txq              10136 drivers/net/wireless/intel/ipw2x00/ipw2200.c 	txq->txb[q->first_empty] = txb;
txq              10291 drivers/net/wireless/intel/ipw2x00/ipw2200.c 	struct clx2_tx_queue *txq = &priv->txq[tx_id];
txq              10293 drivers/net/wireless/intel/ipw2x00/ipw2200.c 	struct clx2_tx_queue *txq = &priv->txq[0];
txq              10296 drivers/net/wireless/intel/ipw2x00/ipw2200.c 	if (ipw_tx_queue_space(&txq->q) < txq->q.high_mark)
txq              1193 drivers/net/wireless/intel/ipw2x00/ipw2200.h 	struct clx2_tx_queue txq[4];
txq               454 drivers/net/wireless/intel/iwlegacy/3945-mac.c 	struct il_tx_queue *txq = NULL;
txq               513 drivers/net/wireless/intel/iwlegacy/3945-mac.c 	txq = &il->txq[txq_id];
txq               514 drivers/net/wireless/intel/iwlegacy/3945-mac.c 	q = &txq->q;
txq               523 drivers/net/wireless/intel/iwlegacy/3945-mac.c 	txq->skbs[q->write_ptr] = skb;
txq               526 drivers/net/wireless/intel/iwlegacy/3945-mac.c 	out_cmd = txq->cmd[idx];
txq               527 drivers/net/wireless/intel/iwlegacy/3945-mac.c 	out_meta = &txq->meta[idx];
txq               595 drivers/net/wireless/intel/iwlegacy/3945-mac.c 	il->ops->txq_attach_buf_to_tfd(il, txq, txcmd_phys, firstlen, 1, 0);
txq               599 drivers/net/wireless/intel/iwlegacy/3945-mac.c 		il->ops->txq_attach_buf_to_tfd(il, txq, phys_addr, secondlen, 0,
txq               603 drivers/net/wireless/intel/iwlegacy/3945-mac.c 		txq->need_update = 1;
txq               606 drivers/net/wireless/intel/iwlegacy/3945-mac.c 		txq->need_update = 0;
txq               619 drivers/net/wireless/intel/iwlegacy/3945-mac.c 	il_txq_update_write_ptr(il, txq);
txq               625 drivers/net/wireless/intel/iwlegacy/3945-mac.c 			txq->need_update = 1;
txq               626 drivers/net/wireless/intel/iwlegacy/3945-mac.c 			il_txq_update_write_ptr(il, txq);
txq               630 drivers/net/wireless/intel/iwlegacy/3945-mac.c 		il_stop_queue(il, txq);
txq              1471 drivers/net/wireless/intel/iwlegacy/3945-mac.c 		il_txq_update_write_ptr(il, &il->txq[0]);
txq              1472 drivers/net/wireless/intel/iwlegacy/3945-mac.c 		il_txq_update_write_ptr(il, &il->txq[1]);
txq              1473 drivers/net/wireless/intel/iwlegacy/3945-mac.c 		il_txq_update_write_ptr(il, &il->txq[2]);
txq              1474 drivers/net/wireless/intel/iwlegacy/3945-mac.c 		il_txq_update_write_ptr(il, &il->txq[3]);
txq              1475 drivers/net/wireless/intel/iwlegacy/3945-mac.c 		il_txq_update_write_ptr(il, &il->txq[4]);
txq              2463 drivers/net/wireless/intel/iwlegacy/3945-mac.c 	if (test_bit(S_EXIT_PENDING, &il->status) || il->txq == NULL)
txq               274 drivers/net/wireless/intel/iwlegacy/3945.c 	struct il_tx_queue *txq = &il->txq[txq_id];
txq               275 drivers/net/wireless/intel/iwlegacy/3945.c 	struct il_queue *q = &txq->q;
txq               283 drivers/net/wireless/intel/iwlegacy/3945.c 		skb = txq->skbs[txq->q.read_ptr];
txq               285 drivers/net/wireless/intel/iwlegacy/3945.c 		txq->skbs[txq->q.read_ptr] = NULL;
txq               286 drivers/net/wireless/intel/iwlegacy/3945.c 		il->ops->txq_free_tfd(il, txq);
txq               291 drivers/net/wireless/intel/iwlegacy/3945.c 		il_wake_queue(il, txq);
txq               304 drivers/net/wireless/intel/iwlegacy/3945.c 	struct il_tx_queue *txq = &il->txq[txq_id];
txq               311 drivers/net/wireless/intel/iwlegacy/3945.c 	if (idx >= txq->q.n_bd || il_queue_used(&txq->q, idx) == 0) {
txq               314 drivers/net/wireless/intel/iwlegacy/3945.c 		       txq->q.n_bd, txq->q.write_ptr, txq->q.read_ptr);
txq               331 drivers/net/wireless/intel/iwlegacy/3945.c 	txq->time_stamp = jiffies;
txq               332 drivers/net/wireless/intel/iwlegacy/3945.c 	info = IEEE80211_SKB_CB(txq->skbs[txq->q.read_ptr]);
txq               597 drivers/net/wireless/intel/iwlegacy/3945.c il3945_hw_txq_attach_buf_to_tfd(struct il_priv *il, struct il_tx_queue *txq,
txq               604 drivers/net/wireless/intel/iwlegacy/3945.c 	q = &txq->q;
txq               605 drivers/net/wireless/intel/iwlegacy/3945.c 	tfd_tmp = (struct il3945_tfd *)txq->tfds;
txq               636 drivers/net/wireless/intel/iwlegacy/3945.c il3945_hw_txq_free_tfd(struct il_priv *il, struct il_tx_queue *txq)
txq               638 drivers/net/wireless/intel/iwlegacy/3945.c 	struct il3945_tfd *tfd_tmp = (struct il3945_tfd *)txq->tfds;
txq               639 drivers/net/wireless/intel/iwlegacy/3945.c 	int idx = txq->q.read_ptr;
txq               655 drivers/net/wireless/intel/iwlegacy/3945.c 		pci_unmap_single(dev, dma_unmap_addr(&txq->meta[idx], mapping),
txq               656 drivers/net/wireless/intel/iwlegacy/3945.c 				 dma_unmap_len(&txq->meta[idx], len),
txq               667 drivers/net/wireless/intel/iwlegacy/3945.c 	if (txq->skbs) {
txq               668 drivers/net/wireless/intel/iwlegacy/3945.c 		struct sk_buff *skb = txq->skbs[txq->q.read_ptr];
txq               673 drivers/net/wireless/intel/iwlegacy/3945.c 			txq->skbs[txq->q.read_ptr] = NULL;
txq              1007 drivers/net/wireless/intel/iwlegacy/3945.c 	if (il->txq) {
txq              1868 drivers/net/wireless/intel/iwlegacy/3945.c 	if (test_bit(S_EXIT_PENDING, &il->status) || il->txq == NULL)
txq              2197 drivers/net/wireless/intel/iwlegacy/3945.c il3945_hw_tx_queue_init(struct il_priv *il, struct il_tx_queue *txq)
txq              2199 drivers/net/wireless/intel/iwlegacy/3945.c 	int txq_id = txq->q.id;
txq              2203 drivers/net/wireless/intel/iwlegacy/3945.c 	shared_data->tx_base_ptr[txq_id] = cpu_to_le32((u32) txq->q.dma_addr);
txq               209 drivers/net/wireless/intel/iwlegacy/3945.h int il3945_hw_txq_attach_buf_to_tfd(struct il_priv *il, struct il_tx_queue *txq,
txq               211 drivers/net/wireless/intel/iwlegacy/3945.h void il3945_hw_txq_free_tfd(struct il_priv *il, struct il_tx_queue *txq);
txq               213 drivers/net/wireless/intel/iwlegacy/3945.h int il3945_hw_tx_queue_init(struct il_priv *il, struct il_tx_queue *txq);
txq               216 drivers/net/wireless/intel/iwlegacy/4965-mac.c 	if (!il->txq) {
txq              1648 drivers/net/wireless/intel/iwlegacy/4965-mac.c 	struct il_tx_queue *txq;
txq              1749 drivers/net/wireless/intel/iwlegacy/4965-mac.c 	txq = &il->txq[txq_id];
txq              1750 drivers/net/wireless/intel/iwlegacy/4965-mac.c 	q = &txq->q;
txq              1765 drivers/net/wireless/intel/iwlegacy/4965-mac.c 	txq->skbs[q->write_ptr] = skb;
txq              1768 drivers/net/wireless/intel/iwlegacy/4965-mac.c 	out_cmd = txq->cmd[q->write_ptr];
txq              1769 drivers/net/wireless/intel/iwlegacy/4965-mac.c 	out_meta = &txq->meta[q->write_ptr];
txq              1836 drivers/net/wireless/intel/iwlegacy/4965-mac.c 	il->ops->txq_attach_buf_to_tfd(il, txq, txcmd_phys, firstlen, 1, 0);
txq              1840 drivers/net/wireless/intel/iwlegacy/4965-mac.c 		il->ops->txq_attach_buf_to_tfd(il, txq, phys_addr, secondlen,
txq              1844 drivers/net/wireless/intel/iwlegacy/4965-mac.c 		txq->need_update = 1;
txq              1847 drivers/net/wireless/intel/iwlegacy/4965-mac.c 		txq->need_update = 0;
txq              1869 drivers/net/wireless/intel/iwlegacy/4965-mac.c 		il->ops->txq_update_byte_cnt_tbl(il, txq, le16_to_cpu(tx_cmd->len));
txq              1876 drivers/net/wireless/intel/iwlegacy/4965-mac.c 	il_txq_update_write_ptr(il, txq);
txq              1899 drivers/net/wireless/intel/iwlegacy/4965-mac.c 			txq->need_update = 1;
txq              1900 drivers/net/wireless/intel/iwlegacy/4965-mac.c 			il_txq_update_write_ptr(il, txq);
txq              1903 drivers/net/wireless/intel/iwlegacy/4965-mac.c 			il_stop_queue(il, txq);
txq              1946 drivers/net/wireless/intel/iwlegacy/4965-mac.c 	if (il->txq) {
txq              2051 drivers/net/wireless/intel/iwlegacy/4965-mac.c 	if (!il->txq)
txq              2187 drivers/net/wireless/intel/iwlegacy/4965-mac.c 	il->txq[txq_id].q.read_ptr = (ssn_idx & 0xff);
txq              2188 drivers/net/wireless/intel/iwlegacy/4965-mac.c 	il->txq[txq_id].q.write_ptr = (ssn_idx & 0xff);
txq              2208 drivers/net/wireless/intel/iwlegacy/4965-mac.c 	il4965_tx_queue_set_status(il, &il->txq[txq_id], tx_fifo, 1);
txq              2256 drivers/net/wireless/intel/iwlegacy/4965-mac.c 	il_set_swq_id(&il->txq[txq_id], il4965_get_ac_from_tid(tid), txq_id);
txq              2299 drivers/net/wireless/intel/iwlegacy/4965-mac.c 	il->txq[txq_id].q.read_ptr = (ssn_idx & 0xff);
txq              2300 drivers/net/wireless/intel/iwlegacy/4965-mac.c 	il->txq[txq_id].q.write_ptr = (ssn_idx & 0xff);
txq              2306 drivers/net/wireless/intel/iwlegacy/4965-mac.c 	il4965_tx_queue_set_status(il, &il->txq[txq_id], tx_fifo, 0);
txq              2354 drivers/net/wireless/intel/iwlegacy/4965-mac.c 	write_ptr = il->txq[txq_id].q.write_ptr;
txq              2355 drivers/net/wireless/intel/iwlegacy/4965-mac.c 	read_ptr = il->txq[txq_id].q.read_ptr;
txq              2392 drivers/net/wireless/intel/iwlegacy/4965-mac.c 	struct il_queue *q = &il->txq[txq_id].q;
txq              2457 drivers/net/wireless/intel/iwlegacy/4965-mac.c 	struct il_tx_queue *txq = &il->txq[txq_id];
txq              2458 drivers/net/wireless/intel/iwlegacy/4965-mac.c 	struct il_queue *q = &txq->q;
txq              2473 drivers/net/wireless/intel/iwlegacy/4965-mac.c 		skb = txq->skbs[txq->q.read_ptr];
txq              2484 drivers/net/wireless/intel/iwlegacy/4965-mac.c 		txq->skbs[txq->q.read_ptr] = NULL;
txq              2485 drivers/net/wireless/intel/iwlegacy/4965-mac.c 		il->ops->txq_free_tfd(il, txq);
txq              2548 drivers/net/wireless/intel/iwlegacy/4965-mac.c 	info = IEEE80211_SKB_CB(il->txq[scd_flow].skbs[agg->start_idx]);
txq              2675 drivers/net/wireless/intel/iwlegacy/4965-mac.c 		info = IEEE80211_SKB_CB(il->txq[txq_id].skbs[idx]);
txq              2708 drivers/net/wireless/intel/iwlegacy/4965-mac.c 			skb = il->txq[txq_id].skbs[idx];
txq              2765 drivers/net/wireless/intel/iwlegacy/4965-mac.c 	struct il_tx_queue *txq = &il->txq[txq_id];
txq              2777 drivers/net/wireless/intel/iwlegacy/4965-mac.c 	if (idx >= txq->q.n_bd || il_queue_used(&txq->q, idx) == 0) {
txq              2780 drivers/net/wireless/intel/iwlegacy/4965-mac.c 		       txq->q.n_bd, txq->q.write_ptr, txq->q.read_ptr);
txq              2784 drivers/net/wireless/intel/iwlegacy/4965-mac.c 	txq->time_stamp = jiffies;
txq              2786 drivers/net/wireless/intel/iwlegacy/4965-mac.c 	skb = txq->skbs[txq->q.read_ptr];
txq              2797 drivers/net/wireless/intel/iwlegacy/4965-mac.c 	if (txq->sched_retry && unlikely(sta_id == IL_INVALID_STATION)) {
txq              2816 drivers/net/wireless/intel/iwlegacy/4965-mac.c 	if (txq->sched_retry) {
txq              2830 drivers/net/wireless/intel/iwlegacy/4965-mac.c 		if (txq->q.read_ptr != (scd_ssn & 0xff)) {
txq              2831 drivers/net/wireless/intel/iwlegacy/4965-mac.c 			idx = il_queue_dec_wrap(scd_ssn & 0xff, txq->q.n_bd);
txq              2840 drivers/net/wireless/intel/iwlegacy/4965-mac.c 			    il_queue_space(&txq->q) > txq->q.low_mark &&
txq              2842 drivers/net/wireless/intel/iwlegacy/4965-mac.c 				il_wake_queue(il, txq);
txq              2864 drivers/net/wireless/intel/iwlegacy/4965-mac.c 		    il_queue_space(&txq->q) > txq->q.low_mark)
txq              2865 drivers/net/wireless/intel/iwlegacy/4965-mac.c 			il_wake_queue(il, txq);
txq              2910 drivers/net/wireless/intel/iwlegacy/4965-mac.c 	struct il_tx_queue *txq = NULL;
txq              2929 drivers/net/wireless/intel/iwlegacy/4965-mac.c 	txq = &il->txq[scd_flow];
txq              2946 drivers/net/wireless/intel/iwlegacy/4965-mac.c 	idx = il_queue_dec_wrap(ba_resp_scd_ssn & 0xff, txq->q.n_bd);
txq              2966 drivers/net/wireless/intel/iwlegacy/4965-mac.c 	if (txq->q.read_ptr != (ba_resp_scd_ssn & 0xff)) {
txq              2971 drivers/net/wireless/intel/iwlegacy/4965-mac.c 		if (il_queue_space(&txq->q) > txq->q.low_mark &&
txq              2974 drivers/net/wireless/intel/iwlegacy/4965-mac.c 			il_wake_queue(il, txq);
txq              3915 drivers/net/wireless/intel/iwlegacy/4965-mac.c il4965_hw_txq_free_tfd(struct il_priv *il, struct il_tx_queue *txq)
txq              3917 drivers/net/wireless/intel/iwlegacy/4965-mac.c 	struct il_tfd *tfd_tmp = (struct il_tfd *)txq->tfds;
txq              3920 drivers/net/wireless/intel/iwlegacy/4965-mac.c 	int idx = txq->q.read_ptr;
txq              3937 drivers/net/wireless/intel/iwlegacy/4965-mac.c 		pci_unmap_single(dev, dma_unmap_addr(&txq->meta[idx], mapping),
txq              3938 drivers/net/wireless/intel/iwlegacy/4965-mac.c 				 dma_unmap_len(&txq->meta[idx], len),
txq              3948 drivers/net/wireless/intel/iwlegacy/4965-mac.c 	if (txq->skbs) {
txq              3949 drivers/net/wireless/intel/iwlegacy/4965-mac.c 		struct sk_buff *skb = txq->skbs[txq->q.read_ptr];
txq              3954 drivers/net/wireless/intel/iwlegacy/4965-mac.c 			txq->skbs[txq->q.read_ptr] = NULL;
txq              3960 drivers/net/wireless/intel/iwlegacy/4965-mac.c il4965_hw_txq_attach_buf_to_tfd(struct il_priv *il, struct il_tx_queue *txq,
txq              3967 drivers/net/wireless/intel/iwlegacy/4965-mac.c 	q = &txq->q;
txq              3968 drivers/net/wireless/intel/iwlegacy/4965-mac.c 	tfd_tmp = (struct il_tfd *)txq->tfds;
txq              4000 drivers/net/wireless/intel/iwlegacy/4965-mac.c il4965_hw_tx_queue_init(struct il_priv *il, struct il_tx_queue *txq)
txq              4002 drivers/net/wireless/intel/iwlegacy/4965-mac.c 	int txq_id = txq->q.id;
txq              4005 drivers/net/wireless/intel/iwlegacy/4965-mac.c 	il_wr(il, FH49_MEM_CBBC_QUEUE(txq_id), txq->q.dma_addr >> 8);
txq              4478 drivers/net/wireless/intel/iwlegacy/4965-mac.c 			il_txq_update_write_ptr(il, &il->txq[i]);
txq              5254 drivers/net/wireless/intel/iwlegacy/4965-mac.c 		il4965_tx_queue_set_status(il, &il->txq[i], ac, 0);
txq              6291 drivers/net/wireless/intel/iwlegacy/4965-mac.c il4965_tx_queue_set_status(struct il_priv *il, struct il_tx_queue *txq,
txq              6294 drivers/net/wireless/intel/iwlegacy/4965-mac.c 	int txq_id = txq->q.id;
txq              6307 drivers/net/wireless/intel/iwlegacy/4965-mac.c 	txq->sched_retry = scd_retry;
txq              1535 drivers/net/wireless/intel/iwlegacy/4965.c il4965_txq_update_byte_cnt_tbl(struct il_priv *il, struct il_tx_queue *txq,
txq              1539 drivers/net/wireless/intel/iwlegacy/4965.c 	int txq_id = txq->q.id;
txq              1540 drivers/net/wireless/intel/iwlegacy/4965.c 	int write_ptr = txq->q.write_ptr;
txq                57 drivers/net/wireless/intel/iwlegacy/4965.h void il4965_hw_txq_free_tfd(struct il_priv *il, struct il_tx_queue *txq);
txq                58 drivers/net/wireless/intel/iwlegacy/4965.h int il4965_hw_txq_attach_buf_to_tfd(struct il_priv *il, struct il_tx_queue *txq,
txq                60 drivers/net/wireless/intel/iwlegacy/4965.h int il4965_hw_tx_queue_init(struct il_priv *il, struct il_tx_queue *txq);
txq                89 drivers/net/wireless/intel/iwlegacy/4965.h void il4965_tx_queue_set_status(struct il_priv *il, struct il_tx_queue *txq,
txq               364 drivers/net/wireless/intel/iwlegacy/common.c 		il->txq[il->cmd_queue].meta[cmd_idx].flags &= ~CMD_WANT_SKB;
txq              2710 drivers/net/wireless/intel/iwlegacy/common.c il_txq_update_write_ptr(struct il_priv *il, struct il_tx_queue *txq)
txq              2713 drivers/net/wireless/intel/iwlegacy/common.c 	int txq_id = txq->q.id;
txq              2715 drivers/net/wireless/intel/iwlegacy/common.c 	if (txq->need_update == 0)
txq              2733 drivers/net/wireless/intel/iwlegacy/common.c 		il_wr(il, HBUS_TARG_WRPTR, txq->q.write_ptr | (txq_id << 8));
txq              2741 drivers/net/wireless/intel/iwlegacy/common.c 		_il_wr(il, HBUS_TARG_WRPTR, txq->q.write_ptr | (txq_id << 8));
txq              2742 drivers/net/wireless/intel/iwlegacy/common.c 	txq->need_update = 0;
txq              2752 drivers/net/wireless/intel/iwlegacy/common.c 	struct il_tx_queue *txq = &il->txq[txq_id];
txq              2753 drivers/net/wireless/intel/iwlegacy/common.c 	struct il_queue *q = &txq->q;
txq              2759 drivers/net/wireless/intel/iwlegacy/common.c 		il->ops->txq_free_tfd(il, txq);
txq              2776 drivers/net/wireless/intel/iwlegacy/common.c 	struct il_tx_queue *txq = &il->txq[txq_id];
txq              2783 drivers/net/wireless/intel/iwlegacy/common.c 	if (txq->cmd) {
txq              2785 drivers/net/wireless/intel/iwlegacy/common.c 			kfree(txq->cmd[i]);
txq              2789 drivers/net/wireless/intel/iwlegacy/common.c 	if (txq->q.n_bd)
txq              2790 drivers/net/wireless/intel/iwlegacy/common.c 		dma_free_coherent(dev, il->hw_params.tfd_size * txq->q.n_bd,
txq              2791 drivers/net/wireless/intel/iwlegacy/common.c 				  txq->tfds, txq->q.dma_addr);
txq              2794 drivers/net/wireless/intel/iwlegacy/common.c 	kfree(txq->skbs);
txq              2795 drivers/net/wireless/intel/iwlegacy/common.c 	txq->skbs = NULL;
txq              2798 drivers/net/wireless/intel/iwlegacy/common.c 	kfree(txq->cmd);
txq              2799 drivers/net/wireless/intel/iwlegacy/common.c 	kfree(txq->meta);
txq              2800 drivers/net/wireless/intel/iwlegacy/common.c 	txq->cmd = NULL;
txq              2801 drivers/net/wireless/intel/iwlegacy/common.c 	txq->meta = NULL;
txq              2804 drivers/net/wireless/intel/iwlegacy/common.c 	memset(txq, 0, sizeof(*txq));
txq              2814 drivers/net/wireless/intel/iwlegacy/common.c 	struct il_tx_queue *txq = &il->txq[il->cmd_queue];
txq              2815 drivers/net/wireless/intel/iwlegacy/common.c 	struct il_queue *q = &txq->q;
txq              2824 drivers/net/wireless/intel/iwlegacy/common.c 		if (txq->meta[i].flags & CMD_MAPPED) {
txq              2826 drivers/net/wireless/intel/iwlegacy/common.c 					 dma_unmap_addr(&txq->meta[i], mapping),
txq              2827 drivers/net/wireless/intel/iwlegacy/common.c 					 dma_unmap_len(&txq->meta[i], len),
txq              2829 drivers/net/wireless/intel/iwlegacy/common.c 			txq->meta[i].flags = 0;
txq              2836 drivers/net/wireless/intel/iwlegacy/common.c 	if (txq->meta[i].flags & CMD_MAPPED) {
txq              2838 drivers/net/wireless/intel/iwlegacy/common.c 				 dma_unmap_addr(&txq->meta[i], mapping),
txq              2839 drivers/net/wireless/intel/iwlegacy/common.c 				 dma_unmap_len(&txq->meta[i], len),
txq              2841 drivers/net/wireless/intel/iwlegacy/common.c 		txq->meta[i].flags = 0;
txq              2857 drivers/net/wireless/intel/iwlegacy/common.c 	struct il_tx_queue *txq = &il->txq[il->cmd_queue];
txq              2864 drivers/net/wireless/intel/iwlegacy/common.c 	if (txq->cmd) {
txq              2866 drivers/net/wireless/intel/iwlegacy/common.c 			kfree(txq->cmd[i]);
txq              2870 drivers/net/wireless/intel/iwlegacy/common.c 	if (txq->q.n_bd)
txq              2871 drivers/net/wireless/intel/iwlegacy/common.c 		dma_free_coherent(dev, il->hw_params.tfd_size * txq->q.n_bd,
txq              2872 drivers/net/wireless/intel/iwlegacy/common.c 				  txq->tfds, txq->q.dma_addr);
txq              2875 drivers/net/wireless/intel/iwlegacy/common.c 	kfree(txq->cmd);
txq              2876 drivers/net/wireless/intel/iwlegacy/common.c 	kfree(txq->meta);
txq              2877 drivers/net/wireless/intel/iwlegacy/common.c 	txq->cmd = NULL;
txq              2878 drivers/net/wireless/intel/iwlegacy/common.c 	txq->meta = NULL;
txq              2881 drivers/net/wireless/intel/iwlegacy/common.c 	memset(txq, 0, sizeof(*txq));
txq              2965 drivers/net/wireless/intel/iwlegacy/common.c il_tx_queue_alloc(struct il_priv *il, struct il_tx_queue *txq, u32 id)
txq              2973 drivers/net/wireless/intel/iwlegacy/common.c 		txq->skbs = kcalloc(TFD_QUEUE_SIZE_MAX,
txq              2976 drivers/net/wireless/intel/iwlegacy/common.c 		if (!txq->skbs) {
txq              2981 drivers/net/wireless/intel/iwlegacy/common.c 		txq->skbs = NULL;
txq              2985 drivers/net/wireless/intel/iwlegacy/common.c 	txq->tfds =
txq              2986 drivers/net/wireless/intel/iwlegacy/common.c 	    dma_alloc_coherent(dev, tfd_sz, &txq->q.dma_addr, GFP_KERNEL);
txq              2987 drivers/net/wireless/intel/iwlegacy/common.c 	if (!txq->tfds)
txq              2990 drivers/net/wireless/intel/iwlegacy/common.c 	txq->q.id = id;
txq              2995 drivers/net/wireless/intel/iwlegacy/common.c 	kfree(txq->skbs);
txq              2996 drivers/net/wireless/intel/iwlegacy/common.c 	txq->skbs = NULL;
txq              3009 drivers/net/wireless/intel/iwlegacy/common.c 	struct il_tx_queue *txq = &il->txq[txq_id];
txq              3027 drivers/net/wireless/intel/iwlegacy/common.c 	txq->meta =
txq              3029 drivers/net/wireless/intel/iwlegacy/common.c 	txq->cmd =
txq              3032 drivers/net/wireless/intel/iwlegacy/common.c 	if (!txq->meta || !txq->cmd)
txq              3041 drivers/net/wireless/intel/iwlegacy/common.c 		txq->cmd[i] = kmalloc(len, GFP_KERNEL);
txq              3042 drivers/net/wireless/intel/iwlegacy/common.c 		if (!txq->cmd[i])
txq              3047 drivers/net/wireless/intel/iwlegacy/common.c 	ret = il_tx_queue_alloc(il, txq, txq_id);
txq              3051 drivers/net/wireless/intel/iwlegacy/common.c 	txq->need_update = 0;
txq              3059 drivers/net/wireless/intel/iwlegacy/common.c 		il_set_swq_id(txq, txq_id, txq_id);
txq              3062 drivers/net/wireless/intel/iwlegacy/common.c 	il_queue_init(il, &txq->q, slots, txq_id);
txq              3065 drivers/net/wireless/intel/iwlegacy/common.c 	il->ops->txq_init(il, txq);
txq              3070 drivers/net/wireless/intel/iwlegacy/common.c 		kfree(txq->cmd[i]);
txq              3072 drivers/net/wireless/intel/iwlegacy/common.c 	kfree(txq->meta);
txq              3073 drivers/net/wireless/intel/iwlegacy/common.c 	txq->meta = NULL;
txq              3074 drivers/net/wireless/intel/iwlegacy/common.c 	kfree(txq->cmd);
txq              3075 drivers/net/wireless/intel/iwlegacy/common.c 	txq->cmd = NULL;
txq              3085 drivers/net/wireless/intel/iwlegacy/common.c 	struct il_tx_queue *txq = &il->txq[txq_id];
txq              3095 drivers/net/wireless/intel/iwlegacy/common.c 	memset(txq->meta, 0, sizeof(struct il_cmd_meta) * actual_slots);
txq              3096 drivers/net/wireless/intel/iwlegacy/common.c 	txq->need_update = 0;
txq              3099 drivers/net/wireless/intel/iwlegacy/common.c 	il_queue_init(il, &txq->q, slots, txq_id);
txq              3102 drivers/net/wireless/intel/iwlegacy/common.c 	il->ops->txq_init(il, txq);
txq              3120 drivers/net/wireless/intel/iwlegacy/common.c 	struct il_tx_queue *txq = &il->txq[il->cmd_queue];
txq              3121 drivers/net/wireless/intel/iwlegacy/common.c 	struct il_queue *q = &txq->q;
txq              3159 drivers/net/wireless/intel/iwlegacy/common.c 	out_cmd = txq->cmd[idx];
txq              3160 drivers/net/wireless/intel/iwlegacy/common.c 	out_meta = &txq->meta[idx];
txq              3218 drivers/net/wireless/intel/iwlegacy/common.c 	txq->need_update = 1;
txq              3222 drivers/net/wireless/intel/iwlegacy/common.c 		il->ops->txq_update_byte_cnt_tbl(il, txq, 0);
txq              3224 drivers/net/wireless/intel/iwlegacy/common.c 	il->ops->txq_attach_buf_to_tfd(il, txq, phys_addr, fix_size, 1,
txq              3229 drivers/net/wireless/intel/iwlegacy/common.c 	il_txq_update_write_ptr(il, txq);
txq              3246 drivers/net/wireless/intel/iwlegacy/common.c 	struct il_tx_queue *txq = &il->txq[txq_id];
txq              3247 drivers/net/wireless/intel/iwlegacy/common.c 	struct il_queue *q = &txq->q;
txq              3288 drivers/net/wireless/intel/iwlegacy/common.c 	struct il_tx_queue *txq = &il->txq[il->cmd_queue];
txq              3297 drivers/net/wireless/intel/iwlegacy/common.c 	     txq_id, il->cmd_queue, sequence, il->txq[il->cmd_queue].q.read_ptr,
txq              3298 drivers/net/wireless/intel/iwlegacy/common.c 	     il->txq[il->cmd_queue].q.write_ptr)) {
txq              3303 drivers/net/wireless/intel/iwlegacy/common.c 	cmd_idx = il_get_cmd_idx(&txq->q, idx, huge);
txq              3304 drivers/net/wireless/intel/iwlegacy/common.c 	cmd = txq->cmd[cmd_idx];
txq              3305 drivers/net/wireless/intel/iwlegacy/common.c 	meta = &txq->meta[cmd_idx];
txq              3307 drivers/net/wireless/intel/iwlegacy/common.c 	txq->time_stamp = jiffies;
txq              4639 drivers/net/wireless/intel/iwlegacy/common.c 	if (!il->txq)
txq              4640 drivers/net/wireless/intel/iwlegacy/common.c 		il->txq =
txq              4644 drivers/net/wireless/intel/iwlegacy/common.c 	if (!il->txq) {
txq              4655 drivers/net/wireless/intel/iwlegacy/common.c 	kfree(il->txq);
txq              4656 drivers/net/wireless/intel/iwlegacy/common.c 	il->txq = NULL;
txq              4763 drivers/net/wireless/intel/iwlegacy/common.c 	if (il->txq == NULL)
txq              4772 drivers/net/wireless/intel/iwlegacy/common.c 		q = &il->txq[i].q;
txq              4796 drivers/net/wireless/intel/iwlegacy/common.c 	struct il_tx_queue *txq = &il->txq[cnt];
txq              4797 drivers/net/wireless/intel/iwlegacy/common.c 	struct il_queue *q = &txq->q;
txq              4803 drivers/net/wireless/intel/iwlegacy/common.c 		txq->time_stamp = now;
txq              4808 drivers/net/wireless/intel/iwlegacy/common.c 	    txq->time_stamp +
txq              4813 drivers/net/wireless/intel/iwlegacy/common.c 		       jiffies_to_msecs(now - txq->time_stamp));
txq              1257 drivers/net/wireless/intel/iwlegacy/common.h 	struct il_tx_queue *txq;
txq              1534 drivers/net/wireless/intel/iwlegacy/common.h 					 struct il_tx_queue *txq,
txq              1537 drivers/net/wireless/intel/iwlegacy/common.h 				      struct il_tx_queue *txq, dma_addr_t addr,
txq              1539 drivers/net/wireless/intel/iwlegacy/common.h 	void (*txq_free_tfd) (struct il_priv *il, struct il_tx_queue *txq);
txq              1540 drivers/net/wireless/intel/iwlegacy/common.h 	int (*txq_init) (struct il_priv *il, struct il_tx_queue *txq);
txq              1751 drivers/net/wireless/intel/iwlegacy/common.h void il_txq_update_write_ptr(struct il_priv *il, struct il_tx_queue *txq);
txq              2248 drivers/net/wireless/intel/iwlegacy/common.h il_set_swq_id(struct il_tx_queue *txq, u8 ac, u8 hwq)
txq              2253 drivers/net/wireless/intel/iwlegacy/common.h 	txq->swq_id = (hwq << 2) | ac;
txq              2270 drivers/net/wireless/intel/iwlegacy/common.h il_wake_queue(struct il_priv *il, struct il_tx_queue *txq)
txq              2272 drivers/net/wireless/intel/iwlegacy/common.h 	u8 queue = txq->swq_id;
txq              2281 drivers/net/wireless/intel/iwlegacy/common.h il_stop_queue(struct il_priv *il, struct il_tx_queue *txq)
txq              2283 drivers/net/wireless/intel/iwlegacy/common.h 	u8 queue = txq->swq_id;
txq               817 drivers/net/wireless/intel/iwlegacy/debug.c 	struct il_tx_queue *txq;
txq               826 drivers/net/wireless/intel/iwlegacy/debug.c 	if (!il->txq) {
txq               835 drivers/net/wireless/intel/iwlegacy/debug.c 		txq = &il->txq[cnt];
txq               836 drivers/net/wireless/intel/iwlegacy/debug.c 		q = &txq->q;
txq               843 drivers/net/wireless/intel/iwlegacy/debug.c 			      txq->swq_id, txq->swq_id & 3,
txq               844 drivers/net/wireless/intel/iwlegacy/debug.c 			      (txq->swq_id >> 2) & 0x1f);
txq               807 drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c void iwl_mvm_mac_itxq_xmit(struct ieee80211_hw *hw, struct ieee80211_txq *txq)
txq               810 drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c 	struct iwl_mvm_txq *mvmtxq = iwl_mvm_txq_from_mac80211(txq);
txq               840 drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c 			skb = ieee80211_tx_dequeue(hw, txq);
txq               843 drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c 				if (txq->sta)
txq               846 drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c 						     txq->sta->addr,
txq               847 drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c 						     txq->tid);
txq               851 drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c 			iwl_mvm_tx_skb(mvm, skb, txq->sta);
txq               858 drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c 				      struct ieee80211_txq *txq)
txq               861 drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c 	struct iwl_mvm_txq *mvmtxq = iwl_mvm_txq_from_mac80211(txq);
txq               880 drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c 	if (!txq->sta || mvmtxq->txq_id != IWL_MVM_INVALID_QUEUE) {
txq               886 drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c 		if (unlikely(txq->sta && !list_empty(&mvmtxq->list)))
txq               889 drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c 		iwl_mvm_mac_itxq_xmit(hw, txq);
txq               794 drivers/net/wireless/intel/iwlwifi/mvm/mvm.h iwl_mvm_txq_from_mac80211(struct ieee80211_txq *txq)
txq               796 drivers/net/wireless/intel/iwlwifi/mvm/mvm.h 	return (void *)txq->drv_priv;
txq               805 drivers/net/wireless/intel/iwlwifi/mvm/mvm.h 	return (void *)sta->txq[tid]->drv_priv;
txq              1520 drivers/net/wireless/intel/iwlwifi/mvm/mvm.h void iwl_mvm_mac_itxq_xmit(struct ieee80211_hw *hw, struct ieee80211_txq *txq);
txq              1096 drivers/net/wireless/intel/iwlwifi/mvm/ops.c 	struct ieee80211_txq *txq;
txq              1140 drivers/net/wireless/intel/iwlwifi/mvm/ops.c 		txq = sta->txq[tid];
txq              1141 drivers/net/wireless/intel/iwlwifi/mvm/ops.c 		mvmtxq = iwl_mvm_txq_from_mac80211(txq);
txq              1145 drivers/net/wireless/intel/iwlwifi/mvm/ops.c 			iwl_mvm_mac_itxq_xmit(mvm->hw, txq);
txq               634 drivers/net/wireless/intel/iwlwifi/mvm/sta.c 				  bool force, struct iwl_mvm_txq *txq)
txq               670 drivers/net/wireless/intel/iwlwifi/mvm/sta.c 	txq->stopped = true;
txq               713 drivers/net/wireless/intel/iwlwifi/mvm/sta.c 	txq->stopped = false;
txq              1386 drivers/net/wireless/intel/iwlwifi/mvm/sta.c 		struct ieee80211_txq *txq;
txq              1392 drivers/net/wireless/intel/iwlwifi/mvm/sta.c 		txq = container_of((void *)mvmtxq, struct ieee80211_txq,
txq              1394 drivers/net/wireless/intel/iwlwifi/mvm/sta.c 		tid = txq->tid;
txq              1398 drivers/net/wireless/intel/iwlwifi/mvm/sta.c 		iwl_mvm_sta_alloc_queue(mvm, txq->sta, txq->ac, tid);
txq              1401 drivers/net/wireless/intel/iwlwifi/mvm/sta.c 		iwl_mvm_mac_itxq_xmit(mvm->hw, txq);
txq              1644 drivers/net/wireless/intel/iwlwifi/mvm/sta.c 	for (i = 0; i < ARRAY_SIZE(sta->txq); i++) {
txq              1646 drivers/net/wireless/intel/iwlwifi/mvm/sta.c 			iwl_mvm_txq_from_mac80211(sta->txq[i]);
txq              1805 drivers/net/wireless/intel/iwlwifi/mvm/sta.c 	for (i = 0; i < ARRAY_SIZE(sta->txq); i++) {
txq              1807 drivers/net/wireless/intel/iwlwifi/mvm/sta.c 			iwl_mvm_txq_from_mac80211(sta->txq[i]);
txq              2020 drivers/net/wireless/intel/iwlwifi/mvm/sta.c 		int txq;
txq              2022 drivers/net/wireless/intel/iwlwifi/mvm/sta.c 		txq = iwl_mvm_enable_aux_snif_queue_tvqm(mvm, sta->sta_id);
txq              2023 drivers/net/wireless/intel/iwlwifi/mvm/sta.c 		if (txq < 0) {
txq              2025 drivers/net/wireless/intel/iwlwifi/mvm/sta.c 			return txq;
txq              2028 drivers/net/wireless/intel/iwlwifi/mvm/sta.c 		*queue = txq;
txq              1772 drivers/net/wireless/intel/iwlwifi/mvm/tx.c 			       int txq, int index,
txq              1800 drivers/net/wireless/intel/iwlwifi/mvm/tx.c 	if (tid_data->txq_id != txq) {
txq              1815 drivers/net/wireless/intel/iwlwifi/mvm/tx.c 	iwl_trans_reclaim(mvm->trans, txq, index, &reclaimed_skbs);
txq              1897 drivers/net/wireless/intel/iwlwifi/mvm/tx.c 	int sta_id, tid, txq, index;
txq              1962 drivers/net/wireless/intel/iwlwifi/mvm/tx.c 	txq = le16_to_cpu(ba_notif->scd_flow);
txq              1984 drivers/net/wireless/intel/iwlwifi/mvm/tx.c 	iwl_mvm_tx_reclaim(mvm, sta_id, tid, txq, index, &ba_info,
txq              1994 drivers/net/wireless/intel/iwlwifi/mvm/tx.c 			   le64_to_cpu(ba_notif->bitmap), txq, index,
txq               152 drivers/net/wireless/intel/iwlwifi/pcie/ctxt-info-gen3.c 		cpu_to_le64(trans_pcie->txq[trans_pcie->cmd_queue]->dma_addr);
txq               211 drivers/net/wireless/intel/iwlwifi/pcie/ctxt-info.c 		cpu_to_le64(trans_pcie->txq[trans_pcie->cmd_queue]->dma_addr);
txq               390 drivers/net/wireless/intel/iwlwifi/pcie/internal.h iwl_pcie_get_first_tb_dma(struct iwl_txq *txq, int idx)
txq               392 drivers/net/wireless/intel/iwlwifi/pcie/internal.h 	return txq->first_tb_dma +
txq               552 drivers/net/wireless/intel/iwlwifi/pcie/internal.h 	struct iwl_txq *txq[IWL_MAX_TVQM_QUEUES];
txq               691 drivers/net/wireless/intel/iwlwifi/pcie/internal.h 				  struct iwl_txq *txq);
txq               698 drivers/net/wireless/intel/iwlwifi/pcie/internal.h 				  struct iwl_txq *txq);
txq               706 drivers/net/wireless/intel/iwlwifi/pcie/internal.h 				   struct iwl_txq *txq, u16 byte_cnt,
txq               914 drivers/net/wireless/intel/iwlwifi/pcie/internal.h 				     struct iwl_txq *txq, int idx)
txq               919 drivers/net/wireless/intel/iwlwifi/pcie/internal.h 		idx = iwl_pcie_get_cmd_index(txq, idx);
txq               921 drivers/net/wireless/intel/iwlwifi/pcie/internal.h 	return txq->tfds + trans_pcie->tfd_size * idx;
txq               976 drivers/net/wireless/intel/iwlwifi/pcie/internal.h 				  struct iwl_txq *txq)
txq               980 drivers/net/wireless/intel/iwlwifi/pcie/internal.h 	if (test_and_clear_bit(txq->id, trans_pcie->queue_stopped)) {
txq               981 drivers/net/wireless/intel/iwlwifi/pcie/internal.h 		IWL_DEBUG_TX_QUEUES(trans, "Wake hwq %d\n", txq->id);
txq               982 drivers/net/wireless/intel/iwlwifi/pcie/internal.h 		iwl_op_mode_queue_not_full(trans->op_mode, txq->id);
txq               987 drivers/net/wireless/intel/iwlwifi/pcie/internal.h 				  struct iwl_txq *txq)
txq               991 drivers/net/wireless/intel/iwlwifi/pcie/internal.h 	if (!test_and_set_bit(txq->id, trans_pcie->queue_stopped)) {
txq               992 drivers/net/wireless/intel/iwlwifi/pcie/internal.h 		iwl_op_mode_queue_full(trans->op_mode, txq->id);
txq               993 drivers/net/wireless/intel/iwlwifi/pcie/internal.h 		IWL_DEBUG_TX_QUEUES(trans, "Stop hwq %d\n", txq->id);
txq               996 drivers/net/wireless/intel/iwlwifi/pcie/internal.h 				    txq->id);
txq              1075 drivers/net/wireless/intel/iwlwifi/pcie/internal.h void iwl_pcie_txq_free_tfd(struct iwl_trans *trans, struct iwl_txq *txq);
txq              1079 drivers/net/wireless/intel/iwlwifi/pcie/internal.h int iwl_pcie_txq_init(struct iwl_trans *trans, struct iwl_txq *txq,
txq              1082 drivers/net/wireless/intel/iwlwifi/pcie/internal.h 		       struct iwl_txq *txq, int slots_num,  bool cmd_queue);
txq              1101 drivers/net/wireless/intel/iwlwifi/pcie/internal.h 				   struct iwl_txq *txq);
txq              1106 drivers/net/wireless/intel/iwlwifi/pcie/internal.h 				      struct iwl_txq *txq,
txq              1244 drivers/net/wireless/intel/iwlwifi/pcie/rx.c 	struct iwl_txq *txq = trans_pcie->txq[trans_pcie->cmd_queue];
txq              1319 drivers/net/wireless/intel/iwlwifi/pcie/rx.c 		cmd_index = iwl_pcie_get_cmd_index(txq, index);
txq              1329 drivers/net/wireless/intel/iwlwifi/pcie/rx.c 			kzfree(txq->entries[cmd_index].free_buf);
txq              1330 drivers/net/wireless/intel/iwlwifi/pcie/rx.c 			txq->entries[cmd_index].free_buf = NULL;
txq              1610 drivers/net/wireless/intel/iwlwifi/pcie/rx.c 		if (!trans_pcie->txq[i])
txq              1612 drivers/net/wireless/intel/iwlwifi/pcie/rx.c 		del_timer(&trans_pcie->txq[i]->stuck_timer);
txq              2222 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 		struct iwl_txq *txq = trans_pcie->txq[queue];
txq              2225 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 		spin_lock_bh(&txq->lock);
txq              2229 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 		if (txq->frozen == freeze)
txq              2235 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 		txq->frozen = freeze;
txq              2237 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 		if (txq->read_ptr == txq->write_ptr)
txq              2242 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 						txq->stuck_timer.expires))) {
txq              2250 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 			txq->frozen_expiry_remainder =
txq              2251 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 				txq->stuck_timer.expires - now;
txq              2252 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 			del_timer(&txq->stuck_timer);
txq              2260 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 		mod_timer(&txq->stuck_timer,
txq              2261 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 			  now + txq->frozen_expiry_remainder);
txq              2264 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 		spin_unlock_bh(&txq->lock);
txq              2274 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 		struct iwl_txq *txq = trans_pcie->txq[i];
txq              2279 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 		spin_lock_bh(&txq->lock);
txq              2281 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 		if (!block && !(WARN_ON_ONCE(!txq->block))) {
txq              2282 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 			txq->block--;
txq              2283 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 			if (!txq->block) {
txq              2285 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 					    txq->write_ptr | (i << 8));
txq              2288 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 			txq->block++;
txq              2291 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 		spin_unlock_bh(&txq->lock);
txq              2297 drivers/net/wireless/intel/iwlwifi/pcie/trans.c void iwl_trans_pcie_log_scd_error(struct iwl_trans *trans, struct iwl_txq *txq)
txq              2299 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 	u32 txq_id = txq->id;
txq              2306 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 			txq->read_ptr, txq->write_ptr);
txq              2318 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 		jiffies_to_msecs(txq->wd_timeout),
txq              2319 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 		txq->read_ptr, txq->write_ptr,
txq              2346 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 	struct iwl_txq *txq;
txq              2359 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 	txq = trans_pcie->txq[txq_idx];
txq              2361 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 	spin_lock_bh(&txq->lock);
txq              2362 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 	overflow_tx = txq->overflow_tx ||
txq              2363 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 		      !skb_queue_empty(&txq->overflow_q);
txq              2364 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 	spin_unlock_bh(&txq->lock);
txq              2366 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 	wr_ptr = READ_ONCE(txq->write_ptr);
txq              2368 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 	while ((txq->read_ptr != READ_ONCE(txq->write_ptr) ||
txq              2372 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 		u8 write_ptr = READ_ONCE(txq->write_ptr);
txq              2387 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 		spin_lock_bh(&txq->lock);
txq              2388 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 		overflow_tx = txq->overflow_tx ||
txq              2389 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 			      !skb_queue_empty(&txq->overflow_q);
txq              2390 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 		spin_unlock_bh(&txq->lock);
txq              2393 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 	if (txq->read_ptr != txq->write_ptr) {
txq              2396 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 		iwl_trans_pcie_log_scd_error(trans, txq);
txq              2551 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 	struct iwl_txq *txq;
txq              2571 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 		txq = trans_pcie->txq[cnt];
txq              2574 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 				cnt, txq->read_ptr, txq->write_ptr,
txq              2577 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 				 txq->need_update, txq->frozen,
txq              3229 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 	struct iwl_txq *cmdq = trans_pcie->txq[trans_pcie->cmd_queue];
txq                79 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	for (txq_id = 0; txq_id < ARRAY_SIZE(trans_pcie->txq); txq_id++) {
txq                80 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 		if (!trans_pcie->txq[txq_id])
txq                90 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 				   struct iwl_txq *txq, u16 byte_cnt,
txq                93 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	struct iwlagn_scd_bc_tbl *scd_bc_tbl = txq->bc_tbl.addr;
txq                95 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	struct iwl_gen3_bc_tbl *scd_bc_tbl_gen3 = txq->bc_tbl.addr;
txq                96 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	int idx = iwl_pcie_get_cmd_index(txq, txq->write_ptr);
txq               101 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	if (WARN(idx >= txq->n_window, "%d >= %d\n", idx, txq->n_window))
txq               136 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 				  struct iwl_txq *txq)
txq               138 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	lockdep_assert_held(&txq->lock);
txq               140 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	IWL_DEBUG_TX(trans, "Q:%d WR: 0x%x\n", txq->id, txq->write_ptr);
txq               146 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	iwl_write32(trans, HBUS_TARG_WRPTR, txq->write_ptr | (txq->id << 16));
txq               187 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c static void iwl_pcie_gen2_free_tfd(struct iwl_trans *trans, struct iwl_txq *txq)
txq               192 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	int idx = iwl_pcie_get_cmd_index(txq, txq->read_ptr);
txq               194 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	lockdep_assert_held(&txq->lock);
txq               196 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	iwl_pcie_gen2_tfd_unmap(trans, &txq->entries[idx].meta,
txq               197 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 				iwl_pcie_get_tfd(trans, txq, idx));
txq               200 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	if (txq->entries) {
txq               203 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 		skb = txq->entries[idx].skb;
txq               211 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 			txq->entries[idx].skb = NULL;
txq               374 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 					  struct iwl_txq *txq,
txq               381 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	int idx = iwl_pcie_get_cmd_index(txq, txq->write_ptr);
txq               382 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	struct iwl_tfh_tfd *tfd = iwl_pcie_get_tfd(trans, txq, idx);
txq               387 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	tb_phys = iwl_pcie_get_first_tb_dma(txq, idx);
txq               419 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	memcpy(&txq->first_tb_bufs[idx], dev_cmd, IWL_FIRST_TB_SIZE);
txq               463 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 				    struct iwl_txq *txq,
txq               471 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	int idx = iwl_pcie_get_cmd_index(txq, txq->write_ptr);
txq               472 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	struct iwl_tfh_tfd *tfd = iwl_pcie_get_tfd(trans, txq, idx);
txq               478 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	tb_phys = iwl_pcie_get_first_tb_dma(txq, idx);
txq               481 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	memcpy(&txq->first_tb_bufs[idx], dev_cmd, IWL_FIRST_TB_SIZE);
txq               551 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 					    struct iwl_txq *txq,
txq               557 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	int idx = iwl_pcie_get_cmd_index(txq, txq->write_ptr);
txq               558 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	struct iwl_tfh_tfd *tfd = iwl_pcie_get_tfd(trans, txq, idx);
txq               584 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 		return iwl_pcie_gen2_build_tx_amsdu(trans, txq, dev_cmd, skb,
txq               587 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	return iwl_pcie_gen2_build_tx(trans, txq, dev_cmd, skb, out_meta,
txq               596 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	struct iwl_txq *txq = trans_pcie->txq[txq_id];
txq               610 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	spin_lock(&txq->lock);
txq               612 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	if (iwl_queue_space(trans, txq) < txq->high_mark) {
txq               613 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 		iwl_stop_queue(trans, txq);
txq               616 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 		if (unlikely(iwl_queue_space(trans, txq) < 3)) {
txq               623 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 			__skb_queue_tail(&txq->overflow_q, skb);
txq               624 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 			spin_unlock(&txq->lock);
txq               629 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	idx = iwl_pcie_get_cmd_index(txq, txq->write_ptr);
txq               632 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	txq->entries[idx].skb = skb;
txq               633 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	txq->entries[idx].cmd = dev_cmd;
txq               640 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	out_meta = &txq->entries[idx].meta;
txq               643 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	tfd = iwl_pcie_gen2_build_tfd(trans, txq, dev_cmd, skb, out_meta);
txq               645 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 		spin_unlock(&txq->lock);
txq               662 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	iwl_pcie_gen2_update_byte_tbl(trans_pcie, txq, cmd_len,
txq               666 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	if (txq->read_ptr == txq->write_ptr && txq->wd_timeout)
txq               667 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 		mod_timer(&txq->stuck_timer, jiffies + txq->wd_timeout);
txq               670 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	txq->write_ptr = iwl_queue_inc_wrap(trans, txq->write_ptr);
txq               671 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	iwl_pcie_gen2_txq_inc_wr_ptr(trans, txq);
txq               676 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	spin_unlock(&txq->lock);
txq               695 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	struct iwl_txq *txq = trans_pcie->txq[trans_pcie->cmd_queue];
txq               776 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	spin_lock_bh(&txq->lock);
txq               778 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	idx = iwl_pcie_get_cmd_index(txq, txq->write_ptr);
txq               779 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	tfd = iwl_pcie_get_tfd(trans, txq, txq->write_ptr);
txq               782 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	if (iwl_queue_space(trans, txq) < ((cmd->flags & CMD_ASYNC) ? 2 : 1)) {
txq               783 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 		spin_unlock_bh(&txq->lock);
txq               791 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	out_cmd = txq->entries[idx].cmd;
txq               792 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	out_meta = &txq->entries[idx].meta;
txq               808 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 					 INDEX_TO_SEQ(txq->write_ptr));
txq               855 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 		     cmd_size, txq->write_ptr, idx, trans_pcie->cmd_queue);
txq               859 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	memcpy(&txq->first_tb_bufs[idx], out_cmd, tb0_size);
txq               860 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	iwl_pcie_gen2_set_tb(trans, tfd, iwl_pcie_get_first_tb_dma(txq, idx),
txq               901 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	if (WARN_ON_ONCE(txq->entries[idx].free_buf))
txq               902 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 		kzfree(txq->entries[idx].free_buf);
txq               903 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	txq->entries[idx].free_buf = dup_buf;
txq               908 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	if (txq->read_ptr == txq->write_ptr && txq->wd_timeout)
txq               909 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 		mod_timer(&txq->stuck_timer, jiffies + txq->wd_timeout);
txq               913 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	txq->write_ptr = iwl_queue_inc_wrap(trans, txq->write_ptr);
txq               914 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	iwl_pcie_gen2_txq_inc_wr_ptr(trans, txq);
txq               918 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	spin_unlock_bh(&txq->lock);
txq               932 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	struct iwl_txq *txq = trans_pcie->txq[trans_pcie->cmd_queue];
txq               963 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 			txq->read_ptr, txq->write_ptr);
txq              1004 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 		txq->entries[cmd_idx].meta.flags &= ~CMD_WANT_SKB;
txq              1051 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	struct iwl_txq *txq = trans_pcie->txq[txq_id];
txq              1053 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	spin_lock_bh(&txq->lock);
txq              1054 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	while (txq->write_ptr != txq->read_ptr) {
txq              1056 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 				   txq_id, txq->read_ptr);
txq              1059 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 			int idx = iwl_pcie_get_cmd_index(txq, txq->read_ptr);
txq              1060 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 			struct sk_buff *skb = txq->entries[idx].skb;
txq              1067 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 		iwl_pcie_gen2_free_tfd(trans, txq);
txq              1068 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 		txq->read_ptr = iwl_queue_inc_wrap(trans, txq->read_ptr);
txq              1071 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	while (!skb_queue_empty(&txq->overflow_q)) {
txq              1072 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 		struct sk_buff *skb = __skb_dequeue(&txq->overflow_q);
txq              1077 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	spin_unlock_bh(&txq->lock);
txq              1080 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	iwl_wake_queue(trans, txq);
txq              1084 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 				   struct iwl_txq *txq)
txq              1090 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	if (txq->tfds) {
txq              1092 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 				  trans_pcie->tfd_size * txq->n_window,
txq              1093 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 				  txq->tfds, txq->dma_addr);
txq              1095 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 				  sizeof(*txq->first_tb_bufs) * txq->n_window,
txq              1096 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 				  txq->first_tb_bufs, txq->first_tb_dma);
txq              1099 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	kfree(txq->entries);
txq              1100 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	iwl_pcie_free_dma_ptr(trans, &txq->bc_tbl);
txq              1101 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	kfree(txq);
txq              1115 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	struct iwl_txq *txq = trans_pcie->txq[txq_id];
txq              1118 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	if (WARN_ON(!txq))
txq              1125 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 		for (i = 0; i < txq->n_window; i++) {
txq              1126 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 			kzfree(txq->entries[i].cmd);
txq              1127 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 			kzfree(txq->entries[i].free_buf);
txq              1129 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	del_timer_sync(&txq->stuck_timer);
txq              1131 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	iwl_pcie_gen2_txq_free_memory(trans, txq);
txq              1133 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	trans_pcie->txq[txq_id] = NULL;
txq              1144 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	struct iwl_txq *txq;
txq              1145 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	txq = kzalloc(sizeof(*txq), GFP_KERNEL);
txq              1146 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	if (!txq)
txq              1148 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	ret = iwl_pcie_alloc_dma_ptr(trans, &txq->bc_tbl,
txq              1155 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 		kfree(txq);
txq              1159 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	ret = iwl_pcie_txq_alloc(trans, txq, size, false);
txq              1164 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	ret = iwl_pcie_txq_init(trans, txq, size, false);
txq              1170 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	txq->wd_timeout = msecs_to_jiffies(timeout);
txq              1172 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	*intxq = txq;
txq              1176 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	iwl_pcie_gen2_txq_free_memory(trans, txq);
txq              1181 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 				      struct iwl_txq *txq,
txq              1199 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	if (qid >= ARRAY_SIZE(trans_pcie->txq)) {
txq              1211 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	txq->id = qid;
txq              1212 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	trans_pcie->txq[qid] = txq;
txq              1216 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	txq->read_ptr = wr_ptr;
txq              1217 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	txq->write_ptr = wr_ptr;
txq              1226 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	iwl_pcie_gen2_txq_free_memory(trans, txq);
txq              1235 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	struct iwl_txq *txq = NULL;
txq              1249 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	ret = iwl_trans_pcie_dyn_txq_alloc_dma(trans, &txq, size, timeout);
txq              1253 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	cmd.tfdq_addr = cpu_to_le64(txq->dma_addr);
txq              1254 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	cmd.byte_cnt_addr = cpu_to_le64(txq->bc_tbl.dma);
txq              1261 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	return iwl_trans_pcie_txq_alloc_response(trans, txq, &hcmd);
txq              1264 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	iwl_pcie_gen2_txq_free_memory(trans, txq);
txq              1286 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	iwl_pcie_gen2_txq_free_memory(trans, trans_pcie->txq[queue]);
txq              1287 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	trans_pcie->txq[queue] = NULL;
txq              1300 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	for (i = 0; i < ARRAY_SIZE(trans_pcie->txq); i++) {
txq              1301 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 		if (!trans_pcie->txq[i])
txq              1315 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	if (!trans_pcie->txq[txq_id]) {
txq              1321 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 		trans_pcie->txq[txq_id] = queue;
txq              1328 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 		queue = trans_pcie->txq[txq_id];
txq              1337 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	trans_pcie->txq[txq_id]->id = txq_id;
txq               185 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	struct iwl_txq *txq = from_timer(txq, t, stuck_timer);
txq               186 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	struct iwl_trans_pcie *trans_pcie = txq->trans_pcie;
txq               189 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	spin_lock(&txq->lock);
txq               191 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	if (txq->read_ptr == txq->write_ptr) {
txq               192 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 		spin_unlock(&txq->lock);
txq               195 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	spin_unlock(&txq->lock);
txq               197 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	iwl_trans_pcie_log_scd_error(trans, txq);
txq               206 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 					     struct iwl_txq *txq, u16 byte_cnt,
txq               211 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	int write_ptr = txq->write_ptr;
txq               212 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	int txq_id = txq->id;
txq               216 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	struct iwl_device_tx_cmd *dev_cmd = txq->entries[txq->write_ptr].cmd;
txq               251 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 					    struct iwl_txq *txq)
txq               256 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	int txq_id = txq->id;
txq               257 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	int read_ptr = txq->read_ptr;
txq               260 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	struct iwl_device_tx_cmd *dev_cmd = txq->entries[read_ptr].cmd;
txq               281 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 				    struct iwl_txq *txq)
txq               285 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	int txq_id = txq->id;
txq               287 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	lockdep_assert_held(&txq->lock);
txq               310 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 			txq->need_update = true;
txq               319 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	IWL_DEBUG_TX(trans, "Q:%d WR: 0x%x\n", txq_id, txq->write_ptr);
txq               320 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	if (!txq->block)
txq               322 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 			    txq->write_ptr | (txq_id << 8));
txq               331 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 		struct iwl_txq *txq = trans_pcie->txq[i];
txq               336 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 		spin_lock_bh(&txq->lock);
txq               337 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 		if (txq->need_update) {
txq               338 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 			iwl_pcie_txq_inc_wr_ptr(trans, txq);
txq               339 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 			txq->need_update = false;
txq               341 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 		spin_unlock_bh(&txq->lock);
txq               405 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 			       struct iwl_txq *txq, int index)
txq               409 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	void *tfd = iwl_pcie_get_tfd(trans, txq, index);
txq               460 drivers/net/wireless/intel/iwlwifi/pcie/tx.c void iwl_pcie_txq_free_tfd(struct iwl_trans *trans, struct iwl_txq *txq)
txq               465 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	int rd_ptr = txq->read_ptr;
txq               466 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	int idx = iwl_pcie_get_cmd_index(txq, rd_ptr);
txq               468 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	lockdep_assert_held(&txq->lock);
txq               473 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	iwl_pcie_tfd_unmap(trans, &txq->entries[idx].meta, txq, rd_ptr);
txq               476 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	if (txq->entries) {
txq               479 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 		skb = txq->entries[idx].skb;
txq               487 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 			txq->entries[idx].skb = NULL;
txq               492 drivers/net/wireless/intel/iwlwifi/pcie/tx.c static int iwl_pcie_txq_build_tfd(struct iwl_trans *trans, struct iwl_txq *txq,
txq               499 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	tfd = txq->tfds + trans_pcie->tfd_size * txq->write_ptr;
txq               522 drivers/net/wireless/intel/iwlwifi/pcie/tx.c int iwl_pcie_txq_alloc(struct iwl_trans *trans, struct iwl_txq *txq,
txq               531 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	if (WARN_ON(txq->entries || txq->tfds))
txq               537 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	timer_setup(&txq->stuck_timer, iwl_pcie_txq_stuck_timer, 0);
txq               538 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	txq->trans_pcie = trans_pcie;
txq               540 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	txq->n_window = slots_num;
txq               542 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	txq->entries = kcalloc(slots_num,
txq               546 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	if (!txq->entries)
txq               551 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 			txq->entries[i].cmd =
txq               554 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 			if (!txq->entries[i].cmd)
txq               560 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	txq->tfds = dma_alloc_coherent(trans->dev, tfd_sz,
txq               561 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 				       &txq->dma_addr, GFP_KERNEL);
txq               562 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	if (!txq->tfds)
txq               565 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	BUILD_BUG_ON(IWL_FIRST_TB_SIZE_ALIGN != sizeof(*txq->first_tb_bufs));
txq               567 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	tb0_buf_sz = sizeof(*txq->first_tb_bufs) * slots_num;
txq               569 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	txq->first_tb_bufs = dma_alloc_coherent(trans->dev, tb0_buf_sz,
txq               570 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 					      &txq->first_tb_dma,
txq               572 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	if (!txq->first_tb_bufs)
txq               577 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	dma_free_coherent(trans->dev, tfd_sz, txq->tfds, txq->dma_addr);
txq               579 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	if (txq->entries && cmd_queue)
txq               581 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 			kfree(txq->entries[i].cmd);
txq               582 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	kfree(txq->entries);
txq               583 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	txq->entries = NULL;
txq               589 drivers/net/wireless/intel/iwlwifi/pcie/tx.c int iwl_pcie_txq_init(struct iwl_trans *trans, struct iwl_txq *txq,
txq               596 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	txq->need_update = false;
txq               606 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	ret = iwl_queue_init(txq, slots_num);
txq               610 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	spin_lock_init(&txq->lock);
txq               615 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 		lockdep_set_class(&txq->lock, &iwl_pcie_cmd_queue_lock_class);
txq               618 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	__skb_queue_head_init(&txq->overflow_q);
txq               658 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	struct iwl_txq *txq = trans_pcie->txq[txq_id];
txq               660 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	spin_lock_bh(&txq->lock);
txq               661 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	while (txq->write_ptr != txq->read_ptr) {
txq               663 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 				   txq_id, txq->read_ptr);
txq               666 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 			struct sk_buff *skb = txq->entries[txq->read_ptr].skb;
txq               673 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 		iwl_pcie_txq_free_tfd(trans, txq);
txq               674 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 		txq->read_ptr = iwl_queue_inc_wrap(trans, txq->read_ptr);
txq               676 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 		if (txq->read_ptr == txq->write_ptr) {
txq               686 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	while (!skb_queue_empty(&txq->overflow_q)) {
txq               687 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 		struct sk_buff *skb = __skb_dequeue(&txq->overflow_q);
txq               692 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	spin_unlock_bh(&txq->lock);
txq               695 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	iwl_wake_queue(trans, txq);
txq               709 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	struct iwl_txq *txq = trans_pcie->txq[txq_id];
txq               713 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	if (WARN_ON(!txq))
txq               720 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 		for (i = 0; i < txq->n_window; i++) {
txq               721 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 			kzfree(txq->entries[i].cmd);
txq               722 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 			kzfree(txq->entries[i].free_buf);
txq               726 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	if (txq->tfds) {
txq               730 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 				  txq->tfds, txq->dma_addr);
txq               731 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 		txq->dma_addr = 0;
txq               732 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 		txq->tfds = NULL;
txq               735 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 				  sizeof(*txq->first_tb_bufs) * txq->n_window,
txq               736 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 				  txq->first_tb_bufs, txq->first_tb_dma);
txq               739 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	kfree(txq->entries);
txq               740 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	txq->entries = NULL;
txq               742 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	del_timer_sync(&txq->stuck_timer);
txq               745 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	memset(txq, 0, sizeof(*txq));
txq               819 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 		struct iwl_txq *txq = trans_pcie->txq[txq_id];
txq               823 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 					   txq->dma_addr);
txq               827 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 					   txq->dma_addr >> 8);
txq               829 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 		txq->read_ptr = 0;
txq               830 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 		txq->write_ptr = 0;
txq               928 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 			trans_pcie->txq[txq_id] = NULL;
txq               997 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 		trans_pcie->txq[txq_id] = &trans_pcie->txq_memory[txq_id];
txq               998 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 		ret = iwl_pcie_txq_alloc(trans, trans_pcie->txq[txq_id],
txq              1004 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 		trans_pcie->txq[txq_id]->id = txq_id;
txq              1051 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 		ret = iwl_pcie_txq_init(trans, trans_pcie->txq[txq_id],
txq              1065 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 				   trans_pcie->txq[txq_id]->dma_addr >> 8);
txq              1081 drivers/net/wireless/intel/iwlwifi/pcie/tx.c static inline void iwl_pcie_txq_progress(struct iwl_txq *txq)
txq              1083 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	lockdep_assert_held(&txq->lock);
txq              1085 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	if (!txq->wd_timeout)
txq              1092 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	if (txq->frozen)
txq              1099 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	if (txq->read_ptr == txq->write_ptr)
txq              1100 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 		del_timer(&txq->stuck_timer);
txq              1102 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 		mod_timer(&txq->stuck_timer, jiffies + txq->wd_timeout);
txq              1110 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	struct iwl_txq *txq = trans_pcie->txq[txq_id];
txq              1111 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	int tfd_num = iwl_pcie_get_cmd_index(txq, ssn);
txq              1112 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	int read_ptr = iwl_pcie_get_cmd_index(txq, txq->read_ptr);
txq              1119 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	spin_lock_bh(&txq->lock);
txq              1131 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 			   txq_id, txq->read_ptr, tfd_num, ssn);
txq              1137 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	if (!iwl_queue_used(txq, last_to_free)) {
txq              1142 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 			txq->write_ptr, txq->read_ptr);
txq              1151 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	     txq->read_ptr = iwl_queue_inc_wrap(trans, txq->read_ptr),
txq              1152 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	     read_ptr = iwl_pcie_get_cmd_index(txq, txq->read_ptr)) {
txq              1153 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 		struct sk_buff *skb = txq->entries[read_ptr].skb;
txq              1162 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 		txq->entries[read_ptr].skb = NULL;
txq              1165 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 			iwl_pcie_txq_inval_byte_cnt_tbl(trans, txq);
txq              1167 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 		iwl_pcie_txq_free_tfd(trans, txq);
txq              1170 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	iwl_pcie_txq_progress(txq);
txq              1172 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	if (iwl_queue_space(trans, txq) > txq->low_mark &&
txq              1177 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 		skb_queue_splice_init(&txq->overflow_q, &overflow_skbs);
txq              1186 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 		txq->overflow_tx = true;
txq              1195 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 		spin_unlock_bh(&txq->lock);
txq              1212 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 		if (iwl_queue_space(trans, txq) > txq->low_mark)
txq              1213 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 			iwl_wake_queue(trans, txq);
txq              1215 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 		spin_lock_bh(&txq->lock);
txq              1216 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 		txq->overflow_tx = false;
txq              1220 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	spin_unlock_bh(&txq->lock);
txq              1227 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	struct iwl_txq *txq = trans_pcie->txq[txq_id];
txq              1229 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	spin_lock_bh(&txq->lock);
txq              1231 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	txq->write_ptr = ptr;
txq              1232 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	txq->read_ptr = txq->write_ptr;
txq              1234 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	spin_unlock_bh(&txq->lock);
txq              1287 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	struct iwl_txq *txq = trans_pcie->txq[txq_id];
txq              1292 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	lockdep_assert_held(&txq->lock);
txq              1294 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	idx = iwl_pcie_get_cmd_index(txq, idx);
txq              1295 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	r = iwl_pcie_get_cmd_index(txq, txq->read_ptr);
txq              1298 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	    (!iwl_queue_used(txq, idx))) {
txq              1303 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 			  txq->write_ptr, txq->read_ptr);
txq              1309 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 		txq->read_ptr = iwl_queue_inc_wrap(trans, txq->read_ptr);
txq              1313 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 				idx, txq->write_ptr, r);
txq              1318 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	if (txq->read_ptr == txq->write_ptr) {
txq              1324 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	iwl_pcie_txq_progress(txq);
txq              1361 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	struct iwl_txq *txq = trans_pcie->txq[txq_id];
txq              1368 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	txq->wd_timeout = msecs_to_jiffies(wdg_timeout);
txq              1393 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 			txq->ampdu = true;
txq              1402 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 			ssn = txq->read_ptr;
txq              1417 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 			!((ssn - txq->write_ptr) & 0x3f) &&
txq              1418 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 			(ssn != txq->write_ptr);
txq              1425 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	txq->read_ptr = (ssn & 0xff);
txq              1426 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	txq->write_ptr = (ssn & 0xff);
txq              1472 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	struct iwl_txq *txq = trans_pcie->txq[txq_id];
txq              1474 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	txq->ampdu = !shared_mode;
txq              1485 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	trans_pcie->txq[txq_id]->frozen_expiry_remainder = 0;
txq              1486 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	trans_pcie->txq[txq_id]->frozen = false;
txq              1508 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	trans_pcie->txq[txq_id]->ampdu = false;
txq              1528 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	struct iwl_txq *txq = trans_pcie->txq[trans_pcie->cmd_queue];
txq              1625 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	spin_lock_bh(&txq->lock);
txq              1627 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	if (iwl_queue_space(trans, txq) < ((cmd->flags & CMD_ASYNC) ? 2 : 1)) {
txq              1628 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 		spin_unlock_bh(&txq->lock);
txq              1636 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	idx = iwl_pcie_get_cmd_index(txq, txq->write_ptr);
txq              1637 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	out_cmd = txq->entries[idx].cmd;
txq              1638 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	out_meta = &txq->entries[idx].meta;
txq              1655 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 						 INDEX_TO_SEQ(txq->write_ptr));
txq              1663 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 						 INDEX_TO_SEQ(txq->write_ptr));
txq              1713 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 		     cmd_size, txq->write_ptr, idx, trans_pcie->cmd_queue);
txq              1717 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	memcpy(&txq->first_tb_bufs[idx], &out_cmd->hdr, tb0_size);
txq              1718 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	iwl_pcie_txq_build_tfd(trans, txq,
txq              1719 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 			       iwl_pcie_get_first_tb_dma(txq, idx),
txq              1729 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 			iwl_pcie_tfd_unmap(trans, out_meta, txq,
txq              1730 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 					   txq->write_ptr);
txq              1735 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 		iwl_pcie_txq_build_tfd(trans, txq, phys_addr,
txq              1753 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 			iwl_pcie_tfd_unmap(trans, out_meta, txq,
txq              1754 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 					   txq->write_ptr);
txq              1759 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 		iwl_pcie_txq_build_tfd(trans, txq, phys_addr, cmdlen[i], false);
txq              1764 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	if (WARN_ON_ONCE(txq->entries[idx].free_buf))
txq              1765 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 		kzfree(txq->entries[idx].free_buf);
txq              1766 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	txq->entries[idx].free_buf = dup_buf;
txq              1771 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	if (txq->read_ptr == txq->write_ptr && txq->wd_timeout)
txq              1772 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 		mod_timer(&txq->stuck_timer, jiffies + txq->wd_timeout);
txq              1783 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	txq->write_ptr = iwl_queue_inc_wrap(trans, txq->write_ptr);
txq              1784 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	iwl_pcie_txq_inc_wr_ptr(trans, txq);
txq              1789 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	spin_unlock_bh(&txq->lock);
txq              1813 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	struct iwl_txq *txq = trans_pcie->txq[trans_pcie->cmd_queue];
txq              1820 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 		 txq_id, trans_pcie->cmd_queue, sequence, txq->read_ptr,
txq              1821 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 		 txq->write_ptr)) {
txq              1826 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	spin_lock_bh(&txq->lock);
txq              1828 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	cmd_index = iwl_pcie_get_cmd_index(txq, index);
txq              1829 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	cmd = txq->entries[cmd_index].cmd;
txq              1830 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	meta = &txq->entries[cmd_index].meta;
txq              1834 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	iwl_pcie_tfd_unmap(trans, meta, txq, index);
txq              1864 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	spin_unlock_bh(&txq->lock);
txq              1892 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	struct iwl_txq *txq = trans_pcie->txq[trans_pcie->cmd_queue];
txq              1928 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 			txq->read_ptr, txq->write_ptr);
txq              1972 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 		txq->entries[cmd_idx].meta.flags &= ~CMD_WANT_SKB;
txq              2004 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 			     struct iwl_txq *txq, u8 hdr_len,
txq              2025 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 		iwl_pcie_txq_build_tfd(trans, txq, tb_phys, head_tb_len, false);
txq              2045 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 		tb_idx = iwl_pcie_txq_build_tfd(trans, txq, tb_phys,
txq              2100 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 				   struct iwl_txq *txq, u8 hdr_len,
txq              2106 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	struct iwl_trans_pcie *trans_pcie = txq->trans_pcie;
txq              2122 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 			     iwl_pcie_get_tfd(trans, txq, txq->write_ptr),
txq              2223 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 		iwl_pcie_txq_build_tfd(trans, txq, hdr_tb_phys,
txq              2249 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 			iwl_pcie_txq_build_tfd(trans, txq, tb_phys,
txq              2283 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 				   struct iwl_txq *txq, u8 hdr_len,
txq              2302 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	struct iwl_txq *txq;
txq              2313 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	txq = trans_pcie->txq[txq_id];
txq              2346 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	spin_lock(&txq->lock);
txq              2348 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	if (iwl_queue_space(trans, txq) < txq->high_mark) {
txq              2349 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 		iwl_stop_queue(trans, txq);
txq              2352 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 		if (unlikely(iwl_queue_space(trans, txq) < 3)) {
txq              2359 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 			__skb_queue_tail(&txq->overflow_q, skb);
txq              2361 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 			spin_unlock(&txq->lock);
txq              2372 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	WARN_ONCE(txq->ampdu &&
txq              2373 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 		  (wifi_seq & 0xff) != txq->write_ptr,
txq              2375 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 		  txq_id, wifi_seq, txq->write_ptr);
txq              2378 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	txq->entries[txq->write_ptr].skb = skb;
txq              2379 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	txq->entries[txq->write_ptr].cmd = dev_cmd;
txq              2383 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 			    INDEX_TO_SEQ(txq->write_ptr)));
txq              2385 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	tb0_phys = iwl_pcie_get_first_tb_dma(txq, txq->write_ptr);
txq              2393 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	out_meta = &txq->entries[txq->write_ptr].meta;
txq              2421 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	iwl_pcie_txq_build_tfd(trans, txq, tb0_phys,
txq              2432 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	iwl_pcie_txq_build_tfd(trans, txq, tb1_phys, tb1_len, false);
txq              2435 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 			     iwl_pcie_get_tfd(trans, txq,
txq              2436 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 					      txq->write_ptr),
txq              2448 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 		if (unlikely(iwl_fill_data_tbs_amsdu(trans, skb, txq, hdr_len,
txq              2455 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 		if (unlikely(iwl_fill_data_tbs(trans, skb, txq, hdr_len,
txq              2460 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 			if (unlikely(iwl_fill_data_tbs(trans, frag, txq, 0,
txq              2467 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	memcpy(&txq->first_tb_bufs[txq->write_ptr], dev_cmd, IWL_FIRST_TB_SIZE);
txq              2469 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	tfd = iwl_pcie_get_tfd(trans, txq, txq->write_ptr);
txq              2471 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	iwl_pcie_txq_update_byte_cnt_tbl(trans, txq, le16_to_cpu(tx_cmd->len),
txq              2477 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	if (txq->read_ptr == txq->write_ptr && txq->wd_timeout) {
txq              2484 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 		if (!txq->frozen)
txq              2485 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 			mod_timer(&txq->stuck_timer,
txq              2486 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 				  jiffies + txq->wd_timeout);
txq              2488 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 			txq->frozen_expiry_remainder = txq->wd_timeout;
txq              2492 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	txq->write_ptr = iwl_queue_inc_wrap(trans, txq->write_ptr);
txq              2494 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 		iwl_pcie_txq_inc_wr_ptr(trans, txq);
txq              2500 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	spin_unlock(&txq->lock);
txq              2503 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	iwl_pcie_tfd_unmap(trans, out_meta, txq, txq->write_ptr);
txq              2504 drivers/net/wireless/intel/iwlwifi/pcie/tx.c 	spin_unlock(&txq->lock);
txq                81 drivers/net/wireless/marvell/mwifiex/debugfs.c 	struct netdev_queue *txq;
txq               142 drivers/net/wireless/marvell/mwifiex/debugfs.c 		txq = netdev_get_tx_queue(netdev, i);
txq               143 drivers/net/wireless/marvell/mwifiex/debugfs.c 		p += sprintf(p, " %d:%s", i, netif_tx_queue_stopped(txq) ?
txq               791 drivers/net/wireless/marvell/mwifiex/main.c 	struct netdev_queue *txq;
txq               795 drivers/net/wireless/marvell/mwifiex/main.c 		txq = netdev_get_tx_queue(priv->netdev, index);
txq               796 drivers/net/wireless/marvell/mwifiex/main.c 		if (!netif_tx_queue_stopped(txq)) {
txq               797 drivers/net/wireless/marvell/mwifiex/main.c 			netif_tx_stop_queue(txq);
txq              1099 drivers/net/wireless/marvell/mwifiex/main.c 	struct netdev_queue *txq;
txq              1159 drivers/net/wireless/marvell/mwifiex/main.c 			txq = netdev_get_tx_queue(priv->netdev, idx);
txq              1161 drivers/net/wireless/marvell/mwifiex/main.c 				     netif_tx_queue_stopped(txq) ?
txq               282 drivers/net/wireless/marvell/mwifiex/txrx.c 	struct netdev_queue *txq;
txq               319 drivers/net/wireless/marvell/mwifiex/txrx.c 		txq = netdev_get_tx_queue(priv->netdev, index);
txq               320 drivers/net/wireless/marvell/mwifiex/txrx.c 		if (netif_tx_queue_stopped(txq)) {
txq               321 drivers/net/wireless/marvell/mwifiex/txrx.c 			netif_tx_wake_queue(txq);
txq               248 drivers/net/wireless/marvell/mwl8k.c 	struct mwl8k_tx_queue txq[MWL8K_MAX_TX_QUEUES];
txq              1453 drivers/net/wireless/marvell/mwl8k.c 	struct mwl8k_tx_queue *txq = priv->txq + index;
txq              1457 drivers/net/wireless/marvell/mwl8k.c 	txq->len = 0;
txq              1458 drivers/net/wireless/marvell/mwl8k.c 	txq->head = 0;
txq              1459 drivers/net/wireless/marvell/mwl8k.c 	txq->tail = 0;
txq              1463 drivers/net/wireless/marvell/mwl8k.c 	txq->txd = pci_zalloc_consistent(priv->pdev, size, &txq->txd_dma);
txq              1464 drivers/net/wireless/marvell/mwl8k.c 	if (txq->txd == NULL) {
txq              1469 drivers/net/wireless/marvell/mwl8k.c 	txq->skb = kcalloc(MWL8K_TX_DESCS, sizeof(*txq->skb), GFP_KERNEL);
txq              1470 drivers/net/wireless/marvell/mwl8k.c 	if (txq->skb == NULL) {
txq              1471 drivers/net/wireless/marvell/mwl8k.c 		pci_free_consistent(priv->pdev, size, txq->txd, txq->txd_dma);
txq              1479 drivers/net/wireless/marvell/mwl8k.c 		tx_desc = txq->txd + i;
txq              1484 drivers/net/wireless/marvell/mwl8k.c 			cpu_to_le32(txq->txd_dma + nexti * sizeof(*tx_desc));
txq              1505 drivers/net/wireless/marvell/mwl8k.c 		struct mwl8k_tx_queue *txq = priv->txq + i;
txq              1512 drivers/net/wireless/marvell/mwl8k.c 			struct mwl8k_tx_desc *tx_desc = txq->txd + desc;
txq              1529 drivers/net/wireless/marvell/mwl8k.c 			  txq->len, txq->head, txq->tail,
txq              1670 drivers/net/wireless/marvell/mwl8k.c 	struct mwl8k_tx_queue *txq = priv->txq + index;
txq              1674 drivers/net/wireless/marvell/mwl8k.c 	while (txq->len > 0 && limit--) {
txq              1687 drivers/net/wireless/marvell/mwl8k.c 		tx = txq->head;
txq              1688 drivers/net/wireless/marvell/mwl8k.c 		tx_desc = txq->txd + tx;
txq              1699 drivers/net/wireless/marvell/mwl8k.c 		txq->head = (tx + 1) % MWL8K_TX_DESCS;
txq              1700 drivers/net/wireless/marvell/mwl8k.c 		BUG_ON(txq->len == 0);
txq              1701 drivers/net/wireless/marvell/mwl8k.c 		txq->len--;
txq              1706 drivers/net/wireless/marvell/mwl8k.c 		skb = txq->skb[tx];
txq              1707 drivers/net/wireless/marvell/mwl8k.c 		txq->skb[tx] = NULL;
txq              1767 drivers/net/wireless/marvell/mwl8k.c 	struct mwl8k_tx_queue *txq = priv->txq + index;
txq              1769 drivers/net/wireless/marvell/mwl8k.c 	if (txq->txd == NULL)
txq              1774 drivers/net/wireless/marvell/mwl8k.c 	kfree(txq->skb);
txq              1775 drivers/net/wireless/marvell/mwl8k.c 	txq->skb = NULL;
txq              1779 drivers/net/wireless/marvell/mwl8k.c 			    txq->txd, txq->txd_dma);
txq              1780 drivers/net/wireless/marvell/mwl8k.c 	txq->txd = NULL;
txq              1902 drivers/net/wireless/marvell/mwl8k.c 	struct mwl8k_tx_queue *txq;
txq              2061 drivers/net/wireless/marvell/mwl8k.c 	txq = priv->txq + index;
txq              2071 drivers/net/wireless/marvell/mwl8k.c 	if (txq->len >= MWL8K_TX_DESCS - 2) {
txq              2072 drivers/net/wireless/marvell/mwl8k.c 		if (!mgmtframe || txq->len == MWL8K_TX_DESCS) {
txq              2087 drivers/net/wireless/marvell/mwl8k.c 	BUG_ON(txq->skb[txq->tail] != NULL);
txq              2088 drivers/net/wireless/marvell/mwl8k.c 	txq->skb[txq->tail] = skb;
txq              2090 drivers/net/wireless/marvell/mwl8k.c 	tx = txq->txd + txq->tail;
txq              2111 drivers/net/wireless/marvell/mwl8k.c 	txq->len++;
txq              2114 drivers/net/wireless/marvell/mwl8k.c 	txq->tail++;
txq              2115 drivers/net/wireless/marvell/mwl8k.c 	if (txq->tail == MWL8K_TX_DESCS)
txq              2116 drivers/net/wireless/marvell/mwl8k.c 		txq->tail = 0;
txq              2473 drivers/net/wireless/marvell/mwl8k.c 		cmd->tx_queue_ptrs[i] = cpu_to_le32(priv->txq[i].txd_dma);
txq              2647 drivers/net/wireless/marvell/mwl8k.c 		cmd->tx_queue_ptrs[i] = cpu_to_le32(priv->txq[j].txd_dma);
txq              3435 drivers/net/wireless/marvell/mwl8k.c 			__u8 txq;
txq              3448 drivers/net/wireless/marvell/mwl8k.c 			__u8 txq;
txq              3482 drivers/net/wireless/marvell/mwl8k.c 		cmd->ap.txq = qnum;
txq              3487 drivers/net/wireless/marvell/mwl8k.c 		cmd->sta.txq = qnum;
txq              5856 drivers/net/wireless/marvell/mwl8k.c 			iowrite32(priv->txq[i].txd_dma,
txq               613 drivers/net/wireless/mediatek/mt76/mac80211.c 	for (i = 0; i < ARRAY_SIZE(sta->txq); i++) {
txq               616 drivers/net/wireless/mediatek/mt76/mac80211.c 		if (!sta->txq[i])
txq               619 drivers/net/wireless/mediatek/mt76/mac80211.c 		mtxq = (struct mt76_txq *)sta->txq[i]->drv_priv;
txq               621 drivers/net/wireless/mediatek/mt76/mac80211.c 			ieee80211_schedule_txq(dev->hw, sta->txq[i]);
txq               675 drivers/net/wireless/mediatek/mt76/mac80211.c 	for (i = 0; i < ARRAY_SIZE(sta->txq); i++) {
txq               678 drivers/net/wireless/mediatek/mt76/mac80211.c 		if (!sta->txq[i])
txq               681 drivers/net/wireless/mediatek/mt76/mac80211.c 		mtxq = (struct mt76_txq *)sta->txq[i]->drv_priv;
txq               684 drivers/net/wireless/mediatek/mt76/mac80211.c 		mt76_txq_init(dev, sta->txq[i]);
txq               712 drivers/net/wireless/mediatek/mt76/mac80211.c 	for (i = 0; i < ARRAY_SIZE(sta->txq); i++)
txq               713 drivers/net/wireless/mediatek/mt76/mac80211.c 		mt76_txq_remove(dev, sta->txq[i]);
txq               706 drivers/net/wireless/mediatek/mt76/mt76.h void mt76_txq_init(struct mt76_dev *dev, struct ieee80211_txq *txq);
txq               707 drivers/net/wireless/mediatek/mt76/mt76.h void mt76_txq_remove(struct mt76_dev *dev, struct ieee80211_txq *txq);
txq               708 drivers/net/wireless/mediatek/mt76/mt76.h void mt76_wake_tx_queue(struct ieee80211_hw *hw, struct ieee80211_txq *txq);
txq                74 drivers/net/wireless/mediatek/mt76/mt7603/main.c 	mtxq = (struct mt76_txq *)vif->txq->drv_priv;
txq                76 drivers/net/wireless/mediatek/mt76/mt7603/main.c 	mt76_txq_init(&dev->mt76, vif->txq);
txq                99 drivers/net/wireless/mediatek/mt76/mt7603/main.c 	mt76_txq_remove(&dev->mt76, vif->txq);
txq               552 drivers/net/wireless/mediatek/mt76/mt7603/main.c 	struct ieee80211_txq *txq = sta->txq[params->tid];
txq               559 drivers/net/wireless/mediatek/mt76/mt7603/main.c 	if (!txq)
txq               562 drivers/net/wireless/mediatek/mt76/mt7603/main.c 	mtxq = (struct mt76_txq *)txq->drv_priv;
txq               104 drivers/net/wireless/mediatek/mt76/mt7615/main.c 	mtxq = (struct mt76_txq *)vif->txq->drv_priv;
txq               106 drivers/net/wireless/mediatek/mt76/mt7615/main.c 	mt76_txq_init(&dev->mt76, vif->txq);
txq               126 drivers/net/wireless/mediatek/mt76/mt7615/main.c 	mt76_txq_remove(&dev->mt76, vif->txq);
txq               447 drivers/net/wireless/mediatek/mt76/mt7615/main.c 	struct ieee80211_txq *txq = sta->txq[params->tid];
txq               453 drivers/net/wireless/mediatek/mt76/mt7615/main.c 	if (!txq)
txq               456 drivers/net/wireless/mediatek/mt76/mt7615/main.c 	mtxq = (struct mt76_txq *)txq->drv_priv;
txq               299 drivers/net/wireless/mediatek/mt76/mt76x02_util.c 	mtxq = (struct mt76_txq *)vif->txq->drv_priv;
txq               302 drivers/net/wireless/mediatek/mt76/mt76x02_util.c 	mt76_txq_init(&dev->mt76, vif->txq);
txq               352 drivers/net/wireless/mediatek/mt76/mt76x02_util.c 	mt76_txq_remove(&dev->mt76, vif->txq);
txq               364 drivers/net/wireless/mediatek/mt76/mt76x02_util.c 	struct ieee80211_txq *txq = sta->txq[params->tid];
txq               369 drivers/net/wireless/mediatek/mt76/mt76x02_util.c 	if (!txq)
txq               372 drivers/net/wireless/mediatek/mt76/mt76x02_util.c 	mtxq = (struct mt76_txq *)txq->drv_priv;
txq                78 drivers/net/wireless/mediatek/mt76/tx.c mt76_txq_get_qid(struct ieee80211_txq *txq)
txq                80 drivers/net/wireless/mediatek/mt76/tx.c 	if (!txq->sta)
txq                83 drivers/net/wireless/mediatek/mt76/tx.c 	return txq->ac;
txq               266 drivers/net/wireless/mediatek/mt76/tx.c 		struct ieee80211_txq *txq;
txq               271 drivers/net/wireless/mediatek/mt76/tx.c 		txq = sta->txq[tid];
txq               272 drivers/net/wireless/mediatek/mt76/tx.c 		mtxq = (struct mt76_txq *)txq->drv_priv;
txq               296 drivers/net/wireless/mediatek/mt76/tx.c 	struct ieee80211_txq *txq = mtxq_to_txq(mtxq);
txq               304 drivers/net/wireless/mediatek/mt76/tx.c 			ieee80211_sta_set_buffered(txq->sta, tid, false);
txq               309 drivers/net/wireless/mediatek/mt76/tx.c 	skb = ieee80211_tx_dequeue(dev->hw, txq);
txq               345 drivers/net/wireless/mediatek/mt76/tx.c 		struct ieee80211_txq *txq = sta->txq[i];
txq               346 drivers/net/wireless/mediatek/mt76/tx.c 		struct mt76_txq *mtxq = (struct mt76_txq *)txq->drv_priv;
txq               383 drivers/net/wireless/mediatek/mt76/tx.c 	struct ieee80211_txq *txq = mtxq_to_txq(mtxq);
txq               384 drivers/net/wireless/mediatek/mt76/tx.c 	enum mt76_txq_id qid = mt76_txq_get_qid(txq);
txq               408 drivers/net/wireless/mediatek/mt76/tx.c 		ieee80211_get_tx_rates(txq->vif, txq->sta, skb,
txq               419 drivers/net/wireless/mediatek/mt76/tx.c 	idx = dev->queue_ops->tx_queue_skb(dev, qid, skb, wcid, txq->sta);
txq               454 drivers/net/wireless/mediatek/mt76/tx.c 						   txq->sta);
txq               477 drivers/net/wireless/mediatek/mt76/tx.c 	struct ieee80211_txq *txq;
txq               494 drivers/net/wireless/mediatek/mt76/tx.c 		txq = ieee80211_next_txq(dev->hw, qid);
txq               495 drivers/net/wireless/mediatek/mt76/tx.c 		if (!txq)
txq               498 drivers/net/wireless/mediatek/mt76/tx.c 		mtxq = (struct mt76_txq *)txq->drv_priv;
txq               504 drivers/net/wireless/mediatek/mt76/tx.c 			struct ieee80211_txq *txq = mtxq_to_txq(mtxq);
txq               505 drivers/net/wireless/mediatek/mt76/tx.c 			struct ieee80211_sta *sta = txq->sta;
txq               506 drivers/net/wireless/mediatek/mt76/tx.c 			struct ieee80211_vif *vif = txq->vif;
txq               508 drivers/net/wireless/mediatek/mt76/tx.c 			u8 tid = txq->tid;
txq               519 drivers/net/wireless/mediatek/mt76/tx.c 		ieee80211_return_txq(dev->hw, txq, !empty);
txq               570 drivers/net/wireless/mediatek/mt76/tx.c 	for (i = 0; i < ARRAY_SIZE(sta->txq); i++) {
txq               571 drivers/net/wireless/mediatek/mt76/tx.c 		struct ieee80211_txq *txq = sta->txq[i];
txq               575 drivers/net/wireless/mediatek/mt76/tx.c 		if (!txq)
txq               578 drivers/net/wireless/mediatek/mt76/tx.c 		mtxq = (struct mt76_txq *)txq->drv_priv;
txq               588 drivers/net/wireless/mediatek/mt76/tx.c void mt76_wake_tx_queue(struct ieee80211_hw *hw, struct ieee80211_txq *txq)
txq               599 drivers/net/wireless/mediatek/mt76/tx.c void mt76_txq_remove(struct mt76_dev *dev, struct ieee80211_txq *txq)
txq               604 drivers/net/wireless/mediatek/mt76/tx.c 	if (!txq)
txq               607 drivers/net/wireless/mediatek/mt76/tx.c 	mtxq = (struct mt76_txq *)txq->drv_priv;
txq               614 drivers/net/wireless/mediatek/mt76/tx.c void mt76_txq_init(struct mt76_dev *dev, struct ieee80211_txq *txq)
txq               616 drivers/net/wireless/mediatek/mt76/tx.c 	struct mt76_txq *mtxq = (struct mt76_txq *)txq->drv_priv;
txq               620 drivers/net/wireless/mediatek/mt76/tx.c 	mtxq->swq = &dev->q_tx[mt76_txq_get_qid(txq)];
txq                98 drivers/net/xen-netback/rx.c 			struct netdev_queue *txq;
txq               100 drivers/net/xen-netback/rx.c 			txq = netdev_get_tx_queue(queue->vif->dev, queue->id);
txq               101 drivers/net/xen-netback/rx.c 			netif_tx_wake_queue(txq);
txq               874 drivers/s390/net/qeth_core.h static inline u16 qeth_iqd_translate_txq(struct net_device *dev, u16 txq)
txq               876 drivers/s390/net/qeth_core.h 	if (txq == QETH_IQD_MCAST_TXQ)
txq               878 drivers/s390/net/qeth_core.h 	if (txq == dev->num_tx_queues - 1)
txq               880 drivers/s390/net/qeth_core.h 	return txq;
txq              3499 drivers/s390/net/qeth_core_main.c 	struct netdev_queue *txq;
txq              3520 drivers/s390/net/qeth_core_main.c 	txq = netdev_get_tx_queue(dev, __queue);
txq              3525 drivers/s390/net/qeth_core_main.c 	if (netif_tx_queue_stopped(txq) && !qeth_out_queue_is_full(queue))
txq              3526 drivers/s390/net/qeth_core_main.c 		netif_tx_wake_queue(txq);
txq              3835 drivers/s390/net/qeth_core_main.c 	struct netdev_queue *txq;
txq              3839 drivers/s390/net/qeth_core_main.c 	txq = netdev_get_tx_queue(card->dev, skb_get_queue_mapping(skb));
txq              3864 drivers/s390/net/qeth_core_main.c 		netif_tx_stop_queue(txq);
txq              3872 drivers/s390/net/qeth_core_main.c 	flush = __netdev_tx_sent_queue(txq, bytes,
txq              3881 drivers/s390/net/qeth_core_main.c 		netif_tx_start_queue(txq);
txq              3892 drivers/s390/net/qeth_core_main.c 	struct netdev_queue *txq;
txq              3914 drivers/s390/net/qeth_core_main.c 	txq = netdev_get_tx_queue(card->dev, skb_get_queue_mapping(skb));
txq              3950 drivers/s390/net/qeth_core_main.c 		netif_tx_stop_queue(txq);
txq              3996 drivers/s390/net/qeth_core_main.c 		netif_tx_start_queue(txq);
txq              5257 drivers/s390/net/qeth_core_main.c 	struct netdev_queue *txq;
txq              5259 drivers/s390/net/qeth_core_main.c 	txq = netdev_get_tx_queue(dev, qeth_iqd_translate_txq(dev, queue_no));
txq              5302 drivers/s390/net/qeth_core_main.c 		netdev_tx_completed_queue(txq, packets, bytes);
txq              5311 drivers/s390/net/qeth_core_main.c 		if (netif_tx_queue_stopped(txq) &&
txq              5313 drivers/s390/net/qeth_core_main.c 			netif_tx_wake_queue(txq);
txq               592 drivers/s390/net/qeth_l2_main.c 	u16 txq = skb_get_queue_mapping(skb);
txq               599 drivers/s390/net/qeth_l2_main.c 		txq = qeth_iqd_translate_txq(dev, txq);
txq               600 drivers/s390/net/qeth_l2_main.c 	queue = card->qdio.out_qs[txq];
txq              2044 drivers/s390/net/qeth_l3_main.c 	u16 txq = skb_get_queue_mapping(skb);
txq              2052 drivers/s390/net/qeth_l3_main.c 		queue = card->qdio.out_qs[qeth_iqd_translate_txq(dev, txq)];
txq              2061 drivers/s390/net/qeth_l3_main.c 		queue = card->qdio.out_qs[txq];
txq              5926 drivers/scsi/lpfc/lpfc_bsg.c 		list_for_each_entry_safe(check_iocb, next_iocb, &pring->txq,
txq              5963 drivers/scsi/lpfc/lpfc_bsg.c 		list_for_each_entry_safe(check_iocb, next_iocb, &pring->txq,
txq              8083 drivers/scsi/lpfc/lpfc_els.c 	list_for_each_entry_safe(piocb, tmp_iocb, &pring->txq, list) {
txq              9973 drivers/scsi/lpfc/lpfc_els.c 			if (pring && !list_empty(&pring->txq))
txq              4752 drivers/scsi/lpfc/lpfc_hbadisc.c 	list_for_each_entry_safe(iocb, next_iocb, &pring->txq, list) {
txq              5710 drivers/scsi/lpfc/lpfc_hbadisc.c 	list_for_each_entry_safe(iocb, next_iocb, &pring->txq, list) {
txq               263 drivers/scsi/lpfc/lpfc_nportdisc.c 	list_for_each_entry_safe(iocb, next_iocb, &pring->txq, list) {
txq               571 drivers/scsi/lpfc/lpfc_scsi.c 		if (!list_empty(&pring->txq))
txq              1292 drivers/scsi/lpfc/lpfc_sli.c 			if (!list_empty(&pring->txq))
txq              1625 drivers/scsi/lpfc/lpfc_sli.c 	list_remove_head((&pring->txq), cmd_iocb, struct lpfc_iocbq, list);
txq              1902 drivers/scsi/lpfc/lpfc_sli.c 	    (!list_empty(&pring->txq)) &&
txq              3941 drivers/scsi/lpfc/lpfc_sli.c 		list_splice_init(&pring->txq, &completions);
txq              3952 drivers/scsi/lpfc/lpfc_sli.c 		list_splice_init(&pring->txq, &completions);
txq              4008 drivers/scsi/lpfc/lpfc_sli.c 	LIST_HEAD(txq);
txq              4032 drivers/scsi/lpfc/lpfc_sli.c 			list_splice_init(&pring->txq, &txq);
txq              4043 drivers/scsi/lpfc/lpfc_sli.c 			lpfc_sli_cancel_iocbs(phba, &txq,
txq              4056 drivers/scsi/lpfc/lpfc_sli.c 		list_splice_init(&pring->txq, &txq);
txq              4067 drivers/scsi/lpfc/lpfc_sli.c 		lpfc_sli_cancel_iocbs(phba, &txq, IOSTAT_LOCAL_REJECT,
txq              8987 drivers/scsi/lpfc/lpfc_sli.c 	list_add_tail(&piocb->list, &pring->txq);
txq              9911 drivers/scsi/lpfc/lpfc_sli.c 			if (!list_empty(&pring->txq)) {
txq              10501 drivers/scsi/lpfc/lpfc_sli.c 		INIT_LIST_HEAD(&pring->txq);
txq              10510 drivers/scsi/lpfc/lpfc_sli.c 	INIT_LIST_HEAD(&pring->txq);
txq              10520 drivers/scsi/lpfc/lpfc_sli.c 		INIT_LIST_HEAD(&pring->txq);
txq              10562 drivers/scsi/lpfc/lpfc_sli.c 		INIT_LIST_HEAD(&pring->txq);
txq              10672 drivers/scsi/lpfc/lpfc_sli.c 						 &pring->txq, list) {
txq              10698 drivers/scsi/lpfc/lpfc_sli.c 						 &pring->txq, list) {
txq              10772 drivers/scsi/lpfc/lpfc_sli.c 			list_splice_init(&pring->txq, &completions);
txq              10780 drivers/scsi/lpfc/lpfc_sli.c 			list_splice_init(&pring->txq, &completions);
txq              11912 drivers/scsi/lpfc/lpfc_sli.c 			list_for_each_entry(iocb, &pring->txq, list) {
txq              13257 drivers/scsi/lpfc/lpfc_sli.c 		if (!list_empty(&pring->txq))
txq              19622 drivers/scsi/lpfc/lpfc_sli.c 	if (unlikely(!pring) || list_empty(&pring->txq))
txq              19626 drivers/scsi/lpfc/lpfc_sli.c 	list_for_each_entry(piocbq, &pring->txq, list) {
txq              19635 drivers/scsi/lpfc/lpfc_sli.c 	while (!list_empty(&pring->txq)) {
txq               235 drivers/scsi/lpfc/lpfc_sli.h 	struct list_head txq;
txq               197 drivers/target/iscsi/cxgbit/cxgbit.h 	struct sk_buff_head txq;
txq               635 drivers/target/iscsi/cxgbit/cxgbit_cm.c 	__skb_queue_tail(&csk->txq, skb);
txq               666 drivers/target/iscsi/cxgbit/cxgbit_cm.c 	__skb_queue_purge(&csk->txq);
txq               781 drivers/target/iscsi/cxgbit/cxgbit_cm.c 	__skb_queue_purge(&csk->txq);
txq              1356 drivers/target/iscsi/cxgbit/cxgbit_cm.c 	skb_queue_head_init(&csk->txq);
txq              1751 drivers/target/iscsi/cxgbit/cxgbit_cm.c 	__skb_queue_purge(&csk->txq);
txq              1878 drivers/target/iscsi/cxgbit/cxgbit_cm.c 	if (skb_queue_len(&csk->txq))
txq               194 drivers/target/iscsi/cxgbit/cxgbit_target.c 		 FW_OFLD_TX_DATA_WR_SHOVE_V(skb_peek(&csk->txq) ? 0 : 1));
txq               206 drivers/target/iscsi/cxgbit/cxgbit_target.c 	while (csk->wr_cred && ((skb = skb_peek(&csk->txq)) != NULL)) {
txq               244 drivers/target/iscsi/cxgbit/cxgbit_target.c 		__skb_unlink(skb, &csk->txq);
txq               347 drivers/target/iscsi/cxgbit/cxgbit_target.c 	skb_queue_splice_tail_init(&csk->ppodq, &csk->txq);
txq               348 drivers/target/iscsi/cxgbit/cxgbit_target.c 	__skb_queue_tail(&csk->txq, skb);
txq              2062 drivers/tty/n_gsm.c 	struct gsm_msg *txq, *ntxq;
txq              2090 drivers/tty/n_gsm.c 	list_for_each_entry_safe(txq, ntxq, &gsm->tx_list, list)
txq              2091 drivers/tty/n_gsm.c 		kfree(txq);
txq               322 include/linux/avf/virtchnl.h 	struct virtchnl_txq_info txq;
txq              2097 include/linux/netdevice.h int netdev_txq_to_tc(struct net_device *dev, unsigned int txq);
txq              3064 include/linux/netdevice.h void netif_schedule_queue(struct netdev_queue *txq);
txq              3095 include/linux/netdevice.h 		struct netdev_queue *txq = netdev_get_tx_queue(dev, i);
txq              3096 include/linux/netdevice.h 		netif_tx_start_queue(txq);
txq              3119 include/linux/netdevice.h 		struct netdev_queue *txq = netdev_get_tx_queue(dev, i);
txq              3120 include/linux/netdevice.h 		netif_tx_wake_queue(txq);
txq              3378 include/linux/netdevice.h 	struct netdev_queue *txq = netdev_get_tx_queue(dev, queue_index);
txq              3380 include/linux/netdevice.h 	netif_tx_start_queue(txq);
txq              3392 include/linux/netdevice.h 	struct netdev_queue *txq = netdev_get_tx_queue(dev, queue_index);
txq              3393 include/linux/netdevice.h 	netif_tx_stop_queue(txq);
txq              3406 include/linux/netdevice.h 	struct netdev_queue *txq = netdev_get_tx_queue(dev, queue_index);
txq              3408 include/linux/netdevice.h 	return netif_tx_queue_stopped(txq);
txq              3426 include/linux/netdevice.h 	struct netdev_queue *txq = netdev_get_tx_queue(dev, queue_index);
txq              3428 include/linux/netdevice.h 	netif_tx_wake_queue(txq);
txq              3547 include/linux/netdevice.h int netif_set_real_num_tx_queues(struct net_device *dev, unsigned int txq);
txq              3694 include/linux/netdevice.h 				    struct netdev_queue *txq, int *ret);
txq              3904 include/linux/netdevice.h static inline void __netif_tx_lock(struct netdev_queue *txq, int cpu)
txq              3906 include/linux/netdevice.h 	spin_lock(&txq->_xmit_lock);
txq              3907 include/linux/netdevice.h 	txq->xmit_lock_owner = cpu;
txq              3910 include/linux/netdevice.h static inline bool __netif_tx_acquire(struct netdev_queue *txq)
txq              3912 include/linux/netdevice.h 	__acquire(&txq->_xmit_lock);
txq              3916 include/linux/netdevice.h static inline void __netif_tx_release(struct netdev_queue *txq)
txq              3918 include/linux/netdevice.h 	__release(&txq->_xmit_lock);
txq              3921 include/linux/netdevice.h static inline void __netif_tx_lock_bh(struct netdev_queue *txq)
txq              3923 include/linux/netdevice.h 	spin_lock_bh(&txq->_xmit_lock);
txq              3924 include/linux/netdevice.h 	txq->xmit_lock_owner = smp_processor_id();
txq              3927 include/linux/netdevice.h static inline bool __netif_tx_trylock(struct netdev_queue *txq)
txq              3929 include/linux/netdevice.h 	bool ok = spin_trylock(&txq->_xmit_lock);
txq              3931 include/linux/netdevice.h 		txq->xmit_lock_owner = smp_processor_id();
txq              3935 include/linux/netdevice.h static inline void __netif_tx_unlock(struct netdev_queue *txq)
txq              3937 include/linux/netdevice.h 	txq->xmit_lock_owner = -1;
txq              3938 include/linux/netdevice.h 	spin_unlock(&txq->_xmit_lock);
txq              3941 include/linux/netdevice.h static inline void __netif_tx_unlock_bh(struct netdev_queue *txq)
txq              3943 include/linux/netdevice.h 	txq->xmit_lock_owner = -1;
txq              3944 include/linux/netdevice.h 	spin_unlock_bh(&txq->_xmit_lock);
txq              3947 include/linux/netdevice.h static inline void txq_trans_update(struct netdev_queue *txq)
txq              3949 include/linux/netdevice.h 	if (txq->xmit_lock_owner != -1)
txq              3950 include/linux/netdevice.h 		txq->trans_start = jiffies;
txq              3956 include/linux/netdevice.h 	struct netdev_queue *txq = netdev_get_tx_queue(dev, 0);
txq              3958 include/linux/netdevice.h 	if (txq->trans_start != jiffies)
txq              3959 include/linux/netdevice.h 		txq->trans_start = jiffies;
txq              3976 include/linux/netdevice.h 		struct netdev_queue *txq = netdev_get_tx_queue(dev, i);
txq              3984 include/linux/netdevice.h 		__netif_tx_lock(txq, cpu);
txq              3985 include/linux/netdevice.h 		set_bit(__QUEUE_STATE_FROZEN, &txq->state);
txq              3986 include/linux/netdevice.h 		__netif_tx_unlock(txq);
txq              4001 include/linux/netdevice.h 		struct netdev_queue *txq = netdev_get_tx_queue(dev, i);
txq              4007 include/linux/netdevice.h 		clear_bit(__QUEUE_STATE_FROZEN, &txq->state);
txq              4008 include/linux/netdevice.h 		netif_schedule_queue(txq);
txq              4019 include/linux/netdevice.h #define HARD_TX_LOCK(dev, txq, cpu) {			\
txq              4021 include/linux/netdevice.h 		__netif_tx_lock(txq, cpu);		\
txq              4023 include/linux/netdevice.h 		__netif_tx_acquire(txq);		\
txq              4027 include/linux/netdevice.h #define HARD_TX_TRYLOCK(dev, txq)			\
txq              4029 include/linux/netdevice.h 		__netif_tx_trylock(txq) :		\
txq              4030 include/linux/netdevice.h 		__netif_tx_acquire(txq))
txq              4032 include/linux/netdevice.h #define HARD_TX_UNLOCK(dev, txq) {			\
txq              4034 include/linux/netdevice.h 		__netif_tx_unlock(txq);			\
txq              4036 include/linux/netdevice.h 		__netif_tx_release(txq);		\
txq              4048 include/linux/netdevice.h 		struct netdev_queue *txq = netdev_get_tx_queue(dev, i);
txq              4050 include/linux/netdevice.h 		__netif_tx_lock(txq, cpu);
txq              4051 include/linux/netdevice.h 		netif_tx_stop_queue(txq);
txq              4052 include/linux/netdevice.h 		__netif_tx_unlock(txq);
txq              4428 include/linux/netdevice.h 					    struct netdev_queue *txq, bool more)
txq              4435 include/linux/netdevice.h 		txq_trans_update(txq);
txq                41 include/linux/netpoll.h 	struct sk_buff_head txq;
txq               142 include/linux/ti_wilink_st.h 	struct sk_buff_head txq, tx_waitq;
txq                59 include/linux/usb/usbnet.h 	struct sk_buff_head	txq;
txq              1620 include/net/mac80211.h 	struct ieee80211_txq *txq;
txq              2014 include/net/mac80211.h 	struct ieee80211_txq *txq[IEEE80211_NUM_TIDS + 1];
txq              4029 include/net/mac80211.h 			      struct ieee80211_txq *txq);
txq              6256 include/net/mac80211.h 				     struct ieee80211_txq *txq);
txq              6270 include/net/mac80211.h 						      struct ieee80211_txq *txq)
txq              6275 include/net/mac80211.h 	skb = ieee80211_tx_dequeue(hw, txq);
txq              6310 include/net/mac80211.h 			      struct ieee80211_txq *txq, bool force);
txq              6325 include/net/mac80211.h ieee80211_schedule_txq(struct ieee80211_hw *hw, struct ieee80211_txq *txq)
txq              6327 include/net/mac80211.h 	__ieee80211_schedule_txq(hw, txq, true);
txq              6341 include/net/mac80211.h ieee80211_return_txq(struct ieee80211_hw *hw, struct ieee80211_txq *txq,
txq              6344 include/net/mac80211.h 	__ieee80211_schedule_txq(hw, txq, force);
txq              6372 include/net/mac80211.h 				struct ieee80211_txq *txq);
txq              6385 include/net/mac80211.h void ieee80211_txq_get_depth(struct ieee80211_txq *txq,
txq               113 include/net/pkt_sched.h 		     struct net_device *dev, struct netdev_queue *txq,
txq               185 include/net/sch_generic.h static inline int qdisc_avail_bulklimit(const struct netdev_queue *txq)
txq               189 include/net/sch_generic.h 	return dql_avail(&txq->dql);
txq               726 include/net/sch_generic.h 		struct netdev_queue *txq = netdev_get_tx_queue(dev, i);
txq               727 include/net/sch_generic.h 		const struct Qdisc *q = rcu_dereference(txq->qdisc);
txq               744 include/net/sch_generic.h 		struct netdev_queue *txq = netdev_get_tx_queue(dev, i);
txq               745 include/net/sch_generic.h 		if (rcu_access_pointer(txq->qdisc) != txq->qdisc_sleeping)
txq               757 include/net/sch_generic.h 		struct netdev_queue *txq = netdev_get_tx_queue(dev, i);
txq               758 include/net/sch_generic.h 		if (rcu_access_pointer(txq->qdisc) != &noop_qdisc)
txq                14 include/trace/events/qdisc.h 	TP_PROTO(struct Qdisc *qdisc, const struct netdev_queue *txq,
txq                17 include/trace/events/qdisc.h 	TP_ARGS(qdisc, txq, packets, skb),
txq                21 include/trace/events/qdisc.h 		__field(const	struct netdev_queue *,	txq	)
txq                33 include/trace/events/qdisc.h 		__entry->txq		= txq;
txq                36 include/trace/events/qdisc.h 		__entry->ifindex	= txq->dev ? txq->dev->ifindex : 0;
txq                39 include/trace/events/qdisc.h 		__entry->txq_state	= txq->state;
txq               172 net/caif/caif_dev.c 	struct netdev_queue *txq;
txq               192 net/caif/caif_dev.c 		txq = netdev_get_tx_queue(skb->dev, 0);
txq               193 net/caif/caif_dev.c 		sch = rcu_dereference_bh(txq->qdisc);
txq              2016 net/core/dev.c static void netif_setup_tc(struct net_device *dev, unsigned int txq)
txq              2022 net/core/dev.c 	if (tc->offset + tc->count > txq) {
txq              2033 net/core/dev.c 		if (tc->offset + tc->count > txq) {
txq              2041 net/core/dev.c int netdev_txq_to_tc(struct net_device *dev, unsigned int txq)
txq              2049 net/core/dev.c 			if ((txq - tc->offset) < tc->count)
txq              2454 net/core/dev.c 	struct netdev_queue *txq = &dev->_tx[dev->num_tx_queues];
txq              2457 net/core/dev.c 	while (txq-- != &dev->_tx[0]) {
txq              2458 net/core/dev.c 		if (txq->sb_dev)
txq              2459 net/core/dev.c 			netdev_unbind_sb_channel(dev, txq->sb_dev);
txq              2509 net/core/dev.c 	struct netdev_queue *txq = &dev->_tx[dev->num_tx_queues];
txq              2517 net/core/dev.c 	while (txq-- != &dev->_tx[0]) {
txq              2518 net/core/dev.c 		if (txq->sb_dev == sb_dev)
txq              2519 net/core/dev.c 			txq->sb_dev = NULL;
txq              2574 net/core/dev.c int netif_set_real_num_tx_queues(struct net_device *dev, unsigned int txq)
txq              2579 net/core/dev.c 	disabling = txq < dev->real_num_tx_queues;
txq              2581 net/core/dev.c 	if (txq < 1 || txq > dev->num_tx_queues)
txq              2589 net/core/dev.c 						  txq);
txq              2594 net/core/dev.c 			netif_setup_tc(dev, txq);
txq              2596 net/core/dev.c 		dev->real_num_tx_queues = txq;
txq              2600 net/core/dev.c 			qdisc_reset_all_tx_gt(dev, txq);
txq              2602 net/core/dev.c 			netif_reset_xps_queues_gt(dev, txq);
txq              2606 net/core/dev.c 		dev->real_num_tx_queues = txq;
txq              2689 net/core/dev.c void netif_schedule_queue(struct netdev_queue *txq)
txq              2692 net/core/dev.c 	if (!(txq->state & QUEUE_STATE_ANY_XOFF)) {
txq              2693 net/core/dev.c 		struct Qdisc *q = rcu_dereference(txq->qdisc);
txq              3190 net/core/dev.c 		    struct netdev_queue *txq, bool more)
txq              3200 net/core/dev.c 	rc = netdev_start_xmit(skb, dev, txq, more);
txq              3207 net/core/dev.c 				    struct netdev_queue *txq, int *ret)
txq              3216 net/core/dev.c 		rc = xmit_one(skb, dev, txq, next != NULL);
txq              3223 net/core/dev.c 		if (netif_tx_queue_stopped(txq) && skb) {
txq              3380 net/core/dev.c 				 struct netdev_queue *txq)
txq              3422 net/core/dev.c 		if (sch_direct_xmit(skb, q, dev, txq, root_lock, true)) {
txq              3703 net/core/dev.c 	struct netdev_queue *txq;
txq              3739 net/core/dev.c 	txq = netdev_core_pick_tx(dev, skb, sb_dev);
txq              3740 net/core/dev.c 	q = rcu_dereference_bh(txq->qdisc);
txq              3744 net/core/dev.c 		rc = __dev_xmit_skb(skb, q, dev, txq);
txq              3763 net/core/dev.c 		if (txq->xmit_lock_owner != cpu) {
txq              3771 net/core/dev.c 			HARD_TX_LOCK(dev, txq, cpu);
txq              3773 net/core/dev.c 			if (!netif_xmit_stopped(txq)) {
txq              3775 net/core/dev.c 				skb = dev_hard_start_xmit(skb, dev, txq, &rc);
txq              3778 net/core/dev.c 					HARD_TX_UNLOCK(dev, txq);
txq              3782 net/core/dev.c 			HARD_TX_UNLOCK(dev, txq);
txq              3822 net/core/dev.c 	struct netdev_queue *txq;
txq              3835 net/core/dev.c 	txq = skb_get_tx_queue(dev, skb);
txq              3839 net/core/dev.c 	HARD_TX_LOCK(dev, txq, smp_processor_id());
txq              3840 net/core/dev.c 	if (!netif_xmit_frozen_or_drv_stopped(txq))
txq              3841 net/core/dev.c 		ret = netdev_start_xmit(skb, dev, txq, false);
txq              3842 net/core/dev.c 	HARD_TX_UNLOCK(dev, txq);
txq              4347 net/core/dev.c 	struct netdev_queue *txq;
txq              4351 net/core/dev.c 	txq = netdev_core_pick_tx(dev, skb, NULL);
txq              4353 net/core/dev.c 	HARD_TX_LOCK(dev, txq, cpu);
txq              4354 net/core/dev.c 	if (!netif_xmit_stopped(txq)) {
txq              4355 net/core/dev.c 		rc = netdev_start_xmit(skb, dev, txq, 0);
txq              4359 net/core/dev.c 	HARD_TX_UNLOCK(dev, txq);
txq              8939 net/core/dev.c 		struct netdev_queue *txq = netdev_get_tx_queue(dev, i);
txq              8941 net/core/dev.c 		netif_tx_stop_queue(txq);
txq              1525 net/core/net-sysfs.c 	int error = 0, txq = 0, rxq = 0, real_rx = 0, real_tx = 0;
txq              1544 net/core/net-sysfs.c 	txq = real_tx;
txq              1549 net/core/net-sysfs.c 	netdev_queue_update_kobjects(dev, txq, 0);
txq                73 net/core/netpoll.c 			      struct netdev_queue *txq)
txq                92 net/core/netpoll.c 	status = netdev_start_xmit(skb, dev, txq, false);
txq               105 net/core/netpoll.c 	while ((skb = skb_dequeue(&npinfo->txq))) {
txq               107 net/core/netpoll.c 		struct netdev_queue *txq;
txq               122 net/core/netpoll.c 		txq = netdev_get_tx_queue(dev, q_index);
txq               123 net/core/netpoll.c 		HARD_TX_LOCK(dev, txq, smp_processor_id());
txq               124 net/core/netpoll.c 		if (netif_xmit_frozen_or_stopped(txq) ||
txq               125 net/core/netpoll.c 		    !dev_xmit_complete(netpoll_start_xmit(skb, dev, txq))) {
txq               126 net/core/netpoll.c 			skb_queue_head(&npinfo->txq, skb);
txq               127 net/core/netpoll.c 			HARD_TX_UNLOCK(dev, txq);
txq               133 net/core/netpoll.c 		HARD_TX_UNLOCK(dev, txq);
txq               324 net/core/netpoll.c 	if (skb_queue_len(&npinfo->txq) == 0 && !netpoll_owner_active(dev)) {
txq               325 net/core/netpoll.c 		struct netdev_queue *txq;
txq               327 net/core/netpoll.c 		txq = netdev_core_pick_tx(dev, skb, NULL);
txq               332 net/core/netpoll.c 			if (HARD_TX_TRYLOCK(dev, txq)) {
txq               333 net/core/netpoll.c 				if (!netif_xmit_stopped(txq))
txq               334 net/core/netpoll.c 					status = netpoll_start_xmit(skb, dev, txq);
txq               336 net/core/netpoll.c 				HARD_TX_UNLOCK(dev, txq);
txq               356 net/core/netpoll.c 		skb_queue_tail(&npinfo->txq, skb);
txq               608 net/core/netpoll.c 		skb_queue_head_init(&npinfo->txq);
txq               781 net/core/netpoll.c 	skb_queue_purge(&npinfo->txq);
txq               787 net/core/netpoll.c 	__skb_queue_purge(&npinfo->txq);
txq              3303 net/core/pktgen.c 	struct netdev_queue *txq;
txq              3400 net/core/pktgen.c 	txq = skb_get_tx_queue(odev, pkt_dev->skb);
txq              3404 net/core/pktgen.c 	HARD_TX_LOCK(odev, txq, smp_processor_id());
txq              3406 net/core/pktgen.c 	if (unlikely(netif_xmit_frozen_or_drv_stopped(txq))) {
txq              3414 net/core/pktgen.c 	ret = netdev_start_xmit(pkt_dev->skb, odev, txq, --burst > 0);
txq              3422 net/core/pktgen.c 		if (burst > 0 && !netif_xmit_frozen_or_drv_stopped(txq))
txq              3443 net/core/pktgen.c 	HARD_TX_UNLOCK(odev, txq);
txq               192 net/mac80211/agg-tx.c 	struct ieee80211_txq *txq = sta->sta.txq[tid];
txq               197 net/mac80211/agg-tx.c 	if (!txq)
txq               200 net/mac80211/agg-tx.c 	txqi = to_txq_info(txq);
txq               201 net/mac80211/agg-tx.c 	sdata = vif_to_sdata(txq->vif);
txq               213 net/mac80211/agg-tx.c 	struct ieee80211_txq *txq = sta->sta.txq[tid];
txq               216 net/mac80211/agg-tx.c 	if (!txq)
txq               219 net/mac80211/agg-tx.c 	txqi = to_txq_info(txq);
txq               918 net/mac80211/agg-tx.c 	struct ieee80211_txq *txq;
txq               928 net/mac80211/agg-tx.c 	txq = sta->sta.txq[tid];
txq               929 net/mac80211/agg-tx.c 	if (!amsdu && txq)
txq               930 net/mac80211/agg-tx.c 		set_bit(IEEE80211_TXQ_NO_AMSDU, &to_txq_info(txq)->flags);
txq              3908 net/mac80211/cfg.c 		if (!sdata->vif.txq) {
txq              3912 net/mac80211/cfg.c 		ieee80211_fill_txq_stats(txqstats, to_txq_info(sdata->vif.txq));
txq               493 net/mac80211/debugfs_netdev.c 	if (!sdata->vif.txq)
txq               496 net/mac80211/debugfs_netdev.c 	txqi = to_txq_info(sdata->vif.txq);
txq               505 net/mac80211/debugfs_netdev.c 			txqi->txq.ac,
txq               164 net/mac80211/debugfs_sta.c 	for (i = 0; i < ARRAY_SIZE(sta->sta.txq); i++) {
txq               165 net/mac80211/debugfs_sta.c 		if (!sta->sta.txq[i])
txq               167 net/mac80211/debugfs_sta.c 		txqi = to_txq_info(sta->sta.txq[i]);
txq               170 net/mac80211/debugfs_sta.c 			       txqi->txq.tid,
txq               171 net/mac80211/debugfs_sta.c 			       txqi->txq.ac,
txq              1201 net/mac80211/driver-ops.h 				     struct txq_info *txq)
txq              1203 net/mac80211/driver-ops.h 	struct ieee80211_sub_if_data *sdata = vif_to_sdata(txq->txq.vif);
txq              1211 net/mac80211/driver-ops.h 	trace_drv_wake_tx_queue(local, sdata, txq);
txq              1212 net/mac80211/driver-ops.h 	local->ops->wake_tx_queue(&local->hw, &txq->txq);
txq              1218 net/mac80211/driver-ops.h 	ieee80211_schedule_txq(&local->hw, &txqi->txq);
txq               851 net/mac80211/ieee80211_i.h 	struct ieee80211_txq txq;
txq              1541 net/mac80211/ieee80211_i.h static inline struct txq_info *to_txq_info(struct ieee80211_txq *txq)
txq              1543 net/mac80211/ieee80211_i.h 	return container_of(txq, struct txq_info, txq);
txq              1546 net/mac80211/ieee80211_i.h static inline bool txq_has_queue(struct ieee80211_txq *txq)
txq              1548 net/mac80211/ieee80211_i.h 	struct txq_info *txqi = to_txq_info(txq);
txq              2088 net/mac80211/ieee80211_i.h 			struct txq_info *txq, int tid);
txq              1906 net/mac80211/iface.c 	if (sdata->vif.txq)
txq              1907 net/mac80211/iface.c 		ieee80211_txq_purge(sdata->local, to_txq_info(sdata->vif.txq));
txq              1583 net/mac80211/rx.c 	if (!sta->sta.txq[0])
txq              1587 net/mac80211/rx.c 		struct ieee80211_txq *txq = sta->sta.txq[tid];
txq              1588 net/mac80211/rx.c 		struct txq_info *txqi = to_txq_info(txq);
txq              1590 net/mac80211/rx.c 		spin_lock(&local->active_txq_lock[txq->ac]);
txq              1593 net/mac80211/rx.c 		spin_unlock(&local->active_txq_lock[txq->ac]);
txq              1595 net/mac80211/rx.c 		if (txq_has_queue(txq))
txq               110 net/mac80211/sta_info.c 	if (sta->sta.txq[0]) {
txq               111 net/mac80211/sta_info.c 		for (i = 0; i < ARRAY_SIZE(sta->sta.txq); i++) {
txq               114 net/mac80211/sta_info.c 			if (!sta->sta.txq[i])
txq               117 net/mac80211/sta_info.c 			txqi = to_txq_info(sta->sta.txq[i]);
txq               252 net/mac80211/sta_info.c 	if (sta->sta.txq[0])
txq               253 net/mac80211/sta_info.c 		kfree(to_txq_info(sta->sta.txq[0]));
txq               379 net/mac80211/sta_info.c 		txq_data = kcalloc(ARRAY_SIZE(sta->sta.txq), size, gfp);
txq               383 net/mac80211/sta_info.c 		for (i = 0; i < ARRAY_SIZE(sta->sta.txq); i++) {
txq               384 net/mac80211/sta_info.c 			struct txq_info *txq = txq_data + i * size;
txq               387 net/mac80211/sta_info.c 			ieee80211_txq_init(sdata, sta, txq, i);
txq               489 net/mac80211/sta_info.c 	if (sta->sta.txq[0])
txq               490 net/mac80211/sta_info.c 		kfree(to_txq_info(sta->sta.txq[0]));
txq              1300 net/mac80211/sta_info.c 	for (i = 0; i < ARRAY_SIZE(sta->sta.txq); i++) {
txq              1301 net/mac80211/sta_info.c 		if (!sta->sta.txq[i] || !txq_has_queue(sta->sta.txq[i]))
txq              1304 net/mac80211/sta_info.c 		schedule_and_wake_txq(local, to_txq_info(sta->sta.txq[i]));
txq              1738 net/mac80211/sta_info.c 		if (!sta->sta.txq[0])
txq              1741 net/mac80211/sta_info.c 		for (tid = 0; tid < ARRAY_SIZE(sta->sta.txq); tid++) {
txq              1742 net/mac80211/sta_info.c 			if (!sta->sta.txq[tid] ||
txq              1744 net/mac80211/sta_info.c 			    txq_has_queue(sta->sta.txq[tid]))
txq              2158 net/mac80211/sta_info.c 					 to_txq_info(sta->sta.txq[tid]));
txq               276 net/mac80211/tdls.c 	struct ieee80211_tx_queue_params *txq;
txq               297 net/mac80211/tdls.c 		txq = &sdata->tx_conf[ieee80211_ac_from_wmm(i)];
txq               298 net/mac80211/tdls.c 		wmm->ac[i].aci_aifsn = ieee80211_wmm_aci_aifsn(txq->aifs,
txq               299 net/mac80211/tdls.c 							       txq->acm, i);
txq               300 net/mac80211/tdls.c 		wmm->ac[i].cw = ieee80211_wmm_ecw(txq->cw_min, txq->cw_max);
txq               301 net/mac80211/tdls.c 		wmm->ac[i].txop_limit = cpu_to_le16(txq->txop);
txq              2660 net/mac80211/trace.h 		 struct txq_info *txq),
txq              2662 net/mac80211/trace.h 	TP_ARGS(local, sdata, txq),
txq              2673 net/mac80211/trace.h 		struct ieee80211_sta *sta = txq->txq.sta;
txq              2678 net/mac80211/trace.h 		__entry->ac = txq->txq.ac;
txq              2679 net/mac80211/trace.h 		__entry->tid = txq->txq.tid;
txq              1099 net/mac80211/tx.c 	} else if (!tx->sta->sta.txq[tid]) {
txq              1253 net/mac80211/tx.c 	struct ieee80211_txq *txq = NULL;
txq              1268 net/mac80211/tx.c 			txq = sta->sta.txq[IEEE80211_NUM_TIDS];
txq              1276 net/mac80211/tx.c 		txq = sta->sta.txq[tid];
txq              1278 net/mac80211/tx.c 		txq = vif->txq;
txq              1281 net/mac80211/tx.c 	if (!txq)
txq              1284 net/mac80211/tx.c 	return to_txq_info(txq);
txq              1314 net/mac80211/tx.c 	local = vif_to_sdata(txqi->txq.vif)->local;
txq              1333 net/mac80211/tx.c 	local = vif_to_sdata(txqi->txq.vif)->local;
txq              1353 net/mac80211/tx.c 	if (txqi->txq.sta) {
txq              1354 net/mac80211/tx.c 		struct sta_info *sta = container_of(txqi->txq.sta,
txq              1438 net/mac80211/tx.c 	if (!ap->vif.txq)
txq              1441 net/mac80211/tx.c 	txqi = to_txq_info(ap->vif.txq);
txq              1461 net/mac80211/tx.c 	txqi->txq.vif = &sdata->vif;
txq              1464 net/mac80211/tx.c 		sdata->vif.txq = &txqi->txq;
txq              1465 net/mac80211/tx.c 		txqi->txq.tid = 0;
txq              1466 net/mac80211/tx.c 		txqi->txq.ac = IEEE80211_AC_BE;
txq              1482 net/mac80211/tx.c 		txqi->txq.ac = IEEE80211_AC_VO;
txq              1484 net/mac80211/tx.c 		txqi->txq.ac = ieee80211_ac_from_tid(tid);
txq              1487 net/mac80211/tx.c 	txqi->txq.sta = &sta->sta;
txq              1488 net/mac80211/tx.c 	txqi->txq.tid = tid;
txq              1489 net/mac80211/tx.c 	sta->sta.txq[tid] = &txqi->txq;
txq              1503 net/mac80211/tx.c 	spin_lock_bh(&local->active_txq_lock[txqi->txq.ac]);
txq              1505 net/mac80211/tx.c 	spin_unlock_bh(&local->active_txq_lock[txqi->txq.ac]);
txq              3236 net/mac80211/tx.c 	struct ieee80211_txq *txq = sta->sta.txq[tid];
txq              3258 net/mac80211/tx.c 	if (!txq)
txq              3261 net/mac80211/tx.c 	txqi = to_txq_info(txq);
txq              3552 net/mac80211/tx.c 				     struct ieee80211_txq *txq)
txq              3555 net/mac80211/tx.c 	struct txq_info *txqi = container_of(txq, struct txq_info, txq);
txq              3563 net/mac80211/tx.c 	struct ieee80211_vif *vif = txq->vif;
txq              3574 net/mac80211/tx.c 	if (vif->txqs_stopped[ieee80211_ac_from_tid(txq->tid)]) {
txq              3599 net/mac80211/tx.c 	if (txq->sta) {
txq              3600 net/mac80211/tx.c 		tx.sta = container_of(txq->sta, struct sta_info, sta);
txq              3636 net/mac80211/tx.c 		struct sta_info *sta = container_of(txq->sta, struct sta_info,
txq              3719 net/mac80211/tx.c 	if (txqi->txq.sta) {
txq              3720 net/mac80211/tx.c 		struct sta_info *sta = container_of(txqi->txq.sta,
txq              3723 net/mac80211/tx.c 		if (sta->airtime[txqi->txq.ac].deficit < 0) {
txq              3724 net/mac80211/tx.c 			sta->airtime[txqi->txq.ac].deficit +=
txq              3727 net/mac80211/tx.c 				       &local->active_txqs[txqi->txq.ac]);
txq              3738 net/mac80211/tx.c 	ret = &txqi->txq;
txq              3747 net/mac80211/tx.c 			      struct ieee80211_txq *txq,
txq              3751 net/mac80211/tx.c 	struct txq_info *txqi = to_txq_info(txq);
txq              3753 net/mac80211/tx.c 	spin_lock_bh(&local->active_txq_lock[txq->ac]);
txq              3765 net/mac80211/tx.c 		if (txqi->txq.sta &&
txq              3769 net/mac80211/tx.c 				 &local->active_txqs[txq->ac]);
txq              3772 net/mac80211/tx.c 				      &local->active_txqs[txq->ac]);
txq              3775 net/mac80211/tx.c 	spin_unlock_bh(&local->active_txq_lock[txq->ac]);
txq              3780 net/mac80211/tx.c 				struct ieee80211_txq *txq)
txq              3783 net/mac80211/tx.c 	struct txq_info *iter, *tmp, *txqi = to_txq_info(txq);
txq              3785 net/mac80211/tx.c 	u8 ac = txq->ac;
txq              3789 net/mac80211/tx.c 	if (!txqi->txq.sta)
txq              3800 net/mac80211/tx.c 		if (!iter->txq.sta) {
txq              3805 net/mac80211/tx.c 		sta = container_of(iter->txq.sta, struct sta_info, sta);
txq              3811 net/mac80211/tx.c 	sta = container_of(txqi->txq.sta, struct sta_info, sta);
txq               262 net/mac80211/util.c 		for (i = 0; i < ARRAY_SIZE(sta->sta.txq); i++) {
txq               263 net/mac80211/util.c 			struct ieee80211_txq *txq = sta->sta.txq[i];
txq               265 net/mac80211/util.c 			if (!txq)
txq               268 net/mac80211/util.c 			txqi = to_txq_info(txq);
txq               270 net/mac80211/util.c 			if (ac != txq->ac)
txq               283 net/mac80211/util.c 	if (!vif->txq)
txq               286 net/mac80211/util.c 	txqi = to_txq_info(vif->txq);
txq               289 net/mac80211/util.c 	    (ps && atomic_read(&ps->num_sta_ps)) || ac != vif->txq->ac)
txq              3983 net/mac80211/util.c void ieee80211_txq_get_depth(struct ieee80211_txq *txq,
txq              3987 net/mac80211/util.c 	struct txq_info *txqi = to_txq_info(txq);
txq                53 net/sched/sch_generic.c 	const struct netdev_queue *txq = q->dev_queue;
txq                65 net/sched/sch_generic.c 		txq = skb_get_tx_queue(txq->dev, skb);
txq                66 net/sched/sch_generic.c 		if (!netif_xmit_frozen_or_stopped(txq)) {
txq               154 net/sched/sch_generic.c 				 const struct netdev_queue *txq,
txq               157 net/sched/sch_generic.c 	int bytelimit = qdisc_avail_bulklimit(txq) - skb->len;
txq               205 net/sched/sch_generic.c 	const struct netdev_queue *txq = q->dev_queue;
txq               233 net/sched/sch_generic.c 		txq = skb_get_tx_queue(txq->dev, skb);
txq               234 net/sched/sch_generic.c 		if (!netif_xmit_frozen_or_stopped(txq)) {
txq               254 net/sched/sch_generic.c 	    netif_xmit_frozen_or_stopped(txq))
txq               267 net/sched/sch_generic.c 			try_bulk_dequeue_skb(q, skb, txq, packets);
txq               272 net/sched/sch_generic.c 	trace_qdisc_dequeue(q, txq, *packets, skb);
txq               286 net/sched/sch_generic.c 		     struct net_device *dev, struct netdev_queue *txq,
txq               311 net/sched/sch_generic.c 		HARD_TX_LOCK(dev, txq, smp_processor_id());
txq               312 net/sched/sch_generic.c 		if (!netif_xmit_frozen_or_stopped(txq))
txq               313 net/sched/sch_generic.c 			skb = dev_hard_start_xmit(skb, dev, txq, &ret);
txq               315 net/sched/sch_generic.c 		HARD_TX_UNLOCK(dev, txq);
txq               360 net/sched/sch_generic.c 	struct netdev_queue *txq;
txq               374 net/sched/sch_generic.c 	txq = skb_get_tx_queue(dev, skb);
txq               376 net/sched/sch_generic.c 	return sch_direct_xmit(skb, q, dev, txq, root_lock, validate);
txq               432 net/sched/sch_generic.c 				struct netdev_queue *txq;
txq               434 net/sched/sch_generic.c 				txq = netdev_get_tx_queue(dev, i);
txq               435 net/sched/sch_generic.c 				trans_start = txq->trans_start;
txq               436 net/sched/sch_generic.c 				if (netif_xmit_stopped(txq) &&
txq               440 net/sched/sch_generic.c 					txq->trans_timeout++;
txq              1056 net/sched/sch_generic.c 	struct netdev_queue *txq;
txq              1059 net/sched/sch_generic.c 	txq = netdev_get_tx_queue(dev, 0);
txq              1064 net/sched/sch_generic.c 		dev->qdisc = txq->qdisc_sleeping;
txq              1067 net/sched/sch_generic.c 		qdisc = qdisc_create_dflt(txq, &mq_qdisc_ops, TC_H_ROOT, NULL);
txq               146 net/sched/sch_teql.c 						struct netdev_queue *txq;
txq               149 net/sched/sch_teql.c 						txq = netdev_get_tx_queue(master->dev, 0);
txq               152 net/sched/sch_teql.c 						root_lock = qdisc_root_sleeping_lock(rtnl_dereference(txq->qdisc));
txq               154 net/sched/sch_teql.c 						qdisc_reset(rtnl_dereference(txq->qdisc));
txq               217 net/sched/sch_teql.c 	       struct net_device *dev, struct netdev_queue *txq,
txq               257 net/sched/sch_teql.c 			       struct netdev_queue *txq)
txq               262 net/sched/sch_teql.c 	if (rcu_access_pointer(txq->qdisc) == &noop_qdisc)
txq               269 net/sched/sch_teql.c 	res = __teql_resolve(skb, skb_res, dev, txq, dst);
txq               295 net/xfrm/xfrm_device.c 	struct netdev_queue *txq;
txq               300 net/xfrm/xfrm_device.c 	txq = netdev_core_pick_tx(dev, skb, NULL);
txq               302 net/xfrm/xfrm_device.c 	HARD_TX_LOCK(dev, txq, smp_processor_id());
txq               303 net/xfrm/xfrm_device.c 	if (!netif_xmit_frozen_or_stopped(txq))
txq               304 net/xfrm/xfrm_device.c 		skb = dev_hard_start_xmit(skb, dev, txq, &ret);
txq               305 net/xfrm/xfrm_device.c 	HARD_TX_UNLOCK(dev, txq);