txq 613 drivers/atm/ambassador.c amb_txq * txq = &dev->txq; txq 621 drivers/atm/ambassador.c spin_lock_irqsave (&txq->lock, flags); txq 623 drivers/atm/ambassador.c if (txq->pending < txq->maximum) { txq 624 drivers/atm/ambassador.c PRINTD (DBG_TX, "TX in slot %p", txq->in.ptr); txq 626 drivers/atm/ambassador.c *txq->in.ptr = *tx; txq 627 drivers/atm/ambassador.c txq->pending++; txq 628 drivers/atm/ambassador.c txq->in.ptr = NEXTQ (txq->in.ptr, txq->in.start, txq->in.limit); txq 630 drivers/atm/ambassador.c wr_mem (dev, offsetof(amb_mem, mb.adapter.tx_address), virt_to_bus (txq->in.ptr)); txq 633 drivers/atm/ambassador.c if (txq->pending > txq->high) txq 634 drivers/atm/ambassador.c txq->high = txq->pending; txq 635 drivers/atm/ambassador.c spin_unlock_irqrestore (&txq->lock, flags); txq 638 drivers/atm/ambassador.c txq->filled++; txq 639 drivers/atm/ambassador.c spin_unlock_irqrestore (&txq->lock, flags); txq 645 drivers/atm/ambassador.c amb_txq * txq = &dev->txq; txq 650 drivers/atm/ambassador.c spin_lock_irqsave (&txq->lock, flags); txq 652 drivers/atm/ambassador.c if (txq->pending && txq->out.ptr->handle) { txq 654 drivers/atm/ambassador.c tx_complete (dev, txq->out.ptr); txq 656 drivers/atm/ambassador.c txq->out.ptr->handle = 0; txq 658 drivers/atm/ambassador.c txq->pending--; txq 659 drivers/atm/ambassador.c txq->out.ptr = NEXTQ (txq->out.ptr, txq->out.start, txq->out.limit); txq 661 drivers/atm/ambassador.c spin_unlock_irqrestore (&txq->lock, flags); txq 665 drivers/atm/ambassador.c spin_unlock_irqrestore (&txq->lock, flags); txq 1435 drivers/atm/ambassador.c amb_txq * t = &dev->txq; txq 1549 drivers/atm/ambassador.c amb_txq * txq = &dev->txq; txq 1551 drivers/atm/ambassador.c txq->pending = 0; txq 1552 drivers/atm/ambassador.c txq->high = 0; txq 1553 drivers/atm/ambassador.c txq->filled = 0; txq 1554 drivers/atm/ambassador.c txq->maximum = txs - 1; txq 1556 drivers/atm/ambassador.c txq->in.start = in; txq 1557 drivers/atm/ambassador.c txq->in.ptr = in; txq 1558 drivers/atm/ambassador.c txq->in.limit = in + txs; txq 1560 drivers/atm/ambassador.c memory = txq->in.limit; txq 1563 drivers/atm/ambassador.c txq->out.start = out; txq 1564 drivers/atm/ambassador.c txq->out.ptr = out; txq 1565 drivers/atm/ambassador.c txq->out.limit = out + txs; txq 1567 drivers/atm/ambassador.c memory = txq->out.limit; txq 1981 drivers/atm/ambassador.c a.tx_start = bus_addr (dev->txq.in.start); txq 1982 drivers/atm/ambassador.c a.tx_end = bus_addr (dev->txq.in.limit); txq 1983 drivers/atm/ambassador.c a.txcom_start = bus_addr (dev->txq.out.start); txq 1984 drivers/atm/ambassador.c a.txcom_end = bus_addr (dev->txq.out.limit); txq 2153 drivers/atm/ambassador.c spin_lock_init (&dev->txq.lock); txq 619 drivers/atm/ambassador.h amb_txq txq; txq 1393 drivers/atm/firestream.c static int init_q(struct fs_dev *dev, struct queue *txq, int queue, txq 1420 drivers/atm/firestream.c txq->sa = p; txq 1421 drivers/atm/firestream.c txq->ea = p; txq 1422 drivers/atm/firestream.c txq->offset = queue; txq 1523 drivers/atm/firestream.c static void free_queue(struct fs_dev *dev, struct queue *txq) txq 1527 drivers/atm/firestream.c write_fs (dev, Q_SA(txq->offset), 0); txq 1528 drivers/atm/firestream.c write_fs (dev, Q_EA(txq->offset), 0); txq 1529 drivers/atm/firestream.c write_fs (dev, Q_RP(txq->offset), 0); txq 1530 drivers/atm/firestream.c write_fs (dev, Q_WP(txq->offset), 0); txq 1533 drivers/atm/firestream.c fs_dprintk (FS_DEBUG_ALLOC, "Free queue: %p\n", txq->sa); txq 1534 drivers/atm/firestream.c kfree (txq->sa); txq 772 drivers/atm/fore200e.c struct host_txq* txq = &fore200e->host_txq; txq 782 drivers/atm/fore200e.c entry = &txq->host_entry[ txq->tail ]; txq 789 drivers/atm/fore200e.c entry, txq->tail, entry->vc_map, entry->skb); txq 856 drivers/atm/fore200e.c FORE200E_NEXT_ENTRY(txq->tail, QUEUE_SIZE_TX); txq 1472 drivers/atm/fore200e.c struct host_txq* txq; txq 1494 drivers/atm/fore200e.c txq = &fore200e->host_txq; txq 1561 drivers/atm/fore200e.c entry = &txq->host_entry[ txq->head ]; txq 1563 drivers/atm/fore200e.c if ((*entry->status != STATUS_FREE) || (txq->txing >= QUEUE_SIZE_TX - 2)) { txq 1613 drivers/atm/fore200e.c FORE200E_NEXT_ENTRY(txq->head, QUEUE_SIZE_TX); txq 1614 drivers/atm/fore200e.c txq->txing++; txq 2197 drivers/atm/fore200e.c struct host_txq* txq = &fore200e->host_txq; txq 2205 drivers/atm/fore200e.c &txq->status, txq 2214 drivers/atm/fore200e.c &txq->tpd, txq 2219 drivers/atm/fore200e.c fore200e_dma_chunk_free(fore200e, &txq->status); txq 2229 drivers/atm/fore200e.c txq->host_entry[ i ].status = txq 2230 drivers/atm/fore200e.c FORE200E_INDEX(txq->status.align_addr, enum status, i); txq 2231 drivers/atm/fore200e.c txq->host_entry[ i ].tpd = txq 2232 drivers/atm/fore200e.c FORE200E_INDEX(txq->tpd.align_addr, struct tpd, i); txq 2233 drivers/atm/fore200e.c txq->host_entry[ i ].tpd_dma = txq 2234 drivers/atm/fore200e.c FORE200E_DMA_INDEX(txq->tpd.dma_addr, struct tpd, i); txq 2235 drivers/atm/fore200e.c txq->host_entry[ i ].cp_entry = &cp_entry[ i ]; txq 2237 drivers/atm/fore200e.c *txq->host_entry[ i ].status = STATUS_FREE; txq 2239 drivers/atm/fore200e.c fore200e->bus->write(FORE200E_DMA_INDEX(txq->status.dma_addr, enum status, i), txq 2249 drivers/atm/fore200e.c txq->head = 0; txq 2250 drivers/atm/fore200e.c txq->tail = 0; txq 72 drivers/bluetooth/bluecard_cs.c struct sk_buff_head txq; txq 259 drivers/bluetooth/bluecard_cs.c skb = skb_dequeue(&(info->txq)); txq 324 drivers/bluetooth/bluecard_cs.c skb_queue_head(&(info->txq), skb); txq 597 drivers/bluetooth/bluecard_cs.c skb_queue_tail(&(info->txq), skb); txq 614 drivers/bluetooth/bluecard_cs.c skb_queue_purge(&(info->txq)); txq 670 drivers/bluetooth/bluecard_cs.c skb_queue_tail(&(info->txq), skb); txq 692 drivers/bluetooth/bluecard_cs.c skb_queue_head_init(&(info->txq)); txq 775 drivers/bluetooth/bluecard_cs.c skb_queue_purge(&(info->txq)); txq 77 drivers/bluetooth/bt3c_cs.c struct sk_buff_head txq; txq 196 drivers/bluetooth/bt3c_cs.c skb = skb_dequeue(&(info->txq)); txq 390 drivers/bluetooth/bt3c_cs.c skb_queue_purge(&(info->txq)); txq 429 drivers/bluetooth/bt3c_cs.c skb_queue_tail(&(info->txq), skb); txq 546 drivers/bluetooth/bt3c_cs.c skb_queue_head_init(&(info->txq)); txq 162 drivers/bluetooth/btmtksdio.c struct sk_buff_head txq; txq 314 drivers/bluetooth/btmtksdio.c while ((skb = skb_dequeue(&bdev->txq))) { txq 318 drivers/bluetooth/btmtksdio.c skb_queue_head(&bdev->txq, skb); txq 651 drivers/bluetooth/btmtksdio.c skb_queue_purge(&bdev->txq); txq 918 drivers/bluetooth/btmtksdio.c skb_queue_tail(&bdev->txq, skb); txq 944 drivers/bluetooth/btmtksdio.c skb_queue_head_init(&bdev->txq); txq 132 drivers/bluetooth/btmtkuart.c struct sk_buff_head txq; txq 379 drivers/bluetooth/btmtkuart.c struct sk_buff *skb = skb_dequeue(&bdev->txq); txq 391 drivers/bluetooth/btmtkuart.c skb_queue_head(&bdev->txq, skb); txq 627 drivers/bluetooth/btmtkuart.c skb_queue_purge(&bdev->txq); txq 892 drivers/bluetooth/btmtkuart.c skb_queue_tail(&bdev->txq, skb); txq 990 drivers/bluetooth/btmtkuart.c skb_queue_head_init(&bdev->txq); txq 49 drivers/bluetooth/btsdio.c struct sk_buff_head txq; txq 101 drivers/bluetooth/btsdio.c while ((skb = skb_dequeue(&data->txq))) { txq 105 drivers/bluetooth/btsdio.c skb_queue_head(&data->txq, skb); txq 231 drivers/bluetooth/btsdio.c skb_queue_purge(&data->txq); txq 259 drivers/bluetooth/btsdio.c skb_queue_tail(&data->txq, skb); txq 302 drivers/bluetooth/btsdio.c skb_queue_head_init(&data->txq); txq 75 drivers/bluetooth/dtl1_cs.c struct sk_buff_head txq; txq 156 drivers/bluetooth/dtl1_cs.c skb = skb_dequeue(&(info->txq)); txq 168 drivers/bluetooth/dtl1_cs.c skb_queue_head(&(info->txq), skb); txq 370 drivers/bluetooth/dtl1_cs.c skb_queue_purge(&(info->txq)); txq 421 drivers/bluetooth/dtl1_cs.c skb_queue_tail(&(info->txq), s); txq 443 drivers/bluetooth/dtl1_cs.c skb_queue_head_init(&(info->txq)); txq 24 drivers/bluetooth/hci_ag6xx.c struct sk_buff_head txq; txq 43 drivers/bluetooth/hci_ag6xx.c skb_queue_head_init(&ag6xx->txq); txq 55 drivers/bluetooth/hci_ag6xx.c skb_queue_purge(&ag6xx->txq); txq 69 drivers/bluetooth/hci_ag6xx.c skb_queue_purge(&ag6xx->txq); txq 78 drivers/bluetooth/hci_ag6xx.c skb = skb_dequeue(&ag6xx->txq); txq 91 drivers/bluetooth/hci_ag6xx.c skb_queue_tail(&ag6xx->txq, skb); txq 35 drivers/bluetooth/hci_ath.c struct sk_buff_head txq; txq 108 drivers/bluetooth/hci_ath.c skb_queue_head_init(&ath->txq); txq 124 drivers/bluetooth/hci_ath.c skb_queue_purge(&ath->txq); txq 142 drivers/bluetooth/hci_ath.c skb_queue_purge(&ath->txq); txq 232 drivers/bluetooth/hci_ath.c skb_queue_tail(&ath->txq, skb); txq 244 drivers/bluetooth/hci_ath.c return skb_dequeue(&ath->txq); txq 120 drivers/bluetooth/hci_bcm.c struct sk_buff_head txq; txq 395 drivers/bluetooth/hci_bcm.c skb_queue_tail(&bcm->txq, skb); txq 416 drivers/bluetooth/hci_bcm.c skb_queue_head_init(&bcm->txq); txq 504 drivers/bluetooth/hci_bcm.c skb_queue_purge(&bcm->txq); txq 518 drivers/bluetooth/hci_bcm.c skb_queue_purge(&bcm->txq); txq 664 drivers/bluetooth/hci_bcm.c skb_queue_tail(&bcm->txq, skb); txq 683 drivers/bluetooth/hci_bcm.c skb = skb_dequeue(&bcm->txq); txq 37 drivers/bluetooth/hci_h4.c struct sk_buff_head txq; txq 51 drivers/bluetooth/hci_h4.c skb_queue_head_init(&h4->txq); txq 64 drivers/bluetooth/hci_h4.c skb_queue_purge(&h4->txq); txq 78 drivers/bluetooth/hci_h4.c skb_queue_purge(&h4->txq); txq 97 drivers/bluetooth/hci_h4.c skb_queue_tail(&h4->txq, skb); txq 131 drivers/bluetooth/hci_h4.c return skb_dequeue(&h4->txq); txq 69 drivers/bluetooth/hci_intel.c struct sk_buff_head txq; txq 179 drivers/bluetooth/hci_intel.c skb_queue_head(&intel->txq, skb); txq 223 drivers/bluetooth/hci_intel.c skb_queue_head(&intel->txq, skb); txq 262 drivers/bluetooth/hci_intel.c skb_queue_head(&intel->txq, skb); txq 401 drivers/bluetooth/hci_intel.c skb_queue_head_init(&intel->txq); txq 424 drivers/bluetooth/hci_intel.c skb_queue_purge(&intel->txq); txq 438 drivers/bluetooth/hci_intel.c skb_queue_purge(&intel->txq); txq 517 drivers/bluetooth/hci_intel.c skb_queue_tail(&intel->txq, skb); txq 1026 drivers/bluetooth/hci_intel.c skb_queue_tail(&intel->txq, skb); txq 1036 drivers/bluetooth/hci_intel.c skb = skb_dequeue(&intel->txq); txq 75 drivers/bluetooth/hci_ll.c struct sk_buff_head txq; txq 105 drivers/bluetooth/hci_ll.c skb_queue_tail(&ll->txq, skb); txq 121 drivers/bluetooth/hci_ll.c skb_queue_head_init(&ll->txq); txq 147 drivers/bluetooth/hci_ll.c skb_queue_purge(&ll->txq); txq 160 drivers/bluetooth/hci_ll.c skb_queue_purge(&ll->txq); txq 191 drivers/bluetooth/hci_ll.c skb_queue_tail(&ll->txq, skb); txq 327 drivers/bluetooth/hci_ll.c skb_queue_tail(&ll->txq, skb); txq 448 drivers/bluetooth/hci_ll.c return skb_dequeue(&ll->txq); txq 38 drivers/bluetooth/hci_mrvl.c struct sk_buff_head txq; txq 69 drivers/bluetooth/hci_mrvl.c skb_queue_head_init(&mrvl->txq); txq 98 drivers/bluetooth/hci_mrvl.c skb_queue_purge(&mrvl->txq); txq 113 drivers/bluetooth/hci_mrvl.c skb_queue_purge(&mrvl->txq); txq 124 drivers/bluetooth/hci_mrvl.c skb = skb_dequeue(&mrvl->txq); txq 140 drivers/bluetooth/hci_mrvl.c skb_queue_tail(&mrvl->txq, skb); txq 157 drivers/bluetooth/hci_mrvl.c skb_queue_tail(&mrvl->txq, skb); txq 135 drivers/bluetooth/hci_nokia.c struct sk_buff_head txq; txq 482 drivers/bluetooth/hci_nokia.c skb_queue_purge(&btdev->txq); txq 496 drivers/bluetooth/hci_nokia.c skb_queue_purge(&btdev->txq); txq 526 drivers/bluetooth/hci_nokia.c skb_queue_tail(&btdev->txq, skb); txq 648 drivers/bluetooth/hci_nokia.c struct sk_buff *result = skb_dequeue(&btdev->txq); txq 741 drivers/bluetooth/hci_nokia.c skb_queue_head_init(&btdev->txq); txq 85 drivers/bluetooth/hci_qca.c struct sk_buff_head txq; txq 301 drivers/bluetooth/hci_qca.c skb_queue_tail(&qca->txq, skb); txq 485 drivers/bluetooth/hci_qca.c skb_queue_head_init(&qca->txq); txq 598 drivers/bluetooth/hci_qca.c skb_queue_purge(&qca->txq); txq 614 drivers/bluetooth/hci_qca.c skb_queue_purge(&qca->txq); txq 742 drivers/bluetooth/hci_qca.c skb_queue_tail(&qca->txq, skb); txq 786 drivers/bluetooth/hci_qca.c skb_queue_tail(&qca->txq, skb); txq 795 drivers/bluetooth/hci_qca.c skb_queue_tail(&qca->txq, skb); txq 959 drivers/bluetooth/hci_qca.c return skb_dequeue(&qca->txq); txq 1020 drivers/bluetooth/hci_qca.c skb_queue_tail(&qca->txq, skb); txq 1025 drivers/bluetooth/hci_qca.c while (!skb_queue_empty(&qca->txq)) txq 585 drivers/crypto/chelsio/chcr_algo.c struct sge_uld_txq *txq; txq 589 drivers/crypto/chelsio/chcr_algo.c txq = &txq_info->uldtxq[idx]; txq 590 drivers/crypto/chelsio/chcr_algo.c spin_lock(&txq->sendq.lock); txq 591 drivers/crypto/chelsio/chcr_algo.c if (txq->full) txq 593 drivers/crypto/chelsio/chcr_algo.c spin_unlock(&txq->sendq.lock); txq 569 drivers/crypto/chelsio/chcr_ipsec.c netif_tx_stop_queue(q->txq); txq 658 drivers/crypto/chelsio/chcr_ipsec.c netif_tx_stop_queue(q->txq); txq 190 drivers/crypto/chelsio/chtls/chtls.h struct sk_buff_head txq; txq 65 drivers/crypto/chelsio/chtls/chtls_cm.c skb_queue_head_init(&csk->txq); txq 145 drivers/crypto/chelsio/chtls/chtls_cm.c while ((skb = __skb_dequeue(&csk->txq))) { txq 2133 drivers/crypto/chelsio/chtls/chtls_cm.c if (skb_queue_len(&csk->txq) && chtls_push_frames(csk, 0)) txq 313 drivers/crypto/chelsio/chtls/chtls_hw.c if (skb_queue_len(&csk->txq)) txq 400 drivers/crypto/chelsio/chtls/chtls_io.c skb_queue_empty(&csk->txq))); txq 536 drivers/crypto/chelsio/chtls/chtls_io.c skb_queue_empty(&csk->txq))); txq 615 drivers/crypto/chelsio/chtls/chtls_io.c while (csk->wr_credits && (skb = skb_peek(&csk->txq)) && txq 617 drivers/crypto/chelsio/chtls/chtls_io.c skb_queue_len(&csk->txq) > 1)) { txq 660 drivers/crypto/chelsio/chtls/chtls_io.c __skb_unlink(skb, &csk->txq); txq 768 drivers/crypto/chelsio/chtls/chtls_io.c int qlen = skb_queue_len(&csk->txq); txq 771 drivers/crypto/chelsio/chtls/chtls_io.c struct sk_buff *skb = skb_peek_tail(&csk->txq); txq 826 drivers/crypto/chelsio/chtls/chtls_io.c __skb_queue_tail(&csk->txq, skb); txq 881 drivers/crypto/chelsio/chtls/chtls_io.c if (skb_queue_len(&csk->txq) == 1) txq 1016 drivers/crypto/chelsio/chtls/chtls_io.c skb = skb_peek_tail(&csk->txq); txq 1186 drivers/crypto/chelsio/chtls/chtls_io.c __skb_unlink(skb, &csk->txq); txq 1224 drivers/crypto/chelsio/chtls/chtls_io.c struct sk_buff *skb = skb_peek_tail(&csk->txq); txq 1435 drivers/crypto/chelsio/chtls/chtls_io.c skb_queue_len(&csk->txq) && txq 1741 drivers/crypto/chelsio/chtls/chtls_io.c skb_queue_len(&csk->txq) && txq 206 drivers/misc/ti-st/st_core.c skb_queue_tail(&st_gdata->txq, waiting_skb); txq 396 drivers/misc/ti-st/st_core.c return skb_dequeue(&st_gdata->txq); txq 418 drivers/misc/ti-st/st_core.c skb_queue_tail(&st_gdata->txq, skb); txq 770 drivers/misc/ti-st/st_core.c skb_queue_purge(&st_gdata->txq); txq 866 drivers/misc/ti-st/st_core.c skb_queue_head_init(&st_gdata->txq); txq 898 drivers/misc/ti-st/st_core.c skb_queue_purge(&st_gdata->txq); txq 4184 drivers/net/bonding/bond_main.c u16 txq = skb_rx_queue_recorded(skb) ? skb_get_rx_queue(skb) : 0; txq 4189 drivers/net/bonding/bond_main.c if (unlikely(txq >= dev->real_num_tx_queues)) { txq 4191 drivers/net/bonding/bond_main.c txq -= dev->real_num_tx_queues; txq 4192 drivers/net/bonding/bond_main.c } while (txq >= dev->real_num_tx_queues); txq 4194 drivers/net/bonding/bond_main.c return txq; txq 546 drivers/net/ethernet/alacritech/slic.h struct slic_tx_queue txq; txq 114 drivers/net/ethernet/alacritech/slicoss.c static unsigned int slic_get_free_tx_descs(struct slic_tx_queue *txq) txq 118 drivers/net/ethernet/alacritech/slicoss.c return slic_get_free_queue_descs(txq->put_idx, txq->done_idx, txq->len); txq 345 drivers/net/ethernet/alacritech/slicoss.c struct slic_tx_queue *txq = &sdev->txq; txq 360 drivers/net/ethernet/alacritech/slicoss.c txq->done_idx = idx; txq 361 drivers/net/ethernet/alacritech/slicoss.c buff = &txq->txbuffs[idx]; txq 388 drivers/net/ethernet/alacritech/slicoss.c (slic_get_free_tx_descs(txq) >= SLIC_MIN_TX_WAKEUP_DESCS)) txq 839 drivers/net/ethernet/alacritech/slicoss.c struct slic_tx_queue *txq = &sdev->txq; txq 845 drivers/net/ethernet/alacritech/slicoss.c txq->len = SLIC_NUM_TX_DESCS; txq 846 drivers/net/ethernet/alacritech/slicoss.c txq->put_idx = 0; txq 847 drivers/net/ethernet/alacritech/slicoss.c txq->done_idx = 0; txq 849 drivers/net/ethernet/alacritech/slicoss.c txq->txbuffs = kcalloc(txq->len, sizeof(*buff), GFP_KERNEL); txq 850 drivers/net/ethernet/alacritech/slicoss.c if (!txq->txbuffs) txq 853 drivers/net/ethernet/alacritech/slicoss.c txq->dma_pool = dma_pool_create("slic_pool", &sdev->pdev->dev, txq 856 drivers/net/ethernet/alacritech/slicoss.c if (!txq->dma_pool) { txq 862 drivers/net/ethernet/alacritech/slicoss.c for (i = 0; i < txq->len; i++) { txq 863 drivers/net/ethernet/alacritech/slicoss.c buff = &txq->txbuffs[i]; txq 864 drivers/net/ethernet/alacritech/slicoss.c desc = dma_pool_zalloc(txq->dma_pool, GFP_KERNEL, txq 884 drivers/net/ethernet/alacritech/slicoss.c buff = &txq->txbuffs[i]; txq 885 drivers/net/ethernet/alacritech/slicoss.c dma_pool_free(txq->dma_pool, buff->desc, buff->desc_paddr); txq 887 drivers/net/ethernet/alacritech/slicoss.c dma_pool_destroy(txq->dma_pool); txq 890 drivers/net/ethernet/alacritech/slicoss.c kfree(txq->txbuffs); txq 897 drivers/net/ethernet/alacritech/slicoss.c struct slic_tx_queue *txq = &sdev->txq; txq 901 drivers/net/ethernet/alacritech/slicoss.c for (i = 0; i < txq->len; i++) { txq 902 drivers/net/ethernet/alacritech/slicoss.c buff = &txq->txbuffs[i]; txq 903 drivers/net/ethernet/alacritech/slicoss.c dma_pool_free(txq->dma_pool, buff->desc, buff->desc_paddr); txq 912 drivers/net/ethernet/alacritech/slicoss.c dma_pool_destroy(txq->dma_pool); txq 914 drivers/net/ethernet/alacritech/slicoss.c kfree(txq->txbuffs); txq 1411 drivers/net/ethernet/alacritech/slicoss.c struct slic_tx_queue *txq = &sdev->txq; txq 1418 drivers/net/ethernet/alacritech/slicoss.c if (unlikely(slic_get_free_tx_descs(txq) < SLIC_MAX_REQ_TX_DESCS)) { txq 1420 drivers/net/ethernet/alacritech/slicoss.c slic_get_free_tx_descs(txq)); txq 1432 drivers/net/ethernet/alacritech/slicoss.c buff = &txq->txbuffs[txq->put_idx]; txq 1443 drivers/net/ethernet/alacritech/slicoss.c txq->put_idx = slic_next_queue_idx(txq->put_idx, txq->len); txq 1451 drivers/net/ethernet/alacritech/slicoss.c if (slic_get_free_tx_descs(txq) < SLIC_MAX_REQ_TX_DESCS) txq 755 drivers/net/ethernet/amazon/ena/ena_netdev.c struct netdev_queue *txq; txq 765 drivers/net/ethernet/amazon/ena/ena_netdev.c txq = netdev_get_tx_queue(tx_ring->netdev, tx_ring->qid); txq 809 drivers/net/ethernet/amazon/ena/ena_netdev.c netdev_tx_completed_queue(txq, tx_pkts, tx_bytes); txq 822 drivers/net/ethernet/amazon/ena/ena_netdev.c if (unlikely(netif_tx_queue_stopped(txq) && above_thresh)) { txq 823 drivers/net/ethernet/amazon/ena/ena_netdev.c __netif_tx_lock(txq, smp_processor_id()); txq 827 drivers/net/ethernet/amazon/ena/ena_netdev.c if (netif_tx_queue_stopped(txq) && above_thresh && txq 829 drivers/net/ethernet/amazon/ena/ena_netdev.c netif_tx_wake_queue(txq); txq 834 drivers/net/ethernet/amazon/ena/ena_netdev.c __netif_tx_unlock(txq); txq 2252 drivers/net/ethernet/amazon/ena/ena_netdev.c struct netdev_queue *txq; txq 2261 drivers/net/ethernet/amazon/ena/ena_netdev.c txq = netdev_get_tx_queue(dev, qid); txq 2320 drivers/net/ethernet/amazon/ena/ena_netdev.c netdev_tx_sent_queue(txq, skb->len); txq 2343 drivers/net/ethernet/amazon/ena/ena_netdev.c netif_tx_stop_queue(txq); txq 2360 drivers/net/ethernet/amazon/ena/ena_netdev.c netif_tx_wake_queue(txq); txq 2367 drivers/net/ethernet/amazon/ena/ena_netdev.c if (netif_xmit_stopped(txq) || !netdev_xmit_more()) { txq 1437 drivers/net/ethernet/amd/xgbe/xgbe-drv.c struct netdev_queue *txq; txq 1468 drivers/net/ethernet/amd/xgbe/xgbe-drv.c txq = netdev_get_tx_queue(netdev, channel->queue_index); txq 1469 drivers/net/ethernet/amd/xgbe/xgbe-drv.c netdev_tx_reset_queue(txq); txq 2019 drivers/net/ethernet/amd/xgbe/xgbe-drv.c struct netdev_queue *txq; txq 2025 drivers/net/ethernet/amd/xgbe/xgbe-drv.c txq = netdev_get_tx_queue(netdev, channel->queue_index); txq 2064 drivers/net/ethernet/amd/xgbe/xgbe-drv.c netdev_tx_sent_queue(txq, packet->tx_bytes); txq 2621 drivers/net/ethernet/amd/xgbe/xgbe-drv.c struct netdev_queue *txq; txq 2637 drivers/net/ethernet/amd/xgbe/xgbe-drv.c txq = netdev_get_tx_queue(netdev, channel->queue_index); txq 2670 drivers/net/ethernet/amd/xgbe/xgbe-drv.c netdev_tx_completed_queue(txq, tx_packets, tx_bytes); txq 2675 drivers/net/ethernet/amd/xgbe/xgbe-drv.c netif_tx_wake_queue(txq); txq 865 drivers/net/ethernet/apm/xgene/xgene_enet_main.c struct netdev_queue *txq; txq 871 drivers/net/ethernet/apm/xgene/xgene_enet_main.c txq = netdev_get_tx_queue(ndev, i); txq 872 drivers/net/ethernet/apm/xgene/xgene_enet_main.c txq->trans_start = jiffies; txq 873 drivers/net/ethernet/apm/xgene/xgene_enet_main.c netif_tx_start_queue(txq); txq 97 drivers/net/ethernet/atheros/alx/alx.h struct alx_tx_queue *txq; txq 378 drivers/net/ethernet/atheros/alx/hw.c u32 rxq, txq, val; txq 383 drivers/net/ethernet/atheros/alx/hw.c txq = alx_read_mem32(hw, ALX_TXQ0); txq 384 drivers/net/ethernet/atheros/alx/hw.c alx_write_mem32(hw, ALX_TXQ0, txq & ~ALX_TXQ0_EN); txq 609 drivers/net/ethernet/atheros/alx/hw.c u32 mac, txq, rxq; txq 613 drivers/net/ethernet/atheros/alx/hw.c txq = alx_read_mem32(hw, ALX_TXQ0); txq 614 drivers/net/ethernet/atheros/alx/hw.c alx_write_mem32(hw, ALX_TXQ0, txq | ALX_TXQ0_EN); txq 54 drivers/net/ethernet/atheros/alx/main.c static void alx_free_txbuf(struct alx_tx_queue *txq, int entry) txq 56 drivers/net/ethernet/atheros/alx/main.c struct alx_buffer *txb = &txq->bufs[entry]; txq 59 drivers/net/ethernet/atheros/alx/main.c dma_unmap_single(txq->dev, txq 150 drivers/net/ethernet/atheros/alx/main.c return alx->qnapi[r_idx]->txq; txq 153 drivers/net/ethernet/atheros/alx/main.c static struct netdev_queue *alx_get_tx_queue(const struct alx_tx_queue *txq) txq 155 drivers/net/ethernet/atheros/alx/main.c return netdev_get_tx_queue(txq->netdev, txq->queue_idx); txq 158 drivers/net/ethernet/atheros/alx/main.c static inline int alx_tpd_avail(struct alx_tx_queue *txq) txq 160 drivers/net/ethernet/atheros/alx/main.c if (txq->write_idx >= txq->read_idx) txq 161 drivers/net/ethernet/atheros/alx/main.c return txq->count + txq->read_idx - txq->write_idx - 1; txq 162 drivers/net/ethernet/atheros/alx/main.c return txq->read_idx - txq->write_idx - 1; txq 165 drivers/net/ethernet/atheros/alx/main.c static bool alx_clean_tx_irq(struct alx_tx_queue *txq) txq 173 drivers/net/ethernet/atheros/alx/main.c alx = netdev_priv(txq->netdev); txq 174 drivers/net/ethernet/atheros/alx/main.c tx_queue = alx_get_tx_queue(txq); txq 176 drivers/net/ethernet/atheros/alx/main.c sw_read_idx = txq->read_idx; txq 177 drivers/net/ethernet/atheros/alx/main.c hw_read_idx = alx_read_mem16(&alx->hw, txq->c_reg); txq 183 drivers/net/ethernet/atheros/alx/main.c skb = txq->bufs[sw_read_idx].skb; txq 190 drivers/net/ethernet/atheros/alx/main.c alx_free_txbuf(txq, sw_read_idx); txq 192 drivers/net/ethernet/atheros/alx/main.c if (++sw_read_idx == txq->count) txq 195 drivers/net/ethernet/atheros/alx/main.c txq->read_idx = sw_read_idx; txq 201 drivers/net/ethernet/atheros/alx/main.c alx_tpd_avail(txq) > txq->count / 4) txq 306 drivers/net/ethernet/atheros/alx/main.c if (np->txq) txq 307 drivers/net/ethernet/atheros/alx/main.c tx_complete = alx_clean_tx_irq(np->txq); txq 460 drivers/net/ethernet/atheros/alx/main.c if (np->txq) { txq 461 drivers/net/ethernet/atheros/alx/main.c np->txq->read_idx = 0; txq 462 drivers/net/ethernet/atheros/alx/main.c np->txq->write_idx = 0; txq 464 drivers/net/ethernet/atheros/alx/main.c txring_header_reg[np->txq->queue_idx], txq 465 drivers/net/ethernet/atheros/alx/main.c np->txq->tpd_dma); txq 489 drivers/net/ethernet/atheros/alx/main.c static void alx_free_txring_buf(struct alx_tx_queue *txq) txq 493 drivers/net/ethernet/atheros/alx/main.c if (!txq->bufs) txq 496 drivers/net/ethernet/atheros/alx/main.c for (i = 0; i < txq->count; i++) txq 497 drivers/net/ethernet/atheros/alx/main.c alx_free_txbuf(txq, i); txq 499 drivers/net/ethernet/atheros/alx/main.c memset(txq->bufs, 0, txq->count * sizeof(struct alx_buffer)); txq 500 drivers/net/ethernet/atheros/alx/main.c memset(txq->tpd, 0, txq->count * sizeof(struct alx_txd)); txq 501 drivers/net/ethernet/atheros/alx/main.c txq->write_idx = 0; txq 502 drivers/net/ethernet/atheros/alx/main.c txq->read_idx = 0; txq 504 drivers/net/ethernet/atheros/alx/main.c netdev_tx_reset_queue(alx_get_tx_queue(txq)); txq 539 drivers/net/ethernet/atheros/alx/main.c if (alx->qnapi[i] && alx->qnapi[i]->txq) txq 540 drivers/net/ethernet/atheros/alx/main.c alx_free_txring_buf(alx->qnapi[i]->txq); txq 617 drivers/net/ethernet/atheros/alx/main.c static int alx_alloc_tx_ring(struct alx_priv *alx, struct alx_tx_queue *txq, txq 620 drivers/net/ethernet/atheros/alx/main.c txq->bufs = kcalloc(txq->count, sizeof(struct alx_buffer), GFP_KERNEL); txq 621 drivers/net/ethernet/atheros/alx/main.c if (!txq->bufs) txq 624 drivers/net/ethernet/atheros/alx/main.c txq->tpd = alx->descmem.virt + offset; txq 625 drivers/net/ethernet/atheros/alx/main.c txq->tpd_dma = alx->descmem.dma + offset; txq 626 drivers/net/ethernet/atheros/alx/main.c offset += sizeof(struct alx_txd) * txq->count; txq 674 drivers/net/ethernet/atheros/alx/main.c offset = alx_alloc_tx_ring(alx, alx->qnapi[i]->txq, offset); txq 697 drivers/net/ethernet/atheros/alx/main.c if (alx->qnapi[i] && alx->qnapi[i]->txq) txq 698 drivers/net/ethernet/atheros/alx/main.c kfree(alx->qnapi[i]->txq->bufs); txq 721 drivers/net/ethernet/atheros/alx/main.c kfree(np->txq); txq 743 drivers/net/ethernet/atheros/alx/main.c struct alx_tx_queue *txq; txq 762 drivers/net/ethernet/atheros/alx/main.c txq = kzalloc(sizeof(*txq), GFP_KERNEL); txq 763 drivers/net/ethernet/atheros/alx/main.c if (!txq) txq 766 drivers/net/ethernet/atheros/alx/main.c np->txq = txq; txq 767 drivers/net/ethernet/atheros/alx/main.c txq->p_reg = tx_pidx_reg[i]; txq 768 drivers/net/ethernet/atheros/alx/main.c txq->c_reg = tx_cidx_reg[i]; txq 769 drivers/net/ethernet/atheros/alx/main.c txq->queue_idx = i; txq 770 drivers/net/ethernet/atheros/alx/main.c txq->count = alx->tx_ringsz; txq 771 drivers/net/ethernet/atheros/alx/main.c txq->netdev = alx->dev; txq 772 drivers/net/ethernet/atheros/alx/main.c txq->dev = &alx->hw.pdev->dev; txq 868 drivers/net/ethernet/atheros/alx/main.c if (np->txq && np->rxq) txq 870 drivers/net/ethernet/atheros/alx/main.c np->txq->queue_idx); txq 871 drivers/net/ethernet/atheros/alx/main.c else if (np->txq) txq 873 drivers/net/ethernet/atheros/alx/main.c np->txq->queue_idx); txq 1436 drivers/net/ethernet/atheros/alx/main.c static int alx_map_tx_skb(struct alx_tx_queue *txq, struct sk_buff *skb) txq 1440 drivers/net/ethernet/atheros/alx/main.c int maplen, f, first_idx = txq->write_idx; txq 1442 drivers/net/ethernet/atheros/alx/main.c first_tpd = &txq->tpd[txq->write_idx]; txq 1446 drivers/net/ethernet/atheros/alx/main.c if (++txq->write_idx == txq->count) txq 1447 drivers/net/ethernet/atheros/alx/main.c txq->write_idx = 0; txq 1449 drivers/net/ethernet/atheros/alx/main.c tpd = &txq->tpd[txq->write_idx]; txq 1456 drivers/net/ethernet/atheros/alx/main.c dma = dma_map_single(txq->dev, skb->data, maplen, txq 1458 drivers/net/ethernet/atheros/alx/main.c if (dma_mapping_error(txq->dev, dma)) txq 1461 drivers/net/ethernet/atheros/alx/main.c dma_unmap_len_set(&txq->bufs[txq->write_idx], size, maplen); txq 1462 drivers/net/ethernet/atheros/alx/main.c dma_unmap_addr_set(&txq->bufs[txq->write_idx], dma, dma); txq 1470 drivers/net/ethernet/atheros/alx/main.c if (++txq->write_idx == txq->count) txq 1471 drivers/net/ethernet/atheros/alx/main.c txq->write_idx = 0; txq 1472 drivers/net/ethernet/atheros/alx/main.c tpd = &txq->tpd[txq->write_idx]; txq 1477 drivers/net/ethernet/atheros/alx/main.c dma = skb_frag_dma_map(txq->dev, frag, 0, txq 1479 drivers/net/ethernet/atheros/alx/main.c if (dma_mapping_error(txq->dev, dma)) txq 1481 drivers/net/ethernet/atheros/alx/main.c dma_unmap_len_set(&txq->bufs[txq->write_idx], size, maplen); txq 1482 drivers/net/ethernet/atheros/alx/main.c dma_unmap_addr_set(&txq->bufs[txq->write_idx], dma, dma); txq 1490 drivers/net/ethernet/atheros/alx/main.c txq->bufs[txq->write_idx].skb = skb; txq 1492 drivers/net/ethernet/atheros/alx/main.c if (++txq->write_idx == txq->count) txq 1493 drivers/net/ethernet/atheros/alx/main.c txq->write_idx = 0; txq 1499 drivers/net/ethernet/atheros/alx/main.c while (f != txq->write_idx) { txq 1500 drivers/net/ethernet/atheros/alx/main.c alx_free_txbuf(txq, f); txq 1501 drivers/net/ethernet/atheros/alx/main.c if (++f == txq->count) txq 1508 drivers/net/ethernet/atheros/alx/main.c struct alx_tx_queue *txq) txq 1514 drivers/net/ethernet/atheros/alx/main.c alx = netdev_priv(txq->netdev); txq 1516 drivers/net/ethernet/atheros/alx/main.c if (alx_tpd_avail(txq) < alx_tpd_req(skb)) { txq 1517 drivers/net/ethernet/atheros/alx/main.c netif_tx_stop_queue(alx_get_tx_queue(txq)); txq 1521 drivers/net/ethernet/atheros/alx/main.c first = &txq->tpd[txq->write_idx]; txq 1530 drivers/net/ethernet/atheros/alx/main.c if (alx_map_tx_skb(txq, skb) < 0) txq 1533 drivers/net/ethernet/atheros/alx/main.c netdev_tx_sent_queue(alx_get_tx_queue(txq), skb->len); txq 1537 drivers/net/ethernet/atheros/alx/main.c alx_write_mem16(&alx->hw, txq->p_reg, txq->write_idx); txq 1539 drivers/net/ethernet/atheros/alx/main.c if (alx_tpd_avail(txq) < txq->count / 8) txq 1540 drivers/net/ethernet/atheros/alx/main.c netif_tx_stop_queue(alx_get_tx_queue(txq)); txq 1190 drivers/net/ethernet/atheros/atl1c/atl1c_main.c u32 mac, txq, rxq; txq 1196 drivers/net/ethernet/atheros/atl1c/atl1c_main.c AT_READ_REG(hw, REG_TXQ_CTRL, &txq); txq 1200 drivers/net/ethernet/atheros/atl1c/atl1c_main.c txq |= TXQ_CTRL_EN; txq 1214 drivers/net/ethernet/atheros/atl1c/atl1c_main.c AT_WRITE_REG(hw, REG_TXQ_CTRL, txq); txq 933 drivers/net/ethernet/broadcom/bcmsysport.c struct netdev_queue *txq; txq 937 drivers/net/ethernet/broadcom/bcmsysport.c txq = netdev_get_tx_queue(priv->netdev, ring->index); txq 942 drivers/net/ethernet/broadcom/bcmsysport.c netif_tx_wake_queue(txq); txq 1281 drivers/net/ethernet/broadcom/bcmsysport.c struct netdev_queue *txq; txq 1290 drivers/net/ethernet/broadcom/bcmsysport.c txq = netdev_get_tx_queue(dev, queue); txq 1296 drivers/net/ethernet/broadcom/bcmsysport.c netif_tx_stop_queue(txq); txq 1347 drivers/net/ethernet/broadcom/bcmsysport.c netif_tx_stop_queue(txq); txq 2424 drivers/net/ethernet/broadcom/bcmsysport.c u32 txq, rxq; txq 2436 drivers/net/ethernet/broadcom/bcmsysport.c if (of_property_read_u32(dn, "systemport,num-txq", &txq)) txq 2437 drivers/net/ethernet/broadcom/bcmsysport.c txq = TDMA_NUM_RINGS; txq 2442 drivers/net/ethernet/broadcom/bcmsysport.c if (!txq || txq > TDMA_NUM_RINGS) txq 2445 drivers/net/ethernet/broadcom/bcmsysport.c dev = alloc_etherdev_mqs(sizeof(*priv), txq, rxq); txq 2453 drivers/net/ethernet/broadcom/bcmsysport.c priv->tx_rings = devm_kcalloc(&pdev->dev, txq, txq 2559 drivers/net/ethernet/broadcom/bcmsysport.c priv->irq0, priv->irq1, txq, rxq); txq 2854 drivers/net/ethernet/broadcom/bnx2.c struct netdev_queue *txq; txq 2857 drivers/net/ethernet/broadcom/bnx2.c txq = netdev_get_tx_queue(bp->dev, index); txq 2919 drivers/net/ethernet/broadcom/bnx2.c netdev_tx_completed_queue(txq, tx_pkt, tx_bytes); txq 2930 drivers/net/ethernet/broadcom/bnx2.c if (unlikely(netif_tx_queue_stopped(txq)) && txq 2932 drivers/net/ethernet/broadcom/bnx2.c __netif_tx_lock(txq, smp_processor_id()); txq 2933 drivers/net/ethernet/broadcom/bnx2.c if ((netif_tx_queue_stopped(txq)) && txq 2935 drivers/net/ethernet/broadcom/bnx2.c netif_tx_wake_queue(txq); txq 2936 drivers/net/ethernet/broadcom/bnx2.c __netif_tx_unlock(txq); txq 6606 drivers/net/ethernet/broadcom/bnx2.c struct netdev_queue *txq; txq 6612 drivers/net/ethernet/broadcom/bnx2.c txq = netdev_get_tx_queue(dev, i); txq 6616 drivers/net/ethernet/broadcom/bnx2.c netif_tx_stop_queue(txq); txq 6716 drivers/net/ethernet/broadcom/bnx2.c netdev_tx_sent_queue(txq, skb->len); txq 6727 drivers/net/ethernet/broadcom/bnx2.c netif_tx_stop_queue(txq); txq 6736 drivers/net/ethernet/broadcom/bnx2.c netif_tx_wake_queue(txq); txq 275 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c struct netdev_queue *txq; txq 284 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c txq = netdev_get_tx_queue(bp->dev, txdata->txq_index); txq 306 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c netdev_tx_completed_queue(txq, pkts_compl, bytes_compl); txq 322 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c if (unlikely(netif_tx_queue_stopped(txq))) { txq 333 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c __netif_tx_lock(txq, smp_processor_id()); txq 335 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c if ((netif_tx_queue_stopped(txq)) && txq 338 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c netif_tx_wake_queue(txq); txq 340 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c __netif_tx_unlock(txq); txq 3737 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c struct netdev_queue *txq; txq 3762 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c txq = netdev_get_tx_queue(dev, txq_index); txq 3790 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c netif_tx_stop_queue(txq); txq 4154 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c netdev_tx_sent_queue(txq, skb->len); txq 4177 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c netif_tx_stop_queue(txq); txq 4186 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c netif_tx_wake_queue(txq); txq 2512 drivers/net/ethernet/broadcom/bnx2x/bnx2x_ethtool.c struct netdev_queue *txq = netdev_get_tx_queue(bp->dev, txq 2579 drivers/net/ethernet/broadcom/bnx2x/bnx2x_ethtool.c netdev_tx_sent_queue(txq, skb->len); txq 654 drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.c req->txq.txq_addr = fp->txdata_ptr[FIRST_TX_COS_INDEX]->tx_desc_mapping; txq 655 drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.c req->txq.vf_sb = fp_idx; txq 656 drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.c req->txq.sb_index = HC_INDEX_ETH_TX_CQ_CONS_COS0; txq 657 drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.c req->txq.hc_rate = bp->tx_ticks ? 1000000/bp->tx_ticks : 0; txq 658 drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.c req->txq.flags = flags; txq 659 drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.c req->txq.traffic_type = LLFC_TRAFFIC_TYPE_NW; txq 1562 drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.c q->sb_idx = setup_q->txq.vf_sb; txq 1565 drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.c init_p->tx.hc_rate = setup_q->txq.hc_rate; txq 1566 drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.c init_p->tx.sb_cq_index = setup_q->txq.sb_index; txq 1568 drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.c bnx2x_vf_mbx_set_q_flags(bp, setup_q->txq.flags, txq 1572 drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.c bnx2x_vf_mbx_set_q_flags(bp, setup_q->txq.flags, txq 1578 drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.c txq_params->dscr_map = setup_q->txq.txq_addr; txq 1579 drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.c txq_params->sb_cq_index = setup_q->txq.sb_index; txq 1580 drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.c txq_params->traffic_type = setup_q->txq.traffic_type; txq 288 drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.h } txq; txq 350 drivers/net/ethernet/broadcom/bnxt/bnxt.c struct netdev_queue *txq; txq 366 drivers/net/ethernet/broadcom/bnxt/bnxt.c txq = netdev_get_tx_queue(dev, i); txq 372 drivers/net/ethernet/broadcom/bnxt/bnxt.c netif_tx_stop_queue(txq); txq 457 drivers/net/ethernet/broadcom/bnxt/bnxt.c netdev_tx_sent_queue(txq, skb->len); txq 565 drivers/net/ethernet/broadcom/bnxt/bnxt.c netdev_tx_sent_queue(txq, skb->len); txq 573 drivers/net/ethernet/broadcom/bnxt/bnxt.c if (!netdev_xmit_more() || netif_xmit_stopped(txq)) txq 582 drivers/net/ethernet/broadcom/bnxt/bnxt.c netif_tx_stop_queue(txq); txq 591 drivers/net/ethernet/broadcom/bnxt/bnxt.c netif_tx_wake_queue(txq); txq 622 drivers/net/ethernet/broadcom/bnxt/bnxt.c struct netdev_queue *txq = netdev_get_tx_queue(bp->dev, txr->txq_index); txq 664 drivers/net/ethernet/broadcom/bnxt/bnxt.c netdev_tx_completed_queue(txq, nr_pkts, tx_bytes); txq 674 drivers/net/ethernet/broadcom/bnxt/bnxt.c if (unlikely(netif_tx_queue_stopped(txq)) && txq 676 drivers/net/ethernet/broadcom/bnxt/bnxt.c __netif_tx_lock(txq, smp_processor_id()); txq 677 drivers/net/ethernet/broadcom/bnxt/bnxt.c if (netif_tx_queue_stopped(txq) && txq 680 drivers/net/ethernet/broadcom/bnxt/bnxt.c netif_tx_wake_queue(txq); txq 681 drivers/net/ethernet/broadcom/bnxt/bnxt.c __netif_tx_unlock(txq); txq 1449 drivers/net/ethernet/broadcom/genet/bcmgenet.c struct netdev_queue *txq; txq 1454 drivers/net/ethernet/broadcom/genet/bcmgenet.c txq = netdev_get_tx_queue(ring->priv->dev, ring->queue); txq 1455 drivers/net/ethernet/broadcom/genet/bcmgenet.c netif_tx_wake_queue(txq); txq 1552 drivers/net/ethernet/broadcom/genet/bcmgenet.c struct netdev_queue *txq; txq 1575 drivers/net/ethernet/broadcom/genet/bcmgenet.c txq = netdev_get_tx_queue(dev, ring->queue); txq 1581 drivers/net/ethernet/broadcom/genet/bcmgenet.c if (!netif_tx_queue_stopped(txq)) { txq 1582 drivers/net/ethernet/broadcom/genet/bcmgenet.c netif_tx_stop_queue(txq); txq 1663 drivers/net/ethernet/broadcom/genet/bcmgenet.c netdev_tx_sent_queue(txq, GENET_CB(skb)->bytes_sent); txq 1666 drivers/net/ethernet/broadcom/genet/bcmgenet.c netif_tx_stop_queue(txq); txq 1668 drivers/net/ethernet/broadcom/genet/bcmgenet.c if (!netdev_xmit_more() || netif_xmit_stopped(txq)) txq 2521 drivers/net/ethernet/broadcom/genet/bcmgenet.c struct netdev_queue *txq; txq 2532 drivers/net/ethernet/broadcom/genet/bcmgenet.c txq = netdev_get_tx_queue(priv->dev, priv->tx_rings[i].queue); txq 2533 drivers/net/ethernet/broadcom/genet/bcmgenet.c netdev_tx_reset_queue(txq); txq 2536 drivers/net/ethernet/broadcom/genet/bcmgenet.c txq = netdev_get_tx_queue(priv->dev, priv->tx_rings[DESC_INDEX].queue); txq 2537 drivers/net/ethernet/broadcom/genet/bcmgenet.c netdev_tx_reset_queue(txq); txq 3019 drivers/net/ethernet/broadcom/genet/bcmgenet.c struct netdev_queue *txq; txq 3026 drivers/net/ethernet/broadcom/genet/bcmgenet.c txq = netdev_get_tx_queue(priv->dev, ring->queue); txq 3038 drivers/net/ethernet/broadcom/genet/bcmgenet.c txq_stopped = netif_tx_queue_stopped(txq); txq 6550 drivers/net/ethernet/broadcom/tg3.c struct netdev_queue *txq; txq 6557 drivers/net/ethernet/broadcom/tg3.c txq = netdev_get_tx_queue(tp->dev, index); txq 6624 drivers/net/ethernet/broadcom/tg3.c netdev_tx_completed_queue(txq, pkts_compl, bytes_compl); txq 6635 drivers/net/ethernet/broadcom/tg3.c if (unlikely(netif_tx_queue_stopped(txq) && txq 6637 drivers/net/ethernet/broadcom/tg3.c __netif_tx_lock(txq, smp_processor_id()); txq 6638 drivers/net/ethernet/broadcom/tg3.c if (netif_tx_queue_stopped(txq) && txq 6640 drivers/net/ethernet/broadcom/tg3.c netif_tx_wake_queue(txq); txq 6641 drivers/net/ethernet/broadcom/tg3.c __netif_tx_unlock(txq); txq 7875 drivers/net/ethernet/broadcom/tg3.c struct netdev_queue *txq, struct sk_buff *skb) txq 7882 drivers/net/ethernet/broadcom/tg3.c netif_tx_stop_queue(txq); txq 7893 drivers/net/ethernet/broadcom/tg3.c netif_tx_wake_queue(txq); txq 7923 drivers/net/ethernet/broadcom/tg3.c struct netdev_queue *txq; txq 7930 drivers/net/ethernet/broadcom/tg3.c txq = netdev_get_tx_queue(dev, skb_get_queue_mapping(skb)); txq 7943 drivers/net/ethernet/broadcom/tg3.c if (!netif_tx_queue_stopped(txq)) { txq 7944 drivers/net/ethernet/broadcom/tg3.c netif_tx_stop_queue(txq); txq 7974 drivers/net/ethernet/broadcom/tg3.c return tg3_tso_bug(tp, tnapi, txq, skb); txq 7982 drivers/net/ethernet/broadcom/tg3.c return tg3_tso_bug(tp, tnapi, txq, skb); txq 8125 drivers/net/ethernet/broadcom/tg3.c return tg3_tso_bug(tp, tnapi, txq, skb); txq 8139 drivers/net/ethernet/broadcom/tg3.c netdev_tx_sent_queue(txq, skb->len); txq 8146 drivers/net/ethernet/broadcom/tg3.c netif_tx_stop_queue(txq); txq 8155 drivers/net/ethernet/broadcom/tg3.c netif_tx_wake_queue(txq); txq 8158 drivers/net/ethernet/broadcom/tg3.c if (!netdev_xmit_more() || netif_xmit_stopped(txq)) { txq 2870 drivers/net/ethernet/brocade/bna/bna_tx_rx.c struct bna_txq *txq; txq 2873 drivers/net/ethernet/brocade/bna/bna_tx_rx.c list_for_each_entry(txq, &tx->txq_q, qe) { txq 2874 drivers/net/ethernet/brocade/bna/bna_tx_rx.c txq->tcb->priority = txq->priority; txq 2876 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_ib_start(tx->bna, &txq->ib, is_regular); txq 3089 drivers/net/ethernet/brocade/bna/bna_tx_rx.c struct bna_txq *txq = NULL; txq 3099 drivers/net/ethernet/brocade/bna/bna_tx_rx.c txq = txq ? list_next_entry(txq, qe) txq 3101 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bfi_enet_datapath_q_init(&cfg_req->q_cfg[i].q.q, &txq->qpt); txq 3102 drivers/net/ethernet/brocade/bna/bna_tx_rx.c cfg_req->q_cfg[i].q.priority = txq->priority; txq 3105 drivers/net/ethernet/brocade/bna/bna_tx_rx.c txq->ib.ib_seg_host_addr.lsb; txq 3107 drivers/net/ethernet/brocade/bna/bna_tx_rx.c txq->ib.ib_seg_host_addr.msb; txq 3109 drivers/net/ethernet/brocade/bna/bna_tx_rx.c htons((u16)txq->ib.intr_vector); txq 3116 drivers/net/ethernet/brocade/bna/bna_tx_rx.c cfg_req->ib_cfg.msix = (txq->ib.intr_type == BNA_INTR_T_MSIX) txq 3119 drivers/net/ethernet/brocade/bna/bna_tx_rx.c htonl((u32)txq->ib.coalescing_timeo); txq 3121 drivers/net/ethernet/brocade/bna/bna_tx_rx.c htonl((u32)txq->ib.interpkt_timeo); txq 3122 drivers/net/ethernet/brocade/bna/bna_tx_rx.c cfg_req->ib_cfg.inter_pkt_count = (u8)txq->ib.interpkt_count; txq 3151 drivers/net/ethernet/brocade/bna/bna_tx_rx.c struct bna_txq *txq; txq 3154 drivers/net/ethernet/brocade/bna/bna_tx_rx.c list_for_each_entry(txq, &tx->txq_q, qe) txq 3155 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_ib_stop(tx->bna, &txq->ib); txq 3161 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_txq_qpt_setup(struct bna_txq *txq, int page_count, int page_size, txq 3171 drivers/net/ethernet/brocade/bna/bna_tx_rx.c txq->qpt.hw_qpt_ptr.lsb = qpt_mem->dma.lsb; txq 3172 drivers/net/ethernet/brocade/bna/bna_tx_rx.c txq->qpt.hw_qpt_ptr.msb = qpt_mem->dma.msb; txq 3173 drivers/net/ethernet/brocade/bna/bna_tx_rx.c txq->qpt.kv_qpt_ptr = qpt_mem->kva; txq 3174 drivers/net/ethernet/brocade/bna/bna_tx_rx.c txq->qpt.page_count = page_count; txq 3175 drivers/net/ethernet/brocade/bna/bna_tx_rx.c txq->qpt.page_size = page_size; txq 3177 drivers/net/ethernet/brocade/bna/bna_tx_rx.c txq->tcb->sw_qpt = (void **) swqpt_mem->kva; txq 3178 drivers/net/ethernet/brocade/bna/bna_tx_rx.c txq->tcb->sw_q = page_mem->kva; txq 3184 drivers/net/ethernet/brocade/bna/bna_tx_rx.c txq->tcb->sw_qpt[i] = kva; txq 3188 drivers/net/ethernet/brocade/bna/bna_tx_rx.c ((struct bna_dma_addr *)txq->qpt.kv_qpt_ptr)[i].lsb = txq 3190 drivers/net/ethernet/brocade/bna/bna_tx_rx.c ((struct bna_dma_addr *)txq->qpt.kv_qpt_ptr)[i].msb = txq 3217 drivers/net/ethernet/brocade/bna/bna_tx_rx.c struct bna_txq *txq; txq 3221 drivers/net/ethernet/brocade/bna/bna_tx_rx.c txq = list_first_entry(&tx->txq_q, struct bna_txq, qe); txq 3222 drivers/net/ethernet/brocade/bna/bna_tx_rx.c txq->tcb = NULL; txq 3223 drivers/net/ethernet/brocade/bna/bna_tx_rx.c txq->tx = NULL; txq 3224 drivers/net/ethernet/brocade/bna/bna_tx_rx.c list_move_tail(&txq->qe, &tx_mod->txq_free_q); txq 3273 drivers/net/ethernet/brocade/bna/bna_tx_rx.c struct bna_txq *txq = NULL; txq 3281 drivers/net/ethernet/brocade/bna/bna_tx_rx.c for (i = 0, txq = list_first_entry(&tx->txq_q, struct bna_txq, qe); txq 3282 drivers/net/ethernet/brocade/bna/bna_tx_rx.c i < tx->num_txq; i++, txq = list_next_entry(txq, qe)) { txq 3284 drivers/net/ethernet/brocade/bna/bna_tx_rx.c txq->tcb->i_dbell->doorbell_addr = txq 3287 drivers/net/ethernet/brocade/bna/bna_tx_rx.c txq->tcb->q_dbell = txq 3290 drivers/net/ethernet/brocade/bna/bna_tx_rx.c txq->hw_id = cfg_rsp->q_handles[i].hw_qid; txq 3293 drivers/net/ethernet/brocade/bna/bna_tx_rx.c (*txq->tcb->hw_consumer_index) = 0; txq 3294 drivers/net/ethernet/brocade/bna/bna_tx_rx.c txq->tcb->producer_index = txq->tcb->consumer_index = 0; txq 3371 drivers/net/ethernet/brocade/bna/bna_tx_rx.c struct bna_txq *txq; txq 3401 drivers/net/ethernet/brocade/bna/bna_tx_rx.c txq = list_first_entry(&tx_mod->txq_free_q, struct bna_txq, qe); txq 3402 drivers/net/ethernet/brocade/bna/bna_tx_rx.c list_move_tail(&txq->qe, &tx->txq_q); txq 3403 drivers/net/ethernet/brocade/bna/bna_tx_rx.c txq->tx = tx; txq 3441 drivers/net/ethernet/brocade/bna/bna_tx_rx.c list_for_each_entry(txq, &tx->txq_q, qe) { txq 3442 drivers/net/ethernet/brocade/bna/bna_tx_rx.c txq->tcb = (struct bna_tcb *) txq 3444 drivers/net/ethernet/brocade/bna/bna_tx_rx.c txq->tx_packets = 0; txq 3445 drivers/net/ethernet/brocade/bna/bna_tx_rx.c txq->tx_bytes = 0; txq 3448 drivers/net/ethernet/brocade/bna/bna_tx_rx.c txq->ib.ib_seg_host_addr.lsb = txq 3450 drivers/net/ethernet/brocade/bna/bna_tx_rx.c txq->ib.ib_seg_host_addr.msb = txq 3452 drivers/net/ethernet/brocade/bna/bna_tx_rx.c txq->ib.ib_seg_host_addr_kva = txq 3454 drivers/net/ethernet/brocade/bna/bna_tx_rx.c txq->ib.intr_type = intr_info->intr_type; txq 3455 drivers/net/ethernet/brocade/bna/bna_tx_rx.c txq->ib.intr_vector = (intr_info->num == 1) ? txq 3459 drivers/net/ethernet/brocade/bna/bna_tx_rx.c txq->ib.intr_vector = BIT(txq->ib.intr_vector); txq 3460 drivers/net/ethernet/brocade/bna/bna_tx_rx.c txq->ib.coalescing_timeo = tx_cfg->coalescing_timeo; txq 3461 drivers/net/ethernet/brocade/bna/bna_tx_rx.c txq->ib.interpkt_timeo = BFI_TX_INTERPKT_TIMEO; txq 3462 drivers/net/ethernet/brocade/bna/bna_tx_rx.c txq->ib.interpkt_count = BFI_TX_INTERPKT_COUNT; txq 3466 drivers/net/ethernet/brocade/bna/bna_tx_rx.c txq->tcb->q_depth = tx_cfg->txq_depth; txq 3467 drivers/net/ethernet/brocade/bna/bna_tx_rx.c txq->tcb->unmap_q = (void *) txq 3469 drivers/net/ethernet/brocade/bna/bna_tx_rx.c txq->tcb->hw_consumer_index = txq 3470 drivers/net/ethernet/brocade/bna/bna_tx_rx.c (u32 *)txq->ib.ib_seg_host_addr_kva; txq 3471 drivers/net/ethernet/brocade/bna/bna_tx_rx.c txq->tcb->i_dbell = &txq->ib.door_bell; txq 3472 drivers/net/ethernet/brocade/bna/bna_tx_rx.c txq->tcb->intr_type = txq->ib.intr_type; txq 3473 drivers/net/ethernet/brocade/bna/bna_tx_rx.c txq->tcb->intr_vector = txq->ib.intr_vector; txq 3474 drivers/net/ethernet/brocade/bna/bna_tx_rx.c txq->tcb->txq = txq; txq 3475 drivers/net/ethernet/brocade/bna/bna_tx_rx.c txq->tcb->bnad = bnad; txq 3476 drivers/net/ethernet/brocade/bna/bna_tx_rx.c txq->tcb->id = i; txq 3479 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_txq_qpt_setup(txq, page_count, PAGE_SIZE, txq 3487 drivers/net/ethernet/brocade/bna/bna_tx_rx.c (tx->tcb_setup_cbfn)(bna->bnad, txq->tcb); txq 3490 drivers/net/ethernet/brocade/bna/bna_tx_rx.c txq->priority = txq->tcb->id; txq 3492 drivers/net/ethernet/brocade/bna/bna_tx_rx.c txq->priority = tx_mod->default_prio; txq 3513 drivers/net/ethernet/brocade/bna/bna_tx_rx.c struct bna_txq *txq; txq 3515 drivers/net/ethernet/brocade/bna/bna_tx_rx.c list_for_each_entry(txq, &tx->txq_q, qe) txq 3517 drivers/net/ethernet/brocade/bna/bna_tx_rx.c (tx->tcb_destroy_cbfn)(tx->bna->bnad, txq->tcb); txq 3587 drivers/net/ethernet/brocade/bna/bna_tx_rx.c tx_mod->txq = (struct bna_txq *) txq 3598 drivers/net/ethernet/brocade/bna/bna_tx_rx.c list_add_tail(&tx_mod->txq[i].qe, &tx_mod->txq_free_q); txq 3663 drivers/net/ethernet/brocade/bna/bna_tx_rx.c struct bna_txq *txq; txq 3665 drivers/net/ethernet/brocade/bna/bna_tx_rx.c list_for_each_entry(txq, &tx->txq_q, qe) txq 3666 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_ib_coalescing_timeo_set(&txq->ib, coalescing_timeo); txq 426 drivers/net/ethernet/brocade/bna/bna_types.h struct bna_txq *txq; txq 515 drivers/net/ethernet/brocade/bna/bna_types.h struct bna_txq *txq; /* BFI_MAX_TXQ entries */ txq 197 drivers/net/ethernet/brocade/bna/bnad.c tcb->txq->tx_packets += sent_packets; txq 198 drivers/net/ethernet/brocade/bna/bnad.c tcb->txq->tx_bytes += sent_bytes; txq 1002 drivers/net/ethernet/brocade/bna/bnad.c (struct bnad_tx_info *)tcb->txq->tx->priv; txq 1012 drivers/net/ethernet/brocade/bna/bnad.c (struct bnad_tx_info *)tcb->txq->tx->priv; txq 2414 drivers/net/ethernet/brocade/bna/bnad.c bnad->tx_info[i].tcb[j]->txq->tx_packets; txq 2416 drivers/net/ethernet/brocade/bna/bnad.c bnad->tx_info[i].tcb[j]->txq->tx_bytes; txq 827 drivers/net/ethernet/brocade/bna/bnad_ethtool.c bnad->tx_info[i].tcb[j]->txq) { txq 829 drivers/net/ethernet/brocade/bna/bnad_ethtool.c buf[bi++] = tcb->txq->tx_packets; txq 830 drivers/net/ethernet/brocade/bna/bnad_ethtool.c buf[bi++] = tcb->txq->tx_bytes; txq 182 drivers/net/ethernet/cavium/liquidio/lio_core.c void octeon_report_tx_completion_to_bql(void *txq, unsigned int pkts_compl, txq 185 drivers/net/ethernet/cavium/liquidio/lio_core.c struct netdev_queue *netdev_queue = txq; txq 224 drivers/net/ethernet/cavium/liquidio/lio_core.c struct netdev_queue *txq; txq 243 drivers/net/ethernet/cavium/liquidio/lio_core.c txq = netdev_get_tx_queue(skb->dev, skb_get_queue_mapping(skb)); txq 244 drivers/net/ethernet/cavium/liquidio/lio_core.c netdev_tx_sent_queue(txq, skb->len); txq 246 drivers/net/ethernet/cavium/liquidio/lio_core.c return netif_xmit_stopped(txq); txq 1071 drivers/net/ethernet/cavium/liquidio/lio_ethtool.c lio->txq = lio->linfo.txpciq[0].s.q_no; txq 3655 drivers/net/ethernet/cavium/liquidio/lio_main.c lio->txq = lio->linfo.txpciq[0].s.q_no; txq 3666 drivers/net/ethernet/cavium/liquidio/lio_main.c lio->tx_qsize = octeon_get_tx_qsize(octeon_dev, lio->txq); txq 2153 drivers/net/ethernet/cavium/liquidio/lio_vf_main.c lio->txq = lio->linfo.txpciq[0].s.q_no; txq 2156 drivers/net/ethernet/cavium/liquidio/lio_vf_main.c lio->tx_qsize = octeon_get_tx_qsize(octeon_dev, lio->txq); txq 386 drivers/net/ethernet/cavium/liquidio/lio_vf_rep.c if (octnet_iq_is_full(vf_rep->oct, parent_lio->txq)) { txq 417 drivers/net/ethernet/cavium/liquidio/lio_vf_rep.c sc->iq_no = parent_lio->txq; txq 70 drivers/net/ethernet/cavium/liquidio/octeon_main.h void octeon_report_tx_completion_to_bql(void *txq, unsigned int pkts_compl, txq 105 drivers/net/ethernet/cavium/liquidio/octeon_network.h int txq; txq 858 drivers/net/ethernet/cavium/thunder/nicvf_main.c struct netdev_queue *txq; txq 939 drivers/net/ethernet/cavium/thunder/nicvf_main.c txq = netdev_get_tx_queue(netdev, txq_idx); txq 941 drivers/net/ethernet/cavium/thunder/nicvf_main.c netdev_tx_completed_queue(txq, tx_pkts, tx_bytes); txq 945 drivers/net/ethernet/cavium/thunder/nicvf_main.c if (netif_tx_queue_stopped(txq) && netif_carrier_ok(netdev)) { txq 946 drivers/net/ethernet/cavium/thunder/nicvf_main.c netif_tx_wake_queue(txq); txq 1258 drivers/net/ethernet/cavium/thunder/nicvf_main.c struct netdev_queue *txq = netdev_get_tx_queue(netdev, qid); txq 1292 drivers/net/ethernet/cavium/thunder/nicvf_main.c if (!netif_tx_queue_stopped(txq) && txq 1294 drivers/net/ethernet/cavium/thunder/nicvf_main.c netif_tx_stop_queue(txq); txq 1301 drivers/net/ethernet/cavium/thunder/nicvf_main.c netif_tx_wake_queue(txq); txq 1468 drivers/net/ethernet/cavium/thunder/nicvf_queues.c struct netdev_queue *txq; txq 1470 drivers/net/ethernet/cavium/thunder/nicvf_queues.c txq = netdev_get_tx_queue(nic->pnicvf->netdev, txq 1473 drivers/net/ethernet/cavium/thunder/nicvf_queues.c netdev_tx_sent_queue(txq, skb->len); txq 204 drivers/net/ethernet/chelsio/cxgb3/adapter.h struct sge_txq txq[SGE_TXQ_PER_SET]; txq 601 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c t3_write_reg(adap, A_SG_KDOORBELL, F_SELEGRCNTX | V_EGRCNTX(qs->txq[j].cntxt_id)); txq 179 drivers/net/ethernet/chelsio/cxgb3/sge.c return container_of(q, struct sge_qset, txq[qidx]); txq 658 drivers/net/ethernet/chelsio/cxgb3/sge.c memset(q->txq, 0, sizeof(struct sge_txq) * SGE_TXQ_PER_SET); txq 695 drivers/net/ethernet/chelsio/cxgb3/sge.c if (q->txq[i].desc) { txq 697 drivers/net/ethernet/chelsio/cxgb3/sge.c t3_sge_enable_ecntxt(adapter, q->txq[i].cntxt_id, 0); txq 699 drivers/net/ethernet/chelsio/cxgb3/sge.c if (q->txq[i].sdesc) { txq 700 drivers/net/ethernet/chelsio/cxgb3/sge.c free_tx_desc(adapter, &q->txq[i], txq 701 drivers/net/ethernet/chelsio/cxgb3/sge.c q->txq[i].in_use); txq 702 drivers/net/ethernet/chelsio/cxgb3/sge.c kfree(q->txq[i].sdesc); txq 705 drivers/net/ethernet/chelsio/cxgb3/sge.c q->txq[i].size * txq 707 drivers/net/ethernet/chelsio/cxgb3/sge.c q->txq[i].desc, q->txq[i].phys_addr); txq 708 drivers/net/ethernet/chelsio/cxgb3/sge.c __skb_queue_purge(&q->txq[i].sendq); txq 735 drivers/net/ethernet/chelsio/cxgb3/sge.c qs->txq[TXQ_ETH].cntxt_id = FW_TUNNEL_SGEEC_START + id; txq 736 drivers/net/ethernet/chelsio/cxgb3/sge.c qs->txq[TXQ_ETH].token = FW_TUNNEL_TID_START + id; txq 737 drivers/net/ethernet/chelsio/cxgb3/sge.c qs->txq[TXQ_OFLD].cntxt_id = FW_OFLD_SGEEC_START + id; txq 738 drivers/net/ethernet/chelsio/cxgb3/sge.c qs->txq[TXQ_CTRL].cntxt_id = FW_CTRL_SGEEC_START + id; txq 739 drivers/net/ethernet/chelsio/cxgb3/sge.c qs->txq[TXQ_CTRL].token = FW_CTRL_TID_START + id; txq 1248 drivers/net/ethernet/chelsio/cxgb3/sge.c static inline void t3_stop_tx_queue(struct netdev_queue *txq, txq 1251 drivers/net/ethernet/chelsio/cxgb3/sge.c netif_tx_stop_queue(txq); txq 1269 drivers/net/ethernet/chelsio/cxgb3/sge.c struct netdev_queue *txq; txq 1285 drivers/net/ethernet/chelsio/cxgb3/sge.c q = &qs->txq[TXQ_ETH]; txq 1286 drivers/net/ethernet/chelsio/cxgb3/sge.c txq = netdev_get_tx_queue(dev, qidx); txq 1294 drivers/net/ethernet/chelsio/cxgb3/sge.c t3_stop_tx_queue(txq, qs, q); txq 1311 drivers/net/ethernet/chelsio/cxgb3/sge.c t3_stop_tx_queue(txq, qs, q); txq 1316 drivers/net/ethernet/chelsio/cxgb3/sge.c netif_tx_start_queue(txq); txq 1527 drivers/net/ethernet/chelsio/cxgb3/sge.c struct sge_txq *q = &qs->txq[TXQ_CTRL]; txq 1567 drivers/net/ethernet/chelsio/cxgb3/sge.c ret = ctrl_xmit(adap, &adap->sge.qs[0].txq[TXQ_CTRL], skb); txq 1744 drivers/net/ethernet/chelsio/cxgb3/sge.c struct sge_txq *q = &qs->txq[TXQ_OFLD]; txq 1838 drivers/net/ethernet/chelsio/cxgb3/sge.c return ctrl_xmit(adap, &qs->txq[TXQ_CTRL], skb); txq 1840 drivers/net/ethernet/chelsio/cxgb3/sge.c return ofld_xmit(adap, &qs->txq[TXQ_OFLD], skb); txq 1987 drivers/net/ethernet/chelsio/cxgb3/sge.c should_restart_tx(&qs->txq[TXQ_ETH]) && txq 1989 drivers/net/ethernet/chelsio/cxgb3/sge.c qs->txq[TXQ_ETH].restarts++; txq 1995 drivers/net/ethernet/chelsio/cxgb3/sge.c should_restart_tx(&qs->txq[TXQ_OFLD]) && txq 1997 drivers/net/ethernet/chelsio/cxgb3/sge.c qs->txq[TXQ_OFLD].restarts++; txq 1998 drivers/net/ethernet/chelsio/cxgb3/sge.c tasklet_schedule(&qs->txq[TXQ_OFLD].qresume_tsk); txq 2001 drivers/net/ethernet/chelsio/cxgb3/sge.c should_restart_tx(&qs->txq[TXQ_CTRL]) && txq 2003 drivers/net/ethernet/chelsio/cxgb3/sge.c qs->txq[TXQ_CTRL].restarts++; txq 2004 drivers/net/ethernet/chelsio/cxgb3/sge.c tasklet_schedule(&qs->txq[TXQ_CTRL].qresume_tsk); txq 2220 drivers/net/ethernet/chelsio/cxgb3/sge.c clear_bit(TXQ_RUNNING, &qs->txq[TXQ_ETH].flags); txq 2225 drivers/net/ethernet/chelsio/cxgb3/sge.c qs->txq[TXQ_ETH].processed += credits; txq 2229 drivers/net/ethernet/chelsio/cxgb3/sge.c qs->txq[TXQ_CTRL].processed += credits; txq 2233 drivers/net/ethernet/chelsio/cxgb3/sge.c clear_bit(TXQ_RUNNING, &qs->txq[TXQ_OFLD].flags); txq 2237 drivers/net/ethernet/chelsio/cxgb3/sge.c qs->txq[TXQ_OFLD].processed += credits; txq 2254 drivers/net/ethernet/chelsio/cxgb3/sge.c struct sge_txq *txq = &qs->txq[TXQ_ETH]; txq 2256 drivers/net/ethernet/chelsio/cxgb3/sge.c if (txq->cleaned + txq->in_use != txq->processed && txq 2257 drivers/net/ethernet/chelsio/cxgb3/sge.c !test_and_set_bit(TXQ_LAST_PKT_DB, &txq->flags)) { txq 2258 drivers/net/ethernet/chelsio/cxgb3/sge.c set_bit(TXQ_RUNNING, &txq->flags); txq 2260 drivers/net/ethernet/chelsio/cxgb3/sge.c V_EGRCNTX(txq->cntxt_id)); txq 2265 drivers/net/ethernet/chelsio/cxgb3/sge.c struct sge_txq *txq = &qs->txq[TXQ_OFLD]; txq 2267 drivers/net/ethernet/chelsio/cxgb3/sge.c if (txq->cleaned + txq->in_use != txq->processed && txq 2268 drivers/net/ethernet/chelsio/cxgb3/sge.c !test_and_set_bit(TXQ_LAST_PKT_DB, &txq->flags)) { txq 2269 drivers/net/ethernet/chelsio/cxgb3/sge.c set_bit(TXQ_RUNNING, &txq->flags); txq 2271 drivers/net/ethernet/chelsio/cxgb3/sge.c V_EGRCNTX(txq->cntxt_id)); txq 2930 drivers/net/ethernet/chelsio/cxgb3/sge.c tbd[TXQ_ETH] = reclaim_completed_tx(adap, &qs->txq[TXQ_ETH], txq 2935 drivers/net/ethernet/chelsio/cxgb3/sge.c if (spin_trylock(&qs->txq[TXQ_OFLD].lock)) { txq 2936 drivers/net/ethernet/chelsio/cxgb3/sge.c tbd[TXQ_OFLD] = reclaim_completed_tx(adap, &qs->txq[TXQ_OFLD], txq 2938 drivers/net/ethernet/chelsio/cxgb3/sge.c spin_unlock(&qs->txq[TXQ_OFLD].lock); txq 3074 drivers/net/ethernet/chelsio/cxgb3/sge.c q->txq[i].desc = alloc_ring(adapter->pdev, p->txq_size[i], txq 3076 drivers/net/ethernet/chelsio/cxgb3/sge.c &q->txq[i].phys_addr, txq 3077 drivers/net/ethernet/chelsio/cxgb3/sge.c &q->txq[i].sdesc); txq 3078 drivers/net/ethernet/chelsio/cxgb3/sge.c if (!q->txq[i].desc) txq 3081 drivers/net/ethernet/chelsio/cxgb3/sge.c q->txq[i].gen = 1; txq 3082 drivers/net/ethernet/chelsio/cxgb3/sge.c q->txq[i].size = p->txq_size[i]; txq 3083 drivers/net/ethernet/chelsio/cxgb3/sge.c spin_lock_init(&q->txq[i].lock); txq 3084 drivers/net/ethernet/chelsio/cxgb3/sge.c skb_queue_head_init(&q->txq[i].sendq); txq 3087 drivers/net/ethernet/chelsio/cxgb3/sge.c tasklet_init(&q->txq[TXQ_OFLD].qresume_tsk, restart_offloadq, txq 3089 drivers/net/ethernet/chelsio/cxgb3/sge.c tasklet_init(&q->txq[TXQ_CTRL].qresume_tsk, restart_ctrlq, txq 3101 drivers/net/ethernet/chelsio/cxgb3/sge.c q->txq[TXQ_ETH].stop_thres = nports * txq 3142 drivers/net/ethernet/chelsio/cxgb3/sge.c ret = t3_sge_init_ecntxt(adapter, q->txq[TXQ_ETH].cntxt_id, USE_GTS, txq 3143 drivers/net/ethernet/chelsio/cxgb3/sge.c SGE_CNTXT_ETH, id, q->txq[TXQ_ETH].phys_addr, txq 3144 drivers/net/ethernet/chelsio/cxgb3/sge.c q->txq[TXQ_ETH].size, q->txq[TXQ_ETH].token, txq 3150 drivers/net/ethernet/chelsio/cxgb3/sge.c ret = t3_sge_init_ecntxt(adapter, q->txq[TXQ_OFLD].cntxt_id, txq 3152 drivers/net/ethernet/chelsio/cxgb3/sge.c q->txq[TXQ_OFLD].phys_addr, txq 3153 drivers/net/ethernet/chelsio/cxgb3/sge.c q->txq[TXQ_OFLD].size, 0, 1, 0); txq 3159 drivers/net/ethernet/chelsio/cxgb3/sge.c ret = t3_sge_init_ecntxt(adapter, q->txq[TXQ_CTRL].cntxt_id, 0, txq 3161 drivers/net/ethernet/chelsio/cxgb3/sge.c q->txq[TXQ_CTRL].phys_addr, txq 3162 drivers/net/ethernet/chelsio/cxgb3/sge.c q->txq[TXQ_CTRL].size, txq 3163 drivers/net/ethernet/chelsio/cxgb3/sge.c q->txq[TXQ_CTRL].token, 1, 0); txq 3295 drivers/net/ethernet/chelsio/cxgb3/sge.c tasklet_kill(&qs->txq[TXQ_OFLD].qresume_tsk); txq 3296 drivers/net/ethernet/chelsio/cxgb3/sge.c tasklet_kill(&qs->txq[TXQ_CTRL].qresume_tsk); txq 3003 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c struct sge_txq *txq = (struct sge_txq *)q; \ txq 3004 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c QDESC_GET(txq, txq->desc, type, label); \ txq 237 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h static inline void cudbg_fill_qdesc_txq(const struct sge_txq *txq, txq 242 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h entry->qid = txq->cntxt_id; txq 244 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h entry->num_desc = txq->size; txq 245 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h entry->data_size = txq->size * sizeof(struct tx_desc); txq 246 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h memcpy(entry->data, txq->desc, entry->data_size); txq 759 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h struct netdev_queue *txq; /* associated netdev TX queue */ txq 1419 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h int t4_sge_alloc_eth_txq(struct adapter *adap, struct sge_eth_txq *txq, txq 1422 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h int t4_sge_alloc_ctrl_txq(struct adapter *adap, struct sge_ctrl_txq *txq, txq 1427 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h int t4_sge_alloc_uld_txq(struct adapter *adap, struct sge_uld_txq *txq, txq 1783 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h unsigned int vf, unsigned int txq, unsigned int txq_eth_ctrl, txq 931 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c struct sge_eth_txq *txq; txq 933 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c txq = &adap->sge.ethtxq[pi->first_qset]; txq 939 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c return adap->sge.dbqtimer_val[txq->dbqtimerix]; txq 989 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c struct sge_eth_txq *txq; txq 1015 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c txq = &s->ethtxq[pi->first_qset]; txq 1016 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c if (txq->dbqtimerix == min_timerix) txq 1019 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c for (qix = 0; qix < pi->nqsets; qix++, txq++) { txq 1024 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c FW_PARAMS_PARAM_YZ_V(txq->q.cntxt_id)); txq 1031 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c txq->dbqtimerix = min_timerix; txq 236 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c struct sge_eth_txq *txq = &adap->sge.ethtxq[pi->first_qset]; txq 242 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c for (i = 0; i < pi->nqsets; i++, txq++) { txq 249 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c FW_PARAMS_PARAM_YZ_V(txq->q.cntxt_id)); txq 265 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c txq->dcb_prio = enable ? value : 0; txq 576 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c struct sge_txq *txq; txq 578 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c txq = q->adap->sge.egr_map[qid - q->adap->sge.egr_start]; txq 579 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c txq->restarts++; txq 580 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c if (txq->q_type == CXGB4_TXQ_ETH) { txq 583 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c eq = container_of(txq, struct sge_eth_txq, q); txq 588 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c oq = container_of(txq, struct sge_uld_txq, q); txq 1064 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c int txq; txq 1081 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c txq = 0; txq 1083 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c txq = (vlan_tci & VLAN_PRIO_MASK) >> VLAN_PRIO_SHIFT; txq 1086 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c txq = skb->priority & 0x7; txq 1089 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c return txq; txq 1094 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c txq = (skb_rx_queue_recorded(skb) txq 1098 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c while (unlikely(txq >= dev->real_num_tx_queues)) txq 1099 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c txq -= dev->real_num_tx_queues; txq 1101 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c return txq; txq 2064 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c struct sge_uld_txq *txq = &txq_info->uldtxq[i]; txq 2066 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c disable_txq_db(&txq->q); txq 2086 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c struct sge_uld_txq *txq = &txq_info->uldtxq[i]; txq 2088 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c enable_txq_db(adap, &txq->q); txq 2167 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c struct sge_uld_txq *txq = &txq_info->uldtxq[i]; txq 2169 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c sync_txq_pidx(adap, &txq->q); txq 464 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c struct sge_uld_txq *txq = &txq_info->uldtxq[i]; txq 466 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c if (txq && txq->q.desc) { txq 467 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c tasklet_kill(&txq->qresume_tsk); txq 469 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c txq->q.cntxt_id); txq 470 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c free_tx_desc(adap, &txq->q, txq->q.in_use, false); txq 471 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c kfree(txq->q.sdesc); txq 472 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c __skb_queue_purge(&txq->sendq); txq 473 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c free_txq(adap, &txq->q); txq 488 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c struct sge_uld_txq *txq = &txq_info->uldtxq[i]; txq 490 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c txq->q.size = 1024; txq 491 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c err = t4_sge_alloc_uld_txq(adap, txq, adap->port[i / j], txq 148 drivers/net/ethernet/chelsio/cxgb4/sched.c struct sge_eth_txq *txq; txq 156 drivers/net/ethernet/chelsio/cxgb4/sched.c txq = &adap->sge.ethtxq[pi->first_qset + p->queue]; txq 157 drivers/net/ethernet/chelsio/cxgb4/sched.c qid = txq->q.cntxt_id; txq 190 drivers/net/ethernet/chelsio/cxgb4/sched.c struct sge_eth_txq *txq; txq 201 drivers/net/ethernet/chelsio/cxgb4/sched.c txq = &adap->sge.ethtxq[pi->first_qset + p->queue]; txq 202 drivers/net/ethernet/chelsio/cxgb4/sched.c qid = txq->q.cntxt_id; txq 1169 drivers/net/ethernet/chelsio/cxgb4/sge.c netif_tx_stop_queue(q->txq); txq 1331 drivers/net/ethernet/chelsio/cxgb4/sge.c if (!q->in_use || !__netif_tx_trylock(eq->txq)) txq 1347 drivers/net/ethernet/chelsio/cxgb4/sge.c if (netif_tx_queue_stopped(eq->txq) && hw_in_use < (q->size / 2)) { txq 1348 drivers/net/ethernet/chelsio/cxgb4/sge.c netif_tx_wake_queue(eq->txq); txq 1352 drivers/net/ethernet/chelsio/cxgb4/sge.c __netif_tx_unlock(eq->txq); txq 1714 drivers/net/ethernet/chelsio/cxgb4/sge.c struct sge_eth_txq *txq; txq 1743 drivers/net/ethernet/chelsio/cxgb4/sge.c txq = &adapter->sge.ethtxq[pi->first_qset + qidx]; txq 1748 drivers/net/ethernet/chelsio/cxgb4/sge.c reclaim_completed_tx(adapter, &txq->q, -1, true); txq 1756 drivers/net/ethernet/chelsio/cxgb4/sge.c credits = txq_avail(&txq->q) - ndesc; txq 1764 drivers/net/ethernet/chelsio/cxgb4/sge.c eth_txq_stop(txq); txq 1777 drivers/net/ethernet/chelsio/cxgb4/sge.c txq->mapping_err++; txq 1791 drivers/net/ethernet/chelsio/cxgb4/sge.c eth_txq_stop(txq); txq 1801 drivers/net/ethernet/chelsio/cxgb4/sge.c wr = (void *)&txq->q.desc[txq->q.pidx]; txq 1853 drivers/net/ethernet/chelsio/cxgb4/sge.c txq->tso++; txq 1854 drivers/net/ethernet/chelsio/cxgb4/sge.c txq->tx_cso += ssi->gso_segs; txq 1872 drivers/net/ethernet/chelsio/cxgb4/sge.c txq->tx_cso++; txq 1882 drivers/net/ethernet/chelsio/cxgb4/sge.c txq->vlan_ins++; txq 1901 drivers/net/ethernet/chelsio/cxgb4/sge.c cxgb4_inline_tx_skb(skb, &txq->q, cpl + 1); txq 1941 drivers/net/ethernet/chelsio/cxgb4/sge.c struct sge_txq *tq = &txq->q; txq 1969 drivers/net/ethernet/chelsio/cxgb4/sge.c txq_advance(&txq->q, ndesc); txq 1971 drivers/net/ethernet/chelsio/cxgb4/sge.c cxgb4_ring_tx_db(adapter, &txq->q, ndesc); txq 2251 drivers/net/ethernet/chelsio/cxgb4/sge.c struct sge_txq *txq; txq 2301 drivers/net/ethernet/chelsio/cxgb4/sge.c txq = &q->q; txq 2306 drivers/net/ethernet/chelsio/cxgb4/sge.c left = (u8 *)end - (u8 *)txq->stat; txq 2307 drivers/net/ethernet/chelsio/cxgb4/sge.c end = (void *)txq->desc + left; txq 2314 drivers/net/ethernet/chelsio/cxgb4/sge.c if (pos == (u64 *)txq->stat) { txq 2315 drivers/net/ethernet/chelsio/cxgb4/sge.c left = (u8 *)end - (u8 *)txq->stat; txq 2316 drivers/net/ethernet/chelsio/cxgb4/sge.c end = (void *)txq->desc + left; txq 2317 drivers/net/ethernet/chelsio/cxgb4/sge.c pos = (void *)txq->desc; txq 2431 drivers/net/ethernet/chelsio/cxgb4/sge.c struct sge_uld_txq *txq; txq 2447 drivers/net/ethernet/chelsio/cxgb4/sge.c txq = &txq_info->uldtxq[idx]; txq 2448 drivers/net/ethernet/chelsio/cxgb4/sge.c return ofld_xmit(txq, skb); txq 2560 drivers/net/ethernet/chelsio/cxgb4/sge.c struct sge_uld_txq *txq; txq 2573 drivers/net/ethernet/chelsio/cxgb4/sge.c txq = &txq_info->uldtxq[idx]; txq 2575 drivers/net/ethernet/chelsio/cxgb4/sge.c ret = ofld_xmit_direct(txq, src, len); txq 2893 drivers/net/ethernet/chelsio/cxgb4/sge.c struct sge_eth_txq *txq; txq 2914 drivers/net/ethernet/chelsio/cxgb4/sge.c txq = &s->ethtxq[pi->first_qset + rspq->idx]; txq 2915 drivers/net/ethernet/chelsio/cxgb4/sge.c t4_sge_eth_txq_egress_update(adapter, txq, -1); txq 3431 drivers/net/ethernet/chelsio/cxgb4/sge.c struct sge_uld_txq *txq = s->egr_map[id]; txq 3434 drivers/net/ethernet/chelsio/cxgb4/sge.c tasklet_schedule(&txq->qresume_tsk); txq 3728 drivers/net/ethernet/chelsio/cxgb4/sge.c int t4_sge_alloc_eth_txq(struct adapter *adap, struct sge_eth_txq *txq, txq 3739 drivers/net/ethernet/chelsio/cxgb4/sge.c nentries = txq->q.size + s->stat_len / sizeof(struct tx_desc); txq 3741 drivers/net/ethernet/chelsio/cxgb4/sge.c txq->q.desc = alloc_ring(adap->pdev_dev, txq->q.size, txq 3743 drivers/net/ethernet/chelsio/cxgb4/sge.c &txq->q.phys_addr, &txq->q.sdesc, s->stat_len, txq 3745 drivers/net/ethernet/chelsio/cxgb4/sge.c if (!txq->q.desc) txq 3779 drivers/net/ethernet/chelsio/cxgb4/sge.c c.eqaddr = cpu_to_be64(txq->q.phys_addr); txq 3790 drivers/net/ethernet/chelsio/cxgb4/sge.c FW_EQ_ETH_CMD_TIMERIX_V(txq->dbqtimerix)); txq 3794 drivers/net/ethernet/chelsio/cxgb4/sge.c kfree(txq->q.sdesc); txq 3795 drivers/net/ethernet/chelsio/cxgb4/sge.c txq->q.sdesc = NULL; txq 3798 drivers/net/ethernet/chelsio/cxgb4/sge.c txq->q.desc, txq->q.phys_addr); txq 3799 drivers/net/ethernet/chelsio/cxgb4/sge.c txq->q.desc = NULL; txq 3803 drivers/net/ethernet/chelsio/cxgb4/sge.c txq->q.q_type = CXGB4_TXQ_ETH; txq 3804 drivers/net/ethernet/chelsio/cxgb4/sge.c init_txq(adap, &txq->q, FW_EQ_ETH_CMD_EQID_G(ntohl(c.eqid_pkd))); txq 3805 drivers/net/ethernet/chelsio/cxgb4/sge.c txq->txq = netdevq; txq 3806 drivers/net/ethernet/chelsio/cxgb4/sge.c txq->tso = txq->tx_cso = txq->vlan_ins = 0; txq 3807 drivers/net/ethernet/chelsio/cxgb4/sge.c txq->mapping_err = 0; txq 3808 drivers/net/ethernet/chelsio/cxgb4/sge.c txq->dbqt = dbqt; txq 3813 drivers/net/ethernet/chelsio/cxgb4/sge.c int t4_sge_alloc_ctrl_txq(struct adapter *adap, struct sge_ctrl_txq *txq, txq 3824 drivers/net/ethernet/chelsio/cxgb4/sge.c nentries = txq->q.size + s->stat_len / sizeof(struct tx_desc); txq 3826 drivers/net/ethernet/chelsio/cxgb4/sge.c txq->q.desc = alloc_ring(adap->pdev_dev, nentries, txq 3827 drivers/net/ethernet/chelsio/cxgb4/sge.c sizeof(struct tx_desc), 0, &txq->q.phys_addr, txq 3829 drivers/net/ethernet/chelsio/cxgb4/sge.c if (!txq->q.desc) txq 3851 drivers/net/ethernet/chelsio/cxgb4/sge.c c.eqaddr = cpu_to_be64(txq->q.phys_addr); txq 3857 drivers/net/ethernet/chelsio/cxgb4/sge.c txq->q.desc, txq->q.phys_addr); txq 3858 drivers/net/ethernet/chelsio/cxgb4/sge.c txq->q.desc = NULL; txq 3862 drivers/net/ethernet/chelsio/cxgb4/sge.c txq->q.q_type = CXGB4_TXQ_CTRL; txq 3863 drivers/net/ethernet/chelsio/cxgb4/sge.c init_txq(adap, &txq->q, FW_EQ_CTRL_CMD_EQID_G(ntohl(c.cmpliqid_eqid))); txq 3864 drivers/net/ethernet/chelsio/cxgb4/sge.c txq->adap = adap; txq 3865 drivers/net/ethernet/chelsio/cxgb4/sge.c skb_queue_head_init(&txq->sendq); txq 3866 drivers/net/ethernet/chelsio/cxgb4/sge.c tasklet_init(&txq->qresume_tsk, restart_ctrlq, (unsigned long)txq); txq 3867 drivers/net/ethernet/chelsio/cxgb4/sge.c txq->full = 0; txq 3883 drivers/net/ethernet/chelsio/cxgb4/sge.c int t4_sge_alloc_uld_txq(struct adapter *adap, struct sge_uld_txq *txq, txq 3895 drivers/net/ethernet/chelsio/cxgb4/sge.c nentries = txq->q.size + s->stat_len / sizeof(struct tx_desc); txq 3897 drivers/net/ethernet/chelsio/cxgb4/sge.c txq->q.desc = alloc_ring(adap->pdev_dev, txq->q.size, txq 3899 drivers/net/ethernet/chelsio/cxgb4/sge.c &txq->q.phys_addr, &txq->q.sdesc, s->stat_len, txq 3901 drivers/net/ethernet/chelsio/cxgb4/sge.c if (!txq->q.desc) txq 3924 drivers/net/ethernet/chelsio/cxgb4/sge.c c.eqaddr = cpu_to_be64(txq->q.phys_addr); txq 3928 drivers/net/ethernet/chelsio/cxgb4/sge.c kfree(txq->q.sdesc); txq 3929 drivers/net/ethernet/chelsio/cxgb4/sge.c txq->q.sdesc = NULL; txq 3932 drivers/net/ethernet/chelsio/cxgb4/sge.c txq->q.desc, txq->q.phys_addr); txq 3933 drivers/net/ethernet/chelsio/cxgb4/sge.c txq->q.desc = NULL; txq 3937 drivers/net/ethernet/chelsio/cxgb4/sge.c txq->q.q_type = CXGB4_TXQ_ULD; txq 3938 drivers/net/ethernet/chelsio/cxgb4/sge.c init_txq(adap, &txq->q, FW_EQ_OFLD_CMD_EQID_G(ntohl(c.eqid_pkd))); txq 3939 drivers/net/ethernet/chelsio/cxgb4/sge.c txq->adap = adap; txq 3940 drivers/net/ethernet/chelsio/cxgb4/sge.c skb_queue_head_init(&txq->sendq); txq 3941 drivers/net/ethernet/chelsio/cxgb4/sge.c tasklet_init(&txq->qresume_tsk, restart_ofldq, (unsigned long)txq); txq 3942 drivers/net/ethernet/chelsio/cxgb4/sge.c txq->full = 0; txq 3943 drivers/net/ethernet/chelsio/cxgb4/sge.c txq->mapping_err = 0; txq 4036 drivers/net/ethernet/chelsio/cxgb4/sge.c __netif_tx_lock_bh(etq->txq); txq 4038 drivers/net/ethernet/chelsio/cxgb4/sge.c __netif_tx_unlock_bh(etq->txq); txq 4114 drivers/net/ethernet/chelsio/cxgb4/sge.c struct sge_uld_txq *txq = txq_info->uldtxq; txq 4117 drivers/net/ethernet/chelsio/cxgb4/sge.c if (txq->q.desc) txq 4118 drivers/net/ethernet/chelsio/cxgb4/sge.c tasklet_kill(&txq->qresume_tsk); txq 4128 drivers/net/ethernet/chelsio/cxgb4/sge.c struct sge_uld_txq *txq = txq_info->uldtxq; txq 4131 drivers/net/ethernet/chelsio/cxgb4/sge.c if (txq->q.desc) txq 4132 drivers/net/ethernet/chelsio/cxgb4/sge.c tasklet_kill(&txq->qresume_tsk); txq 7587 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c unsigned int vf, unsigned int txq, unsigned int txq_eth_ctrl, txq 7603 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c FW_PFVF_CMD_NEQ_V(txq)); txq 260 drivers/net/ethernet/chelsio/cxgb4vf/adapter.h struct netdev_queue *txq; /* associated netdev TX queue */ txq 538 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c struct sge_eth_txq *txq; txq 560 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c txq = container_of(tq, struct sge_eth_txq, q); txq 572 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c txq->q.restarts++; txq 573 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c netif_tx_wake_queue(txq->txq); txq 635 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c struct sge_eth_txq *txq = &s->ethtxq[pi->first_qset]; txq 638 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c for (qs = 0; qs < pi->nqsets; qs++, rxq++, txq++) { txq 645 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c err = t4vf_sge_alloc_eth_txq(adapter, txq, dev, txq 666 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c struct sge_eth_txq *txq = &s->ethtxq[pi->first_qset]; txq 669 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c for (qs = 0; qs < pi->nqsets; qs++, rxq++, txq++) { txq 671 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c EQ_MAP(s, txq->q.abs_id) = &txq->q; txq 1804 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c const struct sge_eth_txq *txq = &adapter->sge.ethtxq[pi->first_qset]; txq 1809 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c for (qs = 0; qs < pi->nqsets; qs++, rxq++, txq++) { txq 1810 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c stats->tso += txq->tso; txq 1811 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c stats->tx_csum += txq->tx_cso; txq 1814 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c stats->vlan_ins += txq->vlan_ins; txq 2070 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c #define T(s, v) S3("u", s, txq[qs].v) txq 2075 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c const struct sge_eth_txq *txq = &adapter->sge.ethtxq[r * QPL]; txq 2224 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c #define T3(fmt, s, v) S3(fmt, s, txq[qs].v) txq 2232 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c const struct sge_eth_txq *txq = &adapter->sge.ethtxq[r * QPL]; txq 2793 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c struct sge_eth_txq *txq = &s->ethtxq[qs]; txq 2797 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c txq->q.size = 1024; txq 1133 drivers/net/ethernet/chelsio/cxgb4vf/sge.c static void txq_stop(struct sge_eth_txq *txq) txq 1135 drivers/net/ethernet/chelsio/cxgb4vf/sge.c netif_tx_stop_queue(txq->txq); txq 1136 drivers/net/ethernet/chelsio/cxgb4vf/sge.c txq->q.stops++; txq 1164 drivers/net/ethernet/chelsio/cxgb4vf/sge.c struct sge_eth_txq *txq; txq 1198 drivers/net/ethernet/chelsio/cxgb4vf/sge.c txq = &adapter->sge.ethtxq[pi->first_qset + qidx]; txq 1208 drivers/net/ethernet/chelsio/cxgb4vf/sge.c reclaim_completed_tx(adapter, &txq->q, true); txq 1217 drivers/net/ethernet/chelsio/cxgb4vf/sge.c credits = txq_avail(&txq->q) - ndesc; txq 1226 drivers/net/ethernet/chelsio/cxgb4vf/sge.c txq_stop(txq); txq 1240 drivers/net/ethernet/chelsio/cxgb4vf/sge.c txq->mapping_err++; txq 1255 drivers/net/ethernet/chelsio/cxgb4vf/sge.c txq_stop(txq); txq 1266 drivers/net/ethernet/chelsio/cxgb4vf/sge.c wr = (void *)&txq->q.desc[txq->q.pidx]; txq 1322 drivers/net/ethernet/chelsio/cxgb4vf/sge.c txq->tso++; txq 1323 drivers/net/ethernet/chelsio/cxgb4vf/sge.c txq->tx_cso += ssi->gso_segs; txq 1340 drivers/net/ethernet/chelsio/cxgb4vf/sge.c txq->tx_cso++; txq 1350 drivers/net/ethernet/chelsio/cxgb4vf/sge.c txq->vlan_ins++; txq 1365 drivers/net/ethernet/chelsio/cxgb4vf/sge.c T4_TRACE5(adapter->tb[txq->q.cntxt_id & 7], txq 1367 drivers/net/ethernet/chelsio/cxgb4vf/sge.c ndesc, credits, txq->q.pidx, skb->len, ssi->nr_frags); txq 1379 drivers/net/ethernet/chelsio/cxgb4vf/sge.c inline_tx_skb(skb, &txq->q, cpl + 1); txq 1420 drivers/net/ethernet/chelsio/cxgb4vf/sge.c struct sge_txq *tq = &txq->q; txq 1449 drivers/net/ethernet/chelsio/cxgb4vf/sge.c txq_advance(&txq->q, ndesc); txq 1451 drivers/net/ethernet/chelsio/cxgb4vf/sge.c ring_tx_db(adapter, &txq->q, ndesc); txq 2134 drivers/net/ethernet/chelsio/cxgb4vf/sge.c struct sge_eth_txq *txq = &s->ethtxq[i]; txq 2136 drivers/net/ethernet/chelsio/cxgb4vf/sge.c if (reclaimable(&txq->q) && __netif_tx_trylock(txq->txq)) { txq 2137 drivers/net/ethernet/chelsio/cxgb4vf/sge.c int avail = reclaimable(&txq->q); txq 2142 drivers/net/ethernet/chelsio/cxgb4vf/sge.c free_tx_desc(adapter, &txq->q, avail, true); txq 2143 drivers/net/ethernet/chelsio/cxgb4vf/sge.c txq->q.in_use -= avail; txq 2144 drivers/net/ethernet/chelsio/cxgb4vf/sge.c __netif_tx_unlock(txq->txq); txq 2412 drivers/net/ethernet/chelsio/cxgb4vf/sge.c int t4vf_sge_alloc_eth_txq(struct adapter *adapter, struct sge_eth_txq *txq, txq 2426 drivers/net/ethernet/chelsio/cxgb4vf/sge.c nentries = txq->q.size + s->stat_len / sizeof(struct tx_desc); txq 2432 drivers/net/ethernet/chelsio/cxgb4vf/sge.c txq->q.desc = alloc_ring(adapter->pdev_dev, txq->q.size, txq 2435 drivers/net/ethernet/chelsio/cxgb4vf/sge.c &txq->q.phys_addr, &txq->q.sdesc, s->stat_len); txq 2436 drivers/net/ethernet/chelsio/cxgb4vf/sge.c if (!txq->q.desc) txq 2468 drivers/net/ethernet/chelsio/cxgb4vf/sge.c cmd.eqaddr = cpu_to_be64(txq->q.phys_addr); txq 2480 drivers/net/ethernet/chelsio/cxgb4vf/sge.c kfree(txq->q.sdesc); txq 2481 drivers/net/ethernet/chelsio/cxgb4vf/sge.c txq->q.sdesc = NULL; txq 2484 drivers/net/ethernet/chelsio/cxgb4vf/sge.c txq->q.desc, txq->q.phys_addr); txq 2485 drivers/net/ethernet/chelsio/cxgb4vf/sge.c txq->q.desc = NULL; txq 2489 drivers/net/ethernet/chelsio/cxgb4vf/sge.c txq->q.in_use = 0; txq 2490 drivers/net/ethernet/chelsio/cxgb4vf/sge.c txq->q.cidx = 0; txq 2491 drivers/net/ethernet/chelsio/cxgb4vf/sge.c txq->q.pidx = 0; txq 2492 drivers/net/ethernet/chelsio/cxgb4vf/sge.c txq->q.stat = (void *)&txq->q.desc[txq->q.size]; txq 2493 drivers/net/ethernet/chelsio/cxgb4vf/sge.c txq->q.cntxt_id = FW_EQ_ETH_CMD_EQID_G(be32_to_cpu(rpl.eqid_pkd)); txq 2494 drivers/net/ethernet/chelsio/cxgb4vf/sge.c txq->q.bar2_addr = bar2_address(adapter, txq 2495 drivers/net/ethernet/chelsio/cxgb4vf/sge.c txq->q.cntxt_id, txq 2497 drivers/net/ethernet/chelsio/cxgb4vf/sge.c &txq->q.bar2_qid); txq 2498 drivers/net/ethernet/chelsio/cxgb4vf/sge.c txq->q.abs_id = txq 2500 drivers/net/ethernet/chelsio/cxgb4vf/sge.c txq->txq = devq; txq 2501 drivers/net/ethernet/chelsio/cxgb4vf/sge.c txq->tso = 0; txq 2502 drivers/net/ethernet/chelsio/cxgb4vf/sge.c txq->tx_cso = 0; txq 2503 drivers/net/ethernet/chelsio/cxgb4vf/sge.c txq->vlan_ins = 0; txq 2504 drivers/net/ethernet/chelsio/cxgb4vf/sge.c txq->q.stops = 0; txq 2505 drivers/net/ethernet/chelsio/cxgb4vf/sge.c txq->q.restarts = 0; txq 2506 drivers/net/ethernet/chelsio/cxgb4vf/sge.c txq->mapping_err = 0; txq 2567 drivers/net/ethernet/chelsio/cxgb4vf/sge.c struct sge_eth_txq *txq = s->ethtxq; txq 2572 drivers/net/ethernet/chelsio/cxgb4vf/sge.c for (qs = 0; qs < adapter->sge.ethqsets; qs++, rxq++, txq++) { txq 2575 drivers/net/ethernet/chelsio/cxgb4vf/sge.c if (txq->q.desc) { txq 2576 drivers/net/ethernet/chelsio/cxgb4vf/sge.c t4vf_eth_eq_free(adapter, txq->q.cntxt_id); txq 2577 drivers/net/ethernet/chelsio/cxgb4vf/sge.c free_tx_desc(adapter, &txq->q, txq->q.in_use, true); txq 2578 drivers/net/ethernet/chelsio/cxgb4vf/sge.c kfree(txq->q.sdesc); txq 2579 drivers/net/ethernet/chelsio/cxgb4vf/sge.c free_txq(adapter, &txq->q); txq 861 drivers/net/ethernet/cisco/enic/enic_main.c struct netdev_queue *txq; txq 870 drivers/net/ethernet/cisco/enic/enic_main.c txq = netdev_get_tx_queue(netdev, txq_map); txq 888 drivers/net/ethernet/cisco/enic/enic_main.c netif_tx_stop_queue(txq); txq 898 drivers/net/ethernet/cisco/enic/enic_main.c netif_tx_stop_queue(txq); txq 900 drivers/net/ethernet/cisco/enic/enic_main.c if (!netdev_xmit_more() || netif_xmit_stopped(txq)) txq 125 drivers/net/ethernet/cortina/gemini.c struct gmac_txq txq[TX_QUEUE_NUM]; txq 555 drivers/net/ethernet/cortina/gemini.c struct gmac_txq *txq = port->txq; txq 589 drivers/net/ethernet/cortina/gemini.c txq->ring = desc_ring; txq 590 drivers/net/ethernet/cortina/gemini.c txq->skb = skb_tab; txq 591 drivers/net/ethernet/cortina/gemini.c txq->noirq_packets = 0; txq 597 drivers/net/ethernet/cortina/gemini.c txq->cptr = r; txq 599 drivers/net/ethernet/cortina/gemini.c txq++; txq 607 drivers/net/ethernet/cortina/gemini.c static void gmac_clean_txq(struct net_device *netdev, struct gmac_txq *txq, txq 613 drivers/net/ethernet/cortina/gemini.c unsigned int c = txq->cptr; txq 629 drivers/net/ethernet/cortina/gemini.c txd = txq->ring + c; txq 639 drivers/net/ethernet/cortina/gemini.c dev_kfree_skb(txq->skb[c]); txq 676 drivers/net/ethernet/cortina/gemini.c txq->cptr = c; txq 695 drivers/net/ethernet/cortina/gemini.c gmac_clean_txq(netdev, port->txq + i, r); txq 699 drivers/net/ethernet/cortina/gemini.c kfree(port->txq->skb); txq 701 drivers/net/ethernet/cortina/gemini.c n_txq * sizeof(*port->txq->ring) << port->txq_order, txq 702 drivers/net/ethernet/cortina/gemini.c port->txq->ring, port->txq_dma_base); txq 1111 drivers/net/ethernet/cortina/gemini.c unsigned int txq, int en) txq 1119 drivers/net/ethernet/cortina/gemini.c mask = GMAC0_IRQ0_TXQ0_INTS << (6 * netdev->dev_id + txq); txq 1138 drivers/net/ethernet/cortina/gemini.c struct gmac_txq *txq, unsigned short *desc) txq 1193 drivers/net/ethernet/cortina/gemini.c txq->skb[w] = skb; txq 1201 drivers/net/ethernet/cortina/gemini.c txd = txq->ring + w; txq 1221 drivers/net/ethernet/cortina/gemini.c dma_unmap_page(geth->dev, txq->ring[w].word2.buf_adr, txq 1222 drivers/net/ethernet/cortina/gemini.c txq->ring[w].word0.bits.buffer_size, txq 1235 drivers/net/ethernet/cortina/gemini.c struct gmac_txq *txq; txq 1244 drivers/net/ethernet/cortina/gemini.c txq = &port->txq[txq_num]; txq 1252 drivers/net/ethernet/cortina/gemini.c d = txq->cptr - w - 1; txq 1256 drivers/net/ethernet/cortina/gemini.c gmac_clean_txq(netdev, txq, r); txq 1257 drivers/net/ethernet/cortina/gemini.c d = txq->cptr - w - 1; txq 1263 drivers/net/ethernet/cortina/gemini.c d = txq->cptr + nfrags + 16; txq 1265 drivers/net/ethernet/cortina/gemini.c txq->ring[d].word3.bits.eofie = 1; txq 1275 drivers/net/ethernet/cortina/gemini.c if (gmac_map_tx_bufs(netdev, skb, txq, &w)) { txq 1283 drivers/net/ethernet/cortina/gemini.c if (gmac_map_tx_bufs(netdev, skb, txq, &w)) txq 1289 drivers/net/ethernet/cortina/gemini.c gmac_clean_txq(netdev, txq, r); txq 1361 drivers/net/ethernet/emulex/benet/be_cmds.c struct be_queue_info *txq = &txo->q; txq 1363 drivers/net/ethernet/emulex/benet/be_cmds.c struct be_dma_mem *q_mem = &txq->dma_mem; txq 1385 drivers/net/ethernet/emulex/benet/be_cmds.c req->queue_size = be_encoded_q_len(txq->len); txq 1393 drivers/net/ethernet/emulex/benet/be_cmds.c txq->id = le16_to_cpu(resp->cid); txq 1398 drivers/net/ethernet/emulex/benet/be_cmds.c txq->created = true; txq 934 drivers/net/ethernet/emulex/benet/be_main.c struct be_queue_info *txq = &txo->q; txq 935 drivers/net/ethernet/emulex/benet/be_main.c struct be_eth_hdr_wrb *hdr = queue_index_node(txq, head); txq 943 drivers/net/ethernet/emulex/benet/be_main.c atomic_add(num_frags, &txq->used); txq 953 drivers/net/ethernet/emulex/benet/be_main.c struct be_queue_info *txq = &txo->q; txq 955 drivers/net/ethernet/emulex/benet/be_main.c wrb = queue_head_node(txq); txq 957 drivers/net/ethernet/emulex/benet/be_main.c queue_head_inc(txq); txq 970 drivers/net/ethernet/emulex/benet/be_main.c struct be_queue_info *txq = &txo->q; txq 973 drivers/net/ethernet/emulex/benet/be_main.c txq->head = head; txq 976 drivers/net/ethernet/emulex/benet/be_main.c queue_head_inc(txq); txq 978 drivers/net/ethernet/emulex/benet/be_main.c wrb = queue_head_node(txq); txq 982 drivers/net/ethernet/emulex/benet/be_main.c queue_head_inc(txq); txq 985 drivers/net/ethernet/emulex/benet/be_main.c txq->head = head; txq 1227 drivers/net/ethernet/emulex/benet/be_main.c struct be_queue_info *txq = &txo->q; txq 1228 drivers/net/ethernet/emulex/benet/be_main.c struct be_eth_hdr_wrb *hdr = queue_index_node(txq, txo->last_req_hdr); txq 1236 drivers/net/ethernet/emulex/benet/be_main.c wrb_fill_dummy(queue_head_node(txq)); txq 1237 drivers/net/ethernet/emulex/benet/be_main.c queue_head_inc(txq); txq 1238 drivers/net/ethernet/emulex/benet/be_main.c atomic_inc(&txq->used); txq 2752 drivers/net/ethernet/emulex/benet/be_main.c struct be_queue_info *txq = &txo->q; txq 2760 drivers/net/ethernet/emulex/benet/be_main.c if (sent_skbs[txq->tail]) { txq 2764 drivers/net/ethernet/emulex/benet/be_main.c skb = sent_skbs[txq->tail]; txq 2765 drivers/net/ethernet/emulex/benet/be_main.c sent_skbs[txq->tail] = NULL; txq 2766 drivers/net/ethernet/emulex/benet/be_main.c queue_tail_inc(txq); /* skip hdr wrb */ txq 2770 drivers/net/ethernet/emulex/benet/be_main.c wrb = queue_tail_node(txq); txq 2771 drivers/net/ethernet/emulex/benet/be_main.c frag_index = txq->tail; txq 2775 drivers/net/ethernet/emulex/benet/be_main.c queue_tail_inc(txq); txq 2872 drivers/net/ethernet/emulex/benet/be_main.c struct be_queue_info *txq; txq 2884 drivers/net/ethernet/emulex/benet/be_main.c txq = &txo->q; txq 2893 drivers/net/ethernet/emulex/benet/be_main.c atomic_sub(num_wrbs, &txq->used); txq 2909 drivers/net/ethernet/emulex/benet/be_main.c txq = &txo->q; txq 2911 drivers/net/ethernet/emulex/benet/be_main.c if (atomic_read(&txq->used)) { txq 2913 drivers/net/ethernet/emulex/benet/be_main.c i, atomic_read(&txq->used)); txq 2914 drivers/net/ethernet/emulex/benet/be_main.c notified_idx = txq->tail; txq 2915 drivers/net/ethernet/emulex/benet/be_main.c end_idx = txq->tail; txq 2916 drivers/net/ethernet/emulex/benet/be_main.c index_adv(&end_idx, atomic_read(&txq->used) - 1, txq 2917 drivers/net/ethernet/emulex/benet/be_main.c txq->len); txq 2922 drivers/net/ethernet/emulex/benet/be_main.c atomic_sub(num_wrbs, &txq->used); txq 2923 drivers/net/ethernet/emulex/benet/be_main.c BUG_ON(atomic_read(&txq->used)); txq 2928 drivers/net/ethernet/emulex/benet/be_main.c txq->head = notified_idx; txq 2929 drivers/net/ethernet/emulex/benet/be_main.c txq->tail = notified_idx; txq 672 drivers/net/ethernet/faraday/ftgmac100.c struct netdev_queue *txq; txq 674 drivers/net/ethernet/faraday/ftgmac100.c txq = netdev_get_tx_queue(netdev, 0); txq 675 drivers/net/ethernet/faraday/ftgmac100.c __netif_tx_lock(txq, smp_processor_id()); txq 679 drivers/net/ethernet/faraday/ftgmac100.c __netif_tx_unlock(txq); txq 2059 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c struct netdev_queue *txq; txq 2109 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c txq = netdev_get_tx_queue(net_dev, queue_mapping); txq 2112 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c txq->trans_start = jiffies; txq 279 drivers/net/ethernet/freescale/fec_main.c #define IS_TSO_HEADER(txq, addr) \ txq 280 drivers/net/ethernet/freescale/fec_main.c ((addr >= txq->tso_hdrs_dma) && \ txq 281 drivers/net/ethernet/freescale/fec_main.c (addr < txq->tso_hdrs_dma + txq->bd.ring_size * TSO_HEADER_SIZE)) txq 305 drivers/net/ethernet/freescale/fec_main.c static int fec_enet_get_free_txdesc_num(struct fec_enet_priv_tx_q *txq) txq 309 drivers/net/ethernet/freescale/fec_main.c entries = (((const char *)txq->dirty_tx - txq 310 drivers/net/ethernet/freescale/fec_main.c (const char *)txq->bd.cur) >> txq->bd.dsize_log2) - 1; txq 312 drivers/net/ethernet/freescale/fec_main.c return entries >= 0 ? entries : entries + txq->bd.ring_size; txq 338 drivers/net/ethernet/freescale/fec_main.c struct fec_enet_priv_tx_q *txq; txq 344 drivers/net/ethernet/freescale/fec_main.c txq = fep->tx_queue[0]; txq 345 drivers/net/ethernet/freescale/fec_main.c bdp = txq->bd.base; txq 350 drivers/net/ethernet/freescale/fec_main.c bdp == txq->bd.cur ? 'S' : ' ', txq 351 drivers/net/ethernet/freescale/fec_main.c bdp == txq->dirty_tx ? 'H' : ' ', txq 355 drivers/net/ethernet/freescale/fec_main.c txq->tx_skbuff[index]); txq 356 drivers/net/ethernet/freescale/fec_main.c bdp = fec_enet_get_nextdesc(bdp, &txq->bd); txq 358 drivers/net/ethernet/freescale/fec_main.c } while (bdp != txq->bd.base); txq 384 drivers/net/ethernet/freescale/fec_main.c fec_enet_txq_submit_frag_skb(struct fec_enet_priv_tx_q *txq, txq 389 drivers/net/ethernet/freescale/fec_main.c struct bufdesc *bdp = txq->bd.cur; txq 403 drivers/net/ethernet/freescale/fec_main.c bdp = fec_enet_get_nextdesc(bdp, &txq->bd); txq 424 drivers/net/ethernet/freescale/fec_main.c estatus |= FEC_TX_BD_FTYPE(txq->bd.qid); txq 433 drivers/net/ethernet/freescale/fec_main.c index = fec_enet_get_bd_index(bdp, &txq->bd); txq 436 drivers/net/ethernet/freescale/fec_main.c memcpy(txq->tx_bounce[index], bufaddr, frag_len); txq 437 drivers/net/ethernet/freescale/fec_main.c bufaddr = txq->tx_bounce[index]; txq 462 drivers/net/ethernet/freescale/fec_main.c bdp = txq->bd.cur; txq 464 drivers/net/ethernet/freescale/fec_main.c bdp = fec_enet_get_nextdesc(bdp, &txq->bd); txq 471 drivers/net/ethernet/freescale/fec_main.c static int fec_enet_txq_submit_skb(struct fec_enet_priv_tx_q *txq, txq 485 drivers/net/ethernet/freescale/fec_main.c entries_free = fec_enet_get_free_txdesc_num(txq); txq 500 drivers/net/ethernet/freescale/fec_main.c bdp = txq->bd.cur; txq 509 drivers/net/ethernet/freescale/fec_main.c index = fec_enet_get_bd_index(bdp, &txq->bd); txq 512 drivers/net/ethernet/freescale/fec_main.c memcpy(txq->tx_bounce[index], skb->data, buflen); txq 513 drivers/net/ethernet/freescale/fec_main.c bufaddr = txq->tx_bounce[index]; txq 529 drivers/net/ethernet/freescale/fec_main.c last_bdp = fec_enet_txq_submit_frag_skb(txq, skb, ndev); txq 557 drivers/net/ethernet/freescale/fec_main.c estatus |= FEC_TX_BD_FTYPE(txq->bd.qid); txq 566 drivers/net/ethernet/freescale/fec_main.c index = fec_enet_get_bd_index(last_bdp, &txq->bd); txq 568 drivers/net/ethernet/freescale/fec_main.c txq->tx_skbuff[index] = skb; txq 582 drivers/net/ethernet/freescale/fec_main.c bdp = fec_enet_get_nextdesc(last_bdp, &txq->bd); txq 590 drivers/net/ethernet/freescale/fec_main.c txq->bd.cur = bdp; txq 593 drivers/net/ethernet/freescale/fec_main.c writel(0, txq->bd.reg_desc_active); txq 599 drivers/net/ethernet/freescale/fec_main.c fec_enet_txq_put_data_tso(struct fec_enet_priv_tx_q *txq, struct sk_buff *skb, txq 617 drivers/net/ethernet/freescale/fec_main.c memcpy(txq->tx_bounce[index], data, size); txq 618 drivers/net/ethernet/freescale/fec_main.c data = txq->tx_bounce[index]; txq 637 drivers/net/ethernet/freescale/fec_main.c estatus |= FEC_TX_BD_FTYPE(txq->bd.qid); txq 659 drivers/net/ethernet/freescale/fec_main.c fec_enet_txq_put_hdr_tso(struct fec_enet_priv_tx_q *txq, txq 675 drivers/net/ethernet/freescale/fec_main.c bufaddr = txq->tso_hdrs + index * TSO_HEADER_SIZE; txq 676 drivers/net/ethernet/freescale/fec_main.c dmabuf = txq->tso_hdrs_dma + index * TSO_HEADER_SIZE; txq 679 drivers/net/ethernet/freescale/fec_main.c memcpy(txq->tx_bounce[index], skb->data, hdr_len); txq 680 drivers/net/ethernet/freescale/fec_main.c bufaddr = txq->tx_bounce[index]; txq 700 drivers/net/ethernet/freescale/fec_main.c estatus |= FEC_TX_BD_FTYPE(txq->bd.qid); txq 712 drivers/net/ethernet/freescale/fec_main.c static int fec_enet_txq_submit_tso(struct fec_enet_priv_tx_q *txq, txq 719 drivers/net/ethernet/freescale/fec_main.c struct bufdesc *bdp = txq->bd.cur; txq 724 drivers/net/ethernet/freescale/fec_main.c if (tso_count_descs(skb) >= fec_enet_get_free_txdesc_num(txq)) { txq 744 drivers/net/ethernet/freescale/fec_main.c index = fec_enet_get_bd_index(bdp, &txq->bd); txq 749 drivers/net/ethernet/freescale/fec_main.c hdr = txq->tso_hdrs + index * TSO_HEADER_SIZE; txq 751 drivers/net/ethernet/freescale/fec_main.c ret = fec_enet_txq_put_hdr_tso(txq, skb, ndev, bdp, index); txq 759 drivers/net/ethernet/freescale/fec_main.c bdp = fec_enet_get_nextdesc(bdp, &txq->bd); txq 760 drivers/net/ethernet/freescale/fec_main.c index = fec_enet_get_bd_index(bdp, &txq->bd); txq 761 drivers/net/ethernet/freescale/fec_main.c ret = fec_enet_txq_put_data_tso(txq, skb, ndev, txq 773 drivers/net/ethernet/freescale/fec_main.c bdp = fec_enet_get_nextdesc(bdp, &txq->bd); txq 777 drivers/net/ethernet/freescale/fec_main.c txq->tx_skbuff[index] = skb; txq 780 drivers/net/ethernet/freescale/fec_main.c txq->bd.cur = bdp; txq 784 drivers/net/ethernet/freescale/fec_main.c !readl(txq->bd.reg_desc_active) || txq 785 drivers/net/ethernet/freescale/fec_main.c !readl(txq->bd.reg_desc_active) || txq 786 drivers/net/ethernet/freescale/fec_main.c !readl(txq->bd.reg_desc_active) || txq 787 drivers/net/ethernet/freescale/fec_main.c !readl(txq->bd.reg_desc_active)) txq 788 drivers/net/ethernet/freescale/fec_main.c writel(0, txq->bd.reg_desc_active); txq 803 drivers/net/ethernet/freescale/fec_main.c struct fec_enet_priv_tx_q *txq; txq 808 drivers/net/ethernet/freescale/fec_main.c txq = fep->tx_queue[queue]; txq 812 drivers/net/ethernet/freescale/fec_main.c ret = fec_enet_txq_submit_tso(txq, skb, ndev); txq 814 drivers/net/ethernet/freescale/fec_main.c ret = fec_enet_txq_submit_skb(txq, skb, ndev); txq 818 drivers/net/ethernet/freescale/fec_main.c entries_free = fec_enet_get_free_txdesc_num(txq); txq 819 drivers/net/ethernet/freescale/fec_main.c if (entries_free <= txq->tx_stop_threshold) txq 830 drivers/net/ethernet/freescale/fec_main.c struct fec_enet_priv_tx_q *txq; txq 860 drivers/net/ethernet/freescale/fec_main.c txq = fep->tx_queue[q]; txq 861 drivers/net/ethernet/freescale/fec_main.c bdp = txq->bd.base; txq 862 drivers/net/ethernet/freescale/fec_main.c txq->bd.cur = bdp; txq 864 drivers/net/ethernet/freescale/fec_main.c for (i = 0; i < txq->bd.ring_size; i++) { txq 868 drivers/net/ethernet/freescale/fec_main.c !IS_TSO_HEADER(txq, fec32_to_cpu(bdp->cbd_bufaddr))) txq 873 drivers/net/ethernet/freescale/fec_main.c if (txq->tx_skbuff[i]) { txq 874 drivers/net/ethernet/freescale/fec_main.c dev_kfree_skb_any(txq->tx_skbuff[i]); txq 875 drivers/net/ethernet/freescale/fec_main.c txq->tx_skbuff[i] = NULL; txq 878 drivers/net/ethernet/freescale/fec_main.c bdp = fec_enet_get_nextdesc(bdp, &txq->bd); txq 882 drivers/net/ethernet/freescale/fec_main.c bdp = fec_enet_get_prevdesc(bdp, &txq->bd); txq 884 drivers/net/ethernet/freescale/fec_main.c txq->dirty_tx = bdp; txq 900 drivers/net/ethernet/freescale/fec_main.c struct fec_enet_priv_tx_q *txq; txq 916 drivers/net/ethernet/freescale/fec_main.c txq = fep->tx_queue[i]; txq 917 drivers/net/ethernet/freescale/fec_main.c writel(txq->bd.dma, fep->hwp + FEC_X_DES_START(i)); txq 929 drivers/net/ethernet/freescale/fec_main.c struct fec_enet_priv_tx_q *txq; txq 933 drivers/net/ethernet/freescale/fec_main.c txq = fep->tx_queue[i]; txq 935 drivers/net/ethernet/freescale/fec_main.c for (j = 0; j < txq->bd.ring_size; j++) { txq 936 drivers/net/ethernet/freescale/fec_main.c if (txq->tx_skbuff[j]) { txq 937 drivers/net/ethernet/freescale/fec_main.c dev_kfree_skb_any(txq->tx_skbuff[j]); txq 938 drivers/net/ethernet/freescale/fec_main.c txq->tx_skbuff[j] = NULL; txq 1248 drivers/net/ethernet/freescale/fec_main.c struct fec_enet_priv_tx_q *txq; txq 1257 drivers/net/ethernet/freescale/fec_main.c txq = fep->tx_queue[queue_id]; txq 1260 drivers/net/ethernet/freescale/fec_main.c bdp = txq->dirty_tx; txq 1263 drivers/net/ethernet/freescale/fec_main.c bdp = fec_enet_get_nextdesc(bdp, &txq->bd); txq 1265 drivers/net/ethernet/freescale/fec_main.c while (bdp != READ_ONCE(txq->bd.cur)) { txq 1272 drivers/net/ethernet/freescale/fec_main.c index = fec_enet_get_bd_index(bdp, &txq->bd); txq 1274 drivers/net/ethernet/freescale/fec_main.c skb = txq->tx_skbuff[index]; txq 1275 drivers/net/ethernet/freescale/fec_main.c txq->tx_skbuff[index] = NULL; txq 1276 drivers/net/ethernet/freescale/fec_main.c if (!IS_TSO_HEADER(txq, fec32_to_cpu(bdp->cbd_bufaddr))) txq 1327 drivers/net/ethernet/freescale/fec_main.c txq->dirty_tx = bdp; txq 1330 drivers/net/ethernet/freescale/fec_main.c bdp = fec_enet_get_nextdesc(bdp, &txq->bd); txq 1335 drivers/net/ethernet/freescale/fec_main.c entries_free = fec_enet_get_free_txdesc_num(txq); txq 1336 drivers/net/ethernet/freescale/fec_main.c if (entries_free >= txq->tx_wake_threshold) txq 1342 drivers/net/ethernet/freescale/fec_main.c if (bdp != txq->bd.cur && txq 1343 drivers/net/ethernet/freescale/fec_main.c readl(txq->bd.reg_desc_active) == 0) txq 1344 drivers/net/ethernet/freescale/fec_main.c writel(0, txq->bd.reg_desc_active); txq 2748 drivers/net/ethernet/freescale/fec_main.c struct fec_enet_priv_tx_q *txq; txq 2770 drivers/net/ethernet/freescale/fec_main.c txq = fep->tx_queue[q]; txq 2771 drivers/net/ethernet/freescale/fec_main.c bdp = txq->bd.base; txq 2772 drivers/net/ethernet/freescale/fec_main.c for (i = 0; i < txq->bd.ring_size; i++) { txq 2773 drivers/net/ethernet/freescale/fec_main.c kfree(txq->tx_bounce[i]); txq 2774 drivers/net/ethernet/freescale/fec_main.c txq->tx_bounce[i] = NULL; txq 2775 drivers/net/ethernet/freescale/fec_main.c skb = txq->tx_skbuff[i]; txq 2776 drivers/net/ethernet/freescale/fec_main.c txq->tx_skbuff[i] = NULL; txq 2786 drivers/net/ethernet/freescale/fec_main.c struct fec_enet_priv_tx_q *txq; txq 2790 drivers/net/ethernet/freescale/fec_main.c txq = fep->tx_queue[i]; txq 2792 drivers/net/ethernet/freescale/fec_main.c txq->bd.ring_size * TSO_HEADER_SIZE, txq 2793 drivers/net/ethernet/freescale/fec_main.c txq->tso_hdrs, txq 2794 drivers/net/ethernet/freescale/fec_main.c txq->tso_hdrs_dma); txq 2808 drivers/net/ethernet/freescale/fec_main.c struct fec_enet_priv_tx_q *txq; txq 2811 drivers/net/ethernet/freescale/fec_main.c txq = kzalloc(sizeof(*txq), GFP_KERNEL); txq 2812 drivers/net/ethernet/freescale/fec_main.c if (!txq) { txq 2817 drivers/net/ethernet/freescale/fec_main.c fep->tx_queue[i] = txq; txq 2818 drivers/net/ethernet/freescale/fec_main.c txq->bd.ring_size = TX_RING_SIZE; txq 2821 drivers/net/ethernet/freescale/fec_main.c txq->tx_stop_threshold = FEC_MAX_SKB_DESCS; txq 2822 drivers/net/ethernet/freescale/fec_main.c txq->tx_wake_threshold = txq 2823 drivers/net/ethernet/freescale/fec_main.c (txq->bd.ring_size - txq->tx_stop_threshold) / 2; txq 2825 drivers/net/ethernet/freescale/fec_main.c txq->tso_hdrs = dma_alloc_coherent(&fep->pdev->dev, txq 2826 drivers/net/ethernet/freescale/fec_main.c txq->bd.ring_size * TSO_HEADER_SIZE, txq 2827 drivers/net/ethernet/freescale/fec_main.c &txq->tso_hdrs_dma, txq 2829 drivers/net/ethernet/freescale/fec_main.c if (!txq->tso_hdrs) { txq 2901 drivers/net/ethernet/freescale/fec_main.c struct fec_enet_priv_tx_q *txq; txq 2903 drivers/net/ethernet/freescale/fec_main.c txq = fep->tx_queue[queue]; txq 2904 drivers/net/ethernet/freescale/fec_main.c bdp = txq->bd.base; txq 2905 drivers/net/ethernet/freescale/fec_main.c for (i = 0; i < txq->bd.ring_size; i++) { txq 2906 drivers/net/ethernet/freescale/fec_main.c txq->tx_bounce[i] = kmalloc(FEC_ENET_TX_FRSIZE, GFP_KERNEL); txq 2907 drivers/net/ethernet/freescale/fec_main.c if (!txq->tx_bounce[i]) txq 2918 drivers/net/ethernet/freescale/fec_main.c bdp = fec_enet_get_nextdesc(bdp, &txq->bd); txq 2922 drivers/net/ethernet/freescale/fec_main.c bdp = fec_enet_get_prevdesc(bdp, &txq->bd); txq 3290 drivers/net/ethernet/freescale/fec_main.c struct fec_enet_priv_tx_q *txq = fep->tx_queue[i]; txq 3291 drivers/net/ethernet/freescale/fec_main.c unsigned size = dsize * txq->bd.ring_size; txq 3293 drivers/net/ethernet/freescale/fec_main.c txq->bd.qid = i; txq 3294 drivers/net/ethernet/freescale/fec_main.c txq->bd.base = cbd_base; txq 3295 drivers/net/ethernet/freescale/fec_main.c txq->bd.cur = cbd_base; txq 3296 drivers/net/ethernet/freescale/fec_main.c txq->bd.dma = bd_dma; txq 3297 drivers/net/ethernet/freescale/fec_main.c txq->bd.dsize = dsize; txq 3298 drivers/net/ethernet/freescale/fec_main.c txq->bd.dsize_log2 = dsize_log2; txq 3299 drivers/net/ethernet/freescale/fec_main.c txq->bd.reg_desc_active = fep->hwp + offset_des_active_txq[i]; txq 3302 drivers/net/ethernet/freescale/fec_main.c txq->bd.last = (struct bufdesc *)(((void *)cbd_base) - dsize); txq 1156 drivers/net/ethernet/freescale/gianfar.c struct netdev_queue *txq; txq 1159 drivers/net/ethernet/freescale/gianfar.c txq = netdev_get_tx_queue(tx_queue->dev, tx_queue->qindex); txq 1162 drivers/net/ethernet/freescale/gianfar.c netdev_tx_reset_queue(txq); txq 1797 drivers/net/ethernet/freescale/gianfar.c struct netdev_queue *txq; txq 1810 drivers/net/ethernet/freescale/gianfar.c txq = netdev_get_tx_queue(dev, rq); txq 1855 drivers/net/ethernet/freescale/gianfar.c netif_tx_stop_queue(txq); txq 1979 drivers/net/ethernet/freescale/gianfar.c netdev_tx_sent_queue(txq, bytes_sent); txq 2011 drivers/net/ethernet/freescale/gianfar.c netif_tx_stop_queue(txq); txq 2186 drivers/net/ethernet/freescale/gianfar.c struct netdev_queue *txq; txq 2202 drivers/net/ethernet/freescale/gianfar.c txq = netdev_get_tx_queue(dev, tqi); txq 2282 drivers/net/ethernet/freescale/gianfar.c netif_tx_queue_stopped(txq) && txq 2290 drivers/net/ethernet/freescale/gianfar.c netdev_tx_completed_queue(txq, howmany, bytes_sent); txq 120 drivers/net/ethernet/hisilicon/hisi_femac.c struct hisi_femac_queue txq; txq 147 drivers/net/ethernet/hisilicon/hisi_femac.c dma_addr = priv->txq.dma_phys[pos]; txq 155 drivers/net/ethernet/hisilicon/hisi_femac.c struct hisi_femac_queue *txq = &priv->txq; txq 163 drivers/net/ethernet/hisilicon/hisi_femac.c skb = txq->skb[txq->tail]; txq 169 drivers/net/ethernet/hisilicon/hisi_femac.c hisi_femac_tx_dma_unmap(priv, skb, txq->tail); txq 177 drivers/net/ethernet/hisilicon/hisi_femac.c txq->skb[txq->tail] = NULL; txq 178 drivers/net/ethernet/hisilicon/hisi_femac.c txq->tail = (txq->tail + 1) % txq->num; txq 372 drivers/net/ethernet/hisilicon/hisi_femac.c ret = hisi_femac_init_queue(priv->dev, &priv->txq, TXQ_NUM); txq 387 drivers/net/ethernet/hisilicon/hisi_femac.c struct hisi_femac_queue *txq = &priv->txq; txq 412 drivers/net/ethernet/hisilicon/hisi_femac.c pos = txq->tail; txq 413 drivers/net/ethernet/hisilicon/hisi_femac.c while (pos != txq->head) { txq 414 drivers/net/ethernet/hisilicon/hisi_femac.c skb = txq->skb[pos]; txq 417 drivers/net/ethernet/hisilicon/hisi_femac.c pos, txq->head); txq 422 drivers/net/ethernet/hisilicon/hisi_femac.c txq->skb[pos] = NULL; txq 423 drivers/net/ethernet/hisilicon/hisi_femac.c pos = (pos + 1) % txq->num; txq 425 drivers/net/ethernet/hisilicon/hisi_femac.c txq->tail = pos; txq 503 drivers/net/ethernet/hisilicon/hisi_femac.c struct hisi_femac_queue *txq = &priv->txq; txq 517 drivers/net/ethernet/hisilicon/hisi_femac.c if (unlikely(!CIRC_SPACE(txq->head, txq->tail, txq 518 drivers/net/ethernet/hisilicon/hisi_femac.c txq->num))) { txq 533 drivers/net/ethernet/hisilicon/hisi_femac.c txq->dma_phys[txq->head] = addr; txq 535 drivers/net/ethernet/hisilicon/hisi_femac.c txq->skb[txq->head] = skb; txq 536 drivers/net/ethernet/hisilicon/hisi_femac.c txq->head = (txq->head + 1) % txq->num; txq 94 drivers/net/ethernet/huawei/hinic/hinic_main.c static void update_tx_stats(struct hinic_dev *nic_dev, struct hinic_txq *txq) txq 101 drivers/net/ethernet/huawei/hinic/hinic_main.c hinic_txq_get_stats(txq, &tx_stats); txq 112 drivers/net/ethernet/huawei/hinic/hinic_main.c hinic_txq_clean_stats(txq); txq 76 drivers/net/ethernet/huawei/hinic/hinic_tx.c void hinic_txq_clean_stats(struct hinic_txq *txq) txq 78 drivers/net/ethernet/huawei/hinic/hinic_tx.c struct hinic_txq_stats *txq_stats = &txq->txq_stats; txq 95 drivers/net/ethernet/huawei/hinic/hinic_tx.c void hinic_txq_get_stats(struct hinic_txq *txq, struct hinic_txq_stats *stats) txq 97 drivers/net/ethernet/huawei/hinic/hinic_tx.c struct hinic_txq_stats *txq_stats = &txq->txq_stats; txq 117 drivers/net/ethernet/huawei/hinic/hinic_tx.c static void txq_stats_init(struct hinic_txq *txq) txq 119 drivers/net/ethernet/huawei/hinic/hinic_tx.c struct hinic_txq_stats *txq_stats = &txq->txq_stats; txq 122 drivers/net/ethernet/huawei/hinic/hinic_tx.c hinic_txq_clean_stats(txq); txq 470 drivers/net/ethernet/huawei/hinic/hinic_tx.c struct hinic_txq *txq; txq 473 drivers/net/ethernet/huawei/hinic/hinic_tx.c txq = &nic_dev->txqs[q_id]; txq 474 drivers/net/ethernet/huawei/hinic/hinic_tx.c qp = container_of(txq->sq, struct hinic_qp, sq); txq 487 drivers/net/ethernet/huawei/hinic/hinic_tx.c u64_stats_update_begin(&txq->txq_stats.syncp); txq 488 drivers/net/ethernet/huawei/hinic/hinic_tx.c txq->txq_stats.big_frags_pkts++; txq 489 drivers/net/ethernet/huawei/hinic/hinic_tx.c u64_stats_update_end(&txq->txq_stats.syncp); txq 492 drivers/net/ethernet/huawei/hinic/hinic_tx.c if (nr_sges > txq->max_sges) { txq 497 drivers/net/ethernet/huawei/hinic/hinic_tx.c err = tx_map_skb(nic_dev, skb, txq->sges); txq 503 drivers/net/ethernet/huawei/hinic/hinic_tx.c sq_wqe = hinic_sq_get_wqe(txq->sq, wqe_size, &prod_idx); txq 510 drivers/net/ethernet/huawei/hinic/hinic_tx.c sq_wqe = hinic_sq_get_wqe(txq->sq, wqe_size, &prod_idx); txq 516 drivers/net/ethernet/huawei/hinic/hinic_tx.c tx_unmap_skb(nic_dev, skb, txq->sges); txq 518 drivers/net/ethernet/huawei/hinic/hinic_tx.c u64_stats_update_begin(&txq->txq_stats.syncp); txq 519 drivers/net/ethernet/huawei/hinic/hinic_tx.c txq->txq_stats.tx_busy++; txq 520 drivers/net/ethernet/huawei/hinic/hinic_tx.c u64_stats_update_end(&txq->txq_stats.syncp); txq 527 drivers/net/ethernet/huawei/hinic/hinic_tx.c hinic_sq_prepare_wqe(txq->sq, prod_idx, sq_wqe, txq->sges, nr_sges); txq 533 drivers/net/ethernet/huawei/hinic/hinic_tx.c hinic_sq_write_wqe(txq->sq, prod_idx, sq_wqe, skb, wqe_size); txq 538 drivers/net/ethernet/huawei/hinic/hinic_tx.c hinic_sq_write_db(txq->sq, prod_idx, wqe_size, 0); txq 543 drivers/net/ethernet/huawei/hinic/hinic_tx.c hinic_sq_return_wqe(txq->sq, wqe_size); txq 544 drivers/net/ethernet/huawei/hinic/hinic_tx.c tx_unmap_skb(nic_dev, skb, txq->sges); txq 550 drivers/net/ethernet/huawei/hinic/hinic_tx.c u64_stats_update_begin(&txq->txq_stats.syncp); txq 551 drivers/net/ethernet/huawei/hinic/hinic_tx.c txq->txq_stats.tx_dropped++; txq 552 drivers/net/ethernet/huawei/hinic/hinic_tx.c u64_stats_update_end(&txq->txq_stats.syncp); txq 575 drivers/net/ethernet/huawei/hinic/hinic_tx.c static void free_all_tx_skbs(struct hinic_txq *txq) txq 577 drivers/net/ethernet/huawei/hinic/hinic_tx.c struct hinic_dev *nic_dev = netdev_priv(txq->netdev); txq 578 drivers/net/ethernet/huawei/hinic/hinic_tx.c struct hinic_sq *sq = txq->sq; txq 592 drivers/net/ethernet/huawei/hinic/hinic_tx.c hinic_sq_get_sges(sq_wqe, txq->free_sges, nr_sges); txq 596 drivers/net/ethernet/huawei/hinic/hinic_tx.c tx_free_skb(nic_dev, skb, txq->free_sges); txq 609 drivers/net/ethernet/huawei/hinic/hinic_tx.c struct hinic_txq *txq = container_of(napi, struct hinic_txq, napi); txq 610 drivers/net/ethernet/huawei/hinic/hinic_tx.c struct hinic_qp *qp = container_of(txq->sq, struct hinic_qp, sq); txq 611 drivers/net/ethernet/huawei/hinic/hinic_tx.c struct hinic_dev *nic_dev = netdev_priv(txq->netdev); txq 613 drivers/net/ethernet/huawei/hinic/hinic_tx.c struct hinic_sq *sq = txq->sq; txq 647 drivers/net/ethernet/huawei/hinic/hinic_tx.c hinic_sq_get_sges(sq_wqe, txq->free_sges, nr_sges); txq 651 drivers/net/ethernet/huawei/hinic/hinic_tx.c tx_free_skb(nic_dev, skb, txq->free_sges); txq 656 drivers/net/ethernet/huawei/hinic/hinic_tx.c netdev_txq = netdev_get_tx_queue(txq->netdev, qp->q_id); txq 664 drivers/net/ethernet/huawei/hinic/hinic_tx.c u64_stats_update_begin(&txq->txq_stats.syncp); txq 665 drivers/net/ethernet/huawei/hinic/hinic_tx.c txq->txq_stats.tx_wake++; txq 666 drivers/net/ethernet/huawei/hinic/hinic_tx.c u64_stats_update_end(&txq->txq_stats.syncp); txq 669 drivers/net/ethernet/huawei/hinic/hinic_tx.c u64_stats_update_begin(&txq->txq_stats.syncp); txq 670 drivers/net/ethernet/huawei/hinic/hinic_tx.c txq->txq_stats.bytes += tx_bytes; txq 671 drivers/net/ethernet/huawei/hinic/hinic_tx.c txq->txq_stats.pkts += pkts; txq 672 drivers/net/ethernet/huawei/hinic/hinic_tx.c u64_stats_update_end(&txq->txq_stats.syncp); txq 685 drivers/net/ethernet/huawei/hinic/hinic_tx.c static void tx_napi_add(struct hinic_txq *txq, int weight) txq 687 drivers/net/ethernet/huawei/hinic/hinic_tx.c netif_napi_add(txq->netdev, &txq->napi, free_tx_poll, weight); txq 688 drivers/net/ethernet/huawei/hinic/hinic_tx.c napi_enable(&txq->napi); txq 691 drivers/net/ethernet/huawei/hinic/hinic_tx.c static void tx_napi_del(struct hinic_txq *txq) txq 693 drivers/net/ethernet/huawei/hinic/hinic_tx.c napi_disable(&txq->napi); txq 694 drivers/net/ethernet/huawei/hinic/hinic_tx.c netif_napi_del(&txq->napi); txq 699 drivers/net/ethernet/huawei/hinic/hinic_tx.c struct hinic_txq *txq = data; txq 702 drivers/net/ethernet/huawei/hinic/hinic_tx.c nic_dev = netdev_priv(txq->netdev); txq 706 drivers/net/ethernet/huawei/hinic/hinic_tx.c txq->sq->msix_entry, txq 709 drivers/net/ethernet/huawei/hinic/hinic_tx.c hinic_hwdev_msix_cnt_set(nic_dev->hwdev, txq->sq->msix_entry); txq 711 drivers/net/ethernet/huawei/hinic/hinic_tx.c napi_schedule(&txq->napi); txq 715 drivers/net/ethernet/huawei/hinic/hinic_tx.c static int tx_request_irq(struct hinic_txq *txq) txq 717 drivers/net/ethernet/huawei/hinic/hinic_tx.c struct hinic_dev *nic_dev = netdev_priv(txq->netdev); txq 721 drivers/net/ethernet/huawei/hinic/hinic_tx.c struct hinic_sq *sq = txq->sq; txq 724 drivers/net/ethernet/huawei/hinic/hinic_tx.c tx_napi_add(txq, nic_dev->tx_weight); txq 731 drivers/net/ethernet/huawei/hinic/hinic_tx.c err = request_irq(sq->irq, tx_irq, 0, txq->irq_name, txq); txq 734 drivers/net/ethernet/huawei/hinic/hinic_tx.c tx_napi_del(txq); txq 741 drivers/net/ethernet/huawei/hinic/hinic_tx.c static void tx_free_irq(struct hinic_txq *txq) txq 743 drivers/net/ethernet/huawei/hinic/hinic_tx.c struct hinic_sq *sq = txq->sq; txq 745 drivers/net/ethernet/huawei/hinic/hinic_tx.c free_irq(sq->irq, txq); txq 746 drivers/net/ethernet/huawei/hinic/hinic_tx.c tx_napi_del(txq); txq 757 drivers/net/ethernet/huawei/hinic/hinic_tx.c int hinic_init_txq(struct hinic_txq *txq, struct hinic_sq *sq, txq 766 drivers/net/ethernet/huawei/hinic/hinic_tx.c txq->netdev = netdev; txq 767 drivers/net/ethernet/huawei/hinic/hinic_tx.c txq->sq = sq; txq 769 drivers/net/ethernet/huawei/hinic/hinic_tx.c txq_stats_init(txq); txq 771 drivers/net/ethernet/huawei/hinic/hinic_tx.c txq->max_sges = HINIC_MAX_SQ_BUFDESCS; txq 773 drivers/net/ethernet/huawei/hinic/hinic_tx.c sges_size = txq->max_sges * sizeof(*txq->sges); txq 774 drivers/net/ethernet/huawei/hinic/hinic_tx.c txq->sges = devm_kzalloc(&netdev->dev, sges_size, GFP_KERNEL); txq 775 drivers/net/ethernet/huawei/hinic/hinic_tx.c if (!txq->sges) txq 778 drivers/net/ethernet/huawei/hinic/hinic_tx.c sges_size = txq->max_sges * sizeof(*txq->free_sges); txq 779 drivers/net/ethernet/huawei/hinic/hinic_tx.c txq->free_sges = devm_kzalloc(&netdev->dev, sges_size, GFP_KERNEL); txq 780 drivers/net/ethernet/huawei/hinic/hinic_tx.c if (!txq->free_sges) { txq 786 drivers/net/ethernet/huawei/hinic/hinic_tx.c txq->irq_name = devm_kzalloc(&netdev->dev, irqname_len, GFP_KERNEL); txq 787 drivers/net/ethernet/huawei/hinic/hinic_tx.c if (!txq->irq_name) { txq 792 drivers/net/ethernet/huawei/hinic/hinic_tx.c sprintf(txq->irq_name, "hinic_txq%d", qp->q_id); txq 799 drivers/net/ethernet/huawei/hinic/hinic_tx.c err = tx_request_irq(txq); txq 809 drivers/net/ethernet/huawei/hinic/hinic_tx.c devm_kfree(&netdev->dev, txq->irq_name); txq 812 drivers/net/ethernet/huawei/hinic/hinic_tx.c devm_kfree(&netdev->dev, txq->free_sges); txq 815 drivers/net/ethernet/huawei/hinic/hinic_tx.c devm_kfree(&netdev->dev, txq->sges); txq 823 drivers/net/ethernet/huawei/hinic/hinic_tx.c void hinic_clean_txq(struct hinic_txq *txq) txq 825 drivers/net/ethernet/huawei/hinic/hinic_tx.c struct net_device *netdev = txq->netdev; txq 827 drivers/net/ethernet/huawei/hinic/hinic_tx.c tx_free_irq(txq); txq 829 drivers/net/ethernet/huawei/hinic/hinic_tx.c free_all_tx_skbs(txq); txq 831 drivers/net/ethernet/huawei/hinic/hinic_tx.c devm_kfree(&netdev->dev, txq->irq_name); txq 832 drivers/net/ethernet/huawei/hinic/hinic_tx.c devm_kfree(&netdev->dev, txq->free_sges); txq 833 drivers/net/ethernet/huawei/hinic/hinic_tx.c devm_kfree(&netdev->dev, txq->sges); txq 43 drivers/net/ethernet/huawei/hinic/hinic_tx.h void hinic_txq_clean_stats(struct hinic_txq *txq); txq 45 drivers/net/ethernet/huawei/hinic/hinic_tx.h void hinic_txq_get_stats(struct hinic_txq *txq, struct hinic_txq_stats *stats); txq 49 drivers/net/ethernet/huawei/hinic/hinic_tx.h int hinic_init_txq(struct hinic_txq *txq, struct hinic_sq *sq, txq 52 drivers/net/ethernet/huawei/hinic/hinic_tx.h void hinic_clean_txq(struct hinic_txq *txq); txq 806 drivers/net/ethernet/ibm/ehea/ehea_main.c struct netdev_queue *txq = netdev_get_tx_queue(pr->port->netdev, txq 857 drivers/net/ethernet/ibm/ehea/ehea_main.c if (unlikely(netif_tx_queue_stopped(txq) && txq 859 drivers/net/ethernet/ibm/ehea/ehea_main.c __netif_tx_lock(txq, smp_processor_id()); txq 860 drivers/net/ethernet/ibm/ehea/ehea_main.c if (netif_tx_queue_stopped(txq) && txq 862 drivers/net/ethernet/ibm/ehea/ehea_main.c netif_tx_wake_queue(txq); txq 863 drivers/net/ethernet/ibm/ehea/ehea_main.c __netif_tx_unlock(txq); txq 2018 drivers/net/ethernet/ibm/ehea/ehea_main.c struct netdev_queue *txq; txq 2021 drivers/net/ethernet/ibm/ehea/ehea_main.c txq = netdev_get_tx_queue(dev, skb_get_queue_mapping(skb)); txq 2071 drivers/net/ethernet/ibm/ehea/ehea_main.c netif_tx_stop_queue(txq); txq 2079 drivers/net/ethernet/ibm/ehea/ehea_main.c netif_tx_stop_queue(txq); txq 1420 drivers/net/ethernet/ibm/ibmvnic.c struct netdev_queue *txq; txq 1454 drivers/net/ethernet/ibm/ibmvnic.c txq = netdev_get_tx_queue(netdev, skb_get_queue_mapping(skb)); txq 1608 drivers/net/ethernet/ibm/ibmvnic.c txq->trans_start = jiffies; txq 2130 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c qpi->txq.queue_id)) { txq 2135 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c vsi_queue_id = qpi->txq.queue_id; txq 2137 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c if (qpi->txq.vsi_id != qci->vsi_id || txq 2156 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c &qpi->txq)) { txq 272 drivers/net/ethernet/intel/iavf/iavf_virtchnl.c vqpi->txq.vsi_id = vqci->vsi_id; txq 273 drivers/net/ethernet/intel/iavf/iavf_virtchnl.c vqpi->txq.queue_id = i; txq 274 drivers/net/ethernet/intel/iavf/iavf_virtchnl.c vqpi->txq.ring_len = adapter->tx_rings[i].count; txq 275 drivers/net/ethernet/intel/iavf/iavf_virtchnl.c vqpi->txq.dma_ring_addr = adapter->tx_rings[i].dma; txq 1727 drivers/net/ethernet/intel/ice/ice_lib.c struct ice_aqc_add_txqs_perq *txq; txq 1763 drivers/net/ethernet/intel/ice/ice_lib.c txq = &qg_buf->txqs[0]; txq 1764 drivers/net/ethernet/intel/ice/ice_lib.c if (pf_q == le16_to_cpu(txq->txq_id)) txq 1765 drivers/net/ethernet/intel/ice/ice_lib.c ring->txq_teid = le32_to_cpu(txq->q_teid); txq 1927 drivers/net/ethernet/intel/ice/ice_lib.c ice_cfg_txq_interrupt(struct ice_vsi *vsi, u16 txq, u16 msix_idx, u16 itr_idx) txq 1930 drivers/net/ethernet/intel/ice/ice_lib.c ice_cfg_txq_interrupt(struct ice_vsi *vsi, u16 txq, u16 msix_idx, u16 itr_idx) txq 1942 drivers/net/ethernet/intel/ice/ice_lib.c wr32(hw, QINT_TQCTL(vsi->txq_map[txq]), val); txq 1988 drivers/net/ethernet/intel/ice/ice_lib.c u32 txq = 0, rxq = 0; txq 2012 drivers/net/ethernet/intel/ice/ice_lib.c ice_cfg_txq_interrupt(vsi, txq, reg_idx, txq 2014 drivers/net/ethernet/intel/ice/ice_lib.c txq++; txq 2681 drivers/net/ethernet/intel/ice/ice_lib.c u32 txq = 0; txq 2692 drivers/net/ethernet/intel/ice/ice_lib.c wr32(hw, QINT_TQCTL(vsi->txq_map[txq]), 0); txq 2693 drivers/net/ethernet/intel/ice/ice_lib.c txq++; txq 38 drivers/net/ethernet/intel/ice/ice_lib.h ice_cfg_txq_interrupt(struct ice_vsi *vsi, u16 txq, u16 msix_idx, u16 itr_idx); txq 2267 drivers/net/ethernet/intel/ice/ice_virtchnl_pf.c if (qpi->txq.vsi_id != qci->vsi_id || txq 2269 drivers/net/ethernet/intel/ice/ice_virtchnl_pf.c qpi->rxq.queue_id != qpi->txq.queue_id || txq 2270 drivers/net/ethernet/intel/ice/ice_virtchnl_pf.c qpi->txq.headwb_enabled || txq 2271 drivers/net/ethernet/intel/ice/ice_virtchnl_pf.c !ice_vc_isvalid_ring_len(qpi->txq.ring_len) || txq 2273 drivers/net/ethernet/intel/ice/ice_virtchnl_pf.c !ice_vc_isvalid_q_id(vf, qci->vsi_id, qpi->txq.queue_id)) { txq 2278 drivers/net/ethernet/intel/ice/ice_virtchnl_pf.c if (qpi->txq.ring_len > 0) { txq 2280 drivers/net/ethernet/intel/ice/ice_virtchnl_pf.c vsi->tx_rings[i]->dma = qpi->txq.dma_ring_addr; txq 2281 drivers/net/ethernet/intel/ice/ice_virtchnl_pf.c vsi->tx_rings[i]->count = qpi->txq.ring_len; txq 8507 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c int txq; txq 8513 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c txq = vdev->tc_to_txq[tc].offset; txq 8514 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c txq += reciprocal_scale(skb_get_hash(skb), txq 8517 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c return txq; txq 8538 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c txq = skb_rx_queue_recorded(skb) ? skb_get_rx_queue(skb) : txq 8541 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c while (txq >= f->indices) txq 8542 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c txq -= f->indices; txq 8544 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c return txq + f->offset; txq 174 drivers/net/ethernet/lantiq_etop.c struct netdev_queue *txq = txq 190 drivers/net/ethernet/lantiq_etop.c if (netif_tx_queue_stopped(txq)) txq 191 drivers/net/ethernet/lantiq_etop.c netif_tx_start_queue(txq); txq 458 drivers/net/ethernet/lantiq_etop.c struct netdev_queue *txq = netdev_get_tx_queue(dev, queue); txq 471 drivers/net/ethernet/lantiq_etop.c netif_tx_stop_queue(txq); txq 492 drivers/net/ethernet/lantiq_etop.c netif_tx_stop_queue(txq); txq 178 drivers/net/ethernet/marvell/mv643xx_eth.c #define IS_TSO_HEADER(txq, addr) \ txq 179 drivers/net/ethernet/marvell/mv643xx_eth.c ((addr >= txq->tso_hdrs_dma) && \ txq 180 drivers/net/ethernet/marvell/mv643xx_eth.c (addr < txq->tso_hdrs_dma + txq->tx_ring_size * TSO_HEADER_SIZE)) txq 407 drivers/net/ethernet/marvell/mv643xx_eth.c struct tx_queue txq[8]; txq 445 drivers/net/ethernet/marvell/mv643xx_eth.c static struct mv643xx_eth_private *txq_to_mp(struct tx_queue *txq) txq 447 drivers/net/ethernet/marvell/mv643xx_eth.c return container_of(txq, struct mv643xx_eth_private, txq[txq->index]); txq 466 drivers/net/ethernet/marvell/mv643xx_eth.c static void txq_reset_hw_ptr(struct tx_queue *txq) txq 468 drivers/net/ethernet/marvell/mv643xx_eth.c struct mv643xx_eth_private *mp = txq_to_mp(txq); txq 471 drivers/net/ethernet/marvell/mv643xx_eth.c addr = (u32)txq->tx_desc_dma; txq 472 drivers/net/ethernet/marvell/mv643xx_eth.c addr += txq->tx_curr_desc * sizeof(struct tx_desc); txq 473 drivers/net/ethernet/marvell/mv643xx_eth.c wrlp(mp, TXQ_CURRENT_DESC_PTR(txq->index), addr); txq 476 drivers/net/ethernet/marvell/mv643xx_eth.c static void txq_enable(struct tx_queue *txq) txq 478 drivers/net/ethernet/marvell/mv643xx_eth.c struct mv643xx_eth_private *mp = txq_to_mp(txq); txq 479 drivers/net/ethernet/marvell/mv643xx_eth.c wrlp(mp, TXQ_COMMAND, 1 << txq->index); txq 482 drivers/net/ethernet/marvell/mv643xx_eth.c static void txq_disable(struct tx_queue *txq) txq 484 drivers/net/ethernet/marvell/mv643xx_eth.c struct mv643xx_eth_private *mp = txq_to_mp(txq); txq 485 drivers/net/ethernet/marvell/mv643xx_eth.c u8 mask = 1 << txq->index; txq 492 drivers/net/ethernet/marvell/mv643xx_eth.c static void txq_maybe_wake(struct tx_queue *txq) txq 494 drivers/net/ethernet/marvell/mv643xx_eth.c struct mv643xx_eth_private *mp = txq_to_mp(txq); txq 495 drivers/net/ethernet/marvell/mv643xx_eth.c struct netdev_queue *nq = netdev_get_tx_queue(mp->dev, txq->index); txq 499 drivers/net/ethernet/marvell/mv643xx_eth.c if (txq->tx_desc_count <= txq->tx_wake_threshold) txq 730 drivers/net/ethernet/marvell/mv643xx_eth.c txq_put_data_tso(struct net_device *dev, struct tx_queue *txq, txq 738 drivers/net/ethernet/marvell/mv643xx_eth.c tx_index = txq->tx_curr_desc++; txq 739 drivers/net/ethernet/marvell/mv643xx_eth.c if (txq->tx_curr_desc == txq->tx_ring_size) txq 740 drivers/net/ethernet/marvell/mv643xx_eth.c txq->tx_curr_desc = 0; txq 741 drivers/net/ethernet/marvell/mv643xx_eth.c desc = &txq->tx_desc_area[tx_index]; txq 742 drivers/net/ethernet/marvell/mv643xx_eth.c txq->tx_desc_mapping[tx_index] = DESC_DMA_MAP_SINGLE; txq 749 drivers/net/ethernet/marvell/mv643xx_eth.c memcpy(txq->tso_hdrs + tx_index * TSO_HEADER_SIZE, txq 751 drivers/net/ethernet/marvell/mv643xx_eth.c desc->buf_ptr = txq->tso_hdrs_dma txq 755 drivers/net/ethernet/marvell/mv643xx_eth.c txq->tx_desc_mapping[tx_index] = DESC_DMA_MAP_SINGLE; txq 778 drivers/net/ethernet/marvell/mv643xx_eth.c txq_put_hdr_tso(struct sk_buff *skb, struct tx_queue *txq, int length, txq 781 drivers/net/ethernet/marvell/mv643xx_eth.c struct mv643xx_eth_private *mp = txq_to_mp(txq); txq 790 drivers/net/ethernet/marvell/mv643xx_eth.c tx_index = txq->tx_curr_desc; txq 791 drivers/net/ethernet/marvell/mv643xx_eth.c desc = &txq->tx_desc_area[tx_index]; txq 802 drivers/net/ethernet/marvell/mv643xx_eth.c desc->buf_ptr = txq->tso_hdrs_dma + txq 803 drivers/net/ethernet/marvell/mv643xx_eth.c txq->tx_curr_desc * TSO_HEADER_SIZE; txq 815 drivers/net/ethernet/marvell/mv643xx_eth.c txq->tx_curr_desc++; txq 816 drivers/net/ethernet/marvell/mv643xx_eth.c if (txq->tx_curr_desc == txq->tx_ring_size) txq 817 drivers/net/ethernet/marvell/mv643xx_eth.c txq->tx_curr_desc = 0; txq 820 drivers/net/ethernet/marvell/mv643xx_eth.c static int txq_submit_tso(struct tx_queue *txq, struct sk_buff *skb, txq 823 drivers/net/ethernet/marvell/mv643xx_eth.c struct mv643xx_eth_private *mp = txq_to_mp(txq); txq 832 drivers/net/ethernet/marvell/mv643xx_eth.c if ((txq->tx_desc_count + tso_count_descs(skb)) >= txq->tx_ring_size) { txq 837 drivers/net/ethernet/marvell/mv643xx_eth.c first_tx_desc = &txq->tx_desc_area[txq->tx_curr_desc]; txq 852 drivers/net/ethernet/marvell/mv643xx_eth.c hdr = txq->tso_hdrs + txq->tx_curr_desc * TSO_HEADER_SIZE; txq 854 drivers/net/ethernet/marvell/mv643xx_eth.c txq_put_hdr_tso(skb, txq, data_left, &first_cmd_sts, txq 862 drivers/net/ethernet/marvell/mv643xx_eth.c ret = txq_put_data_tso(dev, txq, skb, tso.data, size, txq 872 drivers/net/ethernet/marvell/mv643xx_eth.c __skb_queue_tail(&txq->tx_skb, skb); txq 880 drivers/net/ethernet/marvell/mv643xx_eth.c mp->work_tx_end &= ~(1 << txq->index); txq 884 drivers/net/ethernet/marvell/mv643xx_eth.c txq_enable(txq); txq 885 drivers/net/ethernet/marvell/mv643xx_eth.c txq->tx_desc_count += desc_count; txq 894 drivers/net/ethernet/marvell/mv643xx_eth.c static void txq_submit_frag_skb(struct tx_queue *txq, struct sk_buff *skb) txq 896 drivers/net/ethernet/marvell/mv643xx_eth.c struct mv643xx_eth_private *mp = txq_to_mp(txq); txq 906 drivers/net/ethernet/marvell/mv643xx_eth.c tx_index = txq->tx_curr_desc++; txq 907 drivers/net/ethernet/marvell/mv643xx_eth.c if (txq->tx_curr_desc == txq->tx_ring_size) txq 908 drivers/net/ethernet/marvell/mv643xx_eth.c txq->tx_curr_desc = 0; txq 909 drivers/net/ethernet/marvell/mv643xx_eth.c desc = &txq->tx_desc_area[tx_index]; txq 910 drivers/net/ethernet/marvell/mv643xx_eth.c txq->tx_desc_mapping[tx_index] = DESC_DMA_MAP_PAGE; txq 932 drivers/net/ethernet/marvell/mv643xx_eth.c static int txq_submit_skb(struct tx_queue *txq, struct sk_buff *skb, txq 935 drivers/net/ethernet/marvell/mv643xx_eth.c struct mv643xx_eth_private *mp = txq_to_mp(txq); txq 946 drivers/net/ethernet/marvell/mv643xx_eth.c if (txq->tx_ring_size - txq->tx_desc_count < MAX_SKB_FRAGS + 1) { txq 957 drivers/net/ethernet/marvell/mv643xx_eth.c tx_index = txq->tx_curr_desc++; txq 958 drivers/net/ethernet/marvell/mv643xx_eth.c if (txq->tx_curr_desc == txq->tx_ring_size) txq 959 drivers/net/ethernet/marvell/mv643xx_eth.c txq->tx_curr_desc = 0; txq 960 drivers/net/ethernet/marvell/mv643xx_eth.c desc = &txq->tx_desc_area[tx_index]; txq 961 drivers/net/ethernet/marvell/mv643xx_eth.c txq->tx_desc_mapping[tx_index] = DESC_DMA_MAP_SINGLE; txq 964 drivers/net/ethernet/marvell/mv643xx_eth.c txq_submit_frag_skb(txq, skb); txq 976 drivers/net/ethernet/marvell/mv643xx_eth.c __skb_queue_tail(&txq->tx_skb, skb); txq 985 drivers/net/ethernet/marvell/mv643xx_eth.c mp->work_tx_end &= ~(1 << txq->index); txq 989 drivers/net/ethernet/marvell/mv643xx_eth.c txq_enable(txq); txq 991 drivers/net/ethernet/marvell/mv643xx_eth.c txq->tx_desc_count += nr_frags + 1; txq 1000 drivers/net/ethernet/marvell/mv643xx_eth.c struct tx_queue *txq; txq 1004 drivers/net/ethernet/marvell/mv643xx_eth.c txq = mp->txq + queue; txq 1016 drivers/net/ethernet/marvell/mv643xx_eth.c ret = txq_submit_tso(txq, skb, dev); txq 1018 drivers/net/ethernet/marvell/mv643xx_eth.c ret = txq_submit_skb(txq, skb, dev); txq 1020 drivers/net/ethernet/marvell/mv643xx_eth.c txq->tx_bytes += length; txq 1021 drivers/net/ethernet/marvell/mv643xx_eth.c txq->tx_packets++; txq 1023 drivers/net/ethernet/marvell/mv643xx_eth.c if (txq->tx_desc_count >= txq->tx_stop_threshold) txq 1026 drivers/net/ethernet/marvell/mv643xx_eth.c txq->tx_dropped++; txq 1035 drivers/net/ethernet/marvell/mv643xx_eth.c static void txq_kick(struct tx_queue *txq) txq 1037 drivers/net/ethernet/marvell/mv643xx_eth.c struct mv643xx_eth_private *mp = txq_to_mp(txq); txq 1038 drivers/net/ethernet/marvell/mv643xx_eth.c struct netdev_queue *nq = netdev_get_tx_queue(mp->dev, txq->index); txq 1044 drivers/net/ethernet/marvell/mv643xx_eth.c if (rdlp(mp, TXQ_COMMAND) & (1 << txq->index)) txq 1047 drivers/net/ethernet/marvell/mv643xx_eth.c hw_desc_ptr = rdlp(mp, TXQ_CURRENT_DESC_PTR(txq->index)); txq 1048 drivers/net/ethernet/marvell/mv643xx_eth.c expected_ptr = (u32)txq->tx_desc_dma + txq 1049 drivers/net/ethernet/marvell/mv643xx_eth.c txq->tx_curr_desc * sizeof(struct tx_desc); txq 1052 drivers/net/ethernet/marvell/mv643xx_eth.c txq_enable(txq); txq 1057 drivers/net/ethernet/marvell/mv643xx_eth.c mp->work_tx_end &= ~(1 << txq->index); txq 1060 drivers/net/ethernet/marvell/mv643xx_eth.c static int txq_reclaim(struct tx_queue *txq, int budget, int force) txq 1062 drivers/net/ethernet/marvell/mv643xx_eth.c struct mv643xx_eth_private *mp = txq_to_mp(txq); txq 1063 drivers/net/ethernet/marvell/mv643xx_eth.c struct netdev_queue *nq = netdev_get_tx_queue(mp->dev, txq->index); txq 1069 drivers/net/ethernet/marvell/mv643xx_eth.c while (reclaimed < budget && txq->tx_desc_count > 0) { txq 1075 drivers/net/ethernet/marvell/mv643xx_eth.c tx_index = txq->tx_used_desc; txq 1076 drivers/net/ethernet/marvell/mv643xx_eth.c desc = &txq->tx_desc_area[tx_index]; txq 1077 drivers/net/ethernet/marvell/mv643xx_eth.c desc_dma_map = txq->tx_desc_mapping[tx_index]; txq 1087 drivers/net/ethernet/marvell/mv643xx_eth.c txq->tx_used_desc = tx_index + 1; txq 1088 drivers/net/ethernet/marvell/mv643xx_eth.c if (txq->tx_used_desc == txq->tx_ring_size) txq 1089 drivers/net/ethernet/marvell/mv643xx_eth.c txq->tx_used_desc = 0; txq 1092 drivers/net/ethernet/marvell/mv643xx_eth.c txq->tx_desc_count--; txq 1094 drivers/net/ethernet/marvell/mv643xx_eth.c if (!IS_TSO_HEADER(txq, desc->buf_ptr)) { txq 1109 drivers/net/ethernet/marvell/mv643xx_eth.c struct sk_buff *skb = __skb_dequeue(&txq->tx_skb); txq 1125 drivers/net/ethernet/marvell/mv643xx_eth.c mp->work_tx &= ~(1 << txq->index); txq 1168 drivers/net/ethernet/marvell/mv643xx_eth.c static void txq_set_rate(struct tx_queue *txq, int rate, int burst) txq 1170 drivers/net/ethernet/marvell/mv643xx_eth.c struct mv643xx_eth_private *mp = txq_to_mp(txq); txq 1182 drivers/net/ethernet/marvell/mv643xx_eth.c wrlp(mp, TXQ_BW_TOKENS(txq->index), token_rate << 14); txq 1183 drivers/net/ethernet/marvell/mv643xx_eth.c wrlp(mp, TXQ_BW_CONF(txq->index), (bucket_size << 10) | token_rate); txq 1186 drivers/net/ethernet/marvell/mv643xx_eth.c static void txq_set_fixed_prio_mode(struct tx_queue *txq) txq 1188 drivers/net/ethernet/marvell/mv643xx_eth.c struct mv643xx_eth_private *mp = txq_to_mp(txq); txq 1207 drivers/net/ethernet/marvell/mv643xx_eth.c val |= 1 << txq->index; txq 1265 drivers/net/ethernet/marvell/mv643xx_eth.c struct tx_queue *txq = mp->txq + i; txq 1267 drivers/net/ethernet/marvell/mv643xx_eth.c tx_packets += txq->tx_packets; txq 1268 drivers/net/ethernet/marvell/mv643xx_eth.c tx_bytes += txq->tx_bytes; txq 1269 drivers/net/ethernet/marvell/mv643xx_eth.c tx_dropped += txq->tx_dropped; txq 2030 drivers/net/ethernet/marvell/mv643xx_eth.c struct tx_queue *txq = mp->txq + index; txq 2036 drivers/net/ethernet/marvell/mv643xx_eth.c txq->index = index; txq 2038 drivers/net/ethernet/marvell/mv643xx_eth.c txq->tx_ring_size = mp->tx_ring_size; txq 2044 drivers/net/ethernet/marvell/mv643xx_eth.c txq->tx_stop_threshold = txq->tx_ring_size - MV643XX_MAX_SKB_DESCS; txq 2045 drivers/net/ethernet/marvell/mv643xx_eth.c txq->tx_wake_threshold = txq->tx_stop_threshold / 2; txq 2047 drivers/net/ethernet/marvell/mv643xx_eth.c txq->tx_desc_count = 0; txq 2048 drivers/net/ethernet/marvell/mv643xx_eth.c txq->tx_curr_desc = 0; txq 2049 drivers/net/ethernet/marvell/mv643xx_eth.c txq->tx_used_desc = 0; txq 2051 drivers/net/ethernet/marvell/mv643xx_eth.c size = txq->tx_ring_size * sizeof(struct tx_desc); txq 2054 drivers/net/ethernet/marvell/mv643xx_eth.c txq->tx_desc_area = ioremap(mp->tx_desc_sram_addr, txq 2056 drivers/net/ethernet/marvell/mv643xx_eth.c txq->tx_desc_dma = mp->tx_desc_sram_addr; txq 2058 drivers/net/ethernet/marvell/mv643xx_eth.c txq->tx_desc_area = dma_alloc_coherent(mp->dev->dev.parent, txq 2059 drivers/net/ethernet/marvell/mv643xx_eth.c size, &txq->tx_desc_dma, txq 2063 drivers/net/ethernet/marvell/mv643xx_eth.c if (txq->tx_desc_area == NULL) { txq 2068 drivers/net/ethernet/marvell/mv643xx_eth.c memset(txq->tx_desc_area, 0, size); txq 2070 drivers/net/ethernet/marvell/mv643xx_eth.c txq->tx_desc_area_size = size; txq 2072 drivers/net/ethernet/marvell/mv643xx_eth.c tx_desc = txq->tx_desc_area; txq 2073 drivers/net/ethernet/marvell/mv643xx_eth.c for (i = 0; i < txq->tx_ring_size; i++) { txq 2078 drivers/net/ethernet/marvell/mv643xx_eth.c if (nexti == txq->tx_ring_size) txq 2082 drivers/net/ethernet/marvell/mv643xx_eth.c txd->next_desc_ptr = txq->tx_desc_dma + txq 2086 drivers/net/ethernet/marvell/mv643xx_eth.c txq->tx_desc_mapping = kcalloc(txq->tx_ring_size, sizeof(char), txq 2088 drivers/net/ethernet/marvell/mv643xx_eth.c if (!txq->tx_desc_mapping) { txq 2094 drivers/net/ethernet/marvell/mv643xx_eth.c txq->tso_hdrs = dma_alloc_coherent(mp->dev->dev.parent, txq 2095 drivers/net/ethernet/marvell/mv643xx_eth.c txq->tx_ring_size * TSO_HEADER_SIZE, txq 2096 drivers/net/ethernet/marvell/mv643xx_eth.c &txq->tso_hdrs_dma, GFP_KERNEL); txq 2097 drivers/net/ethernet/marvell/mv643xx_eth.c if (txq->tso_hdrs == NULL) { txq 2101 drivers/net/ethernet/marvell/mv643xx_eth.c skb_queue_head_init(&txq->tx_skb); txq 2106 drivers/net/ethernet/marvell/mv643xx_eth.c kfree(txq->tx_desc_mapping); txq 2109 drivers/net/ethernet/marvell/mv643xx_eth.c iounmap(txq->tx_desc_area); txq 2111 drivers/net/ethernet/marvell/mv643xx_eth.c dma_free_coherent(mp->dev->dev.parent, txq->tx_desc_area_size, txq 2112 drivers/net/ethernet/marvell/mv643xx_eth.c txq->tx_desc_area, txq->tx_desc_dma); txq 2116 drivers/net/ethernet/marvell/mv643xx_eth.c static void txq_deinit(struct tx_queue *txq) txq 2118 drivers/net/ethernet/marvell/mv643xx_eth.c struct mv643xx_eth_private *mp = txq_to_mp(txq); txq 2120 drivers/net/ethernet/marvell/mv643xx_eth.c txq_disable(txq); txq 2121 drivers/net/ethernet/marvell/mv643xx_eth.c txq_reclaim(txq, txq->tx_ring_size, 1); txq 2123 drivers/net/ethernet/marvell/mv643xx_eth.c BUG_ON(txq->tx_used_desc != txq->tx_curr_desc); txq 2125 drivers/net/ethernet/marvell/mv643xx_eth.c if (txq->index == 0 && txq 2126 drivers/net/ethernet/marvell/mv643xx_eth.c txq->tx_desc_area_size <= mp->tx_desc_sram_size) txq 2127 drivers/net/ethernet/marvell/mv643xx_eth.c iounmap(txq->tx_desc_area); txq 2129 drivers/net/ethernet/marvell/mv643xx_eth.c dma_free_coherent(mp->dev->dev.parent, txq->tx_desc_area_size, txq 2130 drivers/net/ethernet/marvell/mv643xx_eth.c txq->tx_desc_area, txq->tx_desc_dma); txq 2131 drivers/net/ethernet/marvell/mv643xx_eth.c kfree(txq->tx_desc_mapping); txq 2133 drivers/net/ethernet/marvell/mv643xx_eth.c if (txq->tso_hdrs) txq 2135 drivers/net/ethernet/marvell/mv643xx_eth.c txq->tx_ring_size * TSO_HEADER_SIZE, txq 2136 drivers/net/ethernet/marvell/mv643xx_eth.c txq->tso_hdrs, txq->tso_hdrs_dma); txq 2206 drivers/net/ethernet/marvell/mv643xx_eth.c struct tx_queue *txq = mp->txq + i; txq 2208 drivers/net/ethernet/marvell/mv643xx_eth.c txq_reclaim(txq, txq->tx_ring_size, 1); txq 2209 drivers/net/ethernet/marvell/mv643xx_eth.c txq_reset_hw_ptr(txq); txq 2282 drivers/net/ethernet/marvell/mv643xx_eth.c txq_kick(mp->txq + queue); txq 2284 drivers/net/ethernet/marvell/mv643xx_eth.c work_done += txq_reclaim(mp->txq + queue, work_tbd, 0); txq 2285 drivers/net/ethernet/marvell/mv643xx_eth.c txq_maybe_wake(mp->txq + queue); txq 2349 drivers/net/ethernet/marvell/mv643xx_eth.c struct tx_queue *txq = mp->txq + i; txq 2351 drivers/net/ethernet/marvell/mv643xx_eth.c txq_reset_hw_ptr(txq); txq 2352 drivers/net/ethernet/marvell/mv643xx_eth.c txq_set_rate(txq, 1000000000, 16777216); txq 2353 drivers/net/ethernet/marvell/mv643xx_eth.c txq_set_fixed_prio_mode(txq); txq 2460 drivers/net/ethernet/marvell/mv643xx_eth.c txq_deinit(mp->txq + i); txq 2492 drivers/net/ethernet/marvell/mv643xx_eth.c txq_disable(mp->txq + i); txq 2536 drivers/net/ethernet/marvell/mv643xx_eth.c txq_deinit(mp->txq + i); txq 126 drivers/net/ethernet/marvell/mvneta.c #define MVNETA_CPU_TXQ_ACCESS(txq) BIT(txq + 8) txq 325 drivers/net/ethernet/marvell/mvneta.c #define IS_TSO_HEADER(txq, addr) \ txq 326 drivers/net/ethernet/marvell/mvneta.c ((addr >= txq->tso_hdrs_phys) && \ txq 327 drivers/net/ethernet/marvell/mvneta.c (addr < txq->tso_hdrs_phys + txq->size * TSO_HEADER_SIZE)) txq 669 drivers/net/ethernet/marvell/mvneta.c static void mvneta_txq_inc_get(struct mvneta_tx_queue *txq) txq 671 drivers/net/ethernet/marvell/mvneta.c txq->txq_get_index++; txq 672 drivers/net/ethernet/marvell/mvneta.c if (txq->txq_get_index == txq->size) txq 673 drivers/net/ethernet/marvell/mvneta.c txq->txq_get_index = 0; txq 677 drivers/net/ethernet/marvell/mvneta.c static void mvneta_txq_inc_put(struct mvneta_tx_queue *txq) txq 679 drivers/net/ethernet/marvell/mvneta.c txq->txq_put_index++; txq 680 drivers/net/ethernet/marvell/mvneta.c if (txq->txq_put_index == txq->size) txq 681 drivers/net/ethernet/marvell/mvneta.c txq->txq_put_index = 0; txq 860 drivers/net/ethernet/marvell/mvneta.c struct mvneta_tx_queue *txq, txq 865 drivers/net/ethernet/marvell/mvneta.c pend_desc += txq->pending; txq 870 drivers/net/ethernet/marvell/mvneta.c mvreg_write(pp, MVNETA_TXQ_UPDATE_REG(txq->id), val); txq 873 drivers/net/ethernet/marvell/mvneta.c txq->pending = 0; txq 878 drivers/net/ethernet/marvell/mvneta.c mvneta_txq_next_desc_get(struct mvneta_tx_queue *txq) txq 880 drivers/net/ethernet/marvell/mvneta.c int tx_desc = txq->next_desc_to_proc; txq 882 drivers/net/ethernet/marvell/mvneta.c txq->next_desc_to_proc = MVNETA_QUEUE_NEXT_DESC(txq, tx_desc); txq 883 drivers/net/ethernet/marvell/mvneta.c return txq->descs + tx_desc; txq 889 drivers/net/ethernet/marvell/mvneta.c static void mvneta_txq_desc_put(struct mvneta_tx_queue *txq) txq 891 drivers/net/ethernet/marvell/mvneta.c if (txq->next_desc_to_proc == 0) txq 892 drivers/net/ethernet/marvell/mvneta.c txq->next_desc_to_proc = txq->last_desc - 1; txq 894 drivers/net/ethernet/marvell/mvneta.c txq->next_desc_to_proc--; txq 1155 drivers/net/ethernet/marvell/mvneta.c struct mvneta_tx_queue *txq = &pp->txqs[queue]; txq 1156 drivers/net/ethernet/marvell/mvneta.c if (txq->descs) txq 1393 drivers/net/ethernet/marvell/mvneta.c int rxq, txq; txq 1399 drivers/net/ethernet/marvell/mvneta.c for (txq = 0; txq < txq_number; txq++) txq 1400 drivers/net/ethernet/marvell/mvneta.c if ((txq % max_cpu) == cpu) txq 1401 drivers/net/ethernet/marvell/mvneta.c txq_map |= MVNETA_CPU_TXQ_ACCESS(txq); txq 1604 drivers/net/ethernet/marvell/mvneta.c struct mvneta_tx_queue *txq, u32 value) txq 1608 drivers/net/ethernet/marvell/mvneta.c val = mvreg_read(pp, MVNETA_TXQ_SIZE_REG(txq->id)); txq 1613 drivers/net/ethernet/marvell/mvneta.c mvreg_write(pp, MVNETA_TXQ_SIZE_REG(txq->id), val); txq 1630 drivers/net/ethernet/marvell/mvneta.c struct mvneta_tx_queue *txq, txq 1638 drivers/net/ethernet/marvell/mvneta.c mvreg_write(pp, MVNETA_TXQ_UPDATE_REG(txq->id), val); txq 1643 drivers/net/ethernet/marvell/mvneta.c mvreg_write(pp, MVNETA_TXQ_UPDATE_REG(txq->id), val); txq 1648 drivers/net/ethernet/marvell/mvneta.c struct mvneta_tx_queue *txq) txq 1653 drivers/net/ethernet/marvell/mvneta.c val = mvreg_read(pp, MVNETA_TXQ_STATUS_REG(txq->id)); txq 1664 drivers/net/ethernet/marvell/mvneta.c struct mvneta_tx_queue *txq) txq 1669 drivers/net/ethernet/marvell/mvneta.c sent_desc = mvneta_txq_sent_desc_num_get(pp, txq); txq 1673 drivers/net/ethernet/marvell/mvneta.c mvneta_txq_sent_desc_dec(pp, txq, sent_desc); txq 1768 drivers/net/ethernet/marvell/mvneta.c struct mvneta_tx_queue *txq, int num, txq 1775 drivers/net/ethernet/marvell/mvneta.c struct mvneta_tx_desc *tx_desc = txq->descs + txq 1776 drivers/net/ethernet/marvell/mvneta.c txq->txq_get_index; txq 1777 drivers/net/ethernet/marvell/mvneta.c struct sk_buff *skb = txq->tx_skb[txq->txq_get_index]; txq 1784 drivers/net/ethernet/marvell/mvneta.c mvneta_txq_inc_get(txq); txq 1786 drivers/net/ethernet/marvell/mvneta.c if (!IS_TSO_HEADER(txq, tx_desc->buf_phys_addr)) txq 1800 drivers/net/ethernet/marvell/mvneta.c struct mvneta_tx_queue *txq) txq 1802 drivers/net/ethernet/marvell/mvneta.c struct netdev_queue *nq = netdev_get_tx_queue(pp->dev, txq->id); txq 1805 drivers/net/ethernet/marvell/mvneta.c tx_done = mvneta_txq_sent_desc_proc(pp, txq); txq 1809 drivers/net/ethernet/marvell/mvneta.c mvneta_txq_bufs_free(pp, txq, tx_done, nq); txq 1811 drivers/net/ethernet/marvell/mvneta.c txq->count -= tx_done; txq 1814 drivers/net/ethernet/marvell/mvneta.c if (txq->count <= txq->tx_wake_threshold) txq 2241 drivers/net/ethernet/marvell/mvneta.c struct mvneta_port *pp, struct mvneta_tx_queue *txq) txq 2246 drivers/net/ethernet/marvell/mvneta.c txq->tx_skb[txq->txq_put_index] = NULL; txq 2247 drivers/net/ethernet/marvell/mvneta.c tx_desc = mvneta_txq_next_desc_get(txq); txq 2251 drivers/net/ethernet/marvell/mvneta.c tx_desc->buf_phys_addr = txq->tso_hdrs_phys + txq 2252 drivers/net/ethernet/marvell/mvneta.c txq->txq_put_index * TSO_HEADER_SIZE; txq 2253 drivers/net/ethernet/marvell/mvneta.c mvneta_txq_inc_put(txq); txq 2257 drivers/net/ethernet/marvell/mvneta.c mvneta_tso_put_data(struct net_device *dev, struct mvneta_tx_queue *txq, txq 2263 drivers/net/ethernet/marvell/mvneta.c tx_desc = mvneta_txq_next_desc_get(txq); txq 2269 drivers/net/ethernet/marvell/mvneta.c mvneta_txq_desc_put(txq); txq 2274 drivers/net/ethernet/marvell/mvneta.c txq->tx_skb[txq->txq_put_index] = NULL; txq 2282 drivers/net/ethernet/marvell/mvneta.c txq->tx_skb[txq->txq_put_index] = skb; txq 2284 drivers/net/ethernet/marvell/mvneta.c mvneta_txq_inc_put(txq); txq 2289 drivers/net/ethernet/marvell/mvneta.c struct mvneta_tx_queue *txq) txq 2299 drivers/net/ethernet/marvell/mvneta.c if ((txq->count + tso_count_descs(skb)) >= txq->size) txq 2319 drivers/net/ethernet/marvell/mvneta.c hdr = txq->tso_hdrs + txq->txq_put_index * TSO_HEADER_SIZE; txq 2322 drivers/net/ethernet/marvell/mvneta.c mvneta_tso_put_hdr(skb, pp, txq); txq 2330 drivers/net/ethernet/marvell/mvneta.c if (mvneta_tso_put_data(dev, txq, skb, txq 2348 drivers/net/ethernet/marvell/mvneta.c struct mvneta_tx_desc *tx_desc = txq->descs + i; txq 2349 drivers/net/ethernet/marvell/mvneta.c if (!IS_TSO_HEADER(txq, tx_desc->buf_phys_addr)) txq 2354 drivers/net/ethernet/marvell/mvneta.c mvneta_txq_desc_put(txq); txq 2361 drivers/net/ethernet/marvell/mvneta.c struct mvneta_tx_queue *txq) txq 2370 drivers/net/ethernet/marvell/mvneta.c tx_desc = mvneta_txq_next_desc_get(txq); txq 2379 drivers/net/ethernet/marvell/mvneta.c mvneta_txq_desc_put(txq); txq 2386 drivers/net/ethernet/marvell/mvneta.c txq->tx_skb[txq->txq_put_index] = skb; txq 2390 drivers/net/ethernet/marvell/mvneta.c txq->tx_skb[txq->txq_put_index] = NULL; txq 2392 drivers/net/ethernet/marvell/mvneta.c mvneta_txq_inc_put(txq); txq 2402 drivers/net/ethernet/marvell/mvneta.c tx_desc = txq->descs + i; txq 2407 drivers/net/ethernet/marvell/mvneta.c mvneta_txq_desc_put(txq); txq 2418 drivers/net/ethernet/marvell/mvneta.c struct mvneta_tx_queue *txq = &pp->txqs[txq_id]; txq 2428 drivers/net/ethernet/marvell/mvneta.c frags = mvneta_tx_tso(skb, dev, txq); txq 2435 drivers/net/ethernet/marvell/mvneta.c tx_desc = mvneta_txq_next_desc_get(txq); txq 2446 drivers/net/ethernet/marvell/mvneta.c mvneta_txq_desc_put(txq); txq 2455 drivers/net/ethernet/marvell/mvneta.c txq->tx_skb[txq->txq_put_index] = skb; txq 2456 drivers/net/ethernet/marvell/mvneta.c mvneta_txq_inc_put(txq); txq 2460 drivers/net/ethernet/marvell/mvneta.c txq->tx_skb[txq->txq_put_index] = NULL; txq 2461 drivers/net/ethernet/marvell/mvneta.c mvneta_txq_inc_put(txq); txq 2464 drivers/net/ethernet/marvell/mvneta.c if (mvneta_tx_frag_process(pp, skb, txq)) { txq 2469 drivers/net/ethernet/marvell/mvneta.c mvneta_txq_desc_put(txq); txq 2482 drivers/net/ethernet/marvell/mvneta.c txq->count += frags; txq 2483 drivers/net/ethernet/marvell/mvneta.c if (txq->count >= txq->tx_stop_threshold) txq 2487 drivers/net/ethernet/marvell/mvneta.c txq->pending + frags > MVNETA_TXQ_DEC_SENT_MASK) txq 2488 drivers/net/ethernet/marvell/mvneta.c mvneta_txq_pend_desc_add(pp, txq, frags); txq 2490 drivers/net/ethernet/marvell/mvneta.c txq->pending += frags; txq 2507 drivers/net/ethernet/marvell/mvneta.c struct mvneta_tx_queue *txq) txq 2510 drivers/net/ethernet/marvell/mvneta.c struct netdev_queue *nq = netdev_get_tx_queue(pp->dev, txq->id); txq 2511 drivers/net/ethernet/marvell/mvneta.c int tx_done = txq->count; txq 2513 drivers/net/ethernet/marvell/mvneta.c mvneta_txq_bufs_free(pp, txq, tx_done, nq); txq 2516 drivers/net/ethernet/marvell/mvneta.c txq->count = 0; txq 2517 drivers/net/ethernet/marvell/mvneta.c txq->txq_put_index = 0; txq 2518 drivers/net/ethernet/marvell/mvneta.c txq->txq_get_index = 0; txq 2526 drivers/net/ethernet/marvell/mvneta.c struct mvneta_tx_queue *txq; txq 2531 drivers/net/ethernet/marvell/mvneta.c txq = mvneta_tx_done_policy(pp, cause_tx_done); txq 2533 drivers/net/ethernet/marvell/mvneta.c nq = netdev_get_tx_queue(pp->dev, txq->id); txq 2536 drivers/net/ethernet/marvell/mvneta.c if (txq->count) txq 2537 drivers/net/ethernet/marvell/mvneta.c mvneta_txq_done(pp, txq); txq 2540 drivers/net/ethernet/marvell/mvneta.c cause_tx_done &= ~((1 << txq->id)); txq 2984 drivers/net/ethernet/marvell/mvneta.c struct mvneta_tx_queue *txq) txq 2988 drivers/net/ethernet/marvell/mvneta.c txq->size = pp->tx_ring_size; txq 2994 drivers/net/ethernet/marvell/mvneta.c txq->tx_stop_threshold = txq->size - MVNETA_MAX_SKB_DESCS; txq 2995 drivers/net/ethernet/marvell/mvneta.c txq->tx_wake_threshold = txq->tx_stop_threshold / 2; txq 2998 drivers/net/ethernet/marvell/mvneta.c txq->descs = dma_alloc_coherent(pp->dev->dev.parent, txq 2999 drivers/net/ethernet/marvell/mvneta.c txq->size * MVNETA_DESC_ALIGNED_SIZE, txq 3000 drivers/net/ethernet/marvell/mvneta.c &txq->descs_phys, GFP_KERNEL); txq 3001 drivers/net/ethernet/marvell/mvneta.c if (!txq->descs) txq 3004 drivers/net/ethernet/marvell/mvneta.c txq->last_desc = txq->size - 1; txq 3006 drivers/net/ethernet/marvell/mvneta.c txq->tx_skb = kmalloc_array(txq->size, sizeof(*txq->tx_skb), txq 3008 drivers/net/ethernet/marvell/mvneta.c if (!txq->tx_skb) { txq 3010 drivers/net/ethernet/marvell/mvneta.c txq->size * MVNETA_DESC_ALIGNED_SIZE, txq 3011 drivers/net/ethernet/marvell/mvneta.c txq->descs, txq->descs_phys); txq 3016 drivers/net/ethernet/marvell/mvneta.c txq->tso_hdrs = dma_alloc_coherent(pp->dev->dev.parent, txq 3017 drivers/net/ethernet/marvell/mvneta.c txq->size * TSO_HEADER_SIZE, txq 3018 drivers/net/ethernet/marvell/mvneta.c &txq->tso_hdrs_phys, GFP_KERNEL); txq 3019 drivers/net/ethernet/marvell/mvneta.c if (!txq->tso_hdrs) { txq 3020 drivers/net/ethernet/marvell/mvneta.c kfree(txq->tx_skb); txq 3022 drivers/net/ethernet/marvell/mvneta.c txq->size * MVNETA_DESC_ALIGNED_SIZE, txq 3023 drivers/net/ethernet/marvell/mvneta.c txq->descs, txq->descs_phys); txq 3029 drivers/net/ethernet/marvell/mvneta.c cpu = txq->id % num_present_cpus(); txq 3032 drivers/net/ethernet/marvell/mvneta.c cpumask_set_cpu(cpu, &txq->affinity_mask); txq 3033 drivers/net/ethernet/marvell/mvneta.c netif_set_xps_queue(pp->dev, &txq->affinity_mask, txq->id); txq 3039 drivers/net/ethernet/marvell/mvneta.c struct mvneta_tx_queue *txq) txq 3042 drivers/net/ethernet/marvell/mvneta.c mvreg_write(pp, MVETH_TXQ_TOKEN_CFG_REG(txq->id), 0x03ffffff); txq 3043 drivers/net/ethernet/marvell/mvneta.c mvreg_write(pp, MVETH_TXQ_TOKEN_COUNT_REG(txq->id), 0x3fffffff); txq 3046 drivers/net/ethernet/marvell/mvneta.c mvreg_write(pp, MVNETA_TXQ_BASE_ADDR_REG(txq->id), txq->descs_phys); txq 3047 drivers/net/ethernet/marvell/mvneta.c mvreg_write(pp, MVNETA_TXQ_SIZE_REG(txq->id), txq->size); txq 3049 drivers/net/ethernet/marvell/mvneta.c mvneta_tx_done_pkts_coal_set(pp, txq, txq->done_pkts_coal); txq 3054 drivers/net/ethernet/marvell/mvneta.c struct mvneta_tx_queue *txq) txq 3058 drivers/net/ethernet/marvell/mvneta.c ret = mvneta_txq_sw_init(pp, txq); txq 3062 drivers/net/ethernet/marvell/mvneta.c mvneta_txq_hw_init(pp, txq); txq 3069 drivers/net/ethernet/marvell/mvneta.c struct mvneta_tx_queue *txq) txq 3071 drivers/net/ethernet/marvell/mvneta.c struct netdev_queue *nq = netdev_get_tx_queue(pp->dev, txq->id); txq 3073 drivers/net/ethernet/marvell/mvneta.c kfree(txq->tx_skb); txq 3075 drivers/net/ethernet/marvell/mvneta.c if (txq->tso_hdrs) txq 3077 drivers/net/ethernet/marvell/mvneta.c txq->size * TSO_HEADER_SIZE, txq 3078 drivers/net/ethernet/marvell/mvneta.c txq->tso_hdrs, txq->tso_hdrs_phys); txq 3079 drivers/net/ethernet/marvell/mvneta.c if (txq->descs) txq 3081 drivers/net/ethernet/marvell/mvneta.c txq->size * MVNETA_DESC_ALIGNED_SIZE, txq 3082 drivers/net/ethernet/marvell/mvneta.c txq->descs, txq->descs_phys); txq 3086 drivers/net/ethernet/marvell/mvneta.c txq->descs = NULL; txq 3087 drivers/net/ethernet/marvell/mvneta.c txq->last_desc = 0; txq 3088 drivers/net/ethernet/marvell/mvneta.c txq->next_desc_to_proc = 0; txq 3089 drivers/net/ethernet/marvell/mvneta.c txq->descs_phys = 0; txq 3093 drivers/net/ethernet/marvell/mvneta.c struct mvneta_tx_queue *txq) txq 3096 drivers/net/ethernet/marvell/mvneta.c mvreg_write(pp, MVETH_TXQ_TOKEN_CFG_REG(txq->id), 0); txq 3097 drivers/net/ethernet/marvell/mvneta.c mvreg_write(pp, MVETH_TXQ_TOKEN_COUNT_REG(txq->id), 0); txq 3100 drivers/net/ethernet/marvell/mvneta.c mvreg_write(pp, MVNETA_TXQ_BASE_ADDR_REG(txq->id), 0); txq 3101 drivers/net/ethernet/marvell/mvneta.c mvreg_write(pp, MVNETA_TXQ_SIZE_REG(txq->id), 0); txq 3105 drivers/net/ethernet/marvell/mvneta.c struct mvneta_tx_queue *txq) txq 3107 drivers/net/ethernet/marvell/mvneta.c mvneta_txq_sw_deinit(pp, txq); txq 3108 drivers/net/ethernet/marvell/mvneta.c mvneta_txq_hw_deinit(pp, txq); txq 3994 drivers/net/ethernet/marvell/mvneta.c struct mvneta_tx_queue *txq = &pp->txqs[queue]; txq 3995 drivers/net/ethernet/marvell/mvneta.c txq->done_pkts_coal = c->tx_max_coalesced_frames; txq 3996 drivers/net/ethernet/marvell/mvneta.c mvneta_tx_done_pkts_coal_set(pp, txq, txq->done_pkts_coal); txq 4389 drivers/net/ethernet/marvell/mvneta.c struct mvneta_tx_queue *txq = &pp->txqs[queue]; txq 4390 drivers/net/ethernet/marvell/mvneta.c txq->id = queue; txq 4391 drivers/net/ethernet/marvell/mvneta.c txq->size = pp->tx_ring_size; txq 4392 drivers/net/ethernet/marvell/mvneta.c txq->done_pkts_coal = MVNETA_TXDONE_COAL_PKTS; txq 4788 drivers/net/ethernet/marvell/mvneta.c struct mvneta_tx_queue *txq = &pp->txqs[queue]; txq 4790 drivers/net/ethernet/marvell/mvneta.c mvneta_txq_hw_deinit(pp, txq); txq 4840 drivers/net/ethernet/marvell/mvneta.c struct mvneta_tx_queue *txq = &pp->txqs[queue]; txq 4842 drivers/net/ethernet/marvell/mvneta.c txq->next_desc_to_proc = 0; txq 4843 drivers/net/ethernet/marvell/mvneta.c mvneta_txq_hw_init(pp, txq); txq 189 drivers/net/ethernet/marvell/mvpp2/mvpp2.h #define MVPP2_TXQ_SENT_REG(txq) (0x3c00 + 4 * (txq)) txq 338 drivers/net/ethernet/marvell/mvpp2/mvpp2.h #define MVPP22_CTRS_TX_CTR(port, txq) ((txq) | ((port) << 3) | BIT(7)) txq 207 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c unsigned int txq) txq 210 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c tx_desc->pp21.phys_txq = txq; txq 212 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c tx_desc->pp22.phys_txq = txq; txq 322 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c static inline int mvpp2_txq_phys(int port, int txq) txq 324 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c return (MVPP2_MAX_TCONT + port) * MVPP2_MAX_TXQ + txq; txq 1791 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c struct mvpp2_tx_queue *txq = port->txqs[queue]; txq 1793 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c if (txq->descs) txq 1897 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c mvpp2_txq_next_desc_get(struct mvpp2_tx_queue *txq) txq 1899 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c int tx_desc = txq->next_desc_to_proc; txq 1901 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c txq->next_desc_to_proc = MVPP2_QUEUE_NEXT_DESC(txq, tx_desc); txq 1902 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c return txq->descs + tx_desc; txq 1949 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c struct mvpp2_tx_queue *txq, int num) txq 1955 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c val = (txq->id << MVPP2_TXQ_RSVD_REQ_Q_OFFSET) | num; txq 1967 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c struct mvpp2_tx_queue *txq, txq 1986 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c txq_pcpu_aux = per_cpu_ptr(txq->pcpu, thread); txq 1995 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c (txq->size - (MVPP2_MAX_THREADS * MVPP2_CPU_DESC_CHUNK))) txq 1998 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c txq_pcpu->reserved_num += mvpp2_txq_alloc_reserved_desc(port, txq, req); txq 2009 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c static void mvpp2_txq_desc_put(struct mvpp2_tx_queue *txq) txq 2011 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c if (txq->next_desc_to_proc == 0) txq 2012 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c txq->next_desc_to_proc = txq->last_desc - 1; txq 2014 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c txq->next_desc_to_proc--; txq 2059 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c struct mvpp2_tx_queue *txq) txq 2066 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c MVPP2_TXQ_SENT_REG(txq->id)); txq 2097 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c int txq, tx_port_num; txq 2126 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c for (txq = 0; txq < port->ntxqs; txq++) { txq 2128 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c MVPP2_TXQ_SCHED_TOKEN_SIZE_REG(txq)); txq 2136 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c MVPP2_TXQ_SCHED_TOKEN_SIZE_REG(txq), txq 2162 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c struct mvpp2_tx_queue *txq) txq 2167 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c if (txq->done_pkts_coal > MVPP2_TXQ_THRESH_MASK) txq 2168 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c txq->done_pkts_coal = MVPP2_TXQ_THRESH_MASK; txq 2170 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c val = (txq->done_pkts_coal << MVPP2_TXQ_THRESH_OFFSET); txq 2171 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c mvpp2_thread_write(port->priv, thread, MVPP2_TXQ_NUM_REG, txq->id); txq 2231 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c struct mvpp2_tx_queue *txq, txq 2267 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c static void mvpp2_txq_done(struct mvpp2_port *port, struct mvpp2_tx_queue *txq, txq 2270 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c struct netdev_queue *nq = netdev_get_tx_queue(port->dev, txq->log_id); txq 2276 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c tx_done = mvpp2_txq_sent_desc_proc(port, txq); txq 2279 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c mvpp2_txq_bufs_free(port, txq, txq_pcpu, tx_done); txq 2291 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c struct mvpp2_tx_queue *txq; txq 2296 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c txq = mvpp2_get_tx_queue(port, cause); txq 2297 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c if (!txq) txq 2300 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c txq_pcpu = per_cpu_ptr(txq->pcpu, thread); txq 2303 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c mvpp2_txq_done(port, txq, txq_pcpu); txq 2307 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c cause &= ~(1 << txq->log_id); txq 2454 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c struct mvpp2_tx_queue *txq) txq 2461 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c txq->size = port->tx_ring_size; txq 2464 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c txq->descs = dma_alloc_coherent(port->dev->dev.parent, txq 2465 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c txq->size * MVPP2_DESC_ALIGNED_SIZE, txq 2466 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c &txq->descs_dma, GFP_KERNEL); txq 2467 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c if (!txq->descs) txq 2470 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c txq->last_desc = txq->size - 1; txq 2474 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c mvpp2_thread_write(port->priv, thread, MVPP2_TXQ_NUM_REG, txq->id); txq 2476 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c txq->descs_dma); txq 2478 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c txq->size & MVPP2_TXQ_DESC_SIZE_MASK); txq 2481 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c txq->id << MVPP2_TXQ_RSVD_CLR_OFFSET); txq 2493 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c (txq->log_id * desc_per_txq); txq 2504 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c val = mvpp2_read(port->priv, MVPP2_TXQ_SCHED_REFILL_REG(txq->log_id)); txq 2508 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c mvpp2_write(port->priv, MVPP2_TXQ_SCHED_REFILL_REG(txq->log_id), val); txq 2511 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c mvpp2_write(port->priv, MVPP2_TXQ_SCHED_TOKEN_SIZE_REG(txq->log_id), txq 2515 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c txq_pcpu = per_cpu_ptr(txq->pcpu, thread); txq 2516 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c txq_pcpu->size = txq->size; txq 2529 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c txq_pcpu->stop_threshold = txq->size - MVPP2_MAX_SKB_DESCS; txq 2546 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c struct mvpp2_tx_queue *txq) txq 2552 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c txq_pcpu = per_cpu_ptr(txq->pcpu, thread); txq 2564 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c if (txq->descs) txq 2566 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c txq->size * MVPP2_DESC_ALIGNED_SIZE, txq 2567 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c txq->descs, txq->descs_dma); txq 2569 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c txq->descs = NULL; txq 2570 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c txq->last_desc = 0; txq 2571 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c txq->next_desc_to_proc = 0; txq 2572 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c txq->descs_dma = 0; txq 2575 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c mvpp2_write(port->priv, MVPP2_TXQ_SCHED_TOKEN_CNTR_REG(txq->log_id), 0); txq 2579 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c mvpp2_thread_write(port->priv, thread, MVPP2_TXQ_NUM_REG, txq->id); txq 2586 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c static void mvpp2_txq_clean(struct mvpp2_port *port, struct mvpp2_tx_queue *txq) txq 2593 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c mvpp2_thread_write(port->priv, thread, MVPP2_TXQ_NUM_REG, txq->id); txq 2606 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c port->id, txq->log_id); txq 2622 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c txq_pcpu = per_cpu_ptr(txq->pcpu, thread); txq 2625 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c mvpp2_txq_bufs_free(port, txq, txq_pcpu, txq_pcpu->count); txq 2637 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c struct mvpp2_tx_queue *txq; txq 2648 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c txq = port->txqs[queue]; txq 2649 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c mvpp2_txq_clean(port, txq); txq 2650 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c mvpp2_txq_deinit(port, txq); txq 2688 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c struct mvpp2_tx_queue *txq; txq 2692 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c txq = port->txqs[queue]; txq 2693 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c err = mvpp2_txq_init(port, txq); txq 2705 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c txq = port->txqs[queue]; txq 2706 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c mvpp2_tx_pkts_coal_set(port, txq); txq 3017 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c tx_desc_unmap_put(struct mvpp2_port *port, struct mvpp2_tx_queue *txq, txq 3021 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c struct mvpp2_txq_pcpu *txq_pcpu = per_cpu_ptr(txq->pcpu, thread); txq 3030 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c mvpp2_txq_desc_put(txq); txq 3036 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c struct mvpp2_tx_queue *txq) txq 3039 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c struct mvpp2_txq_pcpu *txq_pcpu = per_cpu_ptr(txq->pcpu, thread); txq 3049 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c mvpp2_txdesc_txq_set(port, tx_desc, txq->id); txq 3056 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c mvpp2_txq_desc_put(txq); txq 3080 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c tx_desc = txq->descs + i; txq 3081 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c tx_desc_unmap_put(port, txq, tx_desc); txq 3089 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c struct mvpp2_tx_queue *txq, txq 3098 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c mvpp2_txdesc_txq_set(port, tx_desc, txq->id); txq 3113 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c struct mvpp2_tx_queue *txq, txq 3122 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c mvpp2_txdesc_txq_set(port, tx_desc, txq->id); txq 3128 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c mvpp2_txq_desc_put(txq); txq 3149 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c struct mvpp2_tx_queue *txq, txq 3160 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c mvpp2_txq_reserved_desc_num_proc(port, txq, txq_pcpu, txq 3175 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c mvpp2_tso_put_hdr(skb, dev, txq, aggr_txq, txq_pcpu, hdr_sz); txq 3182 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c if (mvpp2_tso_put_data(skb, dev, &tso, txq, aggr_txq, txq 3193 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c struct mvpp2_tx_desc *tx_desc = txq->descs + i; txq 3194 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c tx_desc_unmap_put(port, txq, tx_desc); txq 3203 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c struct mvpp2_tx_queue *txq, *aggr_txq; txq 3216 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c txq = port->txqs[txq_id]; txq 3217 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c txq_pcpu = per_cpu_ptr(txq->pcpu, thread); txq 3224 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c frags = mvpp2_tx_tso(skb, dev, txq, aggr_txq, txq_pcpu); txq 3231 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c mvpp2_txq_reserved_desc_num_proc(port, txq, txq_pcpu, frags)) { txq 3238 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c mvpp2_txdesc_txq_set(port, tx_desc, txq->id); txq 3244 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c mvpp2_txq_desc_put(txq); txq 3265 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c if (mvpp2_tx_frag_process(port, skb, aggr_txq, txq)) { txq 3266 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c tx_desc_unmap_put(port, txq, tx_desc); txq 3297 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c if (!port->has_tx_irqs && txq_pcpu->count >= txq->done_pkts_coal) txq 3298 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c mvpp2_txq_done(port, txq, txq_pcpu); txq 4060 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c struct mvpp2_tx_queue *txq = port->txqs[queue]; txq 4062 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c txq->done_pkts_coal = c->tx_max_coalesced_frames; txq 4065 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c mvpp2_tx_pkts_coal_set(port, txq); txq 4572 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c struct mvpp2_tx_queue *txq; txq 4574 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c txq = devm_kzalloc(dev, sizeof(*txq), GFP_KERNEL); txq 4575 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c if (!txq) { txq 4580 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c txq->pcpu = alloc_percpu(struct mvpp2_txq_pcpu); txq 4581 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c if (!txq->pcpu) { txq 4586 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c txq->id = queue_phy_id; txq 4587 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c txq->log_id = queue; txq 4588 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c txq->done_pkts_coal = MVPP2_TXDONE_COAL_PKTS_THRESH; txq 4590 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c txq_pcpu = per_cpu_ptr(txq->pcpu, thread); txq 4594 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c port->txqs[queue] = txq; txq 408 drivers/net/ethernet/mellanox/mlx5/core/en.h struct netdev_queue *txq; txq 1071 drivers/net/ethernet/mellanox/mlx5/core/en.h void mlx5e_tx_disable_queue(struct netdev_queue *txq); txq 59 drivers/net/ethernet/mellanox/mlx5/core/en/reporter_tx.c mlx5e_tx_disable_queue(sq->txq); txq 124 drivers/net/ethernet/mellanox/mlx5/core/en/reporter_tx.c jiffies_to_usecs(jiffies - sq->txq->trans_start)); txq 152 drivers/net/ethernet/mellanox/mlx5/core/en/reporter_tx.c bool stopped = netif_xmit_stopped(sq->txq); txq 1350 drivers/net/ethernet/mellanox/mlx5/core/en_main.c sq->txq = netdev_get_tx_queue(sq->channel->netdev, sq->txq_ix); txq 1352 drivers/net/ethernet/mellanox/mlx5/core/en_main.c netdev_tx_reset_queue(sq->txq); txq 1353 drivers/net/ethernet/mellanox/mlx5/core/en_main.c netif_tx_start_queue(sq->txq); txq 1356 drivers/net/ethernet/mellanox/mlx5/core/en_main.c void mlx5e_tx_disable_queue(struct netdev_queue *txq) txq 1358 drivers/net/ethernet/mellanox/mlx5/core/en_main.c __netif_tx_lock_bh(txq); txq 1359 drivers/net/ethernet/mellanox/mlx5/core/en_main.c netif_tx_stop_queue(txq); txq 1360 drivers/net/ethernet/mellanox/mlx5/core/en_main.c __netif_tx_unlock_bh(txq); txq 1372 drivers/net/ethernet/mellanox/mlx5/core/en_main.c mlx5e_tx_disable_queue(sq->txq); txq 259 drivers/net/ethernet/mellanox/mlx5/core/en_tx.c netif_tx_stop_queue(sq->txq); txq 263 drivers/net/ethernet/mellanox/mlx5/core/en_tx.c send_doorbell = __netdev_tx_sent_queue(sq->txq, num_bytes, txq 525 drivers/net/ethernet/mellanox/mlx5/core/en_tx.c netdev_tx_completed_queue(sq->txq, npkts, nbytes); txq 527 drivers/net/ethernet/mellanox/mlx5/core/en_tx.c if (netif_tx_queue_stopped(sq->txq) && txq 530 drivers/net/ethernet/mellanox/mlx5/core/en_tx.c netif_tx_wake_queue(sq->txq); txq 575 drivers/net/ethernet/mellanox/mlx5/core/en_tx.c netdev_tx_completed_queue(sq->txq, npkts, nbytes); txq 127 drivers/net/ethernet/micrel/ks8851.c struct sk_buff_head txq; txq 760 drivers/net/ethernet/micrel/ks8851.c bool last = skb_queue_empty(&ks->txq); txq 765 drivers/net/ethernet/micrel/ks8851.c txb = skb_dequeue(&ks->txq); txq 766 drivers/net/ethernet/micrel/ks8851.c last = skb_queue_empty(&ks->txq); txq 906 drivers/net/ethernet/micrel/ks8851.c while (!skb_queue_empty(&ks->txq)) { txq 907 drivers/net/ethernet/micrel/ks8851.c struct sk_buff *txb = skb_dequeue(&ks->txq); txq 950 drivers/net/ethernet/micrel/ks8851.c skb_queue_tail(&ks->txq, skb); txq 1519 drivers/net/ethernet/micrel/ks8851.c skb_queue_head_init(&ks->txq); txq 109 drivers/net/ethernet/neterion/vxge/vxge-main.c if (__netif_tx_trylock(fifo->txq)) { txq 112 drivers/net/ethernet/neterion/vxge/vxge-main.c __netif_tx_unlock(fifo->txq); txq 621 drivers/net/ethernet/neterion/vxge/vxge-main.c if (netif_tx_queue_stopped(fifo->txq)) txq 622 drivers/net/ethernet/neterion/vxge/vxge-main.c netif_tx_wake_queue(fifo->txq); txq 863 drivers/net/ethernet/neterion/vxge/vxge-main.c if (netif_tx_queue_stopped(fifo->txq)) txq 878 drivers/net/ethernet/neterion/vxge/vxge-main.c netif_tx_stop_queue(fifo->txq); txq 991 drivers/net/ethernet/neterion/vxge/vxge-main.c netif_tx_stop_queue(fifo->txq); txq 1569 drivers/net/ethernet/neterion/vxge/vxge-main.c if (netif_tx_queue_stopped(vpath->fifo.txq)) txq 1570 drivers/net/ethernet/neterion/vxge/vxge-main.c netif_tx_wake_queue(vpath->fifo.txq); txq 2077 drivers/net/ethernet/neterion/vxge/vxge-main.c vpath->fifo.txq = txq 2080 drivers/net/ethernet/neterion/vxge/vxge-main.c vpath->fifo.txq = txq 2638 drivers/net/ethernet/neterion/vxge/vxge-main.c netif_tx_stop_queue(vpath->fifo.txq); txq 3600 drivers/net/ethernet/neterion/vxge/vxge-main.c netif_tx_stop_queue(vpath->fifo.txq); txq 240 drivers/net/ethernet/neterion/vxge/vxge-main.h struct netdev_queue *txq; txq 962 drivers/net/ethernet/nxp/lpc_eth.c struct netdev_queue *txq = netdev_get_tx_queue(ndev, 0); txq 964 drivers/net/ethernet/nxp/lpc_eth.c __netif_tx_lock(txq, smp_processor_id()); txq 966 drivers/net/ethernet/nxp/lpc_eth.c __netif_tx_unlock(txq); txq 648 drivers/net/ethernet/qlogic/netxen/netxen_nic.h struct netdev_queue *txq; txq 563 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c __netif_tx_lock_bh(tx_ring->txq); txq 568 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c netif_tx_stop_queue(tx_ring->txq); txq 572 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c netif_tx_wake_queue(tx_ring->txq); txq 574 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c __netif_tx_unlock_bh(tx_ring->txq); txq 596 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c __netif_tx_unlock_bh(tx_ring->txq); txq 197 drivers/net/ethernet/qlogic/netxen/netxen_nic_init.c tx_ring->txq = netdev_get_tx_queue(netdev, 0); txq 415 drivers/net/ethernet/qlogic/qede/qede.h #define QEDE_TXQ_XDP_TO_IDX(edev, txq) ((txq)->index - \ txq 421 drivers/net/ethernet/qlogic/qede/qede.h #define QEDE_TXQ_TO_NDEV_TXQ_ID(edev, txq) ((QEDE_TSS_COUNT(edev) * \ txq 422 drivers/net/ethernet/qlogic/qede/qede.h (txq)->cos) + (txq)->index) txq 424 drivers/net/ethernet/qlogic/qede/qede.h (&((edev)->fp_array[QEDE_NDEV_TXQ_ID_TO_FP_ID(edev, idx)].txq \ txq 426 drivers/net/ethernet/qlogic/qede/qede.h #define QEDE_FP_TC0_TXQ(fp) (&((fp)->txq[0])) txq 466 drivers/net/ethernet/qlogic/qede/qede.h struct qede_tx_queue *txq; txq 527 drivers/net/ethernet/qlogic/qede/qede.h struct qede_tx_queue *txq, int *len); txq 567 drivers/net/ethernet/qlogic/qede/qede.h int qede_txq_has_work(struct qede_tx_queue *txq); txq 221 drivers/net/ethernet/qlogic/qede/qede_ethtool.c struct qede_tx_queue *txq, u8 **buf) txq 226 drivers/net/ethernet/qlogic/qede/qede_ethtool.c if (txq->is_xdp) txq 228 drivers/net/ethernet/qlogic/qede/qede_ethtool.c QEDE_TXQ_XDP_TO_IDX(edev, txq), txq 231 drivers/net/ethernet/qlogic/qede/qede_ethtool.c sprintf(*buf, "%d_%d: %s", txq->index, txq->cos, txq 276 drivers/net/ethernet/qlogic/qede/qede_ethtool.c &fp->txq[cos], &buf); txq 311 drivers/net/ethernet/qlogic/qede/qede_ethtool.c static void qede_get_ethtool_stats_txq(struct qede_tx_queue *txq, u64 **buf) txq 316 drivers/net/ethernet/qlogic/qede/qede_ethtool.c **buf = *((u64 *)(((void *)txq) + qede_tqstats_arr[i].offset)); txq 356 drivers/net/ethernet/qlogic/qede/qede_ethtool.c qede_get_ethtool_stats_txq(&fp->txq[cos], &buf); txq 840 drivers/net/ethernet/qlogic/qede/qede_ethtool.c struct qede_tx_queue *txq; txq 849 drivers/net/ethernet/qlogic/qede/qede_ethtool.c txq = QEDE_FP_TC0_TXQ(fp); txq 850 drivers/net/ethernet/qlogic/qede/qede_ethtool.c tx_handle = txq->handle; txq 908 drivers/net/ethernet/qlogic/qede/qede_ethtool.c struct qede_tx_queue *txq; txq 914 drivers/net/ethernet/qlogic/qede/qede_ethtool.c txq = QEDE_FP_TC0_TXQ(fp); txq 918 drivers/net/ethernet/qlogic/qede/qede_ethtool.c txq->handle); txq 1493 drivers/net/ethernet/qlogic/qede/qede_ethtool.c struct qede_tx_queue *txq = NULL; txq 1503 drivers/net/ethernet/qlogic/qede/qede_ethtool.c txq = QEDE_FP_TC0_TXQ(fp); txq 1508 drivers/net/ethernet/qlogic/qede/qede_ethtool.c if (!txq) { txq 1514 drivers/net/ethernet/qlogic/qede/qede_ethtool.c idx = txq->sw_tx_prod; txq 1515 drivers/net/ethernet/qlogic/qede/qede_ethtool.c txq->sw_tx_ring.skbs[idx].skb = skb; txq 1516 drivers/net/ethernet/qlogic/qede/qede_ethtool.c first_bd = qed_chain_produce(&txq->tx_pbl); txq 1535 drivers/net/ethernet/qlogic/qede/qede_ethtool.c txq->sw_tx_prod = (txq->sw_tx_prod + 1) % txq->num_tx_buffers; txq 1537 drivers/net/ethernet/qlogic/qede/qede_ethtool.c val = qed_chain_get_prod_idx(&txq->tx_pbl); txq 1538 drivers/net/ethernet/qlogic/qede/qede_ethtool.c txq->tx_db.data.bd_prod = cpu_to_le16(val); txq 1545 drivers/net/ethernet/qlogic/qede/qede_ethtool.c writel(txq->tx_db.raw, txq->doorbell_addr); txq 1548 drivers/net/ethernet/qlogic/qede/qede_ethtool.c if (qede_txq_has_work(txq)) txq 1553 drivers/net/ethernet/qlogic/qede/qede_ethtool.c if (!qede_txq_has_work(txq)) { txq 1558 drivers/net/ethernet/qlogic/qede/qede_ethtool.c first_bd = (struct eth_tx_1st_bd *)qed_chain_consume(&txq->tx_pbl); txq 1561 drivers/net/ethernet/qlogic/qede/qede_ethtool.c txq->sw_tx_cons = (txq->sw_tx_cons + 1) % txq->num_tx_buffers; txq 1562 drivers/net/ethernet/qlogic/qede/qede_ethtool.c txq->sw_tx_ring.skbs[idx].skb = NULL; txq 100 drivers/net/ethernet/qlogic/qede/qede_fp.c int qede_free_tx_pkt(struct qede_dev *edev, struct qede_tx_queue *txq, int *len) txq 102 drivers/net/ethernet/qlogic/qede/qede_fp.c u16 idx = txq->sw_tx_cons; txq 103 drivers/net/ethernet/qlogic/qede/qede_fp.c struct sk_buff *skb = txq->sw_tx_ring.skbs[idx].skb; txq 108 drivers/net/ethernet/qlogic/qede/qede_fp.c bool data_split = txq->sw_tx_ring.skbs[idx].flags & QEDE_TSO_SPLIT_BD; txq 114 drivers/net/ethernet/qlogic/qede/qede_fp.c idx, txq->sw_tx_cons, txq->sw_tx_prod); txq 120 drivers/net/ethernet/qlogic/qede/qede_fp.c first_bd = (struct eth_tx_1st_bd *)qed_chain_consume(&txq->tx_pbl); txq 128 drivers/net/ethernet/qlogic/qede/qede_fp.c qed_chain_consume(&txq->tx_pbl); txq 138 drivers/net/ethernet/qlogic/qede/qede_fp.c qed_chain_consume(&txq->tx_pbl); txq 144 drivers/net/ethernet/qlogic/qede/qede_fp.c qed_chain_consume(&txq->tx_pbl); txq 148 drivers/net/ethernet/qlogic/qede/qede_fp.c txq->sw_tx_ring.skbs[idx].skb = NULL; txq 149 drivers/net/ethernet/qlogic/qede/qede_fp.c txq->sw_tx_ring.skbs[idx].flags = 0; txq 155 drivers/net/ethernet/qlogic/qede/qede_fp.c static void qede_free_failed_tx_pkt(struct qede_tx_queue *txq, txq 159 drivers/net/ethernet/qlogic/qede/qede_fp.c u16 idx = txq->sw_tx_prod; txq 160 drivers/net/ethernet/qlogic/qede/qede_fp.c struct sk_buff *skb = txq->sw_tx_ring.skbs[idx].skb; txq 165 drivers/net/ethernet/qlogic/qede/qede_fp.c qed_chain_set_prod(&txq->tx_pbl, txq 166 drivers/net/ethernet/qlogic/qede/qede_fp.c le16_to_cpu(txq->tx_db.data.bd_prod), first_bd); txq 168 drivers/net/ethernet/qlogic/qede/qede_fp.c first_bd = (struct eth_tx_1st_bd *)qed_chain_produce(&txq->tx_pbl); txq 172 drivers/net/ethernet/qlogic/qede/qede_fp.c qed_chain_produce(&txq->tx_pbl); txq 177 drivers/net/ethernet/qlogic/qede/qede_fp.c dma_unmap_single(txq->dev, BD_UNMAP_ADDR(first_bd), txq 183 drivers/net/ethernet/qlogic/qede/qede_fp.c qed_chain_produce(&txq->tx_pbl); txq 185 drivers/net/ethernet/qlogic/qede/qede_fp.c dma_unmap_page(txq->dev, txq 191 drivers/net/ethernet/qlogic/qede/qede_fp.c qed_chain_set_prod(&txq->tx_pbl, txq 192 drivers/net/ethernet/qlogic/qede/qede_fp.c le16_to_cpu(txq->tx_db.data.bd_prod), first_bd); txq 196 drivers/net/ethernet/qlogic/qede/qede_fp.c txq->sw_tx_ring.skbs[idx].skb = NULL; txq 197 drivers/net/ethernet/qlogic/qede/qede_fp.c txq->sw_tx_ring.skbs[idx].flags = 0; txq 267 drivers/net/ethernet/qlogic/qede/qede_fp.c static int map_frag_to_bd(struct qede_tx_queue *txq, txq 273 drivers/net/ethernet/qlogic/qede/qede_fp.c mapping = skb_frag_dma_map(txq->dev, frag, 0, txq 275 drivers/net/ethernet/qlogic/qede/qede_fp.c if (unlikely(dma_mapping_error(txq->dev, mapping))) txq 314 drivers/net/ethernet/qlogic/qede/qede_fp.c static inline void qede_update_tx_producer(struct qede_tx_queue *txq) txq 321 drivers/net/ethernet/qlogic/qede/qede_fp.c writel(txq->tx_db.raw, txq->doorbell_addr); txq 333 drivers/net/ethernet/qlogic/qede/qede_fp.c struct qede_tx_queue *txq = fp->xdp_tx; txq 335 drivers/net/ethernet/qlogic/qede/qede_fp.c u16 idx = txq->sw_tx_prod; txq 338 drivers/net/ethernet/qlogic/qede/qede_fp.c if (!qed_chain_get_elem_left(&txq->tx_pbl)) { txq 339 drivers/net/ethernet/qlogic/qede/qede_fp.c txq->stopped_cnt++; txq 343 drivers/net/ethernet/qlogic/qede/qede_fp.c first_bd = (struct eth_tx_1st_bd *)qed_chain_produce(&txq->tx_pbl); txq 365 drivers/net/ethernet/qlogic/qede/qede_fp.c txq->sw_tx_ring.xdp[idx].page = metadata->data; txq 366 drivers/net/ethernet/qlogic/qede/qede_fp.c txq->sw_tx_ring.xdp[idx].mapping = metadata->mapping; txq 367 drivers/net/ethernet/qlogic/qede/qede_fp.c txq->sw_tx_prod = (txq->sw_tx_prod + 1) % txq->num_tx_buffers; txq 375 drivers/net/ethernet/qlogic/qede/qede_fp.c int qede_txq_has_work(struct qede_tx_queue *txq) txq 381 drivers/net/ethernet/qlogic/qede/qede_fp.c hw_bd_cons = le16_to_cpu(*txq->hw_cons_ptr); txq 382 drivers/net/ethernet/qlogic/qede/qede_fp.c if (qed_chain_get_cons_idx(&txq->tx_pbl) == hw_bd_cons + 1) txq 385 drivers/net/ethernet/qlogic/qede/qede_fp.c return hw_bd_cons != qed_chain_get_cons_idx(&txq->tx_pbl); txq 388 drivers/net/ethernet/qlogic/qede/qede_fp.c static void qede_xdp_tx_int(struct qede_dev *edev, struct qede_tx_queue *txq) txq 392 drivers/net/ethernet/qlogic/qede/qede_fp.c hw_bd_cons = le16_to_cpu(*txq->hw_cons_ptr); txq 395 drivers/net/ethernet/qlogic/qede/qede_fp.c while (hw_bd_cons != qed_chain_get_cons_idx(&txq->tx_pbl)) { txq 396 drivers/net/ethernet/qlogic/qede/qede_fp.c qed_chain_consume(&txq->tx_pbl); txq 397 drivers/net/ethernet/qlogic/qede/qede_fp.c idx = txq->sw_tx_cons; txq 400 drivers/net/ethernet/qlogic/qede/qede_fp.c txq->sw_tx_ring.xdp[idx].mapping, txq 402 drivers/net/ethernet/qlogic/qede/qede_fp.c __free_page(txq->sw_tx_ring.xdp[idx].page); txq 404 drivers/net/ethernet/qlogic/qede/qede_fp.c txq->sw_tx_cons = (txq->sw_tx_cons + 1) % txq->num_tx_buffers; txq 405 drivers/net/ethernet/qlogic/qede/qede_fp.c txq->xmit_pkts++; txq 409 drivers/net/ethernet/qlogic/qede/qede_fp.c static int qede_tx_int(struct qede_dev *edev, struct qede_tx_queue *txq) txq 416 drivers/net/ethernet/qlogic/qede/qede_fp.c netdev_txq = netdev_get_tx_queue(edev->ndev, txq->ndev_txq_id); txq 418 drivers/net/ethernet/qlogic/qede/qede_fp.c hw_bd_cons = le16_to_cpu(*txq->hw_cons_ptr); txq 421 drivers/net/ethernet/qlogic/qede/qede_fp.c while (hw_bd_cons != qed_chain_get_cons_idx(&txq->tx_pbl)) { txq 424 drivers/net/ethernet/qlogic/qede/qede_fp.c rc = qede_free_tx_pkt(edev, txq, &len); txq 428 drivers/net/ethernet/qlogic/qede/qede_fp.c qed_chain_get_cons_idx(&txq->tx_pbl)); txq 434 drivers/net/ethernet/qlogic/qede/qede_fp.c txq->sw_tx_cons = (txq->sw_tx_cons + 1) % txq->num_tx_buffers; txq 435 drivers/net/ethernet/qlogic/qede/qede_fp.c txq->xmit_pkts++; txq 466 drivers/net/ethernet/qlogic/qede/qede_fp.c (qed_chain_get_elem_left(&txq->tx_pbl) txq 1363 drivers/net/ethernet/qlogic/qede/qede_fp.c if (qede_txq_has_work(&fp->txq[cos])) txq 1385 drivers/net/ethernet/qlogic/qede/qede_fp.c if (qede_txq_has_work(&fp->txq[cos])) txq 1386 drivers/net/ethernet/qlogic/qede/qede_fp.c qede_tx_int(edev, &fp->txq[cos]); txq 1433 drivers/net/ethernet/qlogic/qede/qede_fp.c struct qede_tx_queue *txq; txq 1450 drivers/net/ethernet/qlogic/qede/qede_fp.c txq = QEDE_NDEV_TXQ_ID_TO_TXQ(edev, txq_index); txq 1453 drivers/net/ethernet/qlogic/qede/qede_fp.c WARN_ON(qed_chain_get_elem_left(&txq->tx_pbl) < (MAX_SKB_FRAGS + 1)); txq 1460 drivers/net/ethernet/qlogic/qede/qede_fp.c txq->tx_mem_alloc_err++; txq 1469 drivers/net/ethernet/qlogic/qede/qede_fp.c idx = txq->sw_tx_prod; txq 1470 drivers/net/ethernet/qlogic/qede/qede_fp.c txq->sw_tx_ring.skbs[idx].skb = skb; txq 1472 drivers/net/ethernet/qlogic/qede/qede_fp.c qed_chain_produce(&txq->tx_pbl); txq 1481 drivers/net/ethernet/qlogic/qede/qede_fp.c mapping = dma_map_single(txq->dev, skb->data, txq 1483 drivers/net/ethernet/qlogic/qede/qede_fp.c if (unlikely(dma_mapping_error(txq->dev, mapping))) { txq 1485 drivers/net/ethernet/qlogic/qede/qede_fp.c qede_free_failed_tx_pkt(txq, first_bd, 0, false); txq 1486 drivers/net/ethernet/qlogic/qede/qede_fp.c qede_update_tx_producer(txq); txq 1497 drivers/net/ethernet/qlogic/qede/qede_fp.c qed_chain_produce(&txq->tx_pbl); txq 1502 drivers/net/ethernet/qlogic/qede/qede_fp.c qed_chain_produce(&txq->tx_pbl); txq 1535 drivers/net/ethernet/qlogic/qede/qede_fp.c if (unlikely(txq->is_legacy)) txq 1592 drivers/net/ethernet/qlogic/qede/qede_fp.c txq->sw_tx_ring.skbs[idx].flags |= QEDE_TSO_SPLIT_BD; txq 1609 drivers/net/ethernet/qlogic/qede/qede_fp.c rc = map_frag_to_bd(txq, txq 1613 drivers/net/ethernet/qlogic/qede/qede_fp.c qede_free_failed_tx_pkt(txq, first_bd, nbd, data_split); txq 1614 drivers/net/ethernet/qlogic/qede/qede_fp.c qede_update_tx_producer(txq); txq 1629 drivers/net/ethernet/qlogic/qede/qede_fp.c qed_chain_produce(&txq->tx_pbl); txq 1633 drivers/net/ethernet/qlogic/qede/qede_fp.c rc = map_frag_to_bd(txq, txq 1637 drivers/net/ethernet/qlogic/qede/qede_fp.c qede_free_failed_tx_pkt(txq, first_bd, nbd, data_split); txq 1638 drivers/net/ethernet/qlogic/qede/qede_fp.c qede_update_tx_producer(txq); txq 1653 drivers/net/ethernet/qlogic/qede/qede_fp.c txq->sw_tx_prod = (txq->sw_tx_prod + 1) % txq->num_tx_buffers; txq 1656 drivers/net/ethernet/qlogic/qede/qede_fp.c txq->tx_db.data.bd_prod = txq 1657 drivers/net/ethernet/qlogic/qede/qede_fp.c cpu_to_le16(qed_chain_get_prod_idx(&txq->tx_pbl)); txq 1660 drivers/net/ethernet/qlogic/qede/qede_fp.c qede_update_tx_producer(txq); txq 1662 drivers/net/ethernet/qlogic/qede/qede_fp.c if (unlikely(qed_chain_get_elem_left(&txq->tx_pbl) txq 1665 drivers/net/ethernet/qlogic/qede/qede_fp.c qede_update_tx_producer(txq); txq 1668 drivers/net/ethernet/qlogic/qede/qede_fp.c txq->stopped_cnt++; txq 1677 drivers/net/ethernet/qlogic/qede/qede_fp.c if ((qed_chain_get_elem_left(&txq->tx_pbl) >= txq 856 drivers/net/ethernet/qlogic/qede/qede_main.c kfree(fp->txq); txq 906 drivers/net/ethernet/qlogic/qede/qede_main.c fp->txq = kcalloc(edev->dev_info.num_tc, txq 907 drivers/net/ethernet/qlogic/qede/qede_main.c sizeof(*fp->txq), GFP_KERNEL); txq 908 drivers/net/ethernet/qlogic/qede/qede_main.c if (!fp->txq) txq 1461 drivers/net/ethernet/qlogic/qede/qede_main.c static void qede_free_mem_txq(struct qede_dev *edev, struct qede_tx_queue *txq) txq 1464 drivers/net/ethernet/qlogic/qede/qede_main.c if (txq->is_xdp) txq 1465 drivers/net/ethernet/qlogic/qede/qede_main.c kfree(txq->sw_tx_ring.xdp); txq 1467 drivers/net/ethernet/qlogic/qede/qede_main.c kfree(txq->sw_tx_ring.skbs); txq 1470 drivers/net/ethernet/qlogic/qede/qede_main.c edev->ops->common->chain_free(edev->cdev, &txq->tx_pbl); txq 1474 drivers/net/ethernet/qlogic/qede/qede_main.c static int qede_alloc_mem_txq(struct qede_dev *edev, struct qede_tx_queue *txq) txq 1479 drivers/net/ethernet/qlogic/qede/qede_main.c txq->num_tx_buffers = edev->q_num_tx_buffers; txq 1482 drivers/net/ethernet/qlogic/qede/qede_main.c if (txq->is_xdp) { txq 1483 drivers/net/ethernet/qlogic/qede/qede_main.c size = sizeof(*txq->sw_tx_ring.xdp) * txq->num_tx_buffers; txq 1484 drivers/net/ethernet/qlogic/qede/qede_main.c txq->sw_tx_ring.xdp = kzalloc(size, GFP_KERNEL); txq 1485 drivers/net/ethernet/qlogic/qede/qede_main.c if (!txq->sw_tx_ring.xdp) txq 1488 drivers/net/ethernet/qlogic/qede/qede_main.c size = sizeof(*txq->sw_tx_ring.skbs) * txq->num_tx_buffers; txq 1489 drivers/net/ethernet/qlogic/qede/qede_main.c txq->sw_tx_ring.skbs = kzalloc(size, GFP_KERNEL); txq 1490 drivers/net/ethernet/qlogic/qede/qede_main.c if (!txq->sw_tx_ring.skbs) txq 1498 drivers/net/ethernet/qlogic/qede/qede_main.c txq->num_tx_buffers, txq 1500 drivers/net/ethernet/qlogic/qede/qede_main.c &txq->tx_pbl, NULL); txq 1507 drivers/net/ethernet/qlogic/qede/qede_main.c qede_free_mem_txq(edev, txq); txq 1526 drivers/net/ethernet/qlogic/qede/qede_main.c qede_free_mem_txq(edev, &fp->txq[cos]); txq 1557 drivers/net/ethernet/qlogic/qede/qede_main.c rc = qede_alloc_mem_txq(edev, &fp->txq[cos]); txq 1600 drivers/net/ethernet/qlogic/qede/qede_main.c struct qede_tx_queue *txq) txq 1606 drivers/net/ethernet/qlogic/qede/qede_main.c netdev_txq = netdev_get_tx_queue(edev->ndev, txq->ndev_txq_id); txq 1608 drivers/net/ethernet/qlogic/qede/qede_main.c while (qed_chain_get_cons_idx(&txq->tx_pbl) != txq 1609 drivers/net/ethernet/qlogic/qede/qede_main.c qed_chain_get_prod_idx(&txq->tx_pbl)) { txq 1612 drivers/net/ethernet/qlogic/qede/qede_main.c txq->index, qed_chain_get_cons_idx(&txq->tx_pbl), txq 1613 drivers/net/ethernet/qlogic/qede/qede_main.c qed_chain_get_prod_idx(&txq->tx_pbl)); txq 1615 drivers/net/ethernet/qlogic/qede/qede_main.c rc = qede_free_tx_pkt(edev, txq, &len); txq 1619 drivers/net/ethernet/qlogic/qede/qede_main.c txq->index, txq 1620 drivers/net/ethernet/qlogic/qede/qede_main.c qed_chain_get_cons_idx(&txq->tx_pbl), txq 1621 drivers/net/ethernet/qlogic/qede/qede_main.c qed_chain_get_prod_idx(&txq->tx_pbl)); txq 1627 drivers/net/ethernet/qlogic/qede/qede_main.c txq->sw_tx_cons++; txq 1646 drivers/net/ethernet/qlogic/qede/qede_main.c &fp->txq[cos]); txq 1688 drivers/net/ethernet/qlogic/qede/qede_main.c struct qede_tx_queue *txq = &fp->txq[cos]; txq 1691 drivers/net/ethernet/qlogic/qede/qede_main.c txq->cos = cos; txq 1692 drivers/net/ethernet/qlogic/qede/qede_main.c txq->index = txq_index; txq 1693 drivers/net/ethernet/qlogic/qede/qede_main.c ndev_tx_id = QEDE_TXQ_TO_NDEV_TXQ_ID(edev, txq); txq 1694 drivers/net/ethernet/qlogic/qede/qede_main.c txq->ndev_txq_id = ndev_tx_id; txq 1697 drivers/net/ethernet/qlogic/qede/qede_main.c txq->is_legacy = 1; txq 1698 drivers/net/ethernet/qlogic/qede/qede_main.c txq->dev = &edev->pdev->dev; txq 1849 drivers/net/ethernet/qlogic/qede/qede_main.c struct qede_tx_queue *txq, bool allow_drain) txq 1853 drivers/net/ethernet/qlogic/qede/qede_main.c while (txq->sw_tx_cons != txq->sw_tx_prod) { txq 1858 drivers/net/ethernet/qlogic/qede/qede_main.c txq->index); txq 1862 drivers/net/ethernet/qlogic/qede/qede_main.c return qede_drain_txq(edev, txq, false); txq 1866 drivers/net/ethernet/qlogic/qede/qede_main.c txq->index, txq->sw_tx_prod, txq 1867 drivers/net/ethernet/qlogic/qede/qede_main.c txq->sw_tx_cons); txq 1882 drivers/net/ethernet/qlogic/qede/qede_main.c struct qede_tx_queue *txq, int rss_id) txq 1885 drivers/net/ethernet/qlogic/qede/qede_main.c edev->ops->common->db_recovery_del(edev->cdev, txq->doorbell_addr, txq 1886 drivers/net/ethernet/qlogic/qede/qede_main.c &txq->tx_db); txq 1888 drivers/net/ethernet/qlogic/qede/qede_main.c return edev->ops->q_tx_stop(edev->cdev, rss_id, txq->handle); txq 1924 drivers/net/ethernet/qlogic/qede/qede_main.c rc = qede_drain_txq(edev, &fp->txq[cos], true); txq 1946 drivers/net/ethernet/qlogic/qede/qede_main.c rc = qede_stop_txq(edev, &fp->txq[cos], i); txq 1981 drivers/net/ethernet/qlogic/qede/qede_main.c struct qede_tx_queue *txq, u8 rss_id, u16 sb_idx) txq 1983 drivers/net/ethernet/qlogic/qede/qede_main.c dma_addr_t phys_table = qed_chain_get_pbl_phys(&txq->tx_pbl); txq 1984 drivers/net/ethernet/qlogic/qede/qede_main.c u32 page_cnt = qed_chain_get_page_cnt(&txq->tx_pbl); txq 1995 drivers/net/ethernet/qlogic/qede/qede_main.c if (txq->is_xdp) txq 1996 drivers/net/ethernet/qlogic/qede/qede_main.c params.queue_id = QEDE_TXQ_XDP_TO_IDX(edev, txq); txq 1998 drivers/net/ethernet/qlogic/qede/qede_main.c params.queue_id = txq->index; txq 2002 drivers/net/ethernet/qlogic/qede/qede_main.c params.tc = txq->cos; txq 2007 drivers/net/ethernet/qlogic/qede/qede_main.c DP_ERR(edev, "Start TXQ #%d failed %d\n", txq->index, rc); txq 2011 drivers/net/ethernet/qlogic/qede/qede_main.c txq->doorbell_addr = ret_params.p_doorbell; txq 2012 drivers/net/ethernet/qlogic/qede/qede_main.c txq->handle = ret_params.p_handle; txq 2015 drivers/net/ethernet/qlogic/qede/qede_main.c txq->hw_cons_ptr = &fp->sb_info->sb_virt->pi_array[sb_idx]; txq 2018 drivers/net/ethernet/qlogic/qede/qede_main.c SET_FIELD(txq->tx_db.data.params, ETH_DB_DATA_DEST, DB_DEST_XCM); txq 2019 drivers/net/ethernet/qlogic/qede/qede_main.c SET_FIELD(txq->tx_db.data.params, ETH_DB_DATA_AGG_CMD, DB_AGG_CMD_SET); txq 2020 drivers/net/ethernet/qlogic/qede/qede_main.c SET_FIELD(txq->tx_db.data.params, ETH_DB_DATA_AGG_VAL_SEL, txq 2022 drivers/net/ethernet/qlogic/qede/qede_main.c txq->tx_db.data.agg_flags = DQ_XCM_ETH_DQ_CF_CMD; txq 2025 drivers/net/ethernet/qlogic/qede/qede_main.c rc = edev->ops->common->db_recovery_add(edev->cdev, txq->doorbell_addr, txq 2026 drivers/net/ethernet/qlogic/qede/qede_main.c &txq->tx_db, DB_REC_WIDTH_32B, txq 2130 drivers/net/ethernet/qlogic/qede/qede_main.c rc = qede_start_txq(edev, fp, &fp->txq[cos], i, txq 2489 drivers/net/ethernet/qlogic/qede/qede_main.c static bool qede_is_txq_full(struct qede_dev *edev, struct qede_tx_queue *txq) txq 2493 drivers/net/ethernet/qlogic/qede/qede_main.c netdev_txq = netdev_get_tx_queue(edev->ndev, txq->ndev_txq_id); txq 2558 drivers/net/ethernet/qlogic/qede/qede_main.c struct qede_tx_queue *txq = QEDE_FP_TC0_TXQ(fp); txq 2560 drivers/net/ethernet/qlogic/qede/qede_main.c if (txq->sw_tx_cons != txq->sw_tx_prod) txq 2562 drivers/net/ethernet/qlogic/qede/qede_main.c if (qede_is_txq_full(edev, txq)) txq 638 drivers/net/ethernet/qlogic/qlcnic/qlcnic.h struct netdev_queue *txq; txq 396 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c __netif_tx_lock_bh(tx_ring->txq); txq 401 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c netif_tx_stop_queue(tx_ring->txq); txq 405 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c netif_tx_wake_queue(tx_ring->txq); txq 408 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c __netif_tx_unlock_bh(tx_ring->txq); txq 432 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c __netif_tx_unlock_bh(tx_ring->txq); txq 693 drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c netif_tx_stop_queue(tx_ring->txq); txq 695 drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c netif_tx_start_queue(tx_ring->txq); txq 930 drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c if (netif_tx_queue_stopped(tx_ring->txq) && txq 933 drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c netif_tx_wake_queue(tx_ring->txq); txq 2413 drivers/net/ethernet/qlogic/qlcnic/qlcnic_main.c tx_ring->txq = netdev_get_tx_queue(netdev, ring); txq 454 drivers/net/ethernet/samsung/sxgbe/sxgbe_common.h struct sxgbe_tx_queue *txq[SXGBE_TX_QUEUES]; txq 333 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c priv->hw->desc->init_tx_desc(&priv->txq[j]->dma_tx[i]); txq 566 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c priv->txq[queue_num], tx_rsize); txq 575 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c priv->txq[queue_num]->priv_ptr = priv; txq 599 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c free_tx_ring(priv->device, priv->txq[queue_num], tx_rsize); txq 635 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c struct sxgbe_tx_queue *tqueue = priv->txq[queue_num]; txq 651 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c free_tx_ring(priv->device, priv->txq[queue_num], tx_rsize); txq 665 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c priv->txq[queue_num] = devm_kmalloc(priv->device, txq 667 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c if (!priv->txq[queue_num]) txq 796 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c struct sxgbe_tx_queue *tqueue = priv->txq[queue_num]; txq 815 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c struct sxgbe_tx_queue *tx_ring = priv->txq[queue_num]; txq 960 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c (priv->txq[queue_num])->dma_tx_phy, txq 1024 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c struct sxgbe_tx_queue *p = priv->txq[queue_num]; txq 1038 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c struct sxgbe_tx_queue *p = priv->txq[queue_num]; txq 1122 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c (priv->txq[queue_num])->irq_no, txq 1124 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c dev->name, priv->txq[queue_num]); txq 1274 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c struct sxgbe_tx_queue *tqueue = priv->txq[txq_index]; txq 1623 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c struct sxgbe_tx_queue *txq = (struct sxgbe_tx_queue *)dev_id; txq 1624 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c struct sxgbe_priv_data *priv = txq->priv_ptr; txq 1627 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c status = priv->hw->dma->tx_dma_int_status(priv->ioaddr, txq->queue_no, txq 1635 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c sxgbe_restart_tx_queue(priv, txq->queue_no); txq 1644 drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c txq->queue_no, priv->tx_tc); txq 127 drivers/net/ethernet/samsung/sxgbe/sxgbe_platform.c priv->txq[i]->irq_no = irq_of_parse_and_map(node, chan++); txq 128 drivers/net/ethernet/samsung/sxgbe/sxgbe_platform.c if (priv->txq[i]->irq_no <= 0) { txq 160 drivers/net/ethernet/samsung/sxgbe/sxgbe_platform.c irq_dispose_mapping(priv->txq[i]->irq_no); txq 284 drivers/net/ethernet/sfc/ptp.c struct sk_buff_head txq; txq 1365 drivers/net/ethernet/sfc/ptp.c skb_queue_purge(&efx->ptp_data->txq); txq 1418 drivers/net/ethernet/sfc/ptp.c while ((skb = skb_dequeue(&ptp_data->txq))) txq 1462 drivers/net/ethernet/sfc/ptp.c skb_queue_head_init(&ptp->txq); txq 1570 drivers/net/ethernet/sfc/ptp.c skb_queue_purge(&efx->ptp_data->txq); txq 1708 drivers/net/ethernet/sfc/ptp.c skb_queue_tail(&ptp->txq, skb); txq 3586 drivers/net/ethernet/sun/niu.c struct netdev_queue *txq; txq 3592 drivers/net/ethernet/sun/niu.c txq = netdev_get_tx_queue(np->dev, index); txq 3616 drivers/net/ethernet/sun/niu.c if (unlikely(netif_tx_queue_stopped(txq) && txq 3618 drivers/net/ethernet/sun/niu.c __netif_tx_lock(txq, smp_processor_id()); txq 3619 drivers/net/ethernet/sun/niu.c if (netif_tx_queue_stopped(txq) && txq 3621 drivers/net/ethernet/sun/niu.c netif_tx_wake_queue(txq); txq 3622 drivers/net/ethernet/sun/niu.c __netif_tx_unlock(txq); txq 6612 drivers/net/ethernet/sun/niu.c struct netdev_queue *txq; txq 6622 drivers/net/ethernet/sun/niu.c txq = netdev_get_tx_queue(dev, i); txq 6625 drivers/net/ethernet/sun/niu.c netif_tx_stop_queue(txq); txq 6716 drivers/net/ethernet/sun/niu.c netif_tx_stop_queue(txq); txq 6718 drivers/net/ethernet/sun/niu.c netif_tx_wake_queue(txq); txq 703 drivers/net/ethernet/sun/sungem.c struct netdev_queue *txq = netdev_get_tx_queue(dev, 0); txq 705 drivers/net/ethernet/sun/sungem.c __netif_tx_lock(txq, smp_processor_id()); txq 709 drivers/net/ethernet/sun/sungem.c __netif_tx_unlock(txq); txq 894 drivers/net/ethernet/sun/sungem.c struct netdev_queue *txq = netdev_get_tx_queue(dev, 0); txq 902 drivers/net/ethernet/sun/sungem.c __netif_tx_lock(txq, smp_processor_id()); txq 904 drivers/net/ethernet/sun/sungem.c __netif_tx_unlock(txq); txq 1349 drivers/net/ethernet/sun/sungem.c struct netdev_queue *txq = netdev_get_tx_queue(gp->dev, 0); txq 1379 drivers/net/ethernet/sun/sungem.c __netif_tx_lock(txq, smp_processor_id()); txq 1441 drivers/net/ethernet/sun/sungem.c __netif_tx_unlock(txq); txq 692 drivers/net/ethernet/sun/sunvnet_common.c struct netdev_queue *txq; txq 726 drivers/net/ethernet/sun/sunvnet_common.c txq = netdev_get_tx_queue(dev, port->q_index); txq 727 drivers/net/ethernet/sun/sunvnet_common.c if (unlikely(netif_tx_queue_stopped(txq) && txq 761 drivers/net/ethernet/sun/sunvnet_common.c struct netdev_queue *txq; txq 763 drivers/net/ethernet/sun/sunvnet_common.c txq = netdev_get_tx_queue(VNET_PORT_TO_NET_DEVICE(port), txq 765 drivers/net/ethernet/sun/sunvnet_common.c __netif_tx_lock(txq, smp_processor_id()); txq 766 drivers/net/ethernet/sun/sunvnet_common.c if (likely(netif_tx_queue_stopped(txq))) txq 767 drivers/net/ethernet/sun/sunvnet_common.c netif_tx_wake_queue(txq); txq 768 drivers/net/ethernet/sun/sunvnet_common.c __netif_tx_unlock(txq); txq 1257 drivers/net/ethernet/sun/sunvnet_common.c struct netdev_queue *txq; txq 1259 drivers/net/ethernet/sun/sunvnet_common.c txq = netdev_get_tx_queue(dev, port->q_index); txq 1260 drivers/net/ethernet/sun/sunvnet_common.c netif_tx_stop_queue(txq); txq 1263 drivers/net/ethernet/sun/sunvnet_common.c netif_tx_wake_queue(txq); txq 1337 drivers/net/ethernet/sun/sunvnet_common.c struct netdev_queue *txq; txq 1397 drivers/net/ethernet/sun/sunvnet_common.c txq = netdev_get_tx_queue(dev, i); txq 1399 drivers/net/ethernet/sun/sunvnet_common.c if (!netif_tx_queue_stopped(txq)) { txq 1400 drivers/net/ethernet/sun/sunvnet_common.c netif_tx_stop_queue(txq); txq 1515 drivers/net/ethernet/sun/sunvnet_common.c netif_tx_stop_queue(txq); txq 1518 drivers/net/ethernet/sun/sunvnet_common.c netif_tx_wake_queue(txq); txq 592 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c struct netdev_queue *txq; txq 608 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c txq = netdev_get_tx_queue(netdev, channel->queue_index); txq 609 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c netdev_tx_reset_queue(txq); txq 707 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c struct netdev_queue *txq; txq 717 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c txq = netdev_get_tx_queue(netdev, channel->queue_index); txq 753 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c netdev_tx_sent_queue(txq, tx_pkt_info->tx_bytes); txq 1044 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c struct netdev_queue *txq; txq 1060 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c txq = netdev_get_tx_queue(netdev, channel->queue_index); txq 1094 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c netdev_tx_completed_queue(txq, tx_packets, tx_bytes); txq 1099 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c netif_tx_wake_queue(txq); txq 378 drivers/net/ethernet/ti/cpsw.c struct netdev_queue *txq; txq 399 drivers/net/ethernet/ti/cpsw.c txq = netdev_get_tx_queue(ndev, ch); txq 400 drivers/net/ethernet/ti/cpsw.c if (unlikely(netif_tx_queue_stopped(txq))) txq 401 drivers/net/ethernet/ti/cpsw.c netif_tx_wake_queue(txq); txq 1801 drivers/net/ethernet/ti/cpsw.c struct netdev_queue *txq; txq 1820 drivers/net/ethernet/ti/cpsw.c txq = netdev_get_tx_queue(ndev, q_idx); txq 1833 drivers/net/ethernet/ti/cpsw.c netif_tx_stop_queue(txq); txq 1839 drivers/net/ethernet/ti/cpsw.c netif_tx_wake_queue(txq); txq 1845 drivers/net/ethernet/ti/cpsw.c netif_tx_stop_queue(txq); txq 1851 drivers/net/ethernet/ti/cpsw.c netif_tx_wake_queue(txq); txq 84 drivers/net/ethernet/ti/cpts.c skb_queue_walk_safe(&cpts->txq, skb, tmp) { txq 87 drivers/net/ethernet/ti/cpts.c __skb_unlink(skb, &cpts->txq); txq 108 drivers/net/ethernet/ti/cpts.c skb_queue_walk_safe(&cpts->txq, skb, tmp) { txq 121 drivers/net/ethernet/ti/cpts.c __skb_unlink(skb, &cpts->txq); txq 131 drivers/net/ethernet/ti/cpts.c __skb_unlink(skb, &cpts->txq); txq 303 drivers/net/ethernet/ti/cpts.c if (!skb_queue_empty(&cpts->txq)) { txq 305 drivers/net/ethernet/ti/cpts.c if (!skb_queue_empty(&cpts->txq)) txq 409 drivers/net/ethernet/ti/cpts.c __skb_queue_tail(&cpts->txq, skb); txq 451 drivers/net/ethernet/ti/cpts.c skb_queue_head_init(&cpts->txq); txq 493 drivers/net/ethernet/ti/cpts.c skb_queue_purge(&cpts->txq); txq 116 drivers/net/ethernet/ti/cpts.h struct sk_buff_head txq; txq 710 drivers/net/hyperv/netvsc.c struct netdev_queue *txq = netdev_get_tx_queue(ndev, q_idx); txq 712 drivers/net/hyperv/netvsc.c if (netif_tx_queue_stopped(txq) && !net_device->tx_disable && txq 715 drivers/net/hyperv/netvsc.c netif_tx_wake_queue(txq); txq 818 drivers/net/hyperv/netvsc.c struct netdev_queue *txq = netdev_get_tx_queue(ndev, packet->q_idx); txq 861 drivers/net/hyperv/netvsc.c netif_tx_stop_queue(txq); txq 865 drivers/net/hyperv/netvsc.c netif_tx_stop_queue(txq); txq 874 drivers/net/hyperv/netvsc.c if (netif_tx_queue_stopped(txq) && txq 877 drivers/net/hyperv/netvsc.c netif_tx_wake_queue(txq); txq 347 drivers/net/hyperv/netvsc_drv.c u16 txq; txq 355 drivers/net/hyperv/netvsc_drv.c txq = vf_ops->ndo_select_queue(vf_netdev, skb, sb_dev); txq 357 drivers/net/hyperv/netvsc_drv.c txq = netdev_pick_tx(vf_netdev, skb, NULL); txq 363 drivers/net/hyperv/netvsc_drv.c qdisc_skb_cb(skb)->slave_dev_queue_mapping = txq; txq 365 drivers/net/hyperv/netvsc_drv.c txq = netvsc_pick_tx(ndev, skb); txq 369 drivers/net/hyperv/netvsc_drv.c while (unlikely(txq >= ndev->real_num_tx_queues)) txq 370 drivers/net/hyperv/netvsc_drv.c txq -= ndev->real_num_tx_queues; txq 372 drivers/net/hyperv/netvsc_drv.c return txq; txq 65 drivers/net/ifb.c struct netdev_queue *txq; txq 68 drivers/net/ifb.c txq = netdev_get_tx_queue(txp->dev, txp->txqnum); txq 71 drivers/net/ifb.c if (!__netif_tx_trylock(txq)) txq 74 drivers/net/ifb.c __netif_tx_unlock(txq); txq 107 drivers/net/ifb.c if (__netif_tx_trylock(txq)) { txq 111 drivers/net/ifb.c if (netif_tx_queue_stopped(txq)) txq 112 drivers/net/ifb.c netif_tx_wake_queue(txq); txq 114 drivers/net/ifb.c __netif_tx_unlock(txq); txq 117 drivers/net/ifb.c __netif_tx_unlock(txq); txq 123 drivers/net/net_failover.c u16 txq; txq 130 drivers/net/net_failover.c txq = ops->ndo_select_queue(primary_dev, skb, sb_dev); txq 132 drivers/net/net_failover.c txq = netdev_pick_tx(primary_dev, skb, NULL); txq 136 drivers/net/net_failover.c return txq; txq 139 drivers/net/net_failover.c txq = skb_rx_queue_recorded(skb) ? skb_get_rx_queue(skb) : 0; txq 144 drivers/net/net_failover.c if (unlikely(txq >= dev->real_num_tx_queues)) { txq 146 drivers/net/net_failover.c txq -= dev->real_num_tx_queues; txq 147 drivers/net/net_failover.c } while (txq >= dev->real_num_tx_queues); txq 150 drivers/net/net_failover.c return txq; txq 1736 drivers/net/team/team.c u16 txq = skb_rx_queue_recorded(skb) ? skb_get_rx_queue(skb) : 0; txq 1743 drivers/net/team/team.c if (unlikely(txq >= dev->real_num_tx_queues)) { txq 1745 drivers/net/team/team.c txq -= dev->real_num_tx_queues; txq 1746 drivers/net/team/team.c } while (txq >= dev->real_num_tx_queues); txq 1748 drivers/net/team/team.c return txq; txq 568 drivers/net/tun.c u32 txq = 0; txq 573 drivers/net/tun.c txq = __skb_get_hash_symmetric(skb); txq 574 drivers/net/tun.c e = tun_flow_find(&tun->flows[tun_hashfn(txq)], txq); txq 576 drivers/net/tun.c tun_flow_save_rps_rxhash(e, txq); txq 577 drivers/net/tun.c txq = e->queue_index; txq 580 drivers/net/tun.c txq = ((u64)txq * numqueues) >> 32; txq 583 drivers/net/tun.c return txq; txq 1065 drivers/net/tun.c int txq = skb->queue_mapping; txq 1070 drivers/net/tun.c tfile = rcu_dereference(tun->tfiles[txq]); txq 372 drivers/net/usb/lan78xx.c struct sk_buff_head txq; txq 2675 drivers/net/usb/lan78xx.c temp = unlink_urbs(dev, &dev->txq) + unlink_urbs(dev, &dev->rxq); txq 2679 drivers/net/usb/lan78xx.c !skb_queue_empty(&dev->txq) && txq 2829 drivers/net/usb/lan78xx.c defer_bh(dev, skb, &dev->txq, tx_done); txq 3382 drivers/net/usb/lan78xx.c spin_lock_irqsave(&dev->txq.lock, flags); txq 3385 drivers/net/usb/lan78xx.c spin_unlock_irqrestore(&dev->txq.lock, flags); txq 3405 drivers/net/usb/lan78xx.c spin_unlock_irqrestore(&dev->txq.lock, flags); txq 3415 drivers/net/usb/lan78xx.c lan78xx_queue_skb(&dev->txq, skb, tx_start); txq 3416 drivers/net/usb/lan78xx.c if (skb_queue_len(&dev->txq) >= dev->tx_qlen) txq 3431 drivers/net/usb/lan78xx.c spin_unlock_irqrestore(&dev->txq.lock, flags); txq 3463 drivers/net/usb/lan78xx.c if (skb_queue_len(&dev->txq) < dev->tx_qlen) txq 3519 drivers/net/usb/lan78xx.c unlink_urbs(dev, &dev->txq); txq 3668 drivers/net/usb/lan78xx.c unlink_urbs(dev, &dev->txq); txq 3740 drivers/net/usb/lan78xx.c skb_queue_head_init(&dev->txq); txq 4017 drivers/net/usb/lan78xx.c spin_lock_irq(&dev->txq.lock); txq 4019 drivers/net/usb/lan78xx.c if ((skb_queue_len(&dev->txq) || txq 4022 drivers/net/usb/lan78xx.c spin_unlock_irq(&dev->txq.lock); txq 4027 drivers/net/usb/lan78xx.c spin_unlock_irq(&dev->txq.lock); txq 4121 drivers/net/usb/lan78xx.c spin_lock_irq(&dev->txq.lock); txq 4131 drivers/net/usb/lan78xx.c lan78xx_queue_skb(&dev->txq, skb, tx_start); txq 4136 drivers/net/usb/lan78xx.c spin_unlock_irq(&dev->txq.lock); txq 4139 drivers/net/usb/lan78xx.c if (!(skb_queue_len(&dev->txq) >= dev->tx_qlen)) txq 786 drivers/net/usb/usbnet.c temp = unlink_urbs(dev, &dev->txq) + txq 791 drivers/net/usb/usbnet.c wait_skb_queue_empty(&dev->txq); txq 1140 drivers/net/usb/usbnet.c unlink_urbs (dev, &dev->txq); txq 1291 drivers/net/usb/usbnet.c (void) defer_bh(dev, skb, &dev->txq, tx_done); txq 1300 drivers/net/usb/usbnet.c unlink_urbs (dev, &dev->txq); txq 1425 drivers/net/usb/usbnet.c spin_lock_irqsave(&dev->txq.lock, flags); txq 1428 drivers/net/usb/usbnet.c spin_unlock_irqrestore(&dev->txq.lock, flags); txq 1433 drivers/net/usb/usbnet.c spin_unlock_irqrestore(&dev->txq.lock, flags); txq 1445 drivers/net/usb/usbnet.c spin_unlock_irqrestore(&dev->txq.lock, flags); txq 1464 drivers/net/usb/usbnet.c __usbnet_queue_skb(&dev->txq, skb, tx_start); txq 1465 drivers/net/usb/usbnet.c if (dev->txq.qlen >= TX_QLEN (dev)) txq 1468 drivers/net/usb/usbnet.c spin_unlock_irqrestore (&dev->txq.lock, flags); txq 1549 drivers/net/usb/usbnet.c if (dev->txq.qlen + dev->rxq.qlen + dev->done.qlen == 0) txq 1571 drivers/net/usb/usbnet.c if (dev->txq.qlen < TX_QLEN (dev)) txq 1700 drivers/net/usb/usbnet.c skb_queue_head_init (&dev->txq); txq 1853 drivers/net/usb/usbnet.c spin_lock_irq(&dev->txq.lock); txq 1855 drivers/net/usb/usbnet.c if (dev->txq.qlen && PMSG_IS_AUTO(message)) { txq 1857 drivers/net/usb/usbnet.c spin_unlock_irq(&dev->txq.lock); txq 1861 drivers/net/usb/usbnet.c spin_unlock_irq(&dev->txq.lock); txq 1892 drivers/net/usb/usbnet.c spin_lock_irq(&dev->txq.lock); txq 1904 drivers/net/usb/usbnet.c __skb_queue_tail(&dev->txq, skb); txq 1910 drivers/net/usb/usbnet.c spin_unlock_irq(&dev->txq.lock); txq 1921 drivers/net/usb/usbnet.c if (!(dev->txq.qlen >= TX_QLEN(dev))) txq 267 drivers/net/virtio_net.c static int txq2vq(int txq) txq 269 drivers/net/virtio_net.c return txq * 2 + 1; txq 1405 drivers/net/virtio_net.c struct netdev_queue *txq = netdev_get_tx_queue(vi->dev, index); txq 1410 drivers/net/virtio_net.c if (__netif_tx_trylock(txq)) { txq 1412 drivers/net/virtio_net.c __netif_tx_unlock(txq); txq 1416 drivers/net/virtio_net.c netif_tx_wake_queue(txq); txq 1485 drivers/net/virtio_net.c struct netdev_queue *txq; txq 1493 drivers/net/virtio_net.c txq = netdev_get_tx_queue(vi->dev, index); txq 1494 drivers/net/virtio_net.c __netif_tx_lock(txq, raw_smp_processor_id()); txq 1496 drivers/net/virtio_net.c __netif_tx_unlock(txq); txq 1501 drivers/net/virtio_net.c netif_tx_wake_queue(txq); txq 1559 drivers/net/virtio_net.c struct netdev_queue *txq = netdev_get_tx_queue(dev, qnum); txq 1616 drivers/net/virtio_net.c if (kick || netif_xmit_stopped(txq)) { txq 1946 drivers/net/vmxnet3/vmxnet3_drv.c struct vmxnet3_tx_queue *txq = &adapter->tx_queue[i]; txq 1947 drivers/net/vmxnet3/vmxnet3_drv.c vmxnet3_tq_tx_complete(txq, adapter); txq 111 drivers/net/vrf.c struct netdev_queue *txq; txq 117 drivers/net/vrf.c txq = netdev_get_tx_queue(dev, 0); txq 118 drivers/net/vrf.c qdisc = rcu_access_pointer(txq->qdisc); txq 192 drivers/net/wan/dlci.c struct netdev_queue *txq = skb_get_tx_queue(dev, skb); txq 193 drivers/net/wan/dlci.c netdev_start_xmit(skb, dlp->slave, txq, false); txq 439 drivers/net/wan/farsync.c struct sk_buff *txq[FST_TXQ_DEPTH]; /* The queue */ txq 1362 drivers/net/wan/farsync.c skb = port->txq[port->txqs]; txq 2330 drivers/net/wan/farsync.c port->txq[port->txqe] = skb; txq 121 drivers/net/wireless/ath/ath10k/core.h struct ieee80211_txq *txq; txq 2270 drivers/net/wireless/ath/ath10k/htt.h struct ieee80211_txq *txq); txq 2272 drivers/net/wireless/ath/ath10k/htt.h struct ieee80211_txq *txq); txq 3042 drivers/net/wireless/ath/ath10k/htt_rx.c struct ieee80211_txq *txq; txq 3114 drivers/net/wireless/ath/ath10k/htt_rx.c txq = ath10k_mac_txq_lookup(ar, peer_id, tid); txq 3121 drivers/net/wireless/ath/ath10k/htt_rx.c if (unlikely(!txq)) { txq 3130 drivers/net/wireless/ath/ath10k/htt_rx.c ieee80211_txq_schedule_start(hw, txq->ac); txq 3131 drivers/net/wireless/ath/ath10k/htt_rx.c may_tx = ieee80211_txq_may_transmit(hw, txq); txq 3137 drivers/net/wireless/ath/ath10k/htt_rx.c ret = ath10k_mac_tx_push_txq(hw, txq); txq 3144 drivers/net/wireless/ath/ath10k/htt_rx.c ieee80211_return_txq(hw, txq, false); txq 3145 drivers/net/wireless/ath/ath10k/htt_rx.c ieee80211_txq_schedule_end(hw, txq->ac); txq 3150 drivers/net/wireless/ath/ath10k/htt_rx.c ath10k_htt_tx_txq_recalc(hw, txq); txq 3205 drivers/net/wireless/ath/ath10k/htt_rx.c struct ieee80211_txq *txq; txq 3278 drivers/net/wireless/ath/ath10k/htt_rx.c txq = ath10k_mac_txq_lookup(ar, peer_id, tid); txq 3285 drivers/net/wireless/ath/ath10k/htt_rx.c if (unlikely(!txq)) { txq 3292 drivers/net/wireless/ath/ath10k/htt_rx.c artxq = (void *)txq->drv_priv; txq 38 drivers/net/wireless/ath/ath10k/htt_tx.c struct ieee80211_txq *txq) txq 42 drivers/net/wireless/ath/ath10k/htt_tx.c struct ath10k_vif *arvif = (void *)txq->vif->drv_priv; txq 59 drivers/net/wireless/ath/ath10k/htt_tx.c if (txq->sta) { txq 60 drivers/net/wireless/ath/ath10k/htt_tx.c arsta = (void *)txq->sta->drv_priv; txq 66 drivers/net/wireless/ath/ath10k/htt_tx.c tid = txq->tid; txq 70 drivers/net/wireless/ath/ath10k/htt_tx.c ieee80211_txq_get_depth(txq, &frame_cnt, &byte_cnt); txq 116 drivers/net/wireless/ath/ath10k/htt_tx.c struct ieee80211_txq *txq) txq 121 drivers/net/wireless/ath/ath10k/htt_tx.c __ath10k_htt_tx_txq_recalc(hw, txq); txq 133 drivers/net/wireless/ath/ath10k/htt_tx.c struct ieee80211_txq *txq) txq 138 drivers/net/wireless/ath/ath10k/htt_tx.c __ath10k_htt_tx_txq_recalc(hw, txq); txq 3577 drivers/net/wireless/ath/ath10k/mac.c struct ieee80211_txq *txq, txq 3607 drivers/net/wireless/ath/ath10k/mac.c cb->txq = txq; txq 3929 drivers/net/wireless/ath/ath10k/mac.c static void ath10k_mac_txq_init(struct ieee80211_txq *txq) txq 3933 drivers/net/wireless/ath/ath10k/mac.c if (!txq) txq 3936 drivers/net/wireless/ath/ath10k/mac.c artxq = (void *)txq->drv_priv; txq 3940 drivers/net/wireless/ath/ath10k/mac.c static void ath10k_mac_txq_unref(struct ath10k *ar, struct ieee80211_txq *txq) txq 3946 drivers/net/wireless/ath/ath10k/mac.c if (!txq) txq 3952 drivers/net/wireless/ath/ath10k/mac.c if (cb->txq == txq) txq 3953 drivers/net/wireless/ath/ath10k/mac.c cb->txq = NULL; txq 3974 drivers/net/wireless/ath/ath10k/mac.c return peer->sta->txq[tid]; txq 3976 drivers/net/wireless/ath/ath10k/mac.c return peer->vif->txq; txq 3982 drivers/net/wireless/ath/ath10k/mac.c struct ieee80211_txq *txq) txq 3985 drivers/net/wireless/ath/ath10k/mac.c struct ath10k_txq *artxq = (void *)txq->drv_priv; txq 4009 drivers/net/wireless/ath/ath10k/mac.c struct ieee80211_txq *txq, txq 4016 drivers/net/wireless/ath/ath10k/mac.c if (!txq || !txq->sta) txq 4023 drivers/net/wireless/ath/ath10k/mac.c arsta = (struct ath10k_sta *)txq->sta->drv_priv; txq 4047 drivers/net/wireless/ath/ath10k/mac.c struct ieee80211_txq *txq) txq 4051 drivers/net/wireless/ath/ath10k/mac.c struct ath10k_txq *artxq = (void *)txq->drv_priv; txq 4052 drivers/net/wireless/ath/ath10k/mac.c struct ieee80211_vif *vif = txq->vif; txq 4053 drivers/net/wireless/ath/ath10k/mac.c struct ieee80211_sta *sta = txq->sta; txq 4070 drivers/net/wireless/ath/ath10k/mac.c skb = ieee80211_tx_dequeue(hw, txq); txq 4079 drivers/net/wireless/ath/ath10k/mac.c airtime = ath10k_mac_update_airtime(ar, txq, skb); txq 4080 drivers/net/wireless/ath/ath10k/mac.c ath10k_mac_tx_h_fill_cb(ar, vif, txq, skb, airtime); txq 4124 drivers/net/wireless/ath/ath10k/mac.c struct ieee80211_txq *txq; txq 4128 drivers/net/wireless/ath/ath10k/mac.c while ((txq = ieee80211_next_txq(hw, ac))) { txq 4129 drivers/net/wireless/ath/ath10k/mac.c while (ath10k_mac_tx_can_push(hw, txq)) { txq 4130 drivers/net/wireless/ath/ath10k/mac.c ret = ath10k_mac_tx_push_txq(hw, txq); txq 4134 drivers/net/wireless/ath/ath10k/mac.c ieee80211_return_txq(hw, txq, false); txq 4135 drivers/net/wireless/ath/ath10k/mac.c ath10k_htt_tx_txq_update(hw, txq); txq 4339 drivers/net/wireless/ath/ath10k/mac.c struct ieee80211_txq *txq = NULL; txq 4349 drivers/net/wireless/ath/ath10k/mac.c airtime = ath10k_mac_update_airtime(ar, txq, skb); txq 4350 drivers/net/wireless/ath/ath10k/mac.c ath10k_mac_tx_h_fill_cb(ar, vif, txq, skb, airtime); txq 4398 drivers/net/wireless/ath/ath10k/mac.c struct ieee80211_txq *txq) txq 4404 drivers/net/wireless/ath/ath10k/mac.c ath10k_htt_tx_txq_update(hw, txq); txq 4408 drivers/net/wireless/ath/ath10k/mac.c ac = txq->ac; txq 4410 drivers/net/wireless/ath/ath10k/mac.c txq = ieee80211_next_txq(hw, ac); txq 4411 drivers/net/wireless/ath/ath10k/mac.c if (!txq) txq 4414 drivers/net/wireless/ath/ath10k/mac.c while (ath10k_mac_tx_can_push(hw, txq)) { txq 4415 drivers/net/wireless/ath/ath10k/mac.c ret = ath10k_mac_tx_push_txq(hw, txq); txq 4419 drivers/net/wireless/ath/ath10k/mac.c ieee80211_return_txq(hw, txq, false); txq 4420 drivers/net/wireless/ath/ath10k/mac.c ath10k_htt_tx_txq_update(hw, txq); txq 5150 drivers/net/wireless/ath/ath10k/mac.c ath10k_mac_txq_init(vif->txq); txq 5584 drivers/net/wireless/ath/ath10k/mac.c ath10k_mac_txq_unref(ar, vif->txq); txq 5601 drivers/net/wireless/ath/ath10k/mac.c ath10k_mac_txq_unref(ar, vif->txq); txq 6471 drivers/net/wireless/ath/ath10k/mac.c for (i = 0; i < ARRAY_SIZE(sta->txq); i++) txq 6472 drivers/net/wireless/ath/ath10k/mac.c ath10k_mac_txq_init(sta->txq[i]); txq 6637 drivers/net/wireless/ath/ath10k/mac.c for (i = 0; i < ARRAY_SIZE(sta->txq); i++) txq 6638 drivers/net/wireless/ath/ath10k/mac.c ath10k_mac_txq_unref(ar, sta->txq[i]); txq 69 drivers/net/wireless/ath/ath10k/mac.h struct ieee80211_txq *txq); txq 49 drivers/net/wireless/ath/ath10k/txrx.c struct ieee80211_txq *txq; txq 74 drivers/net/wireless/ath/ath10k/txrx.c txq = skb_cb->txq; txq 76 drivers/net/wireless/ath/ath10k/txrx.c if (txq) { txq 77 drivers/net/wireless/ath/ath10k/txrx.c artxq = (void *)txq->drv_priv; txq 87 drivers/net/wireless/ath/ath10k/txrx.c if (txq && txq->sta && skb_cb->airtime_est) txq 88 drivers/net/wireless/ath/ath10k/txrx.c ieee80211_sta_register_airtime(txq->sta, txq->tid, txq 733 drivers/net/wireless/ath/ath5k/base.c struct ath5k_txq *txq, int padsize, txq 830 drivers/net/wireless/ath/ath5k/base.c spin_lock_bh(&txq->lock); txq 831 drivers/net/wireless/ath/ath5k/base.c list_add_tail(&bf->list, &txq->q); txq 832 drivers/net/wireless/ath/ath5k/base.c txq->txq_len++; txq 833 drivers/net/wireless/ath/ath5k/base.c if (txq->link == NULL) /* is this first packet? */ txq 834 drivers/net/wireless/ath/ath5k/base.c ath5k_hw_set_txdp(ah, txq->qnum, bf->daddr); txq 836 drivers/net/wireless/ath/ath5k/base.c *txq->link = bf->daddr; txq 838 drivers/net/wireless/ath/ath5k/base.c txq->link = &ds->ds_link; txq 839 drivers/net/wireless/ath/ath5k/base.c ath5k_hw_start_tx_dma(ah, txq->qnum); txq 840 drivers/net/wireless/ath/ath5k/base.c spin_unlock_bh(&txq->lock); txq 977 drivers/net/wireless/ath/ath5k/base.c struct ath5k_txq *txq; txq 1010 drivers/net/wireless/ath/ath5k/base.c txq = &ah->txqs[qnum]; txq 1011 drivers/net/wireless/ath/ath5k/base.c if (!txq->setup) { txq 1012 drivers/net/wireless/ath/ath5k/base.c txq->qnum = qnum; txq 1013 drivers/net/wireless/ath/ath5k/base.c txq->link = NULL; txq 1014 drivers/net/wireless/ath/ath5k/base.c INIT_LIST_HEAD(&txq->q); txq 1015 drivers/net/wireless/ath/ath5k/base.c spin_lock_init(&txq->lock); txq 1016 drivers/net/wireless/ath/ath5k/base.c txq->setup = true; txq 1017 drivers/net/wireless/ath/ath5k/base.c txq->txq_len = 0; txq 1018 drivers/net/wireless/ath/ath5k/base.c txq->txq_max = ATH5K_TXQ_LEN_MAX; txq 1019 drivers/net/wireless/ath/ath5k/base.c txq->txq_poll_mark = false; txq 1020 drivers/net/wireless/ath/ath5k/base.c txq->txq_stuck = 0; txq 1112 drivers/net/wireless/ath/ath5k/base.c struct ath5k_txq *txq; txq 1118 drivers/net/wireless/ath/ath5k/base.c txq = &ah->txqs[i]; txq 1119 drivers/net/wireless/ath/ath5k/base.c spin_lock_bh(&txq->lock); txq 1120 drivers/net/wireless/ath/ath5k/base.c list_for_each_entry_safe(bf, bf0, &txq->q, list) { txq 1128 drivers/net/wireless/ath/ath5k/base.c txq->txq_len--; txq 1131 drivers/net/wireless/ath/ath5k/base.c txq->link = NULL; txq 1132 drivers/net/wireless/ath/ath5k/base.c txq->txq_poll_mark = false; txq 1133 drivers/net/wireless/ath/ath5k/base.c spin_unlock_bh(&txq->lock); txq 1141 drivers/net/wireless/ath/ath5k/base.c struct ath5k_txq *txq = ah->txqs; txq 1144 drivers/net/wireless/ath/ath5k/base.c for (i = 0; i < ARRAY_SIZE(ah->txqs); i++, txq++) txq 1145 drivers/net/wireless/ath/ath5k/base.c if (txq->setup) { txq 1146 drivers/net/wireless/ath/ath5k/base.c ath5k_hw_release_tx_queue(ah, txq->qnum); txq 1147 drivers/net/wireless/ath/ath5k/base.c txq->setup = false; txq 1611 drivers/net/wireless/ath/ath5k/base.c struct ath5k_txq *txq, struct ieee80211_tx_control *control) txq 1618 drivers/net/wireless/ath/ath5k/base.c trace_ath5k_tx(ah, skb, txq); txq 1631 drivers/net/wireless/ath/ath5k/base.c if (txq->txq_len >= txq->txq_max && txq 1632 drivers/net/wireless/ath/ath5k/base.c txq->qnum <= AR5K_TX_QUEUE_ID_DATA_MAX) txq 1633 drivers/net/wireless/ath/ath5k/base.c ieee80211_stop_queue(hw, txq->qnum); txq 1651 drivers/net/wireless/ath/ath5k/base.c if (ath5k_txbuf_setup(ah, bf, txq, padsize, control)) { txq 1667 drivers/net/wireless/ath/ath5k/base.c struct ath5k_txq *txq, struct ath5k_tx_status *ts, txq 1727 drivers/net/wireless/ath/ath5k/base.c trace_ath5k_tx_complete(ah, skb, txq, ts); txq 1732 drivers/net/wireless/ath/ath5k/base.c ath5k_tx_processq(struct ath5k_hw *ah, struct ath5k_txq *txq) txq 1740 drivers/net/wireless/ath/ath5k/base.c spin_lock(&txq->lock); txq 1741 drivers/net/wireless/ath/ath5k/base.c list_for_each_entry_safe(bf, bf0, &txq->q, list) { txq 1743 drivers/net/wireless/ath/ath5k/base.c txq->txq_poll_mark = false; txq 1755 drivers/net/wireless/ath/ath5k/base.c "queue %u\n", ret, txq->qnum); txq 1764 drivers/net/wireless/ath/ath5k/base.c ath5k_tx_frame_completed(ah, skb, txq, &ts, bf); txq 1773 drivers/net/wireless/ath/ath5k/base.c if (ath5k_hw_get_txdp(ah, txq->qnum) != bf->daddr) { txq 1777 drivers/net/wireless/ath/ath5k/base.c txq->txq_len--; txq 1781 drivers/net/wireless/ath/ath5k/base.c spin_unlock(&txq->lock); txq 1782 drivers/net/wireless/ath/ath5k/base.c if (txq->txq_len < ATH5K_TXQ_LEN_LOW && txq->qnum < 4) txq 1783 drivers/net/wireless/ath/ath5k/base.c ieee80211_wake_queue(ah->hw, txq->qnum); txq 2465 drivers/net/wireless/ath/ath5k/base.c struct ath5k_txq *txq; txq 2476 drivers/net/wireless/ath/ath5k/base.c txq = &ah->txqs[i]; txq 2477 drivers/net/wireless/ath/ath5k/base.c spin_lock_bh(&txq->lock); txq 2478 drivers/net/wireless/ath/ath5k/base.c if (txq->txq_len > 1) { txq 2479 drivers/net/wireless/ath/ath5k/base.c if (txq->txq_poll_mark) { txq 2482 drivers/net/wireless/ath/ath5k/base.c txq->qnum); txq 2484 drivers/net/wireless/ath/ath5k/base.c txq->txq_stuck++; txq 2485 drivers/net/wireless/ath/ath5k/base.c spin_unlock_bh(&txq->lock); txq 2488 drivers/net/wireless/ath/ath5k/base.c txq->txq_poll_mark = true; txq 2491 drivers/net/wireless/ath/ath5k/base.c spin_unlock_bh(&txq->lock); txq 2986 drivers/net/wireless/ath/ath5k/base.c struct ath5k_txq *txq; txq 3036 drivers/net/wireless/ath/ath5k/base.c txq = ath5k_txq_setup(ah, AR5K_TX_QUEUE_DATA, AR5K_WME_AC_VO); txq 3037 drivers/net/wireless/ath/ath5k/base.c if (IS_ERR(txq)) { txq 3039 drivers/net/wireless/ath/ath5k/base.c ret = PTR_ERR(txq); txq 3042 drivers/net/wireless/ath/ath5k/base.c txq = ath5k_txq_setup(ah, AR5K_TX_QUEUE_DATA, AR5K_WME_AC_VI); txq 3043 drivers/net/wireless/ath/ath5k/base.c if (IS_ERR(txq)) { txq 3045 drivers/net/wireless/ath/ath5k/base.c ret = PTR_ERR(txq); txq 3048 drivers/net/wireless/ath/ath5k/base.c txq = ath5k_txq_setup(ah, AR5K_TX_QUEUE_DATA, AR5K_WME_AC_BE); txq 3049 drivers/net/wireless/ath/ath5k/base.c if (IS_ERR(txq)) { txq 3051 drivers/net/wireless/ath/ath5k/base.c ret = PTR_ERR(txq); txq 3054 drivers/net/wireless/ath/ath5k/base.c txq = ath5k_txq_setup(ah, AR5K_TX_QUEUE_DATA, AR5K_WME_AC_BK); txq 3055 drivers/net/wireless/ath/ath5k/base.c if (IS_ERR(txq)) { txq 3057 drivers/net/wireless/ath/ath5k/base.c ret = PTR_ERR(txq); txq 3063 drivers/net/wireless/ath/ath5k/base.c txq = ath5k_txq_setup(ah, AR5K_TX_QUEUE_DATA, AR5K_WME_AC_BE); txq 3064 drivers/net/wireless/ath/ath5k/base.c if (IS_ERR(txq)) { txq 3066 drivers/net/wireless/ath/ath5k/base.c ret = PTR_ERR(txq); txq 108 drivers/net/wireless/ath/ath5k/base.h struct ath5k_txq *txq, struct ieee80211_tx_control *control); txq 835 drivers/net/wireless/ath/ath5k/debug.c struct ath5k_txq *txq; txq 843 drivers/net/wireless/ath/ath5k/debug.c txq = &ah->txqs[i]; txq 846 drivers/net/wireless/ath/ath5k/debug.c "%02d: %ssetup\n", i, txq->setup ? "" : "not "); txq 848 drivers/net/wireless/ath/ath5k/debug.c if (!txq->setup) txq 852 drivers/net/wireless/ath/ath5k/debug.c spin_lock_bh(&txq->lock); txq 853 drivers/net/wireless/ath/ath5k/debug.c list_for_each_entry_safe(bf, bf0, &txq->q, list) txq 855 drivers/net/wireless/ath/ath5k/debug.c spin_unlock_bh(&txq->lock); txq 858 drivers/net/wireless/ath/ath5k/debug.c " len: %d bufs: %d\n", txq->txq_len, n); txq 860 drivers/net/wireless/ath/ath5k/debug.c " stuck: %d\n", txq->txq_stuck); txq 272 drivers/net/wireless/ath/ath6kl/debug.c get_queue_depth(&ep_dist->htc_ep->txq)); txq 744 drivers/net/wireless/ath/ath6kl/debug.c get_queue_depth(&ep_list->htc_ep->txq)); txq 511 drivers/net/wireless/ath/ath6kl/htc.h struct list_head txq; txq 426 drivers/net/wireless/ath/ath6kl/htc_mbox.c endpoint->cred_dist.txq_depth = get_queue_depth(&endpoint->txq); txq 439 drivers/net/wireless/ath/ath6kl/htc_mbox.c struct list_head *txq) txq 441 drivers/net/wireless/ath/ath6kl/htc_mbox.c if (list_empty(txq)) txq 446 drivers/net/wireless/ath/ath6kl/htc_mbox.c endpoint->eid, get_queue_depth(txq)); txq 448 drivers/net/wireless/ath/ath6kl/htc_mbox.c ath6kl_tx_complete(endpoint->target, txq); txq 611 drivers/net/wireless/ath/ath6kl/htc_mbox.c if (list_empty(&endpoint->txq)) txq 613 drivers/net/wireless/ath/ath6kl/htc_mbox.c packet = list_first_entry(&endpoint->txq, struct htc_packet, txq 618 drivers/net/wireless/ath/ath6kl/htc_mbox.c packet, get_queue_depth(&endpoint->txq)); txq 628 drivers/net/wireless/ath/ath6kl/htc_mbox.c packet = list_first_entry(&endpoint->txq, struct htc_packet, txq 861 drivers/net/wireless/ath/ath6kl/htc_mbox.c struct list_head txq; txq 882 drivers/net/wireless/ath/ath6kl/htc_mbox.c INIT_LIST_HEAD(&txq); txq 889 drivers/net/wireless/ath/ath6kl/htc_mbox.c if (list_empty(&endpoint->txq)) txq 892 drivers/net/wireless/ath/ath6kl/htc_mbox.c ath6kl_htc_tx_pkts_get(target, endpoint, &txq); txq 894 drivers/net/wireless/ath/ath6kl/htc_mbox.c if (list_empty(&txq)) txq 905 drivers/net/wireless/ath/ath6kl/htc_mbox.c (get_queue_depth(&txq) >= txq 911 drivers/net/wireless/ath/ath6kl/htc_mbox.c ath6kl_htc_tx_bundle(endpoint, &txq, txq 918 drivers/net/wireless/ath/ath6kl/htc_mbox.c if (list_empty(&txq)) txq 921 drivers/net/wireless/ath/ath6kl/htc_mbox.c packet = list_first_entry(&txq, struct htc_packet, txq 976 drivers/net/wireless/ath/ath6kl/htc_mbox.c txq_depth = get_queue_depth(&endpoint->txq); txq 997 drivers/net/wireless/ath/ath6kl/htc_mbox.c list_add_tail(&tx_pkt->list, &endpoint->txq); txq 1020 drivers/net/wireless/ath/ath6kl/htc_mbox.c if (!list_empty(&endpoint->txq)) { txq 1025 drivers/net/wireless/ath/ath6kl/htc_mbox.c get_queue_depth(&endpoint->txq)); txq 1167 drivers/net/wireless/ath/ath6kl/htc_mbox.c list_for_each_entry_safe(packet, tmp_pkt, &endpoint->txq, list) { txq 1233 drivers/net/wireless/ath/ath6kl/htc_mbox.c get_queue_depth(&endpoint->txq); txq 1649 drivers/net/wireless/ath/ath6kl/htc_mbox.c get_queue_depth(&endpoint->txq); txq 2588 drivers/net/wireless/ath/ath6kl/htc_mbox.c INIT_LIST_HEAD(&endpoint->txq); txq 106 drivers/net/wireless/ath/ath6kl/htc_pipe.c if (list_empty(&ep->txq)) txq 110 drivers/net/wireless/ath/ath6kl/htc_pipe.c packet = list_first_entry(&ep->txq, struct htc_packet, list); txq 114 drivers/net/wireless/ath/ath6kl/htc_pipe.c __func__, packet, get_queue_depth(&ep->txq)); txq 159 drivers/net/wireless/ath/ath6kl/htc_pipe.c packet = list_first_entry(&ep->txq, struct htc_packet, list); txq 183 drivers/net/wireless/ath/ath6kl/htc_pipe.c if (list_empty(&ep->txq)) txq 186 drivers/net/wireless/ath/ath6kl/htc_pipe.c packet = list_first_entry(&ep->txq, struct htc_packet, list); txq 191 drivers/net/wireless/ath/ath6kl/htc_pipe.c __func__, packet, get_queue_depth(&ep->txq)); txq 302 drivers/net/wireless/ath/ath6kl/htc_pipe.c struct list_head *txq) txq 312 drivers/net/wireless/ath/ath6kl/htc_pipe.c __func__, txq, txq 313 drivers/net/wireless/ath/ath6kl/htc_pipe.c (txq == NULL) ? 0 : get_queue_depth(txq)); txq 323 drivers/net/wireless/ath/ath6kl/htc_pipe.c if (txq != NULL) { txq 324 drivers/net/wireless/ath/ath6kl/htc_pipe.c if (list_empty(txq)) { txq 330 drivers/net/wireless/ath/ath6kl/htc_pipe.c txqueue_depth = get_queue_depth(&ep->txq); txq 335 drivers/net/wireless/ath/ath6kl/htc_pipe.c overflow = get_queue_depth(txq); txq 339 drivers/net/wireless/ath/ath6kl/htc_pipe.c overflow += get_queue_depth(txq); txq 358 drivers/net/wireless/ath/ath6kl/htc_pipe.c list_splice_tail_init(txq, &send_queue); txq 360 drivers/net/wireless/ath/ath6kl/htc_pipe.c good_pkts = get_queue_depth(txq) - overflow; txq 370 drivers/net/wireless/ath/ath6kl/htc_pipe.c packet = list_first_entry(txq, txq 383 drivers/net/wireless/ath/ath6kl/htc_pipe.c txq, list) { txq 421 drivers/net/wireless/ath/ath6kl/htc_pipe.c list_splice_tail_init(&send_queue, &ep->txq); txq 451 drivers/net/wireless/ath/ath6kl/htc_pipe.c if (get_queue_depth(&ep->txq) == 0) txq 675 drivers/net/wireless/ath/ath6kl/htc_pipe.c if (ep->cred_dist.credits && get_queue_depth(&ep->txq)) { txq 697 drivers/net/wireless/ath/ath6kl/htc_pipe.c while (get_queue_depth(&ep->txq)) { txq 698 drivers/net/wireless/ath/ath6kl/htc_pipe.c packet = list_first_entry(&ep->txq, struct htc_packet, list); txq 1183 drivers/net/wireless/ath/ath6kl/htc_pipe.c INIT_LIST_HEAD(&ep->txq); txq 177 drivers/net/wireless/ath/ath9k/ath9k.h s8 txq; txq 235 drivers/net/wireless/ath/ath9k/ath9k.h struct ath_txq *txq; txq 273 drivers/net/wireless/ath/ath9k/ath9k.h struct ath_txq *txq; txq 289 drivers/net/wireless/ath/ath9k/ath9k.h struct ath_txq txq[ATH9K_NUM_TX_QUEUES]; txq 431 drivers/net/wireless/ath/ath9k/ath9k.h struct ieee80211_txq *txq; txq 435 drivers/net/wireless/ath/ath9k/ath9k.h txq = sta->txq[tidno % ARRAY_SIZE(sta->txq)]; txq 437 drivers/net/wireless/ath/ath9k/ath9k.h txq = vif->txq; txq 439 drivers/net/wireless/ath/ath9k/ath9k.h return (struct ath_atx_tid *) txq->drv_priv; txq 565 drivers/net/wireless/ath/ath9k/ath9k.h static inline void ath_txq_lock(struct ath_softc *sc, struct ath_txq *txq) txq 567 drivers/net/wireless/ath/ath9k/ath9k.h spin_lock_bh(&txq->axq_lock); txq 569 drivers/net/wireless/ath/ath9k/ath9k.h static inline void ath_txq_unlock(struct ath_softc *sc, struct ath_txq *txq) txq 571 drivers/net/wireless/ath/ath9k/ath9k.h spin_unlock_bh(&txq->axq_lock); txq 581 drivers/net/wireless/ath/ath9k/ath9k.h void ath_txq_unlock_complete(struct ath_softc *sc, struct ath_txq *txq); txq 582 drivers/net/wireless/ath/ath9k/ath9k.h void ath_tx_cleanupq(struct ath_softc *sc, struct ath_txq *txq); txq 584 drivers/net/wireless/ath/ath9k/ath9k.h void ath_draintxq(struct ath_softc *sc, struct ath_txq *txq); txq 587 drivers/net/wireless/ath/ath9k/ath9k.h void ath_txq_schedule(struct ath_softc *sc, struct ath_txq *txq); txq 38 drivers/net/wireless/ath/ath9k/beacon.c struct ath_txq *txq; txq 50 drivers/net/wireless/ath/ath9k/beacon.c txq = sc->tx.txq_map[IEEE80211_AC_BE]; txq 51 drivers/net/wireless/ath/ath9k/beacon.c ath9k_hw_get_txq_props(ah, txq->axq_qnum, &qi_be); txq 1016 drivers/net/wireless/ath/ath9k/channel.c txctl.txq = sc->tx.txq_map[IEEE80211_AC_VO]; txq 1137 drivers/net/wireless/ath/ath9k/channel.c txctl.txq = sc->tx.txq_map[IEEE80211_AC_VO]; txq 623 drivers/net/wireless/ath/ath9k/debug.c static void print_queue(struct ath_softc *sc, struct ath_txq *txq, txq 626 drivers/net/wireless/ath/ath9k/debug.c ath_txq_lock(sc, txq); txq 628 drivers/net/wireless/ath/ath9k/debug.c seq_printf(file, "%s: %d ", "qnum", txq->axq_qnum); txq 629 drivers/net/wireless/ath/ath9k/debug.c seq_printf(file, "%s: %2d ", "qdepth", txq->axq_depth); txq 630 drivers/net/wireless/ath/ath9k/debug.c seq_printf(file, "%s: %2d ", "ampdu-depth", txq->axq_ampdu_depth); txq 631 drivers/net/wireless/ath/ath9k/debug.c seq_printf(file, "%s: %3d\n", "pending", txq->pending_frames); txq 633 drivers/net/wireless/ath/ath9k/debug.c ath_txq_unlock(sc, txq); txq 640 drivers/net/wireless/ath/ath9k/debug.c struct ath_txq *txq; txq 647 drivers/net/wireless/ath/ath9k/debug.c txq = sc->tx.txq_map[i]; txq 649 drivers/net/wireless/ath/ath9k/debug.c print_queue(sc, txq, file); txq 783 drivers/net/wireless/ath/ath9k/debug.c struct ath_tx_status *ts, struct ath_txq *txq, txq 786 drivers/net/wireless/ath/ath9k/debug.c int qnum = txq->axq_qnum; txq 259 drivers/net/wireless/ath/ath9k/debug.h struct ath_tx_status *ts, struct ath_txq *txq, txq 296 drivers/net/wireless/ath/ath9k/debug.h struct ath_txq *txq, txq 29 drivers/net/wireless/ath/ath9k/debug_sta.c struct ath_txq *txq; txq 57 drivers/net/wireless/ath/ath9k/debug_sta.c txq = tid->txq; txq 58 drivers/net/wireless/ath/ath9k/debug_sta.c ath_txq_lock(sc, txq); txq 71 drivers/net/wireless/ath/ath9k/debug_sta.c ath_txq_unlock(sc, txq); txq 414 drivers/net/wireless/ath/ath9k/gpio.c struct ath_txq *txq; txq 429 drivers/net/wireless/ath/ath9k/gpio.c txq = sc->tx.txq_map[IEEE80211_AC_BE]; txq 430 drivers/net/wireless/ath/ath9k/gpio.c ath9k_hw_init_btcoex_hw(sc->sc_ah, txq->axq_qnum); txq 821 drivers/net/wireless/ath/ath9k/hw.h struct ath9k_tx_queue_info txq[ATH9K_NUM_TX_QUEUES]; txq 784 drivers/net/wireless/ath/ath9k/init.c ath_tx_cleanupq(sc, &sc->tx.txq[i]); txq 1109 drivers/net/wireless/ath/ath9k/init.c ath_tx_cleanupq(sc, &sc->tx.txq[i]); txq 25 drivers/net/wireless/ath/ath9k/link.c struct ath_txq *txq; txq 32 drivers/net/wireless/ath/ath9k/link.c txq = sc->tx.txq_map[i]; txq 34 drivers/net/wireless/ath/ath9k/link.c ath_txq_lock(sc, txq); txq 35 drivers/net/wireless/ath/ath9k/link.c if (txq->axq_depth) { txq 36 drivers/net/wireless/ath/ath9k/link.c if (txq->axq_tx_inprogress) { txq 37 drivers/net/wireless/ath/ath9k/link.c ath_txq_unlock(sc, txq); txq 41 drivers/net/wireless/ath/ath9k/link.c txq->axq_tx_inprogress = true; txq 43 drivers/net/wireless/ath/ath9k/link.c ath_txq_unlock(sc, txq); txq 182 drivers/net/wireless/ath/ath9k/link.c txctl.txq = sc->tx.txq_map[IEEE80211_AC_BE]; txq 203 drivers/net/wireless/ath/ath9k/mac.c qi = &ah->txq[q]; txq 267 drivers/net/wireless/ath/ath9k/mac.c qi = &ah->txq[q]; txq 323 drivers/net/wireless/ath/ath9k/mac.c qi = &ah->txq[q]; txq 351 drivers/net/wireless/ath/ath9k/mac.c qi = &ah->txq[q]; txq 373 drivers/net/wireless/ath/ath9k/mac.c qi = &ah->txq[q]; txq 57 drivers/net/wireless/ath/ath9k/main.c static bool ath9k_has_pending_frames(struct ath_softc *sc, struct ath_txq *txq, txq 62 drivers/net/wireless/ath/ath9k/main.c spin_lock_bh(&txq->axq_lock); txq 64 drivers/net/wireless/ath/ath9k/main.c if (txq->axq_depth) { txq 72 drivers/net/wireless/ath/ath9k/main.c if (txq->mac80211_qnum >= 0) { txq 75 drivers/net/wireless/ath/ath9k/main.c acq = &sc->cur_chan->acq[txq->mac80211_qnum]; txq 80 drivers/net/wireless/ath/ath9k/main.c spin_unlock_bh(&txq->axq_lock); txq 805 drivers/net/wireless/ath/ath9k/main.c txctl.txq = sc->tx.txq_map[skb_get_queue_mapping(skb)]; txq 812 drivers/net/wireless/ath/ath9k/main.c TX_STAT_INC(sc, txctl.txq->axq_qnum, txfailed); txq 1636 drivers/net/wireless/ath/ath9k/main.c struct ath_txq *txq; txq 1643 drivers/net/wireless/ath/ath9k/main.c txq = sc->tx.txq_map[queue]; txq 1657 drivers/net/wireless/ath/ath9k/main.c queue, txq->axq_qnum, params->aifs, params->cw_min, txq 1661 drivers/net/wireless/ath/ath9k/main.c ret = ath_txq_update(sc, txq->axq_qnum, &qi); txq 2048 drivers/net/wireless/ath/ath9k/main.c npend = ath9k_has_pending_frames(sc, &sc->tx.txq[i], txq 126 drivers/net/wireless/ath/ath9k/tx99.c txctl.txq = sc->tx.txq_map[IEEE80211_AC_VO]; txq 50 drivers/net/wireless/ath/ath9k/xmit.c static void ath_tx_send_normal(struct ath_softc *sc, struct ath_txq *txq, txq 53 drivers/net/wireless/ath/ath9k/xmit.c int tx_flags, struct ath_txq *txq, txq 56 drivers/net/wireless/ath/ath9k/xmit.c struct ath_txq *txq, struct list_head *bf_q, txq 59 drivers/net/wireless/ath/ath9k/xmit.c static void ath_tx_txqaddbuf(struct ath_softc *sc, struct ath_txq *txq, txq 67 drivers/net/wireless/ath/ath9k/xmit.c struct ath_txq *txq, txq 101 drivers/net/wireless/ath/ath9k/xmit.c void ath_txq_unlock_complete(struct ath_softc *sc, struct ath_txq *txq) txq 102 drivers/net/wireless/ath/ath9k/xmit.c __releases(&txq->axq_lock) txq 109 drivers/net/wireless/ath/ath9k/xmit.c skb_queue_splice_init(&txq->complete_q, &q); txq 110 drivers/net/wireless/ath/ath9k/xmit.c spin_unlock_bh(&txq->axq_lock); txq 129 drivers/net/wireless/ath/ath9k/xmit.c struct ath_txq *txq = tid->txq; txq 135 drivers/net/wireless/ath/ath9k/xmit.c ath_txq_lock(sc, txq); txq 136 drivers/net/wireless/ath/ath9k/xmit.c ath_txq_schedule(sc, txq); txq 137 drivers/net/wireless/ath/ath9k/xmit.c ath_txq_unlock(sc, txq); txq 164 drivers/net/wireless/ath/ath9k/xmit.c static void ath_txq_skb_done(struct ath_softc *sc, struct ath_txq *txq, txq 168 drivers/net/wireless/ath/ath9k/xmit.c int q = fi->txq; txq 173 drivers/net/wireless/ath/ath9k/xmit.c txq = sc->tx.txq_map[q]; txq 174 drivers/net/wireless/ath/ath9k/xmit.c if (WARN_ON(--txq->pending_frames < 0)) txq 175 drivers/net/wireless/ath/ath9k/xmit.c txq->pending_frames = 0; txq 189 drivers/net/wireless/ath/ath9k/xmit.c struct ieee80211_txq *txq = container_of((void*)tid, struct ieee80211_txq, drv_priv); txq 193 drivers/net/wireless/ath/ath9k/xmit.c .txq = tid->txq, txq 200 drivers/net/wireless/ath/ath9k/xmit.c skb = ieee80211_tx_dequeue(hw, txq); txq 211 drivers/net/wireless/ath/ath9k/xmit.c if (tid->txq == sc->tx.txq_map[q]) { txq 213 drivers/net/wireless/ath/ath9k/xmit.c fi->txq = q; txq 214 drivers/net/wireless/ath/ath9k/xmit.c ++tid->txq->pending_frames; txq 234 drivers/net/wireless/ath/ath9k/xmit.c struct ath_txq *txq = tid->txq; txq 250 drivers/net/wireless/ath/ath9k/xmit.c ath_txq_skb_done(sc, txq, skb); txq 261 drivers/net/wireless/ath/ath9k/xmit.c ath_tx_complete_buf(sc, bf, txq, &bf_head, NULL, &ts, 0); txq 265 drivers/net/wireless/ath/ath9k/xmit.c ath_txq_unlock(sc, txq); txq 267 drivers/net/wireless/ath/ath9k/xmit.c ath_txq_lock(sc, txq); txq 316 drivers/net/wireless/ath/ath9k/xmit.c static void ath_tid_drain(struct ath_softc *sc, struct ath_txq *txq, txq 335 drivers/net/wireless/ath/ath9k/xmit.c ath_tx_complete(sc, skb, ATH_TX_ERROR, txq, NULL); txq 340 drivers/net/wireless/ath/ath9k/xmit.c ath_tx_complete_buf(sc, bf, txq, &bf_head, NULL, &ts, 0); txq 344 drivers/net/wireless/ath/ath9k/xmit.c static void ath_tx_set_retry(struct ath_softc *sc, struct ath_txq *txq, txq 352 drivers/net/wireless/ath/ath9k/xmit.c TX_STAT_INC(sc, txq->axq_qnum, a_retries); txq 439 drivers/net/wireless/ath/ath9k/xmit.c static void ath_tx_complete_aggr(struct ath_softc *sc, struct ath_txq *txq, txq 479 drivers/net/wireless/ath/ath9k/xmit.c ath_tx_complete_buf(sc, bf, txq, &bf_head, NULL, ts, 0); txq 552 drivers/net/wireless/ath/ath9k/xmit.c ath_tx_set_retry(sc, txq, bf->bf_mpdu, txq 588 drivers/net/wireless/ath/ath9k/xmit.c ath_tx_complete_buf(sc, bf, txq, &bf_head, sta, ts, txq 608 drivers/net/wireless/ath/ath9k/xmit.c ath_tx_complete_buf(sc, bf, txq, txq 648 drivers/net/wireless/ath/ath9k/xmit.c ath_txq_unlock(sc, txq); txq 650 drivers/net/wireless/ath/ath9k/xmit.c ath_txq_lock(sc, txq); txq 681 drivers/net/wireless/ath/ath9k/xmit.c static void ath_tx_process_buffer(struct ath_softc *sc, struct ath_txq *txq, txq 694 drivers/net/wireless/ath/ath9k/xmit.c txq->axq_tx_inprogress = false; txq 696 drivers/net/wireless/ath/ath9k/xmit.c txq->axq_depth--; txq 698 drivers/net/wireless/ath/ath9k/xmit.c txq->axq_ampdu_depth--; txq 722 drivers/net/wireless/ath/ath9k/xmit.c ath_tx_complete_buf(sc, bf, txq, bf_head, sta, ts, txok); txq 724 drivers/net/wireless/ath/ath9k/xmit.c ath_tx_complete_aggr(sc, txq, bf, bf_head, sta, tid, ts, txok); txq 727 drivers/net/wireless/ath/ath9k/xmit.c ath_txq_schedule(sc, txq); txq 760 drivers/net/wireless/ath/ath9k/xmit.c int q = tid->txq->mac80211_qnum; txq 894 drivers/net/wireless/ath/ath9k/xmit.c ath_tx_get_tid_subframe(struct ath_softc *sc, struct ath_txq *txq, txq 912 drivers/net/wireless/ath/ath9k/xmit.c bf = ath_tx_setup_buffer(sc, txq, tid, skb); txq 917 drivers/net/wireless/ath/ath9k/xmit.c ath_txq_skb_done(sc, txq, skb); txq 968 drivers/net/wireless/ath/ath9k/xmit.c ath_tx_complete_buf(sc, bf, txq, &bf_head, NULL, &ts, 0); txq 983 drivers/net/wireless/ath/ath9k/xmit.c ath_tx_form_aggr(struct ath_softc *sc, struct ath_txq *txq, txq 1041 drivers/net/wireless/ath/ath9k/xmit.c ret = ath_tx_get_tid_subframe(sc, txq, tid, &bf); txq 1056 drivers/net/wireless/ath/ath9k/xmit.c TX_STAT_INC(sc, txq->axq_qnum, a_aggr); txq 1343 drivers/net/wireless/ath/ath9k/xmit.c struct ath_txq *txq, int len) txq 1354 drivers/net/wireless/ath/ath9k/xmit.c info.qcu = txq->axq_qnum; txq 1374 drivers/net/wireless/ath/ath9k/xmit.c txq == sc->tx.uapsdq) txq 1435 drivers/net/wireless/ath/ath9k/xmit.c ath_tx_form_burst(struct ath_softc *sc, struct ath_txq *txq, txq 1454 drivers/net/wireless/ath/ath9k/xmit.c ret = ath_tx_get_tid_subframe(sc, txq, tid, &bf); txq 1468 drivers/net/wireless/ath/ath9k/xmit.c static int ath_tx_sched_aggr(struct ath_softc *sc, struct ath_txq *txq, txq 1479 drivers/net/wireless/ath/ath9k/xmit.c ret = ath_tx_get_tid_subframe(sc, txq, tid, &bf); txq 1485 drivers/net/wireless/ath/ath9k/xmit.c if ((aggr && txq->axq_ampdu_depth >= ATH_AGGR_MIN_QDEPTH) || txq 1486 drivers/net/wireless/ath/ath9k/xmit.c (!aggr && txq->axq_depth >= ATH_NON_AGGR_MIN_QDEPTH)) { txq 1493 drivers/net/wireless/ath/ath9k/xmit.c aggr_len = ath_tx_form_aggr(sc, txq, tid, &bf_q, bf); txq 1495 drivers/net/wireless/ath/ath9k/xmit.c ath_tx_form_burst(sc, txq, tid, &bf_q, bf); txq 1505 drivers/net/wireless/ath/ath9k/xmit.c ath_tx_fill_desc(sc, bf, txq, aggr_len); txq 1506 drivers/net/wireless/ath/ath9k/xmit.c ath_tx_txqaddbuf(sc, txq, &bf_q, false); txq 1515 drivers/net/wireless/ath/ath9k/xmit.c struct ath_txq *txq; txq 1523 drivers/net/wireless/ath/ath9k/xmit.c txq = txtid->txq; txq 1525 drivers/net/wireless/ath/ath9k/xmit.c ath_txq_lock(sc, txq); txq 1545 drivers/net/wireless/ath/ath9k/xmit.c ath_txq_unlock_complete(sc, txq); txq 1555 drivers/net/wireless/ath/ath9k/xmit.c struct ath_txq *txq = txtid->txq; txq 1559 drivers/net/wireless/ath/ath9k/xmit.c ath_txq_lock(sc, txq); txq 1562 drivers/net/wireless/ath/ath9k/xmit.c ath_txq_unlock_complete(sc, txq); txq 1587 drivers/net/wireless/ath/ath9k/xmit.c struct ath_txq *txq; txq 1594 drivers/net/wireless/ath/ath9k/xmit.c txq = tid->txq; txq 1596 drivers/net/wireless/ath/ath9k/xmit.c ath_txq_lock(sc, txq); txq 1600 drivers/net/wireless/ath/ath9k/xmit.c ath_txq_schedule(sc, txq); txq 1602 drivers/net/wireless/ath/ath9k/xmit.c ath_txq_unlock_complete(sc, txq); txq 1631 drivers/net/wireless/ath/ath9k/xmit.c struct ath_txq *txq = sc->tx.uapsdq; txq 1647 drivers/net/wireless/ath/ath9k/xmit.c ath_txq_lock(sc, tid->txq); txq 1665 drivers/net/wireless/ath/ath9k/xmit.c TX_STAT_INC(sc, txq->axq_qnum, a_queued_hw); txq 1670 drivers/net/wireless/ath/ath9k/xmit.c ath_txq_unlock_complete(sc, tid->txq); txq 1683 drivers/net/wireless/ath/ath9k/xmit.c ath_txq_lock(sc, txq); txq 1684 drivers/net/wireless/ath/ath9k/xmit.c ath_tx_fill_desc(sc, bf, txq, 0); txq 1685 drivers/net/wireless/ath/ath9k/xmit.c ath_tx_txqaddbuf(sc, txq, &bf_q, false); txq 1686 drivers/net/wireless/ath/ath9k/xmit.c ath_txq_unlock(sc, txq); txq 1745 drivers/net/wireless/ath/ath9k/xmit.c struct ath_txq *txq = &sc->tx.txq[axq_qnum]; txq 1747 drivers/net/wireless/ath/ath9k/xmit.c txq->axq_qnum = axq_qnum; txq 1748 drivers/net/wireless/ath/ath9k/xmit.c txq->mac80211_qnum = -1; txq 1749 drivers/net/wireless/ath/ath9k/xmit.c txq->axq_link = NULL; txq 1750 drivers/net/wireless/ath/ath9k/xmit.c __skb_queue_head_init(&txq->complete_q); txq 1751 drivers/net/wireless/ath/ath9k/xmit.c INIT_LIST_HEAD(&txq->axq_q); txq 1752 drivers/net/wireless/ath/ath9k/xmit.c spin_lock_init(&txq->axq_lock); txq 1753 drivers/net/wireless/ath/ath9k/xmit.c txq->axq_depth = 0; txq 1754 drivers/net/wireless/ath/ath9k/xmit.c txq->axq_ampdu_depth = 0; txq 1755 drivers/net/wireless/ath/ath9k/xmit.c txq->axq_tx_inprogress = false; txq 1758 drivers/net/wireless/ath/ath9k/xmit.c txq->txq_headidx = txq->txq_tailidx = 0; txq 1760 drivers/net/wireless/ath/ath9k/xmit.c INIT_LIST_HEAD(&txq->txq_fifo[i]); txq 1762 drivers/net/wireless/ath/ath9k/xmit.c return &sc->tx.txq[axq_qnum]; txq 1772 drivers/net/wireless/ath/ath9k/xmit.c BUG_ON(sc->tx.txq[qnum].axq_qnum != qnum); txq 1807 drivers/net/wireless/ath/ath9k/xmit.c static void ath_drain_txq_list(struct ath_softc *sc, struct ath_txq *txq, txq 1830 drivers/net/wireless/ath/ath9k/xmit.c ath_tx_process_buffer(sc, txq, &ts, bf, &bf_head); txq 1840 drivers/net/wireless/ath/ath9k/xmit.c void ath_draintxq(struct ath_softc *sc, struct ath_txq *txq) txq 1843 drivers/net/wireless/ath/ath9k/xmit.c ath_txq_lock(sc, txq); txq 1846 drivers/net/wireless/ath/ath9k/xmit.c int idx = txq->txq_tailidx; txq 1848 drivers/net/wireless/ath/ath9k/xmit.c while (!list_empty(&txq->txq_fifo[idx])) { txq 1849 drivers/net/wireless/ath/ath9k/xmit.c ath_drain_txq_list(sc, txq, &txq->txq_fifo[idx]); txq 1853 drivers/net/wireless/ath/ath9k/xmit.c txq->txq_tailidx = idx; txq 1856 drivers/net/wireless/ath/ath9k/xmit.c txq->axq_link = NULL; txq 1857 drivers/net/wireless/ath/ath9k/xmit.c txq->axq_tx_inprogress = false; txq 1858 drivers/net/wireless/ath/ath9k/xmit.c ath_drain_txq_list(sc, txq, &txq->axq_q); txq 1860 drivers/net/wireless/ath/ath9k/xmit.c ath_txq_unlock_complete(sc, txq); txq 1868 drivers/net/wireless/ath/ath9k/xmit.c struct ath_txq *txq; txq 1882 drivers/net/wireless/ath/ath9k/xmit.c if (!sc->tx.txq[i].axq_depth) txq 1885 drivers/net/wireless/ath/ath9k/xmit.c if (ath9k_hw_numtxpending(ah, sc->tx.txq[i].axq_qnum)) txq 1899 drivers/net/wireless/ath/ath9k/xmit.c txq = &sc->tx.txq[i]; txq 1900 drivers/net/wireless/ath/ath9k/xmit.c ath_draintxq(sc, txq); txq 1906 drivers/net/wireless/ath/ath9k/xmit.c void ath_tx_cleanupq(struct ath_softc *sc, struct ath_txq *txq) txq 1908 drivers/net/wireless/ath/ath9k/xmit.c ath9k_hw_releasetxqueue(sc->sc_ah, txq->axq_qnum); txq 1909 drivers/net/wireless/ath/ath9k/xmit.c sc->tx.txqsetup &= ~(1<<txq->axq_qnum); txq 1915 drivers/net/wireless/ath/ath9k/xmit.c void ath_txq_schedule(struct ath_softc *sc, struct ath_txq *txq) txq 1923 drivers/net/wireless/ath/ath9k/xmit.c if (txq->mac80211_qnum < 0) txq 1929 drivers/net/wireless/ath/ath9k/xmit.c ieee80211_txq_schedule_start(hw, txq->mac80211_qnum); txq 1936 drivers/net/wireless/ath/ath9k/xmit.c while ((queue = ieee80211_next_txq(hw, txq->mac80211_qnum))) { txq 1941 drivers/net/wireless/ath/ath9k/xmit.c ret = ath_tx_sched_aggr(sc, txq, tid); txq 1951 drivers/net/wireless/ath/ath9k/xmit.c ieee80211_txq_schedule_end(hw, txq->mac80211_qnum); txq 1956 drivers/net/wireless/ath/ath9k/xmit.c struct ath_txq *txq; txq 1960 drivers/net/wireless/ath/ath9k/xmit.c txq = sc->tx.txq_map[i]; txq 1962 drivers/net/wireless/ath/ath9k/xmit.c spin_lock_bh(&txq->axq_lock); txq 1963 drivers/net/wireless/ath/ath9k/xmit.c ath_txq_schedule(sc, txq); txq 1964 drivers/net/wireless/ath/ath9k/xmit.c spin_unlock_bh(&txq->axq_lock); txq 1976 drivers/net/wireless/ath/ath9k/xmit.c static void ath_tx_txqaddbuf(struct ath_softc *sc, struct ath_txq *txq, txq 1998 drivers/net/wireless/ath/ath9k/xmit.c txq->axq_qnum, txq->axq_depth); txq 2000 drivers/net/wireless/ath/ath9k/xmit.c if (edma && list_empty(&txq->txq_fifo[txq->txq_headidx])) { txq 2001 drivers/net/wireless/ath/ath9k/xmit.c list_splice_tail_init(head, &txq->txq_fifo[txq->txq_headidx]); txq 2002 drivers/net/wireless/ath/ath9k/xmit.c INCR(txq->txq_headidx, ATH_TXFIFO_DEPTH); txq 2005 drivers/net/wireless/ath/ath9k/xmit.c list_splice_tail_init(head, &txq->axq_q); txq 2007 drivers/net/wireless/ath/ath9k/xmit.c if (txq->axq_link) { txq 2008 drivers/net/wireless/ath/ath9k/xmit.c ath9k_hw_set_desc_link(ah, txq->axq_link, bf->bf_daddr); txq 2010 drivers/net/wireless/ath/ath9k/xmit.c txq->axq_qnum, txq->axq_link, txq 2015 drivers/net/wireless/ath/ath9k/xmit.c txq->axq_link = bf_last->bf_desc; txq 2019 drivers/net/wireless/ath/ath9k/xmit.c TX_STAT_INC(sc, txq->axq_qnum, puttxbuf); txq 2020 drivers/net/wireless/ath/ath9k/xmit.c ath9k_hw_puttxbuf(ah, txq->axq_qnum, bf->bf_daddr); txq 2022 drivers/net/wireless/ath/ath9k/xmit.c txq->axq_qnum, ito64(bf->bf_daddr), bf->bf_desc); txq 2026 drivers/net/wireless/ath/ath9k/xmit.c TX_STAT_INC(sc, txq->axq_qnum, txstart); txq 2027 drivers/net/wireless/ath/ath9k/xmit.c ath9k_hw_txstart(ah, txq->axq_qnum); txq 2032 drivers/net/wireless/ath/ath9k/xmit.c txq->axq_depth++; txq 2034 drivers/net/wireless/ath/ath9k/xmit.c txq->axq_ampdu_depth++; txq 2043 drivers/net/wireless/ath/ath9k/xmit.c static void ath_tx_send_normal(struct ath_softc *sc, struct ath_txq *txq, txq 2061 drivers/net/wireless/ath/ath9k/xmit.c ath_tx_fill_desc(sc, bf, txq, fi->framelen); txq 2062 drivers/net/wireless/ath/ath9k/xmit.c ath_tx_txqaddbuf(sc, txq, &bf_head, false); txq 2063 drivers/net/wireless/ath/ath9k/xmit.c TX_STAT_INC(sc, txq->axq_qnum, queued); txq 2107 drivers/net/wireless/ath/ath9k/xmit.c fi->txq = -1; txq 2145 drivers/net/wireless/ath/ath9k/xmit.c struct ath_txq *txq, txq 2273 drivers/net/wireless/ath/ath9k/xmit.c struct ath_txq *txq = txctl->txq; txq 2294 drivers/net/wireless/ath/ath9k/xmit.c txq = sc->tx.uapsdq; txq 2301 drivers/net/wireless/ath/ath9k/xmit.c ath_txq_lock(sc, txq); txq 2302 drivers/net/wireless/ath/ath9k/xmit.c if (txq == sc->tx.txq_map[q]) { txq 2303 drivers/net/wireless/ath/ath9k/xmit.c fi->txq = q; txq 2304 drivers/net/wireless/ath/ath9k/xmit.c ++txq->pending_frames; txq 2307 drivers/net/wireless/ath/ath9k/xmit.c bf = ath_tx_setup_buffer(sc, txq, tid, skb); txq 2309 drivers/net/wireless/ath/ath9k/xmit.c ath_txq_skb_done(sc, txq, skb); txq 2323 drivers/net/wireless/ath/ath9k/xmit.c ath_tx_send_normal(sc, txq, tid, skb); txq 2326 drivers/net/wireless/ath/ath9k/xmit.c ath_txq_unlock(sc, txq); txq 2336 drivers/net/wireless/ath/ath9k/xmit.c .txq = sc->beacon.cabq txq 2355 drivers/net/wireless/ath/ath9k/xmit.c bf = ath_tx_setup_buffer(sc, txctl.txq, NULL, skb); txq 2386 drivers/net/wireless/ath/ath9k/xmit.c ath_txq_lock(sc, txctl.txq); txq 2387 drivers/net/wireless/ath/ath9k/xmit.c ath_tx_fill_desc(sc, bf, txctl.txq, 0); txq 2388 drivers/net/wireless/ath/ath9k/xmit.c ath_tx_txqaddbuf(sc, txctl.txq, &bf_q, false); txq 2389 drivers/net/wireless/ath/ath9k/xmit.c TX_STAT_INC(sc, txctl.txq->axq_qnum, queued); txq 2390 drivers/net/wireless/ath/ath9k/xmit.c ath_txq_unlock(sc, txctl.txq); txq 2398 drivers/net/wireless/ath/ath9k/xmit.c int tx_flags, struct ath_txq *txq, txq 2433 drivers/net/wireless/ath/ath9k/xmit.c if ((sc->ps_flags & PS_WAIT_FOR_TX_ACK) && !txq->axq_depth) { txq 2444 drivers/net/wireless/ath/ath9k/xmit.c ath_txq_skb_done(sc, txq, skb); txq 2446 drivers/net/wireless/ath/ath9k/xmit.c __skb_queue_tail(&txq->complete_q, skb); txq 2450 drivers/net/wireless/ath/ath9k/xmit.c struct ath_txq *txq, struct list_head *bf_q, txq 2478 drivers/net/wireless/ath/ath9k/xmit.c ath_debug_stat_tx(sc, bf, ts, txq, tx_flags); txq 2479 drivers/net/wireless/ath/ath9k/xmit.c ath_tx_complete(sc, skb, tx_flags, txq, sta); txq 2553 drivers/net/wireless/ath/ath9k/xmit.c static void ath_tx_processq(struct ath_softc *sc, struct ath_txq *txq) txq 2564 drivers/net/wireless/ath/ath9k/xmit.c txq->axq_qnum, ath9k_hw_gettxbuf(sc->sc_ah, txq->axq_qnum), txq 2565 drivers/net/wireless/ath/ath9k/xmit.c txq->axq_link); txq 2567 drivers/net/wireless/ath/ath9k/xmit.c ath_txq_lock(sc, txq); txq 2572 drivers/net/wireless/ath/ath9k/xmit.c if (list_empty(&txq->axq_q)) { txq 2573 drivers/net/wireless/ath/ath9k/xmit.c txq->axq_link = NULL; txq 2574 drivers/net/wireless/ath/ath9k/xmit.c ath_txq_schedule(sc, txq); txq 2577 drivers/net/wireless/ath/ath9k/xmit.c bf = list_first_entry(&txq->axq_q, struct ath_buf, list); txq 2590 drivers/net/wireless/ath/ath9k/xmit.c if (list_is_last(&bf_held->list, &txq->axq_q)) txq 2605 drivers/net/wireless/ath/ath9k/xmit.c TX_STAT_INC(sc, txq->axq_qnum, txprocdesc); txq 2616 drivers/net/wireless/ath/ath9k/xmit.c &txq->axq_q, lastbf->list.prev); txq 2623 drivers/net/wireless/ath/ath9k/xmit.c ath_tx_process_buffer(sc, txq, &ts, bf, &bf_head); txq 2625 drivers/net/wireless/ath/ath9k/xmit.c ath_txq_unlock_complete(sc, txq); txq 2637 drivers/net/wireless/ath/ath9k/xmit.c ath_tx_processq(sc, &sc->tx.txq[i]); txq 2647 drivers/net/wireless/ath/ath9k/xmit.c struct ath_txq *txq; txq 2680 drivers/net/wireless/ath/ath9k/xmit.c txq = &sc->tx.txq[ts.qid]; txq 2682 drivers/net/wireless/ath/ath9k/xmit.c ath_txq_lock(sc, txq); txq 2684 drivers/net/wireless/ath/ath9k/xmit.c TX_STAT_INC(sc, txq->axq_qnum, txprocdesc); txq 2686 drivers/net/wireless/ath/ath9k/xmit.c fifo_list = &txq->txq_fifo[txq->txq_tailidx]; txq 2688 drivers/net/wireless/ath/ath9k/xmit.c ath_txq_unlock(sc, txq); txq 2704 drivers/net/wireless/ath/ath9k/xmit.c INCR(txq->txq_tailidx, ATH_TXFIFO_DEPTH); txq 2706 drivers/net/wireless/ath/ath9k/xmit.c if (!list_empty(&txq->axq_q)) { txq 2710 drivers/net/wireless/ath/ath9k/xmit.c txq->axq_link = NULL; txq 2711 drivers/net/wireless/ath/ath9k/xmit.c list_splice_tail_init(&txq->axq_q, &bf_q); txq 2712 drivers/net/wireless/ath/ath9k/xmit.c ath_tx_txqaddbuf(sc, txq, &bf_q, true); txq 2721 drivers/net/wireless/ath/ath9k/xmit.c ath_tx_process_buffer(sc, txq, &ts, bf, &bf_head); txq 2722 drivers/net/wireless/ath/ath9k/xmit.c ath_txq_unlock_complete(sc, txq); txq 2804 drivers/net/wireless/ath/ath9k/xmit.c tid->txq = sc->tx.txq_map[acno]; txq 2814 drivers/net/wireless/ath/ath9k/xmit.c struct ath_txq *txq; txq 2821 drivers/net/wireless/ath/ath9k/xmit.c txq = tid->txq; txq 2823 drivers/net/wireless/ath/ath9k/xmit.c ath_txq_lock(sc, txq); txq 2828 drivers/net/wireless/ath/ath9k/xmit.c ath_tid_drain(sc, txq, tid); txq 2831 drivers/net/wireless/ath/ath9k/xmit.c ath_txq_unlock(sc, txq); txq 2869 drivers/net/wireless/ath/ath9k/xmit.c bf = ath_tx_setup_buffer(sc, txctl->txq, NULL, skb); txq 2878 drivers/net/wireless/ath/ath9k/xmit.c ath9k_hw_tx99_start(sc->sc_ah, txctl->txq->axq_qnum); txq 2880 drivers/net/wireless/ath/ath9k/xmit.c ath_tx_send_normal(sc, txctl->txq, NULL, skb); txq 1495 drivers/net/wireless/ath/wil6210/debugfs.c struct netdev_queue *txq = netdev_get_tx_queue(ndev, i); txq 1496 drivers/net/wireless/ath/wil6210/debugfs.c unsigned long state = txq->state; txq 829 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c struct pktq *txq; txq 834 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c txq = brcmf_bus_gettxq(fws->drvr->bus_if); txq 835 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c if (IS_ERR(txq)) { txq 840 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c for (prec = 0; prec < txq->num_prec; prec++) { txq 841 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c skb = brcmu_pktq_pdeq_match(txq, prec, fn, &ifidx); txq 848 drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c skb = brcmu_pktq_pdeq_match(txq, prec, fn, &ifidx); txq 444 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c struct pktq txq; /* Queue length used for flow-control */ txq 2315 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c brcmu_pktq_mlen(&bus->txq, ~bus->flowcontrol)); txq 2319 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c pkt = brcmu_pktq_mdeq(&bus->txq, tx_prec_map, txq 2351 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c bus->txoff && (pktq_len(&bus->txq) < TXLOW)) { txq 2486 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c brcmu_pktq_flush(&bus->txq, true, NULL, NULL); txq 2677 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c brcmu_pktq_mlen(&bus->txq, ~bus->flowcontrol) && txlimit && txq 2700 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c brcmu_pktq_mlen(&bus->txq, ~bus->flowcontrol) && txq 2712 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c return &bus->txq; txq 2777 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c brcmf_dbg(TRACE, "deferring pktq len %d\n", pktq_len(&bus->txq)); txq 2784 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c if (!brcmf_sdio_prec_enq(&bus->txq, pkt, prec)) { txq 2792 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c if (pktq_len(&bus->txq) >= TXHI) { txq 2799 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c if (pktq_plen(&bus->txq, prec) > qcount[prec]) txq 2800 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c qcount[prec] = pktq_plen(&bus->txq, prec); txq 4014 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c brcmu_pktq_init(&bus->txq, (PRIOMASK + 1), TXQLEN); txq 1229 drivers/net/wireless/cisco/airo.c struct sk_buff_head txq;// tx queue used by mpi350 code txq 1932 drivers/net/wireless/cisco/airo.c npacks = skb_queue_len (&ai->txq); txq 1940 drivers/net/wireless/cisco/airo.c skb_queue_tail (&ai->txq, skb); txq 1945 drivers/net/wireless/cisco/airo.c skb_queue_tail (&ai->txq, skb); txq 1975 drivers/net/wireless/cisco/airo.c if ((skb = skb_dequeue(&ai->txq)) == NULL) { txq 2402 drivers/net/wireless/cisco/airo.c if (test_bit(FLAG_MPI, &ai->flags) && !skb_queue_empty(&ai->txq)) { txq 2404 drivers/net/wireless/cisco/airo.c for (;(skb = skb_dequeue(&ai->txq));) txq 2812 drivers/net/wireless/cisco/airo.c skb_queue_head_init (&ai->txq); txq 3441 drivers/net/wireless/cisco/airo.c if (!skb_queue_empty(&ai->txq)) { txq 2808 drivers/net/wireless/intel/ipw2x00/ipw2100.c struct ipw2100_bd_queue *txq = &priv->tx_queue; txq 2822 drivers/net/wireless/intel/ipw2x00/ipw2100.c tbd = &txq->drv[packet->index]; txq 2829 drivers/net/wireless/intel/ipw2x00/ipw2100.c e = txq->oldest; txq 2836 drivers/net/wireless/intel/ipw2x00/ipw2100.c e = txq->oldest + frag_num; txq 2837 drivers/net/wireless/intel/ipw2x00/ipw2100.c e %= txq->entries; txq 2854 drivers/net/wireless/intel/ipw2x00/ipw2100.c if (w != txq->next) txq 2890 drivers/net/wireless/intel/ipw2x00/ipw2100.c i = txq->oldest; txq 2892 drivers/net/wireless/intel/ipw2x00/ipw2100.c &txq->drv[i], txq 2893 drivers/net/wireless/intel/ipw2x00/ipw2100.c (u32) (txq->nic + i * sizeof(struct ipw2100_bd)), txq 2894 drivers/net/wireless/intel/ipw2x00/ipw2100.c txq->drv[i].host_addr, txq->drv[i].buf_length); txq 2897 drivers/net/wireless/intel/ipw2x00/ipw2100.c i = (i + 1) % txq->entries; txq 2900 drivers/net/wireless/intel/ipw2x00/ipw2100.c &txq->drv[i], txq 2901 drivers/net/wireless/intel/ipw2x00/ipw2100.c (u32) (txq->nic + i * txq 2903 drivers/net/wireless/intel/ipw2x00/ipw2100.c (u32) txq->drv[i].host_addr, txq 2904 drivers/net/wireless/intel/ipw2x00/ipw2100.c txq->drv[i].buf_length); txq 2911 drivers/net/wireless/intel/ipw2x00/ipw2100.c if (txq->drv[txq->oldest].status.info.fields.txType != 0) txq 2915 drivers/net/wireless/intel/ipw2x00/ipw2100.c priv->net_dev->name, txq->oldest, packet->index); txq 2919 drivers/net/wireless/intel/ipw2x00/ipw2100.c tbd = &txq->drv[(packet->index + 1 + i) % txq->entries]; txq 2922 drivers/net/wireless/intel/ipw2x00/ipw2100.c (packet->index + 1 + i) % txq->entries, txq 2948 drivers/net/wireless/intel/ipw2x00/ipw2100.c if (txq->drv[txq->oldest].status.info.fields.txType != 1) txq 2952 drivers/net/wireless/intel/ipw2x00/ipw2100.c priv->net_dev->name, txq->oldest, packet->index); txq 2971 drivers/net/wireless/intel/ipw2x00/ipw2100.c txq->oldest = (e + 1) % txq->entries; txq 2973 drivers/net/wireless/intel/ipw2x00/ipw2100.c txq->available += descriptors_used; txq 2974 drivers/net/wireless/intel/ipw2x00/ipw2100.c SET_STAT(&priv->txq_stat, txq->available); txq 3000 drivers/net/wireless/intel/ipw2x00/ipw2100.c struct ipw2100_bd_queue *txq = &priv->tx_queue; txq 3002 drivers/net/wireless/intel/ipw2x00/ipw2100.c int next = txq->next; txq 3011 drivers/net/wireless/intel/ipw2x00/ipw2100.c if (txq->available <= 3) { txq 3023 drivers/net/wireless/intel/ipw2x00/ipw2100.c &txq->drv[txq->next], txq 3024 drivers/net/wireless/intel/ipw2x00/ipw2100.c (u32) (txq->nic + txq->next * txq 3027 drivers/net/wireless/intel/ipw2x00/ipw2100.c packet->index = txq->next; txq 3029 drivers/net/wireless/intel/ipw2x00/ipw2100.c tbd = &txq->drv[txq->next]; txq 3042 drivers/net/wireless/intel/ipw2x00/ipw2100.c txq->next++; txq 3043 drivers/net/wireless/intel/ipw2x00/ipw2100.c txq->next %= txq->entries; txq 3044 drivers/net/wireless/intel/ipw2x00/ipw2100.c txq->available--; txq 3051 drivers/net/wireless/intel/ipw2x00/ipw2100.c if (txq->next != next) { txq 3057 drivers/net/wireless/intel/ipw2x00/ipw2100.c txq->next); txq 3069 drivers/net/wireless/intel/ipw2x00/ipw2100.c struct ipw2100_bd_queue *txq = &priv->tx_queue; txq 3071 drivers/net/wireless/intel/ipw2x00/ipw2100.c int next = txq->next; txq 3095 drivers/net/wireless/intel/ipw2x00/ipw2100.c if (txq->available <= 3 + packet->info.d_struct.txb->nr_frags) { txq 3103 drivers/net/wireless/intel/ipw2x00/ipw2100.c tbd = &txq->drv[txq->next]; txq 3105 drivers/net/wireless/intel/ipw2x00/ipw2100.c packet->index = txq->next; txq 3142 drivers/net/wireless/intel/ipw2x00/ipw2100.c txq->next++; txq 3143 drivers/net/wireless/intel/ipw2x00/ipw2100.c txq->next %= txq->entries; txq 3154 drivers/net/wireless/intel/ipw2x00/ipw2100.c tbd = &txq->drv[txq->next]; txq 3181 drivers/net/wireless/intel/ipw2x00/ipw2100.c txq->next, tbd->host_addr, txq 3189 drivers/net/wireless/intel/ipw2x00/ipw2100.c txq->next++; txq 3190 drivers/net/wireless/intel/ipw2x00/ipw2100.c txq->next %= txq->entries; txq 3193 drivers/net/wireless/intel/ipw2x00/ipw2100.c txq->available -= 1 + packet->info.d_struct.txb->nr_frags; txq 3194 drivers/net/wireless/intel/ipw2x00/ipw2100.c SET_STAT(&priv->txq_stat, txq->available); txq 3200 drivers/net/wireless/intel/ipw2x00/ipw2100.c if (txq->next != next) { txq 3205 drivers/net/wireless/intel/ipw2x00/ipw2100.c txq->next); txq 199 drivers/net/wireless/intel/ipw2x00/ipw2200.c struct clx2_tx_queue *txq, int qindex); txq 1991 drivers/net/wireless/intel/ipw2x00/ipw2200.c rc = ipw_queue_tx_reclaim(priv, &priv->txq[0], 0); txq 1997 drivers/net/wireless/intel/ipw2x00/ipw2200.c rc = ipw_queue_tx_reclaim(priv, &priv->txq[1], 1); txq 2003 drivers/net/wireless/intel/ipw2x00/ipw2200.c rc = ipw_queue_tx_reclaim(priv, &priv->txq[2], 2); txq 2009 drivers/net/wireless/intel/ipw2x00/ipw2200.c rc = ipw_queue_tx_reclaim(priv, &priv->txq[3], 3); txq 3800 drivers/net/wireless/intel/ipw2x00/ipw2200.c struct clx2_tx_queue *txq) txq 3802 drivers/net/wireless/intel/ipw2x00/ipw2200.c struct tfd_frame *bd = &txq->bd[txq->q.last_used]; txq 3824 drivers/net/wireless/intel/ipw2x00/ipw2200.c if (txq->txb[txq->q.last_used]) { txq 3825 drivers/net/wireless/intel/ipw2x00/ipw2200.c libipw_txb_free(txq->txb[txq->q.last_used]); txq 3826 drivers/net/wireless/intel/ipw2x00/ipw2200.c txq->txb[txq->q.last_used] = NULL; txq 3840 drivers/net/wireless/intel/ipw2x00/ipw2200.c static void ipw_queue_tx_free(struct ipw_priv *priv, struct clx2_tx_queue *txq) txq 3842 drivers/net/wireless/intel/ipw2x00/ipw2200.c struct clx2_queue *q = &txq->q; txq 3851 drivers/net/wireless/intel/ipw2x00/ipw2200.c ipw_queue_tx_free_tfd(priv, txq); txq 3855 drivers/net/wireless/intel/ipw2x00/ipw2200.c pci_free_consistent(dev, sizeof(txq->bd[0]) * q->n_bd, txq->bd, txq 3857 drivers/net/wireless/intel/ipw2x00/ipw2200.c kfree(txq->txb); txq 3860 drivers/net/wireless/intel/ipw2x00/ipw2200.c memset(txq, 0, sizeof(*txq)); txq 3874 drivers/net/wireless/intel/ipw2x00/ipw2200.c ipw_queue_tx_free(priv, &priv->txq[0]); txq 3875 drivers/net/wireless/intel/ipw2x00/ipw2200.c ipw_queue_tx_free(priv, &priv->txq[1]); txq 3876 drivers/net/wireless/intel/ipw2x00/ipw2200.c ipw_queue_tx_free(priv, &priv->txq[2]); txq 3877 drivers/net/wireless/intel/ipw2x00/ipw2200.c ipw_queue_tx_free(priv, &priv->txq[3]); txq 4951 drivers/net/wireless/intel/ipw2x00/ipw2200.c rc = ipw_queue_tx_init(priv, &priv->txq[0], nTx, txq 4959 drivers/net/wireless/intel/ipw2x00/ipw2200.c rc = ipw_queue_tx_init(priv, &priv->txq[1], nTx, txq 4967 drivers/net/wireless/intel/ipw2x00/ipw2200.c rc = ipw_queue_tx_init(priv, &priv->txq[2], nTx, txq 4975 drivers/net/wireless/intel/ipw2x00/ipw2200.c rc = ipw_queue_tx_init(priv, &priv->txq[3], nTx, txq 5007 drivers/net/wireless/intel/ipw2x00/ipw2200.c struct clx2_tx_queue *txq, int qindex) txq 5011 drivers/net/wireless/intel/ipw2x00/ipw2200.c struct clx2_queue *q = &txq->q; txq 5022 drivers/net/wireless/intel/ipw2x00/ipw2200.c ipw_queue_tx_free_tfd(priv, txq); txq 5039 drivers/net/wireless/intel/ipw2x00/ipw2200.c struct clx2_tx_queue *txq = &priv->txq_cmd; txq 5040 drivers/net/wireless/intel/ipw2x00/ipw2200.c struct clx2_queue *q = &txq->q; txq 5048 drivers/net/wireless/intel/ipw2x00/ipw2200.c tfd = &txq->bd[q->first_empty]; txq 5049 drivers/net/wireless/intel/ipw2x00/ipw2200.c txq->txb[q->first_empty] = NULL; txq 10101 drivers/net/wireless/intel/ipw2x00/ipw2200.c struct clx2_tx_queue *txq = &priv->txq[tx_id]; txq 10103 drivers/net/wireless/intel/ipw2x00/ipw2200.c struct clx2_tx_queue *txq = &priv->txq[0]; txq 10105 drivers/net/wireless/intel/ipw2x00/ipw2200.c struct clx2_queue *q = &txq->q; txq 10135 drivers/net/wireless/intel/ipw2x00/ipw2200.c tfd = &txq->bd[q->first_empty]; txq 10136 drivers/net/wireless/intel/ipw2x00/ipw2200.c txq->txb[q->first_empty] = txb; txq 10291 drivers/net/wireless/intel/ipw2x00/ipw2200.c struct clx2_tx_queue *txq = &priv->txq[tx_id]; txq 10293 drivers/net/wireless/intel/ipw2x00/ipw2200.c struct clx2_tx_queue *txq = &priv->txq[0]; txq 10296 drivers/net/wireless/intel/ipw2x00/ipw2200.c if (ipw_tx_queue_space(&txq->q) < txq->q.high_mark) txq 1193 drivers/net/wireless/intel/ipw2x00/ipw2200.h struct clx2_tx_queue txq[4]; txq 454 drivers/net/wireless/intel/iwlegacy/3945-mac.c struct il_tx_queue *txq = NULL; txq 513 drivers/net/wireless/intel/iwlegacy/3945-mac.c txq = &il->txq[txq_id]; txq 514 drivers/net/wireless/intel/iwlegacy/3945-mac.c q = &txq->q; txq 523 drivers/net/wireless/intel/iwlegacy/3945-mac.c txq->skbs[q->write_ptr] = skb; txq 526 drivers/net/wireless/intel/iwlegacy/3945-mac.c out_cmd = txq->cmd[idx]; txq 527 drivers/net/wireless/intel/iwlegacy/3945-mac.c out_meta = &txq->meta[idx]; txq 595 drivers/net/wireless/intel/iwlegacy/3945-mac.c il->ops->txq_attach_buf_to_tfd(il, txq, txcmd_phys, firstlen, 1, 0); txq 599 drivers/net/wireless/intel/iwlegacy/3945-mac.c il->ops->txq_attach_buf_to_tfd(il, txq, phys_addr, secondlen, 0, txq 603 drivers/net/wireless/intel/iwlegacy/3945-mac.c txq->need_update = 1; txq 606 drivers/net/wireless/intel/iwlegacy/3945-mac.c txq->need_update = 0; txq 619 drivers/net/wireless/intel/iwlegacy/3945-mac.c il_txq_update_write_ptr(il, txq); txq 625 drivers/net/wireless/intel/iwlegacy/3945-mac.c txq->need_update = 1; txq 626 drivers/net/wireless/intel/iwlegacy/3945-mac.c il_txq_update_write_ptr(il, txq); txq 630 drivers/net/wireless/intel/iwlegacy/3945-mac.c il_stop_queue(il, txq); txq 1471 drivers/net/wireless/intel/iwlegacy/3945-mac.c il_txq_update_write_ptr(il, &il->txq[0]); txq 1472 drivers/net/wireless/intel/iwlegacy/3945-mac.c il_txq_update_write_ptr(il, &il->txq[1]); txq 1473 drivers/net/wireless/intel/iwlegacy/3945-mac.c il_txq_update_write_ptr(il, &il->txq[2]); txq 1474 drivers/net/wireless/intel/iwlegacy/3945-mac.c il_txq_update_write_ptr(il, &il->txq[3]); txq 1475 drivers/net/wireless/intel/iwlegacy/3945-mac.c il_txq_update_write_ptr(il, &il->txq[4]); txq 2463 drivers/net/wireless/intel/iwlegacy/3945-mac.c if (test_bit(S_EXIT_PENDING, &il->status) || il->txq == NULL) txq 274 drivers/net/wireless/intel/iwlegacy/3945.c struct il_tx_queue *txq = &il->txq[txq_id]; txq 275 drivers/net/wireless/intel/iwlegacy/3945.c struct il_queue *q = &txq->q; txq 283 drivers/net/wireless/intel/iwlegacy/3945.c skb = txq->skbs[txq->q.read_ptr]; txq 285 drivers/net/wireless/intel/iwlegacy/3945.c txq->skbs[txq->q.read_ptr] = NULL; txq 286 drivers/net/wireless/intel/iwlegacy/3945.c il->ops->txq_free_tfd(il, txq); txq 291 drivers/net/wireless/intel/iwlegacy/3945.c il_wake_queue(il, txq); txq 304 drivers/net/wireless/intel/iwlegacy/3945.c struct il_tx_queue *txq = &il->txq[txq_id]; txq 311 drivers/net/wireless/intel/iwlegacy/3945.c if (idx >= txq->q.n_bd || il_queue_used(&txq->q, idx) == 0) { txq 314 drivers/net/wireless/intel/iwlegacy/3945.c txq->q.n_bd, txq->q.write_ptr, txq->q.read_ptr); txq 331 drivers/net/wireless/intel/iwlegacy/3945.c txq->time_stamp = jiffies; txq 332 drivers/net/wireless/intel/iwlegacy/3945.c info = IEEE80211_SKB_CB(txq->skbs[txq->q.read_ptr]); txq 597 drivers/net/wireless/intel/iwlegacy/3945.c il3945_hw_txq_attach_buf_to_tfd(struct il_priv *il, struct il_tx_queue *txq, txq 604 drivers/net/wireless/intel/iwlegacy/3945.c q = &txq->q; txq 605 drivers/net/wireless/intel/iwlegacy/3945.c tfd_tmp = (struct il3945_tfd *)txq->tfds; txq 636 drivers/net/wireless/intel/iwlegacy/3945.c il3945_hw_txq_free_tfd(struct il_priv *il, struct il_tx_queue *txq) txq 638 drivers/net/wireless/intel/iwlegacy/3945.c struct il3945_tfd *tfd_tmp = (struct il3945_tfd *)txq->tfds; txq 639 drivers/net/wireless/intel/iwlegacy/3945.c int idx = txq->q.read_ptr; txq 655 drivers/net/wireless/intel/iwlegacy/3945.c pci_unmap_single(dev, dma_unmap_addr(&txq->meta[idx], mapping), txq 656 drivers/net/wireless/intel/iwlegacy/3945.c dma_unmap_len(&txq->meta[idx], len), txq 667 drivers/net/wireless/intel/iwlegacy/3945.c if (txq->skbs) { txq 668 drivers/net/wireless/intel/iwlegacy/3945.c struct sk_buff *skb = txq->skbs[txq->q.read_ptr]; txq 673 drivers/net/wireless/intel/iwlegacy/3945.c txq->skbs[txq->q.read_ptr] = NULL; txq 1007 drivers/net/wireless/intel/iwlegacy/3945.c if (il->txq) { txq 1868 drivers/net/wireless/intel/iwlegacy/3945.c if (test_bit(S_EXIT_PENDING, &il->status) || il->txq == NULL) txq 2197 drivers/net/wireless/intel/iwlegacy/3945.c il3945_hw_tx_queue_init(struct il_priv *il, struct il_tx_queue *txq) txq 2199 drivers/net/wireless/intel/iwlegacy/3945.c int txq_id = txq->q.id; txq 2203 drivers/net/wireless/intel/iwlegacy/3945.c shared_data->tx_base_ptr[txq_id] = cpu_to_le32((u32) txq->q.dma_addr); txq 209 drivers/net/wireless/intel/iwlegacy/3945.h int il3945_hw_txq_attach_buf_to_tfd(struct il_priv *il, struct il_tx_queue *txq, txq 211 drivers/net/wireless/intel/iwlegacy/3945.h void il3945_hw_txq_free_tfd(struct il_priv *il, struct il_tx_queue *txq); txq 213 drivers/net/wireless/intel/iwlegacy/3945.h int il3945_hw_tx_queue_init(struct il_priv *il, struct il_tx_queue *txq); txq 216 drivers/net/wireless/intel/iwlegacy/4965-mac.c if (!il->txq) { txq 1648 drivers/net/wireless/intel/iwlegacy/4965-mac.c struct il_tx_queue *txq; txq 1749 drivers/net/wireless/intel/iwlegacy/4965-mac.c txq = &il->txq[txq_id]; txq 1750 drivers/net/wireless/intel/iwlegacy/4965-mac.c q = &txq->q; txq 1765 drivers/net/wireless/intel/iwlegacy/4965-mac.c txq->skbs[q->write_ptr] = skb; txq 1768 drivers/net/wireless/intel/iwlegacy/4965-mac.c out_cmd = txq->cmd[q->write_ptr]; txq 1769 drivers/net/wireless/intel/iwlegacy/4965-mac.c out_meta = &txq->meta[q->write_ptr]; txq 1836 drivers/net/wireless/intel/iwlegacy/4965-mac.c il->ops->txq_attach_buf_to_tfd(il, txq, txcmd_phys, firstlen, 1, 0); txq 1840 drivers/net/wireless/intel/iwlegacy/4965-mac.c il->ops->txq_attach_buf_to_tfd(il, txq, phys_addr, secondlen, txq 1844 drivers/net/wireless/intel/iwlegacy/4965-mac.c txq->need_update = 1; txq 1847 drivers/net/wireless/intel/iwlegacy/4965-mac.c txq->need_update = 0; txq 1869 drivers/net/wireless/intel/iwlegacy/4965-mac.c il->ops->txq_update_byte_cnt_tbl(il, txq, le16_to_cpu(tx_cmd->len)); txq 1876 drivers/net/wireless/intel/iwlegacy/4965-mac.c il_txq_update_write_ptr(il, txq); txq 1899 drivers/net/wireless/intel/iwlegacy/4965-mac.c txq->need_update = 1; txq 1900 drivers/net/wireless/intel/iwlegacy/4965-mac.c il_txq_update_write_ptr(il, txq); txq 1903 drivers/net/wireless/intel/iwlegacy/4965-mac.c il_stop_queue(il, txq); txq 1946 drivers/net/wireless/intel/iwlegacy/4965-mac.c if (il->txq) { txq 2051 drivers/net/wireless/intel/iwlegacy/4965-mac.c if (!il->txq) txq 2187 drivers/net/wireless/intel/iwlegacy/4965-mac.c il->txq[txq_id].q.read_ptr = (ssn_idx & 0xff); txq 2188 drivers/net/wireless/intel/iwlegacy/4965-mac.c il->txq[txq_id].q.write_ptr = (ssn_idx & 0xff); txq 2208 drivers/net/wireless/intel/iwlegacy/4965-mac.c il4965_tx_queue_set_status(il, &il->txq[txq_id], tx_fifo, 1); txq 2256 drivers/net/wireless/intel/iwlegacy/4965-mac.c il_set_swq_id(&il->txq[txq_id], il4965_get_ac_from_tid(tid), txq_id); txq 2299 drivers/net/wireless/intel/iwlegacy/4965-mac.c il->txq[txq_id].q.read_ptr = (ssn_idx & 0xff); txq 2300 drivers/net/wireless/intel/iwlegacy/4965-mac.c il->txq[txq_id].q.write_ptr = (ssn_idx & 0xff); txq 2306 drivers/net/wireless/intel/iwlegacy/4965-mac.c il4965_tx_queue_set_status(il, &il->txq[txq_id], tx_fifo, 0); txq 2354 drivers/net/wireless/intel/iwlegacy/4965-mac.c write_ptr = il->txq[txq_id].q.write_ptr; txq 2355 drivers/net/wireless/intel/iwlegacy/4965-mac.c read_ptr = il->txq[txq_id].q.read_ptr; txq 2392 drivers/net/wireless/intel/iwlegacy/4965-mac.c struct il_queue *q = &il->txq[txq_id].q; txq 2457 drivers/net/wireless/intel/iwlegacy/4965-mac.c struct il_tx_queue *txq = &il->txq[txq_id]; txq 2458 drivers/net/wireless/intel/iwlegacy/4965-mac.c struct il_queue *q = &txq->q; txq 2473 drivers/net/wireless/intel/iwlegacy/4965-mac.c skb = txq->skbs[txq->q.read_ptr]; txq 2484 drivers/net/wireless/intel/iwlegacy/4965-mac.c txq->skbs[txq->q.read_ptr] = NULL; txq 2485 drivers/net/wireless/intel/iwlegacy/4965-mac.c il->ops->txq_free_tfd(il, txq); txq 2548 drivers/net/wireless/intel/iwlegacy/4965-mac.c info = IEEE80211_SKB_CB(il->txq[scd_flow].skbs[agg->start_idx]); txq 2675 drivers/net/wireless/intel/iwlegacy/4965-mac.c info = IEEE80211_SKB_CB(il->txq[txq_id].skbs[idx]); txq 2708 drivers/net/wireless/intel/iwlegacy/4965-mac.c skb = il->txq[txq_id].skbs[idx]; txq 2765 drivers/net/wireless/intel/iwlegacy/4965-mac.c struct il_tx_queue *txq = &il->txq[txq_id]; txq 2777 drivers/net/wireless/intel/iwlegacy/4965-mac.c if (idx >= txq->q.n_bd || il_queue_used(&txq->q, idx) == 0) { txq 2780 drivers/net/wireless/intel/iwlegacy/4965-mac.c txq->q.n_bd, txq->q.write_ptr, txq->q.read_ptr); txq 2784 drivers/net/wireless/intel/iwlegacy/4965-mac.c txq->time_stamp = jiffies; txq 2786 drivers/net/wireless/intel/iwlegacy/4965-mac.c skb = txq->skbs[txq->q.read_ptr]; txq 2797 drivers/net/wireless/intel/iwlegacy/4965-mac.c if (txq->sched_retry && unlikely(sta_id == IL_INVALID_STATION)) { txq 2816 drivers/net/wireless/intel/iwlegacy/4965-mac.c if (txq->sched_retry) { txq 2830 drivers/net/wireless/intel/iwlegacy/4965-mac.c if (txq->q.read_ptr != (scd_ssn & 0xff)) { txq 2831 drivers/net/wireless/intel/iwlegacy/4965-mac.c idx = il_queue_dec_wrap(scd_ssn & 0xff, txq->q.n_bd); txq 2840 drivers/net/wireless/intel/iwlegacy/4965-mac.c il_queue_space(&txq->q) > txq->q.low_mark && txq 2842 drivers/net/wireless/intel/iwlegacy/4965-mac.c il_wake_queue(il, txq); txq 2864 drivers/net/wireless/intel/iwlegacy/4965-mac.c il_queue_space(&txq->q) > txq->q.low_mark) txq 2865 drivers/net/wireless/intel/iwlegacy/4965-mac.c il_wake_queue(il, txq); txq 2910 drivers/net/wireless/intel/iwlegacy/4965-mac.c struct il_tx_queue *txq = NULL; txq 2929 drivers/net/wireless/intel/iwlegacy/4965-mac.c txq = &il->txq[scd_flow]; txq 2946 drivers/net/wireless/intel/iwlegacy/4965-mac.c idx = il_queue_dec_wrap(ba_resp_scd_ssn & 0xff, txq->q.n_bd); txq 2966 drivers/net/wireless/intel/iwlegacy/4965-mac.c if (txq->q.read_ptr != (ba_resp_scd_ssn & 0xff)) { txq 2971 drivers/net/wireless/intel/iwlegacy/4965-mac.c if (il_queue_space(&txq->q) > txq->q.low_mark && txq 2974 drivers/net/wireless/intel/iwlegacy/4965-mac.c il_wake_queue(il, txq); txq 3915 drivers/net/wireless/intel/iwlegacy/4965-mac.c il4965_hw_txq_free_tfd(struct il_priv *il, struct il_tx_queue *txq) txq 3917 drivers/net/wireless/intel/iwlegacy/4965-mac.c struct il_tfd *tfd_tmp = (struct il_tfd *)txq->tfds; txq 3920 drivers/net/wireless/intel/iwlegacy/4965-mac.c int idx = txq->q.read_ptr; txq 3937 drivers/net/wireless/intel/iwlegacy/4965-mac.c pci_unmap_single(dev, dma_unmap_addr(&txq->meta[idx], mapping), txq 3938 drivers/net/wireless/intel/iwlegacy/4965-mac.c dma_unmap_len(&txq->meta[idx], len), txq 3948 drivers/net/wireless/intel/iwlegacy/4965-mac.c if (txq->skbs) { txq 3949 drivers/net/wireless/intel/iwlegacy/4965-mac.c struct sk_buff *skb = txq->skbs[txq->q.read_ptr]; txq 3954 drivers/net/wireless/intel/iwlegacy/4965-mac.c txq->skbs[txq->q.read_ptr] = NULL; txq 3960 drivers/net/wireless/intel/iwlegacy/4965-mac.c il4965_hw_txq_attach_buf_to_tfd(struct il_priv *il, struct il_tx_queue *txq, txq 3967 drivers/net/wireless/intel/iwlegacy/4965-mac.c q = &txq->q; txq 3968 drivers/net/wireless/intel/iwlegacy/4965-mac.c tfd_tmp = (struct il_tfd *)txq->tfds; txq 4000 drivers/net/wireless/intel/iwlegacy/4965-mac.c il4965_hw_tx_queue_init(struct il_priv *il, struct il_tx_queue *txq) txq 4002 drivers/net/wireless/intel/iwlegacy/4965-mac.c int txq_id = txq->q.id; txq 4005 drivers/net/wireless/intel/iwlegacy/4965-mac.c il_wr(il, FH49_MEM_CBBC_QUEUE(txq_id), txq->q.dma_addr >> 8); txq 4478 drivers/net/wireless/intel/iwlegacy/4965-mac.c il_txq_update_write_ptr(il, &il->txq[i]); txq 5254 drivers/net/wireless/intel/iwlegacy/4965-mac.c il4965_tx_queue_set_status(il, &il->txq[i], ac, 0); txq 6291 drivers/net/wireless/intel/iwlegacy/4965-mac.c il4965_tx_queue_set_status(struct il_priv *il, struct il_tx_queue *txq, txq 6294 drivers/net/wireless/intel/iwlegacy/4965-mac.c int txq_id = txq->q.id; txq 6307 drivers/net/wireless/intel/iwlegacy/4965-mac.c txq->sched_retry = scd_retry; txq 1535 drivers/net/wireless/intel/iwlegacy/4965.c il4965_txq_update_byte_cnt_tbl(struct il_priv *il, struct il_tx_queue *txq, txq 1539 drivers/net/wireless/intel/iwlegacy/4965.c int txq_id = txq->q.id; txq 1540 drivers/net/wireless/intel/iwlegacy/4965.c int write_ptr = txq->q.write_ptr; txq 57 drivers/net/wireless/intel/iwlegacy/4965.h void il4965_hw_txq_free_tfd(struct il_priv *il, struct il_tx_queue *txq); txq 58 drivers/net/wireless/intel/iwlegacy/4965.h int il4965_hw_txq_attach_buf_to_tfd(struct il_priv *il, struct il_tx_queue *txq, txq 60 drivers/net/wireless/intel/iwlegacy/4965.h int il4965_hw_tx_queue_init(struct il_priv *il, struct il_tx_queue *txq); txq 89 drivers/net/wireless/intel/iwlegacy/4965.h void il4965_tx_queue_set_status(struct il_priv *il, struct il_tx_queue *txq, txq 364 drivers/net/wireless/intel/iwlegacy/common.c il->txq[il->cmd_queue].meta[cmd_idx].flags &= ~CMD_WANT_SKB; txq 2710 drivers/net/wireless/intel/iwlegacy/common.c il_txq_update_write_ptr(struct il_priv *il, struct il_tx_queue *txq) txq 2713 drivers/net/wireless/intel/iwlegacy/common.c int txq_id = txq->q.id; txq 2715 drivers/net/wireless/intel/iwlegacy/common.c if (txq->need_update == 0) txq 2733 drivers/net/wireless/intel/iwlegacy/common.c il_wr(il, HBUS_TARG_WRPTR, txq->q.write_ptr | (txq_id << 8)); txq 2741 drivers/net/wireless/intel/iwlegacy/common.c _il_wr(il, HBUS_TARG_WRPTR, txq->q.write_ptr | (txq_id << 8)); txq 2742 drivers/net/wireless/intel/iwlegacy/common.c txq->need_update = 0; txq 2752 drivers/net/wireless/intel/iwlegacy/common.c struct il_tx_queue *txq = &il->txq[txq_id]; txq 2753 drivers/net/wireless/intel/iwlegacy/common.c struct il_queue *q = &txq->q; txq 2759 drivers/net/wireless/intel/iwlegacy/common.c il->ops->txq_free_tfd(il, txq); txq 2776 drivers/net/wireless/intel/iwlegacy/common.c struct il_tx_queue *txq = &il->txq[txq_id]; txq 2783 drivers/net/wireless/intel/iwlegacy/common.c if (txq->cmd) { txq 2785 drivers/net/wireless/intel/iwlegacy/common.c kfree(txq->cmd[i]); txq 2789 drivers/net/wireless/intel/iwlegacy/common.c if (txq->q.n_bd) txq 2790 drivers/net/wireless/intel/iwlegacy/common.c dma_free_coherent(dev, il->hw_params.tfd_size * txq->q.n_bd, txq 2791 drivers/net/wireless/intel/iwlegacy/common.c txq->tfds, txq->q.dma_addr); txq 2794 drivers/net/wireless/intel/iwlegacy/common.c kfree(txq->skbs); txq 2795 drivers/net/wireless/intel/iwlegacy/common.c txq->skbs = NULL; txq 2798 drivers/net/wireless/intel/iwlegacy/common.c kfree(txq->cmd); txq 2799 drivers/net/wireless/intel/iwlegacy/common.c kfree(txq->meta); txq 2800 drivers/net/wireless/intel/iwlegacy/common.c txq->cmd = NULL; txq 2801 drivers/net/wireless/intel/iwlegacy/common.c txq->meta = NULL; txq 2804 drivers/net/wireless/intel/iwlegacy/common.c memset(txq, 0, sizeof(*txq)); txq 2814 drivers/net/wireless/intel/iwlegacy/common.c struct il_tx_queue *txq = &il->txq[il->cmd_queue]; txq 2815 drivers/net/wireless/intel/iwlegacy/common.c struct il_queue *q = &txq->q; txq 2824 drivers/net/wireless/intel/iwlegacy/common.c if (txq->meta[i].flags & CMD_MAPPED) { txq 2826 drivers/net/wireless/intel/iwlegacy/common.c dma_unmap_addr(&txq->meta[i], mapping), txq 2827 drivers/net/wireless/intel/iwlegacy/common.c dma_unmap_len(&txq->meta[i], len), txq 2829 drivers/net/wireless/intel/iwlegacy/common.c txq->meta[i].flags = 0; txq 2836 drivers/net/wireless/intel/iwlegacy/common.c if (txq->meta[i].flags & CMD_MAPPED) { txq 2838 drivers/net/wireless/intel/iwlegacy/common.c dma_unmap_addr(&txq->meta[i], mapping), txq 2839 drivers/net/wireless/intel/iwlegacy/common.c dma_unmap_len(&txq->meta[i], len), txq 2841 drivers/net/wireless/intel/iwlegacy/common.c txq->meta[i].flags = 0; txq 2857 drivers/net/wireless/intel/iwlegacy/common.c struct il_tx_queue *txq = &il->txq[il->cmd_queue]; txq 2864 drivers/net/wireless/intel/iwlegacy/common.c if (txq->cmd) { txq 2866 drivers/net/wireless/intel/iwlegacy/common.c kfree(txq->cmd[i]); txq 2870 drivers/net/wireless/intel/iwlegacy/common.c if (txq->q.n_bd) txq 2871 drivers/net/wireless/intel/iwlegacy/common.c dma_free_coherent(dev, il->hw_params.tfd_size * txq->q.n_bd, txq 2872 drivers/net/wireless/intel/iwlegacy/common.c txq->tfds, txq->q.dma_addr); txq 2875 drivers/net/wireless/intel/iwlegacy/common.c kfree(txq->cmd); txq 2876 drivers/net/wireless/intel/iwlegacy/common.c kfree(txq->meta); txq 2877 drivers/net/wireless/intel/iwlegacy/common.c txq->cmd = NULL; txq 2878 drivers/net/wireless/intel/iwlegacy/common.c txq->meta = NULL; txq 2881 drivers/net/wireless/intel/iwlegacy/common.c memset(txq, 0, sizeof(*txq)); txq 2965 drivers/net/wireless/intel/iwlegacy/common.c il_tx_queue_alloc(struct il_priv *il, struct il_tx_queue *txq, u32 id) txq 2973 drivers/net/wireless/intel/iwlegacy/common.c txq->skbs = kcalloc(TFD_QUEUE_SIZE_MAX, txq 2976 drivers/net/wireless/intel/iwlegacy/common.c if (!txq->skbs) { txq 2981 drivers/net/wireless/intel/iwlegacy/common.c txq->skbs = NULL; txq 2985 drivers/net/wireless/intel/iwlegacy/common.c txq->tfds = txq 2986 drivers/net/wireless/intel/iwlegacy/common.c dma_alloc_coherent(dev, tfd_sz, &txq->q.dma_addr, GFP_KERNEL); txq 2987 drivers/net/wireless/intel/iwlegacy/common.c if (!txq->tfds) txq 2990 drivers/net/wireless/intel/iwlegacy/common.c txq->q.id = id; txq 2995 drivers/net/wireless/intel/iwlegacy/common.c kfree(txq->skbs); txq 2996 drivers/net/wireless/intel/iwlegacy/common.c txq->skbs = NULL; txq 3009 drivers/net/wireless/intel/iwlegacy/common.c struct il_tx_queue *txq = &il->txq[txq_id]; txq 3027 drivers/net/wireless/intel/iwlegacy/common.c txq->meta = txq 3029 drivers/net/wireless/intel/iwlegacy/common.c txq->cmd = txq 3032 drivers/net/wireless/intel/iwlegacy/common.c if (!txq->meta || !txq->cmd) txq 3041 drivers/net/wireless/intel/iwlegacy/common.c txq->cmd[i] = kmalloc(len, GFP_KERNEL); txq 3042 drivers/net/wireless/intel/iwlegacy/common.c if (!txq->cmd[i]) txq 3047 drivers/net/wireless/intel/iwlegacy/common.c ret = il_tx_queue_alloc(il, txq, txq_id); txq 3051 drivers/net/wireless/intel/iwlegacy/common.c txq->need_update = 0; txq 3059 drivers/net/wireless/intel/iwlegacy/common.c il_set_swq_id(txq, txq_id, txq_id); txq 3062 drivers/net/wireless/intel/iwlegacy/common.c il_queue_init(il, &txq->q, slots, txq_id); txq 3065 drivers/net/wireless/intel/iwlegacy/common.c il->ops->txq_init(il, txq); txq 3070 drivers/net/wireless/intel/iwlegacy/common.c kfree(txq->cmd[i]); txq 3072 drivers/net/wireless/intel/iwlegacy/common.c kfree(txq->meta); txq 3073 drivers/net/wireless/intel/iwlegacy/common.c txq->meta = NULL; txq 3074 drivers/net/wireless/intel/iwlegacy/common.c kfree(txq->cmd); txq 3075 drivers/net/wireless/intel/iwlegacy/common.c txq->cmd = NULL; txq 3085 drivers/net/wireless/intel/iwlegacy/common.c struct il_tx_queue *txq = &il->txq[txq_id]; txq 3095 drivers/net/wireless/intel/iwlegacy/common.c memset(txq->meta, 0, sizeof(struct il_cmd_meta) * actual_slots); txq 3096 drivers/net/wireless/intel/iwlegacy/common.c txq->need_update = 0; txq 3099 drivers/net/wireless/intel/iwlegacy/common.c il_queue_init(il, &txq->q, slots, txq_id); txq 3102 drivers/net/wireless/intel/iwlegacy/common.c il->ops->txq_init(il, txq); txq 3120 drivers/net/wireless/intel/iwlegacy/common.c struct il_tx_queue *txq = &il->txq[il->cmd_queue]; txq 3121 drivers/net/wireless/intel/iwlegacy/common.c struct il_queue *q = &txq->q; txq 3159 drivers/net/wireless/intel/iwlegacy/common.c out_cmd = txq->cmd[idx]; txq 3160 drivers/net/wireless/intel/iwlegacy/common.c out_meta = &txq->meta[idx]; txq 3218 drivers/net/wireless/intel/iwlegacy/common.c txq->need_update = 1; txq 3222 drivers/net/wireless/intel/iwlegacy/common.c il->ops->txq_update_byte_cnt_tbl(il, txq, 0); txq 3224 drivers/net/wireless/intel/iwlegacy/common.c il->ops->txq_attach_buf_to_tfd(il, txq, phys_addr, fix_size, 1, txq 3229 drivers/net/wireless/intel/iwlegacy/common.c il_txq_update_write_ptr(il, txq); txq 3246 drivers/net/wireless/intel/iwlegacy/common.c struct il_tx_queue *txq = &il->txq[txq_id]; txq 3247 drivers/net/wireless/intel/iwlegacy/common.c struct il_queue *q = &txq->q; txq 3288 drivers/net/wireless/intel/iwlegacy/common.c struct il_tx_queue *txq = &il->txq[il->cmd_queue]; txq 3297 drivers/net/wireless/intel/iwlegacy/common.c txq_id, il->cmd_queue, sequence, il->txq[il->cmd_queue].q.read_ptr, txq 3298 drivers/net/wireless/intel/iwlegacy/common.c il->txq[il->cmd_queue].q.write_ptr)) { txq 3303 drivers/net/wireless/intel/iwlegacy/common.c cmd_idx = il_get_cmd_idx(&txq->q, idx, huge); txq 3304 drivers/net/wireless/intel/iwlegacy/common.c cmd = txq->cmd[cmd_idx]; txq 3305 drivers/net/wireless/intel/iwlegacy/common.c meta = &txq->meta[cmd_idx]; txq 3307 drivers/net/wireless/intel/iwlegacy/common.c txq->time_stamp = jiffies; txq 4639 drivers/net/wireless/intel/iwlegacy/common.c if (!il->txq) txq 4640 drivers/net/wireless/intel/iwlegacy/common.c il->txq = txq 4644 drivers/net/wireless/intel/iwlegacy/common.c if (!il->txq) { txq 4655 drivers/net/wireless/intel/iwlegacy/common.c kfree(il->txq); txq 4656 drivers/net/wireless/intel/iwlegacy/common.c il->txq = NULL; txq 4763 drivers/net/wireless/intel/iwlegacy/common.c if (il->txq == NULL) txq 4772 drivers/net/wireless/intel/iwlegacy/common.c q = &il->txq[i].q; txq 4796 drivers/net/wireless/intel/iwlegacy/common.c struct il_tx_queue *txq = &il->txq[cnt]; txq 4797 drivers/net/wireless/intel/iwlegacy/common.c struct il_queue *q = &txq->q; txq 4803 drivers/net/wireless/intel/iwlegacy/common.c txq->time_stamp = now; txq 4808 drivers/net/wireless/intel/iwlegacy/common.c txq->time_stamp + txq 4813 drivers/net/wireless/intel/iwlegacy/common.c jiffies_to_msecs(now - txq->time_stamp)); txq 1257 drivers/net/wireless/intel/iwlegacy/common.h struct il_tx_queue *txq; txq 1534 drivers/net/wireless/intel/iwlegacy/common.h struct il_tx_queue *txq, txq 1537 drivers/net/wireless/intel/iwlegacy/common.h struct il_tx_queue *txq, dma_addr_t addr, txq 1539 drivers/net/wireless/intel/iwlegacy/common.h void (*txq_free_tfd) (struct il_priv *il, struct il_tx_queue *txq); txq 1540 drivers/net/wireless/intel/iwlegacy/common.h int (*txq_init) (struct il_priv *il, struct il_tx_queue *txq); txq 1751 drivers/net/wireless/intel/iwlegacy/common.h void il_txq_update_write_ptr(struct il_priv *il, struct il_tx_queue *txq); txq 2248 drivers/net/wireless/intel/iwlegacy/common.h il_set_swq_id(struct il_tx_queue *txq, u8 ac, u8 hwq) txq 2253 drivers/net/wireless/intel/iwlegacy/common.h txq->swq_id = (hwq << 2) | ac; txq 2270 drivers/net/wireless/intel/iwlegacy/common.h il_wake_queue(struct il_priv *il, struct il_tx_queue *txq) txq 2272 drivers/net/wireless/intel/iwlegacy/common.h u8 queue = txq->swq_id; txq 2281 drivers/net/wireless/intel/iwlegacy/common.h il_stop_queue(struct il_priv *il, struct il_tx_queue *txq) txq 2283 drivers/net/wireless/intel/iwlegacy/common.h u8 queue = txq->swq_id; txq 817 drivers/net/wireless/intel/iwlegacy/debug.c struct il_tx_queue *txq; txq 826 drivers/net/wireless/intel/iwlegacy/debug.c if (!il->txq) { txq 835 drivers/net/wireless/intel/iwlegacy/debug.c txq = &il->txq[cnt]; txq 836 drivers/net/wireless/intel/iwlegacy/debug.c q = &txq->q; txq 843 drivers/net/wireless/intel/iwlegacy/debug.c txq->swq_id, txq->swq_id & 3, txq 844 drivers/net/wireless/intel/iwlegacy/debug.c (txq->swq_id >> 2) & 0x1f); txq 807 drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c void iwl_mvm_mac_itxq_xmit(struct ieee80211_hw *hw, struct ieee80211_txq *txq) txq 810 drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c struct iwl_mvm_txq *mvmtxq = iwl_mvm_txq_from_mac80211(txq); txq 840 drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c skb = ieee80211_tx_dequeue(hw, txq); txq 843 drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c if (txq->sta) txq 846 drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c txq->sta->addr, txq 847 drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c txq->tid); txq 851 drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c iwl_mvm_tx_skb(mvm, skb, txq->sta); txq 858 drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c struct ieee80211_txq *txq) txq 861 drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c struct iwl_mvm_txq *mvmtxq = iwl_mvm_txq_from_mac80211(txq); txq 880 drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c if (!txq->sta || mvmtxq->txq_id != IWL_MVM_INVALID_QUEUE) { txq 886 drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c if (unlikely(txq->sta && !list_empty(&mvmtxq->list))) txq 889 drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c iwl_mvm_mac_itxq_xmit(hw, txq); txq 794 drivers/net/wireless/intel/iwlwifi/mvm/mvm.h iwl_mvm_txq_from_mac80211(struct ieee80211_txq *txq) txq 796 drivers/net/wireless/intel/iwlwifi/mvm/mvm.h return (void *)txq->drv_priv; txq 805 drivers/net/wireless/intel/iwlwifi/mvm/mvm.h return (void *)sta->txq[tid]->drv_priv; txq 1520 drivers/net/wireless/intel/iwlwifi/mvm/mvm.h void iwl_mvm_mac_itxq_xmit(struct ieee80211_hw *hw, struct ieee80211_txq *txq); txq 1096 drivers/net/wireless/intel/iwlwifi/mvm/ops.c struct ieee80211_txq *txq; txq 1140 drivers/net/wireless/intel/iwlwifi/mvm/ops.c txq = sta->txq[tid]; txq 1141 drivers/net/wireless/intel/iwlwifi/mvm/ops.c mvmtxq = iwl_mvm_txq_from_mac80211(txq); txq 1145 drivers/net/wireless/intel/iwlwifi/mvm/ops.c iwl_mvm_mac_itxq_xmit(mvm->hw, txq); txq 634 drivers/net/wireless/intel/iwlwifi/mvm/sta.c bool force, struct iwl_mvm_txq *txq) txq 670 drivers/net/wireless/intel/iwlwifi/mvm/sta.c txq->stopped = true; txq 713 drivers/net/wireless/intel/iwlwifi/mvm/sta.c txq->stopped = false; txq 1386 drivers/net/wireless/intel/iwlwifi/mvm/sta.c struct ieee80211_txq *txq; txq 1392 drivers/net/wireless/intel/iwlwifi/mvm/sta.c txq = container_of((void *)mvmtxq, struct ieee80211_txq, txq 1394 drivers/net/wireless/intel/iwlwifi/mvm/sta.c tid = txq->tid; txq 1398 drivers/net/wireless/intel/iwlwifi/mvm/sta.c iwl_mvm_sta_alloc_queue(mvm, txq->sta, txq->ac, tid); txq 1401 drivers/net/wireless/intel/iwlwifi/mvm/sta.c iwl_mvm_mac_itxq_xmit(mvm->hw, txq); txq 1644 drivers/net/wireless/intel/iwlwifi/mvm/sta.c for (i = 0; i < ARRAY_SIZE(sta->txq); i++) { txq 1646 drivers/net/wireless/intel/iwlwifi/mvm/sta.c iwl_mvm_txq_from_mac80211(sta->txq[i]); txq 1805 drivers/net/wireless/intel/iwlwifi/mvm/sta.c for (i = 0; i < ARRAY_SIZE(sta->txq); i++) { txq 1807 drivers/net/wireless/intel/iwlwifi/mvm/sta.c iwl_mvm_txq_from_mac80211(sta->txq[i]); txq 2020 drivers/net/wireless/intel/iwlwifi/mvm/sta.c int txq; txq 2022 drivers/net/wireless/intel/iwlwifi/mvm/sta.c txq = iwl_mvm_enable_aux_snif_queue_tvqm(mvm, sta->sta_id); txq 2023 drivers/net/wireless/intel/iwlwifi/mvm/sta.c if (txq < 0) { txq 2025 drivers/net/wireless/intel/iwlwifi/mvm/sta.c return txq; txq 2028 drivers/net/wireless/intel/iwlwifi/mvm/sta.c *queue = txq; txq 1772 drivers/net/wireless/intel/iwlwifi/mvm/tx.c int txq, int index, txq 1800 drivers/net/wireless/intel/iwlwifi/mvm/tx.c if (tid_data->txq_id != txq) { txq 1815 drivers/net/wireless/intel/iwlwifi/mvm/tx.c iwl_trans_reclaim(mvm->trans, txq, index, &reclaimed_skbs); txq 1897 drivers/net/wireless/intel/iwlwifi/mvm/tx.c int sta_id, tid, txq, index; txq 1962 drivers/net/wireless/intel/iwlwifi/mvm/tx.c txq = le16_to_cpu(ba_notif->scd_flow); txq 1984 drivers/net/wireless/intel/iwlwifi/mvm/tx.c iwl_mvm_tx_reclaim(mvm, sta_id, tid, txq, index, &ba_info, txq 1994 drivers/net/wireless/intel/iwlwifi/mvm/tx.c le64_to_cpu(ba_notif->bitmap), txq, index, txq 152 drivers/net/wireless/intel/iwlwifi/pcie/ctxt-info-gen3.c cpu_to_le64(trans_pcie->txq[trans_pcie->cmd_queue]->dma_addr); txq 211 drivers/net/wireless/intel/iwlwifi/pcie/ctxt-info.c cpu_to_le64(trans_pcie->txq[trans_pcie->cmd_queue]->dma_addr); txq 390 drivers/net/wireless/intel/iwlwifi/pcie/internal.h iwl_pcie_get_first_tb_dma(struct iwl_txq *txq, int idx) txq 392 drivers/net/wireless/intel/iwlwifi/pcie/internal.h return txq->first_tb_dma + txq 552 drivers/net/wireless/intel/iwlwifi/pcie/internal.h struct iwl_txq *txq[IWL_MAX_TVQM_QUEUES]; txq 691 drivers/net/wireless/intel/iwlwifi/pcie/internal.h struct iwl_txq *txq); txq 698 drivers/net/wireless/intel/iwlwifi/pcie/internal.h struct iwl_txq *txq); txq 706 drivers/net/wireless/intel/iwlwifi/pcie/internal.h struct iwl_txq *txq, u16 byte_cnt, txq 914 drivers/net/wireless/intel/iwlwifi/pcie/internal.h struct iwl_txq *txq, int idx) txq 919 drivers/net/wireless/intel/iwlwifi/pcie/internal.h idx = iwl_pcie_get_cmd_index(txq, idx); txq 921 drivers/net/wireless/intel/iwlwifi/pcie/internal.h return txq->tfds + trans_pcie->tfd_size * idx; txq 976 drivers/net/wireless/intel/iwlwifi/pcie/internal.h struct iwl_txq *txq) txq 980 drivers/net/wireless/intel/iwlwifi/pcie/internal.h if (test_and_clear_bit(txq->id, trans_pcie->queue_stopped)) { txq 981 drivers/net/wireless/intel/iwlwifi/pcie/internal.h IWL_DEBUG_TX_QUEUES(trans, "Wake hwq %d\n", txq->id); txq 982 drivers/net/wireless/intel/iwlwifi/pcie/internal.h iwl_op_mode_queue_not_full(trans->op_mode, txq->id); txq 987 drivers/net/wireless/intel/iwlwifi/pcie/internal.h struct iwl_txq *txq) txq 991 drivers/net/wireless/intel/iwlwifi/pcie/internal.h if (!test_and_set_bit(txq->id, trans_pcie->queue_stopped)) { txq 992 drivers/net/wireless/intel/iwlwifi/pcie/internal.h iwl_op_mode_queue_full(trans->op_mode, txq->id); txq 993 drivers/net/wireless/intel/iwlwifi/pcie/internal.h IWL_DEBUG_TX_QUEUES(trans, "Stop hwq %d\n", txq->id); txq 996 drivers/net/wireless/intel/iwlwifi/pcie/internal.h txq->id); txq 1075 drivers/net/wireless/intel/iwlwifi/pcie/internal.h void iwl_pcie_txq_free_tfd(struct iwl_trans *trans, struct iwl_txq *txq); txq 1079 drivers/net/wireless/intel/iwlwifi/pcie/internal.h int iwl_pcie_txq_init(struct iwl_trans *trans, struct iwl_txq *txq, txq 1082 drivers/net/wireless/intel/iwlwifi/pcie/internal.h struct iwl_txq *txq, int slots_num, bool cmd_queue); txq 1101 drivers/net/wireless/intel/iwlwifi/pcie/internal.h struct iwl_txq *txq); txq 1106 drivers/net/wireless/intel/iwlwifi/pcie/internal.h struct iwl_txq *txq, txq 1244 drivers/net/wireless/intel/iwlwifi/pcie/rx.c struct iwl_txq *txq = trans_pcie->txq[trans_pcie->cmd_queue]; txq 1319 drivers/net/wireless/intel/iwlwifi/pcie/rx.c cmd_index = iwl_pcie_get_cmd_index(txq, index); txq 1329 drivers/net/wireless/intel/iwlwifi/pcie/rx.c kzfree(txq->entries[cmd_index].free_buf); txq 1330 drivers/net/wireless/intel/iwlwifi/pcie/rx.c txq->entries[cmd_index].free_buf = NULL; txq 1610 drivers/net/wireless/intel/iwlwifi/pcie/rx.c if (!trans_pcie->txq[i]) txq 1612 drivers/net/wireless/intel/iwlwifi/pcie/rx.c del_timer(&trans_pcie->txq[i]->stuck_timer); txq 2222 drivers/net/wireless/intel/iwlwifi/pcie/trans.c struct iwl_txq *txq = trans_pcie->txq[queue]; txq 2225 drivers/net/wireless/intel/iwlwifi/pcie/trans.c spin_lock_bh(&txq->lock); txq 2229 drivers/net/wireless/intel/iwlwifi/pcie/trans.c if (txq->frozen == freeze) txq 2235 drivers/net/wireless/intel/iwlwifi/pcie/trans.c txq->frozen = freeze; txq 2237 drivers/net/wireless/intel/iwlwifi/pcie/trans.c if (txq->read_ptr == txq->write_ptr) txq 2242 drivers/net/wireless/intel/iwlwifi/pcie/trans.c txq->stuck_timer.expires))) { txq 2250 drivers/net/wireless/intel/iwlwifi/pcie/trans.c txq->frozen_expiry_remainder = txq 2251 drivers/net/wireless/intel/iwlwifi/pcie/trans.c txq->stuck_timer.expires - now; txq 2252 drivers/net/wireless/intel/iwlwifi/pcie/trans.c del_timer(&txq->stuck_timer); txq 2260 drivers/net/wireless/intel/iwlwifi/pcie/trans.c mod_timer(&txq->stuck_timer, txq 2261 drivers/net/wireless/intel/iwlwifi/pcie/trans.c now + txq->frozen_expiry_remainder); txq 2264 drivers/net/wireless/intel/iwlwifi/pcie/trans.c spin_unlock_bh(&txq->lock); txq 2274 drivers/net/wireless/intel/iwlwifi/pcie/trans.c struct iwl_txq *txq = trans_pcie->txq[i]; txq 2279 drivers/net/wireless/intel/iwlwifi/pcie/trans.c spin_lock_bh(&txq->lock); txq 2281 drivers/net/wireless/intel/iwlwifi/pcie/trans.c if (!block && !(WARN_ON_ONCE(!txq->block))) { txq 2282 drivers/net/wireless/intel/iwlwifi/pcie/trans.c txq->block--; txq 2283 drivers/net/wireless/intel/iwlwifi/pcie/trans.c if (!txq->block) { txq 2285 drivers/net/wireless/intel/iwlwifi/pcie/trans.c txq->write_ptr | (i << 8)); txq 2288 drivers/net/wireless/intel/iwlwifi/pcie/trans.c txq->block++; txq 2291 drivers/net/wireless/intel/iwlwifi/pcie/trans.c spin_unlock_bh(&txq->lock); txq 2297 drivers/net/wireless/intel/iwlwifi/pcie/trans.c void iwl_trans_pcie_log_scd_error(struct iwl_trans *trans, struct iwl_txq *txq) txq 2299 drivers/net/wireless/intel/iwlwifi/pcie/trans.c u32 txq_id = txq->id; txq 2306 drivers/net/wireless/intel/iwlwifi/pcie/trans.c txq->read_ptr, txq->write_ptr); txq 2318 drivers/net/wireless/intel/iwlwifi/pcie/trans.c jiffies_to_msecs(txq->wd_timeout), txq 2319 drivers/net/wireless/intel/iwlwifi/pcie/trans.c txq->read_ptr, txq->write_ptr, txq 2346 drivers/net/wireless/intel/iwlwifi/pcie/trans.c struct iwl_txq *txq; txq 2359 drivers/net/wireless/intel/iwlwifi/pcie/trans.c txq = trans_pcie->txq[txq_idx]; txq 2361 drivers/net/wireless/intel/iwlwifi/pcie/trans.c spin_lock_bh(&txq->lock); txq 2362 drivers/net/wireless/intel/iwlwifi/pcie/trans.c overflow_tx = txq->overflow_tx || txq 2363 drivers/net/wireless/intel/iwlwifi/pcie/trans.c !skb_queue_empty(&txq->overflow_q); txq 2364 drivers/net/wireless/intel/iwlwifi/pcie/trans.c spin_unlock_bh(&txq->lock); txq 2366 drivers/net/wireless/intel/iwlwifi/pcie/trans.c wr_ptr = READ_ONCE(txq->write_ptr); txq 2368 drivers/net/wireless/intel/iwlwifi/pcie/trans.c while ((txq->read_ptr != READ_ONCE(txq->write_ptr) || txq 2372 drivers/net/wireless/intel/iwlwifi/pcie/trans.c u8 write_ptr = READ_ONCE(txq->write_ptr); txq 2387 drivers/net/wireless/intel/iwlwifi/pcie/trans.c spin_lock_bh(&txq->lock); txq 2388 drivers/net/wireless/intel/iwlwifi/pcie/trans.c overflow_tx = txq->overflow_tx || txq 2389 drivers/net/wireless/intel/iwlwifi/pcie/trans.c !skb_queue_empty(&txq->overflow_q); txq 2390 drivers/net/wireless/intel/iwlwifi/pcie/trans.c spin_unlock_bh(&txq->lock); txq 2393 drivers/net/wireless/intel/iwlwifi/pcie/trans.c if (txq->read_ptr != txq->write_ptr) { txq 2396 drivers/net/wireless/intel/iwlwifi/pcie/trans.c iwl_trans_pcie_log_scd_error(trans, txq); txq 2551 drivers/net/wireless/intel/iwlwifi/pcie/trans.c struct iwl_txq *txq; txq 2571 drivers/net/wireless/intel/iwlwifi/pcie/trans.c txq = trans_pcie->txq[cnt]; txq 2574 drivers/net/wireless/intel/iwlwifi/pcie/trans.c cnt, txq->read_ptr, txq->write_ptr, txq 2577 drivers/net/wireless/intel/iwlwifi/pcie/trans.c txq->need_update, txq->frozen, txq 3229 drivers/net/wireless/intel/iwlwifi/pcie/trans.c struct iwl_txq *cmdq = trans_pcie->txq[trans_pcie->cmd_queue]; txq 79 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c for (txq_id = 0; txq_id < ARRAY_SIZE(trans_pcie->txq); txq_id++) { txq 80 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c if (!trans_pcie->txq[txq_id]) txq 90 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c struct iwl_txq *txq, u16 byte_cnt, txq 93 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c struct iwlagn_scd_bc_tbl *scd_bc_tbl = txq->bc_tbl.addr; txq 95 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c struct iwl_gen3_bc_tbl *scd_bc_tbl_gen3 = txq->bc_tbl.addr; txq 96 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c int idx = iwl_pcie_get_cmd_index(txq, txq->write_ptr); txq 101 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c if (WARN(idx >= txq->n_window, "%d >= %d\n", idx, txq->n_window)) txq 136 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c struct iwl_txq *txq) txq 138 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c lockdep_assert_held(&txq->lock); txq 140 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c IWL_DEBUG_TX(trans, "Q:%d WR: 0x%x\n", txq->id, txq->write_ptr); txq 146 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c iwl_write32(trans, HBUS_TARG_WRPTR, txq->write_ptr | (txq->id << 16)); txq 187 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c static void iwl_pcie_gen2_free_tfd(struct iwl_trans *trans, struct iwl_txq *txq) txq 192 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c int idx = iwl_pcie_get_cmd_index(txq, txq->read_ptr); txq 194 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c lockdep_assert_held(&txq->lock); txq 196 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c iwl_pcie_gen2_tfd_unmap(trans, &txq->entries[idx].meta, txq 197 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c iwl_pcie_get_tfd(trans, txq, idx)); txq 200 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c if (txq->entries) { txq 203 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c skb = txq->entries[idx].skb; txq 211 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c txq->entries[idx].skb = NULL; txq 374 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c struct iwl_txq *txq, txq 381 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c int idx = iwl_pcie_get_cmd_index(txq, txq->write_ptr); txq 382 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c struct iwl_tfh_tfd *tfd = iwl_pcie_get_tfd(trans, txq, idx); txq 387 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c tb_phys = iwl_pcie_get_first_tb_dma(txq, idx); txq 419 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c memcpy(&txq->first_tb_bufs[idx], dev_cmd, IWL_FIRST_TB_SIZE); txq 463 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c struct iwl_txq *txq, txq 471 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c int idx = iwl_pcie_get_cmd_index(txq, txq->write_ptr); txq 472 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c struct iwl_tfh_tfd *tfd = iwl_pcie_get_tfd(trans, txq, idx); txq 478 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c tb_phys = iwl_pcie_get_first_tb_dma(txq, idx); txq 481 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c memcpy(&txq->first_tb_bufs[idx], dev_cmd, IWL_FIRST_TB_SIZE); txq 551 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c struct iwl_txq *txq, txq 557 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c int idx = iwl_pcie_get_cmd_index(txq, txq->write_ptr); txq 558 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c struct iwl_tfh_tfd *tfd = iwl_pcie_get_tfd(trans, txq, idx); txq 584 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c return iwl_pcie_gen2_build_tx_amsdu(trans, txq, dev_cmd, skb, txq 587 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c return iwl_pcie_gen2_build_tx(trans, txq, dev_cmd, skb, out_meta, txq 596 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c struct iwl_txq *txq = trans_pcie->txq[txq_id]; txq 610 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c spin_lock(&txq->lock); txq 612 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c if (iwl_queue_space(trans, txq) < txq->high_mark) { txq 613 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c iwl_stop_queue(trans, txq); txq 616 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c if (unlikely(iwl_queue_space(trans, txq) < 3)) { txq 623 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c __skb_queue_tail(&txq->overflow_q, skb); txq 624 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c spin_unlock(&txq->lock); txq 629 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c idx = iwl_pcie_get_cmd_index(txq, txq->write_ptr); txq 632 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c txq->entries[idx].skb = skb; txq 633 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c txq->entries[idx].cmd = dev_cmd; txq 640 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c out_meta = &txq->entries[idx].meta; txq 643 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c tfd = iwl_pcie_gen2_build_tfd(trans, txq, dev_cmd, skb, out_meta); txq 645 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c spin_unlock(&txq->lock); txq 662 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c iwl_pcie_gen2_update_byte_tbl(trans_pcie, txq, cmd_len, txq 666 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c if (txq->read_ptr == txq->write_ptr && txq->wd_timeout) txq 667 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c mod_timer(&txq->stuck_timer, jiffies + txq->wd_timeout); txq 670 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c txq->write_ptr = iwl_queue_inc_wrap(trans, txq->write_ptr); txq 671 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c iwl_pcie_gen2_txq_inc_wr_ptr(trans, txq); txq 676 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c spin_unlock(&txq->lock); txq 695 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c struct iwl_txq *txq = trans_pcie->txq[trans_pcie->cmd_queue]; txq 776 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c spin_lock_bh(&txq->lock); txq 778 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c idx = iwl_pcie_get_cmd_index(txq, txq->write_ptr); txq 779 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c tfd = iwl_pcie_get_tfd(trans, txq, txq->write_ptr); txq 782 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c if (iwl_queue_space(trans, txq) < ((cmd->flags & CMD_ASYNC) ? 2 : 1)) { txq 783 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c spin_unlock_bh(&txq->lock); txq 791 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c out_cmd = txq->entries[idx].cmd; txq 792 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c out_meta = &txq->entries[idx].meta; txq 808 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c INDEX_TO_SEQ(txq->write_ptr)); txq 855 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c cmd_size, txq->write_ptr, idx, trans_pcie->cmd_queue); txq 859 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c memcpy(&txq->first_tb_bufs[idx], out_cmd, tb0_size); txq 860 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c iwl_pcie_gen2_set_tb(trans, tfd, iwl_pcie_get_first_tb_dma(txq, idx), txq 901 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c if (WARN_ON_ONCE(txq->entries[idx].free_buf)) txq 902 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c kzfree(txq->entries[idx].free_buf); txq 903 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c txq->entries[idx].free_buf = dup_buf; txq 908 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c if (txq->read_ptr == txq->write_ptr && txq->wd_timeout) txq 909 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c mod_timer(&txq->stuck_timer, jiffies + txq->wd_timeout); txq 913 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c txq->write_ptr = iwl_queue_inc_wrap(trans, txq->write_ptr); txq 914 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c iwl_pcie_gen2_txq_inc_wr_ptr(trans, txq); txq 918 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c spin_unlock_bh(&txq->lock); txq 932 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c struct iwl_txq *txq = trans_pcie->txq[trans_pcie->cmd_queue]; txq 963 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c txq->read_ptr, txq->write_ptr); txq 1004 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c txq->entries[cmd_idx].meta.flags &= ~CMD_WANT_SKB; txq 1051 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c struct iwl_txq *txq = trans_pcie->txq[txq_id]; txq 1053 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c spin_lock_bh(&txq->lock); txq 1054 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c while (txq->write_ptr != txq->read_ptr) { txq 1056 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c txq_id, txq->read_ptr); txq 1059 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c int idx = iwl_pcie_get_cmd_index(txq, txq->read_ptr); txq 1060 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c struct sk_buff *skb = txq->entries[idx].skb; txq 1067 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c iwl_pcie_gen2_free_tfd(trans, txq); txq 1068 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c txq->read_ptr = iwl_queue_inc_wrap(trans, txq->read_ptr); txq 1071 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c while (!skb_queue_empty(&txq->overflow_q)) { txq 1072 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c struct sk_buff *skb = __skb_dequeue(&txq->overflow_q); txq 1077 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c spin_unlock_bh(&txq->lock); txq 1080 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c iwl_wake_queue(trans, txq); txq 1084 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c struct iwl_txq *txq) txq 1090 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c if (txq->tfds) { txq 1092 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c trans_pcie->tfd_size * txq->n_window, txq 1093 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c txq->tfds, txq->dma_addr); txq 1095 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c sizeof(*txq->first_tb_bufs) * txq->n_window, txq 1096 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c txq->first_tb_bufs, txq->first_tb_dma); txq 1099 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c kfree(txq->entries); txq 1100 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c iwl_pcie_free_dma_ptr(trans, &txq->bc_tbl); txq 1101 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c kfree(txq); txq 1115 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c struct iwl_txq *txq = trans_pcie->txq[txq_id]; txq 1118 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c if (WARN_ON(!txq)) txq 1125 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c for (i = 0; i < txq->n_window; i++) { txq 1126 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c kzfree(txq->entries[i].cmd); txq 1127 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c kzfree(txq->entries[i].free_buf); txq 1129 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c del_timer_sync(&txq->stuck_timer); txq 1131 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c iwl_pcie_gen2_txq_free_memory(trans, txq); txq 1133 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c trans_pcie->txq[txq_id] = NULL; txq 1144 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c struct iwl_txq *txq; txq 1145 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c txq = kzalloc(sizeof(*txq), GFP_KERNEL); txq 1146 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c if (!txq) txq 1148 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c ret = iwl_pcie_alloc_dma_ptr(trans, &txq->bc_tbl, txq 1155 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c kfree(txq); txq 1159 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c ret = iwl_pcie_txq_alloc(trans, txq, size, false); txq 1164 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c ret = iwl_pcie_txq_init(trans, txq, size, false); txq 1170 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c txq->wd_timeout = msecs_to_jiffies(timeout); txq 1172 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c *intxq = txq; txq 1176 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c iwl_pcie_gen2_txq_free_memory(trans, txq); txq 1181 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c struct iwl_txq *txq, txq 1199 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c if (qid >= ARRAY_SIZE(trans_pcie->txq)) { txq 1211 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c txq->id = qid; txq 1212 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c trans_pcie->txq[qid] = txq; txq 1216 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c txq->read_ptr = wr_ptr; txq 1217 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c txq->write_ptr = wr_ptr; txq 1226 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c iwl_pcie_gen2_txq_free_memory(trans, txq); txq 1235 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c struct iwl_txq *txq = NULL; txq 1249 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c ret = iwl_trans_pcie_dyn_txq_alloc_dma(trans, &txq, size, timeout); txq 1253 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c cmd.tfdq_addr = cpu_to_le64(txq->dma_addr); txq 1254 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c cmd.byte_cnt_addr = cpu_to_le64(txq->bc_tbl.dma); txq 1261 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c return iwl_trans_pcie_txq_alloc_response(trans, txq, &hcmd); txq 1264 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c iwl_pcie_gen2_txq_free_memory(trans, txq); txq 1286 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c iwl_pcie_gen2_txq_free_memory(trans, trans_pcie->txq[queue]); txq 1287 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c trans_pcie->txq[queue] = NULL; txq 1300 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c for (i = 0; i < ARRAY_SIZE(trans_pcie->txq); i++) { txq 1301 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c if (!trans_pcie->txq[i]) txq 1315 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c if (!trans_pcie->txq[txq_id]) { txq 1321 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c trans_pcie->txq[txq_id] = queue; txq 1328 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c queue = trans_pcie->txq[txq_id]; txq 1337 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c trans_pcie->txq[txq_id]->id = txq_id; txq 185 drivers/net/wireless/intel/iwlwifi/pcie/tx.c struct iwl_txq *txq = from_timer(txq, t, stuck_timer); txq 186 drivers/net/wireless/intel/iwlwifi/pcie/tx.c struct iwl_trans_pcie *trans_pcie = txq->trans_pcie; txq 189 drivers/net/wireless/intel/iwlwifi/pcie/tx.c spin_lock(&txq->lock); txq 191 drivers/net/wireless/intel/iwlwifi/pcie/tx.c if (txq->read_ptr == txq->write_ptr) { txq 192 drivers/net/wireless/intel/iwlwifi/pcie/tx.c spin_unlock(&txq->lock); txq 195 drivers/net/wireless/intel/iwlwifi/pcie/tx.c spin_unlock(&txq->lock); txq 197 drivers/net/wireless/intel/iwlwifi/pcie/tx.c iwl_trans_pcie_log_scd_error(trans, txq); txq 206 drivers/net/wireless/intel/iwlwifi/pcie/tx.c struct iwl_txq *txq, u16 byte_cnt, txq 211 drivers/net/wireless/intel/iwlwifi/pcie/tx.c int write_ptr = txq->write_ptr; txq 212 drivers/net/wireless/intel/iwlwifi/pcie/tx.c int txq_id = txq->id; txq 216 drivers/net/wireless/intel/iwlwifi/pcie/tx.c struct iwl_device_tx_cmd *dev_cmd = txq->entries[txq->write_ptr].cmd; txq 251 drivers/net/wireless/intel/iwlwifi/pcie/tx.c struct iwl_txq *txq) txq 256 drivers/net/wireless/intel/iwlwifi/pcie/tx.c int txq_id = txq->id; txq 257 drivers/net/wireless/intel/iwlwifi/pcie/tx.c int read_ptr = txq->read_ptr; txq 260 drivers/net/wireless/intel/iwlwifi/pcie/tx.c struct iwl_device_tx_cmd *dev_cmd = txq->entries[read_ptr].cmd; txq 281 drivers/net/wireless/intel/iwlwifi/pcie/tx.c struct iwl_txq *txq) txq 285 drivers/net/wireless/intel/iwlwifi/pcie/tx.c int txq_id = txq->id; txq 287 drivers/net/wireless/intel/iwlwifi/pcie/tx.c lockdep_assert_held(&txq->lock); txq 310 drivers/net/wireless/intel/iwlwifi/pcie/tx.c txq->need_update = true; txq 319 drivers/net/wireless/intel/iwlwifi/pcie/tx.c IWL_DEBUG_TX(trans, "Q:%d WR: 0x%x\n", txq_id, txq->write_ptr); txq 320 drivers/net/wireless/intel/iwlwifi/pcie/tx.c if (!txq->block) txq 322 drivers/net/wireless/intel/iwlwifi/pcie/tx.c txq->write_ptr | (txq_id << 8)); txq 331 drivers/net/wireless/intel/iwlwifi/pcie/tx.c struct iwl_txq *txq = trans_pcie->txq[i]; txq 336 drivers/net/wireless/intel/iwlwifi/pcie/tx.c spin_lock_bh(&txq->lock); txq 337 drivers/net/wireless/intel/iwlwifi/pcie/tx.c if (txq->need_update) { txq 338 drivers/net/wireless/intel/iwlwifi/pcie/tx.c iwl_pcie_txq_inc_wr_ptr(trans, txq); txq 339 drivers/net/wireless/intel/iwlwifi/pcie/tx.c txq->need_update = false; txq 341 drivers/net/wireless/intel/iwlwifi/pcie/tx.c spin_unlock_bh(&txq->lock); txq 405 drivers/net/wireless/intel/iwlwifi/pcie/tx.c struct iwl_txq *txq, int index) txq 409 drivers/net/wireless/intel/iwlwifi/pcie/tx.c void *tfd = iwl_pcie_get_tfd(trans, txq, index); txq 460 drivers/net/wireless/intel/iwlwifi/pcie/tx.c void iwl_pcie_txq_free_tfd(struct iwl_trans *trans, struct iwl_txq *txq) txq 465 drivers/net/wireless/intel/iwlwifi/pcie/tx.c int rd_ptr = txq->read_ptr; txq 466 drivers/net/wireless/intel/iwlwifi/pcie/tx.c int idx = iwl_pcie_get_cmd_index(txq, rd_ptr); txq 468 drivers/net/wireless/intel/iwlwifi/pcie/tx.c lockdep_assert_held(&txq->lock); txq 473 drivers/net/wireless/intel/iwlwifi/pcie/tx.c iwl_pcie_tfd_unmap(trans, &txq->entries[idx].meta, txq, rd_ptr); txq 476 drivers/net/wireless/intel/iwlwifi/pcie/tx.c if (txq->entries) { txq 479 drivers/net/wireless/intel/iwlwifi/pcie/tx.c skb = txq->entries[idx].skb; txq 487 drivers/net/wireless/intel/iwlwifi/pcie/tx.c txq->entries[idx].skb = NULL; txq 492 drivers/net/wireless/intel/iwlwifi/pcie/tx.c static int iwl_pcie_txq_build_tfd(struct iwl_trans *trans, struct iwl_txq *txq, txq 499 drivers/net/wireless/intel/iwlwifi/pcie/tx.c tfd = txq->tfds + trans_pcie->tfd_size * txq->write_ptr; txq 522 drivers/net/wireless/intel/iwlwifi/pcie/tx.c int iwl_pcie_txq_alloc(struct iwl_trans *trans, struct iwl_txq *txq, txq 531 drivers/net/wireless/intel/iwlwifi/pcie/tx.c if (WARN_ON(txq->entries || txq->tfds)) txq 537 drivers/net/wireless/intel/iwlwifi/pcie/tx.c timer_setup(&txq->stuck_timer, iwl_pcie_txq_stuck_timer, 0); txq 538 drivers/net/wireless/intel/iwlwifi/pcie/tx.c txq->trans_pcie = trans_pcie; txq 540 drivers/net/wireless/intel/iwlwifi/pcie/tx.c txq->n_window = slots_num; txq 542 drivers/net/wireless/intel/iwlwifi/pcie/tx.c txq->entries = kcalloc(slots_num, txq 546 drivers/net/wireless/intel/iwlwifi/pcie/tx.c if (!txq->entries) txq 551 drivers/net/wireless/intel/iwlwifi/pcie/tx.c txq->entries[i].cmd = txq 554 drivers/net/wireless/intel/iwlwifi/pcie/tx.c if (!txq->entries[i].cmd) txq 560 drivers/net/wireless/intel/iwlwifi/pcie/tx.c txq->tfds = dma_alloc_coherent(trans->dev, tfd_sz, txq 561 drivers/net/wireless/intel/iwlwifi/pcie/tx.c &txq->dma_addr, GFP_KERNEL); txq 562 drivers/net/wireless/intel/iwlwifi/pcie/tx.c if (!txq->tfds) txq 565 drivers/net/wireless/intel/iwlwifi/pcie/tx.c BUILD_BUG_ON(IWL_FIRST_TB_SIZE_ALIGN != sizeof(*txq->first_tb_bufs)); txq 567 drivers/net/wireless/intel/iwlwifi/pcie/tx.c tb0_buf_sz = sizeof(*txq->first_tb_bufs) * slots_num; txq 569 drivers/net/wireless/intel/iwlwifi/pcie/tx.c txq->first_tb_bufs = dma_alloc_coherent(trans->dev, tb0_buf_sz, txq 570 drivers/net/wireless/intel/iwlwifi/pcie/tx.c &txq->first_tb_dma, txq 572 drivers/net/wireless/intel/iwlwifi/pcie/tx.c if (!txq->first_tb_bufs) txq 577 drivers/net/wireless/intel/iwlwifi/pcie/tx.c dma_free_coherent(trans->dev, tfd_sz, txq->tfds, txq->dma_addr); txq 579 drivers/net/wireless/intel/iwlwifi/pcie/tx.c if (txq->entries && cmd_queue) txq 581 drivers/net/wireless/intel/iwlwifi/pcie/tx.c kfree(txq->entries[i].cmd); txq 582 drivers/net/wireless/intel/iwlwifi/pcie/tx.c kfree(txq->entries); txq 583 drivers/net/wireless/intel/iwlwifi/pcie/tx.c txq->entries = NULL; txq 589 drivers/net/wireless/intel/iwlwifi/pcie/tx.c int iwl_pcie_txq_init(struct iwl_trans *trans, struct iwl_txq *txq, txq 596 drivers/net/wireless/intel/iwlwifi/pcie/tx.c txq->need_update = false; txq 606 drivers/net/wireless/intel/iwlwifi/pcie/tx.c ret = iwl_queue_init(txq, slots_num); txq 610 drivers/net/wireless/intel/iwlwifi/pcie/tx.c spin_lock_init(&txq->lock); txq 615 drivers/net/wireless/intel/iwlwifi/pcie/tx.c lockdep_set_class(&txq->lock, &iwl_pcie_cmd_queue_lock_class); txq 618 drivers/net/wireless/intel/iwlwifi/pcie/tx.c __skb_queue_head_init(&txq->overflow_q); txq 658 drivers/net/wireless/intel/iwlwifi/pcie/tx.c struct iwl_txq *txq = trans_pcie->txq[txq_id]; txq 660 drivers/net/wireless/intel/iwlwifi/pcie/tx.c spin_lock_bh(&txq->lock); txq 661 drivers/net/wireless/intel/iwlwifi/pcie/tx.c while (txq->write_ptr != txq->read_ptr) { txq 663 drivers/net/wireless/intel/iwlwifi/pcie/tx.c txq_id, txq->read_ptr); txq 666 drivers/net/wireless/intel/iwlwifi/pcie/tx.c struct sk_buff *skb = txq->entries[txq->read_ptr].skb; txq 673 drivers/net/wireless/intel/iwlwifi/pcie/tx.c iwl_pcie_txq_free_tfd(trans, txq); txq 674 drivers/net/wireless/intel/iwlwifi/pcie/tx.c txq->read_ptr = iwl_queue_inc_wrap(trans, txq->read_ptr); txq 676 drivers/net/wireless/intel/iwlwifi/pcie/tx.c if (txq->read_ptr == txq->write_ptr) { txq 686 drivers/net/wireless/intel/iwlwifi/pcie/tx.c while (!skb_queue_empty(&txq->overflow_q)) { txq 687 drivers/net/wireless/intel/iwlwifi/pcie/tx.c struct sk_buff *skb = __skb_dequeue(&txq->overflow_q); txq 692 drivers/net/wireless/intel/iwlwifi/pcie/tx.c spin_unlock_bh(&txq->lock); txq 695 drivers/net/wireless/intel/iwlwifi/pcie/tx.c iwl_wake_queue(trans, txq); txq 709 drivers/net/wireless/intel/iwlwifi/pcie/tx.c struct iwl_txq *txq = trans_pcie->txq[txq_id]; txq 713 drivers/net/wireless/intel/iwlwifi/pcie/tx.c if (WARN_ON(!txq)) txq 720 drivers/net/wireless/intel/iwlwifi/pcie/tx.c for (i = 0; i < txq->n_window; i++) { txq 721 drivers/net/wireless/intel/iwlwifi/pcie/tx.c kzfree(txq->entries[i].cmd); txq 722 drivers/net/wireless/intel/iwlwifi/pcie/tx.c kzfree(txq->entries[i].free_buf); txq 726 drivers/net/wireless/intel/iwlwifi/pcie/tx.c if (txq->tfds) { txq 730 drivers/net/wireless/intel/iwlwifi/pcie/tx.c txq->tfds, txq->dma_addr); txq 731 drivers/net/wireless/intel/iwlwifi/pcie/tx.c txq->dma_addr = 0; txq 732 drivers/net/wireless/intel/iwlwifi/pcie/tx.c txq->tfds = NULL; txq 735 drivers/net/wireless/intel/iwlwifi/pcie/tx.c sizeof(*txq->first_tb_bufs) * txq->n_window, txq 736 drivers/net/wireless/intel/iwlwifi/pcie/tx.c txq->first_tb_bufs, txq->first_tb_dma); txq 739 drivers/net/wireless/intel/iwlwifi/pcie/tx.c kfree(txq->entries); txq 740 drivers/net/wireless/intel/iwlwifi/pcie/tx.c txq->entries = NULL; txq 742 drivers/net/wireless/intel/iwlwifi/pcie/tx.c del_timer_sync(&txq->stuck_timer); txq 745 drivers/net/wireless/intel/iwlwifi/pcie/tx.c memset(txq, 0, sizeof(*txq)); txq 819 drivers/net/wireless/intel/iwlwifi/pcie/tx.c struct iwl_txq *txq = trans_pcie->txq[txq_id]; txq 823 drivers/net/wireless/intel/iwlwifi/pcie/tx.c txq->dma_addr); txq 827 drivers/net/wireless/intel/iwlwifi/pcie/tx.c txq->dma_addr >> 8); txq 829 drivers/net/wireless/intel/iwlwifi/pcie/tx.c txq->read_ptr = 0; txq 830 drivers/net/wireless/intel/iwlwifi/pcie/tx.c txq->write_ptr = 0; txq 928 drivers/net/wireless/intel/iwlwifi/pcie/tx.c trans_pcie->txq[txq_id] = NULL; txq 997 drivers/net/wireless/intel/iwlwifi/pcie/tx.c trans_pcie->txq[txq_id] = &trans_pcie->txq_memory[txq_id]; txq 998 drivers/net/wireless/intel/iwlwifi/pcie/tx.c ret = iwl_pcie_txq_alloc(trans, trans_pcie->txq[txq_id], txq 1004 drivers/net/wireless/intel/iwlwifi/pcie/tx.c trans_pcie->txq[txq_id]->id = txq_id; txq 1051 drivers/net/wireless/intel/iwlwifi/pcie/tx.c ret = iwl_pcie_txq_init(trans, trans_pcie->txq[txq_id], txq 1065 drivers/net/wireless/intel/iwlwifi/pcie/tx.c trans_pcie->txq[txq_id]->dma_addr >> 8); txq 1081 drivers/net/wireless/intel/iwlwifi/pcie/tx.c static inline void iwl_pcie_txq_progress(struct iwl_txq *txq) txq 1083 drivers/net/wireless/intel/iwlwifi/pcie/tx.c lockdep_assert_held(&txq->lock); txq 1085 drivers/net/wireless/intel/iwlwifi/pcie/tx.c if (!txq->wd_timeout) txq 1092 drivers/net/wireless/intel/iwlwifi/pcie/tx.c if (txq->frozen) txq 1099 drivers/net/wireless/intel/iwlwifi/pcie/tx.c if (txq->read_ptr == txq->write_ptr) txq 1100 drivers/net/wireless/intel/iwlwifi/pcie/tx.c del_timer(&txq->stuck_timer); txq 1102 drivers/net/wireless/intel/iwlwifi/pcie/tx.c mod_timer(&txq->stuck_timer, jiffies + txq->wd_timeout); txq 1110 drivers/net/wireless/intel/iwlwifi/pcie/tx.c struct iwl_txq *txq = trans_pcie->txq[txq_id]; txq 1111 drivers/net/wireless/intel/iwlwifi/pcie/tx.c int tfd_num = iwl_pcie_get_cmd_index(txq, ssn); txq 1112 drivers/net/wireless/intel/iwlwifi/pcie/tx.c int read_ptr = iwl_pcie_get_cmd_index(txq, txq->read_ptr); txq 1119 drivers/net/wireless/intel/iwlwifi/pcie/tx.c spin_lock_bh(&txq->lock); txq 1131 drivers/net/wireless/intel/iwlwifi/pcie/tx.c txq_id, txq->read_ptr, tfd_num, ssn); txq 1137 drivers/net/wireless/intel/iwlwifi/pcie/tx.c if (!iwl_queue_used(txq, last_to_free)) { txq 1142 drivers/net/wireless/intel/iwlwifi/pcie/tx.c txq->write_ptr, txq->read_ptr); txq 1151 drivers/net/wireless/intel/iwlwifi/pcie/tx.c txq->read_ptr = iwl_queue_inc_wrap(trans, txq->read_ptr), txq 1152 drivers/net/wireless/intel/iwlwifi/pcie/tx.c read_ptr = iwl_pcie_get_cmd_index(txq, txq->read_ptr)) { txq 1153 drivers/net/wireless/intel/iwlwifi/pcie/tx.c struct sk_buff *skb = txq->entries[read_ptr].skb; txq 1162 drivers/net/wireless/intel/iwlwifi/pcie/tx.c txq->entries[read_ptr].skb = NULL; txq 1165 drivers/net/wireless/intel/iwlwifi/pcie/tx.c iwl_pcie_txq_inval_byte_cnt_tbl(trans, txq); txq 1167 drivers/net/wireless/intel/iwlwifi/pcie/tx.c iwl_pcie_txq_free_tfd(trans, txq); txq 1170 drivers/net/wireless/intel/iwlwifi/pcie/tx.c iwl_pcie_txq_progress(txq); txq 1172 drivers/net/wireless/intel/iwlwifi/pcie/tx.c if (iwl_queue_space(trans, txq) > txq->low_mark && txq 1177 drivers/net/wireless/intel/iwlwifi/pcie/tx.c skb_queue_splice_init(&txq->overflow_q, &overflow_skbs); txq 1186 drivers/net/wireless/intel/iwlwifi/pcie/tx.c txq->overflow_tx = true; txq 1195 drivers/net/wireless/intel/iwlwifi/pcie/tx.c spin_unlock_bh(&txq->lock); txq 1212 drivers/net/wireless/intel/iwlwifi/pcie/tx.c if (iwl_queue_space(trans, txq) > txq->low_mark) txq 1213 drivers/net/wireless/intel/iwlwifi/pcie/tx.c iwl_wake_queue(trans, txq); txq 1215 drivers/net/wireless/intel/iwlwifi/pcie/tx.c spin_lock_bh(&txq->lock); txq 1216 drivers/net/wireless/intel/iwlwifi/pcie/tx.c txq->overflow_tx = false; txq 1220 drivers/net/wireless/intel/iwlwifi/pcie/tx.c spin_unlock_bh(&txq->lock); txq 1227 drivers/net/wireless/intel/iwlwifi/pcie/tx.c struct iwl_txq *txq = trans_pcie->txq[txq_id]; txq 1229 drivers/net/wireless/intel/iwlwifi/pcie/tx.c spin_lock_bh(&txq->lock); txq 1231 drivers/net/wireless/intel/iwlwifi/pcie/tx.c txq->write_ptr = ptr; txq 1232 drivers/net/wireless/intel/iwlwifi/pcie/tx.c txq->read_ptr = txq->write_ptr; txq 1234 drivers/net/wireless/intel/iwlwifi/pcie/tx.c spin_unlock_bh(&txq->lock); txq 1287 drivers/net/wireless/intel/iwlwifi/pcie/tx.c struct iwl_txq *txq = trans_pcie->txq[txq_id]; txq 1292 drivers/net/wireless/intel/iwlwifi/pcie/tx.c lockdep_assert_held(&txq->lock); txq 1294 drivers/net/wireless/intel/iwlwifi/pcie/tx.c idx = iwl_pcie_get_cmd_index(txq, idx); txq 1295 drivers/net/wireless/intel/iwlwifi/pcie/tx.c r = iwl_pcie_get_cmd_index(txq, txq->read_ptr); txq 1298 drivers/net/wireless/intel/iwlwifi/pcie/tx.c (!iwl_queue_used(txq, idx))) { txq 1303 drivers/net/wireless/intel/iwlwifi/pcie/tx.c txq->write_ptr, txq->read_ptr); txq 1309 drivers/net/wireless/intel/iwlwifi/pcie/tx.c txq->read_ptr = iwl_queue_inc_wrap(trans, txq->read_ptr); txq 1313 drivers/net/wireless/intel/iwlwifi/pcie/tx.c idx, txq->write_ptr, r); txq 1318 drivers/net/wireless/intel/iwlwifi/pcie/tx.c if (txq->read_ptr == txq->write_ptr) { txq 1324 drivers/net/wireless/intel/iwlwifi/pcie/tx.c iwl_pcie_txq_progress(txq); txq 1361 drivers/net/wireless/intel/iwlwifi/pcie/tx.c struct iwl_txq *txq = trans_pcie->txq[txq_id]; txq 1368 drivers/net/wireless/intel/iwlwifi/pcie/tx.c txq->wd_timeout = msecs_to_jiffies(wdg_timeout); txq 1393 drivers/net/wireless/intel/iwlwifi/pcie/tx.c txq->ampdu = true; txq 1402 drivers/net/wireless/intel/iwlwifi/pcie/tx.c ssn = txq->read_ptr; txq 1417 drivers/net/wireless/intel/iwlwifi/pcie/tx.c !((ssn - txq->write_ptr) & 0x3f) && txq 1418 drivers/net/wireless/intel/iwlwifi/pcie/tx.c (ssn != txq->write_ptr); txq 1425 drivers/net/wireless/intel/iwlwifi/pcie/tx.c txq->read_ptr = (ssn & 0xff); txq 1426 drivers/net/wireless/intel/iwlwifi/pcie/tx.c txq->write_ptr = (ssn & 0xff); txq 1472 drivers/net/wireless/intel/iwlwifi/pcie/tx.c struct iwl_txq *txq = trans_pcie->txq[txq_id]; txq 1474 drivers/net/wireless/intel/iwlwifi/pcie/tx.c txq->ampdu = !shared_mode; txq 1485 drivers/net/wireless/intel/iwlwifi/pcie/tx.c trans_pcie->txq[txq_id]->frozen_expiry_remainder = 0; txq 1486 drivers/net/wireless/intel/iwlwifi/pcie/tx.c trans_pcie->txq[txq_id]->frozen = false; txq 1508 drivers/net/wireless/intel/iwlwifi/pcie/tx.c trans_pcie->txq[txq_id]->ampdu = false; txq 1528 drivers/net/wireless/intel/iwlwifi/pcie/tx.c struct iwl_txq *txq = trans_pcie->txq[trans_pcie->cmd_queue]; txq 1625 drivers/net/wireless/intel/iwlwifi/pcie/tx.c spin_lock_bh(&txq->lock); txq 1627 drivers/net/wireless/intel/iwlwifi/pcie/tx.c if (iwl_queue_space(trans, txq) < ((cmd->flags & CMD_ASYNC) ? 2 : 1)) { txq 1628 drivers/net/wireless/intel/iwlwifi/pcie/tx.c spin_unlock_bh(&txq->lock); txq 1636 drivers/net/wireless/intel/iwlwifi/pcie/tx.c idx = iwl_pcie_get_cmd_index(txq, txq->write_ptr); txq 1637 drivers/net/wireless/intel/iwlwifi/pcie/tx.c out_cmd = txq->entries[idx].cmd; txq 1638 drivers/net/wireless/intel/iwlwifi/pcie/tx.c out_meta = &txq->entries[idx].meta; txq 1655 drivers/net/wireless/intel/iwlwifi/pcie/tx.c INDEX_TO_SEQ(txq->write_ptr)); txq 1663 drivers/net/wireless/intel/iwlwifi/pcie/tx.c INDEX_TO_SEQ(txq->write_ptr)); txq 1713 drivers/net/wireless/intel/iwlwifi/pcie/tx.c cmd_size, txq->write_ptr, idx, trans_pcie->cmd_queue); txq 1717 drivers/net/wireless/intel/iwlwifi/pcie/tx.c memcpy(&txq->first_tb_bufs[idx], &out_cmd->hdr, tb0_size); txq 1718 drivers/net/wireless/intel/iwlwifi/pcie/tx.c iwl_pcie_txq_build_tfd(trans, txq, txq 1719 drivers/net/wireless/intel/iwlwifi/pcie/tx.c iwl_pcie_get_first_tb_dma(txq, idx), txq 1729 drivers/net/wireless/intel/iwlwifi/pcie/tx.c iwl_pcie_tfd_unmap(trans, out_meta, txq, txq 1730 drivers/net/wireless/intel/iwlwifi/pcie/tx.c txq->write_ptr); txq 1735 drivers/net/wireless/intel/iwlwifi/pcie/tx.c iwl_pcie_txq_build_tfd(trans, txq, phys_addr, txq 1753 drivers/net/wireless/intel/iwlwifi/pcie/tx.c iwl_pcie_tfd_unmap(trans, out_meta, txq, txq 1754 drivers/net/wireless/intel/iwlwifi/pcie/tx.c txq->write_ptr); txq 1759 drivers/net/wireless/intel/iwlwifi/pcie/tx.c iwl_pcie_txq_build_tfd(trans, txq, phys_addr, cmdlen[i], false); txq 1764 drivers/net/wireless/intel/iwlwifi/pcie/tx.c if (WARN_ON_ONCE(txq->entries[idx].free_buf)) txq 1765 drivers/net/wireless/intel/iwlwifi/pcie/tx.c kzfree(txq->entries[idx].free_buf); txq 1766 drivers/net/wireless/intel/iwlwifi/pcie/tx.c txq->entries[idx].free_buf = dup_buf; txq 1771 drivers/net/wireless/intel/iwlwifi/pcie/tx.c if (txq->read_ptr == txq->write_ptr && txq->wd_timeout) txq 1772 drivers/net/wireless/intel/iwlwifi/pcie/tx.c mod_timer(&txq->stuck_timer, jiffies + txq->wd_timeout); txq 1783 drivers/net/wireless/intel/iwlwifi/pcie/tx.c txq->write_ptr = iwl_queue_inc_wrap(trans, txq->write_ptr); txq 1784 drivers/net/wireless/intel/iwlwifi/pcie/tx.c iwl_pcie_txq_inc_wr_ptr(trans, txq); txq 1789 drivers/net/wireless/intel/iwlwifi/pcie/tx.c spin_unlock_bh(&txq->lock); txq 1813 drivers/net/wireless/intel/iwlwifi/pcie/tx.c struct iwl_txq *txq = trans_pcie->txq[trans_pcie->cmd_queue]; txq 1820 drivers/net/wireless/intel/iwlwifi/pcie/tx.c txq_id, trans_pcie->cmd_queue, sequence, txq->read_ptr, txq 1821 drivers/net/wireless/intel/iwlwifi/pcie/tx.c txq->write_ptr)) { txq 1826 drivers/net/wireless/intel/iwlwifi/pcie/tx.c spin_lock_bh(&txq->lock); txq 1828 drivers/net/wireless/intel/iwlwifi/pcie/tx.c cmd_index = iwl_pcie_get_cmd_index(txq, index); txq 1829 drivers/net/wireless/intel/iwlwifi/pcie/tx.c cmd = txq->entries[cmd_index].cmd; txq 1830 drivers/net/wireless/intel/iwlwifi/pcie/tx.c meta = &txq->entries[cmd_index].meta; txq 1834 drivers/net/wireless/intel/iwlwifi/pcie/tx.c iwl_pcie_tfd_unmap(trans, meta, txq, index); txq 1864 drivers/net/wireless/intel/iwlwifi/pcie/tx.c spin_unlock_bh(&txq->lock); txq 1892 drivers/net/wireless/intel/iwlwifi/pcie/tx.c struct iwl_txq *txq = trans_pcie->txq[trans_pcie->cmd_queue]; txq 1928 drivers/net/wireless/intel/iwlwifi/pcie/tx.c txq->read_ptr, txq->write_ptr); txq 1972 drivers/net/wireless/intel/iwlwifi/pcie/tx.c txq->entries[cmd_idx].meta.flags &= ~CMD_WANT_SKB; txq 2004 drivers/net/wireless/intel/iwlwifi/pcie/tx.c struct iwl_txq *txq, u8 hdr_len, txq 2025 drivers/net/wireless/intel/iwlwifi/pcie/tx.c iwl_pcie_txq_build_tfd(trans, txq, tb_phys, head_tb_len, false); txq 2045 drivers/net/wireless/intel/iwlwifi/pcie/tx.c tb_idx = iwl_pcie_txq_build_tfd(trans, txq, tb_phys, txq 2100 drivers/net/wireless/intel/iwlwifi/pcie/tx.c struct iwl_txq *txq, u8 hdr_len, txq 2106 drivers/net/wireless/intel/iwlwifi/pcie/tx.c struct iwl_trans_pcie *trans_pcie = txq->trans_pcie; txq 2122 drivers/net/wireless/intel/iwlwifi/pcie/tx.c iwl_pcie_get_tfd(trans, txq, txq->write_ptr), txq 2223 drivers/net/wireless/intel/iwlwifi/pcie/tx.c iwl_pcie_txq_build_tfd(trans, txq, hdr_tb_phys, txq 2249 drivers/net/wireless/intel/iwlwifi/pcie/tx.c iwl_pcie_txq_build_tfd(trans, txq, tb_phys, txq 2283 drivers/net/wireless/intel/iwlwifi/pcie/tx.c struct iwl_txq *txq, u8 hdr_len, txq 2302 drivers/net/wireless/intel/iwlwifi/pcie/tx.c struct iwl_txq *txq; txq 2313 drivers/net/wireless/intel/iwlwifi/pcie/tx.c txq = trans_pcie->txq[txq_id]; txq 2346 drivers/net/wireless/intel/iwlwifi/pcie/tx.c spin_lock(&txq->lock); txq 2348 drivers/net/wireless/intel/iwlwifi/pcie/tx.c if (iwl_queue_space(trans, txq) < txq->high_mark) { txq 2349 drivers/net/wireless/intel/iwlwifi/pcie/tx.c iwl_stop_queue(trans, txq); txq 2352 drivers/net/wireless/intel/iwlwifi/pcie/tx.c if (unlikely(iwl_queue_space(trans, txq) < 3)) { txq 2359 drivers/net/wireless/intel/iwlwifi/pcie/tx.c __skb_queue_tail(&txq->overflow_q, skb); txq 2361 drivers/net/wireless/intel/iwlwifi/pcie/tx.c spin_unlock(&txq->lock); txq 2372 drivers/net/wireless/intel/iwlwifi/pcie/tx.c WARN_ONCE(txq->ampdu && txq 2373 drivers/net/wireless/intel/iwlwifi/pcie/tx.c (wifi_seq & 0xff) != txq->write_ptr, txq 2375 drivers/net/wireless/intel/iwlwifi/pcie/tx.c txq_id, wifi_seq, txq->write_ptr); txq 2378 drivers/net/wireless/intel/iwlwifi/pcie/tx.c txq->entries[txq->write_ptr].skb = skb; txq 2379 drivers/net/wireless/intel/iwlwifi/pcie/tx.c txq->entries[txq->write_ptr].cmd = dev_cmd; txq 2383 drivers/net/wireless/intel/iwlwifi/pcie/tx.c INDEX_TO_SEQ(txq->write_ptr))); txq 2385 drivers/net/wireless/intel/iwlwifi/pcie/tx.c tb0_phys = iwl_pcie_get_first_tb_dma(txq, txq->write_ptr); txq 2393 drivers/net/wireless/intel/iwlwifi/pcie/tx.c out_meta = &txq->entries[txq->write_ptr].meta; txq 2421 drivers/net/wireless/intel/iwlwifi/pcie/tx.c iwl_pcie_txq_build_tfd(trans, txq, tb0_phys, txq 2432 drivers/net/wireless/intel/iwlwifi/pcie/tx.c iwl_pcie_txq_build_tfd(trans, txq, tb1_phys, tb1_len, false); txq 2435 drivers/net/wireless/intel/iwlwifi/pcie/tx.c iwl_pcie_get_tfd(trans, txq, txq 2436 drivers/net/wireless/intel/iwlwifi/pcie/tx.c txq->write_ptr), txq 2448 drivers/net/wireless/intel/iwlwifi/pcie/tx.c if (unlikely(iwl_fill_data_tbs_amsdu(trans, skb, txq, hdr_len, txq 2455 drivers/net/wireless/intel/iwlwifi/pcie/tx.c if (unlikely(iwl_fill_data_tbs(trans, skb, txq, hdr_len, txq 2460 drivers/net/wireless/intel/iwlwifi/pcie/tx.c if (unlikely(iwl_fill_data_tbs(trans, frag, txq, 0, txq 2467 drivers/net/wireless/intel/iwlwifi/pcie/tx.c memcpy(&txq->first_tb_bufs[txq->write_ptr], dev_cmd, IWL_FIRST_TB_SIZE); txq 2469 drivers/net/wireless/intel/iwlwifi/pcie/tx.c tfd = iwl_pcie_get_tfd(trans, txq, txq->write_ptr); txq 2471 drivers/net/wireless/intel/iwlwifi/pcie/tx.c iwl_pcie_txq_update_byte_cnt_tbl(trans, txq, le16_to_cpu(tx_cmd->len), txq 2477 drivers/net/wireless/intel/iwlwifi/pcie/tx.c if (txq->read_ptr == txq->write_ptr && txq->wd_timeout) { txq 2484 drivers/net/wireless/intel/iwlwifi/pcie/tx.c if (!txq->frozen) txq 2485 drivers/net/wireless/intel/iwlwifi/pcie/tx.c mod_timer(&txq->stuck_timer, txq 2486 drivers/net/wireless/intel/iwlwifi/pcie/tx.c jiffies + txq->wd_timeout); txq 2488 drivers/net/wireless/intel/iwlwifi/pcie/tx.c txq->frozen_expiry_remainder = txq->wd_timeout; txq 2492 drivers/net/wireless/intel/iwlwifi/pcie/tx.c txq->write_ptr = iwl_queue_inc_wrap(trans, txq->write_ptr); txq 2494 drivers/net/wireless/intel/iwlwifi/pcie/tx.c iwl_pcie_txq_inc_wr_ptr(trans, txq); txq 2500 drivers/net/wireless/intel/iwlwifi/pcie/tx.c spin_unlock(&txq->lock); txq 2503 drivers/net/wireless/intel/iwlwifi/pcie/tx.c iwl_pcie_tfd_unmap(trans, out_meta, txq, txq->write_ptr); txq 2504 drivers/net/wireless/intel/iwlwifi/pcie/tx.c spin_unlock(&txq->lock); txq 81 drivers/net/wireless/marvell/mwifiex/debugfs.c struct netdev_queue *txq; txq 142 drivers/net/wireless/marvell/mwifiex/debugfs.c txq = netdev_get_tx_queue(netdev, i); txq 143 drivers/net/wireless/marvell/mwifiex/debugfs.c p += sprintf(p, " %d:%s", i, netif_tx_queue_stopped(txq) ? txq 791 drivers/net/wireless/marvell/mwifiex/main.c struct netdev_queue *txq; txq 795 drivers/net/wireless/marvell/mwifiex/main.c txq = netdev_get_tx_queue(priv->netdev, index); txq 796 drivers/net/wireless/marvell/mwifiex/main.c if (!netif_tx_queue_stopped(txq)) { txq 797 drivers/net/wireless/marvell/mwifiex/main.c netif_tx_stop_queue(txq); txq 1099 drivers/net/wireless/marvell/mwifiex/main.c struct netdev_queue *txq; txq 1159 drivers/net/wireless/marvell/mwifiex/main.c txq = netdev_get_tx_queue(priv->netdev, idx); txq 1161 drivers/net/wireless/marvell/mwifiex/main.c netif_tx_queue_stopped(txq) ? txq 282 drivers/net/wireless/marvell/mwifiex/txrx.c struct netdev_queue *txq; txq 319 drivers/net/wireless/marvell/mwifiex/txrx.c txq = netdev_get_tx_queue(priv->netdev, index); txq 320 drivers/net/wireless/marvell/mwifiex/txrx.c if (netif_tx_queue_stopped(txq)) { txq 321 drivers/net/wireless/marvell/mwifiex/txrx.c netif_tx_wake_queue(txq); txq 248 drivers/net/wireless/marvell/mwl8k.c struct mwl8k_tx_queue txq[MWL8K_MAX_TX_QUEUES]; txq 1453 drivers/net/wireless/marvell/mwl8k.c struct mwl8k_tx_queue *txq = priv->txq + index; txq 1457 drivers/net/wireless/marvell/mwl8k.c txq->len = 0; txq 1458 drivers/net/wireless/marvell/mwl8k.c txq->head = 0; txq 1459 drivers/net/wireless/marvell/mwl8k.c txq->tail = 0; txq 1463 drivers/net/wireless/marvell/mwl8k.c txq->txd = pci_zalloc_consistent(priv->pdev, size, &txq->txd_dma); txq 1464 drivers/net/wireless/marvell/mwl8k.c if (txq->txd == NULL) { txq 1469 drivers/net/wireless/marvell/mwl8k.c txq->skb = kcalloc(MWL8K_TX_DESCS, sizeof(*txq->skb), GFP_KERNEL); txq 1470 drivers/net/wireless/marvell/mwl8k.c if (txq->skb == NULL) { txq 1471 drivers/net/wireless/marvell/mwl8k.c pci_free_consistent(priv->pdev, size, txq->txd, txq->txd_dma); txq 1479 drivers/net/wireless/marvell/mwl8k.c tx_desc = txq->txd + i; txq 1484 drivers/net/wireless/marvell/mwl8k.c cpu_to_le32(txq->txd_dma + nexti * sizeof(*tx_desc)); txq 1505 drivers/net/wireless/marvell/mwl8k.c struct mwl8k_tx_queue *txq = priv->txq + i; txq 1512 drivers/net/wireless/marvell/mwl8k.c struct mwl8k_tx_desc *tx_desc = txq->txd + desc; txq 1529 drivers/net/wireless/marvell/mwl8k.c txq->len, txq->head, txq->tail, txq 1670 drivers/net/wireless/marvell/mwl8k.c struct mwl8k_tx_queue *txq = priv->txq + index; txq 1674 drivers/net/wireless/marvell/mwl8k.c while (txq->len > 0 && limit--) { txq 1687 drivers/net/wireless/marvell/mwl8k.c tx = txq->head; txq 1688 drivers/net/wireless/marvell/mwl8k.c tx_desc = txq->txd + tx; txq 1699 drivers/net/wireless/marvell/mwl8k.c txq->head = (tx + 1) % MWL8K_TX_DESCS; txq 1700 drivers/net/wireless/marvell/mwl8k.c BUG_ON(txq->len == 0); txq 1701 drivers/net/wireless/marvell/mwl8k.c txq->len--; txq 1706 drivers/net/wireless/marvell/mwl8k.c skb = txq->skb[tx]; txq 1707 drivers/net/wireless/marvell/mwl8k.c txq->skb[tx] = NULL; txq 1767 drivers/net/wireless/marvell/mwl8k.c struct mwl8k_tx_queue *txq = priv->txq + index; txq 1769 drivers/net/wireless/marvell/mwl8k.c if (txq->txd == NULL) txq 1774 drivers/net/wireless/marvell/mwl8k.c kfree(txq->skb); txq 1775 drivers/net/wireless/marvell/mwl8k.c txq->skb = NULL; txq 1779 drivers/net/wireless/marvell/mwl8k.c txq->txd, txq->txd_dma); txq 1780 drivers/net/wireless/marvell/mwl8k.c txq->txd = NULL; txq 1902 drivers/net/wireless/marvell/mwl8k.c struct mwl8k_tx_queue *txq; txq 2061 drivers/net/wireless/marvell/mwl8k.c txq = priv->txq + index; txq 2071 drivers/net/wireless/marvell/mwl8k.c if (txq->len >= MWL8K_TX_DESCS - 2) { txq 2072 drivers/net/wireless/marvell/mwl8k.c if (!mgmtframe || txq->len == MWL8K_TX_DESCS) { txq 2087 drivers/net/wireless/marvell/mwl8k.c BUG_ON(txq->skb[txq->tail] != NULL); txq 2088 drivers/net/wireless/marvell/mwl8k.c txq->skb[txq->tail] = skb; txq 2090 drivers/net/wireless/marvell/mwl8k.c tx = txq->txd + txq->tail; txq 2111 drivers/net/wireless/marvell/mwl8k.c txq->len++; txq 2114 drivers/net/wireless/marvell/mwl8k.c txq->tail++; txq 2115 drivers/net/wireless/marvell/mwl8k.c if (txq->tail == MWL8K_TX_DESCS) txq 2116 drivers/net/wireless/marvell/mwl8k.c txq->tail = 0; txq 2473 drivers/net/wireless/marvell/mwl8k.c cmd->tx_queue_ptrs[i] = cpu_to_le32(priv->txq[i].txd_dma); txq 2647 drivers/net/wireless/marvell/mwl8k.c cmd->tx_queue_ptrs[i] = cpu_to_le32(priv->txq[j].txd_dma); txq 3435 drivers/net/wireless/marvell/mwl8k.c __u8 txq; txq 3448 drivers/net/wireless/marvell/mwl8k.c __u8 txq; txq 3482 drivers/net/wireless/marvell/mwl8k.c cmd->ap.txq = qnum; txq 3487 drivers/net/wireless/marvell/mwl8k.c cmd->sta.txq = qnum; txq 5856 drivers/net/wireless/marvell/mwl8k.c iowrite32(priv->txq[i].txd_dma, txq 613 drivers/net/wireless/mediatek/mt76/mac80211.c for (i = 0; i < ARRAY_SIZE(sta->txq); i++) { txq 616 drivers/net/wireless/mediatek/mt76/mac80211.c if (!sta->txq[i]) txq 619 drivers/net/wireless/mediatek/mt76/mac80211.c mtxq = (struct mt76_txq *)sta->txq[i]->drv_priv; txq 621 drivers/net/wireless/mediatek/mt76/mac80211.c ieee80211_schedule_txq(dev->hw, sta->txq[i]); txq 675 drivers/net/wireless/mediatek/mt76/mac80211.c for (i = 0; i < ARRAY_SIZE(sta->txq); i++) { txq 678 drivers/net/wireless/mediatek/mt76/mac80211.c if (!sta->txq[i]) txq 681 drivers/net/wireless/mediatek/mt76/mac80211.c mtxq = (struct mt76_txq *)sta->txq[i]->drv_priv; txq 684 drivers/net/wireless/mediatek/mt76/mac80211.c mt76_txq_init(dev, sta->txq[i]); txq 712 drivers/net/wireless/mediatek/mt76/mac80211.c for (i = 0; i < ARRAY_SIZE(sta->txq); i++) txq 713 drivers/net/wireless/mediatek/mt76/mac80211.c mt76_txq_remove(dev, sta->txq[i]); txq 706 drivers/net/wireless/mediatek/mt76/mt76.h void mt76_txq_init(struct mt76_dev *dev, struct ieee80211_txq *txq); txq 707 drivers/net/wireless/mediatek/mt76/mt76.h void mt76_txq_remove(struct mt76_dev *dev, struct ieee80211_txq *txq); txq 708 drivers/net/wireless/mediatek/mt76/mt76.h void mt76_wake_tx_queue(struct ieee80211_hw *hw, struct ieee80211_txq *txq); txq 74 drivers/net/wireless/mediatek/mt76/mt7603/main.c mtxq = (struct mt76_txq *)vif->txq->drv_priv; txq 76 drivers/net/wireless/mediatek/mt76/mt7603/main.c mt76_txq_init(&dev->mt76, vif->txq); txq 99 drivers/net/wireless/mediatek/mt76/mt7603/main.c mt76_txq_remove(&dev->mt76, vif->txq); txq 552 drivers/net/wireless/mediatek/mt76/mt7603/main.c struct ieee80211_txq *txq = sta->txq[params->tid]; txq 559 drivers/net/wireless/mediatek/mt76/mt7603/main.c if (!txq) txq 562 drivers/net/wireless/mediatek/mt76/mt7603/main.c mtxq = (struct mt76_txq *)txq->drv_priv; txq 104 drivers/net/wireless/mediatek/mt76/mt7615/main.c mtxq = (struct mt76_txq *)vif->txq->drv_priv; txq 106 drivers/net/wireless/mediatek/mt76/mt7615/main.c mt76_txq_init(&dev->mt76, vif->txq); txq 126 drivers/net/wireless/mediatek/mt76/mt7615/main.c mt76_txq_remove(&dev->mt76, vif->txq); txq 447 drivers/net/wireless/mediatek/mt76/mt7615/main.c struct ieee80211_txq *txq = sta->txq[params->tid]; txq 453 drivers/net/wireless/mediatek/mt76/mt7615/main.c if (!txq) txq 456 drivers/net/wireless/mediatek/mt76/mt7615/main.c mtxq = (struct mt76_txq *)txq->drv_priv; txq 299 drivers/net/wireless/mediatek/mt76/mt76x02_util.c mtxq = (struct mt76_txq *)vif->txq->drv_priv; txq 302 drivers/net/wireless/mediatek/mt76/mt76x02_util.c mt76_txq_init(&dev->mt76, vif->txq); txq 352 drivers/net/wireless/mediatek/mt76/mt76x02_util.c mt76_txq_remove(&dev->mt76, vif->txq); txq 364 drivers/net/wireless/mediatek/mt76/mt76x02_util.c struct ieee80211_txq *txq = sta->txq[params->tid]; txq 369 drivers/net/wireless/mediatek/mt76/mt76x02_util.c if (!txq) txq 372 drivers/net/wireless/mediatek/mt76/mt76x02_util.c mtxq = (struct mt76_txq *)txq->drv_priv; txq 78 drivers/net/wireless/mediatek/mt76/tx.c mt76_txq_get_qid(struct ieee80211_txq *txq) txq 80 drivers/net/wireless/mediatek/mt76/tx.c if (!txq->sta) txq 83 drivers/net/wireless/mediatek/mt76/tx.c return txq->ac; txq 266 drivers/net/wireless/mediatek/mt76/tx.c struct ieee80211_txq *txq; txq 271 drivers/net/wireless/mediatek/mt76/tx.c txq = sta->txq[tid]; txq 272 drivers/net/wireless/mediatek/mt76/tx.c mtxq = (struct mt76_txq *)txq->drv_priv; txq 296 drivers/net/wireless/mediatek/mt76/tx.c struct ieee80211_txq *txq = mtxq_to_txq(mtxq); txq 304 drivers/net/wireless/mediatek/mt76/tx.c ieee80211_sta_set_buffered(txq->sta, tid, false); txq 309 drivers/net/wireless/mediatek/mt76/tx.c skb = ieee80211_tx_dequeue(dev->hw, txq); txq 345 drivers/net/wireless/mediatek/mt76/tx.c struct ieee80211_txq *txq = sta->txq[i]; txq 346 drivers/net/wireless/mediatek/mt76/tx.c struct mt76_txq *mtxq = (struct mt76_txq *)txq->drv_priv; txq 383 drivers/net/wireless/mediatek/mt76/tx.c struct ieee80211_txq *txq = mtxq_to_txq(mtxq); txq 384 drivers/net/wireless/mediatek/mt76/tx.c enum mt76_txq_id qid = mt76_txq_get_qid(txq); txq 408 drivers/net/wireless/mediatek/mt76/tx.c ieee80211_get_tx_rates(txq->vif, txq->sta, skb, txq 419 drivers/net/wireless/mediatek/mt76/tx.c idx = dev->queue_ops->tx_queue_skb(dev, qid, skb, wcid, txq->sta); txq 454 drivers/net/wireless/mediatek/mt76/tx.c txq->sta); txq 477 drivers/net/wireless/mediatek/mt76/tx.c struct ieee80211_txq *txq; txq 494 drivers/net/wireless/mediatek/mt76/tx.c txq = ieee80211_next_txq(dev->hw, qid); txq 495 drivers/net/wireless/mediatek/mt76/tx.c if (!txq) txq 498 drivers/net/wireless/mediatek/mt76/tx.c mtxq = (struct mt76_txq *)txq->drv_priv; txq 504 drivers/net/wireless/mediatek/mt76/tx.c struct ieee80211_txq *txq = mtxq_to_txq(mtxq); txq 505 drivers/net/wireless/mediatek/mt76/tx.c struct ieee80211_sta *sta = txq->sta; txq 506 drivers/net/wireless/mediatek/mt76/tx.c struct ieee80211_vif *vif = txq->vif; txq 508 drivers/net/wireless/mediatek/mt76/tx.c u8 tid = txq->tid; txq 519 drivers/net/wireless/mediatek/mt76/tx.c ieee80211_return_txq(dev->hw, txq, !empty); txq 570 drivers/net/wireless/mediatek/mt76/tx.c for (i = 0; i < ARRAY_SIZE(sta->txq); i++) { txq 571 drivers/net/wireless/mediatek/mt76/tx.c struct ieee80211_txq *txq = sta->txq[i]; txq 575 drivers/net/wireless/mediatek/mt76/tx.c if (!txq) txq 578 drivers/net/wireless/mediatek/mt76/tx.c mtxq = (struct mt76_txq *)txq->drv_priv; txq 588 drivers/net/wireless/mediatek/mt76/tx.c void mt76_wake_tx_queue(struct ieee80211_hw *hw, struct ieee80211_txq *txq) txq 599 drivers/net/wireless/mediatek/mt76/tx.c void mt76_txq_remove(struct mt76_dev *dev, struct ieee80211_txq *txq) txq 604 drivers/net/wireless/mediatek/mt76/tx.c if (!txq) txq 607 drivers/net/wireless/mediatek/mt76/tx.c mtxq = (struct mt76_txq *)txq->drv_priv; txq 614 drivers/net/wireless/mediatek/mt76/tx.c void mt76_txq_init(struct mt76_dev *dev, struct ieee80211_txq *txq) txq 616 drivers/net/wireless/mediatek/mt76/tx.c struct mt76_txq *mtxq = (struct mt76_txq *)txq->drv_priv; txq 620 drivers/net/wireless/mediatek/mt76/tx.c mtxq->swq = &dev->q_tx[mt76_txq_get_qid(txq)]; txq 98 drivers/net/xen-netback/rx.c struct netdev_queue *txq; txq 100 drivers/net/xen-netback/rx.c txq = netdev_get_tx_queue(queue->vif->dev, queue->id); txq 101 drivers/net/xen-netback/rx.c netif_tx_wake_queue(txq); txq 874 drivers/s390/net/qeth_core.h static inline u16 qeth_iqd_translate_txq(struct net_device *dev, u16 txq) txq 876 drivers/s390/net/qeth_core.h if (txq == QETH_IQD_MCAST_TXQ) txq 878 drivers/s390/net/qeth_core.h if (txq == dev->num_tx_queues - 1) txq 880 drivers/s390/net/qeth_core.h return txq; txq 3499 drivers/s390/net/qeth_core_main.c struct netdev_queue *txq; txq 3520 drivers/s390/net/qeth_core_main.c txq = netdev_get_tx_queue(dev, __queue); txq 3525 drivers/s390/net/qeth_core_main.c if (netif_tx_queue_stopped(txq) && !qeth_out_queue_is_full(queue)) txq 3526 drivers/s390/net/qeth_core_main.c netif_tx_wake_queue(txq); txq 3835 drivers/s390/net/qeth_core_main.c struct netdev_queue *txq; txq 3839 drivers/s390/net/qeth_core_main.c txq = netdev_get_tx_queue(card->dev, skb_get_queue_mapping(skb)); txq 3864 drivers/s390/net/qeth_core_main.c netif_tx_stop_queue(txq); txq 3872 drivers/s390/net/qeth_core_main.c flush = __netdev_tx_sent_queue(txq, bytes, txq 3881 drivers/s390/net/qeth_core_main.c netif_tx_start_queue(txq); txq 3892 drivers/s390/net/qeth_core_main.c struct netdev_queue *txq; txq 3914 drivers/s390/net/qeth_core_main.c txq = netdev_get_tx_queue(card->dev, skb_get_queue_mapping(skb)); txq 3950 drivers/s390/net/qeth_core_main.c netif_tx_stop_queue(txq); txq 3996 drivers/s390/net/qeth_core_main.c netif_tx_start_queue(txq); txq 5257 drivers/s390/net/qeth_core_main.c struct netdev_queue *txq; txq 5259 drivers/s390/net/qeth_core_main.c txq = netdev_get_tx_queue(dev, qeth_iqd_translate_txq(dev, queue_no)); txq 5302 drivers/s390/net/qeth_core_main.c netdev_tx_completed_queue(txq, packets, bytes); txq 5311 drivers/s390/net/qeth_core_main.c if (netif_tx_queue_stopped(txq) && txq 5313 drivers/s390/net/qeth_core_main.c netif_tx_wake_queue(txq); txq 592 drivers/s390/net/qeth_l2_main.c u16 txq = skb_get_queue_mapping(skb); txq 599 drivers/s390/net/qeth_l2_main.c txq = qeth_iqd_translate_txq(dev, txq); txq 600 drivers/s390/net/qeth_l2_main.c queue = card->qdio.out_qs[txq]; txq 2044 drivers/s390/net/qeth_l3_main.c u16 txq = skb_get_queue_mapping(skb); txq 2052 drivers/s390/net/qeth_l3_main.c queue = card->qdio.out_qs[qeth_iqd_translate_txq(dev, txq)]; txq 2061 drivers/s390/net/qeth_l3_main.c queue = card->qdio.out_qs[txq]; txq 5926 drivers/scsi/lpfc/lpfc_bsg.c list_for_each_entry_safe(check_iocb, next_iocb, &pring->txq, txq 5963 drivers/scsi/lpfc/lpfc_bsg.c list_for_each_entry_safe(check_iocb, next_iocb, &pring->txq, txq 8083 drivers/scsi/lpfc/lpfc_els.c list_for_each_entry_safe(piocb, tmp_iocb, &pring->txq, list) { txq 9973 drivers/scsi/lpfc/lpfc_els.c if (pring && !list_empty(&pring->txq)) txq 4752 drivers/scsi/lpfc/lpfc_hbadisc.c list_for_each_entry_safe(iocb, next_iocb, &pring->txq, list) { txq 5710 drivers/scsi/lpfc/lpfc_hbadisc.c list_for_each_entry_safe(iocb, next_iocb, &pring->txq, list) { txq 263 drivers/scsi/lpfc/lpfc_nportdisc.c list_for_each_entry_safe(iocb, next_iocb, &pring->txq, list) { txq 571 drivers/scsi/lpfc/lpfc_scsi.c if (!list_empty(&pring->txq)) txq 1292 drivers/scsi/lpfc/lpfc_sli.c if (!list_empty(&pring->txq)) txq 1625 drivers/scsi/lpfc/lpfc_sli.c list_remove_head((&pring->txq), cmd_iocb, struct lpfc_iocbq, list); txq 1902 drivers/scsi/lpfc/lpfc_sli.c (!list_empty(&pring->txq)) && txq 3941 drivers/scsi/lpfc/lpfc_sli.c list_splice_init(&pring->txq, &completions); txq 3952 drivers/scsi/lpfc/lpfc_sli.c list_splice_init(&pring->txq, &completions); txq 4008 drivers/scsi/lpfc/lpfc_sli.c LIST_HEAD(txq); txq 4032 drivers/scsi/lpfc/lpfc_sli.c list_splice_init(&pring->txq, &txq); txq 4043 drivers/scsi/lpfc/lpfc_sli.c lpfc_sli_cancel_iocbs(phba, &txq, txq 4056 drivers/scsi/lpfc/lpfc_sli.c list_splice_init(&pring->txq, &txq); txq 4067 drivers/scsi/lpfc/lpfc_sli.c lpfc_sli_cancel_iocbs(phba, &txq, IOSTAT_LOCAL_REJECT, txq 8987 drivers/scsi/lpfc/lpfc_sli.c list_add_tail(&piocb->list, &pring->txq); txq 9911 drivers/scsi/lpfc/lpfc_sli.c if (!list_empty(&pring->txq)) { txq 10501 drivers/scsi/lpfc/lpfc_sli.c INIT_LIST_HEAD(&pring->txq); txq 10510 drivers/scsi/lpfc/lpfc_sli.c INIT_LIST_HEAD(&pring->txq); txq 10520 drivers/scsi/lpfc/lpfc_sli.c INIT_LIST_HEAD(&pring->txq); txq 10562 drivers/scsi/lpfc/lpfc_sli.c INIT_LIST_HEAD(&pring->txq); txq 10672 drivers/scsi/lpfc/lpfc_sli.c &pring->txq, list) { txq 10698 drivers/scsi/lpfc/lpfc_sli.c &pring->txq, list) { txq 10772 drivers/scsi/lpfc/lpfc_sli.c list_splice_init(&pring->txq, &completions); txq 10780 drivers/scsi/lpfc/lpfc_sli.c list_splice_init(&pring->txq, &completions); txq 11912 drivers/scsi/lpfc/lpfc_sli.c list_for_each_entry(iocb, &pring->txq, list) { txq 13257 drivers/scsi/lpfc/lpfc_sli.c if (!list_empty(&pring->txq)) txq 19622 drivers/scsi/lpfc/lpfc_sli.c if (unlikely(!pring) || list_empty(&pring->txq)) txq 19626 drivers/scsi/lpfc/lpfc_sli.c list_for_each_entry(piocbq, &pring->txq, list) { txq 19635 drivers/scsi/lpfc/lpfc_sli.c while (!list_empty(&pring->txq)) { txq 235 drivers/scsi/lpfc/lpfc_sli.h struct list_head txq; txq 197 drivers/target/iscsi/cxgbit/cxgbit.h struct sk_buff_head txq; txq 635 drivers/target/iscsi/cxgbit/cxgbit_cm.c __skb_queue_tail(&csk->txq, skb); txq 666 drivers/target/iscsi/cxgbit/cxgbit_cm.c __skb_queue_purge(&csk->txq); txq 781 drivers/target/iscsi/cxgbit/cxgbit_cm.c __skb_queue_purge(&csk->txq); txq 1356 drivers/target/iscsi/cxgbit/cxgbit_cm.c skb_queue_head_init(&csk->txq); txq 1751 drivers/target/iscsi/cxgbit/cxgbit_cm.c __skb_queue_purge(&csk->txq); txq 1878 drivers/target/iscsi/cxgbit/cxgbit_cm.c if (skb_queue_len(&csk->txq)) txq 194 drivers/target/iscsi/cxgbit/cxgbit_target.c FW_OFLD_TX_DATA_WR_SHOVE_V(skb_peek(&csk->txq) ? 0 : 1)); txq 206 drivers/target/iscsi/cxgbit/cxgbit_target.c while (csk->wr_cred && ((skb = skb_peek(&csk->txq)) != NULL)) { txq 244 drivers/target/iscsi/cxgbit/cxgbit_target.c __skb_unlink(skb, &csk->txq); txq 347 drivers/target/iscsi/cxgbit/cxgbit_target.c skb_queue_splice_tail_init(&csk->ppodq, &csk->txq); txq 348 drivers/target/iscsi/cxgbit/cxgbit_target.c __skb_queue_tail(&csk->txq, skb); txq 2062 drivers/tty/n_gsm.c struct gsm_msg *txq, *ntxq; txq 2090 drivers/tty/n_gsm.c list_for_each_entry_safe(txq, ntxq, &gsm->tx_list, list) txq 2091 drivers/tty/n_gsm.c kfree(txq); txq 322 include/linux/avf/virtchnl.h struct virtchnl_txq_info txq; txq 2097 include/linux/netdevice.h int netdev_txq_to_tc(struct net_device *dev, unsigned int txq); txq 3064 include/linux/netdevice.h void netif_schedule_queue(struct netdev_queue *txq); txq 3095 include/linux/netdevice.h struct netdev_queue *txq = netdev_get_tx_queue(dev, i); txq 3096 include/linux/netdevice.h netif_tx_start_queue(txq); txq 3119 include/linux/netdevice.h struct netdev_queue *txq = netdev_get_tx_queue(dev, i); txq 3120 include/linux/netdevice.h netif_tx_wake_queue(txq); txq 3378 include/linux/netdevice.h struct netdev_queue *txq = netdev_get_tx_queue(dev, queue_index); txq 3380 include/linux/netdevice.h netif_tx_start_queue(txq); txq 3392 include/linux/netdevice.h struct netdev_queue *txq = netdev_get_tx_queue(dev, queue_index); txq 3393 include/linux/netdevice.h netif_tx_stop_queue(txq); txq 3406 include/linux/netdevice.h struct netdev_queue *txq = netdev_get_tx_queue(dev, queue_index); txq 3408 include/linux/netdevice.h return netif_tx_queue_stopped(txq); txq 3426 include/linux/netdevice.h struct netdev_queue *txq = netdev_get_tx_queue(dev, queue_index); txq 3428 include/linux/netdevice.h netif_tx_wake_queue(txq); txq 3547 include/linux/netdevice.h int netif_set_real_num_tx_queues(struct net_device *dev, unsigned int txq); txq 3694 include/linux/netdevice.h struct netdev_queue *txq, int *ret); txq 3904 include/linux/netdevice.h static inline void __netif_tx_lock(struct netdev_queue *txq, int cpu) txq 3906 include/linux/netdevice.h spin_lock(&txq->_xmit_lock); txq 3907 include/linux/netdevice.h txq->xmit_lock_owner = cpu; txq 3910 include/linux/netdevice.h static inline bool __netif_tx_acquire(struct netdev_queue *txq) txq 3912 include/linux/netdevice.h __acquire(&txq->_xmit_lock); txq 3916 include/linux/netdevice.h static inline void __netif_tx_release(struct netdev_queue *txq) txq 3918 include/linux/netdevice.h __release(&txq->_xmit_lock); txq 3921 include/linux/netdevice.h static inline void __netif_tx_lock_bh(struct netdev_queue *txq) txq 3923 include/linux/netdevice.h spin_lock_bh(&txq->_xmit_lock); txq 3924 include/linux/netdevice.h txq->xmit_lock_owner = smp_processor_id(); txq 3927 include/linux/netdevice.h static inline bool __netif_tx_trylock(struct netdev_queue *txq) txq 3929 include/linux/netdevice.h bool ok = spin_trylock(&txq->_xmit_lock); txq 3931 include/linux/netdevice.h txq->xmit_lock_owner = smp_processor_id(); txq 3935 include/linux/netdevice.h static inline void __netif_tx_unlock(struct netdev_queue *txq) txq 3937 include/linux/netdevice.h txq->xmit_lock_owner = -1; txq 3938 include/linux/netdevice.h spin_unlock(&txq->_xmit_lock); txq 3941 include/linux/netdevice.h static inline void __netif_tx_unlock_bh(struct netdev_queue *txq) txq 3943 include/linux/netdevice.h txq->xmit_lock_owner = -1; txq 3944 include/linux/netdevice.h spin_unlock_bh(&txq->_xmit_lock); txq 3947 include/linux/netdevice.h static inline void txq_trans_update(struct netdev_queue *txq) txq 3949 include/linux/netdevice.h if (txq->xmit_lock_owner != -1) txq 3950 include/linux/netdevice.h txq->trans_start = jiffies; txq 3956 include/linux/netdevice.h struct netdev_queue *txq = netdev_get_tx_queue(dev, 0); txq 3958 include/linux/netdevice.h if (txq->trans_start != jiffies) txq 3959 include/linux/netdevice.h txq->trans_start = jiffies; txq 3976 include/linux/netdevice.h struct netdev_queue *txq = netdev_get_tx_queue(dev, i); txq 3984 include/linux/netdevice.h __netif_tx_lock(txq, cpu); txq 3985 include/linux/netdevice.h set_bit(__QUEUE_STATE_FROZEN, &txq->state); txq 3986 include/linux/netdevice.h __netif_tx_unlock(txq); txq 4001 include/linux/netdevice.h struct netdev_queue *txq = netdev_get_tx_queue(dev, i); txq 4007 include/linux/netdevice.h clear_bit(__QUEUE_STATE_FROZEN, &txq->state); txq 4008 include/linux/netdevice.h netif_schedule_queue(txq); txq 4019 include/linux/netdevice.h #define HARD_TX_LOCK(dev, txq, cpu) { \ txq 4021 include/linux/netdevice.h __netif_tx_lock(txq, cpu); \ txq 4023 include/linux/netdevice.h __netif_tx_acquire(txq); \ txq 4027 include/linux/netdevice.h #define HARD_TX_TRYLOCK(dev, txq) \ txq 4029 include/linux/netdevice.h __netif_tx_trylock(txq) : \ txq 4030 include/linux/netdevice.h __netif_tx_acquire(txq)) txq 4032 include/linux/netdevice.h #define HARD_TX_UNLOCK(dev, txq) { \ txq 4034 include/linux/netdevice.h __netif_tx_unlock(txq); \ txq 4036 include/linux/netdevice.h __netif_tx_release(txq); \ txq 4048 include/linux/netdevice.h struct netdev_queue *txq = netdev_get_tx_queue(dev, i); txq 4050 include/linux/netdevice.h __netif_tx_lock(txq, cpu); txq 4051 include/linux/netdevice.h netif_tx_stop_queue(txq); txq 4052 include/linux/netdevice.h __netif_tx_unlock(txq); txq 4428 include/linux/netdevice.h struct netdev_queue *txq, bool more) txq 4435 include/linux/netdevice.h txq_trans_update(txq); txq 41 include/linux/netpoll.h struct sk_buff_head txq; txq 142 include/linux/ti_wilink_st.h struct sk_buff_head txq, tx_waitq; txq 59 include/linux/usb/usbnet.h struct sk_buff_head txq; txq 1620 include/net/mac80211.h struct ieee80211_txq *txq; txq 2014 include/net/mac80211.h struct ieee80211_txq *txq[IEEE80211_NUM_TIDS + 1]; txq 4029 include/net/mac80211.h struct ieee80211_txq *txq); txq 6256 include/net/mac80211.h struct ieee80211_txq *txq); txq 6270 include/net/mac80211.h struct ieee80211_txq *txq) txq 6275 include/net/mac80211.h skb = ieee80211_tx_dequeue(hw, txq); txq 6310 include/net/mac80211.h struct ieee80211_txq *txq, bool force); txq 6325 include/net/mac80211.h ieee80211_schedule_txq(struct ieee80211_hw *hw, struct ieee80211_txq *txq) txq 6327 include/net/mac80211.h __ieee80211_schedule_txq(hw, txq, true); txq 6341 include/net/mac80211.h ieee80211_return_txq(struct ieee80211_hw *hw, struct ieee80211_txq *txq, txq 6344 include/net/mac80211.h __ieee80211_schedule_txq(hw, txq, force); txq 6372 include/net/mac80211.h struct ieee80211_txq *txq); txq 6385 include/net/mac80211.h void ieee80211_txq_get_depth(struct ieee80211_txq *txq, txq 113 include/net/pkt_sched.h struct net_device *dev, struct netdev_queue *txq, txq 185 include/net/sch_generic.h static inline int qdisc_avail_bulklimit(const struct netdev_queue *txq) txq 189 include/net/sch_generic.h return dql_avail(&txq->dql); txq 726 include/net/sch_generic.h struct netdev_queue *txq = netdev_get_tx_queue(dev, i); txq 727 include/net/sch_generic.h const struct Qdisc *q = rcu_dereference(txq->qdisc); txq 744 include/net/sch_generic.h struct netdev_queue *txq = netdev_get_tx_queue(dev, i); txq 745 include/net/sch_generic.h if (rcu_access_pointer(txq->qdisc) != txq->qdisc_sleeping) txq 757 include/net/sch_generic.h struct netdev_queue *txq = netdev_get_tx_queue(dev, i); txq 758 include/net/sch_generic.h if (rcu_access_pointer(txq->qdisc) != &noop_qdisc) txq 14 include/trace/events/qdisc.h TP_PROTO(struct Qdisc *qdisc, const struct netdev_queue *txq, txq 17 include/trace/events/qdisc.h TP_ARGS(qdisc, txq, packets, skb), txq 21 include/trace/events/qdisc.h __field(const struct netdev_queue *, txq ) txq 33 include/trace/events/qdisc.h __entry->txq = txq; txq 36 include/trace/events/qdisc.h __entry->ifindex = txq->dev ? txq->dev->ifindex : 0; txq 39 include/trace/events/qdisc.h __entry->txq_state = txq->state; txq 172 net/caif/caif_dev.c struct netdev_queue *txq; txq 192 net/caif/caif_dev.c txq = netdev_get_tx_queue(skb->dev, 0); txq 193 net/caif/caif_dev.c sch = rcu_dereference_bh(txq->qdisc); txq 2016 net/core/dev.c static void netif_setup_tc(struct net_device *dev, unsigned int txq) txq 2022 net/core/dev.c if (tc->offset + tc->count > txq) { txq 2033 net/core/dev.c if (tc->offset + tc->count > txq) { txq 2041 net/core/dev.c int netdev_txq_to_tc(struct net_device *dev, unsigned int txq) txq 2049 net/core/dev.c if ((txq - tc->offset) < tc->count) txq 2454 net/core/dev.c struct netdev_queue *txq = &dev->_tx[dev->num_tx_queues]; txq 2457 net/core/dev.c while (txq-- != &dev->_tx[0]) { txq 2458 net/core/dev.c if (txq->sb_dev) txq 2459 net/core/dev.c netdev_unbind_sb_channel(dev, txq->sb_dev); txq 2509 net/core/dev.c struct netdev_queue *txq = &dev->_tx[dev->num_tx_queues]; txq 2517 net/core/dev.c while (txq-- != &dev->_tx[0]) { txq 2518 net/core/dev.c if (txq->sb_dev == sb_dev) txq 2519 net/core/dev.c txq->sb_dev = NULL; txq 2574 net/core/dev.c int netif_set_real_num_tx_queues(struct net_device *dev, unsigned int txq) txq 2579 net/core/dev.c disabling = txq < dev->real_num_tx_queues; txq 2581 net/core/dev.c if (txq < 1 || txq > dev->num_tx_queues) txq 2589 net/core/dev.c txq); txq 2594 net/core/dev.c netif_setup_tc(dev, txq); txq 2596 net/core/dev.c dev->real_num_tx_queues = txq; txq 2600 net/core/dev.c qdisc_reset_all_tx_gt(dev, txq); txq 2602 net/core/dev.c netif_reset_xps_queues_gt(dev, txq); txq 2606 net/core/dev.c dev->real_num_tx_queues = txq; txq 2689 net/core/dev.c void netif_schedule_queue(struct netdev_queue *txq) txq 2692 net/core/dev.c if (!(txq->state & QUEUE_STATE_ANY_XOFF)) { txq 2693 net/core/dev.c struct Qdisc *q = rcu_dereference(txq->qdisc); txq 3190 net/core/dev.c struct netdev_queue *txq, bool more) txq 3200 net/core/dev.c rc = netdev_start_xmit(skb, dev, txq, more); txq 3207 net/core/dev.c struct netdev_queue *txq, int *ret) txq 3216 net/core/dev.c rc = xmit_one(skb, dev, txq, next != NULL); txq 3223 net/core/dev.c if (netif_tx_queue_stopped(txq) && skb) { txq 3380 net/core/dev.c struct netdev_queue *txq) txq 3422 net/core/dev.c if (sch_direct_xmit(skb, q, dev, txq, root_lock, true)) { txq 3703 net/core/dev.c struct netdev_queue *txq; txq 3739 net/core/dev.c txq = netdev_core_pick_tx(dev, skb, sb_dev); txq 3740 net/core/dev.c q = rcu_dereference_bh(txq->qdisc); txq 3744 net/core/dev.c rc = __dev_xmit_skb(skb, q, dev, txq); txq 3763 net/core/dev.c if (txq->xmit_lock_owner != cpu) { txq 3771 net/core/dev.c HARD_TX_LOCK(dev, txq, cpu); txq 3773 net/core/dev.c if (!netif_xmit_stopped(txq)) { txq 3775 net/core/dev.c skb = dev_hard_start_xmit(skb, dev, txq, &rc); txq 3778 net/core/dev.c HARD_TX_UNLOCK(dev, txq); txq 3782 net/core/dev.c HARD_TX_UNLOCK(dev, txq); txq 3822 net/core/dev.c struct netdev_queue *txq; txq 3835 net/core/dev.c txq = skb_get_tx_queue(dev, skb); txq 3839 net/core/dev.c HARD_TX_LOCK(dev, txq, smp_processor_id()); txq 3840 net/core/dev.c if (!netif_xmit_frozen_or_drv_stopped(txq)) txq 3841 net/core/dev.c ret = netdev_start_xmit(skb, dev, txq, false); txq 3842 net/core/dev.c HARD_TX_UNLOCK(dev, txq); txq 4347 net/core/dev.c struct netdev_queue *txq; txq 4351 net/core/dev.c txq = netdev_core_pick_tx(dev, skb, NULL); txq 4353 net/core/dev.c HARD_TX_LOCK(dev, txq, cpu); txq 4354 net/core/dev.c if (!netif_xmit_stopped(txq)) { txq 4355 net/core/dev.c rc = netdev_start_xmit(skb, dev, txq, 0); txq 4359 net/core/dev.c HARD_TX_UNLOCK(dev, txq); txq 8939 net/core/dev.c struct netdev_queue *txq = netdev_get_tx_queue(dev, i); txq 8941 net/core/dev.c netif_tx_stop_queue(txq); txq 1525 net/core/net-sysfs.c int error = 0, txq = 0, rxq = 0, real_rx = 0, real_tx = 0; txq 1544 net/core/net-sysfs.c txq = real_tx; txq 1549 net/core/net-sysfs.c netdev_queue_update_kobjects(dev, txq, 0); txq 73 net/core/netpoll.c struct netdev_queue *txq) txq 92 net/core/netpoll.c status = netdev_start_xmit(skb, dev, txq, false); txq 105 net/core/netpoll.c while ((skb = skb_dequeue(&npinfo->txq))) { txq 107 net/core/netpoll.c struct netdev_queue *txq; txq 122 net/core/netpoll.c txq = netdev_get_tx_queue(dev, q_index); txq 123 net/core/netpoll.c HARD_TX_LOCK(dev, txq, smp_processor_id()); txq 124 net/core/netpoll.c if (netif_xmit_frozen_or_stopped(txq) || txq 125 net/core/netpoll.c !dev_xmit_complete(netpoll_start_xmit(skb, dev, txq))) { txq 126 net/core/netpoll.c skb_queue_head(&npinfo->txq, skb); txq 127 net/core/netpoll.c HARD_TX_UNLOCK(dev, txq); txq 133 net/core/netpoll.c HARD_TX_UNLOCK(dev, txq); txq 324 net/core/netpoll.c if (skb_queue_len(&npinfo->txq) == 0 && !netpoll_owner_active(dev)) { txq 325 net/core/netpoll.c struct netdev_queue *txq; txq 327 net/core/netpoll.c txq = netdev_core_pick_tx(dev, skb, NULL); txq 332 net/core/netpoll.c if (HARD_TX_TRYLOCK(dev, txq)) { txq 333 net/core/netpoll.c if (!netif_xmit_stopped(txq)) txq 334 net/core/netpoll.c status = netpoll_start_xmit(skb, dev, txq); txq 336 net/core/netpoll.c HARD_TX_UNLOCK(dev, txq); txq 356 net/core/netpoll.c skb_queue_tail(&npinfo->txq, skb); txq 608 net/core/netpoll.c skb_queue_head_init(&npinfo->txq); txq 781 net/core/netpoll.c skb_queue_purge(&npinfo->txq); txq 787 net/core/netpoll.c __skb_queue_purge(&npinfo->txq); txq 3303 net/core/pktgen.c struct netdev_queue *txq; txq 3400 net/core/pktgen.c txq = skb_get_tx_queue(odev, pkt_dev->skb); txq 3404 net/core/pktgen.c HARD_TX_LOCK(odev, txq, smp_processor_id()); txq 3406 net/core/pktgen.c if (unlikely(netif_xmit_frozen_or_drv_stopped(txq))) { txq 3414 net/core/pktgen.c ret = netdev_start_xmit(pkt_dev->skb, odev, txq, --burst > 0); txq 3422 net/core/pktgen.c if (burst > 0 && !netif_xmit_frozen_or_drv_stopped(txq)) txq 3443 net/core/pktgen.c HARD_TX_UNLOCK(odev, txq); txq 192 net/mac80211/agg-tx.c struct ieee80211_txq *txq = sta->sta.txq[tid]; txq 197 net/mac80211/agg-tx.c if (!txq) txq 200 net/mac80211/agg-tx.c txqi = to_txq_info(txq); txq 201 net/mac80211/agg-tx.c sdata = vif_to_sdata(txq->vif); txq 213 net/mac80211/agg-tx.c struct ieee80211_txq *txq = sta->sta.txq[tid]; txq 216 net/mac80211/agg-tx.c if (!txq) txq 219 net/mac80211/agg-tx.c txqi = to_txq_info(txq); txq 918 net/mac80211/agg-tx.c struct ieee80211_txq *txq; txq 928 net/mac80211/agg-tx.c txq = sta->sta.txq[tid]; txq 929 net/mac80211/agg-tx.c if (!amsdu && txq) txq 930 net/mac80211/agg-tx.c set_bit(IEEE80211_TXQ_NO_AMSDU, &to_txq_info(txq)->flags); txq 3908 net/mac80211/cfg.c if (!sdata->vif.txq) { txq 3912 net/mac80211/cfg.c ieee80211_fill_txq_stats(txqstats, to_txq_info(sdata->vif.txq)); txq 493 net/mac80211/debugfs_netdev.c if (!sdata->vif.txq) txq 496 net/mac80211/debugfs_netdev.c txqi = to_txq_info(sdata->vif.txq); txq 505 net/mac80211/debugfs_netdev.c txqi->txq.ac, txq 164 net/mac80211/debugfs_sta.c for (i = 0; i < ARRAY_SIZE(sta->sta.txq); i++) { txq 165 net/mac80211/debugfs_sta.c if (!sta->sta.txq[i]) txq 167 net/mac80211/debugfs_sta.c txqi = to_txq_info(sta->sta.txq[i]); txq 170 net/mac80211/debugfs_sta.c txqi->txq.tid, txq 171 net/mac80211/debugfs_sta.c txqi->txq.ac, txq 1201 net/mac80211/driver-ops.h struct txq_info *txq) txq 1203 net/mac80211/driver-ops.h struct ieee80211_sub_if_data *sdata = vif_to_sdata(txq->txq.vif); txq 1211 net/mac80211/driver-ops.h trace_drv_wake_tx_queue(local, sdata, txq); txq 1212 net/mac80211/driver-ops.h local->ops->wake_tx_queue(&local->hw, &txq->txq); txq 1218 net/mac80211/driver-ops.h ieee80211_schedule_txq(&local->hw, &txqi->txq); txq 851 net/mac80211/ieee80211_i.h struct ieee80211_txq txq; txq 1541 net/mac80211/ieee80211_i.h static inline struct txq_info *to_txq_info(struct ieee80211_txq *txq) txq 1543 net/mac80211/ieee80211_i.h return container_of(txq, struct txq_info, txq); txq 1546 net/mac80211/ieee80211_i.h static inline bool txq_has_queue(struct ieee80211_txq *txq) txq 1548 net/mac80211/ieee80211_i.h struct txq_info *txqi = to_txq_info(txq); txq 2088 net/mac80211/ieee80211_i.h struct txq_info *txq, int tid); txq 1906 net/mac80211/iface.c if (sdata->vif.txq) txq 1907 net/mac80211/iface.c ieee80211_txq_purge(sdata->local, to_txq_info(sdata->vif.txq)); txq 1583 net/mac80211/rx.c if (!sta->sta.txq[0]) txq 1587 net/mac80211/rx.c struct ieee80211_txq *txq = sta->sta.txq[tid]; txq 1588 net/mac80211/rx.c struct txq_info *txqi = to_txq_info(txq); txq 1590 net/mac80211/rx.c spin_lock(&local->active_txq_lock[txq->ac]); txq 1593 net/mac80211/rx.c spin_unlock(&local->active_txq_lock[txq->ac]); txq 1595 net/mac80211/rx.c if (txq_has_queue(txq)) txq 110 net/mac80211/sta_info.c if (sta->sta.txq[0]) { txq 111 net/mac80211/sta_info.c for (i = 0; i < ARRAY_SIZE(sta->sta.txq); i++) { txq 114 net/mac80211/sta_info.c if (!sta->sta.txq[i]) txq 117 net/mac80211/sta_info.c txqi = to_txq_info(sta->sta.txq[i]); txq 252 net/mac80211/sta_info.c if (sta->sta.txq[0]) txq 253 net/mac80211/sta_info.c kfree(to_txq_info(sta->sta.txq[0])); txq 379 net/mac80211/sta_info.c txq_data = kcalloc(ARRAY_SIZE(sta->sta.txq), size, gfp); txq 383 net/mac80211/sta_info.c for (i = 0; i < ARRAY_SIZE(sta->sta.txq); i++) { txq 384 net/mac80211/sta_info.c struct txq_info *txq = txq_data + i * size; txq 387 net/mac80211/sta_info.c ieee80211_txq_init(sdata, sta, txq, i); txq 489 net/mac80211/sta_info.c if (sta->sta.txq[0]) txq 490 net/mac80211/sta_info.c kfree(to_txq_info(sta->sta.txq[0])); txq 1300 net/mac80211/sta_info.c for (i = 0; i < ARRAY_SIZE(sta->sta.txq); i++) { txq 1301 net/mac80211/sta_info.c if (!sta->sta.txq[i] || !txq_has_queue(sta->sta.txq[i])) txq 1304 net/mac80211/sta_info.c schedule_and_wake_txq(local, to_txq_info(sta->sta.txq[i])); txq 1738 net/mac80211/sta_info.c if (!sta->sta.txq[0]) txq 1741 net/mac80211/sta_info.c for (tid = 0; tid < ARRAY_SIZE(sta->sta.txq); tid++) { txq 1742 net/mac80211/sta_info.c if (!sta->sta.txq[tid] || txq 1744 net/mac80211/sta_info.c txq_has_queue(sta->sta.txq[tid])) txq 2158 net/mac80211/sta_info.c to_txq_info(sta->sta.txq[tid])); txq 276 net/mac80211/tdls.c struct ieee80211_tx_queue_params *txq; txq 297 net/mac80211/tdls.c txq = &sdata->tx_conf[ieee80211_ac_from_wmm(i)]; txq 298 net/mac80211/tdls.c wmm->ac[i].aci_aifsn = ieee80211_wmm_aci_aifsn(txq->aifs, txq 299 net/mac80211/tdls.c txq->acm, i); txq 300 net/mac80211/tdls.c wmm->ac[i].cw = ieee80211_wmm_ecw(txq->cw_min, txq->cw_max); txq 301 net/mac80211/tdls.c wmm->ac[i].txop_limit = cpu_to_le16(txq->txop); txq 2660 net/mac80211/trace.h struct txq_info *txq), txq 2662 net/mac80211/trace.h TP_ARGS(local, sdata, txq), txq 2673 net/mac80211/trace.h struct ieee80211_sta *sta = txq->txq.sta; txq 2678 net/mac80211/trace.h __entry->ac = txq->txq.ac; txq 2679 net/mac80211/trace.h __entry->tid = txq->txq.tid; txq 1099 net/mac80211/tx.c } else if (!tx->sta->sta.txq[tid]) { txq 1253 net/mac80211/tx.c struct ieee80211_txq *txq = NULL; txq 1268 net/mac80211/tx.c txq = sta->sta.txq[IEEE80211_NUM_TIDS]; txq 1276 net/mac80211/tx.c txq = sta->sta.txq[tid]; txq 1278 net/mac80211/tx.c txq = vif->txq; txq 1281 net/mac80211/tx.c if (!txq) txq 1284 net/mac80211/tx.c return to_txq_info(txq); txq 1314 net/mac80211/tx.c local = vif_to_sdata(txqi->txq.vif)->local; txq 1333 net/mac80211/tx.c local = vif_to_sdata(txqi->txq.vif)->local; txq 1353 net/mac80211/tx.c if (txqi->txq.sta) { txq 1354 net/mac80211/tx.c struct sta_info *sta = container_of(txqi->txq.sta, txq 1438 net/mac80211/tx.c if (!ap->vif.txq) txq 1441 net/mac80211/tx.c txqi = to_txq_info(ap->vif.txq); txq 1461 net/mac80211/tx.c txqi->txq.vif = &sdata->vif; txq 1464 net/mac80211/tx.c sdata->vif.txq = &txqi->txq; txq 1465 net/mac80211/tx.c txqi->txq.tid = 0; txq 1466 net/mac80211/tx.c txqi->txq.ac = IEEE80211_AC_BE; txq 1482 net/mac80211/tx.c txqi->txq.ac = IEEE80211_AC_VO; txq 1484 net/mac80211/tx.c txqi->txq.ac = ieee80211_ac_from_tid(tid); txq 1487 net/mac80211/tx.c txqi->txq.sta = &sta->sta; txq 1488 net/mac80211/tx.c txqi->txq.tid = tid; txq 1489 net/mac80211/tx.c sta->sta.txq[tid] = &txqi->txq; txq 1503 net/mac80211/tx.c spin_lock_bh(&local->active_txq_lock[txqi->txq.ac]); txq 1505 net/mac80211/tx.c spin_unlock_bh(&local->active_txq_lock[txqi->txq.ac]); txq 3236 net/mac80211/tx.c struct ieee80211_txq *txq = sta->sta.txq[tid]; txq 3258 net/mac80211/tx.c if (!txq) txq 3261 net/mac80211/tx.c txqi = to_txq_info(txq); txq 3552 net/mac80211/tx.c struct ieee80211_txq *txq) txq 3555 net/mac80211/tx.c struct txq_info *txqi = container_of(txq, struct txq_info, txq); txq 3563 net/mac80211/tx.c struct ieee80211_vif *vif = txq->vif; txq 3574 net/mac80211/tx.c if (vif->txqs_stopped[ieee80211_ac_from_tid(txq->tid)]) { txq 3599 net/mac80211/tx.c if (txq->sta) { txq 3600 net/mac80211/tx.c tx.sta = container_of(txq->sta, struct sta_info, sta); txq 3636 net/mac80211/tx.c struct sta_info *sta = container_of(txq->sta, struct sta_info, txq 3719 net/mac80211/tx.c if (txqi->txq.sta) { txq 3720 net/mac80211/tx.c struct sta_info *sta = container_of(txqi->txq.sta, txq 3723 net/mac80211/tx.c if (sta->airtime[txqi->txq.ac].deficit < 0) { txq 3724 net/mac80211/tx.c sta->airtime[txqi->txq.ac].deficit += txq 3727 net/mac80211/tx.c &local->active_txqs[txqi->txq.ac]); txq 3738 net/mac80211/tx.c ret = &txqi->txq; txq 3747 net/mac80211/tx.c struct ieee80211_txq *txq, txq 3751 net/mac80211/tx.c struct txq_info *txqi = to_txq_info(txq); txq 3753 net/mac80211/tx.c spin_lock_bh(&local->active_txq_lock[txq->ac]); txq 3765 net/mac80211/tx.c if (txqi->txq.sta && txq 3769 net/mac80211/tx.c &local->active_txqs[txq->ac]); txq 3772 net/mac80211/tx.c &local->active_txqs[txq->ac]); txq 3775 net/mac80211/tx.c spin_unlock_bh(&local->active_txq_lock[txq->ac]); txq 3780 net/mac80211/tx.c struct ieee80211_txq *txq) txq 3783 net/mac80211/tx.c struct txq_info *iter, *tmp, *txqi = to_txq_info(txq); txq 3785 net/mac80211/tx.c u8 ac = txq->ac; txq 3789 net/mac80211/tx.c if (!txqi->txq.sta) txq 3800 net/mac80211/tx.c if (!iter->txq.sta) { txq 3805 net/mac80211/tx.c sta = container_of(iter->txq.sta, struct sta_info, sta); txq 3811 net/mac80211/tx.c sta = container_of(txqi->txq.sta, struct sta_info, sta); txq 262 net/mac80211/util.c for (i = 0; i < ARRAY_SIZE(sta->sta.txq); i++) { txq 263 net/mac80211/util.c struct ieee80211_txq *txq = sta->sta.txq[i]; txq 265 net/mac80211/util.c if (!txq) txq 268 net/mac80211/util.c txqi = to_txq_info(txq); txq 270 net/mac80211/util.c if (ac != txq->ac) txq 283 net/mac80211/util.c if (!vif->txq) txq 286 net/mac80211/util.c txqi = to_txq_info(vif->txq); txq 289 net/mac80211/util.c (ps && atomic_read(&ps->num_sta_ps)) || ac != vif->txq->ac) txq 3983 net/mac80211/util.c void ieee80211_txq_get_depth(struct ieee80211_txq *txq, txq 3987 net/mac80211/util.c struct txq_info *txqi = to_txq_info(txq); txq 53 net/sched/sch_generic.c const struct netdev_queue *txq = q->dev_queue; txq 65 net/sched/sch_generic.c txq = skb_get_tx_queue(txq->dev, skb); txq 66 net/sched/sch_generic.c if (!netif_xmit_frozen_or_stopped(txq)) { txq 154 net/sched/sch_generic.c const struct netdev_queue *txq, txq 157 net/sched/sch_generic.c int bytelimit = qdisc_avail_bulklimit(txq) - skb->len; txq 205 net/sched/sch_generic.c const struct netdev_queue *txq = q->dev_queue; txq 233 net/sched/sch_generic.c txq = skb_get_tx_queue(txq->dev, skb); txq 234 net/sched/sch_generic.c if (!netif_xmit_frozen_or_stopped(txq)) { txq 254 net/sched/sch_generic.c netif_xmit_frozen_or_stopped(txq)) txq 267 net/sched/sch_generic.c try_bulk_dequeue_skb(q, skb, txq, packets); txq 272 net/sched/sch_generic.c trace_qdisc_dequeue(q, txq, *packets, skb); txq 286 net/sched/sch_generic.c struct net_device *dev, struct netdev_queue *txq, txq 311 net/sched/sch_generic.c HARD_TX_LOCK(dev, txq, smp_processor_id()); txq 312 net/sched/sch_generic.c if (!netif_xmit_frozen_or_stopped(txq)) txq 313 net/sched/sch_generic.c skb = dev_hard_start_xmit(skb, dev, txq, &ret); txq 315 net/sched/sch_generic.c HARD_TX_UNLOCK(dev, txq); txq 360 net/sched/sch_generic.c struct netdev_queue *txq; txq 374 net/sched/sch_generic.c txq = skb_get_tx_queue(dev, skb); txq 376 net/sched/sch_generic.c return sch_direct_xmit(skb, q, dev, txq, root_lock, validate); txq 432 net/sched/sch_generic.c struct netdev_queue *txq; txq 434 net/sched/sch_generic.c txq = netdev_get_tx_queue(dev, i); txq 435 net/sched/sch_generic.c trans_start = txq->trans_start; txq 436 net/sched/sch_generic.c if (netif_xmit_stopped(txq) && txq 440 net/sched/sch_generic.c txq->trans_timeout++; txq 1056 net/sched/sch_generic.c struct netdev_queue *txq; txq 1059 net/sched/sch_generic.c txq = netdev_get_tx_queue(dev, 0); txq 1064 net/sched/sch_generic.c dev->qdisc = txq->qdisc_sleeping; txq 1067 net/sched/sch_generic.c qdisc = qdisc_create_dflt(txq, &mq_qdisc_ops, TC_H_ROOT, NULL); txq 146 net/sched/sch_teql.c struct netdev_queue *txq; txq 149 net/sched/sch_teql.c txq = netdev_get_tx_queue(master->dev, 0); txq 152 net/sched/sch_teql.c root_lock = qdisc_root_sleeping_lock(rtnl_dereference(txq->qdisc)); txq 154 net/sched/sch_teql.c qdisc_reset(rtnl_dereference(txq->qdisc)); txq 217 net/sched/sch_teql.c struct net_device *dev, struct netdev_queue *txq, txq 257 net/sched/sch_teql.c struct netdev_queue *txq) txq 262 net/sched/sch_teql.c if (rcu_access_pointer(txq->qdisc) == &noop_qdisc) txq 269 net/sched/sch_teql.c res = __teql_resolve(skb, skb_res, dev, txq, dst); txq 295 net/xfrm/xfrm_device.c struct netdev_queue *txq; txq 300 net/xfrm/xfrm_device.c txq = netdev_core_pick_tx(dev, skb, NULL); txq 302 net/xfrm/xfrm_device.c HARD_TX_LOCK(dev, txq, smp_processor_id()); txq 303 net/xfrm/xfrm_device.c if (!netif_xmit_frozen_or_stopped(txq)) txq 304 net/xfrm/xfrm_device.c skb = dev_hard_start_xmit(skb, dev, txq, &ret); txq 305 net/xfrm/xfrm_device.c HARD_TX_UNLOCK(dev, txq);