tq 270 drivers/input/keyboard/lkkbd.c struct work_struct tq; tq 457 drivers/input/keyboard/lkkbd.c schedule_work(&lk->tq); tq 568 drivers/input/keyboard/lkkbd.c struct lkkbd *lk = container_of(work, struct lkkbd, tq); tq 623 drivers/input/keyboard/lkkbd.c INIT_WORK(&lk->tq, lkkbd_reinit); tq 63 drivers/input/keyboard/sunkbd.c struct work_struct tq; tq 102 drivers/input/keyboard/sunkbd.c schedule_work(&sunkbd->tq); tq 209 drivers/input/keyboard/sunkbd.c struct sunkbd *sunkbd = container_of(work, struct sunkbd, tq); tq 252 drivers/input/keyboard/sunkbd.c INIT_WORK(&sunkbd->tq, sunkbd_reinit); tq 190 drivers/input/serio/hp_sdc.c curr = hp_sdc.tq[hp_sdc.rcurr]; tq 316 drivers/input/serio/hp_sdc.c curr = hp_sdc.tq[hp_sdc.rcurr]; tq 378 drivers/input/serio/hp_sdc.c if (hp_sdc.tq[curridx] != NULL) tq 392 drivers/input/serio/hp_sdc.c if (hp_sdc.tq[curridx] != NULL) tq 412 drivers/input/serio/hp_sdc.c curr = hp_sdc.tq[curridx]; tq 416 drivers/input/serio/hp_sdc.c hp_sdc.tq[curridx] = NULL; tq 430 drivers/input/serio/hp_sdc.c hp_sdc.tq[curridx] = NULL; tq 573 drivers/input/serio/hp_sdc.c hp_sdc.tq[curridx] = NULL; tq 609 drivers/input/serio/hp_sdc.c if (hp_sdc.tq[i] == this) tq 617 drivers/input/serio/hp_sdc.c if (hp_sdc.tq[i] == NULL) { tq 618 drivers/input/serio/hp_sdc.c hp_sdc.tq[i] = this; tq 652 drivers/input/serio/hp_sdc.c if (hp_sdc.tq[i] == this) tq 653 drivers/input/serio/hp_sdc.c hp_sdc.tq[i] = NULL; tq 853 drivers/input/serio/hp_sdc.c memset(&hp_sdc.tq, 0, sizeof(hp_sdc.tq)); tq 205 drivers/net/can/dev.c bt->tq = (u32)v64; tq 259 drivers/net/can/dev.c brp64 = (u64)priv->clock.freq * (u64)bt->tq; tq 310 drivers/net/can/dev.c if (!bt->tq && bt->bitrate && btc) tq 312 drivers/net/can/dev.c else if (bt->tq && !bt->bitrate && btc) tq 314 drivers/net/can/dev.c else if (!bt->tq && bt->bitrate && bitrate_const) tq 140 drivers/net/can/usb/ucan.c __le32 tq; /* Time quanta (TQ) in nanoseconds */ tq 1249 drivers/net/can/usb/ucan.c cmd_set_bittiming->tq = cpu_to_le32(up->can.bittiming.tq); tq 1941 drivers/net/ethernet/chelsio/cxgb4/sge.c struct sge_txq *tq = &txq->q; tq 1950 drivers/net/ethernet/chelsio/cxgb4/sge.c if (unlikely((void *)sgl == (void *)tq->stat)) { tq 1951 drivers/net/ethernet/chelsio/cxgb4/sge.c sgl = (void *)tq->desc; tq 1952 drivers/net/ethernet/chelsio/cxgb4/sge.c end = (void *)((void *)tq->desc + tq 1953 drivers/net/ethernet/chelsio/cxgb4/sge.c ((void *)end - (void *)tq->stat)); tq 1956 drivers/net/ethernet/chelsio/cxgb4/sge.c cxgb4_write_sgl(skb, tq, sgl, end, 0, addr); tq 1959 drivers/net/ethernet/chelsio/cxgb4/sge.c last_desc = tq->pidx + ndesc - 1; tq 1960 drivers/net/ethernet/chelsio/cxgb4/sge.c if (last_desc >= tq->size) tq 1961 drivers/net/ethernet/chelsio/cxgb4/sge.c last_desc -= tq->size; tq 1962 drivers/net/ethernet/chelsio/cxgb4/sge.c tq->sdesc[last_desc].skb = skb; tq 1963 drivers/net/ethernet/chelsio/cxgb4/sge.c tq->sdesc[last_desc].sgl = sgl; tq 537 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c struct sge_txq *tq; tq 554 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c tq = s->egr_map[eq_idx]; tq 555 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c if (unlikely(tq == NULL)) { tq 560 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c txq = container_of(tq, struct sge_eth_txq, q); tq 561 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c if (unlikely(tq->abs_id != qid)) { tq 564 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c qid, tq->abs_id); tq 233 drivers/net/ethernet/chelsio/cxgb4vf/sge.c static inline unsigned int txq_avail(const struct sge_txq *tq) tq 235 drivers/net/ethernet/chelsio/cxgb4vf/sge.c return tq->size - 1 - tq->in_use; tq 307 drivers/net/ethernet/chelsio/cxgb4vf/sge.c const struct ulptx_sgl *sgl, const struct sge_txq *tq) tq 326 drivers/net/ethernet/chelsio/cxgb4vf/sge.c if (likely((u8 *)(p + 1) <= (u8 *)tq->stat)) { tq 333 drivers/net/ethernet/chelsio/cxgb4vf/sge.c } else if ((u8 *)p == (u8 *)tq->stat) { tq 334 drivers/net/ethernet/chelsio/cxgb4vf/sge.c p = (const struct ulptx_sge_pair *)tq->desc; tq 336 drivers/net/ethernet/chelsio/cxgb4vf/sge.c } else if ((u8 *)p + 8 == (u8 *)tq->stat) { tq 337 drivers/net/ethernet/chelsio/cxgb4vf/sge.c const __be64 *addr = (const __be64 *)tq->desc; tq 345 drivers/net/ethernet/chelsio/cxgb4vf/sge.c const __be64 *addr = (const __be64 *)tq->desc; tq 357 drivers/net/ethernet/chelsio/cxgb4vf/sge.c if ((u8 *)p == (u8 *)tq->stat) tq 358 drivers/net/ethernet/chelsio/cxgb4vf/sge.c p = (const struct ulptx_sge_pair *)tq->desc; tq 359 drivers/net/ethernet/chelsio/cxgb4vf/sge.c addr = ((u8 *)p + 16 <= (u8 *)tq->stat tq 361 drivers/net/ethernet/chelsio/cxgb4vf/sge.c : *(const __be64 *)tq->desc); tq 377 drivers/net/ethernet/chelsio/cxgb4vf/sge.c static void free_tx_desc(struct adapter *adapter, struct sge_txq *tq, tq 381 drivers/net/ethernet/chelsio/cxgb4vf/sge.c unsigned int cidx = tq->cidx; tq 386 drivers/net/ethernet/chelsio/cxgb4vf/sge.c sdesc = &tq->sdesc[cidx]; tq 394 drivers/net/ethernet/chelsio/cxgb4vf/sge.c unmap_sgl(dev, sdesc->skb, sdesc->sgl, tq); tq 400 drivers/net/ethernet/chelsio/cxgb4vf/sge.c if (++cidx == tq->size) { tq 402 drivers/net/ethernet/chelsio/cxgb4vf/sge.c sdesc = tq->sdesc; tq 405 drivers/net/ethernet/chelsio/cxgb4vf/sge.c tq->cidx = cidx; tq 411 drivers/net/ethernet/chelsio/cxgb4vf/sge.c static inline int reclaimable(const struct sge_txq *tq) tq 413 drivers/net/ethernet/chelsio/cxgb4vf/sge.c int hw_cidx = be16_to_cpu(tq->stat->cidx); tq 414 drivers/net/ethernet/chelsio/cxgb4vf/sge.c int reclaimable = hw_cidx - tq->cidx; tq 416 drivers/net/ethernet/chelsio/cxgb4vf/sge.c reclaimable += tq->size; tq 431 drivers/net/ethernet/chelsio/cxgb4vf/sge.c struct sge_txq *tq, tq 434 drivers/net/ethernet/chelsio/cxgb4vf/sge.c int avail = reclaimable(tq); tq 444 drivers/net/ethernet/chelsio/cxgb4vf/sge.c free_tx_desc(adapter, tq, avail, unmap); tq 445 drivers/net/ethernet/chelsio/cxgb4vf/sge.c tq->in_use -= avail; tq 901 drivers/net/ethernet/chelsio/cxgb4vf/sge.c static void write_sgl(const struct sk_buff *skb, struct sge_txq *tq, tq 930 drivers/net/ethernet/chelsio/cxgb4vf/sge.c to = (u8 *)end > (u8 *)tq->stat ? buf : sgl->sge; tq 943 drivers/net/ethernet/chelsio/cxgb4vf/sge.c if (unlikely((u8 *)end > (u8 *)tq->stat)) { tq 944 drivers/net/ethernet/chelsio/cxgb4vf/sge.c unsigned int part0 = (u8 *)tq->stat - (u8 *)sgl->sge, part1; tq 948 drivers/net/ethernet/chelsio/cxgb4vf/sge.c part1 = (u8 *)end - (u8 *)tq->stat; tq 949 drivers/net/ethernet/chelsio/cxgb4vf/sge.c memcpy(tq->desc, (u8 *)buf + part0, part1); tq 950 drivers/net/ethernet/chelsio/cxgb4vf/sge.c end = (void *)tq->desc + part1; tq 964 drivers/net/ethernet/chelsio/cxgb4vf/sge.c static inline void ring_tx_db(struct adapter *adapter, struct sge_txq *tq, tq 975 drivers/net/ethernet/chelsio/cxgb4vf/sge.c if (unlikely(tq->bar2_addr == NULL)) { tq 979 drivers/net/ethernet/chelsio/cxgb4vf/sge.c QID_V(tq->cntxt_id) | val); tq 995 drivers/net/ethernet/chelsio/cxgb4vf/sge.c if (n == 1 && tq->bar2_qid == 0) { tq 996 drivers/net/ethernet/chelsio/cxgb4vf/sge.c unsigned int index = (tq->pidx tq 997 drivers/net/ethernet/chelsio/cxgb4vf/sge.c ? (tq->pidx - 1) tq 998 drivers/net/ethernet/chelsio/cxgb4vf/sge.c : (tq->size - 1)); tq 999 drivers/net/ethernet/chelsio/cxgb4vf/sge.c __be64 *src = (__be64 *)&tq->desc[index]; tq 1000 drivers/net/ethernet/chelsio/cxgb4vf/sge.c __be64 __iomem *dst = (__be64 __iomem *)(tq->bar2_addr + tq 1024 drivers/net/ethernet/chelsio/cxgb4vf/sge.c writel(val | QID_V(tq->bar2_qid), tq 1025 drivers/net/ethernet/chelsio/cxgb4vf/sge.c tq->bar2_addr + SGE_UDB_KDOORBELL); tq 1052 drivers/net/ethernet/chelsio/cxgb4vf/sge.c static void inline_tx_skb(const struct sk_buff *skb, const struct sge_txq *tq, tq 1056 drivers/net/ethernet/chelsio/cxgb4vf/sge.c int left = (void *)tq->stat - pos; tq 1066 drivers/net/ethernet/chelsio/cxgb4vf/sge.c skb_copy_bits(skb, left, tq->desc, skb->len - left); tq 1067 drivers/net/ethernet/chelsio/cxgb4vf/sge.c pos = (void *)tq->desc + (skb->len - left); tq 1142 drivers/net/ethernet/chelsio/cxgb4vf/sge.c static inline void txq_advance(struct sge_txq *tq, unsigned int n) tq 1144 drivers/net/ethernet/chelsio/cxgb4vf/sge.c tq->in_use += n; tq 1145 drivers/net/ethernet/chelsio/cxgb4vf/sge.c tq->pidx += n; tq 1146 drivers/net/ethernet/chelsio/cxgb4vf/sge.c if (tq->pidx >= tq->size) tq 1147 drivers/net/ethernet/chelsio/cxgb4vf/sge.c tq->pidx -= tq->size; tq 1420 drivers/net/ethernet/chelsio/cxgb4vf/sge.c struct sge_txq *tq = &txq->q; tq 1430 drivers/net/ethernet/chelsio/cxgb4vf/sge.c if (unlikely((void *)sgl == (void *)tq->stat)) { tq 1431 drivers/net/ethernet/chelsio/cxgb4vf/sge.c sgl = (void *)tq->desc; tq 1432 drivers/net/ethernet/chelsio/cxgb4vf/sge.c end = ((void *)tq->desc + ((void *)end - (void *)tq->stat)); tq 1435 drivers/net/ethernet/chelsio/cxgb4vf/sge.c write_sgl(skb, tq, sgl, end, 0, addr); tq 1438 drivers/net/ethernet/chelsio/cxgb4vf/sge.c last_desc = tq->pidx + ndesc - 1; tq 1439 drivers/net/ethernet/chelsio/cxgb4vf/sge.c if (last_desc >= tq->size) tq 1440 drivers/net/ethernet/chelsio/cxgb4vf/sge.c last_desc -= tq->size; tq 1441 drivers/net/ethernet/chelsio/cxgb4vf/sge.c tq->sdesc[last_desc].skb = skb; tq 1442 drivers/net/ethernet/chelsio/cxgb4vf/sge.c tq->sdesc[last_desc].sgl = sgl; tq 2513 drivers/net/ethernet/chelsio/cxgb4vf/sge.c static void free_txq(struct adapter *adapter, struct sge_txq *tq) tq 2518 drivers/net/ethernet/chelsio/cxgb4vf/sge.c tq->size * sizeof(*tq->desc) + s->stat_len, tq 2519 drivers/net/ethernet/chelsio/cxgb4vf/sge.c tq->desc, tq->phys_addr); tq 2520 drivers/net/ethernet/chelsio/cxgb4vf/sge.c tq->cntxt_id = 0; tq 2521 drivers/net/ethernet/chelsio/cxgb4vf/sge.c tq->sdesc = NULL; tq 2522 drivers/net/ethernet/chelsio/cxgb4vf/sge.c tq->desc = NULL; tq 51 drivers/net/ifb.c struct sk_buff_head tq; tq 69 drivers/net/ifb.c skb = skb_peek(&txp->tq); tq 73 drivers/net/ifb.c skb_queue_splice_tail_init(&txp->rq, &txp->tq); tq 77 drivers/net/ifb.c while ((skb = __skb_dequeue(&txp->tq)) != NULL) { tq 92 drivers/net/ifb.c if (skb_queue_len(&txp->tq) != 0) tq 170 drivers/net/ifb.c __skb_queue_head_init(&txp->tq); tq 204 drivers/net/ifb.c __skb_queue_purge(&txp->tq); tq 103 drivers/net/vmxnet3/vmxnet3_drv.c vmxnet3_tq_stopped(struct vmxnet3_tx_queue *tq, struct vmxnet3_adapter *adapter) tq 105 drivers/net/vmxnet3/vmxnet3_drv.c return tq->stopped; tq 110 drivers/net/vmxnet3/vmxnet3_drv.c vmxnet3_tq_start(struct vmxnet3_tx_queue *tq, struct vmxnet3_adapter *adapter) tq 112 drivers/net/vmxnet3/vmxnet3_drv.c tq->stopped = false; tq 113 drivers/net/vmxnet3/vmxnet3_drv.c netif_start_subqueue(adapter->netdev, tq - adapter->tx_queue); tq 118 drivers/net/vmxnet3/vmxnet3_drv.c vmxnet3_tq_wake(struct vmxnet3_tx_queue *tq, struct vmxnet3_adapter *adapter) tq 120 drivers/net/vmxnet3/vmxnet3_drv.c tq->stopped = false; tq 121 drivers/net/vmxnet3/vmxnet3_drv.c netif_wake_subqueue(adapter->netdev, (tq - adapter->tx_queue)); tq 126 drivers/net/vmxnet3/vmxnet3_drv.c vmxnet3_tq_stop(struct vmxnet3_tx_queue *tq, struct vmxnet3_adapter *adapter) tq 128 drivers/net/vmxnet3/vmxnet3_drv.c tq->stopped = true; tq 129 drivers/net/vmxnet3/vmxnet3_drv.c tq->num_stop++; tq 130 drivers/net/vmxnet3/vmxnet3_drv.c netif_stop_subqueue(adapter->netdev, (tq - adapter->tx_queue)); tq 329 drivers/net/vmxnet3/vmxnet3_drv.c vmxnet3_unmap_pkt(u32 eop_idx, struct vmxnet3_tx_queue *tq, tq 336 drivers/net/vmxnet3/vmxnet3_drv.c BUG_ON(tq->buf_info[eop_idx].sop_idx != tq->tx_ring.next2comp); tq 337 drivers/net/vmxnet3/vmxnet3_drv.c BUG_ON(VMXNET3_TXDESC_GET_EOP(&(tq->tx_ring.base[eop_idx].txd)) != 1); tq 339 drivers/net/vmxnet3/vmxnet3_drv.c skb = tq->buf_info[eop_idx].skb; tq 341 drivers/net/vmxnet3/vmxnet3_drv.c tq->buf_info[eop_idx].skb = NULL; tq 343 drivers/net/vmxnet3/vmxnet3_drv.c VMXNET3_INC_RING_IDX_ONLY(eop_idx, tq->tx_ring.size); tq 345 drivers/net/vmxnet3/vmxnet3_drv.c while (tq->tx_ring.next2comp != eop_idx) { tq 346 drivers/net/vmxnet3/vmxnet3_drv.c vmxnet3_unmap_tx_buf(tq->buf_info + tq->tx_ring.next2comp, tq 354 drivers/net/vmxnet3/vmxnet3_drv.c vmxnet3_cmd_ring_adv_next2comp(&tq->tx_ring); tq 364 drivers/net/vmxnet3/vmxnet3_drv.c vmxnet3_tq_tx_complete(struct vmxnet3_tx_queue *tq, tq 370 drivers/net/vmxnet3/vmxnet3_drv.c gdesc = tq->comp_ring.base + tq->comp_ring.next2proc; tq 371 drivers/net/vmxnet3/vmxnet3_drv.c while (VMXNET3_TCD_GET_GEN(&gdesc->tcd) == tq->comp_ring.gen) { tq 378 drivers/net/vmxnet3/vmxnet3_drv.c &gdesc->tcd), tq, adapter->pdev, tq 381 drivers/net/vmxnet3/vmxnet3_drv.c vmxnet3_comp_ring_adv_next2proc(&tq->comp_ring); tq 382 drivers/net/vmxnet3/vmxnet3_drv.c gdesc = tq->comp_ring.base + tq->comp_ring.next2proc; tq 386 drivers/net/vmxnet3/vmxnet3_drv.c spin_lock(&tq->tx_lock); tq 387 drivers/net/vmxnet3/vmxnet3_drv.c if (unlikely(vmxnet3_tq_stopped(tq, adapter) && tq 388 drivers/net/vmxnet3/vmxnet3_drv.c vmxnet3_cmd_ring_desc_avail(&tq->tx_ring) > tq 389 drivers/net/vmxnet3/vmxnet3_drv.c VMXNET3_WAKE_QUEUE_THRESHOLD(tq) && tq 391 drivers/net/vmxnet3/vmxnet3_drv.c vmxnet3_tq_wake(tq, adapter); tq 393 drivers/net/vmxnet3/vmxnet3_drv.c spin_unlock(&tq->tx_lock); tq 400 drivers/net/vmxnet3/vmxnet3_drv.c vmxnet3_tq_cleanup(struct vmxnet3_tx_queue *tq, tq 405 drivers/net/vmxnet3/vmxnet3_drv.c while (tq->tx_ring.next2comp != tq->tx_ring.next2fill) { tq 408 drivers/net/vmxnet3/vmxnet3_drv.c tbi = tq->buf_info + tq->tx_ring.next2comp; tq 415 drivers/net/vmxnet3/vmxnet3_drv.c vmxnet3_cmd_ring_adv_next2comp(&tq->tx_ring); tq 419 drivers/net/vmxnet3/vmxnet3_drv.c for (i = 0; i < tq->tx_ring.size; i++) { tq 420 drivers/net/vmxnet3/vmxnet3_drv.c BUG_ON(tq->buf_info[i].skb != NULL || tq 421 drivers/net/vmxnet3/vmxnet3_drv.c tq->buf_info[i].map_type != VMXNET3_MAP_NONE); tq 424 drivers/net/vmxnet3/vmxnet3_drv.c tq->tx_ring.gen = VMXNET3_INIT_GEN; tq 425 drivers/net/vmxnet3/vmxnet3_drv.c tq->tx_ring.next2fill = tq->tx_ring.next2comp = 0; tq 427 drivers/net/vmxnet3/vmxnet3_drv.c tq->comp_ring.gen = VMXNET3_INIT_GEN; tq 428 drivers/net/vmxnet3/vmxnet3_drv.c tq->comp_ring.next2proc = 0; tq 433 drivers/net/vmxnet3/vmxnet3_drv.c vmxnet3_tq_destroy(struct vmxnet3_tx_queue *tq, tq 436 drivers/net/vmxnet3/vmxnet3_drv.c if (tq->tx_ring.base) { tq 437 drivers/net/vmxnet3/vmxnet3_drv.c dma_free_coherent(&adapter->pdev->dev, tq->tx_ring.size * tq 439 drivers/net/vmxnet3/vmxnet3_drv.c tq->tx_ring.base, tq->tx_ring.basePA); tq 440 drivers/net/vmxnet3/vmxnet3_drv.c tq->tx_ring.base = NULL; tq 442 drivers/net/vmxnet3/vmxnet3_drv.c if (tq->data_ring.base) { tq 444 drivers/net/vmxnet3/vmxnet3_drv.c tq->data_ring.size * tq->txdata_desc_size, tq 445 drivers/net/vmxnet3/vmxnet3_drv.c tq->data_ring.base, tq->data_ring.basePA); tq 446 drivers/net/vmxnet3/vmxnet3_drv.c tq->data_ring.base = NULL; tq 448 drivers/net/vmxnet3/vmxnet3_drv.c if (tq->comp_ring.base) { tq 449 drivers/net/vmxnet3/vmxnet3_drv.c dma_free_coherent(&adapter->pdev->dev, tq->comp_ring.size * tq 451 drivers/net/vmxnet3/vmxnet3_drv.c tq->comp_ring.base, tq->comp_ring.basePA); tq 452 drivers/net/vmxnet3/vmxnet3_drv.c tq->comp_ring.base = NULL; tq 454 drivers/net/vmxnet3/vmxnet3_drv.c if (tq->buf_info) { tq 456 drivers/net/vmxnet3/vmxnet3_drv.c tq->tx_ring.size * sizeof(tq->buf_info[0]), tq 457 drivers/net/vmxnet3/vmxnet3_drv.c tq->buf_info, tq->buf_info_pa); tq 458 drivers/net/vmxnet3/vmxnet3_drv.c tq->buf_info = NULL; tq 475 drivers/net/vmxnet3/vmxnet3_drv.c vmxnet3_tq_init(struct vmxnet3_tx_queue *tq, tq 481 drivers/net/vmxnet3/vmxnet3_drv.c memset(tq->tx_ring.base, 0, tq->tx_ring.size * tq 483 drivers/net/vmxnet3/vmxnet3_drv.c tq->tx_ring.next2fill = tq->tx_ring.next2comp = 0; tq 484 drivers/net/vmxnet3/vmxnet3_drv.c tq->tx_ring.gen = VMXNET3_INIT_GEN; tq 486 drivers/net/vmxnet3/vmxnet3_drv.c memset(tq->data_ring.base, 0, tq 487 drivers/net/vmxnet3/vmxnet3_drv.c tq->data_ring.size * tq->txdata_desc_size); tq 490 drivers/net/vmxnet3/vmxnet3_drv.c memset(tq->comp_ring.base, 0, tq->comp_ring.size * tq 492 drivers/net/vmxnet3/vmxnet3_drv.c tq->comp_ring.next2proc = 0; tq 493 drivers/net/vmxnet3/vmxnet3_drv.c tq->comp_ring.gen = VMXNET3_INIT_GEN; tq 496 drivers/net/vmxnet3/vmxnet3_drv.c memset(tq->buf_info, 0, sizeof(tq->buf_info[0]) * tq->tx_ring.size); tq 497 drivers/net/vmxnet3/vmxnet3_drv.c for (i = 0; i < tq->tx_ring.size; i++) tq 498 drivers/net/vmxnet3/vmxnet3_drv.c tq->buf_info[i].map_type = VMXNET3_MAP_NONE; tq 505 drivers/net/vmxnet3/vmxnet3_drv.c vmxnet3_tq_create(struct vmxnet3_tx_queue *tq, tq 510 drivers/net/vmxnet3/vmxnet3_drv.c BUG_ON(tq->tx_ring.base || tq->data_ring.base || tq 511 drivers/net/vmxnet3/vmxnet3_drv.c tq->comp_ring.base || tq->buf_info); tq 513 drivers/net/vmxnet3/vmxnet3_drv.c tq->tx_ring.base = dma_alloc_coherent(&adapter->pdev->dev, tq 514 drivers/net/vmxnet3/vmxnet3_drv.c tq->tx_ring.size * sizeof(struct Vmxnet3_TxDesc), tq 515 drivers/net/vmxnet3/vmxnet3_drv.c &tq->tx_ring.basePA, GFP_KERNEL); tq 516 drivers/net/vmxnet3/vmxnet3_drv.c if (!tq->tx_ring.base) { tq 521 drivers/net/vmxnet3/vmxnet3_drv.c tq->data_ring.base = dma_alloc_coherent(&adapter->pdev->dev, tq 522 drivers/net/vmxnet3/vmxnet3_drv.c tq->data_ring.size * tq->txdata_desc_size, tq 523 drivers/net/vmxnet3/vmxnet3_drv.c &tq->data_ring.basePA, GFP_KERNEL); tq 524 drivers/net/vmxnet3/vmxnet3_drv.c if (!tq->data_ring.base) { tq 529 drivers/net/vmxnet3/vmxnet3_drv.c tq->comp_ring.base = dma_alloc_coherent(&adapter->pdev->dev, tq 530 drivers/net/vmxnet3/vmxnet3_drv.c tq->comp_ring.size * sizeof(struct Vmxnet3_TxCompDesc), tq 531 drivers/net/vmxnet3/vmxnet3_drv.c &tq->comp_ring.basePA, GFP_KERNEL); tq 532 drivers/net/vmxnet3/vmxnet3_drv.c if (!tq->comp_ring.base) { tq 537 drivers/net/vmxnet3/vmxnet3_drv.c sz = tq->tx_ring.size * sizeof(tq->buf_info[0]); tq 538 drivers/net/vmxnet3/vmxnet3_drv.c tq->buf_info = dma_alloc_coherent(&adapter->pdev->dev, sz, tq 539 drivers/net/vmxnet3/vmxnet3_drv.c &tq->buf_info_pa, GFP_KERNEL); tq 540 drivers/net/vmxnet3/vmxnet3_drv.c if (!tq->buf_info) tq 546 drivers/net/vmxnet3/vmxnet3_drv.c vmxnet3_tq_destroy(tq, adapter); tq 675 drivers/net/vmxnet3/vmxnet3_drv.c struct vmxnet3_tx_queue *tq, struct pci_dev *pdev, tq 687 drivers/net/vmxnet3/vmxnet3_drv.c dw2 = (tq->tx_ring.gen ^ 0x1) << VMXNET3_TXD_GEN_SHIFT; tq 689 drivers/net/vmxnet3/vmxnet3_drv.c ctx->sop_txd = tq->tx_ring.base + tq->tx_ring.next2fill; tq 694 drivers/net/vmxnet3/vmxnet3_drv.c ctx->sop_txd->txd.addr = cpu_to_le64(tq->data_ring.basePA + tq 695 drivers/net/vmxnet3/vmxnet3_drv.c tq->tx_ring.next2fill * tq 696 drivers/net/vmxnet3/vmxnet3_drv.c tq->txdata_desc_size); tq 700 drivers/net/vmxnet3/vmxnet3_drv.c tbi = tq->buf_info + tq->tx_ring.next2fill; tq 705 drivers/net/vmxnet3/vmxnet3_drv.c tq->tx_ring.next2fill, tq 708 drivers/net/vmxnet3/vmxnet3_drv.c vmxnet3_cmd_ring_adv_next2fill(&tq->tx_ring); tq 711 drivers/net/vmxnet3/vmxnet3_drv.c dw2 = tq->tx_ring.gen << VMXNET3_TXD_GEN_SHIFT; tq 728 drivers/net/vmxnet3/vmxnet3_drv.c tbi = tq->buf_info + tq->tx_ring.next2fill; tq 738 drivers/net/vmxnet3/vmxnet3_drv.c gdesc = tq->tx_ring.base + tq->tx_ring.next2fill; tq 739 drivers/net/vmxnet3/vmxnet3_drv.c BUG_ON(gdesc->txd.gen == tq->tx_ring.gen); tq 747 drivers/net/vmxnet3/vmxnet3_drv.c tq->tx_ring.next2fill, le64_to_cpu(gdesc->txd.addr), tq 749 drivers/net/vmxnet3/vmxnet3_drv.c vmxnet3_cmd_ring_adv_next2fill(&tq->tx_ring); tq 750 drivers/net/vmxnet3/vmxnet3_drv.c dw2 = tq->tx_ring.gen << VMXNET3_TXD_GEN_SHIFT; tq 763 drivers/net/vmxnet3/vmxnet3_drv.c tbi = tq->buf_info + tq->tx_ring.next2fill; tq 780 drivers/net/vmxnet3/vmxnet3_drv.c gdesc = tq->tx_ring.base + tq->tx_ring.next2fill; tq 781 drivers/net/vmxnet3/vmxnet3_drv.c BUG_ON(gdesc->txd.gen == tq->tx_ring.gen); tq 789 drivers/net/vmxnet3/vmxnet3_drv.c tq->tx_ring.next2fill, le64_to_cpu(gdesc->txd.addr), tq 791 drivers/net/vmxnet3/vmxnet3_drv.c vmxnet3_cmd_ring_adv_next2fill(&tq->tx_ring); tq 792 drivers/net/vmxnet3/vmxnet3_drv.c dw2 = tq->tx_ring.gen << VMXNET3_TXD_GEN_SHIFT; tq 803 drivers/net/vmxnet3/vmxnet3_drv.c tbi->sop_idx = ctx->sop_txd - tq->tx_ring.base; tq 838 drivers/net/vmxnet3/vmxnet3_drv.c vmxnet3_parse_hdr(struct sk_buff *skb, struct vmxnet3_tx_queue *tq, tq 881 drivers/net/vmxnet3/vmxnet3_drv.c tq->txdata_desc_size, tq 893 drivers/net/vmxnet3/vmxnet3_drv.c if (unlikely(ctx->copy_size > tq->txdata_desc_size)) { tq 894 drivers/net/vmxnet3/vmxnet3_drv.c tq->stats.oversized_hdr++; tq 915 drivers/net/vmxnet3/vmxnet3_drv.c vmxnet3_copy_hdr(struct sk_buff *skb, struct vmxnet3_tx_queue *tq, tq 921 drivers/net/vmxnet3/vmxnet3_drv.c tdd = (struct Vmxnet3_TxDataDesc *)((u8 *)tq->data_ring.base + tq 922 drivers/net/vmxnet3/vmxnet3_drv.c tq->tx_ring.next2fill * tq 923 drivers/net/vmxnet3/vmxnet3_drv.c tq->txdata_desc_size); tq 928 drivers/net/vmxnet3/vmxnet3_drv.c ctx->copy_size, tq->tx_ring.next2fill); tq 979 drivers/net/vmxnet3/vmxnet3_drv.c vmxnet3_tq_xmit(struct sk_buff *skb, struct vmxnet3_tx_queue *tq, tq 1004 drivers/net/vmxnet3/vmxnet3_drv.c tq->stats.drop_tso++; tq 1007 drivers/net/vmxnet3/vmxnet3_drv.c tq->stats.copy_skb_header++; tq 1017 drivers/net/vmxnet3/vmxnet3_drv.c tq->stats.drop_too_many_frags++; tq 1020 drivers/net/vmxnet3/vmxnet3_drv.c tq->stats.linearized++; tq 1027 drivers/net/vmxnet3/vmxnet3_drv.c ret = vmxnet3_parse_hdr(skb, tq, &ctx, adapter); tq 1034 drivers/net/vmxnet3/vmxnet3_drv.c tq->stats.drop_oversized_hdr++; tq 1042 drivers/net/vmxnet3/vmxnet3_drv.c tq->stats.drop_oversized_hdr++; tq 1048 drivers/net/vmxnet3/vmxnet3_drv.c tq->stats.drop_hdr_inspect_err++; tq 1052 drivers/net/vmxnet3/vmxnet3_drv.c spin_lock_irqsave(&tq->tx_lock, flags); tq 1054 drivers/net/vmxnet3/vmxnet3_drv.c if (count > vmxnet3_cmd_ring_desc_avail(&tq->tx_ring)) { tq 1055 drivers/net/vmxnet3/vmxnet3_drv.c tq->stats.tx_ring_full++; tq 1059 drivers/net/vmxnet3/vmxnet3_drv.c tq->tx_ring.next2comp, tq->tx_ring.next2fill); tq 1061 drivers/net/vmxnet3/vmxnet3_drv.c vmxnet3_tq_stop(tq, adapter); tq 1062 drivers/net/vmxnet3/vmxnet3_drv.c spin_unlock_irqrestore(&tq->tx_lock, flags); tq 1067 drivers/net/vmxnet3/vmxnet3_drv.c vmxnet3_copy_hdr(skb, tq, &ctx, adapter); tq 1070 drivers/net/vmxnet3/vmxnet3_drv.c if (vmxnet3_map_pkt(skb, &ctx, tq, adapter->pdev, adapter)) tq 1084 drivers/net/vmxnet3/vmxnet3_drv.c tx_num_deferred = le32_to_cpu(tq->shared->txNumDeferred); tq 1102 drivers/net/vmxnet3/vmxnet3_drv.c le32_add_cpu(&tq->shared->txNumDeferred, num_pkts); tq 1129 drivers/net/vmxnet3/vmxnet3_drv.c tq->tx_ring.base), le64_to_cpu(gdesc->txd.addr), tq 1132 drivers/net/vmxnet3/vmxnet3_drv.c spin_unlock_irqrestore(&tq->tx_lock, flags); tq 1134 drivers/net/vmxnet3/vmxnet3_drv.c if (tx_num_deferred >= le32_to_cpu(tq->shared->txThreshold)) { tq 1135 drivers/net/vmxnet3/vmxnet3_drv.c tq->shared->txNumDeferred = 0; tq 1137 drivers/net/vmxnet3/vmxnet3_drv.c VMXNET3_REG_TXPROD + tq->qid * 8, tq 1138 drivers/net/vmxnet3/vmxnet3_drv.c tq->tx_ring.next2fill); tq 1144 drivers/net/vmxnet3/vmxnet3_drv.c spin_unlock_irqrestore(&tq->tx_lock, flags); tq 1146 drivers/net/vmxnet3/vmxnet3_drv.c tq->stats.drop_total++; tq 1911 drivers/net/vmxnet3/vmxnet3_drv.c struct vmxnet3_tx_queue *tq = tq 1913 drivers/net/vmxnet3/vmxnet3_drv.c vmxnet3_tq_tx_complete(tq, adapter); tq 1936 drivers/net/vmxnet3/vmxnet3_drv.c struct vmxnet3_tx_queue *tq = data; tq 1937 drivers/net/vmxnet3/vmxnet3_drv.c struct vmxnet3_adapter *adapter = tq->adapter; tq 1940 drivers/net/vmxnet3/vmxnet3_drv.c vmxnet3_disable_intr(adapter, tq->comp_ring.intr_idx); tq 1950 drivers/net/vmxnet3/vmxnet3_drv.c vmxnet3_tq_tx_complete(tq, adapter); tq 1952 drivers/net/vmxnet3/vmxnet3_drv.c vmxnet3_enable_intr(adapter, tq->comp_ring.intr_idx); tq 2444 drivers/net/vmxnet3/vmxnet3_drv.c struct vmxnet3_tx_queue *tq = &adapter->tx_queue[i]; tq 2447 drivers/net/vmxnet3/vmxnet3_drv.c tqc->txRingBasePA = cpu_to_le64(tq->tx_ring.basePA); tq 2448 drivers/net/vmxnet3/vmxnet3_drv.c tqc->dataRingBasePA = cpu_to_le64(tq->data_ring.basePA); tq 2449 drivers/net/vmxnet3/vmxnet3_drv.c tqc->compRingBasePA = cpu_to_le64(tq->comp_ring.basePA); tq 2450 drivers/net/vmxnet3/vmxnet3_drv.c tqc->ddPA = cpu_to_le64(tq->buf_info_pa); tq 2451 drivers/net/vmxnet3/vmxnet3_drv.c tqc->txRingSize = cpu_to_le32(tq->tx_ring.size); tq 2452 drivers/net/vmxnet3/vmxnet3_drv.c tqc->dataRingSize = cpu_to_le32(tq->data_ring.size); tq 2453 drivers/net/vmxnet3/vmxnet3_drv.c tqc->txDataRingDescSize = cpu_to_le32(tq->txdata_desc_size); tq 2454 drivers/net/vmxnet3/vmxnet3_drv.c tqc->compRingSize = cpu_to_le32(tq->comp_ring.size); tq 2458 drivers/net/vmxnet3/vmxnet3_drv.c tqc->intrIdx = tq->comp_ring.intr_idx; tq 2827 drivers/net/vmxnet3/vmxnet3_drv.c struct vmxnet3_tx_queue *tq = &adapter->tx_queue[i]; tq 2828 drivers/net/vmxnet3/vmxnet3_drv.c tq->tx_ring.size = tx_ring_size; tq 2829 drivers/net/vmxnet3/vmxnet3_drv.c tq->data_ring.size = tx_ring_size; tq 2830 drivers/net/vmxnet3/vmxnet3_drv.c tq->comp_ring.size = tx_ring_size; tq 2831 drivers/net/vmxnet3/vmxnet3_drv.c tq->txdata_desc_size = txdata_desc_size; tq 2832 drivers/net/vmxnet3/vmxnet3_drv.c tq->shared = &adapter->tqd_start[i].ctrl; tq 2833 drivers/net/vmxnet3/vmxnet3_drv.c tq->stopped = true; tq 2834 drivers/net/vmxnet3/vmxnet3_drv.c tq->adapter = adapter; tq 2835 drivers/net/vmxnet3/vmxnet3_drv.c tq->qid = i; tq 2836 drivers/net/vmxnet3/vmxnet3_drv.c err = vmxnet3_tq_create(tq, adapter); tq 392 drivers/net/vmxnet3/vmxnet3_ethtool.c struct vmxnet3_tx_queue *tq = &adapter->tx_queue[i]; tq 397 drivers/net/vmxnet3/vmxnet3_ethtool.c buf[j++] = VMXNET3_GET_ADDR_LO(tq->tx_ring.basePA); tq 398 drivers/net/vmxnet3/vmxnet3_ethtool.c buf[j++] = VMXNET3_GET_ADDR_HI(tq->tx_ring.basePA); tq 399 drivers/net/vmxnet3/vmxnet3_ethtool.c buf[j++] = tq->tx_ring.size; tq 400 drivers/net/vmxnet3/vmxnet3_ethtool.c buf[j++] = tq->tx_ring.next2fill; tq 401 drivers/net/vmxnet3/vmxnet3_ethtool.c buf[j++] = tq->tx_ring.next2comp; tq 402 drivers/net/vmxnet3/vmxnet3_ethtool.c buf[j++] = tq->tx_ring.gen; tq 404 drivers/net/vmxnet3/vmxnet3_ethtool.c buf[j++] = VMXNET3_GET_ADDR_LO(tq->data_ring.basePA); tq 405 drivers/net/vmxnet3/vmxnet3_ethtool.c buf[j++] = VMXNET3_GET_ADDR_HI(tq->data_ring.basePA); tq 406 drivers/net/vmxnet3/vmxnet3_ethtool.c buf[j++] = tq->data_ring.size; tq 407 drivers/net/vmxnet3/vmxnet3_ethtool.c buf[j++] = tq->txdata_desc_size; tq 409 drivers/net/vmxnet3/vmxnet3_ethtool.c buf[j++] = VMXNET3_GET_ADDR_LO(tq->comp_ring.basePA); tq 410 drivers/net/vmxnet3/vmxnet3_ethtool.c buf[j++] = VMXNET3_GET_ADDR_HI(tq->comp_ring.basePA); tq 411 drivers/net/vmxnet3/vmxnet3_ethtool.c buf[j++] = tq->comp_ring.size; tq 412 drivers/net/vmxnet3/vmxnet3_ethtool.c buf[j++] = tq->comp_ring.next2proc; tq 413 drivers/net/vmxnet3/vmxnet3_ethtool.c buf[j++] = tq->comp_ring.gen; tq 415 drivers/net/vmxnet3/vmxnet3_ethtool.c buf[j++] = tq->stopped; tq 404 drivers/net/vmxnet3/vmxnet3_int.h #define VMXNET3_WAKE_QUEUE_THRESHOLD(tq) (5) tq 286 drivers/net/wireless/ath/ath5k/qcu.c struct ath5k_txq_info *tq = &ah->ah_txq[queue]; tq 292 drivers/net/wireless/ath/ath5k/qcu.c (tq->tqi_cw_min << AR5K_NODCU_RETRY_LMT_CW_MIN_S) tq 326 drivers/net/wireless/ath/ath5k/qcu.c struct ath5k_txq_info *tq = &ah->ah_txq[queue]; tq 333 drivers/net/wireless/ath/ath5k/qcu.c (tq->tqi_type == AR5K_TX_QUEUE_INACTIVE)) tq 341 drivers/net/wireless/ath/ath5k/qcu.c AR5K_REG_SM(tq->tqi_cw_min, AR5K_DCU_LCL_IFS_CW_MIN) | tq 342 drivers/net/wireless/ath/ath5k/qcu.c AR5K_REG_SM(tq->tqi_cw_max, AR5K_DCU_LCL_IFS_CW_MAX) | tq 343 drivers/net/wireless/ath/ath5k/qcu.c AR5K_REG_SM(tq->tqi_aifs, AR5K_DCU_LCL_IFS_AIFS), tq 366 drivers/net/wireless/ath/ath5k/qcu.c if (tq->tqi_cbr_period) { tq 367 drivers/net/wireless/ath/ath5k/qcu.c ath5k_hw_reg_write(ah, AR5K_REG_SM(tq->tqi_cbr_period, tq 369 drivers/net/wireless/ath/ath5k/qcu.c AR5K_REG_SM(tq->tqi_cbr_overflow_limit, tq 376 drivers/net/wireless/ath/ath5k/qcu.c if (tq->tqi_cbr_overflow_limit) tq 382 drivers/net/wireless/ath/ath5k/qcu.c if (tq->tqi_ready_time && (tq->tqi_type != AR5K_TX_QUEUE_CAB)) tq 383 drivers/net/wireless/ath/ath5k/qcu.c ath5k_hw_reg_write(ah, AR5K_REG_SM(tq->tqi_ready_time, tq 388 drivers/net/wireless/ath/ath5k/qcu.c if (tq->tqi_burst_time) { tq 389 drivers/net/wireless/ath/ath5k/qcu.c ath5k_hw_reg_write(ah, AR5K_REG_SM(tq->tqi_burst_time, tq 394 drivers/net/wireless/ath/ath5k/qcu.c if (tq->tqi_flags & AR5K_TXQ_FLAG_RDYTIME_EXP_POLICY_ENABLE) tq 400 drivers/net/wireless/ath/ath5k/qcu.c if (tq->tqi_flags & AR5K_TXQ_FLAG_BACKOFF_DISABLE) tq 405 drivers/net/wireless/ath/ath5k/qcu.c if (tq->tqi_flags & AR5K_TXQ_FLAG_FRAG_BURST_BACKOFF_ENABLE) tq 412 drivers/net/wireless/ath/ath5k/qcu.c switch (tq->tqi_type) { tq 434 drivers/net/wireless/ath/ath5k/qcu.c ath5k_hw_reg_write(ah, ((tq->tqi_ready_time - tq 462 drivers/net/wireless/ath/ath5k/qcu.c if (tq->tqi_flags & AR5K_TXQ_FLAG_TXOKINT_ENABLE) tq 465 drivers/net/wireless/ath/ath5k/qcu.c if (tq->tqi_flags & AR5K_TXQ_FLAG_TXERRINT_ENABLE) tq 468 drivers/net/wireless/ath/ath5k/qcu.c if (tq->tqi_flags & AR5K_TXQ_FLAG_TXURNINT_ENABLE) tq 471 drivers/net/wireless/ath/ath5k/qcu.c if (tq->tqi_flags & AR5K_TXQ_FLAG_TXDESCINT_ENABLE) tq 474 drivers/net/wireless/ath/ath5k/qcu.c if (tq->tqi_flags & AR5K_TXQ_FLAG_TXEOLINT_ENABLE) tq 477 drivers/net/wireless/ath/ath5k/qcu.c if (tq->tqi_flags & AR5K_TXQ_FLAG_CBRORNINT_ENABLE) tq 480 drivers/net/wireless/ath/ath5k/qcu.c if (tq->tqi_flags & AR5K_TXQ_FLAG_CBRURNINT_ENABLE) tq 483 drivers/net/wireless/ath/ath5k/qcu.c if (tq->tqi_flags & AR5K_TXQ_FLAG_QTRIGINT_ENABLE) tq 486 drivers/net/wireless/ath/ath5k/qcu.c if (tq->tqi_flags & AR5K_TXQ_FLAG_TXNOFRMINT_ENABLE) tq 300 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(tq-full, 0x08), tq 403 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(tq-bank-conflict-issue-stall, 0x0b), tq 404 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(tq-full, 0x0c), tq 1735 drivers/tty/ipwireless/hardware.c struct ipw_tx_packet *tp, *tq; tq 1745 drivers/tty/ipwireless/hardware.c list_for_each_entry_safe(tp, tq, &hw->tx_queue[i], queue) { tq 281 include/linux/hp_sdc.h hp_sdc_transaction *tq[HP_SDC_QUEUE_LEN]; /* All pending read/writes */ tq 217 include/uapi/linux/batadv_packet.h __u8 tq; tq 34 include/uapi/linux/can/netlink.h __u32 tq; /* Time quanta (TQ) in nanoseconds */ tq 260 ipc/sem.c struct sem_queue *q, *tq; tq 270 ipc/sem.c list_for_each_entry_safe(q, tq, &sma->pending_alter, list) { tq 1127 ipc/sem.c struct sem_queue *q, *tq; tq 1144 ipc/sem.c list_for_each_entry_safe(q, tq, &sma->pending_const, list) { tq 1149 ipc/sem.c list_for_each_entry_safe(q, tq, &sma->pending_alter, list) { tq 1155 ipc/sem.c list_for_each_entry_safe(q, tq, &sem->pending_const, list) { tq 1159 ipc/sem.c list_for_each_entry_safe(q, tq, &sem->pending_alter, list) { tq 219 net/batman-adv/bat_iv_ogm.c batadv_ogm_packet->tq = BATADV_TQ_MAX_VALUE; tq 295 net/batman-adv/bat_iv_ogm.c static u8 batadv_hop_penalty(u8 tq, const struct batadv_priv *bat_priv) tq 300 net/batman-adv/bat_iv_ogm.c new_tq = tq * (BATADV_TQ_MAX_VALUE - hop_penalty); tq 374 net/batman-adv/bat_iv_ogm.c batadv_ogm_packet->tq, batadv_ogm_packet->ttl, tq 716 net/batman-adv/bat_iv_ogm.c batadv_ogm_packet->tq = batadv_hop_penalty(batadv_ogm_packet->tq, tq 721 net/batman-adv/bat_iv_ogm.c batadv_ogm_packet->tq, batadv_ogm_packet->ttl); tq 999 net/batman-adv/bat_iv_ogm.c batadv_ogm_packet->tq); tq 1171 net/batman-adv/bat_iv_ogm.c combined_tq = batadv_ogm_packet->tq * tq 1178 net/batman-adv/bat_iv_ogm.c batadv_ogm_packet->tq = combined_tq; tq 1184 net/batman-adv/bat_iv_ogm.c batadv_ogm_packet->tq, if_incoming->net_dev->name, tq 1190 net/batman-adv/bat_iv_ogm.c if (batadv_ogm_packet->tq >= BATADV_TQ_TOTAL_BIDRECT_LIMIT) tq 1357 net/batman-adv/bat_iv_ogm.c if (ogm_packet->tq == 0) { tq 1604 net/batman-adv/bat_iv_ogm.c ogm_packet->tq, ogm_packet->ttl, tq 778 net/batman-adv/network-coding.c if (ogm_packet->tq < bat_priv->nc.min_tq) tq 1010 net/batman-adv/network-coding.c static u8 batadv_nc_random_weight_tq(u8 tq) tq 1013 net/batman-adv/network-coding.c u8 rand_tq = prandom_u32_max(BATADV_TQ_MAX_VALUE + 1 - tq); tq 284 net/tipc/trace.h TP_PROTO(struct tipc_link *r, u16 f, u16 t, struct sk_buff_head *tq), tq 286 net/tipc/trace.h TP_ARGS(r, f, t, tq), tq 301 net/tipc/trace.h __entry->len = skb_queue_len(tq); tq 302 net/tipc/trace.h __entry->fseqno = msg_seqno(buf_msg(skb_peek(tq))); tq 303 net/tipc/trace.h __entry->lseqno = msg_seqno(buf_msg(skb_peek_tail(tq))); tq 312 net/tipc/trace.h TP_PROTO(struct tipc_link *r, u16 f, u16 t, struct sk_buff_head *tq), tq 313 net/tipc/trace.h TP_ARGS(r, f, t, tq) tq 317 net/tipc/trace.h TP_PROTO(struct tipc_link *r, u16 f, u16 t, struct sk_buff_head *tq), tq 318 net/tipc/trace.h TP_ARGS(r, f, t, tq),