tx 238 arch/mips/include/asm/ip32/mace.h volatile unsigned long tx; tx 1064 arch/mips/include/asm/octeon/cvmx-agl-defs.h uint64_t tx:2; tx 1070 arch/mips/include/asm/octeon/cvmx-agl-defs.h uint64_t tx:2; tx 1077 arch/mips/include/asm/octeon/cvmx-agl-defs.h uint64_t tx:1; tx 1083 arch/mips/include/asm/octeon/cvmx-agl-defs.h uint64_t tx:1; tx 216 arch/mips/include/asm/processor.h u64 tx[4]; tx 55 arch/mips/netlogic/xlp/cop2-ex.c : "r"(r->tx), "r"(r->rx)); tx 85 arch/mips/netlogic/xlp/cop2-ex.c : : "m"(*r), "r"(r->tx), "r"(r->rx)); tx 27 arch/s390/include/uapi/asm/guarded_storage.h __u8 tx : 1; tx 4226 arch/x86/events/intel/core.c EVENT_ATTR_STR(tx-start, tx_start, "event=0xc9,umask=0x1"); tx 4227 arch/x86/events/intel/core.c EVENT_ATTR_STR(tx-commit, tx_commit, "event=0xc9,umask=0x2"); tx 4228 arch/x86/events/intel/core.c EVENT_ATTR_STR(tx-abort, tx_abort, "event=0xc9,umask=0x4"); tx 4229 arch/x86/events/intel/core.c EVENT_ATTR_STR(tx-capacity, tx_capacity, "event=0x54,umask=0x2"); tx 4230 arch/x86/events/intel/core.c EVENT_ATTR_STR(tx-conflict, tx_conflict, "event=0x54,umask=0x1"); tx 4272 arch/x86/events/intel/core.c EVENT_ATTR_STR(tx-capacity-read, tx_capacity_read, "event=0x54,umask=0x80"); tx 4273 arch/x86/events/intel/core.c EVENT_ATTR_STR(tx-capacity-write, tx_capacity_write, "event=0x54,umask=0x2"); tx 39 crypto/async_tx/async_memcpy.c struct dma_async_tx_descriptor *tx = NULL; tx 61 crypto/async_tx/async_memcpy.c tx = device->device_prep_dma_memcpy(chan, unmap->addr[1], tx 66 crypto/async_tx/async_memcpy.c if (tx) { tx 69 crypto/async_tx/async_memcpy.c dma_set_unmap(tx, unmap); tx 70 crypto/async_tx/async_memcpy.c async_tx_submit(chan, tx, submit); tx 91 crypto/async_tx/async_memcpy.c return tx; tx 41 crypto/async_tx/async_pq.c struct dma_async_tx_descriptor *tx = NULL; tx 78 crypto/async_tx/async_pq.c tx = dma->device_prep_dma_pq(chan, dma_dest, tx 83 crypto/async_tx/async_pq.c if (likely(tx)) tx 89 crypto/async_tx/async_pq.c dma_set_unmap(tx, unmap); tx 90 crypto/async_tx/async_pq.c async_tx_submit(chan, tx, submit); tx 91 crypto/async_tx/async_pq.c submit->depend_tx = tx; tx 100 crypto/async_tx/async_pq.c return tx; tx 183 crypto/async_tx/async_pq.c struct dma_async_tx_descriptor *tx; tx 228 crypto/async_tx/async_pq.c tx = do_async_gen_syndrome(chan, coefs, j, unmap, dma_flags, submit); tx 230 crypto/async_tx/async_pq.c return tx; tx 287 crypto/async_tx/async_pq.c struct dma_async_tx_descriptor *tx; tx 342 crypto/async_tx/async_pq.c tx = device->device_prep_dma_pq_val(chan, pq, tx 348 crypto/async_tx/async_pq.c if (likely(tx)) tx 354 crypto/async_tx/async_pq.c dma_set_unmap(tx, unmap); tx 355 crypto/async_tx/async_pq.c async_tx_submit(chan, tx, submit); tx 379 crypto/async_tx/async_pq.c tx = NULL; tx 384 crypto/async_tx/async_pq.c tx = async_xor(spare, blocks, offset, disks-2, len, submit); tx 385 crypto/async_tx/async_pq.c async_tx_quiesce(&tx); tx 395 crypto/async_tx/async_pq.c tx = async_gen_syndrome(blocks, offset, disks, len, submit); tx 396 crypto/async_tx/async_pq.c async_tx_quiesce(&tx); tx 410 crypto/async_tx/async_pq.c tx = NULL; tx 414 crypto/async_tx/async_pq.c return tx; tx 35 crypto/async_tx/async_raid6_recov.c struct dma_async_tx_descriptor *tx; tx 50 crypto/async_tx/async_raid6_recov.c tx = dma->device_prep_dma_pq(chan, pq, unmap->addr, 2, coef, tx 52 crypto/async_tx/async_raid6_recov.c if (tx) { tx 53 crypto/async_tx/async_raid6_recov.c dma_set_unmap(tx, unmap); tx 54 crypto/async_tx/async_raid6_recov.c async_tx_submit(chan, tx, submit); tx 56 crypto/async_tx/async_raid6_recov.c return tx; tx 99 crypto/async_tx/async_raid6_recov.c struct dma_async_tx_descriptor *tx; tx 115 crypto/async_tx/async_raid6_recov.c tx = dma->device_prep_dma_pq(chan, dma_dest, unmap->addr, tx 118 crypto/async_tx/async_raid6_recov.c if (tx) { tx 119 crypto/async_tx/async_raid6_recov.c dma_set_unmap(tx, unmap); tx 121 crypto/async_tx/async_raid6_recov.c async_tx_submit(chan, tx, submit); tx 122 crypto/async_tx/async_raid6_recov.c return tx; tx 149 crypto/async_tx/async_raid6_recov.c struct dma_async_tx_descriptor *tx = NULL; tx 170 crypto/async_tx/async_raid6_recov.c init_async_submit(submit, ASYNC_TX_FENCE, tx, NULL, NULL, scribble); tx 171 crypto/async_tx/async_raid6_recov.c tx = async_sum_product(b, srcs, coef, bytes, submit); tx 176 crypto/async_tx/async_raid6_recov.c init_async_submit(submit, flags | ASYNC_TX_XOR_ZERO_DST, tx, cb_fn, tx 178 crypto/async_tx/async_raid6_recov.c tx = async_xor(a, srcs, 0, 2, bytes, submit); tx 180 crypto/async_tx/async_raid6_recov.c return tx; tx 188 crypto/async_tx/async_raid6_recov.c struct dma_async_tx_descriptor *tx = NULL; tx 221 crypto/async_tx/async_raid6_recov.c init_async_submit(submit, ASYNC_TX_FENCE, tx, NULL, NULL, scribble); tx 222 crypto/async_tx/async_raid6_recov.c tx = async_memcpy(dp, g, 0, 0, bytes, submit); tx 223 crypto/async_tx/async_raid6_recov.c init_async_submit(submit, ASYNC_TX_FENCE, tx, NULL, NULL, scribble); tx 224 crypto/async_tx/async_raid6_recov.c tx = async_mult(dq, g, raid6_gfexp[good], bytes, submit); tx 229 crypto/async_tx/async_raid6_recov.c init_async_submit(submit, ASYNC_TX_FENCE|ASYNC_TX_XOR_DROP_DST, tx, tx 231 crypto/async_tx/async_raid6_recov.c tx = async_xor(dp, srcs, 0, 2, bytes, submit); tx 236 crypto/async_tx/async_raid6_recov.c init_async_submit(submit, ASYNC_TX_FENCE|ASYNC_TX_XOR_DROP_DST, tx, tx 238 crypto/async_tx/async_raid6_recov.c tx = async_xor(dq, srcs, 0, 2, bytes, submit); tx 245 crypto/async_tx/async_raid6_recov.c init_async_submit(submit, ASYNC_TX_FENCE, tx, NULL, NULL, scribble); tx 246 crypto/async_tx/async_raid6_recov.c tx = async_sum_product(dq, srcs, coef, bytes, submit); tx 251 crypto/async_tx/async_raid6_recov.c init_async_submit(submit, flags | ASYNC_TX_XOR_DROP_DST, tx, cb_fn, tx 253 crypto/async_tx/async_raid6_recov.c tx = async_xor(dp, srcs, 0, 2, bytes, submit); tx 255 crypto/async_tx/async_raid6_recov.c return tx; tx 262 crypto/async_tx/async_raid6_recov.c struct dma_async_tx_descriptor *tx = NULL; tx 285 crypto/async_tx/async_raid6_recov.c init_async_submit(submit, ASYNC_TX_FENCE, tx, NULL, NULL, scribble); tx 286 crypto/async_tx/async_raid6_recov.c tx = async_gen_syndrome(blocks, 0, disks, bytes, submit); tx 297 crypto/async_tx/async_raid6_recov.c init_async_submit(submit, ASYNC_TX_FENCE|ASYNC_TX_XOR_DROP_DST, tx, tx 299 crypto/async_tx/async_raid6_recov.c tx = async_xor(dp, srcs, 0, 2, bytes, submit); tx 304 crypto/async_tx/async_raid6_recov.c init_async_submit(submit, ASYNC_TX_FENCE|ASYNC_TX_XOR_DROP_DST, tx, tx 306 crypto/async_tx/async_raid6_recov.c tx = async_xor(dq, srcs, 0, 2, bytes, submit); tx 313 crypto/async_tx/async_raid6_recov.c init_async_submit(submit, ASYNC_TX_FENCE, tx, NULL, NULL, scribble); tx 314 crypto/async_tx/async_raid6_recov.c tx = async_sum_product(dq, srcs, coef, bytes, submit); tx 319 crypto/async_tx/async_raid6_recov.c init_async_submit(submit, flags | ASYNC_TX_XOR_DROP_DST, tx, cb_fn, tx 321 crypto/async_tx/async_raid6_recov.c tx = async_xor(dp, srcs, 0, 2, bytes, submit); tx 323 crypto/async_tx/async_raid6_recov.c return tx; tx 412 crypto/async_tx/async_raid6_recov.c struct dma_async_tx_descriptor *tx = NULL; tx 476 crypto/async_tx/async_raid6_recov.c init_async_submit(submit, ASYNC_TX_FENCE, tx, NULL, NULL, tx 478 crypto/async_tx/async_raid6_recov.c tx = async_memcpy(p, g, 0, 0, bytes, submit); tx 480 crypto/async_tx/async_raid6_recov.c init_async_submit(submit, ASYNC_TX_FENCE, tx, NULL, NULL, tx 482 crypto/async_tx/async_raid6_recov.c tx = async_mult(dq, g, raid6_gfexp[good], bytes, submit); tx 484 crypto/async_tx/async_raid6_recov.c init_async_submit(submit, ASYNC_TX_FENCE, tx, NULL, NULL, tx 486 crypto/async_tx/async_raid6_recov.c tx = async_gen_syndrome(blocks, 0, disks, bytes, submit); tx 498 crypto/async_tx/async_raid6_recov.c init_async_submit(submit, ASYNC_TX_FENCE|ASYNC_TX_XOR_DROP_DST, tx, tx 500 crypto/async_tx/async_raid6_recov.c tx = async_xor(dq, srcs, 0, 2, bytes, submit); tx 502 crypto/async_tx/async_raid6_recov.c init_async_submit(submit, ASYNC_TX_FENCE, tx, NULL, NULL, scribble); tx 503 crypto/async_tx/async_raid6_recov.c tx = async_mult(dq, dq, coef, bytes, submit); tx 507 crypto/async_tx/async_raid6_recov.c init_async_submit(submit, flags | ASYNC_TX_XOR_DROP_DST, tx, cb_fn, tx 509 crypto/async_tx/async_raid6_recov.c tx = async_xor(p, srcs, 0, 2, bytes, submit); tx 511 crypto/async_tx/async_raid6_recov.c return tx; tx 66 crypto/async_tx/async_tx.c struct dma_async_tx_descriptor *tx) tx 74 crypto/async_tx/async_tx.c if (txd_parent(depend_tx) && depend_tx->chan == tx->chan) { tx 75 crypto/async_tx/async_tx.c txd_chain(depend_tx, tx); tx 100 crypto/async_tx/async_tx.c txd_chain(intr_tx, tx); tx 121 crypto/async_tx/async_tx.c tx->tx_submit(tx); tx 143 crypto/async_tx/async_tx.c async_tx_submit(struct dma_chan *chan, struct dma_async_tx_descriptor *tx, tx 148 crypto/async_tx/async_tx.c tx->callback = submit->cb_fn; tx 149 crypto/async_tx/async_tx.c tx->callback_param = submit->cb_param; tx 161 crypto/async_tx/async_tx.c txd_parent(tx)); tx 173 crypto/async_tx/async_tx.c txd_chain(depend_tx, tx); tx 192 crypto/async_tx/async_tx.c async_tx_channel_switch(depend_tx, tx); tx 195 crypto/async_tx/async_tx.c txd_clear_parent(tx); tx 196 crypto/async_tx/async_tx.c tx->tx_submit(tx); tx 200 crypto/async_tx/async_tx.c txd_clear_parent(tx); tx 201 crypto/async_tx/async_tx.c tx->tx_submit(tx); tx 205 crypto/async_tx/async_tx.c async_tx_ack(tx); tx 225 crypto/async_tx/async_tx.c struct dma_async_tx_descriptor *tx; tx 238 crypto/async_tx/async_tx.c tx = device ? device->device_prep_dma_interrupt(chan, 0) : NULL; tx 240 crypto/async_tx/async_tx.c tx = NULL; tx 242 crypto/async_tx/async_tx.c if (tx) { tx 245 crypto/async_tx/async_tx.c async_tx_submit(chan, tx, submit); tx 255 crypto/async_tx/async_tx.c return tx; tx 263 crypto/async_tx/async_tx.c void async_tx_quiesce(struct dma_async_tx_descriptor **tx) tx 265 crypto/async_tx/async_tx.c if (*tx) { tx 269 crypto/async_tx/async_tx.c BUG_ON(async_tx_test_ack(*tx)); tx 270 crypto/async_tx/async_tx.c if (dma_wait_for_async_tx(*tx) != DMA_COMPLETE) tx 273 crypto/async_tx/async_tx.c async_tx_ack(*tx); tx 274 crypto/async_tx/async_tx.c *tx = NULL; tx 27 crypto/async_tx/async_xor.c struct dma_async_tx_descriptor *tx = NULL; tx 65 crypto/async_tx/async_xor.c tx = dma->device_prep_dma_xor(chan, dma_dest, src_list, tx 69 crypto/async_tx/async_xor.c if (unlikely(!tx)) tx 73 crypto/async_tx/async_xor.c while (unlikely(!tx)) { tx 75 crypto/async_tx/async_xor.c tx = dma->device_prep_dma_xor(chan, dma_dest, tx 82 crypto/async_tx/async_xor.c dma_set_unmap(tx, unmap); tx 83 crypto/async_tx/async_xor.c async_tx_submit(chan, tx, submit); tx 84 crypto/async_tx/async_xor.c submit->depend_tx = tx; tx 96 crypto/async_tx/async_xor.c return tx; tx 175 crypto/async_tx/async_xor.c struct dma_async_tx_descriptor *tx; tx 195 crypto/async_tx/async_xor.c tx = do_async_xor(chan, unmap, submit); tx 197 crypto/async_tx/async_xor.c return tx; tx 262 crypto/async_tx/async_xor.c struct dma_async_tx_descriptor *tx = NULL; tx 289 crypto/async_tx/async_xor.c tx = device->device_prep_dma_xor_val(chan, unmap->addr, src_cnt, tx 292 crypto/async_tx/async_xor.c if (unlikely(!tx)) { tx 295 crypto/async_tx/async_xor.c while (!tx) { tx 297 crypto/async_tx/async_xor.c tx = device->device_prep_dma_xor_val(chan, tx 302 crypto/async_tx/async_xor.c dma_set_unmap(tx, unmap); tx 303 crypto/async_tx/async_xor.c async_tx_submit(chan, tx, submit); tx 315 crypto/async_tx/async_xor.c tx = async_xor(dest, src_list, offset, src_cnt, len, submit); tx 317 crypto/async_tx/async_xor.c async_tx_quiesce(&tx); tx 326 crypto/async_tx/async_xor.c return tx; tx 59 crypto/async_tx/raid6test.c struct dma_async_tx_descriptor *tx = NULL; tx 69 crypto/async_tx/raid6test.c tx = async_gen_syndrome(ptrs, 0, disks, bytes, &submit); tx 89 crypto/async_tx/raid6test.c tx = async_xor(dest, blocks, 0, count, bytes, &submit); tx 91 crypto/async_tx/raid6test.c init_async_submit(&submit, 0, tx, NULL, NULL, addr_conv); tx 92 crypto/async_tx/raid6test.c tx = async_gen_syndrome(ptrs, 0, disks, bytes, &submit); tx 98 crypto/async_tx/raid6test.c tx = async_raid6_datap_recov(disks, bytes, faila, ptrs, &submit); tx 102 crypto/async_tx/raid6test.c tx = async_raid6_2data_recov(disks, bytes, faila, failb, ptrs, &submit); tx 106 crypto/async_tx/raid6test.c init_async_submit(&submit, ASYNC_TX_ACK, tx, callback, &cmp, addr_conv); tx 107 crypto/async_tx/raid6test.c tx = async_syndrome_val(ptrs, 0, disks, bytes, &result, spare, &submit); tx 108 crypto/async_tx/raid6test.c async_tx_issue_pending(tx); tx 146 crypto/async_tx/raid6test.c struct dma_async_tx_descriptor *tx; tx 165 crypto/async_tx/raid6test.c tx = async_gen_syndrome(dataptrs, 0, disks, PAGE_SIZE, &submit); tx 166 crypto/async_tx/raid6test.c async_tx_issue_pending(tx); tx 1267 crypto/ecc.c u64 tx[ECC_MAX_DIGITS]; tx 1271 crypto/ecc.c vli_set(tx, point->x, ndigits); tx 1273 crypto/ecc.c apply_z(tx, ty, z, curve->p, ndigits); tx 1274 crypto/ecc.c vli_mod_sub(tz, rx, tx, curve->p, ndigits); tx 1275 crypto/ecc.c xycz_add(tx, ty, rx, ry, curve->p, ndigits); tx 237 drivers/acpi/processor_throttling.c if (p_limit->thermal.tx > target_state) tx 238 drivers/acpi/processor_throttling.c target_state = p_limit->thermal.tx; tx 239 drivers/acpi/processor_throttling.c if (p_limit->user.tx > target_state) tx 240 drivers/acpi/processor_throttling.c target_state = p_limit->user.tx; tx 333 drivers/acpi/processor_throttling.c if (limit->thermal.tx > target_state) tx 334 drivers/acpi/processor_throttling.c target_state = limit->thermal.tx; tx 335 drivers/acpi/processor_throttling.c if (limit->user.tx > target_state) tx 336 drivers/acpi/processor_throttling.c target_state = limit->user.tx; tx 534 drivers/acpi/processor_throttling.c struct acpi_processor_tx_tss *tx = tx 539 drivers/acpi/processor_throttling.c state.pointer = tx; tx 552 drivers/acpi/processor_throttling.c if (!tx->freqpercentage) { tx 836 drivers/acpi/processor_throttling.c struct acpi_processor_tx_tss *tx = tx 839 drivers/acpi/processor_throttling.c if (tx->control == value) tx 851 drivers/acpi/processor_throttling.c struct acpi_processor_tx_tss *tx = tx 854 drivers/acpi/processor_throttling.c *value = tx->control; tx 397 drivers/ata/pata_arasan_cf.c struct dma_async_tx_descriptor *tx; tx 403 drivers/ata/pata_arasan_cf.c tx = chan->device->device_prep_dma_memcpy(chan, dest, src, len, flags); tx 404 drivers/ata/pata_arasan_cf.c if (!tx) { tx 409 drivers/ata/pata_arasan_cf.c tx->callback = dma_callback; tx 410 drivers/ata/pata_arasan_cf.c tx->callback_param = acdev; tx 411 drivers/ata/pata_arasan_cf.c cookie = tx->tx_submit(tx); tx 50 drivers/ata/pata_pxa.c struct dma_async_tx_descriptor *tx; tx 57 drivers/ata/pata_pxa.c tx = dmaengine_prep_slave_sg(pd->dma_chan, qc->sg, qc->n_elem, dir, tx 59 drivers/ata/pata_pxa.c if (!tx) { tx 63 drivers/ata/pata_pxa.c tx->callback = pxa_ata_dma_irq; tx 64 drivers/ata/pata_pxa.c tx->callback_param = pd; tx 65 drivers/ata/pata_pxa.c pd->dma_cookie = dmaengine_submit(tx); tx 118 drivers/atm/adummy.c atomic_inc(&vcc->stats->tx); tx 435 drivers/atm/ambassador.c static void tx_complete (amb_dev * dev, tx_out * tx) { tx 436 drivers/atm/ambassador.c tx_simple * tx_descr = bus_to_virt (tx->handle); tx 439 drivers/atm/ambassador.c PRINTD (DBG_FLOW|DBG_TX, "tx_complete %p %p", dev, tx); tx 442 drivers/atm/ambassador.c atomic_inc(&ATM_SKB(skb)->vcc->stats->tx); tx 612 drivers/atm/ambassador.c static int tx_give (amb_dev * dev, tx_in * tx) { tx 626 drivers/atm/ambassador.c *txq->in.ptr = *tx; tx 1300 drivers/atm/ambassador.c tx_in tx; tx 1357 drivers/atm/ambassador.c tx.vc = cpu_to_be16 (vcc->tx_frame_bits | vc); tx 1358 drivers/atm/ambassador.c tx.tx_descr_length = cpu_to_be16 (sizeof(tx_frag)+sizeof(tx_frag_end)); tx 1359 drivers/atm/ambassador.c tx.tx_descr_addr = cpu_to_be32 (virt_to_bus (&tx_descr->tx_frag)); tx 1361 drivers/atm/ambassador.c while (tx_give (dev, &tx)) tx 229 drivers/atm/atmtcp.c atomic_inc(&vcc->stats->tx); tx 316 drivers/atm/atmtcp.c atomic_inc(&vcc->stats->tx); tx 189 drivers/atm/eni.c if (eni_dev->tx[i].send) tx 191 drivers/atm/eni.c eni_dev->tx[i].send,eni_dev->tx[i].words*4); tx 1038 drivers/atm/eni.c struct eni_tx *tx; tx 1052 drivers/atm/eni.c tx = eni_vcc->tx; tx 1053 drivers/atm/eni.c NULLCHECK(tx); tx 1092 drivers/atm/eni.c if (!NEPMOK(tx->tx_pos,size+TX_GAP, tx 1093 drivers/atm/eni.c eni_in(MID_TX_RDPTR(tx->index)),tx->words)) { tx 1110 drivers/atm/eni.c DPRINTK("dma_wr is %d, tx_pos is %ld\n",dma_wr,tx->tx_pos); tx 1122 drivers/atm/eni.c eni_dev->dma[j++] = (((tx->tx_pos+TX_DESCR_SIZE) & (tx->words-1)) << tx 1123 drivers/atm/eni.c MID_DMA_COUNT_SHIFT) | (tx->index << MID_DMA_CHAN_SHIFT) | tx 1127 drivers/atm/eni.c if (aal5) put_dma(tx->index,eni_dev->dma,&j,paddr,skb->len); tx 1128 drivers/atm/eni.c else put_dma(tx->index,eni_dev->dma,&j,paddr+4,skb->len-4); tx 1133 drivers/atm/eni.c put_dma(tx->index,eni_dev->dma,&j,(unsigned long) tx 1137 drivers/atm/eni.c put_dma(tx->index,eni_dev->dma,&j,(unsigned long) tx 1143 drivers/atm/eni.c put_dma(tx->index, eni_dev->dma, &j, eni_dev->zero.dma, tx 1147 drivers/atm/eni.c eni_dev->dma[j++] = (((tx->tx_pos+size) & (tx->words-1)) << tx 1148 drivers/atm/eni.c MID_DMA_COUNT_SHIFT) | (tx->index << MID_DMA_CHAN_SHIFT) | tx 1154 drivers/atm/eni.c (aal5 ? MID_SEG_AAL5 : 0) | (tx->prescaler << MID_SEG_PR_SHIFT) | tx 1155 drivers/atm/eni.c (tx->resolution << MID_SEG_RATE_SHIFT) | tx 1156 drivers/atm/eni.c (size/(ATM_CELL_PAYLOAD/4)),tx->send+tx->tx_pos*4); tx 1161 drivers/atm/eni.c tx->send+((tx->tx_pos+1) & (tx->words-1))*4); tx 1164 drivers/atm/eni.c writel(skb->len,tx->send+ tx 1165 drivers/atm/eni.c ((tx->tx_pos+size-AAL5_TRAILER) & (tx->words-1))*4); tx 1172 drivers/atm/eni.c ENI_PRV_POS(skb) = tx->tx_pos; tx 1175 drivers/atm/eni.c tx->tx_pos = (tx->tx_pos+size) & (tx->words-1); tx 1176 drivers/atm/eni.c DPRINTK("dma_wr set to %d, tx_pos is now %ld\n",dma_wr,tx->tx_pos); tx 1186 drivers/atm/eni.c struct eni_tx *tx; tx 1193 drivers/atm/eni.c tx = &ENI_DEV(dev)->tx[i]; tx 1194 drivers/atm/eni.c if (tx->send) tx 1195 drivers/atm/eni.c while ((skb = skb_dequeue(&tx->backlog))) { tx 1199 drivers/atm/eni.c skb_queue_head(&tx->backlog,skb); tx 1213 drivers/atm/eni.c struct eni_tx *tx; tx 1221 drivers/atm/eni.c tx = ENI_VCC(vcc)->tx; tx 1222 drivers/atm/eni.c NULLCHECK(ENI_VCC(vcc)->tx); tx 1224 drivers/atm/eni.c (unsigned) eni_in(MID_TX_DESCRSTART(tx->index))); tx 1225 drivers/atm/eni.c if (ENI_VCC(vcc)->txing < tx->words && ENI_PRV_POS(skb) == tx 1226 drivers/atm/eni.c eni_in(MID_TX_DESCRSTART(tx->index))) { tx 1235 drivers/atm/eni.c atomic_inc(&vcc->stats->tx); tx 1247 drivers/atm/eni.c if (!eni_dev->tx[i].send) return eni_dev->tx+i; tx 1294 drivers/atm/eni.c struct eni_tx *tx; tx 1313 drivers/atm/eni.c eni_vcc->tx = eni_dev->ubr; tx 1319 drivers/atm/eni.c new_tx = !eni_vcc->tx; tx 1321 drivers/atm/eni.c if (!new_tx) tx = eni_vcc->tx; tx 1325 drivers/atm/eni.c tx = alloc_tx(eni_dev,unlimited); tx 1326 drivers/atm/eni.c if (!tx) { tx 1330 drivers/atm/eni.c DPRINTK("got chan %d\n",tx->index); tx 1331 drivers/atm/eni.c tx->reserved = tx->shaping = 0; tx 1332 drivers/atm/eni.c tx->send = mem; tx 1333 drivers/atm/eni.c tx->words = size >> 2; tx 1334 drivers/atm/eni.c skb_queue_head_init(&tx->backlog); tx 1337 drivers/atm/eni.c ((tx->send-eni_dev->ram) >> (MID_LOC_SKIP+2)), tx 1338 drivers/atm/eni.c MID_TX_PLACE(tx->index)); tx 1339 drivers/atm/eni.c tx->tx_pos = eni_in(MID_TX_DESCRSTART(tx->index)) & tx 1342 drivers/atm/eni.c error = comp_tx(eni_dev,&rate,tx->reserved,&pre,&res,unlimited); tx 1346 drivers/atm/eni.c if (!error && !ubr && rate > eni_dev->tx_bw+tx->reserved) tx 1348 drivers/atm/eni.c if (!error && set_rsv && !set_shp && rate < tx->shaping) tx 1350 drivers/atm/eni.c if (!error && !set_rsv && rate > tx->reserved && !ubr) tx 1354 drivers/atm/eni.c tx->send = NULL; tx 1361 drivers/atm/eni.c eni_dev->tx_bw += tx->reserved; tx 1362 drivers/atm/eni.c tx->reserved = rate; tx 1366 drivers/atm/eni.c if (unlimited && new_tx) eni_dev->ubr = tx; tx 1367 drivers/atm/eni.c tx->prescaler = pre; tx 1368 drivers/atm/eni.c tx->resolution = res; tx 1369 drivers/atm/eni.c tx->shaping = rate; tx 1371 drivers/atm/eni.c if (set_shp) eni_vcc->tx = tx; tx 1372 drivers/atm/eni.c DPRINTK("rsv %d shp %d\n",tx->reserved,tx->shaping); tx 1379 drivers/atm/eni.c ENI_VCC(vcc)->tx = NULL; tx 1399 drivers/atm/eni.c if (!eni_vcc->tx) return; tx 1409 drivers/atm/eni.c txing = skb_peek(&eni_vcc->tx->backlog) || eni_vcc->txing; tx 1418 drivers/atm/eni.c if (eni_vcc->tx != eni_dev->ubr) { tx 1424 drivers/atm/eni.c while (eni_in(MID_TX_RDPTR(eni_vcc->tx->index)) != tx 1425 drivers/atm/eni.c eni_in(MID_TX_DESCRSTART(eni_vcc->tx->index))) tx 1427 drivers/atm/eni.c eni_free_mem(eni_dev,eni_vcc->tx->send,eni_vcc->tx->words << 2); tx 1428 drivers/atm/eni.c eni_vcc->tx->send = NULL; tx 1429 drivers/atm/eni.c eni_dev->tx_bw += eni_vcc->tx->reserved; tx 1431 drivers/atm/eni.c eni_vcc->tx = NULL; tx 1449 drivers/atm/eni.c eni_dev->tx[i].send = NULL; tx 1450 drivers/atm/eni.c eni_dev->tx[i].index = i; tx 1933 drivers/atm/eni.c eni_vcc->tx = NULL; /* for eni_close after open_rx */ tx 1961 drivers/atm/eni.c struct eni_tx *tx = ENI_VCC(vcc)->tx; tx 1966 drivers/atm/eni.c if (tx == eni_dev->ubr) return -EBADFD; tx 1970 drivers/atm/eni.c if ((flgs & ATM_MF_DEC_RSV) && rate && rate < tx->reserved) rsv = 1; tx 1971 drivers/atm/eni.c if ((flgs & ATM_MF_INC_RSV) && (!rate || rate > tx->reserved)) rsv = 1; tx 1972 drivers/atm/eni.c if ((flgs & ATM_MF_DEC_SHP) && rate && rate < tx->shaping) shp = 1; tx 1973 drivers/atm/eni.c if ((flgs & ATM_MF_INC_SHP) && (!rate || rate > tx->shaping)) shp = 1; tx 1987 drivers/atm/eni.c dsc = tx->send+ENI_PRV_POS(skb)*4; tx 1989 drivers/atm/eni.c (tx->prescaler << MID_SEG_PR_SHIFT) | tx 1990 drivers/atm/eni.c (tx->resolution << MID_SEG_RATE_SHIFT), dsc); tx 2015 drivers/atm/eni.c if ((mult.tx && mult.tx <= 100) || (mult.rx &&mult.rx <= 100) || tx 2016 drivers/atm/eni.c mult.tx > 65536 || mult.rx > 65536) tx 2018 drivers/atm/eni.c if (mult.tx) eni_dev->tx_mult = mult.tx; tx 2056 drivers/atm/eni.c if (!ENI_VCC(vcc)->tx) { tx 2080 drivers/atm/eni.c skb_queue_tail(&ENI_VCC(vcc)->tx->backlog,skb); tx 2160 drivers/atm/eni.c struct eni_tx *tx = eni_dev->tx+i; tx 2162 drivers/atm/eni.c if (!tx->send) continue; tx 2166 drivers/atm/eni.c (unsigned long) (tx->send - eni_dev->ram), tx 2167 drivers/atm/eni.c tx->send-eni_dev->ram+tx->words*4-1,tx->words*4, tx 2168 drivers/atm/eni.c tx->reserved,tx->shaping, tx 2169 drivers/atm/eni.c tx == eni_dev->ubr ? " (UBR)" : ""); tx 2173 drivers/atm/eni.c skb_queue_len(&tx->backlog)); tx 2195 drivers/atm/eni.c if (eni_vcc->tx) length += sprintf(page+length,", "); tx 2197 drivers/atm/eni.c if (eni_vcc->tx) tx 2199 drivers/atm/eni.c eni_vcc->tx->index,eni_vcc->txing); tx 59 drivers/atm/eni.h struct eni_tx *tx; /* TXer, NULL if none */ tx 85 drivers/atm/eni.h struct eni_tx tx[NR_CHAN]; /* TX channels */ tx 741 drivers/atm/firestream.c atomic_inc(&ATM_SKB(skb)->vcc->stats->tx); tx 848 drivers/atm/fore200e.c atomic_inc(&vcc->stats->tx); tx 2605 drivers/atm/he.c atomic_inc(&vcc->stats->tx); tx 1159 drivers/atm/horizon.c atomic_inc(&ATM_SKB(skb)->vcc->stats->tx); tx 814 drivers/atm/idt77252.c atomic_inc(&vcc->stats->tx); tx 2948 drivers/atm/iphase.c atomic_inc(&vcc->stats->tx); tx 3053 drivers/atm/iphase.c atomic_inc(&vcc->stats->tx); tx 3060 drivers/atm/iphase.c if (atomic_read(&vcc->stats->tx) % 20 == 0) { tx 241 drivers/atm/lanai.c } tx; tx 719 drivers/atm/lanai.c dma_addr_t dmaaddr = lvcc->tx.buf.dmaaddr; tx 727 drivers/atm/lanai.c (lvcc->tx.atmvcc->qos.txtp.traffic_class == ATM_CBR) ? tx 732 drivers/atm/lanai.c TXADDR1_SET_SIZE(lanai_buf_size_cardorder(&lvcc->tx.buf)), tx 773 drivers/atm/lanai.c while ((skb = skb_dequeue(&lvcc->tx.backlog)) != NULL) tx 774 drivers/atm/lanai.c lanai_free_skb(lvcc->tx.atmvcc, skb); tx 784 drivers/atm/lanai.c (((lanai_buf_size(&lvcc->tx.buf) / 1024) * HZ) >> 7); tx 789 drivers/atm/lanai.c (lvcc->tx.atmvcc->qos.txtp.traffic_class != ATM_CBR || tx 800 drivers/atm/lanai.c lvcc->tx.atmvcc->dev->number, lvcc->vci); tx 1138 drivers/atm/lanai.c r -= ((unsigned long) lvcc->tx.buf.ptr) - tx 1139 drivers/atm/lanai.c ((unsigned long) lvcc->tx.buf.start); tx 1142 drivers/atm/lanai.c r += lanai_buf_size(&lvcc->tx.buf); tx 1149 drivers/atm/lanai.c return !skb_queue_empty(&lvcc->tx.backlog); tx 1163 drivers/atm/lanai.c APRINTK((((unsigned long) lvcc->tx.buf.ptr) & 15) == 0, tx 1164 drivers/atm/lanai.c "vcc_tx_add_aal5_descriptor: bad ptr=%p\n", lvcc->tx.buf.ptr); tx 1165 drivers/atm/lanai.c lvcc->tx.buf.ptr += 4; /* Hope the values REALLY don't matter */ tx 1166 drivers/atm/lanai.c pos = ((unsigned char *) lvcc->tx.buf.ptr) - tx 1167 drivers/atm/lanai.c (unsigned char *) lvcc->tx.buf.start; tx 1171 drivers/atm/lanai.c lvcc->tx.buf.start, lvcc->tx.buf.ptr, lvcc->tx.buf.end); tx 1172 drivers/atm/lanai.c pos = (pos + len) & (lanai_buf_size(&lvcc->tx.buf) - 1); tx 1176 drivers/atm/lanai.c lvcc->tx.buf.start, lvcc->tx.buf.ptr, lvcc->tx.buf.end); tx 1177 drivers/atm/lanai.c lvcc->tx.buf.ptr[-1] = tx 1179 drivers/atm/lanai.c ((lvcc->tx.atmvcc->atm_options & ATM_ATMOPT_CLP) ? tx 1181 drivers/atm/lanai.c if (lvcc->tx.buf.ptr >= lvcc->tx.buf.end) tx 1182 drivers/atm/lanai.c lvcc->tx.buf.ptr = lvcc->tx.buf.start; tx 1189 drivers/atm/lanai.c APRINTK((((unsigned long) lvcc->tx.buf.ptr) & 15) == 8, tx 1190 drivers/atm/lanai.c "vcc_tx_add_aal5_trailer: bad ptr=%p\n", lvcc->tx.buf.ptr); tx 1191 drivers/atm/lanai.c lvcc->tx.buf.ptr += 2; tx 1192 drivers/atm/lanai.c lvcc->tx.buf.ptr[-2] = cpu_to_be32((uu << 24) | (cpi << 16) | len); tx 1193 drivers/atm/lanai.c if (lvcc->tx.buf.ptr >= lvcc->tx.buf.end) tx 1194 drivers/atm/lanai.c lvcc->tx.buf.ptr = lvcc->tx.buf.start; tx 1202 drivers/atm/lanai.c e = ((unsigned char *) lvcc->tx.buf.ptr) + n; tx 1203 drivers/atm/lanai.c m = e - (unsigned char *) lvcc->tx.buf.end; tx 1206 drivers/atm/lanai.c memcpy(lvcc->tx.buf.ptr, src, n - m); tx 1208 drivers/atm/lanai.c memcpy(lvcc->tx.buf.start, src + n - m, m); tx 1209 drivers/atm/lanai.c e = ((unsigned char *) lvcc->tx.buf.start) + m; tx 1211 drivers/atm/lanai.c lvcc->tx.buf.ptr = (u32 *) e; tx 1220 drivers/atm/lanai.c e = ((unsigned char *) lvcc->tx.buf.ptr) + n; tx 1221 drivers/atm/lanai.c m = e - (unsigned char *) lvcc->tx.buf.end; tx 1224 drivers/atm/lanai.c memset(lvcc->tx.buf.ptr, 0, n - m); tx 1226 drivers/atm/lanai.c memset(lvcc->tx.buf.start, 0, m); tx 1227 drivers/atm/lanai.c e = ((unsigned char *) lvcc->tx.buf.start) + m; tx 1229 drivers/atm/lanai.c lvcc->tx.buf.ptr = (u32 *) e; tx 1236 drivers/atm/lanai.c int i, ptr = ((unsigned char *) lvcc->tx.buf.ptr) - tx 1237 drivers/atm/lanai.c (unsigned char *) lvcc->tx.buf.start; tx 1240 drivers/atm/lanai.c ptr, lvcc->vci, lvcc->tx.buf.start, lvcc->tx.buf.ptr, tx 1241 drivers/atm/lanai.c lvcc->tx.buf.end); tx 1293 drivers/atm/lanai.c lanai_free_skb(lvcc->tx.atmvcc, skb); tx 1294 drivers/atm/lanai.c atomic_inc(&lvcc->tx.atmvcc->stats->tx); tx 1308 drivers/atm/lanai.c skb = skb_dequeue(&lvcc->tx.backlog); tx 1314 drivers/atm/lanai.c skb_queue_head(&lvcc->tx.backlog, skb); tx 1340 drivers/atm/lanai.c skb_queue_tail(&lvcc->tx.backlog, skb); tx 1358 drivers/atm/lanai.c lanai_free_skb(lvcc->tx.atmvcc, skb); tx 1470 drivers/atm/lanai.c skb_queue_head_init(&lvcc->tx.backlog); tx 1512 drivers/atm/lanai.c lvcc->tx.unqueue = vcc_tx_unqueue_aal0; tx 1516 drivers/atm/lanai.c lvcc->tx.unqueue = vcc_tx_unqueue_aal5; tx 1520 drivers/atm/lanai.c return lanai_get_sized_buffer(lanai, &lvcc->tx.buf, max_sdu, tx 1634 drivers/atm/lanai.c if (unlikely(lvcc->tx.atmvcc == NULL)) { tx 1642 drivers/atm/lanai.c lvcc->tx.endptr = SERVICE_GET_END(s); tx 1703 drivers/atm/lanai.c lvcc->tx.unqueue(lanai, lvcc, lvcc->tx.endptr); tx 1750 drivers/atm/lanai.c if (lvcc == NULL || lvcc->tx.atmvcc == NULL || tx 1756 drivers/atm/lanai.c lvcc->tx.unqueue(lanai, lvcc, endptr); tx 1999 drivers/atm/lanai.c lvcc->tx.atmvcc != NULL && lvcc->tx.atmvcc != atmvcc) tx 2288 drivers/atm/lanai.c if (lvcc->tx.atmvcc == atmvcc) { tx 2295 drivers/atm/lanai.c lanai_buf_deallocate(&lvcc->tx.buf, lanai->pci); tx 2296 drivers/atm/lanai.c lvcc->tx.atmvcc = NULL; tx 2356 drivers/atm/lanai.c APRINTK(lvcc->tx.atmvcc == NULL, "tx.atmvcc!=NULL, vci=%d\n", tx 2361 drivers/atm/lanai.c lvcc->tx.atmvcc = atmvcc; tx 2376 drivers/atm/lanai.c if (atmvcc == lvcc->tx.atmvcc) { tx 2395 drivers/atm/lanai.c lvcc->tx.atmvcc != atmvcc)) tx 2519 drivers/atm/lanai.c if (lvcc->tx.atmvcc != NULL) tx 2522 drivers/atm/lanai.c lvcc->tx.atmvcc->qos.aal == ATM_AAL5 ? 5 : 0, tx 2523 drivers/atm/lanai.c lanai_buf_size(&lvcc->tx.buf), tx 2524 drivers/atm/lanai.c lvcc->tx.atmvcc == lanai->cbrvcc ? 'C' : 'U', tx 1257 drivers/atm/nicstar.c if (vcc->qos.txtp.traffic_class != ATM_NONE && vc->tx) tx 1379 drivers/atm/nicstar.c vc->tx = 1; tx 1461 drivers/atm/nicstar.c vc->tx = 0; tx 1642 drivers/atm/nicstar.c if (!vc->tx) { tx 1712 drivers/atm/nicstar.c atomic_inc(&vcc->stats->tx); tx 698 drivers/atm/nicstar.h volatile unsigned int tx:1; /* TX vc? */ tx 1121 drivers/atm/solos-pci.c atomic_inc(&vcc->stats->tx); tx 739 drivers/atm/zatm.c atomic_inc(&vcc->stats->tx); tx 206 drivers/block/aoe/aoenet.c kts.fn = tx; tx 390 drivers/bus/moxtet.c .tx_buf = moxtet->tx, tx 433 drivers/bus/moxtet.c moxtet->tx[moxtet->count - mdev->idx] = val; tx 435 drivers/bus/moxtet.c ret = spi_write(to_spi_device(moxtet->dev), moxtet->tx, tx 452 drivers/bus/moxtet.c return moxtet->tx[moxtet->count - mdev->idx]; tx 502 drivers/bus/moxtet.c p = hex_byte_pack(p, moxtet->tx[moxtet->count - i]); tx 538 drivers/bus/moxtet.c moxtet->tx[moxtet->count - i] = bin[i]; tx 540 drivers/bus/moxtet.c err = spi_write(to_spi_device(moxtet->dev), moxtet->tx, tx 2207 drivers/char/pcmcia/synclink_cs.c icount->tx = cnow.tx; tx 2583 drivers/char/pcmcia/synclink_cs.c info->icount.tx, info->icount.rx); tx 31 drivers/clk/mediatek/clk-apmixed.c struct mtk_ref2usb_tx *tx = to_mtk_ref2usb_tx(hw); tx 33 drivers/clk/mediatek/clk-apmixed.c return (readl(tx->base_addr) & REF2USB_EN_MASK) == REF2USB_EN_MASK; tx 38 drivers/clk/mediatek/clk-apmixed.c struct mtk_ref2usb_tx *tx = to_mtk_ref2usb_tx(hw); tx 41 drivers/clk/mediatek/clk-apmixed.c val = readl(tx->base_addr); tx 44 drivers/clk/mediatek/clk-apmixed.c writel(val, tx->base_addr); tx 48 drivers/clk/mediatek/clk-apmixed.c writel(val, tx->base_addr); tx 51 drivers/clk/mediatek/clk-apmixed.c writel(val, tx->base_addr); tx 58 drivers/clk/mediatek/clk-apmixed.c struct mtk_ref2usb_tx *tx = to_mtk_ref2usb_tx(hw); tx 61 drivers/clk/mediatek/clk-apmixed.c val = readl(tx->base_addr); tx 63 drivers/clk/mediatek/clk-apmixed.c writel(val, tx->base_addr); tx 75 drivers/clk/mediatek/clk-apmixed.c struct mtk_ref2usb_tx *tx; tx 79 drivers/clk/mediatek/clk-apmixed.c tx = kzalloc(sizeof(*tx), GFP_KERNEL); tx 80 drivers/clk/mediatek/clk-apmixed.c if (!tx) tx 83 drivers/clk/mediatek/clk-apmixed.c tx->base_addr = reg; tx 84 drivers/clk/mediatek/clk-apmixed.c tx->hw.init = &init; tx 91 drivers/clk/mediatek/clk-apmixed.c clk = clk_register(NULL, &tx->hw); tx 95 drivers/clk/mediatek/clk-apmixed.c kfree(tx); tx 50 drivers/clk/tegra/clk-bpmp.c } tx; tx 77 drivers/clk/tegra/clk-bpmp.c memcpy(req + 4, clk->tx.data, clk->tx.size); tx 81 drivers/clk/tegra/clk-bpmp.c msg.tx.data = &request; tx 82 drivers/clk/tegra/clk-bpmp.c msg.tx.size = sizeof(request); tx 155 drivers/clk/tegra/clk-bpmp.c msg.tx.data = &request; tx 156 drivers/clk/tegra/clk-bpmp.c msg.tx.size = sizeof(request); tx 182 drivers/clk/tegra/clk-bpmp.c msg.tx.data = &request; tx 183 drivers/clk/tegra/clk-bpmp.c msg.tx.size = sizeof(request); tx 208 drivers/clk/tegra/clk-bpmp.c msg.tx.data = &request; tx 209 drivers/clk/tegra/clk-bpmp.c msg.tx.size = sizeof(request); tx 264 drivers/clk/tegra/clk-bpmp.c msg.tx.data = &request; tx 265 drivers/clk/tegra/clk-bpmp.c msg.tx.size = sizeof(request); tx 128 drivers/cpufreq/tegra186-cpufreq.c msg.tx.data = &req; tx 129 drivers/cpufreq/tegra186-cpufreq.c msg.tx.size = sizeof(req); tx 586 drivers/crypto/omap-sham.c struct dma_async_tx_descriptor *tx; tx 610 drivers/crypto/omap-sham.c tx = dmaengine_prep_slave_sg(dd->dma_lch, ctx->sg, ctx->sg_len, tx 614 drivers/crypto/omap-sham.c if (!tx) { tx 619 drivers/crypto/omap-sham.c tx->callback = omap_sham_dma_callback; tx 620 drivers/crypto/omap-sham.c tx->callback_param = dd; tx 632 drivers/crypto/omap-sham.c dmaengine_submit(tx); tx 200 drivers/dma/altera-msgdma.c #define tx_to_desc(tx) container_of(tx, struct msgdma_sw_desc, async_tx) tx 301 drivers/dma/altera-msgdma.c static dma_cookie_t msgdma_tx_submit(struct dma_async_tx_descriptor *tx) tx 303 drivers/dma/altera-msgdma.c struct msgdma_device *mdev = to_mdev(tx->chan); tx 308 drivers/dma/altera-msgdma.c new = tx_to_desc(tx); tx 310 drivers/dma/altera-msgdma.c cookie = dma_cookie_assign(tx); tx 320 drivers/dma/amba-pl08x.c static inline struct pl08x_txd *to_pl08x_txd(struct dma_async_tx_descriptor *tx) tx 322 drivers/dma/amba-pl08x.c return container_of(tx, struct pl08x_txd, vd.tx); tx 532 drivers/dma/amba-pl08x.c struct pl08x_txd *txd = to_pl08x_txd(&vd->tx); tx 1510 drivers/dma/amba-pl08x.c struct pl08x_txd *txd = to_pl08x_txd(&vd->tx); tx 1511 drivers/dma/amba-pl08x.c struct pl08x_dma_chan *plchan = to_pl08x_chan(vd->tx.chan); tx 1513 drivers/dma/amba-pl08x.c dma_descriptor_unmap(&vd->tx); tx 1580 drivers/dma/amba-pl08x.c struct pl08x_txd *txd = to_pl08x_txd(&vd->tx); tx 2316 drivers/dma/amba-pl08x.c struct pl08x_txd *tx; tx 2326 drivers/dma/amba-pl08x.c tx = plchan->at; tx 2327 drivers/dma/amba-pl08x.c if (tx && tx->cyclic) { tx 2328 drivers/dma/amba-pl08x.c vchan_cyclic_callback(&tx->vd); tx 2329 drivers/dma/amba-pl08x.c } else if (tx) { tx 2336 drivers/dma/amba-pl08x.c tx->done = true; tx 2337 drivers/dma/amba-pl08x.c vchan_cookie_complete(&tx->vd); tx 59 drivers/dma/at_hdmac.c static dma_cookie_t atc_tx_submit(struct dma_async_tx_descriptor *tx); tx 667 drivers/dma/at_hdmac.c static dma_cookie_t atc_tx_submit(struct dma_async_tx_descriptor *tx) tx 669 drivers/dma/at_hdmac.c struct at_desc *desc = txd_to_at_desc(tx); tx 670 drivers/dma/at_hdmac.c struct at_dma_chan *atchan = to_at_dma_chan(tx->chan); tx 675 drivers/dma/at_hdmac.c cookie = dma_cookie_assign(tx); tx 678 drivers/dma/at_hdmac.c dev_vdbg(chan2dev(tx->chan), "tx_submit: started %u\n", tx 683 drivers/dma/at_hdmac.c dev_vdbg(chan2dev(tx->chan), "tx_submit: queued %u\n", tx 420 drivers/dma/at_xdmac.c static dma_cookie_t at_xdmac_tx_submit(struct dma_async_tx_descriptor *tx) tx 422 drivers/dma/at_xdmac.c struct at_xdmac_desc *desc = txd_to_at_desc(tx); tx 423 drivers/dma/at_xdmac.c struct at_xdmac_chan *atchan = to_at_xdmac_chan(tx->chan); tx 428 drivers/dma/at_xdmac.c cookie = dma_cookie_assign(tx); tx 430 drivers/dma/at_xdmac.c dev_vdbg(chan2dev(tx->chan), "%s: atchan 0x%p, add desc 0x%p to xfers_list\n", tx 93 drivers/dma/bcm-sba-raid.c #define to_sba_request(tx) \ tx 94 drivers/dma/bcm-sba-raid.c container_of(tx, struct sba_request, tx) tx 121 drivers/dma/bcm-sba-raid.c struct dma_async_tx_descriptor tx; tx 211 drivers/dma/bcm-sba-raid.c if (async_tx_test_ack(&req->tx)) { tx 235 drivers/dma/bcm-sba-raid.c dma_async_tx_descriptor_init(&req->tx, &sba->dma_chan); tx 236 drivers/dma/bcm-sba-raid.c async_tx_ack(&req->tx); tx 419 drivers/dma/bcm-sba-raid.c struct dma_async_tx_descriptor *tx; tx 424 drivers/dma/bcm-sba-raid.c tx = &first->tx; tx 426 drivers/dma/bcm-sba-raid.c WARN_ON(tx->cookie < 0); tx 427 drivers/dma/bcm-sba-raid.c if (tx->cookie > 0) { tx 429 drivers/dma/bcm-sba-raid.c dma_cookie_complete(tx); tx 431 drivers/dma/bcm-sba-raid.c dmaengine_desc_get_callback_invoke(tx, NULL); tx 432 drivers/dma/bcm-sba-raid.c dma_descriptor_unmap(tx); tx 433 drivers/dma/bcm-sba-raid.c tx->callback = NULL; tx 434 drivers/dma/bcm-sba-raid.c tx->callback_result = NULL; tx 437 drivers/dma/bcm-sba-raid.c dma_run_dependencies(tx); tx 467 drivers/dma/bcm-sba-raid.c if (async_tx_test_ack(&req->tx)) tx 523 drivers/dma/bcm-sba-raid.c static dma_cookie_t sba_tx_submit(struct dma_async_tx_descriptor *tx) tx 530 drivers/dma/bcm-sba-raid.c if (unlikely(!tx)) tx 533 drivers/dma/bcm-sba-raid.c sba = to_sba_device(tx->chan); tx 534 drivers/dma/bcm-sba-raid.c req = to_sba_request(tx); tx 538 drivers/dma/bcm-sba-raid.c cookie = dma_cookie_assign(tx); tx 569 drivers/dma/bcm-sba-raid.c dma_addr_t resp_dma = req->tx.phys; tx 643 drivers/dma/bcm-sba-raid.c req->tx.flags = flags; tx 644 drivers/dma/bcm-sba-raid.c req->tx.cookie = -EBUSY; tx 646 drivers/dma/bcm-sba-raid.c return &req->tx; tx 657 drivers/dma/bcm-sba-raid.c dma_addr_t resp_dma = req->tx.phys; tx 729 drivers/dma/bcm-sba-raid.c req->tx.flags = flags; tx 730 drivers/dma/bcm-sba-raid.c req->tx.cookie = -EBUSY; tx 765 drivers/dma/bcm-sba-raid.c return (first) ? &first->tx : NULL; tx 777 drivers/dma/bcm-sba-raid.c dma_addr_t resp_dma = req->tx.phys; tx 868 drivers/dma/bcm-sba-raid.c req->tx.flags = flags; tx 869 drivers/dma/bcm-sba-raid.c req->tx.cookie = -EBUSY; tx 908 drivers/dma/bcm-sba-raid.c return (first) ? &first->tx : NULL; tx 922 drivers/dma/bcm-sba-raid.c dma_addr_t resp_dma = req->tx.phys; tx 1080 drivers/dma/bcm-sba-raid.c req->tx.flags = flags; tx 1081 drivers/dma/bcm-sba-raid.c req->tx.cookie = -EBUSY; tx 1097 drivers/dma/bcm-sba-raid.c dma_addr_t resp_dma = req->tx.phys; tx 1335 drivers/dma/bcm-sba-raid.c req->tx.flags = flags; tx 1336 drivers/dma/bcm-sba-raid.c req->tx.cookie = -EBUSY; tx 1432 drivers/dma/bcm-sba-raid.c return (first) ? &first->tx : NULL; tx 1520 drivers/dma/bcm-sba-raid.c dma_async_tx_descriptor_init(&req->tx, &sba->dma_chan); tx 1521 drivers/dma/bcm-sba-raid.c async_tx_ack(&req->tx); tx 1522 drivers/dma/bcm-sba-raid.c req->tx.tx_submit = sba_tx_submit; tx 1523 drivers/dma/bcm-sba-raid.c req->tx.phys = sba->resp_dma_base + i * sba->hw_resp_size; tx 203 drivers/dma/bcm2835-dma.c return container_of(t, struct bcm2835_desc, vd.tx); tx 452 drivers/dma/bcm2835-dma.c c->desc = d = to_bcm2835_dma_desc(&vd->tx); tx 580 drivers/dma/bcm2835-dma.c bcm2835_dma_desc_size(to_bcm2835_dma_desc(&vd->tx)); tx 581 drivers/dma/bcm2835-dma.c } else if (c->desc && c->desc->vd.tx.cookie == cookie) { tx 800 drivers/dma/bcm2835-dma.c if (c->desc->vd.tx.flags & DMA_PREP_INTERRUPT) tx 2195 drivers/dma/coh901318.c coh901318_tx_submit(struct dma_async_tx_descriptor *tx) tx 2197 drivers/dma/coh901318.c struct coh901318_desc *cohd = container_of(tx, struct coh901318_desc, tx 2199 drivers/dma/coh901318.c struct coh901318_chan *cohc = to_coh901318_chan(tx->chan); tx 2204 drivers/dma/coh901318.c cookie = dma_cookie_assign(tx); tx 253 drivers/dma/dma-axi-dmac.c if (chan->hw_cyclic && desc->cyclic && !desc->vdesc.tx.callback && tx 257 drivers/dma/dma-jz4780.c struct jz4780_dma_chan *jzchan = to_jz4780_dma_chan(vdesc->tx.chan); tx 495 drivers/dma/dma-jz4780.c if (jzchan->desc->type == DMA_CYCLIC && vdesc->tx.callback) { tx 653 drivers/dma/dma-jz4780.c } else if (cookie == jzchan->desc->vdesc.tx.cookie) { tx 1297 drivers/dma/dmaengine.c void dma_async_tx_descriptor_init(struct dma_async_tx_descriptor *tx, tx 1300 drivers/dma/dmaengine.c tx->chan = chan; tx 1302 drivers/dma/dmaengine.c spin_lock_init(&tx->lock); tx 1311 drivers/dma/dmaengine.c dma_wait_for_async_tx(struct dma_async_tx_descriptor *tx) tx 1315 drivers/dma/dmaengine.c if (!tx) tx 1318 drivers/dma/dmaengine.c while (tx->cookie == -EBUSY) { tx 1320 drivers/dma/dmaengine.c dev_err(tx->chan->device->dev, tx 1327 drivers/dma/dmaengine.c return dma_sync_wait(tx->chan, tx->cookie); tx 1335 drivers/dma/dmaengine.c void dma_run_dependencies(struct dma_async_tx_descriptor *tx) tx 1337 drivers/dma/dmaengine.c struct dma_async_tx_descriptor *dep = txd_next(tx); tx 1345 drivers/dma/dmaengine.c txd_clear_next(tx); tx 29 drivers/dma/dmaengine.h static inline dma_cookie_t dma_cookie_assign(struct dma_async_tx_descriptor *tx) tx 31 drivers/dma/dmaengine.h struct dma_chan *chan = tx->chan; tx 37 drivers/dma/dmaengine.h tx->cookie = chan->cookie = cookie; tx 52 drivers/dma/dmaengine.h static inline void dma_cookie_complete(struct dma_async_tx_descriptor *tx) tx 54 drivers/dma/dmaengine.h BUG_ON(tx->cookie < DMA_MIN_COOKIE); tx 55 drivers/dma/dmaengine.h tx->chan->completed_cookie = tx->cookie; tx 56 drivers/dma/dmaengine.h tx->cookie = 0; tx 106 drivers/dma/dmaengine.h dmaengine_desc_get_callback(struct dma_async_tx_descriptor *tx, tx 109 drivers/dma/dmaengine.h cb->callback = tx->callback; tx 110 drivers/dma/dmaengine.h cb->callback_result = tx->callback_result; tx 111 drivers/dma/dmaengine.h cb->callback_param = tx->callback_param; tx 152 drivers/dma/dmaengine.h dmaengine_desc_get_callback_invoke(struct dma_async_tx_descriptor *tx, tx 157 drivers/dma/dmaengine.h dmaengine_desc_get_callback(tx, &cb); tx 667 drivers/dma/dmatest.c struct dma_async_tx_descriptor *tx = NULL; tx 762 drivers/dma/dmatest.c tx = dev->device_prep_dma_memcpy(chan, tx 766 drivers/dma/dmatest.c tx = dev->device_prep_dma_memset(chan, tx 771 drivers/dma/dmatest.c tx = dev->device_prep_dma_xor(chan, tx 778 drivers/dma/dmatest.c tx = dev->device_prep_dma_pq(chan, dma_pq, srcs, tx 783 drivers/dma/dmatest.c if (!tx) { tx 792 drivers/dma/dmatest.c tx->callback = dmatest_callback; tx 793 drivers/dma/dmatest.c tx->callback_param = done; tx 795 drivers/dma/dmatest.c cookie = tx->tx_submit(tx); tx 213 drivers/dma/dw-axi-dmac/dw-axi-dmac-platform.c desc->vd.tx.phys = phys; tx 228 drivers/dma/dw-axi-dmac/dw-axi-dmac-platform.c dma_pool_free(dw->desc_pool, child, child->vd.tx.phys); tx 232 drivers/dma/dw-axi-dmac/dw-axi-dmac-platform.c dma_pool_free(dw->desc_pool, desc, desc->vd.tx.phys); tx 298 drivers/dma/dw-axi-dmac/dw-axi-dmac-platform.c write_chan_llp(chan, first->vd.tx.phys | lms); tx 321 drivers/dma/dw-axi-dmac/dw-axi-dmac-platform.c vd->tx.cookie); tx 495 drivers/dma/dw-axi-dmac/dw-axi-dmac-platform.c write_desc_llp(prev, desc->vd.tx.phys | lms); tx 560 drivers/dma/dw-axi-dmac/dw-axi-dmac-platform.c axi_chan_name(chan), vd->tx.cookie, status); tx 56 drivers/dma/dw/core.c static dma_cookie_t dwc_tx_submit(struct dma_async_tx_descriptor *tx) tx 58 drivers/dma/dw/core.c struct dw_desc *desc = txd_to_dw_desc(tx); tx 59 drivers/dma/dw/core.c struct dw_dma_chan *dwc = to_dw_dma_chan(tx->chan); tx 64 drivers/dma/dw/core.c cookie = dma_cookie_assign(tx); tx 74 drivers/dma/dw/core.c dev_vdbg(chan2dev(tx->chan), "%s: queued %u\n", tx 829 drivers/dma/ep93xx_dma.c static dma_cookie_t ep93xx_dma_tx_submit(struct dma_async_tx_descriptor *tx) tx 831 drivers/dma/ep93xx_dma.c struct ep93xx_dma_chan *edmac = to_ep93xx_dma_chan(tx->chan); tx 837 drivers/dma/ep93xx_dma.c cookie = dma_cookie_assign(tx); tx 839 drivers/dma/ep93xx_dma.c desc = container_of(tx, struct ep93xx_dma_desc, txd); tx 325 drivers/dma/fsl-edma-common.c if (fsl_chan->edesc && cookie == fsl_chan->edesc->vdesc.tx.cookie) tx 83 drivers/dma/fsl_raid.c #define to_fsl_re_dma_desc(tx) container_of(tx, struct fsl_re_desc, async_tx) tx 86 drivers/dma/fsl_raid.c static dma_cookie_t fsl_re_tx_submit(struct dma_async_tx_descriptor *tx) tx 93 drivers/dma/fsl_raid.c desc = to_fsl_re_dma_desc(tx); tx 94 drivers/dma/fsl_raid.c re_chan = container_of(tx->chan, struct fsl_re_chan, chan); tx 97 drivers/dma/fsl_raid.c cookie = dma_cookie_assign(tx); tx 430 drivers/dma/fsl_raid.c struct dma_async_tx_descriptor *tx; tx 438 drivers/dma/fsl_raid.c tx = fsl_re_prep_dma_genq(chan, dest[1], dma_src, 2, coef, len, tx 440 drivers/dma/fsl_raid.c if (tx) tx 441 drivers/dma/fsl_raid.c desc = to_fsl_re_dma_desc(tx); tx 443 drivers/dma/fsl_raid.c return tx; tx 406 drivers/dma/fsldma.c static dma_cookie_t fsl_dma_tx_submit(struct dma_async_tx_descriptor *tx) tx 408 drivers/dma/fsldma.c struct fsldma_chan *chan = to_fsl_chan(tx->chan); tx 409 drivers/dma/fsldma.c struct fsl_desc_sw *desc = tx_to_fsl_desc(tx); tx 192 drivers/dma/fsldma.h #define tx_to_fsl_desc(tx) container_of(tx, struct fsl_desc_sw, async_tx) tx 330 drivers/dma/hsu/hsu.c if (hsuc->desc && cookie == hsuc->desc->vdesc.tx.cookie) { tx 224 drivers/dma/idma64.c struct idma64_chan *idma64c = to_idma64_chan(vdesc->tx.chan); tx 380 drivers/dma/idma64.c if (idma64c->desc && cookie == idma64c->desc->vdesc.tx.cookie) { tx 176 drivers/dma/img-mdc-dma.c struct virt_dma_desc *vdesc = container_of(t, struct virt_dma_desc, tx); tx 279 drivers/dma/img-mdc-dma.c struct mdc_tx_desc *mdesc = to_mdc_desc(&vd->tx); tx 537 drivers/dma/img-mdc-dma.c mdesc = to_mdc_desc(&vd->tx); tx 591 drivers/dma/img-mdc-dma.c mdesc = to_mdc_desc(&vd->tx); tx 593 drivers/dma/img-mdc-dma.c } else if (mchan->desc && mchan->desc->vd.tx.cookie == cookie) { tx 744 drivers/dma/imx-dma.c static dma_cookie_t imxdma_tx_submit(struct dma_async_tx_descriptor *tx) tx 746 drivers/dma/imx-dma.c struct imxdma_channel *imxdmac = to_imxdma_chan(tx->chan); tx 753 drivers/dma/imx-dma.c cookie = dma_cookie_assign(tx); tx 748 drivers/dma/imx-sdma.c return container_of(t, struct sdma_desc, vd.tx); tx 762 drivers/dma/imx-sdma.c sdmac->desc = desc = to_sdma_desc(&vd->tx); tx 817 drivers/dma/imx-sdma.c dmaengine_desc_get_callback_invoke(&desc->vd.tx, NULL); tx 1665 drivers/dma/imx-sdma.c desc = to_sdma_desc(&vd->tx); tx 1671 drivers/dma/imx-sdma.c } else if (sdmac->desc && sdmac->desc->vd.tx.cookie == cookie) { tx 297 drivers/dma/ioat/dma.c static dma_cookie_t ioat_tx_submit_unlock(struct dma_async_tx_descriptor *tx) tx 300 drivers/dma/ioat/dma.c struct dma_chan *c = tx->chan; tx 304 drivers/dma/ioat/dma.c cookie = dma_cookie_assign(tx); tx 599 drivers/dma/ioat/dma.c struct dma_async_tx_descriptor *tx; tx 609 drivers/dma/ioat/dma.c tx = &desc->txd; tx 610 drivers/dma/ioat/dma.c if (tx->cookie) { tx 611 drivers/dma/ioat/dma.c dma_cookie_complete(tx); tx 612 drivers/dma/ioat/dma.c dma_descriptor_unmap(tx); tx 613 drivers/dma/ioat/dma.c dmaengine_desc_get_callback_invoke(tx, NULL); tx 614 drivers/dma/ioat/dma.c tx->callback = NULL; tx 615 drivers/dma/ioat/dma.c tx->callback_result = NULL; tx 618 drivers/dma/ioat/dma.c if (tx->phys == phys_complete) tx 722 drivers/dma/ioat/dma.c struct dma_async_tx_descriptor *tx; tx 727 drivers/dma/ioat/dma.c tx = &desc->txd; tx 728 drivers/dma/ioat/dma.c if (tx->cookie) { tx 731 drivers/dma/ioat/dma.c dma_cookie_complete(tx); tx 732 drivers/dma/ioat/dma.c dma_descriptor_unmap(tx); tx 734 drivers/dma/ioat/dma.c dmaengine_desc_get_callback_invoke(tx, &res); tx 735 drivers/dma/ioat/dma.c tx->callback = NULL; tx 736 drivers/dma/ioat/dma.c tx->callback_result = NULL; tx 763 drivers/dma/ioat/dma.c struct dma_async_tx_descriptor *tx; tx 832 drivers/dma/ioat/dma.c tx = &desc->txd; tx 833 drivers/dma/ioat/dma.c if (tx->cookie) { tx 834 drivers/dma/ioat/dma.c dma_cookie_complete(tx); tx 835 drivers/dma/ioat/dma.c dma_descriptor_unmap(tx); tx 836 drivers/dma/ioat/dma.c dmaengine_desc_get_callback_invoke(tx, &res); tx 837 drivers/dma/ioat/dma.c tx->callback = NULL; tx 838 drivers/dma/ioat/dma.c tx->callback_result = NULL; tx 218 drivers/dma/ioat/dma.h struct dma_async_tx_descriptor *tx, int id) tx 224 drivers/dma/ioat/dma.h (unsigned long long) tx->phys, tx 225 drivers/dma/ioat/dma.h (unsigned long long) hw->next, tx->cookie, tx->flags, tx 308 drivers/dma/ioat/init.c struct dma_async_tx_descriptor *tx; tx 351 drivers/dma/ioat/init.c tx = ioat_dma->dma_dev.device_prep_dma_memcpy(dma_chan, dma_dest, tx 354 drivers/dma/ioat/init.c if (!tx) { tx 360 drivers/dma/ioat/init.c async_tx_ack(tx); tx 362 drivers/dma/ioat/init.c tx->callback = ioat_dma_test_callback; tx 363 drivers/dma/ioat/init.c tx->callback_param = &cmp; tx 364 drivers/dma/ioat/init.c cookie = tx->tx_submit(tx); tx 793 drivers/dma/ioat/init.c struct dma_async_tx_descriptor *tx; tx 867 drivers/dma/ioat/init.c tx = dma->device_prep_dma_xor(dma_chan, dest_dma, dma_srcs, tx 871 drivers/dma/ioat/init.c if (!tx) { tx 877 drivers/dma/ioat/init.c async_tx_ack(tx); tx 879 drivers/dma/ioat/init.c tx->callback = ioat_dma_test_callback; tx 880 drivers/dma/ioat/init.c tx->callback_param = &cmp; tx 881 drivers/dma/ioat/init.c cookie = tx->tx_submit(tx); tx 936 drivers/dma/ioat/init.c tx = dma->device_prep_dma_xor_val(dma_chan, dma_srcs, tx 939 drivers/dma/ioat/init.c if (!tx) { tx 945 drivers/dma/ioat/init.c async_tx_ack(tx); tx 947 drivers/dma/ioat/init.c tx->callback = ioat_dma_test_callback; tx 948 drivers/dma/ioat/init.c tx->callback_param = &cmp; tx 949 drivers/dma/ioat/init.c cookie = tx->tx_submit(tx); tx 989 drivers/dma/ioat/init.c tx = dma->device_prep_dma_xor_val(dma_chan, dma_srcs, tx 992 drivers/dma/ioat/init.c if (!tx) { tx 998 drivers/dma/ioat/init.c async_tx_ack(tx); tx 1000 drivers/dma/ioat/init.c tx->callback = ioat_dma_test_callback; tx 1001 drivers/dma/ioat/init.c tx->callback_param = &cmp; tx 1002 drivers/dma/ioat/init.c cookie = tx->tx_submit(tx); tx 31 drivers/dma/iop-adma.c #define tx_to_iop_adma_slot(tx) \ tx 32 drivers/dma/iop-adma.c container_of(tx, struct iop_adma_desc_slot, async_tx) tx 55 drivers/dma/iop-adma.c struct dma_async_tx_descriptor *tx = &desc->async_tx; tx 57 drivers/dma/iop-adma.c BUG_ON(tx->cookie < 0); tx 58 drivers/dma/iop-adma.c if (tx->cookie > 0) { tx 59 drivers/dma/iop-adma.c cookie = tx->cookie; tx 60 drivers/dma/iop-adma.c tx->cookie = 0; tx 65 drivers/dma/iop-adma.c dmaengine_desc_get_callback_invoke(tx, NULL); tx 67 drivers/dma/iop-adma.c dma_descriptor_unmap(tx); tx 73 drivers/dma/iop-adma.c dma_run_dependencies(tx); tx 361 drivers/dma/iop-adma.c iop_adma_tx_submit(struct dma_async_tx_descriptor *tx) tx 363 drivers/dma/iop-adma.c struct iop_adma_desc_slot *sw_desc = tx_to_iop_adma_slot(tx); tx 364 drivers/dma/iop-adma.c struct iop_adma_chan *iop_chan = to_iop_adma_chan(tx->chan); tx 374 drivers/dma/iop-adma.c cookie = dma_cookie_assign(tx); tx 838 drivers/dma/iop-adma.c struct dma_async_tx_descriptor *tx; tx 870 drivers/dma/iop-adma.c tx = iop_adma_prep_dma_memcpy(dma_chan, dest_dma, src_dma, tx 874 drivers/dma/iop-adma.c cookie = iop_adma_tx_submit(tx); tx 914 drivers/dma/iop-adma.c struct dma_async_tx_descriptor *tx; tx 970 drivers/dma/iop-adma.c tx = iop_adma_prep_dma_xor(dma_chan, dest_dma, dma_srcs, tx 974 drivers/dma/iop-adma.c cookie = iop_adma_tx_submit(tx); tx 1016 drivers/dma/iop-adma.c tx = iop_adma_prep_dma_xor_val(dma_chan, dma_srcs, tx 1021 drivers/dma/iop-adma.c cookie = iop_adma_tx_submit(tx); tx 1045 drivers/dma/iop-adma.c tx = iop_adma_prep_dma_xor_val(dma_chan, dma_srcs, tx 1050 drivers/dma/iop-adma.c cookie = iop_adma_tx_submit(tx); tx 1092 drivers/dma/iop-adma.c struct dma_async_tx_descriptor *tx; tx 1139 drivers/dma/iop-adma.c tx = iop_adma_prep_dma_pq(dma_chan, pq_dest, pq_src, tx 1145 drivers/dma/iop-adma.c cookie = iop_adma_tx_submit(tx); tx 1177 drivers/dma/iop-adma.c tx = iop_adma_prep_dma_pq_val(dma_chan, &pq_src[IOP_ADMA_NUM_SRC_TEST], tx 1182 drivers/dma/iop-adma.c cookie = iop_adma_tx_submit(tx); tx 1209 drivers/dma/iop-adma.c tx = iop_adma_prep_dma_pq_val(dma_chan, &pq_src[IOP_ADMA_NUM_SRC_TEST], tx 1214 drivers/dma/iop-adma.c cookie = iop_adma_tx_submit(tx); tx 819 drivers/dma/ipu/ipu_idmac.c static dma_cookie_t idmac_tx_submit(struct dma_async_tx_descriptor *tx) tx 821 drivers/dma/ipu/ipu_idmac.c struct idmac_tx_desc *desc = to_tx_desc(tx); tx 822 drivers/dma/ipu/ipu_idmac.c struct idmac_channel *ichan = to_idmac_chan(tx->chan); tx 823 drivers/dma/ipu/ipu_idmac.c struct idmac *idmac = to_idmac(tx->chan->device); tx 833 drivers/dma/ipu/ipu_idmac.c dev_err(dev, "Descriptor %p not prepared!\n", tx); tx 839 drivers/dma/ipu/ipu_idmac.c async_tx_clear_ack(tx); tx 866 drivers/dma/ipu/ipu_idmac.c cookie = dma_cookie_assign(tx); tx 897 drivers/dma/ipu/ipu_idmac.c tx->cookie = cookie; tx 715 drivers/dma/k3dma.c struct k3_dma_dev *d = to_k3_dma(vd->tx.chan->device); tx 372 drivers/dma/mediatek/mtk-cqdma.c dma_run_dependencies(&cvd->vd.tx); tx 428 drivers/dma/mediatek/mtk-cqdma.c if (vd->tx.cookie == cookie) { tx 435 drivers/dma/mediatek/mtk-cqdma.c if (vd->tx.cookie == cookie) tx 492 drivers/dma/mediatek/mtk-cqdma.c struct dma_async_tx_descriptor *tx = NULL, *prev_tx = NULL; tx 529 drivers/dma/mediatek/mtk-cqdma.c tx = vchan_tx_prep(to_virt_chan(c), &cvd[i]->vd, flags); tx 530 drivers/dma/mediatek/mtk-cqdma.c tx->next = NULL; tx 535 drivers/dma/mediatek/mtk-cqdma.c prev_tx->next = tx; tx 545 drivers/dma/mediatek/mtk-cqdma.c prev_tx = tx; tx 548 drivers/dma/mediatek/mtk-cqdma.c return &cvd[0]->vd.tx; tx 584 drivers/dma/mediatek/mtk-hsdma.c hvc = to_hsdma_vchan(cb->vd->tx.chan); tx 667 drivers/dma/mediatek/mtk-hsdma.c if (vd->tx.cookie == cookie) tx 671 drivers/dma/mediatek/mtk-hsdma.c if (vd->tx.cookie == cookie) tx 118 drivers/dma/mediatek/mtk-uart-apdma.c return container_of(t, struct mtk_uart_apdma_desc, vd.tx); tx 134 drivers/dma/mediatek/mtk-uart-apdma.c struct dma_chan *chan = vd->tx.chan; tx 371 drivers/dma/mediatek/mtk-uart-apdma.c c->desc = to_mtk_uart_apdma_desc(&vd->tx); tx 80 drivers/dma/mic_x100_dma.c struct dma_async_tx_descriptor *tx; tx 93 drivers/dma/mic_x100_dma.c tx = &ch->tx_array[last_tail]; tx 94 drivers/dma/mic_x100_dma.c if (tx->cookie) { tx 95 drivers/dma/mic_x100_dma.c dma_cookie_complete(tx); tx 96 drivers/dma/mic_x100_dma.c dmaengine_desc_get_callback_invoke(tx, NULL); tx 97 drivers/dma/mic_x100_dma.c tx->callback = NULL; tx 238 drivers/dma/mic_x100_dma.c static dma_cookie_t mic_dma_tx_submit_unlock(struct dma_async_tx_descriptor *tx) tx 240 drivers/dma/mic_x100_dma.c struct mic_dma_chan *mic_ch = to_mic_dma_chan(tx->chan); tx 243 drivers/dma/mic_x100_dma.c dma_cookie_assign(tx); tx 244 drivers/dma/mic_x100_dma.c cookie = tx->cookie; tx 261 drivers/dma/mic_x100_dma.c struct dma_async_tx_descriptor *tx = &ch->tx_array[idx]; tx 263 drivers/dma/mic_x100_dma.c dma_async_tx_descriptor_init(tx, &ch->api_ch); tx 264 drivers/dma/mic_x100_dma.c tx->tx_submit = mic_dma_tx_submit_unlock; tx 265 drivers/dma/mic_x100_dma.c return tx; tx 499 drivers/dma/mic_x100_dma.c struct dma_async_tx_descriptor *tx; tx 503 drivers/dma/mic_x100_dma.c tx = mic_dma_prep_memcpy_lock(&ch->api_ch, 0, 0, 0, DMA_PREP_FENCE); tx 504 drivers/dma/mic_x100_dma.c if (!tx) { tx 509 drivers/dma/mic_x100_dma.c cookie = tx->tx_submit(tx); tx 132 drivers/dma/mmp_pdma.c #define tx_to_mmp_pdma_desc(tx) \ tx 133 drivers/dma/mmp_pdma.c container_of(tx, struct mmp_pdma_desc_sw, async_tx) tx 341 drivers/dma/mmp_pdma.c static dma_cookie_t mmp_pdma_tx_submit(struct dma_async_tx_descriptor *tx) tx 343 drivers/dma/mmp_pdma.c struct mmp_pdma_chan *chan = to_mmp_pdma_chan(tx->chan); tx 344 drivers/dma/mmp_pdma.c struct mmp_pdma_desc_sw *desc = tx_to_mmp_pdma_desc(tx); tx 372 drivers/dma/mmp_tdma.c static dma_cookie_t mmp_tdma_tx_submit(struct dma_async_tx_descriptor *tx) tx 374 drivers/dma/mmp_tdma.c struct mmp_tdma_chan *tdmac = to_mmp_tdma_chan(tx->chan); tx 178 drivers/dma/moxart-dma.c return container_of(t, struct moxart_desc, vd.tx); tx 433 drivers/dma/moxart-dma.c ch->desc = to_moxart_dma_desc(&vd->tx); tx 495 drivers/dma/moxart-dma.c d = to_moxart_dma_desc(&vd->tx); tx 497 drivers/dma/moxart-dma.c } else if (ch->desc && ch->desc->vd.tx.cookie == cookie) { tx 42 drivers/dma/mv_xor.c #define to_mv_xor_slot(tx) \ tx 43 drivers/dma/mv_xor.c container_of(tx, struct mv_xor_desc_slot, async_tx) tx 382 drivers/dma/mv_xor.c mv_xor_tx_submit(struct dma_async_tx_descriptor *tx) tx 384 drivers/dma/mv_xor.c struct mv_xor_desc_slot *sw_desc = to_mv_xor_slot(tx); tx 385 drivers/dma/mv_xor.c struct mv_xor_chan *mv_chan = to_mv_xor_chan(tx->chan); tx 395 drivers/dma/mv_xor.c cookie = dma_cookie_assign(tx); tx 768 drivers/dma/mv_xor.c struct dma_async_tx_descriptor *tx; tx 823 drivers/dma/mv_xor.c tx = mv_xor_prep_dma_memcpy(dma_chan, dest_dma, src_dma, tx 825 drivers/dma/mv_xor.c if (!tx) { tx 832 drivers/dma/mv_xor.c cookie = mv_xor_tx_submit(tx); tx 841 drivers/dma/mv_xor.c async_tx_ack(tx); tx 879 drivers/dma/mv_xor.c struct dma_async_tx_descriptor *tx; tx 956 drivers/dma/mv_xor.c tx = mv_xor_prep_dma_xor(dma_chan, dest_dma, dma_srcs, tx 958 drivers/dma/mv_xor.c if (!tx) { tx 965 drivers/dma/mv_xor.c cookie = mv_xor_tx_submit(tx); tx 974 drivers/dma/mv_xor.c async_tx_ack(tx); tx 299 drivers/dma/mv_xor_v2.c mv_xor_v2_tx_submit(struct dma_async_tx_descriptor *tx) tx 304 drivers/dma/mv_xor_v2.c container_of(tx, struct mv_xor_v2_sw_desc, async_tx); tx 306 drivers/dma/mv_xor_v2.c container_of(tx->chan, struct mv_xor_v2_device, dmachan); tx 314 drivers/dma/mv_xor_v2.c cookie = dma_cookie_assign(tx); tx 318 drivers/dma/mxs-dma.c static dma_cookie_t mxs_dma_tx_submit(struct dma_async_tx_descriptor *tx) tx 320 drivers/dma/mxs-dma.c return dma_cookie_assign(tx); tx 666 drivers/dma/nbpfaxi.c static dma_cookie_t nbpf_tx_submit(struct dma_async_tx_descriptor *tx) tx 668 drivers/dma/nbpfaxi.c struct nbpf_desc *desc = container_of(tx, struct nbpf_desc, async_tx); tx 674 drivers/dma/nbpfaxi.c cookie = dma_cookie_assign(tx); tx 293 drivers/dma/owl-dma.c static inline struct owl_dma_txd *to_owl_txd(struct dma_async_tx_descriptor *tx) tx 295 drivers/dma/owl-dma.c return container_of(tx, struct owl_dma_txd, vd.tx); tx 500 drivers/dma/owl-dma.c struct owl_dma_txd *txd = to_owl_txd(&vd->tx); tx 651 drivers/dma/owl-dma.c struct owl_dma *od = to_owl_dma(vd->tx.chan->device); tx 652 drivers/dma/owl-dma.c struct owl_dma_txd *txd = to_owl_txd(&vd->tx); tx 783 drivers/dma/owl-dma.c txd = to_owl_txd(&vd->tx); tx 2025 drivers/dma/pl330.c to_desc(struct dma_async_tx_descriptor *tx) tx 2027 drivers/dma/pl330.c return container_of(tx, struct dma_pl330_desc, txd); tx 2489 drivers/dma/pl330.c static dma_cookie_t pl330_tx_submit(struct dma_async_tx_descriptor *tx) tx 2491 drivers/dma/pl330.c struct dma_pl330_desc *desc, *last = to_desc(tx); tx 2492 drivers/dma/pl330.c struct dma_pl330_chan *pch = to_pchan(tx->chan); tx 1227 drivers/dma/ppc4xx/adma.c ppc440spe_adma_tx_submit(struct dma_async_tx_descriptor *tx); tx 1229 drivers/dma/ppc4xx/adma.c static void ppc440spe_adma_set_dest(struct ppc440spe_adma_desc_slot *tx, tx 1232 drivers/dma/ppc4xx/adma.c ppc440spe_adma_memcpy_xor_set_src(struct ppc440spe_adma_desc_slot *tx, tx 1236 drivers/dma/ppc4xx/adma.c ppc440spe_adma_pq_set_dest(struct ppc440spe_adma_desc_slot *tx, tx 1239 drivers/dma/ppc4xx/adma.c ppc440spe_adma_pq_set_src(struct ppc440spe_adma_desc_slot *tx, tx 1242 drivers/dma/ppc4xx/adma.c ppc440spe_adma_pq_set_src_mult(struct ppc440spe_adma_desc_slot *tx, tx 1245 drivers/dma/ppc4xx/adma.c ppc440spe_adma_pqzero_sum_set_dest(struct ppc440spe_adma_desc_slot *tx, tx 1897 drivers/dma/ppc4xx/adma.c static dma_cookie_t ppc440spe_adma_tx_submit(struct dma_async_tx_descriptor *tx) tx 1900 drivers/dma/ppc4xx/adma.c struct ppc440spe_adma_chan *chan = to_ppc440spe_adma_chan(tx->chan); tx 1906 drivers/dma/ppc4xx/adma.c sw_desc = tx_to_ppc440spe_adma_slot(tx); tx 1913 drivers/dma/ppc4xx/adma.c cookie = dma_cookie_assign(tx); tx 2768 drivers/dma/ppc4xx/adma.c struct dma_async_tx_descriptor *tx; tx 2776 drivers/dma/ppc4xx/adma.c tx = ppc440spe_adma_prep_dma_pqzero_sum(chan, pq, &src[1], tx 2779 drivers/dma/ppc4xx/adma.c return tx; tx 22 drivers/dma/ppc4xx/adma.h #define tx_to_ppc440spe_adma_slot(tx) \ tx 23 drivers/dma/ppc4xx/adma.h container_of(tx, struct ppc440spe_adma_desc_slot, async_tx) tx 136 drivers/dma/pxa_dma.c #define tx_to_pxad_desc(tx) \ tx 137 drivers/dma/pxa_dma.c container_of(tx, struct pxad_desc_sw, async_tx) tx 624 drivers/dma/pxa_dma.c __func__, vd, vd->tx.cookie, vd_completed, tx 626 drivers/dma/pxa_dma.c last_started = vd->tx.cookie; tx 776 drivers/dma/pxa_dma.c static dma_cookie_t pxad_tx_submit(struct dma_async_tx_descriptor *tx) tx 778 drivers/dma/pxa_dma.c struct virt_dma_chan *vc = to_virt_chan(tx->chan); tx 781 drivers/dma/pxa_dma.c *vd = container_of(tx, struct virt_dma_desc, tx); tx 785 drivers/dma/pxa_dma.c set_updater_desc(to_pxad_sw_desc(vd), tx->flags); tx 788 drivers/dma/pxa_dma.c cookie = dma_cookie_assign(tx); tx 839 drivers/dma/pxa_dma.c "%s(): txd %p[%x]", __func__, vd_first, vd_first->tx.cookie); tx 852 drivers/dma/pxa_dma.c struct dma_async_tx_descriptor *tx; tx 856 drivers/dma/pxa_dma.c tx = vchan_tx_prep(vc, vd, tx_flags); tx 857 drivers/dma/pxa_dma.c tx->tx_submit = pxad_tx_submit; tx 860 drivers/dma/pxa_dma.c vc, vd, vd->tx.cookie, tx 863 drivers/dma/pxa_dma.c return tx; tx 1103 drivers/dma/pxa_dma.c vd, vd->tx.cookie, is_desc_completed(vd)); tx 931 drivers/dma/qcom/bam_dma.c if (async_desc->vd.tx.cookie != cookie) tx 1028 drivers/dma/qcom/bam_dma.c dmaengine_desc_get_callback(&async_desc->vd.tx, &cb); tx 410 drivers/dma/s3c24xx-dma.c struct s3c24xx_txd *to_s3c24xx_txd(struct dma_async_tx_descriptor *tx) tx 412 drivers/dma/s3c24xx-dma.c return container_of(tx, struct s3c24xx_txd, vd.tx); tx 507 drivers/dma/s3c24xx-dma.c struct s3c24xx_txd *txd = to_s3c24xx_txd(&vd->tx); tx 637 drivers/dma/s3c24xx-dma.c struct s3c24xx_txd *txd = to_s3c24xx_txd(&vd->tx); tx 638 drivers/dma/s3c24xx-dma.c struct s3c24xx_dma_chan *s3cchan = to_s3c24xx_dma_chan(vd->tx.chan); tx 641 drivers/dma/s3c24xx-dma.c dma_descriptor_unmap(&vd->tx); tx 783 drivers/dma/s3c24xx-dma.c txd = to_s3c24xx_txd(&vd->tx); tx 155 drivers/dma/sa11x0-dma.c p->num, &txd->vd, txd->vd.tx.cookie, txd->ddar); tx 450 drivers/dma/sa11x0-dma.c if (p->txd_done && p->txd_done->vd.tx.cookie == cookie) tx 452 drivers/dma/sa11x0-dma.c else if (p->txd_load && p->txd_load->vd.tx.cookie == cookie) tx 460 drivers/dma/sh/rcar-dmac.c static dma_cookie_t rcar_dmac_tx_submit(struct dma_async_tx_descriptor *tx) tx 462 drivers/dma/sh/rcar-dmac.c struct rcar_dmac_chan *chan = to_rcar_dmac_chan(tx->chan); tx 463 drivers/dma/sh/rcar-dmac.c struct rcar_dmac_desc *desc = to_rcar_dmac_desc(tx); tx 469 drivers/dma/sh/rcar-dmac.c cookie = dma_cookie_assign(tx); tx 472 drivers/dma/sh/rcar-dmac.c chan->index, tx->cookie, desc); tx 70 drivers/dma/sh/shdma-base.c static dma_cookie_t shdma_tx_submit(struct dma_async_tx_descriptor *tx) tx 73 drivers/dma/sh/shdma-base.c container_of(tx, struct shdma_desc, async_tx); tx 74 drivers/dma/sh/shdma-base.c struct shdma_chan *schan = to_shdma_chan(tx->chan); tx 75 drivers/dma/sh/shdma-base.c dma_async_tx_callback callback = tx->callback; tx 83 drivers/dma/sh/shdma-base.c cookie = dma_cookie_assign(tx); tx 99 drivers/dma/sh/shdma-base.c chunk->async_tx.callback_param = tx->callback_param; tx 108 drivers/dma/sh/shdma-base.c tx->cookie, &chunk->async_tx, schan->id); tx 337 drivers/dma/sh/shdma-base.c struct dma_async_tx_descriptor *tx = &desc->async_tx; tx 339 drivers/dma/sh/shdma-base.c BUG_ON(tx->cookie > 0 && tx->cookie != desc->cookie); tx 353 drivers/dma/sh/shdma-base.c if (tx->cookie > 0) tx 354 drivers/dma/sh/shdma-base.c cookie = tx->cookie; tx 366 drivers/dma/sh/shdma-base.c if (desc->mark == DESC_COMPLETED && tx->callback) { tx 368 drivers/dma/sh/shdma-base.c dmaengine_desc_get_callback(tx, &cb); tx 369 drivers/dma/sh/shdma-base.c callback = tx->callback; tx 371 drivers/dma/sh/shdma-base.c tx->cookie, tx, schan->id); tx 376 drivers/dma/sh/shdma-base.c if (tx->cookie > 0 || tx->cookie == -EBUSY) { tx 378 drivers/dma/sh/shdma-base.c BUG_ON(tx->cookie < 0); tx 381 drivers/dma/sh/shdma-base.c head_acked = async_tx_test_ack(tx); tx 394 drivers/dma/sh/shdma-base.c tx, tx->cookie); tx 884 drivers/dma/sh/shdma-base.c struct dma_async_tx_descriptor *tx = &sdesc->async_tx; tx 887 drivers/dma/sh/shdma-base.c dmaengine_desc_get_callback_invoke(tx, NULL); tx 57 drivers/dma/sh/shdma.h #define tx_to_sh_desc(tx) container_of(tx, struct sh_desc, async_tx) tx 568 drivers/dma/sh/usb-dmac.c struct usb_dmac_chan *chan = to_usb_dmac_chan(vd->tx.chan); tx 588 drivers/dma/sh/usb-dmac.c desc->done_cookie = desc->vd.tx.cookie; tx 662 drivers/dma/sprd-dma.c } else if (schan->cur_desc && schan->cur_desc->vd.tx.cookie == cookie) { tx 594 drivers/dma/st_fdma.c if (fchan->fdesc && cookie == fchan->fdesc->vdesc.tx.cookie) tx 1465 drivers/dma/ste_dma40.c static dma_cookie_t d40_tx_submit(struct dma_async_tx_descriptor *tx) tx 1467 drivers/dma/ste_dma40.c struct d40_chan *d40c = container_of(tx->chan, tx 1470 drivers/dma/ste_dma40.c struct d40_desc *d40d = container_of(tx, struct d40_desc, txd); tx 1475 drivers/dma/ste_dma40.c cookie = dma_cookie_assign(tx); tx 1156 drivers/dma/stm32-dma.c if (chan->desc && cookie == chan->desc->vdesc.tx.cookie) tx 371 drivers/dma/stm32-mdma.c struct stm32_mdma_chan *chan = to_stm32_mdma_chan(vdesc->tx.chan); tx 1319 drivers/dma/stm32-mdma.c if (chan->desc && cookie == chan->desc->vdesc.tx.cookie) tx 221 drivers/dma/sun6i-dma.c to_sun6i_desc(struct dma_async_tx_descriptor *tx) tx 223 drivers/dma/sun6i-dma.c return container_of(tx, struct sun6i_desc, vd.tx); tx 403 drivers/dma/sun6i-dma.c struct sun6i_desc *txd = to_sun6i_desc(&vd->tx); tx 404 drivers/dma/sun6i-dma.c struct sun6i_dma_dev *sdev = to_sun6i_dma_dev(vd->tx.chan->device); tx 445 drivers/dma/sun6i-dma.c pchan->desc = to_sun6i_desc(&desc->tx); tx 947 drivers/dma/sun6i-dma.c txd = to_sun6i_desc(&vd->tx); tx 269 drivers/dma/tegra20-apb-dma.c static dma_cookie_t tegra_dma_tx_submit(struct dma_async_tx_descriptor *tx); tx 198 drivers/dma/tegra210-adma.c return container_of(td, struct tegra_adma_desc, vd.tx); tx 362 drivers/dma/tegra210-adma.c desc = to_tegra_adma_desc(&vd->tx); tx 526 drivers/dma/tegra210-adma.c desc = to_tegra_adma_desc(&vd->tx); tx 528 drivers/dma/tegra210-adma.c } else if (tdc->desc && tdc->desc->vd.tx.cookie == cookie) { tx 362 drivers/dma/ti/cppi41.c static dma_cookie_t cppi41_tx_submit(struct dma_async_tx_descriptor *tx) tx 366 drivers/dma/ti/cppi41.c cookie = dma_cookie_assign(tx); tx 751 drivers/dma/ti/edma.c static inline struct edma_desc *to_edma_desc(struct dma_async_tx_descriptor *tx) tx 753 drivers/dma/ti/edma.c return container_of(tx, struct edma_desc, vdesc.tx); tx 776 drivers/dma/ti/edma.c echan->edesc = to_edma_desc(&vdesc->tx); tx 1844 drivers/dma/ti/edma.c if (echan->edesc && echan->edesc->vdesc.tx.cookie == cookie) { tx 1851 drivers/dma/ti/edma.c txstate->residue = to_edma_desc(&vdesc->tx)->residue; tx 1862 drivers/dma/ti/edma.c echan->edesc->vdesc.tx.cookie == cookie) { tx 223 drivers/dma/ti/omap-dma.c return container_of(t, struct omap_desc, vd.tx); tx 228 drivers/dma/ti/omap-dma.c struct omap_desc *d = to_omap_dma_desc(&vd->tx); tx 231 drivers/dma/ti/omap-dma.c struct omap_dmadev *od = to_omap_dma_dev(vd->tx.chan->device); tx 542 drivers/dma/ti/omap-dma.c c->desc = d = to_omap_dma_desc(&vd->tx); tx 826 drivers/dma/ti/omap-dma.c if (c->desc && c->desc->vd.tx.cookie == cookie) tx 848 drivers/dma/ti/omap-dma.c to_omap_dma_desc(&vd->tx)); tx 187 drivers/dma/txx9dmac.c static dma_cookie_t txx9dmac_tx_submit(struct dma_async_tx_descriptor *tx); tx 690 drivers/dma/txx9dmac.c static dma_cookie_t txx9dmac_tx_submit(struct dma_async_tx_descriptor *tx) tx 692 drivers/dma/txx9dmac.c struct txx9dmac_desc *desc = txd_to_txx9dmac_desc(tx); tx 693 drivers/dma/txx9dmac.c struct txx9dmac_chan *dc = to_txx9dmac_chan(tx->chan); tx 697 drivers/dma/txx9dmac.c cookie = dma_cookie_assign(tx); tx 699 drivers/dma/txx9dmac.c dev_vdbg(chan2dev(tx->chan), "tx_submit: queued %u %p\n", tx 304 drivers/dma/uniphier-mdmac.c if (mc->md && mc->md->vd.tx.cookie == cookie) { tx 14 drivers/dma/virt-dma.c static struct virt_dma_desc *to_virt_desc(struct dma_async_tx_descriptor *tx) tx 16 drivers/dma/virt-dma.c return container_of(tx, struct virt_dma_desc, tx); tx 19 drivers/dma/virt-dma.c dma_cookie_t vchan_tx_submit(struct dma_async_tx_descriptor *tx) tx 21 drivers/dma/virt-dma.c struct virt_dma_chan *vc = to_virt_chan(tx->chan); tx 22 drivers/dma/virt-dma.c struct virt_dma_desc *vd = to_virt_desc(tx); tx 27 drivers/dma/virt-dma.c cookie = dma_cookie_assign(tx); tx 49 drivers/dma/virt-dma.c int vchan_tx_desc_free(struct dma_async_tx_descriptor *tx) tx 51 drivers/dma/virt-dma.c struct virt_dma_chan *vc = to_virt_chan(tx->chan); tx 52 drivers/dma/virt-dma.c struct virt_dma_desc *vd = to_virt_desc(tx); tx 60 drivers/dma/virt-dma.c vc, vd, vd->tx.cookie); tx 72 drivers/dma/virt-dma.c if (vd->tx.cookie == cookie) tx 95 drivers/dma/virt-dma.c dmaengine_desc_get_callback(&vd->tx, &cb); tx 104 drivers/dma/virt-dma.c dmaengine_desc_get_callback(&vd->tx, &cb); tx 117 drivers/dma/virt-dma.c if (dmaengine_desc_test_reuse(&vd->tx)) { tx 16 drivers/dma/virt-dma.h struct dma_async_tx_descriptor tx; tx 61 drivers/dma/virt-dma.h dma_async_tx_descriptor_init(&vd->tx, &vc->chan); tx 62 drivers/dma/virt-dma.h vd->tx.flags = tx_flags; tx 63 drivers/dma/virt-dma.h vd->tx.tx_submit = vchan_tx_submit; tx 64 drivers/dma/virt-dma.h vd->tx.desc_free = vchan_tx_desc_free; tx 73 drivers/dma/virt-dma.h return &vd->tx; tx 96 drivers/dma/virt-dma.h struct virt_dma_chan *vc = to_virt_chan(vd->tx.chan); tx 99 drivers/dma/virt-dma.h cookie = vd->tx.cookie; tx 100 drivers/dma/virt-dma.h dma_cookie_complete(&vd->tx); tx 114 drivers/dma/virt-dma.h struct virt_dma_chan *vc = to_virt_chan(vd->tx.chan); tx 116 drivers/dma/virt-dma.h if (dmaengine_desc_test_reuse(&vd->tx)) tx 128 drivers/dma/virt-dma.h struct virt_dma_chan *vc = to_virt_chan(vd->tx.chan); tx 142 drivers/dma/virt-dma.h struct virt_dma_chan *vc = to_virt_chan(vd->tx.chan); tx 193 drivers/dma/virt-dma.h dmaengine_desc_clear_reuse(&vd->tx); tx 188 drivers/dma/xgene-dma.c #define to_dma_desc_sw(tx) \ tx 189 drivers/dma/xgene-dma.c container_of(tx, struct xgene_dma_desc_sw, tx) tx 241 drivers/dma/xgene-dma.c struct dma_async_tx_descriptor tx; tx 454 drivers/dma/xgene-dma.c static dma_cookie_t xgene_dma_tx_submit(struct dma_async_tx_descriptor *tx) tx 460 drivers/dma/xgene-dma.c if (unlikely(!tx)) tx 463 drivers/dma/xgene-dma.c chan = to_dma_chan(tx->chan); tx 464 drivers/dma/xgene-dma.c desc = to_dma_desc_sw(tx); tx 468 drivers/dma/xgene-dma.c cookie = dma_cookie_assign(tx); tx 483 drivers/dma/xgene-dma.c dma_pool_free(chan->desc_pool, desc, desc->tx.phys); tx 499 drivers/dma/xgene-dma.c desc->tx.phys = phys; tx 500 drivers/dma/xgene-dma.c desc->tx.tx_submit = xgene_dma_tx_submit; tx 501 drivers/dma/xgene-dma.c dma_async_tx_descriptor_init(&desc->tx, &chan->dma_chan); tx 521 drivers/dma/xgene-dma.c if (async_tx_test_ack(&desc->tx)) tx 537 drivers/dma/xgene-dma.c struct dma_async_tx_descriptor *tx = &desc->tx; tx 546 drivers/dma/xgene-dma.c if (tx->cookie == 0) tx 549 drivers/dma/xgene-dma.c dma_cookie_complete(tx); tx 550 drivers/dma/xgene-dma.c dma_descriptor_unmap(tx); tx 553 drivers/dma/xgene-dma.c dmaengine_desc_get_callback_invoke(tx, NULL); tx 556 drivers/dma/xgene-dma.c dma_run_dependencies(tx); tx 578 drivers/dma/xgene-dma.c if (!async_tx_test_ack(&desc->tx)) { tx 588 drivers/dma/xgene-dma.c dma_pool_free(chan->desc_pool, desc, desc->tx.phys); tx 862 drivers/dma/xgene-dma.c new->tx.cookie = 0; tx 863 drivers/dma/xgene-dma.c async_tx_ack(&new->tx); tx 869 drivers/dma/xgene-dma.c new->tx.flags = flags; /* client is in control of this ack */ tx 870 drivers/dma/xgene-dma.c new->tx.cookie = -EBUSY; tx 873 drivers/dma/xgene-dma.c return &new->tx; tx 920 drivers/dma/xgene-dma.c new->tx.cookie = 0; tx 921 drivers/dma/xgene-dma.c async_tx_ack(&new->tx); tx 946 drivers/dma/xgene-dma.c new->tx.flags = flags; /* client is in control of this ack */ tx 947 drivers/dma/xgene-dma.c new->tx.cookie = -EBUSY; tx 950 drivers/dma/xgene-dma.c return &new->tx; tx 454 drivers/dma/xilinx/xilinx_dma.c #define to_dma_tx_descriptor(tx) \ tx 455 drivers/dma/xilinx/xilinx_dma.c container_of(tx, struct xilinx_dma_tx_descriptor, async_tx) tx 1588 drivers/dma/xilinx/xilinx_dma.c static dma_cookie_t xilinx_dma_tx_submit(struct dma_async_tx_descriptor *tx) tx 1590 drivers/dma/xilinx/xilinx_dma.c struct xilinx_dma_tx_descriptor *desc = to_dma_tx_descriptor(tx); tx 1591 drivers/dma/xilinx/xilinx_dma.c struct xilinx_dma_chan *chan = to_xilinx_chan(tx->chan); tx 1613 drivers/dma/xilinx/xilinx_dma.c cookie = dma_cookie_assign(tx); tx 145 drivers/dma/xilinx/zynqmp_dma.c #define tx_to_desc(tx) container_of(tx, struct zynqmp_dma_desc_sw, \ tx 369 drivers/dma/xilinx/zynqmp_dma.c static dma_cookie_t zynqmp_dma_tx_submit(struct dma_async_tx_descriptor *tx) tx 371 drivers/dma/xilinx/zynqmp_dma.c struct zynqmp_dma_chan *chan = to_chan(tx->chan); tx 376 drivers/dma/xilinx/zynqmp_dma.c new = tx_to_desc(tx); tx 378 drivers/dma/xilinx/zynqmp_dma.c cookie = dma_cookie_assign(tx); tx 719 drivers/dma/zx_dma.c struct zx_dma_dev *d = to_zx_dma(vd->tx.chan->device); tx 154 drivers/firmware/arm_scmi/base.c num_skip = t->tx.buf; tx 207 drivers/firmware/arm_scmi/base.c put_unaligned_le32(id, t->tx.buf); tx 110 drivers/firmware/arm_scmi/clock.c put_unaligned_le32(clk_id, t->tx.buf); tx 141 drivers/firmware/arm_scmi/clock.c clk_desc = t->tx.buf; tx 207 drivers/firmware/arm_scmi/clock.c put_unaligned_le32(clk_id, t->tx.buf); tx 235 drivers/firmware/arm_scmi/clock.c cfg = t->tx.buf; tx 265 drivers/firmware/arm_scmi/clock.c cfg = t->tx.buf; tx 94 drivers/firmware/arm_scmi/common.h struct scmi_msg tx; tx 262 drivers/firmware/arm_scmi/driver.c iowrite32(sizeof(mem->msg_header) + t->tx.len, &mem->length); tx 264 drivers/firmware/arm_scmi/driver.c if (t->tx.buf) tx 265 drivers/firmware/arm_scmi/driver.c memcpy_toio(mem->msg_payload, t->tx.buf, t->tx.len); tx 549 drivers/firmware/arm_scmi/driver.c xfer->tx.len = tx_size; tx 707 drivers/firmware/arm_scmi/driver.c xfer->tx.buf = xfer->rx.buf; tx 723 drivers/firmware/arm_scmi/driver.c int prot_id, bool tx) tx 732 drivers/firmware/arm_scmi/driver.c const char *desc = tx ? "Tx" : "Rx"; tx 735 drivers/firmware/arm_scmi/driver.c idx = tx ? 0 : 1; tx 736 drivers/firmware/arm_scmi/driver.c idr = tx ? &info->tx_idr : &info->rx_idr; tx 754 drivers/firmware/arm_scmi/driver.c cl->tx_prepare = tx ? scmi_tx_prepare : NULL; tx 756 drivers/firmware/arm_scmi/driver.c cl->knows_txdone = tx; tx 198 drivers/firmware/arm_scmi/perf.c put_unaligned_le32(domain, t->tx.buf); tx 253 drivers/firmware/arm_scmi/perf.c dom_info = t->tx.buf; tx 343 drivers/firmware/arm_scmi/perf.c limits = t->tx.buf; tx 382 drivers/firmware/arm_scmi/perf.c put_unaligned_le32(domain, t->tx.buf); tx 424 drivers/firmware/arm_scmi/perf.c lvl = t->tx.buf; tx 461 drivers/firmware/arm_scmi/perf.c put_unaligned_le32(domain, t->tx.buf); tx 518 drivers/firmware/arm_scmi/perf.c info = t->tx.buf; tx 99 drivers/firmware/arm_scmi/power.c put_unaligned_le32(domain, t->tx.buf); tx 128 drivers/firmware/arm_scmi/power.c st = t->tx.buf; tx 150 drivers/firmware/arm_scmi/power.c put_unaligned_le32(domain, t->tx.buf); tx 91 drivers/firmware/arm_scmi/reset.c put_unaligned_le32(domain, t->tx.buf); tx 150 drivers/firmware/arm_scmi/reset.c dom = t->tx.buf; tx 123 drivers/firmware/arm_scmi/sensors.c put_unaligned_le32(desc_index, t->tx.buf); tx 180 drivers/firmware/arm_scmi/sensors.c cfg = t->tx.buf; tx 204 drivers/firmware/arm_scmi/sensors.c trip = t->tx.buf; tx 231 drivers/firmware/arm_scmi/sensors.c sensor = t->tx.buf; tx 116 drivers/firmware/tegra/bpmp-debugfs.c .tx = { tx 151 drivers/firmware/tegra/bpmp-debugfs.c .tx = { tx 173 drivers/firmware/tegra/bpmp-debugfs.c .tx = { tx 23 drivers/firmware/tegra/bpmp-tegra186.c } tx, rx; tx 126 drivers/firmware/tegra/bpmp-tegra186.c priv->tx.virt + offset, priv->tx.phys + offset, tx 176 drivers/firmware/tegra/bpmp-tegra186.c priv->tx.pool = of_gen_pool_get(bpmp->dev->of_node, "shmem", 0); tx 177 drivers/firmware/tegra/bpmp-tegra186.c if (!priv->tx.pool) { tx 182 drivers/firmware/tegra/bpmp-tegra186.c priv->tx.virt = gen_pool_dma_alloc(priv->tx.pool, 4096, &priv->tx.phys); tx 183 drivers/firmware/tegra/bpmp-tegra186.c if (!priv->tx.virt) { tx 256 drivers/firmware/tegra/bpmp-tegra186.c gen_pool_free(priv->tx.pool, (unsigned long)priv->tx.virt, 4096); tx 275 drivers/firmware/tegra/bpmp-tegra186.c gen_pool_free(priv->tx.pool, (unsigned long)priv->tx.virt, 4096); tx 95 drivers/firmware/tegra/bpmp.c return (msg->tx.size <= MSG_DATA_MIN_SZ) && tx 97 drivers/firmware/tegra/bpmp.c (msg->tx.size == 0 || msg->tx.data) && tx 333 drivers/firmware/tegra/bpmp.c msg->tx.data, msg->tx.size); tx 367 drivers/firmware/tegra/bpmp.c channel = tegra_bpmp_write_threaded(bpmp, msg->mrq, msg->tx.data, tx 368 drivers/firmware/tegra/bpmp.c msg->tx.size); tx 508 drivers/firmware/tegra/bpmp.c .tx = { tx 558 drivers/firmware/tegra/bpmp.c msg.tx.data = &request; tx 559 drivers/firmware/tegra/bpmp.c msg.tx.size = sizeof(request); tx 602 drivers/firmware/tegra/bpmp.c msg.tx.data = &request; tx 603 drivers/firmware/tegra/bpmp.c msg.tx.size = sizeof(request); tx 62 drivers/firmware/tegra/ivc.c } tx; tx 97 drivers/firmware/tegra/ivc.c u32 tx = READ_ONCE(header->tx.count); tx 110 drivers/firmware/tegra/ivc.c if (tx - rx > ivc->num_frames) tx 113 drivers/firmware/tegra/ivc.c return tx == rx; tx 119 drivers/firmware/tegra/ivc.c u32 tx = READ_ONCE(header->tx.count); tx 126 drivers/firmware/tegra/ivc.c return tx - rx >= ivc->num_frames; tx 132 drivers/firmware/tegra/ivc.c u32 tx = READ_ONCE(header->tx.count); tx 141 drivers/firmware/tegra/ivc.c return tx - rx; tx 146 drivers/firmware/tegra/ivc.c WRITE_ONCE(ivc->tx.channel->tx.count, tx 147 drivers/firmware/tegra/ivc.c READ_ONCE(ivc->tx.channel->tx.count) + 1); tx 149 drivers/firmware/tegra/ivc.c if (ivc->tx.position == ivc->num_frames - 1) tx 150 drivers/firmware/tegra/ivc.c ivc->tx.position = 0; tx 152 drivers/firmware/tegra/ivc.c ivc->tx.position++; tx 168 drivers/firmware/tegra/ivc.c unsigned int offset = offsetof(struct tegra_ivc_header, tx.count); tx 178 drivers/firmware/tegra/ivc.c if (ivc->tx.channel->tx.state != TEGRA_IVC_STATE_ESTABLISHED) tx 203 drivers/firmware/tegra/ivc.c if (ivc->tx.channel->tx.state != TEGRA_IVC_STATE_ESTABLISHED) tx 206 drivers/firmware/tegra/ivc.c if (!tegra_ivc_full(ivc, ivc->tx.channel)) tx 209 drivers/firmware/tegra/ivc.c tegra_ivc_invalidate(ivc, ivc->tx.phys + offset); tx 211 drivers/firmware/tegra/ivc.c if (tegra_ivc_full(ivc, ivc->tx.channel)) tx 294 drivers/firmware/tegra/ivc.c unsigned int tx = offsetof(struct tegra_ivc_header, tx.count); tx 321 drivers/firmware/tegra/ivc.c tegra_ivc_invalidate(ivc, ivc->rx.phys + tx); tx 339 drivers/firmware/tegra/ivc.c return tegra_ivc_frame_virt(ivc, ivc->tx.channel, ivc->tx.position); tx 346 drivers/firmware/tegra/ivc.c unsigned int tx = offsetof(struct tegra_ivc_header, tx.count); tx 354 drivers/firmware/tegra/ivc.c tegra_ivc_flush_frame(ivc, ivc->tx.phys, ivc->tx.position, 0, tx 364 drivers/firmware/tegra/ivc.c tegra_ivc_flush(ivc, ivc->tx.phys + tx); tx 377 drivers/firmware/tegra/ivc.c tegra_ivc_invalidate(ivc, ivc->tx.phys + rx); tx 379 drivers/firmware/tegra/ivc.c if (tegra_ivc_available(ivc, ivc->tx.channel) == 1) tx 388 drivers/firmware/tegra/ivc.c unsigned int offset = offsetof(struct tegra_ivc_header, tx.count); tx 390 drivers/firmware/tegra/ivc.c ivc->tx.channel->tx.state = TEGRA_IVC_STATE_SYNC; tx 391 drivers/firmware/tegra/ivc.c tegra_ivc_flush(ivc, ivc->tx.phys + offset); tx 418 drivers/firmware/tegra/ivc.c unsigned int offset = offsetof(struct tegra_ivc_header, tx.count); tx 423 drivers/firmware/tegra/ivc.c state = READ_ONCE(ivc->rx.channel->tx.state); tx 426 drivers/firmware/tegra/ivc.c offset = offsetof(struct tegra_ivc_header, tx.count); tx 439 drivers/firmware/tegra/ivc.c ivc->tx.channel->tx.count = 0; tx 442 drivers/firmware/tegra/ivc.c ivc->tx.position = 0; tx 455 drivers/firmware/tegra/ivc.c ivc->tx.channel->tx.state = TEGRA_IVC_STATE_ACK; tx 456 drivers/firmware/tegra/ivc.c tegra_ivc_flush(ivc, ivc->tx.phys + offset); tx 463 drivers/firmware/tegra/ivc.c } else if (ivc->tx.channel->tx.state == TEGRA_IVC_STATE_SYNC && tx 465 drivers/firmware/tegra/ivc.c offset = offsetof(struct tegra_ivc_header, tx.count); tx 478 drivers/firmware/tegra/ivc.c ivc->tx.channel->tx.count = 0; tx 481 drivers/firmware/tegra/ivc.c ivc->tx.position = 0; tx 495 drivers/firmware/tegra/ivc.c ivc->tx.channel->tx.state = TEGRA_IVC_STATE_ESTABLISHED; tx 496 drivers/firmware/tegra/ivc.c tegra_ivc_flush(ivc, ivc->tx.phys + offset); tx 503 drivers/firmware/tegra/ivc.c } else if (ivc->tx.channel->tx.state == TEGRA_IVC_STATE_ACK) { tx 504 drivers/firmware/tegra/ivc.c offset = offsetof(struct tegra_ivc_header, tx.count); tx 519 drivers/firmware/tegra/ivc.c ivc->tx.channel->tx.state = TEGRA_IVC_STATE_ESTABLISHED; tx 520 drivers/firmware/tegra/ivc.c tegra_ivc_flush(ivc, ivc->tx.phys + offset); tx 536 drivers/firmware/tegra/ivc.c if (ivc->tx.channel->tx.state != TEGRA_IVC_STATE_ESTABLISHED) tx 561 drivers/firmware/tegra/ivc.c static int tegra_ivc_check_params(unsigned long rx, unsigned long tx, tx 564 drivers/firmware/tegra/ivc.c BUILD_BUG_ON(!IS_ALIGNED(offsetof(struct tegra_ivc_header, tx.count), tx 590 drivers/firmware/tegra/ivc.c if (!IS_ALIGNED(tx, TEGRA_IVC_ALIGN)) { tx 591 drivers/firmware/tegra/ivc.c pr_err("IVC channel start not aligned: %#lx\n", tx); tx 595 drivers/firmware/tegra/ivc.c if (rx < tx) { tx 596 drivers/firmware/tegra/ivc.c if (rx + frame_size * num_frames > tx) { tx 598 drivers/firmware/tegra/ivc.c rx, frame_size * num_frames, tx); tx 602 drivers/firmware/tegra/ivc.c if (tx + frame_size * num_frames > rx) { tx 604 drivers/firmware/tegra/ivc.c tx, frame_size * num_frames, rx); tx 613 drivers/firmware/tegra/ivc.c dma_addr_t rx_phys, void *tx, dma_addr_t tx_phys, tx 631 drivers/firmware/tegra/ivc.c err = tegra_ivc_check_params((unsigned long)rx, (unsigned long)tx, tx 644 drivers/firmware/tegra/ivc.c ivc->tx.phys = dma_map_single(peer, tx, queue_size, tx 646 drivers/firmware/tegra/ivc.c if (dma_mapping_error(peer, ivc->tx.phys)) { tx 653 drivers/firmware/tegra/ivc.c ivc->tx.phys = tx_phys; tx 657 drivers/firmware/tegra/ivc.c ivc->tx.channel = tx; tx 668 drivers/firmware/tegra/ivc.c ivc->tx.position = 0; tx 683 drivers/firmware/tegra/ivc.c dma_unmap_single(ivc->peer, ivc->tx.phys, size, tx 69 drivers/fpga/machxo2-spi.c struct spi_transfer rx, tx; tx 74 drivers/fpga/machxo2-spi.c memset(&tx, 0, sizeof(tx)); tx 75 drivers/fpga/machxo2-spi.c tx.tx_buf = cmd; tx 76 drivers/fpga/machxo2-spi.c tx.len = sizeof(cmd); tx 80 drivers/fpga/machxo2-spi.c spi_message_add_tail(&tx, &msg); tx 139 drivers/fpga/machxo2-spi.c struct spi_transfer tx[2]; tx 144 drivers/fpga/machxo2-spi.c memset(tx, 0, sizeof(tx)); tx 146 drivers/fpga/machxo2-spi.c tx[0].tx_buf = &erase; tx 147 drivers/fpga/machxo2-spi.c tx[0].len = sizeof(erase); tx 148 drivers/fpga/machxo2-spi.c spi_message_add_tail(&tx[0], &msg); tx 158 drivers/fpga/machxo2-spi.c tx[1].tx_buf = &refresh; tx 159 drivers/fpga/machxo2-spi.c tx[1].len = sizeof(refresh); tx 160 drivers/fpga/machxo2-spi.c tx[1].delay_usecs = MACHXO2_REFRESH_USEC; tx 161 drivers/fpga/machxo2-spi.c spi_message_add_tail(&tx[1], &msg); tx 192 drivers/fpga/machxo2-spi.c struct spi_transfer tx[3]; tx 207 drivers/fpga/machxo2-spi.c memset(tx, 0, sizeof(tx)); tx 209 drivers/fpga/machxo2-spi.c tx[0].tx_buf = &enable; tx 210 drivers/fpga/machxo2-spi.c tx[0].len = sizeof(enable); tx 211 drivers/fpga/machxo2-spi.c tx[0].delay_usecs = MACHXO2_LOW_DELAY_USEC; tx 212 drivers/fpga/machxo2-spi.c spi_message_add_tail(&tx[0], &msg); tx 214 drivers/fpga/machxo2-spi.c tx[1].tx_buf = &erase; tx 215 drivers/fpga/machxo2-spi.c tx[1].len = sizeof(erase); tx 216 drivers/fpga/machxo2-spi.c spi_message_add_tail(&tx[1], &msg); tx 231 drivers/fpga/machxo2-spi.c tx[2].tx_buf = &initaddr; tx 232 drivers/fpga/machxo2-spi.c tx[2].len = sizeof(initaddr); tx 233 drivers/fpga/machxo2-spi.c spi_message_add_tail(&tx[2], &msg); tx 253 drivers/fpga/machxo2-spi.c struct spi_transfer tx; tx 268 drivers/fpga/machxo2-spi.c memset(&tx, 0, sizeof(tx)); tx 270 drivers/fpga/machxo2-spi.c tx.tx_buf = payload; tx 271 drivers/fpga/machxo2-spi.c tx.len = MACHXO2_BUF_SIZE; tx 272 drivers/fpga/machxo2-spi.c tx.delay_usecs = MACHXO2_HIGH_DELAY_USEC; tx 273 drivers/fpga/machxo2-spi.c spi_message_add_tail(&tx, &msg); tx 291 drivers/fpga/machxo2-spi.c struct spi_transfer tx[2]; tx 297 drivers/fpga/machxo2-spi.c memset(tx, 0, sizeof(tx)); tx 299 drivers/fpga/machxo2-spi.c tx[0].tx_buf = &progdone; tx 300 drivers/fpga/machxo2-spi.c tx[0].len = sizeof(progdone); tx 301 drivers/fpga/machxo2-spi.c spi_message_add_tail(&tx[0], &msg); tx 318 drivers/fpga/machxo2-spi.c tx[1].tx_buf = &refresh; tx 319 drivers/fpga/machxo2-spi.c tx[1].len = sizeof(refresh); tx 320 drivers/fpga/machxo2-spi.c tx[1].delay_usecs = MACHXO2_REFRESH_USEC; tx 321 drivers/fpga/machxo2-spi.c spi_message_add_tail(&tx[1], &msg); tx 78 drivers/gpu/drm/amd/powerplay/hwmgr/smu8_hwmgr.h #define DDI_POWERGATING_ARG(phyID, lanemask, rx, tx, core) \ tx 82 drivers/gpu/drm/amd/powerplay/hwmgr/smu8_hwmgr.h ((tx) ? DISPLAYPHY_TX_SELECT : 0) | \ tx 472 drivers/gpu/drm/drm_mipi_dsi.c const u8 *tx = msg->tx_buf; tx 474 drivers/gpu/drm/drm_mipi_dsi.c packet->header[1] = (msg->tx_len > 0) ? tx[0] : 0; tx 475 drivers/gpu/drm/drm_mipi_dsi.c packet->header[2] = (msg->tx_len > 1) ? tx[1] : 0; tx 536 drivers/gpu/drm/drm_mipi_dsi.c u8 tx[2] = { value & 0xff, value >> 8 }; tx 540 drivers/gpu/drm/drm_mipi_dsi.c .tx_len = sizeof(tx), tx 541 drivers/gpu/drm/drm_mipi_dsi.c .tx_buf = tx, tx 698 drivers/gpu/drm/drm_mipi_dsi.c u8 *tx; tx 703 drivers/gpu/drm/drm_mipi_dsi.c tx = kmalloc(size, GFP_KERNEL); tx 704 drivers/gpu/drm/drm_mipi_dsi.c if (!tx) tx 708 drivers/gpu/drm/drm_mipi_dsi.c tx[0] = cmd; tx 709 drivers/gpu/drm/drm_mipi_dsi.c memcpy(&tx[1], data, len); tx 711 drivers/gpu/drm/drm_mipi_dsi.c tx = &cmd; tx 715 drivers/gpu/drm/drm_mipi_dsi.c err = mipi_dsi_dcs_write_buffer(dsi, tx, size); tx 718 drivers/gpu/drm/drm_mipi_dsi.c kfree(tx); tx 57 drivers/gpu/drm/i915/display/intel_lvds.c int tx; tx 168 drivers/gpu/drm/i915/display/intel_lvds.c pps->tx = REG_FIELD_GET(PANEL_LIGHT_OFF_DELAY_MASK, val); tx 184 drivers/gpu/drm/i915/display/intel_lvds.c pps->t1_t2 == 0 && pps->t5 == 0 && pps->t3 == 0 && pps->tx == 0) { tx 192 drivers/gpu/drm/i915/display/intel_lvds.c pps->tx = 200 * 10; tx 197 drivers/gpu/drm/i915/display/intel_lvds.c pps->t1_t2, pps->t3, pps->t4, pps->t5, pps->tx, tx 219 drivers/gpu/drm/i915/display/intel_lvds.c REG_FIELD_PREP(PANEL_LIGHT_OFF_DELAY_MASK, pps->tx)); tx 179 drivers/gpu/drm/mcde/mcde_dsi.c const u8 *tx = msg->tx_buf; tx 226 drivers/gpu/drm/mcde/mcde_dsi.c val |= tx[i] << (i & 3) * 8; tx 232 drivers/gpu/drm/mcde/mcde_dsi.c val |= tx[i + 4] << (i & 3) * 8; tx 238 drivers/gpu/drm/mcde/mcde_dsi.c val |= tx[i + 8] << (i & 3) * 8; tx 244 drivers/gpu/drm/mcde/mcde_dsi.c val |= tx[i + 12] << (i & 3) * 8; tx 2071 drivers/gpu/drm/msm/dsi/dsi_host.c u8 tx[2] = {pkt_size & 0xff, pkt_size >> 8}; tx 2076 drivers/gpu/drm/msm/dsi/dsi_host.c .tx_buf = tx, tx 133 drivers/gpu/drm/nouveau/nvkm/subdev/i2c/base.c u32 hi, lo, rq, tx; tx 138 drivers/gpu/drm/nouveau/nvkm/subdev/i2c/base.c i2c->func->aux_stat(i2c, &hi, &lo, &rq, &tx); tx 139 drivers/gpu/drm/nouveau/nvkm/subdev/i2c/base.c if (!hi && !lo && !rq && !tx) tx 147 drivers/gpu/drm/nouveau/nvkm/subdev/i2c/base.c if (tx & aux->intr) mask |= NVKM_I2C_DONE; tx 28 drivers/gpu/drm/nouveau/nvkm/subdev/i2c/g94.c g94_aux_stat(struct nvkm_i2c *i2c, u32 *hi, u32 *lo, u32 *rq, u32 *tx) tx 33 drivers/gpu/drm/nouveau/nvkm/subdev/i2c/g94.c for (i = 0, *hi = *lo = *rq = *tx = 0; i < 8; i++) { tx 37 drivers/gpu/drm/nouveau/nvkm/subdev/i2c/g94.c if ((stat & (8 << (i * 4)))) *tx |= 1 << i; tx 28 drivers/gpu/drm/nouveau/nvkm/subdev/i2c/gk104.c gk104_aux_stat(struct nvkm_i2c *i2c, u32 *hi, u32 *lo, u32 *rq, u32 *tx) tx 33 drivers/gpu/drm/nouveau/nvkm/subdev/i2c/gk104.c for (i = 0, *hi = *lo = *rq = *tx = 0; i < 8; i++) { tx 37 drivers/gpu/drm/nouveau/nvkm/subdev/i2c/gk104.c if ((stat & (8 << (i * 4)))) *tx |= 1 << i; tx 86 drivers/gpu/drm/omapdrm/omap_dmm_tiler.c struct dma_async_tx_descriptor *tx; tx 90 drivers/gpu/drm/omapdrm/omap_dmm_tiler.c tx = dma_dev->device_prep_dma_memcpy(dmm->wa_dma_chan, dst, src, 4, 0); tx 91 drivers/gpu/drm/omapdrm/omap_dmm_tiler.c if (!tx) { tx 96 drivers/gpu/drm/omapdrm/omap_dmm_tiler.c cookie = tx->tx_submit(tx); tx 101 drivers/gpu/drm/tiny/repaper.c const void *tx, void *rx, size_t len) tx 117 drivers/gpu/drm/tiny/repaper.c if (tx && len <= 32) { tx 118 drivers/gpu/drm/tiny/repaper.c txbuf = kmemdup(tx, len, GFP_KERNEL); tx 133 drivers/gpu/drm/tiny/repaper.c tr[1].tx_buf = txbuf ? txbuf : tx; tx 551 drivers/gpu/drm/vc4/vc4_dsi.c struct dma_async_tx_descriptor *tx; tx 563 drivers/gpu/drm/vc4/vc4_dsi.c tx = chan->device->device_prep_dma_memcpy(chan, tx 567 drivers/gpu/drm/vc4/vc4_dsi.c if (!tx) { tx 572 drivers/gpu/drm/vc4/vc4_dsi.c cookie = tx->tx_submit(tx); tx 52 drivers/hid/hid-led.c } tx; tx 291 drivers/hid/hid-led.c union delcom_packet dp = { .tx.major_cmd = 101, .tx.minor_cmd = 12 }; tx 293 drivers/hid/hid-led.c dp.tx.data_lsb = 1 << delcom_get_lednum(led); tx 294 drivers/hid/hid-led.c dp.tx.data_msb = 0; tx 301 drivers/hid/hid-led.c union delcom_packet dp = { .tx.major_cmd = 101, .tx.minor_cmd = 34 }; tx 303 drivers/hid/hid-led.c dp.tx.data_lsb = delcom_get_lednum(led); tx 304 drivers/hid/hid-led.c dp.tx.data_msb = led->cdev.brightness; tx 162 drivers/hwmon/sht21.c u8 tx[2]; tx 170 drivers/hwmon/sht21.c .buf = tx, tx 181 drivers/hwmon/sht21.c tx[0] = SHT21_READ_SNB_CMD1; tx 182 drivers/hwmon/sht21.c tx[1] = SHT21_READ_SNB_CMD2; tx 191 drivers/hwmon/sht21.c tx[0] = SHT21_READ_SNAC_CMD1; tx 192 drivers/hwmon/sht21.c tx[1] = SHT21_READ_SNAC_CMD2; tx 179 drivers/i2c/busses/i2c-cpm.c struct i2c_msg *pmsg, int num, int tx, int rx) tx 188 drivers/i2c/busses/i2c-cpm.c tbdf = cpm->tbase + tx; tx 193 drivers/i2c/busses/i2c-cpm.c tb = cpm->txbuf[tx]; tx 207 drivers/i2c/busses/i2c-cpm.c if (tx + 1 == num) tx 238 drivers/i2c/busses/i2c-cpm.c struct i2c_msg *pmsg, int tx, int rx) tx 246 drivers/i2c/busses/i2c-cpm.c tbdf = cpm->tbase + tx; tx 249 drivers/i2c/busses/i2c-cpm.c tb = cpm->txbuf[tx]; tx 275 drivers/i2c/busses/i2c-cpm.c dev_dbg(&adap->dev, "tx sc %d 0x%04x\n", tx, tx 58 drivers/i2c/busses/i2c-dln2.c } tx; tx 60 drivers/i2c/busses/i2c-dln2.c tx.port = dln2->port; tx 67 drivers/i2c/busses/i2c-dln2.c return dln2_transfer_tx(dln2->pdev, cmd, &tx, sizeof(tx)); tx 81 drivers/i2c/busses/i2c-dln2.c } __packed *tx = dln2->buf; tx 84 drivers/i2c/busses/i2c-dln2.c BUILD_BUG_ON(sizeof(*tx) > DLN2_I2C_BUF_SIZE); tx 86 drivers/i2c/busses/i2c-dln2.c tx->port = dln2->port; tx 87 drivers/i2c/busses/i2c-dln2.c tx->addr = addr; tx 88 drivers/i2c/busses/i2c-dln2.c tx->mem_addr_len = 0; tx 89 drivers/i2c/busses/i2c-dln2.c tx->mem_addr = 0; tx 90 drivers/i2c/busses/i2c-dln2.c tx->buf_len = cpu_to_le16(data_len); tx 91 drivers/i2c/busses/i2c-dln2.c memcpy(tx->buf, data, data_len); tx 93 drivers/i2c/busses/i2c-dln2.c len = sizeof(*tx) + data_len - DLN2_I2C_MAX_XFER_SIZE; tx 94 drivers/i2c/busses/i2c-dln2.c ret = dln2_transfer_tx(dln2->pdev, DLN2_I2C_WRITE, tx, len); tx 111 drivers/i2c/busses/i2c-dln2.c } __packed tx; tx 120 drivers/i2c/busses/i2c-dln2.c tx.port = dln2->port; tx 121 drivers/i2c/busses/i2c-dln2.c tx.addr = addr; tx 122 drivers/i2c/busses/i2c-dln2.c tx.mem_addr_len = 0; tx 123 drivers/i2c/busses/i2c-dln2.c tx.mem_addr = 0; tx 124 drivers/i2c/busses/i2c-dln2.c tx.buf_len = cpu_to_le16(data_len); tx 126 drivers/i2c/busses/i2c-dln2.c ret = dln2_transfer(dln2->pdev, DLN2_I2C_READ, &tx, sizeof(tx), tx 210 drivers/i2c/busses/i2c-tegra-bpmp.c msg.tx.data = request; tx 211 drivers/i2c/busses/i2c-tegra-bpmp.c msg.tx.size = sizeof(*request); tx 171 drivers/iio/accel/sca3000.c u8 tx[6] ____cacheline_aligned; tx 279 drivers/iio/accel/sca3000.c st->tx[0] = SCA3000_WRITE_REG(address); tx 280 drivers/iio/accel/sca3000.c st->tx[1] = val; tx 281 drivers/iio/accel/sca3000.c return spi_write(st->us, st->tx, 2); tx 291 drivers/iio/accel/sca3000.c .tx_buf = st->tx, tx 297 drivers/iio/accel/sca3000.c st->tx[0] = SCA3000_READ_REG(reg_address_high); tx 333 drivers/iio/accel/sca3000.c .tx_buf = st->tx, tx 337 drivers/iio/accel/sca3000.c .tx_buf = st->tx + 2, tx 340 drivers/iio/accel/sca3000.c .tx_buf = st->tx + 4, tx 343 drivers/iio/accel/sca3000.c st->tx[0] = SCA3000_WRITE_REG(SCA3000_REG_UNLOCK_ADDR); tx 344 drivers/iio/accel/sca3000.c st->tx[1] = 0x00; tx 345 drivers/iio/accel/sca3000.c st->tx[2] = SCA3000_WRITE_REG(SCA3000_REG_UNLOCK_ADDR); tx 346 drivers/iio/accel/sca3000.c st->tx[3] = 0x50; tx 347 drivers/iio/accel/sca3000.c st->tx[4] = SCA3000_WRITE_REG(SCA3000_REG_UNLOCK_ADDR); tx 348 drivers/iio/accel/sca3000.c st->tx[5] = 0xA0; tx 973 drivers/iio/accel/sca3000.c .tx_buf = st->tx, tx 980 drivers/iio/accel/sca3000.c st->tx[0] = SCA3000_READ_REG(reg_address_high); tx 67 drivers/iio/adc/ad7949.c struct spi_transfer tx[] = { tx 77 drivers/iio/adc/ad7949.c spi_message_init_with_transfers(&msg, tx, 1); tx 96 drivers/iio/adc/ad7949.c struct spi_transfer tx[] = { tx 122 drivers/iio/adc/ad7949.c spi_message_init_with_transfers(&msg, tx, 1); tx 793 drivers/iio/adc/max1363.c u8 tx[2] = { st->setupbyte, tx 800 drivers/iio/adc/max1363.c st->send(st->client, tx, 2); tx 76 drivers/iio/gyro/adxrs450.c __be32 tx ____cacheline_aligned; tx 93 drivers/iio/gyro/adxrs450.c u32 tx; tx 97 drivers/iio/gyro/adxrs450.c .tx_buf = &st->tx, tx 99 drivers/iio/gyro/adxrs450.c .len = sizeof(st->tx), tx 109 drivers/iio/gyro/adxrs450.c tx = ADXRS450_READ_DATA | (reg_address << 17); tx 111 drivers/iio/gyro/adxrs450.c if (!(hweight32(tx) & 1)) tx 112 drivers/iio/gyro/adxrs450.c tx |= ADXRS450_P; tx 114 drivers/iio/gyro/adxrs450.c st->tx = cpu_to_be32(tx); tx 141 drivers/iio/gyro/adxrs450.c u32 tx; tx 145 drivers/iio/gyro/adxrs450.c tx = ADXRS450_WRITE_DATA | (reg_address << 17) | (val << 1); tx 147 drivers/iio/gyro/adxrs450.c if (!(hweight32(tx) & 1)) tx 148 drivers/iio/gyro/adxrs450.c tx |= ADXRS450_P; tx 150 drivers/iio/gyro/adxrs450.c st->tx = cpu_to_be32(tx); tx 151 drivers/iio/gyro/adxrs450.c ret = spi_write(st->us, &st->tx, sizeof(st->tx)); tx 171 drivers/iio/gyro/adxrs450.c .tx_buf = &st->tx, tx 173 drivers/iio/gyro/adxrs450.c .len = sizeof(st->tx), tx 183 drivers/iio/gyro/adxrs450.c st->tx = cpu_to_be32(ADXRS450_SENSOR_DATA); tx 208 drivers/iio/gyro/adxrs450.c u32 tx; tx 210 drivers/iio/gyro/adxrs450.c .tx_buf = &st->tx, tx 213 drivers/iio/gyro/adxrs450.c .len = sizeof(st->tx), tx 217 drivers/iio/gyro/adxrs450.c tx = ADXRS450_SENSOR_DATA; tx 219 drivers/iio/gyro/adxrs450.c tx |= (ADXRS450_CHK | ADXRS450_P); tx 220 drivers/iio/gyro/adxrs450.c st->tx = cpu_to_be32(tx); tx 25 drivers/iio/gyro/itg3200_buffer.c u8 tx = 0x80 | ITG3200_REG_TEMP_OUT_H; tx 31 drivers/iio/gyro/itg3200_buffer.c .buf = &tx, tx 214 drivers/iio/health/afe4403.c u8 tx[4] = {AFE440X_CONTROL0, 0x0, 0x0, AFE440X_CONTROL0_READ}; tx 219 drivers/iio/health/afe4403.c ret = spi_write_then_read(afe->spi, tx, 4, NULL, 0); tx 232 drivers/iio/health/afe4403.c tx[3] = AFE440X_CONTROL0_WRITE; tx 233 drivers/iio/health/afe4403.c ret = spi_write_then_read(afe->spi, tx, 4, NULL, 0); tx 313 drivers/iio/health/afe4403.c u8 tx[4] = {AFE440X_CONTROL0, 0x0, 0x0, AFE440X_CONTROL0_READ}; tx 317 drivers/iio/health/afe4403.c ret = spi_write_then_read(afe->spi, tx, 4, NULL, 0); tx 335 drivers/iio/health/afe4403.c tx[3] = AFE440X_CONTROL0_WRITE; tx 336 drivers/iio/health/afe4403.c ret = spi_write_then_read(afe->spi, tx, 4, NULL, 0); tx 37 drivers/iio/imu/adis.c .tx_buf = adis->tx, tx 45 drivers/iio/imu/adis.c .tx_buf = adis->tx + 2, tx 53 drivers/iio/imu/adis.c .tx_buf = adis->tx + 4, tx 61 drivers/iio/imu/adis.c .tx_buf = adis->tx + 6, tx 66 drivers/iio/imu/adis.c .tx_buf = adis->tx + 8, tx 78 drivers/iio/imu/adis.c adis->tx[0] = ADIS_WRITE_REG(ADIS_REG_PAGE_ID); tx 79 drivers/iio/imu/adis.c adis->tx[1] = page; tx 85 drivers/iio/imu/adis.c adis->tx[8] = ADIS_WRITE_REG(reg + 3); tx 86 drivers/iio/imu/adis.c adis->tx[9] = (value >> 24) & 0xff; tx 87 drivers/iio/imu/adis.c adis->tx[6] = ADIS_WRITE_REG(reg + 2); tx 88 drivers/iio/imu/adis.c adis->tx[7] = (value >> 16) & 0xff; tx 91 drivers/iio/imu/adis.c adis->tx[4] = ADIS_WRITE_REG(reg + 1); tx 92 drivers/iio/imu/adis.c adis->tx[5] = (value >> 8) & 0xff; tx 95 drivers/iio/imu/adis.c adis->tx[2] = ADIS_WRITE_REG(reg); tx 96 drivers/iio/imu/adis.c adis->tx[3] = value & 0xff; tx 137 drivers/iio/imu/adis.c .tx_buf = adis->tx, tx 145 drivers/iio/imu/adis.c .tx_buf = adis->tx + 2, tx 153 drivers/iio/imu/adis.c .tx_buf = adis->tx + 4, tx 173 drivers/iio/imu/adis.c adis->tx[0] = ADIS_WRITE_REG(ADIS_REG_PAGE_ID); tx 174 drivers/iio/imu/adis.c adis->tx[1] = page; tx 180 drivers/iio/imu/adis.c adis->tx[2] = ADIS_READ_REG(reg + 2); tx 181 drivers/iio/imu/adis.c adis->tx[3] = 0; tx 185 drivers/iio/imu/adis.c adis->tx[4] = ADIS_READ_REG(reg); tx 186 drivers/iio/imu/adis.c adis->tx[5] = 0; tx 27 drivers/iio/imu/adis_buffer.c u8 *tx; tx 44 drivers/iio/imu/adis_buffer.c tx = adis->buffer + burst_length; tx 45 drivers/iio/imu/adis_buffer.c tx[0] = ADIS_READ_REG(adis->burst->reg_cmd); tx 46 drivers/iio/imu/adis_buffer.c tx[1] = 0; tx 48 drivers/iio/imu/adis_buffer.c adis->xfer[0].tx_buf = tx; tx 69 drivers/iio/imu/adis_buffer.c __be16 *tx, *rx; tx 91 drivers/iio/imu/adis_buffer.c tx = rx + scan_count; tx 102 drivers/iio/imu/adis_buffer.c adis->xfer[j].tx_buf = &tx[j]; tx 113 drivers/iio/imu/adis_buffer.c *tx++ = cpu_to_be16((chan->address + 2) << 8); tx 114 drivers/iio/imu/adis_buffer.c *tx++ = cpu_to_be16(chan->address << 8); tx 134 drivers/iio/imu/adis_buffer.c adis->tx[0] = ADIS_WRITE_REG(ADIS_REG_PAGE_ID); tx 135 drivers/iio/imu/adis_buffer.c adis->tx[1] = 0; tx 136 drivers/iio/imu/adis_buffer.c spi_write(adis->spi, adis->tx, 2); tx 19 drivers/iio/pressure/mpl115_spi.c u8 tx[4]; tx 42 drivers/iio/pressure/mpl115_spi.c .tx_buf = buf->tx, tx 48 drivers/iio/pressure/mpl115_spi.c buf->tx[0] = MPL115_SPI_READ(address); tx 49 drivers/iio/pressure/mpl115_spi.c buf->tx[2] = MPL115_SPI_READ(address + 1); tx 63 drivers/iio/pressure/mpl115_spi.c .tx_buf = buf->tx, tx 67 drivers/iio/pressure/mpl115_spi.c buf->tx[0] = MPL115_SPI_WRITE(address); tx 68 drivers/iio/pressure/mpl115_spi.c buf->tx[1] = value; tx 7333 drivers/infiniband/hw/hfi1/chip.c u16 tx, rx; tx 7346 drivers/infiniband/hw/hfi1/chip.c tx = nibble_to_count(enable_lane_tx); tx 7374 drivers/infiniband/hw/hfi1/chip.c enable_lane_tx, tx, enable_lane_rx, rx); tx 7375 drivers/infiniband/hw/hfi1/chip.c *tx_width = link_width_to_bits(dd, tx); tx 7395 drivers/infiniband/hw/hfi1/chip.c u16 widths, tx, rx; tx 7400 drivers/infiniband/hw/hfi1/chip.c tx = widths >> 12; tx 7403 drivers/infiniband/hw/hfi1/chip.c *tx_width = link_width_to_bits(dd, tx); tx 7625 drivers/infiniband/hw/hfi1/chip.c u16 tx, rx; tx 7651 drivers/infiniband/hw/hfi1/chip.c get_link_widths(ppd->dd, &tx, &rx); tx 7652 drivers/infiniband/hw/hfi1/chip.c ppd->link_width_downgrade_tx_active = tx; tx 46 drivers/infiniband/hw/hfi1/iowait.c struct sdma_txreq *tx, tx 144 drivers/infiniband/hw/hfi1/iowait.h struct sdma_txreq *tx, tx 176 drivers/infiniband/hw/hfi1/iowait.h struct sdma_txreq *tx, tx 328 drivers/infiniband/hw/hfi1/iowait.h struct sdma_txreq *tx = NULL; tx 331 drivers/infiniband/hw/hfi1/iowait.h tx = list_first_entry( tx 335 drivers/infiniband/hw/hfi1/iowait.h list_del_init(&tx->list); tx 337 drivers/infiniband/hw/hfi1/iowait.h return tx; tx 343 drivers/infiniband/hw/hfi1/iowait.h struct sdma_txreq *tx = NULL; tx 346 drivers/infiniband/hw/hfi1/iowait.h tx = list_first_entry(&w->tx_head, struct sdma_txreq, tx 348 drivers/infiniband/hw/hfi1/iowait.h num_desc = tx->num_desc; tx 349 drivers/infiniband/hw/hfi1/iowait.h if (tx->flags & SDMA_TXREQ_F_VIP) tx 366 drivers/infiniband/hw/hfi1/iowait.h struct sdma_txreq *tx = NULL; tx 369 drivers/infiniband/hw/hfi1/iowait.h tx = list_first_entry(&w->tx_head, struct sdma_txreq, tx 371 drivers/infiniband/hw/hfi1/iowait.h if (tx->flags & SDMA_TXREQ_F_VIP) tx 152 drivers/infiniband/hw/hfi1/qp.c struct sdma_txreq *tx; tx 154 drivers/infiniband/hw/hfi1/qp.c tx = list_first_entry( tx 158 drivers/infiniband/hw/hfi1/qp.c list_del_init(&tx->list); tx 160 drivers/infiniband/hw/hfi1/qp.c container_of(tx, struct verbs_txreq, txreq)); tx 495 drivers/infiniband/hw/hfi1/qp.c struct verbs_txreq *tx = container_of(stx, struct verbs_txreq, txreq); tx 501 drivers/infiniband/hw/hfi1/qp.c qp = tx->qp; tx 535 drivers/infiniband/hw/hfi1/qp.c hfi1_put_txreq(tx); tx 365 drivers/infiniband/hw/hfi1/sdma.c struct sdma_txreq *tx, tx 369 drivers/infiniband/hw/hfi1/sdma.c struct iowait *wait = tx->wait; tx 370 drivers/infiniband/hw/hfi1/sdma.c callback_t complete = tx->complete; tx 373 drivers/infiniband/hw/hfi1/sdma.c trace_hfi1_sdma_out_sn(sde, tx->sn); tx 374 drivers/infiniband/hw/hfi1/sdma.c if (WARN_ON_ONCE(sde->head_sn != tx->sn)) tx 376 drivers/infiniband/hw/hfi1/sdma.c sde->head_sn, tx->sn); tx 379 drivers/infiniband/hw/hfi1/sdma.c __sdma_txclean(sde->dd, tx); tx 381 drivers/infiniband/hw/hfi1/sdma.c (*complete)(tx, res); tx 1658 drivers/infiniband/hw/hfi1/sdma.c static inline u8 ahg_mode(struct sdma_txreq *tx) tx 1660 drivers/infiniband/hw/hfi1/sdma.c return (tx->descp[0].qw[1] & SDMA_DESC1_HEADER_MODE_SMASK) tx 1677 drivers/infiniband/hw/hfi1/sdma.c struct sdma_txreq *tx) tx 1681 drivers/infiniband/hw/hfi1/sdma.c if (tx->num_desc) { tx 1682 drivers/infiniband/hw/hfi1/sdma.c u8 skip = 0, mode = ahg_mode(tx); tx 1685 drivers/infiniband/hw/hfi1/sdma.c sdma_unmap_desc(dd, &tx->descp[0]); tx 1689 drivers/infiniband/hw/hfi1/sdma.c for (i = 1 + skip; i < tx->num_desc; i++) tx 1690 drivers/infiniband/hw/hfi1/sdma.c sdma_unmap_desc(dd, &tx->descp[i]); tx 1691 drivers/infiniband/hw/hfi1/sdma.c tx->num_desc = 0; tx 1693 drivers/infiniband/hw/hfi1/sdma.c kfree(tx->coalesce_buf); tx 1694 drivers/infiniband/hw/hfi1/sdma.c tx->coalesce_buf = NULL; tx 1696 drivers/infiniband/hw/hfi1/sdma.c if (unlikely(tx->desc_limit > ARRAY_SIZE(tx->descs))) { tx 1697 drivers/infiniband/hw/hfi1/sdma.c tx->desc_limit = ARRAY_SIZE(tx->descs); tx 1698 drivers/infiniband/hw/hfi1/sdma.c kfree(tx->descp); tx 2307 drivers/infiniband/hw/hfi1/sdma.c static inline u16 submit_tx(struct sdma_engine *sde, struct sdma_txreq *tx) tx 2311 drivers/infiniband/hw/hfi1/sdma.c struct sdma_desc *descp = tx->descp; tx 2312 drivers/infiniband/hw/hfi1/sdma.c u8 skip = 0, mode = ahg_mode(tx); tx 2323 drivers/infiniband/hw/hfi1/sdma.c for (i = 1; i < tx->num_desc; i++, descp++) { tx 2340 drivers/infiniband/hw/hfi1/sdma.c tx->next_descq_idx = tail; tx 2342 drivers/infiniband/hw/hfi1/sdma.c tx->sn = sde->tail_sn++; tx 2343 drivers/infiniband/hw/hfi1/sdma.c trace_hfi1_sdma_in_sn(sde, tx->sn); tx 2346 drivers/infiniband/hw/hfi1/sdma.c sde->tx_ring[sde->tx_tail++ & sde->sdma_mask] = tx; tx 2347 drivers/infiniband/hw/hfi1/sdma.c sde->desc_avail -= tx->num_desc; tx 2357 drivers/infiniband/hw/hfi1/sdma.c struct sdma_txreq *tx, tx 2363 drivers/infiniband/hw/hfi1/sdma.c if (tx->num_desc <= sde->desc_avail) tx 2371 drivers/infiniband/hw/hfi1/sdma.c ret = wait->iow->sleep(sde, wait, tx, seq, pkts_sent); tx 2397 drivers/infiniband/hw/hfi1/sdma.c struct sdma_txreq *tx, tx 2405 drivers/infiniband/hw/hfi1/sdma.c if (unlikely(tx->tlen)) tx 2407 drivers/infiniband/hw/hfi1/sdma.c tx->wait = iowait_ioww_to_iow(wait); tx 2412 drivers/infiniband/hw/hfi1/sdma.c if (unlikely(tx->num_desc > sde->desc_avail)) tx 2414 drivers/infiniband/hw/hfi1/sdma.c tail = submit_tx(sde, tx); tx 2424 drivers/infiniband/hw/hfi1/sdma.c tx->next_descq_idx = 0; tx 2426 drivers/infiniband/hw/hfi1/sdma.c tx->sn = sde->tail_sn++; tx 2427 drivers/infiniband/hw/hfi1/sdma.c trace_hfi1_sdma_in_sn(sde, tx->sn); tx 2430 drivers/infiniband/hw/hfi1/sdma.c list_add_tail(&tx->list, &sde->flushlist); tx 2432 drivers/infiniband/hw/hfi1/sdma.c iowait_inc_wait_count(wait, tx->num_desc); tx 2437 drivers/infiniband/hw/hfi1/sdma.c ret = sdma_check_progress(sde, wait, tx, pkts_sent); tx 2477 drivers/infiniband/hw/hfi1/sdma.c struct sdma_txreq *tx, *tx_next; tx 2485 drivers/infiniband/hw/hfi1/sdma.c list_for_each_entry_safe(tx, tx_next, tx_list, list) { tx 2486 drivers/infiniband/hw/hfi1/sdma.c tx->wait = iowait_ioww_to_iow(wait); tx 2489 drivers/infiniband/hw/hfi1/sdma.c if (unlikely(tx->num_desc > sde->desc_avail)) tx 2491 drivers/infiniband/hw/hfi1/sdma.c if (unlikely(tx->tlen)) { tx 2495 drivers/infiniband/hw/hfi1/sdma.c list_del_init(&tx->list); tx 2496 drivers/infiniband/hw/hfi1/sdma.c tail = submit_tx(sde, tx); tx 2518 drivers/infiniband/hw/hfi1/sdma.c list_for_each_entry_safe(tx, tx_next, tx_list, list) { tx 2519 drivers/infiniband/hw/hfi1/sdma.c tx->wait = iowait_ioww_to_iow(wait); tx 2520 drivers/infiniband/hw/hfi1/sdma.c list_del_init(&tx->list); tx 2521 drivers/infiniband/hw/hfi1/sdma.c tx->next_descq_idx = 0; tx 2523 drivers/infiniband/hw/hfi1/sdma.c tx->sn = sde->tail_sn++; tx 2524 drivers/infiniband/hw/hfi1/sdma.c trace_hfi1_sdma_in_sn(sde, tx->sn); tx 2526 drivers/infiniband/hw/hfi1/sdma.c list_add_tail(&tx->list, &sde->flushlist); tx 2528 drivers/infiniband/hw/hfi1/sdma.c iowait_inc_wait_count(wait, tx->num_desc); tx 2535 drivers/infiniband/hw/hfi1/sdma.c ret = sdma_check_progress(sde, wait, tx, submit_count > 0); tx 3056 drivers/infiniband/hw/hfi1/sdma.c static int _extend_sdma_tx_descs(struct hfi1_devdata *dd, struct sdma_txreq *tx) tx 3061 drivers/infiniband/hw/hfi1/sdma.c if (unlikely((tx->num_desc == (MAX_DESC - 1)))) { tx 3063 drivers/infiniband/hw/hfi1/sdma.c if (!tx->tlen) { tx 3064 drivers/infiniband/hw/hfi1/sdma.c tx->desc_limit = MAX_DESC; tx 3065 drivers/infiniband/hw/hfi1/sdma.c } else if (!tx->coalesce_buf) { tx 3067 drivers/infiniband/hw/hfi1/sdma.c tx->coalesce_buf = kmalloc(tx->tlen + sizeof(u32), tx 3069 drivers/infiniband/hw/hfi1/sdma.c if (!tx->coalesce_buf) tx 3071 drivers/infiniband/hw/hfi1/sdma.c tx->coalesce_idx = 0; tx 3076 drivers/infiniband/hw/hfi1/sdma.c if (unlikely(tx->num_desc == MAX_DESC)) tx 3079 drivers/infiniband/hw/hfi1/sdma.c tx->descp = kmalloc_array( tx 3083 drivers/infiniband/hw/hfi1/sdma.c if (!tx->descp) tx 3087 drivers/infiniband/hw/hfi1/sdma.c tx->desc_limit = MAX_DESC - 1; tx 3089 drivers/infiniband/hw/hfi1/sdma.c for (i = 0; i < tx->num_desc; i++) tx 3090 drivers/infiniband/hw/hfi1/sdma.c tx->descp[i] = tx->descs[i]; tx 3093 drivers/infiniband/hw/hfi1/sdma.c __sdma_txclean(dd, tx); tx 3113 drivers/infiniband/hw/hfi1/sdma.c int ext_coal_sdma_tx_descs(struct hfi1_devdata *dd, struct sdma_txreq *tx, tx 3120 drivers/infiniband/hw/hfi1/sdma.c rval = _extend_sdma_tx_descs(dd, tx); tx 3122 drivers/infiniband/hw/hfi1/sdma.c __sdma_txclean(dd, tx); tx 3127 drivers/infiniband/hw/hfi1/sdma.c if (tx->coalesce_buf) { tx 3129 drivers/infiniband/hw/hfi1/sdma.c __sdma_txclean(dd, tx); tx 3137 drivers/infiniband/hw/hfi1/sdma.c __sdma_txclean(dd, tx); tx 3141 drivers/infiniband/hw/hfi1/sdma.c memcpy(tx->coalesce_buf + tx->coalesce_idx, kvaddr, len); tx 3142 drivers/infiniband/hw/hfi1/sdma.c tx->coalesce_idx += len; tx 3147 drivers/infiniband/hw/hfi1/sdma.c if (tx->tlen - tx->coalesce_idx) tx 3151 drivers/infiniband/hw/hfi1/sdma.c pad_len = tx->packet_len & (sizeof(u32) - 1); tx 3154 drivers/infiniband/hw/hfi1/sdma.c memset(tx->coalesce_buf + tx->coalesce_idx, 0, pad_len); tx 3156 drivers/infiniband/hw/hfi1/sdma.c tx->packet_len += pad_len; tx 3157 drivers/infiniband/hw/hfi1/sdma.c tx->tlen += pad_len; tx 3162 drivers/infiniband/hw/hfi1/sdma.c tx->coalesce_buf, tx 3163 drivers/infiniband/hw/hfi1/sdma.c tx->tlen, tx 3167 drivers/infiniband/hw/hfi1/sdma.c __sdma_txclean(dd, tx); tx 3172 drivers/infiniband/hw/hfi1/sdma.c tx->desc_limit = MAX_DESC; tx 3173 drivers/infiniband/hw/hfi1/sdma.c return _sdma_txadd_daddr(dd, SDMA_MAP_SINGLE, tx, tx 3174 drivers/infiniband/hw/hfi1/sdma.c addr, tx->tlen); tx 3201 drivers/infiniband/hw/hfi1/sdma.c int _pad_sdma_tx_descs(struct hfi1_devdata *dd, struct sdma_txreq *tx) tx 3205 drivers/infiniband/hw/hfi1/sdma.c tx->num_desc++; tx 3206 drivers/infiniband/hw/hfi1/sdma.c if ((unlikely(tx->num_desc == tx->desc_limit))) { tx 3207 drivers/infiniband/hw/hfi1/sdma.c rval = _extend_sdma_tx_descs(dd, tx); tx 3209 drivers/infiniband/hw/hfi1/sdma.c __sdma_txclean(dd, tx); tx 3215 drivers/infiniband/hw/hfi1/sdma.c tx, tx 3218 drivers/infiniband/hw/hfi1/sdma.c sizeof(u32) - (tx->packet_len & (sizeof(u32) - 1))); tx 3219 drivers/infiniband/hw/hfi1/sdma.c _sdma_close_tx(dd, tx); tx 3231 drivers/infiniband/hw/hfi1/sdma.c struct sdma_txreq *tx, tx 3248 drivers/infiniband/hw/hfi1/sdma.c tx->num_desc++; tx 3252 drivers/infiniband/hw/hfi1/sdma.c tx->num_desc++; tx 3253 drivers/infiniband/hw/hfi1/sdma.c tx->descs[2].qw[0] = 0; tx 3254 drivers/infiniband/hw/hfi1/sdma.c tx->descs[2].qw[1] = 0; tx 3257 drivers/infiniband/hw/hfi1/sdma.c tx->num_desc++; tx 3258 drivers/infiniband/hw/hfi1/sdma.c tx->descs[1].qw[0] = 0; tx 3259 drivers/infiniband/hw/hfi1/sdma.c tx->descs[1].qw[1] = 0; tx 3263 drivers/infiniband/hw/hfi1/sdma.c tx->descs[0].qw[1] |= tx 3275 drivers/infiniband/hw/hfi1/sdma.c tx->descs[desc].qw[!!(i & 2)] |= tx 481 drivers/infiniband/hw/hfi1/sdma.h struct sdma_txreq *tx, tx 540 drivers/infiniband/hw/hfi1/sdma.h struct sdma_txreq *tx, tx 553 drivers/infiniband/hw/hfi1/sdma.h tx->desc_limit = ARRAY_SIZE(tx->descs); tx 554 drivers/infiniband/hw/hfi1/sdma.h tx->descp = &tx->descs[0]; tx 555 drivers/infiniband/hw/hfi1/sdma.h INIT_LIST_HEAD(&tx->list); tx 556 drivers/infiniband/hw/hfi1/sdma.h tx->num_desc = 0; tx 557 drivers/infiniband/hw/hfi1/sdma.h tx->flags = flags; tx 558 drivers/infiniband/hw/hfi1/sdma.h tx->complete = cb; tx 559 drivers/infiniband/hw/hfi1/sdma.h tx->coalesce_buf = NULL; tx 560 drivers/infiniband/hw/hfi1/sdma.h tx->wait = NULL; tx 561 drivers/infiniband/hw/hfi1/sdma.h tx->packet_len = tlen; tx 562 drivers/infiniband/hw/hfi1/sdma.h tx->tlen = tx->packet_len; tx 563 drivers/infiniband/hw/hfi1/sdma.h tx->descs[0].qw[0] = SDMA_DESC0_FIRST_DESC_FLAG; tx 564 drivers/infiniband/hw/hfi1/sdma.h tx->descs[0].qw[1] = 0; tx 566 drivers/infiniband/hw/hfi1/sdma.h tx->descs[0].qw[1] |= tx 572 drivers/infiniband/hw/hfi1/sdma.h _sdma_txreq_ahgadd(tx, num_ahg, ahg_entry, ahg, ahg_hlen); tx 608 drivers/infiniband/hw/hfi1/sdma.h struct sdma_txreq *tx, tx 613 drivers/infiniband/hw/hfi1/sdma.h return sdma_txinit_ahg(tx, flags, tlen, 0, 0, NULL, 0, cb); tx 636 drivers/infiniband/hw/hfi1/sdma.h struct sdma_txreq *tx, tx 641 drivers/infiniband/hw/hfi1/sdma.h struct sdma_desc *desc = &tx->descp[tx->num_desc]; tx 643 drivers/infiniband/hw/hfi1/sdma.h if (!tx->num_desc) { tx 659 drivers/infiniband/hw/hfi1/sdma.h int ext_coal_sdma_tx_descs(struct hfi1_devdata *dd, struct sdma_txreq *tx, tx 665 drivers/infiniband/hw/hfi1/sdma.h static inline void sdma_txclean(struct hfi1_devdata *dd, struct sdma_txreq *tx) tx 667 drivers/infiniband/hw/hfi1/sdma.h if (tx->num_desc) tx 668 drivers/infiniband/hw/hfi1/sdma.h __sdma_txclean(dd, tx); tx 673 drivers/infiniband/hw/hfi1/sdma.h struct sdma_txreq *tx) tx 675 drivers/infiniband/hw/hfi1/sdma.h tx->descp[tx->num_desc].qw[0] |= tx 677 drivers/infiniband/hw/hfi1/sdma.h tx->descp[tx->num_desc].qw[1] |= tx 679 drivers/infiniband/hw/hfi1/sdma.h if (tx->flags & SDMA_TXREQ_F_URGENT) tx 680 drivers/infiniband/hw/hfi1/sdma.h tx->descp[tx->num_desc].qw[1] |= tx 688 drivers/infiniband/hw/hfi1/sdma.h struct sdma_txreq *tx, tx 695 drivers/infiniband/hw/hfi1/sdma.h tx, tx 698 drivers/infiniband/hw/hfi1/sdma.h WARN_ON(len > tx->tlen); tx 699 drivers/infiniband/hw/hfi1/sdma.h tx->tlen -= len; tx 701 drivers/infiniband/hw/hfi1/sdma.h if (!tx->tlen) { tx 702 drivers/infiniband/hw/hfi1/sdma.h if (tx->packet_len & (sizeof(u32) - 1)) { tx 703 drivers/infiniband/hw/hfi1/sdma.h rval = _pad_sdma_tx_descs(dd, tx); tx 707 drivers/infiniband/hw/hfi1/sdma.h _sdma_close_tx(dd, tx); tx 710 drivers/infiniband/hw/hfi1/sdma.h tx->num_desc++; tx 732 drivers/infiniband/hw/hfi1/sdma.h struct sdma_txreq *tx, tx 740 drivers/infiniband/hw/hfi1/sdma.h if ((unlikely(tx->num_desc == tx->desc_limit))) { tx 741 drivers/infiniband/hw/hfi1/sdma.h rval = ext_coal_sdma_tx_descs(dd, tx, SDMA_MAP_PAGE, tx 755 drivers/infiniband/hw/hfi1/sdma.h __sdma_txclean(dd, tx); tx 760 drivers/infiniband/hw/hfi1/sdma.h dd, SDMA_MAP_PAGE, tx, addr, len); tx 781 drivers/infiniband/hw/hfi1/sdma.h struct sdma_txreq *tx, tx 787 drivers/infiniband/hw/hfi1/sdma.h if ((unlikely(tx->num_desc == tx->desc_limit))) { tx 788 drivers/infiniband/hw/hfi1/sdma.h rval = ext_coal_sdma_tx_descs(dd, tx, SDMA_MAP_NONE, tx 794 drivers/infiniband/hw/hfi1/sdma.h return _sdma_txadd_daddr(dd, SDMA_MAP_NONE, tx, addr, len); tx 815 drivers/infiniband/hw/hfi1/sdma.h struct sdma_txreq *tx, tx 822 drivers/infiniband/hw/hfi1/sdma.h if ((unlikely(tx->num_desc == tx->desc_limit))) { tx 823 drivers/infiniband/hw/hfi1/sdma.h rval = ext_coal_sdma_tx_descs(dd, tx, SDMA_MAP_SINGLE, tx 836 drivers/infiniband/hw/hfi1/sdma.h __sdma_txclean(dd, tx); tx 841 drivers/infiniband/hw/hfi1/sdma.h dd, SDMA_MAP_SINGLE, tx, addr, len); tx 848 drivers/infiniband/hw/hfi1/sdma.h struct sdma_txreq *tx, tx 899 drivers/infiniband/hw/hfi1/sdma.h struct sdma_txreq *tx) tx 903 drivers/infiniband/hw/hfi1/sdma.h if (tx->num_desc > sde->desc_avail) tx 131 drivers/infiniband/hw/hfi1/sdma_txreq.h static inline int sdma_txreq_built(struct sdma_txreq *tx) tx 133 drivers/infiniband/hw/hfi1/sdma_txreq.h return tx->num_desc; tx 91 drivers/infiniband/hw/hfi1/user_sdma.c struct user_sdma_txreq *tx, u32 datalen); tx 93 drivers/infiniband/hw/hfi1/user_sdma.c struct user_sdma_txreq *tx, u32 len); tx 632 drivers/infiniband/hw/hfi1/user_sdma.c struct user_sdma_txreq *tx) tx 653 drivers/infiniband/hw/hfi1/user_sdma.c (sizeof(tx->hdr) - 4)); tx 701 drivers/infiniband/hw/hfi1/user_sdma.c struct user_sdma_txreq *tx, tx 717 drivers/infiniband/hw/hfi1/user_sdma.c memcpy(&tx->hdr, &req->hdr, sizeof(tx->hdr)); tx 720 drivers/infiniband/hw/hfi1/user_sdma.c tx->hdr.pbc[0] = cpu_to_le16(pbclen); tx 722 drivers/infiniband/hw/hfi1/user_sdma.c ret = check_header_template(req, &tx->hdr, lrhlen, datalen); tx 725 drivers/infiniband/hw/hfi1/user_sdma.c ret = sdma_txinit_ahg(&tx->txreq, SDMA_TXREQ_F_AHG_COPY, tx 726 drivers/infiniband/hw/hfi1/user_sdma.c sizeof(tx->hdr) + datalen, req->ahg_idx, tx 730 drivers/infiniband/hw/hfi1/user_sdma.c ret = sdma_txadd_kvaddr(pq->dd, &tx->txreq, &tx->hdr, sizeof(tx->hdr)); tx 732 drivers/infiniband/hw/hfi1/user_sdma.c sdma_txclean(pq->dd, &tx->txreq); tx 737 drivers/infiniband/hw/hfi1/user_sdma.c struct user_sdma_txreq *tx, tx 756 drivers/infiniband/hw/hfi1/user_sdma.c ret = sdma_txadd_page(pq->dd, &tx->txreq, iovec->pages[pageidx], tx 783 drivers/infiniband/hw/hfi1/user_sdma.c struct user_sdma_txreq *tx = NULL; tx 820 drivers/infiniband/hw/hfi1/user_sdma.c tx = kmem_cache_alloc(pq->txreq_cache, GFP_KERNEL); tx 821 drivers/infiniband/hw/hfi1/user_sdma.c if (!tx) tx 824 drivers/infiniband/hw/hfi1/user_sdma.c tx->flags = 0; tx 825 drivers/infiniband/hw/hfi1/user_sdma.c tx->req = req; tx 826 drivers/infiniband/hw/hfi1/user_sdma.c INIT_LIST_HEAD(&tx->list); tx 833 drivers/infiniband/hw/hfi1/user_sdma.c tx->flags |= (TXREQ_FLAGS_REQ_ACK | tx 852 drivers/infiniband/hw/hfi1/user_sdma.c datalen = compute_data_length(req, tx); tx 868 drivers/infiniband/hw/hfi1/user_sdma.c tx->flags |= TXREQ_FLAGS_REQ_DISABLE_SH; tx 874 drivers/infiniband/hw/hfi1/user_sdma.c ret = user_sdma_txadd_ahg(req, tx, datalen); tx 880 drivers/infiniband/hw/hfi1/user_sdma.c changes = set_txreq_header_ahg(req, tx, tx 888 drivers/infiniband/hw/hfi1/user_sdma.c ret = sdma_txinit(&tx->txreq, 0, sizeof(req->hdr) + tx 898 drivers/infiniband/hw/hfi1/user_sdma.c ret = set_txreq_header(req, tx, datalen); tx 909 drivers/infiniband/hw/hfi1/user_sdma.c ret = user_sdma_txadd(req, tx, iovec, datalen, tx 924 drivers/infiniband/hw/hfi1/user_sdma.c list_add_tail(&tx->txreq.list, &req->txps); tx 930 drivers/infiniband/hw/hfi1/user_sdma.c tx->seqnum = req->seqnum++; tx 951 drivers/infiniband/hw/hfi1/user_sdma.c sdma_txclean(pq->dd, &tx->txreq); tx 953 drivers/infiniband/hw/hfi1/user_sdma.c kmem_cache_free(pq->txreq_cache, tx); tx 1157 drivers/infiniband/hw/hfi1/user_sdma.c struct user_sdma_txreq *tx, u32 datalen) tx 1160 drivers/infiniband/hw/hfi1/user_sdma.c struct hfi1_pkt_header *hdr = &tx->hdr; tx 1214 drivers/infiniband/hw/hfi1/user_sdma.c if (unlikely(tx->flags & TXREQ_FLAGS_REQ_ACK)) tx 1249 drivers/infiniband/hw/hfi1/user_sdma.c if (unlikely(tx->flags & TXREQ_FLAGS_REQ_DISABLE_SH)) tx 1267 drivers/infiniband/hw/hfi1/user_sdma.c return sdma_txadd_kvaddr(pq->dd, &tx->txreq, hdr, sizeof(*hdr)); tx 1271 drivers/infiniband/hw/hfi1/user_sdma.c struct user_sdma_txreq *tx, u32 datalen) tx 1301 drivers/infiniband/hw/hfi1/user_sdma.c if (unlikely(tx->flags & TXREQ_FLAGS_REQ_ACK)) tx 1357 drivers/infiniband/hw/hfi1/user_sdma.c if (unlikely(tx->flags & TXREQ_FLAGS_REQ_DISABLE_SH)) { tx 1378 drivers/infiniband/hw/hfi1/user_sdma.c sdma_txinit_ahg(&tx->txreq, tx 1399 drivers/infiniband/hw/hfi1/user_sdma.c struct user_sdma_txreq *tx = tx 1406 drivers/infiniband/hw/hfi1/user_sdma.c if (!tx->req) tx 1409 drivers/infiniband/hw/hfi1/user_sdma.c req = tx->req; tx 1420 drivers/infiniband/hw/hfi1/user_sdma.c req->seqcomp = tx->seqnum; tx 1421 drivers/infiniband/hw/hfi1/user_sdma.c kmem_cache_free(pq->txreq_cache, tx); tx 1446 drivers/infiniband/hw/hfi1/user_sdma.c struct user_sdma_txreq *tx = tx 1450 drivers/infiniband/hw/hfi1/user_sdma.c kmem_cache_free(req->pq->txreq_cache, tx); tx 630 drivers/infiniband/hw/hfi1/verbs.c struct verbs_txreq *tx = tx 632 drivers/infiniband/hw/hfi1/verbs.c struct rvt_qp *qp = tx->qp; tx 635 drivers/infiniband/hw/hfi1/verbs.c if (tx->wqe) { tx 636 drivers/infiniband/hw/hfi1/verbs.c rvt_send_complete(qp, tx->wqe, IB_WC_SUCCESS); tx 640 drivers/infiniband/hw/hfi1/verbs.c hdr = &tx->phdr.hdr; tx 647 drivers/infiniband/hw/hfi1/verbs.c hfi1_put_txreq(tx); tx 698 drivers/infiniband/hw/hfi1/verbs.c struct verbs_txreq *tx) tx 700 drivers/infiniband/hw/hfi1/verbs.c struct rvt_sge_state *ss = tx->ss; tx 712 drivers/infiniband/hw/hfi1/verbs.c &tx->txreq, tx 762 drivers/infiniband/hw/hfi1/verbs.c struct verbs_txreq *tx, tx 767 drivers/infiniband/hw/hfi1/verbs.c struct hfi1_sdma_header *phdr = &tx->phdr; tx 768 drivers/infiniband/hw/hfi1/verbs.c u16 hdrbytes = (tx->hdr_dwords + sizeof(pbc) / 4) << 2; tx 771 drivers/infiniband/hw/hfi1/verbs.c if (tx->phdr.hdr.hdr_type) { tx 781 drivers/infiniband/hw/hfi1/verbs.c &tx->txreq, tx 795 drivers/infiniband/hw/hfi1/verbs.c &tx->txreq, tx 802 drivers/infiniband/hw/hfi1/verbs.c &tx->txreq, tx 814 drivers/infiniband/hw/hfi1/verbs.c if (tx->ss) { tx 815 drivers/infiniband/hw/hfi1/verbs.c ret = build_verbs_ulp_payload(sde, length, tx); tx 822 drivers/infiniband/hw/hfi1/verbs.c ret = sdma_txadd_daddr(sde->dd, &tx->txreq, tx 848 drivers/infiniband/hw/hfi1/verbs.c struct verbs_txreq *tx; tx 863 drivers/infiniband/hw/hfi1/verbs.c tx = ps->s_txreq; tx 864 drivers/infiniband/hw/hfi1/verbs.c if (!sdma_txreq_built(&tx->txreq)) { tx 888 drivers/infiniband/hw/hfi1/verbs.c tx->wqe = qp->s_wqe; tx 889 drivers/infiniband/hw/hfi1/verbs.c ret = build_verbs_tx_desc(tx->sde, len, tx, ahg_info, pbc); tx 893 drivers/infiniband/hw/hfi1/verbs.c ret = sdma_send_txreq(tx->sde, ps->wait, &tx->txreq, ps->pkts_sent); tx 1219 drivers/infiniband/hw/hfi1/verbs.c struct verbs_txreq *tx = ps->s_txreq; tx 1232 drivers/infiniband/hw/hfi1/verbs.c (tx->s_cur_size + priv->s_running_pkt_size) / 2; tx 1237 drivers/infiniband/hw/hfi1/verbs.c !sdma_txreq_built(&tx->txreq)) tx 402 drivers/infiniband/hw/hfi1/verbs.h void hfi1_put_txreq(struct verbs_txreq *tx); tx 55 drivers/infiniband/hw/hfi1/verbs_txreq.c void hfi1_put_txreq(struct verbs_txreq *tx) tx 63 drivers/infiniband/hw/hfi1/verbs_txreq.c qp = tx->qp; tx 66 drivers/infiniband/hw/hfi1/verbs_txreq.c if (tx->mr) tx 67 drivers/infiniband/hw/hfi1/verbs_txreq.c rvt_put_mr(tx->mr); tx 69 drivers/infiniband/hw/hfi1/verbs_txreq.c sdma_txclean(dd_from_dev(dev), &tx->txreq); tx 72 drivers/infiniband/hw/hfi1/verbs_txreq.c kmem_cache_free(dev->verbs_txreq_cache, tx); tx 97 drivers/infiniband/hw/hfi1/verbs_txreq.c struct verbs_txreq *tx = NULL; tx 103 drivers/infiniband/hw/hfi1/verbs_txreq.c tx = kmem_cache_alloc(dev->verbs_txreq_cache, VERBS_TXREQ_GFP); tx 104 drivers/infiniband/hw/hfi1/verbs_txreq.c if (tx) tx 119 drivers/infiniband/hw/hfi1/verbs_txreq.c return tx; tx 80 drivers/infiniband/hw/hfi1/verbs_txreq.h struct verbs_txreq *tx; tx 83 drivers/infiniband/hw/hfi1/verbs_txreq.h tx = kmem_cache_alloc(dev->verbs_txreq_cache, VERBS_TXREQ_GFP); tx 84 drivers/infiniband/hw/hfi1/verbs_txreq.h if (unlikely(!tx)) { tx 86 drivers/infiniband/hw/hfi1/verbs_txreq.h tx = __get_txreq(dev, qp); tx 87 drivers/infiniband/hw/hfi1/verbs_txreq.h if (!tx) tx 88 drivers/infiniband/hw/hfi1/verbs_txreq.h return tx; tx 90 drivers/infiniband/hw/hfi1/verbs_txreq.h tx->qp = qp; tx 91 drivers/infiniband/hw/hfi1/verbs_txreq.h tx->mr = NULL; tx 92 drivers/infiniband/hw/hfi1/verbs_txreq.h tx->sde = priv->s_sde; tx 93 drivers/infiniband/hw/hfi1/verbs_txreq.h tx->psc = priv->s_sendcontext; tx 95 drivers/infiniband/hw/hfi1/verbs_txreq.h tx->txreq.num_desc = 0; tx 97 drivers/infiniband/hw/hfi1/verbs_txreq.h tx->phdr.hdr.hdr_type = priv->hdr_type; tx 98 drivers/infiniband/hw/hfi1/verbs_txreq.h tx->txreq.flags = 0; tx 99 drivers/infiniband/hw/hfi1/verbs_txreq.h return tx; tx 102 drivers/infiniband/hw/hfi1/verbs_txreq.h static inline struct sdma_txreq *get_sdma_txreq(struct verbs_txreq *tx) tx 104 drivers/infiniband/hw/hfi1/verbs_txreq.h return &tx->txreq; tx 122 drivers/infiniband/hw/hfi1/verbs_txreq.h void hfi1_put_txreq(struct verbs_txreq *tx); tx 83 drivers/infiniband/hw/hfi1/vnic_sdma.c struct vnic_txreq *tx = container_of(txreq, struct vnic_txreq, txreq); tx 84 drivers/infiniband/hw/hfi1/vnic_sdma.c struct hfi1_vnic_sdma *vnic_sdma = tx->sdma; tx 87 drivers/infiniband/hw/hfi1/vnic_sdma.c dev_kfree_skb_any(tx->skb); tx 88 drivers/infiniband/hw/hfi1/vnic_sdma.c kmem_cache_free(vnic_sdma->dd->vnic.txreq_cache, tx); tx 92 drivers/infiniband/hw/hfi1/vnic_sdma.c struct vnic_txreq *tx) tx 98 drivers/infiniband/hw/hfi1/vnic_sdma.c &tx->txreq, tx 99 drivers/infiniband/hw/hfi1/vnic_sdma.c tx->skb->data, tx 100 drivers/infiniband/hw/hfi1/vnic_sdma.c skb_headlen(tx->skb)); tx 104 drivers/infiniband/hw/hfi1/vnic_sdma.c for (i = 0; i < skb_shinfo(tx->skb)->nr_frags; i++) { tx 105 drivers/infiniband/hw/hfi1/vnic_sdma.c skb_frag_t *frag = &skb_shinfo(tx->skb)->frags[i]; tx 109 drivers/infiniband/hw/hfi1/vnic_sdma.c &tx->txreq, tx 117 drivers/infiniband/hw/hfi1/vnic_sdma.c if (tx->plen) tx 118 drivers/infiniband/hw/hfi1/vnic_sdma.c ret = sdma_txadd_kvaddr(sde->dd, &tx->txreq, tx 119 drivers/infiniband/hw/hfi1/vnic_sdma.c tx->pad + HFI1_VNIC_MAX_PAD - tx->plen, tx 120 drivers/infiniband/hw/hfi1/vnic_sdma.c tx->plen); tx 127 drivers/infiniband/hw/hfi1/vnic_sdma.c struct vnic_txreq *tx, tx 134 drivers/infiniband/hw/hfi1/vnic_sdma.c &tx->txreq, tx 136 drivers/infiniband/hw/hfi1/vnic_sdma.c hdrbytes + tx->skb->len + tx->plen, tx 146 drivers/infiniband/hw/hfi1/vnic_sdma.c tx->pbc_val = cpu_to_le64(pbc); tx 149 drivers/infiniband/hw/hfi1/vnic_sdma.c &tx->txreq, tx 150 drivers/infiniband/hw/hfi1/vnic_sdma.c &tx->pbc_val, tx 156 drivers/infiniband/hw/hfi1/vnic_sdma.c ret = build_vnic_ulp_payload(sde, tx); tx 173 drivers/infiniband/hw/hfi1/vnic_sdma.c struct vnic_txreq *tx; tx 182 drivers/infiniband/hw/hfi1/vnic_sdma.c tx = kmem_cache_alloc(dd->vnic.txreq_cache, GFP_ATOMIC); tx 183 drivers/infiniband/hw/hfi1/vnic_sdma.c if (unlikely(!tx)) { tx 188 drivers/infiniband/hw/hfi1/vnic_sdma.c tx->sdma = vnic_sdma; tx 189 drivers/infiniband/hw/hfi1/vnic_sdma.c tx->skb = skb; tx 190 drivers/infiniband/hw/hfi1/vnic_sdma.c hfi1_vnic_update_pad(tx->pad, plen); tx 191 drivers/infiniband/hw/hfi1/vnic_sdma.c tx->plen = plen; tx 192 drivers/infiniband/hw/hfi1/vnic_sdma.c ret = build_vnic_tx_desc(sde, tx, pbc); tx 197 drivers/infiniband/hw/hfi1/vnic_sdma.c &tx->txreq, vnic_sdma->pkts_sent); tx 209 drivers/infiniband/hw/hfi1/vnic_sdma.c sdma_txclean(dd, &tx->txreq); tx 210 drivers/infiniband/hw/hfi1/vnic_sdma.c kmem_cache_free(dd->vnic.txreq_cache, tx); tx 490 drivers/infiniband/hw/qib/qib_sdma.c struct qib_verbs_txreq *tx) tx 492 drivers/infiniband/hw/qib/qib_sdma.c struct qib_qp_priv *priv = tx->qp->priv; tx 496 drivers/infiniband/hw/qib/qib_sdma.c tx->txreq.start_idx = 0; tx 497 drivers/infiniband/hw/qib/qib_sdma.c tx->txreq.next_descq_idx = 0; tx 498 drivers/infiniband/hw/qib/qib_sdma.c list_add_tail(&tx->txreq.list, &ppd->sdma_activelist); tx 513 drivers/infiniband/hw/qib/qib_sdma.c struct qib_verbs_txreq *tx) tx 530 drivers/infiniband/hw/qib/qib_sdma.c complete_sdma_err_req(ppd, tx); tx 534 drivers/infiniband/hw/qib/qib_sdma.c if (tx->txreq.sg_count > qib_sdma_descq_freecnt(ppd)) { tx 543 drivers/infiniband/hw/qib/qib_sdma.c dwoffset = tx->hdr_dwords; tx 544 drivers/infiniband/hw/qib/qib_sdma.c make_sdma_desc(ppd, sdmadesc, (u64) tx->txreq.addr, dwoffset, 0); tx 547 drivers/infiniband/hw/qib/qib_sdma.c if (tx->txreq.flags & QIB_SDMA_TXREQ_F_USELARGEBUF) tx 563 drivers/infiniband/hw/qib/qib_sdma.c tx->txreq.start_idx = tail; tx 580 drivers/infiniband/hw/qib/qib_sdma.c if (tx->txreq.flags & QIB_SDMA_TXREQ_F_USELARGEBUF) tx 601 drivers/infiniband/hw/qib/qib_sdma.c if (tx->txreq.flags & QIB_SDMA_TXREQ_F_HEADTOHOST) tx 603 drivers/infiniband/hw/qib/qib_sdma.c if (tx->txreq.flags & QIB_SDMA_TXREQ_F_INTREQ) tx 605 drivers/infiniband/hw/qib/qib_sdma.c priv = tx->qp->priv; tx 607 drivers/infiniband/hw/qib/qib_sdma.c tx->txreq.next_descq_idx = tail; tx 609 drivers/infiniband/hw/qib/qib_sdma.c ppd->sdma_descq_added += tx->txreq.sg_count; tx 610 drivers/infiniband/hw/qib/qib_sdma.c list_add_tail(&tx->txreq.list, &ppd->sdma_activelist); tx 623 drivers/infiniband/hw/qib/qib_sdma.c qp = tx->qp; tx 625 drivers/infiniband/hw/qib/qib_sdma.c qib_put_txreq(tx); tx 640 drivers/infiniband/hw/qib/qib_sdma.c qp = tx->qp; tx 651 drivers/infiniband/hw/qib/qib_sdma.c tx->ss = ss; tx 652 drivers/infiniband/hw/qib/qib_sdma.c tx->dwords = dwords; tx 653 drivers/infiniband/hw/qib/qib_sdma.c priv->s_tx = tx; tx 670 drivers/infiniband/hw/qib/qib_sdma.c qib_put_txreq(tx); tx 564 drivers/infiniband/hw/qib/qib_verbs.c struct qib_verbs_txreq *tx; tx 576 drivers/infiniband/hw/qib/qib_verbs.c tx = list_entry(l, struct qib_verbs_txreq, txreq.list); tx 587 drivers/infiniband/hw/qib/qib_verbs.c tx = ERR_PTR(-EBUSY); tx 589 drivers/infiniband/hw/qib/qib_verbs.c return tx; tx 595 drivers/infiniband/hw/qib/qib_verbs.c struct qib_verbs_txreq *tx; tx 605 drivers/infiniband/hw/qib/qib_verbs.c tx = list_entry(l, struct qib_verbs_txreq, txreq.list); tx 609 drivers/infiniband/hw/qib/qib_verbs.c tx = __get_txreq(dev, qp); tx 611 drivers/infiniband/hw/qib/qib_verbs.c return tx; tx 614 drivers/infiniband/hw/qib/qib_verbs.c void qib_put_txreq(struct qib_verbs_txreq *tx) tx 621 drivers/infiniband/hw/qib/qib_verbs.c qp = tx->qp; tx 624 drivers/infiniband/hw/qib/qib_verbs.c if (tx->mr) { tx 625 drivers/infiniband/hw/qib/qib_verbs.c rvt_put_mr(tx->mr); tx 626 drivers/infiniband/hw/qib/qib_verbs.c tx->mr = NULL; tx 628 drivers/infiniband/hw/qib/qib_verbs.c if (tx->txreq.flags & QIB_SDMA_TXREQ_F_FREEBUF) { tx 629 drivers/infiniband/hw/qib/qib_verbs.c tx->txreq.flags &= ~QIB_SDMA_TXREQ_F_FREEBUF; tx 631 drivers/infiniband/hw/qib/qib_verbs.c tx->txreq.addr, tx->hdr_dwords << 2, tx 633 drivers/infiniband/hw/qib/qib_verbs.c kfree(tx->align_buf); tx 639 drivers/infiniband/hw/qib/qib_verbs.c list_add(&tx->txreq.list, &dev->txreq_free); tx 714 drivers/infiniband/hw/qib/qib_verbs.c struct qib_verbs_txreq *tx = tx 716 drivers/infiniband/hw/qib/qib_verbs.c struct rvt_qp *qp = tx->qp; tx 720 drivers/infiniband/hw/qib/qib_verbs.c if (tx->wqe) tx 721 drivers/infiniband/hw/qib/qib_verbs.c rvt_send_complete(qp, tx->wqe, IB_WC_SUCCESS); tx 725 drivers/infiniband/hw/qib/qib_verbs.c if (tx->txreq.flags & QIB_SDMA_TXREQ_F_FREEBUF) tx 726 drivers/infiniband/hw/qib/qib_verbs.c hdr = &tx->align_buf->hdr; tx 730 drivers/infiniband/hw/qib/qib_verbs.c hdr = &dev->pio_hdrs[tx->hdr_inx].hdr; tx 744 drivers/infiniband/hw/qib/qib_verbs.c qib_put_txreq(tx); tx 780 drivers/infiniband/hw/qib/qib_verbs.c struct qib_verbs_txreq *tx; tx 786 drivers/infiniband/hw/qib/qib_verbs.c tx = priv->s_tx; tx 787 drivers/infiniband/hw/qib/qib_verbs.c if (tx) { tx 790 drivers/infiniband/hw/qib/qib_verbs.c ret = qib_sdma_verbs_send(ppd, tx->ss, tx->dwords, tx); tx 794 drivers/infiniband/hw/qib/qib_verbs.c tx = get_txreq(dev, qp); tx 795 drivers/infiniband/hw/qib/qib_verbs.c if (IS_ERR(tx)) tx 800 drivers/infiniband/hw/qib/qib_verbs.c tx->qp = qp; tx 801 drivers/infiniband/hw/qib/qib_verbs.c tx->wqe = qp->s_wqe; tx 802 drivers/infiniband/hw/qib/qib_verbs.c tx->mr = qp->s_rdma_mr; tx 805 drivers/infiniband/hw/qib/qib_verbs.c tx->txreq.callback = sdma_complete; tx 807 drivers/infiniband/hw/qib/qib_verbs.c tx->txreq.flags = QIB_SDMA_TXREQ_F_HEADTOHOST; tx 809 drivers/infiniband/hw/qib/qib_verbs.c tx->txreq.flags = QIB_SDMA_TXREQ_F_INTREQ; tx 811 drivers/infiniband/hw/qib/qib_verbs.c tx->txreq.flags |= QIB_SDMA_TXREQ_F_USELARGEBUF; tx 824 drivers/infiniband/hw/qib/qib_verbs.c phdr = &dev->pio_hdrs[tx->hdr_inx]; tx 828 drivers/infiniband/hw/qib/qib_verbs.c tx->txreq.flags |= QIB_SDMA_TXREQ_F_FREEDESC; tx 829 drivers/infiniband/hw/qib/qib_verbs.c tx->txreq.sg_count = ndesc; tx 830 drivers/infiniband/hw/qib/qib_verbs.c tx->txreq.addr = dev->pio_hdrs_phys + tx 831 drivers/infiniband/hw/qib/qib_verbs.c tx->hdr_inx * sizeof(struct qib_pio_header); tx 832 drivers/infiniband/hw/qib/qib_verbs.c tx->hdr_dwords = hdrwords + 2; /* add PBC length */ tx 833 drivers/infiniband/hw/qib/qib_verbs.c ret = qib_sdma_verbs_send(ppd, ss, dwords, tx); tx 838 drivers/infiniband/hw/qib/qib_verbs.c tx->hdr_dwords = plen + 1; tx 839 drivers/infiniband/hw/qib/qib_verbs.c phdr = kmalloc(tx->hdr_dwords << 2, GFP_ATOMIC); tx 847 drivers/infiniband/hw/qib/qib_verbs.c tx->txreq.addr = dma_map_single(&dd->pcidev->dev, phdr, tx 848 drivers/infiniband/hw/qib/qib_verbs.c tx->hdr_dwords << 2, DMA_TO_DEVICE); tx 849 drivers/infiniband/hw/qib/qib_verbs.c if (dma_mapping_error(&dd->pcidev->dev, tx->txreq.addr)) tx 851 drivers/infiniband/hw/qib/qib_verbs.c tx->align_buf = phdr; tx 852 drivers/infiniband/hw/qib/qib_verbs.c tx->txreq.flags |= QIB_SDMA_TXREQ_F_FREEBUF; tx 853 drivers/infiniband/hw/qib/qib_verbs.c tx->txreq.sg_count = 1; tx 854 drivers/infiniband/hw/qib/qib_verbs.c ret = qib_sdma_verbs_send(ppd, NULL, 0, tx); tx 860 drivers/infiniband/hw/qib/qib_verbs.c qib_put_txreq(tx); tx 867 drivers/infiniband/hw/qib/qib_verbs.c ret = PTR_ERR(tx); tx 1532 drivers/infiniband/hw/qib/qib_verbs.c struct qib_verbs_txreq *tx; tx 1534 drivers/infiniband/hw/qib/qib_verbs.c tx = kzalloc(sizeof(*tx), GFP_KERNEL); tx 1535 drivers/infiniband/hw/qib/qib_verbs.c if (!tx) { tx 1539 drivers/infiniband/hw/qib/qib_verbs.c tx->hdr_inx = i; tx 1540 drivers/infiniband/hw/qib/qib_verbs.c list_add(&tx->txreq.list, &dev->txreq_free); tx 1630 drivers/infiniband/hw/qib/qib_verbs.c struct qib_verbs_txreq *tx; tx 1633 drivers/infiniband/hw/qib/qib_verbs.c tx = list_entry(l, struct qib_verbs_txreq, txreq.list); tx 1634 drivers/infiniband/hw/qib/qib_verbs.c kfree(tx); tx 1666 drivers/infiniband/hw/qib/qib_verbs.c struct qib_verbs_txreq *tx; tx 1669 drivers/infiniband/hw/qib/qib_verbs.c tx = list_entry(l, struct qib_verbs_txreq, txreq.list); tx 1670 drivers/infiniband/hw/qib/qib_verbs.c kfree(tx); tx 290 drivers/infiniband/hw/qib/qib_verbs.h void qib_put_txreq(struct qib_verbs_txreq *tx); tx 59 drivers/infiniband/hw/vmw_pvrdma/pvrdma_ring.h struct pvrdma_ring tx; /* Tx ring. */ tx 492 drivers/infiniband/sw/siw/siw.h #define tx_qp(tx) container_of(tx, struct siw_qp, tx_ctx) tx 649 drivers/infiniband/ulp/ipoib/ipoib.h static inline void ipoib_cm_set(struct ipoib_neigh *neigh, struct ipoib_cm_tx *tx) tx 651 drivers/infiniband/ulp/ipoib/ipoib.h neigh->cm = tx; tx 666 drivers/infiniband/ulp/ipoib/ipoib.h void ipoib_cm_send(struct net_device *dev, struct sk_buff *skb, struct ipoib_cm_tx *tx); tx 674 drivers/infiniband/ulp/ipoib/ipoib.h void ipoib_cm_destroy_tx(struct ipoib_cm_tx *tx); tx 706 drivers/infiniband/ulp/ipoib/ipoib.h static inline void ipoib_cm_set(struct ipoib_neigh *neigh, struct ipoib_cm_tx *tx) tx 721 drivers/infiniband/ulp/ipoib/ipoib.h void ipoib_cm_send(struct net_device *dev, struct sk_buff *skb, struct ipoib_cm_tx *tx) tx 758 drivers/infiniband/ulp/ipoib/ipoib.h void ipoib_cm_destroy_tx(struct ipoib_cm_tx *tx) tx 697 drivers/infiniband/ulp/ipoib/ipoib_cm.c struct ipoib_cm_tx *tx, tx 705 drivers/infiniband/ulp/ipoib/ipoib_cm.c return ib_post_send(tx->qp, &priv->tx_wr.wr, NULL); tx 708 drivers/infiniband/ulp/ipoib/ipoib_cm.c void ipoib_cm_send(struct net_device *dev, struct sk_buff *skb, struct ipoib_cm_tx *tx) tx 713 drivers/infiniband/ulp/ipoib/ipoib_cm.c unsigned int usable_sge = tx->max_send_sge - !!skb_headlen(skb); tx 715 drivers/infiniband/ulp/ipoib/ipoib_cm.c if (unlikely(skb->len > tx->mtu)) { tx 717 drivers/infiniband/ulp/ipoib/ipoib_cm.c skb->len, tx->mtu); tx 720 drivers/infiniband/ulp/ipoib/ipoib_cm.c ipoib_cm_skb_too_long(dev, skb, tx->mtu - IPOIB_ENCAP_LEN); tx 741 drivers/infiniband/ulp/ipoib/ipoib_cm.c tx->tx_head, skb->len, tx->qp->qp_num); tx 750 drivers/infiniband/ulp/ipoib/ipoib_cm.c tx_req = &tx->tx_ring[tx->tx_head & (ipoib_sendq_size - 1)]; tx 762 drivers/infiniband/ulp/ipoib/ipoib_cm.c tx->qp->qp_num); tx 778 drivers/infiniband/ulp/ipoib/ipoib_cm.c rc = post_send(priv, tx, tx->tx_head & (ipoib_sendq_size - 1), tx_req); tx 789 drivers/infiniband/ulp/ipoib/ipoib_cm.c ++tx->tx_head; tx 797 drivers/infiniband/ulp/ipoib/ipoib_cm.c struct ipoib_cm_tx *tx = wc->qp->qp_context; tx 811 drivers/infiniband/ulp/ipoib/ipoib_cm.c tx_req = &tx->tx_ring[wr_id]; tx 823 drivers/infiniband/ulp/ipoib/ipoib_cm.c ++tx->tx_tail; tx 850 drivers/infiniband/ulp/ipoib/ipoib_cm.c neigh = tx->neigh; tx 856 drivers/infiniband/ulp/ipoib/ipoib_cm.c tx->neigh = NULL; tx 859 drivers/infiniband/ulp/ipoib/ipoib_cm.c if (test_and_clear_bit(IPOIB_FLAG_INITIALIZED, &tx->flags)) { tx 860 drivers/infiniband/ulp/ipoib/ipoib_cm.c list_move(&tx->list, &priv->cm.reap_list); tx 864 drivers/infiniband/ulp/ipoib/ipoib_cm.c clear_bit(IPOIB_FLAG_OPER_UP, &tx->flags); tx 1057 drivers/infiniband/ulp/ipoib/ipoib_cm.c static struct ib_qp *ipoib_cm_create_tx_qp(struct net_device *dev, struct ipoib_cm_tx *tx) tx 1068 drivers/infiniband/ulp/ipoib/ipoib_cm.c .qp_context = tx, tx 1078 drivers/infiniband/ulp/ipoib/ipoib_cm.c tx->max_send_sge = attr.cap.max_send_sge; tx 1256 drivers/infiniband/ulp/ipoib/ipoib_cm.c struct ipoib_cm_tx *tx = cm_id->context; tx 1257 drivers/infiniband/ulp/ipoib/ipoib_cm.c struct ipoib_dev_priv *priv = ipoib_priv(tx->dev); tx 1281 drivers/infiniband/ulp/ipoib/ipoib_cm.c neigh = tx->neigh; tx 1287 drivers/infiniband/ulp/ipoib/ipoib_cm.c tx->neigh = NULL; tx 1290 drivers/infiniband/ulp/ipoib/ipoib_cm.c if (test_and_clear_bit(IPOIB_FLAG_INITIALIZED, &tx->flags)) { tx 1291 drivers/infiniband/ulp/ipoib/ipoib_cm.c list_move(&tx->list, &priv->cm.reap_list); tx 1309 drivers/infiniband/ulp/ipoib/ipoib_cm.c struct ipoib_cm_tx *tx; tx 1311 drivers/infiniband/ulp/ipoib/ipoib_cm.c tx = kzalloc(sizeof(*tx), GFP_ATOMIC); tx 1312 drivers/infiniband/ulp/ipoib/ipoib_cm.c if (!tx) tx 1315 drivers/infiniband/ulp/ipoib/ipoib_cm.c neigh->cm = tx; tx 1316 drivers/infiniband/ulp/ipoib/ipoib_cm.c tx->neigh = neigh; tx 1317 drivers/infiniband/ulp/ipoib/ipoib_cm.c tx->dev = dev; tx 1318 drivers/infiniband/ulp/ipoib/ipoib_cm.c list_add(&tx->list, &priv->cm.start_list); tx 1319 drivers/infiniband/ulp/ipoib/ipoib_cm.c set_bit(IPOIB_FLAG_INITIALIZED, &tx->flags); tx 1321 drivers/infiniband/ulp/ipoib/ipoib_cm.c return tx; tx 1324 drivers/infiniband/ulp/ipoib/ipoib_cm.c void ipoib_cm_destroy_tx(struct ipoib_cm_tx *tx) tx 1326 drivers/infiniband/ulp/ipoib/ipoib_cm.c struct ipoib_dev_priv *priv = ipoib_priv(tx->dev); tx 1328 drivers/infiniband/ulp/ipoib/ipoib_cm.c if (test_and_clear_bit(IPOIB_FLAG_INITIALIZED, &tx->flags)) { tx 1330 drivers/infiniband/ulp/ipoib/ipoib_cm.c list_move(&tx->list, &priv->cm.reap_list); tx 1333 drivers/infiniband/ulp/ipoib/ipoib_cm.c tx->neigh->daddr + 4); tx 1334 drivers/infiniband/ulp/ipoib/ipoib_cm.c tx->neigh = NULL; tx 80 drivers/input/gameport/gameport.c unsigned int i, t, tx; tx 87 drivers/input/gameport/gameport.c tx = ~0; tx 99 drivers/input/gameport/gameport.c if (t < tx) tx 100 drivers/input/gameport/gameport.c tx = t; tx 105 drivers/input/gameport/gameport.c if (tx) tx 106 drivers/input/gameport/gameport.c t /= tx; tx 114 drivers/input/gameport/gameport.c unsigned int i, t, t1, t2, t3, tx; tx 120 drivers/input/gameport/gameport.c tx = 1 << 30; tx 130 drivers/input/gameport/gameport.c if ((t = DELTA(t2,t1) - DELTA(t3,t2)) < tx) tx = t; tx 134 drivers/input/gameport/gameport.c return 59659 / (tx < 1 ? 1 : tx); tx 139 drivers/input/gameport/gameport.c unsigned long tx, t1, t2, flags; tx 144 drivers/input/gameport/gameport.c tx = 1 << 30; tx 153 drivers/input/gameport/gameport.c if (t2 - t1 < tx) tx = t2 - t1; tx 158 drivers/input/gameport/gameport.c (unsigned long)HZ / (1000 / 50)) / (tx < 1 ? 1 : tx); tx 377 drivers/input/joystick/analog.c unsigned int i, t, tx; tx 397 drivers/input/joystick/analog.c tx = ~0; tx 410 drivers/input/joystick/analog.c if (t < tx) tx = t; tx 413 drivers/input/joystick/analog.c port->loop = tx / 50; tx 216 drivers/input/rmi4/rmi_f54.c u8 tx = drv_data->num_tx_electrodes ? : f54->num_tx_electrodes; tx 221 drivers/input/rmi4/rmi_f54.c size = rx * tx; tx 228 drivers/input/rmi4/rmi_f54.c size = sizeof(u16) * rx * tx; tx 403 drivers/input/rmi4/rmi_f54.c u8 tx = drv_data->num_tx_electrodes ? : f54->num_tx_electrodes; tx 419 drivers/input/rmi4/rmi_f54.c f->height = tx; tx 651 drivers/input/rmi4/rmi_f54.c u8 rx, tx; tx 668 drivers/input/rmi4/rmi_f54.c tx = f54->num_tx_electrodes; tx 670 drivers/input/rmi4/rmi_f54.c array3_size(tx, rx, sizeof(u16)), tx 62 drivers/input/serio/maceps2.c port->tx = val; tx 23 drivers/input/touchscreen/tsc2004.c u8 tx = TSC200X_CMD | TSC200X_CMD_12BIT | cmd; tx 27 drivers/input/touchscreen/tsc2004.c data = i2c_smbus_write_byte(i2c, tx); tx 26 drivers/input/touchscreen/tsc2005.c u8 tx = TSC200X_CMD | TSC200X_CMD_12BIT | cmd; tx 28 drivers/input/touchscreen/tsc2005.c .tx_buf = &tx, tx 106 drivers/ipack/devices/ipoctal.c stats->tx = 0; tx 140 drivers/ipack/devices/ipoctal.c icount->tx = channel->stats.tx; tx 203 drivers/ipack/devices/ipoctal.c channel->stats.tx++; tx 30 drivers/ipack/devices/ipoctal.h unsigned long tx; tx 236 drivers/leds/trigger/ledtrig-netdev.c static DEVICE_ATTR_RW(tx); tx 990 drivers/mailbox/bcm-pdc-mailbox.c struct pdc_ring_alloc tx; tx 994 drivers/mailbox/bcm-pdc-mailbox.c tx.vbase = dma_pool_zalloc(pdcs->ring_pool, GFP_KERNEL, &tx.dmabase); tx 995 drivers/mailbox/bcm-pdc-mailbox.c if (unlikely(!tx.vbase)) { tx 1007 drivers/mailbox/bcm-pdc-mailbox.c dev_dbg(dev, " - base DMA addr of tx ring %pad", &tx.dmabase); tx 1008 drivers/mailbox/bcm-pdc-mailbox.c dev_dbg(dev, " - base virtual addr of tx ring %p", tx.vbase); tx 1012 drivers/mailbox/bcm-pdc-mailbox.c memcpy(&pdcs->tx_ring_alloc, &tx, sizeof(tx)); tx 1078 drivers/mailbox/bcm-pdc-mailbox.c dma_pool_free(pdcs->ring_pool, tx.vbase, tx.dmabase); tx 39 drivers/md/raid5-log.h struct dma_async_tx_descriptor *tx); tx 159 drivers/md/raid5-ppl.c struct dma_async_tx_descriptor *tx) tx 189 drivers/md/raid5-ppl.c return tx; tx 192 drivers/md/raid5-ppl.c init_async_submit(&submit, ASYNC_TX_FENCE|ASYNC_TX_XOR_ZERO_DST, tx, tx 196 drivers/md/raid5-ppl.c tx = async_memcpy(sh->ppl_page, srcs[0], 0, 0, PAGE_SIZE, tx 199 drivers/md/raid5-ppl.c tx = async_xor(sh->ppl_page, srcs, 0, count, PAGE_SIZE, tx 202 drivers/md/raid5-ppl.c return tx; tx 719 drivers/md/raid5-ppl.c struct dma_async_tx_descriptor *tx; tx 724 drivers/md/raid5-ppl.c tx = async_xor(page1, xor_srcs, 0, 2, size, &submit); tx 726 drivers/md/raid5-ppl.c async_tx_quiesce(&tx); tx 1229 drivers/md/raid5.c sector_t sector, struct dma_async_tx_descriptor *tx, tx 1246 drivers/md/raid5.c init_async_submit(&submit, flags, tx, NULL, NULL, NULL); tx 1274 drivers/md/raid5.c tx = async_memcpy(*page, bio_page, page_offset, tx 1277 drivers/md/raid5.c tx = async_memcpy(bio_page, *page, b_offset, tx 1281 drivers/md/raid5.c submit.depend_tx = tx; tx 1288 drivers/md/raid5.c return tx; tx 1330 drivers/md/raid5.c struct dma_async_tx_descriptor *tx = NULL; tx 1348 drivers/md/raid5.c tx = async_copy_data(0, rbi, &dev->page, tx 1349 drivers/md/raid5.c dev->sector, tx, sh, 0); tx 1356 drivers/md/raid5.c init_async_submit(&submit, ASYNC_TX_ACK, tx, ops_complete_biofill, sh, NULL); tx 1413 drivers/md/raid5.c struct dma_async_tx_descriptor *tx; tx 1432 drivers/md/raid5.c tx = async_memcpy(xor_dest, xor_srcs[0], 0, 0, STRIPE_SIZE, &submit); tx 1434 drivers/md/raid5.c tx = async_xor(xor_dest, xor_srcs, 0, count, STRIPE_SIZE, &submit); tx 1436 drivers/md/raid5.c return tx; tx 1493 drivers/md/raid5.c struct dma_async_tx_descriptor *tx; tx 1525 drivers/md/raid5.c tx = async_gen_syndrome(blocks, 0, count+2, STRIPE_SIZE, &submit); tx 1538 drivers/md/raid5.c tx = async_xor(dest, blocks, 0, count, STRIPE_SIZE, &submit); tx 1541 drivers/md/raid5.c return tx; tx 1555 drivers/md/raid5.c struct dma_async_tx_descriptor *tx; tx 1624 drivers/md/raid5.c tx = async_xor(dest, blocks, 0, count, STRIPE_SIZE, tx 1628 drivers/md/raid5.c init_async_submit(&submit, ASYNC_TX_FENCE, tx, tx 1669 drivers/md/raid5.c struct dma_async_tx_descriptor *tx) tx 1692 drivers/md/raid5.c init_async_submit(&submit, ASYNC_TX_FENCE|ASYNC_TX_XOR_DROP_DST, tx, tx 1694 drivers/md/raid5.c tx = async_xor(xor_dest, xor_srcs, 0, count, STRIPE_SIZE, &submit); tx 1696 drivers/md/raid5.c return tx; tx 1701 drivers/md/raid5.c struct dma_async_tx_descriptor *tx) tx 1712 drivers/md/raid5.c init_async_submit(&submit, ASYNC_TX_FENCE|ASYNC_TX_PQ_XOR_DST, tx, tx 1714 drivers/md/raid5.c tx = async_gen_syndrome(blocks, 0, count+2, STRIPE_SIZE, &submit); tx 1716 drivers/md/raid5.c return tx; tx 1720 drivers/md/raid5.c ops_run_biodrain(struct stripe_head *sh, struct dma_async_tx_descriptor *tx) tx 1763 drivers/md/raid5.c tx = async_copy_data(1, wbi, &dev->page, tx 1764 drivers/md/raid5.c dev->sector, tx, sh, tx 1787 drivers/md/raid5.c return tx; tx 1839 drivers/md/raid5.c struct dma_async_tx_descriptor *tx) tx 1904 drivers/md/raid5.c init_async_submit(&submit, flags, tx, ops_complete_reconstruct, head_sh, tx 1908 drivers/md/raid5.c init_async_submit(&submit, flags, tx, NULL, NULL, tx 1913 drivers/md/raid5.c tx = async_memcpy(xor_dest, xor_srcs[0], 0, 0, STRIPE_SIZE, &submit); tx 1915 drivers/md/raid5.c tx = async_xor(xor_dest, xor_srcs, 0, count, STRIPE_SIZE, &submit); tx 1926 drivers/md/raid5.c struct dma_async_tx_descriptor *tx) tx 1970 drivers/md/raid5.c init_async_submit(&submit, txflags, tx, ops_complete_reconstruct, tx 1973 drivers/md/raid5.c init_async_submit(&submit, 0, tx, NULL, NULL, tx 1975 drivers/md/raid5.c tx = async_gen_syndrome(blocks, 0, count+2, STRIPE_SIZE, &submit); tx 2003 drivers/md/raid5.c struct dma_async_tx_descriptor *tx; tx 2023 drivers/md/raid5.c tx = async_xor_val(xor_dest, xor_srcs, 0, count, STRIPE_SIZE, tx 2027 drivers/md/raid5.c init_async_submit(&submit, ASYNC_TX_ACK, tx, ops_complete_check, sh, NULL); tx 2028 drivers/md/raid5.c tx = async_trigger_callback(&submit); tx 2055 drivers/md/raid5.c struct dma_async_tx_descriptor *tx = NULL; tx 2070 drivers/md/raid5.c tx = ops_run_compute5(sh, percpu); tx 2073 drivers/md/raid5.c tx = ops_run_compute6_1(sh, percpu); tx 2075 drivers/md/raid5.c tx = ops_run_compute6_2(sh, percpu); tx 2078 drivers/md/raid5.c if (tx && !test_bit(STRIPE_OP_RECONSTRUCT, &ops_request)) tx 2079 drivers/md/raid5.c async_tx_ack(tx); tx 2084 drivers/md/raid5.c tx = ops_run_prexor5(sh, percpu, tx); tx 2086 drivers/md/raid5.c tx = ops_run_prexor6(sh, percpu, tx); tx 2090 drivers/md/raid5.c tx = ops_run_partial_parity(sh, percpu, tx); tx 2093 drivers/md/raid5.c tx = ops_run_biodrain(sh, tx); tx 2099 drivers/md/raid5.c ops_run_reconstruct5(sh, percpu, tx); tx 2101 drivers/md/raid5.c ops_run_reconstruct6(sh, percpu, tx); tx 4310 drivers/md/raid5.c struct dma_async_tx_descriptor *tx = NULL; tx 4337 drivers/md/raid5.c init_async_submit(&submit, 0, tx, NULL, NULL, NULL); tx 4338 drivers/md/raid5.c tx = async_memcpy(sh2->dev[dd_idx].page, tx 4357 drivers/md/raid5.c async_tx_quiesce(&tx); tx 19 drivers/media/dvb-frontends/cxd2880/cxd2880_spi_device.c struct spi_transfer tx; tx 27 drivers/media/dvb-frontends/cxd2880/cxd2880_spi_device.c memset(&tx, 0, sizeof(tx)); tx 28 drivers/media/dvb-frontends/cxd2880/cxd2880_spi_device.c tx.tx_buf = data; tx 29 drivers/media/dvb-frontends/cxd2880/cxd2880_spi_device.c tx.len = size; tx 32 drivers/media/dvb-frontends/cxd2880/cxd2880_spi_device.c spi_message_add_tail(&tx, &msg); tx 2014 drivers/media/dvb-frontends/dib7000p.c u8 *tx, *rx; tx 2021 drivers/media/dvb-frontends/dib7000p.c tx = kzalloc(2, GFP_KERNEL); tx 2022 drivers/media/dvb-frontends/dib7000p.c if (!tx) tx 2030 drivers/media/dvb-frontends/dib7000p.c msg[0].buf = tx; tx 2033 drivers/media/dvb-frontends/dib7000p.c tx[0] = 0x03; tx 2034 drivers/media/dvb-frontends/dib7000p.c tx[1] = 0x00; tx 2057 drivers/media/dvb-frontends/dib7000p.c kfree(tx); tx 208 drivers/media/dvb-frontends/dib9000.c static int dib9000_risc_apb_access_read(struct dib9000_state *state, u32 address, u16 attribute, const u8 * tx, u32 txlen, u8 * b, u32 len); tx 1012 drivers/media/dvb-frontends/dib9000.c static int dib9000_risc_apb_access_read(struct dib9000_state *state, u32 address, u16 attribute, const u8 * tx, u32 txlen, u8 * b, u32 len) tx 263 drivers/media/dvb-frontends/dibx000_common.c static int dibx000_i2c_gate_ctrl(struct dibx000_i2c_master *mst, u8 tx[4], tx 277 drivers/media/dvb-frontends/dibx000_common.c tx[0] = (((mst->base_reg + 1) >> 8) & 0xff); tx 278 drivers/media/dvb-frontends/dibx000_common.c tx[1] = ((mst->base_reg + 1) & 0xff); tx 279 drivers/media/dvb-frontends/dibx000_common.c tx[2] = val >> 8; tx 280 drivers/media/dvb-frontends/dibx000_common.c tx[3] = val & 0xff; tx 413 drivers/media/dvb-frontends/dibx000_common.c u8 tx[4]; tx 414 drivers/media/dvb-frontends/dibx000_common.c struct i2c_msg m = {.addr = mst->i2c_addr,.buf = tx,.len = 4 }; tx 416 drivers/media/dvb-frontends/dibx000_common.c dibx000_i2c_gate_ctrl(mst, tx, 0, 0); tx 285 drivers/media/i2c/adv748x/adv748x-afe.c ret = adv748x_tx_power(afe->tx, enable); tx 319 drivers/media/i2c/adv748x/adv748x-afe.c struct v4l2_subdev *tx; tx 321 drivers/media/i2c/adv748x/adv748x-afe.c tx = adv748x_get_remote_sd(&afe->pads[ADV748X_AFE_SOURCE]); tx 322 drivers/media/i2c/adv748x/adv748x-afe.c if (!tx) tx 330 drivers/media/i2c/adv748x/adv748x-afe.c return adv748x_csi2_set_pixelrate(tx, 14318180); tx 237 drivers/media/i2c/adv748x/adv748x-core.c static int adv748x_power_up_tx(struct adv748x_csi2 *tx) tx 239 drivers/media/i2c/adv748x/adv748x-core.c struct adv748x_state *state = tx->state; tx 240 drivers/media/i2c/adv748x/adv748x-core.c u8 page = is_txa(tx) ? ADV748X_PAGE_TXA : ADV748X_PAGE_TXB; tx 244 drivers/media/i2c/adv748x/adv748x-core.c adv748x_write_check(state, page, 0x00, 0x80 | tx->num_lanes, &ret); tx 247 drivers/media/i2c/adv748x/adv748x-core.c adv748x_write_check(state, page, 0x00, 0xa0 | tx->num_lanes, &ret); tx 250 drivers/media/i2c/adv748x/adv748x-core.c if (tx->src == &state->hdmi.sd) { tx 273 drivers/media/i2c/adv748x/adv748x-core.c adv748x_write_check(state, page, 0x00, 0x20 | tx->num_lanes, &ret); tx 284 drivers/media/i2c/adv748x/adv748x-core.c static int adv748x_power_down_tx(struct adv748x_csi2 *tx) tx 286 drivers/media/i2c/adv748x/adv748x-core.c struct adv748x_state *state = tx->state; tx 287 drivers/media/i2c/adv748x/adv748x-core.c u8 page = is_txa(tx) ? ADV748X_PAGE_TXA : ADV748X_PAGE_TXB; tx 295 drivers/media/i2c/adv748x/adv748x-core.c adv748x_write_check(state, page, 0x00, 0x80 | tx->num_lanes, &ret); tx 306 drivers/media/i2c/adv748x/adv748x-core.c int adv748x_tx_power(struct adv748x_csi2 *tx, bool on) tx 310 drivers/media/i2c/adv748x/adv748x-core.c if (!is_tx_enabled(tx)) tx 313 drivers/media/i2c/adv748x/adv748x-core.c val = tx_read(tx, ADV748X_CSI_FS_AS_LS); tx 325 drivers/media/i2c/adv748x/adv748x-core.c return on ? adv748x_power_up_tx(tx) : adv748x_power_down_tx(tx); tx 338 drivers/media/i2c/adv748x/adv748x-core.c struct adv748x_csi2 *tx = adv748x_sd_to_csi2(sd); tx 346 drivers/media/i2c/adv748x/adv748x-core.c if (enable && tx->src) tx 351 drivers/media/i2c/adv748x/adv748x-core.c state->afe.tx = enable ? tx : NULL; tx 353 drivers/media/i2c/adv748x/adv748x-core.c state->hdmi.tx = enable ? tx : NULL; tx 355 drivers/media/i2c/adv748x/adv748x-core.c tx->src = enable ? rsd : NULL; tx 357 drivers/media/i2c/adv748x/adv748x-core.c if (state->afe.tx) { tx 360 drivers/media/i2c/adv748x/adv748x-core.c if (is_txa(tx)) tx 366 drivers/media/i2c/adv748x/adv748x-core.c if (state->hdmi.tx) tx 17 drivers/media/i2c/adv748x/adv748x-csi2.c static int adv748x_csi2_set_virtual_channel(struct adv748x_csi2 *tx, tx 20 drivers/media/i2c/adv748x/adv748x-csi2.c return tx_write(tx, ADV748X_CSI_VC_REF, vc << ADV748X_CSI_VC_REF_SHIFT); tx 35 drivers/media/i2c/adv748x/adv748x-csi2.c static int adv748x_csi2_register_link(struct adv748x_csi2 *tx, tx 50 drivers/media/i2c/adv748x/adv748x-csi2.c &tx->sd.entity, ADV748X_CSI2_SINK, tx 56 drivers/media/i2c/adv748x/adv748x-csi2.c tx->src = src; tx 71 drivers/media/i2c/adv748x/adv748x-csi2.c struct adv748x_csi2 *tx = adv748x_sd_to_csi2(sd); tx 72 drivers/media/i2c/adv748x/adv748x-csi2.c struct adv748x_state *state = tx->state; tx 75 drivers/media/i2c/adv748x/adv748x-csi2.c adv_dbg(state, "Registered %s (%s)", is_txa(tx) ? "TXA":"TXB", tx 85 drivers/media/i2c/adv748x/adv748x-csi2.c ret = adv748x_csi2_register_link(tx, sd->v4l2_dev, tx 88 drivers/media/i2c/adv748x/adv748x-csi2.c is_txb(tx)); tx 93 drivers/media/i2c/adv748x/adv748x-csi2.c if (is_txb(tx)) tx 94 drivers/media/i2c/adv748x/adv748x-csi2.c state->afe.tx = tx; tx 98 drivers/media/i2c/adv748x/adv748x-csi2.c if (is_txb(tx) || !is_hdmi_enabled(state)) tx 101 drivers/media/i2c/adv748x/adv748x-csi2.c ret = adv748x_csi2_register_link(tx, sd->v4l2_dev, &state->hdmi.sd, tx 107 drivers/media/i2c/adv748x/adv748x-csi2.c state->hdmi.tx = tx; tx 122 drivers/media/i2c/adv748x/adv748x-csi2.c struct adv748x_csi2 *tx = adv748x_sd_to_csi2(sd); tx 125 drivers/media/i2c/adv748x/adv748x-csi2.c src = adv748x_get_remote_sd(&tx->pads[ADV748X_CSI2_SINK]); tx 148 drivers/media/i2c/adv748x/adv748x-csi2.c struct adv748x_csi2 *tx = adv748x_sd_to_csi2(sd); tx 153 drivers/media/i2c/adv748x/adv748x-csi2.c return &tx->format; tx 160 drivers/media/i2c/adv748x/adv748x-csi2.c struct adv748x_csi2 *tx = adv748x_sd_to_csi2(sd); tx 161 drivers/media/i2c/adv748x/adv748x-csi2.c struct adv748x_state *state = tx->state; tx 182 drivers/media/i2c/adv748x/adv748x-csi2.c struct adv748x_csi2 *tx = adv748x_sd_to_csi2(sd); tx 183 drivers/media/i2c/adv748x/adv748x-csi2.c struct adv748x_state *state = tx->state; tx 237 drivers/media/i2c/adv748x/adv748x-csi2.c struct adv748x_csi2 *tx = adv748x_sd_to_csi2(sd); tx 239 drivers/media/i2c/adv748x/adv748x-csi2.c if (!tx->pixel_rate) tx 242 drivers/media/i2c/adv748x/adv748x-csi2.c return v4l2_ctrl_s_ctrl_int64(tx->pixel_rate, rate); tx 259 drivers/media/i2c/adv748x/adv748x-csi2.c static int adv748x_csi2_init_controls(struct adv748x_csi2 *tx) tx 262 drivers/media/i2c/adv748x/adv748x-csi2.c v4l2_ctrl_handler_init(&tx->ctrl_hdl, 1); tx 264 drivers/media/i2c/adv748x/adv748x-csi2.c tx->pixel_rate = v4l2_ctrl_new_std(&tx->ctrl_hdl, tx 269 drivers/media/i2c/adv748x/adv748x-csi2.c tx->sd.ctrl_handler = &tx->ctrl_hdl; tx 270 drivers/media/i2c/adv748x/adv748x-csi2.c if (tx->ctrl_hdl.error) { tx 271 drivers/media/i2c/adv748x/adv748x-csi2.c v4l2_ctrl_handler_free(&tx->ctrl_hdl); tx 272 drivers/media/i2c/adv748x/adv748x-csi2.c return tx->ctrl_hdl.error; tx 275 drivers/media/i2c/adv748x/adv748x-csi2.c return v4l2_ctrl_handler_setup(&tx->ctrl_hdl); tx 278 drivers/media/i2c/adv748x/adv748x-csi2.c int adv748x_csi2_init(struct adv748x_state *state, struct adv748x_csi2 *tx) tx 282 drivers/media/i2c/adv748x/adv748x-csi2.c if (!is_tx_enabled(tx)) tx 286 drivers/media/i2c/adv748x/adv748x-csi2.c adv748x_csi2_set_virtual_channel(tx, 0); tx 288 drivers/media/i2c/adv748x/adv748x-csi2.c adv748x_subdev_init(&tx->sd, state, &adv748x_csi2_ops, tx 290 drivers/media/i2c/adv748x/adv748x-csi2.c is_txa(tx) ? "txa" : "txb"); tx 293 drivers/media/i2c/adv748x/adv748x-csi2.c tx->sd.fwnode = of_fwnode_handle(state->endpoints[tx->port]); tx 296 drivers/media/i2c/adv748x/adv748x-csi2.c tx->sd.internal_ops = &adv748x_csi2_internal_ops; tx 298 drivers/media/i2c/adv748x/adv748x-csi2.c tx->pads[ADV748X_CSI2_SINK].flags = MEDIA_PAD_FL_SINK; tx 299 drivers/media/i2c/adv748x/adv748x-csi2.c tx->pads[ADV748X_CSI2_SOURCE].flags = MEDIA_PAD_FL_SOURCE; tx 301 drivers/media/i2c/adv748x/adv748x-csi2.c ret = media_entity_pads_init(&tx->sd.entity, ADV748X_CSI2_NR_PADS, tx 302 drivers/media/i2c/adv748x/adv748x-csi2.c tx->pads); tx 306 drivers/media/i2c/adv748x/adv748x-csi2.c ret = adv748x_csi2_init_controls(tx); tx 310 drivers/media/i2c/adv748x/adv748x-csi2.c ret = v4l2_async_register_subdev(&tx->sd); tx 317 drivers/media/i2c/adv748x/adv748x-csi2.c v4l2_ctrl_handler_free(&tx->ctrl_hdl); tx 319 drivers/media/i2c/adv748x/adv748x-csi2.c media_entity_cleanup(&tx->sd.entity); tx 324 drivers/media/i2c/adv748x/adv748x-csi2.c void adv748x_csi2_cleanup(struct adv748x_csi2 *tx) tx 326 drivers/media/i2c/adv748x/adv748x-csi2.c if (!is_tx_enabled(tx)) tx 329 drivers/media/i2c/adv748x/adv748x-csi2.c v4l2_async_unregister_subdev(&tx->sd); tx 330 drivers/media/i2c/adv748x/adv748x-csi2.c media_entity_cleanup(&tx->sd.entity); tx 331 drivers/media/i2c/adv748x/adv748x-csi2.c v4l2_ctrl_handler_free(&tx->ctrl_hdl); tx 361 drivers/media/i2c/adv748x/adv748x-hdmi.c ret = adv748x_tx_power(hdmi->tx, enable); tx 399 drivers/media/i2c/adv748x/adv748x-hdmi.c struct v4l2_subdev *tx; tx 402 drivers/media/i2c/adv748x/adv748x-hdmi.c tx = adv748x_get_remote_sd(&hdmi->pads[ADV748X_HDMI_SOURCE]); tx 403 drivers/media/i2c/adv748x/adv748x-hdmi.c if (!tx) tx 408 drivers/media/i2c/adv748x/adv748x-hdmi.c return adv748x_csi2_set_pixelrate(tx, timings.bt.pixelclock); tx 124 drivers/media/i2c/adv748x/adv748x.h struct adv748x_csi2 *tx; tx 156 drivers/media/i2c/adv748x/adv748x.h struct adv748x_csi2 *tx; tx 433 drivers/media/i2c/adv748x/adv748x.h int adv748x_tx_power(struct adv748x_csi2 *tx, bool on); tx 438 drivers/media/i2c/adv748x/adv748x.h int adv748x_csi2_init(struct adv748x_state *state, struct adv748x_csi2 *tx); tx 439 drivers/media/i2c/adv748x/adv748x.h void adv748x_csi2_cleanup(struct adv748x_csi2 *tx); tx 93 drivers/media/i2c/st-mipid02.c struct v4l2_fwnode_endpoint tx; tx 489 drivers/media/i2c/st-mipid02.c struct v4l2_fwnode_endpoint *ep = &bridge->tx; tx 930 drivers/media/i2c/st-mipid02.c bridge->tx = ep; tx 277 drivers/media/platform/exynos-gsc/gsc-core.c u32 tx, u32 ty, u32 *hratio) tx 295 drivers/media/platform/exynos-gsc/gsc-core.c gsc_cal_prescaler_ratio(var, s_frame->crop.width, tx, wratio); tx 641 drivers/media/platform/exynos-gsc/gsc-core.c int tx, ty; tx 655 drivers/media/platform/exynos-gsc/gsc-core.c tx = d_frame->crop.height; tx 657 drivers/media/platform/exynos-gsc/gsc-core.c tx = d_frame->crop.width; tx 661 drivers/media/platform/exynos-gsc/gsc-core.c if (tx <= 0 || ty <= 0) { tx 662 drivers/media/platform/exynos-gsc/gsc-core.c dev_err(dev, "Invalid target size: %dx%d", tx, ty); tx 667 drivers/media/platform/exynos-gsc/gsc-core.c tx, &sc->pre_hratio); tx 681 drivers/media/platform/exynos-gsc/gsc-core.c tx, ty, &sc->pre_vratio); tx 686 drivers/media/platform/exynos-gsc/gsc-core.c sc->main_hratio = (s_frame->crop.width << 16) / tx; tx 690 drivers/media/platform/exynos-gsc/gsc-core.c s_frame->crop.width, s_frame->crop.height, tx, ty); tx 397 drivers/media/platform/exynos-gsc/gsc-core.h u32 *wratio, u32 tx, u32 ty, u32 *hratio); tx 232 drivers/media/platform/exynos4-is/fimc-core.c int tx, ty, sx, sy; tx 237 drivers/media/platform/exynos4-is/fimc-core.c tx = d_frame->height; tx 239 drivers/media/platform/exynos4-is/fimc-core.c tx = d_frame->width; tx 242 drivers/media/platform/exynos4-is/fimc-core.c if (tx <= 0 || ty <= 0) { tx 243 drivers/media/platform/exynos4-is/fimc-core.c dev_err(dev, "Invalid target size: %dx%d\n", tx, ty); tx 256 drivers/media/platform/exynos4-is/fimc-core.c ret = fimc_get_scaler_factor(sx, tx, &sc->pre_hratio, &sc->hfactor); tx 268 drivers/media/platform/exynos4-is/fimc-core.c sc->main_hratio = (sx << 14) / (tx << sc->hfactor); tx 271 drivers/media/platform/exynos4-is/fimc-core.c sc->main_hratio = (sx << 8) / (tx << sc->hfactor); tx 276 drivers/media/platform/exynos4-is/fimc-core.c sc->scaleup_h = (tx >= sx) ? 1 : 0; tx 210 drivers/media/platform/m2m-deinterlace.c struct dma_async_tx_descriptor *tx; tx 328 drivers/media/platform/m2m-deinterlace.c tx = dmadev->device_prep_interleaved_dma(chan, ctx->xt, flags); tx 329 drivers/media/platform/m2m-deinterlace.c if (tx == NULL) { tx 335 drivers/media/platform/m2m-deinterlace.c tx->callback = dma_callback; tx 336 drivers/media/platform/m2m-deinterlace.c tx->callback_param = ctx; tx 339 drivers/media/platform/m2m-deinterlace.c ctx->cookie = dmaengine_submit(tx); tx 233 drivers/media/platform/omap/omap_vout_vrfb.c struct dma_async_tx_descriptor *tx; tx 270 drivers/media/platform/omap/omap_vout_vrfb.c tx = dmaengine_prep_interleaved_dma(chan, xt, flags); tx 271 drivers/media/platform/omap/omap_vout_vrfb.c if (tx == NULL) { tx 276 drivers/media/platform/omap/omap_vout_vrfb.c tx->callback = omap_vout_vrfb_dma_tx_callback; tx 277 drivers/media/platform/omap/omap_vout_vrfb.c tx->callback_param = &vout->vrfb_dma_tx; tx 279 drivers/media/platform/omap/omap_vout_vrfb.c cookie = dmaengine_submit(tx); tx 176 drivers/media/platform/omap3isp/isphist.c struct dma_async_tx_descriptor *tx; tx 203 drivers/media/platform/omap3isp/isphist.c tx = dmaengine_prep_slave_single(hist->dma_ch, dma_addr, tx 206 drivers/media/platform/omap3isp/isphist.c if (tx == NULL) { tx 212 drivers/media/platform/omap3isp/isphist.c tx->callback = hist_dma_cb; tx 213 drivers/media/platform/omap3isp/isphist.c tx->callback_param = hist; tx 214 drivers/media/platform/omap3isp/isphist.c cookie = tx->tx_submit(tx); tx 864 drivers/media/platform/pxa_camera.c struct dma_async_tx_descriptor *tx; tx 866 drivers/media/platform/pxa_camera.c tx = dmaengine_prep_slave_sg(dma_chan, sg, sglen, DMA_DEV_TO_MEM, tx 868 drivers/media/platform/pxa_camera.c if (!tx) { tx 874 drivers/media/platform/pxa_camera.c tx->callback_param = pcdev; tx 877 drivers/media/platform/pxa_camera.c tx->callback = pxa_camera_dma_irq_y; tx 880 drivers/media/platform/pxa_camera.c tx->callback = pxa_camera_dma_irq_u; tx 883 drivers/media/platform/pxa_camera.c tx->callback = pxa_camera_dma_irq_v; tx 887 drivers/media/platform/pxa_camera.c buf->descs[channel] = tx; tx 892 drivers/media/platform/pxa_camera.c __func__, buf, tx); tx 93 drivers/media/platform/s5p-cec/s5p_cec.c cec->tx = STATE_NACK; tx 96 drivers/media/platform/s5p-cec/s5p_cec.c cec->tx = STATE_ERROR; tx 99 drivers/media/platform/s5p-cec/s5p_cec.c cec->tx = STATE_DONE; tx 132 drivers/media/platform/s5p-cec/s5p_cec.c switch (cec->tx) { tx 135 drivers/media/platform/s5p-cec/s5p_cec.c cec->tx = STATE_IDLE; tx 141 drivers/media/platform/s5p-cec/s5p_cec.c cec->tx = STATE_IDLE; tx 147 drivers/media/platform/s5p-cec/s5p_cec.c cec->tx = STATE_IDLE; tx 72 drivers/media/platform/s5p-cec/s5p_cec.h enum cec_state tx; tx 57 drivers/media/radio/radio-keene.c u8 tx; tx 109 drivers/media/radio/radio-keene.c radio->buffer[2] = radio->tx; tx 253 drivers/media/radio/radio-keene.c radio->tx = db2tx[(ctrl->val - (s32)ctrl->minimum) / (s32)ctrl->step]; tx 336 drivers/media/radio/radio-keene.c radio->tx = 0x32; tx 225 drivers/media/radio/wl128x/fmdrv_tx.c struct fmtx_data *tx = &fmdev->tx_data; tx 230 drivers/media/radio/wl128x/fmdrv_tx.c payload = tx->audio_io; tx 243 drivers/media/radio/wl128x/fmdrv_tx.c struct fmtx_data *tx = &fmdev->tx_data; tx 273 drivers/media/radio/wl128x/fmdrv_tx.c tx->xmit_state = new_xmit_state; tx 282 drivers/media/radio/wl128x/fmdrv_tx.c struct fmtx_data *tx = &fmdev->tx_data; tx 291 drivers/media/radio/wl128x/fmdrv_tx.c tx->pwr_lvl = new_pwr_lvl; tx 309 drivers/media/radio/wl128x/fmdrv_tx.c tx->pwr_lvl = new_pwr_lvl; tx 320 drivers/media/radio/wl128x/fmdrv_tx.c struct fmtx_data *tx = &fmdev->tx_data; tx 344 drivers/media/radio/wl128x/fmdrv_tx.c tx->preemph = payload; tx 372 drivers/media/radio/wl128x/fmdrv_tx.c struct fmtx_data *tx = &fmdev->tx_data; tx 388 drivers/media/radio/wl128x/fmdrv_tx.c tx->tx_frq = (unsigned long)freq_to_set; tx 389 drivers/media/radio/wl128x/fmdrv_tx.c fmdbg("tx: freq_to_set %ld\n", (long int)tx->tx_frq); tx 400 drivers/media/radio/wl128x/fmdrv_tx.c fm_tx_set_pwr_lvl(fmdev, tx->pwr_lvl); tx 401 drivers/media/radio/wl128x/fmdrv_tx.c fm_tx_set_preemph_filter(fmdev, tx->preemph); tx 403 drivers/media/radio/wl128x/fmdrv_tx.c tx->audio_io = 0x01; /* I2S */ tx 408 drivers/media/radio/wl128x/fmdrv_tx.c tx->aud_mode = FM_STEREO_MODE; tx 409 drivers/media/radio/wl128x/fmdrv_tx.c tx->rds.flag = FM_RDS_DISABLE; tx 62 drivers/media/rc/fintek-cir.h } tx; tx 121 drivers/media/rc/imon.c } tx; tx 596 drivers/media/rc/imon.c reinit_completion(&ictx->tx.finished); tx 597 drivers/media/rc/imon.c ictx->tx.busy = true; tx 602 drivers/media/rc/imon.c ictx->tx.busy = false; tx 609 drivers/media/rc/imon.c &ictx->tx.finished); tx 616 drivers/media/rc/imon.c retval = ictx->tx.status; tx 936 drivers/media/rc/imon.c if (copy_from_user(ictx->tx.data_buf, buf, n_bytes)) { tx 943 drivers/media/rc/imon.c ictx->tx.data_buf[i] = ' '; tx 946 drivers/media/rc/imon.c ictx->tx.data_buf[i] = 0xFF; tx 952 drivers/media/rc/imon.c memcpy(ictx->usb_tx_buf, ictx->tx.data_buf + offset, 7); tx 1050 drivers/media/rc/imon.c ictx->tx.status = urb->status; tx 1053 drivers/media/rc/imon.c ictx->tx.busy = false; tx 1055 drivers/media/rc/imon.c complete(&ictx->tx.finished); tx 2200 drivers/media/rc/imon.c init_completion(&ictx->tx.finished); tx 2478 drivers/media/rc/imon.c if (ictx->tx.busy) { tx 2480 drivers/media/rc/imon.c complete(&ictx->tx.finished); tx 65 drivers/media/spi/cxd2880-spi.c struct spi_transfer tx = {}; tx 72 drivers/media/spi/cxd2880-spi.c tx.tx_buf = data; tx 73 drivers/media/spi/cxd2880-spi.c tx.len = size; tx 76 drivers/media/spi/cxd2880-spi.c spi_message_add_tail(&tx, &msg); tx 146 drivers/media/spi/gs1662.c struct spi_transfer tx[] = { tx 159 drivers/media/spi/gs1662.c spi_message_add_tail(&tx[0], &msg); tx 160 drivers/media/spi/gs1662.c spi_message_add_tail(&tx[1], &msg); tx 174 drivers/media/spi/gs1662.c struct spi_transfer tx[] = { tx 187 drivers/media/spi/gs1662.c spi_message_add_tail(&tx[0], &msg); tx 188 drivers/media/spi/gs1662.c spi_message_add_tail(&tx[1], &msg); tx 59 drivers/media/usb/dvb-usb/dib0700.h extern int dib0700_ctrl_rd(struct dvb_usb_device *d, u8 *tx, u8 txlen, u8 *rx, u8 rxlen); tx 53 drivers/media/usb/dvb-usb/dib0700_core.c static int dib0700_ctrl_wr(struct dvb_usb_device *d, u8 *tx, u8 txlen) tx 58 drivers/media/usb/dvb-usb/dib0700_core.c debug_dump(tx, txlen, deb_data); tx 61 drivers/media/usb/dvb-usb/dib0700_core.c tx[0], USB_TYPE_VENDOR | USB_DIR_OUT, 0, 0, tx, txlen, tx 71 drivers/media/usb/dvb-usb/dib0700_core.c int dib0700_ctrl_rd(struct dvb_usb_device *d, u8 *tx, u8 txlen, u8 *rx, u8 rxlen) tx 86 drivers/media/usb/dvb-usb/dib0700_core.c debug_dump(tx,txlen,deb_data); tx 88 drivers/media/usb/dvb-usb/dib0700_core.c value = ((txlen - 2) << 8) | tx[1]; tx 91 drivers/media/usb/dvb-usb/dib0700_core.c index |= (tx[2] << 8); tx 93 drivers/media/usb/dvb-usb/dib0700_core.c index |= tx[3]; tx 95 drivers/media/usb/dvb-usb/dib0700_core.c status = usb_control_msg(d->udev, usb_rcvctrlpipe(d->udev,0), tx[0], tx 85 drivers/media/usb/dvb-usb/technisat-usb2.c u8 device_addr, u8 *tx, u8 txlen, u8 *rx, u8 rxlen) tx 95 drivers/media/usb/dvb-usb/technisat-usb2.c debug_dump(tx, txlen, deb_i2c); tx 117 drivers/media/usb/dvb-usb/technisat-usb2.c memcpy(&b[2], tx, txlen); tx 242 drivers/media/usb/hackrf/hackrf.c const bool tx = test_bit(TX_ON, &dev->flags); tx 264 drivers/media/usb/hackrf/hackrf.c if (!rx && !tx) { tx 275 drivers/media/usb/hackrf/hackrf.c } else if (tx && test_and_clear_bit(TX_DAC_FREQUENCY, &dev->flags)) { tx 314 drivers/media/usb/hackrf/hackrf.c } else if (tx && test_and_clear_bit(TX_BANDWIDTH, &dev->flags)) { tx 351 drivers/media/usb/hackrf/hackrf.c } else if (tx && test_and_clear_bit(TX_RF_FREQUENCY, &dev->flags)) { tx 386 drivers/media/usb/hackrf/hackrf.c if (tx && test_and_clear_bit(TX_RF_GAIN, &dev->flags)) { tx 420 drivers/media/usb/hackrf/hackrf.c if (tx && test_and_clear_bit(TX_LNA_GAIN, &dev->flags)) { tx 33 drivers/mfd/ipaq-micro.c struct ipaq_micro_txdev *tx = µ->tx; tx 40 drivers/mfd/ipaq-micro.c tx->buf[bp++] = CHAR_SOF; tx 43 drivers/mfd/ipaq-micro.c tx->buf[bp++] = checksum; tx 46 drivers/mfd/ipaq-micro.c tx->buf[bp++] = msg->tx_data[i]; tx 50 drivers/mfd/ipaq-micro.c tx->buf[bp++] = checksum; tx 51 drivers/mfd/ipaq-micro.c tx->len = bp; tx 52 drivers/mfd/ipaq-micro.c tx->index = 0; tx 283 drivers/mfd/ipaq-micro.c struct ipaq_micro_txdev *tx = µ->tx; tx 286 drivers/mfd/ipaq-micro.c while ((tx->index < tx->len) && tx 288 drivers/mfd/ipaq-micro.c writel(tx->buf[tx->index], micro->base + UTDR); tx 289 drivers/mfd/ipaq-micro.c tx->index++; tx 335 drivers/mfd/ipaq-micro.c struct ipaq_micro_txdev *tx = µ->tx; tx 357 drivers/mfd/ipaq-micro.c } while (((tx->index < tx->len) && (status & UTSR0_TFS)) || tx 262 drivers/misc/echo/echo.c int16_t oslec_update(struct oslec_state *ec, int16_t tx, int16_t rx) tx 275 drivers/misc/echo/echo.c ec->tx = tx; tx 277 drivers/misc/echo/echo.c tx >>= 1; tx 333 drivers/misc/echo/echo.c new = (int)tx * (int)tx; tx 344 drivers/misc/echo/echo.c ec->ltxacc += abs(tx) - ec->ltx; tx 352 drivers/misc/echo/echo.c echo_value = fir16(&ec->fir_state, tx); tx 359 drivers/misc/echo/echo.c echo_value = fir16(&ec->fir_state_bg, tx); tx 554 drivers/misc/echo/echo.c int16_t oslec_hpf_tx(struct oslec_state *ec, int16_t tx) tx 560 drivers/misc/echo/echo.c tmp = tx << 15; tx 578 drivers/misc/echo/echo.c tx = tmp1; tx 582 drivers/misc/echo/echo.c return tx; tx 118 drivers/misc/echo/echo.h int16_t tx; tx 70 drivers/misc/echo/oslec.h int16_t oslec_update(struct oslec_state *ec, int16_t tx, int16_t rx); tx 79 drivers/misc/echo/oslec.h int16_t oslec_hpf_tx(struct oslec_state *ec, int16_t tx); tx 413 drivers/misc/mic/scif/scif_dma.c struct dma_async_tx_descriptor *tx = NULL; tx 426 drivers/misc/mic/scif/scif_dma.c tx = ddev->device_prep_dma_memcpy(chan, 0, 0, 0, flags); tx 427 drivers/misc/mic/scif/scif_dma.c if (!tx) { tx 433 drivers/misc/mic/scif/scif_dma.c cookie = tx->tx_submit(tx); tx 479 drivers/misc/mic/scif/scif_dma.c struct dma_async_tx_descriptor *tx = NULL; tx 493 drivers/misc/mic/scif/scif_dma.c tx = ddev->device_prep_dma_memcpy(chan, 0, 0, 0, flags); tx 494 drivers/misc/mic/scif/scif_dma.c if (!tx) { tx 501 drivers/misc/mic/scif/scif_dma.c tx->callback = scif_dma_callback; tx 502 drivers/misc/mic/scif/scif_dma.c tx->callback_param = &done_wait; tx 503 drivers/misc/mic/scif/scif_dma.c cookie = tx->tx_submit(tx); tx 880 drivers/misc/mic/scif/scif_dma.c struct dma_async_tx_descriptor *tx; tx 954 drivers/misc/mic/scif/scif_dma.c tx = tx 961 drivers/misc/mic/scif/scif_dma.c if (!tx) { tx 965 drivers/misc/mic/scif/scif_dma.c cookie = tx->tx_submit(tx); tx 978 drivers/misc/mic/scif/scif_dma.c tx = tx 983 drivers/misc/mic/scif/scif_dma.c if (!tx) { tx 987 drivers/misc/mic/scif/scif_dma.c cookie = tx->tx_submit(tx); tx 994 drivers/misc/mic/scif/scif_dma.c tx = tx 999 drivers/misc/mic/scif/scif_dma.c if (!tx) { tx 1003 drivers/misc/mic/scif/scif_dma.c cookie = tx->tx_submit(tx); tx 1011 drivers/misc/mic/scif/scif_dma.c tx = dev->device_prep_dma_memcpy(chan, temp_dma_addr, tx 1013 drivers/misc/mic/scif/scif_dma.c if (!tx) { tx 1017 drivers/misc/mic/scif/scif_dma.c cookie = tx->tx_submit(tx); tx 1061 drivers/misc/mic/scif/scif_dma.c tx = dev->device_prep_dma_memcpy(chan, 0, 0, 0, DMA_PREP_INTERRUPT); tx 1062 drivers/misc/mic/scif/scif_dma.c if (!tx) { tx 1066 drivers/misc/mic/scif/scif_dma.c tx->callback = &scif_rma_completion_cb; tx 1067 drivers/misc/mic/scif/scif_dma.c tx->callback_param = comp_cb; tx 1068 drivers/misc/mic/scif/scif_dma.c cookie = tx->tx_submit(tx); tx 1101 drivers/misc/mic/scif/scif_dma.c struct dma_async_tx_descriptor *tx; tx 1143 drivers/misc/mic/scif/scif_dma.c tx = dev->device_prep_dma_memcpy(chan, dst_dma_addr, tx 1147 drivers/misc/mic/scif/scif_dma.c if (!tx) { tx 1151 drivers/misc/mic/scif/scif_dma.c cookie = tx->tx_submit(tx); tx 1164 drivers/misc/mic/scif/scif_dma.c tx = dev->device_prep_dma_memcpy(chan, dst_dma_addr, tx 1166 drivers/misc/mic/scif/scif_dma.c if (!tx) { tx 1170 drivers/misc/mic/scif/scif_dma.c cookie = tx->tx_submit(tx); tx 1177 drivers/misc/mic/scif/scif_dma.c tx = dev->device_prep_dma_memcpy(chan, dst_dma_addr, tx 1179 drivers/misc/mic/scif/scif_dma.c if (!tx) { tx 1183 drivers/misc/mic/scif/scif_dma.c cookie = tx->tx_submit(tx); tx 1221 drivers/misc/mic/scif/scif_dma.c struct dma_async_tx_descriptor *tx; tx 1314 drivers/misc/mic/scif/scif_dma.c tx = dev->device_prep_dma_memcpy(chan, dst_dma_addr, tx 1319 drivers/misc/mic/scif/scif_dma.c if (!tx) { tx 1323 drivers/misc/mic/scif/scif_dma.c cookie = tx->tx_submit(tx); tx 1337 drivers/misc/mic/scif/scif_dma.c tx = dev->device_prep_dma_memcpy(chan, dst_dma_addr, tx 1340 drivers/misc/mic/scif/scif_dma.c if (!tx) { tx 1344 drivers/misc/mic/scif/scif_dma.c cookie = tx->tx_submit(tx); tx 1351 drivers/misc/mic/scif/scif_dma.c tx = dev->device_prep_dma_memcpy(chan, dst_dma_addr, tx 1354 drivers/misc/mic/scif/scif_dma.c if (!tx) { tx 1358 drivers/misc/mic/scif/scif_dma.c cookie = tx->tx_submit(tx); tx 202 drivers/misc/mic/scif/scif_fence.c struct dma_async_tx_descriptor *tx; tx 209 drivers/misc/mic/scif/scif_fence.c tx = ddev->device_prep_dma_memcpy(chan, 0, 0, 0, DMA_PREP_FENCE); tx 210 drivers/misc/mic/scif/scif_fence.c if (!tx) { tx 216 drivers/misc/mic/scif/scif_fence.c cookie = tx->tx_submit(tx); tx 229 drivers/misc/mic/scif/scif_fence.c tx = ddev->device_prep_dma_imm_data(chan, dst, val, 0); tx 243 drivers/misc/mic/scif/scif_fence.c tx = ddev->device_prep_dma_memcpy(chan, dst, src, sizeof(val), tx 246 drivers/misc/mic/scif/scif_fence.c if (!tx) { tx 261 drivers/misc/mic/scif/scif_fence.c tx->callback = scif_prog_signal_cb; tx 262 drivers/misc/mic/scif/scif_fence.c tx->callback_param = cb_arg; tx 264 drivers/misc/mic/scif/scif_fence.c cookie = tx->tx_submit(tx); tx 591 drivers/misc/mic/scif/scif_fence.c struct dma_async_tx_descriptor *tx; tx 595 drivers/misc/mic/scif/scif_fence.c tx = ddev->device_prep_dma_memcpy(chan, 0, 0, 0, DMA_PREP_FENCE); tx 596 drivers/misc/mic/scif/scif_fence.c if (!tx) { tx 602 drivers/misc/mic/scif/scif_fence.c cookie = tx->tx_submit(tx); tx 610 drivers/misc/mic/scif/scif_fence.c tx = ddev->device_prep_dma_interrupt(chan, DMA_PREP_INTERRUPT); tx 611 drivers/misc/mic/scif/scif_fence.c if (!tx) { tx 617 drivers/misc/mic/scif/scif_fence.c tx->callback = scif_fence_mark_cb; tx 618 drivers/misc/mic/scif/scif_fence.c tx->callback_param = ep; tx 619 drivers/misc/mic/scif/scif_fence.c *mark = cookie = tx->tx_submit(tx); tx 474 drivers/misc/mic/vop/vop_vringh.c struct dma_async_tx_descriptor *tx; tx 483 drivers/misc/mic/vop/vop_vringh.c tx = ddev->device_prep_dma_memcpy(vop_ch, dst, src, len, tx 485 drivers/misc/mic/vop/vop_vringh.c if (!tx) { tx 491 drivers/misc/mic/vop/vop_vringh.c cookie = tx->tx_submit(tx); tx 56 drivers/mmc/core/sdio_uart.c __u32 tx; tx 434 drivers/mmc/core/sdio_uart.c port->icount.tx++; tx 451 drivers/mmc/core/sdio_uart.c port->icount.tx++; tx 974 drivers/mmc/core/sdio_uart.c port->icount.tx, port->icount.rx); tx 983 drivers/mmc/host/omap.c struct dma_async_tx_descriptor *tx; tx 1039 drivers/mmc/host/omap.c tx = dmaengine_prep_slave_sg(c, data->sg, host->sg_len, tx 1042 drivers/mmc/host/omap.c if (!tx) tx 1047 drivers/mmc/host/omap.c tx->callback = mmc_omap_dma_callback; tx 1048 drivers/mmc/host/omap.c tx->callback_param = host; tx 1049 drivers/mmc/host/omap.c dmaengine_submit(tx); tx 1242 drivers/mmc/host/omap_hsmmc.c struct dma_async_tx_descriptor *tx; tx 1281 drivers/mmc/host/omap_hsmmc.c tx = dmaengine_prep_slave_sg(chan, data->sg, data->sg_len, tx 1284 drivers/mmc/host/omap_hsmmc.c if (!tx) { tx 1290 drivers/mmc/host/omap_hsmmc.c tx->callback = omap_hsmmc_dma_callback; tx 1291 drivers/mmc/host/omap_hsmmc.c tx->callback_param = host; tx 1294 drivers/mmc/host/omap_hsmmc.c dmaengine_submit(tx); tx 161 drivers/mmc/host/pxamci.c struct dma_async_tx_descriptor *tx; tx 209 drivers/mmc/host/pxamci.c tx = dmaengine_prep_slave_sg(chan, data->sg, host->dma_len, direction, tx 211 drivers/mmc/host/pxamci.c if (!tx) { tx 217 drivers/mmc/host/pxamci.c tx->callback = pxamci_dma_irq; tx 218 drivers/mmc/host/pxamci.c tx->callback_param = host; tx 221 drivers/mmc/host/pxamci.c host->dma_cookie = dmaengine_submit(tx); tx 579 drivers/mmc/host/rtsx_usb_sdmmc.c static int sd_change_phase(struct rtsx_usb_sdmmc *host, u8 sample_point, int tx) tx 585 drivers/mmc/host/rtsx_usb_sdmmc.c __func__, tx ? "TX" : "RX", sample_point); tx 591 drivers/mmc/host/rtsx_usb_sdmmc.c if (tx) tx 328 drivers/mtd/nand/onenand/omap2.c struct dma_async_tx_descriptor *tx; tx 331 drivers/mtd/nand/onenand/omap2.c tx = dmaengine_prep_dma_memcpy(c->dma_chan, dst, src, count, tx 333 drivers/mtd/nand/onenand/omap2.c if (!tx) { tx 340 drivers/mtd/nand/onenand/omap2.c tx->callback = omap2_onenand_dma_complete_func; tx 341 drivers/mtd/nand/onenand/omap2.c tx->callback_param = &c->dma_done; tx 343 drivers/mtd/nand/onenand/omap2.c cookie = tx->tx_submit(tx); tx 369 drivers/mtd/nand/raw/atmel/nand-controller.c struct dma_async_tx_descriptor *tx; tx 387 drivers/mtd/nand/raw/atmel/nand-controller.c tx = dmaengine_prep_dma_memcpy(nc->dmac, dst_dma, src_dma, len, tx 389 drivers/mtd/nand/raw/atmel/nand-controller.c if (!tx) { tx 394 drivers/mtd/nand/raw/atmel/nand-controller.c tx->callback = atmel_nand_dma_transfer_finished; tx 395 drivers/mtd/nand/raw/atmel/nand-controller.c tx->callback_param = &finished; tx 397 drivers/mtd/nand/raw/atmel/nand-controller.c cookie = dmaengine_submit(tx); tx 462 drivers/mtd/nand/raw/fsmc_nand.c struct dma_async_tx_descriptor *tx; tx 487 drivers/mtd/nand/raw/fsmc_nand.c tx = dma_dev->device_prep_dma_memcpy(chan, dma_dst, dma_src, tx 489 drivers/mtd/nand/raw/fsmc_nand.c if (!tx) { tx 495 drivers/mtd/nand/raw/fsmc_nand.c tx->callback = dma_complete; tx 496 drivers/mtd/nand/raw/fsmc_nand.c tx->callback_param = host; tx 497 drivers/mtd/nand/raw/fsmc_nand.c cookie = tx->tx_submit(tx); tx 823 drivers/mtd/nand/raw/marvell_nand.c struct dma_async_tx_descriptor *tx; tx 832 drivers/mtd/nand/raw/marvell_nand.c tx = dmaengine_prep_slave_sg(nfc->dma_chan, &sg, 1, tx 836 drivers/mtd/nand/raw/marvell_nand.c if (!tx) { tx 842 drivers/mtd/nand/raw/marvell_nand.c cookie = dmaengine_submit(tx); tx 1331 drivers/mtd/nand/raw/nandsim.c ssize_t tx; tx 1339 drivers/mtd/nand/raw/nandsim.c tx = kernel_read(file, buf, count, &pos); tx 1342 drivers/mtd/nand/raw/nandsim.c return tx; tx 1347 drivers/mtd/nand/raw/nandsim.c ssize_t tx; tx 1355 drivers/mtd/nand/raw/nandsim.c tx = kernel_write(file, buf, count, &pos); tx 1358 drivers/mtd/nand/raw/nandsim.c return tx; tx 1419 drivers/mtd/nand/raw/nandsim.c ssize_t tx; tx 1426 drivers/mtd/nand/raw/nandsim.c tx = read_file(ns, ns->cfile, ns->buf.byte, num, pos); tx 1427 drivers/mtd/nand/raw/nandsim.c if (tx != num) { tx 1428 drivers/mtd/nand/raw/nandsim.c NS_ERR("read_page: read error for page %d ret %ld\n", ns->regs.row, (long)tx); tx 1489 drivers/mtd/nand/raw/nandsim.c ssize_t tx; tx 1500 drivers/mtd/nand/raw/nandsim.c tx = read_file(ns, ns->cfile, pg_off, num, off); tx 1501 drivers/mtd/nand/raw/nandsim.c if (tx != num) { tx 1502 drivers/mtd/nand/raw/nandsim.c NS_ERR("prog_page: read error for page %d ret %ld\n", ns->regs.row, (long)tx); tx 1510 drivers/mtd/nand/raw/nandsim.c tx = write_file(ns, ns->cfile, ns->file_buf, ns->geom.pgszoob, pos); tx 1511 drivers/mtd/nand/raw/nandsim.c if (tx != ns->geom.pgszoob) { tx 1512 drivers/mtd/nand/raw/nandsim.c NS_ERR("prog_page: write error for page %d ret %ld\n", ns->regs.row, (long)tx); tx 1517 drivers/mtd/nand/raw/nandsim.c tx = write_file(ns, ns->cfile, pg_off, num, off); tx 1518 drivers/mtd/nand/raw/nandsim.c if (tx != num) { tx 1519 drivers/mtd/nand/raw/nandsim.c NS_ERR("prog_page: write error for page %d ret %ld\n", ns->regs.row, (long)tx); tx 458 drivers/mtd/nand/raw/omap2.c struct dma_async_tx_descriptor *tx; tx 478 drivers/mtd/nand/raw/omap2.c tx = dmaengine_prep_slave_sg(info->dma, &sg, n, tx 481 drivers/mtd/nand/raw/omap2.c if (!tx) tx 484 drivers/mtd/nand/raw/omap2.c tx->callback = omap_nand_dma_callback; tx 485 drivers/mtd/nand/raw/omap2.c tx->callback_param = &info->comp; tx 486 drivers/mtd/nand/raw/omap2.c dmaengine_submit(tx); tx 971 drivers/mtd/spi-nor/cadence-quadspi.c struct dma_async_tx_descriptor *tx; tx 985 drivers/mtd/spi-nor/cadence-quadspi.c tx = dmaengine_prep_dma_memcpy(cqspi->rx_chan, dma_dst, dma_src, tx 987 drivers/mtd/spi-nor/cadence-quadspi.c if (!tx) { tx 993 drivers/mtd/spi-nor/cadence-quadspi.c tx->callback = cqspi_rx_dma_callback; tx 994 drivers/mtd/spi-nor/cadence-quadspi.c tx->callback_param = cqspi; tx 995 drivers/mtd/spi-nor/cadence-quadspi.c cookie = tx->tx_submit(tx); tx 154 drivers/mtd/spi-nor/mtk-quadspi.c u8 *tx, int txlen, u8 *rx, int rxlen) tx 173 drivers/mtd/spi-nor/mtk-quadspi.c writeb(tx[i], mtk_nor->base + MTK_NOR_PRG_REG(idx)); tx 223 drivers/net/can/grcan.c struct grcan_dma_buffer tx; tx 527 drivers/net/can/grcan.c dma->tx.size); tx 558 drivers/net/can/grcan.c txrd = grcan_ring_add(txrd, GRCAN_MSG_SIZE, dma->tx.size); tx 695 drivers/net/can/grcan.c if (grcan_txspace(dma->tx.size, txwr, tx 863 drivers/net/can/grcan.c if (grcan_txspace(priv->dma.tx.size, txwr, priv->eskbp) && tx 937 drivers/net/can/grcan.c struct grcan_dma_buffer *large = rsize > tsize ? &dma->rx : &dma->tx; tx 938 drivers/net/can/grcan.c struct grcan_dma_buffer *small = rsize > tsize ? &dma->tx : &dma->rx; tx 960 drivers/net/can/grcan.c dma->tx.size = tsize; tx 982 drivers/net/can/grcan.c grcan_write_reg(®s->txaddr, priv->dma.tx.handle); tx 983 drivers/net/can/grcan.c grcan_write_reg(®s->txsize, priv->dma.tx.size); tx 1056 drivers/net/can/grcan.c priv->echo_skb = kcalloc(dma->tx.size, sizeof(*priv->echo_skb), tx 1062 drivers/net/can/grcan.c priv->can.echo_skb_max = dma->tx.size; tx 1065 drivers/net/can/grcan.c priv->txdlc = kcalloc(dma->tx.size, sizeof(*priv->txdlc), GFP_KERNEL); tx 1307 drivers/net/can/grcan.c if (grcan_txspace(dma->tx.size, txwr, priv->eskbp)) tx 1379 drivers/net/can/grcan.c space = grcan_txspace(dma->tx.size, txwr, priv->eskbp); tx 1382 drivers/net/can/grcan.c slot = dma->tx.buf + txwr; tx 1434 drivers/net/can/grcan.c if (unlikely(grcan_ring_sub(txwr, txrd, dma->tx.size) == 1)) { tx 1460 drivers/net/can/grcan.c grcan_ring_add(txwr, GRCAN_MSG_SIZE, dma->tx.size)); tx 335 drivers/net/can/kvaser_pciefd.c static int kvaser_pciefd_spi_cmd(struct kvaser_pciefd *pcie, const u8 *tx, tx 349 drivers/net/can/kvaser_pciefd.c iowrite32(*tx++, pcie->reg_base + KVASER_PCIEFD_SPI_TX_REG); tx 234 drivers/net/can/mscan/mscan.c out_be16(®s->tx.idr3_2, can_id); tx 246 drivers/net/can/mscan/mscan.c out_be16(®s->tx.idr1_0, can_id); tx 249 drivers/net/can/mscan/mscan.c void __iomem *data = ®s->tx.dsr1_0; tx 261 drivers/net/can/mscan/mscan.c out_8(®s->tx.dlr, frame->can_dlc); tx 262 drivers/net/can/mscan/mscan.c out_8(®s->tx.tbpr, priv->cur_pri); tx 448 drivers/net/can/mscan/mscan.c stats->tx_bytes += in_8(®s->tx.dlr); tx 217 drivers/net/can/mscan/mscan.h } tx; tx 29 drivers/net/can/softing/softing.h } tx; tx 64 drivers/net/can/softing/softing.h } tx; tx 334 drivers/net/can/softing/softing_fw.c card->tx.pending = 0; tx 455 drivers/net/can/softing/softing_fw.c priv->tx.pending = 0; tx 456 drivers/net/can/softing/softing_fw.c priv->tx.echo_put = 0; tx 457 drivers/net/can/softing/softing_fw.c priv->tx.echo_get = 0; tx 470 drivers/net/can/softing/softing_fw.c card->tx.pending = 0; tx 69 drivers/net/can/softing/softing_main.c (card->tx.pending >= TXMAX) || tx 70 drivers/net/can/softing/softing_main.c (priv->tx.pending >= TX_ECHO_SKB_MAX)) tx 104 drivers/net/can/softing/softing_main.c card->tx.last_bus = priv->index; tx 105 drivers/net/can/softing/softing_main.c ++card->tx.pending; tx 106 drivers/net/can/softing/softing_main.c ++priv->tx.pending; tx 107 drivers/net/can/softing/softing_main.c can_put_echo_skb(skb, dev, priv->tx.echo_put); tx 108 drivers/net/can/softing/softing_main.c ++priv->tx.echo_put; tx 109 drivers/net/can/softing/softing_main.c if (priv->tx.echo_put >= TX_ECHO_SKB_MAX) tx 110 drivers/net/can/softing/softing_main.c priv->tx.echo_put = 0; tx 115 drivers/net/can/softing/softing_main.c if (card->tx.pending >= TXMAX) { tx 284 drivers/net/can/softing/softing_main.c skb = priv->can.echo_skb[priv->tx.echo_get]; tx 287 drivers/net/can/softing/softing_main.c can_get_echo_skb(netdev, priv->tx.echo_get); tx 288 drivers/net/can/softing/softing_main.c ++priv->tx.echo_get; tx 289 drivers/net/can/softing/softing_main.c if (priv->tx.echo_get >= TX_ECHO_SKB_MAX) tx 290 drivers/net/can/softing/softing_main.c priv->tx.echo_get = 0; tx 291 drivers/net/can/softing/softing_main.c if (priv->tx.pending) tx 292 drivers/net/can/softing/softing_main.c --priv->tx.pending; tx 293 drivers/net/can/softing/softing_main.c if (card->tx.pending) tx 294 drivers/net/can/softing/softing_main.c --card->tx.pending; tx 333 drivers/net/can/softing/softing_main.c offset = card->tx.last_bus; tx 335 drivers/net/can/softing/softing_main.c if (card->tx.pending >= TXMAX) tx 344 drivers/net/can/softing/softing_main.c if (priv->tx.pending >= TX_ECHO_SKB_MAX) tx 186 drivers/net/can/usb/ems_usb.c u8 tx; tx 167 drivers/net/can/usb/esd_usb2.c struct tx_msg tx; tx 739 drivers/net/can/usb/esd_usb2.c msg->msg.tx.net = priv->index; tx 740 drivers/net/can/usb/esd_usb2.c msg->msg.tx.dlc = cf->can_dlc; tx 741 drivers/net/can/usb/esd_usb2.c msg->msg.tx.id = cpu_to_le32(cf->can_id & CAN_ERR_MASK); tx 744 drivers/net/can/usb/esd_usb2.c msg->msg.tx.dlc |= ESD_RTR; tx 747 drivers/net/can/usb/esd_usb2.c msg->msg.tx.id |= cpu_to_le32(ESD_EXTID); tx 750 drivers/net/can/usb/esd_usb2.c msg->msg.tx.data[i] = cf->data[i]; tx 775 drivers/net/can/usb/esd_usb2.c msg->msg.tx.hnd = 0x80000000 | i; /* returned in TX done message */ tx 18 drivers/net/dsa/sja1105/sja1105_spi.c const void *tx, void *rx, int size) tx 22 drivers/net/dsa/sja1105/sja1105_spi.c .tx_buf = tx, tx 910 drivers/net/dsa/vitesse-vsc73xx-core.c bool tx) tx 916 drivers/net/dsa/vitesse-vsc73xx-core.c if (tx) { tx 954 drivers/net/ethernet/3com/3c574_cs.c u8 rx, tx, up; tx 976 drivers/net/ethernet/3com/3c574_cs.c tx = inw(ioaddr + 12); tx 1511 drivers/net/ethernet/3com/typhoon.c struct tx_desc *tx; tx 1517 drivers/net/ethernet/3com/typhoon.c tx = (struct tx_desc *) (txRing->ringBase + lastRead); tx 1518 drivers/net/ethernet/3com/typhoon.c type = tx->flags & TYPHOON_TYPE_MASK; tx 1523 drivers/net/ethernet/3com/typhoon.c unsigned long ptr = tx->tx_addr; tx 1529 drivers/net/ethernet/3com/typhoon.c skb_dma = (dma_addr_t) le32_to_cpu(tx->frag.addr); tx 1530 drivers/net/ethernet/3com/typhoon.c dma_len = le16_to_cpu(tx->len); tx 1535 drivers/net/ethernet/3com/typhoon.c tx->flags = 0; tx 2332 drivers/net/ethernet/alteon/acenic.c struct tx_desc __iomem *tx; tx 2333 drivers/net/ethernet/alteon/acenic.c tx = (__force struct tx_desc __iomem *) &ap->tx_ring[i]; tx 2334 drivers/net/ethernet/alteon/acenic.c writel(0, &tx->addr.addrhi); tx 2335 drivers/net/ethernet/alteon/acenic.c writel(0, &tx->addr.addrlo); tx 2336 drivers/net/ethernet/alteon/acenic.c writel(0, &tx->flagsize); tx 654 drivers/net/ethernet/amazon/ena/ena_admin_defs.h u32 tx; tx 56 drivers/net/ethernet/amazon/ena/ena_ethtool.c ENA_STAT_ENTRY(stat, tx) tx 3238 drivers/net/ethernet/amazon/ena/ena_netdev.c if (feat->offload.tx & tx 3242 drivers/net/ethernet/amazon/ena/ena_netdev.c if (feat->offload.tx & tx 3246 drivers/net/ethernet/amazon/ena/ena_netdev.c if (feat->offload.tx & ENA_ADMIN_FEATURE_OFFLOAD_DESC_TSO_IPV4_MASK) tx 3249 drivers/net/ethernet/amazon/ena/ena_netdev.c if (feat->offload.tx & ENA_ADMIN_FEATURE_OFFLOAD_DESC_TSO_IPV6_MASK) tx 3252 drivers/net/ethernet/amazon/ena/ena_netdev.c if (feat->offload.tx & ENA_ADMIN_FEATURE_OFFLOAD_DESC_TSO_ECN_MASK) tx 262 drivers/net/ethernet/amd/sunlance.c void (*tx)(struct net_device *); tx 839 drivers/net/ethernet/amd/sunlance.c lp->tx(dev); tx 1370 drivers/net/ethernet/amd/sunlance.c lp->tx = lance_tx_pio; tx 1382 drivers/net/ethernet/amd/sunlance.c lp->tx = lance_tx_dvma; tx 419 drivers/net/ethernet/amd/xgbe/xgbe-desc.c memset(&ring->tx, 0, sizeof(ring->tx)); tx 510 drivers/net/ethernet/amd/xgbe/xgbe-desc.c memset(&rdata->tx, 0, sizeof(rdata->tx)); tx 551 drivers/net/ethernet/amd/xgbe/xgbe-desc.c if ((tso && (packet->mss != ring->tx.cur_mss)) || tx 552 drivers/net/ethernet/amd/xgbe/xgbe-desc.c (vlan && (packet->vlan_ctag != ring->tx.cur_vlan_ctag))) tx 1663 drivers/net/ethernet/amd/xgbe/xgbe-dev.c ring->tx.xmit_more = 0; tx 1695 drivers/net/ethernet/amd/xgbe/xgbe-dev.c if (tso && (packet->mss != ring->tx.cur_mss)) tx 1700 drivers/net/ethernet/amd/xgbe/xgbe-dev.c if (vlan && (packet->vlan_ctag != ring->tx.cur_vlan_ctag)) tx 1747 drivers/net/ethernet/amd/xgbe/xgbe-dev.c ring->tx.cur_mss = packet->mss; tx 1767 drivers/net/ethernet/amd/xgbe/xgbe-dev.c ring->tx.cur_vlan_ctag = packet->vlan_ctag; tx 1865 drivers/net/ethernet/amd/xgbe/xgbe-dev.c rdata->tx.packets = tx_packets; tx 1866 drivers/net/ethernet/amd/xgbe/xgbe-dev.c rdata->tx.bytes = tx_bytes; tx 1895 drivers/net/ethernet/amd/xgbe/xgbe-dev.c ring->tx.xmit_more = 1; tx 282 drivers/net/ethernet/amd/xgbe/xgbe-drv.c ring->tx.queue_stopped = 1; tx 287 drivers/net/ethernet/amd/xgbe/xgbe-drv.c if (ring->tx.xmit_more) tx 1851 drivers/net/ethernet/amd/xgbe/xgbe-drv.c if (skb_shinfo(skb)->gso_size != ring->tx.cur_mss) { tx 1873 drivers/net/ethernet/amd/xgbe/xgbe-drv.c if (skb_vlan_tag_get(skb) != ring->tx.cur_vlan_ctag) tx 2655 drivers/net/ethernet/amd/xgbe/xgbe-drv.c tx_packets += rdata->tx.packets; tx 2656 drivers/net/ethernet/amd/xgbe/xgbe-drv.c tx_bytes += rdata->tx.bytes; tx 2672 drivers/net/ethernet/amd/xgbe/xgbe-drv.c if ((ring->tx.queue_stopped == 1) && tx 2674 drivers/net/ethernet/amd/xgbe/xgbe-drv.c ring->tx.queue_stopped = 0; tx 660 drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c unsigned int rx, tx; tx 689 drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c tx = __rounddown_pow_of_two(ringparam->tx_pending); tx 690 drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c if (tx != ringparam->tx_pending) tx 693 drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c tx); tx 696 drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c (tx == pdata->tx_desc_count)) tx 700 drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c pdata->tx_desc_count = tx; tx 712 drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c unsigned int rx, tx, combined; tx 722 drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c tx = min(pdata->hw_feat.tx_ch_cnt, pdata->tx_max_channel_count); tx 723 drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c tx = min(tx, pdata->channel_irq_count); tx 724 drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c tx = min(tx, pdata->tx_max_q_count); tx 726 drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c combined = min(rx, tx); tx 730 drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c channels->max_tx = tx ? tx - 1 : 0; tx 734 drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c tx = pdata->new_tx_ring_count ? : pdata->tx_ring_count; tx 736 drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c combined = min(rx, tx); tx 738 drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c tx -= combined; tx 742 drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c channels->tx_count = tx; tx 757 drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c unsigned int rx, rx_curr, tx, tx_curr, combined; tx 767 drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c tx = min(pdata->hw_feat.tx_ch_cnt, pdata->tx_max_channel_count); tx 768 drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c tx = min(tx, pdata->tx_max_q_count); tx 769 drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c tx = min(tx, pdata->channel_irq_count); tx 771 drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c combined = min(rx, tx); tx 814 drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c if ((channels->combined_count + channels->tx_count) > tx) { tx 817 drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c channels->combined_count + channels->tx_count, tx); tx 823 drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c tx = channels->combined_count + channels->tx_count; tx 828 drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c if ((rx == rx_curr) && (tx == tx_curr)) tx 832 drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c pdata->new_tx_ring_count = tx; tx 430 drivers/net/ethernet/amd/xgbe/xgbe.h struct xgbe_tx_ring_data tx; /* Tx-related data */ tx 488 drivers/net/ethernet/amd/xgbe/xgbe.h } tx; tx 607 drivers/net/ethernet/apm/xgene/xgene_enet_hw.c u32 *rx, u32 *tx) tx 613 drivers/net/ethernet/apm/xgene/xgene_enet_hw.c *tx = ECM_DROP_COUNT(count); tx 149 drivers/net/ethernet/apm/xgene/xgene_enet_main.h void (*get_drop_cnt)(struct xgene_enet_pdata *pdata, u32 *rx, u32 *tx); tx 87 drivers/net/ethernet/apm/xgene/xgene_enet_sgmac.c u32 *rx, u32 *tx) tx 96 drivers/net/ethernet/apm/xgene/xgene_enet_sgmac.c *tx = ECM_DROP_COUNT(count); tx 172 drivers/net/ethernet/apm/xgene/xgene_enet_xgmac.c u32 *rx, u32 *tx) tx 178 drivers/net/ethernet/apm/xgene/xgene_enet_xgmac.c *tx = ECM_DROP_COUNT(count); tx 208 drivers/net/ethernet/aquantia/atlantic/aq_ring.c ring->stats.tx.queue_restarts++; tx 248 drivers/net/ethernet/aquantia/atlantic/aq_ring.c self->stats.tx.bytes += buff->skb->len; tx 108 drivers/net/ethernet/aquantia/atlantic/aq_ring.h struct aq_ring_stats_tx_s tx; tx 343 drivers/net/ethernet/aquantia/atlantic/aq_vec.c struct aq_ring_stats_tx_s *tx = &ring[AQ_VEC_TX_ID].stats.tx; tx 355 drivers/net/ethernet/aquantia/atlantic/aq_vec.c stats_tx->packets += tx->packets; tx 356 drivers/net/ethernet/aquantia/atlantic/aq_vec.c stats_tx->bytes += tx->bytes; tx 357 drivers/net/ethernet/aquantia/atlantic/aq_vec.c stats_tx->errors += tx->errors; tx 358 drivers/net/ethernet/aquantia/atlantic/aq_vec.c stats_tx->queue_restarts += tx->queue_restarts; tx 129 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_llh.c u32 tx) tx 163 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_llh.c aq_hw_write_reg_bit(aq_hw, itr_imr_txten_adr[tx], tx 164 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_llh.c itr_imr_txten_msk[tx], tx 165 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_llh.c itr_imr_txten_shift[tx], tx 209 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_llh.c void hw_atl_itr_irq_map_tx_set(struct aq_hw_s *aq_hw, u32 irq_map_tx, u32 tx) tx 243 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_llh.c aq_hw_write_reg_bit(aq_hw, itr_imr_txt_adr[tx], tx 244 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_llh.c itr_imr_txt_msk[tx], tx 245 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_llh.c itr_imr_txt_shift[tx], tx 106 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_llh.h u32 tx); tx 112 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_llh.h void hw_atl_itr_irq_map_tx_set(struct aq_hw_s *aq_hw, u32 irq_map_tx, u32 tx); tx 255 drivers/net/ethernet/atheros/ag71xx.c } tx; tx 649 drivers/net/ethernet/atheros/ag71xx.c skb = ring->buf[i].tx.skb; tx 669 drivers/net/ethernet/atheros/ag71xx.c ring->buf[i].tx.skb = NULL; tx 671 drivers/net/ethernet/atheros/ag71xx.c bytes_compl += ring->buf[i].tx.len; tx 706 drivers/net/ethernet/atheros/ag71xx.c u32 rx, tx; tx 711 drivers/net/ethernet/atheros/ag71xx.c tx = ag71xx_rr(ag, AG71XX_REG_TX_CTRL) & TX_CTRL_TXE; tx 712 drivers/net/ethernet/atheros/ag71xx.c if (!rx && !tx) tx 972 drivers/net/ethernet/atheros/ag71xx.c if (ring->buf[i].tx.skb) { tx 973 drivers/net/ethernet/atheros/ag71xx.c bytes_compl += ring->buf[i].tx.len; tx 975 drivers/net/ethernet/atheros/ag71xx.c dev_kfree_skb_any(ring->buf[i].tx.skb); tx 977 drivers/net/ethernet/atheros/ag71xx.c ring->buf[i].tx.skb = NULL; tx 1001 drivers/net/ethernet/atheros/ag71xx.c ring->buf[i].tx.skb = NULL; tx 1133 drivers/net/ethernet/atheros/ag71xx.c struct ag71xx_ring *tx = &ag->tx_ring; tx 1137 drivers/net/ethernet/atheros/ag71xx.c ring_size = BIT(tx->order) + BIT(rx->order); tx 1138 drivers/net/ethernet/atheros/ag71xx.c tx_size = BIT(tx->order); tx 1140 drivers/net/ethernet/atheros/ag71xx.c tx->buf = kcalloc(ring_size, sizeof(*tx->buf), GFP_KERNEL); tx 1141 drivers/net/ethernet/atheros/ag71xx.c if (!tx->buf) tx 1144 drivers/net/ethernet/atheros/ag71xx.c tx->descs_cpu = dma_alloc_coherent(&ag->pdev->dev, tx 1146 drivers/net/ethernet/atheros/ag71xx.c &tx->descs_dma, GFP_KERNEL); tx 1147 drivers/net/ethernet/atheros/ag71xx.c if (!tx->descs_cpu) { tx 1148 drivers/net/ethernet/atheros/ag71xx.c kfree(tx->buf); tx 1149 drivers/net/ethernet/atheros/ag71xx.c tx->buf = NULL; tx 1153 drivers/net/ethernet/atheros/ag71xx.c rx->buf = &tx->buf[tx_size]; tx 1154 drivers/net/ethernet/atheros/ag71xx.c rx->descs_cpu = ((void *)tx->descs_cpu) + tx_size * AG71XX_DESC_SIZE; tx 1155 drivers/net/ethernet/atheros/ag71xx.c rx->descs_dma = tx->descs_dma + tx_size * AG71XX_DESC_SIZE; tx 1163 drivers/net/ethernet/atheros/ag71xx.c struct ag71xx_ring *tx = &ag->tx_ring; tx 1167 drivers/net/ethernet/atheros/ag71xx.c ring_size = BIT(tx->order) + BIT(rx->order); tx 1169 drivers/net/ethernet/atheros/ag71xx.c if (tx->descs_cpu) tx 1171 drivers/net/ethernet/atheros/ag71xx.c tx->descs_cpu, tx->descs_dma); tx 1173 drivers/net/ethernet/atheros/ag71xx.c kfree(tx->buf); tx 1175 drivers/net/ethernet/atheros/ag71xx.c tx->descs_cpu = NULL; tx 1177 drivers/net/ethernet/atheros/ag71xx.c tx->buf = NULL; tx 1358 drivers/net/ethernet/atheros/ag71xx.c ring->buf[i].tx.len = skb->len; tx 1359 drivers/net/ethernet/atheros/ag71xx.c ring->buf[i].tx.skb = skb; tx 1130 drivers/net/ethernet/aurora/nb8800.c u32 rx, tx; tx 1134 drivers/net/ethernet/aurora/nb8800.c tx = nb8800_read_stat(dev, i | 0x80); tx 1136 drivers/net/ethernet/aurora/nb8800.c st[i + NB8800_NUM_STATS / 2] = tx; tx 235 drivers/net/ethernet/broadcom/bcmsysport.c STAT_MIB_TX("tx_64_octets", mib.tx.pkt_cnt.cnt_64), tx 236 drivers/net/ethernet/broadcom/bcmsysport.c STAT_MIB_TX("tx_65_127_oct", mib.tx.pkt_cnt.cnt_127), tx 237 drivers/net/ethernet/broadcom/bcmsysport.c STAT_MIB_TX("tx_128_255_oct", mib.tx.pkt_cnt.cnt_255), tx 238 drivers/net/ethernet/broadcom/bcmsysport.c STAT_MIB_TX("tx_256_511_oct", mib.tx.pkt_cnt.cnt_511), tx 239 drivers/net/ethernet/broadcom/bcmsysport.c STAT_MIB_TX("tx_512_1023_oct", mib.tx.pkt_cnt.cnt_1023), tx 240 drivers/net/ethernet/broadcom/bcmsysport.c STAT_MIB_TX("tx_1024_1518_oct", mib.tx.pkt_cnt.cnt_1518), tx 241 drivers/net/ethernet/broadcom/bcmsysport.c STAT_MIB_TX("tx_vlan_1519_1522_oct", mib.tx.pkt_cnt.cnt_mgv), tx 242 drivers/net/ethernet/broadcom/bcmsysport.c STAT_MIB_TX("tx_1522_2047_oct", mib.tx.pkt_cnt.cnt_2047), tx 243 drivers/net/ethernet/broadcom/bcmsysport.c STAT_MIB_TX("tx_2048_4095_oct", mib.tx.pkt_cnt.cnt_4095), tx 244 drivers/net/ethernet/broadcom/bcmsysport.c STAT_MIB_TX("tx_4096_9216_oct", mib.tx.pkt_cnt.cnt_9216), tx 245 drivers/net/ethernet/broadcom/bcmsysport.c STAT_MIB_TX("tx_pkts", mib.tx.pkts), tx 246 drivers/net/ethernet/broadcom/bcmsysport.c STAT_MIB_TX("tx_multicast", mib.tx.mca), tx 247 drivers/net/ethernet/broadcom/bcmsysport.c STAT_MIB_TX("tx_broadcast", mib.tx.bca), tx 248 drivers/net/ethernet/broadcom/bcmsysport.c STAT_MIB_TX("tx_pause", mib.tx.pf), tx 249 drivers/net/ethernet/broadcom/bcmsysport.c STAT_MIB_TX("tx_control", mib.tx.cf), tx 250 drivers/net/ethernet/broadcom/bcmsysport.c STAT_MIB_TX("tx_fcs_err", mib.tx.fcs), tx 251 drivers/net/ethernet/broadcom/bcmsysport.c STAT_MIB_TX("tx_oversize", mib.tx.ovr), tx 252 drivers/net/ethernet/broadcom/bcmsysport.c STAT_MIB_TX("tx_defer", mib.tx.drf), tx 253 drivers/net/ethernet/broadcom/bcmsysport.c STAT_MIB_TX("tx_excess_defer", mib.tx.edf), tx 254 drivers/net/ethernet/broadcom/bcmsysport.c STAT_MIB_TX("tx_single_col", mib.tx.scl), tx 255 drivers/net/ethernet/broadcom/bcmsysport.c STAT_MIB_TX("tx_multi_col", mib.tx.mcl), tx 256 drivers/net/ethernet/broadcom/bcmsysport.c STAT_MIB_TX("tx_late_col", mib.tx.lcl), tx 257 drivers/net/ethernet/broadcom/bcmsysport.c STAT_MIB_TX("tx_excess_col", mib.tx.ecl), tx 258 drivers/net/ethernet/broadcom/bcmsysport.c STAT_MIB_TX("tx_frags", mib.tx.frg), tx 259 drivers/net/ethernet/broadcom/bcmsysport.c STAT_MIB_TX("tx_total_col", mib.tx.ncl), tx 260 drivers/net/ethernet/broadcom/bcmsysport.c STAT_MIB_TX("tx_jabber", mib.tx.jbr), tx 261 drivers/net/ethernet/broadcom/bcmsysport.c STAT_MIB_TX("tx_bytes", mib.tx.bytes), tx 262 drivers/net/ethernet/broadcom/bcmsysport.c STAT_MIB_TX("tx_good_pkts", mib.tx.pok), tx 263 drivers/net/ethernet/broadcom/bcmsysport.c STAT_MIB_TX("tx_unicast", mib.tx.uc), tx 590 drivers/net/ethernet/broadcom/bcmsysport.h struct bcm_sysport_tx_counters tx; tx 7338 drivers/net/ethernet/broadcom/bnx2.c bnx2_change_ring_size(struct bnx2 *bp, u32 rx, u32 tx, bool reset_irq) tx 7357 drivers/net/ethernet/broadcom/bnx2.c bp->tx_ring_size = tx; tx 1980 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c int rc, tx, rx; tx 1982 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c tx = BNX2X_NUM_ETH_QUEUES(bp) * bp->max_cos; tx 1988 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c tx++; tx 1991 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c rc = netif_set_real_num_tx_queues(bp->dev, tx); tx 2003 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c tx, rx); tx 3741 drivers/net/ethernet/broadcom/bnx2x/bnx2x_hsi.h struct client_init_tx_data tx; tx 4814 drivers/net/ethernet/broadcom/bnx2x/bnx2x_hsi.h struct client_init_tx_data tx; tx 12159 drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c u32 rx = 0, tx = 0, i; tx 12170 drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c tx = REG_RD(bp, shmem_base + tx 12178 drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c tx = REG_RD(bp, shmem_base + tx 12186 drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c phy->tx_preemphasis[i << 1] = ((tx>>16) & 0xffff); tx 12187 drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c phy->tx_preemphasis[(i << 1) + 1] = (tx & 0xffff); tx 8727 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c __set_bit(BNX2X_Q_FLG_HC, &init_params->tx.flags); tx 8733 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c __set_bit(BNX2X_Q_FLG_HC_EN, &init_params->tx.flags); tx 8738 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c init_params->tx.hc_rate = bp->tx_ticks ? tx 8742 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c init_params->rx.fw_sb_id = init_params->tx.fw_sb_id = tx 8750 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c init_params->tx.sb_cq_index = HC_INDEX_ETH_FIRST_TX_CQ_CONS; tx 4956 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c &data->tx, tx 4981 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c &data->tx, tx 4986 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c data->tx.tx_bd_page_base.lo, tx 4987 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c data->tx.tx_bd_page_base.hi); tx 5011 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c test_bit(BNX2X_Q_FLG_HC, &init->tx.flags)) { tx 5012 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c hc_usec = init->tx.hc_rate ? 1000000 / init->tx.hc_rate : 0; tx 5014 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c bnx2x_update_coalesce_sb_index(bp, init->tx.fw_sb_id, tx 5015 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c init->tx.sb_cq_index, tx 5016 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c !test_bit(BNX2X_Q_FLG_HC_EN, &init->tx.flags), tx 886 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.h } tx; tx 136 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c init_params->tx.sb_cq_index, tx 137 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c init_params->tx.hc_rate, tx 181 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c if (test_bit(BNX2X_Q_FLG_HC, &init_p->tx.flags)) tx 182 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c __set_bit(BNX2X_Q_FLG_HC_EN, &init_p->tx.flags); tx 186 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c init_p->tx.fw_sb_id = vf_igu_sb(vf, q->sb_idx); tx 1565 drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.c init_p->tx.hc_rate = setup_q->txq.hc_rate; tx 1566 drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.c init_p->tx.sb_cq_index = setup_q->txq.sb_index; tx 1569 drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.c &init_p->tx.flags); tx 5650 drivers/net/ethernet/broadcom/bnxt/bnxt.c static int bnxt_trim_rings(struct bnxt *bp, int *rx, int *tx, int max, tx 5685 drivers/net/ethernet/broadcom/bnxt/bnxt.c int tx = hw_resc->resv_tx_rings; tx 5689 drivers/net/ethernet/broadcom/bnxt/bnxt.c if (cp < (rx + tx)) { tx 5690 drivers/net/ethernet/broadcom/bnxt/bnxt.c bnxt_trim_rings(bp, &rx, &tx, cp, false); tx 5694 drivers/net/ethernet/broadcom/bnxt/bnxt.c hw_resc->resv_tx_rings = tx; tx 5864 drivers/net/ethernet/broadcom/bnxt/bnxt.c static int bnxt_hwrm_reserve_rings(struct bnxt *bp, int tx, int rx, int grp, tx 5868 drivers/net/ethernet/broadcom/bnxt/bnxt.c return bnxt_hwrm_reserve_pf_rings(bp, tx, rx, grp, cp, stat, tx 5871 drivers/net/ethernet/broadcom/bnxt/bnxt.c return bnxt_hwrm_reserve_vf_rings(bp, tx, rx, grp, cp, stat, tx 5950 drivers/net/ethernet/broadcom/bnxt/bnxt.c int tx = bp->tx_nr_rings; tx 5968 drivers/net/ethernet/broadcom/bnxt/bnxt.c rc = bnxt_hwrm_reserve_rings(bp, tx, rx, grp, cp, stat, vnic); tx 5972 drivers/net/ethernet/broadcom/bnxt/bnxt.c tx = hw_resc->resv_tx_rings; tx 6001 drivers/net/ethernet/broadcom/bnxt/bnxt.c rc = bnxt_trim_rings(bp, &rx_rings, &tx, cp, sh); tx 6004 drivers/net/ethernet/broadcom/bnxt/bnxt.c cp = sh ? max_t(int, tx, rx_rings) : tx + rx_rings; tx 6005 drivers/net/ethernet/broadcom/bnxt/bnxt.c bp->tx_nr_rings = tx; tx 6009 drivers/net/ethernet/broadcom/bnxt/bnxt.c if (!tx || !rx || !cp || !grp || !vnic || !stat) tx 7854 drivers/net/ethernet/broadcom/bnxt/bnxt.c static int bnxt_trim_rings(struct bnxt *bp, int *rx, int *tx, int max, tx 7857 drivers/net/ethernet/broadcom/bnxt/bnxt.c int _rx = *rx, _tx = *tx; tx 7861 drivers/net/ethernet/broadcom/bnxt/bnxt.c *tx = min_t(int, _tx, max); tx 7873 drivers/net/ethernet/broadcom/bnxt/bnxt.c *tx = _tx; tx 9498 drivers/net/ethernet/broadcom/bnxt/bnxt.c struct tx_port_stats *tx = bp->hw_tx_port_stats; tx 9507 drivers/net/ethernet/broadcom/bnxt/bnxt.c stats->collisions = le64_to_cpu(tx->tx_total_collisions); tx 9508 drivers/net/ethernet/broadcom/bnxt/bnxt.c stats->tx_fifo_errors = le64_to_cpu(tx->tx_fifo_underruns); tx 9509 drivers/net/ethernet/broadcom/bnxt/bnxt.c stats->tx_errors = le64_to_cpu(tx->tx_err); tx 10342 drivers/net/ethernet/broadcom/bnxt/bnxt.c int bnxt_check_rings(struct bnxt *bp, int tx, int rx, bool sh, int tcs, tx 10360 drivers/net/ethernet/broadcom/bnxt/bnxt.c tx_rings_needed = tx * tx_sets + tx_xdp; tx 11494 drivers/net/ethernet/broadcom/bnxt/bnxt.c int rx, tx, cp; tx 11496 drivers/net/ethernet/broadcom/bnxt/bnxt.c _bnxt_get_max_rings(bp, &rx, &tx, &cp); tx 11498 drivers/net/ethernet/broadcom/bnxt/bnxt.c *max_tx = tx; tx 11499 drivers/net/ethernet/broadcom/bnxt/bnxt.c if (!rx || !tx || !cp) tx 1997 drivers/net/ethernet/broadcom/bnxt/bnxt.h int bnxt_check_rings(struct bnxt *bp, int tx, int rx, bool sh, int tcs, tx 302 drivers/net/ethernet/broadcom/bnxt/bnxt_xdp.c int rx, tx; tx 305 drivers/net/ethernet/broadcom/bnxt/bnxt_xdp.c bnxt_get_max_rings(bp, &rx, &tx, true); tx 4961 drivers/net/ethernet/broadcom/cnic.c data->tx.tx_bd_page_base.hi = cpu_to_le32(val); tx 4966 drivers/net/ethernet/broadcom/cnic.c data->tx.tx_bd_page_base.lo = cpu_to_le32(val); tx 4969 drivers/net/ethernet/broadcom/cnic.c data->tx.tx_sb_index_number = HC_SP_INDEX_ETH_ISCSI_CQ_CONS; tx 4970 drivers/net/ethernet/broadcom/cnic.c data->tx.tx_status_block_id = BNX2X_DEF_SB_ID; tx 817 drivers/net/ethernet/broadcom/genet/bcmgenet.c STAT_GENET_MIB_TX("tx_64_octets", mib.tx.pkt_cnt.cnt_64), tx 818 drivers/net/ethernet/broadcom/genet/bcmgenet.c STAT_GENET_MIB_TX("tx_65_127_oct", mib.tx.pkt_cnt.cnt_127), tx 819 drivers/net/ethernet/broadcom/genet/bcmgenet.c STAT_GENET_MIB_TX("tx_128_255_oct", mib.tx.pkt_cnt.cnt_255), tx 820 drivers/net/ethernet/broadcom/genet/bcmgenet.c STAT_GENET_MIB_TX("tx_256_511_oct", mib.tx.pkt_cnt.cnt_511), tx 821 drivers/net/ethernet/broadcom/genet/bcmgenet.c STAT_GENET_MIB_TX("tx_512_1023_oct", mib.tx.pkt_cnt.cnt_1023), tx 822 drivers/net/ethernet/broadcom/genet/bcmgenet.c STAT_GENET_MIB_TX("tx_1024_1518_oct", mib.tx.pkt_cnt.cnt_1518), tx 823 drivers/net/ethernet/broadcom/genet/bcmgenet.c STAT_GENET_MIB_TX("tx_vlan_1519_1522_oct", mib.tx.pkt_cnt.cnt_mgv), tx 824 drivers/net/ethernet/broadcom/genet/bcmgenet.c STAT_GENET_MIB_TX("tx_1522_2047_oct", mib.tx.pkt_cnt.cnt_2047), tx 825 drivers/net/ethernet/broadcom/genet/bcmgenet.c STAT_GENET_MIB_TX("tx_2048_4095_oct", mib.tx.pkt_cnt.cnt_4095), tx 826 drivers/net/ethernet/broadcom/genet/bcmgenet.c STAT_GENET_MIB_TX("tx_4096_9216_oct", mib.tx.pkt_cnt.cnt_9216), tx 827 drivers/net/ethernet/broadcom/genet/bcmgenet.c STAT_GENET_MIB_TX("tx_pkts", mib.tx.pkts), tx 828 drivers/net/ethernet/broadcom/genet/bcmgenet.c STAT_GENET_MIB_TX("tx_multicast", mib.tx.mca), tx 829 drivers/net/ethernet/broadcom/genet/bcmgenet.c STAT_GENET_MIB_TX("tx_broadcast", mib.tx.bca), tx 830 drivers/net/ethernet/broadcom/genet/bcmgenet.c STAT_GENET_MIB_TX("tx_pause", mib.tx.pf), tx 831 drivers/net/ethernet/broadcom/genet/bcmgenet.c STAT_GENET_MIB_TX("tx_control", mib.tx.cf), tx 832 drivers/net/ethernet/broadcom/genet/bcmgenet.c STAT_GENET_MIB_TX("tx_fcs_err", mib.tx.fcs), tx 833 drivers/net/ethernet/broadcom/genet/bcmgenet.c STAT_GENET_MIB_TX("tx_oversize", mib.tx.ovr), tx 834 drivers/net/ethernet/broadcom/genet/bcmgenet.c STAT_GENET_MIB_TX("tx_defer", mib.tx.drf), tx 835 drivers/net/ethernet/broadcom/genet/bcmgenet.c STAT_GENET_MIB_TX("tx_excess_defer", mib.tx.edf), tx 836 drivers/net/ethernet/broadcom/genet/bcmgenet.c STAT_GENET_MIB_TX("tx_single_col", mib.tx.scl), tx 837 drivers/net/ethernet/broadcom/genet/bcmgenet.c STAT_GENET_MIB_TX("tx_multi_col", mib.tx.mcl), tx 838 drivers/net/ethernet/broadcom/genet/bcmgenet.c STAT_GENET_MIB_TX("tx_late_col", mib.tx.lcl), tx 839 drivers/net/ethernet/broadcom/genet/bcmgenet.c STAT_GENET_MIB_TX("tx_excess_col", mib.tx.ecl), tx 840 drivers/net/ethernet/broadcom/genet/bcmgenet.c STAT_GENET_MIB_TX("tx_frags", mib.tx.frg), tx 841 drivers/net/ethernet/broadcom/genet/bcmgenet.c STAT_GENET_MIB_TX("tx_total_col", mib.tx.ncl), tx 842 drivers/net/ethernet/broadcom/genet/bcmgenet.c STAT_GENET_MIB_TX("tx_jabber", mib.tx.jbr), tx 843 drivers/net/ethernet/broadcom/genet/bcmgenet.c STAT_GENET_MIB_TX("tx_bytes", mib.tx.bytes), tx 844 drivers/net/ethernet/broadcom/genet/bcmgenet.c STAT_GENET_MIB_TX("tx_good_pkts", mib.tx.pok), tx 845 drivers/net/ethernet/broadcom/genet/bcmgenet.c STAT_GENET_MIB_TX("tx_unicast", mib.tx.uc), tx 136 drivers/net/ethernet/broadcom/genet/bcmgenet.h struct bcmgenet_tx_counters tx; tx 289 drivers/net/ethernet/brocade/bna/bna.h void bna_bfi_tx_enet_start_rsp(struct bna_tx *tx, tx 291 drivers/net/ethernet/brocade/bna/bna.h void bna_bfi_tx_enet_stop_rsp(struct bna_tx *tx, tx 312 drivers/net/ethernet/brocade/bna/bna.h void bna_tx_destroy(struct bna_tx *tx); tx 313 drivers/net/ethernet/brocade/bna/bna.h void bna_tx_enable(struct bna_tx *tx); tx 314 drivers/net/ethernet/brocade/bna/bna.h void bna_tx_disable(struct bna_tx *tx, enum bna_cleanup_type type, tx 316 drivers/net/ethernet/brocade/bna/bna.h void bna_tx_cleanup_complete(struct bna_tx *tx); tx 317 drivers/net/ethernet/brocade/bna/bna.h void bna_tx_coalescing_timeo_set(struct bna_tx *tx, int coalescing_timeo); tx 276 drivers/net/ethernet/brocade/bna/bna_enet.c struct bna_tx *tx; tx 322 drivers/net/ethernet/brocade/bna/bna_enet.c bna_tx_from_rid(bna, msghdr->enet_id, tx); tx 323 drivers/net/ethernet/brocade/bna/bna_enet.c if (tx) tx 324 drivers/net/ethernet/brocade/bna/bna_enet.c bna_bfi_tx_enet_start_rsp(tx, msghdr); tx 328 drivers/net/ethernet/brocade/bna/bna_enet.c bna_tx_from_rid(bna, msghdr->enet_id, tx); tx 329 drivers/net/ethernet/brocade/bna/bna_enet.c if (tx) tx 330 drivers/net/ethernet/brocade/bna/bna_enet.c bna_bfi_tx_enet_stop_rsp(tx, msghdr); tx 2758 drivers/net/ethernet/brocade/bna/bna_tx_rx.c #define call_tx_stop_cbfn(tx) \ tx 2760 drivers/net/ethernet/brocade/bna/bna_tx_rx.c if ((tx)->stop_cbfn) { \ tx 2763 drivers/net/ethernet/brocade/bna/bna_tx_rx.c cbfn = (tx)->stop_cbfn; \ tx 2764 drivers/net/ethernet/brocade/bna/bna_tx_rx.c cbarg = (tx)->stop_cbarg; \ tx 2765 drivers/net/ethernet/brocade/bna/bna_tx_rx.c (tx)->stop_cbfn = NULL; \ tx 2766 drivers/net/ethernet/brocade/bna/bna_tx_rx.c (tx)->stop_cbarg = NULL; \ tx 2767 drivers/net/ethernet/brocade/bna/bna_tx_rx.c cbfn(cbarg, (tx)); \ tx 2771 drivers/net/ethernet/brocade/bna/bna_tx_rx.c static void bna_tx_mod_cb_tx_stopped(void *tx_mod, struct bna_tx *tx); tx 2772 drivers/net/ethernet/brocade/bna/bna_tx_rx.c static void bna_bfi_tx_enet_start(struct bna_tx *tx); tx 2773 drivers/net/ethernet/brocade/bna/bna_tx_rx.c static void bna_tx_enet_stop(struct bna_tx *tx); tx 2800 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_sm_stopped_entry(struct bna_tx *tx) tx 2802 drivers/net/ethernet/brocade/bna/bna_tx_rx.c call_tx_stop_cbfn(tx); tx 2806 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_sm_stopped(struct bna_tx *tx, enum bna_tx_event event) tx 2810 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bfa_fsm_set_state(tx, bna_tx_sm_start_wait); tx 2814 drivers/net/ethernet/brocade/bna/bna_tx_rx.c call_tx_stop_cbfn(tx); tx 2831 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_sm_start_wait_entry(struct bna_tx *tx) tx 2833 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_bfi_tx_enet_start(tx); tx 2837 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_sm_start_wait(struct bna_tx *tx, enum bna_tx_event event) tx 2841 drivers/net/ethernet/brocade/bna/bna_tx_rx.c tx->flags &= ~BNA_TX_F_BW_UPDATED; tx 2842 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bfa_fsm_set_state(tx, bna_tx_sm_stop_wait); tx 2846 drivers/net/ethernet/brocade/bna/bna_tx_rx.c tx->flags &= ~BNA_TX_F_BW_UPDATED; tx 2847 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bfa_fsm_set_state(tx, bna_tx_sm_stopped); tx 2851 drivers/net/ethernet/brocade/bna/bna_tx_rx.c if (tx->flags & BNA_TX_F_BW_UPDATED) { tx 2852 drivers/net/ethernet/brocade/bna/bna_tx_rx.c tx->flags &= ~BNA_TX_F_BW_UPDATED; tx 2853 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bfa_fsm_set_state(tx, bna_tx_sm_prio_stop_wait); tx 2855 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bfa_fsm_set_state(tx, bna_tx_sm_started); tx 2859 drivers/net/ethernet/brocade/bna/bna_tx_rx.c tx->flags |= BNA_TX_F_BW_UPDATED; tx 2868 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_sm_started_entry(struct bna_tx *tx) tx 2871 drivers/net/ethernet/brocade/bna/bna_tx_rx.c int is_regular = (tx->type == BNA_TX_T_REGULAR); tx 2873 drivers/net/ethernet/brocade/bna/bna_tx_rx.c list_for_each_entry(txq, &tx->txq_q, qe) { tx 2876 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_ib_start(tx->bna, &txq->ib, is_regular); tx 2878 drivers/net/ethernet/brocade/bna/bna_tx_rx.c tx->tx_resume_cbfn(tx->bna->bnad, tx); tx 2882 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_sm_started(struct bna_tx *tx, enum bna_tx_event event) tx 2886 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bfa_fsm_set_state(tx, bna_tx_sm_stop_wait); tx 2887 drivers/net/ethernet/brocade/bna/bna_tx_rx.c tx->tx_stall_cbfn(tx->bna->bnad, tx); tx 2888 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_enet_stop(tx); tx 2892 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bfa_fsm_set_state(tx, bna_tx_sm_failed); tx 2893 drivers/net/ethernet/brocade/bna/bna_tx_rx.c tx->tx_stall_cbfn(tx->bna->bnad, tx); tx 2894 drivers/net/ethernet/brocade/bna/bna_tx_rx.c tx->tx_cleanup_cbfn(tx->bna->bnad, tx); tx 2898 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bfa_fsm_set_state(tx, bna_tx_sm_prio_stop_wait); tx 2907 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_sm_stop_wait_entry(struct bna_tx *tx) tx 2912 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_sm_stop_wait(struct bna_tx *tx, enum bna_tx_event event) tx 2917 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bfa_fsm_set_state(tx, bna_tx_sm_cleanup_wait); tx 2918 drivers/net/ethernet/brocade/bna/bna_tx_rx.c tx->tx_cleanup_cbfn(tx->bna->bnad, tx); tx 2926 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_enet_stop(tx); tx 2939 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_sm_cleanup_wait_entry(struct bna_tx *tx) tx 2944 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_sm_cleanup_wait(struct bna_tx *tx, enum bna_tx_event event) tx 2953 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bfa_fsm_set_state(tx, bna_tx_sm_stopped); tx 2962 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_sm_prio_stop_wait_entry(struct bna_tx *tx) tx 2964 drivers/net/ethernet/brocade/bna/bna_tx_rx.c tx->tx_stall_cbfn(tx->bna->bnad, tx); tx 2965 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_enet_stop(tx); tx 2969 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_sm_prio_stop_wait(struct bna_tx *tx, enum bna_tx_event event) tx 2973 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bfa_fsm_set_state(tx, bna_tx_sm_stop_wait); tx 2977 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bfa_fsm_set_state(tx, bna_tx_sm_failed); tx 2978 drivers/net/ethernet/brocade/bna/bna_tx_rx.c tx->tx_cleanup_cbfn(tx->bna->bnad, tx); tx 2982 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bfa_fsm_set_state(tx, bna_tx_sm_prio_cleanup_wait); tx 2995 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_sm_prio_cleanup_wait_entry(struct bna_tx *tx) tx 2997 drivers/net/ethernet/brocade/bna/bna_tx_rx.c tx->tx_cleanup_cbfn(tx->bna->bnad, tx); tx 3001 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_sm_prio_cleanup_wait(struct bna_tx *tx, enum bna_tx_event event) tx 3005 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bfa_fsm_set_state(tx, bna_tx_sm_cleanup_wait); tx 3009 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bfa_fsm_set_state(tx, bna_tx_sm_failed); tx 3017 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bfa_fsm_set_state(tx, bna_tx_sm_start_wait); tx 3026 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_sm_failed_entry(struct bna_tx *tx) tx 3031 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_sm_failed(struct bna_tx *tx, enum bna_tx_event event) tx 3035 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bfa_fsm_set_state(tx, bna_tx_sm_quiesce_wait); tx 3039 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bfa_fsm_set_state(tx, bna_tx_sm_cleanup_wait); tx 3047 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bfa_fsm_set_state(tx, bna_tx_sm_stopped); tx 3056 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_sm_quiesce_wait_entry(struct bna_tx *tx) tx 3061 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_sm_quiesce_wait(struct bna_tx *tx, enum bna_tx_event event) tx 3065 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bfa_fsm_set_state(tx, bna_tx_sm_cleanup_wait); tx 3069 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bfa_fsm_set_state(tx, bna_tx_sm_failed); tx 3073 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bfa_fsm_set_state(tx, bna_tx_sm_start_wait); tx 3086 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_bfi_tx_enet_start(struct bna_tx *tx) tx 3088 drivers/net/ethernet/brocade/bna/bna_tx_rx.c struct bfi_enet_tx_cfg_req *cfg_req = &tx->bfi_enet_cmd.cfg_req; tx 3093 drivers/net/ethernet/brocade/bna/bna_tx_rx.c BFI_ENET_H2I_TX_CFG_SET_REQ, 0, tx->rid); tx 3097 drivers/net/ethernet/brocade/bna/bna_tx_rx.c cfg_req->num_queues = tx->num_txq; tx 3098 drivers/net/ethernet/brocade/bna/bna_tx_rx.c for (i = 0; i < tx->num_txq; i++) { tx 3100 drivers/net/ethernet/brocade/bna/bna_tx_rx.c : list_first_entry(&tx->txq_q, struct bna_txq, qe); tx 3125 drivers/net/ethernet/brocade/bna/bna_tx_rx.c cfg_req->tx_cfg.vlan_id = htons((u16)tx->txf_vlan_id); tx 3129 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bfa_msgq_cmd_set(&tx->msgq_cmd, NULL, NULL, tx 3131 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bfa_msgq_cmd_post(&tx->bna->msgq, &tx->msgq_cmd); tx 3135 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_bfi_tx_enet_stop(struct bna_tx *tx) tx 3137 drivers/net/ethernet/brocade/bna/bna_tx_rx.c struct bfi_enet_req *req = &tx->bfi_enet_cmd.req; tx 3140 drivers/net/ethernet/brocade/bna/bna_tx_rx.c BFI_ENET_H2I_TX_CFG_CLR_REQ, 0, tx->rid); tx 3143 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bfa_msgq_cmd_set(&tx->msgq_cmd, NULL, NULL, sizeof(struct bfi_enet_req), tx 3145 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bfa_msgq_cmd_post(&tx->bna->msgq, &tx->msgq_cmd); tx 3149 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_enet_stop(struct bna_tx *tx) tx 3154 drivers/net/ethernet/brocade/bna/bna_tx_rx.c list_for_each_entry(txq, &tx->txq_q, qe) tx 3155 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_ib_stop(tx->bna, &txq->ib); tx 3157 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_bfi_tx_enet_stop(tx); tx 3199 drivers/net/ethernet/brocade/bna/bna_tx_rx.c struct bna_tx *tx = NULL; tx 3204 drivers/net/ethernet/brocade/bna/bna_tx_rx.c tx = list_first_entry(&tx_mod->tx_free_q, struct bna_tx, qe); tx 3206 drivers/net/ethernet/brocade/bna/bna_tx_rx.c tx = list_last_entry(&tx_mod->tx_free_q, struct bna_tx, qe); tx 3207 drivers/net/ethernet/brocade/bna/bna_tx_rx.c list_del(&tx->qe); tx 3208 drivers/net/ethernet/brocade/bna/bna_tx_rx.c tx->type = type; tx 3210 drivers/net/ethernet/brocade/bna/bna_tx_rx.c return tx; tx 3214 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_free(struct bna_tx *tx) tx 3216 drivers/net/ethernet/brocade/bna/bna_tx_rx.c struct bna_tx_mod *tx_mod = &tx->bna->tx_mod; tx 3220 drivers/net/ethernet/brocade/bna/bna_tx_rx.c while (!list_empty(&tx->txq_q)) { tx 3221 drivers/net/ethernet/brocade/bna/bna_tx_rx.c txq = list_first_entry(&tx->txq_q, struct bna_txq, qe); tx 3223 drivers/net/ethernet/brocade/bna/bna_tx_rx.c txq->tx = NULL; tx 3228 drivers/net/ethernet/brocade/bna/bna_tx_rx.c if (qe == &tx->qe) { tx 3229 drivers/net/ethernet/brocade/bna/bna_tx_rx.c list_del(&tx->qe); tx 3234 drivers/net/ethernet/brocade/bna/bna_tx_rx.c tx->bna = NULL; tx 3235 drivers/net/ethernet/brocade/bna/bna_tx_rx.c tx->priv = NULL; tx 3238 drivers/net/ethernet/brocade/bna/bna_tx_rx.c if (((struct bna_tx *)qe)->rid < tx->rid) tx 3241 drivers/net/ethernet/brocade/bna/bna_tx_rx.c list_add(&tx->qe, qe); tx 3245 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_start(struct bna_tx *tx) tx 3247 drivers/net/ethernet/brocade/bna/bna_tx_rx.c tx->flags |= BNA_TX_F_ENET_STARTED; tx 3248 drivers/net/ethernet/brocade/bna/bna_tx_rx.c if (tx->flags & BNA_TX_F_ENABLED) tx 3249 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bfa_fsm_send_event(tx, TX_E_START); tx 3253 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_stop(struct bna_tx *tx) tx 3255 drivers/net/ethernet/brocade/bna/bna_tx_rx.c tx->stop_cbfn = bna_tx_mod_cb_tx_stopped; tx 3256 drivers/net/ethernet/brocade/bna/bna_tx_rx.c tx->stop_cbarg = &tx->bna->tx_mod; tx 3258 drivers/net/ethernet/brocade/bna/bna_tx_rx.c tx->flags &= ~BNA_TX_F_ENET_STARTED; tx 3259 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bfa_fsm_send_event(tx, TX_E_STOP); tx 3263 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_fail(struct bna_tx *tx) tx 3265 drivers/net/ethernet/brocade/bna/bna_tx_rx.c tx->flags &= ~BNA_TX_F_ENET_STARTED; tx 3266 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bfa_fsm_send_event(tx, TX_E_FAIL); tx 3270 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_bfi_tx_enet_start_rsp(struct bna_tx *tx, struct bfi_msgq_mhdr *msghdr) tx 3272 drivers/net/ethernet/brocade/bna/bna_tx_rx.c struct bfi_enet_tx_cfg_rsp *cfg_rsp = &tx->bfi_enet_cmd.cfg_rsp; tx 3276 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bfa_msgq_rsp_copy(&tx->bna->msgq, (u8 *)cfg_rsp, tx 3279 drivers/net/ethernet/brocade/bna/bna_tx_rx.c tx->hw_id = cfg_rsp->hw_id; tx 3281 drivers/net/ethernet/brocade/bna/bna_tx_rx.c for (i = 0, txq = list_first_entry(&tx->txq_q, struct bna_txq, qe); tx 3282 drivers/net/ethernet/brocade/bna/bna_tx_rx.c i < tx->num_txq; i++, txq = list_next_entry(txq, qe)) { tx 3285 drivers/net/ethernet/brocade/bna/bna_tx_rx.c tx->bna->pcidev.pci_bar_kva tx 3288 drivers/net/ethernet/brocade/bna/bna_tx_rx.c tx->bna->pcidev.pci_bar_kva tx 3297 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bfa_fsm_send_event(tx, TX_E_STARTED); tx 3301 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_bfi_tx_enet_stop_rsp(struct bna_tx *tx, struct bfi_msgq_mhdr *msghdr) tx 3303 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bfa_fsm_send_event(tx, TX_E_STOPPED); tx 3309 drivers/net/ethernet/brocade/bna/bna_tx_rx.c struct bna_tx *tx; tx 3311 drivers/net/ethernet/brocade/bna/bna_tx_rx.c list_for_each_entry(tx, &tx_mod->tx_active_q, qe) tx 3312 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bfa_fsm_send_event(tx, TX_E_BW_UPDATE); tx 3370 drivers/net/ethernet/brocade/bna/bna_tx_rx.c struct bna_tx *tx; tx 3388 drivers/net/ethernet/brocade/bna/bna_tx_rx.c tx = bna_tx_get(tx_mod, tx_cfg->tx_type); tx 3389 drivers/net/ethernet/brocade/bna/bna_tx_rx.c if (!tx) tx 3391 drivers/net/ethernet/brocade/bna/bna_tx_rx.c tx->bna = bna; tx 3392 drivers/net/ethernet/brocade/bna/bna_tx_rx.c tx->priv = priv; tx 3396 drivers/net/ethernet/brocade/bna/bna_tx_rx.c INIT_LIST_HEAD(&tx->txq_q); tx 3402 drivers/net/ethernet/brocade/bna/bna_tx_rx.c list_move_tail(&txq->qe, &tx->txq_q); tx 3403 drivers/net/ethernet/brocade/bna/bna_tx_rx.c txq->tx = tx; tx 3412 drivers/net/ethernet/brocade/bna/bna_tx_rx.c tx->tcb_setup_cbfn = tx_cbfn->tcb_setup_cbfn; tx 3413 drivers/net/ethernet/brocade/bna/bna_tx_rx.c tx->tcb_destroy_cbfn = tx_cbfn->tcb_destroy_cbfn; tx 3415 drivers/net/ethernet/brocade/bna/bna_tx_rx.c tx->tx_stall_cbfn = tx_cbfn->tx_stall_cbfn; tx 3416 drivers/net/ethernet/brocade/bna/bna_tx_rx.c tx->tx_resume_cbfn = tx_cbfn->tx_resume_cbfn; tx 3417 drivers/net/ethernet/brocade/bna/bna_tx_rx.c tx->tx_cleanup_cbfn = tx_cbfn->tx_cleanup_cbfn; tx 3419 drivers/net/ethernet/brocade/bna/bna_tx_rx.c list_add_tail(&tx->qe, &tx_mod->tx_active_q); tx 3421 drivers/net/ethernet/brocade/bna/bna_tx_rx.c tx->num_txq = tx_cfg->num_txq; tx 3423 drivers/net/ethernet/brocade/bna/bna_tx_rx.c tx->flags = 0; tx 3424 drivers/net/ethernet/brocade/bna/bna_tx_rx.c if (tx->bna->tx_mod.flags & BNA_TX_MOD_F_ENET_STARTED) { tx 3425 drivers/net/ethernet/brocade/bna/bna_tx_rx.c switch (tx->type) { tx 3427 drivers/net/ethernet/brocade/bna/bna_tx_rx.c if (!(tx->bna->tx_mod.flags & tx 3429 drivers/net/ethernet/brocade/bna/bna_tx_rx.c tx->flags |= BNA_TX_F_ENET_STARTED; tx 3432 drivers/net/ethernet/brocade/bna/bna_tx_rx.c if (tx->bna->tx_mod.flags & BNA_TX_MOD_F_ENET_LOOPBACK) tx 3433 drivers/net/ethernet/brocade/bna/bna_tx_rx.c tx->flags |= BNA_TX_F_ENET_STARTED; tx 3441 drivers/net/ethernet/brocade/bna/bna_tx_rx.c list_for_each_entry(txq, &tx->txq_q, qe) { tx 3486 drivers/net/ethernet/brocade/bna/bna_tx_rx.c if (tx->tcb_setup_cbfn) tx 3487 drivers/net/ethernet/brocade/bna/bna_tx_rx.c (tx->tcb_setup_cbfn)(bna->bnad, txq->tcb); tx 3497 drivers/net/ethernet/brocade/bna/bna_tx_rx.c tx->txf_vlan_id = 0; tx 3499 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bfa_fsm_set_state(tx, bna_tx_sm_stopped); tx 3501 drivers/net/ethernet/brocade/bna/bna_tx_rx.c tx_mod->rid_mask |= BIT(tx->rid); tx 3503 drivers/net/ethernet/brocade/bna/bna_tx_rx.c return tx; tx 3506 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_free(tx); tx 3511 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_destroy(struct bna_tx *tx) tx 3515 drivers/net/ethernet/brocade/bna/bna_tx_rx.c list_for_each_entry(txq, &tx->txq_q, qe) tx 3516 drivers/net/ethernet/brocade/bna/bna_tx_rx.c if (tx->tcb_destroy_cbfn) tx 3517 drivers/net/ethernet/brocade/bna/bna_tx_rx.c (tx->tcb_destroy_cbfn)(tx->bna->bnad, txq->tcb); tx 3519 drivers/net/ethernet/brocade/bna/bna_tx_rx.c tx->bna->tx_mod.rid_mask &= ~BIT(tx->rid); tx 3520 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_free(tx); tx 3524 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_enable(struct bna_tx *tx) tx 3526 drivers/net/ethernet/brocade/bna/bna_tx_rx.c if (tx->fsm != (bfa_sm_t)bna_tx_sm_stopped) tx 3529 drivers/net/ethernet/brocade/bna/bna_tx_rx.c tx->flags |= BNA_TX_F_ENABLED; tx 3531 drivers/net/ethernet/brocade/bna/bna_tx_rx.c if (tx->flags & BNA_TX_F_ENET_STARTED) tx 3532 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bfa_fsm_send_event(tx, TX_E_START); tx 3536 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_disable(struct bna_tx *tx, enum bna_cleanup_type type, tx 3540 drivers/net/ethernet/brocade/bna/bna_tx_rx.c (*cbfn)(tx->bna->bnad, tx); tx 3544 drivers/net/ethernet/brocade/bna/bna_tx_rx.c tx->stop_cbfn = cbfn; tx 3545 drivers/net/ethernet/brocade/bna/bna_tx_rx.c tx->stop_cbarg = tx->bna->bnad; tx 3547 drivers/net/ethernet/brocade/bna/bna_tx_rx.c tx->flags &= ~BNA_TX_F_ENABLED; tx 3549 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bfa_fsm_send_event(tx, TX_E_STOP); tx 3553 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_cleanup_complete(struct bna_tx *tx) tx 3555 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bfa_fsm_send_event(tx, TX_E_CLEANUP_DONE); tx 3559 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_mod_cb_tx_stopped(void *arg, struct bna_tx *tx) tx 3585 drivers/net/ethernet/brocade/bna/bna_tx_rx.c tx_mod->tx = (struct bna_tx *) tx 3596 drivers/net/ethernet/brocade/bna/bna_tx_rx.c tx_mod->tx[i].rid = i; tx 3597 drivers/net/ethernet/brocade/bna/bna_tx_rx.c list_add_tail(&tx_mod->tx[i].qe, &tx_mod->tx_free_q); tx 3616 drivers/net/ethernet/brocade/bna/bna_tx_rx.c struct bna_tx *tx; tx 3622 drivers/net/ethernet/brocade/bna/bna_tx_rx.c list_for_each_entry(tx, &tx_mod->tx_active_q, qe) tx 3623 drivers/net/ethernet/brocade/bna/bna_tx_rx.c if (tx->type == type) tx 3624 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_start(tx); tx 3630 drivers/net/ethernet/brocade/bna/bna_tx_rx.c struct bna_tx *tx; tx 3639 drivers/net/ethernet/brocade/bna/bna_tx_rx.c list_for_each_entry(tx, &tx_mod->tx_active_q, qe) tx 3640 drivers/net/ethernet/brocade/bna/bna_tx_rx.c if (tx->type == type) { tx 3642 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_stop(tx); tx 3651 drivers/net/ethernet/brocade/bna/bna_tx_rx.c struct bna_tx *tx; tx 3656 drivers/net/ethernet/brocade/bna/bna_tx_rx.c list_for_each_entry(tx, &tx_mod->tx_active_q, qe) tx 3657 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_fail(tx); tx 3661 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_coalescing_timeo_set(struct bna_tx *tx, int coalescing_timeo) tx 3665 drivers/net/ethernet/brocade/bna/bna_tx_rx.c list_for_each_entry(txq, &tx->txq_q, qe) tx 448 drivers/net/ethernet/brocade/bna/bna_types.h struct bna_tx *tx; tx 480 drivers/net/ethernet/brocade/bna/bna_types.h void (*stop_cbfn)(void *arg, struct bna_tx *tx); tx 514 drivers/net/ethernet/brocade/bna/bna_types.h struct bna_tx *tx; /* BFI_MAX_TXQ entries */ tx 991 drivers/net/ethernet/brocade/bna/bnad.c bnad_cb_tx_disabled(void *arg, struct bna_tx *tx) tx 1002 drivers/net/ethernet/brocade/bna/bnad.c (struct bnad_tx_info *)tcb->txq->tx->priv; tx 1012 drivers/net/ethernet/brocade/bna/bnad.c (struct bnad_tx_info *)tcb->txq->tx->priv; tx 1038 drivers/net/ethernet/brocade/bna/bnad.c bnad_cb_tx_stall(struct bnad *bnad, struct bna_tx *tx) tx 1041 drivers/net/ethernet/brocade/bna/bnad.c (struct bnad_tx_info *)tx->priv; tx 1057 drivers/net/ethernet/brocade/bna/bnad.c bnad_cb_tx_resume(struct bnad *bnad, struct bna_tx *tx) tx 1059 drivers/net/ethernet/brocade/bna/bnad.c struct bnad_tx_info *tx_info = (struct bnad_tx_info *)tx->priv; tx 1129 drivers/net/ethernet/brocade/bna/bnad.c bna_tx_cleanup_complete(tx_info->tx); tx 1134 drivers/net/ethernet/brocade/bna/bnad.c bnad_cb_tx_cleanup(struct bnad *bnad, struct bna_tx *tx) tx 1136 drivers/net/ethernet/brocade/bna/bnad.c struct bnad_tx_info *tx_info = (struct bnad_tx_info *)tx->priv; tx 1917 drivers/net/ethernet/brocade/bna/bnad.c if (!tx_info->tx) tx 1922 drivers/net/ethernet/brocade/bna/bnad.c bna_tx_disable(tx_info->tx, BNA_HARD_CLEANUP, bnad_cb_tx_disabled); tx 1931 drivers/net/ethernet/brocade/bna/bnad.c bna_tx_destroy(tx_info->tx); tx 1934 drivers/net/ethernet/brocade/bna/bnad.c tx_info->tx = NULL; tx 1958 drivers/net/ethernet/brocade/bna/bnad.c struct bna_tx *tx; tx 1987 drivers/net/ethernet/brocade/bna/bnad.c tx = bna_tx_create(&bnad->bna, bnad, tx_config, &tx_cbfn, res_info, tx 1990 drivers/net/ethernet/brocade/bna/bnad.c if (!tx) { tx 1994 drivers/net/ethernet/brocade/bna/bnad.c tx_info->tx = tx; tx 2008 drivers/net/ethernet/brocade/bna/bnad.c bna_tx_enable(tx); tx 2015 drivers/net/ethernet/brocade/bna/bnad.c bna_tx_destroy(tx_info->tx); tx 2017 drivers/net/ethernet/brocade/bna/bnad.c tx_info->tx = NULL; tx 2300 drivers/net/ethernet/brocade/bna/bnad.c if (!tx_info->tx) tx 2303 drivers/net/ethernet/brocade/bna/bnad.c bna_tx_coalescing_timeo_set(tx_info->tx, bnad->tx_coalescing_timeo); tx 204 drivers/net/ethernet/brocade/bna/bnad.h struct bna_tx *tx; /* 1:1 between tx_info & tx */ tx 480 drivers/net/ethernet/brocade/bna/bnad_ethtool.c if (!bnad->tx_info[i].tx) tx 682 drivers/net/ethernet/brocade/bna/bnad_ethtool.c if (!bnad->tx_info[i].tx) tx 745 drivers/net/ethernet/brocade/bna/bnad_ethtool.c if (!bnad->tx_info[i].tx) tx 823 drivers/net/ethernet/brocade/bna/bnad_ethtool.c if (!bnad->tx_info[i].tx) tx 644 drivers/net/ethernet/calxeda/xgmac.c static int xgmac_set_flow_ctrl(struct xgmac_priv *priv, int rx, int tx) tx 650 drivers/net/ethernet/calxeda/xgmac.c priv->tx_pause = tx; tx 652 drivers/net/ethernet/calxeda/xgmac.c if (rx || tx) { tx 655 drivers/net/ethernet/calxeda/xgmac.c if (tx) tx 2696 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c #define T3(fmt_spec, s, v) S3(fmt_spec, s, tx[i].v) tx 2697 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c #define T(s, v) S3("u", s, tx[i].v) tx 2706 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c const struct sge_eth_txq *tx = &s->ethtxq[base_qset]; tx 2722 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 4*(7-tx[i].dcb_prio)) & 0xf); tx 2725 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 1*(7-tx[i].dcb_prio)) & 0x1); tx 2764 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c const struct sge_uld_txq *tx; tx 2767 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c tx = &utxq_info->uldtxq[r * 4]; tx 2913 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c const struct sge_uld_txq *tx; tx 2917 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c tx = &utxq_info->uldtxq[r * 4]; tx 2946 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c const struct sge_ctrl_txq *tx = &s->ctrlq[r * 4]; tx 244 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c const struct sge_eth_txq *tx = &adap->sge.ethtxq[p->first_qset]; tx 248 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c for (i = 0; i < p->nqsets; i++, rx++, tx++) { tx 249 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c s->tso += tx->tso; tx 250 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c s->tx_csum += tx->tx_cso; tx 253 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c s->vlan_ins += tx->vlan_ins; tx 282 drivers/net/ethernet/cisco/enic/enic_ethtool.c *(data++) = ((u64 *)&vstats->tx)[enic_tx_stats[i].index]; tx 924 drivers/net/ethernet/cisco/enic/enic_main.c net_stats->tx_packets = stats->tx.tx_frames_ok; tx 925 drivers/net/ethernet/cisco/enic/enic_main.c net_stats->tx_bytes = stats->tx.tx_bytes_ok; tx 926 drivers/net/ethernet/cisco/enic/enic_main.c net_stats->tx_errors = stats->tx.tx_errors; tx 927 drivers/net/ethernet/cisco/enic/enic_main.c net_stats->tx_dropped = stats->tx.tx_drops; tx 71 drivers/net/ethernet/cisco/enic/vnic_stats.h struct vnic_tx_stats tx; tx 272 drivers/net/ethernet/cortina/gemini.c static void gmac_set_flow_control(struct net_device *netdev, bool tx, bool rx) tx 282 drivers/net/ethernet/cortina/gemini.c if (tx) tx 523 drivers/net/ethernet/dec/tulip/interrupt.c int tx = 0; tx 640 drivers/net/ethernet/dec/tulip/interrupt.c tx++; tx 731 drivers/net/ethernet/dec/tulip/interrupt.c if (tx > maxtx || rx > maxrx || oi > maxoi) { tx 734 drivers/net/ethernet/dec/tulip/interrupt.c csr5, tp->nir, tx, rx, oi); tx 304 drivers/net/ethernet/freescale/fman/fman_port.c struct fman_port_tx_bmi_regs tx; tx 569 drivers/net/ethernet/freescale/fman/fman_port.c struct fman_port_tx_bmi_regs __iomem *regs = &port->bmi_regs->tx; tx 1029 drivers/net/ethernet/freescale/fman/fman_port.c iowrite32be(0xFFFFFF, &port->bmi_regs->tx.fmbm_tcfqid); tx 1031 drivers/net/ethernet/freescale/fman/fman_port.c &port->bmi_regs->tx.fmbm_tfene); tx 1405 drivers/net/ethernet/freescale/fman/fman_port.c iowrite32be(reg, &port->bmi_regs->tx.fmbm_tfp); tx 1605 drivers/net/ethernet/freescale/fman/fman_port.c bmi_cfg_reg = &port->bmi_regs->tx.fmbm_tcfg; tx 1606 drivers/net/ethernet/freescale/fman/fman_port.c bmi_status_reg = &port->bmi_regs->tx.fmbm_tst; tx 1680 drivers/net/ethernet/freescale/fman/fman_port.c bmi_cfg_reg = &port->bmi_regs->tx.fmbm_tcfg; tx 334 drivers/net/ethernet/freescale/fman/mac.c int fman_set_mac_active_pause(struct mac_device *mac_dev, bool rx, bool tx) tx 345 drivers/net/ethernet/freescale/fman/mac.c if (tx != mac_dev->tx_pause_active) { tx 346 drivers/net/ethernet/freescale/fman/mac.c u16 pause_time = (tx ? FSL_FM_PAUSE_TIME_ENABLE : tx 352 drivers/net/ethernet/freescale/fman/mac.c mac_dev->tx_pause_active = tx; tx 96 drivers/net/ethernet/freescale/fman/mac.h int fman_set_mac_active_pause(struct mac_device *mac_dev, bool rx, bool tx); tx 148 drivers/net/ethernet/google/gve/gve.h struct gve_tx_ring *tx; /* tx rings on this block */ tx 166 drivers/net/ethernet/google/gve/gve.h struct gve_tx_ring *tx; /* array of tx_cfg.num_queues */ tx 439 drivers/net/ethernet/google/gve/gve.h struct gve_tx_ring *tx); tx 195 drivers/net/ethernet/google/gve/gve_adminq.c struct gve_tx_ring *tx = &priv->tx[queue_index]; tx 203 drivers/net/ethernet/google/gve/gve_adminq.c .queue_resources_addr = cpu_to_be64(tx->q_resources_bus), tx 204 drivers/net/ethernet/google/gve/gve_adminq.c .tx_ring_addr = cpu_to_be64(tx->bus), tx 205 drivers/net/ethernet/google/gve/gve_adminq.c .queue_page_list_id = cpu_to_be32(tx->tx_fifo.qpl->id), tx 206 drivers/net/ethernet/google/gve/gve_adminq.c .ntfy_id = cpu_to_be32(tx->ntfy_id), tx 285 drivers/net/ethernet/google/gve/gve_adminq.c if (priv->tx_desc_cnt * sizeof(priv->tx->desc[0]) < PAGE_SIZE) { tx 115 drivers/net/ethernet/google/gve/gve_ethtool.c if (priv->tx) { tx 118 drivers/net/ethernet/google/gve/gve_ethtool.c u64_stats_fetch_begin(&priv->tx[ring].statss); tx 119 drivers/net/ethernet/google/gve/gve_ethtool.c tx_pkts += priv->tx[ring].pkt_done; tx 120 drivers/net/ethernet/google/gve/gve_ethtool.c tx_bytes += priv->tx[ring].bytes_done; tx 121 drivers/net/ethernet/google/gve/gve_ethtool.c } while (u64_stats_fetch_retry(&priv->tx[ring].statss, tx 148 drivers/net/ethernet/google/gve/gve_ethtool.c if (priv->tx) { tx 150 drivers/net/ethernet/google/gve/gve_ethtool.c struct gve_tx_ring *tx = &priv->tx[ring]; tx 152 drivers/net/ethernet/google/gve/gve_ethtool.c data[i++] = tx->req; tx 153 drivers/net/ethernet/google/gve/gve_ethtool.c data[i++] = tx->done; tx 154 drivers/net/ethernet/google/gve/gve_ethtool.c data[i++] = tx->wake_queue; tx 155 drivers/net/ethernet/google/gve/gve_ethtool.c data[i++] = tx->stop_queue; tx 157 drivers/net/ethernet/google/gve/gve_ethtool.c tx)); tx 46 drivers/net/ethernet/google/gve/gve_main.c if (priv->tx) { tx 50 drivers/net/ethernet/google/gve/gve_main.c u64_stats_fetch_begin(&priv->tx[ring].statss); tx 51 drivers/net/ethernet/google/gve/gve_main.c s->tx_packets += priv->tx[ring].pkt_done; tx 52 drivers/net/ethernet/google/gve/gve_main.c s->tx_bytes += priv->tx[ring].bytes_done; tx 53 drivers/net/ethernet/google/gve/gve_main.c } while (u64_stats_fetch_retry(&priv->tx[ring].statss, tx 109 drivers/net/ethernet/google/gve/gve_main.c if (block->tx) tx 125 drivers/net/ethernet/google/gve/gve_main.c if (block->tx) tx 415 drivers/net/ethernet/google/gve/gve_main.c priv->tx = kvzalloc(priv->tx_cfg.num_queues * sizeof(*priv->tx), tx 417 drivers/net/ethernet/google/gve/gve_main.c if (!priv->tx) tx 434 drivers/net/ethernet/google/gve/gve_main.c u64_stats_init(&priv->tx[i].statss); tx 453 drivers/net/ethernet/google/gve/gve_main.c kvfree(priv->tx); tx 454 drivers/net/ethernet/google/gve/gve_main.c priv->tx = NULL; tx 497 drivers/net/ethernet/google/gve/gve_main.c if (priv->tx) { tx 503 drivers/net/ethernet/google/gve/gve_main.c kvfree(priv->tx); tx 504 drivers/net/ethernet/google/gve/gve_main.c priv->tx = NULL; tx 139 drivers/net/ethernet/google/gve/gve_tx.c block->tx = NULL; tx 142 drivers/net/ethernet/google/gve/gve_tx.c static int gve_clean_tx_done(struct gve_priv *priv, struct gve_tx_ring *tx, tx 147 drivers/net/ethernet/google/gve/gve_tx.c struct gve_tx_ring *tx = &priv->tx[idx]; tx 153 drivers/net/ethernet/google/gve/gve_tx.c slots = tx->mask + 1; tx 154 drivers/net/ethernet/google/gve/gve_tx.c gve_clean_tx_done(priv, tx, tx->req, false); tx 155 drivers/net/ethernet/google/gve/gve_tx.c netdev_tx_reset_queue(tx->netdev_txq); tx 157 drivers/net/ethernet/google/gve/gve_tx.c dma_free_coherent(hdev, sizeof(*tx->q_resources), tx 158 drivers/net/ethernet/google/gve/gve_tx.c tx->q_resources, tx->q_resources_bus); tx 159 drivers/net/ethernet/google/gve/gve_tx.c tx->q_resources = NULL; tx 161 drivers/net/ethernet/google/gve/gve_tx.c gve_tx_fifo_release(priv, &tx->tx_fifo); tx 162 drivers/net/ethernet/google/gve/gve_tx.c gve_unassign_qpl(priv, tx->tx_fifo.qpl->id); tx 163 drivers/net/ethernet/google/gve/gve_tx.c tx->tx_fifo.qpl = NULL; tx 165 drivers/net/ethernet/google/gve/gve_tx.c bytes = sizeof(*tx->desc) * slots; tx 166 drivers/net/ethernet/google/gve/gve_tx.c dma_free_coherent(hdev, bytes, tx->desc, tx->bus); tx 167 drivers/net/ethernet/google/gve/gve_tx.c tx->desc = NULL; tx 169 drivers/net/ethernet/google/gve/gve_tx.c vfree(tx->info); tx 170 drivers/net/ethernet/google/gve/gve_tx.c tx->info = NULL; tx 179 drivers/net/ethernet/google/gve/gve_tx.c struct gve_tx_ring *tx = &priv->tx[queue_idx]; tx 181 drivers/net/ethernet/google/gve/gve_tx.c block->tx = tx; tx 182 drivers/net/ethernet/google/gve/gve_tx.c tx->ntfy_id = ntfy_idx; tx 187 drivers/net/ethernet/google/gve/gve_tx.c struct gve_tx_ring *tx = &priv->tx[idx]; tx 193 drivers/net/ethernet/google/gve/gve_tx.c memset(tx, 0, sizeof(*tx)); tx 194 drivers/net/ethernet/google/gve/gve_tx.c tx->q_num = idx; tx 196 drivers/net/ethernet/google/gve/gve_tx.c tx->mask = slots - 1; tx 199 drivers/net/ethernet/google/gve/gve_tx.c tx->info = vzalloc(sizeof(*tx->info) * slots); tx 200 drivers/net/ethernet/google/gve/gve_tx.c if (!tx->info) tx 204 drivers/net/ethernet/google/gve/gve_tx.c bytes = sizeof(*tx->desc) * slots; tx 205 drivers/net/ethernet/google/gve/gve_tx.c tx->desc = dma_alloc_coherent(hdev, bytes, &tx->bus, GFP_KERNEL); tx 206 drivers/net/ethernet/google/gve/gve_tx.c if (!tx->desc) tx 209 drivers/net/ethernet/google/gve/gve_tx.c tx->tx_fifo.qpl = gve_assign_tx_qpl(priv); tx 212 drivers/net/ethernet/google/gve/gve_tx.c if (gve_tx_fifo_init(priv, &tx->tx_fifo)) tx 215 drivers/net/ethernet/google/gve/gve_tx.c tx->q_resources = tx 217 drivers/net/ethernet/google/gve/gve_tx.c sizeof(*tx->q_resources), tx 218 drivers/net/ethernet/google/gve/gve_tx.c &tx->q_resources_bus, tx 220 drivers/net/ethernet/google/gve/gve_tx.c if (!tx->q_resources) tx 224 drivers/net/ethernet/google/gve/gve_tx.c (unsigned long)tx->bus); tx 225 drivers/net/ethernet/google/gve/gve_tx.c tx->netdev_txq = netdev_get_tx_queue(priv->dev, idx); tx 231 drivers/net/ethernet/google/gve/gve_tx.c gve_tx_fifo_release(priv, &tx->tx_fifo); tx 233 drivers/net/ethernet/google/gve/gve_tx.c dma_free_coherent(hdev, bytes, tx->desc, tx->bus); tx 234 drivers/net/ethernet/google/gve/gve_tx.c tx->desc = NULL; tx 236 drivers/net/ethernet/google/gve/gve_tx.c vfree(tx->info); tx 237 drivers/net/ethernet/google/gve/gve_tx.c tx->info = NULL; tx 280 drivers/net/ethernet/google/gve/gve_tx.c static inline u32 gve_tx_avail(struct gve_tx_ring *tx) tx 282 drivers/net/ethernet/google/gve/gve_tx.c return tx->mask + 1 - (tx->req - tx->done); tx 285 drivers/net/ethernet/google/gve/gve_tx.c static inline int gve_skb_fifo_bytes_required(struct gve_tx_ring *tx, tx 295 drivers/net/ethernet/google/gve/gve_tx.c pad_bytes = gve_tx_fifo_pad_alloc_one_frag(&tx->tx_fifo, tx 313 drivers/net/ethernet/google/gve/gve_tx.c static inline bool gve_can_tx(struct gve_tx_ring *tx, int bytes_required) tx 315 drivers/net/ethernet/google/gve/gve_tx.c return (gve_tx_avail(tx) >= MAX_TX_DESC_NEEDED && tx 316 drivers/net/ethernet/google/gve/gve_tx.c gve_tx_fifo_can_alloc(&tx->tx_fifo, bytes_required)); tx 320 drivers/net/ethernet/google/gve/gve_tx.c static int gve_maybe_stop_tx(struct gve_tx_ring *tx, struct sk_buff *skb) tx 324 drivers/net/ethernet/google/gve/gve_tx.c bytes_required = gve_skb_fifo_bytes_required(tx, skb); tx 325 drivers/net/ethernet/google/gve/gve_tx.c if (likely(gve_can_tx(tx, bytes_required))) tx 329 drivers/net/ethernet/google/gve/gve_tx.c tx->stop_queue++; tx 330 drivers/net/ethernet/google/gve/gve_tx.c netif_tx_stop_queue(tx->netdev_txq); tx 345 drivers/net/ethernet/google/gve/gve_tx.c if (likely(!gve_can_tx(tx, bytes_required))) tx 348 drivers/net/ethernet/google/gve/gve_tx.c netif_tx_start_queue(tx->netdev_txq); tx 349 drivers/net/ethernet/google/gve/gve_tx.c tx->wake_queue++; tx 407 drivers/net/ethernet/google/gve/gve_tx.c static int gve_tx_add_skb(struct gve_tx_ring *tx, struct sk_buff *skb, tx 414 drivers/net/ethernet/google/gve/gve_tx.c u32 idx = tx->req & tx->mask; tx 420 drivers/net/ethernet/google/gve/gve_tx.c info = &tx->info[idx]; tx 421 drivers/net/ethernet/google/gve/gve_tx.c pkt_desc = &tx->desc[idx]; tx 436 drivers/net/ethernet/google/gve/gve_tx.c pad_bytes = gve_tx_fifo_pad_alloc_one_frag(&tx->tx_fifo, hlen); tx 437 drivers/net/ethernet/google/gve/gve_tx.c hdr_nfrags = gve_tx_alloc_fifo(&tx->tx_fifo, hlen + pad_bytes, tx 440 drivers/net/ethernet/google/gve/gve_tx.c payload_nfrags = gve_tx_alloc_fifo(&tx->tx_fifo, skb->len - hlen, tx 448 drivers/net/ethernet/google/gve/gve_tx.c tx->tx_fifo.base + info->iov[hdr_nfrags - 1].iov_offset, tx 450 drivers/net/ethernet/google/gve/gve_tx.c gve_dma_sync_for_device(dev, tx->tx_fifo.qpl->page_buses, tx 456 drivers/net/ethernet/google/gve/gve_tx.c next_idx = (tx->req + 1 + i - payload_iov) & tx->mask; tx 457 drivers/net/ethernet/google/gve/gve_tx.c seg_desc = &tx->desc[next_idx]; tx 464 drivers/net/ethernet/google/gve/gve_tx.c tx->tx_fifo.base + info->iov[i].iov_offset, tx 466 drivers/net/ethernet/google/gve/gve_tx.c gve_dma_sync_for_device(dev, tx->tx_fifo.qpl->page_buses, tx 478 drivers/net/ethernet/google/gve/gve_tx.c struct gve_tx_ring *tx; tx 483 drivers/net/ethernet/google/gve/gve_tx.c tx = &priv->tx[skb_get_queue_mapping(skb)]; tx 484 drivers/net/ethernet/google/gve/gve_tx.c if (unlikely(gve_maybe_stop_tx(tx, skb))) { tx 490 drivers/net/ethernet/google/gve/gve_tx.c gve_tx_put_doorbell(priv, tx->q_resources, tx->req); tx 493 drivers/net/ethernet/google/gve/gve_tx.c nsegs = gve_tx_add_skb(tx, skb, &priv->pdev->dev); tx 495 drivers/net/ethernet/google/gve/gve_tx.c netdev_tx_sent_queue(tx->netdev_txq, skb->len); tx 499 drivers/net/ethernet/google/gve/gve_tx.c tx->req += nsegs; tx 501 drivers/net/ethernet/google/gve/gve_tx.c if (!netif_xmit_stopped(tx->netdev_txq) && netdev_xmit_more()) tx 504 drivers/net/ethernet/google/gve/gve_tx.c gve_tx_put_doorbell(priv, tx->q_resources, tx->req); tx 510 drivers/net/ethernet/google/gve/gve_tx.c static int gve_clean_tx_done(struct gve_priv *priv, struct gve_tx_ring *tx, tx 521 drivers/net/ethernet/google/gve/gve_tx.c idx = tx->done & tx->mask; tx 524 drivers/net/ethernet/google/gve/gve_tx.c tx->q_num, __func__, idx, tx->req, tx->done); tx 525 drivers/net/ethernet/google/gve/gve_tx.c info = &tx->info[idx]; tx 542 drivers/net/ethernet/google/gve/gve_tx.c tx->done++; tx 545 drivers/net/ethernet/google/gve/gve_tx.c gve_tx_free_fifo(&tx->tx_fifo, space_freed); tx 546 drivers/net/ethernet/google/gve/gve_tx.c u64_stats_update_begin(&tx->statss); tx 547 drivers/net/ethernet/google/gve/gve_tx.c tx->bytes_done += bytes; tx 548 drivers/net/ethernet/google/gve/gve_tx.c tx->pkt_done += pkts; tx 549 drivers/net/ethernet/google/gve/gve_tx.c u64_stats_update_end(&tx->statss); tx 550 drivers/net/ethernet/google/gve/gve_tx.c netdev_tx_completed_queue(tx->netdev_txq, pkts, bytes); tx 557 drivers/net/ethernet/google/gve/gve_tx.c if (try_to_wake && netif_tx_queue_stopped(tx->netdev_txq) && tx 558 drivers/net/ethernet/google/gve/gve_tx.c likely(gve_can_tx(tx, GVE_TX_START_THRESH))) { tx 559 drivers/net/ethernet/google/gve/gve_tx.c tx->wake_queue++; tx 560 drivers/net/ethernet/google/gve/gve_tx.c netif_tx_wake_queue(tx->netdev_txq); tx 567 drivers/net/ethernet/google/gve/gve_tx.c struct gve_tx_ring *tx) tx 569 drivers/net/ethernet/google/gve/gve_tx.c u32 counter_index = be32_to_cpu((tx->q_resources->counter_index)); tx 577 drivers/net/ethernet/google/gve/gve_tx.c struct gve_tx_ring *tx = block->tx; tx 587 drivers/net/ethernet/google/gve/gve_tx.c tx->last_nic_done = gve_tx_load_event_counter(priv, tx); tx 588 drivers/net/ethernet/google/gve/gve_tx.c nic_done = be32_to_cpu(tx->last_nic_done); tx 593 drivers/net/ethernet/google/gve/gve_tx.c to_do = min_t(u32, (nic_done - tx->done), budget); tx 594 drivers/net/ethernet/google/gve/gve_tx.c gve_clean_tx_done(priv, tx, to_do, true); tx 597 drivers/net/ethernet/google/gve/gve_tx.c repoll |= (nic_done != tx->done); tx 330 drivers/net/ethernet/hisilicon/hix5hd2_gmac.c static void hix5hd2_set_desc_depth(struct hix5hd2_priv *priv, int rx, int tx) tx 341 drivers/net/ethernet/hisilicon/hix5hd2_gmac.c writel_relaxed(tx << 3, priv->base + TX_BQ_DEPTH); tx 345 drivers/net/ethernet/hisilicon/hix5hd2_gmac.c writel_relaxed(tx << 3, priv->base + TX_RQ_DEPTH); tx 194 drivers/net/ethernet/hisilicon/hns/hnae.h } tx; tx 96 drivers/net/ethernet/hisilicon/hns/hns_dsaf_gmac.c static void hns_gmac_get_en(void *mac_drv, u32 *rx, u32 *tx) tx 102 drivers/net/ethernet/hisilicon/hns/hns_dsaf_gmac.c *tx = dsaf_get_bit(porten, GMAC_PORT_TX_EN_B); tx 511 drivers/net/ethernet/hisilicon/hns/hns_dsaf_gmac.c u32 tx; tx 522 drivers/net/ethernet/hisilicon/hns/hns_dsaf_gmac.c hns_gmac_get_en(mac_drv, &rx, &tx); tx 523 drivers/net/ethernet/hisilicon/hns/hns_dsaf_gmac.c mac_info->port_en = rx && tx; tx 60 drivers/net/ethernet/hisilicon/hns/hns_enet.c desc->tx.send_size = cpu_to_le16((u16)send_sz); tx 113 drivers/net/ethernet/hisilicon/hns/hns_enet.c desc->tx.ip_offset = ip_offset; tx 114 drivers/net/ethernet/hisilicon/hns/hns_enet.c desc->tx.tse_vlan_snap_v6_sctp_nth = tvsvsn; tx 115 drivers/net/ethernet/hisilicon/hns/hns_enet.c desc->tx.mss = cpu_to_le16(mss); tx 116 drivers/net/ethernet/hisilicon/hns/hns_enet.c desc->tx.l4_len = l4_len; tx 117 drivers/net/ethernet/hisilicon/hns/hns_enet.c desc->tx.paylen = cpu_to_le16(paylen); tx 123 drivers/net/ethernet/hisilicon/hns/hns_enet.c desc->tx.bn_pid = bn_pid; tx 124 drivers/net/ethernet/hisilicon/hns/hns_enet.c desc->tx.ra_ri_cs_fe_vld = rrcfv; tx 162 drivers/net/ethernet/hisilicon/hns/hns_enet.c desc->tx.send_size = cpu_to_le16((u16)size); tx 199 drivers/net/ethernet/hisilicon/hns/hns_enet.c desc->tx.asid_bufnum_pid = cpu_to_le16(asid_bufnum_pid); tx 200 drivers/net/ethernet/hisilicon/hns/hns_enet.c desc->tx.flag_ipoffset = cpu_to_le32(flag_ipoffset); tx 201 drivers/net/ethernet/hisilicon/hns3/hns3_debugfs.c dev_info(dev, "(TX)vlan_tag: %u\n", tx_desc->tx.vlan_tag); tx 202 drivers/net/ethernet/hisilicon/hns3/hns3_debugfs.c dev_info(dev, "(TX)send_size: %u\n", tx_desc->tx.send_size); tx 203 drivers/net/ethernet/hisilicon/hns3/hns3_debugfs.c dev_info(dev, "(TX)vlan_tso: %u\n", tx_desc->tx.type_cs_vlan_tso); tx 204 drivers/net/ethernet/hisilicon/hns3/hns3_debugfs.c dev_info(dev, "(TX)l2_len: %u\n", tx_desc->tx.l2_len); tx 205 drivers/net/ethernet/hisilicon/hns3/hns3_debugfs.c dev_info(dev, "(TX)l3_len: %u\n", tx_desc->tx.l3_len); tx 206 drivers/net/ethernet/hisilicon/hns3/hns3_debugfs.c dev_info(dev, "(TX)l4_len: %u\n", tx_desc->tx.l4_len); tx 207 drivers/net/ethernet/hisilicon/hns3/hns3_debugfs.c dev_info(dev, "(TX)vlan_tag: %u\n", tx_desc->tx.outer_vlan_tag); tx 208 drivers/net/ethernet/hisilicon/hns3/hns3_debugfs.c dev_info(dev, "(TX)tv: %u\n", tx_desc->tx.tv); tx 209 drivers/net/ethernet/hisilicon/hns3/hns3_debugfs.c dev_info(dev, "(TX)vlan_msec: %u\n", tx_desc->tx.ol_type_vlan_msec); tx 210 drivers/net/ethernet/hisilicon/hns3/hns3_debugfs.c dev_info(dev, "(TX)ol2_len: %u\n", tx_desc->tx.ol2_len); tx 211 drivers/net/ethernet/hisilicon/hns3/hns3_debugfs.c dev_info(dev, "(TX)ol3_len: %u\n", tx_desc->tx.ol3_len); tx 212 drivers/net/ethernet/hisilicon/hns3/hns3_debugfs.c dev_info(dev, "(TX)ol4_len: %u\n", tx_desc->tx.ol4_len); tx 213 drivers/net/ethernet/hisilicon/hns3/hns3_debugfs.c dev_info(dev, "(TX)paylen: %u\n", tx_desc->tx.paylen); tx 214 drivers/net/ethernet/hisilicon/hns3/hns3_debugfs.c dev_info(dev, "(TX)vld_ra_ri: %u\n", tx_desc->tx.bdtp_fe_sc_vld_ra_ri); tx 215 drivers/net/ethernet/hisilicon/hns3/hns3_debugfs.c dev_info(dev, "(TX)mss: %u\n", tx_desc->tx.mss); tx 1095 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c desc->tx.ol_type_vlan_len_msec = tx 1097 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c desc->tx.type_cs_vlan_tso_len = cpu_to_le32(type_cs_vlan_tso); tx 1098 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c desc->tx.paylen = cpu_to_le32(paylen); tx 1099 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c desc->tx.mss = cpu_to_le16(mss); tx 1100 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c desc->tx.vlan_tag = cpu_to_le16(inner_vtag); tx 1101 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c desc->tx.outer_vlan_tag = cpu_to_le16(out_vtag); tx 1148 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c desc->tx.send_size = cpu_to_le16(size); tx 1150 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c desc->tx.bdtp_fe_sc_vld_ra_ri = tx 1173 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c desc->tx.send_size = cpu_to_le16((k == frag_buf_num - 1) ? tx 1178 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c desc->tx.bdtp_fe_sc_vld_ra_ri = tx 4089 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c ring->desc[ring->next_to_clean].tx.bdtp_fe_sc_vld_ra_ri = 0; tx 277 drivers/net/ethernet/hisilicon/hns3/hns3_enet.h } tx; tx 161 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_tm.c int hclge_mac_pause_en_cfg(struct hclge_dev *hdev, bool tx, bool rx) tx 167 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_tm.c desc.data[0] = cpu_to_le32((tx ? HCLGE_TX_MAC_PAUSE_EN_MSK : 0) | tx 153 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_tm.h int hclge_mac_pause_en_cfg(struct hclge_dev *hdev, bool tx, bool rx); tx 673 drivers/net/ethernet/i825xx/ether1.c tx_t tx; tx 701 drivers/net/ethernet/i825xx/ether1.c tx.tx_status = 0; tx 702 drivers/net/ethernet/i825xx/ether1.c tx.tx_command = CMD_TX | CMD_INTR; tx 703 drivers/net/ethernet/i825xx/ether1.c tx.tx_link = nopaddr; tx 704 drivers/net/ethernet/i825xx/ether1.c tx.tx_tbdoffset = tbdaddr; tx 714 drivers/net/ethernet/i825xx/ether1.c ether1_writebuffer (dev, &tx, txaddr, TX_SIZE); tx 189 drivers/net/ethernet/intel/fm10k/fm10k.h struct fm10k_ring_container rx, tx; tx 165 drivers/net/ethernet/intel/fm10k/fm10k_debugfs.c for (i = 0; i < q_vector->tx.count; i++) { tx 166 drivers/net/ethernet/intel/fm10k/fm10k_debugfs.c struct fm10k_ring *ring = &q_vector->tx.ring[i]; tx 680 drivers/net/ethernet/intel/fm10k/fm10k_ethtool.c qv->tx.itr = tx_itr; tx 1196 drivers/net/ethernet/intel/fm10k/fm10k_main.c unsigned int budget = q_vector->tx.work_limit; tx 1284 drivers/net/ethernet/intel/fm10k/fm10k_main.c q_vector->tx.total_bytes += total_bytes; tx 1285 drivers/net/ethernet/intel/fm10k/fm10k_main.c q_vector->tx.total_packets += total_packets; tx 1416 drivers/net/ethernet/intel/fm10k/fm10k_main.c fm10k_update_itr(&q_vector->tx); tx 1422 drivers/net/ethernet/intel/fm10k/fm10k_main.c itr |= (q_vector->tx.itr & FM10K_ITR_MAX); tx 1439 drivers/net/ethernet/intel/fm10k/fm10k_main.c fm10k_for_each_ring(ring, q_vector->tx) { tx 1628 drivers/net/ethernet/intel/fm10k/fm10k_main.c q_vector->tx.ring = ring; tx 1629 drivers/net/ethernet/intel/fm10k/fm10k_main.c q_vector->tx.work_limit = FM10K_DEFAULT_TX_WORK; tx 1630 drivers/net/ethernet/intel/fm10k/fm10k_main.c q_vector->tx.itr = interface->tx_itr; tx 1631 drivers/net/ethernet/intel/fm10k/fm10k_main.c q_vector->tx.itr_scale = interface->hw.mac.itr_scale; tx 1632 drivers/net/ethernet/intel/fm10k/fm10k_main.c q_vector->tx.count = txr_count; tx 1708 drivers/net/ethernet/intel/fm10k/fm10k_main.c fm10k_for_each_ring(ring, q_vector->tx) tx 333 drivers/net/ethernet/intel/fm10k/fm10k_mbx.c struct fm10k_mbx_fifo *fifo = &mbx->tx; tx 386 drivers/net/ethernet/intel/fm10k/fm10k_mbx.c struct fm10k_mbx_fifo *fifo = &mbx->tx; tx 622 drivers/net/ethernet/intel/fm10k/fm10k_mbx.c head = fm10k_fifo_head_offset(&mbx->tx, mbx->pulled); tx 625 drivers/net/ethernet/intel/fm10k/fm10k_mbx.c mbx->local = fm10k_fifo_crc(&mbx->tx, head, len, mbx->local); tx 678 drivers/net/ethernet/intel/fm10k/fm10k_mbx.c u16 fifo_unused = fm10k_fifo_unused(&mbx->tx); tx 691 drivers/net/ethernet/intel/fm10k/fm10k_mbx.c return fm10k_fifo_empty(&mbx->tx); tx 754 drivers/net/ethernet/intel/fm10k/fm10k_mbx.c err = fm10k_fifo_enqueue(&mbx->tx, msg); tx 761 drivers/net/ethernet/intel/fm10k/fm10k_mbx.c err = fm10k_fifo_enqueue(&mbx->tx, msg); tx 856 drivers/net/ethernet/intel/fm10k/fm10k_mbx.c struct fm10k_mbx_fifo *fifo = &mbx->tx; tx 1070 drivers/net/ethernet/intel/fm10k/fm10k_mbx.c while (fm10k_fifo_head_len(&mbx->tx) && mbx->pulled) { tx 1071 drivers/net/ethernet/intel/fm10k/fm10k_mbx.c len = fm10k_fifo_head_drop(&mbx->tx); tx 1106 drivers/net/ethernet/intel/fm10k/fm10k_mbx.c for (len = fm10k_fifo_head_len(&mbx->tx); tx 1108 drivers/net/ethernet/intel/fm10k/fm10k_mbx.c len = fm10k_fifo_head_len(&mbx->tx)) { tx 1109 drivers/net/ethernet/intel/fm10k/fm10k_mbx.c fm10k_fifo_head_drop(&mbx->tx); tx 1417 drivers/net/ethernet/intel/fm10k/fm10k_mbx.c fm10k_fifo_drop_all(&mbx->tx); tx 1607 drivers/net/ethernet/intel/fm10k/fm10k_mbx.c fm10k_fifo_init(&mbx->tx, mbx->buffer, FM10K_MBX_TX_BUFFER_SIZE); tx 1758 drivers/net/ethernet/intel/fm10k/fm10k_mbx.c fm10k_fifo_drop_all(&mbx->tx); tx 1920 drivers/net/ethernet/intel/fm10k/fm10k_mbx.c struct fm10k_mbx_fifo *fifo = &mbx->tx; tx 2165 drivers/net/ethernet/intel/fm10k/fm10k_mbx.c fm10k_fifo_init(&mbx->tx, mbx->buffer, FM10K_MBX_TX_BUFFER_SIZE); tx 257 drivers/net/ethernet/intel/fm10k/fm10k_mbx.h struct fm10k_mbx_fifo tx; tx 722 drivers/net/ethernet/intel/fm10k/fm10k_pci.c if (!qv->tx.count && !qv->rx.count) tx 1187 drivers/net/ethernet/intel/fm10k/fm10k_pci.c if (q_vector->rx.count || q_vector->tx.count) tx 1734 drivers/net/ethernet/intel/fm10k/fm10k_pci.c if (!q_vector->tx.count && !q_vector->rx.count) tx 1768 drivers/net/ethernet/intel/fm10k/fm10k_pci.c if (q_vector->tx.count && q_vector->rx.count) { tx 1775 drivers/net/ethernet/intel/fm10k/fm10k_pci.c } else if (q_vector->tx.count) { tx 1818 drivers/net/ethernet/intel/fm10k/fm10k_pci.c if (!q_vector->tx.count && !q_vector->rx.count) tx 864 drivers/net/ethernet/intel/i40e/i40e.h struct i40e_ring_container tx; tx 2839 drivers/net/ethernet/intel/i40e/i40e_ethtool.c q_vector->tx.target_itr = ITR_TO_REG(tx_ring->itr_setting); tx 3540 drivers/net/ethernet/intel/i40e/i40e_main.c q_vector->tx.next_update = jiffies + 1; tx 3541 drivers/net/ethernet/intel/i40e/i40e_main.c q_vector->tx.target_itr = tx 3544 drivers/net/ethernet/intel/i40e/i40e_main.c q_vector->tx.target_itr >> 1); tx 3545 drivers/net/ethernet/intel/i40e/i40e_main.c q_vector->tx.current_itr = q_vector->tx.target_itr; tx 3651 drivers/net/ethernet/intel/i40e/i40e_main.c q_vector->tx.next_update = jiffies + 1; tx 3652 drivers/net/ethernet/intel/i40e/i40e_main.c q_vector->tx.target_itr = ITR_TO_REG(vsi->tx_rings[0]->itr_setting); tx 3653 drivers/net/ethernet/intel/i40e/i40e_main.c wr32(hw, I40E_PFINT_ITR0(I40E_TX_ITR), q_vector->tx.target_itr >> 1); tx 3654 drivers/net/ethernet/intel/i40e/i40e_main.c q_vector->tx.current_itr = q_vector->tx.target_itr; tx 3725 drivers/net/ethernet/intel/i40e/i40e_main.c if (!q_vector->tx.ring && !q_vector->rx.ring) tx 3783 drivers/net/ethernet/intel/i40e/i40e_main.c if (q_vector->tx.ring && q_vector->rx.ring) { tx 3790 drivers/net/ethernet/intel/i40e/i40e_main.c } else if (q_vector->tx.ring) { tx 4146 drivers/net/ethernet/intel/i40e/i40e_main.c if (!q_vector->tx.ring) tx 4149 drivers/net/ethernet/intel/i40e/i40e_main.c vsi = q_vector->tx.ring->vsi; tx 4150 drivers/net/ethernet/intel/i40e/i40e_main.c i40e_clean_fdir_tx_irq(q_vector->tx.ring, vsi->work_limit); tx 4168 drivers/net/ethernet/intel/i40e/i40e_main.c tx_ring->next = q_vector->tx.ring; tx 4169 drivers/net/ethernet/intel/i40e/i40e_main.c q_vector->tx.ring = tx_ring; tx 4170 drivers/net/ethernet/intel/i40e/i40e_main.c q_vector->tx.count++; tx 4177 drivers/net/ethernet/intel/i40e/i40e_main.c xdp_ring->next = q_vector->tx.ring; tx 4178 drivers/net/ethernet/intel/i40e/i40e_main.c q_vector->tx.ring = xdp_ring; tx 4179 drivers/net/ethernet/intel/i40e/i40e_main.c q_vector->tx.count++; tx 4221 drivers/net/ethernet/intel/i40e/i40e_main.c q_vector->tx.count = 0; tx 4223 drivers/net/ethernet/intel/i40e/i40e_main.c q_vector->tx.ring = NULL; tx 4733 drivers/net/ethernet/intel/i40e/i40e_main.c i40e_for_each_ring(ring, q_vector->tx) tx 4819 drivers/net/ethernet/intel/i40e/i40e_main.c if (q_vector->rx.ring || q_vector->tx.ring) tx 4838 drivers/net/ethernet/intel/i40e/i40e_main.c if (q_vector->rx.ring || q_vector->tx.ring) tx 12638 drivers/net/ethernet/intel/i40e/i40e_main.c if (q_vector->rx.ring || q_vector->tx.ring) { tx 908 drivers/net/ethernet/intel/i40e/i40e_txrx.c u16 flags = q_vector->tx.ring[0].flags; tx 1053 drivers/net/ethernet/intel/i40e/i40e_txrx.c (q_vector->tx.target_itr & I40E_ITR_ADAPTIVE_LATENCY)) { tx 1092 drivers/net/ethernet/intel/i40e/i40e_txrx.c itr = min(q_vector->tx.current_itr, q_vector->rx.current_itr); tx 2515 drivers/net/ethernet/intel/i40e/i40e_txrx.c i40e_update_itr(q_vector, &q_vector->tx); tx 2532 drivers/net/ethernet/intel/i40e/i40e_txrx.c } else if ((q_vector->tx.target_itr < q_vector->tx.current_itr) || tx 2534 drivers/net/ethernet/intel/i40e/i40e_txrx.c (q_vector->tx.target_itr - q_vector->tx.current_itr))) { tx 2539 drivers/net/ethernet/intel/i40e/i40e_txrx.c q_vector->tx.target_itr); tx 2540 drivers/net/ethernet/intel/i40e/i40e_txrx.c q_vector->tx.current_itr = q_vector->tx.target_itr; tx 2587 drivers/net/ethernet/intel/i40e/i40e_txrx.c i40e_for_each_ring(ring, q_vector->tx) { tx 2643 drivers/net/ethernet/intel/i40e/i40e_txrx.c q_vector->tx.ring[0].tx_stats.tx_force_wb++; tx 55 drivers/net/ethernet/intel/i40e/i40e_txrx_common.h tx_ring->q_vector->tx.total_bytes += total_bytes; tx 56 drivers/net/ethernet/intel/i40e/i40e_txrx_common.h tx_ring->q_vector->tx.total_packets += total_packets; tx 98 drivers/net/ethernet/intel/iavf/iavf.h struct iavf_ring_container tx; tx 734 drivers/net/ethernet/intel/iavf/iavf_ethtool.c q_vector->tx.target_itr = ITR_TO_REG(tx_ring->itr_setting); tx 287 drivers/net/ethernet/intel/iavf/iavf_main.c if (!q_vector->tx.ring && !q_vector->rx.ring) tx 335 drivers/net/ethernet/intel/iavf/iavf_main.c tx_ring->next = q_vector->tx.ring; tx 337 drivers/net/ethernet/intel/iavf/iavf_main.c q_vector->tx.ring = tx_ring; tx 338 drivers/net/ethernet/intel/iavf/iavf_main.c q_vector->tx.count++; tx 339 drivers/net/ethernet/intel/iavf/iavf_main.c q_vector->tx.next_update = jiffies + 1; tx 340 drivers/net/ethernet/intel/iavf/iavf_main.c q_vector->tx.target_itr = ITR_TO_REG(tx_ring->itr_setting); tx 343 drivers/net/ethernet/intel/iavf/iavf_main.c q_vector->tx.target_itr >> 1); tx 344 drivers/net/ethernet/intel/iavf/iavf_main.c q_vector->tx.current_itr = q_vector->tx.target_itr; tx 431 drivers/net/ethernet/intel/iavf/iavf_main.c if (q_vector->tx.ring && q_vector->rx.ring) { tx 438 drivers/net/ethernet/intel/iavf/iavf_main.c } else if (q_vector->tx.ring) { tx 285 drivers/net/ethernet/intel/iavf/iavf_txrx.c tx_ring->q_vector->tx.total_bytes += total_bytes; tx 286 drivers/net/ethernet/intel/iavf/iavf_txrx.c tx_ring->q_vector->tx.total_packets += total_packets; tx 335 drivers/net/ethernet/intel/iavf/iavf_txrx.c u16 flags = q_vector->tx.ring[0].flags; tx 463 drivers/net/ethernet/intel/iavf/iavf_txrx.c (q_vector->tx.target_itr & IAVF_ITR_ADAPTIVE_LATENCY)) { tx 502 drivers/net/ethernet/intel/iavf/iavf_txrx.c itr = min(q_vector->tx.current_itr, q_vector->rx.current_itr); tx 1650 drivers/net/ethernet/intel/iavf/iavf_txrx.c iavf_update_itr(q_vector, &q_vector->tx); tx 1667 drivers/net/ethernet/intel/iavf/iavf_txrx.c } else if ((q_vector->tx.target_itr < q_vector->tx.current_itr) || tx 1669 drivers/net/ethernet/intel/iavf/iavf_txrx.c (q_vector->tx.target_itr - q_vector->tx.current_itr))) { tx 1674 drivers/net/ethernet/intel/iavf/iavf_txrx.c q_vector->tx.target_itr); tx 1675 drivers/net/ethernet/intel/iavf/iavf_txrx.c q_vector->tx.current_itr = q_vector->tx.target_itr; tx 1722 drivers/net/ethernet/intel/iavf/iavf_txrx.c iavf_for_each_ring(ring, q_vector->tx) { tx 1772 drivers/net/ethernet/intel/iavf/iavf_txrx.c q_vector->tx.ring[0].tx_stats.tx_force_wb++; tx 296 drivers/net/ethernet/intel/ice/ice.h struct ice_ring_container tx; tx 3152 drivers/net/ethernet/intel/ice/ice_ethtool.c &vsi->tx_rings[q_num]->q_vector->tx)) tx 3160 drivers/net/ethernet/intel/ice/ice_ethtool.c &vsi->tx_rings[q_num]->q_vector->tx)) tx 3327 drivers/net/ethernet/intel/ice/ice_ethtool.c &vsi->tx_rings[q_num]->q_vector->tx, tx 3337 drivers/net/ethernet/intel/ice/ice_ethtool.c &vsi->tx_rings[q_num]->q_vector->tx, tx 523 drivers/net/ethernet/intel/ice/ice_lib.c if (!q_vector->tx.ring && !q_vector->rx.ring) tx 1118 drivers/net/ethernet/intel/ice/ice_lib.c ice_for_each_ring(ring, q_vector->tx) tx 1372 drivers/net/ethernet/intel/ice/ice_lib.c q_vector->tx.ring = NULL; tx 1373 drivers/net/ethernet/intel/ice/ice_lib.c q_vector->tx.itr_idx = ICE_TX_ITR; tx 1380 drivers/net/ethernet/intel/ice/ice_lib.c tx_ring->next = q_vector->tx.ring; tx 1381 drivers/net/ethernet/intel/ice/ice_lib.c q_vector->tx.ring = tx_ring; tx 1901 drivers/net/ethernet/intel/ice/ice_lib.c struct ice_ring_container *rc = &q_vector->tx; tx 2013 drivers/net/ethernet/intel/ice/ice_lib.c q_vector->tx.itr_idx); tx 2457 drivers/net/ethernet/intel/ice/ice_lib.c void ice_cfg_sw_lldp(struct ice_vsi *vsi, bool tx, bool create) tx 2472 drivers/net/ethernet/intel/ice/ice_lib.c if (tx) { tx 2493 drivers/net/ethernet/intel/ice/ice_lib.c create ? "adding" : "removing", tx ? "TX" : "RX", tx 72 drivers/net/ethernet/intel/ice/ice_lib.h void ice_cfg_sw_lldp(struct ice_vsi *vsi, bool tx, bool create); tx 1617 drivers/net/ethernet/intel/ice/ice_main.c if (q_vector->tx.ring && q_vector->rx.ring) { tx 1624 drivers/net/ethernet/intel/ice/ice_main.c } else if (q_vector->tx.ring) { tx 3528 drivers/net/ethernet/intel/ice/ice_main.c if (q_vector->rx.ring || q_vector->tx.ring) tx 3895 drivers/net/ethernet/intel/ice/ice_main.c if (q_vector->rx.ring || q_vector->tx.ring) tx 195 drivers/net/ethernet/intel/ice/ice_txrx.c tx_ring->q_vector->tx.total_bytes += total_bytes; tx 196 drivers/net/ethernet/intel/ice/ice_txrx.c tx_ring->q_vector->tx.total_pkts += total_pkts; tx 1241 drivers/net/ethernet/intel/ice/ice_txrx.c (q_vector->tx.target_itr & ICE_ITR_ADAPTIVE_LATENCY)) { tx 1280 drivers/net/ethernet/intel/ice/ice_txrx.c itr = min(q_vector->tx.current_itr, q_vector->rx.current_itr); tx 1364 drivers/net/ethernet/intel/ice/ice_txrx.c struct ice_ring_container *tx = &q_vector->tx; tx 1387 drivers/net/ethernet/intel/ice/ice_txrx.c ice_update_itr(q_vector, tx); tx 1403 drivers/net/ethernet/intel/ice/ice_txrx.c } else if ((tx->target_itr < tx->current_itr) || tx 1405 drivers/net/ethernet/intel/ice/ice_txrx.c (tx->target_itr - tx->current_itr))) { tx 1409 drivers/net/ethernet/intel/ice/ice_txrx.c itr_val = ice_buildreg_itr(tx->itr_idx, tx->target_itr); tx 1410 drivers/net/ethernet/intel/ice/ice_txrx.c tx->current_itr = tx->target_itr; tx 1486 drivers/net/ethernet/intel/ice/ice_txrx.c ice_for_each_ring(ring, q_vector->tx) tx 2209 drivers/net/ethernet/intel/ice/ice_virtchnl_pf.c q_vector->tx.itr_idx = map->txitr_idx; tx 2212 drivers/net/ethernet/intel/ice/ice_virtchnl_pf.c q_vector->tx.itr_idx); tx 302 drivers/net/ethernet/intel/igb/igb.h struct igb_ring_container rx, tx; tx 2236 drivers/net/ethernet/intel/igb/igb_ethtool.c q_vector->tx.work_limit = adapter->tx_work_limit; tx 800 drivers/net/ethernet/intel/igb/igb_main.c if (q_vector->tx.ring) tx 801 drivers/net/ethernet/intel/igb/igb_main.c tx_queue = q_vector->tx.ring->reg_idx; tx 958 drivers/net/ethernet/intel/igb/igb_main.c if (q_vector->rx.ring && q_vector->tx.ring) tx 961 drivers/net/ethernet/intel/igb/igb_main.c else if (q_vector->tx.ring) tx 963 drivers/net/ethernet/intel/igb/igb_main.c q_vector->tx.ring->queue_index); tx 1031 drivers/net/ethernet/intel/igb/igb_main.c if (q_vector->tx.ring) tx 1032 drivers/net/ethernet/intel/igb/igb_main.c adapter->tx_ring[q_vector->tx.ring->queue_index] = NULL; tx 1226 drivers/net/ethernet/intel/igb/igb_main.c q_vector->tx.work_limit = adapter->tx_work_limit; tx 1255 drivers/net/ethernet/intel/igb/igb_main.c igb_add_ring(ring, &q_vector->tx); tx 5482 drivers/net/ethernet/intel/igb/igb_main.c packets = q_vector->tx.total_packets; tx 5485 drivers/net/ethernet/intel/igb/igb_main.c q_vector->tx.total_bytes / packets); tx 5517 drivers/net/ethernet/intel/igb/igb_main.c q_vector->tx.total_bytes = 0; tx 5518 drivers/net/ethernet/intel/igb/igb_main.c q_vector->tx.total_packets = 0; tx 5602 drivers/net/ethernet/intel/igb/igb_main.c igb_update_itr(q_vector, &q_vector->tx); tx 5605 drivers/net/ethernet/intel/igb/igb_main.c current_itr = max(q_vector->rx.itr, q_vector->tx.itr); tx 6642 drivers/net/ethernet/intel/igb/igb_main.c if (q_vector->tx.ring) tx 6643 drivers/net/ethernet/intel/igb/igb_main.c igb_update_tx_dca(adapter, q_vector->tx.ring, cpu); tx 7714 drivers/net/ethernet/intel/igb/igb_main.c if (q_vector->tx.ring) tx 7748 drivers/net/ethernet/intel/igb/igb_main.c struct igb_ring *tx_ring = q_vector->tx.ring; tx 7752 drivers/net/ethernet/intel/igb/igb_main.c unsigned int budget = q_vector->tx.work_limit; tx 7841 drivers/net/ethernet/intel/igb/igb_main.c q_vector->tx.total_bytes += total_bytes; tx 7842 drivers/net/ethernet/intel/igb/igb_main.c q_vector->tx.total_packets += total_packets; tx 292 drivers/net/ethernet/intel/igc/igc.h struct igc_ring_container rx, tx; tx 864 drivers/net/ethernet/intel/igc/igc_ethtool.c q_vector->tx.work_limit = adapter->tx_work_limit; tx 1705 drivers/net/ethernet/intel/igc/igc_main.c unsigned int budget = q_vector->tx.work_limit; tx 1706 drivers/net/ethernet/intel/igc/igc_main.c struct igc_ring *tx_ring = q_vector->tx.ring; tx 1798 drivers/net/ethernet/intel/igc/igc_main.c q_vector->tx.total_bytes += total_bytes; tx 1799 drivers/net/ethernet/intel/igc/igc_main.c q_vector->tx.total_packets += total_packets; tx 2600 drivers/net/ethernet/intel/igc/igc_main.c if (q_vector->tx.ring) tx 2601 drivers/net/ethernet/intel/igc/igc_main.c tx_queue = q_vector->tx.ring->reg_idx; tx 2707 drivers/net/ethernet/intel/igc/igc_main.c if (q_vector->rx.ring && q_vector->tx.ring) tx 2710 drivers/net/ethernet/intel/igc/igc_main.c else if (q_vector->tx.ring) tx 2712 drivers/net/ethernet/intel/igc/igc_main.c q_vector->tx.ring->queue_index); tx 2760 drivers/net/ethernet/intel/igc/igc_main.c if (q_vector->tx.ring) tx 2761 drivers/net/ethernet/intel/igc/igc_main.c adapter->tx_ring[q_vector->tx.ring->queue_index] = NULL; tx 3112 drivers/net/ethernet/intel/igc/igc_main.c packets = q_vector->tx.total_packets; tx 3115 drivers/net/ethernet/intel/igc/igc_main.c q_vector->tx.total_bytes / packets); tx 3147 drivers/net/ethernet/intel/igc/igc_main.c q_vector->tx.total_bytes = 0; tx 3148 drivers/net/ethernet/intel/igc/igc_main.c q_vector->tx.total_packets = 0; tx 3312 drivers/net/ethernet/intel/igc/igc_main.c igc_update_itr(q_vector, &q_vector->tx); tx 3315 drivers/net/ethernet/intel/igc/igc_main.c current_itr = max(q_vector->rx.itr, q_vector->tx.itr); tx 3393 drivers/net/ethernet/intel/igc/igc_main.c if (q_vector->tx.ring) tx 3539 drivers/net/ethernet/intel/igc/igc_main.c q_vector->tx.work_limit = adapter->tx_work_limit; tx 3568 drivers/net/ethernet/intel/igc/igc_main.c igc_add_ring(ring, &q_vector->tx); tx 456 drivers/net/ethernet/intel/ixgbe/ixgbe.h struct ixgbe_ring_container rx, tx; tx 30 drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c int tx = DCB_TX_CONFIG; tx 48 drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c if (dst->path[tx].prio_type != src->path[tx].prio_type) { tx 49 drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c dst->path[tx].prio_type = src->path[tx].prio_type; tx 53 drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c if (dst->path[tx].bwg_id != src->path[tx].bwg_id) { tx 54 drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c dst->path[tx].bwg_id = src->path[tx].bwg_id; tx 58 drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c if (dst->path[tx].bwg_percent != src->path[tx].bwg_percent) { tx 59 drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c dst->path[tx].bwg_percent = src->path[tx].bwg_percent; tx 63 drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c if (dst->path[tx].up_to_tc_bitmap != tx 64 drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c src->path[tx].up_to_tc_bitmap) { tx 65 drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c dst->path[tx].up_to_tc_bitmap = tx 66 drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c src->path[tx].up_to_tc_bitmap; tx 95 drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c if (dcfg->bw_percentage[tx][j] != scfg->bw_percentage[tx][j]) { tx 96 drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c dcfg->bw_percentage[tx][j] = scfg->bw_percentage[tx][j]; tx 2285 drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c if (adapter->q_vector[0]->tx.count && adapter->q_vector[0]->rx.count) tx 2336 drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c if (adapter->q_vector[0]->tx.count && adapter->q_vector[0]->rx.count) { tx 2370 drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c if (adapter->q_vector[0]->tx.count && adapter->q_vector[0]->rx.count) tx 2390 drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c if (q_vector->tx.count && !q_vector->rx.count) tx 91 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c unsigned int *tx, unsigned int *rx) tx 96 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c *tx = 0; tx 102 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c *tx = tc << 2; /* 0, 4, 8, 12, 16, 20, 24, 28 */ tx 118 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c *tx = tc << 5; /* 0, 32, 64 */ tx 120 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c *tx = (tc + 2) << 4; /* 80, 96 */ tx 122 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c *tx = (tc + 8) << 3; /* 104, 112, 120 */ tx 131 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c *tx = tc << 6; /* 0, 64 */ tx 133 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c *tx = (tc + 4) << 4; /* 96, 112 */ tx 884 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c q_vector->tx.work_limit = adapter->tx_work_limit; tx 887 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c q_vector->tx.itr = IXGBE_ITR_ADAPTIVE_MAX_USECS | tx 919 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c ixgbe_add_ring(ring, &q_vector->tx); tx 945 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c ixgbe_add_ring(ring, &q_vector->tx); tx 1023 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c ixgbe_for_each_ring(ring, q_vector->tx) { tx 1119 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c unsigned int budget = q_vector->tx.work_limit; tx 1211 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c q_vector->tx.total_bytes += total_bytes; tx 1212 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c q_vector->tx.total_packets += total_packets; tx 1352 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c ixgbe_for_each_ring(ring, q_vector->tx) tx 2459 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c ixgbe_for_each_ring(ring, q_vector->tx) tx 2720 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c ixgbe_update_itr(q_vector, &q_vector->tx); tx 2724 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c new_itr = min(q_vector->rx.itr, q_vector->tx.itr); tx 3143 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c if (q_vector->rx.ring || q_vector->tx.ring) tx 3170 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c ixgbe_for_each_ring(ring, q_vector->tx) { tx 3235 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c if (q_vector->tx.ring && q_vector->rx.ring) { tx 3242 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c } else if (q_vector->tx.ring) { tx 3406 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c if (!q_vector->rx.ring && !q_vector->tx.ring) tx 7347 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c if (qv->rx.ring || qv->tx.ring) tx 686 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c q_vector->tx.total_bytes += total_bytes; tx 687 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c q_vector->tx.total_packets += total_packets; tx 695 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c return ixgbe_xmit_zc(tx_ring, q_vector->tx.work_limit); tx 803 drivers/net/ethernet/intel/ixgbevf/ethtool.c if (adapter->q_vector[0]->tx.count && adapter->q_vector[0]->rx.count) tx 824 drivers/net/ethernet/intel/ixgbevf/ethtool.c if (adapter->q_vector[0]->tx.count && tx 856 drivers/net/ethernet/intel/ixgbevf/ethtool.c if (q_vector->tx.count && !q_vector->rx.count) tx 239 drivers/net/ethernet/intel/ixgbevf/ixgbevf.h struct ixgbevf_ring_container rx, tx; tx 365 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c q_vector->tx.total_bytes += total_bytes; tx 366 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c q_vector->tx.total_packets += total_packets; tx 1271 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c ixgbevf_for_each_ring(ring, q_vector->tx) { tx 1359 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c ixgbevf_for_each_ring(ring, q_vector->tx) tx 1362 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c if (q_vector->tx.ring && !q_vector->rx.ring) { tx 1462 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c ixgbevf_update_itr(q_vector, &q_vector->tx); tx 1465 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c current_itr = max(q_vector->rx.itr, q_vector->tx.itr); tx 1518 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c if (q_vector->rx.ring || q_vector->tx.ring) tx 1542 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c if (q_vector->tx.ring && q_vector->rx.ring) { tx 1549 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c } else if (q_vector->tx.ring) { tx 1629 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c !adapter->q_vector[i]->tx.ring) tx 2732 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c ixgbevf_add_ring(ring, &q_vector->tx); tx 2760 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c ixgbevf_add_ring(ring, &q_vector->tx); tx 2824 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c ixgbevf_for_each_ring(ring, q_vector->tx) { tx 3207 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c if (qv->rx.ring || qv->tx.ring) tx 558 drivers/net/ethernet/marvell/octeontx2/af/mbox.h } tx; tx 2226 drivers/net/ethernet/marvell/sky2.c static enum flow_control sky2_flow(int rx, int tx) tx 2229 drivers/net/ethernet/marvell/sky2.c return tx ? FC_BOTH : FC_RX; tx 2231 drivers/net/ethernet/marvell/sky2.c return tx ? FC_TX : FC_NONE; tx 72 drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_rxtx.c struct mlx5e_ipsec_tx_metadata tx; tx 223 drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_rxtx.c mdata->content.tx.mss_inv = mlx5e_ipsec_mss_inv(skb); tx 224 drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_rxtx.c mdata->content.tx.seq = htons(ntohl(tcph->seq) & 0xFFFF); tx 228 drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_rxtx.c mdata->content.tx.esp_next_proto = xo->proto; tx 231 drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_rxtx.c mdata->syndrome, mdata->content.tx.esp_next_proto, tx 232 drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_rxtx.c ntohs(mdata->content.tx.mss_inv), tx 233 drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_rxtx.c ntohs(mdata->content.tx.seq)); tx 421 drivers/net/ethernet/mellanox/mlx5/core/fpga/tls.c if (MLX5_GET(tls_extended_cap, buf, tx)) tx 686 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_action.c action->dest_tbl.tbl->tx.s_anchor->chunk->icm_addr; tx 245 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_domain.c dmn->info.tx.ste_type = MLX5DR_STE_TYPE_TX; tx 246 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_domain.c dmn->info.tx.default_icm_addr = dmn->info.caps.nic_tx_allow_address; tx 247 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_domain.c dmn->info.tx.drop_icm_addr = dmn->info.caps.nic_tx_drop_address; tx 257 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_domain.c dmn->info.tx.ste_type = MLX5DR_STE_TYPE_TX; tx 265 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_domain.c dmn->info.tx.default_icm_addr = vport_cap->icm_address_tx; tx 268 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_domain.c dmn->info.tx.drop_icm_addr = dmn->info.caps.esw_tx_drop_address; tx 477 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c ret = dr_matcher_connect(dmn, &matcher->tx, tx 478 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c next_matcher ? &next_matcher->tx : NULL, tx 479 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c prev_matcher ? &prev_matcher->tx : NULL); tx 504 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c dr_matcher_uninit_nic(&matcher->tx); tx 516 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c dr_matcher_uninit_nic(&matcher->tx); tx 581 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c ret = dr_matcher_init_nic(matcher, &matcher->tx); tx 619 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c matcher->tx.nic_tbl = &tbl->tx; tx 620 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c ret = dr_matcher_init_nic(matcher, &matcher->tx); tx 624 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c matcher->tx.nic_tbl = &tbl->tx; tx 738 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c ret = dr_matcher_disconnect(dmn, &tbl->tx, tx 739 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c next_matcher ? &next_matcher->tx : NULL, tx 740 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c prev_matcher ? &prev_matcher->tx : NULL); tx 948 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_rule.c dr_rule_destroy_rule_nic(rule, &rule->tx); tx 961 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_rule.c dr_rule_destroy_rule_nic(rule, &rule->tx); tx 1155 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_rule.c ret = dr_rule_create_rule_nic(rule, &rule->tx, ©_param, tx 1199 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_rule.c rule->tx.nic_matcher = &matcher->tx; tx 1200 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_rule.c ret = dr_rule_create_rule_nic(rule, &rule->tx, ¶m, tx 1205 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_rule.c rule->tx.nic_matcher = &matcher->tx; tx 51 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_table.c last_htbl = last_matcher->tx.e_anchor; tx 53 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_table.c last_htbl = tbl->tx.s_anchor; tx 55 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_table.c tbl->tx.default_icm_addr = action ? tx 56 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_table.c action->dest_tbl.tbl->tx.s_anchor->chunk->icm_addr : tx 57 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_table.c tbl->tx.nic_dmn->default_icm_addr; tx 60 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_table.c info.miss_icm_addr = tbl->tx.default_icm_addr; tx 63 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_table.c tbl->tx.nic_dmn, tx 93 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_table.c dr_table_uninit_nic(&tbl->tx); tx 105 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_table.c dr_table_uninit_nic(&tbl->tx); tx 159 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_table.c ret = dr_table_init_nic(tbl->dmn, &tbl->tx); tx 186 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_table.c tbl->tx.nic_dmn = &tbl->dmn->info.tx; tx 187 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_table.c ret = dr_table_init_nic(tbl->dmn, &tbl->tx); tx 192 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_table.c tbl->tx.nic_dmn = &tbl->dmn->info.tx; tx 221 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_table.c if (tbl->tx.s_anchor) tx 222 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_table.c icm_addr_tx = tbl->tx.s_anchor->chunk->icm_addr; tx 639 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_types.h struct mlx5dr_domain_rx_tx tx; tx 671 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_types.h struct mlx5dr_table_rx_tx tx; tx 696 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_types.h struct mlx5dr_matcher_rx_tx tx; tx 781 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_types.h struct mlx5dr_rule_rx_tx tx; tx 82 drivers/net/ethernet/mellanox/mlxsw/spectrum.c MLXSW_ITEM32(tx, hdr, version, 0x00, 28, 4); tx 89 drivers/net/ethernet/mellanox/mlxsw/spectrum.c MLXSW_ITEM32(tx, hdr, ctl, 0x00, 26, 2); tx 94 drivers/net/ethernet/mellanox/mlxsw/spectrum.c MLXSW_ITEM32(tx, hdr, proto, 0x00, 21, 3); tx 99 drivers/net/ethernet/mellanox/mlxsw/spectrum.c MLXSW_ITEM32(tx, hdr, rx_is_router, 0x00, 19, 1); tx 105 drivers/net/ethernet/mellanox/mlxsw/spectrum.c MLXSW_ITEM32(tx, hdr, fid_valid, 0x00, 16, 1); tx 110 drivers/net/ethernet/mellanox/mlxsw/spectrum.c MLXSW_ITEM32(tx, hdr, swid, 0x00, 12, 3); tx 116 drivers/net/ethernet/mellanox/mlxsw/spectrum.c MLXSW_ITEM32(tx, hdr, control_tclass, 0x00, 6, 1); tx 121 drivers/net/ethernet/mellanox/mlxsw/spectrum.c MLXSW_ITEM32(tx, hdr, etclass, 0x00, 0, 4); tx 131 drivers/net/ethernet/mellanox/mlxsw/spectrum.c MLXSW_ITEM32(tx, hdr, port_mid, 0x04, 16, 16); tx 138 drivers/net/ethernet/mellanox/mlxsw/spectrum.c MLXSW_ITEM32(tx, hdr, fid, 0x08, 0, 16); tx 144 drivers/net/ethernet/mellanox/mlxsw/spectrum.c MLXSW_ITEM32(tx, hdr, type, 0x0C, 0, 4); tx 58 drivers/net/ethernet/mellanox/mlxsw/switchx2.c MLXSW_ITEM32(tx, hdr, version, 0x00, 28, 4); tx 65 drivers/net/ethernet/mellanox/mlxsw/switchx2.c MLXSW_ITEM32(tx, hdr, ctl, 0x00, 26, 2); tx 70 drivers/net/ethernet/mellanox/mlxsw/switchx2.c MLXSW_ITEM32(tx, hdr, proto, 0x00, 21, 3); tx 77 drivers/net/ethernet/mellanox/mlxsw/switchx2.c MLXSW_ITEM32(tx, hdr, etclass, 0x00, 18, 3); tx 82 drivers/net/ethernet/mellanox/mlxsw/switchx2.c MLXSW_ITEM32(tx, hdr, swid, 0x00, 12, 3); tx 92 drivers/net/ethernet/mellanox/mlxsw/switchx2.c MLXSW_ITEM32(tx, hdr, port_mid, 0x04, 16, 16); tx 97 drivers/net/ethernet/mellanox/mlxsw/switchx2.c MLXSW_ITEM32(tx, hdr, ctclass3, 0x04, 14, 1); tx 103 drivers/net/ethernet/mellanox/mlxsw/switchx2.c MLXSW_ITEM32(tx, hdr, rdq, 0x04, 9, 5); tx 108 drivers/net/ethernet/mellanox/mlxsw/switchx2.c MLXSW_ITEM32(tx, hdr, cpu_sig, 0x04, 0, 9); tx 113 drivers/net/ethernet/mellanox/mlxsw/switchx2.c MLXSW_ITEM32(tx, hdr, sig, 0x0C, 16, 16); tx 118 drivers/net/ethernet/mellanox/mlxsw/switchx2.c MLXSW_ITEM32(tx, hdr, stclass, 0x0C, 13, 3); tx 123 drivers/net/ethernet/mellanox/mlxsw/switchx2.c MLXSW_ITEM32(tx, hdr, emad, 0x0C, 5, 1); tx 129 drivers/net/ethernet/mellanox/mlxsw/switchx2.c MLXSW_ITEM32(tx, hdr, type, 0x0C, 0, 4); tx 922 drivers/net/ethernet/micrel/ksz884x.c struct ksz_desc_tx_stat tx; tx 928 drivers/net/ethernet/micrel/ksz884x.c struct ksz_desc_tx_buf tx; tx 1598 drivers/net/ethernet/micrel/ksz884x.c desc->sw.ctrl.tx.hw_owned = 1; tx 1642 drivers/net/ethernet/micrel/ksz884x.c desc->sw.buf.tx.buf_size = len; tx 3204 drivers/net/ethernet/micrel/ksz884x.c static void set_flow_ctrl(struct ksz_hw *hw, int rx, int tx) tx 3215 drivers/net/ethernet/micrel/ksz884x.c if (tx) tx 3231 drivers/net/ethernet/micrel/ksz884x.c int tx; tx 3236 drivers/net/ethernet/micrel/ksz884x.c rx = tx = 0; tx 3238 drivers/net/ethernet/micrel/ksz884x.c rx = tx = 1; tx 3241 drivers/net/ethernet/micrel/ksz884x.c rx = tx = 1; tx 3245 drivers/net/ethernet/micrel/ksz884x.c tx = 1; tx 3252 drivers/net/ethernet/micrel/ksz884x.c set_flow_ctrl(hw, rx, tx); tx 3978 drivers/net/ethernet/micrel/ksz884x.c hw->tx_desc_info.cur->sw.buf.tx.first_seg = 1; tx 4004 drivers/net/ethernet/micrel/ksz884x.c cur->sw.buf.tx.last_seg = 1; tx 4008 drivers/net/ethernet/micrel/ksz884x.c cur->sw.buf.tx.intr = 1; tx 4014 drivers/net/ethernet/micrel/ksz884x.c cur->sw.buf.tx.dest_port = hw->dst_ports; tx 4711 drivers/net/ethernet/micrel/ksz884x.c (desc)->sw.buf.tx.csum_gen_tcp = 1; tx 4712 drivers/net/ethernet/micrel/ksz884x.c (desc)->sw.buf.tx.csum_gen_udp = 1; tx 4751 drivers/net/ethernet/micrel/ksz884x.c if (status.tx.hw_owned) { tx 90 drivers/net/ethernet/microchip/enc28j60.c struct spi_transfer tx = { tx 104 drivers/net/ethernet/microchip/enc28j60.c spi_message_add_tail(&tx, &msg); tx 155 drivers/net/ethernet/microchip/lan743x_main.c struct lan743x_tx *tx = context; tx 156 drivers/net/ethernet/microchip/lan743x_main.c struct lan743x_adapter *adapter = tx->adapter; tx 163 drivers/net/ethernet/microchip/lan743x_main.c INT_BIT_DMA_TX_(tx->channel_number)); tx 166 drivers/net/ethernet/microchip/lan743x_main.c if (int_sts & INT_BIT_DMA_TX_(tx->channel_number)) { tx 167 drivers/net/ethernet/microchip/lan743x_main.c u32 ioc_bit = DMAC_INT_BIT_TX_IOC_(tx->channel_number); tx 184 drivers/net/ethernet/microchip/lan743x_main.c napi_schedule(&tx->napi); tx 192 drivers/net/ethernet/microchip/lan743x_main.c INT_BIT_DMA_TX_(tx->channel_number)); tx 259 drivers/net/ethernet/microchip/lan743x_main.c lan743x_tx_isr(&adapter->tx[channel], tx 606 drivers/net/ethernet/microchip/lan743x_main.c &adapter->tx[index]); tx 1209 drivers/net/ethernet/microchip/lan743x_main.c static void lan743x_tx_release_desc(struct lan743x_tx *tx, tx 1217 drivers/net/ethernet/microchip/lan743x_main.c descriptor = &tx->ring_cpu_ptr[descriptor_index]; tx 1218 drivers/net/ethernet/microchip/lan743x_main.c buffer_info = &tx->buffer_info[descriptor_index]; tx 1233 drivers/net/ethernet/microchip/lan743x_main.c dma_unmap_page(&tx->adapter->pdev->dev, tx 1238 drivers/net/ethernet/microchip/lan743x_main.c dma_unmap_single(&tx->adapter->pdev->dev, tx 1255 drivers/net/ethernet/microchip/lan743x_main.c lan743x_ptp_unrequest_tx_timestamp(tx->adapter); tx 1260 drivers/net/ethernet/microchip/lan743x_main.c lan743x_ptp_tx_timestamp_skb(tx->adapter, tx 1275 drivers/net/ethernet/microchip/lan743x_main.c static int lan743x_tx_next_index(struct lan743x_tx *tx, int index) tx 1277 drivers/net/ethernet/microchip/lan743x_main.c return ((++index) % tx->ring_size); tx 1280 drivers/net/ethernet/microchip/lan743x_main.c static void lan743x_tx_release_completed_descriptors(struct lan743x_tx *tx) tx 1282 drivers/net/ethernet/microchip/lan743x_main.c while ((*tx->head_cpu_ptr) != (tx->last_head)) { tx 1283 drivers/net/ethernet/microchip/lan743x_main.c lan743x_tx_release_desc(tx, tx->last_head, false); tx 1284 drivers/net/ethernet/microchip/lan743x_main.c tx->last_head = lan743x_tx_next_index(tx, tx->last_head); tx 1288 drivers/net/ethernet/microchip/lan743x_main.c static void lan743x_tx_release_all_descriptors(struct lan743x_tx *tx) tx 1292 drivers/net/ethernet/microchip/lan743x_main.c original_head = tx->last_head; tx 1294 drivers/net/ethernet/microchip/lan743x_main.c lan743x_tx_release_desc(tx, tx->last_head, true); tx 1295 drivers/net/ethernet/microchip/lan743x_main.c tx->last_head = lan743x_tx_next_index(tx, tx->last_head); tx 1296 drivers/net/ethernet/microchip/lan743x_main.c } while (tx->last_head != original_head); tx 1297 drivers/net/ethernet/microchip/lan743x_main.c memset(tx->ring_cpu_ptr, 0, tx 1298 drivers/net/ethernet/microchip/lan743x_main.c sizeof(*tx->ring_cpu_ptr) * (tx->ring_size)); tx 1299 drivers/net/ethernet/microchip/lan743x_main.c memset(tx->buffer_info, 0, tx 1300 drivers/net/ethernet/microchip/lan743x_main.c sizeof(*tx->buffer_info) * (tx->ring_size)); tx 1303 drivers/net/ethernet/microchip/lan743x_main.c static int lan743x_tx_get_desc_cnt(struct lan743x_tx *tx, tx 1316 drivers/net/ethernet/microchip/lan743x_main.c static int lan743x_tx_get_avail_desc(struct lan743x_tx *tx) tx 1318 drivers/net/ethernet/microchip/lan743x_main.c int last_head = tx->last_head; tx 1319 drivers/net/ethernet/microchip/lan743x_main.c int last_tail = tx->last_tail; tx 1322 drivers/net/ethernet/microchip/lan743x_main.c return tx->ring_size - last_tail + last_head - 1; tx 1327 drivers/net/ethernet/microchip/lan743x_main.c void lan743x_tx_set_timestamping_mode(struct lan743x_tx *tx, tx 1332 drivers/net/ethernet/microchip/lan743x_main.c tx->ts_flags |= TX_TS_FLAG_TIMESTAMPING_ENABLED; tx 1334 drivers/net/ethernet/microchip/lan743x_main.c tx->ts_flags &= ~TX_TS_FLAG_TIMESTAMPING_ENABLED; tx 1336 drivers/net/ethernet/microchip/lan743x_main.c tx->ts_flags |= TX_TS_FLAG_ONE_STEP_SYNC; tx 1338 drivers/net/ethernet/microchip/lan743x_main.c tx->ts_flags &= ~TX_TS_FLAG_ONE_STEP_SYNC; tx 1341 drivers/net/ethernet/microchip/lan743x_main.c static int lan743x_tx_frame_start(struct lan743x_tx *tx, tx 1353 drivers/net/ethernet/microchip/lan743x_main.c struct lan743x_adapter *adapter = tx->adapter; tx 1357 drivers/net/ethernet/microchip/lan743x_main.c tx->frame_flags |= TX_FRAME_FLAG_IN_PROGRESS; tx 1358 drivers/net/ethernet/microchip/lan743x_main.c tx->frame_first = tx->last_tail; tx 1359 drivers/net/ethernet/microchip/lan743x_main.c tx->frame_tail = tx->frame_first; tx 1361 drivers/net/ethernet/microchip/lan743x_main.c tx_descriptor = &tx->ring_cpu_ptr[tx->frame_tail]; tx 1362 drivers/net/ethernet/microchip/lan743x_main.c buffer_info = &tx->buffer_info[tx->frame_tail]; tx 1378 drivers/net/ethernet/microchip/lan743x_main.c tx->frame_data0 = (first_buffer_length & tx 1384 drivers/net/ethernet/microchip/lan743x_main.c tx->frame_data0 |= TX_DESC_DATA0_TSE_; tx 1387 drivers/net/ethernet/microchip/lan743x_main.c tx->frame_data0 |= TX_DESC_DATA0_ICE_ | tx 1395 drivers/net/ethernet/microchip/lan743x_main.c static void lan743x_tx_frame_add_lso(struct lan743x_tx *tx, tx 1406 drivers/net/ethernet/microchip/lan743x_main.c tx->frame_data0 |= TX_DESC_DATA0_EXT_; tx 1408 drivers/net/ethernet/microchip/lan743x_main.c tx->frame_data0 |= TX_DESC_DATA0_LS_; tx 1409 drivers/net/ethernet/microchip/lan743x_main.c tx->frame_data0 |= TX_DESC_DATA0_IOC_; tx 1411 drivers/net/ethernet/microchip/lan743x_main.c tx_descriptor = &tx->ring_cpu_ptr[tx->frame_tail]; tx 1412 drivers/net/ethernet/microchip/lan743x_main.c tx_descriptor->data0 = tx->frame_data0; tx 1415 drivers/net/ethernet/microchip/lan743x_main.c tx->frame_tail = lan743x_tx_next_index(tx, tx->frame_tail); tx 1416 drivers/net/ethernet/microchip/lan743x_main.c tx_descriptor = &tx->ring_cpu_ptr[tx->frame_tail]; tx 1417 drivers/net/ethernet/microchip/lan743x_main.c buffer_info = &tx->buffer_info[tx->frame_tail]; tx 1429 drivers/net/ethernet/microchip/lan743x_main.c tx->frame_data0 = (frame_length & TX_DESC_DATA0_EXT_PAY_LENGTH_MASK_) | tx 1436 drivers/net/ethernet/microchip/lan743x_main.c static int lan743x_tx_frame_add_fragment(struct lan743x_tx *tx, tx 1445 drivers/net/ethernet/microchip/lan743x_main.c struct lan743x_adapter *adapter = tx->adapter; tx 1455 drivers/net/ethernet/microchip/lan743x_main.c tx_descriptor = &tx->ring_cpu_ptr[tx->frame_tail]; tx 1456 drivers/net/ethernet/microchip/lan743x_main.c tx_descriptor->data0 = tx->frame_data0; tx 1459 drivers/net/ethernet/microchip/lan743x_main.c tx->frame_tail = lan743x_tx_next_index(tx, tx->frame_tail); tx 1460 drivers/net/ethernet/microchip/lan743x_main.c tx_descriptor = &tx->ring_cpu_ptr[tx->frame_tail]; tx 1461 drivers/net/ethernet/microchip/lan743x_main.c buffer_info = &tx->buffer_info[tx->frame_tail]; tx 1469 drivers/net/ethernet/microchip/lan743x_main.c desc_index = tx->frame_first; tx 1470 drivers/net/ethernet/microchip/lan743x_main.c while (desc_index != tx->frame_tail) { tx 1471 drivers/net/ethernet/microchip/lan743x_main.c lan743x_tx_release_desc(tx, desc_index, true); tx 1472 drivers/net/ethernet/microchip/lan743x_main.c desc_index = lan743x_tx_next_index(tx, desc_index); tx 1475 drivers/net/ethernet/microchip/lan743x_main.c tx->frame_flags &= ~TX_FRAME_FLAG_IN_PROGRESS; tx 1476 drivers/net/ethernet/microchip/lan743x_main.c tx->frame_first = 0; tx 1477 drivers/net/ethernet/microchip/lan743x_main.c tx->frame_data0 = 0; tx 1478 drivers/net/ethernet/microchip/lan743x_main.c tx->frame_tail = 0; tx 1493 drivers/net/ethernet/microchip/lan743x_main.c tx->frame_data0 = (fragment_length & TX_DESC_DATA0_BUF_LENGTH_MASK_) | tx 1501 drivers/net/ethernet/microchip/lan743x_main.c static void lan743x_tx_frame_end(struct lan743x_tx *tx, tx 1511 drivers/net/ethernet/microchip/lan743x_main.c struct lan743x_adapter *adapter = tx->adapter; tx 1515 drivers/net/ethernet/microchip/lan743x_main.c if ((tx->frame_data0 & TX_DESC_DATA0_DTYPE_MASK_) == tx 1517 drivers/net/ethernet/microchip/lan743x_main.c tx->frame_data0 |= TX_DESC_DATA0_LS_; tx 1518 drivers/net/ethernet/microchip/lan743x_main.c tx->frame_data0 |= TX_DESC_DATA0_IOC_; tx 1521 drivers/net/ethernet/microchip/lan743x_main.c tx_descriptor = &tx->ring_cpu_ptr[tx->frame_tail]; tx 1522 drivers/net/ethernet/microchip/lan743x_main.c buffer_info = &tx->buffer_info[tx->frame_tail]; tx 1529 drivers/net/ethernet/microchip/lan743x_main.c tx_descriptor->data0 = tx->frame_data0; tx 1530 drivers/net/ethernet/microchip/lan743x_main.c tx->frame_tail = lan743x_tx_next_index(tx, tx->frame_tail); tx 1531 drivers/net/ethernet/microchip/lan743x_main.c tx->last_tail = tx->frame_tail; tx 1535 drivers/net/ethernet/microchip/lan743x_main.c if (tx->vector_flags & LAN743X_VECTOR_FLAG_VECTOR_ENABLE_AUTO_SET) tx 1537 drivers/net/ethernet/microchip/lan743x_main.c if (tx->vector_flags & LAN743X_VECTOR_FLAG_SOURCE_ENABLE_AUTO_SET) tx 1541 drivers/net/ethernet/microchip/lan743x_main.c lan743x_csr_write(adapter, TX_TAIL(tx->channel_number), tx 1542 drivers/net/ethernet/microchip/lan743x_main.c tx_tail_flags | tx->frame_tail); tx 1543 drivers/net/ethernet/microchip/lan743x_main.c tx->frame_flags &= ~TX_FRAME_FLAG_IN_PROGRESS; tx 1546 drivers/net/ethernet/microchip/lan743x_main.c static netdev_tx_t lan743x_tx_xmit_frame(struct lan743x_tx *tx, tx 1560 drivers/net/ethernet/microchip/lan743x_main.c required_number_of_descriptors = lan743x_tx_get_desc_cnt(tx, skb); tx 1562 drivers/net/ethernet/microchip/lan743x_main.c spin_lock_irqsave(&tx->ring_lock, irq_flags); tx 1564 drivers/net/ethernet/microchip/lan743x_main.c lan743x_tx_get_avail_desc(tx)) { tx 1565 drivers/net/ethernet/microchip/lan743x_main.c if (required_number_of_descriptors > (tx->ring_size - 1)) { tx 1569 drivers/net/ethernet/microchip/lan743x_main.c tx->overflow_skb = skb; tx 1570 drivers/net/ethernet/microchip/lan743x_main.c netif_stop_queue(tx->adapter->netdev); tx 1577 drivers/net/ethernet/microchip/lan743x_main.c (tx->ts_flags & TX_TS_FLAG_TIMESTAMPING_ENABLED) && tx 1578 drivers/net/ethernet/microchip/lan743x_main.c (lan743x_ptp_request_tx_timestamp(tx->adapter))) { tx 1581 drivers/net/ethernet/microchip/lan743x_main.c if (tx->ts_flags & TX_TS_FLAG_ONE_STEP_SYNC) tx 1594 drivers/net/ethernet/microchip/lan743x_main.c if (lan743x_tx_frame_start(tx, tx 1604 drivers/net/ethernet/microchip/lan743x_main.c lan743x_tx_frame_add_lso(tx, frame_length, nr_frags); tx 1612 drivers/net/ethernet/microchip/lan743x_main.c if (lan743x_tx_frame_add_fragment(tx, frag, frame_length)) { tx 1624 drivers/net/ethernet/microchip/lan743x_main.c lan743x_tx_frame_end(tx, skb, do_timestamp, ignore_sync); tx 1627 drivers/net/ethernet/microchip/lan743x_main.c spin_unlock_irqrestore(&tx->ring_lock, irq_flags); tx 1633 drivers/net/ethernet/microchip/lan743x_main.c struct lan743x_tx *tx = container_of(napi, struct lan743x_tx, napi); tx 1634 drivers/net/ethernet/microchip/lan743x_main.c struct lan743x_adapter *adapter = tx->adapter; tx 1640 drivers/net/ethernet/microchip/lan743x_main.c ioc_bit = DMAC_INT_BIT_TX_IOC_(tx->channel_number); tx 1642 drivers/net/ethernet/microchip/lan743x_main.c if (tx->vector_flags & LAN743X_VECTOR_FLAG_SOURCE_STATUS_W2C) tx 1644 drivers/net/ethernet/microchip/lan743x_main.c spin_lock_irqsave(&tx->ring_lock, irq_flags); tx 1647 drivers/net/ethernet/microchip/lan743x_main.c lan743x_tx_release_completed_descriptors(tx); tx 1649 drivers/net/ethernet/microchip/lan743x_main.c if (tx->overflow_skb) { tx 1650 drivers/net/ethernet/microchip/lan743x_main.c if (lan743x_tx_get_desc_cnt(tx, tx->overflow_skb) <= tx 1651 drivers/net/ethernet/microchip/lan743x_main.c lan743x_tx_get_avail_desc(tx)) tx 1657 drivers/net/ethernet/microchip/lan743x_main.c spin_unlock_irqrestore(&tx->ring_lock, irq_flags); tx 1661 drivers/net/ethernet/microchip/lan743x_main.c lan743x_tx_xmit_frame(tx, tx->overflow_skb); tx 1662 drivers/net/ethernet/microchip/lan743x_main.c tx->overflow_skb = NULL; tx 1671 drivers/net/ethernet/microchip/lan743x_main.c INT_BIT_DMA_TX_(tx->channel_number)); tx 1678 drivers/net/ethernet/microchip/lan743x_main.c static void lan743x_tx_ring_cleanup(struct lan743x_tx *tx) tx 1680 drivers/net/ethernet/microchip/lan743x_main.c if (tx->head_cpu_ptr) { tx 1681 drivers/net/ethernet/microchip/lan743x_main.c pci_free_consistent(tx->adapter->pdev, tx 1682 drivers/net/ethernet/microchip/lan743x_main.c sizeof(*tx->head_cpu_ptr), tx 1683 drivers/net/ethernet/microchip/lan743x_main.c (void *)(tx->head_cpu_ptr), tx 1684 drivers/net/ethernet/microchip/lan743x_main.c tx->head_dma_ptr); tx 1685 drivers/net/ethernet/microchip/lan743x_main.c tx->head_cpu_ptr = NULL; tx 1686 drivers/net/ethernet/microchip/lan743x_main.c tx->head_dma_ptr = 0; tx 1688 drivers/net/ethernet/microchip/lan743x_main.c kfree(tx->buffer_info); tx 1689 drivers/net/ethernet/microchip/lan743x_main.c tx->buffer_info = NULL; tx 1691 drivers/net/ethernet/microchip/lan743x_main.c if (tx->ring_cpu_ptr) { tx 1692 drivers/net/ethernet/microchip/lan743x_main.c pci_free_consistent(tx->adapter->pdev, tx 1693 drivers/net/ethernet/microchip/lan743x_main.c tx->ring_allocation_size, tx 1694 drivers/net/ethernet/microchip/lan743x_main.c tx->ring_cpu_ptr, tx 1695 drivers/net/ethernet/microchip/lan743x_main.c tx->ring_dma_ptr); tx 1696 drivers/net/ethernet/microchip/lan743x_main.c tx->ring_allocation_size = 0; tx 1697 drivers/net/ethernet/microchip/lan743x_main.c tx->ring_cpu_ptr = NULL; tx 1698 drivers/net/ethernet/microchip/lan743x_main.c tx->ring_dma_ptr = 0; tx 1700 drivers/net/ethernet/microchip/lan743x_main.c tx->ring_size = 0; tx 1703 drivers/net/ethernet/microchip/lan743x_main.c static int lan743x_tx_ring_init(struct lan743x_tx *tx) tx 1710 drivers/net/ethernet/microchip/lan743x_main.c tx->ring_size = LAN743X_TX_RING_SIZE; tx 1711 drivers/net/ethernet/microchip/lan743x_main.c if (tx->ring_size & ~TX_CFG_B_TX_RING_LEN_MASK_) { tx 1715 drivers/net/ethernet/microchip/lan743x_main.c ring_allocation_size = ALIGN(tx->ring_size * tx 1719 drivers/net/ethernet/microchip/lan743x_main.c cpu_ptr = pci_zalloc_consistent(tx->adapter->pdev, tx 1726 drivers/net/ethernet/microchip/lan743x_main.c tx->ring_allocation_size = ring_allocation_size; tx 1727 drivers/net/ethernet/microchip/lan743x_main.c tx->ring_cpu_ptr = (struct lan743x_tx_descriptor *)cpu_ptr; tx 1728 drivers/net/ethernet/microchip/lan743x_main.c tx->ring_dma_ptr = dma_ptr; tx 1730 drivers/net/ethernet/microchip/lan743x_main.c cpu_ptr = kcalloc(tx->ring_size, sizeof(*tx->buffer_info), GFP_KERNEL); tx 1735 drivers/net/ethernet/microchip/lan743x_main.c tx->buffer_info = (struct lan743x_tx_buffer_info *)cpu_ptr; tx 1737 drivers/net/ethernet/microchip/lan743x_main.c cpu_ptr = pci_zalloc_consistent(tx->adapter->pdev, tx 1738 drivers/net/ethernet/microchip/lan743x_main.c sizeof(*tx->head_cpu_ptr), &dma_ptr); tx 1744 drivers/net/ethernet/microchip/lan743x_main.c tx->head_cpu_ptr = cpu_ptr; tx 1745 drivers/net/ethernet/microchip/lan743x_main.c tx->head_dma_ptr = dma_ptr; tx 1746 drivers/net/ethernet/microchip/lan743x_main.c if (tx->head_dma_ptr & 0x3) { tx 1754 drivers/net/ethernet/microchip/lan743x_main.c lan743x_tx_ring_cleanup(tx); tx 1758 drivers/net/ethernet/microchip/lan743x_main.c static void lan743x_tx_close(struct lan743x_tx *tx) tx 1760 drivers/net/ethernet/microchip/lan743x_main.c struct lan743x_adapter *adapter = tx->adapter; tx 1764 drivers/net/ethernet/microchip/lan743x_main.c DMAC_CMD_STOP_T_(tx->channel_number)); tx 1765 drivers/net/ethernet/microchip/lan743x_main.c lan743x_dmac_tx_wait_till_stopped(adapter, tx->channel_number); tx 1769 drivers/net/ethernet/microchip/lan743x_main.c DMAC_INT_BIT_TX_IOC_(tx->channel_number)); tx 1771 drivers/net/ethernet/microchip/lan743x_main.c INT_BIT_DMA_TX_(tx->channel_number)); tx 1772 drivers/net/ethernet/microchip/lan743x_main.c napi_disable(&tx->napi); tx 1773 drivers/net/ethernet/microchip/lan743x_main.c netif_napi_del(&tx->napi); tx 1776 drivers/net/ethernet/microchip/lan743x_main.c FCT_TX_CTL_DIS_(tx->channel_number)); tx 1778 drivers/net/ethernet/microchip/lan743x_main.c FCT_TX_CTL_EN_(tx->channel_number), tx 1781 drivers/net/ethernet/microchip/lan743x_main.c lan743x_tx_release_all_descriptors(tx); tx 1783 drivers/net/ethernet/microchip/lan743x_main.c if (tx->overflow_skb) { tx 1784 drivers/net/ethernet/microchip/lan743x_main.c dev_kfree_skb(tx->overflow_skb); tx 1785 drivers/net/ethernet/microchip/lan743x_main.c tx->overflow_skb = NULL; tx 1788 drivers/net/ethernet/microchip/lan743x_main.c lan743x_tx_ring_cleanup(tx); tx 1791 drivers/net/ethernet/microchip/lan743x_main.c static int lan743x_tx_open(struct lan743x_tx *tx) tx 1797 drivers/net/ethernet/microchip/lan743x_main.c adapter = tx->adapter; tx 1798 drivers/net/ethernet/microchip/lan743x_main.c ret = lan743x_tx_ring_init(tx); tx 1804 drivers/net/ethernet/microchip/lan743x_main.c FCT_TX_CTL_RESET_(tx->channel_number)); tx 1806 drivers/net/ethernet/microchip/lan743x_main.c FCT_TX_CTL_RESET_(tx->channel_number), tx 1811 drivers/net/ethernet/microchip/lan743x_main.c FCT_TX_CTL_EN_(tx->channel_number)); tx 1815 drivers/net/ethernet/microchip/lan743x_main.c DMAC_CMD_TX_SWR_(tx->channel_number)); tx 1817 drivers/net/ethernet/microchip/lan743x_main.c DMAC_CMD_TX_SWR_(tx->channel_number), tx 1822 drivers/net/ethernet/microchip/lan743x_main.c TX_BASE_ADDRH(tx->channel_number), tx 1823 drivers/net/ethernet/microchip/lan743x_main.c DMA_ADDR_HIGH32(tx->ring_dma_ptr)); tx 1825 drivers/net/ethernet/microchip/lan743x_main.c TX_BASE_ADDRL(tx->channel_number), tx 1826 drivers/net/ethernet/microchip/lan743x_main.c DMA_ADDR_LOW32(tx->ring_dma_ptr)); tx 1829 drivers/net/ethernet/microchip/lan743x_main.c data = lan743x_csr_read(adapter, TX_CFG_B(tx->channel_number)); tx 1831 drivers/net/ethernet/microchip/lan743x_main.c data |= ((tx->ring_size) & TX_CFG_B_TX_RING_LEN_MASK_); tx 1834 drivers/net/ethernet/microchip/lan743x_main.c lan743x_csr_write(adapter, TX_CFG_B(tx->channel_number), data); tx 1844 drivers/net/ethernet/microchip/lan743x_main.c lan743x_csr_write(adapter, TX_CFG_A(tx->channel_number), data); tx 1848 drivers/net/ethernet/microchip/lan743x_main.c TX_HEAD_WRITEBACK_ADDRH(tx->channel_number), tx 1849 drivers/net/ethernet/microchip/lan743x_main.c DMA_ADDR_HIGH32(tx->head_dma_ptr)); tx 1851 drivers/net/ethernet/microchip/lan743x_main.c TX_HEAD_WRITEBACK_ADDRL(tx->channel_number), tx 1852 drivers/net/ethernet/microchip/lan743x_main.c DMA_ADDR_LOW32(tx->head_dma_ptr)); tx 1855 drivers/net/ethernet/microchip/lan743x_main.c tx->last_head = lan743x_csr_read(adapter, TX_HEAD(tx->channel_number)); tx 1858 drivers/net/ethernet/microchip/lan743x_main.c tx->last_tail = 0; tx 1859 drivers/net/ethernet/microchip/lan743x_main.c lan743x_csr_write(adapter, TX_TAIL(tx->channel_number), tx 1860 drivers/net/ethernet/microchip/lan743x_main.c (u32)(tx->last_tail)); tx 1861 drivers/net/ethernet/microchip/lan743x_main.c tx->vector_flags = lan743x_intr_get_vector_flags(adapter, tx 1863 drivers/net/ethernet/microchip/lan743x_main.c (tx->channel_number)); tx 1865 drivers/net/ethernet/microchip/lan743x_main.c &tx->napi, lan743x_tx_napi_poll, tx 1866 drivers/net/ethernet/microchip/lan743x_main.c tx->ring_size - 1); tx 1867 drivers/net/ethernet/microchip/lan743x_main.c napi_enable(&tx->napi); tx 1870 drivers/net/ethernet/microchip/lan743x_main.c if (tx->vector_flags & LAN743X_VECTOR_FLAG_SOURCE_ENABLE_AUTO_CLEAR) tx 1872 drivers/net/ethernet/microchip/lan743x_main.c if (tx->vector_flags & LAN743X_VECTOR_FLAG_SOURCE_STATUS_AUTO_CLEAR) tx 1874 drivers/net/ethernet/microchip/lan743x_main.c if (tx->vector_flags & LAN743X_VECTOR_FLAG_SOURCE_STATUS_R2C) tx 1876 drivers/net/ethernet/microchip/lan743x_main.c if (tx->vector_flags & LAN743X_VECTOR_FLAG_SOURCE_ENABLE_R2C) tx 1878 drivers/net/ethernet/microchip/lan743x_main.c lan743x_csr_write(adapter, TX_CFG_C(tx->channel_number), data); tx 1880 drivers/net/ethernet/microchip/lan743x_main.c if (!(tx->vector_flags & LAN743X_VECTOR_FLAG_SOURCE_ENABLE_AUTO_SET)) tx 1882 drivers/net/ethernet/microchip/lan743x_main.c INT_BIT_DMA_TX_(tx->channel_number)); tx 1884 drivers/net/ethernet/microchip/lan743x_main.c DMAC_INT_BIT_TX_IOC_(tx->channel_number)); tx 1888 drivers/net/ethernet/microchip/lan743x_main.c DMAC_CMD_START_T_(tx->channel_number)); tx 2465 drivers/net/ethernet/microchip/lan743x_main.c lan743x_tx_close(&adapter->tx[0]); tx 2511 drivers/net/ethernet/microchip/lan743x_main.c ret = lan743x_tx_open(&adapter->tx[0]); tx 2544 drivers/net/ethernet/microchip/lan743x_main.c return lan743x_tx_xmit_frame(&adapter->tx[0], skb); tx 2671 drivers/net/ethernet/microchip/lan743x_main.c struct lan743x_tx *tx; tx 2706 drivers/net/ethernet/microchip/lan743x_main.c tx = &adapter->tx[0]; tx 2707 drivers/net/ethernet/microchip/lan743x_main.c tx->adapter = adapter; tx 2708 drivers/net/ethernet/microchip/lan743x_main.c tx->channel_number = 0; tx 2709 drivers/net/ethernet/microchip/lan743x_main.c spin_lock_init(&tx->ring_lock); tx 668 drivers/net/ethernet/microchip/lan743x_main.h void lan743x_tx_set_timestamping_mode(struct lan743x_tx *tx, tx 718 drivers/net/ethernet/microchip/lan743x_main.h struct lan743x_tx tx[LAN743X_MAX_TX_CHANNELS]; tx 1132 drivers/net/ethernet/microchip/lan743x_ptp.c lan743x_tx_set_timestamping_mode(&adapter->tx[index], tx 1139 drivers/net/ethernet/microchip/lan743x_ptp.c lan743x_tx_set_timestamping_mode(&adapter->tx[index], tx 1146 drivers/net/ethernet/microchip/lan743x_ptp.c lan743x_tx_set_timestamping_mode(&adapter->tx[index], tx 173 drivers/net/ethernet/myricom/myri10ge/myri10ge.c struct myri10ge_tx_buf tx; /* transmit ring */ tx 1049 drivers/net/ethernet/myricom/myri10ge/myri10ge.c ss->tx.req = 0; tx 1050 drivers/net/ethernet/myricom/myri10ge/myri10ge.c ss->tx.done = 0; tx 1051 drivers/net/ethernet/myricom/myri10ge/myri10ge.c ss->tx.pkt_start = 0; tx 1052 drivers/net/ethernet/myricom/myri10ge/myri10ge.c ss->tx.pkt_done = 0; tx 1057 drivers/net/ethernet/myricom/myri10ge/myri10ge.c ss->tx.wake_queue = 0; tx 1058 drivers/net/ethernet/myricom/myri10ge/myri10ge.c ss->tx.stop_queue = 0; tx 1390 drivers/net/ethernet/myricom/myri10ge/myri10ge.c struct myri10ge_tx_buf *tx = &ss->tx; tx 1395 drivers/net/ethernet/myricom/myri10ge/myri10ge.c while (tx->pkt_done != mcp_index) { tx 1396 drivers/net/ethernet/myricom/myri10ge/myri10ge.c idx = tx->done & tx->mask; tx 1397 drivers/net/ethernet/myricom/myri10ge/myri10ge.c skb = tx->info[idx].skb; tx 1400 drivers/net/ethernet/myricom/myri10ge/myri10ge.c tx->info[idx].skb = NULL; tx 1401 drivers/net/ethernet/myricom/myri10ge/myri10ge.c if (tx->info[idx].last) { tx 1402 drivers/net/ethernet/myricom/myri10ge/myri10ge.c tx->pkt_done++; tx 1403 drivers/net/ethernet/myricom/myri10ge/myri10ge.c tx->info[idx].last = 0; tx 1405 drivers/net/ethernet/myricom/myri10ge/myri10ge.c tx->done++; tx 1406 drivers/net/ethernet/myricom/myri10ge/myri10ge.c len = dma_unmap_len(&tx->info[idx], len); tx 1407 drivers/net/ethernet/myricom/myri10ge/myri10ge.c dma_unmap_len_set(&tx->info[idx], len, 0); tx 1414 drivers/net/ethernet/myricom/myri10ge/myri10ge.c dma_unmap_addr(&tx->info[idx], tx 1420 drivers/net/ethernet/myricom/myri10ge/myri10ge.c dma_unmap_addr(&tx->info[idx], tx 1438 drivers/net/ethernet/myricom/myri10ge/myri10ge.c if (tx->req == tx->done) { tx 1439 drivers/net/ethernet/myricom/myri10ge/myri10ge.c tx->queue_active = 0; tx 1440 drivers/net/ethernet/myricom/myri10ge/myri10ge.c put_be32(htonl(1), tx->send_stop); tx 1448 drivers/net/ethernet/myricom/myri10ge/myri10ge.c tx->req - tx->done < (tx->mask >> 1) && tx 1450 drivers/net/ethernet/myricom/myri10ge/myri10ge.c tx->wake_queue++; tx 1555 drivers/net/ethernet/myricom/myri10ge/myri10ge.c struct myri10ge_tx_buf *tx = &ss->tx; tx 1589 drivers/net/ethernet/myricom/myri10ge/myri10ge.c if (send_done_count != tx->pkt_done) tx 1720 drivers/net/ethernet/myricom/myri10ge/myri10ge.c ring->tx_max_pending = mgp->ss[0].tx.mask + 1; tx 1844 drivers/net/ethernet/myricom/myri10ge/myri10ge.c data[i++] = (unsigned int)ss->tx.pkt_start; tx 1845 drivers/net/ethernet/myricom/myri10ge/myri10ge.c data[i++] = (unsigned int)ss->tx.pkt_done; tx 1846 drivers/net/ethernet/myricom/myri10ge/myri10ge.c data[i++] = (unsigned int)ss->tx.req; tx 1847 drivers/net/ethernet/myricom/myri10ge/myri10ge.c data[i++] = (unsigned int)ss->tx.done; tx 1850 drivers/net/ethernet/myricom/myri10ge/myri10ge.c data[i++] = (unsigned int)ss->tx.wake_queue; tx 1851 drivers/net/ethernet/myricom/myri10ge/myri10ge.c data[i++] = (unsigned int)ss->tx.stop_queue; tx 1852 drivers/net/ethernet/myricom/myri10ge/myri10ge.c data[i++] = (unsigned int)ss->tx.linearized; tx 1962 drivers/net/ethernet/myricom/myri10ge/myri10ge.c ss->tx.mask = tx_ring_entries - 1; tx 1970 drivers/net/ethernet/myricom/myri10ge/myri10ge.c * sizeof(*ss->tx.req_list); tx 1971 drivers/net/ethernet/myricom/myri10ge/myri10ge.c ss->tx.req_bytes = kzalloc(bytes, GFP_KERNEL); tx 1972 drivers/net/ethernet/myricom/myri10ge/myri10ge.c if (ss->tx.req_bytes == NULL) tx 1976 drivers/net/ethernet/myricom/myri10ge/myri10ge.c ss->tx.req_list = (struct mcp_kreq_ether_send *) tx 1977 drivers/net/ethernet/myricom/myri10ge/myri10ge.c ALIGN((unsigned long)ss->tx.req_bytes, 8); tx 1978 drivers/net/ethernet/myricom/myri10ge/myri10ge.c ss->tx.queue_active = 0; tx 1992 drivers/net/ethernet/myricom/myri10ge/myri10ge.c bytes = tx_ring_entries * sizeof(*ss->tx.info); tx 1993 drivers/net/ethernet/myricom/myri10ge/myri10ge.c ss->tx.info = kzalloc(bytes, GFP_KERNEL); tx 1994 drivers/net/ethernet/myricom/myri10ge/myri10ge.c if (ss->tx.info == NULL) tx 2062 drivers/net/ethernet/myricom/myri10ge/myri10ge.c kfree(ss->tx.info); tx 2071 drivers/net/ethernet/myricom/myri10ge/myri10ge.c kfree(ss->tx.req_bytes); tx 2072 drivers/net/ethernet/myricom/myri10ge/myri10ge.c ss->tx.req_bytes = NULL; tx 2073 drivers/net/ethernet/myricom/myri10ge/myri10ge.c ss->tx.req_list = NULL; tx 2083 drivers/net/ethernet/myricom/myri10ge/myri10ge.c struct myri10ge_tx_buf *tx; tx 2087 drivers/net/ethernet/myricom/myri10ge/myri10ge.c if (ss->tx.req_list == NULL) tx 2110 drivers/net/ethernet/myricom/myri10ge/myri10ge.c tx = &ss->tx; tx 2111 drivers/net/ethernet/myricom/myri10ge/myri10ge.c while (tx->done != tx->req) { tx 2112 drivers/net/ethernet/myricom/myri10ge/myri10ge.c idx = tx->done & tx->mask; tx 2113 drivers/net/ethernet/myricom/myri10ge/myri10ge.c skb = tx->info[idx].skb; tx 2116 drivers/net/ethernet/myricom/myri10ge/myri10ge.c tx->info[idx].skb = NULL; tx 2117 drivers/net/ethernet/myricom/myri10ge/myri10ge.c tx->done++; tx 2118 drivers/net/ethernet/myricom/myri10ge/myri10ge.c len = dma_unmap_len(&tx->info[idx], len); tx 2119 drivers/net/ethernet/myricom/myri10ge/myri10ge.c dma_unmap_len_set(&tx->info[idx], len, 0); tx 2125 drivers/net/ethernet/myricom/myri10ge/myri10ge.c dma_unmap_addr(&tx->info[idx], tx 2131 drivers/net/ethernet/myricom/myri10ge/myri10ge.c dma_unmap_addr(&tx->info[idx], tx 2140 drivers/net/ethernet/myricom/myri10ge/myri10ge.c kfree(ss->tx.info); tx 2146 drivers/net/ethernet/myricom/myri10ge/myri10ge.c kfree(ss->tx.req_bytes); tx 2147 drivers/net/ethernet/myricom/myri10ge/myri10ge.c ss->tx.req_bytes = NULL; tx 2148 drivers/net/ethernet/myricom/myri10ge/myri10ge.c ss->tx.req_list = NULL; tx 2246 drivers/net/ethernet/myricom/myri10ge/myri10ge.c ss->tx.lanai = (struct mcp_kreq_ether_send __iomem *) tx 2260 drivers/net/ethernet/myricom/myri10ge/myri10ge.c ss->tx.send_go = (__iomem __be32 *) tx 2262 drivers/net/ethernet/myricom/myri10ge/myri10ge.c ss->tx.send_stop = (__iomem __be32 *) tx 2485 drivers/net/ethernet/myricom/myri10ge/myri10ge.c if (mgp->ss[0].tx.req_bytes == NULL) tx 2522 drivers/net/ethernet/myricom/myri10ge/myri10ge.c myri10ge_submit_req_backwards(struct myri10ge_tx_buf *tx, tx 2526 drivers/net/ethernet/myricom/myri10ge/myri10ge.c starting_slot = tx->req; tx 2529 drivers/net/ethernet/myricom/myri10ge/myri10ge.c idx = (starting_slot + cnt) & tx->mask; tx 2530 drivers/net/ethernet/myricom/myri10ge/myri10ge.c myri10ge_pio_copy(&tx->lanai[idx], &src[cnt], sizeof(*src)); tx 2543 drivers/net/ethernet/myricom/myri10ge/myri10ge.c myri10ge_submit_req(struct myri10ge_tx_buf *tx, struct mcp_kreq_ether_send *src, tx 2551 drivers/net/ethernet/myricom/myri10ge/myri10ge.c idx = tx->req & tx->mask; tx 2556 drivers/net/ethernet/myricom/myri10ge/myri10ge.c dst = dstp = &tx->lanai[idx]; tx 2559 drivers/net/ethernet/myricom/myri10ge/myri10ge.c if ((idx + cnt) < tx->mask) { tx 2569 drivers/net/ethernet/myricom/myri10ge/myri10ge.c myri10ge_submit_req_backwards(tx, src, cnt); tx 2581 drivers/net/ethernet/myricom/myri10ge/myri10ge.c tx->req += cnt; tx 2586 drivers/net/ethernet/myricom/myri10ge/myri10ge.c struct myri10ge_tx_buf *tx, int idx) tx 2592 drivers/net/ethernet/myricom/myri10ge/myri10ge.c last_idx = (idx + 1) & tx->mask; tx 2593 drivers/net/ethernet/myricom/myri10ge/myri10ge.c idx = tx->req & tx->mask; tx 2595 drivers/net/ethernet/myricom/myri10ge/myri10ge.c len = dma_unmap_len(&tx->info[idx], len); tx 2597 drivers/net/ethernet/myricom/myri10ge/myri10ge.c if (tx->info[idx].skb != NULL) tx 2599 drivers/net/ethernet/myricom/myri10ge/myri10ge.c dma_unmap_addr(&tx->info[idx], tx 2604 drivers/net/ethernet/myricom/myri10ge/myri10ge.c dma_unmap_addr(&tx->info[idx], tx 2607 drivers/net/ethernet/myricom/myri10ge/myri10ge.c dma_unmap_len_set(&tx->info[idx], len, 0); tx 2608 drivers/net/ethernet/myricom/myri10ge/myri10ge.c tx->info[idx].skb = NULL; tx 2610 drivers/net/ethernet/myricom/myri10ge/myri10ge.c idx = (idx + 1) & tx->mask; tx 2630 drivers/net/ethernet/myricom/myri10ge/myri10ge.c struct myri10ge_tx_buf *tx; tx 2645 drivers/net/ethernet/myricom/myri10ge/myri10ge.c tx = &ss->tx; tx 2648 drivers/net/ethernet/myricom/myri10ge/myri10ge.c req = tx->req_list; tx 2649 drivers/net/ethernet/myricom/myri10ge/myri10ge.c avail = tx->mask - 1 - (tx->req - tx->done); tx 2661 drivers/net/ethernet/myricom/myri10ge/myri10ge.c tx->stop_queue++; tx 2733 drivers/net/ethernet/myricom/myri10ge/myri10ge.c idx = tx->req & tx->mask; tx 2734 drivers/net/ethernet/myricom/myri10ge/myri10ge.c tx->info[idx].skb = skb; tx 2735 drivers/net/ethernet/myricom/myri10ge/myri10ge.c dma_unmap_addr_set(&tx->info[idx], bus, bus); tx 2736 drivers/net/ethernet/myricom/myri10ge/myri10ge.c dma_unmap_len_set(&tx->info[idx], len, len); tx 2839 drivers/net/ethernet/myricom/myri10ge/myri10ge.c myri10ge_unmap_tx_dma(mgp, tx, idx); tx 2842 drivers/net/ethernet/myricom/myri10ge/myri10ge.c idx = (count + tx->req) & tx->mask; tx 2843 drivers/net/ethernet/myricom/myri10ge/myri10ge.c dma_unmap_addr_set(&tx->info[idx], bus, bus); tx 2844 drivers/net/ethernet/myricom/myri10ge/myri10ge.c dma_unmap_len_set(&tx->info[idx], len, len); tx 2854 drivers/net/ethernet/myricom/myri10ge/myri10ge.c idx = ((count - 1) + tx->req) & tx->mask; tx 2855 drivers/net/ethernet/myricom/myri10ge/myri10ge.c tx->info[idx].last = 1; tx 2856 drivers/net/ethernet/myricom/myri10ge/myri10ge.c myri10ge_submit_req(tx, tx->req_list, count); tx 2859 drivers/net/ethernet/myricom/myri10ge/myri10ge.c if ((mgp->dev->real_num_tx_queues > 1) && tx->queue_active == 0) { tx 2860 drivers/net/ethernet/myricom/myri10ge/myri10ge.c tx->queue_active = 1; tx 2861 drivers/net/ethernet/myricom/myri10ge/myri10ge.c put_be32(htonl(1), tx->send_go); tx 2864 drivers/net/ethernet/myricom/myri10ge/myri10ge.c tx->pkt_start++; tx 2866 drivers/net/ethernet/myricom/myri10ge/myri10ge.c tx->stop_queue++; tx 2872 drivers/net/ethernet/myricom/myri10ge/myri10ge.c myri10ge_unmap_tx_dma(mgp, tx, idx); tx 2882 drivers/net/ethernet/myricom/myri10ge/myri10ge.c tx->linearized++; tx 3375 drivers/net/ethernet/myricom/myri10ge/myri10ge.c if (ss->tx.req != ss->tx.done && tx 3376 drivers/net/ethernet/myricom/myri10ge/myri10ge.c ss->tx.done == ss->watchdog_tx_done && tx 3386 drivers/net/ethernet/myricom/myri10ge/myri10ge.c slice, ss->tx.queue_active, ss->tx.req, tx 3387 drivers/net/ethernet/myricom/myri10ge/myri10ge.c ss->tx.done, ss->tx.pkt_start, tx 3388 drivers/net/ethernet/myricom/myri10ge/myri10ge.c ss->tx.pkt_done, tx 3395 drivers/net/ethernet/myricom/myri10ge/myri10ge.c if (ss->watchdog_tx_done != ss->tx.done || tx 3399 drivers/net/ethernet/myricom/myri10ge/myri10ge.c ss->watchdog_tx_done = ss->tx.done; tx 3400 drivers/net/ethernet/myricom/myri10ge/myri10ge.c ss->watchdog_tx_req = ss->tx.req; tx 1903 drivers/net/ethernet/neterion/vxge/vxge-config.c u32 port, u32 *tx, u32 *rx) tx 1925 drivers/net/ethernet/neterion/vxge/vxge-config.c *tx = 1; tx 1938 drivers/net/ethernet/neterion/vxge/vxge-config.c u32 port, u32 tx, u32 rx) tx 1959 drivers/net/ethernet/neterion/vxge/vxge-config.c if (tx) tx 1888 drivers/net/ethernet/neterion/vxge/vxge-config.h u32 *tx, tx 1894 drivers/net/ethernet/neterion/vxge/vxge-config.h u32 tx, tx 150 drivers/net/ethernet/netronome/nfp/nfp_net_debugfs.c struct dentry *queues, *tx, *rx, *xdp; tx 167 drivers/net/ethernet/netronome/nfp/nfp_net_debugfs.c tx = debugfs_create_dir("tx", queues); tx 180 drivers/net/ethernet/netronome/nfp/nfp_net_debugfs.c debugfs_create_file(name, 0400, tx, tx 111 drivers/net/ethernet/pasemi/pasemi_mac.c return mac->tx; tx 1079 drivers/net/ethernet/pasemi/pasemi_mac.c mac->tx = pasemi_mac_setup_tx_resources(dev); tx 1081 drivers/net/ethernet/pasemi/pasemi_mac.c if (!mac->tx) tx 1104 drivers/net/ethernet/pasemi/pasemi_mac.c write_iob_reg(PAS_IOB_DMA_TXCH_CFG(mac->tx->chan.chno), tx 1169 drivers/net/ethernet/pasemi/pasemi_mac.c ret = request_irq(mac->tx->chan.irq, pasemi_mac_tx_intr, 0, tx 1170 drivers/net/ethernet/pasemi/pasemi_mac.c mac->tx_irq_name, mac->tx); tx 1173 drivers/net/ethernet/pasemi/pasemi_mac.c mac->tx->chan.irq, ret); tx 1191 drivers/net/ethernet/pasemi/pasemi_mac.c timer_setup(&mac->tx->clean_timer, pasemi_mac_tx_timer, 0); tx 1192 drivers/net/ethernet/pasemi/pasemi_mac.c mod_timer(&mac->tx->clean_timer, jiffies + HZ); tx 1197 drivers/net/ethernet/pasemi/pasemi_mac.c free_irq(mac->tx->chan.irq, mac->tx); tx 1202 drivers/net/ethernet/pasemi/pasemi_mac.c if (mac->tx) tx 1287 drivers/net/ethernet/pasemi/pasemi_mac.c del_timer_sync(&mac->tx->clean_timer); tx 1319 drivers/net/ethernet/pasemi/pasemi_mac.c free_irq(mac->tx->chan.irq, mac->tx); tx 1583 drivers/net/ethernet/pasemi/pasemi_mac.c disable_irq(mac->tx->chan.irq); tx 1584 drivers/net/ethernet/pasemi/pasemi_mac.c pasemi_mac_tx_intr(mac->tx->chan.irq, mac->tx); tx 1585 drivers/net/ethernet/pasemi/pasemi_mac.c enable_irq(mac->tx->chan.irq); tx 1804 drivers/net/ethernet/pasemi/pasemi_mac.c pasemi_dma_free_chan(&mac->tx->chan); tx 76 drivers/net/ethernet/pasemi/pasemi_mac.h struct pasemi_mac_txring *tx; tx 94 drivers/net/ethernet/pasemi/pasemi_mac.h #define TX_DESC(tx, num) ((tx)->chan.ring_virt[(num) & (TX_RING_SIZE-1)]) tx 95 drivers/net/ethernet/pasemi/pasemi_mac.h #define TX_DESC_INFO(tx, num) ((tx)->ring_info[(num) & (TX_RING_SIZE-1)]) tx 77 drivers/net/ethernet/pasemi/pasemi_mac_ethtool.c ering->tx_pending = RING_USED(mac->tx)/2; tx 56 drivers/net/ethernet/pensando/ionic/ionic_lif.h struct ionic_tx_stats tx; tx 81 drivers/net/ethernet/pensando/ionic/ionic_lif.h #define q_to_tx_stats(q) (&q_to_qcq(q)->stats->tx) tx 185 drivers/net/ethernet/pensando/ionic/ionic_lif.h #define lif_to_txstats(lif, i) ((lif)->txqcqs[i].stats->tx) tx 257 drivers/net/ethernet/pensando/ionic/ionic_lif.h qcq->stats->tx.sg_cntr[num_sg_elems]++; tx 87 drivers/net/ethernet/pensando/ionic/ionic_stats.c tstats = &txqcq->stats->tx; tx 5734 drivers/net/ethernet/qlogic/qed/qed_hsi.h u8 tx; tx 2353 drivers/net/ethernet/qlogic/qed/qed_int.c u8 timer_res, u16 sb_id, bool tx) tx 2371 drivers/net/ethernet/qlogic/qed/qed_int.c if (tx) tx 430 drivers/net/ethernet/qlogic/qed/qed_int.h u8 timer_res, u16 sb_id, bool tx); tx 93 drivers/net/ethernet/qlogic/qed/qed_l2.c u8 rx = 0, tx = 0; tx 96 drivers/net/ethernet/qlogic/qed/qed_l2.c qed_vf_get_num_txqs(p_hwfn, &tx); tx 98 drivers/net/ethernet/qlogic/qed/qed_l2.c p_l2_info->queues = max_t(u8, rx, tx); tx 1281 drivers/net/ethernet/qlogic/qed/qed_l2.c p_ramrod->filter_cmd_hdr.tx = p_filter_cmd->is_tx_filter ? 1 : 0; tx 1510 drivers/net/ethernet/qlogic/qlcnic/qlcnic.h struct __qlcnic_esw_statistics tx; tx 1315 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c struct qlcnic_host_tx_ring *tx, int ring) tx 1327 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c tx->producer = 0; tx 1328 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c tx->sw_consumer = 0; tx 1329 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c *(tx->hw_consumer) = 0; tx 1334 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c mbx.phys_addr_low = LSD(tx->phys_addr); tx 1335 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c mbx.phys_addr_high = MSD(tx->phys_addr); tx 1336 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c mbx.cnsmr_index_low = LSD(tx->hw_cons_phys_addr); tx 1337 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c mbx.cnsmr_index_high = MSD(tx->hw_cons_phys_addr); tx 1338 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c mbx.size = tx->num_desc; tx 1378 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c tx->crb_cmd_producer = ahw->pci_base0 + mbx_out->host_prod; tx 1379 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c tx->ctx_id = mbx_out->ctx_id; tx 1383 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c tx->crb_intr_mask = ahw->pci_base0 + intr_mask; tx 1387 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c tx->ctx_id, mbx_out->state); tx 1393 drivers/net/ethernet/qlogic/qlcnic/qlcnic_ethtool.c QLCNIC_QUERY_TX_COUNTER, &port_stats.tx); tx 1397 drivers/net/ethernet/qlogic/qlcnic/qlcnic_ethtool.c qlcnic_fill_stats(data, &port_stats.tx, QLCNIC_ESW_STATS); tx 799 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sysfs.c &port_stats.tx); tx 834 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sysfs.c &esw_stats.tx); tx 221 drivers/net/ethernet/smsc/smc911x.h struct dma_async_tx_descriptor *tx; tx 233 drivers/net/ethernet/smsc/smc911x.h tx = dmaengine_prep_slave_single(dma, rx_dmabuf, rx_dmalen, tx 235 drivers/net/ethernet/smsc/smc911x.h if (tx) { tx 236 drivers/net/ethernet/smsc/smc911x.h tx->callback = smc911x_rx_dma_irq; tx 237 drivers/net/ethernet/smsc/smc911x.h tx->callback_param = lp; tx 238 drivers/net/ethernet/smsc/smc911x.h dmaengine_submit(tx); tx 253 drivers/net/ethernet/smsc/smc911x.h struct dma_async_tx_descriptor *tx; tx 265 drivers/net/ethernet/smsc/smc911x.h tx = dmaengine_prep_slave_single(dma, tx_dmabuf, tx_dmalen, tx 267 drivers/net/ethernet/smsc/smc911x.h if (tx) { tx 268 drivers/net/ethernet/smsc/smc911x.h tx->callback = smc911x_tx_dma_irq; tx 269 drivers/net/ethernet/smsc/smc911x.h tx->callback_param = lp; tx 270 drivers/net/ethernet/smsc/smc911x.h dmaengine_submit(tx); tx 302 drivers/net/ethernet/smsc/smc91x.h struct dma_async_tx_descriptor *tx; tx 308 drivers/net/ethernet/smsc/smc91x.h tx = dmaengine_prep_slave_single(lp->dma_chan, dmabuf, len, tx 310 drivers/net/ethernet/smsc/smc91x.h if (tx) { tx 311 drivers/net/ethernet/smsc/smc91x.h cookie = dmaengine_submit(tx); tx 278 drivers/net/ethernet/socionext/sni_ave.c struct ave_desc_info tx; tx 302 drivers/net/ethernet/socionext/sni_ave.c addr = ((id == AVE_DESCID_TX) ? priv->tx.daddr : priv->rx.daddr) tx 320 drivers/net/ethernet/socionext/sni_ave.c addr = ((id == AVE_DESCID_TX) ? priv->tx.daddr : priv->rx.daddr) tx 687 drivers/net/ethernet/socionext/sni_ave.c proc_idx = priv->tx.proc_idx; tx 688 drivers/net/ethernet/socionext/sni_ave.c done_idx = priv->tx.done_idx; tx 689 drivers/net/ethernet/socionext/sni_ave.c ndesc = priv->tx.ndesc; tx 715 drivers/net/ethernet/socionext/sni_ave.c if (priv->tx.desc[done_idx].skbs) { tx 716 drivers/net/ethernet/socionext/sni_ave.c ave_dma_unmap(ndev, &priv->tx.desc[done_idx], tx 718 drivers/net/ethernet/socionext/sni_ave.c dev_consume_skb_any(priv->tx.desc[done_idx].skbs); tx 719 drivers/net/ethernet/socionext/sni_ave.c priv->tx.desc[done_idx].skbs = NULL; tx 725 drivers/net/ethernet/socionext/sni_ave.c priv->tx.done_idx = done_idx; tx 1275 drivers/net/ethernet/socionext/sni_ave.c priv->tx.desc = kcalloc(priv->tx.ndesc, sizeof(*priv->tx.desc), tx 1277 drivers/net/ethernet/socionext/sni_ave.c if (!priv->tx.desc) { tx 1285 drivers/net/ethernet/socionext/sni_ave.c kfree(priv->tx.desc); tx 1291 drivers/net/ethernet/socionext/sni_ave.c priv->tx.proc_idx = 0; tx 1292 drivers/net/ethernet/socionext/sni_ave.c priv->tx.done_idx = 0; tx 1293 drivers/net/ethernet/socionext/sni_ave.c for (entry = 0; entry < priv->tx.ndesc; entry++) { tx 1298 drivers/net/ethernet/socionext/sni_ave.c (((priv->tx.ndesc * priv->desc_size) << 16) & AVE_TXDC_SIZE), tx 1368 drivers/net/ethernet/socionext/sni_ave.c for (entry = 0; entry < priv->tx.ndesc; entry++) { tx 1369 drivers/net/ethernet/socionext/sni_ave.c if (!priv->tx.desc[entry].skbs) tx 1372 drivers/net/ethernet/socionext/sni_ave.c ave_dma_unmap(ndev, &priv->tx.desc[entry], DMA_TO_DEVICE); tx 1373 drivers/net/ethernet/socionext/sni_ave.c dev_kfree_skb_any(priv->tx.desc[entry].skbs); tx 1374 drivers/net/ethernet/socionext/sni_ave.c priv->tx.desc[entry].skbs = NULL; tx 1376 drivers/net/ethernet/socionext/sni_ave.c priv->tx.proc_idx = 0; tx 1377 drivers/net/ethernet/socionext/sni_ave.c priv->tx.done_idx = 0; tx 1391 drivers/net/ethernet/socionext/sni_ave.c kfree(priv->tx.desc); tx 1404 drivers/net/ethernet/socionext/sni_ave.c proc_idx = priv->tx.proc_idx; tx 1405 drivers/net/ethernet/socionext/sni_ave.c done_idx = priv->tx.done_idx; tx 1406 drivers/net/ethernet/socionext/sni_ave.c ndesc = priv->tx.ndesc; tx 1424 drivers/net/ethernet/socionext/sni_ave.c ret = ave_dma_map(ndev, &priv->tx.desc[proc_idx], tx 1432 drivers/net/ethernet/socionext/sni_ave.c priv->tx.desc[proc_idx].skbs = skb; tx 1450 drivers/net/ethernet/socionext/sni_ave.c priv->tx.proc_idx = (proc_idx + 1) % ndesc; tx 1624 drivers/net/ethernet/socionext/sni_ave.c priv->tx.daddr = AVE_TXDM_64; tx 1629 drivers/net/ethernet/socionext/sni_ave.c priv->tx.daddr = AVE_TXDM_32; tx 1637 drivers/net/ethernet/socionext/sni_ave.c priv->tx.ndesc = AVE_NR_TXDESC; tx 74 drivers/net/ethernet/stmicro/stmmac/dwmac-rk.c #define DELAY_ENABLE(soc, tx, rx) \ tx 75 drivers/net/ethernet/stmicro/stmmac/dwmac-rk.c (((tx) ? soc##_GMAC_TXCLK_DLY_ENABLE : soc##_GMAC_TXCLK_DLY_DISABLE) | \ tx 18 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c u32 tx, rx; tx 20 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c tx = readl(ioaddr + XGMAC_TX_CONFIG); tx 23 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c tx |= XGMAC_CORE_INIT_TX; tx 27 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c tx |= XGMAC_CONFIG_TE; tx 28 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c tx &= ~hw->link.speed_mask; tx 32 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c tx |= hw->link.xgmii.speed10000; tx 35 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c tx |= hw->link.speed2500; tx 39 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c tx |= hw->link.speed1000; tx 44 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c writel(tx, ioaddr + XGMAC_TX_CONFIG); tx 51 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c u32 tx = readl(ioaddr + XGMAC_TX_CONFIG); tx 55 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c tx |= XGMAC_CONFIG_TE; tx 58 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c tx &= ~XGMAC_CONFIG_TE; tx 62 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c writel(tx, ioaddr + XGMAC_TX_CONFIG); tx 61 drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c memset(&desc_data->tx, 0, sizeof(desc_data->tx)); tx 452 drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c memset(&ring->tx, 0, sizeof(ring->tx)); tx 527 drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c if ((tso && (pkt_info->mss != ring->tx.cur_mss)) || tx 528 drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c (vlan && (pkt_info->vlan_ctag != ring->tx.cur_vlan_ctag))) tx 699 drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c ring->tx.xmit_more = 0; tx 727 drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c if (tso && (pkt_info->mss != ring->tx.cur_mss)) tx 732 drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c if (vlan && (pkt_info->vlan_ctag != ring->tx.cur_vlan_ctag)) tx 789 drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c ring->tx.cur_mss = pkt_info->mss; tx 818 drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c ring->tx.cur_vlan_ctag = pkt_info->vlan_ctag; tx 973 drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c desc_data->tx.packets = pkt_info->tx_packets; tx 974 drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c desc_data->tx.bytes = pkt_info->tx_bytes; tx 1003 drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c ring->tx.xmit_more = 1; tx 49 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c ring->tx.queue_stopped = 1; tx 54 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c if (ring->tx.xmit_more) tx 134 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c if (skb_shinfo(skb)->gso_size != ring->tx.cur_mss) { tx 161 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c if (skb_vlan_tag_get(skb) != ring->tx.cur_vlan_ctag) tx 1079 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c tx_packets += desc_data->tx.packets; tx 1080 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c tx_bytes += desc_data->tx.bytes; tx 1096 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c if ((ring->tx.queue_stopped == 1) && tx 1098 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c ring->tx.queue_stopped = 0; tx 297 drivers/net/ethernet/synopsys/dwc-xlgmac.h struct xlgmac_tx_desc_data tx; tx 351 drivers/net/ethernet/synopsys/dwc-xlgmac.h } tx; tx 1336 drivers/net/ethernet/ti/netcp_core.c config.u.tx.filt_einfo = false; tx 1337 drivers/net/ethernet/ti/netcp_core.c config.u.tx.filt_pswords = false; tx 1338 drivers/net/ethernet/ti/netcp_core.c config.u.tx.priority = DMA_PRIO_MED_L; tx 776 drivers/net/ethernet/toshiba/ps3_gelic_net.c card->vlan[type].tx); tx 1572 drivers/net/ethernet/toshiba/ps3_gelic_net.c int tx; tx 1576 drivers/net/ethernet/toshiba/ps3_gelic_net.c .tx = GELIC_LV1_VLAN_TX_ETHERNET_0, tx 1580 drivers/net/ethernet/toshiba/ps3_gelic_net.c .tx = GELIC_LV1_VLAN_TX_WIRELESS, tx 1589 drivers/net/ethernet/toshiba/ps3_gelic_net.c vlan_id_ix[i].tx, tx 1595 drivers/net/ethernet/toshiba/ps3_gelic_net.c vlan_id_ix[i].tx, status); tx 1596 drivers/net/ethernet/toshiba/ps3_gelic_net.c card->vlan[i].tx = 0; tx 1600 drivers/net/ethernet/toshiba/ps3_gelic_net.c card->vlan[i].tx = (u16)v1; tx 1612 drivers/net/ethernet/toshiba/ps3_gelic_net.c card->vlan[i].tx = 0; tx 1619 drivers/net/ethernet/toshiba/ps3_gelic_net.c i, card->vlan[i].tx, card->vlan[i].rx); tx 1622 drivers/net/ethernet/toshiba/ps3_gelic_net.c if (card->vlan[GELIC_PORT_ETHERNET_0].tx) { tx 1623 drivers/net/ethernet/toshiba/ps3_gelic_net.c BUG_ON(!card->vlan[GELIC_PORT_WIRELESS].tx); tx 1630 drivers/net/ethernet/toshiba/ps3_gelic_net.c card->vlan[GELIC_PORT_WIRELESS].tx = 0; tx 257 drivers/net/ethernet/toshiba/ps3_gelic_net.h u16 tx; tx 2585 drivers/net/ethernet/toshiba/ps3_gelic_wireless.c if (!card->vlan[GELIC_PORT_WIRELESS].tx) tx 2627 drivers/net/ethernet/toshiba/ps3_gelic_wireless.c if (!card->vlan[GELIC_PORT_WIRELESS].tx) tx 614 drivers/net/ethernet/tundra/tsi108_eth.c int tx; tx 619 drivers/net/ethernet/tundra/tsi108_eth.c tx = data->txtail; tx 621 drivers/net/ethernet/tundra/tsi108_eth.c if (data->txring[tx].misc & TSI108_TX_OWN) tx 624 drivers/net/ethernet/tundra/tsi108_eth.c skb = data->txskbs[tx]; tx 626 drivers/net/ethernet/tundra/tsi108_eth.c if (!(data->txring[tx].misc & TSI108_TX_OK)) tx 628 drivers/net/ethernet/tundra/tsi108_eth.c dev->name, data->txring[tx].misc); tx 633 drivers/net/ethernet/tundra/tsi108_eth.c if (data->txring[tx].misc & TSI108_TX_EOF) { tx 678 drivers/net/ethernet/tundra/tsi108_eth.c int tx = data->txhead; tx 690 drivers/net/ethernet/tundra/tsi108_eth.c if ((tx % TSI108_TX_INT_FREQ == 0) && tx 694 drivers/net/ethernet/tundra/tsi108_eth.c data->txskbs[tx] = skb; tx 697 drivers/net/ethernet/tundra/tsi108_eth.c data->txring[tx].buf0 = dma_map_single(&data->pdev->dev, tx 700 drivers/net/ethernet/tundra/tsi108_eth.c data->txring[tx].len = skb_headlen(skb); tx 705 drivers/net/ethernet/tundra/tsi108_eth.c data->txring[tx].buf0 = tx 709 drivers/net/ethernet/tundra/tsi108_eth.c data->txring[tx].len = skb_frag_size(frag); tx 723 drivers/net/ethernet/tundra/tsi108_eth.c data->txring[tx].misc = misc | TSI108_TX_OWN; tx 1401 drivers/net/ethernet/tundra/tsi108_eth.c int tx = data->txtail; tx 1403 drivers/net/ethernet/tundra/tsi108_eth.c skb = data->txskbs[tx]; tx 1399 drivers/net/ethernet/via/via-velocity.c for (i = 0; i < vptr->tx.numq; i++) { tx 1400 drivers/net/ethernet/via/via-velocity.c writel(vptr->tx.pool_dma[i], ®s->TDBaseLo[i]); tx 1475 drivers/net/ethernet/via/via-velocity.c pool = dma_alloc_coherent(vptr->dev, tx_ring_size * vptr->tx.numq + tx 1489 drivers/net/ethernet/via/via-velocity.c for (i = 0; i < vptr->tx.numq; i++) { tx 1490 drivers/net/ethernet/via/via-velocity.c vptr->tx.rings[i] = pool; tx 1491 drivers/net/ethernet/via/via-velocity.c vptr->tx.pool_dma[i] = pool_dma; tx 1648 drivers/net/ethernet/via/via-velocity.c for (j = 0; j < vptr->tx.numq; j++) { tx 1650 drivers/net/ethernet/via/via-velocity.c vptr->tx.infos[j] = kcalloc(vptr->options.numtx, tx 1653 drivers/net/ethernet/via/via-velocity.c if (!vptr->tx.infos[j]) { tx 1655 drivers/net/ethernet/via/via-velocity.c kfree(vptr->tx.infos[j]); tx 1659 drivers/net/ethernet/via/via-velocity.c vptr->tx.tail[j] = vptr->tx.curr[j] = vptr->tx.used[j] = 0; tx 1673 drivers/net/ethernet/via/via-velocity.c vptr->options.numtx * sizeof(struct tx_desc) * vptr->tx.numq; tx 1743 drivers/net/ethernet/via/via-velocity.c struct velocity_td_info *td_info = &(vptr->tx.infos[q][n]); tx 1773 drivers/net/ethernet/via/via-velocity.c for (j = 0; j < vptr->tx.numq; j++) { tx 1774 drivers/net/ethernet/via/via-velocity.c if (vptr->tx.infos[j] == NULL) tx 1779 drivers/net/ethernet/via/via-velocity.c kfree(vptr->tx.infos[j]); tx 1780 drivers/net/ethernet/via/via-velocity.c vptr->tx.infos[j] = NULL; tx 1898 drivers/net/ethernet/via/via-velocity.c for (qnum = 0; qnum < vptr->tx.numq; qnum++) { tx 1899 drivers/net/ethernet/via/via-velocity.c for (idx = vptr->tx.tail[qnum]; vptr->tx.used[qnum] > 0; tx 1905 drivers/net/ethernet/via/via-velocity.c td = &(vptr->tx.rings[qnum][idx]); tx 1906 drivers/net/ethernet/via/via-velocity.c tdinfo = &(vptr->tx.infos[qnum][idx]); tx 1930 drivers/net/ethernet/via/via-velocity.c vptr->tx.used[qnum]--; tx 1932 drivers/net/ethernet/via/via-velocity.c vptr->tx.tail[qnum] = idx; tx 2286 drivers/net/ethernet/via/via-velocity.c struct tx_info tx; tx 2298 drivers/net/ethernet/via/via-velocity.c tmp_vptr->tx.numq = vptr->tx.numq; tx 2312 drivers/net/ethernet/via/via-velocity.c tx = vptr->tx; tx 2315 drivers/net/ethernet/via/via-velocity.c vptr->tx = tmp_vptr->tx; tx 2318 drivers/net/ethernet/via/via-velocity.c tmp_vptr->tx = tx; tx 2540 drivers/net/ethernet/via/via-velocity.c index = vptr->tx.curr[qnum]; tx 2541 drivers/net/ethernet/via/via-velocity.c td_ptr = &(vptr->tx.rings[qnum][index]); tx 2542 drivers/net/ethernet/via/via-velocity.c tdinfo = &(vptr->tx.infos[qnum][index]); tx 2597 drivers/net/ethernet/via/via-velocity.c vptr->tx.used[qnum]++; tx 2598 drivers/net/ethernet/via/via-velocity.c vptr->tx.curr[qnum] = (index + 1) % vptr->options.numtx; tx 2603 drivers/net/ethernet/via/via-velocity.c td_ptr = &(vptr->tx.rings[qnum][prev]); tx 2642 drivers/net/ethernet/via/via-velocity.c vptr->tx.numq = info->txqueue; tx 3216 drivers/net/ethernet/via/via-velocity.c for (i = 0; i < vptr->tx.numq; i++) { tx 3217 drivers/net/ethernet/via/via-velocity.c if (vptr->tx.used[i]) tx 1423 drivers/net/ethernet/via/via-velocity.h #define AVAIL_TD(p,q) ((p)->options.numtx-((p)->tx.used[(q)])) tx 1451 drivers/net/ethernet/via/via-velocity.h } tx; tx 309 drivers/net/fddi/skfp/fplustm.c smc->hw.fp.tx[QUEUE_S] = queue = &smc->hw.fp.tx_q[QUEUE_S] ; tx 320 drivers/net/fddi/skfp/fplustm.c smc->hw.fp.tx[QUEUE_A0] = queue = &smc->hw.fp.tx_q[QUEUE_A0] ; tx 185 drivers/net/fddi/skfp/h/fplustm.h struct s_smt_tx_queue *tx[USED_QUEUES] ; tx 362 drivers/net/fddi/skfp/hwmtm.c queue = smc->hw.fp.tx[QUEUE_A0] ; tx 376 drivers/net/fddi/skfp/hwmtm.c queue = smc->hw.fp.tx[QUEUE_S] ; tx 563 drivers/net/fddi/skfp/hwmtm.c phys = repair_txd_ring(smc,smc->hw.fp.tx[QUEUE_A0]) ; tx 568 drivers/net/fddi/skfp/hwmtm.c phys = repair_txd_ring(smc,smc->hw.fp.tx[QUEUE_S]) ; tx 1550 drivers/net/fddi/skfp/hwmtm.c smc->os.hwm.tx_p = smc->hw.fp.tx[frame_status & QUEUE_A0] ; tx 1848 drivers/net/fddi/skfp/hwmtm.c queue = smc->hw.fp.tx[QUEUE_A0] ; tx 1950 drivers/net/fddi/skfp/hwmtm.c queue = smc->hw.fp.tx[i] ; tx 2041 drivers/net/fddi/skfp/hwmtm.c queue = smc->hw.fp.tx[i] ; tx 2065 drivers/net/fddi/skfp/hwmtm.c queue = smc->hw.fp.tx[i] ; tx 317 drivers/net/fddi/skfp/rmt.c int tx ; tx 323 drivers/net/fddi/skfp/rmt.c if ((tx = sm_mac_get_tx_state(smc)) == 4 || tx == 5) { tx 342 drivers/net/fddi/skfp/rmt.c tx, smc->r.bn_flag); tx 395 drivers/net/fddi/skfp/rmt.c int tx ; tx 401 drivers/net/fddi/skfp/rmt.c if ((tx = sm_mac_get_tx_state(smc)) == 4 || tx == 5) { tx 420 drivers/net/fddi/skfp/rmt.c tx, smc->r.bn_flag); tx 246 drivers/net/fjes/fjes_hw.c result = fjes_hw_alloc_epbuf(&buf_pair->tx); tx 255 drivers/net/fjes/fjes_hw.c fjes_hw_setup_epbuf(&buf_pair->tx, mac, tx 295 drivers/net/fjes/fjes_hw.c fjes_hw_free_epbuf(&hw->ep_shm_info[epidx].tx); tx 493 drivers/net/fjes/fjes_hw.c buf_pair->tx.size, tx 498 drivers/net/fjes/fjes_hw.c req_buf->share_buffer.buffer[idx++] = buf_pair->tx.size; tx 499 drivers/net/fjes/fjes_hw.c page_count = buf_pair->tx.size / EP_BUFFER_INFO_SIZE; tx 501 drivers/net/fjes/fjes_hw.c addr = ((u8 *)(buf_pair->tx.buffer)) + tx 734 drivers/net/fjes/fjes_hw.c return (hw->ep_shm_info[src_epid].tx.info->v1i.rx_status & tx 784 drivers/net/fjes/fjes_hw.c hw->ep_shm_info[epidx].tx.info->v1i.rx_status |= tx 1062 drivers/net/fjes/fjes_hw.c fjes_hw_setup_epbuf(&hw->ep_shm_info[epidx].tx, tx 1117 drivers/net/fjes/fjes_hw.c &hw->ep_shm_info[epidx].tx, tx 1132 drivers/net/fjes/fjes_hw.c hw->ep_shm_info[epidx].tx. tx 1161 drivers/net/fjes/fjes_hw.c tx.info->v1i.rx_status |= tx 252 drivers/net/fjes/fjes_hw.h } tx, rx; tx 331 drivers/net/fjes/fjes_main.c .tx.info->v1i.rx_status &= tx 404 drivers/net/fjes/fjes_main.c fjes_hw_setup_epbuf(&buf_pair->tx, netdev->dev_addr, tx 459 drivers/net/fjes/fjes_main.c fjes_hw_setup_epbuf(&buf_pair->tx, tx 521 drivers/net/fjes/fjes_main.c info = adapter->hw.ep_shm_info[epid].tx.info; tx 576 drivers/net/fjes/fjes_main.c hw->ep_shm_info[epid].tx.info->v1i.tx_status; tx 580 drivers/net/fjes/fjes_main.c hw->ep_shm_info[epid].tx.info->v1i.tx_status = tx 610 drivers/net/fjes/fjes_main.c retval = fjes_hw_epbuf_tx_pkt_send(&adapter->hw.ep_shm_info[dest].tx, tx 615 drivers/net/fjes/fjes_main.c adapter->hw.ep_shm_info[dest].tx.info->v1i.tx_status = tx 841 drivers/net/fjes/fjes_main.c hw->ep_shm_info[epidx].tx.info->v1i.rx_status &= tx 864 drivers/net/fjes/fjes_main.c fjes_hw_setup_epbuf(&hw->ep_shm_info[epidx].tx, tx 868 drivers/net/fjes/fjes_main.c hw->ep_shm_info[epidx].tx.info->v1i.rx_status |= tx 894 drivers/net/fjes/fjes_main.c &adapter->hw.ep_shm_info[epid].tx, vid)) tx 896 drivers/net/fjes/fjes_main.c &adapter->hw.ep_shm_info[epid].tx, vid); tx 912 drivers/net/fjes/fjes_main.c fjes_hw_del_vlan_id(&adapter->hw.ep_shm_info[epid].tx, vid); tx 935 drivers/net/fjes/fjes_main.c hw->ep_shm_info[src_epid].tx.info->v1i.rx_status |= tx 973 drivers/net/fjes/fjes_main.c hw->ep_shm_info[src_epid].tx.info->v1i.rx_status |= tx 1129 drivers/net/fjes/fjes_main.c .tx.info->v1i.rx_status |= FJES_RX_POLL_WORK; tx 1194 drivers/net/fjes/fjes_main.c adapter->hw.ep_shm_info[epidx].tx tx 1442 drivers/net/fjes/fjes_main.c fjes_hw_setup_epbuf(&hw->ep_shm_info[epidx].tx, tx 1487 drivers/net/fjes/fjes_main.c &hw->ep_shm_info[epidx].tx, tx 1499 drivers/net/fjes/fjes_main.c hw->ep_shm_info[epidx].tx.info->v1i.rx_status &= tx 100 drivers/net/fjes/fjes_trace.h __field(u64, tx) tx 106 drivers/net/fjes/fjes_trace.h void *tx, *rx; tx 108 drivers/net/fjes/fjes_trace.h tx = (void *)buf_pair->tx.buffer; tx 112 drivers/net/fjes/fjes_trace.h __entry->tx_size = buf_pair->tx.size; tx 114 drivers/net/fjes/fjes_trace.h __entry->tx = page_to_phys(vmalloc_to_page(tx)) + tx 115 drivers/net/fjes/fjes_trace.h offset_in_page(tx); tx 120 drivers/net/fjes/fjes_trace.h __entry->length, __entry->epid, __entry->tx, __entry->tx_size, tx 285 drivers/net/fjes/fjes_trace.h hw->ep_shm_info[src_epid].tx.info->v1i.rx_status; tx 305 drivers/net/fjes/fjes_trace.h __entry->rx_status = hw->ep_shm_info[src_epid].tx.info->v1i.rx_status; tx 329 drivers/net/fjes/fjes_trace.h hw->ep_shm_info[src_epid].tx.info->v1i.rx_status; tx 350 drivers/net/fjes/fjes_trace.h hw->ep_shm_info[src_epid].tx.info->v1i.rx_status; tx 203 drivers/net/hamradio/scc.c static void scc_key_trx (struct scc_channel *scc, char tx); tx 899 drivers/net/hamradio/scc.c static void scc_key_trx(struct scc_channel *scc, char tx) tx 904 drivers/net/hamradio/scc.c Outb(scc->ctrl + 4, scc->option | (tx? 0x80 : 0)); tx 909 drivers/net/hamradio/scc.c time_const = (unsigned) (scc->clock / (scc->modem.speed * (tx? 2:64))) - 2; tx 913 drivers/net/hamradio/scc.c if (tx) tx 921 drivers/net/hamradio/scc.c if (tx) tx 957 drivers/net/hamradio/scc.c if (tx) tx 105 drivers/net/ieee802154/at86rf230.c struct at86rf230_state_change tx; tx 908 drivers/net/ieee802154/at86rf230.c struct at86rf230_state_change *ctx = &lp->tx; tx 1712 drivers/net/ieee802154/at86rf230.c at86rf230_setup_spi_messages(lp, &lp->tx); tx 34 drivers/net/netdevsim/ipsec.c ipsec->count, ipsec->tx); tx 270 drivers/net/netdevsim/ipsec.c ipsec->tx++; tx 47 drivers/net/netdevsim/netdevsim.h u32 tx; tx 2094 drivers/net/phy/phy_device.c void phy_set_sym_pause(struct phy_device *phydev, bool rx, bool tx, tx 2099 drivers/net/phy/phy_device.c if (rx && tx && autoneg) tx 2118 drivers/net/phy/phy_device.c void phy_set_asym_pause(struct phy_device *phydev, bool rx, bool tx) tx 2136 drivers/net/phy/phy_device.c if (tx) tx 1610 drivers/net/usb/hso.c icount->tx = cnow.tx; tx 338 drivers/net/veth.c struct pcpu_lstats tx; tx 340 drivers/net/veth.c tot->tx_dropped = veth_stats_tx(&tx, dev); tx 341 drivers/net/veth.c tot->tx_bytes = tx.bytes; tx 342 drivers/net/veth.c tot->tx_packets = tx.packets; tx 352 drivers/net/veth.c tot->rx_dropped += veth_stats_tx(&tx, peer); tx 353 drivers/net/veth.c tot->rx_bytes += tx.bytes; tx 354 drivers/net/veth.c tot->rx_packets += tx.packets; tx 328 drivers/net/wan/ixp4xx_hss.c int tx, txdone, rx, rxfree; tx 889 drivers/net/wan/ixp4xx_hss.c queue_put_desc(queue_ids[port->id].tx, tx_desc_phys(port, n), desc); tx 927 drivers/net/wan/ixp4xx_hss.c err = qmgr_request_queue(queue_ids[port->id].tx, TX_DESCS, 0, 0, tx 946 drivers/net/wan/ixp4xx_hss.c qmgr_release_queue(queue_ids[port->id].tx); tx 961 drivers/net/wan/ixp4xx_hss.c qmgr_release_queue(queue_ids[port->id].tx); tx 1132 drivers/net/wan/ixp4xx_hss.c while (queue_get_desc(queue_ids[port->id].tx, port, 1) >= 0) tx 485 drivers/net/wan/z85230.c .tx = z8530_tx, tx 605 drivers/net/wan/z85230.c .tx = z8530_dma_tx, tx 611 drivers/net/wan/z85230.c .tx = z8530_dma_tx, tx 679 drivers/net/wan/z85230.c .tx = z8530_tx_clear, tx 739 drivers/net/wan/z85230.c irqs->tx(&dev->chanA); tx 751 drivers/net/wan/z85230.c irqs->tx(&dev->chanB); tx 256 drivers/net/wan/z85230.h void (*tx)(struct z8530_channel *); tx 204 drivers/net/wimax/i2400m/debugfs.c d_level_register_debugfs("dl_", tx, dentry); tx 982 drivers/net/wimax/i2400m/driver.c D_SUBMODULE_DEFINE(tx), tx 252 drivers/net/wimax/i2400m/tx.c #define D_SUBMODULE tx tx 74 drivers/net/wimax/i2400m/usb-tx.c #define D_SUBMODULE tx tx 365 drivers/net/wimax/i2400m/usb.c D_SUBMODULE_DEFINE(tx), tx 381 drivers/net/wimax/i2400m/usb.c d_level_register_debugfs("dl_", tx, dentry); tx 1766 drivers/net/wireless/admtek/adm8211.c .tx = adm8211_tx, tx 1353 drivers/net/wireless/ath/ar5523/ar5523.c .tx = ar5523_tx, tx 8184 drivers/net/wireless/ath/ath10k/mac.c .tx = ath10k_mac_op_tx, tx 1390 drivers/net/wireless/ath/ath10k/wmi-tlv.c ath10k_wmi_pull_pdev_stats_tx(&src->tx, dst); tx 2952 drivers/net/wireless/ath/ath10k/wmi.c ath10k_wmi_pull_pdev_stats_tx(&src->tx, dst); tx 3006 drivers/net/wireless/ath/ath10k/wmi.c ath10k_wmi_pull_pdev_stats_tx(&src->tx, dst); tx 3067 drivers/net/wireless/ath/ath10k/wmi.c ath10k_wmi_pull_pdev_stats_tx(&src->tx, dst); tx 3144 drivers/net/wireless/ath/ath10k/wmi.c ath10k_wmi_pull_pdev_stats_tx(&src->tx, dst); tx 3236 drivers/net/wireless/ath/ath10k/wmi.c ath10k_wmi_10_4_pull_pdev_stats_tx(&src->tx, dst); tx 5670 drivers/net/wireless/ath/ath10k/wmi.c u64 busy, total, tx, rx, rx_bss; tx 5683 drivers/net/wireless/ath/ath10k/wmi.c tx = __le64_to_cpu(ev->cycle_tx); tx 5689 drivers/net/wireless/ath/ath10k/wmi.c freq, noise_floor, busy, total, tx, rx, rx_bss); tx 5705 drivers/net/wireless/ath/ath10k/wmi.c survey->time_tx = div_u64(tx, cc_freq_hz); tx 4639 drivers/net/wireless/ath/ath10k/wmi.h struct wmi_pdev_stats_tx tx; tx 4655 drivers/net/wireless/ath/ath10k/wmi.h struct wmi_pdev_stats_tx tx; tx 4668 drivers/net/wireless/ath/ath10k/wmi.h struct wmi_pdev_stats_tx tx; tx 4679 drivers/net/wireless/ath/ath10k/wmi.h struct wmi_10_4_pdev_stats_tx tx; tx 740 drivers/net/wireless/ath/ath5k/mac80211-ops.c u32 *tx, u32 *tx_max, u32 *rx, u32 *rx_max) tx 744 drivers/net/wireless/ath/ath5k/mac80211-ops.c *tx = ah->txqs[AR5K_TX_QUEUE_ID_DATA_MIN].txq_max; tx 751 drivers/net/wireless/ath/ath5k/mac80211-ops.c static int ath5k_set_ringparam(struct ieee80211_hw *hw, u32 tx, u32 rx) tx 761 drivers/net/wireless/ath/ath5k/mac80211-ops.c if (!tx || tx > ATH5K_TXQ_LEN_MAX) tx 771 drivers/net/wireless/ath/ath5k/mac80211-ops.c ah->txqs[qnum].txq_max = tx; tx 781 drivers/net/wireless/ath/ath5k/mac80211-ops.c .tx = ath5k_tx, tx 3408 drivers/net/wireless/ath/ath6kl/cfg80211.c .tx = BIT(IEEE80211_STYPE_ACTION >> 4) | tx 3414 drivers/net/wireless/ath/ath6kl/cfg80211.c .tx = BIT(IEEE80211_STYPE_ACTION >> 4) | tx 3420 drivers/net/wireless/ath/ath6kl/cfg80211.c .tx = BIT(IEEE80211_STYPE_ACTION >> 4) | tx 3426 drivers/net/wireless/ath/ath6kl/cfg80211.c .tx = BIT(IEEE80211_STYPE_ACTION >> 4) | tx 48 drivers/net/wireless/ath/ath6kl/htc-ops.h return target->dev->ar->htc_ops->tx(target, packet); tx 317 drivers/net/wireless/ath/ath6kl/htc.h struct htc_tx_packet_info tx; tx 555 drivers/net/wireless/ath/ath6kl/htc.h int (*tx)(struct htc_target *target, struct htc_packet *packet); tx 643 drivers/net/wireless/ath/ath6kl/htc.h packet->info.tx.tag = tag; tx 384 drivers/net/wireless/ath/ath6kl/htc_mbox.c bool tx) tx 389 drivers/net/wireless/ath/ath6kl/htc_mbox.c buf_list = tx ? &target->free_ctrl_txbuf : &target->free_ctrl_rxbuf; tx 402 drivers/net/wireless/ath/ath6kl/htc_mbox.c if (tx) tx 420 drivers/net/wireless/ath/ath6kl/htc_mbox.c packet->info.tx.cred_used); tx 425 drivers/net/wireless/ath/ath6kl/htc_mbox.c packet->info.tx.cred_used; tx 458 drivers/net/wireless/ath/ath6kl/htc_mbox.c packet->info.tx.seqno); tx 523 drivers/net/wireless/ath/ath6kl/htc_mbox.c send_len, packet->info.tx.seqno, padded_len, tx 633 drivers/net/wireless/ath/ath6kl/htc_mbox.c packet->info.tx.cred_used = req_cred; tx 641 drivers/net/wireless/ath/ath6kl/htc_mbox.c packet->info.tx.flags = flags; tx 642 drivers/net/wireless/ath/ath6kl/htc_mbox.c packet->info.tx.seqno = endpoint->seqno; tx 715 drivers/net/wireless/ath/ath6kl/htc_mbox.c flags = packet->info.tx.flags | HTC_FLAGS_SEND_BUNDLE; tx 717 drivers/net/wireless/ath/ath6kl/htc_mbox.c cred_pad, packet->info.tx.seqno); tx 728 drivers/net/wireless/ath/ath6kl/htc_mbox.c i, packet, packet->info.tx.seqno, len, rem_scat); tx 925 drivers/net/wireless/ath/ath6kl/htc_mbox.c ath6kl_htc_tx_prep_pkt(packet, packet->info.tx.flags, tx 926 drivers/net/wireless/ath/ath6kl/htc_mbox.c 0, packet->info.tx.seqno); tx 1169 drivers/net/wireless/ath/ath6kl/htc_mbox.c (tag == packet->info.tx.tag)) tx 1181 drivers/net/wireless/ath/ath6kl/htc_mbox.c packet->endpoint, packet->info.tx.tag); tx 2920 drivers/net/wireless/ath/ath6kl/htc_mbox.c .tx = ath6kl_htc_mbox_tx, tx 31 drivers/net/wireless/ath/ath6kl/htc_pipe.c if (packet->info.tx.flags & HTC_FLAGS_TX_FIXUP_NETBUF) { tx 33 drivers/net/wireless/ath/ath6kl/htc_pipe.c packet->info.tx.flags &= ~HTC_FLAGS_TX_FIXUP_NETBUF; tx 163 drivers/net/wireless/ath/ath6kl/htc_pipe.c packet->info.tx.cred_used = credits_required; tx 165 drivers/net/wireless/ath/ath6kl/htc_pipe.c packet->info.tx.flags = send_flags; tx 166 drivers/net/wireless/ath/ath6kl/htc_pipe.c packet->info.tx.seqno = ep->seqno; tx 192 drivers/net/wireless/ath/ath6kl/htc_pipe.c packet->info.tx.seqno = ep->seqno; tx 193 drivers/net/wireless/ath/ath6kl/htc_pipe.c packet->info.tx.flags = 0; tx 194 drivers/net/wireless/ath/ath6kl/htc_pipe.c packet->info.tx.cred_used = 0; tx 238 drivers/net/wireless/ath/ath6kl/htc_pipe.c packet->info.tx.flags |= HTC_FLAGS_TX_FIXUP_NETBUF; tx 242 drivers/net/wireless/ath/ath6kl/htc_pipe.c htc_hdr->flags = packet->info.tx.flags; tx 245 drivers/net/wireless/ath/ath6kl/htc_pipe.c htc_hdr->ctrl[1] = (u8) packet->info.tx.seqno; tx 272 drivers/net/wireless/ath/ath6kl/htc_pipe.c ep->cred_dist.credits += packet->info.tx.cred_used; tx 1709 drivers/net/wireless/ath/ath6kl/htc_pipe.c .tx = ath6kl_htc_pipe_tx, tx 715 drivers/net/wireless/ath/ath6kl/main.c stats->tx_pkt += le32_to_cpu(tgt_stats->stats.tx.pkt); tx 716 drivers/net/wireless/ath/ath6kl/main.c stats->tx_byte += le32_to_cpu(tgt_stats->stats.tx.byte); tx 717 drivers/net/wireless/ath/ath6kl/main.c stats->tx_ucast_pkt += le32_to_cpu(tgt_stats->stats.tx.ucast_pkt); tx 718 drivers/net/wireless/ath/ath6kl/main.c stats->tx_ucast_byte += le32_to_cpu(tgt_stats->stats.tx.ucast_byte); tx 719 drivers/net/wireless/ath/ath6kl/main.c stats->tx_mcast_pkt += le32_to_cpu(tgt_stats->stats.tx.mcast_pkt); tx 720 drivers/net/wireless/ath/ath6kl/main.c stats->tx_mcast_byte += le32_to_cpu(tgt_stats->stats.tx.mcast_byte); tx 721 drivers/net/wireless/ath/ath6kl/main.c stats->tx_bcast_pkt += le32_to_cpu(tgt_stats->stats.tx.bcast_pkt); tx 722 drivers/net/wireless/ath/ath6kl/main.c stats->tx_bcast_byte += le32_to_cpu(tgt_stats->stats.tx.bcast_byte); tx 724 drivers/net/wireless/ath/ath6kl/main.c le32_to_cpu(tgt_stats->stats.tx.rts_success_cnt); tx 728 drivers/net/wireless/ath/ath6kl/main.c le32_to_cpu(tgt_stats->stats.tx.pkt_per_ac[ac]); tx 730 drivers/net/wireless/ath/ath6kl/main.c stats->tx_err += le32_to_cpu(tgt_stats->stats.tx.err); tx 731 drivers/net/wireless/ath/ath6kl/main.c stats->tx_fail_cnt += le32_to_cpu(tgt_stats->stats.tx.fail_cnt); tx 732 drivers/net/wireless/ath/ath6kl/main.c stats->tx_retry_cnt += le32_to_cpu(tgt_stats->stats.tx.retry_cnt); tx 734 drivers/net/wireless/ath/ath6kl/main.c le32_to_cpu(tgt_stats->stats.tx.mult_retry_cnt); tx 736 drivers/net/wireless/ath/ath6kl/main.c le32_to_cpu(tgt_stats->stats.tx.rts_fail_cnt); tx 738 drivers/net/wireless/ath/ath6kl/main.c rate = a_sle32_to_cpu(tgt_stats->stats.tx.ucast_rate); tx 92 drivers/net/wireless/ath/ath6kl/trace.h __field(unsigned int, tx) tx 106 drivers/net/wireless/ath/ath6kl/trace.h __entry->tx = 1; tx 108 drivers/net/wireless/ath/ath6kl/trace.h __entry->tx = 0; tx 113 drivers/net/wireless/ath/ath6kl/trace.h __entry->tx ? "tx" : "rx", tx 127 drivers/net/wireless/ath/ath6kl/trace.h __field(unsigned int, tx) tx 147 drivers/net/wireless/ath/ath6kl/trace.h __entry->tx = 1; tx 149 drivers/net/wireless/ath/ath6kl/trace.h __entry->tx = 0; tx 166 drivers/net/wireless/ath/ath6kl/trace.h __entry->tx ? "tx" : "rx", tx 607 drivers/net/wireless/ath/ath6kl/txrx.c if (packet->info.tx.tag == ATH6KL_CONTROL_PKT_TAG) tx 1832 drivers/net/wireless/ath/ath6kl/wmi.h struct tx_stats tx; tx 676 drivers/net/wireless/ath/ath9k/ar9003_phy.c void ar9003_hw_set_chain_masks(struct ath_hw *ah, u8 rx, u8 tx) tx 685 drivers/net/wireless/ath/ath9k/ar9003_phy.c if ((ah->caps.hw_caps & ATH9K_HW_CAP_APM) && (tx == 0x7)) tx 686 drivers/net/wireless/ath/ath9k/ar9003_phy.c tx = 3; tx 688 drivers/net/wireless/ath/ath9k/ar9003_phy.c REG_WRITE(ah, AR_SELFGEN_MASK, tx); tx 83 drivers/net/wireless/ath/ath9k/ath9k.h #define ATH_TXQ_SETUP(sc, i) ((sc)->tx.txqsetup & (1<<i)) tx 1026 drivers/net/wireless/ath/ath9k/ath9k.h struct ath_tx tx; tx 50 drivers/net/wireless/ath/ath9k/beacon.c txq = sc->tx.txq_map[IEEE80211_AC_BE]; tx 1016 drivers/net/wireless/ath/ath9k/channel.c txctl.txq = sc->tx.txq_map[IEEE80211_AC_VO]; tx 1137 drivers/net/wireless/ath/ath9k/channel.c txctl.txq = sc->tx.txq_map[IEEE80211_AC_VO]; tx 647 drivers/net/wireless/ath/ath9k/debug.c txq = sc->tx.txq_map[i]; tx 195 drivers/net/wireless/ath/ath9k/debug.h #define PR_QNUM(_n) sc->tx.txq_map[_n]->axq_qnum tx 429 drivers/net/wireless/ath/ath9k/gpio.c txq = sc->tx.txq_map[IEEE80211_AC_BE]; tx 162 drivers/net/wireless/ath/ath9k/hif_usb.c spin_lock_irqsave(&hif_dev->tx.tx_lock, flags); tx 163 drivers/net/wireless/ath/ath9k/hif_usb.c if (hif_dev->tx.flags & HIF_USB_TX_FLUSH) { tx 164 drivers/net/wireless/ath/ath9k/hif_usb.c spin_unlock_irqrestore(&hif_dev->tx.tx_lock, flags); tx 169 drivers/net/wireless/ath/ath9k/hif_usb.c spin_unlock_irqrestore(&hif_dev->tx.tx_lock, flags); tx 279 drivers/net/wireless/ath/ath9k/hif_usb.c spin_lock(&hif_dev->tx.tx_lock); tx 280 drivers/net/wireless/ath/ath9k/hif_usb.c if (hif_dev->tx.flags & HIF_USB_TX_FLUSH) { tx 281 drivers/net/wireless/ath/ath9k/hif_usb.c spin_unlock(&hif_dev->tx.tx_lock); tx 285 drivers/net/wireless/ath/ath9k/hif_usb.c spin_unlock(&hif_dev->tx.tx_lock); tx 300 drivers/net/wireless/ath/ath9k/hif_usb.c spin_lock(&hif_dev->tx.tx_lock); tx 301 drivers/net/wireless/ath/ath9k/hif_usb.c list_move_tail(&tx_buf->list, &hif_dev->tx.tx_buf); tx 302 drivers/net/wireless/ath/ath9k/hif_usb.c hif_dev->tx.tx_buf_cnt++; tx 303 drivers/net/wireless/ath/ath9k/hif_usb.c if (!(hif_dev->tx.flags & HIF_USB_TX_STOP)) tx 306 drivers/net/wireless/ath/ath9k/hif_usb.c spin_unlock(&hif_dev->tx.tx_lock); tx 319 drivers/net/wireless/ath/ath9k/hif_usb.c if (hif_dev->tx.tx_skb_cnt == 0) tx 323 drivers/net/wireless/ath/ath9k/hif_usb.c if (list_empty(&hif_dev->tx.tx_buf)) tx 326 drivers/net/wireless/ath/ath9k/hif_usb.c tx_buf = list_first_entry(&hif_dev->tx.tx_buf, struct tx_buf, list); tx 327 drivers/net/wireless/ath/ath9k/hif_usb.c list_move_tail(&tx_buf->list, &hif_dev->tx.tx_pending); tx 328 drivers/net/wireless/ath/ath9k/hif_usb.c hif_dev->tx.tx_buf_cnt--; tx 330 drivers/net/wireless/ath/ath9k/hif_usb.c tx_skb_cnt = min_t(u16, hif_dev->tx.tx_skb_cnt, MAX_TX_AGGR_NUM); tx 333 drivers/net/wireless/ath/ath9k/hif_usb.c nskb = __skb_dequeue(&hif_dev->tx.tx_skb_queue); tx 338 drivers/net/wireless/ath/ath9k/hif_usb.c hif_dev->tx.tx_skb_cnt--; tx 369 drivers/net/wireless/ath/ath9k/hif_usb.c list_move_tail(&tx_buf->list, &hif_dev->tx.tx_buf); tx 370 drivers/net/wireless/ath/ath9k/hif_usb.c hif_dev->tx.tx_buf_cnt++; tx 385 drivers/net/wireless/ath/ath9k/hif_usb.c spin_lock_irqsave(&hif_dev->tx.tx_lock, flags); tx 387 drivers/net/wireless/ath/ath9k/hif_usb.c if (hif_dev->tx.flags & HIF_USB_TX_STOP) { tx 388 drivers/net/wireless/ath/ath9k/hif_usb.c spin_unlock_irqrestore(&hif_dev->tx.tx_lock, flags); tx 393 drivers/net/wireless/ath/ath9k/hif_usb.c if (hif_dev->tx.tx_skb_cnt > MAX_TX_BUF_NUM) { tx 394 drivers/net/wireless/ath/ath9k/hif_usb.c spin_unlock_irqrestore(&hif_dev->tx.tx_lock, flags); tx 398 drivers/net/wireless/ath/ath9k/hif_usb.c spin_unlock_irqrestore(&hif_dev->tx.tx_lock, flags); tx 408 drivers/net/wireless/ath/ath9k/hif_usb.c spin_lock_irqsave(&hif_dev->tx.tx_lock, flags); tx 412 drivers/net/wireless/ath/ath9k/hif_usb.c __skb_queue_tail(&hif_dev->tx.tx_skb_queue, skb); tx 413 drivers/net/wireless/ath/ath9k/hif_usb.c hif_dev->tx.tx_skb_cnt++; tx 417 drivers/net/wireless/ath/ath9k/hif_usb.c if ((hif_dev->tx.tx_buf_cnt == MAX_TX_URB_NUM) && tx 418 drivers/net/wireless/ath/ath9k/hif_usb.c (hif_dev->tx.tx_skb_cnt < 2)) { tx 422 drivers/net/wireless/ath/ath9k/hif_usb.c spin_unlock_irqrestore(&hif_dev->tx.tx_lock, flags); tx 434 drivers/net/wireless/ath/ath9k/hif_usb.c spin_lock_irqsave(&hif_dev->tx.tx_lock, flags); tx 435 drivers/net/wireless/ath/ath9k/hif_usb.c hif_dev->tx.flags &= ~HIF_USB_TX_STOP; tx 436 drivers/net/wireless/ath/ath9k/hif_usb.c spin_unlock_irqrestore(&hif_dev->tx.tx_lock, flags); tx 445 drivers/net/wireless/ath/ath9k/hif_usb.c spin_lock_irqsave(&hif_dev->tx.tx_lock, flags); tx 446 drivers/net/wireless/ath/ath9k/hif_usb.c ath9k_skb_queue_complete(hif_dev, &hif_dev->tx.tx_skb_queue, false); tx 447 drivers/net/wireless/ath/ath9k/hif_usb.c hif_dev->tx.tx_skb_cnt = 0; tx 448 drivers/net/wireless/ath/ath9k/hif_usb.c hif_dev->tx.flags |= HIF_USB_TX_STOP; tx 449 drivers/net/wireless/ath/ath9k/hif_usb.c spin_unlock_irqrestore(&hif_dev->tx.tx_lock, flags); tx 453 drivers/net/wireless/ath/ath9k/hif_usb.c &hif_dev->tx.tx_pending, list) { tx 501 drivers/net/wireless/ath/ath9k/hif_usb.c spin_lock_irqsave(&hif_dev->tx.tx_lock, flags); tx 503 drivers/net/wireless/ath/ath9k/hif_usb.c skb_queue_walk_safe(&hif_dev->tx.tx_skb_queue, skb, tmp) { tx 505 drivers/net/wireless/ath/ath9k/hif_usb.c __skb_unlink(skb, &hif_dev->tx.tx_skb_queue); tx 508 drivers/net/wireless/ath/ath9k/hif_usb.c hif_dev->tx.tx_skb_cnt--; tx 513 drivers/net/wireless/ath/ath9k/hif_usb.c spin_unlock_irqrestore(&hif_dev->tx.tx_lock, flags); tx 764 drivers/net/wireless/ath/ath9k/hif_usb.c &hif_dev->tx.tx_buf, list) { tx 772 drivers/net/wireless/ath/ath9k/hif_usb.c spin_lock_irqsave(&hif_dev->tx.tx_lock, flags); tx 773 drivers/net/wireless/ath/ath9k/hif_usb.c hif_dev->tx.flags |= HIF_USB_TX_FLUSH; tx 774 drivers/net/wireless/ath/ath9k/hif_usb.c spin_unlock_irqrestore(&hif_dev->tx.tx_lock, flags); tx 777 drivers/net/wireless/ath/ath9k/hif_usb.c &hif_dev->tx.tx_pending, list) { tx 793 drivers/net/wireless/ath/ath9k/hif_usb.c INIT_LIST_HEAD(&hif_dev->tx.tx_buf); tx 794 drivers/net/wireless/ath/ath9k/hif_usb.c INIT_LIST_HEAD(&hif_dev->tx.tx_pending); tx 795 drivers/net/wireless/ath/ath9k/hif_usb.c spin_lock_init(&hif_dev->tx.tx_lock); tx 796 drivers/net/wireless/ath/ath9k/hif_usb.c __skb_queue_head_init(&hif_dev->tx.tx_skb_queue); tx 815 drivers/net/wireless/ath/ath9k/hif_usb.c list_add_tail(&tx_buf->list, &hif_dev->tx.tx_buf); tx 818 drivers/net/wireless/ath/ath9k/hif_usb.c hif_dev->tx.tx_buf_cnt = MAX_TX_URB_NUM; tx 123 drivers/net/wireless/ath/ath9k/hif_usb.h struct hif_usb_tx tx; tx 499 drivers/net/wireless/ath/ath9k/htc.h struct ath9k_htc_tx tx; tx 180 drivers/net/wireless/ath/ath9k/htc_drv_beacon.c spin_lock_bh(&priv->tx.tx_lock); tx 181 drivers/net/wireless/ath/ath9k/htc_drv_beacon.c priv->tx.queued_cnt++; tx 182 drivers/net/wireless/ath/ath9k/htc_drv_beacon.c spin_unlock_bh(&priv->tx.tx_lock); tx 296 drivers/net/wireless/ath/ath9k/htc_drv_debug.c spin_lock_bh(&priv->tx.tx_lock); tx 300 drivers/net/wireless/ath/ath9k/htc_drv_debug.c MAX_TX_BUF_NUM, priv->tx.tx_slot, tx 301 drivers/net/wireless/ath/ath9k/htc_drv_debug.c bitmap_weight(priv->tx.tx_slot, MAX_TX_BUF_NUM)); tx 302 drivers/net/wireless/ath/ath9k/htc_drv_debug.c spin_unlock_bh(&priv->tx.tx_lock); tx 321 drivers/net/wireless/ath/ath9k/htc_drv_debug.c "Mgmt endpoint", skb_queue_len(&priv->tx.mgmt_ep_queue)); tx 324 drivers/net/wireless/ath/ath9k/htc_drv_debug.c "Cab endpoint", skb_queue_len(&priv->tx.cab_ep_queue)); tx 327 drivers/net/wireless/ath/ath9k/htc_drv_debug.c "Data BE endpoint", skb_queue_len(&priv->tx.data_be_queue)); tx 330 drivers/net/wireless/ath/ath9k/htc_drv_debug.c "Data BK endpoint", skb_queue_len(&priv->tx.data_bk_queue)); tx 333 drivers/net/wireless/ath/ath9k/htc_drv_debug.c "Data VI endpoint", skb_queue_len(&priv->tx.data_vi_queue)); tx 336 drivers/net/wireless/ath/ath9k/htc_drv_debug.c "Data VO endpoint", skb_queue_len(&priv->tx.data_vo_queue)); tx 339 drivers/net/wireless/ath/ath9k/htc_drv_debug.c "Failed queue", skb_queue_len(&priv->tx.tx_failed)); tx 341 drivers/net/wireless/ath/ath9k/htc_drv_debug.c spin_lock_bh(&priv->tx.tx_lock); tx 343 drivers/net/wireless/ath/ath9k/htc_drv_debug.c "Queued count", priv->tx.queued_cnt); tx 344 drivers/net/wireless/ath/ath9k/htc_drv_debug.c spin_unlock_bh(&priv->tx.tx_lock); tx 122 drivers/net/wireless/ath/ath9k/htc_drv_init.c void (*tx) (void *, tx 135 drivers/net/wireless/ath/ath9k/htc_drv_init.c req.ep_callbacks.tx = tx; tx 645 drivers/net/wireless/ath/ath9k/htc_drv_init.c spin_lock_init(&priv->tx.tx_lock); tx 655 drivers/net/wireless/ath/ath9k/htc_drv_init.c timer_setup(&priv->tx.cleanup_timer, ath9k_htc_tx_cleanup_timer, 0); tx 201 drivers/net/wireless/ath/ath9k/htc_drv_main.c del_timer_sync(&priv->tx.cleanup_timer); tx 233 drivers/net/wireless/ath/ath9k/htc_drv_main.c mod_timer(&priv->tx.cleanup_timer, tx 263 drivers/net/wireless/ath/ath9k/htc_drv_main.c del_timer_sync(&priv->tx.cleanup_timer); tx 311 drivers/net/wireless/ath/ath9k/htc_drv_main.c mod_timer(&priv->tx.cleanup_timer, tx 735 drivers/net/wireless/ath/ath9k/htc_drv_main.c spin_lock_bh(&priv->tx.tx_lock); tx 737 drivers/net/wireless/ath/ath9k/htc_drv_main.c spin_unlock_bh(&priv->tx.tx_lock); tx 960 drivers/net/wireless/ath/ath9k/htc_drv_main.c spin_lock_bh(&priv->tx.tx_lock); tx 961 drivers/net/wireless/ath/ath9k/htc_drv_main.c priv->tx.flags &= ~ATH9K_HTC_OP_TX_QUEUES_STOP; tx 962 drivers/net/wireless/ath/ath9k/htc_drv_main.c spin_unlock_bh(&priv->tx.tx_lock); tx 966 drivers/net/wireless/ath/ath9k/htc_drv_main.c mod_timer(&priv->tx.cleanup_timer, tx 1000 drivers/net/wireless/ath/ath9k/htc_drv_main.c del_timer_sync(&priv->tx.cleanup_timer); tx 1691 drivers/net/wireless/ath/ath9k/htc_drv_main.c spin_lock_bh(&priv->tx.tx_lock); tx 1693 drivers/net/wireless/ath/ath9k/htc_drv_main.c spin_unlock_bh(&priv->tx.tx_lock); tx 1870 drivers/net/wireless/ath/ath9k/htc_drv_main.c .tx = ath9k_htc_tx, tx 58 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c spin_lock_bh(&priv->tx.tx_lock); tx 59 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c priv->tx.queued_cnt++; tx 60 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c if ((priv->tx.queued_cnt >= ATH9K_HTC_TX_THRESHOLD) && tx 61 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c !(priv->tx.flags & ATH9K_HTC_OP_TX_QUEUES_STOP)) { tx 62 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c priv->tx.flags |= ATH9K_HTC_OP_TX_QUEUES_STOP; tx 65 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c spin_unlock_bh(&priv->tx.tx_lock); tx 70 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c spin_lock_bh(&priv->tx.tx_lock); tx 71 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c if ((priv->tx.queued_cnt < ATH9K_HTC_TX_THRESHOLD) && tx 72 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c (priv->tx.flags & ATH9K_HTC_OP_TX_QUEUES_STOP)) { tx 73 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c priv->tx.flags &= ~ATH9K_HTC_OP_TX_QUEUES_STOP; tx 76 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c spin_unlock_bh(&priv->tx.tx_lock); tx 83 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c spin_lock_bh(&priv->tx.tx_lock); tx 84 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c slot = find_first_zero_bit(priv->tx.tx_slot, MAX_TX_BUF_NUM); tx 86 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c spin_unlock_bh(&priv->tx.tx_lock); tx 89 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c __set_bit(slot, priv->tx.tx_slot); tx 90 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c spin_unlock_bh(&priv->tx.tx_lock); tx 97 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c spin_lock_bh(&priv->tx.tx_lock); tx 98 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c __clear_bit(slot, priv->tx.tx_slot); tx 99 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c spin_unlock_bh(&priv->tx.tx_lock); tx 137 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c epid_queue = &priv->tx.mgmt_ep_queue; tx 139 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c epid_queue = &priv->tx.cab_ep_queue; tx 141 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c epid_queue = &priv->tx.data_be_queue; tx 143 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c epid_queue = &priv->tx.data_bk_queue; tx 145 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c epid_queue = &priv->tx.data_vi_queue; tx 147 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c epid_queue = &priv->tx.data_vo_queue; tx 392 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c spin_lock_bh(&priv->tx.tx_lock); tx 395 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c spin_unlock_bh(&priv->tx.tx_lock); tx 430 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c spin_lock_bh(&priv->tx.tx_lock); tx 432 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c spin_unlock_bh(&priv->tx.tx_lock); tx 504 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c spin_lock_bh(&priv->tx.tx_lock); tx 505 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c if (WARN_ON(--priv->tx.queued_cnt < 0)) tx 506 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c priv->tx.queued_cnt = 0; tx 507 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c spin_unlock_bh(&priv->tx.tx_lock); tx 538 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c spin_lock_bh(&priv->tx.tx_lock); tx 539 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c priv->tx.flags |= ATH9K_HTC_OP_TX_DRAIN; tx 540 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c spin_unlock_bh(&priv->tx.tx_lock); tx 550 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c ath9k_htc_tx_drainq(priv, &priv->tx.mgmt_ep_queue); tx 551 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c ath9k_htc_tx_drainq(priv, &priv->tx.cab_ep_queue); tx 552 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c ath9k_htc_tx_drainq(priv, &priv->tx.data_be_queue); tx 553 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c ath9k_htc_tx_drainq(priv, &priv->tx.data_bk_queue); tx 554 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c ath9k_htc_tx_drainq(priv, &priv->tx.data_vi_queue); tx 555 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c ath9k_htc_tx_drainq(priv, &priv->tx.data_vo_queue); tx 556 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c ath9k_htc_tx_drainq(priv, &priv->tx.tx_failed); tx 568 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c spin_lock_bh(&priv->tx.tx_lock); tx 569 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c priv->tx.flags &= ~ATH9K_HTC_OP_TX_DRAIN; tx 570 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c spin_unlock_bh(&priv->tx.tx_lock); tx 577 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c spin_lock(&priv->tx.tx_lock); tx 578 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c if (priv->tx.flags & ATH9K_HTC_OP_TX_DRAIN) { tx 579 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c spin_unlock(&priv->tx.tx_lock); tx 582 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c spin_unlock(&priv->tx.tx_lock); tx 584 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c ath9k_htc_tx_drainq(priv, &priv->tx.tx_failed); tx 696 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c skb_queue_tail(&priv->tx.tx_failed, skb); tx 757 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c struct ath9k_htc_priv *priv = from_timer(priv, t, tx.cleanup_timer); tx 788 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c ath9k_htc_tx_cleanup_queue(priv, &priv->tx.mgmt_ep_queue); tx 789 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c ath9k_htc_tx_cleanup_queue(priv, &priv->tx.cab_ep_queue); tx 790 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c ath9k_htc_tx_cleanup_queue(priv, &priv->tx.data_be_queue); tx 791 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c ath9k_htc_tx_cleanup_queue(priv, &priv->tx.data_bk_queue); tx 792 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c ath9k_htc_tx_cleanup_queue(priv, &priv->tx.data_vi_queue); tx 793 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c ath9k_htc_tx_cleanup_queue(priv, &priv->tx.data_vo_queue); tx 798 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c mod_timer(&priv->tx.cleanup_timer, tx 804 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c skb_queue_head_init(&priv->tx.mgmt_ep_queue); tx 805 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c skb_queue_head_init(&priv->tx.cab_ep_queue); tx 806 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c skb_queue_head_init(&priv->tx.data_be_queue); tx 807 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c skb_queue_head_init(&priv->tx.data_bk_queue); tx 808 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c skb_queue_head_init(&priv->tx.data_vi_queue); tx 809 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c skb_queue_head_init(&priv->tx.data_vo_queue); tx 810 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c skb_queue_head_init(&priv->tx.tx_failed); tx 345 drivers/net/wireless/ath/ath9k/htc_hst.c if (endpoint->ep_callbacks.tx) { tx 346 drivers/net/wireless/ath/ath9k/htc_hst.c endpoint->ep_callbacks.tx(endpoint->ep_callbacks.priv, tx 94 drivers/net/wireless/ath/ath9k/htc_hst.h void (*tx) (void *, struct sk_buff *, enum htc_endpoint_id, bool txok); tx 1127 drivers/net/wireless/ath/ath9k/hw.h void ar9003_hw_set_chain_masks(struct ath_hw *ah, u8 rx, u8 tx); tx 414 drivers/net/wireless/ath/ath9k/init.c sc->tx.uapsdq = ath_txq_setup(sc, ATH9K_TX_QUEUE_UAPSD, 0); tx 417 drivers/net/wireless/ath/ath9k/init.c sc->tx.txq_map[i] = ath_txq_setup(sc, ATH9K_TX_QUEUE_DATA, i); tx 418 drivers/net/wireless/ath/ath9k/init.c sc->tx.txq_map[i]->mac80211_qnum = i; tx 784 drivers/net/wireless/ath/ath9k/init.c ath_tx_cleanupq(sc, &sc->tx.txq[i]); tx 1109 drivers/net/wireless/ath/ath9k/init.c ath_tx_cleanupq(sc, &sc->tx.txq[i]); tx 32 drivers/net/wireless/ath/ath9k/link.c txq = sc->tx.txq_map[i]; tx 182 drivers/net/wireless/ath/ath9k/link.c txctl.txq = sc->tx.txq_map[IEEE80211_AC_BE]; tx 299 drivers/net/wireless/ath/ath9k/mac.h } tx; tx 317 drivers/net/wireless/ath/ath9k/mac.h #define ds_ctl2 u.tx.ctl2 tx 318 drivers/net/wireless/ath/ath9k/mac.h #define ds_ctl3 u.tx.ctl3 tx 319 drivers/net/wireless/ath/ath9k/mac.h #define ds_ctl4 u.tx.ctl4 tx 320 drivers/net/wireless/ath/ath9k/mac.h #define ds_ctl5 u.tx.ctl5 tx 321 drivers/net/wireless/ath/ath9k/mac.h #define ds_ctl6 u.tx.ctl6 tx 322 drivers/net/wireless/ath/ath9k/mac.h #define ds_ctl7 u.tx.ctl7 tx 323 drivers/net/wireless/ath/ath9k/mac.h #define ds_ctl8 u.tx.ctl8 tx 324 drivers/net/wireless/ath/ath9k/mac.h #define ds_ctl9 u.tx.ctl9 tx 325 drivers/net/wireless/ath/ath9k/mac.h #define ds_ctl10 u.tx.ctl10 tx 326 drivers/net/wireless/ath/ath9k/mac.h #define ds_ctl11 u.tx.ctl11 tx 328 drivers/net/wireless/ath/ath9k/mac.h #define ds_txstatus0 u.tx.status0 tx 329 drivers/net/wireless/ath/ath9k/mac.h #define ds_txstatus1 u.tx.status1 tx 330 drivers/net/wireless/ath/ath9k/mac.h #define ds_txstatus2 u.tx.status2 tx 331 drivers/net/wireless/ath/ath9k/mac.h #define ds_txstatus3 u.tx.status3 tx 332 drivers/net/wireless/ath/ath9k/mac.h #define ds_txstatus4 u.tx.status4 tx 333 drivers/net/wireless/ath/ath9k/mac.h #define ds_txstatus5 u.tx.status5 tx 334 drivers/net/wireless/ath/ath9k/mac.h #define ds_txstatus6 u.tx.status6 tx 335 drivers/net/wireless/ath/ath9k/mac.h #define ds_txstatus7 u.tx.status7 tx 336 drivers/net/wireless/ath/ath9k/mac.h #define ds_txstatus8 u.tx.status8 tx 337 drivers/net/wireless/ath/ath9k/mac.h #define ds_txstatus9 u.tx.status9 tx 805 drivers/net/wireless/ath/ath9k/main.c txctl.txq = sc->tx.txq_map[skb_get_queue_mapping(skb)]; tx 1643 drivers/net/wireless/ath/ath9k/main.c txq = sc->tx.txq_map[queue]; tx 2048 drivers/net/wireless/ath/ath9k/main.c npend = ath9k_has_pending_frames(sc, &sc->tx.txq[i], tx 2664 drivers/net/wireless/ath/ath9k/main.c .tx = ath9k_tx, tx 126 drivers/net/wireless/ath/ath9k/tx99.c txctl.txq = sc->tx.txq_map[IEEE80211_AC_VO]; tx 173 drivers/net/wireless/ath/ath9k/wmi.c spin_lock_bh(&priv->tx.tx_lock); tx 174 drivers/net/wireless/ath/ath9k/wmi.c if (priv->tx.flags & ATH9K_HTC_OP_TX_DRAIN) { tx 175 drivers/net/wireless/ath/ath9k/wmi.c spin_unlock_bh(&priv->tx.tx_lock); tx 178 drivers/net/wireless/ath/ath9k/wmi.c spin_unlock_bh(&priv->tx.tx_lock); tx 264 drivers/net/wireless/ath/ath9k/wmi.c connect.ep_callbacks.tx = ath9k_wmi_ctrl_tx; tx 173 drivers/net/wireless/ath/ath9k/xmit.c txq = sc->tx.txq_map[q]; tx 211 drivers/net/wireless/ath/ath9k/xmit.c if (tid->txq == sc->tx.txq_map[q]) { tx 368 drivers/net/wireless/ath/ath9k/xmit.c spin_lock_bh(&sc->tx.txbuflock); tx 370 drivers/net/wireless/ath/ath9k/xmit.c if (unlikely(list_empty(&sc->tx.txbuf))) { tx 371 drivers/net/wireless/ath/ath9k/xmit.c spin_unlock_bh(&sc->tx.txbuflock); tx 375 drivers/net/wireless/ath/ath9k/xmit.c bf = list_first_entry(&sc->tx.txbuf, struct ath_buf, list); tx 378 drivers/net/wireless/ath/ath9k/xmit.c spin_unlock_bh(&sc->tx.txbuflock); tx 385 drivers/net/wireless/ath/ath9k/xmit.c spin_lock_bh(&sc->tx.txbuflock); tx 386 drivers/net/wireless/ath/ath9k/xmit.c list_add_tail(&bf->list, &sc->tx.txbuf); tx 387 drivers/net/wireless/ath/ath9k/xmit.c spin_unlock_bh(&sc->tx.txbuflock); tx 792 drivers/net/wireless/ath/ath9k/xmit.c frmlen = sc->tx.max_aggr_framelen[q][modeidx][rates[i].idx]; tx 1118 drivers/net/wireless/ath/ath9k/xmit.c cur_ht20 = sc->tx.max_aggr_framelen[queue][MCS_HT20]; tx 1119 drivers/net/wireless/ath/ath9k/xmit.c cur_ht20_sgi = sc->tx.max_aggr_framelen[queue][MCS_HT20_SGI]; tx 1120 drivers/net/wireless/ath/ath9k/xmit.c cur_ht40 = sc->tx.max_aggr_framelen[queue][MCS_HT40]; tx 1121 drivers/net/wireless/ath/ath9k/xmit.c cur_ht40_sgi = sc->tx.max_aggr_framelen[queue][MCS_HT40_SGI]; tx 1374 drivers/net/wireless/ath/ath9k/xmit.c txq == sc->tx.uapsdq) tx 1631 drivers/net/wireless/ath/ath9k/xmit.c struct ath_txq *txq = sc->tx.uapsdq; tx 1649 drivers/net/wireless/ath/ath9k/xmit.c ret = ath_tx_get_tid_subframe(sc, sc->tx.uapsdq, tx 1745 drivers/net/wireless/ath/ath9k/xmit.c struct ath_txq *txq = &sc->tx.txq[axq_qnum]; tx 1756 drivers/net/wireless/ath/ath9k/xmit.c sc->tx.txqsetup |= 1<<axq_qnum; tx 1762 drivers/net/wireless/ath/ath9k/xmit.c return &sc->tx.txq[axq_qnum]; tx 1772 drivers/net/wireless/ath/ath9k/xmit.c BUG_ON(sc->tx.txq[qnum].axq_qnum != qnum); tx 1882 drivers/net/wireless/ath/ath9k/xmit.c if (!sc->tx.txq[i].axq_depth) tx 1885 drivers/net/wireless/ath/ath9k/xmit.c if (ath9k_hw_numtxpending(ah, sc->tx.txq[i].axq_qnum)) tx 1899 drivers/net/wireless/ath/ath9k/xmit.c txq = &sc->tx.txq[i]; tx 1909 drivers/net/wireless/ath/ath9k/xmit.c sc->tx.txqsetup &= ~(1<<txq->axq_qnum); tx 1960 drivers/net/wireless/ath/ath9k/xmit.c txq = sc->tx.txq_map[i]; tx 2294 drivers/net/wireless/ath/ath9k/xmit.c txq = sc->tx.uapsdq; tx 2302 drivers/net/wireless/ath/ath9k/xmit.c if (txq == sc->tx.txq_map[q]) { tx 2490 drivers/net/wireless/ath/ath9k/xmit.c spin_lock_irqsave(&sc->tx.txbuflock, flags); tx 2491 drivers/net/wireless/ath/ath9k/xmit.c list_splice_tail_init(bf_q, &sc->tx.txbuf); tx 2492 drivers/net/wireless/ath/ath9k/xmit.c spin_unlock_irqrestore(&sc->tx.txbuflock, flags); tx 2637 drivers/net/wireless/ath/ath9k/xmit.c ath_tx_processq(sc, &sc->tx.txq[i]); tx 2680 drivers/net/wireless/ath/ath9k/xmit.c txq = &sc->tx.txq[ts.qid]; tx 2763 drivers/net/wireless/ath/ath9k/xmit.c spin_lock_init(&sc->tx.txbuflock); tx 2765 drivers/net/wireless/ath/ath9k/xmit.c error = ath_descdma_setup(sc, &sc->tx.txdma, &sc->tx.txbuf, tx 2804 drivers/net/wireless/ath/ath9k/xmit.c tid->txq = sc->tx.txq_map[acno]; tx 1753 drivers/net/wireless/ath/carl9170/main.c .tx = carl9170_op_tx, tx 144 drivers/net/wireless/ath/hw.c u32 cycles, busy, rx, tx; tx 154 drivers/net/wireless/ath/hw.c tx = REG_READ(ah, AR_TFCNT); tx 169 drivers/net/wireless/ath/hw.c common->cc_ani.tx_frame += tx; tx 174 drivers/net/wireless/ath/hw.c common->cc_survey.tx_frame += tx; tx 1149 drivers/net/wireless/ath/wcn36xx/main.c .tx = wcn36xx_tx, tx 286 drivers/net/wireless/ath/wil6210/cfg80211.c .tx = BIT(IEEE80211_STYPE_ACTION >> 4) | tx 292 drivers/net/wireless/ath/wil6210/cfg80211.c .tx = BIT(IEEE80211_STYPE_ACTION >> 4) | tx 307 drivers/net/wireless/ath/wil6210/cfg80211.c .tx = BIT(IEEE80211_STYPE_ACTION >> 4) | tx 313 drivers/net/wireless/ath/wil6210/cfg80211.c .tx = BIT(IEEE80211_STYPE_ACTION >> 4) | tx 319 drivers/net/wireless/ath/wil6210/cfg80211.c .tx = BIT(IEEE80211_STYPE_ACTION >> 4) | tx 73 drivers/net/wireless/ath/wil6210/debugfs.c &ring->va[idx].tx.enhanced; tx 133 drivers/net/wireless/ath/wil6210/debugfs.c &ring->va[i].tx.legacy; tx 375 drivers/net/wireless/ath/wil6210/debugfs.c offsetof(struct wil6210_mbox_ctl, tx)); tx 1083 drivers/net/wireless/ath/wil6210/debugfs.c bool tx; tx 1096 drivers/net/wireless/ath/wil6210/debugfs.c tx = ring_idx > 0; /* desc ring 0 is reserved for RX */ tx 1103 drivers/net/wireless/ath/wil6210/debugfs.c tx = (ring_idx < WIL6210_MAX_TX_RINGS); tx 1106 drivers/net/wireless/ath/wil6210/debugfs.c ring = tx ? &wil->ring_tx[ring_idx] : &wil->ring_rx; tx 1109 drivers/net/wireless/ath/wil6210/debugfs.c if (tx) tx 1117 drivers/net/wireless/ath/wil6210/debugfs.c if (tx) tx 1129 drivers/net/wireless/ath/wil6210/debugfs.c d = &ring->va[txdesc_idx].tx.legacy; tx 1134 drivers/net/wireless/ath/wil6210/debugfs.c if (tx) { tx 1151 drivers/net/wireless/ath/wil6210/debugfs.c if (tx) tx 1179 drivers/net/wireless/ath/wil6210/debugfs.c bool tx; tx 1189 drivers/net/wireless/ath/wil6210/debugfs.c tx = !sring->is_rx; tx 1192 drivers/net/wireless/ath/wil6210/debugfs.c seq_printf(s, "No %cX status ring\n", tx ? 'T' : 'R'); tx 1198 drivers/net/wireless/ath/wil6210/debugfs.c tx ? 'T' : 'R', status_msg_idx, sring->size); tx 1205 drivers/net/wireless/ath/wil6210/debugfs.c tx ? 'T' : 'R', sring_idx, status_msg_idx); tx 1209 drivers/net/wireless/ath/wil6210/debugfs.c if (!tx && !wil->use_compressed_rx_status) tx 1478 drivers/net/wireless/ath/wil6210/debugfs.c int tx = atomic_xchg(&wil->isr_count_tx, 0); tx 1487 drivers/net/wireless/ath/wil6210/debugfs.c seq_printf(s, "Tx irqs:packets : %8d : %8ld\n", tx, txf - txf_old); tx 522 drivers/net/wireless/ath/wil6210/interrupt.c wil_mbox_ring_le2cpus(&wil->mbox_ctl.tx); tx 535 drivers/net/wireless/ath/wil6210/interrupt.c if (wil->mbox_ctl.tx.entry_size < min_size) { tx 537 drivers/net/wireless/ath/wil6210/interrupt.c wil->mbox_ctl.tx.entry_size); tx 170 drivers/net/wireless/ath/wil6210/txrx.c &vring->va[i].tx.legacy; tx 226 drivers/net/wireless/ath/wil6210/txrx.c &vring->va[vring->swtail].tx.legacy; tx 1825 drivers/net/wireless/ath/wil6210/txrx.c _hdr_desc = &vring->va[i].tx.legacy; tx 1891 drivers/net/wireless/ath/wil6210/txrx.c _desc = &vring->va[i].tx.legacy; tx 2020 drivers/net/wireless/ath/wil6210/txrx.c d = (struct vring_tx_desc *)&vring->va[i].tx.legacy; tx 2021 drivers/net/wireless/ath/wil6210/txrx.c _desc = &vring->va[i].tx.legacy; tx 2063 drivers/net/wireless/ath/wil6210/txrx.c _d = &ring->va[i].tx.legacy; tx 2103 drivers/net/wireless/ath/wil6210/txrx.c _d = &ring->va[i].tx.legacy; tx 2178 drivers/net/wireless/ath/wil6210/txrx.c _d = &ring->va[i].tx.legacy; tx 2469 drivers/net/wireless/ath/wil6210/txrx.c _d = &vring->va[lf].tx.legacy; tx 2481 drivers/net/wireless/ath/wil6210/txrx.c _d = &vring->va[vring->swtail].tx.legacy; tx 497 drivers/net/wireless/ath/wil6210/txrx.h union wil_tx_desc tx; tx 468 drivers/net/wireless/ath/wil6210/txrx_edma.c &ring->va[ring->swtail].tx.enhanced; tx 1235 drivers/net/wireless/ath/wil6210/txrx_edma.c &ring->va[ring->swtail].tx.enhanced; tx 1363 drivers/net/wireless/ath/wil6210/txrx_edma.c &ring->va[i].tx.enhanced; tx 1538 drivers/net/wireless/ath/wil6210/txrx_edma.c &ring->va[i].tx.enhanced; tx 480 drivers/net/wireless/ath/wil6210/wil6210.h struct wil6210_mbox_ring tx; tx 660 drivers/net/wireless/ath/wil6210/wmi.c struct wil6210_mbox_ring *r = &wil->mbox_ctl.tx; tx 721 drivers/net/wireless/ath/wil6210/wmi.c offsetof(struct wil6210_mbox_ctl, tx.tail)); tx 751 drivers/net/wireless/ath/wil6210/wmi.c wil_w(wil, RGF_MBOX + offsetof(struct wil6210_mbox_ctl, tx.head), tx 2182 drivers/net/wireless/atmel/at76c50x-usb.c .tx = at76_mac80211_tx, tx 289 drivers/net/wireless/broadcom/b43/dma.c ring->tx ? "TX" : "RX", ring->index); tx 304 drivers/net/wireless/broadcom/b43/dma.c B43_WARN_ON(!ring->tx); tx 348 drivers/net/wireless/broadcom/b43/dma.c unsigned char *buf, size_t len, int tx) tx 352 drivers/net/wireless/broadcom/b43/dma.c if (tx) { tx 365 drivers/net/wireless/broadcom/b43/dma.c dma_addr_t addr, size_t len, int tx) tx 367 drivers/net/wireless/broadcom/b43/dma.c if (tx) { tx 380 drivers/net/wireless/broadcom/b43/dma.c B43_WARN_ON(ring->tx); tx 389 drivers/net/wireless/broadcom/b43/dma.c B43_WARN_ON(ring->tx); tx 399 drivers/net/wireless/broadcom/b43/dma.c if (ring->tx) tx 599 drivers/net/wireless/broadcom/b43/dma.c B43_WARN_ON(ring->tx); tx 681 drivers/net/wireless/broadcom/b43/dma.c if (ring->tx) { tx 755 drivers/net/wireless/broadcom/b43/dma.c if (ring->tx) { tx 786 drivers/net/wireless/broadcom/b43/dma.c B43_WARN_ON(!ring->tx); tx 789 drivers/net/wireless/broadcom/b43/dma.c if (ring->tx) { tx 865 drivers/net/wireless/broadcom/b43/dma.c ring->tx = true; tx 1360 drivers/net/wireless/broadcom/b43/dma.c B43_WARN_ON(!ring->tx); tx 1430 drivers/net/wireless/broadcom/b43/dma.c B43_WARN_ON(!ring->tx); tx 1672 drivers/net/wireless/broadcom/b43/dma.c B43_WARN_ON(ring->tx); tx 1693 drivers/net/wireless/broadcom/b43/dma.c B43_WARN_ON(ring->tx); tx 1709 drivers/net/wireless/broadcom/b43/dma.c B43_WARN_ON(!ring->tx); tx 1715 drivers/net/wireless/broadcom/b43/dma.c B43_WARN_ON(!ring->tx); tx 255 drivers/net/wireless/broadcom/b43/dma.h bool tx; tx 5171 drivers/net/wireless/broadcom/b43/main.c .tx = b43_op_tx, tx 746 drivers/net/wireless/broadcom/b43/phy_lp.c static void lpphy_set_trsw_over(struct b43_wldev *dev, bool tx, bool rx) tx 748 drivers/net/wireless/broadcom/b43/phy_lp.c u16 trsw = (tx << 1) | rx; tx 1839 drivers/net/wireless/broadcom/b43/phy_lp.c static int lpphy_rx_iq_cal(struct b43_wldev *dev, bool noise, bool tx, tx 1875 drivers/net/wireless/broadcom/b43/phy_lp.c tx = true; tx 1880 drivers/net/wireless/broadcom/b43/phy_lp.c lpphy_set_trsw_over(dev, tx, rx); tx 32 drivers/net/wireless/broadcom/b43/radio_2056.c const struct b2056_inittab_entry *tx; tx 3011 drivers/net/wireless/broadcom/b43/radio_2056.c .tx = prefix##_tx, \ tx 10229 drivers/net/wireless/broadcom/b43/radio_2056.c B2056_TX0, pts->tx, pts->tx_length); tx 10231 drivers/net/wireless/broadcom/b43/radio_2056.c B2056_TX1, pts->tx, pts->tx_length); tx 149 drivers/net/wireless/broadcom/b43legacy/dma.c ring->tx ? "TX" : "RX", tx 165 drivers/net/wireless/broadcom/b43legacy/dma.c B43legacy_WARN_ON(!ring->tx); tx 250 drivers/net/wireless/broadcom/b43legacy/dma.c int tx) tx 254 drivers/net/wireless/broadcom/b43legacy/dma.c if (tx) tx 270 drivers/net/wireless/broadcom/b43legacy/dma.c int tx) tx 272 drivers/net/wireless/broadcom/b43legacy/dma.c if (tx) tx 287 drivers/net/wireless/broadcom/b43legacy/dma.c B43legacy_WARN_ON(ring->tx); tx 298 drivers/net/wireless/broadcom/b43legacy/dma.c B43legacy_WARN_ON(ring->tx); tx 450 drivers/net/wireless/broadcom/b43legacy/dma.c B43legacy_WARN_ON(ring->tx); tx 535 drivers/net/wireless/broadcom/b43legacy/dma.c if (ring->tx) { tx 571 drivers/net/wireless/broadcom/b43legacy/dma.c if (ring->tx) { tx 593 drivers/net/wireless/broadcom/b43legacy/dma.c B43legacy_WARN_ON(!ring->tx); tx 596 drivers/net/wireless/broadcom/b43legacy/dma.c if (ring->tx) tx 689 drivers/net/wireless/broadcom/b43legacy/dma.c ring->tx = true; tx 735 drivers/net/wireless/broadcom/b43legacy/dma.c (ring->tx) ? "TX" : "RX", ring->max_used_slots, tx 1076 drivers/net/wireless/broadcom/b43legacy/dma.c B43legacy_WARN_ON(!ring->tx); tx 1134 drivers/net/wireless/broadcom/b43legacy/dma.c B43legacy_WARN_ON(!ring->tx); tx 1345 drivers/net/wireless/broadcom/b43legacy/dma.c B43legacy_WARN_ON(ring->tx); tx 1361 drivers/net/wireless/broadcom/b43legacy/dma.c B43legacy_WARN_ON(!ring->tx); tx 1367 drivers/net/wireless/broadcom/b43legacy/dma.c B43legacy_WARN_ON(!ring->tx); tx 149 drivers/net/wireless/broadcom/b43legacy/dma.h bool tx; tx 3533 drivers/net/wireless/broadcom/b43legacy/main.c .tx = b43legacy_op_tx, tx 1291 drivers/net/wireless/broadcom/b43legacy/phy.c u16 tx) tx 1300 drivers/net/wireless/broadcom/b43legacy/phy.c if (tx == 3) tx 6403 drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c .tx = 0xffff, tx 6408 drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c .tx = 0xffff, tx 6413 drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c .tx = 0xffff, tx 6423 drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c .tx = 0xffff, tx 6428 drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c .tx = 0xffff, tx 305 drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c static void brcmf_p2p_print_actframe(bool tx, void *frame, u32 frame_len) tx 319 drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c (tx) ? "TX" : "RX"); tx 323 drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c (tx) ? "TX" : "RX"); tx 327 drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c (tx) ? "TX" : "RX"); tx 331 drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c (tx) ? "TX" : "RX"); tx 335 drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c (tx) ? "TX" : "RX"); tx 339 drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c (tx) ? "TX" : "RX"); tx 343 drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c (tx) ? "TX" : "RX"); tx 347 drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c (tx) ? "TX" : "RX"); tx 351 drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c (tx) ? "TX" : "RX"); tx 355 drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c (tx) ? "TX" : "RX"); tx 363 drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c (tx) ? "TX" : "RX"); tx 367 drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c (tx) ? "TX" : "RX"); tx 371 drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c (tx) ? "TX" : "RX"); tx 375 drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c (tx) ? "TX" : "RX"); tx 379 drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c (tx) ? "TX" : "RX"); tx 387 drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c (tx) ? "TX" : "RX"); tx 391 drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c (tx) ? "TX" : "RX"); tx 395 drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c (tx) ? "TX" : "RX"); tx 399 drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c (tx) ? "TX" : "RX"); tx 403 drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c (tx) ? "TX" : "RX"); tx 411 drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c static void brcmf_p2p_print_actframe(bool tx, void *frame, u32 frame_len) tx 70 drivers/net/wireless/broadcom/brcm80211/brcmsmac/d11.h struct pio2regs tx; tx 82 drivers/net/wireless/broadcom/brcm80211/brcmsmac/d11.h struct pio4regs tx; tx 966 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c .tx = brcms_ops_tx, tx 1152 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c static void wlc_lcnphy_set_trsw_override(struct brcms_phy *pi, bool tx, bool rx) tx 1157 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c (0x1 << 0), (tx ? (0x1 << 1) : 0) | (rx ? (0x1 << 0) : 0)); tx 969 drivers/net/wireless/cisco/airo.c u32 tx; // Tx sequence number tx 1319 drivers/net/wireless/cisco/airo.c cur->tx = 0; /* Tx sequence numbers */ tx 1418 drivers/net/wireless/cisco/airo.c mic->seq = htonl(context->tx); tx 1419 drivers/net/wireless/cisco/airo.c context->tx += 2; tx 307 drivers/net/wireless/intel/iwlegacy/3945-debug.c struct iwl39_stats_tx *tx, *accum_tx, *delta_tx, *max_tx; tx 323 drivers/net/wireless/intel/iwlegacy/3945-debug.c tx = &il->_3945.stats.tx; tx 324 drivers/net/wireless/intel/iwlegacy/3945-debug.c accum_tx = &il->_3945.accum_stats.tx; tx 325 drivers/net/wireless/intel/iwlegacy/3945-debug.c delta_tx = &il->_3945.delta_stats.tx; tx 326 drivers/net/wireless/intel/iwlegacy/3945-debug.c max_tx = &il->_3945.max_delta.tx; tx 336 drivers/net/wireless/intel/iwlegacy/3945-debug.c le32_to_cpu(tx->preamble_cnt), accum_tx->preamble_cnt, tx 341 drivers/net/wireless/intel/iwlegacy/3945-debug.c le32_to_cpu(tx->rx_detected_cnt), tx 347 drivers/net/wireless/intel/iwlegacy/3945-debug.c le32_to_cpu(tx->bt_prio_defer_cnt), tx 353 drivers/net/wireless/intel/iwlegacy/3945-debug.c le32_to_cpu(tx->bt_prio_kill_cnt), tx 359 drivers/net/wireless/intel/iwlegacy/3945-debug.c le32_to_cpu(tx->few_bytes_cnt), accum_tx->few_bytes_cnt, tx 364 drivers/net/wireless/intel/iwlegacy/3945-debug.c le32_to_cpu(tx->cts_timeout), accum_tx->cts_timeout, tx 369 drivers/net/wireless/intel/iwlegacy/3945-debug.c le32_to_cpu(tx->ack_timeout), accum_tx->ack_timeout, tx 374 drivers/net/wireless/intel/iwlegacy/3945-debug.c le32_to_cpu(tx->expected_ack_cnt), tx 380 drivers/net/wireless/intel/iwlegacy/3945-debug.c le32_to_cpu(tx->actual_ack_cnt), accum_tx->actual_ack_cnt, tx 1493 drivers/net/wireless/intel/iwlegacy/3945-mac.c il->isr_stats.tx++; tx 3445 drivers/net/wireless/intel/iwlegacy/3945-mac.c .tx = il3945_mac_tx, tx 2423 drivers/net/wireless/intel/iwlegacy/3945.c tx_beacon_cmd->tx.sta_id = il->hw_params.bcast_id; tx 2424 drivers/net/wireless/intel/iwlegacy/3945.c tx_beacon_cmd->tx.stop_time.life_time = TX_CMD_LIFE_TIME_INFINITE; tx 2431 drivers/net/wireless/intel/iwlegacy/3945.c tx_beacon_cmd->tx.len = cpu_to_le16((u16) frame_size); tx 2433 drivers/net/wireless/intel/iwlegacy/3945.c tx_beacon_cmd->tx.rate = rate; tx 2434 drivers/net/wireless/intel/iwlegacy/3945.c tx_beacon_cmd->tx.tx_flags = tx 2438 drivers/net/wireless/intel/iwlegacy/3945.c tx_beacon_cmd->tx.supp_rates[0] = tx 2441 drivers/net/wireless/intel/iwlegacy/3945.c tx_beacon_cmd->tx.supp_rates[1] = (IL_CCK_BASIC_RATES_MASK & 0xF); tx 461 drivers/net/wireless/intel/iwlegacy/4965-debug.c struct stats_tx *tx, *accum_tx, *delta_tx, *max_tx; tx 476 drivers/net/wireless/intel/iwlegacy/4965-debug.c tx = &il->_4965.stats.tx; tx 477 drivers/net/wireless/intel/iwlegacy/4965-debug.c accum_tx = &il->_4965.accum_stats.tx; tx 478 drivers/net/wireless/intel/iwlegacy/4965-debug.c delta_tx = &il->_4965.delta_stats.tx; tx 479 drivers/net/wireless/intel/iwlegacy/4965-debug.c max_tx = &il->_4965.max_delta.tx; tx 485 drivers/net/wireless/intel/iwlegacy/4965-debug.c le32_to_cpu(tx->preamble_cnt), accum_tx->preamble_cnt, tx 489 drivers/net/wireless/intel/iwlegacy/4965-debug.c le32_to_cpu(tx->rx_detected_cnt), tx 494 drivers/net/wireless/intel/iwlegacy/4965-debug.c le32_to_cpu(tx->bt_prio_defer_cnt), tx 499 drivers/net/wireless/intel/iwlegacy/4965-debug.c le32_to_cpu(tx->bt_prio_kill_cnt), tx 504 drivers/net/wireless/intel/iwlegacy/4965-debug.c le32_to_cpu(tx->few_bytes_cnt), accum_tx->few_bytes_cnt, tx 508 drivers/net/wireless/intel/iwlegacy/4965-debug.c le32_to_cpu(tx->cts_timeout), accum_tx->cts_timeout, tx 512 drivers/net/wireless/intel/iwlegacy/4965-debug.c le32_to_cpu(tx->ack_timeout), accum_tx->ack_timeout, tx 516 drivers/net/wireless/intel/iwlegacy/4965-debug.c le32_to_cpu(tx->expected_ack_cnt), tx 521 drivers/net/wireless/intel/iwlegacy/4965-debug.c le32_to_cpu(tx->actual_ack_cnt), accum_tx->actual_ack_cnt, tx 525 drivers/net/wireless/intel/iwlegacy/4965-debug.c le32_to_cpu(tx->dump_msdu_cnt), accum_tx->dump_msdu_cnt, tx 530 drivers/net/wireless/intel/iwlegacy/4965-debug.c le32_to_cpu(tx->burst_abort_next_frame_mismatch_cnt), tx 537 drivers/net/wireless/intel/iwlegacy/4965-debug.c le32_to_cpu(tx->burst_abort_missing_next_frame_cnt), tx 544 drivers/net/wireless/intel/iwlegacy/4965-debug.c le32_to_cpu(tx->cts_timeout_collision), tx 551 drivers/net/wireless/intel/iwlegacy/4965-debug.c le32_to_cpu(tx->ack_or_ba_timeout_collision), tx 557 drivers/net/wireless/intel/iwlegacy/4965-debug.c le32_to_cpu(tx->agg.ba_timeout), accum_tx->agg.ba_timeout, tx 562 drivers/net/wireless/intel/iwlegacy/4965-debug.c le32_to_cpu(tx->agg.ba_reschedule_frames), tx 569 drivers/net/wireless/intel/iwlegacy/4965-debug.c le32_to_cpu(tx->agg.scd_query_agg_frame_cnt), tx 576 drivers/net/wireless/intel/iwlegacy/4965-debug.c le32_to_cpu(tx->agg.scd_query_no_agg), tx 582 drivers/net/wireless/intel/iwlegacy/4965-debug.c le32_to_cpu(tx->agg.scd_query_agg), tx 588 drivers/net/wireless/intel/iwlegacy/4965-debug.c le32_to_cpu(tx->agg.scd_query_mismatch), tx 594 drivers/net/wireless/intel/iwlegacy/4965-debug.c le32_to_cpu(tx->agg.frame_not_ready), tx 600 drivers/net/wireless/intel/iwlegacy/4965-debug.c le32_to_cpu(tx->agg.underrun), accum_tx->agg.underrun, tx 604 drivers/net/wireless/intel/iwlegacy/4965-debug.c le32_to_cpu(tx->agg.bt_prio_kill), tx 609 drivers/net/wireless/intel/iwlegacy/4965-debug.c le32_to_cpu(tx->agg.rx_ba_rsp_cnt), tx 1770 drivers/net/wireless/intel/iwlegacy/4965-mac.c tx_cmd = &out_cmd->cmd.tx; tx 3813 drivers/net/wireless/intel/iwlegacy/4965-mac.c tx_beacon_cmd->tx.len = cpu_to_le16((u16) frame_size); tx 3814 drivers/net/wireless/intel/iwlegacy/4965-mac.c tx_beacon_cmd->tx.sta_id = il->hw_params.bcast_id; tx 3815 drivers/net/wireless/intel/iwlegacy/4965-mac.c tx_beacon_cmd->tx.stop_time.life_time = TX_CMD_LIFE_TIME_INFINITE; tx 3816 drivers/net/wireless/intel/iwlegacy/4965-mac.c tx_beacon_cmd->tx.tx_flags = tx 3830 drivers/net/wireless/intel/iwlegacy/4965-mac.c tx_beacon_cmd->tx.rate_n_flags = cpu_to_le32(rate | rate_flags); tx 4495 drivers/net/wireless/intel/iwlegacy/4965-mac.c il->isr_stats.tx++; tx 6314 drivers/net/wireless/intel/iwlegacy/4965-mac.c .tx = il4965_mac_tx, tx 2663 drivers/net/wireless/intel/iwlegacy/commands.h struct il3945_tx_cmd tx; tx 2671 drivers/net/wireless/intel/iwlegacy/commands.h struct il_tx_cmd tx; tx 2980 drivers/net/wireless/intel/iwlegacy/commands.h struct iwl39_stats_tx tx; tx 2987 drivers/net/wireless/intel/iwlegacy/commands.h struct stats_tx tx; tx 556 drivers/net/wireless/intel/iwlegacy/common.h struct il_tx_cmd tx; tx 1017 drivers/net/wireless/intel/iwlegacy/common.h u32 tx; tx 704 drivers/net/wireless/intel/iwlegacy/debug.c il->isr_stats.tx); tx 2474 drivers/net/wireless/intel/iwlwifi/dvm/commands.h struct iwl_tx_cmd tx; tx 2760 drivers/net/wireless/intel/iwlwifi/dvm/commands.h struct statistics_tx tx; tx 2767 drivers/net/wireless/intel/iwlwifi/dvm/commands.h struct statistics_tx tx; tx 3835 drivers/net/wireless/intel/iwlwifi/dvm/commands.h u8 tx[IWLAGN_MIC_KEY_SIZE]; tx 3848 drivers/net/wireless/intel/iwlwifi/dvm/commands.h struct iwlagn_p1k_cache tx; tx 1112 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c struct statistics_tx *tx, *accum_tx, *delta_tx, *max_tx; tx 1127 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c tx = &priv->statistics.tx; tx 1128 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c accum_tx = &priv->accum_stats.tx; tx 1129 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c delta_tx = &priv->delta_stats.tx; tx 1130 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c max_tx = &priv->max_delta_stats.tx; tx 1137 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c le32_to_cpu(tx->preamble_cnt), tx 1142 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c le32_to_cpu(tx->rx_detected_cnt), tx 1147 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c le32_to_cpu(tx->bt_prio_defer_cnt), tx 1153 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c le32_to_cpu(tx->bt_prio_kill_cnt), tx 1159 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c le32_to_cpu(tx->few_bytes_cnt), tx 1164 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c le32_to_cpu(tx->cts_timeout), accum_tx->cts_timeout, tx 1168 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c le32_to_cpu(tx->ack_timeout), tx 1173 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c le32_to_cpu(tx->expected_ack_cnt), tx 1179 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c le32_to_cpu(tx->actual_ack_cnt), tx 1185 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c le32_to_cpu(tx->dump_msdu_cnt), tx 1191 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c le32_to_cpu(tx->burst_abort_next_frame_mismatch_cnt), tx 1197 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c le32_to_cpu(tx->burst_abort_missing_next_frame_cnt), tx 1203 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c le32_to_cpu(tx->cts_timeout_collision), tx 1209 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c le32_to_cpu(tx->ack_or_ba_timeout_collision), tx 1215 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c le32_to_cpu(tx->agg.ba_timeout), tx 1221 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c le32_to_cpu(tx->agg.ba_reschedule_frames), tx 1227 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c le32_to_cpu(tx->agg.scd_query_agg_frame_cnt), tx 1233 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c le32_to_cpu(tx->agg.scd_query_no_agg), tx 1239 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c le32_to_cpu(tx->agg.scd_query_agg), tx 1245 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c le32_to_cpu(tx->agg.scd_query_mismatch), tx 1251 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c le32_to_cpu(tx->agg.frame_not_ready), tx 1257 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c le32_to_cpu(tx->agg.underrun), tx 1262 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c le32_to_cpu(tx->agg.bt_prio_kill), tx 1268 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c le32_to_cpu(tx->agg.rx_ba_rsp_cnt), tx 1273 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c if (tx->tx_power.ant_a || tx->tx_power.ant_b || tx->tx_power.ant_c) { tx 1277 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c tx->tx_power.ant_a) tx 1280 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c tx->tx_power.ant_a); tx 1282 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c tx->tx_power.ant_b) tx 1285 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c tx->tx_power.ant_b); tx 1287 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c tx->tx_power.ant_c) tx 1290 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c tx->tx_power.ant_c); tx 771 drivers/net/wireless/intel/iwlwifi/dvm/dev.h struct statistics_tx tx; tx 785 drivers/net/wireless/intel/iwlwifi/dvm/dev.h struct statistics_tx tx; tx 923 drivers/net/wireless/intel/iwlwifi/dvm/lib.c iwlagn_convert_p1k(p1k, data->tkip->tx.p1k); tx 925 drivers/net/wireless/intel/iwlwifi/dvm/lib.c memcpy(data->tkip->mic_keys.tx, tx 1576 drivers/net/wireless/intel/iwlwifi/dvm/mac80211.c .tx = iwlagn_mac_tx, tx 222 drivers/net/wireless/intel/iwlwifi/dvm/main.c tx_beacon_cmd->tx.len = cpu_to_le16((u16)frame_size); tx 223 drivers/net/wireless/intel/iwlwifi/dvm/main.c tx_beacon_cmd->tx.sta_id = priv->beacon_ctx->bcast_sta_id; tx 224 drivers/net/wireless/intel/iwlwifi/dvm/main.c tx_beacon_cmd->tx.stop_time.life_time = TX_CMD_LIFE_TIME_INFINITE; tx 225 drivers/net/wireless/intel/iwlwifi/dvm/main.c tx_beacon_cmd->tx.tx_flags = TX_CMD_FLG_SEQ_CTL_MSK | tx 256 drivers/net/wireless/intel/iwlwifi/dvm/main.c tx_beacon_cmd->tx.rate_n_flags = tx 215 drivers/net/wireless/intel/iwlwifi/dvm/rx.c struct statistics_tx *tx, tx 312 drivers/net/wireless/intel/iwlwifi/dvm/rx.c struct statistics_tx *tx, tx 328 drivers/net/wireless/intel/iwlwifi/dvm/rx.c ACCUM(tx); tx 341 drivers/net/wireless/intel/iwlwifi/dvm/rx.c struct statistics_tx *tx, tx 361 drivers/net/wireless/intel/iwlwifi/dvm/rx.c struct statistics_tx *tx; tx 378 drivers/net/wireless/intel/iwlwifi/dvm/rx.c tx = &stats->tx; tx 396 drivers/net/wireless/intel/iwlwifi/dvm/rx.c tx = &stats->tx; tx 411 drivers/net/wireless/intel/iwlwifi/dvm/rx.c rx_ofdm_ht, rx_cck, tx, bt_activity); tx 413 drivers/net/wireless/intel/iwlwifi/dvm/rx.c iwlagn_recover_from_statistics(priv, rx_ofdm, rx_ofdm_ht, tx, stamp); tx 421 drivers/net/wireless/intel/iwlwifi/dvm/rx.c memcpy(&priv->statistics.tx, tx, sizeof(*tx)); tx 452 drivers/net/wireless/intel/iwlwifi/fw/api/d3.h u8 tx[IWL_MIC_KEY_SIZE]; tx 466 drivers/net/wireless/intel/iwlwifi/fw/api/d3.h struct iwl_p1k_cache tx; tx 421 drivers/net/wireless/intel/iwlwifi/fw/api/stats.h struct mvm_statistics_tx_v4 tx; tx 428 drivers/net/wireless/intel/iwlwifi/fw/api/stats.h struct mvm_statistics_tx_v4 tx; tx 436 drivers/net/wireless/intel/iwlwifi/fw/api/stats.h struct mvm_statistics_tx tx; tx 783 drivers/net/wireless/intel/iwlwifi/fw/api/tx.h struct iwl_tx_cmd tx; tx 802 drivers/net/wireless/intel/iwlwifi/fw/api/tx.h struct iwl_tx_cmd tx; tx 558 drivers/net/wireless/intel/iwlwifi/iwl-trans.h int (*tx)(struct iwl_trans *trans, struct sk_buff *skb, tx 961 drivers/net/wireless/intel/iwlwifi/iwl-trans.h return trans->ops->tx(trans, skb, dev_cmd, queue); tx 264 drivers/net/wireless/intel/iwlwifi/mvm/d3.c iwl_mvm_convert_p1k(p1k, data->tkip->tx.p1k); tx 266 drivers/net/wireless/intel/iwlwifi/mvm/d3.c memcpy(data->tkip->mic_keys.tx, tx 870 drivers/net/wireless/intel/iwlwifi/mvm/mac-ctxt.c struct iwl_tx_cmd *tx) tx 880 drivers/net/wireless/intel/iwlwifi/mvm/mac-ctxt.c tx->len = cpu_to_le16((u16)beacon->len); tx 881 drivers/net/wireless/intel/iwlwifi/mvm/mac-ctxt.c tx->sta_id = mvmvif->bcast_sta.sta_id; tx 882 drivers/net/wireless/intel/iwlwifi/mvm/mac-ctxt.c tx->life_time = cpu_to_le32(TX_CMD_LIFE_TIME_INFINITE); tx 887 drivers/net/wireless/intel/iwlwifi/mvm/mac-ctxt.c tx->tx_flags = cpu_to_le32(tx_flags); tx 893 drivers/net/wireless/intel/iwlwifi/mvm/mac-ctxt.c tx->rate_n_flags = tx 899 drivers/net/wireless/intel/iwlwifi/mvm/mac-ctxt.c tx->rate_n_flags |= cpu_to_le32(iwl_mvm_mac80211_idx_to_hwrate(rate)); tx 901 drivers/net/wireless/intel/iwlwifi/mvm/mac-ctxt.c tx->rate_n_flags |= cpu_to_le32(RATE_MCS_CCK_MSK); tx 931 drivers/net/wireless/intel/iwlwifi/mvm/mac-ctxt.c iwl_mvm_mac_ctxt_set_tx(mvm, vif, beacon, &beacon_cmd.tx); tx 951 drivers/net/wireless/intel/iwlwifi/mvm/mac-ctxt.c iwl_mvm_mac_ctxt_set_tx(mvm, vif, beacon, &beacon_cmd.tx); tx 5009 drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c .tx = iwl_mvm_mac_tx, tx 615 drivers/net/wireless/intel/iwlwifi/mvm/mvm.h } tx; tx 1042 drivers/net/wireless/intel/iwlwifi/mvm/tx.c mdata->tx.airtime += airtime; tx 1057 drivers/net/wireless/intel/iwlwifi/mvm/tx.c mdata->tx.pkts[ac]++; tx 1248 drivers/net/wireless/intel/iwlwifi/mvm/utils.c u32 airtime = mdata->rx.airtime + mdata->tx.airtime; tx 1260 drivers/net/wireless/intel/iwlwifi/mvm/utils.c mdata->tx.pkts[ac]; tx 1271 drivers/net/wireless/intel/iwlwifi/mvm/utils.c memset(&mdata->tx.pkts, 0, sizeof(mdata->tx.pkts)); tx 1282 drivers/net/wireless/intel/iwlwifi/mvm/utils.c memset(&mdata->tx.airtime, 0, sizeof(mdata->tx.airtime)); tx 1392 drivers/net/wireless/intel/iwlwifi/mvm/utils.c memset(&mdata->tx.pkts, 0, sizeof(mdata->tx.pkts)); tx 1394 drivers/net/wireless/intel/iwlwifi/mvm/utils.c memset(&mdata->tx.airtime, 0, sizeof(mdata->tx.airtime)); tx 132 drivers/net/wireless/intel/iwlwifi/pcie/internal.h u32 tx; tx 1950 drivers/net/wireless/intel/iwlwifi/pcie/rx.c isr_stats->tx++; tx 2153 drivers/net/wireless/intel/iwlwifi/pcie/rx.c isr_stats->tx++; tx 2684 drivers/net/wireless/intel/iwlwifi/pcie/trans.c isr_stats->tx); tx 3419 drivers/net/wireless/intel/iwlwifi/pcie/trans.c .tx = iwl_trans_pcie_tx, tx 3446 drivers/net/wireless/intel/iwlwifi/pcie/trans.c .tx = iwl_trans_pcie_gen2_tx, tx 33 drivers/net/wireless/intersil/hostap/hostap.h const struct hfa384x_tx_frame *tx); tx 389 drivers/net/wireless/intersil/hostap/hostap_80211_tx.c struct hostap_tx_data tx; tx 398 drivers/net/wireless/intersil/hostap/hostap_80211_tx.c tx.skb = skb; tx 399 drivers/net/wireless/intersil/hostap/hostap_80211_tx.c tx.sta_ptr = NULL; tx 414 drivers/net/wireless/intersil/hostap/hostap_80211_tx.c tx.crypt = local->crypt_info.crypt[local->crypt_info.tx_keyidx]; tx 415 drivers/net/wireless/intersil/hostap/hostap_80211_tx.c tx.host_encrypt = 1; tx 417 drivers/net/wireless/intersil/hostap/hostap_80211_tx.c tx.crypt = NULL; tx 418 drivers/net/wireless/intersil/hostap/hostap_80211_tx.c tx.host_encrypt = 0; tx 433 drivers/net/wireless/intersil/hostap/hostap_80211_tx.c tx_ret = hostap_handle_sta_tx(local, &tx); tx 434 drivers/net/wireless/intersil/hostap/hostap_80211_tx.c skb = tx.skb; tx 483 drivers/net/wireless/intersil/hostap/hostap_80211_tx.c tx.crypt = NULL; tx 486 drivers/net/wireless/intersil/hostap/hostap_80211_tx.c if (local->ieee_802_1x && meta->ethertype == ETH_P_PAE && tx.crypt && tx 491 drivers/net/wireless/intersil/hostap/hostap_80211_tx.c tx.crypt = NULL; /* no encryption for IEEE 802.1X frames */ tx 494 drivers/net/wireless/intersil/hostap/hostap_80211_tx.c if (tx.crypt && (!tx.crypt->ops || !tx.crypt->ops->encrypt_mpdu)) tx 495 drivers/net/wireless/intersil/hostap/hostap_80211_tx.c tx.crypt = NULL; tx 496 drivers/net/wireless/intersil/hostap/hostap_80211_tx.c else if ((tx.crypt || tx 515 drivers/net/wireless/intersil/hostap/hostap_80211_tx.c if (tx.crypt) { tx 516 drivers/net/wireless/intersil/hostap/hostap_80211_tx.c skb = hostap_tx_encrypt(skb, tx.crypt); tx 535 drivers/net/wireless/intersil/hostap/hostap_80211_tx.c if (local->func->tx == NULL || local->func->tx(skb, dev)) { tx 548 drivers/net/wireless/intersil/hostap/hostap_80211_tx.c if (tx.sta_ptr) tx 549 drivers/net/wireless/intersil/hostap/hostap_80211_tx.c hostap_handle_sta_release(tx.sta_ptr); tx 2682 drivers/net/wireless/intersil/hostap/hostap_ap.c ap_tx_ret hostap_handle_sta_tx(local_info_t *local, struct hostap_tx_data *tx) tx 2685 drivers/net/wireless/intersil/hostap/hostap_ap.c struct sk_buff *skb = tx->skb; tx 2802 drivers/net/wireless/intersil/hostap/hostap_ap.c sta->crypt && tx->host_encrypt) { tx 2803 drivers/net/wireless/intersil/hostap/hostap_ap.c tx->crypt = sta->crypt; tx 2804 drivers/net/wireless/intersil/hostap/hostap_ap.c tx->sta_ptr = sta; /* hostap_handle_sta_release() will tx 236 drivers/net/wireless/intersil/hostap/hostap_ap.h ap_tx_ret hostap_handle_sta_tx(local_info_t *local, struct hostap_tx_data *tx); tx 3100 drivers/net/wireless/intersil/hostap/hostap_hw.c local->func->tx = prism2_tx_80211; tx 549 drivers/net/wireless/intersil/hostap/hostap_main.c void hostap_dump_tx_header(const char *name, const struct hfa384x_tx_frame *tx) tx 555 drivers/net/wireless/intersil/hostap/hostap_main.c name, __le16_to_cpu(tx->status), tx->retry_count, tx->tx_rate, tx 556 drivers/net/wireless/intersil/hostap/hostap_main.c __le16_to_cpu(tx->tx_control), jiffies); tx 558 drivers/net/wireless/intersil/hostap/hostap_main.c fc = __le16_to_cpu(tx->frame_control); tx 563 drivers/net/wireless/intersil/hostap/hostap_main.c __le16_to_cpu(tx->duration_id), __le16_to_cpu(tx->seq_ctrl), tx 564 drivers/net/wireless/intersil/hostap/hostap_main.c __le16_to_cpu(tx->data_len), tx 569 drivers/net/wireless/intersil/hostap/hostap_main.c tx->addr1, tx->addr2, tx->addr3, tx->addr4); tx 572 drivers/net/wireless/intersil/hostap/hostap_main.c tx->dst_addr, tx->src_addr, tx 573 drivers/net/wireless/intersil/hostap/hostap_main.c __be16_to_cpu(tx->len)); tx 600 drivers/net/wireless/intersil/hostap/hostap_wlan.h int (*tx)(struct sk_buff *skb, struct net_device *dev); tx 313 drivers/net/wireless/intersil/p54/main.c priv->survey_raw.tx = 0; tx 706 drivers/net/wireless/intersil/p54/main.c .tx = p54_tx_80211, tx 167 drivers/net/wireless/intersil/p54/p54.h void (*tx)(struct ieee80211_hw *dev, struct sk_buff *skb); tx 212 drivers/net/wireless/intersil/p54/p54.h u64 tx; tx 615 drivers/net/wireless/intersil/p54/p54pci.c priv->common.tx = p54p_tx; tx 645 drivers/net/wireless/intersil/p54/p54spi.c priv->common.tx = p54spi_op_tx; tx 1040 drivers/net/wireless/intersil/p54/p54usb.c priv->common.tx = p54u_tx_lm87; tx 1046 drivers/net/wireless/intersil/p54/p54usb.c priv->common.tx = p54u_tx_net2280; tx 163 drivers/net/wireless/intersil/p54/txrx.c priv->tx(priv->hw, skb); tx 522 drivers/net/wireless/intersil/p54/txrx.c unsigned int i, rssi, tx, cca, dtime, dtotal, dcca, dtx, drssi, unit; tx 552 drivers/net/wireless/intersil/p54/txrx.c tx = le32_to_cpu(stats->sample_tx); tx 559 drivers/net/wireless/intersil/p54/txrx.c dtx = tx - priv->survey_raw.cached_tx; tx 577 drivers/net/wireless/intersil/p54/txrx.c priv->survey_raw.tx += dtx * unit; tx 578 drivers/net/wireless/intersil/p54/txrx.c priv->survey_raw.cached_tx = tx; tx 597 drivers/net/wireless/intersil/p54/txrx.c survey->time_tx = priv->survey_raw.tx; tx 598 drivers/net/wireless/intersil/p54/txrx.c survey->time_busy = priv->survey_raw.tx + tx 2337 drivers/net/wireless/mac80211_hwsim.c .tx = mac80211_hwsim_tx, \ tx 476 drivers/net/wireless/marvell/libertas_tf/main.c .tx = lbtf_op_tx, tx 1668 drivers/net/wireless/marvell/mwifiex/cfg80211.c .tx = BIT(IEEE80211_STYPE_ACTION >> 4) | tx 1674 drivers/net/wireless/marvell/mwifiex/cfg80211.c .tx = BIT(IEEE80211_STYPE_ACTION >> 4) | tx 1680 drivers/net/wireless/marvell/mwifiex/cfg80211.c .tx = BIT(IEEE80211_STYPE_ACTION >> 4) | tx 1686 drivers/net/wireless/marvell/mwifiex/cfg80211.c .tx = BIT(IEEE80211_STYPE_ACTION >> 4) | tx 1675 drivers/net/wireless/marvell/mwl8k.c int tx; tx 1687 drivers/net/wireless/marvell/mwl8k.c tx = txq->head; tx 1688 drivers/net/wireless/marvell/mwl8k.c tx_desc = txq->txd + tx; tx 1699 drivers/net/wireless/marvell/mwl8k.c txq->head = (tx + 1) % MWL8K_TX_DESCS; tx 1706 drivers/net/wireless/marvell/mwl8k.c skb = txq->skb[tx]; tx 1707 drivers/net/wireless/marvell/mwl8k.c txq->skb[tx] = NULL; tx 1903 drivers/net/wireless/marvell/mwl8k.c struct mwl8k_tx_desc *tx; tx 2090 drivers/net/wireless/marvell/mwl8k.c tx = txq->txd + txq->tail; tx 2091 drivers/net/wireless/marvell/mwl8k.c tx->data_rate = txdatarate; tx 2092 drivers/net/wireless/marvell/mwl8k.c tx->tx_priority = txpriority; tx 2093 drivers/net/wireless/marvell/mwl8k.c tx->qos_control = cpu_to_le16(qos); tx 2094 drivers/net/wireless/marvell/mwl8k.c tx->pkt_phys_addr = cpu_to_le32(dma); tx 2095 drivers/net/wireless/marvell/mwl8k.c tx->pkt_len = cpu_to_le16(skb->len); tx 2096 drivers/net/wireless/marvell/mwl8k.c tx->rate_info = 0; tx 2098 drivers/net/wireless/marvell/mwl8k.c tx->peer_id = MWL8K_STA(sta)->peer_id; tx 2100 drivers/net/wireless/marvell/mwl8k.c tx->peer_id = 0; tx 2103 drivers/net/wireless/marvell/mwl8k.c tx->timestamp = cpu_to_le32(ioread32(priv->regs + tx 2106 drivers/net/wireless/marvell/mwl8k.c tx->timestamp = 0; tx 2109 drivers/net/wireless/marvell/mwl8k.c tx->status = cpu_to_le32(MWL8K_TXD_STATUS_FW_OWNED | txstatus); tx 3537 drivers/net/wireless/marvell/mwl8k.c static int mwl8k_cmd_mimo_config(struct ieee80211_hw *hw, __u8 rx, __u8 tx) tx 3550 drivers/net/wireless/marvell/mwl8k.c cmd->tx_antenna_map = tx; tx 5605 drivers/net/wireless/marvell/mwl8k.c .tx = mwl8k_tx, tx 657 drivers/net/wireless/mediatek/mt76/mt7603/main.c .tx = mt7603_tx, tx 493 drivers/net/wireless/mediatek/mt76/mt7615/main.c .tx = mt7615_tx, tx 74 drivers/net/wireless/mediatek/mt76/mt76x0/pci.c .tx = mt76x02_tx, tx 120 drivers/net/wireless/mediatek/mt76/mt76x0/usb.c .tx = mt76x02_tx, tx 157 drivers/net/wireless/mediatek/mt76/mt76x2/pci_main.c .tx = mt76x02_tx, tx 104 drivers/net/wireless/mediatek/mt76/mt76x2/usb_main.c .tx = mt76x02_tx, tx 409 drivers/net/wireless/mediatek/mt7601u/main.c .tx = mt7601u_tx, tx 56 drivers/net/wireless/quantenna/qtnfmac/cfg80211.c .tx = BIT(IEEE80211_STYPE_ACTION >> 4) | tx 63 drivers/net/wireless/quantenna/qtnfmac/cfg80211.c .tx = BIT(IEEE80211_STYPE_ACTION >> 4), tx 401 drivers/net/wireless/ralink/rt2x00/rt2400pci.c ant->tx == ANTENNA_SW_DIVERSITY); tx 409 drivers/net/wireless/ralink/rt2x00/rt2400pci.c switch (ant->tx) { tx 766 drivers/net/wireless/ralink/rt2x00/rt2400pci.c rt2x00_set_field32(®, TXCSR2_TXD_SIZE, rt2x00dev->tx[0].desc_size); tx 767 drivers/net/wireless/ralink/rt2x00/rt2400pci.c rt2x00_set_field32(®, TXCSR2_NUM_TXD, rt2x00dev->tx[1].limit); tx 769 drivers/net/wireless/ralink/rt2x00/rt2400pci.c rt2x00_set_field32(®, TXCSR2_NUM_PRIO, rt2x00dev->tx[0].limit); tx 772 drivers/net/wireless/ralink/rt2x00/rt2400pci.c entry_priv = rt2x00dev->tx[1].entries[0].priv_data; tx 778 drivers/net/wireless/ralink/rt2x00/rt2400pci.c entry_priv = rt2x00dev->tx[0].entries[0].priv_data; tx 1492 drivers/net/wireless/ralink/rt2x00/rt2400pci.c rt2x00dev->default_ant.tx = tx 1503 drivers/net/wireless/ralink/rt2x00/rt2400pci.c if (rt2x00dev->default_ant.tx == ANTENNA_SW_DIVERSITY) tx 1504 drivers/net/wireless/ralink/rt2x00/rt2400pci.c rt2x00dev->default_ant.tx = ANTENNA_HW_DIVERSITY; tx 1675 drivers/net/wireless/ralink/rt2x00/rt2400pci.c rt2x00dev->tx->cw_min, rt2x00dev->tx->cw_max); tx 1705 drivers/net/wireless/ralink/rt2x00/rt2400pci.c .tx = rt2x00mac_tx, tx 408 drivers/net/wireless/ralink/rt2x00/rt2500pci.c ant->tx == ANTENNA_SW_DIVERSITY); tx 417 drivers/net/wireless/ralink/rt2x00/rt2500pci.c switch (ant->tx) { tx 851 drivers/net/wireless/ralink/rt2x00/rt2500pci.c rt2x00_set_field32(®, TXCSR2_TXD_SIZE, rt2x00dev->tx[0].desc_size); tx 852 drivers/net/wireless/ralink/rt2x00/rt2500pci.c rt2x00_set_field32(®, TXCSR2_NUM_TXD, rt2x00dev->tx[1].limit); tx 854 drivers/net/wireless/ralink/rt2x00/rt2500pci.c rt2x00_set_field32(®, TXCSR2_NUM_PRIO, rt2x00dev->tx[0].limit); tx 857 drivers/net/wireless/ralink/rt2x00/rt2500pci.c entry_priv = rt2x00dev->tx[1].entries[0].priv_data; tx 863 drivers/net/wireless/ralink/rt2x00/rt2500pci.c entry_priv = rt2x00dev->tx[0].entries[0].priv_data; tx 1651 drivers/net/wireless/ralink/rt2x00/rt2500pci.c rt2x00dev->default_ant.tx = tx 2004 drivers/net/wireless/ralink/rt2x00/rt2500pci.c .tx = rt2x00mac_tx, tx 513 drivers/net/wireless/ralink/rt2x00/rt2500usb.c ant->tx == ANTENNA_SW_DIVERSITY); tx 523 drivers/net/wireless/ralink/rt2x00/rt2500usb.c switch (ant->tx) { tx 1461 drivers/net/wireless/ralink/rt2x00/rt2500usb.c rt2x00dev->default_ant.tx = tx 1472 drivers/net/wireless/ralink/rt2x00/rt2500usb.c if (rt2x00dev->default_ant.tx == ANTENNA_SW_DIVERSITY) tx 1473 drivers/net/wireless/ralink/rt2x00/rt2500usb.c rt2x00dev->default_ant.tx = ANTENNA_HW_DIVERSITY; tx 1797 drivers/net/wireless/ralink/rt2x00/rt2500usb.c .tx = rt2x00mac_tx, tx 9479 drivers/net/wireless/ralink/rt2x00/rt2800lib.c rt2x00dev->default_ant.tx = ANTENNA_A; tx 9483 drivers/net/wireless/ralink/rt2x00/rt2800lib.c rt2x00dev->default_ant.tx = ANTENNA_A; tx 9488 drivers/net/wireless/ralink/rt2x00/rt2800lib.c rt2x00dev->default_ant.tx = ANTENNA_A; tx 9495 drivers/net/wireless/ralink/rt2x00/rt2800lib.c rt2x00dev->default_ant.tx = ANTENNA_HW_DIVERSITY; /* Unused */ tx 665 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c entry_priv = rt2x00dev->tx[0].entries[0].priv_data; tx 669 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c rt2x00dev->tx[0].limit); tx 673 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c entry_priv = rt2x00dev->tx[1].entries[0].priv_data; tx 677 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c rt2x00dev->tx[1].limit); tx 681 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c entry_priv = rt2x00dev->tx[2].entries[0].priv_data; tx 685 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c rt2x00dev->tx[2].limit); tx 689 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c entry_priv = rt2x00dev->tx[3].entries[0].priv_data; tx 693 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c rt2x00dev->tx[3].limit); tx 290 drivers/net/wireless/ralink/rt2x00/rt2800pci.c .tx = rt2x00mac_tx, tx 135 drivers/net/wireless/ralink/rt2x00/rt2800soc.c .tx = rt2x00mac_tx, tx 632 drivers/net/wireless/ralink/rt2x00/rt2800usb.c .tx = rt2x00mac_tx, tx 202 drivers/net/wireless/ralink/rt2x00/rt2x00.h enum antenna tx; tx 962 drivers/net/wireless/ralink/rt2x00/rt2x00.h struct data_queue *tx; tx 1304 drivers/net/wireless/ralink/rt2x00/rt2x00.h if (queue < rt2x00dev->ops->tx_queues && rt2x00dev->tx) tx 1305 drivers/net/wireless/ralink/rt2x00/rt2x00.h return &rt2x00dev->tx[queue]; tx 1480 drivers/net/wireless/ralink/rt2x00/rt2x00.h u32 *tx, u32 *tx_max, u32 *rx, u32 *rx_max); tx 131 drivers/net/wireless/ralink/rt2x00/rt2x00config.c if (config.tx == ANTENNA_SW_DIVERSITY) { tx 134 drivers/net/wireless/ralink/rt2x00/rt2x00config.c if (def->tx == ANTENNA_SW_DIVERSITY) tx 135 drivers/net/wireless/ralink/rt2x00/rt2x00config.c config.tx = ANTENNA_B; tx 137 drivers/net/wireless/ralink/rt2x00/rt2x00config.c config.tx = def->tx; tx 139 drivers/net/wireless/ralink/rt2x00/rt2x00config.c } else if (config.tx == ANTENNA_SW_DIVERSITY) tx 140 drivers/net/wireless/ralink/rt2x00/rt2x00config.c config.tx = active->tx; tx 1154 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c rt2x00dev->tx->limit * tx 1351 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c if (WARN_ON(!rt2x00dev->tx)) tx 1355 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c return rt2x00dev->tx[0].winfo_size + rt2x00dev->tx[0].desc_size; tx 1357 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c return rt2x00dev->tx[0].winfo_size; tx 103 drivers/net/wireless/ralink/rt2x00/rt2x00link.c new_ant.tx = other_antenna; tx 143 drivers/net/wireless/ralink/rt2x00/rt2x00link.c new_ant.tx = (new_ant.tx == ANTENNA_A) ? ANTENNA_B : ANTENNA_A; tx 735 drivers/net/wireless/ralink/rt2x00/rt2x00mac.c if (tx_ant == 3 && def->tx == ANTENNA_SW_DIVERSITY) { tx 745 drivers/net/wireless/ralink/rt2x00/rt2x00mac.c setup.tx = tx_ant; tx 767 drivers/net/wireless/ralink/rt2x00/rt2x00mac.c *tx_ant = active->tx; tx 779 drivers/net/wireless/ralink/rt2x00/rt2x00mac.c u32 *tx, u32 *tx_max, u32 *rx, u32 *rx_max) tx 785 drivers/net/wireless/ralink/rt2x00/rt2x00mac.c *tx += queue->length; tx 1256 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c rt2x00dev->tx = &queue[1]; tx 1286 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c rt2x00dev->tx = NULL; tx 505 drivers/net/wireless/ralink/rt2x00/rt2x00queue.h &(__dev)->tx[(__dev)->ops->tx_queues] tx 552 drivers/net/wireless/ralink/rt2x00/rt2x00queue.h queue_loop(__entry, (__dev)->tx, tx_queue_end(__dev)) tx 563 drivers/net/wireless/ralink/rt2x00/rt2x00queue.h queue_loop(__entry, (__dev)->tx, queue_end(__dev)) tx 601 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c struct data_queue *queue = rt2x00dev->tx; tx 628 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c if (!rt2x00dev->rx->usb_endpoint || !rt2x00dev->tx->usb_endpoint) { tx 716 drivers/net/wireless/ralink/rt2x00/rt61pci.c ant->tx == ANTENNA_SW_DIVERSITY); tx 1337 drivers/net/wireless/ralink/rt2x00/rt61pci.c rt2x00dev->tx[0].limit); tx 1339 drivers/net/wireless/ralink/rt2x00/rt61pci.c rt2x00dev->tx[1].limit); tx 1341 drivers/net/wireless/ralink/rt2x00/rt61pci.c rt2x00dev->tx[2].limit); tx 1343 drivers/net/wireless/ralink/rt2x00/rt61pci.c rt2x00dev->tx[3].limit); tx 1348 drivers/net/wireless/ralink/rt2x00/rt61pci.c rt2x00dev->tx[0].desc_size / 4); tx 1351 drivers/net/wireless/ralink/rt2x00/rt61pci.c entry_priv = rt2x00dev->tx[0].entries[0].priv_data; tx 1357 drivers/net/wireless/ralink/rt2x00/rt61pci.c entry_priv = rt2x00dev->tx[1].entries[0].priv_data; tx 1363 drivers/net/wireless/ralink/rt2x00/rt61pci.c entry_priv = rt2x00dev->tx[2].entries[0].priv_data; tx 1369 drivers/net/wireless/ralink/rt2x00/rt61pci.c entry_priv = rt2x00dev->tx[3].entries[0].priv_data; tx 2081 drivers/net/wireless/ralink/rt2x00/rt61pci.c for (i = 0; i < rt2x00dev->tx->limit; i++) { tx 2439 drivers/net/wireless/ralink/rt2x00/rt61pci.c rt2x00dev->default_ant.tx = tx 2484 drivers/net/wireless/ralink/rt2x00/rt61pci.c rt2x00dev->default_ant.tx = tx 2488 drivers/net/wireless/ralink/rt2x00/rt61pci.c rt2x00dev->default_ant.tx = ANTENNA_SW_DIVERSITY; tx 2870 drivers/net/wireless/ralink/rt2x00/rt61pci.c .tx = rt2x00mac_tx, tx 697 drivers/net/wireless/ralink/rt2x00/rt73usb.c ant->tx == ANTENNA_SW_DIVERSITY); tx 1879 drivers/net/wireless/ralink/rt2x00/rt73usb.c rt2x00dev->default_ant.tx = tx 2293 drivers/net/wireless/ralink/rt2x00/rt73usb.c .tx = rt2x00mac_tx, tx 1601 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c .tx = rtl8180_tx, tx 1379 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c .tx = rtl8187_tx, tx 5891 drivers/net/wireless/realtek/rtl8xxxu/rtl8xxxu_core.c .tx = rtl8xxxu_tx, tx 1879 drivers/net/wireless/realtek/rtlwifi/core.c .tx = rtl_op_tx, tx 10 drivers/net/wireless/realtek/rtw88/hci.h int (*tx)(struct rtw_dev *rtwdev, tx 32 drivers/net/wireless/realtek/rtw88/hci.h return rtwdev->hci.ops->tx(rtwdev, pkt_info, skb); tx 516 drivers/net/wireless/realtek/rtw88/mac80211.c .tx = rtw_ops_tx, tx 1146 drivers/net/wireless/realtek/rtw88/pci.c .tx = rtw_pci_tx, tx 1975 drivers/net/wireless/rsi/rsi_91x_mac80211.c .tx = rsi_mac80211_tx, tx 239 drivers/net/wireless/st/cw1200/bh.c int *tx) tx 328 drivers/net/wireless/st/cw1200/bh.c *tx = 1; tx 423 drivers/net/wireless/st/cw1200/bh.c int rx, tx, term, suspend; tx 458 drivers/net/wireless/st/cw1200/bh.c tx = atomic_xchg(&priv->bh_tx, 0); tx 462 drivers/net/wireless/st/cw1200/bh.c (rx || tx || term || suspend || priv->bh_error); tx 466 drivers/net/wireless/st/cw1200/bh.c rx, tx, term, suspend, priv->bh_error, status); tx 480 drivers/net/wireless/st/cw1200/bh.c if (priv->hw_bufs_used && (!rx || !tx)) { tx 545 drivers/net/wireless/st/cw1200/bh.c tx += pending_tx; tx 553 drivers/net/wireless/st/cw1200/bh.c ret = cw1200_bh_rx_helper(priv, &ctrl_reg, &tx); tx 558 drivers/net/wireless/st/cw1200/bh.c ret = cw1200_bh_rx_helper(priv, &ctrl_reg, &tx); tx 564 drivers/net/wireless/st/cw1200/bh.c tx: tx 565 drivers/net/wireless/st/cw1200/bh.c if (tx) { tx 566 drivers/net/wireless/st/cw1200/bh.c tx = 0; tx 576 drivers/net/wireless/st/cw1200/bh.c pending_tx = tx; tx 583 drivers/net/wireless/st/cw1200/bh.c tx = ret; tx 595 drivers/net/wireless/st/cw1200/bh.c if (tx) tx 596 drivers/net/wireless/st/cw1200/bh.c goto tx; tx 266 drivers/net/wireless/st/cw1200/debug.c d->tx); tx 14 drivers/net/wireless/st/cw1200/debug.h int tx; tx 35 drivers/net/wireless/st/cw1200/debug.h ++priv->debug->tx; tx 211 drivers/net/wireless/st/cw1200/main.c .tx = cw1200_tx, tx 331 drivers/net/wireless/st/cw1200/queue.c struct wsm_tx **tx, tx 349 drivers/net/wireless/st/cw1200/queue.c *tx = (struct wsm_tx *)item->skb->data; tx 352 drivers/net/wireless/st/cw1200/queue.c (*tx)->packet_id = item->packet_id; tx 84 drivers/net/wireless/st/cw1200/queue.h struct wsm_tx **tx, tx 151 drivers/net/wireless/st/cw1200/sta.c int tx = 0; tx 174 drivers/net/wireless/st/cw1200/sta.c tx = 1; tx 182 drivers/net/wireless/st/cw1200/sta.c tx = 1; tx 190 drivers/net/wireless/st/cw1200/sta.c tx = 0; tx 193 drivers/net/wireless/st/cw1200/sta.c if (tx) { tx 898 drivers/net/wireless/st/cw1200/txrx.c struct ieee80211_tx_info *tx = IEEE80211_SKB_CB(skb); tx 920 drivers/net/wireless/st/cw1200/txrx.c tx->flags |= IEEE80211_TX_STAT_ACK; tx 936 drivers/net/wireless/st/cw1200/txrx.c if (tx->status.rates[i].count >= tx_count) { tx 937 drivers/net/wireless/st/cw1200/txrx.c tx->status.rates[i].count = tx_count; tx 940 drivers/net/wireless/st/cw1200/txrx.c tx_count -= tx->status.rates[i].count; tx 941 drivers/net/wireless/st/cw1200/txrx.c if (tx->status.rates[i].flags & IEEE80211_TX_RC_MCS) tx 942 drivers/net/wireless/st/cw1200/txrx.c tx->status.rates[i].flags |= ht_flags; tx 946 drivers/net/wireless/st/cw1200/txrx.c tx->status.rates[i].count = 0; tx 947 drivers/net/wireless/st/cw1200/txrx.c tx->status.rates[i].idx = -1; tx 951 drivers/net/wireless/st/cw1200/txrx.c if (tx->control.hw_key) { tx 952 drivers/net/wireless/st/cw1200/txrx.c skb_trim(skb, skb->len - tx->control.hw_key->icv_len); tx 953 drivers/net/wireless/st/cw1200/txrx.c if (tx->control.hw_key->cipher == WLAN_CIPHER_SUITE_TKIP) tx 1063 drivers/net/wireless/ti/wl1251/acx.h struct acx_tx_statistics tx; tx 101 drivers/net/wireless/ti/wl1251/debugfs.c DEBUGFS_FWSTATS_FILE(tx, internal_desc_overflow, 20, "%u"); tx 241 drivers/net/wireless/ti/wl1251/debugfs.c DEBUGFS_FWSTATS_DEL(tx, internal_desc_overflow); tx 340 drivers/net/wireless/ti/wl1251/debugfs.c DEBUGFS_FWSTATS_ADD(tx, internal_desc_overflow); tx 1360 drivers/net/wireless/ti/wl1251/main.c .tx = wl1251_op_tx, tx 244 drivers/net/wireless/ti/wl12xx/acx.h struct wl12xx_acx_tx_statistics tx; tx 19 drivers/net/wireless/ti/wl12xx/debugfs.c WL12XX_DEBUGFS_FWSTATS_FILE(tx, internal_desc_overflow, "%u"); tx 119 drivers/net/wireless/ti/wl12xx/debugfs.c DEBUGFS_FWSTATS_ADD(tx, internal_desc_overflow); tx 115 drivers/net/wireless/ti/wl12xx/main.c .tx = { tx 1659 drivers/net/wireless/ti/wl12xx/main.c thold = wl->conf.tx.fast_link_thold; tx 1661 drivers/net/wireless/ti/wl12xx/main.c thold = wl->conf.tx.slow_link_thold; tx 270 drivers/net/wireless/ti/wl18xx/acx.h struct wl18xx_acx_tx_stats tx; tx 45 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_prepared_descs, "%u"); tx 46 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_cmplt, "%u"); tx 47 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_template_prepared, "%u"); tx 48 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_data_prepared, "%u"); tx 49 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_template_programmed, "%u"); tx 50 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_data_programmed, "%u"); tx 51 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_burst_programmed, "%u"); tx 52 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_starts, "%u"); tx 53 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_stop, "%u"); tx 54 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_start_templates, "%u"); tx 55 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_start_int_templates, "%u"); tx 56 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_start_fw_gen, "%u"); tx 57 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_start_data, "%u"); tx 58 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_start_null_frame, "%u"); tx 59 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_exch, "%u"); tx 60 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_retry_template, "%u"); tx 61 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_retry_data, "%u"); tx 62 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE_ARRAY(tx, tx_retry_per_rate, tx 64 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_exch_pending, "%u"); tx 65 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_exch_expiry, "%u"); tx 66 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_done_template, "%u"); tx 67 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_done_data, "%u"); tx 68 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_done_int_template, "%u"); tx 69 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_cfe1, "%u"); tx 70 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_cfe2, "%u"); tx 71 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, frag_called, "%u"); tx 72 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, frag_mpdu_alloc_failed, "%u"); tx 73 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, frag_init_called, "%u"); tx 74 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, frag_in_process_called, "%u"); tx 75 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, frag_tkip_called, "%u"); tx 76 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, frag_key_not_found, "%u"); tx 77 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, frag_need_fragmentation, "%u"); tx 78 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, frag_bad_mblk_num, "%u"); tx 79 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, frag_failed, "%u"); tx 80 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, frag_cache_hit, "%u"); tx 81 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, frag_cache_miss, "%u"); tx 436 drivers/net/wireless/ti/wl18xx/debugfs.c DEBUGFS_FWSTATS_ADD(tx, tx_prepared_descs); tx 437 drivers/net/wireless/ti/wl18xx/debugfs.c DEBUGFS_FWSTATS_ADD(tx, tx_cmplt); tx 438 drivers/net/wireless/ti/wl18xx/debugfs.c DEBUGFS_FWSTATS_ADD(tx, tx_template_prepared); tx 439 drivers/net/wireless/ti/wl18xx/debugfs.c DEBUGFS_FWSTATS_ADD(tx, tx_data_prepared); tx 440 drivers/net/wireless/ti/wl18xx/debugfs.c DEBUGFS_FWSTATS_ADD(tx, tx_template_programmed); tx 441 drivers/net/wireless/ti/wl18xx/debugfs.c DEBUGFS_FWSTATS_ADD(tx, tx_data_programmed); tx 442 drivers/net/wireless/ti/wl18xx/debugfs.c DEBUGFS_FWSTATS_ADD(tx, tx_burst_programmed); tx 443 drivers/net/wireless/ti/wl18xx/debugfs.c DEBUGFS_FWSTATS_ADD(tx, tx_starts); tx 444 drivers/net/wireless/ti/wl18xx/debugfs.c DEBUGFS_FWSTATS_ADD(tx, tx_stop); tx 445 drivers/net/wireless/ti/wl18xx/debugfs.c DEBUGFS_FWSTATS_ADD(tx, tx_start_templates); tx 446 drivers/net/wireless/ti/wl18xx/debugfs.c DEBUGFS_FWSTATS_ADD(tx, tx_start_int_templates); tx 447 drivers/net/wireless/ti/wl18xx/debugfs.c DEBUGFS_FWSTATS_ADD(tx, tx_start_fw_gen); tx 448 drivers/net/wireless/ti/wl18xx/debugfs.c DEBUGFS_FWSTATS_ADD(tx, tx_start_data); tx 449 drivers/net/wireless/ti/wl18xx/debugfs.c DEBUGFS_FWSTATS_ADD(tx, tx_start_null_frame); tx 450 drivers/net/wireless/ti/wl18xx/debugfs.c DEBUGFS_FWSTATS_ADD(tx, tx_exch); tx 451 drivers/net/wireless/ti/wl18xx/debugfs.c DEBUGFS_FWSTATS_ADD(tx, tx_retry_template); tx 452 drivers/net/wireless/ti/wl18xx/debugfs.c DEBUGFS_FWSTATS_ADD(tx, tx_retry_data); tx 453 drivers/net/wireless/ti/wl18xx/debugfs.c DEBUGFS_FWSTATS_ADD(tx, tx_retry_per_rate); tx 454 drivers/net/wireless/ti/wl18xx/debugfs.c DEBUGFS_FWSTATS_ADD(tx, tx_exch_pending); tx 455 drivers/net/wireless/ti/wl18xx/debugfs.c DEBUGFS_FWSTATS_ADD(tx, tx_exch_expiry); tx 456 drivers/net/wireless/ti/wl18xx/debugfs.c DEBUGFS_FWSTATS_ADD(tx, tx_done_template); tx 457 drivers/net/wireless/ti/wl18xx/debugfs.c DEBUGFS_FWSTATS_ADD(tx, tx_done_data); tx 458 drivers/net/wireless/ti/wl18xx/debugfs.c DEBUGFS_FWSTATS_ADD(tx, tx_done_int_template); tx 459 drivers/net/wireless/ti/wl18xx/debugfs.c DEBUGFS_FWSTATS_ADD(tx, tx_cfe1); tx 460 drivers/net/wireless/ti/wl18xx/debugfs.c DEBUGFS_FWSTATS_ADD(tx, tx_cfe2); tx 461 drivers/net/wireless/ti/wl18xx/debugfs.c DEBUGFS_FWSTATS_ADD(tx, frag_called); tx 462 drivers/net/wireless/ti/wl18xx/debugfs.c DEBUGFS_FWSTATS_ADD(tx, frag_mpdu_alloc_failed); tx 463 drivers/net/wireless/ti/wl18xx/debugfs.c DEBUGFS_FWSTATS_ADD(tx, frag_init_called); tx 464 drivers/net/wireless/ti/wl18xx/debugfs.c DEBUGFS_FWSTATS_ADD(tx, frag_in_process_called); tx 465 drivers/net/wireless/ti/wl18xx/debugfs.c DEBUGFS_FWSTATS_ADD(tx, frag_tkip_called); tx 466 drivers/net/wireless/ti/wl18xx/debugfs.c DEBUGFS_FWSTATS_ADD(tx, frag_key_not_found); tx 467 drivers/net/wireless/ti/wl18xx/debugfs.c DEBUGFS_FWSTATS_ADD(tx, frag_need_fragmentation); tx 468 drivers/net/wireless/ti/wl18xx/debugfs.c DEBUGFS_FWSTATS_ADD(tx, frag_bad_mblk_num); tx 469 drivers/net/wireless/ti/wl18xx/debugfs.c DEBUGFS_FWSTATS_ADD(tx, frag_failed); tx 470 drivers/net/wireless/ti/wl18xx/debugfs.c DEBUGFS_FWSTATS_ADD(tx, frag_cache_hit); tx 471 drivers/net/wireless/ti/wl18xx/debugfs.c DEBUGFS_FWSTATS_ADD(tx, frag_cache_miss); tx 256 drivers/net/wireless/ti/wl18xx/main.c .tx = { tx 551 drivers/net/wireless/ti/wlcore/acx.c detection->tx_energy_detection = wl->conf.tx.tx_energy_detection; tx 725 drivers/net/wireless/ti/wlcore/acx.c struct conf_tx_rate_class *c = &wl->conf.tx.sta_rc_conf; tx 906 drivers/net/wireless/ti/wlcore/acx.c frag_threshold = wl->conf.tx.frag_threshold; tx 943 drivers/net/wireless/ti/wlcore/acx.c acx->tx_compl_timeout = cpu_to_le16(wl->conf.tx.tx_compl_timeout); tx 944 drivers/net/wireless/ti/wlcore/acx.c acx->tx_compl_threshold = cpu_to_le16(wl->conf.tx.tx_compl_threshold); tx 1538 drivers/net/wireless/ti/wlcore/acx.c acx->max_tx_retry = cpu_to_le16(wl->conf.tx.max_tx_retries); tx 657 drivers/net/wireless/ti/wlcore/cmd.c cmd->ap.aging_period = cpu_to_le16(wl->conf.tx.ap_aging_period); tx 1035 drivers/net/wireless/ti/wlcore/cmd.c cmd->short_retry_limit = wl->conf.tx.tmpl_short_retry_limit; tx 1036 drivers/net/wireless/ti/wlcore/cmd.c cmd->long_retry_limit = wl->conf.tx.tmpl_long_retry_limit; tx 1133 drivers/net/wireless/ti/wlcore/conf.h struct conf_tx_settings tx; tx 247 drivers/net/wireless/ti/wlcore/event.c u32 num_packets = wl->conf.tx.max_tx_retries; tx 468 drivers/net/wireless/ti/wlcore/init.c for (i = 0; i < wl->conf.tx.ac_conf_count; i++) { tx 611 drivers/net/wireless/ti/wlcore/init.c BUG_ON(wl->conf.tx.tid_conf_count != wl->conf.tx.ac_conf_count); tx 612 drivers/net/wireless/ti/wlcore/init.c for (i = 0; i < wl->conf.tx.tid_conf_count; i++) { tx 613 drivers/net/wireless/ti/wlcore/init.c conf_ac = &wl->conf.tx.ac_conf[i]; tx 620 drivers/net/wireless/ti/wlcore/init.c conf_tid = &wl->conf.tx.tid_conf[i]; tx 211 drivers/net/wireless/ti/wlcore/main.c msecs_to_jiffies(wl->conf.tx.tx_watchdog_timeout)); tx 272 drivers/net/wireless/ti/wlcore/main.c wl->conf.tx.tx_watchdog_timeout); tx 283 drivers/net/wireless/ti/wlcore/main.c wl->conf.tx.tx_watchdog_timeout); tx 297 drivers/net/wireless/ti/wlcore/main.c wl->conf.tx.tx_watchdog_timeout, tx 304 drivers/net/wireless/ti/wlcore/main.c wl->conf.tx.tx_watchdog_timeout); tx 2279 drivers/net/wireless/ti/wlcore/main.c wlvif->bitrate_masks[NL80211_BAND_2GHZ] = wl->conf.tx.basic_rate; tx 2280 drivers/net/wireless/ti/wlcore/main.c wlvif->bitrate_masks[NL80211_BAND_5GHZ] = wl->conf.tx.basic_rate_5; tx 6009 drivers/net/wireless/ti/wlcore/main.c .tx = wl1271_op_tx, tx 1346 drivers/net/wireless/zydas/zd1211rw/zd_mac.c .tx = zd_op_tx, tx 885 drivers/net/wireless/zydas/zd1211rw/zd_usb.c struct zd_usb_tx *tx = &usb->tx; tx 888 drivers/net/wireless/zydas/zd1211rw/zd_usb.c atomic_set(&tx->enabled, 0); tx 891 drivers/net/wireless/zydas/zd1211rw/zd_usb.c usb_kill_anchored_urbs(&tx->submitted); tx 893 drivers/net/wireless/zydas/zd1211rw/zd_usb.c spin_lock_irqsave(&tx->lock, flags); tx 894 drivers/net/wireless/zydas/zd1211rw/zd_usb.c WARN_ON(!skb_queue_empty(&tx->submitted_skbs)); tx 895 drivers/net/wireless/zydas/zd1211rw/zd_usb.c WARN_ON(tx->submitted_urbs != 0); tx 896 drivers/net/wireless/zydas/zd1211rw/zd_usb.c tx->submitted_urbs = 0; tx 897 drivers/net/wireless/zydas/zd1211rw/zd_usb.c spin_unlock_irqrestore(&tx->lock, flags); tx 914 drivers/net/wireless/zydas/zd1211rw/zd_usb.c struct zd_usb_tx *tx = &usb->tx; tx 916 drivers/net/wireless/zydas/zd1211rw/zd_usb.c spin_lock_irqsave(&tx->lock, flags); tx 917 drivers/net/wireless/zydas/zd1211rw/zd_usb.c atomic_set(&tx->enabled, 1); tx 918 drivers/net/wireless/zydas/zd1211rw/zd_usb.c tx->submitted_urbs = 0; tx 920 drivers/net/wireless/zydas/zd1211rw/zd_usb.c tx->stopped = 0; tx 921 drivers/net/wireless/zydas/zd1211rw/zd_usb.c spin_unlock_irqrestore(&tx->lock, flags); tx 926 drivers/net/wireless/zydas/zd1211rw/zd_usb.c struct zd_usb_tx *tx = &usb->tx; tx 929 drivers/net/wireless/zydas/zd1211rw/zd_usb.c spin_lock_irqsave(&tx->lock, flags); tx 930 drivers/net/wireless/zydas/zd1211rw/zd_usb.c --tx->submitted_urbs; tx 931 drivers/net/wireless/zydas/zd1211rw/zd_usb.c if (tx->stopped && tx->submitted_urbs <= ZD_USB_TX_LOW) { tx 933 drivers/net/wireless/zydas/zd1211rw/zd_usb.c tx->stopped = 0; tx 935 drivers/net/wireless/zydas/zd1211rw/zd_usb.c spin_unlock_irqrestore(&tx->lock, flags); tx 940 drivers/net/wireless/zydas/zd1211rw/zd_usb.c struct zd_usb_tx *tx = &usb->tx; tx 943 drivers/net/wireless/zydas/zd1211rw/zd_usb.c spin_lock_irqsave(&tx->lock, flags); tx 944 drivers/net/wireless/zydas/zd1211rw/zd_usb.c ++tx->submitted_urbs; tx 945 drivers/net/wireless/zydas/zd1211rw/zd_usb.c if (!tx->stopped && tx->submitted_urbs > ZD_USB_TX_HIGH) { tx 947 drivers/net/wireless/zydas/zd1211rw/zd_usb.c tx->stopped = 1; tx 949 drivers/net/wireless/zydas/zd1211rw/zd_usb.c spin_unlock_irqrestore(&tx->lock, flags); tx 965 drivers/net/wireless/zydas/zd1211rw/zd_usb.c struct zd_usb_tx *tx; tx 974 drivers/net/wireless/zydas/zd1211rw/zd_usb.c tx = &usb->tx; tx 992 drivers/net/wireless/zydas/zd1211rw/zd_usb.c skb_unlink(skb, &usb->tx.submitted_skbs); tx 998 drivers/net/wireless/zydas/zd1211rw/zd_usb.c usb_anchor_urb(urb, &tx->submitted); tx 1025 drivers/net/wireless/zydas/zd1211rw/zd_usb.c struct zd_usb_tx *tx = &usb->tx; tx 1027 drivers/net/wireless/zydas/zd1211rw/zd_usb.c if (!atomic_read(&tx->enabled)) { tx 1042 drivers/net/wireless/zydas/zd1211rw/zd_usb.c skb_queue_tail(&tx->submitted_skbs, skb); tx 1043 drivers/net/wireless/zydas/zd1211rw/zd_usb.c usb_anchor_urb(urb, &tx->submitted); tx 1049 drivers/net/wireless/zydas/zd1211rw/zd_usb.c skb_unlink(skb, &tx->submitted_skbs); tx 1062 drivers/net/wireless/zydas/zd1211rw/zd_usb.c struct zd_usb_tx *tx = &usb->tx; tx 1063 drivers/net/wireless/zydas/zd1211rw/zd_usb.c struct sk_buff_head *q = &tx->submitted_skbs; tx 1087 drivers/net/wireless/zydas/zd1211rw/zd_usb.c container_of(work, struct zd_usb, tx.watchdog_work.work); tx 1088 drivers/net/wireless/zydas/zd1211rw/zd_usb.c struct zd_usb_tx *tx = &usb->tx; tx 1090 drivers/net/wireless/zydas/zd1211rw/zd_usb.c if (!atomic_read(&tx->enabled) || !tx->watchdog_enabled) tx 1103 drivers/net/wireless/zydas/zd1211rw/zd_usb.c queue_delayed_work(zd_workqueue, &tx->watchdog_work, tx 1109 drivers/net/wireless/zydas/zd1211rw/zd_usb.c struct zd_usb_tx *tx = &usb->tx; tx 1111 drivers/net/wireless/zydas/zd1211rw/zd_usb.c if (!tx->watchdog_enabled) { tx 1113 drivers/net/wireless/zydas/zd1211rw/zd_usb.c queue_delayed_work(zd_workqueue, &tx->watchdog_work, tx 1115 drivers/net/wireless/zydas/zd1211rw/zd_usb.c tx->watchdog_enabled = 1; tx 1121 drivers/net/wireless/zydas/zd1211rw/zd_usb.c struct zd_usb_tx *tx = &usb->tx; tx 1123 drivers/net/wireless/zydas/zd1211rw/zd_usb.c if (tx->watchdog_enabled) { tx 1125 drivers/net/wireless/zydas/zd1211rw/zd_usb.c tx->watchdog_enabled = 0; tx 1126 drivers/net/wireless/zydas/zd1211rw/zd_usb.c cancel_delayed_work_sync(&tx->watchdog_work); tx 1189 drivers/net/wireless/zydas/zd1211rw/zd_usb.c struct zd_usb_tx *tx = &usb->tx; tx 1191 drivers/net/wireless/zydas/zd1211rw/zd_usb.c spin_lock_init(&tx->lock); tx 1192 drivers/net/wireless/zydas/zd1211rw/zd_usb.c atomic_set(&tx->enabled, 0); tx 1193 drivers/net/wireless/zydas/zd1211rw/zd_usb.c tx->stopped = 0; tx 1194 drivers/net/wireless/zydas/zd1211rw/zd_usb.c skb_queue_head_init(&tx->submitted_skbs); tx 1195 drivers/net/wireless/zydas/zd1211rw/zd_usb.c init_usb_anchor(&tx->submitted); tx 1196 drivers/net/wireless/zydas/zd1211rw/zd_usb.c tx->submitted_urbs = 0; tx 1197 drivers/net/wireless/zydas/zd1211rw/zd_usb.c tx->watchdog_enabled = 0; tx 1198 drivers/net/wireless/zydas/zd1211rw/zd_usb.c INIT_DELAYED_WORK(&tx->watchdog_work, zd_tx_watchdog_handler); tx 209 drivers/net/wireless/zydas/zd1211rw/zd_usb.h struct zd_usb_tx tx; tx 149 drivers/net/xen-netback/common.h struct xen_netif_tx_back_ring tx; tx 84 drivers/net/xen-netback/interface.c if (RING_HAS_UNCONSUMED_REQUESTS(&queue->tx)) tx 161 drivers/net/xen-netback/netback.c RING_FINAL_CHECK_FOR_REQUESTS(&queue->tx, more_to_do); tx 197 drivers/net/xen-netback/netback.c RING_IDX cons = queue->tx.req_cons; tx 207 drivers/net/xen-netback/netback.c RING_COPY_REQUEST(&queue->tx, cons++, txp); tx 210 drivers/net/xen-netback/netback.c queue->tx.req_cons = cons; tx 228 drivers/net/xen-netback/netback.c RING_IDX cons = queue->tx.req_cons; tx 276 drivers/net/xen-netback/netback.c RING_COPY_REQUEST(&queue->tx, cons + slots, txp); tx 604 drivers/net/xen-netback/netback.c RING_IDX cons = queue->tx.req_cons; tx 613 drivers/net/xen-netback/netback.c RING_COPY_REQUEST(&queue->tx, cons, &extra); tx 615 drivers/net/xen-netback/netback.c queue->tx.req_cons = ++cons; tx 809 drivers/net/xen-netback/netback.c if (queue->tx.sring->req_prod - queue->tx.req_cons > tx 814 drivers/net/xen-netback/netback.c queue->tx.sring->req_prod, queue->tx.req_cons, tx 820 drivers/net/xen-netback/netback.c work_to_do = RING_HAS_UNCONSUMED_REQUESTS(&queue->tx); tx 824 drivers/net/xen-netback/netback.c idx = queue->tx.req_cons; tx 826 drivers/net/xen-netback/netback.c RING_COPY_REQUEST(&queue->tx, idx, &txreq); tx 836 drivers/net/xen-netback/netback.c queue->tx.req_cons = ++idx; tx 844 drivers/net/xen-netback/netback.c idx = queue->tx.req_cons; tx 1018 drivers/net/xen-netback/netback.c queue->tx.req_cons = idx; tx 1379 drivers/net/xen-netback/netback.c RING_IDX i = queue->tx.rsp_prod_pvt; tx 1382 drivers/net/xen-netback/netback.c resp = RING_GET_RESPONSE(&queue->tx, i); tx 1387 drivers/net/xen-netback/netback.c RING_GET_RESPONSE(&queue->tx, ++i)->status = XEN_NETIF_RSP_NULL; tx 1389 drivers/net/xen-netback/netback.c queue->tx.rsp_prod_pvt = ++i; tx 1396 drivers/net/xen-netback/netback.c RING_PUSH_RESPONSES_AND_CHECK_NOTIFY(&queue->tx, notify); tx 1428 drivers/net/xen-netback/netback.c if (likely(RING_HAS_UNCONSUMED_REQUESTS(&queue->tx))) tx 1441 drivers/net/xen-netback/netback.c if (queue->tx.sring) tx 1443 drivers/net/xen-netback/netback.c queue->tx.sring); tx 1465 drivers/net/xen-netback/netback.c BACK_RING_INIT(&queue->tx, txs, XEN_PAGE_SIZE); tx 29 drivers/net/xen-netback/xenbus.c struct xen_netif_tx_back_ring *tx_ring = &queue->tx; tx 117 drivers/net/xen-netfront.c struct xen_netif_tx_front_ring tx; tx 241 drivers/net/xen-netfront.c return (queue->tx.req_prod_pvt - queue->tx.rsp_cons) < tx 384 drivers/net/xen-netfront.c prod = queue->tx.sring->rsp_prod; tx 387 drivers/net/xen-netfront.c for (cons = queue->tx.rsp_cons; cons != prod; cons++) { tx 390 drivers/net/xen-netfront.c txrsp = RING_GET_RESPONSE(&queue->tx, cons); tx 412 drivers/net/xen-netfront.c queue->tx.rsp_cons = prod; tx 414 drivers/net/xen-netfront.c RING_FINAL_CHECK_FOR_RESPONSES(&queue->tx, more_to_do); tx 424 drivers/net/xen-netfront.c struct xen_netif_tx_request *tx; /* Last request */ tx 433 drivers/net/xen-netfront.c struct xen_netif_tx_request *tx; tx 441 drivers/net/xen-netfront.c tx = RING_GET_REQUEST(&queue->tx, queue->tx.req_prod_pvt++); tx 452 drivers/net/xen-netfront.c tx->id = id; tx 453 drivers/net/xen-netfront.c tx->gref = ref; tx 454 drivers/net/xen-netfront.c tx->offset = offset; tx 455 drivers/net/xen-netfront.c tx->size = len; tx 456 drivers/net/xen-netfront.c tx->flags = 0; tx 458 drivers/net/xen-netfront.c info->tx = tx; tx 459 drivers/net/xen-netfront.c info->size += tx->size; tx 475 drivers/net/xen-netfront.c return info.tx; tx 483 drivers/net/xen-netfront.c info->tx->flags |= XEN_NETTXF_more_data; tx 489 drivers/net/xen-netfront.c struct netfront_queue *queue, struct xen_netif_tx_request *tx, tx 496 drivers/net/xen-netfront.c .tx = tx, tx 516 drivers/net/xen-netfront.c return info.tx; tx 569 drivers/net/xen-netfront.c struct xen_netif_tx_request *tx, *first_tx; tx 635 drivers/net/xen-netfront.c first_tx = tx = xennet_make_first_txreq(queue, skb, tx 637 drivers/net/xen-netfront.c offset += tx->size; tx 642 drivers/net/xen-netfront.c len -= tx->size; tx 646 drivers/net/xen-netfront.c tx->flags |= XEN_NETTXF_csum_blank | XEN_NETTXF_data_validated; tx 649 drivers/net/xen-netfront.c tx->flags |= XEN_NETTXF_data_validated; tx 656 drivers/net/xen-netfront.c RING_GET_REQUEST(&queue->tx, queue->tx.req_prod_pvt++); tx 658 drivers/net/xen-netfront.c tx->flags |= XEN_NETTXF_extra_info; tx 672 drivers/net/xen-netfront.c tx = xennet_make_txreqs(queue, tx, skb, page, offset, len); tx 677 drivers/net/xen-netfront.c tx = xennet_make_txreqs(queue, tx, skb, skb_frag_page(frag), tx 685 drivers/net/xen-netfront.c RING_PUSH_REQUESTS_AND_CHECK_NOTIFY(&queue->tx, notify); tx 1415 drivers/net/xen-netfront.c xennet_end_access(queue->tx_ring_ref, queue->tx.sring); tx 1420 drivers/net/xen-netfront.c queue->tx.sring = NULL; tx 1543 drivers/net/xen-netfront.c queue->tx.sring = NULL; tx 1552 drivers/net/xen-netfront.c FRONT_RING_INIT(&queue->tx, txs, XEN_PAGE_SIZE); tx 781 drivers/ntb/test/ntb_perf.c struct dma_async_tx_descriptor *tx; tx 819 drivers/ntb/test/ntb_perf.c tx = dmaengine_prep_dma_memcpy(pthr->dma_chan, unmap->addr[1], tx 821 drivers/ntb/test/ntb_perf.c if (!tx) tx 823 drivers/ntb/test/ntb_perf.c } while (!tx && (try++ < DMA_TRIES)); tx 825 drivers/ntb/test/ntb_perf.c if (!tx) { tx 830 drivers/ntb/test/ntb_perf.c tx->callback = perf_dma_copy_callback; tx 831 drivers/ntb/test/ntb_perf.c tx->callback_param = pthr; tx 832 drivers/ntb/test/ntb_perf.c dma_set_unmap(tx, unmap); tx 834 drivers/ntb/test/ntb_perf.c ret = dma_submit_error(dmaengine_submit(tx)); tx 1014 drivers/pci/controller/dwc/pcie-tegra194.c msg.tx.data = &req; tx 1015 drivers/pci/controller/dwc/pcie-tegra194.c msg.tx.size = sizeof(req); tx 953 drivers/phy/qualcomm/phy-qcom-qmp.c void __iomem *tx; tx 1447 drivers/phy/qualcomm/phy-qcom-qmp.c void __iomem *tx = qphy->tx; tx 1504 drivers/phy/qualcomm/phy-qcom-qmp.c qcom_qmp_phy_configure(tx, cfg->regs, cfg->tx_tbl, cfg->tx_tbl_num); tx 1881 drivers/phy/qualcomm/phy-qcom-qmp.c qphy->tx = of_iomap(np, 0); tx 1882 drivers/phy/qualcomm/phy-qcom-qmp.c if (!qphy->tx) tx 1908 drivers/phy/qualcomm/phy-qcom-qmp.c qphy->tx2 = qphy->tx + QMP_PHY_LEGACY_LANE_STRIDE; tx 353 drivers/pinctrl/pinctrl-mcp23s08.c u8 tx[2]; tx 358 drivers/pinctrl/pinctrl-mcp23s08.c tx[0] = mcp->addr | 0x01; tx 359 drivers/pinctrl/pinctrl-mcp23s08.c tx[1] = *((u8 *) reg); tx 361 drivers/pinctrl/pinctrl-mcp23s08.c return spi_write_then_read(spi, tx, sizeof(tx), val, val_size); tx 1295 drivers/pinctrl/sh-pfc/pfc-r8a7778.c #define CAN_PFC_DATA(name, tx, rx) SH_PFC_MUX2(name, tx, rx) tx 1338 drivers/pinctrl/sh-pfc/pfc-r8a7778.c #define SCIF_PFC_DAT(name, tx, rx) SH_PFC_MUX2(name, tx, rx) tx 1370 drivers/pinctrl/sh-pfc/pfc-r8a7778.c #define HSPI_PFC_DAT(name, clk, cs, rx, tx) SH_PFC_MUX4(name, clk, cs, rx, tx) tx 399 drivers/platform/mellanox/mlxbf-tmfifo.c int rx, tx; tx 402 drivers/platform/mellanox/mlxbf-tmfifo.c tx = !test_and_set_bit(MLXBF_TM_TX_LWM_IRQ, &fifo->pend_events); tx 404 drivers/platform/mellanox/mlxbf-tmfifo.c if (rx || tx) tx 148 drivers/ps3/ps3-vuart.c u64 tx; tx 151 drivers/ps3/ps3-vuart.c PARAM_TX_TRIGGER, &tx); tx 152 drivers/ps3/ps3-vuart.c trig->tx = tx; tx 181 drivers/ps3/ps3-vuart.c trig->tx, trig->rx); tx 186 drivers/ps3/ps3-vuart.c int ps3_vuart_set_triggers(struct ps3_system_bus_device *dev, unsigned int tx, tx 193 drivers/ps3/ps3-vuart.c PARAM_TX_TRIGGER, tx); tx 220 drivers/ps3/ps3-vuart.c tx, rx); tx 75 drivers/ps3/vuart.h unsigned long tx; tx 80 drivers/ps3/vuart.h int ps3_vuart_set_triggers(struct ps3_system_bus_device *dev, unsigned int tx, tx 115 drivers/ptp/ptp_clock.c static int ptp_clock_adjtime(struct posix_clock *pc, struct __kernel_timex *tx) tx 123 drivers/ptp/ptp_clock.c if (tx->modes & ADJ_SETOFFSET) { tx 128 drivers/ptp/ptp_clock.c ts.tv_sec = tx->time.tv_sec; tx 129 drivers/ptp/ptp_clock.c ts.tv_nsec = tx->time.tv_usec; tx 131 drivers/ptp/ptp_clock.c if (!(tx->modes & ADJ_NANO)) tx 140 drivers/ptp/ptp_clock.c } else if (tx->modes & ADJ_FREQUENCY) { tx 141 drivers/ptp/ptp_clock.c s32 ppb = scaled_ppm_to_ppb(tx->freq); tx 145 drivers/ptp/ptp_clock.c err = ops->adjfine(ops, tx->freq); tx 148 drivers/ptp/ptp_clock.c ptp->dialed_frequency = tx->freq; tx 149 drivers/ptp/ptp_clock.c } else if (tx->modes == 0) { tx 150 drivers/ptp/ptp_clock.c tx->freq = ptp->dialed_frequency; tx 700 drivers/rapidio/devices/rio_mport_cdev.c struct dma_async_tx_descriptor *tx; tx 719 drivers/rapidio/devices/rio_mport_cdev.c tx = prep_dma_xfer(chan, xfer, sgt, nents, dir, tx 722 drivers/rapidio/devices/rio_mport_cdev.c if (!tx) { tx 728 drivers/rapidio/devices/rio_mport_cdev.c } else if (IS_ERR(tx)) { tx 729 drivers/rapidio/devices/rio_mport_cdev.c ret = PTR_ERR(tx); tx 736 drivers/rapidio/devices/rio_mport_cdev.c tx->callback = dma_xfer_callback; tx 737 drivers/rapidio/devices/rio_mport_cdev.c tx->callback_param = req; tx 742 drivers/rapidio/devices/rio_mport_cdev.c cookie = dmaengine_submit(tx); tx 30 drivers/reset/tegra/reset-bpmp.c msg.tx.data = &request; tx 31 drivers/reset/tegra/reset-bpmp.c msg.tx.size = sizeof(request); tx 1580 drivers/rpmsg/qcom_glink_native.c struct qcom_glink_pipe *tx, tx 1592 drivers/rpmsg/qcom_glink_native.c glink->tx_pipe = tx; tx 32 drivers/rpmsg/qcom_glink_native.h struct qcom_glink_pipe *tx, tx 186 drivers/rpmsg/qcom_glink_rpm.c struct glink_rpm_pipe *tx) tx 235 drivers/rpmsg/qcom_glink_rpm.c tx->native.length = size; tx 237 drivers/rpmsg/qcom_glink_rpm.c tx->tail = msg_ram + offset; tx 238 drivers/rpmsg/qcom_glink_rpm.c tx->head = msg_ram + offset + sizeof(u32); tx 239 drivers/rpmsg/qcom_glink_rpm.c tx->fifo = msg_ram + offset + 2 * sizeof(u32); tx 244 drivers/rpmsg/qcom_glink_rpm.c if (!rx->fifo || !tx->fifo) { tx 249 drivers/rpmsg/qcom_smd.c struct smd_channel_info tx; tx 271 drivers/rpmsg/qcom_smd.c struct smd_channel_info_word tx; tx 311 drivers/rpmsg/qcom_smd.c BUILD_BUG_ON(sizeof(channel->info->tx.param) != sizeof(u8)); \ tx 313 drivers/rpmsg/qcom_smd.c le32_to_cpu(channel->info_word->tx.param) : \ tx 314 drivers/rpmsg/qcom_smd.c channel->info->tx.param; \ tx 319 drivers/rpmsg/qcom_smd.c BUILD_BUG_ON(sizeof(channel->info->tx.param) != sizeof(u32)); \ tx 321 drivers/rpmsg/qcom_smd.c channel->info_word->tx.param : \ tx 322 drivers/rpmsg/qcom_smd.c channel->info->tx.param); \ tx 327 drivers/rpmsg/qcom_smd.c BUILD_BUG_ON(sizeof(channel->info->tx.param) != sizeof(u8)); \ tx 329 drivers/rpmsg/qcom_smd.c channel->info_word->tx.param = cpu_to_le32(value); \ tx 331 drivers/rpmsg/qcom_smd.c channel->info->tx.param = value; \ tx 336 drivers/rpmsg/qcom_smd.c BUILD_BUG_ON(sizeof(channel->info->tx.param) != sizeof(u32)); \ tx 338 drivers/rpmsg/qcom_smd.c channel->info_word->tx.param = cpu_to_le32(value); \ tx 340 drivers/rpmsg/qcom_smd.c channel->info->tx.param = cpu_to_le32(value); \ tx 495 drivers/rtc/rtc-ds1305.c u8 *addr, size_t count, char *tx, char *rx) tx 506 drivers/rtc/rtc-ds1305.c x->tx_buf = tx; tx 64 drivers/rtc/rtc-mcp795.c u8 tx[2]; tx 66 drivers/rtc/rtc-mcp795.c tx[0] = MCP795_READ; tx 67 drivers/rtc/rtc-mcp795.c tx[1] = addr; tx 68 drivers/rtc/rtc-mcp795.c ret = spi_write_then_read(spi, tx, sizeof(tx), buf, count); tx 81 drivers/rtc/rtc-mcp795.c u8 tx[257]; tx 83 drivers/rtc/rtc-mcp795.c tx[0] = MCP795_WRITE; tx 84 drivers/rtc/rtc-mcp795.c tx[1] = addr; tx 85 drivers/rtc/rtc-mcp795.c memcpy(&tx[2], data, count); tx 87 drivers/rtc/rtc-mcp795.c ret = spi_write(spi, tx, 2 + count); tx 226 drivers/scsi/cxgbi/libcxgbi.h struct cxgbi_skb_tx_cb tx; tx 239 drivers/scsi/cxgbi/libcxgbi.h #define cxgbi_skcb_tx_wr_next(skb) (CXGBI_SKB_CB(skb)->tx.wr_next) tx 229 drivers/scsi/fnic/fnic_main.c stats->tx_frames = vs->tx.tx_unicast_frames_ok; tx 230 drivers/scsi/fnic/fnic_main.c stats->tx_words = vs->tx.tx_unicast_bytes_ok / 4; tx 233 drivers/scsi/fnic/fnic_main.c stats->error_frames = vs->tx.tx_errors + vs->rx.rx_errors; tx 234 drivers/scsi/fnic/fnic_main.c stats->dumped_frames = vs->tx.tx_drops + vs->rx.rx_drop; tx 64 drivers/scsi/fnic/vnic_stats.h struct vnic_tx_stats tx; tx 205 drivers/scsi/mvsas/mv_chips.h static inline void mvs_start_delivery(struct mvs_info *mvi, u32 tx) tx 208 drivers/scsi/mvsas/mv_chips.h mw32(MVS_TX_PROD_IDX, tx); tx 113 drivers/scsi/mvsas/mv_init.c if (mvi->tx) tx 115 drivers/scsi/mvsas/mv_init.c sizeof(*mvi->tx) * MVS_CHIP_SLOT_SZ, tx 116 drivers/scsi/mvsas/mv_init.c mvi->tx, mvi->tx_dma); tx 239 drivers/scsi/mvsas/mv_init.c mvi->tx = dma_alloc_coherent(mvi->dev, tx 240 drivers/scsi/mvsas/mv_init.c sizeof(*mvi->tx) * MVS_CHIP_SLOT_SZ, tx 242 drivers/scsi/mvsas/mv_init.c if (!mvi->tx) tx 244 drivers/scsi/mvsas/mv_init.c memset(mvi->tx, 0, sizeof(*mvi->tx) * MVS_CHIP_SLOT_SZ); tx 379 drivers/scsi/mvsas/mv_sas.c slot->tx = mvi->tx_prod; tx 380 drivers/scsi/mvsas/mv_sas.c mvi->tx[mvi->tx_prod] = cpu_to_le32((TXQ_CMD_SMP << TXQ_CMD_SHIFT) | tx 452 drivers/scsi/mvsas/mv_sas.c slot->tx = mvi->tx_prod; tx 457 drivers/scsi/mvsas/mv_sas.c mvi->tx[mvi->tx_prod] = cpu_to_le32(del_q); tx 580 drivers/scsi/mvsas/mv_sas.c slot->tx = mvi->tx_prod; tx 581 drivers/scsi/mvsas/mv_sas.c mvi->tx[mvi->tx_prod] = cpu_to_le32(TXQ_MODE_I | tag | tx 121 drivers/scsi/mvsas/mv_sas.h void (*start_delivery)(struct mvs_info *mvi, u32 tx); tx 313 drivers/scsi/mvsas/mv_sas.h u32 tx; tx 349 drivers/scsi/mvsas/mv_sas.h __le32 *tx; tx 64 drivers/scsi/snic/vnic_stats.h struct vnic_tx_stats tx; tx 110 drivers/slimbus/qcom-ctrl.c struct slim_ctrl_buf tx; tx 155 drivers/slimbus/qcom-ctrl.c spin_lock_irqsave(&ctrl->tx.lock, flags); tx 156 drivers/slimbus/qcom-ctrl.c idx = ctrl->tx.head; tx 157 drivers/slimbus/qcom-ctrl.c ctrl->tx.head = (ctrl->tx.head + 1) % ctrl->tx.n; tx 158 drivers/slimbus/qcom-ctrl.c spin_unlock_irqrestore(&ctrl->tx.lock, flags); tx 309 drivers/slimbus/qcom-ctrl.c spin_lock_irqsave(&ctrl->tx.lock, flags); tx 310 drivers/slimbus/qcom-ctrl.c if (((ctrl->tx.head + 1) % ctrl->tx.n) == ctrl->tx.tail) { tx 311 drivers/slimbus/qcom-ctrl.c spin_unlock_irqrestore(&ctrl->tx.lock, flags); tx 315 drivers/slimbus/qcom-ctrl.c idx = ctrl->tx.tail; tx 317 drivers/slimbus/qcom-ctrl.c ctrl->tx.tail = (ctrl->tx.tail + 1) % ctrl->tx.n; tx 319 drivers/slimbus/qcom-ctrl.c spin_unlock_irqrestore(&ctrl->tx.lock, flags); tx 321 drivers/slimbus/qcom-ctrl.c return ctrl->tx.base + (idx * ctrl->tx.sl_sz); tx 537 drivers/slimbus/qcom-ctrl.c ctrl->tx.n = QCOM_TX_MSGS; tx 538 drivers/slimbus/qcom-ctrl.c ctrl->tx.sl_sz = SLIM_MSGQ_BUF_LEN; tx 547 drivers/slimbus/qcom-ctrl.c spin_lock_init(&ctrl->tx.lock); tx 578 drivers/slimbus/qcom-ctrl.c ctrl->tx.base = devm_kcalloc(&pdev->dev, ctrl->tx.n, ctrl->tx.sl_sz, tx 580 drivers/slimbus/qcom-ctrl.c if (!ctrl->tx.base) { tx 415 drivers/slimbus/slimbus.h struct slim_msg_txn *tx); tx 141 drivers/soc/fsl/qbman/qman_test_stash.c struct qman_fq tx; tx 286 drivers/soc/fsl/qbman/qman_test_stash.c if (qman_enqueue(&handler->tx, &dqrr->fd)) { tx 302 drivers/soc/fsl/qbman/qman_test_stash.c if (qman_enqueue(&handler->tx, &dqrr->fd)) { tx 366 drivers/soc/fsl/qbman/qman_test_stash.c qman_destroy_fq(&handler->tx); tx 420 drivers/soc/fsl/qbman/qman_test_stash.c memset(&handler->tx, 0, sizeof(handler->tx)); tx 422 drivers/soc/fsl/qbman/qman_test_stash.c &handler->tx); tx 541 drivers/soc/fsl/qbman/qman_test_stash.c err = qman_enqueue(&special_handler->tx, &fd); tx 46 drivers/soc/tegra/powergate-bpmp.c msg.tx.data = &request; tx 47 drivers/soc/tegra/powergate-bpmp.c msg.tx.size = sizeof(request); tx 74 drivers/soc/tegra/powergate-bpmp.c msg.tx.data = &request; tx 75 drivers/soc/tegra/powergate-bpmp.c msg.tx.size = sizeof(request); tx 102 drivers/soc/tegra/powergate-bpmp.c msg.tx.data = &request; tx 103 drivers/soc/tegra/powergate-bpmp.c msg.tx.size = sizeof(request); tx 132 drivers/soc/tegra/powergate-bpmp.c msg.tx.data = &request; tx 133 drivers/soc/tegra/powergate-bpmp.c msg.tx.size = sizeof(request); tx 159 drivers/soc/ti/knav_dma.c if (cfg->u.tx.filt_pswords) tx 161 drivers/soc/ti/knav_dma.c if (cfg->u.tx.filt_einfo) tx 168 drivers/soc/ti/knav_dma.c writel_relaxed(cfg->u.tx.priority, &chan->reg_tx_sched->prio); tx 330 drivers/soc/ti/knav_dma.c chan->cfg.u.tx.filt_einfo, tx 331 drivers/soc/ti/knav_dma.c chan->cfg.u.tx.filt_pswords, tx 332 drivers/soc/ti/knav_dma.c chan->cfg.u.tx.priority); tx 52 drivers/spi/spi-altera.c const unsigned char *tx; tx 80 drivers/spi/spi-altera.c if (hw->tx) { tx 83 drivers/spi/spi-altera.c txd = hw->tx[hw->count]; tx 86 drivers/spi/spi-altera.c txd = (hw->tx[hw->count * 2] tx 87 drivers/spi/spi-altera.c | (hw->tx[hw->count * 2 + 1] << 8)); tx 120 drivers/spi/spi-altera.c hw->tx = t->tx_buf; tx 49 drivers/spi/spi-au1550.c const u8 *tx; tx 316 drivers/spi/spi-au1550.c hw->tx = t->tx_buf; tx 366 drivers/spi/spi-au1550.c hw->tx = hw->rx; tx 375 drivers/spi/spi-au1550.c res = au1xxx_dbdma_put_source(hw->dma_tx_ch, virt_to_phys(hw->tx), tx 485 drivers/spi/spi-au1550.c if (hw->tx) { \ tx 486 drivers/spi/spi-au1550.c fifoword = *(u##size *)hw->tx & (u32)(mask); \ tx 487 drivers/spi/spi-au1550.c hw->tx += (size) / 8; \ tx 508 drivers/spi/spi-au1550.c hw->tx = t->tx_buf; tx 162 drivers/spi/spi-bcm63xx-hsspi.c const u8 *tx = t->tx_buf; tx 168 drivers/spi/spi-bcm63xx-hsspi.c if (tx && rx) tx 170 drivers/spi/spi-bcm63xx-hsspi.c else if (tx) tx 190 drivers/spi/spi-bcm63xx-hsspi.c if (tx) { tx 191 drivers/spi/spi-bcm63xx-hsspi.c memcpy_toio(bs->fifo + HSSPI_OPCODE_LEN, tx, curr_step); tx 192 drivers/spi/spi-bcm63xx-hsspi.c tx += curr_step; tx 66 drivers/spi/spi-bitbang.c const u8 *tx = t->tx_buf; tx 72 drivers/spi/spi-bitbang.c if (tx) tx 73 drivers/spi/spi-bitbang.c word = *tx++; tx 94 drivers/spi/spi-bitbang.c const u16 *tx = t->tx_buf; tx 100 drivers/spi/spi-bitbang.c if (tx) tx 101 drivers/spi/spi-bitbang.c word = *tx++; tx 122 drivers/spi/spi-bitbang.c const u32 *tx = t->tx_buf; tx 128 drivers/spi/spi-bitbang.c if (tx) tx 129 drivers/spi/spi-bitbang.c word = *tx++; tx 43 drivers/spi/spi-cavium-octeon.c p->regs.tx = 0x10; tx 49 drivers/spi/spi-cavium-thunderx.c p->regs.tx = 0x1010; tx 13 drivers/spi/spi-cavium.h int tx; tx 28 drivers/spi/spi-cavium.h #define OCTEON_SPI_TX(x) (x->regs.tx) tx 115 drivers/spi/spi-davinci.c const void *tx; tx 157 drivers/spi/spi-davinci.c if (dspi->tx) { tx 158 drivers/spi/spi-davinci.c const u8 *tx = dspi->tx; tx 160 drivers/spi/spi-davinci.c data = *tx++; tx 161 drivers/spi/spi-davinci.c dspi->tx = tx; tx 170 drivers/spi/spi-davinci.c if (dspi->tx) { tx 171 drivers/spi/spi-davinci.c const u16 *tx = dspi->tx; tx 173 drivers/spi/spi-davinci.c data = *tx++; tx 174 drivers/spi/spi-davinci.c dspi->tx = tx; tx 590 drivers/spi/spi-davinci.c dspi->tx = t->tx_buf; tx 107 drivers/spi/spi-dln2.c } tx; tx 108 drivers/spi/spi-dln2.c unsigned len = sizeof(tx); tx 110 drivers/spi/spi-dln2.c tx.port = dln2->port; tx 114 drivers/spi/spi-dln2.c len -= sizeof(tx.wait_for_completion); tx 116 drivers/spi/spi-dln2.c tx.wait_for_completion = DLN2_TRANSFERS_WAIT_COMPLETE; tx 120 drivers/spi/spi-dln2.c return dln2_transfer_tx(dln2->pdev, cmd, &tx, len); tx 136 drivers/spi/spi-dln2.c } tx; tx 138 drivers/spi/spi-dln2.c tx.port = dln2->port; tx 145 drivers/spi/spi-dln2.c tx.cs = ~cs_mask; tx 147 drivers/spi/spi-dln2.c return dln2_transfer_tx(dln2->pdev, DLN2_SPI_SET_SS, &tx, sizeof(tx)); tx 166 drivers/spi/spi-dln2.c } tx; tx 169 drivers/spi/spi-dln2.c tx.port = dln2->port; tx 170 drivers/spi/spi-dln2.c tx.cs = cs_mask; tx 173 drivers/spi/spi-dln2.c return dln2_transfer_tx(dln2->pdev, cmd, &tx, sizeof(tx)); tx 188 drivers/spi/spi-dln2.c } tx; tx 194 drivers/spi/spi-dln2.c tx.port = dln2->port; tx 195 drivers/spi/spi-dln2.c ret = dln2_transfer(dln2->pdev, DLN2_SPI_GET_SS_COUNT, &tx, sizeof(tx), tx 214 drivers/spi/spi-dln2.c } tx; tx 220 drivers/spi/spi-dln2.c tx.port = dln2->port; tx 222 drivers/spi/spi-dln2.c ret = dln2_transfer(dln2->pdev, cmd, &tx, sizeof(tx), &rx, &rx_len); tx 264 drivers/spi/spi-dln2.c } __packed tx; tx 270 drivers/spi/spi-dln2.c tx.port = dln2->port; tx 271 drivers/spi/spi-dln2.c tx.speed = cpu_to_le32(speed); tx 273 drivers/spi/spi-dln2.c ret = dln2_transfer(dln2->pdev, DLN2_SPI_SET_FREQUENCY, &tx, sizeof(tx), tx 291 drivers/spi/spi-dln2.c } tx; tx 293 drivers/spi/spi-dln2.c tx.port = dln2->port; tx 294 drivers/spi/spi-dln2.c tx.mode = mode; tx 296 drivers/spi/spi-dln2.c return dln2_transfer_tx(dln2->pdev, DLN2_SPI_SET_MODE, &tx, sizeof(tx)); tx 307 drivers/spi/spi-dln2.c } tx; tx 309 drivers/spi/spi-dln2.c tx.port = dln2->port; tx 310 drivers/spi/spi-dln2.c tx.bpw = bpw; tx 313 drivers/spi/spi-dln2.c &tx, sizeof(tx)); tx 322 drivers/spi/spi-dln2.c } tx; tx 330 drivers/spi/spi-dln2.c tx.port = dln2->port; tx 333 drivers/spi/spi-dln2.c &tx, sizeof(tx), rx, &rx_len); tx 426 drivers/spi/spi-dln2.c } __packed *tx = dln2->buf; tx 429 drivers/spi/spi-dln2.c BUILD_BUG_ON(sizeof(*tx) > DLN2_SPI_BUF_SIZE); tx 434 drivers/spi/spi-dln2.c tx->port = dln2->port; tx 435 drivers/spi/spi-dln2.c tx->size = cpu_to_le16(data_len); tx 436 drivers/spi/spi-dln2.c tx->attr = attr; tx 438 drivers/spi/spi-dln2.c dln2_spi_copy_to_buf(tx->buf, data, data_len, dln2->bpw); tx 440 drivers/spi/spi-dln2.c tx_len = sizeof(*tx) + data_len - DLN2_SPI_MAX_XFER_SIZE; tx 441 drivers/spi/spi-dln2.c return dln2_transfer_tx(dln2->pdev, DLN2_SPI_WRITE, tx, tx_len); tx 455 drivers/spi/spi-dln2.c } __packed tx; tx 467 drivers/spi/spi-dln2.c tx.port = dln2->port; tx 468 drivers/spi/spi-dln2.c tx.size = cpu_to_le16(data_len); tx 469 drivers/spi/spi-dln2.c tx.attr = attr; tx 471 drivers/spi/spi-dln2.c ret = dln2_transfer(dln2->pdev, DLN2_SPI_READ, &tx, sizeof(tx), tx 497 drivers/spi/spi-dln2.c } __packed *tx; tx 504 drivers/spi/spi-dln2.c BUILD_BUG_ON(sizeof(*tx) > DLN2_SPI_BUF_SIZE || tx 515 drivers/spi/spi-dln2.c tx = dln2->buf; tx 518 drivers/spi/spi-dln2.c tx->port = dln2->port; tx 519 drivers/spi/spi-dln2.c tx->size = cpu_to_le16(data_len); tx 520 drivers/spi/spi-dln2.c tx->attr = attr; tx 522 drivers/spi/spi-dln2.c dln2_spi_copy_to_buf(tx->buf, tx_data, data_len, dln2->bpw); tx 524 drivers/spi/spi-dln2.c tx_len = sizeof(*tx) + data_len - DLN2_SPI_MAX_XFER_SIZE; tx 527 drivers/spi/spi-dln2.c ret = dln2_transfer(dln2->pdev, DLN2_SPI_READ_WRITE, tx, tx_len, tx 41 drivers/spi/spi-dw-mid.c struct dw_dma_slave *tx = dws->dma_tx; tx 64 drivers/spi/spi-dw-mid.c tx->dma_dev = &dma_dev->dev; tx 65 drivers/spi/spi-dw-mid.c dws->txchan = dma_request_channel(mask, mid_spi_dma_chan_filter, tx); tx 156 drivers/spi/spi-dw.c tx_left = (dws->tx_end - dws->tx) / dws->n_bytes; tx 167 drivers/spi/spi-dw.c rxtx_gap = ((dws->rx_end - dws->rx) - (dws->tx_end - dws->tx)) tx 192 drivers/spi/spi-dw.c txw = *(u8 *)(dws->tx); tx 194 drivers/spi/spi-dw.c txw = *(u16 *)(dws->tx); tx 197 drivers/spi/spi-dw.c dws->tx += dws->n_bytes; tx 301 drivers/spi/spi-dw.c dws->tx = (void *)transfer->tx_buf; tx 302 drivers/spi/spi-dw.c dws->tx_end = dws->tx + transfer->len; tx 340 drivers/spi/spi-dw.c if (dws->rx && dws->tx) tx 121 drivers/spi/spi-dw.h void *tx; tx 89 drivers/spi/spi-ep93xx.c size_t tx; tx 183 drivers/spi/spi-ep93xx.c val = ((u16 *)xfer->tx_buf)[espi->tx]; tx 184 drivers/spi/spi-ep93xx.c espi->tx += 2; tx 187 drivers/spi/spi-ep93xx.c val = ((u8 *)xfer->tx_buf)[espi->tx]; tx 188 drivers/spi/spi-ep93xx.c espi->tx += 1; tx 234 drivers/spi/spi-ep93xx.c while (espi->fifo_level < SPI_FIFO_SIZE && espi->tx < xfer->len) { tx 494 drivers/spi/spi-ep93xx.c espi->tx = 0; tx 125 drivers/spi/spi-fsl-cpm.c void *nonconst_tx = (void *)mspi->tx; /* shut up gcc */ tx 186 drivers/spi/spi-fsl-dspi.c const void *tx; tx 204 drivers/spi/spi-fsl-dspi.c if (dspi->tx) { tx 206 drivers/spi/spi-fsl-dspi.c txdata = *(u8 *)dspi->tx; tx 208 drivers/spi/spi-fsl-dspi.c txdata = *(u16 *)dspi->tx; tx 210 drivers/spi/spi-fsl-dspi.c txdata = *(u32 *)dspi->tx; tx 211 drivers/spi/spi-fsl-dspi.c dspi->tx += dspi->bytes_per_word; tx 748 drivers/spi/spi-fsl-dspi.c dspi->tx = transfer->tx_buf; tx 42 drivers/spi/spi-fsl-lib.c const type *tx = mpc8xxx_spi->tx; \ tx 43 drivers/spi/spi-fsl-lib.c if (!tx) \ tx 45 drivers/spi/spi-fsl-lib.c data = *tx++ << mpc8xxx_spi->tx_shift; \ tx 46 drivers/spi/spi-fsl-lib.c mpc8xxx_spi->tx = tx; \ tx 25 drivers/spi/spi-fsl-lib.h const void *tx; tx 110 drivers/spi/spi-fsl-lpspi.c void (*tx)(struct fsl_lpspi_data *); tx 254 drivers/spi/spi-fsl-lpspi.c fsl_lpspi->tx(fsl_lpspi); tx 358 drivers/spi/spi-fsl-lpspi.c struct dma_slave_config rx = {}, tx = {}; tx 376 drivers/spi/spi-fsl-lpspi.c tx.direction = DMA_MEM_TO_DEV; tx 377 drivers/spi/spi-fsl-lpspi.c tx.dst_addr = fsl_lpspi->base_phys + IMX7ULP_TDR; tx 378 drivers/spi/spi-fsl-lpspi.c tx.dst_addr_width = buswidth; tx 379 drivers/spi/spi-fsl-lpspi.c tx.dst_maxburst = 1; tx 380 drivers/spi/spi-fsl-lpspi.c ret = dmaengine_slave_config(controller->dma_tx, &tx); tx 457 drivers/spi/spi-fsl-lpspi.c fsl_lpspi->tx = fsl_lpspi_buf_tx_u8; tx 460 drivers/spi/spi-fsl-lpspi.c fsl_lpspi->tx = fsl_lpspi_buf_tx_u16; tx 463 drivers/spi/spi-fsl-lpspi.c fsl_lpspi->tx = fsl_lpspi_buf_tx_u32; tx 572 drivers/spi/spi-fsl-lpspi.c struct sg_table *tx = &transfer->tx_sg, *rx = &transfer->rx_sg; tx 592 drivers/spi/spi-fsl-lpspi.c tx->sgl, tx->nents, DMA_MEM_TO_DEV, tx 335 drivers/spi/spi-fsl-spi.c mpc8xxx_spi->tx = t->tx_buf; tx 98 drivers/spi/spi-imx.c void (*tx)(struct spi_imx_data *); tx 1112 drivers/spi/spi-imx.c spi_imx->tx(spi_imx); tx 1154 drivers/spi/spi-imx.c struct dma_slave_config rx = {}, tx = {}; tx 1171 drivers/spi/spi-imx.c tx.direction = DMA_MEM_TO_DEV; tx 1172 drivers/spi/spi-imx.c tx.dst_addr = spi_imx->base_phys + MXC_CSPITXDATA; tx 1173 drivers/spi/spi-imx.c tx.dst_addr_width = buswidth; tx 1174 drivers/spi/spi-imx.c tx.dst_maxburst = spi_imx->wml; tx 1175 drivers/spi/spi-imx.c ret = dmaengine_slave_config(master->dma_tx, &tx); tx 1215 drivers/spi/spi-imx.c spi_imx->tx = spi_imx_buf_tx_swap; tx 1221 drivers/spi/spi-imx.c spi_imx->tx = spi_imx_buf_tx_u8; tx 1224 drivers/spi/spi-imx.c spi_imx->tx = spi_imx_buf_tx_u16; tx 1227 drivers/spi/spi-imx.c spi_imx->tx = spi_imx_buf_tx_u32; tx 1239 drivers/spi/spi-imx.c spi_imx->tx = mx53_ecspi_tx_slave; tx 1340 drivers/spi/spi-imx.c struct sg_table *tx = &transfer->tx_sg, *rx = &transfer->rx_sg; tx 1384 drivers/spi/spi-imx.c tx->sgl, tx->nents, DMA_MEM_TO_DEV, tx 106 drivers/spi/spi-jcore.c const unsigned char *tx; tx 114 drivers/spi/spi-jcore.c tx = t->tx_buf; tx 122 drivers/spi/spi-jcore.c writel(tx ? *tx++ : 0, data_reg); tx 178 drivers/spi/spi-lantiq-ssc.c const u8 *tx; tx 487 drivers/spi/spi-lantiq-ssc.c tx8 = spi->tx; tx 490 drivers/spi/spi-lantiq-ssc.c spi->tx++; tx 493 drivers/spi/spi-lantiq-ssc.c tx16 = (u16 *) spi->tx; tx 496 drivers/spi/spi-lantiq-ssc.c spi->tx += 2; tx 499 drivers/spi/spi-lantiq-ssc.c tx32 = (u32 *) spi->tx; tx 502 drivers/spi/spi-lantiq-ssc.c spi->tx += 4; tx 617 drivers/spi/spi-lantiq-ssc.c if (spi->tx) { tx 685 drivers/spi/spi-lantiq-ssc.c spi->tx = t->tx_buf; tx 699 drivers/spi/spi-lantiq-ssc.c if (!spi->tx) tx 560 drivers/spi/spi-loopback-test.c void *tx, void *rx) tx 622 drivers/spi/spi-loopback-test.c void *tx, void *rx) tx 649 drivers/spi/spi-loopback-test.c *ptr = tx + off; tx 749 drivers/spi/spi-loopback-test.c void *tx, void *rx) tx 767 drivers/spi/spi-loopback-test.c (void *)tx, rx); tx 773 drivers/spi/spi-loopback-test.c (void *)tx, rx); tx 788 drivers/spi/spi-loopback-test.c ret = test->execute_msg(spi, test, tx, rx); tx 790 drivers/spi/spi-loopback-test.c ret = spi_test_execute_msg(spi, test, tx, rx); tx 811 drivers/spi/spi-loopback-test.c void *tx, void *rx, tx 876 drivers/spi/spi-loopback-test.c return _spi_test_run_iter(spi, &test, tx, rx); tx 890 drivers/spi/spi-loopback-test.c void *tx, void *rx) tx 932 drivers/spi/spi-loopback-test.c ret = spi_test_check_loopback_result(spi, msg, tx, rx); tx 963 drivers/spi/spi-loopback-test.c void *tx, void *rx) tx 1001 drivers/spi/spi-loopback-test.c tx, rx, tx 1026 drivers/spi/spi-loopback-test.c char *rx = NULL, *tx = NULL; tx 1042 drivers/spi/spi-loopback-test.c tx = vmalloc(SPI_TEST_MAX_SIZE_PLUS); tx 1044 drivers/spi/spi-loopback-test.c tx = kzalloc(SPI_TEST_MAX_SIZE_PLUS, GFP_KERNEL); tx 1045 drivers/spi/spi-loopback-test.c if (!tx) { tx 1058 drivers/spi/spi-loopback-test.c ret = test->run_test(spi, test, tx, rx); tx 1060 drivers/spi/spi-loopback-test.c ret = spi_test_run_test(spi, test, tx, rx); tx 1072 drivers/spi/spi-loopback-test.c kvfree(tx); tx 105 drivers/spi/spi-lp8841-rtc.c const u8 *tx = t->tx_buf; tx 110 drivers/spi/spi-lp8841-rtc.c if (tx) { tx 114 drivers/spi/spi-lp8841-rtc.c word = *tx++; tx 102 drivers/spi/spi-mem.c static int spi_check_buswidth_req(struct spi_mem *mem, u8 buswidth, bool tx) tx 111 drivers/spi/spi-mem.c if ((tx && (mode & (SPI_TX_DUAL | SPI_TX_QUAD))) || tx 112 drivers/spi/spi-mem.c (!tx && (mode & (SPI_RX_DUAL | SPI_RX_QUAD)))) tx 118 drivers/spi/spi-mem.c if ((tx && (mode & SPI_TX_QUAD)) || tx 119 drivers/spi/spi-mem.c (!tx && (mode & SPI_RX_QUAD))) tx 125 drivers/spi/spi-mem.c if ((tx && (mode & SPI_TX_OCTAL)) || tx 126 drivers/spi/spi-mem.c (!tx && (mode & SPI_RX_OCTAL))) tx 191 drivers/spi/spi-omap-100k.c const u8 *tx; tx 194 drivers/spi/spi-omap-100k.c tx = xfer->tx_buf; tx 198 drivers/spi/spi-omap-100k.c spi100k_write_data(spi->master, word_len, *tx++); tx 204 drivers/spi/spi-omap-100k.c const u16 *tx; tx 207 drivers/spi/spi-omap-100k.c tx = xfer->tx_buf; tx 211 drivers/spi/spi-omap-100k.c spi100k_write_data(spi->master, word_len, *tx++); tx 217 drivers/spi/spi-omap-100k.c const u32 *tx; tx 220 drivers/spi/spi-omap-100k.c tx = xfer->tx_buf; tx 224 drivers/spi/spi-omap-100k.c spi100k_write_data(spi->master, word_len, *tx); tx 405 drivers/spi/spi-omap2-mcspi.c struct dma_async_tx_descriptor *tx; tx 409 drivers/spi/spi-omap2-mcspi.c tx = dmaengine_prep_slave_sg(mcspi_dma->dma_tx, xfer->tx_sg.sgl, tx 413 drivers/spi/spi-omap2-mcspi.c if (tx) { tx 414 drivers/spi/spi-omap2-mcspi.c tx->callback = omap2_mcspi_tx_callback; tx 415 drivers/spi/spi-omap2-mcspi.c tx->callback_param = spi; tx 416 drivers/spi/spi-omap2-mcspi.c dmaengine_submit(tx); tx 466 drivers/spi/spi-omap2-mcspi.c struct dma_async_tx_descriptor *tx; tx 502 drivers/spi/spi-omap2-mcspi.c tx = dmaengine_prep_slave_sg(mcspi_dma->dma_rx, tx 507 drivers/spi/spi-omap2-mcspi.c if (tx) { tx 508 drivers/spi/spi-omap2-mcspi.c tx->callback = omap2_mcspi_rx_callback; tx 509 drivers/spi/spi-omap2-mcspi.c tx->callback_param = spi; tx 510 drivers/spi/spi-omap2-mcspi.c dmaengine_submit(tx); tx 588 drivers/spi/spi-omap2-mcspi.c const u8 *tx; tx 621 drivers/spi/spi-omap2-mcspi.c tx = xfer->tx_buf; tx 627 drivers/spi/spi-omap2-mcspi.c if (tx) { tx 639 drivers/spi/spi-omap2-mcspi.c if (tx != NULL) { tx 718 drivers/spi/spi-omap2-mcspi.c const u8 *tx; tx 721 drivers/spi/spi-omap2-mcspi.c tx = xfer->tx_buf; tx 725 drivers/spi/spi-omap2-mcspi.c if (tx != NULL) { tx 732 drivers/spi/spi-omap2-mcspi.c word_len, *tx); tx 733 drivers/spi/spi-omap2-mcspi.c writel_relaxed(*tx++, tx_reg); tx 742 drivers/spi/spi-omap2-mcspi.c if (c == 1 && tx == NULL && tx 755 drivers/spi/spi-omap2-mcspi.c } else if (c == 0 && tx == NULL) { tx 766 drivers/spi/spi-omap2-mcspi.c const u16 *tx; tx 769 drivers/spi/spi-omap2-mcspi.c tx = xfer->tx_buf; tx 772 drivers/spi/spi-omap2-mcspi.c if (tx != NULL) { tx 779 drivers/spi/spi-omap2-mcspi.c word_len, *tx); tx 780 drivers/spi/spi-omap2-mcspi.c writel_relaxed(*tx++, tx_reg); tx 789 drivers/spi/spi-omap2-mcspi.c if (c == 2 && tx == NULL && tx 802 drivers/spi/spi-omap2-mcspi.c } else if (c == 0 && tx == NULL) { tx 813 drivers/spi/spi-omap2-mcspi.c const u32 *tx; tx 816 drivers/spi/spi-omap2-mcspi.c tx = xfer->tx_buf; tx 819 drivers/spi/spi-omap2-mcspi.c if (tx != NULL) { tx 826 drivers/spi/spi-omap2-mcspi.c word_len, *tx); tx 827 drivers/spi/spi-omap2-mcspi.c writel_relaxed(*tx++, tx_reg); tx 836 drivers/spi/spi-omap2-mcspi.c if (c == 4 && tx == NULL && tx 849 drivers/spi/spi-omap2-mcspi.c } else if (c == 0 && tx == NULL) { tx 463 drivers/spi/spi-orion.c const u8 *tx = xfer->tx_buf; tx 467 drivers/spi/spi-orion.c if (orion_spi_write_read_8bit(spi, &tx, &rx) < 0) tx 474 drivers/spi/spi-orion.c const u16 *tx = xfer->tx_buf; tx 478 drivers/spi/spi-orion.c if (orion_spi_write_read_16bit(spi, &tx, &rx) < 0) tx 114 drivers/spi/spi-pic32.c const void *tx; tx 165 drivers/spi/spi-pic32.c tx_left = (pic32s->tx_end - pic32s->tx) / n_bytes; tx 177 drivers/spi/spi-pic32.c (pic32s->tx_end - pic32s->tx)) / n_bytes; tx 209 drivers/spi/spi-pic32.c v = *(__type *)(pic32s->tx); \ tx 211 drivers/spi/spi-pic32.c pic32s->tx += sizeof(__type); \ tx 287 drivers/spi/spi-pic32.c if (pic32s->tx_end == pic32s->tx) tx 536 drivers/spi/spi-pic32.c pic32s->tx = (const void *)transfer->tx_buf; tx 538 drivers/spi/spi-pic32.c pic32s->tx_end = pic32s->tx + transfer->len; tx 378 drivers/spi/spi-pl022.c void *tx; tx 685 drivers/spi/spi-pl022.c __func__, pl022->rx, pl022->rx_end, pl022->tx, pl022->tx_end); tx 714 drivers/spi/spi-pl022.c && (pl022->tx < pl022->tx_end)) { tx 720 drivers/spi/spi-pl022.c writew(*(u8 *) (pl022->tx), SSP_DR(pl022->virtbase)); tx 723 drivers/spi/spi-pl022.c writew((*(u16 *) (pl022->tx)), SSP_DR(pl022->virtbase)); tx 726 drivers/spi/spi-pl022.c writel(*(u32 *) (pl022->tx), SSP_DR(pl022->virtbase)); tx 729 drivers/spi/spi-pl022.c pl022->tx += (pl022->cur_chip->n_bytes); tx 1046 drivers/spi/spi-pl022.c setup_dma_scatter(pl022, pl022->tx, tx 1303 drivers/spi/spi-pl022.c if (pl022->tx == pl022->tx_end) { tx 1358 drivers/spi/spi-pl022.c pl022->tx = (void *)transfer->tx_buf; tx 1359 drivers/spi/spi-pl022.c pl022->tx_end = pl022->tx + pl022->cur_transfer->len; tx 1363 drivers/spi/spi-pl022.c pl022->tx ? pl022->cur_chip->write : WRITING_NULL; tx 1548 drivers/spi/spi-pl022.c while (pl022->tx < pl022->tx_end || pl022->rx < pl022->rx_end) { tx 127 drivers/spi/spi-ppc4xx.c const unsigned char *tx; tx 152 drivers/spi/spi-ppc4xx.c hw->tx = t->tx_buf; tx 158 drivers/spi/spi-ppc4xx.c data = hw->tx ? hw->tx[0] : 0; tx 335 drivers/spi/spi-ppc4xx.c data = hw->tx ? hw->tx[count] : 0; tx 106 drivers/spi/spi-pxa2xx-pci.c struct dw_dma_slave *tx, *rx; tx 131 drivers/spi/spi-pxa2xx-pci.c tx = c->tx_param; tx 132 drivers/spi/spi-pxa2xx-pci.c tx->dma_dev = &dma_dev->dev; tx 490 drivers/spi/spi-pxa2xx.c || (drv_data->tx == drv_data->tx_end)) tx 494 drivers/spi/spi-pxa2xx.c drv_data->tx += n_bytes; tx 515 drivers/spi/spi-pxa2xx.c || (drv_data->tx == drv_data->tx_end)) tx 518 drivers/spi/spi-pxa2xx.c pxa2xx_spi_write(drv_data, SSDR, *(u8 *)(drv_data->tx)); tx 519 drivers/spi/spi-pxa2xx.c ++drv_data->tx; tx 538 drivers/spi/spi-pxa2xx.c || (drv_data->tx == drv_data->tx_end)) tx 541 drivers/spi/spi-pxa2xx.c pxa2xx_spi_write(drv_data, SSDR, *(u16 *)(drv_data->tx)); tx 542 drivers/spi/spi-pxa2xx.c drv_data->tx += 2; tx 561 drivers/spi/spi-pxa2xx.c || (drv_data->tx == drv_data->tx_end)) tx 564 drivers/spi/spi-pxa2xx.c pxa2xx_spi_write(drv_data, SSDR, *(u32 *)(drv_data->tx)); tx 565 drivers/spi/spi-pxa2xx.c drv_data->tx += 4; tx 669 drivers/spi/spi-pxa2xx.c if (drv_data->tx == drv_data->tx_end) { tx 982 drivers/spi/spi-pxa2xx.c drv_data->tx = (void *)transfer->tx_buf; tx 983 drivers/spi/spi-pxa2xx.c drv_data->tx_end = drv_data->tx + transfer->len; tx 986 drivers/spi/spi-pxa2xx.c drv_data->write = drv_data->tx ? chip->write : null_writer; tx 1109 drivers/spi/spi-pxa2xx.c drv_data->tx += tx_level; tx 50 drivers/spi/spi-pxa2xx.h void *tx; tx 170 drivers/spi/spi-rockchip.c const void *tx; tx 277 drivers/spi/spi-rockchip.c txw = *(u8 *)rs->tx; tx 279 drivers/spi/spi-rockchip.c txw = *(u16 *)rs->tx; tx 282 drivers/spi/spi-rockchip.c rs->tx += rs->n_bytes; tx 341 drivers/spi/spi-rockchip.c rs->tx = xfer->tx_buf; tx 343 drivers/spi/spi-rockchip.c rs->tx_left = rs->tx ? xfer->len / rs->n_bytes : 0; tx 493 drivers/spi/spi-rspi.c static int rspi_pio_transfer(struct rspi_data *rspi, const u8 *tx, u8 *rx, tx 497 drivers/spi/spi-rspi.c if (tx) { tx 498 drivers/spi/spi-rspi.c int ret = rspi_data_out(rspi, *tx++); tx 521 drivers/spi/spi-rspi.c static int rspi_dma_transfer(struct rspi_data *rspi, struct sg_table *tx, tx 551 drivers/spi/spi-rspi.c if (tx) { tx 552 drivers/spi/spi-rspi.c desc_tx = dmaengine_prep_slave_sg(rspi->ctlr->dma_tx, tx->sgl, tx 553 drivers/spi/spi-rspi.c tx->nents, DMA_MEM_TO_DEV, tx 580 drivers/spi/spi-rspi.c if (tx) tx 591 drivers/spi/spi-rspi.c if (tx) tx 603 drivers/spi/spi-rspi.c if (tx) tx 611 drivers/spi/spi-rspi.c if (tx) tx 733 drivers/spi/spi-rspi.c static int qspi_trigger_transfer_out_in(struct rspi_data *rspi, const u8 *tx, tx 748 drivers/spi/spi-rspi.c rspi_write_data(rspi, *tx++); tx 781 drivers/spi/spi-rspi.c const u8 *tx = xfer->tx_buf; tx 800 drivers/spi/spi-rspi.c rspi_write_data(rspi, *tx++); tx 70 drivers/spi/spi-s3c24xx.c const unsigned char *tx; tx 209 drivers/spi/spi-s3c24xx.c return hw->tx ? hw->tx[count] : 0; tx 275 drivers/spi/spi-s3c24xx.c if (hw->tx && !hw->rx) tx 277 drivers/spi/spi-s3c24xx.c else if (hw->rx && !hw->tx) tx 284 drivers/spi/spi-s3c24xx.c regs.uregs[fiq_rtx] = (long)hw->tx + 1; tx 402 drivers/spi/spi-s3c24xx.c hw->tx = t->tx_buf; tx 148 drivers/spi/spi-sh-hspi.c u32 tx; tx 174 drivers/spi/spi-sh-hspi.c tx = 0; tx 176 drivers/spi/spi-sh-hspi.c tx = (u32)((u8 *)t->tx_buf)[i]; tx 178 drivers/spi/spi-sh-hspi.c hspi_write(hspi, SPTBR, tx); tx 743 drivers/spi/spi-sh-msiof.c static int sh_msiof_dma_once(struct sh_msiof_spi_priv *p, const void *tx, tx 767 drivers/spi/spi-sh-msiof.c if (tx) { tx 792 drivers/spi/spi-sh-msiof.c sh_msiof_spi_set_mode_regs(p, tx, rx, 32, len / 4); tx 797 drivers/spi/spi-sh-msiof.c if (tx) tx 804 drivers/spi/spi-sh-msiof.c if (tx) tx 813 drivers/spi/spi-sh-msiof.c if (tx) { tx 854 drivers/spi/spi-sh-msiof.c if (tx) tx 260 drivers/spi/spi-sirf.c const void *tx; tx 330 drivers/spi/spi-sirf.c const u8 *tx = sspi->tx; tx 332 drivers/spi/spi-sirf.c if (tx) { tx 333 drivers/spi/spi-sirf.c data = *tx++; tx 334 drivers/spi/spi-sirf.c sspi->tx = tx; tx 358 drivers/spi/spi-sirf.c const u16 *tx = sspi->tx; tx 360 drivers/spi/spi-sirf.c if (tx) { tx 361 drivers/spi/spi-sirf.c data = *tx++; tx 362 drivers/spi/spi-sirf.c sspi->tx = tx; tx 388 drivers/spi/spi-sirf.c const u32 *tx = sspi->tx; tx 390 drivers/spi/spi-sirf.c if (tx) { tx 391 drivers/spi/spi-sirf.c data = *tx++; tx 392 drivers/spi/spi-sirf.c sspi->tx = tx; tx 469 drivers/spi/spi-sirf.c memcpy(&cmd, sspi->tx, t->len); tx 556 drivers/spi/spi-sirf.c sspi->src_start = dma_map_single(&spi->dev, (void *)sspi->tx, t->len, tx 709 drivers/spi/spi-sirf.c sspi->tx = t->tx_buf; tx 211 drivers/spi/spi-stm32.c const struct stm32_spi_reg tx; tx 331 drivers/spi/spi-stm32.c .tx = { STM32F4_SPI_DR }, tx 350 drivers/spi/spi-stm32.c .tx = { STM32H7_SPI_TXDR }, tx 1141 drivers/spi/spi-stm32.c dma_conf->dst_addr = spi->phys_addr + spi->cfg->regs->tx.reg; tx 78 drivers/spi/spi-test.h void *tx, void *rx); tx 80 drivers/spi/spi-test.h void *tx, void *rx); tx 109 drivers/spi/spi-test.h void *tx, void *rx); tx 114 drivers/spi/spi-test.h void *tx, void *rx); tx 401 drivers/spi/spi-ti-qspi.c struct dma_async_tx_descriptor *tx; tx 404 drivers/spi/spi-ti-qspi.c tx = dmaengine_prep_dma_memcpy(chan, dma_dst, dma_src, len, flags); tx 405 drivers/spi/spi-ti-qspi.c if (!tx) { tx 410 drivers/spi/spi-ti-qspi.c tx->callback = ti_qspi_dma_callback; tx 411 drivers/spi/spi-ti-qspi.c tx->callback_param = qspi; tx 412 drivers/spi/spi-ti-qspi.c cookie = tx->tx_submit(tx); tx 171 drivers/staging/comedi/drivers/ni_usb6501.c u8 *tx = devpriv->usb_tx_buf; tx 183 drivers/staging/comedi/drivers/ni_usb6501.c memcpy(tx, READ_PORT_REQUEST, request_size); tx 184 drivers/staging/comedi/drivers/ni_usb6501.c tx[14] = val & 0xff; tx 189 drivers/staging/comedi/drivers/ni_usb6501.c memcpy(tx, WRITE_PORT_REQUEST, request_size); tx 190 drivers/staging/comedi/drivers/ni_usb6501.c tx[14] = val & 0xff; tx 191 drivers/staging/comedi/drivers/ni_usb6501.c tx[17] = *bitmap; tx 196 drivers/staging/comedi/drivers/ni_usb6501.c memcpy(tx, SET_PORT_DIR_REQUEST, request_size); tx 197 drivers/staging/comedi/drivers/ni_usb6501.c tx[14] = val & 0xff; tx 198 drivers/staging/comedi/drivers/ni_usb6501.c tx[15] = (val >> 8) & 0xff; tx 199 drivers/staging/comedi/drivers/ni_usb6501.c tx[16] = (val >> 16) & 0xff; tx 253 drivers/staging/comedi/drivers/ni_usb6501.c u8 *tx = devpriv->usb_tx_buf; tx 265 drivers/staging/comedi/drivers/ni_usb6501.c memcpy(tx, START_COUNTER_REQUEST, request_size); tx 270 drivers/staging/comedi/drivers/ni_usb6501.c memcpy(tx, STOP_COUNTER_REQUEST, request_size); tx 275 drivers/staging/comedi/drivers/ni_usb6501.c memcpy(tx, READ_COUNTER_REQUEST, request_size); tx 280 drivers/staging/comedi/drivers/ni_usb6501.c memcpy(tx, WRITE_COUNTER_REQUEST, request_size); tx 283 drivers/staging/comedi/drivers/ni_usb6501.c *((__be32 *)&tx[12]) = cpu_to_be32(*val); tx 683 drivers/staging/fwserial/fwserial.c port->icount.tx += txn->dma_pended.len; tx 1390 drivers/staging/fwserial/fwserial.c icount->tx = port->icount.tx + stats.xchars; tx 1408 drivers/staging/fwserial/fwserial.c port->icount.tx + stats.xchars, port->icount.rx); tx 145 drivers/staging/gdm724x/gdm_usb.c static struct usb_tx_sdu *get_tx_sdu_struct(struct tx_cxt *tx, int *no_spc) tx 149 drivers/staging/gdm724x/gdm_usb.c if (list_empty(&tx->free_list)) tx 152 drivers/staging/gdm724x/gdm_usb.c t_sdu = list_entry(tx->free_list.next, struct usb_tx_sdu, list); tx 155 drivers/staging/gdm724x/gdm_usb.c tx->avail_count--; tx 157 drivers/staging/gdm724x/gdm_usb.c *no_spc = list_empty(&tx->free_list) ? 1 : 0; tx 162 drivers/staging/gdm724x/gdm_usb.c static void put_tx_struct(struct tx_cxt *tx, struct usb_tx_sdu *t_sdu) tx 164 drivers/staging/gdm724x/gdm_usb.c list_add_tail(&t_sdu->list, &tx->free_list); tx 165 drivers/staging/gdm724x/gdm_usb.c tx->avail_count++; tx 247 drivers/staging/gdm724x/gdm_usb.c struct tx_cxt *tx = &udev->tx; tx 253 drivers/staging/gdm724x/gdm_usb.c spin_lock_irqsave(&tx->lock, flags); tx 254 drivers/staging/gdm724x/gdm_usb.c list_for_each_entry_safe(t_sdu, t_sdu_next, &tx->sdu_list, list) { tx 259 drivers/staging/gdm724x/gdm_usb.c list_for_each_entry_safe(t, t_next, &tx->hci_list, list) { tx 264 drivers/staging/gdm724x/gdm_usb.c list_for_each_entry_safe(t_sdu, t_sdu_next, &tx->free_list, list) { tx 268 drivers/staging/gdm724x/gdm_usb.c spin_unlock_irqrestore(&tx->lock, flags); tx 300 drivers/staging/gdm724x/gdm_usb.c struct tx_cxt *tx = &udev->tx; tx 310 drivers/staging/gdm724x/gdm_usb.c INIT_LIST_HEAD(&tx->sdu_list); tx 311 drivers/staging/gdm724x/gdm_usb.c INIT_LIST_HEAD(&tx->hci_list); tx 312 drivers/staging/gdm724x/gdm_usb.c INIT_LIST_HEAD(&tx->free_list); tx 316 drivers/staging/gdm724x/gdm_usb.c spin_lock_init(&tx->lock); tx 321 drivers/staging/gdm724x/gdm_usb.c tx->avail_count = 0; tx 333 drivers/staging/gdm724x/gdm_usb.c list_add(&t_sdu->list, &tx->free_list); tx 334 drivers/staging/gdm724x/gdm_usb.c tx->avail_count++; tx 542 drivers/staging/gdm724x/gdm_usb.c struct tx_cxt *tx = t->tx; tx 543 drivers/staging/gdm724x/gdm_usb.c struct lte_udev *udev = container_of(tx, struct lte_udev, tx); tx 556 drivers/staging/gdm724x/gdm_usb.c spin_lock_irqsave(&tx->lock, flags); tx 559 drivers/staging/gdm724x/gdm_usb.c spin_unlock_irqrestore(&tx->lock, flags); tx 590 drivers/staging/gdm724x/gdm_usb.c struct tx_cxt *tx = &udev->tx; tx 600 drivers/staging/gdm724x/gdm_usb.c spin_lock_irqsave(&tx->lock, flags); tx 601 drivers/staging/gdm724x/gdm_usb.c if (list_empty(&tx->sdu_list)) { tx 602 drivers/staging/gdm724x/gdm_usb.c spin_unlock_irqrestore(&tx->lock, flags); tx 606 drivers/staging/gdm724x/gdm_usb.c t_sdu = list_entry(tx->sdu_list.next, struct usb_tx_sdu, list); tx 608 drivers/staging/gdm724x/gdm_usb.c spin_unlock_irqrestore(&tx->lock, flags); tx 613 drivers/staging/gdm724x/gdm_usb.c spin_unlock_irqrestore(&tx->lock, flags); tx 620 drivers/staging/gdm724x/gdm_usb.c if (tx->avail_count > 10) tx 623 drivers/staging/gdm724x/gdm_usb.c spin_lock_irqsave(&tx->lock, flags); tx 624 drivers/staging/gdm724x/gdm_usb.c put_tx_struct(tx, t_sdu); tx 625 drivers/staging/gdm724x/gdm_usb.c spin_unlock_irqrestore(&tx->lock, flags); tx 639 drivers/staging/gdm724x/gdm_usb.c struct tx_cxt *tx = &udev->tx; tx 651 drivers/staging/gdm724x/gdm_usb.c spin_lock_irqsave(&tx->lock, flags); tx 653 drivers/staging/gdm724x/gdm_usb.c spin_unlock_irqrestore(&tx->lock, flags); tx 658 drivers/staging/gdm724x/gdm_usb.c if (!list_empty(&tx->hci_list)) { tx 659 drivers/staging/gdm724x/gdm_usb.c t = list_entry(tx->hci_list.next, struct usb_tx, list); tx 664 drivers/staging/gdm724x/gdm_usb.c } else if (!list_empty(&tx->sdu_list)) { tx 667 drivers/staging/gdm724x/gdm_usb.c spin_unlock_irqrestore(&tx->lock, flags); tx 673 drivers/staging/gdm724x/gdm_usb.c spin_unlock_irqrestore(&tx->lock, flags); tx 677 drivers/staging/gdm724x/gdm_usb.c t->tx = tx; tx 684 drivers/staging/gdm724x/gdm_usb.c spin_unlock_irqrestore(&tx->lock, flags); tx 687 drivers/staging/gdm724x/gdm_usb.c spin_unlock_irqrestore(&tx->lock, flags); tx 706 drivers/staging/gdm724x/gdm_usb.c struct tx_cxt *tx = &udev->tx; tx 718 drivers/staging/gdm724x/gdm_usb.c spin_lock_irqsave(&tx->lock, flags); tx 719 drivers/staging/gdm724x/gdm_usb.c t_sdu = get_tx_sdu_struct(tx, &no_spc); tx 720 drivers/staging/gdm724x/gdm_usb.c spin_unlock_irqrestore(&tx->lock, flags); tx 747 drivers/staging/gdm724x/gdm_usb.c spin_lock_irqsave(&tx->lock, flags); tx 748 drivers/staging/gdm724x/gdm_usb.c list_add_tail(&t_sdu->list, &tx->sdu_list); tx 750 drivers/staging/gdm724x/gdm_usb.c spin_unlock_irqrestore(&tx->lock, flags); tx 762 drivers/staging/gdm724x/gdm_usb.c struct tx_cxt *tx = &udev->tx; tx 781 drivers/staging/gdm724x/gdm_usb.c t->tx = tx; tx 784 drivers/staging/gdm724x/gdm_usb.c spin_lock_irqsave(&tx->lock, flags); tx 785 drivers/staging/gdm724x/gdm_usb.c list_add_tail(&t->list, &tx->hci_list); tx 787 drivers/staging/gdm724x/gdm_usb.c spin_unlock_irqrestore(&tx->lock, flags); tx 939 drivers/staging/gdm724x/gdm_usb.c struct tx_cxt *tx; tx 967 drivers/staging/gdm724x/gdm_usb.c tx = &udev->tx; tx 968 drivers/staging/gdm724x/gdm_usb.c spin_lock_irqsave(&tx->lock, flags); tx 970 drivers/staging/gdm724x/gdm_usb.c spin_unlock_irqrestore(&tx->lock, flags); tx 42 drivers/staging/gdm724x/gdm_usb.h struct tx_cxt *tx; tx 86 drivers/staging/gdm724x/gdm_usb.h struct tx_cxt tx; tx 28 drivers/staging/iio/meter/ade7854-i2c.c st->tx[0] = (reg_address >> 8) & 0xFF; tx 29 drivers/staging/iio/meter/ade7854-i2c.c st->tx[1] = reg_address & 0xFF; tx 33 drivers/staging/iio/meter/ade7854-i2c.c st->tx[2] = val & 0xFF; tx 37 drivers/staging/iio/meter/ade7854-i2c.c st->tx[2] = (val >> 8) & 0xFF; tx 38 drivers/staging/iio/meter/ade7854-i2c.c st->tx[3] = val & 0xFF; tx 42 drivers/staging/iio/meter/ade7854-i2c.c st->tx[2] = (val >> 16) & 0xFF; tx 43 drivers/staging/iio/meter/ade7854-i2c.c st->tx[3] = (val >> 8) & 0xFF; tx 44 drivers/staging/iio/meter/ade7854-i2c.c st->tx[4] = val & 0xFF; tx 48 drivers/staging/iio/meter/ade7854-i2c.c st->tx[2] = (val >> 24) & 0xFF; tx 49 drivers/staging/iio/meter/ade7854-i2c.c st->tx[3] = (val >> 16) & 0xFF; tx 50 drivers/staging/iio/meter/ade7854-i2c.c st->tx[4] = (val >> 8) & 0xFF; tx 51 drivers/staging/iio/meter/ade7854-i2c.c st->tx[5] = val & 0xFF; tx 59 drivers/staging/iio/meter/ade7854-i2c.c ret = i2c_master_send(st->i2c, st->tx, count); tx 77 drivers/staging/iio/meter/ade7854-i2c.c st->tx[0] = (reg_address >> 8) & 0xFF; tx 78 drivers/staging/iio/meter/ade7854-i2c.c st->tx[1] = reg_address & 0xFF; tx 80 drivers/staging/iio/meter/ade7854-i2c.c ret = i2c_master_send(st->i2c, st->tx, 2); tx 26 drivers/staging/iio/meter/ade7854-spi.c .tx_buf = st->tx, tx 32 drivers/staging/iio/meter/ade7854-spi.c st->tx[0] = ADE7854_WRITE_REG; tx 33 drivers/staging/iio/meter/ade7854-spi.c st->tx[1] = (reg_address >> 8) & 0xFF; tx 34 drivers/staging/iio/meter/ade7854-spi.c st->tx[2] = reg_address & 0xFF; tx 37 drivers/staging/iio/meter/ade7854-spi.c st->tx[3] = val & 0xFF; tx 41 drivers/staging/iio/meter/ade7854-spi.c st->tx[3] = (val >> 8) & 0xFF; tx 42 drivers/staging/iio/meter/ade7854-spi.c st->tx[4] = val & 0xFF; tx 46 drivers/staging/iio/meter/ade7854-spi.c st->tx[3] = (val >> 16) & 0xFF; tx 47 drivers/staging/iio/meter/ade7854-spi.c st->tx[4] = (val >> 8) & 0xFF; tx 48 drivers/staging/iio/meter/ade7854-spi.c st->tx[5] = val & 0xFF; tx 52 drivers/staging/iio/meter/ade7854-spi.c st->tx[3] = (val >> 24) & 0xFF; tx 53 drivers/staging/iio/meter/ade7854-spi.c st->tx[4] = (val >> 16) & 0xFF; tx 54 drivers/staging/iio/meter/ade7854-spi.c st->tx[5] = (val >> 8) & 0xFF; tx 55 drivers/staging/iio/meter/ade7854-spi.c st->tx[6] = val & 0xFF; tx 79 drivers/staging/iio/meter/ade7854-spi.c .tx_buf = st->tx, tx 91 drivers/staging/iio/meter/ade7854-spi.c st->tx[0] = ADE7854_READ_REG; tx 92 drivers/staging/iio/meter/ade7854-spi.c st->tx[1] = (reg_address >> 8) & 0xFF; tx 93 drivers/staging/iio/meter/ade7854-spi.c st->tx[2] = reg_address & 0xFF; tx 165 drivers/staging/iio/meter/ade7854.h u8 tx[ADE7854_MAX_TX] ____cacheline_aligned; tx 98 drivers/staging/iio/resolver/ad2s1210.c u8 tx[2] ____cacheline_aligned; tx 121 drivers/staging/iio/resolver/ad2s1210.c st->tx[0] = data; tx 122 drivers/staging/iio/resolver/ad2s1210.c ret = spi_write(st->sdev, st->tx, 1); tx 137 drivers/staging/iio/resolver/ad2s1210.c .tx_buf = &st->tx[0], tx 142 drivers/staging/iio/resolver/ad2s1210.c .tx_buf = &st->tx[1], tx 148 drivers/staging/iio/resolver/ad2s1210.c st->tx[0] = address | AD2S1210_MSB_IS_HIGH; tx 149 drivers/staging/iio/resolver/ad2s1210.c st->tx[1] = AD2S1210_REG_FAULT; tx 138 drivers/staging/kpc2000/kpc2000_spi.c unsigned int tx : 1; /* Tx Status */ tx 215 drivers/staging/kpc2000/kpc2000_spi.c const u8 *tx = transfer->tx_buf; tx 218 drivers/staging/kpc2000/kpc2000_spi.c if (tx) { tx 220 drivers/staging/kpc2000/kpc2000_spi.c char val = *tx++; tx 105 drivers/staging/media/soc_camera/imx074.c unsigned char tx[3]; tx 109 drivers/staging/media/soc_camera/imx074.c msg.buf = tx; tx 113 drivers/staging/media/soc_camera/imx074.c tx[0] = addr >> 8; tx 114 drivers/staging/media/soc_camera/imx074.c tx[1] = addr & 0xff; tx 115 drivers/staging/media/soc_camera/imx074.c tx[2] = data; tx 66 drivers/staging/most/net/net.c struct net_dev_channel tx; tx 199 drivers/staging/most/net/net.c if (most_start_channel(nd->iface, nd->tx.ch_id, &comp)) { tx 213 drivers/staging/most/net/net.c nd->iface->request_netinfo(nd->iface, nd->tx.ch_id, on_netinfo); tx 226 drivers/staging/most/net/net.c nd->iface->request_netinfo(nd->iface, nd->tx.ch_id, NULL); tx 228 drivers/staging/most/net/net.c most_stop_channel(nd->iface, nd->tx.ch_id, &comp); tx 240 drivers/staging/most/net/net.c mbo = most_get_mbo(nd->iface, nd->tx.ch_id, &comp); tx 297 drivers/staging/most/net/net.c if (nd && nd->rx.linked && nd->tx.linked) tx 339 drivers/staging/most/net/net.c ch = ccfg->direction == MOST_CH_TX ? &nd->tx : &nd->rx; tx 341 drivers/staging/most/net/net.c ch = ccfg->direction == MOST_CH_TX ? &nd->tx : &nd->rx; tx 379 drivers/staging/most/net/net.c } else if (nd->tx.linked && channel_idx == nd->tx.ch_id) { tx 380 drivers/staging/most/net/net.c ch = &nd->tx; tx 386 drivers/staging/most/net/net.c if (nd->rx.linked && nd->tx.linked) { tx 418 drivers/staging/most/net/net.c if (nd->tx.ch_id != channel_idx) tx 473 drivers/staging/nvec/nvec.c if (nvec->tx->pos != nvec->tx->size) { tx 475 drivers/staging/nvec/nvec.c nvec->tx->pos = 0; tx 554 drivers/staging/nvec/nvec.c nvec->tx = &nvec->tx_scratch; tx 555 drivers/staging/nvec/nvec.c list_add_tail(&nvec->tx->node, &nvec->tx_data); tx 557 drivers/staging/nvec/nvec.c nvec->tx = list_first_entry(&nvec->tx_data, struct nvec_msg, tx 559 drivers/staging/nvec/nvec.c nvec->tx->pos = 0; tx 564 drivers/staging/nvec/nvec.c (uint)nvec->tx->size, nvec->tx->data[1]); tx 639 drivers/staging/nvec/nvec.c to_send = nvec->tx->data[0]; tx 640 drivers/staging/nvec/nvec.c nvec->tx->pos = 1; tx 654 drivers/staging/nvec/nvec.c } else if (nvec->tx && nvec->tx->pos < nvec->tx->size) { tx 655 drivers/staging/nvec/nvec.c to_send = nvec->tx->data[nvec->tx->pos++]; tx 659 drivers/staging/nvec/nvec.c nvec->tx, tx 660 drivers/staging/nvec/nvec.c (uint)(nvec->tx ? nvec->tx->pos : 0), tx 661 drivers/staging/nvec/nvec.c (uint)(nvec->tx ? nvec->tx->size : 0)); tx 149 drivers/staging/nvec/nvec.h struct nvec_msg *tx; tx 671 drivers/staging/ralink-gdma/ralink-gdma.c if (desc && (cookie == desc->vdesc.tx.cookie)) { tx 1201 drivers/staging/rtl8188eu/core/rtw_ap.c void update_beacon(struct adapter *padapter, u8 ie_id, u8 *oui, u8 tx) tx 1235 drivers/staging/rtl8188eu/core/rtw_ap.c if (tx) tx 23 drivers/staging/rtl8188eu/include/rtw_ap.h u8 *oui, u8 tx); tx 1748 drivers/staging/rtl8723bs/core/rtw_ap.c void update_beacon(struct adapter *padapter, u8 ie_id, u8 *oui, u8 tx) tx 1820 drivers/staging/rtl8723bs/core/rtw_ap.c if (tx) { tx 205 drivers/staging/rtl8723bs/core/rtw_pwrctrl.c void traffic_check_for_leave_lps(struct adapter *padapter, u8 tx, u32 tx_packets) tx 214 drivers/staging/rtl8723bs/core/rtw_pwrctrl.c if (tx) { /* from tx */ tx 248 drivers/staging/rtl8723bs/core/rtw_pwrctrl.c rtw_lps_ctrl_wk_cmd(padapter, LPS_CTRL_LEAVE, tx?0:1); tx 13 drivers/staging/rtl8723bs/include/rtw_ap.h void update_beacon(struct adapter *padapter, u8 ie_id, u8 *oui, u8 tx); tx 225 drivers/staging/rtl8723bs/include/rtw_mp.h struct mp_tx tx; tx 348 drivers/staging/rtl8723bs/include/rtw_pwrctrl.h void traffic_check_for_leave_lps(struct adapter *padapter, u8 tx, u32 tx_packets); tx 174 drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c .tx = 0xffff, tx 178 drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c .tx = 0xffff, tx 183 drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c .tx = 0xffff, tx 194 drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c .tx = 0xffff, tx 204 drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c .tx = 0xffff, tx 209 drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c .tx = 0xffff, tx 1621 drivers/staging/vt6655/device_main.c .tx = vnt_tx_80211, tx 936 drivers/staging/vt6656/main_usb.c .tx = vnt_tx_80211, tx 588 drivers/staging/vt6656/rxtx.c union vnt_tx_data_head *head = &tx_head->tx_rts.tx.head; tx 607 drivers/staging/vt6656/rxtx.c head = &tx_head->tx_rts.tx.mic.head; tx 620 drivers/staging/vt6656/rxtx.c union vnt_tx_data_head *head = &tx_head->tx_cts.tx.head; tx 634 drivers/staging/vt6656/rxtx.c head = &tx_head->tx_cts.tx.mic.head; tx 648 drivers/staging/vt6656/rxtx.c union vnt_tx_data_head *head = &tx_head->tx_ab.tx.head; tx 657 drivers/staging/vt6656/rxtx.c head = &tx_head->tx_ab.tx.mic.head; tx 692 drivers/staging/vt6656/rxtx.c &tx_buffer->tx_head.tx_rts.tx.mic.hdr; tx 699 drivers/staging/vt6656/rxtx.c *mic_hdr = &tx_buffer->tx_head.tx_cts.tx.mic.hdr; tx 705 drivers/staging/vt6656/rxtx.c *mic_hdr = &tx_buffer->tx_head.tx_ab.tx.mic.hdr; tx 198 drivers/staging/vt6656/rxtx.h union vnt_tx tx; tx 202 drivers/staging/vt6656/rxtx.h union vnt_tx tx; tx 206 drivers/staging/vt6656/rxtx.h union vnt_tx tx; tx 35 drivers/staging/wilc1000/wilc_wfi_cfgoperations.c .tx = 0xffff, tx 40 drivers/staging/wilc1000/wilc_wfi_cfgoperations.c .tx = 0xffff, tx 50 drivers/staging/wilc1000/wilc_wfi_cfgoperations.c .tx = 0xffff, tx 85 drivers/target/iscsi/cxgbit/cxgbit.h struct cxgbit_skb_tx_cb tx; tx 99 drivers/target/iscsi/cxgbit/cxgbit.h #define cxgbit_skcb_submode(skb) (CXGBIT_SKB_CB(skb)->tx.submode) tx 101 drivers/target/iscsi/cxgbit/cxgbit.h #define cxgbit_skcb_tx_extralen(skb) (CXGBIT_SKB_CB(skb)->tx.extra_len) tx 47 drivers/thermal/tegra/tegra-bpmp-thermal.c msg.tx.data = &req; tx 48 drivers/thermal/tegra/tegra-bpmp-thermal.c msg.tx.size = sizeof(req); tx 76 drivers/thermal/tegra/tegra-bpmp-thermal.c msg.tx.data = &req; tx 77 drivers/thermal/tegra/tegra-bpmp-thermal.c msg.tx.size = sizeof(req); tx 135 drivers/thermal/tegra/tegra-bpmp-thermal.c msg.tx.data = &req; tx 136 drivers/thermal/tegra/tegra-bpmp-thermal.c msg.tx.size = sizeof(req); tx 27 drivers/thunderbolt/ctl.c struct tb_ring *tx; tx 364 drivers/thunderbolt/ctl.c res = tb_ring_tx(ctl->tx, &pkg->frame); tx 619 drivers/thunderbolt/ctl.c ctl->tx = tb_ring_alloc_tx(nhi, 0, 10, RING_FLAG_NO_SUSPEND); tx 620 drivers/thunderbolt/ctl.c if (!ctl->tx) tx 658 drivers/thunderbolt/ctl.c if (ctl->tx) tx 659 drivers/thunderbolt/ctl.c tb_ring_free(ctl->tx); tx 677 drivers/thunderbolt/ctl.c tb_ring_start(ctl->tx); /* is used to ack hotplug packets, start first */ tx 700 drivers/thunderbolt/ctl.c tb_ring_stop(ctl->tx); tx 345 drivers/tty/amiserial.c info->icount.tx++; tx 361 drivers/tty/amiserial.c info->icount.tx++; tx 1195 drivers/tty/amiserial.c icount->tx = cnow.tx; tx 1516 drivers/tty/amiserial.c seq_printf(m, " tx:%d rx:%d", state->icount.tx, state->icount.rx); tx 1649 drivers/tty/amiserial.c state->icount.rx = state->icount.tx = 0; tx 601 drivers/tty/cyclades.c info->icount.tx++; tx 656 drivers/tty/cyclades.c info->icount.tx++; tx 664 drivers/tty/cyclades.c info->icount.tx++; tx 1043 drivers/tty/cyclades.c info->icount.tx++; tx 1058 drivers/tty/cyclades.c info->icount.tx += small_count; tx 1073 drivers/tty/cyclades.c info->icount.tx++; tx 2752 drivers/tty/cyclades.c sic->tx = cnow.tx; tx 1105 drivers/tty/ipwireless/hardware.c int tx = 0; tx 1183 drivers/tty/ipwireless/hardware.c tx = 1; tx 1185 drivers/tty/ipwireless/hardware.c if (tx) tx 1191 drivers/tty/ipwireless/hardware.c if (tx || rx) tx 176 drivers/tty/isicom.c static DEFINE_TIMER(tx, isicom_tx); tx 500 drivers/tty/isicom.c mod_timer(&tx, jiffies + msecs_to_jiffies(10)); tx 1669 drivers/tty/isicom.c mod_timer(&tx, jiffies + 1); tx 1682 drivers/tty/isicom.c del_timer_sync(&tx); tx 1840 drivers/tty/mxser.c icount->tx = cnow.tx; tx 2169 drivers/tty/mxser.c port->icount.tx++; tx 2198 drivers/tty/mxser.c port->icount.tx += (cnt - port->xmit_cnt); tx 1738 drivers/tty/nozomi.c icount->tx = cnow.tx; tx 127 drivers/tty/serial/21285.c port->icount.tx++; tx 139 drivers/tty/serial/21285.c port->icount.tx++; tx 31 drivers/tty/serial/8250/8250_dma.c p->port.icount.tx += dma->tx_size; tx 914 drivers/tty/serial/8250/8250_omap.c p->port.icount.tx += dma->tx_size; tx 1731 drivers/tty/serial/8250/8250_port.c port->icount.tx++; tx 1748 drivers/tty/serial/8250/8250_port.c port->icount.tx++; tx 149 drivers/tty/serial/altera_jtaguart.c port->icount.tx++; tx 166 drivers/tty/serial/altera_jtaguart.c port->icount.tx++; tx 260 drivers/tty/serial/altera_uart.c port->icount.tx++; tx 271 drivers/tty/serial/altera_uart.c port->icount.tx++; tx 178 drivers/tty/serial/amba-pl010.c uap->port.icount.tx++; tx 191 drivers/tty/serial/amba-pl010.c uap->port.icount.tx++; tx 684 drivers/tty/serial/amba-pl011.c uap->port.icount.tx += count; tx 792 drivers/tty/serial/amba-pl011.c uap->port.icount.tx++; tx 1386 drivers/tty/serial/amba-pl011.c uap->port.icount.tx++; tx 136 drivers/tty/serial/apbuart.c port->icount.tx++; tx 151 drivers/tty/serial/apbuart.c port->icount.tx++; tx 354 drivers/tty/serial/ar933x_uart.c up->port.icount.tx++; tx 365 drivers/tty/serial/ar933x_uart.c up->port.icount.tx++; tx 168 drivers/tty/serial/arc_uart.c port->icount.tx++; tx 174 drivers/tty/serial/arc_uart.c port->icount.tx++; tx 837 drivers/tty/serial/atmel_serial.c port->icount.tx++; tx 847 drivers/tty/serial/atmel_serial.c port->icount.tx++; tx 876 drivers/tty/serial/atmel_serial.c port->icount.tx += atmel_port->tx_len; tx 1465 drivers/tty/serial/atmel_serial.c port->icount.tx += pdc->ofs; tx 317 drivers/tty/serial/bcm63xx_uart.c port->icount.tx++; tx 341 drivers/tty/serial/bcm63xx_uart.c port->icount.tx++; tx 156 drivers/tty/serial/clps711x.c port->icount.tx++; tx 174 drivers/tty/serial/clps711x.c port->icount.tx++; tx 704 drivers/tty/serial/cpm_uart/cpm_uart_core.c port->icount.tx++; tx 724 drivers/tty/serial/cpm_uart/cpm_uart_core.c port->icount.tx++; tx 193 drivers/tty/serial/digicolor-usart.c port->icount.tx++; tx 206 drivers/tty/serial/digicolor-usart.c port->icount.tx++; tx 269 drivers/tty/serial/dz.c dport->port.icount.tx++; tx 288 drivers/tty/serial/dz.c dport->port.icount.tx++; tx 144 drivers/tty/serial/efm32-uart.c port->icount.tx++; tx 151 drivers/tty/serial/efm32-uart.c port->icount.tx++; tx 182 drivers/tty/serial/fsl_linflexuart.c sport->icount.tx++; tx 474 drivers/tty/serial/fsl_lpuart.c sport->port.icount.tx += sport->dma_tx_bytes; tx 684 drivers/tty/serial/fsl_lpuart.c sport->port.icount.tx++; tx 698 drivers/tty/serial/fsl_lpuart.c sport->port.icount.tx++; tx 715 drivers/tty/serial/fsl_lpuart.c sport->port.icount.tx++; tx 731 drivers/tty/serial/fsl_lpuart.c sport->port.icount.tx++; tx 707 drivers/tty/serial/icom.c icom_port->uart_port.icount.tx += count; tx 503 drivers/tty/serial/imx.c sport->port.icount.tx++; tx 538 drivers/tty/serial/imx.c sport->port.icount.tx++; tx 566 drivers/tty/serial/imx.c sport->port.icount.tx += sport->tx_bytes; tx 398 drivers/tty/serial/ip22zilog.c up->port.icount.tx++; tx 417 drivers/tty/serial/ip22zilog.c up->port.icount.tx++; tx 605 drivers/tty/serial/ip22zilog.c port->icount.tx++; tx 617 drivers/tty/serial/ip22zilog.c port->icount.tx++; tx 237 drivers/tty/serial/lantiq.c port->icount.tx++; tx 248 drivers/tty/serial/lantiq.c port->icount.tx++; tx 290 drivers/tty/serial/lpc32xx_hs.c port->icount.tx++; tx 304 drivers/tty/serial/lpc32xx_hs.c port->icount.tx++; tx 191 drivers/tty/serial/max3100.c static int max3100_sr(struct max3100_port *s, u16 tx, u16 *rx) tx 202 drivers/tty/serial/max3100.c etx = cpu_to_be16(tx); tx 212 drivers/tty/serial/max3100.c dev_dbg(&s->spi->dev, "%04x - %04x\n", tx, *rx); tx 260 drivers/tty/serial/max3100.c u16 tx, rx; tx 287 drivers/tty/serial/max3100.c tx = 0xffff; tx 289 drivers/tty/serial/max3100.c tx = s->port.x_char; tx 290 drivers/tty/serial/max3100.c s->port.icount.tx++; tx 294 drivers/tty/serial/max3100.c tx = xmit->buf[xmit->tail]; tx 297 drivers/tty/serial/max3100.c s->port.icount.tx++; tx 299 drivers/tty/serial/max3100.c if (tx != 0xffff) { tx 300 drivers/tty/serial/max3100.c max3100_calc_parity(s, &tx); tx 301 drivers/tty/serial/max3100.c tx |= MAX3100_WD | (s->rts ? MAX3100_RTS : 0); tx 302 drivers/tty/serial/max3100.c max3100_sr(s, tx, &rx); tx 571 drivers/tty/serial/max3100.c u16 tx, rx; tx 573 drivers/tty/serial/max3100.c tx = MAX3100_WC | MAX3100_SHDN; tx 574 drivers/tty/serial/max3100.c max3100_sr(s, tx, &rx); tx 616 drivers/tty/serial/max3100.c u16 tx, rx; tx 617 drivers/tty/serial/max3100.c tx = 0x4001; tx 618 drivers/tty/serial/max3100.c max3100_sr(s, tx, &rx); tx 741 drivers/tty/serial/max3100.c u16 tx, rx; tx 804 drivers/tty/serial/max3100.c tx = MAX3100_WC | MAX3100_SHDN; tx 805 drivers/tty/serial/max3100.c max3100_sr(max3100s[i], tx, &rx); tx 860 drivers/tty/serial/max3100.c u16 tx, rx; tx 862 drivers/tty/serial/max3100.c tx = MAX3100_WC | MAX3100_SHDN; tx 863 drivers/tty/serial/max3100.c max3100_sr(s, tx, &rx); tx 755 drivers/tty/serial/max310x.c port->icount.tx++; tx 782 drivers/tty/serial/max310x.c port->icount.tx += to_send; tx 338 drivers/tty/serial/mcf.c port->icount.tx++; tx 347 drivers/tty/serial/mcf.c port->icount.tx++; tx 359 drivers/tty/serial/men_z135_uart.c port->icount.tx += n; tx 151 drivers/tty/serial/meson_uart.c port->icount.tx++; tx 162 drivers/tty/serial/meson_uart.c port->icount.tx++; tx 90 drivers/tty/serial/milbeaut_usio.c port->icount.tx++; tx 106 drivers/tty/serial/milbeaut_usio.c port->icount.tx++; tx 1447 drivers/tty/serial/mpc52xx_uart.c port->icount.tx++; tx 1462 drivers/tty/serial/mpc52xx_uart.c port->icount.tx++; tx 138 drivers/tty/serial/mps2-uart.c port->icount.tx++; tx 147 drivers/tty/serial/mps2-uart.c port->icount.tx++; tx 320 drivers/tty/serial/mps2-uart.c MAKE_NAME(-tx), mps_port); tx 458 drivers/tty/serial/msm_serial.c port->icount.tx += count; tx 845 drivers/tty/serial/msm_serial.c port->icount.tx++; tx 883 drivers/tty/serial/msm_serial.c port->icount.tx++; tx 192 drivers/tty/serial/mux.c port->icount.tx++; tx 206 drivers/tty/serial/mux.c port->icount.tx++; tx 200 drivers/tty/serial/mvebu-uart.c port->icount.tx++; tx 306 drivers/tty/serial/mvebu-uart.c port->icount.tx++; tx 319 drivers/tty/serial/mvebu-uart.c port->icount.tx++; tx 623 drivers/tty/serial/mxs-auart.c s->port.icount.tx++; tx 629 drivers/tty/serial/mxs-auart.c s->port.icount.tx++; tx 374 drivers/tty/serial/omap-serial.c up->port.icount.tx++; tx 386 drivers/tty/serial/omap-serial.c up->port.icount.tx++; tx 190 drivers/tty/serial/owl-uart.c port->icount.tx++; tx 201 drivers/tty/serial/owl-uart.c port->icount.tx++; tx 787 drivers/tty/serial/pch_uart.c port->icount.tx += sg_dma_len(sg); tx 905 drivers/tty/serial/pch_uart.c port->icount.tx++; tx 915 drivers/tty/serial/pch_uart.c port->icount.tx += tx_size; tx 962 drivers/tty/serial/pch_uart.c port->icount.tx++; tx 285 drivers/tty/serial/pic32_uart.c port->icount.tx++; tx 314 drivers/tty/serial/pic32_uart.c port->icount.tx++; tx 404 drivers/tty/serial/pmac_zilog.c uap->port.icount.tx++; tx 424 drivers/tty/serial/pmac_zilog.c uap->port.icount.tx++; tx 636 drivers/tty/serial/pmac_zilog.c port->icount.tx++; tx 646 drivers/tty/serial/pmac_zilog.c port->icount.tx++; tx 252 drivers/tty/serial/pnx8xxx_uart.c sport->port.icount.tx++; tx 275 drivers/tty/serial/pnx8xxx_uart.c sport->port.icount.tx++; tx 185 drivers/tty/serial/pxa.c up->port.icount.tx++; tx 198 drivers/tty/serial/pxa.c up->port.icount.tx++; tx 740 drivers/tty/serial/qcom_geni_serial.c uport->icount.tx += tx_bytes; tx 344 drivers/tty/serial/rda-uart.c port->icount.tx++; tx 355 drivers/tty/serial/rda-uart.c port->icount.tx++; tx 446 drivers/tty/serial/rp2.c up->port.icount.tx++; tx 455 drivers/tty/serial/rp2.c up->port.icount.tx++; tx 243 drivers/tty/serial/sa1100.c sport->port.icount.tx++; tx 266 drivers/tty/serial/sa1100.c sport->port.icount.tx++; tx 182 drivers/tty/serial/samsung.c port->icount.tx += count; tx 217 drivers/tty/serial/samsung.c port->icount.tx += count; tx 728 drivers/tty/serial/samsung.c port->icount.tx++; tx 755 drivers/tty/serial/samsung.c port->icount.tx++; tx 396 drivers/tty/serial/sb1250-duart.c sport->port.icount.tx++; tx 408 drivers/tty/serial/sb1250-duart.c sport->port.icount.tx++; tx 641 drivers/tty/serial/sc16is7xx.c port->icount.tx++; tx 663 drivers/tty/serial/sc16is7xx.c port->icount.tx += to_send; tx 452 drivers/tty/serial/sccnxp.c port->icount.tx++; tx 476 drivers/tty/serial/sccnxp.c port->icount.tx++; tx 494 drivers/tty/serial/serial-tegra.c tup->uport.icount.tx++; tx 1273 drivers/tty/serial/serial_core.c icount->tx = cnow.tx; tx 1857 drivers/tty/serial/serial_core.c uport->icount.tx, uport->icount.rx); tx 350 drivers/tty/serial/serial_txx9.c up->port.icount.tx++; tx 363 drivers/tty/serial/serial_txx9.c up->port.icount.tx++; tx 834 drivers/tty/serial/sh-sci.c port->icount.tx++; tx 1195 drivers/tty/serial/sh-sci.c port->icount.tx += s->tx_dma_len; tx 307 drivers/tty/serial/sifive.c ssp->port.icount.tx++; tx 319 drivers/tty/serial/sifive.c ssp->port.icount.tx++; tx 435 drivers/tty/serial/sirfsoc_uart.c port->icount.tx++; tx 453 drivers/tty/serial/sirfsoc_uart.c port->icount.tx += sirfport->transfer_size; tx 223 drivers/tty/serial/sprd_serial.c port->icount.tx += trans_len; tx 261 drivers/tty/serial/sprd_serial.c port->icount.tx += sp->tx_dma.trans_len; tx 333 drivers/tty/serial/sprd_serial.c port->icount.tx++; tx 638 drivers/tty/serial/sprd_serial.c port->icount.tx++; tx 652 drivers/tty/serial/sprd_serial.c port->icount.tx++; tx 255 drivers/tty/serial/st-asc.c port->icount.tx++; tx 280 drivers/tty/serial/st-asc.c port->icount.tx++; tx 342 drivers/tty/serial/stm32-usart.c port->icount.tx++; tx 410 drivers/tty/serial/stm32-usart.c port->icount.tx += count; tx 424 drivers/tty/serial/stm32-usart.c port->icount.tx++; tx 55 drivers/tty/serial/sunhv.c port->icount.tx++; tx 71 drivers/tty/serial/sunhv.c port->icount.tx += sent; tx 274 drivers/tty/serial/sunsab.c up->port.icount.tx++; tx 461 drivers/tty/serial/sunsab.c up->port.icount.tx++; tx 407 drivers/tty/serial/sunsu.c up->port.icount.tx++; tx 424 drivers/tty/serial/sunsu.c up->port.icount.tx++; tx 496 drivers/tty/serial/sunzilog.c up->port.icount.tx++; tx 516 drivers/tty/serial/sunzilog.c up->port.icount.tx++; tx 705 drivers/tty/serial/sunzilog.c port->icount.tx++; tx 717 drivers/tty/serial/sunzilog.c port->icount.tx++; tx 31 drivers/tty/serial/tegra-tcu.c struct mbox_chan *tx, *rx; tx 59 drivers/tty/serial/tegra-tcu.c mbox_send_message(tcu->tx, msg); tx 60 drivers/tty/serial/tegra-tcu.c mbox_flush(tcu->tx, 1000); tx 191 drivers/tty/serial/tegra-tcu.c tcu->tx = mbox_request_channel_byname(&tcu->tx_client, "tx"); tx 192 drivers/tty/serial/tegra-tcu.c if (IS_ERR(tcu->tx)) { tx 193 drivers/tty/serial/tegra-tcu.c err = PTR_ERR(tcu->tx); tx 261 drivers/tty/serial/tegra-tcu.c mbox_free_channel(tcu->tx); tx 276 drivers/tty/serial/tegra-tcu.c mbox_free_channel(tcu->tx); tx 107 drivers/tty/serial/timbuart.c port->icount.tx++; tx 113 drivers/tty/serial/timbuart.c port->icount.tx, tx 186 drivers/tty/serial/uartlite.c port->icount.tx++; tx 195 drivers/tty/serial/uartlite.c port->icount.tx++; tx 354 drivers/tty/serial/ucc_uart.c port->icount.tx++; tx 374 drivers/tty/serial/ucc_uart.c port->icount.tx++; tx 386 drivers/tty/serial/vr41xx_siu.c port->icount.tx++; tx 399 drivers/tty/serial/vr41xx_siu.c port->icount.tx++; tx 202 drivers/tty/serial/vt8500_serial.c port->icount.tx++; tx 217 drivers/tty/serial/vt8500_serial.c port->icount.tx++; tx 328 drivers/tty/serial/xilinx_uartps.c port->icount.tx++; tx 617 drivers/tty/serial/zs.c zport->port.icount.tx++; tx 631 drivers/tty/serial/zs.c zport->port.icount.tx++; tx 2931 drivers/tty/synclink.c icount->tx = cnow.tx; tx 3483 drivers/tty/synclink.c info->icount.tx, info->icount.rx); tx 5654 drivers/tty/synclink.c info->icount.tx += 2; tx 5670 drivers/tty/synclink.c info->icount.tx++; tx 1113 drivers/tty/synclink_gt.c icount->tx = cnow.tx; tx 1275 drivers/tty/synclink_gt.c info->icount.tx, info->icount.rx); tx 1322 drivers/tty/synclinkmp.c icount->tx = cnow.tx; tx 1388 drivers/tty/synclinkmp.c info->icount.tx, info->icount.rx); tx 4304 drivers/tty/synclinkmp.c info->icount.tx += 2; tx 4318 drivers/tty/synclinkmp.c info->icount.tx++; tx 601 drivers/usb/atm/usbatm.c atomic_inc(&vcc->stats->tx); tx 745 drivers/usb/atm/usbatm.c atomic_read(&atm_dev->stats.aal5.tx), tx 148 drivers/usb/host/max3421-hcd.c struct max3421_dma_buf *tx; tx 364 drivers/usb/host/max3421-hcd.c max3421_hcd->tx->data[0] = tx 368 drivers/usb/host/max3421-hcd.c transfer.tx_buf = max3421_hcd->tx->data; tx 390 drivers/usb/host/max3421-hcd.c max3421_hcd->tx->data[0] = tx 393 drivers/usb/host/max3421-hcd.c max3421_hcd->tx->data[1] = val; tx 395 drivers/usb/host/max3421-hcd.c transfer.tx_buf = max3421_hcd->tx->data; tx 414 drivers/usb/host/max3421-hcd.c max3421_hcd->tx->data[0] = tx 417 drivers/usb/host/max3421-hcd.c transfer[0].tx_buf = max3421_hcd->tx->data; tx 440 drivers/usb/host/max3421-hcd.c max3421_hcd->tx->data[0] = tx 444 drivers/usb/host/max3421-hcd.c transfer[0].tx_buf = max3421_hcd->tx->data; tx 1904 drivers/usb/host/max3421-hcd.c max3421_hcd->tx = kmalloc(sizeof(*max3421_hcd->tx), GFP_KERNEL); tx 1905 drivers/usb/host/max3421-hcd.c if (!max3421_hcd->tx) tx 1940 drivers/usb/host/max3421-hcd.c kfree(max3421_hcd->tx); tx 98 drivers/usb/musb/cppi_dma.c static void cppi_reset_tx(struct cppi_tx_stateram __iomem *tx, u32 ptr) tx 100 drivers/usb/musb/cppi_dma.c musb_writel(&tx->tx_head, 0, 0); tx 101 drivers/usb/musb/cppi_dma.c musb_writel(&tx->tx_buf, 0, 0); tx 102 drivers/usb/musb/cppi_dma.c musb_writel(&tx->tx_current, 0, 0); tx 103 drivers/usb/musb/cppi_dma.c musb_writel(&tx->tx_buf_current, 0, 0); tx 104 drivers/usb/musb/cppi_dma.c musb_writel(&tx->tx_info, 0, 0); tx 105 drivers/usb/musb/cppi_dma.c musb_writel(&tx->tx_rem_len, 0, 0); tx 107 drivers/usb/musb/cppi_dma.c musb_writel(&tx->tx_complete, 0, ptr); tx 161 drivers/usb/musb/cppi_dma.c for (i = 0; i < ARRAY_SIZE(controller->tx); i++) { tx 162 drivers/usb/musb/cppi_dma.c controller->tx[i].transmit = true; tx 163 drivers/usb/musb/cppi_dma.c controller->tx[i].index = i; tx 171 drivers/usb/musb/cppi_dma.c for (i = 0; i < ARRAY_SIZE(controller->tx); i++) tx 172 drivers/usb/musb/cppi_dma.c cppi_pool_init(controller, controller->tx + i); tx 180 drivers/usb/musb/cppi_dma.c for (i = 0; i < ARRAY_SIZE(controller->tx); i++) { tx 181 drivers/usb/musb/cppi_dma.c struct cppi_channel *tx_ch = controller->tx + i; tx 182 drivers/usb/musb/cppi_dma.c struct cppi_tx_stateram __iomem *tx; tx 186 drivers/usb/musb/cppi_dma.c tx = tibase + DAVINCI_TXCPPI_STATERAM_OFFSET(i); tx 187 drivers/usb/musb/cppi_dma.c tx_ch->state_ram = tx; tx 188 drivers/usb/musb/cppi_dma.c cppi_reset_tx(tx, 0); tx 238 drivers/usb/musb/cppi_dma.c for (i = 0; i < ARRAY_SIZE(controller->tx); i++) { tx 240 drivers/usb/musb/cppi_dma.c controller->tx[i].last_processed = NULL; tx 241 drivers/usb/musb/cppi_dma.c cppi_pool_free(controller->tx + i); tx 301 drivers/usb/musb/cppi_dma.c if (index >= ARRAY_SIZE(controller->tx)) { tx 305 drivers/usb/musb/cppi_dma.c cppi_ch = controller->tx + index; tx 385 drivers/usb/musb/cppi_dma.c struct cppi_tx_stateram __iomem *tx = c->state_ram; tx 396 drivers/usb/musb/cppi_dma.c musb_readl(&tx->tx_head, 0), tx 397 drivers/usb/musb/cppi_dma.c musb_readl(&tx->tx_buf, 0), tx 398 drivers/usb/musb/cppi_dma.c musb_readl(&tx->tx_current, 0), tx 399 drivers/usb/musb/cppi_dma.c musb_readl(&tx->tx_buf_current, 0), tx 401 drivers/usb/musb/cppi_dma.c musb_readl(&tx->tx_info, 0), tx 402 drivers/usb/musb/cppi_dma.c musb_readl(&tx->tx_rem_len, 0), tx 404 drivers/usb/musb/cppi_dma.c musb_readl(&tx->tx_complete, 0) tx 561 drivers/usb/musb/cppi_dma.c cppi_next_tx_segment(struct musb *musb, struct cppi_channel *tx) tx 563 drivers/usb/musb/cppi_dma.c unsigned maxpacket = tx->maxpacket; tx 564 drivers/usb/musb/cppi_dma.c dma_addr_t addr = tx->buf_dma + tx->offset; tx 565 drivers/usb/musb/cppi_dma.c size_t length = tx->buf_len - tx->offset; tx 569 drivers/usb/musb/cppi_dma.c struct cppi_tx_stateram __iomem *tx_ram = tx->state_ram; tx 595 drivers/usb/musb/cppi_dma.c tx->index, tx 601 drivers/usb/musb/cppi_dma.c cppi_rndis_update(tx, 0, musb->ctrl_base, rndis); tx 609 drivers/usb/musb/cppi_dma.c bd = tx->freelist; tx 610 drivers/usb/musb/cppi_dma.c tx->head = bd; tx 611 drivers/usb/musb/cppi_dma.c tx->last_processed = NULL; tx 627 drivers/usb/musb/cppi_dma.c bd->hw_bufp = tx->buf_dma + tx->offset; tx 632 drivers/usb/musb/cppi_dma.c if ((tx->offset + maxpacket) <= tx->buf_len) { tx 633 drivers/usb/musb/cppi_dma.c tx->offset += maxpacket; tx 641 drivers/usb/musb/cppi_dma.c partial_len = tx->buf_len - tx->offset; tx 642 drivers/usb/musb/cppi_dma.c tx->offset = tx->buf_len; tx 656 drivers/usb/musb/cppi_dma.c tx->tail = bd; tx 664 drivers/usb/musb/cppi_dma.c musb_writel(&tx_ram->tx_head, 0, (u32)tx->freelist->dma); tx 666 drivers/usb/musb/cppi_dma.c cppi_dump_tx(5, tx, "/S"); tx 1147 drivers/usb/musb/cppi_dma.c u32 rx, tx; tx 1157 drivers/usb/musb/cppi_dma.c tx = musb_readl(tibase, DAVINCI_TXCPPI_MASKED_REG); tx 1160 drivers/usb/musb/cppi_dma.c if (!tx && !rx) { tx 1166 drivers/usb/musb/cppi_dma.c musb_dbg(musb, "CPPI IRQ Tx%x Rx%x", tx, rx); tx 1169 drivers/usb/musb/cppi_dma.c for (index = 0; tx; tx = tx >> 1, index++) { tx 1175 drivers/usb/musb/cppi_dma.c if (!(tx & 1)) tx 1178 drivers/usb/musb/cppi_dma.c tx_ch = cppi->tx + index; tx 116 drivers/usb/musb/cppi_dma.h struct cppi_channel tx[4]; tx 62 drivers/usb/musb/musb_gadget.c request->tx tx 75 drivers/usb/musb/musb_gadget.c request->tx tx 100 drivers/usb/musb/musb_gadget.c request->tx tx 108 drivers/usb/musb/musb_gadget.c request->tx tx 1173 drivers/usb/musb/musb_gadget.c if (req->tx) tx 1228 drivers/usb/musb/musb_gadget.c request->tx = musb_ep->is_in; tx 58 drivers/usb/musb/musb_gadget.h u8 tx; /* endpoint direction */ tx 927 drivers/usb/musb/musb_gadget_ep0.c req->tx = ep->is_in; tx 226 drivers/usb/musb/musb_trace.h __entry->is_tx = req->tx; tx 34 drivers/usb/musb/tusb6010_omap.c u8 tx; tx 71 drivers/usb/musb/tusb6010_omap.c if (chdat->tx) tx 112 drivers/usb/musb/tusb6010_omap.c chdat->epnum, chdat->tx ? "tx" : "rx"); tx 114 drivers/usb/musb/tusb6010_omap.c if (chdat->tx) tx 124 drivers/usb/musb/tusb6010_omap.c chdat->tx ? "tx" : "rx", remaining); tx 139 drivers/usb/musb/tusb6010_omap.c if (chdat->tx) { tx 158 drivers/usb/musb/tusb6010_omap.c musb_dma_completion(musb, chdat->epnum, chdat->tx); tx 168 drivers/usb/musb/tusb6010_omap.c if (chdat->tx) { tx 219 drivers/usb/musb/tusb6010_omap.c if (chdat->tx) tx 227 drivers/usb/musb/tusb6010_omap.c chdat->tx ? "tx" : "rx", dma_remaining); tx 260 drivers/usb/musb/tusb6010_omap.c if (chdat->tx) { tx 287 drivers/usb/musb/tusb6010_omap.c chdat->epnum, chdat->tx ? "tx" : "rx", &dma_addr, tx 317 drivers/usb/musb/tusb6010_omap.c chdat->epnum, chdat->tx ? "tx" : "rx", tx 327 drivers/usb/musb/tusb6010_omap.c if (chdat->tx) { tx 344 drivers/usb/musb/tusb6010_omap.c if (chdat->tx) { tx 396 drivers/usb/musb/tusb6010_omap.c if (chdat->tx) tx 424 drivers/usb/musb/tusb6010_omap.c u8 tx) tx 438 drivers/usb/musb/tusb6010_omap.c dev_dbg(musb->controller, "Not allowing DMA for ep0 %s\n", tx ? "tx" : "rx"); tx 461 drivers/usb/musb/tusb6010_omap.c if (tx) tx 462 drivers/usb/musb/tusb6010_omap.c chdat->tx = 1; tx 464 drivers/usb/musb/tusb6010_omap.c chdat->tx = 0; tx 484 drivers/usb/musb/tusb6010_omap.c chdat->tx ? "tx" : "rx", tx 2466 drivers/usb/serial/ftdi_sio.c port->icount.tx += c; tx 2474 drivers/usb/serial/ftdi_sio.c port->icount.tx += count; tx 562 drivers/usb/serial/generic.c icount->tx = cnow.tx; tx 1372 drivers/usb/serial/io_edgeport.c edge_port->port->icount.tx += count; tx 1384 drivers/usb/serial/io_edgeport.c edge_port->port->icount.tx -= count; tx 2078 drivers/usb/serial/io_ti.c edge_port->port->icount.tx += count; tx 1375 drivers/usb/serial/mos7840.c port->icount.tx += transfer_size; tx 1376 drivers/usb/serial/mos7840.c dev_dbg(&port->dev, "icount.tx is %d:\n", port->icount.tx); tx 1334 drivers/usb/serial/ti_usb_3410_5052.c port->icount.tx += count; tx 388 drivers/video/fbdev/mmp/hw/mmp_ctrl.h #define CFG_TXBITS(tx) (((tx) - 1)<<8) /* 0x1F~0x1 */ tx 396 drivers/video/fbdev/mmp/hw/mmp_ctrl.h #define CFG_TXBITSTO0(tx) ((tx)<<4) tx 179 drivers/video/fbdev/pxa168fb.h #define CFG_TXBITS(tx) ((tx) << 8) /* 0x1F~0x1 */ tx 187 drivers/video/fbdev/pxa168fb.h #define CFG_TXBITSTO0(tx) ((tx) << 4) tx 589 fs/dlm/lowcomms.c bool tx, bool rx) tx 593 fs/dlm/lowcomms.c if (tx && !closing && cancel_work_sync(&con->swork)) { tx 195 include/acpi/processor.h int tx; /* throttle level */ tx 88 include/linux/async_tx.h static inline void async_tx_issue_pending(struct dma_async_tx_descriptor *tx) tx 90 include/linux/async_tx.h if (likely(tx)) { tx 91 include/linux/async_tx.h struct dma_chan *chan = tx->chan; tx 112 include/linux/async_tx.h static inline void async_tx_issue_pending(struct dma_async_tx_descriptor *tx) tx 147 include/linux/async_tx.h struct dma_async_tx_descriptor *tx, tx 152 include/linux/async_tx.h args->depend_tx = tx; tx 158 include/linux/async_tx.h void async_tx_submit(struct dma_chan *chan, struct dma_async_tx_descriptor *tx, tx 194 include/linux/async_tx.h void async_tx_quiesce(struct dma_async_tx_descriptor **tx); tx 105 include/linux/cyclades.h __u32 cts, dsr, rng, dcd, tx, rx; tx 171 include/linux/dma/ipu-dma.h #define to_tx_desc(tx) container_of(tx, struct idmac_tx_desc, txd) tx 501 include/linux/dmaengine.h dma_cookie_t (*tx_submit)(struct dma_async_tx_descriptor *tx); tx 502 include/linux/dmaengine.h int (*desc_free)(struct dma_async_tx_descriptor *tx); tx 515 include/linux/dmaengine.h static inline void dma_set_unmap(struct dma_async_tx_descriptor *tx, tx 519 include/linux/dmaengine.h tx->unmap = unmap; tx 526 include/linux/dmaengine.h static inline void dma_set_unmap(struct dma_async_tx_descriptor *tx, tx 540 include/linux/dmaengine.h static inline void dma_descriptor_unmap(struct dma_async_tx_descriptor *tx) tx 542 include/linux/dmaengine.h if (tx->unmap) { tx 543 include/linux/dmaengine.h dmaengine_unmap_put(tx->unmap); tx 544 include/linux/dmaengine.h tx->unmap = NULL; tx 1184 include/linux/dmaengine.h void dma_async_tx_descriptor_init(struct dma_async_tx_descriptor *tx, tx 1187 include/linux/dmaengine.h static inline void async_tx_ack(struct dma_async_tx_descriptor *tx) tx 1189 include/linux/dmaengine.h tx->flags |= DMA_CTRL_ACK; tx 1192 include/linux/dmaengine.h static inline void async_tx_clear_ack(struct dma_async_tx_descriptor *tx) tx 1194 include/linux/dmaengine.h tx->flags &= ~DMA_CTRL_ACK; tx 1197 include/linux/dmaengine.h static inline bool async_tx_test_ack(struct dma_async_tx_descriptor *tx) tx 1199 include/linux/dmaengine.h return (tx->flags & DMA_CTRL_ACK) == DMA_CTRL_ACK; tx 1202 include/linux/dmaengine.h #define dma_cap_set(tx, mask) __dma_cap_set((tx), &(mask)) tx 1209 include/linux/dmaengine.h #define dma_cap_clear(tx, mask) __dma_cap_clear((tx), &(mask)) tx 1222 include/linux/dmaengine.h #define dma_has_cap(tx, mask) __dma_has_cap((tx), &(mask)) tx 1304 include/linux/dmaengine.h enum dma_status dma_wait_for_async_tx(struct dma_async_tx_descriptor *tx); tx 1325 include/linux/dmaengine.h static inline enum dma_status dma_wait_for_async_tx(struct dma_async_tx_descriptor *tx) tx 1366 include/linux/dmaengine.h static inline int dmaengine_desc_set_reuse(struct dma_async_tx_descriptor *tx) tx 1371 include/linux/dmaengine.h ret = dma_get_slave_caps(tx->chan, &caps); tx 1376 include/linux/dmaengine.h tx->flags |= DMA_CTRL_REUSE; tx 1383 include/linux/dmaengine.h static inline void dmaengine_desc_clear_reuse(struct dma_async_tx_descriptor *tx) tx 1385 include/linux/dmaengine.h tx->flags &= ~DMA_CTRL_REUSE; tx 1388 include/linux/dmaengine.h static inline bool dmaengine_desc_test_reuse(struct dma_async_tx_descriptor *tx) tx 1390 include/linux/dmaengine.h return (tx->flags & DMA_CTRL_REUSE) == DMA_CTRL_REUSE; tx 1407 include/linux/dmaengine.h void dma_run_dependencies(struct dma_async_tx_descriptor *tx); tx 70 include/linux/iio/imu/adis.h uint8_t tx[10] ____cacheline_aligned; tx 114 include/linux/mfd/ipaq-micro.h struct ipaq_micro_txdev tx; /* transmit ISR state */ tx 403 include/linux/mlx5/mlx5_ifc_fpga.h u8 tx[0x1]; tx 45 include/linux/moxtet.h u8 tx[TURRIS_MOX_MAX_MODULES]; tx 1165 include/linux/phy.h void phy_set_sym_pause(struct phy_device *phydev, bool rx, bool tx, tx 1167 include/linux/phy.h void phy_set_asym_pause(struct phy_device *phydev, bool rx, bool tx); tx 41 include/linux/posix-clock.h int (*clock_adjtime)(struct posix_clock *pc, struct __kernel_timex *tx); tx 21 include/linux/serial.h __u32 cts, dsr, rng, dcd, tx, rx; tx 94 include/linux/serial_core.h __u32 tx; tx 133 include/linux/soc/ti/knav_dma.h struct knav_dma_tx_cfg tx; tx 929 include/linux/syscalls.h struct __kernel_timex __user *tx); tx 931 include/linux/syscalls.h struct old_timex32 __user *tx); tx 4155 include/net/cfg80211.h u16 tx, rx; tx 3757 include/net/mac80211.h void (*tx)(struct ieee80211_hw *hw, tx 3931 include/net/mac80211.h int (*set_ringparam)(struct ieee80211_hw *hw, u32 tx, u32 rx); tx 3933 include/net/mac80211.h u32 *tx, u32 *tx_max, u32 *rx, u32 *rx_max); tx 861 include/net/tcp.h } tx; /* only used for outgoing skbs */ tx 268 include/net/tls.h struct cipher_context tx; tx 358 include/net/tls.h int tls_set_sw_offload(struct sock *sk, struct tls_context *ctx, int tx); tx 430 include/net/tls.h static inline u16 tls_user_config(struct tls_context *ctx, bool tx) tx 432 include/net/tls.h u16 config = tx ? ctx->tx_conf : ctx->rx_conf; tx 517 include/net/tls.h ctx->tx.iv + TLS_CIPHER_AES_GCM_128_SALT_SIZE, iv_size); tx 95 include/net/xdp_sock.h struct xsk_queue *tx ____cacheline_aligned_in_smp; tx 202 include/soc/fsl/qe/immap_qe.h u8 tx[0x400]; tx 100 include/soc/tegra/bpmp.h } tx; tx 21 include/soc/tegra/ivc.h } rx, tx; tx 95 include/soc/tegra/ivc.h dma_addr_t rx_phys, void *tx, dma_addr_t tx_phys, tx 15 include/uapi/linux/atm_eni.h int tx,rx; /* values are in percent and must be > 100 */ tx 35 include/uapi/linux/atmdev.h __HANDLE_ITEM(tx); /* TX okay */ \ tx 52 include/uapi/linux/if_xdp.h struct xdp_ring_offset tx; tx 104 include/uapi/linux/serial.h int rx, tx; tx 214 include/uapi/linux/synclink.h __u32 cts, dsr, rng, dcd, tx, rx; tx 226 kernel/time/posix-clock.c static int pc_clock_adjtime(clockid_t id, struct __kernel_timex *tx) tx 241 kernel/time/posix-clock.c err = cd.clk->ops.clock_adjtime(cd.clk, tx); tx 11 kernel/time/posix-timers.h int (*clock_adj)(const clockid_t which_clock, struct __kernel_timex *tx); tx 48 net/atm/proc.c atomic_read(&stats->tx), atomic_read(&stats->tx_err), tx 275 net/can/j1939/j1939-priv.h unsigned int tx; tx 704 net/can/j1939/transport.c session->pkt.tx = 0; tx 751 net/can/j1939/transport.c session->pkt.tx = session->pkt.tx_acked; tx 783 net/can/j1939/transport.c while (session->pkt.tx < pkt_end) { tx 784 net/can/j1939/transport.c dat[0] = session->pkt.tx - session->pkt.dpo + 1; tx 785 net/can/j1939/transport.c offset = (session->pkt.tx * 7) - skcb->offset; tx 803 net/can/j1939/transport.c session->pkt.tx++; tx 807 net/can/j1939/transport.c if (session->pkt.tx < session->pkt.total && pdelay) { tx 1375 net/can/j1939/transport.c session->pkt.tx = session->pkt.tx_acked; tx 1572 net/can/j1939/transport.c session->pkt.tx = 0; tx 8916 net/core/dev.c struct netdev_queue *tx; tx 8917 net/core/dev.c size_t sz = count * sizeof(*tx); tx 8922 net/core/dev.c tx = kvzalloc(sz, GFP_KERNEL | __GFP_RETRY_MAYFAIL); tx 8923 net/core/dev.c if (!tx) tx 8926 net/core/dev.c dev->_tx = tx; tx 1448 net/dccp/feat.c } tx, rx; tx 1474 net/dccp/feat.c if (ccid_get_builtin_ccids(&tx.val, &tx.len)) tx 1477 net/dccp/feat.c kfree(tx.val); tx 1481 net/dccp/feat.c if (!dccp_feat_prefer(sysctl_dccp_tx_ccid, tx.val, tx.len) || tx 1485 net/dccp/feat.c rc = __feat_register_sp(fn, DCCPF_CCID, true, false, tx.val, tx.len); tx 1492 net/dccp/feat.c kfree(tx.val); tx 1354 net/ipv4/tcp_input.c if (unlikely(TCP_SKB_CB(prev)->tx.delivered_mstamp)) tx 1355 net/ipv4/tcp_input.c TCP_SKB_CB(prev)->tx.delivered_mstamp = 0; tx 3123 net/ipv4/tcp_input.c last_in_flight = TCP_SKB_CB(skb)->tx.in_flight; tx 1038 net/ipv4/tcp_output.c TCP_SKB_CB(skb)->tx.in_flight = TCP_SKB_CB(skb)->end_seq tx 1376 net/ipv4/tcp_output.c TCP_SKB_CB(buff)->tx = TCP_SKB_CB(skb)->tx; tx 65 net/ipv4/tcp_rate.c TCP_SKB_CB(skb)->tx.first_tx_mstamp = tp->first_tx_mstamp; tx 66 net/ipv4/tcp_rate.c TCP_SKB_CB(skb)->tx.delivered_mstamp = tp->delivered_mstamp; tx 67 net/ipv4/tcp_rate.c TCP_SKB_CB(skb)->tx.delivered = tp->delivered; tx 68 net/ipv4/tcp_rate.c TCP_SKB_CB(skb)->tx.is_app_limited = tp->app_limited ? 1 : 0; tx 84 net/ipv4/tcp_rate.c if (!scb->tx.delivered_mstamp) tx 88 net/ipv4/tcp_rate.c after(scb->tx.delivered, rs->prior_delivered)) { tx 89 net/ipv4/tcp_rate.c rs->prior_delivered = scb->tx.delivered; tx 90 net/ipv4/tcp_rate.c rs->prior_mstamp = scb->tx.delivered_mstamp; tx 91 net/ipv4/tcp_rate.c rs->is_app_limited = scb->tx.is_app_limited; tx 98 net/ipv4/tcp_rate.c scb->tx.first_tx_mstamp); tx 106 net/ipv4/tcp_rate.c scb->tx.delivered_mstamp = 0; tx 59 net/mac80211/agg-rx.c u16 initiator, u16 reason, bool tx) tx 95 net/mac80211/agg-rx.c if (initiator == WLAN_BACK_RECIPIENT && tx) tx 118 net/mac80211/agg-rx.c u16 initiator, u16 reason, bool tx) tx 121 net/mac80211/agg-rx.c ___ieee80211_stop_rx_ba_session(sta, tid, initiator, reason, tx); tx 271 net/mac80211/agg-rx.c u16 buf_size, bool tx, bool auto_seq, tx 453 net/mac80211/agg-rx.c if (tx) tx 462 net/mac80211/agg-rx.c u16 tid, u16 buf_size, bool tx, tx 469 net/mac80211/agg-rx.c buf_size, tx, auto_seq, addbaext); tx 303 net/mac80211/debugfs_sta.c bool start, tx; tx 321 net/mac80211/debugfs_sta.c tx = true; tx 323 net/mac80211/debugfs_sta.c tx = false; tx 332 net/mac80211/debugfs_sta.c if (!tx) tx 345 net/mac80211/debugfs_sta.c if (!buf || !tx || !start) tx 353 net/mac80211/debugfs_sta.c if (tx) { tx 36 net/mac80211/driver-ops.h local->ops->tx(&local->hw, control, skb); tx 711 net/mac80211/driver-ops.h u32 tx, u32 rx) tx 717 net/mac80211/driver-ops.h trace_drv_set_ringparam(local, tx, rx); tx 719 net/mac80211/driver-ops.h ret = local->ops->set_ringparam(&local->hw, tx, rx); tx 726 net/mac80211/driver-ops.h u32 *tx, u32 *tx_max, u32 *rx, u32 *rx_max) tx 730 net/mac80211/driver-ops.h trace_drv_get_ringparam(local, tx, tx_max, rx, rx_max); tx 732 net/mac80211/driver-ops.h local->ops->get_ringparam(&local->hw, tx, tx_max, rx, rx_max); tx 1595 net/mac80211/ieee80211_i.h void ieee80211_tx_set_protected(struct ieee80211_tx_data *tx); tx 1810 net/mac80211/ieee80211_i.h u16 buf_size, bool tx, bool auto_seq, tx 75 net/mac80211/key.h struct tkip_ctx tx; tx 411 net/mac80211/main.c .tx = 0xffff, tx 418 net/mac80211/main.c .tx = 0xffff, tx 423 net/mac80211/main.c .tx = 0xffff, tx 434 net/mac80211/main.c .tx = 0xffff, tx 444 net/mac80211/main.c .tx = 0xffff, tx 449 net/mac80211/main.c .tx = 0xffff, tx 459 net/mac80211/main.c .tx = 0xffff, tx 465 net/mac80211/main.c .tx = 0xffff, tx 516 net/mac80211/main.c if (WARN_ON(!ops->tx || !ops->start || !ops->stop || !ops->config || tx 252 net/mac80211/mesh.h bool tx, bool acked); tx 534 net/mac80211/mesh_ps.c bool tx, bool acked) tx 539 net/mac80211/mesh_ps.c if (tx) { tx 2223 net/mac80211/mlme.c u16 stype, u16 reason, bool tx, tx 2232 net/mac80211/mlme.c if (WARN_ON_ONCE(tx && !frame_buf)) tx 2266 net/mac80211/mlme.c if (tx) tx 2270 net/mac80211/mlme.c if (tx || frame_buf) { tx 2283 net/mac80211/mlme.c tx, frame_buf); tx 2287 net/mac80211/mlme.c if (tx) tx 2653 net/mac80211/mlme.c const u8 *buf, size_t len, bool tx, tx 2658 net/mac80211/mlme.c .u.mlme.data = tx ? DEAUTH_TX_EVENT : DEAUTH_RX_EVENT, tx 2662 net/mac80211/mlme.c if (tx) tx 2675 net/mac80211/mlme.c bool tx; tx 2683 net/mac80211/mlme.c tx = !sdata->csa_block_tx; tx 2692 net/mac80211/mlme.c tx, frame_buf); tx 2703 net/mac80211/mlme.c ieee80211_report_disconnect(sdata, frame_buf, sizeof(frame_buf), tx, tx 4158 net/mac80211/mlme.c u8 *bssid, u8 reason, bool tx) tx 4163 net/mac80211/mlme.c tx, frame_buf); tx 5541 net/mac80211/mlme.c bool tx = !req->local_state_change; tx 5553 net/mac80211/mlme.c req->reason_code, tx, tx 5573 net/mac80211/mlme.c req->reason_code, tx, tx 5590 net/mac80211/mlme.c req->reason_code, tx, frame_buf); tx 155 net/mac80211/tkip.c struct tkip_ctx *ctx = &key->u.tkip.tx; tx 176 net/mac80211/tkip.c struct tkip_ctx *ctx = &key->u.tkip.tx; tx 202 net/mac80211/tkip.c struct tkip_ctx *ctx = &key->u.tkip.tx; tx 1252 net/mac80211/trace.h TP_PROTO(struct ieee80211_local *local, u32 tx, u32 rx), tx 1254 net/mac80211/trace.h TP_ARGS(local, tx, rx), tx 1258 net/mac80211/trace.h __field(u32, tx) tx 1264 net/mac80211/trace.h __entry->tx = tx; tx 1270 net/mac80211/trace.h LOCAL_PR_ARG, __entry->tx, __entry->rx tx 1275 net/mac80211/trace.h TP_PROTO(struct ieee80211_local *local, u32 *tx, u32 *tx_max, tx 1278 net/mac80211/trace.h TP_ARGS(local, tx, tx_max, rx, rx_max), tx 1282 net/mac80211/trace.h __field(u32, tx) tx 1290 net/mac80211/trace.h __entry->tx = *tx; tx 1299 net/mac80211/trace.h __entry->tx, __entry->tx_max, __entry->rx, __entry->rx_max tx 51 net/mac80211/tx.c static __le16 ieee80211_duration(struct ieee80211_tx_data *tx, tx 57 net/mac80211/tx.c struct ieee80211_local *local = tx->local; tx 65 net/mac80211/tx.c if (tx->rate.flags & (IEEE80211_TX_RC_MCS | IEEE80211_TX_RC_VHT_MCS)) tx 69 net/mac80211/tx.c chanctx_conf = rcu_dereference(tx->sdata->vif.chanctx_conf); tx 77 net/mac80211/tx.c if (WARN_ON_ONCE(tx->rate.idx < 0)) tx 81 net/mac80211/tx.c txrate = &sband->bitrates[tx->rate.idx]; tx 150 net/mac80211/tx.c if (tx->sdata->vif.bss_conf.basic_rates & BIT(i)) tx 156 net/mac80211/tx.c if (tx->sdata->flags & IEEE80211_SDATA_OPERATING_GMODE) tx 191 net/mac80211/tx.c tx->sdata->vif.bss_conf.use_short_preamble, tx 201 net/mac80211/tx.c tx->sdata->vif.bss_conf.use_short_preamble, tx 210 net/mac80211/tx.c ieee80211_tx_h_dynamic_ps(struct ieee80211_tx_data *tx) tx 212 net/mac80211/tx.c struct ieee80211_local *local = tx->local; tx 214 net/mac80211/tx.c struct ieee80211_tx_info *info = IEEE80211_SKB_CB(tx->skb); tx 240 net/mac80211/tx.c if (tx->sdata->vif.type != NL80211_IFTYPE_STATION) tx 246 net/mac80211/tx.c ifmgd = &tx->sdata->u.mgd; tx 263 net/mac80211/tx.c skb_get_queue_mapping(tx->skb) == IEEE80211_AC_VO) tx 287 net/mac80211/tx.c ieee80211_tx_h_check_assoc(struct ieee80211_tx_data *tx) tx 290 net/mac80211/tx.c struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)tx->skb->data; tx 291 net/mac80211/tx.c struct ieee80211_tx_info *info = IEEE80211_SKB_CB(tx->skb); tx 297 net/mac80211/tx.c if (unlikely(test_bit(SCAN_SW_SCANNING, &tx->local->scanning)) && tx 298 net/mac80211/tx.c test_bit(SDATA_STATE_OFFCHANNEL, &tx->sdata->state) && tx 314 net/mac80211/tx.c if (tx->sdata->vif.type == NL80211_IFTYPE_OCB) tx 317 net/mac80211/tx.c if (tx->sdata->vif.type == NL80211_IFTYPE_WDS) tx 320 net/mac80211/tx.c if (tx->flags & IEEE80211_TX_PS_BUFFERED) tx 323 net/mac80211/tx.c if (tx->sta) tx 324 net/mac80211/tx.c assoc = test_sta_flag(tx->sta, WLAN_STA_ASSOC); tx 326 net/mac80211/tx.c if (likely(tx->flags & IEEE80211_TX_UNICAST)) { tx 330 net/mac80211/tx.c sdata_info(tx->sdata, tx 334 net/mac80211/tx.c I802_DEBUG_INC(tx->local->tx_handlers_drop_not_assoc); tx 338 net/mac80211/tx.c ieee80211_vif_get_num_mcast_if(tx->sdata) == 0)) { tx 401 net/mac80211/tx.c ieee80211_tx_h_multicast_ps_buf(struct ieee80211_tx_data *tx) tx 403 net/mac80211/tx.c struct ieee80211_tx_info *info = IEEE80211_SKB_CB(tx->skb); tx 404 net/mac80211/tx.c struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)tx->skb->data; tx 416 net/mac80211/tx.c if (tx->sdata->vif.type == NL80211_IFTYPE_AP || tx 417 net/mac80211/tx.c tx->sdata->vif.type == NL80211_IFTYPE_AP_VLAN) { tx 418 net/mac80211/tx.c if (!tx->sdata->bss) tx 421 net/mac80211/tx.c ps = &tx->sdata->bss->ps; tx 422 net/mac80211/tx.c } else if (ieee80211_vif_is_mesh(&tx->sdata->vif)) { tx 423 net/mac80211/tx.c ps = &tx->sdata->u.mesh.ps; tx 436 net/mac80211/tx.c if (ieee80211_hw_check(&tx->local->hw, QUEUE_CONTROL)) tx 437 net/mac80211/tx.c info->hw_queue = tx->sdata->vif.cab_queue; tx 446 net/mac80211/tx.c if (!ieee80211_hw_check(&tx->local->hw, HOST_BROADCAST_PS_BUFFERING)) tx 450 net/mac80211/tx.c if (tx->local->total_ps_buffered >= TOTAL_MAX_TX_BUFFER) tx 451 net/mac80211/tx.c purge_old_ps_buffers(tx->local); tx 454 net/mac80211/tx.c ps_dbg(tx->sdata, tx 456 net/mac80211/tx.c ieee80211_free_txskb(&tx->local->hw, skb_dequeue(&ps->bc_buf)); tx 458 net/mac80211/tx.c tx->local->total_ps_buffered++; tx 460 net/mac80211/tx.c skb_queue_tail(&ps->bc_buf, tx->skb); tx 481 net/mac80211/tx.c ieee80211_tx_h_unicast_ps_buf(struct ieee80211_tx_data *tx) tx 483 net/mac80211/tx.c struct sta_info *sta = tx->sta; tx 484 net/mac80211/tx.c struct ieee80211_tx_info *info = IEEE80211_SKB_CB(tx->skb); tx 485 net/mac80211/tx.c struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)tx->skb->data; tx 486 net/mac80211/tx.c struct ieee80211_local *local = tx->local; tx 495 net/mac80211/tx.c int ac = skb_get_queue_mapping(tx->skb); tx 505 net/mac80211/tx.c if (tx->local->total_ps_buffered >= TOTAL_MAX_TX_BUFFER) tx 506 net/mac80211/tx.c purge_old_ps_buffers(tx->local); tx 524 net/mac80211/tx.c ps_dbg(tx->sdata, tx 529 net/mac80211/tx.c tx->local->total_ps_buffered++; tx 532 net/mac80211/tx.c info->control.vif = &tx->sdata->vif; tx 535 net/mac80211/tx.c skb_queue_tail(&sta->ps_tx_buf[ac], tx->skb); tx 551 net/mac80211/tx.c ps_dbg(tx->sdata, tx 560 net/mac80211/tx.c ieee80211_tx_h_ps_buf(struct ieee80211_tx_data *tx) tx 562 net/mac80211/tx.c if (unlikely(tx->flags & IEEE80211_TX_PS_BUFFERED)) tx 565 net/mac80211/tx.c if (tx->flags & IEEE80211_TX_UNICAST) tx 566 net/mac80211/tx.c return ieee80211_tx_h_unicast_ps_buf(tx); tx 568 net/mac80211/tx.c return ieee80211_tx_h_multicast_ps_buf(tx); tx 572 net/mac80211/tx.c ieee80211_tx_h_check_control_port_protocol(struct ieee80211_tx_data *tx) tx 574 net/mac80211/tx.c struct ieee80211_tx_info *info = IEEE80211_SKB_CB(tx->skb); tx 576 net/mac80211/tx.c if (unlikely(tx->sdata->control_port_protocol == tx->skb->protocol)) { tx 577 net/mac80211/tx.c if (tx->sdata->control_port_no_encrypt) tx 587 net/mac80211/tx.c ieee80211_tx_h_select_key(struct ieee80211_tx_data *tx) tx 590 net/mac80211/tx.c struct ieee80211_tx_info *info = IEEE80211_SKB_CB(tx->skb); tx 591 net/mac80211/tx.c struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)tx->skb->data; tx 594 net/mac80211/tx.c tx->key = NULL; tx 598 net/mac80211/tx.c if (tx->sta && tx 599 net/mac80211/tx.c (key = rcu_dereference(tx->sta->ptk[tx->sta->ptk_idx]))) tx 600 net/mac80211/tx.c tx->key = key; tx 601 net/mac80211/tx.c else if (ieee80211_is_group_privacy_action(tx->skb) && tx 602 net/mac80211/tx.c (key = rcu_dereference(tx->sdata->default_multicast_key))) tx 603 net/mac80211/tx.c tx->key = key; tx 606 net/mac80211/tx.c ieee80211_is_robust_mgmt_frame(tx->skb) && tx 607 net/mac80211/tx.c (key = rcu_dereference(tx->sdata->default_mgmt_key))) tx 608 net/mac80211/tx.c tx->key = key; tx 610 net/mac80211/tx.c (key = rcu_dereference(tx->sdata->default_multicast_key))) tx 611 net/mac80211/tx.c tx->key = key; tx 613 net/mac80211/tx.c (key = rcu_dereference(tx->sdata->default_unicast_key))) tx 614 net/mac80211/tx.c tx->key = key; tx 616 net/mac80211/tx.c tx->key = NULL; tx 618 net/mac80211/tx.c if (tx->key) { tx 623 net/mac80211/tx.c switch (tx->key->conf.cipher) { tx 628 net/mac80211/tx.c tx->key = NULL; tx 635 net/mac80211/tx.c !ieee80211_use_mfp(hdr->frame_control, tx->sta, tx 636 net/mac80211/tx.c tx->skb) && tx 637 net/mac80211/tx.c !ieee80211_is_group_privacy_action(tx->skb)) tx 638 net/mac80211/tx.c tx->key = NULL; tx 640 net/mac80211/tx.c skip_hw = (tx->key->conf.flags & tx 649 net/mac80211/tx.c tx->key = NULL; tx 653 net/mac80211/tx.c if (unlikely(tx->key && tx->key->flags & KEY_FLAG_TAINTED && tx 657 net/mac80211/tx.c if (!skip_hw && tx->key && tx 658 net/mac80211/tx.c tx->key->flags & KEY_FLAG_UPLOADED_TO_HARDWARE) tx 659 net/mac80211/tx.c info->control.hw_key = &tx->key->conf; tx 660 net/mac80211/tx.c } else if (!ieee80211_is_mgmt(hdr->frame_control) && tx->sta && tx 661 net/mac80211/tx.c test_sta_flag(tx->sta, WLAN_STA_USES_ENCRYPTION)) { tx 669 net/mac80211/tx.c ieee80211_tx_h_rate_ctrl(struct ieee80211_tx_data *tx) tx 671 net/mac80211/tx.c struct ieee80211_tx_info *info = IEEE80211_SKB_CB(tx->skb); tx 672 net/mac80211/tx.c struct ieee80211_hdr *hdr = (void *)tx->skb->data; tx 681 net/mac80211/tx.c sband = tx->local->hw.wiphy->bands[info->band]; tx 683 net/mac80211/tx.c len = min_t(u32, tx->skb->len + FCS_LEN, tx 684 net/mac80211/tx.c tx->local->hw.wiphy->frag_threshold); tx 687 net/mac80211/tx.c txrc.hw = &tx->local->hw; tx 689 net/mac80211/tx.c txrc.bss_conf = &tx->sdata->vif.bss_conf; tx 690 net/mac80211/tx.c txrc.skb = tx->skb; tx 692 net/mac80211/tx.c txrc.rate_idx_mask = tx->sdata->rc_rateidx_mask[info->band]; tx 694 net/mac80211/tx.c if (tx->sdata->rc_has_mcs_mask[info->band]) tx 696 net/mac80211/tx.c tx->sdata->rc_rateidx_mcs_mask[info->band]; tx 698 net/mac80211/tx.c txrc.bss = (tx->sdata->vif.type == NL80211_IFTYPE_AP || tx 699 net/mac80211/tx.c tx->sdata->vif.type == NL80211_IFTYPE_MESH_POINT || tx 700 net/mac80211/tx.c tx->sdata->vif.type == NL80211_IFTYPE_ADHOC || tx 701 net/mac80211/tx.c tx->sdata->vif.type == NL80211_IFTYPE_OCB); tx 704 net/mac80211/tx.c if (len > tx->local->hw.wiphy->rts_threshold) { tx 709 net/mac80211/tx.c info->control.use_cts_prot = tx->sdata->vif.bss_conf.use_cts_prot; tx 717 net/mac80211/tx.c if (tx->sdata->vif.bss_conf.use_short_preamble && tx 719 net/mac80211/tx.c (tx->sta && test_sta_flag(tx->sta, WLAN_STA_SHORT_PREAMBLE)))) tx 728 net/mac80211/tx.c if (tx->sta) tx 729 net/mac80211/tx.c assoc = test_sta_flag(tx->sta, WLAN_STA_ASSOC); tx 735 net/mac80211/tx.c if (WARN(test_bit(SCAN_SW_SCANNING, &tx->local->scanning) && assoc && tx 736 net/mac80211/tx.c !rate_usable_index_exists(sband, &tx->sta->sta), tx 740 net/mac80211/tx.c tx->sdata->name, hdr->addr1, tx 748 net/mac80211/tx.c rate_control_get_rate(tx->sdata, tx->sta, &txrc); tx 750 net/mac80211/tx.c if (tx->sta && !info->control.skip_table) tx 751 net/mac80211/tx.c ratetbl = rcu_dereference(tx->sta->sta.rates); tx 764 net/mac80211/tx.c tx->rate = rate; tx 769 net/mac80211/tx.c tx->rate = info->control.rates[0]; tx 773 net/mac80211/tx.c txrc.reported_rate = tx->rate; tx 774 net/mac80211/tx.c if (tx->sta && ieee80211_is_data(hdr->frame_control)) tx 775 net/mac80211/tx.c tx->sta->tx_stats.last_rate = txrc.reported_rate; tx 776 net/mac80211/tx.c } else if (tx->sta) tx 777 net/mac80211/tx.c tx->sta->tx_stats.last_rate = txrc.reported_rate; tx 804 net/mac80211/tx.c ieee80211_tx_h_sequence(struct ieee80211_tx_data *tx) tx 806 net/mac80211/tx.c struct ieee80211_tx_info *info = IEEE80211_SKB_CB(tx->skb); tx 807 net/mac80211/tx.c struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)tx->skb->data; tx 835 net/mac80211/tx.c if (tx->flags & IEEE80211_TX_NO_SEQNO) tx 840 net/mac80211/tx.c hdr->seq_ctrl = cpu_to_le16(tx->sdata->sequence_number); tx 841 net/mac80211/tx.c tx->sdata->sequence_number += 0x10; tx 842 net/mac80211/tx.c if (tx->sta) tx 843 net/mac80211/tx.c tx->sta->tx_stats.msdu[IEEE80211_NUM_TIDS]++; tx 852 net/mac80211/tx.c if (!tx->sta) tx 857 net/mac80211/tx.c tx->sta->tx_stats.msdu[tid]++; tx 859 net/mac80211/tx.c hdr->seq_ctrl = ieee80211_tx_next_seq(tx->sta, tid); tx 864 net/mac80211/tx.c static int ieee80211_fragment(struct ieee80211_tx_data *tx, tx 868 net/mac80211/tx.c struct ieee80211_local *local = tx->local; tx 888 net/mac80211/tx.c tx->sdata->encrypt_headroom + tx 893 net/mac80211/tx.c __skb_queue_tail(&tx->skbs, tmp); tx 896 net/mac80211/tx.c local->tx_headroom + tx->sdata->encrypt_headroom); tx 925 net/mac80211/tx.c ieee80211_tx_h_fragment(struct ieee80211_tx_data *tx) tx 927 net/mac80211/tx.c struct sk_buff *skb = tx->skb; tx 930 net/mac80211/tx.c int frag_threshold = tx->local->hw.wiphy->frag_threshold; tx 935 net/mac80211/tx.c __skb_queue_tail(&tx->skbs, skb); tx 936 net/mac80211/tx.c tx->skb = NULL; tx 941 net/mac80211/tx.c if (ieee80211_hw_check(&tx->local->hw, SUPPORTS_TX_FRAG)) tx 966 net/mac80211/tx.c if (ieee80211_fragment(tx, skb, hdrlen, frag_threshold)) tx 972 net/mac80211/tx.c skb_queue_walk(&tx->skbs, skb) { tx 978 net/mac80211/tx.c if (!skb_queue_is_last(&tx->skbs, skb)) { tx 1000 net/mac80211/tx.c ieee80211_tx_h_stats(struct ieee80211_tx_data *tx) tx 1005 net/mac80211/tx.c if (!tx->sta) tx 1008 net/mac80211/tx.c skb_queue_walk(&tx->skbs, skb) { tx 1010 net/mac80211/tx.c tx->sta->tx_stats.bytes[ac] += skb->len; tx 1013 net/mac80211/tx.c tx->sta->tx_stats.packets[ac]++; tx 1019 net/mac80211/tx.c ieee80211_tx_h_encrypt(struct ieee80211_tx_data *tx) tx 1021 net/mac80211/tx.c if (!tx->key) tx 1024 net/mac80211/tx.c switch (tx->key->conf.cipher) { tx 1027 net/mac80211/tx.c return ieee80211_crypto_wep_encrypt(tx); tx 1029 net/mac80211/tx.c return ieee80211_crypto_tkip_encrypt(tx); tx 1032 net/mac80211/tx.c tx, IEEE80211_CCMP_MIC_LEN); tx 1035 net/mac80211/tx.c tx, IEEE80211_CCMP_256_MIC_LEN); tx 1037 net/mac80211/tx.c return ieee80211_crypto_aes_cmac_encrypt(tx); tx 1039 net/mac80211/tx.c return ieee80211_crypto_aes_cmac_256_encrypt(tx); tx 1042 net/mac80211/tx.c return ieee80211_crypto_aes_gmac_encrypt(tx); tx 1045 net/mac80211/tx.c return ieee80211_crypto_gcmp_encrypt(tx); tx 1047 net/mac80211/tx.c return ieee80211_crypto_hw_encrypt(tx); tx 1054 net/mac80211/tx.c ieee80211_tx_h_calculate_duration(struct ieee80211_tx_data *tx) tx 1061 net/mac80211/tx.c skb_queue_walk(&tx->skbs, skb) { tx 1065 net/mac80211/tx.c if (!skb_queue_is_last(&tx->skbs, skb)) { tx 1066 net/mac80211/tx.c struct sk_buff *next = skb_queue_next(&tx->skbs, skb); tx 1073 net/mac80211/tx.c ieee80211_duration(tx, skb, group_addr, next_len); tx 1081 net/mac80211/tx.c static bool ieee80211_tx_prep_agg(struct ieee80211_tx_data *tx, tx 1099 net/mac80211/tx.c } else if (!tx->sta->sta.txq[tid]) { tx 1100 net/mac80211/tx.c spin_lock(&tx->sta->lock); tx 1119 net/mac80211/tx.c tid_tx = rcu_dereference_protected_tid_tx(tx->sta, tid); tx 1129 net/mac80211/tx.c clear_sta_flag(tx->sta, WLAN_STA_SP); tx 1130 net/mac80211/tx.c ps_dbg(tx->sta->sdata, tx 1132 net/mac80211/tx.c tx->sta->sta.addr, tx->sta->sta.aid); tx 1134 net/mac80211/tx.c info->control.vif = &tx->sdata->vif; tx 1141 net/mac80211/tx.c spin_unlock(&tx->sta->lock); tx 1144 net/mac80211/tx.c ieee80211_free_txskb(&tx->local->hw, purge_skb); tx 1161 net/mac80211/tx.c struct ieee80211_tx_data *tx, tx 1169 net/mac80211/tx.c memset(tx, 0, sizeof(*tx)); tx 1170 net/mac80211/tx.c tx->skb = skb; tx 1171 net/mac80211/tx.c tx->local = local; tx 1172 net/mac80211/tx.c tx->sdata = sdata; tx 1173 net/mac80211/tx.c __skb_queue_head_init(&tx->skbs); tx 1186 net/mac80211/tx.c tx->sta = sta; tx 1189 net/mac80211/tx.c tx->sta = rcu_dereference(sdata->u.vlan.sta); tx 1190 net/mac80211/tx.c if (!tx->sta && sdata->wdev.use_4addr) tx 1194 net/mac80211/tx.c tx->sdata->control_port_protocol == tx->skb->protocol) { tx 1195 net/mac80211/tx.c tx->sta = sta_info_get_bss(sdata, hdr->addr1); tx 1197 net/mac80211/tx.c if (!tx->sta && !is_multicast_ether_addr(hdr->addr1)) tx 1198 net/mac80211/tx.c tx->sta = sta_info_get(sdata, hdr->addr1); tx 1201 net/mac80211/tx.c if (tx->sta && ieee80211_is_data_qos(hdr->frame_control) && tx 1209 net/mac80211/tx.c tid_tx = rcu_dereference(tx->sta->ampdu_mlme.tid_tx[tid]); tx 1213 net/mac80211/tx.c queued = ieee80211_tx_prep_agg(tx, skb, info, tx 1222 net/mac80211/tx.c tx->flags &= ~IEEE80211_TX_UNICAST; tx 1225 net/mac80211/tx.c tx->flags |= IEEE80211_TX_UNICAST; tx 1228 net/mac80211/tx.c if (!(tx->flags & IEEE80211_TX_UNICAST) || tx 1234 net/mac80211/tx.c if (!tx->sta) tx 1236 net/mac80211/tx.c else if (test_and_clear_sta_flag(tx->sta, WLAN_STA_CLEAR_PS_FILT)) { tx 1238 net/mac80211/tx.c ieee80211_check_fast_xmit(tx->sta); tx 1771 net/mac80211/tx.c static int invoke_tx_handlers_early(struct ieee80211_tx_data *tx) tx 1777 net/mac80211/tx.c res = txh(tx); \ tx 1787 net/mac80211/tx.c if (!ieee80211_hw_check(&tx->local->hw, HAS_RATE_CONTROL)) tx 1792 net/mac80211/tx.c I802_DEBUG_INC(tx->local->tx_handlers_drop); tx 1793 net/mac80211/tx.c if (tx->skb) tx 1794 net/mac80211/tx.c ieee80211_free_txskb(&tx->local->hw, tx->skb); tx 1796 net/mac80211/tx.c ieee80211_purge_tx_queue(&tx->local->hw, &tx->skbs); tx 1799 net/mac80211/tx.c I802_DEBUG_INC(tx->local->tx_handlers_queued); tx 1810 net/mac80211/tx.c static int invoke_tx_handlers_late(struct ieee80211_tx_data *tx) tx 1812 net/mac80211/tx.c struct ieee80211_tx_info *info = IEEE80211_SKB_CB(tx->skb); tx 1816 net/mac80211/tx.c __skb_queue_tail(&tx->skbs, tx->skb); tx 1817 net/mac80211/tx.c tx->skb = NULL; tx 1827 net/mac80211/tx.c if (!ieee80211_hw_check(&tx->local->hw, HAS_RATE_CONTROL)) tx 1833 net/mac80211/tx.c I802_DEBUG_INC(tx->local->tx_handlers_drop); tx 1834 net/mac80211/tx.c if (tx->skb) tx 1835 net/mac80211/tx.c ieee80211_free_txskb(&tx->local->hw, tx->skb); tx 1837 net/mac80211/tx.c ieee80211_purge_tx_queue(&tx->local->hw, &tx->skbs); tx 1840 net/mac80211/tx.c I802_DEBUG_INC(tx->local->tx_handlers_queued); tx 1847 net/mac80211/tx.c static int invoke_tx_handlers(struct ieee80211_tx_data *tx) tx 1849 net/mac80211/tx.c int r = invoke_tx_handlers_early(tx); tx 1853 net/mac80211/tx.c return invoke_tx_handlers_late(tx); tx 1862 net/mac80211/tx.c struct ieee80211_tx_data tx; tx 1865 net/mac80211/tx.c if (ieee80211_tx_prepare(sdata, &tx, NULL, skb) == TX_DROP) tx 1872 net/mac80211/tx.c if (invoke_tx_handlers(&tx)) tx 1876 net/mac80211/tx.c if (tx.sta) tx 1877 net/mac80211/tx.c *sta = &tx.sta->sta; tx 1883 net/mac80211/tx.c skb2 = __skb_dequeue(&tx.skbs); tx 1884 net/mac80211/tx.c if (WARN_ON(skb2 != skb || !skb_queue_empty(&tx.skbs))) { tx 1886 net/mac80211/tx.c ieee80211_purge_tx_queue(hw, &tx.skbs); tx 1902 net/mac80211/tx.c struct ieee80211_tx_data tx; tx 1915 net/mac80211/tx.c res_prepare = ieee80211_tx_prepare(sdata, &tx, sta, skb); tx 1917 net/mac80211/tx.c tx.flags |= txdata_flags; tx 1932 net/mac80211/tx.c if (invoke_tx_handlers_early(&tx)) tx 1935 net/mac80211/tx.c if (ieee80211_queue_skb(local, sdata, tx.sta, tx.skb)) tx 1938 net/mac80211/tx.c if (!invoke_tx_handlers_late(&tx)) tx 1939 net/mac80211/tx.c result = __ieee80211_tx(local, &tx.skbs, led_len, tx 1940 net/mac80211/tx.c tx.sta, txpending); tx 3434 net/mac80211/tx.c struct ieee80211_tx_data tx; tx 3515 net/mac80211/tx.c __skb_queue_head_init(&tx.skbs); tx 3517 net/mac80211/tx.c tx.flags = IEEE80211_TX_UNICAST; tx 3518 net/mac80211/tx.c tx.local = local; tx 3519 net/mac80211/tx.c tx.sdata = sdata; tx 3520 net/mac80211/tx.c tx.sta = sta; tx 3521 net/mac80211/tx.c tx.key = fast_tx->key; tx 3524 net/mac80211/tx.c tx.skb = skb; tx 3525 net/mac80211/tx.c r = ieee80211_tx_h_rate_ctrl(&tx); tx 3526 net/mac80211/tx.c skb = tx.skb; tx 3527 net/mac80211/tx.c tx.skb = NULL; tx 3546 net/mac80211/tx.c __skb_queue_tail(&tx.skbs, skb); tx 3547 net/mac80211/tx.c ieee80211_tx_frags(local, &sdata->vif, &sta->sta, &tx.skbs, false); tx 3561 net/mac80211/tx.c struct ieee80211_tx_data tx; tx 3593 net/mac80211/tx.c memset(&tx, 0, sizeof(tx)); tx 3594 net/mac80211/tx.c __skb_queue_head_init(&tx.skbs); tx 3595 net/mac80211/tx.c tx.local = local; tx 3596 net/mac80211/tx.c tx.skb = skb; tx 3597 net/mac80211/tx.c tx.sdata = vif_to_sdata(info->control.vif); tx 3600 net/mac80211/tx.c tx.sta = container_of(txq->sta, struct sta_info, sta); tx 3606 net/mac80211/tx.c !ieee80211_vif_is_mesh(&tx.sdata->vif) && tx 3607 net/mac80211/tx.c tx.sdata->vif.type != NL80211_IFTYPE_OCB && tx 3609 net/mac80211/tx.c !test_sta_flag(tx.sta, WLAN_STA_AUTHORIZED) && tx 3612 net/mac80211/tx.c !ether_addr_equal(tx.sdata->vif.addr, tx 3624 net/mac80211/tx.c r = ieee80211_tx_h_select_key(&tx); tx 3640 net/mac80211/tx.c if (tx.key && tx 3641 net/mac80211/tx.c (tx.key->conf.flags & IEEE80211_KEY_FLAG_GENERATE_IV)) tx 3645 net/mac80211/tx.c tx.key, skb); tx 3647 net/mac80211/tx.c if (invoke_tx_handlers_late(&tx)) tx 3650 net/mac80211/tx.c skb = __skb_dequeue(&tx.skbs); tx 3652 net/mac80211/tx.c if (!skb_queue_empty(&tx.skbs)) { tx 3654 net/mac80211/tx.c skb_queue_splice_tail(&tx.skbs, &txqi->frags); tx 3667 net/mac80211/tx.c switch (tx.sdata->vif.type) { tx 3669 net/mac80211/tx.c if (tx.sdata->u.mntr.flags & MONITOR_FLAG_ACTIVE) { tx 3670 net/mac80211/tx.c vif = &tx.sdata->vif; tx 3673 net/mac80211/tx.c tx.sdata = rcu_dereference(local->monitor_sdata); tx 3674 net/mac80211/tx.c if (tx.sdata) { tx 3675 net/mac80211/tx.c vif = &tx.sdata->vif; tx 3686 net/mac80211/tx.c tx.sdata = container_of(tx.sdata->bss, tx 3690 net/mac80211/tx.c vif = &tx.sdata->vif; tx 4070 net/mac80211/tx.c struct ieee80211_tx_data tx = { tx 4089 net/mac80211/tx.c tx.sta = sta_info_get(sdata, hdr->addr1); tx 4090 net/mac80211/tx.c tx.skb = skb; tx 4092 net/mac80211/tx.c if (ieee80211_tx_h_select_key(&tx) != TX_CONTINUE) { tx 4898 net/mac80211/tx.c struct ieee80211_tx_data tx; tx 4947 net/mac80211/tx.c if (!ieee80211_tx_prepare(sdata, &tx, NULL, skb)) tx 4954 net/mac80211/tx.c tx.flags |= IEEE80211_TX_PS_BUFFERED; tx 4957 net/mac80211/tx.c if (invoke_tx_handlers(&tx)) tx 49 net/mac80211/util.c void ieee80211_tx_set_protected(struct ieee80211_tx_data *tx) tx 54 net/mac80211/util.c skb_queue_walk(&tx->skbs, skb) { tx 272 net/mac80211/wep.c static int wep_encrypt_skb(struct ieee80211_tx_data *tx, struct sk_buff *skb) tx 278 net/mac80211/wep.c if (ieee80211_wep_encrypt(tx->local, skb, tx->key->conf.key, tx 279 net/mac80211/wep.c tx->key->conf.keylen, tx 280 net/mac80211/wep.c tx->key->conf.keyidx)) tx 284 net/mac80211/wep.c if (!ieee80211_wep_add_iv(tx->local, skb, tx 285 net/mac80211/wep.c tx->key->conf.keylen, tx 286 net/mac80211/wep.c tx->key->conf.keyidx)) tx 294 net/mac80211/wep.c ieee80211_crypto_wep_encrypt(struct ieee80211_tx_data *tx) tx 298 net/mac80211/wep.c ieee80211_tx_set_protected(tx); tx 300 net/mac80211/wep.c skb_queue_walk(&tx->skbs, skb) { tx 301 net/mac80211/wep.c if (wep_encrypt_skb(tx, skb) < 0) { tx 302 net/mac80211/wep.c I802_DEBUG_INC(tx->local->tx_handlers_drop_wep); tx 28 net/mac80211/wep.h ieee80211_crypto_wep_encrypt(struct ieee80211_tx_data *tx); tx 29 net/mac80211/wpa.c ieee80211_tx_h_michael_mic_add(struct ieee80211_tx_data *tx) tx 35 net/mac80211/wpa.c struct sk_buff *skb = tx->skb; tx 40 net/mac80211/wpa.c if (!tx->key || tx->key->conf.cipher != WLAN_CIPHER_SUITE_TKIP || tx 58 net/mac80211/wpa.c ieee80211_hw_check(&tx->local->hw, SUPPORTS_TX_FRAG)) && tx 59 net/mac80211/wpa.c !(tx->key->conf.flags & (IEEE80211_KEY_FLAG_GENERATE_MMIC | tx 78 net/mac80211/wpa.c if (tx->key->conf.flags & IEEE80211_KEY_FLAG_PUT_MIC_SPACE) { tx 84 net/mac80211/wpa.c key = &tx->key->conf.key[NL80211_TKIP_DATA_OFFSET_TX_MIC_KEY]; tx 193 net/mac80211/wpa.c static int tkip_encrypt_skb(struct ieee80211_tx_data *tx, struct sk_buff *skb) tx 196 net/mac80211/wpa.c struct ieee80211_key *key = tx->key; tx 242 net/mac80211/wpa.c return ieee80211_tkip_encrypt_data(&tx->local->wep_tx_ctx, tx 248 net/mac80211/wpa.c ieee80211_crypto_tkip_encrypt(struct ieee80211_tx_data *tx) tx 252 net/mac80211/wpa.c ieee80211_tx_set_protected(tx); tx 254 net/mac80211/wpa.c skb_queue_walk(&tx->skbs, skb) { tx 255 net/mac80211/wpa.c if (tkip_encrypt_skb(tx, skb) < 0) tx 405 net/mac80211/wpa.c static int ccmp_encrypt_skb(struct ieee80211_tx_data *tx, struct sk_buff *skb, tx 409 net/mac80211/wpa.c struct ieee80211_key *key = tx->key; tx 477 net/mac80211/wpa.c ieee80211_crypto_ccmp_encrypt(struct ieee80211_tx_data *tx, tx 482 net/mac80211/wpa.c ieee80211_tx_set_protected(tx); tx 484 net/mac80211/wpa.c skb_queue_walk(&tx->skbs, skb) { tx 485 net/mac80211/wpa.c if (ccmp_encrypt_skb(tx, skb, mic_len) < 0) tx 637 net/mac80211/wpa.c static int gcmp_encrypt_skb(struct ieee80211_tx_data *tx, struct sk_buff *skb) tx 640 net/mac80211/wpa.c struct ieee80211_key *key = tx->key; tx 708 net/mac80211/wpa.c ieee80211_crypto_gcmp_encrypt(struct ieee80211_tx_data *tx) tx 712 net/mac80211/wpa.c ieee80211_tx_set_protected(tx); tx 714 net/mac80211/wpa.c skb_queue_walk(&tx->skbs, skb) { tx 715 net/mac80211/wpa.c if (gcmp_encrypt_skb(tx, skb) < 0) tx 796 net/mac80211/wpa.c ieee80211_crypto_cs_encrypt(struct ieee80211_tx_data *tx, tx 800 net/mac80211/wpa.c struct ieee80211_key *key = tx->key; tx 933 net/mac80211/wpa.c ieee80211_crypto_aes_cmac_encrypt(struct ieee80211_tx_data *tx) tx 937 net/mac80211/wpa.c struct ieee80211_key *key = tx->key; tx 942 net/mac80211/wpa.c if (WARN_ON(skb_queue_len(&tx->skbs) != 1)) tx 945 net/mac80211/wpa.c skb = skb_peek(&tx->skbs); tx 981 net/mac80211/wpa.c ieee80211_crypto_aes_cmac_256_encrypt(struct ieee80211_tx_data *tx) tx 985 net/mac80211/wpa.c struct ieee80211_key *key = tx->key; tx 990 net/mac80211/wpa.c if (WARN_ON(skb_queue_len(&tx->skbs) != 1)) tx 993 net/mac80211/wpa.c skb = skb_peek(&tx->skbs); tx 1124 net/mac80211/wpa.c ieee80211_crypto_aes_gmac_encrypt(struct ieee80211_tx_data *tx) tx 1128 net/mac80211/wpa.c struct ieee80211_key *key = tx->key; tx 1135 net/mac80211/wpa.c if (WARN_ON(skb_queue_len(&tx->skbs) != 1)) tx 1138 net/mac80211/wpa.c skb = skb_peek(&tx->skbs); tx 1233 net/mac80211/wpa.c ieee80211_crypto_hw_encrypt(struct ieee80211_tx_data *tx) tx 1239 net/mac80211/wpa.c skb_queue_walk(&tx->skbs, skb) { tx 1246 net/mac80211/wpa.c if (tx->key->flags & KEY_FLAG_CIPHER_SCHEME) { tx 1247 net/mac80211/wpa.c res = ieee80211_crypto_cs_encrypt(tx, skb); tx 1253 net/mac80211/wpa.c ieee80211_tx_set_protected(tx); tx 14 net/mac80211/wpa.h ieee80211_tx_h_michael_mic_add(struct ieee80211_tx_data *tx); tx 19 net/mac80211/wpa.h ieee80211_crypto_tkip_encrypt(struct ieee80211_tx_data *tx); tx 24 net/mac80211/wpa.h ieee80211_crypto_ccmp_encrypt(struct ieee80211_tx_data *tx, tx 31 net/mac80211/wpa.h ieee80211_crypto_aes_cmac_encrypt(struct ieee80211_tx_data *tx); tx 33 net/mac80211/wpa.h ieee80211_crypto_aes_cmac_256_encrypt(struct ieee80211_tx_data *tx); tx 39 net/mac80211/wpa.h ieee80211_crypto_aes_gmac_encrypt(struct ieee80211_tx_data *tx); tx 43 net/mac80211/wpa.h ieee80211_crypto_hw_encrypt(struct ieee80211_tx_data *tx); tx 48 net/mac80211/wpa.h ieee80211_crypto_gcmp_encrypt(struct ieee80211_tx_data *tx); tx 177 net/nfc/nci/spi.c struct spi_transfer tx, rx; tx 183 net/nfc/nci/spi.c memset(&tx, 0, sizeof(struct spi_transfer)); tx 186 net/nfc/nci/spi.c tx.tx_buf = req; tx 187 net/nfc/nci/spi.c tx.len = 2; tx 188 net/nfc/nci/spi.c tx.cs_change = 0; tx 189 net/nfc/nci/spi.c tx.speed_hz = nspi->xfer_speed_hz; tx 190 net/nfc/nci/spi.c spi_message_add_tail(&tx, &m); tx 57 net/tls/tls_device.c kfree(ctx->tx.rec_seq); tx 58 net/tls/tls_device.c kfree(ctx->tx.iv); tx 217 net/tls/tls_device.c rcd_sn = tls_ctx->tx.rec_seq; tx 273 net/tls/tls_device.c tls_advance_record_sn(sk, prot, &ctx->tx); tx 967 net/tls/tls_device.c ctx->tx.iv = kmalloc(iv_size + TLS_CIPHER_AES_GCM_128_SALT_SIZE, tx 969 net/tls/tls_device.c if (!ctx->tx.iv) { tx 974 net/tls/tls_device.c memcpy(ctx->tx.iv + TLS_CIPHER_AES_GCM_128_SALT_SIZE, iv, iv_size); tx 977 net/tls/tls_device.c ctx->tx.rec_seq = kmemdup(rec_seq, rec_seq_size, GFP_KERNEL); tx 978 net/tls/tls_device.c if (!ctx->tx.rec_seq) { tx 988 net/tls/tls_device.c memcpy(&rcd_sn, ctx->tx.rec_seq, sizeof(rcd_sn)); tx 1065 net/tls/tls_device.c kfree(ctx->tx.rec_seq); tx 1067 net/tls/tls_device.c kfree(ctx->tx.iv); tx 278 net/tls/tls_main.c kfree(ctx->tx.rec_seq); tx 279 net/tls/tls_main.c kfree(ctx->tx.iv); tx 376 net/tls/tls_main.c ctx->tx.iv + TLS_CIPHER_AES_GCM_128_SALT_SIZE, tx 378 net/tls/tls_main.c memcpy(crypto_info_aes_gcm_128->rec_seq, ctx->tx.rec_seq, tx 400 net/tls/tls_main.c ctx->tx.iv + TLS_CIPHER_AES_GCM_256_SALT_SIZE, tx 402 net/tls/tls_main.c memcpy(crypto_info_aes_gcm_256->rec_seq, ctx->tx.rec_seq, tx 448 net/tls/tls_main.c unsigned int optlen, int tx) tx 462 net/tls/tls_main.c if (tx) { tx 526 net/tls/tls_main.c if (tx) { tx 547 net/tls/tls_main.c if (tx) tx 552 net/tls/tls_main.c if (tx) { tx 502 net/tls/tls_sw.c memcpy(&rec->iv_data[iv_offset], tls_ctx->tx.iv, tx 505 net/tls/tls_sw.c xor_iv_with_seq(prot->version, rec->iv_data, tls_ctx->tx.rec_seq); tx 541 net/tls/tls_sw.c tls_advance_record_sn(sk, prot, &tls_ctx->tx); tx 748 net/tls/tls_sw.c tls_ctx->tx.rec_seq, prot->rec_seq_size, tx 2258 net/tls/tls_sw.c int tls_set_sw_offload(struct sock *sk, struct tls_context *ctx, int tx) tx 2282 net/tls/tls_sw.c if (tx) { tx 2308 net/tls/tls_sw.c if (tx) { tx 2312 net/tls/tls_sw.c cctx = &ctx->tx; tx 2471 net/tls/tls_sw.c if (tx) { tx 605 net/wireless/mlme.c if (!(wdev->wiphy->mgmt_stypes[wdev->iftype].tx & BIT(stype >> 4))) tx 1678 net/wireless/nl80211.c stypes = mgmt_stypes[ift].tx; tx 12297 net/wireless/nl80211.c bool tx) tx 12319 net/wireless/nl80211.c if (tx) { tx 1590 net/wireless/trace.h TP_PROTO(struct wiphy *wiphy, int ret, u32 tx, u32 rx), tx 1591 net/wireless/trace.h TP_ARGS(wiphy, ret, tx, rx), tx 1595 net/wireless/trace.h __field(u32, tx) tx 1601 net/wireless/trace.h __entry->tx = tx; tx 1605 net/wireless/trace.h WIPHY_PR_ARG, __entry->ret, __entry->tx, __entry->rx) tx 1609 net/wireless/trace.h TP_PROTO(struct wiphy *wiphy, u32 tx, u32 tx_max, tx 1611 net/wireless/trace.h TP_ARGS(wiphy, tx, tx_max, rx, rx_max), tx 1614 net/wireless/trace.h __field(u32, tx) tx 1621 net/wireless/trace.h __entry->tx = tx; tx 1627 net/wireless/trace.h WIPHY_PR_ARG, __entry->tx, __entry->tx_max, __entry->rx, tx 1632 net/wireless/trace.h TP_PROTO(struct wiphy *wiphy, u32 tx, u32 rx), tx 1633 net/wireless/trace.h TP_ARGS(wiphy, rx, tx), tx 1636 net/wireless/trace.h __field(u32, tx) tx 1641 net/wireless/trace.h __entry->tx = tx; tx 1645 net/wireless/trace.h WIPHY_PR_ARG, __entry->tx, __entry->rx) tx 1649 net/wireless/trace.h TP_PROTO(struct wiphy *wiphy, u32 tx, u32 rx), tx 1650 net/wireless/trace.h TP_ARGS(wiphy, rx, tx) tx 30 net/xdp/xdp_umem.c if (!xs->tx) tx 42 net/xdp/xdp_umem.c if (!xs->tx) tx 77 net/xdp/xsk.c xs->tx->ring->flags |= XDP_RING_NEED_WAKEUP; tx 104 net/xdp/xsk.c xs->tx->ring->flags &= ~XDP_RING_NEED_WAKEUP; tx 292 net/xdp/xsk.c if (!xskq_peek_desc(xs->tx, desc, umem)) tx 298 net/xdp/xsk.c xskq_discard_desc(xs->tx); tx 353 net/xdp/xsk.c while (xskq_peek_desc(xs->tx, &desc, xs->umem)) { tx 386 net/xdp/xsk.c xskq_discard_desc(xs->tx); tx 411 net/xdp/xsk.c if (unlikely(!xs->tx)) tx 454 net/xdp/xsk.c if (xs->tx && !xskq_full_desc(xs->tx)) tx 564 net/xdp/xsk.c xskq_destroy(xs->tx); tx 642 net/xdp/xsk.c if (!xs->rx && !xs->tx) { tx 708 net/xdp/xsk.c xskq_set_umem(xs->tx, xs->umem->size, xs->umem->chunk_mask); tx 761 net/xdp/xsk.c q = (optname == XDP_TX_RING) ? &xs->tx : &xs->rx; tx 765 net/xdp/xsk.c xs->tx->ring->flags |= XDP_RING_NEED_WAKEUP; tx 873 net/xdp/xsk.c stats.tx_invalid_descs = xskq_nb_invalid_descs(xs->tx); tx 902 net/xdp/xsk.c &off.tx); tx 909 net/xdp/xsk.c off.tx.flags = offsetof(struct xdp_rxtx_ring, tx 920 net/xdp/xsk.c xsk_enter_rxtx_offsets(&off_v1.tx); tx 979 net/xdp/xsk.c q = READ_ONCE(xs->tx); tx 15 net/xdp/xsk.h struct xdp_ring_offset_v1 tx; tx 42 net/xdp/xsk_diag.c if (!err && xs->tx) tx 43 net/xdp/xsk_diag.c err = xsk_diag_put_ring(xs->tx, XDP_DIAG_TX_RING, nlskb); tx 89 samples/bpf/xdpsock_user.c struct xsk_ring_prod tx; tx 329 samples/bpf/xdpsock_user.c &xsk->rx, &xsk->tx, &cfg); tx 497 samples/bpf/xdpsock_user.c if (!opt_need_wakeup || xsk_ring_prod__needs_wakeup(&xsk->tx)) tx 537 samples/bpf/xdpsock_user.c if (!opt_need_wakeup || xsk_ring_prod__needs_wakeup(&xsk->tx)) tx 615 samples/bpf/xdpsock_user.c if (xsk_ring_prod__reserve(&xsk->tx, BATCH_SIZE, &idx) == BATCH_SIZE) { tx 619 samples/bpf/xdpsock_user.c xsk_ring_prod__tx_desc(&xsk->tx, idx + i)->addr = tx 621 samples/bpf/xdpsock_user.c xsk_ring_prod__tx_desc(&xsk->tx, idx + i)->len = tx 625 samples/bpf/xdpsock_user.c xsk_ring_prod__submit(&xsk->tx, BATCH_SIZE); tx 676 samples/bpf/xdpsock_user.c ret = xsk_ring_prod__reserve(&xsk->tx, rcvd, &idx_tx); tx 680 samples/bpf/xdpsock_user.c if (xsk_ring_prod__needs_wakeup(&xsk->tx)) tx 682 samples/bpf/xdpsock_user.c ret = xsk_ring_prod__reserve(&xsk->tx, rcvd, &idx_tx); tx 696 samples/bpf/xdpsock_user.c xsk_ring_prod__tx_desc(&xsk->tx, idx_tx)->addr = orig; tx 697 samples/bpf/xdpsock_user.c xsk_ring_prod__tx_desc(&xsk->tx, idx_tx++)->len = len; tx 700 samples/bpf/xdpsock_user.c xsk_ring_prod__submit(&xsk->tx, rcvd); tx 377 samples/mic/mpssd/mpssd.c static inline void txrx_prepare(int type, bool tx, struct mic_vring *vr, tx 380 samples/mic/mpssd/mpssd.c copy->vr_idx = tx ? 0 : 1; tx 890 scripts/kconfig/gconf.c gint tx = (gint) event->x; tx 894 scripts/kconfig/gconf.c gtk_tree_view_get_path_at_pos(view, tx, ty, &path, &column, &cx, tx 1006 scripts/kconfig/gconf.c gint tx = (gint) event->x; tx 1010 scripts/kconfig/gconf.c gtk_tree_view_get_path_at_pos(view, tx, ty, &path, &column, &cx, tx 472 sound/firewire/amdtp-stream.c params->header_length = s->ctx_data.tx.ctx_header_size; tx 473 sound/firewire/amdtp-stream.c params->payload_length = s->ctx_data.tx.max_ctx_payload_length; tx 590 sound/firewire/amdtp-stream.c if (*data_blocks > 0 && s->ctx_data.tx.dbc_interval > 0) tx 591 sound/firewire/amdtp-stream.c dbc_interval = s->ctx_data.tx.dbc_interval; tx 623 sound/firewire/amdtp-stream.c if (*payload_length > s->ctx_data.tx.ctx_header_size + tx 624 sound/firewire/amdtp-stream.c s->ctx_data.tx.max_ctx_payload_length) { tx 627 sound/firewire/amdtp-stream.c *payload_length, s->ctx_data.tx.max_ctx_payload_length); tx 715 sound/firewire/amdtp-stream.c s->ctx_data.tx.ctx_header_size / sizeof(*ctx_header); tx 832 sound/firewire/amdtp-stream.c packets = header_length / s->ctx_data.tx.ctx_header_size; tx 975 sound/firewire/amdtp-stream.c s->ctx_data.tx.max_ctx_payload_length = max_ctx_payload_size; tx 976 sound/firewire/amdtp-stream.c s->ctx_data.tx.ctx_header_size = ctx_header_size; tx 134 sound/firewire/amdtp-stream.h } tx; tx 98 sound/firewire/dice/dice-proc.c } tx; tx 172 sound/firewire/dice/dice-proc.c quadlets = min_t(u32, tx_rx_header.size, sizeof(buf.tx) / 4); tx 174 sound/firewire/dice/dice-proc.c if (dice_proc_read_mem(dice, &buf.tx, sections[2] + 2 + tx 179 sound/firewire/dice/dice-proc.c snd_iprintf(buffer, " iso channel: %d\n", (int)buf.tx.iso); tx 181 sound/firewire/dice/dice-proc.c buf.tx.number_audio); tx 182 sound/firewire/dice/dice-proc.c snd_iprintf(buffer, " midi ports: %u\n", buf.tx.number_midi); tx 183 sound/firewire/dice/dice-proc.c snd_iprintf(buffer, " speed: S%u\n", 100u << buf.tx.speed); tx 185 sound/firewire/dice/dice-proc.c dice_proc_fixup_string(buf.tx.names, TX_NAMES_SIZE); tx 186 sound/firewire/dice/dice-proc.c snd_iprintf(buffer, " names: %s\n", buf.tx.names); tx 190 sound/firewire/dice/dice-proc.c buf.tx.ac3_caps); tx 192 sound/firewire/dice/dice-proc.c buf.tx.ac3_enable); tx 58 sound/firewire/fireworks/fireworks_stream.c efw->tx_stream.ctx_data.tx.dbc_interval = 8; tx 154 sound/soc/bcm/bcm2835-i2s.c bool tx, bool rx) tx 164 sound/soc/bcm/bcm2835-i2s.c off = tx ? BCM2835_I2S_TXON : 0; tx 167 sound/soc/bcm/bcm2835-i2s.c clr = tx ? BCM2835_I2S_TXCLR : 0; tx 266 sound/soc/codecs/cs42xx8.c bool tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK; tx 273 sound/soc/codecs/cs42xx8.c if (tx) tx 276 sound/soc/codecs/cs42xx8.c rate[tx] = params_rate(params); tx 277 sound/soc/codecs/cs42xx8.c rate[!tx] = cs42xx8->rate[!tx]; tx 279 sound/soc/codecs/cs42xx8.c ratio[tx] = rate[tx] > 0 ? cs42xx8->sysclk / rate[tx] : 0; tx 280 sound/soc/codecs/cs42xx8.c ratio[!tx] = rate[!tx] > 0 ? cs42xx8->sysclk / rate[!tx] : 0; tx 303 sound/soc/codecs/cs42xx8.c condition1 = ((fm[tx] == CS42XX8_FM_AUTO) ? tx 304 sound/soc/codecs/cs42xx8.c (cs42xx8_ratios[i].ratio[0] == ratio[tx] || tx 305 sound/soc/codecs/cs42xx8.c cs42xx8_ratios[i].ratio[1] == ratio[tx] || tx 306 sound/soc/codecs/cs42xx8.c cs42xx8_ratios[i].ratio[2] == ratio[tx]) : tx 307 sound/soc/codecs/cs42xx8.c (cs42xx8_ratios[i].ratio[fm[tx]] == ratio[tx])) && tx 311 sound/soc/codecs/cs42xx8.c if (!ratio[tx]) tx 315 sound/soc/codecs/cs42xx8.c condition2 = ((fm[!tx] == CS42XX8_FM_AUTO) ? tx 316 sound/soc/codecs/cs42xx8.c (cs42xx8_ratios[i].ratio[0] == ratio[!tx] || tx 317 sound/soc/codecs/cs42xx8.c cs42xx8_ratios[i].ratio[1] == ratio[!tx] || tx 318 sound/soc/codecs/cs42xx8.c cs42xx8_ratios[i].ratio[2] == ratio[!tx]) : tx 319 sound/soc/codecs/cs42xx8.c (cs42xx8_ratios[i].ratio[fm[!tx]] == ratio[!tx])); tx 321 sound/soc/codecs/cs42xx8.c if (!ratio[!tx]) tx 337 sound/soc/codecs/cs42xx8.c cs42xx8->rate[tx] = params_rate(params); tx 343 sound/soc/codecs/cs42xx8.c CS42XX8_FUNCMOD_xC_FM_MASK(tx) | mask, tx 344 sound/soc/codecs/cs42xx8.c CS42XX8_FUNCMOD_xC_FM(tx, fm[tx]) | val); tx 354 sound/soc/codecs/cs42xx8.c bool tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK; tx 357 sound/soc/codecs/cs42xx8.c cs42xx8->rate[tx] = 0; tx 360 sound/soc/codecs/cs42xx8.c CS42XX8_FUNCMOD_xC_FM_MASK(tx), tx 361 sound/soc/codecs/cs42xx8.c CS42XX8_FUNCMOD_xC_FM(tx, CS42XX8_FM_AUTO)); tx 3972 sound/soc/codecs/wcd9335.c bool tx; tx 3982 sound/soc/codecs/wcd9335.c tx = (j >= 16 ? true : false); tx 3983 sound/soc/codecs/wcd9335.c port_id = (tx ? j - 16 : j); tx 3987 sound/soc/codecs/wcd9335.c if (!tx) tx 4006 sound/soc/codecs/wcd9335.c __func__, (tx ? "TX" : "RX"), port_id, val); tx 4011 sound/soc/codecs/wcd9335.c __func__, (tx ? "TX" : "RX"), port_id, val); tx 4015 sound/soc/codecs/wcd9335.c if (!tx) tx 816 sound/soc/codecs/wm8960.c bool tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK; tx 848 sound/soc/codecs/wm8960.c if (tx) { tx 861 sound/soc/codecs/wm8960.c wm8960->is_stream_in_use[tx] = true; tx 863 sound/soc/codecs/wm8960.c if (!wm8960->is_stream_in_use[!tx]) tx 874 sound/soc/codecs/wm8960.c bool tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK; tx 876 sound/soc/codecs/wm8960.c wm8960->is_stream_in_use[tx] = false; tx 143 sound/soc/fsl/fsl-asoc-card.c bool tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK; tx 162 sound/soc/fsl/fsl-asoc-card.c ret = snd_soc_dai_set_sysclk(rtd->cpu_dai, cpu_priv->sysclk_id[tx], tx 163 sound/soc/fsl/fsl-asoc-card.c cpu_priv->sysclk_freq[tx], tx 164 sound/soc/fsl/fsl-asoc-card.c cpu_priv->sysclk_dir[tx]); tx 134 sound/soc/fsl/fsl_asrc_dma.c bool tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK; tx 149 sound/soc/fsl/fsl_asrc_dma.c u8 dir = tx ? OUT : IN; tx 203 sound/soc/fsl/fsl_asrc_dma.c tmp_chan = dma_request_slave_channel(dev_be, tx ? "tx" : "rx"); tx 233 sound/soc/fsl/fsl_asrc_dma.c if (tx) { tx 138 sound/soc/fsl/fsl_esai.c static int fsl_esai_divisor_cal(struct snd_soc_dai *dai, bool tx, u32 ratio, tx 207 sound/soc/fsl/fsl_esai.c regmap_update_bits(esai_priv->regmap, REG_ESAI_xCCR(tx), tx 216 sound/soc/fsl/fsl_esai.c regmap_update_bits(esai_priv->regmap, REG_ESAI_xCCR(tx), tx 238 sound/soc/fsl/fsl_esai.c bool tx = (clk_id <= ESAI_HCKT_EXTAL || esai_priv->synchronous); tx 246 sound/soc/fsl/fsl_esai.c in ? "in" : "out", tx ? 'T' : 'R'); tx 251 sound/soc/fsl/fsl_esai.c if (freq == esai_priv->hck_rate[tx] && dir == esai_priv->hck_dir[tx]) tx 255 sound/soc/fsl/fsl_esai.c esai_priv->sck_div[tx] = true; tx 258 sound/soc/fsl/fsl_esai.c regmap_update_bits(esai_priv->regmap, REG_ESAI_xCCR(tx), tx 297 sound/soc/fsl/fsl_esai.c tx ? 'T' : 'R'); tx 304 sound/soc/fsl/fsl_esai.c ecr |= tx ? ESAI_ECR_ETO : ESAI_ECR_ERO; tx 309 sound/soc/fsl/fsl_esai.c tx ? 'T' : 'R'); tx 313 sound/soc/fsl/fsl_esai.c ret = fsl_esai_divisor_cal(dai, tx, ratio, false, 0); tx 317 sound/soc/fsl/fsl_esai.c esai_priv->sck_div[tx] = false; tx 320 sound/soc/fsl/fsl_esai.c esai_priv->hck_dir[tx] = dir; tx 321 sound/soc/fsl/fsl_esai.c esai_priv->hck_rate[tx] = freq; tx 324 sound/soc/fsl/fsl_esai.c tx ? ESAI_ECR_ETI | ESAI_ECR_ETO : tx 333 sound/soc/fsl/fsl_esai.c static int fsl_esai_set_bclk(struct snd_soc_dai *dai, bool tx, u32 freq) tx 336 sound/soc/fsl/fsl_esai.c u32 hck_rate = esai_priv->hck_rate[tx]; tx 341 sound/soc/fsl/fsl_esai.c if (esai_priv->slave_mode || esai_priv->sck_rate[tx] == freq) tx 354 sound/soc/fsl/fsl_esai.c tx ? 'T' : 'R'); tx 359 sound/soc/fsl/fsl_esai.c if (!esai_priv->sck_div[tx] && (ratio > 16 || ratio == 0)) { tx 364 sound/soc/fsl/fsl_esai.c ret = fsl_esai_divisor_cal(dai, tx, ratio, true, tx 365 sound/soc/fsl/fsl_esai.c esai_priv->sck_div[tx] ? 0 : ratio); tx 370 sound/soc/fsl/fsl_esai.c esai_priv->sck_rate[tx] = freq; tx 509 sound/soc/fsl/fsl_esai.c bool tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK; tx 523 sound/soc/fsl/fsl_esai.c ret = fsl_esai_set_bclk(dai, esai_priv->synchronous || tx, bclk); tx 530 sound/soc/fsl/fsl_esai.c regmap_update_bits(esai_priv->regmap, REG_ESAI_xCR(tx), mask, val); tx 532 sound/soc/fsl/fsl_esai.c if (!tx && esai_priv->synchronous) tx 536 sound/soc/fsl/fsl_esai.c regmap_update_bits(esai_priv->regmap, REG_ESAI_xCR(tx), tx 540 sound/soc/fsl/fsl_esai.c regmap_update_bits(esai_priv->regmap, REG_ESAI_xFCR(tx), tx 544 sound/soc/fsl/fsl_esai.c (tx ? ESAI_xFCR_TE_MASK | ESAI_xFCR_TIEN : ESAI_xFCR_RE_MASK); tx 546 sound/soc/fsl/fsl_esai.c (tx ? ESAI_xFCR_TE(pins) | ESAI_xFCR_TIEN : ESAI_xFCR_RE(pins)); tx 548 sound/soc/fsl/fsl_esai.c regmap_update_bits(esai_priv->regmap, REG_ESAI_xFCR(tx), mask, val); tx 550 sound/soc/fsl/fsl_esai.c if (tx) tx 618 sound/soc/fsl/fsl_esai.c static void fsl_esai_trigger_start(struct fsl_esai *esai_priv, bool tx) tx 620 sound/soc/fsl/fsl_esai.c u8 i, channels = esai_priv->channels[tx]; tx 624 sound/soc/fsl/fsl_esai.c regmap_update_bits(esai_priv->regmap, REG_ESAI_xFCR(tx), tx 628 sound/soc/fsl/fsl_esai.c for (i = 0; tx && i < channels; i++) tx 643 sound/soc/fsl/fsl_esai.c regmap_update_bits(esai_priv->regmap, REG_ESAI_xCR(tx), tx 644 sound/soc/fsl/fsl_esai.c tx ? ESAI_xCR_TE_MASK : ESAI_xCR_RE_MASK, tx 645 sound/soc/fsl/fsl_esai.c tx ? ESAI_xCR_TE(pins) : ESAI_xCR_RE(pins)); tx 646 sound/soc/fsl/fsl_esai.c mask = tx ? esai_priv->tx_mask : esai_priv->rx_mask; tx 648 sound/soc/fsl/fsl_esai.c regmap_update_bits(esai_priv->regmap, REG_ESAI_xSMB(tx), tx 650 sound/soc/fsl/fsl_esai.c regmap_update_bits(esai_priv->regmap, REG_ESAI_xSMA(tx), tx 654 sound/soc/fsl/fsl_esai.c regmap_update_bits(esai_priv->regmap, REG_ESAI_xCR(tx), tx 658 sound/soc/fsl/fsl_esai.c static void fsl_esai_trigger_stop(struct fsl_esai *esai_priv, bool tx) tx 660 sound/soc/fsl/fsl_esai.c regmap_update_bits(esai_priv->regmap, REG_ESAI_xCR(tx), tx 663 sound/soc/fsl/fsl_esai.c regmap_update_bits(esai_priv->regmap, REG_ESAI_xCR(tx), tx 664 sound/soc/fsl/fsl_esai.c tx ? ESAI_xCR_TE_MASK : ESAI_xCR_RE_MASK, 0); tx 665 sound/soc/fsl/fsl_esai.c regmap_update_bits(esai_priv->regmap, REG_ESAI_xSMA(tx), tx 667 sound/soc/fsl/fsl_esai.c regmap_update_bits(esai_priv->regmap, REG_ESAI_xSMB(tx), tx 671 sound/soc/fsl/fsl_esai.c regmap_update_bits(esai_priv->regmap, REG_ESAI_xFCR(tx), tx 673 sound/soc/fsl/fsl_esai.c regmap_update_bits(esai_priv->regmap, REG_ESAI_xFCR(tx), tx 680 sound/soc/fsl/fsl_esai.c bool tx = true, rx = false, enabled[2]; tx 688 sound/soc/fsl/fsl_esai.c enabled[tx] = tfcr & ESAI_xFCR_xFEN; tx 692 sound/soc/fsl/fsl_esai.c fsl_esai_trigger_stop(esai_priv, tx); tx 718 sound/soc/fsl/fsl_esai.c if (enabled[tx]) tx 719 sound/soc/fsl/fsl_esai.c fsl_esai_trigger_start(esai_priv, tx); tx 730 sound/soc/fsl/fsl_esai.c bool tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK; tx 733 sound/soc/fsl/fsl_esai.c esai_priv->channels[tx] = substream->runtime->channels; tx 740 sound/soc/fsl/fsl_esai.c fsl_esai_trigger_start(esai_priv, tx); tx 747 sound/soc/fsl/fsl_esai.c fsl_esai_trigger_stop(esai_priv, tx); tx 22 sound/soc/fsl/fsl_esai.h #define REG_ESAI_xFCR(tx) (tx ? REG_ESAI_TFCR : REG_ESAI_RFCR) tx 23 sound/soc/fsl/fsl_esai.h #define REG_ESAI_xFSR(tx) (tx ? REG_ESAI_TFSR : REG_ESAI_RFSR) tx 41 sound/soc/fsl/fsl_esai.h #define REG_ESAI_xCR(tx) (tx ? REG_ESAI_TCR : REG_ESAI_RCR) tx 42 sound/soc/fsl/fsl_esai.h #define REG_ESAI_xCCR(tx) (tx ? REG_ESAI_TCCR : REG_ESAI_RCCR) tx 47 sound/soc/fsl/fsl_esai.h #define REG_ESAI_xSMA(tx) (tx ? REG_ESAI_TSMA : REG_ESAI_RSMA) tx 48 sound/soc/fsl/fsl_esai.h #define REG_ESAI_xSMB(tx) (tx ? REG_ESAI_TSMB : REG_ESAI_RSMB) tx 155 sound/soc/fsl/fsl_sai.c bool tx = fsl_dir == FSL_FMT_TRANSMITTER; tx 175 sound/soc/fsl/fsl_sai.c regmap_update_bits(sai->regmap, FSL_SAI_xCR2(tx, ofs), tx 209 sound/soc/fsl/fsl_sai.c bool tx = fsl_dir == FSL_FMT_TRANSMITTER; tx 303 sound/soc/fsl/fsl_sai.c regmap_update_bits(sai->regmap, FSL_SAI_xCR2(tx, ofs), tx 305 sound/soc/fsl/fsl_sai.c regmap_update_bits(sai->regmap, FSL_SAI_xCR4(tx, ofs), tx 329 sound/soc/fsl/fsl_sai.c static int fsl_sai_set_bclk(struct snd_soc_dai *dai, bool tx, u32 freq) tx 369 sound/soc/fsl/fsl_sai.c sai->mclk_id[tx] = id; tx 379 sound/soc/fsl/fsl_sai.c tx ? 'T' : 'R', freq); tx 394 sound/soc/fsl/fsl_sai.c (!tx && !sai->synchronous[RX])) { tx 397 sound/soc/fsl/fsl_sai.c FSL_SAI_CR2_MSEL(sai->mclk_id[tx])); tx 401 sound/soc/fsl/fsl_sai.c (tx && !sai->synchronous[TX])) { tx 404 sound/soc/fsl/fsl_sai.c FSL_SAI_CR2_MSEL(sai->mclk_id[tx])); tx 410 sound/soc/fsl/fsl_sai.c sai->mclk_id[tx], savediv, savesub); tx 421 sound/soc/fsl/fsl_sai.c bool tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK; tx 437 sound/soc/fsl/fsl_sai.c ret = fsl_sai_set_bclk(cpu_dai, tx, tx 441 sound/soc/fsl/fsl_sai.c ret = fsl_sai_set_bclk(cpu_dai, tx, tx 449 sound/soc/fsl/fsl_sai.c ret = clk_prepare_enable(sai->mclk_clk[sai->mclk_id[tx]]); tx 478 sound/soc/fsl/fsl_sai.c if (!sai->synchronous[TX] && sai->synchronous[RX] && !tx) { tx 487 sound/soc/fsl/fsl_sai.c } else if (!sai->synchronous[RX] && sai->synchronous[TX] && tx) { tx 499 sound/soc/fsl/fsl_sai.c regmap_update_bits(sai->regmap, FSL_SAI_xCR4(tx, ofs), tx 502 sound/soc/fsl/fsl_sai.c regmap_update_bits(sai->regmap, FSL_SAI_xCR5(tx, ofs), tx 505 sound/soc/fsl/fsl_sai.c regmap_write(sai->regmap, FSL_SAI_xMR(tx), ~0UL - ((1 << channels) - 1)); tx 514 sound/soc/fsl/fsl_sai.c bool tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK; tx 518 sound/soc/fsl/fsl_sai.c clk_disable_unprepare(sai->mclk_clk[sai->mclk_id[tx]]); tx 532 sound/soc/fsl/fsl_sai.c bool tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK; tx 553 sound/soc/fsl/fsl_sai.c regmap_update_bits(sai->regmap, FSL_SAI_xCSR(tx, ofs), tx 561 sound/soc/fsl/fsl_sai.c regmap_update_bits(sai->regmap, FSL_SAI_xCSR(tx, ofs), tx 567 sound/soc/fsl/fsl_sai.c regmap_update_bits(sai->regmap, FSL_SAI_xCSR(tx, ofs), tx 569 sound/soc/fsl/fsl_sai.c regmap_update_bits(sai->regmap, FSL_SAI_xCSR(tx, ofs), tx 573 sound/soc/fsl/fsl_sai.c regmap_read(sai->regmap, FSL_SAI_xCSR(!tx, ofs), &xcsr); tx 585 sound/soc/fsl/fsl_sai.c FSL_SAI_xCSR(tx, ofs), &xcsr); tx 624 sound/soc/fsl/fsl_sai.c bool tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK; tx 627 sound/soc/fsl/fsl_sai.c regmap_update_bits(sai->regmap, FSL_SAI_xCR3(tx, ofs), tx 638 sound/soc/fsl/fsl_sai.c tx ? sai->dma_params_tx.maxburst : tx 652 sound/soc/fsl/fsl_sai.c bool tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK; tx 654 sound/soc/fsl/fsl_sai.c regmap_update_bits(sai->regmap, FSL_SAI_xCR3(tx, ofs), tx 64 sound/soc/fsl/fsl_sai.h #define FSL_SAI_xCSR(tx, ofs) (tx ? FSL_SAI_TCSR(ofs) : FSL_SAI_RCSR(ofs)) tx 65 sound/soc/fsl/fsl_sai.h #define FSL_SAI_xCR1(tx, ofs) (tx ? FSL_SAI_TCR1(ofs) : FSL_SAI_RCR1(ofs)) tx 66 sound/soc/fsl/fsl_sai.h #define FSL_SAI_xCR2(tx, ofs) (tx ? FSL_SAI_TCR2(ofs) : FSL_SAI_RCR2(ofs)) tx 67 sound/soc/fsl/fsl_sai.h #define FSL_SAI_xCR3(tx, ofs) (tx ? FSL_SAI_TCR3(ofs) : FSL_SAI_RCR3(ofs)) tx 68 sound/soc/fsl/fsl_sai.h #define FSL_SAI_xCR4(tx, ofs) (tx ? FSL_SAI_TCR4(ofs) : FSL_SAI_RCR4(ofs)) tx 69 sound/soc/fsl/fsl_sai.h #define FSL_SAI_xCR5(tx, ofs) (tx ? FSL_SAI_TCR5(ofs) : FSL_SAI_RCR5(ofs)) tx 70 sound/soc/fsl/fsl_sai.h #define FSL_SAI_xDR(tx, ofs) (tx ? FSL_SAI_TDR(ofs) : FSL_SAI_RDR(ofs)) tx 71 sound/soc/fsl/fsl_sai.h #define FSL_SAI_xFR(tx, ofs) (tx ? FSL_SAI_TFR(ofs) : FSL_SAI_RFR(ofs)) tx 72 sound/soc/fsl/fsl_sai.h #define FSL_SAI_xMR(tx) (tx ? FSL_SAI_TMR : FSL_SAI_RMR) tx 36 sound/soc/fsl/fsl_spdif.c #define SIE_INTR_FOR(tx) (tx ? INTR_FOR_PLAYBACK : INTR_FOR_CAPTURE) tx 602 sound/soc/fsl/fsl_spdif.c bool tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK; tx 603 sound/soc/fsl/fsl_spdif.c u32 intr = SIE_INTR_FOR(tx); tx 604 sound/soc/fsl/fsl_spdif.c u32 dmaen = SCR_DMA_xX_EN(tx); tx 93 sound/soc/fsl/fsl_spdif.h #define SCR_DMA_xX_EN(tx) (tx ? SCR_DMA_TX_EN : SCR_DMA_RX_EN) tx 405 sound/soc/fsl/fsl_ssi.c static void fsl_ssi_config_enable(struct fsl_ssi *ssi, bool tx) tx 408 sound/soc/fsl/fsl_ssi.c int dir = tx ? TX : RX; tx 413 sound/soc/fsl/fsl_ssi.c SSI_SOR_xX_CLR(tx), SSI_SOR_xX_CLR(tx)); tx 450 sound/soc/fsl/fsl_ssi.c if (ssi->use_dma && tx) { tx 505 sound/soc/fsl/fsl_ssi.c static void fsl_ssi_config_disable(struct fsl_ssi *ssi, bool tx) tx 509 sound/soc/fsl/fsl_ssi.c int adir = tx ? RX : TX; tx 510 sound/soc/fsl/fsl_ssi.c int dir = tx ? TX : RX; tx 563 sound/soc/fsl/fsl_ssi.c SSI_SOR_xX_CLR(tx), SSI_SOR_xX_CLR(tx)); tx 676 sound/soc/fsl/fsl_ssi.c bool tx2, tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK; tx 765 sound/soc/fsl/fsl_ssi.c tx2 = tx || ssi->synchronous; tx 794 sound/soc/fsl/fsl_ssi.c bool tx2, tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK; tx 847 sound/soc/fsl/fsl_ssi.c tx2 = tx || ssi->synchronous; tx 1063 sound/soc/fsl/fsl_ssi.c bool tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK; tx 1076 sound/soc/fsl/fsl_ssi.c if (tx && fsl_ssi_is_ac97(ssi)) tx 1078 sound/soc/fsl/fsl_ssi.c fsl_ssi_config_enable(ssi, tx); tx 1084 sound/soc/fsl/fsl_ssi.c fsl_ssi_config_disable(ssi, tx); tx 33 sound/soc/fsl/fsl_ssi.h #define REG_SSI_SxCR(tx) ((tx) ? REG_SSI_STCR : REG_SSI_SRCR) tx 38 sound/soc/fsl/fsl_ssi.h #define REG_SSI_SxCCR(tx) ((tx) ? REG_SSI_STCCR : REG_SSI_SRCCR) tx 67 sound/soc/fsl/fsl_ssi.h #define REG_SSI_SxMSK(tx) ((tx) ? REG_SSI_STMSK : REG_SSI_SRMSK) tx 250 sound/soc/fsl/fsl_ssi.h #define SSI_SOR_xX_CLR(tx) ((tx) ? SSI_SOR_TX_CLR : SSI_SOR_RX_CLR) tx 78 sound/soc/fsl/imx-audmix.c bool tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK; tx 84 sound/soc/fsl/imx-audmix.c fmt |= tx ? SND_SOC_DAIFMT_CBS_CFS : SND_SOC_DAIFMT_CBM_CFM; tx 85 sound/soc/fsl/imx-audmix.c dir = tx ? SND_SOC_CLOCK_OUT : SND_SOC_CLOCK_IN; tx 117 sound/soc/fsl/imx-audmix.c bool tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK; tx 121 sound/soc/fsl/imx-audmix.c if (!tx) tx 158 sound/soc/intel/atom/sst-atom-controls.c unsigned int is_tx = e->tx; tx 217 sound/soc/intel/atom/sst-atom-controls.c unsigned int is_tx = e->tx; tx 838 sound/soc/intel/atom/sst-atom-controls.h bool tx; tx 847 sound/soc/intel/atom/sst-atom-controls.h (struct sst_enum){ .reg = s_ch_no, .tx = is_tx, .max = 4+1, .texts = xtexts, } tx 214 sound/soc/intel/baytrail/sst-baytrail-ipc.c u64 header = msg->tx.header; tx 625 sound/soc/intel/baytrail/sst-baytrail-ipc.c if (msg->tx.header & IPC_HEADER_LARGE(true)) tx 626 sound/soc/intel/baytrail/sst-baytrail-ipc.c sst_dsp_outbox_write(ipc->dsp, msg->tx.data, msg->tx.size); tx 628 sound/soc/intel/baytrail/sst-baytrail-ipc.c sst_dsp_shim_write64_unlocked(ipc->dsp, SST_IPCX, msg->tx.header); tx 650 sound/soc/intel/baytrail/sst-baytrail-ipc.c *(u32 *)msg->tx.data = (u32)(msg->tx.header & (u32)-1); tx 651 sound/soc/intel/baytrail/sst-baytrail-ipc.c memcpy(msg->tx.data + sizeof(u32), tx_data, tx_size); tx 652 sound/soc/intel/baytrail/sst-baytrail-ipc.c msg->tx.size += sizeof(u32); tx 93 sound/soc/intel/common/sst-ipc.c msg->tx.header = request.header; tx 94 sound/soc/intel/common/sst-ipc.c msg->tx.size = request.size; tx 125 sound/soc/intel/common/sst-ipc.c ipc->msg[i].tx.data = kzalloc(ipc->tx_data_max_size, GFP_KERNEL); tx 126 sound/soc/intel/common/sst-ipc.c if (ipc->msg[i].tx.data == NULL) tx 131 sound/soc/intel/common/sst-ipc.c kfree(ipc->msg[i].tx.data); tx 143 sound/soc/intel/common/sst-ipc.c kfree(ipc->msg[i-1].tx.data); tx 236 sound/soc/intel/common/sst-ipc.c if ((msg->tx.header & mask) == header) tx 310 sound/soc/intel/common/sst-ipc.c kfree(ipc->msg[i].tx.data); tx 28 sound/soc/intel/common/sst-ipc.h struct sst_ipc_message tx; tx 514 sound/soc/intel/haswell/sst-haswell-ipc.c u32 header = msg->tx.header & ~(IPC_STATUS_MASK | IPC_GLB_REPLY_MASK); tx 2056 sound/soc/intel/haswell/sst-haswell-ipc.c sst_dsp_outbox_write(ipc->dsp, msg->tx.data, msg->tx.size); tx 2057 sound/soc/intel/haswell/sst-haswell-ipc.c sst_dsp_ipc_msg_tx(ipc->dsp, msg->tx.header); tx 2078 sound/soc/intel/haswell/sst-haswell-ipc.c memcpy(msg->tx.data, tx_data, tx_size); tx 369 sound/soc/intel/skylake/cnl-sst.c struct skl_ipc_header *header = (struct skl_ipc_header *)(&msg->tx.header); tx 371 sound/soc/intel/skylake/cnl-sst.c if (msg->tx.size) tx 372 sound/soc/intel/skylake/cnl-sst.c sst_dsp_outbox_write(ipc->dsp, msg->tx.data, msg->tx.size); tx 284 sound/soc/intel/skylake/skl-sst-ipc.c memcpy(msg->tx.data, tx_data, tx_size); tx 298 sound/soc/intel/skylake/skl-sst-ipc.c struct skl_ipc_header *header = (struct skl_ipc_header *)(&msg->tx.header); tx 300 sound/soc/intel/skylake/skl-sst-ipc.c if (msg->tx.size) tx 301 sound/soc/intel/skylake/skl-sst-ipc.c sst_dsp_outbox_write(ipc->dsp, msg->tx.data, msg->tx.size); tx 135 sound/soc/mediatek/common/mtk-btcvsd.c struct mtk_btcvsd_snd_stream *tx; tx 212 sound/soc/mediatek/common/mtk-btcvsd.c bt->tx->state, bt->rx->state, bt->irq_disabled); tx 216 sound/soc/mediatek/common/mtk-btcvsd.c if (bt->tx->state == BT_SCO_STATE_IDLE && tx 234 sound/soc/mediatek/common/mtk-btcvsd.c memset(bt->tx, 0, sizeof(*bt->tx)); tx 237 sound/soc/mediatek/common/mtk-btcvsd.c bt->tx->packet_size = BTCVSD_TX_PACKET_SIZE; tx 238 sound/soc/mediatek/common/mtk-btcvsd.c bt->tx->buf_size = BTCVSD_TX_BUF_SIZE; tx 239 sound/soc/mediatek/common/mtk-btcvsd.c bt->tx->timeout = 0; tx 240 sound/soc/mediatek/common/mtk-btcvsd.c bt->tx->rw_cnt = 0; tx 241 sound/soc/mediatek/common/mtk-btcvsd.c bt->tx->stream = SNDRV_PCM_STREAM_PLAYBACK; tx 261 sound/soc/mediatek/common/mtk-btcvsd.c ts->time_stamp_us = bt->tx->time_stamp; tx 262 sound/soc/mediatek/common/mtk-btcvsd.c ts->data_count_equi_time = bt->tx->buf_data_equivalent_time; tx 327 sound/soc/mediatek/common/mtk-btcvsd.c memset(bt->tx->temp_packet_buf, 170, SCO_PACKET_180); tx 329 sound/soc/mediatek/common/mtk-btcvsd.c memcpy(bt->tx->temp_packet_buf, tx 334 sound/soc/mediatek/common/mtk-btcvsd.c num_valid_addr = bt->tx->buffer_info.num_valid_addr; tx 343 sound/soc/mediatek/common/mtk-btcvsd.c bt->tx->buffer_info.bt_sram_addr[i]); tx 345 sound/soc/mediatek/common/mtk-btcvsd.c dst = (void *)bt->tx->buffer_info.bt_sram_addr[i]; tx 348 sound/soc/mediatek/common/mtk-btcvsd.c bt->tx->temp_packet_buf, dst, tx 349 sound/soc/mediatek/common/mtk-btcvsd.c bt->tx->buffer_info.packet_length, tx 350 sound/soc/mediatek/common/mtk-btcvsd.c bt->tx->buffer_info.packet_num); tx 435 sound/soc/mediatek/common/mtk-btcvsd.c memcpy(bt->tx->temp_packet_buf + (bt->tx->packet_size * i), tx 437 sound/soc/mediatek/common/mtk-btcvsd.c (bt->tx->packet_r % SCO_TX_PACKER_BUF_NUM) * tx 438 sound/soc/mediatek/common/mtk-btcvsd.c bt->tx->packet_size), tx 439 sound/soc/mediatek/common/mtk-btcvsd.c bt->tx->packet_size); tx 441 sound/soc/mediatek/common/mtk-btcvsd.c bt->tx->packet_r++; tx 447 sound/soc/mediatek/common/mtk-btcvsd.c if (!bt->tx->mute) { tx 449 sound/soc/mediatek/common/mtk-btcvsd.c bt->tx->temp_packet_buf, dst, tx 454 sound/soc/mediatek/common/mtk-btcvsd.c bt->tx->buffer_info.packet_length = packet_length; tx 455 sound/soc/mediatek/common/mtk-btcvsd.c bt->tx->buffer_info.packet_num = packet_num; tx 456 sound/soc/mediatek/common/mtk-btcvsd.c for (i = 0; i < bt->tx->buffer_info.num_valid_addr; i++) { tx 457 sound/soc/mediatek/common/mtk-btcvsd.c if (bt->tx->buffer_info.bt_sram_addr[i] == ap_addr_tx) { tx 466 sound/soc/mediatek/common/mtk-btcvsd.c bt->tx->buffer_info.num_valid_addr++; tx 467 sound/soc/mediatek/common/mtk-btcvsd.c next_idx = bt->tx->buffer_info.num_valid_addr - 1; tx 468 sound/soc/mediatek/common/mtk-btcvsd.c bt->tx->buffer_info.bt_sram_addr[next_idx] = ap_addr_tx; tx 472 sound/soc/mediatek/common/mtk-btcvsd.c bt->tx->buffer_info.num_valid_addr); tx 475 sound/soc/mediatek/common/mtk-btcvsd.c if (bt->tx->mute) tx 489 sound/soc/mediatek/common/mtk-btcvsd.c bt->tx->state != BT_SCO_STATE_RUNNING && tx 490 sound/soc/mediatek/common/mtk-btcvsd.c bt->tx->state != BT_SCO_STATE_ENDING && tx 491 sound/soc/mediatek/common/mtk-btcvsd.c bt->tx->state != BT_SCO_STATE_LOOPBACK) { tx 493 sound/soc/mediatek/common/mtk-btcvsd.c __func__, bt->rx->state, bt->tx->state); tx 517 sound/soc/mediatek/common/mtk-btcvsd.c if (bt->tx->state == BT_SCO_STATE_LOOPBACK) { tx 542 sound/soc/mediatek/common/mtk-btcvsd.c bt->tx->temp_packet_buf, tx 546 sound/soc/mediatek/common/mtk-btcvsd.c bt->tx->temp_packet_buf, dst, tx 550 sound/soc/mediatek/common/mtk-btcvsd.c bt->tx->rw_cnt++; tx 585 sound/soc/mediatek/common/mtk-btcvsd.c bt->tx->timeout = 0; tx 586 sound/soc/mediatek/common/mtk-btcvsd.c if ((bt->tx->state == BT_SCO_STATE_RUNNING || tx 587 sound/soc/mediatek/common/mtk-btcvsd.c bt->tx->state == BT_SCO_STATE_ENDING) && tx 588 sound/soc/mediatek/common/mtk-btcvsd.c bt->tx->trigger_start) { tx 589 sound/soc/mediatek/common/mtk-btcvsd.c if (bt->tx->xrun) { tx 593 sound/soc/mediatek/common/mtk-btcvsd.c if (bt->tx->packet_w - bt->tx->packet_r >= tx 595 sound/soc/mediatek/common/mtk-btcvsd.c bt->tx->xrun = 0; tx 601 sound/soc/mediatek/common/mtk-btcvsd.c if ((!bt->tx->xrun && tx 602 sound/soc/mediatek/common/mtk-btcvsd.c (bt->tx->packet_w - bt->tx->packet_r >= buf_cnt_tx)) || tx 603 sound/soc/mediatek/common/mtk-btcvsd.c bt->tx->state == BT_SCO_STATE_ENDING) { tx 609 sound/soc/mediatek/common/mtk-btcvsd.c bt->tx->rw_cnt++; tx 611 sound/soc/mediatek/common/mtk-btcvsd.c bt->tx->xrun = 1; tx 624 sound/soc/mediatek/common/mtk-btcvsd.c if (bt->tx->state == BT_SCO_STATE_RUNNING || tx 625 sound/soc/mediatek/common/mtk-btcvsd.c bt->tx->state == BT_SCO_STATE_ENDING) { tx 626 sound/soc/mediatek/common/mtk-btcvsd.c bt->tx->wait_flag = 1; tx 628 sound/soc/mediatek/common/mtk-btcvsd.c snd_pcm_period_elapsed(bt->tx->substream); tx 786 sound/soc/mediatek/common/mtk-btcvsd.c unsigned int packet_size = bt->tx->packet_size; tx 792 sound/soc/mediatek/common/mtk-btcvsd.c bt->tx->time_stamp = sched_clock(); tx 793 sound/soc/mediatek/common/mtk-btcvsd.c bt->tx->buf_data_equivalent_time = tx 794 sound/soc/mediatek/common/mtk-btcvsd.c (unsigned long long)(bt->tx->packet_w - bt->tx->packet_r) * tx 798 sound/soc/mediatek/common/mtk-btcvsd.c bt->tx->buf_data_equivalent_time *= 1000; tx 803 sound/soc/mediatek/common/mtk-btcvsd.c avail = bt->tx->buf_size - tx 804 sound/soc/mediatek/common/mtk-btcvsd.c (bt->tx->packet_w - bt->tx->packet_r) * packet_size; tx 806 sound/soc/mediatek/common/mtk-btcvsd.c cur_write_idx = (bt->tx->packet_w % SCO_TX_PACKER_BUF_NUM) * tx 834 sound/soc/mediatek/common/mtk-btcvsd.c cont = bt->tx->buf_size - cur_write_idx; tx 848 sound/soc/mediatek/common/mtk-btcvsd.c bt->tx->packet_w += write_size / packet_size; tx 861 sound/soc/mediatek/common/mtk-btcvsd.c return bt->tx; tx 893 sound/soc/mediatek/common/mtk-btcvsd.c bt->tx->substream = substream; tx 926 sound/soc/mediatek/common/mtk-btcvsd.c params_buffer_bytes(hw_params) % bt->tx->packet_size != 0) { tx 1013 sound/soc/mediatek/common/mtk-btcvsd.c bt_stream = bt->tx; tx 1021 sound/soc/mediatek/common/mtk-btcvsd.c bt->tx->packet_r : bt->rx->packet_w; tx 1113 sound/soc/mediatek/common/mtk-btcvsd.c bool lpbk_en = bt->tx->state == BT_SCO_STATE_LOOPBACK; tx 1126 sound/soc/mediatek/common/mtk-btcvsd.c mtk_btcvsd_snd_set_state(bt, bt->tx, BT_SCO_STATE_LOOPBACK); tx 1129 sound/soc/mediatek/common/mtk-btcvsd.c mtk_btcvsd_snd_set_state(bt, bt->tx, BT_SCO_STATE_RUNNING); tx 1141 sound/soc/mediatek/common/mtk-btcvsd.c if (!bt->tx) { tx 1146 sound/soc/mediatek/common/mtk-btcvsd.c ucontrol->value.integer.value[0] = bt->tx->mute; tx 1156 sound/soc/mediatek/common/mtk-btcvsd.c if (!bt->tx) tx 1159 sound/soc/mediatek/common/mtk-btcvsd.c bt->tx->mute = ucontrol->value.integer.value[0]; tx 1223 sound/soc/mediatek/common/mtk-btcvsd.c if (!bt->tx) tx 1226 sound/soc/mediatek/common/mtk-btcvsd.c ucontrol->value.integer.value[0] = bt->tx->rw_cnt ? 1 : 0; tx 1236 sound/soc/mediatek/common/mtk-btcvsd.c ucontrol->value.integer.value[0] = bt->tx->timeout; tx 1323 sound/soc/mediatek/common/mtk-btcvsd.c btcvsd->tx = devm_kzalloc(btcvsd->dev, sizeof(*btcvsd->tx), GFP_KERNEL); tx 1324 sound/soc/mediatek/common/mtk-btcvsd.c if (!btcvsd->tx) tx 1394 sound/soc/mediatek/common/mtk-btcvsd.c mtk_btcvsd_snd_set_state(btcvsd, btcvsd->tx, BT_SCO_STATE_IDLE); tx 19 sound/soc/meson/axg-card.c u32 tx; tx 215 sound/soc/meson/axg-card.c be->codec_masks[i].tx, tx 339 sound/soc/meson/axg-card.c u32 tx, rx; tx 349 sound/soc/meson/axg-card.c for (i = 0, tx = 0; i < AXG_TDM_NUM_LANES; i++) { tx 352 sound/soc/meson/axg-card.c tx = max(tx, be->tx_mask[i]); tx 356 sound/soc/meson/axg-card.c if (!tx) tx 370 sound/soc/meson/axg-card.c if (!tx && !rx) { tx 381 sound/soc/meson/axg-card.c be->slots = fls(max(tx, rx)); tx 382 sound/soc/meson/axg-card.c } else if (be->slots < fls(max(tx, rx)) || be->slots > 32) { tx 415 sound/soc/meson/axg-card.c &codec_mask->tx); tx 40 sound/soc/meson/axg-tdm-interface.c struct axg_tdm_stream *tx = (struct axg_tdm_stream *) tx 83 sound/soc/meson/axg-tdm-interface.c if (tx) { tx 84 sound/soc/meson/axg-tdm-interface.c tx->mask = tx_mask; tx 16 sound/soc/samsung/dma.h const char *tx, const char *rx, tx 18 sound/soc/samsung/dmaengine.c const char *tx, const char *rx, tx 31 sound/soc/samsung/dmaengine.c pcm_conf->chan_names[SNDRV_PCM_STREAM_PLAYBACK] = tx; tx 419 sound/soc/ti/omap-mcbsp.c int tx = (stream == SNDRV_PCM_STREAM_PLAYBACK); tx 420 sound/soc/ti/omap-mcbsp.c int rx = !tx; tx 440 sound/soc/ti/omap-mcbsp.c tx &= 1; tx 442 sound/soc/ti/omap-mcbsp.c MCBSP_WRITE(mcbsp, SPCR2, w | tx); tx 465 sound/soc/ti/omap-mcbsp.c w &= ~(tx ? XDISABLE : 0); tx 478 sound/soc/ti/omap-mcbsp.c int tx = (stream == SNDRV_PCM_STREAM_PLAYBACK); tx 479 sound/soc/ti/omap-mcbsp.c int rx = !tx; tx 484 sound/soc/ti/omap-mcbsp.c tx &= 1; tx 487 sound/soc/ti/omap-mcbsp.c w |= (tx ? XDISABLE : 0); tx 491 sound/soc/ti/omap-mcbsp.c MCBSP_WRITE(mcbsp, SPCR2, w & ~tx); tx 834 sound/soc/ti/omap-mcbsp.c int tx = (substream->stream == SNDRV_PCM_STREAM_PLAYBACK); tx 835 sound/soc/ti/omap-mcbsp.c int stream1 = tx ? SNDRV_PCM_STREAM_PLAYBACK : SNDRV_PCM_STREAM_CAPTURE; tx 836 sound/soc/ti/omap-mcbsp.c int stream2 = tx ? SNDRV_PCM_STREAM_CAPTURE : SNDRV_PCM_STREAM_PLAYBACK; tx 857 sound/soc/ti/omap-mcbsp.c int tx = (substream->stream == SNDRV_PCM_STREAM_PLAYBACK); tx 858 sound/soc/ti/omap-mcbsp.c int stream1 = tx ? SNDRV_PCM_STREAM_PLAYBACK : SNDRV_PCM_STREAM_CAPTURE; tx 859 sound/soc/ti/omap-mcbsp.c int stream2 = tx ? SNDRV_PCM_STREAM_CAPTURE : SNDRV_PCM_STREAM_PLAYBACK; tx 268 sound/soc/ti/omap-mcpdm.c int tx = (substream->stream == SNDRV_PCM_STREAM_PLAYBACK); tx 269 sound/soc/ti/omap-mcpdm.c int stream1 = tx ? SNDRV_PCM_STREAM_PLAYBACK : SNDRV_PCM_STREAM_CAPTURE; tx 270 sound/soc/ti/omap-mcpdm.c int stream2 = tx ? SNDRV_PCM_STREAM_CAPTURE : SNDRV_PCM_STREAM_PLAYBACK; tx 380 sound/soc/ti/omap-mcpdm.c int tx = (substream->stream == SNDRV_PCM_STREAM_PLAYBACK); tx 381 sound/soc/ti/omap-mcpdm.c int stream1 = tx ? SNDRV_PCM_STREAM_PLAYBACK : SNDRV_PCM_STREAM_CAPTURE; tx 382 sound/soc/ti/omap-mcpdm.c int stream2 = tx ? SNDRV_PCM_STREAM_CAPTURE : SNDRV_PCM_STREAM_PLAYBACK; tx 52 tools/include/uapi/linux/if_xdp.h struct xdp_ring_offset tx; tx 58 tools/lib/bpf/xsk.c struct xsk_ring_prod *tx; tx 86 tools/lib/bpf/xsk.c struct xdp_ring_offset_v1 tx; tx 166 tools/lib/bpf/xsk.c off->tx.producer = off_v1.tx.producer; tx 167 tools/lib/bpf/xsk.c off->tx.consumer = off_v1.tx.consumer; tx 168 tools/lib/bpf/xsk.c off->tx.desc = off_v1.tx.desc; tx 169 tools/lib/bpf/xsk.c off->tx.flags = off_v1.tx.consumer + sizeof(__u32); tx 560 tools/lib/bpf/xsk.c struct xsk_ring_cons *rx, struct xsk_ring_prod *tx, tx 569 tools/lib/bpf/xsk.c if (!umem || !xsk_ptr || !rx || !tx) tx 615 tools/lib/bpf/xsk.c if (tx) { tx 650 tools/lib/bpf/xsk.c if (tx) { tx 651 tools/lib/bpf/xsk.c tx_map = mmap(NULL, off.tx.desc + tx 660 tools/lib/bpf/xsk.c tx->mask = xsk->config.tx_size - 1; tx 661 tools/lib/bpf/xsk.c tx->size = xsk->config.tx_size; tx 662 tools/lib/bpf/xsk.c tx->producer = tx_map + off.tx.producer; tx 663 tools/lib/bpf/xsk.c tx->consumer = tx_map + off.tx.consumer; tx 664 tools/lib/bpf/xsk.c tx->flags = tx_map + off.tx.flags; tx 665 tools/lib/bpf/xsk.c tx->ring = tx_map + off.tx.desc; tx 666 tools/lib/bpf/xsk.c tx->cached_cons = xsk->config.tx_size; tx 668 tools/lib/bpf/xsk.c xsk->tx = tx; tx 693 tools/lib/bpf/xsk.c if (tx) tx 694 tools/lib/bpf/xsk.c munmap(tx_map, off.tx.desc + tx 753 tools/lib/bpf/xsk.c if (xsk->tx) { tx 754 tools/lib/bpf/xsk.c munmap(xsk->tx->ring - off.tx.desc, tx 755 tools/lib/bpf/xsk.c off.tx.desc + xsk->config.tx_size * desc_sz); tx 64 tools/lib/bpf/xsk.h static inline struct xdp_desc *xsk_ring_prod__tx_desc(struct xsk_ring_prod *tx, tx 67 tools/lib/bpf/xsk.h struct xdp_desc *descs = (struct xdp_desc *)tx->ring; tx 69 tools/lib/bpf/xsk.h return &descs[idx & tx->mask]; tx 235 tools/lib/bpf/xsk.h struct xsk_ring_prod *tx, tx 90 tools/perf/util/stat.c ID(TRANSACTION_START, cpu/tx-start/), tx 113 tools/spi/spidev_test.c static void transfer(int fd, uint8_t const *tx, uint8_t const *rx, size_t len) tx 118 tools/spi/spidev_test.c .tx_buf = (unsigned long)tx, tx 146 tools/spi/spidev_test.c hex_dump(tx, len, 32, "TX"); tx 300 tools/spi/spidev_test.c uint8_t *tx; tx 303 tools/spi/spidev_test.c tx = malloc(size); tx 304 tools/spi/spidev_test.c if (!tx) tx 311 tools/spi/spidev_test.c size = unescape((char *)tx, str, size); tx 312 tools/spi/spidev_test.c transfer(fd, tx, rx, size); tx 314 tools/spi/spidev_test.c free(tx); tx 322 tools/spi/spidev_test.c uint8_t *tx; tx 332 tools/spi/spidev_test.c tx = malloc(sb.st_size); tx 333 tools/spi/spidev_test.c if (!tx) tx 340 tools/spi/spidev_test.c bytes = read(tx_fd, tx, sb.st_size); tx 344 tools/spi/spidev_test.c transfer(fd, tx, rx, sb.st_size); tx 346 tools/spi/spidev_test.c free(tx); tx 369 tools/spi/spidev_test.c uint8_t *tx; tx 373 tools/spi/spidev_test.c tx = malloc(len); tx 374 tools/spi/spidev_test.c if (!tx) tx 377 tools/spi/spidev_test.c tx[i] = random(); tx 383 tools/spi/spidev_test.c transfer(fd, tx, rx, len); tx 389 tools/spi/spidev_test.c if (memcmp(tx, rx, len)) { tx 391 tools/spi/spidev_test.c hex_dump(tx, len, 32, "TX"); tx 398 tools/spi/spidev_test.c free(tx); tx 514 tools/testing/selftests/bpf/test_flow_dissector.c int fdt = -1, fdr = -1, len, tx = 0, rx = 0; tx 529 tools/testing/selftests/bpf/test_flow_dissector.c tx += do_tx(fdt, buf, len); tx 539 tools/testing/selftests/bpf/test_flow_dissector.c fprintf(stderr, "pkts: tx=%u rx=%u\n", tx, rx); tx 540 tools/testing/selftests/bpf/test_flow_dissector.c tx = 0; tx 545 tools/testing/selftests/bpf/test_flow_dissector.c if (tx == cfg_num_pkt) tx 551 tools/testing/selftests/bpf/test_flow_dissector.c if (rx < tx) { tx 553 tools/testing/selftests/bpf/test_flow_dissector.c while (rx < tx) { tx 563 tools/testing/selftests/bpf/test_flow_dissector.c fprintf(stderr, "pkts: tx=%u rx=%u\n", tx, rx); tx 577 tools/testing/selftests/bpf/test_flow_dissector.c return rx != tx; tx 447 tools/testing/selftests/bpf/test_sockmap.c struct msg_stats *s, bool tx, tx 455 tools/testing/selftests/bpf/test_sockmap.c err = msg_alloc_iov(&msg, iov_count, iov_length, data, tx); tx 459 tools/testing/selftests/bpf/test_sockmap.c err = msg_alloc_iov(&msg_peek, iov_count, iov_length, data, tx); tx 464 tools/testing/selftests/bpf/test_sockmap.c if (tx) { tx 446 tools/testing/selftests/net/psock_tpacket.c struct tpacket3_hdr *tx = next; tx 448 tools/testing/selftests/net/psock_tpacket.c tx->tp_snaplen = packet_len; tx 449 tools/testing/selftests/net/psock_tpacket.c tx->tp_len = packet_len; tx 450 tools/testing/selftests/net/psock_tpacket.c tx->tp_next_offset = 0; tx 452 tools/testing/selftests/net/psock_tpacket.c memcpy((uint8_t *)tx + TPACKET3_HDRLEN - tx 455 tools/testing/selftests/net/psock_tpacket.c total_bytes += tx->tp_snaplen; tx 41 tools/testing/selftests/ptp/testptp.c static int clock_adjtime(clockid_t id, struct timex *tx) tx 43 tools/testing/selftests/ptp/testptp.c return syscall(__NR_clock_adjtime, id, tx); tx 152 tools/testing/selftests/ptp/testptp.c struct timex tx; tx 284 tools/testing/selftests/ptp/testptp.c memset(&tx, 0, sizeof(tx)); tx 285 tools/testing/selftests/ptp/testptp.c tx.modes = ADJ_FREQUENCY; tx 286 tools/testing/selftests/ptp/testptp.c tx.freq = ppb_to_scaled_ppm(adjfreq); tx 287 tools/testing/selftests/ptp/testptp.c if (clock_adjtime(clkid, &tx)) { tx 295 tools/testing/selftests/ptp/testptp.c memset(&tx, 0, sizeof(tx)); tx 296 tools/testing/selftests/ptp/testptp.c tx.modes = ADJ_SETOFFSET; tx 297 tools/testing/selftests/ptp/testptp.c tx.time.tv_sec = adjtime; tx 298 tools/testing/selftests/ptp/testptp.c tx.time.tv_usec = 0; tx 299 tools/testing/selftests/ptp/testptp.c if (clock_adjtime(clkid, &tx) < 0) { tx 38 tools/testing/selftests/timers/change_skew.c struct timex tx; tx 41 tools/testing/selftests/timers/change_skew.c tx.modes = ADJ_FREQUENCY; tx 42 tools/testing/selftests/timers/change_skew.c tx.freq = ppm << 16; tx 44 tools/testing/selftests/timers/change_skew.c ret = adjtimex(&tx); tx 60 tools/testing/selftests/timers/change_skew.c struct timex tx; tx 69 tools/testing/selftests/timers/change_skew.c tx.modes = ADJ_OFFSET; tx 70 tools/testing/selftests/timers/change_skew.c tx.offset = 0; tx 71 tools/testing/selftests/timers/change_skew.c ret = adjtimex(&tx); tx 86 tools/testing/selftests/timers/change_skew.c tx.modes = ADJ_FREQUENCY; tx 87 tools/testing/selftests/timers/change_skew.c tx.offset = 0; tx 88 tools/testing/selftests/timers/change_skew.c adjtimex(&tx); tx 97 tools/testing/selftests/timers/leap-a-day.c struct timex tx; tx 106 tools/testing/selftests/timers/leap-a-day.c tx.modes = ADJ_STATUS; tx 107 tools/testing/selftests/timers/leap-a-day.c tx.status = STA_PLL; tx 108 tools/testing/selftests/timers/leap-a-day.c ret = adjtimex(&tx); tx 111 tools/testing/selftests/timers/leap-a-day.c tx.modes = ADJ_MAXERROR; tx 112 tools/testing/selftests/timers/leap-a-day.c tx.maxerror = 0; tx 113 tools/testing/selftests/timers/leap-a-day.c ret = adjtimex(&tx); tx 116 tools/testing/selftests/timers/leap-a-day.c tx.modes = ADJ_STATUS; tx 117 tools/testing/selftests/timers/leap-a-day.c tx.status = 0; tx 118 tools/testing/selftests/timers/leap-a-day.c ret = adjtimex(&tx); tx 132 tools/testing/selftests/timers/leap-a-day.c struct timex tx; tx 135 tools/testing/selftests/timers/leap-a-day.c tx.modes = 0; tx 136 tools/testing/selftests/timers/leap-a-day.c ret = adjtimex(&tx); tx 138 tools/testing/selftests/timers/leap-a-day.c if (tx.time.tv_sec < next_leap) { tx 142 tools/testing/selftests/timers/leap-a-day.c tx.time.tv_sec, tx 143 tools/testing/selftests/timers/leap-a-day.c tx.time.tv_usec, tx 144 tools/testing/selftests/timers/leap-a-day.c tx.tai, tx 151 tools/testing/selftests/timers/leap-a-day.c tx.time.tv_sec, tx 152 tools/testing/selftests/timers/leap-a-day.c tx.time.tv_usec, tx 153 tools/testing/selftests/timers/leap-a-day.c tx.tai, tx 239 tools/testing/selftests/timers/leap-a-day.c struct timex tx; tx 262 tools/testing/selftests/timers/leap-a-day.c tx.modes = ADJ_STATUS; tx 264 tools/testing/selftests/timers/leap-a-day.c tx.status = STA_INS; tx 266 tools/testing/selftests/timers/leap-a-day.c tx.status = STA_DEL; tx 267 tools/testing/selftests/timers/leap-a-day.c ret = adjtimex(&tx); tx 275 tools/testing/selftests/timers/leap-a-day.c tx.modes = 0; tx 276 tools/testing/selftests/timers/leap-a-day.c ret = adjtimex(&tx); tx 277 tools/testing/selftests/timers/leap-a-day.c if (tx.status != STA_INS && tx.status != STA_DEL) { tx 315 tools/testing/selftests/timers/leap-a-day.c tx.modes = 0; tx 316 tools/testing/selftests/timers/leap-a-day.c ret = adjtimex(&tx); tx 317 tools/testing/selftests/timers/leap-a-day.c if (tx.status != STA_INS && tx.status != STA_DEL) { tx 319 tools/testing/selftests/timers/leap-a-day.c tx.modes = ADJ_STATUS; tx 321 tools/testing/selftests/timers/leap-a-day.c tx.status = STA_INS; tx 323 tools/testing/selftests/timers/leap-a-day.c tx.status = STA_DEL; tx 324 tools/testing/selftests/timers/leap-a-day.c ret = adjtimex(&tx); tx 328 tools/testing/selftests/timers/leap-a-day.c now = tx.time.tv_sec; tx 334 tools/testing/selftests/timers/leap-a-day.c tx.modes = 0; tx 335 tools/testing/selftests/timers/leap-a-day.c ret = adjtimex(&tx); tx 344 tools/testing/selftests/timers/leap-a-day.c ctime_r(&tx.time.tv_sec, buf); tx 349 tools/testing/selftests/timers/leap-a-day.c tx.time.tv_usec, tx 350 tools/testing/selftests/timers/leap-a-day.c tx.tai, tx 353 tools/testing/selftests/timers/leap-a-day.c now = tx.time.tv_sec; tx 30 tools/testing/selftests/timers/leapcrash.c struct timex tx; tx 39 tools/testing/selftests/timers/leapcrash.c tx.modes = ADJ_STATUS; tx 40 tools/testing/selftests/timers/leapcrash.c tx.status = STA_PLL; tx 41 tools/testing/selftests/timers/leapcrash.c ret = adjtimex(&tx); tx 43 tools/testing/selftests/timers/leapcrash.c tx.modes = ADJ_STATUS; tx 44 tools/testing/selftests/timers/leapcrash.c tx.status = 0; tx 45 tools/testing/selftests/timers/leapcrash.c ret = adjtimex(&tx); tx 60 tools/testing/selftests/timers/leapcrash.c struct timex tx; tx 92 tools/testing/selftests/timers/leapcrash.c tx.modes = 0; tx 93 tools/testing/selftests/timers/leapcrash.c adjtimex(&tx); tx 96 tools/testing/selftests/timers/leapcrash.c while (tx.time.tv_sec < next_leap + 1) { tx 98 tools/testing/selftests/timers/leapcrash.c tx.modes = ADJ_STATUS; tx 99 tools/testing/selftests/timers/leapcrash.c tx.status = STA_INS; tx 100 tools/testing/selftests/timers/leapcrash.c adjtimex(&tx); tx 30 tools/testing/selftests/timers/set-tai.c struct timex tx; tx 32 tools/testing/selftests/timers/set-tai.c memset(&tx, 0, sizeof(tx)); tx 34 tools/testing/selftests/timers/set-tai.c tx.modes = ADJ_TAI; tx 35 tools/testing/selftests/timers/set-tai.c tx.constant = offset; tx 37 tools/testing/selftests/timers/set-tai.c return adjtimex(&tx); tx 42 tools/testing/selftests/timers/set-tai.c struct timex tx; tx 44 tools/testing/selftests/timers/set-tai.c memset(&tx, 0, sizeof(tx)); tx 46 tools/testing/selftests/timers/set-tai.c adjtimex(&tx); tx 47 tools/testing/selftests/timers/set-tai.c return tx.tai; tx 43 tools/testing/selftests/timers/skew_consistency.c struct timex tx; tx 59 tools/testing/selftests/timers/skew_consistency.c tx.modes = ADJ_FREQUENCY; tx 60 tools/testing/selftests/timers/skew_consistency.c tx.freq = ppm << 16; tx 61 tools/testing/selftests/timers/skew_consistency.c adjtimex(&tx); tx 66 tools/testing/selftests/timers/skew_consistency.c tx.modes = ADJ_FREQUENCY; tx 67 tools/testing/selftests/timers/skew_consistency.c tx.offset = 0; tx 68 tools/testing/selftests/timers/skew_consistency.c adjtimex(&tx); tx 43 tools/testing/selftests/timers/valid-adjtimex.c static int clock_adjtime(clockid_t id, struct timex *tx) tx 45 tools/testing/selftests/timers/valid-adjtimex.c return syscall(__NR_clock_adjtime, id, tx); tx 52 tools/testing/selftests/timers/valid-adjtimex.c struct timex tx; tx 55 tools/testing/selftests/timers/valid-adjtimex.c tx.modes = ADJ_STATUS; tx 56 tools/testing/selftests/timers/valid-adjtimex.c tx.status = 0; tx 57 tools/testing/selftests/timers/valid-adjtimex.c ret = adjtimex(&tx); tx 116 tools/testing/selftests/timers/valid-adjtimex.c struct timex tx; tx 122 tools/testing/selftests/timers/valid-adjtimex.c memset(&tx, 0, sizeof(struct timex)); tx 128 tools/testing/selftests/timers/valid-adjtimex.c tx.modes = ADJ_FREQUENCY; tx 129 tools/testing/selftests/timers/valid-adjtimex.c tx.freq = valid_freq[i]; tx 131 tools/testing/selftests/timers/valid-adjtimex.c ret = adjtimex(&tx); tx 139 tools/testing/selftests/timers/valid-adjtimex.c tx.modes = 0; tx 140 tools/testing/selftests/timers/valid-adjtimex.c ret = adjtimex(&tx); tx 141 tools/testing/selftests/timers/valid-adjtimex.c if (tx.freq != valid_freq[i]) { tx 143 tools/testing/selftests/timers/valid-adjtimex.c tx.freq, valid_freq[i]); tx 147 tools/testing/selftests/timers/valid-adjtimex.c tx.modes = ADJ_FREQUENCY; tx 148 tools/testing/selftests/timers/valid-adjtimex.c tx.freq = outofrange_freq[i]; tx 150 tools/testing/selftests/timers/valid-adjtimex.c ret = adjtimex(&tx); tx 158 tools/testing/selftests/timers/valid-adjtimex.c tx.modes = 0; tx 159 tools/testing/selftests/timers/valid-adjtimex.c ret = adjtimex(&tx); tx 160 tools/testing/selftests/timers/valid-adjtimex.c if (tx.freq == outofrange_freq[i]) { tx 163 tools/testing/selftests/timers/valid-adjtimex.c tx.freq); tx 172 tools/testing/selftests/timers/valid-adjtimex.c tx.modes = ADJ_FREQUENCY; tx 173 tools/testing/selftests/timers/valid-adjtimex.c tx.freq = invalid_freq[i]; tx 174 tools/testing/selftests/timers/valid-adjtimex.c ret = adjtimex(&tx); tx 188 tools/testing/selftests/timers/valid-adjtimex.c tx.modes = ADJ_FREQUENCY; tx 189 tools/testing/selftests/timers/valid-adjtimex.c tx.freq = 0; tx 190 tools/testing/selftests/timers/valid-adjtimex.c ret = adjtimex(&tx);