tx_cfg            390 drivers/hsi/clients/hsi_char.c 	tmp = cl->tx_cfg;
tx_cfg            391 drivers/hsi/clients/hsi_char.c 	cl->tx_cfg.mode = txc->mode;
tx_cfg            392 drivers/hsi/clients/hsi_char.c 	cl->tx_cfg.num_hw_channels = txc->channels;
tx_cfg            393 drivers/hsi/clients/hsi_char.c 	cl->tx_cfg.speed = txc->speed;
tx_cfg            394 drivers/hsi/clients/hsi_char.c 	cl->tx_cfg.arb_mode = txc->arb_mode;
tx_cfg            397 drivers/hsi/clients/hsi_char.c 		cl->tx_cfg = tmp;
tx_cfg            406 drivers/hsi/clients/hsi_char.c 	txc->mode = cl->tx_cfg.mode;
tx_cfg            407 drivers/hsi/clients/hsi_char.c 	txc->channels = cl->tx_cfg.num_hw_channels;
tx_cfg            408 drivers/hsi/clients/hsi_char.c 	txc->speed = cl->tx_cfg.speed;
tx_cfg            409 drivers/hsi/clients/hsi_char.c 	txc->arb_mode = cl->tx_cfg.arb_mode;
tx_cfg            482 drivers/hsi/clients/hsi_char.c 	if (channel->ch >= channel->cl->tx_cfg.num_hw_channels)
tx_cfg            178 drivers/hsi/clients/nokia-modem.c 	ssip.tx_cfg = cl->tx_cfg;
tx_cfg            201 drivers/hsi/clients/nokia-modem.c 	cmtspeech.tx_cfg = cl->tx_cfg;
tx_cfg            468 drivers/hsi/controllers/omap_ssi_port.c 	if (cl->tx_cfg.speed)
tx_cfg            469 drivers/hsi/controllers/omap_ssi_port.c 		omap_ssi->max_speed = cl->tx_cfg.speed;
tx_cfg            473 drivers/hsi/controllers/omap_ssi_port.c 						cl->tx_cfg.speed, div);
tx_cfg            485 drivers/hsi/controllers/omap_ssi_port.c 	writel_relaxed(cl->tx_cfg.num_hw_channels, sst + SSI_SST_CHANNELS_REG);
tx_cfg            486 drivers/hsi/controllers/omap_ssi_port.c 	writel_relaxed(cl->tx_cfg.arb_mode, sst + SSI_SST_ARBMODE_REG);
tx_cfg            487 drivers/hsi/controllers/omap_ssi_port.c 	writel_relaxed(cl->tx_cfg.mode, sst + SSI_SST_MODE_REG);
tx_cfg            498 drivers/hsi/controllers/omap_ssi_port.c 				  cl->tx_cfg.num_hw_channels);
tx_cfg            503 drivers/hsi/controllers/omap_ssi_port.c 	omap_port->sst.channels = cl->tx_cfg.num_hw_channels;
tx_cfg            504 drivers/hsi/controllers/omap_ssi_port.c 	omap_port->sst.arb_mode = cl->tx_cfg.arb_mode;
tx_cfg            505 drivers/hsi/controllers/omap_ssi_port.c 	omap_port->sst.mode = cl->tx_cfg.mode;
tx_cfg             62 drivers/hsi/hsi_core.c 	kfree(cl->tx_cfg.channels);
tx_cfg             77 drivers/hsi/hsi_core.c 	cl->tx_cfg = info->tx_cfg;
tx_cfg             78 drivers/hsi/hsi_core.c 	if (cl->tx_cfg.channels) {
tx_cfg             79 drivers/hsi/hsi_core.c 		size = cl->tx_cfg.num_channels * sizeof(*cl->tx_cfg.channels);
tx_cfg             80 drivers/hsi/hsi_core.c 		cl->tx_cfg.channels = kmemdup(info->tx_cfg.channels, size,
tx_cfg             82 drivers/hsi/hsi_core.c 		if (!cl->tx_cfg.channels)
tx_cfg            109 drivers/hsi/hsi_core.c 	kfree(cl->tx_cfg.channels);
tx_cfg            223 drivers/hsi/hsi_core.c 						 &cl->tx_cfg.mode);
tx_cfg            228 drivers/hsi/hsi_core.c 		cl->tx_cfg.mode = mode;
tx_cfg            232 drivers/hsi/hsi_core.c 				   &cl->tx_cfg.speed);
tx_cfg            235 drivers/hsi/hsi_core.c 	cl->rx_cfg.speed = cl->tx_cfg.speed;
tx_cfg            256 drivers/hsi/hsi_core.c 	cl->tx_cfg.num_channels = cells;
tx_cfg            263 drivers/hsi/hsi_core.c 	cl->tx_cfg.channels = kcalloc(cells, sizeof(channel), GFP_KERNEL);
tx_cfg            264 drivers/hsi/hsi_core.c 	if (!cl->tx_cfg.channels) {
tx_cfg            285 drivers/hsi/hsi_core.c 		cl->tx_cfg.channels[i] = channel;
tx_cfg            289 drivers/hsi/hsi_core.c 	cl->tx_cfg.num_hw_channels = max_chan + 1;
tx_cfg            304 drivers/hsi/hsi_core.c 	kfree(cl->tx_cfg.channels);
tx_cfg            428 drivers/net/ethernet/brocade/bna/bfi_enet.h 	struct bfi_enet_tx_cfg	tx_cfg;
tx_cfg            309 drivers/net/ethernet/brocade/bna/bna.h 			       struct bna_tx_config *tx_cfg,
tx_cfg           3124 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	cfg_req->tx_cfg.vlan_mode = BFI_ENET_TX_VLAN_WI;
tx_cfg           3125 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	cfg_req->tx_cfg.vlan_id = htons((u16)tx->txf_vlan_id);
tx_cfg           3126 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	cfg_req->tx_cfg.admit_tagged_frame = BNA_STATUS_T_ENABLED;
tx_cfg           3127 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	cfg_req->tx_cfg.apply_vlan_filter = BNA_STATUS_T_DISABLED;
tx_cfg           3364 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		struct bna_tx_config *tx_cfg,
tx_cfg           3383 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	if ((intr_info->num != 1) && (intr_info->num != tx_cfg->num_txq))
tx_cfg           3388 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	tx = bna_tx_get(tx_mod, tx_cfg->tx_type);
tx_cfg           3397 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	for (i = 0; i < tx_cfg->num_txq; i++) {
tx_cfg           3421 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	tx->num_txq = tx_cfg->num_txq;
tx_cfg           3460 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		txq->ib.coalescing_timeo = tx_cfg->coalescing_timeo;
tx_cfg           3466 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		txq->tcb->q_depth = tx_cfg->txq_depth;
tx_cfg           3489 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		if (tx_cfg->num_txq == BFI_TX_MAX_PRIO)
tx_cfg            187 drivers/net/ethernet/google/gve/gve.h 	struct gve_queue_config tx_cfg;
tx_cfg            358 drivers/net/ethernet/google/gve/gve.h 	return priv->tx_cfg.num_queues;
tx_cfg             61 drivers/net/ethernet/google/gve/gve_ethtool.c 	for (i = 0; i < priv->tx_cfg.num_queues; i++) {
tx_cfg             83 drivers/net/ethernet/google/gve/gve_ethtool.c 		       (priv->tx_cfg.num_queues * NUM_GVE_TX_CNTS);
tx_cfg            114 drivers/net/ethernet/google/gve/gve_ethtool.c 	     ring < priv->tx_cfg.num_queues; ring++) {
tx_cfg            149 drivers/net/ethernet/google/gve/gve_ethtool.c 		for (ring = 0; ring < priv->tx_cfg.num_queues; ring++) {
tx_cfg            160 drivers/net/ethernet/google/gve/gve_ethtool.c 		i += priv->tx_cfg.num_queues * NUM_GVE_TX_CNTS;
tx_cfg            170 drivers/net/ethernet/google/gve/gve_ethtool.c 	cmd->max_tx = priv->tx_cfg.max_queues;
tx_cfg            174 drivers/net/ethernet/google/gve/gve_ethtool.c 	cmd->tx_count = priv->tx_cfg.num_queues;
tx_cfg            183 drivers/net/ethernet/google/gve/gve_ethtool.c 	struct gve_queue_config new_tx_cfg = priv->tx_cfg;
tx_cfg            199 drivers/net/ethernet/google/gve/gve_ethtool.c 		priv->tx_cfg.num_queues = new_tx;
tx_cfg             47 drivers/net/ethernet/google/gve/gve_main.c 		for (ring = 0; ring < priv->tx_cfg.num_queues; ring++) {
tx_cfg            164 drivers/net/ethernet/google/gve/gve_main.c 		priv->tx_cfg.max_queues = min_t(int, priv->tx_cfg.max_queues,
tx_cfg            170 drivers/net/ethernet/google/gve/gve_main.c 			vecs_enabled, priv->tx_cfg.max_queues,
tx_cfg            172 drivers/net/ethernet/google/gve/gve_main.c 		if (priv->tx_cfg.num_queues > priv->tx_cfg.max_queues)
tx_cfg            173 drivers/net/ethernet/google/gve/gve_main.c 			priv->tx_cfg.num_queues = priv->tx_cfg.max_queues;
tx_cfg            374 drivers/net/ethernet/google/gve/gve_main.c 	for (i = 0; i < priv->tx_cfg.num_queues; i++) {
tx_cfg            415 drivers/net/ethernet/google/gve/gve_main.c 	priv->tx = kvzalloc(priv->tx_cfg.num_queues * sizeof(*priv->tx),
tx_cfg            433 drivers/net/ethernet/google/gve/gve_main.c 	for (i = 0; i < priv->tx_cfg.num_queues; i++) {
tx_cfg            463 drivers/net/ethernet/google/gve/gve_main.c 	for (i = 0; i < priv->tx_cfg.num_queues; i++) {
tx_cfg            498 drivers/net/ethernet/google/gve/gve_main.c 		for (i = 0; i < priv->tx_cfg.num_queues; i++) {
tx_cfg            683 drivers/net/ethernet/google/gve/gve_main.c 	err = netif_set_real_num_tx_queues(dev, priv->tx_cfg.num_queues);
tx_cfg            771 drivers/net/ethernet/google/gve/gve_main.c 		priv->tx_cfg = new_tx_config;
tx_cfg            781 drivers/net/ethernet/google/gve/gve_main.c 	priv->tx_cfg = new_tx_config;
tx_cfg            803 drivers/net/ethernet/google/gve/gve_main.c 	for (idx = 0; idx < priv->tx_cfg.num_queues; idx++) {
tx_cfg            830 drivers/net/ethernet/google/gve/gve_main.c 	for (idx = 0; idx < priv->tx_cfg.num_queues; idx++) {
tx_cfg            953 drivers/net/ethernet/google/gve/gve_main.c 	priv->tx_cfg.max_queues =
tx_cfg            954 drivers/net/ethernet/google/gve/gve_main.c 		min_t(int, priv->tx_cfg.max_queues, priv->num_ntfy_blks / 2);
tx_cfg            958 drivers/net/ethernet/google/gve/gve_main.c 	priv->tx_cfg.num_queues = priv->tx_cfg.max_queues;
tx_cfg            961 drivers/net/ethernet/google/gve/gve_main.c 		priv->tx_cfg.num_queues = min_t(int, priv->default_num_queues,
tx_cfg            962 drivers/net/ethernet/google/gve/gve_main.c 						priv->tx_cfg.num_queues);
tx_cfg            968 drivers/net/ethernet/google/gve/gve_main.c 		   priv->tx_cfg.num_queues, priv->rx_cfg.num_queues);
tx_cfg            970 drivers/net/ethernet/google/gve/gve_main.c 		   priv->tx_cfg.max_queues, priv->rx_cfg.max_queues);
tx_cfg           1161 drivers/net/ethernet/google/gve/gve_main.c 	priv->tx_cfg.max_queues = max_tx_queues;
tx_cfg            246 drivers/net/ethernet/google/gve/gve_tx.c 	for (i = 0; i < priv->tx_cfg.num_queues; i++) {
tx_cfg            269 drivers/net/ethernet/google/gve/gve_tx.c 	for (i = 0; i < priv->tx_cfg.num_queues; i++)
tx_cfg            481 drivers/net/ethernet/google/gve/gve_tx.c 	WARN(skb_get_queue_mapping(skb) > priv->tx_cfg.num_queues,
tx_cfg           1273 drivers/net/ethernet/micrel/ksz884x.c 	u32 tx_cfg;
tx_cfg           3207 drivers/net/ethernet/micrel/ksz884x.c 	u32 tx_cfg;
tx_cfg           3210 drivers/net/ethernet/micrel/ksz884x.c 	tx_cfg = hw->tx_cfg;
tx_cfg           3216 drivers/net/ethernet/micrel/ksz884x.c 		hw->tx_cfg |= DMA_TX_FLOW_ENABLE;
tx_cfg           3218 drivers/net/ethernet/micrel/ksz884x.c 		hw->tx_cfg &= ~DMA_TX_FLOW_ENABLE;
tx_cfg           3222 drivers/net/ethernet/micrel/ksz884x.c 		if (tx_cfg != hw->tx_cfg)
tx_cfg           3223 drivers/net/ethernet/micrel/ksz884x.c 			writel(hw->tx_cfg, hw->io + KS_DMA_TX_CTRL);
tx_cfg           3260 drivers/net/ethernet/micrel/ksz884x.c 		u32 cfg = hw->tx_cfg;
tx_cfg           3264 drivers/net/ethernet/micrel/ksz884x.c 			hw->tx_cfg &= ~DMA_TX_FLOW_ENABLE;
tx_cfg           3265 drivers/net/ethernet/micrel/ksz884x.c 		if (hw->enabled && cfg != hw->tx_cfg)
tx_cfg           3266 drivers/net/ethernet/micrel/ksz884x.c 			writel(hw->tx_cfg, hw->io + KS_DMA_TX_CTRL);
tx_cfg           3772 drivers/net/ethernet/micrel/ksz884x.c 	hw->tx_cfg = (DMA_TX_PAD_ENABLE | DMA_TX_CRC_ENABLE |
tx_cfg           3920 drivers/net/ethernet/micrel/ksz884x.c 	writel(hw->tx_cfg, hw->io + KS_DMA_TX_CTRL);
tx_cfg           3931 drivers/net/ethernet/micrel/ksz884x.c 	writel((hw->tx_cfg & ~DMA_TX_ENABLE), hw->io + KS_DMA_TX_CTRL);
tx_cfg           6337 drivers/net/ethernet/micrel/ksz884x.c 			(hw->tx_cfg & DMA_TX_FLOW_ENABLE) ? 1 : 0;
tx_cfg            581 drivers/net/ethernet/neterion/s2io.c 		struct tx_fifo_config *tx_cfg = &config->tx_cfg[i];
tx_cfg            583 drivers/net/ethernet/neterion/s2io.c 		size += tx_cfg->fifo_len;
tx_cfg            594 drivers/net/ethernet/neterion/s2io.c 		struct tx_fifo_config *tx_cfg = &config->tx_cfg[i];
tx_cfg            596 drivers/net/ethernet/neterion/s2io.c 		size = tx_cfg->fifo_len;
tx_cfg            613 drivers/net/ethernet/neterion/s2io.c 		struct tx_fifo_config *tx_cfg = &config->tx_cfg[i];
tx_cfg            614 drivers/net/ethernet/neterion/s2io.c 		int fifo_len = tx_cfg->fifo_len;
tx_cfg            625 drivers/net/ethernet/neterion/s2io.c 		int page_num = TXD_MEM_PAGE_CNT(config->tx_cfg[i].fifo_len,
tx_cfg            628 drivers/net/ethernet/neterion/s2io.c 		struct tx_fifo_config *tx_cfg = &config->tx_cfg[i];
tx_cfg            631 drivers/net/ethernet/neterion/s2io.c 		fifo->tx_curr_put_info.fifo_len = tx_cfg->fifo_len - 1;
tx_cfg            633 drivers/net/ethernet/neterion/s2io.c 		fifo->tx_curr_get_info.fifo_len = tx_cfg->fifo_len - 1;
tx_cfg            672 drivers/net/ethernet/neterion/s2io.c 				if (l == tx_cfg->fifo_len)
tx_cfg            685 drivers/net/ethernet/neterion/s2io.c 		struct tx_fifo_config *tx_cfg = &config->tx_cfg[i];
tx_cfg            687 drivers/net/ethernet/neterion/s2io.c 		size = tx_cfg->fifo_len;
tx_cfg            896 drivers/net/ethernet/neterion/s2io.c 		struct tx_fifo_config *tx_cfg = &config->tx_cfg[i];
tx_cfg            898 drivers/net/ethernet/neterion/s2io.c 		page_num = TXD_MEM_PAGE_CNT(tx_cfg->fifo_len, lst_per_page);
tx_cfg            928 drivers/net/ethernet/neterion/s2io.c 		swstats->mem_freed += tx_cfg->fifo_len *
tx_cfg            985 drivers/net/ethernet/neterion/s2io.c 		struct tx_fifo_config *tx_cfg = &config->tx_cfg[i];
tx_cfg            988 drivers/net/ethernet/neterion/s2io.c 			swstats->mem_freed += tx_cfg->fifo_len *
tx_cfg           1278 drivers/net/ethernet/neterion/s2io.c 		struct tx_fifo_config *tx_cfg = &config->tx_cfg[i];
tx_cfg           1280 drivers/net/ethernet/neterion/s2io.c 		val64 |= vBIT(tx_cfg->fifo_len - 1, ((j * 32) + 19), 13) |
tx_cfg           1281 drivers/net/ethernet/neterion/s2io.c 			vBIT(tx_cfg->fifo_priority, ((j * 32) + 5), 3);
tx_cfg           2368 drivers/net/ethernet/neterion/s2io.c 		struct tx_fifo_config *tx_cfg = &config->tx_cfg[i];
tx_cfg           2373 drivers/net/ethernet/neterion/s2io.c 		for (j = 0; j < tx_cfg->fifo_len; j++) {
tx_cfg           5474 drivers/net/ethernet/neterion/s2io.c 		tx_desc_count += sp->config.tx_cfg[i].fifo_len;
tx_cfg           7776 drivers/net/ethernet/neterion/s2io.c 		struct tx_fifo_config *tx_cfg = &config->tx_cfg[i];
tx_cfg           7778 drivers/net/ethernet/neterion/s2io.c 		tx_cfg->fifo_len = tx_fifo_len[i];
tx_cfg           7779 drivers/net/ethernet/neterion/s2io.c 		tx_cfg->fifo_priority = i;
tx_cfg           7793 drivers/net/ethernet/neterion/s2io.c 		struct tx_fifo_config *tx_cfg = &config->tx_cfg[i];
tx_cfg           7795 drivers/net/ethernet/neterion/s2io.c 		tx_cfg->f_no_snoop = (NO_SNOOP_TXD | NO_SNOOP_TXD_BUFFER);
tx_cfg           7796 drivers/net/ethernet/neterion/s2io.c 		if (tx_cfg->fifo_len < 65) {
tx_cfg            436 drivers/net/ethernet/neterion/s2io.h 	struct tx_fifo_config tx_cfg[MAX_TX_FIFOS];	/*Per-Tx FIFO config */
tx_cfg            155 drivers/net/ethernet/samsung/sxgbe/sxgbe_core.c 	u32 tx_cfg = readl(ioaddr + SXGBE_CORE_TX_CONFIG_REG);
tx_cfg            158 drivers/net/ethernet/samsung/sxgbe/sxgbe_core.c 	tx_cfg &= ~0x60000000;
tx_cfg            159 drivers/net/ethernet/samsung/sxgbe/sxgbe_core.c 	tx_cfg |= (speed << SXGBE_SPEED_LSHIFT);
tx_cfg            162 drivers/net/ethernet/samsung/sxgbe/sxgbe_core.c 	writel(tx_cfg, ioaddr + SXGBE_CORE_TX_CONFIG_REG);
tx_cfg           1948 drivers/net/ethernet/sfc/falcon/farch.c 	ef4_oword_t tx_cfg;
tx_cfg           1950 drivers/net/ethernet/sfc/falcon/farch.c 	ef4_reado(efx, &tx_cfg, FR_AZ_TX_CFG);
tx_cfg           1955 drivers/net/ethernet/sfc/falcon/farch.c 			tx_cfg, FRF_CZ_TX_ETH_FILTER_FULL_SEARCH_RANGE,
tx_cfg           1959 drivers/net/ethernet/sfc/falcon/farch.c 			tx_cfg, FRF_CZ_TX_ETH_FILTER_WILD_SEARCH_RANGE,
tx_cfg           1964 drivers/net/ethernet/sfc/falcon/farch.c 	ef4_writeo(efx, &tx_cfg, FR_AZ_TX_CFG);
tx_cfg           2008 drivers/net/ethernet/sfc/farch.c 	efx_oword_t tx_cfg;
tx_cfg           2010 drivers/net/ethernet/sfc/farch.c 	efx_reado(efx, &tx_cfg, FR_AZ_TX_CFG);
tx_cfg           2015 drivers/net/ethernet/sfc/farch.c 			tx_cfg, FRF_CZ_TX_ETH_FILTER_FULL_SEARCH_RANGE,
tx_cfg           2019 drivers/net/ethernet/sfc/farch.c 			tx_cfg, FRF_CZ_TX_ETH_FILTER_WILD_SEARCH_RANGE,
tx_cfg           2024 drivers/net/ethernet/sfc/farch.c 	efx_writeo(efx, &tx_cfg, FR_AZ_TX_CFG);
tx_cfg            393 drivers/net/ethernet/sun/niu.c 	u32 tx_cfg, rx_cfg;
tx_cfg            396 drivers/net/ethernet/sun/niu.c 	tx_cfg = (PLL_TX_CFG_ENTX | PLL_TX_CFG_SWING_1375MV);
tx_cfg            407 drivers/net/ethernet/sun/niu.c 		tx_cfg |= PLL_TX_CFG_ENTEST;
tx_cfg            413 drivers/net/ethernet/sun/niu.c 		int err = esr2_set_tx_cfg(np, i, tx_cfg);
tx_cfg            433 drivers/net/ethernet/sun/niu.c 	u32 tx_cfg, rx_cfg;
tx_cfg            437 drivers/net/ethernet/sun/niu.c 	tx_cfg = (PLL_TX_CFG_ENTX | PLL_TX_CFG_SWING_1375MV |
tx_cfg            452 drivers/net/ethernet/sun/niu.c 		tx_cfg |= PLL_TX_CFG_ENTEST;
tx_cfg            481 drivers/net/ethernet/sun/niu.c 		err = esr2_set_tx_cfg(np, i, tx_cfg);
tx_cfg            527 drivers/net/ethernet/sun/niu.c 	u32 tx_cfg, rx_cfg, pll_cfg, pll_sts;
tx_cfg            533 drivers/net/ethernet/sun/niu.c 	tx_cfg = (PLL_TX_CFG_ENTX | PLL_TX_CFG_SWING_1375MV);
tx_cfg            544 drivers/net/ethernet/sun/niu.c 		tx_cfg |= PLL_TX_CFG_ENTEST;
tx_cfg            573 drivers/net/ethernet/sun/niu.c 		err = esr2_set_tx_cfg(np, i, tx_cfg);
tx_cfg           2272 drivers/net/wireless/marvell/mwifiex/sta_cmd.c 	struct mwifiex_ds_11n_tx_cfg tx_cfg;
tx_cfg           2432 drivers/net/wireless/marvell/mwifiex/sta_cmd.c 	tx_cfg.tx_htcap = MWIFIEX_FW_DEF_HTTXCFG;
tx_cfg           2434 drivers/net/wireless/marvell/mwifiex/sta_cmd.c 			       HostCmd_ACT_GEN_SET, 0, &tx_cfg, true);
tx_cfg            380 drivers/soc/qcom/qcom-geni-se.c 			    bool msb_to_lsb, bool tx_cfg, bool rx_cfg)
tx_cfg            413 drivers/soc/qcom/qcom-geni-se.c 	if (tx_cfg) {
tx_cfg            110 drivers/staging/pi433/pi433_if.c 	struct pi433_tx_cfg	tx_cfg;
tx_cfg            283 drivers/staging/pi433/pi433_if.c rf69_set_tx_cfg(struct pi433_device *dev, struct pi433_tx_cfg *tx_cfg)
tx_cfg            287 drivers/staging/pi433/pi433_if.c 	ret = rf69_set_frequency(dev->spi, tx_cfg->frequency);
tx_cfg            290 drivers/staging/pi433/pi433_if.c 	ret = rf69_set_bit_rate(dev->spi, tx_cfg->bit_rate);
tx_cfg            293 drivers/staging/pi433/pi433_if.c 	ret = rf69_set_modulation(dev->spi, tx_cfg->modulation);
tx_cfg            296 drivers/staging/pi433/pi433_if.c 	ret = rf69_set_deviation(dev->spi, tx_cfg->dev_frequency);
tx_cfg            299 drivers/staging/pi433/pi433_if.c 	ret = rf69_set_pa_ramp(dev->spi, tx_cfg->pa_ramp);
tx_cfg            302 drivers/staging/pi433/pi433_if.c 	ret = rf69_set_modulation_shaping(dev->spi, tx_cfg->mod_shaping);
tx_cfg            305 drivers/staging/pi433/pi433_if.c 	ret = rf69_set_tx_start_condition(dev->spi, tx_cfg->tx_start_condition);
tx_cfg            310 drivers/staging/pi433/pi433_if.c 	if (tx_cfg->enable_preamble == OPTION_ON) {
tx_cfg            312 drivers/staging/pi433/pi433_if.c 					       tx_cfg->preamble_length);
tx_cfg            321 drivers/staging/pi433/pi433_if.c 	if (tx_cfg->enable_sync == OPTION_ON) {
tx_cfg            322 drivers/staging/pi433/pi433_if.c 		ret = rf69_set_sync_size(dev->spi, tx_cfg->sync_length);
tx_cfg            325 drivers/staging/pi433/pi433_if.c 		ret = rf69_set_sync_values(dev->spi, tx_cfg->sync_pattern);
tx_cfg            337 drivers/staging/pi433/pi433_if.c 	if (tx_cfg->enable_length_byte == OPTION_ON) {
tx_cfg            347 drivers/staging/pi433/pi433_if.c 	if (tx_cfg->enable_crc == OPTION_ON) {
tx_cfg            560 drivers/staging/pi433/pi433_if.c 	struct pi433_tx_cfg tx_cfg;
tx_cfg            581 drivers/staging/pi433/pi433_if.c 		retval = kfifo_out(&device->tx_fifo, &tx_cfg, sizeof(tx_cfg));
tx_cfg            582 drivers/staging/pi433/pi433_if.c 		if (retval != sizeof(tx_cfg)) {
tx_cfg            585 drivers/staging/pi433/pi433_if.c 				retval, (unsigned int)sizeof(tx_cfg));
tx_cfg            598 drivers/staging/pi433/pi433_if.c 		if (tx_cfg.fixed_message_length != 0)
tx_cfg            599 drivers/staging/pi433/pi433_if.c 			size = tx_cfg.fixed_message_length;
tx_cfg            602 drivers/staging/pi433/pi433_if.c 		if (tx_cfg.enable_length_byte == OPTION_ON)
tx_cfg            606 drivers/staging/pi433/pi433_if.c 		if (tx_cfg.enable_address_byte == OPTION_ON)
tx_cfg            614 drivers/staging/pi433/pi433_if.c 		if (tx_cfg.enable_length_byte  == OPTION_ON)
tx_cfg            622 drivers/staging/pi433/pi433_if.c 		if (tx_cfg.enable_address_byte == OPTION_ON)
tx_cfg            623 drivers/staging/pi433/pi433_if.c 			device->buffer[position++] = tx_cfg.address_byte;
tx_cfg            665 drivers/staging/pi433/pi433_if.c 		if (tx_cfg.enable_length_byte == OPTION_ON) {
tx_cfg            666 drivers/staging/pi433/pi433_if.c 			retval = rf69_set_payload_length(spi, size * tx_cfg.repetitions);
tx_cfg            676 drivers/staging/pi433/pi433_if.c 		retval = rf69_set_tx_cfg(device, &tx_cfg);
tx_cfg            703 drivers/staging/pi433/pi433_if.c 		repetitions = tx_cfg.repetitions;
tx_cfg            834 drivers/staging/pi433/pi433_if.c 	required = sizeof(instance->tx_cfg) + sizeof(size_t) + count;
tx_cfg            843 drivers/staging/pi433/pi433_if.c 	retval = kfifo_in(&device->tx_fifo, &instance->tx_cfg,
tx_cfg            844 drivers/staging/pi433/pi433_if.c 			  sizeof(instance->tx_cfg));
tx_cfg            845 drivers/staging/pi433/pi433_if.c 	if (retval != sizeof(instance->tx_cfg))
tx_cfg            876 drivers/staging/pi433/pi433_if.c 	struct pi433_tx_cfg	tx_cfg;
tx_cfg            891 drivers/staging/pi433/pi433_if.c 		if (copy_to_user(argp, &instance->tx_cfg,
tx_cfg            896 drivers/staging/pi433/pi433_if.c 		if (copy_from_user(&tx_cfg, argp, sizeof(struct pi433_tx_cfg)))
tx_cfg            899 drivers/staging/pi433/pi433_if.c 		memcpy(&instance->tx_cfg, &tx_cfg, sizeof(struct pi433_tx_cfg));
tx_cfg            962 drivers/staging/pi433/pi433_if.c 	instance->tx_cfg.bit_rate = 4711;
tx_cfg            470 drivers/usb/gadget/udc/bcm63xx_udc.c 		const struct iudma_ch_cfg *tx_cfg = &iudma_defaults[i + 1];
tx_cfg            484 drivers/usb/gadget/udc/bcm63xx_udc.c 			((tx_fifo_slot + tx_cfg->n_fifo_slots - 1) <<
tx_cfg            486 drivers/usb/gadget/udc/bcm63xx_udc.c 		tx_fifo_slot += tx_cfg->n_fifo_slots;
tx_cfg            489 drivers/usb/gadget/udc/bcm63xx_udc.c 			    is_hs ? tx_cfg->max_pkt_hs : tx_cfg->max_pkt_fs,
tx_cfg            103 include/linux/hsi/hsi.h 	struct hsi_config	tx_cfg;
tx_cfg            128 include/linux/hsi/hsi.h 	struct hsi_config	tx_cfg;
tx_cfg            223 include/linux/hsi/hsi.h 	struct hsi_config		tx_cfg;
tx_cfg            398 include/linux/qcom-geni-se.h 			    bool msb_to_lsb, bool tx_cfg, bool rx_cfg);