rx_cfg            355 drivers/hsi/clients/hsi_char.c 	tmp = cl->rx_cfg;
rx_cfg            356 drivers/hsi/clients/hsi_char.c 	cl->rx_cfg.mode = rxc->mode;
rx_cfg            357 drivers/hsi/clients/hsi_char.c 	cl->rx_cfg.num_hw_channels = rxc->channels;
rx_cfg            358 drivers/hsi/clients/hsi_char.c 	cl->rx_cfg.flow = rxc->flow;
rx_cfg            361 drivers/hsi/clients/hsi_char.c 		cl->rx_cfg = tmp;
rx_cfg            372 drivers/hsi/clients/hsi_char.c 	rxc->mode = cl->rx_cfg.mode;
rx_cfg            373 drivers/hsi/clients/hsi_char.c 	rxc->channels = cl->rx_cfg.num_hw_channels;
rx_cfg            374 drivers/hsi/clients/hsi_char.c 	rxc->flow = cl->rx_cfg.flow;
rx_cfg            425 drivers/hsi/clients/hsi_char.c 	if (channel->ch >= channel->cl->rx_cfg.num_hw_channels)
rx_cfg            179 drivers/hsi/clients/nokia-modem.c 	ssip.rx_cfg = cl->rx_cfg;
rx_cfg            202 drivers/hsi/clients/nokia-modem.c 	cmtspeech.rx_cfg = cl->rx_cfg;
rx_cfg            490 drivers/hsi/controllers/omap_ssi_port.c 	writel_relaxed(cl->rx_cfg.num_hw_channels, ssr + SSI_SSR_CHANNELS_REG);
rx_cfg            494 drivers/hsi/controllers/omap_ssi_port.c 		(cl->rx_cfg.mode != SSI_MODE_FRAME))
rx_cfg            496 drivers/hsi/controllers/omap_ssi_port.c 	writel_relaxed(cl->rx_cfg.mode, ssr + SSI_SSR_MODE_REG);
rx_cfg            497 drivers/hsi/controllers/omap_ssi_port.c 	omap_port->channels = max(cl->rx_cfg.num_hw_channels,
rx_cfg            509 drivers/hsi/controllers/omap_ssi_port.c 	omap_port->ssr.channels = cl->rx_cfg.num_hw_channels;
rx_cfg            510 drivers/hsi/controllers/omap_ssi_port.c 	omap_port->ssr.mode = cl->rx_cfg.mode;
rx_cfg             63 drivers/hsi/hsi_core.c 	kfree(cl->rx_cfg.channels);
rx_cfg             86 drivers/hsi/hsi_core.c 	cl->rx_cfg = info->rx_cfg;
rx_cfg             87 drivers/hsi/hsi_core.c 	if (cl->rx_cfg.channels) {
rx_cfg             88 drivers/hsi/hsi_core.c 		size = cl->rx_cfg.num_channels * sizeof(*cl->rx_cfg.channels);
rx_cfg             89 drivers/hsi/hsi_core.c 		cl->rx_cfg.channels = kmemdup(info->rx_cfg.channels, size,
rx_cfg             91 drivers/hsi/hsi_core.c 		if (!cl->rx_cfg.channels)
rx_cfg            218 drivers/hsi/hsi_core.c 						 &cl->rx_cfg.mode);
rx_cfg            227 drivers/hsi/hsi_core.c 		cl->rx_cfg.mode = mode;
rx_cfg            235 drivers/hsi/hsi_core.c 	cl->rx_cfg.speed = cl->tx_cfg.speed;
rx_cfg            238 drivers/hsi/hsi_core.c 					 &cl->rx_cfg.flow);
rx_cfg            243 drivers/hsi/hsi_core.c 					     &cl->rx_cfg.arb_mode);
rx_cfg            255 drivers/hsi/hsi_core.c 	cl->rx_cfg.num_channels = cells;
rx_cfg            257 drivers/hsi/hsi_core.c 	cl->rx_cfg.channels = kcalloc(cells, sizeof(channel), GFP_KERNEL);
rx_cfg            258 drivers/hsi/hsi_core.c 	if (!cl->rx_cfg.channels) {
rx_cfg            284 drivers/hsi/hsi_core.c 		cl->rx_cfg.channels[i] = channel;
rx_cfg            288 drivers/hsi/hsi_core.c 	cl->rx_cfg.num_hw_channels = max_chan + 1;
rx_cfg            306 drivers/hsi/hsi_core.c 	kfree(cl->rx_cfg.channels);
rx_cfg            742 drivers/hsi/hsi_core.c 	if (!cl->rx_cfg.channels)
rx_cfg            745 drivers/hsi/hsi_core.c 	for (i = 0; i < cl->rx_cfg.num_channels; i++)
rx_cfg            746 drivers/hsi/hsi_core.c 		if (!strcmp(cl->rx_cfg.channels[i].name, name))
rx_cfg            747 drivers/hsi/hsi_core.c 			return cl->rx_cfg.channels[i].id;
rx_cfg            900 drivers/net/ethernet/apple/bmac.c 	unsigned short rx_cfg;
rx_cfg            902 drivers/net/ethernet/apple/bmac.c 	rx_cfg = bmread(dev, RXCFG);
rx_cfg            903 drivers/net/ethernet/apple/bmac.c 	rx_cfg &= ~RxMACEnable;
rx_cfg            904 drivers/net/ethernet/apple/bmac.c 	bmwrite(dev, RXCFG, rx_cfg);
rx_cfg            906 drivers/net/ethernet/apple/bmac.c 		rx_cfg = bmread(dev, RXCFG);
rx_cfg            907 drivers/net/ethernet/apple/bmac.c 	}  while (rx_cfg & RxMACEnable);
rx_cfg            913 drivers/net/ethernet/apple/bmac.c 	unsigned short rx_cfg;
rx_cfg            915 drivers/net/ethernet/apple/bmac.c 	rx_cfg = bmread(dev, RXCFG);
rx_cfg            916 drivers/net/ethernet/apple/bmac.c 	rx_cfg |= RxMACEnable;
rx_cfg            917 drivers/net/ethernet/apple/bmac.c 	if (hash_enable) rx_cfg |= RxHashFilterEnable;
rx_cfg            918 drivers/net/ethernet/apple/bmac.c 	else rx_cfg &= ~RxHashFilterEnable;
rx_cfg            919 drivers/net/ethernet/apple/bmac.c 	if (promisc_enable) rx_cfg |= RxPromiscEnable;
rx_cfg            920 drivers/net/ethernet/apple/bmac.c 	else rx_cfg &= ~RxPromiscEnable;
rx_cfg            924 drivers/net/ethernet/apple/bmac.c 	bmwrite(dev, RXCFG, rx_cfg );
rx_cfg            925 drivers/net/ethernet/apple/bmac.c 	return rx_cfg;
rx_cfg            972 drivers/net/ethernet/apple/bmac.c 	unsigned short rx_cfg;
rx_cfg            983 drivers/net/ethernet/apple/bmac.c 		rx_cfg = bmac_rx_on(dev, 1, 0);
rx_cfg            986 drivers/net/ethernet/apple/bmac.c 		rx_cfg = bmread(dev, RXCFG);
rx_cfg            987 drivers/net/ethernet/apple/bmac.c 		rx_cfg |= RxPromiscEnable;
rx_cfg            988 drivers/net/ethernet/apple/bmac.c 		bmwrite(dev, RXCFG, rx_cfg);
rx_cfg            989 drivers/net/ethernet/apple/bmac.c 		rx_cfg = bmac_rx_on(dev, 0, 1);
rx_cfg            990 drivers/net/ethernet/apple/bmac.c 		XXDEBUG(("bmac: promisc mode enabled, rx_cfg=%#08x\n", rx_cfg));
rx_cfg            995 drivers/net/ethernet/apple/bmac.c 			rx_cfg = bmac_rx_on(dev, 0, 0);
rx_cfg            996 drivers/net/ethernet/apple/bmac.c 			XXDEBUG(("bmac: multi disabled, rx_cfg=%#08x\n", rx_cfg));
rx_cfg           1001 drivers/net/ethernet/apple/bmac.c 			rx_cfg = bmac_rx_on(dev, 1, 0);
rx_cfg           1002 drivers/net/ethernet/apple/bmac.c 			XXDEBUG(("bmac: multi enabled, rx_cfg=%#08x\n", rx_cfg));
rx_cfg           1014 drivers/net/ethernet/apple/bmac.c 	unsigned short rx_cfg;
rx_cfg           1023 drivers/net/ethernet/apple/bmac.c 		rx_cfg = bmread(dev, RXCFG);
rx_cfg           1024 drivers/net/ethernet/apple/bmac.c 		rx_cfg |= RxPromiscEnable;
rx_cfg           1025 drivers/net/ethernet/apple/bmac.c 		bmwrite(dev, RXCFG, rx_cfg);
rx_cfg           1029 drivers/net/ethernet/apple/bmac.c 		rx_cfg = bmread(dev, RXCFG);
rx_cfg           1030 drivers/net/ethernet/apple/bmac.c 		rx_cfg &= ~RxPromiscEnable;
rx_cfg           1031 drivers/net/ethernet/apple/bmac.c 		bmwrite(dev, RXCFG, rx_cfg);
rx_cfg            501 drivers/net/ethernet/brocade/bna/bfi_enet.h 	struct bfi_enet_rx_cfg	rx_cfg;
rx_cfg            346 drivers/net/ethernet/brocade/bna/bna.h 			       struct bna_rx_config *rx_cfg,
rx_cfg           1625 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	cfg_req->rx_cfg.frame_size = bna_enet_mtu_get(&rx->bna->enet);
rx_cfg           1651 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 				cfg_req->rx_cfg.multi_buffer =
rx_cfg           1690 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		cfg_req->rx_cfg.rxq_type = BFI_ENET_RXQ_LARGE_SMALL;
rx_cfg           1694 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		cfg_req->rx_cfg.rxq_type = BFI_ENET_RXQ_HDS;
rx_cfg           1695 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		cfg_req->rx_cfg.hds.type = rx->hds_cfg.hdr_type;
rx_cfg           1696 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		cfg_req->rx_cfg.hds.force_offset = rx->hds_cfg.forced_offset;
rx_cfg           1697 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		cfg_req->rx_cfg.hds.max_header_size = rx->hds_cfg.forced_offset;
rx_cfg           1701 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		cfg_req->rx_cfg.rxq_type = BFI_ENET_RXQ_SINGLE;
rx_cfg           1707 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	cfg_req->rx_cfg.strip_vlan = rx->rxf.vlan_strip_status;
rx_cfg           1741 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_rx_res_check(struct bna_rx_mod *rx_mod, struct bna_rx_config *rx_cfg)
rx_cfg           1748 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	if (rx_cfg->rxp_type == BNA_RXP_SINGLE) {
rx_cfg           1749 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		if ((rx_mod->rxp_free_count < rx_cfg->num_paths) ||
rx_cfg           1750 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 			(rx_mod->rxq_free_count < rx_cfg->num_paths))
rx_cfg           1753 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		if ((rx_mod->rxp_free_count < rx_cfg->num_paths) ||
rx_cfg           1754 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 			(rx_mod->rxq_free_count < (2 * rx_cfg->num_paths)))
rx_cfg           2251 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		struct bna_rx_config *rx_cfg,
rx_cfg           2280 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	if (!bna_rx_res_check(rx_mod, rx_cfg))
rx_cfg           2307 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	rx = bna_rx_get(rx_mod, rx_cfg->rx_type);
rx_cfg           2338 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	rx->num_paths = rx_cfg->num_paths;
rx_cfg           2343 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		rxp->type = rx_cfg->rxp_type;
rx_cfg           2348 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		if (BNA_RXP_SINGLE == rx_cfg->rxp_type)
rx_cfg           2371 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		rxp->cq.ib.coalescing_timeo = rx_cfg->coalescing_timeo;
rx_cfg           2385 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		q0->rcb->q_depth = rx_cfg->q0_depth;
rx_cfg           2386 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		q0->q_depth = rx_cfg->q0_depth;
rx_cfg           2387 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		q0->multi_buffer = rx_cfg->q0_multi_buf;
rx_cfg           2388 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		q0->buffer_size = rx_cfg->q0_buf_size;
rx_cfg           2389 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		q0->num_vecs = rx_cfg->q0_num_vecs;
rx_cfg           2412 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 			q1->rcb->q_depth = rx_cfg->q1_depth;
rx_cfg           2413 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 			q1->q_depth = rx_cfg->q1_depth;
rx_cfg           2419 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 			q1->buffer_size = (rx_cfg->rxp_type == BNA_RXP_HDS) ?
rx_cfg           2420 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 					rx_cfg->hds_config.forced_offset
rx_cfg           2421 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 					: rx_cfg->q1_buf_size;
rx_cfg           2437 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		cq_depth = rx_cfg->q0_depth +
rx_cfg           2438 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 			((rx_cfg->rxp_type == BNA_RXP_SINGLE) ?
rx_cfg           2439 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 			 0 : rx_cfg->q1_depth);
rx_cfg           2471 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	rx->hds_cfg = rx_cfg->hds_config;
rx_cfg           2473 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	bna_rxf_init(&rx->rxf, rx, rx_cfg, res_info);
rx_cfg           1198 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c static void t3_gate_rx_traffic(struct cmac *mac, u32 *rx_cfg,
rx_cfg           1205 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c 	*rx_cfg = t3_read_reg(mac->adapter, A_XGM_RX_CFG);
rx_cfg           1220 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c static void t3_open_rx_traffic(struct cmac *mac, u32 rx_cfg,
rx_cfg           1226 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c 			 rx_cfg);
rx_cfg           1251 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c 		u32 rx_cfg, rx_hash_high, rx_hash_low;
rx_cfg           1255 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c 		t3_gate_rx_traffic(mac, &rx_cfg, &rx_hash_high, &rx_hash_low);
rx_cfg           1264 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c 		t3_open_rx_traffic(mac, rx_cfg, rx_hash_high, rx_hash_low);
rx_cfg           1304 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c 	u32 rx_cfg, rx_hash_high, rx_hash_low;
rx_cfg           1306 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c 	t3_gate_rx_traffic(mac, &rx_cfg, &rx_hash_high, &rx_hash_low);
rx_cfg           1314 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c 	t3_open_rx_traffic(mac, rx_cfg, rx_hash_high, rx_hash_low);
rx_cfg            188 drivers/net/ethernet/google/gve/gve.h 	struct gve_queue_config rx_cfg;
rx_cfg            365 drivers/net/ethernet/google/gve/gve.h 	return priv->rx_cfg.num_queues;
rx_cfg             55 drivers/net/ethernet/google/gve/gve_ethtool.c 	for (i = 0; i < priv->rx_cfg.num_queues; i++) {
rx_cfg             82 drivers/net/ethernet/google/gve/gve_ethtool.c 		       (priv->rx_cfg.num_queues * NUM_GVE_RX_CNTS) +
rx_cfg            102 drivers/net/ethernet/google/gve/gve_ethtool.c 	     ring < priv->rx_cfg.num_queues; ring++) {
rx_cfg            138 drivers/net/ethernet/google/gve/gve_ethtool.c 		for (ring = 0; ring < priv->rx_cfg.num_queues; ring++) {
rx_cfg            145 drivers/net/ethernet/google/gve/gve_ethtool.c 		i += priv->rx_cfg.num_queues * NUM_GVE_RX_CNTS;
rx_cfg            169 drivers/net/ethernet/google/gve/gve_ethtool.c 	cmd->max_rx = priv->rx_cfg.max_queues;
rx_cfg            173 drivers/net/ethernet/google/gve/gve_ethtool.c 	cmd->rx_count = priv->rx_cfg.num_queues;
rx_cfg            184 drivers/net/ethernet/google/gve/gve_ethtool.c 	struct gve_queue_config new_rx_cfg = priv->rx_cfg;
rx_cfg            200 drivers/net/ethernet/google/gve/gve_ethtool.c 		priv->rx_cfg.num_queues = new_rx;
rx_cfg             36 drivers/net/ethernet/google/gve/gve_main.c 		for (ring = 0; ring < priv->rx_cfg.num_queues; ring++) {
rx_cfg            166 drivers/net/ethernet/google/gve/gve_main.c 		priv->rx_cfg.max_queues = min_t(int, priv->rx_cfg.max_queues,
rx_cfg            171 drivers/net/ethernet/google/gve/gve_main.c 			priv->rx_cfg.max_queues);
rx_cfg            174 drivers/net/ethernet/google/gve/gve_main.c 		if (priv->rx_cfg.num_queues > priv->rx_cfg.max_queues)
rx_cfg            175 drivers/net/ethernet/google/gve/gve_main.c 			priv->rx_cfg.num_queues = priv->rx_cfg.max_queues;
rx_cfg            386 drivers/net/ethernet/google/gve/gve_main.c 	for (i = 0; i < priv->rx_cfg.num_queues; i++) {
rx_cfg            423 drivers/net/ethernet/google/gve/gve_main.c 	priv->rx = kvzalloc(priv->rx_cfg.num_queues * sizeof(*priv->rx),
rx_cfg            439 drivers/net/ethernet/google/gve/gve_main.c 	for (i = 0; i < priv->rx_cfg.num_queues; i++) {
rx_cfg            476 drivers/net/ethernet/google/gve/gve_main.c 	for (i = 0; i < priv->rx_cfg.num_queues; i++) {
rx_cfg            507 drivers/net/ethernet/google/gve/gve_main.c 		for (i = 0; i < priv->rx_cfg.num_queues; i++) {
rx_cfg            686 drivers/net/ethernet/google/gve/gve_main.c 	err = netif_set_real_num_rx_queues(dev, priv->rx_cfg.num_queues);
rx_cfg            772 drivers/net/ethernet/google/gve/gve_main.c 		priv->rx_cfg = new_rx_config;
rx_cfg            782 drivers/net/ethernet/google/gve/gve_main.c 	priv->rx_cfg = new_rx_config;
rx_cfg            809 drivers/net/ethernet/google/gve/gve_main.c 	for (idx = 0; idx < priv->rx_cfg.num_queues; idx++) {
rx_cfg            837 drivers/net/ethernet/google/gve/gve_main.c 	for (idx = 0; idx < priv->rx_cfg.num_queues; idx++) {
rx_cfg            955 drivers/net/ethernet/google/gve/gve_main.c 	priv->rx_cfg.max_queues =
rx_cfg            956 drivers/net/ethernet/google/gve/gve_main.c 		min_t(int, priv->rx_cfg.max_queues, priv->num_ntfy_blks / 2);
rx_cfg            959 drivers/net/ethernet/google/gve/gve_main.c 	priv->rx_cfg.num_queues = priv->rx_cfg.max_queues;
rx_cfg            963 drivers/net/ethernet/google/gve/gve_main.c 		priv->rx_cfg.num_queues = min_t(int, priv->default_num_queues,
rx_cfg            964 drivers/net/ethernet/google/gve/gve_main.c 						priv->rx_cfg.num_queues);
rx_cfg            968 drivers/net/ethernet/google/gve/gve_main.c 		   priv->tx_cfg.num_queues, priv->rx_cfg.num_queues);
rx_cfg            970 drivers/net/ethernet/google/gve/gve_main.c 		   priv->tx_cfg.max_queues, priv->rx_cfg.max_queues);
rx_cfg           1162 drivers/net/ethernet/google/gve/gve_main.c 	priv->rx_cfg.max_queues = max_rx_queues;
rx_cfg            185 drivers/net/ethernet/google/gve/gve_rx.c 	for (i = 0; i < priv->rx_cfg.num_queues; i++) {
rx_cfg            208 drivers/net/ethernet/google/gve/gve_rx.c 	for (i = 0; i < priv->rx_cfg.num_queues; i++)
rx_cfg            444 drivers/net/ethernet/marvell/octeontx2/af/mbox.h 	u64 rx_cfg;   /* See NIX_AF_LF(0..127)_RX_CFG */
rx_cfg            897 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c 	rvu_write64(rvu, blkaddr, NIX_AF_LFX_RX_CFG(nixlf), req->rx_cfg);
rx_cfg           1274 drivers/net/ethernet/micrel/ksz884x.c 	u32 rx_cfg;
rx_cfg           3206 drivers/net/ethernet/micrel/ksz884x.c 	u32 rx_cfg;
rx_cfg           3209 drivers/net/ethernet/micrel/ksz884x.c 	rx_cfg = hw->rx_cfg;
rx_cfg           3212 drivers/net/ethernet/micrel/ksz884x.c 		hw->rx_cfg |= DMA_RX_FLOW_ENABLE;
rx_cfg           3214 drivers/net/ethernet/micrel/ksz884x.c 		hw->rx_cfg &= ~DMA_RX_FLOW_ENABLE;
rx_cfg           3220 drivers/net/ethernet/micrel/ksz884x.c 		if (rx_cfg != hw->rx_cfg)
rx_cfg           3221 drivers/net/ethernet/micrel/ksz884x.c 			writel(hw->rx_cfg, hw->io + KS_DMA_RX_CTRL);
rx_cfg           3776 drivers/net/ethernet/micrel/ksz884x.c 	hw->rx_cfg = (DMA_RX_BROADCAST | DMA_RX_UNICAST |
rx_cfg           3778 drivers/net/ethernet/micrel/ksz884x.c 	hw->rx_cfg |= KS884X_DMA_RX_MULTICAST;
rx_cfg           3781 drivers/net/ethernet/micrel/ksz884x.c 	hw->rx_cfg |= (DMA_RX_CSUM_TCP | DMA_RX_CSUM_IP);
rx_cfg           3784 drivers/net/ethernet/micrel/ksz884x.c 		hw->rx_cfg |= DMA_RX_ALL_MULTICAST;
rx_cfg           3786 drivers/net/ethernet/micrel/ksz884x.c 		hw->rx_cfg |= DMA_RX_PROMISCUOUS;
rx_cfg           3885 drivers/net/ethernet/micrel/ksz884x.c 	writel(hw->rx_cfg, hw->io + KS_DMA_RX_CTRL);
rx_cfg           3909 drivers/net/ethernet/micrel/ksz884x.c 	writel((hw->rx_cfg & ~DMA_RX_ENABLE), hw->io + KS_DMA_RX_CTRL);
rx_cfg           4203 drivers/net/ethernet/micrel/ksz884x.c 		hw->rx_cfg |= DMA_RX_ALL_MULTICAST;
rx_cfg           4205 drivers/net/ethernet/micrel/ksz884x.c 		hw->rx_cfg &= ~DMA_RX_ALL_MULTICAST;
rx_cfg           4224 drivers/net/ethernet/micrel/ksz884x.c 		hw->rx_cfg |= DMA_RX_PROMISCUOUS;
rx_cfg           4226 drivers/net/ethernet/micrel/ksz884x.c 		hw->rx_cfg &= ~DMA_RX_PROMISCUOUS;
rx_cfg           5017 drivers/net/ethernet/micrel/ksz884x.c 	if (hw->rx_cfg & (DMA_RX_CSUM_UDP | DMA_RX_CSUM_TCP))
rx_cfg           5213 drivers/net/ethernet/micrel/ksz884x.c 		if (hw->enabled && (hw->rx_cfg & DMA_RX_ENABLE)) {
rx_cfg           5424 drivers/net/ethernet/micrel/ksz884x.c 		hw->rx_cfg |= DMA_RX_ERROR;
rx_cfg           5427 drivers/net/ethernet/micrel/ksz884x.c 		hw->rx_cfg &= ~DMA_RX_ERROR;
rx_cfg           6335 drivers/net/ethernet/micrel/ksz884x.c 			(hw->rx_cfg & DMA_RX_FLOW_ENABLE) ? 1 : 0;
rx_cfg           6589 drivers/net/ethernet/micrel/ksz884x.c 		hw->rx_cfg |= DMA_RX_CSUM_TCP | DMA_RX_CSUM_IP;
rx_cfg           6591 drivers/net/ethernet/micrel/ksz884x.c 		hw->rx_cfg &= ~(DMA_RX_CSUM_TCP | DMA_RX_CSUM_IP);
rx_cfg           6594 drivers/net/ethernet/micrel/ksz884x.c 		writel(hw->rx_cfg, hw->io + KS_DMA_RX_CTRL);
rx_cfg            697 drivers/net/ethernet/neterion/s2io.c 		struct rx_ring_config *rx_cfg = &config->rx_cfg[i];
rx_cfg            700 drivers/net/ethernet/neterion/s2io.c 		if (rx_cfg->num_rxd % (rxd_count[nic->rxd_mode] + 1)) {
rx_cfg            706 drivers/net/ethernet/neterion/s2io.c 		size += rx_cfg->num_rxd;
rx_cfg            707 drivers/net/ethernet/neterion/s2io.c 		ring->block_count = rx_cfg->num_rxd /
rx_cfg            709 drivers/net/ethernet/neterion/s2io.c 		ring->pkt_cnt = rx_cfg->num_rxd - ring->block_count;
rx_cfg            717 drivers/net/ethernet/neterion/s2io.c 		struct rx_ring_config *rx_cfg = &config->rx_cfg[i];
rx_cfg            722 drivers/net/ethernet/neterion/s2io.c 		ring->rx_curr_get_info.ring_len = rx_cfg->num_rxd - 1;
rx_cfg            725 drivers/net/ethernet/neterion/s2io.c 		ring->rx_curr_put_info.ring_len = rx_cfg->num_rxd - 1;
rx_cfg            729 drivers/net/ethernet/neterion/s2io.c 		blk_cnt = rx_cfg->num_rxd / (rxd_count[nic->rxd_mode] + 1);
rx_cfg            789 drivers/net/ethernet/neterion/s2io.c 			struct rx_ring_config *rx_cfg = &config->rx_cfg[i];
rx_cfg            792 drivers/net/ethernet/neterion/s2io.c 			blk_cnt = rx_cfg->num_rxd /
rx_cfg            954 drivers/net/ethernet/neterion/s2io.c 			struct rx_ring_config *rx_cfg = &config->rx_cfg[i];
rx_cfg            957 drivers/net/ethernet/neterion/s2io.c 			blk_cnt = rx_cfg->num_rxd /
rx_cfg           1340 drivers/net/ethernet/neterion/s2io.c 		struct rx_ring_config *rx_cfg = &config->rx_cfg[i];
rx_cfg           1342 drivers/net/ethernet/neterion/s2io.c 		val64 |= vBIT(rx_cfg->ring_priority, (5 + (i * 8)), 3);
rx_cfg           5469 drivers/net/ethernet/neterion/s2io.c 		rx_desc_count += sp->config.rx_cfg[i].num_rxd;
rx_cfg           6878 drivers/net/ethernet/neterion/s2io.c 		struct rx_ring_config *rx_cfg = &config->rx_cfg[i];
rx_cfg           6881 drivers/net/ethernet/neterion/s2io.c 		blk_cnt = rx_cfg->num_rxd / (rxd_count[sp->rxd_mode] + 1);
rx_cfg           7807 drivers/net/ethernet/neterion/s2io.c 		struct rx_ring_config *rx_cfg = &config->rx_cfg[i];
rx_cfg           7810 drivers/net/ethernet/neterion/s2io.c 		rx_cfg->num_rxd = rx_ring_sz[i] * (rxd_count[sp->rxd_mode] + 1);
rx_cfg           7811 drivers/net/ethernet/neterion/s2io.c 		rx_cfg->ring_priority = i;
rx_cfg           7820 drivers/net/ethernet/neterion/s2io.c 		struct rx_ring_config *rx_cfg = &config->rx_cfg[i];
rx_cfg           7822 drivers/net/ethernet/neterion/s2io.c 		rx_cfg->ring_org = RING_ORG_BUFF1;
rx_cfg           7823 drivers/net/ethernet/neterion/s2io.c 		rx_cfg->f_no_snoop = (NO_SNOOP_RXD | NO_SNOOP_RXD_BUFFER);
rx_cfg            450 drivers/net/ethernet/neterion/s2io.h 	struct rx_ring_config rx_cfg[MAX_RX_RINGS];	/*Per-Rx Ring config */
rx_cfg            393 drivers/net/ethernet/sun/niu.c 	u32 tx_cfg, rx_cfg;
rx_cfg            397 drivers/net/ethernet/sun/niu.c 	rx_cfg = (PLL_RX_CFG_ENRX | PLL_RX_CFG_TERM_0P8VDDT |
rx_cfg            408 drivers/net/ethernet/sun/niu.c 		rx_cfg |= PLL_RX_CFG_ENTEST;
rx_cfg            419 drivers/net/ethernet/sun/niu.c 		int err = esr2_set_rx_cfg(np, i, rx_cfg);
rx_cfg            433 drivers/net/ethernet/sun/niu.c 	u32 tx_cfg, rx_cfg;
rx_cfg            439 drivers/net/ethernet/sun/niu.c 	rx_cfg = (PLL_RX_CFG_ENRX | PLL_RX_CFG_TERM_0P8VDDT |
rx_cfg            444 drivers/net/ethernet/sun/niu.c 		rx_cfg |= PLL_RX_CFG_EQ_LP_ADAPTIVE;
rx_cfg            453 drivers/net/ethernet/sun/niu.c 		rx_cfg |= PLL_RX_CFG_ENTEST;
rx_cfg            487 drivers/net/ethernet/sun/niu.c 		err = esr2_set_rx_cfg(np, i, rx_cfg);
rx_cfg            527 drivers/net/ethernet/sun/niu.c 	u32 tx_cfg, rx_cfg, pll_cfg, pll_sts;
rx_cfg            534 drivers/net/ethernet/sun/niu.c 	rx_cfg = (PLL_RX_CFG_ENRX | PLL_RX_CFG_TERM_0P8VDDT |
rx_cfg            545 drivers/net/ethernet/sun/niu.c 		rx_cfg |= PLL_RX_CFG_ENTEST;
rx_cfg            579 drivers/net/ethernet/sun/niu.c 		err = esr2_set_rx_cfg(np, i, rx_cfg);
rx_cfg            162 drivers/net/wireless/intel/iwlwifi/pcie/ctxt-info.c 	struct iwl_context_info_rbd_cfg *rx_cfg;
rx_cfg            204 drivers/net/wireless/intel/iwlwifi/pcie/ctxt-info.c 	rx_cfg = &ctxt_info->rbd_cfg;
rx_cfg            205 drivers/net/wireless/intel/iwlwifi/pcie/ctxt-info.c 	rx_cfg->free_rbd_addr = cpu_to_le64(trans_pcie->rxq->bd_dma);
rx_cfg            206 drivers/net/wireless/intel/iwlwifi/pcie/ctxt-info.c 	rx_cfg->used_rbd_addr = cpu_to_le64(trans_pcie->rxq->used_bd_dma);
rx_cfg            207 drivers/net/wireless/intel/iwlwifi/pcie/ctxt-info.c 	rx_cfg->status_wr_ptr = cpu_to_le64(trans_pcie->rxq->rb_stts_dma);
rx_cfg            380 drivers/soc/qcom/qcom-geni-se.c 			    bool msb_to_lsb, bool tx_cfg, bool rx_cfg)
rx_cfg            417 drivers/soc/qcom/qcom-geni-se.c 	if (rx_cfg) {
rx_cfg             89 drivers/staging/pi433/pi433_if.c 	struct pi433_rx_cfg	rx_cfg;
rx_cfg            158 drivers/staging/pi433/pi433_if.c rf69_set_rx_cfg(struct pi433_device *dev, struct pi433_rx_cfg *rx_cfg)
rx_cfg            164 drivers/staging/pi433/pi433_if.c 	ret = rf69_set_frequency(dev->spi, rx_cfg->frequency);
rx_cfg            167 drivers/staging/pi433/pi433_if.c 	ret = rf69_set_bit_rate(dev->spi, rx_cfg->bit_rate);
rx_cfg            170 drivers/staging/pi433/pi433_if.c 	ret = rf69_set_modulation(dev->spi, rx_cfg->modulation);
rx_cfg            173 drivers/staging/pi433/pi433_if.c 	ret = rf69_set_antenna_impedance(dev->spi, rx_cfg->antenna_impedance);
rx_cfg            176 drivers/staging/pi433/pi433_if.c 	ret = rf69_set_rssi_threshold(dev->spi, rx_cfg->rssi_threshold);
rx_cfg            179 drivers/staging/pi433/pi433_if.c 	ret = rf69_set_ook_threshold_dec(dev->spi, rx_cfg->threshold_decrement);
rx_cfg            182 drivers/staging/pi433/pi433_if.c 	ret = rf69_set_bandwidth(dev->spi, rx_cfg->bw_mantisse,
rx_cfg            183 drivers/staging/pi433/pi433_if.c 				 rx_cfg->bw_exponent);
rx_cfg            186 drivers/staging/pi433/pi433_if.c 	ret = rf69_set_bandwidth_during_afc(dev->spi, rx_cfg->bw_mantisse,
rx_cfg            187 drivers/staging/pi433/pi433_if.c 					    rx_cfg->bw_exponent);
rx_cfg            190 drivers/staging/pi433/pi433_if.c 	ret = rf69_set_dagc(dev->spi, rx_cfg->dagc);
rx_cfg            194 drivers/staging/pi433/pi433_if.c 	dev->rx_bytes_to_drop = rx_cfg->bytes_to_drop;
rx_cfg            198 drivers/staging/pi433/pi433_if.c 	if (rx_cfg->enable_sync == OPTION_ON) {
rx_cfg            216 drivers/staging/pi433/pi433_if.c 	if (rx_cfg->enable_length_byte == OPTION_ON) {
rx_cfg            226 drivers/staging/pi433/pi433_if.c 					 rx_cfg->enable_address_filtering);
rx_cfg            230 drivers/staging/pi433/pi433_if.c 	if (rx_cfg->enable_crc == OPTION_ON) {
rx_cfg            241 drivers/staging/pi433/pi433_if.c 	ret = rf69_set_sync_size(dev->spi, rx_cfg->sync_length);
rx_cfg            244 drivers/staging/pi433/pi433_if.c 	if (rx_cfg->enable_length_byte == OPTION_ON) {
rx_cfg            248 drivers/staging/pi433/pi433_if.c 	} else if (rx_cfg->fixed_message_length != 0) {
rx_cfg            249 drivers/staging/pi433/pi433_if.c 		payload_length = rx_cfg->fixed_message_length;
rx_cfg            250 drivers/staging/pi433/pi433_if.c 		if (rx_cfg->enable_length_byte  == OPTION_ON)
rx_cfg            252 drivers/staging/pi433/pi433_if.c 		if (rx_cfg->enable_address_filtering != filtering_off)
rx_cfg            264 drivers/staging/pi433/pi433_if.c 	if (rx_cfg->enable_sync == OPTION_ON) {
rx_cfg            265 drivers/staging/pi433/pi433_if.c 		ret = rf69_set_sync_values(dev->spi, rx_cfg->sync_pattern);
rx_cfg            269 drivers/staging/pi433/pi433_if.c 	if (rx_cfg->enable_address_filtering != filtering_off) {
rx_cfg            270 drivers/staging/pi433/pi433_if.c 		ret = rf69_set_node_address(dev->spi, rx_cfg->node_address);
rx_cfg            274 drivers/staging/pi433/pi433_if.c 						 rx_cfg->broadcast_address);
rx_cfg            372 drivers/staging/pi433/pi433_if.c 	retval = rf69_set_rx_cfg(dev, &dev->rx_cfg);
rx_cfg            461 drivers/staging/pi433/pi433_if.c 	if (dev->rx_cfg.fixed_message_length != 0) {
rx_cfg            462 drivers/staging/pi433/pi433_if.c 		if (dev->rx_cfg.fixed_message_length > dev->rx_buffer_size) {
rx_cfg            466 drivers/staging/pi433/pi433_if.c 		bytes_total = dev->rx_cfg.fixed_message_length;
rx_cfg            476 drivers/staging/pi433/pi433_if.c 	if (dev->rx_cfg.enable_length_byte == OPTION_ON) {
rx_cfg            493 drivers/staging/pi433/pi433_if.c 	if (dev->rx_cfg.enable_address_filtering != filtering_off) {
rx_cfg            903 drivers/staging/pi433/pi433_if.c 		if (copy_to_user(argp, &device->rx_cfg,
rx_cfg            916 drivers/staging/pi433/pi433_if.c 		if (copy_from_user(&device->rx_cfg, argp,
rx_cfg            469 drivers/usb/gadget/udc/bcm63xx_udc.c 		const struct iudma_ch_cfg *rx_cfg = &iudma_defaults[i];
rx_cfg            475 drivers/usb/gadget/udc/bcm63xx_udc.c 			((rx_fifo_slot + rx_cfg->n_fifo_slots - 1) <<
rx_cfg            477 drivers/usb/gadget/udc/bcm63xx_udc.c 		rx_fifo_slot += rx_cfg->n_fifo_slots;
rx_cfg            480 drivers/usb/gadget/udc/bcm63xx_udc.c 			    is_hs ? rx_cfg->max_pkt_hs : rx_cfg->max_pkt_fs,
rx_cfg            104 include/linux/hsi/hsi.h 	struct hsi_config	rx_cfg;
rx_cfg            129 include/linux/hsi/hsi.h 	struct hsi_config	rx_cfg;
rx_cfg            224 include/linux/hsi/hsi.h 	struct hsi_config		rx_cfg;
rx_cfg            398 include/linux/qcom-geni-se.h 			    bool msb_to_lsb, bool tx_cfg, bool rx_cfg);