rx_cfg 355 drivers/hsi/clients/hsi_char.c tmp = cl->rx_cfg; rx_cfg 356 drivers/hsi/clients/hsi_char.c cl->rx_cfg.mode = rxc->mode; rx_cfg 357 drivers/hsi/clients/hsi_char.c cl->rx_cfg.num_hw_channels = rxc->channels; rx_cfg 358 drivers/hsi/clients/hsi_char.c cl->rx_cfg.flow = rxc->flow; rx_cfg 361 drivers/hsi/clients/hsi_char.c cl->rx_cfg = tmp; rx_cfg 372 drivers/hsi/clients/hsi_char.c rxc->mode = cl->rx_cfg.mode; rx_cfg 373 drivers/hsi/clients/hsi_char.c rxc->channels = cl->rx_cfg.num_hw_channels; rx_cfg 374 drivers/hsi/clients/hsi_char.c rxc->flow = cl->rx_cfg.flow; rx_cfg 425 drivers/hsi/clients/hsi_char.c if (channel->ch >= channel->cl->rx_cfg.num_hw_channels) rx_cfg 179 drivers/hsi/clients/nokia-modem.c ssip.rx_cfg = cl->rx_cfg; rx_cfg 202 drivers/hsi/clients/nokia-modem.c cmtspeech.rx_cfg = cl->rx_cfg; rx_cfg 490 drivers/hsi/controllers/omap_ssi_port.c writel_relaxed(cl->rx_cfg.num_hw_channels, ssr + SSI_SSR_CHANNELS_REG); rx_cfg 494 drivers/hsi/controllers/omap_ssi_port.c (cl->rx_cfg.mode != SSI_MODE_FRAME)) rx_cfg 496 drivers/hsi/controllers/omap_ssi_port.c writel_relaxed(cl->rx_cfg.mode, ssr + SSI_SSR_MODE_REG); rx_cfg 497 drivers/hsi/controllers/omap_ssi_port.c omap_port->channels = max(cl->rx_cfg.num_hw_channels, rx_cfg 509 drivers/hsi/controllers/omap_ssi_port.c omap_port->ssr.channels = cl->rx_cfg.num_hw_channels; rx_cfg 510 drivers/hsi/controllers/omap_ssi_port.c omap_port->ssr.mode = cl->rx_cfg.mode; rx_cfg 63 drivers/hsi/hsi_core.c kfree(cl->rx_cfg.channels); rx_cfg 86 drivers/hsi/hsi_core.c cl->rx_cfg = info->rx_cfg; rx_cfg 87 drivers/hsi/hsi_core.c if (cl->rx_cfg.channels) { rx_cfg 88 drivers/hsi/hsi_core.c size = cl->rx_cfg.num_channels * sizeof(*cl->rx_cfg.channels); rx_cfg 89 drivers/hsi/hsi_core.c cl->rx_cfg.channels = kmemdup(info->rx_cfg.channels, size, rx_cfg 91 drivers/hsi/hsi_core.c if (!cl->rx_cfg.channels) rx_cfg 218 drivers/hsi/hsi_core.c &cl->rx_cfg.mode); rx_cfg 227 drivers/hsi/hsi_core.c cl->rx_cfg.mode = mode; rx_cfg 235 drivers/hsi/hsi_core.c cl->rx_cfg.speed = cl->tx_cfg.speed; rx_cfg 238 drivers/hsi/hsi_core.c &cl->rx_cfg.flow); rx_cfg 243 drivers/hsi/hsi_core.c &cl->rx_cfg.arb_mode); rx_cfg 255 drivers/hsi/hsi_core.c cl->rx_cfg.num_channels = cells; rx_cfg 257 drivers/hsi/hsi_core.c cl->rx_cfg.channels = kcalloc(cells, sizeof(channel), GFP_KERNEL); rx_cfg 258 drivers/hsi/hsi_core.c if (!cl->rx_cfg.channels) { rx_cfg 284 drivers/hsi/hsi_core.c cl->rx_cfg.channels[i] = channel; rx_cfg 288 drivers/hsi/hsi_core.c cl->rx_cfg.num_hw_channels = max_chan + 1; rx_cfg 306 drivers/hsi/hsi_core.c kfree(cl->rx_cfg.channels); rx_cfg 742 drivers/hsi/hsi_core.c if (!cl->rx_cfg.channels) rx_cfg 745 drivers/hsi/hsi_core.c for (i = 0; i < cl->rx_cfg.num_channels; i++) rx_cfg 746 drivers/hsi/hsi_core.c if (!strcmp(cl->rx_cfg.channels[i].name, name)) rx_cfg 747 drivers/hsi/hsi_core.c return cl->rx_cfg.channels[i].id; rx_cfg 900 drivers/net/ethernet/apple/bmac.c unsigned short rx_cfg; rx_cfg 902 drivers/net/ethernet/apple/bmac.c rx_cfg = bmread(dev, RXCFG); rx_cfg 903 drivers/net/ethernet/apple/bmac.c rx_cfg &= ~RxMACEnable; rx_cfg 904 drivers/net/ethernet/apple/bmac.c bmwrite(dev, RXCFG, rx_cfg); rx_cfg 906 drivers/net/ethernet/apple/bmac.c rx_cfg = bmread(dev, RXCFG); rx_cfg 907 drivers/net/ethernet/apple/bmac.c } while (rx_cfg & RxMACEnable); rx_cfg 913 drivers/net/ethernet/apple/bmac.c unsigned short rx_cfg; rx_cfg 915 drivers/net/ethernet/apple/bmac.c rx_cfg = bmread(dev, RXCFG); rx_cfg 916 drivers/net/ethernet/apple/bmac.c rx_cfg |= RxMACEnable; rx_cfg 917 drivers/net/ethernet/apple/bmac.c if (hash_enable) rx_cfg |= RxHashFilterEnable; rx_cfg 918 drivers/net/ethernet/apple/bmac.c else rx_cfg &= ~RxHashFilterEnable; rx_cfg 919 drivers/net/ethernet/apple/bmac.c if (promisc_enable) rx_cfg |= RxPromiscEnable; rx_cfg 920 drivers/net/ethernet/apple/bmac.c else rx_cfg &= ~RxPromiscEnable; rx_cfg 924 drivers/net/ethernet/apple/bmac.c bmwrite(dev, RXCFG, rx_cfg ); rx_cfg 925 drivers/net/ethernet/apple/bmac.c return rx_cfg; rx_cfg 972 drivers/net/ethernet/apple/bmac.c unsigned short rx_cfg; rx_cfg 983 drivers/net/ethernet/apple/bmac.c rx_cfg = bmac_rx_on(dev, 1, 0); rx_cfg 986 drivers/net/ethernet/apple/bmac.c rx_cfg = bmread(dev, RXCFG); rx_cfg 987 drivers/net/ethernet/apple/bmac.c rx_cfg |= RxPromiscEnable; rx_cfg 988 drivers/net/ethernet/apple/bmac.c bmwrite(dev, RXCFG, rx_cfg); rx_cfg 989 drivers/net/ethernet/apple/bmac.c rx_cfg = bmac_rx_on(dev, 0, 1); rx_cfg 990 drivers/net/ethernet/apple/bmac.c XXDEBUG(("bmac: promisc mode enabled, rx_cfg=%#08x\n", rx_cfg)); rx_cfg 995 drivers/net/ethernet/apple/bmac.c rx_cfg = bmac_rx_on(dev, 0, 0); rx_cfg 996 drivers/net/ethernet/apple/bmac.c XXDEBUG(("bmac: multi disabled, rx_cfg=%#08x\n", rx_cfg)); rx_cfg 1001 drivers/net/ethernet/apple/bmac.c rx_cfg = bmac_rx_on(dev, 1, 0); rx_cfg 1002 drivers/net/ethernet/apple/bmac.c XXDEBUG(("bmac: multi enabled, rx_cfg=%#08x\n", rx_cfg)); rx_cfg 1014 drivers/net/ethernet/apple/bmac.c unsigned short rx_cfg; rx_cfg 1023 drivers/net/ethernet/apple/bmac.c rx_cfg = bmread(dev, RXCFG); rx_cfg 1024 drivers/net/ethernet/apple/bmac.c rx_cfg |= RxPromiscEnable; rx_cfg 1025 drivers/net/ethernet/apple/bmac.c bmwrite(dev, RXCFG, rx_cfg); rx_cfg 1029 drivers/net/ethernet/apple/bmac.c rx_cfg = bmread(dev, RXCFG); rx_cfg 1030 drivers/net/ethernet/apple/bmac.c rx_cfg &= ~RxPromiscEnable; rx_cfg 1031 drivers/net/ethernet/apple/bmac.c bmwrite(dev, RXCFG, rx_cfg); rx_cfg 501 drivers/net/ethernet/brocade/bna/bfi_enet.h struct bfi_enet_rx_cfg rx_cfg; rx_cfg 346 drivers/net/ethernet/brocade/bna/bna.h struct bna_rx_config *rx_cfg, rx_cfg 1625 drivers/net/ethernet/brocade/bna/bna_tx_rx.c cfg_req->rx_cfg.frame_size = bna_enet_mtu_get(&rx->bna->enet); rx_cfg 1651 drivers/net/ethernet/brocade/bna/bna_tx_rx.c cfg_req->rx_cfg.multi_buffer = rx_cfg 1690 drivers/net/ethernet/brocade/bna/bna_tx_rx.c cfg_req->rx_cfg.rxq_type = BFI_ENET_RXQ_LARGE_SMALL; rx_cfg 1694 drivers/net/ethernet/brocade/bna/bna_tx_rx.c cfg_req->rx_cfg.rxq_type = BFI_ENET_RXQ_HDS; rx_cfg 1695 drivers/net/ethernet/brocade/bna/bna_tx_rx.c cfg_req->rx_cfg.hds.type = rx->hds_cfg.hdr_type; rx_cfg 1696 drivers/net/ethernet/brocade/bna/bna_tx_rx.c cfg_req->rx_cfg.hds.force_offset = rx->hds_cfg.forced_offset; rx_cfg 1697 drivers/net/ethernet/brocade/bna/bna_tx_rx.c cfg_req->rx_cfg.hds.max_header_size = rx->hds_cfg.forced_offset; rx_cfg 1701 drivers/net/ethernet/brocade/bna/bna_tx_rx.c cfg_req->rx_cfg.rxq_type = BFI_ENET_RXQ_SINGLE; rx_cfg 1707 drivers/net/ethernet/brocade/bna/bna_tx_rx.c cfg_req->rx_cfg.strip_vlan = rx->rxf.vlan_strip_status; rx_cfg 1741 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_rx_res_check(struct bna_rx_mod *rx_mod, struct bna_rx_config *rx_cfg) rx_cfg 1748 drivers/net/ethernet/brocade/bna/bna_tx_rx.c if (rx_cfg->rxp_type == BNA_RXP_SINGLE) { rx_cfg 1749 drivers/net/ethernet/brocade/bna/bna_tx_rx.c if ((rx_mod->rxp_free_count < rx_cfg->num_paths) || rx_cfg 1750 drivers/net/ethernet/brocade/bna/bna_tx_rx.c (rx_mod->rxq_free_count < rx_cfg->num_paths)) rx_cfg 1753 drivers/net/ethernet/brocade/bna/bna_tx_rx.c if ((rx_mod->rxp_free_count < rx_cfg->num_paths) || rx_cfg 1754 drivers/net/ethernet/brocade/bna/bna_tx_rx.c (rx_mod->rxq_free_count < (2 * rx_cfg->num_paths))) rx_cfg 2251 drivers/net/ethernet/brocade/bna/bna_tx_rx.c struct bna_rx_config *rx_cfg, rx_cfg 2280 drivers/net/ethernet/brocade/bna/bna_tx_rx.c if (!bna_rx_res_check(rx_mod, rx_cfg)) rx_cfg 2307 drivers/net/ethernet/brocade/bna/bna_tx_rx.c rx = bna_rx_get(rx_mod, rx_cfg->rx_type); rx_cfg 2338 drivers/net/ethernet/brocade/bna/bna_tx_rx.c rx->num_paths = rx_cfg->num_paths; rx_cfg 2343 drivers/net/ethernet/brocade/bna/bna_tx_rx.c rxp->type = rx_cfg->rxp_type; rx_cfg 2348 drivers/net/ethernet/brocade/bna/bna_tx_rx.c if (BNA_RXP_SINGLE == rx_cfg->rxp_type) rx_cfg 2371 drivers/net/ethernet/brocade/bna/bna_tx_rx.c rxp->cq.ib.coalescing_timeo = rx_cfg->coalescing_timeo; rx_cfg 2385 drivers/net/ethernet/brocade/bna/bna_tx_rx.c q0->rcb->q_depth = rx_cfg->q0_depth; rx_cfg 2386 drivers/net/ethernet/brocade/bna/bna_tx_rx.c q0->q_depth = rx_cfg->q0_depth; rx_cfg 2387 drivers/net/ethernet/brocade/bna/bna_tx_rx.c q0->multi_buffer = rx_cfg->q0_multi_buf; rx_cfg 2388 drivers/net/ethernet/brocade/bna/bna_tx_rx.c q0->buffer_size = rx_cfg->q0_buf_size; rx_cfg 2389 drivers/net/ethernet/brocade/bna/bna_tx_rx.c q0->num_vecs = rx_cfg->q0_num_vecs; rx_cfg 2412 drivers/net/ethernet/brocade/bna/bna_tx_rx.c q1->rcb->q_depth = rx_cfg->q1_depth; rx_cfg 2413 drivers/net/ethernet/brocade/bna/bna_tx_rx.c q1->q_depth = rx_cfg->q1_depth; rx_cfg 2419 drivers/net/ethernet/brocade/bna/bna_tx_rx.c q1->buffer_size = (rx_cfg->rxp_type == BNA_RXP_HDS) ? rx_cfg 2420 drivers/net/ethernet/brocade/bna/bna_tx_rx.c rx_cfg->hds_config.forced_offset rx_cfg 2421 drivers/net/ethernet/brocade/bna/bna_tx_rx.c : rx_cfg->q1_buf_size; rx_cfg 2437 drivers/net/ethernet/brocade/bna/bna_tx_rx.c cq_depth = rx_cfg->q0_depth + rx_cfg 2438 drivers/net/ethernet/brocade/bna/bna_tx_rx.c ((rx_cfg->rxp_type == BNA_RXP_SINGLE) ? rx_cfg 2439 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 0 : rx_cfg->q1_depth); rx_cfg 2471 drivers/net/ethernet/brocade/bna/bna_tx_rx.c rx->hds_cfg = rx_cfg->hds_config; rx_cfg 2473 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_rxf_init(&rx->rxf, rx, rx_cfg, res_info); rx_cfg 1198 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c static void t3_gate_rx_traffic(struct cmac *mac, u32 *rx_cfg, rx_cfg 1205 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c *rx_cfg = t3_read_reg(mac->adapter, A_XGM_RX_CFG); rx_cfg 1220 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c static void t3_open_rx_traffic(struct cmac *mac, u32 rx_cfg, rx_cfg 1226 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c rx_cfg); rx_cfg 1251 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c u32 rx_cfg, rx_hash_high, rx_hash_low; rx_cfg 1255 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c t3_gate_rx_traffic(mac, &rx_cfg, &rx_hash_high, &rx_hash_low); rx_cfg 1264 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c t3_open_rx_traffic(mac, rx_cfg, rx_hash_high, rx_hash_low); rx_cfg 1304 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c u32 rx_cfg, rx_hash_high, rx_hash_low; rx_cfg 1306 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c t3_gate_rx_traffic(mac, &rx_cfg, &rx_hash_high, &rx_hash_low); rx_cfg 1314 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c t3_open_rx_traffic(mac, rx_cfg, rx_hash_high, rx_hash_low); rx_cfg 188 drivers/net/ethernet/google/gve/gve.h struct gve_queue_config rx_cfg; rx_cfg 365 drivers/net/ethernet/google/gve/gve.h return priv->rx_cfg.num_queues; rx_cfg 55 drivers/net/ethernet/google/gve/gve_ethtool.c for (i = 0; i < priv->rx_cfg.num_queues; i++) { rx_cfg 82 drivers/net/ethernet/google/gve/gve_ethtool.c (priv->rx_cfg.num_queues * NUM_GVE_RX_CNTS) + rx_cfg 102 drivers/net/ethernet/google/gve/gve_ethtool.c ring < priv->rx_cfg.num_queues; ring++) { rx_cfg 138 drivers/net/ethernet/google/gve/gve_ethtool.c for (ring = 0; ring < priv->rx_cfg.num_queues; ring++) { rx_cfg 145 drivers/net/ethernet/google/gve/gve_ethtool.c i += priv->rx_cfg.num_queues * NUM_GVE_RX_CNTS; rx_cfg 169 drivers/net/ethernet/google/gve/gve_ethtool.c cmd->max_rx = priv->rx_cfg.max_queues; rx_cfg 173 drivers/net/ethernet/google/gve/gve_ethtool.c cmd->rx_count = priv->rx_cfg.num_queues; rx_cfg 184 drivers/net/ethernet/google/gve/gve_ethtool.c struct gve_queue_config new_rx_cfg = priv->rx_cfg; rx_cfg 200 drivers/net/ethernet/google/gve/gve_ethtool.c priv->rx_cfg.num_queues = new_rx; rx_cfg 36 drivers/net/ethernet/google/gve/gve_main.c for (ring = 0; ring < priv->rx_cfg.num_queues; ring++) { rx_cfg 166 drivers/net/ethernet/google/gve/gve_main.c priv->rx_cfg.max_queues = min_t(int, priv->rx_cfg.max_queues, rx_cfg 171 drivers/net/ethernet/google/gve/gve_main.c priv->rx_cfg.max_queues); rx_cfg 174 drivers/net/ethernet/google/gve/gve_main.c if (priv->rx_cfg.num_queues > priv->rx_cfg.max_queues) rx_cfg 175 drivers/net/ethernet/google/gve/gve_main.c priv->rx_cfg.num_queues = priv->rx_cfg.max_queues; rx_cfg 386 drivers/net/ethernet/google/gve/gve_main.c for (i = 0; i < priv->rx_cfg.num_queues; i++) { rx_cfg 423 drivers/net/ethernet/google/gve/gve_main.c priv->rx = kvzalloc(priv->rx_cfg.num_queues * sizeof(*priv->rx), rx_cfg 439 drivers/net/ethernet/google/gve/gve_main.c for (i = 0; i < priv->rx_cfg.num_queues; i++) { rx_cfg 476 drivers/net/ethernet/google/gve/gve_main.c for (i = 0; i < priv->rx_cfg.num_queues; i++) { rx_cfg 507 drivers/net/ethernet/google/gve/gve_main.c for (i = 0; i < priv->rx_cfg.num_queues; i++) { rx_cfg 686 drivers/net/ethernet/google/gve/gve_main.c err = netif_set_real_num_rx_queues(dev, priv->rx_cfg.num_queues); rx_cfg 772 drivers/net/ethernet/google/gve/gve_main.c priv->rx_cfg = new_rx_config; rx_cfg 782 drivers/net/ethernet/google/gve/gve_main.c priv->rx_cfg = new_rx_config; rx_cfg 809 drivers/net/ethernet/google/gve/gve_main.c for (idx = 0; idx < priv->rx_cfg.num_queues; idx++) { rx_cfg 837 drivers/net/ethernet/google/gve/gve_main.c for (idx = 0; idx < priv->rx_cfg.num_queues; idx++) { rx_cfg 955 drivers/net/ethernet/google/gve/gve_main.c priv->rx_cfg.max_queues = rx_cfg 956 drivers/net/ethernet/google/gve/gve_main.c min_t(int, priv->rx_cfg.max_queues, priv->num_ntfy_blks / 2); rx_cfg 959 drivers/net/ethernet/google/gve/gve_main.c priv->rx_cfg.num_queues = priv->rx_cfg.max_queues; rx_cfg 963 drivers/net/ethernet/google/gve/gve_main.c priv->rx_cfg.num_queues = min_t(int, priv->default_num_queues, rx_cfg 964 drivers/net/ethernet/google/gve/gve_main.c priv->rx_cfg.num_queues); rx_cfg 968 drivers/net/ethernet/google/gve/gve_main.c priv->tx_cfg.num_queues, priv->rx_cfg.num_queues); rx_cfg 970 drivers/net/ethernet/google/gve/gve_main.c priv->tx_cfg.max_queues, priv->rx_cfg.max_queues); rx_cfg 1162 drivers/net/ethernet/google/gve/gve_main.c priv->rx_cfg.max_queues = max_rx_queues; rx_cfg 185 drivers/net/ethernet/google/gve/gve_rx.c for (i = 0; i < priv->rx_cfg.num_queues; i++) { rx_cfg 208 drivers/net/ethernet/google/gve/gve_rx.c for (i = 0; i < priv->rx_cfg.num_queues; i++) rx_cfg 444 drivers/net/ethernet/marvell/octeontx2/af/mbox.h u64 rx_cfg; /* See NIX_AF_LF(0..127)_RX_CFG */ rx_cfg 897 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c rvu_write64(rvu, blkaddr, NIX_AF_LFX_RX_CFG(nixlf), req->rx_cfg); rx_cfg 1274 drivers/net/ethernet/micrel/ksz884x.c u32 rx_cfg; rx_cfg 3206 drivers/net/ethernet/micrel/ksz884x.c u32 rx_cfg; rx_cfg 3209 drivers/net/ethernet/micrel/ksz884x.c rx_cfg = hw->rx_cfg; rx_cfg 3212 drivers/net/ethernet/micrel/ksz884x.c hw->rx_cfg |= DMA_RX_FLOW_ENABLE; rx_cfg 3214 drivers/net/ethernet/micrel/ksz884x.c hw->rx_cfg &= ~DMA_RX_FLOW_ENABLE; rx_cfg 3220 drivers/net/ethernet/micrel/ksz884x.c if (rx_cfg != hw->rx_cfg) rx_cfg 3221 drivers/net/ethernet/micrel/ksz884x.c writel(hw->rx_cfg, hw->io + KS_DMA_RX_CTRL); rx_cfg 3776 drivers/net/ethernet/micrel/ksz884x.c hw->rx_cfg = (DMA_RX_BROADCAST | DMA_RX_UNICAST | rx_cfg 3778 drivers/net/ethernet/micrel/ksz884x.c hw->rx_cfg |= KS884X_DMA_RX_MULTICAST; rx_cfg 3781 drivers/net/ethernet/micrel/ksz884x.c hw->rx_cfg |= (DMA_RX_CSUM_TCP | DMA_RX_CSUM_IP); rx_cfg 3784 drivers/net/ethernet/micrel/ksz884x.c hw->rx_cfg |= DMA_RX_ALL_MULTICAST; rx_cfg 3786 drivers/net/ethernet/micrel/ksz884x.c hw->rx_cfg |= DMA_RX_PROMISCUOUS; rx_cfg 3885 drivers/net/ethernet/micrel/ksz884x.c writel(hw->rx_cfg, hw->io + KS_DMA_RX_CTRL); rx_cfg 3909 drivers/net/ethernet/micrel/ksz884x.c writel((hw->rx_cfg & ~DMA_RX_ENABLE), hw->io + KS_DMA_RX_CTRL); rx_cfg 4203 drivers/net/ethernet/micrel/ksz884x.c hw->rx_cfg |= DMA_RX_ALL_MULTICAST; rx_cfg 4205 drivers/net/ethernet/micrel/ksz884x.c hw->rx_cfg &= ~DMA_RX_ALL_MULTICAST; rx_cfg 4224 drivers/net/ethernet/micrel/ksz884x.c hw->rx_cfg |= DMA_RX_PROMISCUOUS; rx_cfg 4226 drivers/net/ethernet/micrel/ksz884x.c hw->rx_cfg &= ~DMA_RX_PROMISCUOUS; rx_cfg 5017 drivers/net/ethernet/micrel/ksz884x.c if (hw->rx_cfg & (DMA_RX_CSUM_UDP | DMA_RX_CSUM_TCP)) rx_cfg 5213 drivers/net/ethernet/micrel/ksz884x.c if (hw->enabled && (hw->rx_cfg & DMA_RX_ENABLE)) { rx_cfg 5424 drivers/net/ethernet/micrel/ksz884x.c hw->rx_cfg |= DMA_RX_ERROR; rx_cfg 5427 drivers/net/ethernet/micrel/ksz884x.c hw->rx_cfg &= ~DMA_RX_ERROR; rx_cfg 6335 drivers/net/ethernet/micrel/ksz884x.c (hw->rx_cfg & DMA_RX_FLOW_ENABLE) ? 1 : 0; rx_cfg 6589 drivers/net/ethernet/micrel/ksz884x.c hw->rx_cfg |= DMA_RX_CSUM_TCP | DMA_RX_CSUM_IP; rx_cfg 6591 drivers/net/ethernet/micrel/ksz884x.c hw->rx_cfg &= ~(DMA_RX_CSUM_TCP | DMA_RX_CSUM_IP); rx_cfg 6594 drivers/net/ethernet/micrel/ksz884x.c writel(hw->rx_cfg, hw->io + KS_DMA_RX_CTRL); rx_cfg 697 drivers/net/ethernet/neterion/s2io.c struct rx_ring_config *rx_cfg = &config->rx_cfg[i]; rx_cfg 700 drivers/net/ethernet/neterion/s2io.c if (rx_cfg->num_rxd % (rxd_count[nic->rxd_mode] + 1)) { rx_cfg 706 drivers/net/ethernet/neterion/s2io.c size += rx_cfg->num_rxd; rx_cfg 707 drivers/net/ethernet/neterion/s2io.c ring->block_count = rx_cfg->num_rxd / rx_cfg 709 drivers/net/ethernet/neterion/s2io.c ring->pkt_cnt = rx_cfg->num_rxd - ring->block_count; rx_cfg 717 drivers/net/ethernet/neterion/s2io.c struct rx_ring_config *rx_cfg = &config->rx_cfg[i]; rx_cfg 722 drivers/net/ethernet/neterion/s2io.c ring->rx_curr_get_info.ring_len = rx_cfg->num_rxd - 1; rx_cfg 725 drivers/net/ethernet/neterion/s2io.c ring->rx_curr_put_info.ring_len = rx_cfg->num_rxd - 1; rx_cfg 729 drivers/net/ethernet/neterion/s2io.c blk_cnt = rx_cfg->num_rxd / (rxd_count[nic->rxd_mode] + 1); rx_cfg 789 drivers/net/ethernet/neterion/s2io.c struct rx_ring_config *rx_cfg = &config->rx_cfg[i]; rx_cfg 792 drivers/net/ethernet/neterion/s2io.c blk_cnt = rx_cfg->num_rxd / rx_cfg 954 drivers/net/ethernet/neterion/s2io.c struct rx_ring_config *rx_cfg = &config->rx_cfg[i]; rx_cfg 957 drivers/net/ethernet/neterion/s2io.c blk_cnt = rx_cfg->num_rxd / rx_cfg 1340 drivers/net/ethernet/neterion/s2io.c struct rx_ring_config *rx_cfg = &config->rx_cfg[i]; rx_cfg 1342 drivers/net/ethernet/neterion/s2io.c val64 |= vBIT(rx_cfg->ring_priority, (5 + (i * 8)), 3); rx_cfg 5469 drivers/net/ethernet/neterion/s2io.c rx_desc_count += sp->config.rx_cfg[i].num_rxd; rx_cfg 6878 drivers/net/ethernet/neterion/s2io.c struct rx_ring_config *rx_cfg = &config->rx_cfg[i]; rx_cfg 6881 drivers/net/ethernet/neterion/s2io.c blk_cnt = rx_cfg->num_rxd / (rxd_count[sp->rxd_mode] + 1); rx_cfg 7807 drivers/net/ethernet/neterion/s2io.c struct rx_ring_config *rx_cfg = &config->rx_cfg[i]; rx_cfg 7810 drivers/net/ethernet/neterion/s2io.c rx_cfg->num_rxd = rx_ring_sz[i] * (rxd_count[sp->rxd_mode] + 1); rx_cfg 7811 drivers/net/ethernet/neterion/s2io.c rx_cfg->ring_priority = i; rx_cfg 7820 drivers/net/ethernet/neterion/s2io.c struct rx_ring_config *rx_cfg = &config->rx_cfg[i]; rx_cfg 7822 drivers/net/ethernet/neterion/s2io.c rx_cfg->ring_org = RING_ORG_BUFF1; rx_cfg 7823 drivers/net/ethernet/neterion/s2io.c rx_cfg->f_no_snoop = (NO_SNOOP_RXD | NO_SNOOP_RXD_BUFFER); rx_cfg 450 drivers/net/ethernet/neterion/s2io.h struct rx_ring_config rx_cfg[MAX_RX_RINGS]; /*Per-Rx Ring config */ rx_cfg 393 drivers/net/ethernet/sun/niu.c u32 tx_cfg, rx_cfg; rx_cfg 397 drivers/net/ethernet/sun/niu.c rx_cfg = (PLL_RX_CFG_ENRX | PLL_RX_CFG_TERM_0P8VDDT | rx_cfg 408 drivers/net/ethernet/sun/niu.c rx_cfg |= PLL_RX_CFG_ENTEST; rx_cfg 419 drivers/net/ethernet/sun/niu.c int err = esr2_set_rx_cfg(np, i, rx_cfg); rx_cfg 433 drivers/net/ethernet/sun/niu.c u32 tx_cfg, rx_cfg; rx_cfg 439 drivers/net/ethernet/sun/niu.c rx_cfg = (PLL_RX_CFG_ENRX | PLL_RX_CFG_TERM_0P8VDDT | rx_cfg 444 drivers/net/ethernet/sun/niu.c rx_cfg |= PLL_RX_CFG_EQ_LP_ADAPTIVE; rx_cfg 453 drivers/net/ethernet/sun/niu.c rx_cfg |= PLL_RX_CFG_ENTEST; rx_cfg 487 drivers/net/ethernet/sun/niu.c err = esr2_set_rx_cfg(np, i, rx_cfg); rx_cfg 527 drivers/net/ethernet/sun/niu.c u32 tx_cfg, rx_cfg, pll_cfg, pll_sts; rx_cfg 534 drivers/net/ethernet/sun/niu.c rx_cfg = (PLL_RX_CFG_ENRX | PLL_RX_CFG_TERM_0P8VDDT | rx_cfg 545 drivers/net/ethernet/sun/niu.c rx_cfg |= PLL_RX_CFG_ENTEST; rx_cfg 579 drivers/net/ethernet/sun/niu.c err = esr2_set_rx_cfg(np, i, rx_cfg); rx_cfg 162 drivers/net/wireless/intel/iwlwifi/pcie/ctxt-info.c struct iwl_context_info_rbd_cfg *rx_cfg; rx_cfg 204 drivers/net/wireless/intel/iwlwifi/pcie/ctxt-info.c rx_cfg = &ctxt_info->rbd_cfg; rx_cfg 205 drivers/net/wireless/intel/iwlwifi/pcie/ctxt-info.c rx_cfg->free_rbd_addr = cpu_to_le64(trans_pcie->rxq->bd_dma); rx_cfg 206 drivers/net/wireless/intel/iwlwifi/pcie/ctxt-info.c rx_cfg->used_rbd_addr = cpu_to_le64(trans_pcie->rxq->used_bd_dma); rx_cfg 207 drivers/net/wireless/intel/iwlwifi/pcie/ctxt-info.c rx_cfg->status_wr_ptr = cpu_to_le64(trans_pcie->rxq->rb_stts_dma); rx_cfg 380 drivers/soc/qcom/qcom-geni-se.c bool msb_to_lsb, bool tx_cfg, bool rx_cfg) rx_cfg 417 drivers/soc/qcom/qcom-geni-se.c if (rx_cfg) { rx_cfg 89 drivers/staging/pi433/pi433_if.c struct pi433_rx_cfg rx_cfg; rx_cfg 158 drivers/staging/pi433/pi433_if.c rf69_set_rx_cfg(struct pi433_device *dev, struct pi433_rx_cfg *rx_cfg) rx_cfg 164 drivers/staging/pi433/pi433_if.c ret = rf69_set_frequency(dev->spi, rx_cfg->frequency); rx_cfg 167 drivers/staging/pi433/pi433_if.c ret = rf69_set_bit_rate(dev->spi, rx_cfg->bit_rate); rx_cfg 170 drivers/staging/pi433/pi433_if.c ret = rf69_set_modulation(dev->spi, rx_cfg->modulation); rx_cfg 173 drivers/staging/pi433/pi433_if.c ret = rf69_set_antenna_impedance(dev->spi, rx_cfg->antenna_impedance); rx_cfg 176 drivers/staging/pi433/pi433_if.c ret = rf69_set_rssi_threshold(dev->spi, rx_cfg->rssi_threshold); rx_cfg 179 drivers/staging/pi433/pi433_if.c ret = rf69_set_ook_threshold_dec(dev->spi, rx_cfg->threshold_decrement); rx_cfg 182 drivers/staging/pi433/pi433_if.c ret = rf69_set_bandwidth(dev->spi, rx_cfg->bw_mantisse, rx_cfg 183 drivers/staging/pi433/pi433_if.c rx_cfg->bw_exponent); rx_cfg 186 drivers/staging/pi433/pi433_if.c ret = rf69_set_bandwidth_during_afc(dev->spi, rx_cfg->bw_mantisse, rx_cfg 187 drivers/staging/pi433/pi433_if.c rx_cfg->bw_exponent); rx_cfg 190 drivers/staging/pi433/pi433_if.c ret = rf69_set_dagc(dev->spi, rx_cfg->dagc); rx_cfg 194 drivers/staging/pi433/pi433_if.c dev->rx_bytes_to_drop = rx_cfg->bytes_to_drop; rx_cfg 198 drivers/staging/pi433/pi433_if.c if (rx_cfg->enable_sync == OPTION_ON) { rx_cfg 216 drivers/staging/pi433/pi433_if.c if (rx_cfg->enable_length_byte == OPTION_ON) { rx_cfg 226 drivers/staging/pi433/pi433_if.c rx_cfg->enable_address_filtering); rx_cfg 230 drivers/staging/pi433/pi433_if.c if (rx_cfg->enable_crc == OPTION_ON) { rx_cfg 241 drivers/staging/pi433/pi433_if.c ret = rf69_set_sync_size(dev->spi, rx_cfg->sync_length); rx_cfg 244 drivers/staging/pi433/pi433_if.c if (rx_cfg->enable_length_byte == OPTION_ON) { rx_cfg 248 drivers/staging/pi433/pi433_if.c } else if (rx_cfg->fixed_message_length != 0) { rx_cfg 249 drivers/staging/pi433/pi433_if.c payload_length = rx_cfg->fixed_message_length; rx_cfg 250 drivers/staging/pi433/pi433_if.c if (rx_cfg->enable_length_byte == OPTION_ON) rx_cfg 252 drivers/staging/pi433/pi433_if.c if (rx_cfg->enable_address_filtering != filtering_off) rx_cfg 264 drivers/staging/pi433/pi433_if.c if (rx_cfg->enable_sync == OPTION_ON) { rx_cfg 265 drivers/staging/pi433/pi433_if.c ret = rf69_set_sync_values(dev->spi, rx_cfg->sync_pattern); rx_cfg 269 drivers/staging/pi433/pi433_if.c if (rx_cfg->enable_address_filtering != filtering_off) { rx_cfg 270 drivers/staging/pi433/pi433_if.c ret = rf69_set_node_address(dev->spi, rx_cfg->node_address); rx_cfg 274 drivers/staging/pi433/pi433_if.c rx_cfg->broadcast_address); rx_cfg 372 drivers/staging/pi433/pi433_if.c retval = rf69_set_rx_cfg(dev, &dev->rx_cfg); rx_cfg 461 drivers/staging/pi433/pi433_if.c if (dev->rx_cfg.fixed_message_length != 0) { rx_cfg 462 drivers/staging/pi433/pi433_if.c if (dev->rx_cfg.fixed_message_length > dev->rx_buffer_size) { rx_cfg 466 drivers/staging/pi433/pi433_if.c bytes_total = dev->rx_cfg.fixed_message_length; rx_cfg 476 drivers/staging/pi433/pi433_if.c if (dev->rx_cfg.enable_length_byte == OPTION_ON) { rx_cfg 493 drivers/staging/pi433/pi433_if.c if (dev->rx_cfg.enable_address_filtering != filtering_off) { rx_cfg 903 drivers/staging/pi433/pi433_if.c if (copy_to_user(argp, &device->rx_cfg, rx_cfg 916 drivers/staging/pi433/pi433_if.c if (copy_from_user(&device->rx_cfg, argp, rx_cfg 469 drivers/usb/gadget/udc/bcm63xx_udc.c const struct iudma_ch_cfg *rx_cfg = &iudma_defaults[i]; rx_cfg 475 drivers/usb/gadget/udc/bcm63xx_udc.c ((rx_fifo_slot + rx_cfg->n_fifo_slots - 1) << rx_cfg 477 drivers/usb/gadget/udc/bcm63xx_udc.c rx_fifo_slot += rx_cfg->n_fifo_slots; rx_cfg 480 drivers/usb/gadget/udc/bcm63xx_udc.c is_hs ? rx_cfg->max_pkt_hs : rx_cfg->max_pkt_fs, rx_cfg 104 include/linux/hsi/hsi.h struct hsi_config rx_cfg; rx_cfg 129 include/linux/hsi/hsi.h struct hsi_config rx_cfg; rx_cfg 224 include/linux/hsi/hsi.h struct hsi_config rx_cfg; rx_cfg 398 include/linux/qcom-geni-se.h bool msb_to_lsb, bool tx_cfg, bool rx_cfg);