rx_buff 106 arch/mips/include/asm/ip32/mace.h volatile unsigned long rx_buff; rx_buff 205 drivers/isdn/mISDN/dsp.h u8 rx_buff[CMX_BUFF_SIZE]; rx_buff 247 drivers/isdn/mISDN/dsp_cmx.c memset(dsp->rx_buff, dsp_silence, sizeof(dsp->rx_buff)); rx_buff 1255 drivers/isdn/mISDN/dsp_cmx.c memset(dsp->rx_buff, dsp_silence, sizeof(dsp->rx_buff)); rx_buff 1276 drivers/isdn/mISDN/dsp_cmx.c memset(dsp->rx_buff, dsp_silence, sizeof(dsp->rx_buff)); rx_buff 1288 drivers/isdn/mISDN/dsp_cmx.c d = dsp->rx_buff; rx_buff 1369 drivers/isdn/mISDN/dsp_cmx.c q = dsp->rx_buff; /* received data */ rx_buff 1465 drivers/isdn/mISDN/dsp_cmx.c o_q = other->rx_buff; /* received data */ rx_buff 1715 drivers/isdn/mISDN/dsp_cmx.c q = dsp->rx_buff; rx_buff 1738 drivers/isdn/mISDN/dsp_cmx.c p = dsp->rx_buff; rx_buff 842 drivers/isdn/mISDN/dsp_core.c memset(dsp->rx_buff, 0, sizeof(dsp->rx_buff)); rx_buff 115 drivers/mailbox/ti-msgmgr.c u32 *rx_buff; rx_buff 247 drivers/mailbox/ti-msgmgr.c message.buf = (u8 *)qinst->rx_buff; rx_buff 263 drivers/mailbox/ti-msgmgr.c for (data_reg = qinst->queue_buff_start, word_data = qinst->rx_buff, rx_buff 481 drivers/mailbox/ti-msgmgr.c qinst->rx_buff = kzalloc(d->max_message_size, GFP_KERNEL); rx_buff 482 drivers/mailbox/ti-msgmgr.c if (!qinst->rx_buff) rx_buff 487 drivers/mailbox/ti-msgmgr.c kfree(qinst->rx_buff); rx_buff 505 drivers/mailbox/ti-msgmgr.c kfree(qinst->rx_buff); rx_buff 91 drivers/net/ethernet/amd/ariadne.c volatile u_short *rx_buff[RX_RING_SIZE]; rx_buff 103 drivers/net/ethernet/amd/ariadne.c u_short rx_buff[RX_RING_SIZE][PKT_BUF_SIZE / sizeof(u_short)]; rx_buff 148 drivers/net/ethernet/amd/ariadne.c offsetof(struct lancedata, rx_buff[i])); rx_buff 150 drivers/net/ethernet/amd/ariadne.c offsetof(struct lancedata, rx_buff[i])) | rx_buff 155 drivers/net/ethernet/amd/ariadne.c priv->rx_buff[i] = lancedata->rx_buff[i]; rx_buff 157 drivers/net/ethernet/amd/ariadne.c i, &lancedata->rx_ring[i], lancedata->rx_buff[i]); rx_buff 213 drivers/net/ethernet/amd/ariadne.c (const void *)priv->rx_buff[entry], rx_buff 871 drivers/net/ethernet/amd/lance.c void *rx_buff; rx_buff 876 drivers/net/ethernet/amd/lance.c rx_buff = skb->data; rx_buff 878 drivers/net/ethernet/amd/lance.c rx_buff = kmalloc(PKT_BUF_SZ, GFP_DMA | gfp); rx_buff 879 drivers/net/ethernet/amd/lance.c if (rx_buff == NULL) rx_buff 882 drivers/net/ethernet/amd/lance.c lp->rx_ring[i].base = (u32)isa_virt_to_bus(rx_buff) | 0x80000000; rx_buff 152 drivers/net/ethernet/arc/emac.h struct buffer_state rx_buff[RX_BD_NUM]; rx_buff 183 drivers/net/ethernet/arc/emac_main.c struct buffer_state *rx_buff = &priv->rx_buff[*last_rx_bd]; rx_buff 241 drivers/net/ethernet/arc/emac_main.c dma_unmap_single(&ndev->dev, dma_unmap_addr(rx_buff, addr), rx_buff 242 drivers/net/ethernet/arc/emac_main.c dma_unmap_len(rx_buff, len), DMA_FROM_DEVICE); rx_buff 247 drivers/net/ethernet/arc/emac_main.c skb_put(rx_buff->skb, pktlen); rx_buff 248 drivers/net/ethernet/arc/emac_main.c rx_buff->skb->dev = ndev; rx_buff 249 drivers/net/ethernet/arc/emac_main.c rx_buff->skb->protocol = eth_type_trans(rx_buff->skb, ndev); rx_buff 251 drivers/net/ethernet/arc/emac_main.c netif_receive_skb(rx_buff->skb); rx_buff 253 drivers/net/ethernet/arc/emac_main.c rx_buff->skb = skb; rx_buff 254 drivers/net/ethernet/arc/emac_main.c dma_unmap_addr_set(rx_buff, addr, addr); rx_buff 255 drivers/net/ethernet/arc/emac_main.c dma_unmap_len_set(rx_buff, len, EMAC_BUFFER_SIZE); rx_buff 442 drivers/net/ethernet/arc/emac_main.c struct buffer_state *rx_buff = &priv->rx_buff[*last_rx_bd]; rx_buff 444 drivers/net/ethernet/arc/emac_main.c rx_buff->skb = netdev_alloc_skb_ip_align(ndev, rx_buff 446 drivers/net/ethernet/arc/emac_main.c if (unlikely(!rx_buff->skb)) rx_buff 449 drivers/net/ethernet/arc/emac_main.c addr = dma_map_single(&ndev->dev, (void *)rx_buff->skb->data, rx_buff 453 drivers/net/ethernet/arc/emac_main.c dev_kfree_skb(rx_buff->skb); rx_buff 456 drivers/net/ethernet/arc/emac_main.c dma_unmap_addr_set(rx_buff, addr, addr); rx_buff 457 drivers/net/ethernet/arc/emac_main.c dma_unmap_len_set(rx_buff, len, EMAC_BUFFER_SIZE); rx_buff 587 drivers/net/ethernet/arc/emac_main.c struct buffer_state *rx_buff = &priv->rx_buff[i]; rx_buff 589 drivers/net/ethernet/arc/emac_main.c if (rx_buff->skb) { rx_buff 591 drivers/net/ethernet/arc/emac_main.c dma_unmap_addr(rx_buff, addr), rx_buff 592 drivers/net/ethernet/arc/emac_main.c dma_unmap_len(rx_buff, len), rx_buff 596 drivers/net/ethernet/arc/emac_main.c dev_kfree_skb_irq(rx_buff->skb); rx_buff 601 drivers/net/ethernet/arc/emac_main.c rx_buff->skb = NULL; rx_buff 1125 drivers/net/ethernet/freescale/gianfar.c struct gfar_rx_buff *rxb = &rx_queue->rx_buff[i]; rx_buff 1141 drivers/net/ethernet/freescale/gianfar.c kfree(rx_queue->rx_buff); rx_buff 1142 drivers/net/ethernet/freescale/gianfar.c rx_queue->rx_buff = NULL; rx_buff 1167 drivers/net/ethernet/freescale/gianfar.c if (rx_queue->rx_buff) rx_buff 1276 drivers/net/ethernet/freescale/gianfar.c rxb = &rx_queue->rx_buff[i]; rx_buff 1298 drivers/net/ethernet/freescale/gianfar.c rxb = rx_queue->rx_buff; rx_buff 1422 drivers/net/ethernet/freescale/gianfar.c rx_queue->rx_buff = kcalloc(rx_queue->rx_ring_size, rx_buff 1423 drivers/net/ethernet/freescale/gianfar.c sizeof(*rx_queue->rx_buff), rx_buff 1425 drivers/net/ethernet/freescale/gianfar.c if (!rx_queue->rx_buff) rx_buff 2420 drivers/net/ethernet/freescale/gianfar.c new_rxb = &rxq->rx_buff[nta]; rx_buff 2438 drivers/net/ethernet/freescale/gianfar.c struct gfar_rx_buff *rxb = &rx_queue->rx_buff[rx_queue->next_to_clean]; rx_buff 1008 drivers/net/ethernet/freescale/gianfar.h struct gfar_rx_buff *rx_buff __aligned(SMP_CACHE_BYTES); rx_buff 276 drivers/net/ethernet/ibm/ibmvnic.c if (pool->rx_buff[index].skb) rx_buff 284 drivers/net/ethernet/ibm/ibmvnic.c pool->rx_buff[index].data = dst; rx_buff 287 drivers/net/ethernet/ibm/ibmvnic.c pool->rx_buff[index].dma = dma_addr; rx_buff 288 drivers/net/ethernet/ibm/ibmvnic.c pool->rx_buff[index].skb = skb; rx_buff 289 drivers/net/ethernet/ibm/ibmvnic.c pool->rx_buff[index].pool_index = pool->index; rx_buff 290 drivers/net/ethernet/ibm/ibmvnic.c pool->rx_buff[index].size = pool->buff_size; rx_buff 295 drivers/net/ethernet/ibm/ibmvnic.c cpu_to_be64((u64)&pool->rx_buff[index]); rx_buff 325 drivers/net/ethernet/ibm/ibmvnic.c pool->rx_buff[index].skb = NULL; rx_buff 446 drivers/net/ethernet/ibm/ibmvnic.c memset(rx_pool->rx_buff, 0, rx_buff 474 drivers/net/ethernet/ibm/ibmvnic.c if (!rx_pool->rx_buff) rx_buff 478 drivers/net/ethernet/ibm/ibmvnic.c if (rx_pool->rx_buff[j].skb) { rx_buff 479 drivers/net/ethernet/ibm/ibmvnic.c dev_kfree_skb_any(rx_pool->rx_buff[j].skb); rx_buff 480 drivers/net/ethernet/ibm/ibmvnic.c rx_pool->rx_buff[j].skb = NULL; rx_buff 484 drivers/net/ethernet/ibm/ibmvnic.c kfree(rx_pool->rx_buff); rx_buff 536 drivers/net/ethernet/ibm/ibmvnic.c rx_pool->rx_buff = kcalloc(rx_pool->size, rx_buff 539 drivers/net/ethernet/ibm/ibmvnic.c if (!rx_pool->rx_buff) { rx_buff 1111 drivers/net/ethernet/ibm/ibmvnic.c struct ibmvnic_rx_buff *rx_buff; rx_buff 1125 drivers/net/ethernet/ibm/ibmvnic.c if (!rx_pool || !rx_pool->rx_buff) rx_buff 1130 drivers/net/ethernet/ibm/ibmvnic.c rx_buff = &rx_pool->rx_buff[j]; rx_buff 1131 drivers/net/ethernet/ibm/ibmvnic.c if (rx_buff && rx_buff->skb) { rx_buff 1132 drivers/net/ethernet/ibm/ibmvnic.c dev_kfree_skb_any(rx_buff->skb); rx_buff 1133 drivers/net/ethernet/ibm/ibmvnic.c rx_buff->skb = NULL; rx_buff 2201 drivers/net/ethernet/ibm/ibmvnic.c struct ibmvnic_rx_buff *rx_buff) rx_buff 2203 drivers/net/ethernet/ibm/ibmvnic.c struct ibmvnic_rx_pool *pool = &adapter->rx_pool[rx_buff->pool_index]; rx_buff 2205 drivers/net/ethernet/ibm/ibmvnic.c rx_buff->skb = NULL; rx_buff 2207 drivers/net/ethernet/ibm/ibmvnic.c pool->free_map[pool->next_alloc] = (int)(rx_buff - pool->rx_buff); rx_buff 2223 drivers/net/ethernet/ibm/ibmvnic.c struct ibmvnic_rx_buff *rx_buff; rx_buff 2239 drivers/net/ethernet/ibm/ibmvnic.c rx_buff = rx_buff 2248 drivers/net/ethernet/ibm/ibmvnic.c dev_kfree_skb_any(rx_buff->skb); rx_buff 2249 drivers/net/ethernet/ibm/ibmvnic.c remove_buff_from_pool(adapter, rx_buff); rx_buff 2251 drivers/net/ethernet/ibm/ibmvnic.c } else if (!rx_buff->skb) { rx_buff 2254 drivers/net/ethernet/ibm/ibmvnic.c remove_buff_from_pool(adapter, rx_buff); rx_buff 2261 drivers/net/ethernet/ibm/ibmvnic.c skb = rx_buff->skb; rx_buff 2262 drivers/net/ethernet/ibm/ibmvnic.c skb_copy_to_linear_data(skb, rx_buff->data + offset, rx_buff 2275 drivers/net/ethernet/ibm/ibmvnic.c remove_buff_from_pool(adapter, rx_buff); rx_buff 919 drivers/net/ethernet/ibm/ibmvnic.h struct ibmvnic_rx_buff *rx_buff; rx_buff 348 drivers/net/hamradio/scc.c if(scc->rx_buff != NULL) /* did we receive something? */ rx_buff 351 drivers/net/hamradio/scc.c dev_kfree_skb_irq(scc->rx_buff); rx_buff 352 drivers/net/hamradio/scc.c scc->rx_buff = NULL; rx_buff 529 drivers/net/hamradio/scc.c skb = scc->rx_buff; rx_buff 543 drivers/net/hamradio/scc.c scc->rx_buff = skb; rx_buff 553 drivers/net/hamradio/scc.c scc->rx_buff = NULL; rx_buff 574 drivers/net/hamradio/scc.c skb = scc->rx_buff; rx_buff 583 drivers/net/hamradio/scc.c scc->rx_buff = skb = NULL; rx_buff 595 drivers/net/hamradio/scc.c scc->rx_buff = NULL; rx_buff 599 drivers/net/hamradio/scc.c scc->rx_buff = NULL; rx_buff 173 drivers/net/wireless/ath/wil6210/txrx_edma.c struct wil_rx_buff *rx_buff; rx_buff 204 drivers/net/wireless/ath/wil6210/txrx_edma.c rx_buff = list_first_entry(free, struct wil_rx_buff, list); rx_buff 205 drivers/net/wireless/ath/wil6210/txrx_edma.c buff_id = rx_buff->id; rx_buff 208 drivers/net/wireless/ath/wil6210/txrx_edma.c list_move(&rx_buff->list, active); rx_buff 287 drivers/net/wireless/ath/wil6210/txrx_edma.c struct wil_rx_buff *rx_buff = rx_buff 289 drivers/net/wireless/ath/wil6210/txrx_edma.c struct sk_buff *skb = rx_buff->skb; rx_buff 292 drivers/net/wireless/ath/wil6210/txrx_edma.c wil_err(wil, "No Rx skb at buff_id %d\n", rx_buff->id); rx_buff 294 drivers/net/wireless/ath/wil6210/txrx_edma.c rx_buff->skb = NULL; rx_buff 302 drivers/net/wireless/ath/wil6210/txrx_edma.c list_move(&rx_buff->list, &wil->rx_buff_mgmt.free); rx_buff 167 drivers/ntb/ntb_transport.c void *rx_buff; rx_buff 503 drivers/ntb/ntb_transport.c "rx_buff - \t0x%p\n", qp->rx_buff); rx_buff 642 drivers/ntb/ntb_transport.c qp->rx_buff = mw->virt_addr + rx_size * (qp_num / mw_count); rx_buff 645 drivers/ntb/ntb_transport.c qp->remote_rx_info = qp->rx_buff + rx_size; rx_buff 673 drivers/ntb/ntb_transport.c void *offset = (qp->rx_buff + qp->rx_max_frame * (i + 1) - rx_buff 1490 drivers/ntb/ntb_transport.c void *offset = qp->rx_buff + qp->rx_max_frame * rx_buff 1622 drivers/ntb/ntb_transport.c offset = qp->rx_buff + qp->rx_max_frame * qp->rx_index; rx_buff 177 drivers/s390/net/netiucv.c struct sk_buff *rx_buff; rx_buff 692 drivers/s390/net/netiucv.c conn->rx_buff->data = conn->rx_buff->head; rx_buff 693 drivers/s390/net/netiucv.c skb_reset_tail_pointer(conn->rx_buff); rx_buff 694 drivers/s390/net/netiucv.c conn->rx_buff->len = 0; rx_buff 695 drivers/s390/net/netiucv.c rc = iucv_message_receive(conn->path, msg, 0, conn->rx_buff->data, rx_buff 702 drivers/s390/net/netiucv.c netiucv_unpack_skb(conn, conn->rx_buff); rx_buff 1859 drivers/s390/net/netiucv.c conn->rx_buff = alloc_skb(conn->max_buffsize, GFP_KERNEL | GFP_DMA); rx_buff 1860 drivers/s390/net/netiucv.c if (!conn->rx_buff) rx_buff 1890 drivers/s390/net/netiucv.c kfree_skb(conn->rx_buff); rx_buff 1917 drivers/s390/net/netiucv.c kfree_skb(conn->rx_buff); rx_buff 33 drivers/spi/spi-tle62x0.c unsigned char rx_buff[4]; rx_buff 62 drivers/spi/spi-tle62x0.c .rx_buf = st->rx_buff, rx_buff 101 drivers/spi/spi-tle62x0.c unsigned char *buff = st->rx_buff; rx_buff 75 include/linux/scc.h struct sk_buff *rx_buff; /* pointer to frame currently received */