Searched refs:rx (Results 1 - 200 of 1709) sorted by relevance

123456789

/linux-4.4.14/net/ieee802154/6lowpan/
H A DMakefile3 ieee802154_6lowpan-y := core.o rx.o reassembly.o tx.o
/linux-4.4.14/arch/arm/include/debug/
H A D8250.S18 .macro store, rd, rx:vararg
20 str \rd, \rx
24 .macro load, rd, rx:vararg
25 ldr \rd, \rx
29 .macro store, rd, rx:vararg
30 strb \rd, \rx
33 .macro load, rd, rx:vararg
34 ldrb \rd, \rx
40 .macro senduart,rd,rx
41 store \rd, [\rx, #UART_TX << UART_SHIFT]
44 .macro busyuart,rd,rx
45 1002: load \rd, [\rx, #UART_LSR << UART_SHIFT]
51 .macro waituart,rd,rx
53 1001: load \rd, [\rx, #UART_MSR << UART_SHIFT]
H A Dsamsung.S16 .macro fifo_level_s5pv210 rd, rx
17 ldr \rd, [\rx, # S3C2410_UFSTAT]
21 .macro fifo_full_s5pv210 rd, rx
22 ldr \rd, [\rx, # S3C2410_UFSTAT]
29 .macro fifo_level_s3c2440 rd, rx
30 ldr \rd, [\rx, # S3C2410_UFSTAT]
38 .macro fifo_full_s3c2440 rd, rx
39 ldr \rd, [\rx, # S3C2410_UFSTAT]
47 .macro senduart,rd,rx
48 strb \rd, [\rx, # S3C2410_UTXH]
51 .macro busyuart, rd, rx
52 ldr \rd, [\rx, # S3C2410_UFCON]
57 fifo_full \rd, \rx
63 ldr \rd, [\rx, # S3C2410_UTRSTAT]
70 .macro waituart,rd,rx
71 ldr \rd, [\rx, # S3C2410_UFCON]
76 fifo_level \rd, \rx
82 ldr \rd, [\rx, # S3C2410_UTRSTAT]
H A Dsirf.S26 .macro senduart,rd,rx
27 str \rd, [\rx, #SIRF_LLUART_TXFIFO_DATA]
30 .macro busyuart,rd,rx
33 .macro waituart,rd,rx
34 1001: ldr \rd, [\rx, #SIRF_LLUART_TXFIFO_STATUS]
H A Defm32.S19 .macro addruart, rx, tmp, tmp2
20 ldr \rx, =(CONFIG_DEBUG_UART_PHYS)
28 str \tmp, [\rx, #UARTn_CMD]
31 .macro senduart,rd,rx
32 strb \rd, [\rx, #UARTn_TXDATA]
35 .macro waituart,rd,rx
36 1001: ldr \rd, [\rx, #UARTn_STATUS]
41 .macro busyuart,rd,rx
42 1001: ldr \rd, [\rx, UARTn_STATUS]
H A Dmeson.S21 .macro senduart,rd,rx
22 str \rd, [\rx, #MESON_AO_UART_WFIFO]
25 .macro busyuart,rd,rx
26 1002: ldr \rd, [\rx, #MESON_AO_UART_STATUS]
31 .macro waituart,rd,rx
32 1001: ldr \rd, [\rx, #MESON_AO_UART_STATUS]
H A Dnetx.S22 .macro senduart,rd,rx
23 str \rd, [\rx, #UART_DATA]
26 .macro busyuart,rd,rx
27 1002: ldr \rd, [\rx, #UART_FLAG]
32 .macro waituart,rd,rx
33 1001: ldr \rd, [\rx, #UART_FLAG]
H A Drenesas-scif.S35 .macro waituart, rd, rx
36 1001: ldrh \rd, [\rx, #FSR]
41 .macro senduart, rd, rx
42 strb \rd, [\rx, #FTDR]
43 ldrh \rd, [\rx, #FSR]
45 strh \rd, [\rx, #FSR]
48 .macro busyuart, rd, rx
49 1001: ldrh \rd, [\rx, #FSR]
H A Ddigicolor.S24 .macro senduart,rd,rx
25 strb \rd, [\rx, #UA0_EMI_REC]
28 .macro waituart,rd,rx
31 .macro busyuart,rd,rx
32 1001: ldrb \rd, [\rx, #UA0_STATUS]
H A Dvt8500.S24 .macro senduart,rd,rx
25 strb \rd, [\rx, #0]
28 .macro busyuart,rd,rx
29 1001: ldr \rd, [\rx, #0x1c]
34 .macro waituart,rd,rx
H A Dat91.S30 .macro senduart,rd,rx
31 strb \rd, [\rx, #(AT91_DBGU_THR)] @ Write to Transmitter Holding Register
34 .macro waituart,rd,rx
35 1001: ldr \rd, [\rx, #(AT91_DBGU_SR)] @ Read Status Register
40 .macro busyuart,rd,rx
41 1001: ldr \rd, [\rx, #(AT91_DBGU_SR)] @ Read Status Register
H A Dks8695.S26 .macro senduart, rd, rx
27 str \rd, [\rx, #KS8695_URTH] @ Write to Transmit Holding Register
30 .macro busyuart, rd, rx
31 1001: ldr \rd, [\rx, #KS8695_URLS] @ Read Line Status Register
36 .macro waituart, rd, rx
37 1001: ldr \rd, [\rx, #KS8695_URLS] @ Read Line Status Register
H A Dpl01x.S29 .macro senduart,rd,rx
30 strb \rd, [\rx, #UART01x_DR]
33 .macro waituart,rd,rx
34 1001: ldr \rd, [\rx, #UART01x_FR]
40 .macro busyuart,rd,rx
41 1001: ldr \rd, [\rx, #UART01x_FR]
H A Dsti.S47 .macro senduart,rd,rx
48 strb \rd, [\rx, #ASC_TX_BUF_OFF]
51 .macro waituart,rd,rx
52 1001: ldr \rd, [\rx, #ASC_STA_OFF]
57 .macro busyuart,rd,rx
58 1001: ldr \rd, [\rx, #ASC_STA_OFF]
H A Dclps711x.S27 .macro waituart,rd,rx
30 .macro senduart,rd,rx
31 str \rd, [\rx, #UARTDR]
34 .macro busyuart,rd,rx
35 1001: ldr \rd, [\rx, #SYSFLG]
H A Dimx.S36 .macro senduart,rd,rx
37 str \rd, [\rx, #0x40] @ TXDATA
40 .macro waituart,rd,rx
43 .macro busyuart,rd,rx
44 1002: ldr \rd, [\rx, #0x98] @ SR2
H A Dvf.S26 .macro senduart, rd, rx
27 strb \rd, [\rx, #0x7] @ Data Register
30 .macro busyuart, rd, rx
31 1001: ldrb \rd, [\rx, #0x4] @ Status Register 1
36 .macro waituart,rd,rx
H A Ds3c24xx.S24 .macro fifo_full_s3c2410 rd, rx
25 ldr \rd, [\rx, # S3C2410_UFSTAT]
29 .macro fifo_level_s3c2410 rd, rx
30 ldr \rd, [\rx, # S3C2410_UFSTAT]
H A Dsa1100.S54 .macro senduart,rd,rx
55 str \rd, [\rx, #UTDR]
58 .macro waituart,rd,rx
59 1001: ldr \rd, [\rx, #UTSR1]
64 .macro busyuart,rd,rx
65 1001: ldr \rd, [\rx, #UTSR1]
H A Dzynq.S40 .macro senduart,rd,rx
41 strb \rd, [\rx, #UART_FIFO_OFFSET] @ TXDATA
44 .macro waituart,rd,rx
45 1001: ldr \rd, [\rx, #UART_SR_OFFSET]
51 .macro busyuart,rd,rx
52 1002: ldr \rd, [\rx, #UART_SR_OFFSET] @ get status register
H A Domap2plus.S174 .macro senduart,rd,rx
175 orr \rd, \rd, \rx, lsl #24 @ preserve LSR reg offset
176 bic \rx, \rx, #0xff @ get base (THR) reg address
177 strb \rd, [\rx] @ send lower byte of rd
178 orr \rx, \rx, \rd, lsr #24 @ restore original rx (LSR)
182 .macro busyuart,rd,rx
183 1001: ldrb \rd, [\rx] @ rx contains UART_LSR address
189 .macro waituart,rd,rx
H A Dtegra.S174 .macro senduart, rd, rx
175 cmp \rx, #0
176 strneb \rd, [\rx, #UART_TX << UART_SHIFT]
180 .macro busyuart, rd, rx
181 cmp \rx, #0
183 1001: ldrb \rd, [\rx, #UART_LSR << UART_SHIFT]
190 .macro waituart, rd, rx
192 cmp \rx, #0
194 1001: ldrb \rd, [\rx, #UART_MSR << UART_SHIFT]
/linux-4.4.14/arch/unicore32/kernel/
H A Ddebug-macro.S17 .macro put_word_ocd, rd, rx=r16
18 1001: movc \rx, p1.c0, #0
19 cand.a \rx, #2
26 .macro addruart, rx
29 .macro senduart, rd, rx
30 put_word_ocd \rd, \rx
33 .macro busyuart, rd, rx
36 .macro waituart, rd, rx
47 .macro addruart,rx
48 mrc p0, #0, \rx, c1, c0
49 tst \rx, #1 @ MMU enabled?
50 moveq \rx, #0xee000000 @ physical base address
51 movne \rx, #0x6e000000 @ virtual address
59 str r1, [\rx, #UART_LCR_OFFSET]
62 str r1, [\rx, #UART_DLH_OFFSET]
64 str r1, [\rx, #UART_DLL_OFFSET]
66 str r1, [\rx, #UART_FCR_OFFSET]
68 str r1, [\rx, #UART_LCR_OFFSET]
70 str r1, [\rx, #UART_IER_OFFSET]
73 .macro senduart,rd,rx
74 str \rd, [\rx, #UART_THR_OFFSET]
77 .macro waituart,rd,rx
78 1001: ldr \rd, [\rx, #UART_LSR_OFFSET]
83 .macro busyuart,rd,rx
84 1001: ldr \rd, [\rx, #UART_LSR_OFFSET]
/linux-4.4.14/drivers/net/wimax/i2400m/
H A DMakefile13 rx.o
21 usb-rx.o \
/linux-4.4.14/net/mac80211/
H A Dwpa.h19 ieee80211_rx_h_michael_mic_verify(struct ieee80211_rx_data *rx);
24 ieee80211_crypto_tkip_decrypt(struct ieee80211_rx_data *rx);
30 ieee80211_crypto_ccmp_decrypt(struct ieee80211_rx_data *rx,
38 ieee80211_crypto_aes_cmac_decrypt(struct ieee80211_rx_data *rx);
40 ieee80211_crypto_aes_cmac_256_decrypt(struct ieee80211_rx_data *rx);
44 ieee80211_crypto_aes_gmac_decrypt(struct ieee80211_rx_data *rx);
48 ieee80211_crypto_hw_decrypt(struct ieee80211_rx_data *rx);
53 ieee80211_crypto_gcmp_decrypt(struct ieee80211_rx_data *rx);
H A Drx.c594 static void ieee80211_parse_qos(struct ieee80211_rx_data *rx) ieee80211_parse_qos() argument
596 struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)rx->skb->data; ieee80211_parse_qos()
597 struct ieee80211_rx_status *status = IEEE80211_SKB_RXCB(rx->skb); ieee80211_parse_qos()
629 rx->seqno_idx = seqno_idx; ieee80211_parse_qos()
630 rx->security_idx = security_idx; ieee80211_parse_qos()
633 rx->skb->priority = (tid > 7) ? 0 : tid; ieee80211_parse_qos()
661 static void ieee80211_verify_alignment(struct ieee80211_rx_data *rx) ieee80211_verify_alignment() argument
664 WARN_ONCE((unsigned long)rx->skb->data & 1, ieee80211_verify_alignment()
665 "unaligned packet at 0x%p\n", rx->skb->data); ieee80211_verify_alignment()
670 /* rx handlers */
744 static ieee80211_rx_result ieee80211_rx_mesh_check(struct ieee80211_rx_data *rx) ieee80211_rx_mesh_check() argument
746 struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)rx->skb->data; ieee80211_rx_mesh_check()
747 char *dev_addr = rx->sdata->vif.addr; ieee80211_rx_mesh_check()
768 if (!rx->sta || sta_plink_state(rx->sta) != NL80211_PLINK_ESTAB) { ieee80211_rx_mesh_check()
778 if (rx->skb->len < IEEE80211_MIN_ACTION_SIZE) ieee80211_rx_mesh_check()
1027 static void ieee80211_rx_reorder_ampdu(struct ieee80211_rx_data *rx, ieee80211_rx_reorder_ampdu() argument
1030 struct sk_buff *skb = rx->skb; ieee80211_rx_reorder_ampdu()
1031 struct ieee80211_local *local = rx->local; ieee80211_rx_reorder_ampdu()
1033 struct sta_info *sta = rx->sta; ieee80211_rx_reorder_ampdu()
1043 * filter the QoS data rx stream according to ieee80211_rx_reorder_ampdu()
1073 /* if this mpdu is fragmented - terminate rx aggregation session */ ieee80211_rx_reorder_ampdu()
1077 skb_queue_tail(&rx->sdata->skb_queue, skb); ieee80211_rx_reorder_ampdu()
1078 ieee80211_queue_work(&local->hw, &rx->sdata->work); ieee80211_rx_reorder_ampdu()
1089 if (ieee80211_sta_manage_reorder_buf(rx->sdata, tid_agg_rx, skb, ieee80211_rx_reorder_ampdu()
1098 ieee80211_rx_h_check_dup(struct ieee80211_rx_data *rx) ieee80211_rx_h_check_dup() argument
1100 struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)rx->skb->data; ieee80211_rx_h_check_dup()
1101 struct ieee80211_rx_status *status = IEEE80211_SKB_RXCB(rx->skb); ieee80211_rx_h_check_dup()
1108 if (rx->skb->len < 24) ieee80211_rx_h_check_dup()
1116 if (!rx->sta) ieee80211_rx_h_check_dup()
1120 rx->sta->last_seq_ctrl[rx->seqno_idx] == hdr->seq_ctrl)) { ieee80211_rx_h_check_dup()
1121 I802_DEBUG_INC(rx->local->dot11FrameDuplicateCount); ieee80211_rx_h_check_dup()
1122 rx->sta->rx_stats.num_duplicates++; ieee80211_rx_h_check_dup()
1125 rx->sta->last_seq_ctrl[rx->seqno_idx] = hdr->seq_ctrl; ieee80211_rx_h_check_dup()
1132 ieee80211_rx_h_check(struct ieee80211_rx_data *rx) ieee80211_rx_h_check() argument
1134 struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)rx->skb->data; ieee80211_rx_h_check()
1145 if (ieee80211_vif_is_mesh(&rx->sdata->vif)) ieee80211_rx_h_check()
1146 return ieee80211_rx_mesh_check(rx); ieee80211_rx_h_check()
1150 rx->sdata->vif.type != NL80211_IFTYPE_ADHOC && ieee80211_rx_h_check()
1151 rx->sdata->vif.type != NL80211_IFTYPE_WDS && ieee80211_rx_h_check()
1152 rx->sdata->vif.type != NL80211_IFTYPE_OCB && ieee80211_rx_h_check()
1153 (!rx->sta || !test_sta_flag(rx->sta, WLAN_STA_ASSOC)))) { ieee80211_rx_h_check()
1159 if (rx->sta && rx->sdata->vif.type == NL80211_IFTYPE_STATION && ieee80211_rx_h_check()
1166 if (rx->skb->len < hdrlen + 8) ieee80211_rx_h_check()
1169 skb_copy_bits(rx->skb, hdrlen + 6, &ethertype, 2); ieee80211_rx_h_check()
1170 if (ethertype == rx->sdata->control_port_protocol) ieee80211_rx_h_check()
1174 if (rx->sdata->vif.type == NL80211_IFTYPE_AP && ieee80211_rx_h_check()
1175 cfg80211_rx_spurious_frame(rx->sdata->dev, ieee80211_rx_h_check()
1188 ieee80211_rx_h_check_more_data(struct ieee80211_rx_data *rx) ieee80211_rx_h_check_more_data() argument
1194 local = rx->local; ieee80211_rx_h_check_more_data()
1195 skb = rx->skb; ieee80211_rx_h_check_more_data()
1215 ieee80211_send_pspoll(local, rx->sdata); ieee80211_rx_h_check_more_data()
1300 ieee80211_rx_h_uapsd_and_pspoll(struct ieee80211_rx_data *rx) ieee80211_rx_h_uapsd_and_pspoll() argument
1302 struct ieee80211_sub_if_data *sdata = rx->sdata; ieee80211_rx_h_uapsd_and_pspoll()
1303 struct ieee80211_hdr *hdr = (void *)rx->skb->data; ieee80211_rx_h_uapsd_and_pspoll()
1304 struct ieee80211_rx_status *status = IEEE80211_SKB_RXCB(rx->skb); ieee80211_rx_h_uapsd_and_pspoll()
1307 if (!rx->sta) ieee80211_rx_h_uapsd_and_pspoll()
1327 if (!test_sta_flag(rx->sta, WLAN_STA_PS_STA)) ieee80211_rx_h_uapsd_and_pspoll()
1331 if (!test_sta_flag(rx->sta, WLAN_STA_SP)) { ieee80211_rx_h_uapsd_and_pspoll()
1332 if (!test_sta_flag(rx->sta, WLAN_STA_PS_DRIVER)) ieee80211_rx_h_uapsd_and_pspoll()
1333 ieee80211_sta_ps_deliver_poll_response(rx->sta); ieee80211_rx_h_uapsd_and_pspoll()
1335 set_sta_flag(rx->sta, WLAN_STA_PSPOLL); ieee80211_rx_h_uapsd_and_pspoll()
1340 dev_kfree_skb(rx->skb); ieee80211_rx_h_uapsd_and_pspoll()
1358 if (!(rx->sta->sta.uapsd_queues & BIT(ac))) ieee80211_rx_h_uapsd_and_pspoll()
1362 if (test_sta_flag(rx->sta, WLAN_STA_SP)) ieee80211_rx_h_uapsd_and_pspoll()
1365 if (!test_sta_flag(rx->sta, WLAN_STA_PS_DRIVER)) ieee80211_rx_h_uapsd_and_pspoll()
1366 ieee80211_sta_ps_deliver_uapsd(rx->sta); ieee80211_rx_h_uapsd_and_pspoll()
1368 set_sta_flag(rx->sta, WLAN_STA_UAPSD); ieee80211_rx_h_uapsd_and_pspoll()
1375 ieee80211_rx_h_sta_process(struct ieee80211_rx_data *rx) ieee80211_rx_h_sta_process() argument
1377 struct sta_info *sta = rx->sta; ieee80211_rx_h_sta_process()
1378 struct sk_buff *skb = rx->skb; ieee80211_rx_h_sta_process()
1394 if (rx->sdata->vif.type == NL80211_IFTYPE_ADHOC) { ieee80211_rx_h_sta_process()
1395 u8 *bssid = ieee80211_get_bssid(hdr, rx->skb->len, ieee80211_rx_h_sta_process()
1397 if (ether_addr_equal(bssid, rx->sdata->u.ibss.bssid) && ieee80211_rx_h_sta_process()
1412 } else if (rx->sdata->vif.type == NL80211_IFTYPE_OCB) { ieee80211_rx_h_sta_process()
1428 if (rx->sdata->vif.type == NL80211_IFTYPE_STATION) ieee80211_rx_h_sta_process()
1429 ieee80211_sta_rx_notify(rx->sdata, hdr); ieee80211_rx_h_sta_process()
1432 sta->rx_stats.bytes += rx->skb->len; ieee80211_rx_h_sta_process()
1459 (rx->sdata->vif.type == NL80211_IFTYPE_AP || ieee80211_rx_h_sta_process()
1460 rx->sdata->vif.type == NL80211_IFTYPE_AP_VLAN) && ieee80211_rx_h_sta_process()
1477 if (ieee80211_vif_is_mesh(&rx->sdata->vif)) ieee80211_rx_h_sta_process()
1486 I802_DEBUG_INC(rx->local->rx_handlers_drop_nullfunc); ieee80211_rx_h_sta_process()
1495 (rx->sdata->vif.type == NL80211_IFTYPE_AP || ieee80211_rx_h_sta_process()
1496 (rx->sdata->vif.type == NL80211_IFTYPE_AP_VLAN && ieee80211_rx_h_sta_process()
1497 !rx->sdata->u.vlan.sta))) { ieee80211_rx_h_sta_process()
1500 rx->sdata->dev, sta->sta.addr, ieee80211_rx_h_sta_process()
1509 dev_kfree_skb(rx->skb); ieee80211_rx_h_sta_process()
1517 ieee80211_rx_h_decrypt(struct ieee80211_rx_data *rx) ieee80211_rx_h_decrypt() argument
1519 struct sk_buff *skb = rx->skb; ieee80211_rx_h_decrypt()
1558 rx->key = NULL; ieee80211_rx_h_decrypt()
1561 if (rx->sta) { ieee80211_rx_h_decrypt()
1562 int keyid = rx->sta->ptk_idx; ieee80211_rx_h_decrypt()
1564 if (ieee80211_has_protected(fc) && rx->sta->cipher_scheme) { ieee80211_rx_h_decrypt()
1565 cs = rx->sta->cipher_scheme; ieee80211_rx_h_decrypt()
1566 keyid = iwl80211_get_cs_keyid(cs, rx->skb); ieee80211_rx_h_decrypt()
1570 sta_ptk = rcu_dereference(rx->sta->ptk[keyid]); ieee80211_rx_h_decrypt()
1574 mmie_keyidx = ieee80211_get_mmie_keyidx(rx->skb); ieee80211_rx_h_decrypt()
1577 rx->key = sta_ptk; ieee80211_rx_h_decrypt()
1593 if (rx->sta) ieee80211_rx_h_decrypt()
1594 rx->key = rcu_dereference(rx->sta->gtk[mmie_keyidx]); ieee80211_rx_h_decrypt()
1595 if (!rx->key) ieee80211_rx_h_decrypt()
1596 rx->key = rcu_dereference(rx->sdata->keys[mmie_keyidx]); ieee80211_rx_h_decrypt()
1600 * need to set rx->key if there is a key that could have been ieee80211_rx_h_decrypt()
1605 struct ieee80211_sub_if_data *sdata = rx->sdata; ieee80211_rx_h_decrypt()
1610 (key = rcu_dereference(rx->sdata->default_mgmt_key))) ieee80211_rx_h_decrypt()
1611 rx->key = key; ieee80211_rx_h_decrypt()
1613 if (rx->sta) { ieee80211_rx_h_decrypt()
1615 key = rcu_dereference(rx->sta->gtk[i]); ieee80211_rx_h_decrypt()
1628 rx->key = key; ieee80211_rx_h_decrypt()
1650 keyidx = iwl80211_get_cs_keyid(cs, rx->skb); ieee80211_rx_h_decrypt()
1655 if (rx->skb->len < 8 + hdrlen) ieee80211_rx_h_decrypt()
1661 skb_copy_bits(rx->skb, hdrlen + 3, &keyid, 1); ieee80211_rx_h_decrypt()
1666 if (is_multicast_ether_addr(hdr->addr1) && rx->sta) ieee80211_rx_h_decrypt()
1667 rx->key = rcu_dereference(rx->sta->gtk[keyidx]); ieee80211_rx_h_decrypt()
1670 if (!rx->key) { ieee80211_rx_h_decrypt()
1671 rx->key = rcu_dereference(rx->sdata->keys[keyidx]); ieee80211_rx_h_decrypt()
1678 if (rx->key && ieee80211_rx_h_decrypt()
1679 rx->key->conf.cipher != WLAN_CIPHER_SUITE_WEP40 && ieee80211_rx_h_decrypt()
1680 rx->key->conf.cipher != WLAN_CIPHER_SUITE_WEP104 && ieee80211_rx_h_decrypt()
1682 rx->key = NULL; ieee80211_rx_h_decrypt()
1686 if (rx->key) { ieee80211_rx_h_decrypt()
1687 if (unlikely(rx->key->flags & KEY_FLAG_TAINTED)) ieee80211_rx_h_decrypt()
1695 switch (rx->key->conf.cipher) { ieee80211_rx_h_decrypt()
1698 result = ieee80211_crypto_wep_decrypt(rx); ieee80211_rx_h_decrypt()
1701 result = ieee80211_crypto_tkip_decrypt(rx); ieee80211_rx_h_decrypt()
1705 rx, IEEE80211_CCMP_MIC_LEN); ieee80211_rx_h_decrypt()
1709 rx, IEEE80211_CCMP_256_MIC_LEN); ieee80211_rx_h_decrypt()
1712 result = ieee80211_crypto_aes_cmac_decrypt(rx); ieee80211_rx_h_decrypt()
1715 result = ieee80211_crypto_aes_cmac_256_decrypt(rx); ieee80211_rx_h_decrypt()
1719 result = ieee80211_crypto_aes_gmac_decrypt(rx); ieee80211_rx_h_decrypt()
1723 result = ieee80211_crypto_gcmp_decrypt(rx); ieee80211_rx_h_decrypt()
1726 result = ieee80211_crypto_hw_decrypt(rx); ieee80211_rx_h_decrypt()
1807 ieee80211_rx_h_defragment(struct ieee80211_rx_data *rx) ieee80211_rx_h_defragment() argument
1817 hdr = (struct ieee80211_hdr *)rx->skb->data; ieee80211_rx_h_defragment()
1827 I802_DEBUG_INC(rx->local->dot11MulticastReceivedFrameCount); ieee80211_rx_h_defragment()
1834 I802_DEBUG_INC(rx->local->rx_handlers_fragments); ieee80211_rx_h_defragment()
1836 if (skb_linearize(rx->skb)) ieee80211_rx_h_defragment()
1844 hdr = (struct ieee80211_hdr *)rx->skb->data; ieee80211_rx_h_defragment()
1849 entry = ieee80211_reassemble_add(rx->sdata, frag, seq, ieee80211_rx_h_defragment()
1850 rx->seqno_idx, &(rx->skb)); ieee80211_rx_h_defragment()
1851 if (rx->key && ieee80211_rx_h_defragment()
1852 (rx->key->conf.cipher == WLAN_CIPHER_SUITE_CCMP || ieee80211_rx_h_defragment()
1853 rx->key->conf.cipher == WLAN_CIPHER_SUITE_CCMP_256 || ieee80211_rx_h_defragment()
1854 rx->key->conf.cipher == WLAN_CIPHER_SUITE_GCMP || ieee80211_rx_h_defragment()
1855 rx->key->conf.cipher == WLAN_CIPHER_SUITE_GCMP_256) && ieee80211_rx_h_defragment()
1857 int queue = rx->security_idx; ieee80211_rx_h_defragment()
1864 rx->key->u.ccmp.rx_pn[queue], ieee80211_rx_h_defragment()
1870 BUILD_BUG_ON(sizeof(rx->key->u.ccmp.rx_pn[queue]) != ieee80211_rx_h_defragment()
1871 sizeof(rx->key->u.gcmp.rx_pn[queue])); ieee80211_rx_h_defragment()
1881 entry = ieee80211_reassemble_find(rx->sdata, frag, seq, ieee80211_rx_h_defragment()
1882 rx->seqno_idx, hdr); ieee80211_rx_h_defragment()
1884 I802_DEBUG_INC(rx->local->rx_handlers_drop_defrag); ieee80211_rx_h_defragment()
1898 if (!rx->key || ieee80211_rx_h_defragment()
1899 (rx->key->conf.cipher != WLAN_CIPHER_SUITE_CCMP && ieee80211_rx_h_defragment()
1900 rx->key->conf.cipher != WLAN_CIPHER_SUITE_CCMP_256 && ieee80211_rx_h_defragment()
1901 rx->key->conf.cipher != WLAN_CIPHER_SUITE_GCMP && ieee80211_rx_h_defragment()
1902 rx->key->conf.cipher != WLAN_CIPHER_SUITE_GCMP_256)) ieee80211_rx_h_defragment()
1910 queue = rx->security_idx; ieee80211_rx_h_defragment()
1911 rpn = rx->key->u.ccmp.rx_pn[queue]; ieee80211_rx_h_defragment()
1917 skb_pull(rx->skb, ieee80211_hdrlen(fc)); ieee80211_rx_h_defragment()
1918 __skb_queue_tail(&entry->skb_list, rx->skb); ieee80211_rx_h_defragment()
1920 entry->extra_len += rx->skb->len; ieee80211_rx_h_defragment()
1922 rx->skb = NULL; ieee80211_rx_h_defragment()
1926 rx->skb = __skb_dequeue(&entry->skb_list); ieee80211_rx_h_defragment()
1927 if (skb_tailroom(rx->skb) < entry->extra_len) { ieee80211_rx_h_defragment()
1928 I802_DEBUG_INC(rx->local->rx_expand_skb_head_defrag); ieee80211_rx_h_defragment()
1929 if (unlikely(pskb_expand_head(rx->skb, 0, entry->extra_len, ieee80211_rx_h_defragment()
1931 I802_DEBUG_INC(rx->local->rx_handlers_drop_defrag); ieee80211_rx_h_defragment()
1937 memcpy(skb_put(rx->skb, skb->len), skb->data, skb->len); ieee80211_rx_h_defragment()
1942 status = IEEE80211_SKB_RXCB(rx->skb); ieee80211_rx_h_defragment()
1945 ieee80211_led_rx(rx->local); ieee80211_rx_h_defragment()
1947 if (rx->sta) ieee80211_rx_h_defragment()
1948 rx->sta->rx_stats.packets++; ieee80211_rx_h_defragment()
1952 static int ieee80211_802_1x_port_control(struct ieee80211_rx_data *rx) ieee80211_802_1x_port_control() argument
1954 if (unlikely(!rx->sta || !test_sta_flag(rx->sta, WLAN_STA_AUTHORIZED))) ieee80211_802_1x_port_control()
1960 static int ieee80211_drop_unencrypted(struct ieee80211_rx_data *rx, __le16 fc) ieee80211_drop_unencrypted() argument
1962 struct sk_buff *skb = rx->skb; ieee80211_drop_unencrypted()
1975 ieee80211_is_data(fc) && rx->key)) ieee80211_drop_unencrypted()
1981 static int ieee80211_drop_unencrypted_mgmt(struct ieee80211_rx_data *rx) ieee80211_drop_unencrypted_mgmt() argument
1983 struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)rx->skb->data; ieee80211_drop_unencrypted_mgmt()
1984 struct ieee80211_rx_status *status = IEEE80211_SKB_RXCB(rx->skb); ieee80211_drop_unencrypted_mgmt()
1994 if (rx->sta && test_sta_flag(rx->sta, WLAN_STA_MFP)) { ieee80211_drop_unencrypted_mgmt()
1996 ieee80211_is_unicast_robust_mgmt_frame(rx->skb) && ieee80211_drop_unencrypted_mgmt()
1997 rx->key)) { ieee80211_drop_unencrypted_mgmt()
2000 cfg80211_rx_unprot_mlme_mgmt(rx->sdata->dev, ieee80211_drop_unencrypted_mgmt()
2001 rx->skb->data, ieee80211_drop_unencrypted_mgmt()
2002 rx->skb->len); ieee80211_drop_unencrypted_mgmt()
2006 if (unlikely(ieee80211_is_multicast_robust_mgmt_frame(rx->skb) && ieee80211_drop_unencrypted_mgmt()
2007 ieee80211_get_mmie_keyidx(rx->skb) < 0)) { ieee80211_drop_unencrypted_mgmt()
2010 cfg80211_rx_unprot_mlme_mgmt(rx->sdata->dev, ieee80211_drop_unencrypted_mgmt()
2011 rx->skb->data, ieee80211_drop_unencrypted_mgmt()
2012 rx->skb->len); ieee80211_drop_unencrypted_mgmt()
2019 if (unlikely(ieee80211_is_action(fc) && !rx->key && ieee80211_drop_unencrypted_mgmt()
2020 ieee80211_is_robust_mgmt_frame(rx->skb))) ieee80211_drop_unencrypted_mgmt()
2028 __ieee80211_data_to_8023(struct ieee80211_rx_data *rx, bool *port_control) __ieee80211_data_to_8023() argument
2030 struct ieee80211_sub_if_data *sdata = rx->sdata; __ieee80211_data_to_8023()
2031 struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)rx->skb->data; __ieee80211_data_to_8023()
2054 ret = ieee80211_data_to_8023(rx->skb, sdata->vif.addr, sdata->vif.type); __ieee80211_data_to_8023()
2058 ehdr = (struct ethhdr *) rx->skb->data; __ieee80211_data_to_8023()
2059 if (ehdr->h_proto == rx->sdata->control_port_protocol) __ieee80211_data_to_8023()
2068 * requires that rx->skb is a frame with ethernet header
2070 static bool ieee80211_frame_allowed(struct ieee80211_rx_data *rx, __le16 fc) ieee80211_frame_allowed() argument
2074 struct ethhdr *ehdr = (struct ethhdr *) rx->skb->data; ieee80211_frame_allowed()
2080 if (ehdr->h_proto == rx->sdata->control_port_protocol && ieee80211_frame_allowed()
2081 (ether_addr_equal(ehdr->h_dest, rx->sdata->vif.addr) || ieee80211_frame_allowed()
2085 if (ieee80211_802_1x_port_control(rx) || ieee80211_frame_allowed()
2086 ieee80211_drop_unencrypted(rx, fc)) ieee80211_frame_allowed()
2093 * requires that rx->skb is a frame with ethernet header
2096 ieee80211_deliver_skb(struct ieee80211_rx_data *rx) ieee80211_deliver_skb() argument
2098 struct ieee80211_sub_if_data *sdata = rx->sdata; ieee80211_deliver_skb()
2101 struct ethhdr *ehdr = (struct ethhdr *) rx->skb->data; ieee80211_deliver_skb()
2104 skb = rx->skb; ieee80211_deliver_skb()
2167 if (rx->napi) ieee80211_deliver_skb()
2168 napi_gro_receive(rx->napi, skb); ieee80211_deliver_skb()
2188 ieee80211_rx_h_amsdu(struct ieee80211_rx_data *rx) ieee80211_rx_h_amsdu() argument
2190 struct net_device *dev = rx->sdata->dev; ieee80211_rx_h_amsdu()
2191 struct sk_buff *skb = rx->skb; ieee80211_rx_h_amsdu()
2195 struct ieee80211_rx_status *status = IEEE80211_SKB_RXCB(rx->skb); ieee80211_rx_h_amsdu()
2207 rx->sdata->vif.type == NL80211_IFTYPE_AP_VLAN && ieee80211_rx_h_amsdu()
2208 !rx->sdata->u.vlan.sta) ieee80211_rx_h_amsdu()
2212 ((rx->sdata->vif.type == NL80211_IFTYPE_AP_VLAN && ieee80211_rx_h_amsdu()
2213 rx->sdata->u.vlan.sta) || ieee80211_rx_h_amsdu()
2214 (rx->sdata->vif.type == NL80211_IFTYPE_STATION && ieee80211_rx_h_amsdu()
2215 rx->sdata->u.mgd.use_4addr))) ieee80211_rx_h_amsdu()
2225 rx->sdata->vif.type, ieee80211_rx_h_amsdu()
2226 rx->local->hw.extra_tx_headroom, true); ieee80211_rx_h_amsdu()
2229 rx->skb = __skb_dequeue(&frame_list); ieee80211_rx_h_amsdu()
2231 if (!ieee80211_frame_allowed(rx, fc)) { ieee80211_rx_h_amsdu()
2232 dev_kfree_skb(rx->skb); ieee80211_rx_h_amsdu()
2236 ieee80211_deliver_skb(rx); ieee80211_rx_h_amsdu()
2244 ieee80211_rx_h_mesh_fwding(struct ieee80211_rx_data *rx) ieee80211_rx_h_mesh_fwding() argument
2249 struct sk_buff *skb = rx->skb, *fwd_skb; ieee80211_rx_h_mesh_fwding()
2250 struct ieee80211_local *local = rx->local; ieee80211_rx_h_mesh_fwding()
2251 struct ieee80211_sub_if_data *sdata = rx->sdata; ieee80211_rx_h_mesh_fwding()
2259 if (!pskb_may_pull(rx->skb, hdrlen + 6)) ieee80211_rx_h_mesh_fwding()
2265 if (!pskb_may_pull(rx->skb, ieee80211_rx_h_mesh_fwding()
2273 if (ieee80211_drop_unencrypted(rx, hdr->frame_control)) ieee80211_rx_h_mesh_fwding()
2279 mesh_rmc_check(rx->sdata, hdr->addr3, mesh_hdr)) ieee80211_rx_h_mesh_fwding()
2350 info->control.vif = &rx->sdata->vif; ieee80211_rx_h_mesh_fwding()
2381 ieee80211_rx_h_data(struct ieee80211_rx_data *rx) ieee80211_rx_h_data() argument
2383 struct ieee80211_sub_if_data *sdata = rx->sdata; ieee80211_rx_h_data()
2384 struct ieee80211_local *local = rx->local; ieee80211_rx_h_data()
2386 struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)rx->skb->data; ieee80211_rx_h_data()
2397 if (rx->sta) { ieee80211_rx_h_data()
2403 rx->sta->rx_stats.msdu[rx->seqno_idx]++; ieee80211_rx_h_data()
2412 if (rx->sta && ieee80211_rx_h_data()
2413 !test_and_set_sta_flag(rx->sta, WLAN_STA_4ADDR_EVENT)) ieee80211_rx_h_data()
2415 rx->sdata->dev, rx->sta->sta.addr, GFP_ATOMIC); ieee80211_rx_h_data()
2419 err = __ieee80211_data_to_8023(rx, &port_control); ieee80211_rx_h_data()
2423 if (!ieee80211_frame_allowed(rx, fc)) ieee80211_rx_h_data()
2427 if (unlikely(((struct ethhdr *)rx->skb->data)->h_proto == ieee80211_rx_h_data()
2429 struct ieee80211_tdls_data *tf = (void *)rx->skb->data; ieee80211_rx_h_data()
2431 if (pskb_may_pull(rx->skb, ieee80211_rx_h_data()
2437 skb_queue_tail(&local->skb_queue_tdls_chsw, rx->skb); ieee80211_rx_h_data()
2439 if (rx->sta) ieee80211_rx_h_data()
2440 rx->sta->rx_stats.packets++; ieee80211_rx_h_data()
2446 if (rx->sdata->vif.type == NL80211_IFTYPE_AP_VLAN && ieee80211_rx_h_data()
2451 rx->sdata = sdata; ieee80211_rx_h_data()
2454 rx->skb->dev = dev; ieee80211_rx_h_data()
2458 ((struct ethhdr *)rx->skb->data)->h_dest) && ieee80211_rx_h_data()
2465 ieee80211_deliver_skb(rx); ieee80211_rx_h_data()
2471 ieee80211_rx_h_ctrl(struct ieee80211_rx_data *rx, struct sk_buff_head *frames) ieee80211_rx_h_ctrl() argument
2473 struct sk_buff *skb = rx->skb; ieee80211_rx_h_ctrl()
2490 if (!rx->sta) ieee80211_rx_h_ctrl()
2499 tid_agg_rx = rcu_dereference(rx->sta->ampdu_mlme.tid_rx[tid]); ieee80211_rx_h_ctrl()
2506 event.u.ba.sta = &rx->sta->sta; ieee80211_rx_h_ctrl()
2515 ieee80211_release_reorder_frames(rx->sdata, tid_agg_rx, ieee80211_rx_h_ctrl()
2519 drv_event_callback(rx->local, rx->sdata, &event); ieee80211_rx_h_ctrl()
2580 ieee80211_rx_h_mgmt_check(struct ieee80211_rx_data *rx) ieee80211_rx_h_mgmt_check() argument
2582 struct ieee80211_mgmt *mgmt = (struct ieee80211_mgmt *) rx->skb->data; ieee80211_rx_h_mgmt_check()
2583 struct ieee80211_rx_status *status = IEEE80211_SKB_RXCB(rx->skb); ieee80211_rx_h_mgmt_check()
2590 if (rx->skb->len < 24) ieee80211_rx_h_mgmt_check()
2596 if (rx->sdata->vif.type == NL80211_IFTYPE_AP && ieee80211_rx_h_mgmt_check()
2598 !(rx->flags & IEEE80211_RX_BEACON_REPORTED)) { ieee80211_rx_h_mgmt_check()
2601 if (ieee80211_hw_check(&rx->local->hw, SIGNAL_DBM)) ieee80211_rx_h_mgmt_check()
2604 cfg80211_report_obss_beacon(rx->local->hw.wiphy, ieee80211_rx_h_mgmt_check()
2605 rx->skb->data, rx->skb->len, ieee80211_rx_h_mgmt_check()
2607 rx->flags |= IEEE80211_RX_BEACON_REPORTED; ieee80211_rx_h_mgmt_check()
2610 if (ieee80211_drop_unencrypted_mgmt(rx)) ieee80211_rx_h_mgmt_check()
2617 ieee80211_rx_h_action(struct ieee80211_rx_data *rx) ieee80211_rx_h_action() argument
2619 struct ieee80211_local *local = rx->local; ieee80211_rx_h_action()
2620 struct ieee80211_sub_if_data *sdata = rx->sdata; ieee80211_rx_h_action()
2621 struct ieee80211_mgmt *mgmt = (struct ieee80211_mgmt *) rx->skb->data; ieee80211_rx_h_action()
2622 struct ieee80211_rx_status *status = IEEE80211_SKB_RXCB(rx->skb); ieee80211_rx_h_action()
2623 int len = rx->skb->len; ieee80211_rx_h_action()
2632 if (!rx->sta && mgmt->u.action.category != WLAN_CATEGORY_PUBLIC && ieee80211_rx_h_action()
2640 if (!rx->sta->sta.ht_cap.ht_supported) ieee80211_rx_h_action()
2675 if (rx->sta->sta.smps_mode == smps_mode) ieee80211_rx_h_action()
2677 rx->sta->sta.smps_mode = smps_mode; ieee80211_rx_h_action()
2679 sband = rx->local->hw.wiphy->bands[status->band]; ieee80211_rx_h_action()
2681 rate_control_rate_update(local, sband, rx->sta, ieee80211_rx_h_action()
2691 if (!(rx->sta->sta.ht_cap.cap & ieee80211_rx_h_action()
2698 max_bw = ieee80211_sta_cap_rx_bw(rx->sta); ieee80211_rx_h_action()
2701 rx->sta->cur_max_bandwidth = max_bw; ieee80211_rx_h_action()
2702 new_bw = ieee80211_sta_cur_vht_bw(rx->sta); ieee80211_rx_h_action()
2704 if (rx->sta->sta.bandwidth == new_bw) ieee80211_rx_h_action()
2707 rx->sta->sta.bandwidth = new_bw; ieee80211_rx_h_action()
2708 sband = rx->local->hw.wiphy->bands[status->band]; ieee80211_rx_h_action()
2710 rate_control_rate_update(local, sband, rx->sta, ieee80211_rx_h_action()
2724 if (!rx->sta) ieee80211_rx_h_action()
2757 ieee80211_vht_handle_opmode(rx->sdata, rx->sta, ieee80211_rx_h_action()
2900 if (rx->sta) ieee80211_rx_h_action()
2901 rx->sta->rx_stats.packets++; ieee80211_rx_h_action()
2902 dev_kfree_skb(rx->skb); ieee80211_rx_h_action()
2906 rx->skb->pkt_type = IEEE80211_SDATA_QUEUE_TYPE_FRAME; ieee80211_rx_h_action()
2907 skb_queue_tail(&sdata->skb_queue, rx->skb); ieee80211_rx_h_action()
2909 if (rx->sta) ieee80211_rx_h_action()
2910 rx->sta->rx_stats.packets++; ieee80211_rx_h_action()
2915 ieee80211_rx_h_userspace_mgmt(struct ieee80211_rx_data *rx) ieee80211_rx_h_userspace_mgmt() argument
2917 struct ieee80211_rx_status *status = IEEE80211_SKB_RXCB(rx->skb); ieee80211_rx_h_userspace_mgmt()
2931 if (ieee80211_hw_check(&rx->local->hw, SIGNAL_DBM)) ieee80211_rx_h_userspace_mgmt()
2934 if (cfg80211_rx_mgmt(&rx->sdata->wdev, status->freq, sig, ieee80211_rx_h_userspace_mgmt()
2935 rx->skb->data, rx->skb->len, 0)) { ieee80211_rx_h_userspace_mgmt()
2936 if (rx->sta) ieee80211_rx_h_userspace_mgmt()
2937 rx->sta->rx_stats.packets++; ieee80211_rx_h_userspace_mgmt()
2938 dev_kfree_skb(rx->skb); ieee80211_rx_h_userspace_mgmt()
2946 ieee80211_rx_h_action_return(struct ieee80211_rx_data *rx) ieee80211_rx_h_action_return() argument
2948 struct ieee80211_local *local = rx->local; ieee80211_rx_h_action_return()
2949 struct ieee80211_mgmt *mgmt = (struct ieee80211_mgmt *) rx->skb->data; ieee80211_rx_h_action_return()
2951 struct ieee80211_sub_if_data *sdata = rx->sdata; ieee80211_rx_h_action_return()
2952 struct ieee80211_rx_status *status = IEEE80211_SKB_RXCB(rx->skb); ieee80211_rx_h_action_return()
2979 nskb = skb_copy_expand(rx->skb, local->hw.extra_tx_headroom, 0, ieee80211_rx_h_action_return()
2986 memcpy(nmgmt->sa, rx->sdata->vif.addr, ETH_ALEN); ieee80211_rx_h_action_return()
2990 if (rx->sdata->vif.type == NL80211_IFTYPE_P2P_DEVICE) { ieee80211_rx_h_action_return()
3001 __ieee80211_tx_skb_tid_band(rx->sdata, nskb, 7, ieee80211_rx_h_action_return()
3004 dev_kfree_skb(rx->skb); ieee80211_rx_h_action_return()
3009 ieee80211_rx_h_mgmt(struct ieee80211_rx_data *rx) ieee80211_rx_h_mgmt() argument
3011 struct ieee80211_sub_if_data *sdata = rx->sdata; ieee80211_rx_h_mgmt()
3012 struct ieee80211_mgmt *mgmt = (void *)rx->skb->data; ieee80211_rx_h_mgmt()
3052 rx->skb->pkt_type = IEEE80211_SDATA_QUEUE_TYPE_FRAME; ieee80211_rx_h_mgmt()
3053 skb_queue_tail(&sdata->skb_queue, rx->skb); ieee80211_rx_h_mgmt()
3054 ieee80211_queue_work(&rx->local->hw, &sdata->work); ieee80211_rx_h_mgmt()
3055 if (rx->sta) ieee80211_rx_h_mgmt()
3056 rx->sta->rx_stats.packets++; ieee80211_rx_h_mgmt()
3061 static void ieee80211_rx_cooked_monitor(struct ieee80211_rx_data *rx, ieee80211_rx_cooked_monitor() argument
3065 struct ieee80211_local *local = rx->local; ieee80211_rx_cooked_monitor()
3066 struct sk_buff *skb = rx->skb, *skb2; ieee80211_rx_cooked_monitor()
3075 if (rx->flags & IEEE80211_RX_CMNTR) ieee80211_rx_cooked_monitor()
3077 rx->flags |= IEEE80211_RX_CMNTR; ieee80211_rx_cooked_monitor()
3131 static void ieee80211_rx_handlers_result(struct ieee80211_rx_data *rx, ieee80211_rx_handlers_result() argument
3136 I802_DEBUG_INC(rx->sdata->local->rx_handlers_drop); ieee80211_rx_handlers_result()
3137 if (rx->sta) ieee80211_rx_handlers_result()
3138 rx->sta->rx_stats.dropped++; ieee80211_rx_handlers_result()
3145 status = IEEE80211_SKB_RXCB((rx->skb)); ieee80211_rx_handlers_result()
3147 sband = rx->local->hw.wiphy->bands[status->band]; ieee80211_rx_handlers_result()
3152 ieee80211_rx_cooked_monitor(rx, rate); ieee80211_rx_handlers_result()
3156 I802_DEBUG_INC(rx->sdata->local->rx_handlers_drop); ieee80211_rx_handlers_result()
3157 if (rx->sta) ieee80211_rx_handlers_result()
3158 rx->sta->rx_stats.dropped++; ieee80211_rx_handlers_result()
3159 dev_kfree_skb(rx->skb); ieee80211_rx_handlers_result()
3162 I802_DEBUG_INC(rx->sdata->local->rx_handlers_queued); ieee80211_rx_handlers_result()
3167 static void ieee80211_rx_handlers(struct ieee80211_rx_data *rx, ieee80211_rx_handlers() argument
3175 res = rxh(rx); \ ieee80211_rx_handlers()
3186 spin_lock_bh(&rx->local->rx_path_lock); ieee80211_rx_handlers()
3194 rx->skb = skb; ieee80211_rx_handlers()
3204 if (ieee80211_vif_is_mesh(&rx->sdata->vif)) ieee80211_rx_handlers()
3211 res = ieee80211_rx_h_ctrl(rx, frames); ieee80211_rx_handlers()
3222 ieee80211_rx_handlers_result(rx, res); ieee80211_rx_handlers()
3227 spin_unlock_bh(&rx->local->rx_path_lock); ieee80211_rx_handlers()
3230 static void ieee80211_invoke_rx_handlers(struct ieee80211_rx_data *rx) ieee80211_invoke_rx_handlers() argument
3239 res = rxh(rx); \ ieee80211_invoke_rx_handlers()
3247 ieee80211_rx_reorder_ampdu(rx, &reorder_release); ieee80211_invoke_rx_handlers()
3249 ieee80211_rx_handlers(rx, &reorder_release); ieee80211_invoke_rx_handlers()
3253 ieee80211_rx_handlers_result(rx, res); ieee80211_invoke_rx_handlers()
3265 struct ieee80211_rx_data rx = { ieee80211_release_reorder_timeout() local
3292 drv_event_callback(rx.local, rx.sdata, &event); ieee80211_release_reorder_timeout()
3295 ieee80211_rx_handlers(&rx, &frames); ieee80211_release_reorder_timeout()
3300 static bool ieee80211_accept_frame(struct ieee80211_rx_data *rx) ieee80211_accept_frame() argument
3302 struct ieee80211_sub_if_data *sdata = rx->sdata; ieee80211_accept_frame()
3303 struct sk_buff *skb = rx->skb; ieee80211_accept_frame()
3329 if (!rx->sta) { ieee80211_accept_frame()
3349 if (!rx->sta) { ieee80211_accept_frame()
3417 static bool ieee80211_prepare_and_rx_handle(struct ieee80211_rx_data *rx, ieee80211_prepare_and_rx_handle() argument
3420 struct ieee80211_local *local = rx->local; ieee80211_prepare_and_rx_handle()
3421 struct ieee80211_sub_if_data *sdata = rx->sdata; ieee80211_prepare_and_rx_handle()
3423 rx->skb = skb; ieee80211_prepare_and_rx_handle()
3425 if (!ieee80211_accept_frame(rx)) ieee80211_prepare_and_rx_handle()
3438 rx->skb = skb; ieee80211_prepare_and_rx_handle()
3441 ieee80211_invoke_rx_handlers(rx); ieee80211_prepare_and_rx_handle()
3457 struct ieee80211_rx_data rx; __ieee80211_rx_handle_packet() local
3464 memset(&rx, 0, sizeof(rx)); __ieee80211_rx_handle_packet()
3465 rx.skb = skb; __ieee80211_rx_handle_packet()
3466 rx.local = local; __ieee80211_rx_handle_packet()
3467 rx.napi = napi; __ieee80211_rx_handle_packet()
3488 ieee80211_parse_qos(&rx); __ieee80211_rx_handle_packet()
3489 ieee80211_verify_alignment(&rx); __ieee80211_rx_handle_packet()
3508 rx.sta = prev_sta; __ieee80211_rx_handle_packet()
3509 rx.sdata = prev_sta->sdata; __ieee80211_rx_handle_packet()
3510 ieee80211_prepare_and_rx_handle(&rx, skb, false); __ieee80211_rx_handle_packet()
3516 rx.sta = prev_sta; __ieee80211_rx_handle_packet()
3517 rx.sdata = prev_sta->sdata; __ieee80211_rx_handle_packet()
3519 if (ieee80211_prepare_and_rx_handle(&rx, skb, true)) __ieee80211_rx_handle_packet()
3546 rx.sta = sta_info_get_bss(prev, hdr->addr2); __ieee80211_rx_handle_packet()
3547 rx.sdata = prev; __ieee80211_rx_handle_packet()
3548 ieee80211_prepare_and_rx_handle(&rx, skb, false); __ieee80211_rx_handle_packet()
3554 rx.sta = sta_info_get_bss(prev, hdr->addr2); __ieee80211_rx_handle_packet()
3555 rx.sdata = prev; __ieee80211_rx_handle_packet()
3557 if (ieee80211_prepare_and_rx_handle(&rx, skb, true)) __ieee80211_rx_handle_packet()
3679 /* This is a version of the rx handler that can be called from hard irq
H A Dwpa.c87 ieee80211_rx_h_michael_mic_verify(struct ieee80211_rx_data *rx) ieee80211_rx_h_michael_mic_verify() argument
93 struct sk_buff *skb = rx->skb; ieee80211_rx_h_michael_mic_verify()
114 if (!(status->flag & RX_FLAG_IV_STRIPPED) && rx->key && ieee80211_rx_h_michael_mic_verify()
115 rx->key->conf.cipher == WLAN_CIPHER_SUITE_TKIP) ieee80211_rx_h_michael_mic_verify()
126 if (!rx->key || rx->key->conf.cipher != WLAN_CIPHER_SUITE_TKIP || ieee80211_rx_h_michael_mic_verify()
130 if (rx->sdata->vif.type == NL80211_IFTYPE_AP && rx->key->conf.keyidx) { ieee80211_rx_h_michael_mic_verify()
147 if (skb_linearize(rx->skb)) ieee80211_rx_h_michael_mic_verify()
153 key = &rx->key->conf.key[NL80211_TKIP_DATA_OFFSET_RX_MIC_KEY]; ieee80211_rx_h_michael_mic_verify()
163 rx->key->u.tkip.rx[rx->security_idx].iv32 = rx->tkip_iv32; ieee80211_rx_h_michael_mic_verify()
164 rx->key->u.tkip.rx[rx->security_idx].iv16 = rx->tkip_iv16; ieee80211_rx_h_michael_mic_verify()
169 rx->key->u.tkip.mic_failures++; ieee80211_rx_h_michael_mic_verify()
177 cfg80211_michael_mic_failure(rx->sdata->dev, hdr->addr2, ieee80211_rx_h_michael_mic_verify()
181 rx->key ? rx->key->conf.keyidx : -1, ieee80211_rx_h_michael_mic_verify()
261 ieee80211_crypto_tkip_decrypt(struct ieee80211_rx_data *rx) ieee80211_crypto_tkip_decrypt() argument
263 struct ieee80211_hdr *hdr = (struct ieee80211_hdr *) rx->skb->data; ieee80211_crypto_tkip_decrypt()
265 struct ieee80211_key *key = rx->key; ieee80211_crypto_tkip_decrypt()
266 struct sk_buff *skb = rx->skb; ieee80211_crypto_tkip_decrypt()
274 if (!rx->sta || skb->len - hdrlen < 12) ieee80211_crypto_tkip_decrypt()
278 if (skb_linearize(rx->skb)) ieee80211_crypto_tkip_decrypt()
290 res = ieee80211_tkip_decrypt_data(rx->local->wep_rx_tfm, ieee80211_crypto_tkip_decrypt()
292 skb->len - hdrlen, rx->sta->sta.addr, ieee80211_crypto_tkip_decrypt()
293 hdr->addr1, hwaccel, rx->security_idx, ieee80211_crypto_tkip_decrypt()
294 &rx->tkip_iv32, ieee80211_crypto_tkip_decrypt()
295 &rx->tkip_iv16); ieee80211_crypto_tkip_decrypt()
492 ieee80211_crypto_ccmp_decrypt(struct ieee80211_rx_data *rx, ieee80211_crypto_ccmp_decrypt() argument
495 struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)rx->skb->data; ieee80211_crypto_ccmp_decrypt()
497 struct ieee80211_key *key = rx->key; ieee80211_crypto_ccmp_decrypt()
498 struct sk_buff *skb = rx->skb; ieee80211_crypto_ccmp_decrypt()
511 if (!rx->sta || data_len < 0) ieee80211_crypto_ccmp_decrypt()
515 if (!pskb_may_pull(rx->skb, hdrlen + IEEE80211_CCMP_HDR_LEN)) ieee80211_crypto_ccmp_decrypt()
518 if (skb_linearize(rx->skb)) ieee80211_crypto_ccmp_decrypt()
525 queue = rx->security_idx; ieee80211_crypto_ccmp_decrypt()
718 ieee80211_crypto_gcmp_decrypt(struct ieee80211_rx_data *rx) ieee80211_crypto_gcmp_decrypt() argument
720 struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)rx->skb->data; ieee80211_crypto_gcmp_decrypt()
722 struct ieee80211_key *key = rx->key; ieee80211_crypto_gcmp_decrypt()
723 struct sk_buff *skb = rx->skb; ieee80211_crypto_gcmp_decrypt()
737 if (!rx->sta || data_len < 0) ieee80211_crypto_gcmp_decrypt()
741 if (!pskb_may_pull(rx->skb, hdrlen + IEEE80211_GCMP_HDR_LEN)) ieee80211_crypto_gcmp_decrypt()
744 if (skb_linearize(rx->skb)) ieee80211_crypto_gcmp_decrypt()
751 queue = rx->security_idx; ieee80211_crypto_gcmp_decrypt()
830 ieee80211_crypto_cs_decrypt(struct ieee80211_rx_data *rx) ieee80211_crypto_cs_decrypt() argument
832 struct ieee80211_key *key = rx->key; ieee80211_crypto_cs_decrypt()
833 struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)rx->skb->data; ieee80211_crypto_cs_decrypt()
836 struct ieee80211_rx_status *status = IEEE80211_SKB_RXCB(rx->skb); ieee80211_crypto_cs_decrypt()
842 if (!rx->sta || !rx->sta->cipher_scheme || ieee80211_crypto_cs_decrypt()
849 cs = rx->sta->cipher_scheme; ieee80211_crypto_cs_decrypt()
851 data_len = rx->skb->len - hdrlen - cs->hdr_len; ieee80211_crypto_cs_decrypt()
862 if (skb_linearize(rx->skb)) ieee80211_crypto_cs_decrypt()
865 hdr = (struct ieee80211_hdr *)rx->skb->data; ieee80211_crypto_cs_decrypt()
868 skb_pn = rx->skb->data + hdrlen + cs->pn_off; ieee80211_crypto_cs_decrypt()
876 if (pskb_trim(rx->skb, rx->skb->len - cs->mic_len)) ieee80211_crypto_cs_decrypt()
879 memmove(rx->skb->data + cs->hdr_len, rx->skb->data, hdrlen); ieee80211_crypto_cs_decrypt()
880 skb_pull(rx->skb, cs->hdr_len); ieee80211_crypto_cs_decrypt()
1012 ieee80211_crypto_aes_cmac_decrypt(struct ieee80211_rx_data *rx) ieee80211_crypto_aes_cmac_decrypt() argument
1014 struct sk_buff *skb = rx->skb; ieee80211_crypto_aes_cmac_decrypt()
1016 struct ieee80211_key *key = rx->key; ieee80211_crypto_aes_cmac_decrypt()
1062 ieee80211_crypto_aes_cmac_256_decrypt(struct ieee80211_rx_data *rx) ieee80211_crypto_aes_cmac_256_decrypt() argument
1064 struct sk_buff *skb = rx->skb; ieee80211_crypto_aes_cmac_256_decrypt()
1066 struct ieee80211_key *key = rx->key; ieee80211_crypto_aes_cmac_256_decrypt()
1161 ieee80211_crypto_aes_gmac_decrypt(struct ieee80211_rx_data *rx) ieee80211_crypto_aes_gmac_decrypt() argument
1163 struct sk_buff *skb = rx->skb; ieee80211_crypto_aes_gmac_decrypt()
1165 struct ieee80211_key *key = rx->key; ieee80211_crypto_aes_gmac_decrypt()
1242 ieee80211_crypto_hw_decrypt(struct ieee80211_rx_data *rx) ieee80211_crypto_hw_decrypt() argument
1244 if (rx->sta && rx->sta->cipher_scheme) ieee80211_crypto_hw_decrypt()
1245 return ieee80211_crypto_cs_decrypt(rx); ieee80211_crypto_hw_decrypt()
H A DMakefile11 ht.o agg-tx.o agg-rx.o \
24 rx.o \
H A Dtkip.c268 if (key->u.tkip.rx[queue].state != TKIP_STATE_NOT_INIT && ieee80211_tkip_decrypt_data()
269 (iv32 < key->u.tkip.rx[queue].iv32 || ieee80211_tkip_decrypt_data()
270 (iv32 == key->u.tkip.rx[queue].iv32 && ieee80211_tkip_decrypt_data()
271 iv16 <= key->u.tkip.rx[queue].iv16))) ieee80211_tkip_decrypt_data()
276 key->u.tkip.rx[queue].state = TKIP_STATE_PHASE1_HW_UPLOADED; ieee80211_tkip_decrypt_data()
280 if (key->u.tkip.rx[queue].state == TKIP_STATE_NOT_INIT || ieee80211_tkip_decrypt_data()
281 key->u.tkip.rx[queue].iv32 != iv32) { ieee80211_tkip_decrypt_data()
283 tkip_mixing_phase1(tk, &key->u.tkip.rx[queue], ta, iv32); ieee80211_tkip_decrypt_data()
287 key->u.tkip.rx[queue].state != TKIP_STATE_PHASE1_HW_UPLOADED) { ieee80211_tkip_decrypt_data()
294 iv32, key->u.tkip.rx[queue].p1k); ieee80211_tkip_decrypt_data()
295 key->u.tkip.rx[queue].state = TKIP_STATE_PHASE1_HW_UPLOADED; ieee80211_tkip_decrypt_data()
298 tkip_mixing_phase2(tk, &key->u.tkip.rx[queue], iv16, rc4key); ieee80211_tkip_decrypt_data()
H A Dwep.c275 ieee80211_crypto_wep_decrypt(struct ieee80211_rx_data *rx) ieee80211_crypto_wep_decrypt() argument
277 struct sk_buff *skb = rx->skb; ieee80211_crypto_wep_decrypt()
286 if (skb_linearize(rx->skb)) ieee80211_crypto_wep_decrypt()
288 if (ieee80211_wep_decrypt(rx->local, rx->skb, rx->key)) ieee80211_crypto_wep_decrypt()
291 if (!pskb_may_pull(rx->skb, ieee80211_hdrlen(fc) + ieee80211_crypto_wep_decrypt()
294 ieee80211_wep_remove_iv(rx->local, rx->skb, rx->key); ieee80211_crypto_wep_decrypt()
296 if (pskb_trim(rx->skb, rx->skb->len - IEEE80211_WEP_ICV_LEN)) ieee80211_crypto_wep_decrypt()
H A Dwep.h30 ieee80211_crypto_wep_decrypt(struct ieee80211_rx_data *rx);
/linux-4.4.14/drivers/net/wireless/ath/carl9170/
H A DMakefile1 carl9170-objs := main.o usb.o cmd.o mac.o phy.o led.o fw.o tx.o rx.o
/linux-4.4.14/net/mac802154/
H A DMakefile2 mac802154-objs := main.o rx.o tx.o mac_cmd.o mib.o \
/linux-4.4.14/net/rxrpc/
H A Daf_rxrpc.c80 static int rxrpc_validate_address(struct rxrpc_sock *rx, rxrpc_validate_address() argument
98 if (srx->transport.family != rx->proto) rxrpc_validate_address()
127 struct rxrpc_sock *rx = rxrpc_sk(sk), *prx; rxrpc_bind() local
131 _enter("%p,%p,%d", rx, saddr, len); rxrpc_bind()
133 ret = rxrpc_validate_address(rx, srx, len); rxrpc_bind()
137 lock_sock(&rx->sk); rxrpc_bind()
139 if (rx->sk.sk_state != RXRPC_UNCONNECTED) { rxrpc_bind()
144 memcpy(&rx->srx, srx, sizeof(rx->srx)); rxrpc_bind()
147 local = rxrpc_lookup_local(&rx->srx); rxrpc_bind()
153 rx->local = local; rxrpc_bind()
162 rx->service_id = service_id; rxrpc_bind()
163 list_add_tail(&rx->listen_link, &local->services); rxrpc_bind()
166 rx->sk.sk_state = RXRPC_SERVER_BOUND; rxrpc_bind()
168 rx->sk.sk_state = RXRPC_CLIENT_BOUND; rxrpc_bind()
171 release_sock(&rx->sk); rxrpc_bind()
179 release_sock(&rx->sk); rxrpc_bind()
191 struct rxrpc_sock *rx = rxrpc_sk(sk); rxrpc_listen() local
194 _enter("%p,%d", rx, backlog); rxrpc_listen()
196 lock_sock(&rx->sk); rxrpc_listen()
198 switch (rx->sk.sk_state) { rxrpc_listen()
208 ASSERT(rx->local != NULL); rxrpc_listen()
210 rx->sk.sk_state = RXRPC_SERVER_LISTENING; rxrpc_listen()
215 release_sock(&rx->sk); rxrpc_listen()
230 struct rxrpc_sock *rx = rxrpc_sk(sock->sk); rxrpc_name_to_transport() local
233 _enter("%p,%p,%d,%d", rx, addr, addr_len, flags); rxrpc_name_to_transport()
235 ASSERT(rx->local != NULL); rxrpc_name_to_transport()
236 ASSERT(rx->sk.sk_state > RXRPC_UNCONNECTED); rxrpc_name_to_transport()
238 if (rx->srx.transport_type != srx->transport_type) rxrpc_name_to_transport()
240 if (rx->srx.transport.family != srx->transport.family) rxrpc_name_to_transport()
249 trans = rxrpc_get_transport(rx->local, peer, gfp); rxrpc_name_to_transport()
278 struct rxrpc_sock *rx = rxrpc_sk(sock->sk); rxrpc_kernel_begin_call() local
283 lock_sock(&rx->sk); rxrpc_kernel_begin_call()
294 trans = rx->trans; rxrpc_kernel_begin_call()
302 service_id = rx->service_id; rxrpc_kernel_begin_call()
307 key = rx->key; rxrpc_kernel_begin_call()
311 bundle = rxrpc_get_bundle(rx, trans, key, service_id, gfp); rxrpc_kernel_begin_call()
317 call = rxrpc_get_client_call(rx, trans, bundle, user_call_ID, true, rxrpc_kernel_begin_call()
323 release_sock(&rx->sk); rxrpc_kernel_begin_call()
360 struct rxrpc_sock *rx = rxrpc_sk(sock->sk); rxrpc_kernel_intercept_rx_messages() local
363 rx->interceptor = interceptor; rxrpc_kernel_intercept_rx_messages()
380 struct rxrpc_sock *rx = rxrpc_sk(sk); rxrpc_connect() local
383 _enter("%p,%p,%d,%d", rx, addr, addr_len, flags); rxrpc_connect()
385 ret = rxrpc_validate_address(rx, srx, addr_len); rxrpc_connect()
391 lock_sock(&rx->sk); rxrpc_connect()
393 switch (rx->sk.sk_state) { rxrpc_connect()
396 ASSERTCMP(rx->local, ==, NULL); rxrpc_connect()
397 rx->srx.srx_family = AF_RXRPC; rxrpc_connect()
398 rx->srx.srx_service = 0; rxrpc_connect()
399 rx->srx.transport_type = srx->transport_type; rxrpc_connect()
400 rx->srx.transport_len = sizeof(sa_family_t); rxrpc_connect()
401 rx->srx.transport.family = srx->transport.family; rxrpc_connect()
402 local = rxrpc_lookup_local(&rx->srx); rxrpc_connect()
404 release_sock(&rx->sk); rxrpc_connect()
407 rx->local = local; rxrpc_connect()
408 rx->sk.sk_state = RXRPC_CLIENT_BOUND; rxrpc_connect()
412 release_sock(&rx->sk); rxrpc_connect()
415 release_sock(&rx->sk); rxrpc_connect()
422 release_sock(&rx->sk); rxrpc_connect()
427 rx->trans = trans; rxrpc_connect()
428 rx->service_id = htons(srx->srx_service); rxrpc_connect()
429 rx->sk.sk_state = RXRPC_CLIENT_CONNECTED; rxrpc_connect()
431 release_sock(&rx->sk); rxrpc_connect()
447 struct rxrpc_sock *rx = rxrpc_sk(sock->sk); rxrpc_sendmsg() local
450 _enter(",{%d},,%zu", rx->sk.sk_state, len); rxrpc_sendmsg()
456 ret = rxrpc_validate_address(rx, m->msg_name, m->msg_namelen); rxrpc_sendmsg()
464 lock_sock(&rx->sk); rxrpc_sendmsg()
476 trans = rx->trans; rxrpc_sendmsg()
481 switch (rx->sk.sk_state) { rxrpc_sendmsg()
484 ret = rxrpc_server_sendmsg(rx, m, len); rxrpc_sendmsg()
494 ret = rxrpc_client_sendmsg(rx, trans, m, len); rxrpc_sendmsg()
502 release_sock(&rx->sk); rxrpc_sendmsg()
515 struct rxrpc_sock *rx = rxrpc_sk(sock->sk); rxrpc_setsockopt() local
521 lock_sock(&rx->sk); rxrpc_setsockopt()
531 if (rx->sk.sk_state != RXRPC_UNCONNECTED) rxrpc_setsockopt()
533 set_bit(RXRPC_SOCK_EXCLUSIVE_CONN, &rx->flags); rxrpc_setsockopt()
538 if (rx->key) rxrpc_setsockopt()
541 if (rx->sk.sk_state != RXRPC_UNCONNECTED) rxrpc_setsockopt()
543 ret = rxrpc_request_key(rx, optval, optlen); rxrpc_setsockopt()
548 if (rx->key) rxrpc_setsockopt()
551 if (rx->sk.sk_state != RXRPC_UNCONNECTED) rxrpc_setsockopt()
553 ret = rxrpc_server_keyring(rx, optval, optlen); rxrpc_setsockopt()
561 if (rx->sk.sk_state != RXRPC_UNCONNECTED) rxrpc_setsockopt()
570 rx->min_sec_level = min_sec_level; rxrpc_setsockopt()
581 release_sock(&rx->sk); rxrpc_setsockopt()
617 struct rxrpc_sock *rx; rxrpc_create() local
645 rx = rxrpc_sk(sk); rxrpc_create()
646 rx->proto = protocol; rxrpc_create()
647 rx->calls = RB_ROOT; rxrpc_create()
649 INIT_LIST_HEAD(&rx->listen_link); rxrpc_create()
650 INIT_LIST_HEAD(&rx->secureq); rxrpc_create()
651 INIT_LIST_HEAD(&rx->acceptq); rxrpc_create()
652 rwlock_init(&rx->call_lock); rxrpc_create()
653 memset(&rx->srx, 0, sizeof(rx->srx)); rxrpc_create()
655 _leave(" = 0 [%p]", rx); rxrpc_create()
683 struct rxrpc_sock *rx = rxrpc_sk(sk); rxrpc_release_sock() local
695 ASSERTCMP(rx->listen_link.next, !=, LIST_POISON1); rxrpc_release_sock()
697 if (!list_empty(&rx->listen_link)) { rxrpc_release_sock()
698 write_lock_bh(&rx->local->services_lock); rxrpc_release_sock()
699 list_del(&rx->listen_link); rxrpc_release_sock()
700 write_unlock_bh(&rx->local->services_lock); rxrpc_release_sock()
704 rxrpc_release_calls_on_socket(rx); rxrpc_release_sock()
708 if (rx->conn) { rxrpc_release_sock()
709 rxrpc_put_connection(rx->conn); rxrpc_release_sock()
710 rx->conn = NULL; rxrpc_release_sock()
713 if (rx->bundle) { rxrpc_release_sock()
714 rxrpc_put_bundle(rx->trans, rx->bundle); rxrpc_release_sock()
715 rx->bundle = NULL; rxrpc_release_sock()
717 if (rx->trans) { rxrpc_release_sock()
718 rxrpc_put_transport(rx->trans); rxrpc_release_sock()
719 rx->trans = NULL; rxrpc_release_sock()
721 if (rx->local) { rxrpc_release_sock()
722 rxrpc_put_local(rx->local); rxrpc_release_sock()
723 rx->local = NULL; rxrpc_release_sock()
726 key_put(rx->key); rxrpc_release_sock()
727 rx->key = NULL; rxrpc_release_sock()
728 key_put(rx->securities); rxrpc_release_sock()
729 rx->securities = NULL; rxrpc_release_sock()
H A Dar-recvmsg.c23 void rxrpc_remove_user_ID(struct rxrpc_sock *rx, struct rxrpc_call *call) rxrpc_remove_user_ID() argument
28 write_lock_bh(&rx->call_lock); rxrpc_remove_user_ID()
31 write_unlock_bh(&rx->call_lock); rxrpc_remove_user_ID()
51 struct rxrpc_sock *rx = rxrpc_sk(sock->sk); rxrpc_recvmsg() local
66 timeo = sock_rcvtimeo(&rx->sk, flags & MSG_DONTWAIT); rxrpc_recvmsg()
69 lock_sock(&rx->sk); rxrpc_recvmsg()
74 if (RB_EMPTY_ROOT(&rx->calls)) { rxrpc_recvmsg()
77 if (rx->sk.sk_state != RXRPC_SERVER_LISTENING) { rxrpc_recvmsg()
78 release_sock(&rx->sk); rxrpc_recvmsg()
86 skb = skb_peek(&rx->sk.sk_receive_queue); rxrpc_recvmsg()
94 release_sock(&rx->sk); rxrpc_recvmsg()
95 prepare_to_wait_exclusive(sk_sleep(&rx->sk), &wait, rxrpc_recvmsg()
97 ret = sock_error(&rx->sk); rxrpc_recvmsg()
101 if (skb_queue_empty(&rx->sk.sk_receive_queue)) { rxrpc_recvmsg()
106 finish_wait(sk_sleep(&rx->sk), &wait); rxrpc_recvmsg()
107 lock_sock(&rx->sk); rxrpc_recvmsg()
124 if (skb_dequeue(&rx->sk.sk_receive_queue) != skb) rxrpc_recvmsg()
135 release_sock(&rx->sk); rxrpc_recvmsg()
153 sock_recv_timestamp(msg, &rx->sk, skb); rxrpc_recvmsg()
215 if (skb_dequeue(&rx->sk.sk_receive_queue) != rxrpc_recvmsg()
235 if (skb == (struct sk_buff *) &rx->sk.sk_receive_queue) rxrpc_recvmsg()
241 if (skb_dequeue(&rx->sk.sk_receive_queue) != skb) rxrpc_recvmsg()
249 release_sock(&rx->sk); rxrpc_recvmsg()
268 if (skb_dequeue(&rx->sk.sk_receive_queue) != skb) rxrpc_recvmsg()
320 if (skb_dequeue(&rx->sk.sk_receive_queue) != skb) rxrpc_recvmsg()
323 rxrpc_remove_user_ID(rx, call); rxrpc_recvmsg()
326 release_sock(&rx->sk); rxrpc_recvmsg()
335 release_sock(&rx->sk); rxrpc_recvmsg()
345 finish_wait(sk_sleep(&rx->sk), &wait); rxrpc_recvmsg()
H A Dar-accept.c74 struct rxrpc_sock *rx, rxrpc_accept_incoming_call()
123 call = rxrpc_incoming_call(rx, conn, &sp->hdr, GFP_NOIO); rxrpc_accept_incoming_call()
133 if (rx->sk.sk_state == RXRPC_CLOSE) rxrpc_accept_incoming_call()
136 write_lock(&rx->call_lock); rxrpc_accept_incoming_call()
144 list_add_tail(&call->accept_link, &rx->secureq); rxrpc_accept_incoming_call()
152 list_add_tail(&call->accept_link, &rx->acceptq); rxrpc_accept_incoming_call()
171 write_unlock(&rx->call_lock); rxrpc_accept_incoming_call()
213 struct rxrpc_sock *rx; rxrpc_accept_incoming_calls() local
261 list_for_each_entry(rx, &local->services, listen_link) { rxrpc_accept_incoming_calls()
262 if (rx->service_id == service_id && rxrpc_accept_incoming_calls()
263 rx->sk.sk_state != RXRPC_CLOSE) rxrpc_accept_incoming_calls()
270 _debug("found service %hd", ntohs(rx->service_id)); rxrpc_accept_incoming_calls()
271 if (sk_acceptq_is_full(&rx->sk)) rxrpc_accept_incoming_calls()
273 sk_acceptq_added(&rx->sk); rxrpc_accept_incoming_calls()
274 sock_hold(&rx->sk); rxrpc_accept_incoming_calls()
277 ret = rxrpc_accept_incoming_call(local, rx, skb, &srx); rxrpc_accept_incoming_calls()
279 sk_acceptq_removed(&rx->sk); rxrpc_accept_incoming_calls()
280 sock_put(&rx->sk); rxrpc_accept_incoming_calls()
319 struct rxrpc_call *rxrpc_accept_call(struct rxrpc_sock *rx, rxrpc_accept_call() argument
330 write_lock(&rx->call_lock); rxrpc_accept_call()
333 if (list_empty(&rx->acceptq)) rxrpc_accept_call()
338 pp = &rx->calls.rb_node; rxrpc_accept_call()
353 call = list_entry(rx->acceptq.next, struct rxrpc_call, accept_link); rxrpc_accept_call()
355 sk_acceptq_removed(&rx->sk); rxrpc_accept_call()
379 rb_insert_color(&call->sock_node, &rx->calls); rxrpc_accept_call()
388 write_unlock(&rx->call_lock); rxrpc_accept_call()
404 write_unlock(&rx->call_lock); rxrpc_accept_call()
413 int rxrpc_reject_call(struct rxrpc_sock *rx) rxrpc_reject_call() argument
422 write_lock(&rx->call_lock); rxrpc_reject_call()
425 if (list_empty(&rx->acceptq)) rxrpc_reject_call()
429 call = list_entry(rx->acceptq.next, struct rxrpc_call, accept_link); rxrpc_reject_call()
431 sk_acceptq_removed(&rx->sk); rxrpc_reject_call()
467 write_unlock(&rx->call_lock); rxrpc_reject_call()
73 rxrpc_accept_incoming_call(struct rxrpc_local *local, struct rxrpc_sock *rx, struct sk_buff *skb, struct sockaddr_rxrpc *srx) rxrpc_accept_incoming_call() argument
H A Dar-call.c252 struct rxrpc_sock *rx, rxrpc_alloc_client_call()
262 ASSERT(rx != NULL); rxrpc_alloc_client_call()
270 sock_hold(&rx->sk); rxrpc_alloc_client_call()
271 call->socket = rx; rxrpc_alloc_client_call()
274 ret = rxrpc_connect_call(rx, trans, bundle, call, gfp); rxrpc_alloc_client_call()
281 call->proto = rx->proto; rxrpc_alloc_client_call()
315 struct rxrpc_call *rxrpc_get_client_call(struct rxrpc_sock *rx, rxrpc_get_client_call() argument
326 rx, trans ? trans->debug_id : -1, bundle ? bundle->debug_id : -1, rxrpc_get_client_call()
331 read_lock(&rx->call_lock); rxrpc_get_client_call()
333 p = rx->calls.rb_node; rxrpc_get_client_call()
345 read_unlock(&rx->call_lock); rxrpc_get_client_call()
352 candidate = rxrpc_alloc_client_call(rx, trans, bundle, gfp); rxrpc_get_client_call()
361 write_lock(&rx->call_lock); rxrpc_get_client_call()
363 pp = &rx->calls.rb_node; rxrpc_get_client_call()
383 rb_insert_color(&call->sock_node, &rx->calls); rxrpc_get_client_call()
384 write_unlock(&rx->call_lock); rxrpc_get_client_call()
398 read_unlock(&rx->call_lock); rxrpc_get_client_call()
405 write_unlock(&rx->call_lock); rxrpc_get_client_call()
415 struct rxrpc_call *rxrpc_incoming_call(struct rxrpc_sock *rx, rxrpc_incoming_call() argument
426 ASSERT(rx != NULL); rxrpc_incoming_call()
432 candidate->socket = rx; rxrpc_incoming_call()
510 sock_hold(&rx->sk); rxrpc_incoming_call()
523 call->proto = rx->proto; rxrpc_incoming_call()
574 struct rxrpc_call *rxrpc_find_server_call(struct rxrpc_sock *rx, rxrpc_find_server_call() argument
580 _enter("%p,%lx", rx, user_call_ID); rxrpc_find_server_call()
584 read_lock(&rx->call_lock); rxrpc_find_server_call()
586 p = rx->calls.rb_node; rxrpc_find_server_call()
598 read_unlock(&rx->call_lock); rxrpc_find_server_call()
605 read_unlock(&rx->call_lock); rxrpc_find_server_call()
616 struct rxrpc_sock *rx = call->socket; rxrpc_release_call() local
633 write_lock_bh(&rx->call_lock); rxrpc_release_call()
639 sk_acceptq_removed(&rx->sk); rxrpc_release_call()
641 rb_erase(&call->sock_node, &rx->calls); rxrpc_release_call()
645 write_unlock_bh(&rx->call_lock); rxrpc_release_call()
780 void rxrpc_release_calls_on_socket(struct rxrpc_sock *rx) rxrpc_release_calls_on_socket() argument
785 _enter("%p", rx); rxrpc_release_calls_on_socket()
787 read_lock_bh(&rx->call_lock); rxrpc_release_calls_on_socket()
790 for (p = rb_first(&rx->calls); p; p = rb_next(p)) { rxrpc_release_calls_on_socket()
796 list_for_each_entry(call, &rx->secureq, accept_link) { rxrpc_release_calls_on_socket()
800 list_for_each_entry(call, &rx->acceptq, accept_link) { rxrpc_release_calls_on_socket()
804 read_unlock_bh(&rx->call_lock); rxrpc_release_calls_on_socket()
251 rxrpc_alloc_client_call( struct rxrpc_sock *rx, struct rxrpc_transport *trans, struct rxrpc_conn_bundle *bundle, gfp_t gfp) rxrpc_alloc_client_call() argument
H A Dar-connection.c69 struct rxrpc_conn_bundle *rxrpc_get_bundle(struct rxrpc_sock *rx, rxrpc_get_bundle() argument
79 rx, key_serial(key), trans->debug_id, ntohs(service_id)); rxrpc_get_bundle()
81 if (rx->trans == trans && rx->bundle) { rxrpc_get_bundle()
82 atomic_inc(&rx->bundle->usage); rxrpc_get_bundle()
83 return rx->bundle; rxrpc_get_bundle()
139 if (!rx->bundle && rx->sk.sk_state == RXRPC_CLIENT_CONNECTED) { rxrpc_get_bundle()
141 rx->bundle = bundle; rxrpc_get_bundle()
151 if (!rx->bundle && rx->sk.sk_state == RXRPC_CLIENT_CONNECTED) { rxrpc_get_bundle()
153 rx->bundle = bundle; rxrpc_get_bundle()
164 if (!rx->bundle && rx->sk.sk_state == RXRPC_CLIENT_CONNECTED) { rxrpc_get_bundle()
166 rx->bundle = bundle; rxrpc_get_bundle()
335 static int rxrpc_connect_exclusive(struct rxrpc_sock *rx, rxrpc_connect_exclusive() argument
346 conn = rx->conn; rxrpc_connect_exclusive()
365 conn->security_level = rx->min_sec_level; rxrpc_connect_exclusive()
366 conn->key = key_get(rx->key); rxrpc_connect_exclusive()
387 rx->conn = conn; rxrpc_connect_exclusive()
429 int rxrpc_connect_call(struct rxrpc_sock *rx, rxrpc_connect_call() argument
440 _enter("%p,%lx,", rx, call->user_call_ID); rxrpc_connect_call()
442 if (test_bit(RXRPC_SOCK_EXCLUSIVE_CONN, &rx->flags)) rxrpc_connect_call()
443 return rxrpc_connect_exclusive(rx, trans, bundle->service_id, rxrpc_connect_call()
542 candidate->security_level = rx->min_sec_level; rxrpc_connect_call()
H A Dar-output.c26 static int rxrpc_send_data(struct rxrpc_sock *rx,
33 static int rxrpc_sendmsg_cmsg(struct rxrpc_sock *rx, struct msghdr *msg, rxrpc_sendmsg_cmsg() argument
131 int rxrpc_client_sendmsg(struct rxrpc_sock *rx, struct rxrpc_transport *trans, rxrpc_client_sendmsg() argument
147 ret = rxrpc_sendmsg_cmsg(rx, msg, &user_call_ID, &cmd, &abort_code, rxrpc_client_sendmsg()
154 service_id = rx->service_id; rxrpc_client_sendmsg()
160 key = rx->key; rxrpc_client_sendmsg()
161 if (key && !rx->key->payload.data[0]) rxrpc_client_sendmsg()
163 bundle = rxrpc_get_bundle(rx, trans, key, service_id, rxrpc_client_sendmsg()
169 call = rxrpc_get_client_call(rx, trans, bundle, user_call_ID, rxrpc_client_sendmsg()
192 ret = rxrpc_send_data(rx, call, msg, len); rxrpc_client_sendmsg()
272 int rxrpc_server_sendmsg(struct rxrpc_sock *rx, struct msghdr *msg, size_t len) rxrpc_server_sendmsg() argument
282 ret = rxrpc_sendmsg_cmsg(rx, msg, &user_call_ID, &cmd, &abort_code, rxrpc_server_sendmsg()
288 call = rxrpc_accept_call(rx, user_call_ID); rxrpc_server_sendmsg()
295 call = rxrpc_find_server_call(rx, user_call_ID); rxrpc_server_sendmsg()
313 ret = rxrpc_send_data(rx, call, msg, len); rxrpc_server_sendmsg()
396 static int rxrpc_wait_for_tx_window(struct rxrpc_sock *rx, rxrpc_wait_for_tx_window() argument
420 release_sock(&rx->sk); rxrpc_wait_for_tx_window()
422 lock_sock(&rx->sk); rxrpc_wait_for_tx_window()
494 /* attempt to cancel the rx-ACK timer, deferring reply transmission if rxrpc_queue_packet()
520 static int rxrpc_send_data(struct rxrpc_sock *rx, rxrpc_send_data() argument
526 struct sock *sk = &rx->sk; rxrpc_send_data()
556 ret = rxrpc_wait_for_tx_window(rx, call, rxrpc_send_data()
H A Dar-security.c167 struct rxrpc_sock *rx; rxrpc_init_server_conn_security() local
184 list_for_each_entry(rx, &local->services, listen_link) { rxrpc_init_server_conn_security()
185 if (rx->service_id == conn->service_id) rxrpc_init_server_conn_security()
196 if (!rx->securities) { rxrpc_init_server_conn_security()
204 kref = keyring_search(make_key_ref(rx->securities, 1UL), rxrpc_init_server_conn_security()
/linux-4.4.14/drivers/net/ethernet/pasemi/
H A Dpasemi_mac_ethtool.c31 { "rx-drops" },
32 { "rx-bytes" },
33 { "rx-packets" },
34 { "rx-broadcast-packets" },
35 { "rx-multicast-packets" },
36 { "rx-crc-errors" },
37 { "rx-undersize-errors" },
38 { "rx-oversize-errors" },
39 { "rx-short-fragment-errors" },
40 { "rx-jabber-errors" },
41 { "rx-64-byte-packets" },
42 { "rx-65-127-byte-packets" },
43 { "rx-128-255-byte-packets" },
44 { "rx-256-511-byte-packets" },
45 { "rx-512-1023-byte-packets" },
46 { "rx-1024-1518-byte-packets" },
47 { "rx-pause-frames" },
117 ering->rx_pending = RING_USED(mac->rx)/4; pasemi_mac_ethtool_get_ringparam()
H A Dpasemi_mac.h93 struct pasemi_mac_rxring *rx; member in struct:pasemi_mac
96 char rx_irq_name[10]; /* "eth%d rx" */
112 #define RX_DESC(rx, num) ((rx)->chan.ring_virt[(num) & (RX_RING_SIZE-1)])
113 #define RX_DESC_INFO(rx, num) ((rx)->ring_info[(num) & (RX_RING_SIZE-1)])
114 #define RX_BUFF(rx, num) ((rx)->buffers[(num) & (RX_RING_SIZE-1)])
H A Dpasemi_mac.c122 return mac->rx; rx_ring()
480 mac->rx = ring; pasemi_mac_setup_rx_resources()
591 struct pasemi_mac_rxring *rx = rx_ring(mac); pasemi_mac_free_rx_buffers() local
596 info = &RX_DESC_INFO(rx, i); pasemi_mac_free_rx_buffers()
609 RX_BUFF(rx, i) = 0; pasemi_mac_free_rx_buffers()
621 mac->rx = NULL; pasemi_mac_free_rx_resources()
628 struct pasemi_mac_rxring *rx = rx_ring(mac); pasemi_mac_replenish_rx_ring() local
636 struct pasemi_mac_buffer *info = &RX_DESC_INFO(rx, fill); pasemi_mac_replenish_rx_ring()
637 u64 *buff = &RX_BUFF(rx, fill); pasemi_mac_replenish_rx_ring()
675 struct pasemi_mac_rxring *rx = rx_ring(mac); pasemi_mac_restart_rx_intr() local
681 pcnt = *rx->chan.status & PAS_STATUS_PCNT_M; pasemi_mac_restart_rx_intr()
685 if (*rx->chan.status & PAS_STATUS_TIMER) pasemi_mac_restart_rx_intr()
688 write_iob_reg(PAS_IOB_DMA_RXCH_RESET(mac->rx->chan.chno), reg); pasemi_mac_restart_rx_intr()
716 printk(KERN_ERR "pasemi_mac: rx error. macrx %016llx, rx status %llx\n", pasemi_mac_rx_error()
740 static int pasemi_mac_clean_rx(struct pasemi_mac_rxring *rx, pasemi_mac_clean_rx() argument
743 const struct pasemi_dmachan *chan = &rx->chan; pasemi_mac_clean_rx()
744 struct pasemi_mac *mac = rx->mac; pasemi_mac_clean_rx()
757 spin_lock(&rx->lock); pasemi_mac_clean_rx()
759 n = rx->next_to_clean; pasemi_mac_clean_rx()
761 prefetch(&RX_DESC(rx, n)); pasemi_mac_clean_rx()
764 macrx = RX_DESC(rx, n); pasemi_mac_clean_rx()
765 prefetch(&RX_DESC(rx, n+4)); pasemi_mac_clean_rx()
778 eval = (RX_DESC(rx, n+1) & XCT_RXRES_8B_EVAL_M) >> pasemi_mac_clean_rx()
782 dma = (RX_DESC(rx, n+2) & XCT_PTR_ADDR_M); pasemi_mac_clean_rx()
783 info = &RX_DESC_INFO(rx, buf_index); pasemi_mac_clean_rx()
823 RX_DESC(rx, n) = 0; pasemi_mac_clean_rx()
824 RX_DESC(rx, n+1) = 0; pasemi_mac_clean_rx()
829 RX_BUFF(rx, buf_index) = 0; pasemi_mac_clean_rx()
848 write_dma_reg(PAS_DMA_RXCHAN_INCR(mac->rx->chan.chno), count << 1); pasemi_mac_clean_rx()
1152 write_iob_reg(PAS_IOB_DMA_RXCH_CFG(mac->rx->chan.chno), pasemi_mac_open()
1159 PAS_MAC_IPC_CHNL_DCHNO(mac->rx->chan.chno) | pasemi_mac_open()
1160 PAS_MAC_IPC_CHNL_BCH(mac->rx->chan.chno)); pasemi_mac_open()
1162 /* enable rx if */ pasemi_mac_open()
1170 /* enable rx channel */ pasemi_mac_open()
1228 snprintf(mac->rx_irq_name, sizeof(mac->rx_irq_name), "%s rx", pasemi_mac_open()
1231 ret = request_irq(mac->rx->chan.irq, pasemi_mac_rx_intr, 0, pasemi_mac_open()
1232 mac->rx_irq_name, mac->rx); pasemi_mac_open()
1235 mac->rx->chan.irq, ret); pasemi_mac_open()
1302 "Failed to stop rx channel, ccmdsta 08%x\n", sta); pasemi_mac_pause_rxchan()
1321 "Failed to stop rx interface, rcmdsta %08x\n", sta); pasemi_mac_pause_rxint()
1372 free_irq(mac->rx->chan.irq, mac->rx); pasemi_mac_close()
1640 disable_irq(mac->rx->chan.irq); pasemi_mac_netpoll()
1641 pasemi_mac_rx_intr(mac->rx->chan.irq, mac->rx); pasemi_mac_netpoll()
1642 enable_irq(mac->rx->chan.irq); pasemi_mac_netpoll()
1662 * interface ring, then finally re-fill the rx ring with pasemi_mac_change_mtu()
1865 pasemi_dma_free_chan(&mac->rx->chan); pasemi_mac_remove()
/linux-4.4.14/drivers/staging/gdm724x/
H A Dgdm_mux.c123 static struct mux_rx *get_rx_struct(struct rx_cxt *rx) get_rx_struct() argument
128 spin_lock_irqsave(&rx->free_list_lock, flags); get_rx_struct()
130 if (list_empty(&rx->rx_free_list)) { get_rx_struct()
131 spin_unlock_irqrestore(&rx->free_list_lock, flags); get_rx_struct()
135 r = list_entry(rx->rx_free_list.prev, struct mux_rx, free_list); get_rx_struct()
138 spin_unlock_irqrestore(&rx->free_list_lock, flags); get_rx_struct()
143 static void put_rx_struct(struct rx_cxt *rx, struct mux_rx *r) put_rx_struct() argument
147 spin_lock_irqsave(&rx->free_list_lock, flags); put_rx_struct()
148 list_add_tail(&r->free_list, &rx->rx_free_list); put_rx_struct()
149 spin_unlock_irqrestore(&rx->free_list_lock, flags); put_rx_struct()
223 struct rx_cxt *rx = &mux_dev->rx; do_rx() local
228 spin_lock_irqsave(&rx->to_host_lock, flags); do_rx()
229 if (list_empty(&rx->to_host_list)) { do_rx()
230 spin_unlock_irqrestore(&rx->to_host_lock, flags); do_rx()
233 r = list_entry(rx->to_host_list.next, struct mux_rx, do_rx()
236 spin_unlock_irqrestore(&rx->to_host_lock, flags); do_rx()
242 put_rx_struct(rx, r); do_rx()
246 static void remove_rx_submit_list(struct mux_rx *r, struct rx_cxt *rx) remove_rx_submit_list() argument
251 spin_lock_irqsave(&rx->submit_list_lock, flags); remove_rx_submit_list()
252 list_for_each_entry_safe(r_remove, r_remove_next, &rx->rx_submit_list, remove_rx_submit_list()
257 spin_unlock_irqrestore(&rx->submit_list_lock, flags); remove_rx_submit_list()
264 struct rx_cxt *rx = &mux_dev->rx; gdm_mux_rcv_complete() local
267 remove_rx_submit_list(r, rx); gdm_mux_rcv_complete()
273 put_rx_struct(rx, r); gdm_mux_rcv_complete()
276 spin_lock_irqsave(&rx->to_host_lock, flags); gdm_mux_rcv_complete()
277 list_add_tail(&r->to_host_list, &rx->to_host_list); gdm_mux_rcv_complete()
279 spin_unlock_irqrestore(&rx->to_host_lock, flags); gdm_mux_rcv_complete()
289 struct rx_cxt *rx = &mux_dev->rx; gdm_mux_recv() local
298 r = get_rx_struct(rx); gdm_mux_recv()
317 spin_lock_irqsave(&rx->submit_list_lock, flags); gdm_mux_recv()
318 list_add_tail(&r->rx_submit_list, &rx->rx_submit_list); gdm_mux_recv()
319 spin_unlock_irqrestore(&rx->submit_list_lock, flags); gdm_mux_recv()
324 spin_lock_irqsave(&rx->submit_list_lock, flags); gdm_mux_recv()
326 spin_unlock_irqrestore(&rx->submit_list_lock, flags); gdm_mux_recv()
328 put_rx_struct(rx, r); gdm_mux_recv()
443 struct rx_cxt *rx = &mux_dev->rx; release_usb() local
449 spin_lock_irqsave(&rx->submit_list_lock, flags); release_usb()
450 list_for_each_entry_safe(r, r_next, &rx->rx_submit_list, release_usb()
452 spin_unlock_irqrestore(&rx->submit_list_lock, flags); release_usb()
454 spin_lock_irqsave(&rx->submit_list_lock, flags); release_usb()
456 spin_unlock_irqrestore(&rx->submit_list_lock, flags); release_usb()
458 spin_lock_irqsave(&rx->free_list_lock, flags); release_usb()
459 list_for_each_entry_safe(r, r_next, &rx->rx_free_list, free_list) { release_usb()
463 spin_unlock_irqrestore(&rx->free_list_lock, flags); release_usb()
465 spin_lock_irqsave(&rx->to_host_lock, flags); release_usb()
466 list_for_each_entry_safe(r, r_next, &rx->to_host_list, to_host_list) { release_usb()
472 spin_unlock_irqrestore(&rx->to_host_lock, flags); release_usb()
478 struct rx_cxt *rx = &mux_dev->rx; init_usb() local
483 INIT_LIST_HEAD(&rx->to_host_list); init_usb()
484 INIT_LIST_HEAD(&rx->rx_submit_list); init_usb()
485 INIT_LIST_HEAD(&rx->rx_free_list); init_usb()
486 spin_lock_init(&rx->to_host_lock); init_usb()
487 spin_lock_init(&rx->submit_list_lock); init_usb()
488 spin_lock_init(&rx->free_list_lock); init_usb()
497 list_add(&r->free_list, &rx->rx_free_list); init_usb()
597 struct rx_cxt *rx; gdm_mux_suspend() local
603 rx = &mux_dev->rx; gdm_mux_suspend()
612 spin_lock_irqsave(&rx->submit_list_lock, flags); gdm_mux_suspend()
613 list_for_each_entry_safe(r, r_next, &rx->rx_submit_list, gdm_mux_suspend()
615 spin_unlock_irqrestore(&rx->submit_list_lock, flags); gdm_mux_suspend()
617 spin_lock_irqsave(&rx->submit_list_lock, flags); gdm_mux_suspend()
619 spin_unlock_irqrestore(&rx->submit_list_lock, flags); gdm_mux_suspend()
H A Dgdm_usb.c221 static struct usb_rx *get_rx_struct(struct rx_cxt *rx, int *no_spc) get_rx_struct() argument
226 spin_lock_irqsave(&rx->rx_lock, flags); get_rx_struct()
228 if (list_empty(&rx->free_list)) { get_rx_struct()
229 spin_unlock_irqrestore(&rx->rx_lock, flags); get_rx_struct()
233 r = list_entry(rx->free_list.next, struct usb_rx, free_list); get_rx_struct()
236 rx->avail_count--; get_rx_struct()
238 *no_spc = list_empty(&rx->free_list) ? 1 : 0; get_rx_struct()
240 spin_unlock_irqrestore(&rx->rx_lock, flags); get_rx_struct()
245 static void put_rx_struct(struct rx_cxt *rx, struct usb_rx *r) put_rx_struct() argument
249 spin_lock_irqsave(&rx->rx_lock, flags); put_rx_struct()
251 list_add_tail(&r->free_list, &rx->free_list); put_rx_struct()
252 rx->avail_count++; put_rx_struct()
254 spin_unlock_irqrestore(&rx->rx_lock, flags); put_rx_struct()
259 struct rx_cxt *rx = &udev->rx; release_usb() local
283 spin_lock_irqsave(&rx->submit_lock, flags); release_usb()
284 list_for_each_entry_safe(r, r_next, &rx->rx_submit_list, release_usb()
286 spin_unlock_irqrestore(&rx->submit_lock, flags); release_usb()
288 spin_lock_irqsave(&rx->submit_lock, flags); release_usb()
290 spin_unlock_irqrestore(&rx->submit_lock, flags); release_usb()
292 spin_lock_irqsave(&rx->rx_lock, flags); release_usb()
293 list_for_each_entry_safe(r, r_next, &rx->free_list, free_list) { release_usb()
297 spin_unlock_irqrestore(&rx->rx_lock, flags); release_usb()
299 spin_lock_irqsave(&rx->to_host_lock, flags); release_usb()
300 list_for_each_entry_safe(r, r_next, &rx->to_host_list, to_host_list) { release_usb()
306 spin_unlock_irqrestore(&rx->to_host_lock, flags); release_usb()
314 struct rx_cxt *rx = &udev->rx; init_usb() local
326 INIT_LIST_HEAD(&rx->rx_submit_list); init_usb()
327 INIT_LIST_HEAD(&rx->free_list); init_usb()
328 INIT_LIST_HEAD(&rx->to_host_list); init_usb()
330 spin_lock_init(&rx->rx_lock); init_usb()
331 spin_lock_init(&rx->submit_lock); init_usb()
332 spin_lock_init(&rx->to_host_lock); init_usb()
335 rx->avail_count = 0; init_usb()
357 list_add(&r->free_list, &rx->free_list); init_usb()
358 rx->avail_count++; init_usb()
394 struct rx_cxt *rx = &udev->rx; do_rx() local
403 spin_lock_irqsave(&rx->to_host_lock, flags); do_rx()
404 if (list_empty(&rx->to_host_list)) { do_rx()
405 spin_unlock_irqrestore(&rx->to_host_lock, flags); do_rx()
408 r = list_entry(rx->to_host_list.next, do_rx()
411 spin_unlock_irqrestore(&rx->to_host_lock, flags); do_rx()
441 put_rx_struct(rx, r); do_rx()
450 static void remove_rx_submit_list(struct usb_rx *r, struct rx_cxt *rx) remove_rx_submit_list() argument
455 spin_lock_irqsave(&rx->submit_lock, flags); remove_rx_submit_list()
457 &rx->rx_submit_list, rx_submit_list) { remove_rx_submit_list()
463 spin_unlock_irqrestore(&rx->submit_lock, flags); remove_rx_submit_list()
469 struct rx_cxt *rx = r->rx; gdm_usb_rcv_complete() local
471 struct lte_udev *udev = container_of(r->rx, struct lte_udev, rx); gdm_usb_rcv_complete()
474 remove_rx_submit_list(r, rx); gdm_usb_rcv_complete()
477 spin_lock_irqsave(&rx->to_host_lock, flags); gdm_usb_rcv_complete()
478 list_add_tail(&r->to_host_list, &rx->to_host_list); gdm_usb_rcv_complete()
480 spin_unlock_irqrestore(&rx->to_host_lock, flags); gdm_usb_rcv_complete()
486 put_rx_struct(rx, r); gdm_usb_rcv_complete()
500 struct rx_cxt *rx = &udev->rx; gdm_usb_recv() local
511 r = get_rx_struct(rx, &no_spc); gdm_usb_recv()
521 r->rx = rx; gdm_usb_recv()
531 spin_lock_irqsave(&rx->submit_lock, flags); gdm_usb_recv()
532 list_add_tail(&r->rx_submit_list, &rx->rx_submit_list); gdm_usb_recv()
533 spin_unlock_irqrestore(&rx->submit_lock, flags); gdm_usb_recv()
541 spin_lock_irqsave(&rx->submit_lock, flags); gdm_usb_recv()
543 spin_unlock_irqrestore(&rx->submit_lock, flags); gdm_usb_recv()
546 put_rx_struct(rx, r); gdm_usb_recv()
923 struct rx_cxt *rx; gdm_usb_suspend() local
930 rx = &udev->rx; gdm_usb_suspend()
938 spin_lock_irqsave(&rx->submit_lock, flags); gdm_usb_suspend()
939 list_for_each_entry_safe(r, r_next, &rx->rx_submit_list, gdm_usb_suspend()
941 spin_unlock_irqrestore(&rx->submit_lock, flags); gdm_usb_suspend()
943 spin_lock_irqsave(&rx->submit_lock, flags); gdm_usb_suspend()
945 spin_unlock_irqrestore(&rx->submit_lock, flags); gdm_usb_suspend()
955 struct rx_cxt *rx; gdm_usb_resume() local
962 rx = &udev->rx; gdm_usb_resume()
970 spin_lock_irqsave(&rx->rx_lock, flags); gdm_usb_resume()
971 issue_count = rx->avail_count - MAX_RX_SUBMIT_COUNT; gdm_usb_resume()
972 spin_unlock_irqrestore(&rx->rx_lock, flags); gdm_usb_resume()
H A Dgdm_usb.h68 struct rx_cxt *rx; member in struct:usb_rx
98 struct rx_cxt rx; member in struct:lte_udev
/linux-4.4.14/drivers/net/wireless/ti/wl18xx/
H A Ddebugfs.c95 WL18XX_DEBUGFS_FWSTATS_FILE(rx, rx_beacon_early_term, "%u");
96 WL18XX_DEBUGFS_FWSTATS_FILE(rx, rx_out_of_mpdu_nodes, "%u");
97 WL18XX_DEBUGFS_FWSTATS_FILE(rx, rx_hdr_overflow, "%u");
98 WL18XX_DEBUGFS_FWSTATS_FILE(rx, rx_dropped_frame, "%u");
99 WL18XX_DEBUGFS_FWSTATS_FILE(rx, rx_done, "%u");
100 WL18XX_DEBUGFS_FWSTATS_FILE(rx, rx_defrag, "%u");
101 WL18XX_DEBUGFS_FWSTATS_FILE(rx, rx_defrag_end, "%u");
102 WL18XX_DEBUGFS_FWSTATS_FILE(rx, rx_cmplt, "%u");
103 WL18XX_DEBUGFS_FWSTATS_FILE(rx, rx_pre_complt, "%u");
104 WL18XX_DEBUGFS_FWSTATS_FILE(rx, rx_cmplt_task, "%u");
105 WL18XX_DEBUGFS_FWSTATS_FILE(rx, rx_phy_hdr, "%u");
106 WL18XX_DEBUGFS_FWSTATS_FILE(rx, rx_timeout, "%u");
107 WL18XX_DEBUGFS_FWSTATS_FILE(rx, rx_rts_timeout, "%u");
108 WL18XX_DEBUGFS_FWSTATS_FILE(rx, rx_timeout_wa, "%u");
109 WL18XX_DEBUGFS_FWSTATS_FILE(rx, defrag_called, "%u");
110 WL18XX_DEBUGFS_FWSTATS_FILE(rx, defrag_init_called, "%u");
111 WL18XX_DEBUGFS_FWSTATS_FILE(rx, defrag_in_process_called, "%u");
112 WL18XX_DEBUGFS_FWSTATS_FILE(rx, defrag_tkip_called, "%u");
113 WL18XX_DEBUGFS_FWSTATS_FILE(rx, defrag_need_defrag, "%u");
114 WL18XX_DEBUGFS_FWSTATS_FILE(rx, defrag_decrypt_failed, "%u");
115 WL18XX_DEBUGFS_FWSTATS_FILE(rx, decrypt_key_not_found, "%u");
116 WL18XX_DEBUGFS_FWSTATS_FILE(rx, defrag_need_decrypt, "%u");
117 WL18XX_DEBUGFS_FWSTATS_FILE(rx, rx_tkip_replays, "%u");
118 WL18XX_DEBUGFS_FWSTATS_FILE(rx, rx_xfr, "%u");
423 DEBUGFS_FWSTATS_ADD(rx, rx_beacon_early_term); wl18xx_debugfs_add_files()
424 DEBUGFS_FWSTATS_ADD(rx, rx_out_of_mpdu_nodes); wl18xx_debugfs_add_files()
425 DEBUGFS_FWSTATS_ADD(rx, rx_hdr_overflow); wl18xx_debugfs_add_files()
426 DEBUGFS_FWSTATS_ADD(rx, rx_dropped_frame); wl18xx_debugfs_add_files()
427 DEBUGFS_FWSTATS_ADD(rx, rx_done); wl18xx_debugfs_add_files()
428 DEBUGFS_FWSTATS_ADD(rx, rx_defrag); wl18xx_debugfs_add_files()
429 DEBUGFS_FWSTATS_ADD(rx, rx_defrag_end); wl18xx_debugfs_add_files()
430 DEBUGFS_FWSTATS_ADD(rx, rx_cmplt); wl18xx_debugfs_add_files()
431 DEBUGFS_FWSTATS_ADD(rx, rx_pre_complt); wl18xx_debugfs_add_files()
432 DEBUGFS_FWSTATS_ADD(rx, rx_cmplt_task); wl18xx_debugfs_add_files()
433 DEBUGFS_FWSTATS_ADD(rx, rx_phy_hdr); wl18xx_debugfs_add_files()
434 DEBUGFS_FWSTATS_ADD(rx, rx_timeout); wl18xx_debugfs_add_files()
435 DEBUGFS_FWSTATS_ADD(rx, rx_rts_timeout); wl18xx_debugfs_add_files()
436 DEBUGFS_FWSTATS_ADD(rx, rx_timeout_wa); wl18xx_debugfs_add_files()
437 DEBUGFS_FWSTATS_ADD(rx, defrag_called); wl18xx_debugfs_add_files()
438 DEBUGFS_FWSTATS_ADD(rx, defrag_init_called); wl18xx_debugfs_add_files()
439 DEBUGFS_FWSTATS_ADD(rx, defrag_in_process_called); wl18xx_debugfs_add_files()
440 DEBUGFS_FWSTATS_ADD(rx, defrag_tkip_called); wl18xx_debugfs_add_files()
441 DEBUGFS_FWSTATS_ADD(rx, defrag_need_defrag); wl18xx_debugfs_add_files()
442 DEBUGFS_FWSTATS_ADD(rx, defrag_decrypt_failed); wl18xx_debugfs_add_files()
443 DEBUGFS_FWSTATS_ADD(rx, decrypt_key_not_found); wl18xx_debugfs_add_files()
444 DEBUGFS_FWSTATS_ADD(rx, defrag_need_decrypt); wl18xx_debugfs_add_files()
445 DEBUGFS_FWSTATS_ADD(rx, rx_tkip_replays); wl18xx_debugfs_add_files()
446 DEBUGFS_FWSTATS_ADD(rx, rx_xfr); wl18xx_debugfs_add_files()
H A Devent.h65 /* rx ba constraint */
81 /* rx BA win size indicated by RX_BA_WIN_SIZE_CHANGE_EVENT_ID */
/linux-4.4.14/drivers/scsi/aacraid/
H A DMakefile6 dpcsup.o rx.o sa.o rkt.o nark.o src.o
H A Dnark.c46 iounmap(dev->regs.rx); aac_nark_ioremap()
47 dev->regs.rx = NULL; aac_nark_ioremap()
53 dev->regs.rx = ioremap((u64)pci_resource_start(dev->pdev, 0) | aac_nark_ioremap()
57 if (dev->regs.rx == NULL) aac_nark_ioremap()
61 iounmap(dev->regs.rx); aac_nark_ioremap()
62 dev->regs.rx = NULL; aac_nark_ioremap()
/linux-4.4.14/drivers/net/ethernet/brocade/bna/
H A Dbna_tx_rx.c192 bfi_msgq_mhdr_set(req->mh, BFI_MC_ENET, req_type, 0, rxf->rx->rid); bna_bfi_ucast_req()
198 bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd); bna_bfi_ucast_req()
208 0, rxf->rx->rid); bna_bfi_mcast_add_req()
214 bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd); bna_bfi_mcast_add_req()
224 0, rxf->rx->rid); bna_bfi_mcast_del_req()
230 bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd); bna_bfi_mcast_del_req()
239 BFI_ENET_H2I_MAC_MCAST_FILTER_REQ, 0, rxf->rx->rid); bna_bfi_mcast_filter_req()
245 bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd); bna_bfi_mcast_filter_req()
254 BFI_ENET_H2I_RX_PROMISCUOUS_REQ, 0, rxf->rx->rid); bna_bfi_rx_promisc_req()
260 bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd); bna_bfi_rx_promisc_req()
271 BFI_ENET_H2I_RX_VLAN_SET_REQ, 0, rxf->rx->rid); bna_bfi_rx_vlan_filter_set()
285 bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd); bna_bfi_rx_vlan_filter_set()
294 BFI_ENET_H2I_RX_VLAN_STRIP_ENABLE_REQ, 0, rxf->rx->rid); bna_bfi_vlan_strip_enable()
300 bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd); bna_bfi_vlan_strip_enable()
309 BFI_ENET_H2I_RIT_CFG_REQ, 0, rxf->rx->rid); bna_bfi_rit_cfg()
316 bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd); bna_bfi_rit_cfg()
326 BFI_ENET_H2I_RSS_CFG_REQ, 0, rxf->rx->rid); bna_bfi_rss_cfg()
336 bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd); bna_bfi_rss_cfg()
345 BFI_ENET_H2I_RSS_ENABLE_REQ, 0, rxf->rx->rid); bna_bfi_rss_enable()
351 bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd); bna_bfi_rss_enable()
392 mchandle = bna_mcam_mod_handle_get(&rxf->rx->bna->mcam_mod); bna_rxf_mchandle_attach()
419 bna_mcam_mod_handle_put(&rxf->rx->bna->mcam_mod, mchandle); bna_rxf_mcast_del()
437 list_move_tail(&mac->qe, bna_mcam_mod_del_q(rxf->rx->bna)); bna_rxf_mcast_cfg_apply()
485 list_move_tail(&mac->qe, bna_mcam_mod_del_q(rxf->rx->bna)); bna_rxf_mcast_cfg_reset()
569 struct bna_rx *rx = rxf->rx; bna_rit_init() local
574 list_for_each_entry(rxp, &rx->rxp_q, qe) { bna_rit_init()
617 struct bna_rx *rx, bna_rxf_init()
621 rxf->rx = rx; bna_rxf_init()
669 list_move_tail(&mac->qe, bna_ucam_mod_free_q(rxf->rx->bna)); bna_rxf_uninit()
674 bna_ucam_mod_free_q(rxf->rx->bna)); bna_rxf_uninit()
681 list_move_tail(&mac->qe, bna_mcam_mod_free_q(rxf->rx->bna)); bna_rxf_uninit()
686 if (rxf->rx->bna->promisc_rid == rxf->rx->rid) bna_rxf_uninit()
687 rxf->rx->bna->promisc_rid = BFI_INVALID_RID; bna_rxf_uninit()
688 if (rxf->rx->bna->default_mode_rid == rxf->rx->rid) bna_rxf_uninit()
689 rxf->rx->bna->default_mode_rid = BFI_INVALID_RID; bna_rxf_uninit()
694 rxf->rx = NULL; bna_rxf_uninit()
698 bna_rx_cb_rxf_started(struct bna_rx *rx) bna_rx_cb_rxf_started() argument
700 bfa_fsm_send_event(rx, RX_E_RXF_STARTED); bna_rx_cb_rxf_started()
707 rxf->start_cbarg = rxf->rx; bna_rxf_start()
712 bna_rx_cb_rxf_stopped(struct bna_rx *rx) bna_rx_cb_rxf_stopped() argument
714 bfa_fsm_send_event(rx, RX_E_RXF_STOPPED); bna_rx_cb_rxf_stopped()
721 rxf->stop_cbarg = rxf->rx; bna_rxf_stop()
732 bna_rx_ucast_set(struct bna_rx *rx, const u8 *ucmac) bna_rx_ucast_set() argument
734 struct bna_rxf *rxf = &rx->rxf; bna_rx_ucast_set()
738 bna_cam_mod_mac_get(bna_ucam_mod_free_q(rxf->rx->bna)); bna_rx_ucast_set()
746 rxf->cam_fltr_cbarg = rx->bna->bnad; bna_rx_ucast_set()
754 bna_rx_mcast_add(struct bna_rx *rx, const u8 *addr, bna_rx_mcast_add() argument
757 struct bna_rxf *rxf = &rx->rxf; bna_rx_mcast_add()
764 cbfn(rx->bna->bnad, rx); bna_rx_mcast_add()
768 mac = bna_cam_mod_mac_get(bna_mcam_mod_free_q(rxf->rx->bna)); bna_rx_mcast_add()
775 rxf->cam_fltr_cbarg = rx->bna->bnad; bna_rx_mcast_add()
783 bna_rx_ucast_listset(struct bna_rx *rx, int count, const u8 *uclist) bna_rx_ucast_listset() argument
785 struct bna_ucam_mod *ucam_mod = &rx->bna->ucam_mod; bna_rx_ucast_listset()
786 struct bna_rxf *rxf = &rx->rxf; bna_rx_ucast_listset()
841 bna_rx_mcast_listset(struct bna_rx *rx, int count, const u8 *mclist) bna_rx_mcast_listset() argument
843 struct bna_mcam_mod *mcam_mod = &rx->bna->mcam_mod; bna_rx_mcast_listset()
844 struct bna_rxf *rxf = &rx->rxf; bna_rx_mcast_listset()
901 bna_rx_mcast_delall(struct bna_rx *rx) bna_rx_mcast_delall() argument
903 struct bna_rxf *rxf = &rx->rxf; bna_rx_mcast_delall()
911 list_move_tail(&mac->qe, bna_mcam_mod_free_q(rxf->rx->bna)); bna_rx_mcast_delall()
919 del_mac = bna_cam_mod_mac_get(bna_mcam_mod_del_q(rxf->rx->bna)); bna_rx_mcast_delall()
923 list_add_tail(&mac->qe, bna_mcam_mod_free_q(rxf->rx->bna)); bna_rx_mcast_delall()
932 bna_rx_vlan_add(struct bna_rx *rx, int vlan_id) bna_rx_vlan_add() argument
934 struct bna_rxf *rxf = &rx->rxf; bna_rx_vlan_add()
947 bna_rx_vlan_del(struct bna_rx *rx, int vlan_id) bna_rx_vlan_del() argument
949 struct bna_rxf *rxf = &rx->rxf; bna_rx_vlan_del()
971 list_move_tail(&mac->qe, bna_ucam_mod_del_q(rxf->rx->bna)); bna_rxf_ucast_cfg_apply()
1009 bna_ucam_mod_del_q(rxf->rx->bna)); bna_rxf_ucast_cfg_reset()
1014 bna_ucam_mod_del_q(rxf->rx->bna)); bna_rxf_ucast_cfg_reset()
1047 struct bna *bna = rxf->rx->bna; bna_rxf_promisc_cfg_apply()
1075 struct bna *bna = rxf->rx->bna; bna_rxf_promisc_cfg_reset()
1161 struct bna *bna = rxf->rx->bna; bna_rxf_promisc_enable()
1177 bna->promisc_rid = rxf->rx->rid; bna_rxf_promisc_enable()
1187 struct bna *bna = rxf->rx->bna; bna_rxf_promisc_disable()
1278 #define call_rx_stop_cbfn(rx) \
1280 if ((rx)->stop_cbfn) { \
1283 cbfn = (rx)->stop_cbfn; \
1284 cbarg = (rx)->stop_cbarg; \
1285 (rx)->stop_cbfn = NULL; \
1286 (rx)->stop_cbarg = NULL; \
1287 cbfn(cbarg, rx); \
1291 #define call_rx_stall_cbfn(rx) \
1293 if ((rx)->rx_stall_cbfn) \
1294 (rx)->rx_stall_cbfn((rx)->bna->bnad, (rx)); \
1309 static void bna_bfi_rx_enet_start(struct bna_rx *rx);
1310 static void bna_rx_enet_stop(struct bna_rx *rx);
1311 static void bna_rx_mod_cb_rx_stopped(void *arg, struct bna_rx *rx);
1334 static void bna_rx_sm_stopped_entry(struct bna_rx *rx) bna_rx_sm_stopped_entry() argument
1336 call_rx_stop_cbfn(rx); bna_rx_sm_stopped_entry()
1339 static void bna_rx_sm_stopped(struct bna_rx *rx, bna_rx_sm_stopped() argument
1344 bfa_fsm_set_state(rx, bna_rx_sm_start_wait); bna_rx_sm_stopped()
1348 call_rx_stop_cbfn(rx); bna_rx_sm_stopped()
1361 static void bna_rx_sm_start_wait_entry(struct bna_rx *rx) bna_rx_sm_start_wait_entry() argument
1363 bna_bfi_rx_enet_start(rx); bna_rx_sm_start_wait_entry()
1367 bna_rx_sm_stop_wait_entry(struct bna_rx *rx) bna_rx_sm_stop_wait_entry() argument
1372 bna_rx_sm_stop_wait(struct bna_rx *rx, enum bna_rx_event event) bna_rx_sm_stop_wait() argument
1377 bfa_fsm_set_state(rx, bna_rx_sm_cleanup_wait); bna_rx_sm_stop_wait()
1378 rx->rx_cleanup_cbfn(rx->bna->bnad, rx); bna_rx_sm_stop_wait()
1382 bna_rx_enet_stop(rx); bna_rx_sm_stop_wait()
1391 static void bna_rx_sm_start_wait(struct bna_rx *rx, bna_rx_sm_start_wait() argument
1396 bfa_fsm_set_state(rx, bna_rx_sm_start_stop_wait); bna_rx_sm_start_wait()
1400 bfa_fsm_set_state(rx, bna_rx_sm_stopped); bna_rx_sm_start_wait()
1404 bfa_fsm_set_state(rx, bna_rx_sm_rxf_start_wait); bna_rx_sm_start_wait()
1413 static void bna_rx_sm_rxf_start_wait_entry(struct bna_rx *rx) bna_rx_sm_rxf_start_wait_entry() argument
1415 rx->rx_post_cbfn(rx->bna->bnad, rx); bna_rx_sm_rxf_start_wait_entry()
1416 bna_rxf_start(&rx->rxf); bna_rx_sm_rxf_start_wait_entry()
1420 bna_rx_sm_rxf_stop_wait_entry(struct bna_rx *rx) bna_rx_sm_rxf_stop_wait_entry() argument
1425 bna_rx_sm_rxf_stop_wait(struct bna_rx *rx, enum bna_rx_event event) bna_rx_sm_rxf_stop_wait() argument
1429 bfa_fsm_set_state(rx, bna_rx_sm_cleanup_wait); bna_rx_sm_rxf_stop_wait()
1430 bna_rxf_fail(&rx->rxf); bna_rx_sm_rxf_stop_wait()
1431 call_rx_stall_cbfn(rx); bna_rx_sm_rxf_stop_wait()
1432 rx->rx_cleanup_cbfn(rx->bna->bnad, rx); bna_rx_sm_rxf_stop_wait()
1436 bna_rxf_stop(&rx->rxf); bna_rx_sm_rxf_stop_wait()
1440 bfa_fsm_set_state(rx, bna_rx_sm_stop_wait); bna_rx_sm_rxf_stop_wait()
1441 call_rx_stall_cbfn(rx); bna_rx_sm_rxf_stop_wait()
1442 bna_rx_enet_stop(rx); bna_rx_sm_rxf_stop_wait()
1453 bna_rx_sm_start_stop_wait_entry(struct bna_rx *rx) bna_rx_sm_start_stop_wait_entry() argument
1458 bna_rx_sm_start_stop_wait(struct bna_rx *rx, enum bna_rx_event event) bna_rx_sm_start_stop_wait() argument
1463 bfa_fsm_set_state(rx, bna_rx_sm_stopped); bna_rx_sm_start_stop_wait()
1467 bna_rx_enet_stop(rx); bna_rx_sm_start_stop_wait()
1476 bna_rx_sm_started_entry(struct bna_rx *rx) bna_rx_sm_started_entry() argument
1479 int is_regular = (rx->type == BNA_RX_T_REGULAR); bna_rx_sm_started_entry()
1482 list_for_each_entry(rxp, &rx->rxp_q, qe) bna_rx_sm_started_entry()
1483 bna_ib_start(rx->bna, &rxp->cq.ib, is_regular); bna_rx_sm_started_entry()
1485 bna_ethport_cb_rx_started(&rx->bna->ethport); bna_rx_sm_started_entry()
1489 bna_rx_sm_started(struct bna_rx *rx, enum bna_rx_event event) bna_rx_sm_started() argument
1493 bfa_fsm_set_state(rx, bna_rx_sm_rxf_stop_wait); bna_rx_sm_started()
1494 bna_ethport_cb_rx_stopped(&rx->bna->ethport); bna_rx_sm_started()
1495 bna_rxf_stop(&rx->rxf); bna_rx_sm_started()
1499 bfa_fsm_set_state(rx, bna_rx_sm_failed); bna_rx_sm_started()
1500 bna_ethport_cb_rx_stopped(&rx->bna->ethport); bna_rx_sm_started()
1501 bna_rxf_fail(&rx->rxf); bna_rx_sm_started()
1502 call_rx_stall_cbfn(rx); bna_rx_sm_started()
1503 rx->rx_cleanup_cbfn(rx->bna->bnad, rx); bna_rx_sm_started()
1512 static void bna_rx_sm_rxf_start_wait(struct bna_rx *rx, bna_rx_sm_rxf_start_wait() argument
1517 bfa_fsm_set_state(rx, bna_rx_sm_rxf_stop_wait); bna_rx_sm_rxf_start_wait()
1521 bfa_fsm_set_state(rx, bna_rx_sm_failed); bna_rx_sm_rxf_start_wait()
1522 bna_rxf_fail(&rx->rxf); bna_rx_sm_rxf_start_wait()
1523 call_rx_stall_cbfn(rx); bna_rx_sm_rxf_start_wait()
1524 rx->rx_cleanup_cbfn(rx->bna->bnad, rx); bna_rx_sm_rxf_start_wait()
1528 bfa_fsm_set_state(rx, bna_rx_sm_started); bna_rx_sm_rxf_start_wait()
1538 bna_rx_sm_cleanup_wait_entry(struct bna_rx *rx) bna_rx_sm_cleanup_wait_entry() argument
1543 bna_rx_sm_cleanup_wait(struct bna_rx *rx, enum bna_rx_event event) bna_rx_sm_cleanup_wait() argument
1552 bfa_fsm_set_state(rx, bna_rx_sm_stopped); bna_rx_sm_cleanup_wait()
1562 bna_rx_sm_failed_entry(struct bna_rx *rx) bna_rx_sm_failed_entry() argument
1567 bna_rx_sm_failed(struct bna_rx *rx, enum bna_rx_event event) bna_rx_sm_failed() argument
1571 bfa_fsm_set_state(rx, bna_rx_sm_quiesce_wait); bna_rx_sm_failed()
1575 bfa_fsm_set_state(rx, bna_rx_sm_cleanup_wait); bna_rx_sm_failed()
1585 bfa_fsm_set_state(rx, bna_rx_sm_stopped); bna_rx_sm_failed()
1594 bna_rx_sm_quiesce_wait_entry(struct bna_rx *rx) bna_rx_sm_quiesce_wait_entry() argument
1599 bna_rx_sm_quiesce_wait(struct bna_rx *rx, enum bna_rx_event event) bna_rx_sm_quiesce_wait() argument
1603 bfa_fsm_set_state(rx, bna_rx_sm_cleanup_wait); bna_rx_sm_quiesce_wait()
1607 bfa_fsm_set_state(rx, bna_rx_sm_failed); bna_rx_sm_quiesce_wait()
1611 bfa_fsm_set_state(rx, bna_rx_sm_start_wait); bna_rx_sm_quiesce_wait()
1621 bna_bfi_rx_enet_start(struct bna_rx *rx) bna_bfi_rx_enet_start() argument
1623 struct bfi_enet_rx_cfg_req *cfg_req = &rx->bfi_enet_cmd.cfg_req; bna_bfi_rx_enet_start()
1629 BFI_ENET_H2I_RX_CFG_SET_REQ, 0, rx->rid); bna_bfi_rx_enet_start()
1633 cfg_req->rx_cfg.frame_size = bna_enet_mtu_get(&rx->bna->enet); bna_bfi_rx_enet_start()
1634 cfg_req->num_queue_sets = rx->num_paths; bna_bfi_rx_enet_start()
1635 for (i = 0; i < rx->num_paths; i++) { bna_bfi_rx_enet_start()
1637 : list_first_entry(&rx->rxp_q, struct bna_rxp, qe); bna_bfi_rx_enet_start()
1655 * a new rx with new set of resources. bna_bfi_rx_enet_start()
1663 bna_enet_mtu_get(&rx->bna->enet); bna_bfi_rx_enet_start()
1703 cfg_req->rx_cfg.hds.type = rx->hds_cfg.hdr_type; bna_bfi_rx_enet_start()
1704 cfg_req->rx_cfg.hds.force_offset = rx->hds_cfg.forced_offset; bna_bfi_rx_enet_start()
1705 cfg_req->rx_cfg.hds.max_header_size = rx->hds_cfg.forced_offset; bna_bfi_rx_enet_start()
1715 cfg_req->rx_cfg.strip_vlan = rx->rxf.vlan_strip_status; bna_bfi_rx_enet_start()
1717 bfa_msgq_cmd_set(&rx->msgq_cmd, NULL, NULL, bna_bfi_rx_enet_start()
1719 bfa_msgq_cmd_post(&rx->bna->msgq, &rx->msgq_cmd); bna_bfi_rx_enet_start()
1723 bna_bfi_rx_enet_stop(struct bna_rx *rx) bna_bfi_rx_enet_stop() argument
1725 struct bfi_enet_req *req = &rx->bfi_enet_cmd.req; bna_bfi_rx_enet_stop()
1728 BFI_ENET_H2I_RX_CFG_CLR_REQ, 0, rx->rid); bna_bfi_rx_enet_stop()
1731 bfa_msgq_cmd_set(&rx->msgq_cmd, NULL, NULL, sizeof(struct bfi_enet_req), bna_bfi_rx_enet_stop()
1733 bfa_msgq_cmd_post(&rx->bna->msgq, &rx->msgq_cmd); bna_bfi_rx_enet_stop()
1737 bna_rx_enet_stop(struct bna_rx *rx) bna_rx_enet_stop() argument
1742 list_for_each_entry(rxp, &rx->rxp_q, qe) bna_rx_enet_stop()
1743 bna_ib_stop(rx->bna, &rxp->cq.ib); bna_rx_enet_stop()
1745 bna_bfi_rx_enet_stop(rx); bna_rx_enet_stop()
1810 struct bna_rx *rx = NULL; bna_rx_get() local
1814 rx = list_first_entry(&rx_mod->rx_free_q, struct bna_rx, qe); bna_rx_get()
1816 rx = list_last_entry(&rx_mod->rx_free_q, struct bna_rx, qe); bna_rx_get()
1819 list_move_tail(&rx->qe, &rx_mod->rx_active_q); bna_rx_get()
1820 rx->type = type; bna_rx_get()
1822 return rx; bna_rx_get()
1826 bna_rx_put(struct bna_rx_mod *rx_mod, struct bna_rx *rx) bna_rx_put() argument
1831 if (((struct bna_rx *)qe)->rid < rx->rid) bna_rx_put()
1834 list_add(&rx->qe, qe); bna_rx_put()
1938 bna_rx_mod_cb_rx_stopped(void *arg, struct bna_rx *rx) bna_rx_mod_cb_rx_stopped() argument
1956 bna_rx_start(struct bna_rx *rx) bna_rx_start() argument
1958 rx->rx_flags |= BNA_RX_F_ENET_STARTED; bna_rx_start()
1959 if (rx->rx_flags & BNA_RX_F_ENABLED) bna_rx_start()
1960 bfa_fsm_send_event(rx, RX_E_START); bna_rx_start()
1964 bna_rx_stop(struct bna_rx *rx) bna_rx_stop() argument
1966 rx->rx_flags &= ~BNA_RX_F_ENET_STARTED; bna_rx_stop()
1967 if (rx->fsm == (bfa_fsm_t) bna_rx_sm_stopped) bna_rx_stop()
1968 bna_rx_mod_cb_rx_stopped(&rx->bna->rx_mod, rx); bna_rx_stop()
1970 rx->stop_cbfn = bna_rx_mod_cb_rx_stopped; bna_rx_stop()
1971 rx->stop_cbarg = &rx->bna->rx_mod; bna_rx_stop()
1972 bfa_fsm_send_event(rx, RX_E_STOP); bna_rx_stop()
1977 bna_rx_fail(struct bna_rx *rx) bna_rx_fail() argument
1980 rx->rx_flags &= ~BNA_RX_F_ENET_STARTED; bna_rx_fail()
1981 bfa_fsm_send_event(rx, RX_E_FAIL); bna_rx_fail()
1987 struct bna_rx *rx; bna_rx_mod_start() local
1993 list_for_each_entry(rx, &rx_mod->rx_active_q, qe) bna_rx_mod_start()
1994 if (rx->type == type) bna_rx_mod_start()
1995 bna_rx_start(rx); bna_rx_mod_start()
2001 struct bna_rx *rx; bna_rx_mod_stop() local
2010 list_for_each_entry(rx, &rx_mod->rx_active_q, qe) bna_rx_mod_stop()
2011 if (rx->type == type) { bna_rx_mod_stop()
2013 bna_rx_stop(rx); bna_rx_mod_stop()
2022 struct bna_rx *rx; bna_rx_mod_fail() local
2027 list_for_each_entry(rx, &rx_mod->rx_active_q, qe) bna_rx_mod_fail()
2028 bna_rx_fail(rx); bna_rx_mod_fail()
2042 rx_mod->rx = (struct bna_rx *) bna_rx_mod_init()
2060 rx_ptr = &rx_mod->rx[index]; bna_rx_mod_init()
2094 bna_bfi_rx_enet_start_rsp(struct bna_rx *rx, struct bfi_msgq_mhdr *msghdr) bna_bfi_rx_enet_start_rsp() argument
2096 struct bfi_enet_rx_cfg_rsp *cfg_rsp = &rx->bfi_enet_cmd.cfg_rsp; bna_bfi_rx_enet_start_rsp()
2101 bfa_msgq_rsp_copy(&rx->bna->msgq, (u8 *)cfg_rsp, bna_bfi_rx_enet_start_rsp()
2104 rx->hw_id = cfg_rsp->hw_id; bna_bfi_rx_enet_start_rsp()
2106 for (i = 0, rxp = list_first_entry(&rx->rxp_q, struct bna_rxp, qe); bna_bfi_rx_enet_start_rsp()
2107 i < rx->num_paths; i++, rxp = list_next_entry(rxp, qe)) { bna_bfi_rx_enet_start_rsp()
2112 rx->bna->pcidev.pci_bar_kva bna_bfi_rx_enet_start_rsp()
2116 rx->bna->pcidev.pci_bar_kva bna_bfi_rx_enet_start_rsp()
2121 rx->bna->pcidev.pci_bar_kva bna_bfi_rx_enet_start_rsp()
2134 bfa_fsm_send_event(rx, RX_E_STARTED); bna_bfi_rx_enet_start_rsp()
2138 bna_bfi_rx_enet_stop_rsp(struct bna_rx *rx, struct bfi_msgq_mhdr *msghdr) bna_bfi_rx_enet_stop_rsp() argument
2140 bfa_fsm_send_event(rx, RX_E_STOPPED); bna_bfi_rx_enet_stop_rsp()
2265 struct bna_rx *rx; bna_rx_create() local
2315 rx = bna_rx_get(rx_mod, rx_cfg->rx_type); bna_rx_create()
2316 rx->bna = bna; bna_rx_create()
2317 rx->rx_flags = 0; bna_rx_create()
2318 INIT_LIST_HEAD(&rx->rxp_q); bna_rx_create()
2319 rx->stop_cbfn = NULL; bna_rx_create()
2320 rx->stop_cbarg = NULL; bna_rx_create()
2321 rx->priv = priv; bna_rx_create()
2323 rx->rcb_setup_cbfn = rx_cbfn->rcb_setup_cbfn; bna_rx_create()
2324 rx->rcb_destroy_cbfn = rx_cbfn->rcb_destroy_cbfn; bna_rx_create()
2325 rx->ccb_setup_cbfn = rx_cbfn->ccb_setup_cbfn; bna_rx_create()
2326 rx->ccb_destroy_cbfn = rx_cbfn->ccb_destroy_cbfn; bna_rx_create()
2327 rx->rx_stall_cbfn = rx_cbfn->rx_stall_cbfn; bna_rx_create()
2329 rx->rx_cleanup_cbfn = rx_cbfn->rx_cleanup_cbfn; bna_rx_create()
2330 rx->rx_post_cbfn = rx_cbfn->rx_post_cbfn; bna_rx_create()
2332 if (rx->bna->rx_mod.flags & BNA_RX_MOD_F_ENET_STARTED) { bna_rx_create()
2333 switch (rx->type) { bna_rx_create()
2335 if (!(rx->bna->rx_mod.flags & bna_rx_create()
2337 rx->rx_flags |= BNA_RX_F_ENET_STARTED; bna_rx_create()
2340 if (rx->bna->rx_mod.flags & BNA_RX_MOD_F_ENET_LOOPBACK) bna_rx_create()
2341 rx->rx_flags |= BNA_RX_F_ENET_STARTED; bna_rx_create()
2346 rx->num_paths = rx_cfg->num_paths; bna_rx_create()
2348 i < rx->num_paths; i++) { bna_rx_create()
2350 list_add_tail(&rxp->qe, &rx->rxp_q); bna_rx_create()
2352 rxp->rx = rx; bna_rx_create()
2353 rxp->cq.rx = rx; bna_rx_create()
2387 q0->rx = rx; bna_rx_create()
2408 if (rx->rcb_setup_cbfn) bna_rx_create()
2409 rx->rcb_setup_cbfn(bnad, q0->rcb); bna_rx_create()
2414 q1->rx = rx; bna_rx_create()
2438 if (rx->rcb_setup_cbfn) bna_rx_create()
2439 rx->rcb_setup_cbfn(bnad, q1->rcb); bna_rx_create()
2475 if (rx->ccb_setup_cbfn) bna_rx_create()
2476 rx->ccb_setup_cbfn(bnad, rxp->cq.ccb); bna_rx_create()
2479 rx->hds_cfg = rx_cfg->hds_config; bna_rx_create()
2481 bna_rxf_init(&rx->rxf, rx, rx_cfg, res_info); bna_rx_create()
2483 bfa_fsm_set_state(rx, bna_rx_sm_stopped); bna_rx_create()
2485 rx_mod->rid_mask |= BIT(rx->rid); bna_rx_create()
2487 return rx; bna_rx_create()
2491 bna_rx_destroy(struct bna_rx *rx) bna_rx_destroy() argument
2493 struct bna_rx_mod *rx_mod = &rx->bna->rx_mod; bna_rx_destroy()
2499 bna_rxf_uninit(&rx->rxf); bna_rx_destroy()
2501 while (!list_empty(&rx->rxp_q)) { bna_rx_destroy()
2502 rxp = list_first_entry(&rx->rxp_q, struct bna_rxp, qe); bna_rx_destroy()
2505 if (rx->rcb_destroy_cbfn) bna_rx_destroy()
2506 rx->rcb_destroy_cbfn(rx->bna->bnad, q0->rcb); bna_rx_destroy()
2509 q0->rx = NULL; bna_rx_destroy()
2513 if (rx->rcb_destroy_cbfn) bna_rx_destroy()
2514 rx->rcb_destroy_cbfn(rx->bna->bnad, q1->rcb); bna_rx_destroy()
2517 q1->rx = NULL; bna_rx_destroy()
2523 if (rx->ccb_destroy_cbfn) bna_rx_destroy()
2524 rx->ccb_destroy_cbfn(rx->bna->bnad, rxp->cq.ccb); bna_rx_destroy()
2526 rxp->rx = NULL; bna_rx_destroy()
2531 if (qe == &rx->qe) { bna_rx_destroy()
2532 list_del(&rx->qe); bna_rx_destroy()
2536 rx_mod->rid_mask &= ~BIT(rx->rid); bna_rx_destroy()
2538 rx->bna = NULL; bna_rx_destroy()
2539 rx->priv = NULL; bna_rx_destroy()
2540 bna_rx_put(rx_mod, rx); bna_rx_destroy()
2544 bna_rx_enable(struct bna_rx *rx) bna_rx_enable() argument
2546 if (rx->fsm != (bfa_sm_t)bna_rx_sm_stopped) bna_rx_enable()
2549 rx->rx_flags |= BNA_RX_F_ENABLED; bna_rx_enable()
2550 if (rx->rx_flags & BNA_RX_F_ENET_STARTED) bna_rx_enable()
2551 bfa_fsm_send_event(rx, RX_E_START); bna_rx_enable()
2555 bna_rx_disable(struct bna_rx *rx, enum bna_cleanup_type type, bna_rx_disable() argument
2560 (*cbfn)(rx->bna->bnad, rx); bna_rx_disable()
2562 rx->stop_cbfn = cbfn; bna_rx_disable()
2563 rx->stop_cbarg = rx->bna->bnad; bna_rx_disable()
2565 rx->rx_flags &= ~BNA_RX_F_ENABLED; bna_rx_disable()
2567 bfa_fsm_send_event(rx, RX_E_STOP); bna_rx_disable()
2572 bna_rx_cleanup_complete(struct bna_rx *rx) bna_rx_cleanup_complete() argument
2574 bfa_fsm_send_event(rx, RX_E_CLEANUP_DONE); bna_rx_cleanup_complete()
2578 bna_rx_vlan_strip_enable(struct bna_rx *rx) bna_rx_vlan_strip_enable() argument
2580 struct bna_rxf *rxf = &rx->rxf; bna_rx_vlan_strip_enable()
2590 bna_rx_vlan_strip_disable(struct bna_rx *rx) bna_rx_vlan_strip_disable() argument
2592 struct bna_rxf *rxf = &rx->rxf; bna_rx_vlan_strip_disable()
2602 bna_rx_mode_set(struct bna_rx *rx, enum bna_rxmode new_mode, bna_rx_mode_set() argument
2605 struct bna_rxf *rxf = &rx->rxf; bna_rx_mode_set()
2612 if ((rx->bna->promisc_rid != BFI_INVALID_RID) && bna_rx_mode_set()
2613 (rx->bna->promisc_rid != rxf->rx->rid)) bna_rx_mode_set()
2617 if (rx->bna->default_mode_rid != BFI_INVALID_RID) bna_rx_mode_set()
2627 if ((rx->bna->default_mode_rid != BFI_INVALID_RID) && bna_rx_mode_set()
2628 (rx->bna->default_mode_rid != rxf->rx->rid)) { bna_rx_mode_set()
2633 if (rx->bna->promisc_rid != BFI_INVALID_RID) bna_rx_mode_set()
2659 rxf->cam_fltr_cbarg = rx->bna->bnad; bna_rx_mode_set()
2670 bna_rx_vlanfilter_enable(struct bna_rx *rx) bna_rx_vlanfilter_enable() argument
2672 struct bna_rxf *rxf = &rx->rxf; bna_rx_vlanfilter_enable()
2682 bna_rx_coalescing_timeo_set(struct bna_rx *rx, int coalescing_timeo) bna_rx_coalescing_timeo_set() argument
2686 list_for_each_entry(rxp, &rx->rxp_q, qe) { bna_rx_coalescing_timeo_set()
2705 struct bna *bna = ccb->cq->rx->bna; bna_rx_dim_update()
616 bna_rxf_init(struct bna_rxf *rxf, struct bna_rx *rx, struct bna_rx_config *q_config, struct bna_res_info *res_info) bna_rxf_init() argument
H A Dbna.h118 cbfn(cbarg, rxf->rx); \
330 void bna_bfi_rx_enet_start_rsp(struct bna_rx *rx,
332 void bna_bfi_rx_enet_stop_rsp(struct bna_rx *rx,
357 void bna_rx_destroy(struct bna_rx *rx);
358 void bna_rx_enable(struct bna_rx *rx);
359 void bna_rx_disable(struct bna_rx *rx, enum bna_cleanup_type type,
361 void bna_rx_cleanup_complete(struct bna_rx *rx);
362 void bna_rx_coalescing_timeo_set(struct bna_rx *rx, int coalescing_timeo);
365 enum bna_cb_status bna_rx_ucast_set(struct bna_rx *rx, const u8 *ucmac);
366 enum bna_cb_status bna_rx_ucast_listset(struct bna_rx *rx, int count,
368 enum bna_cb_status bna_rx_mcast_add(struct bna_rx *rx, const u8 *mcmac,
371 enum bna_cb_status bna_rx_mcast_listset(struct bna_rx *rx, int count,
374 bna_rx_mcast_delall(struct bna_rx *rx);
376 bna_rx_mode_set(struct bna_rx *rx, enum bna_rxmode rxmode,
378 void bna_rx_vlan_add(struct bna_rx *rx, int vlan_id);
379 void bna_rx_vlan_del(struct bna_rx *rx, int vlan_id);
380 void bna_rx_vlanfilter_enable(struct bna_rx *rx);
381 void bna_rx_vlan_strip_enable(struct bna_rx *rx);
382 void bna_rx_vlan_strip_disable(struct bna_rx *rx);
H A Dbna_types.h582 struct bna_rx *rx; member in struct:bna_rxq
648 struct bna_rx *rx; member in struct:bna_cq
701 struct bna_rx *rx; member in struct:bna_rxp
724 void (*start_cbfn) (struct bna_rx *rx);
728 void (*stop_cbfn) (struct bna_rx *rx);
738 void (*cam_fltr_cbfn)(struct bnad *bnad, struct bna_rx *rx);
776 struct bna_rx *rx; member in struct:bna_rxf
816 void (*stop_cbfn)(void *arg, struct bna_rx *rx);
835 /* Rx module - keeps track of free, active rx objects */
838 struct bna_rx *rx; /* BFI_MAX_RXQ entries */ member in struct:bna_rx_mod
H A Dbnad.c850 if (!rx_info->rx) bnad_isr()
1037 (struct bnad_rx_info *)ccb->cq->rx->priv; bnad_cb_ccb_setup()
1047 (struct bnad_rx_info *)ccb->cq->rx->priv; bnad_cb_ccb_destroy()
1165 bnad_cb_rx_stall(struct bnad *bnad, struct bna_rx *rx) bnad_cb_rx_stall() argument
1167 struct bnad_rx_info *rx_info = (struct bnad_rx_info *)rx->priv; bnad_cb_rx_stall()
1219 bna_rx_cleanup_complete(rx_info->rx); bnad_rx_cleanup()
1224 bnad_cb_rx_cleanup(struct bnad *bnad, struct bna_rx *rx) bnad_cb_rx_cleanup() argument
1226 struct bnad_rx_info *rx_info = (struct bnad_rx_info *)rx->priv; bnad_cb_rx_cleanup()
1247 bnad_cb_rx_post(struct bnad *bnad, struct bna_rx *rx) bnad_cb_rx_post() argument
1249 struct bnad_rx_info *rx_info = (struct bnad_rx_info *)rx->priv; bnad_cb_rx_post()
1277 bnad_cb_rx_disabled(void *arg, struct bna_rx *rx) bnad_cb_rx_disabled() argument
1285 bnad_cb_rx_mcast_add(struct bnad *bnad, struct bna_rx *rx) bnad_cb_rx_mcast_add() argument
1772 if (!rx_info->rx) bnad_dim_timeout()
2041 /* Setup the rx config for bna_rx_create */
2125 /* destroy and create new rx objects */ bnad_reinit_rx()
2127 if (!bnad->rx_info[rx_id].rx) bnad_reinit_rx()
2146 /* restore rx configuration */ bnad_reinit_rx()
2147 if (bnad->rx_info[0].rx && !err) { bnad_reinit_rx()
2169 if (!rx_info->rx) bnad_destroy_rx()
2186 bna_rx_disable(rx_info->rx, BNA_HARD_CLEANUP, bnad_cb_rx_disabled); bnad_destroy_rx()
2196 bna_rx_destroy(rx_info->rx); bnad_destroy_rx()
2198 rx_info->rx = NULL; bnad_destroy_rx()
2224 struct bna_rx *rx; bnad_setup_rx() local
2260 rx = bna_rx_create(&bnad->bna, bnad, rx_config, &rx_cbfn, res_info, bnad_setup_rx()
2262 if (!rx) { bnad_setup_rx()
2267 rx_info->rx = rx; bnad_setup_rx()
2294 bna_rx_vlanfilter_enable(rx); bnad_setup_rx()
2300 bna_rx_enable(rx); bnad_setup_rx()
2332 if (!rx_info->rx) bnad_rx_coalescing_timeo_set()
2334 bna_rx_coalescing_timeo_set(rx_info->rx, bnad_rx_coalescing_timeo_set()
2351 if (!bnad->rx_info[0].rx) bnad_mac_addr_set_locked()
2354 ret = bna_rx_ucast_set(bnad->rx_info[0].rx, mac_addr); bnad_mac_addr_set_locked()
2372 ret = bna_rx_mcast_add(rx_info->rx, bnad_bcast_addr, bnad_enable_default_bcast()
2396 bna_rx_vlan_add(bnad->rx_info[rx_id].rx, vid); bnad_restore_vlans()
3149 bna_rx_ucast_listset(bnad->rx_info[0].rx, 0, NULL); bnad_set_rx_ucast_fltr()
3166 ret = bna_rx_ucast_listset(bnad->rx_info[0].rx, entry, mac_list);
3177 bna_rx_ucast_listset(bnad->rx_info[0].rx, 0, NULL);
3206 ret = bna_rx_mcast_listset(bnad->rx_info[0].rx, mc_count + 1, mac_list); bnad_set_rx_mcast_fltr()
3216 bna_rx_mcast_delall(bnad->rx_info[0].rx); bnad_set_rx_mcast_fltr()
3228 if (bnad->rx_info[0].rx == NULL) { bnad_set_rx_mode()
3255 bna_rx_mode_set(bnad->rx_info[0].rx, new_mode, mode_mask); bnad_set_rx_mode()
3327 /* rx_count > 0 - new rx created bnad_change_mtu()
3344 if (!bnad->rx_info[0].rx) bnad_vlan_rx_add_vid()
3350 bna_rx_vlan_add(bnad->rx_info[0].rx, vid); bnad_vlan_rx_add_vid()
3365 if (!bnad->rx_info[0].rx) bnad_vlan_rx_kill_vid()
3372 bna_rx_vlan_del(bnad->rx_info[0].rx, vid); bnad_vlan_rx_kill_vid()
3391 bna_rx_vlan_strip_enable(bnad->rx_info[0].rx); bnad_set_features()
3393 bna_rx_vlan_strip_disable(bnad->rx_info[0].rx); bnad_set_features()
3424 if (!rx_info->rx) bnad_netpoll()
/linux-4.4.14/drivers/staging/gdm72xx/
H A Dgdm_usb.c82 static struct usb_rx *alloc_rx_struct(struct rx_cxt *rx) alloc_rx_struct() argument
98 r->rx_cxt = rx; alloc_rx_struct()
136 static struct usb_rx *get_rx_struct(struct rx_cxt *rx) get_rx_struct() argument
140 if (list_empty(&rx->free_list)) { get_rx_struct()
141 r = alloc_rx_struct(rx); get_rx_struct()
145 list_add(&r->list, &rx->free_list); get_rx_struct()
148 r = list_entry(rx->free_list.next, struct usb_rx, list); get_rx_struct()
149 list_move_tail(&r->list, &rx->used_list); get_rx_struct()
155 static void put_rx_struct(struct rx_cxt *rx, struct usb_rx *r) put_rx_struct() argument
157 list_move(&r->list, &rx->free_list); put_rx_struct()
163 struct rx_cxt *rx = &udev->rx; release_usb() local
187 spin_lock_irqsave(&rx->lock, flags); release_usb()
189 list_for_each_entry_safe(r, r_next, &rx->free_list, list) { release_usb()
194 list_for_each_entry_safe(r, r_next, &rx->used_list, list) { release_usb()
199 spin_unlock_irqrestore(&rx->lock, flags); release_usb()
206 struct rx_cxt *rx = &udev->rx; init_usb() local
218 INIT_LIST_HEAD(&rx->free_list); init_usb()
219 INIT_LIST_HEAD(&rx->used_list); init_usb()
222 spin_lock_init(&rx->lock); init_usb()
236 r = alloc_rx_struct(rx); init_usb()
242 spin_lock_irqsave(&rx->lock, flags); init_usb()
243 list_add(&r->list, &rx->free_list); init_usb()
244 spin_unlock_irqrestore(&rx->lock, flags); init_usb()
359 struct rx_cxt *rx; gdm_usb_send() local
362 rx = &udev->rx; gdm_usb_send()
364 spin_lock_irqsave(&rx->lock, flags2); gdm_usb_send()
365 list_for_each_entry(r, &rx->used_list, list) gdm_usb_send()
367 spin_unlock_irqrestore(&rx->lock, flags2); gdm_usb_send()
407 struct rx_cxt *rx = r->rx_cxt; gdm_usb_rcv_complete() local
408 struct usbwm_dev *udev = container_of(r->rx_cxt, struct usbwm_dev, rx); gdm_usb_rcv_complete()
451 spin_lock_irqsave(&rx->lock, flags2); gdm_usb_rcv_complete()
452 put_rx_struct(rx, r); gdm_usb_rcv_complete()
453 spin_unlock_irqrestore(&rx->lock, flags2); gdm_usb_rcv_complete()
468 struct rx_cxt *rx = &udev->rx; gdm_usb_receive() local
477 spin_lock_irqsave(&rx->lock, flags); gdm_usb_receive()
478 r = get_rx_struct(rx); gdm_usb_receive()
479 spin_unlock_irqrestore(&rx->lock, flags); gdm_usb_receive()
646 struct rx_cxt *rx; gdm_suspend() local
655 rx = &udev->rx; gdm_suspend()
657 spin_lock_irqsave(&rx->lock, flags); gdm_suspend()
659 list_for_each_entry(r, &rx->used_list, list) gdm_suspend()
662 spin_unlock_irqrestore(&rx->lock, flags); gdm_suspend()
671 struct rx_cxt *rx; gdm_resume() local
680 rx = &udev->rx; gdm_resume()
682 spin_lock_irqsave(&rx->lock, flags); gdm_resume()
684 list_for_each_entry(r, &rx->used_list, list) gdm_resume()
687 spin_unlock_irqrestore(&rx->lock, flags); gdm_resume()
699 struct rx_cxt *rx; k_mode_thread() local
710 rx = &udev->rx; k_mode_thread()
719 spin_lock_irqsave(&rx->lock, flags); k_mode_thread()
721 list_for_each_entry(r, &rx->used_list, list) k_mode_thread()
724 spin_unlock_irqrestore(&rx->lock, flags); k_mode_thread()
H A Dgdm_sdio.c67 static struct sdio_rx *alloc_rx_struct(struct rx_cxt *rx) alloc_rx_struct() argument
72 r->rx_cxt = rx; alloc_rx_struct()
105 static struct sdio_rx *get_rx_struct(struct rx_cxt *rx) get_rx_struct() argument
109 if (list_empty(&rx->free_list)) get_rx_struct()
112 r = list_entry(rx->free_list.prev, struct sdio_rx, list); get_rx_struct()
119 static void put_rx_struct(struct rx_cxt *rx, struct sdio_rx *r) put_rx_struct() argument
121 list_add_tail(&r->list, &rx->free_list); put_rx_struct()
127 struct rx_cxt *rx = &sdev->rx; release_sdio() local
148 kfree(rx->rx_buf); release_sdio()
150 list_for_each_entry_safe(r, r_next, &rx->free_list, list) { release_sdio()
155 list_for_each_entry_safe(r, r_next, &rx->req_list, list) { release_sdio()
165 struct rx_cxt *rx = &sdev->rx; init_sdio() local
188 INIT_LIST_HEAD(&rx->free_list); init_sdio()
189 INIT_LIST_HEAD(&rx->req_list); init_sdio()
191 spin_lock_init(&rx->lock); init_sdio()
194 r = alloc_rx_struct(rx); init_sdio()
199 list_add(&r->list, &rx->free_list); init_sdio()
202 rx->rx_buf = kmalloc(RX_BUF_SIZE, GFP_KERNEL); init_sdio()
203 if (!rx->rx_buf) init_sdio()
477 struct rx_cxt *rx = &sdev->rx; gdm_sdio_irq() local
517 memcpy(rx->rx_buf, hdr + TYPE_A_HEADER_SIZE, gdm_sdio_irq()
520 buf = rx->rx_buf + TYPE_A_LOOKAHEAD_SIZE - TYPE_A_HEADER_SIZE; gdm_sdio_irq()
549 dev_dbg(&func->dev, "sdio_receive: %*ph\n", len, rx->rx_buf); gdm_sdio_irq()
551 len = control_sdu_tx_flow(sdev, rx->rx_buf, len); gdm_sdio_irq()
553 spin_lock_irqsave(&rx->lock, flags); gdm_sdio_irq()
555 if (!list_empty(&rx->req_list)) { gdm_sdio_irq()
556 r = list_entry(rx->req_list.next, struct sdio_rx, list); gdm_sdio_irq()
557 spin_unlock_irqrestore(&rx->lock, flags); gdm_sdio_irq()
559 r->callback(r->cb_data, rx->rx_buf, len); gdm_sdio_irq()
560 spin_lock_irqsave(&rx->lock, flags); gdm_sdio_irq()
562 put_rx_struct(rx, r); gdm_sdio_irq()
565 spin_unlock_irqrestore(&rx->lock, flags); gdm_sdio_irq()
578 struct rx_cxt *rx = &sdev->rx; gdm_sdio_receive() local
582 spin_lock_irqsave(&rx->lock, flags); gdm_sdio_receive()
583 r = get_rx_struct(rx); gdm_sdio_receive()
585 spin_unlock_irqrestore(&rx->lock, flags); gdm_sdio_receive()
592 list_add_tail(&r->list, &rx->req_list); gdm_sdio_receive()
593 spin_unlock_irqrestore(&rx->lock, flags); gdm_sdio_receive()
/linux-4.4.14/drivers/media/radio/wl128x/
H A Dfmdrv_rx.c30 fmdev->rx.rds.flag = FM_RDS_DISABLE; fm_rx_reset_rds_cache()
31 fmdev->rx.rds.last_blk_idx = 0; fm_rx_reset_rds_cache()
32 fmdev->rx.rds.wr_idx = 0; fm_rx_reset_rds_cache()
33 fmdev->rx.rds.rd_idx = 0; fm_rx_reset_rds_cache()
35 if (fmdev->rx.af_mode == FM_RX_RDS_AF_SWITCH_MODE_ON) fm_rx_reset_rds_cache()
41 fmdev->rx.stat_info.picode = FM_NO_PI_CODE; fm_rx_reset_station_info()
42 fmdev->rx.stat_info.afcache_size = 0; fm_rx_reset_station_info()
43 fmdev->rx.stat_info.af_list_max = 0; fm_rx_reset_station_info()
54 if (freq < fmdev->rx.region.bot_freq || freq > fmdev->rx.region.top_freq) { fm_rx_set_freq()
75 payload = (freq - fmdev->rx.region.bot_freq) / FM_FREQ_MUL; fm_rx_set_freq()
120 curr_frq_in_khz = (fmdev->rx.region.bot_freq + ((u32)curr_frq * FM_FREQ_MUL)); fm_rx_set_freq()
128 fmdev->rx.freq = curr_frq_in_khz; fm_rx_set_freq()
164 fmdev->rx.region.chanl_space = spacing * FM_FREQ_MUL; fm_rx_set_channel_spacing()
193 last_frq = (fmdev->rx.region.top_freq - fmdev->rx.region.bot_freq) / FM_FREQ_MUL; fm_rx_seek()
196 space_idx = fmdev->rx.region.chanl_space / FM_FREQ_MUL; fm_rx_seek()
269 fmdev->rx.freq = seek_upward ? fm_rx_seek()
270 fmdev->rx.region.top_freq : fm_rx_seek()
271 fmdev->rx.region.bot_freq; fm_rx_seek()
273 fmdev->rx.freq = seek_upward ? fm_rx_seek()
274 fmdev->rx.region.bot_freq : fm_rx_seek()
275 fmdev->rx.region.top_freq; fm_rx_seek()
277 next_frq = (fmdev->rx.freq - fm_rx_seek()
278 fmdev->rx.region.bot_freq) / FM_FREQ_MUL; fm_rx_seek()
289 fmdev->rx.freq = (fmdev->rx.region.bot_freq + fm_rx_seek()
321 fmdev->rx.volume = vol_to_set; fm_rx_set_volume()
336 *curr_vol = fmdev->rx.volume / FM_RX_VOLUME_GAIN_STEP; fm_rx_get_volume()
345 *bot_freq = fmdev->rx.region.bot_freq; fm_rx_get_band_freq_range()
348 *top_freq = fmdev->rx.region.top_freq; fm_rx_get_band_freq_range()
356 *region = fmdev->rx.region.fm_band; fm_rx_get_region()
372 if (fmdev->rx.region.fm_band == region_to_set) { fm_rx_set_region()
387 if (fmdev->rx.freq < fmdev->rx.region.bot_freq) fm_rx_set_region()
388 new_frq = fmdev->rx.region.bot_freq; fm_rx_set_region()
389 else if (fmdev->rx.freq > fmdev->rx.region.top_freq) fm_rx_set_region()
390 new_frq = fmdev->rx.region.top_freq; fm_rx_set_region()
413 *curr_mute_mode = fmdev->rx.mute_mode; fm_rx_get_mute_mode()
424 switch (fmdev->rx.mute_mode) { fm_config_rx_mute_reg()
437 if (fmdev->rx.rf_depend_mute == FM_RX_RF_DEPENDENT_MUTE_ON) fm_config_rx_mute_reg()
457 if (fmdev->rx.mute_mode == mute_mode_toset) fm_rx_set_mute_mode()
460 org_state = fmdev->rx.mute_mode; fm_rx_set_mute_mode()
461 fmdev->rx.mute_mode = mute_mode_toset; fm_rx_set_mute_mode()
465 fmdev->rx.mute_mode = org_state; fm_rx_set_mute_mode()
483 *curr_mute_mode = fmdev->rx.rf_depend_mute; fm_rx_get_rfdepend_softmute()
502 if (fmdev->rx.rf_depend_mute == rfdepend_mute) fm_rx_set_rfdepend_softmute()
505 org_state = fmdev->rx.rf_depend_mute; fm_rx_set_rfdepend_softmute()
506 fmdev->rx.rf_depend_mute = rfdepend_mute; fm_rx_set_rfdepend_softmute()
510 fmdev->rx.rf_depend_mute = org_state; fm_rx_set_rfdepend_softmute()
559 fmdev->rx.rssi_threshold = rssi_lvl_toset; fm_rx_set_rssi_threshold()
575 *curr_rssi_lvl = fmdev->rx.rssi_threshold; fm_rx_get_rssi_threshold()
641 fmerr("Invalid rx de-emphasis mode (%d)\n", mode); fm_rx_set_deemphasis_mode()
651 fmdev->rx.deemphasis_mode = mode; fm_rx_set_deemphasis_mode()
667 *curr_deemphasis_mode = fmdev->rx.deemphasis_mode; fm_rx_get_deemph_mode()
684 && fmdev->rx.rds.flag == FM_RDS_DISABLE) { fm_rx_set_rds_mode()
723 fmdev->rx.rds.flag = FM_RDS_ENABLE; fm_rx_set_rds_mode()
725 && fmdev->rx.rds.flag == FM_RDS_ENABLE) { fm_rx_set_rds_mode()
734 fmdev->rx.rds.last_blk_idx = 0; fm_rx_set_rds_mode()
735 fmdev->rx.rds.wr_idx = 0; fm_rx_set_rds_mode()
736 fmdev->rx.rds.rd_idx = 0; fm_rx_set_rds_mode()
741 fmdev->rx.rds.flag = FM_RDS_DISABLE; fm_rx_set_rds_mode()
758 *curr_rds_en_dis = fmdev->rx.rds.flag; fm_rx_get_rds_mode()
783 fmdev->rx.rds_mode = rds_mode; fm_rx_set_rds_system()
814 fmdev->rx.af_mode = af_mode; fm_rx_set_af_switch()
830 *af_mode = fmdev->rx.af_mode; fm_rx_get_af_switch()
H A Dfmdrv_common.c253 fmdev->rx.region = region_configs[region_to_set]; fmc_update_region_info()
638 struct tuned_station_info *stat_info = &fmdev->rx.stat_info; fm_rx_update_af_cache()
639 u8 reg_idx = fmdev->rx.region.fm_band; fm_rx_update_af_cache()
645 fmdev->rx.stat_info.af_list_max = (af - FM_RDS_1_AF_FOLLOWS + 1); fm_rx_update_af_cache()
646 fmdev->rx.stat_info.afcache_size = 0; fm_rx_update_af_cache()
647 fmdbg("No of expected AF : %d\n", fmdev->rx.stat_info.af_list_max); fm_rx_update_af_cache()
658 freq = fmdev->rx.region.bot_freq + (af * 100); fm_rx_update_af_cache()
659 if (freq == fmdev->rx.freq) { fm_rx_update_af_cache()
661 fmdev->rx.freq, freq); fm_rx_update_af_cache()
713 struct fm_rds *rds = &fmdev->rx.rds; fm_irq_handle_rdsdata_getcmd_resp()
765 if (fmdev->rx.stat_info.picode != cur_picode) fm_irq_handle_rdsdata_getcmd_resp()
766 fmdev->rx.stat_info.picode = cur_picode; fm_irq_handle_rdsdata_getcmd_resp()
859 if ((fmdev->rx.af_mode == FM_RX_RDS_AF_SWITCH_MODE_ON) && fm_irq_handle_low_rssi_start()
861 (fmdev->rx.freq != FM_UNDEFINED_FREQ) && fm_irq_handle_low_rssi_start()
862 (fmdev->rx.stat_info.afcache_size != 0)) { fm_irq_handle_low_rssi_start()
868 fmdev->rx.afjump_idx = 0; fm_irq_handle_low_rssi_start()
869 fmdev->rx.freq_before_jump = fmdev->rx.freq; fm_irq_handle_low_rssi_start()
884 payload = fmdev->rx.stat_info.picode; fm_irq_afjump_set_pi()
918 fmdbg("Swtich to %d KHz\n", fmdev->rx.stat_info.af_cache[fmdev->rx.afjump_idx]); fm_irq_afjump_setfreq()
919 frq_index = (fmdev->rx.stat_info.af_cache[fmdev->rx.afjump_idx] - fm_irq_afjump_setfreq()
920 fmdev->rx.region.bot_freq) / FM_FREQ_MUL; fm_irq_afjump_setfreq()
990 curr_freq = fmdev->rx.region.bot_freq + ((u32)read_freq * FM_FREQ_MUL); fm_irq_afjump_rd_freq_resp()
992 jumped_freq = fmdev->rx.stat_info.af_cache[fmdev->rx.afjump_idx]; fm_irq_afjump_rd_freq_resp()
995 if ((curr_freq != fmdev->rx.freq_before_jump) && (curr_freq == jumped_freq)) { fm_irq_afjump_rd_freq_resp()
997 fmdev->rx.freq = curr_freq; fm_irq_afjump_rd_freq_resp()
1001 if (fmdev->rx.af_mode == FM_RX_RDS_AF_SWITCH_MODE_ON) fm_irq_afjump_rd_freq_resp()
1006 fmdev->rx.afjump_idx++; fm_irq_afjump_rd_freq_resp()
1009 if (fmdev->rx.afjump_idx >= fmdev->rx.stat_info.afcache_size) { fm_irq_afjump_rd_freq_resp()
1061 poll_wait(file, &fmdev->rx.rds.read_queue, pts); fmc_is_rds_data_available()
1062 if (fmdev->rx.rds.rd_idx != fmdev->rx.rds.wr_idx) fmc_is_rds_data_available()
1077 if (fmdev->rx.rds.wr_idx == fmdev->rx.rds.rd_idx) { fmc_transfer_rds_from_internal_buff()
1081 ret = wait_event_interruptible(fmdev->rx.rds.read_queue, fmc_transfer_rds_from_internal_buff()
1082 (fmdev->rx.rds.wr_idx != fmdev->rx.rds.rd_idx)); fmc_transfer_rds_from_internal_buff()
1095 if (fmdev->rx.rds.wr_idx == fmdev->rx.rds.rd_idx) { fmc_transfer_rds_from_internal_buff()
1099 memcpy(tmpbuf, &fmdev->rx.rds.buff[fmdev->rx.rds.rd_idx], fmc_transfer_rds_from_internal_buff()
1101 fmdev->rx.rds.rd_idx += FM_RDS_BLK_SIZE; fmc_transfer_rds_from_internal_buff()
1102 if (fmdev->rx.rds.rd_idx >= fmdev->rx.rds.buf_size) fmc_transfer_rds_from_internal_buff()
1103 fmdev->rx.rds.rd_idx = 0; fmc_transfer_rds_from_internal_buff()
1133 if (fmdev->rx.freq == FM_UNDEFINED_FREQ) { fmc_get_freq()
1144 *cur_tuned_frq = fmdev->rx.freq; fmc_get_freq()
1422 fmdbg("Loading default rx configuration..\n"); fmc_set_mode()
1568 fmdev->rx.region = region_configs[default_radio_region]; fmc_prepare()
1570 fmdev->rx.mute_mode = FM_MUTE_OFF; fmc_prepare()
1571 fmdev->rx.rf_depend_mute = FM_RX_RF_DEPENDENT_MUTE_OFF; fmc_prepare()
1572 fmdev->rx.rds.flag = FM_RDS_DISABLE; fmc_prepare()
1573 fmdev->rx.freq = FM_UNDEFINED_FREQ; fmc_prepare()
1574 fmdev->rx.rds_mode = FM_RDS_SYSTEM_RDS; fmc_prepare()
1575 fmdev->rx.af_mode = FM_RX_RDS_AF_SWITCH_MODE_OFF; fmc_prepare()
1579 init_waitqueue_head(&fmdev->rx.rds.read_queue); fmc_prepare()
1601 wake_up_interruptible(&fmdev->rx.rds.read_queue); fmc_release()
1610 fmdev->rx.freq = 0; fmc_release()
1642 fmdev->rx.rds.buf_size = default_rds_buf * FM_RDS_BLK_SIZE; fm_drv_init()
1643 fmdev->rx.rds.buff = kzalloc(fmdev->rx.rds.buf_size, GFP_KERNEL); fm_drv_init()
1644 if (NULL == fmdev->rx.rds.buff) { fm_drv_init()
1660 kfree(fmdev->rx.rds.buff); fm_drv_init()
1674 kfree(fmdev->rx.rds.buff); fm_drv_exit()
/linux-4.4.14/drivers/net/wireless/ti/wl1251/
H A DMakefile1 wl1251-objs = main.o event.o tx.o rx.o ps.o cmd.o \
H A Ddebugfs.c122 DEBUGFS_FWSTATS_FILE(rx, out_of_mem, 20, "%u");
123 DEBUGFS_FWSTATS_FILE(rx, hdr_overflow, 20, "%u");
124 DEBUGFS_FWSTATS_FILE(rx, hw_stuck, 20, "%u");
125 DEBUGFS_FWSTATS_FILE(rx, dropped, 20, "%u");
126 DEBUGFS_FWSTATS_FILE(rx, fcs_err, 20, "%u");
127 DEBUGFS_FWSTATS_FILE(rx, xfr_hint_trig, 20, "%u");
128 DEBUGFS_FWSTATS_FILE(rx, path_reset, 20, "%u");
129 DEBUGFS_FWSTATS_FILE(rx, reset_counter, 20, "%u");
262 DEBUGFS_FWSTATS_DEL(rx, out_of_mem); wl1251_debugfs_delete_files()
263 DEBUGFS_FWSTATS_DEL(rx, hdr_overflow); wl1251_debugfs_delete_files()
264 DEBUGFS_FWSTATS_DEL(rx, hw_stuck); wl1251_debugfs_delete_files()
265 DEBUGFS_FWSTATS_DEL(rx, dropped); wl1251_debugfs_delete_files()
266 DEBUGFS_FWSTATS_DEL(rx, fcs_err); wl1251_debugfs_delete_files()
267 DEBUGFS_FWSTATS_DEL(rx, xfr_hint_trig); wl1251_debugfs_delete_files()
268 DEBUGFS_FWSTATS_DEL(rx, path_reset); wl1251_debugfs_delete_files()
269 DEBUGFS_FWSTATS_DEL(rx, reset_counter); wl1251_debugfs_delete_files()
363 DEBUGFS_FWSTATS_ADD(rx, out_of_mem); wl1251_debugfs_add_files()
364 DEBUGFS_FWSTATS_ADD(rx, hdr_overflow); wl1251_debugfs_add_files()
365 DEBUGFS_FWSTATS_ADD(rx, hw_stuck); wl1251_debugfs_add_files()
366 DEBUGFS_FWSTATS_ADD(rx, dropped); wl1251_debugfs_add_files()
367 DEBUGFS_FWSTATS_ADD(rx, fcs_err); wl1251_debugfs_add_files()
368 DEBUGFS_FWSTATS_ADD(rx, xfr_hint_trig); wl1251_debugfs_add_files()
369 DEBUGFS_FWSTATS_ADD(rx, path_reset); wl1251_debugfs_add_files()
370 DEBUGFS_FWSTATS_ADD(rx, reset_counter); wl1251_debugfs_add_files()
/linux-4.4.14/drivers/net/wireless/ti/wlcore/
H A DMakefile1 wlcore-objs = main.o cmd.o io.o event.o tx.o rx.o ps.o acx.o \
H A Drx.c30 #include "rx.h"
168 /* skb length not including rx descriptor */ wl1271_rx_handle_data()
181 * Copy packets from aggregation buffer to the skbs without rx wl1271_rx_handle_data()
202 wl1271_debug(DEBUG_RX, "rx skb 0x%p: %d B %s seq %d hlid %d", skb, wl1271_rx_handle_data()
314 "enabled rx filter %d", index); wl1271_rx_filter_enable()
321 wl1271_error("Failed to %s rx data filter %d (err=%d)", wl1271_rx_filter_enable()
/linux-4.4.14/arch/arm/kernel/
H A Ddebug.S28 .macro addruart_current, rx, tmp1, tmp2
29 addruart \tmp1, \tmp2, \rx
30 mrc p15, 0, \rx, c1, c0
31 tst \rx, #1
32 moveq \rx, \tmp1
33 movne \rx, \tmp2
37 .macro addruart_current, rx, tmp1, tmp2
38 addruart \rx, \tmp1, \tmp2
/linux-4.4.14/sound/arm/
H A Daaci.h77 #define SR_RXTOFE (1 << 11) /* rx timeout fifo empty */
78 #define SR_TXTO (1 << 10) /* rx timeout fifo nonempty */
80 #define SR_RXO (1 << 8) /* rx overrun */
82 #define SR_RXB (1 << 6) /* rx busy */
84 #define SR_RXFF (1 << 4) /* rx fifo full */
86 #define SR_RXHF (1 << 2) /* rx fifo half full */
88 #define SR_RXFE (1 << 0) /* rx fifo empty */
93 #define ISR_RXTOFEINTR (1 << 6) /* rx fifo empty */
95 #define ISR_ORINTR (1 << 4) /* rx overflow */
96 #define ISR_RXINTR (1 << 3) /* rx fifo */
115 #define ISR_RXTOFE (1 << 6) /* rx timeout fifo empty */
117 #define ISR_OR (1 << 4) /* rx fifo overrun */
118 #define ISR_RX (1 << 3) /* rx interrupt status */
120 #define ISR_RXTO (1 << 1) /* rx timeout */
126 #define IE_RXTOFE (1 << 6) /* rx timeout fifo empty */
128 #define IE_OR (1 << 4) /* rx fifo overrun */
129 #define IE_RX (1 << 3) /* rx interrupt status */
131 #define IE_RXTO (1 << 1) /* rx timeout */
140 #define SLFR_12RXV (1 << 10) /* slot 12 rx valid */
142 #define SLFR_2RXV (1 << 8) /* slot 2 rx valid */
144 #define SLFR_1RXV (1 << 6) /* slot 1 rx valid */
146 #define SLFR_12RXB (1 << 4) /* slot 12 rx busy */
148 #define SLFR_2RXB (1 << 2) /* slot 2 rx busy */
150 #define SLFR_1RXB (1 << 0) /* slot 1 rx busy */
/linux-4.4.14/drivers/usb/musb/
H A Dcppi_dma.c83 /* zero out entire rx state RAM entry for the channel */ cppi_reset_rx()
84 static void cppi_reset_rx(struct cppi_rx_stateram __iomem *rx) cppi_reset_rx() argument
86 musb_writel(&rx->rx_skipbytes, 0, 0); cppi_reset_rx()
87 musb_writel(&rx->rx_head, 0, 0); cppi_reset_rx()
88 musb_writel(&rx->rx_sop, 0, 0); cppi_reset_rx()
89 musb_writel(&rx->rx_current, 0, 0); cppi_reset_rx()
90 musb_writel(&rx->rx_buf_current, 0, 0); cppi_reset_rx()
91 musb_writel(&rx->rx_len_len, 0, 0); cppi_reset_rx()
92 musb_writel(&rx->rx_cnt_cnt, 0, 0); cppi_reset_rx()
163 for (i = 0; i < ARRAY_SIZE(controller->rx); i++) { cppi_controller_start()
164 controller->rx[i].transmit = false; cppi_controller_start()
165 controller->rx[i].index = i; cppi_controller_start()
171 for (i = 0; i < ARRAY_SIZE(controller->rx); i++) cppi_controller_start()
172 cppi_pool_init(controller, controller->rx + i); cppi_controller_start()
177 /* initialise tx/rx channel head pointers to zero */ cppi_controller_start()
188 for (i = 0; i < ARRAY_SIZE(controller->rx); i++) { cppi_controller_start()
189 struct cppi_channel *rx_ch = controller->rx + i; cppi_controller_start()
190 struct cppi_rx_stateram __iomem *rx; cppi_controller_start() local
194 rx = tibase + DAVINCI_RXCPPI_STATERAM_OFFSET(i); cppi_controller_start()
195 rx_ch->state_ram = rx; cppi_controller_start()
196 cppi_reset_rx(rx); cppi_controller_start()
205 /* enable tx/rx CPPI control */ cppi_controller_start()
209 /* disable RNDIS mode, also host rx RNDIS autorequest */ cppi_controller_start()
241 for (i = 0; i < ARRAY_SIZE(controller->rx); i++) cppi_controller_stop()
242 cppi_pool_free(controller->rx + i); cppi_controller_stop()
248 /*disable tx/rx cppi */ cppi_controller_stop()
305 if (index >= ARRAY_SIZE(controller->rx)) { cppi_channel_allocate()
309 cppi_ch = controller->rx + index; cppi_channel_allocate()
353 struct cppi_rx_stateram __iomem *rx = c->state_ram; cppi_dump_rx() local
367 musb_readl(&rx->rx_skipbytes, 0), cppi_dump_rx()
368 musb_readl(&rx->rx_head, 0), cppi_dump_rx()
369 musb_readl(&rx->rx_sop, 0), cppi_dump_rx()
370 musb_readl(&rx->rx_current, 0), cppi_dump_rx()
372 musb_readl(&rx->rx_buf_current, 0), cppi_dump_rx()
373 musb_readl(&rx->rx_len_len, 0), cppi_dump_rx()
374 musb_readl(&rx->rx_cnt_cnt, 0), cppi_dump_rx()
375 musb_readl(&rx->rx_complete, 0) cppi_dump_rx()
438 static void cppi_dump_rxq(int level, const char *tag, struct cppi_channel *rx) cppi_dump_rxq() argument
442 cppi_dump_rx(level, rx, tag); cppi_dump_rxq()
443 if (rx->last_processed) cppi_dump_rxq()
444 cppi_dump_rxbd("last", rx->last_processed); cppi_dump_rxq()
445 for (bd = rx->head; bd; bd = bd->next) cppi_dump_rxq()
453 static inline int cppi_autoreq_update(struct cppi_channel *rx, cppi_autoreq_update() argument
464 val = tmp & ~((0x3) << (rx->index * 2)); cppi_autoreq_update()
472 val |= ((0x3) << (rx->index * 2)); cppi_autoreq_update()
476 val |= ((0x1) << (rx->index * 2)); cppi_autoreq_update()
495 if (n_bds && rx->channel.actual_len) { cppi_autoreq_update()
496 void __iomem *regs = rx->hw_ep->regs; cppi_autoreq_update()
730 * try rx rndis mode
748 * @rx: dma channel
761 cppi_next_rx_segment(struct musb *musb, struct cppi_channel *rx, int onepacket) cppi_next_rx_segment() argument
763 unsigned maxpacket = rx->maxpacket; cppi_next_rx_segment()
764 dma_addr_t addr = rx->buf_dma + rx->offset; cppi_next_rx_segment()
765 size_t length = rx->buf_len - rx->offset; cppi_next_rx_segment()
771 struct cppi_rx_stateram __iomem *rx_ram = rx->state_ram; cppi_next_rx_segment()
810 n_bds = cppi_autoreq_update(rx, tibase, onepacket, n_bds); cppi_next_rx_segment()
812 cppi_rndis_update(rx, 1, musb->ctrl_base, is_rndis); cppi_next_rx_segment()
818 rx->index, maxpacket, cppi_next_rx_segment()
824 DAVINCI_RXCPPI_BUFCNT0_REG + (rx->index * 4)) cppi_next_rx_segment()
827 rx->channel.actual_len, rx->buf_len); cppi_next_rx_segment()
832 bd = cppi_bd_alloc(rx); cppi_next_rx_segment()
833 rx->head = bd; cppi_next_rx_segment()
840 bd = cppi_bd_alloc(rx); cppi_next_rx_segment()
856 rx->offset += bd_len; cppi_next_rx_segment()
867 WARNING("rx dma%d -- no BDs? need %d\n", rx->index, n_bds); cppi_next_rx_segment()
870 WARNING("rx dma%d -- only %d of %d BDs\n", rx->index, i, n_bds); cppi_next_rx_segment()
875 bd = rx->head; cppi_next_rx_segment()
876 rx->tail = tail; cppi_next_rx_segment()
886 for (d = rx->head; d; d = d->next) cppi_next_rx_segment()
890 tail = rx->last_processed; cppi_next_rx_segment()
896 core_rxirq_enable(tibase, rx->index + 1); cppi_next_rx_segment()
911 DAVINCI_RXCPPI_BUFCNT0_REG + (rx->index * 4)) cppi_next_rx_segment()
916 DAVINCI_RXCPPI_BUFCNT0_REG + (rx->index * 4), cppi_next_rx_segment()
920 DAVINCI_RXCPPI_BUFCNT0_REG + (rx->index * 4), cppi_next_rx_segment()
924 DAVINCI_RXCPPI_BUFCNT0_REG + (rx->index * 4)) cppi_next_rx_segment()
928 rx->index, i, n_bds); cppi_next_rx_segment()
930 DAVINCI_RXCPPI_BUFCNT0_REG + (rx->index * 4), cppi_next_rx_segment()
934 cppi_dump_rx(4, rx, "/S"); cppi_next_rx_segment()
1004 struct cppi_channel *rx = &cppi->rx[ch]; cppi_rx_scan() local
1005 struct cppi_rx_stateram __iomem *state = rx->state_ram; cppi_rx_scan()
1007 struct cppi_descriptor *last = rx->last_processed; cppi_rx_scan()
1012 void __iomem *regs = rx->hw_ep->regs; cppi_rx_scan()
1015 cppi_dump_rx(6, rx, "/K"); cppi_rx_scan()
1017 bd = last ? last->next : rx->head; cppi_rx_scan()
1036 rx->channel.actual_len); cppi_rx_scan()
1054 dev_dbg(musb->controller, "rx short %d/%d (%d)\n", cppi_rx_scan()
1056 rx->channel.actual_len); cppi_rx_scan()
1075 rx->channel.actual_len += len; cppi_rx_scan()
1077 cppi_bd_free(rx, last); cppi_rx_scan()
1084 rx->last_processed = last; cppi_rx_scan()
1090 if (safe2ack == 0 || safe2ack == rx->last_processed->dma) cppi_rx_scan()
1093 cppi_bd_free(rx, last); cppi_rx_scan()
1094 rx->last_processed = NULL; cppi_rx_scan()
1099 WARN_ON(rx->head); cppi_rx_scan()
1101 musb_ep_select(cppi->mregs, rx->index + 1); cppi_rx_scan()
1105 rx->index, cppi_rx_scan()
1106 rx->head, rx->tail, cppi_rx_scan()
1107 rx->last_processed cppi_rx_scan()
1109 rx->last_processed->dma cppi_rx_scan()
1113 cppi_dump_rxq(4, "/what?", rx); cppi_rx_scan()
1119 rx->head = bd; cppi_rx_scan()
1124 csr = musb_readw(rx->hw_ep->regs, MUSB_RXCSR); cppi_rx_scan()
1131 csr = musb_readw(rx->hw_ep->regs, MUSB_RXCSR); cppi_rx_scan()
1134 rx->head = NULL; cppi_rx_scan()
1135 rx->tail = NULL; cppi_rx_scan()
1138 cppi_dump_rx(6, rx, completed ? "/completed" : "/cleaned"); cppi_rx_scan()
1148 u32 rx, tx; cppi_interrupt() local
1159 rx = musb_readl(tibase, DAVINCI_RXCPPI_MASKED_REG); cppi_interrupt()
1161 if (!tx && !rx) { cppi_interrupt()
1167 dev_dbg(musb->controller, "CPPI IRQ Tx%x Rx%x\n", tx, rx); cppi_interrupt()
1223 * REVISIT use the same ack strategy as rx cppi_interrupt()
1260 for (index = 0; rx; rx = rx >> 1, index++) { cppi_interrupt()
1262 if (rx & 1) { cppi_interrupt()
1265 rx_ch = cppi->rx + index; cppi_interrupt()
H A Dcppi_dma.h51 #define CPPI_ZERO_SET ((u32)(1 << 23)) /* rx saw zlp; tx issues one */
52 #define CPPI_RXABT_MASK ((u32)(1 << 19)) /* need more rx buffers */
124 struct cppi_channel rx[4]; member in struct:cppi
/linux-4.4.14/drivers/staging/most/hdm-i2c/
H A Dhdm_i2c.c54 struct rx { struct in struct:hdm_i2c
59 } rx; member in struct:hdm_i2c
96 schedule_delayed_work(&dev->rx.dwork, configure_channel()
126 mutex_lock(&dev->rx.list_mutex); enqueue()
127 list_add_tail(&mbo->list, &dev->rx.list); enqueue()
128 mutex_unlock(&dev->rx.list_mutex); enqueue()
129 wake_up_interruptible(&dev->rx.waitq); enqueue()
169 mutex_lock(&dev->rx.list_mutex); poison_channel()
170 while (!list_empty(&dev->rx.list)) { poison_channel()
171 mbo = list_first_mbo(&dev->rx.list); poison_channel()
173 mutex_unlock(&dev->rx.list_mutex); poison_channel()
179 mutex_lock(&dev->rx.list_mutex); poison_channel()
181 mutex_unlock(&dev->rx.list_mutex); poison_channel()
182 wake_up_interruptible(&dev->rx.waitq); poison_channel()
225 if (wait_event_interruptible(dev->rx.waitq, do_rx_work()
227 !list_empty(&dev->rx.list))) { do_rx_work()
235 mutex_lock(&dev->rx.list_mutex); do_rx_work()
238 if (!list_empty(&dev->rx.list)) do_rx_work()
241 mutex_unlock(&dev->rx.list_mutex); do_rx_work()
244 mbo = list_first_mbo(&dev->rx.list); do_rx_work()
246 mutex_unlock(&dev->rx.list_mutex); do_rx_work()
262 struct hdm_i2c *dev = container_of(work, struct hdm_i2c, rx.dwork.work); pending_rx_work()
268 schedule_delayed_work(&dev->rx.dwork, pending_rx_work()
300 schedule_delayed_work(&dev->rx.dwork, 0); most_irq_handler()
335 dev->capabilities[CH_RX].name_suffix = "rx"; i2c_probe()
348 INIT_LIST_HEAD(&dev->rx.list); i2c_probe()
349 mutex_init(&dev->rx.list_mutex); i2c_probe()
350 init_waitqueue_head(&dev->rx.waitq); i2c_probe()
352 INIT_DELAYED_WORK(&dev->rx.dwork, pending_rx_work); i2c_probe()
403 cancel_delayed_work_sync(&dev->rx.dwork); i2c_remove()
/linux-4.4.14/drivers/net/ethernet/freescale/
H A Ducc_geth_ethtool.c44 "rx-64-frames",
45 "rx-65-127-frames",
46 "rx-128-255-frames",
51 "rx-frames",
52 "rx-bytes-ok",
53 "rx-bytes-all",
54 "rx-multicast-frames",
55 "rx-broadcast-frames",
58 "rx-dropped-frames",
77 "rx-crc-errors",
78 "rx-alignment-errors",
79 "rx-in-range-length-errors",
80 "rx-out-of-range-length-errors",
81 "rx-too-long-frames",
82 "rx-runt",
83 "rx-very-long-event",
84 "rx-symbol-errors",
85 "rx-busy-drop-frames",
88 "rx-mismatch-drop-frames",
89 "rx-small-than-64",
90 "rx-256-511-frames",
91 "rx-512-1023-frames",
92 "rx-1024-1518-frames",
93 "rx-jumbo-frames",
94 "rx-mac-error-loss",
95 "rx-pause-frames",
97 "rx-vlan-removed",
98 "rx-vlan-replaced",
99 "rx-vlan-inserted",
100 "rx-ip-checksum-errors",
H A Dgianfar_ethtool.c65 "rx-allocation-errors",
66 "rx-large-frame-errors",
67 "rx-short-frame-errors",
68 "rx-non-octet-errors",
69 "rx-crc-errors",
70 "rx-overrun-errors",
71 "rx-busy-errors",
72 "rx-babbling-errors",
73 "rx-truncated-frames",
79 "tx-rx-64-frames",
80 "tx-rx-65-127-frames",
81 "tx-rx-128-255-frames",
82 "tx-rx-256-511-frames",
83 "tx-rx-512-1023-frames",
84 "tx-rx-1024-1518-frames",
85 "tx-rx-1519-1522-good-vlan",
86 "rx-bytes",
87 "rx-packets",
88 "rx-fcs-errors",
91 "rx-control-frame-packets",
92 "rx-pause-frame-packets",
93 "rx-unknown-op-code",
94 "rx-alignment-error",
95 "rx-frame-length-error",
96 "rx-code-error",
97 "rx-carrier-sense-error",
98 "rx-undersize-packets",
99 "rx-oversize-packets",
100 "rx-fragmented-frames",
101 "rx-jabber-frames",
102 "rx-dropped-frames",
334 * normal {rx,tx}_* coalescing parameters are used. gfar_gcoalesce()
338 * is above pkt_rate_high, the {rx,tx}_*_high parameters are gfar_gcoalesce()
400 /* Set up rx coalescing */ gfar_scoalesce()
445 * rx, rx_mini, and rx_jumbo rings are the same size, as mini and
/linux-4.4.14/drivers/staging/iio/meter/
H A Dade7854-i2c.c116 ret = i2c_master_recv(st->i2c, st->rx, 1); ade7854_i2c_read_reg_8()
120 *val = st->rx[0]; ade7854_i2c_read_reg_8()
142 ret = i2c_master_recv(st->i2c, st->rx, 2); ade7854_i2c_read_reg_16()
146 *val = (st->rx[0] << 8) | st->rx[1]; ade7854_i2c_read_reg_16()
168 ret = i2c_master_recv(st->i2c, st->rx, 3); ade7854_i2c_read_reg_24()
172 *val = (st->rx[0] << 16) | (st->rx[1] << 8) | st->rx[2]; ade7854_i2c_read_reg_24()
194 ret = i2c_master_recv(st->i2c, st->rx, 3); ade7854_i2c_read_reg_32()
198 *val = (st->rx[0] << 24) | (st->rx[1] << 16) | ade7854_i2c_read_reg_32()
199 (st->rx[2] << 8) | st->rx[3]; ade7854_i2c_read_reg_32()
H A Dade7759.h42 * @buf_lock: mutex to protect tx and rx
44 * @rx: receive buffer
50 u8 rx[ADE7759_MAX_RX]; member in struct:ade7759_state
H A Dade7753.h62 * @rx: receive buffer
63 * @buf_lock: mutex to protect tx and rx
69 u8 rx[ADE7753_MAX_RX]; member in struct:ade7753_state
H A Dade7754.h79 * @buf_lock: mutex to protect tx and rx
81 * @rx: receive buffer
87 u8 rx[ADE7754_MAX_RX]; member in struct:ade7754_state
H A Dade7854-spi.c137 .rx_buf = st->rx, ade7854_spi_read_reg_8()
155 *val = st->rx[0]; ade7854_spi_read_reg_8()
175 .rx_buf = st->rx, ade7854_spi_read_reg_16()
192 *val = be16_to_cpup((const __be16 *)st->rx); ade7854_spi_read_reg_16()
212 .rx_buf = st->rx, ade7854_spi_read_reg_24()
230 *val = (st->rx[0] << 16) | (st->rx[1] << 8) | st->rx[2]; ade7854_spi_read_reg_24()
250 .rx_buf = st->rx, ade7854_spi_read_reg_32()
268 *val = be32_to_cpup((const __be32 *)st->rx); ade7854_spi_read_reg_32()
H A Dade7758_core.c114 .rx_buf = st->rx, ade7758_spi_read_reg_8()
130 *val = st->rx[0]; ade7758_spi_read_reg_8()
153 .rx_buf = st->rx, ade7758_spi_read_reg_16()
172 *val = (st->rx[0] << 8) | st->rx[1]; ade7758_spi_read_reg_16()
195 .rx_buf = st->rx, ade7758_spi_read_reg_24()
213 *val = (st->rx[0] << 16) | (st->rx[1] << 8) | st->rx[2]; ade7758_spi_read_reg_24()
832 st->rx = kcalloc(ADE7758_MAX_RX, sizeof(*st->rx), GFP_KERNEL); ade7758_probe()
833 if (!st->rx) ade7758_probe()
879 kfree(st->rx); ade7758_probe()
893 kfree(st->rx); ade7758_remove()
H A Dade7758.h113 * @rx: receive buffer
114 * @buf_lock: mutex to protect tx and rx
120 u8 *rx; member in struct:ade7758_state
H A Dade7854.h149 * @buf_lock: mutex to protect tx and rx
151 * @rx: receive buffer
167 u8 rx[ADE7854_MAX_RX]; member in struct:ade7854_state
/linux-4.4.14/drivers/mfd/
H A Dipaq-micro.c149 struct ipaq_micro_rxdev *rx = &micro->rx; micro_process_char() local
151 switch (rx->state) { micro_process_char()
154 rx->state = STATE_ID; /* Next byte is the id and len */ micro_process_char()
157 rx->id = (ch & 0xf0) >> 4; micro_process_char()
158 rx->len = (ch & 0x0f); micro_process_char()
159 rx->index = 0; micro_process_char()
160 rx->chksum = ch; micro_process_char()
161 rx->state = (rx->len > 0) ? STATE_DATA : STATE_CHKSUM; micro_process_char()
164 rx->chksum += ch; micro_process_char()
165 rx->buf[rx->index] = ch; micro_process_char()
166 if (++rx->index == rx->len) micro_process_char()
167 rx->state = STATE_CHKSUM; micro_process_char()
170 if (ch == rx->chksum) micro_process_char()
171 micro_rx_msg(micro, rx->id, rx->len, rx->buf); micro_process_char()
172 rx->state = STATE_SOF; micro_process_char()
184 dev_err(micro->dev, "rx: parity error\n"); micro_rx_chars()
186 dev_err(micro->dev, "rx: framing error\n"); micro_rx_chars()
188 dev_err(micro->dev, "rx: overrun error\n"); micro_rx_chars()
308 struct ipaq_micro_rxdev *rx = &micro->rx; micro_reset_comm() local
315 rx->state = STATE_SOF; /* Reset the state machine */ micro_reset_comm()
/linux-4.4.14/drivers/net/usb/
H A Dasix_common.c54 struct asix_rx_fixup_info *rx) asix_rx_fixup_internal()
68 if (rx->remaining && (rx->remaining + sizeof(u32) <= skb->len)) { asix_rx_fixup_internal()
69 offset = ((rx->remaining + 1) & 0xfffe); asix_rx_fixup_internal()
70 rx->header = get_unaligned_le32(skb->data + offset); asix_rx_fixup_internal()
73 size = (u16)(rx->header & 0x7ff); asix_rx_fixup_internal()
74 if (size != ((~rx->header >> 16) & 0x7ff)) { asix_rx_fixup_internal()
76 rx->remaining); asix_rx_fixup_internal()
77 if (rx->ax_skb) { asix_rx_fixup_internal()
78 kfree_skb(rx->ax_skb); asix_rx_fixup_internal()
79 rx->ax_skb = NULL; asix_rx_fixup_internal()
85 rx->remaining = 0; asix_rx_fixup_internal()
93 if (!rx->remaining) { asix_rx_fixup_internal()
95 rx->header = get_unaligned_le16( asix_rx_fixup_internal()
97 rx->split_head = true; asix_rx_fixup_internal()
102 if (rx->split_head == true) { asix_rx_fixup_internal()
103 rx->header |= (get_unaligned_le16( asix_rx_fixup_internal()
105 rx->split_head = false; asix_rx_fixup_internal()
108 rx->header = get_unaligned_le32(skb->data + asix_rx_fixup_internal()
114 size = (u16)(rx->header & 0x7ff); asix_rx_fixup_internal()
115 if (size != ((~rx->header >> 16) & 0x7ff)) { asix_rx_fixup_internal()
117 rx->header, offset); asix_rx_fixup_internal()
131 rx->ax_skb = netdev_alloc_skb_ip_align(dev->net, size); asix_rx_fixup_internal()
133 rx->remaining = size; asix_rx_fixup_internal()
136 if (rx->remaining > skb->len - offset) { asix_rx_fixup_internal()
138 rx->remaining -= copy_length; asix_rx_fixup_internal()
140 copy_length = rx->remaining; asix_rx_fixup_internal()
141 rx->remaining = 0; asix_rx_fixup_internal()
144 if (rx->ax_skb) { asix_rx_fixup_internal()
145 data = skb_put(rx->ax_skb, copy_length); asix_rx_fixup_internal()
147 if (!rx->remaining) asix_rx_fixup_internal()
148 usbnet_skb_return(dev, rx->ax_skb); asix_rx_fixup_internal()
166 struct asix_rx_fixup_info *rx = &dp->rx_fixup_info; asix_rx_fixup_common() local
168 return asix_rx_fixup_internal(dev, skb, rx); asix_rx_fixup_common()
53 asix_rx_fixup_internal(struct usbnet *dev, struct sk_buff *skb, struct asix_rx_fixup_info *rx) asix_rx_fixup_internal() argument
/linux-4.4.14/drivers/net/ethernet/stmicro/stmmac/
H A Dnorm_desc.c79 if (unlikely(p->des01.rx.last_descriptor == 0)) { ndesc_get_rx_status()
86 if (unlikely(p->des01.rx.error_summary)) { ndesc_get_rx_status()
87 if (unlikely(p->des01.rx.descriptor_error)) ndesc_get_rx_status()
89 if (unlikely(p->des01.rx.sa_filter_fail)) ndesc_get_rx_status()
91 if (unlikely(p->des01.rx.overflow_error)) ndesc_get_rx_status()
93 if (unlikely(p->des01.rx.ipc_csum_error)) ndesc_get_rx_status()
95 if (unlikely(p->des01.rx.collision)) { ndesc_get_rx_status()
99 if (unlikely(p->des01.rx.crc_error)) { ndesc_get_rx_status()
105 if (unlikely(p->des01.rx.dribbling)) ndesc_get_rx_status()
108 if (unlikely(p->des01.rx.length_error)) { ndesc_get_rx_status()
112 if (unlikely(p->des01.rx.mii_error)) { ndesc_get_rx_status()
117 if (p->des01.rx.vlan_tag) ndesc_get_rx_status()
127 p->des01.rx.own = 1; ndesc_init_rx_desc()
128 p->des01.rx.buffer1_size = BUF_SIZE_2KiB - 1; ndesc_init_rx_desc()
136 p->des01.rx.disable_ic = 1; ndesc_init_rx_desc()
155 return p->des01.rx.own; ndesc_get_rx_owner()
165 p->des01.rx.own = 1; ndesc_set_rx_owner()
216 return p->des01.rx.frame_length - 2; ndesc_get_rx_frame_len()
218 return p->des01.rx.frame_length; ndesc_get_rx_frame_len()
H A Ddescs_com.h66 p->des01.rx.buffer2_size = BUF_SIZE_2KiB - 1; ndesc_rx_set_on_ring()
68 p->des01.rx.end_ring = 1; ndesc_rx_set_on_ring()
117 p->des01.rx.second_address_chained = 1; ndesc_rx_set_on_chain()
/linux-4.4.14/arch/s390/kernel/
H A Duprobes.c259 union split_register *rx; handle_insn_ril() local
266 rx = (union split_register *) &regs->gprs[insn->reg]; handle_insn_ril()
274 rx->u64 = (unsigned long)uptr; handle_insn_ril()
281 rc = emu_load_ril((u16 __user *)uptr, &rx->u32[1]); handle_insn_ril()
284 rc = emu_load_ril((s16 __user *)uptr, &rx->u64); handle_insn_ril()
287 rc = emu_load_ril((s16 __user *)uptr, &rx->u32[1]); handle_insn_ril()
290 rc = emu_load_ril((u16 __user *)uptr, &rx->u64); handle_insn_ril()
293 rc = emu_load_ril((u64 __user *)uptr, &rx->u64); handle_insn_ril()
296 rc = emu_load_ril((s32 __user *)uptr, &rx->u64); handle_insn_ril()
299 rc = emu_load_ril((u32 __user *)uptr, &rx->u32[1]); handle_insn_ril()
302 rc = emu_load_ril((u32 __user *)uptr, &rx->u64); handle_insn_ril()
305 rc = emu_store_ril(regs, (u16 __user *)uptr, &rx->u16[3]); handle_insn_ril()
308 rc = emu_store_ril(regs, (u64 __user *)uptr, &rx->u64); handle_insn_ril()
311 rc = emu_store_ril(regs, (u32 __user *)uptr, &rx->u32[1]); handle_insn_ril()
322 rc = emu_cmp_ril(regs, (s16 __user *)uptr, &rx->s64); handle_insn_ril()
325 rc = emu_cmp_ril(regs, (s16 __user *)uptr, &rx->s32[1]); handle_insn_ril()
328 rc = emu_cmp_ril(regs, (u16 __user *)uptr, &rx->u64); handle_insn_ril()
331 rc = emu_cmp_ril(regs, (u16 __user *)uptr, &rx->u32[1]); handle_insn_ril()
334 rc = emu_cmp_ril(regs, (s64 __user *)uptr, &rx->s64); handle_insn_ril()
337 rc = emu_cmp_ril(regs, (u64 __user *)uptr, &rx->u64); handle_insn_ril()
340 rc = emu_cmp_ril(regs, (s32 __user *)uptr, &rx->s64); handle_insn_ril()
343 rc = emu_cmp_ril(regs, (s32 __user *)uptr, &rx->s32[1]); handle_insn_ril()
346 rc = emu_cmp_ril(regs, (u32 __user *)uptr, &rx->u64); handle_insn_ril()
349 rc = emu_cmp_ril(regs, (u32 __user *)uptr, &rx->u32[1]); handle_insn_ril()
H A Ddiag.c138 static inline int __diag14(unsigned long rx, unsigned long ry1, __diag14() argument
152 : "d" (rx), "d" (_ry1) __diag14()
158 int diag14(unsigned long rx, unsigned long ry1, unsigned long subcode) diag14() argument
161 return __diag14(rx, ry1, subcode); diag14()
/linux-4.4.14/drivers/net/wireless/ath/ath9k/
H A Drecv.c34 * buffer (or rx fifo). This can incorrectly acknowledge packets
55 * setup rx descriptors. The rx_bufsize here tells the hardware ath_rx_buf_link()
63 if (sc->rx.rxlink) ath_rx_buf_link()
64 *sc->rx.rxlink = bf->bf_daddr; ath_rx_buf_link()
68 sc->rx.rxlink = &ds->ds_link; ath_rx_buf_link()
74 if (sc->rx.buf_hold) ath_rx_buf_relink()
75 ath_rx_buf_link(sc, sc->rx.buf_hold, flush); ath_rx_buf_relink()
77 sc->rx.buf_hold = bf; ath_rx_buf_relink()
84 sc->rx.defant = antenna; ath_setdefantenna()
85 sc->rx.rxotherant = 0; ath_setdefantenna()
95 /* configure rx filter */ ath_opmode_init()
118 rx_edma = &sc->rx.rx_edma[qtype]; ath_rx_edma_buf_link()
122 bf = list_first_entry(&sc->rx.rxbuf, struct ath_rxbuf, list); ath_rx_edma_buf_link()
144 if (list_empty(&sc->rx.rxbuf)) { ath_rx_addbuffer_edma()
145 ath_dbg(common, QUEUE, "No free rx buf available\n"); ath_rx_addbuffer_edma()
149 list_for_each_entry_safe(bf, tbf, &sc->rx.rxbuf, list) ath_rx_addbuffer_edma()
162 rx_edma = &sc->rx.rx_edma[qtype]; ath_rx_remove_buffer()
167 list_add_tail(&bf->list, &sc->rx.rxbuf); ath_rx_remove_buffer()
180 list_for_each_entry(bf, &sc->rx.rxbuf, list) { ath_rx_edma_cleanup()
210 ath_rx_edma_init_queue(&sc->rx.rx_edma[ATH9K_RX_QUEUE_LP], ath_rx_edma_init()
212 ath_rx_edma_init_queue(&sc->rx.rx_edma[ATH9K_RX_QUEUE_HP], ath_rx_edma_init()
220 INIT_LIST_HEAD(&sc->rx.rxbuf); ath_rx_edma_init()
246 list_add_tail(&bf->list, &sc->rx.rxbuf); ath_rx_edma_init()
289 /* Initialize rx descriptors */ ath_rx_init()
291 error = ath_descdma_setup(sc, &sc->rx.rxdma, &sc->rx.rxbuf, ath_rx_init()
292 "rx", nbufs, 1, 0); ath_rx_init()
295 "failed to allocate rx descriptors: %d\n", ath_rx_init()
300 list_for_each_entry(bf, &sc->rx.rxbuf, list) { ath_rx_init()
323 sc->rx.rxlink = NULL; ath_rx_init()
343 list_for_each_entry(bf, &sc->rx.rxbuf, list) { ath_rx_cleanup()
447 if (list_empty(&sc->rx.rxbuf)) ath_startrecv()
450 sc->rx.buf_hold = NULL; ath_startrecv()
451 sc->rx.rxlink = NULL; ath_startrecv()
452 list_for_each_entry_safe(bf, tbf, &sc->rx.rxbuf, list) { ath_startrecv()
457 if (list_empty(&sc->rx.rxbuf)) ath_startrecv()
460 bf = list_first_entry(&sc->rx.rxbuf, struct ath_rxbuf, list); ath_startrecv()
490 sc->rx.rxlink = NULL; ath_stoprecv()
628 struct ath_rx_edma *rx_edma = &sc->rx.rx_edma[qtype]; ath_edma_get_buffers()
656 list_add_tail(&bf->list, &sc->rx.rxbuf); ath_edma_get_buffers()
665 list_add_tail(&bf->list, &sc->rx.rxbuf); ath_edma_get_buffers()
700 if (list_empty(&sc->rx.rxbuf)) { ath_get_next_rx_buf()
701 sc->rx.rxlink = NULL; ath_get_next_rx_buf()
705 bf = list_first_entry(&sc->rx.rxbuf, struct ath_rxbuf, list); ath_get_next_rx_buf()
706 if (bf == sc->rx.buf_hold) ath_get_next_rx_buf()
720 * a self-linked list to avoid rx overruns. ath_get_next_rx_buf()
729 if (list_is_last(&bf->list, &sc->rx.rxbuf)) { ath_get_next_rx_buf()
730 sc->rx.rxlink = NULL; ath_get_next_rx_buf()
814 bool discard_current = sc->rx.discard_next; ath9k_rx_skb_preprocess()
823 sc->rx.discard_next = false; ath9k_rx_skb_preprocess()
924 sc->rx.num_pkts++; ath9k_rx_skb_preprocess()
930 sc->rx.discard_next = rx_stats->rs_more; ath9k_rx_skb_preprocess()
955 * Change the default rx antenna if rx diversity ath9k_antenna_check()
958 if (sc->rx.defant != rs->rs_antenna) { ath9k_antenna_check()
959 if (++sc->rx.rxotherant >= 3) ath9k_antenna_check()
962 sc->rx.rxotherant = 0; ath9k_antenna_check()
979 rxs->ampdu_reference = sc->rx.ampdu_ref; ath9k_apply_ampdu_details()
983 sc->rx.ampdu_ref++; ath9k_apply_ampdu_details()
1039 if (sc->rx.frag) ath_rx_tasklet()
1040 hdr_skb = sc->rx.frag; ath_rx_tasklet()
1058 * skb and put it at the tail of the sc->rx.rxbuf list for ath_rx_tasklet()
1095 if (sc->rx.frag) { ath_rx_tasklet()
1097 dev_kfree_skb_any(sc->rx.frag); ath_rx_tasklet()
1102 sc->rx.frag = skb; ath_rx_tasklet()
1106 if (sc->rx.frag) { ath_rx_tasklet()
1115 sc->rx.frag = NULL; ath_rx_tasklet()
1145 if (sc->rx.frag) { ath_rx_tasklet()
1146 dev_kfree_skb_any(sc->rx.frag); ath_rx_tasklet()
1147 sc->rx.frag = NULL; ath_rx_tasklet()
1150 list_add_tail(&bf->list, &sc->rx.rxbuf); ath_rx_tasklet()
/linux-4.4.14/drivers/staging/media/lirc/
H A Dlirc_zilog.c112 struct IR_rx *rx; member in struct:IR
181 * ir->rx set to NULL and deallocated - happens before ir->rx->ir put() release_ir_device()
182 * ir->rx->task kthread stopped - happens before ir->rx->ir put() release_ir_device()
214 struct IR_rx *rx; get_ir_rx() local
217 rx = ir->rx; get_ir_rx()
218 if (rx != NULL) get_ir_rx()
219 kref_get(&rx->ref); get_ir_rx()
221 return rx; get_ir_rx()
224 static void destroy_rx_kthread(struct IR_rx *rx, bool ir_devices_lock_held) destroy_rx_kthread() argument
227 if (!IS_ERR_OR_NULL(rx->task)) { destroy_rx_kthread()
228 kthread_stop(rx->task); destroy_rx_kthread()
229 rx->task = NULL; destroy_rx_kthread()
230 /* Put the ir ptr that ir_probe() gave to the rx poll thread */ destroy_rx_kthread()
231 put_ir_device(rx->ir, ir_devices_lock_held); destroy_rx_kthread()
237 struct IR_rx *rx = container_of(ref, struct IR_rx, ref); release_ir_rx() local
238 struct IR *ir = rx->ir; release_ir_rx()
247 /* Don't put_ir_device(rx->ir) here; lock can't be freed yet */ release_ir_rx()
248 ir->rx = NULL; release_ir_rx()
249 /* Don't do the kfree(rx) here; we still need to kill the poll thread */ release_ir_rx()
252 static int put_ir_rx(struct IR_rx *rx, bool ir_devices_lock_held) put_ir_rx() argument
255 struct IR *ir = rx->ir; put_ir_rx()
258 released = kref_put(&rx->ref, release_ir_rx); put_ir_rx()
260 /* Destroy the rx kthread while not holding the spinlock */ put_ir_rx()
262 destroy_rx_kthread(rx, ir_devices_lock_held); put_ir_rx()
263 kfree(rx); put_ir_rx()
267 /* Do a reference put() for the rx->ir reference, if we released rx */ put_ir_rx()
321 struct IR_rx *rx; add_to_buf() local
329 rx = get_ir_rx(ir); add_to_buf()
330 if (rx == NULL) add_to_buf()
333 /* Ensure our rx->c i2c_client remains valid for the duration */ add_to_buf()
334 mutex_lock(&rx->client_lock); add_to_buf()
335 if (rx->c == NULL) { add_to_buf()
336 mutex_unlock(&rx->client_lock); add_to_buf()
337 put_ir_rx(rx, false); add_to_buf()
369 ret = i2c_master_send(rx->c, sendbuf, 1); add_to_buf()
405 ret = i2c_master_recv(rx->c, keybuf, sizeof(keybuf)); add_to_buf()
412 rx->b[0] = keybuf[3]; add_to_buf()
413 rx->b[1] = keybuf[4]; add_to_buf()
414 rx->b[2] = keybuf[5]; add_to_buf()
417 rx->b[0], rx->b[1]); add_to_buf()
421 if (rx->hdpvr_data_fmt) { add_to_buf()
429 } else if ((rx->b[0] & 0x80) == 0) { add_to_buf()
435 code = (((__u16)rx->b[0] & 0x7f) << 6) | (rx->b[1] >> 2); add_to_buf()
446 mutex_unlock(&rx->client_lock); add_to_buf()
449 put_ir_rx(rx, false); add_to_buf()
895 struct IR_rx *rx; read() local
907 rx = get_ir_rx(ir); read()
908 if (rx == NULL) read()
969 put_ir_rx(rx, false); read()
1211 struct IR_rx *rx; poll() local
1217 rx = get_ir_rx(ir); poll()
1218 if (rx == NULL) { poll()
1417 struct IR_rx *rx = i2c_get_clientdata(client); ir_remove() local
1419 if (rx != NULL) { ir_remove()
1420 mutex_lock(&rx->client_lock); ir_remove()
1421 rx->c = NULL; ir_remove()
1422 mutex_unlock(&rx->client_lock); ir_remove()
1423 put_ir_rx(rx, false); ir_remove()
1451 struct IR_rx *rx; ir_probe() local
1514 rx = get_ir_rx(ir); ir_probe()
1550 if (rx == NULL && !tx_only) { ir_probe()
1561 rx = kzalloc(sizeof(struct IR_rx), GFP_KERNEL); ir_probe()
1562 if (rx == NULL) { ir_probe()
1566 kref_init(&rx->ref); ir_probe()
1567 ir->rx = rx; ir_probe()
1570 mutex_init(&rx->client_lock); ir_probe()
1571 rx->c = client; ir_probe()
1572 rx->hdpvr_data_fmt = ir_probe()
1576 rx->ir = get_ir_device(ir, true); ir_probe()
1578 /* An rx ref goes to the i2c_client */ ir_probe()
1586 /* An ir ref goes to the new rx polling kthread */ ir_probe()
1587 rx->task = kthread_run(lirc_thread, get_ir_device(ir, true), ir_probe()
1588 "zilog-rx-i2c-%d", adap->nr); ir_probe()
1589 if (IS_ERR(rx->task)) { ir_probe()
1590 ret = PTR_ERR(rx->task); ir_probe()
1596 /* Failure exit, so put back rx ref from i2c_client */ ir_probe()
1598 put_ir_rx(rx, true); ir_probe()
1626 if (rx != NULL) ir_probe()
1627 put_ir_rx(rx, true); ir_probe()
1638 if (rx != NULL) ir_probe()
1639 put_ir_rx(rx, true); ir_probe()
H A Dlirc_imon.c101 } rx; member in struct:imon_context
498 context->rx.count = 0; ir_open()
499 context->rx.initial_space = 1; ir_open()
500 context->rx.prev_bit = 0; ir_open()
555 int value = context->rx.count; submit_data()
562 if (context->rx.prev_bit) submit_data()
612 if (buf[7] == 1 && context->rx.initial_space) { imon_incoming_packet()
614 context->rx.prev_bit = 0; imon_incoming_packet()
615 context->rx.count = 4; imon_incoming_packet()
617 context->rx.count = 0; imon_incoming_packet()
625 if (curr_bit != context->rx.prev_bit) { imon_incoming_packet()
626 if (context->rx.count) { imon_incoming_packet()
628 context->rx.count = 0; imon_incoming_packet()
630 context->rx.prev_bit = curr_bit; imon_incoming_packet()
632 ++context->rx.count; imon_incoming_packet()
638 if (context->rx.count) { imon_incoming_packet()
640 context->rx.count = 0; imon_incoming_packet()
642 context->rx.initial_space = context->rx.prev_bit; imon_incoming_packet()
/linux-4.4.14/drivers/media/platform/vivid/
H A DMakefile3 vivid-radio-rx.o vivid-radio-tx.o vivid-radio-common.o \
/linux-4.4.14/drivers/net/wireless/ti/wl12xx/
H A Ddebugfs.c35 WL12XX_DEBUGFS_FWSTATS_FILE(rx, out_of_mem, "%u");
36 WL12XX_DEBUGFS_FWSTATS_FILE(rx, hdr_overflow, "%u");
37 WL12XX_DEBUGFS_FWSTATS_FILE(rx, hw_stuck, "%u");
38 WL12XX_DEBUGFS_FWSTATS_FILE(rx, dropped, "%u");
39 WL12XX_DEBUGFS_FWSTATS_FILE(rx, fcs_err, "%u");
40 WL12XX_DEBUGFS_FWSTATS_FILE(rx, xfr_hint_trig, "%u");
41 WL12XX_DEBUGFS_FWSTATS_FILE(rx, path_reset, "%u");
42 WL12XX_DEBUGFS_FWSTATS_FILE(rx, reset_counter, "%u");
145 DEBUGFS_FWSTATS_ADD(rx, out_of_mem); wl12xx_debugfs_add_files()
146 DEBUGFS_FWSTATS_ADD(rx, hdr_overflow); wl12xx_debugfs_add_files()
147 DEBUGFS_FWSTATS_ADD(rx, hw_stuck); wl12xx_debugfs_add_files()
148 DEBUGFS_FWSTATS_ADD(rx, dropped); wl12xx_debugfs_add_files()
149 DEBUGFS_FWSTATS_ADD(rx, fcs_err); wl12xx_debugfs_add_files()
150 DEBUGFS_FWSTATS_ADD(rx, xfr_hint_trig); wl12xx_debugfs_add_files()
151 DEBUGFS_FWSTATS_ADD(rx, path_reset); wl12xx_debugfs_add_files()
152 DEBUGFS_FWSTATS_ADD(rx, reset_counter); wl12xx_debugfs_add_files()
/linux-4.4.14/drivers/net/ethernet/broadcom/
H A Dbcm63xx_enet.h29 * hardware maximum rx/tx packet size including FCS, max mtu is
30 * actually 2047, but if we set max rx size register to 2047 we won't
207 /* hw view of rx & tx dma ring */
211 /* allocated size (in bytes) for rx & tx dma ring */
218 /* dma channel id for rx */
221 /* number of dma desc in rx ring */
224 /* cpu view of rx dma ring */
227 /* current number of armed descriptor given to hardware for rx */
230 /* next rx descriptor to fetch from hardware */
233 /* next dirty rx descriptor to refill */
236 /* size of allocated rx skbs */
239 /* list of skb given to hw for rx */
242 /* used when rx skb allocation failed, so we defer rx queue
246 /* lock rx_timeout against rx normal operation */
259 /* cpu view of rx dma ring */
H A Dbcm63xx_enet.c235 * refill rx queue
284 /* If rx ring is still empty, set a timer to try allocating bcm_enet_refill_rx()
287 dev_warn(&priv->pdev->dev, "unable to refill rx ring\n"); bcm_enet_refill_rx()
296 * timer callback to defer refill rx queue in case we're OOM
312 * extract packet from rx queue
418 /* kick rx dma */ bcm_enet_receive_queue()
508 /* rx queue is not yet empty/clean */ bcm_enet_poll()
512 /* no more packet in rx/tx queue, remove device from poll bcm_enet_poll()
516 /* restore rx/tx interrupt */ bcm_enet_poll()
552 * rx/tx dma interrupt handler
562 /* mask rx/tx interrupts */ bcm_enet_isr_dma()
684 * Change rx mode (promiscuous/allmulti) and update multicast list
764 /* rx flow control (pause frame handling) */ bcm_enet_set_flow()
839 phydev->pause == 1 ? "rx&tx" : "off"); bcm_enet_adjust_phy_link()
861 priv->pause_rx ? "rx" : "off", bcm_enet_adjust_link()
866 * open callback, allocate dma rings & buffers and start rx operation
949 /* allocate rx dma ring */ bcm_enet_open()
983 /* init & fill rx ring with skbs */ bcm_enet_open()
1004 dev_err(kdev, "cannot allocate rx skb queue\n"); bcm_enet_open()
1009 /* write rx & tx ring addresses */ bcm_enet_open()
1022 /* clear remaining state ram for rx & tx channel */ bcm_enet_open()
1035 /* set max rx/tx length */ bcm_enet_open()
1061 * kick rx dma channel */ bcm_enet_open()
1074 /* watch "packet transferred" interrupt in rx and tx */ bcm_enet_open()
1080 /* make sure we enable napi before rx interrupt */ bcm_enet_open()
1213 /* free the rx skb ring */ bcm_enet_stop()
1513 /* rx/tx ring is actually only limited by memory */ bcm_enet_get_ringparam()
1648 * align rx buffer size to dma burst len, account FCS since compute_hw_mtu()
1774 /* get rx & tx dma channel id for this mac */ bcm_enet_probe()
1882 /* init rx timeout (used for oom) */ bcm_enet_probe()
2140 * open callback, allocate dma rings & buffers and start rx operation
2170 /* allocate rx dma ring */ bcm_enetsw_open()
2174 dev_err(kdev, "cannot allocate rx ring %u\n", size); bcm_enetsw_open()
2199 dev_err(kdev, "cannot allocate rx skb queue\n"); bcm_enetsw_open()
2209 /* init & fill rx ring with skbs */ bcm_enetsw_open()
2213 dev_err(kdev, "cannot allocate rx skb queue\n"); bcm_enetsw_open()
2261 dev_err(kdev, "cannot allocate rx skb queue\n"); bcm_enetsw_open()
2266 /* write rx & tx ring addresses */ bcm_enetsw_open()
2272 /* clear remaining state ram for rx & tx channel */ bcm_enetsw_open()
2293 * kick rx dma channel bcm_enetsw_open()
2300 /* watch "packet transferred" interrupt in rx and tx */ bcm_enetsw_open()
2306 /* make sure we enable napi before rx interrupt */ bcm_enetsw_open()
2424 /* free the rx skb ring */ bcm_enetsw_stop()
2661 /* rx/tx ring is actually only limited by memory */ bcm_enetsw_get_ringparam()
2781 /* init rx timeout (used for oom) */ bcm_enetsw_probe()
/linux-4.4.14/include/uapi/linux/
H A Datm_eni.h14 int tx,rx; /* values are in percent and must be > 100 */ member in struct:eni_multipliers
H A Dtipc_config.h74 #define TIPC_CMD_NOOP 0x0000 /* tx none, rx none */
75 #define TIPC_CMD_GET_NODES 0x0001 /* tx net_addr, rx node_info(s) */
76 #define TIPC_CMD_GET_MEDIA_NAMES 0x0002 /* tx none, rx media_name(s) */
77 #define TIPC_CMD_GET_BEARER_NAMES 0x0003 /* tx none, rx bearer_name(s) */
78 #define TIPC_CMD_GET_LINKS 0x0004 /* tx net_addr, rx link_info(s) */
79 #define TIPC_CMD_SHOW_NAME_TABLE 0x0005 /* tx name_tbl_query, rx ultra_string */
80 #define TIPC_CMD_SHOW_PORTS 0x0006 /* tx none, rx ultra_string */
81 #define TIPC_CMD_SHOW_LINK_STATS 0x000B /* tx link_name, rx ultra_string */
82 #define TIPC_CMD_SHOW_STATS 0x000F /* tx unsigned, rx ultra_string */
91 #define TIPC_CMD_GET_REMOTE_MNG 0x4003 /* tx none, rx unsigned */
92 #define TIPC_CMD_GET_MAX_PORTS 0x4004 /* tx none, rx unsigned */
99 #define TIPC_CMD_GET_NETID 0x400B /* tx none, rx unsigned */
101 #define TIPC_CMD_ENABLE_BEARER 0x4101 /* tx bearer_config, rx none */
102 #define TIPC_CMD_DISABLE_BEARER 0x4102 /* tx bearer_name, rx none */
103 #define TIPC_CMD_SET_LINK_TOL 0x4107 /* tx link_config, rx none */
104 #define TIPC_CMD_SET_LINK_PRI 0x4108 /* tx link_config, rx none */
105 #define TIPC_CMD_SET_LINK_WINDOW 0x4109 /* tx link_config, rx none */
108 #define TIPC_CMD_RESET_LINK_STATS 0x410C /* tx link_name, rx none */
116 #define TIPC_CMD_SET_NODE_ADDR 0x8001 /* tx net_addr, rx none */
117 #define TIPC_CMD_SET_REMOTE_MNG 0x8003 /* tx unsigned, rx none */
118 #define TIPC_CMD_SET_MAX_PORTS 0x8004 /* tx unsigned, rx none */
125 #define TIPC_CMD_SET_NETID 0x800B /* tx unsigned, rx none */
133 #define TIPC_CMD_NOT_NET_ADMIN 0xC001 /* tx none, rx none */
/linux-4.4.14/arch/mips/include/asm/mach-bcm63xx/
H A Dbcm63xx_iudma.h7 * rx/tx dma descriptor
/linux-4.4.14/arch/arm/mach-omap1/
H A Dmcbsp.c105 .name = "rx",
115 .name = "rx",
132 .name = "rx",
142 .name = "rx",
182 .name = "rx",
192 .name = "rx",
209 .name = "rx",
219 .name = "rx",
236 .name = "rx",
246 .name = "rx",
289 .name = "rx",
299 .name = "rx",
316 .name = "rx",
326 .name = "rx",
343 .name = "rx",
353 .name = "rx",
/linux-4.4.14/drivers/spi/
H A Dspi-dln2.c194 } rx; dln2_spi_get_cs_num() local
195 unsigned rx_len = sizeof(rx); dln2_spi_get_cs_num()
199 &rx, &rx_len); dln2_spi_get_cs_num()
202 if (rx_len < sizeof(rx)) dln2_spi_get_cs_num()
205 *cs_num = le16_to_cpu(rx.cs_count); dln2_spi_get_cs_num()
220 } rx; dln2_spi_get_speed() local
221 unsigned rx_len = sizeof(rx); dln2_spi_get_speed()
225 ret = dln2_transfer(dln2->pdev, cmd, &tx, sizeof(tx), &rx, &rx_len); dln2_spi_get_speed()
228 if (rx_len < sizeof(rx)) dln2_spi_get_speed()
231 *freq = le32_to_cpu(rx.speed); dln2_spi_get_speed()
270 } rx; dln2_spi_set_speed() local
271 int rx_len = sizeof(rx); dln2_spi_set_speed()
277 &rx, &rx_len); dln2_spi_set_speed()
280 if (rx_len < sizeof(rx)) dln2_spi_set_speed()
329 } *rx = dln2->buf; dln2_spi_get_supported_frame_sizes() local
330 unsigned rx_len = sizeof(*rx); dln2_spi_get_supported_frame_sizes()
336 &tx, sizeof(tx), rx, &rx_len); dln2_spi_get_supported_frame_sizes()
339 if (rx_len < sizeof(*rx)) dln2_spi_get_supported_frame_sizes()
341 if (rx->count > ARRAY_SIZE(rx->frame_sizes)) dln2_spi_get_supported_frame_sizes()
345 for (i = 0; i < rx->count; i++) dln2_spi_get_supported_frame_sizes()
346 *bpw_mask |= BIT(rx->frame_sizes[i] - 1); dln2_spi_get_supported_frame_sizes()
462 } __packed *rx = dln2->buf; dln2_spi_read_one() local
463 unsigned rx_len = sizeof(*rx); dln2_spi_read_one()
465 BUILD_BUG_ON(sizeof(*rx) > DLN2_SPI_BUF_SIZE); dln2_spi_read_one()
475 rx, &rx_len); dln2_spi_read_one()
478 if (rx_len < sizeof(rx->size) + data_len) dln2_spi_read_one()
480 if (le16_to_cpu(rx->size) != data_len) dln2_spi_read_one()
483 dln2_spi_copy_from_buf(data, rx->buf, data_len, dln2->bpw); dln2_spi_read_one()
504 } __packed *rx; dln2_spi_read_write_one() local
508 sizeof(*rx) > DLN2_SPI_BUF_SIZE); dln2_spi_read_write_one()
515 * safe to use the same buffer for both tx and rx. When DLN2 sends the dln2_spi_read_write_one()
516 * response back, with the rx data, we don't need the tx buffer anymore. dln2_spi_read_write_one()
519 rx = dln2->buf; dln2_spi_read_write_one()
528 rx_len = sizeof(*rx); dln2_spi_read_write_one()
531 rx, &rx_len); dln2_spi_read_write_one()
534 if (rx_len < sizeof(rx->size) + data_len) dln2_spi_read_write_one()
536 if (le16_to_cpu(rx->size) != data_len) dln2_spi_read_write_one()
539 dln2_spi_copy_from_buf(rx_data, rx->buf, data_len, dln2->bpw); dln2_spi_read_write_one()
H A Dspi-adi-v3.c74 void *rx; member in struct:adi_spi_master
106 u32 tx_dummy_val; /* tx value for rx only transfer */
228 /* discard invalid rx data and empty rfifo */ dummy_read()
251 while (drv_data->rx < drv_data->rx_end) { adi_spi_u8_read()
255 *(u8 *)(drv_data->rx++) = ioread32(&drv_data->regs->rfifo); adi_spi_u8_read()
262 while (drv_data->rx < drv_data->rx_end) { adi_spi_u8_duplex()
266 *(u8 *)(drv_data->rx++) = ioread32(&drv_data->regs->rfifo); adi_spi_u8_duplex()
293 while (drv_data->rx < drv_data->rx_end) { adi_spi_u16_read()
297 *(u16 *)drv_data->rx = ioread32(&drv_data->regs->rfifo); adi_spi_u16_read()
298 drv_data->rx += 2; adi_spi_u16_read()
305 while (drv_data->rx < drv_data->rx_end) { adi_spi_u16_duplex()
310 *(u16 *)drv_data->rx = ioread32(&drv_data->regs->rfifo); adi_spi_u16_duplex()
311 drv_data->rx += 2; adi_spi_u16_duplex()
338 while (drv_data->rx < drv_data->rx_end) { adi_spi_u32_read()
342 *(u32 *)drv_data->rx = ioread32(&drv_data->regs->rfifo); adi_spi_u32_read()
343 drv_data->rx += 4; adi_spi_u32_read()
350 while (drv_data->rx < drv_data->rx_end) { adi_spi_u32_duplex()
355 *(u32 *)drv_data->rx = ioread32(&drv_data->regs->rfifo); adi_spi_u32_duplex()
356 drv_data->rx += 4; adi_spi_u32_duplex()
405 drv->rx = t->rx_buf; adi_spi_setup_transfer()
406 drv->rx_end = drv->rx + t->len; adi_spi_setup_transfer()
408 drv->rx = NULL; adi_spi_setup_transfer()
466 if (!drv_data->rx) { adi_spi_dma_xfer()
476 rx_buf = drv_data->rx; adi_spi_dma_xfer()
483 rx_buf = drv_data->rx; adi_spi_dma_xfer()
534 if (!drv_data->rx) { adi_spi_pio_xfer()
542 if (drv_data->rx != drv_data->rx_end) adi_spi_pio_xfer()
783 "spi rx dma error: %d\n", dma_stat); adi_spi_rx_dma_isr()
789 "dma interrupt missing: tx=%d,rx=%d\n", adi_spi_rx_dma_isr()
826 dev_err(dev, "can not get rx dma resource\n"); adi_spi_probe()
864 /* request tx and rx dma */ adi_spi_probe()
H A Dspi-fsl-lib.c36 type *rx = mpc8xxx_spi->rx; \
37 *rx++ = (type)(data >> mpc8xxx_spi->rx_shift); \
38 mpc8xxx_spi->rx = rx; \
H A Dspi-altera.c61 unsigned char *rx; member in struct:altera_spi
124 hw->rx = t->rx_buf; altera_spi_txrx()
153 if (hw->rx) { altera_spi_txrx()
156 hw->rx[hw->count] = rxd; altera_spi_txrx()
159 hw->rx[hw->count * 2] = rxd; altera_spi_txrx()
160 hw->rx[hw->count * 2 + 1] = rxd >> 8; altera_spi_txrx()
178 if (hw->rx) { altera_spi_irq()
181 hw->rx[hw->count] = rxd; altera_spi_irq()
184 hw->rx[hw->count * 2] = rxd; altera_spi_irq()
185 hw->rx[hw->count * 2 + 1] = rxd >> 8; altera_spi_irq()
H A Dspi-dw.c157 * Another concern is about the tx/rx mismatch, we tx_max()
160 * data which is out of tx/rx fifo and inside the tx_max()
164 rxtx_gap = ((dws->rx_end - dws->rx) - (dws->tx_end - dws->tx)) tx_max()
170 /* Return the max entries we should read out of rx fifo */ rx_max()
173 u32 rx_left = (dws->rx_end - dws->rx) / dws->n_bytes; rx_max()
203 /* Care rx only if the transfer's original "rx" is not null */ dw_reader()
206 *(u8 *)(dws->rx) = rxw; dw_reader()
208 *(u16 *)(dws->rx) = rxw; dw_reader()
210 dws->rx += dws->n_bytes; dw_reader()
235 if (dws->rx_end == dws->rx) { interrupt_transfer()
274 } while (dws->rx_end > dws->rx); poll_transfer()
294 dws->rx = transfer->rx_buf; dw_spi_transfer_one()
295 dws->rx_end = dws->rx + transfer->len; dw_spi_transfer_one()
330 if (dws->rx && dws->tx) dw_spi_transfer_one()
332 else if (dws->rx) dw_spi_transfer_one()
448 /* Restart the controller, disable all interrupts, clean rx fifo */ spi_hw_init()
H A Dspi-omap2-mcspi.c566 u8 *rx; omap2_mcspi_txrx_dma() local
610 rx = xfer->rx_buf; omap2_mcspi_txrx_dma()
616 if (rx != NULL) omap2_mcspi_txrx_dma()
636 if (rx == NULL) { omap2_mcspi_txrx_dma()
688 u8 *rx; omap2_mcspi_txrx_pio() local
691 rx = xfer->rx_buf; omap2_mcspi_txrx_pio()
706 if (rx != NULL) { omap2_mcspi_txrx_pio()
716 *rx++ = readl_relaxed(rx_reg); omap2_mcspi_txrx_pio()
718 word_len, *(rx - 1)); omap2_mcspi_txrx_pio()
730 *rx++ = readl_relaxed(rx_reg); omap2_mcspi_txrx_pio()
732 word_len, *(rx - 1)); omap2_mcspi_txrx_pio()
736 u16 *rx; omap2_mcspi_txrx_pio() local
739 rx = xfer->rx_buf; omap2_mcspi_txrx_pio()
753 if (rx != NULL) { omap2_mcspi_txrx_pio()
763 *rx++ = readl_relaxed(rx_reg); omap2_mcspi_txrx_pio()
765 word_len, *(rx - 1)); omap2_mcspi_txrx_pio()
777 *rx++ = readl_relaxed(rx_reg); omap2_mcspi_txrx_pio()
779 word_len, *(rx - 1)); omap2_mcspi_txrx_pio()
783 u32 *rx; omap2_mcspi_txrx_pio() local
786 rx = xfer->rx_buf; omap2_mcspi_txrx_pio()
800 if (rx != NULL) { omap2_mcspi_txrx_pio()
810 *rx++ = readl_relaxed(rx_reg); omap2_mcspi_txrx_pio()
812 word_len, *(rx - 1)); omap2_mcspi_txrx_pio()
824 *rx++ = readl_relaxed(rx_reg); omap2_mcspi_txrx_pio()
826 word_len, *(rx - 1)); omap2_mcspi_txrx_pio()
840 /* disable chan to purge rx datas received in TX_ONLY transfer, omap2_mcspi_txrx_pio()
841 * otherwise these rx datas will affect the direct following omap2_mcspi_txrx_pio()
959 * for both rx and tx. Otherwise we'll do PIO for both rx and tx.
1265 rx_buf ? "rx" : "", omap2_mcspi_transfer_one()
1440 sprintf(dma_rx_ch_name, "rx%d", i); omap2_mcspi_probe()
H A Dspi-bfin5xx.c89 void *rx; member in struct:bfin_spi_master_data
221 /* used to kick off transfer in rx mode and read unwanted RX data */ bfin_spi_dummy_read()
250 while (drv_data->rx < drv_data->rx_end) { bfin_spi_u8_reader()
254 *(u8 *) (drv_data->rx++) = bfin_read(&drv_data->regs->rdbr); bfin_spi_u8_reader()
263 while (drv_data->rx < drv_data->rx_end) { bfin_spi_u8_duplex()
267 *(u8 *) (drv_data->rx++) = bfin_read(&drv_data->regs->rdbr); bfin_spi_u8_duplex()
301 while (drv_data->rx < drv_data->rx_end) { bfin_spi_u16_reader()
305 *(u16 *) (drv_data->rx) = bfin_read(&drv_data->regs->rdbr); bfin_spi_u16_reader()
306 drv_data->rx += 2; bfin_spi_u16_reader()
315 while (drv_data->rx < drv_data->rx_end) { bfin_spi_u16_duplex()
320 *(u16 *) (drv_data->rx) = bfin_read(&drv_data->regs->rdbr); bfin_spi_u16_duplex()
321 drv_data->rx += 2; bfin_spi_u16_duplex()
392 (drv_data->rx && drv_data->rx >= (drv_data->rx_end - n_bytes))) { bfin_spi_pio_irq_handler()
394 if (drv_data->rx) { bfin_spi_pio_irq_handler()
397 u16 *buf = (u16 *)drv_data->rx; bfin_spi_pio_irq_handler()
401 u8 *buf = (u8 *)drv_data->rx; bfin_spi_pio_irq_handler()
405 drv_data->rx += n_bytes; bfin_spi_pio_irq_handler()
421 if (drv_data->rx && drv_data->tx) { bfin_spi_pio_irq_handler()
425 u16 *buf = (u16 *)drv_data->rx; bfin_spi_pio_irq_handler()
432 u8 *buf = (u8 *)drv_data->rx; bfin_spi_pio_irq_handler()
439 } else if (drv_data->rx) { bfin_spi_pio_irq_handler()
443 u16 *buf = (u16 *)drv_data->rx; bfin_spi_pio_irq_handler()
449 u8 *buf = (u8 *)drv_data->rx; bfin_spi_pio_irq_handler()
475 if (drv_data->rx) bfin_spi_pio_irq_handler()
476 drv_data->rx += n_bytes; bfin_spi_pio_irq_handler()
494 if (drv_data->rx != NULL) { bfin_spi_dma_irq_handler()
628 drv_data->rx = transfer->rx_buf; bfin_spi_pump_transfers()
629 drv_data->rx_end = drv_data->rx + transfer->len; bfin_spi_pump_transfers()
633 drv_data->rx = NULL; bfin_spi_pump_transfers()
726 /* In dma mode, rx or tx must be NULL in one transfer */ bfin_spi_pump_transfers()
728 if (drv_data->rx != NULL) { bfin_spi_pump_transfers()
731 drv_data->rx, drv_data->len_in_bytes); bfin_spi_pump_transfers()
734 if (bfin_addr_dcacheable((unsigned long) drv_data->rx)) bfin_spi_pump_transfers()
735 invalidate_dcache_range((unsigned long) drv_data->rx, bfin_spi_pump_transfers()
736 (unsigned long) (drv_data->rx + bfin_spi_pump_transfers()
740 dma_start_addr = (unsigned long)drv_data->rx; bfin_spi_pump_transfers()
824 (drv_data->rx_end - drv_data->rx)); bfin_spi_pump_transfers()
841 } else if (drv_data->rx != NULL) { bfin_spi_pump_transfers()
847 if (drv_data->rx != drv_data->rx_end) bfin_spi_pump_transfers()
H A Dspi-sirf.c151 * only if the rx/tx buffer and transfer size are 4-bytes aligned, we use dma
260 /* rx & tx bufs from the spi_transfer */
262 void *rx; member in struct:sirfsoc_spi
264 /* place received word into rx buffer */
273 /* rx & tx DMA channels */
281 * if tx size is not more than 4 and rx size is NULL, use
316 u8 *rx = sspi->rx; spi_sirfsoc_rx_word_u8() local
320 if (rx) { spi_sirfsoc_rx_word_u8()
321 *rx++ = (u8) data; spi_sirfsoc_rx_word_u8()
322 sspi->rx = rx; spi_sirfsoc_rx_word_u8()
344 u16 *rx = sspi->rx; spi_sirfsoc_rx_word_u16() local
348 if (rx) { spi_sirfsoc_rx_word_u16()
349 *rx++ = (u16) data; spi_sirfsoc_rx_word_u16()
350 sspi->rx = rx; spi_sirfsoc_rx_word_u16()
373 u32 *rx = sspi->rx; spi_sirfsoc_rx_word_u32() local
377 if (rx) { spi_sirfsoc_rx_word_u32()
378 *rx++ = (u32) data; spi_sirfsoc_rx_word_u32()
379 sspi->rx = rx; spi_sirfsoc_rx_word_u32()
548 sspi->dst_start = dma_map_single(&spi->dev, sspi->rx, t->len, spi_sirfsoc_dma_transfer()
586 * we get rx data by writing tx data, so if rx is done, tx has spi_sirfsoc_dma_transfer()
711 sspi->rx = t->rx_buf; spi_sirfsoc_transfer()
1132 sspi->rx_chan = dma_request_slave_channel(&pdev->dev, "rx"); spi_sirfsoc_probe()
1134 dev_err(&pdev->dev, "can not allocate rx dma channel\n"); spi_sirfsoc_probe()
H A Dspi-bitbang.c72 u8 *rx = t->rx_buf; bitbang_txrx_8() local
80 if (rx) bitbang_txrx_8()
81 *rx++ = word; bitbang_txrx_8()
98 u16 *rx = t->rx_buf; bitbang_txrx_16() local
106 if (rx) bitbang_txrx_16()
107 *rx++ = word; bitbang_txrx_16()
124 u32 *rx = t->rx_buf; bitbang_txrx_32() local
132 if (rx) bitbang_txrx_32()
133 *rx++ = word; bitbang_txrx_32()
H A Dspi-omap-100k.c199 u8 *rx; omap1_spi100k_txrx_pio() local
202 rx = xfer->rx_buf; omap1_spi100k_txrx_pio()
209 *rx++ = spi100k_read_data(spi->master, word_len); omap1_spi100k_txrx_pio()
212 u16 *rx; omap1_spi100k_txrx_pio() local
215 rx = xfer->rx_buf; omap1_spi100k_txrx_pio()
222 *rx++ = spi100k_read_data(spi->master, word_len); omap1_spi100k_txrx_pio()
225 u32 *rx; omap1_spi100k_txrx_pio() local
228 rx = xfer->rx_buf; omap1_spi100k_txrx_pio()
235 *rx = spi100k_read_data(spi->master, word_len); omap1_spi100k_txrx_pio()
H A Dspi-rockchip.c188 void *rx; member in struct:rockchip_spi
259 u32 rx_left = (rs->rx_end - rs->rx) / rs->n_bytes; rx_max()
372 *(u8 *)(rs->rx) = (u8)rxw; rockchip_spi_pio_reader()
374 *(u16 *)(rs->rx) = (u16)rxw; rockchip_spi_pio_reader()
375 rs->rx += rs->n_bytes; rockchip_spi_pio_reader()
389 if (rs->rx) { rockchip_spi_pio_transfer()
390 remain = rs->rx_end - rs->rx; rockchip_spi_pio_transfer()
453 if (rs->rx) { rockchip_spi_prepare_dma()
486 /* rx must be started before tx due to spi instinct */ rockchip_spi_prepare_dma()
521 if (rs->rx) rockchip_spi_config()
589 rs->rx = xfer->rx_buf; rockchip_spi_transfer_one()
590 rs->rx_end = rs->rx + xfer->len; rockchip_spi_transfer_one()
596 if (rs->tx && rs->rx) rockchip_spi_transfer_one()
600 else if (rs->rx) rockchip_spi_transfer_one()
613 /* rx: dma must be prepared first */ rockchip_spi_transfer_one()
695 if (!of_property_read_u32(pdev->dev.of_node, "rx-sample-delay-ns", rockchip_spi_probe()
728 rs->dma_rx.ch = dma_request_slave_channel(rs->dev, "rx"); rockchip_spi_probe()
H A Dspi-dw-mid.c50 struct dw_dma_slave *rx = dws->dma_rx; mid_spi_dma_init() local
64 /* 1. Init rx channel */ mid_spi_dma_init()
65 rx->dma_dev = &dma_dev->dev; mid_spi_dma_init()
66 dws->rxchan = dma_request_channel(mask, mid_spi_dma_chan_filter, rx); mid_spi_dma_init()
182 * dws->dma_chan_busy is set before the dma transfer starts, callback for rx
258 /* rx must be started before tx due to spi instinct */ mid_spi_dma_transfer()
H A Dspi-au1550.c59 u8 *rx; member in struct:au1550_spi
281 * for dma spi transfers, we have to setup rx channel, otherwise there is
284 * and if only tx dma channel is set up (and rx fifo overflow event masked)
285 * spi master done event irq is not generated unless rx fifo is empty (emptied)
286 * so we need rx tmp buffer to use for rx dma if user does not provide one
326 hw->rx = t->rx_buf; au1550_spi_dma_txrxb()
335 * use rx buffer in place of tx if tx buffer was not provided au1550_spi_dma_txrxb()
336 * use temp rx buffer (preallocated or realloc to fit) for rx dma au1550_spi_dma_txrxb()
354 dev_err(hw->dev, "rx dma map error\n"); au1550_spi_dma_txrxb()
366 hw->rx = hw->dma_rx_tmpbuf; au1550_spi_dma_txrxb()
375 hw->tx = hw->rx; au1550_spi_dma_txrxb()
379 res = au1xxx_dbdma_put_dest(hw->dma_rx_ch, virt_to_phys(hw->rx), au1550_spi_dma_txrxb()
382 dev_err(hw->dev, "rx dma put dest error\n"); au1550_spi_dma_txrxb()
483 if (hw->rx) { \
484 *(u##size *)hw->rx = (u##size)fifoword; \
485 hw->rx += (size) / 8; \
518 hw->rx = t->rx_buf; au1550_spi_pio_txrxb()
586 * while there is something to read from rx fifo au1550_spi_pio_irq_callback()
839 "Cannot allocate rx dma channel\n"); au1550_spi_probe()
847 "Cannot allocate rx dma descriptors\n"); au1550_spi_probe()
856 "Cannot allocate initial rx dma tmp buffer\n"); au1550_spi_probe()
H A Dspi-rspi.c487 static int rspi_pio_transfer(struct rspi_data *rspi, const u8 *tx, u8 *rx, rspi_pio_transfer() argument
496 if (rx) { rspi_pio_transfer()
500 *rx++ = ret; rspi_pio_transfer()
516 struct sg_table *rx) rspi_dma_transfer()
525 if (rx) { rspi_dma_transfer()
527 rx->sgl, rx->nents, DMA_FROM_DEVICE, rspi_dma_transfer()
554 if (rx) { rspi_dma_transfer()
576 if (rx && rspi->rx_irq != other_irq) rspi_dma_transfer()
583 if (rx) rspi_dma_transfer()
597 if (rx) rspi_dma_transfer()
605 if (rx && rspi->rx_irq != other_irq) rspi_dma_transfer()
611 if (rx) rspi_dma_transfer()
726 u8 *rx, unsigned int len) qspi_trigger_transfer_out_in()
749 *rx++ = rspi_read_data(rspi); qspi_trigger_transfer_out_in()
751 ret = rspi_pio_transfer(rspi, tx, rx, n); qspi_trigger_transfer_out_in()
1013 dir == DMA_MEM_TO_DEV ? "tx" : "rx"); rspi_request_dma_chan()
1238 ret = platform_get_irq_byname(pdev, "rx"); rspi_probe()
1263 "rx", rspi); rspi_probe()
515 rspi_dma_transfer(struct rspi_data *rspi, struct sg_table *tx, struct sg_table *rx) rspi_dma_transfer() argument
725 qspi_trigger_transfer_out_in(struct rspi_data *rspi, const u8 *tx, u8 *rx, unsigned int len) qspi_trigger_transfer_out_in() argument
/linux-4.4.14/drivers/net/wireless/zd1211rw/
H A Dzd_usb.c665 struct zd_usb_rx *rx; rx_urb_complete() local
688 rx = &usb->rx; rx_urb_complete()
690 tasklet_schedule(&rx->reset_timer_tasklet); rx_urb_complete()
692 if (length%rx->usb_packet_size > rx->usb_packet_size-4) { rx_urb_complete()
695 ZD_ASSERT(length <= ARRAY_SIZE(rx->fragment)); rx_urb_complete()
696 spin_lock(&rx->lock); rx_urb_complete()
697 memcpy(rx->fragment, buffer, length); rx_urb_complete()
698 rx->fragment_length = length; rx_urb_complete()
699 spin_unlock(&rx->lock); rx_urb_complete()
703 spin_lock(&rx->lock); rx_urb_complete()
704 if (rx->fragment_length > 0) { rx_urb_complete()
706 ZD_ASSERT(length + rx->fragment_length <= rx_urb_complete()
707 ARRAY_SIZE(rx->fragment)); rx_urb_complete()
709 memcpy(rx->fragment+rx->fragment_length, buffer, length); rx_urb_complete()
710 handle_rx_packet(usb, rx->fragment, rx_urb_complete()
711 rx->fragment_length + length); rx_urb_complete()
712 rx->fragment_length = 0; rx_urb_complete()
713 spin_unlock(&rx->lock); rx_urb_complete()
715 spin_unlock(&rx->lock); rx_urb_complete()
761 struct zd_usb_rx *rx = &usb->rx; __zd_usb_enable_rx() local
777 spin_lock_irq(&rx->lock); __zd_usb_enable_rx()
778 if (rx->urbs) { __zd_usb_enable_rx()
779 spin_unlock_irq(&rx->lock); __zd_usb_enable_rx()
783 rx->urbs = urbs; __zd_usb_enable_rx()
784 rx->urbs_count = RX_URBS_COUNT; __zd_usb_enable_rx()
785 spin_unlock_irq(&rx->lock); __zd_usb_enable_rx()
798 spin_lock_irq(&rx->lock); __zd_usb_enable_rx()
799 rx->urbs = NULL; __zd_usb_enable_rx()
800 rx->urbs_count = 0; __zd_usb_enable_rx()
801 spin_unlock_irq(&rx->lock); __zd_usb_enable_rx()
813 struct zd_usb_rx *rx = &usb->rx; zd_usb_enable_rx() local
815 mutex_lock(&rx->setup_mutex); zd_usb_enable_rx()
817 mutex_unlock(&rx->setup_mutex); zd_usb_enable_rx()
830 struct zd_usb_rx *rx = &usb->rx; __zd_usb_disable_rx() local
832 spin_lock_irqsave(&rx->lock, flags); __zd_usb_disable_rx()
833 urbs = rx->urbs; __zd_usb_disable_rx()
834 count = rx->urbs_count; __zd_usb_disable_rx()
835 spin_unlock_irqrestore(&rx->lock, flags); __zd_usb_disable_rx()
845 spin_lock_irqsave(&rx->lock, flags); __zd_usb_disable_rx()
846 rx->urbs = NULL; __zd_usb_disable_rx()
847 rx->urbs_count = 0; __zd_usb_disable_rx()
848 spin_unlock_irqrestore(&rx->lock, flags); __zd_usb_disable_rx()
853 struct zd_usb_rx *rx = &usb->rx; zd_usb_disable_rx() local
855 mutex_lock(&rx->setup_mutex); zd_usb_disable_rx()
857 mutex_unlock(&rx->setup_mutex); zd_usb_disable_rx()
859 tasklet_kill(&rx->reset_timer_tasklet); zd_usb_disable_rx()
860 cancel_delayed_work_sync(&rx->idle_work); zd_usb_disable_rx()
866 struct zd_usb_rx *rx = &usb->rx; zd_usb_reset_rx() local
869 mutex_lock(&rx->setup_mutex); zd_usb_reset_rx()
871 spin_lock_irqsave(&rx->lock, flags); zd_usb_reset_rx()
872 do_reset = rx->urbs != NULL; zd_usb_reset_rx()
873 spin_unlock_irqrestore(&rx->lock, flags); zd_usb_reset_rx()
880 mutex_unlock(&rx->setup_mutex); zd_usb_reset_rx()
1142 container_of(work, struct zd_usb, rx.idle_work.work); zd_rx_idle_timer_handler()
1150 /* 30 seconds since last rx, reset rx */ zd_rx_idle_timer_handler()
1163 struct zd_usb_rx *rx = &usb->rx; zd_usb_reset_rx_idle_timer() local
1165 mod_delayed_work(zd_workqueue, &rx->idle_work, ZD_RX_IDLE_INTERVAL); zd_usb_reset_rx_idle_timer()
1181 struct zd_usb_rx *rx = &usb->rx; init_usb_rx() local
1183 spin_lock_init(&rx->lock); init_usb_rx()
1184 mutex_init(&rx->setup_mutex); init_usb_rx()
1186 rx->usb_packet_size = 512; init_usb_rx()
1188 rx->usb_packet_size = 64; init_usb_rx()
1190 ZD_ASSERT(rx->fragment_length == 0); init_usb_rx()
1191 INIT_DELAYED_WORK(&rx->idle_work, zd_rx_idle_timer_handler); init_usb_rx()
1192 rx->reset_timer_tasklet.func = zd_usb_reset_rx_idle_timer_tasklet; init_usb_rx()
1193 rx->reset_timer_tasklet.data = (unsigned long)usb; init_usb_rx()
/linux-4.4.14/net/nfc/nci/
H A Dspi.c190 struct spi_transfer tx, rx; __nci_spi_read() local
205 memset(&rx, 0, sizeof(struct spi_transfer)); __nci_spi_read()
206 rx.rx_buf = resp_hdr; __nci_spi_read()
207 rx.len = 2; __nci_spi_read()
208 rx.cs_change = 1; __nci_spi_read()
209 rx.speed_hz = nspi->xfer_speed_hz; __nci_spi_read()
210 spi_message_add_tail(&rx, &m); __nci_spi_read()
228 memset(&rx, 0, sizeof(struct spi_transfer)); __nci_spi_read()
229 rx.rx_buf = skb_put(skb, rx_len); __nci_spi_read()
230 rx.len = rx_len; __nci_spi_read()
231 rx.cs_change = 0; __nci_spi_read()
232 rx.delay_usecs = nspi->xfer_udelay; __nci_spi_read()
233 rx.speed_hz = nspi->xfer_speed_hz; __nci_spi_read()
234 spi_message_add_tail(&rx, &m); __nci_spi_read()
/linux-4.4.14/drivers/staging/iio/resolver/
H A Dad2s1200.c41 u8 rx[2] ____cacheline_aligned;
60 ret = spi_read(st->sdev, st->rx, 2); ad2s1200_read_raw()
68 *val = (((u16)(st->rx[0])) << 4) | ((st->rx[1] & 0xF0) >> 4); ad2s1200_read_raw()
71 vel = (((s16)(st->rx[0])) << 4) | ((st->rx[1] & 0xF0) >> 4); ad2s1200_read_raw()
H A Dad2s90.c25 u8 rx[2] ____cacheline_aligned;
38 ret = spi_read(st->sdev, st->rx, 2); ad2s90_read_raw()
41 *val = (((u16)(st->rx[0])) << 4) | ((st->rx[1] & 0xF0) >> 4); ad2s90_read_raw()
/linux-4.4.14/drivers/media/pci/cobalt/
H A Dcobalt-v4l2.c266 int rx = s->video_channel; cobalt_dma_start_streaming() local
268 COBALT_CVI_EVCNT(cobalt, rx); cobalt_dma_start_streaming()
300 int rx = s->video_channel; cobalt_start_streaming() local
301 struct m00389_cvi_regmap __iomem *cvi = COBALT_CVI(cobalt, rx); cobalt_start_streaming()
302 struct m00460_evcnt_regmap __iomem *evcnt = COBALT_CVI_EVCNT(cobalt, rx); cobalt_start_streaming()
317 fw = COBALT_CVI_FREEWHEEL(cobalt, rx); cobalt_start_streaming()
318 vmr = COBALT_CVI_VMR(cobalt, rx); cobalt_start_streaming()
319 clkloss = COBALT_CVI_CLK_LOSS(cobalt, rx); cobalt_start_streaming()
368 int rx = s->video_channel; cobalt_dma_stop_streaming() local
370 COBALT_CVI_EVCNT(cobalt, rx); cobalt_dma_stop_streaming()
406 int rx = s->video_channel; cobalt_stop_streaming() local
428 fw = COBALT_CVI_FREEWHEEL(cobalt, rx); cobalt_stop_streaming()
429 vmr = COBALT_CVI_VMR(cobalt, rx); cobalt_stop_streaming()
430 clkloss = COBALT_CVI_CLK_LOSS(cobalt, rx); cobalt_stop_streaming()
519 int rx = s->video_channel; cobalt_video_input_status_show() local
524 cvi = COBALT_CVI(cobalt, rx); cobalt_video_input_status_show()
525 vmr = COBALT_CVI_VMR(cobalt, rx); cobalt_video_input_status_show()
526 fw = COBALT_CVI_FREEWHEEL(cobalt, rx); cobalt_video_input_status_show()
527 clkloss = COBALT_CVI_CLK_LOSS(cobalt, rx); cobalt_video_input_status_show()
528 packer = COBALT_CVI_PACKER(cobalt, rx); cobalt_video_input_status_show()
533 cobalt_info("rx%d: cvi resolution: %dx%d\n", rx, cobalt_video_input_status_show()
535 cobalt_info("rx%d: cvi control: %s%s%s\n", rx, cobalt_video_input_status_show()
542 cobalt_info("rx%d: cvi status: %s%s\n", rx, cobalt_video_input_status_show()
548 cobalt_info("rx%d: Measurements: %s%s%s%s%s%s%s\n", rx, cobalt_video_input_status_show()
563 cobalt_info("rx%d: irq_status: 0x%02x irq_triggers: 0x%02x\n", rx, cobalt_video_input_status_show()
566 cobalt_info("rx%d: vsync: %d\n", rx, ioread32(&vmr->vsync_time)); cobalt_video_input_status_show()
567 cobalt_info("rx%d: vbp: %d\n", rx, ioread32(&vmr->vback_porch)); cobalt_video_input_status_show()
568 cobalt_info("rx%d: vact: %d\n", rx, ioread32(&vmr->vactive_area)); cobalt_video_input_status_show()
569 cobalt_info("rx%d: vfb: %d\n", rx, ioread32(&vmr->vfront_porch)); cobalt_video_input_status_show()
570 cobalt_info("rx%d: hsync: %d\n", rx, ioread32(&vmr->hsync_time)); cobalt_video_input_status_show()
571 cobalt_info("rx%d: hbp: %d\n", rx, ioread32(&vmr->hback_porch)); cobalt_video_input_status_show()
572 cobalt_info("rx%d: hact: %d\n", rx, ioread32(&vmr->hactive_area)); cobalt_video_input_status_show()
573 cobalt_info("rx%d: hfb: %d\n", rx, ioread32(&vmr->hfront_porch)); cobalt_video_input_status_show()
574 cobalt_info("rx%d: Freewheeling: %s%s%s\n", rx, cobalt_video_input_status_show()
582 cobalt_info("rx%d: Clock Loss Detection: %s%s\n", rx, cobalt_video_input_status_show()
587 cobalt_info("rx%d: Packer: %x\n", rx, ioread32(&packer->control)); cobalt_video_input_status_show()
H A Dcobalt-irq.c30 int rx = s->video_channel; cobalt_dma_stream_queue_handler() local
32 COBALT_CVI_FREEWHEEL(s->cobalt, rx); cobalt_dma_stream_queue_handler()
34 COBALT_CVI_VMR(s->cobalt, rx); cobalt_dma_stream_queue_handler()
36 COBALT_CVI(s->cobalt, rx); cobalt_dma_stream_queue_handler()
38 COBALT_CVI_CLK_LOSS(s->cobalt, rx); cobalt_dma_stream_queue_handler()
180 cobalt_info("full rx FIFO %d\n", i); cobalt_irq_handler()
/linux-4.4.14/sound/firewire/dice/
H A Ddice-proc.c61 "global", "tx", "rx", "ext_sync", "unused2" dice_proc_read()
108 } rx; dice_proc_read() member in union:__anon15218
199 quadlets = min_t(u32, tx_rx_header.size, sizeof(buf.rx) / 4); dice_proc_read()
201 if (dice_proc_read_mem(dice, &buf.rx, sections[4] + 2 + dice_proc_read()
205 snd_iprintf(buffer, "rx %u:\n", stream); dice_proc_read()
206 snd_iprintf(buffer, " iso channel: %d\n", (int)buf.rx.iso); dice_proc_read()
207 snd_iprintf(buffer, " sequence start: %u\n", buf.rx.seq_start); dice_proc_read()
209 buf.rx.number_audio); dice_proc_read()
210 snd_iprintf(buffer, " midi ports: %u\n", buf.rx.number_midi); dice_proc_read()
212 dice_proc_fixup_string(buf.rx.names, RX_NAMES_SIZE); dice_proc_read()
213 snd_iprintf(buffer, " names: %s\n", buf.rx.names); dice_proc_read()
217 buf.rx.ac3_caps); dice_proc_read()
219 buf.rx.ac3_enable); dice_proc_read()
/linux-4.4.14/drivers/ipack/devices/
H A Dipoctal.h27 * @rx: Number of received bytes
35 unsigned long rx; member in struct:ipoctal_stats
/linux-4.4.14/arch/arm/mach-omap2/
H A Domap_hwmod_2xxx_3xxx_ipblock_data.c51 { .name = "rx", .dma_req = 50, },
57 { .name = "rx", .dma_req = 52, },
63 { .name = "rx", .dma_req = 54, },
70 { .name = "rx", .dma_req = 28 },
76 { .name = "rx", .dma_req = 30 },
101 { .name = "rx", .dma_req = 32 },
107 { .name = "rx", .dma_req = 34 },
113 { .name = "rx", .dma_req = 18 },
/linux-4.4.14/drivers/staging/octeon/
H A DMakefile18 octeon-ethernet-y += ethernet-rx.o
/linux-4.4.14/drivers/net/wireless/iwlwifi/
H A DMakefile9 iwlwifi-objs += pcie/drv.o pcie/rx.o pcie/tx.o pcie/trans.o
/linux-4.4.14/include/linux/
H A Dserial.h21 __u32 cts, dsr, rng, dcd, tx, rx; member in struct:async_icount
H A Dtimb_dma.h29 * @rx: true if this channel handles data in the direction to
38 bool rx; member in struct:timb_dma_platform_data_channel
/linux-4.4.14/drivers/tty/serial/
H A Dmax3100.c116 int rx_enabled; /* if we should rx chars */
196 static int max3100_sr(struct max3100_port *s, u16 tx, u16 *rx) max3100_sr() argument
215 *rx = be16_to_cpu(erx); max3100_sr()
216 s->tx_empty = (*rx & MAX3100_T) > 0; max3100_sr()
217 dev_dbg(&s->spi->dev, "%04x - %04x\n", tx, *rx); max3100_sr()
221 static int max3100_handlerx(struct max3100_port *s, u16 rx) max3100_handlerx() argument
226 if (rx & MAX3100_R && s->rx_enabled) { max3100_handlerx()
228 ch = rx & (s->parity & MAX3100_7BIT ? 0x7f : 0xff); max3100_handlerx()
229 if (rx & MAX3100_RAFE) { max3100_handlerx()
235 if (max3100_check_parity(s, rx)) { max3100_handlerx()
236 s->port.icount.rx++; max3100_handlerx()
244 s->port.icount.rx++; max3100_handlerx()
252 cts = (rx & MAX3100_CTS) > 0; max3100_handlerx()
265 u16 tx, rx; max3100_work() local
282 max3100_sr(s, MAX3100_WC | conf, &rx); max3100_work()
285 (s->rts ? MAX3100_RTS : 0), &rx); max3100_work()
286 rxchars += max3100_handlerx(s, rx); max3100_work()
289 max3100_sr(s, MAX3100_RD, &rx); max3100_work()
290 rxchars += max3100_handlerx(s, rx); max3100_work()
292 if (rx & MAX3100_T) { max3100_work()
308 max3100_sr(s, tx, &rx); max3100_work()
309 rxchars += max3100_handlerx(s, rx); max3100_work()
322 ((rx & MAX3100_R) || max3100_work()
578 u16 tx, rx; max3100_shutdown() local
581 max3100_sr(s, tx, &rx); max3100_shutdown()
623 u16 tx, rx; max3100_startup() local
625 max3100_sr(s, tx, &rx); max3100_startup()
748 u16 tx, rx; max3100_probe() local
814 max3100_sr(max3100s[i], tx, &rx); max3100_probe()
869 u16 tx, rx; max3100_suspend() local
872 max3100_sr(s, tx, &rx); max3100_suspend()
H A Dbcm63xx_uart.c41 * rx interrupt mask / stat
44 * - rx fifo full
45 * - rx fifo above threshold
46 * - rx fifo not empty for too long
180 * serial core request to stop rx, called before port shutdown
232 * read all chars in rx fifo and send them to core
268 port->icount.rx++; bcm_uart_do_rx()
399 * enable rx & tx operation on uart
411 * disable rx & tx operation on uart
424 * clear all unread data in rx fifo and unsent data in tx fifo
430 /* empty rx and tx fifo */ bcm_uart_flush()
441 * serial core request to initialize uart and start rx operation
456 /* set rx/tx fifo thresh to fifo half size */ bcm_uart_startup()
463 /* set rx fifo timeout to 1 char time */ bcm_uart_startup()
475 /* register irq and enable rx interrupts */ bcm_uart_startup()
/linux-4.4.14/drivers/staging/iio/accel/
H A Dsca3000_ring.c86 u8 *rx; sca3000_read_first_n_hw_rb() local
103 num_available = st->rx[0]; sca3000_read_first_n_hw_rb()
115 &rx, num_read); sca3000_read_first_n_hw_rb()
120 *(((u16 *)rx) + i) = be16_to_cpup((__be16 *)rx + i); sca3000_read_first_n_hw_rb()
122 if (copy_to_user(buf, rx, num_read)) sca3000_read_first_n_hw_rb()
124 kfree(rx); sca3000_read_first_n_hw_rb()
151 val = st->rx[0]; sca3000_query_ring_int()
183 st->rx[0] | this_attr->address); sca3000_set_ring_int()
187 st->rx[0] & ~this_attr->address); sca3000_set_ring_int()
300 (st->rx[0] | SCA3000_RING_BUF_ENABLE)); __sca3000_hw_ring_state_set()
304 (st->rx[0] & ~SCA3000_RING_BUF_ENABLE)); __sca3000_hw_ring_state_set()
H A Dsca3000_core.c100 .rx_buf = st->rx, sca3000_read_data_short()
121 return !(st->rx[0] & SCA3000_LOCKED); sca3000_reg_lock_on()
220 return st->rx[0]; sca3000_read_ctrl_reg()
242 st->rx[0] & SCA3000_REVID_MAJOR_MASK, sca3000_show_rev()
243 st->rx[0] & SCA3000_REVID_MINOR_MASK); sca3000_show_rev()
302 st->rx[0] &= 0x03; sca3000_show_measurement_mode()
303 switch (st->rx[0]) { sca3000_show_measurement_mode()
361 st->rx[0] &= ~mask; sca3000_store_measurement_mode()
362 st->rx[0] |= (val & mask); sca3000_store_measurement_mode()
363 ret = sca3000_write_reg(st, SCA3000_REG_ADDR_MODE, st->rx[0]); sca3000_store_measurement_mode()
471 *val = (be16_to_cpup((__be16 *)st->rx) >> 3) & 0x1FFF; sca3000_read_raw()
483 *val = ((st->rx[0] & 0x3F) << 3) | sca3000_read_raw()
484 ((st->rx[1] & 0xE0) >> 5); sca3000_read_raw()
522 val = st->rx[0]; sca3000_read_av_freq()
566 switch (0x03 & st->rx[0]) { __sca3000_get_base_freq()
786 val = st->rx[0]; sca3000_event_handler()
852 if ((st->rx[0] & protect_mask) != SCA3000_MEAS_MODE_MOT_DET) { sca3000_read_event_config()
881 val = st->rx[0]; sca3000_query_free_fall_mode()
917 if (val && !(st->rx[0] & protect_mask)) sca3000_set_free_fall_mode()
919 (st->rx[0] | SCA3000_FREE_FALL_DETECT)); sca3000_set_free_fall_mode()
921 else if (!val && (st->rx[0] & protect_mask)) sca3000_set_free_fall_mode()
923 (st->rx[0] & ~protect_mask)); sca3000_set_free_fall_mode()
984 ((st->rx[0] & protect_mask) != SCA3000_MEAS_MODE_MOT_DET)) sca3000_write_event_config()
986 (st->rx[0] & ~protect_mask) sca3000_write_event_config()
990 ((st->rx[0] & protect_mask) == SCA3000_MEAS_MODE_MOT_DET)) sca3000_write_event_config()
992 (st->rx[0] & ~protect_mask)); sca3000_write_event_config()
1077 (st->rx[0] & SCA3000_MODE_PROT_MASK)); sca3000_clean_setup()
1164 (st->rx[0] & sca3000_stop_all_interrupts()
H A Dadis16220.h129 * @rx: receive buffer
130 * @buf_lock: mutex to protect tx and rx
137 u8 rx[ADIS16220_MAX_RX]; member in struct:adis16220_state
/linux-4.4.14/drivers/net/
H A Dxen-netfront.c86 /* IRQ name is queue name with "-tx" or "-rx" appended */
111 char rx_irq_name[IRQ_NAME_SIZE]; /* DEVNAME-qN-rx */
118 * {tx,rx}_skbs store outstanding skbuffs. Free tx_skb entries
136 struct xen_netif_rx_front_ring rx; member in struct:netfront_queue
163 struct xen_netif_rx_response rx; member in struct:netfront_rx_info
282 RING_IDX req_prod = queue->rx.req_prod_pvt; xennet_alloc_rx_buffers()
288 for (req_prod = queue->rx.req_prod_pvt; xennet_alloc_rx_buffers()
289 req_prod - queue->rx.rsp_cons < NET_RX_RING_SIZE; xennet_alloc_rx_buffers()
312 req = RING_GET_REQUEST(&queue->rx, req_prod); xennet_alloc_rx_buffers()
321 queue->rx.req_prod_pvt = req_prod; xennet_alloc_rx_buffers()
324 if (req_prod - queue->rx.rsp_cons < NET_RX_SLOTS_MIN) { xennet_alloc_rx_buffers()
331 RING_PUSH_REQUESTS_AND_CHECK_NOTIFY(&queue->rx, notify); xennet_alloc_rx_buffers()
350 queue->rx.sring->rsp_event = queue->rx.rsp_cons + 1; xennet_open()
351 if (RING_HAS_UNCONSUMED_RESPONSES(&queue->rx)) xennet_open()
709 int new = xennet_rxidx(queue->rx.req_prod_pvt); xennet_move_rx_slot()
714 RING_GET_REQUEST(&queue->rx, queue->rx.req_prod_pvt)->id = new; xennet_move_rx_slot()
715 RING_GET_REQUEST(&queue->rx, queue->rx.req_prod_pvt)->gref = ref; xennet_move_rx_slot()
716 queue->rx.req_prod_pvt++; xennet_move_rx_slot()
726 RING_IDX cons = queue->rx.rsp_cons; xennet_get_extras()
741 RING_GET_RESPONSE(&queue->rx, ++cons); xennet_get_extras()
759 queue->rx.rsp_cons = cons; xennet_get_extras()
767 struct xen_netif_rx_response *rx = &rinfo->rx; xennet_get_responses() local
770 RING_IDX cons = queue->rx.rsp_cons; xennet_get_responses()
773 int max = MAX_SKB_FRAGS + (rx->status <= RX_COPY_THRESHOLD); xennet_get_responses()
778 if (rx->flags & XEN_NETRXF_extra_info) { xennet_get_responses()
780 cons = queue->rx.rsp_cons; xennet_get_responses()
784 if (unlikely(rx->status < 0 || xennet_get_responses()
785 rx->offset + rx->status > XEN_PAGE_SIZE)) { xennet_get_responses()
787 dev_warn(dev, "rx->offset: %u, size: %d\n", xennet_get_responses()
788 rx->offset, rx->status); xennet_get_responses()
801 dev_warn(dev, "Bad rx response id %d.\n", xennet_get_responses()
802 rx->id); xennet_get_responses()
815 if (!(rx->flags & XEN_NETRXF_more_data)) xennet_get_responses()
825 rx = RING_GET_RESPONSE(&queue->rx, cons + slots); xennet_get_responses()
838 queue->rx.rsp_cons = cons + slots; xennet_get_responses()
877 RING_IDX cons = queue->rx.rsp_cons; xennet_fill_frags()
881 struct xen_netif_rx_response *rx = xennet_fill_frags() local
882 RING_GET_RESPONSE(&queue->rx, ++cons); xennet_fill_frags()
894 rx->offset, rx->status, PAGE_SIZE); xennet_fill_frags()
969 struct xen_netif_rx_response *rx = &rinfo.rx; xennet_poll() local
984 rp = queue->rx.sring->rsp_prod; xennet_poll()
987 i = queue->rx.rsp_cons; xennet_poll()
990 memcpy(rx, RING_GET_RESPONSE(&queue->rx, i), sizeof(*rx)); xennet_poll()
1000 i = queue->rx.rsp_cons; xennet_poll()
1012 queue->rx.rsp_cons += skb_queue_len(&tmpq); xennet_poll()
1017 NETFRONT_SKB_CB(skb)->pull_to = rx->status; xennet_poll()
1021 skb_shinfo(skb)->frags[0].page_offset = rx->offset; xennet_poll()
1022 skb_frag_size_set(&skb_shinfo(skb)->frags[0], rx->status); xennet_poll()
1023 skb->data_len = rx->status; xennet_poll()
1024 skb->len += rx->status; xennet_poll()
1028 if (rx->flags & XEN_NETRXF_csum_blank) xennet_poll()
1030 else if (rx->flags & XEN_NETRXF_data_validated) xennet_poll()
1035 queue->rx.rsp_cons = ++i; xennet_poll()
1050 RING_FINAL_CHECK_FOR_RESPONSES(&queue->rx, more_to_do); xennet_poll()
1236 RING_HAS_UNCONSUMED_RESPONSES(&queue->rx))) xennet_rx_interrupt()
1413 xennet_end_access(queue->rx_ring_ref, queue->rx.sring); xennet_disconnect_backend()
1418 queue->rx.sring = NULL; xennet_disconnect_backend()
1506 "%s-rx", queue->name); setup_netfront_split()
1539 queue->rx.sring = NULL; setup_netfront()
1559 xenbus_dev_fatal(dev, err, "allocating rx ring page"); setup_netfront()
1563 FRONT_RING_INIT(&queue->rx, rxs, XEN_PAGE_SIZE); setup_netfront()
1639 /* A grant for every rx ring slot */ xennet_init_queue()
1642 pr_alert("can't alloc rx grant refs\n"); xennet_init_queue()
1691 err = xenbus_printf(*xbt, path, "rx-ring-ref", "%u", write_queue_xenstore_keys()
1694 message = "writing rx-ring-ref"; write_queue_xenstore_keys()
1719 "event-channel-rx", "%u", queue->rx_evtchn); write_queue_xenstore_keys()
1721 message = "writing event-channel-rx"; write_queue_xenstore_keys()
1898 err = xenbus_printf(xbt, dev->nodename, "request-rx-copy", "%u", talk_to_netback()
1901 message = "writing request-rx-copy"; talk_to_netback()
1905 err = xenbus_printf(xbt, dev->nodename, "feature-rx-notify", "%d", 1); talk_to_netback()
1907 message = "writing feature-rx-notify"; talk_to_netback()
1968 "feature-rx-copy", "%u", &feature_rx_copy); xennet_connect()
/linux-4.4.14/fs/jffs2/
H A Dxattr.c137 struct jffs2_raw_xattr rx; do_verify_xattr_datum() local
148 rc = jffs2_flash_read(c, offset, sizeof(rx), &readlen, (char *)&rx); do_verify_xattr_datum()
149 if (rc || readlen != sizeof(rx)) { do_verify_xattr_datum()
151 rc, sizeof(rx), readlen, offset); do_verify_xattr_datum()
154 crc = crc32(0, &rx, sizeof(rx) - 4); do_verify_xattr_datum()
155 if (crc != je32_to_cpu(rx.node_crc)) { do_verify_xattr_datum()
157 offset, je32_to_cpu(rx.hdr_crc), crc); do_verify_xattr_datum()
161 totlen = PAD(sizeof(rx) + rx.name_len + 1 + je16_to_cpu(rx.value_len)); do_verify_xattr_datum()
162 if (je16_to_cpu(rx.magic) != JFFS2_MAGIC_BITMASK do_verify_xattr_datum()
163 || je16_to_cpu(rx.nodetype) != JFFS2_NODETYPE_XATTR do_verify_xattr_datum()
164 || je32_to_cpu(rx.totlen) != totlen do_verify_xattr_datum()
165 || je32_to_cpu(rx.xid) != xd->xid do_verify_xattr_datum()
166 || je32_to_cpu(rx.version) != xd->version) { do_verify_xattr_datum()
169 offset, je16_to_cpu(rx.magic), JFFS2_MAGIC_BITMASK, do_verify_xattr_datum()
170 je16_to_cpu(rx.nodetype), JFFS2_NODETYPE_XATTR, do_verify_xattr_datum()
171 je32_to_cpu(rx.totlen), totlen, do_verify_xattr_datum()
172 je32_to_cpu(rx.xid), xd->xid, do_verify_xattr_datum()
173 je32_to_cpu(rx.version), xd->version); do_verify_xattr_datum()
177 xd->xprefix = rx.xprefix; do_verify_xattr_datum()
178 xd->name_len = rx.name_len; do_verify_xattr_datum()
179 xd->value_len = je16_to_cpu(rx.value_len); do_verify_xattr_datum()
180 xd->data_crc = je32_to_cpu(rx.data_crc); do_verify_xattr_datum()
287 struct jffs2_raw_xattr rx; save_xattr_datum() local
296 vecs[0].iov_base = &rx; save_xattr_datum()
297 vecs[0].iov_len = sizeof(rx); save_xattr_datum()
303 memset(&rx, 0, sizeof(rx)); save_xattr_datum()
304 rx.magic = cpu_to_je16(JFFS2_MAGIC_BITMASK); save_xattr_datum()
305 rx.nodetype = cpu_to_je16(JFFS2_NODETYPE_XATTR); save_xattr_datum()
306 rx.totlen = cpu_to_je32(PAD(totlen)); save_xattr_datum()
307 rx.hdr_crc = cpu_to_je32(crc32(0, &rx, sizeof(struct jffs2_unknown_node) - 4)); save_xattr_datum()
309 rx.xid = cpu_to_je32(xd->xid); save_xattr_datum()
310 rx.version = cpu_to_je32(++xd->version); save_xattr_datum()
311 rx.xprefix = xd->xprefix; save_xattr_datum()
312 rx.name_len = xd->name_len; save_xattr_datum()
313 rx.value_len = cpu_to_je16(xd->value_len); save_xattr_datum()
314 rx.data_crc = cpu_to_je32(crc32(0, vecs[1].iov_base, vecs[1].iov_len)); save_xattr_datum()
315 rx.node_crc = cpu_to_je32(crc32(0, &rx, sizeof(struct jffs2_raw_xattr) - 4)); save_xattr_datum()
/linux-4.4.14/include/net/
H A Dbond_3ad.h50 /* rx machine states(43.4.11 in the 802.3ad standard) */
53 AD_RX_INITIALIZE, /* rx Machine */
54 AD_RX_PORT_DISABLED, /* rx Machine */
55 AD_RX_LACP_DISABLED, /* rx Machine */
56 AD_RX_EXPIRED, /* rx Machine */
57 AD_RX_DEFAULTED, /* rx Machine */
58 AD_RX_CURRENT /* rx Machine */
92 /* rx indication types */
98 /* rx marker indication types */
231 rx_states_t sm_rx_state; /* state machine rx state */
232 u16 sm_rx_timer_counter; /* state machine rx timer counter */
/linux-4.4.14/drivers/iio/accel/
H A Dkxsd9.c48 * @buf_lock: protect the rx and tx buffers.
50 * @rx: single rx buffer storage
56 u8 rx[KXSD9_STATE_RX_SIZE] ____cacheline_aligned;
108 .rx_buf = st->rx, kxsd9_read()
116 ret = (((u16)(st->rx[0])) << 8) | (st->rx[1] & 0xF0); kxsd9_read()
/linux-4.4.14/drivers/staging/wlan-ng/
H A Dp80211netdev.c139 * the device is supposed to be ready for tx and rx. In our
541 netdev_dbg(dev, "rx'd ioctl, cmd=%d, len=%d\n", cmd, req->len); p80211knetdev_do_ioctl()
732 /* Set up the rx queue */ wlan_setup()
848 /* Now to clean out the rx queue */ unregister_wlandev()
934 netdev_dbg(wlandev->netdev, "rx'd mgmt:\n"); p80211_rx_typedrop()
935 wlandev->rx.mgmt++; p80211_rx_typedrop()
939 wlandev->rx.assocreq++; p80211_rx_typedrop()
943 wlandev->rx.assocresp++; p80211_rx_typedrop()
947 wlandev->rx.reassocreq++; p80211_rx_typedrop()
951 wlandev->rx.reassocresp++; p80211_rx_typedrop()
955 wlandev->rx.probereq++; p80211_rx_typedrop()
959 wlandev->rx.proberesp++; p80211_rx_typedrop()
963 wlandev->rx.beacon++; p80211_rx_typedrop()
967 wlandev->rx.atim++; p80211_rx_typedrop()
971 wlandev->rx.disassoc++; p80211_rx_typedrop()
975 wlandev->rx.authen++; p80211_rx_typedrop()
979 wlandev->rx.deauthen++; p80211_rx_typedrop()
983 wlandev->rx.mgmt_unknown++; p80211_rx_typedrop()
996 netdev_dbg(wlandev->netdev, "rx'd ctl:\n"); p80211_rx_typedrop()
997 wlandev->rx.ctl++; p80211_rx_typedrop()
1001 wlandev->rx.pspoll++; p80211_rx_typedrop()
1005 wlandev->rx.rts++; p80211_rx_typedrop()
1009 wlandev->rx.cts++; p80211_rx_typedrop()
1013 wlandev->rx.ack++; p80211_rx_typedrop()
1017 wlandev->rx.cfend++; p80211_rx_typedrop()
1021 wlandev->rx.cfendcfack++; p80211_rx_typedrop()
1025 wlandev->rx.ctl_unknown++; p80211_rx_typedrop()
1033 wlandev->rx.data++; p80211_rx_typedrop()
1036 wlandev->rx.dataonly++; p80211_rx_typedrop()
1039 wlandev->rx.data_cfack++; p80211_rx_typedrop()
1042 wlandev->rx.data_cfpoll++; p80211_rx_typedrop()
1045 wlandev->rx.data__cfack_cfpoll++; p80211_rx_typedrop()
1048 netdev_dbg(wlandev->netdev, "rx'd data:null\n"); p80211_rx_typedrop()
1049 wlandev->rx.null++; p80211_rx_typedrop()
1052 netdev_dbg(wlandev->netdev, "rx'd data:cfack\n"); p80211_rx_typedrop()
1053 wlandev->rx.cfack++; p80211_rx_typedrop()
1056 netdev_dbg(wlandev->netdev, "rx'd data:cfpoll\n"); p80211_rx_typedrop()
1057 wlandev->rx.cfpoll++; p80211_rx_typedrop()
1060 netdev_dbg(wlandev->netdev, "rx'd data:cfack_cfpoll\n"); p80211_rx_typedrop()
1061 wlandev->rx.cfack_cfpoll++; p80211_rx_typedrop()
1065 wlandev->rx.data_unknown++; p80211_rx_typedrop()
/linux-4.4.14/drivers/staging/rtl8723au/include/
H A Dieee80211.h101 /* tx: cck only , rx: cck only, hw: cck */
103 /* tx: ofdm only, rx: ofdm & cck, hw: cck & ofdm */
105 /* tx: ofdm only, rx: ofdm only, hw: ofdm only */
107 /* tx: MCS only, rx: MCS & cck, hw: MCS & cck */
109 /* tx: MCS only, rx: MCS & ofdm, hw: ofdm only */
115 /* tx: cck & ofdm, rx: cck & ofdm & MCS, hw: cck & ofdm */
117 /* tx: ofdm & MCS, rx: ofdm & cck & MCS, hw: cck & ofdm */
119 /* tx: ofdm & MCS, rx: ofdm & MCS, hw: ofdm only */
121 /* tx: ofdm & cck & MCS, rx: ofdm & cck & MCS, hw: ofdm & cck */
123 /* tx: ofdm & MCS, rx: ofdm & MCS, hw: ofdm only */
H A Drtw_ht.h28 u32 rx_ampdu_maxlen; /* for rx reordering ctrl win_sz, updated when join_callback. */
/linux-4.4.14/drivers/net/wireless/rt2x00/
H A Drt2x00config.c130 if (config.rx == ANTENNA_SW_DIVERSITY) { rt2x00lib_config_antenna()
133 if (def->rx == ANTENNA_SW_DIVERSITY) rt2x00lib_config_antenna()
134 config.rx = ANTENNA_B; rt2x00lib_config_antenna()
136 config.rx = def->rx; rt2x00lib_config_antenna()
138 } else if (config.rx == ANTENNA_SW_DIVERSITY) rt2x00lib_config_antenna()
139 config.rx = active->rx; rt2x00lib_config_antenna()
158 rt2x00queue_stop_queue(rt2x00dev->rx); rt2x00lib_config_antenna()
172 rt2x00queue_start_queue(rt2x00dev->rx); rt2x00lib_config_antenna()
/linux-4.4.14/drivers/net/wireless/ath/ath10k/
H A Dhtt_rx.c31 /* when under memory pressure rx ring refill may fail and needs a retry */
147 * Make sure the rx buffer is updated before available buffer __ath10k_htt_rx_ring_fill_n()
148 * index to avoid any potential rx ring corruption. __ath10k_htt_rx_ring_fill_n()
190 * another A-MPDU rx, no special recovery is needed. ath10k_htt_rx_msdu_buff_replenish()
258 ath10k_warn(ar, "tried to pop sk_buff from an empty rx ring\n"); ath10k_htt_rx_netbuf_pop()
276 ath10k_dbg_dump(ar, ATH10K_DBG_HTT_DUMP, NULL, "htt rx netbuf pop: ", ath10k_htt_rx_netbuf_pop()
314 * rx data. ath10k_htt_rx_amsdu_pop()
327 * Copy the FW rx descriptor for this MSDU from the rx ath10k_htt_rx_amsdu_pop()
329 * same rx indication message definition as LL, and simply ath10k_htt_rx_amsdu_pop()
330 * appends new info (fields from the HW rx desc, and the ath10k_htt_rx_amsdu_pop()
331 * MSDU payload itself). So, the offset into the rx ath10k_htt_rx_amsdu_pop()
333 * offset of the per-MSDU FW rx desc info within the ath10k_htt_rx_amsdu_pop()
334 * message, and how many bytes of the per-MSDU FW rx desc ath10k_htt_rx_amsdu_pop()
336 * the host, since for a big-endian host, the rx ind ath10k_htt_rx_amsdu_pop()
337 * message contents, including the per-MSDU rx desc bytes, ath10k_htt_rx_amsdu_pop()
344 * per-MSDU rx descriptors. Just to be sure, verify ath10k_htt_rx_amsdu_pop()
382 /* Note: Chained buffers do not contain rx descriptor */ ath10k_htt_rx_amsdu_pop()
416 * refill the rx ring with 1 PPDU's worth of rx buffers (something ath10k_htt_rx_amsdu_pop()
417 * like 32 x 3 buffers), rather than one MPDU's worth of rx buffers ath10k_htt_rx_amsdu_pop()
419 * SW to tell us when it is done pulling all the PPDU's rx buffers ath10k_htt_rx_amsdu_pop()
420 * out of the rx ring, and then refill it just once. ath10k_htt_rx_amsdu_pop()
453 ath10k_dbg_dump(ar, ATH10K_DBG_HTT_DUMP, NULL, "htt rx netbuf pop: ", ath10k_htt_rx_pop_paddr()
527 ath10k_warn(ar, "htt rx ring size is not power of 2\n"); ath10k_htt_rx_alloc()
576 ath10k_dbg(ar, ATH10K_DBG_BOOT, "htt rx ring size %d fill_level %d\n", ath10k_htt_rx_alloc()
739 "rx desc msdu payload: ", ath10k_htt_rx_h_rates()
982 "rx skb %p len %u peer %pM %s %s sn %u %s%s%s%s%s %srate_idx %u vht_nss %u freq %u band %u flag 0x%x fcs-err %i mic-err %i amsdu-more %i\n", ath10k_process_rx()
1003 ath10k_dbg_dump(ar, ATH10K_DBG_HTT_DUMP, NULL, "rx skb: ", ath10k_process_rx()
1067 * the case for frames that come in via fragmented rx indication. ath10k_htt_rx_h_undecap_raw()
1451 /* TODO: bump some rx-oom error stat */ ath10k_unchain_msdu()
1536 ath10k_dbg(ar, ATH10K_DBG_HTT, "htt rx mgmt ctrl\n"); ath10k_htt_rx_amsdu_allowed()
1541 ath10k_dbg(ar, ATH10K_DBG_HTT, "htt rx cac running\n"); ath10k_htt_rx_amsdu_allowed()
1562 struct htt_rx_indication *rx) ath10k_htt_rx_handler()
1578 fw_desc_len = __le16_to_cpu(rx->prefix.fw_rx_desc_bytes); ath10k_htt_rx_handler()
1579 fw_desc = (u8 *)&rx->fw_desc; ath10k_htt_rx_handler()
1581 num_mpdu_ranges = MS(__le32_to_cpu(rx->hdr.info1), ath10k_htt_rx_handler()
1583 mpdu_ranges = htt_rx_ind_get_mpdu_ranges(rx); ath10k_htt_rx_handler()
1585 ath10k_dbg_dump(ar, ATH10K_DBG_HTT_DUMP, NULL, "htt rx ind: ", ath10k_htt_rx_handler()
1586 rx, sizeof(*rx) + ath10k_htt_rx_handler()
1598 ath10k_warn(ar, "rx ring became corrupted: %d\n", ret); ath10k_htt_rx_handler()
1639 ath10k_dbg(ar, ATH10K_DBG_HTT_DUMP, "htt rx frag ahead\n"); ath10k_htt_rx_frag_handler()
1642 ath10k_warn(ar, "failed to pop amsdu from httr rx ring for fragmented rx %d\n", ath10k_htt_rx_frag_handler()
1661 "expecting more fragmented rx in one indication %d\n", ath10k_htt_rx_frag_handler()
1716 "htt rx addba tid %hu peer_id %hu size %hhu\n", ath10k_htt_rx_addba()
1737 "htt rx start rx ba session sta %pM tid %hu size %hhu\n", ath10k_htt_rx_addba()
1756 "htt rx delba tid %hu peer_id %hu\n", ath10k_htt_rx_delba()
1777 "htt rx stop rx ba session sta %pM tid %hu\n", ath10k_htt_rx_delba()
1840 struct htt_rx_offload_msdu *rx; ath10k_htt_rx_h_rx_offload() local
1849 rx = (void *)msdu->data; ath10k_htt_rx_h_rx_offload()
1851 skb_put(msdu, sizeof(*rx)); ath10k_htt_rx_h_rx_offload()
1852 skb_pull(msdu, sizeof(*rx)); ath10k_htt_rx_h_rx_offload()
1854 if (skb_tailroom(msdu) < __le16_to_cpu(rx->msdu_len)) { ath10k_htt_rx_h_rx_offload()
1855 ath10k_warn(ar, "dropping frame: offloaded rx msdu is too long!\n"); ath10k_htt_rx_h_rx_offload()
1860 skb_put(msdu, __le16_to_cpu(rx->msdu_len)); ath10k_htt_rx_h_rx_offload()
1862 /* Offloaded rx header length isn't multiple of 2 nor 4 so the ath10k_htt_rx_h_rx_offload()
1880 ath10k_htt_rx_h_channel(ar, status, NULL, rx->vdev_id); ath10k_htt_rx_h_rx_offload()
1917 "htt rx in ord vdev %i peer %i tid %i offload %i frag %i msdu count %i\n", ath10k_htt_rx_in_ord_ind()
1921 ath10k_warn(ar, "dropping invalid in order rx indication\n"); ath10k_htt_rx_in_ord_ind()
1948 * frames from different PPDUs meaning reported rx rate ath10k_htt_rx_in_ord_ind()
1951 * should still give an idea about rx rate to the user. ath10k_htt_rx_in_ord_ind()
1982 ath10k_dbg(ar, ATH10K_DBG_HTT, "htt rx, msg_type: 0x%0X\n", ath10k_htt_t2h_msg_handler()
1986 ath10k_dbg(ar, ATH10K_DBG_HTT, "htt rx, unsupported msg_type: 0x%0X\n max: 0x%0X", ath10k_htt_t2h_msg_handler()
1561 ath10k_htt_rx_handler(struct ath10k_htt *htt, struct htt_rx_indication *rx) ath10k_htt_rx_handler() argument
/linux-4.4.14/drivers/staging/wilc1000/
H A Dwilc_wfi_cfgoperations.h64 .rx = BIT(IEEE80211_STYPE_ACTION >> 4) |
69 .rx = BIT(IEEE80211_STYPE_ASSOC_REQ >> 4) |
79 .rx = BIT(IEEE80211_STYPE_ACTION >> 4) |
/linux-4.4.14/drivers/thunderbolt/
H A Dnhi_regs.h82 * three bitfields: tx, rx, rx overflow
91 * two bitfields: rx, tx
H A Dctl.c31 struct tb_ring *rx; member in struct:tb_ctl
389 ring_rx(pkg->ctl->rx, &pkg->frame); /* tb_ctl_rx_submit()
391 * All rx packets are referenced tb_ctl_rx_submit()
411 goto rx; tb_ctl_rx_callback()
419 goto rx; tb_ctl_rx_callback()
425 goto rx; tb_ctl_rx_callback()
429 goto rx; tb_ctl_rx_callback()
433 rx: tb_ctl_rx_callback()
495 ctl->rx = ring_alloc_rx(nhi, 0, 10); tb_ctl_alloc()
496 if (!ctl->rx) tb_ctl_alloc()
523 if (ctl->rx) tb_ctl_free()
524 ring_free(ctl->rx); tb_ctl_free()
546 ring_start(ctl->rx); tb_ctl_start()
561 ring_stop(ctl->rx); tb_ctl_stop()
/linux-4.4.14/drivers/iio/imu/
H A Dadis_buffer.c30 __be16 *tx, *rx; adis_update_scan_mode() local
45 rx = adis->buffer; adis_update_scan_mode()
46 tx = rx + scan_count; adis_update_scan_mode()
59 adis->xfer[j].rx_buf = &rx[j - 1]; adis_update_scan_mode()
/linux-4.4.14/drivers/hwmon/
H A Dad7314.c41 u16 rx ____cacheline_aligned;
48 ret = spi_read(chip->spi_dev, (u8 *)&chip->rx, sizeof(chip->rx)); ad7314_spi_read()
54 return be16_to_cpu(chip->rx); ad7314_spi_read()
/linux-4.4.14/arch/sh/kernel/cpu/sh3/
H A Dserial-sh7720.c24 /* Clear PTCR bit 5-2; enable only tx and rx */ sh7720_sci_init_pins()
/linux-4.4.14/drivers/soc/ti/
H A Dknav_dma.c166 if (cfg->u.rx.einfo_present) chan_start()
168 if (cfg->u.rx.psinfo_present) chan_start()
170 if (cfg->u.rx.err_mode == DMA_RETRY) chan_start()
172 v |= (cfg->u.rx.desc_type & DESC_TYPE_MASK) << DESC_TYPE_SHIFT; chan_start()
173 if (cfg->u.rx.psinfo_at_sop) chan_start()
175 v |= (cfg->u.rx.sop_offset & CHAN_SOP_OFF_MASK) chan_start()
177 v |= cfg->u.rx.dst_q & CHAN_QNUM_MASK; chan_start()
183 v = cfg->u.rx.fdq[0] << 16; chan_start()
184 v |= cfg->u.rx.fdq[1] & CHAN_QNUM_MASK; chan_start()
187 v = cfg->u.rx.fdq[2] << 16; chan_start()
188 v |= cfg->u.rx.fdq[3] & CHAN_QNUM_MASK; chan_start()
318 ((chan->direction == DMA_MEM_TO_DEV) ? "tx chan" : "rx flow"), dma_debug_show_channels()
328 chan->cfg.u.rx.einfo_present, dma_debug_show_channels()
329 chan->cfg.u.rx.psinfo_present, dma_debug_show_channels()
330 chan->cfg.u.rx.desc_type); dma_debug_show_channels()
332 chan->cfg.u.rx.dst_q, dma_debug_show_channels()
333 chan->cfg.u.rx.thresh); dma_debug_show_channels()
335 seq_printf(s, "[%d]", chan->cfg.u.rx.fdq[i]); dma_debug_show_channels()
562 dev_dbg(kdev->dev, "rx flow(%d) (%p)\n", chan->flow, chan->reg_rx_flow); pktdma_init_rx_chan()
685 ret = of_property_read_u32(node, "ti,rx-retry-timeout", &timeout); dma_init()
687 dev_dbg(kdev->dev, "unspecified rx timeout using value %d\n", dma_init()
722 dev_info(kdev->dev, "DMA %s registered %d logical channels, flows %d, tx chans: %d, rx chans: %d%s\n", dma_init()
/linux-4.4.14/drivers/iio/gyro/
H A Dadxrs450.c70 * @buf_lock: mutex to protect tx and rx
72 * @rx: receive buffer
78 __be32 rx; member in struct:adxrs450_state
103 .rx_buf = &st->rx, adxrs450_spi_read_reg_16()
105 .len = sizeof(st->rx), adxrs450_spi_read_reg_16()
123 *val = (be32_to_cpu(st->rx) >> 5) & 0xFFFF; adxrs450_spi_read_reg_16()
177 .rx_buf = &st->rx, adxrs450_spi_sensor_data()
179 .len = sizeof(st->rx), adxrs450_spi_sensor_data()
192 *val = (be32_to_cpu(st->rx) >> 10) & 0xFFFF; adxrs450_spi_sensor_data()
212 .rx_buf = &st->rx, adxrs450_spi_initial()
228 *val = be32_to_cpu(st->rx); adxrs450_spi_initial()
/linux-4.4.14/drivers/net/ethernet/intel/ixgbe/
H A Dixgbe_dcb_nl.c56 int rx = DCB_RX_CONFIG; ixgbe_copy_dcb_cfg() local
95 if (dst->path[rx].prio_type != src->path[rx].prio_type) { ixgbe_copy_dcb_cfg()
96 dst->path[rx].prio_type = src->path[rx].prio_type; ixgbe_copy_dcb_cfg()
100 if (dst->path[rx].bwg_id != src->path[rx].bwg_id) { ixgbe_copy_dcb_cfg()
101 dst->path[rx].bwg_id = src->path[rx].bwg_id; ixgbe_copy_dcb_cfg()
105 if (dst->path[rx].bwg_percent != src->path[rx].bwg_percent) { ixgbe_copy_dcb_cfg()
106 dst->path[rx].bwg_percent = src->path[rx].bwg_percent; ixgbe_copy_dcb_cfg()
110 if (dst->path[rx].up_to_tc_bitmap != ixgbe_copy_dcb_cfg()
111 src->path[rx].up_to_tc_bitmap) { ixgbe_copy_dcb_cfg()
112 dst->path[rx].up_to_tc_bitmap = ixgbe_copy_dcb_cfg()
113 src->path[rx].up_to_tc_bitmap; ixgbe_copy_dcb_cfg()
124 if (dcfg->bw_percentage[rx][j] != scfg->bw_percentage[rx][j]) { ixgbe_copy_dcb_cfg()
125 dcfg->bw_percentage[rx][j] = scfg->bw_percentage[rx][j]; ixgbe_copy_dcb_cfg()
/linux-4.4.14/drivers/net/wireless/cw1200/
H A Ddebug.h19 int rx; member in struct:cw1200_debug_priv
55 ++priv->debug->rx; cw1200_debug_rxed()
H A Dbh.c291 pr_err("rx blew up, len %zu\n", alloc_len); cw1200_bh_rx_helper()
426 int rx, tx, term, suspend; cw1200_bh() local
460 rx = atomic_xchg(&priv->bh_rx, 0); cw1200_bh()
465 (rx || tx || term || suspend || priv->bh_error); cw1200_bh()
468 pr_debug("[BH] - rx: %d, tx: %d, term: %d, bh_err: %d, suspend: %d, status: %ld\n", cw1200_bh()
469 rx, tx, term, suspend, priv->bh_error, status); cw1200_bh()
483 if (priv->hw_bufs_used && (!rx || !tx)) { cw1200_bh()
487 rx = 1; cw1200_bh()
547 rx: cw1200_bh()
554 /* Don't bother trying to rx unless we have data to read */ cw1200_bh()
577 * after we handle rx.. cw1200_bh()
597 goto rx; cw1200_bh()
/linux-4.4.14/arch/mips/netlogic/xlp/
H A Dcop2-ex.c52 : "r"(r->tx), "r"(r->rx)); nlm_cop2_save()
82 : : "m"(*r), "r"(r->tx), "r"(r->rx)); nlm_cop2_restore()
/linux-4.4.14/drivers/i2c/busses/
H A Di2c-designware-core.h48 * @msg_read_idx: the element index of the current rx message in the msgs
50 * @rx_buf_len: the length of the current rx buffer
51 * @rx_buf: the current rx buffer
58 * @rx_fifo_depth: depth of the hardware rx fifo
59 * @rx_outstanding: current master-rx elements in tx fifo
H A Di2c-dln2.c117 } __packed *rx = dln2->buf; dln2_i2c_read() local
118 unsigned rx_len = sizeof(*rx); dln2_i2c_read()
120 BUILD_BUG_ON(sizeof(*rx) > DLN2_I2C_BUF_SIZE); dln2_i2c_read()
129 rx, &rx_len); dln2_i2c_read()
132 if (rx_len < sizeof(rx->buf_len) + data_len) dln2_i2c_read()
134 if (le16_to_cpu(rx->buf_len) != data_len) dln2_i2c_read()
137 memcpy(data, rx->buf, data_len); dln2_i2c_read()
/linux-4.4.14/include/xen/interface/io/
H A Dnetif.h36 * If the client sends notification for rx requests then it should specify
37 * feature 'feature-rx-notify' via xenbus. Otherwise the backend will assume
48 * "event-channel-tx" and "event-channel-rx" respectively. If frontend
68 * number of tx and rx rings.
76 * event-channel (or event-channel-{tx,rx}) and {tx,rx}-ring-ref keys,
85 * /local/domain/1/device/vif/0/queue-0/rx-ring-ref = "<ring-ref-rx0>"
87 * /local/domain/1/device/vif/0/queue-0/event-channel-rx = "<evtchn-rx0>"
90 * /local/domain/1/device/vif/0/queue-1/rx-ring-ref = "<ring-ref-rx1"
92 * /local/domain/1/device/vif/0/queue-1/event-channel-rx = "<evtchn-rx1>"
/linux-4.4.14/drivers/net/ethernet/toshiba/
H A Dps3_gelic_net.h129 GELIC_DESCR_DMA_BUFFER_FULL = 0x00000000, /* used in rx */
130 GELIC_DESCR_DMA_RESPONSE_ERROR = 0x10000000, /* used in rx, tx */
131 GELIC_DESCR_DMA_PROTECTION_ERROR = 0x20000000, /* used in rx, tx */
132 GELIC_DESCR_DMA_FRAME_END = 0x40000000, /* used in rx */
133 GELIC_DESCR_DMA_FORCE_END = 0x50000000, /* used in rx, tx */
134 GELIC_DESCR_DMA_CARDOWNED = 0xa0000000, /* used in rx, tx */
271 u16 rx; member in struct:gelic_vlan_id
H A Dps3_gelic_net.c257 /* start rx */ gelic_card_up()
281 /* stop rx */ gelic_card_down()
373 * gelic_descr_prepare_rx - reinitializes a rx descriptor
379 * allocates a new rx skb, iommu-maps it and attaches it to the descriptor.
421 "%s:Could not iommu-map rx buffer\n", __func__); gelic_descr_prepare_rx()
431 * gelic_card_release_rx_chain - free all skb of rx descr
456 * gelic_card_fill_rx_chain - fills descriptors/skbs in the rx chains
459 * fills all descriptors in the rx chain: allocates skbs
484 * gelic_card_alloc_rx_skbs - allocates rx skbs in rx descriptor chains
972 * gelic_card_decode_one_descr - processes an rx descriptor
977 * processes an rx descriptor by iommu-unmapping the data buffer and passing
1004 if (card->vlan[i].rx == vid) { gelic_card_decode_one_descr()
1594 int rx; gelic_card_get_vlan_info() member in struct:__anon7812
1598 .rx = GELIC_LV1_VLAN_RX_ETHERNET_0 gelic_card_get_vlan_info()
1602 .rx = GELIC_LV1_VLAN_RX_WIRELESS gelic_card_get_vlan_info()
1618 card->vlan[i].rx = 0; gelic_card_get_vlan_info()
1623 /* rx tag */ gelic_card_get_vlan_info()
1626 vlan_id_ix[i].rx, gelic_card_get_vlan_info()
1631 "get vlan id for rx(%d) failed(%d)\n", gelic_card_get_vlan_info()
1632 vlan_id_ix[i].rx, status); gelic_card_get_vlan_info()
1634 card->vlan[i].rx = 0; gelic_card_get_vlan_info()
1637 card->vlan[i].rx = (u16)v1; gelic_card_get_vlan_info()
1639 dev_dbg(ctodev(card), "vlan_id[%d] tx=%02x rx=%02x\n", gelic_card_get_vlan_info()
1640 i, card->vlan[i].tx, card->vlan[i].rx); gelic_card_get_vlan_info()
1652 card->vlan[GELIC_PORT_WIRELESS].rx = 0; gelic_card_get_vlan_info()
1755 dev_dbg(ctodev(card), "descr rx %p, tx %p, size %#lx, num %#x\n", ps3_gelic_driver_probe()
1758 /* allocate rx skbs */ ps3_gelic_driver_probe()
/linux-4.4.14/drivers/net/fddi/skfp/h/
H A Dfplustm.h72 __le32 rxd_rbadr ; /* physical rx buffer address */
113 HW_PTR rx_bmu_ctl ; /* BMU addr for rx start */
142 u_short rx1_fifo_start ; /* rx queue start address */
143 u_short rx1_fifo_size ; /* rx queue size */
144 u_short rx2_fifo_start ; /* rx queue start address */
145 u_short rx2_fifo_size ; /* rx queue size */
190 struct s_smt_rx_queue *rx[USED_QUEUES] ; member in struct:s_smt_fp
/linux-4.4.14/Documentation/spi/
H A Dspidev_test.c103 static void transfer(int fd, uint8_t const *tx, uint8_t const *rx, size_t len) transfer() argument
109 .rx_buf = (unsigned long)rx, transfer()
137 hex_dump(rx, len, 32, "RX"); transfer()
257 uint8_t *rx; main() local
306 rx = malloc(size); main()
308 transfer(fd, tx, rx, size); main()
309 free(rx); main()
/linux-4.4.14/drivers/staging/most/aim-network/
H A Dnetworking.c72 struct net_dev_channel rx; member in struct:net_dev_context
192 BUG_ON(!nd->tx.linked || !nd->rx.linked); most_nd_open()
194 if (most_start_channel(nd->iface, nd->rx.ch_id, &aim)) { most_nd_open()
201 most_stop_channel(nd->iface, nd->rx.ch_id, &aim); most_nd_open()
227 most_stop_channel(nd->iface, nd->rx.ch_id, &aim); most_nd_stop()
340 ch = ccfg->direction == MOST_CH_TX ? &nd->tx : &nd->rx; aim_probe_channel()
346 if (nd->tx.linked || nd->rx.linked) { aim_probe_channel()
385 if (nd->rx.linked && channel_idx == nd->rx.ch_id) aim_disconnect_channel()
386 ch = &nd->rx; aim_disconnect_channel()
400 if (!nd->rx.linked && !nd->tx.linked) { aim_disconnect_channel()
436 if (!nd || !nd->channels_opened || nd->rx.ch_id != mbo->hdm_channel_id) aim_rx_data()
/linux-4.4.14/drivers/media/usb/dvb-usb/
H A Dttusb2.c123 u8 rx[60];/* (64 -4) */ tt3650_ci_msg() local
124 ret = ttusb2_msg(d, cmd, data, write_len, rx, read_len); tt3650_ci_msg()
126 memcpy(data, rx, read_len); tt3650_ci_msg()
433 u8 rx[9]; /* A CMD_GET_IR_CODE reply is 9 bytes long */ tt3650_rc_query() local
435 ret = ttusb2_msg(d, CMD_GET_IR_CODE, NULL, 0, rx, sizeof(rx)); tt3650_rc_query()
439 if (rx[8] & 0x01) { tt3650_rc_query()
441 st->last_rc_key = RC_SCANCODE_RC5(rx[3], rx[2]); tt3650_rc_query()
442 deb_info("%s: cmd=0x%02x sys=0x%02x\n", __func__, rx[2], rx[3]); tt3650_rc_query()
443 rc_keydown(d->rc_dev, RC_TYPE_RC5, st->last_rc_key, rx[1]); tt3650_rc_query()
H A Dpctv452e.c503 u8 rx[PCTV_ANSWER_LEN]; pctv452e_power_ctrl() local
523 ret = dvb_usb_generic_rw(d, b0, sizeof(b0), rx, PCTV_ANSWER_LEN, 0); pctv452e_power_ctrl()
530 ret = dvb_usb_generic_rw(d, b0, sizeof(b0), rx, PCTV_ANSWER_LEN, 0); pctv452e_power_ctrl()
543 u8 rx[PCTV_ANSWER_LEN]; pctv452e_rc_query() local
554 ret = dvb_usb_generic_rw(d, b, 4, rx, PCTV_ANSWER_LEN, 0); pctv452e_rc_query()
559 info("%s: read: %2d: %*ph: ", __func__, ret, 3, rx); pctv452e_rc_query()
560 for (i = 0; (i < rx[3]) && ((i+3) < PCTV_ANSWER_LEN); i++) pctv452e_rc_query()
561 info(" %02x", rx[i+3]); pctv452e_rc_query()
566 if ((rx[3] == 9) && (rx[12] & 0x01)) { pctv452e_rc_query()
568 state->last_rc_key = RC_SCANCODE_RC5(rx[7], rx[6]); pctv452e_rc_query()
571 __func__, rx[6], rx[7]); pctv452e_rc_query()
/linux-4.4.14/drivers/net/wireless/iwlegacy/
H A D3945-debug.c81 ofdm = &il->_3945.stats.rx.ofdm; il3945_ucode_rx_stats_read()
82 cck = &il->_3945.stats.rx.cck; il3945_ucode_rx_stats_read()
83 general = &il->_3945.stats.rx.general; il3945_ucode_rx_stats_read()
84 accum_ofdm = &il->_3945.accum_stats.rx.ofdm; il3945_ucode_rx_stats_read()
85 accum_cck = &il->_3945.accum_stats.rx.cck; il3945_ucode_rx_stats_read()
86 accum_general = &il->_3945.accum_stats.rx.general; il3945_ucode_rx_stats_read()
87 delta_ofdm = &il->_3945.delta_stats.rx.ofdm; il3945_ucode_rx_stats_read()
88 delta_cck = &il->_3945.delta_stats.rx.cck; il3945_ucode_rx_stats_read()
89 delta_general = &il->_3945.delta_stats.rx.general; il3945_ucode_rx_stats_read()
90 max_ofdm = &il->_3945.max_delta.rx.ofdm; il3945_ucode_rx_stats_read()
91 max_cck = &il->_3945.max_delta.rx.cck; il3945_ucode_rx_stats_read()
92 max_general = &il->_3945.max_delta.rx.general; il3945_ucode_rx_stats_read()
/linux-4.4.14/drivers/net/ethernet/altera/
H A Daltera_sgdma.c86 netdev_err(priv->dev, "error mapping rx descriptor memory\n"); sgdma_initialize()
270 "sgdma rx and rx queue empty!\n"); sgdma_rx_status()
277 /* kick the rx sgdma after reaping this descriptor */ sgdma_rx_status()
359 netdev_err(priv->dev, "no rx buffers available\n"); sgdma_async_read()
366 0, /* read addr 0 for rx dma */ sgdma_async_read()
367 rxbuffer->dma_addr, /* write addr for rx dma */ sgdma_async_read()
369 0, /* EOP: NA for rx dma */ sgdma_async_read()
370 0, /* read fixed: NA for rx dma */ sgdma_async_read()
371 0); /* SOP: NA for rx DMA */ sgdma_async_read()
462 /* adds a tse_buffer to the tail of a rx buffer list
512 /* check and return rx sgdma status without polling
/linux-4.4.14/drivers/net/ethernet/amd/
H A Dau1000_eth.h30 #define NUM_RX_DMA 4 /* Au1x00 has 4 rx hardware descriptors */
127 u32 vaddr; /* virtual address of rx/tx buffers */
128 dma_addr_t dma_addr; /* dma address of rx/tx buffers */
/linux-4.4.14/drivers/staging/lustre/lnet/klnds/o2iblnd/
H A Do2iblnd_cb.c132 kiblnd_drop_rx(kib_rx_t *rx) kiblnd_drop_rx() argument
134 kib_conn_t *conn = rx->rx_conn; kiblnd_drop_rx()
147 kiblnd_post_rx(kib_rx_t *rx, int credit) kiblnd_post_rx() argument
149 kib_conn_t *conn = rx->rx_conn; kiblnd_post_rx()
161 mr = kiblnd_find_dma_mr(conn->ibc_hdev, rx->rx_msgaddr, IBLND_MSG_SIZE); kiblnd_post_rx()
164 rx->rx_sge.lkey = mr->lkey; kiblnd_post_rx()
165 rx->rx_sge.addr = rx->rx_msgaddr; kiblnd_post_rx()
166 rx->rx_sge.length = IBLND_MSG_SIZE; kiblnd_post_rx()
168 rx->rx_wrq.next = NULL; kiblnd_post_rx()
169 rx->rx_wrq.sg_list = &rx->rx_sge; kiblnd_post_rx()
170 rx->rx_wrq.num_sge = 1; kiblnd_post_rx()
171 rx->rx_wrq.wr_id = kiblnd_ptr2wreqid(rx, IBLND_WID_RX); kiblnd_post_rx()
174 LASSERT(rx->rx_nob >= 0); /* not posted */ kiblnd_post_rx()
177 kiblnd_drop_rx(rx); /* No more posts for this rx */ kiblnd_post_rx()
181 rx->rx_nob = -1; /* flag posted */ kiblnd_post_rx()
184 * own this rx (and rx::rx_conn) anymore, LU-5678. kiblnd_post_rx()
187 rc = ib_post_recv(conn->ibc_cmid->qp, &rx->rx_wrq, &bad_wrq); kiblnd_post_rx()
189 CERROR("Can't post rx for %s: %d, bad_wrq: %p\n", kiblnd_post_rx()
191 rx->rx_nob = 0; kiblnd_post_rx()
199 kiblnd_drop_rx(rx); /* No more posts for this rx */ kiblnd_post_rx()
302 kiblnd_handle_rx(kib_rx_t *rx) kiblnd_handle_rx() argument
304 kib_msg_t *msg = rx->rx_msg; kiblnd_handle_rx()
305 kib_conn_t *conn = rx->rx_conn; kiblnd_handle_rx()
334 kiblnd_post_rx(rx, IBLND_POSTRX_NO_CREDIT); kiblnd_handle_rx()
372 msg->ibm_srcnid, rx, 0); kiblnd_handle_rx()
380 msg->ibm_srcnid, rx, 1); kiblnd_handle_rx()
442 msg->ibm_srcnid, rx, 1); kiblnd_handle_rx()
459 kiblnd_post_rx(rx, post_credit); kiblnd_handle_rx()
463 kiblnd_rx_complete(kib_rx_t *rx, int status, int nob) kiblnd_rx_complete() argument
465 kib_msg_t *msg = rx->rx_msg; kiblnd_rx_complete()
466 kib_conn_t *conn = rx->rx_conn; kiblnd_rx_complete()
473 LASSERT(rx->rx_nob < 0); /* was posted */ kiblnd_rx_complete()
474 rx->rx_nob = 0; /* isn't now */ kiblnd_rx_complete()
486 rx->rx_nob = nob; kiblnd_rx_complete()
488 rc = kiblnd_unpack_msg(msg, rx->rx_nob); kiblnd_rx_complete()
490 CERROR("Error %d unpacking rx from %s\n", kiblnd_rx_complete()
499 CERROR("Stale rx from %s\n", kiblnd_rx_complete()
517 list_add_tail(&rx->rx_list, &conn->ibc_early_rxs); kiblnd_rx_complete()
523 kiblnd_handle_rx(rx); kiblnd_rx_complete()
527 CDEBUG(D_NET, "rx %p conn %p\n", rx, conn); kiblnd_rx_complete()
530 kiblnd_drop_rx(rx); /* Don't re-post rx. */ kiblnd_rx_complete()
1587 kiblnd_reply(lnet_ni_t *ni, kib_rx_t *rx, lnet_msg_t *lntmsg) kiblnd_reply() argument
1598 tx = kiblnd_get_idle_tx(ni, rx->rx_conn->ibc_peer->ibp_nid); kiblnd_reply()
1620 rc = kiblnd_init_rdma(rx->rx_conn, tx, kiblnd_reply()
1622 &rx->rx_msg->ibm_u.get.ibgm_rd, kiblnd_reply()
1623 rx->rx_msg->ibm_u.get.ibgm_cookie); kiblnd_reply()
1639 kiblnd_queue_tx(tx, rx->rx_conn); kiblnd_reply()
1653 kib_rx_t *rx = private; kiblnd_recv() local
1654 kib_msg_t *rxmsg = rx->rx_msg; kiblnd_recv()
1655 kib_conn_t *conn = rx->rx_conn; kiblnd_recv()
1672 if (nob > rx->rx_nob) { kiblnd_recv()
1675 nob, rx->rx_nob); kiblnd_recv()
1699 kiblnd_send_completion(rx->rx_conn, IBLND_MSG_PUT_NAK, 0, kiblnd_recv()
1726 kiblnd_send_completion(rx->rx_conn, IBLND_MSG_PUT_NAK, rc, kiblnd_recv()
1749 kiblnd_reply(ni, rx, lntmsg); kiblnd_recv()
1752 kiblnd_send_completion(rx->rx_conn, IBLND_MSG_GET_DONE, kiblnd_recv()
1759 kiblnd_post_rx(rx, post_credit); kiblnd_recv()
1899 kib_rx_t *rx; kiblnd_handle_early_rxs() local
1906 list_for_each_entry_safe(rx, tmp, &conn->ibc_early_rxs, rx_list) { kiblnd_handle_early_rxs()
1907 list_del(&rx->rx_list); kiblnd_handle_early_rxs()
1910 kiblnd_handle_rx(rx); kiblnd_handle_early_rxs()
/linux-4.4.14/net/bluetooth/
H A Decc.c694 u64 rx[2][NUM_ECC_DIGITS]; ecc_point_mult() local
699 vli_set(rx[1], point->x); ecc_point_mult()
702 xycz_initial_double(rx[1], ry[1], rx[0], ry[0], initial_z); ecc_point_mult()
706 xycz_add_c(rx[1 - nb], ry[1 - nb], rx[nb], ry[nb]); ecc_point_mult()
707 xycz_add(rx[nb], ry[nb], rx[1 - nb], ry[1 - nb]); ecc_point_mult()
711 xycz_add_c(rx[1 - nb], ry[1 - nb], rx[nb], ry[nb]); ecc_point_mult()
714 vli_mod_sub(z, rx[1], rx[0], curve_p); /* X1 - X0 */ ecc_point_mult()
719 vli_mod_mult_fast(z, z, rx[1 - nb]); /* Xb * yP / (xP * Yb * (X1 - X0)) */ ecc_point_mult()
722 xycz_add(rx[nb], ry[nb], rx[1 - nb], ry[1 - nb]); ecc_point_mult()
724 apply_z(rx[0], ry[0], z); ecc_point_mult()
726 vli_set(result->x, rx[0]); ecc_point_mult()
/linux-4.4.14/drivers/net/ethernet/intel/
H A De100.c444 struct rx { struct
445 struct rx *next, *prev;
572 struct rx *rxs ____cacheline_aligned;
573 struct rx *rx_to_use;
574 struct rx *rx_to_clean;
1934 static inline void e100_start_receiver(struct nic *nic, struct rx *rx) e100_start_receiver() argument
1940 if (!rx) rx = nic->rxs; e100_start_receiver()
1943 if (rx->skb) { e100_start_receiver()
1944 e100_exec_cmd(nic, ruc_start, rx->dma_addr); e100_start_receiver()
1950 static int e100_rx_alloc_skb(struct nic *nic, struct rx *rx) e100_rx_alloc_skb() argument
1952 if (!(rx->skb = netdev_alloc_skb_ip_align(nic->netdev, RFD_BUF_LEN))) e100_rx_alloc_skb()
1956 skb_copy_to_linear_data(rx->skb, &nic->blank_rfd, sizeof(struct rfd)); e100_rx_alloc_skb()
1957 rx->dma_addr = pci_map_single(nic->pdev, rx->skb->data, e100_rx_alloc_skb()
1960 if (pci_dma_mapping_error(nic->pdev, rx->dma_addr)) { e100_rx_alloc_skb()
1961 dev_kfree_skb_any(rx->skb); e100_rx_alloc_skb()
1962 rx->skb = NULL; e100_rx_alloc_skb()
1963 rx->dma_addr = 0; e100_rx_alloc_skb()
1970 if (rx->prev->skb) { e100_rx_alloc_skb()
1971 struct rfd *prev_rfd = (struct rfd *)rx->prev->skb->data; e100_rx_alloc_skb()
1972 put_unaligned_le32(rx->dma_addr, &prev_rfd->link); e100_rx_alloc_skb()
1973 pci_dma_sync_single_for_device(nic->pdev, rx->prev->dma_addr, e100_rx_alloc_skb()
1980 static int e100_rx_indicate(struct nic *nic, struct rx *rx, e100_rx_indicate() argument
1984 struct sk_buff *skb = rx->skb; e100_rx_indicate()
1993 pci_dma_sync_single_for_cpu(nic->pdev, rx->dma_addr, e100_rx_indicate()
2013 pci_dma_sync_single_for_device(nic->pdev, rx->dma_addr, e100_rx_indicate()
2027 pci_unmap_single(nic->pdev, rx->dma_addr, e100_rx_indicate()
2074 rx->skb = NULL; e100_rx_indicate()
2082 struct rx *rx; e100_rx_clean() local
2084 struct rx *old_before_last_rx, *new_before_last_rx; e100_rx_clean()
2088 for (rx = nic->rx_to_clean; rx->skb; rx = nic->rx_to_clean = rx->next) { e100_rx_clean()
2089 err = e100_rx_indicate(nic, rx, work_done, work_to_do); e100_rx_clean()
2109 for (rx = nic->rx_to_use; !rx->skb; rx = nic->rx_to_use = rx->next) { e100_rx_clean()
2110 if (unlikely(e100_rx_alloc_skb(nic, rx))) e100_rx_clean()
2158 struct rx *rx; e100_rx_clean_list() local
2164 for (rx = nic->rxs, i = 0; i < count; rx++, i++) { e100_rx_clean_list()
2165 if (rx->skb) { e100_rx_clean_list()
2166 pci_unmap_single(nic->pdev, rx->dma_addr, e100_rx_clean_list()
2168 dev_kfree_skb(rx->skb); e100_rx_clean_list()
2180 struct rx *rx; e100_rx_alloc_list() local
2187 if (!(nic->rxs = kcalloc(count, sizeof(struct rx), GFP_ATOMIC))) e100_rx_alloc_list()
2190 for (rx = nic->rxs, i = 0; i < count; rx++, i++) { e100_rx_alloc_list()
2191 rx->next = (i + 1 < count) ? rx + 1 : nic->rxs; e100_rx_alloc_list()
2192 rx->prev = (i == 0) ? nic->rxs + count - 1 : rx - 1; e100_rx_alloc_list()
2193 if (e100_rx_alloc_skb(nic, rx)) { e100_rx_alloc_list()
2205 rx = nic->rxs->prev->prev; e100_rx_alloc_list()
2206 before_last = (struct rfd *)rx->skb->data; e100_rx_alloc_list()
2209 pci_dma_sync_single_for_device(nic->pdev, rx->dma_addr, e100_rx_alloc_list()
2605 netif_info(nic, drv, nic->netdev, "Ring Param settings: rx: %d, tx %d\n", e100_set_ringparam()
2912 /* D100 MAC doesn't allow rx of vlan packets with normal MTU */ e100_probe()
/linux-4.4.14/drivers/staging/nvec/
H A Dnvec.c494 if (nvec->rx->pos != nvec_msg_size(nvec->rx)) { nvec_rx_completed()
496 (uint) nvec_msg_size(nvec->rx), nvec_rx_completed()
497 (uint) nvec->rx->pos); nvec_rx_completed()
499 nvec_msg_free(nvec, nvec->rx); nvec_rx_completed()
503 if (nvec->rx->data[0] == NVEC_BAT) nvec_rx_completed()
513 list_add_tail(&nvec->rx->node, &nvec->rx_data); nvec_rx_completed()
519 if (!nvec_msg_is_event(nvec->rx)) nvec_rx_completed()
618 nvec->rx = nvec_msg_alloc(nvec, NVEC_MSG_RX); nvec_interrupt()
620 if (unlikely(nvec->rx == NULL)) { nvec_interrupt()
624 nvec->rx->data[0] = received; nvec_interrupt()
625 nvec->rx->pos = 1; nvec_interrupt()
632 if (nvec->rx->data[0] != 0x01) { nvec_interrupt()
638 nvec_msg_free(nvec, nvec->rx); nvec_interrupt()
645 BUG_ON(nvec->rx == NULL); nvec_interrupt()
646 nvec->rx->data[1] = received; nvec_interrupt()
647 nvec->rx->pos = 2; nvec_interrupt()
673 else if (nvec->rx && nvec->rx->pos < NVEC_MSG_SIZE) nvec_interrupt()
674 nvec->rx->data[nvec->rx->pos++] = received; nvec_interrupt()
678 nvec->rx, nvec->rx ? nvec->rx->pos : 0, nvec_interrupt()
840 nvec->rx = &nvec->msg_pool[0]; tegra_nvec_probe()
/linux-4.4.14/drivers/net/arcnet/
H A Darc-rawmode.c39 static void rx(struct net_device *dev, int bufnum, rx() function
75 arcnet_dump_skb(dev, skb, "rx"); rx()
156 .rx = rx,
H A Drfc1051.c40 static void rx(struct net_device *dev, int bufnum,
51 .rx = rx,
121 static void rx(struct net_device *dev, int bufnum, rx() function
154 arcnet_dump_skb(dev, skb, "rx"); rx()
/linux-4.4.14/drivers/net/wireless/realtek/rtl818x/rtl8187/
H A Dleds.h50 /* If the LED is radio or tx/rx */
/linux-4.4.14/drivers/net/ethernet/amd/xgbe/
H A Dxgbe-desc.c344 xgbe_set_buffer_data(&rdata->rx.hdr, &ring->rx_hdr_pa, xgbe_map_rx_buffer()
348 xgbe_set_buffer_data(&rdata->rx.buf, &ring->rx_buf_pa, xgbe_map_rx_buffer()
458 if (rdata->rx.hdr.pa.pages) xgbe_unmap_rdata()
459 put_page(rdata->rx.hdr.pa.pages); xgbe_unmap_rdata()
461 if (rdata->rx.hdr.pa_unmap.pages) { xgbe_unmap_rdata()
462 dma_unmap_page(pdata->dev, rdata->rx.hdr.pa_unmap.pages_dma, xgbe_unmap_rdata()
463 rdata->rx.hdr.pa_unmap.pages_len, xgbe_unmap_rdata()
465 put_page(rdata->rx.hdr.pa_unmap.pages); xgbe_unmap_rdata()
468 if (rdata->rx.buf.pa.pages) xgbe_unmap_rdata()
469 put_page(rdata->rx.buf.pa.pages); xgbe_unmap_rdata()
471 if (rdata->rx.buf.pa_unmap.pages) { xgbe_unmap_rdata()
472 dma_unmap_page(pdata->dev, rdata->rx.buf.pa_unmap.pages_dma, xgbe_unmap_rdata()
473 rdata->rx.buf.pa_unmap.pages_len, xgbe_unmap_rdata()
475 put_page(rdata->rx.buf.pa_unmap.pages); xgbe_unmap_rdata()
479 memset(&rdata->rx, 0, sizeof(rdata->rx)); xgbe_unmap_rdata()
/linux-4.4.14/arch/cris/boot/dts/include/dt-bindings/net/
H A Dti-dp83867.h27 /* RGMIIDCTL internal delay for rx and tx */
/linux-4.4.14/net/dccp/
H A Dccid.c143 struct ccid *ccid_new(const u8 id, struct sock *sk, bool rx) ccid_new() argument
151 ccid = kmem_cache_alloc(rx ? ccid_ops->ccid_hc_rx_slab : ccid_new()
156 if (rx) { ccid_new()
170 kmem_cache_free(rx ? ccid_ops->ccid_hc_rx_slab : ccid_new()
H A Dfeat.c41 static int dccp_hdlr_ccid(struct sock *sk, u64 ccid, bool rx) dccp_hdlr_ccid() argument
44 struct ccid *new_ccid = ccid_new(ccid, sk, rx); dccp_hdlr_ccid()
49 if (rx) { dccp_hdlr_ccid()
59 static int dccp_hdlr_seq_win(struct sock *sk, u64 seq_win, bool rx) dccp_hdlr_seq_win() argument
63 if (rx) { dccp_hdlr_seq_win()
75 static int dccp_hdlr_ack_ratio(struct sock *sk, u64 ratio, bool rx) dccp_hdlr_ack_ratio() argument
77 if (rx) dccp_hdlr_ack_ratio()
84 static int dccp_hdlr_ackvec(struct sock *sk, u64 enable, bool rx) dccp_hdlr_ackvec() argument
88 if (rx) { dccp_hdlr_ackvec()
101 static int dccp_hdlr_ndp(struct sock *sk, u64 enable, bool rx) dccp_hdlr_ndp() argument
103 if (!rx) dccp_hdlr_ndp()
110 * `rx' holds when the sending peer informs about his partial coverage via a
117 static int dccp_hdlr_min_cscov(struct sock *sk, u64 cscov, bool rx) dccp_hdlr_min_cscov() argument
121 if (rx) dccp_hdlr_min_cscov()
138 int (*activation_hdlr)(struct sock *sk, u64 val, bool rx);
311 bool rx; __dccp_feat_activate() local
338 rx = (is_local == (dccp_feat_table[idx].rxtx == FEAT_AT_RX)); __dccp_feat_activate()
340 dccp_debug(" -> activating %s %s, %sval=%llu\n", rx ? "RX" : "TX", __dccp_feat_activate()
344 return dccp_feat_table[idx].activation_hdlr(sk, val, rx); __dccp_feat_activate()
1448 } tx, rx; dccp_feat_init() local
1475 ccid_get_builtin_ccids(&rx.val, &rx.len)) dccp_feat_init()
1479 !dccp_feat_prefer(sysctl_dccp_rx_ccid, rx.val, rx.len)) dccp_feat_init()
1486 rc = __feat_register_sp(fn, DCCPF_CCID, false, false, rx.val, rx.len); dccp_feat_init()
1490 kfree(rx.val); dccp_feat_init()
/linux-4.4.14/include/dt-bindings/net/
H A Dti-dp83867.h27 /* RGMIIDCTL internal delay for rx and tx */
/linux-4.4.14/include/linux/can/
H A Dled.h25 /* keep space for interface name + "-tx"/"-rx"/"-rxtx"
/linux-4.4.14/arch/metag/boot/dts/include/dt-bindings/net/
H A Dti-dp83867.h27 /* RGMIIDCTL internal delay for rx and tx */
/linux-4.4.14/arch/mips/boot/dts/include/dt-bindings/net/
H A Dti-dp83867.h27 /* RGMIIDCTL internal delay for rx and tx */
/linux-4.4.14/arch/powerpc/boot/dts/include/dt-bindings/net/
H A Dti-dp83867.h27 /* RGMIIDCTL internal delay for rx and tx */
/linux-4.4.14/arch/arm64/boot/dts/include/dt-bindings/net/
H A Dti-dp83867.h27 /* RGMIIDCTL internal delay for rx and tx */
/linux-4.4.14/arch/arm/boot/dts/include/dt-bindings/net/
H A Dti-dp83867.h27 /* RGMIIDCTL internal delay for rx and tx */
/linux-4.4.14/drivers/net/wireless/ath/ath6kl/
H A Dtestmode.c55 ath6kl_warn("failed to allocate testmode rx skb!\n"); ath6kl_tm_rx_event()
66 ath6kl_warn("nla_put failed on testmode rx skb!\n"); ath6kl_tm_rx_event()
/linux-4.4.14/drivers/usb/atm/
H A Dxusbatm.c39 XUSBATM_PARM(rx_endpoint, unsigned char, byte, "rx endpoint number");
41 XUSBATM_PARM(rx_padding, unsigned char, byte, "rx padding (default 0)");
43 XUSBATM_PARM(rx_altsetting, unsigned char, byte, "rx altsetting (default 0)");
44 XUSBATM_PARM(tx_altsetting, unsigned char, byte, "rx altsetting (default 0)");
104 " rx: ep %02x padd %d alt %2d tx: ep %02x padd %d alt %2d\n", xusbatm_bind()
128 usb_dbg(usbatm, "%s: rx If#=%2d; tx If#=%2d\n", __func__, xusbatm_bind()
H A Dusbatm.h106 int bulk_in; /* bulk rx endpoint */
107 int isoc_in; /* isochronous rx endpoint */
178 unsigned char *cell_buf; /* holds partial rx cell */
/linux-4.4.14/drivers/net/wireless/libertas/
H A Drx.c89 lbs_deb_rx("rx err: frame received with bad length\n"); lbs_process_rxed_packet()
96 lbs_deb_rx("rx data: skb->len - pkt_ptr = %d-%zd = %zd\n", lbs_process_rxed_packet()
149 lbs_deb_rx("rx data: size of actual packet %d\n", skb->len); lbs_process_rxed_packet()
232 lbs_deb_rx("rx err: frame received with bad length\n"); process_rxed_802_11_packet()
239 lbs_deb_rx("rx data: skb->len-sizeof(RxPd) = %d-%zd = %zd\n", process_rxed_802_11_packet()
270 lbs_deb_rx("rx data: size of actual packet %d\n", skb->len); process_rxed_802_11_packet()
/linux-4.4.14/drivers/net/wireless/prism54/
H A Dislpci_dev.h132 * For rx queues, [index, device_curr_frag) contains fragments
135 * rx queue, waiting for data (owned by device). The driver
138 * If device_curr_frag == driver_curr_frag, no more rx buffers are
139 * available, and the rx DMA engine of the device is halted.
147 u32 index_mgmt_rx; /* real index mgmt rx queue */
149 u32 free_data_rx; /* free pointer data rx queue */
/linux-4.4.14/drivers/net/ethernet/intel/fm10k/
H A Dfm10k_debugfs.c139 if (ring < q_vector->rx.ring) fm10k_dbg_desc_open()
166 * folder, a debugfs file is created for each tx and rx ring
185 /* Generate a file for each rx ring in the q_vector */ fm10k_dbg_q_vector_init()
196 /* Generate a file for each rx ring in the q_vector */ fm10k_dbg_q_vector_init()
197 for (i = 0; i < q_vector->rx.count; i++) { fm10k_dbg_q_vector_init()
198 struct fm10k_ring *ring = &q_vector->rx.ring[i]; fm10k_dbg_q_vector_init()
/linux-4.4.14/drivers/net/ethernet/aurora/
H A Dnb8800.h182 /* DMA descriptor and associated data for rx.
193 /* Address of buffer on rx ring */
252 /* Value for rx interrupt time register in NAPI interrupt mode */
255 /* Value for rx interrupt time register in NAPI poll mode */
258 /* Value for config field of rx DMA descriptors */
302 /* DMA base address of rx descriptors, see rx_descs above */
/linux-4.4.14/drivers/usb/gadget/function/
H A Df_phonet.c49 } rx; member in struct:f_phonet
331 spin_lock_irqsave(&fp->rx.lock, flags); pn_rx_complete()
332 skb = fp->rx.skb; pn_rx_complete()
334 skb = fp->rx.skb = netdev_alloc_skb(dev, 12); pn_rx_complete()
336 fp->rx.skb = NULL; pn_rx_complete()
337 spin_unlock_irqrestore(&fp->rx.lock, flags); pn_rx_complete()
396 if (fp->rx.skb) { __pn_reset()
397 dev_kfree_skb_irq(fp->rx.skb); __pn_reset()
398 fp->rx.skb = NULL; __pn_reset()
693 spin_lock_init(&fp->rx.lock); phonet_alloc()
/linux-4.4.14/drivers/atm/
H A Dlanai.c32 * o AAL0 is stubbed in but the actual rx/tx path isn't written yet:
238 } rx; member in struct:lanai_vcc
257 unsigned atm_ovfl; /* # of cells dropped - rx fifo overflow */
700 if (lvcc->rx.atmvcc->qos.aal == ATM_AAL5) { host_vcc_start_rx()
701 dma_addr_t dmaaddr = lvcc->rx.buf.dmaaddr; host_vcc_start_rx()
709 RXADDR1_SET_SIZE(lanai_buf_size_cardorder(&lvcc->rx.buf))| host_vcc_start_rx()
1371 int m = ((const unsigned char *) lvcc->rx.buf.ptr) + n - vcc_rx_memcpy()
1372 ((const unsigned char *) (lvcc->rx.buf.end)); vcc_rx_memcpy()
1375 memcpy(dest, lvcc->rx.buf.ptr, n - m); vcc_rx_memcpy()
1376 memcpy(dest + n - m, lvcc->rx.buf.start, m); vcc_rx_memcpy()
1387 u32 *end = &lvcc->rx.buf.start[endptr * 4]; vcc_rx_aal5()
1388 int n = ((unsigned long) end) - ((unsigned long) lvcc->rx.buf.ptr); vcc_rx_aal5()
1390 n += lanai_buf_size(&lvcc->rx.buf); vcc_rx_aal5()
1391 APRINTK(n >= 0 && n < lanai_buf_size(&lvcc->rx.buf) && !(n & 15), vcc_rx_aal5()
1393 n, lanai_buf_size(&lvcc->rx.buf)); vcc_rx_aal5()
1395 if ((x = &end[-2]) < lvcc->rx.buf.start) vcc_rx_aal5()
1396 x = &lvcc->rx.buf.end[-2]; vcc_rx_aal5()
1407 lvcc->rx.atmvcc->dev->number, lvcc->vci, size, n); vcc_rx_aal5()
1411 skb = atm_alloc_charge(lvcc->rx.atmvcc, size, GFP_ATOMIC); vcc_rx_aal5()
1418 ATM_SKB(skb)->vcc = lvcc->rx.atmvcc; vcc_rx_aal5()
1420 lvcc->rx.atmvcc->push(lvcc->rx.atmvcc, skb); vcc_rx_aal5()
1421 atomic_inc(&lvcc->rx.atmvcc->stats->rx); vcc_rx_aal5()
1423 lvcc->rx.buf.ptr = end; vcc_rx_aal5()
1431 /* Remember to increment lvcc->rx.atmvcc->stats->rx */ vcc_rx_aal0()
1506 return lanai_get_sized_buffer(lanai, &lvcc->rx.buf, lanai_setup_rx_vci_aal5()
1650 if (unlikely(lvcc->rx.atmvcc == NULL)) { handle_service()
1657 if (unlikely(lvcc->rx.atmvcc->qos.aal != ATM_AAL5)) { handle_service()
1662 atomic_inc(&lvcc->rx.atmvcc->stats->rx_err); handle_service()
1673 DPRINTK("got trashed rx pdu on vci %d\n", vci); handle_service()
1674 atomic_inc(&lvcc->rx.atmvcc->stats->rx_err); handle_service()
1677 (((unsigned long) lvcc->rx.buf.ptr) - handle_service()
1678 ((unsigned long) lvcc->rx.buf.start)) + 47; handle_service()
1680 bytes += lanai_buf_size(&lvcc->rx.buf); handle_service()
1686 atomic_inc(&lvcc->rx.atmvcc->stats->rx_err); handle_service()
1693 DPRINTK("got rx crc error on vci %d\n", vci); handle_service()
1694 atomic_inc(&lvcc->rx.atmvcc->stats->rx_err); handle_service()
1696 lvcc->rx.buf.ptr = &lvcc->rx.buf.start[SERVICE_GET_END(s) * 4]; handle_service()
2002 lvcc->rx.atmvcc != NULL && lvcc->rx.atmvcc != atmvcc) vci_is_ok()
2014 if (vci0 != NULL && vci0->rx.atmvcc != NULL) vci_is_ok()
2284 if (lvcc->rx.atmvcc == atmvcc) { lanai_close()
2290 lanai_buf_deallocate(&lvcc->rx.buf, lanai->pci); lanai_close()
2291 lvcc->rx.atmvcc = NULL; lanai_close()
2341 APRINTK(lvcc->rx.atmvcc == NULL, "rx.atmvcc!=NULL, vci=%d\n", lanai_open()
2351 lvcc->rx.atmvcc = atmvcc; lanai_open()
2379 if (atmvcc == lvcc->rx.atmvcc) lanai_open()
2511 if (lvcc->rx.atmvcc != NULL) { lanai_proc_read()
2513 lvcc->rx.atmvcc->qos.aal == ATM_AAL5 ? 5 : 0); lanai_proc_read()
2514 if (lvcc->rx.atmvcc->qos.aal == ATM_AAL5) lanai_proc_read()
2518 lanai_buf_size(&lvcc->rx.buf), lanai_proc_read()
/linux-4.4.14/drivers/net/ethernet/via/
H A Dvia-velocity.c553 vptr->rx.dirty = vptr->rx.filled = vptr->rx.curr = 0; velocity_init_rx_ring_indexes()
575 vptr->rx.ring[i].rdesc0.len |= OWNED_BY_NIC; velocity_rx_reset()
578 writel(vptr->rx.pool_dma, &regs->RDBaseLo); velocity_rx_reset()
1402 writel(vptr->rx.pool_dma, &regs->RDBaseLo); velocity_init_registers()
1447 if (vptr->rx.filled < 4) velocity_give_many_rx_descs()
1452 unusable = vptr->rx.filled & 0x0003; velocity_give_many_rx_descs()
1453 dirty = vptr->rx.dirty - unusable; velocity_give_many_rx_descs()
1454 for (avail = vptr->rx.filled & 0xfffc; avail; avail--) { velocity_give_many_rx_descs()
1456 vptr->rx.ring[dirty].rdesc0.len |= OWNED_BY_NIC; velocity_give_many_rx_descs()
1459 writew(vptr->rx.filled & 0xfffc, &regs->RBRDU); velocity_give_many_rx_descs()
1460 vptr->rx.filled = unusable; velocity_give_many_rx_descs()
1493 vptr->rx.ring = pool; velocity_init_dma_rings()
1494 vptr->rx.pool_dma = pool_dma; velocity_init_dma_rings()
1511 vptr->rx.buf_sz = (mtu <= ETH_DATA_LEN) ? PKT_BUF_SZ : mtu + 32; velocity_set_rxbufsize()
1526 struct rx_desc *rd = &(vptr->rx.ring[idx]); velocity_alloc_rx_buf()
1527 struct velocity_rd_info *rd_info = &(vptr->rx.info[idx]); velocity_alloc_rx_buf()
1529 rd_info->skb = netdev_alloc_skb(vptr->netdev, vptr->rx.buf_sz + 64); velocity_alloc_rx_buf()
1540 vptr->rx.buf_sz, DMA_FROM_DEVICE); velocity_alloc_rx_buf()
1547 rd->size = cpu_to_le16(vptr->rx.buf_sz) | RX_INTEN; velocity_alloc_rx_buf()
1556 int dirty = vptr->rx.dirty, done = 0; velocity_rx_refill()
1559 struct rx_desc *rd = vptr->rx.ring + dirty; velocity_rx_refill()
1565 if (!vptr->rx.info[dirty].skb) { velocity_rx_refill()
1571 } while (dirty != vptr->rx.curr); velocity_rx_refill()
1574 vptr->rx.dirty = dirty; velocity_rx_refill()
1575 vptr->rx.filled += done; velocity_rx_refill()
1592 if (vptr->rx.info == NULL) velocity_free_rd_ring()
1596 struct velocity_rd_info *rd_info = &(vptr->rx.info[i]); velocity_free_rd_ring()
1597 struct rx_desc *rd = vptr->rx.ring + i; velocity_free_rd_ring()
1603 dma_unmap_single(vptr->dev, rd_info->skb_dma, vptr->rx.buf_sz, velocity_free_rd_ring()
1611 kfree(vptr->rx.info); velocity_free_rd_ring()
1612 vptr->rx.info = NULL; velocity_free_rd_ring()
1626 vptr->rx.info = kcalloc(vptr->options.numrx, velocity_init_rd_ring()
1628 if (!vptr->rx.info) velocity_init_rd_ring()
1685 dma_free_coherent(vptr->dev, size, vptr->rx.ring, vptr->rx.pool_dma); velocity_free_dma_rings()
2045 struct velocity_rd_info *rd_info = &(vptr->rx.info[idx]); velocity_receive_frame()
2046 struct rx_desc *rd = &(vptr->rx.ring[idx]); velocity_receive_frame()
2063 vptr->rx.buf_sz, DMA_FROM_DEVICE); velocity_receive_frame()
2070 dma_unmap_single(vptr->dev, rd_info->skb_dma, vptr->rx.buf_sz, velocity_receive_frame()
2074 vptr->rx.buf_sz, DMA_FROM_DEVICE); velocity_receive_frame()
2104 int rd_curr = vptr->rx.curr; velocity_rx_srv()
2108 struct rx_desc *rd = vptr->rx.ring + rd_curr; velocity_rx_srv()
2110 if (!vptr->rx.info[rd_curr].skb) velocity_rx_srv()
2141 vptr->rx.curr = rd_curr; velocity_rx_srv()
2158 * Do rx and tx twice for performance (taken from the VIA velocity_poll()
2305 struct rx_info rx; velocity_change_mtu() local
2331 rx = vptr->rx; velocity_change_mtu()
2334 vptr->rx = tmp_vptr->rx; velocity_change_mtu()
2337 tmp_vptr->rx = rx; velocity_change_mtu()
/linux-4.4.14/drivers/soc/qcom/
H A Dsmd.c165 * @recv_lock: guard for rx info modifications and cb pointer
230 struct smd_channel_info rx; member in struct:smd_channel_info_pair
252 struct smd_channel_info_word rx; member in struct:smd_channel_info_word_pair
257 BUILD_BUG_ON(sizeof(channel->info->rx.param) != sizeof(u8)); \
259 le32_to_cpu(channel->info_word->rx.param) : \
260 channel->info->rx.param; \
265 BUILD_BUG_ON(sizeof(channel->info->rx.param) != sizeof(u32)); \
267 channel->info_word->rx.param : \
268 channel->info->rx.param); \
273 BUILD_BUG_ON(sizeof(channel->info->rx.param) != sizeof(u8)); \
275 channel->info_word->rx.param = cpu_to_le32(value); \
277 channel->info->rx.param = value; \
282 BUILD_BUG_ON(sizeof(channel->info->rx.param) != sizeof(u32)); \
284 channel->info_word->rx.param = cpu_to_le32(value); \
286 channel->info->rx.param = cpu_to_le32(value); \
381 * Calculate the amount of data available in the rx fifo
455 * Read count bytes of data from the rx fifo into buf, but don't advance the
487 * Advance the rx tail by count bytes.
501 * Read out a single packet from the rx fifo and deliver it to the device
641 * Delivers any outstanding packets in the rx fifo, can be used after probe of
1071 /* The channel consist of a rx and tx fifo of equal size */ qcom_smd_create_channel()
/linux-4.4.14/sound/soc/fsl/
H A Dfsl_ssi.c111 struct fsl_ssi_reg_val rx; member in struct:fsl_ssi_rxtx_reg_val
369 * Enable/Disable all rx/tx config flags at once.
379 vals->rx.sier | vals->tx.sier, fsl_ssi_rxtx_config()
380 vals->rx.sier | vals->tx.sier); fsl_ssi_rxtx_config()
382 vals->rx.srcr | vals->tx.srcr, fsl_ssi_rxtx_config()
383 vals->rx.srcr | vals->tx.srcr); fsl_ssi_rxtx_config()
385 vals->rx.stcr | vals->tx.stcr, fsl_ssi_rxtx_config()
386 vals->rx.stcr | vals->tx.stcr); fsl_ssi_rxtx_config()
389 vals->rx.srcr | vals->tx.srcr, 0); fsl_ssi_rxtx_config()
391 vals->rx.stcr | vals->tx.stcr, 0); fsl_ssi_rxtx_config()
393 vals->rx.sier | vals->tx.sier, 0); fsl_ssi_rxtx_config()
419 * ssi_private->rxtx_reg_val.rx or tx as vals parameter.
440 /* Find the other direction values rx or tx which we do not want to fsl_ssi_config()
442 if (&ssi_private->rxtx_reg_val.rx == vals) fsl_ssi_config()
445 avals = &ssi_private->rxtx_reg_val.rx; fsl_ssi_config()
481 * Disabling the necessary flags for one of rx/tx while the fsl_ssi_config()
484 * streams (rx XOR tx) and that are set in the stream that is fsl_ssi_config()
511 fsl_ssi_config(ssi_private, enable, &ssi_private->rxtx_reg_val.rx); fsl_ssi_rx_config()
520 * Setup rx/tx register values used to enable/disable the streams. These will
528 reg->rx.sier = CCSR_SSI_SIER_RFF0_EN; fsl_ssi_setup_reg_vals()
529 reg->rx.srcr = CCSR_SSI_SRCR_RFEN0; fsl_ssi_setup_reg_vals()
530 reg->rx.scr = 0; fsl_ssi_setup_reg_vals()
536 reg->rx.scr = CCSR_SSI_SCR_SSIEN | CCSR_SSI_SCR_RE; fsl_ssi_setup_reg_vals()
537 reg->rx.sier |= CCSR_SSI_SIER_RFF0_EN; fsl_ssi_setup_reg_vals()
543 reg->rx.sier |= CCSR_SSI_SIER_RDMAE; fsl_ssi_setup_reg_vals()
546 reg->rx.sier |= CCSR_SSI_SIER_RIE; fsl_ssi_setup_reg_vals()
550 reg->rx.sier |= FSLSSI_SIER_DBG_RX_FLAGS; fsl_ssi_setup_reg_vals()
/linux-4.4.14/drivers/net/vmxnet3/
H A Dvmxnet3_ethtool.c81 { " LRO pkts rx", offsetof(struct UPT1_RxStats, LROPktsRxOK) },
82 { " LRO byte rx", offsetof(struct UPT1_RxStats, LROBytesRxOK) },
83 { " ucast pkts rx", offsetof(struct UPT1_RxStats, ucastPktsRxOK) },
84 { " ucast bytes rx", offsetof(struct UPT1_RxStats, ucastBytesRxOK) },
85 { " mcast pkts rx", offsetof(struct UPT1_RxStats, mcastPktsRxOK) },
86 { " mcast bytes rx", offsetof(struct UPT1_RxStats, mcastBytesRxOK) },
87 { " bcast pkts rx", offsetof(struct UPT1_RxStats, bcastPktsRxOK) },
88 { " bcast bytes rx", offsetof(struct UPT1_RxStats, bcastBytesRxOK) },
89 { " pkts rx OOB", offsetof(struct UPT1_RxStats, pktsRxOutOfBuf) },
90 { " pkts rx err", offsetof(struct UPT1_RxStats, pktsRxError) },
97 { " drv dropped rx total", offsetof(struct vmxnet3_rq_driver_stats,
103 { " rx buf alloc fail", offsetof(struct vmxnet3_rq_driver_stats,
588 /* recreate the rx queue and the tx queue based on the vmxnet3_set_ringparam()
/linux-4.4.14/drivers/net/ethernet/hisilicon/
H A Dhip04_eth.c301 /* enable tx & rx */ hip04_mac_enable()
306 /* clear rx int */ hip04_mac_enable()
328 /* disable tx & rx */ hip04_mac_disable()
469 /* disable rx interrupt and timer */ hip04_mac_start_xmit()
495 int rx = 0; hip04_rx_poll() local
528 rx++; hip04_rx_poll()
543 if (rx >= budget) hip04_rx_poll()
551 /* enable rx interrupt */ hip04_rx_poll()
559 if (rx < budget && tx_remaining) hip04_rx_poll()
562 return rx; hip04_rx_poll()
581 netdev_err(ndev, "rx drop\n"); hip04_mac_interrupt()
590 /* disable rx interrupt */ hip04_mac_interrupt()
607 /* disable rx interrupt */ tx_done()
/linux-4.4.14/arch/powerpc/include/asm/
H A Dmmu-hash64.h451 * rx = scratch register (clobbered)
453 * - rt and rx must be different registers
458 #define ASM_VSID_SCRAMBLE(rt, rx, size) \
459 lis rx,VSID_MULTIPLIER_##size@h; \
460 ori rx,rx,VSID_MULTIPLIER_##size@l; \
461 mulld rt,rt,rx; /* rt = rt * MULTIPLIER */ \
463 srdi rx,rt,VSID_BITS_##size; \
465 add rt,rt,rx; /* add high and low bits */ \
473 addi rx,rt,1; \
474 srdi rx,rx,VSID_BITS_##size; /* extract 2^VSID_BITS bit */ \
475 add rt,rt,rx
/linux-4.4.14/drivers/media/rc/
H A Dfintek-cir.h66 /* for rx */
97 /* rx settings */
185 #define CIR_CR_CLASS 0x07 /* 0xff = rx-only, 0x66 = rx + 2 tx,
186 0x33 = rx + 1 tx */
/linux-4.4.14/drivers/net/irda/
H A Dvlsi_ir.h177 VLSI_PIO_RINGPTR = 0x02, /* rx/tx ring pointer (u16, ro) */
179 VLSI_PIO_RINGSIZE = 0x06, /* rx/tx ring size (u16, rw) */
210 IRINTR_RPKTINT = 0x10, /* rx-packet transferred from fifo to memory finished */
213 IRINTR_OE_EN = 0x02, /* UART rx fifo overrun error interrupt enable */
214 IRINTR_OE_INT = 0x01 /* UART rx fifo overrun error (read LSR to clear) */
225 /* _both_ ring pointers are indices relative to the _entire_ rx,tx-ring!
227 * at RINGBASE + PTR * sizeof(descr) for rx and tx
231 #define MAX_RING_DESCR 64 /* tx, rx rings may contain up to 64 descr each */
251 * The entire ring area is split in rx and tx areas with each
253 * The rx(tx) ring is located at ringbase+0 (ringbase+64*8).
262 /* bit mask to indicate the ring size to be used for rx and tx.
269 * located at [15:12] for tx and [11:8] for rx ([7:0] unused)
276 #define TX_RX_TO_RINGSIZE(tx,rx) ((SIZE_TO_BITS(tx)<<12)|(SIZE_TO_BITS(rx)<<8))
286 * for both tx, rx rings as follows:
327 IRCFG_RXPOL = 0x0001 /* invert rx polarity when set */
522 /* descriptors for rx/tx ring
538 volatile __le16 rd_count; /* tx/rx count [11:0] */
657 /* sw descriptor rings for rx, tx:
739 * post-processing in vlsi_process_tx/rx() after it was completed
/linux-4.4.14/drivers/net/ethernet/freescale/fs_enet/
H A Dfs_enet.h156 u32 ev_napi_rx; /* mask of NAPI rx events */
157 u32 ev_napi_tx; /* mask of NAPI rx events */
158 u32 ev_rx; /* rx event mask */
162 u16 bd_rx_empty; /* mask of BD rx empty */
163 u16 bd_rx_err; /* mask of BD rx errors */
/linux-4.4.14/arch/s390/kvm/
H A Ddiag.c75 u16 rx = (vcpu->arch.sie_block->ipa & 0xf0) >> 4; __diag_page_ref_service() local
79 vcpu->run->s.regs.gprs[rx]); __diag_page_ref_service()
81 if (vcpu->run->s.regs.gprs[rx] & 7) __diag_page_ref_service()
83 rc = read_guest(vcpu, vcpu->run->s.regs.gprs[rx], rx, &parm, sizeof(parm)); __diag_page_ref_service()
/linux-4.4.14/net/l2tp/
H A Dl2tp_debugfs.c150 seq_printf(m, " %08x rx %ld/%ld/%ld rx %ld/%ld/%ld\n", l2tp_dfs_seq_tunnel_show()
207 seq_printf(m, " %hu/%hu tx %ld/%ld/%ld rx %ld/%ld/%ld\n", l2tp_dfs_seq_session_show()
229 seq_puts(m, " debug tx-pkts/bytes/errs rx-pkts/bytes/errs\n"); l2tp_dfs_seq_show()
236 seq_puts(m, " nr/ns tx-pkts/bytes/errs rx-pkts/bytes/errs\n"); l2tp_dfs_seq_show()
/linux-4.4.14/sound/soc/au1x/
H A Dac97c.c33 #define CFG_RC(x) (((x) & 0x3ff) << 13) /* valid rx slots mask */
40 #define STAT_RU (1 << 9) /* rx underflow */
41 #define STAT_RO (1 << 8) /* rx overflow */
46 #define STAT_RE (1 << 1) /* rx fifo empty */
47 #define STAT_RF (1 << 0) /* rx fifo full */
/linux-4.4.14/drivers/usb/gadget/udc/
H A Domap_udc.h77 /* rx/tx dma channels numbered 1-3 not 0-2 */
104 #define UDC_DATA_DMA (UDC_BASE + 0x48) /* rx/tx fifo addr */
106 /* rx/tx dma control, numbering channels 1-3 not 0-2 */

Completed in 5336 milliseconds

123456789