/linux-4.4.14/net/ieee802154/6lowpan/ |
H A D | Makefile | 3 ieee802154_6lowpan-y := core.o rx.o reassembly.o tx.o
|
/linux-4.4.14/arch/arm/include/debug/ |
H A D | 8250.S | 18 .macro store, rd, rx:vararg 20 str \rd, \rx 24 .macro load, rd, rx:vararg 25 ldr \rd, \rx 29 .macro store, rd, rx:vararg 30 strb \rd, \rx 33 .macro load, rd, rx:vararg 34 ldrb \rd, \rx 40 .macro senduart,rd,rx 41 store \rd, [\rx, #UART_TX << UART_SHIFT] 44 .macro busyuart,rd,rx 45 1002: load \rd, [\rx, #UART_LSR << UART_SHIFT] 51 .macro waituart,rd,rx 53 1001: load \rd, [\rx, #UART_MSR << UART_SHIFT]
|
H A D | samsung.S | 16 .macro fifo_level_s5pv210 rd, rx 17 ldr \rd, [\rx, # S3C2410_UFSTAT] 21 .macro fifo_full_s5pv210 rd, rx 22 ldr \rd, [\rx, # S3C2410_UFSTAT] 29 .macro fifo_level_s3c2440 rd, rx 30 ldr \rd, [\rx, # S3C2410_UFSTAT] 38 .macro fifo_full_s3c2440 rd, rx 39 ldr \rd, [\rx, # S3C2410_UFSTAT] 47 .macro senduart,rd,rx 48 strb \rd, [\rx, # S3C2410_UTXH] 51 .macro busyuart, rd, rx 52 ldr \rd, [\rx, # S3C2410_UFCON] 57 fifo_full \rd, \rx 63 ldr \rd, [\rx, # S3C2410_UTRSTAT] 70 .macro waituart,rd,rx 71 ldr \rd, [\rx, # S3C2410_UFCON] 76 fifo_level \rd, \rx 82 ldr \rd, [\rx, # S3C2410_UTRSTAT]
|
H A D | sirf.S | 26 .macro senduart,rd,rx 27 str \rd, [\rx, #SIRF_LLUART_TXFIFO_DATA] 30 .macro busyuart,rd,rx 33 .macro waituart,rd,rx 34 1001: ldr \rd, [\rx, #SIRF_LLUART_TXFIFO_STATUS]
|
H A D | efm32.S | 19 .macro addruart, rx, tmp, tmp2 20 ldr \rx, =(CONFIG_DEBUG_UART_PHYS) 28 str \tmp, [\rx, #UARTn_CMD] 31 .macro senduart,rd,rx 32 strb \rd, [\rx, #UARTn_TXDATA] 35 .macro waituart,rd,rx 36 1001: ldr \rd, [\rx, #UARTn_STATUS] 41 .macro busyuart,rd,rx 42 1001: ldr \rd, [\rx, UARTn_STATUS]
|
H A D | meson.S | 21 .macro senduart,rd,rx 22 str \rd, [\rx, #MESON_AO_UART_WFIFO] 25 .macro busyuart,rd,rx 26 1002: ldr \rd, [\rx, #MESON_AO_UART_STATUS] 31 .macro waituart,rd,rx 32 1001: ldr \rd, [\rx, #MESON_AO_UART_STATUS]
|
H A D | netx.S | 22 .macro senduart,rd,rx 23 str \rd, [\rx, #UART_DATA] 26 .macro busyuart,rd,rx 27 1002: ldr \rd, [\rx, #UART_FLAG] 32 .macro waituart,rd,rx 33 1001: ldr \rd, [\rx, #UART_FLAG]
|
H A D | renesas-scif.S | 35 .macro waituart, rd, rx 36 1001: ldrh \rd, [\rx, #FSR] 41 .macro senduart, rd, rx 42 strb \rd, [\rx, #FTDR] 43 ldrh \rd, [\rx, #FSR] 45 strh \rd, [\rx, #FSR] 48 .macro busyuart, rd, rx 49 1001: ldrh \rd, [\rx, #FSR]
|
H A D | digicolor.S | 24 .macro senduart,rd,rx 25 strb \rd, [\rx, #UA0_EMI_REC] 28 .macro waituart,rd,rx 31 .macro busyuart,rd,rx 32 1001: ldrb \rd, [\rx, #UA0_STATUS]
|
H A D | vt8500.S | 24 .macro senduart,rd,rx 25 strb \rd, [\rx, #0] 28 .macro busyuart,rd,rx 29 1001: ldr \rd, [\rx, #0x1c] 34 .macro waituart,rd,rx
|
H A D | at91.S | 30 .macro senduart,rd,rx 31 strb \rd, [\rx, #(AT91_DBGU_THR)] @ Write to Transmitter Holding Register 34 .macro waituart,rd,rx 35 1001: ldr \rd, [\rx, #(AT91_DBGU_SR)] @ Read Status Register 40 .macro busyuart,rd,rx 41 1001: ldr \rd, [\rx, #(AT91_DBGU_SR)] @ Read Status Register
|
H A D | ks8695.S | 26 .macro senduart, rd, rx 27 str \rd, [\rx, #KS8695_URTH] @ Write to Transmit Holding Register 30 .macro busyuart, rd, rx 31 1001: ldr \rd, [\rx, #KS8695_URLS] @ Read Line Status Register 36 .macro waituart, rd, rx 37 1001: ldr \rd, [\rx, #KS8695_URLS] @ Read Line Status Register
|
H A D | pl01x.S | 29 .macro senduart,rd,rx 30 strb \rd, [\rx, #UART01x_DR] 33 .macro waituart,rd,rx 34 1001: ldr \rd, [\rx, #UART01x_FR] 40 .macro busyuart,rd,rx 41 1001: ldr \rd, [\rx, #UART01x_FR]
|
H A D | sti.S | 47 .macro senduart,rd,rx 48 strb \rd, [\rx, #ASC_TX_BUF_OFF] 51 .macro waituart,rd,rx 52 1001: ldr \rd, [\rx, #ASC_STA_OFF] 57 .macro busyuart,rd,rx 58 1001: ldr \rd, [\rx, #ASC_STA_OFF]
|
H A D | clps711x.S | 27 .macro waituart,rd,rx 30 .macro senduart,rd,rx 31 str \rd, [\rx, #UARTDR] 34 .macro busyuart,rd,rx 35 1001: ldr \rd, [\rx, #SYSFLG]
|
H A D | imx.S | 36 .macro senduart,rd,rx 37 str \rd, [\rx, #0x40] @ TXDATA 40 .macro waituart,rd,rx 43 .macro busyuart,rd,rx 44 1002: ldr \rd, [\rx, #0x98] @ SR2
|
H A D | vf.S | 26 .macro senduart, rd, rx 27 strb \rd, [\rx, #0x7] @ Data Register 30 .macro busyuart, rd, rx 31 1001: ldrb \rd, [\rx, #0x4] @ Status Register 1 36 .macro waituart,rd,rx
|
H A D | s3c24xx.S | 24 .macro fifo_full_s3c2410 rd, rx 25 ldr \rd, [\rx, # S3C2410_UFSTAT] 29 .macro fifo_level_s3c2410 rd, rx 30 ldr \rd, [\rx, # S3C2410_UFSTAT]
|
H A D | sa1100.S | 54 .macro senduart,rd,rx 55 str \rd, [\rx, #UTDR] 58 .macro waituart,rd,rx 59 1001: ldr \rd, [\rx, #UTSR1] 64 .macro busyuart,rd,rx 65 1001: ldr \rd, [\rx, #UTSR1]
|
H A D | zynq.S | 40 .macro senduart,rd,rx 41 strb \rd, [\rx, #UART_FIFO_OFFSET] @ TXDATA 44 .macro waituart,rd,rx 45 1001: ldr \rd, [\rx, #UART_SR_OFFSET] 51 .macro busyuart,rd,rx 52 1002: ldr \rd, [\rx, #UART_SR_OFFSET] @ get status register
|
H A D | omap2plus.S | 174 .macro senduart,rd,rx 175 orr \rd, \rd, \rx, lsl #24 @ preserve LSR reg offset 176 bic \rx, \rx, #0xff @ get base (THR) reg address 177 strb \rd, [\rx] @ send lower byte of rd 178 orr \rx, \rx, \rd, lsr #24 @ restore original rx (LSR) 182 .macro busyuart,rd,rx 183 1001: ldrb \rd, [\rx] @ rx contains UART_LSR address 189 .macro waituart,rd,rx
|
H A D | tegra.S | 174 .macro senduart, rd, rx 175 cmp \rx, #0 176 strneb \rd, [\rx, #UART_TX << UART_SHIFT] 180 .macro busyuart, rd, rx 181 cmp \rx, #0 183 1001: ldrb \rd, [\rx, #UART_LSR << UART_SHIFT] 190 .macro waituart, rd, rx 192 cmp \rx, #0 194 1001: ldrb \rd, [\rx, #UART_MSR << UART_SHIFT]
|
/linux-4.4.14/arch/unicore32/kernel/ |
H A D | debug-macro.S | 17 .macro put_word_ocd, rd, rx=r16 18 1001: movc \rx, p1.c0, #0 19 cand.a \rx, #2 26 .macro addruart, rx 29 .macro senduart, rd, rx 30 put_word_ocd \rd, \rx 33 .macro busyuart, rd, rx 36 .macro waituart, rd, rx 47 .macro addruart,rx 48 mrc p0, #0, \rx, c1, c0 49 tst \rx, #1 @ MMU enabled? 50 moveq \rx, #0xee000000 @ physical base address 51 movne \rx, #0x6e000000 @ virtual address 59 str r1, [\rx, #UART_LCR_OFFSET] 62 str r1, [\rx, #UART_DLH_OFFSET] 64 str r1, [\rx, #UART_DLL_OFFSET] 66 str r1, [\rx, #UART_FCR_OFFSET] 68 str r1, [\rx, #UART_LCR_OFFSET] 70 str r1, [\rx, #UART_IER_OFFSET] 73 .macro senduart,rd,rx 74 str \rd, [\rx, #UART_THR_OFFSET] 77 .macro waituart,rd,rx 78 1001: ldr \rd, [\rx, #UART_LSR_OFFSET] 83 .macro busyuart,rd,rx 84 1001: ldr \rd, [\rx, #UART_LSR_OFFSET]
|
/linux-4.4.14/drivers/net/wimax/i2400m/ |
H A D | Makefile | 13 rx.o 21 usb-rx.o \
|
/linux-4.4.14/net/mac80211/ |
H A D | wpa.h | 19 ieee80211_rx_h_michael_mic_verify(struct ieee80211_rx_data *rx); 24 ieee80211_crypto_tkip_decrypt(struct ieee80211_rx_data *rx); 30 ieee80211_crypto_ccmp_decrypt(struct ieee80211_rx_data *rx, 38 ieee80211_crypto_aes_cmac_decrypt(struct ieee80211_rx_data *rx); 40 ieee80211_crypto_aes_cmac_256_decrypt(struct ieee80211_rx_data *rx); 44 ieee80211_crypto_aes_gmac_decrypt(struct ieee80211_rx_data *rx); 48 ieee80211_crypto_hw_decrypt(struct ieee80211_rx_data *rx); 53 ieee80211_crypto_gcmp_decrypt(struct ieee80211_rx_data *rx);
|
H A D | rx.c | 594 static void ieee80211_parse_qos(struct ieee80211_rx_data *rx) ieee80211_parse_qos() argument 596 struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)rx->skb->data; ieee80211_parse_qos() 597 struct ieee80211_rx_status *status = IEEE80211_SKB_RXCB(rx->skb); ieee80211_parse_qos() 629 rx->seqno_idx = seqno_idx; ieee80211_parse_qos() 630 rx->security_idx = security_idx; ieee80211_parse_qos() 633 rx->skb->priority = (tid > 7) ? 0 : tid; ieee80211_parse_qos() 661 static void ieee80211_verify_alignment(struct ieee80211_rx_data *rx) ieee80211_verify_alignment() argument 664 WARN_ONCE((unsigned long)rx->skb->data & 1, ieee80211_verify_alignment() 665 "unaligned packet at 0x%p\n", rx->skb->data); ieee80211_verify_alignment() 670 /* rx handlers */ 744 static ieee80211_rx_result ieee80211_rx_mesh_check(struct ieee80211_rx_data *rx) ieee80211_rx_mesh_check() argument 746 struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)rx->skb->data; ieee80211_rx_mesh_check() 747 char *dev_addr = rx->sdata->vif.addr; ieee80211_rx_mesh_check() 768 if (!rx->sta || sta_plink_state(rx->sta) != NL80211_PLINK_ESTAB) { ieee80211_rx_mesh_check() 778 if (rx->skb->len < IEEE80211_MIN_ACTION_SIZE) ieee80211_rx_mesh_check() 1027 static void ieee80211_rx_reorder_ampdu(struct ieee80211_rx_data *rx, ieee80211_rx_reorder_ampdu() argument 1030 struct sk_buff *skb = rx->skb; ieee80211_rx_reorder_ampdu() 1031 struct ieee80211_local *local = rx->local; ieee80211_rx_reorder_ampdu() 1033 struct sta_info *sta = rx->sta; ieee80211_rx_reorder_ampdu() 1043 * filter the QoS data rx stream according to ieee80211_rx_reorder_ampdu() 1073 /* if this mpdu is fragmented - terminate rx aggregation session */ ieee80211_rx_reorder_ampdu() 1077 skb_queue_tail(&rx->sdata->skb_queue, skb); ieee80211_rx_reorder_ampdu() 1078 ieee80211_queue_work(&local->hw, &rx->sdata->work); ieee80211_rx_reorder_ampdu() 1089 if (ieee80211_sta_manage_reorder_buf(rx->sdata, tid_agg_rx, skb, ieee80211_rx_reorder_ampdu() 1098 ieee80211_rx_h_check_dup(struct ieee80211_rx_data *rx) ieee80211_rx_h_check_dup() argument 1100 struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)rx->skb->data; ieee80211_rx_h_check_dup() 1101 struct ieee80211_rx_status *status = IEEE80211_SKB_RXCB(rx->skb); ieee80211_rx_h_check_dup() 1108 if (rx->skb->len < 24) ieee80211_rx_h_check_dup() 1116 if (!rx->sta) ieee80211_rx_h_check_dup() 1120 rx->sta->last_seq_ctrl[rx->seqno_idx] == hdr->seq_ctrl)) { ieee80211_rx_h_check_dup() 1121 I802_DEBUG_INC(rx->local->dot11FrameDuplicateCount); ieee80211_rx_h_check_dup() 1122 rx->sta->rx_stats.num_duplicates++; ieee80211_rx_h_check_dup() 1125 rx->sta->last_seq_ctrl[rx->seqno_idx] = hdr->seq_ctrl; ieee80211_rx_h_check_dup() 1132 ieee80211_rx_h_check(struct ieee80211_rx_data *rx) ieee80211_rx_h_check() argument 1134 struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)rx->skb->data; ieee80211_rx_h_check() 1145 if (ieee80211_vif_is_mesh(&rx->sdata->vif)) ieee80211_rx_h_check() 1146 return ieee80211_rx_mesh_check(rx); ieee80211_rx_h_check() 1150 rx->sdata->vif.type != NL80211_IFTYPE_ADHOC && ieee80211_rx_h_check() 1151 rx->sdata->vif.type != NL80211_IFTYPE_WDS && ieee80211_rx_h_check() 1152 rx->sdata->vif.type != NL80211_IFTYPE_OCB && ieee80211_rx_h_check() 1153 (!rx->sta || !test_sta_flag(rx->sta, WLAN_STA_ASSOC)))) { ieee80211_rx_h_check() 1159 if (rx->sta && rx->sdata->vif.type == NL80211_IFTYPE_STATION && ieee80211_rx_h_check() 1166 if (rx->skb->len < hdrlen + 8) ieee80211_rx_h_check() 1169 skb_copy_bits(rx->skb, hdrlen + 6, ðertype, 2); ieee80211_rx_h_check() 1170 if (ethertype == rx->sdata->control_port_protocol) ieee80211_rx_h_check() 1174 if (rx->sdata->vif.type == NL80211_IFTYPE_AP && ieee80211_rx_h_check() 1175 cfg80211_rx_spurious_frame(rx->sdata->dev, ieee80211_rx_h_check() 1188 ieee80211_rx_h_check_more_data(struct ieee80211_rx_data *rx) ieee80211_rx_h_check_more_data() argument 1194 local = rx->local; ieee80211_rx_h_check_more_data() 1195 skb = rx->skb; ieee80211_rx_h_check_more_data() 1215 ieee80211_send_pspoll(local, rx->sdata); ieee80211_rx_h_check_more_data() 1300 ieee80211_rx_h_uapsd_and_pspoll(struct ieee80211_rx_data *rx) ieee80211_rx_h_uapsd_and_pspoll() argument 1302 struct ieee80211_sub_if_data *sdata = rx->sdata; ieee80211_rx_h_uapsd_and_pspoll() 1303 struct ieee80211_hdr *hdr = (void *)rx->skb->data; ieee80211_rx_h_uapsd_and_pspoll() 1304 struct ieee80211_rx_status *status = IEEE80211_SKB_RXCB(rx->skb); ieee80211_rx_h_uapsd_and_pspoll() 1307 if (!rx->sta) ieee80211_rx_h_uapsd_and_pspoll() 1327 if (!test_sta_flag(rx->sta, WLAN_STA_PS_STA)) ieee80211_rx_h_uapsd_and_pspoll() 1331 if (!test_sta_flag(rx->sta, WLAN_STA_SP)) { ieee80211_rx_h_uapsd_and_pspoll() 1332 if (!test_sta_flag(rx->sta, WLAN_STA_PS_DRIVER)) ieee80211_rx_h_uapsd_and_pspoll() 1333 ieee80211_sta_ps_deliver_poll_response(rx->sta); ieee80211_rx_h_uapsd_and_pspoll() 1335 set_sta_flag(rx->sta, WLAN_STA_PSPOLL); ieee80211_rx_h_uapsd_and_pspoll() 1340 dev_kfree_skb(rx->skb); ieee80211_rx_h_uapsd_and_pspoll() 1358 if (!(rx->sta->sta.uapsd_queues & BIT(ac))) ieee80211_rx_h_uapsd_and_pspoll() 1362 if (test_sta_flag(rx->sta, WLAN_STA_SP)) ieee80211_rx_h_uapsd_and_pspoll() 1365 if (!test_sta_flag(rx->sta, WLAN_STA_PS_DRIVER)) ieee80211_rx_h_uapsd_and_pspoll() 1366 ieee80211_sta_ps_deliver_uapsd(rx->sta); ieee80211_rx_h_uapsd_and_pspoll() 1368 set_sta_flag(rx->sta, WLAN_STA_UAPSD); ieee80211_rx_h_uapsd_and_pspoll() 1375 ieee80211_rx_h_sta_process(struct ieee80211_rx_data *rx) ieee80211_rx_h_sta_process() argument 1377 struct sta_info *sta = rx->sta; ieee80211_rx_h_sta_process() 1378 struct sk_buff *skb = rx->skb; ieee80211_rx_h_sta_process() 1394 if (rx->sdata->vif.type == NL80211_IFTYPE_ADHOC) { ieee80211_rx_h_sta_process() 1395 u8 *bssid = ieee80211_get_bssid(hdr, rx->skb->len, ieee80211_rx_h_sta_process() 1397 if (ether_addr_equal(bssid, rx->sdata->u.ibss.bssid) && ieee80211_rx_h_sta_process() 1412 } else if (rx->sdata->vif.type == NL80211_IFTYPE_OCB) { ieee80211_rx_h_sta_process() 1428 if (rx->sdata->vif.type == NL80211_IFTYPE_STATION) ieee80211_rx_h_sta_process() 1429 ieee80211_sta_rx_notify(rx->sdata, hdr); ieee80211_rx_h_sta_process() 1432 sta->rx_stats.bytes += rx->skb->len; ieee80211_rx_h_sta_process() 1459 (rx->sdata->vif.type == NL80211_IFTYPE_AP || ieee80211_rx_h_sta_process() 1460 rx->sdata->vif.type == NL80211_IFTYPE_AP_VLAN) && ieee80211_rx_h_sta_process() 1477 if (ieee80211_vif_is_mesh(&rx->sdata->vif)) ieee80211_rx_h_sta_process() 1486 I802_DEBUG_INC(rx->local->rx_handlers_drop_nullfunc); ieee80211_rx_h_sta_process() 1495 (rx->sdata->vif.type == NL80211_IFTYPE_AP || ieee80211_rx_h_sta_process() 1496 (rx->sdata->vif.type == NL80211_IFTYPE_AP_VLAN && ieee80211_rx_h_sta_process() 1497 !rx->sdata->u.vlan.sta))) { ieee80211_rx_h_sta_process() 1500 rx->sdata->dev, sta->sta.addr, ieee80211_rx_h_sta_process() 1509 dev_kfree_skb(rx->skb); ieee80211_rx_h_sta_process() 1517 ieee80211_rx_h_decrypt(struct ieee80211_rx_data *rx) ieee80211_rx_h_decrypt() argument 1519 struct sk_buff *skb = rx->skb; ieee80211_rx_h_decrypt() 1558 rx->key = NULL; ieee80211_rx_h_decrypt() 1561 if (rx->sta) { ieee80211_rx_h_decrypt() 1562 int keyid = rx->sta->ptk_idx; ieee80211_rx_h_decrypt() 1564 if (ieee80211_has_protected(fc) && rx->sta->cipher_scheme) { ieee80211_rx_h_decrypt() 1565 cs = rx->sta->cipher_scheme; ieee80211_rx_h_decrypt() 1566 keyid = iwl80211_get_cs_keyid(cs, rx->skb); ieee80211_rx_h_decrypt() 1570 sta_ptk = rcu_dereference(rx->sta->ptk[keyid]); ieee80211_rx_h_decrypt() 1574 mmie_keyidx = ieee80211_get_mmie_keyidx(rx->skb); ieee80211_rx_h_decrypt() 1577 rx->key = sta_ptk; ieee80211_rx_h_decrypt() 1593 if (rx->sta) ieee80211_rx_h_decrypt() 1594 rx->key = rcu_dereference(rx->sta->gtk[mmie_keyidx]); ieee80211_rx_h_decrypt() 1595 if (!rx->key) ieee80211_rx_h_decrypt() 1596 rx->key = rcu_dereference(rx->sdata->keys[mmie_keyidx]); ieee80211_rx_h_decrypt() 1600 * need to set rx->key if there is a key that could have been ieee80211_rx_h_decrypt() 1605 struct ieee80211_sub_if_data *sdata = rx->sdata; ieee80211_rx_h_decrypt() 1610 (key = rcu_dereference(rx->sdata->default_mgmt_key))) ieee80211_rx_h_decrypt() 1611 rx->key = key; ieee80211_rx_h_decrypt() 1613 if (rx->sta) { ieee80211_rx_h_decrypt() 1615 key = rcu_dereference(rx->sta->gtk[i]); ieee80211_rx_h_decrypt() 1628 rx->key = key; ieee80211_rx_h_decrypt() 1650 keyidx = iwl80211_get_cs_keyid(cs, rx->skb); ieee80211_rx_h_decrypt() 1655 if (rx->skb->len < 8 + hdrlen) ieee80211_rx_h_decrypt() 1661 skb_copy_bits(rx->skb, hdrlen + 3, &keyid, 1); ieee80211_rx_h_decrypt() 1666 if (is_multicast_ether_addr(hdr->addr1) && rx->sta) ieee80211_rx_h_decrypt() 1667 rx->key = rcu_dereference(rx->sta->gtk[keyidx]); ieee80211_rx_h_decrypt() 1670 if (!rx->key) { ieee80211_rx_h_decrypt() 1671 rx->key = rcu_dereference(rx->sdata->keys[keyidx]); ieee80211_rx_h_decrypt() 1678 if (rx->key && ieee80211_rx_h_decrypt() 1679 rx->key->conf.cipher != WLAN_CIPHER_SUITE_WEP40 && ieee80211_rx_h_decrypt() 1680 rx->key->conf.cipher != WLAN_CIPHER_SUITE_WEP104 && ieee80211_rx_h_decrypt() 1682 rx->key = NULL; ieee80211_rx_h_decrypt() 1686 if (rx->key) { ieee80211_rx_h_decrypt() 1687 if (unlikely(rx->key->flags & KEY_FLAG_TAINTED)) ieee80211_rx_h_decrypt() 1695 switch (rx->key->conf.cipher) { ieee80211_rx_h_decrypt() 1698 result = ieee80211_crypto_wep_decrypt(rx); ieee80211_rx_h_decrypt() 1701 result = ieee80211_crypto_tkip_decrypt(rx); ieee80211_rx_h_decrypt() 1705 rx, IEEE80211_CCMP_MIC_LEN); ieee80211_rx_h_decrypt() 1709 rx, IEEE80211_CCMP_256_MIC_LEN); ieee80211_rx_h_decrypt() 1712 result = ieee80211_crypto_aes_cmac_decrypt(rx); ieee80211_rx_h_decrypt() 1715 result = ieee80211_crypto_aes_cmac_256_decrypt(rx); ieee80211_rx_h_decrypt() 1719 result = ieee80211_crypto_aes_gmac_decrypt(rx); ieee80211_rx_h_decrypt() 1723 result = ieee80211_crypto_gcmp_decrypt(rx); ieee80211_rx_h_decrypt() 1726 result = ieee80211_crypto_hw_decrypt(rx); ieee80211_rx_h_decrypt() 1807 ieee80211_rx_h_defragment(struct ieee80211_rx_data *rx) ieee80211_rx_h_defragment() argument 1817 hdr = (struct ieee80211_hdr *)rx->skb->data; ieee80211_rx_h_defragment() 1827 I802_DEBUG_INC(rx->local->dot11MulticastReceivedFrameCount); ieee80211_rx_h_defragment() 1834 I802_DEBUG_INC(rx->local->rx_handlers_fragments); ieee80211_rx_h_defragment() 1836 if (skb_linearize(rx->skb)) ieee80211_rx_h_defragment() 1844 hdr = (struct ieee80211_hdr *)rx->skb->data; ieee80211_rx_h_defragment() 1849 entry = ieee80211_reassemble_add(rx->sdata, frag, seq, ieee80211_rx_h_defragment() 1850 rx->seqno_idx, &(rx->skb)); ieee80211_rx_h_defragment() 1851 if (rx->key && ieee80211_rx_h_defragment() 1852 (rx->key->conf.cipher == WLAN_CIPHER_SUITE_CCMP || ieee80211_rx_h_defragment() 1853 rx->key->conf.cipher == WLAN_CIPHER_SUITE_CCMP_256 || ieee80211_rx_h_defragment() 1854 rx->key->conf.cipher == WLAN_CIPHER_SUITE_GCMP || ieee80211_rx_h_defragment() 1855 rx->key->conf.cipher == WLAN_CIPHER_SUITE_GCMP_256) && ieee80211_rx_h_defragment() 1857 int queue = rx->security_idx; ieee80211_rx_h_defragment() 1864 rx->key->u.ccmp.rx_pn[queue], ieee80211_rx_h_defragment() 1870 BUILD_BUG_ON(sizeof(rx->key->u.ccmp.rx_pn[queue]) != ieee80211_rx_h_defragment() 1871 sizeof(rx->key->u.gcmp.rx_pn[queue])); ieee80211_rx_h_defragment() 1881 entry = ieee80211_reassemble_find(rx->sdata, frag, seq, ieee80211_rx_h_defragment() 1882 rx->seqno_idx, hdr); ieee80211_rx_h_defragment() 1884 I802_DEBUG_INC(rx->local->rx_handlers_drop_defrag); ieee80211_rx_h_defragment() 1898 if (!rx->key || ieee80211_rx_h_defragment() 1899 (rx->key->conf.cipher != WLAN_CIPHER_SUITE_CCMP && ieee80211_rx_h_defragment() 1900 rx->key->conf.cipher != WLAN_CIPHER_SUITE_CCMP_256 && ieee80211_rx_h_defragment() 1901 rx->key->conf.cipher != WLAN_CIPHER_SUITE_GCMP && ieee80211_rx_h_defragment() 1902 rx->key->conf.cipher != WLAN_CIPHER_SUITE_GCMP_256)) ieee80211_rx_h_defragment() 1910 queue = rx->security_idx; ieee80211_rx_h_defragment() 1911 rpn = rx->key->u.ccmp.rx_pn[queue]; ieee80211_rx_h_defragment() 1917 skb_pull(rx->skb, ieee80211_hdrlen(fc)); ieee80211_rx_h_defragment() 1918 __skb_queue_tail(&entry->skb_list, rx->skb); ieee80211_rx_h_defragment() 1920 entry->extra_len += rx->skb->len; ieee80211_rx_h_defragment() 1922 rx->skb = NULL; ieee80211_rx_h_defragment() 1926 rx->skb = __skb_dequeue(&entry->skb_list); ieee80211_rx_h_defragment() 1927 if (skb_tailroom(rx->skb) < entry->extra_len) { ieee80211_rx_h_defragment() 1928 I802_DEBUG_INC(rx->local->rx_expand_skb_head_defrag); ieee80211_rx_h_defragment() 1929 if (unlikely(pskb_expand_head(rx->skb, 0, entry->extra_len, ieee80211_rx_h_defragment() 1931 I802_DEBUG_INC(rx->local->rx_handlers_drop_defrag); ieee80211_rx_h_defragment() 1937 memcpy(skb_put(rx->skb, skb->len), skb->data, skb->len); ieee80211_rx_h_defragment() 1942 status = IEEE80211_SKB_RXCB(rx->skb); ieee80211_rx_h_defragment() 1945 ieee80211_led_rx(rx->local); ieee80211_rx_h_defragment() 1947 if (rx->sta) ieee80211_rx_h_defragment() 1948 rx->sta->rx_stats.packets++; ieee80211_rx_h_defragment() 1952 static int ieee80211_802_1x_port_control(struct ieee80211_rx_data *rx) ieee80211_802_1x_port_control() argument 1954 if (unlikely(!rx->sta || !test_sta_flag(rx->sta, WLAN_STA_AUTHORIZED))) ieee80211_802_1x_port_control() 1960 static int ieee80211_drop_unencrypted(struct ieee80211_rx_data *rx, __le16 fc) ieee80211_drop_unencrypted() argument 1962 struct sk_buff *skb = rx->skb; ieee80211_drop_unencrypted() 1975 ieee80211_is_data(fc) && rx->key)) ieee80211_drop_unencrypted() 1981 static int ieee80211_drop_unencrypted_mgmt(struct ieee80211_rx_data *rx) ieee80211_drop_unencrypted_mgmt() argument 1983 struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)rx->skb->data; ieee80211_drop_unencrypted_mgmt() 1984 struct ieee80211_rx_status *status = IEEE80211_SKB_RXCB(rx->skb); ieee80211_drop_unencrypted_mgmt() 1994 if (rx->sta && test_sta_flag(rx->sta, WLAN_STA_MFP)) { ieee80211_drop_unencrypted_mgmt() 1996 ieee80211_is_unicast_robust_mgmt_frame(rx->skb) && ieee80211_drop_unencrypted_mgmt() 1997 rx->key)) { ieee80211_drop_unencrypted_mgmt() 2000 cfg80211_rx_unprot_mlme_mgmt(rx->sdata->dev, ieee80211_drop_unencrypted_mgmt() 2001 rx->skb->data, ieee80211_drop_unencrypted_mgmt() 2002 rx->skb->len); ieee80211_drop_unencrypted_mgmt() 2006 if (unlikely(ieee80211_is_multicast_robust_mgmt_frame(rx->skb) && ieee80211_drop_unencrypted_mgmt() 2007 ieee80211_get_mmie_keyidx(rx->skb) < 0)) { ieee80211_drop_unencrypted_mgmt() 2010 cfg80211_rx_unprot_mlme_mgmt(rx->sdata->dev, ieee80211_drop_unencrypted_mgmt() 2011 rx->skb->data, ieee80211_drop_unencrypted_mgmt() 2012 rx->skb->len); ieee80211_drop_unencrypted_mgmt() 2019 if (unlikely(ieee80211_is_action(fc) && !rx->key && ieee80211_drop_unencrypted_mgmt() 2020 ieee80211_is_robust_mgmt_frame(rx->skb))) ieee80211_drop_unencrypted_mgmt() 2028 __ieee80211_data_to_8023(struct ieee80211_rx_data *rx, bool *port_control) __ieee80211_data_to_8023() argument 2030 struct ieee80211_sub_if_data *sdata = rx->sdata; __ieee80211_data_to_8023() 2031 struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)rx->skb->data; __ieee80211_data_to_8023() 2054 ret = ieee80211_data_to_8023(rx->skb, sdata->vif.addr, sdata->vif.type); __ieee80211_data_to_8023() 2058 ehdr = (struct ethhdr *) rx->skb->data; __ieee80211_data_to_8023() 2059 if (ehdr->h_proto == rx->sdata->control_port_protocol) __ieee80211_data_to_8023() 2068 * requires that rx->skb is a frame with ethernet header 2070 static bool ieee80211_frame_allowed(struct ieee80211_rx_data *rx, __le16 fc) ieee80211_frame_allowed() argument 2074 struct ethhdr *ehdr = (struct ethhdr *) rx->skb->data; ieee80211_frame_allowed() 2080 if (ehdr->h_proto == rx->sdata->control_port_protocol && ieee80211_frame_allowed() 2081 (ether_addr_equal(ehdr->h_dest, rx->sdata->vif.addr) || ieee80211_frame_allowed() 2085 if (ieee80211_802_1x_port_control(rx) || ieee80211_frame_allowed() 2086 ieee80211_drop_unencrypted(rx, fc)) ieee80211_frame_allowed() 2093 * requires that rx->skb is a frame with ethernet header 2096 ieee80211_deliver_skb(struct ieee80211_rx_data *rx) ieee80211_deliver_skb() argument 2098 struct ieee80211_sub_if_data *sdata = rx->sdata; ieee80211_deliver_skb() 2101 struct ethhdr *ehdr = (struct ethhdr *) rx->skb->data; ieee80211_deliver_skb() 2104 skb = rx->skb; ieee80211_deliver_skb() 2167 if (rx->napi) ieee80211_deliver_skb() 2168 napi_gro_receive(rx->napi, skb); ieee80211_deliver_skb() 2188 ieee80211_rx_h_amsdu(struct ieee80211_rx_data *rx) ieee80211_rx_h_amsdu() argument 2190 struct net_device *dev = rx->sdata->dev; ieee80211_rx_h_amsdu() 2191 struct sk_buff *skb = rx->skb; ieee80211_rx_h_amsdu() 2195 struct ieee80211_rx_status *status = IEEE80211_SKB_RXCB(rx->skb); ieee80211_rx_h_amsdu() 2207 rx->sdata->vif.type == NL80211_IFTYPE_AP_VLAN && ieee80211_rx_h_amsdu() 2208 !rx->sdata->u.vlan.sta) ieee80211_rx_h_amsdu() 2212 ((rx->sdata->vif.type == NL80211_IFTYPE_AP_VLAN && ieee80211_rx_h_amsdu() 2213 rx->sdata->u.vlan.sta) || ieee80211_rx_h_amsdu() 2214 (rx->sdata->vif.type == NL80211_IFTYPE_STATION && ieee80211_rx_h_amsdu() 2215 rx->sdata->u.mgd.use_4addr))) ieee80211_rx_h_amsdu() 2225 rx->sdata->vif.type, ieee80211_rx_h_amsdu() 2226 rx->local->hw.extra_tx_headroom, true); ieee80211_rx_h_amsdu() 2229 rx->skb = __skb_dequeue(&frame_list); ieee80211_rx_h_amsdu() 2231 if (!ieee80211_frame_allowed(rx, fc)) { ieee80211_rx_h_amsdu() 2232 dev_kfree_skb(rx->skb); ieee80211_rx_h_amsdu() 2236 ieee80211_deliver_skb(rx); ieee80211_rx_h_amsdu() 2244 ieee80211_rx_h_mesh_fwding(struct ieee80211_rx_data *rx) ieee80211_rx_h_mesh_fwding() argument 2249 struct sk_buff *skb = rx->skb, *fwd_skb; ieee80211_rx_h_mesh_fwding() 2250 struct ieee80211_local *local = rx->local; ieee80211_rx_h_mesh_fwding() 2251 struct ieee80211_sub_if_data *sdata = rx->sdata; ieee80211_rx_h_mesh_fwding() 2259 if (!pskb_may_pull(rx->skb, hdrlen + 6)) ieee80211_rx_h_mesh_fwding() 2265 if (!pskb_may_pull(rx->skb, ieee80211_rx_h_mesh_fwding() 2273 if (ieee80211_drop_unencrypted(rx, hdr->frame_control)) ieee80211_rx_h_mesh_fwding() 2279 mesh_rmc_check(rx->sdata, hdr->addr3, mesh_hdr)) ieee80211_rx_h_mesh_fwding() 2350 info->control.vif = &rx->sdata->vif; ieee80211_rx_h_mesh_fwding() 2381 ieee80211_rx_h_data(struct ieee80211_rx_data *rx) ieee80211_rx_h_data() argument 2383 struct ieee80211_sub_if_data *sdata = rx->sdata; ieee80211_rx_h_data() 2384 struct ieee80211_local *local = rx->local; ieee80211_rx_h_data() 2386 struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)rx->skb->data; ieee80211_rx_h_data() 2397 if (rx->sta) { ieee80211_rx_h_data() 2403 rx->sta->rx_stats.msdu[rx->seqno_idx]++; ieee80211_rx_h_data() 2412 if (rx->sta && ieee80211_rx_h_data() 2413 !test_and_set_sta_flag(rx->sta, WLAN_STA_4ADDR_EVENT)) ieee80211_rx_h_data() 2415 rx->sdata->dev, rx->sta->sta.addr, GFP_ATOMIC); ieee80211_rx_h_data() 2419 err = __ieee80211_data_to_8023(rx, &port_control); ieee80211_rx_h_data() 2423 if (!ieee80211_frame_allowed(rx, fc)) ieee80211_rx_h_data() 2427 if (unlikely(((struct ethhdr *)rx->skb->data)->h_proto == ieee80211_rx_h_data() 2429 struct ieee80211_tdls_data *tf = (void *)rx->skb->data; ieee80211_rx_h_data() 2431 if (pskb_may_pull(rx->skb, ieee80211_rx_h_data() 2437 skb_queue_tail(&local->skb_queue_tdls_chsw, rx->skb); ieee80211_rx_h_data() 2439 if (rx->sta) ieee80211_rx_h_data() 2440 rx->sta->rx_stats.packets++; ieee80211_rx_h_data() 2446 if (rx->sdata->vif.type == NL80211_IFTYPE_AP_VLAN && ieee80211_rx_h_data() 2451 rx->sdata = sdata; ieee80211_rx_h_data() 2454 rx->skb->dev = dev; ieee80211_rx_h_data() 2458 ((struct ethhdr *)rx->skb->data)->h_dest) && ieee80211_rx_h_data() 2465 ieee80211_deliver_skb(rx); ieee80211_rx_h_data() 2471 ieee80211_rx_h_ctrl(struct ieee80211_rx_data *rx, struct sk_buff_head *frames) ieee80211_rx_h_ctrl() argument 2473 struct sk_buff *skb = rx->skb; ieee80211_rx_h_ctrl() 2490 if (!rx->sta) ieee80211_rx_h_ctrl() 2499 tid_agg_rx = rcu_dereference(rx->sta->ampdu_mlme.tid_rx[tid]); ieee80211_rx_h_ctrl() 2506 event.u.ba.sta = &rx->sta->sta; ieee80211_rx_h_ctrl() 2515 ieee80211_release_reorder_frames(rx->sdata, tid_agg_rx, ieee80211_rx_h_ctrl() 2519 drv_event_callback(rx->local, rx->sdata, &event); ieee80211_rx_h_ctrl() 2580 ieee80211_rx_h_mgmt_check(struct ieee80211_rx_data *rx) ieee80211_rx_h_mgmt_check() argument 2582 struct ieee80211_mgmt *mgmt = (struct ieee80211_mgmt *) rx->skb->data; ieee80211_rx_h_mgmt_check() 2583 struct ieee80211_rx_status *status = IEEE80211_SKB_RXCB(rx->skb); ieee80211_rx_h_mgmt_check() 2590 if (rx->skb->len < 24) ieee80211_rx_h_mgmt_check() 2596 if (rx->sdata->vif.type == NL80211_IFTYPE_AP && ieee80211_rx_h_mgmt_check() 2598 !(rx->flags & IEEE80211_RX_BEACON_REPORTED)) { ieee80211_rx_h_mgmt_check() 2601 if (ieee80211_hw_check(&rx->local->hw, SIGNAL_DBM)) ieee80211_rx_h_mgmt_check() 2604 cfg80211_report_obss_beacon(rx->local->hw.wiphy, ieee80211_rx_h_mgmt_check() 2605 rx->skb->data, rx->skb->len, ieee80211_rx_h_mgmt_check() 2607 rx->flags |= IEEE80211_RX_BEACON_REPORTED; ieee80211_rx_h_mgmt_check() 2610 if (ieee80211_drop_unencrypted_mgmt(rx)) ieee80211_rx_h_mgmt_check() 2617 ieee80211_rx_h_action(struct ieee80211_rx_data *rx) ieee80211_rx_h_action() argument 2619 struct ieee80211_local *local = rx->local; ieee80211_rx_h_action() 2620 struct ieee80211_sub_if_data *sdata = rx->sdata; ieee80211_rx_h_action() 2621 struct ieee80211_mgmt *mgmt = (struct ieee80211_mgmt *) rx->skb->data; ieee80211_rx_h_action() 2622 struct ieee80211_rx_status *status = IEEE80211_SKB_RXCB(rx->skb); ieee80211_rx_h_action() 2623 int len = rx->skb->len; ieee80211_rx_h_action() 2632 if (!rx->sta && mgmt->u.action.category != WLAN_CATEGORY_PUBLIC && ieee80211_rx_h_action() 2640 if (!rx->sta->sta.ht_cap.ht_supported) ieee80211_rx_h_action() 2675 if (rx->sta->sta.smps_mode == smps_mode) ieee80211_rx_h_action() 2677 rx->sta->sta.smps_mode = smps_mode; ieee80211_rx_h_action() 2679 sband = rx->local->hw.wiphy->bands[status->band]; ieee80211_rx_h_action() 2681 rate_control_rate_update(local, sband, rx->sta, ieee80211_rx_h_action() 2691 if (!(rx->sta->sta.ht_cap.cap & ieee80211_rx_h_action() 2698 max_bw = ieee80211_sta_cap_rx_bw(rx->sta); ieee80211_rx_h_action() 2701 rx->sta->cur_max_bandwidth = max_bw; ieee80211_rx_h_action() 2702 new_bw = ieee80211_sta_cur_vht_bw(rx->sta); ieee80211_rx_h_action() 2704 if (rx->sta->sta.bandwidth == new_bw) ieee80211_rx_h_action() 2707 rx->sta->sta.bandwidth = new_bw; ieee80211_rx_h_action() 2708 sband = rx->local->hw.wiphy->bands[status->band]; ieee80211_rx_h_action() 2710 rate_control_rate_update(local, sband, rx->sta, ieee80211_rx_h_action() 2724 if (!rx->sta) ieee80211_rx_h_action() 2757 ieee80211_vht_handle_opmode(rx->sdata, rx->sta, ieee80211_rx_h_action() 2900 if (rx->sta) ieee80211_rx_h_action() 2901 rx->sta->rx_stats.packets++; ieee80211_rx_h_action() 2902 dev_kfree_skb(rx->skb); ieee80211_rx_h_action() 2906 rx->skb->pkt_type = IEEE80211_SDATA_QUEUE_TYPE_FRAME; ieee80211_rx_h_action() 2907 skb_queue_tail(&sdata->skb_queue, rx->skb); ieee80211_rx_h_action() 2909 if (rx->sta) ieee80211_rx_h_action() 2910 rx->sta->rx_stats.packets++; ieee80211_rx_h_action() 2915 ieee80211_rx_h_userspace_mgmt(struct ieee80211_rx_data *rx) ieee80211_rx_h_userspace_mgmt() argument 2917 struct ieee80211_rx_status *status = IEEE80211_SKB_RXCB(rx->skb); ieee80211_rx_h_userspace_mgmt() 2931 if (ieee80211_hw_check(&rx->local->hw, SIGNAL_DBM)) ieee80211_rx_h_userspace_mgmt() 2934 if (cfg80211_rx_mgmt(&rx->sdata->wdev, status->freq, sig, ieee80211_rx_h_userspace_mgmt() 2935 rx->skb->data, rx->skb->len, 0)) { ieee80211_rx_h_userspace_mgmt() 2936 if (rx->sta) ieee80211_rx_h_userspace_mgmt() 2937 rx->sta->rx_stats.packets++; ieee80211_rx_h_userspace_mgmt() 2938 dev_kfree_skb(rx->skb); ieee80211_rx_h_userspace_mgmt() 2946 ieee80211_rx_h_action_return(struct ieee80211_rx_data *rx) ieee80211_rx_h_action_return() argument 2948 struct ieee80211_local *local = rx->local; ieee80211_rx_h_action_return() 2949 struct ieee80211_mgmt *mgmt = (struct ieee80211_mgmt *) rx->skb->data; ieee80211_rx_h_action_return() 2951 struct ieee80211_sub_if_data *sdata = rx->sdata; ieee80211_rx_h_action_return() 2952 struct ieee80211_rx_status *status = IEEE80211_SKB_RXCB(rx->skb); ieee80211_rx_h_action_return() 2979 nskb = skb_copy_expand(rx->skb, local->hw.extra_tx_headroom, 0, ieee80211_rx_h_action_return() 2986 memcpy(nmgmt->sa, rx->sdata->vif.addr, ETH_ALEN); ieee80211_rx_h_action_return() 2990 if (rx->sdata->vif.type == NL80211_IFTYPE_P2P_DEVICE) { ieee80211_rx_h_action_return() 3001 __ieee80211_tx_skb_tid_band(rx->sdata, nskb, 7, ieee80211_rx_h_action_return() 3004 dev_kfree_skb(rx->skb); ieee80211_rx_h_action_return() 3009 ieee80211_rx_h_mgmt(struct ieee80211_rx_data *rx) ieee80211_rx_h_mgmt() argument 3011 struct ieee80211_sub_if_data *sdata = rx->sdata; ieee80211_rx_h_mgmt() 3012 struct ieee80211_mgmt *mgmt = (void *)rx->skb->data; ieee80211_rx_h_mgmt() 3052 rx->skb->pkt_type = IEEE80211_SDATA_QUEUE_TYPE_FRAME; ieee80211_rx_h_mgmt() 3053 skb_queue_tail(&sdata->skb_queue, rx->skb); ieee80211_rx_h_mgmt() 3054 ieee80211_queue_work(&rx->local->hw, &sdata->work); ieee80211_rx_h_mgmt() 3055 if (rx->sta) ieee80211_rx_h_mgmt() 3056 rx->sta->rx_stats.packets++; ieee80211_rx_h_mgmt() 3061 static void ieee80211_rx_cooked_monitor(struct ieee80211_rx_data *rx, ieee80211_rx_cooked_monitor() argument 3065 struct ieee80211_local *local = rx->local; ieee80211_rx_cooked_monitor() 3066 struct sk_buff *skb = rx->skb, *skb2; ieee80211_rx_cooked_monitor() 3075 if (rx->flags & IEEE80211_RX_CMNTR) ieee80211_rx_cooked_monitor() 3077 rx->flags |= IEEE80211_RX_CMNTR; ieee80211_rx_cooked_monitor() 3131 static void ieee80211_rx_handlers_result(struct ieee80211_rx_data *rx, ieee80211_rx_handlers_result() argument 3136 I802_DEBUG_INC(rx->sdata->local->rx_handlers_drop); ieee80211_rx_handlers_result() 3137 if (rx->sta) ieee80211_rx_handlers_result() 3138 rx->sta->rx_stats.dropped++; ieee80211_rx_handlers_result() 3145 status = IEEE80211_SKB_RXCB((rx->skb)); ieee80211_rx_handlers_result() 3147 sband = rx->local->hw.wiphy->bands[status->band]; ieee80211_rx_handlers_result() 3152 ieee80211_rx_cooked_monitor(rx, rate); ieee80211_rx_handlers_result() 3156 I802_DEBUG_INC(rx->sdata->local->rx_handlers_drop); ieee80211_rx_handlers_result() 3157 if (rx->sta) ieee80211_rx_handlers_result() 3158 rx->sta->rx_stats.dropped++; ieee80211_rx_handlers_result() 3159 dev_kfree_skb(rx->skb); ieee80211_rx_handlers_result() 3162 I802_DEBUG_INC(rx->sdata->local->rx_handlers_queued); ieee80211_rx_handlers_result() 3167 static void ieee80211_rx_handlers(struct ieee80211_rx_data *rx, ieee80211_rx_handlers() argument 3175 res = rxh(rx); \ ieee80211_rx_handlers() 3186 spin_lock_bh(&rx->local->rx_path_lock); ieee80211_rx_handlers() 3194 rx->skb = skb; ieee80211_rx_handlers() 3204 if (ieee80211_vif_is_mesh(&rx->sdata->vif)) ieee80211_rx_handlers() 3211 res = ieee80211_rx_h_ctrl(rx, frames); ieee80211_rx_handlers() 3222 ieee80211_rx_handlers_result(rx, res); ieee80211_rx_handlers() 3227 spin_unlock_bh(&rx->local->rx_path_lock); ieee80211_rx_handlers() 3230 static void ieee80211_invoke_rx_handlers(struct ieee80211_rx_data *rx) ieee80211_invoke_rx_handlers() argument 3239 res = rxh(rx); \ ieee80211_invoke_rx_handlers() 3247 ieee80211_rx_reorder_ampdu(rx, &reorder_release); ieee80211_invoke_rx_handlers() 3249 ieee80211_rx_handlers(rx, &reorder_release); ieee80211_invoke_rx_handlers() 3253 ieee80211_rx_handlers_result(rx, res); ieee80211_invoke_rx_handlers() 3265 struct ieee80211_rx_data rx = { ieee80211_release_reorder_timeout() local 3292 drv_event_callback(rx.local, rx.sdata, &event); ieee80211_release_reorder_timeout() 3295 ieee80211_rx_handlers(&rx, &frames); ieee80211_release_reorder_timeout() 3300 static bool ieee80211_accept_frame(struct ieee80211_rx_data *rx) ieee80211_accept_frame() argument 3302 struct ieee80211_sub_if_data *sdata = rx->sdata; ieee80211_accept_frame() 3303 struct sk_buff *skb = rx->skb; ieee80211_accept_frame() 3329 if (!rx->sta) { ieee80211_accept_frame() 3349 if (!rx->sta) { ieee80211_accept_frame() 3417 static bool ieee80211_prepare_and_rx_handle(struct ieee80211_rx_data *rx, ieee80211_prepare_and_rx_handle() argument 3420 struct ieee80211_local *local = rx->local; ieee80211_prepare_and_rx_handle() 3421 struct ieee80211_sub_if_data *sdata = rx->sdata; ieee80211_prepare_and_rx_handle() 3423 rx->skb = skb; ieee80211_prepare_and_rx_handle() 3425 if (!ieee80211_accept_frame(rx)) ieee80211_prepare_and_rx_handle() 3438 rx->skb = skb; ieee80211_prepare_and_rx_handle() 3441 ieee80211_invoke_rx_handlers(rx); ieee80211_prepare_and_rx_handle() 3457 struct ieee80211_rx_data rx; __ieee80211_rx_handle_packet() local 3464 memset(&rx, 0, sizeof(rx)); __ieee80211_rx_handle_packet() 3465 rx.skb = skb; __ieee80211_rx_handle_packet() 3466 rx.local = local; __ieee80211_rx_handle_packet() 3467 rx.napi = napi; __ieee80211_rx_handle_packet() 3488 ieee80211_parse_qos(&rx); __ieee80211_rx_handle_packet() 3489 ieee80211_verify_alignment(&rx); __ieee80211_rx_handle_packet() 3508 rx.sta = prev_sta; __ieee80211_rx_handle_packet() 3509 rx.sdata = prev_sta->sdata; __ieee80211_rx_handle_packet() 3510 ieee80211_prepare_and_rx_handle(&rx, skb, false); __ieee80211_rx_handle_packet() 3516 rx.sta = prev_sta; __ieee80211_rx_handle_packet() 3517 rx.sdata = prev_sta->sdata; __ieee80211_rx_handle_packet() 3519 if (ieee80211_prepare_and_rx_handle(&rx, skb, true)) __ieee80211_rx_handle_packet() 3546 rx.sta = sta_info_get_bss(prev, hdr->addr2); __ieee80211_rx_handle_packet() 3547 rx.sdata = prev; __ieee80211_rx_handle_packet() 3548 ieee80211_prepare_and_rx_handle(&rx, skb, false); __ieee80211_rx_handle_packet() 3554 rx.sta = sta_info_get_bss(prev, hdr->addr2); __ieee80211_rx_handle_packet() 3555 rx.sdata = prev; __ieee80211_rx_handle_packet() 3557 if (ieee80211_prepare_and_rx_handle(&rx, skb, true)) __ieee80211_rx_handle_packet() 3679 /* This is a version of the rx handler that can be called from hard irq
|
H A D | wpa.c | 87 ieee80211_rx_h_michael_mic_verify(struct ieee80211_rx_data *rx) ieee80211_rx_h_michael_mic_verify() argument 93 struct sk_buff *skb = rx->skb; ieee80211_rx_h_michael_mic_verify() 114 if (!(status->flag & RX_FLAG_IV_STRIPPED) && rx->key && ieee80211_rx_h_michael_mic_verify() 115 rx->key->conf.cipher == WLAN_CIPHER_SUITE_TKIP) ieee80211_rx_h_michael_mic_verify() 126 if (!rx->key || rx->key->conf.cipher != WLAN_CIPHER_SUITE_TKIP || ieee80211_rx_h_michael_mic_verify() 130 if (rx->sdata->vif.type == NL80211_IFTYPE_AP && rx->key->conf.keyidx) { ieee80211_rx_h_michael_mic_verify() 147 if (skb_linearize(rx->skb)) ieee80211_rx_h_michael_mic_verify() 153 key = &rx->key->conf.key[NL80211_TKIP_DATA_OFFSET_RX_MIC_KEY]; ieee80211_rx_h_michael_mic_verify() 163 rx->key->u.tkip.rx[rx->security_idx].iv32 = rx->tkip_iv32; ieee80211_rx_h_michael_mic_verify() 164 rx->key->u.tkip.rx[rx->security_idx].iv16 = rx->tkip_iv16; ieee80211_rx_h_michael_mic_verify() 169 rx->key->u.tkip.mic_failures++; ieee80211_rx_h_michael_mic_verify() 177 cfg80211_michael_mic_failure(rx->sdata->dev, hdr->addr2, ieee80211_rx_h_michael_mic_verify() 181 rx->key ? rx->key->conf.keyidx : -1, ieee80211_rx_h_michael_mic_verify() 261 ieee80211_crypto_tkip_decrypt(struct ieee80211_rx_data *rx) ieee80211_crypto_tkip_decrypt() argument 263 struct ieee80211_hdr *hdr = (struct ieee80211_hdr *) rx->skb->data; ieee80211_crypto_tkip_decrypt() 265 struct ieee80211_key *key = rx->key; ieee80211_crypto_tkip_decrypt() 266 struct sk_buff *skb = rx->skb; ieee80211_crypto_tkip_decrypt() 274 if (!rx->sta || skb->len - hdrlen < 12) ieee80211_crypto_tkip_decrypt() 278 if (skb_linearize(rx->skb)) ieee80211_crypto_tkip_decrypt() 290 res = ieee80211_tkip_decrypt_data(rx->local->wep_rx_tfm, ieee80211_crypto_tkip_decrypt() 292 skb->len - hdrlen, rx->sta->sta.addr, ieee80211_crypto_tkip_decrypt() 293 hdr->addr1, hwaccel, rx->security_idx, ieee80211_crypto_tkip_decrypt() 294 &rx->tkip_iv32, ieee80211_crypto_tkip_decrypt() 295 &rx->tkip_iv16); ieee80211_crypto_tkip_decrypt() 492 ieee80211_crypto_ccmp_decrypt(struct ieee80211_rx_data *rx, ieee80211_crypto_ccmp_decrypt() argument 495 struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)rx->skb->data; ieee80211_crypto_ccmp_decrypt() 497 struct ieee80211_key *key = rx->key; ieee80211_crypto_ccmp_decrypt() 498 struct sk_buff *skb = rx->skb; ieee80211_crypto_ccmp_decrypt() 511 if (!rx->sta || data_len < 0) ieee80211_crypto_ccmp_decrypt() 515 if (!pskb_may_pull(rx->skb, hdrlen + IEEE80211_CCMP_HDR_LEN)) ieee80211_crypto_ccmp_decrypt() 518 if (skb_linearize(rx->skb)) ieee80211_crypto_ccmp_decrypt() 525 queue = rx->security_idx; ieee80211_crypto_ccmp_decrypt() 718 ieee80211_crypto_gcmp_decrypt(struct ieee80211_rx_data *rx) ieee80211_crypto_gcmp_decrypt() argument 720 struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)rx->skb->data; ieee80211_crypto_gcmp_decrypt() 722 struct ieee80211_key *key = rx->key; ieee80211_crypto_gcmp_decrypt() 723 struct sk_buff *skb = rx->skb; ieee80211_crypto_gcmp_decrypt() 737 if (!rx->sta || data_len < 0) ieee80211_crypto_gcmp_decrypt() 741 if (!pskb_may_pull(rx->skb, hdrlen + IEEE80211_GCMP_HDR_LEN)) ieee80211_crypto_gcmp_decrypt() 744 if (skb_linearize(rx->skb)) ieee80211_crypto_gcmp_decrypt() 751 queue = rx->security_idx; ieee80211_crypto_gcmp_decrypt() 830 ieee80211_crypto_cs_decrypt(struct ieee80211_rx_data *rx) ieee80211_crypto_cs_decrypt() argument 832 struct ieee80211_key *key = rx->key; ieee80211_crypto_cs_decrypt() 833 struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)rx->skb->data; ieee80211_crypto_cs_decrypt() 836 struct ieee80211_rx_status *status = IEEE80211_SKB_RXCB(rx->skb); ieee80211_crypto_cs_decrypt() 842 if (!rx->sta || !rx->sta->cipher_scheme || ieee80211_crypto_cs_decrypt() 849 cs = rx->sta->cipher_scheme; ieee80211_crypto_cs_decrypt() 851 data_len = rx->skb->len - hdrlen - cs->hdr_len; ieee80211_crypto_cs_decrypt() 862 if (skb_linearize(rx->skb)) ieee80211_crypto_cs_decrypt() 865 hdr = (struct ieee80211_hdr *)rx->skb->data; ieee80211_crypto_cs_decrypt() 868 skb_pn = rx->skb->data + hdrlen + cs->pn_off; ieee80211_crypto_cs_decrypt() 876 if (pskb_trim(rx->skb, rx->skb->len - cs->mic_len)) ieee80211_crypto_cs_decrypt() 879 memmove(rx->skb->data + cs->hdr_len, rx->skb->data, hdrlen); ieee80211_crypto_cs_decrypt() 880 skb_pull(rx->skb, cs->hdr_len); ieee80211_crypto_cs_decrypt() 1012 ieee80211_crypto_aes_cmac_decrypt(struct ieee80211_rx_data *rx) ieee80211_crypto_aes_cmac_decrypt() argument 1014 struct sk_buff *skb = rx->skb; ieee80211_crypto_aes_cmac_decrypt() 1016 struct ieee80211_key *key = rx->key; ieee80211_crypto_aes_cmac_decrypt() 1062 ieee80211_crypto_aes_cmac_256_decrypt(struct ieee80211_rx_data *rx) ieee80211_crypto_aes_cmac_256_decrypt() argument 1064 struct sk_buff *skb = rx->skb; ieee80211_crypto_aes_cmac_256_decrypt() 1066 struct ieee80211_key *key = rx->key; ieee80211_crypto_aes_cmac_256_decrypt() 1161 ieee80211_crypto_aes_gmac_decrypt(struct ieee80211_rx_data *rx) ieee80211_crypto_aes_gmac_decrypt() argument 1163 struct sk_buff *skb = rx->skb; ieee80211_crypto_aes_gmac_decrypt() 1165 struct ieee80211_key *key = rx->key; ieee80211_crypto_aes_gmac_decrypt() 1242 ieee80211_crypto_hw_decrypt(struct ieee80211_rx_data *rx) ieee80211_crypto_hw_decrypt() argument 1244 if (rx->sta && rx->sta->cipher_scheme) ieee80211_crypto_hw_decrypt() 1245 return ieee80211_crypto_cs_decrypt(rx); ieee80211_crypto_hw_decrypt()
|
H A D | Makefile | 11 ht.o agg-tx.o agg-rx.o \ 24 rx.o \
|
H A D | tkip.c | 268 if (key->u.tkip.rx[queue].state != TKIP_STATE_NOT_INIT && ieee80211_tkip_decrypt_data() 269 (iv32 < key->u.tkip.rx[queue].iv32 || ieee80211_tkip_decrypt_data() 270 (iv32 == key->u.tkip.rx[queue].iv32 && ieee80211_tkip_decrypt_data() 271 iv16 <= key->u.tkip.rx[queue].iv16))) ieee80211_tkip_decrypt_data() 276 key->u.tkip.rx[queue].state = TKIP_STATE_PHASE1_HW_UPLOADED; ieee80211_tkip_decrypt_data() 280 if (key->u.tkip.rx[queue].state == TKIP_STATE_NOT_INIT || ieee80211_tkip_decrypt_data() 281 key->u.tkip.rx[queue].iv32 != iv32) { ieee80211_tkip_decrypt_data() 283 tkip_mixing_phase1(tk, &key->u.tkip.rx[queue], ta, iv32); ieee80211_tkip_decrypt_data() 287 key->u.tkip.rx[queue].state != TKIP_STATE_PHASE1_HW_UPLOADED) { ieee80211_tkip_decrypt_data() 294 iv32, key->u.tkip.rx[queue].p1k); ieee80211_tkip_decrypt_data() 295 key->u.tkip.rx[queue].state = TKIP_STATE_PHASE1_HW_UPLOADED; ieee80211_tkip_decrypt_data() 298 tkip_mixing_phase2(tk, &key->u.tkip.rx[queue], iv16, rc4key); ieee80211_tkip_decrypt_data()
|
H A D | wep.c | 275 ieee80211_crypto_wep_decrypt(struct ieee80211_rx_data *rx) ieee80211_crypto_wep_decrypt() argument 277 struct sk_buff *skb = rx->skb; ieee80211_crypto_wep_decrypt() 286 if (skb_linearize(rx->skb)) ieee80211_crypto_wep_decrypt() 288 if (ieee80211_wep_decrypt(rx->local, rx->skb, rx->key)) ieee80211_crypto_wep_decrypt() 291 if (!pskb_may_pull(rx->skb, ieee80211_hdrlen(fc) + ieee80211_crypto_wep_decrypt() 294 ieee80211_wep_remove_iv(rx->local, rx->skb, rx->key); ieee80211_crypto_wep_decrypt() 296 if (pskb_trim(rx->skb, rx->skb->len - IEEE80211_WEP_ICV_LEN)) ieee80211_crypto_wep_decrypt()
|
H A D | wep.h | 30 ieee80211_crypto_wep_decrypt(struct ieee80211_rx_data *rx);
|
/linux-4.4.14/drivers/net/wireless/ath/carl9170/ |
H A D | Makefile | 1 carl9170-objs := main.o usb.o cmd.o mac.o phy.o led.o fw.o tx.o rx.o
|
/linux-4.4.14/net/mac802154/ |
H A D | Makefile | 2 mac802154-objs := main.o rx.o tx.o mac_cmd.o mib.o \
|
/linux-4.4.14/net/rxrpc/ |
H A D | af_rxrpc.c | 80 static int rxrpc_validate_address(struct rxrpc_sock *rx, rxrpc_validate_address() argument 98 if (srx->transport.family != rx->proto) rxrpc_validate_address() 127 struct rxrpc_sock *rx = rxrpc_sk(sk), *prx; rxrpc_bind() local 131 _enter("%p,%p,%d", rx, saddr, len); rxrpc_bind() 133 ret = rxrpc_validate_address(rx, srx, len); rxrpc_bind() 137 lock_sock(&rx->sk); rxrpc_bind() 139 if (rx->sk.sk_state != RXRPC_UNCONNECTED) { rxrpc_bind() 144 memcpy(&rx->srx, srx, sizeof(rx->srx)); rxrpc_bind() 147 local = rxrpc_lookup_local(&rx->srx); rxrpc_bind() 153 rx->local = local; rxrpc_bind() 162 rx->service_id = service_id; rxrpc_bind() 163 list_add_tail(&rx->listen_link, &local->services); rxrpc_bind() 166 rx->sk.sk_state = RXRPC_SERVER_BOUND; rxrpc_bind() 168 rx->sk.sk_state = RXRPC_CLIENT_BOUND; rxrpc_bind() 171 release_sock(&rx->sk); rxrpc_bind() 179 release_sock(&rx->sk); rxrpc_bind() 191 struct rxrpc_sock *rx = rxrpc_sk(sk); rxrpc_listen() local 194 _enter("%p,%d", rx, backlog); rxrpc_listen() 196 lock_sock(&rx->sk); rxrpc_listen() 198 switch (rx->sk.sk_state) { rxrpc_listen() 208 ASSERT(rx->local != NULL); rxrpc_listen() 210 rx->sk.sk_state = RXRPC_SERVER_LISTENING; rxrpc_listen() 215 release_sock(&rx->sk); rxrpc_listen() 230 struct rxrpc_sock *rx = rxrpc_sk(sock->sk); rxrpc_name_to_transport() local 233 _enter("%p,%p,%d,%d", rx, addr, addr_len, flags); rxrpc_name_to_transport() 235 ASSERT(rx->local != NULL); rxrpc_name_to_transport() 236 ASSERT(rx->sk.sk_state > RXRPC_UNCONNECTED); rxrpc_name_to_transport() 238 if (rx->srx.transport_type != srx->transport_type) rxrpc_name_to_transport() 240 if (rx->srx.transport.family != srx->transport.family) rxrpc_name_to_transport() 249 trans = rxrpc_get_transport(rx->local, peer, gfp); rxrpc_name_to_transport() 278 struct rxrpc_sock *rx = rxrpc_sk(sock->sk); rxrpc_kernel_begin_call() local 283 lock_sock(&rx->sk); rxrpc_kernel_begin_call() 294 trans = rx->trans; rxrpc_kernel_begin_call() 302 service_id = rx->service_id; rxrpc_kernel_begin_call() 307 key = rx->key; rxrpc_kernel_begin_call() 311 bundle = rxrpc_get_bundle(rx, trans, key, service_id, gfp); rxrpc_kernel_begin_call() 317 call = rxrpc_get_client_call(rx, trans, bundle, user_call_ID, true, rxrpc_kernel_begin_call() 323 release_sock(&rx->sk); rxrpc_kernel_begin_call() 360 struct rxrpc_sock *rx = rxrpc_sk(sock->sk); rxrpc_kernel_intercept_rx_messages() local 363 rx->interceptor = interceptor; rxrpc_kernel_intercept_rx_messages() 380 struct rxrpc_sock *rx = rxrpc_sk(sk); rxrpc_connect() local 383 _enter("%p,%p,%d,%d", rx, addr, addr_len, flags); rxrpc_connect() 385 ret = rxrpc_validate_address(rx, srx, addr_len); rxrpc_connect() 391 lock_sock(&rx->sk); rxrpc_connect() 393 switch (rx->sk.sk_state) { rxrpc_connect() 396 ASSERTCMP(rx->local, ==, NULL); rxrpc_connect() 397 rx->srx.srx_family = AF_RXRPC; rxrpc_connect() 398 rx->srx.srx_service = 0; rxrpc_connect() 399 rx->srx.transport_type = srx->transport_type; rxrpc_connect() 400 rx->srx.transport_len = sizeof(sa_family_t); rxrpc_connect() 401 rx->srx.transport.family = srx->transport.family; rxrpc_connect() 402 local = rxrpc_lookup_local(&rx->srx); rxrpc_connect() 404 release_sock(&rx->sk); rxrpc_connect() 407 rx->local = local; rxrpc_connect() 408 rx->sk.sk_state = RXRPC_CLIENT_BOUND; rxrpc_connect() 412 release_sock(&rx->sk); rxrpc_connect() 415 release_sock(&rx->sk); rxrpc_connect() 422 release_sock(&rx->sk); rxrpc_connect() 427 rx->trans = trans; rxrpc_connect() 428 rx->service_id = htons(srx->srx_service); rxrpc_connect() 429 rx->sk.sk_state = RXRPC_CLIENT_CONNECTED; rxrpc_connect() 431 release_sock(&rx->sk); rxrpc_connect() 447 struct rxrpc_sock *rx = rxrpc_sk(sock->sk); rxrpc_sendmsg() local 450 _enter(",{%d},,%zu", rx->sk.sk_state, len); rxrpc_sendmsg() 456 ret = rxrpc_validate_address(rx, m->msg_name, m->msg_namelen); rxrpc_sendmsg() 464 lock_sock(&rx->sk); rxrpc_sendmsg() 476 trans = rx->trans; rxrpc_sendmsg() 481 switch (rx->sk.sk_state) { rxrpc_sendmsg() 484 ret = rxrpc_server_sendmsg(rx, m, len); rxrpc_sendmsg() 494 ret = rxrpc_client_sendmsg(rx, trans, m, len); rxrpc_sendmsg() 502 release_sock(&rx->sk); rxrpc_sendmsg() 515 struct rxrpc_sock *rx = rxrpc_sk(sock->sk); rxrpc_setsockopt() local 521 lock_sock(&rx->sk); rxrpc_setsockopt() 531 if (rx->sk.sk_state != RXRPC_UNCONNECTED) rxrpc_setsockopt() 533 set_bit(RXRPC_SOCK_EXCLUSIVE_CONN, &rx->flags); rxrpc_setsockopt() 538 if (rx->key) rxrpc_setsockopt() 541 if (rx->sk.sk_state != RXRPC_UNCONNECTED) rxrpc_setsockopt() 543 ret = rxrpc_request_key(rx, optval, optlen); rxrpc_setsockopt() 548 if (rx->key) rxrpc_setsockopt() 551 if (rx->sk.sk_state != RXRPC_UNCONNECTED) rxrpc_setsockopt() 553 ret = rxrpc_server_keyring(rx, optval, optlen); rxrpc_setsockopt() 561 if (rx->sk.sk_state != RXRPC_UNCONNECTED) rxrpc_setsockopt() 570 rx->min_sec_level = min_sec_level; rxrpc_setsockopt() 581 release_sock(&rx->sk); rxrpc_setsockopt() 617 struct rxrpc_sock *rx; rxrpc_create() local 645 rx = rxrpc_sk(sk); rxrpc_create() 646 rx->proto = protocol; rxrpc_create() 647 rx->calls = RB_ROOT; rxrpc_create() 649 INIT_LIST_HEAD(&rx->listen_link); rxrpc_create() 650 INIT_LIST_HEAD(&rx->secureq); rxrpc_create() 651 INIT_LIST_HEAD(&rx->acceptq); rxrpc_create() 652 rwlock_init(&rx->call_lock); rxrpc_create() 653 memset(&rx->srx, 0, sizeof(rx->srx)); rxrpc_create() 655 _leave(" = 0 [%p]", rx); rxrpc_create() 683 struct rxrpc_sock *rx = rxrpc_sk(sk); rxrpc_release_sock() local 695 ASSERTCMP(rx->listen_link.next, !=, LIST_POISON1); rxrpc_release_sock() 697 if (!list_empty(&rx->listen_link)) { rxrpc_release_sock() 698 write_lock_bh(&rx->local->services_lock); rxrpc_release_sock() 699 list_del(&rx->listen_link); rxrpc_release_sock() 700 write_unlock_bh(&rx->local->services_lock); rxrpc_release_sock() 704 rxrpc_release_calls_on_socket(rx); rxrpc_release_sock() 708 if (rx->conn) { rxrpc_release_sock() 709 rxrpc_put_connection(rx->conn); rxrpc_release_sock() 710 rx->conn = NULL; rxrpc_release_sock() 713 if (rx->bundle) { rxrpc_release_sock() 714 rxrpc_put_bundle(rx->trans, rx->bundle); rxrpc_release_sock() 715 rx->bundle = NULL; rxrpc_release_sock() 717 if (rx->trans) { rxrpc_release_sock() 718 rxrpc_put_transport(rx->trans); rxrpc_release_sock() 719 rx->trans = NULL; rxrpc_release_sock() 721 if (rx->local) { rxrpc_release_sock() 722 rxrpc_put_local(rx->local); rxrpc_release_sock() 723 rx->local = NULL; rxrpc_release_sock() 726 key_put(rx->key); rxrpc_release_sock() 727 rx->key = NULL; rxrpc_release_sock() 728 key_put(rx->securities); rxrpc_release_sock() 729 rx->securities = NULL; rxrpc_release_sock()
|
H A D | ar-recvmsg.c | 23 void rxrpc_remove_user_ID(struct rxrpc_sock *rx, struct rxrpc_call *call) rxrpc_remove_user_ID() argument 28 write_lock_bh(&rx->call_lock); rxrpc_remove_user_ID() 31 write_unlock_bh(&rx->call_lock); rxrpc_remove_user_ID() 51 struct rxrpc_sock *rx = rxrpc_sk(sock->sk); rxrpc_recvmsg() local 66 timeo = sock_rcvtimeo(&rx->sk, flags & MSG_DONTWAIT); rxrpc_recvmsg() 69 lock_sock(&rx->sk); rxrpc_recvmsg() 74 if (RB_EMPTY_ROOT(&rx->calls)) { rxrpc_recvmsg() 77 if (rx->sk.sk_state != RXRPC_SERVER_LISTENING) { rxrpc_recvmsg() 78 release_sock(&rx->sk); rxrpc_recvmsg() 86 skb = skb_peek(&rx->sk.sk_receive_queue); rxrpc_recvmsg() 94 release_sock(&rx->sk); rxrpc_recvmsg() 95 prepare_to_wait_exclusive(sk_sleep(&rx->sk), &wait, rxrpc_recvmsg() 97 ret = sock_error(&rx->sk); rxrpc_recvmsg() 101 if (skb_queue_empty(&rx->sk.sk_receive_queue)) { rxrpc_recvmsg() 106 finish_wait(sk_sleep(&rx->sk), &wait); rxrpc_recvmsg() 107 lock_sock(&rx->sk); rxrpc_recvmsg() 124 if (skb_dequeue(&rx->sk.sk_receive_queue) != skb) rxrpc_recvmsg() 135 release_sock(&rx->sk); rxrpc_recvmsg() 153 sock_recv_timestamp(msg, &rx->sk, skb); rxrpc_recvmsg() 215 if (skb_dequeue(&rx->sk.sk_receive_queue) != rxrpc_recvmsg() 235 if (skb == (struct sk_buff *) &rx->sk.sk_receive_queue) rxrpc_recvmsg() 241 if (skb_dequeue(&rx->sk.sk_receive_queue) != skb) rxrpc_recvmsg() 249 release_sock(&rx->sk); rxrpc_recvmsg() 268 if (skb_dequeue(&rx->sk.sk_receive_queue) != skb) rxrpc_recvmsg() 320 if (skb_dequeue(&rx->sk.sk_receive_queue) != skb) rxrpc_recvmsg() 323 rxrpc_remove_user_ID(rx, call); rxrpc_recvmsg() 326 release_sock(&rx->sk); rxrpc_recvmsg() 335 release_sock(&rx->sk); rxrpc_recvmsg() 345 finish_wait(sk_sleep(&rx->sk), &wait); rxrpc_recvmsg()
|
H A D | ar-accept.c | 74 struct rxrpc_sock *rx, rxrpc_accept_incoming_call() 123 call = rxrpc_incoming_call(rx, conn, &sp->hdr, GFP_NOIO); rxrpc_accept_incoming_call() 133 if (rx->sk.sk_state == RXRPC_CLOSE) rxrpc_accept_incoming_call() 136 write_lock(&rx->call_lock); rxrpc_accept_incoming_call() 144 list_add_tail(&call->accept_link, &rx->secureq); rxrpc_accept_incoming_call() 152 list_add_tail(&call->accept_link, &rx->acceptq); rxrpc_accept_incoming_call() 171 write_unlock(&rx->call_lock); rxrpc_accept_incoming_call() 213 struct rxrpc_sock *rx; rxrpc_accept_incoming_calls() local 261 list_for_each_entry(rx, &local->services, listen_link) { rxrpc_accept_incoming_calls() 262 if (rx->service_id == service_id && rxrpc_accept_incoming_calls() 263 rx->sk.sk_state != RXRPC_CLOSE) rxrpc_accept_incoming_calls() 270 _debug("found service %hd", ntohs(rx->service_id)); rxrpc_accept_incoming_calls() 271 if (sk_acceptq_is_full(&rx->sk)) rxrpc_accept_incoming_calls() 273 sk_acceptq_added(&rx->sk); rxrpc_accept_incoming_calls() 274 sock_hold(&rx->sk); rxrpc_accept_incoming_calls() 277 ret = rxrpc_accept_incoming_call(local, rx, skb, &srx); rxrpc_accept_incoming_calls() 279 sk_acceptq_removed(&rx->sk); rxrpc_accept_incoming_calls() 280 sock_put(&rx->sk); rxrpc_accept_incoming_calls() 319 struct rxrpc_call *rxrpc_accept_call(struct rxrpc_sock *rx, rxrpc_accept_call() argument 330 write_lock(&rx->call_lock); rxrpc_accept_call() 333 if (list_empty(&rx->acceptq)) rxrpc_accept_call() 338 pp = &rx->calls.rb_node; rxrpc_accept_call() 353 call = list_entry(rx->acceptq.next, struct rxrpc_call, accept_link); rxrpc_accept_call() 355 sk_acceptq_removed(&rx->sk); rxrpc_accept_call() 379 rb_insert_color(&call->sock_node, &rx->calls); rxrpc_accept_call() 388 write_unlock(&rx->call_lock); rxrpc_accept_call() 404 write_unlock(&rx->call_lock); rxrpc_accept_call() 413 int rxrpc_reject_call(struct rxrpc_sock *rx) rxrpc_reject_call() argument 422 write_lock(&rx->call_lock); rxrpc_reject_call() 425 if (list_empty(&rx->acceptq)) rxrpc_reject_call() 429 call = list_entry(rx->acceptq.next, struct rxrpc_call, accept_link); rxrpc_reject_call() 431 sk_acceptq_removed(&rx->sk); rxrpc_reject_call() 467 write_unlock(&rx->call_lock); rxrpc_reject_call() 73 rxrpc_accept_incoming_call(struct rxrpc_local *local, struct rxrpc_sock *rx, struct sk_buff *skb, struct sockaddr_rxrpc *srx) rxrpc_accept_incoming_call() argument
|
H A D | ar-call.c | 252 struct rxrpc_sock *rx, rxrpc_alloc_client_call() 262 ASSERT(rx != NULL); rxrpc_alloc_client_call() 270 sock_hold(&rx->sk); rxrpc_alloc_client_call() 271 call->socket = rx; rxrpc_alloc_client_call() 274 ret = rxrpc_connect_call(rx, trans, bundle, call, gfp); rxrpc_alloc_client_call() 281 call->proto = rx->proto; rxrpc_alloc_client_call() 315 struct rxrpc_call *rxrpc_get_client_call(struct rxrpc_sock *rx, rxrpc_get_client_call() argument 326 rx, trans ? trans->debug_id : -1, bundle ? bundle->debug_id : -1, rxrpc_get_client_call() 331 read_lock(&rx->call_lock); rxrpc_get_client_call() 333 p = rx->calls.rb_node; rxrpc_get_client_call() 345 read_unlock(&rx->call_lock); rxrpc_get_client_call() 352 candidate = rxrpc_alloc_client_call(rx, trans, bundle, gfp); rxrpc_get_client_call() 361 write_lock(&rx->call_lock); rxrpc_get_client_call() 363 pp = &rx->calls.rb_node; rxrpc_get_client_call() 383 rb_insert_color(&call->sock_node, &rx->calls); rxrpc_get_client_call() 384 write_unlock(&rx->call_lock); rxrpc_get_client_call() 398 read_unlock(&rx->call_lock); rxrpc_get_client_call() 405 write_unlock(&rx->call_lock); rxrpc_get_client_call() 415 struct rxrpc_call *rxrpc_incoming_call(struct rxrpc_sock *rx, rxrpc_incoming_call() argument 426 ASSERT(rx != NULL); rxrpc_incoming_call() 432 candidate->socket = rx; rxrpc_incoming_call() 510 sock_hold(&rx->sk); rxrpc_incoming_call() 523 call->proto = rx->proto; rxrpc_incoming_call() 574 struct rxrpc_call *rxrpc_find_server_call(struct rxrpc_sock *rx, rxrpc_find_server_call() argument 580 _enter("%p,%lx", rx, user_call_ID); rxrpc_find_server_call() 584 read_lock(&rx->call_lock); rxrpc_find_server_call() 586 p = rx->calls.rb_node; rxrpc_find_server_call() 598 read_unlock(&rx->call_lock); rxrpc_find_server_call() 605 read_unlock(&rx->call_lock); rxrpc_find_server_call() 616 struct rxrpc_sock *rx = call->socket; rxrpc_release_call() local 633 write_lock_bh(&rx->call_lock); rxrpc_release_call() 639 sk_acceptq_removed(&rx->sk); rxrpc_release_call() 641 rb_erase(&call->sock_node, &rx->calls); rxrpc_release_call() 645 write_unlock_bh(&rx->call_lock); rxrpc_release_call() 780 void rxrpc_release_calls_on_socket(struct rxrpc_sock *rx) rxrpc_release_calls_on_socket() argument 785 _enter("%p", rx); rxrpc_release_calls_on_socket() 787 read_lock_bh(&rx->call_lock); rxrpc_release_calls_on_socket() 790 for (p = rb_first(&rx->calls); p; p = rb_next(p)) { rxrpc_release_calls_on_socket() 796 list_for_each_entry(call, &rx->secureq, accept_link) { rxrpc_release_calls_on_socket() 800 list_for_each_entry(call, &rx->acceptq, accept_link) { rxrpc_release_calls_on_socket() 804 read_unlock_bh(&rx->call_lock); rxrpc_release_calls_on_socket() 251 rxrpc_alloc_client_call( struct rxrpc_sock *rx, struct rxrpc_transport *trans, struct rxrpc_conn_bundle *bundle, gfp_t gfp) rxrpc_alloc_client_call() argument
|
H A D | ar-connection.c | 69 struct rxrpc_conn_bundle *rxrpc_get_bundle(struct rxrpc_sock *rx, rxrpc_get_bundle() argument 79 rx, key_serial(key), trans->debug_id, ntohs(service_id)); rxrpc_get_bundle() 81 if (rx->trans == trans && rx->bundle) { rxrpc_get_bundle() 82 atomic_inc(&rx->bundle->usage); rxrpc_get_bundle() 83 return rx->bundle; rxrpc_get_bundle() 139 if (!rx->bundle && rx->sk.sk_state == RXRPC_CLIENT_CONNECTED) { rxrpc_get_bundle() 141 rx->bundle = bundle; rxrpc_get_bundle() 151 if (!rx->bundle && rx->sk.sk_state == RXRPC_CLIENT_CONNECTED) { rxrpc_get_bundle() 153 rx->bundle = bundle; rxrpc_get_bundle() 164 if (!rx->bundle && rx->sk.sk_state == RXRPC_CLIENT_CONNECTED) { rxrpc_get_bundle() 166 rx->bundle = bundle; rxrpc_get_bundle() 335 static int rxrpc_connect_exclusive(struct rxrpc_sock *rx, rxrpc_connect_exclusive() argument 346 conn = rx->conn; rxrpc_connect_exclusive() 365 conn->security_level = rx->min_sec_level; rxrpc_connect_exclusive() 366 conn->key = key_get(rx->key); rxrpc_connect_exclusive() 387 rx->conn = conn; rxrpc_connect_exclusive() 429 int rxrpc_connect_call(struct rxrpc_sock *rx, rxrpc_connect_call() argument 440 _enter("%p,%lx,", rx, call->user_call_ID); rxrpc_connect_call() 442 if (test_bit(RXRPC_SOCK_EXCLUSIVE_CONN, &rx->flags)) rxrpc_connect_call() 443 return rxrpc_connect_exclusive(rx, trans, bundle->service_id, rxrpc_connect_call() 542 candidate->security_level = rx->min_sec_level; rxrpc_connect_call()
|
H A D | ar-output.c | 26 static int rxrpc_send_data(struct rxrpc_sock *rx, 33 static int rxrpc_sendmsg_cmsg(struct rxrpc_sock *rx, struct msghdr *msg, rxrpc_sendmsg_cmsg() argument 131 int rxrpc_client_sendmsg(struct rxrpc_sock *rx, struct rxrpc_transport *trans, rxrpc_client_sendmsg() argument 147 ret = rxrpc_sendmsg_cmsg(rx, msg, &user_call_ID, &cmd, &abort_code, rxrpc_client_sendmsg() 154 service_id = rx->service_id; rxrpc_client_sendmsg() 160 key = rx->key; rxrpc_client_sendmsg() 161 if (key && !rx->key->payload.data[0]) rxrpc_client_sendmsg() 163 bundle = rxrpc_get_bundle(rx, trans, key, service_id, rxrpc_client_sendmsg() 169 call = rxrpc_get_client_call(rx, trans, bundle, user_call_ID, rxrpc_client_sendmsg() 192 ret = rxrpc_send_data(rx, call, msg, len); rxrpc_client_sendmsg() 272 int rxrpc_server_sendmsg(struct rxrpc_sock *rx, struct msghdr *msg, size_t len) rxrpc_server_sendmsg() argument 282 ret = rxrpc_sendmsg_cmsg(rx, msg, &user_call_ID, &cmd, &abort_code, rxrpc_server_sendmsg() 288 call = rxrpc_accept_call(rx, user_call_ID); rxrpc_server_sendmsg() 295 call = rxrpc_find_server_call(rx, user_call_ID); rxrpc_server_sendmsg() 313 ret = rxrpc_send_data(rx, call, msg, len); rxrpc_server_sendmsg() 396 static int rxrpc_wait_for_tx_window(struct rxrpc_sock *rx, rxrpc_wait_for_tx_window() argument 420 release_sock(&rx->sk); rxrpc_wait_for_tx_window() 422 lock_sock(&rx->sk); rxrpc_wait_for_tx_window() 494 /* attempt to cancel the rx-ACK timer, deferring reply transmission if rxrpc_queue_packet() 520 static int rxrpc_send_data(struct rxrpc_sock *rx, rxrpc_send_data() argument 526 struct sock *sk = &rx->sk; rxrpc_send_data() 556 ret = rxrpc_wait_for_tx_window(rx, call, rxrpc_send_data()
|
H A D | ar-security.c | 167 struct rxrpc_sock *rx; rxrpc_init_server_conn_security() local 184 list_for_each_entry(rx, &local->services, listen_link) { rxrpc_init_server_conn_security() 185 if (rx->service_id == conn->service_id) rxrpc_init_server_conn_security() 196 if (!rx->securities) { rxrpc_init_server_conn_security() 204 kref = keyring_search(make_key_ref(rx->securities, 1UL), rxrpc_init_server_conn_security()
|
/linux-4.4.14/drivers/net/ethernet/pasemi/ |
H A D | pasemi_mac_ethtool.c | 31 { "rx-drops" }, 32 { "rx-bytes" }, 33 { "rx-packets" }, 34 { "rx-broadcast-packets" }, 35 { "rx-multicast-packets" }, 36 { "rx-crc-errors" }, 37 { "rx-undersize-errors" }, 38 { "rx-oversize-errors" }, 39 { "rx-short-fragment-errors" }, 40 { "rx-jabber-errors" }, 41 { "rx-64-byte-packets" }, 42 { "rx-65-127-byte-packets" }, 43 { "rx-128-255-byte-packets" }, 44 { "rx-256-511-byte-packets" }, 45 { "rx-512-1023-byte-packets" }, 46 { "rx-1024-1518-byte-packets" }, 47 { "rx-pause-frames" }, 117 ering->rx_pending = RING_USED(mac->rx)/4; pasemi_mac_ethtool_get_ringparam()
|
H A D | pasemi_mac.h | 93 struct pasemi_mac_rxring *rx; member in struct:pasemi_mac 96 char rx_irq_name[10]; /* "eth%d rx" */ 112 #define RX_DESC(rx, num) ((rx)->chan.ring_virt[(num) & (RX_RING_SIZE-1)]) 113 #define RX_DESC_INFO(rx, num) ((rx)->ring_info[(num) & (RX_RING_SIZE-1)]) 114 #define RX_BUFF(rx, num) ((rx)->buffers[(num) & (RX_RING_SIZE-1)])
|
H A D | pasemi_mac.c | 122 return mac->rx; rx_ring() 480 mac->rx = ring; pasemi_mac_setup_rx_resources() 591 struct pasemi_mac_rxring *rx = rx_ring(mac); pasemi_mac_free_rx_buffers() local 596 info = &RX_DESC_INFO(rx, i); pasemi_mac_free_rx_buffers() 609 RX_BUFF(rx, i) = 0; pasemi_mac_free_rx_buffers() 621 mac->rx = NULL; pasemi_mac_free_rx_resources() 628 struct pasemi_mac_rxring *rx = rx_ring(mac); pasemi_mac_replenish_rx_ring() local 636 struct pasemi_mac_buffer *info = &RX_DESC_INFO(rx, fill); pasemi_mac_replenish_rx_ring() 637 u64 *buff = &RX_BUFF(rx, fill); pasemi_mac_replenish_rx_ring() 675 struct pasemi_mac_rxring *rx = rx_ring(mac); pasemi_mac_restart_rx_intr() local 681 pcnt = *rx->chan.status & PAS_STATUS_PCNT_M; pasemi_mac_restart_rx_intr() 685 if (*rx->chan.status & PAS_STATUS_TIMER) pasemi_mac_restart_rx_intr() 688 write_iob_reg(PAS_IOB_DMA_RXCH_RESET(mac->rx->chan.chno), reg); pasemi_mac_restart_rx_intr() 716 printk(KERN_ERR "pasemi_mac: rx error. macrx %016llx, rx status %llx\n", pasemi_mac_rx_error() 740 static int pasemi_mac_clean_rx(struct pasemi_mac_rxring *rx, pasemi_mac_clean_rx() argument 743 const struct pasemi_dmachan *chan = &rx->chan; pasemi_mac_clean_rx() 744 struct pasemi_mac *mac = rx->mac; pasemi_mac_clean_rx() 757 spin_lock(&rx->lock); pasemi_mac_clean_rx() 759 n = rx->next_to_clean; pasemi_mac_clean_rx() 761 prefetch(&RX_DESC(rx, n)); pasemi_mac_clean_rx() 764 macrx = RX_DESC(rx, n); pasemi_mac_clean_rx() 765 prefetch(&RX_DESC(rx, n+4)); pasemi_mac_clean_rx() 778 eval = (RX_DESC(rx, n+1) & XCT_RXRES_8B_EVAL_M) >> pasemi_mac_clean_rx() 782 dma = (RX_DESC(rx, n+2) & XCT_PTR_ADDR_M); pasemi_mac_clean_rx() 783 info = &RX_DESC_INFO(rx, buf_index); pasemi_mac_clean_rx() 823 RX_DESC(rx, n) = 0; pasemi_mac_clean_rx() 824 RX_DESC(rx, n+1) = 0; pasemi_mac_clean_rx() 829 RX_BUFF(rx, buf_index) = 0; pasemi_mac_clean_rx() 848 write_dma_reg(PAS_DMA_RXCHAN_INCR(mac->rx->chan.chno), count << 1); pasemi_mac_clean_rx() 1152 write_iob_reg(PAS_IOB_DMA_RXCH_CFG(mac->rx->chan.chno), pasemi_mac_open() 1159 PAS_MAC_IPC_CHNL_DCHNO(mac->rx->chan.chno) | pasemi_mac_open() 1160 PAS_MAC_IPC_CHNL_BCH(mac->rx->chan.chno)); pasemi_mac_open() 1162 /* enable rx if */ pasemi_mac_open() 1170 /* enable rx channel */ pasemi_mac_open() 1228 snprintf(mac->rx_irq_name, sizeof(mac->rx_irq_name), "%s rx", pasemi_mac_open() 1231 ret = request_irq(mac->rx->chan.irq, pasemi_mac_rx_intr, 0, pasemi_mac_open() 1232 mac->rx_irq_name, mac->rx); pasemi_mac_open() 1235 mac->rx->chan.irq, ret); pasemi_mac_open() 1302 "Failed to stop rx channel, ccmdsta 08%x\n", sta); pasemi_mac_pause_rxchan() 1321 "Failed to stop rx interface, rcmdsta %08x\n", sta); pasemi_mac_pause_rxint() 1372 free_irq(mac->rx->chan.irq, mac->rx); pasemi_mac_close() 1640 disable_irq(mac->rx->chan.irq); pasemi_mac_netpoll() 1641 pasemi_mac_rx_intr(mac->rx->chan.irq, mac->rx); pasemi_mac_netpoll() 1642 enable_irq(mac->rx->chan.irq); pasemi_mac_netpoll() 1662 * interface ring, then finally re-fill the rx ring with pasemi_mac_change_mtu() 1865 pasemi_dma_free_chan(&mac->rx->chan); pasemi_mac_remove()
|
/linux-4.4.14/drivers/staging/gdm724x/ |
H A D | gdm_mux.c | 123 static struct mux_rx *get_rx_struct(struct rx_cxt *rx) get_rx_struct() argument 128 spin_lock_irqsave(&rx->free_list_lock, flags); get_rx_struct() 130 if (list_empty(&rx->rx_free_list)) { get_rx_struct() 131 spin_unlock_irqrestore(&rx->free_list_lock, flags); get_rx_struct() 135 r = list_entry(rx->rx_free_list.prev, struct mux_rx, free_list); get_rx_struct() 138 spin_unlock_irqrestore(&rx->free_list_lock, flags); get_rx_struct() 143 static void put_rx_struct(struct rx_cxt *rx, struct mux_rx *r) put_rx_struct() argument 147 spin_lock_irqsave(&rx->free_list_lock, flags); put_rx_struct() 148 list_add_tail(&r->free_list, &rx->rx_free_list); put_rx_struct() 149 spin_unlock_irqrestore(&rx->free_list_lock, flags); put_rx_struct() 223 struct rx_cxt *rx = &mux_dev->rx; do_rx() local 228 spin_lock_irqsave(&rx->to_host_lock, flags); do_rx() 229 if (list_empty(&rx->to_host_list)) { do_rx() 230 spin_unlock_irqrestore(&rx->to_host_lock, flags); do_rx() 233 r = list_entry(rx->to_host_list.next, struct mux_rx, do_rx() 236 spin_unlock_irqrestore(&rx->to_host_lock, flags); do_rx() 242 put_rx_struct(rx, r); do_rx() 246 static void remove_rx_submit_list(struct mux_rx *r, struct rx_cxt *rx) remove_rx_submit_list() argument 251 spin_lock_irqsave(&rx->submit_list_lock, flags); remove_rx_submit_list() 252 list_for_each_entry_safe(r_remove, r_remove_next, &rx->rx_submit_list, remove_rx_submit_list() 257 spin_unlock_irqrestore(&rx->submit_list_lock, flags); remove_rx_submit_list() 264 struct rx_cxt *rx = &mux_dev->rx; gdm_mux_rcv_complete() local 267 remove_rx_submit_list(r, rx); gdm_mux_rcv_complete() 273 put_rx_struct(rx, r); gdm_mux_rcv_complete() 276 spin_lock_irqsave(&rx->to_host_lock, flags); gdm_mux_rcv_complete() 277 list_add_tail(&r->to_host_list, &rx->to_host_list); gdm_mux_rcv_complete() 279 spin_unlock_irqrestore(&rx->to_host_lock, flags); gdm_mux_rcv_complete() 289 struct rx_cxt *rx = &mux_dev->rx; gdm_mux_recv() local 298 r = get_rx_struct(rx); gdm_mux_recv() 317 spin_lock_irqsave(&rx->submit_list_lock, flags); gdm_mux_recv() 318 list_add_tail(&r->rx_submit_list, &rx->rx_submit_list); gdm_mux_recv() 319 spin_unlock_irqrestore(&rx->submit_list_lock, flags); gdm_mux_recv() 324 spin_lock_irqsave(&rx->submit_list_lock, flags); gdm_mux_recv() 326 spin_unlock_irqrestore(&rx->submit_list_lock, flags); gdm_mux_recv() 328 put_rx_struct(rx, r); gdm_mux_recv() 443 struct rx_cxt *rx = &mux_dev->rx; release_usb() local 449 spin_lock_irqsave(&rx->submit_list_lock, flags); release_usb() 450 list_for_each_entry_safe(r, r_next, &rx->rx_submit_list, release_usb() 452 spin_unlock_irqrestore(&rx->submit_list_lock, flags); release_usb() 454 spin_lock_irqsave(&rx->submit_list_lock, flags); release_usb() 456 spin_unlock_irqrestore(&rx->submit_list_lock, flags); release_usb() 458 spin_lock_irqsave(&rx->free_list_lock, flags); release_usb() 459 list_for_each_entry_safe(r, r_next, &rx->rx_free_list, free_list) { release_usb() 463 spin_unlock_irqrestore(&rx->free_list_lock, flags); release_usb() 465 spin_lock_irqsave(&rx->to_host_lock, flags); release_usb() 466 list_for_each_entry_safe(r, r_next, &rx->to_host_list, to_host_list) { release_usb() 472 spin_unlock_irqrestore(&rx->to_host_lock, flags); release_usb() 478 struct rx_cxt *rx = &mux_dev->rx; init_usb() local 483 INIT_LIST_HEAD(&rx->to_host_list); init_usb() 484 INIT_LIST_HEAD(&rx->rx_submit_list); init_usb() 485 INIT_LIST_HEAD(&rx->rx_free_list); init_usb() 486 spin_lock_init(&rx->to_host_lock); init_usb() 487 spin_lock_init(&rx->submit_list_lock); init_usb() 488 spin_lock_init(&rx->free_list_lock); init_usb() 497 list_add(&r->free_list, &rx->rx_free_list); init_usb() 597 struct rx_cxt *rx; gdm_mux_suspend() local 603 rx = &mux_dev->rx; gdm_mux_suspend() 612 spin_lock_irqsave(&rx->submit_list_lock, flags); gdm_mux_suspend() 613 list_for_each_entry_safe(r, r_next, &rx->rx_submit_list, gdm_mux_suspend() 615 spin_unlock_irqrestore(&rx->submit_list_lock, flags); gdm_mux_suspend() 617 spin_lock_irqsave(&rx->submit_list_lock, flags); gdm_mux_suspend() 619 spin_unlock_irqrestore(&rx->submit_list_lock, flags); gdm_mux_suspend()
|
H A D | gdm_usb.c | 221 static struct usb_rx *get_rx_struct(struct rx_cxt *rx, int *no_spc) get_rx_struct() argument 226 spin_lock_irqsave(&rx->rx_lock, flags); get_rx_struct() 228 if (list_empty(&rx->free_list)) { get_rx_struct() 229 spin_unlock_irqrestore(&rx->rx_lock, flags); get_rx_struct() 233 r = list_entry(rx->free_list.next, struct usb_rx, free_list); get_rx_struct() 236 rx->avail_count--; get_rx_struct() 238 *no_spc = list_empty(&rx->free_list) ? 1 : 0; get_rx_struct() 240 spin_unlock_irqrestore(&rx->rx_lock, flags); get_rx_struct() 245 static void put_rx_struct(struct rx_cxt *rx, struct usb_rx *r) put_rx_struct() argument 249 spin_lock_irqsave(&rx->rx_lock, flags); put_rx_struct() 251 list_add_tail(&r->free_list, &rx->free_list); put_rx_struct() 252 rx->avail_count++; put_rx_struct() 254 spin_unlock_irqrestore(&rx->rx_lock, flags); put_rx_struct() 259 struct rx_cxt *rx = &udev->rx; release_usb() local 283 spin_lock_irqsave(&rx->submit_lock, flags); release_usb() 284 list_for_each_entry_safe(r, r_next, &rx->rx_submit_list, release_usb() 286 spin_unlock_irqrestore(&rx->submit_lock, flags); release_usb() 288 spin_lock_irqsave(&rx->submit_lock, flags); release_usb() 290 spin_unlock_irqrestore(&rx->submit_lock, flags); release_usb() 292 spin_lock_irqsave(&rx->rx_lock, flags); release_usb() 293 list_for_each_entry_safe(r, r_next, &rx->free_list, free_list) { release_usb() 297 spin_unlock_irqrestore(&rx->rx_lock, flags); release_usb() 299 spin_lock_irqsave(&rx->to_host_lock, flags); release_usb() 300 list_for_each_entry_safe(r, r_next, &rx->to_host_list, to_host_list) { release_usb() 306 spin_unlock_irqrestore(&rx->to_host_lock, flags); release_usb() 314 struct rx_cxt *rx = &udev->rx; init_usb() local 326 INIT_LIST_HEAD(&rx->rx_submit_list); init_usb() 327 INIT_LIST_HEAD(&rx->free_list); init_usb() 328 INIT_LIST_HEAD(&rx->to_host_list); init_usb() 330 spin_lock_init(&rx->rx_lock); init_usb() 331 spin_lock_init(&rx->submit_lock); init_usb() 332 spin_lock_init(&rx->to_host_lock); init_usb() 335 rx->avail_count = 0; init_usb() 357 list_add(&r->free_list, &rx->free_list); init_usb() 358 rx->avail_count++; init_usb() 394 struct rx_cxt *rx = &udev->rx; do_rx() local 403 spin_lock_irqsave(&rx->to_host_lock, flags); do_rx() 404 if (list_empty(&rx->to_host_list)) { do_rx() 405 spin_unlock_irqrestore(&rx->to_host_lock, flags); do_rx() 408 r = list_entry(rx->to_host_list.next, do_rx() 411 spin_unlock_irqrestore(&rx->to_host_lock, flags); do_rx() 441 put_rx_struct(rx, r); do_rx() 450 static void remove_rx_submit_list(struct usb_rx *r, struct rx_cxt *rx) remove_rx_submit_list() argument 455 spin_lock_irqsave(&rx->submit_lock, flags); remove_rx_submit_list() 457 &rx->rx_submit_list, rx_submit_list) { remove_rx_submit_list() 463 spin_unlock_irqrestore(&rx->submit_lock, flags); remove_rx_submit_list() 469 struct rx_cxt *rx = r->rx; gdm_usb_rcv_complete() local 471 struct lte_udev *udev = container_of(r->rx, struct lte_udev, rx); gdm_usb_rcv_complete() 474 remove_rx_submit_list(r, rx); gdm_usb_rcv_complete() 477 spin_lock_irqsave(&rx->to_host_lock, flags); gdm_usb_rcv_complete() 478 list_add_tail(&r->to_host_list, &rx->to_host_list); gdm_usb_rcv_complete() 480 spin_unlock_irqrestore(&rx->to_host_lock, flags); gdm_usb_rcv_complete() 486 put_rx_struct(rx, r); gdm_usb_rcv_complete() 500 struct rx_cxt *rx = &udev->rx; gdm_usb_recv() local 511 r = get_rx_struct(rx, &no_spc); gdm_usb_recv() 521 r->rx = rx; gdm_usb_recv() 531 spin_lock_irqsave(&rx->submit_lock, flags); gdm_usb_recv() 532 list_add_tail(&r->rx_submit_list, &rx->rx_submit_list); gdm_usb_recv() 533 spin_unlock_irqrestore(&rx->submit_lock, flags); gdm_usb_recv() 541 spin_lock_irqsave(&rx->submit_lock, flags); gdm_usb_recv() 543 spin_unlock_irqrestore(&rx->submit_lock, flags); gdm_usb_recv() 546 put_rx_struct(rx, r); gdm_usb_recv() 923 struct rx_cxt *rx; gdm_usb_suspend() local 930 rx = &udev->rx; gdm_usb_suspend() 938 spin_lock_irqsave(&rx->submit_lock, flags); gdm_usb_suspend() 939 list_for_each_entry_safe(r, r_next, &rx->rx_submit_list, gdm_usb_suspend() 941 spin_unlock_irqrestore(&rx->submit_lock, flags); gdm_usb_suspend() 943 spin_lock_irqsave(&rx->submit_lock, flags); gdm_usb_suspend() 945 spin_unlock_irqrestore(&rx->submit_lock, flags); gdm_usb_suspend() 955 struct rx_cxt *rx; gdm_usb_resume() local 962 rx = &udev->rx; gdm_usb_resume() 970 spin_lock_irqsave(&rx->rx_lock, flags); gdm_usb_resume() 971 issue_count = rx->avail_count - MAX_RX_SUBMIT_COUNT; gdm_usb_resume() 972 spin_unlock_irqrestore(&rx->rx_lock, flags); gdm_usb_resume()
|
H A D | gdm_usb.h | 68 struct rx_cxt *rx; member in struct:usb_rx 98 struct rx_cxt rx; member in struct:lte_udev
|
/linux-4.4.14/drivers/net/wireless/ti/wl18xx/ |
H A D | debugfs.c | 95 WL18XX_DEBUGFS_FWSTATS_FILE(rx, rx_beacon_early_term, "%u"); 96 WL18XX_DEBUGFS_FWSTATS_FILE(rx, rx_out_of_mpdu_nodes, "%u"); 97 WL18XX_DEBUGFS_FWSTATS_FILE(rx, rx_hdr_overflow, "%u"); 98 WL18XX_DEBUGFS_FWSTATS_FILE(rx, rx_dropped_frame, "%u"); 99 WL18XX_DEBUGFS_FWSTATS_FILE(rx, rx_done, "%u"); 100 WL18XX_DEBUGFS_FWSTATS_FILE(rx, rx_defrag, "%u"); 101 WL18XX_DEBUGFS_FWSTATS_FILE(rx, rx_defrag_end, "%u"); 102 WL18XX_DEBUGFS_FWSTATS_FILE(rx, rx_cmplt, "%u"); 103 WL18XX_DEBUGFS_FWSTATS_FILE(rx, rx_pre_complt, "%u"); 104 WL18XX_DEBUGFS_FWSTATS_FILE(rx, rx_cmplt_task, "%u"); 105 WL18XX_DEBUGFS_FWSTATS_FILE(rx, rx_phy_hdr, "%u"); 106 WL18XX_DEBUGFS_FWSTATS_FILE(rx, rx_timeout, "%u"); 107 WL18XX_DEBUGFS_FWSTATS_FILE(rx, rx_rts_timeout, "%u"); 108 WL18XX_DEBUGFS_FWSTATS_FILE(rx, rx_timeout_wa, "%u"); 109 WL18XX_DEBUGFS_FWSTATS_FILE(rx, defrag_called, "%u"); 110 WL18XX_DEBUGFS_FWSTATS_FILE(rx, defrag_init_called, "%u"); 111 WL18XX_DEBUGFS_FWSTATS_FILE(rx, defrag_in_process_called, "%u"); 112 WL18XX_DEBUGFS_FWSTATS_FILE(rx, defrag_tkip_called, "%u"); 113 WL18XX_DEBUGFS_FWSTATS_FILE(rx, defrag_need_defrag, "%u"); 114 WL18XX_DEBUGFS_FWSTATS_FILE(rx, defrag_decrypt_failed, "%u"); 115 WL18XX_DEBUGFS_FWSTATS_FILE(rx, decrypt_key_not_found, "%u"); 116 WL18XX_DEBUGFS_FWSTATS_FILE(rx, defrag_need_decrypt, "%u"); 117 WL18XX_DEBUGFS_FWSTATS_FILE(rx, rx_tkip_replays, "%u"); 118 WL18XX_DEBUGFS_FWSTATS_FILE(rx, rx_xfr, "%u"); 423 DEBUGFS_FWSTATS_ADD(rx, rx_beacon_early_term); wl18xx_debugfs_add_files() 424 DEBUGFS_FWSTATS_ADD(rx, rx_out_of_mpdu_nodes); wl18xx_debugfs_add_files() 425 DEBUGFS_FWSTATS_ADD(rx, rx_hdr_overflow); wl18xx_debugfs_add_files() 426 DEBUGFS_FWSTATS_ADD(rx, rx_dropped_frame); wl18xx_debugfs_add_files() 427 DEBUGFS_FWSTATS_ADD(rx, rx_done); wl18xx_debugfs_add_files() 428 DEBUGFS_FWSTATS_ADD(rx, rx_defrag); wl18xx_debugfs_add_files() 429 DEBUGFS_FWSTATS_ADD(rx, rx_defrag_end); wl18xx_debugfs_add_files() 430 DEBUGFS_FWSTATS_ADD(rx, rx_cmplt); wl18xx_debugfs_add_files() 431 DEBUGFS_FWSTATS_ADD(rx, rx_pre_complt); wl18xx_debugfs_add_files() 432 DEBUGFS_FWSTATS_ADD(rx, rx_cmplt_task); wl18xx_debugfs_add_files() 433 DEBUGFS_FWSTATS_ADD(rx, rx_phy_hdr); wl18xx_debugfs_add_files() 434 DEBUGFS_FWSTATS_ADD(rx, rx_timeout); wl18xx_debugfs_add_files() 435 DEBUGFS_FWSTATS_ADD(rx, rx_rts_timeout); wl18xx_debugfs_add_files() 436 DEBUGFS_FWSTATS_ADD(rx, rx_timeout_wa); wl18xx_debugfs_add_files() 437 DEBUGFS_FWSTATS_ADD(rx, defrag_called); wl18xx_debugfs_add_files() 438 DEBUGFS_FWSTATS_ADD(rx, defrag_init_called); wl18xx_debugfs_add_files() 439 DEBUGFS_FWSTATS_ADD(rx, defrag_in_process_called); wl18xx_debugfs_add_files() 440 DEBUGFS_FWSTATS_ADD(rx, defrag_tkip_called); wl18xx_debugfs_add_files() 441 DEBUGFS_FWSTATS_ADD(rx, defrag_need_defrag); wl18xx_debugfs_add_files() 442 DEBUGFS_FWSTATS_ADD(rx, defrag_decrypt_failed); wl18xx_debugfs_add_files() 443 DEBUGFS_FWSTATS_ADD(rx, decrypt_key_not_found); wl18xx_debugfs_add_files() 444 DEBUGFS_FWSTATS_ADD(rx, defrag_need_decrypt); wl18xx_debugfs_add_files() 445 DEBUGFS_FWSTATS_ADD(rx, rx_tkip_replays); wl18xx_debugfs_add_files() 446 DEBUGFS_FWSTATS_ADD(rx, rx_xfr); wl18xx_debugfs_add_files()
|
H A D | event.h | 65 /* rx ba constraint */ 81 /* rx BA win size indicated by RX_BA_WIN_SIZE_CHANGE_EVENT_ID */
|
/linux-4.4.14/drivers/scsi/aacraid/ |
H A D | Makefile | 6 dpcsup.o rx.o sa.o rkt.o nark.o src.o
|
H A D | nark.c | 46 iounmap(dev->regs.rx); aac_nark_ioremap() 47 dev->regs.rx = NULL; aac_nark_ioremap() 53 dev->regs.rx = ioremap((u64)pci_resource_start(dev->pdev, 0) | aac_nark_ioremap() 57 if (dev->regs.rx == NULL) aac_nark_ioremap() 61 iounmap(dev->regs.rx); aac_nark_ioremap() 62 dev->regs.rx = NULL; aac_nark_ioremap()
|
/linux-4.4.14/drivers/net/ethernet/brocade/bna/ |
H A D | bna_tx_rx.c | 192 bfi_msgq_mhdr_set(req->mh, BFI_MC_ENET, req_type, 0, rxf->rx->rid); bna_bfi_ucast_req() 198 bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd); bna_bfi_ucast_req() 208 0, rxf->rx->rid); bna_bfi_mcast_add_req() 214 bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd); bna_bfi_mcast_add_req() 224 0, rxf->rx->rid); bna_bfi_mcast_del_req() 230 bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd); bna_bfi_mcast_del_req() 239 BFI_ENET_H2I_MAC_MCAST_FILTER_REQ, 0, rxf->rx->rid); bna_bfi_mcast_filter_req() 245 bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd); bna_bfi_mcast_filter_req() 254 BFI_ENET_H2I_RX_PROMISCUOUS_REQ, 0, rxf->rx->rid); bna_bfi_rx_promisc_req() 260 bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd); bna_bfi_rx_promisc_req() 271 BFI_ENET_H2I_RX_VLAN_SET_REQ, 0, rxf->rx->rid); bna_bfi_rx_vlan_filter_set() 285 bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd); bna_bfi_rx_vlan_filter_set() 294 BFI_ENET_H2I_RX_VLAN_STRIP_ENABLE_REQ, 0, rxf->rx->rid); bna_bfi_vlan_strip_enable() 300 bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd); bna_bfi_vlan_strip_enable() 309 BFI_ENET_H2I_RIT_CFG_REQ, 0, rxf->rx->rid); bna_bfi_rit_cfg() 316 bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd); bna_bfi_rit_cfg() 326 BFI_ENET_H2I_RSS_CFG_REQ, 0, rxf->rx->rid); bna_bfi_rss_cfg() 336 bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd); bna_bfi_rss_cfg() 345 BFI_ENET_H2I_RSS_ENABLE_REQ, 0, rxf->rx->rid); bna_bfi_rss_enable() 351 bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd); bna_bfi_rss_enable() 392 mchandle = bna_mcam_mod_handle_get(&rxf->rx->bna->mcam_mod); bna_rxf_mchandle_attach() 419 bna_mcam_mod_handle_put(&rxf->rx->bna->mcam_mod, mchandle); bna_rxf_mcast_del() 437 list_move_tail(&mac->qe, bna_mcam_mod_del_q(rxf->rx->bna)); bna_rxf_mcast_cfg_apply() 485 list_move_tail(&mac->qe, bna_mcam_mod_del_q(rxf->rx->bna)); bna_rxf_mcast_cfg_reset() 569 struct bna_rx *rx = rxf->rx; bna_rit_init() local 574 list_for_each_entry(rxp, &rx->rxp_q, qe) { bna_rit_init() 617 struct bna_rx *rx, bna_rxf_init() 621 rxf->rx = rx; bna_rxf_init() 669 list_move_tail(&mac->qe, bna_ucam_mod_free_q(rxf->rx->bna)); bna_rxf_uninit() 674 bna_ucam_mod_free_q(rxf->rx->bna)); bna_rxf_uninit() 681 list_move_tail(&mac->qe, bna_mcam_mod_free_q(rxf->rx->bna)); bna_rxf_uninit() 686 if (rxf->rx->bna->promisc_rid == rxf->rx->rid) bna_rxf_uninit() 687 rxf->rx->bna->promisc_rid = BFI_INVALID_RID; bna_rxf_uninit() 688 if (rxf->rx->bna->default_mode_rid == rxf->rx->rid) bna_rxf_uninit() 689 rxf->rx->bna->default_mode_rid = BFI_INVALID_RID; bna_rxf_uninit() 694 rxf->rx = NULL; bna_rxf_uninit() 698 bna_rx_cb_rxf_started(struct bna_rx *rx) bna_rx_cb_rxf_started() argument 700 bfa_fsm_send_event(rx, RX_E_RXF_STARTED); bna_rx_cb_rxf_started() 707 rxf->start_cbarg = rxf->rx; bna_rxf_start() 712 bna_rx_cb_rxf_stopped(struct bna_rx *rx) bna_rx_cb_rxf_stopped() argument 714 bfa_fsm_send_event(rx, RX_E_RXF_STOPPED); bna_rx_cb_rxf_stopped() 721 rxf->stop_cbarg = rxf->rx; bna_rxf_stop() 732 bna_rx_ucast_set(struct bna_rx *rx, const u8 *ucmac) bna_rx_ucast_set() argument 734 struct bna_rxf *rxf = &rx->rxf; bna_rx_ucast_set() 738 bna_cam_mod_mac_get(bna_ucam_mod_free_q(rxf->rx->bna)); bna_rx_ucast_set() 746 rxf->cam_fltr_cbarg = rx->bna->bnad; bna_rx_ucast_set() 754 bna_rx_mcast_add(struct bna_rx *rx, const u8 *addr, bna_rx_mcast_add() argument 757 struct bna_rxf *rxf = &rx->rxf; bna_rx_mcast_add() 764 cbfn(rx->bna->bnad, rx); bna_rx_mcast_add() 768 mac = bna_cam_mod_mac_get(bna_mcam_mod_free_q(rxf->rx->bna)); bna_rx_mcast_add() 775 rxf->cam_fltr_cbarg = rx->bna->bnad; bna_rx_mcast_add() 783 bna_rx_ucast_listset(struct bna_rx *rx, int count, const u8 *uclist) bna_rx_ucast_listset() argument 785 struct bna_ucam_mod *ucam_mod = &rx->bna->ucam_mod; bna_rx_ucast_listset() 786 struct bna_rxf *rxf = &rx->rxf; bna_rx_ucast_listset() 841 bna_rx_mcast_listset(struct bna_rx *rx, int count, const u8 *mclist) bna_rx_mcast_listset() argument 843 struct bna_mcam_mod *mcam_mod = &rx->bna->mcam_mod; bna_rx_mcast_listset() 844 struct bna_rxf *rxf = &rx->rxf; bna_rx_mcast_listset() 901 bna_rx_mcast_delall(struct bna_rx *rx) bna_rx_mcast_delall() argument 903 struct bna_rxf *rxf = &rx->rxf; bna_rx_mcast_delall() 911 list_move_tail(&mac->qe, bna_mcam_mod_free_q(rxf->rx->bna)); bna_rx_mcast_delall() 919 del_mac = bna_cam_mod_mac_get(bna_mcam_mod_del_q(rxf->rx->bna)); bna_rx_mcast_delall() 923 list_add_tail(&mac->qe, bna_mcam_mod_free_q(rxf->rx->bna)); bna_rx_mcast_delall() 932 bna_rx_vlan_add(struct bna_rx *rx, int vlan_id) bna_rx_vlan_add() argument 934 struct bna_rxf *rxf = &rx->rxf; bna_rx_vlan_add() 947 bna_rx_vlan_del(struct bna_rx *rx, int vlan_id) bna_rx_vlan_del() argument 949 struct bna_rxf *rxf = &rx->rxf; bna_rx_vlan_del() 971 list_move_tail(&mac->qe, bna_ucam_mod_del_q(rxf->rx->bna)); bna_rxf_ucast_cfg_apply() 1009 bna_ucam_mod_del_q(rxf->rx->bna)); bna_rxf_ucast_cfg_reset() 1014 bna_ucam_mod_del_q(rxf->rx->bna)); bna_rxf_ucast_cfg_reset() 1047 struct bna *bna = rxf->rx->bna; bna_rxf_promisc_cfg_apply() 1075 struct bna *bna = rxf->rx->bna; bna_rxf_promisc_cfg_reset() 1161 struct bna *bna = rxf->rx->bna; bna_rxf_promisc_enable() 1177 bna->promisc_rid = rxf->rx->rid; bna_rxf_promisc_enable() 1187 struct bna *bna = rxf->rx->bna; bna_rxf_promisc_disable() 1278 #define call_rx_stop_cbfn(rx) \ 1280 if ((rx)->stop_cbfn) { \ 1283 cbfn = (rx)->stop_cbfn; \ 1284 cbarg = (rx)->stop_cbarg; \ 1285 (rx)->stop_cbfn = NULL; \ 1286 (rx)->stop_cbarg = NULL; \ 1287 cbfn(cbarg, rx); \ 1291 #define call_rx_stall_cbfn(rx) \ 1293 if ((rx)->rx_stall_cbfn) \ 1294 (rx)->rx_stall_cbfn((rx)->bna->bnad, (rx)); \ 1309 static void bna_bfi_rx_enet_start(struct bna_rx *rx); 1310 static void bna_rx_enet_stop(struct bna_rx *rx); 1311 static void bna_rx_mod_cb_rx_stopped(void *arg, struct bna_rx *rx); 1334 static void bna_rx_sm_stopped_entry(struct bna_rx *rx) bna_rx_sm_stopped_entry() argument 1336 call_rx_stop_cbfn(rx); bna_rx_sm_stopped_entry() 1339 static void bna_rx_sm_stopped(struct bna_rx *rx, bna_rx_sm_stopped() argument 1344 bfa_fsm_set_state(rx, bna_rx_sm_start_wait); bna_rx_sm_stopped() 1348 call_rx_stop_cbfn(rx); bna_rx_sm_stopped() 1361 static void bna_rx_sm_start_wait_entry(struct bna_rx *rx) bna_rx_sm_start_wait_entry() argument 1363 bna_bfi_rx_enet_start(rx); bna_rx_sm_start_wait_entry() 1367 bna_rx_sm_stop_wait_entry(struct bna_rx *rx) bna_rx_sm_stop_wait_entry() argument 1372 bna_rx_sm_stop_wait(struct bna_rx *rx, enum bna_rx_event event) bna_rx_sm_stop_wait() argument 1377 bfa_fsm_set_state(rx, bna_rx_sm_cleanup_wait); bna_rx_sm_stop_wait() 1378 rx->rx_cleanup_cbfn(rx->bna->bnad, rx); bna_rx_sm_stop_wait() 1382 bna_rx_enet_stop(rx); bna_rx_sm_stop_wait() 1391 static void bna_rx_sm_start_wait(struct bna_rx *rx, bna_rx_sm_start_wait() argument 1396 bfa_fsm_set_state(rx, bna_rx_sm_start_stop_wait); bna_rx_sm_start_wait() 1400 bfa_fsm_set_state(rx, bna_rx_sm_stopped); bna_rx_sm_start_wait() 1404 bfa_fsm_set_state(rx, bna_rx_sm_rxf_start_wait); bna_rx_sm_start_wait() 1413 static void bna_rx_sm_rxf_start_wait_entry(struct bna_rx *rx) bna_rx_sm_rxf_start_wait_entry() argument 1415 rx->rx_post_cbfn(rx->bna->bnad, rx); bna_rx_sm_rxf_start_wait_entry() 1416 bna_rxf_start(&rx->rxf); bna_rx_sm_rxf_start_wait_entry() 1420 bna_rx_sm_rxf_stop_wait_entry(struct bna_rx *rx) bna_rx_sm_rxf_stop_wait_entry() argument 1425 bna_rx_sm_rxf_stop_wait(struct bna_rx *rx, enum bna_rx_event event) bna_rx_sm_rxf_stop_wait() argument 1429 bfa_fsm_set_state(rx, bna_rx_sm_cleanup_wait); bna_rx_sm_rxf_stop_wait() 1430 bna_rxf_fail(&rx->rxf); bna_rx_sm_rxf_stop_wait() 1431 call_rx_stall_cbfn(rx); bna_rx_sm_rxf_stop_wait() 1432 rx->rx_cleanup_cbfn(rx->bna->bnad, rx); bna_rx_sm_rxf_stop_wait() 1436 bna_rxf_stop(&rx->rxf); bna_rx_sm_rxf_stop_wait() 1440 bfa_fsm_set_state(rx, bna_rx_sm_stop_wait); bna_rx_sm_rxf_stop_wait() 1441 call_rx_stall_cbfn(rx); bna_rx_sm_rxf_stop_wait() 1442 bna_rx_enet_stop(rx); bna_rx_sm_rxf_stop_wait() 1453 bna_rx_sm_start_stop_wait_entry(struct bna_rx *rx) bna_rx_sm_start_stop_wait_entry() argument 1458 bna_rx_sm_start_stop_wait(struct bna_rx *rx, enum bna_rx_event event) bna_rx_sm_start_stop_wait() argument 1463 bfa_fsm_set_state(rx, bna_rx_sm_stopped); bna_rx_sm_start_stop_wait() 1467 bna_rx_enet_stop(rx); bna_rx_sm_start_stop_wait() 1476 bna_rx_sm_started_entry(struct bna_rx *rx) bna_rx_sm_started_entry() argument 1479 int is_regular = (rx->type == BNA_RX_T_REGULAR); bna_rx_sm_started_entry() 1482 list_for_each_entry(rxp, &rx->rxp_q, qe) bna_rx_sm_started_entry() 1483 bna_ib_start(rx->bna, &rxp->cq.ib, is_regular); bna_rx_sm_started_entry() 1485 bna_ethport_cb_rx_started(&rx->bna->ethport); bna_rx_sm_started_entry() 1489 bna_rx_sm_started(struct bna_rx *rx, enum bna_rx_event event) bna_rx_sm_started() argument 1493 bfa_fsm_set_state(rx, bna_rx_sm_rxf_stop_wait); bna_rx_sm_started() 1494 bna_ethport_cb_rx_stopped(&rx->bna->ethport); bna_rx_sm_started() 1495 bna_rxf_stop(&rx->rxf); bna_rx_sm_started() 1499 bfa_fsm_set_state(rx, bna_rx_sm_failed); bna_rx_sm_started() 1500 bna_ethport_cb_rx_stopped(&rx->bna->ethport); bna_rx_sm_started() 1501 bna_rxf_fail(&rx->rxf); bna_rx_sm_started() 1502 call_rx_stall_cbfn(rx); bna_rx_sm_started() 1503 rx->rx_cleanup_cbfn(rx->bna->bnad, rx); bna_rx_sm_started() 1512 static void bna_rx_sm_rxf_start_wait(struct bna_rx *rx, bna_rx_sm_rxf_start_wait() argument 1517 bfa_fsm_set_state(rx, bna_rx_sm_rxf_stop_wait); bna_rx_sm_rxf_start_wait() 1521 bfa_fsm_set_state(rx, bna_rx_sm_failed); bna_rx_sm_rxf_start_wait() 1522 bna_rxf_fail(&rx->rxf); bna_rx_sm_rxf_start_wait() 1523 call_rx_stall_cbfn(rx); bna_rx_sm_rxf_start_wait() 1524 rx->rx_cleanup_cbfn(rx->bna->bnad, rx); bna_rx_sm_rxf_start_wait() 1528 bfa_fsm_set_state(rx, bna_rx_sm_started); bna_rx_sm_rxf_start_wait() 1538 bna_rx_sm_cleanup_wait_entry(struct bna_rx *rx) bna_rx_sm_cleanup_wait_entry() argument 1543 bna_rx_sm_cleanup_wait(struct bna_rx *rx, enum bna_rx_event event) bna_rx_sm_cleanup_wait() argument 1552 bfa_fsm_set_state(rx, bna_rx_sm_stopped); bna_rx_sm_cleanup_wait() 1562 bna_rx_sm_failed_entry(struct bna_rx *rx) bna_rx_sm_failed_entry() argument 1567 bna_rx_sm_failed(struct bna_rx *rx, enum bna_rx_event event) bna_rx_sm_failed() argument 1571 bfa_fsm_set_state(rx, bna_rx_sm_quiesce_wait); bna_rx_sm_failed() 1575 bfa_fsm_set_state(rx, bna_rx_sm_cleanup_wait); bna_rx_sm_failed() 1585 bfa_fsm_set_state(rx, bna_rx_sm_stopped); bna_rx_sm_failed() 1594 bna_rx_sm_quiesce_wait_entry(struct bna_rx *rx) bna_rx_sm_quiesce_wait_entry() argument 1599 bna_rx_sm_quiesce_wait(struct bna_rx *rx, enum bna_rx_event event) bna_rx_sm_quiesce_wait() argument 1603 bfa_fsm_set_state(rx, bna_rx_sm_cleanup_wait); bna_rx_sm_quiesce_wait() 1607 bfa_fsm_set_state(rx, bna_rx_sm_failed); bna_rx_sm_quiesce_wait() 1611 bfa_fsm_set_state(rx, bna_rx_sm_start_wait); bna_rx_sm_quiesce_wait() 1621 bna_bfi_rx_enet_start(struct bna_rx *rx) bna_bfi_rx_enet_start() argument 1623 struct bfi_enet_rx_cfg_req *cfg_req = &rx->bfi_enet_cmd.cfg_req; bna_bfi_rx_enet_start() 1629 BFI_ENET_H2I_RX_CFG_SET_REQ, 0, rx->rid); bna_bfi_rx_enet_start() 1633 cfg_req->rx_cfg.frame_size = bna_enet_mtu_get(&rx->bna->enet); bna_bfi_rx_enet_start() 1634 cfg_req->num_queue_sets = rx->num_paths; bna_bfi_rx_enet_start() 1635 for (i = 0; i < rx->num_paths; i++) { bna_bfi_rx_enet_start() 1637 : list_first_entry(&rx->rxp_q, struct bna_rxp, qe); bna_bfi_rx_enet_start() 1655 * a new rx with new set of resources. bna_bfi_rx_enet_start() 1663 bna_enet_mtu_get(&rx->bna->enet); bna_bfi_rx_enet_start() 1703 cfg_req->rx_cfg.hds.type = rx->hds_cfg.hdr_type; bna_bfi_rx_enet_start() 1704 cfg_req->rx_cfg.hds.force_offset = rx->hds_cfg.forced_offset; bna_bfi_rx_enet_start() 1705 cfg_req->rx_cfg.hds.max_header_size = rx->hds_cfg.forced_offset; bna_bfi_rx_enet_start() 1715 cfg_req->rx_cfg.strip_vlan = rx->rxf.vlan_strip_status; bna_bfi_rx_enet_start() 1717 bfa_msgq_cmd_set(&rx->msgq_cmd, NULL, NULL, bna_bfi_rx_enet_start() 1719 bfa_msgq_cmd_post(&rx->bna->msgq, &rx->msgq_cmd); bna_bfi_rx_enet_start() 1723 bna_bfi_rx_enet_stop(struct bna_rx *rx) bna_bfi_rx_enet_stop() argument 1725 struct bfi_enet_req *req = &rx->bfi_enet_cmd.req; bna_bfi_rx_enet_stop() 1728 BFI_ENET_H2I_RX_CFG_CLR_REQ, 0, rx->rid); bna_bfi_rx_enet_stop() 1731 bfa_msgq_cmd_set(&rx->msgq_cmd, NULL, NULL, sizeof(struct bfi_enet_req), bna_bfi_rx_enet_stop() 1733 bfa_msgq_cmd_post(&rx->bna->msgq, &rx->msgq_cmd); bna_bfi_rx_enet_stop() 1737 bna_rx_enet_stop(struct bna_rx *rx) bna_rx_enet_stop() argument 1742 list_for_each_entry(rxp, &rx->rxp_q, qe) bna_rx_enet_stop() 1743 bna_ib_stop(rx->bna, &rxp->cq.ib); bna_rx_enet_stop() 1745 bna_bfi_rx_enet_stop(rx); bna_rx_enet_stop() 1810 struct bna_rx *rx = NULL; bna_rx_get() local 1814 rx = list_first_entry(&rx_mod->rx_free_q, struct bna_rx, qe); bna_rx_get() 1816 rx = list_last_entry(&rx_mod->rx_free_q, struct bna_rx, qe); bna_rx_get() 1819 list_move_tail(&rx->qe, &rx_mod->rx_active_q); bna_rx_get() 1820 rx->type = type; bna_rx_get() 1822 return rx; bna_rx_get() 1826 bna_rx_put(struct bna_rx_mod *rx_mod, struct bna_rx *rx) bna_rx_put() argument 1831 if (((struct bna_rx *)qe)->rid < rx->rid) bna_rx_put() 1834 list_add(&rx->qe, qe); bna_rx_put() 1938 bna_rx_mod_cb_rx_stopped(void *arg, struct bna_rx *rx) bna_rx_mod_cb_rx_stopped() argument 1956 bna_rx_start(struct bna_rx *rx) bna_rx_start() argument 1958 rx->rx_flags |= BNA_RX_F_ENET_STARTED; bna_rx_start() 1959 if (rx->rx_flags & BNA_RX_F_ENABLED) bna_rx_start() 1960 bfa_fsm_send_event(rx, RX_E_START); bna_rx_start() 1964 bna_rx_stop(struct bna_rx *rx) bna_rx_stop() argument 1966 rx->rx_flags &= ~BNA_RX_F_ENET_STARTED; bna_rx_stop() 1967 if (rx->fsm == (bfa_fsm_t) bna_rx_sm_stopped) bna_rx_stop() 1968 bna_rx_mod_cb_rx_stopped(&rx->bna->rx_mod, rx); bna_rx_stop() 1970 rx->stop_cbfn = bna_rx_mod_cb_rx_stopped; bna_rx_stop() 1971 rx->stop_cbarg = &rx->bna->rx_mod; bna_rx_stop() 1972 bfa_fsm_send_event(rx, RX_E_STOP); bna_rx_stop() 1977 bna_rx_fail(struct bna_rx *rx) bna_rx_fail() argument 1980 rx->rx_flags &= ~BNA_RX_F_ENET_STARTED; bna_rx_fail() 1981 bfa_fsm_send_event(rx, RX_E_FAIL); bna_rx_fail() 1987 struct bna_rx *rx; bna_rx_mod_start() local 1993 list_for_each_entry(rx, &rx_mod->rx_active_q, qe) bna_rx_mod_start() 1994 if (rx->type == type) bna_rx_mod_start() 1995 bna_rx_start(rx); bna_rx_mod_start() 2001 struct bna_rx *rx; bna_rx_mod_stop() local 2010 list_for_each_entry(rx, &rx_mod->rx_active_q, qe) bna_rx_mod_stop() 2011 if (rx->type == type) { bna_rx_mod_stop() 2013 bna_rx_stop(rx); bna_rx_mod_stop() 2022 struct bna_rx *rx; bna_rx_mod_fail() local 2027 list_for_each_entry(rx, &rx_mod->rx_active_q, qe) bna_rx_mod_fail() 2028 bna_rx_fail(rx); bna_rx_mod_fail() 2042 rx_mod->rx = (struct bna_rx *) bna_rx_mod_init() 2060 rx_ptr = &rx_mod->rx[index]; bna_rx_mod_init() 2094 bna_bfi_rx_enet_start_rsp(struct bna_rx *rx, struct bfi_msgq_mhdr *msghdr) bna_bfi_rx_enet_start_rsp() argument 2096 struct bfi_enet_rx_cfg_rsp *cfg_rsp = &rx->bfi_enet_cmd.cfg_rsp; bna_bfi_rx_enet_start_rsp() 2101 bfa_msgq_rsp_copy(&rx->bna->msgq, (u8 *)cfg_rsp, bna_bfi_rx_enet_start_rsp() 2104 rx->hw_id = cfg_rsp->hw_id; bna_bfi_rx_enet_start_rsp() 2106 for (i = 0, rxp = list_first_entry(&rx->rxp_q, struct bna_rxp, qe); bna_bfi_rx_enet_start_rsp() 2107 i < rx->num_paths; i++, rxp = list_next_entry(rxp, qe)) { bna_bfi_rx_enet_start_rsp() 2112 rx->bna->pcidev.pci_bar_kva bna_bfi_rx_enet_start_rsp() 2116 rx->bna->pcidev.pci_bar_kva bna_bfi_rx_enet_start_rsp() 2121 rx->bna->pcidev.pci_bar_kva bna_bfi_rx_enet_start_rsp() 2134 bfa_fsm_send_event(rx, RX_E_STARTED); bna_bfi_rx_enet_start_rsp() 2138 bna_bfi_rx_enet_stop_rsp(struct bna_rx *rx, struct bfi_msgq_mhdr *msghdr) bna_bfi_rx_enet_stop_rsp() argument 2140 bfa_fsm_send_event(rx, RX_E_STOPPED); bna_bfi_rx_enet_stop_rsp() 2265 struct bna_rx *rx; bna_rx_create() local 2315 rx = bna_rx_get(rx_mod, rx_cfg->rx_type); bna_rx_create() 2316 rx->bna = bna; bna_rx_create() 2317 rx->rx_flags = 0; bna_rx_create() 2318 INIT_LIST_HEAD(&rx->rxp_q); bna_rx_create() 2319 rx->stop_cbfn = NULL; bna_rx_create() 2320 rx->stop_cbarg = NULL; bna_rx_create() 2321 rx->priv = priv; bna_rx_create() 2323 rx->rcb_setup_cbfn = rx_cbfn->rcb_setup_cbfn; bna_rx_create() 2324 rx->rcb_destroy_cbfn = rx_cbfn->rcb_destroy_cbfn; bna_rx_create() 2325 rx->ccb_setup_cbfn = rx_cbfn->ccb_setup_cbfn; bna_rx_create() 2326 rx->ccb_destroy_cbfn = rx_cbfn->ccb_destroy_cbfn; bna_rx_create() 2327 rx->rx_stall_cbfn = rx_cbfn->rx_stall_cbfn; bna_rx_create() 2329 rx->rx_cleanup_cbfn = rx_cbfn->rx_cleanup_cbfn; bna_rx_create() 2330 rx->rx_post_cbfn = rx_cbfn->rx_post_cbfn; bna_rx_create() 2332 if (rx->bna->rx_mod.flags & BNA_RX_MOD_F_ENET_STARTED) { bna_rx_create() 2333 switch (rx->type) { bna_rx_create() 2335 if (!(rx->bna->rx_mod.flags & bna_rx_create() 2337 rx->rx_flags |= BNA_RX_F_ENET_STARTED; bna_rx_create() 2340 if (rx->bna->rx_mod.flags & BNA_RX_MOD_F_ENET_LOOPBACK) bna_rx_create() 2341 rx->rx_flags |= BNA_RX_F_ENET_STARTED; bna_rx_create() 2346 rx->num_paths = rx_cfg->num_paths; bna_rx_create() 2348 i < rx->num_paths; i++) { bna_rx_create() 2350 list_add_tail(&rxp->qe, &rx->rxp_q); bna_rx_create() 2352 rxp->rx = rx; bna_rx_create() 2353 rxp->cq.rx = rx; bna_rx_create() 2387 q0->rx = rx; bna_rx_create() 2408 if (rx->rcb_setup_cbfn) bna_rx_create() 2409 rx->rcb_setup_cbfn(bnad, q0->rcb); bna_rx_create() 2414 q1->rx = rx; bna_rx_create() 2438 if (rx->rcb_setup_cbfn) bna_rx_create() 2439 rx->rcb_setup_cbfn(bnad, q1->rcb); bna_rx_create() 2475 if (rx->ccb_setup_cbfn) bna_rx_create() 2476 rx->ccb_setup_cbfn(bnad, rxp->cq.ccb); bna_rx_create() 2479 rx->hds_cfg = rx_cfg->hds_config; bna_rx_create() 2481 bna_rxf_init(&rx->rxf, rx, rx_cfg, res_info); bna_rx_create() 2483 bfa_fsm_set_state(rx, bna_rx_sm_stopped); bna_rx_create() 2485 rx_mod->rid_mask |= BIT(rx->rid); bna_rx_create() 2487 return rx; bna_rx_create() 2491 bna_rx_destroy(struct bna_rx *rx) bna_rx_destroy() argument 2493 struct bna_rx_mod *rx_mod = &rx->bna->rx_mod; bna_rx_destroy() 2499 bna_rxf_uninit(&rx->rxf); bna_rx_destroy() 2501 while (!list_empty(&rx->rxp_q)) { bna_rx_destroy() 2502 rxp = list_first_entry(&rx->rxp_q, struct bna_rxp, qe); bna_rx_destroy() 2505 if (rx->rcb_destroy_cbfn) bna_rx_destroy() 2506 rx->rcb_destroy_cbfn(rx->bna->bnad, q0->rcb); bna_rx_destroy() 2509 q0->rx = NULL; bna_rx_destroy() 2513 if (rx->rcb_destroy_cbfn) bna_rx_destroy() 2514 rx->rcb_destroy_cbfn(rx->bna->bnad, q1->rcb); bna_rx_destroy() 2517 q1->rx = NULL; bna_rx_destroy() 2523 if (rx->ccb_destroy_cbfn) bna_rx_destroy() 2524 rx->ccb_destroy_cbfn(rx->bna->bnad, rxp->cq.ccb); bna_rx_destroy() 2526 rxp->rx = NULL; bna_rx_destroy() 2531 if (qe == &rx->qe) { bna_rx_destroy() 2532 list_del(&rx->qe); bna_rx_destroy() 2536 rx_mod->rid_mask &= ~BIT(rx->rid); bna_rx_destroy() 2538 rx->bna = NULL; bna_rx_destroy() 2539 rx->priv = NULL; bna_rx_destroy() 2540 bna_rx_put(rx_mod, rx); bna_rx_destroy() 2544 bna_rx_enable(struct bna_rx *rx) bna_rx_enable() argument 2546 if (rx->fsm != (bfa_sm_t)bna_rx_sm_stopped) bna_rx_enable() 2549 rx->rx_flags |= BNA_RX_F_ENABLED; bna_rx_enable() 2550 if (rx->rx_flags & BNA_RX_F_ENET_STARTED) bna_rx_enable() 2551 bfa_fsm_send_event(rx, RX_E_START); bna_rx_enable() 2555 bna_rx_disable(struct bna_rx *rx, enum bna_cleanup_type type, bna_rx_disable() argument 2560 (*cbfn)(rx->bna->bnad, rx); bna_rx_disable() 2562 rx->stop_cbfn = cbfn; bna_rx_disable() 2563 rx->stop_cbarg = rx->bna->bnad; bna_rx_disable() 2565 rx->rx_flags &= ~BNA_RX_F_ENABLED; bna_rx_disable() 2567 bfa_fsm_send_event(rx, RX_E_STOP); bna_rx_disable() 2572 bna_rx_cleanup_complete(struct bna_rx *rx) bna_rx_cleanup_complete() argument 2574 bfa_fsm_send_event(rx, RX_E_CLEANUP_DONE); bna_rx_cleanup_complete() 2578 bna_rx_vlan_strip_enable(struct bna_rx *rx) bna_rx_vlan_strip_enable() argument 2580 struct bna_rxf *rxf = &rx->rxf; bna_rx_vlan_strip_enable() 2590 bna_rx_vlan_strip_disable(struct bna_rx *rx) bna_rx_vlan_strip_disable() argument 2592 struct bna_rxf *rxf = &rx->rxf; bna_rx_vlan_strip_disable() 2602 bna_rx_mode_set(struct bna_rx *rx, enum bna_rxmode new_mode, bna_rx_mode_set() argument 2605 struct bna_rxf *rxf = &rx->rxf; bna_rx_mode_set() 2612 if ((rx->bna->promisc_rid != BFI_INVALID_RID) && bna_rx_mode_set() 2613 (rx->bna->promisc_rid != rxf->rx->rid)) bna_rx_mode_set() 2617 if (rx->bna->default_mode_rid != BFI_INVALID_RID) bna_rx_mode_set() 2627 if ((rx->bna->default_mode_rid != BFI_INVALID_RID) && bna_rx_mode_set() 2628 (rx->bna->default_mode_rid != rxf->rx->rid)) { bna_rx_mode_set() 2633 if (rx->bna->promisc_rid != BFI_INVALID_RID) bna_rx_mode_set() 2659 rxf->cam_fltr_cbarg = rx->bna->bnad; bna_rx_mode_set() 2670 bna_rx_vlanfilter_enable(struct bna_rx *rx) bna_rx_vlanfilter_enable() argument 2672 struct bna_rxf *rxf = &rx->rxf; bna_rx_vlanfilter_enable() 2682 bna_rx_coalescing_timeo_set(struct bna_rx *rx, int coalescing_timeo) bna_rx_coalescing_timeo_set() argument 2686 list_for_each_entry(rxp, &rx->rxp_q, qe) { bna_rx_coalescing_timeo_set() 2705 struct bna *bna = ccb->cq->rx->bna; bna_rx_dim_update() 616 bna_rxf_init(struct bna_rxf *rxf, struct bna_rx *rx, struct bna_rx_config *q_config, struct bna_res_info *res_info) bna_rxf_init() argument
|
H A D | bna.h | 118 cbfn(cbarg, rxf->rx); \ 330 void bna_bfi_rx_enet_start_rsp(struct bna_rx *rx, 332 void bna_bfi_rx_enet_stop_rsp(struct bna_rx *rx, 357 void bna_rx_destroy(struct bna_rx *rx); 358 void bna_rx_enable(struct bna_rx *rx); 359 void bna_rx_disable(struct bna_rx *rx, enum bna_cleanup_type type, 361 void bna_rx_cleanup_complete(struct bna_rx *rx); 362 void bna_rx_coalescing_timeo_set(struct bna_rx *rx, int coalescing_timeo); 365 enum bna_cb_status bna_rx_ucast_set(struct bna_rx *rx, const u8 *ucmac); 366 enum bna_cb_status bna_rx_ucast_listset(struct bna_rx *rx, int count, 368 enum bna_cb_status bna_rx_mcast_add(struct bna_rx *rx, const u8 *mcmac, 371 enum bna_cb_status bna_rx_mcast_listset(struct bna_rx *rx, int count, 374 bna_rx_mcast_delall(struct bna_rx *rx); 376 bna_rx_mode_set(struct bna_rx *rx, enum bna_rxmode rxmode, 378 void bna_rx_vlan_add(struct bna_rx *rx, int vlan_id); 379 void bna_rx_vlan_del(struct bna_rx *rx, int vlan_id); 380 void bna_rx_vlanfilter_enable(struct bna_rx *rx); 381 void bna_rx_vlan_strip_enable(struct bna_rx *rx); 382 void bna_rx_vlan_strip_disable(struct bna_rx *rx);
|
H A D | bna_types.h | 582 struct bna_rx *rx; member in struct:bna_rxq 648 struct bna_rx *rx; member in struct:bna_cq 701 struct bna_rx *rx; member in struct:bna_rxp 724 void (*start_cbfn) (struct bna_rx *rx); 728 void (*stop_cbfn) (struct bna_rx *rx); 738 void (*cam_fltr_cbfn)(struct bnad *bnad, struct bna_rx *rx); 776 struct bna_rx *rx; member in struct:bna_rxf 816 void (*stop_cbfn)(void *arg, struct bna_rx *rx); 835 /* Rx module - keeps track of free, active rx objects */ 838 struct bna_rx *rx; /* BFI_MAX_RXQ entries */ member in struct:bna_rx_mod
|
H A D | bnad.c | 850 if (!rx_info->rx) bnad_isr() 1037 (struct bnad_rx_info *)ccb->cq->rx->priv; bnad_cb_ccb_setup() 1047 (struct bnad_rx_info *)ccb->cq->rx->priv; bnad_cb_ccb_destroy() 1165 bnad_cb_rx_stall(struct bnad *bnad, struct bna_rx *rx) bnad_cb_rx_stall() argument 1167 struct bnad_rx_info *rx_info = (struct bnad_rx_info *)rx->priv; bnad_cb_rx_stall() 1219 bna_rx_cleanup_complete(rx_info->rx); bnad_rx_cleanup() 1224 bnad_cb_rx_cleanup(struct bnad *bnad, struct bna_rx *rx) bnad_cb_rx_cleanup() argument 1226 struct bnad_rx_info *rx_info = (struct bnad_rx_info *)rx->priv; bnad_cb_rx_cleanup() 1247 bnad_cb_rx_post(struct bnad *bnad, struct bna_rx *rx) bnad_cb_rx_post() argument 1249 struct bnad_rx_info *rx_info = (struct bnad_rx_info *)rx->priv; bnad_cb_rx_post() 1277 bnad_cb_rx_disabled(void *arg, struct bna_rx *rx) bnad_cb_rx_disabled() argument 1285 bnad_cb_rx_mcast_add(struct bnad *bnad, struct bna_rx *rx) bnad_cb_rx_mcast_add() argument 1772 if (!rx_info->rx) bnad_dim_timeout() 2041 /* Setup the rx config for bna_rx_create */ 2125 /* destroy and create new rx objects */ bnad_reinit_rx() 2127 if (!bnad->rx_info[rx_id].rx) bnad_reinit_rx() 2146 /* restore rx configuration */ bnad_reinit_rx() 2147 if (bnad->rx_info[0].rx && !err) { bnad_reinit_rx() 2169 if (!rx_info->rx) bnad_destroy_rx() 2186 bna_rx_disable(rx_info->rx, BNA_HARD_CLEANUP, bnad_cb_rx_disabled); bnad_destroy_rx() 2196 bna_rx_destroy(rx_info->rx); bnad_destroy_rx() 2198 rx_info->rx = NULL; bnad_destroy_rx() 2224 struct bna_rx *rx; bnad_setup_rx() local 2260 rx = bna_rx_create(&bnad->bna, bnad, rx_config, &rx_cbfn, res_info, bnad_setup_rx() 2262 if (!rx) { bnad_setup_rx() 2267 rx_info->rx = rx; bnad_setup_rx() 2294 bna_rx_vlanfilter_enable(rx); bnad_setup_rx() 2300 bna_rx_enable(rx); bnad_setup_rx() 2332 if (!rx_info->rx) bnad_rx_coalescing_timeo_set() 2334 bna_rx_coalescing_timeo_set(rx_info->rx, bnad_rx_coalescing_timeo_set() 2351 if (!bnad->rx_info[0].rx) bnad_mac_addr_set_locked() 2354 ret = bna_rx_ucast_set(bnad->rx_info[0].rx, mac_addr); bnad_mac_addr_set_locked() 2372 ret = bna_rx_mcast_add(rx_info->rx, bnad_bcast_addr, bnad_enable_default_bcast() 2396 bna_rx_vlan_add(bnad->rx_info[rx_id].rx, vid); bnad_restore_vlans() 3149 bna_rx_ucast_listset(bnad->rx_info[0].rx, 0, NULL); bnad_set_rx_ucast_fltr() 3166 ret = bna_rx_ucast_listset(bnad->rx_info[0].rx, entry, mac_list); 3177 bna_rx_ucast_listset(bnad->rx_info[0].rx, 0, NULL); 3206 ret = bna_rx_mcast_listset(bnad->rx_info[0].rx, mc_count + 1, mac_list); bnad_set_rx_mcast_fltr() 3216 bna_rx_mcast_delall(bnad->rx_info[0].rx); bnad_set_rx_mcast_fltr() 3228 if (bnad->rx_info[0].rx == NULL) { bnad_set_rx_mode() 3255 bna_rx_mode_set(bnad->rx_info[0].rx, new_mode, mode_mask); bnad_set_rx_mode() 3327 /* rx_count > 0 - new rx created bnad_change_mtu() 3344 if (!bnad->rx_info[0].rx) bnad_vlan_rx_add_vid() 3350 bna_rx_vlan_add(bnad->rx_info[0].rx, vid); bnad_vlan_rx_add_vid() 3365 if (!bnad->rx_info[0].rx) bnad_vlan_rx_kill_vid() 3372 bna_rx_vlan_del(bnad->rx_info[0].rx, vid); bnad_vlan_rx_kill_vid() 3391 bna_rx_vlan_strip_enable(bnad->rx_info[0].rx); bnad_set_features() 3393 bna_rx_vlan_strip_disable(bnad->rx_info[0].rx); bnad_set_features() 3424 if (!rx_info->rx) bnad_netpoll()
|
/linux-4.4.14/drivers/staging/gdm72xx/ |
H A D | gdm_usb.c | 82 static struct usb_rx *alloc_rx_struct(struct rx_cxt *rx) alloc_rx_struct() argument 98 r->rx_cxt = rx; alloc_rx_struct() 136 static struct usb_rx *get_rx_struct(struct rx_cxt *rx) get_rx_struct() argument 140 if (list_empty(&rx->free_list)) { get_rx_struct() 141 r = alloc_rx_struct(rx); get_rx_struct() 145 list_add(&r->list, &rx->free_list); get_rx_struct() 148 r = list_entry(rx->free_list.next, struct usb_rx, list); get_rx_struct() 149 list_move_tail(&r->list, &rx->used_list); get_rx_struct() 155 static void put_rx_struct(struct rx_cxt *rx, struct usb_rx *r) put_rx_struct() argument 157 list_move(&r->list, &rx->free_list); put_rx_struct() 163 struct rx_cxt *rx = &udev->rx; release_usb() local 187 spin_lock_irqsave(&rx->lock, flags); release_usb() 189 list_for_each_entry_safe(r, r_next, &rx->free_list, list) { release_usb() 194 list_for_each_entry_safe(r, r_next, &rx->used_list, list) { release_usb() 199 spin_unlock_irqrestore(&rx->lock, flags); release_usb() 206 struct rx_cxt *rx = &udev->rx; init_usb() local 218 INIT_LIST_HEAD(&rx->free_list); init_usb() 219 INIT_LIST_HEAD(&rx->used_list); init_usb() 222 spin_lock_init(&rx->lock); init_usb() 236 r = alloc_rx_struct(rx); init_usb() 242 spin_lock_irqsave(&rx->lock, flags); init_usb() 243 list_add(&r->list, &rx->free_list); init_usb() 244 spin_unlock_irqrestore(&rx->lock, flags); init_usb() 359 struct rx_cxt *rx; gdm_usb_send() local 362 rx = &udev->rx; gdm_usb_send() 364 spin_lock_irqsave(&rx->lock, flags2); gdm_usb_send() 365 list_for_each_entry(r, &rx->used_list, list) gdm_usb_send() 367 spin_unlock_irqrestore(&rx->lock, flags2); gdm_usb_send() 407 struct rx_cxt *rx = r->rx_cxt; gdm_usb_rcv_complete() local 408 struct usbwm_dev *udev = container_of(r->rx_cxt, struct usbwm_dev, rx); gdm_usb_rcv_complete() 451 spin_lock_irqsave(&rx->lock, flags2); gdm_usb_rcv_complete() 452 put_rx_struct(rx, r); gdm_usb_rcv_complete() 453 spin_unlock_irqrestore(&rx->lock, flags2); gdm_usb_rcv_complete() 468 struct rx_cxt *rx = &udev->rx; gdm_usb_receive() local 477 spin_lock_irqsave(&rx->lock, flags); gdm_usb_receive() 478 r = get_rx_struct(rx); gdm_usb_receive() 479 spin_unlock_irqrestore(&rx->lock, flags); gdm_usb_receive() 646 struct rx_cxt *rx; gdm_suspend() local 655 rx = &udev->rx; gdm_suspend() 657 spin_lock_irqsave(&rx->lock, flags); gdm_suspend() 659 list_for_each_entry(r, &rx->used_list, list) gdm_suspend() 662 spin_unlock_irqrestore(&rx->lock, flags); gdm_suspend() 671 struct rx_cxt *rx; gdm_resume() local 680 rx = &udev->rx; gdm_resume() 682 spin_lock_irqsave(&rx->lock, flags); gdm_resume() 684 list_for_each_entry(r, &rx->used_list, list) gdm_resume() 687 spin_unlock_irqrestore(&rx->lock, flags); gdm_resume() 699 struct rx_cxt *rx; k_mode_thread() local 710 rx = &udev->rx; k_mode_thread() 719 spin_lock_irqsave(&rx->lock, flags); k_mode_thread() 721 list_for_each_entry(r, &rx->used_list, list) k_mode_thread() 724 spin_unlock_irqrestore(&rx->lock, flags); k_mode_thread()
|
H A D | gdm_sdio.c | 67 static struct sdio_rx *alloc_rx_struct(struct rx_cxt *rx) alloc_rx_struct() argument 72 r->rx_cxt = rx; alloc_rx_struct() 105 static struct sdio_rx *get_rx_struct(struct rx_cxt *rx) get_rx_struct() argument 109 if (list_empty(&rx->free_list)) get_rx_struct() 112 r = list_entry(rx->free_list.prev, struct sdio_rx, list); get_rx_struct() 119 static void put_rx_struct(struct rx_cxt *rx, struct sdio_rx *r) put_rx_struct() argument 121 list_add_tail(&r->list, &rx->free_list); put_rx_struct() 127 struct rx_cxt *rx = &sdev->rx; release_sdio() local 148 kfree(rx->rx_buf); release_sdio() 150 list_for_each_entry_safe(r, r_next, &rx->free_list, list) { release_sdio() 155 list_for_each_entry_safe(r, r_next, &rx->req_list, list) { release_sdio() 165 struct rx_cxt *rx = &sdev->rx; init_sdio() local 188 INIT_LIST_HEAD(&rx->free_list); init_sdio() 189 INIT_LIST_HEAD(&rx->req_list); init_sdio() 191 spin_lock_init(&rx->lock); init_sdio() 194 r = alloc_rx_struct(rx); init_sdio() 199 list_add(&r->list, &rx->free_list); init_sdio() 202 rx->rx_buf = kmalloc(RX_BUF_SIZE, GFP_KERNEL); init_sdio() 203 if (!rx->rx_buf) init_sdio() 477 struct rx_cxt *rx = &sdev->rx; gdm_sdio_irq() local 517 memcpy(rx->rx_buf, hdr + TYPE_A_HEADER_SIZE, gdm_sdio_irq() 520 buf = rx->rx_buf + TYPE_A_LOOKAHEAD_SIZE - TYPE_A_HEADER_SIZE; gdm_sdio_irq() 549 dev_dbg(&func->dev, "sdio_receive: %*ph\n", len, rx->rx_buf); gdm_sdio_irq() 551 len = control_sdu_tx_flow(sdev, rx->rx_buf, len); gdm_sdio_irq() 553 spin_lock_irqsave(&rx->lock, flags); gdm_sdio_irq() 555 if (!list_empty(&rx->req_list)) { gdm_sdio_irq() 556 r = list_entry(rx->req_list.next, struct sdio_rx, list); gdm_sdio_irq() 557 spin_unlock_irqrestore(&rx->lock, flags); gdm_sdio_irq() 559 r->callback(r->cb_data, rx->rx_buf, len); gdm_sdio_irq() 560 spin_lock_irqsave(&rx->lock, flags); gdm_sdio_irq() 562 put_rx_struct(rx, r); gdm_sdio_irq() 565 spin_unlock_irqrestore(&rx->lock, flags); gdm_sdio_irq() 578 struct rx_cxt *rx = &sdev->rx; gdm_sdio_receive() local 582 spin_lock_irqsave(&rx->lock, flags); gdm_sdio_receive() 583 r = get_rx_struct(rx); gdm_sdio_receive() 585 spin_unlock_irqrestore(&rx->lock, flags); gdm_sdio_receive() 592 list_add_tail(&r->list, &rx->req_list); gdm_sdio_receive() 593 spin_unlock_irqrestore(&rx->lock, flags); gdm_sdio_receive()
|
/linux-4.4.14/drivers/media/radio/wl128x/ |
H A D | fmdrv_rx.c | 30 fmdev->rx.rds.flag = FM_RDS_DISABLE; fm_rx_reset_rds_cache() 31 fmdev->rx.rds.last_blk_idx = 0; fm_rx_reset_rds_cache() 32 fmdev->rx.rds.wr_idx = 0; fm_rx_reset_rds_cache() 33 fmdev->rx.rds.rd_idx = 0; fm_rx_reset_rds_cache() 35 if (fmdev->rx.af_mode == FM_RX_RDS_AF_SWITCH_MODE_ON) fm_rx_reset_rds_cache() 41 fmdev->rx.stat_info.picode = FM_NO_PI_CODE; fm_rx_reset_station_info() 42 fmdev->rx.stat_info.afcache_size = 0; fm_rx_reset_station_info() 43 fmdev->rx.stat_info.af_list_max = 0; fm_rx_reset_station_info() 54 if (freq < fmdev->rx.region.bot_freq || freq > fmdev->rx.region.top_freq) { fm_rx_set_freq() 75 payload = (freq - fmdev->rx.region.bot_freq) / FM_FREQ_MUL; fm_rx_set_freq() 120 curr_frq_in_khz = (fmdev->rx.region.bot_freq + ((u32)curr_frq * FM_FREQ_MUL)); fm_rx_set_freq() 128 fmdev->rx.freq = curr_frq_in_khz; fm_rx_set_freq() 164 fmdev->rx.region.chanl_space = spacing * FM_FREQ_MUL; fm_rx_set_channel_spacing() 193 last_frq = (fmdev->rx.region.top_freq - fmdev->rx.region.bot_freq) / FM_FREQ_MUL; fm_rx_seek() 196 space_idx = fmdev->rx.region.chanl_space / FM_FREQ_MUL; fm_rx_seek() 269 fmdev->rx.freq = seek_upward ? fm_rx_seek() 270 fmdev->rx.region.top_freq : fm_rx_seek() 271 fmdev->rx.region.bot_freq; fm_rx_seek() 273 fmdev->rx.freq = seek_upward ? fm_rx_seek() 274 fmdev->rx.region.bot_freq : fm_rx_seek() 275 fmdev->rx.region.top_freq; fm_rx_seek() 277 next_frq = (fmdev->rx.freq - fm_rx_seek() 278 fmdev->rx.region.bot_freq) / FM_FREQ_MUL; fm_rx_seek() 289 fmdev->rx.freq = (fmdev->rx.region.bot_freq + fm_rx_seek() 321 fmdev->rx.volume = vol_to_set; fm_rx_set_volume() 336 *curr_vol = fmdev->rx.volume / FM_RX_VOLUME_GAIN_STEP; fm_rx_get_volume() 345 *bot_freq = fmdev->rx.region.bot_freq; fm_rx_get_band_freq_range() 348 *top_freq = fmdev->rx.region.top_freq; fm_rx_get_band_freq_range() 356 *region = fmdev->rx.region.fm_band; fm_rx_get_region() 372 if (fmdev->rx.region.fm_band == region_to_set) { fm_rx_set_region() 387 if (fmdev->rx.freq < fmdev->rx.region.bot_freq) fm_rx_set_region() 388 new_frq = fmdev->rx.region.bot_freq; fm_rx_set_region() 389 else if (fmdev->rx.freq > fmdev->rx.region.top_freq) fm_rx_set_region() 390 new_frq = fmdev->rx.region.top_freq; fm_rx_set_region() 413 *curr_mute_mode = fmdev->rx.mute_mode; fm_rx_get_mute_mode() 424 switch (fmdev->rx.mute_mode) { fm_config_rx_mute_reg() 437 if (fmdev->rx.rf_depend_mute == FM_RX_RF_DEPENDENT_MUTE_ON) fm_config_rx_mute_reg() 457 if (fmdev->rx.mute_mode == mute_mode_toset) fm_rx_set_mute_mode() 460 org_state = fmdev->rx.mute_mode; fm_rx_set_mute_mode() 461 fmdev->rx.mute_mode = mute_mode_toset; fm_rx_set_mute_mode() 465 fmdev->rx.mute_mode = org_state; fm_rx_set_mute_mode() 483 *curr_mute_mode = fmdev->rx.rf_depend_mute; fm_rx_get_rfdepend_softmute() 502 if (fmdev->rx.rf_depend_mute == rfdepend_mute) fm_rx_set_rfdepend_softmute() 505 org_state = fmdev->rx.rf_depend_mute; fm_rx_set_rfdepend_softmute() 506 fmdev->rx.rf_depend_mute = rfdepend_mute; fm_rx_set_rfdepend_softmute() 510 fmdev->rx.rf_depend_mute = org_state; fm_rx_set_rfdepend_softmute() 559 fmdev->rx.rssi_threshold = rssi_lvl_toset; fm_rx_set_rssi_threshold() 575 *curr_rssi_lvl = fmdev->rx.rssi_threshold; fm_rx_get_rssi_threshold() 641 fmerr("Invalid rx de-emphasis mode (%d)\n", mode); fm_rx_set_deemphasis_mode() 651 fmdev->rx.deemphasis_mode = mode; fm_rx_set_deemphasis_mode() 667 *curr_deemphasis_mode = fmdev->rx.deemphasis_mode; fm_rx_get_deemph_mode() 684 && fmdev->rx.rds.flag == FM_RDS_DISABLE) { fm_rx_set_rds_mode() 723 fmdev->rx.rds.flag = FM_RDS_ENABLE; fm_rx_set_rds_mode() 725 && fmdev->rx.rds.flag == FM_RDS_ENABLE) { fm_rx_set_rds_mode() 734 fmdev->rx.rds.last_blk_idx = 0; fm_rx_set_rds_mode() 735 fmdev->rx.rds.wr_idx = 0; fm_rx_set_rds_mode() 736 fmdev->rx.rds.rd_idx = 0; fm_rx_set_rds_mode() 741 fmdev->rx.rds.flag = FM_RDS_DISABLE; fm_rx_set_rds_mode() 758 *curr_rds_en_dis = fmdev->rx.rds.flag; fm_rx_get_rds_mode() 783 fmdev->rx.rds_mode = rds_mode; fm_rx_set_rds_system() 814 fmdev->rx.af_mode = af_mode; fm_rx_set_af_switch() 830 *af_mode = fmdev->rx.af_mode; fm_rx_get_af_switch()
|
H A D | fmdrv_common.c | 253 fmdev->rx.region = region_configs[region_to_set]; fmc_update_region_info() 638 struct tuned_station_info *stat_info = &fmdev->rx.stat_info; fm_rx_update_af_cache() 639 u8 reg_idx = fmdev->rx.region.fm_band; fm_rx_update_af_cache() 645 fmdev->rx.stat_info.af_list_max = (af - FM_RDS_1_AF_FOLLOWS + 1); fm_rx_update_af_cache() 646 fmdev->rx.stat_info.afcache_size = 0; fm_rx_update_af_cache() 647 fmdbg("No of expected AF : %d\n", fmdev->rx.stat_info.af_list_max); fm_rx_update_af_cache() 658 freq = fmdev->rx.region.bot_freq + (af * 100); fm_rx_update_af_cache() 659 if (freq == fmdev->rx.freq) { fm_rx_update_af_cache() 661 fmdev->rx.freq, freq); fm_rx_update_af_cache() 713 struct fm_rds *rds = &fmdev->rx.rds; fm_irq_handle_rdsdata_getcmd_resp() 765 if (fmdev->rx.stat_info.picode != cur_picode) fm_irq_handle_rdsdata_getcmd_resp() 766 fmdev->rx.stat_info.picode = cur_picode; fm_irq_handle_rdsdata_getcmd_resp() 859 if ((fmdev->rx.af_mode == FM_RX_RDS_AF_SWITCH_MODE_ON) && fm_irq_handle_low_rssi_start() 861 (fmdev->rx.freq != FM_UNDEFINED_FREQ) && fm_irq_handle_low_rssi_start() 862 (fmdev->rx.stat_info.afcache_size != 0)) { fm_irq_handle_low_rssi_start() 868 fmdev->rx.afjump_idx = 0; fm_irq_handle_low_rssi_start() 869 fmdev->rx.freq_before_jump = fmdev->rx.freq; fm_irq_handle_low_rssi_start() 884 payload = fmdev->rx.stat_info.picode; fm_irq_afjump_set_pi() 918 fmdbg("Swtich to %d KHz\n", fmdev->rx.stat_info.af_cache[fmdev->rx.afjump_idx]); fm_irq_afjump_setfreq() 919 frq_index = (fmdev->rx.stat_info.af_cache[fmdev->rx.afjump_idx] - fm_irq_afjump_setfreq() 920 fmdev->rx.region.bot_freq) / FM_FREQ_MUL; fm_irq_afjump_setfreq() 990 curr_freq = fmdev->rx.region.bot_freq + ((u32)read_freq * FM_FREQ_MUL); fm_irq_afjump_rd_freq_resp() 992 jumped_freq = fmdev->rx.stat_info.af_cache[fmdev->rx.afjump_idx]; fm_irq_afjump_rd_freq_resp() 995 if ((curr_freq != fmdev->rx.freq_before_jump) && (curr_freq == jumped_freq)) { fm_irq_afjump_rd_freq_resp() 997 fmdev->rx.freq = curr_freq; fm_irq_afjump_rd_freq_resp() 1001 if (fmdev->rx.af_mode == FM_RX_RDS_AF_SWITCH_MODE_ON) fm_irq_afjump_rd_freq_resp() 1006 fmdev->rx.afjump_idx++; fm_irq_afjump_rd_freq_resp() 1009 if (fmdev->rx.afjump_idx >= fmdev->rx.stat_info.afcache_size) { fm_irq_afjump_rd_freq_resp() 1061 poll_wait(file, &fmdev->rx.rds.read_queue, pts); fmc_is_rds_data_available() 1062 if (fmdev->rx.rds.rd_idx != fmdev->rx.rds.wr_idx) fmc_is_rds_data_available() 1077 if (fmdev->rx.rds.wr_idx == fmdev->rx.rds.rd_idx) { fmc_transfer_rds_from_internal_buff() 1081 ret = wait_event_interruptible(fmdev->rx.rds.read_queue, fmc_transfer_rds_from_internal_buff() 1082 (fmdev->rx.rds.wr_idx != fmdev->rx.rds.rd_idx)); fmc_transfer_rds_from_internal_buff() 1095 if (fmdev->rx.rds.wr_idx == fmdev->rx.rds.rd_idx) { fmc_transfer_rds_from_internal_buff() 1099 memcpy(tmpbuf, &fmdev->rx.rds.buff[fmdev->rx.rds.rd_idx], fmc_transfer_rds_from_internal_buff() 1101 fmdev->rx.rds.rd_idx += FM_RDS_BLK_SIZE; fmc_transfer_rds_from_internal_buff() 1102 if (fmdev->rx.rds.rd_idx >= fmdev->rx.rds.buf_size) fmc_transfer_rds_from_internal_buff() 1103 fmdev->rx.rds.rd_idx = 0; fmc_transfer_rds_from_internal_buff() 1133 if (fmdev->rx.freq == FM_UNDEFINED_FREQ) { fmc_get_freq() 1144 *cur_tuned_frq = fmdev->rx.freq; fmc_get_freq() 1422 fmdbg("Loading default rx configuration..\n"); fmc_set_mode() 1568 fmdev->rx.region = region_configs[default_radio_region]; fmc_prepare() 1570 fmdev->rx.mute_mode = FM_MUTE_OFF; fmc_prepare() 1571 fmdev->rx.rf_depend_mute = FM_RX_RF_DEPENDENT_MUTE_OFF; fmc_prepare() 1572 fmdev->rx.rds.flag = FM_RDS_DISABLE; fmc_prepare() 1573 fmdev->rx.freq = FM_UNDEFINED_FREQ; fmc_prepare() 1574 fmdev->rx.rds_mode = FM_RDS_SYSTEM_RDS; fmc_prepare() 1575 fmdev->rx.af_mode = FM_RX_RDS_AF_SWITCH_MODE_OFF; fmc_prepare() 1579 init_waitqueue_head(&fmdev->rx.rds.read_queue); fmc_prepare() 1601 wake_up_interruptible(&fmdev->rx.rds.read_queue); fmc_release() 1610 fmdev->rx.freq = 0; fmc_release() 1642 fmdev->rx.rds.buf_size = default_rds_buf * FM_RDS_BLK_SIZE; fm_drv_init() 1643 fmdev->rx.rds.buff = kzalloc(fmdev->rx.rds.buf_size, GFP_KERNEL); fm_drv_init() 1644 if (NULL == fmdev->rx.rds.buff) { fm_drv_init() 1660 kfree(fmdev->rx.rds.buff); fm_drv_init() 1674 kfree(fmdev->rx.rds.buff); fm_drv_exit()
|
/linux-4.4.14/drivers/net/wireless/ti/wl1251/ |
H A D | Makefile | 1 wl1251-objs = main.o event.o tx.o rx.o ps.o cmd.o \
|
H A D | debugfs.c | 122 DEBUGFS_FWSTATS_FILE(rx, out_of_mem, 20, "%u"); 123 DEBUGFS_FWSTATS_FILE(rx, hdr_overflow, 20, "%u"); 124 DEBUGFS_FWSTATS_FILE(rx, hw_stuck, 20, "%u"); 125 DEBUGFS_FWSTATS_FILE(rx, dropped, 20, "%u"); 126 DEBUGFS_FWSTATS_FILE(rx, fcs_err, 20, "%u"); 127 DEBUGFS_FWSTATS_FILE(rx, xfr_hint_trig, 20, "%u"); 128 DEBUGFS_FWSTATS_FILE(rx, path_reset, 20, "%u"); 129 DEBUGFS_FWSTATS_FILE(rx, reset_counter, 20, "%u"); 262 DEBUGFS_FWSTATS_DEL(rx, out_of_mem); wl1251_debugfs_delete_files() 263 DEBUGFS_FWSTATS_DEL(rx, hdr_overflow); wl1251_debugfs_delete_files() 264 DEBUGFS_FWSTATS_DEL(rx, hw_stuck); wl1251_debugfs_delete_files() 265 DEBUGFS_FWSTATS_DEL(rx, dropped); wl1251_debugfs_delete_files() 266 DEBUGFS_FWSTATS_DEL(rx, fcs_err); wl1251_debugfs_delete_files() 267 DEBUGFS_FWSTATS_DEL(rx, xfr_hint_trig); wl1251_debugfs_delete_files() 268 DEBUGFS_FWSTATS_DEL(rx, path_reset); wl1251_debugfs_delete_files() 269 DEBUGFS_FWSTATS_DEL(rx, reset_counter); wl1251_debugfs_delete_files() 363 DEBUGFS_FWSTATS_ADD(rx, out_of_mem); wl1251_debugfs_add_files() 364 DEBUGFS_FWSTATS_ADD(rx, hdr_overflow); wl1251_debugfs_add_files() 365 DEBUGFS_FWSTATS_ADD(rx, hw_stuck); wl1251_debugfs_add_files() 366 DEBUGFS_FWSTATS_ADD(rx, dropped); wl1251_debugfs_add_files() 367 DEBUGFS_FWSTATS_ADD(rx, fcs_err); wl1251_debugfs_add_files() 368 DEBUGFS_FWSTATS_ADD(rx, xfr_hint_trig); wl1251_debugfs_add_files() 369 DEBUGFS_FWSTATS_ADD(rx, path_reset); wl1251_debugfs_add_files() 370 DEBUGFS_FWSTATS_ADD(rx, reset_counter); wl1251_debugfs_add_files()
|
/linux-4.4.14/drivers/net/wireless/ti/wlcore/ |
H A D | Makefile | 1 wlcore-objs = main.o cmd.o io.o event.o tx.o rx.o ps.o acx.o \
|
H A D | rx.c | 30 #include "rx.h" 168 /* skb length not including rx descriptor */ wl1271_rx_handle_data() 181 * Copy packets from aggregation buffer to the skbs without rx wl1271_rx_handle_data() 202 wl1271_debug(DEBUG_RX, "rx skb 0x%p: %d B %s seq %d hlid %d", skb, wl1271_rx_handle_data() 314 "enabled rx filter %d", index); wl1271_rx_filter_enable() 321 wl1271_error("Failed to %s rx data filter %d (err=%d)", wl1271_rx_filter_enable()
|
/linux-4.4.14/arch/arm/kernel/ |
H A D | debug.S | 28 .macro addruart_current, rx, tmp1, tmp2 29 addruart \tmp1, \tmp2, \rx 30 mrc p15, 0, \rx, c1, c0 31 tst \rx, #1 32 moveq \rx, \tmp1 33 movne \rx, \tmp2 37 .macro addruart_current, rx, tmp1, tmp2 38 addruart \rx, \tmp1, \tmp2
|
/linux-4.4.14/sound/arm/ |
H A D | aaci.h | 77 #define SR_RXTOFE (1 << 11) /* rx timeout fifo empty */ 78 #define SR_TXTO (1 << 10) /* rx timeout fifo nonempty */ 80 #define SR_RXO (1 << 8) /* rx overrun */ 82 #define SR_RXB (1 << 6) /* rx busy */ 84 #define SR_RXFF (1 << 4) /* rx fifo full */ 86 #define SR_RXHF (1 << 2) /* rx fifo half full */ 88 #define SR_RXFE (1 << 0) /* rx fifo empty */ 93 #define ISR_RXTOFEINTR (1 << 6) /* rx fifo empty */ 95 #define ISR_ORINTR (1 << 4) /* rx overflow */ 96 #define ISR_RXINTR (1 << 3) /* rx fifo */ 115 #define ISR_RXTOFE (1 << 6) /* rx timeout fifo empty */ 117 #define ISR_OR (1 << 4) /* rx fifo overrun */ 118 #define ISR_RX (1 << 3) /* rx interrupt status */ 120 #define ISR_RXTO (1 << 1) /* rx timeout */ 126 #define IE_RXTOFE (1 << 6) /* rx timeout fifo empty */ 128 #define IE_OR (1 << 4) /* rx fifo overrun */ 129 #define IE_RX (1 << 3) /* rx interrupt status */ 131 #define IE_RXTO (1 << 1) /* rx timeout */ 140 #define SLFR_12RXV (1 << 10) /* slot 12 rx valid */ 142 #define SLFR_2RXV (1 << 8) /* slot 2 rx valid */ 144 #define SLFR_1RXV (1 << 6) /* slot 1 rx valid */ 146 #define SLFR_12RXB (1 << 4) /* slot 12 rx busy */ 148 #define SLFR_2RXB (1 << 2) /* slot 2 rx busy */ 150 #define SLFR_1RXB (1 << 0) /* slot 1 rx busy */
|
/linux-4.4.14/drivers/usb/musb/ |
H A D | cppi_dma.c | 83 /* zero out entire rx state RAM entry for the channel */ cppi_reset_rx() 84 static void cppi_reset_rx(struct cppi_rx_stateram __iomem *rx) cppi_reset_rx() argument 86 musb_writel(&rx->rx_skipbytes, 0, 0); cppi_reset_rx() 87 musb_writel(&rx->rx_head, 0, 0); cppi_reset_rx() 88 musb_writel(&rx->rx_sop, 0, 0); cppi_reset_rx() 89 musb_writel(&rx->rx_current, 0, 0); cppi_reset_rx() 90 musb_writel(&rx->rx_buf_current, 0, 0); cppi_reset_rx() 91 musb_writel(&rx->rx_len_len, 0, 0); cppi_reset_rx() 92 musb_writel(&rx->rx_cnt_cnt, 0, 0); cppi_reset_rx() 163 for (i = 0; i < ARRAY_SIZE(controller->rx); i++) { cppi_controller_start() 164 controller->rx[i].transmit = false; cppi_controller_start() 165 controller->rx[i].index = i; cppi_controller_start() 171 for (i = 0; i < ARRAY_SIZE(controller->rx); i++) cppi_controller_start() 172 cppi_pool_init(controller, controller->rx + i); cppi_controller_start() 177 /* initialise tx/rx channel head pointers to zero */ cppi_controller_start() 188 for (i = 0; i < ARRAY_SIZE(controller->rx); i++) { cppi_controller_start() 189 struct cppi_channel *rx_ch = controller->rx + i; cppi_controller_start() 190 struct cppi_rx_stateram __iomem *rx; cppi_controller_start() local 194 rx = tibase + DAVINCI_RXCPPI_STATERAM_OFFSET(i); cppi_controller_start() 195 rx_ch->state_ram = rx; cppi_controller_start() 196 cppi_reset_rx(rx); cppi_controller_start() 205 /* enable tx/rx CPPI control */ cppi_controller_start() 209 /* disable RNDIS mode, also host rx RNDIS autorequest */ cppi_controller_start() 241 for (i = 0; i < ARRAY_SIZE(controller->rx); i++) cppi_controller_stop() 242 cppi_pool_free(controller->rx + i); cppi_controller_stop() 248 /*disable tx/rx cppi */ cppi_controller_stop() 305 if (index >= ARRAY_SIZE(controller->rx)) { cppi_channel_allocate() 309 cppi_ch = controller->rx + index; cppi_channel_allocate() 353 struct cppi_rx_stateram __iomem *rx = c->state_ram; cppi_dump_rx() local 367 musb_readl(&rx->rx_skipbytes, 0), cppi_dump_rx() 368 musb_readl(&rx->rx_head, 0), cppi_dump_rx() 369 musb_readl(&rx->rx_sop, 0), cppi_dump_rx() 370 musb_readl(&rx->rx_current, 0), cppi_dump_rx() 372 musb_readl(&rx->rx_buf_current, 0), cppi_dump_rx() 373 musb_readl(&rx->rx_len_len, 0), cppi_dump_rx() 374 musb_readl(&rx->rx_cnt_cnt, 0), cppi_dump_rx() 375 musb_readl(&rx->rx_complete, 0) cppi_dump_rx() 438 static void cppi_dump_rxq(int level, const char *tag, struct cppi_channel *rx) cppi_dump_rxq() argument 442 cppi_dump_rx(level, rx, tag); cppi_dump_rxq() 443 if (rx->last_processed) cppi_dump_rxq() 444 cppi_dump_rxbd("last", rx->last_processed); cppi_dump_rxq() 445 for (bd = rx->head; bd; bd = bd->next) cppi_dump_rxq() 453 static inline int cppi_autoreq_update(struct cppi_channel *rx, cppi_autoreq_update() argument 464 val = tmp & ~((0x3) << (rx->index * 2)); cppi_autoreq_update() 472 val |= ((0x3) << (rx->index * 2)); cppi_autoreq_update() 476 val |= ((0x1) << (rx->index * 2)); cppi_autoreq_update() 495 if (n_bds && rx->channel.actual_len) { cppi_autoreq_update() 496 void __iomem *regs = rx->hw_ep->regs; cppi_autoreq_update() 730 * try rx rndis mode 748 * @rx: dma channel 761 cppi_next_rx_segment(struct musb *musb, struct cppi_channel *rx, int onepacket) cppi_next_rx_segment() argument 763 unsigned maxpacket = rx->maxpacket; cppi_next_rx_segment() 764 dma_addr_t addr = rx->buf_dma + rx->offset; cppi_next_rx_segment() 765 size_t length = rx->buf_len - rx->offset; cppi_next_rx_segment() 771 struct cppi_rx_stateram __iomem *rx_ram = rx->state_ram; cppi_next_rx_segment() 810 n_bds = cppi_autoreq_update(rx, tibase, onepacket, n_bds); cppi_next_rx_segment() 812 cppi_rndis_update(rx, 1, musb->ctrl_base, is_rndis); cppi_next_rx_segment() 818 rx->index, maxpacket, cppi_next_rx_segment() 824 DAVINCI_RXCPPI_BUFCNT0_REG + (rx->index * 4)) cppi_next_rx_segment() 827 rx->channel.actual_len, rx->buf_len); cppi_next_rx_segment() 832 bd = cppi_bd_alloc(rx); cppi_next_rx_segment() 833 rx->head = bd; cppi_next_rx_segment() 840 bd = cppi_bd_alloc(rx); cppi_next_rx_segment() 856 rx->offset += bd_len; cppi_next_rx_segment() 867 WARNING("rx dma%d -- no BDs? need %d\n", rx->index, n_bds); cppi_next_rx_segment() 870 WARNING("rx dma%d -- only %d of %d BDs\n", rx->index, i, n_bds); cppi_next_rx_segment() 875 bd = rx->head; cppi_next_rx_segment() 876 rx->tail = tail; cppi_next_rx_segment() 886 for (d = rx->head; d; d = d->next) cppi_next_rx_segment() 890 tail = rx->last_processed; cppi_next_rx_segment() 896 core_rxirq_enable(tibase, rx->index + 1); cppi_next_rx_segment() 911 DAVINCI_RXCPPI_BUFCNT0_REG + (rx->index * 4)) cppi_next_rx_segment() 916 DAVINCI_RXCPPI_BUFCNT0_REG + (rx->index * 4), cppi_next_rx_segment() 920 DAVINCI_RXCPPI_BUFCNT0_REG + (rx->index * 4), cppi_next_rx_segment() 924 DAVINCI_RXCPPI_BUFCNT0_REG + (rx->index * 4)) cppi_next_rx_segment() 928 rx->index, i, n_bds); cppi_next_rx_segment() 930 DAVINCI_RXCPPI_BUFCNT0_REG + (rx->index * 4), cppi_next_rx_segment() 934 cppi_dump_rx(4, rx, "/S"); cppi_next_rx_segment() 1004 struct cppi_channel *rx = &cppi->rx[ch]; cppi_rx_scan() local 1005 struct cppi_rx_stateram __iomem *state = rx->state_ram; cppi_rx_scan() 1007 struct cppi_descriptor *last = rx->last_processed; cppi_rx_scan() 1012 void __iomem *regs = rx->hw_ep->regs; cppi_rx_scan() 1015 cppi_dump_rx(6, rx, "/K"); cppi_rx_scan() 1017 bd = last ? last->next : rx->head; cppi_rx_scan() 1036 rx->channel.actual_len); cppi_rx_scan() 1054 dev_dbg(musb->controller, "rx short %d/%d (%d)\n", cppi_rx_scan() 1056 rx->channel.actual_len); cppi_rx_scan() 1075 rx->channel.actual_len += len; cppi_rx_scan() 1077 cppi_bd_free(rx, last); cppi_rx_scan() 1084 rx->last_processed = last; cppi_rx_scan() 1090 if (safe2ack == 0 || safe2ack == rx->last_processed->dma) cppi_rx_scan() 1093 cppi_bd_free(rx, last); cppi_rx_scan() 1094 rx->last_processed = NULL; cppi_rx_scan() 1099 WARN_ON(rx->head); cppi_rx_scan() 1101 musb_ep_select(cppi->mregs, rx->index + 1); cppi_rx_scan() 1105 rx->index, cppi_rx_scan() 1106 rx->head, rx->tail, cppi_rx_scan() 1107 rx->last_processed cppi_rx_scan() 1109 rx->last_processed->dma cppi_rx_scan() 1113 cppi_dump_rxq(4, "/what?", rx); cppi_rx_scan() 1119 rx->head = bd; cppi_rx_scan() 1124 csr = musb_readw(rx->hw_ep->regs, MUSB_RXCSR); cppi_rx_scan() 1131 csr = musb_readw(rx->hw_ep->regs, MUSB_RXCSR); cppi_rx_scan() 1134 rx->head = NULL; cppi_rx_scan() 1135 rx->tail = NULL; cppi_rx_scan() 1138 cppi_dump_rx(6, rx, completed ? "/completed" : "/cleaned"); cppi_rx_scan() 1148 u32 rx, tx; cppi_interrupt() local 1159 rx = musb_readl(tibase, DAVINCI_RXCPPI_MASKED_REG); cppi_interrupt() 1161 if (!tx && !rx) { cppi_interrupt() 1167 dev_dbg(musb->controller, "CPPI IRQ Tx%x Rx%x\n", tx, rx); cppi_interrupt() 1223 * REVISIT use the same ack strategy as rx cppi_interrupt() 1260 for (index = 0; rx; rx = rx >> 1, index++) { cppi_interrupt() 1262 if (rx & 1) { cppi_interrupt() 1265 rx_ch = cppi->rx + index; cppi_interrupt()
|
H A D | cppi_dma.h | 51 #define CPPI_ZERO_SET ((u32)(1 << 23)) /* rx saw zlp; tx issues one */ 52 #define CPPI_RXABT_MASK ((u32)(1 << 19)) /* need more rx buffers */ 124 struct cppi_channel rx[4]; member in struct:cppi
|
/linux-4.4.14/drivers/staging/most/hdm-i2c/ |
H A D | hdm_i2c.c | 54 struct rx { struct in struct:hdm_i2c 59 } rx; member in struct:hdm_i2c 96 schedule_delayed_work(&dev->rx.dwork, configure_channel() 126 mutex_lock(&dev->rx.list_mutex); enqueue() 127 list_add_tail(&mbo->list, &dev->rx.list); enqueue() 128 mutex_unlock(&dev->rx.list_mutex); enqueue() 129 wake_up_interruptible(&dev->rx.waitq); enqueue() 169 mutex_lock(&dev->rx.list_mutex); poison_channel() 170 while (!list_empty(&dev->rx.list)) { poison_channel() 171 mbo = list_first_mbo(&dev->rx.list); poison_channel() 173 mutex_unlock(&dev->rx.list_mutex); poison_channel() 179 mutex_lock(&dev->rx.list_mutex); poison_channel() 181 mutex_unlock(&dev->rx.list_mutex); poison_channel() 182 wake_up_interruptible(&dev->rx.waitq); poison_channel() 225 if (wait_event_interruptible(dev->rx.waitq, do_rx_work() 227 !list_empty(&dev->rx.list))) { do_rx_work() 235 mutex_lock(&dev->rx.list_mutex); do_rx_work() 238 if (!list_empty(&dev->rx.list)) do_rx_work() 241 mutex_unlock(&dev->rx.list_mutex); do_rx_work() 244 mbo = list_first_mbo(&dev->rx.list); do_rx_work() 246 mutex_unlock(&dev->rx.list_mutex); do_rx_work() 262 struct hdm_i2c *dev = container_of(work, struct hdm_i2c, rx.dwork.work); pending_rx_work() 268 schedule_delayed_work(&dev->rx.dwork, pending_rx_work() 300 schedule_delayed_work(&dev->rx.dwork, 0); most_irq_handler() 335 dev->capabilities[CH_RX].name_suffix = "rx"; i2c_probe() 348 INIT_LIST_HEAD(&dev->rx.list); i2c_probe() 349 mutex_init(&dev->rx.list_mutex); i2c_probe() 350 init_waitqueue_head(&dev->rx.waitq); i2c_probe() 352 INIT_DELAYED_WORK(&dev->rx.dwork, pending_rx_work); i2c_probe() 403 cancel_delayed_work_sync(&dev->rx.dwork); i2c_remove()
|
/linux-4.4.14/drivers/net/ethernet/freescale/ |
H A D | ucc_geth_ethtool.c | 44 "rx-64-frames", 45 "rx-65-127-frames", 46 "rx-128-255-frames", 51 "rx-frames", 52 "rx-bytes-ok", 53 "rx-bytes-all", 54 "rx-multicast-frames", 55 "rx-broadcast-frames", 58 "rx-dropped-frames", 77 "rx-crc-errors", 78 "rx-alignment-errors", 79 "rx-in-range-length-errors", 80 "rx-out-of-range-length-errors", 81 "rx-too-long-frames", 82 "rx-runt", 83 "rx-very-long-event", 84 "rx-symbol-errors", 85 "rx-busy-drop-frames", 88 "rx-mismatch-drop-frames", 89 "rx-small-than-64", 90 "rx-256-511-frames", 91 "rx-512-1023-frames", 92 "rx-1024-1518-frames", 93 "rx-jumbo-frames", 94 "rx-mac-error-loss", 95 "rx-pause-frames", 97 "rx-vlan-removed", 98 "rx-vlan-replaced", 99 "rx-vlan-inserted", 100 "rx-ip-checksum-errors",
|
H A D | gianfar_ethtool.c | 65 "rx-allocation-errors", 66 "rx-large-frame-errors", 67 "rx-short-frame-errors", 68 "rx-non-octet-errors", 69 "rx-crc-errors", 70 "rx-overrun-errors", 71 "rx-busy-errors", 72 "rx-babbling-errors", 73 "rx-truncated-frames", 79 "tx-rx-64-frames", 80 "tx-rx-65-127-frames", 81 "tx-rx-128-255-frames", 82 "tx-rx-256-511-frames", 83 "tx-rx-512-1023-frames", 84 "tx-rx-1024-1518-frames", 85 "tx-rx-1519-1522-good-vlan", 86 "rx-bytes", 87 "rx-packets", 88 "rx-fcs-errors", 91 "rx-control-frame-packets", 92 "rx-pause-frame-packets", 93 "rx-unknown-op-code", 94 "rx-alignment-error", 95 "rx-frame-length-error", 96 "rx-code-error", 97 "rx-carrier-sense-error", 98 "rx-undersize-packets", 99 "rx-oversize-packets", 100 "rx-fragmented-frames", 101 "rx-jabber-frames", 102 "rx-dropped-frames", 334 * normal {rx,tx}_* coalescing parameters are used. gfar_gcoalesce() 338 * is above pkt_rate_high, the {rx,tx}_*_high parameters are gfar_gcoalesce() 400 /* Set up rx coalescing */ gfar_scoalesce() 445 * rx, rx_mini, and rx_jumbo rings are the same size, as mini and
|
/linux-4.4.14/drivers/staging/iio/meter/ |
H A D | ade7854-i2c.c | 116 ret = i2c_master_recv(st->i2c, st->rx, 1); ade7854_i2c_read_reg_8() 120 *val = st->rx[0]; ade7854_i2c_read_reg_8() 142 ret = i2c_master_recv(st->i2c, st->rx, 2); ade7854_i2c_read_reg_16() 146 *val = (st->rx[0] << 8) | st->rx[1]; ade7854_i2c_read_reg_16() 168 ret = i2c_master_recv(st->i2c, st->rx, 3); ade7854_i2c_read_reg_24() 172 *val = (st->rx[0] << 16) | (st->rx[1] << 8) | st->rx[2]; ade7854_i2c_read_reg_24() 194 ret = i2c_master_recv(st->i2c, st->rx, 3); ade7854_i2c_read_reg_32() 198 *val = (st->rx[0] << 24) | (st->rx[1] << 16) | ade7854_i2c_read_reg_32() 199 (st->rx[2] << 8) | st->rx[3]; ade7854_i2c_read_reg_32()
|
H A D | ade7759.h | 42 * @buf_lock: mutex to protect tx and rx 44 * @rx: receive buffer 50 u8 rx[ADE7759_MAX_RX]; member in struct:ade7759_state
|
H A D | ade7753.h | 62 * @rx: receive buffer 63 * @buf_lock: mutex to protect tx and rx 69 u8 rx[ADE7753_MAX_RX]; member in struct:ade7753_state
|
H A D | ade7754.h | 79 * @buf_lock: mutex to protect tx and rx 81 * @rx: receive buffer 87 u8 rx[ADE7754_MAX_RX]; member in struct:ade7754_state
|
H A D | ade7854-spi.c | 137 .rx_buf = st->rx, ade7854_spi_read_reg_8() 155 *val = st->rx[0]; ade7854_spi_read_reg_8() 175 .rx_buf = st->rx, ade7854_spi_read_reg_16() 192 *val = be16_to_cpup((const __be16 *)st->rx); ade7854_spi_read_reg_16() 212 .rx_buf = st->rx, ade7854_spi_read_reg_24() 230 *val = (st->rx[0] << 16) | (st->rx[1] << 8) | st->rx[2]; ade7854_spi_read_reg_24() 250 .rx_buf = st->rx, ade7854_spi_read_reg_32() 268 *val = be32_to_cpup((const __be32 *)st->rx); ade7854_spi_read_reg_32()
|
H A D | ade7758_core.c | 114 .rx_buf = st->rx, ade7758_spi_read_reg_8() 130 *val = st->rx[0]; ade7758_spi_read_reg_8() 153 .rx_buf = st->rx, ade7758_spi_read_reg_16() 172 *val = (st->rx[0] << 8) | st->rx[1]; ade7758_spi_read_reg_16() 195 .rx_buf = st->rx, ade7758_spi_read_reg_24() 213 *val = (st->rx[0] << 16) | (st->rx[1] << 8) | st->rx[2]; ade7758_spi_read_reg_24() 832 st->rx = kcalloc(ADE7758_MAX_RX, sizeof(*st->rx), GFP_KERNEL); ade7758_probe() 833 if (!st->rx) ade7758_probe() 879 kfree(st->rx); ade7758_probe() 893 kfree(st->rx); ade7758_remove()
|
H A D | ade7758.h | 113 * @rx: receive buffer 114 * @buf_lock: mutex to protect tx and rx 120 u8 *rx; member in struct:ade7758_state
|
H A D | ade7854.h | 149 * @buf_lock: mutex to protect tx and rx 151 * @rx: receive buffer 167 u8 rx[ADE7854_MAX_RX]; member in struct:ade7854_state
|
/linux-4.4.14/drivers/mfd/ |
H A D | ipaq-micro.c | 149 struct ipaq_micro_rxdev *rx = µ->rx; micro_process_char() local 151 switch (rx->state) { micro_process_char() 154 rx->state = STATE_ID; /* Next byte is the id and len */ micro_process_char() 157 rx->id = (ch & 0xf0) >> 4; micro_process_char() 158 rx->len = (ch & 0x0f); micro_process_char() 159 rx->index = 0; micro_process_char() 160 rx->chksum = ch; micro_process_char() 161 rx->state = (rx->len > 0) ? STATE_DATA : STATE_CHKSUM; micro_process_char() 164 rx->chksum += ch; micro_process_char() 165 rx->buf[rx->index] = ch; micro_process_char() 166 if (++rx->index == rx->len) micro_process_char() 167 rx->state = STATE_CHKSUM; micro_process_char() 170 if (ch == rx->chksum) micro_process_char() 171 micro_rx_msg(micro, rx->id, rx->len, rx->buf); micro_process_char() 172 rx->state = STATE_SOF; micro_process_char() 184 dev_err(micro->dev, "rx: parity error\n"); micro_rx_chars() 186 dev_err(micro->dev, "rx: framing error\n"); micro_rx_chars() 188 dev_err(micro->dev, "rx: overrun error\n"); micro_rx_chars() 308 struct ipaq_micro_rxdev *rx = µ->rx; micro_reset_comm() local 315 rx->state = STATE_SOF; /* Reset the state machine */ micro_reset_comm()
|
/linux-4.4.14/drivers/net/usb/ |
H A D | asix_common.c | 54 struct asix_rx_fixup_info *rx) asix_rx_fixup_internal() 68 if (rx->remaining && (rx->remaining + sizeof(u32) <= skb->len)) { asix_rx_fixup_internal() 69 offset = ((rx->remaining + 1) & 0xfffe); asix_rx_fixup_internal() 70 rx->header = get_unaligned_le32(skb->data + offset); asix_rx_fixup_internal() 73 size = (u16)(rx->header & 0x7ff); asix_rx_fixup_internal() 74 if (size != ((~rx->header >> 16) & 0x7ff)) { asix_rx_fixup_internal() 76 rx->remaining); asix_rx_fixup_internal() 77 if (rx->ax_skb) { asix_rx_fixup_internal() 78 kfree_skb(rx->ax_skb); asix_rx_fixup_internal() 79 rx->ax_skb = NULL; asix_rx_fixup_internal() 85 rx->remaining = 0; asix_rx_fixup_internal() 93 if (!rx->remaining) { asix_rx_fixup_internal() 95 rx->header = get_unaligned_le16( asix_rx_fixup_internal() 97 rx->split_head = true; asix_rx_fixup_internal() 102 if (rx->split_head == true) { asix_rx_fixup_internal() 103 rx->header |= (get_unaligned_le16( asix_rx_fixup_internal() 105 rx->split_head = false; asix_rx_fixup_internal() 108 rx->header = get_unaligned_le32(skb->data + asix_rx_fixup_internal() 114 size = (u16)(rx->header & 0x7ff); asix_rx_fixup_internal() 115 if (size != ((~rx->header >> 16) & 0x7ff)) { asix_rx_fixup_internal() 117 rx->header, offset); asix_rx_fixup_internal() 131 rx->ax_skb = netdev_alloc_skb_ip_align(dev->net, size); asix_rx_fixup_internal() 133 rx->remaining = size; asix_rx_fixup_internal() 136 if (rx->remaining > skb->len - offset) { asix_rx_fixup_internal() 138 rx->remaining -= copy_length; asix_rx_fixup_internal() 140 copy_length = rx->remaining; asix_rx_fixup_internal() 141 rx->remaining = 0; asix_rx_fixup_internal() 144 if (rx->ax_skb) { asix_rx_fixup_internal() 145 data = skb_put(rx->ax_skb, copy_length); asix_rx_fixup_internal() 147 if (!rx->remaining) asix_rx_fixup_internal() 148 usbnet_skb_return(dev, rx->ax_skb); asix_rx_fixup_internal() 166 struct asix_rx_fixup_info *rx = &dp->rx_fixup_info; asix_rx_fixup_common() local 168 return asix_rx_fixup_internal(dev, skb, rx); asix_rx_fixup_common() 53 asix_rx_fixup_internal(struct usbnet *dev, struct sk_buff *skb, struct asix_rx_fixup_info *rx) asix_rx_fixup_internal() argument
|
/linux-4.4.14/drivers/net/ethernet/stmicro/stmmac/ |
H A D | norm_desc.c | 79 if (unlikely(p->des01.rx.last_descriptor == 0)) { ndesc_get_rx_status() 86 if (unlikely(p->des01.rx.error_summary)) { ndesc_get_rx_status() 87 if (unlikely(p->des01.rx.descriptor_error)) ndesc_get_rx_status() 89 if (unlikely(p->des01.rx.sa_filter_fail)) ndesc_get_rx_status() 91 if (unlikely(p->des01.rx.overflow_error)) ndesc_get_rx_status() 93 if (unlikely(p->des01.rx.ipc_csum_error)) ndesc_get_rx_status() 95 if (unlikely(p->des01.rx.collision)) { ndesc_get_rx_status() 99 if (unlikely(p->des01.rx.crc_error)) { ndesc_get_rx_status() 105 if (unlikely(p->des01.rx.dribbling)) ndesc_get_rx_status() 108 if (unlikely(p->des01.rx.length_error)) { ndesc_get_rx_status() 112 if (unlikely(p->des01.rx.mii_error)) { ndesc_get_rx_status() 117 if (p->des01.rx.vlan_tag) ndesc_get_rx_status() 127 p->des01.rx.own = 1; ndesc_init_rx_desc() 128 p->des01.rx.buffer1_size = BUF_SIZE_2KiB - 1; ndesc_init_rx_desc() 136 p->des01.rx.disable_ic = 1; ndesc_init_rx_desc() 155 return p->des01.rx.own; ndesc_get_rx_owner() 165 p->des01.rx.own = 1; ndesc_set_rx_owner() 216 return p->des01.rx.frame_length - 2; ndesc_get_rx_frame_len() 218 return p->des01.rx.frame_length; ndesc_get_rx_frame_len()
|
H A D | descs_com.h | 66 p->des01.rx.buffer2_size = BUF_SIZE_2KiB - 1; ndesc_rx_set_on_ring() 68 p->des01.rx.end_ring = 1; ndesc_rx_set_on_ring() 117 p->des01.rx.second_address_chained = 1; ndesc_rx_set_on_chain()
|
/linux-4.4.14/arch/s390/kernel/ |
H A D | uprobes.c | 259 union split_register *rx; handle_insn_ril() local 266 rx = (union split_register *) ®s->gprs[insn->reg]; handle_insn_ril() 274 rx->u64 = (unsigned long)uptr; handle_insn_ril() 281 rc = emu_load_ril((u16 __user *)uptr, &rx->u32[1]); handle_insn_ril() 284 rc = emu_load_ril((s16 __user *)uptr, &rx->u64); handle_insn_ril() 287 rc = emu_load_ril((s16 __user *)uptr, &rx->u32[1]); handle_insn_ril() 290 rc = emu_load_ril((u16 __user *)uptr, &rx->u64); handle_insn_ril() 293 rc = emu_load_ril((u64 __user *)uptr, &rx->u64); handle_insn_ril() 296 rc = emu_load_ril((s32 __user *)uptr, &rx->u64); handle_insn_ril() 299 rc = emu_load_ril((u32 __user *)uptr, &rx->u32[1]); handle_insn_ril() 302 rc = emu_load_ril((u32 __user *)uptr, &rx->u64); handle_insn_ril() 305 rc = emu_store_ril(regs, (u16 __user *)uptr, &rx->u16[3]); handle_insn_ril() 308 rc = emu_store_ril(regs, (u64 __user *)uptr, &rx->u64); handle_insn_ril() 311 rc = emu_store_ril(regs, (u32 __user *)uptr, &rx->u32[1]); handle_insn_ril() 322 rc = emu_cmp_ril(regs, (s16 __user *)uptr, &rx->s64); handle_insn_ril() 325 rc = emu_cmp_ril(regs, (s16 __user *)uptr, &rx->s32[1]); handle_insn_ril() 328 rc = emu_cmp_ril(regs, (u16 __user *)uptr, &rx->u64); handle_insn_ril() 331 rc = emu_cmp_ril(regs, (u16 __user *)uptr, &rx->u32[1]); handle_insn_ril() 334 rc = emu_cmp_ril(regs, (s64 __user *)uptr, &rx->s64); handle_insn_ril() 337 rc = emu_cmp_ril(regs, (u64 __user *)uptr, &rx->u64); handle_insn_ril() 340 rc = emu_cmp_ril(regs, (s32 __user *)uptr, &rx->s64); handle_insn_ril() 343 rc = emu_cmp_ril(regs, (s32 __user *)uptr, &rx->s32[1]); handle_insn_ril() 346 rc = emu_cmp_ril(regs, (u32 __user *)uptr, &rx->u64); handle_insn_ril() 349 rc = emu_cmp_ril(regs, (u32 __user *)uptr, &rx->u32[1]); handle_insn_ril()
|
H A D | diag.c | 138 static inline int __diag14(unsigned long rx, unsigned long ry1, __diag14() argument 152 : "d" (rx), "d" (_ry1) __diag14() 158 int diag14(unsigned long rx, unsigned long ry1, unsigned long subcode) diag14() argument 161 return __diag14(rx, ry1, subcode); diag14()
|
/linux-4.4.14/drivers/net/wireless/ath/ath9k/ |
H A D | recv.c | 34 * buffer (or rx fifo). This can incorrectly acknowledge packets 55 * setup rx descriptors. The rx_bufsize here tells the hardware ath_rx_buf_link() 63 if (sc->rx.rxlink) ath_rx_buf_link() 64 *sc->rx.rxlink = bf->bf_daddr; ath_rx_buf_link() 68 sc->rx.rxlink = &ds->ds_link; ath_rx_buf_link() 74 if (sc->rx.buf_hold) ath_rx_buf_relink() 75 ath_rx_buf_link(sc, sc->rx.buf_hold, flush); ath_rx_buf_relink() 77 sc->rx.buf_hold = bf; ath_rx_buf_relink() 84 sc->rx.defant = antenna; ath_setdefantenna() 85 sc->rx.rxotherant = 0; ath_setdefantenna() 95 /* configure rx filter */ ath_opmode_init() 118 rx_edma = &sc->rx.rx_edma[qtype]; ath_rx_edma_buf_link() 122 bf = list_first_entry(&sc->rx.rxbuf, struct ath_rxbuf, list); ath_rx_edma_buf_link() 144 if (list_empty(&sc->rx.rxbuf)) { ath_rx_addbuffer_edma() 145 ath_dbg(common, QUEUE, "No free rx buf available\n"); ath_rx_addbuffer_edma() 149 list_for_each_entry_safe(bf, tbf, &sc->rx.rxbuf, list) ath_rx_addbuffer_edma() 162 rx_edma = &sc->rx.rx_edma[qtype]; ath_rx_remove_buffer() 167 list_add_tail(&bf->list, &sc->rx.rxbuf); ath_rx_remove_buffer() 180 list_for_each_entry(bf, &sc->rx.rxbuf, list) { ath_rx_edma_cleanup() 210 ath_rx_edma_init_queue(&sc->rx.rx_edma[ATH9K_RX_QUEUE_LP], ath_rx_edma_init() 212 ath_rx_edma_init_queue(&sc->rx.rx_edma[ATH9K_RX_QUEUE_HP], ath_rx_edma_init() 220 INIT_LIST_HEAD(&sc->rx.rxbuf); ath_rx_edma_init() 246 list_add_tail(&bf->list, &sc->rx.rxbuf); ath_rx_edma_init() 289 /* Initialize rx descriptors */ ath_rx_init() 291 error = ath_descdma_setup(sc, &sc->rx.rxdma, &sc->rx.rxbuf, ath_rx_init() 292 "rx", nbufs, 1, 0); ath_rx_init() 295 "failed to allocate rx descriptors: %d\n", ath_rx_init() 300 list_for_each_entry(bf, &sc->rx.rxbuf, list) { ath_rx_init() 323 sc->rx.rxlink = NULL; ath_rx_init() 343 list_for_each_entry(bf, &sc->rx.rxbuf, list) { ath_rx_cleanup() 447 if (list_empty(&sc->rx.rxbuf)) ath_startrecv() 450 sc->rx.buf_hold = NULL; ath_startrecv() 451 sc->rx.rxlink = NULL; ath_startrecv() 452 list_for_each_entry_safe(bf, tbf, &sc->rx.rxbuf, list) { ath_startrecv() 457 if (list_empty(&sc->rx.rxbuf)) ath_startrecv() 460 bf = list_first_entry(&sc->rx.rxbuf, struct ath_rxbuf, list); ath_startrecv() 490 sc->rx.rxlink = NULL; ath_stoprecv() 628 struct ath_rx_edma *rx_edma = &sc->rx.rx_edma[qtype]; ath_edma_get_buffers() 656 list_add_tail(&bf->list, &sc->rx.rxbuf); ath_edma_get_buffers() 665 list_add_tail(&bf->list, &sc->rx.rxbuf); ath_edma_get_buffers() 700 if (list_empty(&sc->rx.rxbuf)) { ath_get_next_rx_buf() 701 sc->rx.rxlink = NULL; ath_get_next_rx_buf() 705 bf = list_first_entry(&sc->rx.rxbuf, struct ath_rxbuf, list); ath_get_next_rx_buf() 706 if (bf == sc->rx.buf_hold) ath_get_next_rx_buf() 720 * a self-linked list to avoid rx overruns. ath_get_next_rx_buf() 729 if (list_is_last(&bf->list, &sc->rx.rxbuf)) { ath_get_next_rx_buf() 730 sc->rx.rxlink = NULL; ath_get_next_rx_buf() 814 bool discard_current = sc->rx.discard_next; ath9k_rx_skb_preprocess() 823 sc->rx.discard_next = false; ath9k_rx_skb_preprocess() 924 sc->rx.num_pkts++; ath9k_rx_skb_preprocess() 930 sc->rx.discard_next = rx_stats->rs_more; ath9k_rx_skb_preprocess() 955 * Change the default rx antenna if rx diversity ath9k_antenna_check() 958 if (sc->rx.defant != rs->rs_antenna) { ath9k_antenna_check() 959 if (++sc->rx.rxotherant >= 3) ath9k_antenna_check() 962 sc->rx.rxotherant = 0; ath9k_antenna_check() 979 rxs->ampdu_reference = sc->rx.ampdu_ref; ath9k_apply_ampdu_details() 983 sc->rx.ampdu_ref++; ath9k_apply_ampdu_details() 1039 if (sc->rx.frag) ath_rx_tasklet() 1040 hdr_skb = sc->rx.frag; ath_rx_tasklet() 1058 * skb and put it at the tail of the sc->rx.rxbuf list for ath_rx_tasklet() 1095 if (sc->rx.frag) { ath_rx_tasklet() 1097 dev_kfree_skb_any(sc->rx.frag); ath_rx_tasklet() 1102 sc->rx.frag = skb; ath_rx_tasklet() 1106 if (sc->rx.frag) { ath_rx_tasklet() 1115 sc->rx.frag = NULL; ath_rx_tasklet() 1145 if (sc->rx.frag) { ath_rx_tasklet() 1146 dev_kfree_skb_any(sc->rx.frag); ath_rx_tasklet() 1147 sc->rx.frag = NULL; ath_rx_tasklet() 1150 list_add_tail(&bf->list, &sc->rx.rxbuf); ath_rx_tasklet()
|
/linux-4.4.14/drivers/staging/media/lirc/ |
H A D | lirc_zilog.c | 112 struct IR_rx *rx; member in struct:IR 181 * ir->rx set to NULL and deallocated - happens before ir->rx->ir put() release_ir_device() 182 * ir->rx->task kthread stopped - happens before ir->rx->ir put() release_ir_device() 214 struct IR_rx *rx; get_ir_rx() local 217 rx = ir->rx; get_ir_rx() 218 if (rx != NULL) get_ir_rx() 219 kref_get(&rx->ref); get_ir_rx() 221 return rx; get_ir_rx() 224 static void destroy_rx_kthread(struct IR_rx *rx, bool ir_devices_lock_held) destroy_rx_kthread() argument 227 if (!IS_ERR_OR_NULL(rx->task)) { destroy_rx_kthread() 228 kthread_stop(rx->task); destroy_rx_kthread() 229 rx->task = NULL; destroy_rx_kthread() 230 /* Put the ir ptr that ir_probe() gave to the rx poll thread */ destroy_rx_kthread() 231 put_ir_device(rx->ir, ir_devices_lock_held); destroy_rx_kthread() 237 struct IR_rx *rx = container_of(ref, struct IR_rx, ref); release_ir_rx() local 238 struct IR *ir = rx->ir; release_ir_rx() 247 /* Don't put_ir_device(rx->ir) here; lock can't be freed yet */ release_ir_rx() 248 ir->rx = NULL; release_ir_rx() 249 /* Don't do the kfree(rx) here; we still need to kill the poll thread */ release_ir_rx() 252 static int put_ir_rx(struct IR_rx *rx, bool ir_devices_lock_held) put_ir_rx() argument 255 struct IR *ir = rx->ir; put_ir_rx() 258 released = kref_put(&rx->ref, release_ir_rx); put_ir_rx() 260 /* Destroy the rx kthread while not holding the spinlock */ put_ir_rx() 262 destroy_rx_kthread(rx, ir_devices_lock_held); put_ir_rx() 263 kfree(rx); put_ir_rx() 267 /* Do a reference put() for the rx->ir reference, if we released rx */ put_ir_rx() 321 struct IR_rx *rx; add_to_buf() local 329 rx = get_ir_rx(ir); add_to_buf() 330 if (rx == NULL) add_to_buf() 333 /* Ensure our rx->c i2c_client remains valid for the duration */ add_to_buf() 334 mutex_lock(&rx->client_lock); add_to_buf() 335 if (rx->c == NULL) { add_to_buf() 336 mutex_unlock(&rx->client_lock); add_to_buf() 337 put_ir_rx(rx, false); add_to_buf() 369 ret = i2c_master_send(rx->c, sendbuf, 1); add_to_buf() 405 ret = i2c_master_recv(rx->c, keybuf, sizeof(keybuf)); add_to_buf() 412 rx->b[0] = keybuf[3]; add_to_buf() 413 rx->b[1] = keybuf[4]; add_to_buf() 414 rx->b[2] = keybuf[5]; add_to_buf() 417 rx->b[0], rx->b[1]); add_to_buf() 421 if (rx->hdpvr_data_fmt) { add_to_buf() 429 } else if ((rx->b[0] & 0x80) == 0) { add_to_buf() 435 code = (((__u16)rx->b[0] & 0x7f) << 6) | (rx->b[1] >> 2); add_to_buf() 446 mutex_unlock(&rx->client_lock); add_to_buf() 449 put_ir_rx(rx, false); add_to_buf() 895 struct IR_rx *rx; read() local 907 rx = get_ir_rx(ir); read() 908 if (rx == NULL) read() 969 put_ir_rx(rx, false); read() 1211 struct IR_rx *rx; poll() local 1217 rx = get_ir_rx(ir); poll() 1218 if (rx == NULL) { poll() 1417 struct IR_rx *rx = i2c_get_clientdata(client); ir_remove() local 1419 if (rx != NULL) { ir_remove() 1420 mutex_lock(&rx->client_lock); ir_remove() 1421 rx->c = NULL; ir_remove() 1422 mutex_unlock(&rx->client_lock); ir_remove() 1423 put_ir_rx(rx, false); ir_remove() 1451 struct IR_rx *rx; ir_probe() local 1514 rx = get_ir_rx(ir); ir_probe() 1550 if (rx == NULL && !tx_only) { ir_probe() 1561 rx = kzalloc(sizeof(struct IR_rx), GFP_KERNEL); ir_probe() 1562 if (rx == NULL) { ir_probe() 1566 kref_init(&rx->ref); ir_probe() 1567 ir->rx = rx; ir_probe() 1570 mutex_init(&rx->client_lock); ir_probe() 1571 rx->c = client; ir_probe() 1572 rx->hdpvr_data_fmt = ir_probe() 1576 rx->ir = get_ir_device(ir, true); ir_probe() 1578 /* An rx ref goes to the i2c_client */ ir_probe() 1586 /* An ir ref goes to the new rx polling kthread */ ir_probe() 1587 rx->task = kthread_run(lirc_thread, get_ir_device(ir, true), ir_probe() 1588 "zilog-rx-i2c-%d", adap->nr); ir_probe() 1589 if (IS_ERR(rx->task)) { ir_probe() 1590 ret = PTR_ERR(rx->task); ir_probe() 1596 /* Failure exit, so put back rx ref from i2c_client */ ir_probe() 1598 put_ir_rx(rx, true); ir_probe() 1626 if (rx != NULL) ir_probe() 1627 put_ir_rx(rx, true); ir_probe() 1638 if (rx != NULL) ir_probe() 1639 put_ir_rx(rx, true); ir_probe()
|
H A D | lirc_imon.c | 101 } rx; member in struct:imon_context 498 context->rx.count = 0; ir_open() 499 context->rx.initial_space = 1; ir_open() 500 context->rx.prev_bit = 0; ir_open() 555 int value = context->rx.count; submit_data() 562 if (context->rx.prev_bit) submit_data() 612 if (buf[7] == 1 && context->rx.initial_space) { imon_incoming_packet() 614 context->rx.prev_bit = 0; imon_incoming_packet() 615 context->rx.count = 4; imon_incoming_packet() 617 context->rx.count = 0; imon_incoming_packet() 625 if (curr_bit != context->rx.prev_bit) { imon_incoming_packet() 626 if (context->rx.count) { imon_incoming_packet() 628 context->rx.count = 0; imon_incoming_packet() 630 context->rx.prev_bit = curr_bit; imon_incoming_packet() 632 ++context->rx.count; imon_incoming_packet() 638 if (context->rx.count) { imon_incoming_packet() 640 context->rx.count = 0; imon_incoming_packet() 642 context->rx.initial_space = context->rx.prev_bit; imon_incoming_packet()
|
/linux-4.4.14/drivers/media/platform/vivid/ |
H A D | Makefile | 3 vivid-radio-rx.o vivid-radio-tx.o vivid-radio-common.o \
|
/linux-4.4.14/drivers/net/wireless/ti/wl12xx/ |
H A D | debugfs.c | 35 WL12XX_DEBUGFS_FWSTATS_FILE(rx, out_of_mem, "%u"); 36 WL12XX_DEBUGFS_FWSTATS_FILE(rx, hdr_overflow, "%u"); 37 WL12XX_DEBUGFS_FWSTATS_FILE(rx, hw_stuck, "%u"); 38 WL12XX_DEBUGFS_FWSTATS_FILE(rx, dropped, "%u"); 39 WL12XX_DEBUGFS_FWSTATS_FILE(rx, fcs_err, "%u"); 40 WL12XX_DEBUGFS_FWSTATS_FILE(rx, xfr_hint_trig, "%u"); 41 WL12XX_DEBUGFS_FWSTATS_FILE(rx, path_reset, "%u"); 42 WL12XX_DEBUGFS_FWSTATS_FILE(rx, reset_counter, "%u"); 145 DEBUGFS_FWSTATS_ADD(rx, out_of_mem); wl12xx_debugfs_add_files() 146 DEBUGFS_FWSTATS_ADD(rx, hdr_overflow); wl12xx_debugfs_add_files() 147 DEBUGFS_FWSTATS_ADD(rx, hw_stuck); wl12xx_debugfs_add_files() 148 DEBUGFS_FWSTATS_ADD(rx, dropped); wl12xx_debugfs_add_files() 149 DEBUGFS_FWSTATS_ADD(rx, fcs_err); wl12xx_debugfs_add_files() 150 DEBUGFS_FWSTATS_ADD(rx, xfr_hint_trig); wl12xx_debugfs_add_files() 151 DEBUGFS_FWSTATS_ADD(rx, path_reset); wl12xx_debugfs_add_files() 152 DEBUGFS_FWSTATS_ADD(rx, reset_counter); wl12xx_debugfs_add_files()
|
/linux-4.4.14/drivers/net/ethernet/broadcom/ |
H A D | bcm63xx_enet.h | 29 * hardware maximum rx/tx packet size including FCS, max mtu is 30 * actually 2047, but if we set max rx size register to 2047 we won't 207 /* hw view of rx & tx dma ring */ 211 /* allocated size (in bytes) for rx & tx dma ring */ 218 /* dma channel id for rx */ 221 /* number of dma desc in rx ring */ 224 /* cpu view of rx dma ring */ 227 /* current number of armed descriptor given to hardware for rx */ 230 /* next rx descriptor to fetch from hardware */ 233 /* next dirty rx descriptor to refill */ 236 /* size of allocated rx skbs */ 239 /* list of skb given to hw for rx */ 242 /* used when rx skb allocation failed, so we defer rx queue 246 /* lock rx_timeout against rx normal operation */ 259 /* cpu view of rx dma ring */
|
H A D | bcm63xx_enet.c | 235 * refill rx queue 284 /* If rx ring is still empty, set a timer to try allocating bcm_enet_refill_rx() 287 dev_warn(&priv->pdev->dev, "unable to refill rx ring\n"); bcm_enet_refill_rx() 296 * timer callback to defer refill rx queue in case we're OOM 312 * extract packet from rx queue 418 /* kick rx dma */ bcm_enet_receive_queue() 508 /* rx queue is not yet empty/clean */ bcm_enet_poll() 512 /* no more packet in rx/tx queue, remove device from poll bcm_enet_poll() 516 /* restore rx/tx interrupt */ bcm_enet_poll() 552 * rx/tx dma interrupt handler 562 /* mask rx/tx interrupts */ bcm_enet_isr_dma() 684 * Change rx mode (promiscuous/allmulti) and update multicast list 764 /* rx flow control (pause frame handling) */ bcm_enet_set_flow() 839 phydev->pause == 1 ? "rx&tx" : "off"); bcm_enet_adjust_phy_link() 861 priv->pause_rx ? "rx" : "off", bcm_enet_adjust_link() 866 * open callback, allocate dma rings & buffers and start rx operation 949 /* allocate rx dma ring */ bcm_enet_open() 983 /* init & fill rx ring with skbs */ bcm_enet_open() 1004 dev_err(kdev, "cannot allocate rx skb queue\n"); bcm_enet_open() 1009 /* write rx & tx ring addresses */ bcm_enet_open() 1022 /* clear remaining state ram for rx & tx channel */ bcm_enet_open() 1035 /* set max rx/tx length */ bcm_enet_open() 1061 * kick rx dma channel */ bcm_enet_open() 1074 /* watch "packet transferred" interrupt in rx and tx */ bcm_enet_open() 1080 /* make sure we enable napi before rx interrupt */ bcm_enet_open() 1213 /* free the rx skb ring */ bcm_enet_stop() 1513 /* rx/tx ring is actually only limited by memory */ bcm_enet_get_ringparam() 1648 * align rx buffer size to dma burst len, account FCS since compute_hw_mtu() 1774 /* get rx & tx dma channel id for this mac */ bcm_enet_probe() 1882 /* init rx timeout (used for oom) */ bcm_enet_probe() 2140 * open callback, allocate dma rings & buffers and start rx operation 2170 /* allocate rx dma ring */ bcm_enetsw_open() 2174 dev_err(kdev, "cannot allocate rx ring %u\n", size); bcm_enetsw_open() 2199 dev_err(kdev, "cannot allocate rx skb queue\n"); bcm_enetsw_open() 2209 /* init & fill rx ring with skbs */ bcm_enetsw_open() 2213 dev_err(kdev, "cannot allocate rx skb queue\n"); bcm_enetsw_open() 2261 dev_err(kdev, "cannot allocate rx skb queue\n"); bcm_enetsw_open() 2266 /* write rx & tx ring addresses */ bcm_enetsw_open() 2272 /* clear remaining state ram for rx & tx channel */ bcm_enetsw_open() 2293 * kick rx dma channel bcm_enetsw_open() 2300 /* watch "packet transferred" interrupt in rx and tx */ bcm_enetsw_open() 2306 /* make sure we enable napi before rx interrupt */ bcm_enetsw_open() 2424 /* free the rx skb ring */ bcm_enetsw_stop() 2661 /* rx/tx ring is actually only limited by memory */ bcm_enetsw_get_ringparam() 2781 /* init rx timeout (used for oom) */ bcm_enetsw_probe()
|
/linux-4.4.14/include/uapi/linux/ |
H A D | atm_eni.h | 14 int tx,rx; /* values are in percent and must be > 100 */ member in struct:eni_multipliers
|
H A D | tipc_config.h | 74 #define TIPC_CMD_NOOP 0x0000 /* tx none, rx none */ 75 #define TIPC_CMD_GET_NODES 0x0001 /* tx net_addr, rx node_info(s) */ 76 #define TIPC_CMD_GET_MEDIA_NAMES 0x0002 /* tx none, rx media_name(s) */ 77 #define TIPC_CMD_GET_BEARER_NAMES 0x0003 /* tx none, rx bearer_name(s) */ 78 #define TIPC_CMD_GET_LINKS 0x0004 /* tx net_addr, rx link_info(s) */ 79 #define TIPC_CMD_SHOW_NAME_TABLE 0x0005 /* tx name_tbl_query, rx ultra_string */ 80 #define TIPC_CMD_SHOW_PORTS 0x0006 /* tx none, rx ultra_string */ 81 #define TIPC_CMD_SHOW_LINK_STATS 0x000B /* tx link_name, rx ultra_string */ 82 #define TIPC_CMD_SHOW_STATS 0x000F /* tx unsigned, rx ultra_string */ 91 #define TIPC_CMD_GET_REMOTE_MNG 0x4003 /* tx none, rx unsigned */ 92 #define TIPC_CMD_GET_MAX_PORTS 0x4004 /* tx none, rx unsigned */ 99 #define TIPC_CMD_GET_NETID 0x400B /* tx none, rx unsigned */ 101 #define TIPC_CMD_ENABLE_BEARER 0x4101 /* tx bearer_config, rx none */ 102 #define TIPC_CMD_DISABLE_BEARER 0x4102 /* tx bearer_name, rx none */ 103 #define TIPC_CMD_SET_LINK_TOL 0x4107 /* tx link_config, rx none */ 104 #define TIPC_CMD_SET_LINK_PRI 0x4108 /* tx link_config, rx none */ 105 #define TIPC_CMD_SET_LINK_WINDOW 0x4109 /* tx link_config, rx none */ 108 #define TIPC_CMD_RESET_LINK_STATS 0x410C /* tx link_name, rx none */ 116 #define TIPC_CMD_SET_NODE_ADDR 0x8001 /* tx net_addr, rx none */ 117 #define TIPC_CMD_SET_REMOTE_MNG 0x8003 /* tx unsigned, rx none */ 118 #define TIPC_CMD_SET_MAX_PORTS 0x8004 /* tx unsigned, rx none */ 125 #define TIPC_CMD_SET_NETID 0x800B /* tx unsigned, rx none */ 133 #define TIPC_CMD_NOT_NET_ADMIN 0xC001 /* tx none, rx none */
|
/linux-4.4.14/arch/mips/include/asm/mach-bcm63xx/ |
H A D | bcm63xx_iudma.h | 7 * rx/tx dma descriptor
|
/linux-4.4.14/arch/arm/mach-omap1/ |
H A D | mcbsp.c | 105 .name = "rx", 115 .name = "rx", 132 .name = "rx", 142 .name = "rx", 182 .name = "rx", 192 .name = "rx", 209 .name = "rx", 219 .name = "rx", 236 .name = "rx", 246 .name = "rx", 289 .name = "rx", 299 .name = "rx", 316 .name = "rx", 326 .name = "rx", 343 .name = "rx", 353 .name = "rx",
|
/linux-4.4.14/drivers/spi/ |
H A D | spi-dln2.c | 194 } rx; dln2_spi_get_cs_num() local 195 unsigned rx_len = sizeof(rx); dln2_spi_get_cs_num() 199 &rx, &rx_len); dln2_spi_get_cs_num() 202 if (rx_len < sizeof(rx)) dln2_spi_get_cs_num() 205 *cs_num = le16_to_cpu(rx.cs_count); dln2_spi_get_cs_num() 220 } rx; dln2_spi_get_speed() local 221 unsigned rx_len = sizeof(rx); dln2_spi_get_speed() 225 ret = dln2_transfer(dln2->pdev, cmd, &tx, sizeof(tx), &rx, &rx_len); dln2_spi_get_speed() 228 if (rx_len < sizeof(rx)) dln2_spi_get_speed() 231 *freq = le32_to_cpu(rx.speed); dln2_spi_get_speed() 270 } rx; dln2_spi_set_speed() local 271 int rx_len = sizeof(rx); dln2_spi_set_speed() 277 &rx, &rx_len); dln2_spi_set_speed() 280 if (rx_len < sizeof(rx)) dln2_spi_set_speed() 329 } *rx = dln2->buf; dln2_spi_get_supported_frame_sizes() local 330 unsigned rx_len = sizeof(*rx); dln2_spi_get_supported_frame_sizes() 336 &tx, sizeof(tx), rx, &rx_len); dln2_spi_get_supported_frame_sizes() 339 if (rx_len < sizeof(*rx)) dln2_spi_get_supported_frame_sizes() 341 if (rx->count > ARRAY_SIZE(rx->frame_sizes)) dln2_spi_get_supported_frame_sizes() 345 for (i = 0; i < rx->count; i++) dln2_spi_get_supported_frame_sizes() 346 *bpw_mask |= BIT(rx->frame_sizes[i] - 1); dln2_spi_get_supported_frame_sizes() 462 } __packed *rx = dln2->buf; dln2_spi_read_one() local 463 unsigned rx_len = sizeof(*rx); dln2_spi_read_one() 465 BUILD_BUG_ON(sizeof(*rx) > DLN2_SPI_BUF_SIZE); dln2_spi_read_one() 475 rx, &rx_len); dln2_spi_read_one() 478 if (rx_len < sizeof(rx->size) + data_len) dln2_spi_read_one() 480 if (le16_to_cpu(rx->size) != data_len) dln2_spi_read_one() 483 dln2_spi_copy_from_buf(data, rx->buf, data_len, dln2->bpw); dln2_spi_read_one() 504 } __packed *rx; dln2_spi_read_write_one() local 508 sizeof(*rx) > DLN2_SPI_BUF_SIZE); dln2_spi_read_write_one() 515 * safe to use the same buffer for both tx and rx. When DLN2 sends the dln2_spi_read_write_one() 516 * response back, with the rx data, we don't need the tx buffer anymore. dln2_spi_read_write_one() 519 rx = dln2->buf; dln2_spi_read_write_one() 528 rx_len = sizeof(*rx); dln2_spi_read_write_one() 531 rx, &rx_len); dln2_spi_read_write_one() 534 if (rx_len < sizeof(rx->size) + data_len) dln2_spi_read_write_one() 536 if (le16_to_cpu(rx->size) != data_len) dln2_spi_read_write_one() 539 dln2_spi_copy_from_buf(rx_data, rx->buf, data_len, dln2->bpw); dln2_spi_read_write_one()
|
H A D | spi-adi-v3.c | 74 void *rx; member in struct:adi_spi_master 106 u32 tx_dummy_val; /* tx value for rx only transfer */ 228 /* discard invalid rx data and empty rfifo */ dummy_read() 251 while (drv_data->rx < drv_data->rx_end) { adi_spi_u8_read() 255 *(u8 *)(drv_data->rx++) = ioread32(&drv_data->regs->rfifo); adi_spi_u8_read() 262 while (drv_data->rx < drv_data->rx_end) { adi_spi_u8_duplex() 266 *(u8 *)(drv_data->rx++) = ioread32(&drv_data->regs->rfifo); adi_spi_u8_duplex() 293 while (drv_data->rx < drv_data->rx_end) { adi_spi_u16_read() 297 *(u16 *)drv_data->rx = ioread32(&drv_data->regs->rfifo); adi_spi_u16_read() 298 drv_data->rx += 2; adi_spi_u16_read() 305 while (drv_data->rx < drv_data->rx_end) { adi_spi_u16_duplex() 310 *(u16 *)drv_data->rx = ioread32(&drv_data->regs->rfifo); adi_spi_u16_duplex() 311 drv_data->rx += 2; adi_spi_u16_duplex() 338 while (drv_data->rx < drv_data->rx_end) { adi_spi_u32_read() 342 *(u32 *)drv_data->rx = ioread32(&drv_data->regs->rfifo); adi_spi_u32_read() 343 drv_data->rx += 4; adi_spi_u32_read() 350 while (drv_data->rx < drv_data->rx_end) { adi_spi_u32_duplex() 355 *(u32 *)drv_data->rx = ioread32(&drv_data->regs->rfifo); adi_spi_u32_duplex() 356 drv_data->rx += 4; adi_spi_u32_duplex() 405 drv->rx = t->rx_buf; adi_spi_setup_transfer() 406 drv->rx_end = drv->rx + t->len; adi_spi_setup_transfer() 408 drv->rx = NULL; adi_spi_setup_transfer() 466 if (!drv_data->rx) { adi_spi_dma_xfer() 476 rx_buf = drv_data->rx; adi_spi_dma_xfer() 483 rx_buf = drv_data->rx; adi_spi_dma_xfer() 534 if (!drv_data->rx) { adi_spi_pio_xfer() 542 if (drv_data->rx != drv_data->rx_end) adi_spi_pio_xfer() 783 "spi rx dma error: %d\n", dma_stat); adi_spi_rx_dma_isr() 789 "dma interrupt missing: tx=%d,rx=%d\n", adi_spi_rx_dma_isr() 826 dev_err(dev, "can not get rx dma resource\n"); adi_spi_probe() 864 /* request tx and rx dma */ adi_spi_probe()
|
H A D | spi-fsl-lib.c | 36 type *rx = mpc8xxx_spi->rx; \ 37 *rx++ = (type)(data >> mpc8xxx_spi->rx_shift); \ 38 mpc8xxx_spi->rx = rx; \
|
H A D | spi-altera.c | 61 unsigned char *rx; member in struct:altera_spi 124 hw->rx = t->rx_buf; altera_spi_txrx() 153 if (hw->rx) { altera_spi_txrx() 156 hw->rx[hw->count] = rxd; altera_spi_txrx() 159 hw->rx[hw->count * 2] = rxd; altera_spi_txrx() 160 hw->rx[hw->count * 2 + 1] = rxd >> 8; altera_spi_txrx() 178 if (hw->rx) { altera_spi_irq() 181 hw->rx[hw->count] = rxd; altera_spi_irq() 184 hw->rx[hw->count * 2] = rxd; altera_spi_irq() 185 hw->rx[hw->count * 2 + 1] = rxd >> 8; altera_spi_irq()
|
H A D | spi-dw.c | 157 * Another concern is about the tx/rx mismatch, we tx_max() 160 * data which is out of tx/rx fifo and inside the tx_max() 164 rxtx_gap = ((dws->rx_end - dws->rx) - (dws->tx_end - dws->tx)) tx_max() 170 /* Return the max entries we should read out of rx fifo */ rx_max() 173 u32 rx_left = (dws->rx_end - dws->rx) / dws->n_bytes; rx_max() 203 /* Care rx only if the transfer's original "rx" is not null */ dw_reader() 206 *(u8 *)(dws->rx) = rxw; dw_reader() 208 *(u16 *)(dws->rx) = rxw; dw_reader() 210 dws->rx += dws->n_bytes; dw_reader() 235 if (dws->rx_end == dws->rx) { interrupt_transfer() 274 } while (dws->rx_end > dws->rx); poll_transfer() 294 dws->rx = transfer->rx_buf; dw_spi_transfer_one() 295 dws->rx_end = dws->rx + transfer->len; dw_spi_transfer_one() 330 if (dws->rx && dws->tx) dw_spi_transfer_one() 332 else if (dws->rx) dw_spi_transfer_one() 448 /* Restart the controller, disable all interrupts, clean rx fifo */ spi_hw_init()
|
H A D | spi-omap2-mcspi.c | 566 u8 *rx; omap2_mcspi_txrx_dma() local 610 rx = xfer->rx_buf; omap2_mcspi_txrx_dma() 616 if (rx != NULL) omap2_mcspi_txrx_dma() 636 if (rx == NULL) { omap2_mcspi_txrx_dma() 688 u8 *rx; omap2_mcspi_txrx_pio() local 691 rx = xfer->rx_buf; omap2_mcspi_txrx_pio() 706 if (rx != NULL) { omap2_mcspi_txrx_pio() 716 *rx++ = readl_relaxed(rx_reg); omap2_mcspi_txrx_pio() 718 word_len, *(rx - 1)); omap2_mcspi_txrx_pio() 730 *rx++ = readl_relaxed(rx_reg); omap2_mcspi_txrx_pio() 732 word_len, *(rx - 1)); omap2_mcspi_txrx_pio() 736 u16 *rx; omap2_mcspi_txrx_pio() local 739 rx = xfer->rx_buf; omap2_mcspi_txrx_pio() 753 if (rx != NULL) { omap2_mcspi_txrx_pio() 763 *rx++ = readl_relaxed(rx_reg); omap2_mcspi_txrx_pio() 765 word_len, *(rx - 1)); omap2_mcspi_txrx_pio() 777 *rx++ = readl_relaxed(rx_reg); omap2_mcspi_txrx_pio() 779 word_len, *(rx - 1)); omap2_mcspi_txrx_pio() 783 u32 *rx; omap2_mcspi_txrx_pio() local 786 rx = xfer->rx_buf; omap2_mcspi_txrx_pio() 800 if (rx != NULL) { omap2_mcspi_txrx_pio() 810 *rx++ = readl_relaxed(rx_reg); omap2_mcspi_txrx_pio() 812 word_len, *(rx - 1)); omap2_mcspi_txrx_pio() 824 *rx++ = readl_relaxed(rx_reg); omap2_mcspi_txrx_pio() 826 word_len, *(rx - 1)); omap2_mcspi_txrx_pio() 840 /* disable chan to purge rx datas received in TX_ONLY transfer, omap2_mcspi_txrx_pio() 841 * otherwise these rx datas will affect the direct following omap2_mcspi_txrx_pio() 959 * for both rx and tx. Otherwise we'll do PIO for both rx and tx. 1265 rx_buf ? "rx" : "", omap2_mcspi_transfer_one() 1440 sprintf(dma_rx_ch_name, "rx%d", i); omap2_mcspi_probe()
|
H A D | spi-bfin5xx.c | 89 void *rx; member in struct:bfin_spi_master_data 221 /* used to kick off transfer in rx mode and read unwanted RX data */ bfin_spi_dummy_read() 250 while (drv_data->rx < drv_data->rx_end) { bfin_spi_u8_reader() 254 *(u8 *) (drv_data->rx++) = bfin_read(&drv_data->regs->rdbr); bfin_spi_u8_reader() 263 while (drv_data->rx < drv_data->rx_end) { bfin_spi_u8_duplex() 267 *(u8 *) (drv_data->rx++) = bfin_read(&drv_data->regs->rdbr); bfin_spi_u8_duplex() 301 while (drv_data->rx < drv_data->rx_end) { bfin_spi_u16_reader() 305 *(u16 *) (drv_data->rx) = bfin_read(&drv_data->regs->rdbr); bfin_spi_u16_reader() 306 drv_data->rx += 2; bfin_spi_u16_reader() 315 while (drv_data->rx < drv_data->rx_end) { bfin_spi_u16_duplex() 320 *(u16 *) (drv_data->rx) = bfin_read(&drv_data->regs->rdbr); bfin_spi_u16_duplex() 321 drv_data->rx += 2; bfin_spi_u16_duplex() 392 (drv_data->rx && drv_data->rx >= (drv_data->rx_end - n_bytes))) { bfin_spi_pio_irq_handler() 394 if (drv_data->rx) { bfin_spi_pio_irq_handler() 397 u16 *buf = (u16 *)drv_data->rx; bfin_spi_pio_irq_handler() 401 u8 *buf = (u8 *)drv_data->rx; bfin_spi_pio_irq_handler() 405 drv_data->rx += n_bytes; bfin_spi_pio_irq_handler() 421 if (drv_data->rx && drv_data->tx) { bfin_spi_pio_irq_handler() 425 u16 *buf = (u16 *)drv_data->rx; bfin_spi_pio_irq_handler() 432 u8 *buf = (u8 *)drv_data->rx; bfin_spi_pio_irq_handler() 439 } else if (drv_data->rx) { bfin_spi_pio_irq_handler() 443 u16 *buf = (u16 *)drv_data->rx; bfin_spi_pio_irq_handler() 449 u8 *buf = (u8 *)drv_data->rx; bfin_spi_pio_irq_handler() 475 if (drv_data->rx) bfin_spi_pio_irq_handler() 476 drv_data->rx += n_bytes; bfin_spi_pio_irq_handler() 494 if (drv_data->rx != NULL) { bfin_spi_dma_irq_handler() 628 drv_data->rx = transfer->rx_buf; bfin_spi_pump_transfers() 629 drv_data->rx_end = drv_data->rx + transfer->len; bfin_spi_pump_transfers() 633 drv_data->rx = NULL; bfin_spi_pump_transfers() 726 /* In dma mode, rx or tx must be NULL in one transfer */ bfin_spi_pump_transfers() 728 if (drv_data->rx != NULL) { bfin_spi_pump_transfers() 731 drv_data->rx, drv_data->len_in_bytes); bfin_spi_pump_transfers() 734 if (bfin_addr_dcacheable((unsigned long) drv_data->rx)) bfin_spi_pump_transfers() 735 invalidate_dcache_range((unsigned long) drv_data->rx, bfin_spi_pump_transfers() 736 (unsigned long) (drv_data->rx + bfin_spi_pump_transfers() 740 dma_start_addr = (unsigned long)drv_data->rx; bfin_spi_pump_transfers() 824 (drv_data->rx_end - drv_data->rx)); bfin_spi_pump_transfers() 841 } else if (drv_data->rx != NULL) { bfin_spi_pump_transfers() 847 if (drv_data->rx != drv_data->rx_end) bfin_spi_pump_transfers()
|
H A D | spi-sirf.c | 151 * only if the rx/tx buffer and transfer size are 4-bytes aligned, we use dma 260 /* rx & tx bufs from the spi_transfer */ 262 void *rx; member in struct:sirfsoc_spi 264 /* place received word into rx buffer */ 273 /* rx & tx DMA channels */ 281 * if tx size is not more than 4 and rx size is NULL, use 316 u8 *rx = sspi->rx; spi_sirfsoc_rx_word_u8() local 320 if (rx) { spi_sirfsoc_rx_word_u8() 321 *rx++ = (u8) data; spi_sirfsoc_rx_word_u8() 322 sspi->rx = rx; spi_sirfsoc_rx_word_u8() 344 u16 *rx = sspi->rx; spi_sirfsoc_rx_word_u16() local 348 if (rx) { spi_sirfsoc_rx_word_u16() 349 *rx++ = (u16) data; spi_sirfsoc_rx_word_u16() 350 sspi->rx = rx; spi_sirfsoc_rx_word_u16() 373 u32 *rx = sspi->rx; spi_sirfsoc_rx_word_u32() local 377 if (rx) { spi_sirfsoc_rx_word_u32() 378 *rx++ = (u32) data; spi_sirfsoc_rx_word_u32() 379 sspi->rx = rx; spi_sirfsoc_rx_word_u32() 548 sspi->dst_start = dma_map_single(&spi->dev, sspi->rx, t->len, spi_sirfsoc_dma_transfer() 586 * we get rx data by writing tx data, so if rx is done, tx has spi_sirfsoc_dma_transfer() 711 sspi->rx = t->rx_buf; spi_sirfsoc_transfer() 1132 sspi->rx_chan = dma_request_slave_channel(&pdev->dev, "rx"); spi_sirfsoc_probe() 1134 dev_err(&pdev->dev, "can not allocate rx dma channel\n"); spi_sirfsoc_probe()
|
H A D | spi-bitbang.c | 72 u8 *rx = t->rx_buf; bitbang_txrx_8() local 80 if (rx) bitbang_txrx_8() 81 *rx++ = word; bitbang_txrx_8() 98 u16 *rx = t->rx_buf; bitbang_txrx_16() local 106 if (rx) bitbang_txrx_16() 107 *rx++ = word; bitbang_txrx_16() 124 u32 *rx = t->rx_buf; bitbang_txrx_32() local 132 if (rx) bitbang_txrx_32() 133 *rx++ = word; bitbang_txrx_32()
|
H A D | spi-omap-100k.c | 199 u8 *rx; omap1_spi100k_txrx_pio() local 202 rx = xfer->rx_buf; omap1_spi100k_txrx_pio() 209 *rx++ = spi100k_read_data(spi->master, word_len); omap1_spi100k_txrx_pio() 212 u16 *rx; omap1_spi100k_txrx_pio() local 215 rx = xfer->rx_buf; omap1_spi100k_txrx_pio() 222 *rx++ = spi100k_read_data(spi->master, word_len); omap1_spi100k_txrx_pio() 225 u32 *rx; omap1_spi100k_txrx_pio() local 228 rx = xfer->rx_buf; omap1_spi100k_txrx_pio() 235 *rx = spi100k_read_data(spi->master, word_len); omap1_spi100k_txrx_pio()
|
H A D | spi-rockchip.c | 188 void *rx; member in struct:rockchip_spi 259 u32 rx_left = (rs->rx_end - rs->rx) / rs->n_bytes; rx_max() 372 *(u8 *)(rs->rx) = (u8)rxw; rockchip_spi_pio_reader() 374 *(u16 *)(rs->rx) = (u16)rxw; rockchip_spi_pio_reader() 375 rs->rx += rs->n_bytes; rockchip_spi_pio_reader() 389 if (rs->rx) { rockchip_spi_pio_transfer() 390 remain = rs->rx_end - rs->rx; rockchip_spi_pio_transfer() 453 if (rs->rx) { rockchip_spi_prepare_dma() 486 /* rx must be started before tx due to spi instinct */ rockchip_spi_prepare_dma() 521 if (rs->rx) rockchip_spi_config() 589 rs->rx = xfer->rx_buf; rockchip_spi_transfer_one() 590 rs->rx_end = rs->rx + xfer->len; rockchip_spi_transfer_one() 596 if (rs->tx && rs->rx) rockchip_spi_transfer_one() 600 else if (rs->rx) rockchip_spi_transfer_one() 613 /* rx: dma must be prepared first */ rockchip_spi_transfer_one() 695 if (!of_property_read_u32(pdev->dev.of_node, "rx-sample-delay-ns", rockchip_spi_probe() 728 rs->dma_rx.ch = dma_request_slave_channel(rs->dev, "rx"); rockchip_spi_probe()
|
H A D | spi-dw-mid.c | 50 struct dw_dma_slave *rx = dws->dma_rx; mid_spi_dma_init() local 64 /* 1. Init rx channel */ mid_spi_dma_init() 65 rx->dma_dev = &dma_dev->dev; mid_spi_dma_init() 66 dws->rxchan = dma_request_channel(mask, mid_spi_dma_chan_filter, rx); mid_spi_dma_init() 182 * dws->dma_chan_busy is set before the dma transfer starts, callback for rx 258 /* rx must be started before tx due to spi instinct */ mid_spi_dma_transfer()
|
H A D | spi-au1550.c | 59 u8 *rx; member in struct:au1550_spi 281 * for dma spi transfers, we have to setup rx channel, otherwise there is 284 * and if only tx dma channel is set up (and rx fifo overflow event masked) 285 * spi master done event irq is not generated unless rx fifo is empty (emptied) 286 * so we need rx tmp buffer to use for rx dma if user does not provide one 326 hw->rx = t->rx_buf; au1550_spi_dma_txrxb() 335 * use rx buffer in place of tx if tx buffer was not provided au1550_spi_dma_txrxb() 336 * use temp rx buffer (preallocated or realloc to fit) for rx dma au1550_spi_dma_txrxb() 354 dev_err(hw->dev, "rx dma map error\n"); au1550_spi_dma_txrxb() 366 hw->rx = hw->dma_rx_tmpbuf; au1550_spi_dma_txrxb() 375 hw->tx = hw->rx; au1550_spi_dma_txrxb() 379 res = au1xxx_dbdma_put_dest(hw->dma_rx_ch, virt_to_phys(hw->rx), au1550_spi_dma_txrxb() 382 dev_err(hw->dev, "rx dma put dest error\n"); au1550_spi_dma_txrxb() 483 if (hw->rx) { \ 484 *(u##size *)hw->rx = (u##size)fifoword; \ 485 hw->rx += (size) / 8; \ 518 hw->rx = t->rx_buf; au1550_spi_pio_txrxb() 586 * while there is something to read from rx fifo au1550_spi_pio_irq_callback() 839 "Cannot allocate rx dma channel\n"); au1550_spi_probe() 847 "Cannot allocate rx dma descriptors\n"); au1550_spi_probe() 856 "Cannot allocate initial rx dma tmp buffer\n"); au1550_spi_probe()
|
H A D | spi-rspi.c | 487 static int rspi_pio_transfer(struct rspi_data *rspi, const u8 *tx, u8 *rx, rspi_pio_transfer() argument 496 if (rx) { rspi_pio_transfer() 500 *rx++ = ret; rspi_pio_transfer() 516 struct sg_table *rx) rspi_dma_transfer() 525 if (rx) { rspi_dma_transfer() 527 rx->sgl, rx->nents, DMA_FROM_DEVICE, rspi_dma_transfer() 554 if (rx) { rspi_dma_transfer() 576 if (rx && rspi->rx_irq != other_irq) rspi_dma_transfer() 583 if (rx) rspi_dma_transfer() 597 if (rx) rspi_dma_transfer() 605 if (rx && rspi->rx_irq != other_irq) rspi_dma_transfer() 611 if (rx) rspi_dma_transfer() 726 u8 *rx, unsigned int len) qspi_trigger_transfer_out_in() 749 *rx++ = rspi_read_data(rspi); qspi_trigger_transfer_out_in() 751 ret = rspi_pio_transfer(rspi, tx, rx, n); qspi_trigger_transfer_out_in() 1013 dir == DMA_MEM_TO_DEV ? "tx" : "rx"); rspi_request_dma_chan() 1238 ret = platform_get_irq_byname(pdev, "rx"); rspi_probe() 1263 "rx", rspi); rspi_probe() 515 rspi_dma_transfer(struct rspi_data *rspi, struct sg_table *tx, struct sg_table *rx) rspi_dma_transfer() argument 725 qspi_trigger_transfer_out_in(struct rspi_data *rspi, const u8 *tx, u8 *rx, unsigned int len) qspi_trigger_transfer_out_in() argument
|
/linux-4.4.14/drivers/net/wireless/zd1211rw/ |
H A D | zd_usb.c | 665 struct zd_usb_rx *rx; rx_urb_complete() local 688 rx = &usb->rx; rx_urb_complete() 690 tasklet_schedule(&rx->reset_timer_tasklet); rx_urb_complete() 692 if (length%rx->usb_packet_size > rx->usb_packet_size-4) { rx_urb_complete() 695 ZD_ASSERT(length <= ARRAY_SIZE(rx->fragment)); rx_urb_complete() 696 spin_lock(&rx->lock); rx_urb_complete() 697 memcpy(rx->fragment, buffer, length); rx_urb_complete() 698 rx->fragment_length = length; rx_urb_complete() 699 spin_unlock(&rx->lock); rx_urb_complete() 703 spin_lock(&rx->lock); rx_urb_complete() 704 if (rx->fragment_length > 0) { rx_urb_complete() 706 ZD_ASSERT(length + rx->fragment_length <= rx_urb_complete() 707 ARRAY_SIZE(rx->fragment)); rx_urb_complete() 709 memcpy(rx->fragment+rx->fragment_length, buffer, length); rx_urb_complete() 710 handle_rx_packet(usb, rx->fragment, rx_urb_complete() 711 rx->fragment_length + length); rx_urb_complete() 712 rx->fragment_length = 0; rx_urb_complete() 713 spin_unlock(&rx->lock); rx_urb_complete() 715 spin_unlock(&rx->lock); rx_urb_complete() 761 struct zd_usb_rx *rx = &usb->rx; __zd_usb_enable_rx() local 777 spin_lock_irq(&rx->lock); __zd_usb_enable_rx() 778 if (rx->urbs) { __zd_usb_enable_rx() 779 spin_unlock_irq(&rx->lock); __zd_usb_enable_rx() 783 rx->urbs = urbs; __zd_usb_enable_rx() 784 rx->urbs_count = RX_URBS_COUNT; __zd_usb_enable_rx() 785 spin_unlock_irq(&rx->lock); __zd_usb_enable_rx() 798 spin_lock_irq(&rx->lock); __zd_usb_enable_rx() 799 rx->urbs = NULL; __zd_usb_enable_rx() 800 rx->urbs_count = 0; __zd_usb_enable_rx() 801 spin_unlock_irq(&rx->lock); __zd_usb_enable_rx() 813 struct zd_usb_rx *rx = &usb->rx; zd_usb_enable_rx() local 815 mutex_lock(&rx->setup_mutex); zd_usb_enable_rx() 817 mutex_unlock(&rx->setup_mutex); zd_usb_enable_rx() 830 struct zd_usb_rx *rx = &usb->rx; __zd_usb_disable_rx() local 832 spin_lock_irqsave(&rx->lock, flags); __zd_usb_disable_rx() 833 urbs = rx->urbs; __zd_usb_disable_rx() 834 count = rx->urbs_count; __zd_usb_disable_rx() 835 spin_unlock_irqrestore(&rx->lock, flags); __zd_usb_disable_rx() 845 spin_lock_irqsave(&rx->lock, flags); __zd_usb_disable_rx() 846 rx->urbs = NULL; __zd_usb_disable_rx() 847 rx->urbs_count = 0; __zd_usb_disable_rx() 848 spin_unlock_irqrestore(&rx->lock, flags); __zd_usb_disable_rx() 853 struct zd_usb_rx *rx = &usb->rx; zd_usb_disable_rx() local 855 mutex_lock(&rx->setup_mutex); zd_usb_disable_rx() 857 mutex_unlock(&rx->setup_mutex); zd_usb_disable_rx() 859 tasklet_kill(&rx->reset_timer_tasklet); zd_usb_disable_rx() 860 cancel_delayed_work_sync(&rx->idle_work); zd_usb_disable_rx() 866 struct zd_usb_rx *rx = &usb->rx; zd_usb_reset_rx() local 869 mutex_lock(&rx->setup_mutex); zd_usb_reset_rx() 871 spin_lock_irqsave(&rx->lock, flags); zd_usb_reset_rx() 872 do_reset = rx->urbs != NULL; zd_usb_reset_rx() 873 spin_unlock_irqrestore(&rx->lock, flags); zd_usb_reset_rx() 880 mutex_unlock(&rx->setup_mutex); zd_usb_reset_rx() 1142 container_of(work, struct zd_usb, rx.idle_work.work); zd_rx_idle_timer_handler() 1150 /* 30 seconds since last rx, reset rx */ zd_rx_idle_timer_handler() 1163 struct zd_usb_rx *rx = &usb->rx; zd_usb_reset_rx_idle_timer() local 1165 mod_delayed_work(zd_workqueue, &rx->idle_work, ZD_RX_IDLE_INTERVAL); zd_usb_reset_rx_idle_timer() 1181 struct zd_usb_rx *rx = &usb->rx; init_usb_rx() local 1183 spin_lock_init(&rx->lock); init_usb_rx() 1184 mutex_init(&rx->setup_mutex); init_usb_rx() 1186 rx->usb_packet_size = 512; init_usb_rx() 1188 rx->usb_packet_size = 64; init_usb_rx() 1190 ZD_ASSERT(rx->fragment_length == 0); init_usb_rx() 1191 INIT_DELAYED_WORK(&rx->idle_work, zd_rx_idle_timer_handler); init_usb_rx() 1192 rx->reset_timer_tasklet.func = zd_usb_reset_rx_idle_timer_tasklet; init_usb_rx() 1193 rx->reset_timer_tasklet.data = (unsigned long)usb; init_usb_rx()
|
/linux-4.4.14/net/nfc/nci/ |
H A D | spi.c | 190 struct spi_transfer tx, rx; __nci_spi_read() local 205 memset(&rx, 0, sizeof(struct spi_transfer)); __nci_spi_read() 206 rx.rx_buf = resp_hdr; __nci_spi_read() 207 rx.len = 2; __nci_spi_read() 208 rx.cs_change = 1; __nci_spi_read() 209 rx.speed_hz = nspi->xfer_speed_hz; __nci_spi_read() 210 spi_message_add_tail(&rx, &m); __nci_spi_read() 228 memset(&rx, 0, sizeof(struct spi_transfer)); __nci_spi_read() 229 rx.rx_buf = skb_put(skb, rx_len); __nci_spi_read() 230 rx.len = rx_len; __nci_spi_read() 231 rx.cs_change = 0; __nci_spi_read() 232 rx.delay_usecs = nspi->xfer_udelay; __nci_spi_read() 233 rx.speed_hz = nspi->xfer_speed_hz; __nci_spi_read() 234 spi_message_add_tail(&rx, &m); __nci_spi_read()
|
/linux-4.4.14/drivers/staging/iio/resolver/ |
H A D | ad2s1200.c | 41 u8 rx[2] ____cacheline_aligned; 60 ret = spi_read(st->sdev, st->rx, 2); ad2s1200_read_raw() 68 *val = (((u16)(st->rx[0])) << 4) | ((st->rx[1] & 0xF0) >> 4); ad2s1200_read_raw() 71 vel = (((s16)(st->rx[0])) << 4) | ((st->rx[1] & 0xF0) >> 4); ad2s1200_read_raw()
|
H A D | ad2s90.c | 25 u8 rx[2] ____cacheline_aligned; 38 ret = spi_read(st->sdev, st->rx, 2); ad2s90_read_raw() 41 *val = (((u16)(st->rx[0])) << 4) | ((st->rx[1] & 0xF0) >> 4); ad2s90_read_raw()
|
/linux-4.4.14/drivers/media/pci/cobalt/ |
H A D | cobalt-v4l2.c | 266 int rx = s->video_channel; cobalt_dma_start_streaming() local 268 COBALT_CVI_EVCNT(cobalt, rx); cobalt_dma_start_streaming() 300 int rx = s->video_channel; cobalt_start_streaming() local 301 struct m00389_cvi_regmap __iomem *cvi = COBALT_CVI(cobalt, rx); cobalt_start_streaming() 302 struct m00460_evcnt_regmap __iomem *evcnt = COBALT_CVI_EVCNT(cobalt, rx); cobalt_start_streaming() 317 fw = COBALT_CVI_FREEWHEEL(cobalt, rx); cobalt_start_streaming() 318 vmr = COBALT_CVI_VMR(cobalt, rx); cobalt_start_streaming() 319 clkloss = COBALT_CVI_CLK_LOSS(cobalt, rx); cobalt_start_streaming() 368 int rx = s->video_channel; cobalt_dma_stop_streaming() local 370 COBALT_CVI_EVCNT(cobalt, rx); cobalt_dma_stop_streaming() 406 int rx = s->video_channel; cobalt_stop_streaming() local 428 fw = COBALT_CVI_FREEWHEEL(cobalt, rx); cobalt_stop_streaming() 429 vmr = COBALT_CVI_VMR(cobalt, rx); cobalt_stop_streaming() 430 clkloss = COBALT_CVI_CLK_LOSS(cobalt, rx); cobalt_stop_streaming() 519 int rx = s->video_channel; cobalt_video_input_status_show() local 524 cvi = COBALT_CVI(cobalt, rx); cobalt_video_input_status_show() 525 vmr = COBALT_CVI_VMR(cobalt, rx); cobalt_video_input_status_show() 526 fw = COBALT_CVI_FREEWHEEL(cobalt, rx); cobalt_video_input_status_show() 527 clkloss = COBALT_CVI_CLK_LOSS(cobalt, rx); cobalt_video_input_status_show() 528 packer = COBALT_CVI_PACKER(cobalt, rx); cobalt_video_input_status_show() 533 cobalt_info("rx%d: cvi resolution: %dx%d\n", rx, cobalt_video_input_status_show() 535 cobalt_info("rx%d: cvi control: %s%s%s\n", rx, cobalt_video_input_status_show() 542 cobalt_info("rx%d: cvi status: %s%s\n", rx, cobalt_video_input_status_show() 548 cobalt_info("rx%d: Measurements: %s%s%s%s%s%s%s\n", rx, cobalt_video_input_status_show() 563 cobalt_info("rx%d: irq_status: 0x%02x irq_triggers: 0x%02x\n", rx, cobalt_video_input_status_show() 566 cobalt_info("rx%d: vsync: %d\n", rx, ioread32(&vmr->vsync_time)); cobalt_video_input_status_show() 567 cobalt_info("rx%d: vbp: %d\n", rx, ioread32(&vmr->vback_porch)); cobalt_video_input_status_show() 568 cobalt_info("rx%d: vact: %d\n", rx, ioread32(&vmr->vactive_area)); cobalt_video_input_status_show() 569 cobalt_info("rx%d: vfb: %d\n", rx, ioread32(&vmr->vfront_porch)); cobalt_video_input_status_show() 570 cobalt_info("rx%d: hsync: %d\n", rx, ioread32(&vmr->hsync_time)); cobalt_video_input_status_show() 571 cobalt_info("rx%d: hbp: %d\n", rx, ioread32(&vmr->hback_porch)); cobalt_video_input_status_show() 572 cobalt_info("rx%d: hact: %d\n", rx, ioread32(&vmr->hactive_area)); cobalt_video_input_status_show() 573 cobalt_info("rx%d: hfb: %d\n", rx, ioread32(&vmr->hfront_porch)); cobalt_video_input_status_show() 574 cobalt_info("rx%d: Freewheeling: %s%s%s\n", rx, cobalt_video_input_status_show() 582 cobalt_info("rx%d: Clock Loss Detection: %s%s\n", rx, cobalt_video_input_status_show() 587 cobalt_info("rx%d: Packer: %x\n", rx, ioread32(&packer->control)); cobalt_video_input_status_show()
|
H A D | cobalt-irq.c | 30 int rx = s->video_channel; cobalt_dma_stream_queue_handler() local 32 COBALT_CVI_FREEWHEEL(s->cobalt, rx); cobalt_dma_stream_queue_handler() 34 COBALT_CVI_VMR(s->cobalt, rx); cobalt_dma_stream_queue_handler() 36 COBALT_CVI(s->cobalt, rx); cobalt_dma_stream_queue_handler() 38 COBALT_CVI_CLK_LOSS(s->cobalt, rx); cobalt_dma_stream_queue_handler() 180 cobalt_info("full rx FIFO %d\n", i); cobalt_irq_handler()
|
/linux-4.4.14/sound/firewire/dice/ |
H A D | dice-proc.c | 61 "global", "tx", "rx", "ext_sync", "unused2" dice_proc_read() 108 } rx; dice_proc_read() member in union:__anon15218 199 quadlets = min_t(u32, tx_rx_header.size, sizeof(buf.rx) / 4); dice_proc_read() 201 if (dice_proc_read_mem(dice, &buf.rx, sections[4] + 2 + dice_proc_read() 205 snd_iprintf(buffer, "rx %u:\n", stream); dice_proc_read() 206 snd_iprintf(buffer, " iso channel: %d\n", (int)buf.rx.iso); dice_proc_read() 207 snd_iprintf(buffer, " sequence start: %u\n", buf.rx.seq_start); dice_proc_read() 209 buf.rx.number_audio); dice_proc_read() 210 snd_iprintf(buffer, " midi ports: %u\n", buf.rx.number_midi); dice_proc_read() 212 dice_proc_fixup_string(buf.rx.names, RX_NAMES_SIZE); dice_proc_read() 213 snd_iprintf(buffer, " names: %s\n", buf.rx.names); dice_proc_read() 217 buf.rx.ac3_caps); dice_proc_read() 219 buf.rx.ac3_enable); dice_proc_read()
|
/linux-4.4.14/drivers/ipack/devices/ |
H A D | ipoctal.h | 27 * @rx: Number of received bytes 35 unsigned long rx; member in struct:ipoctal_stats
|
/linux-4.4.14/arch/arm/mach-omap2/ |
H A D | omap_hwmod_2xxx_3xxx_ipblock_data.c | 51 { .name = "rx", .dma_req = 50, }, 57 { .name = "rx", .dma_req = 52, }, 63 { .name = "rx", .dma_req = 54, }, 70 { .name = "rx", .dma_req = 28 }, 76 { .name = "rx", .dma_req = 30 }, 101 { .name = "rx", .dma_req = 32 }, 107 { .name = "rx", .dma_req = 34 }, 113 { .name = "rx", .dma_req = 18 },
|
/linux-4.4.14/drivers/staging/octeon/ |
H A D | Makefile | 18 octeon-ethernet-y += ethernet-rx.o
|
/linux-4.4.14/drivers/net/wireless/iwlwifi/ |
H A D | Makefile | 9 iwlwifi-objs += pcie/drv.o pcie/rx.o pcie/tx.o pcie/trans.o
|
/linux-4.4.14/include/linux/ |
H A D | serial.h | 21 __u32 cts, dsr, rng, dcd, tx, rx; member in struct:async_icount
|
H A D | timb_dma.h | 29 * @rx: true if this channel handles data in the direction to 38 bool rx; member in struct:timb_dma_platform_data_channel
|
/linux-4.4.14/drivers/tty/serial/ |
H A D | max3100.c | 116 int rx_enabled; /* if we should rx chars */ 196 static int max3100_sr(struct max3100_port *s, u16 tx, u16 *rx) max3100_sr() argument 215 *rx = be16_to_cpu(erx); max3100_sr() 216 s->tx_empty = (*rx & MAX3100_T) > 0; max3100_sr() 217 dev_dbg(&s->spi->dev, "%04x - %04x\n", tx, *rx); max3100_sr() 221 static int max3100_handlerx(struct max3100_port *s, u16 rx) max3100_handlerx() argument 226 if (rx & MAX3100_R && s->rx_enabled) { max3100_handlerx() 228 ch = rx & (s->parity & MAX3100_7BIT ? 0x7f : 0xff); max3100_handlerx() 229 if (rx & MAX3100_RAFE) { max3100_handlerx() 235 if (max3100_check_parity(s, rx)) { max3100_handlerx() 236 s->port.icount.rx++; max3100_handlerx() 244 s->port.icount.rx++; max3100_handlerx() 252 cts = (rx & MAX3100_CTS) > 0; max3100_handlerx() 265 u16 tx, rx; max3100_work() local 282 max3100_sr(s, MAX3100_WC | conf, &rx); max3100_work() 285 (s->rts ? MAX3100_RTS : 0), &rx); max3100_work() 286 rxchars += max3100_handlerx(s, rx); max3100_work() 289 max3100_sr(s, MAX3100_RD, &rx); max3100_work() 290 rxchars += max3100_handlerx(s, rx); max3100_work() 292 if (rx & MAX3100_T) { max3100_work() 308 max3100_sr(s, tx, &rx); max3100_work() 309 rxchars += max3100_handlerx(s, rx); max3100_work() 322 ((rx & MAX3100_R) || max3100_work() 578 u16 tx, rx; max3100_shutdown() local 581 max3100_sr(s, tx, &rx); max3100_shutdown() 623 u16 tx, rx; max3100_startup() local 625 max3100_sr(s, tx, &rx); max3100_startup() 748 u16 tx, rx; max3100_probe() local 814 max3100_sr(max3100s[i], tx, &rx); max3100_probe() 869 u16 tx, rx; max3100_suspend() local 872 max3100_sr(s, tx, &rx); max3100_suspend()
|
H A D | bcm63xx_uart.c | 41 * rx interrupt mask / stat 44 * - rx fifo full 45 * - rx fifo above threshold 46 * - rx fifo not empty for too long 180 * serial core request to stop rx, called before port shutdown 232 * read all chars in rx fifo and send them to core 268 port->icount.rx++; bcm_uart_do_rx() 399 * enable rx & tx operation on uart 411 * disable rx & tx operation on uart 424 * clear all unread data in rx fifo and unsent data in tx fifo 430 /* empty rx and tx fifo */ bcm_uart_flush() 441 * serial core request to initialize uart and start rx operation 456 /* set rx/tx fifo thresh to fifo half size */ bcm_uart_startup() 463 /* set rx fifo timeout to 1 char time */ bcm_uart_startup() 475 /* register irq and enable rx interrupts */ bcm_uart_startup()
|
/linux-4.4.14/drivers/staging/iio/accel/ |
H A D | sca3000_ring.c | 86 u8 *rx; sca3000_read_first_n_hw_rb() local 103 num_available = st->rx[0]; sca3000_read_first_n_hw_rb() 115 &rx, num_read); sca3000_read_first_n_hw_rb() 120 *(((u16 *)rx) + i) = be16_to_cpup((__be16 *)rx + i); sca3000_read_first_n_hw_rb() 122 if (copy_to_user(buf, rx, num_read)) sca3000_read_first_n_hw_rb() 124 kfree(rx); sca3000_read_first_n_hw_rb() 151 val = st->rx[0]; sca3000_query_ring_int() 183 st->rx[0] | this_attr->address); sca3000_set_ring_int() 187 st->rx[0] & ~this_attr->address); sca3000_set_ring_int() 300 (st->rx[0] | SCA3000_RING_BUF_ENABLE)); __sca3000_hw_ring_state_set() 304 (st->rx[0] & ~SCA3000_RING_BUF_ENABLE)); __sca3000_hw_ring_state_set()
|
H A D | sca3000_core.c | 100 .rx_buf = st->rx, sca3000_read_data_short() 121 return !(st->rx[0] & SCA3000_LOCKED); sca3000_reg_lock_on() 220 return st->rx[0]; sca3000_read_ctrl_reg() 242 st->rx[0] & SCA3000_REVID_MAJOR_MASK, sca3000_show_rev() 243 st->rx[0] & SCA3000_REVID_MINOR_MASK); sca3000_show_rev() 302 st->rx[0] &= 0x03; sca3000_show_measurement_mode() 303 switch (st->rx[0]) { sca3000_show_measurement_mode() 361 st->rx[0] &= ~mask; sca3000_store_measurement_mode() 362 st->rx[0] |= (val & mask); sca3000_store_measurement_mode() 363 ret = sca3000_write_reg(st, SCA3000_REG_ADDR_MODE, st->rx[0]); sca3000_store_measurement_mode() 471 *val = (be16_to_cpup((__be16 *)st->rx) >> 3) & 0x1FFF; sca3000_read_raw() 483 *val = ((st->rx[0] & 0x3F) << 3) | sca3000_read_raw() 484 ((st->rx[1] & 0xE0) >> 5); sca3000_read_raw() 522 val = st->rx[0]; sca3000_read_av_freq() 566 switch (0x03 & st->rx[0]) { __sca3000_get_base_freq() 786 val = st->rx[0]; sca3000_event_handler() 852 if ((st->rx[0] & protect_mask) != SCA3000_MEAS_MODE_MOT_DET) { sca3000_read_event_config() 881 val = st->rx[0]; sca3000_query_free_fall_mode() 917 if (val && !(st->rx[0] & protect_mask)) sca3000_set_free_fall_mode() 919 (st->rx[0] | SCA3000_FREE_FALL_DETECT)); sca3000_set_free_fall_mode() 921 else if (!val && (st->rx[0] & protect_mask)) sca3000_set_free_fall_mode() 923 (st->rx[0] & ~protect_mask)); sca3000_set_free_fall_mode() 984 ((st->rx[0] & protect_mask) != SCA3000_MEAS_MODE_MOT_DET)) sca3000_write_event_config() 986 (st->rx[0] & ~protect_mask) sca3000_write_event_config() 990 ((st->rx[0] & protect_mask) == SCA3000_MEAS_MODE_MOT_DET)) sca3000_write_event_config() 992 (st->rx[0] & ~protect_mask)); sca3000_write_event_config() 1077 (st->rx[0] & SCA3000_MODE_PROT_MASK)); sca3000_clean_setup() 1164 (st->rx[0] & sca3000_stop_all_interrupts()
|
H A D | adis16220.h | 129 * @rx: receive buffer 130 * @buf_lock: mutex to protect tx and rx 137 u8 rx[ADIS16220_MAX_RX]; member in struct:adis16220_state
|
/linux-4.4.14/drivers/net/ |
H A D | xen-netfront.c | 86 /* IRQ name is queue name with "-tx" or "-rx" appended */ 111 char rx_irq_name[IRQ_NAME_SIZE]; /* DEVNAME-qN-rx */ 118 * {tx,rx}_skbs store outstanding skbuffs. Free tx_skb entries 136 struct xen_netif_rx_front_ring rx; member in struct:netfront_queue 163 struct xen_netif_rx_response rx; member in struct:netfront_rx_info 282 RING_IDX req_prod = queue->rx.req_prod_pvt; xennet_alloc_rx_buffers() 288 for (req_prod = queue->rx.req_prod_pvt; xennet_alloc_rx_buffers() 289 req_prod - queue->rx.rsp_cons < NET_RX_RING_SIZE; xennet_alloc_rx_buffers() 312 req = RING_GET_REQUEST(&queue->rx, req_prod); xennet_alloc_rx_buffers() 321 queue->rx.req_prod_pvt = req_prod; xennet_alloc_rx_buffers() 324 if (req_prod - queue->rx.rsp_cons < NET_RX_SLOTS_MIN) { xennet_alloc_rx_buffers() 331 RING_PUSH_REQUESTS_AND_CHECK_NOTIFY(&queue->rx, notify); xennet_alloc_rx_buffers() 350 queue->rx.sring->rsp_event = queue->rx.rsp_cons + 1; xennet_open() 351 if (RING_HAS_UNCONSUMED_RESPONSES(&queue->rx)) xennet_open() 709 int new = xennet_rxidx(queue->rx.req_prod_pvt); xennet_move_rx_slot() 714 RING_GET_REQUEST(&queue->rx, queue->rx.req_prod_pvt)->id = new; xennet_move_rx_slot() 715 RING_GET_REQUEST(&queue->rx, queue->rx.req_prod_pvt)->gref = ref; xennet_move_rx_slot() 716 queue->rx.req_prod_pvt++; xennet_move_rx_slot() 726 RING_IDX cons = queue->rx.rsp_cons; xennet_get_extras() 741 RING_GET_RESPONSE(&queue->rx, ++cons); xennet_get_extras() 759 queue->rx.rsp_cons = cons; xennet_get_extras() 767 struct xen_netif_rx_response *rx = &rinfo->rx; xennet_get_responses() local 770 RING_IDX cons = queue->rx.rsp_cons; xennet_get_responses() 773 int max = MAX_SKB_FRAGS + (rx->status <= RX_COPY_THRESHOLD); xennet_get_responses() 778 if (rx->flags & XEN_NETRXF_extra_info) { xennet_get_responses() 780 cons = queue->rx.rsp_cons; xennet_get_responses() 784 if (unlikely(rx->status < 0 || xennet_get_responses() 785 rx->offset + rx->status > XEN_PAGE_SIZE)) { xennet_get_responses() 787 dev_warn(dev, "rx->offset: %u, size: %d\n", xennet_get_responses() 788 rx->offset, rx->status); xennet_get_responses() 801 dev_warn(dev, "Bad rx response id %d.\n", xennet_get_responses() 802 rx->id); xennet_get_responses() 815 if (!(rx->flags & XEN_NETRXF_more_data)) xennet_get_responses() 825 rx = RING_GET_RESPONSE(&queue->rx, cons + slots); xennet_get_responses() 838 queue->rx.rsp_cons = cons + slots; xennet_get_responses() 877 RING_IDX cons = queue->rx.rsp_cons; xennet_fill_frags() 881 struct xen_netif_rx_response *rx = xennet_fill_frags() local 882 RING_GET_RESPONSE(&queue->rx, ++cons); xennet_fill_frags() 894 rx->offset, rx->status, PAGE_SIZE); xennet_fill_frags() 969 struct xen_netif_rx_response *rx = &rinfo.rx; xennet_poll() local 984 rp = queue->rx.sring->rsp_prod; xennet_poll() 987 i = queue->rx.rsp_cons; xennet_poll() 990 memcpy(rx, RING_GET_RESPONSE(&queue->rx, i), sizeof(*rx)); xennet_poll() 1000 i = queue->rx.rsp_cons; xennet_poll() 1012 queue->rx.rsp_cons += skb_queue_len(&tmpq); xennet_poll() 1017 NETFRONT_SKB_CB(skb)->pull_to = rx->status; xennet_poll() 1021 skb_shinfo(skb)->frags[0].page_offset = rx->offset; xennet_poll() 1022 skb_frag_size_set(&skb_shinfo(skb)->frags[0], rx->status); xennet_poll() 1023 skb->data_len = rx->status; xennet_poll() 1024 skb->len += rx->status; xennet_poll() 1028 if (rx->flags & XEN_NETRXF_csum_blank) xennet_poll() 1030 else if (rx->flags & XEN_NETRXF_data_validated) xennet_poll() 1035 queue->rx.rsp_cons = ++i; xennet_poll() 1050 RING_FINAL_CHECK_FOR_RESPONSES(&queue->rx, more_to_do); xennet_poll() 1236 RING_HAS_UNCONSUMED_RESPONSES(&queue->rx))) xennet_rx_interrupt() 1413 xennet_end_access(queue->rx_ring_ref, queue->rx.sring); xennet_disconnect_backend() 1418 queue->rx.sring = NULL; xennet_disconnect_backend() 1506 "%s-rx", queue->name); setup_netfront_split() 1539 queue->rx.sring = NULL; setup_netfront() 1559 xenbus_dev_fatal(dev, err, "allocating rx ring page"); setup_netfront() 1563 FRONT_RING_INIT(&queue->rx, rxs, XEN_PAGE_SIZE); setup_netfront() 1639 /* A grant for every rx ring slot */ xennet_init_queue() 1642 pr_alert("can't alloc rx grant refs\n"); xennet_init_queue() 1691 err = xenbus_printf(*xbt, path, "rx-ring-ref", "%u", write_queue_xenstore_keys() 1694 message = "writing rx-ring-ref"; write_queue_xenstore_keys() 1719 "event-channel-rx", "%u", queue->rx_evtchn); write_queue_xenstore_keys() 1721 message = "writing event-channel-rx"; write_queue_xenstore_keys() 1898 err = xenbus_printf(xbt, dev->nodename, "request-rx-copy", "%u", talk_to_netback() 1901 message = "writing request-rx-copy"; talk_to_netback() 1905 err = xenbus_printf(xbt, dev->nodename, "feature-rx-notify", "%d", 1); talk_to_netback() 1907 message = "writing feature-rx-notify"; talk_to_netback() 1968 "feature-rx-copy", "%u", &feature_rx_copy); xennet_connect()
|
/linux-4.4.14/fs/jffs2/ |
H A D | xattr.c | 137 struct jffs2_raw_xattr rx; do_verify_xattr_datum() local 148 rc = jffs2_flash_read(c, offset, sizeof(rx), &readlen, (char *)&rx); do_verify_xattr_datum() 149 if (rc || readlen != sizeof(rx)) { do_verify_xattr_datum() 151 rc, sizeof(rx), readlen, offset); do_verify_xattr_datum() 154 crc = crc32(0, &rx, sizeof(rx) - 4); do_verify_xattr_datum() 155 if (crc != je32_to_cpu(rx.node_crc)) { do_verify_xattr_datum() 157 offset, je32_to_cpu(rx.hdr_crc), crc); do_verify_xattr_datum() 161 totlen = PAD(sizeof(rx) + rx.name_len + 1 + je16_to_cpu(rx.value_len)); do_verify_xattr_datum() 162 if (je16_to_cpu(rx.magic) != JFFS2_MAGIC_BITMASK do_verify_xattr_datum() 163 || je16_to_cpu(rx.nodetype) != JFFS2_NODETYPE_XATTR do_verify_xattr_datum() 164 || je32_to_cpu(rx.totlen) != totlen do_verify_xattr_datum() 165 || je32_to_cpu(rx.xid) != xd->xid do_verify_xattr_datum() 166 || je32_to_cpu(rx.version) != xd->version) { do_verify_xattr_datum() 169 offset, je16_to_cpu(rx.magic), JFFS2_MAGIC_BITMASK, do_verify_xattr_datum() 170 je16_to_cpu(rx.nodetype), JFFS2_NODETYPE_XATTR, do_verify_xattr_datum() 171 je32_to_cpu(rx.totlen), totlen, do_verify_xattr_datum() 172 je32_to_cpu(rx.xid), xd->xid, do_verify_xattr_datum() 173 je32_to_cpu(rx.version), xd->version); do_verify_xattr_datum() 177 xd->xprefix = rx.xprefix; do_verify_xattr_datum() 178 xd->name_len = rx.name_len; do_verify_xattr_datum() 179 xd->value_len = je16_to_cpu(rx.value_len); do_verify_xattr_datum() 180 xd->data_crc = je32_to_cpu(rx.data_crc); do_verify_xattr_datum() 287 struct jffs2_raw_xattr rx; save_xattr_datum() local 296 vecs[0].iov_base = ℞ save_xattr_datum() 297 vecs[0].iov_len = sizeof(rx); save_xattr_datum() 303 memset(&rx, 0, sizeof(rx)); save_xattr_datum() 304 rx.magic = cpu_to_je16(JFFS2_MAGIC_BITMASK); save_xattr_datum() 305 rx.nodetype = cpu_to_je16(JFFS2_NODETYPE_XATTR); save_xattr_datum() 306 rx.totlen = cpu_to_je32(PAD(totlen)); save_xattr_datum() 307 rx.hdr_crc = cpu_to_je32(crc32(0, &rx, sizeof(struct jffs2_unknown_node) - 4)); save_xattr_datum() 309 rx.xid = cpu_to_je32(xd->xid); save_xattr_datum() 310 rx.version = cpu_to_je32(++xd->version); save_xattr_datum() 311 rx.xprefix = xd->xprefix; save_xattr_datum() 312 rx.name_len = xd->name_len; save_xattr_datum() 313 rx.value_len = cpu_to_je16(xd->value_len); save_xattr_datum() 314 rx.data_crc = cpu_to_je32(crc32(0, vecs[1].iov_base, vecs[1].iov_len)); save_xattr_datum() 315 rx.node_crc = cpu_to_je32(crc32(0, &rx, sizeof(struct jffs2_raw_xattr) - 4)); save_xattr_datum()
|
/linux-4.4.14/include/net/ |
H A D | bond_3ad.h | 50 /* rx machine states(43.4.11 in the 802.3ad standard) */ 53 AD_RX_INITIALIZE, /* rx Machine */ 54 AD_RX_PORT_DISABLED, /* rx Machine */ 55 AD_RX_LACP_DISABLED, /* rx Machine */ 56 AD_RX_EXPIRED, /* rx Machine */ 57 AD_RX_DEFAULTED, /* rx Machine */ 58 AD_RX_CURRENT /* rx Machine */ 92 /* rx indication types */ 98 /* rx marker indication types */ 231 rx_states_t sm_rx_state; /* state machine rx state */ 232 u16 sm_rx_timer_counter; /* state machine rx timer counter */
|
/linux-4.4.14/drivers/iio/accel/ |
H A D | kxsd9.c | 48 * @buf_lock: protect the rx and tx buffers. 50 * @rx: single rx buffer storage 56 u8 rx[KXSD9_STATE_RX_SIZE] ____cacheline_aligned; 108 .rx_buf = st->rx, kxsd9_read() 116 ret = (((u16)(st->rx[0])) << 8) | (st->rx[1] & 0xF0); kxsd9_read()
|
/linux-4.4.14/drivers/staging/wlan-ng/ |
H A D | p80211netdev.c | 139 * the device is supposed to be ready for tx and rx. In our 541 netdev_dbg(dev, "rx'd ioctl, cmd=%d, len=%d\n", cmd, req->len); p80211knetdev_do_ioctl() 732 /* Set up the rx queue */ wlan_setup() 848 /* Now to clean out the rx queue */ unregister_wlandev() 934 netdev_dbg(wlandev->netdev, "rx'd mgmt:\n"); p80211_rx_typedrop() 935 wlandev->rx.mgmt++; p80211_rx_typedrop() 939 wlandev->rx.assocreq++; p80211_rx_typedrop() 943 wlandev->rx.assocresp++; p80211_rx_typedrop() 947 wlandev->rx.reassocreq++; p80211_rx_typedrop() 951 wlandev->rx.reassocresp++; p80211_rx_typedrop() 955 wlandev->rx.probereq++; p80211_rx_typedrop() 959 wlandev->rx.proberesp++; p80211_rx_typedrop() 963 wlandev->rx.beacon++; p80211_rx_typedrop() 967 wlandev->rx.atim++; p80211_rx_typedrop() 971 wlandev->rx.disassoc++; p80211_rx_typedrop() 975 wlandev->rx.authen++; p80211_rx_typedrop() 979 wlandev->rx.deauthen++; p80211_rx_typedrop() 983 wlandev->rx.mgmt_unknown++; p80211_rx_typedrop() 996 netdev_dbg(wlandev->netdev, "rx'd ctl:\n"); p80211_rx_typedrop() 997 wlandev->rx.ctl++; p80211_rx_typedrop() 1001 wlandev->rx.pspoll++; p80211_rx_typedrop() 1005 wlandev->rx.rts++; p80211_rx_typedrop() 1009 wlandev->rx.cts++; p80211_rx_typedrop() 1013 wlandev->rx.ack++; p80211_rx_typedrop() 1017 wlandev->rx.cfend++; p80211_rx_typedrop() 1021 wlandev->rx.cfendcfack++; p80211_rx_typedrop() 1025 wlandev->rx.ctl_unknown++; p80211_rx_typedrop() 1033 wlandev->rx.data++; p80211_rx_typedrop() 1036 wlandev->rx.dataonly++; p80211_rx_typedrop() 1039 wlandev->rx.data_cfack++; p80211_rx_typedrop() 1042 wlandev->rx.data_cfpoll++; p80211_rx_typedrop() 1045 wlandev->rx.data__cfack_cfpoll++; p80211_rx_typedrop() 1048 netdev_dbg(wlandev->netdev, "rx'd data:null\n"); p80211_rx_typedrop() 1049 wlandev->rx.null++; p80211_rx_typedrop() 1052 netdev_dbg(wlandev->netdev, "rx'd data:cfack\n"); p80211_rx_typedrop() 1053 wlandev->rx.cfack++; p80211_rx_typedrop() 1056 netdev_dbg(wlandev->netdev, "rx'd data:cfpoll\n"); p80211_rx_typedrop() 1057 wlandev->rx.cfpoll++; p80211_rx_typedrop() 1060 netdev_dbg(wlandev->netdev, "rx'd data:cfack_cfpoll\n"); p80211_rx_typedrop() 1061 wlandev->rx.cfack_cfpoll++; p80211_rx_typedrop() 1065 wlandev->rx.data_unknown++; p80211_rx_typedrop()
|
/linux-4.4.14/drivers/staging/rtl8723au/include/ |
H A D | ieee80211.h | 101 /* tx: cck only , rx: cck only, hw: cck */ 103 /* tx: ofdm only, rx: ofdm & cck, hw: cck & ofdm */ 105 /* tx: ofdm only, rx: ofdm only, hw: ofdm only */ 107 /* tx: MCS only, rx: MCS & cck, hw: MCS & cck */ 109 /* tx: MCS only, rx: MCS & ofdm, hw: ofdm only */ 115 /* tx: cck & ofdm, rx: cck & ofdm & MCS, hw: cck & ofdm */ 117 /* tx: ofdm & MCS, rx: ofdm & cck & MCS, hw: cck & ofdm */ 119 /* tx: ofdm & MCS, rx: ofdm & MCS, hw: ofdm only */ 121 /* tx: ofdm & cck & MCS, rx: ofdm & cck & MCS, hw: ofdm & cck */ 123 /* tx: ofdm & MCS, rx: ofdm & MCS, hw: ofdm only */
|
H A D | rtw_ht.h | 28 u32 rx_ampdu_maxlen; /* for rx reordering ctrl win_sz, updated when join_callback. */
|
/linux-4.4.14/drivers/net/wireless/rt2x00/ |
H A D | rt2x00config.c | 130 if (config.rx == ANTENNA_SW_DIVERSITY) { rt2x00lib_config_antenna() 133 if (def->rx == ANTENNA_SW_DIVERSITY) rt2x00lib_config_antenna() 134 config.rx = ANTENNA_B; rt2x00lib_config_antenna() 136 config.rx = def->rx; rt2x00lib_config_antenna() 138 } else if (config.rx == ANTENNA_SW_DIVERSITY) rt2x00lib_config_antenna() 139 config.rx = active->rx; rt2x00lib_config_antenna() 158 rt2x00queue_stop_queue(rt2x00dev->rx); rt2x00lib_config_antenna() 172 rt2x00queue_start_queue(rt2x00dev->rx); rt2x00lib_config_antenna()
|
/linux-4.4.14/drivers/net/wireless/ath/ath10k/ |
H A D | htt_rx.c | 31 /* when under memory pressure rx ring refill may fail and needs a retry */ 147 * Make sure the rx buffer is updated before available buffer __ath10k_htt_rx_ring_fill_n() 148 * index to avoid any potential rx ring corruption. __ath10k_htt_rx_ring_fill_n() 190 * another A-MPDU rx, no special recovery is needed. ath10k_htt_rx_msdu_buff_replenish() 258 ath10k_warn(ar, "tried to pop sk_buff from an empty rx ring\n"); ath10k_htt_rx_netbuf_pop() 276 ath10k_dbg_dump(ar, ATH10K_DBG_HTT_DUMP, NULL, "htt rx netbuf pop: ", ath10k_htt_rx_netbuf_pop() 314 * rx data. ath10k_htt_rx_amsdu_pop() 327 * Copy the FW rx descriptor for this MSDU from the rx ath10k_htt_rx_amsdu_pop() 329 * same rx indication message definition as LL, and simply ath10k_htt_rx_amsdu_pop() 330 * appends new info (fields from the HW rx desc, and the ath10k_htt_rx_amsdu_pop() 331 * MSDU payload itself). So, the offset into the rx ath10k_htt_rx_amsdu_pop() 333 * offset of the per-MSDU FW rx desc info within the ath10k_htt_rx_amsdu_pop() 334 * message, and how many bytes of the per-MSDU FW rx desc ath10k_htt_rx_amsdu_pop() 336 * the host, since for a big-endian host, the rx ind ath10k_htt_rx_amsdu_pop() 337 * message contents, including the per-MSDU rx desc bytes, ath10k_htt_rx_amsdu_pop() 344 * per-MSDU rx descriptors. Just to be sure, verify ath10k_htt_rx_amsdu_pop() 382 /* Note: Chained buffers do not contain rx descriptor */ ath10k_htt_rx_amsdu_pop() 416 * refill the rx ring with 1 PPDU's worth of rx buffers (something ath10k_htt_rx_amsdu_pop() 417 * like 32 x 3 buffers), rather than one MPDU's worth of rx buffers ath10k_htt_rx_amsdu_pop() 419 * SW to tell us when it is done pulling all the PPDU's rx buffers ath10k_htt_rx_amsdu_pop() 420 * out of the rx ring, and then refill it just once. ath10k_htt_rx_amsdu_pop() 453 ath10k_dbg_dump(ar, ATH10K_DBG_HTT_DUMP, NULL, "htt rx netbuf pop: ", ath10k_htt_rx_pop_paddr() 527 ath10k_warn(ar, "htt rx ring size is not power of 2\n"); ath10k_htt_rx_alloc() 576 ath10k_dbg(ar, ATH10K_DBG_BOOT, "htt rx ring size %d fill_level %d\n", ath10k_htt_rx_alloc() 739 "rx desc msdu payload: ", ath10k_htt_rx_h_rates() 982 "rx skb %p len %u peer %pM %s %s sn %u %s%s%s%s%s %srate_idx %u vht_nss %u freq %u band %u flag 0x%x fcs-err %i mic-err %i amsdu-more %i\n", ath10k_process_rx() 1003 ath10k_dbg_dump(ar, ATH10K_DBG_HTT_DUMP, NULL, "rx skb: ", ath10k_process_rx() 1067 * the case for frames that come in via fragmented rx indication. ath10k_htt_rx_h_undecap_raw() 1451 /* TODO: bump some rx-oom error stat */ ath10k_unchain_msdu() 1536 ath10k_dbg(ar, ATH10K_DBG_HTT, "htt rx mgmt ctrl\n"); ath10k_htt_rx_amsdu_allowed() 1541 ath10k_dbg(ar, ATH10K_DBG_HTT, "htt rx cac running\n"); ath10k_htt_rx_amsdu_allowed() 1562 struct htt_rx_indication *rx) ath10k_htt_rx_handler() 1578 fw_desc_len = __le16_to_cpu(rx->prefix.fw_rx_desc_bytes); ath10k_htt_rx_handler() 1579 fw_desc = (u8 *)&rx->fw_desc; ath10k_htt_rx_handler() 1581 num_mpdu_ranges = MS(__le32_to_cpu(rx->hdr.info1), ath10k_htt_rx_handler() 1583 mpdu_ranges = htt_rx_ind_get_mpdu_ranges(rx); ath10k_htt_rx_handler() 1585 ath10k_dbg_dump(ar, ATH10K_DBG_HTT_DUMP, NULL, "htt rx ind: ", ath10k_htt_rx_handler() 1586 rx, sizeof(*rx) + ath10k_htt_rx_handler() 1598 ath10k_warn(ar, "rx ring became corrupted: %d\n", ret); ath10k_htt_rx_handler() 1639 ath10k_dbg(ar, ATH10K_DBG_HTT_DUMP, "htt rx frag ahead\n"); ath10k_htt_rx_frag_handler() 1642 ath10k_warn(ar, "failed to pop amsdu from httr rx ring for fragmented rx %d\n", ath10k_htt_rx_frag_handler() 1661 "expecting more fragmented rx in one indication %d\n", ath10k_htt_rx_frag_handler() 1716 "htt rx addba tid %hu peer_id %hu size %hhu\n", ath10k_htt_rx_addba() 1737 "htt rx start rx ba session sta %pM tid %hu size %hhu\n", ath10k_htt_rx_addba() 1756 "htt rx delba tid %hu peer_id %hu\n", ath10k_htt_rx_delba() 1777 "htt rx stop rx ba session sta %pM tid %hu\n", ath10k_htt_rx_delba() 1840 struct htt_rx_offload_msdu *rx; ath10k_htt_rx_h_rx_offload() local 1849 rx = (void *)msdu->data; ath10k_htt_rx_h_rx_offload() 1851 skb_put(msdu, sizeof(*rx)); ath10k_htt_rx_h_rx_offload() 1852 skb_pull(msdu, sizeof(*rx)); ath10k_htt_rx_h_rx_offload() 1854 if (skb_tailroom(msdu) < __le16_to_cpu(rx->msdu_len)) { ath10k_htt_rx_h_rx_offload() 1855 ath10k_warn(ar, "dropping frame: offloaded rx msdu is too long!\n"); ath10k_htt_rx_h_rx_offload() 1860 skb_put(msdu, __le16_to_cpu(rx->msdu_len)); ath10k_htt_rx_h_rx_offload() 1862 /* Offloaded rx header length isn't multiple of 2 nor 4 so the ath10k_htt_rx_h_rx_offload() 1880 ath10k_htt_rx_h_channel(ar, status, NULL, rx->vdev_id); ath10k_htt_rx_h_rx_offload() 1917 "htt rx in ord vdev %i peer %i tid %i offload %i frag %i msdu count %i\n", ath10k_htt_rx_in_ord_ind() 1921 ath10k_warn(ar, "dropping invalid in order rx indication\n"); ath10k_htt_rx_in_ord_ind() 1948 * frames from different PPDUs meaning reported rx rate ath10k_htt_rx_in_ord_ind() 1951 * should still give an idea about rx rate to the user. ath10k_htt_rx_in_ord_ind() 1982 ath10k_dbg(ar, ATH10K_DBG_HTT, "htt rx, msg_type: 0x%0X\n", ath10k_htt_t2h_msg_handler() 1986 ath10k_dbg(ar, ATH10K_DBG_HTT, "htt rx, unsupported msg_type: 0x%0X\n max: 0x%0X", ath10k_htt_t2h_msg_handler() 1561 ath10k_htt_rx_handler(struct ath10k_htt *htt, struct htt_rx_indication *rx) ath10k_htt_rx_handler() argument
|
/linux-4.4.14/drivers/staging/wilc1000/ |
H A D | wilc_wfi_cfgoperations.h | 64 .rx = BIT(IEEE80211_STYPE_ACTION >> 4) | 69 .rx = BIT(IEEE80211_STYPE_ASSOC_REQ >> 4) | 79 .rx = BIT(IEEE80211_STYPE_ACTION >> 4) |
|
/linux-4.4.14/drivers/thunderbolt/ |
H A D | nhi_regs.h | 82 * three bitfields: tx, rx, rx overflow 91 * two bitfields: rx, tx
|
H A D | ctl.c | 31 struct tb_ring *rx; member in struct:tb_ctl 389 ring_rx(pkg->ctl->rx, &pkg->frame); /* tb_ctl_rx_submit() 391 * All rx packets are referenced tb_ctl_rx_submit() 411 goto rx; tb_ctl_rx_callback() 419 goto rx; tb_ctl_rx_callback() 425 goto rx; tb_ctl_rx_callback() 429 goto rx; tb_ctl_rx_callback() 433 rx: tb_ctl_rx_callback() 495 ctl->rx = ring_alloc_rx(nhi, 0, 10); tb_ctl_alloc() 496 if (!ctl->rx) tb_ctl_alloc() 523 if (ctl->rx) tb_ctl_free() 524 ring_free(ctl->rx); tb_ctl_free() 546 ring_start(ctl->rx); tb_ctl_start() 561 ring_stop(ctl->rx); tb_ctl_stop()
|
/linux-4.4.14/drivers/iio/imu/ |
H A D | adis_buffer.c | 30 __be16 *tx, *rx; adis_update_scan_mode() local 45 rx = adis->buffer; adis_update_scan_mode() 46 tx = rx + scan_count; adis_update_scan_mode() 59 adis->xfer[j].rx_buf = &rx[j - 1]; adis_update_scan_mode()
|
/linux-4.4.14/drivers/hwmon/ |
H A D | ad7314.c | 41 u16 rx ____cacheline_aligned; 48 ret = spi_read(chip->spi_dev, (u8 *)&chip->rx, sizeof(chip->rx)); ad7314_spi_read() 54 return be16_to_cpu(chip->rx); ad7314_spi_read()
|
/linux-4.4.14/arch/sh/kernel/cpu/sh3/ |
H A D | serial-sh7720.c | 24 /* Clear PTCR bit 5-2; enable only tx and rx */ sh7720_sci_init_pins()
|
/linux-4.4.14/drivers/soc/ti/ |
H A D | knav_dma.c | 166 if (cfg->u.rx.einfo_present) chan_start() 168 if (cfg->u.rx.psinfo_present) chan_start() 170 if (cfg->u.rx.err_mode == DMA_RETRY) chan_start() 172 v |= (cfg->u.rx.desc_type & DESC_TYPE_MASK) << DESC_TYPE_SHIFT; chan_start() 173 if (cfg->u.rx.psinfo_at_sop) chan_start() 175 v |= (cfg->u.rx.sop_offset & CHAN_SOP_OFF_MASK) chan_start() 177 v |= cfg->u.rx.dst_q & CHAN_QNUM_MASK; chan_start() 183 v = cfg->u.rx.fdq[0] << 16; chan_start() 184 v |= cfg->u.rx.fdq[1] & CHAN_QNUM_MASK; chan_start() 187 v = cfg->u.rx.fdq[2] << 16; chan_start() 188 v |= cfg->u.rx.fdq[3] & CHAN_QNUM_MASK; chan_start() 318 ((chan->direction == DMA_MEM_TO_DEV) ? "tx chan" : "rx flow"), dma_debug_show_channels() 328 chan->cfg.u.rx.einfo_present, dma_debug_show_channels() 329 chan->cfg.u.rx.psinfo_present, dma_debug_show_channels() 330 chan->cfg.u.rx.desc_type); dma_debug_show_channels() 332 chan->cfg.u.rx.dst_q, dma_debug_show_channels() 333 chan->cfg.u.rx.thresh); dma_debug_show_channels() 335 seq_printf(s, "[%d]", chan->cfg.u.rx.fdq[i]); dma_debug_show_channels() 562 dev_dbg(kdev->dev, "rx flow(%d) (%p)\n", chan->flow, chan->reg_rx_flow); pktdma_init_rx_chan() 685 ret = of_property_read_u32(node, "ti,rx-retry-timeout", &timeout); dma_init() 687 dev_dbg(kdev->dev, "unspecified rx timeout using value %d\n", dma_init() 722 dev_info(kdev->dev, "DMA %s registered %d logical channels, flows %d, tx chans: %d, rx chans: %d%s\n", dma_init()
|
/linux-4.4.14/drivers/iio/gyro/ |
H A D | adxrs450.c | 70 * @buf_lock: mutex to protect tx and rx 72 * @rx: receive buffer 78 __be32 rx; member in struct:adxrs450_state 103 .rx_buf = &st->rx, adxrs450_spi_read_reg_16() 105 .len = sizeof(st->rx), adxrs450_spi_read_reg_16() 123 *val = (be32_to_cpu(st->rx) >> 5) & 0xFFFF; adxrs450_spi_read_reg_16() 177 .rx_buf = &st->rx, adxrs450_spi_sensor_data() 179 .len = sizeof(st->rx), adxrs450_spi_sensor_data() 192 *val = (be32_to_cpu(st->rx) >> 10) & 0xFFFF; adxrs450_spi_sensor_data() 212 .rx_buf = &st->rx, adxrs450_spi_initial() 228 *val = be32_to_cpu(st->rx); adxrs450_spi_initial()
|
/linux-4.4.14/drivers/net/ethernet/intel/ixgbe/ |
H A D | ixgbe_dcb_nl.c | 56 int rx = DCB_RX_CONFIG; ixgbe_copy_dcb_cfg() local 95 if (dst->path[rx].prio_type != src->path[rx].prio_type) { ixgbe_copy_dcb_cfg() 96 dst->path[rx].prio_type = src->path[rx].prio_type; ixgbe_copy_dcb_cfg() 100 if (dst->path[rx].bwg_id != src->path[rx].bwg_id) { ixgbe_copy_dcb_cfg() 101 dst->path[rx].bwg_id = src->path[rx].bwg_id; ixgbe_copy_dcb_cfg() 105 if (dst->path[rx].bwg_percent != src->path[rx].bwg_percent) { ixgbe_copy_dcb_cfg() 106 dst->path[rx].bwg_percent = src->path[rx].bwg_percent; ixgbe_copy_dcb_cfg() 110 if (dst->path[rx].up_to_tc_bitmap != ixgbe_copy_dcb_cfg() 111 src->path[rx].up_to_tc_bitmap) { ixgbe_copy_dcb_cfg() 112 dst->path[rx].up_to_tc_bitmap = ixgbe_copy_dcb_cfg() 113 src->path[rx].up_to_tc_bitmap; ixgbe_copy_dcb_cfg() 124 if (dcfg->bw_percentage[rx][j] != scfg->bw_percentage[rx][j]) { ixgbe_copy_dcb_cfg() 125 dcfg->bw_percentage[rx][j] = scfg->bw_percentage[rx][j]; ixgbe_copy_dcb_cfg()
|
/linux-4.4.14/drivers/net/wireless/cw1200/ |
H A D | debug.h | 19 int rx; member in struct:cw1200_debug_priv 55 ++priv->debug->rx; cw1200_debug_rxed()
|
H A D | bh.c | 291 pr_err("rx blew up, len %zu\n", alloc_len); cw1200_bh_rx_helper() 426 int rx, tx, term, suspend; cw1200_bh() local 460 rx = atomic_xchg(&priv->bh_rx, 0); cw1200_bh() 465 (rx || tx || term || suspend || priv->bh_error); cw1200_bh() 468 pr_debug("[BH] - rx: %d, tx: %d, term: %d, bh_err: %d, suspend: %d, status: %ld\n", cw1200_bh() 469 rx, tx, term, suspend, priv->bh_error, status); cw1200_bh() 483 if (priv->hw_bufs_used && (!rx || !tx)) { cw1200_bh() 487 rx = 1; cw1200_bh() 547 rx: cw1200_bh() 554 /* Don't bother trying to rx unless we have data to read */ cw1200_bh() 577 * after we handle rx.. cw1200_bh() 597 goto rx; cw1200_bh()
|
/linux-4.4.14/arch/mips/netlogic/xlp/ |
H A D | cop2-ex.c | 52 : "r"(r->tx), "r"(r->rx)); nlm_cop2_save() 82 : : "m"(*r), "r"(r->tx), "r"(r->rx)); nlm_cop2_restore()
|
/linux-4.4.14/drivers/i2c/busses/ |
H A D | i2c-designware-core.h | 48 * @msg_read_idx: the element index of the current rx message in the msgs 50 * @rx_buf_len: the length of the current rx buffer 51 * @rx_buf: the current rx buffer 58 * @rx_fifo_depth: depth of the hardware rx fifo 59 * @rx_outstanding: current master-rx elements in tx fifo
|
H A D | i2c-dln2.c | 117 } __packed *rx = dln2->buf; dln2_i2c_read() local 118 unsigned rx_len = sizeof(*rx); dln2_i2c_read() 120 BUILD_BUG_ON(sizeof(*rx) > DLN2_I2C_BUF_SIZE); dln2_i2c_read() 129 rx, &rx_len); dln2_i2c_read() 132 if (rx_len < sizeof(rx->buf_len) + data_len) dln2_i2c_read() 134 if (le16_to_cpu(rx->buf_len) != data_len) dln2_i2c_read() 137 memcpy(data, rx->buf, data_len); dln2_i2c_read()
|
/linux-4.4.14/include/xen/interface/io/ |
H A D | netif.h | 36 * If the client sends notification for rx requests then it should specify 37 * feature 'feature-rx-notify' via xenbus. Otherwise the backend will assume 48 * "event-channel-tx" and "event-channel-rx" respectively. If frontend 68 * number of tx and rx rings. 76 * event-channel (or event-channel-{tx,rx}) and {tx,rx}-ring-ref keys, 85 * /local/domain/1/device/vif/0/queue-0/rx-ring-ref = "<ring-ref-rx0>" 87 * /local/domain/1/device/vif/0/queue-0/event-channel-rx = "<evtchn-rx0>" 90 * /local/domain/1/device/vif/0/queue-1/rx-ring-ref = "<ring-ref-rx1" 92 * /local/domain/1/device/vif/0/queue-1/event-channel-rx = "<evtchn-rx1>"
|
/linux-4.4.14/drivers/net/ethernet/toshiba/ |
H A D | ps3_gelic_net.h | 129 GELIC_DESCR_DMA_BUFFER_FULL = 0x00000000, /* used in rx */ 130 GELIC_DESCR_DMA_RESPONSE_ERROR = 0x10000000, /* used in rx, tx */ 131 GELIC_DESCR_DMA_PROTECTION_ERROR = 0x20000000, /* used in rx, tx */ 132 GELIC_DESCR_DMA_FRAME_END = 0x40000000, /* used in rx */ 133 GELIC_DESCR_DMA_FORCE_END = 0x50000000, /* used in rx, tx */ 134 GELIC_DESCR_DMA_CARDOWNED = 0xa0000000, /* used in rx, tx */ 271 u16 rx; member in struct:gelic_vlan_id
|
H A D | ps3_gelic_net.c | 257 /* start rx */ gelic_card_up() 281 /* stop rx */ gelic_card_down() 373 * gelic_descr_prepare_rx - reinitializes a rx descriptor 379 * allocates a new rx skb, iommu-maps it and attaches it to the descriptor. 421 "%s:Could not iommu-map rx buffer\n", __func__); gelic_descr_prepare_rx() 431 * gelic_card_release_rx_chain - free all skb of rx descr 456 * gelic_card_fill_rx_chain - fills descriptors/skbs in the rx chains 459 * fills all descriptors in the rx chain: allocates skbs 484 * gelic_card_alloc_rx_skbs - allocates rx skbs in rx descriptor chains 972 * gelic_card_decode_one_descr - processes an rx descriptor 977 * processes an rx descriptor by iommu-unmapping the data buffer and passing 1004 if (card->vlan[i].rx == vid) { gelic_card_decode_one_descr() 1594 int rx; gelic_card_get_vlan_info() member in struct:__anon7812 1598 .rx = GELIC_LV1_VLAN_RX_ETHERNET_0 gelic_card_get_vlan_info() 1602 .rx = GELIC_LV1_VLAN_RX_WIRELESS gelic_card_get_vlan_info() 1618 card->vlan[i].rx = 0; gelic_card_get_vlan_info() 1623 /* rx tag */ gelic_card_get_vlan_info() 1626 vlan_id_ix[i].rx, gelic_card_get_vlan_info() 1631 "get vlan id for rx(%d) failed(%d)\n", gelic_card_get_vlan_info() 1632 vlan_id_ix[i].rx, status); gelic_card_get_vlan_info() 1634 card->vlan[i].rx = 0; gelic_card_get_vlan_info() 1637 card->vlan[i].rx = (u16)v1; gelic_card_get_vlan_info() 1639 dev_dbg(ctodev(card), "vlan_id[%d] tx=%02x rx=%02x\n", gelic_card_get_vlan_info() 1640 i, card->vlan[i].tx, card->vlan[i].rx); gelic_card_get_vlan_info() 1652 card->vlan[GELIC_PORT_WIRELESS].rx = 0; gelic_card_get_vlan_info() 1755 dev_dbg(ctodev(card), "descr rx %p, tx %p, size %#lx, num %#x\n", ps3_gelic_driver_probe() 1758 /* allocate rx skbs */ ps3_gelic_driver_probe()
|
/linux-4.4.14/drivers/net/fddi/skfp/h/ |
H A D | fplustm.h | 72 __le32 rxd_rbadr ; /* physical rx buffer address */ 113 HW_PTR rx_bmu_ctl ; /* BMU addr for rx start */ 142 u_short rx1_fifo_start ; /* rx queue start address */ 143 u_short rx1_fifo_size ; /* rx queue size */ 144 u_short rx2_fifo_start ; /* rx queue start address */ 145 u_short rx2_fifo_size ; /* rx queue size */ 190 struct s_smt_rx_queue *rx[USED_QUEUES] ; member in struct:s_smt_fp
|
/linux-4.4.14/Documentation/spi/ |
H A D | spidev_test.c | 103 static void transfer(int fd, uint8_t const *tx, uint8_t const *rx, size_t len) transfer() argument 109 .rx_buf = (unsigned long)rx, transfer() 137 hex_dump(rx, len, 32, "RX"); transfer() 257 uint8_t *rx; main() local 306 rx = malloc(size); main() 308 transfer(fd, tx, rx, size); main() 309 free(rx); main()
|
/linux-4.4.14/drivers/staging/most/aim-network/ |
H A D | networking.c | 72 struct net_dev_channel rx; member in struct:net_dev_context 192 BUG_ON(!nd->tx.linked || !nd->rx.linked); most_nd_open() 194 if (most_start_channel(nd->iface, nd->rx.ch_id, &aim)) { most_nd_open() 201 most_stop_channel(nd->iface, nd->rx.ch_id, &aim); most_nd_open() 227 most_stop_channel(nd->iface, nd->rx.ch_id, &aim); most_nd_stop() 340 ch = ccfg->direction == MOST_CH_TX ? &nd->tx : &nd->rx; aim_probe_channel() 346 if (nd->tx.linked || nd->rx.linked) { aim_probe_channel() 385 if (nd->rx.linked && channel_idx == nd->rx.ch_id) aim_disconnect_channel() 386 ch = &nd->rx; aim_disconnect_channel() 400 if (!nd->rx.linked && !nd->tx.linked) { aim_disconnect_channel() 436 if (!nd || !nd->channels_opened || nd->rx.ch_id != mbo->hdm_channel_id) aim_rx_data()
|
/linux-4.4.14/drivers/media/usb/dvb-usb/ |
H A D | ttusb2.c | 123 u8 rx[60];/* (64 -4) */ tt3650_ci_msg() local 124 ret = ttusb2_msg(d, cmd, data, write_len, rx, read_len); tt3650_ci_msg() 126 memcpy(data, rx, read_len); tt3650_ci_msg() 433 u8 rx[9]; /* A CMD_GET_IR_CODE reply is 9 bytes long */ tt3650_rc_query() local 435 ret = ttusb2_msg(d, CMD_GET_IR_CODE, NULL, 0, rx, sizeof(rx)); tt3650_rc_query() 439 if (rx[8] & 0x01) { tt3650_rc_query() 441 st->last_rc_key = RC_SCANCODE_RC5(rx[3], rx[2]); tt3650_rc_query() 442 deb_info("%s: cmd=0x%02x sys=0x%02x\n", __func__, rx[2], rx[3]); tt3650_rc_query() 443 rc_keydown(d->rc_dev, RC_TYPE_RC5, st->last_rc_key, rx[1]); tt3650_rc_query()
|
H A D | pctv452e.c | 503 u8 rx[PCTV_ANSWER_LEN]; pctv452e_power_ctrl() local 523 ret = dvb_usb_generic_rw(d, b0, sizeof(b0), rx, PCTV_ANSWER_LEN, 0); pctv452e_power_ctrl() 530 ret = dvb_usb_generic_rw(d, b0, sizeof(b0), rx, PCTV_ANSWER_LEN, 0); pctv452e_power_ctrl() 543 u8 rx[PCTV_ANSWER_LEN]; pctv452e_rc_query() local 554 ret = dvb_usb_generic_rw(d, b, 4, rx, PCTV_ANSWER_LEN, 0); pctv452e_rc_query() 559 info("%s: read: %2d: %*ph: ", __func__, ret, 3, rx); pctv452e_rc_query() 560 for (i = 0; (i < rx[3]) && ((i+3) < PCTV_ANSWER_LEN); i++) pctv452e_rc_query() 561 info(" %02x", rx[i+3]); pctv452e_rc_query() 566 if ((rx[3] == 9) && (rx[12] & 0x01)) { pctv452e_rc_query() 568 state->last_rc_key = RC_SCANCODE_RC5(rx[7], rx[6]); pctv452e_rc_query() 571 __func__, rx[6], rx[7]); pctv452e_rc_query()
|
/linux-4.4.14/drivers/net/wireless/iwlegacy/ |
H A D | 3945-debug.c | 81 ofdm = &il->_3945.stats.rx.ofdm; il3945_ucode_rx_stats_read() 82 cck = &il->_3945.stats.rx.cck; il3945_ucode_rx_stats_read() 83 general = &il->_3945.stats.rx.general; il3945_ucode_rx_stats_read() 84 accum_ofdm = &il->_3945.accum_stats.rx.ofdm; il3945_ucode_rx_stats_read() 85 accum_cck = &il->_3945.accum_stats.rx.cck; il3945_ucode_rx_stats_read() 86 accum_general = &il->_3945.accum_stats.rx.general; il3945_ucode_rx_stats_read() 87 delta_ofdm = &il->_3945.delta_stats.rx.ofdm; il3945_ucode_rx_stats_read() 88 delta_cck = &il->_3945.delta_stats.rx.cck; il3945_ucode_rx_stats_read() 89 delta_general = &il->_3945.delta_stats.rx.general; il3945_ucode_rx_stats_read() 90 max_ofdm = &il->_3945.max_delta.rx.ofdm; il3945_ucode_rx_stats_read() 91 max_cck = &il->_3945.max_delta.rx.cck; il3945_ucode_rx_stats_read() 92 max_general = &il->_3945.max_delta.rx.general; il3945_ucode_rx_stats_read()
|
/linux-4.4.14/drivers/net/ethernet/altera/ |
H A D | altera_sgdma.c | 86 netdev_err(priv->dev, "error mapping rx descriptor memory\n"); sgdma_initialize() 270 "sgdma rx and rx queue empty!\n"); sgdma_rx_status() 277 /* kick the rx sgdma after reaping this descriptor */ sgdma_rx_status() 359 netdev_err(priv->dev, "no rx buffers available\n"); sgdma_async_read() 366 0, /* read addr 0 for rx dma */ sgdma_async_read() 367 rxbuffer->dma_addr, /* write addr for rx dma */ sgdma_async_read() 369 0, /* EOP: NA for rx dma */ sgdma_async_read() 370 0, /* read fixed: NA for rx dma */ sgdma_async_read() 371 0); /* SOP: NA for rx DMA */ sgdma_async_read() 462 /* adds a tse_buffer to the tail of a rx buffer list 512 /* check and return rx sgdma status without polling
|
/linux-4.4.14/drivers/net/ethernet/amd/ |
H A D | au1000_eth.h | 30 #define NUM_RX_DMA 4 /* Au1x00 has 4 rx hardware descriptors */ 127 u32 vaddr; /* virtual address of rx/tx buffers */ 128 dma_addr_t dma_addr; /* dma address of rx/tx buffers */
|
/linux-4.4.14/drivers/staging/lustre/lnet/klnds/o2iblnd/ |
H A D | o2iblnd_cb.c | 132 kiblnd_drop_rx(kib_rx_t *rx) kiblnd_drop_rx() argument 134 kib_conn_t *conn = rx->rx_conn; kiblnd_drop_rx() 147 kiblnd_post_rx(kib_rx_t *rx, int credit) kiblnd_post_rx() argument 149 kib_conn_t *conn = rx->rx_conn; kiblnd_post_rx() 161 mr = kiblnd_find_dma_mr(conn->ibc_hdev, rx->rx_msgaddr, IBLND_MSG_SIZE); kiblnd_post_rx() 164 rx->rx_sge.lkey = mr->lkey; kiblnd_post_rx() 165 rx->rx_sge.addr = rx->rx_msgaddr; kiblnd_post_rx() 166 rx->rx_sge.length = IBLND_MSG_SIZE; kiblnd_post_rx() 168 rx->rx_wrq.next = NULL; kiblnd_post_rx() 169 rx->rx_wrq.sg_list = &rx->rx_sge; kiblnd_post_rx() 170 rx->rx_wrq.num_sge = 1; kiblnd_post_rx() 171 rx->rx_wrq.wr_id = kiblnd_ptr2wreqid(rx, IBLND_WID_RX); kiblnd_post_rx() 174 LASSERT(rx->rx_nob >= 0); /* not posted */ kiblnd_post_rx() 177 kiblnd_drop_rx(rx); /* No more posts for this rx */ kiblnd_post_rx() 181 rx->rx_nob = -1; /* flag posted */ kiblnd_post_rx() 184 * own this rx (and rx::rx_conn) anymore, LU-5678. kiblnd_post_rx() 187 rc = ib_post_recv(conn->ibc_cmid->qp, &rx->rx_wrq, &bad_wrq); kiblnd_post_rx() 189 CERROR("Can't post rx for %s: %d, bad_wrq: %p\n", kiblnd_post_rx() 191 rx->rx_nob = 0; kiblnd_post_rx() 199 kiblnd_drop_rx(rx); /* No more posts for this rx */ kiblnd_post_rx() 302 kiblnd_handle_rx(kib_rx_t *rx) kiblnd_handle_rx() argument 304 kib_msg_t *msg = rx->rx_msg; kiblnd_handle_rx() 305 kib_conn_t *conn = rx->rx_conn; kiblnd_handle_rx() 334 kiblnd_post_rx(rx, IBLND_POSTRX_NO_CREDIT); kiblnd_handle_rx() 372 msg->ibm_srcnid, rx, 0); kiblnd_handle_rx() 380 msg->ibm_srcnid, rx, 1); kiblnd_handle_rx() 442 msg->ibm_srcnid, rx, 1); kiblnd_handle_rx() 459 kiblnd_post_rx(rx, post_credit); kiblnd_handle_rx() 463 kiblnd_rx_complete(kib_rx_t *rx, int status, int nob) kiblnd_rx_complete() argument 465 kib_msg_t *msg = rx->rx_msg; kiblnd_rx_complete() 466 kib_conn_t *conn = rx->rx_conn; kiblnd_rx_complete() 473 LASSERT(rx->rx_nob < 0); /* was posted */ kiblnd_rx_complete() 474 rx->rx_nob = 0; /* isn't now */ kiblnd_rx_complete() 486 rx->rx_nob = nob; kiblnd_rx_complete() 488 rc = kiblnd_unpack_msg(msg, rx->rx_nob); kiblnd_rx_complete() 490 CERROR("Error %d unpacking rx from %s\n", kiblnd_rx_complete() 499 CERROR("Stale rx from %s\n", kiblnd_rx_complete() 517 list_add_tail(&rx->rx_list, &conn->ibc_early_rxs); kiblnd_rx_complete() 523 kiblnd_handle_rx(rx); kiblnd_rx_complete() 527 CDEBUG(D_NET, "rx %p conn %p\n", rx, conn); kiblnd_rx_complete() 530 kiblnd_drop_rx(rx); /* Don't re-post rx. */ kiblnd_rx_complete() 1587 kiblnd_reply(lnet_ni_t *ni, kib_rx_t *rx, lnet_msg_t *lntmsg) kiblnd_reply() argument 1598 tx = kiblnd_get_idle_tx(ni, rx->rx_conn->ibc_peer->ibp_nid); kiblnd_reply() 1620 rc = kiblnd_init_rdma(rx->rx_conn, tx, kiblnd_reply() 1622 &rx->rx_msg->ibm_u.get.ibgm_rd, kiblnd_reply() 1623 rx->rx_msg->ibm_u.get.ibgm_cookie); kiblnd_reply() 1639 kiblnd_queue_tx(tx, rx->rx_conn); kiblnd_reply() 1653 kib_rx_t *rx = private; kiblnd_recv() local 1654 kib_msg_t *rxmsg = rx->rx_msg; kiblnd_recv() 1655 kib_conn_t *conn = rx->rx_conn; kiblnd_recv() 1672 if (nob > rx->rx_nob) { kiblnd_recv() 1675 nob, rx->rx_nob); kiblnd_recv() 1699 kiblnd_send_completion(rx->rx_conn, IBLND_MSG_PUT_NAK, 0, kiblnd_recv() 1726 kiblnd_send_completion(rx->rx_conn, IBLND_MSG_PUT_NAK, rc, kiblnd_recv() 1749 kiblnd_reply(ni, rx, lntmsg); kiblnd_recv() 1752 kiblnd_send_completion(rx->rx_conn, IBLND_MSG_GET_DONE, kiblnd_recv() 1759 kiblnd_post_rx(rx, post_credit); kiblnd_recv() 1899 kib_rx_t *rx; kiblnd_handle_early_rxs() local 1906 list_for_each_entry_safe(rx, tmp, &conn->ibc_early_rxs, rx_list) { kiblnd_handle_early_rxs() 1907 list_del(&rx->rx_list); kiblnd_handle_early_rxs() 1910 kiblnd_handle_rx(rx); kiblnd_handle_early_rxs()
|
/linux-4.4.14/net/bluetooth/ |
H A D | ecc.c | 694 u64 rx[2][NUM_ECC_DIGITS]; ecc_point_mult() local 699 vli_set(rx[1], point->x); ecc_point_mult() 702 xycz_initial_double(rx[1], ry[1], rx[0], ry[0], initial_z); ecc_point_mult() 706 xycz_add_c(rx[1 - nb], ry[1 - nb], rx[nb], ry[nb]); ecc_point_mult() 707 xycz_add(rx[nb], ry[nb], rx[1 - nb], ry[1 - nb]); ecc_point_mult() 711 xycz_add_c(rx[1 - nb], ry[1 - nb], rx[nb], ry[nb]); ecc_point_mult() 714 vli_mod_sub(z, rx[1], rx[0], curve_p); /* X1 - X0 */ ecc_point_mult() 719 vli_mod_mult_fast(z, z, rx[1 - nb]); /* Xb * yP / (xP * Yb * (X1 - X0)) */ ecc_point_mult() 722 xycz_add(rx[nb], ry[nb], rx[1 - nb], ry[1 - nb]); ecc_point_mult() 724 apply_z(rx[0], ry[0], z); ecc_point_mult() 726 vli_set(result->x, rx[0]); ecc_point_mult()
|
/linux-4.4.14/drivers/net/ethernet/intel/ |
H A D | e100.c | 444 struct rx { struct 445 struct rx *next, *prev; 572 struct rx *rxs ____cacheline_aligned; 573 struct rx *rx_to_use; 574 struct rx *rx_to_clean; 1934 static inline void e100_start_receiver(struct nic *nic, struct rx *rx) e100_start_receiver() argument 1940 if (!rx) rx = nic->rxs; e100_start_receiver() 1943 if (rx->skb) { e100_start_receiver() 1944 e100_exec_cmd(nic, ruc_start, rx->dma_addr); e100_start_receiver() 1950 static int e100_rx_alloc_skb(struct nic *nic, struct rx *rx) e100_rx_alloc_skb() argument 1952 if (!(rx->skb = netdev_alloc_skb_ip_align(nic->netdev, RFD_BUF_LEN))) e100_rx_alloc_skb() 1956 skb_copy_to_linear_data(rx->skb, &nic->blank_rfd, sizeof(struct rfd)); e100_rx_alloc_skb() 1957 rx->dma_addr = pci_map_single(nic->pdev, rx->skb->data, e100_rx_alloc_skb() 1960 if (pci_dma_mapping_error(nic->pdev, rx->dma_addr)) { e100_rx_alloc_skb() 1961 dev_kfree_skb_any(rx->skb); e100_rx_alloc_skb() 1962 rx->skb = NULL; e100_rx_alloc_skb() 1963 rx->dma_addr = 0; e100_rx_alloc_skb() 1970 if (rx->prev->skb) { e100_rx_alloc_skb() 1971 struct rfd *prev_rfd = (struct rfd *)rx->prev->skb->data; e100_rx_alloc_skb() 1972 put_unaligned_le32(rx->dma_addr, &prev_rfd->link); e100_rx_alloc_skb() 1973 pci_dma_sync_single_for_device(nic->pdev, rx->prev->dma_addr, e100_rx_alloc_skb() 1980 static int e100_rx_indicate(struct nic *nic, struct rx *rx, e100_rx_indicate() argument 1984 struct sk_buff *skb = rx->skb; e100_rx_indicate() 1993 pci_dma_sync_single_for_cpu(nic->pdev, rx->dma_addr, e100_rx_indicate() 2013 pci_dma_sync_single_for_device(nic->pdev, rx->dma_addr, e100_rx_indicate() 2027 pci_unmap_single(nic->pdev, rx->dma_addr, e100_rx_indicate() 2074 rx->skb = NULL; e100_rx_indicate() 2082 struct rx *rx; e100_rx_clean() local 2084 struct rx *old_before_last_rx, *new_before_last_rx; e100_rx_clean() 2088 for (rx = nic->rx_to_clean; rx->skb; rx = nic->rx_to_clean = rx->next) { e100_rx_clean() 2089 err = e100_rx_indicate(nic, rx, work_done, work_to_do); e100_rx_clean() 2109 for (rx = nic->rx_to_use; !rx->skb; rx = nic->rx_to_use = rx->next) { e100_rx_clean() 2110 if (unlikely(e100_rx_alloc_skb(nic, rx))) e100_rx_clean() 2158 struct rx *rx; e100_rx_clean_list() local 2164 for (rx = nic->rxs, i = 0; i < count; rx++, i++) { e100_rx_clean_list() 2165 if (rx->skb) { e100_rx_clean_list() 2166 pci_unmap_single(nic->pdev, rx->dma_addr, e100_rx_clean_list() 2168 dev_kfree_skb(rx->skb); e100_rx_clean_list() 2180 struct rx *rx; e100_rx_alloc_list() local 2187 if (!(nic->rxs = kcalloc(count, sizeof(struct rx), GFP_ATOMIC))) e100_rx_alloc_list() 2190 for (rx = nic->rxs, i = 0; i < count; rx++, i++) { e100_rx_alloc_list() 2191 rx->next = (i + 1 < count) ? rx + 1 : nic->rxs; e100_rx_alloc_list() 2192 rx->prev = (i == 0) ? nic->rxs + count - 1 : rx - 1; e100_rx_alloc_list() 2193 if (e100_rx_alloc_skb(nic, rx)) { e100_rx_alloc_list() 2205 rx = nic->rxs->prev->prev; e100_rx_alloc_list() 2206 before_last = (struct rfd *)rx->skb->data; e100_rx_alloc_list() 2209 pci_dma_sync_single_for_device(nic->pdev, rx->dma_addr, e100_rx_alloc_list() 2605 netif_info(nic, drv, nic->netdev, "Ring Param settings: rx: %d, tx %d\n", e100_set_ringparam() 2912 /* D100 MAC doesn't allow rx of vlan packets with normal MTU */ e100_probe()
|
/linux-4.4.14/drivers/staging/nvec/ |
H A D | nvec.c | 494 if (nvec->rx->pos != nvec_msg_size(nvec->rx)) { nvec_rx_completed() 496 (uint) nvec_msg_size(nvec->rx), nvec_rx_completed() 497 (uint) nvec->rx->pos); nvec_rx_completed() 499 nvec_msg_free(nvec, nvec->rx); nvec_rx_completed() 503 if (nvec->rx->data[0] == NVEC_BAT) nvec_rx_completed() 513 list_add_tail(&nvec->rx->node, &nvec->rx_data); nvec_rx_completed() 519 if (!nvec_msg_is_event(nvec->rx)) nvec_rx_completed() 618 nvec->rx = nvec_msg_alloc(nvec, NVEC_MSG_RX); nvec_interrupt() 620 if (unlikely(nvec->rx == NULL)) { nvec_interrupt() 624 nvec->rx->data[0] = received; nvec_interrupt() 625 nvec->rx->pos = 1; nvec_interrupt() 632 if (nvec->rx->data[0] != 0x01) { nvec_interrupt() 638 nvec_msg_free(nvec, nvec->rx); nvec_interrupt() 645 BUG_ON(nvec->rx == NULL); nvec_interrupt() 646 nvec->rx->data[1] = received; nvec_interrupt() 647 nvec->rx->pos = 2; nvec_interrupt() 673 else if (nvec->rx && nvec->rx->pos < NVEC_MSG_SIZE) nvec_interrupt() 674 nvec->rx->data[nvec->rx->pos++] = received; nvec_interrupt() 678 nvec->rx, nvec->rx ? nvec->rx->pos : 0, nvec_interrupt() 840 nvec->rx = &nvec->msg_pool[0]; tegra_nvec_probe()
|
/linux-4.4.14/drivers/net/arcnet/ |
H A D | arc-rawmode.c | 39 static void rx(struct net_device *dev, int bufnum, rx() function 75 arcnet_dump_skb(dev, skb, "rx"); rx() 156 .rx = rx,
|
H A D | rfc1051.c | 40 static void rx(struct net_device *dev, int bufnum, 51 .rx = rx, 121 static void rx(struct net_device *dev, int bufnum, rx() function 154 arcnet_dump_skb(dev, skb, "rx"); rx()
|
/linux-4.4.14/drivers/net/wireless/realtek/rtl818x/rtl8187/ |
H A D | leds.h | 50 /* If the LED is radio or tx/rx */
|
/linux-4.4.14/drivers/net/ethernet/amd/xgbe/ |
H A D | xgbe-desc.c | 344 xgbe_set_buffer_data(&rdata->rx.hdr, &ring->rx_hdr_pa, xgbe_map_rx_buffer() 348 xgbe_set_buffer_data(&rdata->rx.buf, &ring->rx_buf_pa, xgbe_map_rx_buffer() 458 if (rdata->rx.hdr.pa.pages) xgbe_unmap_rdata() 459 put_page(rdata->rx.hdr.pa.pages); xgbe_unmap_rdata() 461 if (rdata->rx.hdr.pa_unmap.pages) { xgbe_unmap_rdata() 462 dma_unmap_page(pdata->dev, rdata->rx.hdr.pa_unmap.pages_dma, xgbe_unmap_rdata() 463 rdata->rx.hdr.pa_unmap.pages_len, xgbe_unmap_rdata() 465 put_page(rdata->rx.hdr.pa_unmap.pages); xgbe_unmap_rdata() 468 if (rdata->rx.buf.pa.pages) xgbe_unmap_rdata() 469 put_page(rdata->rx.buf.pa.pages); xgbe_unmap_rdata() 471 if (rdata->rx.buf.pa_unmap.pages) { xgbe_unmap_rdata() 472 dma_unmap_page(pdata->dev, rdata->rx.buf.pa_unmap.pages_dma, xgbe_unmap_rdata() 473 rdata->rx.buf.pa_unmap.pages_len, xgbe_unmap_rdata() 475 put_page(rdata->rx.buf.pa_unmap.pages); xgbe_unmap_rdata() 479 memset(&rdata->rx, 0, sizeof(rdata->rx)); xgbe_unmap_rdata()
|
/linux-4.4.14/arch/cris/boot/dts/include/dt-bindings/net/ |
H A D | ti-dp83867.h | 27 /* RGMIIDCTL internal delay for rx and tx */
|
/linux-4.4.14/net/dccp/ |
H A D | ccid.c | 143 struct ccid *ccid_new(const u8 id, struct sock *sk, bool rx) ccid_new() argument 151 ccid = kmem_cache_alloc(rx ? ccid_ops->ccid_hc_rx_slab : ccid_new() 156 if (rx) { ccid_new() 170 kmem_cache_free(rx ? ccid_ops->ccid_hc_rx_slab : ccid_new()
|
H A D | feat.c | 41 static int dccp_hdlr_ccid(struct sock *sk, u64 ccid, bool rx) dccp_hdlr_ccid() argument 44 struct ccid *new_ccid = ccid_new(ccid, sk, rx); dccp_hdlr_ccid() 49 if (rx) { dccp_hdlr_ccid() 59 static int dccp_hdlr_seq_win(struct sock *sk, u64 seq_win, bool rx) dccp_hdlr_seq_win() argument 63 if (rx) { dccp_hdlr_seq_win() 75 static int dccp_hdlr_ack_ratio(struct sock *sk, u64 ratio, bool rx) dccp_hdlr_ack_ratio() argument 77 if (rx) dccp_hdlr_ack_ratio() 84 static int dccp_hdlr_ackvec(struct sock *sk, u64 enable, bool rx) dccp_hdlr_ackvec() argument 88 if (rx) { dccp_hdlr_ackvec() 101 static int dccp_hdlr_ndp(struct sock *sk, u64 enable, bool rx) dccp_hdlr_ndp() argument 103 if (!rx) dccp_hdlr_ndp() 110 * `rx' holds when the sending peer informs about his partial coverage via a 117 static int dccp_hdlr_min_cscov(struct sock *sk, u64 cscov, bool rx) dccp_hdlr_min_cscov() argument 121 if (rx) dccp_hdlr_min_cscov() 138 int (*activation_hdlr)(struct sock *sk, u64 val, bool rx); 311 bool rx; __dccp_feat_activate() local 338 rx = (is_local == (dccp_feat_table[idx].rxtx == FEAT_AT_RX)); __dccp_feat_activate() 340 dccp_debug(" -> activating %s %s, %sval=%llu\n", rx ? "RX" : "TX", __dccp_feat_activate() 344 return dccp_feat_table[idx].activation_hdlr(sk, val, rx); __dccp_feat_activate() 1448 } tx, rx; dccp_feat_init() local 1475 ccid_get_builtin_ccids(&rx.val, &rx.len)) dccp_feat_init() 1479 !dccp_feat_prefer(sysctl_dccp_rx_ccid, rx.val, rx.len)) dccp_feat_init() 1486 rc = __feat_register_sp(fn, DCCPF_CCID, false, false, rx.val, rx.len); dccp_feat_init() 1490 kfree(rx.val); dccp_feat_init()
|
/linux-4.4.14/include/dt-bindings/net/ |
H A D | ti-dp83867.h | 27 /* RGMIIDCTL internal delay for rx and tx */
|
/linux-4.4.14/include/linux/can/ |
H A D | led.h | 25 /* keep space for interface name + "-tx"/"-rx"/"-rxtx"
|
/linux-4.4.14/arch/metag/boot/dts/include/dt-bindings/net/ |
H A D | ti-dp83867.h | 27 /* RGMIIDCTL internal delay for rx and tx */
|
/linux-4.4.14/arch/mips/boot/dts/include/dt-bindings/net/ |
H A D | ti-dp83867.h | 27 /* RGMIIDCTL internal delay for rx and tx */
|
/linux-4.4.14/arch/powerpc/boot/dts/include/dt-bindings/net/ |
H A D | ti-dp83867.h | 27 /* RGMIIDCTL internal delay for rx and tx */
|
/linux-4.4.14/arch/arm64/boot/dts/include/dt-bindings/net/ |
H A D | ti-dp83867.h | 27 /* RGMIIDCTL internal delay for rx and tx */
|
/linux-4.4.14/arch/arm/boot/dts/include/dt-bindings/net/ |
H A D | ti-dp83867.h | 27 /* RGMIIDCTL internal delay for rx and tx */
|
/linux-4.4.14/drivers/net/wireless/ath/ath6kl/ |
H A D | testmode.c | 55 ath6kl_warn("failed to allocate testmode rx skb!\n"); ath6kl_tm_rx_event() 66 ath6kl_warn("nla_put failed on testmode rx skb!\n"); ath6kl_tm_rx_event()
|
/linux-4.4.14/drivers/usb/atm/ |
H A D | xusbatm.c | 39 XUSBATM_PARM(rx_endpoint, unsigned char, byte, "rx endpoint number"); 41 XUSBATM_PARM(rx_padding, unsigned char, byte, "rx padding (default 0)"); 43 XUSBATM_PARM(rx_altsetting, unsigned char, byte, "rx altsetting (default 0)"); 44 XUSBATM_PARM(tx_altsetting, unsigned char, byte, "rx altsetting (default 0)"); 104 " rx: ep %02x padd %d alt %2d tx: ep %02x padd %d alt %2d\n", xusbatm_bind() 128 usb_dbg(usbatm, "%s: rx If#=%2d; tx If#=%2d\n", __func__, xusbatm_bind()
|
H A D | usbatm.h | 106 int bulk_in; /* bulk rx endpoint */ 107 int isoc_in; /* isochronous rx endpoint */ 178 unsigned char *cell_buf; /* holds partial rx cell */
|
/linux-4.4.14/drivers/net/wireless/libertas/ |
H A D | rx.c | 89 lbs_deb_rx("rx err: frame received with bad length\n"); lbs_process_rxed_packet() 96 lbs_deb_rx("rx data: skb->len - pkt_ptr = %d-%zd = %zd\n", lbs_process_rxed_packet() 149 lbs_deb_rx("rx data: size of actual packet %d\n", skb->len); lbs_process_rxed_packet() 232 lbs_deb_rx("rx err: frame received with bad length\n"); process_rxed_802_11_packet() 239 lbs_deb_rx("rx data: skb->len-sizeof(RxPd) = %d-%zd = %zd\n", process_rxed_802_11_packet() 270 lbs_deb_rx("rx data: size of actual packet %d\n", skb->len); process_rxed_802_11_packet()
|
/linux-4.4.14/drivers/net/wireless/prism54/ |
H A D | islpci_dev.h | 132 * For rx queues, [index, device_curr_frag) contains fragments 135 * rx queue, waiting for data (owned by device). The driver 138 * If device_curr_frag == driver_curr_frag, no more rx buffers are 139 * available, and the rx DMA engine of the device is halted. 147 u32 index_mgmt_rx; /* real index mgmt rx queue */ 149 u32 free_data_rx; /* free pointer data rx queue */
|
/linux-4.4.14/drivers/net/ethernet/intel/fm10k/ |
H A D | fm10k_debugfs.c | 139 if (ring < q_vector->rx.ring) fm10k_dbg_desc_open() 166 * folder, a debugfs file is created for each tx and rx ring 185 /* Generate a file for each rx ring in the q_vector */ fm10k_dbg_q_vector_init() 196 /* Generate a file for each rx ring in the q_vector */ fm10k_dbg_q_vector_init() 197 for (i = 0; i < q_vector->rx.count; i++) { fm10k_dbg_q_vector_init() 198 struct fm10k_ring *ring = &q_vector->rx.ring[i]; fm10k_dbg_q_vector_init()
|
/linux-4.4.14/drivers/net/ethernet/aurora/ |
H A D | nb8800.h | 182 /* DMA descriptor and associated data for rx. 193 /* Address of buffer on rx ring */ 252 /* Value for rx interrupt time register in NAPI interrupt mode */ 255 /* Value for rx interrupt time register in NAPI poll mode */ 258 /* Value for config field of rx DMA descriptors */ 302 /* DMA base address of rx descriptors, see rx_descs above */
|
/linux-4.4.14/drivers/usb/gadget/function/ |
H A D | f_phonet.c | 49 } rx; member in struct:f_phonet 331 spin_lock_irqsave(&fp->rx.lock, flags); pn_rx_complete() 332 skb = fp->rx.skb; pn_rx_complete() 334 skb = fp->rx.skb = netdev_alloc_skb(dev, 12); pn_rx_complete() 336 fp->rx.skb = NULL; pn_rx_complete() 337 spin_unlock_irqrestore(&fp->rx.lock, flags); pn_rx_complete() 396 if (fp->rx.skb) { __pn_reset() 397 dev_kfree_skb_irq(fp->rx.skb); __pn_reset() 398 fp->rx.skb = NULL; __pn_reset() 693 spin_lock_init(&fp->rx.lock); phonet_alloc()
|
/linux-4.4.14/drivers/atm/ |
H A D | lanai.c | 32 * o AAL0 is stubbed in but the actual rx/tx path isn't written yet: 238 } rx; member in struct:lanai_vcc 257 unsigned atm_ovfl; /* # of cells dropped - rx fifo overflow */ 700 if (lvcc->rx.atmvcc->qos.aal == ATM_AAL5) { host_vcc_start_rx() 701 dma_addr_t dmaaddr = lvcc->rx.buf.dmaaddr; host_vcc_start_rx() 709 RXADDR1_SET_SIZE(lanai_buf_size_cardorder(&lvcc->rx.buf))| host_vcc_start_rx() 1371 int m = ((const unsigned char *) lvcc->rx.buf.ptr) + n - vcc_rx_memcpy() 1372 ((const unsigned char *) (lvcc->rx.buf.end)); vcc_rx_memcpy() 1375 memcpy(dest, lvcc->rx.buf.ptr, n - m); vcc_rx_memcpy() 1376 memcpy(dest + n - m, lvcc->rx.buf.start, m); vcc_rx_memcpy() 1387 u32 *end = &lvcc->rx.buf.start[endptr * 4]; vcc_rx_aal5() 1388 int n = ((unsigned long) end) - ((unsigned long) lvcc->rx.buf.ptr); vcc_rx_aal5() 1390 n += lanai_buf_size(&lvcc->rx.buf); vcc_rx_aal5() 1391 APRINTK(n >= 0 && n < lanai_buf_size(&lvcc->rx.buf) && !(n & 15), vcc_rx_aal5() 1393 n, lanai_buf_size(&lvcc->rx.buf)); vcc_rx_aal5() 1395 if ((x = &end[-2]) < lvcc->rx.buf.start) vcc_rx_aal5() 1396 x = &lvcc->rx.buf.end[-2]; vcc_rx_aal5() 1407 lvcc->rx.atmvcc->dev->number, lvcc->vci, size, n); vcc_rx_aal5() 1411 skb = atm_alloc_charge(lvcc->rx.atmvcc, size, GFP_ATOMIC); vcc_rx_aal5() 1418 ATM_SKB(skb)->vcc = lvcc->rx.atmvcc; vcc_rx_aal5() 1420 lvcc->rx.atmvcc->push(lvcc->rx.atmvcc, skb); vcc_rx_aal5() 1421 atomic_inc(&lvcc->rx.atmvcc->stats->rx); vcc_rx_aal5() 1423 lvcc->rx.buf.ptr = end; vcc_rx_aal5() 1431 /* Remember to increment lvcc->rx.atmvcc->stats->rx */ vcc_rx_aal0() 1506 return lanai_get_sized_buffer(lanai, &lvcc->rx.buf, lanai_setup_rx_vci_aal5() 1650 if (unlikely(lvcc->rx.atmvcc == NULL)) { handle_service() 1657 if (unlikely(lvcc->rx.atmvcc->qos.aal != ATM_AAL5)) { handle_service() 1662 atomic_inc(&lvcc->rx.atmvcc->stats->rx_err); handle_service() 1673 DPRINTK("got trashed rx pdu on vci %d\n", vci); handle_service() 1674 atomic_inc(&lvcc->rx.atmvcc->stats->rx_err); handle_service() 1677 (((unsigned long) lvcc->rx.buf.ptr) - handle_service() 1678 ((unsigned long) lvcc->rx.buf.start)) + 47; handle_service() 1680 bytes += lanai_buf_size(&lvcc->rx.buf); handle_service() 1686 atomic_inc(&lvcc->rx.atmvcc->stats->rx_err); handle_service() 1693 DPRINTK("got rx crc error on vci %d\n", vci); handle_service() 1694 atomic_inc(&lvcc->rx.atmvcc->stats->rx_err); handle_service() 1696 lvcc->rx.buf.ptr = &lvcc->rx.buf.start[SERVICE_GET_END(s) * 4]; handle_service() 2002 lvcc->rx.atmvcc != NULL && lvcc->rx.atmvcc != atmvcc) vci_is_ok() 2014 if (vci0 != NULL && vci0->rx.atmvcc != NULL) vci_is_ok() 2284 if (lvcc->rx.atmvcc == atmvcc) { lanai_close() 2290 lanai_buf_deallocate(&lvcc->rx.buf, lanai->pci); lanai_close() 2291 lvcc->rx.atmvcc = NULL; lanai_close() 2341 APRINTK(lvcc->rx.atmvcc == NULL, "rx.atmvcc!=NULL, vci=%d\n", lanai_open() 2351 lvcc->rx.atmvcc = atmvcc; lanai_open() 2379 if (atmvcc == lvcc->rx.atmvcc) lanai_open() 2511 if (lvcc->rx.atmvcc != NULL) { lanai_proc_read() 2513 lvcc->rx.atmvcc->qos.aal == ATM_AAL5 ? 5 : 0); lanai_proc_read() 2514 if (lvcc->rx.atmvcc->qos.aal == ATM_AAL5) lanai_proc_read() 2518 lanai_buf_size(&lvcc->rx.buf), lanai_proc_read()
|
/linux-4.4.14/drivers/net/ethernet/via/ |
H A D | via-velocity.c | 553 vptr->rx.dirty = vptr->rx.filled = vptr->rx.curr = 0; velocity_init_rx_ring_indexes() 575 vptr->rx.ring[i].rdesc0.len |= OWNED_BY_NIC; velocity_rx_reset() 578 writel(vptr->rx.pool_dma, ®s->RDBaseLo); velocity_rx_reset() 1402 writel(vptr->rx.pool_dma, ®s->RDBaseLo); velocity_init_registers() 1447 if (vptr->rx.filled < 4) velocity_give_many_rx_descs() 1452 unusable = vptr->rx.filled & 0x0003; velocity_give_many_rx_descs() 1453 dirty = vptr->rx.dirty - unusable; velocity_give_many_rx_descs() 1454 for (avail = vptr->rx.filled & 0xfffc; avail; avail--) { velocity_give_many_rx_descs() 1456 vptr->rx.ring[dirty].rdesc0.len |= OWNED_BY_NIC; velocity_give_many_rx_descs() 1459 writew(vptr->rx.filled & 0xfffc, ®s->RBRDU); velocity_give_many_rx_descs() 1460 vptr->rx.filled = unusable; velocity_give_many_rx_descs() 1493 vptr->rx.ring = pool; velocity_init_dma_rings() 1494 vptr->rx.pool_dma = pool_dma; velocity_init_dma_rings() 1511 vptr->rx.buf_sz = (mtu <= ETH_DATA_LEN) ? PKT_BUF_SZ : mtu + 32; velocity_set_rxbufsize() 1526 struct rx_desc *rd = &(vptr->rx.ring[idx]); velocity_alloc_rx_buf() 1527 struct velocity_rd_info *rd_info = &(vptr->rx.info[idx]); velocity_alloc_rx_buf() 1529 rd_info->skb = netdev_alloc_skb(vptr->netdev, vptr->rx.buf_sz + 64); velocity_alloc_rx_buf() 1540 vptr->rx.buf_sz, DMA_FROM_DEVICE); velocity_alloc_rx_buf() 1547 rd->size = cpu_to_le16(vptr->rx.buf_sz) | RX_INTEN; velocity_alloc_rx_buf() 1556 int dirty = vptr->rx.dirty, done = 0; velocity_rx_refill() 1559 struct rx_desc *rd = vptr->rx.ring + dirty; velocity_rx_refill() 1565 if (!vptr->rx.info[dirty].skb) { velocity_rx_refill() 1571 } while (dirty != vptr->rx.curr); velocity_rx_refill() 1574 vptr->rx.dirty = dirty; velocity_rx_refill() 1575 vptr->rx.filled += done; velocity_rx_refill() 1592 if (vptr->rx.info == NULL) velocity_free_rd_ring() 1596 struct velocity_rd_info *rd_info = &(vptr->rx.info[i]); velocity_free_rd_ring() 1597 struct rx_desc *rd = vptr->rx.ring + i; velocity_free_rd_ring() 1603 dma_unmap_single(vptr->dev, rd_info->skb_dma, vptr->rx.buf_sz, velocity_free_rd_ring() 1611 kfree(vptr->rx.info); velocity_free_rd_ring() 1612 vptr->rx.info = NULL; velocity_free_rd_ring() 1626 vptr->rx.info = kcalloc(vptr->options.numrx, velocity_init_rd_ring() 1628 if (!vptr->rx.info) velocity_init_rd_ring() 1685 dma_free_coherent(vptr->dev, size, vptr->rx.ring, vptr->rx.pool_dma); velocity_free_dma_rings() 2045 struct velocity_rd_info *rd_info = &(vptr->rx.info[idx]); velocity_receive_frame() 2046 struct rx_desc *rd = &(vptr->rx.ring[idx]); velocity_receive_frame() 2063 vptr->rx.buf_sz, DMA_FROM_DEVICE); velocity_receive_frame() 2070 dma_unmap_single(vptr->dev, rd_info->skb_dma, vptr->rx.buf_sz, velocity_receive_frame() 2074 vptr->rx.buf_sz, DMA_FROM_DEVICE); velocity_receive_frame() 2104 int rd_curr = vptr->rx.curr; velocity_rx_srv() 2108 struct rx_desc *rd = vptr->rx.ring + rd_curr; velocity_rx_srv() 2110 if (!vptr->rx.info[rd_curr].skb) velocity_rx_srv() 2141 vptr->rx.curr = rd_curr; velocity_rx_srv() 2158 * Do rx and tx twice for performance (taken from the VIA velocity_poll() 2305 struct rx_info rx; velocity_change_mtu() local 2331 rx = vptr->rx; velocity_change_mtu() 2334 vptr->rx = tmp_vptr->rx; velocity_change_mtu() 2337 tmp_vptr->rx = rx; velocity_change_mtu()
|
/linux-4.4.14/drivers/soc/qcom/ |
H A D | smd.c | 165 * @recv_lock: guard for rx info modifications and cb pointer 230 struct smd_channel_info rx; member in struct:smd_channel_info_pair 252 struct smd_channel_info_word rx; member in struct:smd_channel_info_word_pair 257 BUILD_BUG_ON(sizeof(channel->info->rx.param) != sizeof(u8)); \ 259 le32_to_cpu(channel->info_word->rx.param) : \ 260 channel->info->rx.param; \ 265 BUILD_BUG_ON(sizeof(channel->info->rx.param) != sizeof(u32)); \ 267 channel->info_word->rx.param : \ 268 channel->info->rx.param); \ 273 BUILD_BUG_ON(sizeof(channel->info->rx.param) != sizeof(u8)); \ 275 channel->info_word->rx.param = cpu_to_le32(value); \ 277 channel->info->rx.param = value; \ 282 BUILD_BUG_ON(sizeof(channel->info->rx.param) != sizeof(u32)); \ 284 channel->info_word->rx.param = cpu_to_le32(value); \ 286 channel->info->rx.param = cpu_to_le32(value); \ 381 * Calculate the amount of data available in the rx fifo 455 * Read count bytes of data from the rx fifo into buf, but don't advance the 487 * Advance the rx tail by count bytes. 501 * Read out a single packet from the rx fifo and deliver it to the device 641 * Delivers any outstanding packets in the rx fifo, can be used after probe of 1071 /* The channel consist of a rx and tx fifo of equal size */ qcom_smd_create_channel()
|
/linux-4.4.14/sound/soc/fsl/ |
H A D | fsl_ssi.c | 111 struct fsl_ssi_reg_val rx; member in struct:fsl_ssi_rxtx_reg_val 369 * Enable/Disable all rx/tx config flags at once. 379 vals->rx.sier | vals->tx.sier, fsl_ssi_rxtx_config() 380 vals->rx.sier | vals->tx.sier); fsl_ssi_rxtx_config() 382 vals->rx.srcr | vals->tx.srcr, fsl_ssi_rxtx_config() 383 vals->rx.srcr | vals->tx.srcr); fsl_ssi_rxtx_config() 385 vals->rx.stcr | vals->tx.stcr, fsl_ssi_rxtx_config() 386 vals->rx.stcr | vals->tx.stcr); fsl_ssi_rxtx_config() 389 vals->rx.srcr | vals->tx.srcr, 0); fsl_ssi_rxtx_config() 391 vals->rx.stcr | vals->tx.stcr, 0); fsl_ssi_rxtx_config() 393 vals->rx.sier | vals->tx.sier, 0); fsl_ssi_rxtx_config() 419 * ssi_private->rxtx_reg_val.rx or tx as vals parameter. 440 /* Find the other direction values rx or tx which we do not want to fsl_ssi_config() 442 if (&ssi_private->rxtx_reg_val.rx == vals) fsl_ssi_config() 445 avals = &ssi_private->rxtx_reg_val.rx; fsl_ssi_config() 481 * Disabling the necessary flags for one of rx/tx while the fsl_ssi_config() 484 * streams (rx XOR tx) and that are set in the stream that is fsl_ssi_config() 511 fsl_ssi_config(ssi_private, enable, &ssi_private->rxtx_reg_val.rx); fsl_ssi_rx_config() 520 * Setup rx/tx register values used to enable/disable the streams. These will 528 reg->rx.sier = CCSR_SSI_SIER_RFF0_EN; fsl_ssi_setup_reg_vals() 529 reg->rx.srcr = CCSR_SSI_SRCR_RFEN0; fsl_ssi_setup_reg_vals() 530 reg->rx.scr = 0; fsl_ssi_setup_reg_vals() 536 reg->rx.scr = CCSR_SSI_SCR_SSIEN | CCSR_SSI_SCR_RE; fsl_ssi_setup_reg_vals() 537 reg->rx.sier |= CCSR_SSI_SIER_RFF0_EN; fsl_ssi_setup_reg_vals() 543 reg->rx.sier |= CCSR_SSI_SIER_RDMAE; fsl_ssi_setup_reg_vals() 546 reg->rx.sier |= CCSR_SSI_SIER_RIE; fsl_ssi_setup_reg_vals() 550 reg->rx.sier |= FSLSSI_SIER_DBG_RX_FLAGS; fsl_ssi_setup_reg_vals()
|
/linux-4.4.14/drivers/net/vmxnet3/ |
H A D | vmxnet3_ethtool.c | 81 { " LRO pkts rx", offsetof(struct UPT1_RxStats, LROPktsRxOK) }, 82 { " LRO byte rx", offsetof(struct UPT1_RxStats, LROBytesRxOK) }, 83 { " ucast pkts rx", offsetof(struct UPT1_RxStats, ucastPktsRxOK) }, 84 { " ucast bytes rx", offsetof(struct UPT1_RxStats, ucastBytesRxOK) }, 85 { " mcast pkts rx", offsetof(struct UPT1_RxStats, mcastPktsRxOK) }, 86 { " mcast bytes rx", offsetof(struct UPT1_RxStats, mcastBytesRxOK) }, 87 { " bcast pkts rx", offsetof(struct UPT1_RxStats, bcastPktsRxOK) }, 88 { " bcast bytes rx", offsetof(struct UPT1_RxStats, bcastBytesRxOK) }, 89 { " pkts rx OOB", offsetof(struct UPT1_RxStats, pktsRxOutOfBuf) }, 90 { " pkts rx err", offsetof(struct UPT1_RxStats, pktsRxError) }, 97 { " drv dropped rx total", offsetof(struct vmxnet3_rq_driver_stats, 103 { " rx buf alloc fail", offsetof(struct vmxnet3_rq_driver_stats, 588 /* recreate the rx queue and the tx queue based on the vmxnet3_set_ringparam()
|
/linux-4.4.14/drivers/net/ethernet/hisilicon/ |
H A D | hip04_eth.c | 301 /* enable tx & rx */ hip04_mac_enable() 306 /* clear rx int */ hip04_mac_enable() 328 /* disable tx & rx */ hip04_mac_disable() 469 /* disable rx interrupt and timer */ hip04_mac_start_xmit() 495 int rx = 0; hip04_rx_poll() local 528 rx++; hip04_rx_poll() 543 if (rx >= budget) hip04_rx_poll() 551 /* enable rx interrupt */ hip04_rx_poll() 559 if (rx < budget && tx_remaining) hip04_rx_poll() 562 return rx; hip04_rx_poll() 581 netdev_err(ndev, "rx drop\n"); hip04_mac_interrupt() 590 /* disable rx interrupt */ hip04_mac_interrupt() 607 /* disable rx interrupt */ tx_done()
|
/linux-4.4.14/arch/powerpc/include/asm/ |
H A D | mmu-hash64.h | 451 * rx = scratch register (clobbered) 453 * - rt and rx must be different registers 458 #define ASM_VSID_SCRAMBLE(rt, rx, size) \ 459 lis rx,VSID_MULTIPLIER_##size@h; \ 460 ori rx,rx,VSID_MULTIPLIER_##size@l; \ 461 mulld rt,rt,rx; /* rt = rt * MULTIPLIER */ \ 463 srdi rx,rt,VSID_BITS_##size; \ 465 add rt,rt,rx; /* add high and low bits */ \ 473 addi rx,rt,1; \ 474 srdi rx,rx,VSID_BITS_##size; /* extract 2^VSID_BITS bit */ \ 475 add rt,rt,rx
|
/linux-4.4.14/drivers/media/rc/ |
H A D | fintek-cir.h | 66 /* for rx */ 97 /* rx settings */ 185 #define CIR_CR_CLASS 0x07 /* 0xff = rx-only, 0x66 = rx + 2 tx, 186 0x33 = rx + 1 tx */
|
/linux-4.4.14/drivers/net/irda/ |
H A D | vlsi_ir.h | 177 VLSI_PIO_RINGPTR = 0x02, /* rx/tx ring pointer (u16, ro) */ 179 VLSI_PIO_RINGSIZE = 0x06, /* rx/tx ring size (u16, rw) */ 210 IRINTR_RPKTINT = 0x10, /* rx-packet transferred from fifo to memory finished */ 213 IRINTR_OE_EN = 0x02, /* UART rx fifo overrun error interrupt enable */ 214 IRINTR_OE_INT = 0x01 /* UART rx fifo overrun error (read LSR to clear) */ 225 /* _both_ ring pointers are indices relative to the _entire_ rx,tx-ring! 227 * at RINGBASE + PTR * sizeof(descr) for rx and tx 231 #define MAX_RING_DESCR 64 /* tx, rx rings may contain up to 64 descr each */ 251 * The entire ring area is split in rx and tx areas with each 253 * The rx(tx) ring is located at ringbase+0 (ringbase+64*8). 262 /* bit mask to indicate the ring size to be used for rx and tx. 269 * located at [15:12] for tx and [11:8] for rx ([7:0] unused) 276 #define TX_RX_TO_RINGSIZE(tx,rx) ((SIZE_TO_BITS(tx)<<12)|(SIZE_TO_BITS(rx)<<8)) 286 * for both tx, rx rings as follows: 327 IRCFG_RXPOL = 0x0001 /* invert rx polarity when set */ 522 /* descriptors for rx/tx ring 538 volatile __le16 rd_count; /* tx/rx count [11:0] */ 657 /* sw descriptor rings for rx, tx: 739 * post-processing in vlsi_process_tx/rx() after it was completed
|
/linux-4.4.14/drivers/net/ethernet/freescale/fs_enet/ |
H A D | fs_enet.h | 156 u32 ev_napi_rx; /* mask of NAPI rx events */ 157 u32 ev_napi_tx; /* mask of NAPI rx events */ 158 u32 ev_rx; /* rx event mask */ 162 u16 bd_rx_empty; /* mask of BD rx empty */ 163 u16 bd_rx_err; /* mask of BD rx errors */
|
/linux-4.4.14/arch/s390/kvm/ |
H A D | diag.c | 75 u16 rx = (vcpu->arch.sie_block->ipa & 0xf0) >> 4; __diag_page_ref_service() local 79 vcpu->run->s.regs.gprs[rx]); __diag_page_ref_service() 81 if (vcpu->run->s.regs.gprs[rx] & 7) __diag_page_ref_service() 83 rc = read_guest(vcpu, vcpu->run->s.regs.gprs[rx], rx, &parm, sizeof(parm)); __diag_page_ref_service()
|
/linux-4.4.14/net/l2tp/ |
H A D | l2tp_debugfs.c | 150 seq_printf(m, " %08x rx %ld/%ld/%ld rx %ld/%ld/%ld\n", l2tp_dfs_seq_tunnel_show() 207 seq_printf(m, " %hu/%hu tx %ld/%ld/%ld rx %ld/%ld/%ld\n", l2tp_dfs_seq_session_show() 229 seq_puts(m, " debug tx-pkts/bytes/errs rx-pkts/bytes/errs\n"); l2tp_dfs_seq_show() 236 seq_puts(m, " nr/ns tx-pkts/bytes/errs rx-pkts/bytes/errs\n"); l2tp_dfs_seq_show()
|
/linux-4.4.14/sound/soc/au1x/ |
H A D | ac97c.c | 33 #define CFG_RC(x) (((x) & 0x3ff) << 13) /* valid rx slots mask */ 40 #define STAT_RU (1 << 9) /* rx underflow */ 41 #define STAT_RO (1 << 8) /* rx overflow */ 46 #define STAT_RE (1 << 1) /* rx fifo empty */ 47 #define STAT_RF (1 << 0) /* rx fifo full */
|
/linux-4.4.14/drivers/usb/gadget/udc/ |
H A D | omap_udc.h | 77 /* rx/tx dma channels numbered 1-3 not 0-2 */ 104 #define UDC_DATA_DMA (UDC_BASE + 0x48) /* rx/tx fifo addr */ 106 /* rx/tx dma control, numbering channels 1-3 not 0-2 */
|