Lines Matching refs:cqe

354 					 struct eth_end_agg_rx_cqe *cqe)  in bnx2x_update_sge_prod()  argument
367 RX_SGE(le16_to_cpu(cqe->sgl_or_raw_data.sgl[i]))); in bnx2x_update_sge_prod()
370 sge_len - 1, le16_to_cpu(cqe->sgl_or_raw_data.sgl[sge_len - 1])); in bnx2x_update_sge_prod()
375 le16_to_cpu(cqe->sgl_or_raw_data.sgl[sge_len - 1])); in bnx2x_update_sge_prod()
409 const struct eth_fast_path_rx_cqe *cqe, in bnx2x_get_rxhash() argument
414 (cqe->status_flags & ETH_FAST_PATH_RX_CQE_RSS_HASH_FLG)) { in bnx2x_get_rxhash()
417 htype = cqe->status_flags & ETH_FAST_PATH_RX_CQE_RSS_HASH_TYPE; in bnx2x_get_rxhash()
422 return le32_to_cpu(cqe->rss_hash_result); in bnx2x_get_rxhash()
430 struct eth_fast_path_rx_cqe *cqe) in bnx2x_tpa_start() argument
473 le16_to_cpu(cqe->pars_flags.flags); in bnx2x_tpa_start()
474 tpa_info->vlan_tag = le16_to_cpu(cqe->vlan_tag); in bnx2x_tpa_start()
476 tpa_info->len_on_bd = le16_to_cpu(cqe->len_on_bd); in bnx2x_tpa_start()
477 tpa_info->placement_offset = cqe->placement_offset; in bnx2x_tpa_start()
478 tpa_info->rxhash = bnx2x_get_rxhash(bp, cqe, &tpa_info->rxhash_type); in bnx2x_tpa_start()
480 u16 gro_size = le16_to_cpu(cqe->pkt_len_or_gro_seg_len); in bnx2x_tpa_start()
578 struct eth_end_agg_rx_cqe *cqe, in bnx2x_fill_frag_skb() argument
587 frag_size = le16_to_cpu(cqe->pkt_len) - len_on_bd; in bnx2x_fill_frag_skb()
597 le16_to_cpu(cqe->pkt_len), in bnx2x_fill_frag_skb()
598 le16_to_cpu(cqe->num_of_coalesced_segs)); in bnx2x_fill_frag_skb()
604 BNX2X_ERR("cqe->pkt_len = %d\n", cqe->pkt_len); in bnx2x_fill_frag_skb()
612 u16 sge_idx = RX_SGE(le16_to_cpu(cqe->sgl_or_raw_data.sgl[j])); in bnx2x_fill_frag_skb()
742 struct eth_end_agg_rx_cqe *cqe, in bnx2x_tpa_stop() argument
788 skb, cqe, cqe_idx)) { in bnx2x_tpa_stop()
843 void bnx2x_csum_validate(struct sk_buff *skb, union eth_rx_cqe *cqe, in bnx2x_csum_validate() argument
852 if (cqe->fast_path_cqe.status_flags & in bnx2x_csum_validate()
858 if (cqe->fast_path_cqe.type_error_flags & in bnx2x_csum_validate()
872 union eth_rx_cqe *cqe; in bnx2x_rx_int() local
889 cqe = &fp->rx_comp_ring[comp_ring_cons]; in bnx2x_rx_int()
890 cqe_fp = &cqe->fast_path_cqe; in bnx2x_rx_int()
938 bnx2x_sp_event(fp, cqe); in bnx2x_rx_int()
969 queue = cqe->end_agg_cqe.queue_index; in bnx2x_rx_int()
975 frag_size = le16_to_cpu(cqe->end_agg_cqe.pkt_len) - in bnx2x_rx_int()
986 &cqe->end_agg_cqe, comp_ring_cons); in bnx2x_rx_int()
992 bnx2x_update_sge_prod(fp, pages, &cqe->end_agg_cqe); in bnx2x_rx_int()
1062 bnx2x_csum_validate(skb, cqe, fp, in bnx2x_rx_int()
1068 if (unlikely(cqe->fast_path_cqe.type_error_flags & in bnx2x_rx_int()
1101 cqe = &fp->rx_comp_ring[comp_ring_cons]; in bnx2x_rx_int()
1102 cqe_fp = &cqe->fast_path_cqe; in bnx2x_rx_int()