root/drivers/net/ethernet/brocade/bna/bna_tx_rx.c

/* [<][>][^][v][top][bottom][index][help] */

DEFINITIONS

This source file includes following definitions.
  1. bna_ib_coalescing_timeo_set
  2. bna_rxf_sm_stopped_entry
  3. bna_rxf_sm_stopped
  4. bna_rxf_sm_cfg_wait_entry
  5. bna_rxf_sm_cfg_wait
  6. bna_rxf_sm_started_entry
  7. bna_rxf_sm_started
  8. bna_rxf_sm_last_resp_wait_entry
  9. bna_rxf_sm_last_resp_wait
  10. bna_bfi_ucast_req
  11. bna_bfi_mcast_add_req
  12. bna_bfi_mcast_del_req
  13. bna_bfi_mcast_filter_req
  14. bna_bfi_rx_promisc_req
  15. bna_bfi_rx_vlan_filter_set
  16. bna_bfi_vlan_strip_enable
  17. bna_bfi_rit_cfg
  18. bna_bfi_rss_cfg
  19. bna_bfi_rss_enable
  20. bna_rxf_mcmac_get
  21. bna_rxf_mchandle_get
  22. bna_rxf_mchandle_attach
  23. bna_rxf_mcast_del
  24. bna_rxf_mcast_cfg_apply
  25. bna_rxf_vlan_cfg_apply
  26. bna_rxf_mcast_cfg_reset
  27. bna_rxf_rss_cfg_apply
  28. bna_rxf_cfg_apply
  29. bna_rxf_cfg_reset
  30. bna_rit_init
  31. bna_bfi_rxf_cfg_rsp
  32. bna_bfi_rxf_ucast_set_rsp
  33. bna_bfi_rxf_mcast_add_rsp
  34. bna_rxf_init
  35. bna_rxf_uninit
  36. bna_rx_cb_rxf_started
  37. bna_rxf_start
  38. bna_rx_cb_rxf_stopped
  39. bna_rxf_stop
  40. bna_rxf_fail
  41. bna_rx_ucast_set
  42. bna_rx_mcast_add
  43. bna_rx_ucast_listset
  44. bna_rx_mcast_listset
  45. bna_rx_mcast_delall
  46. bna_rx_vlan_add
  47. bna_rx_vlan_del
  48. bna_rxf_ucast_cfg_apply
  49. bna_rxf_ucast_cfg_reset
  50. bna_rxf_promisc_cfg_apply
  51. bna_rxf_promisc_cfg_reset
  52. bna_rxf_allmulti_cfg_apply
  53. bna_rxf_allmulti_cfg_reset
  54. bna_rxf_promisc_enable
  55. bna_rxf_promisc_disable
  56. bna_rxf_allmulti_enable
  57. bna_rxf_allmulti_disable
  58. bna_rxf_vlan_strip_cfg_apply
  59. bna_rx_sm_stopped_entry
  60. bna_rx_sm_stopped
  61. bna_rx_sm_start_wait_entry
  62. bna_rx_sm_stop_wait_entry
  63. bna_rx_sm_stop_wait
  64. bna_rx_sm_start_wait
  65. bna_rx_sm_rxf_start_wait_entry
  66. bna_rx_sm_rxf_stop_wait_entry
  67. bna_rx_sm_rxf_stop_wait
  68. bna_rx_sm_start_stop_wait_entry
  69. bna_rx_sm_start_stop_wait
  70. bna_rx_sm_started_entry
  71. bna_rx_sm_started
  72. bna_rx_sm_rxf_start_wait
  73. bna_rx_sm_cleanup_wait_entry
  74. bna_rx_sm_cleanup_wait
  75. bna_rx_sm_failed_entry
  76. bna_rx_sm_failed
  77. bna_rx_sm_quiesce_wait_entry
  78. bna_rx_sm_quiesce_wait
  79. bna_bfi_rx_enet_start
  80. bna_bfi_rx_enet_stop
  81. bna_rx_enet_stop
  82. bna_rx_res_check
  83. bna_rxq_get
  84. bna_rxq_put
  85. bna_rxp_get
  86. bna_rxp_put
  87. bna_rx_get
  88. bna_rx_put
  89. bna_rxp_add_rxqs
  90. bna_rxq_qpt_setup
  91. bna_rxp_cqpt_setup
  92. bna_rx_mod_cb_rx_stopped
  93. bna_rx_mod_cb_rx_stopped_all
  94. bna_rx_start
  95. bna_rx_stop
  96. bna_rx_fail
  97. bna_rx_mod_start
  98. bna_rx_mod_stop
  99. bna_rx_mod_fail
  100. bna_rx_mod_init
  101. bna_rx_mod_uninit
  102. bna_bfi_rx_enet_start_rsp
  103. bna_bfi_rx_enet_stop_rsp
  104. bna_rx_res_req
  105. bna_rx_create
  106. bna_rx_destroy
  107. bna_rx_enable
  108. bna_rx_disable
  109. bna_rx_cleanup_complete
  110. bna_rx_vlan_strip_enable
  111. bna_rx_vlan_strip_disable
  112. bna_rx_mode_set
  113. bna_rx_vlanfilter_enable
  114. bna_rx_coalescing_timeo_set
  115. bna_rx_dim_reconfig
  116. bna_rx_dim_update
  117. bna_tx_sm_stopped_entry
  118. bna_tx_sm_stopped
  119. bna_tx_sm_start_wait_entry
  120. bna_tx_sm_start_wait
  121. bna_tx_sm_started_entry
  122. bna_tx_sm_started
  123. bna_tx_sm_stop_wait_entry
  124. bna_tx_sm_stop_wait
  125. bna_tx_sm_cleanup_wait_entry
  126. bna_tx_sm_cleanup_wait
  127. bna_tx_sm_prio_stop_wait_entry
  128. bna_tx_sm_prio_stop_wait
  129. bna_tx_sm_prio_cleanup_wait_entry
  130. bna_tx_sm_prio_cleanup_wait
  131. bna_tx_sm_failed_entry
  132. bna_tx_sm_failed
  133. bna_tx_sm_quiesce_wait_entry
  134. bna_tx_sm_quiesce_wait
  135. bna_bfi_tx_enet_start
  136. bna_bfi_tx_enet_stop
  137. bna_tx_enet_stop
  138. bna_txq_qpt_setup
  139. bna_tx_get
  140. bna_tx_free
  141. bna_tx_start
  142. bna_tx_stop
  143. bna_tx_fail
  144. bna_bfi_tx_enet_start_rsp
  145. bna_bfi_tx_enet_stop_rsp
  146. bna_bfi_bw_update_aen
  147. bna_tx_res_req
  148. bna_tx_create
  149. bna_tx_destroy
  150. bna_tx_enable
  151. bna_tx_disable
  152. bna_tx_cleanup_complete
  153. bna_tx_mod_cb_tx_stopped
  154. bna_tx_mod_cb_tx_stopped_all
  155. bna_tx_mod_init
  156. bna_tx_mod_uninit
  157. bna_tx_mod_start
  158. bna_tx_mod_stop
  159. bna_tx_mod_fail
  160. bna_tx_coalescing_timeo_set

   1 // SPDX-License-Identifier: GPL-2.0-only
   2 /*
   3  * Linux network driver for QLogic BR-series Converged Network Adapter.
   4   */
   5 /*
   6  * Copyright (c) 2005-2014 Brocade Communications Systems, Inc.
   7  * Copyright (c) 2014-2015 QLogic Corporation
   8  * All rights reserved
   9  * www.qlogic.com
  10  */
  11 #include "bna.h"
  12 #include "bfi.h"
  13 
  14 /* IB */
  15 static void
  16 bna_ib_coalescing_timeo_set(struct bna_ib *ib, u8 coalescing_timeo)
  17 {
  18         ib->coalescing_timeo = coalescing_timeo;
  19         ib->door_bell.doorbell_ack = BNA_DOORBELL_IB_INT_ACK(
  20                                 (u32)ib->coalescing_timeo, 0);
  21 }
  22 
  23 /* RXF */
  24 
  25 #define bna_rxf_vlan_cfg_soft_reset(rxf)                                \
  26 do {                                                                    \
  27         (rxf)->vlan_pending_bitmask = (u8)BFI_VLAN_BMASK_ALL;           \
  28         (rxf)->vlan_strip_pending = true;                               \
  29 } while (0)
  30 
  31 #define bna_rxf_rss_cfg_soft_reset(rxf)                                 \
  32 do {                                                                    \
  33         if ((rxf)->rss_status == BNA_STATUS_T_ENABLED)                  \
  34                 (rxf)->rss_pending = (BNA_RSS_F_RIT_PENDING |           \
  35                                 BNA_RSS_F_CFG_PENDING |                 \
  36                                 BNA_RSS_F_STATUS_PENDING);              \
  37 } while (0)
  38 
  39 static int bna_rxf_cfg_apply(struct bna_rxf *rxf);
  40 static void bna_rxf_cfg_reset(struct bna_rxf *rxf);
  41 static int bna_rxf_ucast_cfg_apply(struct bna_rxf *rxf);
  42 static int bna_rxf_promisc_cfg_apply(struct bna_rxf *rxf);
  43 static int bna_rxf_allmulti_cfg_apply(struct bna_rxf *rxf);
  44 static int bna_rxf_vlan_strip_cfg_apply(struct bna_rxf *rxf);
  45 static int bna_rxf_ucast_cfg_reset(struct bna_rxf *rxf,
  46                                         enum bna_cleanup_type cleanup);
  47 static int bna_rxf_promisc_cfg_reset(struct bna_rxf *rxf,
  48                                         enum bna_cleanup_type cleanup);
  49 static int bna_rxf_allmulti_cfg_reset(struct bna_rxf *rxf,
  50                                         enum bna_cleanup_type cleanup);
  51 
  52 bfa_fsm_state_decl(bna_rxf, stopped, struct bna_rxf,
  53                         enum bna_rxf_event);
  54 bfa_fsm_state_decl(bna_rxf, cfg_wait, struct bna_rxf,
  55                         enum bna_rxf_event);
  56 bfa_fsm_state_decl(bna_rxf, started, struct bna_rxf,
  57                         enum bna_rxf_event);
  58 bfa_fsm_state_decl(bna_rxf, last_resp_wait, struct bna_rxf,
  59                         enum bna_rxf_event);
  60 
  61 static void
  62 bna_rxf_sm_stopped_entry(struct bna_rxf *rxf)
  63 {
  64         call_rxf_stop_cbfn(rxf);
  65 }
  66 
  67 static void
  68 bna_rxf_sm_stopped(struct bna_rxf *rxf, enum bna_rxf_event event)
  69 {
  70         switch (event) {
  71         case RXF_E_START:
  72                 bfa_fsm_set_state(rxf, bna_rxf_sm_cfg_wait);
  73                 break;
  74 
  75         case RXF_E_STOP:
  76                 call_rxf_stop_cbfn(rxf);
  77                 break;
  78 
  79         case RXF_E_FAIL:
  80                 /* No-op */
  81                 break;
  82 
  83         case RXF_E_CONFIG:
  84                 call_rxf_cam_fltr_cbfn(rxf);
  85                 break;
  86 
  87         default:
  88                 bfa_sm_fault(event);
  89         }
  90 }
  91 
  92 static void
  93 bna_rxf_sm_cfg_wait_entry(struct bna_rxf *rxf)
  94 {
  95         if (!bna_rxf_cfg_apply(rxf)) {
  96                 /* No more pending config updates */
  97                 bfa_fsm_set_state(rxf, bna_rxf_sm_started);
  98         }
  99 }
 100 
 101 static void
 102 bna_rxf_sm_cfg_wait(struct bna_rxf *rxf, enum bna_rxf_event event)
 103 {
 104         switch (event) {
 105         case RXF_E_STOP:
 106                 bfa_fsm_set_state(rxf, bna_rxf_sm_last_resp_wait);
 107                 break;
 108 
 109         case RXF_E_FAIL:
 110                 bna_rxf_cfg_reset(rxf);
 111                 call_rxf_start_cbfn(rxf);
 112                 call_rxf_cam_fltr_cbfn(rxf);
 113                 bfa_fsm_set_state(rxf, bna_rxf_sm_stopped);
 114                 break;
 115 
 116         case RXF_E_CONFIG:
 117                 /* No-op */
 118                 break;
 119 
 120         case RXF_E_FW_RESP:
 121                 if (!bna_rxf_cfg_apply(rxf)) {
 122                         /* No more pending config updates */
 123                         bfa_fsm_set_state(rxf, bna_rxf_sm_started);
 124                 }
 125                 break;
 126 
 127         default:
 128                 bfa_sm_fault(event);
 129         }
 130 }
 131 
 132 static void
 133 bna_rxf_sm_started_entry(struct bna_rxf *rxf)
 134 {
 135         call_rxf_start_cbfn(rxf);
 136         call_rxf_cam_fltr_cbfn(rxf);
 137 }
 138 
 139 static void
 140 bna_rxf_sm_started(struct bna_rxf *rxf, enum bna_rxf_event event)
 141 {
 142         switch (event) {
 143         case RXF_E_STOP:
 144         case RXF_E_FAIL:
 145                 bna_rxf_cfg_reset(rxf);
 146                 bfa_fsm_set_state(rxf, bna_rxf_sm_stopped);
 147                 break;
 148 
 149         case RXF_E_CONFIG:
 150                 bfa_fsm_set_state(rxf, bna_rxf_sm_cfg_wait);
 151                 break;
 152 
 153         default:
 154                 bfa_sm_fault(event);
 155         }
 156 }
 157 
 158 static void
 159 bna_rxf_sm_last_resp_wait_entry(struct bna_rxf *rxf)
 160 {
 161 }
 162 
 163 static void
 164 bna_rxf_sm_last_resp_wait(struct bna_rxf *rxf, enum bna_rxf_event event)
 165 {
 166         switch (event) {
 167         case RXF_E_FAIL:
 168         case RXF_E_FW_RESP:
 169                 bna_rxf_cfg_reset(rxf);
 170                 bfa_fsm_set_state(rxf, bna_rxf_sm_stopped);
 171                 break;
 172 
 173         default:
 174                 bfa_sm_fault(event);
 175         }
 176 }
 177 
 178 static void
 179 bna_bfi_ucast_req(struct bna_rxf *rxf, struct bna_mac *mac,
 180                 enum bfi_enet_h2i_msgs req_type)
 181 {
 182         struct bfi_enet_ucast_req *req = &rxf->bfi_enet_cmd.ucast_req;
 183 
 184         bfi_msgq_mhdr_set(req->mh, BFI_MC_ENET, req_type, 0, rxf->rx->rid);
 185         req->mh.num_entries = htons(
 186         bfi_msgq_num_cmd_entries(sizeof(struct bfi_enet_ucast_req)));
 187         ether_addr_copy(req->mac_addr, mac->addr);
 188         bfa_msgq_cmd_set(&rxf->msgq_cmd, NULL, NULL,
 189                 sizeof(struct bfi_enet_ucast_req), &req->mh);
 190         bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd);
 191 }
 192 
 193 static void
 194 bna_bfi_mcast_add_req(struct bna_rxf *rxf, struct bna_mac *mac)
 195 {
 196         struct bfi_enet_mcast_add_req *req =
 197                 &rxf->bfi_enet_cmd.mcast_add_req;
 198 
 199         bfi_msgq_mhdr_set(req->mh, BFI_MC_ENET, BFI_ENET_H2I_MAC_MCAST_ADD_REQ,
 200                 0, rxf->rx->rid);
 201         req->mh.num_entries = htons(
 202         bfi_msgq_num_cmd_entries(sizeof(struct bfi_enet_mcast_add_req)));
 203         ether_addr_copy(req->mac_addr, mac->addr);
 204         bfa_msgq_cmd_set(&rxf->msgq_cmd, NULL, NULL,
 205                 sizeof(struct bfi_enet_mcast_add_req), &req->mh);
 206         bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd);
 207 }
 208 
 209 static void
 210 bna_bfi_mcast_del_req(struct bna_rxf *rxf, u16 handle)
 211 {
 212         struct bfi_enet_mcast_del_req *req =
 213                 &rxf->bfi_enet_cmd.mcast_del_req;
 214 
 215         bfi_msgq_mhdr_set(req->mh, BFI_MC_ENET, BFI_ENET_H2I_MAC_MCAST_DEL_REQ,
 216                 0, rxf->rx->rid);
 217         req->mh.num_entries = htons(
 218         bfi_msgq_num_cmd_entries(sizeof(struct bfi_enet_mcast_del_req)));
 219         req->handle = htons(handle);
 220         bfa_msgq_cmd_set(&rxf->msgq_cmd, NULL, NULL,
 221                 sizeof(struct bfi_enet_mcast_del_req), &req->mh);
 222         bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd);
 223 }
 224 
 225 static void
 226 bna_bfi_mcast_filter_req(struct bna_rxf *rxf, enum bna_status status)
 227 {
 228         struct bfi_enet_enable_req *req = &rxf->bfi_enet_cmd.req;
 229 
 230         bfi_msgq_mhdr_set(req->mh, BFI_MC_ENET,
 231                 BFI_ENET_H2I_MAC_MCAST_FILTER_REQ, 0, rxf->rx->rid);
 232         req->mh.num_entries = htons(
 233                 bfi_msgq_num_cmd_entries(sizeof(struct bfi_enet_enable_req)));
 234         req->enable = status;
 235         bfa_msgq_cmd_set(&rxf->msgq_cmd, NULL, NULL,
 236                 sizeof(struct bfi_enet_enable_req), &req->mh);
 237         bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd);
 238 }
 239 
 240 static void
 241 bna_bfi_rx_promisc_req(struct bna_rxf *rxf, enum bna_status status)
 242 {
 243         struct bfi_enet_enable_req *req = &rxf->bfi_enet_cmd.req;
 244 
 245         bfi_msgq_mhdr_set(req->mh, BFI_MC_ENET,
 246                 BFI_ENET_H2I_RX_PROMISCUOUS_REQ, 0, rxf->rx->rid);
 247         req->mh.num_entries = htons(
 248                 bfi_msgq_num_cmd_entries(sizeof(struct bfi_enet_enable_req)));
 249         req->enable = status;
 250         bfa_msgq_cmd_set(&rxf->msgq_cmd, NULL, NULL,
 251                 sizeof(struct bfi_enet_enable_req), &req->mh);
 252         bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd);
 253 }
 254 
 255 static void
 256 bna_bfi_rx_vlan_filter_set(struct bna_rxf *rxf, u8 block_idx)
 257 {
 258         struct bfi_enet_rx_vlan_req *req = &rxf->bfi_enet_cmd.vlan_req;
 259         int i;
 260         int j;
 261 
 262         bfi_msgq_mhdr_set(req->mh, BFI_MC_ENET,
 263                 BFI_ENET_H2I_RX_VLAN_SET_REQ, 0, rxf->rx->rid);
 264         req->mh.num_entries = htons(
 265                 bfi_msgq_num_cmd_entries(sizeof(struct bfi_enet_rx_vlan_req)));
 266         req->block_idx = block_idx;
 267         for (i = 0; i < (BFI_ENET_VLAN_BLOCK_SIZE / 32); i++) {
 268                 j = (block_idx * (BFI_ENET_VLAN_BLOCK_SIZE / 32)) + i;
 269                 if (rxf->vlan_filter_status == BNA_STATUS_T_ENABLED)
 270                         req->bit_mask[i] =
 271                                 htonl(rxf->vlan_filter_table[j]);
 272                 else
 273                         req->bit_mask[i] = 0xFFFFFFFF;
 274         }
 275         bfa_msgq_cmd_set(&rxf->msgq_cmd, NULL, NULL,
 276                 sizeof(struct bfi_enet_rx_vlan_req), &req->mh);
 277         bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd);
 278 }
 279 
 280 static void
 281 bna_bfi_vlan_strip_enable(struct bna_rxf *rxf)
 282 {
 283         struct bfi_enet_enable_req *req = &rxf->bfi_enet_cmd.req;
 284 
 285         bfi_msgq_mhdr_set(req->mh, BFI_MC_ENET,
 286                 BFI_ENET_H2I_RX_VLAN_STRIP_ENABLE_REQ, 0, rxf->rx->rid);
 287         req->mh.num_entries = htons(
 288                 bfi_msgq_num_cmd_entries(sizeof(struct bfi_enet_enable_req)));
 289         req->enable = rxf->vlan_strip_status;
 290         bfa_msgq_cmd_set(&rxf->msgq_cmd, NULL, NULL,
 291                 sizeof(struct bfi_enet_enable_req), &req->mh);
 292         bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd);
 293 }
 294 
 295 static void
 296 bna_bfi_rit_cfg(struct bna_rxf *rxf)
 297 {
 298         struct bfi_enet_rit_req *req = &rxf->bfi_enet_cmd.rit_req;
 299 
 300         bfi_msgq_mhdr_set(req->mh, BFI_MC_ENET,
 301                 BFI_ENET_H2I_RIT_CFG_REQ, 0, rxf->rx->rid);
 302         req->mh.num_entries = htons(
 303                 bfi_msgq_num_cmd_entries(sizeof(struct bfi_enet_rit_req)));
 304         req->size = htons(rxf->rit_size);
 305         memcpy(&req->table[0], rxf->rit, rxf->rit_size);
 306         bfa_msgq_cmd_set(&rxf->msgq_cmd, NULL, NULL,
 307                 sizeof(struct bfi_enet_rit_req), &req->mh);
 308         bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd);
 309 }
 310 
 311 static void
 312 bna_bfi_rss_cfg(struct bna_rxf *rxf)
 313 {
 314         struct bfi_enet_rss_cfg_req *req = &rxf->bfi_enet_cmd.rss_req;
 315         int i;
 316 
 317         bfi_msgq_mhdr_set(req->mh, BFI_MC_ENET,
 318                 BFI_ENET_H2I_RSS_CFG_REQ, 0, rxf->rx->rid);
 319         req->mh.num_entries = htons(
 320                 bfi_msgq_num_cmd_entries(sizeof(struct bfi_enet_rss_cfg_req)));
 321         req->cfg.type = rxf->rss_cfg.hash_type;
 322         req->cfg.mask = rxf->rss_cfg.hash_mask;
 323         for (i = 0; i < BFI_ENET_RSS_KEY_LEN; i++)
 324                 req->cfg.key[i] =
 325                         htonl(rxf->rss_cfg.toeplitz_hash_key[i]);
 326         bfa_msgq_cmd_set(&rxf->msgq_cmd, NULL, NULL,
 327                 sizeof(struct bfi_enet_rss_cfg_req), &req->mh);
 328         bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd);
 329 }
 330 
 331 static void
 332 bna_bfi_rss_enable(struct bna_rxf *rxf)
 333 {
 334         struct bfi_enet_enable_req *req = &rxf->bfi_enet_cmd.req;
 335 
 336         bfi_msgq_mhdr_set(req->mh, BFI_MC_ENET,
 337                 BFI_ENET_H2I_RSS_ENABLE_REQ, 0, rxf->rx->rid);
 338         req->mh.num_entries = htons(
 339                 bfi_msgq_num_cmd_entries(sizeof(struct bfi_enet_enable_req)));
 340         req->enable = rxf->rss_status;
 341         bfa_msgq_cmd_set(&rxf->msgq_cmd, NULL, NULL,
 342                 sizeof(struct bfi_enet_enable_req), &req->mh);
 343         bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd);
 344 }
 345 
 346 /* This function gets the multicast MAC that has already been added to CAM */
 347 static struct bna_mac *
 348 bna_rxf_mcmac_get(struct bna_rxf *rxf, const u8 *mac_addr)
 349 {
 350         struct bna_mac *mac;
 351 
 352         list_for_each_entry(mac, &rxf->mcast_active_q, qe)
 353                 if (ether_addr_equal(mac->addr, mac_addr))
 354                         return mac;
 355 
 356         list_for_each_entry(mac, &rxf->mcast_pending_del_q, qe)
 357                 if (ether_addr_equal(mac->addr, mac_addr))
 358                         return mac;
 359 
 360         return NULL;
 361 }
 362 
 363 static struct bna_mcam_handle *
 364 bna_rxf_mchandle_get(struct bna_rxf *rxf, int handle)
 365 {
 366         struct bna_mcam_handle *mchandle;
 367 
 368         list_for_each_entry(mchandle, &rxf->mcast_handle_q, qe)
 369                 if (mchandle->handle == handle)
 370                         return mchandle;
 371 
 372         return NULL;
 373 }
 374 
 375 static void
 376 bna_rxf_mchandle_attach(struct bna_rxf *rxf, u8 *mac_addr, int handle)
 377 {
 378         struct bna_mac *mcmac;
 379         struct bna_mcam_handle *mchandle;
 380 
 381         mcmac = bna_rxf_mcmac_get(rxf, mac_addr);
 382         mchandle = bna_rxf_mchandle_get(rxf, handle);
 383         if (mchandle == NULL) {
 384                 mchandle = bna_mcam_mod_handle_get(&rxf->rx->bna->mcam_mod);
 385                 mchandle->handle = handle;
 386                 mchandle->refcnt = 0;
 387                 list_add_tail(&mchandle->qe, &rxf->mcast_handle_q);
 388         }
 389         mchandle->refcnt++;
 390         mcmac->handle = mchandle;
 391 }
 392 
 393 static int
 394 bna_rxf_mcast_del(struct bna_rxf *rxf, struct bna_mac *mac,
 395                 enum bna_cleanup_type cleanup)
 396 {
 397         struct bna_mcam_handle *mchandle;
 398         int ret = 0;
 399 
 400         mchandle = mac->handle;
 401         if (mchandle == NULL)
 402                 return ret;
 403 
 404         mchandle->refcnt--;
 405         if (mchandle->refcnt == 0) {
 406                 if (cleanup == BNA_HARD_CLEANUP) {
 407                         bna_bfi_mcast_del_req(rxf, mchandle->handle);
 408                         ret = 1;
 409                 }
 410                 list_del(&mchandle->qe);
 411                 bna_mcam_mod_handle_put(&rxf->rx->bna->mcam_mod, mchandle);
 412         }
 413         mac->handle = NULL;
 414 
 415         return ret;
 416 }
 417 
 418 static int
 419 bna_rxf_mcast_cfg_apply(struct bna_rxf *rxf)
 420 {
 421         struct bna_mac *mac = NULL;
 422         int ret;
 423 
 424         /* First delete multicast entries to maintain the count */
 425         while (!list_empty(&rxf->mcast_pending_del_q)) {
 426                 mac = list_first_entry(&rxf->mcast_pending_del_q,
 427                                        struct bna_mac, qe);
 428                 ret = bna_rxf_mcast_del(rxf, mac, BNA_HARD_CLEANUP);
 429                 list_move_tail(&mac->qe, bna_mcam_mod_del_q(rxf->rx->bna));
 430                 if (ret)
 431                         return ret;
 432         }
 433 
 434         /* Add multicast entries */
 435         if (!list_empty(&rxf->mcast_pending_add_q)) {
 436                 mac = list_first_entry(&rxf->mcast_pending_add_q,
 437                                        struct bna_mac, qe);
 438                 list_move_tail(&mac->qe, &rxf->mcast_active_q);
 439                 bna_bfi_mcast_add_req(rxf, mac);
 440                 return 1;
 441         }
 442 
 443         return 0;
 444 }
 445 
 446 static int
 447 bna_rxf_vlan_cfg_apply(struct bna_rxf *rxf)
 448 {
 449         u8 vlan_pending_bitmask;
 450         int block_idx = 0;
 451 
 452         if (rxf->vlan_pending_bitmask) {
 453                 vlan_pending_bitmask = rxf->vlan_pending_bitmask;
 454                 while (!(vlan_pending_bitmask & 0x1)) {
 455                         block_idx++;
 456                         vlan_pending_bitmask >>= 1;
 457                 }
 458                 rxf->vlan_pending_bitmask &= ~BIT(block_idx);
 459                 bna_bfi_rx_vlan_filter_set(rxf, block_idx);
 460                 return 1;
 461         }
 462 
 463         return 0;
 464 }
 465 
 466 static int
 467 bna_rxf_mcast_cfg_reset(struct bna_rxf *rxf, enum bna_cleanup_type cleanup)
 468 {
 469         struct bna_mac *mac;
 470         int ret;
 471 
 472         /* Throw away delete pending mcast entries */
 473         while (!list_empty(&rxf->mcast_pending_del_q)) {
 474                 mac = list_first_entry(&rxf->mcast_pending_del_q,
 475                                        struct bna_mac, qe);
 476                 ret = bna_rxf_mcast_del(rxf, mac, cleanup);
 477                 list_move_tail(&mac->qe, bna_mcam_mod_del_q(rxf->rx->bna));
 478                 if (ret)
 479                         return ret;
 480         }
 481 
 482         /* Move active mcast entries to pending_add_q */
 483         while (!list_empty(&rxf->mcast_active_q)) {
 484                 mac = list_first_entry(&rxf->mcast_active_q,
 485                                        struct bna_mac, qe);
 486                 list_move_tail(&mac->qe, &rxf->mcast_pending_add_q);
 487                 if (bna_rxf_mcast_del(rxf, mac, cleanup))
 488                         return 1;
 489         }
 490 
 491         return 0;
 492 }
 493 
 494 static int
 495 bna_rxf_rss_cfg_apply(struct bna_rxf *rxf)
 496 {
 497         if (rxf->rss_pending) {
 498                 if (rxf->rss_pending & BNA_RSS_F_RIT_PENDING) {
 499                         rxf->rss_pending &= ~BNA_RSS_F_RIT_PENDING;
 500                         bna_bfi_rit_cfg(rxf);
 501                         return 1;
 502                 }
 503 
 504                 if (rxf->rss_pending & BNA_RSS_F_CFG_PENDING) {
 505                         rxf->rss_pending &= ~BNA_RSS_F_CFG_PENDING;
 506                         bna_bfi_rss_cfg(rxf);
 507                         return 1;
 508                 }
 509 
 510                 if (rxf->rss_pending & BNA_RSS_F_STATUS_PENDING) {
 511                         rxf->rss_pending &= ~BNA_RSS_F_STATUS_PENDING;
 512                         bna_bfi_rss_enable(rxf);
 513                         return 1;
 514                 }
 515         }
 516 
 517         return 0;
 518 }
 519 
 520 static int
 521 bna_rxf_cfg_apply(struct bna_rxf *rxf)
 522 {
 523         if (bna_rxf_ucast_cfg_apply(rxf))
 524                 return 1;
 525 
 526         if (bna_rxf_mcast_cfg_apply(rxf))
 527                 return 1;
 528 
 529         if (bna_rxf_promisc_cfg_apply(rxf))
 530                 return 1;
 531 
 532         if (bna_rxf_allmulti_cfg_apply(rxf))
 533                 return 1;
 534 
 535         if (bna_rxf_vlan_cfg_apply(rxf))
 536                 return 1;
 537 
 538         if (bna_rxf_vlan_strip_cfg_apply(rxf))
 539                 return 1;
 540 
 541         if (bna_rxf_rss_cfg_apply(rxf))
 542                 return 1;
 543 
 544         return 0;
 545 }
 546 
 547 static void
 548 bna_rxf_cfg_reset(struct bna_rxf *rxf)
 549 {
 550         bna_rxf_ucast_cfg_reset(rxf, BNA_SOFT_CLEANUP);
 551         bna_rxf_mcast_cfg_reset(rxf, BNA_SOFT_CLEANUP);
 552         bna_rxf_promisc_cfg_reset(rxf, BNA_SOFT_CLEANUP);
 553         bna_rxf_allmulti_cfg_reset(rxf, BNA_SOFT_CLEANUP);
 554         bna_rxf_vlan_cfg_soft_reset(rxf);
 555         bna_rxf_rss_cfg_soft_reset(rxf);
 556 }
 557 
 558 static void
 559 bna_rit_init(struct bna_rxf *rxf, int rit_size)
 560 {
 561         struct bna_rx *rx = rxf->rx;
 562         struct bna_rxp *rxp;
 563         int offset = 0;
 564 
 565         rxf->rit_size = rit_size;
 566         list_for_each_entry(rxp, &rx->rxp_q, qe) {
 567                 rxf->rit[offset] = rxp->cq.ccb->id;
 568                 offset++;
 569         }
 570 }
 571 
 572 void
 573 bna_bfi_rxf_cfg_rsp(struct bna_rxf *rxf, struct bfi_msgq_mhdr *msghdr)
 574 {
 575         bfa_fsm_send_event(rxf, RXF_E_FW_RESP);
 576 }
 577 
 578 void
 579 bna_bfi_rxf_ucast_set_rsp(struct bna_rxf *rxf,
 580                         struct bfi_msgq_mhdr *msghdr)
 581 {
 582         struct bfi_enet_rsp *rsp =
 583                 container_of(msghdr, struct bfi_enet_rsp, mh);
 584 
 585         if (rsp->error) {
 586                 /* Clear ucast from cache */
 587                 rxf->ucast_active_set = 0;
 588         }
 589 
 590         bfa_fsm_send_event(rxf, RXF_E_FW_RESP);
 591 }
 592 
 593 void
 594 bna_bfi_rxf_mcast_add_rsp(struct bna_rxf *rxf,
 595                         struct bfi_msgq_mhdr *msghdr)
 596 {
 597         struct bfi_enet_mcast_add_req *req =
 598                 &rxf->bfi_enet_cmd.mcast_add_req;
 599         struct bfi_enet_mcast_add_rsp *rsp =
 600                 container_of(msghdr, struct bfi_enet_mcast_add_rsp, mh);
 601 
 602         bna_rxf_mchandle_attach(rxf, (u8 *)&req->mac_addr,
 603                 ntohs(rsp->handle));
 604         bfa_fsm_send_event(rxf, RXF_E_FW_RESP);
 605 }
 606 
 607 static void
 608 bna_rxf_init(struct bna_rxf *rxf,
 609                 struct bna_rx *rx,
 610                 struct bna_rx_config *q_config,
 611                 struct bna_res_info *res_info)
 612 {
 613         rxf->rx = rx;
 614 
 615         INIT_LIST_HEAD(&rxf->ucast_pending_add_q);
 616         INIT_LIST_HEAD(&rxf->ucast_pending_del_q);
 617         rxf->ucast_pending_set = 0;
 618         rxf->ucast_active_set = 0;
 619         INIT_LIST_HEAD(&rxf->ucast_active_q);
 620         rxf->ucast_pending_mac = NULL;
 621 
 622         INIT_LIST_HEAD(&rxf->mcast_pending_add_q);
 623         INIT_LIST_HEAD(&rxf->mcast_pending_del_q);
 624         INIT_LIST_HEAD(&rxf->mcast_active_q);
 625         INIT_LIST_HEAD(&rxf->mcast_handle_q);
 626 
 627         rxf->rit = (u8 *)
 628                 res_info[BNA_RX_RES_MEM_T_RIT].res_u.mem_info.mdl[0].kva;
 629         bna_rit_init(rxf, q_config->num_paths);
 630 
 631         rxf->rss_status = q_config->rss_status;
 632         if (rxf->rss_status == BNA_STATUS_T_ENABLED) {
 633                 rxf->rss_cfg = q_config->rss_config;
 634                 rxf->rss_pending |= BNA_RSS_F_CFG_PENDING;
 635                 rxf->rss_pending |= BNA_RSS_F_RIT_PENDING;
 636                 rxf->rss_pending |= BNA_RSS_F_STATUS_PENDING;
 637         }
 638 
 639         rxf->vlan_filter_status = BNA_STATUS_T_DISABLED;
 640         memset(rxf->vlan_filter_table, 0,
 641                         (sizeof(u32) * (BFI_ENET_VLAN_ID_MAX / 32)));
 642         rxf->vlan_filter_table[0] |= 1; /* for pure priority tagged frames */
 643         rxf->vlan_pending_bitmask = (u8)BFI_VLAN_BMASK_ALL;
 644 
 645         rxf->vlan_strip_status = q_config->vlan_strip_status;
 646 
 647         bfa_fsm_set_state(rxf, bna_rxf_sm_stopped);
 648 }
 649 
 650 static void
 651 bna_rxf_uninit(struct bna_rxf *rxf)
 652 {
 653         struct bna_mac *mac;
 654 
 655         rxf->ucast_pending_set = 0;
 656         rxf->ucast_active_set = 0;
 657 
 658         while (!list_empty(&rxf->ucast_pending_add_q)) {
 659                 mac = list_first_entry(&rxf->ucast_pending_add_q,
 660                                        struct bna_mac, qe);
 661                 list_move_tail(&mac->qe, bna_ucam_mod_free_q(rxf->rx->bna));
 662         }
 663 
 664         if (rxf->ucast_pending_mac) {
 665                 list_add_tail(&rxf->ucast_pending_mac->qe,
 666                               bna_ucam_mod_free_q(rxf->rx->bna));
 667                 rxf->ucast_pending_mac = NULL;
 668         }
 669 
 670         while (!list_empty(&rxf->mcast_pending_add_q)) {
 671                 mac = list_first_entry(&rxf->mcast_pending_add_q,
 672                                        struct bna_mac, qe);
 673                 list_move_tail(&mac->qe, bna_mcam_mod_free_q(rxf->rx->bna));
 674         }
 675 
 676         rxf->rxmode_pending = 0;
 677         rxf->rxmode_pending_bitmask = 0;
 678         if (rxf->rx->bna->promisc_rid == rxf->rx->rid)
 679                 rxf->rx->bna->promisc_rid = BFI_INVALID_RID;
 680         if (rxf->rx->bna->default_mode_rid == rxf->rx->rid)
 681                 rxf->rx->bna->default_mode_rid = BFI_INVALID_RID;
 682 
 683         rxf->rss_pending = 0;
 684         rxf->vlan_strip_pending = false;
 685 
 686         rxf->rx = NULL;
 687 }
 688 
 689 static void
 690 bna_rx_cb_rxf_started(struct bna_rx *rx)
 691 {
 692         bfa_fsm_send_event(rx, RX_E_RXF_STARTED);
 693 }
 694 
 695 static void
 696 bna_rxf_start(struct bna_rxf *rxf)
 697 {
 698         rxf->start_cbfn = bna_rx_cb_rxf_started;
 699         rxf->start_cbarg = rxf->rx;
 700         bfa_fsm_send_event(rxf, RXF_E_START);
 701 }
 702 
 703 static void
 704 bna_rx_cb_rxf_stopped(struct bna_rx *rx)
 705 {
 706         bfa_fsm_send_event(rx, RX_E_RXF_STOPPED);
 707 }
 708 
 709 static void
 710 bna_rxf_stop(struct bna_rxf *rxf)
 711 {
 712         rxf->stop_cbfn = bna_rx_cb_rxf_stopped;
 713         rxf->stop_cbarg = rxf->rx;
 714         bfa_fsm_send_event(rxf, RXF_E_STOP);
 715 }
 716 
 717 static void
 718 bna_rxf_fail(struct bna_rxf *rxf)
 719 {
 720         bfa_fsm_send_event(rxf, RXF_E_FAIL);
 721 }
 722 
 723 enum bna_cb_status
 724 bna_rx_ucast_set(struct bna_rx *rx, const u8 *ucmac)
 725 {
 726         struct bna_rxf *rxf = &rx->rxf;
 727 
 728         if (rxf->ucast_pending_mac == NULL) {
 729                 rxf->ucast_pending_mac =
 730                         bna_cam_mod_mac_get(bna_ucam_mod_free_q(rxf->rx->bna));
 731                 if (rxf->ucast_pending_mac == NULL)
 732                         return BNA_CB_UCAST_CAM_FULL;
 733         }
 734 
 735         ether_addr_copy(rxf->ucast_pending_mac->addr, ucmac);
 736         rxf->ucast_pending_set = 1;
 737         rxf->cam_fltr_cbfn = NULL;
 738         rxf->cam_fltr_cbarg = rx->bna->bnad;
 739 
 740         bfa_fsm_send_event(rxf, RXF_E_CONFIG);
 741 
 742         return BNA_CB_SUCCESS;
 743 }
 744 
 745 enum bna_cb_status
 746 bna_rx_mcast_add(struct bna_rx *rx, const u8 *addr,
 747                  void (*cbfn)(struct bnad *, struct bna_rx *))
 748 {
 749         struct bna_rxf *rxf = &rx->rxf;
 750         struct bna_mac *mac;
 751 
 752         /* Check if already added or pending addition */
 753         if (bna_mac_find(&rxf->mcast_active_q, addr) ||
 754                 bna_mac_find(&rxf->mcast_pending_add_q, addr)) {
 755                 if (cbfn)
 756                         cbfn(rx->bna->bnad, rx);
 757                 return BNA_CB_SUCCESS;
 758         }
 759 
 760         mac = bna_cam_mod_mac_get(bna_mcam_mod_free_q(rxf->rx->bna));
 761         if (mac == NULL)
 762                 return BNA_CB_MCAST_LIST_FULL;
 763         ether_addr_copy(mac->addr, addr);
 764         list_add_tail(&mac->qe, &rxf->mcast_pending_add_q);
 765 
 766         rxf->cam_fltr_cbfn = cbfn;
 767         rxf->cam_fltr_cbarg = rx->bna->bnad;
 768 
 769         bfa_fsm_send_event(rxf, RXF_E_CONFIG);
 770 
 771         return BNA_CB_SUCCESS;
 772 }
 773 
 774 enum bna_cb_status
 775 bna_rx_ucast_listset(struct bna_rx *rx, int count, const u8 *uclist)
 776 {
 777         struct bna_ucam_mod *ucam_mod = &rx->bna->ucam_mod;
 778         struct bna_rxf *rxf = &rx->rxf;
 779         struct list_head list_head;
 780         const u8 *mcaddr;
 781         struct bna_mac *mac, *del_mac;
 782         int i;
 783 
 784         /* Purge the pending_add_q */
 785         while (!list_empty(&rxf->ucast_pending_add_q)) {
 786                 mac = list_first_entry(&rxf->ucast_pending_add_q,
 787                                        struct bna_mac, qe);
 788                 list_move_tail(&mac->qe, &ucam_mod->free_q);
 789         }
 790 
 791         /* Schedule active_q entries for deletion */
 792         while (!list_empty(&rxf->ucast_active_q)) {
 793                 mac = list_first_entry(&rxf->ucast_active_q,
 794                                        struct bna_mac, qe);
 795                 del_mac = bna_cam_mod_mac_get(&ucam_mod->del_q);
 796                 ether_addr_copy(del_mac->addr, mac->addr);
 797                 del_mac->handle = mac->handle;
 798                 list_add_tail(&del_mac->qe, &rxf->ucast_pending_del_q);
 799                 list_move_tail(&mac->qe, &ucam_mod->free_q);
 800         }
 801 
 802         /* Allocate nodes */
 803         INIT_LIST_HEAD(&list_head);
 804         for (i = 0, mcaddr = uclist; i < count; i++) {
 805                 mac = bna_cam_mod_mac_get(&ucam_mod->free_q);
 806                 if (mac == NULL)
 807                         goto err_return;
 808                 ether_addr_copy(mac->addr, mcaddr);
 809                 list_add_tail(&mac->qe, &list_head);
 810                 mcaddr += ETH_ALEN;
 811         }
 812 
 813         /* Add the new entries */
 814         while (!list_empty(&list_head)) {
 815                 mac = list_first_entry(&list_head, struct bna_mac, qe);
 816                 list_move_tail(&mac->qe, &rxf->ucast_pending_add_q);
 817         }
 818 
 819         bfa_fsm_send_event(rxf, RXF_E_CONFIG);
 820 
 821         return BNA_CB_SUCCESS;
 822 
 823 err_return:
 824         while (!list_empty(&list_head)) {
 825                 mac = list_first_entry(&list_head, struct bna_mac, qe);
 826                 list_move_tail(&mac->qe, &ucam_mod->free_q);
 827         }
 828 
 829         return BNA_CB_UCAST_CAM_FULL;
 830 }
 831 
 832 enum bna_cb_status
 833 bna_rx_mcast_listset(struct bna_rx *rx, int count, const u8 *mclist)
 834 {
 835         struct bna_mcam_mod *mcam_mod = &rx->bna->mcam_mod;
 836         struct bna_rxf *rxf = &rx->rxf;
 837         struct list_head list_head;
 838         const u8 *mcaddr;
 839         struct bna_mac *mac, *del_mac;
 840         int i;
 841 
 842         /* Purge the pending_add_q */
 843         while (!list_empty(&rxf->mcast_pending_add_q)) {
 844                 mac = list_first_entry(&rxf->mcast_pending_add_q,
 845                                        struct bna_mac, qe);
 846                 list_move_tail(&mac->qe, &mcam_mod->free_q);
 847         }
 848 
 849         /* Schedule active_q entries for deletion */
 850         while (!list_empty(&rxf->mcast_active_q)) {
 851                 mac = list_first_entry(&rxf->mcast_active_q,
 852                                        struct bna_mac, qe);
 853                 del_mac = bna_cam_mod_mac_get(&mcam_mod->del_q);
 854                 ether_addr_copy(del_mac->addr, mac->addr);
 855                 del_mac->handle = mac->handle;
 856                 list_add_tail(&del_mac->qe, &rxf->mcast_pending_del_q);
 857                 mac->handle = NULL;
 858                 list_move_tail(&mac->qe, &mcam_mod->free_q);
 859         }
 860 
 861         /* Allocate nodes */
 862         INIT_LIST_HEAD(&list_head);
 863         for (i = 0, mcaddr = mclist; i < count; i++) {
 864                 mac = bna_cam_mod_mac_get(&mcam_mod->free_q);
 865                 if (mac == NULL)
 866                         goto err_return;
 867                 ether_addr_copy(mac->addr, mcaddr);
 868                 list_add_tail(&mac->qe, &list_head);
 869 
 870                 mcaddr += ETH_ALEN;
 871         }
 872 
 873         /* Add the new entries */
 874         while (!list_empty(&list_head)) {
 875                 mac = list_first_entry(&list_head, struct bna_mac, qe);
 876                 list_move_tail(&mac->qe, &rxf->mcast_pending_add_q);
 877         }
 878 
 879         bfa_fsm_send_event(rxf, RXF_E_CONFIG);
 880 
 881         return BNA_CB_SUCCESS;
 882 
 883 err_return:
 884         while (!list_empty(&list_head)) {
 885                 mac = list_first_entry(&list_head, struct bna_mac, qe);
 886                 list_move_tail(&mac->qe, &mcam_mod->free_q);
 887         }
 888 
 889         return BNA_CB_MCAST_LIST_FULL;
 890 }
 891 
 892 void
 893 bna_rx_mcast_delall(struct bna_rx *rx)
 894 {
 895         struct bna_rxf *rxf = &rx->rxf;
 896         struct bna_mac *mac, *del_mac;
 897         int need_hw_config = 0;
 898 
 899         /* Purge all entries from pending_add_q */
 900         while (!list_empty(&rxf->mcast_pending_add_q)) {
 901                 mac = list_first_entry(&rxf->mcast_pending_add_q,
 902                                        struct bna_mac, qe);
 903                 list_move_tail(&mac->qe, bna_mcam_mod_free_q(rxf->rx->bna));
 904         }
 905 
 906         /* Schedule all entries in active_q for deletion */
 907         while (!list_empty(&rxf->mcast_active_q)) {
 908                 mac = list_first_entry(&rxf->mcast_active_q,
 909                                        struct bna_mac, qe);
 910                 list_del(&mac->qe);
 911                 del_mac = bna_cam_mod_mac_get(bna_mcam_mod_del_q(rxf->rx->bna));
 912                 memcpy(del_mac, mac, sizeof(*del_mac));
 913                 list_add_tail(&del_mac->qe, &rxf->mcast_pending_del_q);
 914                 mac->handle = NULL;
 915                 list_add_tail(&mac->qe, bna_mcam_mod_free_q(rxf->rx->bna));
 916                 need_hw_config = 1;
 917         }
 918 
 919         if (need_hw_config)
 920                 bfa_fsm_send_event(rxf, RXF_E_CONFIG);
 921 }
 922 
 923 void
 924 bna_rx_vlan_add(struct bna_rx *rx, int vlan_id)
 925 {
 926         struct bna_rxf *rxf = &rx->rxf;
 927         int index = (vlan_id >> BFI_VLAN_WORD_SHIFT);
 928         int bit = BIT(vlan_id & BFI_VLAN_WORD_MASK);
 929         int group_id = (vlan_id >> BFI_VLAN_BLOCK_SHIFT);
 930 
 931         rxf->vlan_filter_table[index] |= bit;
 932         if (rxf->vlan_filter_status == BNA_STATUS_T_ENABLED) {
 933                 rxf->vlan_pending_bitmask |= BIT(group_id);
 934                 bfa_fsm_send_event(rxf, RXF_E_CONFIG);
 935         }
 936 }
 937 
 938 void
 939 bna_rx_vlan_del(struct bna_rx *rx, int vlan_id)
 940 {
 941         struct bna_rxf *rxf = &rx->rxf;
 942         int index = (vlan_id >> BFI_VLAN_WORD_SHIFT);
 943         int bit = BIT(vlan_id & BFI_VLAN_WORD_MASK);
 944         int group_id = (vlan_id >> BFI_VLAN_BLOCK_SHIFT);
 945 
 946         rxf->vlan_filter_table[index] &= ~bit;
 947         if (rxf->vlan_filter_status == BNA_STATUS_T_ENABLED) {
 948                 rxf->vlan_pending_bitmask |= BIT(group_id);
 949                 bfa_fsm_send_event(rxf, RXF_E_CONFIG);
 950         }
 951 }
 952 
 953 static int
 954 bna_rxf_ucast_cfg_apply(struct bna_rxf *rxf)
 955 {
 956         struct bna_mac *mac = NULL;
 957 
 958         /* Delete MAC addresses previousely added */
 959         if (!list_empty(&rxf->ucast_pending_del_q)) {
 960                 mac = list_first_entry(&rxf->ucast_pending_del_q,
 961                                        struct bna_mac, qe);
 962                 bna_bfi_ucast_req(rxf, mac, BFI_ENET_H2I_MAC_UCAST_DEL_REQ);
 963                 list_move_tail(&mac->qe, bna_ucam_mod_del_q(rxf->rx->bna));
 964                 return 1;
 965         }
 966 
 967         /* Set default unicast MAC */
 968         if (rxf->ucast_pending_set) {
 969                 rxf->ucast_pending_set = 0;
 970                 ether_addr_copy(rxf->ucast_active_mac.addr,
 971                                 rxf->ucast_pending_mac->addr);
 972                 rxf->ucast_active_set = 1;
 973                 bna_bfi_ucast_req(rxf, &rxf->ucast_active_mac,
 974                         BFI_ENET_H2I_MAC_UCAST_SET_REQ);
 975                 return 1;
 976         }
 977 
 978         /* Add additional MAC entries */
 979         if (!list_empty(&rxf->ucast_pending_add_q)) {
 980                 mac = list_first_entry(&rxf->ucast_pending_add_q,
 981                                        struct bna_mac, qe);
 982                 list_move_tail(&mac->qe, &rxf->ucast_active_q);
 983                 bna_bfi_ucast_req(rxf, mac, BFI_ENET_H2I_MAC_UCAST_ADD_REQ);
 984                 return 1;
 985         }
 986 
 987         return 0;
 988 }
 989 
 990 static int
 991 bna_rxf_ucast_cfg_reset(struct bna_rxf *rxf, enum bna_cleanup_type cleanup)
 992 {
 993         struct bna_mac *mac;
 994 
 995         /* Throw away delete pending ucast entries */
 996         while (!list_empty(&rxf->ucast_pending_del_q)) {
 997                 mac = list_first_entry(&rxf->ucast_pending_del_q,
 998                                        struct bna_mac, qe);
 999                 if (cleanup == BNA_SOFT_CLEANUP)
1000                         list_move_tail(&mac->qe,
1001                                        bna_ucam_mod_del_q(rxf->rx->bna));
1002                 else {
1003                         bna_bfi_ucast_req(rxf, mac,
1004                                           BFI_ENET_H2I_MAC_UCAST_DEL_REQ);
1005                         list_move_tail(&mac->qe,
1006                                        bna_ucam_mod_del_q(rxf->rx->bna));
1007                         return 1;
1008                 }
1009         }
1010 
1011         /* Move active ucast entries to pending_add_q */
1012         while (!list_empty(&rxf->ucast_active_q)) {
1013                 mac = list_first_entry(&rxf->ucast_active_q,
1014                                        struct bna_mac, qe);
1015                 list_move_tail(&mac->qe, &rxf->ucast_pending_add_q);
1016                 if (cleanup == BNA_HARD_CLEANUP) {
1017                         bna_bfi_ucast_req(rxf, mac,
1018                                 BFI_ENET_H2I_MAC_UCAST_DEL_REQ);
1019                         return 1;
1020                 }
1021         }
1022 
1023         if (rxf->ucast_active_set) {
1024                 rxf->ucast_pending_set = 1;
1025                 rxf->ucast_active_set = 0;
1026                 if (cleanup == BNA_HARD_CLEANUP) {
1027                         bna_bfi_ucast_req(rxf, &rxf->ucast_active_mac,
1028                                 BFI_ENET_H2I_MAC_UCAST_CLR_REQ);
1029                         return 1;
1030                 }
1031         }
1032 
1033         return 0;
1034 }
1035 
1036 static int
1037 bna_rxf_promisc_cfg_apply(struct bna_rxf *rxf)
1038 {
1039         struct bna *bna = rxf->rx->bna;
1040 
1041         /* Enable/disable promiscuous mode */
1042         if (is_promisc_enable(rxf->rxmode_pending,
1043                                 rxf->rxmode_pending_bitmask)) {
1044                 /* move promisc configuration from pending -> active */
1045                 promisc_inactive(rxf->rxmode_pending,
1046                                 rxf->rxmode_pending_bitmask);
1047                 rxf->rxmode_active |= BNA_RXMODE_PROMISC;
1048                 bna_bfi_rx_promisc_req(rxf, BNA_STATUS_T_ENABLED);
1049                 return 1;
1050         } else if (is_promisc_disable(rxf->rxmode_pending,
1051                                 rxf->rxmode_pending_bitmask)) {
1052                 /* move promisc configuration from pending -> active */
1053                 promisc_inactive(rxf->rxmode_pending,
1054                                 rxf->rxmode_pending_bitmask);
1055                 rxf->rxmode_active &= ~BNA_RXMODE_PROMISC;
1056                 bna->promisc_rid = BFI_INVALID_RID;
1057                 bna_bfi_rx_promisc_req(rxf, BNA_STATUS_T_DISABLED);
1058                 return 1;
1059         }
1060 
1061         return 0;
1062 }
1063 
1064 static int
1065 bna_rxf_promisc_cfg_reset(struct bna_rxf *rxf, enum bna_cleanup_type cleanup)
1066 {
1067         struct bna *bna = rxf->rx->bna;
1068 
1069         /* Clear pending promisc mode disable */
1070         if (is_promisc_disable(rxf->rxmode_pending,
1071                                 rxf->rxmode_pending_bitmask)) {
1072                 promisc_inactive(rxf->rxmode_pending,
1073                                 rxf->rxmode_pending_bitmask);
1074                 rxf->rxmode_active &= ~BNA_RXMODE_PROMISC;
1075                 bna->promisc_rid = BFI_INVALID_RID;
1076                 if (cleanup == BNA_HARD_CLEANUP) {
1077                         bna_bfi_rx_promisc_req(rxf, BNA_STATUS_T_DISABLED);
1078                         return 1;
1079                 }
1080         }
1081 
1082         /* Move promisc mode config from active -> pending */
1083         if (rxf->rxmode_active & BNA_RXMODE_PROMISC) {
1084                 promisc_enable(rxf->rxmode_pending,
1085                                 rxf->rxmode_pending_bitmask);
1086                 rxf->rxmode_active &= ~BNA_RXMODE_PROMISC;
1087                 if (cleanup == BNA_HARD_CLEANUP) {
1088                         bna_bfi_rx_promisc_req(rxf, BNA_STATUS_T_DISABLED);
1089                         return 1;
1090                 }
1091         }
1092 
1093         return 0;
1094 }
1095 
1096 static int
1097 bna_rxf_allmulti_cfg_apply(struct bna_rxf *rxf)
1098 {
1099         /* Enable/disable allmulti mode */
1100         if (is_allmulti_enable(rxf->rxmode_pending,
1101                                 rxf->rxmode_pending_bitmask)) {
1102                 /* move allmulti configuration from pending -> active */
1103                 allmulti_inactive(rxf->rxmode_pending,
1104                                 rxf->rxmode_pending_bitmask);
1105                 rxf->rxmode_active |= BNA_RXMODE_ALLMULTI;
1106                 bna_bfi_mcast_filter_req(rxf, BNA_STATUS_T_DISABLED);
1107                 return 1;
1108         } else if (is_allmulti_disable(rxf->rxmode_pending,
1109                                         rxf->rxmode_pending_bitmask)) {
1110                 /* move allmulti configuration from pending -> active */
1111                 allmulti_inactive(rxf->rxmode_pending,
1112                                 rxf->rxmode_pending_bitmask);
1113                 rxf->rxmode_active &= ~BNA_RXMODE_ALLMULTI;
1114                 bna_bfi_mcast_filter_req(rxf, BNA_STATUS_T_ENABLED);
1115                 return 1;
1116         }
1117 
1118         return 0;
1119 }
1120 
1121 static int
1122 bna_rxf_allmulti_cfg_reset(struct bna_rxf *rxf, enum bna_cleanup_type cleanup)
1123 {
1124         /* Clear pending allmulti mode disable */
1125         if (is_allmulti_disable(rxf->rxmode_pending,
1126                                 rxf->rxmode_pending_bitmask)) {
1127                 allmulti_inactive(rxf->rxmode_pending,
1128                                 rxf->rxmode_pending_bitmask);
1129                 rxf->rxmode_active &= ~BNA_RXMODE_ALLMULTI;
1130                 if (cleanup == BNA_HARD_CLEANUP) {
1131                         bna_bfi_mcast_filter_req(rxf, BNA_STATUS_T_ENABLED);
1132                         return 1;
1133                 }
1134         }
1135 
1136         /* Move allmulti mode config from active -> pending */
1137         if (rxf->rxmode_active & BNA_RXMODE_ALLMULTI) {
1138                 allmulti_enable(rxf->rxmode_pending,
1139                                 rxf->rxmode_pending_bitmask);
1140                 rxf->rxmode_active &= ~BNA_RXMODE_ALLMULTI;
1141                 if (cleanup == BNA_HARD_CLEANUP) {
1142                         bna_bfi_mcast_filter_req(rxf, BNA_STATUS_T_ENABLED);
1143                         return 1;
1144                 }
1145         }
1146 
1147         return 0;
1148 }
1149 
1150 static int
1151 bna_rxf_promisc_enable(struct bna_rxf *rxf)
1152 {
1153         struct bna *bna = rxf->rx->bna;
1154         int ret = 0;
1155 
1156         if (is_promisc_enable(rxf->rxmode_pending,
1157                                 rxf->rxmode_pending_bitmask) ||
1158                 (rxf->rxmode_active & BNA_RXMODE_PROMISC)) {
1159                 /* Do nothing if pending enable or already enabled */
1160         } else if (is_promisc_disable(rxf->rxmode_pending,
1161                                         rxf->rxmode_pending_bitmask)) {
1162                 /* Turn off pending disable command */
1163                 promisc_inactive(rxf->rxmode_pending,
1164                         rxf->rxmode_pending_bitmask);
1165         } else {
1166                 /* Schedule enable */
1167                 promisc_enable(rxf->rxmode_pending,
1168                                 rxf->rxmode_pending_bitmask);
1169                 bna->promisc_rid = rxf->rx->rid;
1170                 ret = 1;
1171         }
1172 
1173         return ret;
1174 }
1175 
1176 static int
1177 bna_rxf_promisc_disable(struct bna_rxf *rxf)
1178 {
1179         struct bna *bna = rxf->rx->bna;
1180         int ret = 0;
1181 
1182         if (is_promisc_disable(rxf->rxmode_pending,
1183                                 rxf->rxmode_pending_bitmask) ||
1184                 (!(rxf->rxmode_active & BNA_RXMODE_PROMISC))) {
1185                 /* Do nothing if pending disable or already disabled */
1186         } else if (is_promisc_enable(rxf->rxmode_pending,
1187                                         rxf->rxmode_pending_bitmask)) {
1188                 /* Turn off pending enable command */
1189                 promisc_inactive(rxf->rxmode_pending,
1190                                 rxf->rxmode_pending_bitmask);
1191                 bna->promisc_rid = BFI_INVALID_RID;
1192         } else if (rxf->rxmode_active & BNA_RXMODE_PROMISC) {
1193                 /* Schedule disable */
1194                 promisc_disable(rxf->rxmode_pending,
1195                                 rxf->rxmode_pending_bitmask);
1196                 ret = 1;
1197         }
1198 
1199         return ret;
1200 }
1201 
1202 static int
1203 bna_rxf_allmulti_enable(struct bna_rxf *rxf)
1204 {
1205         int ret = 0;
1206 
1207         if (is_allmulti_enable(rxf->rxmode_pending,
1208                         rxf->rxmode_pending_bitmask) ||
1209                         (rxf->rxmode_active & BNA_RXMODE_ALLMULTI)) {
1210                 /* Do nothing if pending enable or already enabled */
1211         } else if (is_allmulti_disable(rxf->rxmode_pending,
1212                                         rxf->rxmode_pending_bitmask)) {
1213                 /* Turn off pending disable command */
1214                 allmulti_inactive(rxf->rxmode_pending,
1215                         rxf->rxmode_pending_bitmask);
1216         } else {
1217                 /* Schedule enable */
1218                 allmulti_enable(rxf->rxmode_pending,
1219                                 rxf->rxmode_pending_bitmask);
1220                 ret = 1;
1221         }
1222 
1223         return ret;
1224 }
1225 
1226 static int
1227 bna_rxf_allmulti_disable(struct bna_rxf *rxf)
1228 {
1229         int ret = 0;
1230 
1231         if (is_allmulti_disable(rxf->rxmode_pending,
1232                                 rxf->rxmode_pending_bitmask) ||
1233                 (!(rxf->rxmode_active & BNA_RXMODE_ALLMULTI))) {
1234                 /* Do nothing if pending disable or already disabled */
1235         } else if (is_allmulti_enable(rxf->rxmode_pending,
1236                                         rxf->rxmode_pending_bitmask)) {
1237                 /* Turn off pending enable command */
1238                 allmulti_inactive(rxf->rxmode_pending,
1239                                 rxf->rxmode_pending_bitmask);
1240         } else if (rxf->rxmode_active & BNA_RXMODE_ALLMULTI) {
1241                 /* Schedule disable */
1242                 allmulti_disable(rxf->rxmode_pending,
1243                                 rxf->rxmode_pending_bitmask);
1244                 ret = 1;
1245         }
1246 
1247         return ret;
1248 }
1249 
1250 static int
1251 bna_rxf_vlan_strip_cfg_apply(struct bna_rxf *rxf)
1252 {
1253         if (rxf->vlan_strip_pending) {
1254                         rxf->vlan_strip_pending = false;
1255                         bna_bfi_vlan_strip_enable(rxf);
1256                         return 1;
1257         }
1258 
1259         return 0;
1260 }
1261 
1262 /* RX */
1263 
1264 #define BNA_GET_RXQS(qcfg)      (((qcfg)->rxp_type == BNA_RXP_SINGLE) ? \
1265         (qcfg)->num_paths : ((qcfg)->num_paths * 2))
1266 
1267 #define SIZE_TO_PAGES(size)     (((size) >> PAGE_SHIFT) + ((((size) &\
1268         (PAGE_SIZE - 1)) + (PAGE_SIZE - 1)) >> PAGE_SHIFT))
1269 
1270 #define call_rx_stop_cbfn(rx)                                           \
1271 do {                                                                \
1272         if ((rx)->stop_cbfn) {                                          \
1273                 void (*cbfn)(void *, struct bna_rx *);    \
1274                 void *cbarg;                                        \
1275                 cbfn = (rx)->stop_cbfn;                          \
1276                 cbarg = (rx)->stop_cbarg;                              \
1277                 (rx)->stop_cbfn = NULL;                                 \
1278                 (rx)->stop_cbarg = NULL;                                \
1279                 cbfn(cbarg, rx);                                        \
1280         }                                                              \
1281 } while (0)
1282 
1283 #define call_rx_stall_cbfn(rx)                                          \
1284 do {                                                                    \
1285         if ((rx)->rx_stall_cbfn)                                        \
1286                 (rx)->rx_stall_cbfn((rx)->bna->bnad, (rx));             \
1287 } while (0)
1288 
1289 #define bfi_enet_datapath_q_init(bfi_q, bna_qpt)                        \
1290 do {                                                                    \
1291         struct bna_dma_addr cur_q_addr =                                \
1292                 *((struct bna_dma_addr *)((bna_qpt)->kv_qpt_ptr));      \
1293         (bfi_q)->pg_tbl.a32.addr_lo = (bna_qpt)->hw_qpt_ptr.lsb;        \
1294         (bfi_q)->pg_tbl.a32.addr_hi = (bna_qpt)->hw_qpt_ptr.msb;        \
1295         (bfi_q)->first_entry.a32.addr_lo = cur_q_addr.lsb;              \
1296         (bfi_q)->first_entry.a32.addr_hi = cur_q_addr.msb;              \
1297         (bfi_q)->pages = htons((u16)(bna_qpt)->page_count);     \
1298         (bfi_q)->page_sz = htons((u16)(bna_qpt)->page_size);\
1299 } while (0)
1300 
1301 static void bna_bfi_rx_enet_start(struct bna_rx *rx);
1302 static void bna_rx_enet_stop(struct bna_rx *rx);
1303 static void bna_rx_mod_cb_rx_stopped(void *arg, struct bna_rx *rx);
1304 
1305 bfa_fsm_state_decl(bna_rx, stopped,
1306         struct bna_rx, enum bna_rx_event);
1307 bfa_fsm_state_decl(bna_rx, start_wait,
1308         struct bna_rx, enum bna_rx_event);
1309 bfa_fsm_state_decl(bna_rx, start_stop_wait,
1310         struct bna_rx, enum bna_rx_event);
1311 bfa_fsm_state_decl(bna_rx, rxf_start_wait,
1312         struct bna_rx, enum bna_rx_event);
1313 bfa_fsm_state_decl(bna_rx, started,
1314         struct bna_rx, enum bna_rx_event);
1315 bfa_fsm_state_decl(bna_rx, rxf_stop_wait,
1316         struct bna_rx, enum bna_rx_event);
1317 bfa_fsm_state_decl(bna_rx, stop_wait,
1318         struct bna_rx, enum bna_rx_event);
1319 bfa_fsm_state_decl(bna_rx, cleanup_wait,
1320         struct bna_rx, enum bna_rx_event);
1321 bfa_fsm_state_decl(bna_rx, failed,
1322         struct bna_rx, enum bna_rx_event);
1323 bfa_fsm_state_decl(bna_rx, quiesce_wait,
1324         struct bna_rx, enum bna_rx_event);
1325 
1326 static void bna_rx_sm_stopped_entry(struct bna_rx *rx)
1327 {
1328         call_rx_stop_cbfn(rx);
1329 }
1330 
1331 static void bna_rx_sm_stopped(struct bna_rx *rx,
1332                                 enum bna_rx_event event)
1333 {
1334         switch (event) {
1335         case RX_E_START:
1336                 bfa_fsm_set_state(rx, bna_rx_sm_start_wait);
1337                 break;
1338 
1339         case RX_E_STOP:
1340                 call_rx_stop_cbfn(rx);
1341                 break;
1342 
1343         case RX_E_FAIL:
1344                 /* no-op */
1345                 break;
1346 
1347         default:
1348                 bfa_sm_fault(event);
1349                 break;
1350         }
1351 }
1352 
1353 static void bna_rx_sm_start_wait_entry(struct bna_rx *rx)
1354 {
1355         bna_bfi_rx_enet_start(rx);
1356 }
1357 
1358 static void
1359 bna_rx_sm_stop_wait_entry(struct bna_rx *rx)
1360 {
1361 }
1362 
1363 static void
1364 bna_rx_sm_stop_wait(struct bna_rx *rx, enum bna_rx_event event)
1365 {
1366         switch (event) {
1367         case RX_E_FAIL:
1368         case RX_E_STOPPED:
1369                 bfa_fsm_set_state(rx, bna_rx_sm_cleanup_wait);
1370                 rx->rx_cleanup_cbfn(rx->bna->bnad, rx);
1371                 break;
1372 
1373         case RX_E_STARTED:
1374                 bna_rx_enet_stop(rx);
1375                 break;
1376 
1377         default:
1378                 bfa_sm_fault(event);
1379                 break;
1380         }
1381 }
1382 
1383 static void bna_rx_sm_start_wait(struct bna_rx *rx,
1384                                 enum bna_rx_event event)
1385 {
1386         switch (event) {
1387         case RX_E_STOP:
1388                 bfa_fsm_set_state(rx, bna_rx_sm_start_stop_wait);
1389                 break;
1390 
1391         case RX_E_FAIL:
1392                 bfa_fsm_set_state(rx, bna_rx_sm_stopped);
1393                 break;
1394 
1395         case RX_E_STARTED:
1396                 bfa_fsm_set_state(rx, bna_rx_sm_rxf_start_wait);
1397                 break;
1398 
1399         default:
1400                 bfa_sm_fault(event);
1401                 break;
1402         }
1403 }
1404 
1405 static void bna_rx_sm_rxf_start_wait_entry(struct bna_rx *rx)
1406 {
1407         rx->rx_post_cbfn(rx->bna->bnad, rx);
1408         bna_rxf_start(&rx->rxf);
1409 }
1410 
1411 static void
1412 bna_rx_sm_rxf_stop_wait_entry(struct bna_rx *rx)
1413 {
1414 }
1415 
1416 static void
1417 bna_rx_sm_rxf_stop_wait(struct bna_rx *rx, enum bna_rx_event event)
1418 {
1419         switch (event) {
1420         case RX_E_FAIL:
1421                 bfa_fsm_set_state(rx, bna_rx_sm_cleanup_wait);
1422                 bna_rxf_fail(&rx->rxf);
1423                 call_rx_stall_cbfn(rx);
1424                 rx->rx_cleanup_cbfn(rx->bna->bnad, rx);
1425                 break;
1426 
1427         case RX_E_RXF_STARTED:
1428                 bna_rxf_stop(&rx->rxf);
1429                 break;
1430 
1431         case RX_E_RXF_STOPPED:
1432                 bfa_fsm_set_state(rx, bna_rx_sm_stop_wait);
1433                 call_rx_stall_cbfn(rx);
1434                 bna_rx_enet_stop(rx);
1435                 break;
1436 
1437         default:
1438                 bfa_sm_fault(event);
1439                 break;
1440         }
1441 
1442 }
1443 
1444 static void
1445 bna_rx_sm_start_stop_wait_entry(struct bna_rx *rx)
1446 {
1447 }
1448 
1449 static void
1450 bna_rx_sm_start_stop_wait(struct bna_rx *rx, enum bna_rx_event event)
1451 {
1452         switch (event) {
1453         case RX_E_FAIL:
1454         case RX_E_STOPPED:
1455                 bfa_fsm_set_state(rx, bna_rx_sm_stopped);
1456                 break;
1457 
1458         case RX_E_STARTED:
1459                 bna_rx_enet_stop(rx);
1460                 break;
1461 
1462         default:
1463                 bfa_sm_fault(event);
1464         }
1465 }
1466 
1467 static void
1468 bna_rx_sm_started_entry(struct bna_rx *rx)
1469 {
1470         struct bna_rxp *rxp;
1471         int is_regular = (rx->type == BNA_RX_T_REGULAR);
1472 
1473         /* Start IB */
1474         list_for_each_entry(rxp, &rx->rxp_q, qe)
1475                 bna_ib_start(rx->bna, &rxp->cq.ib, is_regular);
1476 
1477         bna_ethport_cb_rx_started(&rx->bna->ethport);
1478 }
1479 
1480 static void
1481 bna_rx_sm_started(struct bna_rx *rx, enum bna_rx_event event)
1482 {
1483         switch (event) {
1484         case RX_E_STOP:
1485                 bfa_fsm_set_state(rx, bna_rx_sm_rxf_stop_wait);
1486                 bna_ethport_cb_rx_stopped(&rx->bna->ethport);
1487                 bna_rxf_stop(&rx->rxf);
1488                 break;
1489 
1490         case RX_E_FAIL:
1491                 bfa_fsm_set_state(rx, bna_rx_sm_failed);
1492                 bna_ethport_cb_rx_stopped(&rx->bna->ethport);
1493                 bna_rxf_fail(&rx->rxf);
1494                 call_rx_stall_cbfn(rx);
1495                 rx->rx_cleanup_cbfn(rx->bna->bnad, rx);
1496                 break;
1497 
1498         default:
1499                 bfa_sm_fault(event);
1500                 break;
1501         }
1502 }
1503 
1504 static void bna_rx_sm_rxf_start_wait(struct bna_rx *rx,
1505                                 enum bna_rx_event event)
1506 {
1507         switch (event) {
1508         case RX_E_STOP:
1509                 bfa_fsm_set_state(rx, bna_rx_sm_rxf_stop_wait);
1510                 break;
1511 
1512         case RX_E_FAIL:
1513                 bfa_fsm_set_state(rx, bna_rx_sm_failed);
1514                 bna_rxf_fail(&rx->rxf);
1515                 call_rx_stall_cbfn(rx);
1516                 rx->rx_cleanup_cbfn(rx->bna->bnad, rx);
1517                 break;
1518 
1519         case RX_E_RXF_STARTED:
1520                 bfa_fsm_set_state(rx, bna_rx_sm_started);
1521                 break;
1522 
1523         default:
1524                 bfa_sm_fault(event);
1525                 break;
1526         }
1527 }
1528 
1529 static void
1530 bna_rx_sm_cleanup_wait_entry(struct bna_rx *rx)
1531 {
1532 }
1533 
1534 static void
1535 bna_rx_sm_cleanup_wait(struct bna_rx *rx, enum bna_rx_event event)
1536 {
1537         switch (event) {
1538         case RX_E_FAIL:
1539         case RX_E_RXF_STOPPED:
1540                 /* No-op */
1541                 break;
1542 
1543         case RX_E_CLEANUP_DONE:
1544                 bfa_fsm_set_state(rx, bna_rx_sm_stopped);
1545                 break;
1546 
1547         default:
1548                 bfa_sm_fault(event);
1549                 break;
1550         }
1551 }
1552 
1553 static void
1554 bna_rx_sm_failed_entry(struct bna_rx *rx)
1555 {
1556 }
1557 
1558 static void
1559 bna_rx_sm_failed(struct bna_rx *rx, enum bna_rx_event event)
1560 {
1561         switch (event) {
1562         case RX_E_START:
1563                 bfa_fsm_set_state(rx, bna_rx_sm_quiesce_wait);
1564                 break;
1565 
1566         case RX_E_STOP:
1567                 bfa_fsm_set_state(rx, bna_rx_sm_cleanup_wait);
1568                 break;
1569 
1570         case RX_E_FAIL:
1571         case RX_E_RXF_STARTED:
1572         case RX_E_RXF_STOPPED:
1573                 /* No-op */
1574                 break;
1575 
1576         case RX_E_CLEANUP_DONE:
1577                 bfa_fsm_set_state(rx, bna_rx_sm_stopped);
1578                 break;
1579 
1580         default:
1581                 bfa_sm_fault(event);
1582                 break;
1583 }       }
1584 
1585 static void
1586 bna_rx_sm_quiesce_wait_entry(struct bna_rx *rx)
1587 {
1588 }
1589 
1590 static void
1591 bna_rx_sm_quiesce_wait(struct bna_rx *rx, enum bna_rx_event event)
1592 {
1593         switch (event) {
1594         case RX_E_STOP:
1595                 bfa_fsm_set_state(rx, bna_rx_sm_cleanup_wait);
1596                 break;
1597 
1598         case RX_E_FAIL:
1599                 bfa_fsm_set_state(rx, bna_rx_sm_failed);
1600                 break;
1601 
1602         case RX_E_CLEANUP_DONE:
1603                 bfa_fsm_set_state(rx, bna_rx_sm_start_wait);
1604                 break;
1605 
1606         default:
1607                 bfa_sm_fault(event);
1608                 break;
1609         }
1610 }
1611 
1612 static void
1613 bna_bfi_rx_enet_start(struct bna_rx *rx)
1614 {
1615         struct bfi_enet_rx_cfg_req *cfg_req = &rx->bfi_enet_cmd.cfg_req;
1616         struct bna_rxp *rxp = NULL;
1617         struct bna_rxq *q0 = NULL, *q1 = NULL;
1618         int i;
1619 
1620         bfi_msgq_mhdr_set(cfg_req->mh, BFI_MC_ENET,
1621                 BFI_ENET_H2I_RX_CFG_SET_REQ, 0, rx->rid);
1622         cfg_req->mh.num_entries = htons(
1623                 bfi_msgq_num_cmd_entries(sizeof(struct bfi_enet_rx_cfg_req)));
1624 
1625         cfg_req->rx_cfg.frame_size = bna_enet_mtu_get(&rx->bna->enet);
1626         cfg_req->num_queue_sets = rx->num_paths;
1627         for (i = 0; i < rx->num_paths; i++) {
1628                 rxp = rxp ? list_next_entry(rxp, qe)
1629                         : list_first_entry(&rx->rxp_q, struct bna_rxp, qe);
1630                 GET_RXQS(rxp, q0, q1);
1631                 switch (rxp->type) {
1632                 case BNA_RXP_SLR:
1633                 case BNA_RXP_HDS:
1634                         /* Small RxQ */
1635                         bfi_enet_datapath_q_init(&cfg_req->q_cfg[i].qs.q,
1636                                                 &q1->qpt);
1637                         cfg_req->q_cfg[i].qs.rx_buffer_size =
1638                                 htons((u16)q1->buffer_size);
1639                         /* Fall through */
1640 
1641                 case BNA_RXP_SINGLE:
1642                         /* Large/Single RxQ */
1643                         bfi_enet_datapath_q_init(&cfg_req->q_cfg[i].ql.q,
1644                                                 &q0->qpt);
1645                         if (q0->multi_buffer)
1646                                 /* multi-buffer is enabled by allocating
1647                                  * a new rx with new set of resources.
1648                                  * q0->buffer_size should be initialized to
1649                                  * fragment size.
1650                                  */
1651                                 cfg_req->rx_cfg.multi_buffer =
1652                                         BNA_STATUS_T_ENABLED;
1653                         else
1654                                 q0->buffer_size =
1655                                         bna_enet_mtu_get(&rx->bna->enet);
1656                         cfg_req->q_cfg[i].ql.rx_buffer_size =
1657                                 htons((u16)q0->buffer_size);
1658                         break;
1659 
1660                 default:
1661                         BUG_ON(1);
1662                 }
1663 
1664                 bfi_enet_datapath_q_init(&cfg_req->q_cfg[i].cq.q,
1665                                         &rxp->cq.qpt);
1666 
1667                 cfg_req->q_cfg[i].ib.index_addr.a32.addr_lo =
1668                         rxp->cq.ib.ib_seg_host_addr.lsb;
1669                 cfg_req->q_cfg[i].ib.index_addr.a32.addr_hi =
1670                         rxp->cq.ib.ib_seg_host_addr.msb;
1671                 cfg_req->q_cfg[i].ib.intr.msix_index =
1672                         htons((u16)rxp->cq.ib.intr_vector);
1673         }
1674 
1675         cfg_req->ib_cfg.int_pkt_dma = BNA_STATUS_T_DISABLED;
1676         cfg_req->ib_cfg.int_enabled = BNA_STATUS_T_ENABLED;
1677         cfg_req->ib_cfg.int_pkt_enabled = BNA_STATUS_T_DISABLED;
1678         cfg_req->ib_cfg.continuous_coalescing = BNA_STATUS_T_DISABLED;
1679         cfg_req->ib_cfg.msix = (rxp->cq.ib.intr_type == BNA_INTR_T_MSIX)
1680                                 ? BNA_STATUS_T_ENABLED :
1681                                 BNA_STATUS_T_DISABLED;
1682         cfg_req->ib_cfg.coalescing_timeout =
1683                         htonl((u32)rxp->cq.ib.coalescing_timeo);
1684         cfg_req->ib_cfg.inter_pkt_timeout =
1685                         htonl((u32)rxp->cq.ib.interpkt_timeo);
1686         cfg_req->ib_cfg.inter_pkt_count = (u8)rxp->cq.ib.interpkt_count;
1687 
1688         switch (rxp->type) {
1689         case BNA_RXP_SLR:
1690                 cfg_req->rx_cfg.rxq_type = BFI_ENET_RXQ_LARGE_SMALL;
1691                 break;
1692 
1693         case BNA_RXP_HDS:
1694                 cfg_req->rx_cfg.rxq_type = BFI_ENET_RXQ_HDS;
1695                 cfg_req->rx_cfg.hds.type = rx->hds_cfg.hdr_type;
1696                 cfg_req->rx_cfg.hds.force_offset = rx->hds_cfg.forced_offset;
1697                 cfg_req->rx_cfg.hds.max_header_size = rx->hds_cfg.forced_offset;
1698                 break;
1699 
1700         case BNA_RXP_SINGLE:
1701                 cfg_req->rx_cfg.rxq_type = BFI_ENET_RXQ_SINGLE;
1702                 break;
1703 
1704         default:
1705                 BUG_ON(1);
1706         }
1707         cfg_req->rx_cfg.strip_vlan = rx->rxf.vlan_strip_status;
1708 
1709         bfa_msgq_cmd_set(&rx->msgq_cmd, NULL, NULL,
1710                 sizeof(struct bfi_enet_rx_cfg_req), &cfg_req->mh);
1711         bfa_msgq_cmd_post(&rx->bna->msgq, &rx->msgq_cmd);
1712 }
1713 
1714 static void
1715 bna_bfi_rx_enet_stop(struct bna_rx *rx)
1716 {
1717         struct bfi_enet_req *req = &rx->bfi_enet_cmd.req;
1718 
1719         bfi_msgq_mhdr_set(req->mh, BFI_MC_ENET,
1720                 BFI_ENET_H2I_RX_CFG_CLR_REQ, 0, rx->rid);
1721         req->mh.num_entries = htons(
1722                 bfi_msgq_num_cmd_entries(sizeof(struct bfi_enet_req)));
1723         bfa_msgq_cmd_set(&rx->msgq_cmd, NULL, NULL, sizeof(struct bfi_enet_req),
1724                 &req->mh);
1725         bfa_msgq_cmd_post(&rx->bna->msgq, &rx->msgq_cmd);
1726 }
1727 
1728 static void
1729 bna_rx_enet_stop(struct bna_rx *rx)
1730 {
1731         struct bna_rxp *rxp;
1732 
1733         /* Stop IB */
1734         list_for_each_entry(rxp, &rx->rxp_q, qe)
1735                 bna_ib_stop(rx->bna, &rxp->cq.ib);
1736 
1737         bna_bfi_rx_enet_stop(rx);
1738 }
1739 
1740 static int
1741 bna_rx_res_check(struct bna_rx_mod *rx_mod, struct bna_rx_config *rx_cfg)
1742 {
1743         if ((rx_mod->rx_free_count == 0) ||
1744                 (rx_mod->rxp_free_count == 0) ||
1745                 (rx_mod->rxq_free_count == 0))
1746                 return 0;
1747 
1748         if (rx_cfg->rxp_type == BNA_RXP_SINGLE) {
1749                 if ((rx_mod->rxp_free_count < rx_cfg->num_paths) ||
1750                         (rx_mod->rxq_free_count < rx_cfg->num_paths))
1751                                 return 0;
1752         } else {
1753                 if ((rx_mod->rxp_free_count < rx_cfg->num_paths) ||
1754                         (rx_mod->rxq_free_count < (2 * rx_cfg->num_paths)))
1755                         return 0;
1756         }
1757 
1758         return 1;
1759 }
1760 
1761 static struct bna_rxq *
1762 bna_rxq_get(struct bna_rx_mod *rx_mod)
1763 {
1764         struct bna_rxq *rxq = NULL;
1765 
1766         rxq = list_first_entry(&rx_mod->rxq_free_q, struct bna_rxq, qe);
1767         list_del(&rxq->qe);
1768         rx_mod->rxq_free_count--;
1769 
1770         return rxq;
1771 }
1772 
1773 static void
1774 bna_rxq_put(struct bna_rx_mod *rx_mod, struct bna_rxq *rxq)
1775 {
1776         list_add_tail(&rxq->qe, &rx_mod->rxq_free_q);
1777         rx_mod->rxq_free_count++;
1778 }
1779 
1780 static struct bna_rxp *
1781 bna_rxp_get(struct bna_rx_mod *rx_mod)
1782 {
1783         struct bna_rxp *rxp = NULL;
1784 
1785         rxp = list_first_entry(&rx_mod->rxp_free_q, struct bna_rxp, qe);
1786         list_del(&rxp->qe);
1787         rx_mod->rxp_free_count--;
1788 
1789         return rxp;
1790 }
1791 
1792 static void
1793 bna_rxp_put(struct bna_rx_mod *rx_mod, struct bna_rxp *rxp)
1794 {
1795         list_add_tail(&rxp->qe, &rx_mod->rxp_free_q);
1796         rx_mod->rxp_free_count++;
1797 }
1798 
1799 static struct bna_rx *
1800 bna_rx_get(struct bna_rx_mod *rx_mod, enum bna_rx_type type)
1801 {
1802         struct bna_rx *rx = NULL;
1803 
1804         BUG_ON(list_empty(&rx_mod->rx_free_q));
1805         if (type == BNA_RX_T_REGULAR)
1806                 rx = list_first_entry(&rx_mod->rx_free_q, struct bna_rx, qe);
1807         else
1808                 rx = list_last_entry(&rx_mod->rx_free_q, struct bna_rx, qe);
1809 
1810         rx_mod->rx_free_count--;
1811         list_move_tail(&rx->qe, &rx_mod->rx_active_q);
1812         rx->type = type;
1813 
1814         return rx;
1815 }
1816 
1817 static void
1818 bna_rx_put(struct bna_rx_mod *rx_mod, struct bna_rx *rx)
1819 {
1820         struct list_head *qe;
1821 
1822         list_for_each_prev(qe, &rx_mod->rx_free_q)
1823                 if (((struct bna_rx *)qe)->rid < rx->rid)
1824                         break;
1825 
1826         list_add(&rx->qe, qe);
1827         rx_mod->rx_free_count++;
1828 }
1829 
1830 static void
1831 bna_rxp_add_rxqs(struct bna_rxp *rxp, struct bna_rxq *q0,
1832                 struct bna_rxq *q1)
1833 {
1834         switch (rxp->type) {
1835         case BNA_RXP_SINGLE:
1836                 rxp->rxq.single.only = q0;
1837                 rxp->rxq.single.reserved = NULL;
1838                 break;
1839         case BNA_RXP_SLR:
1840                 rxp->rxq.slr.large = q0;
1841                 rxp->rxq.slr.small = q1;
1842                 break;
1843         case BNA_RXP_HDS:
1844                 rxp->rxq.hds.data = q0;
1845                 rxp->rxq.hds.hdr = q1;
1846                 break;
1847         default:
1848                 break;
1849         }
1850 }
1851 
1852 static void
1853 bna_rxq_qpt_setup(struct bna_rxq *rxq,
1854                 struct bna_rxp *rxp,
1855                 u32 page_count,
1856                 u32 page_size,
1857                 struct bna_mem_descr *qpt_mem,
1858                 struct bna_mem_descr *swqpt_mem,
1859                 struct bna_mem_descr *page_mem)
1860 {
1861         u8 *kva;
1862         u64 dma;
1863         struct bna_dma_addr bna_dma;
1864         int     i;
1865 
1866         rxq->qpt.hw_qpt_ptr.lsb = qpt_mem->dma.lsb;
1867         rxq->qpt.hw_qpt_ptr.msb = qpt_mem->dma.msb;
1868         rxq->qpt.kv_qpt_ptr = qpt_mem->kva;
1869         rxq->qpt.page_count = page_count;
1870         rxq->qpt.page_size = page_size;
1871 
1872         rxq->rcb->sw_qpt = (void **) swqpt_mem->kva;
1873         rxq->rcb->sw_q = page_mem->kva;
1874 
1875         kva = page_mem->kva;
1876         BNA_GET_DMA_ADDR(&page_mem->dma, dma);
1877 
1878         for (i = 0; i < rxq->qpt.page_count; i++) {
1879                 rxq->rcb->sw_qpt[i] = kva;
1880                 kva += PAGE_SIZE;
1881 
1882                 BNA_SET_DMA_ADDR(dma, &bna_dma);
1883                 ((struct bna_dma_addr *)rxq->qpt.kv_qpt_ptr)[i].lsb =
1884                         bna_dma.lsb;
1885                 ((struct bna_dma_addr *)rxq->qpt.kv_qpt_ptr)[i].msb =
1886                         bna_dma.msb;
1887                 dma += PAGE_SIZE;
1888         }
1889 }
1890 
1891 static void
1892 bna_rxp_cqpt_setup(struct bna_rxp *rxp,
1893                 u32 page_count,
1894                 u32 page_size,
1895                 struct bna_mem_descr *qpt_mem,
1896                 struct bna_mem_descr *swqpt_mem,
1897                 struct bna_mem_descr *page_mem)
1898 {
1899         u8 *kva;
1900         u64 dma;
1901         struct bna_dma_addr bna_dma;
1902         int     i;
1903 
1904         rxp->cq.qpt.hw_qpt_ptr.lsb = qpt_mem->dma.lsb;
1905         rxp->cq.qpt.hw_qpt_ptr.msb = qpt_mem->dma.msb;
1906         rxp->cq.qpt.kv_qpt_ptr = qpt_mem->kva;
1907         rxp->cq.qpt.page_count = page_count;
1908         rxp->cq.qpt.page_size = page_size;
1909 
1910         rxp->cq.ccb->sw_qpt = (void **) swqpt_mem->kva;
1911         rxp->cq.ccb->sw_q = page_mem->kva;
1912 
1913         kva = page_mem->kva;
1914         BNA_GET_DMA_ADDR(&page_mem->dma, dma);
1915 
1916         for (i = 0; i < rxp->cq.qpt.page_count; i++) {
1917                 rxp->cq.ccb->sw_qpt[i] = kva;
1918                 kva += PAGE_SIZE;
1919 
1920                 BNA_SET_DMA_ADDR(dma, &bna_dma);
1921                 ((struct bna_dma_addr *)rxp->cq.qpt.kv_qpt_ptr)[i].lsb =
1922                         bna_dma.lsb;
1923                 ((struct bna_dma_addr *)rxp->cq.qpt.kv_qpt_ptr)[i].msb =
1924                         bna_dma.msb;
1925                 dma += PAGE_SIZE;
1926         }
1927 }
1928 
1929 static void
1930 bna_rx_mod_cb_rx_stopped(void *arg, struct bna_rx *rx)
1931 {
1932         struct bna_rx_mod *rx_mod = (struct bna_rx_mod *)arg;
1933 
1934         bfa_wc_down(&rx_mod->rx_stop_wc);
1935 }
1936 
1937 static void
1938 bna_rx_mod_cb_rx_stopped_all(void *arg)
1939 {
1940         struct bna_rx_mod *rx_mod = (struct bna_rx_mod *)arg;
1941 
1942         if (rx_mod->stop_cbfn)
1943                 rx_mod->stop_cbfn(&rx_mod->bna->enet);
1944         rx_mod->stop_cbfn = NULL;
1945 }
1946 
1947 static void
1948 bna_rx_start(struct bna_rx *rx)
1949 {
1950         rx->rx_flags |= BNA_RX_F_ENET_STARTED;
1951         if (rx->rx_flags & BNA_RX_F_ENABLED)
1952                 bfa_fsm_send_event(rx, RX_E_START);
1953 }
1954 
1955 static void
1956 bna_rx_stop(struct bna_rx *rx)
1957 {
1958         rx->rx_flags &= ~BNA_RX_F_ENET_STARTED;
1959         if (rx->fsm == (bfa_fsm_t) bna_rx_sm_stopped)
1960                 bna_rx_mod_cb_rx_stopped(&rx->bna->rx_mod, rx);
1961         else {
1962                 rx->stop_cbfn = bna_rx_mod_cb_rx_stopped;
1963                 rx->stop_cbarg = &rx->bna->rx_mod;
1964                 bfa_fsm_send_event(rx, RX_E_STOP);
1965         }
1966 }
1967 
1968 static void
1969 bna_rx_fail(struct bna_rx *rx)
1970 {
1971         /* Indicate Enet is not enabled, and failed */
1972         rx->rx_flags &= ~BNA_RX_F_ENET_STARTED;
1973         bfa_fsm_send_event(rx, RX_E_FAIL);
1974 }
1975 
1976 void
1977 bna_rx_mod_start(struct bna_rx_mod *rx_mod, enum bna_rx_type type)
1978 {
1979         struct bna_rx *rx;
1980 
1981         rx_mod->flags |= BNA_RX_MOD_F_ENET_STARTED;
1982         if (type == BNA_RX_T_LOOPBACK)
1983                 rx_mod->flags |= BNA_RX_MOD_F_ENET_LOOPBACK;
1984 
1985         list_for_each_entry(rx, &rx_mod->rx_active_q, qe)
1986                 if (rx->type == type)
1987                         bna_rx_start(rx);
1988 }
1989 
1990 void
1991 bna_rx_mod_stop(struct bna_rx_mod *rx_mod, enum bna_rx_type type)
1992 {
1993         struct bna_rx *rx;
1994 
1995         rx_mod->flags &= ~BNA_RX_MOD_F_ENET_STARTED;
1996         rx_mod->flags &= ~BNA_RX_MOD_F_ENET_LOOPBACK;
1997 
1998         rx_mod->stop_cbfn = bna_enet_cb_rx_stopped;
1999 
2000         bfa_wc_init(&rx_mod->rx_stop_wc, bna_rx_mod_cb_rx_stopped_all, rx_mod);
2001 
2002         list_for_each_entry(rx, &rx_mod->rx_active_q, qe)
2003                 if (rx->type == type) {
2004                         bfa_wc_up(&rx_mod->rx_stop_wc);
2005                         bna_rx_stop(rx);
2006                 }
2007 
2008         bfa_wc_wait(&rx_mod->rx_stop_wc);
2009 }
2010 
2011 void
2012 bna_rx_mod_fail(struct bna_rx_mod *rx_mod)
2013 {
2014         struct bna_rx *rx;
2015 
2016         rx_mod->flags &= ~BNA_RX_MOD_F_ENET_STARTED;
2017         rx_mod->flags &= ~BNA_RX_MOD_F_ENET_LOOPBACK;
2018 
2019         list_for_each_entry(rx, &rx_mod->rx_active_q, qe)
2020                 bna_rx_fail(rx);
2021 }
2022 
2023 void bna_rx_mod_init(struct bna_rx_mod *rx_mod, struct bna *bna,
2024                         struct bna_res_info *res_info)
2025 {
2026         int     index;
2027         struct bna_rx *rx_ptr;
2028         struct bna_rxp *rxp_ptr;
2029         struct bna_rxq *rxq_ptr;
2030 
2031         rx_mod->bna = bna;
2032         rx_mod->flags = 0;
2033 
2034         rx_mod->rx = (struct bna_rx *)
2035                 res_info[BNA_MOD_RES_MEM_T_RX_ARRAY].res_u.mem_info.mdl[0].kva;
2036         rx_mod->rxp = (struct bna_rxp *)
2037                 res_info[BNA_MOD_RES_MEM_T_RXP_ARRAY].res_u.mem_info.mdl[0].kva;
2038         rx_mod->rxq = (struct bna_rxq *)
2039                 res_info[BNA_MOD_RES_MEM_T_RXQ_ARRAY].res_u.mem_info.mdl[0].kva;
2040 
2041         /* Initialize the queues */
2042         INIT_LIST_HEAD(&rx_mod->rx_free_q);
2043         rx_mod->rx_free_count = 0;
2044         INIT_LIST_HEAD(&rx_mod->rxq_free_q);
2045         rx_mod->rxq_free_count = 0;
2046         INIT_LIST_HEAD(&rx_mod->rxp_free_q);
2047         rx_mod->rxp_free_count = 0;
2048         INIT_LIST_HEAD(&rx_mod->rx_active_q);
2049 
2050         /* Build RX queues */
2051         for (index = 0; index < bna->ioceth.attr.num_rxp; index++) {
2052                 rx_ptr = &rx_mod->rx[index];
2053 
2054                 INIT_LIST_HEAD(&rx_ptr->rxp_q);
2055                 rx_ptr->bna = NULL;
2056                 rx_ptr->rid = index;
2057                 rx_ptr->stop_cbfn = NULL;
2058                 rx_ptr->stop_cbarg = NULL;
2059 
2060                 list_add_tail(&rx_ptr->qe, &rx_mod->rx_free_q);
2061                 rx_mod->rx_free_count++;
2062         }
2063 
2064         /* build RX-path queue */
2065         for (index = 0; index < bna->ioceth.attr.num_rxp; index++) {
2066                 rxp_ptr = &rx_mod->rxp[index];
2067                 list_add_tail(&rxp_ptr->qe, &rx_mod->rxp_free_q);
2068                 rx_mod->rxp_free_count++;
2069         }
2070 
2071         /* build RXQ queue */
2072         for (index = 0; index < (bna->ioceth.attr.num_rxp * 2); index++) {
2073                 rxq_ptr = &rx_mod->rxq[index];
2074                 list_add_tail(&rxq_ptr->qe, &rx_mod->rxq_free_q);
2075                 rx_mod->rxq_free_count++;
2076         }
2077 }
2078 
2079 void
2080 bna_rx_mod_uninit(struct bna_rx_mod *rx_mod)
2081 {
2082         rx_mod->bna = NULL;
2083 }
2084 
2085 void
2086 bna_bfi_rx_enet_start_rsp(struct bna_rx *rx, struct bfi_msgq_mhdr *msghdr)
2087 {
2088         struct bfi_enet_rx_cfg_rsp *cfg_rsp = &rx->bfi_enet_cmd.cfg_rsp;
2089         struct bna_rxp *rxp = NULL;
2090         struct bna_rxq *q0 = NULL, *q1 = NULL;
2091         int i;
2092 
2093         bfa_msgq_rsp_copy(&rx->bna->msgq, (u8 *)cfg_rsp,
2094                 sizeof(struct bfi_enet_rx_cfg_rsp));
2095 
2096         rx->hw_id = cfg_rsp->hw_id;
2097 
2098         for (i = 0, rxp = list_first_entry(&rx->rxp_q, struct bna_rxp, qe);
2099              i < rx->num_paths; i++, rxp = list_next_entry(rxp, qe)) {
2100                 GET_RXQS(rxp, q0, q1);
2101 
2102                 /* Setup doorbells */
2103                 rxp->cq.ccb->i_dbell->doorbell_addr =
2104                         rx->bna->pcidev.pci_bar_kva
2105                         + ntohl(cfg_rsp->q_handles[i].i_dbell);
2106                 rxp->hw_id = cfg_rsp->q_handles[i].hw_cqid;
2107                 q0->rcb->q_dbell =
2108                         rx->bna->pcidev.pci_bar_kva
2109                         + ntohl(cfg_rsp->q_handles[i].ql_dbell);
2110                 q0->hw_id = cfg_rsp->q_handles[i].hw_lqid;
2111                 if (q1) {
2112                         q1->rcb->q_dbell =
2113                         rx->bna->pcidev.pci_bar_kva
2114                         + ntohl(cfg_rsp->q_handles[i].qs_dbell);
2115                         q1->hw_id = cfg_rsp->q_handles[i].hw_sqid;
2116                 }
2117 
2118                 /* Initialize producer/consumer indexes */
2119                 (*rxp->cq.ccb->hw_producer_index) = 0;
2120                 rxp->cq.ccb->producer_index = 0;
2121                 q0->rcb->producer_index = q0->rcb->consumer_index = 0;
2122                 if (q1)
2123                         q1->rcb->producer_index = q1->rcb->consumer_index = 0;
2124         }
2125 
2126         bfa_fsm_send_event(rx, RX_E_STARTED);
2127 }
2128 
2129 void
2130 bna_bfi_rx_enet_stop_rsp(struct bna_rx *rx, struct bfi_msgq_mhdr *msghdr)
2131 {
2132         bfa_fsm_send_event(rx, RX_E_STOPPED);
2133 }
2134 
2135 void
2136 bna_rx_res_req(struct bna_rx_config *q_cfg, struct bna_res_info *res_info)
2137 {
2138         u32 cq_size, hq_size, dq_size;
2139         u32 cpage_count, hpage_count, dpage_count;
2140         struct bna_mem_info *mem_info;
2141         u32 cq_depth;
2142         u32 hq_depth;
2143         u32 dq_depth;
2144 
2145         dq_depth = q_cfg->q0_depth;
2146         hq_depth = ((q_cfg->rxp_type == BNA_RXP_SINGLE) ? 0 : q_cfg->q1_depth);
2147         cq_depth = roundup_pow_of_two(dq_depth + hq_depth);
2148 
2149         cq_size = cq_depth * BFI_CQ_WI_SIZE;
2150         cq_size = ALIGN(cq_size, PAGE_SIZE);
2151         cpage_count = SIZE_TO_PAGES(cq_size);
2152 
2153         dq_depth = roundup_pow_of_two(dq_depth);
2154         dq_size = dq_depth * BFI_RXQ_WI_SIZE;
2155         dq_size = ALIGN(dq_size, PAGE_SIZE);
2156         dpage_count = SIZE_TO_PAGES(dq_size);
2157 
2158         if (BNA_RXP_SINGLE != q_cfg->rxp_type) {
2159                 hq_depth = roundup_pow_of_two(hq_depth);
2160                 hq_size = hq_depth * BFI_RXQ_WI_SIZE;
2161                 hq_size = ALIGN(hq_size, PAGE_SIZE);
2162                 hpage_count = SIZE_TO_PAGES(hq_size);
2163         } else
2164                 hpage_count = 0;
2165 
2166         res_info[BNA_RX_RES_MEM_T_CCB].res_type = BNA_RES_T_MEM;
2167         mem_info = &res_info[BNA_RX_RES_MEM_T_CCB].res_u.mem_info;
2168         mem_info->mem_type = BNA_MEM_T_KVA;
2169         mem_info->len = sizeof(struct bna_ccb);
2170         mem_info->num = q_cfg->num_paths;
2171 
2172         res_info[BNA_RX_RES_MEM_T_RCB].res_type = BNA_RES_T_MEM;
2173         mem_info = &res_info[BNA_RX_RES_MEM_T_RCB].res_u.mem_info;
2174         mem_info->mem_type = BNA_MEM_T_KVA;
2175         mem_info->len = sizeof(struct bna_rcb);
2176         mem_info->num = BNA_GET_RXQS(q_cfg);
2177 
2178         res_info[BNA_RX_RES_MEM_T_CQPT].res_type = BNA_RES_T_MEM;
2179         mem_info = &res_info[BNA_RX_RES_MEM_T_CQPT].res_u.mem_info;
2180         mem_info->mem_type = BNA_MEM_T_DMA;
2181         mem_info->len = cpage_count * sizeof(struct bna_dma_addr);
2182         mem_info->num = q_cfg->num_paths;
2183 
2184         res_info[BNA_RX_RES_MEM_T_CSWQPT].res_type = BNA_RES_T_MEM;
2185         mem_info = &res_info[BNA_RX_RES_MEM_T_CSWQPT].res_u.mem_info;
2186         mem_info->mem_type = BNA_MEM_T_KVA;
2187         mem_info->len = cpage_count * sizeof(void *);
2188         mem_info->num = q_cfg->num_paths;
2189 
2190         res_info[BNA_RX_RES_MEM_T_CQPT_PAGE].res_type = BNA_RES_T_MEM;
2191         mem_info = &res_info[BNA_RX_RES_MEM_T_CQPT_PAGE].res_u.mem_info;
2192         mem_info->mem_type = BNA_MEM_T_DMA;
2193         mem_info->len = PAGE_SIZE * cpage_count;
2194         mem_info->num = q_cfg->num_paths;
2195 
2196         res_info[BNA_RX_RES_MEM_T_DQPT].res_type = BNA_RES_T_MEM;
2197         mem_info = &res_info[BNA_RX_RES_MEM_T_DQPT].res_u.mem_info;
2198         mem_info->mem_type = BNA_MEM_T_DMA;
2199         mem_info->len = dpage_count * sizeof(struct bna_dma_addr);
2200         mem_info->num = q_cfg->num_paths;
2201 
2202         res_info[BNA_RX_RES_MEM_T_DSWQPT].res_type = BNA_RES_T_MEM;
2203         mem_info = &res_info[BNA_RX_RES_MEM_T_DSWQPT].res_u.mem_info;
2204         mem_info->mem_type = BNA_MEM_T_KVA;
2205         mem_info->len = dpage_count * sizeof(void *);
2206         mem_info->num = q_cfg->num_paths;
2207 
2208         res_info[BNA_RX_RES_MEM_T_DPAGE].res_type = BNA_RES_T_MEM;
2209         mem_info = &res_info[BNA_RX_RES_MEM_T_DPAGE].res_u.mem_info;
2210         mem_info->mem_type = BNA_MEM_T_DMA;
2211         mem_info->len = PAGE_SIZE * dpage_count;
2212         mem_info->num = q_cfg->num_paths;
2213 
2214         res_info[BNA_RX_RES_MEM_T_HQPT].res_type = BNA_RES_T_MEM;
2215         mem_info = &res_info[BNA_RX_RES_MEM_T_HQPT].res_u.mem_info;
2216         mem_info->mem_type = BNA_MEM_T_DMA;
2217         mem_info->len = hpage_count * sizeof(struct bna_dma_addr);
2218         mem_info->num = (hpage_count ? q_cfg->num_paths : 0);
2219 
2220         res_info[BNA_RX_RES_MEM_T_HSWQPT].res_type = BNA_RES_T_MEM;
2221         mem_info = &res_info[BNA_RX_RES_MEM_T_HSWQPT].res_u.mem_info;
2222         mem_info->mem_type = BNA_MEM_T_KVA;
2223         mem_info->len = hpage_count * sizeof(void *);
2224         mem_info->num = (hpage_count ? q_cfg->num_paths : 0);
2225 
2226         res_info[BNA_RX_RES_MEM_T_HPAGE].res_type = BNA_RES_T_MEM;
2227         mem_info = &res_info[BNA_RX_RES_MEM_T_HPAGE].res_u.mem_info;
2228         mem_info->mem_type = BNA_MEM_T_DMA;
2229         mem_info->len = PAGE_SIZE * hpage_count;
2230         mem_info->num = (hpage_count ? q_cfg->num_paths : 0);
2231 
2232         res_info[BNA_RX_RES_MEM_T_IBIDX].res_type = BNA_RES_T_MEM;
2233         mem_info = &res_info[BNA_RX_RES_MEM_T_IBIDX].res_u.mem_info;
2234         mem_info->mem_type = BNA_MEM_T_DMA;
2235         mem_info->len = BFI_IBIDX_SIZE;
2236         mem_info->num = q_cfg->num_paths;
2237 
2238         res_info[BNA_RX_RES_MEM_T_RIT].res_type = BNA_RES_T_MEM;
2239         mem_info = &res_info[BNA_RX_RES_MEM_T_RIT].res_u.mem_info;
2240         mem_info->mem_type = BNA_MEM_T_KVA;
2241         mem_info->len = BFI_ENET_RSS_RIT_MAX;
2242         mem_info->num = 1;
2243 
2244         res_info[BNA_RX_RES_T_INTR].res_type = BNA_RES_T_INTR;
2245         res_info[BNA_RX_RES_T_INTR].res_u.intr_info.intr_type = BNA_INTR_T_MSIX;
2246         res_info[BNA_RX_RES_T_INTR].res_u.intr_info.num = q_cfg->num_paths;
2247 }
2248 
2249 struct bna_rx *
2250 bna_rx_create(struct bna *bna, struct bnad *bnad,
2251                 struct bna_rx_config *rx_cfg,
2252                 const struct bna_rx_event_cbfn *rx_cbfn,
2253                 struct bna_res_info *res_info,
2254                 void *priv)
2255 {
2256         struct bna_rx_mod *rx_mod = &bna->rx_mod;
2257         struct bna_rx *rx;
2258         struct bna_rxp *rxp;
2259         struct bna_rxq *q0;
2260         struct bna_rxq *q1;
2261         struct bna_intr_info *intr_info;
2262         struct bna_mem_descr *hqunmap_mem;
2263         struct bna_mem_descr *dqunmap_mem;
2264         struct bna_mem_descr *ccb_mem;
2265         struct bna_mem_descr *rcb_mem;
2266         struct bna_mem_descr *cqpt_mem;
2267         struct bna_mem_descr *cswqpt_mem;
2268         struct bna_mem_descr *cpage_mem;
2269         struct bna_mem_descr *hqpt_mem;
2270         struct bna_mem_descr *dqpt_mem;
2271         struct bna_mem_descr *hsqpt_mem;
2272         struct bna_mem_descr *dsqpt_mem;
2273         struct bna_mem_descr *hpage_mem;
2274         struct bna_mem_descr *dpage_mem;
2275         u32 dpage_count, hpage_count;
2276         u32 hq_idx, dq_idx, rcb_idx;
2277         u32 cq_depth, i;
2278         u32 page_count;
2279 
2280         if (!bna_rx_res_check(rx_mod, rx_cfg))
2281                 return NULL;
2282 
2283         intr_info = &res_info[BNA_RX_RES_T_INTR].res_u.intr_info;
2284         ccb_mem = &res_info[BNA_RX_RES_MEM_T_CCB].res_u.mem_info.mdl[0];
2285         rcb_mem = &res_info[BNA_RX_RES_MEM_T_RCB].res_u.mem_info.mdl[0];
2286         dqunmap_mem = &res_info[BNA_RX_RES_MEM_T_UNMAPDQ].res_u.mem_info.mdl[0];
2287         hqunmap_mem = &res_info[BNA_RX_RES_MEM_T_UNMAPHQ].res_u.mem_info.mdl[0];
2288         cqpt_mem = &res_info[BNA_RX_RES_MEM_T_CQPT].res_u.mem_info.mdl[0];
2289         cswqpt_mem = &res_info[BNA_RX_RES_MEM_T_CSWQPT].res_u.mem_info.mdl[0];
2290         cpage_mem = &res_info[BNA_RX_RES_MEM_T_CQPT_PAGE].res_u.mem_info.mdl[0];
2291         hqpt_mem = &res_info[BNA_RX_RES_MEM_T_HQPT].res_u.mem_info.mdl[0];
2292         dqpt_mem = &res_info[BNA_RX_RES_MEM_T_DQPT].res_u.mem_info.mdl[0];
2293         hsqpt_mem = &res_info[BNA_RX_RES_MEM_T_HSWQPT].res_u.mem_info.mdl[0];
2294         dsqpt_mem = &res_info[BNA_RX_RES_MEM_T_DSWQPT].res_u.mem_info.mdl[0];
2295         hpage_mem = &res_info[BNA_RX_RES_MEM_T_HPAGE].res_u.mem_info.mdl[0];
2296         dpage_mem = &res_info[BNA_RX_RES_MEM_T_DPAGE].res_u.mem_info.mdl[0];
2297 
2298         page_count = res_info[BNA_RX_RES_MEM_T_CQPT_PAGE].res_u.mem_info.len /
2299                         PAGE_SIZE;
2300 
2301         dpage_count = res_info[BNA_RX_RES_MEM_T_DPAGE].res_u.mem_info.len /
2302                         PAGE_SIZE;
2303 
2304         hpage_count = res_info[BNA_RX_RES_MEM_T_HPAGE].res_u.mem_info.len /
2305                         PAGE_SIZE;
2306 
2307         rx = bna_rx_get(rx_mod, rx_cfg->rx_type);
2308         rx->bna = bna;
2309         rx->rx_flags = 0;
2310         INIT_LIST_HEAD(&rx->rxp_q);
2311         rx->stop_cbfn = NULL;
2312         rx->stop_cbarg = NULL;
2313         rx->priv = priv;
2314 
2315         rx->rcb_setup_cbfn = rx_cbfn->rcb_setup_cbfn;
2316         rx->rcb_destroy_cbfn = rx_cbfn->rcb_destroy_cbfn;
2317         rx->ccb_setup_cbfn = rx_cbfn->ccb_setup_cbfn;
2318         rx->ccb_destroy_cbfn = rx_cbfn->ccb_destroy_cbfn;
2319         rx->rx_stall_cbfn = rx_cbfn->rx_stall_cbfn;
2320         /* Following callbacks are mandatory */
2321         rx->rx_cleanup_cbfn = rx_cbfn->rx_cleanup_cbfn;
2322         rx->rx_post_cbfn = rx_cbfn->rx_post_cbfn;
2323 
2324         if (rx->bna->rx_mod.flags & BNA_RX_MOD_F_ENET_STARTED) {
2325                 switch (rx->type) {
2326                 case BNA_RX_T_REGULAR:
2327                         if (!(rx->bna->rx_mod.flags &
2328                                 BNA_RX_MOD_F_ENET_LOOPBACK))
2329                                 rx->rx_flags |= BNA_RX_F_ENET_STARTED;
2330                         break;
2331                 case BNA_RX_T_LOOPBACK:
2332                         if (rx->bna->rx_mod.flags & BNA_RX_MOD_F_ENET_LOOPBACK)
2333                                 rx->rx_flags |= BNA_RX_F_ENET_STARTED;
2334                         break;
2335                 }
2336         }
2337 
2338         rx->num_paths = rx_cfg->num_paths;
2339         for (i = 0, hq_idx = 0, dq_idx = 0, rcb_idx = 0;
2340                         i < rx->num_paths; i++) {
2341                 rxp = bna_rxp_get(rx_mod);
2342                 list_add_tail(&rxp->qe, &rx->rxp_q);
2343                 rxp->type = rx_cfg->rxp_type;
2344                 rxp->rx = rx;
2345                 rxp->cq.rx = rx;
2346 
2347                 q0 = bna_rxq_get(rx_mod);
2348                 if (BNA_RXP_SINGLE == rx_cfg->rxp_type)
2349                         q1 = NULL;
2350                 else
2351                         q1 = bna_rxq_get(rx_mod);
2352 
2353                 if (1 == intr_info->num)
2354                         rxp->vector = intr_info->idl[0].vector;
2355                 else
2356                         rxp->vector = intr_info->idl[i].vector;
2357 
2358                 /* Setup IB */
2359 
2360                 rxp->cq.ib.ib_seg_host_addr.lsb =
2361                 res_info[BNA_RX_RES_MEM_T_IBIDX].res_u.mem_info.mdl[i].dma.lsb;
2362                 rxp->cq.ib.ib_seg_host_addr.msb =
2363                 res_info[BNA_RX_RES_MEM_T_IBIDX].res_u.mem_info.mdl[i].dma.msb;
2364                 rxp->cq.ib.ib_seg_host_addr_kva =
2365                 res_info[BNA_RX_RES_MEM_T_IBIDX].res_u.mem_info.mdl[i].kva;
2366                 rxp->cq.ib.intr_type = intr_info->intr_type;
2367                 if (intr_info->intr_type == BNA_INTR_T_MSIX)
2368                         rxp->cq.ib.intr_vector = rxp->vector;
2369                 else
2370                         rxp->cq.ib.intr_vector = BIT(rxp->vector);
2371                 rxp->cq.ib.coalescing_timeo = rx_cfg->coalescing_timeo;
2372                 rxp->cq.ib.interpkt_count = BFI_RX_INTERPKT_COUNT;
2373                 rxp->cq.ib.interpkt_timeo = BFI_RX_INTERPKT_TIMEO;
2374 
2375                 bna_rxp_add_rxqs(rxp, q0, q1);
2376 
2377                 /* Setup large Q */
2378 
2379                 q0->rx = rx;
2380                 q0->rxp = rxp;
2381 
2382                 q0->rcb = (struct bna_rcb *) rcb_mem[rcb_idx].kva;
2383                 q0->rcb->unmap_q = (void *)dqunmap_mem[dq_idx].kva;
2384                 rcb_idx++; dq_idx++;
2385                 q0->rcb->q_depth = rx_cfg->q0_depth;
2386                 q0->q_depth = rx_cfg->q0_depth;
2387                 q0->multi_buffer = rx_cfg->q0_multi_buf;
2388                 q0->buffer_size = rx_cfg->q0_buf_size;
2389                 q0->num_vecs = rx_cfg->q0_num_vecs;
2390                 q0->rcb->rxq = q0;
2391                 q0->rcb->bnad = bna->bnad;
2392                 q0->rcb->id = 0;
2393                 q0->rx_packets = q0->rx_bytes = 0;
2394                 q0->rx_packets_with_error = q0->rxbuf_alloc_failed = 0;
2395                 q0->rxbuf_map_failed = 0;
2396 
2397                 bna_rxq_qpt_setup(q0, rxp, dpage_count, PAGE_SIZE,
2398                         &dqpt_mem[i], &dsqpt_mem[i], &dpage_mem[i]);
2399 
2400                 if (rx->rcb_setup_cbfn)
2401                         rx->rcb_setup_cbfn(bnad, q0->rcb);
2402 
2403                 /* Setup small Q */
2404 
2405                 if (q1) {
2406                         q1->rx = rx;
2407                         q1->rxp = rxp;
2408 
2409                         q1->rcb = (struct bna_rcb *) rcb_mem[rcb_idx].kva;
2410                         q1->rcb->unmap_q = (void *)hqunmap_mem[hq_idx].kva;
2411                         rcb_idx++; hq_idx++;
2412                         q1->rcb->q_depth = rx_cfg->q1_depth;
2413                         q1->q_depth = rx_cfg->q1_depth;
2414                         q1->multi_buffer = BNA_STATUS_T_DISABLED;
2415                         q1->num_vecs = 1;
2416                         q1->rcb->rxq = q1;
2417                         q1->rcb->bnad = bna->bnad;
2418                         q1->rcb->id = 1;
2419                         q1->buffer_size = (rx_cfg->rxp_type == BNA_RXP_HDS) ?
2420                                         rx_cfg->hds_config.forced_offset
2421                                         : rx_cfg->q1_buf_size;
2422                         q1->rx_packets = q1->rx_bytes = 0;
2423                         q1->rx_packets_with_error = q1->rxbuf_alloc_failed = 0;
2424                         q1->rxbuf_map_failed = 0;
2425 
2426                         bna_rxq_qpt_setup(q1, rxp, hpage_count, PAGE_SIZE,
2427                                 &hqpt_mem[i], &hsqpt_mem[i],
2428                                 &hpage_mem[i]);
2429 
2430                         if (rx->rcb_setup_cbfn)
2431                                 rx->rcb_setup_cbfn(bnad, q1->rcb);
2432                 }
2433 
2434                 /* Setup CQ */
2435 
2436                 rxp->cq.ccb = (struct bna_ccb *) ccb_mem[i].kva;
2437                 cq_depth = rx_cfg->q0_depth +
2438                         ((rx_cfg->rxp_type == BNA_RXP_SINGLE) ?
2439                          0 : rx_cfg->q1_depth);
2440                 /* if multi-buffer is enabled sum of q0_depth
2441                  * and q1_depth need not be a power of 2
2442                  */
2443                 cq_depth = roundup_pow_of_two(cq_depth);
2444                 rxp->cq.ccb->q_depth = cq_depth;
2445                 rxp->cq.ccb->cq = &rxp->cq;
2446                 rxp->cq.ccb->rcb[0] = q0->rcb;
2447                 q0->rcb->ccb = rxp->cq.ccb;
2448                 if (q1) {
2449                         rxp->cq.ccb->rcb[1] = q1->rcb;
2450                         q1->rcb->ccb = rxp->cq.ccb;
2451                 }
2452                 rxp->cq.ccb->hw_producer_index =
2453                         (u32 *)rxp->cq.ib.ib_seg_host_addr_kva;
2454                 rxp->cq.ccb->i_dbell = &rxp->cq.ib.door_bell;
2455                 rxp->cq.ccb->intr_type = rxp->cq.ib.intr_type;
2456                 rxp->cq.ccb->intr_vector = rxp->cq.ib.intr_vector;
2457                 rxp->cq.ccb->rx_coalescing_timeo =
2458                         rxp->cq.ib.coalescing_timeo;
2459                 rxp->cq.ccb->pkt_rate.small_pkt_cnt = 0;
2460                 rxp->cq.ccb->pkt_rate.large_pkt_cnt = 0;
2461                 rxp->cq.ccb->bnad = bna->bnad;
2462                 rxp->cq.ccb->id = i;
2463 
2464                 bna_rxp_cqpt_setup(rxp, page_count, PAGE_SIZE,
2465                         &cqpt_mem[i], &cswqpt_mem[i], &cpage_mem[i]);
2466 
2467                 if (rx->ccb_setup_cbfn)
2468                         rx->ccb_setup_cbfn(bnad, rxp->cq.ccb);
2469         }
2470 
2471         rx->hds_cfg = rx_cfg->hds_config;
2472 
2473         bna_rxf_init(&rx->rxf, rx, rx_cfg, res_info);
2474 
2475         bfa_fsm_set_state(rx, bna_rx_sm_stopped);
2476 
2477         rx_mod->rid_mask |= BIT(rx->rid);
2478 
2479         return rx;
2480 }
2481 
2482 void
2483 bna_rx_destroy(struct bna_rx *rx)
2484 {
2485         struct bna_rx_mod *rx_mod = &rx->bna->rx_mod;
2486         struct bna_rxq *q0 = NULL;
2487         struct bna_rxq *q1 = NULL;
2488         struct bna_rxp *rxp;
2489         struct list_head *qe;
2490 
2491         bna_rxf_uninit(&rx->rxf);
2492 
2493         while (!list_empty(&rx->rxp_q)) {
2494                 rxp = list_first_entry(&rx->rxp_q, struct bna_rxp, qe);
2495                 list_del(&rxp->qe);
2496                 GET_RXQS(rxp, q0, q1);
2497                 if (rx->rcb_destroy_cbfn)
2498                         rx->rcb_destroy_cbfn(rx->bna->bnad, q0->rcb);
2499                 q0->rcb = NULL;
2500                 q0->rxp = NULL;
2501                 q0->rx = NULL;
2502                 bna_rxq_put(rx_mod, q0);
2503 
2504                 if (q1) {
2505                         if (rx->rcb_destroy_cbfn)
2506                                 rx->rcb_destroy_cbfn(rx->bna->bnad, q1->rcb);
2507                         q1->rcb = NULL;
2508                         q1->rxp = NULL;
2509                         q1->rx = NULL;
2510                         bna_rxq_put(rx_mod, q1);
2511                 }
2512                 rxp->rxq.slr.large = NULL;
2513                 rxp->rxq.slr.small = NULL;
2514 
2515                 if (rx->ccb_destroy_cbfn)
2516                         rx->ccb_destroy_cbfn(rx->bna->bnad, rxp->cq.ccb);
2517                 rxp->cq.ccb = NULL;
2518                 rxp->rx = NULL;
2519                 bna_rxp_put(rx_mod, rxp);
2520         }
2521 
2522         list_for_each(qe, &rx_mod->rx_active_q)
2523                 if (qe == &rx->qe) {
2524                         list_del(&rx->qe);
2525                         break;
2526                 }
2527 
2528         rx_mod->rid_mask &= ~BIT(rx->rid);
2529 
2530         rx->bna = NULL;
2531         rx->priv = NULL;
2532         bna_rx_put(rx_mod, rx);
2533 }
2534 
2535 void
2536 bna_rx_enable(struct bna_rx *rx)
2537 {
2538         if (rx->fsm != (bfa_sm_t)bna_rx_sm_stopped)
2539                 return;
2540 
2541         rx->rx_flags |= BNA_RX_F_ENABLED;
2542         if (rx->rx_flags & BNA_RX_F_ENET_STARTED)
2543                 bfa_fsm_send_event(rx, RX_E_START);
2544 }
2545 
2546 void
2547 bna_rx_disable(struct bna_rx *rx, enum bna_cleanup_type type,
2548                 void (*cbfn)(void *, struct bna_rx *))
2549 {
2550         if (type == BNA_SOFT_CLEANUP) {
2551                 /* h/w should not be accessed. Treat we're stopped */
2552                 (*cbfn)(rx->bna->bnad, rx);
2553         } else {
2554                 rx->stop_cbfn = cbfn;
2555                 rx->stop_cbarg = rx->bna->bnad;
2556 
2557                 rx->rx_flags &= ~BNA_RX_F_ENABLED;
2558 
2559                 bfa_fsm_send_event(rx, RX_E_STOP);
2560         }
2561 }
2562 
2563 void
2564 bna_rx_cleanup_complete(struct bna_rx *rx)
2565 {
2566         bfa_fsm_send_event(rx, RX_E_CLEANUP_DONE);
2567 }
2568 
2569 void
2570 bna_rx_vlan_strip_enable(struct bna_rx *rx)
2571 {
2572         struct bna_rxf *rxf = &rx->rxf;
2573 
2574         if (rxf->vlan_strip_status == BNA_STATUS_T_DISABLED) {
2575                 rxf->vlan_strip_status = BNA_STATUS_T_ENABLED;
2576                 rxf->vlan_strip_pending = true;
2577                 bfa_fsm_send_event(rxf, RXF_E_CONFIG);
2578         }
2579 }
2580 
2581 void
2582 bna_rx_vlan_strip_disable(struct bna_rx *rx)
2583 {
2584         struct bna_rxf *rxf = &rx->rxf;
2585 
2586         if (rxf->vlan_strip_status != BNA_STATUS_T_DISABLED) {
2587                 rxf->vlan_strip_status = BNA_STATUS_T_DISABLED;
2588                 rxf->vlan_strip_pending = true;
2589                 bfa_fsm_send_event(rxf, RXF_E_CONFIG);
2590         }
2591 }
2592 
2593 enum bna_cb_status
2594 bna_rx_mode_set(struct bna_rx *rx, enum bna_rxmode new_mode,
2595                 enum bna_rxmode bitmask)
2596 {
2597         struct bna_rxf *rxf = &rx->rxf;
2598         int need_hw_config = 0;
2599 
2600         /* Error checks */
2601 
2602         if (is_promisc_enable(new_mode, bitmask)) {
2603                 /* If promisc mode is already enabled elsewhere in the system */
2604                 if ((rx->bna->promisc_rid != BFI_INVALID_RID) &&
2605                         (rx->bna->promisc_rid != rxf->rx->rid))
2606                         goto err_return;
2607 
2608                 /* If default mode is already enabled in the system */
2609                 if (rx->bna->default_mode_rid != BFI_INVALID_RID)
2610                         goto err_return;
2611 
2612                 /* Trying to enable promiscuous and default mode together */
2613                 if (is_default_enable(new_mode, bitmask))
2614                         goto err_return;
2615         }
2616 
2617         if (is_default_enable(new_mode, bitmask)) {
2618                 /* If default mode is already enabled elsewhere in the system */
2619                 if ((rx->bna->default_mode_rid != BFI_INVALID_RID) &&
2620                         (rx->bna->default_mode_rid != rxf->rx->rid)) {
2621                                 goto err_return;
2622                 }
2623 
2624                 /* If promiscuous mode is already enabled in the system */
2625                 if (rx->bna->promisc_rid != BFI_INVALID_RID)
2626                         goto err_return;
2627         }
2628 
2629         /* Process the commands */
2630 
2631         if (is_promisc_enable(new_mode, bitmask)) {
2632                 if (bna_rxf_promisc_enable(rxf))
2633                         need_hw_config = 1;
2634         } else if (is_promisc_disable(new_mode, bitmask)) {
2635                 if (bna_rxf_promisc_disable(rxf))
2636                         need_hw_config = 1;
2637         }
2638 
2639         if (is_allmulti_enable(new_mode, bitmask)) {
2640                 if (bna_rxf_allmulti_enable(rxf))
2641                         need_hw_config = 1;
2642         } else if (is_allmulti_disable(new_mode, bitmask)) {
2643                 if (bna_rxf_allmulti_disable(rxf))
2644                         need_hw_config = 1;
2645         }
2646 
2647         /* Trigger h/w if needed */
2648 
2649         if (need_hw_config) {
2650                 rxf->cam_fltr_cbfn = NULL;
2651                 rxf->cam_fltr_cbarg = rx->bna->bnad;
2652                 bfa_fsm_send_event(rxf, RXF_E_CONFIG);
2653         }
2654 
2655         return BNA_CB_SUCCESS;
2656 
2657 err_return:
2658         return BNA_CB_FAIL;
2659 }
2660 
2661 void
2662 bna_rx_vlanfilter_enable(struct bna_rx *rx)
2663 {
2664         struct bna_rxf *rxf = &rx->rxf;
2665 
2666         if (rxf->vlan_filter_status == BNA_STATUS_T_DISABLED) {
2667                 rxf->vlan_filter_status = BNA_STATUS_T_ENABLED;
2668                 rxf->vlan_pending_bitmask = (u8)BFI_VLAN_BMASK_ALL;
2669                 bfa_fsm_send_event(rxf, RXF_E_CONFIG);
2670         }
2671 }
2672 
2673 void
2674 bna_rx_coalescing_timeo_set(struct bna_rx *rx, int coalescing_timeo)
2675 {
2676         struct bna_rxp *rxp;
2677 
2678         list_for_each_entry(rxp, &rx->rxp_q, qe) {
2679                 rxp->cq.ccb->rx_coalescing_timeo = coalescing_timeo;
2680                 bna_ib_coalescing_timeo_set(&rxp->cq.ib, coalescing_timeo);
2681         }
2682 }
2683 
2684 void
2685 bna_rx_dim_reconfig(struct bna *bna, const u32 vector[][BNA_BIAS_T_MAX])
2686 {
2687         int i, j;
2688 
2689         for (i = 0; i < BNA_LOAD_T_MAX; i++)
2690                 for (j = 0; j < BNA_BIAS_T_MAX; j++)
2691                         bna->rx_mod.dim_vector[i][j] = vector[i][j];
2692 }
2693 
2694 void
2695 bna_rx_dim_update(struct bna_ccb *ccb)
2696 {
2697         struct bna *bna = ccb->cq->rx->bna;
2698         u32 load, bias;
2699         u32 pkt_rt, small_rt, large_rt;
2700         u8 coalescing_timeo;
2701 
2702         if ((ccb->pkt_rate.small_pkt_cnt == 0) &&
2703                 (ccb->pkt_rate.large_pkt_cnt == 0))
2704                 return;
2705 
2706         /* Arrive at preconfigured coalescing timeo value based on pkt rate */
2707 
2708         small_rt = ccb->pkt_rate.small_pkt_cnt;
2709         large_rt = ccb->pkt_rate.large_pkt_cnt;
2710 
2711         pkt_rt = small_rt + large_rt;
2712 
2713         if (pkt_rt < BNA_PKT_RATE_10K)
2714                 load = BNA_LOAD_T_LOW_4;
2715         else if (pkt_rt < BNA_PKT_RATE_20K)
2716                 load = BNA_LOAD_T_LOW_3;
2717         else if (pkt_rt < BNA_PKT_RATE_30K)
2718                 load = BNA_LOAD_T_LOW_2;
2719         else if (pkt_rt < BNA_PKT_RATE_40K)
2720                 load = BNA_LOAD_T_LOW_1;
2721         else if (pkt_rt < BNA_PKT_RATE_50K)
2722                 load = BNA_LOAD_T_HIGH_1;
2723         else if (pkt_rt < BNA_PKT_RATE_60K)
2724                 load = BNA_LOAD_T_HIGH_2;
2725         else if (pkt_rt < BNA_PKT_RATE_80K)
2726                 load = BNA_LOAD_T_HIGH_3;
2727         else
2728                 load = BNA_LOAD_T_HIGH_4;
2729 
2730         if (small_rt > (large_rt << 1))
2731                 bias = 0;
2732         else
2733                 bias = 1;
2734 
2735         ccb->pkt_rate.small_pkt_cnt = 0;
2736         ccb->pkt_rate.large_pkt_cnt = 0;
2737 
2738         coalescing_timeo = bna->rx_mod.dim_vector[load][bias];
2739         ccb->rx_coalescing_timeo = coalescing_timeo;
2740 
2741         /* Set it to IB */
2742         bna_ib_coalescing_timeo_set(&ccb->cq->ib, coalescing_timeo);
2743 }
2744 
2745 const u32 bna_napi_dim_vector[BNA_LOAD_T_MAX][BNA_BIAS_T_MAX] = {
2746         {12, 12},
2747         {6, 10},
2748         {5, 10},
2749         {4, 8},
2750         {3, 6},
2751         {3, 6},
2752         {2, 4},
2753         {1, 2},
2754 };
2755 
2756 /* TX */
2757 
2758 #define call_tx_stop_cbfn(tx)                                           \
2759 do {                                                                    \
2760         if ((tx)->stop_cbfn) {                                          \
2761                 void (*cbfn)(void *, struct bna_tx *);          \
2762                 void *cbarg;                                            \
2763                 cbfn = (tx)->stop_cbfn;                                 \
2764                 cbarg = (tx)->stop_cbarg;                               \
2765                 (tx)->stop_cbfn = NULL;                                 \
2766                 (tx)->stop_cbarg = NULL;                                \
2767                 cbfn(cbarg, (tx));                                      \
2768         }                                                               \
2769 } while (0)
2770 
2771 static void bna_tx_mod_cb_tx_stopped(void *tx_mod, struct bna_tx *tx);
2772 static void bna_bfi_tx_enet_start(struct bna_tx *tx);
2773 static void bna_tx_enet_stop(struct bna_tx *tx);
2774 
2775 enum bna_tx_event {
2776         TX_E_START                      = 1,
2777         TX_E_STOP                       = 2,
2778         TX_E_FAIL                       = 3,
2779         TX_E_STARTED                    = 4,
2780         TX_E_STOPPED                    = 5,
2781         TX_E_CLEANUP_DONE               = 7,
2782         TX_E_BW_UPDATE                  = 8,
2783 };
2784 
2785 bfa_fsm_state_decl(bna_tx, stopped, struct bna_tx, enum bna_tx_event);
2786 bfa_fsm_state_decl(bna_tx, start_wait, struct bna_tx, enum bna_tx_event);
2787 bfa_fsm_state_decl(bna_tx, started, struct bna_tx, enum bna_tx_event);
2788 bfa_fsm_state_decl(bna_tx, stop_wait, struct bna_tx, enum bna_tx_event);
2789 bfa_fsm_state_decl(bna_tx, cleanup_wait, struct bna_tx,
2790                         enum bna_tx_event);
2791 bfa_fsm_state_decl(bna_tx, prio_stop_wait, struct bna_tx,
2792                         enum bna_tx_event);
2793 bfa_fsm_state_decl(bna_tx, prio_cleanup_wait, struct bna_tx,
2794                         enum bna_tx_event);
2795 bfa_fsm_state_decl(bna_tx, failed, struct bna_tx, enum bna_tx_event);
2796 bfa_fsm_state_decl(bna_tx, quiesce_wait, struct bna_tx,
2797                         enum bna_tx_event);
2798 
2799 static void
2800 bna_tx_sm_stopped_entry(struct bna_tx *tx)
2801 {
2802         call_tx_stop_cbfn(tx);
2803 }
2804 
2805 static void
2806 bna_tx_sm_stopped(struct bna_tx *tx, enum bna_tx_event event)
2807 {
2808         switch (event) {
2809         case TX_E_START:
2810                 bfa_fsm_set_state(tx, bna_tx_sm_start_wait);
2811                 break;
2812 
2813         case TX_E_STOP:
2814                 call_tx_stop_cbfn(tx);
2815                 break;
2816 
2817         case TX_E_FAIL:
2818                 /* No-op */
2819                 break;
2820 
2821         case TX_E_BW_UPDATE:
2822                 /* No-op */
2823                 break;
2824 
2825         default:
2826                 bfa_sm_fault(event);
2827         }
2828 }
2829 
2830 static void
2831 bna_tx_sm_start_wait_entry(struct bna_tx *tx)
2832 {
2833         bna_bfi_tx_enet_start(tx);
2834 }
2835 
2836 static void
2837 bna_tx_sm_start_wait(struct bna_tx *tx, enum bna_tx_event event)
2838 {
2839         switch (event) {
2840         case TX_E_STOP:
2841                 tx->flags &= ~BNA_TX_F_BW_UPDATED;
2842                 bfa_fsm_set_state(tx, bna_tx_sm_stop_wait);
2843                 break;
2844 
2845         case TX_E_FAIL:
2846                 tx->flags &= ~BNA_TX_F_BW_UPDATED;
2847                 bfa_fsm_set_state(tx, bna_tx_sm_stopped);
2848                 break;
2849 
2850         case TX_E_STARTED:
2851                 if (tx->flags & BNA_TX_F_BW_UPDATED) {
2852                         tx->flags &= ~BNA_TX_F_BW_UPDATED;
2853                         bfa_fsm_set_state(tx, bna_tx_sm_prio_stop_wait);
2854                 } else
2855                         bfa_fsm_set_state(tx, bna_tx_sm_started);
2856                 break;
2857 
2858         case TX_E_BW_UPDATE:
2859                 tx->flags |= BNA_TX_F_BW_UPDATED;
2860                 break;
2861 
2862         default:
2863                 bfa_sm_fault(event);
2864         }
2865 }
2866 
2867 static void
2868 bna_tx_sm_started_entry(struct bna_tx *tx)
2869 {
2870         struct bna_txq *txq;
2871         int is_regular = (tx->type == BNA_TX_T_REGULAR);
2872 
2873         list_for_each_entry(txq, &tx->txq_q, qe) {
2874                 txq->tcb->priority = txq->priority;
2875                 /* Start IB */
2876                 bna_ib_start(tx->bna, &txq->ib, is_regular);
2877         }
2878         tx->tx_resume_cbfn(tx->bna->bnad, tx);
2879 }
2880 
2881 static void
2882 bna_tx_sm_started(struct bna_tx *tx, enum bna_tx_event event)
2883 {
2884         switch (event) {
2885         case TX_E_STOP:
2886                 bfa_fsm_set_state(tx, bna_tx_sm_stop_wait);
2887                 tx->tx_stall_cbfn(tx->bna->bnad, tx);
2888                 bna_tx_enet_stop(tx);
2889                 break;
2890 
2891         case TX_E_FAIL:
2892                 bfa_fsm_set_state(tx, bna_tx_sm_failed);
2893                 tx->tx_stall_cbfn(tx->bna->bnad, tx);
2894                 tx->tx_cleanup_cbfn(tx->bna->bnad, tx);
2895                 break;
2896 
2897         case TX_E_BW_UPDATE:
2898                 bfa_fsm_set_state(tx, bna_tx_sm_prio_stop_wait);
2899                 break;
2900 
2901         default:
2902                 bfa_sm_fault(event);
2903         }
2904 }
2905 
2906 static void
2907 bna_tx_sm_stop_wait_entry(struct bna_tx *tx)
2908 {
2909 }
2910 
2911 static void
2912 bna_tx_sm_stop_wait(struct bna_tx *tx, enum bna_tx_event event)
2913 {
2914         switch (event) {
2915         case TX_E_FAIL:
2916         case TX_E_STOPPED:
2917                 bfa_fsm_set_state(tx, bna_tx_sm_cleanup_wait);
2918                 tx->tx_cleanup_cbfn(tx->bna->bnad, tx);
2919                 break;
2920 
2921         case TX_E_STARTED:
2922                 /**
2923                  * We are here due to start_wait -> stop_wait transition on
2924                  * TX_E_STOP event
2925                  */
2926                 bna_tx_enet_stop(tx);
2927                 break;
2928 
2929         case TX_E_BW_UPDATE:
2930                 /* No-op */
2931                 break;
2932 
2933         default:
2934                 bfa_sm_fault(event);
2935         }
2936 }
2937 
2938 static void
2939 bna_tx_sm_cleanup_wait_entry(struct bna_tx *tx)
2940 {
2941 }
2942 
2943 static void
2944 bna_tx_sm_cleanup_wait(struct bna_tx *tx, enum bna_tx_event event)
2945 {
2946         switch (event) {
2947         case TX_E_FAIL:
2948         case TX_E_BW_UPDATE:
2949                 /* No-op */
2950                 break;
2951 
2952         case TX_E_CLEANUP_DONE:
2953                 bfa_fsm_set_state(tx, bna_tx_sm_stopped);
2954                 break;
2955 
2956         default:
2957                 bfa_sm_fault(event);
2958         }
2959 }
2960 
2961 static void
2962 bna_tx_sm_prio_stop_wait_entry(struct bna_tx *tx)
2963 {
2964         tx->tx_stall_cbfn(tx->bna->bnad, tx);
2965         bna_tx_enet_stop(tx);
2966 }
2967 
2968 static void
2969 bna_tx_sm_prio_stop_wait(struct bna_tx *tx, enum bna_tx_event event)
2970 {
2971         switch (event) {
2972         case TX_E_STOP:
2973                 bfa_fsm_set_state(tx, bna_tx_sm_stop_wait);
2974                 break;
2975 
2976         case TX_E_FAIL:
2977                 bfa_fsm_set_state(tx, bna_tx_sm_failed);
2978                 tx->tx_cleanup_cbfn(tx->bna->bnad, tx);
2979                 break;
2980 
2981         case TX_E_STOPPED:
2982                 bfa_fsm_set_state(tx, bna_tx_sm_prio_cleanup_wait);
2983                 break;
2984 
2985         case TX_E_BW_UPDATE:
2986                 /* No-op */
2987                 break;
2988 
2989         default:
2990                 bfa_sm_fault(event);
2991         }
2992 }
2993 
2994 static void
2995 bna_tx_sm_prio_cleanup_wait_entry(struct bna_tx *tx)
2996 {
2997         tx->tx_cleanup_cbfn(tx->bna->bnad, tx);
2998 }
2999 
3000 static void
3001 bna_tx_sm_prio_cleanup_wait(struct bna_tx *tx, enum bna_tx_event event)
3002 {
3003         switch (event) {
3004         case TX_E_STOP:
3005                 bfa_fsm_set_state(tx, bna_tx_sm_cleanup_wait);
3006                 break;
3007 
3008         case TX_E_FAIL:
3009                 bfa_fsm_set_state(tx, bna_tx_sm_failed);
3010                 break;
3011 
3012         case TX_E_BW_UPDATE:
3013                 /* No-op */
3014                 break;
3015 
3016         case TX_E_CLEANUP_DONE:
3017                 bfa_fsm_set_state(tx, bna_tx_sm_start_wait);
3018                 break;
3019 
3020         default:
3021                 bfa_sm_fault(event);
3022         }
3023 }
3024 
3025 static void
3026 bna_tx_sm_failed_entry(struct bna_tx *tx)
3027 {
3028 }
3029 
3030 static void
3031 bna_tx_sm_failed(struct bna_tx *tx, enum bna_tx_event event)
3032 {
3033         switch (event) {
3034         case TX_E_START:
3035                 bfa_fsm_set_state(tx, bna_tx_sm_quiesce_wait);
3036                 break;
3037 
3038         case TX_E_STOP:
3039                 bfa_fsm_set_state(tx, bna_tx_sm_cleanup_wait);
3040                 break;
3041 
3042         case TX_E_FAIL:
3043                 /* No-op */
3044                 break;
3045 
3046         case TX_E_CLEANUP_DONE:
3047                 bfa_fsm_set_state(tx, bna_tx_sm_stopped);
3048                 break;
3049 
3050         default:
3051                 bfa_sm_fault(event);
3052         }
3053 }
3054 
3055 static void
3056 bna_tx_sm_quiesce_wait_entry(struct bna_tx *tx)
3057 {
3058 }
3059 
3060 static void
3061 bna_tx_sm_quiesce_wait(struct bna_tx *tx, enum bna_tx_event event)
3062 {
3063         switch (event) {
3064         case TX_E_STOP:
3065                 bfa_fsm_set_state(tx, bna_tx_sm_cleanup_wait);
3066                 break;
3067 
3068         case TX_E_FAIL:
3069                 bfa_fsm_set_state(tx, bna_tx_sm_failed);
3070                 break;
3071 
3072         case TX_E_CLEANUP_DONE:
3073                 bfa_fsm_set_state(tx, bna_tx_sm_start_wait);
3074                 break;
3075 
3076         case TX_E_BW_UPDATE:
3077                 /* No-op */
3078                 break;
3079 
3080         default:
3081                 bfa_sm_fault(event);
3082         }
3083 }
3084 
3085 static void
3086 bna_bfi_tx_enet_start(struct bna_tx *tx)
3087 {
3088         struct bfi_enet_tx_cfg_req *cfg_req = &tx->bfi_enet_cmd.cfg_req;
3089         struct bna_txq *txq = NULL;
3090         int i;
3091 
3092         bfi_msgq_mhdr_set(cfg_req->mh, BFI_MC_ENET,
3093                 BFI_ENET_H2I_TX_CFG_SET_REQ, 0, tx->rid);
3094         cfg_req->mh.num_entries = htons(
3095                 bfi_msgq_num_cmd_entries(sizeof(struct bfi_enet_tx_cfg_req)));
3096 
3097         cfg_req->num_queues = tx->num_txq;
3098         for (i = 0; i < tx->num_txq; i++) {
3099                 txq = txq ? list_next_entry(txq, qe)
3100                         : list_first_entry(&tx->txq_q, struct bna_txq, qe);
3101                 bfi_enet_datapath_q_init(&cfg_req->q_cfg[i].q.q, &txq->qpt);
3102                 cfg_req->q_cfg[i].q.priority = txq->priority;
3103 
3104                 cfg_req->q_cfg[i].ib.index_addr.a32.addr_lo =
3105                         txq->ib.ib_seg_host_addr.lsb;
3106                 cfg_req->q_cfg[i].ib.index_addr.a32.addr_hi =
3107                         txq->ib.ib_seg_host_addr.msb;
3108                 cfg_req->q_cfg[i].ib.intr.msix_index =
3109                         htons((u16)txq->ib.intr_vector);
3110         }
3111 
3112         cfg_req->ib_cfg.int_pkt_dma = BNA_STATUS_T_ENABLED;
3113         cfg_req->ib_cfg.int_enabled = BNA_STATUS_T_ENABLED;
3114         cfg_req->ib_cfg.int_pkt_enabled = BNA_STATUS_T_DISABLED;
3115         cfg_req->ib_cfg.continuous_coalescing = BNA_STATUS_T_ENABLED;
3116         cfg_req->ib_cfg.msix = (txq->ib.intr_type == BNA_INTR_T_MSIX)
3117                                 ? BNA_STATUS_T_ENABLED : BNA_STATUS_T_DISABLED;
3118         cfg_req->ib_cfg.coalescing_timeout =
3119                         htonl((u32)txq->ib.coalescing_timeo);
3120         cfg_req->ib_cfg.inter_pkt_timeout =
3121                         htonl((u32)txq->ib.interpkt_timeo);
3122         cfg_req->ib_cfg.inter_pkt_count = (u8)txq->ib.interpkt_count;
3123 
3124         cfg_req->tx_cfg.vlan_mode = BFI_ENET_TX_VLAN_WI;
3125         cfg_req->tx_cfg.vlan_id = htons((u16)tx->txf_vlan_id);
3126         cfg_req->tx_cfg.admit_tagged_frame = BNA_STATUS_T_ENABLED;
3127         cfg_req->tx_cfg.apply_vlan_filter = BNA_STATUS_T_DISABLED;
3128 
3129         bfa_msgq_cmd_set(&tx->msgq_cmd, NULL, NULL,
3130                 sizeof(struct bfi_enet_tx_cfg_req), &cfg_req->mh);
3131         bfa_msgq_cmd_post(&tx->bna->msgq, &tx->msgq_cmd);
3132 }
3133 
3134 static void
3135 bna_bfi_tx_enet_stop(struct bna_tx *tx)
3136 {
3137         struct bfi_enet_req *req = &tx->bfi_enet_cmd.req;
3138 
3139         bfi_msgq_mhdr_set(req->mh, BFI_MC_ENET,
3140                 BFI_ENET_H2I_TX_CFG_CLR_REQ, 0, tx->rid);
3141         req->mh.num_entries = htons(
3142                 bfi_msgq_num_cmd_entries(sizeof(struct bfi_enet_req)));
3143         bfa_msgq_cmd_set(&tx->msgq_cmd, NULL, NULL, sizeof(struct bfi_enet_req),
3144                 &req->mh);
3145         bfa_msgq_cmd_post(&tx->bna->msgq, &tx->msgq_cmd);
3146 }
3147 
3148 static void
3149 bna_tx_enet_stop(struct bna_tx *tx)
3150 {
3151         struct bna_txq *txq;
3152 
3153         /* Stop IB */
3154         list_for_each_entry(txq, &tx->txq_q, qe)
3155                 bna_ib_stop(tx->bna, &txq->ib);
3156 
3157         bna_bfi_tx_enet_stop(tx);
3158 }
3159 
3160 static void
3161 bna_txq_qpt_setup(struct bna_txq *txq, int page_count, int page_size,
3162                 struct bna_mem_descr *qpt_mem,
3163                 struct bna_mem_descr *swqpt_mem,
3164                 struct bna_mem_descr *page_mem)
3165 {
3166         u8 *kva;
3167         u64 dma;
3168         struct bna_dma_addr bna_dma;
3169         int i;
3170 
3171         txq->qpt.hw_qpt_ptr.lsb = qpt_mem->dma.lsb;
3172         txq->qpt.hw_qpt_ptr.msb = qpt_mem->dma.msb;
3173         txq->qpt.kv_qpt_ptr = qpt_mem->kva;
3174         txq->qpt.page_count = page_count;
3175         txq->qpt.page_size = page_size;
3176 
3177         txq->tcb->sw_qpt = (void **) swqpt_mem->kva;
3178         txq->tcb->sw_q = page_mem->kva;
3179 
3180         kva = page_mem->kva;
3181         BNA_GET_DMA_ADDR(&page_mem->dma, dma);
3182 
3183         for (i = 0; i < page_count; i++) {
3184                 txq->tcb->sw_qpt[i] = kva;
3185                 kva += PAGE_SIZE;
3186 
3187                 BNA_SET_DMA_ADDR(dma, &bna_dma);
3188                 ((struct bna_dma_addr *)txq->qpt.kv_qpt_ptr)[i].lsb =
3189                         bna_dma.lsb;
3190                 ((struct bna_dma_addr *)txq->qpt.kv_qpt_ptr)[i].msb =
3191                         bna_dma.msb;
3192                 dma += PAGE_SIZE;
3193         }
3194 }
3195 
3196 static struct bna_tx *
3197 bna_tx_get(struct bna_tx_mod *tx_mod, enum bna_tx_type type)
3198 {
3199         struct bna_tx *tx = NULL;
3200 
3201         if (list_empty(&tx_mod->tx_free_q))
3202                 return NULL;
3203         if (type == BNA_TX_T_REGULAR)
3204                 tx = list_first_entry(&tx_mod->tx_free_q, struct bna_tx, qe);
3205         else
3206                 tx = list_last_entry(&tx_mod->tx_free_q, struct bna_tx, qe);
3207         list_del(&tx->qe);
3208         tx->type = type;
3209 
3210         return tx;
3211 }
3212 
3213 static void
3214 bna_tx_free(struct bna_tx *tx)
3215 {
3216         struct bna_tx_mod *tx_mod = &tx->bna->tx_mod;
3217         struct bna_txq *txq;
3218         struct list_head *qe;
3219 
3220         while (!list_empty(&tx->txq_q)) {
3221                 txq = list_first_entry(&tx->txq_q, struct bna_txq, qe);
3222                 txq->tcb = NULL;
3223                 txq->tx = NULL;
3224                 list_move_tail(&txq->qe, &tx_mod->txq_free_q);
3225         }
3226 
3227         list_for_each(qe, &tx_mod->tx_active_q) {
3228                 if (qe == &tx->qe) {
3229                         list_del(&tx->qe);
3230                         break;
3231                 }
3232         }
3233 
3234         tx->bna = NULL;
3235         tx->priv = NULL;
3236 
3237         list_for_each_prev(qe, &tx_mod->tx_free_q)
3238                 if (((struct bna_tx *)qe)->rid < tx->rid)
3239                         break;
3240 
3241         list_add(&tx->qe, qe);
3242 }
3243 
3244 static void
3245 bna_tx_start(struct bna_tx *tx)
3246 {
3247         tx->flags |= BNA_TX_F_ENET_STARTED;
3248         if (tx->flags & BNA_TX_F_ENABLED)
3249                 bfa_fsm_send_event(tx, TX_E_START);
3250 }
3251 
3252 static void
3253 bna_tx_stop(struct bna_tx *tx)
3254 {
3255         tx->stop_cbfn = bna_tx_mod_cb_tx_stopped;
3256         tx->stop_cbarg = &tx->bna->tx_mod;
3257 
3258         tx->flags &= ~BNA_TX_F_ENET_STARTED;
3259         bfa_fsm_send_event(tx, TX_E_STOP);
3260 }
3261 
3262 static void
3263 bna_tx_fail(struct bna_tx *tx)
3264 {
3265         tx->flags &= ~BNA_TX_F_ENET_STARTED;
3266         bfa_fsm_send_event(tx, TX_E_FAIL);
3267 }
3268 
3269 void
3270 bna_bfi_tx_enet_start_rsp(struct bna_tx *tx, struct bfi_msgq_mhdr *msghdr)
3271 {
3272         struct bfi_enet_tx_cfg_rsp *cfg_rsp = &tx->bfi_enet_cmd.cfg_rsp;
3273         struct bna_txq *txq = NULL;
3274         int i;
3275 
3276         bfa_msgq_rsp_copy(&tx->bna->msgq, (u8 *)cfg_rsp,
3277                 sizeof(struct bfi_enet_tx_cfg_rsp));
3278 
3279         tx->hw_id = cfg_rsp->hw_id;
3280 
3281         for (i = 0, txq = list_first_entry(&tx->txq_q, struct bna_txq, qe);
3282              i < tx->num_txq; i++, txq = list_next_entry(txq, qe)) {
3283                 /* Setup doorbells */
3284                 txq->tcb->i_dbell->doorbell_addr =
3285                         tx->bna->pcidev.pci_bar_kva
3286                         + ntohl(cfg_rsp->q_handles[i].i_dbell);
3287                 txq->tcb->q_dbell =
3288                         tx->bna->pcidev.pci_bar_kva
3289                         + ntohl(cfg_rsp->q_handles[i].q_dbell);
3290                 txq->hw_id = cfg_rsp->q_handles[i].hw_qid;
3291 
3292                 /* Initialize producer/consumer indexes */
3293                 (*txq->tcb->hw_consumer_index) = 0;
3294                 txq->tcb->producer_index = txq->tcb->consumer_index = 0;
3295         }
3296 
3297         bfa_fsm_send_event(tx, TX_E_STARTED);
3298 }
3299 
3300 void
3301 bna_bfi_tx_enet_stop_rsp(struct bna_tx *tx, struct bfi_msgq_mhdr *msghdr)
3302 {
3303         bfa_fsm_send_event(tx, TX_E_STOPPED);
3304 }
3305 
3306 void
3307 bna_bfi_bw_update_aen(struct bna_tx_mod *tx_mod)
3308 {
3309         struct bna_tx *tx;
3310 
3311         list_for_each_entry(tx, &tx_mod->tx_active_q, qe)
3312                 bfa_fsm_send_event(tx, TX_E_BW_UPDATE);
3313 }
3314 
3315 void
3316 bna_tx_res_req(int num_txq, int txq_depth, struct bna_res_info *res_info)
3317 {
3318         u32 q_size;
3319         u32 page_count;
3320         struct bna_mem_info *mem_info;
3321 
3322         res_info[BNA_TX_RES_MEM_T_TCB].res_type = BNA_RES_T_MEM;
3323         mem_info = &res_info[BNA_TX_RES_MEM_T_TCB].res_u.mem_info;
3324         mem_info->mem_type = BNA_MEM_T_KVA;
3325         mem_info->len = sizeof(struct bna_tcb);
3326         mem_info->num = num_txq;
3327 
3328         q_size = txq_depth * BFI_TXQ_WI_SIZE;
3329         q_size = ALIGN(q_size, PAGE_SIZE);
3330         page_count = q_size >> PAGE_SHIFT;
3331 
3332         res_info[BNA_TX_RES_MEM_T_QPT].res_type = BNA_RES_T_MEM;
3333         mem_info = &res_info[BNA_TX_RES_MEM_T_QPT].res_u.mem_info;
3334         mem_info->mem_type = BNA_MEM_T_DMA;
3335         mem_info->len = page_count * sizeof(struct bna_dma_addr);
3336         mem_info->num = num_txq;
3337 
3338         res_info[BNA_TX_RES_MEM_T_SWQPT].res_type = BNA_RES_T_MEM;
3339         mem_info = &res_info[BNA_TX_RES_MEM_T_SWQPT].res_u.mem_info;
3340         mem_info->mem_type = BNA_MEM_T_KVA;
3341         mem_info->len = page_count * sizeof(void *);
3342         mem_info->num = num_txq;
3343 
3344         res_info[BNA_TX_RES_MEM_T_PAGE].res_type = BNA_RES_T_MEM;
3345         mem_info = &res_info[BNA_TX_RES_MEM_T_PAGE].res_u.mem_info;
3346         mem_info->mem_type = BNA_MEM_T_DMA;
3347         mem_info->len = PAGE_SIZE * page_count;
3348         mem_info->num = num_txq;
3349 
3350         res_info[BNA_TX_RES_MEM_T_IBIDX].res_type = BNA_RES_T_MEM;
3351         mem_info = &res_info[BNA_TX_RES_MEM_T_IBIDX].res_u.mem_info;
3352         mem_info->mem_type = BNA_MEM_T_DMA;
3353         mem_info->len = BFI_IBIDX_SIZE;
3354         mem_info->num = num_txq;
3355 
3356         res_info[BNA_TX_RES_INTR_T_TXCMPL].res_type = BNA_RES_T_INTR;
3357         res_info[BNA_TX_RES_INTR_T_TXCMPL].res_u.intr_info.intr_type =
3358                         BNA_INTR_T_MSIX;
3359         res_info[BNA_TX_RES_INTR_T_TXCMPL].res_u.intr_info.num = num_txq;
3360 }
3361 
3362 struct bna_tx *
3363 bna_tx_create(struct bna *bna, struct bnad *bnad,
3364                 struct bna_tx_config *tx_cfg,
3365                 const struct bna_tx_event_cbfn *tx_cbfn,
3366                 struct bna_res_info *res_info, void *priv)
3367 {
3368         struct bna_intr_info *intr_info;
3369         struct bna_tx_mod *tx_mod = &bna->tx_mod;
3370         struct bna_tx *tx;
3371         struct bna_txq *txq;
3372         int page_count;
3373         int i;
3374 
3375         intr_info = &res_info[BNA_TX_RES_INTR_T_TXCMPL].res_u.intr_info;
3376         page_count = (res_info[BNA_TX_RES_MEM_T_PAGE].res_u.mem_info.len) /
3377                                         PAGE_SIZE;
3378 
3379         /**
3380          * Get resources
3381          */
3382 
3383         if ((intr_info->num != 1) && (intr_info->num != tx_cfg->num_txq))
3384                 return NULL;
3385 
3386         /* Tx */
3387 
3388         tx = bna_tx_get(tx_mod, tx_cfg->tx_type);
3389         if (!tx)
3390                 return NULL;
3391         tx->bna = bna;
3392         tx->priv = priv;
3393 
3394         /* TxQs */
3395 
3396         INIT_LIST_HEAD(&tx->txq_q);
3397         for (i = 0; i < tx_cfg->num_txq; i++) {
3398                 if (list_empty(&tx_mod->txq_free_q))
3399                         goto err_return;
3400 
3401                 txq = list_first_entry(&tx_mod->txq_free_q, struct bna_txq, qe);
3402                 list_move_tail(&txq->qe, &tx->txq_q);
3403                 txq->tx = tx;
3404         }
3405 
3406         /*
3407          * Initialize
3408          */
3409 
3410         /* Tx */
3411 
3412         tx->tcb_setup_cbfn = tx_cbfn->tcb_setup_cbfn;
3413         tx->tcb_destroy_cbfn = tx_cbfn->tcb_destroy_cbfn;
3414         /* Following callbacks are mandatory */
3415         tx->tx_stall_cbfn = tx_cbfn->tx_stall_cbfn;
3416         tx->tx_resume_cbfn = tx_cbfn->tx_resume_cbfn;
3417         tx->tx_cleanup_cbfn = tx_cbfn->tx_cleanup_cbfn;
3418 
3419         list_add_tail(&tx->qe, &tx_mod->tx_active_q);
3420 
3421         tx->num_txq = tx_cfg->num_txq;
3422 
3423         tx->flags = 0;
3424         if (tx->bna->tx_mod.flags & BNA_TX_MOD_F_ENET_STARTED) {
3425                 switch (tx->type) {
3426                 case BNA_TX_T_REGULAR:
3427                         if (!(tx->bna->tx_mod.flags &
3428                                 BNA_TX_MOD_F_ENET_LOOPBACK))
3429                                 tx->flags |= BNA_TX_F_ENET_STARTED;
3430                         break;
3431                 case BNA_TX_T_LOOPBACK:
3432                         if (tx->bna->tx_mod.flags & BNA_TX_MOD_F_ENET_LOOPBACK)
3433                                 tx->flags |= BNA_TX_F_ENET_STARTED;
3434                         break;
3435                 }
3436         }
3437 
3438         /* TxQ */
3439 
3440         i = 0;
3441         list_for_each_entry(txq, &tx->txq_q, qe) {
3442                 txq->tcb = (struct bna_tcb *)
3443                 res_info[BNA_TX_RES_MEM_T_TCB].res_u.mem_info.mdl[i].kva;
3444                 txq->tx_packets = 0;
3445                 txq->tx_bytes = 0;
3446 
3447                 /* IB */
3448                 txq->ib.ib_seg_host_addr.lsb =
3449                 res_info[BNA_TX_RES_MEM_T_IBIDX].res_u.mem_info.mdl[i].dma.lsb;
3450                 txq->ib.ib_seg_host_addr.msb =
3451                 res_info[BNA_TX_RES_MEM_T_IBIDX].res_u.mem_info.mdl[i].dma.msb;
3452                 txq->ib.ib_seg_host_addr_kva =
3453                 res_info[BNA_TX_RES_MEM_T_IBIDX].res_u.mem_info.mdl[i].kva;
3454                 txq->ib.intr_type = intr_info->intr_type;
3455                 txq->ib.intr_vector = (intr_info->num == 1) ?
3456                                         intr_info->idl[0].vector :
3457                                         intr_info->idl[i].vector;
3458                 if (intr_info->intr_type == BNA_INTR_T_INTX)
3459                         txq->ib.intr_vector = BIT(txq->ib.intr_vector);
3460                 txq->ib.coalescing_timeo = tx_cfg->coalescing_timeo;
3461                 txq->ib.interpkt_timeo = BFI_TX_INTERPKT_TIMEO;
3462                 txq->ib.interpkt_count = BFI_TX_INTERPKT_COUNT;
3463 
3464                 /* TCB */
3465 
3466                 txq->tcb->q_depth = tx_cfg->txq_depth;
3467                 txq->tcb->unmap_q = (void *)
3468                 res_info[BNA_TX_RES_MEM_T_UNMAPQ].res_u.mem_info.mdl[i].kva;
3469                 txq->tcb->hw_consumer_index =
3470                         (u32 *)txq->ib.ib_seg_host_addr_kva;
3471                 txq->tcb->i_dbell = &txq->ib.door_bell;
3472                 txq->tcb->intr_type = txq->ib.intr_type;
3473                 txq->tcb->intr_vector = txq->ib.intr_vector;
3474                 txq->tcb->txq = txq;
3475                 txq->tcb->bnad = bnad;
3476                 txq->tcb->id = i;
3477 
3478                 /* QPT, SWQPT, Pages */
3479                 bna_txq_qpt_setup(txq, page_count, PAGE_SIZE,
3480                         &res_info[BNA_TX_RES_MEM_T_QPT].res_u.mem_info.mdl[i],
3481                         &res_info[BNA_TX_RES_MEM_T_SWQPT].res_u.mem_info.mdl[i],
3482                         &res_info[BNA_TX_RES_MEM_T_PAGE].
3483                                   res_u.mem_info.mdl[i]);
3484 
3485                 /* Callback to bnad for setting up TCB */
3486                 if (tx->tcb_setup_cbfn)
3487                         (tx->tcb_setup_cbfn)(bna->bnad, txq->tcb);
3488 
3489                 if (tx_cfg->num_txq == BFI_TX_MAX_PRIO)
3490                         txq->priority = txq->tcb->id;
3491                 else
3492                         txq->priority = tx_mod->default_prio;
3493 
3494                 i++;
3495         }
3496 
3497         tx->txf_vlan_id = 0;
3498 
3499         bfa_fsm_set_state(tx, bna_tx_sm_stopped);
3500 
3501         tx_mod->rid_mask |= BIT(tx->rid);
3502 
3503         return tx;
3504 
3505 err_return:
3506         bna_tx_free(tx);
3507         return NULL;
3508 }
3509 
3510 void
3511 bna_tx_destroy(struct bna_tx *tx)
3512 {
3513         struct bna_txq *txq;
3514 
3515         list_for_each_entry(txq, &tx->txq_q, qe)
3516                 if (tx->tcb_destroy_cbfn)
3517                         (tx->tcb_destroy_cbfn)(tx->bna->bnad, txq->tcb);
3518 
3519         tx->bna->tx_mod.rid_mask &= ~BIT(tx->rid);
3520         bna_tx_free(tx);
3521 }
3522 
3523 void
3524 bna_tx_enable(struct bna_tx *tx)
3525 {
3526         if (tx->fsm != (bfa_sm_t)bna_tx_sm_stopped)
3527                 return;
3528 
3529         tx->flags |= BNA_TX_F_ENABLED;
3530 
3531         if (tx->flags & BNA_TX_F_ENET_STARTED)
3532                 bfa_fsm_send_event(tx, TX_E_START);
3533 }
3534 
3535 void
3536 bna_tx_disable(struct bna_tx *tx, enum bna_cleanup_type type,
3537                 void (*cbfn)(void *, struct bna_tx *))
3538 {
3539         if (type == BNA_SOFT_CLEANUP) {
3540                 (*cbfn)(tx->bna->bnad, tx);
3541                 return;
3542         }
3543 
3544         tx->stop_cbfn = cbfn;
3545         tx->stop_cbarg = tx->bna->bnad;
3546 
3547         tx->flags &= ~BNA_TX_F_ENABLED;
3548 
3549         bfa_fsm_send_event(tx, TX_E_STOP);
3550 }
3551 
3552 void
3553 bna_tx_cleanup_complete(struct bna_tx *tx)
3554 {
3555         bfa_fsm_send_event(tx, TX_E_CLEANUP_DONE);
3556 }
3557 
3558 static void
3559 bna_tx_mod_cb_tx_stopped(void *arg, struct bna_tx *tx)
3560 {
3561         struct bna_tx_mod *tx_mod = (struct bna_tx_mod *)arg;
3562 
3563         bfa_wc_down(&tx_mod->tx_stop_wc);
3564 }
3565 
3566 static void
3567 bna_tx_mod_cb_tx_stopped_all(void *arg)
3568 {
3569         struct bna_tx_mod *tx_mod = (struct bna_tx_mod *)arg;
3570 
3571         if (tx_mod->stop_cbfn)
3572                 tx_mod->stop_cbfn(&tx_mod->bna->enet);
3573         tx_mod->stop_cbfn = NULL;
3574 }
3575 
3576 void
3577 bna_tx_mod_init(struct bna_tx_mod *tx_mod, struct bna *bna,
3578                 struct bna_res_info *res_info)
3579 {
3580         int i;
3581 
3582         tx_mod->bna = bna;
3583         tx_mod->flags = 0;
3584 
3585         tx_mod->tx = (struct bna_tx *)
3586                 res_info[BNA_MOD_RES_MEM_T_TX_ARRAY].res_u.mem_info.mdl[0].kva;
3587         tx_mod->txq = (struct bna_txq *)
3588                 res_info[BNA_MOD_RES_MEM_T_TXQ_ARRAY].res_u.mem_info.mdl[0].kva;
3589 
3590         INIT_LIST_HEAD(&tx_mod->tx_free_q);
3591         INIT_LIST_HEAD(&tx_mod->tx_active_q);
3592 
3593         INIT_LIST_HEAD(&tx_mod->txq_free_q);
3594 
3595         for (i = 0; i < bna->ioceth.attr.num_txq; i++) {
3596                 tx_mod->tx[i].rid = i;
3597                 list_add_tail(&tx_mod->tx[i].qe, &tx_mod->tx_free_q);
3598                 list_add_tail(&tx_mod->txq[i].qe, &tx_mod->txq_free_q);
3599         }
3600 
3601         tx_mod->prio_map = BFI_TX_PRIO_MAP_ALL;
3602         tx_mod->default_prio = 0;
3603         tx_mod->iscsi_over_cee = BNA_STATUS_T_DISABLED;
3604         tx_mod->iscsi_prio = -1;
3605 }
3606 
3607 void
3608 bna_tx_mod_uninit(struct bna_tx_mod *tx_mod)
3609 {
3610         tx_mod->bna = NULL;
3611 }
3612 
3613 void
3614 bna_tx_mod_start(struct bna_tx_mod *tx_mod, enum bna_tx_type type)
3615 {
3616         struct bna_tx *tx;
3617 
3618         tx_mod->flags |= BNA_TX_MOD_F_ENET_STARTED;
3619         if (type == BNA_TX_T_LOOPBACK)
3620                 tx_mod->flags |= BNA_TX_MOD_F_ENET_LOOPBACK;
3621 
3622         list_for_each_entry(tx, &tx_mod->tx_active_q, qe)
3623                 if (tx->type == type)
3624                         bna_tx_start(tx);
3625 }
3626 
3627 void
3628 bna_tx_mod_stop(struct bna_tx_mod *tx_mod, enum bna_tx_type type)
3629 {
3630         struct bna_tx *tx;
3631 
3632         tx_mod->flags &= ~BNA_TX_MOD_F_ENET_STARTED;
3633         tx_mod->flags &= ~BNA_TX_MOD_F_ENET_LOOPBACK;
3634 
3635         tx_mod->stop_cbfn = bna_enet_cb_tx_stopped;
3636 
3637         bfa_wc_init(&tx_mod->tx_stop_wc, bna_tx_mod_cb_tx_stopped_all, tx_mod);
3638 
3639         list_for_each_entry(tx, &tx_mod->tx_active_q, qe)
3640                 if (tx->type == type) {
3641                         bfa_wc_up(&tx_mod->tx_stop_wc);
3642                         bna_tx_stop(tx);
3643                 }
3644 
3645         bfa_wc_wait(&tx_mod->tx_stop_wc);
3646 }
3647 
3648 void
3649 bna_tx_mod_fail(struct bna_tx_mod *tx_mod)
3650 {
3651         struct bna_tx *tx;
3652 
3653         tx_mod->flags &= ~BNA_TX_MOD_F_ENET_STARTED;
3654         tx_mod->flags &= ~BNA_TX_MOD_F_ENET_LOOPBACK;
3655 
3656         list_for_each_entry(tx, &tx_mod->tx_active_q, qe)
3657                 bna_tx_fail(tx);
3658 }
3659 
3660 void
3661 bna_tx_coalescing_timeo_set(struct bna_tx *tx, int coalescing_timeo)
3662 {
3663         struct bna_txq *txq;
3664 
3665         list_for_each_entry(txq, &tx->txq_q, qe)
3666                 bna_ib_coalescing_timeo_set(&txq->ib, coalescing_timeo);
3667 }

/* [<][>][^][v][top][bottom][index][help] */