root/drivers/infiniband/hw/hfi1/opfn.c

/* [<][>][^][v][top][bottom][index][help] */

DEFINITIONS

This source file includes following definitions.
  1. hfi1_opfn_extended
  2. opfn_conn_request
  3. opfn_send_conn_request
  4. opfn_schedule_conn_request
  5. opfn_conn_response
  6. opfn_conn_reply
  7. opfn_conn_error
  8. opfn_qp_init
  9. opfn_trigger_conn_request
  10. opfn_init
  11. opfn_exit

   1 // SPDX-License-Identifier: (GPL-2.0 OR BSD-3-Clause)
   2 /*
   3  * Copyright(c) 2018 Intel Corporation.
   4  *
   5  */
   6 #include "hfi.h"
   7 #include "trace.h"
   8 #include "qp.h"
   9 #include "opfn.h"
  10 
  11 #define IB_BTHE_E                 BIT(IB_BTHE_E_SHIFT)
  12 
  13 #define OPFN_CODE(code) BIT((code) - 1)
  14 #define OPFN_MASK(code) OPFN_CODE(STL_VERBS_EXTD_##code)
  15 
  16 struct hfi1_opfn_type {
  17         bool (*request)(struct rvt_qp *qp, u64 *data);
  18         bool (*response)(struct rvt_qp *qp, u64 *data);
  19         bool (*reply)(struct rvt_qp *qp, u64 data);
  20         void (*error)(struct rvt_qp *qp);
  21 };
  22 
  23 static struct hfi1_opfn_type hfi1_opfn_handlers[STL_VERBS_EXTD_MAX] = {
  24         [STL_VERBS_EXTD_TID_RDMA] = {
  25                 .request = tid_rdma_conn_req,
  26                 .response = tid_rdma_conn_resp,
  27                 .reply = tid_rdma_conn_reply,
  28                 .error = tid_rdma_conn_error,
  29         },
  30 };
  31 
  32 static struct workqueue_struct *opfn_wq;
  33 
  34 static void opfn_schedule_conn_request(struct rvt_qp *qp);
  35 
  36 static bool hfi1_opfn_extended(u32 bth1)
  37 {
  38         return !!(bth1 & IB_BTHE_E);
  39 }
  40 
  41 static void opfn_conn_request(struct rvt_qp *qp)
  42 {
  43         struct hfi1_qp_priv *priv = qp->priv;
  44         struct ib_atomic_wr wr;
  45         u16 mask, capcode;
  46         struct hfi1_opfn_type *extd;
  47         u64 data;
  48         unsigned long flags;
  49         int ret = 0;
  50 
  51         trace_hfi1_opfn_state_conn_request(qp);
  52         spin_lock_irqsave(&priv->opfn.lock, flags);
  53         /*
  54          * Exit if the extended bit is not set, or if nothing is requested, or
  55          * if we have completed all requests, or if a previous request is in
  56          * progress
  57          */
  58         if (!priv->opfn.extended || !priv->opfn.requested ||
  59             priv->opfn.requested == priv->opfn.completed || priv->opfn.curr)
  60                 goto done;
  61 
  62         mask = priv->opfn.requested & ~priv->opfn.completed;
  63         capcode = ilog2(mask & ~(mask - 1)) + 1;
  64         if (capcode >= STL_VERBS_EXTD_MAX) {
  65                 priv->opfn.completed |= OPFN_CODE(capcode);
  66                 goto done;
  67         }
  68 
  69         extd = &hfi1_opfn_handlers[capcode];
  70         if (!extd || !extd->request || !extd->request(qp, &data)) {
  71                 /*
  72                  * Either there is no handler for this capability or the request
  73                  * packet could not be generated. Either way, mark it as done so
  74                  * we don't keep attempting to complete it.
  75                  */
  76                 priv->opfn.completed |= OPFN_CODE(capcode);
  77                 goto done;
  78         }
  79 
  80         trace_hfi1_opfn_data_conn_request(qp, capcode, data);
  81         data = (data & ~0xf) | capcode;
  82 
  83         memset(&wr, 0, sizeof(wr));
  84         wr.wr.opcode = IB_WR_OPFN;
  85         wr.remote_addr = HFI1_VERBS_E_ATOMIC_VADDR;
  86         wr.compare_add = data;
  87 
  88         priv->opfn.curr = capcode;      /* A new request is now in progress */
  89         /* Drop opfn.lock before calling ib_post_send() */
  90         spin_unlock_irqrestore(&priv->opfn.lock, flags);
  91 
  92         ret = ib_post_send(&qp->ibqp, &wr.wr, NULL);
  93         if (ret)
  94                 goto err;
  95         trace_hfi1_opfn_state_conn_request(qp);
  96         return;
  97 err:
  98         trace_hfi1_msg_opfn_conn_request(qp, "ib_ost_send failed: ret = ",
  99                                          (u64)ret);
 100         spin_lock_irqsave(&priv->opfn.lock, flags);
 101         /*
 102          * In case of an unexpected error return from ib_post_send
 103          * clear opfn.curr and reschedule to try again
 104          */
 105         priv->opfn.curr = STL_VERBS_EXTD_NONE;
 106         opfn_schedule_conn_request(qp);
 107 done:
 108         spin_unlock_irqrestore(&priv->opfn.lock, flags);
 109 }
 110 
 111 void opfn_send_conn_request(struct work_struct *work)
 112 {
 113         struct hfi1_opfn_data *od;
 114         struct hfi1_qp_priv *qpriv;
 115 
 116         od = container_of(work, struct hfi1_opfn_data, opfn_work);
 117         qpriv = container_of(od, struct hfi1_qp_priv, opfn);
 118 
 119         opfn_conn_request(qpriv->owner);
 120 }
 121 
 122 /*
 123  * When QP s_lock is held in the caller, the OPFN request must be scheduled
 124  * to a different workqueue to avoid double locking QP s_lock in call to
 125  * ib_post_send in opfn_conn_request
 126  */
 127 static void opfn_schedule_conn_request(struct rvt_qp *qp)
 128 {
 129         struct hfi1_qp_priv *priv = qp->priv;
 130 
 131         trace_hfi1_opfn_state_sched_conn_request(qp);
 132         queue_work(opfn_wq, &priv->opfn.opfn_work);
 133 }
 134 
 135 void opfn_conn_response(struct rvt_qp *qp, struct rvt_ack_entry *e,
 136                         struct ib_atomic_eth *ateth)
 137 {
 138         struct hfi1_qp_priv *priv = qp->priv;
 139         u64 data = be64_to_cpu(ateth->compare_data);
 140         struct hfi1_opfn_type *extd;
 141         u8 capcode;
 142         unsigned long flags;
 143 
 144         trace_hfi1_opfn_state_conn_response(qp);
 145         capcode = data & 0xf;
 146         trace_hfi1_opfn_data_conn_response(qp, capcode, data);
 147         if (!capcode || capcode >= STL_VERBS_EXTD_MAX)
 148                 return;
 149 
 150         extd = &hfi1_opfn_handlers[capcode];
 151 
 152         if (!extd || !extd->response) {
 153                 e->atomic_data = capcode;
 154                 return;
 155         }
 156 
 157         spin_lock_irqsave(&priv->opfn.lock, flags);
 158         if (priv->opfn.completed & OPFN_CODE(capcode)) {
 159                 /*
 160                  * We are receiving a request for a feature that has already
 161                  * been negotiated. This may mean that the other side has reset
 162                  */
 163                 priv->opfn.completed &= ~OPFN_CODE(capcode);
 164                 if (extd->error)
 165                         extd->error(qp);
 166         }
 167 
 168         if (extd->response(qp, &data))
 169                 priv->opfn.completed |= OPFN_CODE(capcode);
 170         e->atomic_data = (data & ~0xf) | capcode;
 171         trace_hfi1_opfn_state_conn_response(qp);
 172         spin_unlock_irqrestore(&priv->opfn.lock, flags);
 173 }
 174 
 175 void opfn_conn_reply(struct rvt_qp *qp, u64 data)
 176 {
 177         struct hfi1_qp_priv *priv = qp->priv;
 178         struct hfi1_opfn_type *extd;
 179         u8 capcode;
 180         unsigned long flags;
 181 
 182         trace_hfi1_opfn_state_conn_reply(qp);
 183         capcode = data & 0xf;
 184         trace_hfi1_opfn_data_conn_reply(qp, capcode, data);
 185         if (!capcode || capcode >= STL_VERBS_EXTD_MAX)
 186                 return;
 187 
 188         spin_lock_irqsave(&priv->opfn.lock, flags);
 189         /*
 190          * Either there is no previous request or the reply is not for the
 191          * current request
 192          */
 193         if (!priv->opfn.curr || capcode != priv->opfn.curr)
 194                 goto done;
 195 
 196         extd = &hfi1_opfn_handlers[capcode];
 197 
 198         if (!extd || !extd->reply)
 199                 goto clear;
 200 
 201         if (extd->reply(qp, data))
 202                 priv->opfn.completed |= OPFN_CODE(capcode);
 203 clear:
 204         /*
 205          * Clear opfn.curr to indicate that the previous request is no longer in
 206          * progress
 207          */
 208         priv->opfn.curr = STL_VERBS_EXTD_NONE;
 209         trace_hfi1_opfn_state_conn_reply(qp);
 210 done:
 211         spin_unlock_irqrestore(&priv->opfn.lock, flags);
 212 }
 213 
 214 void opfn_conn_error(struct rvt_qp *qp)
 215 {
 216         struct hfi1_qp_priv *priv = qp->priv;
 217         struct hfi1_opfn_type *extd = NULL;
 218         unsigned long flags;
 219         u16 capcode;
 220 
 221         trace_hfi1_opfn_state_conn_error(qp);
 222         trace_hfi1_msg_opfn_conn_error(qp, "error. qp state ", (u64)qp->state);
 223         /*
 224          * The QP has gone into the Error state. We have to invalidate all
 225          * negotiated feature, including the one in progress (if any). The RC
 226          * QP handling will clean the WQE for the connection request.
 227          */
 228         spin_lock_irqsave(&priv->opfn.lock, flags);
 229         while (priv->opfn.completed) {
 230                 capcode = priv->opfn.completed & ~(priv->opfn.completed - 1);
 231                 extd = &hfi1_opfn_handlers[ilog2(capcode) + 1];
 232                 if (extd->error)
 233                         extd->error(qp);
 234                 priv->opfn.completed &= ~OPFN_CODE(capcode);
 235         }
 236         priv->opfn.extended = 0;
 237         priv->opfn.requested = 0;
 238         priv->opfn.curr = STL_VERBS_EXTD_NONE;
 239         spin_unlock_irqrestore(&priv->opfn.lock, flags);
 240 }
 241 
 242 void opfn_qp_init(struct rvt_qp *qp, struct ib_qp_attr *attr, int attr_mask)
 243 {
 244         struct ib_qp *ibqp = &qp->ibqp;
 245         struct hfi1_qp_priv *priv = qp->priv;
 246         unsigned long flags;
 247 
 248         if (attr_mask & IB_QP_RETRY_CNT)
 249                 priv->s_retry = attr->retry_cnt;
 250 
 251         spin_lock_irqsave(&priv->opfn.lock, flags);
 252         if (ibqp->qp_type == IB_QPT_RC && HFI1_CAP_IS_KSET(TID_RDMA)) {
 253                 struct tid_rdma_params *local = &priv->tid_rdma.local;
 254 
 255                 if (attr_mask & IB_QP_TIMEOUT)
 256                         priv->tid_retry_timeout_jiffies = qp->timeout_jiffies;
 257                 if (qp->pmtu == enum_to_mtu(OPA_MTU_4096) ||
 258                     qp->pmtu == enum_to_mtu(OPA_MTU_8192)) {
 259                         tid_rdma_opfn_init(qp, local);
 260                         /*
 261                          * We only want to set the OPFN requested bit when the
 262                          * QP transitions to RTS.
 263                          */
 264                         if (attr_mask & IB_QP_STATE &&
 265                             attr->qp_state == IB_QPS_RTS) {
 266                                 priv->opfn.requested |= OPFN_MASK(TID_RDMA);
 267                                 /*
 268                                  * If the QP is transitioning to RTS and the
 269                                  * opfn.completed for TID RDMA has already been
 270                                  * set, the QP is being moved *back* into RTS.
 271                                  * We can now renegotiate the TID RDMA
 272                                  * parameters.
 273                                  */
 274                                 if (priv->opfn.completed &
 275                                     OPFN_MASK(TID_RDMA)) {
 276                                         priv->opfn.completed &=
 277                                                 ~OPFN_MASK(TID_RDMA);
 278                                         /*
 279                                          * Since the opfn.completed bit was
 280                                          * already set, it is safe to assume
 281                                          * that the opfn.extended is also set.
 282                                          */
 283                                         opfn_schedule_conn_request(qp);
 284                                 }
 285                         }
 286                 } else {
 287                         memset(local, 0, sizeof(*local));
 288                 }
 289         }
 290         spin_unlock_irqrestore(&priv->opfn.lock, flags);
 291 }
 292 
 293 void opfn_trigger_conn_request(struct rvt_qp *qp, u32 bth1)
 294 {
 295         struct hfi1_qp_priv *priv = qp->priv;
 296 
 297         if (!priv->opfn.extended && hfi1_opfn_extended(bth1) &&
 298             HFI1_CAP_IS_KSET(OPFN)) {
 299                 priv->opfn.extended = 1;
 300                 if (qp->state == IB_QPS_RTS)
 301                         opfn_conn_request(qp);
 302         }
 303 }
 304 
 305 int opfn_init(void)
 306 {
 307         opfn_wq = alloc_workqueue("hfi_opfn",
 308                                   WQ_SYSFS | WQ_HIGHPRI | WQ_CPU_INTENSIVE |
 309                                   WQ_MEM_RECLAIM,
 310                                   HFI1_MAX_ACTIVE_WORKQUEUE_ENTRIES);
 311         if (!opfn_wq)
 312                 return -ENOMEM;
 313 
 314         return 0;
 315 }
 316 
 317 void opfn_exit(void)
 318 {
 319         if (opfn_wq) {
 320                 destroy_workqueue(opfn_wq);
 321                 opfn_wq = NULL;
 322         }
 323 }

/* [<][>][^][v][top][bottom][index][help] */