sq_wr_avail 1330 drivers/infiniband/ulp/srpt/ib_srpt.c atomic_add(ioctx->n_rdma, &ch->sq_wr_avail); sq_wr_avail 1757 drivers/infiniband/ulp/srpt/ib_srpt.c atomic_add(1 + ioctx->n_rdma, &ch->sq_wr_avail); sq_wr_avail 1861 drivers/infiniband/ulp/srpt/ib_srpt.c atomic_set(&ch->sq_wr_avail, qp_init->cap.max_send_wr); sq_wr_avail 2743 drivers/infiniband/ulp/srpt/ib_srpt.c if (atomic_sub_return(ioctx->n_rdma, &ch->sq_wr_avail) < 0) { sq_wr_avail 2763 drivers/infiniband/ulp/srpt/ib_srpt.c atomic_read(&ch->sq_wr_avail)); sq_wr_avail 2769 drivers/infiniband/ulp/srpt/ib_srpt.c atomic_add(ioctx->n_rdma, &ch->sq_wr_avail); sq_wr_avail 2848 drivers/infiniband/ulp/srpt/ib_srpt.c &ch->sq_wr_avail) < 0)) { sq_wr_avail 2880 drivers/infiniband/ulp/srpt/ib_srpt.c atomic_add(1 + ioctx->n_rdma, &ch->sq_wr_avail); sq_wr_avail 311 drivers/infiniband/ulp/srpt/ib_srpt.h atomic_t sq_wr_avail; sq_wr_avail 80 drivers/nvme/target/rdma.c atomic_t sq_wr_avail; sq_wr_avail 503 drivers/nvme/target/rdma.c atomic_add(1 + rsp->n_rdma, &queue->sq_wr_avail); sq_wr_avail 589 drivers/nvme/target/rdma.c atomic_add(rsp->n_rdma, &queue->sq_wr_avail); sq_wr_avail 736 drivers/nvme/target/rdma.c &queue->sq_wr_avail) < 0)) { sq_wr_avail 740 drivers/nvme/target/rdma.c atomic_add(1 + rsp->n_rdma, &queue->sq_wr_avail); sq_wr_avail 1029 drivers/nvme/target/rdma.c atomic_set(&queue->sq_wr_avail, qp_attr.cap.max_send_wr);