sq_wr_avail      1330 drivers/infiniband/ulp/srpt/ib_srpt.c 	atomic_add(ioctx->n_rdma, &ch->sq_wr_avail);
sq_wr_avail      1757 drivers/infiniband/ulp/srpt/ib_srpt.c 	atomic_add(1 + ioctx->n_rdma, &ch->sq_wr_avail);
sq_wr_avail      1861 drivers/infiniband/ulp/srpt/ib_srpt.c 	atomic_set(&ch->sq_wr_avail, qp_init->cap.max_send_wr);
sq_wr_avail      2743 drivers/infiniband/ulp/srpt/ib_srpt.c 	if (atomic_sub_return(ioctx->n_rdma, &ch->sq_wr_avail) < 0) {
sq_wr_avail      2763 drivers/infiniband/ulp/srpt/ib_srpt.c 			 atomic_read(&ch->sq_wr_avail));
sq_wr_avail      2769 drivers/infiniband/ulp/srpt/ib_srpt.c 	atomic_add(ioctx->n_rdma, &ch->sq_wr_avail);
sq_wr_avail      2848 drivers/infiniband/ulp/srpt/ib_srpt.c 			&ch->sq_wr_avail) < 0)) {
sq_wr_avail      2880 drivers/infiniband/ulp/srpt/ib_srpt.c 	atomic_add(1 + ioctx->n_rdma, &ch->sq_wr_avail);
sq_wr_avail       311 drivers/infiniband/ulp/srpt/ib_srpt.h 	atomic_t		sq_wr_avail;
sq_wr_avail        80 drivers/nvme/target/rdma.c 	atomic_t		sq_wr_avail;
sq_wr_avail       503 drivers/nvme/target/rdma.c 	atomic_add(1 + rsp->n_rdma, &queue->sq_wr_avail);
sq_wr_avail       589 drivers/nvme/target/rdma.c 	atomic_add(rsp->n_rdma, &queue->sq_wr_avail);
sq_wr_avail       736 drivers/nvme/target/rdma.c 			&queue->sq_wr_avail) < 0)) {
sq_wr_avail       740 drivers/nvme/target/rdma.c 		atomic_add(1 + rsp->n_rdma, &queue->sq_wr_avail);
sq_wr_avail      1029 drivers/nvme/target/rdma.c 	atomic_set(&queue->sq_wr_avail, qp_attr.cap.max_send_wr);