wr 187 arch/mips/include/asm/pci/bridge.h u64 wr; /* write-only */ wr 39 arch/mips/include/uapi/asm/ucontext.h unsigned long long wr[32]; wr 189 arch/mips/kernel/signal.c err |= _save_msa_all_upper(&msa->wr); wr 199 arch/mips/kernel/signal.c err |= __put_user(val, &msa->wr[i]); wr 236 arch/mips/kernel/signal.c err |= _restore_msa_all_upper(&msa->wr); wr 244 arch/mips/kernel/signal.c err |= __get_user(val, &msa->wr[i]); wr 2050 arch/mips/mm/tlbex.c struct work_registers wr = build_get_work_registers(p); wr 2053 arch/mips/mm/tlbex.c build_get_pmde64(p, l, r, wr.r1, wr.r2); /* get pmd in ptr */ wr 2055 arch/mips/mm/tlbex.c build_get_pgde32(p, wr.r1, wr.r2); /* get pgd in ptr */ wr 2064 arch/mips/mm/tlbex.c build_is_huge_pte(p, r, wr.r1, wr.r2, label_tlb_huge_update); wr 2067 arch/mips/mm/tlbex.c UASM_i_MFC0(p, wr.r1, C0_BADVADDR); wr 2068 arch/mips/mm/tlbex.c UASM_i_LW(p, wr.r2, 0, wr.r2); wr 2069 arch/mips/mm/tlbex.c UASM_i_SRL(p, wr.r1, wr.r1, PAGE_SHIFT + PTE_ORDER - PTE_T_LOG2); wr 2070 arch/mips/mm/tlbex.c uasm_i_andi(p, wr.r1, wr.r1, (PTRS_PER_PTE - 1) << PTE_T_LOG2); wr 2071 arch/mips/mm/tlbex.c UASM_i_ADDU(p, wr.r2, wr.r2, wr.r1); wr 2076 arch/mips/mm/tlbex.c iPTE_LW(p, wr.r1, wr.r2); /* get even pte */ wr 2082 arch/mips/mm/tlbex.c uasm_i_mfc0(p, wr.r3, C0_INDEX); wr 2083 arch/mips/mm/tlbex.c uasm_il_bltz(p, r, wr.r3, label_leave); wr 2087 arch/mips/mm/tlbex.c return wr; wr 2113 arch/mips/mm/tlbex.c struct work_registers wr; wr 2133 arch/mips/mm/tlbex.c wr = build_r4000_tlbchange_handler_head(&p, &l, &r); wr 2134 arch/mips/mm/tlbex.c build_pte_present(&p, &r, wr.r1, wr.r2, wr.r3, label_nopage_tlbl); wr 2144 arch/mips/mm/tlbex.c uasm_il_bbit0(&p, &r, wr.r1, ilog2(_PAGE_VALID), wr 2147 arch/mips/mm/tlbex.c uasm_i_andi(&p, wr.r3, wr.r1, _PAGE_VALID); wr 2148 arch/mips/mm/tlbex.c uasm_il_beqz(&p, &r, wr.r3, label_tlbl_goaround1); wr 2176 arch/mips/mm/tlbex.c uasm_i_bbit0(&p, wr.r2, ilog2(sizeof(pte_t)), 8); wr 2178 arch/mips/mm/tlbex.c uasm_i_andi(&p, wr.r3, wr.r2, sizeof(pte_t)); wr 2179 arch/mips/mm/tlbex.c uasm_i_beqz(&p, wr.r3, 8); wr 2182 arch/mips/mm/tlbex.c UASM_i_MFC0(&p, wr.r3, C0_ENTRYLO0); wr 2184 arch/mips/mm/tlbex.c UASM_i_MFC0(&p, wr.r3, C0_ENTRYLO1); wr 2190 arch/mips/mm/tlbex.c uasm_il_bbit1(&p, &r, wr.r3, 1, label_nopage_tlbl); wr 2194 arch/mips/mm/tlbex.c uasm_i_andi(&p, wr.r3, wr.r3, 2); wr 2195 arch/mips/mm/tlbex.c uasm_il_bnez(&p, &r, wr.r3, label_nopage_tlbl); wr 2200 arch/mips/mm/tlbex.c build_make_valid(&p, &r, wr.r1, wr.r2, wr.r3); wr 2201 arch/mips/mm/tlbex.c build_r4000_tlbchange_handler_tail(&p, &l, &r, wr.r1, wr.r2); wr 2209 arch/mips/mm/tlbex.c iPTE_LW(&p, wr.r1, wr.r2); wr 2210 arch/mips/mm/tlbex.c build_pte_present(&p, &r, wr.r1, wr.r2, wr.r3, label_nopage_tlbl); wr 2219 arch/mips/mm/tlbex.c uasm_il_bbit0(&p, &r, wr.r1, ilog2(_PAGE_VALID), wr 2222 arch/mips/mm/tlbex.c uasm_i_andi(&p, wr.r3, wr.r1, _PAGE_VALID); wr 2223 arch/mips/mm/tlbex.c uasm_il_beqz(&p, &r, wr.r3, label_tlbl_goaround2); wr 2251 arch/mips/mm/tlbex.c uasm_i_bbit0(&p, wr.r2, ilog2(sizeof(pte_t)), 8); wr 2253 arch/mips/mm/tlbex.c uasm_i_andi(&p, wr.r3, wr.r2, sizeof(pte_t)); wr 2254 arch/mips/mm/tlbex.c uasm_i_beqz(&p, wr.r3, 8); wr 2257 arch/mips/mm/tlbex.c UASM_i_MFC0(&p, wr.r3, C0_ENTRYLO0); wr 2259 arch/mips/mm/tlbex.c UASM_i_MFC0(&p, wr.r3, C0_ENTRYLO1); wr 2265 arch/mips/mm/tlbex.c uasm_il_bbit0(&p, &r, wr.r3, 1, label_tlbl_goaround2); wr 2267 arch/mips/mm/tlbex.c uasm_i_andi(&p, wr.r3, wr.r3, 2); wr 2268 arch/mips/mm/tlbex.c uasm_il_beqz(&p, &r, wr.r3, label_tlbl_goaround2); wr 2276 arch/mips/mm/tlbex.c build_restore_pagemask(&p, &r, wr.r3, label_nopage_tlbl, 0); wr 2280 arch/mips/mm/tlbex.c uasm_i_ori(&p, wr.r1, wr.r1, (_PAGE_ACCESSED | _PAGE_VALID)); wr 2281 arch/mips/mm/tlbex.c build_huge_handler_tail(&p, &r, &l, wr.r1, wr.r2, 1); wr 2313 arch/mips/mm/tlbex.c struct work_registers wr; wr 2319 arch/mips/mm/tlbex.c wr = build_r4000_tlbchange_handler_head(&p, &l, &r); wr 2320 arch/mips/mm/tlbex.c build_pte_writable(&p, &r, wr.r1, wr.r2, wr.r3, label_nopage_tlbs); wr 2323 arch/mips/mm/tlbex.c build_make_write(&p, &r, wr.r1, wr.r2, wr.r3); wr 2324 arch/mips/mm/tlbex.c build_r4000_tlbchange_handler_tail(&p, &l, &r, wr.r1, wr.r2); wr 2332 arch/mips/mm/tlbex.c iPTE_LW(&p, wr.r1, wr.r2); wr 2333 arch/mips/mm/tlbex.c build_pte_writable(&p, &r, wr.r1, wr.r2, wr.r3, label_nopage_tlbs); wr 2335 arch/mips/mm/tlbex.c uasm_i_ori(&p, wr.r1, wr.r1, wr 2337 arch/mips/mm/tlbex.c build_huge_handler_tail(&p, &r, &l, wr.r1, wr.r2, 1); wr 2369 arch/mips/mm/tlbex.c struct work_registers wr; wr 2375 arch/mips/mm/tlbex.c wr = build_r4000_tlbchange_handler_head(&p, &l, &r); wr 2376 arch/mips/mm/tlbex.c build_pte_modifiable(&p, &r, wr.r1, wr.r2, wr.r3, label_nopage_tlbm); wr 2380 arch/mips/mm/tlbex.c build_make_write(&p, &r, wr.r1, wr.r2, wr.r3); wr 2381 arch/mips/mm/tlbex.c build_r4000_tlbchange_handler_tail(&p, &l, &r, wr.r1, wr.r2); wr 2389 arch/mips/mm/tlbex.c iPTE_LW(&p, wr.r1, wr.r2); wr 2390 arch/mips/mm/tlbex.c build_pte_modifiable(&p, &r, wr.r1, wr.r2, wr.r3, label_nopage_tlbm); wr 2392 arch/mips/mm/tlbex.c uasm_i_ori(&p, wr.r1, wr.r1, wr 2394 arch/mips/mm/tlbex.c build_huge_handler_tail(&p, &r, &l, wr.r1, wr.r2, 0); wr 63 arch/sparc/include/asm/backoff.h wr tmp, 0, %asr27; \ wr 23 arch/sparc/include/asm/visasm.h 297: wr %g0, FPRS_FEF, %fprs; \ wr 26 arch/sparc/include/asm/visasm.h wr %g0, 0, %fprs; wr 43 arch/sparc/include/asm/visasm.h 297: wr %o5, FPRS_FEF, %fprs; wr 46 arch/sparc/include/asm/visasm.h wr %o5, 0, %fprs; wr 52 arch/sparc/include/asm/winmacro.h wr %scratch, 0x0, %y; wr 4208 arch/x86/kvm/vmx/nested.c u32 vmx_instruction_info, bool wr, int len, gva_t *ret) wr 4292 arch/x86/kvm/vmx/nested.c if (wr) wr 34 arch/x86/kvm/vmx/nested.h u32 vmx_instruction_info, bool wr, int len, gva_t *ret); wr 169 drivers/crypto/atmel-sha.c static const char *atmel_sha_reg_name(u32 offset, char *tmp, size_t sz, bool wr) wr 231 drivers/crypto/atmel-sha.c if (wr) wr 529 drivers/crypto/chelsio/chcr_ipsec.c struct chcr_ipsec_wr *wr; wr 564 drivers/crypto/chelsio/chcr_ipsec.c wr = (struct chcr_ipsec_wr *)pos; wr 565 drivers/crypto/chelsio/chcr_ipsec.c wr->wreq.op_to_compl = htonl(FW_WR_OP_V(FW_ULPTX_WR)); wr 575 drivers/crypto/chelsio/chcr_ipsec.c wr->wreq.flowid_len16 = htonl(wr_mid); wr 578 drivers/crypto/chelsio/chcr_ipsec.c wr->req.ulptx.cmd_dest = FILL_ULPTX_CMD_DEST(pi->port_id, qid); wr 579 drivers/crypto/chelsio/chcr_ipsec.c wr->req.ulptx.len = htonl(ndesc - 1); wr 582 drivers/crypto/chelsio/chcr_ipsec.c wr->req.sc_imm.cmd_more = FILL_CMD_MORE(!immdatalen || sc_more); wr 583 drivers/crypto/chelsio/chcr_ipsec.c wr->req.sc_imm.len = cpu_to_be32(sizeof(struct cpl_tx_sec_pdu) + wr 584 drivers/crypto/chelsio/chcr_ipsec.c sizeof(wr->req.key_ctx) + wr 594 drivers/crypto/chelsio/chcr_ipsec.c wr->req.sec_cpl.op_ivinsrtofst = htonl( wr 601 drivers/crypto/chelsio/chcr_ipsec.c wr->req.sec_cpl.pldlen = htonl(skb->len + esnlen); wr 610 drivers/crypto/chelsio/chcr_ipsec.c wr->req.sec_cpl.aadstart_cipherstop_hi = FILL_SEC_CPL_CIPHERSTOP_HI( wr 615 drivers/crypto/chelsio/chcr_ipsec.c wr->req.sec_cpl.cipherstop_lo_authinsert = wr 619 drivers/crypto/chelsio/chcr_ipsec.c wr->req.sec_cpl.seqno_numivs = wr 625 drivers/crypto/chelsio/chcr_ipsec.c wr->req.sec_cpl.ivgen_hdrlen = FILL_SEC_CPL_IVGEN_HDRLEN(0, 0, 1, wr 305 drivers/crypto/chelsio/chtls/chtls.h struct tls_key_wr wr; wr 329 drivers/crypto/chelsio/chtls/chtls.h struct wr_skb_cb wr; /* reserve space for write request */ wr 270 drivers/crypto/chelsio/chtls/chtls_cm.c req->wr.wr_hi = htonl(FW_WR_OP_V(FW_TP_WR) | wr 272 drivers/crypto/chelsio/chtls/chtls_cm.c sizeof(req->wr))); wr 273 drivers/crypto/chelsio/chtls/chtls_cm.c req->wr.wr_mid = htonl(FW_WR_LEN16_V(DIV_ROUND_UP(sizeof(*req), 16)) | wr 30 drivers/crypto/chelsio/chtls/chtls_hw.c req->wr.wr_mid |= htonl(FW_WR_FLOWID_V(csk->tid)); wr 330 drivers/crypto/chelsio/chtls/chtls_hw.c kwr->wr.op_to_compl = wr 333 drivers/crypto/chelsio/chtls/chtls_hw.c kwr->wr.flowid_len16 = wr 336 drivers/crypto/chelsio/chtls/chtls_hw.c kwr->wr.protocol = 0; wr 337 drivers/crypto/chelsio/chtls/chtls_hw.c kwr->wr.mfs = htons(TLS_MFS); wr 338 drivers/crypto/chelsio/chtls/chtls_hw.c kwr->wr.reneg_to_write_rx = optname; wr 345 drivers/crypto/chelsio/chtls/chtls_hw.c DIV_ROUND_UP(len - sizeof(kwr->wr), 16)); wr 699 drivers/crypto/chelsio/chtls/chtls_io.c req->wr.wr_hi |= htonl(FW_WR_COMPL_F); wr 63 drivers/dma/dw-edma/dw-edma-v0-core.c return &__dw_regs(dw)->type.unroll.ch[ch].wr; wr 190 drivers/dma/dw-edma/dw-edma-v0-debugfs.c dw_edma_debugfs_regs_ch(®s->type.unroll.ch[i].wr, ch_dir); wr 192 drivers/dma/dw-edma/dw-edma-v0-debugfs.c lim[0][i].start = ®s->type.unroll.ch[i].wr; wr 40 drivers/dma/dw-edma/dw-edma-v0-regs.h struct dw_edma_v0_ch_regs wr; /* 0x200 */ wr 126 drivers/gpu/drm/nouveau/nvif/object.c struct nvif_ioctl_wr_v0 wr; wr 129 drivers/gpu/drm/nouveau/nvif/object.c .wr.size = size, wr 130 drivers/gpu/drm/nouveau/nvif/object.c .wr.addr = addr, wr 131 drivers/gpu/drm/nouveau/nvif/object.c .wr.data = data, wr 69 drivers/gpu/drm/udl/udl_dmabuf.c struct scatterlist *rd, *wr; wr 108 drivers/gpu/drm/udl/udl_dmabuf.c wr = sgt->sgl; wr 110 drivers/gpu/drm/udl/udl_dmabuf.c sg_set_page(wr, sg_page(rd), rd->length, rd->offset); wr 112 drivers/gpu/drm/udl/udl_dmabuf.c wr = sg_next(wr); wr 511 drivers/i2c/busses/i2c-ocores.c u32 curr, wr; wr 519 drivers/i2c/busses/i2c-ocores.c wr = (curr & 0xff00) | value; wr 521 drivers/i2c/busses/i2c-ocores.c wr = (((u32)value) << 8) | (curr & 0xff); wr 523 drivers/i2c/busses/i2c-ocores.c wr = value; wr 525 drivers/i2c/busses/i2c-ocores.c iowrite32be(wr, i2c->base + (rreg << i2c->reg_shift)); wr 1460 drivers/ide/pmac.c int wr = !!(cmd->tf_flags & IDE_TFLAG_WRITE); wr 1496 drivers/ide/pmac.c table->command = cpu_to_le16(wr? OUTPUT_MORE: INPUT_MORE); wr 1512 drivers/ide/pmac.c table[-1].command = cpu_to_le16(wr? OUTPUT_LAST: INPUT_LAST); wr 65 drivers/infiniband/core/mad.c struct ib_ud_wr *wr = &mad_send_wr->send_wr; wr 68 drivers/infiniband/core/mad.c rdma_query_ah(wr->ah, &attr); wr 72 drivers/infiniband/core/mad.c ib_query_pkey(dev, pnum, wr->pkey_index, &pkey); wr 74 drivers/infiniband/core/mad.c entry->rqpn = wr->remote_qpn; wr 75 drivers/infiniband/core/mad.c entry->rqkey = wr->remote_qkey; wr 904 drivers/infiniband/core/mad.c send_wr->wr.wr_cqe, drslid, wr 1110 drivers/infiniband/core/mad.c mad_send_wr->send_wr.wr.wr_cqe = &mad_send_wr->mad_list.cqe; wr 1111 drivers/infiniband/core/mad.c mad_send_wr->send_wr.wr.sg_list = mad_send_wr->sg_list; wr 1112 drivers/infiniband/core/mad.c mad_send_wr->send_wr.wr.num_sge = 2; wr 1113 drivers/infiniband/core/mad.c mad_send_wr->send_wr.wr.opcode = IB_WR_SEND; wr 1114 drivers/infiniband/core/mad.c mad_send_wr->send_wr.wr.send_flags = IB_SEND_SIGNALED; wr 1223 drivers/infiniband/core/mad.c mad_send_wr->send_wr.wr.wr_cqe = &mad_send_wr->mad_list.cqe; wr 1251 drivers/infiniband/core/mad.c ret = ib_post_send(mad_agent->qp, &mad_send_wr->send_wr.wr, wr 1901 drivers/infiniband/core/mad.c static inline int rcv_has_same_class(const struct ib_mad_send_wr_private *wr, wr 1904 drivers/infiniband/core/mad.c return ((struct ib_mad_hdr *)(wr->send_buf.mad))->mgmt_class == wr 1909 drivers/infiniband/core/mad.c const struct ib_mad_send_wr_private *wr, wr 1920 drivers/infiniband/core/mad.c send_resp = ib_response_mad((struct ib_mad_hdr *)wr->send_buf.mad); wr 1927 drivers/infiniband/core/mad.c if (rdma_query_ah(wr->send_buf.ah, &attr)) wr 1973 drivers/infiniband/core/mad.c struct ib_mad_send_wr_private *wr; wr 1978 drivers/infiniband/core/mad.c list_for_each_entry(wr, &mad_agent_priv->wait_list, agent_list) { wr 1979 drivers/infiniband/core/mad.c if ((wr->tid == mad_hdr->tid) && wr 1980 drivers/infiniband/core/mad.c rcv_has_same_class(wr, wc) && wr 1986 drivers/infiniband/core/mad.c rcv_has_same_gid(mad_agent_priv, wr, wc))) wr 1987 drivers/infiniband/core/mad.c return (wr->status == IB_WC_SUCCESS) ? wr : NULL; wr 1994 drivers/infiniband/core/mad.c list_for_each_entry(wr, &mad_agent_priv->send_list, agent_list) { wr 1995 drivers/infiniband/core/mad.c if (is_rmpp_data_mad(mad_agent_priv, wr->send_buf.mad) && wr 1996 drivers/infiniband/core/mad.c wr->tid == mad_hdr->tid && wr 1997 drivers/infiniband/core/mad.c wr->timeout && wr 1998 drivers/infiniband/core/mad.c rcv_has_same_class(wr, wc) && wr 2004 drivers/infiniband/core/mad.c rcv_has_same_gid(mad_agent_priv, wr, wc))) wr 2006 drivers/infiniband/core/mad.c return (wr->status == IB_WC_SUCCESS) ? wr : NULL; wr 2567 drivers/infiniband/core/mad.c ret = ib_post_send(qp_info->qp, &queued_send_wr->send_wr.wr, wr 2615 drivers/infiniband/core/mad.c ret = ib_post_send(qp_info->qp, &mad_send_wr->send_wr.wr, wr 2782 drivers/infiniband/core/mad.c local->mad_send_wr->send_wr.wr.wr_cqe, wr 635 drivers/infiniband/core/mad_rmpp.c static inline void adjust_last_ack(struct ib_mad_send_wr_private *wr, wr 640 drivers/infiniband/core/mad_rmpp.c wr->last_ack = seg_num; wr 641 drivers/infiniband/core/mad_rmpp.c list = &wr->last_ack_seg->list; wr 642 drivers/infiniband/core/mad_rmpp.c list_for_each_entry(wr->last_ack_seg, list, list) wr 643 drivers/infiniband/core/mad_rmpp.c if (wr->last_ack_seg->num == seg_num) wr 75 drivers/infiniband/core/rw.c reg->inv_wr.next = ®->reg_wr.wr; wr 106 drivers/infiniband/core/rw.c reg->reg_wr.wr.opcode = IB_WR_REG_MR; wr 146 drivers/infiniband/core/rw.c prev->wr.wr.next = ®->inv_wr; wr 148 drivers/infiniband/core/rw.c prev->wr.wr.next = ®->reg_wr.wr; wr 151 drivers/infiniband/core/rw.c reg->reg_wr.wr.next = ®->wr.wr; wr 153 drivers/infiniband/core/rw.c reg->wr.wr.sg_list = ®->sge; wr 154 drivers/infiniband/core/rw.c reg->wr.wr.num_sge = 1; wr 155 drivers/infiniband/core/rw.c reg->wr.remote_addr = remote_addr; wr 156 drivers/infiniband/core/rw.c reg->wr.rkey = rkey; wr 158 drivers/infiniband/core/rw.c reg->wr.wr.opcode = IB_WR_RDMA_WRITE; wr 160 drivers/infiniband/core/rw.c reg->wr.wr.opcode = IB_WR_RDMA_READ; wr 162 drivers/infiniband/core/rw.c reg->wr.wr.opcode = IB_WR_RDMA_READ_WITH_INV; wr 163 drivers/infiniband/core/rw.c reg->wr.wr.ex.invalidate_rkey = reg->mr->lkey; wr 176 drivers/infiniband/core/rw.c prev->wr.wr.next = NULL; wr 213 drivers/infiniband/core/rw.c rdma_wr->wr.opcode = IB_WR_RDMA_WRITE; wr 215 drivers/infiniband/core/rw.c rdma_wr->wr.opcode = IB_WR_RDMA_READ; wr 218 drivers/infiniband/core/rw.c rdma_wr->wr.num_sge = nr_sge; wr 219 drivers/infiniband/core/rw.c rdma_wr->wr.sg_list = sge; wr 232 drivers/infiniband/core/rw.c rdma_wr->wr.next = i + 1 < ctx->nr_ops ? wr 233 drivers/infiniband/core/rw.c &ctx->map.wrs[i + 1].wr : NULL; wr 249 drivers/infiniband/core/rw.c struct ib_rdma_wr *rdma_wr = &ctx->single.wr; wr 259 drivers/infiniband/core/rw.c rdma_wr->wr.opcode = IB_WR_RDMA_WRITE; wr 261 drivers/infiniband/core/rw.c rdma_wr->wr.opcode = IB_WR_RDMA_READ; wr 262 drivers/infiniband/core/rw.c rdma_wr->wr.sg_list = &ctx->single.sge; wr 263 drivers/infiniband/core/rw.c rdma_wr->wr.num_sge = 1; wr 428 drivers/infiniband/core/rw.c ctx->reg->reg_wr.wr.opcode = IB_WR_REG_MR_INTEGRITY; wr 429 drivers/infiniband/core/rw.c ctx->reg->reg_wr.wr.wr_cqe = NULL; wr 430 drivers/infiniband/core/rw.c ctx->reg->reg_wr.wr.num_sge = 0; wr 431 drivers/infiniband/core/rw.c ctx->reg->reg_wr.wr.send_flags = 0; wr 444 drivers/infiniband/core/rw.c rdma_wr = &ctx->reg->wr; wr 445 drivers/infiniband/core/rw.c rdma_wr->wr.sg_list = &ctx->reg->sge; wr 446 drivers/infiniband/core/rw.c rdma_wr->wr.num_sge = 1; wr 450 drivers/infiniband/core/rw.c rdma_wr->wr.opcode = IB_WR_RDMA_WRITE; wr 452 drivers/infiniband/core/rw.c rdma_wr->wr.opcode = IB_WR_RDMA_READ; wr 453 drivers/infiniband/core/rw.c ctx->reg->reg_wr.wr.next = &rdma_wr->wr; wr 511 drivers/infiniband/core/rw.c ctx->reg[i].wr.wr.opcode != wr 518 drivers/infiniband/core/rw.c first_wr = &ctx->reg[0].reg_wr.wr; wr 519 drivers/infiniband/core/rw.c last_wr = &ctx->reg[ctx->nr_ops - 1].wr.wr; wr 522 drivers/infiniband/core/rw.c first_wr = &ctx->map.wrs[0].wr; wr 523 drivers/infiniband/core/rw.c last_wr = &ctx->map.wrs[ctx->nr_ops - 1].wr; wr 526 drivers/infiniband/core/rw.c first_wr = &ctx->single.wr.wr; wr 527 drivers/infiniband/core/rw.c last_wr = &ctx->single.wr.wr; wr 2004 drivers/infiniband/core/uverbs_cmd.c struct ib_send_wr *wr = NULL, *last, *next; wr 2071 drivers/infiniband/core/uverbs_cmd.c user_wr->wr.ud.ah, attrs); wr 2077 drivers/infiniband/core/uverbs_cmd.c ud->remote_qpn = user_wr->wr.ud.remote_qpn; wr 2078 drivers/infiniband/core/uverbs_cmd.c ud->remote_qkey = user_wr->wr.ud.remote_qkey; wr 2080 drivers/infiniband/core/uverbs_cmd.c next = &ud->wr; wr 2093 drivers/infiniband/core/uverbs_cmd.c rdma->remote_addr = user_wr->wr.rdma.remote_addr; wr 2094 drivers/infiniband/core/uverbs_cmd.c rdma->rkey = user_wr->wr.rdma.rkey; wr 2096 drivers/infiniband/core/uverbs_cmd.c next = &rdma->wr; wr 2108 drivers/infiniband/core/uverbs_cmd.c atomic->remote_addr = user_wr->wr.atomic.remote_addr; wr 2109 drivers/infiniband/core/uverbs_cmd.c atomic->compare_add = user_wr->wr.atomic.compare_add; wr 2110 drivers/infiniband/core/uverbs_cmd.c atomic->swap = user_wr->wr.atomic.swap; wr 2111 drivers/infiniband/core/uverbs_cmd.c atomic->rkey = user_wr->wr.atomic.rkey; wr 2113 drivers/infiniband/core/uverbs_cmd.c next = &atomic->wr; wr 2137 drivers/infiniband/core/uverbs_cmd.c wr = next; wr 2163 drivers/infiniband/core/uverbs_cmd.c ret = qp->device->ops.post_send(qp->real_qp, wr, &bad_wr); wr 2165 drivers/infiniband/core/uverbs_cmd.c for (next = wr; next; next = next->next) { wr 2178 drivers/infiniband/core/uverbs_cmd.c while (wr) { wr 2179 drivers/infiniband/core/uverbs_cmd.c if (is_ud && ud_wr(wr)->ah) wr 2180 drivers/infiniband/core/uverbs_cmd.c uobj_put_obj_read(ud_wr(wr)->ah); wr 2181 drivers/infiniband/core/uverbs_cmd.c next = wr->next; wr 2182 drivers/infiniband/core/uverbs_cmd.c kfree(wr); wr 2183 drivers/infiniband/core/uverbs_cmd.c wr = next; wr 2197 drivers/infiniband/core/uverbs_cmd.c struct ib_recv_wr *wr = NULL, *last, *next; wr 2252 drivers/infiniband/core/uverbs_cmd.c wr = next; wr 2276 drivers/infiniband/core/uverbs_cmd.c return wr; wr 2281 drivers/infiniband/core/uverbs_cmd.c while (wr) { wr 2282 drivers/infiniband/core/uverbs_cmd.c next = wr->next; wr 2283 drivers/infiniband/core/uverbs_cmd.c kfree(wr); wr 2284 drivers/infiniband/core/uverbs_cmd.c wr = next; wr 2294 drivers/infiniband/core/uverbs_cmd.c struct ib_recv_wr *wr, *next; wr 2304 drivers/infiniband/core/uverbs_cmd.c wr = ib_uverbs_unmarshall_recv(&iter, cmd.wr_count, cmd.wqe_size, wr 2306 drivers/infiniband/core/uverbs_cmd.c if (IS_ERR(wr)) wr 2307 drivers/infiniband/core/uverbs_cmd.c return PTR_ERR(wr); wr 2316 drivers/infiniband/core/uverbs_cmd.c ret = qp->device->ops.post_recv(qp->real_qp, wr, &bad_wr); wr 2320 drivers/infiniband/core/uverbs_cmd.c for (next = wr; next; next = next->next) { wr 2331 drivers/infiniband/core/uverbs_cmd.c while (wr) { wr 2332 drivers/infiniband/core/uverbs_cmd.c next = wr->next; wr 2333 drivers/infiniband/core/uverbs_cmd.c kfree(wr); wr 2334 drivers/infiniband/core/uverbs_cmd.c wr = next; wr 2344 drivers/infiniband/core/uverbs_cmd.c struct ib_recv_wr *wr, *next; wr 2354 drivers/infiniband/core/uverbs_cmd.c wr = ib_uverbs_unmarshall_recv(&iter, cmd.wr_count, cmd.wqe_size, wr 2356 drivers/infiniband/core/uverbs_cmd.c if (IS_ERR(wr)) wr 2357 drivers/infiniband/core/uverbs_cmd.c return PTR_ERR(wr); wr 2366 drivers/infiniband/core/uverbs_cmd.c ret = srq->device->ops.post_srq_recv(srq, wr, &bad_wr); wr 2371 drivers/infiniband/core/uverbs_cmd.c for (next = wr; next; next = next->next) { wr 2382 drivers/infiniband/core/uverbs_cmd.c while (wr) { wr 2383 drivers/infiniband/core/uverbs_cmd.c next = wr->next; wr 2384 drivers/infiniband/core/uverbs_cmd.c kfree(wr); wr 2385 drivers/infiniband/core/uverbs_cmd.c wr = next; wr 2636 drivers/infiniband/core/verbs.c .wr = { wr 2653 drivers/infiniband/core/verbs.c ret = ib_post_send(qp, &swr.wr, NULL); wr 1474 drivers/infiniband/hw/bnxt_re/ib_verbs.c int bnxt_re_post_srq_recv(struct ib_srq *ib_srq, const struct ib_recv_wr *wr, wr 1484 drivers/infiniband/hw/bnxt_re/ib_verbs.c while (wr) { wr 1486 drivers/infiniband/hw/bnxt_re/ib_verbs.c wqe.num_sge = wr->num_sge; wr 1487 drivers/infiniband/hw/bnxt_re/ib_verbs.c bnxt_re_build_sgl(wr->sg_list, wqe.sg_list, wr->num_sge); wr 1488 drivers/infiniband/hw/bnxt_re/ib_verbs.c wqe.wr_id = wr->wr_id; wr 1493 drivers/infiniband/hw/bnxt_re/ib_verbs.c *bad_wr = wr; wr 1496 drivers/infiniband/hw/bnxt_re/ib_verbs.c wr = wr->next; wr 1836 drivers/infiniband/hw/bnxt_re/ib_verbs.c const struct ib_send_wr *wr, wr 1840 drivers/infiniband/hw/bnxt_re/ib_verbs.c struct bnxt_re_ah *ah = container_of(ud_wr(wr)->ah, struct bnxt_re_ah, wr 1937 drivers/infiniband/hw/bnxt_re/ib_verbs.c if (wr->opcode == IB_WR_SEND_WITH_IMM) { wr 1943 drivers/infiniband/hw/bnxt_re/ib_verbs.c if (wr->send_flags & IB_SEND_SOLICITED) wr 2010 drivers/infiniband/hw/bnxt_re/ib_verbs.c const struct ib_recv_wr *wr, wr 2056 drivers/infiniband/hw/bnxt_re/ib_verbs.c const struct ib_send_wr *wr, wr 2062 drivers/infiniband/hw/bnxt_re/ib_verbs.c ah = container_of(ud_wr(wr)->ah, struct bnxt_re_ah, ib_ah); wr 2063 drivers/infiniband/hw/bnxt_re/ib_verbs.c wqe->send.q_key = ud_wr(wr)->remote_qkey; wr 2064 drivers/infiniband/hw/bnxt_re/ib_verbs.c wqe->send.dst_qp = ud_wr(wr)->remote_qpn; wr 2067 drivers/infiniband/hw/bnxt_re/ib_verbs.c switch (wr->opcode) { wr 2073 drivers/infiniband/hw/bnxt_re/ib_verbs.c wqe->send.imm_data = wr->ex.imm_data; wr 2077 drivers/infiniband/hw/bnxt_re/ib_verbs.c wqe->send.inv_key = wr->ex.invalidate_rkey; wr 2082 drivers/infiniband/hw/bnxt_re/ib_verbs.c if (wr->send_flags & IB_SEND_SIGNALED) wr 2084 drivers/infiniband/hw/bnxt_re/ib_verbs.c if (wr->send_flags & IB_SEND_FENCE) wr 2086 drivers/infiniband/hw/bnxt_re/ib_verbs.c if (wr->send_flags & IB_SEND_SOLICITED) wr 2088 drivers/infiniband/hw/bnxt_re/ib_verbs.c if (wr->send_flags & IB_SEND_INLINE) wr 2094 drivers/infiniband/hw/bnxt_re/ib_verbs.c static int bnxt_re_build_rdma_wqe(const struct ib_send_wr *wr, wr 2097 drivers/infiniband/hw/bnxt_re/ib_verbs.c switch (wr->opcode) { wr 2103 drivers/infiniband/hw/bnxt_re/ib_verbs.c wqe->rdma.imm_data = wr->ex.imm_data; wr 2107 drivers/infiniband/hw/bnxt_re/ib_verbs.c wqe->rdma.inv_key = wr->ex.invalidate_rkey; wr 2112 drivers/infiniband/hw/bnxt_re/ib_verbs.c wqe->rdma.remote_va = rdma_wr(wr)->remote_addr; wr 2113 drivers/infiniband/hw/bnxt_re/ib_verbs.c wqe->rdma.r_key = rdma_wr(wr)->rkey; wr 2114 drivers/infiniband/hw/bnxt_re/ib_verbs.c if (wr->send_flags & IB_SEND_SIGNALED) wr 2116 drivers/infiniband/hw/bnxt_re/ib_verbs.c if (wr->send_flags & IB_SEND_FENCE) wr 2118 drivers/infiniband/hw/bnxt_re/ib_verbs.c if (wr->send_flags & IB_SEND_SOLICITED) wr 2120 drivers/infiniband/hw/bnxt_re/ib_verbs.c if (wr->send_flags & IB_SEND_INLINE) wr 2126 drivers/infiniband/hw/bnxt_re/ib_verbs.c static int bnxt_re_build_atomic_wqe(const struct ib_send_wr *wr, wr 2129 drivers/infiniband/hw/bnxt_re/ib_verbs.c switch (wr->opcode) { wr 2132 drivers/infiniband/hw/bnxt_re/ib_verbs.c wqe->atomic.cmp_data = atomic_wr(wr)->compare_add; wr 2133 drivers/infiniband/hw/bnxt_re/ib_verbs.c wqe->atomic.swap_data = atomic_wr(wr)->swap; wr 2137 drivers/infiniband/hw/bnxt_re/ib_verbs.c wqe->atomic.cmp_data = atomic_wr(wr)->compare_add; wr 2142 drivers/infiniband/hw/bnxt_re/ib_verbs.c wqe->atomic.remote_va = atomic_wr(wr)->remote_addr; wr 2143 drivers/infiniband/hw/bnxt_re/ib_verbs.c wqe->atomic.r_key = atomic_wr(wr)->rkey; wr 2144 drivers/infiniband/hw/bnxt_re/ib_verbs.c if (wr->send_flags & IB_SEND_SIGNALED) wr 2146 drivers/infiniband/hw/bnxt_re/ib_verbs.c if (wr->send_flags & IB_SEND_FENCE) wr 2148 drivers/infiniband/hw/bnxt_re/ib_verbs.c if (wr->send_flags & IB_SEND_SOLICITED) wr 2153 drivers/infiniband/hw/bnxt_re/ib_verbs.c static int bnxt_re_build_inv_wqe(const struct ib_send_wr *wr, wr 2157 drivers/infiniband/hw/bnxt_re/ib_verbs.c wqe->local_inv.inv_l_key = wr->ex.invalidate_rkey; wr 2164 drivers/infiniband/hw/bnxt_re/ib_verbs.c if (wr->send_flags & IB_SEND_SIGNALED) wr 2166 drivers/infiniband/hw/bnxt_re/ib_verbs.c if (wr->send_flags & IB_SEND_SOLICITED) wr 2172 drivers/infiniband/hw/bnxt_re/ib_verbs.c static int bnxt_re_build_reg_wqe(const struct ib_reg_wr *wr, wr 2175 drivers/infiniband/hw/bnxt_re/ib_verbs.c struct bnxt_re_mr *mr = container_of(wr->mr, struct bnxt_re_mr, ib_mr); wr 2177 drivers/infiniband/hw/bnxt_re/ib_verbs.c int access = wr->access; wr 2192 drivers/infiniband/hw/bnxt_re/ib_verbs.c if (wr->wr.send_flags & IB_SEND_SIGNALED) wr 2206 drivers/infiniband/hw/bnxt_re/ib_verbs.c wqe->frmr.l_key = wr->key; wr 2207 drivers/infiniband/hw/bnxt_re/ib_verbs.c wqe->frmr.length = wr->mr->length; wr 2208 drivers/infiniband/hw/bnxt_re/ib_verbs.c wqe->frmr.pbl_pg_sz_log = (wr->mr->page_size >> PAGE_SHIFT_4K) - 1; wr 2209 drivers/infiniband/hw/bnxt_re/ib_verbs.c wqe->frmr.va = wr->mr->iova; wr 2214 drivers/infiniband/hw/bnxt_re/ib_verbs.c const struct ib_send_wr *wr, wr 2223 drivers/infiniband/hw/bnxt_re/ib_verbs.c for (i = 0; i < wr->num_sge; i++) { wr 2225 drivers/infiniband/hw/bnxt_re/ib_verbs.c wr->sg_list[i].addr; wr 2226 drivers/infiniband/hw/bnxt_re/ib_verbs.c sge_len = wr->sg_list[i].length; wr 2234 drivers/infiniband/hw/bnxt_re/ib_verbs.c sge_len = wr->sg_list[i].length; wr 2237 drivers/infiniband/hw/bnxt_re/ib_verbs.c in_data += wr->sg_list[i].length; wr 2238 drivers/infiniband/hw/bnxt_re/ib_verbs.c wqe->inline_len += wr->sg_list[i].length; wr 2244 drivers/infiniband/hw/bnxt_re/ib_verbs.c const struct ib_send_wr *wr, wr 2249 drivers/infiniband/hw/bnxt_re/ib_verbs.c if (wr->send_flags & IB_SEND_INLINE) wr 2250 drivers/infiniband/hw/bnxt_re/ib_verbs.c payload_sz = bnxt_re_copy_inline_data(rdev, wr, wqe); wr 2252 drivers/infiniband/hw/bnxt_re/ib_verbs.c payload_sz = bnxt_re_build_sgl(wr->sg_list, wqe->sg_list, wr 2276 drivers/infiniband/hw/bnxt_re/ib_verbs.c const struct ib_send_wr *wr) wr 2284 drivers/infiniband/hw/bnxt_re/ib_verbs.c while (wr) { wr 2289 drivers/infiniband/hw/bnxt_re/ib_verbs.c wqe.num_sge = wr->num_sge; wr 2290 drivers/infiniband/hw/bnxt_re/ib_verbs.c if (wr->num_sge > qp->qplib_qp.sq.max_sge) { wr 2297 drivers/infiniband/hw/bnxt_re/ib_verbs.c payload_sz = bnxt_re_copy_wr_payload(qp->rdev, wr, &wqe); wr 2302 drivers/infiniband/hw/bnxt_re/ib_verbs.c wqe.wr_id = wr->wr_id; wr 2306 drivers/infiniband/hw/bnxt_re/ib_verbs.c rc = bnxt_re_build_send_wqe(qp, wr, &wqe); wr 2313 drivers/infiniband/hw/bnxt_re/ib_verbs.c wr->opcode, rc); wr 2316 drivers/infiniband/hw/bnxt_re/ib_verbs.c wr = wr->next; wr 2324 drivers/infiniband/hw/bnxt_re/ib_verbs.c int bnxt_re_post_send(struct ib_qp *ib_qp, const struct ib_send_wr *wr, wr 2333 drivers/infiniband/hw/bnxt_re/ib_verbs.c while (wr) { wr 2338 drivers/infiniband/hw/bnxt_re/ib_verbs.c wqe.num_sge = wr->num_sge; wr 2339 drivers/infiniband/hw/bnxt_re/ib_verbs.c if (wr->num_sge > qp->qplib_qp.sq.max_sge) { wr 2346 drivers/infiniband/hw/bnxt_re/ib_verbs.c payload_sz = bnxt_re_copy_wr_payload(qp->rdev, wr, &wqe); wr 2351 drivers/infiniband/hw/bnxt_re/ib_verbs.c wqe.wr_id = wr->wr_id; wr 2353 drivers/infiniband/hw/bnxt_re/ib_verbs.c switch (wr->opcode) { wr 2357 drivers/infiniband/hw/bnxt_re/ib_verbs.c rc = bnxt_re_build_qp1_send_v2(qp, wr, &wqe, wr 2364 drivers/infiniband/hw/bnxt_re/ib_verbs.c switch (wr->send_flags) { wr 2374 drivers/infiniband/hw/bnxt_re/ib_verbs.c rc = bnxt_re_build_send_wqe(qp, wr, &wqe); wr 2379 drivers/infiniband/hw/bnxt_re/ib_verbs.c rc = bnxt_re_build_rdma_wqe(wr, &wqe); wr 2383 drivers/infiniband/hw/bnxt_re/ib_verbs.c rc = bnxt_re_build_atomic_wqe(wr, &wqe); wr 2391 drivers/infiniband/hw/bnxt_re/ib_verbs.c rc = bnxt_re_build_inv_wqe(wr, &wqe); wr 2394 drivers/infiniband/hw/bnxt_re/ib_verbs.c rc = bnxt_re_build_reg_wqe(reg_wr(wr), &wqe); wr 2399 drivers/infiniband/hw/bnxt_re/ib_verbs.c "WR (%#x) is not supported", wr->opcode); wr 2409 drivers/infiniband/hw/bnxt_re/ib_verbs.c wr->opcode, qp->qplib_qp.state, rc); wr 2410 drivers/infiniband/hw/bnxt_re/ib_verbs.c *bad_wr = wr; wr 2413 drivers/infiniband/hw/bnxt_re/ib_verbs.c wr = wr->next; wr 2424 drivers/infiniband/hw/bnxt_re/ib_verbs.c const struct ib_recv_wr *wr) wr 2430 drivers/infiniband/hw/bnxt_re/ib_verbs.c while (wr) { wr 2435 drivers/infiniband/hw/bnxt_re/ib_verbs.c wqe.num_sge = wr->num_sge; wr 2436 drivers/infiniband/hw/bnxt_re/ib_verbs.c if (wr->num_sge > qp->qplib_qp.rq.max_sge) { wr 2442 drivers/infiniband/hw/bnxt_re/ib_verbs.c bnxt_re_build_sgl(wr->sg_list, wqe.sg_list, wr->num_sge); wr 2443 drivers/infiniband/hw/bnxt_re/ib_verbs.c wqe.wr_id = wr->wr_id; wr 2450 drivers/infiniband/hw/bnxt_re/ib_verbs.c wr = wr->next; wr 2457 drivers/infiniband/hw/bnxt_re/ib_verbs.c int bnxt_re_post_recv(struct ib_qp *ib_qp, const struct ib_recv_wr *wr, wr 2467 drivers/infiniband/hw/bnxt_re/ib_verbs.c while (wr) { wr 2472 drivers/infiniband/hw/bnxt_re/ib_verbs.c wqe.num_sge = wr->num_sge; wr 2473 drivers/infiniband/hw/bnxt_re/ib_verbs.c if (wr->num_sge > qp->qplib_qp.rq.max_sge) { wr 2477 drivers/infiniband/hw/bnxt_re/ib_verbs.c *bad_wr = wr; wr 2481 drivers/infiniband/hw/bnxt_re/ib_verbs.c payload_sz = bnxt_re_build_sgl(wr->sg_list, wqe.sg_list, wr 2482 drivers/infiniband/hw/bnxt_re/ib_verbs.c wr->num_sge); wr 2483 drivers/infiniband/hw/bnxt_re/ib_verbs.c wqe.wr_id = wr->wr_id; wr 2488 drivers/infiniband/hw/bnxt_re/ib_verbs.c rc = bnxt_re_build_qp1_shadow_qp_recv(qp, wr, &wqe, wr 2493 drivers/infiniband/hw/bnxt_re/ib_verbs.c *bad_wr = wr; wr 2503 drivers/infiniband/hw/bnxt_re/ib_verbs.c wr = wr->next; wr 2876 drivers/infiniband/hw/bnxt_re/ib_verbs.c swr = &udwr.wr; wr 168 drivers/infiniband/hw/cxgb3/cxio_hal.h int cxio_peek_cq(struct t3_wq *wr, struct t3_cq *cq, int opcode); wr 177 drivers/infiniband/hw/cxgb3/iwch_cm.c req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_FORWARD)); wr 192 drivers/infiniband/hw/cxgb3/iwch_cm.c req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_FORWARD)); wr 193 drivers/infiniband/hw/cxgb3/iwch_cm.c req->wr.wr_lo = htonl(V_WR_TID(ep->hwtid)); wr 213 drivers/infiniband/hw/cxgb3/iwch_cm.c req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_FORWARD)); wr 214 drivers/infiniband/hw/cxgb3/iwch_cm.c req->wr.wr_lo = htonl(V_WR_TID(ep->hwtid)); wr 400 drivers/infiniband/hw/cxgb3/iwch_cm.c req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_OFLD_CLOSE_CON)); wr 401 drivers/infiniband/hw/cxgb3/iwch_cm.c req->wr.wr_lo = htonl(V_WR_TID(ep->hwtid)); wr 419 drivers/infiniband/hw/cxgb3/iwch_cm.c req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_OFLD_HOST_ABORT_CON_REQ)); wr 420 drivers/infiniband/hw/cxgb3/iwch_cm.c req->wr.wr_lo = htonl(V_WR_TID(ep->hwtid)); wr 457 drivers/infiniband/hw/cxgb3/iwch_cm.c req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_FORWARD)); wr 802 drivers/infiniband/hw/cxgb3/iwch_cm.c req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_FORWARD)); wr 1207 drivers/infiniband/hw/cxgb3/iwch_cm.c req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_FORWARD)); wr 1248 drivers/infiniband/hw/cxgb3/iwch_cm.c req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_FORWARD)); wr 1293 drivers/infiniband/hw/cxgb3/iwch_cm.c rpl->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_FORWARD)); wr 1322 drivers/infiniband/hw/cxgb3/iwch_cm.c rpl->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_FORWARD)); wr 1616 drivers/infiniband/hw/cxgb3/iwch_cm.c rpl->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_OFLD_HOST_ABORT_CON_RPL)); wr 1617 drivers/infiniband/hw/cxgb3/iwch_cm.c rpl->wr.wr_lo = htonl(V_WR_TID(ep->hwtid)); wr 329 drivers/infiniband/hw/cxgb3/iwch_provider.h int iwch_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr, wr 331 drivers/infiniband/hw/cxgb3/iwch_provider.h int iwch_post_receive(struct ib_qp *ibqp, const struct ib_recv_wr *wr, wr 42 drivers/infiniband/hw/cxgb3/iwch_qp.c static int build_rdma_send(union t3_wr *wqe, const struct ib_send_wr *wr, wr 48 drivers/infiniband/hw/cxgb3/iwch_qp.c switch (wr->opcode) { wr 50 drivers/infiniband/hw/cxgb3/iwch_qp.c if (wr->send_flags & IB_SEND_SOLICITED) wr 57 drivers/infiniband/hw/cxgb3/iwch_qp.c if (wr->send_flags & IB_SEND_SOLICITED) wr 61 drivers/infiniband/hw/cxgb3/iwch_qp.c wqe->send.rem_stag = cpu_to_be32(wr->ex.invalidate_rkey); wr 66 drivers/infiniband/hw/cxgb3/iwch_qp.c if (wr->num_sge > T3_MAX_SGE) wr 72 drivers/infiniband/hw/cxgb3/iwch_qp.c for (i = 0; i < wr->num_sge; i++) { wr 73 drivers/infiniband/hw/cxgb3/iwch_qp.c if ((plen + wr->sg_list[i].length) < plen) wr 76 drivers/infiniband/hw/cxgb3/iwch_qp.c plen += wr->sg_list[i].length; wr 77 drivers/infiniband/hw/cxgb3/iwch_qp.c wqe->send.sgl[i].stag = cpu_to_be32(wr->sg_list[i].lkey); wr 78 drivers/infiniband/hw/cxgb3/iwch_qp.c wqe->send.sgl[i].len = cpu_to_be32(wr->sg_list[i].length); wr 79 drivers/infiniband/hw/cxgb3/iwch_qp.c wqe->send.sgl[i].to = cpu_to_be64(wr->sg_list[i].addr); wr 81 drivers/infiniband/hw/cxgb3/iwch_qp.c wqe->send.num_sgle = cpu_to_be32(wr->num_sge); wr 82 drivers/infiniband/hw/cxgb3/iwch_qp.c *flit_cnt = 4 + ((wr->num_sge) << 1); wr 87 drivers/infiniband/hw/cxgb3/iwch_qp.c static int build_rdma_write(union t3_wr *wqe, const struct ib_send_wr *wr, wr 92 drivers/infiniband/hw/cxgb3/iwch_qp.c if (wr->num_sge > T3_MAX_SGE) wr 98 drivers/infiniband/hw/cxgb3/iwch_qp.c wqe->write.stag_sink = cpu_to_be32(rdma_wr(wr)->rkey); wr 99 drivers/infiniband/hw/cxgb3/iwch_qp.c wqe->write.to_sink = cpu_to_be64(rdma_wr(wr)->remote_addr); wr 101 drivers/infiniband/hw/cxgb3/iwch_qp.c if (wr->opcode == IB_WR_RDMA_WRITE_WITH_IMM) { wr 103 drivers/infiniband/hw/cxgb3/iwch_qp.c wqe->write.sgl[0].stag = wr->ex.imm_data; wr 109 drivers/infiniband/hw/cxgb3/iwch_qp.c for (i = 0; i < wr->num_sge; i++) { wr 110 drivers/infiniband/hw/cxgb3/iwch_qp.c if ((plen + wr->sg_list[i].length) < plen) { wr 113 drivers/infiniband/hw/cxgb3/iwch_qp.c plen += wr->sg_list[i].length; wr 115 drivers/infiniband/hw/cxgb3/iwch_qp.c cpu_to_be32(wr->sg_list[i].lkey); wr 117 drivers/infiniband/hw/cxgb3/iwch_qp.c cpu_to_be32(wr->sg_list[i].length); wr 119 drivers/infiniband/hw/cxgb3/iwch_qp.c cpu_to_be64(wr->sg_list[i].addr); wr 121 drivers/infiniband/hw/cxgb3/iwch_qp.c wqe->write.num_sgle = cpu_to_be32(wr->num_sge); wr 122 drivers/infiniband/hw/cxgb3/iwch_qp.c *flit_cnt = 5 + ((wr->num_sge) << 1); wr 128 drivers/infiniband/hw/cxgb3/iwch_qp.c static int build_rdma_read(union t3_wr *wqe, const struct ib_send_wr *wr, wr 131 drivers/infiniband/hw/cxgb3/iwch_qp.c if (wr->num_sge > 1) wr 134 drivers/infiniband/hw/cxgb3/iwch_qp.c if (wr->opcode == IB_WR_RDMA_READ_WITH_INV) wr 140 drivers/infiniband/hw/cxgb3/iwch_qp.c wqe->read.rem_stag = cpu_to_be32(rdma_wr(wr)->rkey); wr 141 drivers/infiniband/hw/cxgb3/iwch_qp.c wqe->read.rem_to = cpu_to_be64(rdma_wr(wr)->remote_addr); wr 142 drivers/infiniband/hw/cxgb3/iwch_qp.c wqe->read.local_stag = cpu_to_be32(wr->sg_list[0].lkey); wr 143 drivers/infiniband/hw/cxgb3/iwch_qp.c wqe->read.local_len = cpu_to_be32(wr->sg_list[0].length); wr 144 drivers/infiniband/hw/cxgb3/iwch_qp.c wqe->read.local_to = cpu_to_be64(wr->sg_list[0].addr); wr 149 drivers/infiniband/hw/cxgb3/iwch_qp.c static int build_memreg(union t3_wr *wqe, const struct ib_reg_wr *wr, wr 152 drivers/infiniband/hw/cxgb3/iwch_qp.c struct iwch_mr *mhp = to_iwch_mr(wr->mr); wr 159 drivers/infiniband/hw/cxgb3/iwch_qp.c wqe->fastreg.stag = cpu_to_be32(wr->key); wr 166 drivers/infiniband/hw/cxgb3/iwch_qp.c V_FR_PAGE_SIZE(ilog2(wr->mr->page_size) - 12) | wr 168 drivers/infiniband/hw/cxgb3/iwch_qp.c V_FR_PERMS(iwch_ib_to_tpt_access(wr->access))); wr 192 drivers/infiniband/hw/cxgb3/iwch_qp.c static int build_inv_stag(union t3_wr *wqe, const struct ib_send_wr *wr, wr 195 drivers/infiniband/hw/cxgb3/iwch_qp.c wqe->local_inv.stag = cpu_to_be32(wr->ex.invalidate_rkey); wr 249 drivers/infiniband/hw/cxgb3/iwch_qp.c const struct ib_recv_wr *wr) wr 255 drivers/infiniband/hw/cxgb3/iwch_qp.c err = iwch_sgl2pbl_map(qhp->rhp, wr->sg_list, wr->num_sge, pbl_addr, wr 263 drivers/infiniband/hw/cxgb3/iwch_qp.c wqe->recv.num_sgle = cpu_to_be32(wr->num_sge); wr 264 drivers/infiniband/hw/cxgb3/iwch_qp.c for (i = 0; i < wr->num_sge; i++) { wr 265 drivers/infiniband/hw/cxgb3/iwch_qp.c wqe->recv.sgl[i].stag = cpu_to_be32(wr->sg_list[i].lkey); wr 266 drivers/infiniband/hw/cxgb3/iwch_qp.c wqe->recv.sgl[i].len = cpu_to_be32(wr->sg_list[i].length); wr 269 drivers/infiniband/hw/cxgb3/iwch_qp.c wqe->recv.sgl[i].to = cpu_to_be64(((u32)wr->sg_list[i].addr) & wr 282 drivers/infiniband/hw/cxgb3/iwch_qp.c qhp->wq.rq_size_log2)].wr_id = wr->wr_id; wr 289 drivers/infiniband/hw/cxgb3/iwch_qp.c const struct ib_recv_wr *wr) wr 311 drivers/infiniband/hw/cxgb3/iwch_qp.c wqe->recv.num_sgle = cpu_to_be32(wr->num_sge); wr 313 drivers/infiniband/hw/cxgb3/iwch_qp.c for (i = 0; i < wr->num_sge; i++) { wr 322 drivers/infiniband/hw/cxgb3/iwch_qp.c if (wr->sg_list[i].length > T3_STAG0_MAX_PBE_LEN) wr 329 drivers/infiniband/hw/cxgb3/iwch_qp.c if (wr->sg_list[i].lkey != 0) wr 332 drivers/infiniband/hw/cxgb3/iwch_qp.c wqe->recv.sgl[i].len = cpu_to_be32(wr->sg_list[i].length); wr 333 drivers/infiniband/hw/cxgb3/iwch_qp.c wqe->recv.sgl[i].to = cpu_to_be64(wr->sg_list[i].addr); wr 345 drivers/infiniband/hw/cxgb3/iwch_qp.c qhp->wq.rq_size_log2)].wr_id = wr->wr_id; wr 351 drivers/infiniband/hw/cxgb3/iwch_qp.c int iwch_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr, wr 380 drivers/infiniband/hw/cxgb3/iwch_qp.c while (wr) { wr 388 drivers/infiniband/hw/cxgb3/iwch_qp.c if (wr->send_flags & IB_SEND_SOLICITED) wr 390 drivers/infiniband/hw/cxgb3/iwch_qp.c if (wr->send_flags & IB_SEND_SIGNALED) wr 394 drivers/infiniband/hw/cxgb3/iwch_qp.c switch (wr->opcode) { wr 397 drivers/infiniband/hw/cxgb3/iwch_qp.c if (wr->send_flags & IB_SEND_FENCE) wr 400 drivers/infiniband/hw/cxgb3/iwch_qp.c err = build_rdma_send(wqe, wr, &t3_wr_flit_cnt); wr 405 drivers/infiniband/hw/cxgb3/iwch_qp.c err = build_rdma_write(wqe, wr, &t3_wr_flit_cnt); wr 411 drivers/infiniband/hw/cxgb3/iwch_qp.c err = build_rdma_read(wqe, wr, &t3_wr_flit_cnt); wr 420 drivers/infiniband/hw/cxgb3/iwch_qp.c err = build_memreg(wqe, reg_wr(wr), &t3_wr_flit_cnt, wr 424 drivers/infiniband/hw/cxgb3/iwch_qp.c if (wr->send_flags & IB_SEND_FENCE) wr 427 drivers/infiniband/hw/cxgb3/iwch_qp.c err = build_inv_stag(wqe, wr, &t3_wr_flit_cnt); wr 431 drivers/infiniband/hw/cxgb3/iwch_qp.c wr->opcode); wr 437 drivers/infiniband/hw/cxgb3/iwch_qp.c sqp->wr_id = wr->wr_id; wr 441 drivers/infiniband/hw/cxgb3/iwch_qp.c sqp->signaled = (wr->send_flags & IB_SEND_SIGNALED); wr 448 drivers/infiniband/hw/cxgb3/iwch_qp.c __func__, (unsigned long long)wr->wr_id, idx, wr 451 drivers/infiniband/hw/cxgb3/iwch_qp.c wr = wr->next; wr 462 drivers/infiniband/hw/cxgb3/iwch_qp.c *bad_wr = wr; wr 466 drivers/infiniband/hw/cxgb3/iwch_qp.c int iwch_post_receive(struct ib_qp *ibqp, const struct ib_recv_wr *wr, wr 485 drivers/infiniband/hw/cxgb3/iwch_qp.c if (!wr) { wr 490 drivers/infiniband/hw/cxgb3/iwch_qp.c while (wr) { wr 491 drivers/infiniband/hw/cxgb3/iwch_qp.c if (wr->num_sge > T3_MAX_SGE) { wr 498 drivers/infiniband/hw/cxgb3/iwch_qp.c if (wr->sg_list[0].lkey) wr 499 drivers/infiniband/hw/cxgb3/iwch_qp.c err = build_rdma_recv(qhp, wqe, wr); wr 501 drivers/infiniband/hw/cxgb3/iwch_qp.c err = build_zero_stag_recv(qhp, wqe, wr); wr 512 drivers/infiniband/hw/cxgb3/iwch_qp.c __func__, (unsigned long long)wr->wr_id, wr 516 drivers/infiniband/hw/cxgb3/iwch_qp.c wr = wr->next; wr 525 drivers/infiniband/hw/cxgb3/iwch_qp.c *bad_wr = wr; wr 971 drivers/infiniband/hw/cxgb4/iw_cxgb4.h int c4iw_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr, wr 973 drivers/infiniband/hw/cxgb4/iw_cxgb4.h int c4iw_post_receive(struct ib_qp *ibqp, const struct ib_recv_wr *wr, wr 1052 drivers/infiniband/hw/cxgb4/iw_cxgb4.h int c4iw_post_srq_recv(struct ib_srq *ibsrq, const struct ib_recv_wr *wr, wr 86 drivers/infiniband/hw/cxgb4/mem.c req->wr.wr_hi = cpu_to_be32(FW_WR_OP_V(FW_ULPTX_WR) | wr 88 drivers/infiniband/hw/cxgb4/mem.c req->wr.wr_lo = wr_waitp ? (__force __be64)(unsigned long)wr_waitp : 0L; wr 89 drivers/infiniband/hw/cxgb4/mem.c req->wr.wr_mid = cpu_to_be32(FW_WR_LEN16_V(DIV_ROUND_UP(wr_len, 16))); wr 94 drivers/infiniband/hw/cxgb4/mem.c req->len16 = cpu_to_be32(DIV_ROUND_UP(wr_len-sizeof(req->wr), 16)); wr 148 drivers/infiniband/hw/cxgb4/mem.c req->wr.wr_hi = cpu_to_be32(FW_WR_OP_V(FW_ULPTX_WR) | wr 150 drivers/infiniband/hw/cxgb4/mem.c req->wr.wr_lo = (__force __be64)(unsigned long)wr_waitp; wr 152 drivers/infiniband/hw/cxgb4/mem.c req->wr.wr_hi = cpu_to_be32(FW_WR_OP_V(FW_ULPTX_WR)); wr 153 drivers/infiniband/hw/cxgb4/mem.c req->wr.wr_mid = cpu_to_be32( wr 159 drivers/infiniband/hw/cxgb4/mem.c req->len16 = cpu_to_be32(DIV_ROUND_UP(wr_len-sizeof(req->wr), wr 414 drivers/infiniband/hw/cxgb4/qp.c const struct ib_send_wr *wr, int max, u32 *plenp) wr 422 drivers/infiniband/hw/cxgb4/qp.c for (i = 0; i < wr->num_sge; i++) { wr 423 drivers/infiniband/hw/cxgb4/qp.c if ((plen + wr->sg_list[i].length) > max) wr 425 drivers/infiniband/hw/cxgb4/qp.c srcp = (u8 *)(unsigned long)wr->sg_list[i].addr; wr 426 drivers/infiniband/hw/cxgb4/qp.c plen += wr->sg_list[i].length; wr 427 drivers/infiniband/hw/cxgb4/qp.c rem = wr->sg_list[i].length; wr 489 drivers/infiniband/hw/cxgb4/qp.c const struct ib_send_wr *wr, u8 *len16) wr 495 drivers/infiniband/hw/cxgb4/qp.c if (wr->num_sge > T4_MAX_SEND_SGE) wr 497 drivers/infiniband/hw/cxgb4/qp.c switch (wr->opcode) { wr 499 drivers/infiniband/hw/cxgb4/qp.c if (wr->send_flags & IB_SEND_SOLICITED) wr 508 drivers/infiniband/hw/cxgb4/qp.c if (wr->send_flags & IB_SEND_SOLICITED) wr 514 drivers/infiniband/hw/cxgb4/qp.c wqe->send.stag_inv = cpu_to_be32(wr->ex.invalidate_rkey); wr 524 drivers/infiniband/hw/cxgb4/qp.c if (wr->num_sge) { wr 525 drivers/infiniband/hw/cxgb4/qp.c if (wr->send_flags & IB_SEND_INLINE) { wr 526 drivers/infiniband/hw/cxgb4/qp.c ret = build_immd(sq, wqe->send.u.immd_src, wr, wr 536 drivers/infiniband/hw/cxgb4/qp.c wr->sg_list, wr->num_sge, &plen); wr 540 drivers/infiniband/hw/cxgb4/qp.c wr->num_sge * sizeof(struct fw_ri_sge); wr 556 drivers/infiniband/hw/cxgb4/qp.c const struct ib_send_wr *wr, u8 *len16) wr 562 drivers/infiniband/hw/cxgb4/qp.c if (wr->num_sge > T4_MAX_SEND_SGE) wr 569 drivers/infiniband/hw/cxgb4/qp.c if (wr->opcode == IB_WR_RDMA_WRITE_WITH_IMM) wr 570 drivers/infiniband/hw/cxgb4/qp.c wqe->write.iw_imm_data.ib_imm_data.imm_data32 = wr->ex.imm_data; wr 573 drivers/infiniband/hw/cxgb4/qp.c wqe->write.stag_sink = cpu_to_be32(rdma_wr(wr)->rkey); wr 574 drivers/infiniband/hw/cxgb4/qp.c wqe->write.to_sink = cpu_to_be64(rdma_wr(wr)->remote_addr); wr 575 drivers/infiniband/hw/cxgb4/qp.c if (wr->num_sge) { wr 576 drivers/infiniband/hw/cxgb4/qp.c if (wr->send_flags & IB_SEND_INLINE) { wr 577 drivers/infiniband/hw/cxgb4/qp.c ret = build_immd(sq, wqe->write.u.immd_src, wr, wr 587 drivers/infiniband/hw/cxgb4/qp.c wr->sg_list, wr->num_sge, &plen); wr 591 drivers/infiniband/hw/cxgb4/qp.c wr->num_sge * sizeof(struct fw_ri_sge); wr 607 drivers/infiniband/hw/cxgb4/qp.c struct ib_send_wr *wr) wr 609 drivers/infiniband/hw/cxgb4/qp.c memcpy((u8 *)immdp->data, (u8 *)(uintptr_t)wr->sg_list->addr, 16); wr 617 drivers/infiniband/hw/cxgb4/qp.c const struct ib_send_wr *wr, u8 *len16) wr 633 drivers/infiniband/hw/cxgb4/qp.c wcwr->stag_sink = cpu_to_be32(rdma_wr(wr)->rkey); wr 634 drivers/infiniband/hw/cxgb4/qp.c wcwr->to_sink = cpu_to_be64(rdma_wr(wr)->remote_addr); wr 635 drivers/infiniband/hw/cxgb4/qp.c if (wr->next->opcode == IB_WR_SEND) wr 638 drivers/infiniband/hw/cxgb4/qp.c wcwr->stag_inv = cpu_to_be32(wr->next->ex.invalidate_rkey); wr 643 drivers/infiniband/hw/cxgb4/qp.c if (wr->next->send_flags & IB_SEND_INLINE) wr 644 drivers/infiniband/hw/cxgb4/qp.c build_immd_cmpl(sq, &wcwr->u_cmpl.immd_src, wr->next); wr 647 drivers/infiniband/hw/cxgb4/qp.c &wcwr->u_cmpl.isgl_src, wr->next->sg_list, 1, NULL); wr 651 drivers/infiniband/hw/cxgb4/qp.c wcwr->u.isgl_src, wr->sg_list, wr->num_sge, &plen); wr 654 drivers/infiniband/hw/cxgb4/qp.c wr->num_sge * sizeof(struct fw_ri_sge); wr 659 drivers/infiniband/hw/cxgb4/qp.c static int build_rdma_read(union t4_wr *wqe, const struct ib_send_wr *wr, wr 662 drivers/infiniband/hw/cxgb4/qp.c if (wr->num_sge > 1) wr 664 drivers/infiniband/hw/cxgb4/qp.c if (wr->num_sge && wr->sg_list[0].length) { wr 665 drivers/infiniband/hw/cxgb4/qp.c wqe->read.stag_src = cpu_to_be32(rdma_wr(wr)->rkey); wr 666 drivers/infiniband/hw/cxgb4/qp.c wqe->read.to_src_hi = cpu_to_be32((u32)(rdma_wr(wr)->remote_addr wr 668 drivers/infiniband/hw/cxgb4/qp.c wqe->read.to_src_lo = cpu_to_be32((u32)rdma_wr(wr)->remote_addr); wr 669 drivers/infiniband/hw/cxgb4/qp.c wqe->read.stag_sink = cpu_to_be32(wr->sg_list[0].lkey); wr 670 drivers/infiniband/hw/cxgb4/qp.c wqe->read.plen = cpu_to_be32(wr->sg_list[0].length); wr 671 drivers/infiniband/hw/cxgb4/qp.c wqe->read.to_sink_hi = cpu_to_be32((u32)(wr->sg_list[0].addr wr 673 drivers/infiniband/hw/cxgb4/qp.c wqe->read.to_sink_lo = cpu_to_be32((u32)(wr->sg_list[0].addr)); wr 689 drivers/infiniband/hw/cxgb4/qp.c static void post_write_cmpl(struct c4iw_qp *qhp, const struct ib_send_wr *wr) wr 691 drivers/infiniband/hw/cxgb4/qp.c bool send_signaled = (wr->next->send_flags & IB_SEND_SIGNALED) || wr 693 drivers/infiniband/hw/cxgb4/qp.c bool write_signaled = (wr->send_flags & IB_SEND_SIGNALED) || wr 707 drivers/infiniband/hw/cxgb4/qp.c build_rdma_write_cmpl(&qhp->wq.sq, &wqe->write_cmpl, wr, &len16); wr 716 drivers/infiniband/hw/cxgb4/qp.c swsqe->wr_id = wr->wr_id; wr 732 drivers/infiniband/hw/cxgb4/qp.c if (wr->next->opcode == IB_WR_SEND) wr 740 drivers/infiniband/hw/cxgb4/qp.c swsqe->wr_id = wr->next->wr_id; wr 759 drivers/infiniband/hw/cxgb4/qp.c const struct ib_recv_wr *wr, u8 *len16) wr 765 drivers/infiniband/hw/cxgb4/qp.c &wqe->recv.isgl, wr->sg_list, wr->num_sge, NULL); wr 769 drivers/infiniband/hw/cxgb4/qp.c sizeof(wqe->recv) + wr->num_sge * sizeof(struct fw_ri_sge), 16); wr 773 drivers/infiniband/hw/cxgb4/qp.c static int build_srq_recv(union t4_recv_wr *wqe, const struct ib_recv_wr *wr, wr 779 drivers/infiniband/hw/cxgb4/qp.c &wqe->recv.isgl, wr->sg_list, wr->num_sge, NULL); wr 783 drivers/infiniband/hw/cxgb4/qp.c wr->num_sge * sizeof(struct fw_ri_sge), 16); wr 788 drivers/infiniband/hw/cxgb4/qp.c const struct ib_reg_wr *wr, struct c4iw_mr *mhp, wr 802 drivers/infiniband/hw/cxgb4/qp.c FW_RI_TPTE_PERM_V(c4iw_ib_to_tpt_access(wr->access)) | wr 804 drivers/infiniband/hw/cxgb4/qp.c FW_RI_TPTE_PS_V(ilog2(wr->mr->page_size) - 12)); wr 820 drivers/infiniband/hw/cxgb4/qp.c const struct ib_reg_wr *wr, struct c4iw_mr *mhp, wr 833 drivers/infiniband/hw/cxgb4/qp.c wqe->fr.pgsz_shift = ilog2(wr->mr->page_size) - 12; wr 835 drivers/infiniband/hw/cxgb4/qp.c wqe->fr.mem_perms = c4iw_ib_to_tpt_access(wr->access); wr 838 drivers/infiniband/hw/cxgb4/qp.c wqe->fr.stag = cpu_to_be32(wr->key); wr 883 drivers/infiniband/hw/cxgb4/qp.c static int build_inv_stag(union t4_wr *wqe, const struct ib_send_wr *wr, wr 886 drivers/infiniband/hw/cxgb4/qp.c wqe->inv.stag_inv = cpu_to_be32(wr->ex.invalidate_rkey); wr 979 drivers/infiniband/hw/cxgb4/qp.c const struct ib_send_wr *wr) wr 990 drivers/infiniband/hw/cxgb4/qp.c opcode = ib_to_fw_opcode(wr->opcode); wr 994 drivers/infiniband/hw/cxgb4/qp.c cqe.u.drain_cookie = wr->wr_id; wr 1018 drivers/infiniband/hw/cxgb4/qp.c const struct ib_send_wr *wr, wr 1023 drivers/infiniband/hw/cxgb4/qp.c while (wr) { wr 1024 drivers/infiniband/hw/cxgb4/qp.c ret = complete_sq_drain_wr(qhp, wr); wr 1026 drivers/infiniband/hw/cxgb4/qp.c *bad_wr = wr; wr 1029 drivers/infiniband/hw/cxgb4/qp.c wr = wr->next; wr 1035 drivers/infiniband/hw/cxgb4/qp.c const struct ib_recv_wr *wr) wr 1045 drivers/infiniband/hw/cxgb4/qp.c cqe.u.drain_cookie = wr->wr_id; wr 1068 drivers/infiniband/hw/cxgb4/qp.c const struct ib_recv_wr *wr) wr 1070 drivers/infiniband/hw/cxgb4/qp.c while (wr) { wr 1071 drivers/infiniband/hw/cxgb4/qp.c complete_rq_drain_wr(qhp, wr); wr 1072 drivers/infiniband/hw/cxgb4/qp.c wr = wr->next; wr 1076 drivers/infiniband/hw/cxgb4/qp.c int c4iw_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr, wr 1101 drivers/infiniband/hw/cxgb4/qp.c err = complete_sq_drain_wrs(qhp, wr, bad_wr); wr 1107 drivers/infiniband/hw/cxgb4/qp.c *bad_wr = wr; wr 1123 drivers/infiniband/hw/cxgb4/qp.c wr && wr->next && !wr->next->next && wr 1124 drivers/infiniband/hw/cxgb4/qp.c wr->opcode == IB_WR_RDMA_WRITE && wr 1125 drivers/infiniband/hw/cxgb4/qp.c wr->sg_list[0].length && wr->num_sge <= T4_WRITE_CMPL_MAX_SGL && wr 1126 drivers/infiniband/hw/cxgb4/qp.c (wr->next->opcode == IB_WR_SEND || wr 1127 drivers/infiniband/hw/cxgb4/qp.c wr->next->opcode == IB_WR_SEND_WITH_INV) && wr 1128 drivers/infiniband/hw/cxgb4/qp.c wr->next->sg_list[0].length == T4_WRITE_CMPL_MAX_CQE && wr 1129 drivers/infiniband/hw/cxgb4/qp.c wr->next->num_sge == 1 && num_wrs >= 2) { wr 1130 drivers/infiniband/hw/cxgb4/qp.c post_write_cmpl(qhp, wr); wr 1135 drivers/infiniband/hw/cxgb4/qp.c while (wr) { wr 1138 drivers/infiniband/hw/cxgb4/qp.c *bad_wr = wr; wr 1145 drivers/infiniband/hw/cxgb4/qp.c if (wr->send_flags & IB_SEND_SOLICITED) wr 1147 drivers/infiniband/hw/cxgb4/qp.c if (wr->send_flags & IB_SEND_SIGNALED || qhp->sq_sig_all) wr 1150 drivers/infiniband/hw/cxgb4/qp.c switch (wr->opcode) { wr 1153 drivers/infiniband/hw/cxgb4/qp.c if (wr->send_flags & IB_SEND_FENCE) wr 1156 drivers/infiniband/hw/cxgb4/qp.c if (wr->opcode == IB_WR_SEND) wr 1160 drivers/infiniband/hw/cxgb4/qp.c err = build_rdma_send(&qhp->wq.sq, wqe, wr, &len16); wr 1172 drivers/infiniband/hw/cxgb4/qp.c err = build_rdma_write(&qhp->wq.sq, wqe, wr, &len16); wr 1178 drivers/infiniband/hw/cxgb4/qp.c if (wr->opcode == IB_WR_RDMA_READ_WITH_INV) { wr 1179 drivers/infiniband/hw/cxgb4/qp.c c4iw_invalidate_mr(rhp, wr->sg_list[0].lkey); wr 1184 drivers/infiniband/hw/cxgb4/qp.c err = build_rdma_read(wqe, wr, &len16); wr 1187 drivers/infiniband/hw/cxgb4/qp.c swsqe->read_len = wr->sg_list[0].length; wr 1192 drivers/infiniband/hw/cxgb4/qp.c struct c4iw_mr *mhp = to_c4iw_mr(reg_wr(wr)->mr); wr 1198 drivers/infiniband/hw/cxgb4/qp.c build_tpte_memreg(&wqe->fr_tpte, reg_wr(wr), wr 1202 drivers/infiniband/hw/cxgb4/qp.c err = build_memreg(&qhp->wq.sq, wqe, reg_wr(wr), wr 1212 drivers/infiniband/hw/cxgb4/qp.c if (wr->send_flags & IB_SEND_FENCE) wr 1216 drivers/infiniband/hw/cxgb4/qp.c err = build_inv_stag(wqe, wr, &len16); wr 1217 drivers/infiniband/hw/cxgb4/qp.c c4iw_invalidate_mr(rhp, wr->ex.invalidate_rkey); wr 1221 drivers/infiniband/hw/cxgb4/qp.c wr->opcode); wr 1225 drivers/infiniband/hw/cxgb4/qp.c *bad_wr = wr; wr 1230 drivers/infiniband/hw/cxgb4/qp.c swsqe->signaled = (wr->send_flags & IB_SEND_SIGNALED) || wr 1233 drivers/infiniband/hw/cxgb4/qp.c swsqe->wr_id = wr->wr_id; wr 1243 drivers/infiniband/hw/cxgb4/qp.c (unsigned long long)wr->wr_id, qhp->wq.sq.pidx, wr 1245 drivers/infiniband/hw/cxgb4/qp.c wr = wr->next; wr 1260 drivers/infiniband/hw/cxgb4/qp.c int c4iw_post_receive(struct ib_qp *ibqp, const struct ib_recv_wr *wr, wr 1280 drivers/infiniband/hw/cxgb4/qp.c complete_rq_drain_wrs(qhp, wr); wr 1286 drivers/infiniband/hw/cxgb4/qp.c *bad_wr = wr; wr 1289 drivers/infiniband/hw/cxgb4/qp.c while (wr) { wr 1290 drivers/infiniband/hw/cxgb4/qp.c if (wr->num_sge > T4_MAX_RECV_SGE) { wr 1292 drivers/infiniband/hw/cxgb4/qp.c *bad_wr = wr; wr 1299 drivers/infiniband/hw/cxgb4/qp.c err = build_rdma_recv(qhp, wqe, wr, &len16); wr 1303 drivers/infiniband/hw/cxgb4/qp.c *bad_wr = wr; wr 1307 drivers/infiniband/hw/cxgb4/qp.c qhp->wq.rq.sw_rq[qhp->wq.rq.pidx].wr_id = wr->wr_id; wr 1324 drivers/infiniband/hw/cxgb4/qp.c (unsigned long long)wr->wr_id, qhp->wq.rq.pidx); wr 1327 drivers/infiniband/hw/cxgb4/qp.c wr = wr->next; wr 1356 drivers/infiniband/hw/cxgb4/qp.c int c4iw_post_srq_recv(struct ib_srq *ibsrq, const struct ib_recv_wr *wr, wr 1374 drivers/infiniband/hw/cxgb4/qp.c while (wr) { wr 1375 drivers/infiniband/hw/cxgb4/qp.c if (wr->num_sge > T4_MAX_RECV_SGE) { wr 1377 drivers/infiniband/hw/cxgb4/qp.c *bad_wr = wr; wr 1382 drivers/infiniband/hw/cxgb4/qp.c err = build_srq_recv(wqe, wr, &len16); wr 1386 drivers/infiniband/hw/cxgb4/qp.c *bad_wr = wr; wr 1401 drivers/infiniband/hw/cxgb4/qp.c defer_srq_wr(&srq->wq, wqe, wr->wr_id, len16); wr 1403 drivers/infiniband/hw/cxgb4/qp.c srq->wq.sw_rq[srq->wq.pidx].wr_id = wr->wr_id; wr 1410 drivers/infiniband/hw/cxgb4/qp.c (unsigned long long)wr->wr_id); wr 1414 drivers/infiniband/hw/cxgb4/qp.c wr = wr->next; wr 44 drivers/infiniband/hw/hfi1/opfn.c struct ib_atomic_wr wr; wr 83 drivers/infiniband/hw/hfi1/opfn.c memset(&wr, 0, sizeof(wr)); wr 84 drivers/infiniband/hw/hfi1/opfn.c wr.wr.opcode = IB_WR_OPFN; wr 85 drivers/infiniband/hw/hfi1/opfn.c wr.remote_addr = HFI1_VERBS_E_ATOMIC_VADDR; wr 86 drivers/infiniband/hw/hfi1/opfn.c wr.compare_add = data; wr 92 drivers/infiniband/hw/hfi1/opfn.c ret = ib_post_send(&qp->ibqp, &wr.wr, NULL); wr 674 drivers/infiniband/hw/hfi1/qp.c wqe ? wqe->wr.opcode : 0, wr 536 drivers/infiniband/hw/hfi1/rc.c if ((wqe->wr.send_flags & IB_SEND_FENCE) && wr 538 drivers/infiniband/hw/hfi1/rc.c (wqe->wr.opcode != IB_WR_TID_RDMA_READ || wr 547 drivers/infiniband/hw/hfi1/rc.c if (wqe->wr.opcode == IB_WR_REG_MR || wr 548 drivers/infiniband/hw/hfi1/rc.c wqe->wr.opcode == IB_WR_LOCAL_INV) { wr 558 drivers/infiniband/hw/hfi1/rc.c if (!(wqe->wr.send_flags & wr 562 drivers/infiniband/hw/hfi1/rc.c wqe->wr.ex.invalidate_rkey); wr 593 drivers/infiniband/hw/hfi1/rc.c switch (wqe->wr.opcode) { wr 605 drivers/infiniband/hw/hfi1/rc.c if (wqe->wr.opcode == IB_WR_SEND) { wr 607 drivers/infiniband/hw/hfi1/rc.c } else if (wqe->wr.opcode == IB_WR_SEND_WITH_IMM) { wr 610 drivers/infiniband/hw/hfi1/rc.c ohdr->u.imm_data = wqe->wr.ex.imm_data; wr 616 drivers/infiniband/hw/hfi1/rc.c wqe->wr.ex.invalidate_rkey); wr 619 drivers/infiniband/hw/hfi1/rc.c if (wqe->wr.send_flags & IB_SEND_SOLICITED) wr 647 drivers/infiniband/hw/hfi1/rc.c if (wqe->wr.opcode == IB_WR_RDMA_WRITE) { wr 653 drivers/infiniband/hw/hfi1/rc.c ohdr->u.rc.imm_data = wqe->wr.ex.imm_data; wr 655 drivers/infiniband/hw/hfi1/rc.c if (wqe->wr.send_flags & IB_SEND_SOLICITED) wr 713 drivers/infiniband/hw/hfi1/rc.c if (__w->wr.opcode != IB_WR_TID_RDMA_WRITE || wr 763 drivers/infiniband/hw/hfi1/rc.c wqe->wr.opcode, wr 803 drivers/infiniband/hw/hfi1/rc.c wqe->wr.opcode, wr 835 drivers/infiniband/hw/hfi1/rc.c qp->s_sge.num_sge = wqe->wr.num_sge; wr 889 drivers/infiniband/hw/hfi1/rc.c if (wqe->wr.opcode == IB_WR_ATOMIC_CMP_AND_SWP || wr 890 drivers/infiniband/hw/hfi1/rc.c wqe->wr.opcode == IB_WR_OPFN) { wr 917 drivers/infiniband/hw/hfi1/rc.c if (wqe->wr.opcode != IB_WR_TID_RDMA_READ) { wr 920 drivers/infiniband/hw/hfi1/rc.c qp->s_sge.num_sge = wqe->wr.num_sge; wr 929 drivers/infiniband/hw/hfi1/rc.c if (wqe->wr.opcode == IB_WR_RDMA_READ || wr 930 drivers/infiniband/hw/hfi1/rc.c wqe->wr.opcode == IB_WR_TID_RDMA_WRITE) wr 932 drivers/infiniband/hw/hfi1/rc.c else if (wqe->wr.opcode == IB_WR_TID_RDMA_READ) wr 962 drivers/infiniband/hw/hfi1/rc.c if (wqe->wr.opcode == IB_WR_SEND) { wr 964 drivers/infiniband/hw/hfi1/rc.c } else if (wqe->wr.opcode == IB_WR_SEND_WITH_IMM) { wr 967 drivers/infiniband/hw/hfi1/rc.c ohdr->u.imm_data = wqe->wr.ex.imm_data; wr 972 drivers/infiniband/hw/hfi1/rc.c ohdr->u.ieth = cpu_to_be32(wqe->wr.ex.invalidate_rkey); wr 975 drivers/infiniband/hw/hfi1/rc.c if (wqe->wr.send_flags & IB_SEND_SOLICITED) wr 1007 drivers/infiniband/hw/hfi1/rc.c if (wqe->wr.opcode == IB_WR_RDMA_WRITE) { wr 1012 drivers/infiniband/hw/hfi1/rc.c ohdr->u.imm_data = wqe->wr.ex.imm_data; wr 1014 drivers/infiniband/hw/hfi1/rc.c if (wqe->wr.send_flags & IB_SEND_SOLICITED) wr 1075 drivers/infiniband/hw/hfi1/rc.c trace_hfi1_tid_req_make_req_write(qp, 0, wqe->wr.opcode, wr 1080 drivers/infiniband/hw/hfi1/rc.c if (wqe->wr.opcode != IB_WR_TID_RDMA_READ) wr 1130 drivers/infiniband/hw/hfi1/rc.c trace_hfi1_tid_req_make_req_read(qp, 0, wqe->wr.opcode, wr 1141 drivers/infiniband/hw/hfi1/rc.c if (wqe->wr.opcode != IB_WR_TID_RDMA_READ || delta == 0 || wr 1173 drivers/infiniband/hw/hfi1/rc.c trace_hfi1_tid_req_make_req_read(qp, 0, wqe->wr.opcode, wr 1180 drivers/infiniband/hw/hfi1/rc.c wqe->wr.opcode != IB_WR_TID_RDMA_WRITE) wr 1463 drivers/infiniband/hw/hfi1/rc.c u32 opcode = wqe->wr.opcode; wr 1481 drivers/infiniband/hw/hfi1/rc.c wqe->wr.opcode, wr 1551 drivers/infiniband/hw/hfi1/rc.c opcode = wqe->wr.opcode; wr 1625 drivers/infiniband/hw/hfi1/rc.c if (wqe->wr.opcode == IB_WR_OPFN) { wr 1638 drivers/infiniband/hw/hfi1/rc.c if (wqe->wr.opcode == IB_WR_TID_RDMA_READ) { wr 1659 drivers/infiniband/hw/hfi1/rc.c if (wqe->wr.opcode == IB_WR_RDMA_READ || wr 1660 drivers/infiniband/hw/hfi1/rc.c wqe->wr.opcode == IB_WR_TID_RDMA_READ) wr 1688 drivers/infiniband/hw/hfi1/rc.c if (wqe->wr.opcode == IB_WR_RDMA_READ || wr 1689 drivers/infiniband/hw/hfi1/rc.c wqe->wr.opcode == IB_WR_TID_RDMA_READ || wr 1690 drivers/infiniband/hw/hfi1/rc.c wqe->wr.opcode == IB_WR_TID_RDMA_WRITE) wr 1826 drivers/infiniband/hw/hfi1/rc.c if (wqe->wr.opcode == IB_WR_TID_RDMA_WRITE && wr 1840 drivers/infiniband/hw/hfi1/rc.c ib_hfi1_wc_opcode[wqe->wr.opcode], wr 1886 drivers/infiniband/hw/hfi1/rc.c ib_hfi1_wc_opcode[wqe->wr.opcode], wr 1917 drivers/infiniband/hw/hfi1/rc.c if (wqe->wr.opcode != IB_WR_TID_RDMA_WRITE) wr 2043 drivers/infiniband/hw/hfi1/rc.c if (wqe->wr.opcode == IB_WR_RDMA_READ && wr 2058 drivers/infiniband/hw/hfi1/rc.c if ((wqe->wr.opcode == IB_WR_RDMA_READ && wr 2060 drivers/infiniband/hw/hfi1/rc.c (wqe->wr.opcode == IB_WR_TID_RDMA_READ && wr 2062 drivers/infiniband/hw/hfi1/rc.c ((wqe->wr.opcode == IB_WR_ATOMIC_CMP_AND_SWP || wr 2063 drivers/infiniband/hw/hfi1/rc.c wqe->wr.opcode == IB_WR_ATOMIC_FETCH_AND_ADD) && wr 2065 drivers/infiniband/hw/hfi1/rc.c (wqe->wr.opcode == IB_WR_TID_RDMA_WRITE && wr 2074 drivers/infiniband/hw/hfi1/rc.c if (wqe->wr.opcode == IB_WR_ATOMIC_CMP_AND_SWP || wr 2075 drivers/infiniband/hw/hfi1/rc.c wqe->wr.opcode == IB_WR_ATOMIC_FETCH_AND_ADD) { wr 2079 drivers/infiniband/hw/hfi1/rc.c if (wqe->wr.opcode == IB_WR_OPFN) wr 2083 drivers/infiniband/hw/hfi1/rc.c (wqe->wr.opcode == IB_WR_RDMA_READ || wr 2084 drivers/infiniband/hw/hfi1/rc.c wqe->wr.opcode == IB_WR_ATOMIC_CMP_AND_SWP || wr 2085 drivers/infiniband/hw/hfi1/rc.c wqe->wr.opcode == IB_WR_ATOMIC_FETCH_AND_ADD)) { wr 2104 drivers/infiniband/hw/hfi1/rc.c if (wqe->wr.opcode == IB_WR_TID_RDMA_WRITE) wr 2117 drivers/infiniband/hw/hfi1/rc.c if (wqe->wr.opcode == IB_WR_TID_RDMA_READ) { wr 2133 drivers/infiniband/hw/hfi1/rc.c if (__w && __w->wr.opcode == IB_WR_TID_RDMA_WRITE && wr 2198 drivers/infiniband/hw/hfi1/rc.c if (wqe->wr.opcode == IB_WR_TID_RDMA_WRITE && wr 2212 drivers/infiniband/hw/hfi1/rc.c if (!(rdi->post_parms[wqe->wr.opcode].flags & wr 2228 drivers/infiniband/hw/hfi1/rc.c if (wqe->wr.opcode == IB_WR_TID_RDMA_WRITE) { wr 2275 drivers/infiniband/hw/hfi1/rc.c if (wqe->wr.opcode == IB_WR_TID_RDMA_READ) wr 2318 drivers/infiniband/hw/hfi1/rc.c if (wqe->wr.opcode == IB_WR_RDMA_READ || wr 2319 drivers/infiniband/hw/hfi1/rc.c wqe->wr.opcode == IB_WR_TID_RDMA_READ || wr 2320 drivers/infiniband/hw/hfi1/rc.c wqe->wr.opcode == IB_WR_TID_RDMA_WRITE || wr 2321 drivers/infiniband/hw/hfi1/rc.c wqe->wr.opcode == IB_WR_ATOMIC_CMP_AND_SWP || wr 2322 drivers/infiniband/hw/hfi1/rc.c wqe->wr.opcode == IB_WR_ATOMIC_FETCH_AND_ADD) wr 2413 drivers/infiniband/hw/hfi1/rc.c if (unlikely(wqe->wr.opcode != IB_WR_RDMA_READ)) wr 2428 drivers/infiniband/hw/hfi1/rc.c if (unlikely(wqe->wr.opcode != IB_WR_RDMA_READ)) wr 2484 drivers/infiniband/hw/hfi1/rc.c if (unlikely(wqe->wr.opcode != IB_WR_RDMA_READ)) wr 1793 drivers/infiniband/hw/hfi1/tid_rdma.c trace_hfi1_tid_req_build_read_req(qp, 0, wqe->wr.opcode, wqe->psn, wr 2422 drivers/infiniband/hw/hfi1/tid_rdma.c if (wqe->wr.opcode == opcode) wr 2521 drivers/infiniband/hw/hfi1/tid_rdma.c trace_hfi1_tid_req_rcv_read_resp(qp, 0, req->e.swqe->wr.opcode, wr 2583 drivers/infiniband/hw/hfi1/tid_rdma.c if (wqe->wr.opcode == IB_WR_TID_RDMA_READ) { wr 2692 drivers/infiniband/hw/hfi1/tid_rdma.c if (wqe->wr.opcode == IB_WR_RDMA_READ || wr 2693 drivers/infiniband/hw/hfi1/tid_rdma.c wqe->wr.opcode == IB_WR_TID_RDMA_READ || wr 2694 drivers/infiniband/hw/hfi1/tid_rdma.c wqe->wr.opcode == IB_WR_ATOMIC_CMP_AND_SWP || wr 2695 drivers/infiniband/hw/hfi1/tid_rdma.c wqe->wr.opcode == IB_WR_ATOMIC_FETCH_AND_ADD) { wr 2699 drivers/infiniband/hw/hfi1/tid_rdma.c if (wqe->wr.opcode == IB_WR_TID_RDMA_READ) { wr 2730 drivers/infiniband/hw/hfi1/tid_rdma.c if (wqe->wr.opcode != IB_WR_TID_RDMA_READ) wr 2734 drivers/infiniband/hw/hfi1/tid_rdma.c trace_hfi1_tid_req_read_kdeth_eflags(qp, 0, wqe->wr.opcode, wqe->psn, wr 3047 drivers/infiniband/hw/hfi1/tid_rdma.c if (wqe->wr.opcode == IB_WR_TID_RDMA_READ) { wr 3054 drivers/infiniband/hw/hfi1/tid_rdma.c trace_hfi1_tid_req_restart_req(qp, 0, wqe->wr.opcode, wr 3065 drivers/infiniband/hw/hfi1/tid_rdma.c if (wqe->wr.opcode == IB_WR_TID_RDMA_READ) wr 3097 drivers/infiniband/hw/hfi1/tid_rdma.c if (wqe->wr.opcode == IB_WR_TID_RDMA_WRITE) { wr 3116 drivers/infiniband/hw/hfi1/tid_rdma.c if (wqe->wr.opcode == IB_WR_TID_RDMA_READ) wr 3123 drivers/infiniband/hw/hfi1/tid_rdma.c trace_hfi1_tid_req_restart_req(qp, 0, wqe->wr.opcode, wqe->psn, wr 3126 drivers/infiniband/hw/hfi1/tid_rdma.c if (wqe->wr.opcode == IB_WR_TID_RDMA_WRITE) { wr 3144 drivers/infiniband/hw/hfi1/tid_rdma.c } while (wqe->wr.opcode != IB_WR_TID_RDMA_WRITE); wr 3177 drivers/infiniband/hw/hfi1/tid_rdma.c if (wqe->wr.opcode != IB_WR_TID_RDMA_READ) wr 3211 drivers/infiniband/hw/hfi1/tid_rdma.c switch (wqe->wr.opcode) { wr 3218 drivers/infiniband/hw/hfi1/tid_rdma.c switch (prev->wr.opcode) { wr 3228 drivers/infiniband/hw/hfi1/tid_rdma.c if (prev->wr.opcode != IB_WR_TID_RDMA_WRITE) wr 3232 drivers/infiniband/hw/hfi1/tid_rdma.c switch (prev->wr.opcode) { wr 3293 drivers/infiniband/hw/hfi1/tid_rdma.c if (wqe->wr.opcode == IB_WR_RDMA_READ) { wr 3295 drivers/infiniband/hw/hfi1/tid_rdma.c wqe->wr.num_sge)) { wr 3299 drivers/infiniband/hw/hfi1/tid_rdma.c } else if (wqe->wr.opcode == IB_WR_RDMA_WRITE) { wr 3316 drivers/infiniband/hw/hfi1/tid_rdma.c wqe->wr.opcode = new_opcode; wr 3323 drivers/infiniband/hw/hfi1/tid_rdma.c if (wqe->wr.opcode == IB_WR_TID_RDMA_READ) { wr 3343 drivers/infiniband/hw/hfi1/tid_rdma.c trace_hfi1_tid_req_setup_tid_wqe(qp, 1, wqe->wr.opcode, wr 4069 drivers/infiniband/hw/hfi1/tid_rdma.c if (unlikely(wqe->wr.opcode != IB_WR_TID_RDMA_WRITE)) wr 4142 drivers/infiniband/hw/hfi1/tid_rdma.c trace_hfi1_tid_req_rcv_write_resp(qp, 0, wqe->wr.opcode, wqe->psn, wr 4173 drivers/infiniband/hw/hfi1/tid_rdma.c if (wqe->wr.opcode == IB_WR_TID_RDMA_WRITE) wr 4526 drivers/infiniband/hw/hfi1/tid_rdma.c if (wqe->wr.opcode != IB_WR_TID_RDMA_WRITE) wr 4530 drivers/infiniband/hw/hfi1/tid_rdma.c trace_hfi1_tid_req_rcv_tid_ack(qp, 0, wqe->wr.opcode, wqe->psn, wr 4547 drivers/infiniband/hw/hfi1/tid_rdma.c trace_hfi1_tid_req_rcv_tid_ack(qp, 0, wqe->wr.opcode, wqe->psn, wr 4558 drivers/infiniband/hw/hfi1/tid_rdma.c if (wqe->wr.opcode != IB_WR_TID_RDMA_WRITE) wr 4566 drivers/infiniband/hw/hfi1/tid_rdma.c trace_hfi1_tid_req_rcv_tid_ack(qp, 0, wqe->wr.opcode, wqe->psn, wr 4574 drivers/infiniband/hw/hfi1/tid_rdma.c if (wqe->wr.opcode == IB_WR_TID_RDMA_WRITE && wr 4811 drivers/infiniband/hw/hfi1/tid_rdma.c qp, 0, wqe->wr.opcode, wqe->psn, wqe->lpsn, req); wr 4988 drivers/infiniband/hw/hfi1/tid_rdma.c if (wqe->wr.opcode == IB_WR_TID_RDMA_WRITE) wr 5057 drivers/infiniband/hw/hfi1/tid_rdma.c trace_hfi1_tid_req_make_tid_pkt(qp, 0, wqe->wr.opcode, wqe->psn, wr 5064 drivers/infiniband/hw/hfi1/tid_rdma.c priv->tid_ss.num_sge = wqe->wr.num_sge; wr 5095 drivers/infiniband/hw/hfi1/tid_rdma.c trace_hfi1_tid_req_make_tid_pkt(qp, 0, wqe->wr.opcode, wr 268 drivers/infiniband/hw/hfi1/tid_rdma.h (wqe->wr.opcode == IB_WR_RDMA_READ || wr 269 drivers/infiniband/hw/hfi1/tid_rdma.h wqe->wr.opcode == IB_WR_RDMA_WRITE) && wr 137 drivers/infiniband/hw/hfi1/trace_rc.h __entry->opcode = wqe->wr.opcode; wr 797 drivers/infiniband/hw/hfi1/trace_tx.h __entry->wr_id = wqe->wr.wr_id; wr 803 drivers/infiniband/hw/hfi1/trace_tx.h __entry->opcode = wqe->wr.opcode; wr 804 drivers/infiniband/hw/hfi1/trace_tx.h __entry->send_flags = wqe->wr.send_flags; wr 129 drivers/infiniband/hw/hfi1/uc.c if (wqe->wr.opcode == IB_WR_REG_MR || wr 130 drivers/infiniband/hw/hfi1/uc.c wqe->wr.opcode == IB_WR_LOCAL_INV) { wr 138 drivers/infiniband/hw/hfi1/uc.c if (!(wqe->wr.send_flags & RVT_SEND_COMPLETION_ONLY)) { wr 140 drivers/infiniband/hw/hfi1/uc.c qp, wqe->wr.ex.invalidate_rkey); wr 155 drivers/infiniband/hw/hfi1/uc.c qp->s_sge.num_sge = wqe->wr.num_sge; wr 159 drivers/infiniband/hw/hfi1/uc.c switch (wqe->wr.opcode) { wr 167 drivers/infiniband/hw/hfi1/uc.c if (wqe->wr.opcode == IB_WR_SEND) { wr 173 drivers/infiniband/hw/hfi1/uc.c ohdr->u.imm_data = wqe->wr.ex.imm_data; wr 176 drivers/infiniband/hw/hfi1/uc.c if (wqe->wr.send_flags & IB_SEND_SOLICITED) wr 196 drivers/infiniband/hw/hfi1/uc.c if (wqe->wr.opcode == IB_WR_RDMA_WRITE) { wr 202 drivers/infiniband/hw/hfi1/uc.c ohdr->u.rc.imm_data = wqe->wr.ex.imm_data; wr 204 drivers/infiniband/hw/hfi1/uc.c if (wqe->wr.send_flags & IB_SEND_SOLICITED) wr 227 drivers/infiniband/hw/hfi1/uc.c if (wqe->wr.opcode == IB_WR_SEND) { wr 232 drivers/infiniband/hw/hfi1/uc.c ohdr->u.imm_data = wqe->wr.ex.imm_data; wr 235 drivers/infiniband/hw/hfi1/uc.c if (wqe->wr.send_flags & IB_SEND_SOLICITED) wr 252 drivers/infiniband/hw/hfi1/uc.c if (wqe->wr.opcode == IB_WR_RDMA_WRITE) { wr 258 drivers/infiniband/hw/hfi1/uc.c ohdr->u.imm_data = wqe->wr.ex.imm_data; wr 260 drivers/infiniband/hw/hfi1/uc.c if (wqe->wr.send_flags & IB_SEND_SOLICITED) wr 152 drivers/infiniband/hw/hfi1/ud.c if (swqe->wr.opcode == IB_WR_SEND_WITH_IMM) { wr 154 drivers/infiniband/hw/hfi1/ud.c wc.ex.imm_data = swqe->wr.ex.imm_data; wr 222 drivers/infiniband/hw/hfi1/ud.c ssge.num_sge = swqe->wr.num_sge; wr 258 drivers/infiniband/hw/hfi1/ud.c rvt_recv_cq(qp, &wc, swqe->wr.send_flags & IB_SEND_SOLICITED); wr 274 drivers/infiniband/hw/hfi1/ud.c if (wqe->wr.opcode == IB_WR_SEND_WITH_IMM) { wr 275 drivers/infiniband/hw/hfi1/ud.c ohdr->u.ud.imm_data = wqe->wr.ex.imm_data; wr 281 drivers/infiniband/hw/hfi1/ud.c if (wqe->wr.send_flags & IB_SEND_SOLICITED) wr 325 drivers/infiniband/hw/hfi1/ud.c if (wqe->wr.opcode == IB_WR_SEND_WITH_IMM) wr 404 drivers/infiniband/hw/hfi1/ud.c if (wqe->wr.opcode == IB_WR_SEND_WITH_IMM) wr 557 drivers/infiniband/hw/hfi1/ud.c qp->s_sge.num_sge = wqe->wr.num_sge; wr 969 drivers/infiniband/hw/hns/hns_roce_device.h int (*post_send)(struct ib_qp *ibqp, const struct ib_send_wr *wr, wr 990 drivers/infiniband/hw/hns/hns_roce_device.h int (*post_srq_recv)(struct ib_srq *ibsrq, const struct ib_recv_wr *wr, wr 1256 drivers/infiniband/hw/hns/hns_roce_device.h __be32 send_ieth(const struct ib_send_wr *wr); wr 62 drivers/infiniband/hw/hns/hns_roce_hw_v1.c const struct ib_send_wr *wr, wr 66 drivers/infiniband/hw/hns/hns_roce_hw_v1.c struct hns_roce_ah *ah = to_hr_ah(ud_wr(wr)->ah); wr 92 drivers/infiniband/hw/hns/hns_roce_hw_v1.c for (nreq = 0; wr; ++nreq, wr = wr->next) { wr 95 drivers/infiniband/hw/hns/hns_roce_hw_v1.c *bad_wr = wr; wr 101 drivers/infiniband/hw/hns/hns_roce_hw_v1.c if (unlikely(wr->num_sge > qp->sq.max_gs)) { wr 103 drivers/infiniband/hw/hns/hns_roce_hw_v1.c wr->num_sge, qp->sq.max_gs); wr 105 drivers/infiniband/hw/hns/hns_roce_hw_v1.c *bad_wr = wr; wr 110 drivers/infiniband/hw/hns/hns_roce_hw_v1.c qp->sq.wrid[wqe_idx] = wr->wr_id; wr 160 drivers/infiniband/hw/hns/hns_roce_hw_v1.c ud_sq_wqe->u32_8 |= (wr->send_flags & IB_SEND_SIGNALED ? wr 162 drivers/infiniband/hw/hns/hns_roce_hw_v1.c (wr->send_flags & IB_SEND_SOLICITED ? wr 164 drivers/infiniband/hw/hns/hns_roce_hw_v1.c ((wr->opcode == IB_WR_SEND_WITH_IMM) ? wr 170 drivers/infiniband/hw/hns/hns_roce_hw_v1.c ud_wr(wr)->remote_qpn); wr 202 drivers/infiniband/hw/hns/hns_roce_hw_v1.c cpu_to_le32((u32)wr->sg_list[0].addr); wr 204 drivers/infiniband/hw/hns/hns_roce_hw_v1.c cpu_to_le32((wr->sg_list[0].addr) >> 32); wr 206 drivers/infiniband/hw/hns/hns_roce_hw_v1.c cpu_to_le32(wr->sg_list[0].lkey); wr 209 drivers/infiniband/hw/hns/hns_roce_hw_v1.c cpu_to_le32((u32)wr->sg_list[1].addr); wr 211 drivers/infiniband/hw/hns/hns_roce_hw_v1.c cpu_to_le32((wr->sg_list[1].addr) >> 32); wr 213 drivers/infiniband/hw/hns/hns_roce_hw_v1.c cpu_to_le32(wr->sg_list[1].lkey); wr 219 drivers/infiniband/hw/hns/hns_roce_hw_v1.c for (i = 0; i < wr->num_sge; i++) wr 220 drivers/infiniband/hw/hns/hns_roce_hw_v1.c tmp_len += wr->sg_list[i].length; wr 228 drivers/infiniband/hw/hns/hns_roce_hw_v1.c switch (wr->opcode) { wr 231 drivers/infiniband/hw/hns/hns_roce_hw_v1.c ctrl->imm_data = wr->ex.imm_data; wr 235 drivers/infiniband/hw/hns/hns_roce_hw_v1.c cpu_to_le32(wr->ex.invalidate_rkey); wr 244 drivers/infiniband/hw/hns/hns_roce_hw_v1.c ctrl->flag |= (wr->send_flags & IB_SEND_SIGNALED ? wr 246 drivers/infiniband/hw/hns/hns_roce_hw_v1.c (wr->send_flags & IB_SEND_SOLICITED ? wr 248 drivers/infiniband/hw/hns/hns_roce_hw_v1.c ((wr->opcode == IB_WR_SEND_WITH_IMM || wr 249 drivers/infiniband/hw/hns/hns_roce_hw_v1.c wr->opcode == IB_WR_RDMA_WRITE_WITH_IMM) ? wr 251 drivers/infiniband/hw/hns/hns_roce_hw_v1.c (wr->send_flags & IB_SEND_FENCE ? wr 256 drivers/infiniband/hw/hns/hns_roce_hw_v1.c switch (wr->opcode) { wr 259 drivers/infiniband/hw/hns/hns_roce_hw_v1.c set_raddr_seg(wqe, rdma_wr(wr)->remote_addr, wr 260 drivers/infiniband/hw/hns/hns_roce_hw_v1.c rdma_wr(wr)->rkey); wr 265 drivers/infiniband/hw/hns/hns_roce_hw_v1.c set_raddr_seg(wqe, rdma_wr(wr)->remote_addr, wr 266 drivers/infiniband/hw/hns/hns_roce_hw_v1.c rdma_wr(wr)->rkey); wr 286 drivers/infiniband/hw/hns/hns_roce_hw_v1.c if (wr->send_flags & IB_SEND_INLINE && wr->num_sge) { wr 290 drivers/infiniband/hw/hns/hns_roce_hw_v1.c *bad_wr = wr; wr 296 drivers/infiniband/hw/hns/hns_roce_hw_v1.c for (i = 0; i < wr->num_sge; i++) { wr 298 drivers/infiniband/hw/hns/hns_roce_hw_v1.c wr->sg_list[i].addr), wr 299 drivers/infiniband/hw/hns/hns_roce_hw_v1.c wr->sg_list[i].length); wr 300 drivers/infiniband/hw/hns/hns_roce_hw_v1.c wqe += wr->sg_list[i].length; wr 305 drivers/infiniband/hw/hns/hns_roce_hw_v1.c for (i = 0; i < wr->num_sge; i++) wr 306 drivers/infiniband/hw/hns/hns_roce_hw_v1.c set_data_seg(dseg + i, wr->sg_list + i); wr 308 drivers/infiniband/hw/hns/hns_roce_hw_v1.c ctrl->flag |= cpu_to_le32(wr->num_sge << wr 346 drivers/infiniband/hw/hns/hns_roce_hw_v1.c const struct ib_recv_wr *wr, wr 365 drivers/infiniband/hw/hns/hns_roce_hw_v1.c for (nreq = 0; wr; ++nreq, wr = wr->next) { wr 369 drivers/infiniband/hw/hns/hns_roce_hw_v1.c *bad_wr = wr; wr 375 drivers/infiniband/hw/hns/hns_roce_hw_v1.c if (unlikely(wr->num_sge > hr_qp->rq.max_gs)) { wr 377 drivers/infiniband/hw/hns/hns_roce_hw_v1.c wr->num_sge, hr_qp->rq.max_gs); wr 379 drivers/infiniband/hw/hns/hns_roce_hw_v1.c *bad_wr = wr; wr 388 drivers/infiniband/hw/hns/hns_roce_hw_v1.c wr->num_sge); wr 392 drivers/infiniband/hw/hns/hns_roce_hw_v1.c for (i = 0; i < wr->num_sge; i++) wr 393 drivers/infiniband/hw/hns/hns_roce_hw_v1.c set_data_seg(scat + i, wr->sg_list + i); wr 395 drivers/infiniband/hw/hns/hns_roce_hw_v1.c hr_qp->rq.wrid[wqe_idx] = wr->wr_id; wr 61 drivers/infiniband/hw/hns/hns_roce_hw_v2.c const struct ib_reg_wr *wr) wr 63 drivers/infiniband/hw/hns/hns_roce_hw_v2.c struct hns_roce_mr *mr = to_hr_mr(wr->mr); wr 68 drivers/infiniband/hw/hns/hns_roce_hw_v2.c wr->access & IB_ACCESS_MW_BIND ? 1 : 0); wr 71 drivers/infiniband/hw/hns/hns_roce_hw_v2.c wr->access & IB_ACCESS_REMOTE_ATOMIC ? 1 : 0); wr 74 drivers/infiniband/hw/hns/hns_roce_hw_v2.c wr->access & IB_ACCESS_REMOTE_READ ? 1 : 0); wr 77 drivers/infiniband/hw/hns/hns_roce_hw_v2.c wr->access & IB_ACCESS_REMOTE_WRITE ? 1 : 0); wr 80 drivers/infiniband/hw/hns/hns_roce_hw_v2.c wr->access & IB_ACCESS_LOCAL_WRITE ? 1 : 0); wr 86 drivers/infiniband/hw/hns/hns_roce_hw_v2.c rc_sq_wqe->byte_16 = cpu_to_le32(wr->mr->length & 0xffffffff); wr 87 drivers/infiniband/hw/hns/hns_roce_hw_v2.c rc_sq_wqe->byte_20 = cpu_to_le32(wr->mr->length >> 32); wr 88 drivers/infiniband/hw/hns/hns_roce_hw_v2.c rc_sq_wqe->rkey = cpu_to_le32(wr->key); wr 89 drivers/infiniband/hw/hns/hns_roce_hw_v2.c rc_sq_wqe->va = cpu_to_le64(wr->mr->iova); wr 101 drivers/infiniband/hw/hns/hns_roce_hw_v2.c const struct ib_atomic_wr *wr) wr 103 drivers/infiniband/hw/hns/hns_roce_hw_v2.c if (wr->wr.opcode == IB_WR_ATOMIC_CMP_AND_SWP) { wr 104 drivers/infiniband/hw/hns/hns_roce_hw_v2.c aseg->fetchadd_swap_data = cpu_to_le64(wr->swap); wr 105 drivers/infiniband/hw/hns/hns_roce_hw_v2.c aseg->cmp_data = cpu_to_le64(wr->compare_add); wr 107 drivers/infiniband/hw/hns/hns_roce_hw_v2.c aseg->fetchadd_swap_data = cpu_to_le64(wr->compare_add); wr 112 drivers/infiniband/hw/hns/hns_roce_hw_v2.c static void set_extend_sge(struct hns_roce_qp *qp, const struct ib_send_wr *wr, wr 127 drivers/infiniband/hw/hns/hns_roce_hw_v2.c sg = wr->sg_list + num_in_wqe; wr 159 drivers/infiniband/hw/hns/hns_roce_hw_v2.c static int set_rwqe_data_seg(struct ib_qp *ibqp, const struct ib_send_wr *wr, wr 171 drivers/infiniband/hw/hns/hns_roce_hw_v2.c if (wr->send_flags & IB_SEND_INLINE && valid_num_sge) { wr 174 drivers/infiniband/hw/hns/hns_roce_hw_v2.c *bad_wr = wr; wr 180 drivers/infiniband/hw/hns/hns_roce_hw_v2.c if (wr->opcode == IB_WR_RDMA_READ) { wr 181 drivers/infiniband/hw/hns/hns_roce_hw_v2.c *bad_wr = wr; wr 186 drivers/infiniband/hw/hns/hns_roce_hw_v2.c for (i = 0; i < wr->num_sge; i++) { wr 187 drivers/infiniband/hw/hns/hns_roce_hw_v2.c memcpy(wqe, ((void *)wr->sg_list[i].addr), wr 188 drivers/infiniband/hw/hns/hns_roce_hw_v2.c wr->sg_list[i].length); wr 189 drivers/infiniband/hw/hns/hns_roce_hw_v2.c wqe += wr->sg_list[i].length; wr 196 drivers/infiniband/hw/hns/hns_roce_hw_v2.c for (i = 0; i < wr->num_sge; i++) { wr 197 drivers/infiniband/hw/hns/hns_roce_hw_v2.c if (likely(wr->sg_list[i].length)) { wr 198 drivers/infiniband/hw/hns/hns_roce_hw_v2.c set_data_seg_v2(dseg, wr->sg_list + i); wr 208 drivers/infiniband/hw/hns/hns_roce_hw_v2.c for (i = 0; i < wr->num_sge && wr 210 drivers/infiniband/hw/hns/hns_roce_hw_v2.c if (likely(wr->sg_list[i].length)) { wr 211 drivers/infiniband/hw/hns/hns_roce_hw_v2.c set_data_seg_v2(dseg, wr->sg_list + i); wr 217 drivers/infiniband/hw/hns/hns_roce_hw_v2.c set_extend_sge(qp, wr, sge_ind, valid_num_sge); wr 234 drivers/infiniband/hw/hns/hns_roce_hw_v2.c const struct ib_send_wr *wr, wr 238 drivers/infiniband/hw/hns/hns_roce_hw_v2.c struct hns_roce_ah *ah = to_hr_ah(ud_wr(wr)->ah); wr 265 drivers/infiniband/hw/hns/hns_roce_hw_v2.c *bad_wr = wr; wr 272 drivers/infiniband/hw/hns/hns_roce_hw_v2.c *bad_wr = wr; wr 279 drivers/infiniband/hw/hns/hns_roce_hw_v2.c for (nreq = 0; wr; ++nreq, wr = wr->next) { wr 282 drivers/infiniband/hw/hns/hns_roce_hw_v2.c *bad_wr = wr; wr 288 drivers/infiniband/hw/hns/hns_roce_hw_v2.c if (unlikely(wr->num_sge > qp->sq.max_gs)) { wr 290 drivers/infiniband/hw/hns/hns_roce_hw_v2.c wr->num_sge, qp->sq.max_gs); wr 292 drivers/infiniband/hw/hns/hns_roce_hw_v2.c *bad_wr = wr; wr 297 drivers/infiniband/hw/hns/hns_roce_hw_v2.c qp->sq.wrid[wqe_idx] = wr->wr_id; wr 303 drivers/infiniband/hw/hns/hns_roce_hw_v2.c for (i = 0; i < wr->num_sge; i++) { wr 304 drivers/infiniband/hw/hns/hns_roce_hw_v2.c if (likely(wr->sg_list[i].length)) { wr 305 drivers/infiniband/hw/hns/hns_roce_hw_v2.c tmp_len += wr->sg_list[i].length; wr 348 drivers/infiniband/hw/hns/hns_roce_hw_v2.c switch (wr->opcode) { wr 352 drivers/infiniband/hw/hns/hns_roce_hw_v2.c cpu_to_le32(be32_to_cpu(wr->ex.imm_data)); wr 362 drivers/infiniband/hw/hns/hns_roce_hw_v2.c (wr->send_flags & IB_SEND_SIGNALED) ? 1 : 0); wr 367 drivers/infiniband/hw/hns/hns_roce_hw_v2.c (wr->send_flags & IB_SEND_SOLICITED) ? 1 : 0); wr 391 drivers/infiniband/hw/hns/hns_roce_hw_v2.c cpu_to_le32(ud_wr(wr)->remote_qkey & 0x80000000 ? wr 392 drivers/infiniband/hw/hns/hns_roce_hw_v2.c qp->qkey : ud_wr(wr)->remote_qkey); wr 396 drivers/infiniband/hw/hns/hns_roce_hw_v2.c ud_wr(wr)->remote_qpn); wr 435 drivers/infiniband/hw/hns/hns_roce_hw_v2.c set_extend_sge(qp, wr, &sge_idx, valid_num_sge); wr 443 drivers/infiniband/hw/hns/hns_roce_hw_v2.c switch (wr->opcode) { wr 447 drivers/infiniband/hw/hns/hns_roce_hw_v2.c cpu_to_le32(be32_to_cpu(wr->ex.imm_data)); wr 451 drivers/infiniband/hw/hns/hns_roce_hw_v2.c cpu_to_le32(wr->ex.invalidate_rkey); wr 460 drivers/infiniband/hw/hns/hns_roce_hw_v2.c (wr->send_flags & IB_SEND_FENCE) ? 1 : 0); wr 464 drivers/infiniband/hw/hns/hns_roce_hw_v2.c (wr->send_flags & IB_SEND_SOLICITED) ? 1 : 0); wr 468 drivers/infiniband/hw/hns/hns_roce_hw_v2.c (wr->send_flags & IB_SEND_SIGNALED) ? 1 : 0); wr 474 drivers/infiniband/hw/hns/hns_roce_hw_v2.c switch (wr->opcode) { wr 478 drivers/infiniband/hw/hns/hns_roce_hw_v2.c cpu_to_le32(rdma_wr(wr)->rkey); wr 480 drivers/infiniband/hw/hns/hns_roce_hw_v2.c cpu_to_le64(rdma_wr(wr)->remote_addr); wr 485 drivers/infiniband/hw/hns/hns_roce_hw_v2.c cpu_to_le32(rdma_wr(wr)->rkey); wr 487 drivers/infiniband/hw/hns/hns_roce_hw_v2.c cpu_to_le64(rdma_wr(wr)->remote_addr); wr 492 drivers/infiniband/hw/hns/hns_roce_hw_v2.c cpu_to_le32(rdma_wr(wr)->rkey); wr 494 drivers/infiniband/hw/hns/hns_roce_hw_v2.c cpu_to_le64(rdma_wr(wr)->remote_addr); wr 510 drivers/infiniband/hw/hns/hns_roce_hw_v2.c cpu_to_le32(wr->ex.invalidate_rkey); wr 515 drivers/infiniband/hw/hns/hns_roce_hw_v2.c set_frmr_seg(rc_sq_wqe, fseg, reg_wr(wr)); wr 520 drivers/infiniband/hw/hns/hns_roce_hw_v2.c cpu_to_le32(atomic_wr(wr)->rkey); wr 522 drivers/infiniband/hw/hns/hns_roce_hw_v2.c cpu_to_le64(atomic_wr(wr)->remote_addr); wr 527 drivers/infiniband/hw/hns/hns_roce_hw_v2.c cpu_to_le32(atomic_wr(wr)->rkey); wr 529 drivers/infiniband/hw/hns/hns_roce_hw_v2.c cpu_to_le64(atomic_wr(wr)->remote_addr); wr 548 drivers/infiniband/hw/hns/hns_roce_hw_v2.c if (wr->opcode == IB_WR_ATOMIC_CMP_AND_SWP || wr 549 drivers/infiniband/hw/hns/hns_roce_hw_v2.c wr->opcode == IB_WR_ATOMIC_FETCH_AND_ADD) { wr 553 drivers/infiniband/hw/hns/hns_roce_hw_v2.c set_data_seg_v2(dseg, wr->sg_list); wr 555 drivers/infiniband/hw/hns/hns_roce_hw_v2.c set_atomic_seg(wqe, atomic_wr(wr)); wr 560 drivers/infiniband/hw/hns/hns_roce_hw_v2.c } else if (wr->opcode != IB_WR_REG_MR) { wr 561 drivers/infiniband/hw/hns/hns_roce_hw_v2.c ret = set_rwqe_data_seg(ibqp, wr, rc_sq_wqe, wr 570 drivers/infiniband/hw/hns/hns_roce_hw_v2.c *bad_wr = wr; wr 606 drivers/infiniband/hw/hns/hns_roce_hw_v2.c *bad_wr = wr; wr 618 drivers/infiniband/hw/hns/hns_roce_hw_v2.c const struct ib_recv_wr *wr, wr 639 drivers/infiniband/hw/hns/hns_roce_hw_v2.c *bad_wr = wr; wr 643 drivers/infiniband/hw/hns/hns_roce_hw_v2.c for (nreq = 0; wr; ++nreq, wr = wr->next) { wr 647 drivers/infiniband/hw/hns/hns_roce_hw_v2.c *bad_wr = wr; wr 653 drivers/infiniband/hw/hns/hns_roce_hw_v2.c if (unlikely(wr->num_sge > hr_qp->rq.max_gs)) { wr 655 drivers/infiniband/hw/hns/hns_roce_hw_v2.c wr->num_sge, hr_qp->rq.max_gs); wr 657 drivers/infiniband/hw/hns/hns_roce_hw_v2.c *bad_wr = wr; wr 663 drivers/infiniband/hw/hns/hns_roce_hw_v2.c for (i = 0; i < wr->num_sge; i++) { wr 664 drivers/infiniband/hw/hns/hns_roce_hw_v2.c if (!wr->sg_list[i].length) wr 666 drivers/infiniband/hw/hns/hns_roce_hw_v2.c set_data_seg_v2(dseg, wr->sg_list + i); wr 679 drivers/infiniband/hw/hns/hns_roce_hw_v2.c (u32)wr->num_sge; wr 680 drivers/infiniband/hw/hns/hns_roce_hw_v2.c for (i = 0; i < wr->num_sge; i++) { wr 682 drivers/infiniband/hw/hns/hns_roce_hw_v2.c (void *)(u64)wr->sg_list[i].addr; wr 683 drivers/infiniband/hw/hns/hns_roce_hw_v2.c sge_list[i].len = wr->sg_list[i].length; wr 687 drivers/infiniband/hw/hns/hns_roce_hw_v2.c hr_qp->rq.wrid[wqe_idx] = wr->wr_id; wr 707 drivers/infiniband/hw/hns/hns_roce_hw_v2.c *bad_wr = wr; wr 6230 drivers/infiniband/hw/hns/hns_roce_hw_v2.c const struct ib_recv_wr *wr, wr 6249 drivers/infiniband/hw/hns/hns_roce_hw_v2.c for (nreq = 0; wr; ++nreq, wr = wr->next) { wr 6250 drivers/infiniband/hw/hns/hns_roce_hw_v2.c if (unlikely(wr->num_sge > srq->max_gs)) { wr 6252 drivers/infiniband/hw/hns/hns_roce_hw_v2.c *bad_wr = wr; wr 6258 drivers/infiniband/hw/hns/hns_roce_hw_v2.c *bad_wr = wr; wr 6265 drivers/infiniband/hw/hns/hns_roce_hw_v2.c *bad_wr = wr; wr 6273 drivers/infiniband/hw/hns/hns_roce_hw_v2.c for (i = 0; i < wr->num_sge; ++i) { wr 6274 drivers/infiniband/hw/hns/hns_roce_hw_v2.c dseg[i].len = cpu_to_le32(wr->sg_list[i].length); wr 6275 drivers/infiniband/hw/hns/hns_roce_hw_v2.c dseg[i].lkey = cpu_to_le32(wr->sg_list[i].lkey); wr 6276 drivers/infiniband/hw/hns/hns_roce_hw_v2.c dseg[i].addr = cpu_to_le64(wr->sg_list[i].addr); wr 6285 drivers/infiniband/hw/hns/hns_roce_hw_v2.c srq->wrid[wqe_idx] = wr->wr_id; wr 508 drivers/infiniband/hw/mlx4/mad.c struct ib_ud_wr wr; wr 640 drivers/infiniband/hw/mlx4/mad.c wr.ah = ah; wr 641 drivers/infiniband/hw/mlx4/mad.c wr.port_num = port; wr 642 drivers/infiniband/hw/mlx4/mad.c wr.remote_qkey = IB_QP_SET_QKEY; wr 643 drivers/infiniband/hw/mlx4/mad.c wr.remote_qpn = dqpn; wr 644 drivers/infiniband/hw/mlx4/mad.c wr.wr.next = NULL; wr 645 drivers/infiniband/hw/mlx4/mad.c wr.wr.wr_id = ((u64) tun_tx_ix) | MLX4_TUN_SET_WRID_QPN(dest_qpt); wr 646 drivers/infiniband/hw/mlx4/mad.c wr.wr.sg_list = &list; wr 647 drivers/infiniband/hw/mlx4/mad.c wr.wr.num_sge = 1; wr 648 drivers/infiniband/hw/mlx4/mad.c wr.wr.opcode = IB_WR_SEND; wr 649 drivers/infiniband/hw/mlx4/mad.c wr.wr.send_flags = IB_SEND_SIGNALED; wr 651 drivers/infiniband/hw/mlx4/mad.c ret = ib_post_send(src_qp, &wr.wr, &bad_wr); wr 1366 drivers/infiniband/hw/mlx4/mad.c struct ib_ud_wr wr; wr 1439 drivers/infiniband/hw/mlx4/mad.c wr.ah = ah; wr 1440 drivers/infiniband/hw/mlx4/mad.c wr.port_num = port; wr 1441 drivers/infiniband/hw/mlx4/mad.c wr.pkey_index = wire_pkey_ix; wr 1442 drivers/infiniband/hw/mlx4/mad.c wr.remote_qkey = qkey; wr 1443 drivers/infiniband/hw/mlx4/mad.c wr.remote_qpn = remote_qpn; wr 1444 drivers/infiniband/hw/mlx4/mad.c wr.wr.next = NULL; wr 1445 drivers/infiniband/hw/mlx4/mad.c wr.wr.wr_id = ((u64) wire_tx_ix) | MLX4_TUN_SET_WRID_QPN(src_qpnum); wr 1446 drivers/infiniband/hw/mlx4/mad.c wr.wr.sg_list = &list; wr 1447 drivers/infiniband/hw/mlx4/mad.c wr.wr.num_sge = 1; wr 1448 drivers/infiniband/hw/mlx4/mad.c wr.wr.opcode = IB_WR_SEND; wr 1449 drivers/infiniband/hw/mlx4/mad.c wr.wr.send_flags = IB_SEND_SIGNALED; wr 1451 drivers/infiniband/hw/mlx4/mad.c ret = ib_post_send(send_qp, &wr.wr, &bad_wr); wr 768 drivers/infiniband/hw/mlx4/mlx4_ib.h int mlx4_ib_post_srq_recv(struct ib_srq *ibsrq, const struct ib_recv_wr *wr, wr 781 drivers/infiniband/hw/mlx4/mlx4_ib.h int mlx4_ib_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr, wr 783 drivers/infiniband/hw/mlx4/mlx4_ib.h int mlx4_ib_post_recv(struct ib_qp *ibqp, const struct ib_recv_wr *wr, wr 2881 drivers/infiniband/hw/mlx4/qp.c const struct ib_ud_wr *wr, wr 2888 drivers/infiniband/hw/mlx4/qp.c struct mlx4_ib_ah *ah = to_mah(wr->ah); wr 2897 drivers/infiniband/hw/mlx4/qp.c if (wr->wr.opcode != IB_WR_SEND) wr 2902 drivers/infiniband/hw/mlx4/qp.c for (i = 0; i < wr->wr.num_sge; ++i) wr 2903 drivers/infiniband/hw/mlx4/qp.c send_size += wr->wr.sg_list[i].length; wr 2928 drivers/infiniband/hw/mlx4/qp.c sqp->ud_header.bth.solicited_event = !!(wr->wr.send_flags & IB_SEND_SOLICITED); wr 2934 drivers/infiniband/hw/mlx4/qp.c sqp->ud_header.bth.destination_qpn = cpu_to_be32(wr->remote_qpn); wr 3032 drivers/infiniband/hw/mlx4/qp.c static int build_mlx_header(struct mlx4_ib_sqp *sqp, const struct ib_ud_wr *wr, wr 3040 drivers/infiniband/hw/mlx4/qp.c struct mlx4_ib_ah *ah = to_mah(wr->ah); wr 3056 drivers/infiniband/hw/mlx4/qp.c for (i = 0; i < wr->wr.num_sge; ++i) wr 3057 drivers/infiniband/hw/mlx4/qp.c send_size += wr->wr.sg_list[i].length; wr 3167 drivers/infiniband/hw/mlx4/qp.c switch (wr->wr.opcode) { wr 3175 drivers/infiniband/hw/mlx4/qp.c sqp->ud_header.immediate_data = wr->wr.ex.imm_data; wr 3216 drivers/infiniband/hw/mlx4/qp.c sqp->ud_header.bth.solicited_event = !!(wr->wr.send_flags & IB_SEND_SOLICITED); wr 3221 drivers/infiniband/hw/mlx4/qp.c err = ib_get_cached_pkey(ib_dev, sqp->qp.port, wr->pkey_index, wr 3227 drivers/infiniband/hw/mlx4/qp.c sqp->ud_header.bth.destination_qpn = cpu_to_be32(wr->remote_qpn); wr 3229 drivers/infiniband/hw/mlx4/qp.c sqp->ud_header.deth.qkey = cpu_to_be32(wr->remote_qkey & 0x80000000 ? wr 3230 drivers/infiniband/hw/mlx4/qp.c sqp->qkey : wr->remote_qkey); wr 3319 drivers/infiniband/hw/mlx4/qp.c const struct ib_reg_wr *wr) wr 3321 drivers/infiniband/hw/mlx4/qp.c struct mlx4_ib_mr *mr = to_mmr(wr->mr); wr 3323 drivers/infiniband/hw/mlx4/qp.c fseg->flags = convert_access(wr->access); wr 3324 drivers/infiniband/hw/mlx4/qp.c fseg->mem_key = cpu_to_be32(wr->key); wr 3349 drivers/infiniband/hw/mlx4/qp.c const struct ib_atomic_wr *wr) wr 3351 drivers/infiniband/hw/mlx4/qp.c if (wr->wr.opcode == IB_WR_ATOMIC_CMP_AND_SWP) { wr 3352 drivers/infiniband/hw/mlx4/qp.c aseg->swap_add = cpu_to_be64(wr->swap); wr 3353 drivers/infiniband/hw/mlx4/qp.c aseg->compare = cpu_to_be64(wr->compare_add); wr 3354 drivers/infiniband/hw/mlx4/qp.c } else if (wr->wr.opcode == IB_WR_MASKED_ATOMIC_FETCH_AND_ADD) { wr 3355 drivers/infiniband/hw/mlx4/qp.c aseg->swap_add = cpu_to_be64(wr->compare_add); wr 3356 drivers/infiniband/hw/mlx4/qp.c aseg->compare = cpu_to_be64(wr->compare_add_mask); wr 3358 drivers/infiniband/hw/mlx4/qp.c aseg->swap_add = cpu_to_be64(wr->compare_add); wr 3365 drivers/infiniband/hw/mlx4/qp.c const struct ib_atomic_wr *wr) wr 3367 drivers/infiniband/hw/mlx4/qp.c aseg->swap_add = cpu_to_be64(wr->swap); wr 3368 drivers/infiniband/hw/mlx4/qp.c aseg->swap_add_mask = cpu_to_be64(wr->swap_mask); wr 3369 drivers/infiniband/hw/mlx4/qp.c aseg->compare = cpu_to_be64(wr->compare_add); wr 3370 drivers/infiniband/hw/mlx4/qp.c aseg->compare_mask = cpu_to_be64(wr->compare_add_mask); wr 3374 drivers/infiniband/hw/mlx4/qp.c const struct ib_ud_wr *wr) wr 3376 drivers/infiniband/hw/mlx4/qp.c memcpy(dseg->av, &to_mah(wr->ah)->av, sizeof (struct mlx4_av)); wr 3377 drivers/infiniband/hw/mlx4/qp.c dseg->dqpn = cpu_to_be32(wr->remote_qpn); wr 3378 drivers/infiniband/hw/mlx4/qp.c dseg->qkey = cpu_to_be32(wr->remote_qkey); wr 3379 drivers/infiniband/hw/mlx4/qp.c dseg->vlan = to_mah(wr->ah)->av.eth.vlan; wr 3380 drivers/infiniband/hw/mlx4/qp.c memcpy(dseg->mac, to_mah(wr->ah)->av.eth.mac, 6); wr 3385 drivers/infiniband/hw/mlx4/qp.c const struct ib_ud_wr *wr, wr 3388 drivers/infiniband/hw/mlx4/qp.c union mlx4_ext_av *av = &to_mah(wr->ah)->av; wr 3407 drivers/infiniband/hw/mlx4/qp.c static void build_tunnel_header(const struct ib_ud_wr *wr, void *wqe, wr 3412 drivers/infiniband/hw/mlx4/qp.c struct mlx4_ib_ah *ah = to_mah(wr->ah); wr 3417 drivers/infiniband/hw/mlx4/qp.c hdr.remote_qpn = cpu_to_be32(wr->remote_qpn); wr 3418 drivers/infiniband/hw/mlx4/qp.c hdr.pkey_index = cpu_to_be16(wr->pkey_index); wr 3419 drivers/infiniband/hw/mlx4/qp.c hdr.qkey = cpu_to_be32(wr->remote_qkey); wr 3492 drivers/infiniband/hw/mlx4/qp.c const struct ib_ud_wr *wr, struct mlx4_ib_qp *qp, wr 3495 drivers/infiniband/hw/mlx4/qp.c unsigned halign = ALIGN(sizeof *wqe + wr->hlen, 16); wr 3501 drivers/infiniband/hw/mlx4/qp.c wr->wr.num_sge > qp->sq.max_gs - (halign >> 4))) wr 3504 drivers/infiniband/hw/mlx4/qp.c memcpy(wqe->header, wr->header, wr->hlen); wr 3506 drivers/infiniband/hw/mlx4/qp.c *lso_hdr_sz = cpu_to_be32(wr->mss << 16 | wr->hlen); wr 3511 drivers/infiniband/hw/mlx4/qp.c static __be32 send_ieth(const struct ib_send_wr *wr) wr 3513 drivers/infiniband/hw/mlx4/qp.c switch (wr->opcode) { wr 3516 drivers/infiniband/hw/mlx4/qp.c return wr->ex.imm_data; wr 3519 drivers/infiniband/hw/mlx4/qp.c return cpu_to_be32(wr->ex.invalidate_rkey); wr 3533 drivers/infiniband/hw/mlx4/qp.c static int _mlx4_ib_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr, wr 3557 drivers/infiniband/hw/mlx4/qp.c struct mlx4_ib_ah *ah = to_mah(ud_wr(wr)->ah); wr 3576 drivers/infiniband/hw/mlx4/qp.c *bad_wr = wr; wr 3583 drivers/infiniband/hw/mlx4/qp.c for (nreq = 0; wr; ++nreq, wr = wr->next) { wr 3589 drivers/infiniband/hw/mlx4/qp.c *bad_wr = wr; wr 3593 drivers/infiniband/hw/mlx4/qp.c if (unlikely(wr->num_sge > qp->sq.max_gs)) { wr 3595 drivers/infiniband/hw/mlx4/qp.c *bad_wr = wr; wr 3600 drivers/infiniband/hw/mlx4/qp.c qp->sq.wrid[(qp->sq.head + nreq) & (qp->sq.wqe_cnt - 1)] = wr->wr_id; wr 3603 drivers/infiniband/hw/mlx4/qp.c (wr->send_flags & IB_SEND_SIGNALED ? wr 3605 drivers/infiniband/hw/mlx4/qp.c (wr->send_flags & IB_SEND_SOLICITED ? wr 3607 drivers/infiniband/hw/mlx4/qp.c ((wr->send_flags & IB_SEND_IP_CSUM) ? wr 3612 drivers/infiniband/hw/mlx4/qp.c ctrl->imm = send_ieth(wr); wr 3620 drivers/infiniband/hw/mlx4/qp.c switch (wr->opcode) { wr 3624 drivers/infiniband/hw/mlx4/qp.c set_raddr_seg(wqe, atomic_wr(wr)->remote_addr, wr 3625 drivers/infiniband/hw/mlx4/qp.c atomic_wr(wr)->rkey); wr 3628 drivers/infiniband/hw/mlx4/qp.c set_atomic_seg(wqe, atomic_wr(wr)); wr 3637 drivers/infiniband/hw/mlx4/qp.c set_raddr_seg(wqe, atomic_wr(wr)->remote_addr, wr 3638 drivers/infiniband/hw/mlx4/qp.c atomic_wr(wr)->rkey); wr 3641 drivers/infiniband/hw/mlx4/qp.c set_masked_atomic_seg(wqe, atomic_wr(wr)); wr 3652 drivers/infiniband/hw/mlx4/qp.c set_raddr_seg(wqe, rdma_wr(wr)->remote_addr, wr 3653 drivers/infiniband/hw/mlx4/qp.c rdma_wr(wr)->rkey); wr 3661 drivers/infiniband/hw/mlx4/qp.c set_local_inv_seg(wqe, wr->ex.invalidate_rkey); wr 3669 drivers/infiniband/hw/mlx4/qp.c set_reg_seg(wqe, reg_wr(wr)); wr 3681 drivers/infiniband/hw/mlx4/qp.c err = build_sriov_qp0_header(to_msqp(qp), ud_wr(wr), wr 3684 drivers/infiniband/hw/mlx4/qp.c *bad_wr = wr; wr 3693 drivers/infiniband/hw/mlx4/qp.c set_datagram_seg(wqe, ud_wr(wr)); wr 3700 drivers/infiniband/hw/mlx4/qp.c set_datagram_seg(wqe, ud_wr(wr)); wr 3704 drivers/infiniband/hw/mlx4/qp.c if (wr->opcode == IB_WR_LSO) { wr 3705 drivers/infiniband/hw/mlx4/qp.c err = build_lso_seg(wqe, ud_wr(wr), qp, &seglen, wr 3708 drivers/infiniband/hw/mlx4/qp.c *bad_wr = wr; wr 3718 drivers/infiniband/hw/mlx4/qp.c err = build_sriov_qp0_header(to_msqp(qp), ud_wr(wr), wr 3721 drivers/infiniband/hw/mlx4/qp.c *bad_wr = wr; wr 3730 drivers/infiniband/hw/mlx4/qp.c build_tunnel_header(ud_wr(wr), wqe, &seglen); wr 3741 drivers/infiniband/hw/mlx4/qp.c ud_wr(wr), wr 3745 drivers/infiniband/hw/mlx4/qp.c build_tunnel_header(ud_wr(wr), wqe, &seglen); wr 3752 drivers/infiniband/hw/mlx4/qp.c err = build_mlx_header(to_msqp(qp), ud_wr(wr), ctrl, wr 3755 drivers/infiniband/hw/mlx4/qp.c *bad_wr = wr; wr 3774 drivers/infiniband/hw/mlx4/qp.c dseg += wr->num_sge - 1; wr 3775 drivers/infiniband/hw/mlx4/qp.c size += wr->num_sge * (sizeof (struct mlx4_wqe_data_seg) / 16); wr 3786 drivers/infiniband/hw/mlx4/qp.c for (i = wr->num_sge - 1; i >= 0; --i, --dseg) wr 3787 drivers/infiniband/hw/mlx4/qp.c set_data_seg(dseg, wr->sg_list + i); wr 3797 drivers/infiniband/hw/mlx4/qp.c ctrl->qpn_vlan.fence_size = (wr->send_flags & IB_SEND_FENCE ? wr 3807 drivers/infiniband/hw/mlx4/qp.c if (wr->opcode < 0 || wr->opcode >= ARRAY_SIZE(mlx4_ib_opcode)) { wr 3808 drivers/infiniband/hw/mlx4/qp.c *bad_wr = wr; wr 3813 drivers/infiniband/hw/mlx4/qp.c ctrl->owner_opcode = mlx4_ib_opcode[wr->opcode] | wr 3821 drivers/infiniband/hw/mlx4/qp.c if (wr->next) wr 3849 drivers/infiniband/hw/mlx4/qp.c int mlx4_ib_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr, wr 3852 drivers/infiniband/hw/mlx4/qp.c return _mlx4_ib_post_send(ibqp, wr, bad_wr, false); wr 3855 drivers/infiniband/hw/mlx4/qp.c static int _mlx4_ib_post_recv(struct ib_qp *ibqp, const struct ib_recv_wr *wr, wr 3874 drivers/infiniband/hw/mlx4/qp.c *bad_wr = wr; wr 3881 drivers/infiniband/hw/mlx4/qp.c for (nreq = 0; wr; ++nreq, wr = wr->next) { wr 3884 drivers/infiniband/hw/mlx4/qp.c *bad_wr = wr; wr 3888 drivers/infiniband/hw/mlx4/qp.c if (unlikely(wr->num_sge > qp->rq.max_gs)) { wr 3890 drivers/infiniband/hw/mlx4/qp.c *bad_wr = wr; wr 3905 drivers/infiniband/hw/mlx4/qp.c scat->lkey = cpu_to_be32(wr->sg_list->lkey); wr 3911 drivers/infiniband/hw/mlx4/qp.c for (i = 0; i < wr->num_sge; ++i) wr 3912 drivers/infiniband/hw/mlx4/qp.c __set_data_seg(scat + i, wr->sg_list + i); wr 3920 drivers/infiniband/hw/mlx4/qp.c qp->rq.wrid[ind] = wr->wr_id; wr 3943 drivers/infiniband/hw/mlx4/qp.c int mlx4_ib_post_recv(struct ib_qp *ibqp, const struct ib_recv_wr *wr, wr 3946 drivers/infiniband/hw/mlx4/qp.c return _mlx4_ib_post_recv(ibqp, wr, bad_wr, false); wr 4484 drivers/infiniband/hw/mlx4/qp.c .wr = { wr 4503 drivers/infiniband/hw/mlx4/qp.c ret = _mlx4_ib_post_send(qp, &swr.wr, &bad_swr, true); wr 300 drivers/infiniband/hw/mlx4/srq.c int mlx4_ib_post_srq_recv(struct ib_srq *ibsrq, const struct ib_recv_wr *wr, wr 315 drivers/infiniband/hw/mlx4/srq.c *bad_wr = wr; wr 320 drivers/infiniband/hw/mlx4/srq.c for (nreq = 0; wr; ++nreq, wr = wr->next) { wr 321 drivers/infiniband/hw/mlx4/srq.c if (unlikely(wr->num_sge > srq->msrq.max_gs)) { wr 323 drivers/infiniband/hw/mlx4/srq.c *bad_wr = wr; wr 329 drivers/infiniband/hw/mlx4/srq.c *bad_wr = wr; wr 333 drivers/infiniband/hw/mlx4/srq.c srq->wrid[srq->head] = wr->wr_id; wr 339 drivers/infiniband/hw/mlx4/srq.c for (i = 0; i < wr->num_sge; ++i) { wr 340 drivers/infiniband/hw/mlx4/srq.c scat[i].byte_count = cpu_to_be32(wr->sg_list[i].length); wr 341 drivers/infiniband/hw/mlx4/srq.c scat[i].lkey = cpu_to_be32(wr->sg_list[i].lkey); wr 342 drivers/infiniband/hw/mlx4/srq.c scat[i].addr = cpu_to_be64(wr->sg_list[i].addr); wr 76 drivers/infiniband/hw/mlx5/gsi.c struct mlx5_ib_gsi_wr *wr; wr 81 drivers/infiniband/hw/mlx5/gsi.c wr = &gsi->outstanding_wrs[index % gsi->cap.max_send_wr]; wr 83 drivers/infiniband/hw/mlx5/gsi.c if (!wr->completed) wr 87 drivers/infiniband/hw/mlx5/gsi.c wr->send_flags & IB_SEND_SIGNALED) wr 88 drivers/infiniband/hw/mlx5/gsi.c WARN_ON_ONCE(mlx5_ib_generate_wc(gsi_cq, &wr->wc)); wr 90 drivers/infiniband/hw/mlx5/gsi.c wr->completed = false; wr 99 drivers/infiniband/hw/mlx5/gsi.c struct mlx5_ib_gsi_wr *wr = wr 105 drivers/infiniband/hw/mlx5/gsi.c wr->completed = true; wr 106 drivers/infiniband/hw/mlx5/gsi.c wr_id = wr->wc.wr_id; wr 107 drivers/infiniband/hw/mlx5/gsi.c wr->wc = *wc; wr 108 drivers/infiniband/hw/mlx5/gsi.c wr->wc.wr_id = wr_id; wr 109 drivers/infiniband/hw/mlx5/gsi.c wr->wc.qp = &gsi->ibqp; wr 415 drivers/infiniband/hw/mlx5/gsi.c struct ib_ud_wr *wr, struct ib_wc *wc) wr 431 drivers/infiniband/hw/mlx5/gsi.c gsi_wr->wc.pkey_index = wr->pkey_index; wr 432 drivers/infiniband/hw/mlx5/gsi.c gsi_wr->wc.wr_id = wr->wr.wr_id; wr 439 drivers/infiniband/hw/mlx5/gsi.c wr->wr.wr_cqe = &gsi_wr->cqe; wr 446 drivers/infiniband/hw/mlx5/gsi.c struct ib_ud_wr *wr) wr 449 drivers/infiniband/hw/mlx5/gsi.c { .wr_id = wr->wr.wr_id }, wr 456 drivers/infiniband/hw/mlx5/gsi.c ret = mlx5_ib_add_outstanding_wr(gsi, wr, &wc); wr 466 drivers/infiniband/hw/mlx5/gsi.c static struct ib_qp *get_tx_qp(struct mlx5_ib_gsi_qp *gsi, struct ib_ud_wr *wr) wr 469 drivers/infiniband/hw/mlx5/gsi.c int qp_index = wr->pkey_index; wr 480 drivers/infiniband/hw/mlx5/gsi.c int mlx5_ib_gsi_post_send(struct ib_qp *qp, const struct ib_send_wr *wr, wr 488 drivers/infiniband/hw/mlx5/gsi.c for (; wr; wr = wr->next) { wr 489 drivers/infiniband/hw/mlx5/gsi.c struct ib_ud_wr cur_wr = *ud_wr(wr); wr 491 drivers/infiniband/hw/mlx5/gsi.c cur_wr.wr.next = NULL; wr 507 drivers/infiniband/hw/mlx5/gsi.c ret = ib_post_send(tx_qp, &cur_wr.wr, bad_wr); wr 520 drivers/infiniband/hw/mlx5/gsi.c *bad_wr = wr; wr 524 drivers/infiniband/hw/mlx5/gsi.c int mlx5_ib_gsi_post_recv(struct ib_qp *qp, const struct ib_recv_wr *wr, wr 529 drivers/infiniband/hw/mlx5/gsi.c return ib_post_recv(gsi->rx_qp, wr, bad_wr); wr 477 drivers/infiniband/hw/mlx5/mlx5_ib.h struct ib_send_wr wr; wr 489 drivers/infiniband/hw/mlx5/mlx5_ib.h static inline const struct mlx5_umr_wr *umr_wr(const struct ib_send_wr *wr) wr 491 drivers/infiniband/hw/mlx5/mlx5_ib.h return container_of(wr, struct mlx5_umr_wr, wr); wr 1116 drivers/infiniband/hw/mlx5/mlx5_ib.h int mlx5_ib_post_srq_recv(struct ib_srq *ibsrq, const struct ib_recv_wr *wr, wr 1130 drivers/infiniband/hw/mlx5/mlx5_ib.h int mlx5_ib_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr, wr 1132 drivers/infiniband/hw/mlx5/mlx5_ib.h int mlx5_ib_post_recv(struct ib_qp *ibqp, const struct ib_recv_wr *wr, wr 1322 drivers/infiniband/hw/mlx5/mlx5_ib.h int mlx5_ib_gsi_post_send(struct ib_qp *qp, const struct ib_send_wr *wr, wr 1324 drivers/infiniband/hw/mlx5/mlx5_ib.h int mlx5_ib_gsi_post_recv(struct ib_qp *qp, const struct ib_recv_wr *wr, wr 831 drivers/infiniband/hw/mlx5/mr.c umrwr->wr.wr_cqe = &umr_context.cqe; wr 834 drivers/infiniband/hw/mlx5/mr.c err = ib_post_send(umrc->qp, &umrwr->wr, &bad); wr 927 drivers/infiniband/hw/mlx5/mr.c struct mlx5_umr_wr wr; wr 987 drivers/infiniband/hw/mlx5/mr.c memset(&wr, 0, sizeof(wr)); wr 988 drivers/infiniband/hw/mlx5/mr.c wr.wr.send_flags = MLX5_IB_SEND_UMR_UPDATE_XLT; wr 990 drivers/infiniband/hw/mlx5/mr.c wr.wr.send_flags |= MLX5_IB_SEND_UMR_FAIL_IF_FREE; wr 991 drivers/infiniband/hw/mlx5/mr.c wr.wr.sg_list = &sg; wr 992 drivers/infiniband/hw/mlx5/mr.c wr.wr.num_sge = 1; wr 993 drivers/infiniband/hw/mlx5/mr.c wr.wr.opcode = MLX5_IB_WR_UMR; wr 995 drivers/infiniband/hw/mlx5/mr.c wr.pd = mr->ibmr.pd; wr 996 drivers/infiniband/hw/mlx5/mr.c wr.mkey = mr->mmkey.key; wr 997 drivers/infiniband/hw/mlx5/mr.c wr.length = mr->mmkey.size; wr 998 drivers/infiniband/hw/mlx5/mr.c wr.virt_addr = mr->mmkey.iova; wr 999 drivers/infiniband/hw/mlx5/mr.c wr.access_flags = mr->access_flags; wr 1000 drivers/infiniband/hw/mlx5/mr.c wr.page_shift = page_shift; wr 1017 drivers/infiniband/hw/mlx5/mr.c wr.wr.send_flags |= wr 1023 drivers/infiniband/hw/mlx5/mr.c wr.wr.send_flags |= wr 1026 drivers/infiniband/hw/mlx5/mr.c wr.wr.send_flags |= wr 1030 drivers/infiniband/hw/mlx5/mr.c wr.offset = idx * desc_size; wr 1031 drivers/infiniband/hw/mlx5/mr.c wr.xlt_size = sg.length; wr 1033 drivers/infiniband/hw/mlx5/mr.c err = mlx5_ib_post_send_wait(dev, &wr); wr 1359 drivers/infiniband/hw/mlx5/mr.c umrwr.wr.send_flags = MLX5_IB_SEND_UMR_DISABLE_MR | wr 1361 drivers/infiniband/hw/mlx5/mr.c umrwr.wr.opcode = MLX5_IB_WR_UMR; wr 1376 drivers/infiniband/hw/mlx5/mr.c umrwr.wr.send_flags = MLX5_IB_SEND_UMR_FAIL_IF_FREE; wr 1378 drivers/infiniband/hw/mlx5/mr.c umrwr.wr.opcode = MLX5_IB_WR_UMR; wr 1384 drivers/infiniband/hw/mlx5/mr.c umrwr.wr.send_flags |= MLX5_IB_SEND_UMR_UPDATE_PD_ACCESS; wr 4095 drivers/infiniband/hw/mlx5/qp.c static void set_eth_seg(const struct ib_send_wr *wr, struct mlx5_ib_qp *qp, wr 4102 drivers/infiniband/hw/mlx5/qp.c if (wr->send_flags & IB_SEND_IP_CSUM) wr 4106 drivers/infiniband/hw/mlx5/qp.c if (wr->opcode == IB_WR_LSO) { wr 4107 drivers/infiniband/hw/mlx5/qp.c struct ib_ud_wr *ud_wr = container_of(wr, struct ib_ud_wr, wr); wr 4144 drivers/infiniband/hw/mlx5/qp.c const struct ib_send_wr *wr) wr 4146 drivers/infiniband/hw/mlx5/qp.c memcpy(&dseg->av, &to_mah(ud_wr(wr)->ah)->av, sizeof(struct mlx5_av)); wr 4147 drivers/infiniband/hw/mlx5/qp.c dseg->av.dqp_dct = cpu_to_be32(ud_wr(wr)->remote_qpn | MLX5_EXTENDED_UD_AV); wr 4148 drivers/infiniband/hw/mlx5/qp.c dseg->av.key.qkey.qkey = cpu_to_be32(ud_wr(wr)->remote_qkey); wr 4292 drivers/infiniband/hw/mlx5/qp.c const struct ib_send_wr *wr, int atomic) wr 4294 drivers/infiniband/hw/mlx5/qp.c const struct mlx5_umr_wr *umrwr = umr_wr(wr); wr 4299 drivers/infiniband/hw/mlx5/qp.c if (wr->send_flags & MLX5_IB_SEND_UMR_FAIL_IF_FREE) wr 4308 drivers/infiniband/hw/mlx5/qp.c if (wr->send_flags & MLX5_IB_SEND_UMR_UPDATE_XLT) { wr 4315 drivers/infiniband/hw/mlx5/qp.c if (wr->send_flags & MLX5_IB_SEND_UMR_UPDATE_TRANSLATION) wr 4317 drivers/infiniband/hw/mlx5/qp.c if (wr->send_flags & MLX5_IB_SEND_UMR_UPDATE_PD_ACCESS) { wr 4321 drivers/infiniband/hw/mlx5/qp.c if (wr->send_flags & MLX5_IB_SEND_UMR_ENABLE_MR) wr 4323 drivers/infiniband/hw/mlx5/qp.c if (wr->send_flags & MLX5_IB_SEND_UMR_DISABLE_MR) wr 4326 drivers/infiniband/hw/mlx5/qp.c if (!wr->num_sge) wr 4370 drivers/infiniband/hw/mlx5/qp.c const struct ib_send_wr *wr) wr 4372 drivers/infiniband/hw/mlx5/qp.c const struct mlx5_umr_wr *umrwr = umr_wr(wr); wr 4375 drivers/infiniband/hw/mlx5/qp.c if (wr->send_flags & MLX5_IB_SEND_UMR_DISABLE_MR) wr 4381 drivers/infiniband/hw/mlx5/qp.c if (wr->send_flags & MLX5_IB_SEND_UMR_UPDATE_TRANSLATION && wr 4403 drivers/infiniband/hw/mlx5/qp.c static __be32 send_ieth(const struct ib_send_wr *wr) wr 4405 drivers/infiniband/hw/mlx5/qp.c switch (wr->opcode) { wr 4408 drivers/infiniband/hw/mlx5/qp.c return wr->ex.imm_data; wr 4411 drivers/infiniband/hw/mlx5/qp.c return cpu_to_be32(wr->ex.invalidate_rkey); wr 4435 drivers/infiniband/hw/mlx5/qp.c static int set_data_inl_seg(struct mlx5_ib_qp *qp, const struct ib_send_wr *wr, wr 4447 drivers/infiniband/hw/mlx5/qp.c for (i = 0; i < wr->num_sge; i++) { wr 4448 drivers/infiniband/hw/mlx5/qp.c size_t len = wr->sg_list[i].length; wr 4449 drivers/infiniband/hw/mlx5/qp.c void *addr = (void *)(unsigned long)(wr->sg_list[i].addr); wr 4734 drivers/infiniband/hw/mlx5/qp.c const struct ib_reg_wr *wr = reg_wr(send_wr); wr 4735 drivers/infiniband/hw/mlx5/qp.c struct mlx5_ib_mr *sig_mr = to_mmr(wr->mr); wr 4743 drivers/infiniband/hw/mlx5/qp.c unlikely(wr->access & IB_ACCESS_REMOTE_ATOMIC) || wr 4766 drivers/infiniband/hw/mlx5/qp.c set_sig_mkey_segment(*seg, wr->mr, wr->access, xlt_size, region_len, wr 4772 drivers/infiniband/hw/mlx5/qp.c ret = set_sig_data_segment(send_wr, wr->mr, sig_attrs, qp, seg, size, wr 4809 drivers/infiniband/hw/mlx5/qp.c const struct ib_reg_wr *wr, wr 4813 drivers/infiniband/hw/mlx5/qp.c struct mlx5_ib_mr *mr = to_mmr(wr->mr); wr 4818 drivers/infiniband/hw/mlx5/qp.c bool atomic = wr->access & IB_ACCESS_REMOTE_ATOMIC; wr 4831 drivers/infiniband/hw/mlx5/qp.c if (unlikely(wr->wr.send_flags & IB_SEND_INLINE)) { wr 4847 drivers/infiniband/hw/mlx5/qp.c set_reg_mkey_seg(*seg, mr, wr->key, wr->access); wr 4898 drivers/infiniband/hw/mlx5/qp.c const struct ib_send_wr *wr, unsigned int *idx, wr 4909 drivers/infiniband/hw/mlx5/qp.c (*ctrl)->imm = send_ieth(wr); wr 4923 drivers/infiniband/hw/mlx5/qp.c const struct ib_send_wr *wr, unsigned *idx, wr 4926 drivers/infiniband/hw/mlx5/qp.c return __begin_wqe(qp, seg, ctrl, wr, idx, size, cur_edge, nreq, wr 4927 drivers/infiniband/hw/mlx5/qp.c wr->send_flags & IB_SEND_SIGNALED, wr 4928 drivers/infiniband/hw/mlx5/qp.c wr->send_flags & IB_SEND_SOLICITED); wr 4962 drivers/infiniband/hw/mlx5/qp.c static int _mlx5_ib_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr, wr 4990 drivers/infiniband/hw/mlx5/qp.c *bad_wr = wr; wr 4995 drivers/infiniband/hw/mlx5/qp.c return mlx5_ib_gsi_post_send(ibqp, wr, bad_wr); wr 5002 drivers/infiniband/hw/mlx5/qp.c for (nreq = 0; wr; nreq++, wr = wr->next) { wr 5003 drivers/infiniband/hw/mlx5/qp.c if (unlikely(wr->opcode >= ARRAY_SIZE(mlx5_ib_opcode))) { wr 5006 drivers/infiniband/hw/mlx5/qp.c *bad_wr = wr; wr 5010 drivers/infiniband/hw/mlx5/qp.c num_sge = wr->num_sge; wr 5014 drivers/infiniband/hw/mlx5/qp.c *bad_wr = wr; wr 5018 drivers/infiniband/hw/mlx5/qp.c err = begin_wqe(qp, &seg, &ctrl, wr, &idx, &size, &cur_edge, wr 5023 drivers/infiniband/hw/mlx5/qp.c *bad_wr = wr; wr 5027 drivers/infiniband/hw/mlx5/qp.c if (wr->opcode == IB_WR_REG_MR || wr 5028 drivers/infiniband/hw/mlx5/qp.c wr->opcode == IB_WR_REG_MR_INTEGRITY) { wr 5032 drivers/infiniband/hw/mlx5/qp.c if (wr->send_flags & IB_SEND_FENCE) { wr 5049 drivers/infiniband/hw/mlx5/qp.c switch (wr->opcode) { wr 5053 drivers/infiniband/hw/mlx5/qp.c set_raddr_seg(seg, rdma_wr(wr)->remote_addr, wr 5054 drivers/infiniband/hw/mlx5/qp.c rdma_wr(wr)->rkey); wr 5064 drivers/infiniband/hw/mlx5/qp.c *bad_wr = wr; wr 5069 drivers/infiniband/hw/mlx5/qp.c ctrl->imm = cpu_to_be32(wr->ex.invalidate_rkey); wr 5076 drivers/infiniband/hw/mlx5/qp.c ctrl->imm = cpu_to_be32(reg_wr(wr)->key); wr 5077 drivers/infiniband/hw/mlx5/qp.c err = set_reg_wr(qp, reg_wr(wr), &seg, &size, wr 5080 drivers/infiniband/hw/mlx5/qp.c *bad_wr = wr; wr 5089 drivers/infiniband/hw/mlx5/qp.c mr = to_mmr(reg_wr(wr)->mr); wr 5097 drivers/infiniband/hw/mlx5/qp.c reg_pi_wr.access = reg_wr(wr)->access; wr 5106 drivers/infiniband/hw/mlx5/qp.c *bad_wr = wr; wr 5110 drivers/infiniband/hw/mlx5/qp.c cur_edge, idx, wr->wr_id, wr 5114 drivers/infiniband/hw/mlx5/qp.c err = begin_wqe(qp, &seg, &ctrl, wr, wr 5120 drivers/infiniband/hw/mlx5/qp.c *bad_wr = wr; wr 5146 drivers/infiniband/hw/mlx5/qp.c err = set_pi_umr_wr(wr, qp, &seg, &size, wr 5150 drivers/infiniband/hw/mlx5/qp.c *bad_wr = wr; wr 5154 drivers/infiniband/hw/mlx5/qp.c wr->wr_id, nreq, fence, wr 5162 drivers/infiniband/hw/mlx5/qp.c err = __begin_wqe(qp, &seg, &ctrl, wr, &idx, wr 5168 drivers/infiniband/hw/mlx5/qp.c *bad_wr = wr; wr 5176 drivers/infiniband/hw/mlx5/qp.c *bad_wr = wr; wr 5180 drivers/infiniband/hw/mlx5/qp.c wr->wr_id, nreq, next_fence, wr 5183 drivers/infiniband/hw/mlx5/qp.c err = __begin_wqe(qp, &seg, &ctrl, wr, &idx, wr 5189 drivers/infiniband/hw/mlx5/qp.c *bad_wr = wr; wr 5197 drivers/infiniband/hw/mlx5/qp.c *bad_wr = wr; wr 5201 drivers/infiniband/hw/mlx5/qp.c wr->wr_id, nreq, next_fence, wr 5215 drivers/infiniband/hw/mlx5/qp.c switch (wr->opcode) { wr 5218 drivers/infiniband/hw/mlx5/qp.c set_raddr_seg(seg, rdma_wr(wr)->remote_addr, wr 5219 drivers/infiniband/hw/mlx5/qp.c rdma_wr(wr)->rkey); wr 5233 drivers/infiniband/hw/mlx5/qp.c *bad_wr = wr; wr 5238 drivers/infiniband/hw/mlx5/qp.c set_datagram_seg(seg, wr); wr 5245 drivers/infiniband/hw/mlx5/qp.c set_datagram_seg(seg, wr); wr 5258 drivers/infiniband/hw/mlx5/qp.c set_eth_seg(wr, qp, &seg, &size, &cur_edge); wr 5264 drivers/infiniband/hw/mlx5/qp.c if (wr->opcode != MLX5_IB_WR_UMR) { wr 5270 drivers/infiniband/hw/mlx5/qp.c ctrl->imm = cpu_to_be32(umr_wr(wr)->mkey); wr 5271 drivers/infiniband/hw/mlx5/qp.c err = set_reg_umr_segment(dev, seg, wr, !!(MLX5_CAP_GEN(mdev, atomic))); wr 5277 drivers/infiniband/hw/mlx5/qp.c set_reg_mkey_segment(seg, wr); wr 5287 drivers/infiniband/hw/mlx5/qp.c if (wr->send_flags & IB_SEND_INLINE && num_sge) { wr 5288 drivers/infiniband/hw/mlx5/qp.c err = set_data_inl_seg(qp, wr, &seg, &size, &cur_edge); wr 5291 drivers/infiniband/hw/mlx5/qp.c *bad_wr = wr; wr 5298 drivers/infiniband/hw/mlx5/qp.c if (likely(wr->sg_list[i].length)) { wr 5301 drivers/infiniband/hw/mlx5/qp.c wr->sg_list + i); wr 5309 drivers/infiniband/hw/mlx5/qp.c finish_wqe(qp, ctrl, seg, size, cur_edge, idx, wr->wr_id, nreq, wr 5310 drivers/infiniband/hw/mlx5/qp.c fence, mlx5_ib_opcode[wr->opcode]); wr 5344 drivers/infiniband/hw/mlx5/qp.c int mlx5_ib_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr, wr 5347 drivers/infiniband/hw/mlx5/qp.c return _mlx5_ib_post_send(ibqp, wr, bad_wr, false); wr 5355 drivers/infiniband/hw/mlx5/qp.c static int _mlx5_ib_post_recv(struct ib_qp *ibqp, const struct ib_recv_wr *wr, wr 5371 drivers/infiniband/hw/mlx5/qp.c *bad_wr = wr; wr 5376 drivers/infiniband/hw/mlx5/qp.c return mlx5_ib_gsi_post_recv(ibqp, wr, bad_wr); wr 5382 drivers/infiniband/hw/mlx5/qp.c for (nreq = 0; wr; nreq++, wr = wr->next) { wr 5385 drivers/infiniband/hw/mlx5/qp.c *bad_wr = wr; wr 5389 drivers/infiniband/hw/mlx5/qp.c if (unlikely(wr->num_sge > qp->rq.max_gs)) { wr 5391 drivers/infiniband/hw/mlx5/qp.c *bad_wr = wr; wr 5399 drivers/infiniband/hw/mlx5/qp.c for (i = 0; i < wr->num_sge; i++) wr 5400 drivers/infiniband/hw/mlx5/qp.c set_data_ptr_seg(scat + i, wr->sg_list + i); wr 5413 drivers/infiniband/hw/mlx5/qp.c qp->rq.wrid[ind] = wr->wr_id; wr 5435 drivers/infiniband/hw/mlx5/qp.c int mlx5_ib_post_recv(struct ib_qp *ibqp, const struct ib_recv_wr *wr, wr 5438 drivers/infiniband/hw/mlx5/qp.c return _mlx5_ib_post_recv(ibqp, wr, bad_wr, false); wr 6446 drivers/infiniband/hw/mlx5/qp.c .wr = { wr 6465 drivers/infiniband/hw/mlx5/qp.c ret = _mlx5_ib_post_send(qp, &swr.wr, &bad_swr, true); wr 420 drivers/infiniband/hw/mlx5/srq.c int mlx5_ib_post_srq_recv(struct ib_srq *ibsrq, const struct ib_recv_wr *wr, wr 437 drivers/infiniband/hw/mlx5/srq.c *bad_wr = wr; wr 441 drivers/infiniband/hw/mlx5/srq.c for (nreq = 0; wr; nreq++, wr = wr->next) { wr 442 drivers/infiniband/hw/mlx5/srq.c if (unlikely(wr->num_sge > srq->msrq.max_gs)) { wr 444 drivers/infiniband/hw/mlx5/srq.c *bad_wr = wr; wr 450 drivers/infiniband/hw/mlx5/srq.c *bad_wr = wr; wr 454 drivers/infiniband/hw/mlx5/srq.c srq->wrid[srq->head] = wr->wr_id; wr 460 drivers/infiniband/hw/mlx5/srq.c for (i = 0; i < wr->num_sge; i++) { wr 461 drivers/infiniband/hw/mlx5/srq.c scat[i].byte_count = cpu_to_be32(wr->sg_list[i].length); wr 462 drivers/infiniband/hw/mlx5/srq.c scat[i].lkey = cpu_to_be32(wr->sg_list[i].lkey); wr 463 drivers/infiniband/hw/mlx5/srq.c scat[i].addr = cpu_to_be64(wr->sg_list[i].addr); wr 523 drivers/infiniband/hw/mthca/mthca_dev.h int mthca_tavor_post_srq_recv(struct ib_srq *srq, const struct ib_recv_wr *wr, wr 525 drivers/infiniband/hw/mthca/mthca_dev.h int mthca_arbel_post_srq_recv(struct ib_srq *srq, const struct ib_recv_wr *wr, wr 534 drivers/infiniband/hw/mthca/mthca_dev.h int mthca_tavor_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr, wr 536 drivers/infiniband/hw/mthca/mthca_dev.h int mthca_tavor_post_receive(struct ib_qp *ibqp, const struct ib_recv_wr *wr, wr 538 drivers/infiniband/hw/mthca/mthca_dev.h int mthca_arbel_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr, wr 540 drivers/infiniband/hw/mthca/mthca_dev.h int mthca_arbel_post_receive(struct ib_qp *ibqp, const struct ib_recv_wr *wr, wr 1500 drivers/infiniband/hw/mthca/mthca_qp.c int ind, const struct ib_ud_wr *wr, wr 1509 drivers/infiniband/hw/mthca/mthca_qp.c mthca_ah_grh_present(to_mah(wr->ah)), 0, 0, 0, wr 1512 drivers/infiniband/hw/mthca/mthca_qp.c err = mthca_read_ah(dev, to_mah(wr->ah), &sqp->ud_header); wr 1523 drivers/infiniband/hw/mthca/mthca_qp.c switch (wr->wr.opcode) { wr 1531 drivers/infiniband/hw/mthca/mthca_qp.c sqp->ud_header.immediate_data = wr->wr.ex.imm_data; wr 1540 drivers/infiniband/hw/mthca/mthca_qp.c sqp->ud_header.bth.solicited_event = !!(wr->wr.send_flags & IB_SEND_SOLICITED); wr 1546 drivers/infiniband/hw/mthca/mthca_qp.c wr->pkey_index, &pkey); wr 1548 drivers/infiniband/hw/mthca/mthca_qp.c sqp->ud_header.bth.destination_qpn = cpu_to_be32(wr->remote_qpn); wr 1550 drivers/infiniband/hw/mthca/mthca_qp.c sqp->ud_header.deth.qkey = cpu_to_be32(wr->remote_qkey & 0x80000000 ? wr 1551 drivers/infiniband/hw/mthca/mthca_qp.c sqp->qkey : wr->remote_qkey); wr 1593 drivers/infiniband/hw/mthca/mthca_qp.c const struct ib_atomic_wr *wr) wr 1595 drivers/infiniband/hw/mthca/mthca_qp.c if (wr->wr.opcode == IB_WR_ATOMIC_CMP_AND_SWP) { wr 1596 drivers/infiniband/hw/mthca/mthca_qp.c aseg->swap_add = cpu_to_be64(wr->swap); wr 1597 drivers/infiniband/hw/mthca/mthca_qp.c aseg->compare = cpu_to_be64(wr->compare_add); wr 1599 drivers/infiniband/hw/mthca/mthca_qp.c aseg->swap_add = cpu_to_be64(wr->compare_add); wr 1606 drivers/infiniband/hw/mthca/mthca_qp.c const struct ib_ud_wr *wr) wr 1608 drivers/infiniband/hw/mthca/mthca_qp.c useg->lkey = cpu_to_be32(to_mah(wr->ah)->key); wr 1609 drivers/infiniband/hw/mthca/mthca_qp.c useg->av_addr = cpu_to_be64(to_mah(wr->ah)->avdma); wr 1610 drivers/infiniband/hw/mthca/mthca_qp.c useg->dqpn = cpu_to_be32(wr->remote_qpn); wr 1611 drivers/infiniband/hw/mthca/mthca_qp.c useg->qkey = cpu_to_be32(wr->remote_qkey); wr 1616 drivers/infiniband/hw/mthca/mthca_qp.c const struct ib_ud_wr *wr) wr 1618 drivers/infiniband/hw/mthca/mthca_qp.c memcpy(useg->av, to_mah(wr->ah)->av, MTHCA_AV_SIZE); wr 1619 drivers/infiniband/hw/mthca/mthca_qp.c useg->dqpn = cpu_to_be32(wr->remote_qpn); wr 1620 drivers/infiniband/hw/mthca/mthca_qp.c useg->qkey = cpu_to_be32(wr->remote_qkey); wr 1623 drivers/infiniband/hw/mthca/mthca_qp.c int mthca_tavor_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr, wr 1653 drivers/infiniband/hw/mthca/mthca_qp.c for (nreq = 0; wr; ++nreq, wr = wr->next) { wr 1660 drivers/infiniband/hw/mthca/mthca_qp.c *bad_wr = wr; wr 1671 drivers/infiniband/hw/mthca/mthca_qp.c ((wr->send_flags & IB_SEND_SIGNALED) ? wr 1673 drivers/infiniband/hw/mthca/mthca_qp.c ((wr->send_flags & IB_SEND_SOLICITED) ? wr 1676 drivers/infiniband/hw/mthca/mthca_qp.c if (wr->opcode == IB_WR_SEND_WITH_IMM || wr 1677 drivers/infiniband/hw/mthca/mthca_qp.c wr->opcode == IB_WR_RDMA_WRITE_WITH_IMM) wr 1678 drivers/infiniband/hw/mthca/mthca_qp.c ((struct mthca_next_seg *) wqe)->imm = wr->ex.imm_data; wr 1685 drivers/infiniband/hw/mthca/mthca_qp.c switch (wr->opcode) { wr 1688 drivers/infiniband/hw/mthca/mthca_qp.c set_raddr_seg(wqe, atomic_wr(wr)->remote_addr, wr 1689 drivers/infiniband/hw/mthca/mthca_qp.c atomic_wr(wr)->rkey); wr 1692 drivers/infiniband/hw/mthca/mthca_qp.c set_atomic_seg(wqe, atomic_wr(wr)); wr 1701 drivers/infiniband/hw/mthca/mthca_qp.c set_raddr_seg(wqe, rdma_wr(wr)->remote_addr, wr 1702 drivers/infiniband/hw/mthca/mthca_qp.c rdma_wr(wr)->rkey); wr 1715 drivers/infiniband/hw/mthca/mthca_qp.c switch (wr->opcode) { wr 1718 drivers/infiniband/hw/mthca/mthca_qp.c set_raddr_seg(wqe, rdma_wr(wr)->remote_addr, wr 1719 drivers/infiniband/hw/mthca/mthca_qp.c rdma_wr(wr)->rkey); wr 1732 drivers/infiniband/hw/mthca/mthca_qp.c set_tavor_ud_seg(wqe, ud_wr(wr)); wr 1738 drivers/infiniband/hw/mthca/mthca_qp.c err = build_mlx_header(dev, to_msqp(qp), ind, ud_wr(wr), wr 1742 drivers/infiniband/hw/mthca/mthca_qp.c *bad_wr = wr; wr 1750 drivers/infiniband/hw/mthca/mthca_qp.c if (wr->num_sge > qp->sq.max_gs) { wr 1753 drivers/infiniband/hw/mthca/mthca_qp.c *bad_wr = wr; wr 1757 drivers/infiniband/hw/mthca/mthca_qp.c for (i = 0; i < wr->num_sge; ++i) { wr 1758 drivers/infiniband/hw/mthca/mthca_qp.c mthca_set_data_seg(wqe, wr->sg_list + i); wr 1772 drivers/infiniband/hw/mthca/mthca_qp.c qp->wrid[ind + qp->rq.max] = wr->wr_id; wr 1774 drivers/infiniband/hw/mthca/mthca_qp.c if (wr->opcode >= ARRAY_SIZE(mthca_opcode)) { wr 1777 drivers/infiniband/hw/mthca/mthca_qp.c *bad_wr = wr; wr 1784 drivers/infiniband/hw/mthca/mthca_qp.c mthca_opcode[wr->opcode]); wr 1788 drivers/infiniband/hw/mthca/mthca_qp.c ((wr->send_flags & IB_SEND_FENCE) ? wr 1793 drivers/infiniband/hw/mthca/mthca_qp.c op0 = mthca_opcode[wr->opcode]; wr 1794 drivers/infiniband/hw/mthca/mthca_qp.c f0 = wr->send_flags & IB_SEND_FENCE ? wr 1821 drivers/infiniband/hw/mthca/mthca_qp.c int mthca_tavor_post_receive(struct ib_qp *ibqp, const struct ib_recv_wr *wr, wr 1849 drivers/infiniband/hw/mthca/mthca_qp.c for (nreq = 0; wr; wr = wr->next) { wr 1856 drivers/infiniband/hw/mthca/mthca_qp.c *bad_wr = wr; wr 1871 drivers/infiniband/hw/mthca/mthca_qp.c if (unlikely(wr->num_sge > qp->rq.max_gs)) { wr 1873 drivers/infiniband/hw/mthca/mthca_qp.c *bad_wr = wr; wr 1877 drivers/infiniband/hw/mthca/mthca_qp.c for (i = 0; i < wr->num_sge; ++i) { wr 1878 drivers/infiniband/hw/mthca/mthca_qp.c mthca_set_data_seg(wqe, wr->sg_list + i); wr 1883 drivers/infiniband/hw/mthca/mthca_qp.c qp->wrid[ind] = wr->wr_id; wr 1926 drivers/infiniband/hw/mthca/mthca_qp.c int mthca_arbel_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr, wr 1957 drivers/infiniband/hw/mthca/mthca_qp.c for (nreq = 0; wr; ++nreq, wr = wr->next) { wr 1990 drivers/infiniband/hw/mthca/mthca_qp.c *bad_wr = wr; wr 1999 drivers/infiniband/hw/mthca/mthca_qp.c ((wr->send_flags & IB_SEND_SIGNALED) ? wr 2001 drivers/infiniband/hw/mthca/mthca_qp.c ((wr->send_flags & IB_SEND_SOLICITED) ? wr 2003 drivers/infiniband/hw/mthca/mthca_qp.c ((wr->send_flags & IB_SEND_IP_CSUM) ? wr 2006 drivers/infiniband/hw/mthca/mthca_qp.c if (wr->opcode == IB_WR_SEND_WITH_IMM || wr 2007 drivers/infiniband/hw/mthca/mthca_qp.c wr->opcode == IB_WR_RDMA_WRITE_WITH_IMM) wr 2008 drivers/infiniband/hw/mthca/mthca_qp.c ((struct mthca_next_seg *) wqe)->imm = wr->ex.imm_data; wr 2015 drivers/infiniband/hw/mthca/mthca_qp.c switch (wr->opcode) { wr 2018 drivers/infiniband/hw/mthca/mthca_qp.c set_raddr_seg(wqe, atomic_wr(wr)->remote_addr, wr 2019 drivers/infiniband/hw/mthca/mthca_qp.c atomic_wr(wr)->rkey); wr 2022 drivers/infiniband/hw/mthca/mthca_qp.c set_atomic_seg(wqe, atomic_wr(wr)); wr 2031 drivers/infiniband/hw/mthca/mthca_qp.c set_raddr_seg(wqe, rdma_wr(wr)->remote_addr, wr 2032 drivers/infiniband/hw/mthca/mthca_qp.c rdma_wr(wr)->rkey); wr 2045 drivers/infiniband/hw/mthca/mthca_qp.c switch (wr->opcode) { wr 2048 drivers/infiniband/hw/mthca/mthca_qp.c set_raddr_seg(wqe, rdma_wr(wr)->remote_addr, wr 2049 drivers/infiniband/hw/mthca/mthca_qp.c rdma_wr(wr)->rkey); wr 2062 drivers/infiniband/hw/mthca/mthca_qp.c set_arbel_ud_seg(wqe, ud_wr(wr)); wr 2068 drivers/infiniband/hw/mthca/mthca_qp.c err = build_mlx_header(dev, to_msqp(qp), ind, ud_wr(wr), wr 2072 drivers/infiniband/hw/mthca/mthca_qp.c *bad_wr = wr; wr 2080 drivers/infiniband/hw/mthca/mthca_qp.c if (wr->num_sge > qp->sq.max_gs) { wr 2083 drivers/infiniband/hw/mthca/mthca_qp.c *bad_wr = wr; wr 2087 drivers/infiniband/hw/mthca/mthca_qp.c for (i = 0; i < wr->num_sge; ++i) { wr 2088 drivers/infiniband/hw/mthca/mthca_qp.c mthca_set_data_seg(wqe, wr->sg_list + i); wr 2102 drivers/infiniband/hw/mthca/mthca_qp.c qp->wrid[ind + qp->rq.max] = wr->wr_id; wr 2104 drivers/infiniband/hw/mthca/mthca_qp.c if (wr->opcode >= ARRAY_SIZE(mthca_opcode)) { wr 2107 drivers/infiniband/hw/mthca/mthca_qp.c *bad_wr = wr; wr 2114 drivers/infiniband/hw/mthca/mthca_qp.c mthca_opcode[wr->opcode]); wr 2118 drivers/infiniband/hw/mthca/mthca_qp.c ((wr->send_flags & IB_SEND_FENCE) ? wr 2123 drivers/infiniband/hw/mthca/mthca_qp.c op0 = mthca_opcode[wr->opcode]; wr 2124 drivers/infiniband/hw/mthca/mthca_qp.c f0 = wr->send_flags & IB_SEND_FENCE ? wr 2160 drivers/infiniband/hw/mthca/mthca_qp.c int mthca_arbel_post_receive(struct ib_qp *ibqp, const struct ib_recv_wr *wr, wr 2178 drivers/infiniband/hw/mthca/mthca_qp.c for (nreq = 0; wr; ++nreq, wr = wr->next) { wr 2185 drivers/infiniband/hw/mthca/mthca_qp.c *bad_wr = wr; wr 2195 drivers/infiniband/hw/mthca/mthca_qp.c if (unlikely(wr->num_sge > qp->rq.max_gs)) { wr 2197 drivers/infiniband/hw/mthca/mthca_qp.c *bad_wr = wr; wr 2201 drivers/infiniband/hw/mthca/mthca_qp.c for (i = 0; i < wr->num_sge; ++i) { wr 2202 drivers/infiniband/hw/mthca/mthca_qp.c mthca_set_data_seg(wqe, wr->sg_list + i); wr 2209 drivers/infiniband/hw/mthca/mthca_qp.c qp->wrid[ind] = wr->wr_id; wr 483 drivers/infiniband/hw/mthca/mthca_srq.c int mthca_tavor_post_srq_recv(struct ib_srq *ibsrq, const struct ib_recv_wr *wr, wr 502 drivers/infiniband/hw/mthca/mthca_srq.c for (nreq = 0; wr; wr = wr->next) { wr 510 drivers/infiniband/hw/mthca/mthca_srq.c *bad_wr = wr; wr 522 drivers/infiniband/hw/mthca/mthca_srq.c if (unlikely(wr->num_sge > srq->max_gs)) { wr 524 drivers/infiniband/hw/mthca/mthca_srq.c *bad_wr = wr; wr 529 drivers/infiniband/hw/mthca/mthca_srq.c for (i = 0; i < wr->num_sge; ++i) { wr 530 drivers/infiniband/hw/mthca/mthca_srq.c mthca_set_data_seg(wqe, wr->sg_list + i); wr 540 drivers/infiniband/hw/mthca/mthca_srq.c srq->wrid[ind] = wr->wr_id; wr 577 drivers/infiniband/hw/mthca/mthca_srq.c int mthca_arbel_post_srq_recv(struct ib_srq *ibsrq, const struct ib_recv_wr *wr, wr 592 drivers/infiniband/hw/mthca/mthca_srq.c for (nreq = 0; wr; ++nreq, wr = wr->next) { wr 600 drivers/infiniband/hw/mthca/mthca_srq.c *bad_wr = wr; wr 609 drivers/infiniband/hw/mthca/mthca_srq.c if (unlikely(wr->num_sge > srq->max_gs)) { wr 611 drivers/infiniband/hw/mthca/mthca_srq.c *bad_wr = wr; wr 615 drivers/infiniband/hw/mthca/mthca_srq.c for (i = 0; i < wr->num_sge; ++i) { wr 616 drivers/infiniband/hw/mthca/mthca_srq.c mthca_set_data_seg(wqe, wr->sg_list + i); wr 623 drivers/infiniband/hw/mthca/mthca_srq.c srq->wrid[ind] = wr->wr_id; wr 1888 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c const struct ib_send_wr *wr) wr 1892 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c struct ocrdma_ah *ah = get_ocrdma_ah(ud_wr(wr)->ah); wr 1894 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c ud_hdr->rsvd_dest_qpn = ud_wr(wr)->remote_qpn; wr 1898 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c ud_hdr->qkey = ud_wr(wr)->remote_qkey; wr 1935 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c const struct ib_send_wr *wr, u32 wqe_size) wr 1940 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c if (wr->send_flags & IB_SEND_INLINE && qp->qp_type != IB_QPT_UD) { wr 1941 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c hdr->total_len = ocrdma_sglist_len(wr->sg_list, wr->num_sge); wr 1949 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c for (i = 0; i < wr->num_sge; i++) { wr 1951 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c (void *)(unsigned long)wr->sg_list[i].addr, wr 1952 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c wr->sg_list[i].length); wr 1953 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c dpp_addr += wr->sg_list[i].length; wr 1961 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c ocrdma_build_sges(hdr, sge, wr->num_sge, wr->sg_list); wr 1962 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c if (wr->num_sge) wr 1963 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c wqe_size += (wr->num_sge * sizeof(struct ocrdma_sge)); wr 1973 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c const struct ib_send_wr *wr) wr 1980 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c ocrdma_build_ud_hdr(qp, hdr, wr); wr 1987 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c status = ocrdma_build_inline_sges(qp, hdr, sge, wr, wqe_size); wr 1992 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c const struct ib_send_wr *wr) wr 1999 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c status = ocrdma_build_inline_sges(qp, hdr, sge, wr, wqe_size); wr 2002 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c ext_rw->addr_lo = rdma_wr(wr)->remote_addr; wr 2003 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c ext_rw->addr_hi = upper_32_bits(rdma_wr(wr)->remote_addr); wr 2004 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c ext_rw->lrkey = rdma_wr(wr)->rkey; wr 2010 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c const struct ib_send_wr *wr) wr 2014 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c u32 wqe_size = ((wr->num_sge + 1) * sizeof(struct ocrdma_sge)) + wr 2017 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c ocrdma_build_sges(hdr, sge, wr->num_sge, wr->sg_list); wr 2022 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c ext_rw->addr_lo = rdma_wr(wr)->remote_addr; wr 2023 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c ext_rw->addr_hi = upper_32_bits(rdma_wr(wr)->remote_addr); wr 2024 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c ext_rw->lrkey = rdma_wr(wr)->rkey; wr 2040 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c const struct ib_reg_wr *wr) wr 2044 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c struct ocrdma_mr *mr = get_ocrdma_mr(wr->mr); wr 2055 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c if (wr->access & IB_ACCESS_LOCAL_WRITE) wr 2057 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c if (wr->access & IB_ACCESS_REMOTE_WRITE) wr 2059 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c if (wr->access & IB_ACCESS_REMOTE_READ) wr 2061 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c hdr->lkey = wr->key; wr 2101 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c int ocrdma_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr, wr 2112 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c *bad_wr = wr; wr 2116 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c while (wr) { wr 2118 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c (wr->opcode != IB_WR_SEND && wr 2119 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c wr->opcode != IB_WR_SEND_WITH_IMM)) { wr 2120 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c *bad_wr = wr; wr 2125 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c wr->num_sge > qp->sq.max_sges) { wr 2126 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c *bad_wr = wr; wr 2132 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c if (wr->send_flags & IB_SEND_SIGNALED || qp->signaled) wr 2134 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c if (wr->send_flags & IB_SEND_FENCE) wr 2137 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c if (wr->send_flags & IB_SEND_SOLICITED) wr 2141 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c switch (wr->opcode) { wr 2144 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c hdr->immdt = ntohl(wr->ex.imm_data); wr 2148 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c ocrdma_build_send(qp, hdr, wr); wr 2153 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c hdr->lkey = wr->ex.invalidate_rkey; wr 2154 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c status = ocrdma_build_send(qp, hdr, wr); wr 2158 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c hdr->immdt = ntohl(wr->ex.imm_data); wr 2162 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c status = ocrdma_build_write(qp, hdr, wr); wr 2165 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c ocrdma_build_read(qp, hdr, wr); wr 2173 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c hdr->lkey = wr->ex.invalidate_rkey; wr 2176 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c status = ocrdma_build_reg(qp, hdr, reg_wr(wr)); wr 2183 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c *bad_wr = wr; wr 2186 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c if (wr->send_flags & IB_SEND_SIGNALED || qp->signaled) wr 2190 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c qp->wqe_wr_id_tbl[qp->sq.head].wrid = wr->wr_id; wr 2200 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c wr = wr->next; wr 2214 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c const struct ib_recv_wr *wr, u16 tag) wr 2218 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c if (wr->num_sge) wr 2219 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c wqe_size = (wr->num_sge * sizeof(*sge)) + sizeof(*rqe); wr 2230 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c ocrdma_build_sges(rqe, sge, wr->num_sge, wr->sg_list); wr 2234 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c int ocrdma_post_recv(struct ib_qp *ibqp, const struct ib_recv_wr *wr, wr 2245 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c *bad_wr = wr; wr 2248 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c while (wr) { wr 2250 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c wr->num_sge > qp->rq.max_sges) { wr 2251 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c *bad_wr = wr; wr 2256 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c ocrdma_build_rqe(rqe, wr, 0); wr 2258 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c qp->rqe_wr_id_tbl[qp->rq.head] = wr->wr_id; wr 2267 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c wr = wr->next; wr 2304 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c int ocrdma_post_srq_recv(struct ib_srq *ibsrq, const struct ib_recv_wr *wr, wr 2316 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c while (wr) { wr 2318 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c wr->num_sge > srq->rq.max_sges) { wr 2320 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c *bad_wr = wr; wr 2325 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c ocrdma_build_rqe(rqe, wr, tag); wr 2327 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c srq->rqe_wr_id_tbl[tag] = wr->wr_id; wr 2334 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c wr = wr->next; wr 543 drivers/infiniband/hw/qedr/qedr_roce_cm.c int qedr_gsi_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr, wr 553 drivers/infiniband/hw/qedr/qedr_roce_cm.c *bad_wr = wr; wr 560 drivers/infiniband/hw/qedr/qedr_roce_cm.c if (wr->num_sge > RDMA_MAX_SGE_PER_SQ_WQE) { wr 562 drivers/infiniband/hw/qedr/qedr_roce_cm.c wr->num_sge, RDMA_MAX_SGE_PER_SQ_WQE); wr 567 drivers/infiniband/hw/qedr/qedr_roce_cm.c if (wr->opcode != IB_WR_SEND) { wr 570 drivers/infiniband/hw/qedr/qedr_roce_cm.c wr->opcode); wr 577 drivers/infiniband/hw/qedr/qedr_roce_cm.c rc = qedr_gsi_build_packet(dev, qp, wr, &pkt); wr 586 drivers/infiniband/hw/qedr/qedr_roce_cm.c qp->wqe_wr_id[qp->sq.prod].wr_id = wr->wr_id; wr 590 drivers/infiniband/hw/qedr/qedr_roce_cm.c wr->opcode, in_irq(), irqs_disabled(), wr->wr_id); wr 594 drivers/infiniband/hw/qedr/qedr_roce_cm.c *bad_wr = wr; wr 599 drivers/infiniband/hw/qedr/qedr_roce_cm.c if (wr->next) { wr 602 drivers/infiniband/hw/qedr/qedr_roce_cm.c *bad_wr = wr->next; wr 609 drivers/infiniband/hw/qedr/qedr_roce_cm.c *bad_wr = wr; wr 613 drivers/infiniband/hw/qedr/qedr_roce_cm.c int qedr_gsi_post_recv(struct ib_qp *ibqp, const struct ib_recv_wr *wr, wr 623 drivers/infiniband/hw/qedr/qedr_roce_cm.c *bad_wr = wr; wr 632 drivers/infiniband/hw/qedr/qedr_roce_cm.c while (wr) { wr 633 drivers/infiniband/hw/qedr/qedr_roce_cm.c if (wr->num_sge > QEDR_GSI_MAX_RECV_SGE) { wr 636 drivers/infiniband/hw/qedr/qedr_roce_cm.c wr->num_sge, QEDR_GSI_MAX_RECV_SGE); wr 642 drivers/infiniband/hw/qedr/qedr_roce_cm.c wr->sg_list[0].addr, wr 643 drivers/infiniband/hw/qedr/qedr_roce_cm.c wr->sg_list[0].length, wr 655 drivers/infiniband/hw/qedr/qedr_roce_cm.c qp->rqe_wr_id[qp->rq.prod].sg_list[0] = wr->sg_list[0]; wr 656 drivers/infiniband/hw/qedr/qedr_roce_cm.c qp->rqe_wr_id[qp->rq.prod].wr_id = wr->wr_id; wr 660 drivers/infiniband/hw/qedr/qedr_roce_cm.c wr = wr->next; wr 668 drivers/infiniband/hw/qedr/qedr_roce_cm.c *bad_wr = wr; wr 49 drivers/infiniband/hw/qedr/qedr_roce_cm.h int qedr_gsi_post_recv(struct ib_qp *ibqp, const struct ib_recv_wr *wr, wr 51 drivers/infiniband/hw/qedr/qedr_roce_cm.h int qedr_gsi_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr, wr 2928 drivers/infiniband/hw/qedr/verbs.c const struct ib_send_wr *wr, wr 2932 drivers/infiniband/hw/qedr/verbs.c u32 data_size = sge_data_len(wr->sg_list, wr->num_sge); wr 2938 drivers/infiniband/hw/qedr/verbs.c *bad_wr = wr; wr 2952 drivers/infiniband/hw/qedr/verbs.c for (i = 0; i < wr->num_sge; i++) { wr 2953 drivers/infiniband/hw/qedr/verbs.c u32 len = wr->sg_list[i].length; wr 2954 drivers/infiniband/hw/qedr/verbs.c void *src = (void *)(uintptr_t)wr->sg_list[i].addr; wr 3013 drivers/infiniband/hw/qedr/verbs.c const struct ib_send_wr *wr) wr 3018 drivers/infiniband/hw/qedr/verbs.c for (i = 0; i < wr->num_sge; i++) { wr 3021 drivers/infiniband/hw/qedr/verbs.c DMA_REGPAIR_LE(sge->addr, wr->sg_list[i].addr); wr 3022 drivers/infiniband/hw/qedr/verbs.c sge->l_key = cpu_to_le32(wr->sg_list[i].lkey); wr 3023 drivers/infiniband/hw/qedr/verbs.c sge->length = cpu_to_le32(wr->sg_list[i].length); wr 3024 drivers/infiniband/hw/qedr/verbs.c data_size += wr->sg_list[i].length; wr 3028 drivers/infiniband/hw/qedr/verbs.c *wqe_size += wr->num_sge; wr 3037 drivers/infiniband/hw/qedr/verbs.c const struct ib_send_wr *wr, wr 3040 drivers/infiniband/hw/qedr/verbs.c rwqe2->r_key = cpu_to_le32(rdma_wr(wr)->rkey); wr 3041 drivers/infiniband/hw/qedr/verbs.c DMA_REGPAIR_LE(rwqe2->remote_va, rdma_wr(wr)->remote_addr); wr 3043 drivers/infiniband/hw/qedr/verbs.c if (wr->send_flags & IB_SEND_INLINE && wr 3044 drivers/infiniband/hw/qedr/verbs.c (wr->opcode == IB_WR_RDMA_WRITE_WITH_IMM || wr 3045 drivers/infiniband/hw/qedr/verbs.c wr->opcode == IB_WR_RDMA_WRITE)) { wr 3049 drivers/infiniband/hw/qedr/verbs.c return qedr_prepare_sq_inline_data(dev, qp, &rwqe->wqe_size, wr, wr 3053 drivers/infiniband/hw/qedr/verbs.c return qedr_prepare_sq_sges(qp, &rwqe->wqe_size, wr); wr 3060 drivers/infiniband/hw/qedr/verbs.c const struct ib_send_wr *wr, wr 3064 drivers/infiniband/hw/qedr/verbs.c if (wr->send_flags & IB_SEND_INLINE) { wr 3068 drivers/infiniband/hw/qedr/verbs.c return qedr_prepare_sq_inline_data(dev, qp, &swqe->wqe_size, wr, wr 3072 drivers/infiniband/hw/qedr/verbs.c return qedr_prepare_sq_sges(qp, &swqe->wqe_size, wr); wr 3077 drivers/infiniband/hw/qedr/verbs.c const struct ib_reg_wr *wr) wr 3079 drivers/infiniband/hw/qedr/verbs.c struct qedr_mr *mr = get_qedr_mr(wr->mr); wr 3085 drivers/infiniband/hw/qedr/verbs.c fwqe1->l_key = wr->key; wr 3090 drivers/infiniband/hw/qedr/verbs.c !!(wr->access & IB_ACCESS_REMOTE_READ)); wr 3092 drivers/infiniband/hw/qedr/verbs.c !!(wr->access & IB_ACCESS_REMOTE_WRITE)); wr 3094 drivers/infiniband/hw/qedr/verbs.c !!(wr->access & IB_ACCESS_REMOTE_ATOMIC)); wr 3097 drivers/infiniband/hw/qedr/verbs.c !!(wr->access & IB_ACCESS_LOCAL_WRITE)); wr 3140 drivers/infiniband/hw/qedr/verbs.c const struct ib_send_wr *wr) wr 3146 drivers/infiniband/hw/qedr/verbs.c err_wr = wr->num_sge > qp->sq.max_sges; wr 3177 drivers/infiniband/hw/qedr/verbs.c static int __qedr_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr, wr 3196 drivers/infiniband/hw/qedr/verbs.c if (!qedr_can_post_send(qp, wr)) { wr 3197 drivers/infiniband/hw/qedr/verbs.c *bad_wr = wr; wr 3203 drivers/infiniband/hw/qedr/verbs.c !!(wr->send_flags & IB_SEND_SIGNALED) || qp->signaled; wr 3207 drivers/infiniband/hw/qedr/verbs.c !!(wr->send_flags & IB_SEND_SOLICITED)); wr 3208 drivers/infiniband/hw/qedr/verbs.c comp = (!!(wr->send_flags & IB_SEND_SIGNALED)) || qp->signaled; wr 3211 drivers/infiniband/hw/qedr/verbs.c !!(wr->send_flags & IB_SEND_FENCE)); wr 3214 drivers/infiniband/hw/qedr/verbs.c qp->wqe_wr_id[qp->sq.prod].opcode = qedr_ib_to_wc_opcode(wr->opcode); wr 3216 drivers/infiniband/hw/qedr/verbs.c switch (wr->opcode) { wr 3220 drivers/infiniband/hw/qedr/verbs.c *bad_wr = wr; wr 3228 drivers/infiniband/hw/qedr/verbs.c swqe->inv_key_or_imm_data = cpu_to_le32(be32_to_cpu(wr->ex.imm_data)); wr 3230 drivers/infiniband/hw/qedr/verbs.c wr, bad_wr); wr 3243 drivers/infiniband/hw/qedr/verbs.c wr, bad_wr); wr 3254 drivers/infiniband/hw/qedr/verbs.c swqe->inv_key_or_imm_data = cpu_to_le32(wr->ex.invalidate_rkey); wr 3256 drivers/infiniband/hw/qedr/verbs.c wr, bad_wr); wr 3266 drivers/infiniband/hw/qedr/verbs.c *bad_wr = wr; wr 3273 drivers/infiniband/hw/qedr/verbs.c rwqe->imm_data = htonl(cpu_to_le32(wr->ex.imm_data)); wr 3276 drivers/infiniband/hw/qedr/verbs.c wr, bad_wr); wr 3289 drivers/infiniband/hw/qedr/verbs.c wr, bad_wr); wr 3306 drivers/infiniband/hw/qedr/verbs.c wr, bad_wr); wr 3319 drivers/infiniband/hw/qedr/verbs.c DMA_REGPAIR_LE(awqe2->remote_va, atomic_wr(wr)->remote_addr); wr 3320 drivers/infiniband/hw/qedr/verbs.c awqe2->r_key = cpu_to_le32(atomic_wr(wr)->rkey); wr 3324 drivers/infiniband/hw/qedr/verbs.c if (wr->opcode == IB_WR_ATOMIC_FETCH_AND_ADD) { wr 3327 drivers/infiniband/hw/qedr/verbs.c atomic_wr(wr)->compare_add); wr 3331 drivers/infiniband/hw/qedr/verbs.c atomic_wr(wr)->swap); wr 3333 drivers/infiniband/hw/qedr/verbs.c atomic_wr(wr)->compare_add); wr 3336 drivers/infiniband/hw/qedr/verbs.c qedr_prepare_sq_sges(qp, NULL, wr); wr 3347 drivers/infiniband/hw/qedr/verbs.c iwqe->inv_l_key = wr->ex.invalidate_rkey; wr 3357 drivers/infiniband/hw/qedr/verbs.c rc = qedr_prepare_reg(qp, fwqe1, reg_wr(wr)); wr 3360 drivers/infiniband/hw/qedr/verbs.c *bad_wr = wr; wr 3368 drivers/infiniband/hw/qedr/verbs.c DP_ERR(dev, "invalid opcode 0x%x!\n", wr->opcode); wr 3370 drivers/infiniband/hw/qedr/verbs.c *bad_wr = wr; wr 3392 drivers/infiniband/hw/qedr/verbs.c int qedr_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr, wr 3403 drivers/infiniband/hw/qedr/verbs.c return qedr_gsi_post_send(ibqp, wr, bad_wr); wr 3412 drivers/infiniband/hw/qedr/verbs.c *bad_wr = wr; wr 3420 drivers/infiniband/hw/qedr/verbs.c while (wr) { wr 3421 drivers/infiniband/hw/qedr/verbs.c rc = __qedr_post_send(ibqp, wr, bad_wr); wr 3425 drivers/infiniband/hw/qedr/verbs.c qp->wqe_wr_id[qp->sq.prod].wr_id = wr->wr_id; wr 3431 drivers/infiniband/hw/qedr/verbs.c wr = wr->next; wr 3468 drivers/infiniband/hw/qedr/verbs.c int qedr_post_srq_recv(struct ib_srq *ibsrq, const struct ib_recv_wr *wr, wr 3484 drivers/infiniband/hw/qedr/verbs.c while (wr) { wr 3489 drivers/infiniband/hw/qedr/verbs.c wr->num_sge > srq->hw_srq.max_sges) { wr 3492 drivers/infiniband/hw/qedr/verbs.c wr->num_sge, srq->hw_srq.max_sges); wr 3494 drivers/infiniband/hw/qedr/verbs.c *bad_wr = wr; wr 3499 drivers/infiniband/hw/qedr/verbs.c num_sge = wr->num_sge; wr 3501 drivers/infiniband/hw/qedr/verbs.c SRQ_HDR_SET(hdr, wr->wr_id, num_sge); wr 3509 drivers/infiniband/hw/qedr/verbs.c wr->num_sge, hw_srq->wqe_prod, wr->wr_id); wr 3511 drivers/infiniband/hw/qedr/verbs.c for (i = 0; i < wr->num_sge; i++) { wr 3515 drivers/infiniband/hw/qedr/verbs.c SRQ_SGE_SET(srq_sge, wr->sg_list[i].addr, wr 3516 drivers/infiniband/hw/qedr/verbs.c wr->sg_list[i].length, wr->sg_list[i].lkey); wr 3541 drivers/infiniband/hw/qedr/verbs.c wr = wr->next; wr 3551 drivers/infiniband/hw/qedr/verbs.c int qedr_post_recv(struct ib_qp *ibqp, const struct ib_recv_wr *wr, wr 3560 drivers/infiniband/hw/qedr/verbs.c return qedr_gsi_post_recv(ibqp, wr, bad_wr); wr 3566 drivers/infiniband/hw/qedr/verbs.c *bad_wr = wr; wr 3570 drivers/infiniband/hw/qedr/verbs.c while (wr) { wr 3575 drivers/infiniband/hw/qedr/verbs.c wr->num_sge > qp->rq.max_sges) { wr 3578 drivers/infiniband/hw/qedr/verbs.c QEDR_MAX_RQE_ELEMENTS_PER_RQE, wr->num_sge, wr 3581 drivers/infiniband/hw/qedr/verbs.c *bad_wr = wr; wr 3584 drivers/infiniband/hw/qedr/verbs.c for (i = 0; i < wr->num_sge; i++) { wr 3594 drivers/infiniband/hw/qedr/verbs.c wr->num_sge); wr 3597 drivers/infiniband/hw/qedr/verbs.c wr->sg_list[i].lkey); wr 3599 drivers/infiniband/hw/qedr/verbs.c RQ_SGE_SET(rqe, wr->sg_list[i].addr, wr 3600 drivers/infiniband/hw/qedr/verbs.c wr->sg_list[i].length, flags); wr 3607 drivers/infiniband/hw/qedr/verbs.c if (!wr->num_sge) { wr 3622 drivers/infiniband/hw/qedr/verbs.c qp->rqe_wr_id[qp->rq.prod].wr_id = wr->wr_id; wr 3644 drivers/infiniband/hw/qedr/verbs.c wr = wr->next; wr 72 drivers/infiniband/hw/qedr/verbs.h int qedr_post_srq_recv(struct ib_srq *ibsrq, const struct ib_recv_wr *wr, wr 437 drivers/infiniband/hw/qib/qib_qp.c wqe->wr.opcode, wr 296 drivers/infiniband/hw/qib/qib_rc.c if ((wqe->wr.send_flags & IB_SEND_FENCE) && wr 312 drivers/infiniband/hw/qib/qib_rc.c switch (wqe->wr.opcode) { wr 323 drivers/infiniband/hw/qib/qib_rc.c if (wqe->wr.opcode == IB_WR_SEND) wr 328 drivers/infiniband/hw/qib/qib_rc.c ohdr->u.imm_data = wqe->wr.ex.imm_data; wr 331 drivers/infiniband/hw/qib/qib_rc.c if (wqe->wr.send_flags & IB_SEND_SOLICITED) wr 358 drivers/infiniband/hw/qib/qib_rc.c if (wqe->rdma_wr.wr.opcode == IB_WR_RDMA_WRITE) wr 364 drivers/infiniband/hw/qib/qib_rc.c wqe->rdma_wr.wr.ex.imm_data; wr 366 drivers/infiniband/hw/qib/qib_rc.c if (wqe->rdma_wr.wr.send_flags & IB_SEND_SOLICITED) wr 420 drivers/infiniband/hw/qib/qib_rc.c if (wqe->atomic_wr.wr.opcode == IB_WR_ATOMIC_CMP_AND_SWP) { wr 449 drivers/infiniband/hw/qib/qib_rc.c qp->s_sge.num_sge = wqe->wr.num_sge; wr 457 drivers/infiniband/hw/qib/qib_rc.c if (wqe->wr.opcode == IB_WR_RDMA_READ) wr 486 drivers/infiniband/hw/qib/qib_rc.c if (wqe->wr.opcode == IB_WR_SEND) wr 491 drivers/infiniband/hw/qib/qib_rc.c ohdr->u.imm_data = wqe->wr.ex.imm_data; wr 494 drivers/infiniband/hw/qib/qib_rc.c if (wqe->wr.send_flags & IB_SEND_SOLICITED) wr 525 drivers/infiniband/hw/qib/qib_rc.c if (wqe->wr.opcode == IB_WR_RDMA_WRITE) wr 530 drivers/infiniband/hw/qib/qib_rc.c ohdr->u.imm_data = wqe->wr.ex.imm_data; wr 532 drivers/infiniband/hw/qib/qib_rc.c if (wqe->wr.send_flags & IB_SEND_SOLICITED) wr 753 drivers/infiniband/hw/qib/qib_rc.c opcode = wqe->wr.opcode; wr 774 drivers/infiniband/hw/qib/qib_rc.c opcode = wqe->wr.opcode; wr 839 drivers/infiniband/hw/qib/qib_rc.c if (wqe->wr.opcode == IB_WR_RDMA_READ) wr 865 drivers/infiniband/hw/qib/qib_rc.c if (wqe->wr.opcode == IB_WR_RDMA_READ) wr 924 drivers/infiniband/hw/qib/qib_rc.c ib_qib_wc_opcode[wqe->wr.opcode], wr 963 drivers/infiniband/hw/qib/qib_rc.c ib_qib_wc_opcode[wqe->wr.opcode], wr 1039 drivers/infiniband/hw/qib/qib_rc.c if (wqe->wr.opcode == IB_WR_RDMA_READ && wr 1054 drivers/infiniband/hw/qib/qib_rc.c if ((wqe->wr.opcode == IB_WR_RDMA_READ && wr 1056 drivers/infiniband/hw/qib/qib_rc.c ((wqe->wr.opcode == IB_WR_ATOMIC_CMP_AND_SWP || wr 1057 drivers/infiniband/hw/qib/qib_rc.c wqe->wr.opcode == IB_WR_ATOMIC_FETCH_AND_ADD) && wr 1076 drivers/infiniband/hw/qib/qib_rc.c if (wqe->wr.opcode == IB_WR_ATOMIC_CMP_AND_SWP || wr 1077 drivers/infiniband/hw/qib/qib_rc.c wqe->wr.opcode == IB_WR_ATOMIC_FETCH_AND_ADD) { wr 1082 drivers/infiniband/hw/qib/qib_rc.c (wqe->wr.opcode == IB_WR_RDMA_READ || wr 1083 drivers/infiniband/hw/qib/qib_rc.c wqe->wr.opcode == IB_WR_ATOMIC_CMP_AND_SWP || wr 1084 drivers/infiniband/hw/qib/qib_rc.c wqe->wr.opcode == IB_WR_ATOMIC_FETCH_AND_ADD)) { wr 1234 drivers/infiniband/hw/qib/qib_rc.c if (wqe->wr.opcode == IB_WR_RDMA_READ || wr 1235 drivers/infiniband/hw/qib/qib_rc.c wqe->wr.opcode == IB_WR_ATOMIC_CMP_AND_SWP || wr 1236 drivers/infiniband/hw/qib/qib_rc.c wqe->wr.opcode == IB_WR_ATOMIC_FETCH_AND_ADD) wr 1356 drivers/infiniband/hw/qib/qib_rc.c if (unlikely(wqe->wr.opcode != IB_WR_RDMA_READ)) wr 1371 drivers/infiniband/hw/qib/qib_rc.c if (unlikely(wqe->wr.opcode != IB_WR_RDMA_READ)) wr 1430 drivers/infiniband/hw/qib/qib_rc.c if (unlikely(wqe->wr.opcode != IB_WR_RDMA_READ)) wr 100 drivers/infiniband/hw/qib/qib_uc.c qp->s_sge.num_sge = wqe->wr.num_sge; wr 104 drivers/infiniband/hw/qib/qib_uc.c switch (wqe->wr.opcode) { wr 112 drivers/infiniband/hw/qib/qib_uc.c if (wqe->wr.opcode == IB_WR_SEND) wr 118 drivers/infiniband/hw/qib/qib_uc.c ohdr->u.imm_data = wqe->wr.ex.imm_data; wr 121 drivers/infiniband/hw/qib/qib_uc.c if (wqe->wr.send_flags & IB_SEND_SOLICITED) wr 141 drivers/infiniband/hw/qib/qib_uc.c if (wqe->wr.opcode == IB_WR_RDMA_WRITE) wr 147 drivers/infiniband/hw/qib/qib_uc.c ohdr->u.rc.imm_data = wqe->wr.ex.imm_data; wr 149 drivers/infiniband/hw/qib/qib_uc.c if (wqe->wr.send_flags & IB_SEND_SOLICITED) wr 171 drivers/infiniband/hw/qib/qib_uc.c if (wqe->wr.opcode == IB_WR_SEND) wr 176 drivers/infiniband/hw/qib/qib_uc.c ohdr->u.imm_data = wqe->wr.ex.imm_data; wr 179 drivers/infiniband/hw/qib/qib_uc.c if (wqe->wr.send_flags & IB_SEND_SOLICITED) wr 195 drivers/infiniband/hw/qib/qib_uc.c if (wqe->wr.opcode == IB_WR_RDMA_WRITE) wr 201 drivers/infiniband/hw/qib/qib_uc.c ohdr->u.imm_data = wqe->wr.ex.imm_data; wr 203 drivers/infiniband/hw/qib/qib_uc.c if (wqe->wr.send_flags & IB_SEND_SOLICITED) wr 128 drivers/infiniband/hw/qib/qib_ud.c if (swqe->wr.opcode == IB_WR_SEND_WITH_IMM) { wr 130 drivers/infiniband/hw/qib/qib_ud.c wc.ex.imm_data = swqe->wr.ex.imm_data; wr 173 drivers/infiniband/hw/qib/qib_ud.c ssge.num_sge = swqe->wr.num_sge; wr 214 drivers/infiniband/hw/qib/qib_ud.c rvt_recv_cq(qp, &wc, swqe->wr.send_flags & IB_SEND_SOLICITED); wr 319 drivers/infiniband/hw/qib/qib_ud.c qp->s_sge.num_sge = wqe->wr.num_sge; wr 338 drivers/infiniband/hw/qib/qib_ud.c if (wqe->wr.opcode == IB_WR_SEND_WITH_IMM) { wr 340 drivers/infiniband/hw/qib/qib_ud.c ohdr->u.ud.imm_data = wqe->wr.ex.imm_data; wr 361 drivers/infiniband/hw/qib/qib_ud.c if (wqe->wr.send_flags & IB_SEND_SOLICITED) wr 312 drivers/infiniband/hw/qib/qib_verbs.h int qib_post_ud_send(struct rvt_qp *qp, const struct ib_send_wr *wr); wr 594 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c const struct ib_reg_wr *wr) wr 596 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c struct pvrdma_user_mr *mr = to_vmr(wr->mr); wr 598 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c wqe_hdr->wr.fast_reg.iova_start = mr->ibmr.iova; wr 599 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c wqe_hdr->wr.fast_reg.pl_pdir_dma = mr->pdir.dir_dma; wr 600 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c wqe_hdr->wr.fast_reg.page_shift = mr->page_shift; wr 601 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c wqe_hdr->wr.fast_reg.page_list_len = mr->npages; wr 602 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c wqe_hdr->wr.fast_reg.length = mr->ibmr.length; wr 603 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c wqe_hdr->wr.fast_reg.access_flags = wr->access; wr 604 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c wqe_hdr->wr.fast_reg.rkey = wr->key; wr 618 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c int pvrdma_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr, wr 633 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c *bad_wr = wr; wr 639 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c while (wr) { wr 646 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c *bad_wr = wr; wr 651 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c if (unlikely(wr->num_sge > qp->sq.max_sg || wr->num_sge < 0)) { wr 654 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c *bad_wr = wr; wr 659 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c if (unlikely(wr->opcode < 0)) { wr 662 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c *bad_wr = wr; wr 687 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c wr->opcode != IB_WR_SEND) { wr 690 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c *bad_wr = wr; wr 695 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c if (wr->opcode != IB_WR_SEND && wr 696 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c wr->opcode != IB_WR_SEND_WITH_IMM) { wr 699 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c *bad_wr = wr; wr 707 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c wqe_hdr->wr_id = wr->wr_id; wr 708 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c wqe_hdr->num_sge = wr->num_sge; wr 709 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c wqe_hdr->opcode = ib_wr_opcode_to_pvrdma(wr->opcode); wr 710 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c wqe_hdr->send_flags = ib_send_flags_to_pvrdma(wr->send_flags); wr 711 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c if (wr->opcode == IB_WR_SEND_WITH_IMM || wr 712 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c wr->opcode == IB_WR_RDMA_WRITE_WITH_IMM) wr 713 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c wqe_hdr->ex.imm_data = wr->ex.imm_data; wr 716 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c *bad_wr = wr; wr 724 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c if (unlikely(!ud_wr(wr)->ah)) { wr 727 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c *bad_wr = wr; wr 736 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c wqe_hdr->wr.ud.remote_qpn = ud_wr(wr)->remote_qpn; wr 737 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c wqe_hdr->wr.ud.remote_qkey = wr 738 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c ud_wr(wr)->remote_qkey & 0x80000000 ? wr 739 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c qp->qkey : ud_wr(wr)->remote_qkey; wr 740 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c wqe_hdr->wr.ud.av = to_vah(ud_wr(wr)->ah)->av; wr 744 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c switch (wr->opcode) { wr 748 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c wqe_hdr->wr.rdma.remote_addr = wr 749 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c rdma_wr(wr)->remote_addr; wr 750 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c wqe_hdr->wr.rdma.rkey = rdma_wr(wr)->rkey; wr 755 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c wr->ex.invalidate_rkey; wr 759 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c wqe_hdr->wr.atomic.remote_addr = wr 760 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c atomic_wr(wr)->remote_addr; wr 761 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c wqe_hdr->wr.atomic.rkey = atomic_wr(wr)->rkey; wr 762 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c wqe_hdr->wr.atomic.compare_add = wr 763 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c atomic_wr(wr)->compare_add; wr 764 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c if (wr->opcode == IB_WR_ATOMIC_CMP_AND_SWP) wr 765 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c wqe_hdr->wr.atomic.swap = wr 766 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c atomic_wr(wr)->swap; wr 769 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c ret = set_reg_seg(wqe_hdr, reg_wr(wr)); wr 773 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c *bad_wr = wr; wr 786 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c *bad_wr = wr; wr 791 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c for (i = 0; i < wr->num_sge; i++) { wr 793 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c sge->addr = wr->sg_list[i].addr; wr 794 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c sge->length = wr->sg_list[i].length; wr 795 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c sge->lkey = wr->sg_list[i].lkey; wr 806 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c wr = wr->next; wr 828 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c int pvrdma_post_recv(struct ib_qp *ibqp, const struct ib_recv_wr *wr, wr 844 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c *bad_wr = wr; wr 850 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c *bad_wr = wr; wr 856 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c while (wr) { wr 859 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c if (unlikely(wr->num_sge > qp->rq.max_sg || wr 860 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c wr->num_sge < 0)) { wr 862 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c *bad_wr = wr; wr 871 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c *bad_wr = wr; wr 878 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c wqe_hdr->wr_id = wr->wr_id; wr 879 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c wqe_hdr->num_sge = wr->num_sge; wr 883 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c for (i = 0; i < wr->num_sge; i++) { wr 884 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c sge->addr = wr->sg_list[i].addr; wr 885 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c sge->length = wr->sg_list[i].length; wr 886 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c sge->lkey = wr->sg_list[i].lkey; wr 897 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c wr = wr->next; wr 436 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.h int pvrdma_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr, wr 438 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.h int pvrdma_post_recv(struct ib_qp *ibqp, const struct ib_recv_wr *wr, wr 664 drivers/infiniband/sw/rdmavt/qp.c for (i = 0; i < wqe->wr.num_sge; i++) { wr 1821 drivers/infiniband/sw/rdmavt/qp.c int rvt_post_recv(struct ib_qp *ibqp, const struct ib_recv_wr *wr, wr 1832 drivers/infiniband/sw/rdmavt/qp.c *bad_wr = wr; wr 1836 drivers/infiniband/sw/rdmavt/qp.c for (; wr; wr = wr->next) { wr 1841 drivers/infiniband/sw/rdmavt/qp.c if ((unsigned)wr->num_sge > qp->r_rq.max_sge) { wr 1842 drivers/infiniband/sw/rdmavt/qp.c *bad_wr = wr; wr 1852 drivers/infiniband/sw/rdmavt/qp.c *bad_wr = wr; wr 1861 drivers/infiniband/sw/rdmavt/qp.c wc.wr_id = wr->wr_id; wr 1866 drivers/infiniband/sw/rdmavt/qp.c wqe->wr_id = wr->wr_id; wr 1867 drivers/infiniband/sw/rdmavt/qp.c wqe->num_sge = wr->num_sge; wr 1868 drivers/infiniband/sw/rdmavt/qp.c for (i = 0; i < wr->num_sge; i++) { wr 1869 drivers/infiniband/sw/rdmavt/qp.c wqe->sg_list[i].addr = wr->sg_list[i].addr; wr 1870 drivers/infiniband/sw/rdmavt/qp.c wqe->sg_list[i].length = wr->sg_list[i].length; wr 1871 drivers/infiniband/sw/rdmavt/qp.c wqe->sg_list[i].lkey = wr->sg_list[i].lkey; wr 1905 drivers/infiniband/sw/rdmavt/qp.c const struct ib_send_wr *wr) wr 1909 drivers/infiniband/sw/rdmavt/qp.c if (wr->opcode >= RVT_OPERATION_MAX || !post_parms[wr->opcode].length) wr 1911 drivers/infiniband/sw/rdmavt/qp.c if (!(post_parms[wr->opcode].qpt_support & BIT(qp->ibqp.qp_type))) wr 1913 drivers/infiniband/sw/rdmavt/qp.c if ((post_parms[wr->opcode].flags & RVT_OPERATION_PRIV) && wr 1916 drivers/infiniband/sw/rdmavt/qp.c if (post_parms[wr->opcode].flags & RVT_OPERATION_ATOMIC_SGE && wr 1917 drivers/infiniband/sw/rdmavt/qp.c (wr->num_sge == 0 || wr 1918 drivers/infiniband/sw/rdmavt/qp.c wr->sg_list[0].length < sizeof(u64) || wr 1919 drivers/infiniband/sw/rdmavt/qp.c wr->sg_list[0].addr & (sizeof(u64) - 1))) wr 1921 drivers/infiniband/sw/rdmavt/qp.c if (post_parms[wr->opcode].flags & RVT_OPERATION_ATOMIC && wr 1924 drivers/infiniband/sw/rdmavt/qp.c len = post_parms[wr->opcode].length; wr 1928 drivers/infiniband/sw/rdmavt/qp.c if (qp->ibqp.pd != ud_wr(wr)->ah->pd) wr 2000 drivers/infiniband/sw/rdmavt/qp.c const struct ib_send_wr *wr, wr 2020 drivers/infiniband/sw/rdmavt/qp.c if (unlikely(wr->num_sge > qp->s_max_sge)) wr 2023 drivers/infiniband/sw/rdmavt/qp.c ret = rvt_qp_valid_operation(qp, rdi->post_parms, wr); wr 2039 drivers/infiniband/sw/rdmavt/qp.c if ((rdi->post_parms[wr->opcode].flags & RVT_OPERATION_LOCAL)) { wr 2040 drivers/infiniband/sw/rdmavt/qp.c switch (wr->opcode) { wr 2043 drivers/infiniband/sw/rdmavt/qp.c reg_wr(wr)->mr, wr 2044 drivers/infiniband/sw/rdmavt/qp.c reg_wr(wr)->key, wr 2045 drivers/infiniband/sw/rdmavt/qp.c reg_wr(wr)->access); wr 2046 drivers/infiniband/sw/rdmavt/qp.c if (ret || !(wr->send_flags & IB_SEND_SIGNALED)) wr 2050 drivers/infiniband/sw/rdmavt/qp.c if ((wr->send_flags & IB_SEND_FENCE) || wr 2055 drivers/infiniband/sw/rdmavt/qp.c qp, wr->ex.invalidate_rkey); wr 2056 drivers/infiniband/sw/rdmavt/qp.c if (ret || !(wr->send_flags & IB_SEND_SIGNALED)) wr 2065 drivers/infiniband/sw/rdmavt/qp.c reserved_op = rdi->post_parms[wr->opcode].flags & wr 2080 drivers/infiniband/sw/rdmavt/qp.c memcpy(&wqe->wr, wr, cplen); wr 2084 drivers/infiniband/sw/rdmavt/qp.c if (wr->num_sge) { wr 2087 drivers/infiniband/sw/rdmavt/qp.c acc = wr->opcode >= IB_WR_RDMA_READ ? wr 2089 drivers/infiniband/sw/rdmavt/qp.c for (i = 0; i < wr->num_sge; i++) { wr 2090 drivers/infiniband/sw/rdmavt/qp.c u32 length = wr->sg_list[i].length; wr 2095 drivers/infiniband/sw/rdmavt/qp.c &wr->sg_list[i], acc); wr 2103 drivers/infiniband/sw/rdmavt/qp.c wqe->wr.num_sge = j; wr 2119 drivers/infiniband/sw/rdmavt/qp.c if (rdi->post_parms[wr->opcode].flags & RVT_OPERATION_LOCAL) { wr 2123 drivers/infiniband/sw/rdmavt/qp.c wqe->wr.send_flags |= RVT_SEND_COMPLETION_ONLY; wr 2143 drivers/infiniband/sw/rdmavt/qp.c if (!(rdi->post_parms[wr->opcode].flags & RVT_OPERATION_LOCAL)) wr 2147 drivers/infiniband/sw/rdmavt/qp.c wqe->wr.send_flags |= RVT_SEND_RESERVE_USED; wr 2150 drivers/infiniband/sw/rdmavt/qp.c wqe->wr.send_flags &= ~RVT_SEND_RESERVE_USED; wr 2153 drivers/infiniband/sw/rdmavt/qp.c trace_rvt_post_one_wr(qp, wqe, wr->num_sge); wr 2182 drivers/infiniband/sw/rdmavt/qp.c int rvt_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr, wr 2208 drivers/infiniband/sw/rdmavt/qp.c call_send = qp->s_head == READ_ONCE(qp->s_last) && !wr->next; wr 2210 drivers/infiniband/sw/rdmavt/qp.c for (; wr; wr = wr->next) { wr 2211 drivers/infiniband/sw/rdmavt/qp.c err = rvt_post_one_wr(qp, wr, &call_send); wr 2213 drivers/infiniband/sw/rdmavt/qp.c *bad_wr = wr; wr 2243 drivers/infiniband/sw/rdmavt/qp.c int rvt_post_srq_recv(struct ib_srq *ibsrq, const struct ib_recv_wr *wr, wr 2250 drivers/infiniband/sw/rdmavt/qp.c for (; wr; wr = wr->next) { wr 2255 drivers/infiniband/sw/rdmavt/qp.c if ((unsigned)wr->num_sge > srq->rq.max_sge) { wr 2256 drivers/infiniband/sw/rdmavt/qp.c *bad_wr = wr; wr 2267 drivers/infiniband/sw/rdmavt/qp.c *bad_wr = wr; wr 2272 drivers/infiniband/sw/rdmavt/qp.c wqe->wr_id = wr->wr_id; wr 2273 drivers/infiniband/sw/rdmavt/qp.c wqe->num_sge = wr->num_sge; wr 2274 drivers/infiniband/sw/rdmavt/qp.c for (i = 0; i < wr->num_sge; i++) { wr 2275 drivers/infiniband/sw/rdmavt/qp.c wqe->sg_list[i].addr = wr->sg_list[i].addr; wr 2276 drivers/infiniband/sw/rdmavt/qp.c wqe->sg_list[i].length = wr->sg_list[i].length; wr 2277 drivers/infiniband/sw/rdmavt/qp.c wqe->sg_list[i].lkey = wr->sg_list[i].lkey; wr 2852 drivers/infiniband/sw/rdmavt/qp.c last = rvt_qp_complete_swqe(qp, wqe, rdi->wc_opcode[wqe->wr.opcode], wr 3043 drivers/infiniband/sw/rdmavt/qp.c sqp->s_sge.num_sge = wqe->wr.num_sge; wr 3045 drivers/infiniband/sw/rdmavt/qp.c switch (wqe->wr.opcode) { wr 3050 drivers/infiniband/sw/rdmavt/qp.c if (!(wqe->wr.send_flags & RVT_SEND_COMPLETION_ONLY)) { wr 3052 drivers/infiniband/sw/rdmavt/qp.c wqe->wr.ex.invalidate_rkey)) wr 3068 drivers/infiniband/sw/rdmavt/qp.c switch (wqe->wr.opcode) { wr 3071 drivers/infiniband/sw/rdmavt/qp.c wqe->wr.ex.invalidate_rkey)) { wr 3074 drivers/infiniband/sw/rdmavt/qp.c wqe->wr.ex.invalidate_rkey; wr 3079 drivers/infiniband/sw/rdmavt/qp.c wc.ex.imm_data = wqe->wr.ex.imm_data; wr 3090 drivers/infiniband/sw/rdmavt/qp.c wc.ex.imm_data = wqe->wr.ex.imm_data; wr 3128 drivers/infiniband/sw/rdmavt/qp.c qp->r_sge.num_sge = wqe->wr.num_sge; wr 3145 drivers/infiniband/sw/rdmavt/qp.c (wqe->wr.opcode == IB_WR_ATOMIC_FETCH_AND_ADD) ? wr 3174 drivers/infiniband/sw/rdmavt/qp.c if (wqe->wr.opcode == IB_WR_RDMA_WRITE_WITH_IMM) wr 3187 drivers/infiniband/sw/rdmavt/qp.c rvt_recv_cq(qp, &wc, wqe->wr.send_flags & IB_SEND_SOLICITED); wr 63 drivers/infiniband/sw/rdmavt/qp.h int rvt_post_recv(struct ib_qp *ibqp, const struct ib_recv_wr *wr, wr 65 drivers/infiniband/sw/rdmavt/qp.h int rvt_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr, wr 67 drivers/infiniband/sw/rdmavt/qp.h int rvt_post_srq_recv(struct ib_srq *ibsrq, const struct ib_recv_wr *wr, wr 203 drivers/infiniband/sw/rdmavt/rc.c ss->num_sge = wqe->wr.num_sge; wr 116 drivers/infiniband/sw/rdmavt/trace_tx.h __entry->wr_id = wqe->wr.wr_id; wr 122 drivers/infiniband/sw/rdmavt/trace_tx.h __entry->opcode = wqe->wr.opcode; wr 129 drivers/infiniband/sw/rdmavt/trace_tx.h __entry->send_flags = wqe->wr.send_flags; wr 130 drivers/infiniband/sw/rdmavt/trace_tx.h __entry->num_sge = wqe->wr.num_sge; wr 175 drivers/infiniband/sw/rdmavt/trace_tx.h __entry->wr_id = wqe->wr.wr_id; wr 181 drivers/infiniband/sw/rdmavt/trace_tx.h __entry->opcode = wqe->wr.opcode; wr 182 drivers/infiniband/sw/rdmavt/trace_tx.h __entry->send_flags = wqe->wr.send_flags; wr 289 drivers/infiniband/sw/rxe/rxe_comp.c if (wqe->wr.opcode != IB_WR_RDMA_READ && wr 290 drivers/infiniband/sw/rxe/rxe_comp.c wqe->wr.opcode != IB_WR_RDMA_READ_WITH_INV) { wr 303 drivers/infiniband/sw/rxe/rxe_comp.c if (wqe->wr.opcode != IB_WR_ATOMIC_CMP_AND_SWP && wr 304 drivers/infiniband/sw/rxe/rxe_comp.c wqe->wr.opcode != IB_WR_ATOMIC_FETCH_AND_ADD) wr 410 drivers/infiniband/sw/rxe/rxe_comp.c wc->wr_id = wqe->wr.wr_id; wr 412 drivers/infiniband/sw/rxe/rxe_comp.c wc->opcode = wr_to_wc_opcode(wqe->wr.opcode); wr 413 drivers/infiniband/sw/rxe/rxe_comp.c if (wqe->wr.opcode == IB_WR_RDMA_WRITE_WITH_IMM || wr 414 drivers/infiniband/sw/rxe/rxe_comp.c wqe->wr.opcode == IB_WR_SEND_WITH_IMM) wr 421 drivers/infiniband/sw/rxe/rxe_comp.c uwc->wr_id = wqe->wr.wr_id; wr 423 drivers/infiniband/sw/rxe/rxe_comp.c uwc->opcode = wr_to_wc_opcode(wqe->wr.opcode); wr 424 drivers/infiniband/sw/rxe/rxe_comp.c if (wqe->wr.opcode == IB_WR_RDMA_WRITE_WITH_IMM || wr 425 drivers/infiniband/sw/rxe/rxe_comp.c wqe->wr.opcode == IB_WR_SEND_WITH_IMM) wr 446 drivers/infiniband/sw/rxe/rxe_comp.c (wqe->wr.send_flags & IB_SEND_SIGNALED) || wr 455 drivers/infiniband/sw/rxe/rxe_comp.c if (wqe->wr.opcode == IB_WR_SEND || wr 456 drivers/infiniband/sw/rxe/rxe_comp.c wqe->wr.opcode == IB_WR_SEND_WITH_IMM || wr 457 drivers/infiniband/sw/rxe/rxe_comp.c wqe->wr.opcode == IB_WR_SEND_WITH_INV) wr 55 drivers/infiniband/sw/rxe/rxe_req.c wqe->wr.opcode); wr 57 drivers/infiniband/sw/rxe/rxe_req.c if (wqe->wr.send_flags & IB_SEND_INLINE) { wr 84 drivers/infiniband/sw/rxe/rxe_req.c mask = wr_opcode_mask(wqe->wr.opcode, qp); wr 93 drivers/infiniband/sw/rxe/rxe_req.c wqe->wr.wr.atomic.remote_addr : wr 95 drivers/infiniband/sw/rxe/rxe_req.c wqe->wr.wr.rdma.remote_addr : wr 184 drivers/infiniband/sw/rxe/rxe_req.c if (unlikely((wqe->wr.send_flags & IB_SEND_FENCE) && wr 190 drivers/infiniband/sw/rxe/rxe_req.c wqe->mask = wr_opcode_mask(wqe->wr.opcode, qp); wr 386 drivers/infiniband/sw/rxe/rxe_req.c struct rxe_send_wr *ibwr = &wqe->wr; wr 423 drivers/infiniband/sw/rxe/rxe_req.c port->pkey_tbl[ibwr->wr.ud.pkey_index] : wr 426 drivers/infiniband/sw/rxe/rxe_req.c qp_num = (pkt->mask & RXE_DETH_MASK) ? ibwr->wr.ud.remote_qpn : wr 439 drivers/infiniband/sw/rxe/rxe_req.c reth_set_rkey(pkt, ibwr->wr.rdma.rkey); wr 454 drivers/infiniband/sw/rxe/rxe_req.c atmeth_set_swap_add(pkt, ibwr->wr.atomic.swap); wr 455 drivers/infiniband/sw/rxe/rxe_req.c atmeth_set_comp(pkt, ibwr->wr.atomic.compare_add); wr 457 drivers/infiniband/sw/rxe/rxe_req.c atmeth_set_swap_add(pkt, ibwr->wr.atomic.compare_add); wr 459 drivers/infiniband/sw/rxe/rxe_req.c atmeth_set_rkey(pkt, ibwr->wr.atomic.rkey); wr 466 drivers/infiniband/sw/rxe/rxe_req.c deth_set_qkey(pkt, ibwr->wr.ud.remote_qkey); wr 487 drivers/infiniband/sw/rxe/rxe_req.c if (wqe->wr.send_flags & IB_SEND_INLINE) { wr 628 drivers/infiniband/sw/rxe/rxe_req.c if (wqe->wr.opcode == IB_WR_LOCAL_INV) { wr 633 drivers/infiniband/sw/rxe/rxe_req.c wqe->wr.ex.invalidate_rkey >> 8); wr 636 drivers/infiniband/sw/rxe/rxe_req.c wqe->wr.ex.invalidate_rkey); wr 645 drivers/infiniband/sw/rxe/rxe_req.c } else if (wqe->wr.opcode == IB_WR_REG_MR) { wr 646 drivers/infiniband/sw/rxe/rxe_req.c struct rxe_mem *rmr = to_rmr(wqe->wr.wr.reg.mr); wr 649 drivers/infiniband/sw/rxe/rxe_req.c rmr->access = wqe->wr.wr.reg.access; wr 650 drivers/infiniband/sw/rxe/rxe_req.c rmr->lkey = wqe->wr.wr.reg.key; wr 651 drivers/infiniband/sw/rxe/rxe_req.c rmr->rkey = wqe->wr.wr.reg.key; wr 652 drivers/infiniband/sw/rxe/rxe_req.c rmr->iova = wqe->wr.wr.reg.mr->iova; wr 658 drivers/infiniband/sw/rxe/rxe_req.c if ((wqe->wr.send_flags & IB_SEND_SIGNALED) || wr 679 drivers/infiniband/sw/rxe/rxe_req.c opcode = next_opcode(qp, wqe, wqe->wr.opcode); wr 387 drivers/infiniband/sw/rxe/rxe_verbs.c static int rxe_post_srq_recv(struct ib_srq *ibsrq, const struct ib_recv_wr *wr, wr 396 drivers/infiniband/sw/rxe/rxe_verbs.c while (wr) { wr 397 drivers/infiniband/sw/rxe/rxe_verbs.c err = post_one_recv(&srq->rq, wr); wr 400 drivers/infiniband/sw/rxe/rxe_verbs.c wr = wr->next; wr 406 drivers/infiniband/sw/rxe/rxe_verbs.c *bad_wr = wr; wr 530 drivers/infiniband/sw/rxe/rxe_verbs.c static void init_send_wr(struct rxe_qp *qp, struct rxe_send_wr *wr, wr 533 drivers/infiniband/sw/rxe/rxe_verbs.c wr->wr_id = ibwr->wr_id; wr 534 drivers/infiniband/sw/rxe/rxe_verbs.c wr->num_sge = ibwr->num_sge; wr 535 drivers/infiniband/sw/rxe/rxe_verbs.c wr->opcode = ibwr->opcode; wr 536 drivers/infiniband/sw/rxe/rxe_verbs.c wr->send_flags = ibwr->send_flags; wr 541 drivers/infiniband/sw/rxe/rxe_verbs.c wr->wr.ud.remote_qpn = ud_wr(ibwr)->remote_qpn; wr 542 drivers/infiniband/sw/rxe/rxe_verbs.c wr->wr.ud.remote_qkey = ud_wr(ibwr)->remote_qkey; wr 544 drivers/infiniband/sw/rxe/rxe_verbs.c wr->wr.ud.pkey_index = ud_wr(ibwr)->pkey_index; wr 545 drivers/infiniband/sw/rxe/rxe_verbs.c if (wr->opcode == IB_WR_SEND_WITH_IMM) wr 546 drivers/infiniband/sw/rxe/rxe_verbs.c wr->ex.imm_data = ibwr->ex.imm_data; wr 548 drivers/infiniband/sw/rxe/rxe_verbs.c switch (wr->opcode) { wr 550 drivers/infiniband/sw/rxe/rxe_verbs.c wr->ex.imm_data = ibwr->ex.imm_data; wr 554 drivers/infiniband/sw/rxe/rxe_verbs.c wr->wr.rdma.remote_addr = rdma_wr(ibwr)->remote_addr; wr 555 drivers/infiniband/sw/rxe/rxe_verbs.c wr->wr.rdma.rkey = rdma_wr(ibwr)->rkey; wr 558 drivers/infiniband/sw/rxe/rxe_verbs.c wr->ex.imm_data = ibwr->ex.imm_data; wr 561 drivers/infiniband/sw/rxe/rxe_verbs.c wr->ex.invalidate_rkey = ibwr->ex.invalidate_rkey; wr 565 drivers/infiniband/sw/rxe/rxe_verbs.c wr->wr.atomic.remote_addr = wr 567 drivers/infiniband/sw/rxe/rxe_verbs.c wr->wr.atomic.compare_add = wr 569 drivers/infiniband/sw/rxe/rxe_verbs.c wr->wr.atomic.swap = atomic_wr(ibwr)->swap; wr 570 drivers/infiniband/sw/rxe/rxe_verbs.c wr->wr.atomic.rkey = atomic_wr(ibwr)->rkey; wr 573 drivers/infiniband/sw/rxe/rxe_verbs.c wr->ex.invalidate_rkey = ibwr->ex.invalidate_rkey; wr 576 drivers/infiniband/sw/rxe/rxe_verbs.c wr->wr.reg.mr = reg_wr(ibwr)->mr; wr 577 drivers/infiniband/sw/rxe/rxe_verbs.c wr->wr.reg.key = reg_wr(ibwr)->key; wr 578 drivers/infiniband/sw/rxe/rxe_verbs.c wr->wr.reg.access = reg_wr(ibwr)->access; wr 595 drivers/infiniband/sw/rxe/rxe_verbs.c init_send_wr(qp, &wqe->wr, ibwr); wr 675 drivers/infiniband/sw/rxe/rxe_verbs.c static int rxe_post_send_kernel(struct rxe_qp *qp, const struct ib_send_wr *wr, wr 683 drivers/infiniband/sw/rxe/rxe_verbs.c while (wr) { wr 684 drivers/infiniband/sw/rxe/rxe_verbs.c mask = wr_opcode_mask(wr->opcode, qp); wr 687 drivers/infiniband/sw/rxe/rxe_verbs.c *bad_wr = wr; wr 691 drivers/infiniband/sw/rxe/rxe_verbs.c if (unlikely((wr->send_flags & IB_SEND_INLINE) && wr 694 drivers/infiniband/sw/rxe/rxe_verbs.c *bad_wr = wr; wr 699 drivers/infiniband/sw/rxe/rxe_verbs.c for (i = 0; i < wr->num_sge; i++) wr 700 drivers/infiniband/sw/rxe/rxe_verbs.c length += wr->sg_list[i].length; wr 702 drivers/infiniband/sw/rxe/rxe_verbs.c err = post_one_send(qp, wr, mask, length); wr 705 drivers/infiniband/sw/rxe/rxe_verbs.c *bad_wr = wr; wr 708 drivers/infiniband/sw/rxe/rxe_verbs.c wr = wr->next; wr 718 drivers/infiniband/sw/rxe/rxe_verbs.c static int rxe_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr, wr 724 drivers/infiniband/sw/rxe/rxe_verbs.c *bad_wr = wr; wr 729 drivers/infiniband/sw/rxe/rxe_verbs.c *bad_wr = wr; wr 738 drivers/infiniband/sw/rxe/rxe_verbs.c return rxe_post_send_kernel(qp, wr, bad_wr); wr 741 drivers/infiniband/sw/rxe/rxe_verbs.c static int rxe_post_recv(struct ib_qp *ibqp, const struct ib_recv_wr *wr, wr 750 drivers/infiniband/sw/rxe/rxe_verbs.c *bad_wr = wr; wr 756 drivers/infiniband/sw/rxe/rxe_verbs.c *bad_wr = wr; wr 763 drivers/infiniband/sw/rxe/rxe_verbs.c while (wr) { wr 764 drivers/infiniband/sw/rxe/rxe_verbs.c err = post_one_recv(rq, wr); wr 766 drivers/infiniband/sw/rxe/rxe_verbs.c *bad_wr = wr; wr 769 drivers/infiniband/sw/rxe/rxe_verbs.c wr = wr->next; wr 689 drivers/infiniband/sw/siw/siw_verbs.c static int siw_sq_flush_wr(struct siw_qp *qp, const struct ib_send_wr *wr, wr 695 drivers/infiniband/sw/siw/siw_verbs.c while (wr) { wr 696 drivers/infiniband/sw/siw/siw_verbs.c sqe.id = wr->wr_id; wr 697 drivers/infiniband/sw/siw/siw_verbs.c sqe.opcode = wr->opcode; wr 701 drivers/infiniband/sw/siw/siw_verbs.c *bad_wr = wr; wr 704 drivers/infiniband/sw/siw/siw_verbs.c wr = wr->next; wr 710 drivers/infiniband/sw/siw/siw_verbs.c static int siw_rq_flush_wr(struct siw_qp *qp, const struct ib_recv_wr *wr, wr 716 drivers/infiniband/sw/siw/siw_verbs.c while (wr) { wr 717 drivers/infiniband/sw/siw/siw_verbs.c rqe.id = wr->wr_id; wr 721 drivers/infiniband/sw/siw/siw_verbs.c *bad_wr = wr; wr 724 drivers/infiniband/sw/siw/siw_verbs.c wr = wr->next; wr 738 drivers/infiniband/sw/siw/siw_verbs.c int siw_post_send(struct ib_qp *base_qp, const struct ib_send_wr *wr, wr 747 drivers/infiniband/sw/siw/siw_verbs.c if (wr && !qp->kernel_verbs) { wr 749 drivers/infiniband/sw/siw/siw_verbs.c *bad_wr = wr; wr 768 drivers/infiniband/sw/siw/siw_verbs.c rv = siw_sq_flush_wr(qp, wr, bad_wr); wr 772 drivers/infiniband/sw/siw/siw_verbs.c *bad_wr = wr; wr 786 drivers/infiniband/sw/siw/siw_verbs.c rv = siw_sq_flush_wr(qp, wr, bad_wr); wr 790 drivers/infiniband/sw/siw/siw_verbs.c *bad_wr = wr; wr 798 drivers/infiniband/sw/siw/siw_verbs.c while (wr) { wr 807 drivers/infiniband/sw/siw/siw_verbs.c if (wr->num_sge > qp->attrs.sq_max_sges) { wr 808 drivers/infiniband/sw/siw/siw_verbs.c siw_dbg_qp(qp, "too many sge's: %d\n", wr->num_sge); wr 812 drivers/infiniband/sw/siw/siw_verbs.c sqe->id = wr->wr_id; wr 814 drivers/infiniband/sw/siw/siw_verbs.c if ((wr->send_flags & IB_SEND_SIGNALED) || wr 818 drivers/infiniband/sw/siw/siw_verbs.c if (wr->send_flags & IB_SEND_FENCE) wr 821 drivers/infiniband/sw/siw/siw_verbs.c switch (wr->opcode) { wr 824 drivers/infiniband/sw/siw/siw_verbs.c if (wr->send_flags & IB_SEND_SOLICITED) wr 827 drivers/infiniband/sw/siw/siw_verbs.c if (!(wr->send_flags & IB_SEND_INLINE)) { wr 828 drivers/infiniband/sw/siw/siw_verbs.c siw_copy_sgl(wr->sg_list, sqe->sge, wr 829 drivers/infiniband/sw/siw/siw_verbs.c wr->num_sge); wr 830 drivers/infiniband/sw/siw/siw_verbs.c sqe->num_sge = wr->num_sge; wr 832 drivers/infiniband/sw/siw/siw_verbs.c rv = siw_copy_inline_sgl(wr, sqe); wr 840 drivers/infiniband/sw/siw/siw_verbs.c if (wr->opcode == IB_WR_SEND) wr 844 drivers/infiniband/sw/siw/siw_verbs.c sqe->rkey = wr->ex.invalidate_rkey; wr 857 drivers/infiniband/sw/siw/siw_verbs.c if (unlikely(wr->num_sge != 1)) { wr 861 drivers/infiniband/sw/siw/siw_verbs.c siw_copy_sgl(wr->sg_list, &sqe->sge[0], 1); wr 865 drivers/infiniband/sw/siw/siw_verbs.c sqe->raddr = rdma_wr(wr)->remote_addr; wr 866 drivers/infiniband/sw/siw/siw_verbs.c sqe->rkey = rdma_wr(wr)->rkey; wr 869 drivers/infiniband/sw/siw/siw_verbs.c if (wr->opcode == IB_WR_RDMA_READ) wr 876 drivers/infiniband/sw/siw/siw_verbs.c if (!(wr->send_flags & IB_SEND_INLINE)) { wr 877 drivers/infiniband/sw/siw/siw_verbs.c siw_copy_sgl(wr->sg_list, &sqe->sge[0], wr 878 drivers/infiniband/sw/siw/siw_verbs.c wr->num_sge); wr 879 drivers/infiniband/sw/siw/siw_verbs.c sqe->num_sge = wr->num_sge; wr 881 drivers/infiniband/sw/siw/siw_verbs.c rv = siw_copy_inline_sgl(wr, sqe); wr 889 drivers/infiniband/sw/siw/siw_verbs.c sqe->raddr = rdma_wr(wr)->remote_addr; wr 890 drivers/infiniband/sw/siw/siw_verbs.c sqe->rkey = rdma_wr(wr)->rkey; wr 895 drivers/infiniband/sw/siw/siw_verbs.c sqe->base_mr = (uintptr_t)reg_wr(wr)->mr; wr 896 drivers/infiniband/sw/siw/siw_verbs.c sqe->rkey = reg_wr(wr)->key; wr 897 drivers/infiniband/sw/siw/siw_verbs.c sqe->access = reg_wr(wr)->access & IWARP_ACCESS_MASK; wr 902 drivers/infiniband/sw/siw/siw_verbs.c sqe->rkey = wr->ex.invalidate_rkey; wr 908 drivers/infiniband/sw/siw/siw_verbs.c wr->opcode); wr 924 drivers/infiniband/sw/siw/siw_verbs.c wr = wr->next; wr 965 drivers/infiniband/sw/siw/siw_verbs.c *bad_wr = wr; wr 978 drivers/infiniband/sw/siw/siw_verbs.c int siw_post_receive(struct ib_qp *base_qp, const struct ib_recv_wr *wr, wr 986 drivers/infiniband/sw/siw/siw_verbs.c *bad_wr = wr; wr 991 drivers/infiniband/sw/siw/siw_verbs.c *bad_wr = wr; wr 1010 drivers/infiniband/sw/siw/siw_verbs.c rv = siw_rq_flush_wr(qp, wr, bad_wr); wr 1014 drivers/infiniband/sw/siw/siw_verbs.c *bad_wr = wr; wr 1028 drivers/infiniband/sw/siw/siw_verbs.c rv = siw_rq_flush_wr(qp, wr, bad_wr); wr 1032 drivers/infiniband/sw/siw/siw_verbs.c *bad_wr = wr; wr 1044 drivers/infiniband/sw/siw/siw_verbs.c while (wr) { wr 1053 drivers/infiniband/sw/siw/siw_verbs.c if (wr->num_sge > qp->attrs.rq_max_sges) { wr 1054 drivers/infiniband/sw/siw/siw_verbs.c siw_dbg_qp(qp, "too many sge's: %d\n", wr->num_sge); wr 1058 drivers/infiniband/sw/siw/siw_verbs.c rqe->id = wr->wr_id; wr 1059 drivers/infiniband/sw/siw/siw_verbs.c rqe->num_sge = wr->num_sge; wr 1060 drivers/infiniband/sw/siw/siw_verbs.c siw_copy_sgl(wr->sg_list, rqe->sge, wr->num_sge); wr 1068 drivers/infiniband/sw/siw/siw_verbs.c wr = wr->next; wr 1076 drivers/infiniband/sw/siw/siw_verbs.c *bad_wr = wr; wr 1744 drivers/infiniband/sw/siw/siw_verbs.c int siw_post_srq_recv(struct ib_srq *base_srq, const struct ib_recv_wr *wr, wr 1764 drivers/infiniband/sw/siw/siw_verbs.c while (wr) { wr 1773 drivers/infiniband/sw/siw/siw_verbs.c if (unlikely(wr->num_sge > srq->max_sge)) { wr 1775 drivers/infiniband/sw/siw/siw_verbs.c "[SRQ]: too many sge's: %d\n", wr->num_sge); wr 1779 drivers/infiniband/sw/siw/siw_verbs.c rqe->id = wr->wr_id; wr 1780 drivers/infiniband/sw/siw/siw_verbs.c rqe->num_sge = wr->num_sge; wr 1781 drivers/infiniband/sw/siw/siw_verbs.c siw_copy_sgl(wr->sg_list, rqe->sge, wr->num_sge); wr 1789 drivers/infiniband/sw/siw/siw_verbs.c wr = wr->next; wr 1795 drivers/infiniband/sw/siw/siw_verbs.c *bad_wr = wr; wr 62 drivers/infiniband/sw/siw/siw_verbs.h int siw_post_send(struct ib_qp *base_qp, const struct ib_send_wr *wr, wr 64 drivers/infiniband/sw/siw/siw_verbs.h int siw_post_receive(struct ib_qp *base_qp, const struct ib_recv_wr *wr, wr 83 drivers/infiniband/sw/siw/siw_verbs.h int siw_post_srq_recv(struct ib_srq *base_srq, const struct ib_recv_wr *wr, wr 561 drivers/infiniband/ulp/ipoib/ipoib.h priv->tx_wr.wr.num_sge = nr_frags + off; wr 118 drivers/infiniband/ulp/ipoib/ipoib_cm.c struct ib_recv_wr *wr, wr 124 drivers/infiniband/ulp/ipoib/ipoib_cm.c wr->wr_id = id | IPOIB_OP_CM | IPOIB_OP_RECV; wr 129 drivers/infiniband/ulp/ipoib/ipoib_cm.c ret = ib_post_recv(rx->qp, wr, NULL); wr 329 drivers/infiniband/ulp/ipoib/ipoib_cm.c struct ib_recv_wr *wr, wr 342 drivers/infiniband/ulp/ipoib/ipoib_cm.c wr->next = NULL; wr 343 drivers/infiniband/ulp/ipoib/ipoib_cm.c wr->sg_list = sge; wr 344 drivers/infiniband/ulp/ipoib/ipoib_cm.c wr->num_sge = priv->cm.num_frags; wr 352 drivers/infiniband/ulp/ipoib/ipoib_cm.c struct ib_recv_wr wr; wr 369 drivers/infiniband/ulp/ipoib/ipoib_cm.c ipoib_cm_init_rx_wr(dev, &t->wr, t->sge); wr 391 drivers/infiniband/ulp/ipoib/ipoib_cm.c ret = ipoib_cm_post_receive_nonsrq(dev, rx, &t->wr, t->sge, i); wr 703 drivers/infiniband/ulp/ipoib/ipoib_cm.c priv->tx_wr.wr.wr_id = wr_id | IPOIB_OP_CM; wr 705 drivers/infiniband/ulp/ipoib/ipoib_cm.c return ib_post_send(tx->qp, &priv->tx_wr.wr, NULL); wr 551 drivers/infiniband/ulp/ipoib/ipoib_ib.c priv->tx_wr.wr.wr_id = wr_id; wr 559 drivers/infiniband/ulp/ipoib/ipoib_ib.c priv->tx_wr.wr.opcode = IB_WR_LSO; wr 561 drivers/infiniband/ulp/ipoib/ipoib_ib.c priv->tx_wr.wr.opcode = IB_WR_SEND; wr 563 drivers/infiniband/ulp/ipoib/ipoib_ib.c return ib_post_send(priv->qp, &priv->tx_wr.wr, NULL); wr 635 drivers/infiniband/ulp/ipoib/ipoib_ib.c priv->tx_wr.wr.send_flags |= IB_SEND_IP_CSUM; wr 637 drivers/infiniband/ulp/ipoib/ipoib_ib.c priv->tx_wr.wr.send_flags &= ~IB_SEND_IP_CSUM; wr 537 drivers/infiniband/ulp/ipoib/ipoib_main.c priv->tx_wr.wr.send_flags &= ~IB_SEND_IP_CSUM; wr 221 drivers/infiniband/ulp/ipoib/ipoib_verbs.c priv->tx_wr.wr.opcode = IB_WR_SEND; wr 222 drivers/infiniband/ulp/ipoib/ipoib_verbs.c priv->tx_wr.wr.sg_list = priv->tx_sge; wr 223 drivers/infiniband/ulp/ipoib/ipoib_verbs.c priv->tx_wr.wr.send_flags = IB_SEND_SIGNALED; wr 209 drivers/infiniband/ulp/iser/iscsi_iser.c tx_desc->reg_wr.wr.next = NULL; wr 390 drivers/infiniband/ulp/iser/iser_memory.c struct ib_reg_wr *wr = &tx_desc->reg_wr; wr 401 drivers/infiniband/ulp/iser/iser_memory.c iser_inv_rkey(&tx_desc->inv_wr, mr, cqe, &wr->wr); wr 413 drivers/infiniband/ulp/iser/iser_memory.c memset(wr, 0, sizeof(*wr)); wr 414 drivers/infiniband/ulp/iser/iser_memory.c wr->wr.next = &tx_desc->send_wr; wr 415 drivers/infiniband/ulp/iser/iser_memory.c wr->wr.opcode = IB_WR_REG_MR_INTEGRITY; wr 416 drivers/infiniband/ulp/iser/iser_memory.c wr->wr.wr_cqe = cqe; wr 417 drivers/infiniband/ulp/iser/iser_memory.c wr->wr.num_sge = 0; wr 418 drivers/infiniband/ulp/iser/iser_memory.c wr->wr.send_flags = 0; wr 419 drivers/infiniband/ulp/iser/iser_memory.c wr->mr = mr; wr 420 drivers/infiniband/ulp/iser/iser_memory.c wr->key = mr->rkey; wr 421 drivers/infiniband/ulp/iser/iser_memory.c wr->access = IB_ACCESS_LOCAL_WRITE | wr 446 drivers/infiniband/ulp/iser/iser_memory.c struct ib_reg_wr *wr = &tx_desc->reg_wr; wr 450 drivers/infiniband/ulp/iser/iser_memory.c iser_inv_rkey(&tx_desc->inv_wr, mr, cqe, &wr->wr); wr 461 drivers/infiniband/ulp/iser/iser_memory.c wr->wr.next = &tx_desc->send_wr; wr 462 drivers/infiniband/ulp/iser/iser_memory.c wr->wr.opcode = IB_WR_REG_MR; wr 463 drivers/infiniband/ulp/iser/iser_memory.c wr->wr.wr_cqe = cqe; wr 464 drivers/infiniband/ulp/iser/iser_memory.c wr->wr.send_flags = 0; wr 465 drivers/infiniband/ulp/iser/iser_memory.c wr->wr.num_sge = 0; wr 466 drivers/infiniband/ulp/iser/iser_memory.c wr->mr = mr; wr 467 drivers/infiniband/ulp/iser/iser_memory.c wr->key = mr->rkey; wr 468 drivers/infiniband/ulp/iser/iser_memory.c wr->access = IB_ACCESS_LOCAL_WRITE | wr 976 drivers/infiniband/ulp/iser/iser_verbs.c struct ib_recv_wr wr; wr 984 drivers/infiniband/ulp/iser/iser_verbs.c wr.wr_cqe = &desc->cqe; wr 985 drivers/infiniband/ulp/iser/iser_verbs.c wr.sg_list = &desc->sge; wr 986 drivers/infiniband/ulp/iser/iser_verbs.c wr.num_sge = 1; wr 987 drivers/infiniband/ulp/iser/iser_verbs.c wr.next = NULL; wr 990 drivers/infiniband/ulp/iser/iser_verbs.c ib_ret = ib_post_recv(ib_conn->qp, &wr, NULL); wr 1004 drivers/infiniband/ulp/iser/iser_verbs.c struct ib_recv_wr *wr; wr 1007 drivers/infiniband/ulp/iser/iser_verbs.c for (wr = ib_conn->rx_wr, i = 0; i < count; i++, wr++) { wr 1010 drivers/infiniband/ulp/iser/iser_verbs.c wr->wr_cqe = &rx_desc->cqe; wr 1011 drivers/infiniband/ulp/iser/iser_verbs.c wr->sg_list = &rx_desc->rx_sg; wr 1012 drivers/infiniband/ulp/iser/iser_verbs.c wr->num_sge = 1; wr 1013 drivers/infiniband/ulp/iser/iser_verbs.c wr->next = wr + 1; wr 1017 drivers/infiniband/ulp/iser/iser_verbs.c wr--; wr 1018 drivers/infiniband/ulp/iser/iser_verbs.c wr->next = NULL; /* mark end of work requests list */ wr 1040 drivers/infiniband/ulp/iser/iser_verbs.c struct ib_send_wr *wr = &tx_desc->send_wr; wr 1048 drivers/infiniband/ulp/iser/iser_verbs.c wr->next = NULL; wr 1049 drivers/infiniband/ulp/iser/iser_verbs.c wr->wr_cqe = &tx_desc->cqe; wr 1050 drivers/infiniband/ulp/iser/iser_verbs.c wr->sg_list = tx_desc->tx_sg; wr 1051 drivers/infiniband/ulp/iser/iser_verbs.c wr->num_sge = tx_desc->num_sge; wr 1052 drivers/infiniband/ulp/iser/iser_verbs.c wr->opcode = IB_WR_SEND; wr 1053 drivers/infiniband/ulp/iser/iser_verbs.c wr->send_flags = signal ? IB_SEND_SIGNALED : 0; wr 1057 drivers/infiniband/ulp/iser/iser_verbs.c else if (tx_desc->reg_wr.wr.next) wr 1058 drivers/infiniband/ulp/iser/iser_verbs.c first_wr = &tx_desc->reg_wr.wr; wr 1060 drivers/infiniband/ulp/iser/iser_verbs.c first_wr = wr; wr 1065 drivers/infiniband/ulp/iser/iser_verbs.c ib_ret, wr->opcode); wr 1229 drivers/infiniband/ulp/srp/ib_srp.c struct ib_send_wr wr = { wr 1237 drivers/infiniband/ulp/srp/ib_srp.c wr.wr_cqe = &req->reg_cqe; wr 1239 drivers/infiniband/ulp/srp/ib_srp.c return ib_post_send(ch->qp, &wr, NULL); wr 1527 drivers/infiniband/ulp/srp/ib_srp.c struct ib_reg_wr wr; wr 1573 drivers/infiniband/ulp/srp/ib_srp.c wr.wr.next = NULL; wr 1574 drivers/infiniband/ulp/srp/ib_srp.c wr.wr.opcode = IB_WR_REG_MR; wr 1575 drivers/infiniband/ulp/srp/ib_srp.c wr.wr.wr_cqe = &req->reg_cqe; wr 1576 drivers/infiniband/ulp/srp/ib_srp.c wr.wr.num_sge = 0; wr 1577 drivers/infiniband/ulp/srp/ib_srp.c wr.wr.send_flags = 0; wr 1578 drivers/infiniband/ulp/srp/ib_srp.c wr.mr = desc->mr; wr 1579 drivers/infiniband/ulp/srp/ib_srp.c wr.key = desc->mr->rkey; wr 1580 drivers/infiniband/ulp/srp/ib_srp.c wr.access = (IB_ACCESS_LOCAL_WRITE | wr 1590 drivers/infiniband/ulp/srp/ib_srp.c err = ib_post_send(ch->qp, &wr.wr, NULL); wr 2074 drivers/infiniband/ulp/srp/ib_srp.c struct ib_send_wr wr; wr 2085 drivers/infiniband/ulp/srp/ib_srp.c wr.next = NULL; wr 2086 drivers/infiniband/ulp/srp/ib_srp.c wr.wr_cqe = &iu->cqe; wr 2087 drivers/infiniband/ulp/srp/ib_srp.c wr.sg_list = &iu->sge[0]; wr 2088 drivers/infiniband/ulp/srp/ib_srp.c wr.num_sge = iu->num_sge; wr 2089 drivers/infiniband/ulp/srp/ib_srp.c wr.opcode = IB_WR_SEND; wr 2090 drivers/infiniband/ulp/srp/ib_srp.c wr.send_flags = IB_SEND_SIGNALED; wr 2092 drivers/infiniband/ulp/srp/ib_srp.c return ib_post_send(ch->qp, &wr, NULL); wr 2098 drivers/infiniband/ulp/srp/ib_srp.c struct ib_recv_wr wr; wr 2107 drivers/infiniband/ulp/srp/ib_srp.c wr.next = NULL; wr 2108 drivers/infiniband/ulp/srp/ib_srp.c wr.wr_cqe = &iu->cqe; wr 2109 drivers/infiniband/ulp/srp/ib_srp.c wr.sg_list = &list; wr 2110 drivers/infiniband/ulp/srp/ib_srp.c wr.num_sge = 1; wr 2112 drivers/infiniband/ulp/srp/ib_srp.c return ib_post_recv(ch->qp, &wr, NULL); wr 825 drivers/infiniband/ulp/srpt/ib_srpt.c struct ib_recv_wr wr; wr 833 drivers/infiniband/ulp/srpt/ib_srpt.c wr.wr_cqe = &ioctx->ioctx.cqe; wr 834 drivers/infiniband/ulp/srpt/ib_srpt.c wr.next = NULL; wr 835 drivers/infiniband/ulp/srpt/ib_srpt.c wr.sg_list = &list; wr 836 drivers/infiniband/ulp/srpt/ib_srpt.c wr.num_sge = 1; wr 839 drivers/infiniband/ulp/srpt/ib_srpt.c return ib_post_srq_recv(sdev->srq, &wr, NULL); wr 841 drivers/infiniband/ulp/srpt/ib_srpt.c return ib_post_recv(ch->qp, &wr, NULL); wr 855 drivers/infiniband/ulp/srpt/ib_srpt.c struct ib_rdma_wr wr = { wr 856 drivers/infiniband/ulp/srpt/ib_srpt.c .wr = { wr 867 drivers/infiniband/ulp/srpt/ib_srpt.c return ib_post_send(ch->qp, &wr.wr, NULL); wr 732 drivers/irqchip/irq-gic-v3-its.c u64 wr = its_cmd_ptr_to_offset(its, its->cmd_write); wr 734 drivers/irqchip/irq-gic-v3-its.c writel_relaxed(wr, its->base + GITS_CWRITER); wr 752 drivers/md/dm-integrity.c static void rw_section_mac(struct dm_integrity_c *ic, unsigned section, bool wr) wr 765 drivers/md/dm-integrity.c if (likely(wr)) wr 224 drivers/media/common/videobuf2/videobuf2-dma-contig.c struct scatterlist *rd, *wr; wr 244 drivers/media/common/videobuf2/videobuf2-dma-contig.c wr = sgt->sgl; wr 246 drivers/media/common/videobuf2/videobuf2-dma-contig.c sg_set_page(wr, sg_page(rd), rd->length, rd->offset); wr 248 drivers/media/common/videobuf2/videobuf2-dma-contig.c wr = sg_next(wr); wr 368 drivers/media/common/videobuf2/videobuf2-dma-sg.c struct scatterlist *rd, *wr; wr 388 drivers/media/common/videobuf2/videobuf2-dma-sg.c wr = sgt->sgl; wr 390 drivers/media/common/videobuf2/videobuf2-dma-sg.c sg_set_page(wr, sg_page(rd), rd->length, rd->offset); wr 392 drivers/media/common/videobuf2/videobuf2-dma-sg.c wr = sg_next(wr); wr 148 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_LOCK1_MASK, DIB3000MB_LOCK1_SEARCH_4); wr 153 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_FFT, DIB3000_TRANSMISSION_MODE_2K); wr 157 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_FFT, DIB3000_TRANSMISSION_MODE_8K); wr 169 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_GUARD_TIME, DIB3000_GUARD_TIME_1_32); wr 173 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_GUARD_TIME, DIB3000_GUARD_TIME_1_16); wr 177 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_GUARD_TIME, DIB3000_GUARD_TIME_1_8); wr 181 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_GUARD_TIME, DIB3000_GUARD_TIME_1_4); wr 193 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_DDS_INV, DIB3000_DDS_INVERSION_OFF); wr 200 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_DDS_INV, DIB3000_DDS_INVERSION_ON); wr 209 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_QAM, DIB3000_CONSTELLATION_QPSK); wr 213 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_QAM, DIB3000_CONSTELLATION_16QAM); wr 217 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_QAM, DIB3000_CONSTELLATION_64QAM); wr 230 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_VIT_ALPHA, DIB3000_ALPHA_1); wr 234 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_VIT_ALPHA, DIB3000_ALPHA_2); wr 238 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_VIT_ALPHA, DIB3000_ALPHA_4); wr 248 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_VIT_HRCH, DIB3000_HRCH_OFF); wr 249 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_VIT_HP, DIB3000_SELECT_HP); wr 252 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_VIT_HRCH, DIB3000_HRCH_ON); wr 253 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_VIT_HP, DIB3000_SELECT_LP); wr 259 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_VIT_CODE_RATE, DIB3000_FEC_1_2); wr 263 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_VIT_CODE_RATE, DIB3000_FEC_2_3); wr 267 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_VIT_CODE_RATE, DIB3000_FEC_3_4); wr 271 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_VIT_CODE_RATE, DIB3000_FEC_5_6); wr 275 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_VIT_CODE_RATE, DIB3000_FEC_7_8); wr 294 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_SEQ, seq); wr 296 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_ISI, seq ? DIB3000MB_ISI_INHIBIT : DIB3000MB_ISI_ACTIVATE); wr 300 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_SYNC_IMPROVEMENT, DIB3000MB_SYNC_IMPROVE_2K_1_8); wr 302 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_SYNC_IMPROVEMENT, DIB3000MB_SYNC_IMPROVE_DEFAULT); wr 305 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_UNK_121, DIB3000MB_UNK_121_2K); wr 307 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_UNK_121, DIB3000MB_UNK_121_DEFAULT); wr 310 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_MOBILE_ALGO, DIB3000MB_MOBILE_ALGO_OFF); wr 311 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_MOBILE_MODE_QAM, DIB3000MB_MOBILE_MODE_QAM_OFF); wr 312 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_MOBILE_MODE, DIB3000MB_MOBILE_MODE_OFF); wr 316 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_ISI, DIB3000MB_ISI_ACTIVATE); wr 318 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_RESTART, DIB3000MB_RESTART_AGC + DIB3000MB_RESTART_CTRL); wr 319 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_RESTART, DIB3000MB_RESTART_OFF); wr 335 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_ISI, DIB3000MB_ISI_INHIBIT); wr 337 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_RESTART, DIB3000MB_RESTART_AUTO_SEARCH); wr 338 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_RESTART, DIB3000MB_RESTART_OFF); wr 357 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_RESTART, DIB3000MB_RESTART_CTRL); wr 358 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_RESTART, DIB3000MB_RESTART_OFF); wr 369 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_POWER_CONTROL, DIB3000MB_POWER_UP); wr 371 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_RESTART, DIB3000MB_RESTART_AGC); wr 373 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_RESET_DEVICE, DIB3000MB_RESET_DEVICE); wr 374 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_RESET_DEVICE, DIB3000MB_RESET_DEVICE_RST); wr 376 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_CLOCK, DIB3000MB_CLOCK_DEFAULT); wr 378 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_ELECT_OUT_MODE, DIB3000MB_ELECT_OUT_MODE_ON); wr 380 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_DDS_FREQ_MSB, DIB3000MB_DDS_FREQ_MSB); wr 381 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_DDS_FREQ_LSB, DIB3000MB_DDS_FREQ_LSB); wr 390 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_PHASE_NOISE, DIB3000MB_PHASE_NOISE_DEFAULT); wr 398 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_LOCK0_MASK, DIB3000MB_LOCK0_DEFAULT); wr 399 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_LOCK1_MASK, DIB3000MB_LOCK1_SEARCH_4); wr 400 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_LOCK2_MASK, DIB3000MB_LOCK2_DEFAULT); wr 401 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_SEQ, dib3000_seq[1][1][1]); wr 405 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_UNK_68, DIB3000MB_UNK_68); wr 406 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_UNK_69, DIB3000MB_UNK_69); wr 407 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_UNK_71, DIB3000MB_UNK_71); wr 408 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_UNK_77, DIB3000MB_UNK_77); wr 409 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_UNK_78, DIB3000MB_UNK_78); wr 410 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_ISI, DIB3000MB_ISI_INHIBIT); wr 411 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_UNK_92, DIB3000MB_UNK_92); wr 412 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_UNK_96, DIB3000MB_UNK_96); wr 413 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_UNK_97, DIB3000MB_UNK_97); wr 414 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_UNK_106, DIB3000MB_UNK_106); wr 415 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_UNK_107, DIB3000MB_UNK_107); wr 416 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_UNK_108, DIB3000MB_UNK_108); wr 417 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_UNK_122, DIB3000MB_UNK_122); wr 418 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_MOBILE_MODE_QAM, DIB3000MB_MOBILE_MODE_QAM_OFF); wr 419 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_BERLEN, DIB3000MB_BERLEN_DEFAULT); wr 423 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_MOBILE_ALGO, DIB3000MB_MOBILE_ALGO_ON); wr 424 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_MULTI_DEMOD_MSB, DIB3000MB_MULTI_DEMOD_MSB); wr 425 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_MULTI_DEMOD_LSB, DIB3000MB_MULTI_DEMOD_LSB); wr 427 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_OUTPUT_MODE, DIB3000MB_OUTPUT_MODE_SLAVE); wr 429 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_FIFO_142, DIB3000MB_FIFO_142); wr 430 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_MPEG2_OUT_MODE, DIB3000MB_MPEG2_OUT_MODE_188); wr 431 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_PID_PARSE, DIB3000MB_PID_PARSE_ACTIVATE); wr 432 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_FIFO, DIB3000MB_FIFO_INHIBIT); wr 433 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_FIFO_146, DIB3000MB_FIFO_146); wr 434 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_FIFO_147, DIB3000MB_FIFO_147); wr 436 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_DATA_IN_DIVERSITY, DIB3000MB_DATA_DIVERSITY_IN_OFF); wr 674 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_POWER_CONTROL, DIB3000MB_POWER_DOWN); wr 705 drivers/media/dvb-frontends/dib3000mb.c wr(index+DIB3000MB_REG_FIRST_PID,pid); wr 715 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_FIFO, DIB3000MB_FIFO_ACTIVATE); wr 717 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_FIFO, DIB3000MB_FIFO_INHIBIT); wr 726 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_PID_PARSE,onoff); wr 734 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_TUNER, DIB3000_TUNER_WRITE_ENABLE(pll_addr)); wr 736 drivers/media/dvb-frontends/dib3000mb.c wr(DIB3000MB_REG_TUNER, DIB3000_TUNER_WRITE_DISABLE(pll_addr)); wr 23 drivers/media/dvb-frontends/dib3000mb_priv.h wr(a[i],v[i]); \ wr 26 drivers/media/dvb-frontends/dib3000mb_priv.h #define set_or(reg,val) wr(reg,rd(reg) | val) wr 28 drivers/media/dvb-frontends/dib3000mb_priv.h #define set_and(reg,val) wr(reg,rd(reg) & val) wr 95 drivers/media/dvb-frontends/lgdt330x.c u8 wr[] = { reg }; wr 100 drivers/media/dvb-frontends/lgdt330x.c .buf = wr, wr 61 drivers/media/pci/solo6x10/solo6x10-i2c.c static void solo_i2c_flush(struct solo_dev *solo_dev, int wr) wr 70 drivers/media/pci/solo6x10/solo6x10-i2c.c if (wr) { wr 28 drivers/media/pci/solo6x10/solo6x10-p2m.c int solo_p2m_dma(struct solo_dev *solo_dev, int wr, wr 41 drivers/media/pci/solo6x10/solo6x10-p2m.c wr ? PCI_DMA_TODEVICE : PCI_DMA_FROMDEVICE); wr 45 drivers/media/pci/solo6x10/solo6x10-p2m.c ret = solo_p2m_dma_t(solo_dev, wr, dma_addr, ext_addr, size, wr 49 drivers/media/pci/solo6x10/solo6x10-p2m.c wr ? PCI_DMA_TODEVICE : PCI_DMA_FROMDEVICE); wr 124 drivers/media/pci/solo6x10/solo6x10-p2m.c void solo_p2m_fill_desc(struct solo_p2m_desc *desc, int wr, wr 133 drivers/media/pci/solo6x10/solo6x10-p2m.c (wr ? SOLO_P2M_WRITE : 0) | SOLO_P2M_TRANS_ON; wr 145 drivers/media/pci/solo6x10/solo6x10-p2m.c int solo_p2m_dma_t(struct solo_dev *solo_dev, int wr, wr 151 drivers/media/pci/solo6x10/solo6x10-p2m.c solo_p2m_fill_desc(&desc[1], wr, dma_addr, ext_addr, size, repeat, wr 341 drivers/media/pci/solo6x10/solo6x10.h int solo_p2m_dma_t(struct solo_dev *solo_dev, int wr, wr 344 drivers/media/pci/solo6x10/solo6x10.h int solo_p2m_dma(struct solo_dev *solo_dev, int wr, wr 347 drivers/media/pci/solo6x10/solo6x10.h void solo_p2m_fill_desc(struct solo_p2m_desc *desc, int wr, wr 162 drivers/media/radio/radio-tea5764.c struct tea5764_write_regs wr; wr 167 drivers/media/radio/radio-tea5764.c .len = sizeof(wr), wr 168 drivers/media/radio/radio-tea5764.c .buf = (void *)&wr wr 171 drivers/media/radio/radio-tea5764.c wr.intreg = r->intreg & 0xff; wr 172 drivers/media/radio/radio-tea5764.c wr.frqset = __cpu_to_be16(r->frqset); wr 173 drivers/media/radio/radio-tea5764.c wr.tnctrl = __cpu_to_be16(r->tnctrl); wr 174 drivers/media/radio/radio-tea5764.c wr.testreg = __cpu_to_be16(r->testreg); wr 175 drivers/media/radio/radio-tea5764.c wr.rdsctrl = __cpu_to_be16(r->rdsctrl); wr 176 drivers/media/radio/radio-tea5764.c wr.rdsbbl = __cpu_to_be16(r->rdsbbl); wr 394 drivers/mtd/nand/raw/mpc5121_nfc.c u8 *buffer, uint size, int wr) wr 431 drivers/mtd/nand/raw/mpc5121_nfc.c if (wr) wr 446 drivers/mtd/nand/raw/mpc5121_nfc.c int wr) wr 460 drivers/mtd/nand/raw/mpc5121_nfc.c mpc5121_nfc_copy_spare(mtd, c, buf, len, wr); wr 471 drivers/mtd/nand/raw/mpc5121_nfc.c if (wr) wr 480 drivers/mtd/nand/raw/mpc5121_nfc.c mpc5121_nfc_buf_copy(mtd, buf, len, wr); wr 1171 drivers/net/can/grcan.c u32 wr, rd, startrd; wr 1180 drivers/net/can/grcan.c wr = grcan_read_reg(®s->rxwr); wr 1181 drivers/net/can/grcan.c if (rd == wr) wr 257 drivers/net/ethernet/apm/xgene/xgene_enet_hw.c void __iomem *addr, *wr, *cmd, *cmd_done; wr 270 drivers/net/ethernet/apm/xgene/xgene_enet_hw.c wr = pdata->mcx_mac_addr + MAC_WRITE_REG_OFFSET; wr 276 drivers/net/ethernet/apm/xgene/xgene_enet_hw.c iowrite32(wr_data, wr); wr 39 drivers/net/ethernet/apm/xgene/xgene_enet_xgmac.c static bool xgene_enet_wr_indirect(void __iomem *addr, void __iomem *wr, wr 47 drivers/net/ethernet/apm/xgene/xgene_enet_xgmac.c iowrite32(wr_data, wr); wr 65 drivers/net/ethernet/apm/xgene/xgene_enet_xgmac.c void __iomem *addr, *wr, *cmd, *cmd_done; wr 68 drivers/net/ethernet/apm/xgene/xgene_enet_xgmac.c wr = pdata->pcs_addr + PCS_WRITE_REG_OFFSET; wr 72 drivers/net/ethernet/apm/xgene/xgene_enet_xgmac.c if (!xgene_enet_wr_indirect(addr, wr, cmd, cmd_done, wr_addr, wr_data)) wr 475 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_FORWARD)); wr 498 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_FORWARD)); wr 520 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_FORWARD)); wr 539 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c greq->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_FORWARD)); wr 913 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_FORWARD)); wr 556 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_FORWARD)); wr 878 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c rpl->wr.wr_hi = wr 880 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c rpl->wr.wr_lo = htonl(V_WR_TID(tid)); wr 1100 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_FORWARD)); wr 100 drivers/net/ethernet/chelsio/cxgb3/l2t.c req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_FORWARD)); wr 1226 drivers/net/ethernet/chelsio/cxgb3/sge.c cpl->wr.wr_hi = htonl(V_WR_BCNTLFLT(skb->len & 7) | wr 1230 drivers/net/ethernet/chelsio/cxgb3/sge.c cpl->wr.wr_lo = htonl(V_WR_LEN(flits) | V_WR_GEN(gen) | wr 300 drivers/net/ethernet/chelsio/cxgb3/t3_cpl.h # define WR_HDR struct work_request_hdr wr wr 1751 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h void t4_mk_filtdelwr(unsigned int ftid, struct fw_filter_wr *wr, int qid); wr 339 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c u32 obq_wr_t4[2 * CIM_NUM_OBQ], *wr; wr 352 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c wr = obq_wr_t4; wr 356 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c wr = obq_wr_t5; wr 372 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c for ( ; i < CIM_NUM_IBQ + cim_num_obq; i++, p += 4, wr += 2) wr 375 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c QUERDADDR_G(p[0]) & 0x3fff, wr[0] - base[i], wr 1459 drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c struct work_request_hdr *wr; wr 1487 drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c wrlen = roundup(sizeof(*wr) + (sizeof(*req) + sizeof(*aligner)) wr 1497 drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c wr = (struct work_request_hdr *)req; wr 1498 drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c wr++; wr 1499 drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c req = (struct cpl_set_tcb_field *)wr; wr 58 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h (w)->wr.wr_hi = htonl(FW_WR_OP_V(FW_TP_WR) | \ wr 59 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h FW_WR_IMMDLEN_V(sizeof(*w) - sizeof(w->wr))); \ wr 60 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h (w)->wr.wr_mid = htonl(FW_WR_LEN16_V(DIV_ROUND_UP(sizeof(*w), 16)) | \ wr 62 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h (w)->wr.wr_lo = cpu_to_be64(0); \ wr 71 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h (w)->wr.wr_hi = htonl(FW_WR_OP_V(FW_ULPTX_WR) | \ wr 73 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h (w)->wr.wr_mid = htonl(FW_WR_LEN16_V(DIV_ROUND_UP(wrlen, 16)) | \ wr 75 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h (w)->wr.wr_lo = cpu_to_be64(0); \ wr 1004 drivers/net/ethernet/chelsio/cxgb4/sge.c u64 *wr = (u64 *)&q->desc[index]; wr 1008 drivers/net/ethernet/chelsio/cxgb4/sge.c wr); wr 1372 drivers/net/ethernet/chelsio/cxgb4/sge.c struct fw_eth_tx_pkt_wr *wr; wr 1481 drivers/net/ethernet/chelsio/cxgb4/sge.c wr = (void *)&q->q.desc[q->q.pidx]; wr 1482 drivers/net/ethernet/chelsio/cxgb4/sge.c wr->equiq_to_len16 = htonl(wr_mid); wr 1483 drivers/net/ethernet/chelsio/cxgb4/sge.c wr->r3 = cpu_to_be64(0); wr 1484 drivers/net/ethernet/chelsio/cxgb4/sge.c end = (u64 *)wr + flits; wr 1489 drivers/net/ethernet/chelsio/cxgb4/sge.c struct cpl_tx_pkt_lso_core *lso = (void *)(wr + 1); wr 1493 drivers/net/ethernet/chelsio/cxgb4/sge.c struct cpl_tx_tnl_lso *tnl_lso = (void *)(wr + 1); wr 1500 drivers/net/ethernet/chelsio/cxgb4/sge.c wr->op_immdlen = htonl(FW_WR_OP_V(FW_ETH_TX_PKT_WR) | wr 1564 drivers/net/ethernet/chelsio/cxgb4/sge.c wr->op_immdlen = htonl(FW_WR_OP_V(op) | wr 1566 drivers/net/ethernet/chelsio/cxgb4/sge.c cpl = (void *)(wr + 1); wr 1709 drivers/net/ethernet/chelsio/cxgb4/sge.c struct fw_eth_tx_pkt_vm_wr *wr; wr 1718 drivers/net/ethernet/chelsio/cxgb4/sge.c const size_t fw_hdr_copy_len = sizeof(wr->ethmacdst) + wr 1719 drivers/net/ethernet/chelsio/cxgb4/sge.c sizeof(wr->ethmacsrc) + wr 1720 drivers/net/ethernet/chelsio/cxgb4/sge.c sizeof(wr->ethtype) + wr 1721 drivers/net/ethernet/chelsio/cxgb4/sge.c sizeof(wr->vlantci); wr 1801 drivers/net/ethernet/chelsio/cxgb4/sge.c wr = (void *)&txq->q.desc[txq->q.pidx]; wr 1802 drivers/net/ethernet/chelsio/cxgb4/sge.c wr->equiq_to_len16 = cpu_to_be32(wr_mid); wr 1803 drivers/net/ethernet/chelsio/cxgb4/sge.c wr->r3[0] = cpu_to_be32(0); wr 1804 drivers/net/ethernet/chelsio/cxgb4/sge.c wr->r3[1] = cpu_to_be32(0); wr 1805 drivers/net/ethernet/chelsio/cxgb4/sge.c skb_copy_from_linear_data(skb, (void *)wr->ethmacdst, fw_hdr_copy_len); wr 1806 drivers/net/ethernet/chelsio/cxgb4/sge.c end = (u64 *)wr + flits; wr 1814 drivers/net/ethernet/chelsio/cxgb4/sge.c struct cpl_tx_pkt_lso_core *lso = (void *)(wr + 1); wr 1819 drivers/net/ethernet/chelsio/cxgb4/sge.c wr->op_immdlen = wr 1861 drivers/net/ethernet/chelsio/cxgb4/sge.c wr->op_immdlen = wr 1868 drivers/net/ethernet/chelsio/cxgb4/sge.c cpl = (void *)(wr + 1); wr 2033 drivers/net/ethernet/chelsio/cxgb4/sge.c static void ctrlq_check_stop(struct sge_ctrl_txq *q, struct fw_wr_hdr *wr) wr 2037 drivers/net/ethernet/chelsio/cxgb4/sge.c wr->lo |= htonl(FW_WR_EQUEQ_F | FW_WR_EQUIQ_F); wr 2054 drivers/net/ethernet/chelsio/cxgb4/sge.c struct fw_wr_hdr *wr; wr 2072 drivers/net/ethernet/chelsio/cxgb4/sge.c wr = (struct fw_wr_hdr *)&q->q.desc[q->q.pidx]; wr 2073 drivers/net/ethernet/chelsio/cxgb4/sge.c cxgb4_inline_tx_skb(skb, &q->q, wr); wr 2077 drivers/net/ethernet/chelsio/cxgb4/sge.c ctrlq_check_stop(q, wr); wr 2103 drivers/net/ethernet/chelsio/cxgb4/sge.c struct fw_wr_hdr *wr; wr 2110 drivers/net/ethernet/chelsio/cxgb4/sge.c wr = (struct fw_wr_hdr *)&q->q.desc[q->q.pidx]; wr 2114 drivers/net/ethernet/chelsio/cxgb4/sge.c cxgb4_inline_tx_skb(skb, &q->q, wr); wr 2120 drivers/net/ethernet/chelsio/cxgb4/sge.c ctrlq_check_stop(q, wr); wr 2221 drivers/net/ethernet/chelsio/cxgb4/sge.c static void ofldtxq_stop(struct sge_uld_txq *q, struct fw_wr_hdr *wr) wr 2223 drivers/net/ethernet/chelsio/cxgb4/sge.c wr->lo |= htonl(FW_WR_EQUEQ_F | FW_WR_EQUIQ_F); wr 6502 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c void t4_mk_filtdelwr(unsigned int ftid, struct fw_filter_wr *wr, int qid) wr 6504 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c memset(wr, 0, sizeof(*wr)); wr 6505 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c wr->op_pkd = cpu_to_be32(FW_WR_OP_V(FW_FILTER_WR)); wr 6506 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c wr->len16_pkd = cpu_to_be32(FW_WR_LEN16_V(sizeof(*wr) / 16)); wr 6507 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c wr->tid_to_iq = cpu_to_be32(FW_FILTER_WR_TID_V(ftid) | wr 6509 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c wr->del_filter_to_l2tix = cpu_to_be32(FW_FILTER_WR_DEL_FILTER_F); wr 6511 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c wr->rx_chan_rx_rpl_iq = wr 249 drivers/net/ethernet/chelsio/cxgb4/t4_msg.h #define WR_HDR struct work_request_hdr wr wr 3571 drivers/net/ethernet/chelsio/cxgb4/t4fw_api.h } wr; wr 1166 drivers/net/ethernet/chelsio/cxgb4vf/sge.c struct fw_eth_tx_pkt_vm_wr *wr; wr 1170 drivers/net/ethernet/chelsio/cxgb4vf/sge.c const size_t fw_hdr_copy_len = (sizeof(wr->ethmacdst) + wr 1171 drivers/net/ethernet/chelsio/cxgb4vf/sge.c sizeof(wr->ethmacsrc) + wr 1172 drivers/net/ethernet/chelsio/cxgb4vf/sge.c sizeof(wr->ethtype) + wr 1173 drivers/net/ethernet/chelsio/cxgb4vf/sge.c sizeof(wr->vlantci)); wr 1266 drivers/net/ethernet/chelsio/cxgb4vf/sge.c wr = (void *)&txq->q.desc[txq->q.pidx]; wr 1267 drivers/net/ethernet/chelsio/cxgb4vf/sge.c wr->equiq_to_len16 = cpu_to_be32(wr_mid); wr 1268 drivers/net/ethernet/chelsio/cxgb4vf/sge.c wr->r3[0] = cpu_to_be32(0); wr 1269 drivers/net/ethernet/chelsio/cxgb4vf/sge.c wr->r3[1] = cpu_to_be32(0); wr 1270 drivers/net/ethernet/chelsio/cxgb4vf/sge.c skb_copy_from_linear_data(skb, (void *)wr->ethmacdst, fw_hdr_copy_len); wr 1271 drivers/net/ethernet/chelsio/cxgb4vf/sge.c end = (u64 *)wr + flits; wr 1280 drivers/net/ethernet/chelsio/cxgb4vf/sge.c struct cpl_tx_pkt_lso_core *lso = (void *)(wr + 1); wr 1285 drivers/net/ethernet/chelsio/cxgb4vf/sge.c wr->op_immdlen = wr 1328 drivers/net/ethernet/chelsio/cxgb4vf/sge.c wr->op_immdlen = wr 1336 drivers/net/ethernet/chelsio/cxgb4vf/sge.c cpl = (void *)(wr + 1); wr 537 drivers/net/ethernet/netronome/nfp/bpf/jit.c __emit_lcsr(struct nfp_prog *nfp_prog, u16 areg, u16 breg, bool wr, u16 addr, wr 545 drivers/net/ethernet/netronome/nfp/bpf/jit.c FIELD_PREP(OP_LCSR_WRITE, wr) | wr 5892 drivers/net/ethernet/sun/niu.c u64 rd, wr, val; wr 5896 drivers/net/ethernet/sun/niu.c wr = nr64_ipp(IPP_DFIFO_WR_PTR); wr 5898 drivers/net/ethernet/sun/niu.c while (--limit >= 0 && (rd != wr)) { wr 5900 drivers/net/ethernet/sun/niu.c wr = nr64_ipp(IPP_DFIFO_WR_PTR); wr 5903 drivers/net/ethernet/sun/niu.c (rd != 0 && wr != 1)) { wr 444 drivers/net/hamradio/hdlcdrv.c s->hdlcrx.hbuf.rd = s->hdlcrx.hbuf.wr = 0; wr 448 drivers/net/hamradio/hdlcdrv.c s->hdlctx.hbuf.rd = s->hdlctx.hbuf.wr = 0; wr 576 drivers/net/hamradio/hdlcdrv.c if (s->bitbuf_channel.rd == s->bitbuf_channel.wr) wr 589 drivers/net/hamradio/hdlcdrv.c if (s->bitbuf_hdlc.rd == s->bitbuf_hdlc.wr) wr 641 drivers/net/hamradio/hdlcdrv.c s->hdlcrx.hbuf.rd = s->hdlcrx.hbuf.wr = 0; wr 646 drivers/net/hamradio/hdlcdrv.c s->hdlctx.hbuf.rd = s->hdlctx.hbuf.wr = 0; wr 656 drivers/net/hamradio/hdlcdrv.c s->bitbuf_channel.rd = s->bitbuf_channel.wr = 0; wr 659 drivers/net/hamradio/hdlcdrv.c s->bitbuf_hdlc.rd = s->bitbuf_hdlc.wr = 0; wr 722 drivers/net/hamradio/scc.c wr(scc,R12,tc & 255); /* brg rate LOW */ wr 723 drivers/net/hamradio/scc.c wr(scc,R13,tc >> 8); /* brg rate HIGH */ wr 743 drivers/net/hamradio/scc.c wr(scc, R14, BRSRC); /* BRG source = PCLK */ wr 800 drivers/net/hamradio/scc.c wr(scc,R4,X1CLK|SDLC); /* *1 clock, SDLC mode */ wr 801 drivers/net/hamradio/scc.c wr(scc,R1,0); /* no W/REQ operation */ wr 802 drivers/net/hamradio/scc.c wr(scc,R3,Rx8|RxCRC_ENAB); /* RX 8 bits/char, CRC, disabled */ wr 803 drivers/net/hamradio/scc.c wr(scc,R5,Tx8|DTR|TxCRC_ENAB); /* TX 8 bits/char, disabled, DTR */ wr 804 drivers/net/hamradio/scc.c wr(scc,R6,0); /* SDLC address zero (not used) */ wr 805 drivers/net/hamradio/scc.c wr(scc,R7,FLAG); /* SDLC flag value */ wr 806 drivers/net/hamradio/scc.c wr(scc,R9,VIS); /* vector includes status */ wr 807 drivers/net/hamradio/scc.c wr(scc,R10,(scc->modem.nrz? NRZ : NRZI)|CRCPS|ABUNDER); /* abort on underrun, preset CRC generator, NRZ(I) */ wr 808 drivers/net/hamradio/scc.c wr(scc,R14, 0); wr 840 drivers/net/hamradio/scc.c wr(scc, R11, RCDPLL|TCDPLL|TRxCOI|TRxCDP); wr 845 drivers/net/hamradio/scc.c wr(scc, R11, ((scc->brand & BAYCOM)? TRxCDP : TRxCBR) | RCDPLL|TCRTxCP|TRxCOI); wr 850 drivers/net/hamradio/scc.c wr(scc, R11, (scc->brand & BAYCOM)? RCTRxCP|TCRTxCP : RCRTxCP|TCTRxCP); wr 861 drivers/net/hamradio/scc.c wr(scc,R7,AUTOEOM); wr 872 drivers/net/hamradio/scc.c wr(scc,R15, BRKIE|TxUIE|(scc->kiss.softdcd? SYNCIE:DCDIE)); wr 930 drivers/net/hamradio/scc.c wr(scc, R11, RCDPLL|TCBR|TRxCOI|TRxCBR); wr 946 drivers/net/hamradio/scc.c wr(scc, R11, RCDPLL|TCDPLL|TRxCOI|TRxCDP); wr 1410 drivers/net/hamradio/scc.c wr(scc, R6, 0); wr 1411 drivers/net/hamradio/scc.c wr(scc, R7, FLAG); wr 1436 drivers/net/hamradio/scc.c wr(scc, R6, 0); wr 1437 drivers/net/hamradio/scc.c wr(scc, R7, pattern); wr 1501 drivers/net/hamradio/scc.c wr(scc, R2, chip*16); /* interrupt vector */ wr 1502 drivers/net/hamradio/scc.c wr(scc, R9, VIS); /* vector includes status */ wr 1606 drivers/net/hamradio/scc.c wr(scc,R1,0); /* disable interrupts */ wr 1607 drivers/net/hamradio/scc.c wr(scc,R3,0); wr 55 drivers/net/phy/mdio-xgene.c void __iomem *addr, *wr, *cmd, *cmd_done; wr 60 drivers/net/phy/mdio-xgene.c wr = pdata->mac_csr_addr + MAC_WRITE_REG_OFFSET; wr 66 drivers/net/phy/mdio-xgene.c iowrite32(data, wr); wr 12 drivers/net/wireless/mediatek/mt76/debugfs.c dev->bus->wr(dev, dev->debugfs_reg, val); wr 86 drivers/net/wireless/mediatek/mt76/mmio.c .wr = mt76_mmio_wr, wr 39 drivers/net/wireless/mediatek/mt76/mt76.h void (*wr)(struct mt76_dev *dev, u32 offset, u32 val); wr 536 drivers/net/wireless/mediatek/mt76/mt76.h #define __mt76_wr(dev, ...) (dev)->bus->wr((dev), __VA_ARGS__) wr 545 drivers/net/wireless/mediatek/mt76/mt76.h #define mt76_wr(dev, ...) (dev)->mt76.bus->wr(&((dev)->mt76), __VA_ARGS__) wr 59 drivers/net/wireless/mediatek/mt76/mt7603/core.c dev->bus_ops->wr(&dev->mt76, MT_MCU_PCIE_REMAP_2, base); wr 429 drivers/net/wireless/mediatek/mt76/mt7603/init.c dev->bus_ops->wr(mdev, addr, val); wr 523 drivers/net/wireless/mediatek/mt76/mt7603/init.c bus_ops->wr = mt7603_wr; wr 951 drivers/net/wireless/mediatek/mt76/usb.c .wr = mt76u_wr, wr 1135 drivers/nvme/host/rdma.c struct ib_send_wr wr = { wr 1144 drivers/nvme/host/rdma.c wr.wr_cqe = &req->reg_cqe; wr 1146 drivers/nvme/host/rdma.c return ib_post_send(queue->qp, &wr, NULL); wr 1246 drivers/nvme/host/rdma.c req->reg_wr.wr.opcode = IB_WR_REG_MR; wr 1247 drivers/nvme/host/rdma.c req->reg_wr.wr.wr_cqe = &req->reg_cqe; wr 1248 drivers/nvme/host/rdma.c req->reg_wr.wr.num_sge = 0; wr 1346 drivers/nvme/host/rdma.c struct ib_send_wr wr; wr 1353 drivers/nvme/host/rdma.c wr.next = NULL; wr 1354 drivers/nvme/host/rdma.c wr.wr_cqe = &qe->cqe; wr 1355 drivers/nvme/host/rdma.c wr.sg_list = sge; wr 1356 drivers/nvme/host/rdma.c wr.num_sge = num_sge; wr 1357 drivers/nvme/host/rdma.c wr.opcode = IB_WR_SEND; wr 1358 drivers/nvme/host/rdma.c wr.send_flags = IB_SEND_SIGNALED; wr 1361 drivers/nvme/host/rdma.c first->next = ≀ wr 1363 drivers/nvme/host/rdma.c first = ≀ wr 1376 drivers/nvme/host/rdma.c struct ib_recv_wr wr; wr 1386 drivers/nvme/host/rdma.c wr.next = NULL; wr 1387 drivers/nvme/host/rdma.c wr.wr_cqe = &qe->cqe; wr 1388 drivers/nvme/host/rdma.c wr.sg_list = &list; wr 1389 drivers/nvme/host/rdma.c wr.num_sge = 1; wr 1391 drivers/nvme/host/rdma.c ret = ib_post_recv(queue->qp, &wr, NULL); wr 1783 drivers/nvme/host/rdma.c req->mr ? &req->reg_wr.wr : NULL); wr 37 drivers/nvme/target/rdma.c struct ib_recv_wr wr; wr 300 drivers/nvme/target/rdma.c c->wr.wr_cqe = &c->cqe; wr 301 drivers/nvme/target/rdma.c c->wr.sg_list = c->sge; wr 302 drivers/nvme/target/rdma.c c->wr.num_sge = admin ? 1 : ndev->inline_page_count + 1; wr 465 drivers/nvme/target/rdma.c ret = ib_post_srq_recv(ndev->srq, &cmd->wr, NULL); wr 467 drivers/nvme/target/rdma.c ret = ib_post_recv(cmd->queue->cm_id->qp, &cmd->wr, NULL); wr 136 drivers/perf/fsl_imx8_ddr_perf.c IMX8_DDR_PMU_EVENT_ATTR(wr-xact-credit, 0x29), wr 345 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(wr-cmd-sent, 0x06), wr 352 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(wr-wra-cmd-sent, 0x0d), wr 354 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(in-wr-collision, 0x0f), wr 360 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(mcu-wr-request, 0x15), wr 366 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(mcu-wr-proceed-all, 0x1b), wr 367 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(mcu-wr-proceed-cancel, 0x1c), wr 438 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(pa-req-buf-alloc-wr, 0x03), wr 443 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(pa-cp-wr-req, 0x08), wr 446 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(ba-wr-req, 0x0b), wr 449 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(pa-wr-invalidate-req-issued-stashable, 0x12), wr 450 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(pa-wr-invalidate-req-issued-nonstashable, 0x13), wr 451 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(pa-wr-back-req-issued-stashable, 0x14), wr 452 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(pa-wr-back-req-issued-nonstashable, 0x15), wr 453 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(pa-ptl-wr-req, 0x16), wr 455 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(pa-wr-back-clean-data, 0x18), wr 456 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(pa-wr-back-cancelled-on-SS, 0x1b), wr 464 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(pa-rd-wr-invalid-snoop, 0x25), wr 465 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(pa-rd-wr-invalid-snoop-hit, 0x26), wr 480 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(pa-axi0-wr-req, 0x02), wr 482 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(pa-axi1-wr-req, 0x04), wr 485 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(ba-axi-wr-req, 0x09), wr 495 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(wr-req-recv, 0x04), wr 496 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(wr-req-recv-2, 0x05), wr 506 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(wr-req-sent-to-mcu, 0x0f), wr 509 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(wr-gack-recv, 0x12), wr 511 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(cancel-wr-gack, 0x14), wr 519 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(mcu-wr-table-full, 0x1c), wr 536 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(wr-sent, 0x05), wr 543 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(wr-wra-sent, 0x0c), wr 549 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(hprd-lprd-wr-req-vld, 0x12), wr 553 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(wr-req-vld, 0x16), wr 554 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(partial-wr-req-vld, 0x17), wr 556 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(wr-retry, 0x19), wr 573 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(wr-enq, 0x2a), wr 3933 drivers/scsi/csiostor/csio_hw.c csio_process_fwevtq_entry(struct csio_hw *hw, void *wr, uint32_t len, wr 3941 drivers/scsi/csiostor/csio_hw.c op = ((struct rss_header *) wr)->opcode; wr 3956 drivers/scsi/csiostor/csio_hw.c msg = (void *)((uintptr_t)wr + sizeof(__be64)); wr 131 drivers/scsi/csiostor/csio_isr.c csio_fwevt_intx_handler(struct csio_hw *hw, void *wr, uint32_t len, wr 146 drivers/scsi/csiostor/csio_isr.c csio_process_scsi_cmpl(struct csio_hw *hw, void *wr, uint32_t len, wr 155 drivers/scsi/csiostor/csio_isr.c ioreq = csio_scsi_cmpl_handler(hw, wr, len, flb, NULL, &scsiwr); wr 285 drivers/scsi/csiostor/csio_isr.c csio_scsi_intx_handler(struct csio_hw *hw, void *wr, uint32_t len, wr 1433 drivers/scsi/csiostor/csio_lnode.c csio_ln_mgmt_wr_handler(struct csio_hw *hw, void *wr, uint32_t len) wr 1440 drivers/scsi/csiostor/csio_lnode.c wr_cmd = (struct fw_fcoe_els_ct_wr *) wr; wr 1490 drivers/scsi/csiostor/csio_lnode.c struct fw_wr_hdr *wr; wr 1526 drivers/scsi/csiostor/csio_lnode.c wr = (struct fw_wr_hdr *) (cmd + 4); wr 1527 drivers/scsi/csiostor/csio_lnode.c if (FW_WR_OP_G(be32_to_cpu(wr->hi)) wr 1595 drivers/scsi/csiostor/csio_lnode.c FW_WR_OP_G(be32_to_cpu((wr->hi)))); wr 1599 drivers/scsi/csiostor/csio_lnode.c wr = (struct fw_wr_hdr *) (cmd); wr 1600 drivers/scsi/csiostor/csio_lnode.c if (FW_WR_OP_G(be32_to_cpu(wr->hi)) == FW_FCOE_ELS_CT_WR) { wr 1601 drivers/scsi/csiostor/csio_lnode.c csio_ln_mgmt_wr_handler(hw, wr, wr 1605 drivers/scsi/csiostor/csio_lnode.c FW_WR_OP_G(be32_to_cpu((wr->hi)))); wr 1685 drivers/scsi/csiostor/csio_lnode.c struct fw_fcoe_els_ct_wr *wr; wr 1688 drivers/scsi/csiostor/csio_lnode.c wr = (struct fw_fcoe_els_ct_wr *)fw_wr; wr 1689 drivers/scsi/csiostor/csio_lnode.c wr->op_immdlen = cpu_to_be32(FW_WR_OP_V(FW_FCOE_ELS_CT_WR) | wr 1693 drivers/scsi/csiostor/csio_lnode.c wr->flowid_len16 = cpu_to_be32(FW_WR_FLOWID_V(flow_id) | wr 1695 drivers/scsi/csiostor/csio_lnode.c wr->els_ct_type = sub_op; wr 1696 drivers/scsi/csiostor/csio_lnode.c wr->ctl_pri = 0; wr 1697 drivers/scsi/csiostor/csio_lnode.c wr->cp_en_class = 0; wr 1698 drivers/scsi/csiostor/csio_lnode.c wr->cookie = io_req->fw_handle; wr 1699 drivers/scsi/csiostor/csio_lnode.c wr->iqid = cpu_to_be16(csio_q_physiqid( wr 1701 drivers/scsi/csiostor/csio_lnode.c wr->fl_to_sp = FW_FCOE_ELS_CT_WR_SP(1); wr 1702 drivers/scsi/csiostor/csio_lnode.c wr->tmo_val = (uint8_t) io_req->tmo; wr 1704 drivers/scsi/csiostor/csio_lnode.c memcpy(wr->l_id, PORT_ID_PTR(port_id), 3); wr 1706 drivers/scsi/csiostor/csio_lnode.c memcpy(wr->r_id, PORT_ID_PTR(port_id), 3); wr 1709 drivers/scsi/csiostor/csio_lnode.c wr->rsp_dmalen = cpu_to_be32(io_req->dma_buf.len); wr 1710 drivers/scsi/csiostor/csio_lnode.c wr->rsp_dmaaddr = cpu_to_be64(io_req->dma_buf.paddr); wr 195 drivers/scsi/csiostor/csio_mb.c const u32 *params, u32 *val, bool wr, wr 207 drivers/scsi/csiostor/csio_mb.c (wr ? FW_CMD_WRITE_F : FW_CMD_READ_F) | wr 213 drivers/scsi/csiostor/csio_mb.c if (wr) { wr 301 drivers/scsi/csiostor/csio_mb.c bool wr, bool init, bool tgt, bool cofld, wr 307 drivers/scsi/csiostor/csio_mb.c CSIO_INIT_MBP(mbp, cmdp, tmo, hw, cbfn, wr ? 0 : 1); wr 311 drivers/scsi/csiostor/csio_mb.c (wr ? FW_CMD_WRITE_F : FW_CMD_READ_F)); wr 315 drivers/scsi/csiostor/csio_mb.c if (!wr) wr 343 drivers/scsi/csiostor/csio_mb.c u8 portid, bool wr, uint32_t fc, uint16_t fw_caps, wr 352 drivers/scsi/csiostor/csio_mb.c (wr ? FW_CMD_EXEC_F : FW_CMD_READ_F) | wr 354 drivers/scsi/csiostor/csio_mb.c if (!wr) { wr 205 drivers/scsi/csiostor/csio_scsi.c struct fw_scsi_cmd_wr *wr = (struct fw_scsi_cmd_wr *)addr; wr 209 drivers/scsi/csiostor/csio_scsi.c wr->op_immdlen = cpu_to_be32(FW_WR_OP_V(FW_SCSI_CMD_WR) | wr 211 drivers/scsi/csiostor/csio_scsi.c wr->flowid_len16 = cpu_to_be32(FW_WR_FLOWID_V(rn->flowid) | wr 215 drivers/scsi/csiostor/csio_scsi.c wr->cookie = (uintptr_t) req; wr 216 drivers/scsi/csiostor/csio_scsi.c wr->iqid = cpu_to_be16(csio_q_physiqid(hw, req->iq_idx)); wr 217 drivers/scsi/csiostor/csio_scsi.c wr->tmo_val = (uint8_t) req->tmo; wr 218 drivers/scsi/csiostor/csio_scsi.c wr->r3 = 0; wr 219 drivers/scsi/csiostor/csio_scsi.c memset(&wr->r5, 0, 8); wr 225 drivers/scsi/csiostor/csio_scsi.c wr->rsp_dmalen = cpu_to_be32(dma_buf->len); wr 226 drivers/scsi/csiostor/csio_scsi.c wr->rsp_dmaaddr = cpu_to_be64(dma_buf->paddr); wr 228 drivers/scsi/csiostor/csio_scsi.c wr->r6 = 0; wr 230 drivers/scsi/csiostor/csio_scsi.c wr->u.fcoe.ctl_pri = 0; wr 231 drivers/scsi/csiostor/csio_scsi.c wr->u.fcoe.cp_en_class = 0; wr 232 drivers/scsi/csiostor/csio_scsi.c wr->u.fcoe.r4_lo[0] = 0; wr 233 drivers/scsi/csiostor/csio_scsi.c wr->u.fcoe.r4_lo[1] = 0; wr 364 drivers/scsi/csiostor/csio_scsi.c struct fw_scsi_read_wr *wr = (struct fw_scsi_read_wr *)wrp; wr 370 drivers/scsi/csiostor/csio_scsi.c wr->op_immdlen = cpu_to_be32(FW_WR_OP_V(FW_SCSI_READ_WR) | wr 372 drivers/scsi/csiostor/csio_scsi.c wr->flowid_len16 = cpu_to_be32(FW_WR_FLOWID_V(rn->flowid) | wr 374 drivers/scsi/csiostor/csio_scsi.c wr->cookie = (uintptr_t)req; wr 375 drivers/scsi/csiostor/csio_scsi.c wr->iqid = cpu_to_be16(csio_q_physiqid(hw, req->iq_idx)); wr 376 drivers/scsi/csiostor/csio_scsi.c wr->tmo_val = (uint8_t)(req->tmo); wr 377 drivers/scsi/csiostor/csio_scsi.c wr->use_xfer_cnt = 1; wr 378 drivers/scsi/csiostor/csio_scsi.c wr->xfer_cnt = cpu_to_be32(scsi_bufflen(scmnd)); wr 379 drivers/scsi/csiostor/csio_scsi.c wr->ini_xfer_cnt = cpu_to_be32(scsi_bufflen(scmnd)); wr 384 drivers/scsi/csiostor/csio_scsi.c wr->rsp_dmalen = cpu_to_be32(dma_buf->len); wr 385 drivers/scsi/csiostor/csio_scsi.c wr->rsp_dmaaddr = cpu_to_be64(dma_buf->paddr); wr 387 drivers/scsi/csiostor/csio_scsi.c wr->r4 = 0; wr 389 drivers/scsi/csiostor/csio_scsi.c wr->u.fcoe.ctl_pri = 0; wr 390 drivers/scsi/csiostor/csio_scsi.c wr->u.fcoe.cp_en_class = 0; wr 391 drivers/scsi/csiostor/csio_scsi.c wr->u.fcoe.r3_lo[0] = 0; wr 392 drivers/scsi/csiostor/csio_scsi.c wr->u.fcoe.r3_lo[1] = 0; wr 417 drivers/scsi/csiostor/csio_scsi.c struct fw_scsi_write_wr *wr = (struct fw_scsi_write_wr *)wrp; wr 423 drivers/scsi/csiostor/csio_scsi.c wr->op_immdlen = cpu_to_be32(FW_WR_OP_V(FW_SCSI_WRITE_WR) | wr 425 drivers/scsi/csiostor/csio_scsi.c wr->flowid_len16 = cpu_to_be32(FW_WR_FLOWID_V(rn->flowid) | wr 427 drivers/scsi/csiostor/csio_scsi.c wr->cookie = (uintptr_t)req; wr 428 drivers/scsi/csiostor/csio_scsi.c wr->iqid = cpu_to_be16(csio_q_physiqid(hw, req->iq_idx)); wr 429 drivers/scsi/csiostor/csio_scsi.c wr->tmo_val = (uint8_t)(req->tmo); wr 430 drivers/scsi/csiostor/csio_scsi.c wr->use_xfer_cnt = 1; wr 431 drivers/scsi/csiostor/csio_scsi.c wr->xfer_cnt = cpu_to_be32(scsi_bufflen(scmnd)); wr 432 drivers/scsi/csiostor/csio_scsi.c wr->ini_xfer_cnt = cpu_to_be32(scsi_bufflen(scmnd)); wr 437 drivers/scsi/csiostor/csio_scsi.c wr->rsp_dmalen = cpu_to_be32(dma_buf->len); wr 438 drivers/scsi/csiostor/csio_scsi.c wr->rsp_dmaaddr = cpu_to_be64(dma_buf->paddr); wr 440 drivers/scsi/csiostor/csio_scsi.c wr->r4 = 0; wr 442 drivers/scsi/csiostor/csio_scsi.c wr->u.fcoe.ctl_pri = 0; wr 443 drivers/scsi/csiostor/csio_scsi.c wr->u.fcoe.cp_en_class = 0; wr 444 drivers/scsi/csiostor/csio_scsi.c wr->u.fcoe.r3_lo[0] = 0; wr 445 drivers/scsi/csiostor/csio_scsi.c wr->u.fcoe.r3_lo[1] = 0; wr 651 drivers/scsi/csiostor/csio_scsi.c struct fw_scsi_abrt_cls_wr *wr = (struct fw_scsi_abrt_cls_wr *)addr; wr 653 drivers/scsi/csiostor/csio_scsi.c wr->op_immdlen = cpu_to_be32(FW_WR_OP_V(FW_SCSI_ABRT_CLS_WR)); wr 654 drivers/scsi/csiostor/csio_scsi.c wr->flowid_len16 = cpu_to_be32(FW_WR_FLOWID_V(rn->flowid) | wr 658 drivers/scsi/csiostor/csio_scsi.c wr->cookie = (uintptr_t) req; wr 659 drivers/scsi/csiostor/csio_scsi.c wr->iqid = cpu_to_be16(csio_q_physiqid(hw, req->iq_idx)); wr 660 drivers/scsi/csiostor/csio_scsi.c wr->tmo_val = (uint8_t) req->tmo; wr 662 drivers/scsi/csiostor/csio_scsi.c wr->sub_opcode_to_chk_all_io = wr 665 drivers/scsi/csiostor/csio_scsi.c wr->r3[0] = 0; wr 666 drivers/scsi/csiostor/csio_scsi.c wr->r3[1] = 0; wr 667 drivers/scsi/csiostor/csio_scsi.c wr->r3[2] = 0; wr 668 drivers/scsi/csiostor/csio_scsi.c wr->r3[3] = 0; wr 670 drivers/scsi/csiostor/csio_scsi.c wr->t_cookie = (uintptr_t) req; wr 1102 drivers/scsi/csiostor/csio_scsi.c csio_scsi_cmpl_handler(struct csio_hw *hw, void *wr, uint32_t len, wr 1112 drivers/scsi/csiostor/csio_scsi.c cpl = (struct cpl_fw6_msg *)((uintptr_t)wr + sizeof(__be64)); wr 766 drivers/scsi/csiostor/csio_wr.c void *wr; wr 775 drivers/scsi/csiostor/csio_wr.c wr = (void *)((uintptr_t)q->vstart + wr 778 drivers/scsi/csiostor/csio_wr.c ftr = (struct csio_iqwr_footer *)((uintptr_t)wr + wr 1042 drivers/scsi/csiostor/csio_wr.c void *wr, uint32_t len_to_qid, wr 1092 drivers/scsi/csiostor/csio_wr.c iq_handler(hw, wr, q->wr_sz - sizeof(struct csio_iqwr_footer), wr 1135 drivers/scsi/csiostor/csio_wr.c void *wr = (void *)((uintptr_t)q->vstart + (q->cidx * q->wr_sz)); wr 1144 drivers/scsi/csiostor/csio_wr.c ftr = (struct csio_iqwr_footer *)((uintptr_t)wr + wr 1153 drivers/scsi/csiostor/csio_wr.c CSIO_DB_ASSERT(((uintptr_t)wr + q->wr_sz) <= wr 1161 drivers/scsi/csiostor/csio_wr.c iq_handler(hw, wr, q->wr_sz - sizeof(*ftr), NULL, priv); wr 1164 drivers/scsi/csiostor/csio_wr.c csio_wr_process_fl(hw, q, wr, wr 1203 drivers/scsi/csiostor/csio_wr.c if (((uintptr_t)wr + q->wr_sz) == (uintptr_t)q->vwrap) { wr 1207 drivers/scsi/csiostor/csio_wr.c wr = q->vstart; wr 1215 drivers/scsi/csiostor/csio_wr.c wr = (void *)((uintptr_t)(q->vstart) + wr 1219 drivers/scsi/csiostor/csio_wr.c ftr = (struct csio_iqwr_footer *)((uintptr_t)wr + wr 165 drivers/scsi/cxgbi/cxgb3i/cxgb3i.c req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_FORWARD)); wr 210 drivers/scsi/cxgbi/cxgb3i/cxgb3i.c req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_OFLD_CLOSE_CON)); wr 211 drivers/scsi/cxgbi/cxgb3i/cxgb3i.c req->wr.wr_lo = htonl(V_WR_TID(tid)); wr 254 drivers/scsi/cxgbi/cxgb3i/cxgb3i.c req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_OFLD_HOST_ABORT_CON_REQ)); wr 255 drivers/scsi/cxgbi/cxgb3i/cxgb3i.c req->wr.wr_lo = htonl(V_WR_TID(csk->tid)); wr 285 drivers/scsi/cxgbi/cxgb3i/cxgb3i.c rpl->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_OFLD_HOST_ABORT_CON_RPL)); wr 286 drivers/scsi/cxgbi/cxgb3i/cxgb3i.c rpl->wr.wr_lo = htonl(V_WR_TID(csk->tid)); wr 313 drivers/scsi/cxgbi/cxgb3i/cxgb3i.c req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_FORWARD)); wr 1075 drivers/scsi/cxgbi/cxgb3i/cxgb3i.c req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_BYPASS)); wr 1159 drivers/scsi/cxgbi/cxgb3i/cxgb3i.c req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_FORWARD)); wr 1195 drivers/scsi/cxgbi/cxgb3i/cxgb3i.c req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_FORWARD)); wr 784 drivers/scsi/cxgbi/cxgb4i/cxgb4i.c req->wr.wr_hi |= htonl(FW_WR_COMPL_F); wr 1898 drivers/scsi/cxgbi/cxgb4i/cxgb4i.c req->wr.wr_hi = htonl(FW_WR_OP_V(FW_ULPTX_WR) | wr 1905 drivers/scsi/cxgbi/cxgb4i/cxgb4i.c req->len16 = htonl(DIV_ROUND_UP(wr_len - sizeof(req->wr), 16)); wr 3698 drivers/scsi/qla2xxx/qla_mbx.c qla2x00_disable_fce_trace(scsi_qla_host_t *vha, uint64_t *wr, uint64_t *rd) wr 3730 drivers/scsi/qla2xxx/qla_mbx.c if (wr) wr 3731 drivers/scsi/qla2xxx/qla_mbx.c *wr = (uint64_t) mcp->mb[5] << 48 | wr 510 drivers/spi/spi-bcm-qspi.c u32 wr = 0; wr 514 drivers/spi/spi-bcm-qspi.c wr = (rd & ~0xff) | (1 << cs); wr 515 drivers/spi/spi-bcm-qspi.c if (rd == wr) wr 517 drivers/spi/spi-bcm-qspi.c bcm_qspi_write(qspi, CHIP_SELECT, 0, wr); wr 29 drivers/staging/fbtft/fb_agm1264k-fl.c #define EPIN gpio.wr wr 112 drivers/staging/fbtft/fbtft-core.c ret = fbtft_request_one_gpio(par, "wr", 0, &par->gpio.wr); wr 1125 drivers/staging/fbtft/fbtft-core.c if (!par->gpio.wr) { wr 145 drivers/staging/fbtft/fbtft-io.c gpiod_set_value(par->gpio.wr, 0); wr 150 drivers/staging/fbtft/fbtft-io.c gpiod_set_value(par->gpio.wr, 0); /* used as delay */ wr 168 drivers/staging/fbtft/fbtft-io.c gpiod_set_value(par->gpio.wr, 1); wr 195 drivers/staging/fbtft/fbtft-io.c gpiod_set_value(par->gpio.wr, 0); wr 200 drivers/staging/fbtft/fbtft-io.c gpiod_set_value(par->gpio.wr, 0); /* used as delay */ wr 218 drivers/staging/fbtft/fbtft-io.c gpiod_set_value(par->gpio.wr, 1); wr 212 drivers/staging/fbtft/fbtft.h struct gpio_desc *wr; wr 81 drivers/target/iscsi/cxgbit/cxgbit_ddp.c req->wr.wr_hi = htonl(FW_WR_OP_V(FW_ULPTX_WR) | wr 88 drivers/target/iscsi/cxgbit/cxgbit_ddp.c req->len16 = htonl(DIV_ROUND_UP(wr_len - sizeof(req->wr), 16)); wr 272 drivers/target/iscsi/cxgbit/cxgbit_target.c req->wr.wr_hi |= htonl(FW_WR_COMPL_F); wr 50 fs/autofs/waitq.c ssize_t wr = 0; wr 56 fs/autofs/waitq.c wr = __kernel_write(file, data, bytes, &file->f_pos); wr 57 fs/autofs/waitq.c if (wr <= 0) wr 59 fs/autofs/waitq.c data += wr; wr 60 fs/autofs/waitq.c bytes -= wr; wr 67 fs/autofs/waitq.c if (wr == -EPIPE && !sigpipe) { wr 75 fs/autofs/waitq.c return bytes == 0 ? 0 : wr < 0 ? wr : -EIO; wr 2510 fs/cifs/smbdirect.c reg_wr = &smbdirect_mr->wr; wr 2511 fs/cifs/smbdirect.c reg_wr->wr.opcode = IB_WR_REG_MR; wr 2513 fs/cifs/smbdirect.c reg_wr->wr.wr_cqe = &smbdirect_mr->cqe; wr 2514 fs/cifs/smbdirect.c reg_wr->wr.num_sge = 0; wr 2515 fs/cifs/smbdirect.c reg_wr->wr.send_flags = IB_SEND_SIGNALED; wr 2527 fs/cifs/smbdirect.c rc = ib_post_send(info->id->qp, ®_wr->wr, NULL); wr 2572 fs/cifs/smbdirect.c struct ib_send_wr *wr; wr 2578 fs/cifs/smbdirect.c wr = &smbdirect_mr->inv_wr; wr 2579 fs/cifs/smbdirect.c wr->opcode = IB_WR_LOCAL_INV; wr 2581 fs/cifs/smbdirect.c wr->wr_cqe = &smbdirect_mr->cqe; wr 2582 fs/cifs/smbdirect.c wr->num_sge = 0; wr 2583 fs/cifs/smbdirect.c wr->ex.invalidate_rkey = smbdirect_mr->mr->rkey; wr 2584 fs/cifs/smbdirect.c wr->send_flags = IB_SEND_SIGNALED; wr 2587 fs/cifs/smbdirect.c rc = ib_post_send(info->id->qp, wr, NULL); wr 297 fs/cifs/smbdirect.h struct ib_reg_wr wr; wr 50 fs/orangefs/file.c loff_t readahead_size, struct orangefs_write_range *wr, int *index_return) wr 89 fs/orangefs/file.c if (type == ORANGEFS_IO_WRITE && wr) { wr 90 fs/orangefs/file.c new_op->upcall.uid = from_kuid(&init_user_ns, wr->uid); wr 91 fs/orangefs/file.c new_op->upcall.gid = from_kgid(&init_user_ns, wr->gid); wr 22 fs/orangefs/inode.c struct orangefs_write_range *wr = NULL; wr 33 fs/orangefs/inode.c wr = (struct orangefs_write_range *)page_private(page); wr 34 fs/orangefs/inode.c WARN_ON(wr->pos >= len); wr 35 fs/orangefs/inode.c off = wr->pos; wr 36 fs/orangefs/inode.c if (off + wr->len > len) wr 39 fs/orangefs/inode.c wlen = wr->len; wr 58 fs/orangefs/inode.c len, wr, NULL); wr 65 fs/orangefs/inode.c if (wr) { wr 66 fs/orangefs/inode.c kfree(wr); wr 98 fs/orangefs/inode.c struct orangefs_write_range *wrp, wr; wr 126 fs/orangefs/inode.c wr.uid = ow->uid; wr 127 fs/orangefs/inode.c wr.gid = ow->gid; wr 129 fs/orangefs/inode.c 0, &wr, NULL); wr 165 fs/orangefs/inode.c struct orangefs_write_range *wr; wr 175 fs/orangefs/inode.c wr = (struct orangefs_write_range *)page_private(page); wr 179 fs/orangefs/inode.c ow->off = wr->pos; wr 180 fs/orangefs/inode.c ow->len = wr->len; wr 181 fs/orangefs/inode.c ow->uid = wr->uid; wr 182 fs/orangefs/inode.c ow->gid = wr->gid; wr 187 fs/orangefs/inode.c if (!uid_eq(ow->uid, wr->uid) || !gid_eq(ow->gid, wr->gid)) { wr 193 fs/orangefs/inode.c if (ow->off + ow->len == wr->pos) { wr 194 fs/orangefs/inode.c ow->len += wr->len; wr 393 fs/orangefs/inode.c struct orangefs_write_range *wr; wr 417 fs/orangefs/inode.c struct orangefs_write_range *wr; wr 418 fs/orangefs/inode.c wr = (struct orangefs_write_range *)page_private(page); wr 419 fs/orangefs/inode.c if (wr->pos + wr->len == pos && wr 420 fs/orangefs/inode.c uid_eq(wr->uid, current_fsuid()) && wr 421 fs/orangefs/inode.c gid_eq(wr->gid, current_fsgid())) { wr 422 fs/orangefs/inode.c wr->len += len; wr 431 fs/orangefs/inode.c wr = kmalloc(sizeof *wr, GFP_KERNEL); wr 432 fs/orangefs/inode.c if (!wr) wr 435 fs/orangefs/inode.c wr->pos = pos; wr 436 fs/orangefs/inode.c wr->len = len; wr 437 fs/orangefs/inode.c wr->uid = current_fsuid(); wr 438 fs/orangefs/inode.c wr->gid = current_fsgid(); wr 440 fs/orangefs/inode.c set_page_private(page, (unsigned long)wr); wr 485 fs/orangefs/inode.c struct orangefs_write_range *wr; wr 486 fs/orangefs/inode.c wr = (struct orangefs_write_range *)page_private(page); wr 495 fs/orangefs/inode.c } else if (page_offset(page) + offset <= wr->pos && wr 496 fs/orangefs/inode.c wr->pos + wr->len <= page_offset(page) + offset + length) { wr 505 fs/orangefs/inode.c } else if (wr->pos < page_offset(page) + offset && wr 506 fs/orangefs/inode.c wr->pos + wr->len <= page_offset(page) + offset + length && wr 507 fs/orangefs/inode.c page_offset(page) + offset < wr->pos + wr->len) { wr 509 fs/orangefs/inode.c x = wr->pos + wr->len - (page_offset(page) + offset); wr 510 fs/orangefs/inode.c WARN_ON(x > wr->len); wr 511 fs/orangefs/inode.c wr->len -= x; wr 512 fs/orangefs/inode.c wr->uid = current_fsuid(); wr 513 fs/orangefs/inode.c wr->gid = current_fsgid(); wr 515 fs/orangefs/inode.c } else if (page_offset(page) + offset <= wr->pos && wr 516 fs/orangefs/inode.c page_offset(page) + offset + length < wr->pos + wr->len && wr 517 fs/orangefs/inode.c wr->pos < page_offset(page) + offset + length) { wr 519 fs/orangefs/inode.c x = page_offset(page) + offset + length - wr->pos; wr 520 fs/orangefs/inode.c WARN_ON(x > wr->len); wr 521 fs/orangefs/inode.c wr->pos += x; wr 522 fs/orangefs/inode.c wr->len -= x; wr 523 fs/orangefs/inode.c wr->uid = current_fsuid(); wr 524 fs/orangefs/inode.c wr->gid = current_fsgid(); wr 526 fs/orangefs/inode.c } else if (wr->pos < page_offset(page) + offset && wr 527 fs/orangefs/inode.c page_offset(page) + offset + length < wr->pos + wr->len) { wr 539 fs/orangefs/inode.c if (!((page_offset(page) + offset + length <= wr->pos) ^ wr 540 fs/orangefs/inode.c (wr->pos + wr->len <= page_offset(page) + offset))) { wr 545 fs/orangefs/inode.c wr->pos, wr->len); wr 726 fs/orangefs/inode.c struct orangefs_write_range *wr; wr 748 fs/orangefs/inode.c wr = (struct orangefs_write_range *)page_private(page); wr 749 fs/orangefs/inode.c if (uid_eq(wr->uid, current_fsuid()) && wr 750 fs/orangefs/inode.c gid_eq(wr->gid, current_fsgid())) { wr 751 fs/orangefs/inode.c wr->pos = page_offset(page); wr 752 fs/orangefs/inode.c wr->len = PAGE_SIZE; wr 761 fs/orangefs/inode.c wr = kmalloc(sizeof *wr, GFP_KERNEL); wr 762 fs/orangefs/inode.c if (!wr) { wr 766 fs/orangefs/inode.c wr->pos = page_offset(page); wr 767 fs/orangefs/inode.c wr->len = PAGE_SIZE; wr 768 fs/orangefs/inode.c wr->uid = current_fsuid(); wr 769 fs/orangefs/inode.c wr->gid = current_fsgid(); wr 771 fs/orangefs/inode.c set_page_private(page, (unsigned long)wr); wr 1889 fs/ubifs/lpt.c int ubifs_lpt_init(struct ubifs_info *c, int rd, int wr) wr 1899 fs/ubifs/lpt.c if (wr) { wr 1908 fs/ubifs/lpt.c if (wr) wr 1930 fs/ubifs/ubifs.h int ubifs_lpt_init(struct ubifs_info *c, int rd, int wr); wr 28 include/linux/hdlcdrv.h unsigned rd, wr; wr 35 include/linux/hdlcdrv.h unsigned int wr; wr 49 include/linux/hdlcdrv.h buf->buffer[buf->wr] = buf->shreg; wr 50 include/linux/hdlcdrv.h buf->wr = (buf->wr+1) % sizeof(buf->buffer); wr 58 include/linux/hdlcdrv.h buf->buffer[buf->wr] = bits & 0xff; wr 59 include/linux/hdlcdrv.h buf->wr = (buf->wr+1) % sizeof(buf->buffer); wr 60 include/linux/hdlcdrv.h buf->buffer[buf->wr] = (bits >> 8) & 0xff; wr 61 include/linux/hdlcdrv.h buf->wr = (buf->wr+1) % sizeof(buf->buffer); wr 164 include/linux/hdlcdrv.h ret = !((HDLCDRV_HDLCBUFFER - 1 + hb->rd - hb->wr) % HDLCDRV_HDLCBUFFER); wr 177 include/linux/hdlcdrv.h ret = (hb->rd == hb->wr); wr 191 include/linux/hdlcdrv.h if (hb->rd == hb->wr) wr 211 include/linux/hdlcdrv.h newp = (hb->wr+1) % HDLCDRV_HDLCBUFFER; wr 213 include/linux/hdlcdrv.h hb->buf[hb->wr] = val & 0xffff; wr 214 include/linux/hdlcdrv.h hb->wr = newp; wr 185 include/linux/sunrpc/svc_rdma.h extern int svc_rdma_send(struct svcxprt_rdma *rdma, struct ib_send_wr *wr); wr 1341 include/rdma/ib_verbs.h struct ib_send_wr wr; wr 1346 include/rdma/ib_verbs.h static inline const struct ib_rdma_wr *rdma_wr(const struct ib_send_wr *wr) wr 1348 include/rdma/ib_verbs.h return container_of(wr, struct ib_rdma_wr, wr); wr 1352 include/rdma/ib_verbs.h struct ib_send_wr wr; wr 1361 include/rdma/ib_verbs.h static inline const struct ib_atomic_wr *atomic_wr(const struct ib_send_wr *wr) wr 1363 include/rdma/ib_verbs.h return container_of(wr, struct ib_atomic_wr, wr); wr 1367 include/rdma/ib_verbs.h struct ib_send_wr wr; wr 1378 include/rdma/ib_verbs.h static inline const struct ib_ud_wr *ud_wr(const struct ib_send_wr *wr) wr 1380 include/rdma/ib_verbs.h return container_of(wr, struct ib_ud_wr, wr); wr 1384 include/rdma/ib_verbs.h struct ib_send_wr wr; wr 1390 include/rdma/ib_verbs.h static inline const struct ib_reg_wr *reg_wr(const struct ib_send_wr *wr) wr 1392 include/rdma/ib_verbs.h return container_of(wr, struct ib_reg_wr, wr); wr 172 include/rdma/rdmavt_qp.h struct ib_ud_wr wr; wr 183 include/rdma/rdmavt_qp.h struct ib_send_wr wr; /* don't use wr.sg_list */ wr 229 include/rdma/rdmavt_qp.h return ibah_to_rvtah(swqe->ud_wr.wr.ah); wr 249 include/rdma/rdmavt_qp.h return swqe->ud_wr.wr.remote_qpn; wr 259 include/rdma/rdmavt_qp.h return swqe->ud_wr.wr.remote_qkey; wr 269 include/rdma/rdmavt_qp.h return swqe->ud_wr.wr.pkey_index; wr 586 include/rdma/rdmavt_qp.h for (i = 0; i < wqe->wr.num_sge; i++) { wr 852 include/rdma/rdmavt_qp.h int flags = wqe->wr.send_flags; wr 863 include/rdma/rdmavt_qp.h wr_id = wqe->wr.wr_id; wr 25 include/rdma/rw.h struct ib_rdma_wr wr; wr 37 include/rdma/rw.h struct ib_rdma_wr wr; wr 24 include/trace/events/ib_mad.h TP_PROTO(struct ib_mad_send_wr_private *wr, wr 26 include/trace/events/ib_mad.h TP_ARGS(wr, qp_info), wr 56 include/trace/events/ib_mad.h __entry->dev_index = wr->mad_agent_priv->agent.device->index; wr 57 include/trace/events/ib_mad.h __entry->port_num = wr->mad_agent_priv->agent.port_num; wr 58 include/trace/events/ib_mad.h __entry->qp_num = wr->mad_agent_priv->qp_info->qp->qp_num; wr 59 include/trace/events/ib_mad.h __entry->agent_priv = wr->mad_agent_priv; wr 60 include/trace/events/ib_mad.h __entry->wrtid = wr->tid; wr 61 include/trace/events/ib_mad.h __entry->max_retries = wr->max_retries; wr 62 include/trace/events/ib_mad.h __entry->retries_left = wr->retries_left; wr 63 include/trace/events/ib_mad.h __entry->retry = wr->retry; wr 64 include/trace/events/ib_mad.h __entry->timeout = wr->timeout; wr 65 include/trace/events/ib_mad.h __entry->length = wr->send_buf.hdr_len + wr 66 include/trace/events/ib_mad.h wr->send_buf.data_len; wr 68 include/trace/events/ib_mad.h ((struct ib_mad_hdr *)wr->send_buf.mad)->base_version; wr 70 include/trace/events/ib_mad.h ((struct ib_mad_hdr *)wr->send_buf.mad)->mgmt_class; wr 72 include/trace/events/ib_mad.h ((struct ib_mad_hdr *)wr->send_buf.mad)->class_version; wr 74 include/trace/events/ib_mad.h ((struct ib_mad_hdr *)wr->send_buf.mad)->method; wr 76 include/trace/events/ib_mad.h ((struct ib_mad_hdr *)wr->send_buf.mad)->status; wr 78 include/trace/events/ib_mad.h ((struct ib_mad_hdr *)wr->send_buf.mad)->class_specific; wr 79 include/trace/events/ib_mad.h __entry->tid = ((struct ib_mad_hdr *)wr->send_buf.mad)->tid; wr 81 include/trace/events/ib_mad.h ((struct ib_mad_hdr *)wr->send_buf.mad)->attr_id; wr 83 include/trace/events/ib_mad.h ((struct ib_mad_hdr *)wr->send_buf.mad)->attr_mod; wr 84 include/trace/events/ib_mad.h create_mad_addr_info(wr, qp_info, __entry); wr 109 include/trace/events/ib_mad.h TP_PROTO(struct ib_mad_send_wr_private *wr, wr 111 include/trace/events/ib_mad.h TP_ARGS(wr, qp_info)); wr 113 include/trace/events/ib_mad.h TP_PROTO(struct ib_mad_send_wr_private *wr, wr 115 include/trace/events/ib_mad.h TP_ARGS(wr, qp_info)); wr 117 include/trace/events/ib_mad.h TP_PROTO(struct ib_mad_send_wr_private *wr, wr 119 include/trace/events/ib_mad.h TP_ARGS(wr, qp_info)); wr 122 include/trace/events/ib_mad.h TP_PROTO(struct ib_mad_send_wr_private *wr, struct ib_wc *wc), wr 123 include/trace/events/ib_mad.h TP_ARGS(wr, wc), wr 145 include/trace/events/ib_mad.h __entry->dev_index = wr->mad_agent_priv->agent.device->index; wr 146 include/trace/events/ib_mad.h __entry->port_num = wr->mad_agent_priv->agent.port_num; wr 147 include/trace/events/ib_mad.h __entry->qp_num = wr->mad_agent_priv->qp_info->qp->qp_num; wr 148 include/trace/events/ib_mad.h __entry->agent_priv = wr->mad_agent_priv; wr 149 include/trace/events/ib_mad.h __entry->wrtid = wr->tid; wr 150 include/trace/events/ib_mad.h __entry->max_retries = wr->max_retries; wr 151 include/trace/events/ib_mad.h __entry->retries_left = wr->retries_left; wr 152 include/trace/events/ib_mad.h __entry->retry = wr->retry; wr 153 include/trace/events/ib_mad.h __entry->timeout = wr->timeout; wr 155 include/trace/events/ib_mad.h ((struct ib_mad_hdr *)wr->send_buf.mad)->base_version; wr 157 include/trace/events/ib_mad.h ((struct ib_mad_hdr *)wr->send_buf.mad)->mgmt_class; wr 159 include/trace/events/ib_mad.h ((struct ib_mad_hdr *)wr->send_buf.mad)->class_version; wr 161 include/trace/events/ib_mad.h ((struct ib_mad_hdr *)wr->send_buf.mad)->method; wr 163 include/trace/events/ib_mad.h ((struct ib_mad_hdr *)wr->send_buf.mad)->status; wr 1641 include/trace/events/rpcrdma.h const struct ib_send_wr *wr wr 1644 include/trace/events/rpcrdma.h TP_ARGS(wr), wr 1653 include/trace/events/rpcrdma.h __entry->cqe = wr->wr_cqe; wr 1654 include/trace/events/rpcrdma.h __entry->num_sge = wr->num_sge; wr 1655 include/trace/events/rpcrdma.h __entry->inv_rkey = (wr->opcode == IB_WR_SEND_WITH_INV) ? wr 1656 include/trace/events/rpcrdma.h wr->ex.invalidate_rkey : 0; wr 1669 include/trace/events/rpcrdma.h const struct ib_recv_wr *wr, wr 1673 include/trace/events/rpcrdma.h TP_ARGS(wr, status), wr 1681 include/trace/events/rpcrdma.h __entry->cqe = wr->wr_cqe; wr 821 include/uapi/rdma/ib_user_verbs.h } wr; wr 105 include/uapi/rdma/rdma_user_rxe.h } wr; wr 134 include/uapi/rdma/rdma_user_rxe.h struct rxe_send_wr wr; wr 273 include/uapi/rdma/vmw_pvrdma-abi.h } wr; wr 292 lib/decompress_unlzma.c static inline size_t INIT get_pos(struct writer *wr) wr 295 lib/decompress_unlzma.c wr->global_pos + wr->buffer_pos; wr 298 lib/decompress_unlzma.c static inline uint8_t INIT peek_old_byte(struct writer *wr, wr 301 lib/decompress_unlzma.c if (!wr->flush) { wr 303 lib/decompress_unlzma.c while (offs > wr->header->dict_size) wr 304 lib/decompress_unlzma.c offs -= wr->header->dict_size; wr 305 lib/decompress_unlzma.c pos = wr->buffer_pos - offs; wr 306 lib/decompress_unlzma.c return wr->buffer[pos]; wr 308 lib/decompress_unlzma.c uint32_t pos = wr->buffer_pos - offs; wr 309 lib/decompress_unlzma.c while (pos >= wr->header->dict_size) wr 310 lib/decompress_unlzma.c pos += wr->header->dict_size; wr 311 lib/decompress_unlzma.c return wr->buffer[pos]; wr 316 lib/decompress_unlzma.c static inline int INIT write_byte(struct writer *wr, uint8_t byte) wr 318 lib/decompress_unlzma.c wr->buffer[wr->buffer_pos++] = wr->previous_byte = byte; wr 319 lib/decompress_unlzma.c if (wr->flush && wr->buffer_pos == wr->header->dict_size) { wr 320 lib/decompress_unlzma.c wr->buffer_pos = 0; wr 321 lib/decompress_unlzma.c wr->global_pos += wr->header->dict_size; wr 322 lib/decompress_unlzma.c if (wr->flush((char *)wr->buffer, wr->header->dict_size) wr 323 lib/decompress_unlzma.c != wr->header->dict_size) wr 330 lib/decompress_unlzma.c static inline int INIT copy_byte(struct writer *wr, uint32_t offs) wr 332 lib/decompress_unlzma.c return write_byte(wr, peek_old_byte(wr, offs)); wr 335 lib/decompress_unlzma.c static inline int INIT copy_bytes(struct writer *wr, wr 339 lib/decompress_unlzma.c if (copy_byte(wr, rep0)) wr 342 lib/decompress_unlzma.c } while (len != 0 && wr->buffer_pos < wr->header->dst_size); wr 347 lib/decompress_unlzma.c static inline int INIT process_bit0(struct writer *wr, struct rc *rc, wr 355 lib/decompress_unlzma.c * (((get_pos(wr) & literal_pos_mask) << lc) wr 356 lib/decompress_unlzma.c + (wr->previous_byte >> (8 - lc)))) wr 360 lib/decompress_unlzma.c int match_byte = peek_old_byte(wr, cst->rep0); wr 388 lib/decompress_unlzma.c return write_byte(wr, mi); wr 391 lib/decompress_unlzma.c static inline int INIT process_bit1(struct writer *wr, struct rc *rc, wr 422 lib/decompress_unlzma.c return copy_byte(wr, cst->rep0); wr 525 lib/decompress_unlzma.c if (cst->rep0 > wr->header->dict_size wr 526 lib/decompress_unlzma.c || cst->rep0 > get_pos(wr)) wr 532 lib/decompress_unlzma.c return copy_bytes(wr, cst->rep0, len); wr 553 lib/decompress_unlzma.c struct writer wr; wr 572 lib/decompress_unlzma.c wr.header = &header; wr 573 lib/decompress_unlzma.c wr.flush = flush; wr 574 lib/decompress_unlzma.c wr.global_pos = 0; wr 575 lib/decompress_unlzma.c wr.previous_byte = 0; wr 576 lib/decompress_unlzma.c wr.buffer_pos = 0; wr 613 lib/decompress_unlzma.c wr.buffer = output; wr 615 lib/decompress_unlzma.c wr.bufsize = MIN(header.dst_size, header.dict_size); wr 616 lib/decompress_unlzma.c wr.buffer = large_malloc(wr.bufsize); wr 618 lib/decompress_unlzma.c if (wr.buffer == NULL) wr 631 lib/decompress_unlzma.c while (get_pos(&wr) < header.dst_size) { wr 632 lib/decompress_unlzma.c int pos_state = get_pos(&wr) & pos_state_mask; wr 636 lib/decompress_unlzma.c if (process_bit0(&wr, &rc, &cst, p, pos_state, prob, wr 642 lib/decompress_unlzma.c if (process_bit1(&wr, &rc, &cst, p, pos_state, prob)) { wr 655 lib/decompress_unlzma.c if (!wr.flush || wr.flush(wr.buffer, wr.buffer_pos) == wr.buffer_pos) wr 661 lib/decompress_unlzma.c large_free(wr.buffer); wr 144 net/9p/trans_fd.c struct file *wr; wr 234 net/9p/trans_fd.c if (ts->rd != ts->wr) wr 235 net/9p/trans_fd.c ret = (ret & ~EPOLLOUT) | (vfs_poll(ts->wr, pt) & ~EPOLLIN); wr 421 net/9p/trans_fd.c if (!(ts->wr->f_flags & O_NONBLOCK)) wr 424 net/9p/trans_fd.c ret = kernel_write(ts->wr, v, len, &ts->wr->f_pos); wr 806 net/9p/trans_fd.c ts->wr = fget(wfd); wr 807 net/9p/trans_fd.c if (!ts->rd || !ts->wr) { wr 810 net/9p/trans_fd.c if (ts->wr) wr 811 net/9p/trans_fd.c fput(ts->wr); wr 841 net/9p/trans_fd.c p->wr = p->rd = file; wr 902 net/9p/trans_fd.c if (ts->wr) wr 903 net/9p/trans_fd.c fput(ts->wr); wr 386 net/9p/trans_rdma.c struct ib_recv_wr wr; wr 401 net/9p/trans_rdma.c wr.next = NULL; wr 402 net/9p/trans_rdma.c wr.wr_cqe = &c->cqe; wr 403 net/9p/trans_rdma.c wr.sg_list = &sge; wr 404 net/9p/trans_rdma.c wr.num_sge = 1; wr 405 net/9p/trans_rdma.c return ib_post_recv(rdma->qp, &wr, NULL); wr 415 net/9p/trans_rdma.c struct ib_send_wr wr; wr 493 net/9p/trans_rdma.c wr.next = NULL; wr 494 net/9p/trans_rdma.c wr.wr_cqe = &c->cqe; wr 495 net/9p/trans_rdma.c wr.opcode = IB_WR_SEND; wr 496 net/9p/trans_rdma.c wr.send_flags = IB_SEND_SIGNALED; wr 497 net/9p/trans_rdma.c wr.sg_list = &sge; wr 498 net/9p/trans_rdma.c wr.num_sge = 1; wr 510 net/9p/trans_rdma.c err = ib_post_send(rdma->qp, &wr, NULL); wr 154 net/rds/ib_frmr.c reg_wr.wr.wr_id = (unsigned long)(void *)ibmr; wr 155 net/rds/ib_frmr.c reg_wr.wr.opcode = IB_WR_REG_MR; wr 156 net/rds/ib_frmr.c reg_wr.wr.num_sge = 0; wr 162 net/rds/ib_frmr.c reg_wr.wr.send_flags = IB_SEND_SIGNALED; wr 164 net/rds/ib_frmr.c ret = ib_post_send(ibmr->ic->i_cm_id->qp, ®_wr.wr, NULL); wr 575 net/rds/ib_recv.c struct ib_send_wr *wr = &ic->i_ack_wr; wr 582 net/rds/ib_recv.c wr->sg_list = sge; wr 583 net/rds/ib_recv.c wr->num_sge = 1; wr 584 net/rds/ib_recv.c wr->opcode = IB_WR_SEND; wr 585 net/rds/ib_recv.c wr->wr_id = RDS_IB_ACK_WR_ID; wr 586 net/rds/ib_recv.c wr->send_flags = IB_SEND_SIGNALED | IB_SEND_SOLICITED; wr 775 net/rds/ib_send.c send->s_atomic_wr.wr.opcode = IB_WR_MASKED_ATOMIC_CMP_AND_SWP; wr 781 net/rds/ib_send.c send->s_atomic_wr.wr.opcode = IB_WR_MASKED_ATOMIC_FETCH_AND_ADD; wr 789 net/rds/ib_send.c send->s_atomic_wr.wr.num_sge = 1; wr 790 net/rds/ib_send.c send->s_atomic_wr.wr.next = NULL; wr 817 net/rds/ib_send.c failed_wr = &send->s_atomic_wr.wr; wr 818 net/rds/ib_send.c ret = ib_post_send(ic->i_cm_id->qp, &send->s_atomic_wr.wr, &failed_wr); wr 821 net/rds/ib_send.c BUG_ON(failed_wr != &send->s_atomic_wr.wr); wr 830 net/rds/ib_send.c if (unlikely(failed_wr != &send->s_atomic_wr.wr)) { wr 832 net/rds/ib_send.c BUG_ON(failed_wr != &send->s_atomic_wr.wr); wr 909 net/rds/ib_send.c send->s_rdma_wr.wr.num_sge = max_sge; wr 912 net/rds/ib_send.c send->s_rdma_wr.wr.num_sge = num_sge; wr 915 net/rds/ib_send.c send->s_rdma_wr.wr.next = NULL; wr 918 net/rds/ib_send.c prev->s_rdma_wr.wr.next = &send->s_rdma_wr.wr; wr 920 net/rds/ib_send.c for (j = 0; j < send->s_rdma_wr.wr.num_sge && wr 935 net/rds/ib_send.c &send->s_rdma_wr.wr, wr 936 net/rds/ib_send.c send->s_rdma_wr.wr.num_sge, wr 937 net/rds/ib_send.c send->s_rdma_wr.wr.next); wr 958 net/rds/ib_send.c failed_wr = &first->s_rdma_wr.wr; wr 959 net/rds/ib_send.c ret = ib_post_send(ic->i_cm_id->qp, &first->s_rdma_wr.wr, &failed_wr); wr 961 net/rds/ib_send.c first, &first->s_rdma_wr.wr, ret, failed_wr); wr 962 net/rds/ib_send.c BUG_ON(failed_wr != &first->s_rdma_wr.wr); wr 971 net/rds/ib_send.c if (unlikely(failed_wr != &first->s_rdma_wr.wr)) { wr 973 net/rds/ib_send.c BUG_ON(failed_wr != &first->s_rdma_wr.wr); wr 275 net/smc/smc_tx.c rdma_wr->wr.wr_id = smc_wr_tx_get_next_wr_id(link); wr 276 net/smc/smc_tx.c rdma_wr->wr.num_sge = num_sges; wr 284 net/smc/smc_tx.c rc = ib_post_send(link->roce_qp, &rdma_wr->wr, NULL); wr 324 net/smc/smc_tx.c wr_rdma_buf->wr_tx_rdma[dstchunk].wr.sg_list; wr 263 net/smc/smc_wr.c link->wr_reg.wr.wr_id = (u64)(uintptr_t)mr; wr 266 net/smc/smc_wr.c rc = ib_post_send(link->roce_qp, &link->wr_reg.wr, NULL); wr 486 net/smc/smc_wr.c lnk->wr_tx_rdmas[i].wr_tx_rdma[0].wr.opcode = IB_WR_RDMA_WRITE; wr 487 net/smc/smc_wr.c lnk->wr_tx_rdmas[i].wr_tx_rdma[1].wr.opcode = IB_WR_RDMA_WRITE; wr 488 net/smc/smc_wr.c lnk->wr_tx_rdmas[i].wr_tx_rdma[0].wr.sg_list = wr 490 net/smc/smc_wr.c lnk->wr_tx_rdmas[i].wr_tx_rdma[1].wr.sg_list = wr 502 net/smc/smc_wr.c lnk->wr_reg.wr.next = NULL; wr 503 net/smc/smc_wr.c lnk->wr_reg.wr.num_sge = 0; wr 504 net/smc/smc_wr.c lnk->wr_reg.wr.send_flags = IB_SEND_SIGNALED; wr 505 net/smc/smc_wr.c lnk->wr_reg.wr.opcode = IB_WR_REG_MR; wr 435 net/sunrpc/xprtrdma/frwr_ops.c frwr->fr_regwr.wr.next = post_wr; wr 436 net/sunrpc/xprtrdma/frwr_ops.c frwr->fr_regwr.wr.wr_cqe = &frwr->fr_cqe; wr 437 net/sunrpc/xprtrdma/frwr_ops.c frwr->fr_regwr.wr.num_sge = 0; wr 438 net/sunrpc/xprtrdma/frwr_ops.c frwr->fr_regwr.wr.opcode = IB_WR_REG_MR; wr 439 net/sunrpc/xprtrdma/frwr_ops.c frwr->fr_regwr.wr.send_flags = 0; wr 441 net/sunrpc/xprtrdma/frwr_ops.c post_wr = &frwr->fr_regwr.wr; wr 288 net/sunrpc/xprtrdma/svc_rdma_sendto.c int svc_rdma_send(struct svcxprt_rdma *rdma, struct ib_send_wr *wr) wr 309 net/sunrpc/xprtrdma/svc_rdma_sendto.c trace_svcrdma_post_send(wr); wr 310 net/sunrpc/xprtrdma/svc_rdma_sendto.c ret = ib_post_send(rdma->sc_qp, wr, NULL); wr 1500 net/sunrpc/xprtrdma/verbs.c struct ib_recv_wr *i, *wr, *bad_wr; wr 1515 net/sunrpc/xprtrdma/verbs.c wr = NULL; wr 1527 net/sunrpc/xprtrdma/verbs.c rep->rr_recv_wr.next = wr; wr 1528 net/sunrpc/xprtrdma/verbs.c wr = &rep->rr_recv_wr; wr 1531 net/sunrpc/xprtrdma/verbs.c if (!wr) wr 1534 net/sunrpc/xprtrdma/verbs.c for (i = wr; i; i = i->next) { wr 1544 net/sunrpc/xprtrdma/verbs.c rc = ib_post_recv(r_xprt->rx_ia.ri_id->qp, wr, wr 1549 net/sunrpc/xprtrdma/verbs.c for (wr = bad_wr; wr;) { wr 1552 net/sunrpc/xprtrdma/verbs.c rep = container_of(wr, struct rpcrdma_rep, rr_recv_wr); wr 1553 net/sunrpc/xprtrdma/verbs.c wr = wr->next; wr 1562 net/sunrpc/xprtrdma/verbs.c for (i = wr; i;) { wr 1913 net/wireless/reg.c const struct ieee80211_regdomain *wr = get_wiphy_regdom(wiphy); wr 1916 net/wireless/reg.c if (is_world_regdom(cr->alpha2) || (wr && is_world_regdom(wr->alpha2))) wr 91 sound/pci/ctxfi/cthardware.h int (*src_set_wr)(void *blk, unsigned int wr); wr 212 sound/pci/ctxfi/cthw20k1.c static int src_set_wr(void *blk, unsigned int wr) wr 216 sound/pci/ctxfi/cthw20k1.c set_field(&ctl->ctl, SRCCTL_WR, wr); wr 212 sound/pci/ctxfi/cthw20k2.c static int src_set_wr(void *blk, unsigned int wr) wr 216 sound/pci/ctxfi/cthw20k2.c set_field(&ctl->ctl, SRCCTL_WR, wr); wr 1245 sound/soc/codecs/wm_adsp.c unsigned int out, rd, wr, vol; wr 1249 sound/soc/codecs/wm_adsp.c wr = SNDRV_CTL_ELEM_ACCESS_TLV_WRITE; wr 1255 sound/soc/codecs/wm_adsp.c wr = SNDRV_CTL_ELEM_ACCESS_WRITE; wr 1264 sound/soc/codecs/wm_adsp.c out |= wr; wr 1268 sound/soc/codecs/wm_adsp.c out |= rd | wr | vol; wr 19 sound/soc/uniphier/aio-core.c static u64 rb_cnt(u64 wr, u64 rd, u64 len) wr 21 sound/soc/uniphier/aio-core.c if (rd <= wr) wr 22 sound/soc/uniphier/aio-core.c return wr - rd; wr 24 sound/soc/uniphier/aio-core.c return len - (rd - wr); wr 27 sound/soc/uniphier/aio-core.c static u64 rb_cnt_to_end(u64 wr, u64 rd, u64 len) wr 29 sound/soc/uniphier/aio-core.c if (rd <= wr) wr 30 sound/soc/uniphier/aio-core.c return wr - rd; wr 35 sound/soc/uniphier/aio-core.c static u64 rb_space(u64 wr, u64 rd, u64 len) wr 37 sound/soc/uniphier/aio-core.c if (rd <= wr) wr 38 sound/soc/uniphier/aio-core.c return len - (wr - rd) - 8; wr 40 sound/soc/uniphier/aio-core.c return rd - wr - 8; wr 43 sound/soc/uniphier/aio-core.c static u64 rb_space_to_end(u64 wr, u64 rd, u64 len) wr 45 sound/soc/uniphier/aio-core.c if (rd > wr) wr 46 sound/soc/uniphier/aio-core.c return rd - wr - 8; wr 48 sound/soc/uniphier/aio-core.c return len - wr; wr 50 sound/soc/uniphier/aio-core.c return len - wr - 8; wr 308 tools/bpf/bpftool/json_writer.c json_writer_t *wr = jsonw_new(stdout); wr 310 tools/bpf/bpftool/json_writer.c jsonw_start_object(wr); wr 311 tools/bpf/bpftool/json_writer.c jsonw_pretty(wr, true); wr 312 tools/bpf/bpftool/json_writer.c jsonw_name(wr, "Vyatta"); wr 313 tools/bpf/bpftool/json_writer.c jsonw_start_object(wr); wr 314 tools/bpf/bpftool/json_writer.c jsonw_string_field(wr, "url", "http://vyatta.com"); wr 315 tools/bpf/bpftool/json_writer.c jsonw_uint_field(wr, "downloads", 2000000ul); wr 316 tools/bpf/bpftool/json_writer.c jsonw_float_field(wr, "stock", 8.16); wr 318 tools/bpf/bpftool/json_writer.c jsonw_name(wr, "ARGV"); wr 319 tools/bpf/bpftool/json_writer.c jsonw_start_array(wr); wr 321 tools/bpf/bpftool/json_writer.c jsonw_string(wr, *++argv); wr 322 tools/bpf/bpftool/json_writer.c jsonw_end_array(wr); wr 324 tools/bpf/bpftool/json_writer.c jsonw_name(wr, "empty"); wr 325 tools/bpf/bpftool/json_writer.c jsonw_start_array(wr); wr 326 tools/bpf/bpftool/json_writer.c jsonw_end_array(wr); wr 328 tools/bpf/bpftool/json_writer.c jsonw_name(wr, "NIL"); wr 329 tools/bpf/bpftool/json_writer.c jsonw_start_object(wr); wr 330 tools/bpf/bpftool/json_writer.c jsonw_end_object(wr); wr 332 tools/bpf/bpftool/json_writer.c jsonw_null_field(wr, "my_null"); wr 334 tools/bpf/bpftool/json_writer.c jsonw_name(wr, "special chars"); wr 335 tools/bpf/bpftool/json_writer.c jsonw_start_array(wr); wr 336 tools/bpf/bpftool/json_writer.c jsonw_string_field(wr, "slash", "/"); wr 337 tools/bpf/bpftool/json_writer.c jsonw_string_field(wr, "newline", "\n"); wr 338 tools/bpf/bpftool/json_writer.c jsonw_string_field(wr, "tab", "\t"); wr 339 tools/bpf/bpftool/json_writer.c jsonw_string_field(wr, "ff", "\f"); wr 340 tools/bpf/bpftool/json_writer.c jsonw_string_field(wr, "quote", "\""); wr 341 tools/bpf/bpftool/json_writer.c jsonw_string_field(wr, "tick", "\'"); wr 342 tools/bpf/bpftool/json_writer.c jsonw_string_field(wr, "backslash", "\\"); wr 343 tools/bpf/bpftool/json_writer.c jsonw_end_array(wr); wr 345 tools/bpf/bpftool/json_writer.c jsonw_end_object(wr); wr 347 tools/bpf/bpftool/json_writer.c jsonw_end_object(wr); wr 348 tools/bpf/bpftool/json_writer.c jsonw_destroy(&wr); wr 33 tools/testing/selftests/breakpoints/breakpoint_test_arm64.c static void child(int size, int wr) wr 35 tools/testing/selftests/breakpoints/breakpoint_test_arm64.c volatile uint8_t *addr = &var[32 + wr]; wr 112 tools/testing/selftests/breakpoints/breakpoint_test_arm64.c static bool arun_test(int wr_size, int wp_size, int wr, int wp) wr 125 tools/testing/selftests/breakpoints/breakpoint_test_arm64.c child(wr_size, wr); wr 204 tools/testing/selftests/breakpoints/breakpoint_test_arm64.c int wr, wp, size; wr 215 tools/testing/selftests/breakpoints/breakpoint_test_arm64.c for (wr = 0; wr <= 32; wr = wr + size) { wr 216 tools/testing/selftests/breakpoints/breakpoint_test_arm64.c for (wp = wr - size; wp <= wr + size; wp = wp + size) { wr 217 tools/testing/selftests/breakpoints/breakpoint_test_arm64.c result = run_test(size, MIN(size, 8), wr, wp); wr 218 tools/testing/selftests/breakpoints/breakpoint_test_arm64.c if ((result && wr == wp) || wr 219 tools/testing/selftests/breakpoints/breakpoint_test_arm64.c (!result && wr != wp)) wr 222 tools/testing/selftests/breakpoints/breakpoint_test_arm64.c size, wr, wp); wr 226 tools/testing/selftests/breakpoints/breakpoint_test_arm64.c size, wr, wp); wr 420 tools/testing/selftests/net/tcp_mmap.c long wr = FILE_SZ - total; wr 422 tools/testing/selftests/net/tcp_mmap.c if (wr > chunk_size) wr 423 tools/testing/selftests/net/tcp_mmap.c wr = chunk_size; wr 425 tools/testing/selftests/net/tcp_mmap.c wr = send(fd, buffer, wr, zflg ? MSG_ZEROCOPY : 0); wr 426 tools/testing/selftests/net/tcp_mmap.c if (wr <= 0) wr 428 tools/testing/selftests/net/tcp_mmap.c total += wr; wr 1735 virt/kvm/arm/vgic/vgic-its.c #define REGISTER_ITS_DESC(off, rd, wr, length, acc) \ wr 1741 virt/kvm/arm/vgic/vgic-its.c .its_write = wr, \ wr 1744 virt/kvm/arm/vgic/vgic-its.c #define REGISTER_ITS_DESC_UACCESS(off, rd, wr, uwr, length, acc)\ wr 1750 virt/kvm/arm/vgic/vgic-its.c .its_write = wr, \ wr 448 virt/kvm/arm/vgic/vgic-mmio-v3.c #define REGISTER_DESC_WITH_BITS_PER_IRQ_SHARED(off, rd, wr, ur, uw, bpi, acc) \ wr 462 virt/kvm/arm/vgic/vgic-mmio-v3.c .write = wr, \ wr 67 virt/kvm/arm/vgic/vgic-mmio.h #define REGISTER_DESC_WITH_BITS_PER_IRQ(off, rd, wr, ur, uw, bpi, acc) \ wr 74 virt/kvm/arm/vgic/vgic-mmio.h .write = wr, \ wr 79 virt/kvm/arm/vgic/vgic-mmio.h #define REGISTER_DESC_WITH_LENGTH(off, rd, wr, length, acc) \ wr 86 virt/kvm/arm/vgic/vgic-mmio.h .write = wr, \ wr 89 virt/kvm/arm/vgic/vgic-mmio.h #define REGISTER_DESC_WITH_LENGTH_UACCESS(off, rd, wr, urd, uwr, length, acc) \ wr 96 virt/kvm/arm/vgic/vgic-mmio.h .write = wr, \