wr                187 arch/mips/include/asm/pci/bridge.h 		u64	wr;			/* write-only */
wr                 39 arch/mips/include/uapi/asm/ucontext.h 	unsigned long long	wr[32];
wr                189 arch/mips/kernel/signal.c 		err |= _save_msa_all_upper(&msa->wr);
wr                199 arch/mips/kernel/signal.c 			err |= __put_user(val, &msa->wr[i]);
wr                236 arch/mips/kernel/signal.c 		err |= _restore_msa_all_upper(&msa->wr);
wr                244 arch/mips/kernel/signal.c 			err |= __get_user(val, &msa->wr[i]);
wr               2050 arch/mips/mm/tlbex.c 	struct work_registers wr = build_get_work_registers(p);
wr               2053 arch/mips/mm/tlbex.c 	build_get_pmde64(p, l, r, wr.r1, wr.r2); /* get pmd in ptr */
wr               2055 arch/mips/mm/tlbex.c 	build_get_pgde32(p, wr.r1, wr.r2); /* get pgd in ptr */
wr               2064 arch/mips/mm/tlbex.c 	build_is_huge_pte(p, r, wr.r1, wr.r2, label_tlb_huge_update);
wr               2067 arch/mips/mm/tlbex.c 	UASM_i_MFC0(p, wr.r1, C0_BADVADDR);
wr               2068 arch/mips/mm/tlbex.c 	UASM_i_LW(p, wr.r2, 0, wr.r2);
wr               2069 arch/mips/mm/tlbex.c 	UASM_i_SRL(p, wr.r1, wr.r1, PAGE_SHIFT + PTE_ORDER - PTE_T_LOG2);
wr               2070 arch/mips/mm/tlbex.c 	uasm_i_andi(p, wr.r1, wr.r1, (PTRS_PER_PTE - 1) << PTE_T_LOG2);
wr               2071 arch/mips/mm/tlbex.c 	UASM_i_ADDU(p, wr.r2, wr.r2, wr.r1);
wr               2076 arch/mips/mm/tlbex.c 	iPTE_LW(p, wr.r1, wr.r2); /* get even pte */
wr               2082 arch/mips/mm/tlbex.c 			uasm_i_mfc0(p, wr.r3, C0_INDEX);
wr               2083 arch/mips/mm/tlbex.c 			uasm_il_bltz(p, r, wr.r3, label_leave);
wr               2087 arch/mips/mm/tlbex.c 	return wr;
wr               2113 arch/mips/mm/tlbex.c 	struct work_registers wr;
wr               2133 arch/mips/mm/tlbex.c 	wr = build_r4000_tlbchange_handler_head(&p, &l, &r);
wr               2134 arch/mips/mm/tlbex.c 	build_pte_present(&p, &r, wr.r1, wr.r2, wr.r3, label_nopage_tlbl);
wr               2144 arch/mips/mm/tlbex.c 			uasm_il_bbit0(&p, &r, wr.r1, ilog2(_PAGE_VALID),
wr               2147 arch/mips/mm/tlbex.c 			uasm_i_andi(&p, wr.r3, wr.r1, _PAGE_VALID);
wr               2148 arch/mips/mm/tlbex.c 			uasm_il_beqz(&p, &r, wr.r3, label_tlbl_goaround1);
wr               2176 arch/mips/mm/tlbex.c 			uasm_i_bbit0(&p, wr.r2, ilog2(sizeof(pte_t)), 8);
wr               2178 arch/mips/mm/tlbex.c 			uasm_i_andi(&p, wr.r3, wr.r2, sizeof(pte_t));
wr               2179 arch/mips/mm/tlbex.c 			uasm_i_beqz(&p, wr.r3, 8);
wr               2182 arch/mips/mm/tlbex.c 		UASM_i_MFC0(&p, wr.r3, C0_ENTRYLO0);
wr               2184 arch/mips/mm/tlbex.c 		UASM_i_MFC0(&p, wr.r3, C0_ENTRYLO1);
wr               2190 arch/mips/mm/tlbex.c 			uasm_il_bbit1(&p, &r, wr.r3, 1, label_nopage_tlbl);
wr               2194 arch/mips/mm/tlbex.c 			uasm_i_andi(&p, wr.r3, wr.r3, 2);
wr               2195 arch/mips/mm/tlbex.c 			uasm_il_bnez(&p, &r, wr.r3, label_nopage_tlbl);
wr               2200 arch/mips/mm/tlbex.c 	build_make_valid(&p, &r, wr.r1, wr.r2, wr.r3);
wr               2201 arch/mips/mm/tlbex.c 	build_r4000_tlbchange_handler_tail(&p, &l, &r, wr.r1, wr.r2);
wr               2209 arch/mips/mm/tlbex.c 	iPTE_LW(&p, wr.r1, wr.r2);
wr               2210 arch/mips/mm/tlbex.c 	build_pte_present(&p, &r, wr.r1, wr.r2, wr.r3, label_nopage_tlbl);
wr               2219 arch/mips/mm/tlbex.c 			uasm_il_bbit0(&p, &r, wr.r1, ilog2(_PAGE_VALID),
wr               2222 arch/mips/mm/tlbex.c 			uasm_i_andi(&p, wr.r3, wr.r1, _PAGE_VALID);
wr               2223 arch/mips/mm/tlbex.c 			uasm_il_beqz(&p, &r, wr.r3, label_tlbl_goaround2);
wr               2251 arch/mips/mm/tlbex.c 			uasm_i_bbit0(&p, wr.r2, ilog2(sizeof(pte_t)), 8);
wr               2253 arch/mips/mm/tlbex.c 			uasm_i_andi(&p, wr.r3, wr.r2, sizeof(pte_t));
wr               2254 arch/mips/mm/tlbex.c 			uasm_i_beqz(&p, wr.r3, 8);
wr               2257 arch/mips/mm/tlbex.c 		UASM_i_MFC0(&p, wr.r3, C0_ENTRYLO0);
wr               2259 arch/mips/mm/tlbex.c 		UASM_i_MFC0(&p, wr.r3, C0_ENTRYLO1);
wr               2265 arch/mips/mm/tlbex.c 			uasm_il_bbit0(&p, &r, wr.r3, 1, label_tlbl_goaround2);
wr               2267 arch/mips/mm/tlbex.c 			uasm_i_andi(&p, wr.r3, wr.r3, 2);
wr               2268 arch/mips/mm/tlbex.c 			uasm_il_beqz(&p, &r, wr.r3, label_tlbl_goaround2);
wr               2276 arch/mips/mm/tlbex.c 		build_restore_pagemask(&p, &r, wr.r3, label_nopage_tlbl, 0);
wr               2280 arch/mips/mm/tlbex.c 	uasm_i_ori(&p, wr.r1, wr.r1, (_PAGE_ACCESSED | _PAGE_VALID));
wr               2281 arch/mips/mm/tlbex.c 	build_huge_handler_tail(&p, &r, &l, wr.r1, wr.r2, 1);
wr               2313 arch/mips/mm/tlbex.c 	struct work_registers wr;
wr               2319 arch/mips/mm/tlbex.c 	wr = build_r4000_tlbchange_handler_head(&p, &l, &r);
wr               2320 arch/mips/mm/tlbex.c 	build_pte_writable(&p, &r, wr.r1, wr.r2, wr.r3, label_nopage_tlbs);
wr               2323 arch/mips/mm/tlbex.c 	build_make_write(&p, &r, wr.r1, wr.r2, wr.r3);
wr               2324 arch/mips/mm/tlbex.c 	build_r4000_tlbchange_handler_tail(&p, &l, &r, wr.r1, wr.r2);
wr               2332 arch/mips/mm/tlbex.c 	iPTE_LW(&p, wr.r1, wr.r2);
wr               2333 arch/mips/mm/tlbex.c 	build_pte_writable(&p, &r, wr.r1, wr.r2, wr.r3, label_nopage_tlbs);
wr               2335 arch/mips/mm/tlbex.c 	uasm_i_ori(&p, wr.r1, wr.r1,
wr               2337 arch/mips/mm/tlbex.c 	build_huge_handler_tail(&p, &r, &l, wr.r1, wr.r2, 1);
wr               2369 arch/mips/mm/tlbex.c 	struct work_registers wr;
wr               2375 arch/mips/mm/tlbex.c 	wr = build_r4000_tlbchange_handler_head(&p, &l, &r);
wr               2376 arch/mips/mm/tlbex.c 	build_pte_modifiable(&p, &r, wr.r1, wr.r2, wr.r3, label_nopage_tlbm);
wr               2380 arch/mips/mm/tlbex.c 	build_make_write(&p, &r, wr.r1, wr.r2, wr.r3);
wr               2381 arch/mips/mm/tlbex.c 	build_r4000_tlbchange_handler_tail(&p, &l, &r, wr.r1, wr.r2);
wr               2389 arch/mips/mm/tlbex.c 	iPTE_LW(&p, wr.r1, wr.r2);
wr               2390 arch/mips/mm/tlbex.c 	build_pte_modifiable(&p, &r, wr.r1, wr.r2,  wr.r3, label_nopage_tlbm);
wr               2392 arch/mips/mm/tlbex.c 	uasm_i_ori(&p, wr.r1, wr.r1,
wr               2394 arch/mips/mm/tlbex.c 	build_huge_handler_tail(&p, &r, &l, wr.r1, wr.r2, 0);
wr                 63 arch/sparc/include/asm/backoff.h 	wr		tmp, 0, %asr27;		\
wr                 23 arch/sparc/include/asm/visasm.h 297:	wr		%g0, FPRS_FEF, %fprs;		\
wr                 26 arch/sparc/include/asm/visasm.h 	wr		%g0, 0, %fprs;
wr                 43 arch/sparc/include/asm/visasm.h 297:	wr		%o5, FPRS_FEF, %fprs;
wr                 46 arch/sparc/include/asm/visasm.h 	wr		%o5, 0, %fprs;
wr                 52 arch/sparc/include/asm/winmacro.h         wr      %scratch, 0x0, %y;
wr               4208 arch/x86/kvm/vmx/nested.c 			u32 vmx_instruction_info, bool wr, int len, gva_t *ret)
wr               4292 arch/x86/kvm/vmx/nested.c 		if (wr)
wr                 34 arch/x86/kvm/vmx/nested.h 			u32 vmx_instruction_info, bool wr, int len, gva_t *ret);
wr                169 drivers/crypto/atmel-sha.c static const char *atmel_sha_reg_name(u32 offset, char *tmp, size_t sz, bool wr)
wr                231 drivers/crypto/atmel-sha.c 		if (wr)
wr                529 drivers/crypto/chelsio/chcr_ipsec.c 	struct chcr_ipsec_wr *wr;
wr                564 drivers/crypto/chelsio/chcr_ipsec.c 	wr = (struct chcr_ipsec_wr *)pos;
wr                565 drivers/crypto/chelsio/chcr_ipsec.c 	wr->wreq.op_to_compl = htonl(FW_WR_OP_V(FW_ULPTX_WR));
wr                575 drivers/crypto/chelsio/chcr_ipsec.c 	wr->wreq.flowid_len16 = htonl(wr_mid);
wr                578 drivers/crypto/chelsio/chcr_ipsec.c 	wr->req.ulptx.cmd_dest = FILL_ULPTX_CMD_DEST(pi->port_id, qid);
wr                579 drivers/crypto/chelsio/chcr_ipsec.c 	wr->req.ulptx.len = htonl(ndesc - 1);
wr                582 drivers/crypto/chelsio/chcr_ipsec.c 	wr->req.sc_imm.cmd_more = FILL_CMD_MORE(!immdatalen || sc_more);
wr                583 drivers/crypto/chelsio/chcr_ipsec.c 	wr->req.sc_imm.len = cpu_to_be32(sizeof(struct cpl_tx_sec_pdu) +
wr                584 drivers/crypto/chelsio/chcr_ipsec.c 					 sizeof(wr->req.key_ctx) +
wr                594 drivers/crypto/chelsio/chcr_ipsec.c 	wr->req.sec_cpl.op_ivinsrtofst = htonl(
wr                601 drivers/crypto/chelsio/chcr_ipsec.c 	wr->req.sec_cpl.pldlen = htonl(skb->len + esnlen);
wr                610 drivers/crypto/chelsio/chcr_ipsec.c 	wr->req.sec_cpl.aadstart_cipherstop_hi = FILL_SEC_CPL_CIPHERSTOP_HI(
wr                615 drivers/crypto/chelsio/chcr_ipsec.c 	wr->req.sec_cpl.cipherstop_lo_authinsert =
wr                619 drivers/crypto/chelsio/chcr_ipsec.c 	wr->req.sec_cpl.seqno_numivs =
wr                625 drivers/crypto/chelsio/chcr_ipsec.c 	wr->req.sec_cpl.ivgen_hdrlen =  FILL_SEC_CPL_IVGEN_HDRLEN(0, 0, 1,
wr                305 drivers/crypto/chelsio/chtls/chtls.h 	struct tls_key_wr wr;
wr                329 drivers/crypto/chelsio/chtls/chtls.h 	struct wr_skb_cb wr;		/* reserve space for write request */
wr                270 drivers/crypto/chelsio/chtls/chtls_cm.c 	req->wr.wr_hi = htonl(FW_WR_OP_V(FW_TP_WR) |
wr                272 drivers/crypto/chelsio/chtls/chtls_cm.c 					      sizeof(req->wr)));
wr                273 drivers/crypto/chelsio/chtls/chtls_cm.c 	req->wr.wr_mid = htonl(FW_WR_LEN16_V(DIV_ROUND_UP(sizeof(*req), 16)) |
wr                 30 drivers/crypto/chelsio/chtls/chtls_hw.c 	req->wr.wr_mid |= htonl(FW_WR_FLOWID_V(csk->tid));
wr                330 drivers/crypto/chelsio/chtls/chtls_hw.c 	kwr->wr.op_to_compl =
wr                333 drivers/crypto/chelsio/chtls/chtls_hw.c 	kwr->wr.flowid_len16 =
wr                336 drivers/crypto/chelsio/chtls/chtls_hw.c 	kwr->wr.protocol = 0;
wr                337 drivers/crypto/chelsio/chtls/chtls_hw.c 	kwr->wr.mfs = htons(TLS_MFS);
wr                338 drivers/crypto/chelsio/chtls/chtls_hw.c 	kwr->wr.reneg_to_write_rx = optname;
wr                345 drivers/crypto/chelsio/chtls/chtls_hw.c 			      DIV_ROUND_UP(len - sizeof(kwr->wr), 16));
wr                699 drivers/crypto/chelsio/chtls/chtls_io.c 				req->wr.wr_hi |= htonl(FW_WR_COMPL_F);
wr                 63 drivers/dma/dw-edma/dw-edma-v0-core.c 		return &__dw_regs(dw)->type.unroll.ch[ch].wr;
wr                190 drivers/dma/dw-edma/dw-edma-v0-debugfs.c 		dw_edma_debugfs_regs_ch(&regs->type.unroll.ch[i].wr, ch_dir);
wr                192 drivers/dma/dw-edma/dw-edma-v0-debugfs.c 		lim[0][i].start = &regs->type.unroll.ch[i].wr;
wr                 40 drivers/dma/dw-edma/dw-edma-v0-regs.h 	struct dw_edma_v0_ch_regs wr;			/* 0x200 */
wr                126 drivers/gpu/drm/nouveau/nvif/object.c 		struct nvif_ioctl_wr_v0 wr;
wr                129 drivers/gpu/drm/nouveau/nvif/object.c 		.wr.size = size,
wr                130 drivers/gpu/drm/nouveau/nvif/object.c 		.wr.addr = addr,
wr                131 drivers/gpu/drm/nouveau/nvif/object.c 		.wr.data = data,
wr                 69 drivers/gpu/drm/udl/udl_dmabuf.c 	struct scatterlist *rd, *wr;
wr                108 drivers/gpu/drm/udl/udl_dmabuf.c 	wr = sgt->sgl;
wr                110 drivers/gpu/drm/udl/udl_dmabuf.c 		sg_set_page(wr, sg_page(rd), rd->length, rd->offset);
wr                112 drivers/gpu/drm/udl/udl_dmabuf.c 		wr = sg_next(wr);
wr                511 drivers/i2c/busses/i2c-ocores.c 	u32 curr, wr;
wr                519 drivers/i2c/busses/i2c-ocores.c 			wr = (curr & 0xff00) | value;
wr                521 drivers/i2c/busses/i2c-ocores.c 			wr = (((u32)value) << 8) | (curr & 0xff);
wr                523 drivers/i2c/busses/i2c-ocores.c 		wr = value;
wr                525 drivers/i2c/busses/i2c-ocores.c 	iowrite32be(wr, i2c->base + (rreg << i2c->reg_shift));
wr               1460 drivers/ide/pmac.c 	int wr = !!(cmd->tf_flags & IDE_TFLAG_WRITE);
wr               1496 drivers/ide/pmac.c 			table->command = cpu_to_le16(wr? OUTPUT_MORE: INPUT_MORE);
wr               1512 drivers/ide/pmac.c 		table[-1].command = cpu_to_le16(wr? OUTPUT_LAST: INPUT_LAST);
wr                 65 drivers/infiniband/core/mad.c 	struct ib_ud_wr *wr = &mad_send_wr->send_wr;
wr                 68 drivers/infiniband/core/mad.c 	rdma_query_ah(wr->ah, &attr);
wr                 72 drivers/infiniband/core/mad.c 	ib_query_pkey(dev, pnum, wr->pkey_index, &pkey);
wr                 74 drivers/infiniband/core/mad.c 	entry->rqpn = wr->remote_qpn;
wr                 75 drivers/infiniband/core/mad.c 	entry->rqkey = wr->remote_qkey;
wr                904 drivers/infiniband/core/mad.c 		     send_wr->wr.wr_cqe, drslid,
wr               1110 drivers/infiniband/core/mad.c 	mad_send_wr->send_wr.wr.wr_cqe = &mad_send_wr->mad_list.cqe;
wr               1111 drivers/infiniband/core/mad.c 	mad_send_wr->send_wr.wr.sg_list = mad_send_wr->sg_list;
wr               1112 drivers/infiniband/core/mad.c 	mad_send_wr->send_wr.wr.num_sge = 2;
wr               1113 drivers/infiniband/core/mad.c 	mad_send_wr->send_wr.wr.opcode = IB_WR_SEND;
wr               1114 drivers/infiniband/core/mad.c 	mad_send_wr->send_wr.wr.send_flags = IB_SEND_SIGNALED;
wr               1223 drivers/infiniband/core/mad.c 	mad_send_wr->send_wr.wr.wr_cqe = &mad_send_wr->mad_list.cqe;
wr               1251 drivers/infiniband/core/mad.c 		ret = ib_post_send(mad_agent->qp, &mad_send_wr->send_wr.wr,
wr               1901 drivers/infiniband/core/mad.c static inline int rcv_has_same_class(const struct ib_mad_send_wr_private *wr,
wr               1904 drivers/infiniband/core/mad.c 	return ((struct ib_mad_hdr *)(wr->send_buf.mad))->mgmt_class ==
wr               1909 drivers/infiniband/core/mad.c 				   const struct ib_mad_send_wr_private *wr,
wr               1920 drivers/infiniband/core/mad.c 	send_resp = ib_response_mad((struct ib_mad_hdr *)wr->send_buf.mad);
wr               1927 drivers/infiniband/core/mad.c 	if (rdma_query_ah(wr->send_buf.ah, &attr))
wr               1973 drivers/infiniband/core/mad.c 	struct ib_mad_send_wr_private *wr;
wr               1978 drivers/infiniband/core/mad.c 	list_for_each_entry(wr, &mad_agent_priv->wait_list, agent_list) {
wr               1979 drivers/infiniband/core/mad.c 		if ((wr->tid == mad_hdr->tid) &&
wr               1980 drivers/infiniband/core/mad.c 		    rcv_has_same_class(wr, wc) &&
wr               1986 drivers/infiniband/core/mad.c 		     rcv_has_same_gid(mad_agent_priv, wr, wc)))
wr               1987 drivers/infiniband/core/mad.c 			return (wr->status == IB_WC_SUCCESS) ? wr : NULL;
wr               1994 drivers/infiniband/core/mad.c 	list_for_each_entry(wr, &mad_agent_priv->send_list, agent_list) {
wr               1995 drivers/infiniband/core/mad.c 		if (is_rmpp_data_mad(mad_agent_priv, wr->send_buf.mad) &&
wr               1996 drivers/infiniband/core/mad.c 		    wr->tid == mad_hdr->tid &&
wr               1997 drivers/infiniband/core/mad.c 		    wr->timeout &&
wr               1998 drivers/infiniband/core/mad.c 		    rcv_has_same_class(wr, wc) &&
wr               2004 drivers/infiniband/core/mad.c 		     rcv_has_same_gid(mad_agent_priv, wr, wc)))
wr               2006 drivers/infiniband/core/mad.c 			return (wr->status == IB_WC_SUCCESS) ? wr : NULL;
wr               2567 drivers/infiniband/core/mad.c 		ret = ib_post_send(qp_info->qp, &queued_send_wr->send_wr.wr,
wr               2615 drivers/infiniband/core/mad.c 			ret = ib_post_send(qp_info->qp, &mad_send_wr->send_wr.wr,
wr               2782 drivers/infiniband/core/mad.c 				     local->mad_send_wr->send_wr.wr.wr_cqe,
wr                635 drivers/infiniband/core/mad_rmpp.c static inline void adjust_last_ack(struct ib_mad_send_wr_private *wr,
wr                640 drivers/infiniband/core/mad_rmpp.c 	wr->last_ack = seg_num;
wr                641 drivers/infiniband/core/mad_rmpp.c 	list = &wr->last_ack_seg->list;
wr                642 drivers/infiniband/core/mad_rmpp.c 	list_for_each_entry(wr->last_ack_seg, list, list)
wr                643 drivers/infiniband/core/mad_rmpp.c 		if (wr->last_ack_seg->num == seg_num)
wr                 75 drivers/infiniband/core/rw.c 		reg->inv_wr.next = &reg->reg_wr.wr;
wr                106 drivers/infiniband/core/rw.c 	reg->reg_wr.wr.opcode = IB_WR_REG_MR;
wr                146 drivers/infiniband/core/rw.c 				prev->wr.wr.next = &reg->inv_wr;
wr                148 drivers/infiniband/core/rw.c 				prev->wr.wr.next = &reg->reg_wr.wr;
wr                151 drivers/infiniband/core/rw.c 		reg->reg_wr.wr.next = &reg->wr.wr;
wr                153 drivers/infiniband/core/rw.c 		reg->wr.wr.sg_list = &reg->sge;
wr                154 drivers/infiniband/core/rw.c 		reg->wr.wr.num_sge = 1;
wr                155 drivers/infiniband/core/rw.c 		reg->wr.remote_addr = remote_addr;
wr                156 drivers/infiniband/core/rw.c 		reg->wr.rkey = rkey;
wr                158 drivers/infiniband/core/rw.c 			reg->wr.wr.opcode = IB_WR_RDMA_WRITE;
wr                160 drivers/infiniband/core/rw.c 			reg->wr.wr.opcode = IB_WR_RDMA_READ;
wr                162 drivers/infiniband/core/rw.c 			reg->wr.wr.opcode = IB_WR_RDMA_READ_WITH_INV;
wr                163 drivers/infiniband/core/rw.c 			reg->wr.wr.ex.invalidate_rkey = reg->mr->lkey;
wr                176 drivers/infiniband/core/rw.c 		prev->wr.wr.next = NULL;
wr                213 drivers/infiniband/core/rw.c 			rdma_wr->wr.opcode = IB_WR_RDMA_WRITE;
wr                215 drivers/infiniband/core/rw.c 			rdma_wr->wr.opcode = IB_WR_RDMA_READ;
wr                218 drivers/infiniband/core/rw.c 		rdma_wr->wr.num_sge = nr_sge;
wr                219 drivers/infiniband/core/rw.c 		rdma_wr->wr.sg_list = sge;
wr                232 drivers/infiniband/core/rw.c 		rdma_wr->wr.next = i + 1 < ctx->nr_ops ?
wr                233 drivers/infiniband/core/rw.c 			&ctx->map.wrs[i + 1].wr : NULL;
wr                249 drivers/infiniband/core/rw.c 	struct ib_rdma_wr *rdma_wr = &ctx->single.wr;
wr                259 drivers/infiniband/core/rw.c 		rdma_wr->wr.opcode = IB_WR_RDMA_WRITE;
wr                261 drivers/infiniband/core/rw.c 		rdma_wr->wr.opcode = IB_WR_RDMA_READ;
wr                262 drivers/infiniband/core/rw.c 	rdma_wr->wr.sg_list = &ctx->single.sge;
wr                263 drivers/infiniband/core/rw.c 	rdma_wr->wr.num_sge = 1;
wr                428 drivers/infiniband/core/rw.c 	ctx->reg->reg_wr.wr.opcode = IB_WR_REG_MR_INTEGRITY;
wr                429 drivers/infiniband/core/rw.c 	ctx->reg->reg_wr.wr.wr_cqe = NULL;
wr                430 drivers/infiniband/core/rw.c 	ctx->reg->reg_wr.wr.num_sge = 0;
wr                431 drivers/infiniband/core/rw.c 	ctx->reg->reg_wr.wr.send_flags = 0;
wr                444 drivers/infiniband/core/rw.c 	rdma_wr = &ctx->reg->wr;
wr                445 drivers/infiniband/core/rw.c 	rdma_wr->wr.sg_list = &ctx->reg->sge;
wr                446 drivers/infiniband/core/rw.c 	rdma_wr->wr.num_sge = 1;
wr                450 drivers/infiniband/core/rw.c 		rdma_wr->wr.opcode = IB_WR_RDMA_WRITE;
wr                452 drivers/infiniband/core/rw.c 		rdma_wr->wr.opcode = IB_WR_RDMA_READ;
wr                453 drivers/infiniband/core/rw.c 	ctx->reg->reg_wr.wr.next = &rdma_wr->wr;
wr                511 drivers/infiniband/core/rw.c 				ctx->reg[i].wr.wr.opcode !=
wr                518 drivers/infiniband/core/rw.c 			first_wr = &ctx->reg[0].reg_wr.wr;
wr                519 drivers/infiniband/core/rw.c 		last_wr = &ctx->reg[ctx->nr_ops - 1].wr.wr;
wr                522 drivers/infiniband/core/rw.c 		first_wr = &ctx->map.wrs[0].wr;
wr                523 drivers/infiniband/core/rw.c 		last_wr = &ctx->map.wrs[ctx->nr_ops - 1].wr;
wr                526 drivers/infiniband/core/rw.c 		first_wr = &ctx->single.wr.wr;
wr                527 drivers/infiniband/core/rw.c 		last_wr = &ctx->single.wr.wr;
wr               2004 drivers/infiniband/core/uverbs_cmd.c 	struct ib_send_wr              *wr = NULL, *last, *next;
wr               2071 drivers/infiniband/core/uverbs_cmd.c 						   user_wr->wr.ud.ah, attrs);
wr               2077 drivers/infiniband/core/uverbs_cmd.c 			ud->remote_qpn = user_wr->wr.ud.remote_qpn;
wr               2078 drivers/infiniband/core/uverbs_cmd.c 			ud->remote_qkey = user_wr->wr.ud.remote_qkey;
wr               2080 drivers/infiniband/core/uverbs_cmd.c 			next = &ud->wr;
wr               2093 drivers/infiniband/core/uverbs_cmd.c 			rdma->remote_addr = user_wr->wr.rdma.remote_addr;
wr               2094 drivers/infiniband/core/uverbs_cmd.c 			rdma->rkey = user_wr->wr.rdma.rkey;
wr               2096 drivers/infiniband/core/uverbs_cmd.c 			next = &rdma->wr;
wr               2108 drivers/infiniband/core/uverbs_cmd.c 			atomic->remote_addr = user_wr->wr.atomic.remote_addr;
wr               2109 drivers/infiniband/core/uverbs_cmd.c 			atomic->compare_add = user_wr->wr.atomic.compare_add;
wr               2110 drivers/infiniband/core/uverbs_cmd.c 			atomic->swap = user_wr->wr.atomic.swap;
wr               2111 drivers/infiniband/core/uverbs_cmd.c 			atomic->rkey = user_wr->wr.atomic.rkey;
wr               2113 drivers/infiniband/core/uverbs_cmd.c 			next = &atomic->wr;
wr               2137 drivers/infiniband/core/uverbs_cmd.c 			wr = next;
wr               2163 drivers/infiniband/core/uverbs_cmd.c 	ret = qp->device->ops.post_send(qp->real_qp, wr, &bad_wr);
wr               2165 drivers/infiniband/core/uverbs_cmd.c 		for (next = wr; next; next = next->next) {
wr               2178 drivers/infiniband/core/uverbs_cmd.c 	while (wr) {
wr               2179 drivers/infiniband/core/uverbs_cmd.c 		if (is_ud && ud_wr(wr)->ah)
wr               2180 drivers/infiniband/core/uverbs_cmd.c 			uobj_put_obj_read(ud_wr(wr)->ah);
wr               2181 drivers/infiniband/core/uverbs_cmd.c 		next = wr->next;
wr               2182 drivers/infiniband/core/uverbs_cmd.c 		kfree(wr);
wr               2183 drivers/infiniband/core/uverbs_cmd.c 		wr = next;
wr               2197 drivers/infiniband/core/uverbs_cmd.c 	struct ib_recv_wr        *wr = NULL, *last, *next;
wr               2252 drivers/infiniband/core/uverbs_cmd.c 			wr = next;
wr               2276 drivers/infiniband/core/uverbs_cmd.c 	return wr;
wr               2281 drivers/infiniband/core/uverbs_cmd.c 	while (wr) {
wr               2282 drivers/infiniband/core/uverbs_cmd.c 		next = wr->next;
wr               2283 drivers/infiniband/core/uverbs_cmd.c 		kfree(wr);
wr               2284 drivers/infiniband/core/uverbs_cmd.c 		wr = next;
wr               2294 drivers/infiniband/core/uverbs_cmd.c 	struct ib_recv_wr              *wr, *next;
wr               2304 drivers/infiniband/core/uverbs_cmd.c 	wr = ib_uverbs_unmarshall_recv(&iter, cmd.wr_count, cmd.wqe_size,
wr               2306 drivers/infiniband/core/uverbs_cmd.c 	if (IS_ERR(wr))
wr               2307 drivers/infiniband/core/uverbs_cmd.c 		return PTR_ERR(wr);
wr               2316 drivers/infiniband/core/uverbs_cmd.c 	ret = qp->device->ops.post_recv(qp->real_qp, wr, &bad_wr);
wr               2320 drivers/infiniband/core/uverbs_cmd.c 		for (next = wr; next; next = next->next) {
wr               2331 drivers/infiniband/core/uverbs_cmd.c 	while (wr) {
wr               2332 drivers/infiniband/core/uverbs_cmd.c 		next = wr->next;
wr               2333 drivers/infiniband/core/uverbs_cmd.c 		kfree(wr);
wr               2334 drivers/infiniband/core/uverbs_cmd.c 		wr = next;
wr               2344 drivers/infiniband/core/uverbs_cmd.c 	struct ib_recv_wr                  *wr, *next;
wr               2354 drivers/infiniband/core/uverbs_cmd.c 	wr = ib_uverbs_unmarshall_recv(&iter, cmd.wr_count, cmd.wqe_size,
wr               2356 drivers/infiniband/core/uverbs_cmd.c 	if (IS_ERR(wr))
wr               2357 drivers/infiniband/core/uverbs_cmd.c 		return PTR_ERR(wr);
wr               2366 drivers/infiniband/core/uverbs_cmd.c 	ret = srq->device->ops.post_srq_recv(srq, wr, &bad_wr);
wr               2371 drivers/infiniband/core/uverbs_cmd.c 		for (next = wr; next; next = next->next) {
wr               2382 drivers/infiniband/core/uverbs_cmd.c 	while (wr) {
wr               2383 drivers/infiniband/core/uverbs_cmd.c 		next = wr->next;
wr               2384 drivers/infiniband/core/uverbs_cmd.c 		kfree(wr);
wr               2385 drivers/infiniband/core/uverbs_cmd.c 		wr = next;
wr               2636 drivers/infiniband/core/verbs.c 		.wr = {
wr               2653 drivers/infiniband/core/verbs.c 	ret = ib_post_send(qp, &swr.wr, NULL);
wr               1474 drivers/infiniband/hw/bnxt_re/ib_verbs.c int bnxt_re_post_srq_recv(struct ib_srq *ib_srq, const struct ib_recv_wr *wr,
wr               1484 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	while (wr) {
wr               1486 drivers/infiniband/hw/bnxt_re/ib_verbs.c 		wqe.num_sge = wr->num_sge;
wr               1487 drivers/infiniband/hw/bnxt_re/ib_verbs.c 		bnxt_re_build_sgl(wr->sg_list, wqe.sg_list, wr->num_sge);
wr               1488 drivers/infiniband/hw/bnxt_re/ib_verbs.c 		wqe.wr_id = wr->wr_id;
wr               1493 drivers/infiniband/hw/bnxt_re/ib_verbs.c 			*bad_wr = wr;
wr               1496 drivers/infiniband/hw/bnxt_re/ib_verbs.c 		wr = wr->next;
wr               1836 drivers/infiniband/hw/bnxt_re/ib_verbs.c 				     const struct ib_send_wr *wr,
wr               1840 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	struct bnxt_re_ah *ah = container_of(ud_wr(wr)->ah, struct bnxt_re_ah,
wr               1937 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	if (wr->opcode == IB_WR_SEND_WITH_IMM) {
wr               1943 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	if (wr->send_flags & IB_SEND_SOLICITED)
wr               2010 drivers/infiniband/hw/bnxt_re/ib_verbs.c 					    const struct ib_recv_wr *wr,
wr               2056 drivers/infiniband/hw/bnxt_re/ib_verbs.c 				  const struct ib_send_wr *wr,
wr               2062 drivers/infiniband/hw/bnxt_re/ib_verbs.c 		ah = container_of(ud_wr(wr)->ah, struct bnxt_re_ah, ib_ah);
wr               2063 drivers/infiniband/hw/bnxt_re/ib_verbs.c 		wqe->send.q_key = ud_wr(wr)->remote_qkey;
wr               2064 drivers/infiniband/hw/bnxt_re/ib_verbs.c 		wqe->send.dst_qp = ud_wr(wr)->remote_qpn;
wr               2067 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	switch (wr->opcode) {
wr               2073 drivers/infiniband/hw/bnxt_re/ib_verbs.c 		wqe->send.imm_data = wr->ex.imm_data;
wr               2077 drivers/infiniband/hw/bnxt_re/ib_verbs.c 		wqe->send.inv_key = wr->ex.invalidate_rkey;
wr               2082 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	if (wr->send_flags & IB_SEND_SIGNALED)
wr               2084 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	if (wr->send_flags & IB_SEND_FENCE)
wr               2086 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	if (wr->send_flags & IB_SEND_SOLICITED)
wr               2088 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	if (wr->send_flags & IB_SEND_INLINE)
wr               2094 drivers/infiniband/hw/bnxt_re/ib_verbs.c static int bnxt_re_build_rdma_wqe(const struct ib_send_wr *wr,
wr               2097 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	switch (wr->opcode) {
wr               2103 drivers/infiniband/hw/bnxt_re/ib_verbs.c 		wqe->rdma.imm_data = wr->ex.imm_data;
wr               2107 drivers/infiniband/hw/bnxt_re/ib_verbs.c 		wqe->rdma.inv_key = wr->ex.invalidate_rkey;
wr               2112 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	wqe->rdma.remote_va = rdma_wr(wr)->remote_addr;
wr               2113 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	wqe->rdma.r_key = rdma_wr(wr)->rkey;
wr               2114 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	if (wr->send_flags & IB_SEND_SIGNALED)
wr               2116 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	if (wr->send_flags & IB_SEND_FENCE)
wr               2118 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	if (wr->send_flags & IB_SEND_SOLICITED)
wr               2120 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	if (wr->send_flags & IB_SEND_INLINE)
wr               2126 drivers/infiniband/hw/bnxt_re/ib_verbs.c static int bnxt_re_build_atomic_wqe(const struct ib_send_wr *wr,
wr               2129 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	switch (wr->opcode) {
wr               2132 drivers/infiniband/hw/bnxt_re/ib_verbs.c 		wqe->atomic.cmp_data = atomic_wr(wr)->compare_add;
wr               2133 drivers/infiniband/hw/bnxt_re/ib_verbs.c 		wqe->atomic.swap_data = atomic_wr(wr)->swap;
wr               2137 drivers/infiniband/hw/bnxt_re/ib_verbs.c 		wqe->atomic.cmp_data = atomic_wr(wr)->compare_add;
wr               2142 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	wqe->atomic.remote_va = atomic_wr(wr)->remote_addr;
wr               2143 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	wqe->atomic.r_key = atomic_wr(wr)->rkey;
wr               2144 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	if (wr->send_flags & IB_SEND_SIGNALED)
wr               2146 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	if (wr->send_flags & IB_SEND_FENCE)
wr               2148 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	if (wr->send_flags & IB_SEND_SOLICITED)
wr               2153 drivers/infiniband/hw/bnxt_re/ib_verbs.c static int bnxt_re_build_inv_wqe(const struct ib_send_wr *wr,
wr               2157 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	wqe->local_inv.inv_l_key = wr->ex.invalidate_rkey;
wr               2164 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	if (wr->send_flags & IB_SEND_SIGNALED)
wr               2166 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	if (wr->send_flags & IB_SEND_SOLICITED)
wr               2172 drivers/infiniband/hw/bnxt_re/ib_verbs.c static int bnxt_re_build_reg_wqe(const struct ib_reg_wr *wr,
wr               2175 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	struct bnxt_re_mr *mr = container_of(wr->mr, struct bnxt_re_mr, ib_mr);
wr               2177 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	int access = wr->access;
wr               2192 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	if (wr->wr.send_flags & IB_SEND_SIGNALED)
wr               2206 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	wqe->frmr.l_key = wr->key;
wr               2207 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	wqe->frmr.length = wr->mr->length;
wr               2208 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	wqe->frmr.pbl_pg_sz_log = (wr->mr->page_size >> PAGE_SHIFT_4K) - 1;
wr               2209 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	wqe->frmr.va = wr->mr->iova;
wr               2214 drivers/infiniband/hw/bnxt_re/ib_verbs.c 				    const struct ib_send_wr *wr,
wr               2223 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	for (i = 0; i < wr->num_sge; i++) {
wr               2225 drivers/infiniband/hw/bnxt_re/ib_verbs.c 				wr->sg_list[i].addr;
wr               2226 drivers/infiniband/hw/bnxt_re/ib_verbs.c 		sge_len = wr->sg_list[i].length;
wr               2234 drivers/infiniband/hw/bnxt_re/ib_verbs.c 		sge_len = wr->sg_list[i].length;
wr               2237 drivers/infiniband/hw/bnxt_re/ib_verbs.c 		in_data += wr->sg_list[i].length;
wr               2238 drivers/infiniband/hw/bnxt_re/ib_verbs.c 		wqe->inline_len += wr->sg_list[i].length;
wr               2244 drivers/infiniband/hw/bnxt_re/ib_verbs.c 				   const struct ib_send_wr *wr,
wr               2249 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	if (wr->send_flags & IB_SEND_INLINE)
wr               2250 drivers/infiniband/hw/bnxt_re/ib_verbs.c 		payload_sz = bnxt_re_copy_inline_data(rdev, wr, wqe);
wr               2252 drivers/infiniband/hw/bnxt_re/ib_verbs.c 		payload_sz = bnxt_re_build_sgl(wr->sg_list, wqe->sg_list,
wr               2276 drivers/infiniband/hw/bnxt_re/ib_verbs.c 				       const struct ib_send_wr *wr)
wr               2284 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	while (wr) {
wr               2289 drivers/infiniband/hw/bnxt_re/ib_verbs.c 		wqe.num_sge = wr->num_sge;
wr               2290 drivers/infiniband/hw/bnxt_re/ib_verbs.c 		if (wr->num_sge > qp->qplib_qp.sq.max_sge) {
wr               2297 drivers/infiniband/hw/bnxt_re/ib_verbs.c 		payload_sz = bnxt_re_copy_wr_payload(qp->rdev, wr, &wqe);
wr               2302 drivers/infiniband/hw/bnxt_re/ib_verbs.c 		wqe.wr_id = wr->wr_id;
wr               2306 drivers/infiniband/hw/bnxt_re/ib_verbs.c 		rc = bnxt_re_build_send_wqe(qp, wr, &wqe);
wr               2313 drivers/infiniband/hw/bnxt_re/ib_verbs.c 				wr->opcode, rc);
wr               2316 drivers/infiniband/hw/bnxt_re/ib_verbs.c 		wr = wr->next;
wr               2324 drivers/infiniband/hw/bnxt_re/ib_verbs.c int bnxt_re_post_send(struct ib_qp *ib_qp, const struct ib_send_wr *wr,
wr               2333 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	while (wr) {
wr               2338 drivers/infiniband/hw/bnxt_re/ib_verbs.c 		wqe.num_sge = wr->num_sge;
wr               2339 drivers/infiniband/hw/bnxt_re/ib_verbs.c 		if (wr->num_sge > qp->qplib_qp.sq.max_sge) {
wr               2346 drivers/infiniband/hw/bnxt_re/ib_verbs.c 		payload_sz = bnxt_re_copy_wr_payload(qp->rdev, wr, &wqe);
wr               2351 drivers/infiniband/hw/bnxt_re/ib_verbs.c 		wqe.wr_id = wr->wr_id;
wr               2353 drivers/infiniband/hw/bnxt_re/ib_verbs.c 		switch (wr->opcode) {
wr               2357 drivers/infiniband/hw/bnxt_re/ib_verbs.c 				rc = bnxt_re_build_qp1_send_v2(qp, wr, &wqe,
wr               2364 drivers/infiniband/hw/bnxt_re/ib_verbs.c 			switch (wr->send_flags) {
wr               2374 drivers/infiniband/hw/bnxt_re/ib_verbs.c 			rc = bnxt_re_build_send_wqe(qp, wr, &wqe);
wr               2379 drivers/infiniband/hw/bnxt_re/ib_verbs.c 			rc = bnxt_re_build_rdma_wqe(wr, &wqe);
wr               2383 drivers/infiniband/hw/bnxt_re/ib_verbs.c 			rc = bnxt_re_build_atomic_wqe(wr, &wqe);
wr               2391 drivers/infiniband/hw/bnxt_re/ib_verbs.c 			rc = bnxt_re_build_inv_wqe(wr, &wqe);
wr               2394 drivers/infiniband/hw/bnxt_re/ib_verbs.c 			rc = bnxt_re_build_reg_wqe(reg_wr(wr), &wqe);
wr               2399 drivers/infiniband/hw/bnxt_re/ib_verbs.c 				"WR (%#x) is not supported", wr->opcode);
wr               2409 drivers/infiniband/hw/bnxt_re/ib_verbs.c 				wr->opcode, qp->qplib_qp.state, rc);
wr               2410 drivers/infiniband/hw/bnxt_re/ib_verbs.c 			*bad_wr = wr;
wr               2413 drivers/infiniband/hw/bnxt_re/ib_verbs.c 		wr = wr->next;
wr               2424 drivers/infiniband/hw/bnxt_re/ib_verbs.c 				       const struct ib_recv_wr *wr)
wr               2430 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	while (wr) {
wr               2435 drivers/infiniband/hw/bnxt_re/ib_verbs.c 		wqe.num_sge = wr->num_sge;
wr               2436 drivers/infiniband/hw/bnxt_re/ib_verbs.c 		if (wr->num_sge > qp->qplib_qp.rq.max_sge) {
wr               2442 drivers/infiniband/hw/bnxt_re/ib_verbs.c 		bnxt_re_build_sgl(wr->sg_list, wqe.sg_list, wr->num_sge);
wr               2443 drivers/infiniband/hw/bnxt_re/ib_verbs.c 		wqe.wr_id = wr->wr_id;
wr               2450 drivers/infiniband/hw/bnxt_re/ib_verbs.c 		wr = wr->next;
wr               2457 drivers/infiniband/hw/bnxt_re/ib_verbs.c int bnxt_re_post_recv(struct ib_qp *ib_qp, const struct ib_recv_wr *wr,
wr               2467 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	while (wr) {
wr               2472 drivers/infiniband/hw/bnxt_re/ib_verbs.c 		wqe.num_sge = wr->num_sge;
wr               2473 drivers/infiniband/hw/bnxt_re/ib_verbs.c 		if (wr->num_sge > qp->qplib_qp.rq.max_sge) {
wr               2477 drivers/infiniband/hw/bnxt_re/ib_verbs.c 			*bad_wr = wr;
wr               2481 drivers/infiniband/hw/bnxt_re/ib_verbs.c 		payload_sz = bnxt_re_build_sgl(wr->sg_list, wqe.sg_list,
wr               2482 drivers/infiniband/hw/bnxt_re/ib_verbs.c 					       wr->num_sge);
wr               2483 drivers/infiniband/hw/bnxt_re/ib_verbs.c 		wqe.wr_id = wr->wr_id;
wr               2488 drivers/infiniband/hw/bnxt_re/ib_verbs.c 			rc = bnxt_re_build_qp1_shadow_qp_recv(qp, wr, &wqe,
wr               2493 drivers/infiniband/hw/bnxt_re/ib_verbs.c 			*bad_wr = wr;
wr               2503 drivers/infiniband/hw/bnxt_re/ib_verbs.c 		wr = wr->next;
wr               2876 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	swr = &udwr.wr;
wr                168 drivers/infiniband/hw/cxgb3/cxio_hal.h int cxio_peek_cq(struct t3_wq *wr, struct t3_cq *cq, int opcode);
wr                177 drivers/infiniband/hw/cxgb3/iwch_cm.c 	req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_FORWARD));
wr                192 drivers/infiniband/hw/cxgb3/iwch_cm.c 	req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_FORWARD));
wr                193 drivers/infiniband/hw/cxgb3/iwch_cm.c 	req->wr.wr_lo = htonl(V_WR_TID(ep->hwtid));
wr                213 drivers/infiniband/hw/cxgb3/iwch_cm.c 	req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_FORWARD));
wr                214 drivers/infiniband/hw/cxgb3/iwch_cm.c 	req->wr.wr_lo = htonl(V_WR_TID(ep->hwtid));
wr                400 drivers/infiniband/hw/cxgb3/iwch_cm.c 	req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_OFLD_CLOSE_CON));
wr                401 drivers/infiniband/hw/cxgb3/iwch_cm.c 	req->wr.wr_lo = htonl(V_WR_TID(ep->hwtid));
wr                419 drivers/infiniband/hw/cxgb3/iwch_cm.c 	req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_OFLD_HOST_ABORT_CON_REQ));
wr                420 drivers/infiniband/hw/cxgb3/iwch_cm.c 	req->wr.wr_lo = htonl(V_WR_TID(ep->hwtid));
wr                457 drivers/infiniband/hw/cxgb3/iwch_cm.c 	req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_FORWARD));
wr                802 drivers/infiniband/hw/cxgb3/iwch_cm.c 	req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_FORWARD));
wr               1207 drivers/infiniband/hw/cxgb3/iwch_cm.c 	req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_FORWARD));
wr               1248 drivers/infiniband/hw/cxgb3/iwch_cm.c 	req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_FORWARD));
wr               1293 drivers/infiniband/hw/cxgb3/iwch_cm.c 	rpl->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_FORWARD));
wr               1322 drivers/infiniband/hw/cxgb3/iwch_cm.c 		rpl->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_FORWARD));
wr               1616 drivers/infiniband/hw/cxgb3/iwch_cm.c 	rpl->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_OFLD_HOST_ABORT_CON_RPL));
wr               1617 drivers/infiniband/hw/cxgb3/iwch_cm.c 	rpl->wr.wr_lo = htonl(V_WR_TID(ep->hwtid));
wr                329 drivers/infiniband/hw/cxgb3/iwch_provider.h int iwch_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr,
wr                331 drivers/infiniband/hw/cxgb3/iwch_provider.h int iwch_post_receive(struct ib_qp *ibqp, const struct ib_recv_wr *wr,
wr                 42 drivers/infiniband/hw/cxgb3/iwch_qp.c static int build_rdma_send(union t3_wr *wqe, const struct ib_send_wr *wr,
wr                 48 drivers/infiniband/hw/cxgb3/iwch_qp.c 	switch (wr->opcode) {
wr                 50 drivers/infiniband/hw/cxgb3/iwch_qp.c 		if (wr->send_flags & IB_SEND_SOLICITED)
wr                 57 drivers/infiniband/hw/cxgb3/iwch_qp.c 		if (wr->send_flags & IB_SEND_SOLICITED)
wr                 61 drivers/infiniband/hw/cxgb3/iwch_qp.c 		wqe->send.rem_stag = cpu_to_be32(wr->ex.invalidate_rkey);
wr                 66 drivers/infiniband/hw/cxgb3/iwch_qp.c 	if (wr->num_sge > T3_MAX_SGE)
wr                 72 drivers/infiniband/hw/cxgb3/iwch_qp.c 	for (i = 0; i < wr->num_sge; i++) {
wr                 73 drivers/infiniband/hw/cxgb3/iwch_qp.c 		if ((plen + wr->sg_list[i].length) < plen)
wr                 76 drivers/infiniband/hw/cxgb3/iwch_qp.c 		plen += wr->sg_list[i].length;
wr                 77 drivers/infiniband/hw/cxgb3/iwch_qp.c 		wqe->send.sgl[i].stag = cpu_to_be32(wr->sg_list[i].lkey);
wr                 78 drivers/infiniband/hw/cxgb3/iwch_qp.c 		wqe->send.sgl[i].len = cpu_to_be32(wr->sg_list[i].length);
wr                 79 drivers/infiniband/hw/cxgb3/iwch_qp.c 		wqe->send.sgl[i].to = cpu_to_be64(wr->sg_list[i].addr);
wr                 81 drivers/infiniband/hw/cxgb3/iwch_qp.c 	wqe->send.num_sgle = cpu_to_be32(wr->num_sge);
wr                 82 drivers/infiniband/hw/cxgb3/iwch_qp.c 	*flit_cnt = 4 + ((wr->num_sge) << 1);
wr                 87 drivers/infiniband/hw/cxgb3/iwch_qp.c static int build_rdma_write(union t3_wr *wqe, const struct ib_send_wr *wr,
wr                 92 drivers/infiniband/hw/cxgb3/iwch_qp.c 	if (wr->num_sge > T3_MAX_SGE)
wr                 98 drivers/infiniband/hw/cxgb3/iwch_qp.c 	wqe->write.stag_sink = cpu_to_be32(rdma_wr(wr)->rkey);
wr                 99 drivers/infiniband/hw/cxgb3/iwch_qp.c 	wqe->write.to_sink = cpu_to_be64(rdma_wr(wr)->remote_addr);
wr                101 drivers/infiniband/hw/cxgb3/iwch_qp.c 	if (wr->opcode == IB_WR_RDMA_WRITE_WITH_IMM) {
wr                103 drivers/infiniband/hw/cxgb3/iwch_qp.c 		wqe->write.sgl[0].stag = wr->ex.imm_data;
wr                109 drivers/infiniband/hw/cxgb3/iwch_qp.c 		for (i = 0; i < wr->num_sge; i++) {
wr                110 drivers/infiniband/hw/cxgb3/iwch_qp.c 			if ((plen + wr->sg_list[i].length) < plen) {
wr                113 drivers/infiniband/hw/cxgb3/iwch_qp.c 			plen += wr->sg_list[i].length;
wr                115 drivers/infiniband/hw/cxgb3/iwch_qp.c 			    cpu_to_be32(wr->sg_list[i].lkey);
wr                117 drivers/infiniband/hw/cxgb3/iwch_qp.c 			    cpu_to_be32(wr->sg_list[i].length);
wr                119 drivers/infiniband/hw/cxgb3/iwch_qp.c 			    cpu_to_be64(wr->sg_list[i].addr);
wr                121 drivers/infiniband/hw/cxgb3/iwch_qp.c 		wqe->write.num_sgle = cpu_to_be32(wr->num_sge);
wr                122 drivers/infiniband/hw/cxgb3/iwch_qp.c 		*flit_cnt = 5 + ((wr->num_sge) << 1);
wr                128 drivers/infiniband/hw/cxgb3/iwch_qp.c static int build_rdma_read(union t3_wr *wqe, const struct ib_send_wr *wr,
wr                131 drivers/infiniband/hw/cxgb3/iwch_qp.c 	if (wr->num_sge > 1)
wr                134 drivers/infiniband/hw/cxgb3/iwch_qp.c 	if (wr->opcode == IB_WR_RDMA_READ_WITH_INV)
wr                140 drivers/infiniband/hw/cxgb3/iwch_qp.c 	wqe->read.rem_stag = cpu_to_be32(rdma_wr(wr)->rkey);
wr                141 drivers/infiniband/hw/cxgb3/iwch_qp.c 	wqe->read.rem_to = cpu_to_be64(rdma_wr(wr)->remote_addr);
wr                142 drivers/infiniband/hw/cxgb3/iwch_qp.c 	wqe->read.local_stag = cpu_to_be32(wr->sg_list[0].lkey);
wr                143 drivers/infiniband/hw/cxgb3/iwch_qp.c 	wqe->read.local_len = cpu_to_be32(wr->sg_list[0].length);
wr                144 drivers/infiniband/hw/cxgb3/iwch_qp.c 	wqe->read.local_to = cpu_to_be64(wr->sg_list[0].addr);
wr                149 drivers/infiniband/hw/cxgb3/iwch_qp.c static int build_memreg(union t3_wr *wqe, const struct ib_reg_wr *wr,
wr                152 drivers/infiniband/hw/cxgb3/iwch_qp.c 	struct iwch_mr *mhp = to_iwch_mr(wr->mr);
wr                159 drivers/infiniband/hw/cxgb3/iwch_qp.c 	wqe->fastreg.stag = cpu_to_be32(wr->key);
wr                166 drivers/infiniband/hw/cxgb3/iwch_qp.c 		V_FR_PAGE_SIZE(ilog2(wr->mr->page_size) - 12) |
wr                168 drivers/infiniband/hw/cxgb3/iwch_qp.c 		V_FR_PERMS(iwch_ib_to_tpt_access(wr->access)));
wr                192 drivers/infiniband/hw/cxgb3/iwch_qp.c static int build_inv_stag(union t3_wr *wqe, const struct ib_send_wr *wr,
wr                195 drivers/infiniband/hw/cxgb3/iwch_qp.c 	wqe->local_inv.stag = cpu_to_be32(wr->ex.invalidate_rkey);
wr                249 drivers/infiniband/hw/cxgb3/iwch_qp.c 			   const struct ib_recv_wr *wr)
wr                255 drivers/infiniband/hw/cxgb3/iwch_qp.c 	err = iwch_sgl2pbl_map(qhp->rhp, wr->sg_list, wr->num_sge, pbl_addr,
wr                263 drivers/infiniband/hw/cxgb3/iwch_qp.c 	wqe->recv.num_sgle = cpu_to_be32(wr->num_sge);
wr                264 drivers/infiniband/hw/cxgb3/iwch_qp.c 	for (i = 0; i < wr->num_sge; i++) {
wr                265 drivers/infiniband/hw/cxgb3/iwch_qp.c 		wqe->recv.sgl[i].stag = cpu_to_be32(wr->sg_list[i].lkey);
wr                266 drivers/infiniband/hw/cxgb3/iwch_qp.c 		wqe->recv.sgl[i].len = cpu_to_be32(wr->sg_list[i].length);
wr                269 drivers/infiniband/hw/cxgb3/iwch_qp.c 		wqe->recv.sgl[i].to = cpu_to_be64(((u32)wr->sg_list[i].addr) &
wr                282 drivers/infiniband/hw/cxgb3/iwch_qp.c 			     qhp->wq.rq_size_log2)].wr_id = wr->wr_id;
wr                289 drivers/infiniband/hw/cxgb3/iwch_qp.c 				const struct ib_recv_wr *wr)
wr                311 drivers/infiniband/hw/cxgb3/iwch_qp.c 	wqe->recv.num_sgle = cpu_to_be32(wr->num_sge);
wr                313 drivers/infiniband/hw/cxgb3/iwch_qp.c 	for (i = 0; i < wr->num_sge; i++) {
wr                322 drivers/infiniband/hw/cxgb3/iwch_qp.c 		if (wr->sg_list[i].length > T3_STAG0_MAX_PBE_LEN)
wr                329 drivers/infiniband/hw/cxgb3/iwch_qp.c 		if (wr->sg_list[i].lkey != 0)
wr                332 drivers/infiniband/hw/cxgb3/iwch_qp.c 		wqe->recv.sgl[i].len = cpu_to_be32(wr->sg_list[i].length);
wr                333 drivers/infiniband/hw/cxgb3/iwch_qp.c 		wqe->recv.sgl[i].to = cpu_to_be64(wr->sg_list[i].addr);
wr                345 drivers/infiniband/hw/cxgb3/iwch_qp.c 			     qhp->wq.rq_size_log2)].wr_id = wr->wr_id;
wr                351 drivers/infiniband/hw/cxgb3/iwch_qp.c int iwch_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr,
wr                380 drivers/infiniband/hw/cxgb3/iwch_qp.c 	while (wr) {
wr                388 drivers/infiniband/hw/cxgb3/iwch_qp.c 		if (wr->send_flags & IB_SEND_SOLICITED)
wr                390 drivers/infiniband/hw/cxgb3/iwch_qp.c 		if (wr->send_flags & IB_SEND_SIGNALED)
wr                394 drivers/infiniband/hw/cxgb3/iwch_qp.c 		switch (wr->opcode) {
wr                397 drivers/infiniband/hw/cxgb3/iwch_qp.c 			if (wr->send_flags & IB_SEND_FENCE)
wr                400 drivers/infiniband/hw/cxgb3/iwch_qp.c 			err = build_rdma_send(wqe, wr, &t3_wr_flit_cnt);
wr                405 drivers/infiniband/hw/cxgb3/iwch_qp.c 			err = build_rdma_write(wqe, wr, &t3_wr_flit_cnt);
wr                411 drivers/infiniband/hw/cxgb3/iwch_qp.c 			err = build_rdma_read(wqe, wr, &t3_wr_flit_cnt);
wr                420 drivers/infiniband/hw/cxgb3/iwch_qp.c 			err = build_memreg(wqe, reg_wr(wr), &t3_wr_flit_cnt,
wr                424 drivers/infiniband/hw/cxgb3/iwch_qp.c 			if (wr->send_flags & IB_SEND_FENCE)
wr                427 drivers/infiniband/hw/cxgb3/iwch_qp.c 			err = build_inv_stag(wqe, wr, &t3_wr_flit_cnt);
wr                431 drivers/infiniband/hw/cxgb3/iwch_qp.c 				 wr->opcode);
wr                437 drivers/infiniband/hw/cxgb3/iwch_qp.c 		sqp->wr_id = wr->wr_id;
wr                441 drivers/infiniband/hw/cxgb3/iwch_qp.c 		sqp->signaled = (wr->send_flags & IB_SEND_SIGNALED);
wr                448 drivers/infiniband/hw/cxgb3/iwch_qp.c 			 __func__, (unsigned long long)wr->wr_id, idx,
wr                451 drivers/infiniband/hw/cxgb3/iwch_qp.c 		wr = wr->next;
wr                462 drivers/infiniband/hw/cxgb3/iwch_qp.c 		*bad_wr = wr;
wr                466 drivers/infiniband/hw/cxgb3/iwch_qp.c int iwch_post_receive(struct ib_qp *ibqp, const struct ib_recv_wr *wr,
wr                485 drivers/infiniband/hw/cxgb3/iwch_qp.c 	if (!wr) {
wr                490 drivers/infiniband/hw/cxgb3/iwch_qp.c 	while (wr) {
wr                491 drivers/infiniband/hw/cxgb3/iwch_qp.c 		if (wr->num_sge > T3_MAX_SGE) {
wr                498 drivers/infiniband/hw/cxgb3/iwch_qp.c 			if (wr->sg_list[0].lkey)
wr                499 drivers/infiniband/hw/cxgb3/iwch_qp.c 				err = build_rdma_recv(qhp, wqe, wr);
wr                501 drivers/infiniband/hw/cxgb3/iwch_qp.c 				err = build_zero_stag_recv(qhp, wqe, wr);
wr                512 drivers/infiniband/hw/cxgb3/iwch_qp.c 			 __func__, (unsigned long long)wr->wr_id,
wr                516 drivers/infiniband/hw/cxgb3/iwch_qp.c 		wr = wr->next;
wr                525 drivers/infiniband/hw/cxgb3/iwch_qp.c 		*bad_wr = wr;
wr                971 drivers/infiniband/hw/cxgb4/iw_cxgb4.h int c4iw_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr,
wr                973 drivers/infiniband/hw/cxgb4/iw_cxgb4.h int c4iw_post_receive(struct ib_qp *ibqp, const struct ib_recv_wr *wr,
wr               1052 drivers/infiniband/hw/cxgb4/iw_cxgb4.h int c4iw_post_srq_recv(struct ib_srq *ibsrq, const struct ib_recv_wr *wr,
wr                 86 drivers/infiniband/hw/cxgb4/mem.c 	req->wr.wr_hi = cpu_to_be32(FW_WR_OP_V(FW_ULPTX_WR) |
wr                 88 drivers/infiniband/hw/cxgb4/mem.c 	req->wr.wr_lo = wr_waitp ? (__force __be64)(unsigned long)wr_waitp : 0L;
wr                 89 drivers/infiniband/hw/cxgb4/mem.c 	req->wr.wr_mid = cpu_to_be32(FW_WR_LEN16_V(DIV_ROUND_UP(wr_len, 16)));
wr                 94 drivers/infiniband/hw/cxgb4/mem.c 	req->len16 = cpu_to_be32(DIV_ROUND_UP(wr_len-sizeof(req->wr), 16));
wr                148 drivers/infiniband/hw/cxgb4/mem.c 			req->wr.wr_hi = cpu_to_be32(FW_WR_OP_V(FW_ULPTX_WR) |
wr                150 drivers/infiniband/hw/cxgb4/mem.c 			req->wr.wr_lo = (__force __be64)(unsigned long)wr_waitp;
wr                152 drivers/infiniband/hw/cxgb4/mem.c 			req->wr.wr_hi = cpu_to_be32(FW_WR_OP_V(FW_ULPTX_WR));
wr                153 drivers/infiniband/hw/cxgb4/mem.c 		req->wr.wr_mid = cpu_to_be32(
wr                159 drivers/infiniband/hw/cxgb4/mem.c 		req->len16 = cpu_to_be32(DIV_ROUND_UP(wr_len-sizeof(req->wr),
wr                414 drivers/infiniband/hw/cxgb4/qp.c 		      const struct ib_send_wr *wr, int max, u32 *plenp)
wr                422 drivers/infiniband/hw/cxgb4/qp.c 	for (i = 0; i < wr->num_sge; i++) {
wr                423 drivers/infiniband/hw/cxgb4/qp.c 		if ((plen + wr->sg_list[i].length) > max)
wr                425 drivers/infiniband/hw/cxgb4/qp.c 		srcp = (u8 *)(unsigned long)wr->sg_list[i].addr;
wr                426 drivers/infiniband/hw/cxgb4/qp.c 		plen += wr->sg_list[i].length;
wr                427 drivers/infiniband/hw/cxgb4/qp.c 		rem = wr->sg_list[i].length;
wr                489 drivers/infiniband/hw/cxgb4/qp.c 			   const struct ib_send_wr *wr, u8 *len16)
wr                495 drivers/infiniband/hw/cxgb4/qp.c 	if (wr->num_sge > T4_MAX_SEND_SGE)
wr                497 drivers/infiniband/hw/cxgb4/qp.c 	switch (wr->opcode) {
wr                499 drivers/infiniband/hw/cxgb4/qp.c 		if (wr->send_flags & IB_SEND_SOLICITED)
wr                508 drivers/infiniband/hw/cxgb4/qp.c 		if (wr->send_flags & IB_SEND_SOLICITED)
wr                514 drivers/infiniband/hw/cxgb4/qp.c 		wqe->send.stag_inv = cpu_to_be32(wr->ex.invalidate_rkey);
wr                524 drivers/infiniband/hw/cxgb4/qp.c 	if (wr->num_sge) {
wr                525 drivers/infiniband/hw/cxgb4/qp.c 		if (wr->send_flags & IB_SEND_INLINE) {
wr                526 drivers/infiniband/hw/cxgb4/qp.c 			ret = build_immd(sq, wqe->send.u.immd_src, wr,
wr                536 drivers/infiniband/hw/cxgb4/qp.c 					 wr->sg_list, wr->num_sge, &plen);
wr                540 drivers/infiniband/hw/cxgb4/qp.c 			       wr->num_sge * sizeof(struct fw_ri_sge);
wr                556 drivers/infiniband/hw/cxgb4/qp.c 			    const struct ib_send_wr *wr, u8 *len16)
wr                562 drivers/infiniband/hw/cxgb4/qp.c 	if (wr->num_sge > T4_MAX_SEND_SGE)
wr                569 drivers/infiniband/hw/cxgb4/qp.c 	if (wr->opcode == IB_WR_RDMA_WRITE_WITH_IMM)
wr                570 drivers/infiniband/hw/cxgb4/qp.c 		wqe->write.iw_imm_data.ib_imm_data.imm_data32 = wr->ex.imm_data;
wr                573 drivers/infiniband/hw/cxgb4/qp.c 	wqe->write.stag_sink = cpu_to_be32(rdma_wr(wr)->rkey);
wr                574 drivers/infiniband/hw/cxgb4/qp.c 	wqe->write.to_sink = cpu_to_be64(rdma_wr(wr)->remote_addr);
wr                575 drivers/infiniband/hw/cxgb4/qp.c 	if (wr->num_sge) {
wr                576 drivers/infiniband/hw/cxgb4/qp.c 		if (wr->send_flags & IB_SEND_INLINE) {
wr                577 drivers/infiniband/hw/cxgb4/qp.c 			ret = build_immd(sq, wqe->write.u.immd_src, wr,
wr                587 drivers/infiniband/hw/cxgb4/qp.c 					 wr->sg_list, wr->num_sge, &plen);
wr                591 drivers/infiniband/hw/cxgb4/qp.c 			       wr->num_sge * sizeof(struct fw_ri_sge);
wr                607 drivers/infiniband/hw/cxgb4/qp.c 			    struct ib_send_wr *wr)
wr                609 drivers/infiniband/hw/cxgb4/qp.c 	memcpy((u8 *)immdp->data, (u8 *)(uintptr_t)wr->sg_list->addr, 16);
wr                617 drivers/infiniband/hw/cxgb4/qp.c 				  const struct ib_send_wr *wr, u8 *len16)
wr                633 drivers/infiniband/hw/cxgb4/qp.c 	wcwr->stag_sink = cpu_to_be32(rdma_wr(wr)->rkey);
wr                634 drivers/infiniband/hw/cxgb4/qp.c 	wcwr->to_sink = cpu_to_be64(rdma_wr(wr)->remote_addr);
wr                635 drivers/infiniband/hw/cxgb4/qp.c 	if (wr->next->opcode == IB_WR_SEND)
wr                638 drivers/infiniband/hw/cxgb4/qp.c 		wcwr->stag_inv = cpu_to_be32(wr->next->ex.invalidate_rkey);
wr                643 drivers/infiniband/hw/cxgb4/qp.c 	if (wr->next->send_flags & IB_SEND_INLINE)
wr                644 drivers/infiniband/hw/cxgb4/qp.c 		build_immd_cmpl(sq, &wcwr->u_cmpl.immd_src, wr->next);
wr                647 drivers/infiniband/hw/cxgb4/qp.c 			   &wcwr->u_cmpl.isgl_src, wr->next->sg_list, 1, NULL);
wr                651 drivers/infiniband/hw/cxgb4/qp.c 		   wcwr->u.isgl_src, wr->sg_list, wr->num_sge, &plen);
wr                654 drivers/infiniband/hw/cxgb4/qp.c 		wr->num_sge * sizeof(struct fw_ri_sge);
wr                659 drivers/infiniband/hw/cxgb4/qp.c static int build_rdma_read(union t4_wr *wqe, const struct ib_send_wr *wr,
wr                662 drivers/infiniband/hw/cxgb4/qp.c 	if (wr->num_sge > 1)
wr                664 drivers/infiniband/hw/cxgb4/qp.c 	if (wr->num_sge && wr->sg_list[0].length) {
wr                665 drivers/infiniband/hw/cxgb4/qp.c 		wqe->read.stag_src = cpu_to_be32(rdma_wr(wr)->rkey);
wr                666 drivers/infiniband/hw/cxgb4/qp.c 		wqe->read.to_src_hi = cpu_to_be32((u32)(rdma_wr(wr)->remote_addr
wr                668 drivers/infiniband/hw/cxgb4/qp.c 		wqe->read.to_src_lo = cpu_to_be32((u32)rdma_wr(wr)->remote_addr);
wr                669 drivers/infiniband/hw/cxgb4/qp.c 		wqe->read.stag_sink = cpu_to_be32(wr->sg_list[0].lkey);
wr                670 drivers/infiniband/hw/cxgb4/qp.c 		wqe->read.plen = cpu_to_be32(wr->sg_list[0].length);
wr                671 drivers/infiniband/hw/cxgb4/qp.c 		wqe->read.to_sink_hi = cpu_to_be32((u32)(wr->sg_list[0].addr
wr                673 drivers/infiniband/hw/cxgb4/qp.c 		wqe->read.to_sink_lo = cpu_to_be32((u32)(wr->sg_list[0].addr));
wr                689 drivers/infiniband/hw/cxgb4/qp.c static void post_write_cmpl(struct c4iw_qp *qhp, const struct ib_send_wr *wr)
wr                691 drivers/infiniband/hw/cxgb4/qp.c 	bool send_signaled = (wr->next->send_flags & IB_SEND_SIGNALED) ||
wr                693 drivers/infiniband/hw/cxgb4/qp.c 	bool write_signaled = (wr->send_flags & IB_SEND_SIGNALED) ||
wr                707 drivers/infiniband/hw/cxgb4/qp.c 	build_rdma_write_cmpl(&qhp->wq.sq, &wqe->write_cmpl, wr, &len16);
wr                716 drivers/infiniband/hw/cxgb4/qp.c 	swsqe->wr_id = wr->wr_id;
wr                732 drivers/infiniband/hw/cxgb4/qp.c 	if (wr->next->opcode == IB_WR_SEND)
wr                740 drivers/infiniband/hw/cxgb4/qp.c 	swsqe->wr_id = wr->next->wr_id;
wr                759 drivers/infiniband/hw/cxgb4/qp.c 			   const struct ib_recv_wr *wr, u8 *len16)
wr                765 drivers/infiniband/hw/cxgb4/qp.c 			 &wqe->recv.isgl, wr->sg_list, wr->num_sge, NULL);
wr                769 drivers/infiniband/hw/cxgb4/qp.c 		sizeof(wqe->recv) + wr->num_sge * sizeof(struct fw_ri_sge), 16);
wr                773 drivers/infiniband/hw/cxgb4/qp.c static int build_srq_recv(union t4_recv_wr *wqe, const struct ib_recv_wr *wr,
wr                779 drivers/infiniband/hw/cxgb4/qp.c 			 &wqe->recv.isgl, wr->sg_list, wr->num_sge, NULL);
wr                783 drivers/infiniband/hw/cxgb4/qp.c 			      wr->num_sge * sizeof(struct fw_ri_sge), 16);
wr                788 drivers/infiniband/hw/cxgb4/qp.c 			      const struct ib_reg_wr *wr, struct c4iw_mr *mhp,
wr                802 drivers/infiniband/hw/cxgb4/qp.c 		FW_RI_TPTE_PERM_V(c4iw_ib_to_tpt_access(wr->access)) |
wr                804 drivers/infiniband/hw/cxgb4/qp.c 		FW_RI_TPTE_PS_V(ilog2(wr->mr->page_size) - 12));
wr                820 drivers/infiniband/hw/cxgb4/qp.c 			const struct ib_reg_wr *wr, struct c4iw_mr *mhp,
wr                833 drivers/infiniband/hw/cxgb4/qp.c 	wqe->fr.pgsz_shift = ilog2(wr->mr->page_size) - 12;
wr                835 drivers/infiniband/hw/cxgb4/qp.c 	wqe->fr.mem_perms = c4iw_ib_to_tpt_access(wr->access);
wr                838 drivers/infiniband/hw/cxgb4/qp.c 	wqe->fr.stag = cpu_to_be32(wr->key);
wr                883 drivers/infiniband/hw/cxgb4/qp.c static int build_inv_stag(union t4_wr *wqe, const struct ib_send_wr *wr,
wr                886 drivers/infiniband/hw/cxgb4/qp.c 	wqe->inv.stag_inv = cpu_to_be32(wr->ex.invalidate_rkey);
wr                979 drivers/infiniband/hw/cxgb4/qp.c 				const struct ib_send_wr *wr)
wr                990 drivers/infiniband/hw/cxgb4/qp.c 	opcode = ib_to_fw_opcode(wr->opcode);
wr                994 drivers/infiniband/hw/cxgb4/qp.c 	cqe.u.drain_cookie = wr->wr_id;
wr               1018 drivers/infiniband/hw/cxgb4/qp.c 				 const struct ib_send_wr *wr,
wr               1023 drivers/infiniband/hw/cxgb4/qp.c 	while (wr) {
wr               1024 drivers/infiniband/hw/cxgb4/qp.c 		ret = complete_sq_drain_wr(qhp, wr);
wr               1026 drivers/infiniband/hw/cxgb4/qp.c 			*bad_wr = wr;
wr               1029 drivers/infiniband/hw/cxgb4/qp.c 		wr = wr->next;
wr               1035 drivers/infiniband/hw/cxgb4/qp.c 				 const struct ib_recv_wr *wr)
wr               1045 drivers/infiniband/hw/cxgb4/qp.c 	cqe.u.drain_cookie = wr->wr_id;
wr               1068 drivers/infiniband/hw/cxgb4/qp.c 				  const struct ib_recv_wr *wr)
wr               1070 drivers/infiniband/hw/cxgb4/qp.c 	while (wr) {
wr               1071 drivers/infiniband/hw/cxgb4/qp.c 		complete_rq_drain_wr(qhp, wr);
wr               1072 drivers/infiniband/hw/cxgb4/qp.c 		wr = wr->next;
wr               1076 drivers/infiniband/hw/cxgb4/qp.c int c4iw_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr,
wr               1101 drivers/infiniband/hw/cxgb4/qp.c 		err = complete_sq_drain_wrs(qhp, wr, bad_wr);
wr               1107 drivers/infiniband/hw/cxgb4/qp.c 		*bad_wr = wr;
wr               1123 drivers/infiniband/hw/cxgb4/qp.c 	    wr && wr->next && !wr->next->next &&
wr               1124 drivers/infiniband/hw/cxgb4/qp.c 	    wr->opcode == IB_WR_RDMA_WRITE &&
wr               1125 drivers/infiniband/hw/cxgb4/qp.c 	    wr->sg_list[0].length && wr->num_sge <= T4_WRITE_CMPL_MAX_SGL &&
wr               1126 drivers/infiniband/hw/cxgb4/qp.c 	    (wr->next->opcode == IB_WR_SEND ||
wr               1127 drivers/infiniband/hw/cxgb4/qp.c 	    wr->next->opcode == IB_WR_SEND_WITH_INV) &&
wr               1128 drivers/infiniband/hw/cxgb4/qp.c 	    wr->next->sg_list[0].length == T4_WRITE_CMPL_MAX_CQE &&
wr               1129 drivers/infiniband/hw/cxgb4/qp.c 	    wr->next->num_sge == 1 && num_wrs >= 2) {
wr               1130 drivers/infiniband/hw/cxgb4/qp.c 		post_write_cmpl(qhp, wr);
wr               1135 drivers/infiniband/hw/cxgb4/qp.c 	while (wr) {
wr               1138 drivers/infiniband/hw/cxgb4/qp.c 			*bad_wr = wr;
wr               1145 drivers/infiniband/hw/cxgb4/qp.c 		if (wr->send_flags & IB_SEND_SOLICITED)
wr               1147 drivers/infiniband/hw/cxgb4/qp.c 		if (wr->send_flags & IB_SEND_SIGNALED || qhp->sq_sig_all)
wr               1150 drivers/infiniband/hw/cxgb4/qp.c 		switch (wr->opcode) {
wr               1153 drivers/infiniband/hw/cxgb4/qp.c 			if (wr->send_flags & IB_SEND_FENCE)
wr               1156 drivers/infiniband/hw/cxgb4/qp.c 			if (wr->opcode == IB_WR_SEND)
wr               1160 drivers/infiniband/hw/cxgb4/qp.c 			err = build_rdma_send(&qhp->wq.sq, wqe, wr, &len16);
wr               1172 drivers/infiniband/hw/cxgb4/qp.c 			err = build_rdma_write(&qhp->wq.sq, wqe, wr, &len16);
wr               1178 drivers/infiniband/hw/cxgb4/qp.c 			if (wr->opcode == IB_WR_RDMA_READ_WITH_INV) {
wr               1179 drivers/infiniband/hw/cxgb4/qp.c 				c4iw_invalidate_mr(rhp, wr->sg_list[0].lkey);
wr               1184 drivers/infiniband/hw/cxgb4/qp.c 			err = build_rdma_read(wqe, wr, &len16);
wr               1187 drivers/infiniband/hw/cxgb4/qp.c 			swsqe->read_len = wr->sg_list[0].length;
wr               1192 drivers/infiniband/hw/cxgb4/qp.c 			struct c4iw_mr *mhp = to_c4iw_mr(reg_wr(wr)->mr);
wr               1198 drivers/infiniband/hw/cxgb4/qp.c 				build_tpte_memreg(&wqe->fr_tpte, reg_wr(wr),
wr               1202 drivers/infiniband/hw/cxgb4/qp.c 				err = build_memreg(&qhp->wq.sq, wqe, reg_wr(wr),
wr               1212 drivers/infiniband/hw/cxgb4/qp.c 			if (wr->send_flags & IB_SEND_FENCE)
wr               1216 drivers/infiniband/hw/cxgb4/qp.c 			err = build_inv_stag(wqe, wr, &len16);
wr               1217 drivers/infiniband/hw/cxgb4/qp.c 			c4iw_invalidate_mr(rhp, wr->ex.invalidate_rkey);
wr               1221 drivers/infiniband/hw/cxgb4/qp.c 				wr->opcode);
wr               1225 drivers/infiniband/hw/cxgb4/qp.c 			*bad_wr = wr;
wr               1230 drivers/infiniband/hw/cxgb4/qp.c 		swsqe->signaled = (wr->send_flags & IB_SEND_SIGNALED) ||
wr               1233 drivers/infiniband/hw/cxgb4/qp.c 		swsqe->wr_id = wr->wr_id;
wr               1243 drivers/infiniband/hw/cxgb4/qp.c 			 (unsigned long long)wr->wr_id, qhp->wq.sq.pidx,
wr               1245 drivers/infiniband/hw/cxgb4/qp.c 		wr = wr->next;
wr               1260 drivers/infiniband/hw/cxgb4/qp.c int c4iw_post_receive(struct ib_qp *ibqp, const struct ib_recv_wr *wr,
wr               1280 drivers/infiniband/hw/cxgb4/qp.c 		complete_rq_drain_wrs(qhp, wr);
wr               1286 drivers/infiniband/hw/cxgb4/qp.c 		*bad_wr = wr;
wr               1289 drivers/infiniband/hw/cxgb4/qp.c 	while (wr) {
wr               1290 drivers/infiniband/hw/cxgb4/qp.c 		if (wr->num_sge > T4_MAX_RECV_SGE) {
wr               1292 drivers/infiniband/hw/cxgb4/qp.c 			*bad_wr = wr;
wr               1299 drivers/infiniband/hw/cxgb4/qp.c 			err = build_rdma_recv(qhp, wqe, wr, &len16);
wr               1303 drivers/infiniband/hw/cxgb4/qp.c 			*bad_wr = wr;
wr               1307 drivers/infiniband/hw/cxgb4/qp.c 		qhp->wq.rq.sw_rq[qhp->wq.rq.pidx].wr_id = wr->wr_id;
wr               1324 drivers/infiniband/hw/cxgb4/qp.c 			 (unsigned long long)wr->wr_id, qhp->wq.rq.pidx);
wr               1327 drivers/infiniband/hw/cxgb4/qp.c 		wr = wr->next;
wr               1356 drivers/infiniband/hw/cxgb4/qp.c int c4iw_post_srq_recv(struct ib_srq *ibsrq, const struct ib_recv_wr *wr,
wr               1374 drivers/infiniband/hw/cxgb4/qp.c 	while (wr) {
wr               1375 drivers/infiniband/hw/cxgb4/qp.c 		if (wr->num_sge > T4_MAX_RECV_SGE) {
wr               1377 drivers/infiniband/hw/cxgb4/qp.c 			*bad_wr = wr;
wr               1382 drivers/infiniband/hw/cxgb4/qp.c 			err = build_srq_recv(wqe, wr, &len16);
wr               1386 drivers/infiniband/hw/cxgb4/qp.c 			*bad_wr = wr;
wr               1401 drivers/infiniband/hw/cxgb4/qp.c 			defer_srq_wr(&srq->wq, wqe, wr->wr_id, len16);
wr               1403 drivers/infiniband/hw/cxgb4/qp.c 			srq->wq.sw_rq[srq->wq.pidx].wr_id = wr->wr_id;
wr               1410 drivers/infiniband/hw/cxgb4/qp.c 				 (unsigned long long)wr->wr_id);
wr               1414 drivers/infiniband/hw/cxgb4/qp.c 		wr = wr->next;
wr                 44 drivers/infiniband/hw/hfi1/opfn.c 	struct ib_atomic_wr wr;
wr                 83 drivers/infiniband/hw/hfi1/opfn.c 	memset(&wr, 0, sizeof(wr));
wr                 84 drivers/infiniband/hw/hfi1/opfn.c 	wr.wr.opcode = IB_WR_OPFN;
wr                 85 drivers/infiniband/hw/hfi1/opfn.c 	wr.remote_addr = HFI1_VERBS_E_ATOMIC_VADDR;
wr                 86 drivers/infiniband/hw/hfi1/opfn.c 	wr.compare_add = data;
wr                 92 drivers/infiniband/hw/hfi1/opfn.c 	ret = ib_post_send(&qp->ibqp, &wr.wr, NULL);
wr                674 drivers/infiniband/hw/hfi1/qp.c 		   wqe ? wqe->wr.opcode : 0,
wr                536 drivers/infiniband/hw/hfi1/rc.c 			if ((wqe->wr.send_flags & IB_SEND_FENCE) &&
wr                538 drivers/infiniband/hw/hfi1/rc.c 			    (wqe->wr.opcode != IB_WR_TID_RDMA_READ ||
wr                547 drivers/infiniband/hw/hfi1/rc.c 			if (wqe->wr.opcode == IB_WR_REG_MR ||
wr                548 drivers/infiniband/hw/hfi1/rc.c 			    wqe->wr.opcode == IB_WR_LOCAL_INV) {
wr                558 drivers/infiniband/hw/hfi1/rc.c 				if (!(wqe->wr.send_flags &
wr                562 drivers/infiniband/hw/hfi1/rc.c 						wqe->wr.ex.invalidate_rkey);
wr                593 drivers/infiniband/hw/hfi1/rc.c 		switch (wqe->wr.opcode) {
wr                605 drivers/infiniband/hw/hfi1/rc.c 			if (wqe->wr.opcode == IB_WR_SEND) {
wr                607 drivers/infiniband/hw/hfi1/rc.c 			} else if (wqe->wr.opcode == IB_WR_SEND_WITH_IMM) {
wr                610 drivers/infiniband/hw/hfi1/rc.c 				ohdr->u.imm_data = wqe->wr.ex.imm_data;
wr                616 drivers/infiniband/hw/hfi1/rc.c 						wqe->wr.ex.invalidate_rkey);
wr                619 drivers/infiniband/hw/hfi1/rc.c 			if (wqe->wr.send_flags & IB_SEND_SOLICITED)
wr                647 drivers/infiniband/hw/hfi1/rc.c 			if (wqe->wr.opcode == IB_WR_RDMA_WRITE) {
wr                653 drivers/infiniband/hw/hfi1/rc.c 				ohdr->u.rc.imm_data = wqe->wr.ex.imm_data;
wr                655 drivers/infiniband/hw/hfi1/rc.c 				if (wqe->wr.send_flags & IB_SEND_SOLICITED)
wr                713 drivers/infiniband/hw/hfi1/rc.c 				if (__w->wr.opcode != IB_WR_TID_RDMA_WRITE ||
wr                763 drivers/infiniband/hw/hfi1/rc.c 							  wqe->wr.opcode,
wr                803 drivers/infiniband/hw/hfi1/rc.c 							 wqe->wr.opcode,
wr                835 drivers/infiniband/hw/hfi1/rc.c 					qp->s_sge.num_sge = wqe->wr.num_sge;
wr                889 drivers/infiniband/hw/hfi1/rc.c 			if (wqe->wr.opcode == IB_WR_ATOMIC_CMP_AND_SWP ||
wr                890 drivers/infiniband/hw/hfi1/rc.c 			    wqe->wr.opcode == IB_WR_OPFN) {
wr                917 drivers/infiniband/hw/hfi1/rc.c 		if (wqe->wr.opcode != IB_WR_TID_RDMA_READ) {
wr                920 drivers/infiniband/hw/hfi1/rc.c 			qp->s_sge.num_sge = wqe->wr.num_sge;
wr                929 drivers/infiniband/hw/hfi1/rc.c 		if (wqe->wr.opcode == IB_WR_RDMA_READ ||
wr                930 drivers/infiniband/hw/hfi1/rc.c 		    wqe->wr.opcode == IB_WR_TID_RDMA_WRITE)
wr                932 drivers/infiniband/hw/hfi1/rc.c 		else if (wqe->wr.opcode == IB_WR_TID_RDMA_READ)
wr                962 drivers/infiniband/hw/hfi1/rc.c 		if (wqe->wr.opcode == IB_WR_SEND) {
wr                964 drivers/infiniband/hw/hfi1/rc.c 		} else if (wqe->wr.opcode == IB_WR_SEND_WITH_IMM) {
wr                967 drivers/infiniband/hw/hfi1/rc.c 			ohdr->u.imm_data = wqe->wr.ex.imm_data;
wr                972 drivers/infiniband/hw/hfi1/rc.c 			ohdr->u.ieth = cpu_to_be32(wqe->wr.ex.invalidate_rkey);
wr                975 drivers/infiniband/hw/hfi1/rc.c 		if (wqe->wr.send_flags & IB_SEND_SOLICITED)
wr               1007 drivers/infiniband/hw/hfi1/rc.c 		if (wqe->wr.opcode == IB_WR_RDMA_WRITE) {
wr               1012 drivers/infiniband/hw/hfi1/rc.c 			ohdr->u.imm_data = wqe->wr.ex.imm_data;
wr               1014 drivers/infiniband/hw/hfi1/rc.c 			if (wqe->wr.send_flags & IB_SEND_SOLICITED)
wr               1075 drivers/infiniband/hw/hfi1/rc.c 		trace_hfi1_tid_req_make_req_write(qp, 0, wqe->wr.opcode,
wr               1080 drivers/infiniband/hw/hfi1/rc.c 		if (wqe->wr.opcode != IB_WR_TID_RDMA_READ)
wr               1130 drivers/infiniband/hw/hfi1/rc.c 		trace_hfi1_tid_req_make_req_read(qp, 0, wqe->wr.opcode,
wr               1141 drivers/infiniband/hw/hfi1/rc.c 		if (wqe->wr.opcode != IB_WR_TID_RDMA_READ || delta == 0 ||
wr               1173 drivers/infiniband/hw/hfi1/rc.c 		trace_hfi1_tid_req_make_req_read(qp, 0, wqe->wr.opcode,
wr               1180 drivers/infiniband/hw/hfi1/rc.c 	    wqe->wr.opcode != IB_WR_TID_RDMA_WRITE)
wr               1463 drivers/infiniband/hw/hfi1/rc.c 	u32 opcode = wqe->wr.opcode;
wr               1481 drivers/infiniband/hw/hfi1/rc.c 								wqe->wr.opcode,
wr               1551 drivers/infiniband/hw/hfi1/rc.c 	opcode = wqe->wr.opcode;
wr               1625 drivers/infiniband/hw/hfi1/rc.c 			if (wqe->wr.opcode == IB_WR_OPFN) {
wr               1638 drivers/infiniband/hw/hfi1/rc.c 				if (wqe->wr.opcode == IB_WR_TID_RDMA_READ) {
wr               1659 drivers/infiniband/hw/hfi1/rc.c 	if (wqe->wr.opcode == IB_WR_RDMA_READ ||
wr               1660 drivers/infiniband/hw/hfi1/rc.c 	    wqe->wr.opcode == IB_WR_TID_RDMA_READ)
wr               1688 drivers/infiniband/hw/hfi1/rc.c 			if (wqe->wr.opcode == IB_WR_RDMA_READ ||
wr               1689 drivers/infiniband/hw/hfi1/rc.c 			    wqe->wr.opcode == IB_WR_TID_RDMA_READ ||
wr               1690 drivers/infiniband/hw/hfi1/rc.c 			    wqe->wr.opcode == IB_WR_TID_RDMA_WRITE)
wr               1826 drivers/infiniband/hw/hfi1/rc.c 		if (wqe->wr.opcode == IB_WR_TID_RDMA_WRITE &&
wr               1840 drivers/infiniband/hw/hfi1/rc.c 				     ib_hfi1_wc_opcode[wqe->wr.opcode],
wr               1886 drivers/infiniband/hw/hfi1/rc.c 				     ib_hfi1_wc_opcode[wqe->wr.opcode],
wr               1917 drivers/infiniband/hw/hfi1/rc.c 	if (wqe->wr.opcode != IB_WR_TID_RDMA_WRITE)
wr               2043 drivers/infiniband/hw/hfi1/rc.c 		if (wqe->wr.opcode == IB_WR_RDMA_READ &&
wr               2058 drivers/infiniband/hw/hfi1/rc.c 		if ((wqe->wr.opcode == IB_WR_RDMA_READ &&
wr               2060 drivers/infiniband/hw/hfi1/rc.c 		    (wqe->wr.opcode == IB_WR_TID_RDMA_READ &&
wr               2062 drivers/infiniband/hw/hfi1/rc.c 		    ((wqe->wr.opcode == IB_WR_ATOMIC_CMP_AND_SWP ||
wr               2063 drivers/infiniband/hw/hfi1/rc.c 		      wqe->wr.opcode == IB_WR_ATOMIC_FETCH_AND_ADD) &&
wr               2065 drivers/infiniband/hw/hfi1/rc.c 		    (wqe->wr.opcode == IB_WR_TID_RDMA_WRITE &&
wr               2074 drivers/infiniband/hw/hfi1/rc.c 		if (wqe->wr.opcode == IB_WR_ATOMIC_CMP_AND_SWP ||
wr               2075 drivers/infiniband/hw/hfi1/rc.c 		    wqe->wr.opcode == IB_WR_ATOMIC_FETCH_AND_ADD) {
wr               2079 drivers/infiniband/hw/hfi1/rc.c 		if (wqe->wr.opcode == IB_WR_OPFN)
wr               2083 drivers/infiniband/hw/hfi1/rc.c 		    (wqe->wr.opcode == IB_WR_RDMA_READ ||
wr               2084 drivers/infiniband/hw/hfi1/rc.c 		     wqe->wr.opcode == IB_WR_ATOMIC_CMP_AND_SWP ||
wr               2085 drivers/infiniband/hw/hfi1/rc.c 		     wqe->wr.opcode == IB_WR_ATOMIC_FETCH_AND_ADD)) {
wr               2104 drivers/infiniband/hw/hfi1/rc.c 		if (wqe->wr.opcode == IB_WR_TID_RDMA_WRITE)
wr               2117 drivers/infiniband/hw/hfi1/rc.c 		if (wqe->wr.opcode == IB_WR_TID_RDMA_READ) {
wr               2133 drivers/infiniband/hw/hfi1/rc.c 			if (__w && __w->wr.opcode == IB_WR_TID_RDMA_WRITE &&
wr               2198 drivers/infiniband/hw/hfi1/rc.c 		if (wqe->wr.opcode == IB_WR_TID_RDMA_WRITE &&
wr               2212 drivers/infiniband/hw/hfi1/rc.c 		if (!(rdi->post_parms[wqe->wr.opcode].flags &
wr               2228 drivers/infiniband/hw/hfi1/rc.c 		if (wqe->wr.opcode == IB_WR_TID_RDMA_WRITE) {
wr               2275 drivers/infiniband/hw/hfi1/rc.c 				if (wqe->wr.opcode == IB_WR_TID_RDMA_READ)
wr               2318 drivers/infiniband/hw/hfi1/rc.c 		if (wqe->wr.opcode == IB_WR_RDMA_READ ||
wr               2319 drivers/infiniband/hw/hfi1/rc.c 		    wqe->wr.opcode == IB_WR_TID_RDMA_READ ||
wr               2320 drivers/infiniband/hw/hfi1/rc.c 		    wqe->wr.opcode == IB_WR_TID_RDMA_WRITE ||
wr               2321 drivers/infiniband/hw/hfi1/rc.c 		    wqe->wr.opcode == IB_WR_ATOMIC_CMP_AND_SWP ||
wr               2322 drivers/infiniband/hw/hfi1/rc.c 		    wqe->wr.opcode == IB_WR_ATOMIC_FETCH_AND_ADD)
wr               2413 drivers/infiniband/hw/hfi1/rc.c 		if (unlikely(wqe->wr.opcode != IB_WR_RDMA_READ))
wr               2428 drivers/infiniband/hw/hfi1/rc.c 		if (unlikely(wqe->wr.opcode != IB_WR_RDMA_READ))
wr               2484 drivers/infiniband/hw/hfi1/rc.c 		if (unlikely(wqe->wr.opcode != IB_WR_RDMA_READ))
wr               1793 drivers/infiniband/hw/hfi1/tid_rdma.c 	trace_hfi1_tid_req_build_read_req(qp, 0, wqe->wr.opcode, wqe->psn,
wr               2422 drivers/infiniband/hw/hfi1/tid_rdma.c 			if (wqe->wr.opcode == opcode)
wr               2521 drivers/infiniband/hw/hfi1/tid_rdma.c 	trace_hfi1_tid_req_rcv_read_resp(qp, 0, req->e.swqe->wr.opcode,
wr               2583 drivers/infiniband/hw/hfi1/tid_rdma.c 		if (wqe->wr.opcode == IB_WR_TID_RDMA_READ) {
wr               2692 drivers/infiniband/hw/hfi1/tid_rdma.c 		if (wqe->wr.opcode == IB_WR_RDMA_READ ||
wr               2693 drivers/infiniband/hw/hfi1/tid_rdma.c 		    wqe->wr.opcode == IB_WR_TID_RDMA_READ ||
wr               2694 drivers/infiniband/hw/hfi1/tid_rdma.c 		    wqe->wr.opcode == IB_WR_ATOMIC_CMP_AND_SWP ||
wr               2695 drivers/infiniband/hw/hfi1/tid_rdma.c 		    wqe->wr.opcode == IB_WR_ATOMIC_FETCH_AND_ADD) {
wr               2699 drivers/infiniband/hw/hfi1/tid_rdma.c 				if (wqe->wr.opcode == IB_WR_TID_RDMA_READ) {
wr               2730 drivers/infiniband/hw/hfi1/tid_rdma.c 	if (wqe->wr.opcode != IB_WR_TID_RDMA_READ)
wr               2734 drivers/infiniband/hw/hfi1/tid_rdma.c 	trace_hfi1_tid_req_read_kdeth_eflags(qp, 0, wqe->wr.opcode, wqe->psn,
wr               3047 drivers/infiniband/hw/hfi1/tid_rdma.c 	if (wqe->wr.opcode == IB_WR_TID_RDMA_READ) {
wr               3054 drivers/infiniband/hw/hfi1/tid_rdma.c 			trace_hfi1_tid_req_restart_req(qp, 0, wqe->wr.opcode,
wr               3065 drivers/infiniband/hw/hfi1/tid_rdma.c 	if (wqe->wr.opcode == IB_WR_TID_RDMA_READ)
wr               3097 drivers/infiniband/hw/hfi1/tid_rdma.c 	if (wqe->wr.opcode == IB_WR_TID_RDMA_WRITE) {
wr               3116 drivers/infiniband/hw/hfi1/tid_rdma.c 	if (wqe->wr.opcode == IB_WR_TID_RDMA_READ)
wr               3123 drivers/infiniband/hw/hfi1/tid_rdma.c 	trace_hfi1_tid_req_restart_req(qp, 0, wqe->wr.opcode, wqe->psn,
wr               3126 drivers/infiniband/hw/hfi1/tid_rdma.c 	if (wqe->wr.opcode == IB_WR_TID_RDMA_WRITE) {
wr               3144 drivers/infiniband/hw/hfi1/tid_rdma.c 			} while (wqe->wr.opcode != IB_WR_TID_RDMA_WRITE);
wr               3177 drivers/infiniband/hw/hfi1/tid_rdma.c 		if (wqe->wr.opcode != IB_WR_TID_RDMA_READ)
wr               3211 drivers/infiniband/hw/hfi1/tid_rdma.c 	switch (wqe->wr.opcode) {
wr               3218 drivers/infiniband/hw/hfi1/tid_rdma.c 		switch (prev->wr.opcode) {
wr               3228 drivers/infiniband/hw/hfi1/tid_rdma.c 		if (prev->wr.opcode != IB_WR_TID_RDMA_WRITE)
wr               3232 drivers/infiniband/hw/hfi1/tid_rdma.c 		switch (prev->wr.opcode) {
wr               3293 drivers/infiniband/hw/hfi1/tid_rdma.c 	if (wqe->wr.opcode == IB_WR_RDMA_READ) {
wr               3295 drivers/infiniband/hw/hfi1/tid_rdma.c 					 wqe->wr.num_sge)) {
wr               3299 drivers/infiniband/hw/hfi1/tid_rdma.c 	} else if (wqe->wr.opcode == IB_WR_RDMA_WRITE) {
wr               3316 drivers/infiniband/hw/hfi1/tid_rdma.c 		wqe->wr.opcode = new_opcode;
wr               3323 drivers/infiniband/hw/hfi1/tid_rdma.c 		if (wqe->wr.opcode == IB_WR_TID_RDMA_READ) {
wr               3343 drivers/infiniband/hw/hfi1/tid_rdma.c 		trace_hfi1_tid_req_setup_tid_wqe(qp, 1, wqe->wr.opcode,
wr               4069 drivers/infiniband/hw/hfi1/tid_rdma.c 	if (unlikely(wqe->wr.opcode != IB_WR_TID_RDMA_WRITE))
wr               4142 drivers/infiniband/hw/hfi1/tid_rdma.c 	trace_hfi1_tid_req_rcv_write_resp(qp, 0, wqe->wr.opcode, wqe->psn,
wr               4173 drivers/infiniband/hw/hfi1/tid_rdma.c 			if (wqe->wr.opcode == IB_WR_TID_RDMA_WRITE)
wr               4526 drivers/infiniband/hw/hfi1/tid_rdma.c 	if (wqe->wr.opcode != IB_WR_TID_RDMA_WRITE)
wr               4530 drivers/infiniband/hw/hfi1/tid_rdma.c 	trace_hfi1_tid_req_rcv_tid_ack(qp, 0, wqe->wr.opcode, wqe->psn,
wr               4547 drivers/infiniband/hw/hfi1/tid_rdma.c 		trace_hfi1_tid_req_rcv_tid_ack(qp, 0, wqe->wr.opcode, wqe->psn,
wr               4558 drivers/infiniband/hw/hfi1/tid_rdma.c 			if (wqe->wr.opcode != IB_WR_TID_RDMA_WRITE)
wr               4566 drivers/infiniband/hw/hfi1/tid_rdma.c 	trace_hfi1_tid_req_rcv_tid_ack(qp, 0, wqe->wr.opcode, wqe->psn,
wr               4574 drivers/infiniband/hw/hfi1/tid_rdma.c 			if (wqe->wr.opcode == IB_WR_TID_RDMA_WRITE &&
wr               4811 drivers/infiniband/hw/hfi1/tid_rdma.c 			   qp, 0, wqe->wr.opcode, wqe->psn, wqe->lpsn, req);
wr               4988 drivers/infiniband/hw/hfi1/tid_rdma.c 		if (wqe->wr.opcode == IB_WR_TID_RDMA_WRITE)
wr               5057 drivers/infiniband/hw/hfi1/tid_rdma.c 	trace_hfi1_tid_req_make_tid_pkt(qp, 0, wqe->wr.opcode, wqe->psn,
wr               5064 drivers/infiniband/hw/hfi1/tid_rdma.c 		priv->tid_ss.num_sge = wqe->wr.num_sge;
wr               5095 drivers/infiniband/hw/hfi1/tid_rdma.c 		trace_hfi1_tid_req_make_tid_pkt(qp, 0, wqe->wr.opcode,
wr                268 drivers/infiniband/hw/hfi1/tid_rdma.h 	    (wqe->wr.opcode == IB_WR_RDMA_READ ||
wr                269 drivers/infiniband/hw/hfi1/tid_rdma.h 	     wqe->wr.opcode == IB_WR_RDMA_WRITE) &&
wr                137 drivers/infiniband/hw/hfi1/trace_rc.h 		__entry->opcode = wqe->wr.opcode;
wr                797 drivers/infiniband/hw/hfi1/trace_tx.h 		__entry->wr_id = wqe->wr.wr_id;
wr                803 drivers/infiniband/hw/hfi1/trace_tx.h 		__entry->opcode = wqe->wr.opcode;
wr                804 drivers/infiniband/hw/hfi1/trace_tx.h 		__entry->send_flags = wqe->wr.send_flags;
wr                129 drivers/infiniband/hw/hfi1/uc.c 		if (wqe->wr.opcode == IB_WR_REG_MR ||
wr                130 drivers/infiniband/hw/hfi1/uc.c 		    wqe->wr.opcode == IB_WR_LOCAL_INV) {
wr                138 drivers/infiniband/hw/hfi1/uc.c 			if (!(wqe->wr.send_flags & RVT_SEND_COMPLETION_ONLY)) {
wr                140 drivers/infiniband/hw/hfi1/uc.c 					qp, wqe->wr.ex.invalidate_rkey);
wr                155 drivers/infiniband/hw/hfi1/uc.c 		qp->s_sge.num_sge = wqe->wr.num_sge;
wr                159 drivers/infiniband/hw/hfi1/uc.c 		switch (wqe->wr.opcode) {
wr                167 drivers/infiniband/hw/hfi1/uc.c 			if (wqe->wr.opcode == IB_WR_SEND) {
wr                173 drivers/infiniband/hw/hfi1/uc.c 				ohdr->u.imm_data = wqe->wr.ex.imm_data;
wr                176 drivers/infiniband/hw/hfi1/uc.c 			if (wqe->wr.send_flags & IB_SEND_SOLICITED)
wr                196 drivers/infiniband/hw/hfi1/uc.c 			if (wqe->wr.opcode == IB_WR_RDMA_WRITE) {
wr                202 drivers/infiniband/hw/hfi1/uc.c 				ohdr->u.rc.imm_data = wqe->wr.ex.imm_data;
wr                204 drivers/infiniband/hw/hfi1/uc.c 				if (wqe->wr.send_flags & IB_SEND_SOLICITED)
wr                227 drivers/infiniband/hw/hfi1/uc.c 		if (wqe->wr.opcode == IB_WR_SEND) {
wr                232 drivers/infiniband/hw/hfi1/uc.c 			ohdr->u.imm_data = wqe->wr.ex.imm_data;
wr                235 drivers/infiniband/hw/hfi1/uc.c 		if (wqe->wr.send_flags & IB_SEND_SOLICITED)
wr                252 drivers/infiniband/hw/hfi1/uc.c 		if (wqe->wr.opcode == IB_WR_RDMA_WRITE) {
wr                258 drivers/infiniband/hw/hfi1/uc.c 			ohdr->u.imm_data = wqe->wr.ex.imm_data;
wr                260 drivers/infiniband/hw/hfi1/uc.c 			if (wqe->wr.send_flags & IB_SEND_SOLICITED)
wr                152 drivers/infiniband/hw/hfi1/ud.c 	if (swqe->wr.opcode == IB_WR_SEND_WITH_IMM) {
wr                154 drivers/infiniband/hw/hfi1/ud.c 		wc.ex.imm_data = swqe->wr.ex.imm_data;
wr                222 drivers/infiniband/hw/hfi1/ud.c 	ssge.num_sge = swqe->wr.num_sge;
wr                258 drivers/infiniband/hw/hfi1/ud.c 	rvt_recv_cq(qp, &wc, swqe->wr.send_flags & IB_SEND_SOLICITED);
wr                274 drivers/infiniband/hw/hfi1/ud.c 	if (wqe->wr.opcode == IB_WR_SEND_WITH_IMM) {
wr                275 drivers/infiniband/hw/hfi1/ud.c 		ohdr->u.ud.imm_data = wqe->wr.ex.imm_data;
wr                281 drivers/infiniband/hw/hfi1/ud.c 	if (wqe->wr.send_flags & IB_SEND_SOLICITED)
wr                325 drivers/infiniband/hw/hfi1/ud.c 	if (wqe->wr.opcode == IB_WR_SEND_WITH_IMM)
wr                404 drivers/infiniband/hw/hfi1/ud.c 		if (wqe->wr.opcode == IB_WR_SEND_WITH_IMM)
wr                557 drivers/infiniband/hw/hfi1/ud.c 	qp->s_sge.num_sge = wqe->wr.num_sge;
wr                969 drivers/infiniband/hw/hns/hns_roce_device.h 	int (*post_send)(struct ib_qp *ibqp, const struct ib_send_wr *wr,
wr                990 drivers/infiniband/hw/hns/hns_roce_device.h 	int (*post_srq_recv)(struct ib_srq *ibsrq, const struct ib_recv_wr *wr,
wr               1256 drivers/infiniband/hw/hns/hns_roce_device.h __be32 send_ieth(const struct ib_send_wr *wr);
wr                 62 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 				 const struct ib_send_wr *wr,
wr                 66 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 	struct hns_roce_ah *ah = to_hr_ah(ud_wr(wr)->ah);
wr                 92 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 	for (nreq = 0; wr; ++nreq, wr = wr->next) {
wr                 95 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 			*bad_wr = wr;
wr                101 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 		if (unlikely(wr->num_sge > qp->sq.max_gs)) {
wr                103 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 				wr->num_sge, qp->sq.max_gs);
wr                105 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 			*bad_wr = wr;
wr                110 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 		qp->sq.wrid[wqe_idx] = wr->wr_id;
wr                160 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 			ud_sq_wqe->u32_8 |= (wr->send_flags & IB_SEND_SIGNALED ?
wr                162 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 				(wr->send_flags & IB_SEND_SOLICITED ?
wr                164 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 				((wr->opcode == IB_WR_SEND_WITH_IMM) ?
wr                170 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 				       ud_wr(wr)->remote_qpn);
wr                202 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 				       cpu_to_le32((u32)wr->sg_list[0].addr);
wr                204 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 				       cpu_to_le32((wr->sg_list[0].addr) >> 32);
wr                206 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 				       cpu_to_le32(wr->sg_list[0].lkey);
wr                209 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 				       cpu_to_le32((u32)wr->sg_list[1].addr);
wr                211 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 				       cpu_to_le32((wr->sg_list[1].addr) >> 32);
wr                213 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 				       cpu_to_le32(wr->sg_list[1].lkey);
wr                219 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 			for (i = 0; i < wr->num_sge; i++)
wr                220 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 				tmp_len += wr->sg_list[i].length;
wr                228 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 			switch (wr->opcode) {
wr                231 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 				ctrl->imm_data = wr->ex.imm_data;
wr                235 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 					cpu_to_le32(wr->ex.invalidate_rkey);
wr                244 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 			ctrl->flag |= (wr->send_flags & IB_SEND_SIGNALED ?
wr                246 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 				      (wr->send_flags & IB_SEND_SOLICITED ?
wr                248 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 				      ((wr->opcode == IB_WR_SEND_WITH_IMM ||
wr                249 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 				      wr->opcode == IB_WR_RDMA_WRITE_WITH_IMM) ?
wr                251 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 				      (wr->send_flags & IB_SEND_FENCE ?
wr                256 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 			switch (wr->opcode) {
wr                259 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 				set_raddr_seg(wqe,  rdma_wr(wr)->remote_addr,
wr                260 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 					       rdma_wr(wr)->rkey);
wr                265 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 				set_raddr_seg(wqe,  rdma_wr(wr)->remote_addr,
wr                266 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 					      rdma_wr(wr)->rkey);
wr                286 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 			if (wr->send_flags & IB_SEND_INLINE && wr->num_sge) {
wr                290 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 					*bad_wr = wr;
wr                296 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 				for (i = 0; i < wr->num_sge; i++) {
wr                298 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 					       wr->sg_list[i].addr),
wr                299 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 					       wr->sg_list[i].length);
wr                300 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 					wqe += wr->sg_list[i].length;
wr                305 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 				for (i = 0; i < wr->num_sge; i++)
wr                306 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 					set_data_seg(dseg + i, wr->sg_list + i);
wr                308 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 				ctrl->flag |= cpu_to_le32(wr->num_sge <<
wr                346 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 				 const struct ib_recv_wr *wr,
wr                365 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 	for (nreq = 0; wr; ++nreq, wr = wr->next) {
wr                369 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 			*bad_wr = wr;
wr                375 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 		if (unlikely(wr->num_sge > hr_qp->rq.max_gs)) {
wr                377 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 				wr->num_sge, hr_qp->rq.max_gs);
wr                379 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 			*bad_wr = wr;
wr                388 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 			       wr->num_sge);
wr                392 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 		for (i = 0; i < wr->num_sge; i++)
wr                393 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 			set_data_seg(scat + i, wr->sg_list + i);
wr                395 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 		hr_qp->rq.wrid[wqe_idx] = wr->wr_id;
wr                 61 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			 const struct ib_reg_wr *wr)
wr                 63 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	struct hns_roce_mr *mr = to_hr_mr(wr->mr);
wr                 68 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		     wr->access & IB_ACCESS_MW_BIND ? 1 : 0);
wr                 71 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		     wr->access & IB_ACCESS_REMOTE_ATOMIC ? 1 : 0);
wr                 74 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		     wr->access & IB_ACCESS_REMOTE_READ ? 1 : 0);
wr                 77 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		     wr->access & IB_ACCESS_REMOTE_WRITE ? 1 : 0);
wr                 80 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		     wr->access & IB_ACCESS_LOCAL_WRITE ? 1 : 0);
wr                 86 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	rc_sq_wqe->byte_16 = cpu_to_le32(wr->mr->length & 0xffffffff);
wr                 87 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	rc_sq_wqe->byte_20 = cpu_to_le32(wr->mr->length >> 32);
wr                 88 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	rc_sq_wqe->rkey = cpu_to_le32(wr->key);
wr                 89 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	rc_sq_wqe->va = cpu_to_le64(wr->mr->iova);
wr                101 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			   const struct ib_atomic_wr *wr)
wr                103 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	if (wr->wr.opcode == IB_WR_ATOMIC_CMP_AND_SWP) {
wr                104 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		aseg->fetchadd_swap_data = cpu_to_le64(wr->swap);
wr                105 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		aseg->cmp_data  = cpu_to_le64(wr->compare_add);
wr                107 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		aseg->fetchadd_swap_data = cpu_to_le64(wr->compare_add);
wr                112 drivers/infiniband/hw/hns/hns_roce_hw_v2.c static void set_extend_sge(struct hns_roce_qp *qp, const struct ib_send_wr *wr,
wr                127 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	sg = wr->sg_list + num_in_wqe;
wr                159 drivers/infiniband/hw/hns/hns_roce_hw_v2.c static int set_rwqe_data_seg(struct ib_qp *ibqp, const struct ib_send_wr *wr,
wr                171 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	if (wr->send_flags & IB_SEND_INLINE && valid_num_sge) {
wr                174 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			*bad_wr = wr;
wr                180 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		if (wr->opcode == IB_WR_RDMA_READ) {
wr                181 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			*bad_wr =  wr;
wr                186 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		for (i = 0; i < wr->num_sge; i++) {
wr                187 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			memcpy(wqe, ((void *)wr->sg_list[i].addr),
wr                188 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			       wr->sg_list[i].length);
wr                189 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			wqe += wr->sg_list[i].length;
wr                196 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			for (i = 0; i < wr->num_sge; i++) {
wr                197 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 				if (likely(wr->sg_list[i].length)) {
wr                198 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 					set_data_seg_v2(dseg, wr->sg_list + i);
wr                208 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			for (i = 0; i < wr->num_sge &&
wr                210 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 				if (likely(wr->sg_list[i].length)) {
wr                211 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 					set_data_seg_v2(dseg, wr->sg_list + i);
wr                217 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			set_extend_sge(qp, wr, sge_ind, valid_num_sge);
wr                234 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 				 const struct ib_send_wr *wr,
wr                238 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	struct hns_roce_ah *ah = to_hr_ah(ud_wr(wr)->ah);
wr                265 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		*bad_wr = wr;
wr                272 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		*bad_wr = wr;
wr                279 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	for (nreq = 0; wr; ++nreq, wr = wr->next) {
wr                282 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			*bad_wr = wr;
wr                288 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		if (unlikely(wr->num_sge > qp->sq.max_gs)) {
wr                290 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 				wr->num_sge, qp->sq.max_gs);
wr                292 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			*bad_wr = wr;
wr                297 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		qp->sq.wrid[wqe_idx] = wr->wr_id;
wr                303 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		for (i = 0; i < wr->num_sge; i++) {
wr                304 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			if (likely(wr->sg_list[i].length)) {
wr                305 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 				tmp_len += wr->sg_list[i].length;
wr                348 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			switch (wr->opcode) {
wr                352 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 				      cpu_to_le32(be32_to_cpu(wr->ex.imm_data));
wr                362 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 				   (wr->send_flags & IB_SEND_SIGNALED) ? 1 : 0);
wr                367 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 				  (wr->send_flags & IB_SEND_SOLICITED) ? 1 : 0);
wr                391 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			     cpu_to_le32(ud_wr(wr)->remote_qkey & 0x80000000 ?
wr                392 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			     qp->qkey : ud_wr(wr)->remote_qkey);
wr                396 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 				       ud_wr(wr)->remote_qpn);
wr                435 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			set_extend_sge(qp, wr, &sge_idx, valid_num_sge);
wr                443 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			switch (wr->opcode) {
wr                447 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 				      cpu_to_le32(be32_to_cpu(wr->ex.imm_data));
wr                451 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 					cpu_to_le32(wr->ex.invalidate_rkey);
wr                460 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 				     (wr->send_flags & IB_SEND_FENCE) ? 1 : 0);
wr                464 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 				  (wr->send_flags & IB_SEND_SOLICITED) ? 1 : 0);
wr                468 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 				   (wr->send_flags & IB_SEND_SIGNALED) ? 1 : 0);
wr                474 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			switch (wr->opcode) {
wr                478 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 					cpu_to_le32(rdma_wr(wr)->rkey);
wr                480 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 					cpu_to_le64(rdma_wr(wr)->remote_addr);
wr                485 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 					cpu_to_le32(rdma_wr(wr)->rkey);
wr                487 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 					cpu_to_le64(rdma_wr(wr)->remote_addr);
wr                492 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 					cpu_to_le32(rdma_wr(wr)->rkey);
wr                494 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 					cpu_to_le64(rdma_wr(wr)->remote_addr);
wr                510 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 					    cpu_to_le32(wr->ex.invalidate_rkey);
wr                515 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 				set_frmr_seg(rc_sq_wqe, fseg, reg_wr(wr));
wr                520 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 					cpu_to_le32(atomic_wr(wr)->rkey);
wr                522 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 					cpu_to_le64(atomic_wr(wr)->remote_addr);
wr                527 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 					cpu_to_le32(atomic_wr(wr)->rkey);
wr                529 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 					cpu_to_le64(atomic_wr(wr)->remote_addr);
wr                548 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			if (wr->opcode == IB_WR_ATOMIC_CMP_AND_SWP ||
wr                549 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			    wr->opcode == IB_WR_ATOMIC_FETCH_AND_ADD) {
wr                553 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 				set_data_seg_v2(dseg, wr->sg_list);
wr                555 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 				set_atomic_seg(wqe, atomic_wr(wr));
wr                560 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			} else if (wr->opcode != IB_WR_REG_MR) {
wr                561 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 				ret = set_rwqe_data_seg(ibqp, wr, rc_sq_wqe,
wr                570 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			*bad_wr = wr;
wr                606 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 				*bad_wr = wr;
wr                618 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 				 const struct ib_recv_wr *wr,
wr                639 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		*bad_wr = wr;
wr                643 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	for (nreq = 0; wr; ++nreq, wr = wr->next) {
wr                647 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			*bad_wr = wr;
wr                653 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		if (unlikely(wr->num_sge > hr_qp->rq.max_gs)) {
wr                655 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 				wr->num_sge, hr_qp->rq.max_gs);
wr                657 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			*bad_wr = wr;
wr                663 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		for (i = 0; i < wr->num_sge; i++) {
wr                664 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			if (!wr->sg_list[i].length)
wr                666 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			set_data_seg_v2(dseg, wr->sg_list + i);
wr                679 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 							       (u32)wr->num_sge;
wr                680 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			for (i = 0; i < wr->num_sge; i++) {
wr                682 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 					       (void *)(u64)wr->sg_list[i].addr;
wr                683 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 				sge_list[i].len = wr->sg_list[i].length;
wr                687 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		hr_qp->rq.wrid[wqe_idx] = wr->wr_id;
wr                707 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 				*bad_wr = wr;
wr               6230 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 				     const struct ib_recv_wr *wr,
wr               6249 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 	for (nreq = 0; wr; ++nreq, wr = wr->next) {
wr               6250 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		if (unlikely(wr->num_sge > srq->max_gs)) {
wr               6252 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			*bad_wr = wr;
wr               6258 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			*bad_wr = wr;
wr               6265 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			*bad_wr = wr;
wr               6273 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		for (i = 0; i < wr->num_sge; ++i) {
wr               6274 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			dseg[i].len = cpu_to_le32(wr->sg_list[i].length);
wr               6275 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			dseg[i].lkey = cpu_to_le32(wr->sg_list[i].lkey);
wr               6276 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 			dseg[i].addr = cpu_to_le64(wr->sg_list[i].addr);
wr               6285 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		srq->wrid[wqe_idx] = wr->wr_id;
wr                508 drivers/infiniband/hw/mlx4/mad.c 	struct ib_ud_wr wr;
wr                640 drivers/infiniband/hw/mlx4/mad.c 	wr.ah = ah;
wr                641 drivers/infiniband/hw/mlx4/mad.c 	wr.port_num = port;
wr                642 drivers/infiniband/hw/mlx4/mad.c 	wr.remote_qkey = IB_QP_SET_QKEY;
wr                643 drivers/infiniband/hw/mlx4/mad.c 	wr.remote_qpn = dqpn;
wr                644 drivers/infiniband/hw/mlx4/mad.c 	wr.wr.next = NULL;
wr                645 drivers/infiniband/hw/mlx4/mad.c 	wr.wr.wr_id = ((u64) tun_tx_ix) | MLX4_TUN_SET_WRID_QPN(dest_qpt);
wr                646 drivers/infiniband/hw/mlx4/mad.c 	wr.wr.sg_list = &list;
wr                647 drivers/infiniband/hw/mlx4/mad.c 	wr.wr.num_sge = 1;
wr                648 drivers/infiniband/hw/mlx4/mad.c 	wr.wr.opcode = IB_WR_SEND;
wr                649 drivers/infiniband/hw/mlx4/mad.c 	wr.wr.send_flags = IB_SEND_SIGNALED;
wr                651 drivers/infiniband/hw/mlx4/mad.c 	ret = ib_post_send(src_qp, &wr.wr, &bad_wr);
wr               1366 drivers/infiniband/hw/mlx4/mad.c 	struct ib_ud_wr wr;
wr               1439 drivers/infiniband/hw/mlx4/mad.c 	wr.ah = ah;
wr               1440 drivers/infiniband/hw/mlx4/mad.c 	wr.port_num = port;
wr               1441 drivers/infiniband/hw/mlx4/mad.c 	wr.pkey_index = wire_pkey_ix;
wr               1442 drivers/infiniband/hw/mlx4/mad.c 	wr.remote_qkey = qkey;
wr               1443 drivers/infiniband/hw/mlx4/mad.c 	wr.remote_qpn = remote_qpn;
wr               1444 drivers/infiniband/hw/mlx4/mad.c 	wr.wr.next = NULL;
wr               1445 drivers/infiniband/hw/mlx4/mad.c 	wr.wr.wr_id = ((u64) wire_tx_ix) | MLX4_TUN_SET_WRID_QPN(src_qpnum);
wr               1446 drivers/infiniband/hw/mlx4/mad.c 	wr.wr.sg_list = &list;
wr               1447 drivers/infiniband/hw/mlx4/mad.c 	wr.wr.num_sge = 1;
wr               1448 drivers/infiniband/hw/mlx4/mad.c 	wr.wr.opcode = IB_WR_SEND;
wr               1449 drivers/infiniband/hw/mlx4/mad.c 	wr.wr.send_flags = IB_SEND_SIGNALED;
wr               1451 drivers/infiniband/hw/mlx4/mad.c 	ret = ib_post_send(send_qp, &wr.wr, &bad_wr);
wr                768 drivers/infiniband/hw/mlx4/mlx4_ib.h int mlx4_ib_post_srq_recv(struct ib_srq *ibsrq, const struct ib_recv_wr *wr,
wr                781 drivers/infiniband/hw/mlx4/mlx4_ib.h int mlx4_ib_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr,
wr                783 drivers/infiniband/hw/mlx4/mlx4_ib.h int mlx4_ib_post_recv(struct ib_qp *ibqp, const struct ib_recv_wr *wr,
wr               2881 drivers/infiniband/hw/mlx4/qp.c 				  const struct ib_ud_wr *wr,
wr               2888 drivers/infiniband/hw/mlx4/qp.c 	struct mlx4_ib_ah *ah = to_mah(wr->ah);
wr               2897 drivers/infiniband/hw/mlx4/qp.c 	if (wr->wr.opcode != IB_WR_SEND)
wr               2902 drivers/infiniband/hw/mlx4/qp.c 	for (i = 0; i < wr->wr.num_sge; ++i)
wr               2903 drivers/infiniband/hw/mlx4/qp.c 		send_size += wr->wr.sg_list[i].length;
wr               2928 drivers/infiniband/hw/mlx4/qp.c 	sqp->ud_header.bth.solicited_event = !!(wr->wr.send_flags & IB_SEND_SOLICITED);
wr               2934 drivers/infiniband/hw/mlx4/qp.c 		sqp->ud_header.bth.destination_qpn = cpu_to_be32(wr->remote_qpn);
wr               3032 drivers/infiniband/hw/mlx4/qp.c static int build_mlx_header(struct mlx4_ib_sqp *sqp, const struct ib_ud_wr *wr,
wr               3040 drivers/infiniband/hw/mlx4/qp.c 	struct mlx4_ib_ah *ah = to_mah(wr->ah);
wr               3056 drivers/infiniband/hw/mlx4/qp.c 	for (i = 0; i < wr->wr.num_sge; ++i)
wr               3057 drivers/infiniband/hw/mlx4/qp.c 		send_size += wr->wr.sg_list[i].length;
wr               3167 drivers/infiniband/hw/mlx4/qp.c 	switch (wr->wr.opcode) {
wr               3175 drivers/infiniband/hw/mlx4/qp.c 		sqp->ud_header.immediate_data    = wr->wr.ex.imm_data;
wr               3216 drivers/infiniband/hw/mlx4/qp.c 	sqp->ud_header.bth.solicited_event = !!(wr->wr.send_flags & IB_SEND_SOLICITED);
wr               3221 drivers/infiniband/hw/mlx4/qp.c 		err = ib_get_cached_pkey(ib_dev, sqp->qp.port, wr->pkey_index,
wr               3227 drivers/infiniband/hw/mlx4/qp.c 	sqp->ud_header.bth.destination_qpn = cpu_to_be32(wr->remote_qpn);
wr               3229 drivers/infiniband/hw/mlx4/qp.c 	sqp->ud_header.deth.qkey = cpu_to_be32(wr->remote_qkey & 0x80000000 ?
wr               3230 drivers/infiniband/hw/mlx4/qp.c 					       sqp->qkey : wr->remote_qkey);
wr               3319 drivers/infiniband/hw/mlx4/qp.c 			const struct ib_reg_wr *wr)
wr               3321 drivers/infiniband/hw/mlx4/qp.c 	struct mlx4_ib_mr *mr = to_mmr(wr->mr);
wr               3323 drivers/infiniband/hw/mlx4/qp.c 	fseg->flags		= convert_access(wr->access);
wr               3324 drivers/infiniband/hw/mlx4/qp.c 	fseg->mem_key		= cpu_to_be32(wr->key);
wr               3349 drivers/infiniband/hw/mlx4/qp.c 			   const struct ib_atomic_wr *wr)
wr               3351 drivers/infiniband/hw/mlx4/qp.c 	if (wr->wr.opcode == IB_WR_ATOMIC_CMP_AND_SWP) {
wr               3352 drivers/infiniband/hw/mlx4/qp.c 		aseg->swap_add = cpu_to_be64(wr->swap);
wr               3353 drivers/infiniband/hw/mlx4/qp.c 		aseg->compare  = cpu_to_be64(wr->compare_add);
wr               3354 drivers/infiniband/hw/mlx4/qp.c 	} else if (wr->wr.opcode == IB_WR_MASKED_ATOMIC_FETCH_AND_ADD) {
wr               3355 drivers/infiniband/hw/mlx4/qp.c 		aseg->swap_add = cpu_to_be64(wr->compare_add);
wr               3356 drivers/infiniband/hw/mlx4/qp.c 		aseg->compare  = cpu_to_be64(wr->compare_add_mask);
wr               3358 drivers/infiniband/hw/mlx4/qp.c 		aseg->swap_add = cpu_to_be64(wr->compare_add);
wr               3365 drivers/infiniband/hw/mlx4/qp.c 				  const struct ib_atomic_wr *wr)
wr               3367 drivers/infiniband/hw/mlx4/qp.c 	aseg->swap_add		= cpu_to_be64(wr->swap);
wr               3368 drivers/infiniband/hw/mlx4/qp.c 	aseg->swap_add_mask	= cpu_to_be64(wr->swap_mask);
wr               3369 drivers/infiniband/hw/mlx4/qp.c 	aseg->compare		= cpu_to_be64(wr->compare_add);
wr               3370 drivers/infiniband/hw/mlx4/qp.c 	aseg->compare_mask	= cpu_to_be64(wr->compare_add_mask);
wr               3374 drivers/infiniband/hw/mlx4/qp.c 			     const struct ib_ud_wr *wr)
wr               3376 drivers/infiniband/hw/mlx4/qp.c 	memcpy(dseg->av, &to_mah(wr->ah)->av, sizeof (struct mlx4_av));
wr               3377 drivers/infiniband/hw/mlx4/qp.c 	dseg->dqpn = cpu_to_be32(wr->remote_qpn);
wr               3378 drivers/infiniband/hw/mlx4/qp.c 	dseg->qkey = cpu_to_be32(wr->remote_qkey);
wr               3379 drivers/infiniband/hw/mlx4/qp.c 	dseg->vlan = to_mah(wr->ah)->av.eth.vlan;
wr               3380 drivers/infiniband/hw/mlx4/qp.c 	memcpy(dseg->mac, to_mah(wr->ah)->av.eth.mac, 6);
wr               3385 drivers/infiniband/hw/mlx4/qp.c 				    const struct ib_ud_wr *wr,
wr               3388 drivers/infiniband/hw/mlx4/qp.c 	union mlx4_ext_av *av = &to_mah(wr->ah)->av;
wr               3407 drivers/infiniband/hw/mlx4/qp.c static void build_tunnel_header(const struct ib_ud_wr *wr, void *wqe,
wr               3412 drivers/infiniband/hw/mlx4/qp.c 	struct mlx4_ib_ah *ah = to_mah(wr->ah);
wr               3417 drivers/infiniband/hw/mlx4/qp.c 	hdr.remote_qpn = cpu_to_be32(wr->remote_qpn);
wr               3418 drivers/infiniband/hw/mlx4/qp.c 	hdr.pkey_index = cpu_to_be16(wr->pkey_index);
wr               3419 drivers/infiniband/hw/mlx4/qp.c 	hdr.qkey = cpu_to_be32(wr->remote_qkey);
wr               3492 drivers/infiniband/hw/mlx4/qp.c 			 const struct ib_ud_wr *wr, struct mlx4_ib_qp *qp,
wr               3495 drivers/infiniband/hw/mlx4/qp.c 	unsigned halign = ALIGN(sizeof *wqe + wr->hlen, 16);
wr               3501 drivers/infiniband/hw/mlx4/qp.c 		     wr->wr.num_sge > qp->sq.max_gs - (halign >> 4)))
wr               3504 drivers/infiniband/hw/mlx4/qp.c 	memcpy(wqe->header, wr->header, wr->hlen);
wr               3506 drivers/infiniband/hw/mlx4/qp.c 	*lso_hdr_sz  = cpu_to_be32(wr->mss << 16 | wr->hlen);
wr               3511 drivers/infiniband/hw/mlx4/qp.c static __be32 send_ieth(const struct ib_send_wr *wr)
wr               3513 drivers/infiniband/hw/mlx4/qp.c 	switch (wr->opcode) {
wr               3516 drivers/infiniband/hw/mlx4/qp.c 		return wr->ex.imm_data;
wr               3519 drivers/infiniband/hw/mlx4/qp.c 		return cpu_to_be32(wr->ex.invalidate_rkey);
wr               3533 drivers/infiniband/hw/mlx4/qp.c static int _mlx4_ib_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr,
wr               3557 drivers/infiniband/hw/mlx4/qp.c 			struct mlx4_ib_ah *ah = to_mah(ud_wr(wr)->ah);
wr               3576 drivers/infiniband/hw/mlx4/qp.c 		*bad_wr = wr;
wr               3583 drivers/infiniband/hw/mlx4/qp.c 	for (nreq = 0; wr; ++nreq, wr = wr->next) {
wr               3589 drivers/infiniband/hw/mlx4/qp.c 			*bad_wr = wr;
wr               3593 drivers/infiniband/hw/mlx4/qp.c 		if (unlikely(wr->num_sge > qp->sq.max_gs)) {
wr               3595 drivers/infiniband/hw/mlx4/qp.c 			*bad_wr = wr;
wr               3600 drivers/infiniband/hw/mlx4/qp.c 		qp->sq.wrid[(qp->sq.head + nreq) & (qp->sq.wqe_cnt - 1)] = wr->wr_id;
wr               3603 drivers/infiniband/hw/mlx4/qp.c 			(wr->send_flags & IB_SEND_SIGNALED ?
wr               3605 drivers/infiniband/hw/mlx4/qp.c 			(wr->send_flags & IB_SEND_SOLICITED ?
wr               3607 drivers/infiniband/hw/mlx4/qp.c 			((wr->send_flags & IB_SEND_IP_CSUM) ?
wr               3612 drivers/infiniband/hw/mlx4/qp.c 		ctrl->imm = send_ieth(wr);
wr               3620 drivers/infiniband/hw/mlx4/qp.c 			switch (wr->opcode) {
wr               3624 drivers/infiniband/hw/mlx4/qp.c 				set_raddr_seg(wqe, atomic_wr(wr)->remote_addr,
wr               3625 drivers/infiniband/hw/mlx4/qp.c 					      atomic_wr(wr)->rkey);
wr               3628 drivers/infiniband/hw/mlx4/qp.c 				set_atomic_seg(wqe, atomic_wr(wr));
wr               3637 drivers/infiniband/hw/mlx4/qp.c 				set_raddr_seg(wqe, atomic_wr(wr)->remote_addr,
wr               3638 drivers/infiniband/hw/mlx4/qp.c 					      atomic_wr(wr)->rkey);
wr               3641 drivers/infiniband/hw/mlx4/qp.c 				set_masked_atomic_seg(wqe, atomic_wr(wr));
wr               3652 drivers/infiniband/hw/mlx4/qp.c 				set_raddr_seg(wqe, rdma_wr(wr)->remote_addr,
wr               3653 drivers/infiniband/hw/mlx4/qp.c 					      rdma_wr(wr)->rkey);
wr               3661 drivers/infiniband/hw/mlx4/qp.c 				set_local_inv_seg(wqe, wr->ex.invalidate_rkey);
wr               3669 drivers/infiniband/hw/mlx4/qp.c 				set_reg_seg(wqe, reg_wr(wr));
wr               3681 drivers/infiniband/hw/mlx4/qp.c 			err =  build_sriov_qp0_header(to_msqp(qp), ud_wr(wr),
wr               3684 drivers/infiniband/hw/mlx4/qp.c 				*bad_wr = wr;
wr               3693 drivers/infiniband/hw/mlx4/qp.c 			set_datagram_seg(wqe, ud_wr(wr));
wr               3700 drivers/infiniband/hw/mlx4/qp.c 			set_datagram_seg(wqe, ud_wr(wr));
wr               3704 drivers/infiniband/hw/mlx4/qp.c 			if (wr->opcode == IB_WR_LSO) {
wr               3705 drivers/infiniband/hw/mlx4/qp.c 				err = build_lso_seg(wqe, ud_wr(wr), qp, &seglen,
wr               3708 drivers/infiniband/hw/mlx4/qp.c 					*bad_wr = wr;
wr               3718 drivers/infiniband/hw/mlx4/qp.c 			err = build_sriov_qp0_header(to_msqp(qp), ud_wr(wr),
wr               3721 drivers/infiniband/hw/mlx4/qp.c 				*bad_wr = wr;
wr               3730 drivers/infiniband/hw/mlx4/qp.c 			build_tunnel_header(ud_wr(wr), wqe, &seglen);
wr               3741 drivers/infiniband/hw/mlx4/qp.c 						ud_wr(wr),
wr               3745 drivers/infiniband/hw/mlx4/qp.c 			build_tunnel_header(ud_wr(wr), wqe, &seglen);
wr               3752 drivers/infiniband/hw/mlx4/qp.c 			err = build_mlx_header(to_msqp(qp), ud_wr(wr), ctrl,
wr               3755 drivers/infiniband/hw/mlx4/qp.c 				*bad_wr = wr;
wr               3774 drivers/infiniband/hw/mlx4/qp.c 		dseg += wr->num_sge - 1;
wr               3775 drivers/infiniband/hw/mlx4/qp.c 		size += wr->num_sge * (sizeof (struct mlx4_wqe_data_seg) / 16);
wr               3786 drivers/infiniband/hw/mlx4/qp.c 		for (i = wr->num_sge - 1; i >= 0; --i, --dseg)
wr               3787 drivers/infiniband/hw/mlx4/qp.c 			set_data_seg(dseg, wr->sg_list + i);
wr               3797 drivers/infiniband/hw/mlx4/qp.c 		ctrl->qpn_vlan.fence_size = (wr->send_flags & IB_SEND_FENCE ?
wr               3807 drivers/infiniband/hw/mlx4/qp.c 		if (wr->opcode < 0 || wr->opcode >= ARRAY_SIZE(mlx4_ib_opcode)) {
wr               3808 drivers/infiniband/hw/mlx4/qp.c 			*bad_wr = wr;
wr               3813 drivers/infiniband/hw/mlx4/qp.c 		ctrl->owner_opcode = mlx4_ib_opcode[wr->opcode] |
wr               3821 drivers/infiniband/hw/mlx4/qp.c 		if (wr->next)
wr               3849 drivers/infiniband/hw/mlx4/qp.c int mlx4_ib_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr,
wr               3852 drivers/infiniband/hw/mlx4/qp.c 	return _mlx4_ib_post_send(ibqp, wr, bad_wr, false);
wr               3855 drivers/infiniband/hw/mlx4/qp.c static int _mlx4_ib_post_recv(struct ib_qp *ibqp, const struct ib_recv_wr *wr,
wr               3874 drivers/infiniband/hw/mlx4/qp.c 		*bad_wr = wr;
wr               3881 drivers/infiniband/hw/mlx4/qp.c 	for (nreq = 0; wr; ++nreq, wr = wr->next) {
wr               3884 drivers/infiniband/hw/mlx4/qp.c 			*bad_wr = wr;
wr               3888 drivers/infiniband/hw/mlx4/qp.c 		if (unlikely(wr->num_sge > qp->rq.max_gs)) {
wr               3890 drivers/infiniband/hw/mlx4/qp.c 			*bad_wr = wr;
wr               3905 drivers/infiniband/hw/mlx4/qp.c 			scat->lkey = cpu_to_be32(wr->sg_list->lkey);
wr               3911 drivers/infiniband/hw/mlx4/qp.c 		for (i = 0; i < wr->num_sge; ++i)
wr               3912 drivers/infiniband/hw/mlx4/qp.c 			__set_data_seg(scat + i, wr->sg_list + i);
wr               3920 drivers/infiniband/hw/mlx4/qp.c 		qp->rq.wrid[ind] = wr->wr_id;
wr               3943 drivers/infiniband/hw/mlx4/qp.c int mlx4_ib_post_recv(struct ib_qp *ibqp, const struct ib_recv_wr *wr,
wr               3946 drivers/infiniband/hw/mlx4/qp.c 	return _mlx4_ib_post_recv(ibqp, wr, bad_wr, false);
wr               4484 drivers/infiniband/hw/mlx4/qp.c 		.wr = {
wr               4503 drivers/infiniband/hw/mlx4/qp.c 	ret = _mlx4_ib_post_send(qp, &swr.wr, &bad_swr, true);
wr                300 drivers/infiniband/hw/mlx4/srq.c int mlx4_ib_post_srq_recv(struct ib_srq *ibsrq, const struct ib_recv_wr *wr,
wr                315 drivers/infiniband/hw/mlx4/srq.c 		*bad_wr = wr;
wr                320 drivers/infiniband/hw/mlx4/srq.c 	for (nreq = 0; wr; ++nreq, wr = wr->next) {
wr                321 drivers/infiniband/hw/mlx4/srq.c 		if (unlikely(wr->num_sge > srq->msrq.max_gs)) {
wr                323 drivers/infiniband/hw/mlx4/srq.c 			*bad_wr = wr;
wr                329 drivers/infiniband/hw/mlx4/srq.c 			*bad_wr = wr;
wr                333 drivers/infiniband/hw/mlx4/srq.c 		srq->wrid[srq->head] = wr->wr_id;
wr                339 drivers/infiniband/hw/mlx4/srq.c 		for (i = 0; i < wr->num_sge; ++i) {
wr                340 drivers/infiniband/hw/mlx4/srq.c 			scat[i].byte_count = cpu_to_be32(wr->sg_list[i].length);
wr                341 drivers/infiniband/hw/mlx4/srq.c 			scat[i].lkey       = cpu_to_be32(wr->sg_list[i].lkey);
wr                342 drivers/infiniband/hw/mlx4/srq.c 			scat[i].addr       = cpu_to_be64(wr->sg_list[i].addr);
wr                 76 drivers/infiniband/hw/mlx5/gsi.c 	struct mlx5_ib_gsi_wr *wr;
wr                 81 drivers/infiniband/hw/mlx5/gsi.c 		wr = &gsi->outstanding_wrs[index % gsi->cap.max_send_wr];
wr                 83 drivers/infiniband/hw/mlx5/gsi.c 		if (!wr->completed)
wr                 87 drivers/infiniband/hw/mlx5/gsi.c 		    wr->send_flags & IB_SEND_SIGNALED)
wr                 88 drivers/infiniband/hw/mlx5/gsi.c 			WARN_ON_ONCE(mlx5_ib_generate_wc(gsi_cq, &wr->wc));
wr                 90 drivers/infiniband/hw/mlx5/gsi.c 		wr->completed = false;
wr                 99 drivers/infiniband/hw/mlx5/gsi.c 	struct mlx5_ib_gsi_wr *wr =
wr                105 drivers/infiniband/hw/mlx5/gsi.c 	wr->completed = true;
wr                106 drivers/infiniband/hw/mlx5/gsi.c 	wr_id = wr->wc.wr_id;
wr                107 drivers/infiniband/hw/mlx5/gsi.c 	wr->wc = *wc;
wr                108 drivers/infiniband/hw/mlx5/gsi.c 	wr->wc.wr_id = wr_id;
wr                109 drivers/infiniband/hw/mlx5/gsi.c 	wr->wc.qp = &gsi->ibqp;
wr                415 drivers/infiniband/hw/mlx5/gsi.c 				      struct ib_ud_wr *wr, struct ib_wc *wc)
wr                431 drivers/infiniband/hw/mlx5/gsi.c 		gsi_wr->wc.pkey_index = wr->pkey_index;
wr                432 drivers/infiniband/hw/mlx5/gsi.c 		gsi_wr->wc.wr_id = wr->wr.wr_id;
wr                439 drivers/infiniband/hw/mlx5/gsi.c 	wr->wr.wr_cqe = &gsi_wr->cqe;
wr                446 drivers/infiniband/hw/mlx5/gsi.c 				    struct ib_ud_wr *wr)
wr                449 drivers/infiniband/hw/mlx5/gsi.c 		{ .wr_id = wr->wr.wr_id },
wr                456 drivers/infiniband/hw/mlx5/gsi.c 	ret = mlx5_ib_add_outstanding_wr(gsi, wr, &wc);
wr                466 drivers/infiniband/hw/mlx5/gsi.c static struct ib_qp *get_tx_qp(struct mlx5_ib_gsi_qp *gsi, struct ib_ud_wr *wr)
wr                469 drivers/infiniband/hw/mlx5/gsi.c 	int qp_index = wr->pkey_index;
wr                480 drivers/infiniband/hw/mlx5/gsi.c int mlx5_ib_gsi_post_send(struct ib_qp *qp, const struct ib_send_wr *wr,
wr                488 drivers/infiniband/hw/mlx5/gsi.c 	for (; wr; wr = wr->next) {
wr                489 drivers/infiniband/hw/mlx5/gsi.c 		struct ib_ud_wr cur_wr = *ud_wr(wr);
wr                491 drivers/infiniband/hw/mlx5/gsi.c 		cur_wr.wr.next = NULL;
wr                507 drivers/infiniband/hw/mlx5/gsi.c 		ret = ib_post_send(tx_qp, &cur_wr.wr, bad_wr);
wr                520 drivers/infiniband/hw/mlx5/gsi.c 	*bad_wr = wr;
wr                524 drivers/infiniband/hw/mlx5/gsi.c int mlx5_ib_gsi_post_recv(struct ib_qp *qp, const struct ib_recv_wr *wr,
wr                529 drivers/infiniband/hw/mlx5/gsi.c 	return ib_post_recv(gsi->rx_qp, wr, bad_wr);
wr                477 drivers/infiniband/hw/mlx5/mlx5_ib.h 	struct ib_send_wr		wr;
wr                489 drivers/infiniband/hw/mlx5/mlx5_ib.h static inline const struct mlx5_umr_wr *umr_wr(const struct ib_send_wr *wr)
wr                491 drivers/infiniband/hw/mlx5/mlx5_ib.h 	return container_of(wr, struct mlx5_umr_wr, wr);
wr               1116 drivers/infiniband/hw/mlx5/mlx5_ib.h int mlx5_ib_post_srq_recv(struct ib_srq *ibsrq, const struct ib_recv_wr *wr,
wr               1130 drivers/infiniband/hw/mlx5/mlx5_ib.h int mlx5_ib_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr,
wr               1132 drivers/infiniband/hw/mlx5/mlx5_ib.h int mlx5_ib_post_recv(struct ib_qp *ibqp, const struct ib_recv_wr *wr,
wr               1322 drivers/infiniband/hw/mlx5/mlx5_ib.h int mlx5_ib_gsi_post_send(struct ib_qp *qp, const struct ib_send_wr *wr,
wr               1324 drivers/infiniband/hw/mlx5/mlx5_ib.h int mlx5_ib_gsi_post_recv(struct ib_qp *qp, const struct ib_recv_wr *wr,
wr                831 drivers/infiniband/hw/mlx5/mr.c 	umrwr->wr.wr_cqe = &umr_context.cqe;
wr                834 drivers/infiniband/hw/mlx5/mr.c 	err = ib_post_send(umrc->qp, &umrwr->wr, &bad);
wr                927 drivers/infiniband/hw/mlx5/mr.c 	struct mlx5_umr_wr wr;
wr                987 drivers/infiniband/hw/mlx5/mr.c 	memset(&wr, 0, sizeof(wr));
wr                988 drivers/infiniband/hw/mlx5/mr.c 	wr.wr.send_flags = MLX5_IB_SEND_UMR_UPDATE_XLT;
wr                990 drivers/infiniband/hw/mlx5/mr.c 		wr.wr.send_flags |= MLX5_IB_SEND_UMR_FAIL_IF_FREE;
wr                991 drivers/infiniband/hw/mlx5/mr.c 	wr.wr.sg_list = &sg;
wr                992 drivers/infiniband/hw/mlx5/mr.c 	wr.wr.num_sge = 1;
wr                993 drivers/infiniband/hw/mlx5/mr.c 	wr.wr.opcode = MLX5_IB_WR_UMR;
wr                995 drivers/infiniband/hw/mlx5/mr.c 	wr.pd = mr->ibmr.pd;
wr                996 drivers/infiniband/hw/mlx5/mr.c 	wr.mkey = mr->mmkey.key;
wr                997 drivers/infiniband/hw/mlx5/mr.c 	wr.length = mr->mmkey.size;
wr                998 drivers/infiniband/hw/mlx5/mr.c 	wr.virt_addr = mr->mmkey.iova;
wr                999 drivers/infiniband/hw/mlx5/mr.c 	wr.access_flags = mr->access_flags;
wr               1000 drivers/infiniband/hw/mlx5/mr.c 	wr.page_shift = page_shift;
wr               1017 drivers/infiniband/hw/mlx5/mr.c 				wr.wr.send_flags |=
wr               1023 drivers/infiniband/hw/mlx5/mr.c 				wr.wr.send_flags |=
wr               1026 drivers/infiniband/hw/mlx5/mr.c 				wr.wr.send_flags |=
wr               1030 drivers/infiniband/hw/mlx5/mr.c 		wr.offset = idx * desc_size;
wr               1031 drivers/infiniband/hw/mlx5/mr.c 		wr.xlt_size = sg.length;
wr               1033 drivers/infiniband/hw/mlx5/mr.c 		err = mlx5_ib_post_send_wait(dev, &wr);
wr               1359 drivers/infiniband/hw/mlx5/mr.c 	umrwr.wr.send_flags = MLX5_IB_SEND_UMR_DISABLE_MR |
wr               1361 drivers/infiniband/hw/mlx5/mr.c 	umrwr.wr.opcode = MLX5_IB_WR_UMR;
wr               1376 drivers/infiniband/hw/mlx5/mr.c 	umrwr.wr.send_flags = MLX5_IB_SEND_UMR_FAIL_IF_FREE;
wr               1378 drivers/infiniband/hw/mlx5/mr.c 	umrwr.wr.opcode = MLX5_IB_WR_UMR;
wr               1384 drivers/infiniband/hw/mlx5/mr.c 		umrwr.wr.send_flags |= MLX5_IB_SEND_UMR_UPDATE_PD_ACCESS;
wr               4095 drivers/infiniband/hw/mlx5/qp.c static void set_eth_seg(const struct ib_send_wr *wr, struct mlx5_ib_qp *qp,
wr               4102 drivers/infiniband/hw/mlx5/qp.c 	if (wr->send_flags & IB_SEND_IP_CSUM)
wr               4106 drivers/infiniband/hw/mlx5/qp.c 	if (wr->opcode == IB_WR_LSO) {
wr               4107 drivers/infiniband/hw/mlx5/qp.c 		struct ib_ud_wr *ud_wr = container_of(wr, struct ib_ud_wr, wr);
wr               4144 drivers/infiniband/hw/mlx5/qp.c 			     const struct ib_send_wr *wr)
wr               4146 drivers/infiniband/hw/mlx5/qp.c 	memcpy(&dseg->av, &to_mah(ud_wr(wr)->ah)->av, sizeof(struct mlx5_av));
wr               4147 drivers/infiniband/hw/mlx5/qp.c 	dseg->av.dqp_dct = cpu_to_be32(ud_wr(wr)->remote_qpn | MLX5_EXTENDED_UD_AV);
wr               4148 drivers/infiniband/hw/mlx5/qp.c 	dseg->av.key.qkey.qkey = cpu_to_be32(ud_wr(wr)->remote_qkey);
wr               4292 drivers/infiniband/hw/mlx5/qp.c 			       const struct ib_send_wr *wr, int atomic)
wr               4294 drivers/infiniband/hw/mlx5/qp.c 	const struct mlx5_umr_wr *umrwr = umr_wr(wr);
wr               4299 drivers/infiniband/hw/mlx5/qp.c 		if (wr->send_flags & MLX5_IB_SEND_UMR_FAIL_IF_FREE)
wr               4308 drivers/infiniband/hw/mlx5/qp.c 	if (wr->send_flags & MLX5_IB_SEND_UMR_UPDATE_XLT) {
wr               4315 drivers/infiniband/hw/mlx5/qp.c 	if (wr->send_flags & MLX5_IB_SEND_UMR_UPDATE_TRANSLATION)
wr               4317 drivers/infiniband/hw/mlx5/qp.c 	if (wr->send_flags & MLX5_IB_SEND_UMR_UPDATE_PD_ACCESS) {
wr               4321 drivers/infiniband/hw/mlx5/qp.c 	if (wr->send_flags & MLX5_IB_SEND_UMR_ENABLE_MR)
wr               4323 drivers/infiniband/hw/mlx5/qp.c 	if (wr->send_flags & MLX5_IB_SEND_UMR_DISABLE_MR)
wr               4326 drivers/infiniband/hw/mlx5/qp.c 	if (!wr->num_sge)
wr               4370 drivers/infiniband/hw/mlx5/qp.c 				 const struct ib_send_wr *wr)
wr               4372 drivers/infiniband/hw/mlx5/qp.c 	const struct mlx5_umr_wr *umrwr = umr_wr(wr);
wr               4375 drivers/infiniband/hw/mlx5/qp.c 	if (wr->send_flags & MLX5_IB_SEND_UMR_DISABLE_MR)
wr               4381 drivers/infiniband/hw/mlx5/qp.c 	if (wr->send_flags & MLX5_IB_SEND_UMR_UPDATE_TRANSLATION &&
wr               4403 drivers/infiniband/hw/mlx5/qp.c static __be32 send_ieth(const struct ib_send_wr *wr)
wr               4405 drivers/infiniband/hw/mlx5/qp.c 	switch (wr->opcode) {
wr               4408 drivers/infiniband/hw/mlx5/qp.c 		return wr->ex.imm_data;
wr               4411 drivers/infiniband/hw/mlx5/qp.c 		return cpu_to_be32(wr->ex.invalidate_rkey);
wr               4435 drivers/infiniband/hw/mlx5/qp.c static int set_data_inl_seg(struct mlx5_ib_qp *qp, const struct ib_send_wr *wr,
wr               4447 drivers/infiniband/hw/mlx5/qp.c 	for (i = 0; i < wr->num_sge; i++) {
wr               4448 drivers/infiniband/hw/mlx5/qp.c 		size_t len  = wr->sg_list[i].length;
wr               4449 drivers/infiniband/hw/mlx5/qp.c 		void *addr = (void *)(unsigned long)(wr->sg_list[i].addr);
wr               4734 drivers/infiniband/hw/mlx5/qp.c 	const struct ib_reg_wr *wr = reg_wr(send_wr);
wr               4735 drivers/infiniband/hw/mlx5/qp.c 	struct mlx5_ib_mr *sig_mr = to_mmr(wr->mr);
wr               4743 drivers/infiniband/hw/mlx5/qp.c 	    unlikely(wr->access & IB_ACCESS_REMOTE_ATOMIC) ||
wr               4766 drivers/infiniband/hw/mlx5/qp.c 	set_sig_mkey_segment(*seg, wr->mr, wr->access, xlt_size, region_len,
wr               4772 drivers/infiniband/hw/mlx5/qp.c 	ret = set_sig_data_segment(send_wr, wr->mr, sig_attrs, qp, seg, size,
wr               4809 drivers/infiniband/hw/mlx5/qp.c 		      const struct ib_reg_wr *wr,
wr               4813 drivers/infiniband/hw/mlx5/qp.c 	struct mlx5_ib_mr *mr = to_mmr(wr->mr);
wr               4818 drivers/infiniband/hw/mlx5/qp.c 	bool atomic = wr->access & IB_ACCESS_REMOTE_ATOMIC;
wr               4831 drivers/infiniband/hw/mlx5/qp.c 	if (unlikely(wr->wr.send_flags & IB_SEND_INLINE)) {
wr               4847 drivers/infiniband/hw/mlx5/qp.c 	set_reg_mkey_seg(*seg, mr, wr->key, wr->access);
wr               4898 drivers/infiniband/hw/mlx5/qp.c 		       const struct ib_send_wr *wr, unsigned int *idx,
wr               4909 drivers/infiniband/hw/mlx5/qp.c 	(*ctrl)->imm = send_ieth(wr);
wr               4923 drivers/infiniband/hw/mlx5/qp.c 		     const struct ib_send_wr *wr, unsigned *idx,
wr               4926 drivers/infiniband/hw/mlx5/qp.c 	return __begin_wqe(qp, seg, ctrl, wr, idx, size, cur_edge, nreq,
wr               4927 drivers/infiniband/hw/mlx5/qp.c 			   wr->send_flags & IB_SEND_SIGNALED,
wr               4928 drivers/infiniband/hw/mlx5/qp.c 			   wr->send_flags & IB_SEND_SOLICITED);
wr               4962 drivers/infiniband/hw/mlx5/qp.c static int _mlx5_ib_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr,
wr               4990 drivers/infiniband/hw/mlx5/qp.c 		*bad_wr = wr;
wr               4995 drivers/infiniband/hw/mlx5/qp.c 		return mlx5_ib_gsi_post_send(ibqp, wr, bad_wr);
wr               5002 drivers/infiniband/hw/mlx5/qp.c 	for (nreq = 0; wr; nreq++, wr = wr->next) {
wr               5003 drivers/infiniband/hw/mlx5/qp.c 		if (unlikely(wr->opcode >= ARRAY_SIZE(mlx5_ib_opcode))) {
wr               5006 drivers/infiniband/hw/mlx5/qp.c 			*bad_wr = wr;
wr               5010 drivers/infiniband/hw/mlx5/qp.c 		num_sge = wr->num_sge;
wr               5014 drivers/infiniband/hw/mlx5/qp.c 			*bad_wr = wr;
wr               5018 drivers/infiniband/hw/mlx5/qp.c 		err = begin_wqe(qp, &seg, &ctrl, wr, &idx, &size, &cur_edge,
wr               5023 drivers/infiniband/hw/mlx5/qp.c 			*bad_wr = wr;
wr               5027 drivers/infiniband/hw/mlx5/qp.c 		if (wr->opcode == IB_WR_REG_MR ||
wr               5028 drivers/infiniband/hw/mlx5/qp.c 		    wr->opcode == IB_WR_REG_MR_INTEGRITY) {
wr               5032 drivers/infiniband/hw/mlx5/qp.c 			if (wr->send_flags & IB_SEND_FENCE) {
wr               5049 drivers/infiniband/hw/mlx5/qp.c 			switch (wr->opcode) {
wr               5053 drivers/infiniband/hw/mlx5/qp.c 				set_raddr_seg(seg, rdma_wr(wr)->remote_addr,
wr               5054 drivers/infiniband/hw/mlx5/qp.c 					      rdma_wr(wr)->rkey);
wr               5064 drivers/infiniband/hw/mlx5/qp.c 				*bad_wr = wr;
wr               5069 drivers/infiniband/hw/mlx5/qp.c 				ctrl->imm = cpu_to_be32(wr->ex.invalidate_rkey);
wr               5076 drivers/infiniband/hw/mlx5/qp.c 				ctrl->imm = cpu_to_be32(reg_wr(wr)->key);
wr               5077 drivers/infiniband/hw/mlx5/qp.c 				err = set_reg_wr(qp, reg_wr(wr), &seg, &size,
wr               5080 drivers/infiniband/hw/mlx5/qp.c 					*bad_wr = wr;
wr               5089 drivers/infiniband/hw/mlx5/qp.c 				mr = to_mmr(reg_wr(wr)->mr);
wr               5097 drivers/infiniband/hw/mlx5/qp.c 					reg_pi_wr.access = reg_wr(wr)->access;
wr               5106 drivers/infiniband/hw/mlx5/qp.c 						*bad_wr = wr;
wr               5110 drivers/infiniband/hw/mlx5/qp.c 						   cur_edge, idx, wr->wr_id,
wr               5114 drivers/infiniband/hw/mlx5/qp.c 					err = begin_wqe(qp, &seg, &ctrl, wr,
wr               5120 drivers/infiniband/hw/mlx5/qp.c 						*bad_wr = wr;
wr               5146 drivers/infiniband/hw/mlx5/qp.c 				err = set_pi_umr_wr(wr, qp, &seg, &size,
wr               5150 drivers/infiniband/hw/mlx5/qp.c 					*bad_wr = wr;
wr               5154 drivers/infiniband/hw/mlx5/qp.c 					   wr->wr_id, nreq, fence,
wr               5162 drivers/infiniband/hw/mlx5/qp.c 				err = __begin_wqe(qp, &seg, &ctrl, wr, &idx,
wr               5168 drivers/infiniband/hw/mlx5/qp.c 					*bad_wr = wr;
wr               5176 drivers/infiniband/hw/mlx5/qp.c 					*bad_wr = wr;
wr               5180 drivers/infiniband/hw/mlx5/qp.c 					   wr->wr_id, nreq, next_fence,
wr               5183 drivers/infiniband/hw/mlx5/qp.c 				err = __begin_wqe(qp, &seg, &ctrl, wr, &idx,
wr               5189 drivers/infiniband/hw/mlx5/qp.c 					*bad_wr = wr;
wr               5197 drivers/infiniband/hw/mlx5/qp.c 					*bad_wr = wr;
wr               5201 drivers/infiniband/hw/mlx5/qp.c 					   wr->wr_id, nreq, next_fence,
wr               5215 drivers/infiniband/hw/mlx5/qp.c 			switch (wr->opcode) {
wr               5218 drivers/infiniband/hw/mlx5/qp.c 				set_raddr_seg(seg, rdma_wr(wr)->remote_addr,
wr               5219 drivers/infiniband/hw/mlx5/qp.c 					      rdma_wr(wr)->rkey);
wr               5233 drivers/infiniband/hw/mlx5/qp.c 				*bad_wr = wr;
wr               5238 drivers/infiniband/hw/mlx5/qp.c 			set_datagram_seg(seg, wr);
wr               5245 drivers/infiniband/hw/mlx5/qp.c 			set_datagram_seg(seg, wr);
wr               5258 drivers/infiniband/hw/mlx5/qp.c 				set_eth_seg(wr, qp, &seg, &size, &cur_edge);
wr               5264 drivers/infiniband/hw/mlx5/qp.c 			if (wr->opcode != MLX5_IB_WR_UMR) {
wr               5270 drivers/infiniband/hw/mlx5/qp.c 			ctrl->imm = cpu_to_be32(umr_wr(wr)->mkey);
wr               5271 drivers/infiniband/hw/mlx5/qp.c 			err = set_reg_umr_segment(dev, seg, wr, !!(MLX5_CAP_GEN(mdev, atomic)));
wr               5277 drivers/infiniband/hw/mlx5/qp.c 			set_reg_mkey_segment(seg, wr);
wr               5287 drivers/infiniband/hw/mlx5/qp.c 		if (wr->send_flags & IB_SEND_INLINE && num_sge) {
wr               5288 drivers/infiniband/hw/mlx5/qp.c 			err = set_data_inl_seg(qp, wr, &seg, &size, &cur_edge);
wr               5291 drivers/infiniband/hw/mlx5/qp.c 				*bad_wr = wr;
wr               5298 drivers/infiniband/hw/mlx5/qp.c 				if (likely(wr->sg_list[i].length)) {
wr               5301 drivers/infiniband/hw/mlx5/qp.c 					 wr->sg_list + i);
wr               5309 drivers/infiniband/hw/mlx5/qp.c 		finish_wqe(qp, ctrl, seg, size, cur_edge, idx, wr->wr_id, nreq,
wr               5310 drivers/infiniband/hw/mlx5/qp.c 			   fence, mlx5_ib_opcode[wr->opcode]);
wr               5344 drivers/infiniband/hw/mlx5/qp.c int mlx5_ib_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr,
wr               5347 drivers/infiniband/hw/mlx5/qp.c 	return _mlx5_ib_post_send(ibqp, wr, bad_wr, false);
wr               5355 drivers/infiniband/hw/mlx5/qp.c static int _mlx5_ib_post_recv(struct ib_qp *ibqp, const struct ib_recv_wr *wr,
wr               5371 drivers/infiniband/hw/mlx5/qp.c 		*bad_wr = wr;
wr               5376 drivers/infiniband/hw/mlx5/qp.c 		return mlx5_ib_gsi_post_recv(ibqp, wr, bad_wr);
wr               5382 drivers/infiniband/hw/mlx5/qp.c 	for (nreq = 0; wr; nreq++, wr = wr->next) {
wr               5385 drivers/infiniband/hw/mlx5/qp.c 			*bad_wr = wr;
wr               5389 drivers/infiniband/hw/mlx5/qp.c 		if (unlikely(wr->num_sge > qp->rq.max_gs)) {
wr               5391 drivers/infiniband/hw/mlx5/qp.c 			*bad_wr = wr;
wr               5399 drivers/infiniband/hw/mlx5/qp.c 		for (i = 0; i < wr->num_sge; i++)
wr               5400 drivers/infiniband/hw/mlx5/qp.c 			set_data_ptr_seg(scat + i, wr->sg_list + i);
wr               5413 drivers/infiniband/hw/mlx5/qp.c 		qp->rq.wrid[ind] = wr->wr_id;
wr               5435 drivers/infiniband/hw/mlx5/qp.c int mlx5_ib_post_recv(struct ib_qp *ibqp, const struct ib_recv_wr *wr,
wr               5438 drivers/infiniband/hw/mlx5/qp.c 	return _mlx5_ib_post_recv(ibqp, wr, bad_wr, false);
wr               6446 drivers/infiniband/hw/mlx5/qp.c 		.wr = {
wr               6465 drivers/infiniband/hw/mlx5/qp.c 	ret = _mlx5_ib_post_send(qp, &swr.wr, &bad_swr, true);
wr                420 drivers/infiniband/hw/mlx5/srq.c int mlx5_ib_post_srq_recv(struct ib_srq *ibsrq, const struct ib_recv_wr *wr,
wr                437 drivers/infiniband/hw/mlx5/srq.c 		*bad_wr = wr;
wr                441 drivers/infiniband/hw/mlx5/srq.c 	for (nreq = 0; wr; nreq++, wr = wr->next) {
wr                442 drivers/infiniband/hw/mlx5/srq.c 		if (unlikely(wr->num_sge > srq->msrq.max_gs)) {
wr                444 drivers/infiniband/hw/mlx5/srq.c 			*bad_wr = wr;
wr                450 drivers/infiniband/hw/mlx5/srq.c 			*bad_wr = wr;
wr                454 drivers/infiniband/hw/mlx5/srq.c 		srq->wrid[srq->head] = wr->wr_id;
wr                460 drivers/infiniband/hw/mlx5/srq.c 		for (i = 0; i < wr->num_sge; i++) {
wr                461 drivers/infiniband/hw/mlx5/srq.c 			scat[i].byte_count = cpu_to_be32(wr->sg_list[i].length);
wr                462 drivers/infiniband/hw/mlx5/srq.c 			scat[i].lkey       = cpu_to_be32(wr->sg_list[i].lkey);
wr                463 drivers/infiniband/hw/mlx5/srq.c 			scat[i].addr       = cpu_to_be64(wr->sg_list[i].addr);
wr                523 drivers/infiniband/hw/mthca/mthca_dev.h int mthca_tavor_post_srq_recv(struct ib_srq *srq, const struct ib_recv_wr *wr,
wr                525 drivers/infiniband/hw/mthca/mthca_dev.h int mthca_arbel_post_srq_recv(struct ib_srq *srq, const struct ib_recv_wr *wr,
wr                534 drivers/infiniband/hw/mthca/mthca_dev.h int mthca_tavor_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr,
wr                536 drivers/infiniband/hw/mthca/mthca_dev.h int mthca_tavor_post_receive(struct ib_qp *ibqp, const struct ib_recv_wr *wr,
wr                538 drivers/infiniband/hw/mthca/mthca_dev.h int mthca_arbel_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr,
wr                540 drivers/infiniband/hw/mthca/mthca_dev.h int mthca_arbel_post_receive(struct ib_qp *ibqp, const struct ib_recv_wr *wr,
wr               1500 drivers/infiniband/hw/mthca/mthca_qp.c 			    int ind, const struct ib_ud_wr *wr,
wr               1509 drivers/infiniband/hw/mthca/mthca_qp.c 			  mthca_ah_grh_present(to_mah(wr->ah)), 0, 0, 0,
wr               1512 drivers/infiniband/hw/mthca/mthca_qp.c 	err = mthca_read_ah(dev, to_mah(wr->ah), &sqp->ud_header);
wr               1523 drivers/infiniband/hw/mthca/mthca_qp.c 	switch (wr->wr.opcode) {
wr               1531 drivers/infiniband/hw/mthca/mthca_qp.c 		sqp->ud_header.immediate_data = wr->wr.ex.imm_data;
wr               1540 drivers/infiniband/hw/mthca/mthca_qp.c 	sqp->ud_header.bth.solicited_event = !!(wr->wr.send_flags & IB_SEND_SOLICITED);
wr               1546 drivers/infiniband/hw/mthca/mthca_qp.c 				   wr->pkey_index, &pkey);
wr               1548 drivers/infiniband/hw/mthca/mthca_qp.c 	sqp->ud_header.bth.destination_qpn = cpu_to_be32(wr->remote_qpn);
wr               1550 drivers/infiniband/hw/mthca/mthca_qp.c 	sqp->ud_header.deth.qkey = cpu_to_be32(wr->remote_qkey & 0x80000000 ?
wr               1551 drivers/infiniband/hw/mthca/mthca_qp.c 					       sqp->qkey : wr->remote_qkey);
wr               1593 drivers/infiniband/hw/mthca/mthca_qp.c 					   const struct ib_atomic_wr *wr)
wr               1595 drivers/infiniband/hw/mthca/mthca_qp.c 	if (wr->wr.opcode == IB_WR_ATOMIC_CMP_AND_SWP) {
wr               1596 drivers/infiniband/hw/mthca/mthca_qp.c 		aseg->swap_add = cpu_to_be64(wr->swap);
wr               1597 drivers/infiniband/hw/mthca/mthca_qp.c 		aseg->compare  = cpu_to_be64(wr->compare_add);
wr               1599 drivers/infiniband/hw/mthca/mthca_qp.c 		aseg->swap_add = cpu_to_be64(wr->compare_add);
wr               1606 drivers/infiniband/hw/mthca/mthca_qp.c 			     const struct ib_ud_wr *wr)
wr               1608 drivers/infiniband/hw/mthca/mthca_qp.c 	useg->lkey    = cpu_to_be32(to_mah(wr->ah)->key);
wr               1609 drivers/infiniband/hw/mthca/mthca_qp.c 	useg->av_addr =	cpu_to_be64(to_mah(wr->ah)->avdma);
wr               1610 drivers/infiniband/hw/mthca/mthca_qp.c 	useg->dqpn    =	cpu_to_be32(wr->remote_qpn);
wr               1611 drivers/infiniband/hw/mthca/mthca_qp.c 	useg->qkey    =	cpu_to_be32(wr->remote_qkey);
wr               1616 drivers/infiniband/hw/mthca/mthca_qp.c 			     const struct ib_ud_wr *wr)
wr               1618 drivers/infiniband/hw/mthca/mthca_qp.c 	memcpy(useg->av, to_mah(wr->ah)->av, MTHCA_AV_SIZE);
wr               1619 drivers/infiniband/hw/mthca/mthca_qp.c 	useg->dqpn = cpu_to_be32(wr->remote_qpn);
wr               1620 drivers/infiniband/hw/mthca/mthca_qp.c 	useg->qkey = cpu_to_be32(wr->remote_qkey);
wr               1623 drivers/infiniband/hw/mthca/mthca_qp.c int mthca_tavor_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr,
wr               1653 drivers/infiniband/hw/mthca/mthca_qp.c 	for (nreq = 0; wr; ++nreq, wr = wr->next) {
wr               1660 drivers/infiniband/hw/mthca/mthca_qp.c 			*bad_wr = wr;
wr               1671 drivers/infiniband/hw/mthca/mthca_qp.c 			((wr->send_flags & IB_SEND_SIGNALED) ?
wr               1673 drivers/infiniband/hw/mthca/mthca_qp.c 			((wr->send_flags & IB_SEND_SOLICITED) ?
wr               1676 drivers/infiniband/hw/mthca/mthca_qp.c 		if (wr->opcode == IB_WR_SEND_WITH_IMM ||
wr               1677 drivers/infiniband/hw/mthca/mthca_qp.c 		    wr->opcode == IB_WR_RDMA_WRITE_WITH_IMM)
wr               1678 drivers/infiniband/hw/mthca/mthca_qp.c 			((struct mthca_next_seg *) wqe)->imm = wr->ex.imm_data;
wr               1685 drivers/infiniband/hw/mthca/mthca_qp.c 			switch (wr->opcode) {
wr               1688 drivers/infiniband/hw/mthca/mthca_qp.c 				set_raddr_seg(wqe, atomic_wr(wr)->remote_addr,
wr               1689 drivers/infiniband/hw/mthca/mthca_qp.c 					      atomic_wr(wr)->rkey);
wr               1692 drivers/infiniband/hw/mthca/mthca_qp.c 				set_atomic_seg(wqe, atomic_wr(wr));
wr               1701 drivers/infiniband/hw/mthca/mthca_qp.c 				set_raddr_seg(wqe, rdma_wr(wr)->remote_addr,
wr               1702 drivers/infiniband/hw/mthca/mthca_qp.c 					      rdma_wr(wr)->rkey);
wr               1715 drivers/infiniband/hw/mthca/mthca_qp.c 			switch (wr->opcode) {
wr               1718 drivers/infiniband/hw/mthca/mthca_qp.c 				set_raddr_seg(wqe, rdma_wr(wr)->remote_addr,
wr               1719 drivers/infiniband/hw/mthca/mthca_qp.c 					      rdma_wr(wr)->rkey);
wr               1732 drivers/infiniband/hw/mthca/mthca_qp.c 			set_tavor_ud_seg(wqe, ud_wr(wr));
wr               1738 drivers/infiniband/hw/mthca/mthca_qp.c 			err = build_mlx_header(dev, to_msqp(qp), ind, ud_wr(wr),
wr               1742 drivers/infiniband/hw/mthca/mthca_qp.c 				*bad_wr = wr;
wr               1750 drivers/infiniband/hw/mthca/mthca_qp.c 		if (wr->num_sge > qp->sq.max_gs) {
wr               1753 drivers/infiniband/hw/mthca/mthca_qp.c 			*bad_wr = wr;
wr               1757 drivers/infiniband/hw/mthca/mthca_qp.c 		for (i = 0; i < wr->num_sge; ++i) {
wr               1758 drivers/infiniband/hw/mthca/mthca_qp.c 			mthca_set_data_seg(wqe, wr->sg_list + i);
wr               1772 drivers/infiniband/hw/mthca/mthca_qp.c 		qp->wrid[ind + qp->rq.max] = wr->wr_id;
wr               1774 drivers/infiniband/hw/mthca/mthca_qp.c 		if (wr->opcode >= ARRAY_SIZE(mthca_opcode)) {
wr               1777 drivers/infiniband/hw/mthca/mthca_qp.c 			*bad_wr = wr;
wr               1784 drivers/infiniband/hw/mthca/mthca_qp.c 				    mthca_opcode[wr->opcode]);
wr               1788 drivers/infiniband/hw/mthca/mthca_qp.c 				    ((wr->send_flags & IB_SEND_FENCE) ?
wr               1793 drivers/infiniband/hw/mthca/mthca_qp.c 			op0   = mthca_opcode[wr->opcode];
wr               1794 drivers/infiniband/hw/mthca/mthca_qp.c 			f0    = wr->send_flags & IB_SEND_FENCE ?
wr               1821 drivers/infiniband/hw/mthca/mthca_qp.c int mthca_tavor_post_receive(struct ib_qp *ibqp, const struct ib_recv_wr *wr,
wr               1849 drivers/infiniband/hw/mthca/mthca_qp.c 	for (nreq = 0; wr; wr = wr->next) {
wr               1856 drivers/infiniband/hw/mthca/mthca_qp.c 			*bad_wr = wr;
wr               1871 drivers/infiniband/hw/mthca/mthca_qp.c 		if (unlikely(wr->num_sge > qp->rq.max_gs)) {
wr               1873 drivers/infiniband/hw/mthca/mthca_qp.c 			*bad_wr = wr;
wr               1877 drivers/infiniband/hw/mthca/mthca_qp.c 		for (i = 0; i < wr->num_sge; ++i) {
wr               1878 drivers/infiniband/hw/mthca/mthca_qp.c 			mthca_set_data_seg(wqe, wr->sg_list + i);
wr               1883 drivers/infiniband/hw/mthca/mthca_qp.c 		qp->wrid[ind] = wr->wr_id;
wr               1926 drivers/infiniband/hw/mthca/mthca_qp.c int mthca_arbel_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr,
wr               1957 drivers/infiniband/hw/mthca/mthca_qp.c 	for (nreq = 0; wr; ++nreq, wr = wr->next) {
wr               1990 drivers/infiniband/hw/mthca/mthca_qp.c 			*bad_wr = wr;
wr               1999 drivers/infiniband/hw/mthca/mthca_qp.c 			((wr->send_flags & IB_SEND_SIGNALED) ?
wr               2001 drivers/infiniband/hw/mthca/mthca_qp.c 			((wr->send_flags & IB_SEND_SOLICITED) ?
wr               2003 drivers/infiniband/hw/mthca/mthca_qp.c 			((wr->send_flags & IB_SEND_IP_CSUM) ?
wr               2006 drivers/infiniband/hw/mthca/mthca_qp.c 		if (wr->opcode == IB_WR_SEND_WITH_IMM ||
wr               2007 drivers/infiniband/hw/mthca/mthca_qp.c 		    wr->opcode == IB_WR_RDMA_WRITE_WITH_IMM)
wr               2008 drivers/infiniband/hw/mthca/mthca_qp.c 			((struct mthca_next_seg *) wqe)->imm = wr->ex.imm_data;
wr               2015 drivers/infiniband/hw/mthca/mthca_qp.c 			switch (wr->opcode) {
wr               2018 drivers/infiniband/hw/mthca/mthca_qp.c 				set_raddr_seg(wqe, atomic_wr(wr)->remote_addr,
wr               2019 drivers/infiniband/hw/mthca/mthca_qp.c 					      atomic_wr(wr)->rkey);
wr               2022 drivers/infiniband/hw/mthca/mthca_qp.c 				set_atomic_seg(wqe, atomic_wr(wr));
wr               2031 drivers/infiniband/hw/mthca/mthca_qp.c 				set_raddr_seg(wqe, rdma_wr(wr)->remote_addr,
wr               2032 drivers/infiniband/hw/mthca/mthca_qp.c 					      rdma_wr(wr)->rkey);
wr               2045 drivers/infiniband/hw/mthca/mthca_qp.c 			switch (wr->opcode) {
wr               2048 drivers/infiniband/hw/mthca/mthca_qp.c 				set_raddr_seg(wqe, rdma_wr(wr)->remote_addr,
wr               2049 drivers/infiniband/hw/mthca/mthca_qp.c 					      rdma_wr(wr)->rkey);
wr               2062 drivers/infiniband/hw/mthca/mthca_qp.c 			set_arbel_ud_seg(wqe, ud_wr(wr));
wr               2068 drivers/infiniband/hw/mthca/mthca_qp.c 			err = build_mlx_header(dev, to_msqp(qp), ind, ud_wr(wr),
wr               2072 drivers/infiniband/hw/mthca/mthca_qp.c 				*bad_wr = wr;
wr               2080 drivers/infiniband/hw/mthca/mthca_qp.c 		if (wr->num_sge > qp->sq.max_gs) {
wr               2083 drivers/infiniband/hw/mthca/mthca_qp.c 			*bad_wr = wr;
wr               2087 drivers/infiniband/hw/mthca/mthca_qp.c 		for (i = 0; i < wr->num_sge; ++i) {
wr               2088 drivers/infiniband/hw/mthca/mthca_qp.c 			mthca_set_data_seg(wqe, wr->sg_list + i);
wr               2102 drivers/infiniband/hw/mthca/mthca_qp.c 		qp->wrid[ind + qp->rq.max] = wr->wr_id;
wr               2104 drivers/infiniband/hw/mthca/mthca_qp.c 		if (wr->opcode >= ARRAY_SIZE(mthca_opcode)) {
wr               2107 drivers/infiniband/hw/mthca/mthca_qp.c 			*bad_wr = wr;
wr               2114 drivers/infiniband/hw/mthca/mthca_qp.c 				    mthca_opcode[wr->opcode]);
wr               2118 drivers/infiniband/hw/mthca/mthca_qp.c 				    ((wr->send_flags & IB_SEND_FENCE) ?
wr               2123 drivers/infiniband/hw/mthca/mthca_qp.c 			op0   = mthca_opcode[wr->opcode];
wr               2124 drivers/infiniband/hw/mthca/mthca_qp.c 			f0    = wr->send_flags & IB_SEND_FENCE ?
wr               2160 drivers/infiniband/hw/mthca/mthca_qp.c int mthca_arbel_post_receive(struct ib_qp *ibqp, const struct ib_recv_wr *wr,
wr               2178 drivers/infiniband/hw/mthca/mthca_qp.c 	for (nreq = 0; wr; ++nreq, wr = wr->next) {
wr               2185 drivers/infiniband/hw/mthca/mthca_qp.c 			*bad_wr = wr;
wr               2195 drivers/infiniband/hw/mthca/mthca_qp.c 		if (unlikely(wr->num_sge > qp->rq.max_gs)) {
wr               2197 drivers/infiniband/hw/mthca/mthca_qp.c 			*bad_wr = wr;
wr               2201 drivers/infiniband/hw/mthca/mthca_qp.c 		for (i = 0; i < wr->num_sge; ++i) {
wr               2202 drivers/infiniband/hw/mthca/mthca_qp.c 			mthca_set_data_seg(wqe, wr->sg_list + i);
wr               2209 drivers/infiniband/hw/mthca/mthca_qp.c 		qp->wrid[ind] = wr->wr_id;
wr                483 drivers/infiniband/hw/mthca/mthca_srq.c int mthca_tavor_post_srq_recv(struct ib_srq *ibsrq, const struct ib_recv_wr *wr,
wr                502 drivers/infiniband/hw/mthca/mthca_srq.c 	for (nreq = 0; wr; wr = wr->next) {
wr                510 drivers/infiniband/hw/mthca/mthca_srq.c 			*bad_wr = wr;
wr                522 drivers/infiniband/hw/mthca/mthca_srq.c 		if (unlikely(wr->num_sge > srq->max_gs)) {
wr                524 drivers/infiniband/hw/mthca/mthca_srq.c 			*bad_wr = wr;
wr                529 drivers/infiniband/hw/mthca/mthca_srq.c 		for (i = 0; i < wr->num_sge; ++i) {
wr                530 drivers/infiniband/hw/mthca/mthca_srq.c 			mthca_set_data_seg(wqe, wr->sg_list + i);
wr                540 drivers/infiniband/hw/mthca/mthca_srq.c 		srq->wrid[ind]  = wr->wr_id;
wr                577 drivers/infiniband/hw/mthca/mthca_srq.c int mthca_arbel_post_srq_recv(struct ib_srq *ibsrq, const struct ib_recv_wr *wr,
wr                592 drivers/infiniband/hw/mthca/mthca_srq.c 	for (nreq = 0; wr; ++nreq, wr = wr->next) {
wr                600 drivers/infiniband/hw/mthca/mthca_srq.c 			*bad_wr = wr;
wr                609 drivers/infiniband/hw/mthca/mthca_srq.c 		if (unlikely(wr->num_sge > srq->max_gs)) {
wr                611 drivers/infiniband/hw/mthca/mthca_srq.c 			*bad_wr = wr;
wr                615 drivers/infiniband/hw/mthca/mthca_srq.c 		for (i = 0; i < wr->num_sge; ++i) {
wr                616 drivers/infiniband/hw/mthca/mthca_srq.c 			mthca_set_data_seg(wqe, wr->sg_list + i);
wr                623 drivers/infiniband/hw/mthca/mthca_srq.c 		srq->wrid[ind]  = wr->wr_id;
wr               1888 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 				const struct ib_send_wr *wr)
wr               1892 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	struct ocrdma_ah *ah = get_ocrdma_ah(ud_wr(wr)->ah);
wr               1894 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	ud_hdr->rsvd_dest_qpn = ud_wr(wr)->remote_qpn;
wr               1898 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		ud_hdr->qkey = ud_wr(wr)->remote_qkey;
wr               1935 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 				    const struct ib_send_wr *wr, u32 wqe_size)
wr               1940 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	if (wr->send_flags & IB_SEND_INLINE && qp->qp_type != IB_QPT_UD) {
wr               1941 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		hdr->total_len = ocrdma_sglist_len(wr->sg_list, wr->num_sge);
wr               1949 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		for (i = 0; i < wr->num_sge; i++) {
wr               1951 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 			       (void *)(unsigned long)wr->sg_list[i].addr,
wr               1952 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 			       wr->sg_list[i].length);
wr               1953 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 			dpp_addr += wr->sg_list[i].length;
wr               1961 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		ocrdma_build_sges(hdr, sge, wr->num_sge, wr->sg_list);
wr               1962 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		if (wr->num_sge)
wr               1963 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 			wqe_size += (wr->num_sge * sizeof(struct ocrdma_sge));
wr               1973 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 			     const struct ib_send_wr *wr)
wr               1980 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		ocrdma_build_ud_hdr(qp, hdr, wr);
wr               1987 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	status = ocrdma_build_inline_sges(qp, hdr, sge, wr, wqe_size);
wr               1992 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 			      const struct ib_send_wr *wr)
wr               1999 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	status = ocrdma_build_inline_sges(qp, hdr, sge, wr, wqe_size);
wr               2002 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	ext_rw->addr_lo = rdma_wr(wr)->remote_addr;
wr               2003 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	ext_rw->addr_hi = upper_32_bits(rdma_wr(wr)->remote_addr);
wr               2004 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	ext_rw->lrkey = rdma_wr(wr)->rkey;
wr               2010 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 			      const struct ib_send_wr *wr)
wr               2014 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	u32 wqe_size = ((wr->num_sge + 1) * sizeof(struct ocrdma_sge)) +
wr               2017 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	ocrdma_build_sges(hdr, sge, wr->num_sge, wr->sg_list);
wr               2022 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	ext_rw->addr_lo = rdma_wr(wr)->remote_addr;
wr               2023 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	ext_rw->addr_hi = upper_32_bits(rdma_wr(wr)->remote_addr);
wr               2024 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	ext_rw->lrkey = rdma_wr(wr)->rkey;
wr               2040 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 			    const struct ib_reg_wr *wr)
wr               2044 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	struct ocrdma_mr *mr = get_ocrdma_mr(wr->mr);
wr               2055 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	if (wr->access & IB_ACCESS_LOCAL_WRITE)
wr               2057 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	if (wr->access & IB_ACCESS_REMOTE_WRITE)
wr               2059 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	if (wr->access & IB_ACCESS_REMOTE_READ)
wr               2061 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	hdr->lkey = wr->key;
wr               2101 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c int ocrdma_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr,
wr               2112 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		*bad_wr = wr;
wr               2116 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	while (wr) {
wr               2118 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		    (wr->opcode != IB_WR_SEND &&
wr               2119 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		     wr->opcode != IB_WR_SEND_WITH_IMM)) {
wr               2120 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 			*bad_wr = wr;
wr               2125 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		    wr->num_sge > qp->sq.max_sges) {
wr               2126 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 			*bad_wr = wr;
wr               2132 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		if (wr->send_flags & IB_SEND_SIGNALED || qp->signaled)
wr               2134 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		if (wr->send_flags & IB_SEND_FENCE)
wr               2137 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		if (wr->send_flags & IB_SEND_SOLICITED)
wr               2141 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		switch (wr->opcode) {
wr               2144 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 			hdr->immdt = ntohl(wr->ex.imm_data);
wr               2148 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 			ocrdma_build_send(qp, hdr, wr);
wr               2153 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 			hdr->lkey = wr->ex.invalidate_rkey;
wr               2154 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 			status = ocrdma_build_send(qp, hdr, wr);
wr               2158 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 			hdr->immdt = ntohl(wr->ex.imm_data);
wr               2162 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 			status = ocrdma_build_write(qp, hdr, wr);
wr               2165 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 			ocrdma_build_read(qp, hdr, wr);
wr               2173 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 			hdr->lkey = wr->ex.invalidate_rkey;
wr               2176 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 			status = ocrdma_build_reg(qp, hdr, reg_wr(wr));
wr               2183 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 			*bad_wr = wr;
wr               2186 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		if (wr->send_flags & IB_SEND_SIGNALED || qp->signaled)
wr               2190 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		qp->wqe_wr_id_tbl[qp->sq.head].wrid = wr->wr_id;
wr               2200 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		wr = wr->next;
wr               2214 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 			     const struct ib_recv_wr *wr, u16 tag)
wr               2218 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	if (wr->num_sge)
wr               2219 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		wqe_size = (wr->num_sge * sizeof(*sge)) + sizeof(*rqe);
wr               2230 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	ocrdma_build_sges(rqe, sge, wr->num_sge, wr->sg_list);
wr               2234 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c int ocrdma_post_recv(struct ib_qp *ibqp, const struct ib_recv_wr *wr,
wr               2245 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		*bad_wr = wr;
wr               2248 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	while (wr) {
wr               2250 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		    wr->num_sge > qp->rq.max_sges) {
wr               2251 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 			*bad_wr = wr;
wr               2256 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		ocrdma_build_rqe(rqe, wr, 0);
wr               2258 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		qp->rqe_wr_id_tbl[qp->rq.head] = wr->wr_id;
wr               2267 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		wr = wr->next;
wr               2304 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c int ocrdma_post_srq_recv(struct ib_srq *ibsrq, const struct ib_recv_wr *wr,
wr               2316 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	while (wr) {
wr               2318 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		    wr->num_sge > srq->rq.max_sges) {
wr               2320 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 			*bad_wr = wr;
wr               2325 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		ocrdma_build_rqe(rqe, wr, tag);
wr               2327 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		srq->rqe_wr_id_tbl[tag] = wr->wr_id;
wr               2334 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		wr = wr->next;
wr                543 drivers/infiniband/hw/qedr/qedr_roce_cm.c int qedr_gsi_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr,
wr                553 drivers/infiniband/hw/qedr/qedr_roce_cm.c 		*bad_wr = wr;
wr                560 drivers/infiniband/hw/qedr/qedr_roce_cm.c 	if (wr->num_sge > RDMA_MAX_SGE_PER_SQ_WQE) {
wr                562 drivers/infiniband/hw/qedr/qedr_roce_cm.c 		       wr->num_sge, RDMA_MAX_SGE_PER_SQ_WQE);
wr                567 drivers/infiniband/hw/qedr/qedr_roce_cm.c 	if (wr->opcode != IB_WR_SEND) {
wr                570 drivers/infiniband/hw/qedr/qedr_roce_cm.c 		       wr->opcode);
wr                577 drivers/infiniband/hw/qedr/qedr_roce_cm.c 	rc = qedr_gsi_build_packet(dev, qp, wr, &pkt);
wr                586 drivers/infiniband/hw/qedr/qedr_roce_cm.c 		qp->wqe_wr_id[qp->sq.prod].wr_id = wr->wr_id;
wr                590 drivers/infiniband/hw/qedr/qedr_roce_cm.c 			 wr->opcode, in_irq(), irqs_disabled(), wr->wr_id);
wr                594 drivers/infiniband/hw/qedr/qedr_roce_cm.c 		*bad_wr = wr;
wr                599 drivers/infiniband/hw/qedr/qedr_roce_cm.c 	if (wr->next) {
wr                602 drivers/infiniband/hw/qedr/qedr_roce_cm.c 		*bad_wr = wr->next;
wr                609 drivers/infiniband/hw/qedr/qedr_roce_cm.c 	*bad_wr = wr;
wr                613 drivers/infiniband/hw/qedr/qedr_roce_cm.c int qedr_gsi_post_recv(struct ib_qp *ibqp, const struct ib_recv_wr *wr,
wr                623 drivers/infiniband/hw/qedr/qedr_roce_cm.c 		*bad_wr = wr;
wr                632 drivers/infiniband/hw/qedr/qedr_roce_cm.c 	while (wr) {
wr                633 drivers/infiniband/hw/qedr/qedr_roce_cm.c 		if (wr->num_sge > QEDR_GSI_MAX_RECV_SGE) {
wr                636 drivers/infiniband/hw/qedr/qedr_roce_cm.c 			       wr->num_sge, QEDR_GSI_MAX_RECV_SGE);
wr                642 drivers/infiniband/hw/qedr/qedr_roce_cm.c 						  wr->sg_list[0].addr,
wr                643 drivers/infiniband/hw/qedr/qedr_roce_cm.c 						  wr->sg_list[0].length,
wr                655 drivers/infiniband/hw/qedr/qedr_roce_cm.c 		qp->rqe_wr_id[qp->rq.prod].sg_list[0] = wr->sg_list[0];
wr                656 drivers/infiniband/hw/qedr/qedr_roce_cm.c 		qp->rqe_wr_id[qp->rq.prod].wr_id = wr->wr_id;
wr                660 drivers/infiniband/hw/qedr/qedr_roce_cm.c 		wr = wr->next;
wr                668 drivers/infiniband/hw/qedr/qedr_roce_cm.c 	*bad_wr = wr;
wr                 49 drivers/infiniband/hw/qedr/qedr_roce_cm.h int qedr_gsi_post_recv(struct ib_qp *ibqp, const struct ib_recv_wr *wr,
wr                 51 drivers/infiniband/hw/qedr/qedr_roce_cm.h int qedr_gsi_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr,
wr               2928 drivers/infiniband/hw/qedr/verbs.c 				       const struct ib_send_wr *wr,
wr               2932 drivers/infiniband/hw/qedr/verbs.c 	u32 data_size = sge_data_len(wr->sg_list, wr->num_sge);
wr               2938 drivers/infiniband/hw/qedr/verbs.c 		*bad_wr = wr;
wr               2952 drivers/infiniband/hw/qedr/verbs.c 	for (i = 0; i < wr->num_sge; i++) {
wr               2953 drivers/infiniband/hw/qedr/verbs.c 		u32 len = wr->sg_list[i].length;
wr               2954 drivers/infiniband/hw/qedr/verbs.c 		void *src = (void *)(uintptr_t)wr->sg_list[i].addr;
wr               3013 drivers/infiniband/hw/qedr/verbs.c 				const struct ib_send_wr *wr)
wr               3018 drivers/infiniband/hw/qedr/verbs.c 	for (i = 0; i < wr->num_sge; i++) {
wr               3021 drivers/infiniband/hw/qedr/verbs.c 		DMA_REGPAIR_LE(sge->addr, wr->sg_list[i].addr);
wr               3022 drivers/infiniband/hw/qedr/verbs.c 		sge->l_key = cpu_to_le32(wr->sg_list[i].lkey);
wr               3023 drivers/infiniband/hw/qedr/verbs.c 		sge->length = cpu_to_le32(wr->sg_list[i].length);
wr               3024 drivers/infiniband/hw/qedr/verbs.c 		data_size += wr->sg_list[i].length;
wr               3028 drivers/infiniband/hw/qedr/verbs.c 		*wqe_size += wr->num_sge;
wr               3037 drivers/infiniband/hw/qedr/verbs.c 				     const struct ib_send_wr *wr,
wr               3040 drivers/infiniband/hw/qedr/verbs.c 	rwqe2->r_key = cpu_to_le32(rdma_wr(wr)->rkey);
wr               3041 drivers/infiniband/hw/qedr/verbs.c 	DMA_REGPAIR_LE(rwqe2->remote_va, rdma_wr(wr)->remote_addr);
wr               3043 drivers/infiniband/hw/qedr/verbs.c 	if (wr->send_flags & IB_SEND_INLINE &&
wr               3044 drivers/infiniband/hw/qedr/verbs.c 	    (wr->opcode == IB_WR_RDMA_WRITE_WITH_IMM ||
wr               3045 drivers/infiniband/hw/qedr/verbs.c 	     wr->opcode == IB_WR_RDMA_WRITE)) {
wr               3049 drivers/infiniband/hw/qedr/verbs.c 		return qedr_prepare_sq_inline_data(dev, qp, &rwqe->wqe_size, wr,
wr               3053 drivers/infiniband/hw/qedr/verbs.c 	return qedr_prepare_sq_sges(qp, &rwqe->wqe_size, wr);
wr               3060 drivers/infiniband/hw/qedr/verbs.c 				     const struct ib_send_wr *wr,
wr               3064 drivers/infiniband/hw/qedr/verbs.c 	if (wr->send_flags & IB_SEND_INLINE) {
wr               3068 drivers/infiniband/hw/qedr/verbs.c 		return qedr_prepare_sq_inline_data(dev, qp, &swqe->wqe_size, wr,
wr               3072 drivers/infiniband/hw/qedr/verbs.c 	return qedr_prepare_sq_sges(qp, &swqe->wqe_size, wr);
wr               3077 drivers/infiniband/hw/qedr/verbs.c 			    const struct ib_reg_wr *wr)
wr               3079 drivers/infiniband/hw/qedr/verbs.c 	struct qedr_mr *mr = get_qedr_mr(wr->mr);
wr               3085 drivers/infiniband/hw/qedr/verbs.c 	fwqe1->l_key = wr->key;
wr               3090 drivers/infiniband/hw/qedr/verbs.c 		   !!(wr->access & IB_ACCESS_REMOTE_READ));
wr               3092 drivers/infiniband/hw/qedr/verbs.c 		   !!(wr->access & IB_ACCESS_REMOTE_WRITE));
wr               3094 drivers/infiniband/hw/qedr/verbs.c 		   !!(wr->access & IB_ACCESS_REMOTE_ATOMIC));
wr               3097 drivers/infiniband/hw/qedr/verbs.c 		   !!(wr->access & IB_ACCESS_LOCAL_WRITE));
wr               3140 drivers/infiniband/hw/qedr/verbs.c 				      const struct ib_send_wr *wr)
wr               3146 drivers/infiniband/hw/qedr/verbs.c 	err_wr = wr->num_sge > qp->sq.max_sges;
wr               3177 drivers/infiniband/hw/qedr/verbs.c static int __qedr_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr,
wr               3196 drivers/infiniband/hw/qedr/verbs.c 	if (!qedr_can_post_send(qp, wr)) {
wr               3197 drivers/infiniband/hw/qedr/verbs.c 		*bad_wr = wr;
wr               3203 drivers/infiniband/hw/qedr/verbs.c 		!!(wr->send_flags & IB_SEND_SIGNALED) || qp->signaled;
wr               3207 drivers/infiniband/hw/qedr/verbs.c 		   !!(wr->send_flags & IB_SEND_SOLICITED));
wr               3208 drivers/infiniband/hw/qedr/verbs.c 	comp = (!!(wr->send_flags & IB_SEND_SIGNALED)) || qp->signaled;
wr               3211 drivers/infiniband/hw/qedr/verbs.c 		   !!(wr->send_flags & IB_SEND_FENCE));
wr               3214 drivers/infiniband/hw/qedr/verbs.c 	qp->wqe_wr_id[qp->sq.prod].opcode = qedr_ib_to_wc_opcode(wr->opcode);
wr               3216 drivers/infiniband/hw/qedr/verbs.c 	switch (wr->opcode) {
wr               3220 drivers/infiniband/hw/qedr/verbs.c 			*bad_wr = wr;
wr               3228 drivers/infiniband/hw/qedr/verbs.c 		swqe->inv_key_or_imm_data = cpu_to_le32(be32_to_cpu(wr->ex.imm_data));
wr               3230 drivers/infiniband/hw/qedr/verbs.c 						   wr, bad_wr);
wr               3243 drivers/infiniband/hw/qedr/verbs.c 						   wr, bad_wr);
wr               3254 drivers/infiniband/hw/qedr/verbs.c 		swqe->inv_key_or_imm_data = cpu_to_le32(wr->ex.invalidate_rkey);
wr               3256 drivers/infiniband/hw/qedr/verbs.c 						   wr, bad_wr);
wr               3266 drivers/infiniband/hw/qedr/verbs.c 			*bad_wr = wr;
wr               3273 drivers/infiniband/hw/qedr/verbs.c 		rwqe->imm_data = htonl(cpu_to_le32(wr->ex.imm_data));
wr               3276 drivers/infiniband/hw/qedr/verbs.c 						   wr, bad_wr);
wr               3289 drivers/infiniband/hw/qedr/verbs.c 						   wr, bad_wr);
wr               3306 drivers/infiniband/hw/qedr/verbs.c 						   wr, bad_wr);
wr               3319 drivers/infiniband/hw/qedr/verbs.c 		DMA_REGPAIR_LE(awqe2->remote_va, atomic_wr(wr)->remote_addr);
wr               3320 drivers/infiniband/hw/qedr/verbs.c 		awqe2->r_key = cpu_to_le32(atomic_wr(wr)->rkey);
wr               3324 drivers/infiniband/hw/qedr/verbs.c 		if (wr->opcode == IB_WR_ATOMIC_FETCH_AND_ADD) {
wr               3327 drivers/infiniband/hw/qedr/verbs.c 				       atomic_wr(wr)->compare_add);
wr               3331 drivers/infiniband/hw/qedr/verbs.c 				       atomic_wr(wr)->swap);
wr               3333 drivers/infiniband/hw/qedr/verbs.c 				       atomic_wr(wr)->compare_add);
wr               3336 drivers/infiniband/hw/qedr/verbs.c 		qedr_prepare_sq_sges(qp, NULL, wr);
wr               3347 drivers/infiniband/hw/qedr/verbs.c 		iwqe->inv_l_key = wr->ex.invalidate_rkey;
wr               3357 drivers/infiniband/hw/qedr/verbs.c 		rc = qedr_prepare_reg(qp, fwqe1, reg_wr(wr));
wr               3360 drivers/infiniband/hw/qedr/verbs.c 			*bad_wr = wr;
wr               3368 drivers/infiniband/hw/qedr/verbs.c 		DP_ERR(dev, "invalid opcode 0x%x!\n", wr->opcode);
wr               3370 drivers/infiniband/hw/qedr/verbs.c 		*bad_wr = wr;
wr               3392 drivers/infiniband/hw/qedr/verbs.c int qedr_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr,
wr               3403 drivers/infiniband/hw/qedr/verbs.c 		return qedr_gsi_post_send(ibqp, wr, bad_wr);
wr               3412 drivers/infiniband/hw/qedr/verbs.c 			*bad_wr = wr;
wr               3420 drivers/infiniband/hw/qedr/verbs.c 	while (wr) {
wr               3421 drivers/infiniband/hw/qedr/verbs.c 		rc = __qedr_post_send(ibqp, wr, bad_wr);
wr               3425 drivers/infiniband/hw/qedr/verbs.c 		qp->wqe_wr_id[qp->sq.prod].wr_id = wr->wr_id;
wr               3431 drivers/infiniband/hw/qedr/verbs.c 		wr = wr->next;
wr               3468 drivers/infiniband/hw/qedr/verbs.c int qedr_post_srq_recv(struct ib_srq *ibsrq, const struct ib_recv_wr *wr,
wr               3484 drivers/infiniband/hw/qedr/verbs.c 	while (wr) {
wr               3489 drivers/infiniband/hw/qedr/verbs.c 		    wr->num_sge > srq->hw_srq.max_sges) {
wr               3492 drivers/infiniband/hw/qedr/verbs.c 			       wr->num_sge, srq->hw_srq.max_sges);
wr               3494 drivers/infiniband/hw/qedr/verbs.c 			*bad_wr = wr;
wr               3499 drivers/infiniband/hw/qedr/verbs.c 		num_sge = wr->num_sge;
wr               3501 drivers/infiniband/hw/qedr/verbs.c 		SRQ_HDR_SET(hdr, wr->wr_id, num_sge);
wr               3509 drivers/infiniband/hw/qedr/verbs.c 			 wr->num_sge, hw_srq->wqe_prod, wr->wr_id);
wr               3511 drivers/infiniband/hw/qedr/verbs.c 		for (i = 0; i < wr->num_sge; i++) {
wr               3515 drivers/infiniband/hw/qedr/verbs.c 			SRQ_SGE_SET(srq_sge, wr->sg_list[i].addr,
wr               3516 drivers/infiniband/hw/qedr/verbs.c 				    wr->sg_list[i].length, wr->sg_list[i].lkey);
wr               3541 drivers/infiniband/hw/qedr/verbs.c 		wr = wr->next;
wr               3551 drivers/infiniband/hw/qedr/verbs.c int qedr_post_recv(struct ib_qp *ibqp, const struct ib_recv_wr *wr,
wr               3560 drivers/infiniband/hw/qedr/verbs.c 		return qedr_gsi_post_recv(ibqp, wr, bad_wr);
wr               3566 drivers/infiniband/hw/qedr/verbs.c 		*bad_wr = wr;
wr               3570 drivers/infiniband/hw/qedr/verbs.c 	while (wr) {
wr               3575 drivers/infiniband/hw/qedr/verbs.c 		    wr->num_sge > qp->rq.max_sges) {
wr               3578 drivers/infiniband/hw/qedr/verbs.c 			       QEDR_MAX_RQE_ELEMENTS_PER_RQE, wr->num_sge,
wr               3581 drivers/infiniband/hw/qedr/verbs.c 			*bad_wr = wr;
wr               3584 drivers/infiniband/hw/qedr/verbs.c 		for (i = 0; i < wr->num_sge; i++) {
wr               3594 drivers/infiniband/hw/qedr/verbs.c 					  wr->num_sge);
wr               3597 drivers/infiniband/hw/qedr/verbs.c 				  wr->sg_list[i].lkey);
wr               3599 drivers/infiniband/hw/qedr/verbs.c 			RQ_SGE_SET(rqe, wr->sg_list[i].addr,
wr               3600 drivers/infiniband/hw/qedr/verbs.c 				   wr->sg_list[i].length, flags);
wr               3607 drivers/infiniband/hw/qedr/verbs.c 		if (!wr->num_sge) {
wr               3622 drivers/infiniband/hw/qedr/verbs.c 		qp->rqe_wr_id[qp->rq.prod].wr_id = wr->wr_id;
wr               3644 drivers/infiniband/hw/qedr/verbs.c 		wr = wr->next;
wr                 72 drivers/infiniband/hw/qedr/verbs.h int qedr_post_srq_recv(struct ib_srq *ibsrq, const struct ib_recv_wr *wr,
wr                437 drivers/infiniband/hw/qib/qib_qp.c 		   wqe->wr.opcode,
wr                296 drivers/infiniband/hw/qib/qib_rc.c 			if ((wqe->wr.send_flags & IB_SEND_FENCE) &&
wr                312 drivers/infiniband/hw/qib/qib_rc.c 		switch (wqe->wr.opcode) {
wr                323 drivers/infiniband/hw/qib/qib_rc.c 			if (wqe->wr.opcode == IB_WR_SEND)
wr                328 drivers/infiniband/hw/qib/qib_rc.c 				ohdr->u.imm_data = wqe->wr.ex.imm_data;
wr                331 drivers/infiniband/hw/qib/qib_rc.c 			if (wqe->wr.send_flags & IB_SEND_SOLICITED)
wr                358 drivers/infiniband/hw/qib/qib_rc.c 			if (wqe->rdma_wr.wr.opcode == IB_WR_RDMA_WRITE)
wr                364 drivers/infiniband/hw/qib/qib_rc.c 					wqe->rdma_wr.wr.ex.imm_data;
wr                366 drivers/infiniband/hw/qib/qib_rc.c 				if (wqe->rdma_wr.wr.send_flags & IB_SEND_SOLICITED)
wr                420 drivers/infiniband/hw/qib/qib_rc.c 			if (wqe->atomic_wr.wr.opcode == IB_WR_ATOMIC_CMP_AND_SWP) {
wr                449 drivers/infiniband/hw/qib/qib_rc.c 		qp->s_sge.num_sge = wqe->wr.num_sge;
wr                457 drivers/infiniband/hw/qib/qib_rc.c 		if (wqe->wr.opcode == IB_WR_RDMA_READ)
wr                486 drivers/infiniband/hw/qib/qib_rc.c 		if (wqe->wr.opcode == IB_WR_SEND)
wr                491 drivers/infiniband/hw/qib/qib_rc.c 			ohdr->u.imm_data = wqe->wr.ex.imm_data;
wr                494 drivers/infiniband/hw/qib/qib_rc.c 		if (wqe->wr.send_flags & IB_SEND_SOLICITED)
wr                525 drivers/infiniband/hw/qib/qib_rc.c 		if (wqe->wr.opcode == IB_WR_RDMA_WRITE)
wr                530 drivers/infiniband/hw/qib/qib_rc.c 			ohdr->u.imm_data = wqe->wr.ex.imm_data;
wr                532 drivers/infiniband/hw/qib/qib_rc.c 			if (wqe->wr.send_flags & IB_SEND_SOLICITED)
wr                753 drivers/infiniband/hw/qib/qib_rc.c 	opcode = wqe->wr.opcode;
wr                774 drivers/infiniband/hw/qib/qib_rc.c 		opcode = wqe->wr.opcode;
wr                839 drivers/infiniband/hw/qib/qib_rc.c 	if (wqe->wr.opcode == IB_WR_RDMA_READ)
wr                865 drivers/infiniband/hw/qib/qib_rc.c 			if (wqe->wr.opcode == IB_WR_RDMA_READ)
wr                924 drivers/infiniband/hw/qib/qib_rc.c 				     ib_qib_wc_opcode[wqe->wr.opcode],
wr                963 drivers/infiniband/hw/qib/qib_rc.c 				     ib_qib_wc_opcode[wqe->wr.opcode],
wr               1039 drivers/infiniband/hw/qib/qib_rc.c 		if (wqe->wr.opcode == IB_WR_RDMA_READ &&
wr               1054 drivers/infiniband/hw/qib/qib_rc.c 		if ((wqe->wr.opcode == IB_WR_RDMA_READ &&
wr               1056 drivers/infiniband/hw/qib/qib_rc.c 		    ((wqe->wr.opcode == IB_WR_ATOMIC_CMP_AND_SWP ||
wr               1057 drivers/infiniband/hw/qib/qib_rc.c 		      wqe->wr.opcode == IB_WR_ATOMIC_FETCH_AND_ADD) &&
wr               1076 drivers/infiniband/hw/qib/qib_rc.c 		if (wqe->wr.opcode == IB_WR_ATOMIC_CMP_AND_SWP ||
wr               1077 drivers/infiniband/hw/qib/qib_rc.c 		    wqe->wr.opcode == IB_WR_ATOMIC_FETCH_AND_ADD) {
wr               1082 drivers/infiniband/hw/qib/qib_rc.c 		    (wqe->wr.opcode == IB_WR_RDMA_READ ||
wr               1083 drivers/infiniband/hw/qib/qib_rc.c 		     wqe->wr.opcode == IB_WR_ATOMIC_CMP_AND_SWP ||
wr               1084 drivers/infiniband/hw/qib/qib_rc.c 		     wqe->wr.opcode == IB_WR_ATOMIC_FETCH_AND_ADD)) {
wr               1234 drivers/infiniband/hw/qib/qib_rc.c 		if (wqe->wr.opcode == IB_WR_RDMA_READ ||
wr               1235 drivers/infiniband/hw/qib/qib_rc.c 		    wqe->wr.opcode == IB_WR_ATOMIC_CMP_AND_SWP ||
wr               1236 drivers/infiniband/hw/qib/qib_rc.c 		    wqe->wr.opcode == IB_WR_ATOMIC_FETCH_AND_ADD)
wr               1356 drivers/infiniband/hw/qib/qib_rc.c 		if (unlikely(wqe->wr.opcode != IB_WR_RDMA_READ))
wr               1371 drivers/infiniband/hw/qib/qib_rc.c 		if (unlikely(wqe->wr.opcode != IB_WR_RDMA_READ))
wr               1430 drivers/infiniband/hw/qib/qib_rc.c 		if (unlikely(wqe->wr.opcode != IB_WR_RDMA_READ))
wr                100 drivers/infiniband/hw/qib/qib_uc.c 		qp->s_sge.num_sge = wqe->wr.num_sge;
wr                104 drivers/infiniband/hw/qib/qib_uc.c 		switch (wqe->wr.opcode) {
wr                112 drivers/infiniband/hw/qib/qib_uc.c 			if (wqe->wr.opcode == IB_WR_SEND)
wr                118 drivers/infiniband/hw/qib/qib_uc.c 				ohdr->u.imm_data = wqe->wr.ex.imm_data;
wr                121 drivers/infiniband/hw/qib/qib_uc.c 			if (wqe->wr.send_flags & IB_SEND_SOLICITED)
wr                141 drivers/infiniband/hw/qib/qib_uc.c 			if (wqe->wr.opcode == IB_WR_RDMA_WRITE)
wr                147 drivers/infiniband/hw/qib/qib_uc.c 				ohdr->u.rc.imm_data = wqe->wr.ex.imm_data;
wr                149 drivers/infiniband/hw/qib/qib_uc.c 				if (wqe->wr.send_flags & IB_SEND_SOLICITED)
wr                171 drivers/infiniband/hw/qib/qib_uc.c 		if (wqe->wr.opcode == IB_WR_SEND)
wr                176 drivers/infiniband/hw/qib/qib_uc.c 			ohdr->u.imm_data = wqe->wr.ex.imm_data;
wr                179 drivers/infiniband/hw/qib/qib_uc.c 		if (wqe->wr.send_flags & IB_SEND_SOLICITED)
wr                195 drivers/infiniband/hw/qib/qib_uc.c 		if (wqe->wr.opcode == IB_WR_RDMA_WRITE)
wr                201 drivers/infiniband/hw/qib/qib_uc.c 			ohdr->u.imm_data = wqe->wr.ex.imm_data;
wr                203 drivers/infiniband/hw/qib/qib_uc.c 			if (wqe->wr.send_flags & IB_SEND_SOLICITED)
wr                128 drivers/infiniband/hw/qib/qib_ud.c 	if (swqe->wr.opcode == IB_WR_SEND_WITH_IMM) {
wr                130 drivers/infiniband/hw/qib/qib_ud.c 		wc.ex.imm_data = swqe->wr.ex.imm_data;
wr                173 drivers/infiniband/hw/qib/qib_ud.c 	ssge.num_sge = swqe->wr.num_sge;
wr                214 drivers/infiniband/hw/qib/qib_ud.c 	rvt_recv_cq(qp, &wc, swqe->wr.send_flags & IB_SEND_SOLICITED);
wr                319 drivers/infiniband/hw/qib/qib_ud.c 	qp->s_sge.num_sge = wqe->wr.num_sge;
wr                338 drivers/infiniband/hw/qib/qib_ud.c 	if (wqe->wr.opcode == IB_WR_SEND_WITH_IMM) {
wr                340 drivers/infiniband/hw/qib/qib_ud.c 		ohdr->u.ud.imm_data = wqe->wr.ex.imm_data;
wr                361 drivers/infiniband/hw/qib/qib_ud.c 	if (wqe->wr.send_flags & IB_SEND_SOLICITED)
wr                312 drivers/infiniband/hw/qib/qib_verbs.h int qib_post_ud_send(struct rvt_qp *qp, const struct ib_send_wr *wr);
wr                594 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 		       const struct ib_reg_wr *wr)
wr                596 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 	struct pvrdma_user_mr *mr = to_vmr(wr->mr);
wr                598 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 	wqe_hdr->wr.fast_reg.iova_start = mr->ibmr.iova;
wr                599 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 	wqe_hdr->wr.fast_reg.pl_pdir_dma = mr->pdir.dir_dma;
wr                600 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 	wqe_hdr->wr.fast_reg.page_shift = mr->page_shift;
wr                601 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 	wqe_hdr->wr.fast_reg.page_list_len = mr->npages;
wr                602 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 	wqe_hdr->wr.fast_reg.length = mr->ibmr.length;
wr                603 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 	wqe_hdr->wr.fast_reg.access_flags = wr->access;
wr                604 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 	wqe_hdr->wr.fast_reg.rkey = wr->key;
wr                618 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c int pvrdma_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr,
wr                633 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 		*bad_wr = wr;
wr                639 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 	while (wr) {
wr                646 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 			*bad_wr = wr;
wr                651 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 		if (unlikely(wr->num_sge > qp->sq.max_sg || wr->num_sge < 0)) {
wr                654 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 			*bad_wr = wr;
wr                659 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 		if (unlikely(wr->opcode < 0)) {
wr                662 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 			*bad_wr = wr;
wr                687 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 			wr->opcode != IB_WR_SEND) {
wr                690 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 			*bad_wr = wr;
wr                695 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 			if (wr->opcode != IB_WR_SEND &&
wr                696 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 			    wr->opcode != IB_WR_SEND_WITH_IMM) {
wr                699 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 				*bad_wr = wr;
wr                707 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 		wqe_hdr->wr_id = wr->wr_id;
wr                708 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 		wqe_hdr->num_sge = wr->num_sge;
wr                709 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 		wqe_hdr->opcode = ib_wr_opcode_to_pvrdma(wr->opcode);
wr                710 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 		wqe_hdr->send_flags = ib_send_flags_to_pvrdma(wr->send_flags);
wr                711 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 		if (wr->opcode == IB_WR_SEND_WITH_IMM ||
wr                712 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 		    wr->opcode == IB_WR_RDMA_WRITE_WITH_IMM)
wr                713 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 			wqe_hdr->ex.imm_data = wr->ex.imm_data;
wr                716 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 			*bad_wr = wr;
wr                724 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 			if (unlikely(!ud_wr(wr)->ah)) {
wr                727 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 				*bad_wr = wr;
wr                736 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 			wqe_hdr->wr.ud.remote_qpn = ud_wr(wr)->remote_qpn;
wr                737 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 			wqe_hdr->wr.ud.remote_qkey =
wr                738 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 				ud_wr(wr)->remote_qkey & 0x80000000 ?
wr                739 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 				qp->qkey : ud_wr(wr)->remote_qkey;
wr                740 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 			wqe_hdr->wr.ud.av = to_vah(ud_wr(wr)->ah)->av;
wr                744 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 			switch (wr->opcode) {
wr                748 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 				wqe_hdr->wr.rdma.remote_addr =
wr                749 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 					rdma_wr(wr)->remote_addr;
wr                750 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 				wqe_hdr->wr.rdma.rkey = rdma_wr(wr)->rkey;
wr                755 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 					wr->ex.invalidate_rkey;
wr                759 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 				wqe_hdr->wr.atomic.remote_addr =
wr                760 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 					atomic_wr(wr)->remote_addr;
wr                761 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 				wqe_hdr->wr.atomic.rkey = atomic_wr(wr)->rkey;
wr                762 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 				wqe_hdr->wr.atomic.compare_add =
wr                763 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 					atomic_wr(wr)->compare_add;
wr                764 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 				if (wr->opcode == IB_WR_ATOMIC_CMP_AND_SWP)
wr                765 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 					wqe_hdr->wr.atomic.swap =
wr                766 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 						atomic_wr(wr)->swap;
wr                769 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 				ret = set_reg_seg(wqe_hdr, reg_wr(wr));
wr                773 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 					*bad_wr = wr;
wr                786 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 			*bad_wr = wr;
wr                791 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 		for (i = 0; i < wr->num_sge; i++) {
wr                793 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 			sge->addr = wr->sg_list[i].addr;
wr                794 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 			sge->length = wr->sg_list[i].length;
wr                795 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 			sge->lkey = wr->sg_list[i].lkey;
wr                806 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 		wr = wr->next;
wr                828 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c int pvrdma_post_recv(struct ib_qp *ibqp, const struct ib_recv_wr *wr,
wr                844 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 		*bad_wr = wr;
wr                850 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 		*bad_wr = wr;
wr                856 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 	while (wr) {
wr                859 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 		if (unlikely(wr->num_sge > qp->rq.max_sg ||
wr                860 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 			     wr->num_sge < 0)) {
wr                862 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 			*bad_wr = wr;
wr                871 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 			*bad_wr = wr;
wr                878 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 		wqe_hdr->wr_id = wr->wr_id;
wr                879 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 		wqe_hdr->num_sge = wr->num_sge;
wr                883 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 		for (i = 0; i < wr->num_sge; i++) {
wr                884 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 			sge->addr = wr->sg_list[i].addr;
wr                885 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 			sge->length = wr->sg_list[i].length;
wr                886 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 			sge->lkey = wr->sg_list[i].lkey;
wr                897 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 		wr = wr->next;
wr                436 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.h int pvrdma_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr,
wr                438 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.h int pvrdma_post_recv(struct ib_qp *ibqp, const struct ib_recv_wr *wr,
wr                664 drivers/infiniband/sw/rdmavt/qp.c 	for (i = 0; i < wqe->wr.num_sge; i++) {
wr               1821 drivers/infiniband/sw/rdmavt/qp.c int rvt_post_recv(struct ib_qp *ibqp, const struct ib_recv_wr *wr,
wr               1832 drivers/infiniband/sw/rdmavt/qp.c 		*bad_wr = wr;
wr               1836 drivers/infiniband/sw/rdmavt/qp.c 	for (; wr; wr = wr->next) {
wr               1841 drivers/infiniband/sw/rdmavt/qp.c 		if ((unsigned)wr->num_sge > qp->r_rq.max_sge) {
wr               1842 drivers/infiniband/sw/rdmavt/qp.c 			*bad_wr = wr;
wr               1852 drivers/infiniband/sw/rdmavt/qp.c 			*bad_wr = wr;
wr               1861 drivers/infiniband/sw/rdmavt/qp.c 			wc.wr_id = wr->wr_id;
wr               1866 drivers/infiniband/sw/rdmavt/qp.c 			wqe->wr_id = wr->wr_id;
wr               1867 drivers/infiniband/sw/rdmavt/qp.c 			wqe->num_sge = wr->num_sge;
wr               1868 drivers/infiniband/sw/rdmavt/qp.c 			for (i = 0; i < wr->num_sge; i++) {
wr               1869 drivers/infiniband/sw/rdmavt/qp.c 				wqe->sg_list[i].addr = wr->sg_list[i].addr;
wr               1870 drivers/infiniband/sw/rdmavt/qp.c 				wqe->sg_list[i].length = wr->sg_list[i].length;
wr               1871 drivers/infiniband/sw/rdmavt/qp.c 				wqe->sg_list[i].lkey = wr->sg_list[i].lkey;
wr               1905 drivers/infiniband/sw/rdmavt/qp.c 	const struct ib_send_wr *wr)
wr               1909 drivers/infiniband/sw/rdmavt/qp.c 	if (wr->opcode >= RVT_OPERATION_MAX || !post_parms[wr->opcode].length)
wr               1911 drivers/infiniband/sw/rdmavt/qp.c 	if (!(post_parms[wr->opcode].qpt_support & BIT(qp->ibqp.qp_type)))
wr               1913 drivers/infiniband/sw/rdmavt/qp.c 	if ((post_parms[wr->opcode].flags & RVT_OPERATION_PRIV) &&
wr               1916 drivers/infiniband/sw/rdmavt/qp.c 	if (post_parms[wr->opcode].flags & RVT_OPERATION_ATOMIC_SGE &&
wr               1917 drivers/infiniband/sw/rdmavt/qp.c 	    (wr->num_sge == 0 ||
wr               1918 drivers/infiniband/sw/rdmavt/qp.c 	     wr->sg_list[0].length < sizeof(u64) ||
wr               1919 drivers/infiniband/sw/rdmavt/qp.c 	     wr->sg_list[0].addr & (sizeof(u64) - 1)))
wr               1921 drivers/infiniband/sw/rdmavt/qp.c 	if (post_parms[wr->opcode].flags & RVT_OPERATION_ATOMIC &&
wr               1924 drivers/infiniband/sw/rdmavt/qp.c 	len = post_parms[wr->opcode].length;
wr               1928 drivers/infiniband/sw/rdmavt/qp.c 		if (qp->ibqp.pd != ud_wr(wr)->ah->pd)
wr               2000 drivers/infiniband/sw/rdmavt/qp.c 			   const struct ib_send_wr *wr,
wr               2020 drivers/infiniband/sw/rdmavt/qp.c 	if (unlikely(wr->num_sge > qp->s_max_sge))
wr               2023 drivers/infiniband/sw/rdmavt/qp.c 	ret = rvt_qp_valid_operation(qp, rdi->post_parms, wr);
wr               2039 drivers/infiniband/sw/rdmavt/qp.c 	if ((rdi->post_parms[wr->opcode].flags & RVT_OPERATION_LOCAL)) {
wr               2040 drivers/infiniband/sw/rdmavt/qp.c 		switch (wr->opcode) {
wr               2043 drivers/infiniband/sw/rdmavt/qp.c 					      reg_wr(wr)->mr,
wr               2044 drivers/infiniband/sw/rdmavt/qp.c 					      reg_wr(wr)->key,
wr               2045 drivers/infiniband/sw/rdmavt/qp.c 					      reg_wr(wr)->access);
wr               2046 drivers/infiniband/sw/rdmavt/qp.c 			if (ret || !(wr->send_flags & IB_SEND_SIGNALED))
wr               2050 drivers/infiniband/sw/rdmavt/qp.c 			if ((wr->send_flags & IB_SEND_FENCE) ||
wr               2055 drivers/infiniband/sw/rdmavt/qp.c 					qp, wr->ex.invalidate_rkey);
wr               2056 drivers/infiniband/sw/rdmavt/qp.c 				if (ret || !(wr->send_flags & IB_SEND_SIGNALED))
wr               2065 drivers/infiniband/sw/rdmavt/qp.c 	reserved_op = rdi->post_parms[wr->opcode].flags &
wr               2080 drivers/infiniband/sw/rdmavt/qp.c 	memcpy(&wqe->wr, wr, cplen);
wr               2084 drivers/infiniband/sw/rdmavt/qp.c 	if (wr->num_sge) {
wr               2087 drivers/infiniband/sw/rdmavt/qp.c 		acc = wr->opcode >= IB_WR_RDMA_READ ?
wr               2089 drivers/infiniband/sw/rdmavt/qp.c 		for (i = 0; i < wr->num_sge; i++) {
wr               2090 drivers/infiniband/sw/rdmavt/qp.c 			u32 length = wr->sg_list[i].length;
wr               2095 drivers/infiniband/sw/rdmavt/qp.c 					  &wr->sg_list[i], acc);
wr               2103 drivers/infiniband/sw/rdmavt/qp.c 		wqe->wr.num_sge = j;
wr               2119 drivers/infiniband/sw/rdmavt/qp.c 	if (rdi->post_parms[wr->opcode].flags & RVT_OPERATION_LOCAL) {
wr               2123 drivers/infiniband/sw/rdmavt/qp.c 			wqe->wr.send_flags |= RVT_SEND_COMPLETION_ONLY;
wr               2143 drivers/infiniband/sw/rdmavt/qp.c 	if (!(rdi->post_parms[wr->opcode].flags & RVT_OPERATION_LOCAL))
wr               2147 drivers/infiniband/sw/rdmavt/qp.c 		wqe->wr.send_flags |= RVT_SEND_RESERVE_USED;
wr               2150 drivers/infiniband/sw/rdmavt/qp.c 		wqe->wr.send_flags &= ~RVT_SEND_RESERVE_USED;
wr               2153 drivers/infiniband/sw/rdmavt/qp.c 	trace_rvt_post_one_wr(qp, wqe, wr->num_sge);
wr               2182 drivers/infiniband/sw/rdmavt/qp.c int rvt_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr,
wr               2208 drivers/infiniband/sw/rdmavt/qp.c 	call_send = qp->s_head == READ_ONCE(qp->s_last) && !wr->next;
wr               2210 drivers/infiniband/sw/rdmavt/qp.c 	for (; wr; wr = wr->next) {
wr               2211 drivers/infiniband/sw/rdmavt/qp.c 		err = rvt_post_one_wr(qp, wr, &call_send);
wr               2213 drivers/infiniband/sw/rdmavt/qp.c 			*bad_wr = wr;
wr               2243 drivers/infiniband/sw/rdmavt/qp.c int rvt_post_srq_recv(struct ib_srq *ibsrq, const struct ib_recv_wr *wr,
wr               2250 drivers/infiniband/sw/rdmavt/qp.c 	for (; wr; wr = wr->next) {
wr               2255 drivers/infiniband/sw/rdmavt/qp.c 		if ((unsigned)wr->num_sge > srq->rq.max_sge) {
wr               2256 drivers/infiniband/sw/rdmavt/qp.c 			*bad_wr = wr;
wr               2267 drivers/infiniband/sw/rdmavt/qp.c 			*bad_wr = wr;
wr               2272 drivers/infiniband/sw/rdmavt/qp.c 		wqe->wr_id = wr->wr_id;
wr               2273 drivers/infiniband/sw/rdmavt/qp.c 		wqe->num_sge = wr->num_sge;
wr               2274 drivers/infiniband/sw/rdmavt/qp.c 		for (i = 0; i < wr->num_sge; i++) {
wr               2275 drivers/infiniband/sw/rdmavt/qp.c 			wqe->sg_list[i].addr = wr->sg_list[i].addr;
wr               2276 drivers/infiniband/sw/rdmavt/qp.c 			wqe->sg_list[i].length = wr->sg_list[i].length;
wr               2277 drivers/infiniband/sw/rdmavt/qp.c 			wqe->sg_list[i].lkey = wr->sg_list[i].lkey;
wr               2852 drivers/infiniband/sw/rdmavt/qp.c 	last = rvt_qp_complete_swqe(qp, wqe, rdi->wc_opcode[wqe->wr.opcode],
wr               3043 drivers/infiniband/sw/rdmavt/qp.c 	sqp->s_sge.num_sge = wqe->wr.num_sge;
wr               3045 drivers/infiniband/sw/rdmavt/qp.c 	switch (wqe->wr.opcode) {
wr               3050 drivers/infiniband/sw/rdmavt/qp.c 		if (!(wqe->wr.send_flags & RVT_SEND_COMPLETION_ONLY)) {
wr               3052 drivers/infiniband/sw/rdmavt/qp.c 						wqe->wr.ex.invalidate_rkey))
wr               3068 drivers/infiniband/sw/rdmavt/qp.c 		switch (wqe->wr.opcode) {
wr               3071 drivers/infiniband/sw/rdmavt/qp.c 						 wqe->wr.ex.invalidate_rkey)) {
wr               3074 drivers/infiniband/sw/rdmavt/qp.c 					wqe->wr.ex.invalidate_rkey;
wr               3079 drivers/infiniband/sw/rdmavt/qp.c 			wc.ex.imm_data = wqe->wr.ex.imm_data;
wr               3090 drivers/infiniband/sw/rdmavt/qp.c 		wc.ex.imm_data = wqe->wr.ex.imm_data;
wr               3128 drivers/infiniband/sw/rdmavt/qp.c 		qp->r_sge.num_sge = wqe->wr.num_sge;
wr               3145 drivers/infiniband/sw/rdmavt/qp.c 			(wqe->wr.opcode == IB_WR_ATOMIC_FETCH_AND_ADD) ?
wr               3174 drivers/infiniband/sw/rdmavt/qp.c 	if (wqe->wr.opcode == IB_WR_RDMA_WRITE_WITH_IMM)
wr               3187 drivers/infiniband/sw/rdmavt/qp.c 	rvt_recv_cq(qp, &wc, wqe->wr.send_flags & IB_SEND_SOLICITED);
wr                 63 drivers/infiniband/sw/rdmavt/qp.h int rvt_post_recv(struct ib_qp *ibqp, const struct ib_recv_wr *wr,
wr                 65 drivers/infiniband/sw/rdmavt/qp.h int rvt_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr,
wr                 67 drivers/infiniband/sw/rdmavt/qp.h int rvt_post_srq_recv(struct ib_srq *ibsrq, const struct ib_recv_wr *wr,
wr                203 drivers/infiniband/sw/rdmavt/rc.c 	ss->num_sge = wqe->wr.num_sge;
wr                116 drivers/infiniband/sw/rdmavt/trace_tx.h 		__entry->wr_id = wqe->wr.wr_id;
wr                122 drivers/infiniband/sw/rdmavt/trace_tx.h 		__entry->opcode = wqe->wr.opcode;
wr                129 drivers/infiniband/sw/rdmavt/trace_tx.h 		__entry->send_flags = wqe->wr.send_flags;
wr                130 drivers/infiniband/sw/rdmavt/trace_tx.h 		__entry->num_sge = wqe->wr.num_sge;
wr                175 drivers/infiniband/sw/rdmavt/trace_tx.h 		__entry->wr_id = wqe->wr.wr_id;
wr                181 drivers/infiniband/sw/rdmavt/trace_tx.h 		__entry->opcode = wqe->wr.opcode;
wr                182 drivers/infiniband/sw/rdmavt/trace_tx.h 		__entry->send_flags = wqe->wr.send_flags;
wr                289 drivers/infiniband/sw/rxe/rxe_comp.c 		if (wqe->wr.opcode != IB_WR_RDMA_READ &&
wr                290 drivers/infiniband/sw/rxe/rxe_comp.c 		    wqe->wr.opcode != IB_WR_RDMA_READ_WITH_INV) {
wr                303 drivers/infiniband/sw/rxe/rxe_comp.c 		if (wqe->wr.opcode != IB_WR_ATOMIC_CMP_AND_SWP &&
wr                304 drivers/infiniband/sw/rxe/rxe_comp.c 		    wqe->wr.opcode != IB_WR_ATOMIC_FETCH_AND_ADD)
wr                410 drivers/infiniband/sw/rxe/rxe_comp.c 		wc->wr_id		= wqe->wr.wr_id;
wr                412 drivers/infiniband/sw/rxe/rxe_comp.c 		wc->opcode		= wr_to_wc_opcode(wqe->wr.opcode);
wr                413 drivers/infiniband/sw/rxe/rxe_comp.c 		if (wqe->wr.opcode == IB_WR_RDMA_WRITE_WITH_IMM ||
wr                414 drivers/infiniband/sw/rxe/rxe_comp.c 		    wqe->wr.opcode == IB_WR_SEND_WITH_IMM)
wr                421 drivers/infiniband/sw/rxe/rxe_comp.c 		uwc->wr_id		= wqe->wr.wr_id;
wr                423 drivers/infiniband/sw/rxe/rxe_comp.c 		uwc->opcode		= wr_to_wc_opcode(wqe->wr.opcode);
wr                424 drivers/infiniband/sw/rxe/rxe_comp.c 		if (wqe->wr.opcode == IB_WR_RDMA_WRITE_WITH_IMM ||
wr                425 drivers/infiniband/sw/rxe/rxe_comp.c 		    wqe->wr.opcode == IB_WR_SEND_WITH_IMM)
wr                446 drivers/infiniband/sw/rxe/rxe_comp.c 	    (wqe->wr.send_flags & IB_SEND_SIGNALED) ||
wr                455 drivers/infiniband/sw/rxe/rxe_comp.c 	if (wqe->wr.opcode == IB_WR_SEND ||
wr                456 drivers/infiniband/sw/rxe/rxe_comp.c 	    wqe->wr.opcode == IB_WR_SEND_WITH_IMM ||
wr                457 drivers/infiniband/sw/rxe/rxe_comp.c 	    wqe->wr.opcode == IB_WR_SEND_WITH_INV)
wr                 55 drivers/infiniband/sw/rxe/rxe_req.c 					     wqe->wr.opcode);
wr                 57 drivers/infiniband/sw/rxe/rxe_req.c 		if (wqe->wr.send_flags & IB_SEND_INLINE) {
wr                 84 drivers/infiniband/sw/rxe/rxe_req.c 		mask = wr_opcode_mask(wqe->wr.opcode, qp);
wr                 93 drivers/infiniband/sw/rxe/rxe_req.c 			     wqe->wr.wr.atomic.remote_addr :
wr                 95 drivers/infiniband/sw/rxe/rxe_req.c 			     wqe->wr.wr.rdma.remote_addr :
wr                184 drivers/infiniband/sw/rxe/rxe_req.c 	if (unlikely((wqe->wr.send_flags & IB_SEND_FENCE) &&
wr                190 drivers/infiniband/sw/rxe/rxe_req.c 	wqe->mask = wr_opcode_mask(wqe->wr.opcode, qp);
wr                386 drivers/infiniband/sw/rxe/rxe_req.c 	struct rxe_send_wr	*ibwr = &wqe->wr;
wr                423 drivers/infiniband/sw/rxe/rxe_req.c 		 port->pkey_tbl[ibwr->wr.ud.pkey_index] :
wr                426 drivers/infiniband/sw/rxe/rxe_req.c 	qp_num = (pkt->mask & RXE_DETH_MASK) ? ibwr->wr.ud.remote_qpn :
wr                439 drivers/infiniband/sw/rxe/rxe_req.c 		reth_set_rkey(pkt, ibwr->wr.rdma.rkey);
wr                454 drivers/infiniband/sw/rxe/rxe_req.c 			atmeth_set_swap_add(pkt, ibwr->wr.atomic.swap);
wr                455 drivers/infiniband/sw/rxe/rxe_req.c 			atmeth_set_comp(pkt, ibwr->wr.atomic.compare_add);
wr                457 drivers/infiniband/sw/rxe/rxe_req.c 			atmeth_set_swap_add(pkt, ibwr->wr.atomic.compare_add);
wr                459 drivers/infiniband/sw/rxe/rxe_req.c 		atmeth_set_rkey(pkt, ibwr->wr.atomic.rkey);
wr                466 drivers/infiniband/sw/rxe/rxe_req.c 			deth_set_qkey(pkt, ibwr->wr.ud.remote_qkey);
wr                487 drivers/infiniband/sw/rxe/rxe_req.c 		if (wqe->wr.send_flags & IB_SEND_INLINE) {
wr                628 drivers/infiniband/sw/rxe/rxe_req.c 		if (wqe->wr.opcode == IB_WR_LOCAL_INV) {
wr                633 drivers/infiniband/sw/rxe/rxe_req.c 						 wqe->wr.ex.invalidate_rkey >> 8);
wr                636 drivers/infiniband/sw/rxe/rxe_req.c 				       wqe->wr.ex.invalidate_rkey);
wr                645 drivers/infiniband/sw/rxe/rxe_req.c 		} else if (wqe->wr.opcode == IB_WR_REG_MR) {
wr                646 drivers/infiniband/sw/rxe/rxe_req.c 			struct rxe_mem *rmr = to_rmr(wqe->wr.wr.reg.mr);
wr                649 drivers/infiniband/sw/rxe/rxe_req.c 			rmr->access = wqe->wr.wr.reg.access;
wr                650 drivers/infiniband/sw/rxe/rxe_req.c 			rmr->lkey = wqe->wr.wr.reg.key;
wr                651 drivers/infiniband/sw/rxe/rxe_req.c 			rmr->rkey = wqe->wr.wr.reg.key;
wr                652 drivers/infiniband/sw/rxe/rxe_req.c 			rmr->iova = wqe->wr.wr.reg.mr->iova;
wr                658 drivers/infiniband/sw/rxe/rxe_req.c 		if ((wqe->wr.send_flags & IB_SEND_SIGNALED) ||
wr                679 drivers/infiniband/sw/rxe/rxe_req.c 	opcode = next_opcode(qp, wqe, wqe->wr.opcode);
wr                387 drivers/infiniband/sw/rxe/rxe_verbs.c static int rxe_post_srq_recv(struct ib_srq *ibsrq, const struct ib_recv_wr *wr,
wr                396 drivers/infiniband/sw/rxe/rxe_verbs.c 	while (wr) {
wr                397 drivers/infiniband/sw/rxe/rxe_verbs.c 		err = post_one_recv(&srq->rq, wr);
wr                400 drivers/infiniband/sw/rxe/rxe_verbs.c 		wr = wr->next;
wr                406 drivers/infiniband/sw/rxe/rxe_verbs.c 		*bad_wr = wr;
wr                530 drivers/infiniband/sw/rxe/rxe_verbs.c static void init_send_wr(struct rxe_qp *qp, struct rxe_send_wr *wr,
wr                533 drivers/infiniband/sw/rxe/rxe_verbs.c 	wr->wr_id = ibwr->wr_id;
wr                534 drivers/infiniband/sw/rxe/rxe_verbs.c 	wr->num_sge = ibwr->num_sge;
wr                535 drivers/infiniband/sw/rxe/rxe_verbs.c 	wr->opcode = ibwr->opcode;
wr                536 drivers/infiniband/sw/rxe/rxe_verbs.c 	wr->send_flags = ibwr->send_flags;
wr                541 drivers/infiniband/sw/rxe/rxe_verbs.c 		wr->wr.ud.remote_qpn = ud_wr(ibwr)->remote_qpn;
wr                542 drivers/infiniband/sw/rxe/rxe_verbs.c 		wr->wr.ud.remote_qkey = ud_wr(ibwr)->remote_qkey;
wr                544 drivers/infiniband/sw/rxe/rxe_verbs.c 			wr->wr.ud.pkey_index = ud_wr(ibwr)->pkey_index;
wr                545 drivers/infiniband/sw/rxe/rxe_verbs.c 		if (wr->opcode == IB_WR_SEND_WITH_IMM)
wr                546 drivers/infiniband/sw/rxe/rxe_verbs.c 			wr->ex.imm_data = ibwr->ex.imm_data;
wr                548 drivers/infiniband/sw/rxe/rxe_verbs.c 		switch (wr->opcode) {
wr                550 drivers/infiniband/sw/rxe/rxe_verbs.c 			wr->ex.imm_data = ibwr->ex.imm_data;
wr                554 drivers/infiniband/sw/rxe/rxe_verbs.c 			wr->wr.rdma.remote_addr = rdma_wr(ibwr)->remote_addr;
wr                555 drivers/infiniband/sw/rxe/rxe_verbs.c 			wr->wr.rdma.rkey	= rdma_wr(ibwr)->rkey;
wr                558 drivers/infiniband/sw/rxe/rxe_verbs.c 			wr->ex.imm_data = ibwr->ex.imm_data;
wr                561 drivers/infiniband/sw/rxe/rxe_verbs.c 			wr->ex.invalidate_rkey = ibwr->ex.invalidate_rkey;
wr                565 drivers/infiniband/sw/rxe/rxe_verbs.c 			wr->wr.atomic.remote_addr =
wr                567 drivers/infiniband/sw/rxe/rxe_verbs.c 			wr->wr.atomic.compare_add =
wr                569 drivers/infiniband/sw/rxe/rxe_verbs.c 			wr->wr.atomic.swap = atomic_wr(ibwr)->swap;
wr                570 drivers/infiniband/sw/rxe/rxe_verbs.c 			wr->wr.atomic.rkey = atomic_wr(ibwr)->rkey;
wr                573 drivers/infiniband/sw/rxe/rxe_verbs.c 			wr->ex.invalidate_rkey = ibwr->ex.invalidate_rkey;
wr                576 drivers/infiniband/sw/rxe/rxe_verbs.c 			wr->wr.reg.mr = reg_wr(ibwr)->mr;
wr                577 drivers/infiniband/sw/rxe/rxe_verbs.c 			wr->wr.reg.key = reg_wr(ibwr)->key;
wr                578 drivers/infiniband/sw/rxe/rxe_verbs.c 			wr->wr.reg.access = reg_wr(ibwr)->access;
wr                595 drivers/infiniband/sw/rxe/rxe_verbs.c 	init_send_wr(qp, &wqe->wr, ibwr);
wr                675 drivers/infiniband/sw/rxe/rxe_verbs.c static int rxe_post_send_kernel(struct rxe_qp *qp, const struct ib_send_wr *wr,
wr                683 drivers/infiniband/sw/rxe/rxe_verbs.c 	while (wr) {
wr                684 drivers/infiniband/sw/rxe/rxe_verbs.c 		mask = wr_opcode_mask(wr->opcode, qp);
wr                687 drivers/infiniband/sw/rxe/rxe_verbs.c 			*bad_wr = wr;
wr                691 drivers/infiniband/sw/rxe/rxe_verbs.c 		if (unlikely((wr->send_flags & IB_SEND_INLINE) &&
wr                694 drivers/infiniband/sw/rxe/rxe_verbs.c 			*bad_wr = wr;
wr                699 drivers/infiniband/sw/rxe/rxe_verbs.c 		for (i = 0; i < wr->num_sge; i++)
wr                700 drivers/infiniband/sw/rxe/rxe_verbs.c 			length += wr->sg_list[i].length;
wr                702 drivers/infiniband/sw/rxe/rxe_verbs.c 		err = post_one_send(qp, wr, mask, length);
wr                705 drivers/infiniband/sw/rxe/rxe_verbs.c 			*bad_wr = wr;
wr                708 drivers/infiniband/sw/rxe/rxe_verbs.c 		wr = wr->next;
wr                718 drivers/infiniband/sw/rxe/rxe_verbs.c static int rxe_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr,
wr                724 drivers/infiniband/sw/rxe/rxe_verbs.c 		*bad_wr = wr;
wr                729 drivers/infiniband/sw/rxe/rxe_verbs.c 		*bad_wr = wr;
wr                738 drivers/infiniband/sw/rxe/rxe_verbs.c 		return rxe_post_send_kernel(qp, wr, bad_wr);
wr                741 drivers/infiniband/sw/rxe/rxe_verbs.c static int rxe_post_recv(struct ib_qp *ibqp, const struct ib_recv_wr *wr,
wr                750 drivers/infiniband/sw/rxe/rxe_verbs.c 		*bad_wr = wr;
wr                756 drivers/infiniband/sw/rxe/rxe_verbs.c 		*bad_wr = wr;
wr                763 drivers/infiniband/sw/rxe/rxe_verbs.c 	while (wr) {
wr                764 drivers/infiniband/sw/rxe/rxe_verbs.c 		err = post_one_recv(rq, wr);
wr                766 drivers/infiniband/sw/rxe/rxe_verbs.c 			*bad_wr = wr;
wr                769 drivers/infiniband/sw/rxe/rxe_verbs.c 		wr = wr->next;
wr                689 drivers/infiniband/sw/siw/siw_verbs.c static int siw_sq_flush_wr(struct siw_qp *qp, const struct ib_send_wr *wr,
wr                695 drivers/infiniband/sw/siw/siw_verbs.c 	while (wr) {
wr                696 drivers/infiniband/sw/siw/siw_verbs.c 		sqe.id = wr->wr_id;
wr                697 drivers/infiniband/sw/siw/siw_verbs.c 		sqe.opcode = wr->opcode;
wr                701 drivers/infiniband/sw/siw/siw_verbs.c 				*bad_wr = wr;
wr                704 drivers/infiniband/sw/siw/siw_verbs.c 		wr = wr->next;
wr                710 drivers/infiniband/sw/siw/siw_verbs.c static int siw_rq_flush_wr(struct siw_qp *qp, const struct ib_recv_wr *wr,
wr                716 drivers/infiniband/sw/siw/siw_verbs.c 	while (wr) {
wr                717 drivers/infiniband/sw/siw/siw_verbs.c 		rqe.id = wr->wr_id;
wr                721 drivers/infiniband/sw/siw/siw_verbs.c 				*bad_wr = wr;
wr                724 drivers/infiniband/sw/siw/siw_verbs.c 		wr = wr->next;
wr                738 drivers/infiniband/sw/siw/siw_verbs.c int siw_post_send(struct ib_qp *base_qp, const struct ib_send_wr *wr,
wr                747 drivers/infiniband/sw/siw/siw_verbs.c 	if (wr && !qp->kernel_verbs) {
wr                749 drivers/infiniband/sw/siw/siw_verbs.c 		*bad_wr = wr;
wr                768 drivers/infiniband/sw/siw/siw_verbs.c 			rv = siw_sq_flush_wr(qp, wr, bad_wr);
wr                772 drivers/infiniband/sw/siw/siw_verbs.c 			*bad_wr = wr;
wr                786 drivers/infiniband/sw/siw/siw_verbs.c 			rv = siw_sq_flush_wr(qp, wr, bad_wr);
wr                790 drivers/infiniband/sw/siw/siw_verbs.c 			*bad_wr = wr;
wr                798 drivers/infiniband/sw/siw/siw_verbs.c 	while (wr) {
wr                807 drivers/infiniband/sw/siw/siw_verbs.c 		if (wr->num_sge > qp->attrs.sq_max_sges) {
wr                808 drivers/infiniband/sw/siw/siw_verbs.c 			siw_dbg_qp(qp, "too many sge's: %d\n", wr->num_sge);
wr                812 drivers/infiniband/sw/siw/siw_verbs.c 		sqe->id = wr->wr_id;
wr                814 drivers/infiniband/sw/siw/siw_verbs.c 		if ((wr->send_flags & IB_SEND_SIGNALED) ||
wr                818 drivers/infiniband/sw/siw/siw_verbs.c 		if (wr->send_flags & IB_SEND_FENCE)
wr                821 drivers/infiniband/sw/siw/siw_verbs.c 		switch (wr->opcode) {
wr                824 drivers/infiniband/sw/siw/siw_verbs.c 			if (wr->send_flags & IB_SEND_SOLICITED)
wr                827 drivers/infiniband/sw/siw/siw_verbs.c 			if (!(wr->send_flags & IB_SEND_INLINE)) {
wr                828 drivers/infiniband/sw/siw/siw_verbs.c 				siw_copy_sgl(wr->sg_list, sqe->sge,
wr                829 drivers/infiniband/sw/siw/siw_verbs.c 					     wr->num_sge);
wr                830 drivers/infiniband/sw/siw/siw_verbs.c 				sqe->num_sge = wr->num_sge;
wr                832 drivers/infiniband/sw/siw/siw_verbs.c 				rv = siw_copy_inline_sgl(wr, sqe);
wr                840 drivers/infiniband/sw/siw/siw_verbs.c 			if (wr->opcode == IB_WR_SEND)
wr                844 drivers/infiniband/sw/siw/siw_verbs.c 				sqe->rkey = wr->ex.invalidate_rkey;
wr                857 drivers/infiniband/sw/siw/siw_verbs.c 			if (unlikely(wr->num_sge != 1)) {
wr                861 drivers/infiniband/sw/siw/siw_verbs.c 			siw_copy_sgl(wr->sg_list, &sqe->sge[0], 1);
wr                865 drivers/infiniband/sw/siw/siw_verbs.c 			sqe->raddr = rdma_wr(wr)->remote_addr;
wr                866 drivers/infiniband/sw/siw/siw_verbs.c 			sqe->rkey = rdma_wr(wr)->rkey;
wr                869 drivers/infiniband/sw/siw/siw_verbs.c 			if (wr->opcode == IB_WR_RDMA_READ)
wr                876 drivers/infiniband/sw/siw/siw_verbs.c 			if (!(wr->send_flags & IB_SEND_INLINE)) {
wr                877 drivers/infiniband/sw/siw/siw_verbs.c 				siw_copy_sgl(wr->sg_list, &sqe->sge[0],
wr                878 drivers/infiniband/sw/siw/siw_verbs.c 					     wr->num_sge);
wr                879 drivers/infiniband/sw/siw/siw_verbs.c 				sqe->num_sge = wr->num_sge;
wr                881 drivers/infiniband/sw/siw/siw_verbs.c 				rv = siw_copy_inline_sgl(wr, sqe);
wr                889 drivers/infiniband/sw/siw/siw_verbs.c 			sqe->raddr = rdma_wr(wr)->remote_addr;
wr                890 drivers/infiniband/sw/siw/siw_verbs.c 			sqe->rkey = rdma_wr(wr)->rkey;
wr                895 drivers/infiniband/sw/siw/siw_verbs.c 			sqe->base_mr = (uintptr_t)reg_wr(wr)->mr;
wr                896 drivers/infiniband/sw/siw/siw_verbs.c 			sqe->rkey = reg_wr(wr)->key;
wr                897 drivers/infiniband/sw/siw/siw_verbs.c 			sqe->access = reg_wr(wr)->access & IWARP_ACCESS_MASK;
wr                902 drivers/infiniband/sw/siw/siw_verbs.c 			sqe->rkey = wr->ex.invalidate_rkey;
wr                908 drivers/infiniband/sw/siw/siw_verbs.c 				   wr->opcode);
wr                924 drivers/infiniband/sw/siw/siw_verbs.c 		wr = wr->next;
wr                965 drivers/infiniband/sw/siw/siw_verbs.c 	*bad_wr = wr;
wr                978 drivers/infiniband/sw/siw/siw_verbs.c int siw_post_receive(struct ib_qp *base_qp, const struct ib_recv_wr *wr,
wr                986 drivers/infiniband/sw/siw/siw_verbs.c 		*bad_wr = wr;
wr                991 drivers/infiniband/sw/siw/siw_verbs.c 		*bad_wr = wr;
wr               1010 drivers/infiniband/sw/siw/siw_verbs.c 			rv = siw_rq_flush_wr(qp, wr, bad_wr);
wr               1014 drivers/infiniband/sw/siw/siw_verbs.c 			*bad_wr = wr;
wr               1028 drivers/infiniband/sw/siw/siw_verbs.c 			rv = siw_rq_flush_wr(qp, wr, bad_wr);
wr               1032 drivers/infiniband/sw/siw/siw_verbs.c 			*bad_wr = wr;
wr               1044 drivers/infiniband/sw/siw/siw_verbs.c 	while (wr) {
wr               1053 drivers/infiniband/sw/siw/siw_verbs.c 		if (wr->num_sge > qp->attrs.rq_max_sges) {
wr               1054 drivers/infiniband/sw/siw/siw_verbs.c 			siw_dbg_qp(qp, "too many sge's: %d\n", wr->num_sge);
wr               1058 drivers/infiniband/sw/siw/siw_verbs.c 		rqe->id = wr->wr_id;
wr               1059 drivers/infiniband/sw/siw/siw_verbs.c 		rqe->num_sge = wr->num_sge;
wr               1060 drivers/infiniband/sw/siw/siw_verbs.c 		siw_copy_sgl(wr->sg_list, rqe->sge, wr->num_sge);
wr               1068 drivers/infiniband/sw/siw/siw_verbs.c 		wr = wr->next;
wr               1076 drivers/infiniband/sw/siw/siw_verbs.c 		*bad_wr = wr;
wr               1744 drivers/infiniband/sw/siw/siw_verbs.c int siw_post_srq_recv(struct ib_srq *base_srq, const struct ib_recv_wr *wr,
wr               1764 drivers/infiniband/sw/siw/siw_verbs.c 	while (wr) {
wr               1773 drivers/infiniband/sw/siw/siw_verbs.c 		if (unlikely(wr->num_sge > srq->max_sge)) {
wr               1775 drivers/infiniband/sw/siw/siw_verbs.c 				   "[SRQ]: too many sge's: %d\n", wr->num_sge);
wr               1779 drivers/infiniband/sw/siw/siw_verbs.c 		rqe->id = wr->wr_id;
wr               1780 drivers/infiniband/sw/siw/siw_verbs.c 		rqe->num_sge = wr->num_sge;
wr               1781 drivers/infiniband/sw/siw/siw_verbs.c 		siw_copy_sgl(wr->sg_list, rqe->sge, wr->num_sge);
wr               1789 drivers/infiniband/sw/siw/siw_verbs.c 		wr = wr->next;
wr               1795 drivers/infiniband/sw/siw/siw_verbs.c 		*bad_wr = wr;
wr                 62 drivers/infiniband/sw/siw/siw_verbs.h int siw_post_send(struct ib_qp *base_qp, const struct ib_send_wr *wr,
wr                 64 drivers/infiniband/sw/siw/siw_verbs.h int siw_post_receive(struct ib_qp *base_qp, const struct ib_recv_wr *wr,
wr                 83 drivers/infiniband/sw/siw/siw_verbs.h int siw_post_srq_recv(struct ib_srq *base_srq, const struct ib_recv_wr *wr,
wr                561 drivers/infiniband/ulp/ipoib/ipoib.h 	priv->tx_wr.wr.num_sge	     = nr_frags + off;
wr                118 drivers/infiniband/ulp/ipoib/ipoib_cm.c 					struct ib_recv_wr *wr,
wr                124 drivers/infiniband/ulp/ipoib/ipoib_cm.c 	wr->wr_id = id | IPOIB_OP_CM | IPOIB_OP_RECV;
wr                129 drivers/infiniband/ulp/ipoib/ipoib_cm.c 	ret = ib_post_recv(rx->qp, wr, NULL);
wr                329 drivers/infiniband/ulp/ipoib/ipoib_cm.c 				struct ib_recv_wr *wr,
wr                342 drivers/infiniband/ulp/ipoib/ipoib_cm.c 	wr->next    = NULL;
wr                343 drivers/infiniband/ulp/ipoib/ipoib_cm.c 	wr->sg_list = sge;
wr                344 drivers/infiniband/ulp/ipoib/ipoib_cm.c 	wr->num_sge = priv->cm.num_frags;
wr                352 drivers/infiniband/ulp/ipoib/ipoib_cm.c 		struct ib_recv_wr wr;
wr                369 drivers/infiniband/ulp/ipoib/ipoib_cm.c 	ipoib_cm_init_rx_wr(dev, &t->wr, t->sge);
wr                391 drivers/infiniband/ulp/ipoib/ipoib_cm.c 		ret = ipoib_cm_post_receive_nonsrq(dev, rx, &t->wr, t->sge, i);
wr                703 drivers/infiniband/ulp/ipoib/ipoib_cm.c 	priv->tx_wr.wr.wr_id	= wr_id | IPOIB_OP_CM;
wr                705 drivers/infiniband/ulp/ipoib/ipoib_cm.c 	return ib_post_send(tx->qp, &priv->tx_wr.wr, NULL);
wr                551 drivers/infiniband/ulp/ipoib/ipoib_ib.c 	priv->tx_wr.wr.wr_id	= wr_id;
wr                559 drivers/infiniband/ulp/ipoib/ipoib_ib.c 		priv->tx_wr.wr.opcode	= IB_WR_LSO;
wr                561 drivers/infiniband/ulp/ipoib/ipoib_ib.c 		priv->tx_wr.wr.opcode	= IB_WR_SEND;
wr                563 drivers/infiniband/ulp/ipoib/ipoib_ib.c 	return ib_post_send(priv->qp, &priv->tx_wr.wr, NULL);
wr                635 drivers/infiniband/ulp/ipoib/ipoib_ib.c 		priv->tx_wr.wr.send_flags |= IB_SEND_IP_CSUM;
wr                637 drivers/infiniband/ulp/ipoib/ipoib_ib.c 		priv->tx_wr.wr.send_flags &= ~IB_SEND_IP_CSUM;
wr                537 drivers/infiniband/ulp/ipoib/ipoib_main.c 		priv->tx_wr.wr.send_flags &= ~IB_SEND_IP_CSUM;
wr                221 drivers/infiniband/ulp/ipoib/ipoib_verbs.c 	priv->tx_wr.wr.opcode		= IB_WR_SEND;
wr                222 drivers/infiniband/ulp/ipoib/ipoib_verbs.c 	priv->tx_wr.wr.sg_list		= priv->tx_sge;
wr                223 drivers/infiniband/ulp/ipoib/ipoib_verbs.c 	priv->tx_wr.wr.send_flags	= IB_SEND_SIGNALED;
wr                209 drivers/infiniband/ulp/iser/iscsi_iser.c 	tx_desc->reg_wr.wr.next = NULL;
wr                390 drivers/infiniband/ulp/iser/iser_memory.c 	struct ib_reg_wr *wr = &tx_desc->reg_wr;
wr                401 drivers/infiniband/ulp/iser/iser_memory.c 		iser_inv_rkey(&tx_desc->inv_wr, mr, cqe, &wr->wr);
wr                413 drivers/infiniband/ulp/iser/iser_memory.c 	memset(wr, 0, sizeof(*wr));
wr                414 drivers/infiniband/ulp/iser/iser_memory.c 	wr->wr.next = &tx_desc->send_wr;
wr                415 drivers/infiniband/ulp/iser/iser_memory.c 	wr->wr.opcode = IB_WR_REG_MR_INTEGRITY;
wr                416 drivers/infiniband/ulp/iser/iser_memory.c 	wr->wr.wr_cqe = cqe;
wr                417 drivers/infiniband/ulp/iser/iser_memory.c 	wr->wr.num_sge = 0;
wr                418 drivers/infiniband/ulp/iser/iser_memory.c 	wr->wr.send_flags = 0;
wr                419 drivers/infiniband/ulp/iser/iser_memory.c 	wr->mr = mr;
wr                420 drivers/infiniband/ulp/iser/iser_memory.c 	wr->key = mr->rkey;
wr                421 drivers/infiniband/ulp/iser/iser_memory.c 	wr->access = IB_ACCESS_LOCAL_WRITE |
wr                446 drivers/infiniband/ulp/iser/iser_memory.c 	struct ib_reg_wr *wr = &tx_desc->reg_wr;
wr                450 drivers/infiniband/ulp/iser/iser_memory.c 		iser_inv_rkey(&tx_desc->inv_wr, mr, cqe, &wr->wr);
wr                461 drivers/infiniband/ulp/iser/iser_memory.c 	wr->wr.next = &tx_desc->send_wr;
wr                462 drivers/infiniband/ulp/iser/iser_memory.c 	wr->wr.opcode = IB_WR_REG_MR;
wr                463 drivers/infiniband/ulp/iser/iser_memory.c 	wr->wr.wr_cqe = cqe;
wr                464 drivers/infiniband/ulp/iser/iser_memory.c 	wr->wr.send_flags = 0;
wr                465 drivers/infiniband/ulp/iser/iser_memory.c 	wr->wr.num_sge = 0;
wr                466 drivers/infiniband/ulp/iser/iser_memory.c 	wr->mr = mr;
wr                467 drivers/infiniband/ulp/iser/iser_memory.c 	wr->key = mr->rkey;
wr                468 drivers/infiniband/ulp/iser/iser_memory.c 	wr->access = IB_ACCESS_LOCAL_WRITE  |
wr                976 drivers/infiniband/ulp/iser/iser_verbs.c 	struct ib_recv_wr wr;
wr                984 drivers/infiniband/ulp/iser/iser_verbs.c 	wr.wr_cqe = &desc->cqe;
wr                985 drivers/infiniband/ulp/iser/iser_verbs.c 	wr.sg_list = &desc->sge;
wr                986 drivers/infiniband/ulp/iser/iser_verbs.c 	wr.num_sge = 1;
wr                987 drivers/infiniband/ulp/iser/iser_verbs.c 	wr.next = NULL;
wr                990 drivers/infiniband/ulp/iser/iser_verbs.c 	ib_ret = ib_post_recv(ib_conn->qp, &wr, NULL);
wr               1004 drivers/infiniband/ulp/iser/iser_verbs.c 	struct ib_recv_wr *wr;
wr               1007 drivers/infiniband/ulp/iser/iser_verbs.c 	for (wr = ib_conn->rx_wr, i = 0; i < count; i++, wr++) {
wr               1010 drivers/infiniband/ulp/iser/iser_verbs.c 		wr->wr_cqe = &rx_desc->cqe;
wr               1011 drivers/infiniband/ulp/iser/iser_verbs.c 		wr->sg_list = &rx_desc->rx_sg;
wr               1012 drivers/infiniband/ulp/iser/iser_verbs.c 		wr->num_sge = 1;
wr               1013 drivers/infiniband/ulp/iser/iser_verbs.c 		wr->next = wr + 1;
wr               1017 drivers/infiniband/ulp/iser/iser_verbs.c 	wr--;
wr               1018 drivers/infiniband/ulp/iser/iser_verbs.c 	wr->next = NULL; /* mark end of work requests list */
wr               1040 drivers/infiniband/ulp/iser/iser_verbs.c 	struct ib_send_wr *wr = &tx_desc->send_wr;
wr               1048 drivers/infiniband/ulp/iser/iser_verbs.c 	wr->next = NULL;
wr               1049 drivers/infiniband/ulp/iser/iser_verbs.c 	wr->wr_cqe = &tx_desc->cqe;
wr               1050 drivers/infiniband/ulp/iser/iser_verbs.c 	wr->sg_list = tx_desc->tx_sg;
wr               1051 drivers/infiniband/ulp/iser/iser_verbs.c 	wr->num_sge = tx_desc->num_sge;
wr               1052 drivers/infiniband/ulp/iser/iser_verbs.c 	wr->opcode = IB_WR_SEND;
wr               1053 drivers/infiniband/ulp/iser/iser_verbs.c 	wr->send_flags = signal ? IB_SEND_SIGNALED : 0;
wr               1057 drivers/infiniband/ulp/iser/iser_verbs.c 	else if (tx_desc->reg_wr.wr.next)
wr               1058 drivers/infiniband/ulp/iser/iser_verbs.c 		first_wr = &tx_desc->reg_wr.wr;
wr               1060 drivers/infiniband/ulp/iser/iser_verbs.c 		first_wr = wr;
wr               1065 drivers/infiniband/ulp/iser/iser_verbs.c 			 ib_ret, wr->opcode);
wr               1229 drivers/infiniband/ulp/srp/ib_srp.c 	struct ib_send_wr wr = {
wr               1237 drivers/infiniband/ulp/srp/ib_srp.c 	wr.wr_cqe = &req->reg_cqe;
wr               1239 drivers/infiniband/ulp/srp/ib_srp.c 	return ib_post_send(ch->qp, &wr, NULL);
wr               1527 drivers/infiniband/ulp/srp/ib_srp.c 	struct ib_reg_wr wr;
wr               1573 drivers/infiniband/ulp/srp/ib_srp.c 	wr.wr.next = NULL;
wr               1574 drivers/infiniband/ulp/srp/ib_srp.c 	wr.wr.opcode = IB_WR_REG_MR;
wr               1575 drivers/infiniband/ulp/srp/ib_srp.c 	wr.wr.wr_cqe = &req->reg_cqe;
wr               1576 drivers/infiniband/ulp/srp/ib_srp.c 	wr.wr.num_sge = 0;
wr               1577 drivers/infiniband/ulp/srp/ib_srp.c 	wr.wr.send_flags = 0;
wr               1578 drivers/infiniband/ulp/srp/ib_srp.c 	wr.mr = desc->mr;
wr               1579 drivers/infiniband/ulp/srp/ib_srp.c 	wr.key = desc->mr->rkey;
wr               1580 drivers/infiniband/ulp/srp/ib_srp.c 	wr.access = (IB_ACCESS_LOCAL_WRITE |
wr               1590 drivers/infiniband/ulp/srp/ib_srp.c 	err = ib_post_send(ch->qp, &wr.wr, NULL);
wr               2074 drivers/infiniband/ulp/srp/ib_srp.c 	struct ib_send_wr wr;
wr               2085 drivers/infiniband/ulp/srp/ib_srp.c 	wr.next       = NULL;
wr               2086 drivers/infiniband/ulp/srp/ib_srp.c 	wr.wr_cqe     = &iu->cqe;
wr               2087 drivers/infiniband/ulp/srp/ib_srp.c 	wr.sg_list    = &iu->sge[0];
wr               2088 drivers/infiniband/ulp/srp/ib_srp.c 	wr.num_sge    = iu->num_sge;
wr               2089 drivers/infiniband/ulp/srp/ib_srp.c 	wr.opcode     = IB_WR_SEND;
wr               2090 drivers/infiniband/ulp/srp/ib_srp.c 	wr.send_flags = IB_SEND_SIGNALED;
wr               2092 drivers/infiniband/ulp/srp/ib_srp.c 	return ib_post_send(ch->qp, &wr, NULL);
wr               2098 drivers/infiniband/ulp/srp/ib_srp.c 	struct ib_recv_wr wr;
wr               2107 drivers/infiniband/ulp/srp/ib_srp.c 	wr.next     = NULL;
wr               2108 drivers/infiniband/ulp/srp/ib_srp.c 	wr.wr_cqe   = &iu->cqe;
wr               2109 drivers/infiniband/ulp/srp/ib_srp.c 	wr.sg_list  = &list;
wr               2110 drivers/infiniband/ulp/srp/ib_srp.c 	wr.num_sge  = 1;
wr               2112 drivers/infiniband/ulp/srp/ib_srp.c 	return ib_post_recv(ch->qp, &wr, NULL);
wr                825 drivers/infiniband/ulp/srpt/ib_srpt.c 	struct ib_recv_wr wr;
wr                833 drivers/infiniband/ulp/srpt/ib_srpt.c 	wr.wr_cqe = &ioctx->ioctx.cqe;
wr                834 drivers/infiniband/ulp/srpt/ib_srpt.c 	wr.next = NULL;
wr                835 drivers/infiniband/ulp/srpt/ib_srpt.c 	wr.sg_list = &list;
wr                836 drivers/infiniband/ulp/srpt/ib_srpt.c 	wr.num_sge = 1;
wr                839 drivers/infiniband/ulp/srpt/ib_srpt.c 		return ib_post_srq_recv(sdev->srq, &wr, NULL);
wr                841 drivers/infiniband/ulp/srpt/ib_srpt.c 		return ib_post_recv(ch->qp, &wr, NULL);
wr                855 drivers/infiniband/ulp/srpt/ib_srpt.c 	struct ib_rdma_wr wr = {
wr                856 drivers/infiniband/ulp/srpt/ib_srpt.c 		.wr = {
wr                867 drivers/infiniband/ulp/srpt/ib_srpt.c 	return ib_post_send(ch->qp, &wr.wr, NULL);
wr                732 drivers/irqchip/irq-gic-v3-its.c 	u64 wr = its_cmd_ptr_to_offset(its, its->cmd_write);
wr                734 drivers/irqchip/irq-gic-v3-its.c 	writel_relaxed(wr, its->base + GITS_CWRITER);
wr                752 drivers/md/dm-integrity.c static void rw_section_mac(struct dm_integrity_c *ic, unsigned section, bool wr)
wr                765 drivers/md/dm-integrity.c 		if (likely(wr))
wr                224 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	struct scatterlist *rd, *wr;
wr                244 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	wr = sgt->sgl;
wr                246 drivers/media/common/videobuf2/videobuf2-dma-contig.c 		sg_set_page(wr, sg_page(rd), rd->length, rd->offset);
wr                248 drivers/media/common/videobuf2/videobuf2-dma-contig.c 		wr = sg_next(wr);
wr                368 drivers/media/common/videobuf2/videobuf2-dma-sg.c 	struct scatterlist *rd, *wr;
wr                388 drivers/media/common/videobuf2/videobuf2-dma-sg.c 	wr = sgt->sgl;
wr                390 drivers/media/common/videobuf2/videobuf2-dma-sg.c 		sg_set_page(wr, sg_page(rd), rd->length, rd->offset);
wr                392 drivers/media/common/videobuf2/videobuf2-dma-sg.c 		wr = sg_next(wr);
wr                148 drivers/media/dvb-frontends/dib3000mb.c 	wr(DIB3000MB_REG_LOCK1_MASK, DIB3000MB_LOCK1_SEARCH_4);
wr                153 drivers/media/dvb-frontends/dib3000mb.c 			wr(DIB3000MB_REG_FFT, DIB3000_TRANSMISSION_MODE_2K);
wr                157 drivers/media/dvb-frontends/dib3000mb.c 			wr(DIB3000MB_REG_FFT, DIB3000_TRANSMISSION_MODE_8K);
wr                169 drivers/media/dvb-frontends/dib3000mb.c 			wr(DIB3000MB_REG_GUARD_TIME, DIB3000_GUARD_TIME_1_32);
wr                173 drivers/media/dvb-frontends/dib3000mb.c 			wr(DIB3000MB_REG_GUARD_TIME, DIB3000_GUARD_TIME_1_16);
wr                177 drivers/media/dvb-frontends/dib3000mb.c 			wr(DIB3000MB_REG_GUARD_TIME, DIB3000_GUARD_TIME_1_8);
wr                181 drivers/media/dvb-frontends/dib3000mb.c 			wr(DIB3000MB_REG_GUARD_TIME, DIB3000_GUARD_TIME_1_4);
wr                193 drivers/media/dvb-frontends/dib3000mb.c 			wr(DIB3000MB_REG_DDS_INV, DIB3000_DDS_INVERSION_OFF);
wr                200 drivers/media/dvb-frontends/dib3000mb.c 			wr(DIB3000MB_REG_DDS_INV, DIB3000_DDS_INVERSION_ON);
wr                209 drivers/media/dvb-frontends/dib3000mb.c 			wr(DIB3000MB_REG_QAM, DIB3000_CONSTELLATION_QPSK);
wr                213 drivers/media/dvb-frontends/dib3000mb.c 			wr(DIB3000MB_REG_QAM, DIB3000_CONSTELLATION_16QAM);
wr                217 drivers/media/dvb-frontends/dib3000mb.c 			wr(DIB3000MB_REG_QAM, DIB3000_CONSTELLATION_64QAM);
wr                230 drivers/media/dvb-frontends/dib3000mb.c 			wr(DIB3000MB_REG_VIT_ALPHA, DIB3000_ALPHA_1);
wr                234 drivers/media/dvb-frontends/dib3000mb.c 			wr(DIB3000MB_REG_VIT_ALPHA, DIB3000_ALPHA_2);
wr                238 drivers/media/dvb-frontends/dib3000mb.c 			wr(DIB3000MB_REG_VIT_ALPHA, DIB3000_ALPHA_4);
wr                248 drivers/media/dvb-frontends/dib3000mb.c 		wr(DIB3000MB_REG_VIT_HRCH, DIB3000_HRCH_OFF);
wr                249 drivers/media/dvb-frontends/dib3000mb.c 		wr(DIB3000MB_REG_VIT_HP, DIB3000_SELECT_HP);
wr                252 drivers/media/dvb-frontends/dib3000mb.c 		wr(DIB3000MB_REG_VIT_HRCH, DIB3000_HRCH_ON);
wr                253 drivers/media/dvb-frontends/dib3000mb.c 		wr(DIB3000MB_REG_VIT_HP, DIB3000_SELECT_LP);
wr                259 drivers/media/dvb-frontends/dib3000mb.c 			wr(DIB3000MB_REG_VIT_CODE_RATE, DIB3000_FEC_1_2);
wr                263 drivers/media/dvb-frontends/dib3000mb.c 			wr(DIB3000MB_REG_VIT_CODE_RATE, DIB3000_FEC_2_3);
wr                267 drivers/media/dvb-frontends/dib3000mb.c 			wr(DIB3000MB_REG_VIT_CODE_RATE, DIB3000_FEC_3_4);
wr                271 drivers/media/dvb-frontends/dib3000mb.c 			wr(DIB3000MB_REG_VIT_CODE_RATE, DIB3000_FEC_5_6);
wr                275 drivers/media/dvb-frontends/dib3000mb.c 			wr(DIB3000MB_REG_VIT_CODE_RATE, DIB3000_FEC_7_8);
wr                294 drivers/media/dvb-frontends/dib3000mb.c 	wr(DIB3000MB_REG_SEQ, seq);
wr                296 drivers/media/dvb-frontends/dib3000mb.c 	wr(DIB3000MB_REG_ISI, seq ? DIB3000MB_ISI_INHIBIT : DIB3000MB_ISI_ACTIVATE);
wr                300 drivers/media/dvb-frontends/dib3000mb.c 			wr(DIB3000MB_REG_SYNC_IMPROVEMENT, DIB3000MB_SYNC_IMPROVE_2K_1_8);
wr                302 drivers/media/dvb-frontends/dib3000mb.c 			wr(DIB3000MB_REG_SYNC_IMPROVEMENT, DIB3000MB_SYNC_IMPROVE_DEFAULT);
wr                305 drivers/media/dvb-frontends/dib3000mb.c 		wr(DIB3000MB_REG_UNK_121, DIB3000MB_UNK_121_2K);
wr                307 drivers/media/dvb-frontends/dib3000mb.c 		wr(DIB3000MB_REG_UNK_121, DIB3000MB_UNK_121_DEFAULT);
wr                310 drivers/media/dvb-frontends/dib3000mb.c 	wr(DIB3000MB_REG_MOBILE_ALGO, DIB3000MB_MOBILE_ALGO_OFF);
wr                311 drivers/media/dvb-frontends/dib3000mb.c 	wr(DIB3000MB_REG_MOBILE_MODE_QAM, DIB3000MB_MOBILE_MODE_QAM_OFF);
wr                312 drivers/media/dvb-frontends/dib3000mb.c 	wr(DIB3000MB_REG_MOBILE_MODE, DIB3000MB_MOBILE_MODE_OFF);
wr                316 drivers/media/dvb-frontends/dib3000mb.c 	wr(DIB3000MB_REG_ISI, DIB3000MB_ISI_ACTIVATE);
wr                318 drivers/media/dvb-frontends/dib3000mb.c 	wr(DIB3000MB_REG_RESTART, DIB3000MB_RESTART_AGC + DIB3000MB_RESTART_CTRL);
wr                319 drivers/media/dvb-frontends/dib3000mb.c 	wr(DIB3000MB_REG_RESTART, DIB3000MB_RESTART_OFF);
wr                335 drivers/media/dvb-frontends/dib3000mb.c 		wr(DIB3000MB_REG_ISI, DIB3000MB_ISI_INHIBIT);
wr                337 drivers/media/dvb-frontends/dib3000mb.c 		wr(DIB3000MB_REG_RESTART, DIB3000MB_RESTART_AUTO_SEARCH);
wr                338 drivers/media/dvb-frontends/dib3000mb.c 		wr(DIB3000MB_REG_RESTART, DIB3000MB_RESTART_OFF);
wr                357 drivers/media/dvb-frontends/dib3000mb.c 		wr(DIB3000MB_REG_RESTART, DIB3000MB_RESTART_CTRL);
wr                358 drivers/media/dvb-frontends/dib3000mb.c 		wr(DIB3000MB_REG_RESTART, DIB3000MB_RESTART_OFF);
wr                369 drivers/media/dvb-frontends/dib3000mb.c 	wr(DIB3000MB_REG_POWER_CONTROL, DIB3000MB_POWER_UP);
wr                371 drivers/media/dvb-frontends/dib3000mb.c 	wr(DIB3000MB_REG_RESTART, DIB3000MB_RESTART_AGC);
wr                373 drivers/media/dvb-frontends/dib3000mb.c 	wr(DIB3000MB_REG_RESET_DEVICE, DIB3000MB_RESET_DEVICE);
wr                374 drivers/media/dvb-frontends/dib3000mb.c 	wr(DIB3000MB_REG_RESET_DEVICE, DIB3000MB_RESET_DEVICE_RST);
wr                376 drivers/media/dvb-frontends/dib3000mb.c 	wr(DIB3000MB_REG_CLOCK, DIB3000MB_CLOCK_DEFAULT);
wr                378 drivers/media/dvb-frontends/dib3000mb.c 	wr(DIB3000MB_REG_ELECT_OUT_MODE, DIB3000MB_ELECT_OUT_MODE_ON);
wr                380 drivers/media/dvb-frontends/dib3000mb.c 	wr(DIB3000MB_REG_DDS_FREQ_MSB, DIB3000MB_DDS_FREQ_MSB);
wr                381 drivers/media/dvb-frontends/dib3000mb.c 	wr(DIB3000MB_REG_DDS_FREQ_LSB, DIB3000MB_DDS_FREQ_LSB);
wr                390 drivers/media/dvb-frontends/dib3000mb.c 	wr(DIB3000MB_REG_PHASE_NOISE, DIB3000MB_PHASE_NOISE_DEFAULT);
wr                398 drivers/media/dvb-frontends/dib3000mb.c 	wr(DIB3000MB_REG_LOCK0_MASK, DIB3000MB_LOCK0_DEFAULT);
wr                399 drivers/media/dvb-frontends/dib3000mb.c 	wr(DIB3000MB_REG_LOCK1_MASK, DIB3000MB_LOCK1_SEARCH_4);
wr                400 drivers/media/dvb-frontends/dib3000mb.c 	wr(DIB3000MB_REG_LOCK2_MASK, DIB3000MB_LOCK2_DEFAULT);
wr                401 drivers/media/dvb-frontends/dib3000mb.c 	wr(DIB3000MB_REG_SEQ, dib3000_seq[1][1][1]);
wr                405 drivers/media/dvb-frontends/dib3000mb.c 	wr(DIB3000MB_REG_UNK_68, DIB3000MB_UNK_68);
wr                406 drivers/media/dvb-frontends/dib3000mb.c 	wr(DIB3000MB_REG_UNK_69, DIB3000MB_UNK_69);
wr                407 drivers/media/dvb-frontends/dib3000mb.c 	wr(DIB3000MB_REG_UNK_71, DIB3000MB_UNK_71);
wr                408 drivers/media/dvb-frontends/dib3000mb.c 	wr(DIB3000MB_REG_UNK_77, DIB3000MB_UNK_77);
wr                409 drivers/media/dvb-frontends/dib3000mb.c 	wr(DIB3000MB_REG_UNK_78, DIB3000MB_UNK_78);
wr                410 drivers/media/dvb-frontends/dib3000mb.c 	wr(DIB3000MB_REG_ISI, DIB3000MB_ISI_INHIBIT);
wr                411 drivers/media/dvb-frontends/dib3000mb.c 	wr(DIB3000MB_REG_UNK_92, DIB3000MB_UNK_92);
wr                412 drivers/media/dvb-frontends/dib3000mb.c 	wr(DIB3000MB_REG_UNK_96, DIB3000MB_UNK_96);
wr                413 drivers/media/dvb-frontends/dib3000mb.c 	wr(DIB3000MB_REG_UNK_97, DIB3000MB_UNK_97);
wr                414 drivers/media/dvb-frontends/dib3000mb.c 	wr(DIB3000MB_REG_UNK_106, DIB3000MB_UNK_106);
wr                415 drivers/media/dvb-frontends/dib3000mb.c 	wr(DIB3000MB_REG_UNK_107, DIB3000MB_UNK_107);
wr                416 drivers/media/dvb-frontends/dib3000mb.c 	wr(DIB3000MB_REG_UNK_108, DIB3000MB_UNK_108);
wr                417 drivers/media/dvb-frontends/dib3000mb.c 	wr(DIB3000MB_REG_UNK_122, DIB3000MB_UNK_122);
wr                418 drivers/media/dvb-frontends/dib3000mb.c 	wr(DIB3000MB_REG_MOBILE_MODE_QAM, DIB3000MB_MOBILE_MODE_QAM_OFF);
wr                419 drivers/media/dvb-frontends/dib3000mb.c 	wr(DIB3000MB_REG_BERLEN, DIB3000MB_BERLEN_DEFAULT);
wr                423 drivers/media/dvb-frontends/dib3000mb.c 	wr(DIB3000MB_REG_MOBILE_ALGO, DIB3000MB_MOBILE_ALGO_ON);
wr                424 drivers/media/dvb-frontends/dib3000mb.c 	wr(DIB3000MB_REG_MULTI_DEMOD_MSB, DIB3000MB_MULTI_DEMOD_MSB);
wr                425 drivers/media/dvb-frontends/dib3000mb.c 	wr(DIB3000MB_REG_MULTI_DEMOD_LSB, DIB3000MB_MULTI_DEMOD_LSB);
wr                427 drivers/media/dvb-frontends/dib3000mb.c 	wr(DIB3000MB_REG_OUTPUT_MODE, DIB3000MB_OUTPUT_MODE_SLAVE);
wr                429 drivers/media/dvb-frontends/dib3000mb.c 	wr(DIB3000MB_REG_FIFO_142, DIB3000MB_FIFO_142);
wr                430 drivers/media/dvb-frontends/dib3000mb.c 	wr(DIB3000MB_REG_MPEG2_OUT_MODE, DIB3000MB_MPEG2_OUT_MODE_188);
wr                431 drivers/media/dvb-frontends/dib3000mb.c 	wr(DIB3000MB_REG_PID_PARSE, DIB3000MB_PID_PARSE_ACTIVATE);
wr                432 drivers/media/dvb-frontends/dib3000mb.c 	wr(DIB3000MB_REG_FIFO, DIB3000MB_FIFO_INHIBIT);
wr                433 drivers/media/dvb-frontends/dib3000mb.c 	wr(DIB3000MB_REG_FIFO_146, DIB3000MB_FIFO_146);
wr                434 drivers/media/dvb-frontends/dib3000mb.c 	wr(DIB3000MB_REG_FIFO_147, DIB3000MB_FIFO_147);
wr                436 drivers/media/dvb-frontends/dib3000mb.c 	wr(DIB3000MB_REG_DATA_IN_DIVERSITY, DIB3000MB_DATA_DIVERSITY_IN_OFF);
wr                674 drivers/media/dvb-frontends/dib3000mb.c 	wr(DIB3000MB_REG_POWER_CONTROL, DIB3000MB_POWER_DOWN);
wr                705 drivers/media/dvb-frontends/dib3000mb.c 	wr(index+DIB3000MB_REG_FIRST_PID,pid);
wr                715 drivers/media/dvb-frontends/dib3000mb.c 		wr(DIB3000MB_REG_FIFO, DIB3000MB_FIFO_ACTIVATE);
wr                717 drivers/media/dvb-frontends/dib3000mb.c 		wr(DIB3000MB_REG_FIFO, DIB3000MB_FIFO_INHIBIT);
wr                726 drivers/media/dvb-frontends/dib3000mb.c 	wr(DIB3000MB_REG_PID_PARSE,onoff);
wr                734 drivers/media/dvb-frontends/dib3000mb.c 		wr(DIB3000MB_REG_TUNER, DIB3000_TUNER_WRITE_ENABLE(pll_addr));
wr                736 drivers/media/dvb-frontends/dib3000mb.c 		wr(DIB3000MB_REG_TUNER, DIB3000_TUNER_WRITE_DISABLE(pll_addr));
wr                 23 drivers/media/dvb-frontends/dib3000mb_priv.h 		wr(a[i],v[i]); \
wr                 26 drivers/media/dvb-frontends/dib3000mb_priv.h #define set_or(reg,val) wr(reg,rd(reg) | val)
wr                 28 drivers/media/dvb-frontends/dib3000mb_priv.h #define set_and(reg,val) wr(reg,rd(reg) & val)
wr                 95 drivers/media/dvb-frontends/lgdt330x.c 	u8 wr[] = { reg };
wr                100 drivers/media/dvb-frontends/lgdt330x.c 			.buf = wr,
wr                 61 drivers/media/pci/solo6x10/solo6x10-i2c.c static void solo_i2c_flush(struct solo_dev *solo_dev, int wr)
wr                 70 drivers/media/pci/solo6x10/solo6x10-i2c.c 	if (wr) {
wr                 28 drivers/media/pci/solo6x10/solo6x10-p2m.c int solo_p2m_dma(struct solo_dev *solo_dev, int wr,
wr                 41 drivers/media/pci/solo6x10/solo6x10-p2m.c 				  wr ? PCI_DMA_TODEVICE : PCI_DMA_FROMDEVICE);
wr                 45 drivers/media/pci/solo6x10/solo6x10-p2m.c 	ret = solo_p2m_dma_t(solo_dev, wr, dma_addr, ext_addr, size,
wr                 49 drivers/media/pci/solo6x10/solo6x10-p2m.c 			 wr ? PCI_DMA_TODEVICE : PCI_DMA_FROMDEVICE);
wr                124 drivers/media/pci/solo6x10/solo6x10-p2m.c void solo_p2m_fill_desc(struct solo_p2m_desc *desc, int wr,
wr                133 drivers/media/pci/solo6x10/solo6x10-p2m.c 		(wr ? SOLO_P2M_WRITE : 0) | SOLO_P2M_TRANS_ON;
wr                145 drivers/media/pci/solo6x10/solo6x10-p2m.c int solo_p2m_dma_t(struct solo_dev *solo_dev, int wr,
wr                151 drivers/media/pci/solo6x10/solo6x10-p2m.c 	solo_p2m_fill_desc(&desc[1], wr, dma_addr, ext_addr, size, repeat,
wr                341 drivers/media/pci/solo6x10/solo6x10.h int solo_p2m_dma_t(struct solo_dev *solo_dev, int wr,
wr                344 drivers/media/pci/solo6x10/solo6x10.h int solo_p2m_dma(struct solo_dev *solo_dev, int wr,
wr                347 drivers/media/pci/solo6x10/solo6x10.h void solo_p2m_fill_desc(struct solo_p2m_desc *desc, int wr,
wr                162 drivers/media/radio/radio-tea5764.c 	struct tea5764_write_regs wr;
wr                167 drivers/media/radio/radio-tea5764.c 			.len = sizeof(wr),
wr                168 drivers/media/radio/radio-tea5764.c 			.buf = (void *)&wr
wr                171 drivers/media/radio/radio-tea5764.c 	wr.intreg  = r->intreg & 0xff;
wr                172 drivers/media/radio/radio-tea5764.c 	wr.frqset  = __cpu_to_be16(r->frqset);
wr                173 drivers/media/radio/radio-tea5764.c 	wr.tnctrl  = __cpu_to_be16(r->tnctrl);
wr                174 drivers/media/radio/radio-tea5764.c 	wr.testreg = __cpu_to_be16(r->testreg);
wr                175 drivers/media/radio/radio-tea5764.c 	wr.rdsctrl = __cpu_to_be16(r->rdsctrl);
wr                176 drivers/media/radio/radio-tea5764.c 	wr.rdsbbl  = __cpu_to_be16(r->rdsbbl);
wr                394 drivers/mtd/nand/raw/mpc5121_nfc.c 						u8 *buffer, uint size, int wr)
wr                431 drivers/mtd/nand/raw/mpc5121_nfc.c 		if (wr)
wr                446 drivers/mtd/nand/raw/mpc5121_nfc.c 									int wr)
wr                460 drivers/mtd/nand/raw/mpc5121_nfc.c 		mpc5121_nfc_copy_spare(mtd, c, buf, len, wr);
wr                471 drivers/mtd/nand/raw/mpc5121_nfc.c 	if (wr)
wr                480 drivers/mtd/nand/raw/mpc5121_nfc.c 		mpc5121_nfc_buf_copy(mtd, buf, len, wr);
wr               1171 drivers/net/can/grcan.c 	u32 wr, rd, startrd;
wr               1180 drivers/net/can/grcan.c 		wr = grcan_read_reg(&regs->rxwr);
wr               1181 drivers/net/can/grcan.c 		if (rd == wr)
wr                257 drivers/net/ethernet/apm/xgene/xgene_enet_hw.c 	void __iomem *addr, *wr, *cmd, *cmd_done;
wr                270 drivers/net/ethernet/apm/xgene/xgene_enet_hw.c 	wr = pdata->mcx_mac_addr + MAC_WRITE_REG_OFFSET;
wr                276 drivers/net/ethernet/apm/xgene/xgene_enet_hw.c 	iowrite32(wr_data, wr);
wr                 39 drivers/net/ethernet/apm/xgene/xgene_enet_xgmac.c static bool xgene_enet_wr_indirect(void __iomem *addr, void __iomem *wr,
wr                 47 drivers/net/ethernet/apm/xgene/xgene_enet_xgmac.c 	iowrite32(wr_data, wr);
wr                 65 drivers/net/ethernet/apm/xgene/xgene_enet_xgmac.c 	void __iomem *addr, *wr, *cmd, *cmd_done;
wr                 68 drivers/net/ethernet/apm/xgene/xgene_enet_xgmac.c 	wr = pdata->pcs_addr + PCS_WRITE_REG_OFFSET;
wr                 72 drivers/net/ethernet/apm/xgene/xgene_enet_xgmac.c 	if (!xgene_enet_wr_indirect(addr, wr, cmd, cmd_done, wr_addr, wr_data))
wr                475 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c 		req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_FORWARD));
wr                498 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c 		req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_FORWARD));
wr                520 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c 		req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_FORWARD));
wr                539 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c 	greq->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_FORWARD));
wr                913 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c 	req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_FORWARD));
wr                556 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c 	req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_FORWARD));
wr                878 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c 		rpl->wr.wr_hi =
wr                880 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c 		rpl->wr.wr_lo = htonl(V_WR_TID(tid));
wr               1100 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c 	req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_FORWARD));
wr                100 drivers/net/ethernet/chelsio/cxgb3/l2t.c 	req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_FORWARD));
wr               1226 drivers/net/ethernet/chelsio/cxgb3/sge.c 			cpl->wr.wr_hi = htonl(V_WR_BCNTLFLT(skb->len & 7) |
wr               1230 drivers/net/ethernet/chelsio/cxgb3/sge.c 			cpl->wr.wr_lo = htonl(V_WR_LEN(flits) | V_WR_GEN(gen) |
wr                300 drivers/net/ethernet/chelsio/cxgb3/t3_cpl.h # define WR_HDR struct work_request_hdr wr
wr               1751 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h void t4_mk_filtdelwr(unsigned int ftid, struct fw_filter_wr *wr, int qid);
wr                339 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 	u32 obq_wr_t4[2 * CIM_NUM_OBQ], *wr;
wr                352 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 			wr = obq_wr_t4;
wr                356 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 			wr = obq_wr_t5;
wr                372 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 	for ( ; i < CIM_NUM_IBQ + cim_num_obq; i++, p += 4, wr += 2)
wr                375 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 			   QUERDADDR_G(p[0]) & 0x3fff, wr[0] - base[i],
wr               1459 drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c 	struct work_request_hdr *wr;
wr               1487 drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c 	wrlen = roundup(sizeof(*wr) + (sizeof(*req) + sizeof(*aligner))
wr               1497 drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c 	wr = (struct work_request_hdr *)req;
wr               1498 drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c 	wr++;
wr               1499 drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c 	req = (struct cpl_set_tcb_field *)wr;
wr                 58 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h 	(w)->wr.wr_hi = htonl(FW_WR_OP_V(FW_TP_WR) | \
wr                 59 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h 			      FW_WR_IMMDLEN_V(sizeof(*w) - sizeof(w->wr))); \
wr                 60 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h 	(w)->wr.wr_mid = htonl(FW_WR_LEN16_V(DIV_ROUND_UP(sizeof(*w), 16)) | \
wr                 62 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h 	(w)->wr.wr_lo = cpu_to_be64(0); \
wr                 71 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h 	(w)->wr.wr_hi = htonl(FW_WR_OP_V(FW_ULPTX_WR) | \
wr                 73 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h 	(w)->wr.wr_mid = htonl(FW_WR_LEN16_V(DIV_ROUND_UP(wrlen, 16)) | \
wr                 75 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h 	(w)->wr.wr_lo = cpu_to_be64(0); \
wr               1004 drivers/net/ethernet/chelsio/cxgb4/sge.c 			u64 *wr = (u64 *)&q->desc[index];
wr               1008 drivers/net/ethernet/chelsio/cxgb4/sge.c 				      wr);
wr               1372 drivers/net/ethernet/chelsio/cxgb4/sge.c 	struct fw_eth_tx_pkt_wr *wr;
wr               1481 drivers/net/ethernet/chelsio/cxgb4/sge.c 	wr = (void *)&q->q.desc[q->q.pidx];
wr               1482 drivers/net/ethernet/chelsio/cxgb4/sge.c 	wr->equiq_to_len16 = htonl(wr_mid);
wr               1483 drivers/net/ethernet/chelsio/cxgb4/sge.c 	wr->r3 = cpu_to_be64(0);
wr               1484 drivers/net/ethernet/chelsio/cxgb4/sge.c 	end = (u64 *)wr + flits;
wr               1489 drivers/net/ethernet/chelsio/cxgb4/sge.c 		struct cpl_tx_pkt_lso_core *lso = (void *)(wr + 1);
wr               1493 drivers/net/ethernet/chelsio/cxgb4/sge.c 		struct cpl_tx_tnl_lso *tnl_lso = (void *)(wr + 1);
wr               1500 drivers/net/ethernet/chelsio/cxgb4/sge.c 		wr->op_immdlen = htonl(FW_WR_OP_V(FW_ETH_TX_PKT_WR) |
wr               1564 drivers/net/ethernet/chelsio/cxgb4/sge.c 		wr->op_immdlen = htonl(FW_WR_OP_V(op) |
wr               1566 drivers/net/ethernet/chelsio/cxgb4/sge.c 		cpl = (void *)(wr + 1);
wr               1709 drivers/net/ethernet/chelsio/cxgb4/sge.c 	struct fw_eth_tx_pkt_vm_wr *wr;
wr               1718 drivers/net/ethernet/chelsio/cxgb4/sge.c 	const size_t fw_hdr_copy_len = sizeof(wr->ethmacdst) +
wr               1719 drivers/net/ethernet/chelsio/cxgb4/sge.c 				       sizeof(wr->ethmacsrc) +
wr               1720 drivers/net/ethernet/chelsio/cxgb4/sge.c 				       sizeof(wr->ethtype) +
wr               1721 drivers/net/ethernet/chelsio/cxgb4/sge.c 				       sizeof(wr->vlantci);
wr               1801 drivers/net/ethernet/chelsio/cxgb4/sge.c 	wr = (void *)&txq->q.desc[txq->q.pidx];
wr               1802 drivers/net/ethernet/chelsio/cxgb4/sge.c 	wr->equiq_to_len16 = cpu_to_be32(wr_mid);
wr               1803 drivers/net/ethernet/chelsio/cxgb4/sge.c 	wr->r3[0] = cpu_to_be32(0);
wr               1804 drivers/net/ethernet/chelsio/cxgb4/sge.c 	wr->r3[1] = cpu_to_be32(0);
wr               1805 drivers/net/ethernet/chelsio/cxgb4/sge.c 	skb_copy_from_linear_data(skb, (void *)wr->ethmacdst, fw_hdr_copy_len);
wr               1806 drivers/net/ethernet/chelsio/cxgb4/sge.c 	end = (u64 *)wr + flits;
wr               1814 drivers/net/ethernet/chelsio/cxgb4/sge.c 		struct cpl_tx_pkt_lso_core *lso = (void *)(wr + 1);
wr               1819 drivers/net/ethernet/chelsio/cxgb4/sge.c 		wr->op_immdlen =
wr               1861 drivers/net/ethernet/chelsio/cxgb4/sge.c 		wr->op_immdlen =
wr               1868 drivers/net/ethernet/chelsio/cxgb4/sge.c 		cpl = (void *)(wr + 1);
wr               2033 drivers/net/ethernet/chelsio/cxgb4/sge.c static void ctrlq_check_stop(struct sge_ctrl_txq *q, struct fw_wr_hdr *wr)
wr               2037 drivers/net/ethernet/chelsio/cxgb4/sge.c 		wr->lo |= htonl(FW_WR_EQUEQ_F | FW_WR_EQUIQ_F);
wr               2054 drivers/net/ethernet/chelsio/cxgb4/sge.c 	struct fw_wr_hdr *wr;
wr               2072 drivers/net/ethernet/chelsio/cxgb4/sge.c 	wr = (struct fw_wr_hdr *)&q->q.desc[q->q.pidx];
wr               2073 drivers/net/ethernet/chelsio/cxgb4/sge.c 	cxgb4_inline_tx_skb(skb, &q->q, wr);
wr               2077 drivers/net/ethernet/chelsio/cxgb4/sge.c 		ctrlq_check_stop(q, wr);
wr               2103 drivers/net/ethernet/chelsio/cxgb4/sge.c 		struct fw_wr_hdr *wr;
wr               2110 drivers/net/ethernet/chelsio/cxgb4/sge.c 		wr = (struct fw_wr_hdr *)&q->q.desc[q->q.pidx];
wr               2114 drivers/net/ethernet/chelsio/cxgb4/sge.c 		cxgb4_inline_tx_skb(skb, &q->q, wr);
wr               2120 drivers/net/ethernet/chelsio/cxgb4/sge.c 			ctrlq_check_stop(q, wr);
wr               2221 drivers/net/ethernet/chelsio/cxgb4/sge.c static void ofldtxq_stop(struct sge_uld_txq *q, struct fw_wr_hdr *wr)
wr               2223 drivers/net/ethernet/chelsio/cxgb4/sge.c 	wr->lo |= htonl(FW_WR_EQUEQ_F | FW_WR_EQUIQ_F);
wr               6502 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c void t4_mk_filtdelwr(unsigned int ftid, struct fw_filter_wr *wr, int qid)
wr               6504 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 	memset(wr, 0, sizeof(*wr));
wr               6505 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 	wr->op_pkd = cpu_to_be32(FW_WR_OP_V(FW_FILTER_WR));
wr               6506 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 	wr->len16_pkd = cpu_to_be32(FW_WR_LEN16_V(sizeof(*wr) / 16));
wr               6507 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 	wr->tid_to_iq = cpu_to_be32(FW_FILTER_WR_TID_V(ftid) |
wr               6509 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 	wr->del_filter_to_l2tix = cpu_to_be32(FW_FILTER_WR_DEL_FILTER_F);
wr               6511 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 		wr->rx_chan_rx_rpl_iq =
wr                249 drivers/net/ethernet/chelsio/cxgb4/t4_msg.h #define WR_HDR struct work_request_hdr wr
wr               3571 drivers/net/ethernet/chelsio/cxgb4/t4fw_api.h 		} wr;
wr               1166 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	struct fw_eth_tx_pkt_vm_wr *wr;
wr               1170 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	const size_t fw_hdr_copy_len = (sizeof(wr->ethmacdst) +
wr               1171 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 					sizeof(wr->ethmacsrc) +
wr               1172 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 					sizeof(wr->ethtype) +
wr               1173 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 					sizeof(wr->vlantci));
wr               1266 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	wr = (void *)&txq->q.desc[txq->q.pidx];
wr               1267 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	wr->equiq_to_len16 = cpu_to_be32(wr_mid);
wr               1268 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	wr->r3[0] = cpu_to_be32(0);
wr               1269 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	wr->r3[1] = cpu_to_be32(0);
wr               1270 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	skb_copy_from_linear_data(skb, (void *)wr->ethmacdst, fw_hdr_copy_len);
wr               1271 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	end = (u64 *)wr + flits;
wr               1280 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		struct cpl_tx_pkt_lso_core *lso = (void *)(wr + 1);
wr               1285 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		wr->op_immdlen =
wr               1328 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		wr->op_immdlen =
wr               1336 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		cpl = (void *)(wr + 1);
wr                537 drivers/net/ethernet/netronome/nfp/bpf/jit.c __emit_lcsr(struct nfp_prog *nfp_prog, u16 areg, u16 breg, bool wr, u16 addr,
wr                545 drivers/net/ethernet/netronome/nfp/bpf/jit.c 		FIELD_PREP(OP_LCSR_WRITE, wr) |
wr               5892 drivers/net/ethernet/sun/niu.c 	u64 rd, wr, val;
wr               5896 drivers/net/ethernet/sun/niu.c 	wr = nr64_ipp(IPP_DFIFO_WR_PTR);
wr               5898 drivers/net/ethernet/sun/niu.c 	while (--limit >= 0 && (rd != wr)) {
wr               5900 drivers/net/ethernet/sun/niu.c 		wr = nr64_ipp(IPP_DFIFO_WR_PTR);
wr               5903 drivers/net/ethernet/sun/niu.c 	    (rd != 0 && wr != 1)) {
wr                444 drivers/net/hamradio/hdlcdrv.c 	s->hdlcrx.hbuf.rd = s->hdlcrx.hbuf.wr = 0;
wr                448 drivers/net/hamradio/hdlcdrv.c 	s->hdlctx.hbuf.rd = s->hdlctx.hbuf.wr = 0;
wr                576 drivers/net/hamradio/hdlcdrv.c 		if (s->bitbuf_channel.rd == s->bitbuf_channel.wr) 
wr                589 drivers/net/hamradio/hdlcdrv.c 		if (s->bitbuf_hdlc.rd == s->bitbuf_hdlc.wr) 
wr                641 drivers/net/hamradio/hdlcdrv.c 	s->hdlcrx.hbuf.rd = s->hdlcrx.hbuf.wr = 0;
wr                646 drivers/net/hamradio/hdlcdrv.c 	s->hdlctx.hbuf.rd = s->hdlctx.hbuf.wr = 0;
wr                656 drivers/net/hamradio/hdlcdrv.c 	s->bitbuf_channel.rd = s->bitbuf_channel.wr = 0;
wr                659 drivers/net/hamradio/hdlcdrv.c 	s->bitbuf_hdlc.rd = s->bitbuf_hdlc.wr = 0;
wr                722 drivers/net/hamradio/scc.c 	wr(scc,R12,tc & 255);		/* brg rate LOW */
wr                723 drivers/net/hamradio/scc.c 	wr(scc,R13,tc >> 8);   		/* brg rate HIGH */
wr                743 drivers/net/hamradio/scc.c 	wr(scc, R14, BRSRC);				/* BRG source = PCLK */
wr                800 drivers/net/hamradio/scc.c 	wr(scc,R4,X1CLK|SDLC);		/* *1 clock, SDLC mode */
wr                801 drivers/net/hamradio/scc.c 	wr(scc,R1,0);			/* no W/REQ operation */
wr                802 drivers/net/hamradio/scc.c 	wr(scc,R3,Rx8|RxCRC_ENAB);	/* RX 8 bits/char, CRC, disabled */	
wr                803 drivers/net/hamradio/scc.c 	wr(scc,R5,Tx8|DTR|TxCRC_ENAB);	/* TX 8 bits/char, disabled, DTR */
wr                804 drivers/net/hamradio/scc.c 	wr(scc,R6,0);			/* SDLC address zero (not used) */
wr                805 drivers/net/hamradio/scc.c 	wr(scc,R7,FLAG);		/* SDLC flag value */
wr                806 drivers/net/hamradio/scc.c 	wr(scc,R9,VIS);			/* vector includes status */
wr                807 drivers/net/hamradio/scc.c 	wr(scc,R10,(scc->modem.nrz? NRZ : NRZI)|CRCPS|ABUNDER); /* abort on underrun, preset CRC generator, NRZ(I) */
wr                808 drivers/net/hamradio/scc.c 	wr(scc,R14, 0);
wr                840 drivers/net/hamradio/scc.c 			wr(scc, R11, RCDPLL|TCDPLL|TRxCOI|TRxCDP);
wr                845 drivers/net/hamradio/scc.c 			wr(scc, R11, ((scc->brand & BAYCOM)? TRxCDP : TRxCBR) | RCDPLL|TCRTxCP|TRxCOI);
wr                850 drivers/net/hamradio/scc.c 			wr(scc, R11, (scc->brand & BAYCOM)? RCTRxCP|TCRTxCP : RCRTxCP|TCTRxCP);
wr                861 drivers/net/hamradio/scc.c 		wr(scc,R7,AUTOEOM);
wr                872 drivers/net/hamradio/scc.c 	wr(scc,R15, BRKIE|TxUIE|(scc->kiss.softdcd? SYNCIE:DCDIE));
wr                930 drivers/net/hamradio/scc.c 			wr(scc, R11, RCDPLL|TCBR|TRxCOI|TRxCBR);
wr                946 drivers/net/hamradio/scc.c 			wr(scc, R11, RCDPLL|TCDPLL|TRxCOI|TRxCDP);
wr               1410 drivers/net/hamradio/scc.c 	wr(scc, R6, 0);
wr               1411 drivers/net/hamradio/scc.c 	wr(scc, R7, FLAG);
wr               1436 drivers/net/hamradio/scc.c 	wr(scc, R6, 0);
wr               1437 drivers/net/hamradio/scc.c 	wr(scc, R7, pattern);
wr               1501 drivers/net/hamradio/scc.c 		wr(scc, R2, chip*16);			/* interrupt vector */
wr               1502 drivers/net/hamradio/scc.c 		wr(scc, R9, VIS);			/* vector includes status */
wr               1606 drivers/net/hamradio/scc.c 	wr(scc,R1,0);			/* disable interrupts */
wr               1607 drivers/net/hamradio/scc.c 	wr(scc,R3,0);
wr                 55 drivers/net/phy/mdio-xgene.c 	void __iomem *addr, *wr, *cmd, *cmd_done;
wr                 60 drivers/net/phy/mdio-xgene.c 	wr = pdata->mac_csr_addr + MAC_WRITE_REG_OFFSET;
wr                 66 drivers/net/phy/mdio-xgene.c 	iowrite32(data, wr);
wr                 12 drivers/net/wireless/mediatek/mt76/debugfs.c 	dev->bus->wr(dev, dev->debugfs_reg, val);
wr                 86 drivers/net/wireless/mediatek/mt76/mmio.c 		.wr = mt76_mmio_wr,
wr                 39 drivers/net/wireless/mediatek/mt76/mt76.h 	void (*wr)(struct mt76_dev *dev, u32 offset, u32 val);
wr                536 drivers/net/wireless/mediatek/mt76/mt76.h #define __mt76_wr(dev, ...)	(dev)->bus->wr((dev), __VA_ARGS__)
wr                545 drivers/net/wireless/mediatek/mt76/mt76.h #define mt76_wr(dev, ...)	(dev)->mt76.bus->wr(&((dev)->mt76), __VA_ARGS__)
wr                 59 drivers/net/wireless/mediatek/mt76/mt7603/core.c 	dev->bus_ops->wr(&dev->mt76, MT_MCU_PCIE_REMAP_2, base);
wr                429 drivers/net/wireless/mediatek/mt76/mt7603/init.c 	dev->bus_ops->wr(mdev, addr, val);
wr                523 drivers/net/wireless/mediatek/mt76/mt7603/init.c 	bus_ops->wr = mt7603_wr;
wr                951 drivers/net/wireless/mediatek/mt76/usb.c 		.wr = mt76u_wr,
wr               1135 drivers/nvme/host/rdma.c 	struct ib_send_wr wr = {
wr               1144 drivers/nvme/host/rdma.c 	wr.wr_cqe = &req->reg_cqe;
wr               1146 drivers/nvme/host/rdma.c 	return ib_post_send(queue->qp, &wr, NULL);
wr               1246 drivers/nvme/host/rdma.c 	req->reg_wr.wr.opcode = IB_WR_REG_MR;
wr               1247 drivers/nvme/host/rdma.c 	req->reg_wr.wr.wr_cqe = &req->reg_cqe;
wr               1248 drivers/nvme/host/rdma.c 	req->reg_wr.wr.num_sge = 0;
wr               1346 drivers/nvme/host/rdma.c 	struct ib_send_wr wr;
wr               1353 drivers/nvme/host/rdma.c 	wr.next       = NULL;
wr               1354 drivers/nvme/host/rdma.c 	wr.wr_cqe     = &qe->cqe;
wr               1355 drivers/nvme/host/rdma.c 	wr.sg_list    = sge;
wr               1356 drivers/nvme/host/rdma.c 	wr.num_sge    = num_sge;
wr               1357 drivers/nvme/host/rdma.c 	wr.opcode     = IB_WR_SEND;
wr               1358 drivers/nvme/host/rdma.c 	wr.send_flags = IB_SEND_SIGNALED;
wr               1361 drivers/nvme/host/rdma.c 		first->next = &wr;
wr               1363 drivers/nvme/host/rdma.c 		first = &wr;
wr               1376 drivers/nvme/host/rdma.c 	struct ib_recv_wr wr;
wr               1386 drivers/nvme/host/rdma.c 	wr.next     = NULL;
wr               1387 drivers/nvme/host/rdma.c 	wr.wr_cqe   = &qe->cqe;
wr               1388 drivers/nvme/host/rdma.c 	wr.sg_list  = &list;
wr               1389 drivers/nvme/host/rdma.c 	wr.num_sge  = 1;
wr               1391 drivers/nvme/host/rdma.c 	ret = ib_post_recv(queue->qp, &wr, NULL);
wr               1783 drivers/nvme/host/rdma.c 			req->mr ? &req->reg_wr.wr : NULL);
wr                 37 drivers/nvme/target/rdma.c 	struct ib_recv_wr	wr;
wr                300 drivers/nvme/target/rdma.c 	c->wr.wr_cqe = &c->cqe;
wr                301 drivers/nvme/target/rdma.c 	c->wr.sg_list = c->sge;
wr                302 drivers/nvme/target/rdma.c 	c->wr.num_sge = admin ? 1 : ndev->inline_page_count + 1;
wr                465 drivers/nvme/target/rdma.c 		ret = ib_post_srq_recv(ndev->srq, &cmd->wr, NULL);
wr                467 drivers/nvme/target/rdma.c 		ret = ib_post_recv(cmd->queue->cm_id->qp, &cmd->wr, NULL);
wr                136 drivers/perf/fsl_imx8_ddr_perf.c 	IMX8_DDR_PMU_EVENT_ATTR(wr-xact-credit, 0x29),
wr                345 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(wr-cmd-sent,			0x06),
wr                352 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(wr-wra-cmd-sent,			0x0d),
wr                354 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(in-wr-collision,			0x0f),
wr                360 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(mcu-wr-request,			0x15),
wr                366 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(mcu-wr-proceed-all,		0x1b),
wr                367 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(mcu-wr-proceed-cancel,		0x1c),
wr                438 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(pa-req-buf-alloc-wr,		0x03),
wr                443 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(pa-cp-wr-req,			0x08),
wr                446 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(ba-wr-req,				0x0b),
wr                449 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(pa-wr-invalidate-req-issued-stashable, 0x12),
wr                450 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(pa-wr-invalidate-req-issued-nonstashable, 0x13),
wr                451 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(pa-wr-back-req-issued-stashable,	0x14),
wr                452 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(pa-wr-back-req-issued-nonstashable, 0x15),
wr                453 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(pa-ptl-wr-req,			0x16),
wr                455 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(pa-wr-back-clean-data,		0x18),
wr                456 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(pa-wr-back-cancelled-on-SS,	0x1b),
wr                464 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(pa-rd-wr-invalid-snoop,		0x25),
wr                465 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(pa-rd-wr-invalid-snoop-hit,	0x26),
wr                480 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(pa-axi0-wr-req,			0x02),
wr                482 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(pa-axi1-wr-req,			0x04),
wr                485 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(ba-axi-wr-req,			0x09),
wr                495 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(wr-req-recv,			0x04),
wr                496 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(wr-req-recv-2,			0x05),
wr                506 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(wr-req-sent-to-mcu,		0x0f),
wr                509 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(wr-gack-recv,			0x12),
wr                511 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(cancel-wr-gack,			0x14),
wr                519 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(mcu-wr-table-full,			0x1c),
wr                536 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(wr-sent,				0x05),
wr                543 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(wr-wra-sent,			0x0c),
wr                549 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(hprd-lprd-wr-req-vld,		0x12),
wr                553 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(wr-req-vld,			0x16),
wr                554 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(partial-wr-req-vld,		0x17),
wr                556 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(wr-retry,				0x19),
wr                573 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(wr-enq,				0x2a),
wr               3933 drivers/scsi/csiostor/csio_hw.c csio_process_fwevtq_entry(struct csio_hw *hw, void *wr, uint32_t len,
wr               3941 drivers/scsi/csiostor/csio_hw.c 	op = ((struct rss_header *) wr)->opcode;
wr               3956 drivers/scsi/csiostor/csio_hw.c 		msg = (void *)((uintptr_t)wr + sizeof(__be64));
wr                131 drivers/scsi/csiostor/csio_isr.c csio_fwevt_intx_handler(struct csio_hw *hw, void *wr, uint32_t len,
wr                146 drivers/scsi/csiostor/csio_isr.c csio_process_scsi_cmpl(struct csio_hw *hw, void *wr, uint32_t len,
wr                155 drivers/scsi/csiostor/csio_isr.c 	ioreq = csio_scsi_cmpl_handler(hw, wr, len, flb, NULL, &scsiwr);
wr                285 drivers/scsi/csiostor/csio_isr.c csio_scsi_intx_handler(struct csio_hw *hw, void *wr, uint32_t len,
wr               1433 drivers/scsi/csiostor/csio_lnode.c csio_ln_mgmt_wr_handler(struct csio_hw *hw, void *wr, uint32_t len)
wr               1440 drivers/scsi/csiostor/csio_lnode.c 	wr_cmd = (struct fw_fcoe_els_ct_wr *) wr;
wr               1490 drivers/scsi/csiostor/csio_lnode.c 	struct fw_wr_hdr *wr;
wr               1526 drivers/scsi/csiostor/csio_lnode.c 		wr = (struct fw_wr_hdr *) (cmd + 4);
wr               1527 drivers/scsi/csiostor/csio_lnode.c 		if (FW_WR_OP_G(be32_to_cpu(wr->hi))
wr               1595 drivers/scsi/csiostor/csio_lnode.c 				  FW_WR_OP_G(be32_to_cpu((wr->hi))));
wr               1599 drivers/scsi/csiostor/csio_lnode.c 		wr = (struct fw_wr_hdr *) (cmd);
wr               1600 drivers/scsi/csiostor/csio_lnode.c 		if (FW_WR_OP_G(be32_to_cpu(wr->hi)) == FW_FCOE_ELS_CT_WR) {
wr               1601 drivers/scsi/csiostor/csio_lnode.c 			csio_ln_mgmt_wr_handler(hw, wr,
wr               1605 drivers/scsi/csiostor/csio_lnode.c 				  FW_WR_OP_G(be32_to_cpu((wr->hi))));
wr               1685 drivers/scsi/csiostor/csio_lnode.c 	struct fw_fcoe_els_ct_wr *wr;
wr               1688 drivers/scsi/csiostor/csio_lnode.c 	wr  = (struct fw_fcoe_els_ct_wr *)fw_wr;
wr               1689 drivers/scsi/csiostor/csio_lnode.c 	wr->op_immdlen = cpu_to_be32(FW_WR_OP_V(FW_FCOE_ELS_CT_WR) |
wr               1693 drivers/scsi/csiostor/csio_lnode.c 	wr->flowid_len16 = cpu_to_be32(FW_WR_FLOWID_V(flow_id) |
wr               1695 drivers/scsi/csiostor/csio_lnode.c 	wr->els_ct_type = sub_op;
wr               1696 drivers/scsi/csiostor/csio_lnode.c 	wr->ctl_pri = 0;
wr               1697 drivers/scsi/csiostor/csio_lnode.c 	wr->cp_en_class = 0;
wr               1698 drivers/scsi/csiostor/csio_lnode.c 	wr->cookie = io_req->fw_handle;
wr               1699 drivers/scsi/csiostor/csio_lnode.c 	wr->iqid = cpu_to_be16(csio_q_physiqid(
wr               1701 drivers/scsi/csiostor/csio_lnode.c 	wr->fl_to_sp =  FW_FCOE_ELS_CT_WR_SP(1);
wr               1702 drivers/scsi/csiostor/csio_lnode.c 	wr->tmo_val = (uint8_t) io_req->tmo;
wr               1704 drivers/scsi/csiostor/csio_lnode.c 	memcpy(wr->l_id, PORT_ID_PTR(port_id), 3);
wr               1706 drivers/scsi/csiostor/csio_lnode.c 	memcpy(wr->r_id, PORT_ID_PTR(port_id), 3);
wr               1709 drivers/scsi/csiostor/csio_lnode.c 	wr->rsp_dmalen = cpu_to_be32(io_req->dma_buf.len);
wr               1710 drivers/scsi/csiostor/csio_lnode.c 	wr->rsp_dmaaddr = cpu_to_be64(io_req->dma_buf.paddr);
wr                195 drivers/scsi/csiostor/csio_mb.c 	       const u32 *params, u32 *val, bool wr,
wr                207 drivers/scsi/csiostor/csio_mb.c 				(wr ? FW_CMD_WRITE_F : FW_CMD_READ_F)	|
wr                213 drivers/scsi/csiostor/csio_mb.c 	if (wr) {
wr                301 drivers/scsi/csiostor/csio_mb.c 		    bool wr, bool init, bool tgt, bool cofld,
wr                307 drivers/scsi/csiostor/csio_mb.c 	CSIO_INIT_MBP(mbp, cmdp, tmo, hw, cbfn, wr ? 0 : 1);
wr                311 drivers/scsi/csiostor/csio_mb.c 				  (wr ? FW_CMD_WRITE_F : FW_CMD_READ_F));
wr                315 drivers/scsi/csiostor/csio_mb.c 	if (!wr)
wr                343 drivers/scsi/csiostor/csio_mb.c 	     u8 portid, bool wr, uint32_t fc, uint16_t fw_caps,
wr                352 drivers/scsi/csiostor/csio_mb.c 				   (wr ? FW_CMD_EXEC_F : FW_CMD_READ_F)	|
wr                354 drivers/scsi/csiostor/csio_mb.c 	if (!wr) {
wr                205 drivers/scsi/csiostor/csio_scsi.c 	struct fw_scsi_cmd_wr *wr = (struct fw_scsi_cmd_wr *)addr;
wr                209 drivers/scsi/csiostor/csio_scsi.c 	wr->op_immdlen = cpu_to_be32(FW_WR_OP_V(FW_SCSI_CMD_WR) |
wr                211 drivers/scsi/csiostor/csio_scsi.c 	wr->flowid_len16 = cpu_to_be32(FW_WR_FLOWID_V(rn->flowid) |
wr                215 drivers/scsi/csiostor/csio_scsi.c 	wr->cookie = (uintptr_t) req;
wr                216 drivers/scsi/csiostor/csio_scsi.c 	wr->iqid = cpu_to_be16(csio_q_physiqid(hw, req->iq_idx));
wr                217 drivers/scsi/csiostor/csio_scsi.c 	wr->tmo_val = (uint8_t) req->tmo;
wr                218 drivers/scsi/csiostor/csio_scsi.c 	wr->r3 = 0;
wr                219 drivers/scsi/csiostor/csio_scsi.c 	memset(&wr->r5, 0, 8);
wr                225 drivers/scsi/csiostor/csio_scsi.c 	wr->rsp_dmalen = cpu_to_be32(dma_buf->len);
wr                226 drivers/scsi/csiostor/csio_scsi.c 	wr->rsp_dmaaddr = cpu_to_be64(dma_buf->paddr);
wr                228 drivers/scsi/csiostor/csio_scsi.c 	wr->r6 = 0;
wr                230 drivers/scsi/csiostor/csio_scsi.c 	wr->u.fcoe.ctl_pri = 0;
wr                231 drivers/scsi/csiostor/csio_scsi.c 	wr->u.fcoe.cp_en_class = 0;
wr                232 drivers/scsi/csiostor/csio_scsi.c 	wr->u.fcoe.r4_lo[0] = 0;
wr                233 drivers/scsi/csiostor/csio_scsi.c 	wr->u.fcoe.r4_lo[1] = 0;
wr                364 drivers/scsi/csiostor/csio_scsi.c 	struct fw_scsi_read_wr *wr = (struct fw_scsi_read_wr *)wrp;
wr                370 drivers/scsi/csiostor/csio_scsi.c 	wr->op_immdlen = cpu_to_be32(FW_WR_OP_V(FW_SCSI_READ_WR) |
wr                372 drivers/scsi/csiostor/csio_scsi.c 	wr->flowid_len16 = cpu_to_be32(FW_WR_FLOWID_V(rn->flowid) |
wr                374 drivers/scsi/csiostor/csio_scsi.c 	wr->cookie = (uintptr_t)req;
wr                375 drivers/scsi/csiostor/csio_scsi.c 	wr->iqid = cpu_to_be16(csio_q_physiqid(hw, req->iq_idx));
wr                376 drivers/scsi/csiostor/csio_scsi.c 	wr->tmo_val = (uint8_t)(req->tmo);
wr                377 drivers/scsi/csiostor/csio_scsi.c 	wr->use_xfer_cnt = 1;
wr                378 drivers/scsi/csiostor/csio_scsi.c 	wr->xfer_cnt = cpu_to_be32(scsi_bufflen(scmnd));
wr                379 drivers/scsi/csiostor/csio_scsi.c 	wr->ini_xfer_cnt = cpu_to_be32(scsi_bufflen(scmnd));
wr                384 drivers/scsi/csiostor/csio_scsi.c 	wr->rsp_dmalen = cpu_to_be32(dma_buf->len);
wr                385 drivers/scsi/csiostor/csio_scsi.c 	wr->rsp_dmaaddr = cpu_to_be64(dma_buf->paddr);
wr                387 drivers/scsi/csiostor/csio_scsi.c 	wr->r4 = 0;
wr                389 drivers/scsi/csiostor/csio_scsi.c 	wr->u.fcoe.ctl_pri = 0;
wr                390 drivers/scsi/csiostor/csio_scsi.c 	wr->u.fcoe.cp_en_class = 0;
wr                391 drivers/scsi/csiostor/csio_scsi.c 	wr->u.fcoe.r3_lo[0] = 0;
wr                392 drivers/scsi/csiostor/csio_scsi.c 	wr->u.fcoe.r3_lo[1] = 0;
wr                417 drivers/scsi/csiostor/csio_scsi.c 	struct fw_scsi_write_wr *wr = (struct fw_scsi_write_wr *)wrp;
wr                423 drivers/scsi/csiostor/csio_scsi.c 	wr->op_immdlen = cpu_to_be32(FW_WR_OP_V(FW_SCSI_WRITE_WR) |
wr                425 drivers/scsi/csiostor/csio_scsi.c 	wr->flowid_len16 = cpu_to_be32(FW_WR_FLOWID_V(rn->flowid) |
wr                427 drivers/scsi/csiostor/csio_scsi.c 	wr->cookie = (uintptr_t)req;
wr                428 drivers/scsi/csiostor/csio_scsi.c 	wr->iqid = cpu_to_be16(csio_q_physiqid(hw, req->iq_idx));
wr                429 drivers/scsi/csiostor/csio_scsi.c 	wr->tmo_val = (uint8_t)(req->tmo);
wr                430 drivers/scsi/csiostor/csio_scsi.c 	wr->use_xfer_cnt = 1;
wr                431 drivers/scsi/csiostor/csio_scsi.c 	wr->xfer_cnt = cpu_to_be32(scsi_bufflen(scmnd));
wr                432 drivers/scsi/csiostor/csio_scsi.c 	wr->ini_xfer_cnt = cpu_to_be32(scsi_bufflen(scmnd));
wr                437 drivers/scsi/csiostor/csio_scsi.c 	wr->rsp_dmalen = cpu_to_be32(dma_buf->len);
wr                438 drivers/scsi/csiostor/csio_scsi.c 	wr->rsp_dmaaddr = cpu_to_be64(dma_buf->paddr);
wr                440 drivers/scsi/csiostor/csio_scsi.c 	wr->r4 = 0;
wr                442 drivers/scsi/csiostor/csio_scsi.c 	wr->u.fcoe.ctl_pri = 0;
wr                443 drivers/scsi/csiostor/csio_scsi.c 	wr->u.fcoe.cp_en_class = 0;
wr                444 drivers/scsi/csiostor/csio_scsi.c 	wr->u.fcoe.r3_lo[0] = 0;
wr                445 drivers/scsi/csiostor/csio_scsi.c 	wr->u.fcoe.r3_lo[1] = 0;
wr                651 drivers/scsi/csiostor/csio_scsi.c 	struct fw_scsi_abrt_cls_wr *wr = (struct fw_scsi_abrt_cls_wr *)addr;
wr                653 drivers/scsi/csiostor/csio_scsi.c 	wr->op_immdlen = cpu_to_be32(FW_WR_OP_V(FW_SCSI_ABRT_CLS_WR));
wr                654 drivers/scsi/csiostor/csio_scsi.c 	wr->flowid_len16 = cpu_to_be32(FW_WR_FLOWID_V(rn->flowid) |
wr                658 drivers/scsi/csiostor/csio_scsi.c 	wr->cookie = (uintptr_t) req;
wr                659 drivers/scsi/csiostor/csio_scsi.c 	wr->iqid = cpu_to_be16(csio_q_physiqid(hw, req->iq_idx));
wr                660 drivers/scsi/csiostor/csio_scsi.c 	wr->tmo_val = (uint8_t) req->tmo;
wr                662 drivers/scsi/csiostor/csio_scsi.c 	wr->sub_opcode_to_chk_all_io =
wr                665 drivers/scsi/csiostor/csio_scsi.c 	wr->r3[0] = 0;
wr                666 drivers/scsi/csiostor/csio_scsi.c 	wr->r3[1] = 0;
wr                667 drivers/scsi/csiostor/csio_scsi.c 	wr->r3[2] = 0;
wr                668 drivers/scsi/csiostor/csio_scsi.c 	wr->r3[3] = 0;
wr                670 drivers/scsi/csiostor/csio_scsi.c 	wr->t_cookie = (uintptr_t) req;
wr               1102 drivers/scsi/csiostor/csio_scsi.c csio_scsi_cmpl_handler(struct csio_hw *hw, void *wr, uint32_t len,
wr               1112 drivers/scsi/csiostor/csio_scsi.c 	cpl = (struct cpl_fw6_msg *)((uintptr_t)wr + sizeof(__be64));
wr                766 drivers/scsi/csiostor/csio_wr.c 	void *wr;
wr                775 drivers/scsi/csiostor/csio_wr.c 		wr = (void *)((uintptr_t)q->vstart +
wr                778 drivers/scsi/csiostor/csio_wr.c 		ftr = (struct csio_iqwr_footer *)((uintptr_t)wr +
wr               1042 drivers/scsi/csiostor/csio_wr.c 		   void *wr, uint32_t len_to_qid,
wr               1092 drivers/scsi/csiostor/csio_wr.c 	iq_handler(hw, wr, q->wr_sz - sizeof(struct csio_iqwr_footer),
wr               1135 drivers/scsi/csiostor/csio_wr.c 	void *wr = (void *)((uintptr_t)q->vstart + (q->cidx * q->wr_sz));
wr               1144 drivers/scsi/csiostor/csio_wr.c 	ftr = (struct csio_iqwr_footer *)((uintptr_t)wr +
wr               1153 drivers/scsi/csiostor/csio_wr.c 		CSIO_DB_ASSERT(((uintptr_t)wr + q->wr_sz) <=
wr               1161 drivers/scsi/csiostor/csio_wr.c 			iq_handler(hw, wr, q->wr_sz - sizeof(*ftr), NULL, priv);
wr               1164 drivers/scsi/csiostor/csio_wr.c 			csio_wr_process_fl(hw, q, wr,
wr               1203 drivers/scsi/csiostor/csio_wr.c 		if (((uintptr_t)wr + q->wr_sz) == (uintptr_t)q->vwrap) {
wr               1207 drivers/scsi/csiostor/csio_wr.c 			wr	= q->vstart;
wr               1215 drivers/scsi/csiostor/csio_wr.c 			wr	= (void *)((uintptr_t)(q->vstart) +
wr               1219 drivers/scsi/csiostor/csio_wr.c 		ftr = (struct csio_iqwr_footer *)((uintptr_t)wr +
wr                165 drivers/scsi/cxgbi/cxgb3i/cxgb3i.c 	req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_FORWARD));
wr                210 drivers/scsi/cxgbi/cxgb3i/cxgb3i.c 	req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_OFLD_CLOSE_CON));
wr                211 drivers/scsi/cxgbi/cxgb3i/cxgb3i.c 	req->wr.wr_lo = htonl(V_WR_TID(tid));
wr                254 drivers/scsi/cxgbi/cxgb3i/cxgb3i.c 	req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_OFLD_HOST_ABORT_CON_REQ));
wr                255 drivers/scsi/cxgbi/cxgb3i/cxgb3i.c 	req->wr.wr_lo = htonl(V_WR_TID(csk->tid));
wr                285 drivers/scsi/cxgbi/cxgb3i/cxgb3i.c 	rpl->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_OFLD_HOST_ABORT_CON_RPL));
wr                286 drivers/scsi/cxgbi/cxgb3i/cxgb3i.c 	rpl->wr.wr_lo = htonl(V_WR_TID(csk->tid));
wr                313 drivers/scsi/cxgbi/cxgb3i/cxgb3i.c 	req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_FORWARD));
wr               1075 drivers/scsi/cxgbi/cxgb3i/cxgb3i.c 	req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_BYPASS));
wr               1159 drivers/scsi/cxgbi/cxgb3i/cxgb3i.c 	req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_FORWARD));
wr               1195 drivers/scsi/cxgbi/cxgb3i/cxgb3i.c 	req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_FORWARD));
wr                784 drivers/scsi/cxgbi/cxgb4i/cxgb4i.c 			req->wr.wr_hi |= htonl(FW_WR_COMPL_F);
wr               1898 drivers/scsi/cxgbi/cxgb4i/cxgb4i.c 	req->wr.wr_hi = htonl(FW_WR_OP_V(FW_ULPTX_WR) |
wr               1905 drivers/scsi/cxgbi/cxgb4i/cxgb4i.c 	req->len16 = htonl(DIV_ROUND_UP(wr_len - sizeof(req->wr), 16));
wr               3698 drivers/scsi/qla2xxx/qla_mbx.c qla2x00_disable_fce_trace(scsi_qla_host_t *vha, uint64_t *wr, uint64_t *rd)
wr               3730 drivers/scsi/qla2xxx/qla_mbx.c 		if (wr)
wr               3731 drivers/scsi/qla2xxx/qla_mbx.c 			*wr = (uint64_t) mcp->mb[5] << 48 |
wr                510 drivers/spi/spi-bcm-qspi.c 	u32 wr = 0;
wr                514 drivers/spi/spi-bcm-qspi.c 		wr = (rd & ~0xff) | (1 << cs);
wr                515 drivers/spi/spi-bcm-qspi.c 		if (rd == wr)
wr                517 drivers/spi/spi-bcm-qspi.c 		bcm_qspi_write(qspi, CHIP_SELECT, 0, wr);
wr                 29 drivers/staging/fbtft/fb_agm1264k-fl.c #define EPIN		gpio.wr
wr                112 drivers/staging/fbtft/fbtft-core.c 	ret = fbtft_request_one_gpio(par, "wr", 0, &par->gpio.wr);
wr               1125 drivers/staging/fbtft/fbtft-core.c 	if (!par->gpio.wr) {
wr                145 drivers/staging/fbtft/fbtft-io.c 		gpiod_set_value(par->gpio.wr, 0);
wr                150 drivers/staging/fbtft/fbtft-io.c 			gpiod_set_value(par->gpio.wr, 0); /* used as delay */
wr                168 drivers/staging/fbtft/fbtft-io.c 		gpiod_set_value(par->gpio.wr, 1);
wr                195 drivers/staging/fbtft/fbtft-io.c 		gpiod_set_value(par->gpio.wr, 0);
wr                200 drivers/staging/fbtft/fbtft-io.c 			gpiod_set_value(par->gpio.wr, 0); /* used as delay */
wr                218 drivers/staging/fbtft/fbtft-io.c 		gpiod_set_value(par->gpio.wr, 1);
wr                212 drivers/staging/fbtft/fbtft.h 		struct gpio_desc *wr;
wr                 81 drivers/target/iscsi/cxgbit/cxgbit_ddp.c 	req->wr.wr_hi = htonl(FW_WR_OP_V(FW_ULPTX_WR) |
wr                 88 drivers/target/iscsi/cxgbit/cxgbit_ddp.c 	req->len16 = htonl(DIV_ROUND_UP(wr_len - sizeof(req->wr), 16));
wr                272 drivers/target/iscsi/cxgbit/cxgbit_target.c 			req->wr.wr_hi |= htonl(FW_WR_COMPL_F);
wr                 50 fs/autofs/waitq.c 	ssize_t wr = 0;
wr                 56 fs/autofs/waitq.c 		wr = __kernel_write(file, data, bytes, &file->f_pos);
wr                 57 fs/autofs/waitq.c 		if (wr <= 0)
wr                 59 fs/autofs/waitq.c 		data += wr;
wr                 60 fs/autofs/waitq.c 		bytes -= wr;
wr                 67 fs/autofs/waitq.c 	if (wr == -EPIPE && !sigpipe) {
wr                 75 fs/autofs/waitq.c 	return bytes == 0 ? 0 : wr < 0 ? wr : -EIO;
wr               2510 fs/cifs/smbdirect.c 	reg_wr = &smbdirect_mr->wr;
wr               2511 fs/cifs/smbdirect.c 	reg_wr->wr.opcode = IB_WR_REG_MR;
wr               2513 fs/cifs/smbdirect.c 	reg_wr->wr.wr_cqe = &smbdirect_mr->cqe;
wr               2514 fs/cifs/smbdirect.c 	reg_wr->wr.num_sge = 0;
wr               2515 fs/cifs/smbdirect.c 	reg_wr->wr.send_flags = IB_SEND_SIGNALED;
wr               2527 fs/cifs/smbdirect.c 	rc = ib_post_send(info->id->qp, &reg_wr->wr, NULL);
wr               2572 fs/cifs/smbdirect.c 	struct ib_send_wr *wr;
wr               2578 fs/cifs/smbdirect.c 		wr = &smbdirect_mr->inv_wr;
wr               2579 fs/cifs/smbdirect.c 		wr->opcode = IB_WR_LOCAL_INV;
wr               2581 fs/cifs/smbdirect.c 		wr->wr_cqe = &smbdirect_mr->cqe;
wr               2582 fs/cifs/smbdirect.c 		wr->num_sge = 0;
wr               2583 fs/cifs/smbdirect.c 		wr->ex.invalidate_rkey = smbdirect_mr->mr->rkey;
wr               2584 fs/cifs/smbdirect.c 		wr->send_flags = IB_SEND_SIGNALED;
wr               2587 fs/cifs/smbdirect.c 		rc = ib_post_send(info->id->qp, wr, NULL);
wr                297 fs/cifs/smbdirect.h 		struct ib_reg_wr	wr;
wr                 50 fs/orangefs/file.c     loff_t readahead_size, struct orangefs_write_range *wr, int *index_return)
wr                 89 fs/orangefs/file.c 	if (type == ORANGEFS_IO_WRITE && wr) {
wr                 90 fs/orangefs/file.c 		new_op->upcall.uid = from_kuid(&init_user_ns, wr->uid);
wr                 91 fs/orangefs/file.c 		new_op->upcall.gid = from_kgid(&init_user_ns, wr->gid);
wr                 22 fs/orangefs/inode.c 	struct orangefs_write_range *wr = NULL;
wr                 33 fs/orangefs/inode.c 		wr = (struct orangefs_write_range *)page_private(page);
wr                 34 fs/orangefs/inode.c 		WARN_ON(wr->pos >= len);
wr                 35 fs/orangefs/inode.c 		off = wr->pos;
wr                 36 fs/orangefs/inode.c 		if (off + wr->len > len)
wr                 39 fs/orangefs/inode.c 			wlen = wr->len;
wr                 58 fs/orangefs/inode.c 	    len, wr, NULL);
wr                 65 fs/orangefs/inode.c 	if (wr) {
wr                 66 fs/orangefs/inode.c 		kfree(wr);
wr                 98 fs/orangefs/inode.c 	struct orangefs_write_range *wrp, wr;
wr                126 fs/orangefs/inode.c 	wr.uid = ow->uid;
wr                127 fs/orangefs/inode.c 	wr.gid = ow->gid;
wr                129 fs/orangefs/inode.c 	    0, &wr, NULL);
wr                165 fs/orangefs/inode.c 	struct orangefs_write_range *wr;
wr                175 fs/orangefs/inode.c 	wr = (struct orangefs_write_range *)page_private(page);
wr                179 fs/orangefs/inode.c 		ow->off = wr->pos;
wr                180 fs/orangefs/inode.c 		ow->len = wr->len;
wr                181 fs/orangefs/inode.c 		ow->uid = wr->uid;
wr                182 fs/orangefs/inode.c 		ow->gid = wr->gid;
wr                187 fs/orangefs/inode.c 	if (!uid_eq(ow->uid, wr->uid) || !gid_eq(ow->gid, wr->gid)) {
wr                193 fs/orangefs/inode.c 	if (ow->off + ow->len == wr->pos) {
wr                194 fs/orangefs/inode.c 		ow->len += wr->len;
wr                393 fs/orangefs/inode.c 	struct orangefs_write_range *wr;
wr                417 fs/orangefs/inode.c 		struct orangefs_write_range *wr;
wr                418 fs/orangefs/inode.c 		wr = (struct orangefs_write_range *)page_private(page);
wr                419 fs/orangefs/inode.c 		if (wr->pos + wr->len == pos &&
wr                420 fs/orangefs/inode.c 		    uid_eq(wr->uid, current_fsuid()) &&
wr                421 fs/orangefs/inode.c 		    gid_eq(wr->gid, current_fsgid())) {
wr                422 fs/orangefs/inode.c 			wr->len += len;
wr                431 fs/orangefs/inode.c 	wr = kmalloc(sizeof *wr, GFP_KERNEL);
wr                432 fs/orangefs/inode.c 	if (!wr)
wr                435 fs/orangefs/inode.c 	wr->pos = pos;
wr                436 fs/orangefs/inode.c 	wr->len = len;
wr                437 fs/orangefs/inode.c 	wr->uid = current_fsuid();
wr                438 fs/orangefs/inode.c 	wr->gid = current_fsgid();
wr                440 fs/orangefs/inode.c 	set_page_private(page, (unsigned long)wr);
wr                485 fs/orangefs/inode.c 	struct orangefs_write_range *wr;
wr                486 fs/orangefs/inode.c 	wr = (struct orangefs_write_range *)page_private(page);
wr                495 fs/orangefs/inode.c 	} else if (page_offset(page) + offset <= wr->pos &&
wr                496 fs/orangefs/inode.c 	    wr->pos + wr->len <= page_offset(page) + offset + length) {
wr                505 fs/orangefs/inode.c 	} else if (wr->pos < page_offset(page) + offset &&
wr                506 fs/orangefs/inode.c 	    wr->pos + wr->len <= page_offset(page) + offset + length &&
wr                507 fs/orangefs/inode.c 	     page_offset(page) + offset < wr->pos + wr->len) {
wr                509 fs/orangefs/inode.c 		x = wr->pos + wr->len - (page_offset(page) + offset);
wr                510 fs/orangefs/inode.c 		WARN_ON(x > wr->len);
wr                511 fs/orangefs/inode.c 		wr->len -= x;
wr                512 fs/orangefs/inode.c 		wr->uid = current_fsuid();
wr                513 fs/orangefs/inode.c 		wr->gid = current_fsgid();
wr                515 fs/orangefs/inode.c 	} else if (page_offset(page) + offset <= wr->pos &&
wr                516 fs/orangefs/inode.c 	    page_offset(page) + offset + length < wr->pos + wr->len &&
wr                517 fs/orangefs/inode.c 	    wr->pos < page_offset(page) + offset + length) {
wr                519 fs/orangefs/inode.c 		x = page_offset(page) + offset + length - wr->pos;
wr                520 fs/orangefs/inode.c 		WARN_ON(x > wr->len);
wr                521 fs/orangefs/inode.c 		wr->pos += x;
wr                522 fs/orangefs/inode.c 		wr->len -= x;
wr                523 fs/orangefs/inode.c 		wr->uid = current_fsuid();
wr                524 fs/orangefs/inode.c 		wr->gid = current_fsgid();
wr                526 fs/orangefs/inode.c 	} else if (wr->pos < page_offset(page) + offset &&
wr                527 fs/orangefs/inode.c 	    page_offset(page) + offset + length < wr->pos + wr->len) {
wr                539 fs/orangefs/inode.c 		if (!((page_offset(page) + offset + length <= wr->pos) ^
wr                540 fs/orangefs/inode.c 		    (wr->pos + wr->len <= page_offset(page) + offset))) {
wr                545 fs/orangefs/inode.c 			    wr->pos, wr->len);
wr                726 fs/orangefs/inode.c 	struct orangefs_write_range *wr;
wr                748 fs/orangefs/inode.c 		wr = (struct orangefs_write_range *)page_private(page);
wr                749 fs/orangefs/inode.c 		if (uid_eq(wr->uid, current_fsuid()) &&
wr                750 fs/orangefs/inode.c 		    gid_eq(wr->gid, current_fsgid())) {
wr                751 fs/orangefs/inode.c 			wr->pos = page_offset(page);
wr                752 fs/orangefs/inode.c 			wr->len = PAGE_SIZE;
wr                761 fs/orangefs/inode.c 	wr = kmalloc(sizeof *wr, GFP_KERNEL);
wr                762 fs/orangefs/inode.c 	if (!wr) {
wr                766 fs/orangefs/inode.c 	wr->pos = page_offset(page);
wr                767 fs/orangefs/inode.c 	wr->len = PAGE_SIZE;
wr                768 fs/orangefs/inode.c 	wr->uid = current_fsuid();
wr                769 fs/orangefs/inode.c 	wr->gid = current_fsgid();
wr                771 fs/orangefs/inode.c 	set_page_private(page, (unsigned long)wr);
wr               1889 fs/ubifs/lpt.c int ubifs_lpt_init(struct ubifs_info *c, int rd, int wr)
wr               1899 fs/ubifs/lpt.c 	if (wr) {
wr               1908 fs/ubifs/lpt.c 	if (wr)
wr               1930 fs/ubifs/ubifs.h int ubifs_lpt_init(struct ubifs_info *c, int rd, int wr);
wr                 28 include/linux/hdlcdrv.h 	unsigned rd, wr;
wr                 35 include/linux/hdlcdrv.h 	unsigned int wr;
wr                 49 include/linux/hdlcdrv.h 		buf->buffer[buf->wr] = buf->shreg;
wr                 50 include/linux/hdlcdrv.h 		buf->wr = (buf->wr+1) % sizeof(buf->buffer);
wr                 58 include/linux/hdlcdrv.h 	buf->buffer[buf->wr] = bits & 0xff;
wr                 59 include/linux/hdlcdrv.h 	buf->wr = (buf->wr+1) % sizeof(buf->buffer);
wr                 60 include/linux/hdlcdrv.h 	buf->buffer[buf->wr] = (bits >> 8) & 0xff;
wr                 61 include/linux/hdlcdrv.h 	buf->wr = (buf->wr+1) % sizeof(buf->buffer);
wr                164 include/linux/hdlcdrv.h 	ret = !((HDLCDRV_HDLCBUFFER - 1 + hb->rd - hb->wr) % HDLCDRV_HDLCBUFFER);
wr                177 include/linux/hdlcdrv.h 	ret = (hb->rd == hb->wr);
wr                191 include/linux/hdlcdrv.h 	if (hb->rd == hb->wr)
wr                211 include/linux/hdlcdrv.h 	newp = (hb->wr+1) % HDLCDRV_HDLCBUFFER;
wr                213 include/linux/hdlcdrv.h 		hb->buf[hb->wr] = val & 0xffff;
wr                214 include/linux/hdlcdrv.h 		hb->wr = newp;
wr                185 include/linux/sunrpc/svc_rdma.h extern int svc_rdma_send(struct svcxprt_rdma *rdma, struct ib_send_wr *wr);
wr               1341 include/rdma/ib_verbs.h 	struct ib_send_wr	wr;
wr               1346 include/rdma/ib_verbs.h static inline const struct ib_rdma_wr *rdma_wr(const struct ib_send_wr *wr)
wr               1348 include/rdma/ib_verbs.h 	return container_of(wr, struct ib_rdma_wr, wr);
wr               1352 include/rdma/ib_verbs.h 	struct ib_send_wr	wr;
wr               1361 include/rdma/ib_verbs.h static inline const struct ib_atomic_wr *atomic_wr(const struct ib_send_wr *wr)
wr               1363 include/rdma/ib_verbs.h 	return container_of(wr, struct ib_atomic_wr, wr);
wr               1367 include/rdma/ib_verbs.h 	struct ib_send_wr	wr;
wr               1378 include/rdma/ib_verbs.h static inline const struct ib_ud_wr *ud_wr(const struct ib_send_wr *wr)
wr               1380 include/rdma/ib_verbs.h 	return container_of(wr, struct ib_ud_wr, wr);
wr               1384 include/rdma/ib_verbs.h 	struct ib_send_wr	wr;
wr               1390 include/rdma/ib_verbs.h static inline const struct ib_reg_wr *reg_wr(const struct ib_send_wr *wr)
wr               1392 include/rdma/ib_verbs.h 	return container_of(wr, struct ib_reg_wr, wr);
wr                172 include/rdma/rdmavt_qp.h 	struct ib_ud_wr wr;
wr                183 include/rdma/rdmavt_qp.h 		struct ib_send_wr wr;   /* don't use wr.sg_list */
wr                229 include/rdma/rdmavt_qp.h 	return ibah_to_rvtah(swqe->ud_wr.wr.ah);
wr                249 include/rdma/rdmavt_qp.h 	return swqe->ud_wr.wr.remote_qpn;
wr                259 include/rdma/rdmavt_qp.h 	return swqe->ud_wr.wr.remote_qkey;
wr                269 include/rdma/rdmavt_qp.h 	return swqe->ud_wr.wr.pkey_index;
wr                586 include/rdma/rdmavt_qp.h 	for (i = 0; i < wqe->wr.num_sge; i++) {
wr                852 include/rdma/rdmavt_qp.h 	int flags = wqe->wr.send_flags;
wr                863 include/rdma/rdmavt_qp.h 		wr_id = wqe->wr.wr_id;
wr                 25 include/rdma/rw.h 			struct ib_rdma_wr	wr;
wr                 37 include/rdma/rw.h 			struct ib_rdma_wr	wr;
wr                 24 include/trace/events/ib_mad.h 	TP_PROTO(struct ib_mad_send_wr_private *wr,
wr                 26 include/trace/events/ib_mad.h 	TP_ARGS(wr, qp_info),
wr                 56 include/trace/events/ib_mad.h 		__entry->dev_index = wr->mad_agent_priv->agent.device->index;
wr                 57 include/trace/events/ib_mad.h 		__entry->port_num = wr->mad_agent_priv->agent.port_num;
wr                 58 include/trace/events/ib_mad.h 		__entry->qp_num = wr->mad_agent_priv->qp_info->qp->qp_num;
wr                 59 include/trace/events/ib_mad.h 		__entry->agent_priv = wr->mad_agent_priv;
wr                 60 include/trace/events/ib_mad.h 		__entry->wrtid = wr->tid;
wr                 61 include/trace/events/ib_mad.h 		__entry->max_retries = wr->max_retries;
wr                 62 include/trace/events/ib_mad.h 		__entry->retries_left = wr->retries_left;
wr                 63 include/trace/events/ib_mad.h 		__entry->retry = wr->retry;
wr                 64 include/trace/events/ib_mad.h 		__entry->timeout = wr->timeout;
wr                 65 include/trace/events/ib_mad.h 		__entry->length = wr->send_buf.hdr_len +
wr                 66 include/trace/events/ib_mad.h 				  wr->send_buf.data_len;
wr                 68 include/trace/events/ib_mad.h 			((struct ib_mad_hdr *)wr->send_buf.mad)->base_version;
wr                 70 include/trace/events/ib_mad.h 			((struct ib_mad_hdr *)wr->send_buf.mad)->mgmt_class;
wr                 72 include/trace/events/ib_mad.h 			((struct ib_mad_hdr *)wr->send_buf.mad)->class_version;
wr                 74 include/trace/events/ib_mad.h 			((struct ib_mad_hdr *)wr->send_buf.mad)->method;
wr                 76 include/trace/events/ib_mad.h 			((struct ib_mad_hdr *)wr->send_buf.mad)->status;
wr                 78 include/trace/events/ib_mad.h 			((struct ib_mad_hdr *)wr->send_buf.mad)->class_specific;
wr                 79 include/trace/events/ib_mad.h 		__entry->tid = ((struct ib_mad_hdr *)wr->send_buf.mad)->tid;
wr                 81 include/trace/events/ib_mad.h 			((struct ib_mad_hdr *)wr->send_buf.mad)->attr_id;
wr                 83 include/trace/events/ib_mad.h 			((struct ib_mad_hdr *)wr->send_buf.mad)->attr_mod;
wr                 84 include/trace/events/ib_mad.h 		create_mad_addr_info(wr, qp_info, __entry);
wr                109 include/trace/events/ib_mad.h 	TP_PROTO(struct ib_mad_send_wr_private *wr,
wr                111 include/trace/events/ib_mad.h 	TP_ARGS(wr, qp_info));
wr                113 include/trace/events/ib_mad.h 	TP_PROTO(struct ib_mad_send_wr_private *wr,
wr                115 include/trace/events/ib_mad.h 	TP_ARGS(wr, qp_info));
wr                117 include/trace/events/ib_mad.h 	TP_PROTO(struct ib_mad_send_wr_private *wr,
wr                119 include/trace/events/ib_mad.h 	TP_ARGS(wr, qp_info));
wr                122 include/trace/events/ib_mad.h 	TP_PROTO(struct ib_mad_send_wr_private *wr, struct ib_wc *wc),
wr                123 include/trace/events/ib_mad.h 	TP_ARGS(wr, wc),
wr                145 include/trace/events/ib_mad.h 		__entry->dev_index = wr->mad_agent_priv->agent.device->index;
wr                146 include/trace/events/ib_mad.h 		__entry->port_num = wr->mad_agent_priv->agent.port_num;
wr                147 include/trace/events/ib_mad.h 		__entry->qp_num = wr->mad_agent_priv->qp_info->qp->qp_num;
wr                148 include/trace/events/ib_mad.h 		__entry->agent_priv = wr->mad_agent_priv;
wr                149 include/trace/events/ib_mad.h 		__entry->wrtid = wr->tid;
wr                150 include/trace/events/ib_mad.h 		__entry->max_retries = wr->max_retries;
wr                151 include/trace/events/ib_mad.h 		__entry->retries_left = wr->retries_left;
wr                152 include/trace/events/ib_mad.h 		__entry->retry = wr->retry;
wr                153 include/trace/events/ib_mad.h 		__entry->timeout = wr->timeout;
wr                155 include/trace/events/ib_mad.h 			((struct ib_mad_hdr *)wr->send_buf.mad)->base_version;
wr                157 include/trace/events/ib_mad.h 			((struct ib_mad_hdr *)wr->send_buf.mad)->mgmt_class;
wr                159 include/trace/events/ib_mad.h 			((struct ib_mad_hdr *)wr->send_buf.mad)->class_version;
wr                161 include/trace/events/ib_mad.h 			((struct ib_mad_hdr *)wr->send_buf.mad)->method;
wr                163 include/trace/events/ib_mad.h 			((struct ib_mad_hdr *)wr->send_buf.mad)->status;
wr               1641 include/trace/events/rpcrdma.h 		const struct ib_send_wr *wr
wr               1644 include/trace/events/rpcrdma.h 	TP_ARGS(wr),
wr               1653 include/trace/events/rpcrdma.h 		__entry->cqe = wr->wr_cqe;
wr               1654 include/trace/events/rpcrdma.h 		__entry->num_sge = wr->num_sge;
wr               1655 include/trace/events/rpcrdma.h 		__entry->inv_rkey = (wr->opcode == IB_WR_SEND_WITH_INV) ?
wr               1656 include/trace/events/rpcrdma.h 					wr->ex.invalidate_rkey : 0;
wr               1669 include/trace/events/rpcrdma.h 		const struct ib_recv_wr *wr,
wr               1673 include/trace/events/rpcrdma.h 	TP_ARGS(wr, status),
wr               1681 include/trace/events/rpcrdma.h 		__entry->cqe = wr->wr_cqe;
wr                821 include/uapi/rdma/ib_user_verbs.h 	} wr;
wr                105 include/uapi/rdma/rdma_user_rxe.h 	} wr;
wr                134 include/uapi/rdma/rdma_user_rxe.h 	struct rxe_send_wr	wr;
wr                273 include/uapi/rdma/vmw_pvrdma-abi.h 	} wr;
wr                292 lib/decompress_unlzma.c static inline size_t INIT get_pos(struct writer *wr)
wr                295 lib/decompress_unlzma.c 		wr->global_pos + wr->buffer_pos;
wr                298 lib/decompress_unlzma.c static inline uint8_t INIT peek_old_byte(struct writer *wr,
wr                301 lib/decompress_unlzma.c 	if (!wr->flush) {
wr                303 lib/decompress_unlzma.c 		while (offs > wr->header->dict_size)
wr                304 lib/decompress_unlzma.c 			offs -= wr->header->dict_size;
wr                305 lib/decompress_unlzma.c 		pos = wr->buffer_pos - offs;
wr                306 lib/decompress_unlzma.c 		return wr->buffer[pos];
wr                308 lib/decompress_unlzma.c 		uint32_t pos = wr->buffer_pos - offs;
wr                309 lib/decompress_unlzma.c 		while (pos >= wr->header->dict_size)
wr                310 lib/decompress_unlzma.c 			pos += wr->header->dict_size;
wr                311 lib/decompress_unlzma.c 		return wr->buffer[pos];
wr                316 lib/decompress_unlzma.c static inline int INIT write_byte(struct writer *wr, uint8_t byte)
wr                318 lib/decompress_unlzma.c 	wr->buffer[wr->buffer_pos++] = wr->previous_byte = byte;
wr                319 lib/decompress_unlzma.c 	if (wr->flush && wr->buffer_pos == wr->header->dict_size) {
wr                320 lib/decompress_unlzma.c 		wr->buffer_pos = 0;
wr                321 lib/decompress_unlzma.c 		wr->global_pos += wr->header->dict_size;
wr                322 lib/decompress_unlzma.c 		if (wr->flush((char *)wr->buffer, wr->header->dict_size)
wr                323 lib/decompress_unlzma.c 				!= wr->header->dict_size)
wr                330 lib/decompress_unlzma.c static inline int INIT copy_byte(struct writer *wr, uint32_t offs)
wr                332 lib/decompress_unlzma.c 	return write_byte(wr, peek_old_byte(wr, offs));
wr                335 lib/decompress_unlzma.c static inline int INIT copy_bytes(struct writer *wr,
wr                339 lib/decompress_unlzma.c 		if (copy_byte(wr, rep0))
wr                342 lib/decompress_unlzma.c 	} while (len != 0 && wr->buffer_pos < wr->header->dst_size);
wr                347 lib/decompress_unlzma.c static inline int INIT process_bit0(struct writer *wr, struct rc *rc,
wr                355 lib/decompress_unlzma.c 		 * (((get_pos(wr) & literal_pos_mask) << lc)
wr                356 lib/decompress_unlzma.c 		    + (wr->previous_byte >> (8 - lc))))
wr                360 lib/decompress_unlzma.c 		int match_byte = peek_old_byte(wr, cst->rep0);
wr                388 lib/decompress_unlzma.c 	return write_byte(wr, mi);
wr                391 lib/decompress_unlzma.c static inline int INIT process_bit1(struct writer *wr, struct rc *rc,
wr                422 lib/decompress_unlzma.c 				return copy_byte(wr, cst->rep0);
wr                525 lib/decompress_unlzma.c 		if (cst->rep0 > wr->header->dict_size
wr                526 lib/decompress_unlzma.c 				|| cst->rep0 > get_pos(wr))
wr                532 lib/decompress_unlzma.c 	return copy_bytes(wr, cst->rep0, len);
wr                553 lib/decompress_unlzma.c 	struct writer wr;
wr                572 lib/decompress_unlzma.c 	wr.header = &header;
wr                573 lib/decompress_unlzma.c 	wr.flush = flush;
wr                574 lib/decompress_unlzma.c 	wr.global_pos = 0;
wr                575 lib/decompress_unlzma.c 	wr.previous_byte = 0;
wr                576 lib/decompress_unlzma.c 	wr.buffer_pos = 0;
wr                613 lib/decompress_unlzma.c 		wr.buffer = output;
wr                615 lib/decompress_unlzma.c 		wr.bufsize = MIN(header.dst_size, header.dict_size);
wr                616 lib/decompress_unlzma.c 		wr.buffer = large_malloc(wr.bufsize);
wr                618 lib/decompress_unlzma.c 	if (wr.buffer == NULL)
wr                631 lib/decompress_unlzma.c 	while (get_pos(&wr) < header.dst_size) {
wr                632 lib/decompress_unlzma.c 		int pos_state =	get_pos(&wr) & pos_state_mask;
wr                636 lib/decompress_unlzma.c 			if (process_bit0(&wr, &rc, &cst, p, pos_state, prob,
wr                642 lib/decompress_unlzma.c 			if (process_bit1(&wr, &rc, &cst, p, pos_state, prob)) {
wr                655 lib/decompress_unlzma.c 	if (!wr.flush || wr.flush(wr.buffer, wr.buffer_pos) == wr.buffer_pos)
wr                661 lib/decompress_unlzma.c 		large_free(wr.buffer);
wr                144 net/9p/trans_fd.c 	struct file *wr;
wr                234 net/9p/trans_fd.c 	if (ts->rd != ts->wr)
wr                235 net/9p/trans_fd.c 		ret = (ret & ~EPOLLOUT) | (vfs_poll(ts->wr, pt) & ~EPOLLIN);
wr                421 net/9p/trans_fd.c 	if (!(ts->wr->f_flags & O_NONBLOCK))
wr                424 net/9p/trans_fd.c 	ret = kernel_write(ts->wr, v, len, &ts->wr->f_pos);
wr                806 net/9p/trans_fd.c 	ts->wr = fget(wfd);
wr                807 net/9p/trans_fd.c 	if (!ts->rd || !ts->wr) {
wr                810 net/9p/trans_fd.c 		if (ts->wr)
wr                811 net/9p/trans_fd.c 			fput(ts->wr);
wr                841 net/9p/trans_fd.c 	p->wr = p->rd = file;
wr                902 net/9p/trans_fd.c 	if (ts->wr)
wr                903 net/9p/trans_fd.c 		fput(ts->wr);
wr                386 net/9p/trans_rdma.c 	struct ib_recv_wr wr;
wr                401 net/9p/trans_rdma.c 	wr.next = NULL;
wr                402 net/9p/trans_rdma.c 	wr.wr_cqe = &c->cqe;
wr                403 net/9p/trans_rdma.c 	wr.sg_list = &sge;
wr                404 net/9p/trans_rdma.c 	wr.num_sge = 1;
wr                405 net/9p/trans_rdma.c 	return ib_post_recv(rdma->qp, &wr, NULL);
wr                415 net/9p/trans_rdma.c 	struct ib_send_wr wr;
wr                493 net/9p/trans_rdma.c 	wr.next = NULL;
wr                494 net/9p/trans_rdma.c 	wr.wr_cqe = &c->cqe;
wr                495 net/9p/trans_rdma.c 	wr.opcode = IB_WR_SEND;
wr                496 net/9p/trans_rdma.c 	wr.send_flags = IB_SEND_SIGNALED;
wr                497 net/9p/trans_rdma.c 	wr.sg_list = &sge;
wr                498 net/9p/trans_rdma.c 	wr.num_sge = 1;
wr                510 net/9p/trans_rdma.c 	err = ib_post_send(rdma->qp, &wr, NULL);
wr                154 net/rds/ib_frmr.c 	reg_wr.wr.wr_id = (unsigned long)(void *)ibmr;
wr                155 net/rds/ib_frmr.c 	reg_wr.wr.opcode = IB_WR_REG_MR;
wr                156 net/rds/ib_frmr.c 	reg_wr.wr.num_sge = 0;
wr                162 net/rds/ib_frmr.c 	reg_wr.wr.send_flags = IB_SEND_SIGNALED;
wr                164 net/rds/ib_frmr.c 	ret = ib_post_send(ibmr->ic->i_cm_id->qp, &reg_wr.wr, NULL);
wr                575 net/rds/ib_recv.c 	struct ib_send_wr *wr = &ic->i_ack_wr;
wr                582 net/rds/ib_recv.c 	wr->sg_list = sge;
wr                583 net/rds/ib_recv.c 	wr->num_sge = 1;
wr                584 net/rds/ib_recv.c 	wr->opcode = IB_WR_SEND;
wr                585 net/rds/ib_recv.c 	wr->wr_id = RDS_IB_ACK_WR_ID;
wr                586 net/rds/ib_recv.c 	wr->send_flags = IB_SEND_SIGNALED | IB_SEND_SOLICITED;
wr                775 net/rds/ib_send.c 		send->s_atomic_wr.wr.opcode = IB_WR_MASKED_ATOMIC_CMP_AND_SWP;
wr                781 net/rds/ib_send.c 		send->s_atomic_wr.wr.opcode = IB_WR_MASKED_ATOMIC_FETCH_AND_ADD;
wr                789 net/rds/ib_send.c 	send->s_atomic_wr.wr.num_sge = 1;
wr                790 net/rds/ib_send.c 	send->s_atomic_wr.wr.next = NULL;
wr                817 net/rds/ib_send.c 	failed_wr = &send->s_atomic_wr.wr;
wr                818 net/rds/ib_send.c 	ret = ib_post_send(ic->i_cm_id->qp, &send->s_atomic_wr.wr, &failed_wr);
wr                821 net/rds/ib_send.c 	BUG_ON(failed_wr != &send->s_atomic_wr.wr);
wr                830 net/rds/ib_send.c 	if (unlikely(failed_wr != &send->s_atomic_wr.wr)) {
wr                832 net/rds/ib_send.c 		BUG_ON(failed_wr != &send->s_atomic_wr.wr);
wr                909 net/rds/ib_send.c 			send->s_rdma_wr.wr.num_sge = max_sge;
wr                912 net/rds/ib_send.c 			send->s_rdma_wr.wr.num_sge = num_sge;
wr                915 net/rds/ib_send.c 		send->s_rdma_wr.wr.next = NULL;
wr                918 net/rds/ib_send.c 			prev->s_rdma_wr.wr.next = &send->s_rdma_wr.wr;
wr                920 net/rds/ib_send.c 		for (j = 0; j < send->s_rdma_wr.wr.num_sge &&
wr                935 net/rds/ib_send.c 			&send->s_rdma_wr.wr,
wr                936 net/rds/ib_send.c 			send->s_rdma_wr.wr.num_sge,
wr                937 net/rds/ib_send.c 			send->s_rdma_wr.wr.next);
wr                958 net/rds/ib_send.c 	failed_wr = &first->s_rdma_wr.wr;
wr                959 net/rds/ib_send.c 	ret = ib_post_send(ic->i_cm_id->qp, &first->s_rdma_wr.wr, &failed_wr);
wr                961 net/rds/ib_send.c 		 first, &first->s_rdma_wr.wr, ret, failed_wr);
wr                962 net/rds/ib_send.c 	BUG_ON(failed_wr != &first->s_rdma_wr.wr);
wr                971 net/rds/ib_send.c 	if (unlikely(failed_wr != &first->s_rdma_wr.wr)) {
wr                973 net/rds/ib_send.c 		BUG_ON(failed_wr != &first->s_rdma_wr.wr);
wr                275 net/smc/smc_tx.c 	rdma_wr->wr.wr_id = smc_wr_tx_get_next_wr_id(link);
wr                276 net/smc/smc_tx.c 	rdma_wr->wr.num_sge = num_sges;
wr                284 net/smc/smc_tx.c 	rc = ib_post_send(link->roce_qp, &rdma_wr->wr, NULL);
wr                324 net/smc/smc_tx.c 			wr_rdma_buf->wr_tx_rdma[dstchunk].wr.sg_list;
wr                263 net/smc/smc_wr.c 	link->wr_reg.wr.wr_id = (u64)(uintptr_t)mr;
wr                266 net/smc/smc_wr.c 	rc = ib_post_send(link->roce_qp, &link->wr_reg.wr, NULL);
wr                486 net/smc/smc_wr.c 		lnk->wr_tx_rdmas[i].wr_tx_rdma[0].wr.opcode = IB_WR_RDMA_WRITE;
wr                487 net/smc/smc_wr.c 		lnk->wr_tx_rdmas[i].wr_tx_rdma[1].wr.opcode = IB_WR_RDMA_WRITE;
wr                488 net/smc/smc_wr.c 		lnk->wr_tx_rdmas[i].wr_tx_rdma[0].wr.sg_list =
wr                490 net/smc/smc_wr.c 		lnk->wr_tx_rdmas[i].wr_tx_rdma[1].wr.sg_list =
wr                502 net/smc/smc_wr.c 	lnk->wr_reg.wr.next = NULL;
wr                503 net/smc/smc_wr.c 	lnk->wr_reg.wr.num_sge = 0;
wr                504 net/smc/smc_wr.c 	lnk->wr_reg.wr.send_flags = IB_SEND_SIGNALED;
wr                505 net/smc/smc_wr.c 	lnk->wr_reg.wr.opcode = IB_WR_REG_MR;
wr                435 net/sunrpc/xprtrdma/frwr_ops.c 		frwr->fr_regwr.wr.next = post_wr;
wr                436 net/sunrpc/xprtrdma/frwr_ops.c 		frwr->fr_regwr.wr.wr_cqe = &frwr->fr_cqe;
wr                437 net/sunrpc/xprtrdma/frwr_ops.c 		frwr->fr_regwr.wr.num_sge = 0;
wr                438 net/sunrpc/xprtrdma/frwr_ops.c 		frwr->fr_regwr.wr.opcode = IB_WR_REG_MR;
wr                439 net/sunrpc/xprtrdma/frwr_ops.c 		frwr->fr_regwr.wr.send_flags = 0;
wr                441 net/sunrpc/xprtrdma/frwr_ops.c 		post_wr = &frwr->fr_regwr.wr;
wr                288 net/sunrpc/xprtrdma/svc_rdma_sendto.c int svc_rdma_send(struct svcxprt_rdma *rdma, struct ib_send_wr *wr)
wr                309 net/sunrpc/xprtrdma/svc_rdma_sendto.c 		trace_svcrdma_post_send(wr);
wr                310 net/sunrpc/xprtrdma/svc_rdma_sendto.c 		ret = ib_post_send(rdma->sc_qp, wr, NULL);
wr               1500 net/sunrpc/xprtrdma/verbs.c 	struct ib_recv_wr *i, *wr, *bad_wr;
wr               1515 net/sunrpc/xprtrdma/verbs.c 	wr = NULL;
wr               1527 net/sunrpc/xprtrdma/verbs.c 		rep->rr_recv_wr.next = wr;
wr               1528 net/sunrpc/xprtrdma/verbs.c 		wr = &rep->rr_recv_wr;
wr               1531 net/sunrpc/xprtrdma/verbs.c 	if (!wr)
wr               1534 net/sunrpc/xprtrdma/verbs.c 	for (i = wr; i; i = i->next) {
wr               1544 net/sunrpc/xprtrdma/verbs.c 	rc = ib_post_recv(r_xprt->rx_ia.ri_id->qp, wr,
wr               1549 net/sunrpc/xprtrdma/verbs.c 		for (wr = bad_wr; wr;) {
wr               1552 net/sunrpc/xprtrdma/verbs.c 			rep = container_of(wr, struct rpcrdma_rep, rr_recv_wr);
wr               1553 net/sunrpc/xprtrdma/verbs.c 			wr = wr->next;
wr               1562 net/sunrpc/xprtrdma/verbs.c 	for (i = wr; i;) {
wr               1913 net/wireless/reg.c 	const struct ieee80211_regdomain *wr = get_wiphy_regdom(wiphy);
wr               1916 net/wireless/reg.c 	if (is_world_regdom(cr->alpha2) || (wr && is_world_regdom(wr->alpha2)))
wr                 91 sound/pci/ctxfi/cthardware.h 	int (*src_set_wr)(void *blk, unsigned int wr);
wr                212 sound/pci/ctxfi/cthw20k1.c static int src_set_wr(void *blk, unsigned int wr)
wr                216 sound/pci/ctxfi/cthw20k1.c 	set_field(&ctl->ctl, SRCCTL_WR, wr);
wr                212 sound/pci/ctxfi/cthw20k2.c static int src_set_wr(void *blk, unsigned int wr)
wr                216 sound/pci/ctxfi/cthw20k2.c 	set_field(&ctl->ctl, SRCCTL_WR, wr);
wr               1245 sound/soc/codecs/wm_adsp.c 	unsigned int out, rd, wr, vol;
wr               1249 sound/soc/codecs/wm_adsp.c 		wr = SNDRV_CTL_ELEM_ACCESS_TLV_WRITE;
wr               1255 sound/soc/codecs/wm_adsp.c 		wr = SNDRV_CTL_ELEM_ACCESS_WRITE;
wr               1264 sound/soc/codecs/wm_adsp.c 			out |= wr;
wr               1268 sound/soc/codecs/wm_adsp.c 		out |= rd | wr | vol;
wr                 19 sound/soc/uniphier/aio-core.c static u64 rb_cnt(u64 wr, u64 rd, u64 len)
wr                 21 sound/soc/uniphier/aio-core.c 	if (rd <= wr)
wr                 22 sound/soc/uniphier/aio-core.c 		return wr - rd;
wr                 24 sound/soc/uniphier/aio-core.c 		return len - (rd - wr);
wr                 27 sound/soc/uniphier/aio-core.c static u64 rb_cnt_to_end(u64 wr, u64 rd, u64 len)
wr                 29 sound/soc/uniphier/aio-core.c 	if (rd <= wr)
wr                 30 sound/soc/uniphier/aio-core.c 		return wr - rd;
wr                 35 sound/soc/uniphier/aio-core.c static u64 rb_space(u64 wr, u64 rd, u64 len)
wr                 37 sound/soc/uniphier/aio-core.c 	if (rd <= wr)
wr                 38 sound/soc/uniphier/aio-core.c 		return len - (wr - rd) - 8;
wr                 40 sound/soc/uniphier/aio-core.c 		return rd - wr - 8;
wr                 43 sound/soc/uniphier/aio-core.c static u64 rb_space_to_end(u64 wr, u64 rd, u64 len)
wr                 45 sound/soc/uniphier/aio-core.c 	if (rd > wr)
wr                 46 sound/soc/uniphier/aio-core.c 		return rd - wr - 8;
wr                 48 sound/soc/uniphier/aio-core.c 		return len - wr;
wr                 50 sound/soc/uniphier/aio-core.c 		return len - wr - 8;
wr                308 tools/bpf/bpftool/json_writer.c 	json_writer_t *wr = jsonw_new(stdout);
wr                310 tools/bpf/bpftool/json_writer.c 	jsonw_start_object(wr);
wr                311 tools/bpf/bpftool/json_writer.c 	jsonw_pretty(wr, true);
wr                312 tools/bpf/bpftool/json_writer.c 	jsonw_name(wr, "Vyatta");
wr                313 tools/bpf/bpftool/json_writer.c 	jsonw_start_object(wr);
wr                314 tools/bpf/bpftool/json_writer.c 	jsonw_string_field(wr, "url", "http://vyatta.com");
wr                315 tools/bpf/bpftool/json_writer.c 	jsonw_uint_field(wr, "downloads", 2000000ul);
wr                316 tools/bpf/bpftool/json_writer.c 	jsonw_float_field(wr, "stock", 8.16);
wr                318 tools/bpf/bpftool/json_writer.c 	jsonw_name(wr, "ARGV");
wr                319 tools/bpf/bpftool/json_writer.c 	jsonw_start_array(wr);
wr                321 tools/bpf/bpftool/json_writer.c 		jsonw_string(wr, *++argv);
wr                322 tools/bpf/bpftool/json_writer.c 	jsonw_end_array(wr);
wr                324 tools/bpf/bpftool/json_writer.c 	jsonw_name(wr, "empty");
wr                325 tools/bpf/bpftool/json_writer.c 	jsonw_start_array(wr);
wr                326 tools/bpf/bpftool/json_writer.c 	jsonw_end_array(wr);
wr                328 tools/bpf/bpftool/json_writer.c 	jsonw_name(wr, "NIL");
wr                329 tools/bpf/bpftool/json_writer.c 	jsonw_start_object(wr);
wr                330 tools/bpf/bpftool/json_writer.c 	jsonw_end_object(wr);
wr                332 tools/bpf/bpftool/json_writer.c 	jsonw_null_field(wr, "my_null");
wr                334 tools/bpf/bpftool/json_writer.c 	jsonw_name(wr, "special chars");
wr                335 tools/bpf/bpftool/json_writer.c 	jsonw_start_array(wr);
wr                336 tools/bpf/bpftool/json_writer.c 	jsonw_string_field(wr, "slash", "/");
wr                337 tools/bpf/bpftool/json_writer.c 	jsonw_string_field(wr, "newline", "\n");
wr                338 tools/bpf/bpftool/json_writer.c 	jsonw_string_field(wr, "tab", "\t");
wr                339 tools/bpf/bpftool/json_writer.c 	jsonw_string_field(wr, "ff", "\f");
wr                340 tools/bpf/bpftool/json_writer.c 	jsonw_string_field(wr, "quote", "\"");
wr                341 tools/bpf/bpftool/json_writer.c 	jsonw_string_field(wr, "tick", "\'");
wr                342 tools/bpf/bpftool/json_writer.c 	jsonw_string_field(wr, "backslash", "\\");
wr                343 tools/bpf/bpftool/json_writer.c 	jsonw_end_array(wr);
wr                345 tools/bpf/bpftool/json_writer.c 	jsonw_end_object(wr);
wr                347 tools/bpf/bpftool/json_writer.c 	jsonw_end_object(wr);
wr                348 tools/bpf/bpftool/json_writer.c 	jsonw_destroy(&wr);
wr                 33 tools/testing/selftests/breakpoints/breakpoint_test_arm64.c static void child(int size, int wr)
wr                 35 tools/testing/selftests/breakpoints/breakpoint_test_arm64.c 	volatile uint8_t *addr = &var[32 + wr];
wr                112 tools/testing/selftests/breakpoints/breakpoint_test_arm64.c static bool arun_test(int wr_size, int wp_size, int wr, int wp)
wr                125 tools/testing/selftests/breakpoints/breakpoint_test_arm64.c 		child(wr_size, wr);
wr                204 tools/testing/selftests/breakpoints/breakpoint_test_arm64.c 	int wr, wp, size;
wr                215 tools/testing/selftests/breakpoints/breakpoint_test_arm64.c 		for (wr = 0; wr <= 32; wr = wr + size) {
wr                216 tools/testing/selftests/breakpoints/breakpoint_test_arm64.c 			for (wp = wr - size; wp <= wr + size; wp = wp + size) {
wr                217 tools/testing/selftests/breakpoints/breakpoint_test_arm64.c 				result = run_test(size, MIN(size, 8), wr, wp);
wr                218 tools/testing/selftests/breakpoints/breakpoint_test_arm64.c 				if ((result && wr == wp) ||
wr                219 tools/testing/selftests/breakpoints/breakpoint_test_arm64.c 				    (!result && wr != wp))
wr                222 tools/testing/selftests/breakpoints/breakpoint_test_arm64.c 						size, wr, wp);
wr                226 tools/testing/selftests/breakpoints/breakpoint_test_arm64.c 						size, wr, wp);
wr                420 tools/testing/selftests/net/tcp_mmap.c 		long wr = FILE_SZ - total;
wr                422 tools/testing/selftests/net/tcp_mmap.c 		if (wr > chunk_size)
wr                423 tools/testing/selftests/net/tcp_mmap.c 			wr = chunk_size;
wr                425 tools/testing/selftests/net/tcp_mmap.c 		wr = send(fd, buffer, wr, zflg ? MSG_ZEROCOPY : 0);
wr                426 tools/testing/selftests/net/tcp_mmap.c 		if (wr <= 0)
wr                428 tools/testing/selftests/net/tcp_mmap.c 		total += wr;
wr               1735 virt/kvm/arm/vgic/vgic-its.c #define REGISTER_ITS_DESC(off, rd, wr, length, acc)		\
wr               1741 virt/kvm/arm/vgic/vgic-its.c 	.its_write = wr,					\
wr               1744 virt/kvm/arm/vgic/vgic-its.c #define REGISTER_ITS_DESC_UACCESS(off, rd, wr, uwr, length, acc)\
wr               1750 virt/kvm/arm/vgic/vgic-its.c 	.its_write = wr,					\
wr                448 virt/kvm/arm/vgic/vgic-mmio-v3.c #define REGISTER_DESC_WITH_BITS_PER_IRQ_SHARED(off, rd, wr, ur, uw, bpi, acc) \
wr                462 virt/kvm/arm/vgic/vgic-mmio-v3.c 		.write = wr,						\
wr                 67 virt/kvm/arm/vgic/vgic-mmio.h #define REGISTER_DESC_WITH_BITS_PER_IRQ(off, rd, wr, ur, uw, bpi, acc)	\
wr                 74 virt/kvm/arm/vgic/vgic-mmio.h 		.write = wr,						\
wr                 79 virt/kvm/arm/vgic/vgic-mmio.h #define REGISTER_DESC_WITH_LENGTH(off, rd, wr, length, acc)		\
wr                 86 virt/kvm/arm/vgic/vgic-mmio.h 		.write = wr,						\
wr                 89 virt/kvm/arm/vgic/vgic-mmio.h #define REGISTER_DESC_WITH_LENGTH_UACCESS(off, rd, wr, urd, uwr, length, acc) \
wr                 96 virt/kvm/arm/vgic/vgic-mmio.h 		.write = wr,						\