Lines Matching refs:cpu_to_be32
104 [IB_WR_SEND] = cpu_to_be32(MLX4_OPCODE_SEND),
105 [IB_WR_LSO] = cpu_to_be32(MLX4_OPCODE_LSO),
106 [IB_WR_SEND_WITH_IMM] = cpu_to_be32(MLX4_OPCODE_SEND_IMM),
107 [IB_WR_RDMA_WRITE] = cpu_to_be32(MLX4_OPCODE_RDMA_WRITE),
108 [IB_WR_RDMA_WRITE_WITH_IMM] = cpu_to_be32(MLX4_OPCODE_RDMA_WRITE_IMM),
109 [IB_WR_RDMA_READ] = cpu_to_be32(MLX4_OPCODE_RDMA_READ),
110 [IB_WR_ATOMIC_CMP_AND_SWP] = cpu_to_be32(MLX4_OPCODE_ATOMIC_CS),
111 [IB_WR_ATOMIC_FETCH_AND_ADD] = cpu_to_be32(MLX4_OPCODE_ATOMIC_FA),
112 [IB_WR_SEND_WITH_INV] = cpu_to_be32(MLX4_OPCODE_SEND_INVAL),
113 [IB_WR_LOCAL_INV] = cpu_to_be32(MLX4_OPCODE_LOCAL_INVAL),
114 [IB_WR_FAST_REG_MR] = cpu_to_be32(MLX4_OPCODE_FMR),
115 [IB_WR_MASKED_ATOMIC_CMP_AND_SWP] = cpu_to_be32(MLX4_OPCODE_MASKED_ATOMIC_CS),
116 [IB_WR_MASKED_ATOMIC_FETCH_AND_ADD] = cpu_to_be32(MLX4_OPCODE_MASKED_ATOMIC_FA),
117 [IB_WR_BIND_MW] = cpu_to_be32(MLX4_OPCODE_BIND_MW),
221 stamp = ind & qp->sq.wqe_cnt ? cpu_to_be32(0x7fffffff) : in stamp_send_wqe()
222 cpu_to_be32(0xffffffff); in stamp_send_wqe()
232 *wqe = cpu_to_be32(0xffffffff); in stamp_send_wqe()
251 av->port_pd = cpu_to_be32((qp->port << 24) | to_mpd(qp->ibqp.pd)->pdn); in post_nop_wqe()
258 inl->byte_count = cpu_to_be32(1 << 31 | (size - s - sizeof *inl)); in post_nop_wqe()
268 ctrl->owner_opcode = cpu_to_be32(MLX4_OPCODE_NOP | MLX4_WQE_CTRL_NEC) | in post_nop_wqe()
269 (n & qp->sq.wqe_cnt ? cpu_to_be32(1 << 31) : 0); in post_nop_wqe()
704 qp->sq_signal_bits = cpu_to_be32(MLX4_WQE_CTRL_CQ_UPDATE); in create_qp_common()
1254 return cpu_to_be32(hw_access_flags); in to_mlx4_access_flags()
1305 cpu_to_be32((ah->grh.traffic_class << 20) | in _mlx4_set_path()
1470 context->flags = cpu_to_be32((to_mlx4_state(new_state) << 28) | in __mlx4_ib_modify_qp()
1474 context->flags |= cpu_to_be32(MLX4_QP_PM_MIGRATED << 11); in __mlx4_ib_modify_qp()
1479 context->flags |= cpu_to_be32(MLX4_QP_PM_MIGRATED << 11); in __mlx4_ib_modify_qp()
1482 context->flags |= cpu_to_be32(MLX4_QP_PM_REARM << 11); in __mlx4_ib_modify_qp()
1485 context->flags |= cpu_to_be32(MLX4_QP_PM_ARMED << 11); in __mlx4_ib_modify_qp()
1520 context->xrcd = cpu_to_be32((u32) qp->xrcdn); in __mlx4_ib_modify_qp()
1522 context->param3 |= cpu_to_be32(1 << 30); in __mlx4_ib_modify_qp()
1526 context->usr_page = cpu_to_be32(to_mucontext(ibqp->uobject->context)->uar.index); in __mlx4_ib_modify_qp()
1528 context->usr_page = cpu_to_be32(dev->priv_uar.index); in __mlx4_ib_modify_qp()
1531 context->remote_qpn = cpu_to_be32(attr->dest_qp_num); in __mlx4_ib_modify_qp()
1598 context->pd = cpu_to_be32(pd->pdn); in __mlx4_ib_modify_qp()
1599 context->cqn_send = cpu_to_be32(send_cq->mcq.cqn); in __mlx4_ib_modify_qp()
1600 context->cqn_recv = cpu_to_be32(recv_cq->mcq.cqn); in __mlx4_ib_modify_qp()
1601 context->params1 = cpu_to_be32(MLX4_IB_ACK_REQ_FREQ << 28); in __mlx4_ib_modify_qp()
1605 context->params1 |= cpu_to_be32(1 << 11); in __mlx4_ib_modify_qp()
1608 context->params1 |= cpu_to_be32(attr->rnr_retry << 13); in __mlx4_ib_modify_qp()
1613 context->params1 |= cpu_to_be32(attr->retry_cnt << 16); in __mlx4_ib_modify_qp()
1620 cpu_to_be32(fls(attr->max_rd_atomic - 1) << 21); in __mlx4_ib_modify_qp()
1625 context->next_send_psn = cpu_to_be32(attr->sq_psn); in __mlx4_ib_modify_qp()
1630 cpu_to_be32(fls(attr->max_dest_rd_atomic - 1) << 21); in __mlx4_ib_modify_qp()
1640 context->params2 |= cpu_to_be32(MLX4_QP_BIT_RIC); in __mlx4_ib_modify_qp()
1643 context->rnr_nextrecvpsn |= cpu_to_be32(attr->min_rnr_timer << 24); in __mlx4_ib_modify_qp()
1647 context->rnr_nextrecvpsn |= cpu_to_be32(attr->rq_psn); in __mlx4_ib_modify_qp()
1653 context->qkey = cpu_to_be32(IB_QP_SET_QKEY); in __mlx4_ib_modify_qp()
1665 context->qkey = cpu_to_be32(attr->qkey); in __mlx4_ib_modify_qp()
1671 context->srqn = cpu_to_be32(1 << 24 | to_msrq(ibqp->srq)->msrq.srqn); in __mlx4_ib_modify_qp()
1718 if (!(context->flags & cpu_to_be32(1 << MLX4_RSS_QPC_FLAG_OFFSET))) in __mlx4_ib_modify_qp()
1719 context->srqn = cpu_to_be32(7 << 28); in __mlx4_ib_modify_qp()
1755 ctrl->owner_opcode = cpu_to_be32(1 << 31); in __mlx4_ib_modify_qp()
2073 mlx->flags &= cpu_to_be32(MLX4_WQE_CTRL_CQ_UPDATE); in build_sriov_qp0_header()
2076 mlx->flags |= cpu_to_be32(MLX4_WQE_MLX_VL15 | 0x1 | MLX4_WQE_MLX_SLR); in build_sriov_qp0_header()
2084 sqp->ud_header.bth.destination_qpn = cpu_to_be32(wr->wr.ud.remote_qpn); in build_sriov_qp0_header()
2087 cpu_to_be32(mdev->dev->caps.qp0_tunnel[sqp->qp.port - 1]); in build_sriov_qp0_header()
2089 sqp->ud_header.bth.psn = cpu_to_be32((sqp->send_psn++) & ((1 << 24) - 1)); in build_sriov_qp0_header()
2097 sqp->ud_header.deth.qkey = cpu_to_be32(qkey); in build_sriov_qp0_header()
2098 sqp->ud_header.deth.source_qpn = cpu_to_be32(sqp->qp.mqp.qpn); in build_sriov_qp0_header()
2114 inl->byte_count = cpu_to_be32(1 << 31 | header_size); in build_sriov_qp0_header()
2118 inl->byte_count = cpu_to_be32(1 << 31 | spc); in build_sriov_qp0_header()
2137 inl->byte_count = cpu_to_be32(1 << 31 | (header_size - spc)); in build_sriov_qp0_header()
2218 ah->av.ib.sl_tclass_flowlabel & cpu_to_be32(0xfffff); in build_mlx_header()
2243 mlx->flags &= cpu_to_be32(MLX4_WQE_CTRL_CQ_UPDATE); in build_mlx_header()
2246 mlx->flags |= cpu_to_be32((!sqp->qp.ibqp.qp_num ? MLX4_WQE_MLX_VL15 : 0) | in build_mlx_header()
2250 if (ah->av.ib.port_pd & cpu_to_be32(0x80000000)) in build_mlx_header()
2251 mlx->flags |= cpu_to_be32(0x1); /* force loopback */ in build_mlx_header()
2294 mlx->flags |= cpu_to_be32(MLX4_WQE_CTRL_FORCE_LOOPBACK); in build_mlx_header()
2312 sqp->ud_header.bth.destination_qpn = cpu_to_be32(wr->wr.ud.remote_qpn); in build_mlx_header()
2313 sqp->ud_header.bth.psn = cpu_to_be32((sqp->send_psn++) & ((1 << 24) - 1)); in build_mlx_header()
2314 sqp->ud_header.deth.qkey = cpu_to_be32(wr->wr.ud.remote_qkey & 0x80000000 ? in build_mlx_header()
2316 sqp->ud_header.deth.source_qpn = cpu_to_be32(sqp->qp.ibqp.qp_num); in build_mlx_header()
2342 inl->byte_count = cpu_to_be32(1 << 31 | header_size); in build_mlx_header()
2346 inl->byte_count = cpu_to_be32(1 << 31 | spc); in build_mlx_header()
2365 inl->byte_count = cpu_to_be32(1 << 31 | (header_size - spc)); in build_mlx_header()
2394 cpu_to_be32(MLX4_WQE_FMR_AND_BIND_PERM_ATOMIC) : 0) | in convert_access()
2396 cpu_to_be32(MLX4_WQE_FMR_AND_BIND_PERM_REMOTE_WRITE) : 0) | in convert_access()
2398 cpu_to_be32(MLX4_WQE_FMR_AND_BIND_PERM_REMOTE_READ) : 0) | in convert_access()
2399 (acc & IB_ACCESS_LOCAL_WRITE ? cpu_to_be32(MLX4_WQE_FMR_PERM_LOCAL_WRITE) : 0) | in convert_access()
2400 cpu_to_be32(MLX4_WQE_FMR_PERM_LOCAL_READ); in convert_access()
2414 fseg->mem_key = cpu_to_be32(wr->wr.fast_reg.rkey); in set_fmr_seg()
2419 fseg->page_size = cpu_to_be32(wr->wr.fast_reg.page_shift); in set_fmr_seg()
2428 cpu_to_be32(MLX4_WQE_FMR_AND_BIND_PERM_REMOTE_READ | in set_bind_seg()
2433 bseg->flags2 |= cpu_to_be32(MLX4_WQE_BIND_TYPE_2); in set_bind_seg()
2435 bseg->flags2 |= cpu_to_be32(MLX4_WQE_BIND_ZERO_BASED); in set_bind_seg()
2436 bseg->new_rkey = cpu_to_be32(wr->wr.bind_mw.rkey); in set_bind_seg()
2437 bseg->lkey = cpu_to_be32(wr->wr.bind_mw.bind_info.mr->lkey); in set_bind_seg()
2445 iseg->mem_key = cpu_to_be32(rkey); in set_local_inv_seg()
2452 rseg->rkey = cpu_to_be32(rkey); in set_raddr_seg()
2484 dseg->dqpn = cpu_to_be32(wr->wr.ud.remote_qpn); in set_datagram_seg()
2485 dseg->qkey = cpu_to_be32(wr->wr.ud.remote_qkey); in set_datagram_seg()
2500 sqp_av.port_pd = av->ib.port_pd | cpu_to_be32(0x80000000); in set_tunnel_datagram_seg()
2503 cpu_to_be32(0xf0000000); in set_tunnel_datagram_seg()
2507 dseg->dqpn = cpu_to_be32(dev->dev->caps.qp1_tunnel[port - 1]); in set_tunnel_datagram_seg()
2509 dseg->dqpn = cpu_to_be32(dev->dev->caps.qp0_tunnel[port - 1]); in set_tunnel_datagram_seg()
2511 dseg->qkey = cpu_to_be32(IB_QP_SET_QKEY); in set_tunnel_datagram_seg()
2523 hdr.remote_qpn = cpu_to_be32(wr->wr.ud.remote_qpn); in build_tunnel_header()
2525 hdr.qkey = cpu_to_be32(wr->wr.ud.remote_qkey); in build_tunnel_header()
2534 inl->byte_count = cpu_to_be32(1 << 31 | sizeof (hdr)); in build_tunnel_header()
2539 inl->byte_count = cpu_to_be32(1 << 31 | spc); in build_tunnel_header()
2544 inl->byte_count = cpu_to_be32(1 << 31 | (sizeof (hdr) - spc)); in build_tunnel_header()
2569 iseg->byte_count = cpu_to_be32((1 << 31) | 4); in set_mlx_icrc_seg()
2574 dseg->lkey = cpu_to_be32(sg->lkey); in set_data_seg()
2587 dseg->byte_count = cpu_to_be32(sg->length); in set_data_seg()
2592 dseg->byte_count = cpu_to_be32(sg->length); in __set_data_seg()
2593 dseg->lkey = cpu_to_be32(sg->lkey); in __set_data_seg()
2604 *blh = cpu_to_be32(1 << 6); in build_lso_seg()
2612 *lso_hdr_sz = cpu_to_be32(wr->wr.ud.mss << 16 | wr->wr.ud.hlen); in build_lso_seg()
2625 return cpu_to_be32(wr->ex.invalidate_rkey); in send_ieth()
2636 inl->byte_count = cpu_to_be32(1 << 31); in add_zero_len_inline()
2691 cpu_to_be32(MLX4_WQE_CTRL_CQ_UPDATE) : 0) | in mlx4_ib_post_send()
2693 cpu_to_be32(MLX4_WQE_CTRL_SOLICITED) : 0) | in mlx4_ib_post_send()
2695 cpu_to_be32(MLX4_WQE_CTRL_IP_CSUM | in mlx4_ib_post_send()
2747 cpu_to_be32(MLX4_WQE_CTRL_STRONG_ORDER); in mlx4_ib_post_send()
2755 cpu_to_be32(MLX4_WQE_CTRL_STRONG_ORDER); in mlx4_ib_post_send()
2763 cpu_to_be32(MLX4_WQE_CTRL_STRONG_ORDER); in mlx4_ib_post_send()
2788 *(__be32 *) wqe |= cpu_to_be32(0x80000000); in mlx4_ib_post_send()
2903 (ind & qp->sq.wqe_cnt ? cpu_to_be32(1 << 31) : 0) | blh; in mlx4_ib_post_send()
3000 cpu_to_be32(sizeof (struct mlx4_ib_proxy_sqp_hdr)); in mlx4_ib_post_recv()
3002 scat->lkey = cpu_to_be32(wr->sg_list->lkey); in mlx4_ib_post_recv()
3013 scat[i].lkey = cpu_to_be32(MLX4_INVALID_LKEY); in mlx4_ib_post_recv()
3032 *qp->db.db = cpu_to_be32(qp->rq.head & 0xffff); in mlx4_ib_post_recv()
3210 qp->sq_signal_bits == cpu_to_be32(MLX4_WQE_CTRL_CQ_UPDATE) ? in mlx4_ib_query_qp()