cq_size 3198 drivers/infiniband/core/mad.c int ret, cq_size; cq_size 3222 drivers/infiniband/core/mad.c cq_size = mad_sendq_size + mad_recvq_size; cq_size 3225 drivers/infiniband/core/mad.c cq_size *= 2; cq_size 3234 drivers/infiniband/core/mad.c port_priv->cq = ib_alloc_cq(port_priv->device, port_priv, cq_size, 0, cq_size 1954 drivers/infiniband/hw/bnxt_re/qplib_fp.c req.cq_size = cpu_to_le32(cq->hwq.max_elements); cq_size 1314 drivers/infiniband/hw/bnxt_re/roce_hsi.h __le32 cq_size; cq_size 880 drivers/infiniband/hw/cxgb4/device.c rdev->status_page->cq_size = rdev->lldi.vr->cq.size; cq_size 865 drivers/infiniband/hw/cxgb4/t4.h u64 cq_size; cq_size 1957 drivers/infiniband/hw/i40iw/i40iw_ctrl.c cq->cq_uk.cq_size = info->num_elem; cq_size 2013 drivers/infiniband/hw/i40iw/i40iw_ctrl.c set_64bit_val(wqe, 0, ccq->cq_uk.cq_size); cq_size 2072 drivers/infiniband/hw/i40iw/i40iw_ctrl.c set_64bit_val(wqe, 0, ccq->cq_uk.cq_size); cq_size 2177 drivers/infiniband/hw/i40iw/i40iw_ctrl.c set_64bit_val(wqe, 0, cq->cq_uk.cq_size); cq_size 2229 drivers/infiniband/hw/i40iw/i40iw_ctrl.c set_64bit_val(wqe, 0, cq->cq_uk.cq_size); cq_size 2270 drivers/infiniband/hw/i40iw/i40iw_ctrl.c u32 cq_size, ceq_id, first_pm_pbl_idx; cq_size 2293 drivers/infiniband/hw/i40iw/i40iw_ctrl.c cq_size = info->cq_resize ? info->cq_size : cq->cq_uk.cq_size; cq_size 2310 drivers/infiniband/hw/i40iw/i40iw_ctrl.c cq->cq_uk.cq_size = cq_size; cq_size 2318 drivers/infiniband/hw/i40iw/i40iw_ctrl.c set_64bit_val(wqe, 0, cq_size); cq_size 641 drivers/infiniband/hw/i40iw/i40iw_puda.c set_64bit_val(wqe, 0, cq->cq_uk.cq_size); cq_size 683 drivers/infiniband/hw/i40iw/i40iw_puda.c cqsize = rsrc->cq_size * (sizeof(struct i40iw_cqe)); cq_size 695 drivers/infiniband/hw/i40iw/i40iw_puda.c info.shadow_read_threshold = rsrc->cq_size >> 2; cq_size 701 drivers/infiniband/hw/i40iw/i40iw_puda.c init_info->cq_size = rsrc->cq_size; cq_size 939 drivers/infiniband/hw/i40iw/i40iw_puda.c rsrc->cq_size = info->rq_size + info->sq_size; cq_size 131 drivers/infiniband/hw/i40iw/i40iw_puda.h u32 cq_size; cq_size 521 drivers/infiniband/hw/i40iw/i40iw_type.h u32 cq_size; cq_size 1059 drivers/infiniband/hw/i40iw/i40iw_uk.c if ((info->cq_size < I40IW_MIN_CQ_SIZE) || cq_size 1060 drivers/infiniband/hw/i40iw/i40iw_uk.c (info->cq_size > I40IW_MAX_CQ_SIZE)) cq_size 1064 drivers/infiniband/hw/i40iw/i40iw_uk.c cq->cq_size = info->cq_size; cq_size 1069 drivers/infiniband/hw/i40iw/i40iw_uk.c I40IW_RING_INIT(cq->cq_ring, cq->cq_size); cq_size 369 drivers/infiniband/hw/i40iw/i40iw_user.h u32 cq_size; cq_size 399 drivers/infiniband/hw/i40iw/i40iw_user.h u32 cq_size; cq_size 1115 drivers/infiniband/hw/i40iw/i40iw_verbs.c ukinfo->cq_size = max(entries, 4); cq_size 1117 drivers/infiniband/hw/i40iw/i40iw_verbs.c iwcq->ibcq.cqe = info.cq_uk_init_info.cq_size; cq_size 1162 drivers/infiniband/hw/i40iw/i40iw_verbs.c rsize = info.cq_uk_init_info.cq_size * sizeof(struct i40iw_cqe); cq_size 1206 drivers/infiniband/hw/i40iw/i40iw_verbs.c resp.cq_size = info.cq_uk_init_info.cq_size; cq_size 109 drivers/infiniband/hw/i40iw/i40iw_verbs.h u16 cq_size; cq_size 1981 drivers/infiniband/hw/mlx4/mad.c int ret, cq_size; cq_size 2007 drivers/infiniband/hw/mlx4/mad.c cq_size = 2 * MLX4_NUM_TUNNEL_BUFS; cq_size 2009 drivers/infiniband/hw/mlx4/mad.c cq_size *= 2; cq_size 2011 drivers/infiniband/hw/mlx4/mad.c cq_attr.cqe = cq_size; cq_size 760 drivers/infiniband/hw/qedr/verbs.c params->cq_size = chain_entries - 1; cq_size 920 drivers/infiniband/hw/qedr/verbs.c cq->icid, cq, params.cq_size); cq_size 2138 drivers/net/ethernet/brocade/bna/bna_tx_rx.c u32 cq_size, hq_size, dq_size; cq_size 2149 drivers/net/ethernet/brocade/bna/bna_tx_rx.c cq_size = cq_depth * BFI_CQ_WI_SIZE; cq_size 2150 drivers/net/ethernet/brocade/bna/bna_tx_rx.c cq_size = ALIGN(cq_size, PAGE_SIZE); cq_size 2151 drivers/net/ethernet/brocade/bna/bna_tx_rx.c cpage_count = SIZE_TO_PAGES(cq_size); cq_size 428 drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c static int mlx5_fpga_conn_create_cq(struct mlx5_fpga_conn *conn, int cq_size) cq_size 442 drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c cq_size = roundup_pow_of_two(cq_size); cq_size 443 drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c MLX5_SET(cqc, temp_cqc, log_cq_size, ilog2(cq_size)); cq_size 473 drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c MLX5_SET(cqc, cqc, log_cq_size, ilog2(cq_size)); cq_size 866 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_send.c int cq_size; cq_size 874 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_send.c cq_size = QUEUE_SIZE + 1; cq_size 875 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_send.c dmn->send_ring->cq = dr_create_cq(dmn->mdev, dmn->uar, cq_size); cq_size 342 drivers/net/ethernet/pensando/ionic/ionic_lif.c u32 q_size, cq_size, sg_size, total_size; cq_size 354 drivers/net/ethernet/pensando/ionic/ionic_lif.c cq_size = num_descs * cq_desc_size; cq_size 357 drivers/net/ethernet/pensando/ionic/ionic_lif.c total_size = ALIGN(q_size, PAGE_SIZE) + ALIGN(cq_size, PAGE_SIZE); cq_size 451 drivers/net/ethernet/pensando/ionic/ionic_lif.c sg_base = (void *)ALIGN((uintptr_t)cq_base + cq_size, cq_size 453 drivers/net/ethernet/pensando/ionic/ionic_lif.c sg_base_pa = ALIGN(cq_base_pa + cq_size, PAGE_SIZE); cq_size 1065 drivers/net/ethernet/qlogic/qed/qed_rdma.c p_ramrod->max_cqes = cpu_to_le32(params->cq_size); cq_size 1424 drivers/staging/qlge/qlge.h u32 cq_size; cq_size 1747 drivers/staging/qlge/qlge_dbg.c pr_err("rx_ring->cq_size = %d\n", rx_ring->cq_size); cq_size 2971 drivers/staging/qlge/qlge_main.c rx_ring->cq_size, cq_size 2987 drivers/staging/qlge/qlge_main.c pci_alloc_consistent(qdev->pdev, rx_ring->cq_size, cq_size 4149 drivers/staging/qlge/qlge_main.c rx_ring->cq_size = cq_size 4166 drivers/staging/qlge/qlge_main.c rx_ring->cq_size = cq_size 278 include/linux/qed/qed_rdma_if.h u32 cq_size; cq_size 92 include/uapi/rdma/i40iw-abi.h __u32 cq_size;