num_sq_entries 65 drivers/net/ethernet/intel/ice/ice_controlq.c (cq->num_sq_entries | cq->sq.len_ena_mask); num_sq_entries 78 drivers/net/ethernet/intel/ice/ice_controlq.c size_t size = cq->num_sq_entries * sizeof(struct ice_aq_desc); num_sq_entries 87 drivers/net/ethernet/intel/ice/ice_controlq.c cq->sq.cmd_buf = devm_kcalloc(ice_hw_to_dev(hw), cq->num_sq_entries, num_sq_entries 218 drivers/net/ethernet/intel/ice/ice_controlq.c cq->sq.dma_head = devm_kcalloc(ice_hw_to_dev(hw), cq->num_sq_entries, num_sq_entries 225 drivers/net/ethernet/intel/ice/ice_controlq.c for (i = 0; i < cq->num_sq_entries; i++) { num_sq_entries 282 drivers/net/ethernet/intel/ice/ice_controlq.c return ice_cfg_cq_regs(hw, &cq->sq, cq->num_sq_entries); num_sq_entries 332 drivers/net/ethernet/intel/ice/ice_controlq.c if (!cq->num_sq_entries || !cq->sq_buf_size) { num_sq_entries 356 drivers/net/ethernet/intel/ice/ice_controlq.c cq->sq.count = cq->num_sq_entries; num_sq_entries 611 drivers/net/ethernet/intel/ice/ice_controlq.c if (!cq->num_rq_entries || !cq->num_sq_entries || num_sq_entries 888 drivers/net/ethernet/intel/ice/ice_controlq.c if (val >= cq->num_sq_entries) { num_sq_entries 88 drivers/net/ethernet/intel/ice/ice_controlq.h u16 num_sq_entries; /* send queue depth */ num_sq_entries 1544 drivers/net/ethernet/intel/ice/ice_main.c hw->adminq.num_sq_entries = ICE_AQ_LEN; num_sq_entries 1548 drivers/net/ethernet/intel/ice/ice_main.c hw->mailboxq.num_sq_entries = ICE_MBXSQ_LEN;