num_sq_entries     65 drivers/net/ethernet/intel/ice/ice_controlq.c 			(cq->num_sq_entries | cq->sq.len_ena_mask);
num_sq_entries     78 drivers/net/ethernet/intel/ice/ice_controlq.c 	size_t size = cq->num_sq_entries * sizeof(struct ice_aq_desc);
num_sq_entries     87 drivers/net/ethernet/intel/ice/ice_controlq.c 	cq->sq.cmd_buf = devm_kcalloc(ice_hw_to_dev(hw), cq->num_sq_entries,
num_sq_entries    218 drivers/net/ethernet/intel/ice/ice_controlq.c 	cq->sq.dma_head = devm_kcalloc(ice_hw_to_dev(hw), cq->num_sq_entries,
num_sq_entries    225 drivers/net/ethernet/intel/ice/ice_controlq.c 	for (i = 0; i < cq->num_sq_entries; i++) {
num_sq_entries    282 drivers/net/ethernet/intel/ice/ice_controlq.c 	return ice_cfg_cq_regs(hw, &cq->sq, cq->num_sq_entries);
num_sq_entries    332 drivers/net/ethernet/intel/ice/ice_controlq.c 	if (!cq->num_sq_entries || !cq->sq_buf_size) {
num_sq_entries    356 drivers/net/ethernet/intel/ice/ice_controlq.c 	cq->sq.count = cq->num_sq_entries;
num_sq_entries    611 drivers/net/ethernet/intel/ice/ice_controlq.c 	if (!cq->num_rq_entries || !cq->num_sq_entries ||
num_sq_entries    888 drivers/net/ethernet/intel/ice/ice_controlq.c 	if (val >= cq->num_sq_entries) {
num_sq_entries     88 drivers/net/ethernet/intel/ice/ice_controlq.h 	u16 num_sq_entries;		/* send queue depth */
num_sq_entries   1544 drivers/net/ethernet/intel/ice/ice_main.c 	hw->adminq.num_sq_entries = ICE_AQ_LEN;
num_sq_entries   1548 drivers/net/ethernet/intel/ice/ice_main.c 	hw->mailboxq.num_sq_entries = ICE_MBXSQ_LEN;