qi                224 arch/um/drivers/vector_kern.c static int vector_advancehead(struct vector_queue *qi, int advance)
qi                228 arch/um/drivers/vector_kern.c 	qi->head =
qi                229 arch/um/drivers/vector_kern.c 		(qi->head + advance)
qi                230 arch/um/drivers/vector_kern.c 			% qi->max_depth;
qi                233 arch/um/drivers/vector_kern.c 	spin_lock(&qi->tail_lock);
qi                234 arch/um/drivers/vector_kern.c 	qi->queue_depth -= advance;
qi                240 arch/um/drivers/vector_kern.c 	if (qi->queue_depth == 0) {
qi                241 arch/um/drivers/vector_kern.c 		qi->head = 0;
qi                242 arch/um/drivers/vector_kern.c 		qi->tail = 0;
qi                244 arch/um/drivers/vector_kern.c 	queue_depth = qi->queue_depth;
qi                245 arch/um/drivers/vector_kern.c 	spin_unlock(&qi->tail_lock);
qi                254 arch/um/drivers/vector_kern.c static int vector_advancetail(struct vector_queue *qi, int advance)
qi                258 arch/um/drivers/vector_kern.c 	qi->tail =
qi                259 arch/um/drivers/vector_kern.c 		(qi->tail + advance)
qi                260 arch/um/drivers/vector_kern.c 			% qi->max_depth;
qi                261 arch/um/drivers/vector_kern.c 	spin_lock(&qi->head_lock);
qi                262 arch/um/drivers/vector_kern.c 	qi->queue_depth += advance;
qi                263 arch/um/drivers/vector_kern.c 	queue_depth = qi->queue_depth;
qi                264 arch/um/drivers/vector_kern.c 	spin_unlock(&qi->head_lock);
qi                309 arch/um/drivers/vector_kern.c static int vector_enqueue(struct vector_queue *qi, struct sk_buff *skb)
qi                311 arch/um/drivers/vector_kern.c 	struct vector_private *vp = netdev_priv(qi->dev);
qi                314 arch/um/drivers/vector_kern.c 	struct mmsghdr *mmsg_vector = qi->mmsg_vector;
qi                317 arch/um/drivers/vector_kern.c 	spin_lock(&qi->tail_lock);
qi                318 arch/um/drivers/vector_kern.c 	spin_lock(&qi->head_lock);
qi                319 arch/um/drivers/vector_kern.c 	queue_depth = qi->queue_depth;
qi                320 arch/um/drivers/vector_kern.c 	spin_unlock(&qi->head_lock);
qi                325 arch/um/drivers/vector_kern.c 	if (queue_depth < qi->max_depth) {
qi                327 arch/um/drivers/vector_kern.c 		*(qi->skbuff_vector + qi->tail) = skb;
qi                328 arch/um/drivers/vector_kern.c 		mmsg_vector += qi->tail;
qi                339 arch/um/drivers/vector_kern.c 		queue_depth = vector_advancetail(qi, 1);
qi                342 arch/um/drivers/vector_kern.c 	spin_unlock(&qi->tail_lock);
qi                345 arch/um/drivers/vector_kern.c 	qi->dev->stats.tx_dropped++;
qi                349 arch/um/drivers/vector_kern.c 		netdev_completed_queue(qi->dev, 1, packet_len);
qi                351 arch/um/drivers/vector_kern.c 	spin_unlock(&qi->tail_lock);
qi                355 arch/um/drivers/vector_kern.c static int consume_vector_skbs(struct vector_queue *qi, int count)
qi                361 arch/um/drivers/vector_kern.c 	for (skb_index = qi->head; skb_index < qi->head + count; skb_index++) {
qi                362 arch/um/drivers/vector_kern.c 		skb = *(qi->skbuff_vector + skb_index);
qi                367 arch/um/drivers/vector_kern.c 		*(qi->skbuff_vector + skb_index) = NULL;
qi                370 arch/um/drivers/vector_kern.c 	qi->dev->stats.tx_bytes += bytes_compl;
qi                371 arch/um/drivers/vector_kern.c 	qi->dev->stats.tx_packets += count;
qi                372 arch/um/drivers/vector_kern.c 	netdev_completed_queue(qi->dev, count, bytes_compl);
qi                373 arch/um/drivers/vector_kern.c 	return vector_advancehead(qi, count);
qi                383 arch/um/drivers/vector_kern.c static int vector_send(struct vector_queue *qi)
qi                385 arch/um/drivers/vector_kern.c 	struct vector_private *vp = netdev_priv(qi->dev);
qi                387 arch/um/drivers/vector_kern.c 	int result = 0, send_len, queue_depth = qi->max_depth;
qi                389 arch/um/drivers/vector_kern.c 	if (spin_trylock(&qi->head_lock)) {
qi                390 arch/um/drivers/vector_kern.c 		if (spin_trylock(&qi->tail_lock)) {
qi                392 arch/um/drivers/vector_kern.c 			queue_depth = qi->queue_depth;
qi                393 arch/um/drivers/vector_kern.c 			spin_unlock(&qi->tail_lock);
qi                397 arch/um/drivers/vector_kern.c 				send_from = qi->mmsg_vector;
qi                398 arch/um/drivers/vector_kern.c 				send_from += qi->head;
qi                400 arch/um/drivers/vector_kern.c 				if (send_len + qi->head > qi->max_depth)
qi                401 arch/um/drivers/vector_kern.c 					send_len = qi->max_depth - qi->head;
qi                427 arch/um/drivers/vector_kern.c 						consume_vector_skbs(qi, result);
qi                437 arch/um/drivers/vector_kern.c 				netif_trans_update(qi->dev);
qi                438 arch/um/drivers/vector_kern.c 				netif_wake_queue(qi->dev);
qi                448 arch/um/drivers/vector_kern.c 		spin_unlock(&qi->head_lock);
qi                459 arch/um/drivers/vector_kern.c static void destroy_queue(struct vector_queue *qi)
qi                463 arch/um/drivers/vector_kern.c 	struct vector_private *vp = netdev_priv(qi->dev);
qi                466 arch/um/drivers/vector_kern.c 	if (qi == NULL)
qi                471 arch/um/drivers/vector_kern.c 	if (qi->skbuff_vector != NULL) {
qi                472 arch/um/drivers/vector_kern.c 		for (i = 0; i < qi->max_depth; i++) {
qi                473 arch/um/drivers/vector_kern.c 			if (*(qi->skbuff_vector + i) != NULL)
qi                474 arch/um/drivers/vector_kern.c 				dev_kfree_skb_any(*(qi->skbuff_vector + i));
qi                476 arch/um/drivers/vector_kern.c 		kfree(qi->skbuff_vector);
qi                479 arch/um/drivers/vector_kern.c 	if (qi->mmsg_vector != NULL) {
qi                480 arch/um/drivers/vector_kern.c 		mmsg_vector = qi->mmsg_vector;
qi                481 arch/um/drivers/vector_kern.c 		for (i = 0; i < qi->max_depth; i++) {
qi                491 arch/um/drivers/vector_kern.c 		kfree(qi->mmsg_vector);
qi                493 arch/um/drivers/vector_kern.c 	kfree(qi);
qi                652 arch/um/drivers/vector_kern.c static void prep_queue_for_rx(struct vector_queue *qi)
qi                654 arch/um/drivers/vector_kern.c 	struct vector_private *vp = netdev_priv(qi->dev);
qi                655 arch/um/drivers/vector_kern.c 	struct mmsghdr *mmsg_vector = qi->mmsg_vector;
qi                656 arch/um/drivers/vector_kern.c 	void **skbuff_vector = qi->skbuff_vector;
qi                659 arch/um/drivers/vector_kern.c 	if (qi->queue_depth == 0)
qi                661 arch/um/drivers/vector_kern.c 	for (i = 0; i < qi->queue_depth; i++) {
qi                671 arch/um/drivers/vector_kern.c 	qi->queue_depth = 0;
qi                937 arch/um/drivers/vector_kern.c 	struct vector_queue *qi = vp->rx_queue;
qi                939 arch/um/drivers/vector_kern.c 	struct mmsghdr *mmsg_vector = qi->mmsg_vector;
qi                940 arch/um/drivers/vector_kern.c 	void **skbuff_vector = qi->skbuff_vector;
qi                947 arch/um/drivers/vector_kern.c 	prep_queue_for_rx(qi);
qi                952 arch/um/drivers/vector_kern.c 		vp->fds->rx_fd, qi->mmsg_vector, qi->max_depth, 0);
qi                965 arch/um/drivers/vector_kern.c 	qi->queue_depth = packet_count;
qi                743 drivers/crypto/caam/ctrl.c 		ctrlpriv->qi = (struct caam_queue_if __iomem __force *)
qi                748 drivers/crypto/caam/ctrl.c 		wr_reg32(&ctrlpriv->qi->qi_control_lo, QICTL_DQEN);
qi                 73 drivers/crypto/caam/intern.h 	struct caam_queue_if __iomem *qi; /* QI control region */
qi                298 drivers/crypto/caam/regs.h 	u32 qi;		/* QI_VERSION */
qi                 26 drivers/gpu/drm/i915/display/intel_bw.c 					  struct intel_qgv_info *qi)
qi                 40 drivers/gpu/drm/i915/display/intel_bw.c 		qi->dram_type = INTEL_DRAM_DDR4;
qi                 43 drivers/gpu/drm/i915/display/intel_bw.c 		qi->dram_type = INTEL_DRAM_DDR3;
qi                 46 drivers/gpu/drm/i915/display/intel_bw.c 		qi->dram_type = INTEL_DRAM_LPDDR3;
qi                 49 drivers/gpu/drm/i915/display/intel_bw.c 		qi->dram_type = INTEL_DRAM_LPDDR3;
qi                 56 drivers/gpu/drm/i915/display/intel_bw.c 	qi->num_channels = (val & 0xf0) >> 4;
qi                 57 drivers/gpu/drm/i915/display/intel_bw.c 	qi->num_points = (val & 0xf00) >> 8;
qi                 59 drivers/gpu/drm/i915/display/intel_bw.c 	qi->t_bl = qi->dram_type == INTEL_DRAM_DDR4 ? 4 : 8;
qi                 91 drivers/gpu/drm/i915/display/intel_bw.c 			      struct intel_qgv_info *qi)
qi                 95 drivers/gpu/drm/i915/display/intel_bw.c 	ret = icl_pcode_read_mem_global_info(dev_priv, qi);
qi                 99 drivers/gpu/drm/i915/display/intel_bw.c 	if (WARN_ON(qi->num_points > ARRAY_SIZE(qi->points)))
qi                100 drivers/gpu/drm/i915/display/intel_bw.c 		qi->num_points = ARRAY_SIZE(qi->points);
qi                102 drivers/gpu/drm/i915/display/intel_bw.c 	for (i = 0; i < qi->num_points; i++) {
qi                103 drivers/gpu/drm/i915/display/intel_bw.c 		struct intel_qgv_point *sp = &qi->points[i];
qi                123 drivers/gpu/drm/i915/display/intel_bw.c static int icl_sagv_max_dclk(const struct intel_qgv_info *qi)
qi                128 drivers/gpu/drm/i915/display/intel_bw.c 	for (i = 0; i < qi->num_points; i++)
qi                129 drivers/gpu/drm/i915/display/intel_bw.c 		dclk = max(dclk, qi->points[i].dclk);
qi                147 drivers/gpu/drm/i915/display/intel_bw.c 	struct intel_qgv_info qi = {};
qi                157 drivers/gpu/drm/i915/display/intel_bw.c 	ret = icl_get_qgv_points(dev_priv, &qi);
qi                162 drivers/gpu/drm/i915/display/intel_bw.c 	num_channels = qi.num_channels;
qi                165 drivers/gpu/drm/i915/display/intel_bw.c 	dclk_max = icl_sagv_max_dclk(&qi);
qi                181 drivers/gpu/drm/i915/display/intel_bw.c 		bi->num_qgv_points = qi.num_points;
qi                183 drivers/gpu/drm/i915/display/intel_bw.c 		for (j = 0; j < qi.num_points; j++) {
qi                184 drivers/gpu/drm/i915/display/intel_bw.c 			const struct intel_qgv_point *sp = &qi.points[j];
qi                194 drivers/gpu/drm/i915/display/intel_bw.c 				   (clpchgroup - 1) * qi.t_bl + sp->t_rdpre);
qi               1130 drivers/iommu/dmar.c 	if (iommu->qi) {
qi               1131 drivers/iommu/dmar.c 		free_page((unsigned long)iommu->qi->desc);
qi               1132 drivers/iommu/dmar.c 		kfree(iommu->qi->desc_status);
qi               1133 drivers/iommu/dmar.c 		kfree(iommu->qi);
qi               1146 drivers/iommu/dmar.c static inline void reclaim_free_desc(struct q_inval *qi)
qi               1148 drivers/iommu/dmar.c 	while (qi->desc_status[qi->free_tail] == QI_DONE ||
qi               1149 drivers/iommu/dmar.c 	       qi->desc_status[qi->free_tail] == QI_ABORT) {
qi               1150 drivers/iommu/dmar.c 		qi->desc_status[qi->free_tail] = QI_FREE;
qi               1151 drivers/iommu/dmar.c 		qi->free_tail = (qi->free_tail + 1) % QI_LENGTH;
qi               1152 drivers/iommu/dmar.c 		qi->free_cnt++;
qi               1160 drivers/iommu/dmar.c 	struct q_inval *qi = iommu->qi;
qi               1164 drivers/iommu/dmar.c 	if (qi->desc_status[wait_index] == QI_ABORT)
qi               1177 drivers/iommu/dmar.c 			struct qi_desc *desc = qi->desc + head;
qi               1187 drivers/iommu/dmar.c 			memcpy(desc, qi->desc + (wait_index << shift),
qi               1208 drivers/iommu/dmar.c 			if (qi->desc_status[head] == QI_IN_USE)
qi               1209 drivers/iommu/dmar.c 				qi->desc_status[head] = QI_ABORT;
qi               1213 drivers/iommu/dmar.c 		if (qi->desc_status[wait_index] == QI_ABORT)
qi               1230 drivers/iommu/dmar.c 	struct q_inval *qi = iommu->qi;
qi               1236 drivers/iommu/dmar.c 	if (!qi)
qi               1242 drivers/iommu/dmar.c 	raw_spin_lock_irqsave(&qi->q_lock, flags);
qi               1243 drivers/iommu/dmar.c 	while (qi->free_cnt < 3) {
qi               1244 drivers/iommu/dmar.c 		raw_spin_unlock_irqrestore(&qi->q_lock, flags);
qi               1246 drivers/iommu/dmar.c 		raw_spin_lock_irqsave(&qi->q_lock, flags);
qi               1249 drivers/iommu/dmar.c 	index = qi->free_head;
qi               1254 drivers/iommu/dmar.c 	qi->desc_status[index] = qi->desc_status[wait_index] = QI_IN_USE;
qi               1257 drivers/iommu/dmar.c 	memcpy(qi->desc + offset, desc, length);
qi               1260 drivers/iommu/dmar.c 	wait_desc.qw1 = virt_to_phys(&qi->desc_status[wait_index]);
qi               1265 drivers/iommu/dmar.c 	memcpy(qi->desc + offset, &wait_desc, length);
qi               1267 drivers/iommu/dmar.c 	qi->free_head = (qi->free_head + 2) % QI_LENGTH;
qi               1268 drivers/iommu/dmar.c 	qi->free_cnt -= 2;
qi               1274 drivers/iommu/dmar.c 	writel(qi->free_head << shift, iommu->reg + DMAR_IQT_REG);
qi               1276 drivers/iommu/dmar.c 	while (qi->desc_status[wait_index] != QI_DONE) {
qi               1288 drivers/iommu/dmar.c 		raw_spin_unlock(&qi->q_lock);
qi               1290 drivers/iommu/dmar.c 		raw_spin_lock(&qi->q_lock);
qi               1293 drivers/iommu/dmar.c 	qi->desc_status[index] = QI_DONE;
qi               1295 drivers/iommu/dmar.c 	reclaim_free_desc(qi);
qi               1296 drivers/iommu/dmar.c 	raw_spin_unlock_irqrestore(&qi->q_lock, flags);
qi               1422 drivers/iommu/dmar.c 	struct q_inval *qi = iommu->qi;
qi               1423 drivers/iommu/dmar.c 	u64 val = virt_to_phys(qi->desc);
qi               1425 drivers/iommu/dmar.c 	qi->free_head = qi->free_tail = 0;
qi               1426 drivers/iommu/dmar.c 	qi->free_cnt = QI_LENGTH;
qi               1458 drivers/iommu/dmar.c 	struct q_inval *qi;
qi               1467 drivers/iommu/dmar.c 	if (iommu->qi)
qi               1470 drivers/iommu/dmar.c 	iommu->qi = kmalloc(sizeof(*qi), GFP_ATOMIC);
qi               1471 drivers/iommu/dmar.c 	if (!iommu->qi)
qi               1474 drivers/iommu/dmar.c 	qi = iommu->qi;
qi               1483 drivers/iommu/dmar.c 		kfree(qi);
qi               1484 drivers/iommu/dmar.c 		iommu->qi = NULL;
qi               1488 drivers/iommu/dmar.c 	qi->desc = page_address(desc_page);
qi               1490 drivers/iommu/dmar.c 	qi->desc_status = kcalloc(QI_LENGTH, sizeof(int), GFP_ATOMIC);
qi               1491 drivers/iommu/dmar.c 	if (!qi->desc_status) {
qi               1492 drivers/iommu/dmar.c 		free_page((unsigned long) qi->desc);
qi               1493 drivers/iommu/dmar.c 		kfree(qi);
qi               1494 drivers/iommu/dmar.c 		iommu->qi = NULL;
qi               1498 drivers/iommu/dmar.c 	raw_spin_lock_init(&qi->q_lock);
qi               1847 drivers/iommu/dmar.c 	if (!iommu->qi)
qi               1350 drivers/iommu/intel-iommu.c 	if (!iommu->qi)
qi               2963 drivers/iommu/intel-iommu.c 	if (!iommu->qi) {
qi               4175 drivers/iommu/intel-iommu.c 		if (iommu->qi)
qi                584 drivers/iommu/intel_irq_remapping.c 	if (!iommu->qi) {
qi               1045 drivers/iommu/intel_irq_remapping.c 		if (iommu->qi)
qi                425 drivers/net/ethernet/intel/ice/ice_controlq.c #define ICE_FREE_CQ_BUFS(hw, qi, ring)					\
qi                429 drivers/net/ethernet/intel/ice/ice_controlq.c 	for (i = 0; i < (qi)->num_##ring##_entries; i++)		\
qi                430 drivers/net/ethernet/intel/ice/ice_controlq.c 		if ((qi)->ring.r.ring##_bi[i].pa) {			\
qi                432 drivers/net/ethernet/intel/ice/ice_controlq.c 					   (qi)->ring.r.ring##_bi[i].size,\
qi                433 drivers/net/ethernet/intel/ice/ice_controlq.c 					   (qi)->ring.r.ring##_bi[i].va,\
qi                434 drivers/net/ethernet/intel/ice/ice_controlq.c 					   (qi)->ring.r.ring##_bi[i].pa);\
qi                435 drivers/net/ethernet/intel/ice/ice_controlq.c 			(qi)->ring.r.ring##_bi[i].va = NULL;		\
qi                436 drivers/net/ethernet/intel/ice/ice_controlq.c 			(qi)->ring.r.ring##_bi[i].pa = 0;		\
qi                437 drivers/net/ethernet/intel/ice/ice_controlq.c 			(qi)->ring.r.ring##_bi[i].size = 0;		\
qi                440 drivers/net/ethernet/intel/ice/ice_controlq.c 	if ((qi)->ring.cmd_buf)						\
qi                441 drivers/net/ethernet/intel/ice/ice_controlq.c 		devm_kfree(ice_hw_to_dev(hw), (qi)->ring.cmd_buf);	\
qi                443 drivers/net/ethernet/intel/ice/ice_controlq.c 	devm_kfree(ice_hw_to_dev(hw), (qi)->ring.dma_head);		\
qi                303 drivers/net/ethernet/pensando/ionic/ionic_txrx.c 	unsigned int qi = rxcq->bound_q->index;
qi                312 drivers/net/ethernet/pensando/ionic/ionic_txrx.c 	txcq = &lif->txqcqs[qi].qcq->cq;
qi                978 drivers/net/wireless/ath/ath5k/base.c 	struct ath5k_txq_info qi = {
qi               1000 drivers/net/wireless/ath/ath5k/base.c 	qi.tqi_flags = AR5K_TXQ_FLAG_TXEOLINT_ENABLE |
qi               1002 drivers/net/wireless/ath/ath5k/base.c 	qnum = ath5k_hw_setup_tx_queue(ah, qtype, &qi);
qi               1028 drivers/net/wireless/ath/ath5k/base.c 	struct ath5k_txq_info qi = {
qi               1038 drivers/net/wireless/ath/ath5k/base.c 	return ath5k_hw_setup_tx_queue(ah, AR5K_TX_QUEUE_BEACON, &qi);
qi               1044 drivers/net/wireless/ath/ath5k/base.c 	struct ath5k_txq_info qi;
qi               1047 drivers/net/wireless/ath/ath5k/base.c 	ret = ath5k_hw_get_tx_queueprops(ah, ah->bhalq, &qi);
qi               1057 drivers/net/wireless/ath/ath5k/base.c 		qi.tqi_aifs = 0;
qi               1058 drivers/net/wireless/ath/ath5k/base.c 		qi.tqi_cw_min = 0;
qi               1059 drivers/net/wireless/ath/ath5k/base.c 		qi.tqi_cw_max = 0;
qi               1064 drivers/net/wireless/ath/ath5k/base.c 		qi.tqi_aifs = 0;
qi               1065 drivers/net/wireless/ath/ath5k/base.c 		qi.tqi_cw_min = 0;
qi               1066 drivers/net/wireless/ath/ath5k/base.c 		qi.tqi_cw_max = 2 * AR5K_TUNE_CWMIN;
qi               1071 drivers/net/wireless/ath/ath5k/base.c 		qi.tqi_aifs, qi.tqi_cw_min, qi.tqi_cw_max);
qi               1073 drivers/net/wireless/ath/ath5k/base.c 	ret = ath5k_hw_set_tx_queueprops(ah, ah->bhalq, &qi);
qi               1084 drivers/net/wireless/ath/ath5k/base.c 	ret = ath5k_hw_get_tx_queueprops(ah, AR5K_TX_QUEUE_ID_CAB, &qi);
qi               1088 drivers/net/wireless/ath/ath5k/base.c 	qi.tqi_ready_time = (ah->bintval * 80) / 100;
qi               1089 drivers/net/wireless/ath/ath5k/base.c 	ret = ath5k_hw_set_tx_queueprops(ah, AR5K_TX_QUEUE_ID_CAB, &qi);
qi                578 drivers/net/wireless/ath/ath5k/mac80211-ops.c 	struct ath5k_txq_info qi;
qi                586 drivers/net/wireless/ath/ath5k/mac80211-ops.c 	ath5k_hw_get_tx_queueprops(ah, queue, &qi);
qi                588 drivers/net/wireless/ath/ath5k/mac80211-ops.c 	qi.tqi_aifs = params->aifs;
qi                589 drivers/net/wireless/ath/ath5k/mac80211-ops.c 	qi.tqi_cw_min = params->cw_min;
qi                590 drivers/net/wireless/ath/ath5k/mac80211-ops.c 	qi.tqi_cw_max = params->cw_max;
qi                591 drivers/net/wireless/ath/ath5k/mac80211-ops.c 	qi.tqi_burst_time = params->txop * 32;
qi                599 drivers/net/wireless/ath/ath5k/mac80211-ops.c 	if (ath5k_hw_set_tx_queueprops(ah, queue, &qi)) {
qi                157 drivers/net/wireless/ath/ath5k/qcu.c 	struct ath5k_txq_info *qi;
qi                161 drivers/net/wireless/ath/ath5k/qcu.c 	qi = &ah->ah_txq[queue];
qi                163 drivers/net/wireless/ath/ath5k/qcu.c 	if (qi->tqi_type == AR5K_TX_QUEUE_INACTIVE)
qi                167 drivers/net/wireless/ath/ath5k/qcu.c 	qi->tqi_type = qinfo->tqi_type;
qi                168 drivers/net/wireless/ath/ath5k/qcu.c 	qi->tqi_subtype = qinfo->tqi_subtype;
qi                169 drivers/net/wireless/ath/ath5k/qcu.c 	qi->tqi_flags = qinfo->tqi_flags;
qi                175 drivers/net/wireless/ath/ath5k/qcu.c 	qi->tqi_aifs = min(qinfo->tqi_aifs, (u8)0xFC);
qi                176 drivers/net/wireless/ath/ath5k/qcu.c 	qi->tqi_cw_min = ath5k_cw_validate(qinfo->tqi_cw_min);
qi                177 drivers/net/wireless/ath/ath5k/qcu.c 	qi->tqi_cw_max = ath5k_cw_validate(qinfo->tqi_cw_max);
qi                178 drivers/net/wireless/ath/ath5k/qcu.c 	qi->tqi_cbr_period = qinfo->tqi_cbr_period;
qi                179 drivers/net/wireless/ath/ath5k/qcu.c 	qi->tqi_cbr_overflow_limit = qinfo->tqi_cbr_overflow_limit;
qi                180 drivers/net/wireless/ath/ath5k/qcu.c 	qi->tqi_burst_time = qinfo->tqi_burst_time;
qi                181 drivers/net/wireless/ath/ath5k/qcu.c 	qi->tqi_ready_time = qinfo->tqi_ready_time;
qi                189 drivers/net/wireless/ath/ath5k/qcu.c 		qi->tqi_flags |= AR5K_TXQ_FLAG_POST_FR_BKOFF_DIS;
qi                 37 drivers/net/wireless/ath/ath9k/beacon.c 	struct ath9k_tx_queue_info qi, qi_be;
qi                 40 drivers/net/wireless/ath/ath9k/beacon.c 	ath9k_hw_get_txq_props(ah, sc->beacon.beaconq, &qi);
qi                 45 drivers/net/wireless/ath/ath9k/beacon.c 		qi.tqi_aifs = 1;
qi                 46 drivers/net/wireless/ath/ath9k/beacon.c 		qi.tqi_cwmin = 0;
qi                 47 drivers/net/wireless/ath/ath9k/beacon.c 		qi.tqi_cwmax = 0;
qi                 52 drivers/net/wireless/ath/ath9k/beacon.c 		qi.tqi_aifs = qi_be.tqi_aifs;
qi                 54 drivers/net/wireless/ath/ath9k/beacon.c 			qi.tqi_cwmin = 2*qi_be.tqi_cwmin;
qi                 56 drivers/net/wireless/ath/ath9k/beacon.c 			qi.tqi_cwmin = 4*qi_be.tqi_cwmin;
qi                 57 drivers/net/wireless/ath/ath9k/beacon.c 		qi.tqi_cwmax = qi_be.tqi_cwmax;
qi                 60 drivers/net/wireless/ath/ath9k/beacon.c 	if (!ath9k_hw_set_txq_props(ah, sc->beacon.beaconq, &qi)) {
qi                 24 drivers/net/wireless/ath/ath9k/htc_drv_beacon.c 	struct ath9k_tx_queue_info qi, qi_be;
qi                 26 drivers/net/wireless/ath/ath9k/htc_drv_beacon.c 	memset(&qi, 0, sizeof(struct ath9k_tx_queue_info));
qi                 29 drivers/net/wireless/ath/ath9k/htc_drv_beacon.c 	ath9k_hw_get_txq_props(ah, priv->beacon.beaconq, &qi);
qi                 33 drivers/net/wireless/ath/ath9k/htc_drv_beacon.c 		qi.tqi_aifs = 1;
qi                 34 drivers/net/wireless/ath/ath9k/htc_drv_beacon.c 		qi.tqi_cwmin = 0;
qi                 35 drivers/net/wireless/ath/ath9k/htc_drv_beacon.c 		qi.tqi_cwmax = 0;
qi                 41 drivers/net/wireless/ath/ath9k/htc_drv_beacon.c 		qi.tqi_aifs = qi_be.tqi_aifs;
qi                 49 drivers/net/wireless/ath/ath9k/htc_drv_beacon.c 			qi.tqi_cwmin = 2*qi_be.tqi_cwmin;
qi                 51 drivers/net/wireless/ath/ath9k/htc_drv_beacon.c 			qi.tqi_cwmin = 4*qi_be.tqi_cwmin;
qi                 53 drivers/net/wireless/ath/ath9k/htc_drv_beacon.c 		qi.tqi_cwmax = qi_be.tqi_cwmax;
qi                 57 drivers/net/wireless/ath/ath9k/htc_drv_beacon.c 	if (!ath9k_hw_set_txq_props(ah, priv->beacon.beaconq, &qi)) {
qi               1376 drivers/net/wireless/ath/ath9k/htc_drv_main.c 	struct ath9k_tx_queue_info qi;
qi               1385 drivers/net/wireless/ath/ath9k/htc_drv_main.c 	memset(&qi, 0, sizeof(struct ath9k_tx_queue_info));
qi               1387 drivers/net/wireless/ath/ath9k/htc_drv_main.c 	qi.tqi_aifs = params->aifs;
qi               1388 drivers/net/wireless/ath/ath9k/htc_drv_main.c 	qi.tqi_cwmin = params->cw_min;
qi               1389 drivers/net/wireless/ath/ath9k/htc_drv_main.c 	qi.tqi_cwmax = params->cw_max;
qi               1390 drivers/net/wireless/ath/ath9k/htc_drv_main.c 	qi.tqi_burstTime = params->txop * 32;
qi               1399 drivers/net/wireless/ath/ath9k/htc_drv_main.c 	ret = ath_htc_txq_update(priv, qnum, &qi);
qi                 31 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 		qi.tqi_subtype = subtype_txq_to_hwq[subtype];	\
qi                 32 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 		qi.tqi_aifs = ATH9K_TXQ_USEDEFAULT;		\
qi                 33 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 		qi.tqi_cwmin = ATH9K_TXQ_USEDEFAULT;		\
qi                 34 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 		qi.tqi_cwmax = ATH9K_TXQ_USEDEFAULT;		\
qi                 35 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 		qi.tqi_physCompBuf = 0;				\
qi                 36 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 		qi.tqi_qflags = TXQ_FLAG_TXEOLINT_ENABLE |	\
qi                193 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	struct ath9k_tx_queue_info qi;
qi                195 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	ath9k_hw_get_txq_props(ah, qnum, &qi);
qi                197 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	qi.tqi_aifs = qinfo->tqi_aifs;
qi                198 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	qi.tqi_cwmin = qinfo->tqi_cwmin / 2; /* XXX */
qi                199 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	qi.tqi_cwmax = qinfo->tqi_cwmax;
qi                200 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	qi.tqi_burstTime = qinfo->tqi_burstTime;
qi                201 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	qi.tqi_readyTime = qinfo->tqi_readyTime;
qi                203 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	if (!ath9k_hw_set_txq_props(ah, qnum, &qi)) {
qi                823 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	struct ath9k_tx_queue_info qi;
qi                826 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	memset(&qi, 0, sizeof(qi));
qi                829 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	qnum = ath9k_hw_setuptxqueue(priv->ah, ATH9K_TX_QUEUE_DATA, &qi);
qi                846 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	struct ath9k_tx_queue_info qi;
qi                848 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	memset(&qi, 0, sizeof(qi));
qi                851 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	return ath9k_hw_setuptxqueue(priv->ah, ATH9K_TX_QUEUE_CAB, &qi);
qi                 22 drivers/net/wireless/ath/ath9k/mac.c 					struct ath9k_tx_queue_info *qi)
qi                201 drivers/net/wireless/ath/ath9k/mac.c 	struct ath9k_tx_queue_info *qi;
qi                203 drivers/net/wireless/ath/ath9k/mac.c 	qi = &ah->txq[q];
qi                204 drivers/net/wireless/ath/ath9k/mac.c 	if (qi->tqi_type == ATH9K_TX_QUEUE_INACTIVE) {
qi                212 drivers/net/wireless/ath/ath9k/mac.c 	qi->tqi_ver = qinfo->tqi_ver;
qi                213 drivers/net/wireless/ath/ath9k/mac.c 	qi->tqi_subtype = qinfo->tqi_subtype;
qi                214 drivers/net/wireless/ath/ath9k/mac.c 	qi->tqi_qflags = qinfo->tqi_qflags;
qi                215 drivers/net/wireless/ath/ath9k/mac.c 	qi->tqi_priority = qinfo->tqi_priority;
qi                217 drivers/net/wireless/ath/ath9k/mac.c 		qi->tqi_aifs = min(qinfo->tqi_aifs, 255U);
qi                219 drivers/net/wireless/ath/ath9k/mac.c 		qi->tqi_aifs = INIT_AIFS;
qi                222 drivers/net/wireless/ath/ath9k/mac.c 		qi->tqi_cwmin = 1;
qi                223 drivers/net/wireless/ath/ath9k/mac.c 		while (qi->tqi_cwmin < cw)
qi                224 drivers/net/wireless/ath/ath9k/mac.c 			qi->tqi_cwmin = (qi->tqi_cwmin << 1) | 1;
qi                226 drivers/net/wireless/ath/ath9k/mac.c 		qi->tqi_cwmin = qinfo->tqi_cwmin;
qi                229 drivers/net/wireless/ath/ath9k/mac.c 		qi->tqi_cwmax = 1;
qi                230 drivers/net/wireless/ath/ath9k/mac.c 		while (qi->tqi_cwmax < cw)
qi                231 drivers/net/wireless/ath/ath9k/mac.c 			qi->tqi_cwmax = (qi->tqi_cwmax << 1) | 1;
qi                233 drivers/net/wireless/ath/ath9k/mac.c 		qi->tqi_cwmax = INIT_CWMAX;
qi                236 drivers/net/wireless/ath/ath9k/mac.c 		qi->tqi_shretry = min((u32) qinfo->tqi_shretry, 15U);
qi                238 drivers/net/wireless/ath/ath9k/mac.c 		qi->tqi_shretry = INIT_SH_RETRY;
qi                240 drivers/net/wireless/ath/ath9k/mac.c 		qi->tqi_lgretry = min((u32) qinfo->tqi_lgretry, 15U);
qi                242 drivers/net/wireless/ath/ath9k/mac.c 		qi->tqi_lgretry = INIT_LG_RETRY;
qi                243 drivers/net/wireless/ath/ath9k/mac.c 	qi->tqi_cbrPeriod = qinfo->tqi_cbrPeriod;
qi                244 drivers/net/wireless/ath/ath9k/mac.c 	qi->tqi_cbrOverflowLimit = qinfo->tqi_cbrOverflowLimit;
qi                245 drivers/net/wireless/ath/ath9k/mac.c 	qi->tqi_burstTime = qinfo->tqi_burstTime;
qi                246 drivers/net/wireless/ath/ath9k/mac.c 	qi->tqi_readyTime = qinfo->tqi_readyTime;
qi                250 drivers/net/wireless/ath/ath9k/mac.c 		if (qi->tqi_type == ATH9K_TX_QUEUE_DATA)
qi                251 drivers/net/wireless/ath/ath9k/mac.c 			qi->tqi_intFlags = ATH9K_TXQ_USE_LOCKOUT_BKOFF_DIS;
qi                265 drivers/net/wireless/ath/ath9k/mac.c 	struct ath9k_tx_queue_info *qi;
qi                267 drivers/net/wireless/ath/ath9k/mac.c 	qi = &ah->txq[q];
qi                268 drivers/net/wireless/ath/ath9k/mac.c 	if (qi->tqi_type == ATH9K_TX_QUEUE_INACTIVE) {
qi                274 drivers/net/wireless/ath/ath9k/mac.c 	qinfo->tqi_qflags = qi->tqi_qflags;
qi                275 drivers/net/wireless/ath/ath9k/mac.c 	qinfo->tqi_ver = qi->tqi_ver;
qi                276 drivers/net/wireless/ath/ath9k/mac.c 	qinfo->tqi_subtype = qi->tqi_subtype;
qi                277 drivers/net/wireless/ath/ath9k/mac.c 	qinfo->tqi_qflags = qi->tqi_qflags;
qi                278 drivers/net/wireless/ath/ath9k/mac.c 	qinfo->tqi_priority = qi->tqi_priority;
qi                279 drivers/net/wireless/ath/ath9k/mac.c 	qinfo->tqi_aifs = qi->tqi_aifs;
qi                280 drivers/net/wireless/ath/ath9k/mac.c 	qinfo->tqi_cwmin = qi->tqi_cwmin;
qi                281 drivers/net/wireless/ath/ath9k/mac.c 	qinfo->tqi_cwmax = qi->tqi_cwmax;
qi                282 drivers/net/wireless/ath/ath9k/mac.c 	qinfo->tqi_shretry = qi->tqi_shretry;
qi                283 drivers/net/wireless/ath/ath9k/mac.c 	qinfo->tqi_lgretry = qi->tqi_lgretry;
qi                284 drivers/net/wireless/ath/ath9k/mac.c 	qinfo->tqi_cbrPeriod = qi->tqi_cbrPeriod;
qi                285 drivers/net/wireless/ath/ath9k/mac.c 	qinfo->tqi_cbrOverflowLimit = qi->tqi_cbrOverflowLimit;
qi                286 drivers/net/wireless/ath/ath9k/mac.c 	qinfo->tqi_burstTime = qi->tqi_burstTime;
qi                287 drivers/net/wireless/ath/ath9k/mac.c 	qinfo->tqi_readyTime = qi->tqi_readyTime;
qi                297 drivers/net/wireless/ath/ath9k/mac.c 	struct ath9k_tx_queue_info *qi;
qi                323 drivers/net/wireless/ath/ath9k/mac.c 	qi = &ah->txq[q];
qi                324 drivers/net/wireless/ath/ath9k/mac.c 	if (qi->tqi_type != ATH9K_TX_QUEUE_INACTIVE) {
qi                328 drivers/net/wireless/ath/ath9k/mac.c 	memset(qi, 0, sizeof(struct ath9k_tx_queue_info));
qi                329 drivers/net/wireless/ath/ath9k/mac.c 	qi->tqi_type = type;
qi                330 drivers/net/wireless/ath/ath9k/mac.c 	qi->tqi_physCompBuf = qinfo->tqi_physCompBuf;
qi                349 drivers/net/wireless/ath/ath9k/mac.c 	struct ath9k_tx_queue_info *qi;
qi                351 drivers/net/wireless/ath/ath9k/mac.c 	qi = &ah->txq[q];
qi                352 drivers/net/wireless/ath/ath9k/mac.c 	if (qi->tqi_type == ATH9K_TX_QUEUE_INACTIVE) {
qi                359 drivers/net/wireless/ath/ath9k/mac.c 	qi->tqi_type = ATH9K_TX_QUEUE_INACTIVE;
qi                361 drivers/net/wireless/ath/ath9k/mac.c 	ath9k_hw_set_txq_interrupts(ah, qi);
qi                370 drivers/net/wireless/ath/ath9k/mac.c 	struct ath9k_tx_queue_info *qi;
qi                373 drivers/net/wireless/ath/ath9k/mac.c 	qi = &ah->txq[q];
qi                374 drivers/net/wireless/ath/ath9k/mac.c 	if (qi->tqi_type == ATH9K_TX_QUEUE_INACTIVE) {
qi                381 drivers/net/wireless/ath/ath9k/mac.c 	if (qi->tqi_cwmin == ATH9K_TXQ_USEDEFAULT) {
qi                386 drivers/net/wireless/ath/ath9k/mac.c 		cwMin = qi->tqi_cwmin;
qi                392 drivers/net/wireless/ath/ath9k/mac.c 		  SM(qi->tqi_cwmax, AR_D_LCL_IFS_CWMAX) |
qi                393 drivers/net/wireless/ath/ath9k/mac.c 		  SM(qi->tqi_aifs, AR_D_LCL_IFS_AIFS));
qi                398 drivers/net/wireless/ath/ath9k/mac.c 		  SM(qi->tqi_shretry, AR_D_RETRY_LIMIT_FR_SH));
qi                409 drivers/net/wireless/ath/ath9k/mac.c 	if (qi->tqi_cbrPeriod) {
qi                411 drivers/net/wireless/ath/ath9k/mac.c 			  SM(qi->tqi_cbrPeriod, AR_Q_CBRCFG_INTERVAL) |
qi                412 drivers/net/wireless/ath/ath9k/mac.c 			  SM(qi->tqi_cbrOverflowLimit, AR_Q_CBRCFG_OVF_THRESH));
qi                414 drivers/net/wireless/ath/ath9k/mac.c 			    (qi->tqi_cbrOverflowLimit ?
qi                417 drivers/net/wireless/ath/ath9k/mac.c 	if (qi->tqi_readyTime && (qi->tqi_type != ATH9K_TX_QUEUE_CAB)) {
qi                419 drivers/net/wireless/ath/ath9k/mac.c 			  SM(qi->tqi_readyTime, AR_Q_RDYTIMECFG_DURATION) |
qi                424 drivers/net/wireless/ath/ath9k/mac.c 		  SM(qi->tqi_burstTime, AR_D_CHNTIME_DUR) |
qi                425 drivers/net/wireless/ath/ath9k/mac.c 		  (qi->tqi_burstTime ? AR_D_CHNTIME_EN : 0));
qi                427 drivers/net/wireless/ath/ath9k/mac.c 	if (qi->tqi_burstTime
qi                428 drivers/net/wireless/ath/ath9k/mac.c 	    && (qi->tqi_qflags & TXQ_FLAG_RDYTIME_EXP_POLICY_ENABLE))
qi                431 drivers/net/wireless/ath/ath9k/mac.c 	if (qi->tqi_qflags & TXQ_FLAG_BACKOFF_DISABLE)
qi                436 drivers/net/wireless/ath/ath9k/mac.c 	if (qi->tqi_qflags & TXQ_FLAG_FRAG_BURST_BACKOFF_ENABLE)
qi                439 drivers/net/wireless/ath/ath9k/mac.c 	switch (qi->tqi_type) {
qi                465 drivers/net/wireless/ath/ath9k/mac.c 				  | SM(qi->tqi_aifs, AR_D_LCL_IFS_AIFS));
qi                475 drivers/net/wireless/ath/ath9k/mac.c 		value = (qi->tqi_readyTime -
qi                497 drivers/net/wireless/ath/ath9k/mac.c 	if (qi->tqi_intFlags & ATH9K_TXQ_USE_LOCKOUT_BKOFF_DIS) {
qi                508 drivers/net/wireless/ath/ath9k/mac.c 	if (qi->tqi_qflags & TXQ_FLAG_TXINT_ENABLE) {
qi                512 drivers/net/wireless/ath/ath9k/mac.c 	if (qi->tqi_qflags & TXQ_FLAG_TXDESCINT_ENABLE)
qi                514 drivers/net/wireless/ath/ath9k/mac.c 	if (qi->tqi_qflags & TXQ_FLAG_TXEOLINT_ENABLE)
qi                516 drivers/net/wireless/ath/ath9k/mac.c 	if (qi->tqi_qflags & TXQ_FLAG_TXURNINT_ENABLE)
qi                518 drivers/net/wireless/ath/ath9k/mac.c 	ath9k_hw_set_txq_interrupts(ah, qi);
qi                744 drivers/net/wireless/ath/ath9k/mac.c 	struct ath9k_tx_queue_info qi;
qi                746 drivers/net/wireless/ath/ath9k/mac.c 	memset(&qi, 0, sizeof(qi));
qi                747 drivers/net/wireless/ath/ath9k/mac.c 	qi.tqi_aifs = 1;
qi                748 drivers/net/wireless/ath/ath9k/mac.c 	qi.tqi_cwmin = 0;
qi                749 drivers/net/wireless/ath/ath9k/mac.c 	qi.tqi_cwmax = 0;
qi                752 drivers/net/wireless/ath/ath9k/mac.c 		qi.tqi_qflags = TXQ_FLAG_TXINT_ENABLE;
qi                754 drivers/net/wireless/ath/ath9k/mac.c 	return ath9k_hw_setuptxqueue(ah, ATH9K_TX_QUEUE_BEACON, &qi);
qi               1637 drivers/net/wireless/ath/ath9k/main.c 	struct ath9k_tx_queue_info qi;
qi               1648 drivers/net/wireless/ath/ath9k/main.c 	memset(&qi, 0, sizeof(struct ath9k_tx_queue_info));
qi               1650 drivers/net/wireless/ath/ath9k/main.c 	qi.tqi_aifs = params->aifs;
qi               1651 drivers/net/wireless/ath/ath9k/main.c 	qi.tqi_cwmin = params->cw_min;
qi               1652 drivers/net/wireless/ath/ath9k/main.c 	qi.tqi_cwmax = params->cw_max;
qi               1653 drivers/net/wireless/ath/ath9k/main.c 	qi.tqi_burstTime = params->txop * 32;
qi               1660 drivers/net/wireless/ath/ath9k/main.c 	ath_update_max_aggr_framelen(sc, queue, qi.tqi_burstTime);
qi               1661 drivers/net/wireless/ath/ath9k/main.c 	ret = ath_txq_update(sc, txq->axq_qnum, &qi);
qi               1696 drivers/net/wireless/ath/ath9k/xmit.c 	struct ath9k_tx_queue_info qi;
qi               1705 drivers/net/wireless/ath/ath9k/xmit.c 	memset(&qi, 0, sizeof(qi));
qi               1706 drivers/net/wireless/ath/ath9k/xmit.c 	qi.tqi_subtype = subtype_txq_to_hwq[subtype];
qi               1707 drivers/net/wireless/ath/ath9k/xmit.c 	qi.tqi_aifs = ATH9K_TXQ_USEDEFAULT;
qi               1708 drivers/net/wireless/ath/ath9k/xmit.c 	qi.tqi_cwmin = ATH9K_TXQ_USEDEFAULT;
qi               1709 drivers/net/wireless/ath/ath9k/xmit.c 	qi.tqi_cwmax = ATH9K_TXQ_USEDEFAULT;
qi               1710 drivers/net/wireless/ath/ath9k/xmit.c 	qi.tqi_physCompBuf = 0;
qi               1728 drivers/net/wireless/ath/ath9k/xmit.c 		qi.tqi_qflags = TXQ_FLAG_TXINT_ENABLE;
qi               1731 drivers/net/wireless/ath/ath9k/xmit.c 			qi.tqi_qflags = TXQ_FLAG_TXDESCINT_ENABLE;
qi               1733 drivers/net/wireless/ath/ath9k/xmit.c 			qi.tqi_qflags = TXQ_FLAG_TXEOLINT_ENABLE |
qi               1736 drivers/net/wireless/ath/ath9k/xmit.c 	axq_qnum = ath9k_hw_setuptxqueue(ah, qtype, &qi);
qi               1770 drivers/net/wireless/ath/ath9k/xmit.c 	struct ath9k_tx_queue_info qi;
qi               1774 drivers/net/wireless/ath/ath9k/xmit.c 	ath9k_hw_get_txq_props(ah, qnum, &qi);
qi               1775 drivers/net/wireless/ath/ath9k/xmit.c 	qi.tqi_aifs = qinfo->tqi_aifs;
qi               1776 drivers/net/wireless/ath/ath9k/xmit.c 	qi.tqi_cwmin = qinfo->tqi_cwmin;
qi               1777 drivers/net/wireless/ath/ath9k/xmit.c 	qi.tqi_cwmax = qinfo->tqi_cwmax;
qi               1778 drivers/net/wireless/ath/ath9k/xmit.c 	qi.tqi_burstTime = qinfo->tqi_burstTime;
qi               1779 drivers/net/wireless/ath/ath9k/xmit.c 	qi.tqi_readyTime = qinfo->tqi_readyTime;
qi               1781 drivers/net/wireless/ath/ath9k/xmit.c 	if (!ath9k_hw_set_txq_props(ah, qnum, &qi)) {
qi               1794 drivers/net/wireless/ath/ath9k/xmit.c 	struct ath9k_tx_queue_info qi;
qi               1798 drivers/net/wireless/ath/ath9k/xmit.c 	ath9k_hw_get_txq_props(sc->sc_ah, qnum, &qi);
qi               1800 drivers/net/wireless/ath/ath9k/xmit.c 	qi.tqi_readyTime = (TU_TO_USEC(cur_conf->beacon_interval) *
qi               1802 drivers/net/wireless/ath/ath9k/xmit.c 	ath_txq_update(sc, qnum, &qi);
qi                373 drivers/net/wireless/ath/carl9170/debug.c #define DEBUGFS_QUEUE_DUMP(q, qi)					\
qi                374 drivers/net/wireless/ath/carl9170/debug.c static char *carl9170_debugfs_##q ##_##qi ##_read(struct ar9170 *ar,	\
qi                377 drivers/net/wireless/ath/carl9170/debug.c 	carl9170_debugfs_queue_dump(ar, buf, len, bufsize, &ar->q[qi]);	\
qi                380 drivers/net/wireless/ath/carl9170/debug.c DEBUGFS_DECLARE_RO_FILE(q##_##qi, 8000);
qi                 31 drivers/regulator/mt6323-regulator.c 	u32 qi;
qi                 57 drivers/regulator/mt6323-regulator.c 	.qi = BIT(13),							\
qi                 80 drivers/regulator/mt6323-regulator.c 	.qi = BIT(15),							\
qi                100 drivers/regulator/mt6323-regulator.c 	.qi = BIT(15),							\
qi                169 drivers/regulator/mt6323-regulator.c 	return (regval & info->qi) ? REGULATOR_STATUS_ON : REGULATOR_STATUS_OFF;
qi                 28 drivers/regulator/mt6358-regulator.c 	u32 qi;
qi                 61 drivers/regulator/mt6358-regulator.c 	.qi = BIT(0),	\
qi                 89 drivers/regulator/mt6358-regulator.c 	.qi = BIT(15),	\
qi                118 drivers/regulator/mt6358-regulator.c 	.qi = BIT(0),	\
qi                137 drivers/regulator/mt6358-regulator.c 	.qi = BIT(15),							\
qi                325 drivers/regulator/mt6358-regulator.c 	return (regval & info->qi) ? REGULATOR_STATUS_ON : REGULATOR_STATUS_OFF;
qi                 31 drivers/regulator/mt6397-regulator.c 	u32 qi;
qi                 59 drivers/regulator/mt6397-regulator.c 	.qi = BIT(13),							\
qi                 85 drivers/regulator/mt6397-regulator.c 	.qi = BIT(15),							\
qi                102 drivers/regulator/mt6397-regulator.c 	.qi = BIT(15),							\
qi                218 drivers/regulator/mt6397-regulator.c 	return (regval & info->qi) ? REGULATOR_STATUS_ON : REGULATOR_STATUS_OFF;
qi               1401 fs/cifs/smb2ops.c 	struct smb_query_info qi;
qi               1426 fs/cifs/smb2ops.c 	if (copy_from_user(&qi, arg, sizeof(struct smb_query_info)))
qi               1429 fs/cifs/smb2ops.c 	if (qi.output_buffer_length > 1024)
qi               1438 fs/cifs/smb2ops.c 	buffer = kmalloc(qi.output_buffer_length, GFP_KERNEL);
qi               1443 fs/cifs/smb2ops.c 			   qi.output_buffer_length)) {
qi               1463 fs/cifs/smb2ops.c 	if (qi.flags & PASSTHRU_FSCTL) {
qi               1464 fs/cifs/smb2ops.c 		switch (qi.info_type & FSCTL_DEVICE_ACCESS_MASK) {
qi               1478 fs/cifs/smb2ops.c 	} else if (qi.flags & PASSTHRU_SET_INFO) {
qi               1490 fs/cifs/smb2ops.c 	if (qi.flags & PASSTHRU_FSCTL) {
qi               1501 fs/cifs/smb2ops.c 					     qi.info_type, true, buffer,
qi               1502 fs/cifs/smb2ops.c 					     qi.output_buffer_length,
qi               1507 fs/cifs/smb2ops.c 	} else if (qi.flags == PASSTHRU_SET_INFO) {
qi               1525 fs/cifs/smb2ops.c 	} else if (qi.flags == PASSTHRU_QUERY_INFO) {
qi               1531 fs/cifs/smb2ops.c 				  COMPOUND_FID, qi.file_info_class,
qi               1532 fs/cifs/smb2ops.c 				  qi.info_type, qi.additional_information,
qi               1533 fs/cifs/smb2ops.c 				  qi.input_buffer_length,
qi               1534 fs/cifs/smb2ops.c 				  qi.output_buffer_length, buffer);
qi               1536 fs/cifs/smb2ops.c 		cifs_tcon_dbg(VFS, "invalid passthru query flags: 0x%x\n", qi.flags);
qi               1561 fs/cifs/smb2ops.c 	if (qi.flags & PASSTHRU_FSCTL) {
qi               1564 fs/cifs/smb2ops.c 		if (le32_to_cpu(io_rsp->OutputCount) < qi.input_buffer_length)
qi               1565 fs/cifs/smb2ops.c 			qi.input_buffer_length = le32_to_cpu(io_rsp->OutputCount);
qi               1566 fs/cifs/smb2ops.c 		if (qi.input_buffer_length > 0 &&
qi               1567 fs/cifs/smb2ops.c 		    le32_to_cpu(io_rsp->OutputOffset) + qi.input_buffer_length > rsp_iov[1].iov_len) {
qi               1571 fs/cifs/smb2ops.c 		if (copy_to_user(&pqi->input_buffer_length, &qi.input_buffer_length,
qi               1572 fs/cifs/smb2ops.c 				 sizeof(qi.input_buffer_length))) {
qi               1578 fs/cifs/smb2ops.c 				 qi.input_buffer_length)) {
qi               1585 fs/cifs/smb2ops.c 		if (le32_to_cpu(qi_rsp->OutputBufferLength) < qi.input_buffer_length)
qi               1586 fs/cifs/smb2ops.c 			qi.input_buffer_length = le32_to_cpu(qi_rsp->OutputBufferLength);
qi               1587 fs/cifs/smb2ops.c 		if (copy_to_user(&pqi->input_buffer_length, &qi.input_buffer_length,
qi               1588 fs/cifs/smb2ops.c 				 sizeof(qi.input_buffer_length))) {
qi               1592 fs/cifs/smb2ops.c 		if (copy_to_user(pqi + 1, qi_rsp->Buffer, qi.input_buffer_length)) {
qi               1601 fs/cifs/smb2ops.c 	if (qi.flags & PASSTHRU_FSCTL)
qi                 83 fs/xfs/scrub/quota.c 	struct xfs_quotainfo	*qi = mp->m_quotainfo;
qi                101 fs/xfs/scrub/quota.c 	offset = id / qi->qi_dqperchunk;
qi                191 fs/xfs/scrub/quota.c 	struct xfs_quotainfo	*qi = sc->mp->m_quotainfo;
qi                202 fs/xfs/scrub/quota.c 	max_dqid_off = ((xfs_dqid_t)-1) / qi->qi_dqperchunk;
qi                230 fs/xfs/scrub/quota.c 	struct xfs_quotainfo	*qi = mp->m_quotainfo;
qi                256 fs/xfs/scrub/quota.c 			sqi.last_id * qi->qi_dqperchunk, &error))
qi                658 fs/xfs/xfs_dquot.c 	struct xfs_quotainfo	*qi,
qi                665 fs/xfs/xfs_dquot.c 	mutex_lock(&qi->qi_tree_lock);
qi                668 fs/xfs/xfs_dquot.c 		mutex_unlock(&qi->qi_tree_lock);
qi                676 fs/xfs/xfs_dquot.c 		mutex_unlock(&qi->qi_tree_lock);
qi                683 fs/xfs/xfs_dquot.c 	mutex_unlock(&qi->qi_tree_lock);
qi                699 fs/xfs/xfs_dquot.c 	struct xfs_quotainfo	*qi,
qi                706 fs/xfs/xfs_dquot.c 	mutex_lock(&qi->qi_tree_lock);
qi                711 fs/xfs/xfs_dquot.c 		mutex_unlock(&qi->qi_tree_lock);
qi                720 fs/xfs/xfs_dquot.c 	qi->qi_dquots++;
qi                721 fs/xfs/xfs_dquot.c 	mutex_unlock(&qi->qi_tree_lock);
qi                766 fs/xfs/xfs_dquot.c 	struct xfs_quotainfo	*qi = mp->m_quotainfo;
qi                767 fs/xfs/xfs_dquot.c 	struct radix_tree_root	*tree = xfs_dquot_tree(qi, type);
qi                776 fs/xfs/xfs_dquot.c 	dqp = xfs_qm_dqget_cache_lookup(mp, qi, tree, id);
qi                786 fs/xfs/xfs_dquot.c 	error = xfs_qm_dqget_cache_insert(mp, qi, tree, id, dqp);
qi                855 fs/xfs/xfs_dquot.c 	struct xfs_quotainfo	*qi = mp->m_quotainfo;
qi                856 fs/xfs/xfs_dquot.c 	struct radix_tree_root	*tree = xfs_dquot_tree(qi, type);
qi                871 fs/xfs/xfs_dquot.c 	dqp = xfs_qm_dqget_cache_lookup(mp, qi, tree, id);
qi                910 fs/xfs/xfs_dquot.c 	error = xfs_qm_dqget_cache_insert(mp, qi, tree, id, dqp);
qi                977 fs/xfs/xfs_dquot.c 		struct xfs_quotainfo	*qi = dqp->q_mount->m_quotainfo;
qi                980 fs/xfs/xfs_dquot.c 		if (list_lru_add(&qi->qi_lru, &dqp->q_lru))
qi                 35 fs/xfs/xfs_qm.c STATIC void	xfs_qm_destroy_quotainos(xfs_quotainfo_t *qi);
qi                 53 fs/xfs/xfs_qm.c 	struct xfs_quotainfo	*qi = mp->m_quotainfo;
qi                 54 fs/xfs/xfs_qm.c 	struct radix_tree_root	*tree = xfs_dquot_tree(qi, type);
qi                 70 fs/xfs/xfs_qm.c 		mutex_lock(&qi->qi_tree_lock);
qi                 74 fs/xfs/xfs_qm.c 			mutex_unlock(&qi->qi_tree_lock);
qi                 92 fs/xfs/xfs_qm.c 		mutex_unlock(&qi->qi_tree_lock);
qi                122 fs/xfs/xfs_qm.c 	struct xfs_quotainfo	*qi = mp->m_quotainfo;
qi                162 fs/xfs/xfs_qm.c 	radix_tree_delete(xfs_dquot_tree(qi, dqp->q_core.d_flags),
qi                164 fs/xfs/xfs_qm.c 	qi->qi_dquots--;
qi                171 fs/xfs/xfs_qm.c 	list_lru_del(&qi->qi_lru, &dqp->q_lru);
qi                499 fs/xfs/xfs_qm.c 	struct xfs_quotainfo	*qi = container_of(shrink,
qi                511 fs/xfs/xfs_qm.c 	freed = list_lru_shrink_walk(&qi->qi_lru, sc,
qi                534 fs/xfs/xfs_qm.c 	struct xfs_quotainfo	*qi = container_of(shrink,
qi                537 fs/xfs/xfs_qm.c 	return list_lru_shrink_count(&qi->qi_lru, sc);
qi                714 fs/xfs/xfs_qm.c 	xfs_quotainfo_t *qi;
qi                716 fs/xfs/xfs_qm.c 	qi = mp->m_quotainfo;
qi                717 fs/xfs/xfs_qm.c 	ASSERT(qi != NULL);
qi                719 fs/xfs/xfs_qm.c 	unregister_shrinker(&qi->qi_shrinker);
qi                720 fs/xfs/xfs_qm.c 	list_lru_destroy(&qi->qi_lru);
qi                721 fs/xfs/xfs_qm.c 	xfs_qm_destroy_quotainos(qi);
qi                722 fs/xfs/xfs_qm.c 	mutex_destroy(&qi->qi_tree_lock);
qi                723 fs/xfs/xfs_qm.c 	mutex_destroy(&qi->qi_quotaofflock);
qi                724 fs/xfs/xfs_qm.c 	kmem_free(qi);
qi               1562 fs/xfs/xfs_qm.c 	xfs_quotainfo_t	*qi)
qi               1564 fs/xfs/xfs_qm.c 	if (qi->qi_uquotaip) {
qi               1565 fs/xfs/xfs_qm.c 		xfs_irele(qi->qi_uquotaip);
qi               1566 fs/xfs/xfs_qm.c 		qi->qi_uquotaip = NULL; /* paranoia */
qi               1568 fs/xfs/xfs_qm.c 	if (qi->qi_gquotaip) {
qi               1569 fs/xfs/xfs_qm.c 		xfs_irele(qi->qi_gquotaip);
qi               1570 fs/xfs/xfs_qm.c 		qi->qi_gquotaip = NULL;
qi               1572 fs/xfs/xfs_qm.c 	if (qi->qi_pquotaip) {
qi               1573 fs/xfs/xfs_qm.c 		xfs_irele(qi->qi_pquotaip);
qi               1574 fs/xfs/xfs_qm.c 		qi->qi_pquotaip = NULL;
qi               1583 fs/xfs/xfs_qm.c 	struct xfs_quotainfo	*qi = mp->m_quotainfo;
qi               1585 fs/xfs/xfs_qm.c 	mutex_lock(&qi->qi_tree_lock);
qi               1586 fs/xfs/xfs_qm.c 	radix_tree_delete(xfs_dquot_tree(qi, dqp->q_core.d_flags),
qi               1589 fs/xfs/xfs_qm.c 	qi->qi_dquots--;
qi               1590 fs/xfs/xfs_qm.c 	mutex_unlock(&qi->qi_tree_lock);
qi                 84 fs/xfs/xfs_qm.h 	struct xfs_quotainfo	*qi,
qi                 89 fs/xfs/xfs_qm.h 		return &qi->qi_uquota_tree;
qi                 91 fs/xfs/xfs_qm.h 		return &qi->qi_gquota_tree;
qi                 93 fs/xfs/xfs_qm.h 		return &qi->qi_pquota_tree;
qi                167 fs/xfs/xfs_qm.h xfs_get_defquota(struct xfs_dquot *dqp, struct xfs_quotainfo *qi)
qi                172 fs/xfs/xfs_qm.h 		defq = &qi->qi_usr_default;
qi                174 fs/xfs/xfs_qm.h 		defq = &qi->qi_grp_default;
qi                177 fs/xfs/xfs_qm.h 		defq = &qi->qi_prj_default;
qi                548 include/linux/intel-iommu.h 	struct q_inval  *qi;            /* Queued invalidation info */