qi 224 arch/um/drivers/vector_kern.c static int vector_advancehead(struct vector_queue *qi, int advance) qi 228 arch/um/drivers/vector_kern.c qi->head = qi 229 arch/um/drivers/vector_kern.c (qi->head + advance) qi 230 arch/um/drivers/vector_kern.c % qi->max_depth; qi 233 arch/um/drivers/vector_kern.c spin_lock(&qi->tail_lock); qi 234 arch/um/drivers/vector_kern.c qi->queue_depth -= advance; qi 240 arch/um/drivers/vector_kern.c if (qi->queue_depth == 0) { qi 241 arch/um/drivers/vector_kern.c qi->head = 0; qi 242 arch/um/drivers/vector_kern.c qi->tail = 0; qi 244 arch/um/drivers/vector_kern.c queue_depth = qi->queue_depth; qi 245 arch/um/drivers/vector_kern.c spin_unlock(&qi->tail_lock); qi 254 arch/um/drivers/vector_kern.c static int vector_advancetail(struct vector_queue *qi, int advance) qi 258 arch/um/drivers/vector_kern.c qi->tail = qi 259 arch/um/drivers/vector_kern.c (qi->tail + advance) qi 260 arch/um/drivers/vector_kern.c % qi->max_depth; qi 261 arch/um/drivers/vector_kern.c spin_lock(&qi->head_lock); qi 262 arch/um/drivers/vector_kern.c qi->queue_depth += advance; qi 263 arch/um/drivers/vector_kern.c queue_depth = qi->queue_depth; qi 264 arch/um/drivers/vector_kern.c spin_unlock(&qi->head_lock); qi 309 arch/um/drivers/vector_kern.c static int vector_enqueue(struct vector_queue *qi, struct sk_buff *skb) qi 311 arch/um/drivers/vector_kern.c struct vector_private *vp = netdev_priv(qi->dev); qi 314 arch/um/drivers/vector_kern.c struct mmsghdr *mmsg_vector = qi->mmsg_vector; qi 317 arch/um/drivers/vector_kern.c spin_lock(&qi->tail_lock); qi 318 arch/um/drivers/vector_kern.c spin_lock(&qi->head_lock); qi 319 arch/um/drivers/vector_kern.c queue_depth = qi->queue_depth; qi 320 arch/um/drivers/vector_kern.c spin_unlock(&qi->head_lock); qi 325 arch/um/drivers/vector_kern.c if (queue_depth < qi->max_depth) { qi 327 arch/um/drivers/vector_kern.c *(qi->skbuff_vector + qi->tail) = skb; qi 328 arch/um/drivers/vector_kern.c mmsg_vector += qi->tail; qi 339 arch/um/drivers/vector_kern.c queue_depth = vector_advancetail(qi, 1); qi 342 arch/um/drivers/vector_kern.c spin_unlock(&qi->tail_lock); qi 345 arch/um/drivers/vector_kern.c qi->dev->stats.tx_dropped++; qi 349 arch/um/drivers/vector_kern.c netdev_completed_queue(qi->dev, 1, packet_len); qi 351 arch/um/drivers/vector_kern.c spin_unlock(&qi->tail_lock); qi 355 arch/um/drivers/vector_kern.c static int consume_vector_skbs(struct vector_queue *qi, int count) qi 361 arch/um/drivers/vector_kern.c for (skb_index = qi->head; skb_index < qi->head + count; skb_index++) { qi 362 arch/um/drivers/vector_kern.c skb = *(qi->skbuff_vector + skb_index); qi 367 arch/um/drivers/vector_kern.c *(qi->skbuff_vector + skb_index) = NULL; qi 370 arch/um/drivers/vector_kern.c qi->dev->stats.tx_bytes += bytes_compl; qi 371 arch/um/drivers/vector_kern.c qi->dev->stats.tx_packets += count; qi 372 arch/um/drivers/vector_kern.c netdev_completed_queue(qi->dev, count, bytes_compl); qi 373 arch/um/drivers/vector_kern.c return vector_advancehead(qi, count); qi 383 arch/um/drivers/vector_kern.c static int vector_send(struct vector_queue *qi) qi 385 arch/um/drivers/vector_kern.c struct vector_private *vp = netdev_priv(qi->dev); qi 387 arch/um/drivers/vector_kern.c int result = 0, send_len, queue_depth = qi->max_depth; qi 389 arch/um/drivers/vector_kern.c if (spin_trylock(&qi->head_lock)) { qi 390 arch/um/drivers/vector_kern.c if (spin_trylock(&qi->tail_lock)) { qi 392 arch/um/drivers/vector_kern.c queue_depth = qi->queue_depth; qi 393 arch/um/drivers/vector_kern.c spin_unlock(&qi->tail_lock); qi 397 arch/um/drivers/vector_kern.c send_from = qi->mmsg_vector; qi 398 arch/um/drivers/vector_kern.c send_from += qi->head; qi 400 arch/um/drivers/vector_kern.c if (send_len + qi->head > qi->max_depth) qi 401 arch/um/drivers/vector_kern.c send_len = qi->max_depth - qi->head; qi 427 arch/um/drivers/vector_kern.c consume_vector_skbs(qi, result); qi 437 arch/um/drivers/vector_kern.c netif_trans_update(qi->dev); qi 438 arch/um/drivers/vector_kern.c netif_wake_queue(qi->dev); qi 448 arch/um/drivers/vector_kern.c spin_unlock(&qi->head_lock); qi 459 arch/um/drivers/vector_kern.c static void destroy_queue(struct vector_queue *qi) qi 463 arch/um/drivers/vector_kern.c struct vector_private *vp = netdev_priv(qi->dev); qi 466 arch/um/drivers/vector_kern.c if (qi == NULL) qi 471 arch/um/drivers/vector_kern.c if (qi->skbuff_vector != NULL) { qi 472 arch/um/drivers/vector_kern.c for (i = 0; i < qi->max_depth; i++) { qi 473 arch/um/drivers/vector_kern.c if (*(qi->skbuff_vector + i) != NULL) qi 474 arch/um/drivers/vector_kern.c dev_kfree_skb_any(*(qi->skbuff_vector + i)); qi 476 arch/um/drivers/vector_kern.c kfree(qi->skbuff_vector); qi 479 arch/um/drivers/vector_kern.c if (qi->mmsg_vector != NULL) { qi 480 arch/um/drivers/vector_kern.c mmsg_vector = qi->mmsg_vector; qi 481 arch/um/drivers/vector_kern.c for (i = 0; i < qi->max_depth; i++) { qi 491 arch/um/drivers/vector_kern.c kfree(qi->mmsg_vector); qi 493 arch/um/drivers/vector_kern.c kfree(qi); qi 652 arch/um/drivers/vector_kern.c static void prep_queue_for_rx(struct vector_queue *qi) qi 654 arch/um/drivers/vector_kern.c struct vector_private *vp = netdev_priv(qi->dev); qi 655 arch/um/drivers/vector_kern.c struct mmsghdr *mmsg_vector = qi->mmsg_vector; qi 656 arch/um/drivers/vector_kern.c void **skbuff_vector = qi->skbuff_vector; qi 659 arch/um/drivers/vector_kern.c if (qi->queue_depth == 0) qi 661 arch/um/drivers/vector_kern.c for (i = 0; i < qi->queue_depth; i++) { qi 671 arch/um/drivers/vector_kern.c qi->queue_depth = 0; qi 937 arch/um/drivers/vector_kern.c struct vector_queue *qi = vp->rx_queue; qi 939 arch/um/drivers/vector_kern.c struct mmsghdr *mmsg_vector = qi->mmsg_vector; qi 940 arch/um/drivers/vector_kern.c void **skbuff_vector = qi->skbuff_vector; qi 947 arch/um/drivers/vector_kern.c prep_queue_for_rx(qi); qi 952 arch/um/drivers/vector_kern.c vp->fds->rx_fd, qi->mmsg_vector, qi->max_depth, 0); qi 965 arch/um/drivers/vector_kern.c qi->queue_depth = packet_count; qi 743 drivers/crypto/caam/ctrl.c ctrlpriv->qi = (struct caam_queue_if __iomem __force *) qi 748 drivers/crypto/caam/ctrl.c wr_reg32(&ctrlpriv->qi->qi_control_lo, QICTL_DQEN); qi 73 drivers/crypto/caam/intern.h struct caam_queue_if __iomem *qi; /* QI control region */ qi 298 drivers/crypto/caam/regs.h u32 qi; /* QI_VERSION */ qi 26 drivers/gpu/drm/i915/display/intel_bw.c struct intel_qgv_info *qi) qi 40 drivers/gpu/drm/i915/display/intel_bw.c qi->dram_type = INTEL_DRAM_DDR4; qi 43 drivers/gpu/drm/i915/display/intel_bw.c qi->dram_type = INTEL_DRAM_DDR3; qi 46 drivers/gpu/drm/i915/display/intel_bw.c qi->dram_type = INTEL_DRAM_LPDDR3; qi 49 drivers/gpu/drm/i915/display/intel_bw.c qi->dram_type = INTEL_DRAM_LPDDR3; qi 56 drivers/gpu/drm/i915/display/intel_bw.c qi->num_channels = (val & 0xf0) >> 4; qi 57 drivers/gpu/drm/i915/display/intel_bw.c qi->num_points = (val & 0xf00) >> 8; qi 59 drivers/gpu/drm/i915/display/intel_bw.c qi->t_bl = qi->dram_type == INTEL_DRAM_DDR4 ? 4 : 8; qi 91 drivers/gpu/drm/i915/display/intel_bw.c struct intel_qgv_info *qi) qi 95 drivers/gpu/drm/i915/display/intel_bw.c ret = icl_pcode_read_mem_global_info(dev_priv, qi); qi 99 drivers/gpu/drm/i915/display/intel_bw.c if (WARN_ON(qi->num_points > ARRAY_SIZE(qi->points))) qi 100 drivers/gpu/drm/i915/display/intel_bw.c qi->num_points = ARRAY_SIZE(qi->points); qi 102 drivers/gpu/drm/i915/display/intel_bw.c for (i = 0; i < qi->num_points; i++) { qi 103 drivers/gpu/drm/i915/display/intel_bw.c struct intel_qgv_point *sp = &qi->points[i]; qi 123 drivers/gpu/drm/i915/display/intel_bw.c static int icl_sagv_max_dclk(const struct intel_qgv_info *qi) qi 128 drivers/gpu/drm/i915/display/intel_bw.c for (i = 0; i < qi->num_points; i++) qi 129 drivers/gpu/drm/i915/display/intel_bw.c dclk = max(dclk, qi->points[i].dclk); qi 147 drivers/gpu/drm/i915/display/intel_bw.c struct intel_qgv_info qi = {}; qi 157 drivers/gpu/drm/i915/display/intel_bw.c ret = icl_get_qgv_points(dev_priv, &qi); qi 162 drivers/gpu/drm/i915/display/intel_bw.c num_channels = qi.num_channels; qi 165 drivers/gpu/drm/i915/display/intel_bw.c dclk_max = icl_sagv_max_dclk(&qi); qi 181 drivers/gpu/drm/i915/display/intel_bw.c bi->num_qgv_points = qi.num_points; qi 183 drivers/gpu/drm/i915/display/intel_bw.c for (j = 0; j < qi.num_points; j++) { qi 184 drivers/gpu/drm/i915/display/intel_bw.c const struct intel_qgv_point *sp = &qi.points[j]; qi 194 drivers/gpu/drm/i915/display/intel_bw.c (clpchgroup - 1) * qi.t_bl + sp->t_rdpre); qi 1130 drivers/iommu/dmar.c if (iommu->qi) { qi 1131 drivers/iommu/dmar.c free_page((unsigned long)iommu->qi->desc); qi 1132 drivers/iommu/dmar.c kfree(iommu->qi->desc_status); qi 1133 drivers/iommu/dmar.c kfree(iommu->qi); qi 1146 drivers/iommu/dmar.c static inline void reclaim_free_desc(struct q_inval *qi) qi 1148 drivers/iommu/dmar.c while (qi->desc_status[qi->free_tail] == QI_DONE || qi 1149 drivers/iommu/dmar.c qi->desc_status[qi->free_tail] == QI_ABORT) { qi 1150 drivers/iommu/dmar.c qi->desc_status[qi->free_tail] = QI_FREE; qi 1151 drivers/iommu/dmar.c qi->free_tail = (qi->free_tail + 1) % QI_LENGTH; qi 1152 drivers/iommu/dmar.c qi->free_cnt++; qi 1160 drivers/iommu/dmar.c struct q_inval *qi = iommu->qi; qi 1164 drivers/iommu/dmar.c if (qi->desc_status[wait_index] == QI_ABORT) qi 1177 drivers/iommu/dmar.c struct qi_desc *desc = qi->desc + head; qi 1187 drivers/iommu/dmar.c memcpy(desc, qi->desc + (wait_index << shift), qi 1208 drivers/iommu/dmar.c if (qi->desc_status[head] == QI_IN_USE) qi 1209 drivers/iommu/dmar.c qi->desc_status[head] = QI_ABORT; qi 1213 drivers/iommu/dmar.c if (qi->desc_status[wait_index] == QI_ABORT) qi 1230 drivers/iommu/dmar.c struct q_inval *qi = iommu->qi; qi 1236 drivers/iommu/dmar.c if (!qi) qi 1242 drivers/iommu/dmar.c raw_spin_lock_irqsave(&qi->q_lock, flags); qi 1243 drivers/iommu/dmar.c while (qi->free_cnt < 3) { qi 1244 drivers/iommu/dmar.c raw_spin_unlock_irqrestore(&qi->q_lock, flags); qi 1246 drivers/iommu/dmar.c raw_spin_lock_irqsave(&qi->q_lock, flags); qi 1249 drivers/iommu/dmar.c index = qi->free_head; qi 1254 drivers/iommu/dmar.c qi->desc_status[index] = qi->desc_status[wait_index] = QI_IN_USE; qi 1257 drivers/iommu/dmar.c memcpy(qi->desc + offset, desc, length); qi 1260 drivers/iommu/dmar.c wait_desc.qw1 = virt_to_phys(&qi->desc_status[wait_index]); qi 1265 drivers/iommu/dmar.c memcpy(qi->desc + offset, &wait_desc, length); qi 1267 drivers/iommu/dmar.c qi->free_head = (qi->free_head + 2) % QI_LENGTH; qi 1268 drivers/iommu/dmar.c qi->free_cnt -= 2; qi 1274 drivers/iommu/dmar.c writel(qi->free_head << shift, iommu->reg + DMAR_IQT_REG); qi 1276 drivers/iommu/dmar.c while (qi->desc_status[wait_index] != QI_DONE) { qi 1288 drivers/iommu/dmar.c raw_spin_unlock(&qi->q_lock); qi 1290 drivers/iommu/dmar.c raw_spin_lock(&qi->q_lock); qi 1293 drivers/iommu/dmar.c qi->desc_status[index] = QI_DONE; qi 1295 drivers/iommu/dmar.c reclaim_free_desc(qi); qi 1296 drivers/iommu/dmar.c raw_spin_unlock_irqrestore(&qi->q_lock, flags); qi 1422 drivers/iommu/dmar.c struct q_inval *qi = iommu->qi; qi 1423 drivers/iommu/dmar.c u64 val = virt_to_phys(qi->desc); qi 1425 drivers/iommu/dmar.c qi->free_head = qi->free_tail = 0; qi 1426 drivers/iommu/dmar.c qi->free_cnt = QI_LENGTH; qi 1458 drivers/iommu/dmar.c struct q_inval *qi; qi 1467 drivers/iommu/dmar.c if (iommu->qi) qi 1470 drivers/iommu/dmar.c iommu->qi = kmalloc(sizeof(*qi), GFP_ATOMIC); qi 1471 drivers/iommu/dmar.c if (!iommu->qi) qi 1474 drivers/iommu/dmar.c qi = iommu->qi; qi 1483 drivers/iommu/dmar.c kfree(qi); qi 1484 drivers/iommu/dmar.c iommu->qi = NULL; qi 1488 drivers/iommu/dmar.c qi->desc = page_address(desc_page); qi 1490 drivers/iommu/dmar.c qi->desc_status = kcalloc(QI_LENGTH, sizeof(int), GFP_ATOMIC); qi 1491 drivers/iommu/dmar.c if (!qi->desc_status) { qi 1492 drivers/iommu/dmar.c free_page((unsigned long) qi->desc); qi 1493 drivers/iommu/dmar.c kfree(qi); qi 1494 drivers/iommu/dmar.c iommu->qi = NULL; qi 1498 drivers/iommu/dmar.c raw_spin_lock_init(&qi->q_lock); qi 1847 drivers/iommu/dmar.c if (!iommu->qi) qi 1350 drivers/iommu/intel-iommu.c if (!iommu->qi) qi 2963 drivers/iommu/intel-iommu.c if (!iommu->qi) { qi 4175 drivers/iommu/intel-iommu.c if (iommu->qi) qi 584 drivers/iommu/intel_irq_remapping.c if (!iommu->qi) { qi 1045 drivers/iommu/intel_irq_remapping.c if (iommu->qi) qi 425 drivers/net/ethernet/intel/ice/ice_controlq.c #define ICE_FREE_CQ_BUFS(hw, qi, ring) \ qi 429 drivers/net/ethernet/intel/ice/ice_controlq.c for (i = 0; i < (qi)->num_##ring##_entries; i++) \ qi 430 drivers/net/ethernet/intel/ice/ice_controlq.c if ((qi)->ring.r.ring##_bi[i].pa) { \ qi 432 drivers/net/ethernet/intel/ice/ice_controlq.c (qi)->ring.r.ring##_bi[i].size,\ qi 433 drivers/net/ethernet/intel/ice/ice_controlq.c (qi)->ring.r.ring##_bi[i].va,\ qi 434 drivers/net/ethernet/intel/ice/ice_controlq.c (qi)->ring.r.ring##_bi[i].pa);\ qi 435 drivers/net/ethernet/intel/ice/ice_controlq.c (qi)->ring.r.ring##_bi[i].va = NULL; \ qi 436 drivers/net/ethernet/intel/ice/ice_controlq.c (qi)->ring.r.ring##_bi[i].pa = 0; \ qi 437 drivers/net/ethernet/intel/ice/ice_controlq.c (qi)->ring.r.ring##_bi[i].size = 0; \ qi 440 drivers/net/ethernet/intel/ice/ice_controlq.c if ((qi)->ring.cmd_buf) \ qi 441 drivers/net/ethernet/intel/ice/ice_controlq.c devm_kfree(ice_hw_to_dev(hw), (qi)->ring.cmd_buf); \ qi 443 drivers/net/ethernet/intel/ice/ice_controlq.c devm_kfree(ice_hw_to_dev(hw), (qi)->ring.dma_head); \ qi 303 drivers/net/ethernet/pensando/ionic/ionic_txrx.c unsigned int qi = rxcq->bound_q->index; qi 312 drivers/net/ethernet/pensando/ionic/ionic_txrx.c txcq = &lif->txqcqs[qi].qcq->cq; qi 978 drivers/net/wireless/ath/ath5k/base.c struct ath5k_txq_info qi = { qi 1000 drivers/net/wireless/ath/ath5k/base.c qi.tqi_flags = AR5K_TXQ_FLAG_TXEOLINT_ENABLE | qi 1002 drivers/net/wireless/ath/ath5k/base.c qnum = ath5k_hw_setup_tx_queue(ah, qtype, &qi); qi 1028 drivers/net/wireless/ath/ath5k/base.c struct ath5k_txq_info qi = { qi 1038 drivers/net/wireless/ath/ath5k/base.c return ath5k_hw_setup_tx_queue(ah, AR5K_TX_QUEUE_BEACON, &qi); qi 1044 drivers/net/wireless/ath/ath5k/base.c struct ath5k_txq_info qi; qi 1047 drivers/net/wireless/ath/ath5k/base.c ret = ath5k_hw_get_tx_queueprops(ah, ah->bhalq, &qi); qi 1057 drivers/net/wireless/ath/ath5k/base.c qi.tqi_aifs = 0; qi 1058 drivers/net/wireless/ath/ath5k/base.c qi.tqi_cw_min = 0; qi 1059 drivers/net/wireless/ath/ath5k/base.c qi.tqi_cw_max = 0; qi 1064 drivers/net/wireless/ath/ath5k/base.c qi.tqi_aifs = 0; qi 1065 drivers/net/wireless/ath/ath5k/base.c qi.tqi_cw_min = 0; qi 1066 drivers/net/wireless/ath/ath5k/base.c qi.tqi_cw_max = 2 * AR5K_TUNE_CWMIN; qi 1071 drivers/net/wireless/ath/ath5k/base.c qi.tqi_aifs, qi.tqi_cw_min, qi.tqi_cw_max); qi 1073 drivers/net/wireless/ath/ath5k/base.c ret = ath5k_hw_set_tx_queueprops(ah, ah->bhalq, &qi); qi 1084 drivers/net/wireless/ath/ath5k/base.c ret = ath5k_hw_get_tx_queueprops(ah, AR5K_TX_QUEUE_ID_CAB, &qi); qi 1088 drivers/net/wireless/ath/ath5k/base.c qi.tqi_ready_time = (ah->bintval * 80) / 100; qi 1089 drivers/net/wireless/ath/ath5k/base.c ret = ath5k_hw_set_tx_queueprops(ah, AR5K_TX_QUEUE_ID_CAB, &qi); qi 578 drivers/net/wireless/ath/ath5k/mac80211-ops.c struct ath5k_txq_info qi; qi 586 drivers/net/wireless/ath/ath5k/mac80211-ops.c ath5k_hw_get_tx_queueprops(ah, queue, &qi); qi 588 drivers/net/wireless/ath/ath5k/mac80211-ops.c qi.tqi_aifs = params->aifs; qi 589 drivers/net/wireless/ath/ath5k/mac80211-ops.c qi.tqi_cw_min = params->cw_min; qi 590 drivers/net/wireless/ath/ath5k/mac80211-ops.c qi.tqi_cw_max = params->cw_max; qi 591 drivers/net/wireless/ath/ath5k/mac80211-ops.c qi.tqi_burst_time = params->txop * 32; qi 599 drivers/net/wireless/ath/ath5k/mac80211-ops.c if (ath5k_hw_set_tx_queueprops(ah, queue, &qi)) { qi 157 drivers/net/wireless/ath/ath5k/qcu.c struct ath5k_txq_info *qi; qi 161 drivers/net/wireless/ath/ath5k/qcu.c qi = &ah->ah_txq[queue]; qi 163 drivers/net/wireless/ath/ath5k/qcu.c if (qi->tqi_type == AR5K_TX_QUEUE_INACTIVE) qi 167 drivers/net/wireless/ath/ath5k/qcu.c qi->tqi_type = qinfo->tqi_type; qi 168 drivers/net/wireless/ath/ath5k/qcu.c qi->tqi_subtype = qinfo->tqi_subtype; qi 169 drivers/net/wireless/ath/ath5k/qcu.c qi->tqi_flags = qinfo->tqi_flags; qi 175 drivers/net/wireless/ath/ath5k/qcu.c qi->tqi_aifs = min(qinfo->tqi_aifs, (u8)0xFC); qi 176 drivers/net/wireless/ath/ath5k/qcu.c qi->tqi_cw_min = ath5k_cw_validate(qinfo->tqi_cw_min); qi 177 drivers/net/wireless/ath/ath5k/qcu.c qi->tqi_cw_max = ath5k_cw_validate(qinfo->tqi_cw_max); qi 178 drivers/net/wireless/ath/ath5k/qcu.c qi->tqi_cbr_period = qinfo->tqi_cbr_period; qi 179 drivers/net/wireless/ath/ath5k/qcu.c qi->tqi_cbr_overflow_limit = qinfo->tqi_cbr_overflow_limit; qi 180 drivers/net/wireless/ath/ath5k/qcu.c qi->tqi_burst_time = qinfo->tqi_burst_time; qi 181 drivers/net/wireless/ath/ath5k/qcu.c qi->tqi_ready_time = qinfo->tqi_ready_time; qi 189 drivers/net/wireless/ath/ath5k/qcu.c qi->tqi_flags |= AR5K_TXQ_FLAG_POST_FR_BKOFF_DIS; qi 37 drivers/net/wireless/ath/ath9k/beacon.c struct ath9k_tx_queue_info qi, qi_be; qi 40 drivers/net/wireless/ath/ath9k/beacon.c ath9k_hw_get_txq_props(ah, sc->beacon.beaconq, &qi); qi 45 drivers/net/wireless/ath/ath9k/beacon.c qi.tqi_aifs = 1; qi 46 drivers/net/wireless/ath/ath9k/beacon.c qi.tqi_cwmin = 0; qi 47 drivers/net/wireless/ath/ath9k/beacon.c qi.tqi_cwmax = 0; qi 52 drivers/net/wireless/ath/ath9k/beacon.c qi.tqi_aifs = qi_be.tqi_aifs; qi 54 drivers/net/wireless/ath/ath9k/beacon.c qi.tqi_cwmin = 2*qi_be.tqi_cwmin; qi 56 drivers/net/wireless/ath/ath9k/beacon.c qi.tqi_cwmin = 4*qi_be.tqi_cwmin; qi 57 drivers/net/wireless/ath/ath9k/beacon.c qi.tqi_cwmax = qi_be.tqi_cwmax; qi 60 drivers/net/wireless/ath/ath9k/beacon.c if (!ath9k_hw_set_txq_props(ah, sc->beacon.beaconq, &qi)) { qi 24 drivers/net/wireless/ath/ath9k/htc_drv_beacon.c struct ath9k_tx_queue_info qi, qi_be; qi 26 drivers/net/wireless/ath/ath9k/htc_drv_beacon.c memset(&qi, 0, sizeof(struct ath9k_tx_queue_info)); qi 29 drivers/net/wireless/ath/ath9k/htc_drv_beacon.c ath9k_hw_get_txq_props(ah, priv->beacon.beaconq, &qi); qi 33 drivers/net/wireless/ath/ath9k/htc_drv_beacon.c qi.tqi_aifs = 1; qi 34 drivers/net/wireless/ath/ath9k/htc_drv_beacon.c qi.tqi_cwmin = 0; qi 35 drivers/net/wireless/ath/ath9k/htc_drv_beacon.c qi.tqi_cwmax = 0; qi 41 drivers/net/wireless/ath/ath9k/htc_drv_beacon.c qi.tqi_aifs = qi_be.tqi_aifs; qi 49 drivers/net/wireless/ath/ath9k/htc_drv_beacon.c qi.tqi_cwmin = 2*qi_be.tqi_cwmin; qi 51 drivers/net/wireless/ath/ath9k/htc_drv_beacon.c qi.tqi_cwmin = 4*qi_be.tqi_cwmin; qi 53 drivers/net/wireless/ath/ath9k/htc_drv_beacon.c qi.tqi_cwmax = qi_be.tqi_cwmax; qi 57 drivers/net/wireless/ath/ath9k/htc_drv_beacon.c if (!ath9k_hw_set_txq_props(ah, priv->beacon.beaconq, &qi)) { qi 1376 drivers/net/wireless/ath/ath9k/htc_drv_main.c struct ath9k_tx_queue_info qi; qi 1385 drivers/net/wireless/ath/ath9k/htc_drv_main.c memset(&qi, 0, sizeof(struct ath9k_tx_queue_info)); qi 1387 drivers/net/wireless/ath/ath9k/htc_drv_main.c qi.tqi_aifs = params->aifs; qi 1388 drivers/net/wireless/ath/ath9k/htc_drv_main.c qi.tqi_cwmin = params->cw_min; qi 1389 drivers/net/wireless/ath/ath9k/htc_drv_main.c qi.tqi_cwmax = params->cw_max; qi 1390 drivers/net/wireless/ath/ath9k/htc_drv_main.c qi.tqi_burstTime = params->txop * 32; qi 1399 drivers/net/wireless/ath/ath9k/htc_drv_main.c ret = ath_htc_txq_update(priv, qnum, &qi); qi 31 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c qi.tqi_subtype = subtype_txq_to_hwq[subtype]; \ qi 32 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c qi.tqi_aifs = ATH9K_TXQ_USEDEFAULT; \ qi 33 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c qi.tqi_cwmin = ATH9K_TXQ_USEDEFAULT; \ qi 34 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c qi.tqi_cwmax = ATH9K_TXQ_USEDEFAULT; \ qi 35 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c qi.tqi_physCompBuf = 0; \ qi 36 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c qi.tqi_qflags = TXQ_FLAG_TXEOLINT_ENABLE | \ qi 193 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c struct ath9k_tx_queue_info qi; qi 195 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c ath9k_hw_get_txq_props(ah, qnum, &qi); qi 197 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c qi.tqi_aifs = qinfo->tqi_aifs; qi 198 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c qi.tqi_cwmin = qinfo->tqi_cwmin / 2; /* XXX */ qi 199 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c qi.tqi_cwmax = qinfo->tqi_cwmax; qi 200 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c qi.tqi_burstTime = qinfo->tqi_burstTime; qi 201 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c qi.tqi_readyTime = qinfo->tqi_readyTime; qi 203 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c if (!ath9k_hw_set_txq_props(ah, qnum, &qi)) { qi 823 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c struct ath9k_tx_queue_info qi; qi 826 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c memset(&qi, 0, sizeof(qi)); qi 829 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c qnum = ath9k_hw_setuptxqueue(priv->ah, ATH9K_TX_QUEUE_DATA, &qi); qi 846 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c struct ath9k_tx_queue_info qi; qi 848 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c memset(&qi, 0, sizeof(qi)); qi 851 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c return ath9k_hw_setuptxqueue(priv->ah, ATH9K_TX_QUEUE_CAB, &qi); qi 22 drivers/net/wireless/ath/ath9k/mac.c struct ath9k_tx_queue_info *qi) qi 201 drivers/net/wireless/ath/ath9k/mac.c struct ath9k_tx_queue_info *qi; qi 203 drivers/net/wireless/ath/ath9k/mac.c qi = &ah->txq[q]; qi 204 drivers/net/wireless/ath/ath9k/mac.c if (qi->tqi_type == ATH9K_TX_QUEUE_INACTIVE) { qi 212 drivers/net/wireless/ath/ath9k/mac.c qi->tqi_ver = qinfo->tqi_ver; qi 213 drivers/net/wireless/ath/ath9k/mac.c qi->tqi_subtype = qinfo->tqi_subtype; qi 214 drivers/net/wireless/ath/ath9k/mac.c qi->tqi_qflags = qinfo->tqi_qflags; qi 215 drivers/net/wireless/ath/ath9k/mac.c qi->tqi_priority = qinfo->tqi_priority; qi 217 drivers/net/wireless/ath/ath9k/mac.c qi->tqi_aifs = min(qinfo->tqi_aifs, 255U); qi 219 drivers/net/wireless/ath/ath9k/mac.c qi->tqi_aifs = INIT_AIFS; qi 222 drivers/net/wireless/ath/ath9k/mac.c qi->tqi_cwmin = 1; qi 223 drivers/net/wireless/ath/ath9k/mac.c while (qi->tqi_cwmin < cw) qi 224 drivers/net/wireless/ath/ath9k/mac.c qi->tqi_cwmin = (qi->tqi_cwmin << 1) | 1; qi 226 drivers/net/wireless/ath/ath9k/mac.c qi->tqi_cwmin = qinfo->tqi_cwmin; qi 229 drivers/net/wireless/ath/ath9k/mac.c qi->tqi_cwmax = 1; qi 230 drivers/net/wireless/ath/ath9k/mac.c while (qi->tqi_cwmax < cw) qi 231 drivers/net/wireless/ath/ath9k/mac.c qi->tqi_cwmax = (qi->tqi_cwmax << 1) | 1; qi 233 drivers/net/wireless/ath/ath9k/mac.c qi->tqi_cwmax = INIT_CWMAX; qi 236 drivers/net/wireless/ath/ath9k/mac.c qi->tqi_shretry = min((u32) qinfo->tqi_shretry, 15U); qi 238 drivers/net/wireless/ath/ath9k/mac.c qi->tqi_shretry = INIT_SH_RETRY; qi 240 drivers/net/wireless/ath/ath9k/mac.c qi->tqi_lgretry = min((u32) qinfo->tqi_lgretry, 15U); qi 242 drivers/net/wireless/ath/ath9k/mac.c qi->tqi_lgretry = INIT_LG_RETRY; qi 243 drivers/net/wireless/ath/ath9k/mac.c qi->tqi_cbrPeriod = qinfo->tqi_cbrPeriod; qi 244 drivers/net/wireless/ath/ath9k/mac.c qi->tqi_cbrOverflowLimit = qinfo->tqi_cbrOverflowLimit; qi 245 drivers/net/wireless/ath/ath9k/mac.c qi->tqi_burstTime = qinfo->tqi_burstTime; qi 246 drivers/net/wireless/ath/ath9k/mac.c qi->tqi_readyTime = qinfo->tqi_readyTime; qi 250 drivers/net/wireless/ath/ath9k/mac.c if (qi->tqi_type == ATH9K_TX_QUEUE_DATA) qi 251 drivers/net/wireless/ath/ath9k/mac.c qi->tqi_intFlags = ATH9K_TXQ_USE_LOCKOUT_BKOFF_DIS; qi 265 drivers/net/wireless/ath/ath9k/mac.c struct ath9k_tx_queue_info *qi; qi 267 drivers/net/wireless/ath/ath9k/mac.c qi = &ah->txq[q]; qi 268 drivers/net/wireless/ath/ath9k/mac.c if (qi->tqi_type == ATH9K_TX_QUEUE_INACTIVE) { qi 274 drivers/net/wireless/ath/ath9k/mac.c qinfo->tqi_qflags = qi->tqi_qflags; qi 275 drivers/net/wireless/ath/ath9k/mac.c qinfo->tqi_ver = qi->tqi_ver; qi 276 drivers/net/wireless/ath/ath9k/mac.c qinfo->tqi_subtype = qi->tqi_subtype; qi 277 drivers/net/wireless/ath/ath9k/mac.c qinfo->tqi_qflags = qi->tqi_qflags; qi 278 drivers/net/wireless/ath/ath9k/mac.c qinfo->tqi_priority = qi->tqi_priority; qi 279 drivers/net/wireless/ath/ath9k/mac.c qinfo->tqi_aifs = qi->tqi_aifs; qi 280 drivers/net/wireless/ath/ath9k/mac.c qinfo->tqi_cwmin = qi->tqi_cwmin; qi 281 drivers/net/wireless/ath/ath9k/mac.c qinfo->tqi_cwmax = qi->tqi_cwmax; qi 282 drivers/net/wireless/ath/ath9k/mac.c qinfo->tqi_shretry = qi->tqi_shretry; qi 283 drivers/net/wireless/ath/ath9k/mac.c qinfo->tqi_lgretry = qi->tqi_lgretry; qi 284 drivers/net/wireless/ath/ath9k/mac.c qinfo->tqi_cbrPeriod = qi->tqi_cbrPeriod; qi 285 drivers/net/wireless/ath/ath9k/mac.c qinfo->tqi_cbrOverflowLimit = qi->tqi_cbrOverflowLimit; qi 286 drivers/net/wireless/ath/ath9k/mac.c qinfo->tqi_burstTime = qi->tqi_burstTime; qi 287 drivers/net/wireless/ath/ath9k/mac.c qinfo->tqi_readyTime = qi->tqi_readyTime; qi 297 drivers/net/wireless/ath/ath9k/mac.c struct ath9k_tx_queue_info *qi; qi 323 drivers/net/wireless/ath/ath9k/mac.c qi = &ah->txq[q]; qi 324 drivers/net/wireless/ath/ath9k/mac.c if (qi->tqi_type != ATH9K_TX_QUEUE_INACTIVE) { qi 328 drivers/net/wireless/ath/ath9k/mac.c memset(qi, 0, sizeof(struct ath9k_tx_queue_info)); qi 329 drivers/net/wireless/ath/ath9k/mac.c qi->tqi_type = type; qi 330 drivers/net/wireless/ath/ath9k/mac.c qi->tqi_physCompBuf = qinfo->tqi_physCompBuf; qi 349 drivers/net/wireless/ath/ath9k/mac.c struct ath9k_tx_queue_info *qi; qi 351 drivers/net/wireless/ath/ath9k/mac.c qi = &ah->txq[q]; qi 352 drivers/net/wireless/ath/ath9k/mac.c if (qi->tqi_type == ATH9K_TX_QUEUE_INACTIVE) { qi 359 drivers/net/wireless/ath/ath9k/mac.c qi->tqi_type = ATH9K_TX_QUEUE_INACTIVE; qi 361 drivers/net/wireless/ath/ath9k/mac.c ath9k_hw_set_txq_interrupts(ah, qi); qi 370 drivers/net/wireless/ath/ath9k/mac.c struct ath9k_tx_queue_info *qi; qi 373 drivers/net/wireless/ath/ath9k/mac.c qi = &ah->txq[q]; qi 374 drivers/net/wireless/ath/ath9k/mac.c if (qi->tqi_type == ATH9K_TX_QUEUE_INACTIVE) { qi 381 drivers/net/wireless/ath/ath9k/mac.c if (qi->tqi_cwmin == ATH9K_TXQ_USEDEFAULT) { qi 386 drivers/net/wireless/ath/ath9k/mac.c cwMin = qi->tqi_cwmin; qi 392 drivers/net/wireless/ath/ath9k/mac.c SM(qi->tqi_cwmax, AR_D_LCL_IFS_CWMAX) | qi 393 drivers/net/wireless/ath/ath9k/mac.c SM(qi->tqi_aifs, AR_D_LCL_IFS_AIFS)); qi 398 drivers/net/wireless/ath/ath9k/mac.c SM(qi->tqi_shretry, AR_D_RETRY_LIMIT_FR_SH)); qi 409 drivers/net/wireless/ath/ath9k/mac.c if (qi->tqi_cbrPeriod) { qi 411 drivers/net/wireless/ath/ath9k/mac.c SM(qi->tqi_cbrPeriod, AR_Q_CBRCFG_INTERVAL) | qi 412 drivers/net/wireless/ath/ath9k/mac.c SM(qi->tqi_cbrOverflowLimit, AR_Q_CBRCFG_OVF_THRESH)); qi 414 drivers/net/wireless/ath/ath9k/mac.c (qi->tqi_cbrOverflowLimit ? qi 417 drivers/net/wireless/ath/ath9k/mac.c if (qi->tqi_readyTime && (qi->tqi_type != ATH9K_TX_QUEUE_CAB)) { qi 419 drivers/net/wireless/ath/ath9k/mac.c SM(qi->tqi_readyTime, AR_Q_RDYTIMECFG_DURATION) | qi 424 drivers/net/wireless/ath/ath9k/mac.c SM(qi->tqi_burstTime, AR_D_CHNTIME_DUR) | qi 425 drivers/net/wireless/ath/ath9k/mac.c (qi->tqi_burstTime ? AR_D_CHNTIME_EN : 0)); qi 427 drivers/net/wireless/ath/ath9k/mac.c if (qi->tqi_burstTime qi 428 drivers/net/wireless/ath/ath9k/mac.c && (qi->tqi_qflags & TXQ_FLAG_RDYTIME_EXP_POLICY_ENABLE)) qi 431 drivers/net/wireless/ath/ath9k/mac.c if (qi->tqi_qflags & TXQ_FLAG_BACKOFF_DISABLE) qi 436 drivers/net/wireless/ath/ath9k/mac.c if (qi->tqi_qflags & TXQ_FLAG_FRAG_BURST_BACKOFF_ENABLE) qi 439 drivers/net/wireless/ath/ath9k/mac.c switch (qi->tqi_type) { qi 465 drivers/net/wireless/ath/ath9k/mac.c | SM(qi->tqi_aifs, AR_D_LCL_IFS_AIFS)); qi 475 drivers/net/wireless/ath/ath9k/mac.c value = (qi->tqi_readyTime - qi 497 drivers/net/wireless/ath/ath9k/mac.c if (qi->tqi_intFlags & ATH9K_TXQ_USE_LOCKOUT_BKOFF_DIS) { qi 508 drivers/net/wireless/ath/ath9k/mac.c if (qi->tqi_qflags & TXQ_FLAG_TXINT_ENABLE) { qi 512 drivers/net/wireless/ath/ath9k/mac.c if (qi->tqi_qflags & TXQ_FLAG_TXDESCINT_ENABLE) qi 514 drivers/net/wireless/ath/ath9k/mac.c if (qi->tqi_qflags & TXQ_FLAG_TXEOLINT_ENABLE) qi 516 drivers/net/wireless/ath/ath9k/mac.c if (qi->tqi_qflags & TXQ_FLAG_TXURNINT_ENABLE) qi 518 drivers/net/wireless/ath/ath9k/mac.c ath9k_hw_set_txq_interrupts(ah, qi); qi 744 drivers/net/wireless/ath/ath9k/mac.c struct ath9k_tx_queue_info qi; qi 746 drivers/net/wireless/ath/ath9k/mac.c memset(&qi, 0, sizeof(qi)); qi 747 drivers/net/wireless/ath/ath9k/mac.c qi.tqi_aifs = 1; qi 748 drivers/net/wireless/ath/ath9k/mac.c qi.tqi_cwmin = 0; qi 749 drivers/net/wireless/ath/ath9k/mac.c qi.tqi_cwmax = 0; qi 752 drivers/net/wireless/ath/ath9k/mac.c qi.tqi_qflags = TXQ_FLAG_TXINT_ENABLE; qi 754 drivers/net/wireless/ath/ath9k/mac.c return ath9k_hw_setuptxqueue(ah, ATH9K_TX_QUEUE_BEACON, &qi); qi 1637 drivers/net/wireless/ath/ath9k/main.c struct ath9k_tx_queue_info qi; qi 1648 drivers/net/wireless/ath/ath9k/main.c memset(&qi, 0, sizeof(struct ath9k_tx_queue_info)); qi 1650 drivers/net/wireless/ath/ath9k/main.c qi.tqi_aifs = params->aifs; qi 1651 drivers/net/wireless/ath/ath9k/main.c qi.tqi_cwmin = params->cw_min; qi 1652 drivers/net/wireless/ath/ath9k/main.c qi.tqi_cwmax = params->cw_max; qi 1653 drivers/net/wireless/ath/ath9k/main.c qi.tqi_burstTime = params->txop * 32; qi 1660 drivers/net/wireless/ath/ath9k/main.c ath_update_max_aggr_framelen(sc, queue, qi.tqi_burstTime); qi 1661 drivers/net/wireless/ath/ath9k/main.c ret = ath_txq_update(sc, txq->axq_qnum, &qi); qi 1696 drivers/net/wireless/ath/ath9k/xmit.c struct ath9k_tx_queue_info qi; qi 1705 drivers/net/wireless/ath/ath9k/xmit.c memset(&qi, 0, sizeof(qi)); qi 1706 drivers/net/wireless/ath/ath9k/xmit.c qi.tqi_subtype = subtype_txq_to_hwq[subtype]; qi 1707 drivers/net/wireless/ath/ath9k/xmit.c qi.tqi_aifs = ATH9K_TXQ_USEDEFAULT; qi 1708 drivers/net/wireless/ath/ath9k/xmit.c qi.tqi_cwmin = ATH9K_TXQ_USEDEFAULT; qi 1709 drivers/net/wireless/ath/ath9k/xmit.c qi.tqi_cwmax = ATH9K_TXQ_USEDEFAULT; qi 1710 drivers/net/wireless/ath/ath9k/xmit.c qi.tqi_physCompBuf = 0; qi 1728 drivers/net/wireless/ath/ath9k/xmit.c qi.tqi_qflags = TXQ_FLAG_TXINT_ENABLE; qi 1731 drivers/net/wireless/ath/ath9k/xmit.c qi.tqi_qflags = TXQ_FLAG_TXDESCINT_ENABLE; qi 1733 drivers/net/wireless/ath/ath9k/xmit.c qi.tqi_qflags = TXQ_FLAG_TXEOLINT_ENABLE | qi 1736 drivers/net/wireless/ath/ath9k/xmit.c axq_qnum = ath9k_hw_setuptxqueue(ah, qtype, &qi); qi 1770 drivers/net/wireless/ath/ath9k/xmit.c struct ath9k_tx_queue_info qi; qi 1774 drivers/net/wireless/ath/ath9k/xmit.c ath9k_hw_get_txq_props(ah, qnum, &qi); qi 1775 drivers/net/wireless/ath/ath9k/xmit.c qi.tqi_aifs = qinfo->tqi_aifs; qi 1776 drivers/net/wireless/ath/ath9k/xmit.c qi.tqi_cwmin = qinfo->tqi_cwmin; qi 1777 drivers/net/wireless/ath/ath9k/xmit.c qi.tqi_cwmax = qinfo->tqi_cwmax; qi 1778 drivers/net/wireless/ath/ath9k/xmit.c qi.tqi_burstTime = qinfo->tqi_burstTime; qi 1779 drivers/net/wireless/ath/ath9k/xmit.c qi.tqi_readyTime = qinfo->tqi_readyTime; qi 1781 drivers/net/wireless/ath/ath9k/xmit.c if (!ath9k_hw_set_txq_props(ah, qnum, &qi)) { qi 1794 drivers/net/wireless/ath/ath9k/xmit.c struct ath9k_tx_queue_info qi; qi 1798 drivers/net/wireless/ath/ath9k/xmit.c ath9k_hw_get_txq_props(sc->sc_ah, qnum, &qi); qi 1800 drivers/net/wireless/ath/ath9k/xmit.c qi.tqi_readyTime = (TU_TO_USEC(cur_conf->beacon_interval) * qi 1802 drivers/net/wireless/ath/ath9k/xmit.c ath_txq_update(sc, qnum, &qi); qi 373 drivers/net/wireless/ath/carl9170/debug.c #define DEBUGFS_QUEUE_DUMP(q, qi) \ qi 374 drivers/net/wireless/ath/carl9170/debug.c static char *carl9170_debugfs_##q ##_##qi ##_read(struct ar9170 *ar, \ qi 377 drivers/net/wireless/ath/carl9170/debug.c carl9170_debugfs_queue_dump(ar, buf, len, bufsize, &ar->q[qi]); \ qi 380 drivers/net/wireless/ath/carl9170/debug.c DEBUGFS_DECLARE_RO_FILE(q##_##qi, 8000); qi 31 drivers/regulator/mt6323-regulator.c u32 qi; qi 57 drivers/regulator/mt6323-regulator.c .qi = BIT(13), \ qi 80 drivers/regulator/mt6323-regulator.c .qi = BIT(15), \ qi 100 drivers/regulator/mt6323-regulator.c .qi = BIT(15), \ qi 169 drivers/regulator/mt6323-regulator.c return (regval & info->qi) ? REGULATOR_STATUS_ON : REGULATOR_STATUS_OFF; qi 28 drivers/regulator/mt6358-regulator.c u32 qi; qi 61 drivers/regulator/mt6358-regulator.c .qi = BIT(0), \ qi 89 drivers/regulator/mt6358-regulator.c .qi = BIT(15), \ qi 118 drivers/regulator/mt6358-regulator.c .qi = BIT(0), \ qi 137 drivers/regulator/mt6358-regulator.c .qi = BIT(15), \ qi 325 drivers/regulator/mt6358-regulator.c return (regval & info->qi) ? REGULATOR_STATUS_ON : REGULATOR_STATUS_OFF; qi 31 drivers/regulator/mt6397-regulator.c u32 qi; qi 59 drivers/regulator/mt6397-regulator.c .qi = BIT(13), \ qi 85 drivers/regulator/mt6397-regulator.c .qi = BIT(15), \ qi 102 drivers/regulator/mt6397-regulator.c .qi = BIT(15), \ qi 218 drivers/regulator/mt6397-regulator.c return (regval & info->qi) ? REGULATOR_STATUS_ON : REGULATOR_STATUS_OFF; qi 1401 fs/cifs/smb2ops.c struct smb_query_info qi; qi 1426 fs/cifs/smb2ops.c if (copy_from_user(&qi, arg, sizeof(struct smb_query_info))) qi 1429 fs/cifs/smb2ops.c if (qi.output_buffer_length > 1024) qi 1438 fs/cifs/smb2ops.c buffer = kmalloc(qi.output_buffer_length, GFP_KERNEL); qi 1443 fs/cifs/smb2ops.c qi.output_buffer_length)) { qi 1463 fs/cifs/smb2ops.c if (qi.flags & PASSTHRU_FSCTL) { qi 1464 fs/cifs/smb2ops.c switch (qi.info_type & FSCTL_DEVICE_ACCESS_MASK) { qi 1478 fs/cifs/smb2ops.c } else if (qi.flags & PASSTHRU_SET_INFO) { qi 1490 fs/cifs/smb2ops.c if (qi.flags & PASSTHRU_FSCTL) { qi 1501 fs/cifs/smb2ops.c qi.info_type, true, buffer, qi 1502 fs/cifs/smb2ops.c qi.output_buffer_length, qi 1507 fs/cifs/smb2ops.c } else if (qi.flags == PASSTHRU_SET_INFO) { qi 1525 fs/cifs/smb2ops.c } else if (qi.flags == PASSTHRU_QUERY_INFO) { qi 1531 fs/cifs/smb2ops.c COMPOUND_FID, qi.file_info_class, qi 1532 fs/cifs/smb2ops.c qi.info_type, qi.additional_information, qi 1533 fs/cifs/smb2ops.c qi.input_buffer_length, qi 1534 fs/cifs/smb2ops.c qi.output_buffer_length, buffer); qi 1536 fs/cifs/smb2ops.c cifs_tcon_dbg(VFS, "invalid passthru query flags: 0x%x\n", qi.flags); qi 1561 fs/cifs/smb2ops.c if (qi.flags & PASSTHRU_FSCTL) { qi 1564 fs/cifs/smb2ops.c if (le32_to_cpu(io_rsp->OutputCount) < qi.input_buffer_length) qi 1565 fs/cifs/smb2ops.c qi.input_buffer_length = le32_to_cpu(io_rsp->OutputCount); qi 1566 fs/cifs/smb2ops.c if (qi.input_buffer_length > 0 && qi 1567 fs/cifs/smb2ops.c le32_to_cpu(io_rsp->OutputOffset) + qi.input_buffer_length > rsp_iov[1].iov_len) { qi 1571 fs/cifs/smb2ops.c if (copy_to_user(&pqi->input_buffer_length, &qi.input_buffer_length, qi 1572 fs/cifs/smb2ops.c sizeof(qi.input_buffer_length))) { qi 1578 fs/cifs/smb2ops.c qi.input_buffer_length)) { qi 1585 fs/cifs/smb2ops.c if (le32_to_cpu(qi_rsp->OutputBufferLength) < qi.input_buffer_length) qi 1586 fs/cifs/smb2ops.c qi.input_buffer_length = le32_to_cpu(qi_rsp->OutputBufferLength); qi 1587 fs/cifs/smb2ops.c if (copy_to_user(&pqi->input_buffer_length, &qi.input_buffer_length, qi 1588 fs/cifs/smb2ops.c sizeof(qi.input_buffer_length))) { qi 1592 fs/cifs/smb2ops.c if (copy_to_user(pqi + 1, qi_rsp->Buffer, qi.input_buffer_length)) { qi 1601 fs/cifs/smb2ops.c if (qi.flags & PASSTHRU_FSCTL) qi 83 fs/xfs/scrub/quota.c struct xfs_quotainfo *qi = mp->m_quotainfo; qi 101 fs/xfs/scrub/quota.c offset = id / qi->qi_dqperchunk; qi 191 fs/xfs/scrub/quota.c struct xfs_quotainfo *qi = sc->mp->m_quotainfo; qi 202 fs/xfs/scrub/quota.c max_dqid_off = ((xfs_dqid_t)-1) / qi->qi_dqperchunk; qi 230 fs/xfs/scrub/quota.c struct xfs_quotainfo *qi = mp->m_quotainfo; qi 256 fs/xfs/scrub/quota.c sqi.last_id * qi->qi_dqperchunk, &error)) qi 658 fs/xfs/xfs_dquot.c struct xfs_quotainfo *qi, qi 665 fs/xfs/xfs_dquot.c mutex_lock(&qi->qi_tree_lock); qi 668 fs/xfs/xfs_dquot.c mutex_unlock(&qi->qi_tree_lock); qi 676 fs/xfs/xfs_dquot.c mutex_unlock(&qi->qi_tree_lock); qi 683 fs/xfs/xfs_dquot.c mutex_unlock(&qi->qi_tree_lock); qi 699 fs/xfs/xfs_dquot.c struct xfs_quotainfo *qi, qi 706 fs/xfs/xfs_dquot.c mutex_lock(&qi->qi_tree_lock); qi 711 fs/xfs/xfs_dquot.c mutex_unlock(&qi->qi_tree_lock); qi 720 fs/xfs/xfs_dquot.c qi->qi_dquots++; qi 721 fs/xfs/xfs_dquot.c mutex_unlock(&qi->qi_tree_lock); qi 766 fs/xfs/xfs_dquot.c struct xfs_quotainfo *qi = mp->m_quotainfo; qi 767 fs/xfs/xfs_dquot.c struct radix_tree_root *tree = xfs_dquot_tree(qi, type); qi 776 fs/xfs/xfs_dquot.c dqp = xfs_qm_dqget_cache_lookup(mp, qi, tree, id); qi 786 fs/xfs/xfs_dquot.c error = xfs_qm_dqget_cache_insert(mp, qi, tree, id, dqp); qi 855 fs/xfs/xfs_dquot.c struct xfs_quotainfo *qi = mp->m_quotainfo; qi 856 fs/xfs/xfs_dquot.c struct radix_tree_root *tree = xfs_dquot_tree(qi, type); qi 871 fs/xfs/xfs_dquot.c dqp = xfs_qm_dqget_cache_lookup(mp, qi, tree, id); qi 910 fs/xfs/xfs_dquot.c error = xfs_qm_dqget_cache_insert(mp, qi, tree, id, dqp); qi 977 fs/xfs/xfs_dquot.c struct xfs_quotainfo *qi = dqp->q_mount->m_quotainfo; qi 980 fs/xfs/xfs_dquot.c if (list_lru_add(&qi->qi_lru, &dqp->q_lru)) qi 35 fs/xfs/xfs_qm.c STATIC void xfs_qm_destroy_quotainos(xfs_quotainfo_t *qi); qi 53 fs/xfs/xfs_qm.c struct xfs_quotainfo *qi = mp->m_quotainfo; qi 54 fs/xfs/xfs_qm.c struct radix_tree_root *tree = xfs_dquot_tree(qi, type); qi 70 fs/xfs/xfs_qm.c mutex_lock(&qi->qi_tree_lock); qi 74 fs/xfs/xfs_qm.c mutex_unlock(&qi->qi_tree_lock); qi 92 fs/xfs/xfs_qm.c mutex_unlock(&qi->qi_tree_lock); qi 122 fs/xfs/xfs_qm.c struct xfs_quotainfo *qi = mp->m_quotainfo; qi 162 fs/xfs/xfs_qm.c radix_tree_delete(xfs_dquot_tree(qi, dqp->q_core.d_flags), qi 164 fs/xfs/xfs_qm.c qi->qi_dquots--; qi 171 fs/xfs/xfs_qm.c list_lru_del(&qi->qi_lru, &dqp->q_lru); qi 499 fs/xfs/xfs_qm.c struct xfs_quotainfo *qi = container_of(shrink, qi 511 fs/xfs/xfs_qm.c freed = list_lru_shrink_walk(&qi->qi_lru, sc, qi 534 fs/xfs/xfs_qm.c struct xfs_quotainfo *qi = container_of(shrink, qi 537 fs/xfs/xfs_qm.c return list_lru_shrink_count(&qi->qi_lru, sc); qi 714 fs/xfs/xfs_qm.c xfs_quotainfo_t *qi; qi 716 fs/xfs/xfs_qm.c qi = mp->m_quotainfo; qi 717 fs/xfs/xfs_qm.c ASSERT(qi != NULL); qi 719 fs/xfs/xfs_qm.c unregister_shrinker(&qi->qi_shrinker); qi 720 fs/xfs/xfs_qm.c list_lru_destroy(&qi->qi_lru); qi 721 fs/xfs/xfs_qm.c xfs_qm_destroy_quotainos(qi); qi 722 fs/xfs/xfs_qm.c mutex_destroy(&qi->qi_tree_lock); qi 723 fs/xfs/xfs_qm.c mutex_destroy(&qi->qi_quotaofflock); qi 724 fs/xfs/xfs_qm.c kmem_free(qi); qi 1562 fs/xfs/xfs_qm.c xfs_quotainfo_t *qi) qi 1564 fs/xfs/xfs_qm.c if (qi->qi_uquotaip) { qi 1565 fs/xfs/xfs_qm.c xfs_irele(qi->qi_uquotaip); qi 1566 fs/xfs/xfs_qm.c qi->qi_uquotaip = NULL; /* paranoia */ qi 1568 fs/xfs/xfs_qm.c if (qi->qi_gquotaip) { qi 1569 fs/xfs/xfs_qm.c xfs_irele(qi->qi_gquotaip); qi 1570 fs/xfs/xfs_qm.c qi->qi_gquotaip = NULL; qi 1572 fs/xfs/xfs_qm.c if (qi->qi_pquotaip) { qi 1573 fs/xfs/xfs_qm.c xfs_irele(qi->qi_pquotaip); qi 1574 fs/xfs/xfs_qm.c qi->qi_pquotaip = NULL; qi 1583 fs/xfs/xfs_qm.c struct xfs_quotainfo *qi = mp->m_quotainfo; qi 1585 fs/xfs/xfs_qm.c mutex_lock(&qi->qi_tree_lock); qi 1586 fs/xfs/xfs_qm.c radix_tree_delete(xfs_dquot_tree(qi, dqp->q_core.d_flags), qi 1589 fs/xfs/xfs_qm.c qi->qi_dquots--; qi 1590 fs/xfs/xfs_qm.c mutex_unlock(&qi->qi_tree_lock); qi 84 fs/xfs/xfs_qm.h struct xfs_quotainfo *qi, qi 89 fs/xfs/xfs_qm.h return &qi->qi_uquota_tree; qi 91 fs/xfs/xfs_qm.h return &qi->qi_gquota_tree; qi 93 fs/xfs/xfs_qm.h return &qi->qi_pquota_tree; qi 167 fs/xfs/xfs_qm.h xfs_get_defquota(struct xfs_dquot *dqp, struct xfs_quotainfo *qi) qi 172 fs/xfs/xfs_qm.h defq = &qi->qi_usr_default; qi 174 fs/xfs/xfs_qm.h defq = &qi->qi_grp_default; qi 177 fs/xfs/xfs_qm.h defq = &qi->qi_prj_default; qi 548 include/linux/intel-iommu.h struct q_inval *qi; /* Queued invalidation info */