qd                309 block/blk-iolatency.c static inline unsigned long scale_amount(unsigned long qd, bool up)
qd                311 block/blk-iolatency.c 	return max(up ? qd >> SCALE_UP_FACTOR : qd >> SCALE_DOWN_FACTOR, 1UL);
qd                327 block/blk-iolatency.c 	unsigned long qd = blkiolat->rqos.q->nr_requests;
qd                328 block/blk-iolatency.c 	unsigned long scale = scale_amount(qd, up);
qd                330 block/blk-iolatency.c 	unsigned long max_scale = qd << 1;
qd                340 block/blk-iolatency.c 		else if (diff > qd)
qd                351 block/blk-iolatency.c 		if (diff > qd) {
qd                367 block/blk-iolatency.c 	unsigned long qd = iolat->blkiolat->rqos.q->nr_requests;
qd                368 block/blk-iolatency.c 	unsigned long scale = scale_amount(qd, up);
qd                371 block/blk-iolatency.c 	if (old > qd)
qd                372 block/blk-iolatency.c 		old = qd;
qd                378 block/blk-iolatency.c 		if (old < qd) {
qd                380 block/blk-iolatency.c 			old = min(old, qd);
qd                623 drivers/atm/firestream.c static int qd;
qd               1193 drivers/atm/firestream.c 	da[qd] = td;
qd               1194 drivers/atm/firestream.c 	dq[qd].flags = td->flags;
qd               1195 drivers/atm/firestream.c 	dq[qd].next  = td->next;
qd               1196 drivers/atm/firestream.c 	dq[qd].bsa   = td->bsa;
qd               1197 drivers/atm/firestream.c 	dq[qd].skb   = td->skb;
qd               1198 drivers/atm/firestream.c 	dq[qd].dev   = td->dev;
qd               1199 drivers/atm/firestream.c 	qd++;
qd               1200 drivers/atm/firestream.c 	if (qd >= 60) qd = 0;
qd               1948 drivers/atm/firestream.c 			i, da[qd], dq[qd].flags, dq[qd].bsa, dq[qd].skb, dq[qd].dev);
qd               1949 drivers/atm/firestream.c 		qd++;
qd               1950 drivers/atm/firestream.c 		if (qd >= 60) qd = 0;
qd                883 drivers/block/xen-blkfront.c 			  const struct blk_mq_queue_data *qd)
qd                892 drivers/block/xen-blkfront.c 	blk_mq_start_request(qd->rq);
qd                897 drivers/block/xen-blkfront.c 	if (blkif_request_flush_invalid(qd->rq, rinfo->dev_info))
qd                900 drivers/block/xen-blkfront.c 	if (blkif_queue_request(qd->rq, rinfo))
qd                 47 drivers/clk/at91/clk-audio-pll.c #define AUDIO_PLL_QDPAD(qd, div)	((AT91_PMC_AUDIO_PLL_QDPAD_EXTDIV(qd) & \
qd               9622 drivers/infiniband/hw/hfi1/chip.c 	struct qsfp_data *qd;
qd               9626 drivers/infiniband/hw/hfi1/chip.c 	qd = container_of(work, struct qsfp_data, qsfp_work);
qd               9627 drivers/infiniband/hw/hfi1/chip.c 	ppd = qd->ppd;
qd               9647 drivers/infiniband/hw/hfi1/chip.c 	if (qd->cache_refresh_required) {
qd               9661 drivers/infiniband/hw/hfi1/chip.c 	if (qd->check_interrupt_flags) {
qd               2980 drivers/infiniband/hw/qib/qib_iba7322.c 		struct qib_qsfp_data *qd;
qd               2991 drivers/infiniband/hw/qib/qib_iba7322.c 			qd = &ppd->cpspec->qsfp_data;
qd               2997 drivers/infiniband/hw/qib/qib_iba7322.c 				qd->t_insert = jiffies;
qd               2998 drivers/infiniband/hw/qib/qib_iba7322.c 				queue_work(ib_wq, &qd->work);
qd               5574 drivers/infiniband/hw/qib/qib_iba7322.c 			struct qib_qsfp_data *qd =
qd               5586 drivers/infiniband/hw/qib/qib_iba7322.c 				qd->t_insert = jiffies;
qd               5587 drivers/infiniband/hw/qib/qib_iba7322.c 				queue_work(ib_wq, &qd->work);
qd               5937 drivers/infiniband/hw/qib/qib_iba7322.c 	struct qib_qsfp_data *qd;
qd               5944 drivers/infiniband/hw/qib/qib_iba7322.c 	qd = container_of(work, struct qib_qsfp_data, work);
qd               5945 drivers/infiniband/hw/qib/qib_iba7322.c 	ppd = qd->ppd;
qd               5946 drivers/infiniband/hw/qib/qib_iba7322.c 	pwrup = qd->t_insert +
qd               5972 drivers/infiniband/hw/qib/qib_iba7322.c 		ret = qib_refresh_qsfp_cache(ppd, &qd->cache);
qd               5981 drivers/infiniband/hw/qib/qib_iba7322.c 			if (QSFP_IS_ACTIVE_FAR(qd->cache.tech))
qd               5983 drivers/infiniband/hw/qib/qib_iba7322.c 			else if (qd->cache.atten[1] >= qib_long_atten &&
qd               5984 drivers/infiniband/hw/qib/qib_iba7322.c 				 QSFP_IS_CU(qd->cache.tech))
qd               6021 drivers/infiniband/hw/qib/qib_iba7322.c 	struct qib_qsfp_data *qd = &ppd->cpspec->qsfp_data;
qd               6026 drivers/infiniband/hw/qib/qib_iba7322.c 	qd->ppd = ppd;
qd               6027 drivers/infiniband/hw/qib/qib_iba7322.c 	qib_qsfp_init(qd, qsfp_7322_event);
qd               7621 drivers/infiniband/hw/qib/qib_iba7322.c 	struct qib_qsfp_cache *qd = &ppd->cpspec->qsfp_data.cache;
qd               7628 drivers/infiniband/hw/qib/qib_iba7322.c 		if (!memcmp(v->oui, qd->oui, QSFP_VOUI_LEN) &&
qd               7630 drivers/infiniband/hw/qib/qib_iba7322.c 		     !memcmp(v->partnum, qd->partnum, QSFP_PN_LEN))) {
qd               7640 drivers/infiniband/hw/qib/qib_iba7322.c 	if (!override && QSFP_IS_ACTIVE(qd->tech)) {
qd               7647 drivers/infiniband/hw/qib/qib_iba7322.c 	if (!override && QSFP_HAS_ATTEN(qd->tech) && (qd->atten[0] ||
qd               7648 drivers/infiniband/hw/qib/qib_iba7322.c 						      qd->atten[1])) {
qd               7649 drivers/infiniband/hw/qib/qib_iba7322.c 		*sdr_dds = get_atten_table(txdds_sdr, qd->atten[0]);
qd               7650 drivers/infiniband/hw/qib/qib_iba7322.c 		*ddr_dds = get_atten_table(txdds_ddr, qd->atten[0]);
qd               7651 drivers/infiniband/hw/qib/qib_iba7322.c 		*qdr_dds = get_atten_table(txdds_qdr, qd->atten[1]);
qd                461 drivers/infiniband/hw/qib/qib_qsfp.c void qib_qsfp_init(struct qib_qsfp_data *qd,
qd                466 drivers/infiniband/hw/qib/qib_qsfp.c 	struct qib_devdata *dd = qd->ppd->dd;
qd                469 drivers/infiniband/hw/qib/qib_qsfp.c 	INIT_WORK(&qd->work, fevent);
qd                478 drivers/infiniband/hw/qib/qib_qsfp.c 	if (qd->ppd->hw_pidx) {
qd                187 drivers/infiniband/hw/qib/qib_qsfp.h extern void qib_qsfp_init(struct qib_qsfp_data *qd,
qd                585 drivers/mailbox/ti-msgmgr.c 				 const struct ti_msgmgr_valid_queue_desc *qd,
qd                591 drivers/mailbox/ti-msgmgr.c 	qinst->proxy_id = qd->proxy_id;
qd                592 drivers/mailbox/ti-msgmgr.c 	qinst->queue_id = qd->queue_id;
qd                626 drivers/mailbox/ti-msgmgr.c 		qinst->is_tx = qd->is_tx;
qd               3073 drivers/s390/block/dasd.c 				    const struct blk_mq_queue_data *qd)
qd               3077 drivers/s390/block/dasd.c 	struct request *req = qd->rq;
qd                283 drivers/s390/block/scm_blk.c 			   const struct blk_mq_queue_data *qd)
qd                288 drivers/s390/block/scm_blk.c 	struct request *req = qd->rq;
qd                323 drivers/s390/block/scm_blk.c 	if (qd->last || scmrq->aob->request.msb_count == nr_requests_per_io) {
qd                334 drivers/s390/block/scm_blk.c 	struct scm_queue *qd = kzalloc(sizeof(*qd), GFP_KERNEL);
qd                336 drivers/s390/block/scm_blk.c 	if (!qd)
qd                339 drivers/s390/block/scm_blk.c 	spin_lock_init(&qd->lock);
qd                340 drivers/s390/block/scm_blk.c 	hctx->driver_data = qd;
qd                347 drivers/s390/block/scm_blk.c 	struct scm_queue *qd = hctx->driver_data;
qd                349 drivers/s390/block/scm_blk.c 	WARN_ON(qd->scmrq);
qd               3351 drivers/scsi/aacraid/aachba.c 	struct aac_query_disk qd;
qd               3357 drivers/scsi/aacraid/aachba.c 	if (copy_from_user(&qd, arg, sizeof (struct aac_query_disk)))
qd               3359 drivers/scsi/aacraid/aachba.c 	if (qd.cnum == -1) {
qd               3360 drivers/scsi/aacraid/aachba.c 		if (qd.id < 0 || qd.id >= dev->maximum_num_containers)
qd               3362 drivers/scsi/aacraid/aachba.c 		qd.cnum = qd.id;
qd               3363 drivers/scsi/aacraid/aachba.c 	} else if ((qd.bus == -1) && (qd.id == -1) && (qd.lun == -1)) {
qd               3364 drivers/scsi/aacraid/aachba.c 		if (qd.cnum < 0 || qd.cnum >= dev->maximum_num_containers)
qd               3366 drivers/scsi/aacraid/aachba.c 		qd.instance = dev->scsi_host_ptr->host_no;
qd               3367 drivers/scsi/aacraid/aachba.c 		qd.bus = 0;
qd               3368 drivers/scsi/aacraid/aachba.c 		qd.id = CONTAINER_TO_ID(qd.cnum);
qd               3369 drivers/scsi/aacraid/aachba.c 		qd.lun = CONTAINER_TO_LUN(qd.cnum);
qd               3373 drivers/scsi/aacraid/aachba.c 	qd.valid = fsa_dev_ptr[qd.cnum].valid != 0;
qd               3374 drivers/scsi/aacraid/aachba.c 	qd.locked = fsa_dev_ptr[qd.cnum].locked;
qd               3375 drivers/scsi/aacraid/aachba.c 	qd.deleted = fsa_dev_ptr[qd.cnum].deleted;
qd               3377 drivers/scsi/aacraid/aachba.c 	if (fsa_dev_ptr[qd.cnum].devname[0] == '\0')
qd               3378 drivers/scsi/aacraid/aachba.c 		qd.unmapped = 1;
qd               3380 drivers/scsi/aacraid/aachba.c 		qd.unmapped = 0;
qd               3382 drivers/scsi/aacraid/aachba.c 	strlcpy(qd.name, fsa_dev_ptr[qd.cnum].devname,
qd               3383 drivers/scsi/aacraid/aachba.c 	  min(sizeof(qd.name), sizeof(fsa_dev_ptr[qd.cnum].devname) + 1));
qd               3385 drivers/scsi/aacraid/aachba.c 	if (copy_to_user(arg, &qd, sizeof (struct aac_query_disk)))
qd                 18 fs/erofs/namei.c 				   const struct erofs_qstr *qd,
qd                 28 fs/erofs/namei.c 	DBG_BUGON(qd->name > qd->end);
qd                 32 fs/erofs/namei.c 	while (qd->name + i < qd->end && qd->name[i] != '\0') {
qd                 33 fs/erofs/namei.c 		if (qn->name[i] != qd->name[i]) {
qd                 35 fs/erofs/namei.c 			return qn->name[i] > qd->name[i] ? 1 : -1;
qd                108 fs/gfs2/quota.c 	struct gfs2_quota_data *qd = container_of(rcu, struct gfs2_quota_data, qd_rcu);
qd                109 fs/gfs2/quota.c 	kmem_cache_free(gfs2_quotad_cachep, qd);
qd                114 fs/gfs2/quota.c 	struct gfs2_quota_data *qd;
qd                118 fs/gfs2/quota.c 		qd = list_entry(list->next, struct gfs2_quota_data, qd_lru);
qd                119 fs/gfs2/quota.c 		sdp = qd->qd_gl->gl_name.ln_sbd;
qd                121 fs/gfs2/quota.c 		list_del(&qd->qd_lru);
qd                125 fs/gfs2/quota.c 		list_del(&qd->qd_list);
qd                128 fs/gfs2/quota.c 		spin_lock_bucket(qd->qd_hash);
qd                129 fs/gfs2/quota.c 		hlist_bl_del_rcu(&qd->qd_hlist);
qd                130 fs/gfs2/quota.c 		spin_unlock_bucket(qd->qd_hash);
qd                132 fs/gfs2/quota.c 		gfs2_assert_warn(sdp, !qd->qd_change);
qd                133 fs/gfs2/quota.c 		gfs2_assert_warn(sdp, !qd->qd_slot_count);
qd                134 fs/gfs2/quota.c 		gfs2_assert_warn(sdp, !qd->qd_bh_count);
qd                136 fs/gfs2/quota.c 		gfs2_glock_put(qd->qd_gl);
qd                140 fs/gfs2/quota.c 		call_rcu(&qd->qd_rcu, gfs2_qd_dealloc);
qd                149 fs/gfs2/quota.c 	struct gfs2_quota_data *qd = list_entry(item, struct gfs2_quota_data, qd_lru);
qd                151 fs/gfs2/quota.c 	if (!spin_trylock(&qd->qd_lockref.lock))
qd                154 fs/gfs2/quota.c 	if (qd->qd_lockref.count == 0) {
qd                155 fs/gfs2/quota.c 		lockref_mark_dead(&qd->qd_lockref);
qd                156 fs/gfs2/quota.c 		list_lru_isolate_move(lru, &qd->qd_lru, dispose);
qd                159 fs/gfs2/quota.c 	spin_unlock(&qd->qd_lockref.lock);
qd                194 fs/gfs2/quota.c static u64 qd2index(struct gfs2_quota_data *qd)
qd                196 fs/gfs2/quota.c 	struct kqid qid = qd->qd_id;
qd                201 fs/gfs2/quota.c static u64 qd2offset(struct gfs2_quota_data *qd)
qd                205 fs/gfs2/quota.c 	offset = qd2index(qd);
qd                213 fs/gfs2/quota.c 	struct gfs2_quota_data *qd;
qd                216 fs/gfs2/quota.c 	qd = kmem_cache_zalloc(gfs2_quotad_cachep, GFP_NOFS);
qd                217 fs/gfs2/quota.c 	if (!qd)
qd                220 fs/gfs2/quota.c 	qd->qd_sbd = sdp;
qd                221 fs/gfs2/quota.c 	qd->qd_lockref.count = 1;
qd                222 fs/gfs2/quota.c 	spin_lock_init(&qd->qd_lockref.lock);
qd                223 fs/gfs2/quota.c 	qd->qd_id = qid;
qd                224 fs/gfs2/quota.c 	qd->qd_slot = -1;
qd                225 fs/gfs2/quota.c 	INIT_LIST_HEAD(&qd->qd_lru);
qd                226 fs/gfs2/quota.c 	qd->qd_hash = hash;
qd                228 fs/gfs2/quota.c 	error = gfs2_glock_get(sdp, qd2index(qd),
qd                229 fs/gfs2/quota.c 			      &gfs2_quota_glops, CREATE, &qd->qd_gl);
qd                233 fs/gfs2/quota.c 	return qd;
qd                236 fs/gfs2/quota.c 	kmem_cache_free(gfs2_quotad_cachep, qd);
qd                244 fs/gfs2/quota.c 	struct gfs2_quota_data *qd;
qd                247 fs/gfs2/quota.c 	hlist_bl_for_each_entry_rcu(qd, h, &qd_hash_table[hash], qd_hlist) {
qd                248 fs/gfs2/quota.c 		if (!qid_eq(qd->qd_id, qid))
qd                250 fs/gfs2/quota.c 		if (qd->qd_sbd != sdp)
qd                252 fs/gfs2/quota.c 		if (lockref_get_not_dead(&qd->qd_lockref)) {
qd                253 fs/gfs2/quota.c 			list_lru_del(&gfs2_qd_lru, &qd->qd_lru);
qd                254 fs/gfs2/quota.c 			return qd;
qd                265 fs/gfs2/quota.c 	struct gfs2_quota_data *qd, *new_qd;
qd                269 fs/gfs2/quota.c 	*qdp = qd = gfs2_qd_search_bucket(hash, sdp, qid);
qd                272 fs/gfs2/quota.c 	if (qd)
qd                281 fs/gfs2/quota.c 	*qdp = qd = gfs2_qd_search_bucket(hash, sdp, qid);
qd                282 fs/gfs2/quota.c 	if (qd == NULL) {
qd                291 fs/gfs2/quota.c 	if (qd) {
qd                300 fs/gfs2/quota.c static void qd_hold(struct gfs2_quota_data *qd)
qd                302 fs/gfs2/quota.c 	struct gfs2_sbd *sdp = qd->qd_gl->gl_name.ln_sbd;
qd                303 fs/gfs2/quota.c 	gfs2_assert(sdp, !__lockref_is_dead(&qd->qd_lockref));
qd                304 fs/gfs2/quota.c 	lockref_get(&qd->qd_lockref);
qd                307 fs/gfs2/quota.c static void qd_put(struct gfs2_quota_data *qd)
qd                309 fs/gfs2/quota.c 	if (lockref_put_or_lock(&qd->qd_lockref))
qd                312 fs/gfs2/quota.c 	qd->qd_lockref.count = 0;
qd                313 fs/gfs2/quota.c 	list_lru_add(&gfs2_qd_lru, &qd->qd_lru);
qd                314 fs/gfs2/quota.c 	spin_unlock(&qd->qd_lockref.lock);
qd                318 fs/gfs2/quota.c static int slot_get(struct gfs2_quota_data *qd)
qd                320 fs/gfs2/quota.c 	struct gfs2_sbd *sdp = qd->qd_sbd;
qd                325 fs/gfs2/quota.c 	if (qd->qd_slot_count != 0)
qd                332 fs/gfs2/quota.c 		qd->qd_slot = bit;
qd                335 fs/gfs2/quota.c 		qd->qd_slot_count++;
qd                342 fs/gfs2/quota.c static void slot_hold(struct gfs2_quota_data *qd)
qd                344 fs/gfs2/quota.c 	struct gfs2_sbd *sdp = qd->qd_sbd;
qd                347 fs/gfs2/quota.c 	gfs2_assert(sdp, qd->qd_slot_count);
qd                348 fs/gfs2/quota.c 	qd->qd_slot_count++;
qd                352 fs/gfs2/quota.c static void slot_put(struct gfs2_quota_data *qd)
qd                354 fs/gfs2/quota.c 	struct gfs2_sbd *sdp = qd->qd_sbd;
qd                357 fs/gfs2/quota.c 	gfs2_assert(sdp, qd->qd_slot_count);
qd                358 fs/gfs2/quota.c 	if (!--qd->qd_slot_count) {
qd                359 fs/gfs2/quota.c 		BUG_ON(!test_and_clear_bit(qd->qd_slot, sdp->sd_quota_bitmap));
qd                360 fs/gfs2/quota.c 		qd->qd_slot = -1;
qd                365 fs/gfs2/quota.c static int bh_get(struct gfs2_quota_data *qd)
qd                367 fs/gfs2/quota.c 	struct gfs2_sbd *sdp = qd->qd_gl->gl_name.ln_sbd;
qd                376 fs/gfs2/quota.c 	if (qd->qd_bh_count++) {
qd                381 fs/gfs2/quota.c 	block = qd->qd_slot / sdp->sd_qc_per_block;
qd                382 fs/gfs2/quota.c 	offset = qd->qd_slot % sdp->sd_qc_per_block;
qd                395 fs/gfs2/quota.c 	qd->qd_bh = bh;
qd                396 fs/gfs2/quota.c 	qd->qd_bh_qc = (struct gfs2_quota_change *)
qd                407 fs/gfs2/quota.c 	qd->qd_bh_count--;
qd                412 fs/gfs2/quota.c static void bh_put(struct gfs2_quota_data *qd)
qd                414 fs/gfs2/quota.c 	struct gfs2_sbd *sdp = qd->qd_gl->gl_name.ln_sbd;
qd                417 fs/gfs2/quota.c 	gfs2_assert(sdp, qd->qd_bh_count);
qd                418 fs/gfs2/quota.c 	if (!--qd->qd_bh_count) {
qd                419 fs/gfs2/quota.c 		brelse(qd->qd_bh);
qd                420 fs/gfs2/quota.c 		qd->qd_bh = NULL;
qd                421 fs/gfs2/quota.c 		qd->qd_bh_qc = NULL;
qd                426 fs/gfs2/quota.c static int qd_check_sync(struct gfs2_sbd *sdp, struct gfs2_quota_data *qd,
qd                429 fs/gfs2/quota.c 	if (test_bit(QDF_LOCKED, &qd->qd_flags) ||
qd                430 fs/gfs2/quota.c 	    !test_bit(QDF_CHANGE, &qd->qd_flags) ||
qd                431 fs/gfs2/quota.c 	    (sync_gen && (qd->qd_sync_gen >= *sync_gen)))
qd                434 fs/gfs2/quota.c 	if (!lockref_get_not_dead(&qd->qd_lockref))
qd                437 fs/gfs2/quota.c 	list_move_tail(&qd->qd_list, &sdp->sd_quota_list);
qd                438 fs/gfs2/quota.c 	set_bit(QDF_LOCKED, &qd->qd_flags);
qd                439 fs/gfs2/quota.c 	qd->qd_change_sync = qd->qd_change;
qd                440 fs/gfs2/quota.c 	slot_hold(qd);
qd                446 fs/gfs2/quota.c 	struct gfs2_quota_data *qd = NULL;
qd                457 fs/gfs2/quota.c 	list_for_each_entry(qd, &sdp->sd_quota_list, qd_list) {
qd                458 fs/gfs2/quota.c 		found = qd_check_sync(sdp, qd, &sdp->sd_quota_sync_gen);
qd                464 fs/gfs2/quota.c 		qd = NULL;
qd                468 fs/gfs2/quota.c 	if (qd) {
qd                469 fs/gfs2/quota.c 		gfs2_assert_warn(sdp, qd->qd_change_sync);
qd                470 fs/gfs2/quota.c 		error = bh_get(qd);
qd                472 fs/gfs2/quota.c 			clear_bit(QDF_LOCKED, &qd->qd_flags);
qd                473 fs/gfs2/quota.c 			slot_put(qd);
qd                474 fs/gfs2/quota.c 			qd_put(qd);
qd                479 fs/gfs2/quota.c 	*qdp = qd;
qd                484 fs/gfs2/quota.c static void qd_unlock(struct gfs2_quota_data *qd)
qd                486 fs/gfs2/quota.c 	gfs2_assert_warn(qd->qd_gl->gl_name.ln_sbd,
qd                487 fs/gfs2/quota.c 			 test_bit(QDF_LOCKED, &qd->qd_flags));
qd                488 fs/gfs2/quota.c 	clear_bit(QDF_LOCKED, &qd->qd_flags);
qd                489 fs/gfs2/quota.c 	bh_put(qd);
qd                490 fs/gfs2/quota.c 	slot_put(qd);
qd                491 fs/gfs2/quota.c 	qd_put(qd);
qd                520 fs/gfs2/quota.c static void qdsb_put(struct gfs2_quota_data *qd)
qd                522 fs/gfs2/quota.c 	bh_put(qd);
qd                523 fs/gfs2/quota.c 	slot_put(qd);
qd                524 fs/gfs2/quota.c 	qd_put(qd);
qd                563 fs/gfs2/quota.c 	struct gfs2_quota_data **qd;
qd                575 fs/gfs2/quota.c 	qd = ip->i_qadata->qa_qd;
qd                581 fs/gfs2/quota.c 	error = qdsb_get(sdp, make_kqid_uid(ip->i_inode.i_uid), qd);
qd                585 fs/gfs2/quota.c 	qd++;
qd                587 fs/gfs2/quota.c 	error = qdsb_get(sdp, make_kqid_gid(ip->i_inode.i_gid), qd);
qd                591 fs/gfs2/quota.c 	qd++;
qd                595 fs/gfs2/quota.c 		error = qdsb_get(sdp, make_kqid_uid(uid), qd);
qd                599 fs/gfs2/quota.c 		qd++;
qd                604 fs/gfs2/quota.c 		error = qdsb_get(sdp, make_kqid_gid(gid), qd);
qd                608 fs/gfs2/quota.c 		qd++;
qd                645 fs/gfs2/quota.c static void do_qc(struct gfs2_quota_data *qd, s64 change)
qd                647 fs/gfs2/quota.c 	struct gfs2_sbd *sdp = qd->qd_gl->gl_name.ln_sbd;
qd                649 fs/gfs2/quota.c 	struct gfs2_quota_change *qc = qd->qd_bh_qc;
qd                653 fs/gfs2/quota.c 	gfs2_trans_add_meta(ip->i_gl, qd->qd_bh);
qd                655 fs/gfs2/quota.c 	if (!test_bit(QDF_CHANGE, &qd->qd_flags)) {
qd                658 fs/gfs2/quota.c 		if (qd->qd_id.type == USRQUOTA)
qd                660 fs/gfs2/quota.c 		qc->qc_id = cpu_to_be32(from_kqid(&init_user_ns, qd->qd_id));
qd                667 fs/gfs2/quota.c 	qd->qd_change = x;
qd                671 fs/gfs2/quota.c 		gfs2_assert_warn(sdp, test_bit(QDF_CHANGE, &qd->qd_flags));
qd                672 fs/gfs2/quota.c 		clear_bit(QDF_CHANGE, &qd->qd_flags);
qd                675 fs/gfs2/quota.c 		slot_put(qd);
qd                676 fs/gfs2/quota.c 		qd_put(qd);
qd                677 fs/gfs2/quota.c 	} else if (!test_and_set_bit(QDF_CHANGE, &qd->qd_flags)) {
qd                678 fs/gfs2/quota.c 		qd_hold(qd);
qd                679 fs/gfs2/quota.c 		slot_hold(qd);
qd                683 fs/gfs2/quota.c 		clear_bit(QDF_QMSG_QUIET, &qd->qd_flags);
qd                811 fs/gfs2/quota.c 			     s64 change, struct gfs2_quota_data *qd,
qd                836 fs/gfs2/quota.c 	qd->qd_qb.qb_value = q.qu_value;
qd                840 fs/gfs2/quota.c 			qd->qd_qb.qb_warn = q.qu_warn;
qd                844 fs/gfs2/quota.c 			qd->qd_qb.qb_limit = q.qu_limit;
qd                848 fs/gfs2/quota.c 			qd->qd_qb.qb_value = q.qu_value;
qd                859 fs/gfs2/quota.c 		set_bit(QDF_REFRESH, &qd->qd_flags);
qd                873 fs/gfs2/quota.c 	struct gfs2_quota_data *qd;
qd                935 fs/gfs2/quota.c 		qd = qda[x];
qd                936 fs/gfs2/quota.c 		offset = qd2offset(qd);
qd                937 fs/gfs2/quota.c 		error = gfs2_adjust_quota(ip, offset, qd->qd_change_sync, qd, NULL);
qd                941 fs/gfs2/quota.c 		do_qc(qd, -qd->qd_change_sync);
qd                942 fs/gfs2/quota.c 		set_bit(QDF_REFRESH, &qd->qd_flags);
qd                963 fs/gfs2/quota.c static int update_qd(struct gfs2_sbd *sdp, struct gfs2_quota_data *qd)
qd                972 fs/gfs2/quota.c 	pos = qd2offset(qd);
qd                977 fs/gfs2/quota.c 	qlvb = (struct gfs2_quota_lvb *)qd->qd_gl->gl_lksb.sb_lvbptr;
qd                983 fs/gfs2/quota.c 	qd->qd_qb = *qlvb;
qd                988 fs/gfs2/quota.c static int do_glock(struct gfs2_quota_data *qd, int force_refresh,
qd                991 fs/gfs2/quota.c 	struct gfs2_sbd *sdp = qd->qd_gl->gl_name.ln_sbd;
qd                997 fs/gfs2/quota.c 	error = gfs2_glock_nq_init(qd->qd_gl, LM_ST_SHARED, 0, q_gh);
qd               1001 fs/gfs2/quota.c 	if (test_and_clear_bit(QDF_REFRESH, &qd->qd_flags))
qd               1004 fs/gfs2/quota.c 	qd->qd_qb = *(struct gfs2_quota_lvb *)qd->qd_gl->gl_lksb.sb_lvbptr;
qd               1006 fs/gfs2/quota.c 	if (force_refresh || qd->qd_qb.qb_magic != cpu_to_be32(GFS2_MAGIC)) {
qd               1008 fs/gfs2/quota.c 		error = gfs2_glock_nq_init(qd->qd_gl, LM_ST_EXCLUSIVE,
qd               1017 fs/gfs2/quota.c 		error = update_qd(sdp, qd);
qd               1039 fs/gfs2/quota.c 	struct gfs2_quota_data *qd;
qd               1054 fs/gfs2/quota.c 		qd = ip->i_qadata->qa_qd[x];
qd               1055 fs/gfs2/quota.c 		error = do_glock(qd, NO_FORCE, &ip->i_qadata->qa_qd_ghs[x]);
qd               1071 fs/gfs2/quota.c static int need_sync(struct gfs2_quota_data *qd)
qd               1073 fs/gfs2/quota.c 	struct gfs2_sbd *sdp = qd->qd_gl->gl_name.ln_sbd;
qd               1079 fs/gfs2/quota.c 	if (!qd->qd_qb.qb_limit)
qd               1083 fs/gfs2/quota.c 	value = qd->qd_change;
qd               1093 fs/gfs2/quota.c 	else if ((s64)be64_to_cpu(qd->qd_qb.qb_value) >=
qd               1094 fs/gfs2/quota.c 		 (s64)be64_to_cpu(qd->qd_qb.qb_limit))
qd               1099 fs/gfs2/quota.c 		value += (s64)be64_to_cpu(qd->qd_qb.qb_value);
qd               1100 fs/gfs2/quota.c 		if (value < (s64)be64_to_cpu(qd->qd_qb.qb_limit))
qd               1119 fs/gfs2/quota.c 		struct gfs2_quota_data *qd;
qd               1122 fs/gfs2/quota.c 		qd = ip->i_qadata->qa_qd[x];
qd               1123 fs/gfs2/quota.c 		sync = need_sync(qd);
qd               1130 fs/gfs2/quota.c 		found = qd_check_sync(sdp, qd, NULL);
qd               1136 fs/gfs2/quota.c 		gfs2_assert_warn(sdp, qd->qd_change_sync);
qd               1137 fs/gfs2/quota.c 		if (bh_get(qd)) {
qd               1138 fs/gfs2/quota.c 			clear_bit(QDF_LOCKED, &qd->qd_flags);
qd               1139 fs/gfs2/quota.c 			slot_put(qd);
qd               1140 fs/gfs2/quota.c 			qd_put(qd);
qd               1144 fs/gfs2/quota.c 		qda[count++] = qd;
qd               1159 fs/gfs2/quota.c static int print_message(struct gfs2_quota_data *qd, char *type)
qd               1161 fs/gfs2/quota.c 	struct gfs2_sbd *sdp = qd->qd_gl->gl_name.ln_sbd;
qd               1165 fs/gfs2/quota.c 		(qd->qd_id.type == USRQUOTA) ? "user" : "group",
qd               1166 fs/gfs2/quota.c 		from_kqid(&init_user_ns, qd->qd_id));
qd               1192 fs/gfs2/quota.c 	struct gfs2_quota_data *qd;
qd               1205 fs/gfs2/quota.c 		qd = ip->i_qadata->qa_qd[x];
qd               1207 fs/gfs2/quota.c 		if (!(qid_eq(qd->qd_id, make_kqid_uid(uid)) ||
qd               1208 fs/gfs2/quota.c 		      qid_eq(qd->qd_id, make_kqid_gid(gid))))
qd               1211 fs/gfs2/quota.c 		warn = (s64)be64_to_cpu(qd->qd_qb.qb_warn);
qd               1212 fs/gfs2/quota.c 		limit = (s64)be64_to_cpu(qd->qd_qb.qb_limit);
qd               1213 fs/gfs2/quota.c 		value = (s64)be64_to_cpu(qd->qd_qb.qb_value);
qd               1215 fs/gfs2/quota.c 		value += qd->qd_change;
qd               1226 fs/gfs2/quota.c 						      &qd->qd_flags)) {
qd               1227 fs/gfs2/quota.c 					print_message(qd, "exceeded");
qd               1228 fs/gfs2/quota.c 					quota_send_warning(qd->qd_id,
qd               1236 fs/gfs2/quota.c 			   time_after_eq(jiffies, qd->qd_last_warn +
qd               1239 fs/gfs2/quota.c 			quota_send_warning(qd->qd_id,
qd               1241 fs/gfs2/quota.c 			error = print_message(qd, "warning");
qd               1242 fs/gfs2/quota.c 			qd->qd_last_warn = jiffies;
qd               1251 fs/gfs2/quota.c 	struct gfs2_quota_data *qd;
qd               1262 fs/gfs2/quota.c 		qd = ip->i_qadata->qa_qd[x];
qd               1264 fs/gfs2/quota.c 		if (qid_eq(qd->qd_id, make_kqid_uid(uid)) ||
qd               1265 fs/gfs2/quota.c 		    qid_eq(qd->qd_id, make_kqid_gid(gid))) {
qd               1266 fs/gfs2/quota.c 			do_qc(qd, change);
qd               1319 fs/gfs2/quota.c 	struct gfs2_quota_data *qd;
qd               1323 fs/gfs2/quota.c 	error = qd_get(sdp, qid, &qd);
qd               1327 fs/gfs2/quota.c 	error = do_glock(qd, FORCE, &q_gh);
qd               1331 fs/gfs2/quota.c 	qd_put(qd);
qd               1385 fs/gfs2/quota.c 			struct gfs2_quota_data *qd;
qd               1397 fs/gfs2/quota.c 			qd = qd_alloc(hash, sdp, qc_id);
qd               1398 fs/gfs2/quota.c 			if (qd == NULL) {
qd               1403 fs/gfs2/quota.c 			set_bit(QDF_CHANGE, &qd->qd_flags);
qd               1404 fs/gfs2/quota.c 			qd->qd_change = qc_change;
qd               1405 fs/gfs2/quota.c 			qd->qd_slot = slot;
qd               1406 fs/gfs2/quota.c 			qd->qd_slot_count = 1;
qd               1410 fs/gfs2/quota.c 			list_add(&qd->qd_list, &sdp->sd_quota_list);
qd               1415 fs/gfs2/quota.c 			hlist_bl_add_head_rcu(&qd->qd_hlist, &qd_hash_table[hash]);
qd               1439 fs/gfs2/quota.c 	struct gfs2_quota_data *qd;
qd               1443 fs/gfs2/quota.c 		qd = list_entry(head->prev, struct gfs2_quota_data, qd_list);
qd               1445 fs/gfs2/quota.c 		list_del(&qd->qd_list);
qd               1448 fs/gfs2/quota.c 		list_lru_del(&gfs2_qd_lru, &qd->qd_lru);
qd               1452 fs/gfs2/quota.c 		spin_lock_bucket(qd->qd_hash);
qd               1453 fs/gfs2/quota.c 		hlist_bl_del_rcu(&qd->qd_hlist);
qd               1454 fs/gfs2/quota.c 		spin_unlock_bucket(qd->qd_hash);
qd               1456 fs/gfs2/quota.c 		gfs2_assert_warn(sdp, !qd->qd_change);
qd               1457 fs/gfs2/quota.c 		gfs2_assert_warn(sdp, !qd->qd_slot_count);
qd               1458 fs/gfs2/quota.c 		gfs2_assert_warn(sdp, !qd->qd_bh_count);
qd               1460 fs/gfs2/quota.c 		gfs2_glock_put(qd->qd_gl);
qd               1461 fs/gfs2/quota.c 		call_rcu(&qd->qd_rcu, gfs2_qd_dealloc);
qd               1615 fs/gfs2/quota.c 	struct gfs2_quota_data *qd;
qd               1628 fs/gfs2/quota.c 	error = qd_get(sdp, qid, &qd);
qd               1631 fs/gfs2/quota.c 	error = do_glock(qd, FORCE, &q_gh);
qd               1635 fs/gfs2/quota.c 	qlvb = (struct gfs2_quota_lvb *)qd->qd_gl->gl_lksb.sb_lvbptr;
qd               1642 fs/gfs2/quota.c 	qd_put(qd);
qd               1654 fs/gfs2/quota.c 	struct gfs2_quota_data *qd;
qd               1672 fs/gfs2/quota.c 	error = qd_get(sdp, qid, &qd);
qd               1681 fs/gfs2/quota.c 	error = gfs2_glock_nq_init(qd->qd_gl, LM_ST_EXCLUSIVE, 0, &q_gh);
qd               1689 fs/gfs2/quota.c 	error = update_qd(sdp, qd);
qd               1695 fs/gfs2/quota.c 	    ((fdq->d_spc_softlimit >> sdp->sd_sb.sb_bsize_shift) == be64_to_cpu(qd->qd_qb.qb_warn)))
qd               1699 fs/gfs2/quota.c 	    ((fdq->d_spc_hardlimit >> sdp->sd_sb.sb_bsize_shift) == be64_to_cpu(qd->qd_qb.qb_limit)))
qd               1703 fs/gfs2/quota.c 	    ((fdq->d_space >> sdp->sd_sb.sb_bsize_shift) == be64_to_cpu(qd->qd_qb.qb_value)))
qd               1709 fs/gfs2/quota.c 	offset = qd2offset(qd);
qd               1732 fs/gfs2/quota.c 	error = gfs2_adjust_quota(ip, offset, 0, qd, fdq);
qd               1734 fs/gfs2/quota.c 		clear_bit(QDF_QMSG_QUIET, &qd->qd_flags);
qd               1747 fs/gfs2/quota.c 	qd_put(qd);
qd                365 include/linux/intel-iommu.h #define QI_DEV_EIOTLB_QDEP(qd)	((u64)((qd) & 0x1f) << 4)