qd 309 block/blk-iolatency.c static inline unsigned long scale_amount(unsigned long qd, bool up) qd 311 block/blk-iolatency.c return max(up ? qd >> SCALE_UP_FACTOR : qd >> SCALE_DOWN_FACTOR, 1UL); qd 327 block/blk-iolatency.c unsigned long qd = blkiolat->rqos.q->nr_requests; qd 328 block/blk-iolatency.c unsigned long scale = scale_amount(qd, up); qd 330 block/blk-iolatency.c unsigned long max_scale = qd << 1; qd 340 block/blk-iolatency.c else if (diff > qd) qd 351 block/blk-iolatency.c if (diff > qd) { qd 367 block/blk-iolatency.c unsigned long qd = iolat->blkiolat->rqos.q->nr_requests; qd 368 block/blk-iolatency.c unsigned long scale = scale_amount(qd, up); qd 371 block/blk-iolatency.c if (old > qd) qd 372 block/blk-iolatency.c old = qd; qd 378 block/blk-iolatency.c if (old < qd) { qd 380 block/blk-iolatency.c old = min(old, qd); qd 623 drivers/atm/firestream.c static int qd; qd 1193 drivers/atm/firestream.c da[qd] = td; qd 1194 drivers/atm/firestream.c dq[qd].flags = td->flags; qd 1195 drivers/atm/firestream.c dq[qd].next = td->next; qd 1196 drivers/atm/firestream.c dq[qd].bsa = td->bsa; qd 1197 drivers/atm/firestream.c dq[qd].skb = td->skb; qd 1198 drivers/atm/firestream.c dq[qd].dev = td->dev; qd 1199 drivers/atm/firestream.c qd++; qd 1200 drivers/atm/firestream.c if (qd >= 60) qd = 0; qd 1948 drivers/atm/firestream.c i, da[qd], dq[qd].flags, dq[qd].bsa, dq[qd].skb, dq[qd].dev); qd 1949 drivers/atm/firestream.c qd++; qd 1950 drivers/atm/firestream.c if (qd >= 60) qd = 0; qd 883 drivers/block/xen-blkfront.c const struct blk_mq_queue_data *qd) qd 892 drivers/block/xen-blkfront.c blk_mq_start_request(qd->rq); qd 897 drivers/block/xen-blkfront.c if (blkif_request_flush_invalid(qd->rq, rinfo->dev_info)) qd 900 drivers/block/xen-blkfront.c if (blkif_queue_request(qd->rq, rinfo)) qd 47 drivers/clk/at91/clk-audio-pll.c #define AUDIO_PLL_QDPAD(qd, div) ((AT91_PMC_AUDIO_PLL_QDPAD_EXTDIV(qd) & \ qd 9622 drivers/infiniband/hw/hfi1/chip.c struct qsfp_data *qd; qd 9626 drivers/infiniband/hw/hfi1/chip.c qd = container_of(work, struct qsfp_data, qsfp_work); qd 9627 drivers/infiniband/hw/hfi1/chip.c ppd = qd->ppd; qd 9647 drivers/infiniband/hw/hfi1/chip.c if (qd->cache_refresh_required) { qd 9661 drivers/infiniband/hw/hfi1/chip.c if (qd->check_interrupt_flags) { qd 2980 drivers/infiniband/hw/qib/qib_iba7322.c struct qib_qsfp_data *qd; qd 2991 drivers/infiniband/hw/qib/qib_iba7322.c qd = &ppd->cpspec->qsfp_data; qd 2997 drivers/infiniband/hw/qib/qib_iba7322.c qd->t_insert = jiffies; qd 2998 drivers/infiniband/hw/qib/qib_iba7322.c queue_work(ib_wq, &qd->work); qd 5574 drivers/infiniband/hw/qib/qib_iba7322.c struct qib_qsfp_data *qd = qd 5586 drivers/infiniband/hw/qib/qib_iba7322.c qd->t_insert = jiffies; qd 5587 drivers/infiniband/hw/qib/qib_iba7322.c queue_work(ib_wq, &qd->work); qd 5937 drivers/infiniband/hw/qib/qib_iba7322.c struct qib_qsfp_data *qd; qd 5944 drivers/infiniband/hw/qib/qib_iba7322.c qd = container_of(work, struct qib_qsfp_data, work); qd 5945 drivers/infiniband/hw/qib/qib_iba7322.c ppd = qd->ppd; qd 5946 drivers/infiniband/hw/qib/qib_iba7322.c pwrup = qd->t_insert + qd 5972 drivers/infiniband/hw/qib/qib_iba7322.c ret = qib_refresh_qsfp_cache(ppd, &qd->cache); qd 5981 drivers/infiniband/hw/qib/qib_iba7322.c if (QSFP_IS_ACTIVE_FAR(qd->cache.tech)) qd 5983 drivers/infiniband/hw/qib/qib_iba7322.c else if (qd->cache.atten[1] >= qib_long_atten && qd 5984 drivers/infiniband/hw/qib/qib_iba7322.c QSFP_IS_CU(qd->cache.tech)) qd 6021 drivers/infiniband/hw/qib/qib_iba7322.c struct qib_qsfp_data *qd = &ppd->cpspec->qsfp_data; qd 6026 drivers/infiniband/hw/qib/qib_iba7322.c qd->ppd = ppd; qd 6027 drivers/infiniband/hw/qib/qib_iba7322.c qib_qsfp_init(qd, qsfp_7322_event); qd 7621 drivers/infiniband/hw/qib/qib_iba7322.c struct qib_qsfp_cache *qd = &ppd->cpspec->qsfp_data.cache; qd 7628 drivers/infiniband/hw/qib/qib_iba7322.c if (!memcmp(v->oui, qd->oui, QSFP_VOUI_LEN) && qd 7630 drivers/infiniband/hw/qib/qib_iba7322.c !memcmp(v->partnum, qd->partnum, QSFP_PN_LEN))) { qd 7640 drivers/infiniband/hw/qib/qib_iba7322.c if (!override && QSFP_IS_ACTIVE(qd->tech)) { qd 7647 drivers/infiniband/hw/qib/qib_iba7322.c if (!override && QSFP_HAS_ATTEN(qd->tech) && (qd->atten[0] || qd 7648 drivers/infiniband/hw/qib/qib_iba7322.c qd->atten[1])) { qd 7649 drivers/infiniband/hw/qib/qib_iba7322.c *sdr_dds = get_atten_table(txdds_sdr, qd->atten[0]); qd 7650 drivers/infiniband/hw/qib/qib_iba7322.c *ddr_dds = get_atten_table(txdds_ddr, qd->atten[0]); qd 7651 drivers/infiniband/hw/qib/qib_iba7322.c *qdr_dds = get_atten_table(txdds_qdr, qd->atten[1]); qd 461 drivers/infiniband/hw/qib/qib_qsfp.c void qib_qsfp_init(struct qib_qsfp_data *qd, qd 466 drivers/infiniband/hw/qib/qib_qsfp.c struct qib_devdata *dd = qd->ppd->dd; qd 469 drivers/infiniband/hw/qib/qib_qsfp.c INIT_WORK(&qd->work, fevent); qd 478 drivers/infiniband/hw/qib/qib_qsfp.c if (qd->ppd->hw_pidx) { qd 187 drivers/infiniband/hw/qib/qib_qsfp.h extern void qib_qsfp_init(struct qib_qsfp_data *qd, qd 585 drivers/mailbox/ti-msgmgr.c const struct ti_msgmgr_valid_queue_desc *qd, qd 591 drivers/mailbox/ti-msgmgr.c qinst->proxy_id = qd->proxy_id; qd 592 drivers/mailbox/ti-msgmgr.c qinst->queue_id = qd->queue_id; qd 626 drivers/mailbox/ti-msgmgr.c qinst->is_tx = qd->is_tx; qd 3073 drivers/s390/block/dasd.c const struct blk_mq_queue_data *qd) qd 3077 drivers/s390/block/dasd.c struct request *req = qd->rq; qd 283 drivers/s390/block/scm_blk.c const struct blk_mq_queue_data *qd) qd 288 drivers/s390/block/scm_blk.c struct request *req = qd->rq; qd 323 drivers/s390/block/scm_blk.c if (qd->last || scmrq->aob->request.msb_count == nr_requests_per_io) { qd 334 drivers/s390/block/scm_blk.c struct scm_queue *qd = kzalloc(sizeof(*qd), GFP_KERNEL); qd 336 drivers/s390/block/scm_blk.c if (!qd) qd 339 drivers/s390/block/scm_blk.c spin_lock_init(&qd->lock); qd 340 drivers/s390/block/scm_blk.c hctx->driver_data = qd; qd 347 drivers/s390/block/scm_blk.c struct scm_queue *qd = hctx->driver_data; qd 349 drivers/s390/block/scm_blk.c WARN_ON(qd->scmrq); qd 3351 drivers/scsi/aacraid/aachba.c struct aac_query_disk qd; qd 3357 drivers/scsi/aacraid/aachba.c if (copy_from_user(&qd, arg, sizeof (struct aac_query_disk))) qd 3359 drivers/scsi/aacraid/aachba.c if (qd.cnum == -1) { qd 3360 drivers/scsi/aacraid/aachba.c if (qd.id < 0 || qd.id >= dev->maximum_num_containers) qd 3362 drivers/scsi/aacraid/aachba.c qd.cnum = qd.id; qd 3363 drivers/scsi/aacraid/aachba.c } else if ((qd.bus == -1) && (qd.id == -1) && (qd.lun == -1)) { qd 3364 drivers/scsi/aacraid/aachba.c if (qd.cnum < 0 || qd.cnum >= dev->maximum_num_containers) qd 3366 drivers/scsi/aacraid/aachba.c qd.instance = dev->scsi_host_ptr->host_no; qd 3367 drivers/scsi/aacraid/aachba.c qd.bus = 0; qd 3368 drivers/scsi/aacraid/aachba.c qd.id = CONTAINER_TO_ID(qd.cnum); qd 3369 drivers/scsi/aacraid/aachba.c qd.lun = CONTAINER_TO_LUN(qd.cnum); qd 3373 drivers/scsi/aacraid/aachba.c qd.valid = fsa_dev_ptr[qd.cnum].valid != 0; qd 3374 drivers/scsi/aacraid/aachba.c qd.locked = fsa_dev_ptr[qd.cnum].locked; qd 3375 drivers/scsi/aacraid/aachba.c qd.deleted = fsa_dev_ptr[qd.cnum].deleted; qd 3377 drivers/scsi/aacraid/aachba.c if (fsa_dev_ptr[qd.cnum].devname[0] == '\0') qd 3378 drivers/scsi/aacraid/aachba.c qd.unmapped = 1; qd 3380 drivers/scsi/aacraid/aachba.c qd.unmapped = 0; qd 3382 drivers/scsi/aacraid/aachba.c strlcpy(qd.name, fsa_dev_ptr[qd.cnum].devname, qd 3383 drivers/scsi/aacraid/aachba.c min(sizeof(qd.name), sizeof(fsa_dev_ptr[qd.cnum].devname) + 1)); qd 3385 drivers/scsi/aacraid/aachba.c if (copy_to_user(arg, &qd, sizeof (struct aac_query_disk))) qd 18 fs/erofs/namei.c const struct erofs_qstr *qd, qd 28 fs/erofs/namei.c DBG_BUGON(qd->name > qd->end); qd 32 fs/erofs/namei.c while (qd->name + i < qd->end && qd->name[i] != '\0') { qd 33 fs/erofs/namei.c if (qn->name[i] != qd->name[i]) { qd 35 fs/erofs/namei.c return qn->name[i] > qd->name[i] ? 1 : -1; qd 108 fs/gfs2/quota.c struct gfs2_quota_data *qd = container_of(rcu, struct gfs2_quota_data, qd_rcu); qd 109 fs/gfs2/quota.c kmem_cache_free(gfs2_quotad_cachep, qd); qd 114 fs/gfs2/quota.c struct gfs2_quota_data *qd; qd 118 fs/gfs2/quota.c qd = list_entry(list->next, struct gfs2_quota_data, qd_lru); qd 119 fs/gfs2/quota.c sdp = qd->qd_gl->gl_name.ln_sbd; qd 121 fs/gfs2/quota.c list_del(&qd->qd_lru); qd 125 fs/gfs2/quota.c list_del(&qd->qd_list); qd 128 fs/gfs2/quota.c spin_lock_bucket(qd->qd_hash); qd 129 fs/gfs2/quota.c hlist_bl_del_rcu(&qd->qd_hlist); qd 130 fs/gfs2/quota.c spin_unlock_bucket(qd->qd_hash); qd 132 fs/gfs2/quota.c gfs2_assert_warn(sdp, !qd->qd_change); qd 133 fs/gfs2/quota.c gfs2_assert_warn(sdp, !qd->qd_slot_count); qd 134 fs/gfs2/quota.c gfs2_assert_warn(sdp, !qd->qd_bh_count); qd 136 fs/gfs2/quota.c gfs2_glock_put(qd->qd_gl); qd 140 fs/gfs2/quota.c call_rcu(&qd->qd_rcu, gfs2_qd_dealloc); qd 149 fs/gfs2/quota.c struct gfs2_quota_data *qd = list_entry(item, struct gfs2_quota_data, qd_lru); qd 151 fs/gfs2/quota.c if (!spin_trylock(&qd->qd_lockref.lock)) qd 154 fs/gfs2/quota.c if (qd->qd_lockref.count == 0) { qd 155 fs/gfs2/quota.c lockref_mark_dead(&qd->qd_lockref); qd 156 fs/gfs2/quota.c list_lru_isolate_move(lru, &qd->qd_lru, dispose); qd 159 fs/gfs2/quota.c spin_unlock(&qd->qd_lockref.lock); qd 194 fs/gfs2/quota.c static u64 qd2index(struct gfs2_quota_data *qd) qd 196 fs/gfs2/quota.c struct kqid qid = qd->qd_id; qd 201 fs/gfs2/quota.c static u64 qd2offset(struct gfs2_quota_data *qd) qd 205 fs/gfs2/quota.c offset = qd2index(qd); qd 213 fs/gfs2/quota.c struct gfs2_quota_data *qd; qd 216 fs/gfs2/quota.c qd = kmem_cache_zalloc(gfs2_quotad_cachep, GFP_NOFS); qd 217 fs/gfs2/quota.c if (!qd) qd 220 fs/gfs2/quota.c qd->qd_sbd = sdp; qd 221 fs/gfs2/quota.c qd->qd_lockref.count = 1; qd 222 fs/gfs2/quota.c spin_lock_init(&qd->qd_lockref.lock); qd 223 fs/gfs2/quota.c qd->qd_id = qid; qd 224 fs/gfs2/quota.c qd->qd_slot = -1; qd 225 fs/gfs2/quota.c INIT_LIST_HEAD(&qd->qd_lru); qd 226 fs/gfs2/quota.c qd->qd_hash = hash; qd 228 fs/gfs2/quota.c error = gfs2_glock_get(sdp, qd2index(qd), qd 229 fs/gfs2/quota.c &gfs2_quota_glops, CREATE, &qd->qd_gl); qd 233 fs/gfs2/quota.c return qd; qd 236 fs/gfs2/quota.c kmem_cache_free(gfs2_quotad_cachep, qd); qd 244 fs/gfs2/quota.c struct gfs2_quota_data *qd; qd 247 fs/gfs2/quota.c hlist_bl_for_each_entry_rcu(qd, h, &qd_hash_table[hash], qd_hlist) { qd 248 fs/gfs2/quota.c if (!qid_eq(qd->qd_id, qid)) qd 250 fs/gfs2/quota.c if (qd->qd_sbd != sdp) qd 252 fs/gfs2/quota.c if (lockref_get_not_dead(&qd->qd_lockref)) { qd 253 fs/gfs2/quota.c list_lru_del(&gfs2_qd_lru, &qd->qd_lru); qd 254 fs/gfs2/quota.c return qd; qd 265 fs/gfs2/quota.c struct gfs2_quota_data *qd, *new_qd; qd 269 fs/gfs2/quota.c *qdp = qd = gfs2_qd_search_bucket(hash, sdp, qid); qd 272 fs/gfs2/quota.c if (qd) qd 281 fs/gfs2/quota.c *qdp = qd = gfs2_qd_search_bucket(hash, sdp, qid); qd 282 fs/gfs2/quota.c if (qd == NULL) { qd 291 fs/gfs2/quota.c if (qd) { qd 300 fs/gfs2/quota.c static void qd_hold(struct gfs2_quota_data *qd) qd 302 fs/gfs2/quota.c struct gfs2_sbd *sdp = qd->qd_gl->gl_name.ln_sbd; qd 303 fs/gfs2/quota.c gfs2_assert(sdp, !__lockref_is_dead(&qd->qd_lockref)); qd 304 fs/gfs2/quota.c lockref_get(&qd->qd_lockref); qd 307 fs/gfs2/quota.c static void qd_put(struct gfs2_quota_data *qd) qd 309 fs/gfs2/quota.c if (lockref_put_or_lock(&qd->qd_lockref)) qd 312 fs/gfs2/quota.c qd->qd_lockref.count = 0; qd 313 fs/gfs2/quota.c list_lru_add(&gfs2_qd_lru, &qd->qd_lru); qd 314 fs/gfs2/quota.c spin_unlock(&qd->qd_lockref.lock); qd 318 fs/gfs2/quota.c static int slot_get(struct gfs2_quota_data *qd) qd 320 fs/gfs2/quota.c struct gfs2_sbd *sdp = qd->qd_sbd; qd 325 fs/gfs2/quota.c if (qd->qd_slot_count != 0) qd 332 fs/gfs2/quota.c qd->qd_slot = bit; qd 335 fs/gfs2/quota.c qd->qd_slot_count++; qd 342 fs/gfs2/quota.c static void slot_hold(struct gfs2_quota_data *qd) qd 344 fs/gfs2/quota.c struct gfs2_sbd *sdp = qd->qd_sbd; qd 347 fs/gfs2/quota.c gfs2_assert(sdp, qd->qd_slot_count); qd 348 fs/gfs2/quota.c qd->qd_slot_count++; qd 352 fs/gfs2/quota.c static void slot_put(struct gfs2_quota_data *qd) qd 354 fs/gfs2/quota.c struct gfs2_sbd *sdp = qd->qd_sbd; qd 357 fs/gfs2/quota.c gfs2_assert(sdp, qd->qd_slot_count); qd 358 fs/gfs2/quota.c if (!--qd->qd_slot_count) { qd 359 fs/gfs2/quota.c BUG_ON(!test_and_clear_bit(qd->qd_slot, sdp->sd_quota_bitmap)); qd 360 fs/gfs2/quota.c qd->qd_slot = -1; qd 365 fs/gfs2/quota.c static int bh_get(struct gfs2_quota_data *qd) qd 367 fs/gfs2/quota.c struct gfs2_sbd *sdp = qd->qd_gl->gl_name.ln_sbd; qd 376 fs/gfs2/quota.c if (qd->qd_bh_count++) { qd 381 fs/gfs2/quota.c block = qd->qd_slot / sdp->sd_qc_per_block; qd 382 fs/gfs2/quota.c offset = qd->qd_slot % sdp->sd_qc_per_block; qd 395 fs/gfs2/quota.c qd->qd_bh = bh; qd 396 fs/gfs2/quota.c qd->qd_bh_qc = (struct gfs2_quota_change *) qd 407 fs/gfs2/quota.c qd->qd_bh_count--; qd 412 fs/gfs2/quota.c static void bh_put(struct gfs2_quota_data *qd) qd 414 fs/gfs2/quota.c struct gfs2_sbd *sdp = qd->qd_gl->gl_name.ln_sbd; qd 417 fs/gfs2/quota.c gfs2_assert(sdp, qd->qd_bh_count); qd 418 fs/gfs2/quota.c if (!--qd->qd_bh_count) { qd 419 fs/gfs2/quota.c brelse(qd->qd_bh); qd 420 fs/gfs2/quota.c qd->qd_bh = NULL; qd 421 fs/gfs2/quota.c qd->qd_bh_qc = NULL; qd 426 fs/gfs2/quota.c static int qd_check_sync(struct gfs2_sbd *sdp, struct gfs2_quota_data *qd, qd 429 fs/gfs2/quota.c if (test_bit(QDF_LOCKED, &qd->qd_flags) || qd 430 fs/gfs2/quota.c !test_bit(QDF_CHANGE, &qd->qd_flags) || qd 431 fs/gfs2/quota.c (sync_gen && (qd->qd_sync_gen >= *sync_gen))) qd 434 fs/gfs2/quota.c if (!lockref_get_not_dead(&qd->qd_lockref)) qd 437 fs/gfs2/quota.c list_move_tail(&qd->qd_list, &sdp->sd_quota_list); qd 438 fs/gfs2/quota.c set_bit(QDF_LOCKED, &qd->qd_flags); qd 439 fs/gfs2/quota.c qd->qd_change_sync = qd->qd_change; qd 440 fs/gfs2/quota.c slot_hold(qd); qd 446 fs/gfs2/quota.c struct gfs2_quota_data *qd = NULL; qd 457 fs/gfs2/quota.c list_for_each_entry(qd, &sdp->sd_quota_list, qd_list) { qd 458 fs/gfs2/quota.c found = qd_check_sync(sdp, qd, &sdp->sd_quota_sync_gen); qd 464 fs/gfs2/quota.c qd = NULL; qd 468 fs/gfs2/quota.c if (qd) { qd 469 fs/gfs2/quota.c gfs2_assert_warn(sdp, qd->qd_change_sync); qd 470 fs/gfs2/quota.c error = bh_get(qd); qd 472 fs/gfs2/quota.c clear_bit(QDF_LOCKED, &qd->qd_flags); qd 473 fs/gfs2/quota.c slot_put(qd); qd 474 fs/gfs2/quota.c qd_put(qd); qd 479 fs/gfs2/quota.c *qdp = qd; qd 484 fs/gfs2/quota.c static void qd_unlock(struct gfs2_quota_data *qd) qd 486 fs/gfs2/quota.c gfs2_assert_warn(qd->qd_gl->gl_name.ln_sbd, qd 487 fs/gfs2/quota.c test_bit(QDF_LOCKED, &qd->qd_flags)); qd 488 fs/gfs2/quota.c clear_bit(QDF_LOCKED, &qd->qd_flags); qd 489 fs/gfs2/quota.c bh_put(qd); qd 490 fs/gfs2/quota.c slot_put(qd); qd 491 fs/gfs2/quota.c qd_put(qd); qd 520 fs/gfs2/quota.c static void qdsb_put(struct gfs2_quota_data *qd) qd 522 fs/gfs2/quota.c bh_put(qd); qd 523 fs/gfs2/quota.c slot_put(qd); qd 524 fs/gfs2/quota.c qd_put(qd); qd 563 fs/gfs2/quota.c struct gfs2_quota_data **qd; qd 575 fs/gfs2/quota.c qd = ip->i_qadata->qa_qd; qd 581 fs/gfs2/quota.c error = qdsb_get(sdp, make_kqid_uid(ip->i_inode.i_uid), qd); qd 585 fs/gfs2/quota.c qd++; qd 587 fs/gfs2/quota.c error = qdsb_get(sdp, make_kqid_gid(ip->i_inode.i_gid), qd); qd 591 fs/gfs2/quota.c qd++; qd 595 fs/gfs2/quota.c error = qdsb_get(sdp, make_kqid_uid(uid), qd); qd 599 fs/gfs2/quota.c qd++; qd 604 fs/gfs2/quota.c error = qdsb_get(sdp, make_kqid_gid(gid), qd); qd 608 fs/gfs2/quota.c qd++; qd 645 fs/gfs2/quota.c static void do_qc(struct gfs2_quota_data *qd, s64 change) qd 647 fs/gfs2/quota.c struct gfs2_sbd *sdp = qd->qd_gl->gl_name.ln_sbd; qd 649 fs/gfs2/quota.c struct gfs2_quota_change *qc = qd->qd_bh_qc; qd 653 fs/gfs2/quota.c gfs2_trans_add_meta(ip->i_gl, qd->qd_bh); qd 655 fs/gfs2/quota.c if (!test_bit(QDF_CHANGE, &qd->qd_flags)) { qd 658 fs/gfs2/quota.c if (qd->qd_id.type == USRQUOTA) qd 660 fs/gfs2/quota.c qc->qc_id = cpu_to_be32(from_kqid(&init_user_ns, qd->qd_id)); qd 667 fs/gfs2/quota.c qd->qd_change = x; qd 671 fs/gfs2/quota.c gfs2_assert_warn(sdp, test_bit(QDF_CHANGE, &qd->qd_flags)); qd 672 fs/gfs2/quota.c clear_bit(QDF_CHANGE, &qd->qd_flags); qd 675 fs/gfs2/quota.c slot_put(qd); qd 676 fs/gfs2/quota.c qd_put(qd); qd 677 fs/gfs2/quota.c } else if (!test_and_set_bit(QDF_CHANGE, &qd->qd_flags)) { qd 678 fs/gfs2/quota.c qd_hold(qd); qd 679 fs/gfs2/quota.c slot_hold(qd); qd 683 fs/gfs2/quota.c clear_bit(QDF_QMSG_QUIET, &qd->qd_flags); qd 811 fs/gfs2/quota.c s64 change, struct gfs2_quota_data *qd, qd 836 fs/gfs2/quota.c qd->qd_qb.qb_value = q.qu_value; qd 840 fs/gfs2/quota.c qd->qd_qb.qb_warn = q.qu_warn; qd 844 fs/gfs2/quota.c qd->qd_qb.qb_limit = q.qu_limit; qd 848 fs/gfs2/quota.c qd->qd_qb.qb_value = q.qu_value; qd 859 fs/gfs2/quota.c set_bit(QDF_REFRESH, &qd->qd_flags); qd 873 fs/gfs2/quota.c struct gfs2_quota_data *qd; qd 935 fs/gfs2/quota.c qd = qda[x]; qd 936 fs/gfs2/quota.c offset = qd2offset(qd); qd 937 fs/gfs2/quota.c error = gfs2_adjust_quota(ip, offset, qd->qd_change_sync, qd, NULL); qd 941 fs/gfs2/quota.c do_qc(qd, -qd->qd_change_sync); qd 942 fs/gfs2/quota.c set_bit(QDF_REFRESH, &qd->qd_flags); qd 963 fs/gfs2/quota.c static int update_qd(struct gfs2_sbd *sdp, struct gfs2_quota_data *qd) qd 972 fs/gfs2/quota.c pos = qd2offset(qd); qd 977 fs/gfs2/quota.c qlvb = (struct gfs2_quota_lvb *)qd->qd_gl->gl_lksb.sb_lvbptr; qd 983 fs/gfs2/quota.c qd->qd_qb = *qlvb; qd 988 fs/gfs2/quota.c static int do_glock(struct gfs2_quota_data *qd, int force_refresh, qd 991 fs/gfs2/quota.c struct gfs2_sbd *sdp = qd->qd_gl->gl_name.ln_sbd; qd 997 fs/gfs2/quota.c error = gfs2_glock_nq_init(qd->qd_gl, LM_ST_SHARED, 0, q_gh); qd 1001 fs/gfs2/quota.c if (test_and_clear_bit(QDF_REFRESH, &qd->qd_flags)) qd 1004 fs/gfs2/quota.c qd->qd_qb = *(struct gfs2_quota_lvb *)qd->qd_gl->gl_lksb.sb_lvbptr; qd 1006 fs/gfs2/quota.c if (force_refresh || qd->qd_qb.qb_magic != cpu_to_be32(GFS2_MAGIC)) { qd 1008 fs/gfs2/quota.c error = gfs2_glock_nq_init(qd->qd_gl, LM_ST_EXCLUSIVE, qd 1017 fs/gfs2/quota.c error = update_qd(sdp, qd); qd 1039 fs/gfs2/quota.c struct gfs2_quota_data *qd; qd 1054 fs/gfs2/quota.c qd = ip->i_qadata->qa_qd[x]; qd 1055 fs/gfs2/quota.c error = do_glock(qd, NO_FORCE, &ip->i_qadata->qa_qd_ghs[x]); qd 1071 fs/gfs2/quota.c static int need_sync(struct gfs2_quota_data *qd) qd 1073 fs/gfs2/quota.c struct gfs2_sbd *sdp = qd->qd_gl->gl_name.ln_sbd; qd 1079 fs/gfs2/quota.c if (!qd->qd_qb.qb_limit) qd 1083 fs/gfs2/quota.c value = qd->qd_change; qd 1093 fs/gfs2/quota.c else if ((s64)be64_to_cpu(qd->qd_qb.qb_value) >= qd 1094 fs/gfs2/quota.c (s64)be64_to_cpu(qd->qd_qb.qb_limit)) qd 1099 fs/gfs2/quota.c value += (s64)be64_to_cpu(qd->qd_qb.qb_value); qd 1100 fs/gfs2/quota.c if (value < (s64)be64_to_cpu(qd->qd_qb.qb_limit)) qd 1119 fs/gfs2/quota.c struct gfs2_quota_data *qd; qd 1122 fs/gfs2/quota.c qd = ip->i_qadata->qa_qd[x]; qd 1123 fs/gfs2/quota.c sync = need_sync(qd); qd 1130 fs/gfs2/quota.c found = qd_check_sync(sdp, qd, NULL); qd 1136 fs/gfs2/quota.c gfs2_assert_warn(sdp, qd->qd_change_sync); qd 1137 fs/gfs2/quota.c if (bh_get(qd)) { qd 1138 fs/gfs2/quota.c clear_bit(QDF_LOCKED, &qd->qd_flags); qd 1139 fs/gfs2/quota.c slot_put(qd); qd 1140 fs/gfs2/quota.c qd_put(qd); qd 1144 fs/gfs2/quota.c qda[count++] = qd; qd 1159 fs/gfs2/quota.c static int print_message(struct gfs2_quota_data *qd, char *type) qd 1161 fs/gfs2/quota.c struct gfs2_sbd *sdp = qd->qd_gl->gl_name.ln_sbd; qd 1165 fs/gfs2/quota.c (qd->qd_id.type == USRQUOTA) ? "user" : "group", qd 1166 fs/gfs2/quota.c from_kqid(&init_user_ns, qd->qd_id)); qd 1192 fs/gfs2/quota.c struct gfs2_quota_data *qd; qd 1205 fs/gfs2/quota.c qd = ip->i_qadata->qa_qd[x]; qd 1207 fs/gfs2/quota.c if (!(qid_eq(qd->qd_id, make_kqid_uid(uid)) || qd 1208 fs/gfs2/quota.c qid_eq(qd->qd_id, make_kqid_gid(gid)))) qd 1211 fs/gfs2/quota.c warn = (s64)be64_to_cpu(qd->qd_qb.qb_warn); qd 1212 fs/gfs2/quota.c limit = (s64)be64_to_cpu(qd->qd_qb.qb_limit); qd 1213 fs/gfs2/quota.c value = (s64)be64_to_cpu(qd->qd_qb.qb_value); qd 1215 fs/gfs2/quota.c value += qd->qd_change; qd 1226 fs/gfs2/quota.c &qd->qd_flags)) { qd 1227 fs/gfs2/quota.c print_message(qd, "exceeded"); qd 1228 fs/gfs2/quota.c quota_send_warning(qd->qd_id, qd 1236 fs/gfs2/quota.c time_after_eq(jiffies, qd->qd_last_warn + qd 1239 fs/gfs2/quota.c quota_send_warning(qd->qd_id, qd 1241 fs/gfs2/quota.c error = print_message(qd, "warning"); qd 1242 fs/gfs2/quota.c qd->qd_last_warn = jiffies; qd 1251 fs/gfs2/quota.c struct gfs2_quota_data *qd; qd 1262 fs/gfs2/quota.c qd = ip->i_qadata->qa_qd[x]; qd 1264 fs/gfs2/quota.c if (qid_eq(qd->qd_id, make_kqid_uid(uid)) || qd 1265 fs/gfs2/quota.c qid_eq(qd->qd_id, make_kqid_gid(gid))) { qd 1266 fs/gfs2/quota.c do_qc(qd, change); qd 1319 fs/gfs2/quota.c struct gfs2_quota_data *qd; qd 1323 fs/gfs2/quota.c error = qd_get(sdp, qid, &qd); qd 1327 fs/gfs2/quota.c error = do_glock(qd, FORCE, &q_gh); qd 1331 fs/gfs2/quota.c qd_put(qd); qd 1385 fs/gfs2/quota.c struct gfs2_quota_data *qd; qd 1397 fs/gfs2/quota.c qd = qd_alloc(hash, sdp, qc_id); qd 1398 fs/gfs2/quota.c if (qd == NULL) { qd 1403 fs/gfs2/quota.c set_bit(QDF_CHANGE, &qd->qd_flags); qd 1404 fs/gfs2/quota.c qd->qd_change = qc_change; qd 1405 fs/gfs2/quota.c qd->qd_slot = slot; qd 1406 fs/gfs2/quota.c qd->qd_slot_count = 1; qd 1410 fs/gfs2/quota.c list_add(&qd->qd_list, &sdp->sd_quota_list); qd 1415 fs/gfs2/quota.c hlist_bl_add_head_rcu(&qd->qd_hlist, &qd_hash_table[hash]); qd 1439 fs/gfs2/quota.c struct gfs2_quota_data *qd; qd 1443 fs/gfs2/quota.c qd = list_entry(head->prev, struct gfs2_quota_data, qd_list); qd 1445 fs/gfs2/quota.c list_del(&qd->qd_list); qd 1448 fs/gfs2/quota.c list_lru_del(&gfs2_qd_lru, &qd->qd_lru); qd 1452 fs/gfs2/quota.c spin_lock_bucket(qd->qd_hash); qd 1453 fs/gfs2/quota.c hlist_bl_del_rcu(&qd->qd_hlist); qd 1454 fs/gfs2/quota.c spin_unlock_bucket(qd->qd_hash); qd 1456 fs/gfs2/quota.c gfs2_assert_warn(sdp, !qd->qd_change); qd 1457 fs/gfs2/quota.c gfs2_assert_warn(sdp, !qd->qd_slot_count); qd 1458 fs/gfs2/quota.c gfs2_assert_warn(sdp, !qd->qd_bh_count); qd 1460 fs/gfs2/quota.c gfs2_glock_put(qd->qd_gl); qd 1461 fs/gfs2/quota.c call_rcu(&qd->qd_rcu, gfs2_qd_dealloc); qd 1615 fs/gfs2/quota.c struct gfs2_quota_data *qd; qd 1628 fs/gfs2/quota.c error = qd_get(sdp, qid, &qd); qd 1631 fs/gfs2/quota.c error = do_glock(qd, FORCE, &q_gh); qd 1635 fs/gfs2/quota.c qlvb = (struct gfs2_quota_lvb *)qd->qd_gl->gl_lksb.sb_lvbptr; qd 1642 fs/gfs2/quota.c qd_put(qd); qd 1654 fs/gfs2/quota.c struct gfs2_quota_data *qd; qd 1672 fs/gfs2/quota.c error = qd_get(sdp, qid, &qd); qd 1681 fs/gfs2/quota.c error = gfs2_glock_nq_init(qd->qd_gl, LM_ST_EXCLUSIVE, 0, &q_gh); qd 1689 fs/gfs2/quota.c error = update_qd(sdp, qd); qd 1695 fs/gfs2/quota.c ((fdq->d_spc_softlimit >> sdp->sd_sb.sb_bsize_shift) == be64_to_cpu(qd->qd_qb.qb_warn))) qd 1699 fs/gfs2/quota.c ((fdq->d_spc_hardlimit >> sdp->sd_sb.sb_bsize_shift) == be64_to_cpu(qd->qd_qb.qb_limit))) qd 1703 fs/gfs2/quota.c ((fdq->d_space >> sdp->sd_sb.sb_bsize_shift) == be64_to_cpu(qd->qd_qb.qb_value))) qd 1709 fs/gfs2/quota.c offset = qd2offset(qd); qd 1732 fs/gfs2/quota.c error = gfs2_adjust_quota(ip, offset, 0, qd, fdq); qd 1734 fs/gfs2/quota.c clear_bit(QDF_QMSG_QUIET, &qd->qd_flags); qd 1747 fs/gfs2/quota.c qd_put(qd); qd 365 include/linux/intel-iommu.h #define QI_DEV_EIOTLB_QDEP(qd) ((u64)((qd) & 0x1f) << 4)