queued 175 block/bfq-cgroup.c if (blkg_rwstat_total(&stats->queued)) queued 217 block/bfq-cgroup.c blkg_rwstat_total(&stats->queued)); queued 225 block/bfq-cgroup.c blkg_rwstat_add(&bfqg->stats.queued, op, 1); queued 233 block/bfq-cgroup.c blkg_rwstat_add(&bfqg->stats.queued, op, -1); queued 438 block/bfq-cgroup.c blkg_rwstat_exit(&stats->queued); queued 455 block/bfq-cgroup.c blkg_rwstat_init(&stats->queued, gfp) || queued 1290 block/bfq-cgroup.c .private = offsetof(struct bfq_group, stats.queued), queued 1333 block/bfq-cgroup.c .private = offsetof(struct bfq_group, stats.queued), queued 423 block/bfq-iosched.c if (bfqd->queued != 0) { queued 1873 block/bfq-iosched.c bfqq->queued[rq_is_sync(rq)]++; queued 1874 block/bfq-iosched.c bfqd->queued++; queued 2166 block/bfq-iosched.c bfqq->queued[sync]--; queued 2167 block/bfq-iosched.c bfqd->queued--; queued 5356 block/bfq-iosched.c bool small_req = bfqq->queued[rq_is_sync(rq)] == 1 && queued 5571 block/bfq-iosched.c if (bfqd->rq_in_driver + bfqd->queued <= BFQ_HW_QUEUE_THRESHOLD) queued 5580 block/bfq-iosched.c bfqq->dispatched + bfqq->queued[0] + bfqq->queued[1] < queued 6240 block/bfq-iosched.c else if (bfqq->queued[0] == 0 && bfqq->queued[1] == 0) queued 261 block/bfq-iosched.h int queued[2]; queued 519 block/bfq-iosched.h int queued; queued 820 block/bfq-iosched.h struct blkg_rwstat queued; queued 579 block/blk-mq-debugfs.c seq_printf(m, "%lu\n", hctx->queued); queued 588 block/blk-mq-debugfs.c hctx->queued = 0; queued 418 block/blk-mq.c data->hctx->queued++; queued 1048 block/blk-mq.c static inline unsigned int queued_to_index(unsigned int queued) queued 1050 block/blk-mq.c if (!queued) queued 1053 block/blk-mq.c return min(BLK_MQ_MAX_DISPATCH_ORDER - 1, ilog2(queued) + 1); queued 1217 block/blk-mq.c int errors, queued; queued 1228 block/blk-mq.c errors = queued = 0; queued 1297 block/blk-mq.c queued++; queued 1300 block/blk-mq.c hctx->dispatched[queued_to_index(queued)]++; queued 1364 block/blk-mq.c return (queued + errors) != 0; queued 80 block/blk-throttle.c struct list_head queued[2]; /* throtl_qnode [READ/WRITE] */ queued 407 block/blk-throttle.c struct list_head *queued) queued 411 block/blk-throttle.c list_add_tail(&qn->node, queued); queued 420 block/blk-throttle.c static struct bio *throtl_peek_queued(struct list_head *queued) queued 422 block/blk-throttle.c struct throtl_qnode *qn = list_first_entry(queued, struct throtl_qnode, node); queued 425 block/blk-throttle.c if (list_empty(queued)) queued 447 block/blk-throttle.c static struct bio *throtl_pop_queued(struct list_head *queued, queued 450 block/blk-throttle.c struct throtl_qnode *qn = list_first_entry(queued, struct throtl_qnode, node); queued 453 block/blk-throttle.c if (list_empty(queued)) queued 466 block/blk-throttle.c list_move_tail(&qn->node, queued); queued 475 block/blk-throttle.c INIT_LIST_HEAD(&sq->queued[0]); queued 476 block/blk-throttle.c INIT_LIST_HEAD(&sq->queued[1]); queued 980 block/blk-throttle.c bio != throtl_peek_queued(&tg->service_queue.queued[rw])); queued 1072 block/blk-throttle.c throtl_qnode_add_bio(bio, qn, &sq->queued[rw]); queued 1084 block/blk-throttle.c bio = throtl_peek_queued(&sq->queued[READ]); queued 1088 block/blk-throttle.c bio = throtl_peek_queued(&sq->queued[WRITE]); queued 1128 block/blk-throttle.c bio = throtl_pop_queued(&sq->queued[rw], &tg_to_put); queued 1145 block/blk-throttle.c &parent_sq->queued[rw]); queued 1166 block/blk-throttle.c while ((bio = throtl_peek_queued(&sq->queued[READ])) && queued 1176 block/blk-throttle.c while ((bio = throtl_peek_queued(&sq->queued[WRITE])) && queued 1318 block/blk-throttle.c while ((bio = throtl_pop_queued(&td_sq->queued[rw], NULL))) queued 2323 block/blk-throttle.c while ((bio = throtl_peek_queued(&sq->queued[READ]))) queued 2325 block/blk-throttle.c while ((bio = throtl_peek_queued(&sq->queued[WRITE]))) queued 2364 block/blk-throttle.c while ((bio = throtl_pop_queued(&td->service_queue.queued[rw], queued 153 drivers/atm/eni.c static int tx_complete = 0,dma_complete = 0,queued = 0,requeued = 0, queued 1179 drivers/atm/eni.c queued++; queued 1466 drivers/atm/eni.c tx_complete,dma_complete,queued,requeued,submitted,backlogged, queued 635 drivers/block/drbd/drbd_worker.c int queued = sk->sk_wmem_queued; queued 637 drivers/block/drbd/drbd_worker.c if (queued > sndbuf / 2) { queued 567 drivers/crypto/inside-secure/safexcel_cipher.c int queued, i, ret = 0; queued 608 drivers/crypto/inside-secure/safexcel_cipher.c queued = totlen_src; queued 649 drivers/crypto/inside-secure/safexcel_cipher.c if (queued - len < 0) queued 650 drivers/crypto/inside-secure/safexcel_cipher.c len = queued; queued 653 drivers/crypto/inside-secure/safexcel_cipher.c !(queued - len), queued 667 drivers/crypto/inside-secure/safexcel_cipher.c queued -= len; queued 668 drivers/crypto/inside-secure/safexcel_cipher.c if (!queued) queued 265 drivers/crypto/inside-secure/safexcel_hash.c u64 queued, len, cache_len; queued 267 drivers/crypto/inside-secure/safexcel_hash.c queued = len = safexcel_queued_len(req); queued 268 drivers/crypto/inside-secure/safexcel_hash.c if (queued <= HASH_CACHE_SIZE) queued 269 drivers/crypto/inside-secure/safexcel_hash.c cache_len = queued; queued 271 drivers/crypto/inside-secure/safexcel_hash.c cache_len = queued - areq->nbytes; queued 277 drivers/crypto/inside-secure/safexcel_hash.c extra = queued & (HASH_CACHE_SIZE - 1); queued 289 drivers/crypto/inside-secure/safexcel_hash.c queued -= extra; queued 292 drivers/crypto/inside-secure/safexcel_hash.c if (!queued) { queued 317 drivers/crypto/inside-secure/safexcel_hash.c queued -= cache_len; queued 318 drivers/crypto/inside-secure/safexcel_hash.c if (!queued) queued 340 drivers/crypto/inside-secure/safexcel_hash.c if (queued < sglen) queued 341 drivers/crypto/inside-secure/safexcel_hash.c sglen = queued; queued 344 drivers/crypto/inside-secure/safexcel_hash.c !(queued - sglen), queued 356 drivers/crypto/inside-secure/safexcel_hash.c queued -= sglen; queued 357 drivers/crypto/inside-secure/safexcel_hash.c if (!queued) queued 197 drivers/dma/mpc512x_dma.c struct list_head queued; queued 262 drivers/dma/mpc512x_dma.c while (!list_empty(&mchan->queued)) { queued 263 drivers/dma/mpc512x_dma.c mdesc = list_first_entry(&mchan->queued, queued 341 drivers/dma/mpc512x_dma.c if (!list_empty(&mchan->queued)) queued 472 drivers/dma/mpc512x_dma.c list_move_tail(&mdesc->node, &mchan->queued); queued 558 drivers/dma/mpc512x_dma.c BUG_ON(!list_empty(&mchan->queued)); queued 889 drivers/dma/mpc512x_dma.c list_splice_tail_init(&mchan->queued, &mchan->free); queued 1004 drivers/dma/mpc512x_dma.c INIT_LIST_HEAD(&mchan->queued); queued 214 drivers/dma/nbpfaxi.c struct list_head queued; queued 603 drivers/dma/nbpfaxi.c if (list_empty(&chan->queued)) queued 606 drivers/dma/nbpfaxi.c list_splice_tail_init(&chan->queued, &chan->active); queued 647 drivers/dma/nbpfaxi.c list_for_each_entry(desc, &chan->queued, node) queued 675 drivers/dma/nbpfaxi.c list_add_tail(&desc->node, &chan->queued); queued 843 drivers/dma/nbpfaxi.c list_splice_init(&chan->queued, &head); queued 1040 drivers/dma/nbpfaxi.c INIT_LIST_HEAD(&chan->queued); queued 178 drivers/dma/qcom/hidma.c bool queued = false; queued 184 drivers/dma/qcom/hidma.c queued = true; queued 194 drivers/dma/qcom/hidma.c if (queued) { queued 219 drivers/dma/qcom/hidma.c INIT_LIST_HEAD(&mchan->queued); queued 244 drivers/dma/qcom/hidma.c list_for_each_entry_safe(qdesc, next, &mchan->queued, node) { queued 338 drivers/dma/qcom/hidma.c list_move_tail(&mdesc->node, &mchan->queued); queued 483 drivers/dma/qcom/hidma.c list_splice_init(&mchan->queued, &list); queued 30 drivers/dma/qcom/hidma.h bool queued; /* flag whether this is pending */ queued 102 drivers/dma/qcom/hidma.h struct list_head queued; queued 31 drivers/dma/qcom/hidma_dbg.c seq_printf(s, "queued = 0x%x\n", tre->queued); queued 160 drivers/dma/qcom/hidma_ll.c tre->queued = 0; queued 223 drivers/dma/qcom/hidma_ll.c tre->queued = 0; queued 539 drivers/dma/qcom/hidma_ll.c tre->queued = 1; queued 101 drivers/dma/sirf-dma.c struct list_head queued; queued 248 drivers/dma/sirf-dma.c sdesc = list_first_entry(&schan->queued, struct sirfsoc_dma_desc, queued 291 drivers/dma/sirf-dma.c if (!list_empty(&schan->queued)) queued 317 drivers/dma/sirf-dma.c if (!list_empty(&schan->queued)) queued 416 drivers/dma/sirf-dma.c list_move_tail(&sdesc->node, &schan->queued); queued 481 drivers/dma/sirf-dma.c list_splice_tail_init(&schan->queued, &schan->free); queued 608 drivers/dma/sirf-dma.c BUG_ON(!list_empty(&schan->queued)); queued 632 drivers/dma/sirf-dma.c if (list_empty(&schan->active) && !list_empty(&schan->queued)) queued 933 drivers/dma/sirf-dma.c INIT_LIST_HEAD(&schan->queued); queued 63 drivers/gpu/drm/drm_flip_work.c list_add_tail(&task->node, &work->queued); queued 107 drivers/gpu/drm/drm_flip_work.c list_splice_tail(&work->queued, &work->commited); queued 108 drivers/gpu/drm/drm_flip_work.c INIT_LIST_HEAD(&work->queued); queued 151 drivers/gpu/drm/drm_flip_work.c INIT_LIST_HEAD(&work->queued); queued 168 drivers/gpu/drm/drm_flip_work.c WARN_ON(!list_empty(&work->queued) || !list_empty(&work->commited)); queued 746 drivers/infiniband/hw/hfi1/user_sdma.c u32 queued = *queued_ptr, data_sent = *data_sent_ptr; queued 755 drivers/infiniband/hw/hfi1/user_sdma.c len = min((datalen - queued), len); queued 763 drivers/infiniband/hw/hfi1/user_sdma.c queued += len; queued 765 drivers/infiniband/hw/hfi1/user_sdma.c if (unlikely(queued < datalen && pageidx == iovec->npages && queued 772 drivers/infiniband/hw/hfi1/user_sdma.c *queued_ptr = queued; queued 809 drivers/infiniband/hw/hfi1/user_sdma.c u32 datalen = 0, queued = 0, data_sent = 0; queued 907 drivers/infiniband/hw/hfi1/user_sdma.c while (queued < datalen && queued 910 drivers/infiniband/hw/hfi1/user_sdma.c &queued, &data_sent, &iov_offset); queued 26 drivers/md/dm-cache-background-tracker.c struct list_head queued; queued 47 drivers/md/dm-cache-background-tracker.c INIT_LIST_HEAD(&b->queued); queued 205 drivers/md/dm-cache-background-tracker.c list_add(&w->list, &b->queued); queued 219 drivers/md/dm-cache-background-tracker.c if (list_empty(&b->queued)) queued 222 drivers/md/dm-cache-background-tracker.c w = list_first_entry(&b->queued, struct bt_work, list); queued 2407 drivers/media/common/videobuf2/videobuf2-core.c unsigned int queued:1; queued 2537 drivers/media/common/videobuf2/videobuf2-core.c fileio->bufs[i].queued = 1; queued 2657 drivers/media/common/videobuf2/videobuf2-core.c buf->queued = 0; queued 2728 drivers/media/common/videobuf2/videobuf2-core.c buf->queued = 1; queued 89 drivers/media/platform/fsl-viu.c struct list_head queued; queued 314 drivers/media/platform/fsl-viu.c if (list_empty(&vidq->queued)) queued 316 drivers/media/platform/fsl-viu.c buf = list_entry(vidq->queued.next, struct viu_buf, vb.queue); queued 500 drivers/media/platform/fsl-viu.c if (!list_empty(&vidq->queued)) { queued 503 drivers/media/platform/fsl-viu.c vidq, &vidq->queued); queued 505 drivers/media/platform/fsl-viu.c dev, &vidq->queued, vidq->queued.next, queued 506 drivers/media/platform/fsl-viu.c vidq->queued.prev); queued 507 drivers/media/platform/fsl-viu.c list_add_tail(&buf->vb.queue, &vidq->queued); queued 531 drivers/media/platform/fsl-viu.c list_add_tail(&buf->vb.queue, &vidq->queued); queued 967 drivers/media/platform/fsl-viu.c } else if (!list_empty(&vidq->queued)) { queued 968 drivers/media/platform/fsl-viu.c buf = list_entry(vidq->queued.next, struct viu_buf, queued 1204 drivers/media/platform/fsl-viu.c list_empty(&dev->vidq.queued)); queued 1434 drivers/media/platform/fsl-viu.c INIT_LIST_HEAD(&viu_dev->vidq.queued); queued 224 drivers/media/platform/vsp1/vsp1_dl.c struct vsp1_dl_list *queued; queued 839 drivers/media/platform/vsp1/vsp1_dl.c if (!dlm->queued) queued 897 drivers/media/platform/vsp1/vsp1_dl.c __vsp1_dl_list_put(dlm->queued); queued 898 drivers/media/platform/vsp1/vsp1_dl.c dlm->queued = dl; queued 1018 drivers/media/platform/vsp1/vsp1_dl.c if (dlm->queued) { queued 1019 drivers/media/platform/vsp1/vsp1_dl.c if (dlm->queued->flags & VSP1_DL_FRAME_END_INTERNAL) queued 1021 drivers/media/platform/vsp1/vsp1_dl.c dlm->queued->flags &= ~VSP1_DL_FRAME_END_INTERNAL; queued 1024 drivers/media/platform/vsp1/vsp1_dl.c dlm->active = dlm->queued; queued 1025 drivers/media/platform/vsp1/vsp1_dl.c dlm->queued = NULL; queued 1036 drivers/media/platform/vsp1/vsp1_dl.c dlm->queued = dlm->pending; queued 1072 drivers/media/platform/vsp1/vsp1_dl.c __vsp1_dl_list_put(dlm->queued); queued 1078 drivers/media/platform/vsp1/vsp1_dl.c dlm->queued = NULL; queued 1484 drivers/media/usb/cx231xx/cx231xx-cards.c INIT_LIST_HEAD(&dev->video_mode.vidq.queued); queued 1488 drivers/media/usb/cx231xx/cx231xx-cards.c INIT_LIST_HEAD(&dev->vbi_mode.vidq.queued); queued 240 drivers/media/usb/cx231xx/cx231xx.h struct list_head queued; queued 1348 drivers/media/usb/tm6000/tm6000-video.c list_empty(&dev->vidq.queued)); queued 1640 drivers/media/usb/tm6000/tm6000-video.c INIT_LIST_HEAD(&dev->vidq.queued); queued 81 drivers/media/usb/tm6000/tm6000.h struct list_head queued; queued 572 drivers/mtd/ftl.c int queued, ret; queued 579 drivers/mtd/ftl.c queued = 0; queued 590 drivers/mtd/ftl.c queued = 1; queued 610 drivers/mtd/ftl.c if (queued) { queued 3195 drivers/net/wireless/ath/ath6kl/cfg80211.c bool more_data, queued; queued 3233 drivers/net/wireless/ath/ath6kl/cfg80211.c queued = ath6kl_mgmt_powersave_ap(vif, id, freq, wait, buf, len, queued 3235 drivers/net/wireless/ath/ath6kl/cfg80211.c if (queued) queued 598 drivers/net/wireless/ath/ath9k/debug.c PR("MPDUs Queued: ", queued); queued 1329 drivers/net/wireless/ath/ath9k/debug.c AWDATA(queued); queued 171 drivers/net/wireless/ath/ath9k/debug.h u32 queued; queued 2063 drivers/net/wireless/ath/ath9k/xmit.c TX_STAT_INC(sc, txq->axq_qnum, queued); queued 2389 drivers/net/wireless/ath/ath9k/xmit.c TX_STAT_INC(sc, txctl.txq->axq_qnum, queued); queued 1432 drivers/net/wireless/ath/carl9170/tx.c goto queued; queued 1440 drivers/net/wireless/ath/carl9170/tx.c goto queued; queued 1445 drivers/net/wireless/ath/carl9170/tx.c queued: queued 42 drivers/net/wireless/mediatek/mt76/debugfs.c i, q->q->queued, q->q->head, q->q->tail, queued 85 drivers/net/wireless/mediatek/mt76/dma.c q->queued++; queued 157 drivers/net/wireless/mediatek/mt76/dma.c while ((q->queued > n_queued) && q->tail != last) { queued 180 drivers/net/wireless/mediatek/mt76/dma.c q->queued -= n_queued; queued 192 drivers/net/wireless/mediatek/mt76/dma.c qid < IEEE80211_NUM_ACS && q->queued < q->ndesc - 8; queued 196 drivers/net/wireless/mediatek/mt76/dma.c if (!q->queued) queued 238 drivers/net/wireless/mediatek/mt76/dma.c if (!q->queued) queued 245 drivers/net/wireless/mediatek/mt76/dma.c q->queued--; queued 339 drivers/net/wireless/mediatek/mt76/dma.c if (q->queued + (tx_info.nbuf + 1) / 2 >= q->ndesc - 1) { queued 372 drivers/net/wireless/mediatek/mt76/dma.c while (q->queued < q->ndesc - 1) { queued 393 drivers/net/wireless/mediatek/mt76/mac80211.c if (q && q->queued) queued 119 drivers/net/wireless/mediatek/mt76/mt76.h int queued; queued 139 drivers/net/wireless/mediatek/mt76/mt7603/beacon.c if (dev->mt76.q_tx[MT_TXQ_BEACON].q->queued > queued 1400 drivers/net/wireless/mediatek/mt76/mt7603/mac.c if (!q->queued) queued 366 drivers/net/wireless/mediatek/mt76/mt76x02_mmio.c if (!q->queued) queued 284 drivers/net/wireless/mediatek/mt76/tx.c if (q->queued > q->ndesc - 8 && !q->stopped) { queued 398 drivers/net/wireless/mediatek/mt76/usb.c if (q->queued > 0) { queued 401 drivers/net/wireless/mediatek/mt76/usb.c q->queued--; queued 521 drivers/net/wireless/mediatek/mt76/usb.c q->queued++; queued 576 drivers/net/wireless/mediatek/mt76/usb.c q->queued = 0; queued 667 drivers/net/wireless/mediatek/mt76/usb.c while (q->queued > n_dequeued) { queued 687 drivers/net/wireless/mediatek/mt76/usb.c q->queued -= n_dequeued; queued 689 drivers/net/wireless/mediatek/mt76/usb.c wake = q->stopped && q->queued < q->ndesc - 8; queued 693 drivers/net/wireless/mediatek/mt76/usb.c if (!q->queued) queued 779 drivers/net/wireless/mediatek/mt76/usb.c if (q->queued == q->ndesc) queued 797 drivers/net/wireless/mediatek/mt76/usb.c q->queued++; queued 901 drivers/net/wireless/mediatek/mt76/usb.c while (q->queued) { queued 904 drivers/net/wireless/mediatek/mt76/usb.c q->queued--; queued 1586 drivers/net/wireless/st/cw1200/wsm.c int queued; queued 1591 drivers/net/wireless/st/cw1200/wsm.c queued = cw1200_queue_get_num_queued(&priv->tx_queue[i], queued 1593 drivers/net/wireless/st/cw1200/wsm.c if (!queued) queued 1595 drivers/net/wireless/st/cw1200/wsm.c *total += queued; queued 1568 drivers/scsi/ncr53c8xx.c u_char queued; queued 4457 drivers/scsi/ncr53c8xx.c cp->queued = 1; queued 5093 drivers/scsi/ncr53c8xx.c if (cp->queued) { queued 5097 drivers/scsi/ncr53c8xx.c if (cp->queued) { queued 5099 drivers/scsi/ncr53c8xx.c cp->queued = 0; queued 7283 drivers/scsi/ncr53c8xx.c if (cp->queued) { queued 7289 drivers/scsi/ncr53c8xx.c if (cp->queued) { queued 7291 drivers/scsi/ncr53c8xx.c cp->queued = 0; queued 292 drivers/scsi/qla2xxx/qla_target.c uint8_t queued = 0; queued 318 drivers/scsi/qla2xxx/qla_target.c if (!queued) { queued 319 drivers/scsi/qla2xxx/qla_target.c queued = 1; queued 1665 drivers/spi/spi.c ctlr->queued = true; queued 2590 drivers/spi/spi.c if (ctlr->queued) { queued 2612 drivers/spi/spi.c if (!ctlr->queued) queued 2627 drivers/spi/spi.c if (!ctlr->queued) queued 464 drivers/tty/hvc/hvc_iucv.c int queued; queued 473 drivers/tty/hvc/hvc_iucv.c queued = hvc_iucv_queue(priv, buf, count); queued 476 drivers/tty/hvc/hvc_iucv.c return queued; queued 255 drivers/tty/serial/amba-pl011.c bool queued; queued 556 drivers/tty/serial/amba-pl011.c if (uap->dmatx.queued) queued 575 drivers/tty/serial/amba-pl011.c uap->dmatx.queued = false; queued 615 drivers/tty/serial/amba-pl011.c uap->dmatx.queued = false; queued 647 drivers/tty/serial/amba-pl011.c uap->dmatx.queued = false; queued 656 drivers/tty/serial/amba-pl011.c uap->dmatx.queued = false; queued 677 drivers/tty/serial/amba-pl011.c uap->dmatx.queued = true; queued 710 drivers/tty/serial/amba-pl011.c if (uap->dmatx.queued) { queued 736 drivers/tty/serial/amba-pl011.c if (uap->dmatx.queued) { queued 761 drivers/tty/serial/amba-pl011.c if (!uap->dmatx.queued) { queued 818 drivers/tty/serial/amba-pl011.c if (uap->dmatx.queued) { queued 821 drivers/tty/serial/amba-pl011.c uap->dmatx.queued = false; queued 1204 drivers/tty/serial/amba-pl011.c if (uap->dmatx.queued) { queued 1207 drivers/tty/serial/amba-pl011.c uap->dmatx.queued = false; queued 224 drivers/usb/dwc3/trace.h __field(u32, queued) queued 130 drivers/usb/mon/mon_bin.c u32 queued; queued 1120 drivers/usb/mon/mon_bin.c if (put_user(nevents, &sp->queued)) queued 1798 drivers/usb/serial/io_ti.c int queued; queued 1800 drivers/usb/serial/io_ti.c queued = tty_insert_flip_string(&port->port, data, length); queued 1801 drivers/usb/serial/io_ti.c if (queued < length) queued 1803 drivers/usb/serial/io_ti.c __func__, length - queued); queued 302 fs/f2fs/f2fs.h unsigned char queued; /* queued discard */ queued 954 fs/f2fs/segment.c dc->queued = 0; queued 987 fs/f2fs/segment.c atomic_sub(dc->queued, &dcc->queued_discard); queued 1194 fs/f2fs/segment.c dc->queued++; queued 517 fs/fuse/dev.c bool queued = false; queued 536 fs/fuse/dev.c queued = true; queued 540 fs/fuse/dev.c return queued; queued 816 fs/ocfs2/cluster/heartbeat.c int queued = 0; queued 834 fs/ocfs2/cluster/heartbeat.c queued = 1; queued 839 fs/ocfs2/cluster/heartbeat.c if (queued) queued 900 fs/ocfs2/cluster/heartbeat.c int queued = 0; queued 994 fs/ocfs2/cluster/heartbeat.c queued = 1; queued 1046 fs/ocfs2/cluster/heartbeat.c queued = 1; queued 1062 fs/ocfs2/cluster/heartbeat.c if (queued) queued 102 fs/xfs/xfs_mru_cache.c unsigned int queued; /* work has been queued */ queued 204 fs/xfs/xfs_mru_cache.c if (!mru->queued) { queued 205 fs/xfs/xfs_mru_cache.c mru->queued = 1; queued 280 fs/xfs/xfs_mru_cache.c mru->queued = next; queued 281 fs/xfs/xfs_mru_cache.c if ((mru->queued > 0)) { queued 388 fs/xfs/xfs_mru_cache.c if (mru->queued) { queued 77 include/drm/drm_flip_work.h struct list_head queued; queued 51 include/linux/blk-mq.h unsigned long queued; queued 301 include/linux/kvm_host.h u32 queued; queued 550 include/linux/spi/spi.h bool queued; queued 1589 kernel/sched/core.c bool queued, running; queued 1593 kernel/sched/core.c queued = task_on_rq_queued(p); queued 1596 kernel/sched/core.c if (queued) { queued 1609 kernel/sched/core.c if (queued) queued 1889 kernel/sched/core.c int running, queued; queued 1928 kernel/sched/core.c queued = task_on_rq_queued(p); queued 1960 kernel/sched/core.c if (unlikely(queued)) { queued 4372 kernel/sched/core.c int prio, oldprio, queued, running, queue_flag = queued 4432 kernel/sched/core.c queued = task_on_rq_queued(p); queued 4434 kernel/sched/core.c if (queued) queued 4472 kernel/sched/core.c if (queued) queued 4495 kernel/sched/core.c bool queued, running; queued 4519 kernel/sched/core.c queued = task_on_rq_queued(p); queued 4521 kernel/sched/core.c if (queued) queued 4532 kernel/sched/core.c if (queued) { queued 4755 kernel/sched/core.c int retval, oldprio, oldpolicy = -1, queued, running; queued 4967 kernel/sched/core.c queued = task_on_rq_queued(p); queued 4969 kernel/sched/core.c if (queued) queued 4979 kernel/sched/core.c if (queued) { queued 6147 kernel/sched/core.c bool queued, running; queued 6152 kernel/sched/core.c queued = task_on_rq_queued(p); queued 6155 kernel/sched/core.c if (queued) queued 6162 kernel/sched/core.c if (queued) queued 7035 kernel/sched/core.c int queued, running, queue_flags = queued 7044 kernel/sched/core.c queued = task_on_rq_queued(tsk); queued 7046 kernel/sched/core.c if (queued) queued 7053 kernel/sched/core.c if (queued) queued 1812 kernel/sched/deadline.c static void task_tick_dl(struct rq *rq, struct task_struct *p, int queued) queued 1822 kernel/sched/deadline.c if (hrtick_enabled(rq) && queued && p->dl.runtime > 0 && queued 4281 kernel/sched/fair.c entity_tick(struct cfs_rq *cfs_rq, struct sched_entity *curr, int queued) queued 4299 kernel/sched/fair.c if (queued) { queued 9979 kernel/sched/fair.c static void task_tick_fair(struct rq *rq, struct task_struct *curr, int queued) queued 9986 kernel/sched/fair.c entity_tick(cfs_rq, se, queued); queued 428 kernel/sched/idle.c static void task_tick_idle(struct rq *rq, struct task_struct *curr, int queued) queued 2315 kernel/sched/rt.c static void task_tick_rt(struct rq *rq, struct task_struct *p, int queued) queued 1753 kernel/sched/sched.h void (*task_tick)(struct rq *rq, struct task_struct *p, int queued); queued 93 kernel/sched/stop_task.c static void task_tick_stop(struct rq *rq, struct task_struct *curr, int queued) queued 379 kernel/stop_machine.c bool queued = false; queued 395 kernel/stop_machine.c queued = true; queued 401 kernel/stop_machine.c return queued; queued 147 net/ax25/ax25_ds_in.c int queued = 0; queued 240 net/ax25/ax25_ds_in.c queued = ax25_rx_iframe(ax25, skb); queued 273 net/ax25/ax25_ds_in.c return queued; queued 281 net/ax25/ax25_ds_in.c int queued = 0, frametype, ns, nr, pf; queued 287 net/ax25/ax25_ds_in.c queued = ax25_ds_state1_machine(ax25, skb, frametype, pf, type); queued 290 net/ax25/ax25_ds_in.c queued = ax25_ds_state2_machine(ax25, skb, frametype, pf, type); queued 293 net/ax25/ax25_ds_in.c queued = ax25_ds_state3_machine(ax25, skb, frametype, ns, nr, pf, type); queued 297 net/ax25/ax25_ds_in.c return queued; queued 103 net/ax25/ax25_in.c int queued = 0; queued 145 net/ax25/ax25_in.c queued = 1; queued 151 net/ax25/ax25_in.c return queued; queued 159 net/ax25/ax25_in.c int queued = 0; queued 167 net/ax25/ax25_in.c queued = ax25_std_frame_in(ax25, skb, type); queued 173 net/ax25/ax25_in.c queued = ax25_ds_frame_in(ax25, skb, type); queued 175 net/ax25/ax25_in.c queued = ax25_std_frame_in(ax25, skb, type); queued 180 net/ax25/ax25_in.c return queued; queued 143 net/ax25/ax25_std_in.c int queued = 0; queued 225 net/ax25/ax25_std_in.c queued = ax25_rx_iframe(ax25, skb); queued 258 net/ax25/ax25_std_in.c return queued; queued 268 net/ax25/ax25_std_in.c int queued = 0; queued 380 net/ax25/ax25_std_in.c queued = ax25_rx_iframe(ax25, skb); queued 413 net/ax25/ax25_std_in.c return queued; queued 421 net/ax25/ax25_std_in.c int queued = 0, frametype, ns, nr, pf; queued 427 net/ax25/ax25_std_in.c queued = ax25_std_state1_machine(ax25, skb, frametype, pf, type); queued 430 net/ax25/ax25_std_in.c queued = ax25_std_state2_machine(ax25, skb, frametype, pf, type); queued 433 net/ax25/ax25_std_in.c queued = ax25_std_state3_machine(ax25, skb, frametype, ns, nr, pf, type); queued 436 net/ax25/ax25_std_in.c queued = ax25_std_state4_machine(ax25, skb, frametype, ns, nr, pf, type); queued 442 net/ax25/ax25_std_in.c return queued; queued 130 net/caif/caif_socket.c bool queued = false; queued 153 net/caif/caif_socket.c queued = !sock_flag(sk, SOCK_DEAD); queued 154 net/caif/caif_socket.c if (queued) queued 158 net/caif/caif_socket.c if (queued) queued 45 net/dccp/input.c int queued = 0; queued 76 net/dccp/input.c queued = 1; queued 86 net/dccp/input.c return queued; queued 91 net/dccp/input.c int queued = 0; queued 101 net/dccp/input.c return queued; queued 113 net/dccp/input.c queued = 1; queued 120 net/dccp/input.c return queued; queued 524 net/dccp/input.c int queued = 0; queued 562 net/dccp/input.c queued = 1; /* packet was queued queued 568 net/dccp/input.c return queued; queued 578 net/dccp/input.c int queued = 0; queued 676 net/dccp/input.c queued = dccp_rcv_request_sent_state_process(sk, skb, dh, len); queued 677 net/dccp/input.c if (queued >= 0) queued 678 net/dccp/input.c return queued; queued 689 net/dccp/input.c queued = dccp_rcv_respond_partopen_state_process(sk, skb, queued 707 net/dccp/input.c if (!queued) { queued 609 net/decnet/dn_nsp_in.c int queued = 0; queued 622 net/decnet/dn_nsp_in.c queued = 1; queued 628 net/decnet/dn_nsp_in.c if (!queued) queued 634 net/decnet/dn_nsp_in.c int queued = 0; queued 648 net/decnet/dn_nsp_in.c queued = 1; queued 659 net/decnet/dn_nsp_in.c if (!queued) queued 6141 net/ipv4/tcp_input.c int queued = 0; queued 6177 net/ipv4/tcp_input.c queued = tcp_rcv_synsent_state_process(sk, skb, th); queued 6178 net/ipv4/tcp_input.c if (queued >= 0) queued 6179 net/ipv4/tcp_input.c return queued; queued 6358 net/ipv4/tcp_input.c queued = 1; queued 6368 net/ipv4/tcp_input.c if (!queued) { queued 249 net/lapb/lapb_in.c int queued = 0; queued 403 net/lapb/lapb_in.c queued = 1; queued 464 net/lapb/lapb_in.c if (!queued) queued 555 net/mac80211/offchannel.c bool queued = false, combine_started = true; queued 634 net/mac80211/offchannel.c queued = true; queued 662 net/mac80211/offchannel.c queued = true; queued 669 net/mac80211/offchannel.c queued = ieee80211_coalesce_hw_started_roc(local, roc, tmp); queued 670 net/mac80211/offchannel.c if (queued) queued 680 net/mac80211/offchannel.c if (!queued) queued 1087 net/mac80211/tx.c bool queued = false; queued 1127 net/mac80211/tx.c queued = true; queued 1151 net/mac80211/tx.c return queued; queued 1211 net/mac80211/tx.c bool queued; queued 1213 net/mac80211/tx.c queued = ieee80211_tx_prep_agg(tx, skb, info, queued 1216 net/mac80211/tx.c if (unlikely(queued)) queued 779 net/netfilter/nfnetlink_queue.c unsigned int queued; queued 816 net/netfilter/nfnetlink_queue.c queued = 0; queued 824 net/netfilter/nfnetlink_queue.c queued++; queued 830 net/netfilter/nfnetlink_queue.c if (queued) { queued 153 net/netrom/nr_in.c int queued = 0; queued 226 net/netrom/nr_in.c queued = 1; queued 273 net/netrom/nr_in.c return queued; queued 280 net/netrom/nr_in.c int queued = 0, frametype; queued 289 net/netrom/nr_in.c queued = nr_state1_machine(sk, skb, frametype); queued 292 net/netrom/nr_in.c queued = nr_state2_machine(sk, skb, frametype); queued 295 net/netrom/nr_in.c queued = nr_state3_machine(sk, skb, frametype); queued 301 net/netrom/nr_in.c return queued; queued 821 net/rds/send.c __be16 dport, int *queued) queued 826 net/rds/send.c if (*queued) queued 877 net/rds/send.c *queued = 1; queued 882 net/rds/send.c return *queued; queued 1111 net/rds/send.c int queued = 0, allocated_mr = 0; queued 1351 net/rds/send.c dport, &queued)) { queued 1363 net/rds/send.c &queued), queued 1365 net/rds/send.c rdsdebug("sendmsg woke queued %d timeo %ld\n", queued, timeo); queued 104 net/rose/rose_in.c int queued = 0; queued 167 net/rose/rose_in.c queued = 1; queued 204 net/rose/rose_in.c return queued; queued 265 net/rose/rose_in.c int queued = 0, frametype, ns, nr, q, d, m; queued 274 net/rose/rose_in.c queued = rose_state1_machine(sk, skb, frametype); queued 277 net/rose/rose_in.c queued = rose_state2_machine(sk, skb, frametype); queued 280 net/rose/rose_in.c queued = rose_state3_machine(sk, skb, frametype, ns, nr, q, d, m); queued 283 net/rose/rose_in.c queued = rose_state4_machine(sk, skb, frametype); queued 286 net/rose/rose_in.c queued = rose_state5_machine(sk, skb, frametype); queued 292 net/rose/rose_in.c return queued; queued 51 net/x25/x25_dev.c int queued = 1; queued 56 net/x25/x25_dev.c queued = x25_process_rx_frame(sk, skb); queued 58 net/x25/x25_dev.c queued = !sk_add_backlog(sk, skb, READ_ONCE(sk->sk_rcvbuf)); queued 62 net/x25/x25_dev.c return queued; queued 210 net/x25/x25_in.c int queued = 0; queued 277 net/x25/x25_in.c queued = 1; queued 315 net/x25/x25_in.c queued = !sock_queue_rcv_skb(sk, skb); queued 319 net/x25/x25_in.c queued = 1; queued 330 net/x25/x25_in.c return queued; queued 389 net/x25/x25_in.c int queued = 0, frametype, ns, nr, q, d, m; queued 398 net/x25/x25_in.c queued = x25_state1_machine(sk, skb, frametype); queued 401 net/x25/x25_in.c queued = x25_state2_machine(sk, skb, frametype); queued 404 net/x25/x25_in.c queued = x25_state3_machine(sk, skb, frametype, ns, nr, q, d, m); queued 407 net/x25/x25_in.c queued = x25_state4_machine(sk, skb, frametype); queued 413 net/x25/x25_in.c return queued; queued 418 net/x25/x25_in.c int queued = x25_process_rx_frame(sk, skb); queued 420 net/x25/x25_in.c if (!queued) queued 127 sound/firewire/fireworks/fireworks_hwdep.c bool queued; queued 132 sound/firewire/fireworks/fireworks_hwdep.c queued = efw->push_ptr != efw->pull_ptr; queued 134 sound/firewire/fireworks/fireworks_hwdep.c while (!dev_lock_changed && !queued) { queued 143 sound/firewire/fireworks/fireworks_hwdep.c queued = efw->push_ptr != efw->pull_ptr; queued 150 sound/firewire/fireworks/fireworks_hwdep.c else if (queued) queued 144 virt/kvm/async_pf.c vcpu->async_pf.queued = 0; queued 163 virt/kvm/async_pf.c vcpu->async_pf.queued--; queued 173 virt/kvm/async_pf.c if (vcpu->async_pf.queued >= ASYNC_PF_PER_VCPU) queued 205 virt/kvm/async_pf.c vcpu->async_pf.queued++; queued 233 virt/kvm/async_pf.c vcpu->async_pf.queued++;