inflight 270 block/blk-iolatency.c atomic_dec(&rqw->inflight); inflight 299 block/blk-iolatency.c atomic_inc(&rqw->inflight); inflight 597 block/blk-iolatency.c int inflight = 0; inflight 619 block/blk-iolatency.c inflight = atomic_dec_return(&rqw->inflight); inflight 620 block/blk-iolatency.c WARN_ON_ONCE(inflight < 0); inflight 96 block/blk-mq.c unsigned int *inflight; inflight 109 block/blk-mq.c mi->inflight[0]++; inflight 116 block/blk-mq.c unsigned inflight[2]; inflight 117 block/blk-mq.c struct mq_inflight mi = { .part = part, .inflight = inflight, }; inflight 119 block/blk-mq.c inflight[0] = inflight[1] = 0; inflight 122 block/blk-mq.c return inflight[0]; inflight 132 block/blk-mq.c mi->inflight[rq_data_dir(rq)]++; inflight 138 block/blk-mq.c unsigned int inflight[2]) inflight 140 block/blk-mq.c struct mq_inflight mi = { .part = part, .inflight = inflight, }; inflight 142 block/blk-mq.c inflight[0] = inflight[1] = 0; inflight 190 block/blk-mq.h unsigned int inflight[2]); inflight 29 block/blk-rq-qos.c return atomic_inc_below(&rq_wait->inflight, limit); inflight 23 block/blk-rq-qos.h atomic_t inflight; inflight 96 block/blk-rq-qos.h atomic_set(&rq_wait->inflight, 0); inflight 130 block/blk-wbt.c int inflight, limit; inflight 132 block/blk-wbt.c inflight = atomic_dec_return(&rqw->inflight); inflight 158 block/blk-wbt.c if (inflight && inflight >= limit) inflight 162 block/blk-wbt.c int diff = limit - inflight; inflight 164 block/blk-wbt.c if (!inflight || diff >= rwb->wb_background / 2) inflight 356 block/blk-wbt.c unsigned int inflight = wbt_inflight(rwb); inflight 362 block/blk-wbt.c inflight); inflight 404 block/blk-wbt.c if (rqd->scale_step || inflight) inflight 753 block/blk-wbt.c atomic_read(&rwb->rq_wait[i].inflight)); inflight 82 block/blk-wbt.h ret += atomic_read(&rwb->rq_wait[i].inflight); inflight 72 block/genhd.c unsigned int inflight; inflight 78 block/genhd.c inflight = 0; inflight 80 block/genhd.c inflight += part_stat_local_read_cpu(part, in_flight[0], cpu) + inflight 83 block/genhd.c if ((int)inflight < 0) inflight 84 block/genhd.c inflight = 0; inflight 86 block/genhd.c return inflight; inflight 90 block/genhd.c unsigned int inflight[2]) inflight 95 block/genhd.c blk_mq_in_flight_rw(q, part, inflight); inflight 99 block/genhd.c inflight[0] = 0; inflight 100 block/genhd.c inflight[1] = 0; inflight 102 block/genhd.c inflight[0] += part_stat_local_read_cpu(part, in_flight[0], cpu); inflight 103 block/genhd.c inflight[1] += part_stat_local_read_cpu(part, in_flight[1], cpu); inflight 105 block/genhd.c if ((int)inflight[0] < 0) inflight 106 block/genhd.c inflight[0] = 0; inflight 107 block/genhd.c if ((int)inflight[1] < 0) inflight 108 block/genhd.c inflight[1] = 0; inflight 1193 block/genhd.c static DEVICE_ATTR(inflight, 0444, part_inflight_show, NULL); inflight 1371 block/genhd.c unsigned int inflight; inflight 1383 block/genhd.c inflight = part_in_flight(gp->queue, hd); inflight 1399 block/genhd.c inflight, inflight 123 block/partition-generic.c unsigned int inflight; inflight 125 block/partition-generic.c inflight = part_in_flight(q, p); inflight 140 block/partition-generic.c inflight, inflight 154 block/partition-generic.c unsigned int inflight[2]; inflight 156 block/partition-generic.c part_in_flight_rw(q, p, inflight); inflight 157 block/partition-generic.c return sprintf(buf, "%8u %8u\n", inflight[0], inflight[1]); inflight 190 block/partition-generic.c static DEVICE_ATTR(inflight, 0444, part_inflight_show, NULL); inflight 746 drivers/block/xen-blkback/blkback.c if (atomic_dec_and_test(&ring->inflight) && atomic_read(&blkif->drain)) { inflight 1067 drivers/block/xen-blkback/blkback.c if (atomic_read(&ring->inflight) == 0) inflight 1360 drivers/block/xen-blkback/blkback.c atomic_inc(&ring->inflight); inflight 259 drivers/block/xen-blkback/common.h atomic_t inflight; inflight 267 drivers/block/xen-blkback/xenbus.c if (atomic_read(&ring->inflight) > 0) { inflight 443 drivers/crypto/cavium/cpt/cpt_hw_types.h u64 inflight:8; inflight 445 drivers/crypto/cavium/cpt/cpt_hw_types.h u64 inflight:8; inflight 390 drivers/crypto/cavium/cpt/cptvf_main.c vqx_inprg.s.inflight = val; inflight 205 drivers/crypto/chelsio/chcr_algo.c atomic_inc(&dev->inflight); inflight 211 drivers/crypto/chelsio/chcr_algo.c atomic_dec(&dev->inflight); inflight 59 drivers/crypto/chelsio/chcr_core.c if (atomic_read(&dev->inflight)) { inflight 63 drivers/crypto/chelsio/chcr_core.c atomic_read(&dev->inflight)); inflight 68 drivers/crypto/chelsio/chcr_core.c atomic_read(&dev->inflight)); inflight 106 drivers/crypto/chelsio/chcr_core.c atomic_set(&dev->inflight, 0); inflight 125 drivers/crypto/chelsio/chcr_core.c atomic_set(&dev->inflight, 0); inflight 249 drivers/crypto/chelsio/chcr_core.c if (atomic_read(&dev->inflight) != 0) { inflight 146 drivers/crypto/chelsio/chcr_core.h atomic_t inflight; inflight 42 drivers/gpu/drm/i915/gt/intel_context_types.h struct intel_engine_cs *inflight; inflight 43 drivers/gpu/drm/i915/gt/intel_context_types.h #define intel_context_inflight(ce) ptr_mask_bits((ce)->inflight, 2) inflight 44 drivers/gpu/drm/i915/gt/intel_context_types.h #define intel_context_inflight_count(ce) ptr_unmask_bits((ce)->inflight, 2) inflight 134 drivers/gpu/drm/i915/gt/intel_engine.h GEM_BUG_ON(execlists->active - execlists->inflight > inflight 486 drivers/gpu/drm/i915/gt/intel_engine_cs.c memset(execlists->inflight, 0, sizeof(execlists->inflight)); inflight 213 drivers/gpu/drm/i915/gt/intel_engine_types.h struct i915_request *inflight[EXECLIST_MAX_PORTS + 1 /* sentinel */]; inflight 582 drivers/gpu/drm/i915/gt/intel_lrc.c old = READ_ONCE(ce->inflight); inflight 585 drivers/gpu/drm/i915/gt/intel_lrc.c WRITE_ONCE(ce->inflight, __execlists_schedule_in(rq)); inflight 588 drivers/gpu/drm/i915/gt/intel_lrc.c } while (!try_cmpxchg(&ce->inflight, &old, ptr_inc(old))); inflight 636 drivers/gpu/drm/i915/gt/intel_lrc.c old = READ_ONCE(ce->inflight); inflight 639 drivers/gpu/drm/i915/gt/intel_lrc.c while (!try_cmpxchg(&ce->inflight, &old, cur)); inflight 888 drivers/gpu/drm/i915/gt/intel_lrc.c const struct intel_engine_cs *inflight; inflight 902 drivers/gpu/drm/i915/gt/intel_lrc.c inflight = intel_context_inflight(&ve->context); inflight 903 drivers/gpu/drm/i915/gt/intel_lrc.c if (inflight && inflight != engine) inflight 1220 drivers/gpu/drm/i915/gt/intel_lrc.c GEM_BUG_ON(READ_ONCE(ve->context.inflight)); inflight 1383 drivers/gpu/drm/i915/gt/intel_lrc.c memset(execlists->inflight, 0, sizeof(execlists->inflight)); inflight 1566 drivers/gpu/drm/i915/gt/intel_lrc.c memcpy(execlists->inflight, inflight 1593 drivers/gpu/drm/i915/gt/intel_lrc.c GEM_BUG_ON(execlists->active - execlists->inflight > inflight 3426 drivers/gpu/drm/i915/gt/intel_lrc.c GEM_BUG_ON(ve->context.inflight); inflight 540 drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c struct i915_request **first = execlists->inflight; inflight 573 drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c port - execlists->inflight); inflight 590 drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c *port = schedule_in(last, port - execlists->inflight); inflight 594 drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c execlists->active = execlists->inflight; inflight 606 drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c for (port = execlists->inflight; (rq = *port); port++) { inflight 612 drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c if (port != execlists->inflight) { inflight 613 drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c int idx = port - execlists->inflight; inflight 614 drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c int rem = ARRAY_SIZE(execlists->inflight) - idx; inflight 615 drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c memmove(execlists->inflight, port, rem * sizeof(*port)); inflight 651 drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c memset(execlists->inflight, 0, sizeof(execlists->inflight)); inflight 1135 drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c BUILD_BUG_ON(ARRAY_SIZE(engine->execlists.inflight) * inflight 196 drivers/gpu/drm/i915/i915_scheduler.c const struct i915_request *inflight; inflight 208 drivers/gpu/drm/i915/i915_scheduler.c inflight = execlists_active(&engine->execlists); inflight 209 drivers/gpu/drm/i915/i915_scheduler.c if (!inflight) inflight 219 drivers/gpu/drm/i915/i915_scheduler.c if (inflight->hw_context == rq->hw_context) inflight 223 drivers/gpu/drm/i915/i915_scheduler.c if (need_preempt(prio, rq_prio(inflight))) inflight 31 drivers/iommu/amd_iommu_v2.c atomic_t inflight; inflight 418 drivers/iommu/amd_iommu_v2.c if (atomic_dec_and_test(&pasid_state->pri[tag].inflight) && inflight 561 drivers/iommu/amd_iommu_v2.c atomic_inc(&pasid_state->pri[tag].inflight); inflight 3561 drivers/scsi/scsi_transport_fc.c int err = 0, inflight = 0; inflight 3566 drivers/scsi/scsi_transport_fc.c inflight = bsg_job_get(job); inflight 3568 drivers/scsi/scsi_transport_fc.c if (inflight && i->f->bsg_timeout) { inflight 3580 drivers/scsi/scsi_transport_fc.c if (inflight) inflight 114 drivers/vhost/scsi.c struct vhost_scsi_inflight *inflight; inflight 233 drivers/vhost/scsi.c struct vhost_scsi_inflight *inflight; inflight 235 drivers/vhost/scsi.c inflight = container_of(kref, struct vhost_scsi_inflight, kref); inflight 236 drivers/vhost/scsi.c complete(&inflight->comp); inflight 269 drivers/vhost/scsi.c struct vhost_scsi_inflight *inflight; inflight 273 drivers/vhost/scsi.c inflight = &svq->inflights[svq->inflight_idx]; inflight 274 drivers/vhost/scsi.c kref_get(&inflight->kref); inflight 276 drivers/vhost/scsi.c return inflight; inflight 279 drivers/vhost/scsi.c static void vhost_scsi_put_inflight(struct vhost_scsi_inflight *inflight) inflight 281 drivers/vhost/scsi.c kref_put(&inflight->kref, vhost_scsi_done_inflight); inflight 339 drivers/vhost/scsi.c vhost_scsi_put_inflight(tv_cmd->inflight); inflight 609 drivers/vhost/scsi.c cmd->inflight = vhost_scsi_get_inflight(vq); inflight 2750 fs/io_uring.c unsigned inflight; inflight 2759 fs/io_uring.c timeout = inflight = 0; inflight 2764 fs/io_uring.c if (inflight) { inflight 2780 fs/io_uring.c inflight = 0; inflight 2787 fs/io_uring.c nr_events = inflight; inflight 2790 fs/io_uring.c inflight -= nr_events; inflight 2791 fs/io_uring.c if (!inflight) inflight 2814 fs/io_uring.c if (inflight || !time_after(jiffies, timeout)) { inflight 2856 fs/io_uring.c inflight += io_submit_sqes(ctx, to_submit, cur_mm != NULL, inflight 403 include/linux/genhd.h unsigned int inflight[2]); inflight 61 include/net/af_unix.h atomic_long_t inflight; inflight 16 include/trace/events/page_pool.h s32 inflight, u32 hold, u32 release), inflight 18 include/trace/events/page_pool.h TP_ARGS(pool, inflight, hold, release), inflight 22 include/trace/events/page_pool.h __field(s32, inflight) inflight 29 include/trace/events/page_pool.h __entry->inflight = inflight; inflight 35 include/trace/events/page_pool.h __entry->pool, __entry->inflight, __entry->hold, __entry->release) inflight 132 include/trace/events/wbt.h int step, unsigned int inflight), inflight 134 include/trace/events/wbt.h TP_ARGS(bdi, status, step, inflight), inflight 140 include/trace/events/wbt.h __field(unsigned int, inflight) inflight 148 include/trace/events/wbt.h __entry->inflight = inflight; inflight 152 include/trace/events/wbt.h __entry->status, __entry->step, __entry->inflight) inflight 747 kernel/rcu/rcutorture.c int inflight; inflight 756 kernel/rcu/rcutorture.c smp_store_release(&rbip->inflight, 0); inflight 802 kernel/rcu/rcutorture.c struct rcu_boost_inflight rbi = { .inflight = 0 }; inflight 846 kernel/rcu/rcutorture.c if (!smp_load_acquire(&rbi.inflight)) { inflight 848 kernel/rcu/rcutorture.c smp_store_release(&rbi.inflight, 1); inflight 866 kernel/rcu/rcutorture.c if (!failed && smp_load_acquire(&rbi.inflight)) inflight 892 kernel/rcu/rcutorture.c while (!kthread_should_stop() || smp_load_acquire(&rbi.inflight)) { inflight 64 net/atm/pppoatm.c atomic_t inflight; inflight 137 net/atm/pppoatm.c atomic_dec(&pvcc->inflight); inflight 242 net/atm/pppoatm.c atomic_inc_not_zero(&pvcc->inflight)) inflight 272 net/atm/pppoatm.c atomic_inc_not_zero(&pvcc->inflight)) inflight 408 net/atm/pppoatm.c atomic_set(&pvcc->inflight, NONE_INFLIGHT); inflight 1196 net/core/net-sysfs.c __ATTR(inflight, 0444, bql_show_inflight, NULL); inflight 199 net/core/page_pool.c s32 inflight; inflight 201 net/core/page_pool.c inflight = _distance(hold_cnt, release_cnt); inflight 203 net/core/page_pool.c trace_page_pool_inflight(pool, inflight, hold_cnt, release_cnt); inflight 204 net/core/page_pool.c WARN(inflight < 0, "Negative(%d) inflight packet-pages", inflight); inflight 206 net/core/page_pool.c return inflight; inflight 373 net/core/page_pool.c int inflight; inflight 376 net/core/page_pool.c inflight = page_pool_inflight(pool); inflight 377 net/core/page_pool.c if (!inflight) inflight 380 net/core/page_pool.c return inflight; inflight 387 net/core/page_pool.c int inflight; inflight 389 net/core/page_pool.c inflight = page_pool_release(pool); inflight 390 net/core/page_pool.c if (!inflight) inflight 398 net/core/page_pool.c __func__, inflight, sec); inflight 412 net/ipv4/tcp_bbr.c u32 inflight; inflight 414 net/ipv4/tcp_bbr.c inflight = bbr_bdp(sk, bw, gain); inflight 415 net/ipv4/tcp_bbr.c inflight = bbr_quantization_budget(sk, inflight); inflight 417 net/ipv4/tcp_bbr.c return inflight; inflight 559 net/ipv4/tcp_bbr.c u32 inflight, bw; inflight 567 net/ipv4/tcp_bbr.c inflight = bbr_packets_in_net_at_edt(sk, rs->prior_in_flight); inflight 578 net/ipv4/tcp_bbr.c inflight >= bbr_inflight(sk, bw, bbr->pacing_gain)); inflight 585 net/ipv4/tcp_bbr.c inflight <= bbr_inflight(sk, bw, BBR_UNIT); inflight 651 net/sctp/output.c size_t datasize, rwnd, inflight, flight_size; inflight 670 net/sctp/output.c inflight = q->outstanding_bytes; inflight 675 net/sctp/output.c if (datasize > rwnd && inflight > 0) inflight 703 net/sctp/output.c if ((sctp_sk(asoc->base.sk)->nodelay || inflight == 0) && inflight 787 net/unix/af_unix.c atomic_long_set(&u->inflight, 0); inflight 169 net/unix/garbage.c atomic_long_dec(&usk->inflight); inflight 174 net/unix/garbage.c atomic_long_inc(&usk->inflight); inflight 179 net/unix/garbage.c atomic_long_inc(&u->inflight); inflight 237 net/unix/garbage.c inflight_refs = atomic_long_read(&u->inflight); inflight 268 net/unix/garbage.c if (atomic_long_read(&u->inflight) > 0) { inflight 56 net/unix/scm.c if (atomic_long_inc_return(&u->inflight) == 1) { inflight 77 net/unix/scm.c BUG_ON(!atomic_long_read(&u->inflight)); inflight 80 net/unix/scm.c if (atomic_long_dec_and_test(&u->inflight)) inflight 80 tools/io_uring/io_uring-bench.c int inflight; inflight 272 tools/io_uring/io_uring-bench.c s->inflight -= reaped; inflight 292 tools/io_uring/io_uring-bench.c if (!prepped && s->inflight < DEPTH) { inflight 293 tools/io_uring/io_uring-bench.c to_prep = min(DEPTH - s->inflight, BATCH_SUBMIT); inflight 296 tools/io_uring/io_uring-bench.c s->inflight += prepped; inflight 300 tools/io_uring/io_uring-bench.c if (to_submit && (s->inflight + to_submit <= DEPTH)) inflight 303 tools/io_uring/io_uring-bench.c to_wait = min(s->inflight + to_submit, BATCH_COMPLETE); inflight 341 tools/io_uring/io_uring-bench.c if (s->inflight) inflight 581 tools/io_uring/io_uring-bench.c this_done - done, rpc, ipc, s->inflight,