requests 106 arch/powerpc/kvm/trace.h __field( __u32, requests ) requests 111 arch/powerpc/kvm/trace.h __entry->requests = vcpu->requests; requests 115 arch/powerpc/kvm/trace.h __entry->cpu_nr, __entry->requests) requests 1098 arch/powerpc/perf/hv-24x7.c req = request_buffer->requests; requests 1146 arch/powerpc/perf/hv-24x7.c req = (void *) request_buffer->requests + i * req_size; requests 65 arch/powerpc/perf/hv-24x7.h struct hv_24x7_request requests[]; requests 86 drivers/acpi/acpica/dbstats.c "%8.2X %8.2X %8.2X %8.2X\n", list->requests, list->hits, requests 87 drivers/acpi/acpica/dbstats.c list->requests - list->hits, list->object_size); requests 224 drivers/acpi/acpica/utcache.c ACPI_MEM_TRACKING(cache->requests++); requests 50 drivers/base/devtmpfs.c } *requests; requests 123 drivers/base/devtmpfs.c req.next = requests; requests 124 drivers/base/devtmpfs.c requests = &req; requests 153 drivers/base/devtmpfs.c req.next = requests; requests 154 drivers/base/devtmpfs.c requests = &req; requests 405 drivers/base/devtmpfs.c while (requests) { requests 406 drivers/base/devtmpfs.c struct req *req = requests; requests 407 drivers/base/devtmpfs.c requests = NULL; requests 221 drivers/block/xen-blkfront.c struct list_head requests; requests 2050 drivers/block/xen-blkfront.c list_for_each_entry_safe(req, n, &info->requests, queuelist) { requests 2082 drivers/block/xen-blkfront.c INIT_LIST_HEAD(&info->requests); requests 2107 drivers/block/xen-blkfront.c list_add(&shadow[j].request->queuelist, &info->requests); requests 758 drivers/crypto/inside-secure/safexcel.c int coal = min_t(int, priv->ring[ring].requests, EIP197_MAX_BATCH_SZ); requests 829 drivers/crypto/inside-secure/safexcel.c priv->ring[ring].requests += nreq; requests 1018 drivers/crypto/inside-secure/safexcel.c priv->ring[ring].requests -= handled; requests 1021 drivers/crypto/inside-secure/safexcel.c if (!priv->ring[ring].requests) requests 1504 drivers/crypto/inside-secure/safexcel.c priv->ring[i].requests = 0; requests 633 drivers/crypto/inside-secure/safexcel.h int requests; requests 2160 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c list_for_each_entry(rq, &tl->requests, link) { requests 2168 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c if (&rq->link == &tl->requests) requests 712 drivers/gpu/drm/i915/gt/intel_engine_cs.c INIT_LIST_HEAD(&engine->active.requests); requests 806 drivers/gpu/drm/i915/gt/intel_engine_cs.c GEM_BUG_ON(!list_empty(&engine->active.requests)); requests 1556 drivers/gpu/drm/i915/gt/intel_engine_cs.c list_for_each_entry(request, &engine->active.requests, sched.link) { requests 318 drivers/gpu/drm/i915/gt/intel_engine_types.h struct list_head requests; requests 350 drivers/gpu/drm/i915/gt/intel_lrc.c if (!list_is_last(&rq->sched.link, &engine->active.requests) && requests 484 drivers/gpu/drm/i915/gt/intel_lrc.c &engine->active.requests, requests 1004 drivers/gpu/drm/i915/gt/intel_lrc.c if (list_is_last(&rq->sched.link, &engine->active.requests)) requests 1016 drivers/gpu/drm/i915/gt/intel_lrc.c if (list_is_last(&rq->sched.link, &engine->active.requests)) requests 1162 drivers/gpu/drm/i915/gt/intel_lrc.c &engine->active.requests)) requests 2445 drivers/gpu/drm/i915/gt/intel_lrc.c list = &rq->timeline->requests; requests 2595 drivers/gpu/drm/i915/gt/intel_lrc.c list_for_each_entry(rq, &engine->active.requests, sched.link) requests 3415 drivers/gpu/drm/i915/gt/intel_lrc.c return &ve->base.execlists.default_priolist.requests[0]; requests 3931 drivers/gpu/drm/i915/gt/intel_lrc.c list_for_each_entry(rq, &engine->active.requests, sched.link) { requests 49 drivers/gpu/drm/i915/gt/intel_reset.c list_for_each_entry_continue(rq, &engine->active.requests, sched.link) requests 790 drivers/gpu/drm/i915/gt/intel_ringbuffer.c list_for_each_entry(pos, &engine->active.requests, sched.link) { requests 918 drivers/gpu/drm/i915/gt/intel_ringbuffer.c list_for_each_entry(request, &engine->active.requests, sched.link) { requests 1879 drivers/gpu/drm/i915/gt/intel_ringbuffer.c GEM_BUG_ON(list_empty(&tl->requests)); requests 1880 drivers/gpu/drm/i915/gt/intel_ringbuffer.c list_for_each_entry(target, &tl->requests, link) { requests 1890 drivers/gpu/drm/i915/gt/intel_ringbuffer.c if (GEM_WARN_ON(&target->link == &tl->requests)) requests 258 drivers/gpu/drm/i915/gt/intel_timeline.c INIT_LIST_HEAD(&timeline->requests); requests 284 drivers/gpu/drm/i915/gt/intel_timeline.c GEM_BUG_ON(!list_empty(&timeline->requests)); requests 59 drivers/gpu/drm/i915/gt/intel_timeline_types.h struct list_head requests; requests 226 drivers/gpu/drm/i915/gt/mock_engine.c list_for_each_entry(request, &engine->active.requests, sched.link) { requests 19 drivers/gpu/drm/i915/gt/selftests/mock_timeline.c INIT_LIST_HEAD(&timeline->requests); requests 708 drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c list_for_each_entry(rq, &engine->active.requests, sched.link) { requests 1259 drivers/gpu/drm/i915/i915_gem.c struct i915_request *requests[I915_NUM_ENGINES] = {}; requests 1312 drivers/gpu/drm/i915/i915_gem.c requests[id] = i915_request_get(rq); requests 1324 drivers/gpu/drm/i915/i915_gem.c for (id = 0; id < ARRAY_SIZE(requests); id++) { requests 1329 drivers/gpu/drm/i915/i915_gem.c rq = requests[id]; requests 1380 drivers/gpu/drm/i915/i915_gem.c for (id = 0; id < ARRAY_SIZE(requests); id++) { requests 1384 drivers/gpu/drm/i915/i915_gem.c rq = requests[id]; requests 765 drivers/gpu/drm/i915/i915_gpu_error.c &ee->requests[j], requests 942 drivers/gpu/drm/i915/i915_gpu_error.c kfree(ee->requests); requests 1197 drivers/gpu/drm/i915/i915_gpu_error.c list_for_each_entry_from(request, &engine->active.requests, sched.link) requests 1202 drivers/gpu/drm/i915/i915_gpu_error.c ee->requests = kcalloc(count, sizeof(*ee->requests), ATOMIC_MAYFAIL); requests 1203 drivers/gpu/drm/i915/i915_gpu_error.c if (!ee->requests) requests 1211 drivers/gpu/drm/i915/i915_gpu_error.c &engine->active.requests, sched.link) { requests 1231 drivers/gpu/drm/i915/i915_gpu_error.c record_request(request, &ee->requests[count++]); requests 152 drivers/gpu/drm/i915/i915_gpu_error.h } *requests, execlist[EXECLIST_MAX_PORTS]; requests 46 drivers/gpu/drm/i915/i915_priolist_types.h struct list_head requests[I915_PRIORITY_COUNT]; requests 243 drivers/gpu/drm/i915/i915_request.c GEM_BUG_ON(!list_is_first(&rq->link, &rq->timeline->requests)); requests 331 drivers/gpu/drm/i915/i915_request.c tmp = list_first_entry(&tl->requests, typeof(*tmp), link); requests 446 drivers/gpu/drm/i915/i915_request.c list_move_tail(&request->sched.link, &engine->active.requests); requests 598 drivers/gpu/drm/i915/i915_request.c list_for_each_entry_safe(rq, rn, &tl->requests, link) requests 608 drivers/gpu/drm/i915/i915_request.c if (list_empty(&tl->requests)) requests 615 drivers/gpu/drm/i915/i915_request.c rq = list_first_entry(&tl->requests, typeof(*rq), link); requests 624 drivers/gpu/drm/i915/i915_request.c rq = list_last_entry(&tl->requests, typeof(*rq), link); requests 778 drivers/gpu/drm/i915/i915_request.c rq = list_first_entry(&tl->requests, typeof(*rq), link); requests 779 drivers/gpu/drm/i915/i915_request.c if (!list_is_last(&rq->link, &tl->requests)) requests 801 drivers/gpu/drm/i915/i915_request.c if (list_is_first(&signal->link, &signal->timeline->requests)) requests 1169 drivers/gpu/drm/i915/i915_request.c list_add_tail(&rq->link, &timeline->requests); requests 62 drivers/gpu/drm/i915/i915_scheduler.c for (i = 0; i < ARRAY_SIZE(p->requests); i++) { requests 63 drivers/gpu/drm/i915/i915_scheduler.c if (list_empty(&p->requests[i])) requests 128 drivers/gpu/drm/i915/i915_scheduler.c for (i = 0; i < ARRAY_SIZE(p->requests); i++) requests 129 drivers/gpu/drm/i915/i915_scheduler.c INIT_LIST_HEAD(&p->requests[i]); requests 136 drivers/gpu/drm/i915/i915_scheduler.c return &p->requests[idx]; requests 17 drivers/gpu/drm/i915/i915_scheduler.h for (idx = 0; idx < ARRAY_SIZE((plist)->requests); idx++) \ requests 18 drivers/gpu/drm/i915/i915_scheduler.h list_for_each_entry(it, &(plist)->requests[idx], sched.link) requests 25 drivers/gpu/drm/i915/i915_scheduler.h &(plist)->requests[idx], \ requests 18 drivers/gpu/drm/i915/selftests/i915_live_selftests.h selftest(requests, i915_request_live_selftests) requests 19 drivers/gpu/drm/i915/selftests/i915_mock_selftests.h selftest(requests, i915_request_mock_selftests) requests 289 drivers/gpu/drm/i915/selftests/i915_request.c struct i915_request **requests; requests 303 drivers/gpu/drm/i915/selftests/i915_request.c requests = kmalloc_array(total, sizeof(*requests), GFP_KERNEL); requests 304 drivers/gpu/drm/i915/selftests/i915_request.c if (!requests) requests 357 drivers/gpu/drm/i915/selftests/i915_request.c requests[n] = i915_request_get(rq); requests 381 drivers/gpu/drm/i915/selftests/i915_request.c struct i915_request *rq = requests[count - 1]; requests 396 drivers/gpu/drm/i915/selftests/i915_request.c struct i915_request *rq = requests[n]; requests 425 drivers/gpu/drm/i915/selftests/i915_request.c kfree(requests); requests 166 drivers/iio/adc/twl4030-madc.c struct twl4030_madc_request requests[TWL4030_MADC_NUM_METHODS]; requests 498 drivers/iio/adc/twl4030-madc.c madc->requests[i].result_pending = 1; requests 501 drivers/iio/adc/twl4030-madc.c r = &madc->requests[i]; requests 523 drivers/iio/adc/twl4030-madc.c r = &madc->requests[i]; requests 624 drivers/iio/adc/twl4030-madc.c if (twl4030_madc->requests[req->method].active) { requests 655 drivers/iio/adc/twl4030-madc.c twl4030_madc->requests[req->method].active = 1; requests 659 drivers/iio/adc/twl4030-madc.c twl4030_madc->requests[req->method].active = 0; requests 664 drivers/iio/adc/twl4030-madc.c twl4030_madc->requests[req->method].active = 0; requests 66 drivers/input/misc/uinput.c struct uinput_request *requests[UINPUT_NUM_REQUESTS]; requests 104 drivers/input/misc/uinput.c if (!udev->requests[id]) { requests 106 drivers/input/misc/uinput.c udev->requests[id] = request; requests 123 drivers/input/misc/uinput.c return udev->requests[id]; requests 139 drivers/input/misc/uinput.c udev->requests[id] = NULL; requests 209 drivers/input/misc/uinput.c request = udev->requests[i]; requests 44 drivers/iommu/virtio-iommu.c struct list_head requests; requests 166 drivers/iommu/virtio-iommu.c while (!list_empty(&viommu->requests)) { requests 257 drivers/iommu/virtio-iommu.c list_add_tail(&req->list, &viommu->requests); requests 1028 drivers/iommu/virtio-iommu.c INIT_LIST_HEAD(&viommu->requests); requests 276 drivers/mailbox/bcm-flexrm-mailbox.c struct brcm_message *requests[RING_MAX_REQ_COUNT]; requests 1008 drivers/mailbox/bcm-flexrm-mailbox.c ring->requests[reqid] = msg; requests 1013 drivers/mailbox/bcm-flexrm-mailbox.c ring->requests[reqid] = NULL; requests 1075 drivers/mailbox/bcm-flexrm-mailbox.c ring->requests[reqid] = NULL; requests 1133 drivers/mailbox/bcm-flexrm-mailbox.c msg = ring->requests[reqid]; requests 1142 drivers/mailbox/bcm-flexrm-mailbox.c ring->requests[reqid] = NULL; requests 1408 drivers/mailbox/bcm-flexrm-mailbox.c msg = ring->requests[reqid]; requests 1413 drivers/mailbox/bcm-flexrm-mailbox.c ring->requests[reqid] = NULL; requests 1569 drivers/mailbox/bcm-flexrm-mailbox.c memset(ring->requests, 0, sizeof(ring->requests)); requests 31 drivers/media/pci/tw686x/tw686x-audio.c void tw686x_audio_irq(struct tw686x_dev *dev, unsigned long requests, requests 37 drivers/media/pci/tw686x/tw686x-audio.c for_each_set_bit(ch, &requests, max_channels(dev)) { requests 1087 drivers/media/pci/tw686x/tw686x-video.c void tw686x_video_irq(struct tw686x_dev *dev, unsigned long requests, requests 1095 drivers/media/pci/tw686x/tw686x-video.c for_each_set_bit(ch, &requests, max_channels(dev)) { requests 172 drivers/media/pci/tw686x/tw686x.h void tw686x_video_irq(struct tw686x_dev *dev, unsigned long requests, requests 178 drivers/media/pci/tw686x/tw686x.h void tw686x_audio_irq(struct tw686x_dev *dev, unsigned long requests, requests 2134 drivers/media/v4l2-core/v4l2-ctrls.c INIT_LIST_HEAD(&hdl->requests); requests 2157 drivers/media/v4l2-core/v4l2-ctrls.c if (!hdl->req_obj.req && !list_empty(&hdl->requests)) { requests 2160 drivers/media/v4l2-core/v4l2-ctrls.c list_for_each_entry_safe(req, next_req, &hdl->requests, requests) { requests 3188 drivers/media/v4l2-core/v4l2-ctrls.c list_del_init(&hdl->requests); requests 3249 drivers/media/v4l2-core/v4l2-ctrls.c list_add_tail(&hdl->requests, &from->requests); requests 727 drivers/mfd/db8500-prcmu.c static int requests[2]; requests 739 drivers/mfd/db8500-prcmu.c if (!div && !requests[clkout]) requests 773 drivers/mfd/db8500-prcmu.c requests[clkout] += (div ? 1 : -1); requests 1056 drivers/mfd/db8500-prcmu.c static unsigned int requests; requests 1061 drivers/mfd/db8500-prcmu.c if (0 != requests++) requests 1065 drivers/mfd/db8500-prcmu.c if (requests == 0) { requests 1068 drivers/mfd/db8500-prcmu.c } else if (1 != requests--) { requests 584 drivers/net/ethernet/broadcom/bnxt/bnxt_dcb.c pfc->requests[i] = le64_to_cpu(*(stats + tx_off)); requests 281 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_dcb.c u64 requests[HNAE3_MAX_TC], indications[HNAE3_MAX_TC]; requests 300 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_dcb.c ret = hclge_pfc_tx_stats_get(hdev, requests); requests 309 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_dcb.c pfc->requests[i] = requests[i]; requests 770 drivers/net/ethernet/intel/fm10k/fm10k_pci.c struct list_head *requests; requests 777 drivers/net/ethernet/intel/fm10k/fm10k_pci.c requests = &interface->macvlan_requests; requests 782 drivers/net/ethernet/intel/fm10k/fm10k_pci.c item = list_first_entry_or_null(requests, requests 808 drivers/net/ethernet/intel/fm10k/fm10k_pci.c list_add(&item->list, requests); requests 89 drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c pfc->requests[i] = pf->stats.priority_xoff_tx[i]; requests 578 drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c pfc->requests[i] = adapter->stats.pxoffrxc[i]; requests 342 drivers/net/ethernet/mellanox/mlx5/core/en_dcbnl.c pfc->requests[i] = PPORT_PER_PRIO_GET(pstats, i, tx_pause); requests 545 drivers/net/ethernet/mellanox/mlxsw/spectrum_dcb.c my_pfc->requests[prio] = mlxsw_reg_ppcnt_tx_pause_get(ppcnt_pl); requests 738 drivers/perf/qcom_l2_pmu.c L2CACHE_EVENT_ATTR(total-requests, L2_EVENT_TOTAL_REQUESTS), requests 629 drivers/s390/scsi/zfcp_sysfs.c ZFCP_SHOST_ATTR(requests, "%llu %llu %llu\n", requests 1419 drivers/usb/dwc2/hcd.c chan->requests++; requests 1507 drivers/usb/dwc2/hcd.c chan->requests++; requests 1567 drivers/usb/dwc2/hcd.c chan->requests++; requests 1585 drivers/usb/dwc2/hcd.c chan->requests++; requests 2622 drivers/usb/dwc2/hcd.c chan->requests = 0; requests 2957 drivers/usb/dwc2/hcd.c qh->channel->requests == qh->channel->multi_count) { requests 3882 drivers/usb/dwc2/hcd.c dev_dbg(hsotg->dev, " requests: %d\n", chan->requests); requests 157 drivers/usb/dwc2/hcd.h u8 requests; requests 568 drivers/usb/gadget/udc/net2280.c td = dma_pool_alloc(ep->dev->requests, gfp_flags, requests 596 drivers/usb/gadget/udc/net2280.c dma_pool_free(ep->dev->requests, req->td, req->td_dma); requests 3581 drivers/usb/gadget/udc/net2280.c if (dev->requests) { requests 3586 drivers/usb/gadget/udc/net2280.c dma_pool_free(dev->requests, dev->ep[i].dummy, requests 3589 drivers/usb/gadget/udc/net2280.c dma_pool_destroy(dev->requests); requests 3727 drivers/usb/gadget/udc/net2280.c dev->requests = dma_pool_create("requests", &pdev->dev, requests 3731 drivers/usb/gadget/udc/net2280.c if (!dev->requests) { requests 3739 drivers/usb/gadget/udc/net2280.c td = dma_pool_alloc(dev->requests, GFP_KERNEL, requests 183 drivers/usb/gadget/udc/net2280.h struct dma_pool *requests; requests 1230 include/acpi/actypes.h u32 requests; requests 152 include/linux/kvm_host.h BUILD_BUG_ON((unsigned)(nr) >= (FIELD_SIZEOF(struct kvm_vcpu, requests) * 8) - KVM_REQUEST_ARCH_BASE); \ requests 272 include/linux/kvm_host.h u64 requests; requests 1207 include/linux/kvm_host.h set_bit(req & KVM_REQUEST_MASK, (void *)&vcpu->requests); requests 1212 include/linux/kvm_host.h return READ_ONCE(vcpu->requests); requests 1217 include/linux/kvm_host.h return test_bit(req & KVM_REQUEST_MASK, (void *)&vcpu->requests); requests 1222 include/linux/kvm_host.h clear_bit(req & KVM_REQUEST_MASK, (void *)&vcpu->requests); requests 343 include/media/v4l2-ctrls.h struct list_head requests; requests 162 include/uapi/linux/dcbnl.h __u64 requests[IEEE_8021QAZ_MAX_TCS]; requests 300 net/ncsi/internal.h struct ncsi_request requests[256]; /* Request table */ requests 349 net/ncsi/ncsi-manage.c int i, limit = ARRAY_SIZE(ndp->requests); requests 355 net/ncsi/ncsi-manage.c if (ndp->requests[i].used) requests 358 net/ncsi/ncsi-manage.c nr = &ndp->requests[i]; requests 367 net/ncsi/ncsi-manage.c if (ndp->requests[i].used) requests 370 net/ncsi/ncsi-manage.c nr = &ndp->requests[i]; requests 1654 net/ncsi/ncsi-manage.c for (i = 0; i < ARRAY_SIZE(ndp->requests); i++) { requests 1655 net/ncsi/ncsi-manage.c ndp->requests[i].id = i; requests 1656 net/ncsi/ncsi-manage.c ndp->requests[i].ndp = ndp; requests 1657 net/ncsi/ncsi-manage.c timer_setup(&ndp->requests[i].timer, ncsi_request_timeout, 0); requests 1147 net/ncsi/ncsi-rsp.c nr = &ndp->requests[hdr->id]; requests 259 tools/testing/nvdimm/test/iomap.c list_for_each_entry(req, &nfit_res->requests, list) requests 299 tools/testing/nvdimm/test/iomap.c list_for_each_entry(req, &nfit_res->requests, list) requests 323 tools/testing/nvdimm/test/iomap.c list_add(&req->list, &nfit_res->requests); requests 1449 tools/testing/nvdimm/test/nfit.c INIT_LIST_HEAD(&nfit_res->requests); requests 19 tools/testing/nvdimm/test/nfit_test.h struct list_head requests;