/linux-4.4.14/net/rose/ |
D | rose_in.c | 107 int queued = 0; in rose_state3_machine() local 169 queued = 1; in rose_state3_machine() 206 return queued; in rose_state3_machine() 266 int queued = 0, frametype, ns, nr, q, d, m; in rose_process_rx_frame() local 275 queued = rose_state1_machine(sk, skb, frametype); in rose_process_rx_frame() 278 queued = rose_state2_machine(sk, skb, frametype); in rose_process_rx_frame() 281 queued = rose_state3_machine(sk, skb, frametype, ns, nr, q, d, m); in rose_process_rx_frame() 284 queued = rose_state4_machine(sk, skb, frametype); in rose_process_rx_frame() 287 queued = rose_state5_machine(sk, skb, frametype); in rose_process_rx_frame() 293 return queued; in rose_process_rx_frame()
|
/linux-4.4.14/net/x25/ |
D | x25_in.c | 206 int queued = 0; in x25_state3_machine() local 273 queued = 1; in x25_state3_machine() 311 queued = !sock_queue_rcv_skb(sk, skb); in x25_state3_machine() 315 queued = 1; in x25_state3_machine() 326 return queued; in x25_state3_machine() 384 int queued = 0, frametype, ns, nr, q, d, m; in x25_process_rx_frame() local 393 queued = x25_state1_machine(sk, skb, frametype); in x25_process_rx_frame() 396 queued = x25_state2_machine(sk, skb, frametype); in x25_process_rx_frame() 399 queued = x25_state3_machine(sk, skb, frametype, ns, nr, q, d, m); in x25_process_rx_frame() 402 queued = x25_state4_machine(sk, skb, frametype); in x25_process_rx_frame() [all …]
|
D | x25_dev.c | 56 int queued = 1; in x25_receive_data() local 61 queued = x25_process_rx_frame(sk, skb); in x25_receive_data() 63 queued = !sk_add_backlog(sk, skb, sk->sk_rcvbuf); in x25_receive_data() 67 return queued; in x25_receive_data()
|
/linux-4.4.14/net/dccp/ |
D | input.c | 49 int queued = 0; in dccp_rcv_close() local 80 queued = 1; in dccp_rcv_close() 90 return queued; in dccp_rcv_close() 95 int queued = 0; in dccp_rcv_closereq() local 105 return queued; in dccp_rcv_closereq() 117 queued = 1; in dccp_rcv_closereq() 124 return queued; in dccp_rcv_closereq() 528 int queued = 0; in dccp_rcv_respond_partopen_state_process() local 565 queued = 1; /* packet was queued in dccp_rcv_respond_partopen_state_process() 571 return queued; in dccp_rcv_respond_partopen_state_process() [all …]
|
/linux-4.4.14/net/netrom/ |
D | nr_in.c | 156 int queued = 0; in nr_state3_machine() local 229 queued = 1; in nr_state3_machine() 276 return queued; in nr_state3_machine() 283 int queued = 0, frametype; in nr_process_rx_frame() local 292 queued = nr_state1_machine(sk, skb, frametype); in nr_process_rx_frame() 295 queued = nr_state2_machine(sk, skb, frametype); in nr_process_rx_frame() 298 queued = nr_state3_machine(sk, skb, frametype); in nr_process_rx_frame() 304 return queued; in nr_process_rx_frame()
|
/linux-4.4.14/net/ax25/ |
D | ax25_std_in.c | 146 int queued = 0; in ax25_std_state3_machine() local 228 queued = ax25_rx_iframe(ax25, skb); in ax25_std_state3_machine() 261 return queued; in ax25_std_state3_machine() 271 int queued = 0; in ax25_std_state4_machine() local 383 queued = ax25_rx_iframe(ax25, skb); in ax25_std_state4_machine() 416 return queued; in ax25_std_state4_machine() 424 int queued = 0, frametype, ns, nr, pf; in ax25_std_frame_in() local 430 queued = ax25_std_state1_machine(ax25, skb, frametype, pf, type); in ax25_std_frame_in() 433 queued = ax25_std_state2_machine(ax25, skb, frametype, pf, type); in ax25_std_frame_in() 436 queued = ax25_std_state3_machine(ax25, skb, frametype, ns, nr, pf, type); in ax25_std_frame_in() [all …]
|
D | ax25_ds_in.c | 150 int queued = 0; in ax25_ds_state3_machine() local 243 queued = ax25_rx_iframe(ax25, skb); in ax25_ds_state3_machine() 276 return queued; in ax25_ds_state3_machine() 284 int queued = 0, frametype, ns, nr, pf; in ax25_ds_frame_in() local 290 queued = ax25_ds_state1_machine(ax25, skb, frametype, pf, type); in ax25_ds_frame_in() 293 queued = ax25_ds_state2_machine(ax25, skb, frametype, pf, type); in ax25_ds_frame_in() 296 queued = ax25_ds_state3_machine(ax25, skb, frametype, ns, nr, pf, type); in ax25_ds_frame_in() 300 return queued; in ax25_ds_frame_in()
|
D | ax25_in.c | 106 int queued = 0; in ax25_rx_iframe() local 148 queued = 1; in ax25_rx_iframe() 154 return queued; in ax25_rx_iframe() 162 int queued = 0; in ax25_process_rx_frame() local 170 queued = ax25_std_frame_in(ax25, skb, type); in ax25_process_rx_frame() 176 queued = ax25_ds_frame_in(ax25, skb, type); in ax25_process_rx_frame() 178 queued = ax25_std_frame_in(ax25, skb, type); in ax25_process_rx_frame() 183 return queued; in ax25_process_rx_frame()
|
/linux-4.4.14/drivers/gpu/drm/ |
D | drm_flip_work.c | 60 list_add_tail(&task->node, &work->queued); in drm_flip_work_queue_task() 104 list_splice_tail(&work->queued, &work->commited); in drm_flip_work_commit() 105 INIT_LIST_HEAD(&work->queued); in drm_flip_work_commit() 148 INIT_LIST_HEAD(&work->queued); in drm_flip_work_init() 165 WARN_ON(!list_empty(&work->queued) || !list_empty(&work->commited)); in drm_flip_work_cleanup()
|
/linux-4.4.14/block/ |
D | blk-throttle.c | 65 struct list_head queued[2]; /* throtl_qnode [READ/WRITE] */ member 251 struct list_head *queued) in throtl_qnode_add_bio() argument 255 list_add_tail(&qn->node, queued); in throtl_qnode_add_bio() 264 static struct bio *throtl_peek_queued(struct list_head *queued) in throtl_peek_queued() argument 266 struct throtl_qnode *qn = list_first_entry(queued, struct throtl_qnode, node); in throtl_peek_queued() 269 if (list_empty(queued)) in throtl_peek_queued() 291 static struct bio *throtl_pop_queued(struct list_head *queued, in throtl_pop_queued() argument 294 struct throtl_qnode *qn = list_first_entry(queued, struct throtl_qnode, node); in throtl_pop_queued() 297 if (list_empty(queued)) in throtl_pop_queued() 310 list_move_tail(&qn->node, queued); in throtl_pop_queued() [all …]
|
D | blk-flush.c | 172 bool queued = false, kicked; in blk_flush_complete_seq() local 193 queued = blk_flush_queue_rq(rq, true); in blk_flush_complete_seq() 217 return kicked | queued; in blk_flush_complete_seq() 224 bool queued = false; in flush_end_io() local 253 queued |= blk_flush_complete_seq(rq, fq, seq, error); in flush_end_io() 267 if (queued || fq->flush_queue_delayed) { in flush_end_io()
|
D | cfq-iosched.c | 122 int queued[2]; member 187 struct blkg_rwstat queued; member 537 if (blkg_rwstat_total(&stats->queued)) in cfqg_stats_set_start_empty_time() 581 blkg_rwstat_total(&stats->queued)); in cfqg_stats_update_avg_queue_size() 665 blkg_rwstat_add(&cfqg->stats.queued, rw, 1); in cfqg_stats_update_io_add() 681 blkg_rwstat_add(&cfqg->stats.queued, rw, -1); in cfqg_stats_update_io_remove() 1533 blkg_rwstat_exit(&stats->queued); in cfqg_stats_exit() 1551 blkg_rwstat_init(&stats->queued, gfp) || in cfqg_stats_init() 2061 .private = offsetof(struct cfq_group, stats.queued), 2102 .private = offsetof(struct cfq_group, stats.queued), [all …]
|
D | blk-mq.c | 731 int queued; in __blk_mq_run_hw_queue() local 765 queued = 0; in __blk_mq_run_hw_queue() 780 queued++; in __blk_mq_run_hw_queue() 805 if (!queued) in __blk_mq_run_hw_queue() 807 else if (queued < (1 << (BLK_MQ_MAX_DISPATCH_ORDER - 1))) in __blk_mq_run_hw_queue() 808 hctx->dispatched[ilog2(queued) + 1]++; in __blk_mq_run_hw_queue() 1195 hctx->queued++; in blk_mq_map_request()
|
D | blk-mq-sysfs.c | 185 return sprintf(page, "%lu\n", hctx->queued); in blk_mq_hw_sysfs_queued_show()
|
/linux-4.4.14/virt/kvm/ |
D | async_pf.c | 138 vcpu->async_pf.queued = 0; in kvm_clear_async_pf_completion_queue() 157 vcpu->async_pf.queued--; in kvm_check_async_pf_completion() 167 if (vcpu->async_pf.queued >= ASYNC_PF_PER_VCPU) in kvm_setup_async_pf() 199 vcpu->async_pf.queued++; in kvm_setup_async_pf() 227 vcpu->async_pf.queued++; in kvm_async_pf_wakeup_all()
|
/linux-4.4.14/fs/xfs/ |
D | xfs_mru_cache.c | 114 unsigned int queued; /* work has been queued */ member 215 if (!mru->queued) { in _xfs_mru_cache_list_insert() 216 mru->queued = 1; in _xfs_mru_cache_list_insert() 291 mru->queued = next; in _xfs_mru_cache_reap() 292 if ((mru->queued > 0)) { in _xfs_mru_cache_reap() 398 if (mru->queued) { in xfs_mru_cache_flush()
|
/linux-4.4.14/Documentation/features/locking/queued-spinlocks/ |
D | arch-support.txt | 2 # Feature name: queued-spinlocks 4 # description: arch supports queued spinlocks
|
/linux-4.4.14/Documentation/features/locking/queued-rwlocks/ |
D | arch-support.txt | 2 # Feature name: queued-rwlocks 4 # description: arch supports queued rwlocks
|
/linux-4.4.14/Documentation/usb/ |
D | ohci.txt | 18 - interrupt transfers can be larger, and can be queued 24 types can be queued. That was also true in "usb-ohci", except for interrupt 26 to overhead in IRQ processing. When interrupt transfers are queued, those
|
D | bulk-streams.txt | 6 queued at once. 17 Once a buffer has been queued to a stream ring, the device is notified (through
|
D | URB.txt | 23 queued the requested action. 129 It immediately returns, either with status 0 (request queued) or some 135 - Too many queued ISO transfers (-EAGAIN) 241 earlier, if you always keep at least one URB queued and your completion
|
D | usbmon.txt | 250 u32 queued; 254 The member "queued" refers to the number of events currently queued in the
|
D | ehci.txt | 67 Transfers of all types can be queued. This means that control transfers
|
/linux-4.4.14/drivers/soc/ti/ |
D | Kconfig | 16 Packets are queued/de-queued by writing/reading descriptor address
|
/linux-4.4.14/Documentation/networking/ |
D | rds.txt | 76 a socket. A message is queued when sendmsg is called, and 101 of queued bytes over the SO_SNDSIZE threshold will return 108 Receives a message that was queued to this socket. The sockets 125 incoming message queued to the socket, or a pending notification, 130 there's room on the send queue (ie the number of bytes queued 141 This allows the application to discard all messages queued to a 226 number of bytes queued equals or exceeds rcvbuf then the socket 269 This is then queued for the individual connection and sent by the 273 the transport to the general code and queued by the general code 298 Dropping a connection while packets are queued will cause queued or
|
D | ipvlan.txt | 40 slave device and packets are switched and queued to the master device to send 48 used before packets are queued on the outbound device. In this mode the slaves
|
D | tcp.txt | 72 one that has been queued already. To add a frame we throw it on the end. Ack 94 Frames are queued for output by tcp_write. We do our best to send the frames
|
D | multiqueue.txt | 45 will be queued to the band associated with the hardware queue.
|
D | netlink_mmap.txt | 191 RX ring only: user-space queued the message for later processing, but 281 /* Frame queued to socket receive queue */
|
D | netdev-FAQ.txt | 114 Q: How can I tell what patches are queued up for backporting to the 142 if it is already queued. If not, then send a mail to netdev, listing
|
D | scaling.txt | 132 and the packet is queued to the tail of that CPU’s backlog queue. At 134 packets have been queued to their backlog queue. The IPI wakes backlog 135 processing on the remote CPU, and any queued packets are then processed
|
D | switchdev.txt | 388 The object created by a driver in "prepare" phase and it is queued up by: 394 cleanup of the queued-up objects.
|
D | timestamping.txt | 132 scheduler. In that case timestamps will be queued onto the error 319 however, the full packet is queued, taking up budget from SO_RCVBUF.
|
D | ppp_generic.txt | 180 fragment if it doesn't have any fragments currently queued up for it 182 the fragment is queued up for the channel to transmit later. This
|
D | altera_tse.txt | 75 intialization. Receive buffers may or may not be queued depending upon the
|
D | stmmac.txt | 59 are not queued so the driver has to scan all the descriptors in the ring during
|
D | ip-sysctl.txt | 91 queued for each unresolved address by other network layers. 97 The maximum number of packets which may be queued for each 707 result in a large amount of packets queued in qdisc/device
|
/linux-4.4.14/Documentation/ABI/testing/ |
D | sysfs-class-net-queues | 67 Indicates the current limit of bytes allowed to be queued 77 queued on this network device transmit queue. See 86 queued on this network device transmit queue. Default value is
|
D | sysfs-ata | 99 queued: Drive supports queued DSM TRIM 100 forced_unqueued: Drive's queued DSM support is known to be
|
D | sysfs-block-dm | 32 reasonable merge candidate can be queued on the request
|
D | sysfs-class-net-statistics | 127 transmitted packets or all packets that have been queued for
|
D | sysfs-power | 195 the sleep state represented by that string is queued up. This
|
/linux-4.4.14/Documentation/ |
D | workqueue.txt | 32 When a new work item gets queued, the worker begins executing again. 86 off of the queue, one after the other. If no work is queued, the 92 which manages worker-pools and processes the queued work items. 96 worker-pools to serve work items queued on unbound workqueues - the 107 When a work item is queued to a workqueue, the target worker-pool is 111 be queued on the worklist of either normal or highpri worker-pool that 148 on code paths that handle memory reclaim are required to be queued on 171 Work items queued to an unbound wq are served by the special 201 Work items of a highpri wq are queued to the highpri 251 behavior. Work items on such wq are always queued to the unbound [all …]
|
D | SubmitChecklist | 92 that it still works with all of the other queued patches and various
|
D | iostats.txt | 65 ultimately handed to the disk, and so it will be counted (and queued)
|
D | vgaarbiter.txt | 183 Airlie finally put this work in shape and queued to Jesse Barnes' PCI tree.
|
D | kernel-per-CPU-kthreads.txt | 246 another way of preventing any callbacks from being queued on the
|
D | kmemcheck.txt | 369 441 * queued once. Changing the restart behaviour to
|
D | kernel-parameters.txt | 1856 * [no]ncqtrim: Turn off queued DSM TRIM. 3159 Set threshold of queued RCU callbacks beyond which 3163 Set threshold of queued RCU callbacks below which 4120 By default, all work items queued to unbound
|
D | memory-barriers.txt | 2228 queued on the semaphore, by virtue of it having a piece of its stack linked to 2266 Once it has queued itself and dropped the semaphore lock, the waiter does not
|
/linux-4.4.14/net/decnet/ |
D | dn_nsp_in.c | 616 int queued = 0; in dn_nsp_otherdata() local 629 queued = 1; in dn_nsp_otherdata() 635 if (!queued) in dn_nsp_otherdata() 641 int queued = 0; in dn_nsp_data() local 655 queued = 1; in dn_nsp_data() 666 if (!queued) in dn_nsp_data()
|
/linux-4.4.14/drivers/media/platform/ |
D | fsl-viu.c | 88 struct list_head queued; member 313 if (list_empty(&vidq->queued)) in restart_video_queue() 315 buf = list_entry(vidq->queued.next, struct viu_buf, vb.queue); in restart_video_queue() 499 if (!list_empty(&vidq->queued)) { in buffer_queue() 503 vidq, &vidq->queued); in buffer_queue() 505 dev, &vidq->queued, vidq->queued.next, in buffer_queue() 506 vidq->queued.prev); in buffer_queue() 507 list_add_tail(&buf->vb.queue, &vidq->queued); in buffer_queue() 533 list_add_tail(&buf->vb.queue, &vidq->queued); in buffer_queue() 976 } else if (!list_empty(&vidq->queued)) { in viu_activate_next_buf() [all …]
|
/linux-4.4.14/net/rds/ |
D | send.c | 812 __be16 dport, int *queued) in rds_send_queue_rm() argument 817 if (*queued) in rds_send_queue_rm() 866 *queued = 1; in rds_send_queue_rm() 871 return *queued; in rds_send_queue_rm() 989 int queued = 0, allocated_mr = 0; in rds_sendmsg() local 1099 dport, &queued)) { in rds_sendmsg() 1111 &queued), in rds_sendmsg() 1113 rdsdebug("sendmsg woke queued %d timeo %ld\n", queued, timeo); in rds_sendmsg()
|
/linux-4.4.14/Documentation/video4linux/ |
D | pxa_camera.txt | 16 When a buffer is queued (pxa_videobuf_ops->buf_queue), the QCI starts. 19 More buffers can be queued while the QCI is started without halting the 82 - first buffer queued for capture 83 Once a first buffer is queued for capture, the QCI is started, but data
|
D | videobuf | 101 When a buffer is queued for I/O, it is passed to buf_queue(), which should
|
D | v4l2-framework.txt | 486 sub-device. After registration events can be queued as usual on the
|
/linux-4.4.14/crypto/ |
D | chainiv.c | 118 int queued; in async_chainiv_schedule_work() local 130 queued = queue_work(kcrypto_wq, &ctx->postponed); in async_chainiv_schedule_work() 131 BUG_ON(!queued); in async_chainiv_schedule_work()
|
/linux-4.4.14/drivers/dma/ |
D | mpc512x_dma.c | 207 struct list_head queued; member 269 while (!list_empty(&mchan->queued)) { in mpc_dma_execute() 270 mdesc = list_first_entry(&mchan->queued, in mpc_dma_execute() 348 if (!list_empty(&mchan->queued)) in mpc_dma_irq_process() 485 list_move_tail(&mdesc->node, &mchan->queued); in mpc_dma_tx_submit() 571 BUG_ON(!list_empty(&mchan->queued)); in mpc_dma_free_chan_resources() 856 list_splice_tail_init(&mchan->queued, &mchan->free); in mpc_dma_device_terminate_all() 972 INIT_LIST_HEAD(&mchan->queued); in mpc_dma_probe()
|
D | sirf-dma.c | 102 struct list_head queued; member 249 sdesc = list_first_entry(&schan->queued, struct sirfsoc_dma_desc, in sirfsoc_dma_execute() 292 if (!list_empty(&schan->queued)) in sirfsoc_dma_irq() 318 if (!list_empty(&schan->queued)) in sirfsoc_dma_irq() 420 list_move_tail(&sdesc->node, &schan->queued); in sirfsoc_dma_tx_submit() 485 list_splice_tail_init(&schan->queued, &schan->free); in sirfsoc_dma_terminate_all() 612 BUG_ON(!list_empty(&schan->queued)); in sirfsoc_dma_free_chan_resources() 636 if (list_empty(&schan->active) && !list_empty(&schan->queued)) in sirfsoc_dma_issue_pending() 938 INIT_LIST_HEAD(&schan->queued); in sirfsoc_dma_probe()
|
D | nbpfaxi.c | 217 struct list_head queued; member 579 if (list_empty(&chan->queued)) in nbpf_issue_pending() 582 list_splice_tail_init(&chan->queued, &chan->active); in nbpf_issue_pending() 623 list_for_each_entry(desc, &chan->queued, node) in nbpf_tx_status() 651 list_add_tail(&desc->node, &chan->queued); in nbpf_tx_submit() 819 list_splice_init(&chan->queued, &head); in nbpf_chan_idle() 1031 INIT_LIST_HEAD(&chan->queued); in nbpf_alloc_chan_resources()
|
/linux-4.4.14/include/drm/ |
D | drm_flip_work.h | 77 struct list_head queued; member
|
/linux-4.4.14/Documentation/scsi/ |
D | hptiop.txt | 84 All queued requests are handled via inbound/outbound queue port. 121 Non-queued requests (reset/flush etc) can be sent via inbound message 129 All queued requests are handled via inbound/outbound list. 160 Non-queued requests (reset communication/reset/flush etc) can be sent via PCIe
|
D | lpfc.txt | 33 In older revisions of the lpfc driver, the driver internally queued i/o 38 the LLDD would simply be queued for a short duration, allowing the device
|
D | sym53c8xx_2.txt | 228 maximum number of queued commands up to 32. The Symbios Setup only allows 231 The maximum number of simultaneous tagged commands queued to a device 239 accept more than 64 simultaneous commands. So, using more than 64 queued 416 Maximum number of queued commands (default answer: 32) 418 that can be queued to a device. The maximum supported value is 255. 436 - enable tagged commands, up to 4 tagged commands queued. 450 #tags will be truncated to the max queued commands configuration parameter.
|
D | scsi_mid_low_api.txt | 373 scsi_block_requests - prevent further commands being queued to given host 384 scsi_unblock_requests - allow further commands to be queued to given host 475 * scsi_block_requests - prevent further commands being queued to given host 684 * scsi_unblock_requests - allow further commands to be queued to given host 847 * will then be queued on current host during eh. 867 * queued on current host during eh. 885 * queued on current host during eh. 903 * queued on current host during eh. 1215 cmd_per_lun - maximum number of commands that can be queued on devices 1256 commands become queued against the LLD than are indicated by [all …]
|
D | ncr53c8xx.txt | 256 maximum number of queued commands up to 32. The Symbios Setup only allows 259 The maximum number of simultaneous tagged commands queued to a device 268 more than 64 simultaneous commands. So, using more than 64 queued commands 577 that can be queued to a device. The maximum supported value is 32. 635 - enable tagged commands, up to 4 tagged commands queued. 700 #tags will be truncated to the max queued commands configuration parameter. 1121 Max number of commands that can be queued to a host. 1124 Max number of commands queued to a host for a device.
|
D | libsas.txt | 314 Returns: -SAS_QUEUE_FULL, -ENOMEM, nothing was queued; 315 0, the task(s) were queued.
|
D | ChangeLog.ncr53c8xx | 358 Count actual number of CCB queued to the controller (future use). 442 - Prepare CCBs for SCSI commands that cannot be queued, instead of
|
D | arcmsr_spec.txt | 86 ** 0x03 : Reset (Abort all queued Command)
|
D | ChangeLog.sym53c8xx | 19 queued to SCRIPTS. This is not always true, notably after a
|
D | ChangeLog.1992-1997 | 730 being queued to more than one host at the same time (used when
|
D | ChangeLog.lpfc | 759 * If we rcv a plogi on a NPort queued up for discovery, clear the
|
/linux-4.4.14/drivers/isdn/pcbit/ |
D | drv.c | 320 chan->queued = 0; in pcbit_block_timer() 348 if (chan->queued >= MAX_QUEUED) in pcbit_xmit() 353 chan->queued); in pcbit_xmit() 373 chan->queued++; in pcbit_xmit() 509 if (chan->queued == MAX_QUEUED) { in pcbit_l3_receive() 515 chan->queued--; in pcbit_l3_receive()
|
D | pcbit.h | 28 unsigned char queued; /* unacked data messages */ member
|
D | layer2.c | 468 if (chan->queued) { in pcbit_fake_conf() 469 chan->queued--; in pcbit_fake_conf()
|
/linux-4.4.14/drivers/tty/serial/ |
D | amba-pl011.c | 157 bool queued; member 435 if (uap->dmatx.queued) in pl011_dma_tx_callback() 454 uap->dmatx.queued = false; in pl011_dma_tx_callback() 494 uap->dmatx.queued = false; in pl011_dma_tx_refill() 526 uap->dmatx.queued = false; in pl011_dma_tx_refill() 535 uap->dmatx.queued = false; in pl011_dma_tx_refill() 556 uap->dmatx.queued = true; in pl011_dma_tx_refill() 589 if (uap->dmatx.queued) { in pl011_dma_tx_irq() 615 if (uap->dmatx.queued) { in pl011_dma_tx_stop() 640 if (!uap->dmatx.queued) { in pl011_dma_tx_start() [all …]
|
/linux-4.4.14/kernel/sched/ |
D | idle_task.c | 54 static void task_tick_idle(struct rq *rq, struct task_struct *curr, int queued) in task_tick_idle() argument
|
D | core.c | 1170 bool queued, running; in do_set_cpus_allowed() local 1174 queued = task_on_rq_queued(p); in do_set_cpus_allowed() 1177 if (queued) { in do_set_cpus_allowed() 1192 if (queued) in do_set_cpus_allowed() 1430 int running, queued; in wait_task_inactive() local 1468 queued = task_on_rq_queued(p); in wait_task_inactive() 1500 if (unlikely(queued)) { in wait_task_inactive() 3368 int oldprio, queued, running, enqueue_flag = ENQUEUE_RESTORE; in rt_mutex_setprio() local 3397 queued = task_on_rq_queued(p); in rt_mutex_setprio() 3399 if (queued) in rt_mutex_setprio() [all …]
|
D | stop_task.c | 77 static void task_tick_stop(struct rq *rq, struct task_struct *curr, int queued) in task_tick_stop() argument
|
D | deadline.c | 1219 static void task_tick_dl(struct rq *rq, struct task_struct *p, int queued) in task_tick_dl() argument 1228 if (hrtick_enabled(rq) && queued && p->dl.runtime > 0 && in task_tick_dl()
|
D | sched.h | 1209 void (*task_tick) (struct rq *rq, struct task_struct *p, int queued);
|
D | fair.c | 3283 entity_tick(struct cfs_rq *cfs_rq, struct sched_entity *curr, int queued) in entity_tick() argument 3301 if (queued) { in entity_tick() 7875 static void task_tick_fair(struct rq *rq, struct task_struct *curr, int queued) in task_tick_fair() argument 7882 entity_tick(cfs_rq, se, queued); in task_tick_fair()
|
D | rt.c | 2209 static void task_tick_rt(struct rq *rq, struct task_struct *p, int queued) in task_tick_rt() argument
|
/linux-4.4.14/Documentation/arm/keystone/ |
D | knav-qmss.txt | 13 management of the packet queues. Packets are queued/de-queued by writing or
|
/linux-4.4.14/drivers/usb/dwc3/ |
D | gadget.h | 75 req->queued = true; in dwc3_gadget_move_request_queued()
|
D | core.h | 634 unsigned queued:1; member
|
D | gadget.c | 240 if (req->queued) { in dwc3_gadget_giveback() 254 req->queued = false; in dwc3_gadget_giveback()
|
/linux-4.4.14/Documentation/block/ |
D | stat.txt | 44 already-queued I/O request. 74 had I/O requests queued.
|
D | biodoc.txt | 787 Returns 1 if the queue can accept a new queued command, 0 if we are 1036 multi-page bios being queued in one shot, we may not need to wait to merge
|
/linux-4.4.14/drivers/isdn/act2000/ |
D | act2000.h | 122 short queued; /* User-Data Bytes in TX queue */ member
|
D | capi.c | 486 chan->queued = 0; in actcapi_disconnect_b3_resp() 609 chan->queued -= m->msg.data_b3_req.datalen; in handle_ack() 613 if (chan->queued < 0) in handle_ack() 614 chan->queued = 0; in handle_ack()
|
D | module.c | 432 if ((chan->queued + len) >= ACT2000_MAX_QUEUED) in act2000_sendbuf() 465 chan->queued += len; in act2000_sendbuf()
|
/linux-4.4.14/drivers/staging/lustre/lustre/osc/ |
D | osc_io.c | 112 int queued = 0; in osc_io_submit() local 172 if (++queued == max_pages) { in osc_io_submit() 173 queued = 0; in osc_io_submit() 181 if (queued > 0) in osc_io_submit()
|
/linux-4.4.14/net/lapb/ |
D | lapb_in.c | 254 int queued = 0; in lapb_state3_machine() local 408 queued = 1; in lapb_state3_machine() 470 if (!queued) in lapb_state3_machine()
|
/linux-4.4.14/arch/x86/kernel/apic/ |
D | apic.c | 1217 unsigned int value, queued; in setup_local_APIC() local 1287 queued = 0; in setup_local_APIC() 1289 queued |= apic_read(APIC_IRR + i*0x10); in setup_local_APIC() 1305 if (queued) { in setup_local_APIC() 1312 } while (queued && max_loops > 0); in setup_local_APIC()
|
/linux-4.4.14/fs/ocfs2/cluster/ |
D | heartbeat.c | 695 int queued = 0; in o2hb_shutdown_slot() local 713 queued = 1; in o2hb_shutdown_slot() 718 if (queued) in o2hb_shutdown_slot() 779 int queued = 0; in o2hb_check_slot() local 873 queued = 1; in o2hb_check_slot() 925 queued = 1; in o2hb_check_slot() 941 if (queued) in o2hb_check_slot()
|
/linux-4.4.14/net/caif/ |
D | caif_socket.c | 130 bool queued = false; in caif_queue_rcv_skb() local 153 queued = !sock_flag(sk, SOCK_DEAD); in caif_queue_rcv_skb() 154 if (queued) in caif_queue_rcv_skb() 158 if (queued) in caif_queue_rcv_skb()
|
/linux-4.4.14/drivers/mtd/ |
D | ftl.c | 611 int queued, ret; in reclaim_block() local 618 queued = 0; in reclaim_block() 629 queued = 1; in reclaim_block() 649 if (queued) { in reclaim_block()
|
/linux-4.4.14/drivers/staging/rdma/hfi1/ |
D | user_sdma.c | 801 u32 datalen = 0, queued = 0, data_sent = 0; in user_sdma_send_pkts() local 929 while (queued < datalen && in user_sdma_send_pkts() 941 len = min((datalen - queued), len); in user_sdma_send_pkts() 960 queued += len; in user_sdma_send_pkts() 962 if (unlikely(queued < datalen && in user_sdma_send_pkts()
|
/linux-4.4.14/Documentation/filesystems/caching/ |
D | fscache.txt | 246 ok=N Number of attr changed requests queued 284 enq=N Number of times async ops queued for processing 288 dfr=N Number of async ops queued for deferred release 403 S Show objects that have work queued 404 s Show objects that don't have work queued
|
D | object.txt | 123 is not masked, the object will be queued for processing (by calling 201 operations the netfs has queued for an object. If creation failed, the
|
D | backend-api.txt | 422 queued and 0 returned. When the read finishes, fscache_end_io() should be 431 queued and released when I/O on the page has been formally ended.
|
D | netfs-api.txt | 869 that, operations will be queued again behind the invalidation operation.
|
/linux-4.4.14/net/sunrpc/ |
D | svc_xprt.c | 327 bool queued = false; in svc_xprt_do_enqueue() local 362 if (!queued) { in svc_xprt_do_enqueue() 390 if (!queued) { in svc_xprt_do_enqueue() 391 queued = true; in svc_xprt_do_enqueue()
|
/linux-4.4.14/include/linux/ |
D | blk-mq.h | 50 unsigned long queued; member
|
D | kvm_host.h | 264 u32 queued; member
|
/linux-4.4.14/Documentation/sound/alsa/ |
D | timestamping.txt | 20 the ring buffer and the amount of queued samples. 27 queued samples have been played out. 49 |< codec delay >|<--hw delay-->|<queued samples>|<---avail->|
|
/linux-4.4.14/Documentation/devicetree/bindings/soc/ti/ |
D | keystone-navigator-qmss.txt | 9 management of the packet queues. Packets are queued/de-queued by writing or
|
/linux-4.4.14/Documentation/dmaengine/ |
D | pxa_dma.txt | 8 is queued even on a running DMA channel. 88 such a transfer tx_submit() will be queued on the submitted queue, and
|
D | client.txt | 158 queue is started and subsequent ones queued up.
|
D | provider.txt | 306 currently queued transfers, but only on subsequent ones 394 that can be queued to buffers before being flushed to
|
/linux-4.4.14/net/netfilter/ |
D | nfnetlink_queue.c | 675 unsigned int queued; in nfqnl_enqueue_packet() local 712 queued = 0; in nfqnl_enqueue_packet() 720 queued++; in nfqnl_enqueue_packet() 726 if (queued) { in nfqnl_enqueue_packet()
|
/linux-4.4.14/drivers/mailbox/ |
D | Kconfig | 5 on-chip processors through queued messages and interrupt driven
|
/linux-4.4.14/Documentation/filesystems/ |
D | gfs2-glocks.txt | 16 The gl_holders list contains all the queued lock requests (not 20 are queued, except for those marked LM_FLAG_PRIORITY which are 215 queue - Number of glock requests queued (qcnt in glstats file)
|
D | dnotify.txt | 34 (SIGRTMIN + <n>) so that the notifications may be queued. This is
|
D | fuse.txt | 161 this flag is set, an INTERRUPT request is queued. 164 request is queued. 167 userspace filesystem will receive queued INTERRUPTs before any others.
|
D | proc.txt | 244 SigQ number of signals queued/max. number for queue
|
/linux-4.4.14/drivers/tty/hvc/ |
D | hvc_iucv.c | 463 int queued; in hvc_iucv_put_chars() local 472 queued = hvc_iucv_queue(priv, buf, count); in hvc_iucv_put_chars() 475 return queued; in hvc_iucv_put_chars()
|
/linux-4.4.14/drivers/net/wireless/ath/ath9k/ |
D | debug.h | 172 u32 queued; member
|
D | debug.c | 598 PR("MPDUs Queued: ", queued); in read_file_xmit() 1268 AWDATA(queued); in ath9k_get_et_stats()
|
D | xmit.c | 2066 TX_STAT_INC(txq->axq_qnum, queued); in ath_tx_send_normal() 2445 TX_STAT_INC(txctl.txq->axq_qnum, queued); in ath_tx_cabq()
|
/linux-4.4.14/drivers/media/v4l2-core/ |
D | videobuf2-v4l2.c | 872 unsigned int queued:1; member 1015 fileio->bufs[i].queued = 1; in __vb2_init_fileio() 1135 buf->queued = 0; in __vb2_perform_fileio() 1213 buf->queued = 1; in __vb2_perform_fileio()
|
/linux-4.4.14/drivers/crypto/caam/ |
D | Kconfig | 72 more descriptor completions are queued without reaching the count
|
/linux-4.4.14/drivers/media/usb/tm6000/ |
D | tm6000.h | 94 struct list_head queued; member
|
D | tm6000-video.c | 1375 "queued=%d\n", list_empty(&dev->vidq.queued)); in __tm6000_open() 1660 INIT_LIST_HEAD(&dev->vidq.queued); in tm6000_v4l2_register()
|
/linux-4.4.14/drivers/net/wireless/cw1200/ |
D | wsm.c | 1591 int queued; in cw1200_get_prio_queue() local 1596 queued = cw1200_queue_get_num_queued(&priv->tx_queue[i], in cw1200_get_prio_queue() 1598 if (!queued) in cw1200_get_prio_queue() 1600 *total += queued; in cw1200_get_prio_queue()
|
/linux-4.4.14/Documentation/sysctl/ |
D | net.txt | 152 Maximum number of packets, queued on the INPUT side, when the interface 197 unix_dgram_qlen limits the max number of datagrams queued in Unix domain
|
D | kernel.txt | 741 of POSIX realtime (queued) signals that can be outstanding 744 rtsig-nr shows the number of RT signals currently queued.
|
/linux-4.4.14/drivers/spi/ |
D | spi.c | 1406 master->queued = true; in spi_master_initialize_queue() 1914 if (master->queued) { in spi_unregister_master() 1933 if (!master->queued) in spi_master_suspend() 1948 if (!master->queued) in spi_master_resume()
|
/linux-4.4.14/include/linux/spi/ |
D | spi.h | 485 bool queued; member
|
/linux-4.4.14/Documentation/s390/ |
D | CommonIO | 94 A write request to this file is blocked until all queued cio actions are
|
/linux-4.4.14/drivers/net/wireless/ath/carl9170/ |
D | tx.c | 1429 goto queued; in carl9170_tx_ampdu_queue() 1437 goto queued; in carl9170_tx_ampdu_queue() 1442 queued: in carl9170_tx_ampdu_queue()
|
/linux-4.4.14/Documentation/RCU/ |
D | rcubarrier.txt | 76 synchronize_rcu(), in particular, if there are no RCU callbacks queued 199 that RCU callbacks are never reordered once queued on one of the per-CPU
|
D | trace.txt | 143 "N" Indicates that there are callbacks queued that are not 148 "R" Indicates that there are callbacks queued that are 151 "W" Indicates that there are callbacks queued that are 154 "D" Indicates that there are callbacks queued that have
|
/linux-4.4.14/Documentation/input/ |
D | atarikbd.txt | 133 scanning the joysticks (samples are not queued). 143 are not queued). 367 joystick events are also queued. 372 causes any accumulated motion to be immediately queued as packets, if the
|
/linux-4.4.14/Documentation/devicetree/bindings/mailbox/ |
D | omap-mailbox.txt | 5 using a queued mailbox interrupt mechanism. The IP block is external to the
|
/linux-4.4.14/drivers/scsi/ |
D | Kconfig | 709 int "maximum number of queued commands" 713 This specifies how many SCSI commands can be maximally queued for 1085 int "Maximum number of queued commands" 1090 that can be queued to any device, when tagged command queuing is 1091 possible. The driver supports up to 256 queued commands per device. 1180 int "maximum number of queued commands" 1185 that can be queued to any device, when tagged command queuing is 1428 int "maximum number of queued commands" 1432 This specifies how many SCSI commands can be maximally queued for
|
D | ncr53c8xx.c | 1580 u_char queued; member 4466 cp->queued = 1; in ncr_put_start_queue() 5102 if (cp->queued) { in ncr_ccb_skipped() 5106 if (cp->queued) { in ncr_ccb_skipped() 5108 cp->queued = 0; in ncr_ccb_skipped() 7291 if (cp->queued) { in ncr_free_ccb() 7297 if (cp->queued) { in ncr_free_ccb() 7299 cp->queued = 0; in ncr_free_ccb()
|
/linux-4.4.14/Documentation/hid/ |
D | hidraw.txt | 41 read() will read a queued report received from the HID device. On USB
|
D | uhid.txt | 120 read() will return a queued output report. No reaction is required to any of
|
/linux-4.4.14/Documentation/cris/ |
D | README | 106 block: queued sectors max/low 9109kB/3036kB, 64 slots per queue
|
/linux-4.4.14/net/mac80211/ |
D | tx.c | 1074 bool queued = false; in ieee80211_tx_prep_agg() local 1114 queued = true; in ieee80211_tx_prep_agg() 1134 return queued; in ieee80211_tx_prep_agg() 1196 bool queued; in ieee80211_tx_prepare() local 1198 queued = ieee80211_tx_prep_agg(tx, skb, info, in ieee80211_tx_prepare() 1201 if (unlikely(queued)) in ieee80211_tx_prepare()
|
D | cfg.c | 2555 bool queued = false; in ieee80211_start_roc_work() local 2638 queued = true; in ieee80211_start_roc_work() 2654 queued = true; in ieee80211_start_roc_work() 2659 queued = true; in ieee80211_start_roc_work() 2672 queued = true; in ieee80211_start_roc_work() 2688 if (!queued) in ieee80211_start_roc_work()
|
/linux-4.4.14/Documentation/power/ |
D | runtime_pm.txt | 246 - if set, there's a pending request (i.e. a work item queued up into pm_wq) 340 success or error code if the request has not been queued up 345 expired then the work item is queued up immediately 351 item is queued up immediately); returns 0 on success, 1 if the device's PM 353 hasn't been scheduled (or queued up if 'delay' is 0); if the execution of 361 error code if the request hasn't been queued up
|
D | pci.txt | 1027 queued by the PM core (for example, after processing a request to resume a
|
/linux-4.4.14/Documentation/nfc/ |
D | nfc-hci.txt | 187 the current llc. In case of shdlc, the frame is queued in shdlc rx queue. 250 queued to HCI rx_queue and will be dispatched from HCI rx worker
|
/linux-4.4.14/drivers/usb/serial/ |
D | io_ti.c | 1796 int queued; in edge_tty_recv() local 1798 queued = tty_insert_flip_string(&port->port, data, length); in edge_tty_recv() 1799 if (queued < length) in edge_tty_recv() 1801 __func__, length - queued); in edge_tty_recv()
|
/linux-4.4.14/Documentation/locking/ |
D | mutex-design.txt | 60 soon. The mutex spinners are queued up using MCS lock so that only
|
D | ww-mutex-design.txt | 23 evict other buffers which are already queued up to the GPU), but for a
|
/linux-4.4.14/drivers/usb/mon/ |
D | mon_bin.c | 127 u32 queued; member 1108 if (put_user(nevents, &sp->queued)) in mon_bin_ioctl()
|
/linux-4.4.14/Documentation/crypto/ |
D | async-tx-api.txt | 77 has been queued to execute asynchronously. Descriptors are recycled
|
/linux-4.4.14/drivers/media/usb/cx231xx/ |
D | cx231xx-cards.c | 1365 INIT_LIST_HEAD(&dev->video_mode.vidq.queued); in cx231xx_init_dev() 1369 INIT_LIST_HEAD(&dev->vbi_mode.vidq.queued); in cx231xx_init_dev()
|
D | cx231xx.h | 249 struct list_head queued; member
|
/linux-4.4.14/drivers/scsi/aic7xxx/ |
D | aic7xxx.reg | 1030 * Input queue for queued SCBs (commands that the seqencer has yet to start) 1041 * Number of queued SCBs 1075 * Number of queued SCBs in the Out FIFO
|
D | aic79xx.seq | 361 * target queue is empty, the SCB can be queued 366 * could allow commands to be queued out of order. 369 * SCB can be queued to the waiting for selection 382 * it can be queued. Since the SCB_SCSIID of the 384 * queued SCB, there is no need to restore the SCBID
|
D | aic79xx.reg | 4048 * queued to the controller, it cannot enter 4050 * selections for any previously queued 4052 * the wait, the MK_MESSAGE SCB is queued
|
D | aic7xxx.seq | 64 * command for which a second SCB has been queued. The sequencer will 104 * We have at least one queued SCB now and we don't have any
|
/linux-4.4.14/drivers/atm/ |
D | eni.c | 152 static int tx_complete = 0,dma_complete = 0,queued = 0,requeued = 0, variable 1177 queued++; in do_tx() 1464 tx_complete,dma_complete,queued,requeued,submitted,backlogged,
|
/linux-4.4.14/Documentation/timers/ |
D | hrtimers.txt | 97 queued timers, without having to walk the rbtree.
|
/linux-4.4.14/Documentation/cgroups/ |
D | blkio-controller.txt | 245 - Total number of requests queued up at any given instant for this 258 (i.e., went from 0 to 1 request queued) to get a timeslice for one of
|
D | memory.txt | 497 writeback - # of bytes of file/anon cache that are queued for syncing to
|
/linux-4.4.14/net/ipv4/ |
D | tcp_input.c | 5751 int queued = 0; in tcp_rcv_state_process() local 5796 queued = tcp_rcv_synsent_state_process(sk, skb, th); in tcp_rcv_state_process() 5797 if (queued >= 0) in tcp_rcv_state_process() 5798 return queued; in tcp_rcv_state_process() 5996 queued = 1; in tcp_rcv_state_process() 6006 if (!queued) { in tcp_rcv_state_process()
|
/linux-4.4.14/Documentation/scheduler/ |
D | sched-design-CFS.txt | 65 rq->cfs.load value, which is the sum of the weights of the tasks queued on the
|
/linux-4.4.14/drivers/block/drbd/ |
D | drbd_worker.c | 605 int queued = sk->sk_wmem_queued; in make_resync_request() local 607 if (queued > sndbuf / 2) { in make_resync_request()
|
/linux-4.4.14/Documentation/ide/ |
D | ChangeLog.ide-tape.1995-2002 | 36 * from the other device can be queued and ide.c will
|
/linux-4.4.14/Documentation/spi/ |
D | spi-summary | 569 method is not used on queued controllers and must be NULL if 577 SPI subsystem, just implement the queued methods specified above. Using
|
/linux-4.4.14/drivers/net/wireless/ath/ath6kl/ |
D | cfg80211.c | 3186 bool more_data, queued; in ath6kl_mgmt_tx() local 3224 queued = ath6kl_mgmt_powersave_ap(vif, id, freq, wait, buf, len, in ath6kl_mgmt_tx() 3226 if (queued) in ath6kl_mgmt_tx()
|
/linux-4.4.14/Documentation/networking/dsa/ |
D | dsa.txt | 163 These frames are then queued for transmission using the master network device
|
/linux-4.4.14/Documentation/DocBook/ |
D | device-drivers.xml.db | 1066 API-spi-get-next-queued-message
|
/linux-4.4.14/Documentation/m68k/ |
D | kernel-options.txt | 584 This is the maximum number of SCSI commands queued internally to the
|
/linux-4.4.14/init/ |
D | Kconfig | 621 they have RCU callbacks queued, and prevents RCU from waking
|
/linux-4.4.14/ |
D | MAINTAINERS | 629 P: Andres Salomon <dilinger@queued.net>
|