/linux-4.1.27/drivers/s390/cio/ |
D | eadm_sch.c | 64 static int eadm_subchannel_start(struct subchannel *sch, struct aob *aob) in eadm_subchannel_start() argument 66 union orb *orb = &get_eadm_private(sch)->orb; in eadm_subchannel_start() 71 orb->eadm.intparm = (u32)(addr_t)sch; in eadm_subchannel_start() 75 EADM_LOG_HEX(6, &sch->schid, sizeof(sch->schid)); in eadm_subchannel_start() 77 cc = ssch(sch->schid, orb); in eadm_subchannel_start() 80 sch->schib.scsw.eadm.actl |= SCSW_ACTL_START_PEND; in eadm_subchannel_start() 91 static int eadm_subchannel_clear(struct subchannel *sch) in eadm_subchannel_clear() argument 95 cc = csch(sch->schid); in eadm_subchannel_clear() 99 sch->schib.scsw.eadm.actl |= SCSW_ACTL_CLEAR_PEND; in eadm_subchannel_clear() 105 struct subchannel *sch = (struct subchannel *) data; in eadm_subchannel_timeout() local [all …]
|
D | cio.c | 90 int cio_set_options(struct subchannel *sch, int flags) in cio_set_options() argument 92 struct io_subchannel_private *priv = to_io_private(sch); in cio_set_options() 101 cio_start_handle_notoper(struct subchannel *sch, __u8 lpm) in cio_start_handle_notoper() argument 106 sch->lpm &= ~lpm; in cio_start_handle_notoper() 108 sch->lpm = 0; in cio_start_handle_notoper() 111 "subchannel 0.%x.%04x!\n", sch->schid.ssid, in cio_start_handle_notoper() 112 sch->schid.sch_no); in cio_start_handle_notoper() 114 if (cio_update_schib(sch)) in cio_start_handle_notoper() 117 sprintf(dbf_text, "no%s", dev_name(&sch->dev)); in cio_start_handle_notoper() 119 CIO_HEX_EVENT(0, &sch->schib, sizeof (struct schib)); in cio_start_handle_notoper() [all …]
|
D | css.c | 68 struct subchannel *sch = to_subchannel(dev); in call_fn_known_sch() local 73 idset_sch_del(cb->set, sch->schid); in call_fn_known_sch() 75 rc = cb->fn_known_sch(sch, cb->data); in call_fn_known_sch() 92 struct subchannel *sch; in call_fn_all_sch() local 95 sch = get_subchannel_by_schid(schid); in call_fn_all_sch() 96 if (sch) { in call_fn_all_sch() 98 rc = cb->fn_known_sch(sch, cb->data); in call_fn_all_sch() 99 put_device(&sch->dev); in call_fn_all_sch() 148 static int css_sch_create_locks(struct subchannel *sch) in css_sch_create_locks() argument 150 sch->lock = kmalloc(sizeof(*sch->lock), GFP_KERNEL); in css_sch_create_locks() [all …]
|
D | device.c | 150 static int io_subchannel_prepare(struct subchannel *sch) in io_subchannel_prepare() argument 157 cdev = sch_get_cdev(sch); in io_subchannel_prepare() 220 struct subchannel *sch = to_subchannel(dev); in chpids_show() local 221 struct chsc_ssd_info *ssd = &sch->ssd_info; in chpids_show() 240 struct subchannel *sch = to_subchannel(dev); in pimpampom_show() local 241 struct pmcw *pmcw = &sch->schib.pmcw; in pimpampom_show() 323 struct subchannel *sch; in ccw_device_set_offline() local 337 sch = to_subchannel(cdev->dev.parent); in ccw_device_set_offline() 358 io_subchannel_quiesce(sch); in ccw_device_set_offline() 592 struct subchannel *sch; in available_show() local [all …]
|
D | device_ops.c | 140 struct subchannel *sch; in ccw_device_clear() local 145 sch = to_subchannel(cdev->dev.parent); in ccw_device_clear() 146 if (!sch->schib.pmcw.ena) in ccw_device_clear() 154 ret = cio_clear(sch); in ccw_device_clear() 188 struct subchannel *sch; in ccw_device_start_key() local 193 sch = to_subchannel(cdev->dev.parent); in ccw_device_start_key() 194 if (!sch->schib.pmcw.ena) in ccw_device_start_key() 209 ((sch->schib.scsw.cmd.stctl & SCSW_STCTL_PRIM_STATUS) && in ccw_device_start_key() 210 !(sch->schib.scsw.cmd.stctl & SCSW_STCTL_SEC_STATUS)) || in ccw_device_start_key() 213 ret = cio_set_options (sch, flags); in ccw_device_start_key() [all …]
|
D | device_fsm.c | 39 struct subchannel *sch; in ccw_timeout_log() local 44 sch = to_subchannel(cdev->dev.parent); in ccw_timeout_log() 45 private = to_io_private(sch); in ccw_timeout_log() 47 cc = stsch_err(sch->schid, &schib); in ccw_timeout_log() 57 dev_name(&sch->dev)); in ccw_timeout_log() 59 "vpm: %02x\n", sch->lpm, sch->opm, sch->vpm); in ccw_timeout_log() 138 struct subchannel *sch; in ccw_device_cancel_halt_clear() local 141 sch = to_subchannel(cdev->dev.parent); in ccw_device_cancel_halt_clear() 142 if (cio_update_schib(sch)) in ccw_device_cancel_halt_clear() 144 if (!sch->schib.pmcw.ena) in ccw_device_cancel_halt_clear() [all …]
|
D | device_pgid.c | 33 struct subchannel *sch = to_subchannel(cdev->dev.parent); in verify_done() local 41 if (sch->config.mp != mpath) { in verify_done() 42 sch->config.mp = mpath; in verify_done() 43 rc = cio_commit_config(sch); in verify_done() 48 sch->vpm); in verify_done() 72 struct subchannel *sch = to_subchannel(cdev->dev.parent); in nop_do() local 75 req->lpm = lpm_adjust(req->lpm, sch->schib.pmcw.pam & sch->opm & in nop_do() 84 verify_done(cdev, sch->vpm ? 0 : -EACCES); in nop_do() 104 struct subchannel *sch = to_subchannel(cdev->dev.parent); in nop_callback() local 109 sch->vpm |= req->lpm; in nop_callback() [all …]
|
D | chsc_sch.c | 57 static void chsc_subchannel_irq(struct subchannel *sch) in chsc_subchannel_irq() argument 59 struct chsc_private *private = dev_get_drvdata(&sch->dev); in chsc_subchannel_irq() 70 sch->schid.ssid, sch->schid.sch_no); in chsc_subchannel_irq() 75 cio_update_schib(sch); in chsc_subchannel_irq() 77 put_device(&sch->dev); in chsc_subchannel_irq() 80 static int chsc_subchannel_probe(struct subchannel *sch) in chsc_subchannel_probe() argument 86 sch->schid.ssid, sch->schid.sch_no); in chsc_subchannel_probe() 87 sch->isc = CHSC_SCH_ISC; in chsc_subchannel_probe() 91 dev_set_drvdata(&sch->dev, private); in chsc_subchannel_probe() 92 ret = cio_enable_subchannel(sch, (u32)(unsigned long)sch); in chsc_subchannel_probe() [all …]
|
D | ccwreq.c | 77 struct subchannel *sch = to_subchannel(cdev->dev.parent); in ccwreq_do() local 89 rc = cio_start(sch, cp, (u8) req->mask); in ccwreq_do() 105 rc = cio_clear(sch); in ccwreq_do() 152 struct subchannel *sch = to_subchannel(cdev->dev.parent); in ccw_request_cancel() local 159 rc = cio_clear(sch); in ccw_request_cancel() 327 struct subchannel *sch = to_subchannel(cdev->dev.parent); in ccw_request_timeout() local 331 if (cio_update_schib(sch)) in ccw_request_timeout() 335 if ((0x80 >> chp) & sch->schib.pmcw.lpum) in ccw_request_timeout() 339 scsw_cstat(&sch->schib.scsw), in ccw_request_timeout() 340 scsw_dstat(&sch->schib.scsw), in ccw_request_timeout() [all …]
|
D | chsc.c | 88 u16 sch; /* subchannel */ member 208 static int s390_subchannel_remove_chpid(struct subchannel *sch, void *data) in s390_subchannel_remove_chpid() argument 210 spin_lock_irq(sch->lock); in s390_subchannel_remove_chpid() 211 if (sch->driver && sch->driver->chp_event) in s390_subchannel_remove_chpid() 212 if (sch->driver->chp_event(sch, data, CHP_OFFLINE) != 0) in s390_subchannel_remove_chpid() 214 spin_unlock_irq(sch->lock); in s390_subchannel_remove_chpid() 218 sch->lpm = 0; in s390_subchannel_remove_chpid() 219 spin_unlock_irq(sch->lock); in s390_subchannel_remove_chpid() 220 css_schedule_eval(sch->schid); in s390_subchannel_remove_chpid() 241 static int __s390_process_res_acc(struct subchannel *sch, void *data) in __s390_process_res_acc() argument [all …]
|
D | cio.h | 120 extern int cio_update_schib(struct subchannel *sch); 121 extern int cio_commit_config(struct subchannel *sch); 123 int cio_tm_start_key(struct subchannel *sch, struct tcw *tcw, u8 lpm, u8 key); 124 int cio_tm_intrg(struct subchannel *sch); 131 extern void cio_tsch(struct subchannel *sch);
|
D | device_status.c | 53 struct subchannel *sch; in ccw_device_path_notoper() local 55 sch = to_subchannel(cdev->dev.parent); in ccw_device_path_notoper() 56 if (cio_update_schib(sch)) in ccw_device_path_notoper() 61 sch->schid.ssid, sch->schid.sch_no, in ccw_device_path_notoper() 62 sch->schib.pmcw.pnom); in ccw_device_path_notoper() 64 sch->lpm &= ~sch->schib.pmcw.pnom; in ccw_device_path_notoper() 311 struct subchannel *sch; in ccw_device_do_sense() local 315 sch = to_subchannel(cdev->dev.parent); in ccw_device_do_sense() 330 sense_ccw = &to_io_private(sch)->sense_ccw; in ccw_device_do_sense() 336 rc = cio_start(sch, sense_ccw, 0xff); in ccw_device_do_sense()
|
D | io_sch.h | 26 static inline struct ccw_device *sch_get_cdev(struct subchannel *sch) in sch_get_cdev() argument 28 struct io_subchannel_private *priv = to_io_private(sch); in sch_get_cdev() 32 static inline void sch_set_cdev(struct subchannel *sch, in sch_set_cdev() argument 35 struct io_subchannel_private *priv = to_io_private(sch); in sch_set_cdev() 119 struct subchannel *sch; member
|
D | cmf.c | 189 struct subchannel *sch; in set_schib() local 191 sch = to_subchannel(cdev->dev.parent); in set_schib() 193 sch->config.mme = mme; in set_schib() 194 sch->config.mbfc = mbfc; in set_schib() 197 sch->config.mba = address; in set_schib() 199 sch->config.mbi = address; in set_schib() 201 return cio_commit_config(sch); in set_schib() 298 struct subchannel *sch; in cmf_copy_block() local 303 sch = to_subchannel(cdev->dev.parent); in cmf_copy_block() 305 if (cio_update_schib(sch)) in cmf_copy_block() [all …]
|
D | device_id.c | 200 struct subchannel *sch = to_subchannel(cdev->dev.parent); in ccw_device_sense_id_start() local 218 req->lpm = sch->schib.pmcw.pam & sch->opm; in ccw_device_sense_id_start()
|
D | css.h | 113 void css_update_ssd_info(struct subchannel *sch); 145 void css_sched_sch_todo(struct subchannel *sch, enum sch_todo todo);
|
D | eadm_sch.h | 14 struct subchannel *sch; member
|
D | chp.h | 61 u8 chp_get_sch_opm(struct subchannel *sch);
|
D | device.h | 87 void io_subchannel_init_config(struct subchannel *sch);
|
D | chsc.h | 204 u16 sch; member
|
D | chp.c | 77 u8 chp_get_sch_opm(struct subchannel *sch) in chp_get_sch_opm() argument 87 chpid.id = sch->schib.pmcw.chpid[i]; in chp_get_sch_opm()
|
D | qdio_setup.c | 321 (ssqd->qdio_ssqd.sch != schid->sch_no)) in qdio_setup_get_ssqd()
|
/linux-4.1.27/drivers/gpio/ |
D | gpio-sch.c | 46 static unsigned sch_gpio_offset(struct sch_gpio *sch, unsigned gpio, in sch_gpio_offset() argument 51 if (gpio >= sch->resume_base) { in sch_gpio_offset() 52 gpio -= sch->resume_base; in sch_gpio_offset() 59 static unsigned sch_gpio_bit(struct sch_gpio *sch, unsigned gpio) in sch_gpio_bit() argument 61 if (gpio >= sch->resume_base) in sch_gpio_bit() 62 gpio -= sch->resume_base; in sch_gpio_bit() 68 struct sch_gpio *sch = to_sch_gpio(gc); in sch_gpio_reg_get() local 72 offset = sch_gpio_offset(sch, gpio, reg); in sch_gpio_reg_get() 73 bit = sch_gpio_bit(sch, gpio); in sch_gpio_reg_get() 75 reg_val = !!(inb(sch->iobase + offset) & BIT(bit)); in sch_gpio_reg_get() [all …]
|
D | Makefile | 79 obj-$(CONFIG_GPIO_SCH) += gpio-sch.o
|
/linux-4.1.27/net/sched/ |
D | sch_mq.c | 25 static void mq_destroy(struct Qdisc *sch) in mq_destroy() argument 27 struct net_device *dev = qdisc_dev(sch); in mq_destroy() 28 struct mq_sched *priv = qdisc_priv(sch); in mq_destroy() 38 static int mq_init(struct Qdisc *sch, struct nlattr *opt) in mq_init() argument 40 struct net_device *dev = qdisc_dev(sch); in mq_init() 41 struct mq_sched *priv = qdisc_priv(sch); in mq_init() 46 if (sch->parent != TC_H_ROOT) in mq_init() 61 TC_H_MAKE(TC_H_MAJ(sch->handle), in mq_init() 69 sch->flags |= TCQ_F_MQROOT; in mq_init() 73 mq_destroy(sch); in mq_init() [all …]
|
D | sch_codel.c | 67 static struct sk_buff *dequeue(struct codel_vars *vars, struct Qdisc *sch) in dequeue() argument 69 struct sk_buff *skb = __skb_dequeue(&sch->q); in dequeue() 75 static struct sk_buff *codel_qdisc_dequeue(struct Qdisc *sch) in codel_qdisc_dequeue() argument 77 struct codel_sched_data *q = qdisc_priv(sch); in codel_qdisc_dequeue() 80 skb = codel_dequeue(sch, &q->params, &q->vars, &q->stats, dequeue); in codel_qdisc_dequeue() 85 if (q->stats.drop_count && sch->q.qlen) { in codel_qdisc_dequeue() 86 qdisc_tree_decrease_qlen(sch, q->stats.drop_count); in codel_qdisc_dequeue() 90 qdisc_bstats_update(sch, skb); in codel_qdisc_dequeue() 94 static int codel_qdisc_enqueue(struct sk_buff *skb, struct Qdisc *sch) in codel_qdisc_enqueue() argument 98 if (likely(qdisc_qlen(sch) < sch->limit)) { in codel_qdisc_enqueue() [all …]
|
D | sch_fifo.c | 22 static int bfifo_enqueue(struct sk_buff *skb, struct Qdisc *sch) in bfifo_enqueue() argument 24 if (likely(sch->qstats.backlog + qdisc_pkt_len(skb) <= sch->limit)) in bfifo_enqueue() 25 return qdisc_enqueue_tail(skb, sch); in bfifo_enqueue() 27 return qdisc_reshape_fail(skb, sch); in bfifo_enqueue() 30 static int pfifo_enqueue(struct sk_buff *skb, struct Qdisc *sch) in pfifo_enqueue() argument 32 if (likely(skb_queue_len(&sch->q) < sch->limit)) in pfifo_enqueue() 33 return qdisc_enqueue_tail(skb, sch); in pfifo_enqueue() 35 return qdisc_reshape_fail(skb, sch); in pfifo_enqueue() 38 static int pfifo_tail_enqueue(struct sk_buff *skb, struct Qdisc *sch) in pfifo_tail_enqueue() argument 40 if (likely(skb_queue_len(&sch->q) < sch->limit)) in pfifo_tail_enqueue() [all …]
|
D | sch_multiq.c | 40 multiq_classify(struct sk_buff *skb, struct Qdisc *sch, int *qerr) in multiq_classify() argument 42 struct multiq_sched_data *q = qdisc_priv(sch); in multiq_classify() 68 multiq_enqueue(struct sk_buff *skb, struct Qdisc *sch) in multiq_enqueue() argument 73 qdisc = multiq_classify(skb, sch, &ret); in multiq_enqueue() 78 qdisc_qstats_drop(sch); in multiq_enqueue() 86 sch->q.qlen++; in multiq_enqueue() 90 qdisc_qstats_drop(sch); in multiq_enqueue() 94 static struct sk_buff *multiq_dequeue(struct Qdisc *sch) in multiq_dequeue() argument 96 struct multiq_sched_data *q = qdisc_priv(sch); in multiq_dequeue() 111 netdev_get_tx_queue(qdisc_dev(sch), q->curband))) { in multiq_dequeue() [all …]
|
D | sch_dsmark.c | 55 static int dsmark_graft(struct Qdisc *sch, unsigned long arg, in dsmark_graft() argument 58 struct dsmark_qdisc_data *p = qdisc_priv(sch); in dsmark_graft() 61 __func__, sch, p, new, old); in dsmark_graft() 64 new = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops, in dsmark_graft() 65 sch->handle); in dsmark_graft() 70 sch_tree_lock(sch); in dsmark_graft() 75 sch_tree_unlock(sch); in dsmark_graft() 80 static struct Qdisc *dsmark_leaf(struct Qdisc *sch, unsigned long arg) in dsmark_leaf() argument 82 struct dsmark_qdisc_data *p = qdisc_priv(sch); in dsmark_leaf() 86 static unsigned long dsmark_get(struct Qdisc *sch, u32 classid) in dsmark_get() argument [all …]
|
D | sch_prio.c | 34 prio_classify(struct sk_buff *skb, struct Qdisc *sch, int *qerr) in prio_classify() argument 36 struct prio_sched_data *q = qdisc_priv(sch); in prio_classify() 43 if (TC_H_MAJ(skb->priority) != sch->handle) { in prio_classify() 70 prio_enqueue(struct sk_buff *skb, struct Qdisc *sch) in prio_enqueue() argument 75 qdisc = prio_classify(skb, sch, &ret); in prio_enqueue() 80 qdisc_qstats_drop(sch); in prio_enqueue() 88 sch->q.qlen++; in prio_enqueue() 92 qdisc_qstats_drop(sch); in prio_enqueue() 96 static struct sk_buff *prio_peek(struct Qdisc *sch) in prio_peek() argument 98 struct prio_sched_data *q = qdisc_priv(sch); in prio_peek() [all …]
|
D | sch_red.c | 59 static int red_enqueue(struct sk_buff *skb, struct Qdisc *sch) in red_enqueue() argument 61 struct red_sched_data *q = qdisc_priv(sch); in red_enqueue() 77 qdisc_qstats_overlimit(sch); in red_enqueue() 87 qdisc_qstats_overlimit(sch); in red_enqueue() 100 sch->q.qlen++; in red_enqueue() 103 qdisc_qstats_drop(sch); in red_enqueue() 108 qdisc_drop(skb, sch); in red_enqueue() 112 static struct sk_buff *red_dequeue(struct Qdisc *sch) in red_dequeue() argument 115 struct red_sched_data *q = qdisc_priv(sch); in red_dequeue() 120 qdisc_bstats_update(sch, skb); in red_dequeue() [all …]
|
D | sch_atm.c | 70 static inline struct atm_flow_data *lookup_flow(struct Qdisc *sch, u32 classid) in lookup_flow() argument 72 struct atm_qdisc_data *p = qdisc_priv(sch); in lookup_flow() 82 static int atm_tc_graft(struct Qdisc *sch, unsigned long arg, in atm_tc_graft() argument 85 struct atm_qdisc_data *p = qdisc_priv(sch); in atm_tc_graft() 89 sch, p, flow, new, old); in atm_tc_graft() 101 static struct Qdisc *atm_tc_leaf(struct Qdisc *sch, unsigned long cl) in atm_tc_leaf() argument 105 pr_debug("atm_tc_leaf(sch %p,flow %p)\n", sch, flow); in atm_tc_leaf() 109 static unsigned long atm_tc_get(struct Qdisc *sch, u32 classid) in atm_tc_get() argument 111 struct atm_qdisc_data *p __maybe_unused = qdisc_priv(sch); in atm_tc_get() 114 pr_debug("atm_tc_get(sch %p,[qdisc %p],classid %x)\n", sch, p, classid); in atm_tc_get() [all …]
|
D | sch_drr.c | 42 static struct drr_class *drr_find_class(struct Qdisc *sch, u32 classid) in drr_find_class() argument 44 struct drr_sched *q = qdisc_priv(sch); in drr_find_class() 65 static int drr_change_class(struct Qdisc *sch, u32 classid, u32 parentid, in drr_change_class() argument 68 struct drr_sched *q = qdisc_priv(sch); in drr_change_class() 87 quantum = psched_mtu(qdisc_dev(sch)); in drr_change_class() 93 qdisc_root_sleeping_lock(sch), in drr_change_class() 99 sch_tree_lock(sch); in drr_change_class() 102 sch_tree_unlock(sch); in drr_change_class() 114 cl->qdisc = qdisc_create_dflt(sch->dev_queue, in drr_change_class() 121 qdisc_root_sleeping_lock(sch), in drr_change_class() [all …]
|
D | sch_plug.c | 89 static int plug_enqueue(struct sk_buff *skb, struct Qdisc *sch) in plug_enqueue() argument 91 struct plug_sched_data *q = qdisc_priv(sch); in plug_enqueue() 93 if (likely(sch->qstats.backlog + skb->len <= q->limit)) { in plug_enqueue() 96 return qdisc_enqueue_tail(skb, sch); in plug_enqueue() 99 return qdisc_reshape_fail(skb, sch); in plug_enqueue() 102 static struct sk_buff *plug_dequeue(struct Qdisc *sch) in plug_dequeue() argument 104 struct plug_sched_data *q = qdisc_priv(sch); in plug_dequeue() 106 if (qdisc_is_throttled(sch)) in plug_dequeue() 114 qdisc_throttled(sch); in plug_dequeue() 120 return qdisc_dequeue_head(sch); in plug_dequeue() [all …]
|
D | sch_mqprio.c | 27 static void mqprio_destroy(struct Qdisc *sch) in mqprio_destroy() argument 29 struct net_device *dev = qdisc_dev(sch); in mqprio_destroy() 30 struct mqprio_sched *priv = qdisc_priv(sch); in mqprio_destroy() 92 static int mqprio_init(struct Qdisc *sch, struct nlattr *opt) in mqprio_init() argument 94 struct net_device *dev = qdisc_dev(sch); in mqprio_init() 95 struct mqprio_sched *priv = qdisc_priv(sch); in mqprio_init() 104 if (sch->parent != TC_H_ROOT) in mqprio_init() 128 TC_H_MAKE(TC_H_MAJ(sch->handle), in mqprio_init() 158 sch->flags |= TCQ_F_MQROOT; in mqprio_init() 162 mqprio_destroy(sch); in mqprio_init() [all …]
|
D | sch_pie.c | 98 static bool drop_early(struct Qdisc *sch, u32 packet_size) in drop_early() argument 100 struct pie_sched_data *q = qdisc_priv(sch); in drop_early() 103 u32 mtu = psched_mtu(qdisc_dev(sch)); in drop_early() 119 if (sch->qstats.backlog < 2 * mtu) in drop_early() 137 static int pie_qdisc_enqueue(struct sk_buff *skb, struct Qdisc *sch) in pie_qdisc_enqueue() argument 139 struct pie_sched_data *q = qdisc_priv(sch); in pie_qdisc_enqueue() 142 if (unlikely(qdisc_qlen(sch) >= sch->limit)) { in pie_qdisc_enqueue() 147 if (!drop_early(sch, skb->len)) { in pie_qdisc_enqueue() 161 if (qdisc_qlen(sch) > q->stats.maxq) in pie_qdisc_enqueue() 162 q->stats.maxq = qdisc_qlen(sch); in pie_qdisc_enqueue() [all …]
|
D | sch_fq_codel.c | 84 static unsigned int fq_codel_classify(struct sk_buff *skb, struct Qdisc *sch, in fq_codel_classify() argument 87 struct fq_codel_sched_data *q = qdisc_priv(sch); in fq_codel_classify() 92 if (TC_H_MAJ(skb->priority) == sch->handle && in fq_codel_classify() 143 static unsigned int fq_codel_drop(struct Qdisc *sch) in fq_codel_drop() argument 145 struct fq_codel_sched_data *q = qdisc_priv(sch); in fq_codel_drop() 165 sch->q.qlen--; in fq_codel_drop() 166 qdisc_qstats_drop(sch); in fq_codel_drop() 167 qdisc_qstats_backlog_dec(sch, skb); in fq_codel_drop() 173 static int fq_codel_enqueue(struct sk_buff *skb, struct Qdisc *sch) in fq_codel_enqueue() argument 175 struct fq_codel_sched_data *q = qdisc_priv(sch); in fq_codel_enqueue() [all …]
|
D | sch_tbf.c | 158 static int tbf_segment(struct sk_buff *skb, struct Qdisc *sch) in tbf_segment() argument 160 struct tbf_sched_data *q = qdisc_priv(sch); in tbf_segment() 168 return qdisc_reshape_fail(skb, sch); in tbf_segment() 178 qdisc_qstats_drop(sch); in tbf_segment() 184 sch->q.qlen += nb; in tbf_segment() 186 qdisc_tree_decrease_qlen(sch, 1 - nb); in tbf_segment() 191 static int tbf_enqueue(struct sk_buff *skb, struct Qdisc *sch) in tbf_enqueue() argument 193 struct tbf_sched_data *q = qdisc_priv(sch); in tbf_enqueue() 198 return tbf_segment(skb, sch); in tbf_enqueue() 199 return qdisc_reshape_fail(skb, sch); in tbf_enqueue() [all …]
|
D | sch_choke.c | 118 static void choke_drop_by_idx(struct Qdisc *sch, unsigned int idx) in choke_drop_by_idx() argument 120 struct choke_sched_data *q = qdisc_priv(sch); in choke_drop_by_idx() 130 qdisc_qstats_backlog_dec(sch, skb); in choke_drop_by_idx() 131 qdisc_drop(skb, sch); in choke_drop_by_idx() 132 qdisc_tree_decrease_qlen(sch, 1); in choke_drop_by_idx() 133 --sch->q.qlen; in choke_drop_by_idx() 201 struct Qdisc *sch, int *qerr) in choke_classify() argument 204 struct choke_sched_data *q = qdisc_priv(sch); in choke_classify() 270 static int choke_enqueue(struct sk_buff *skb, struct Qdisc *sch) in choke_enqueue() argument 273 struct choke_sched_data *q = qdisc_priv(sch); in choke_enqueue() [all …]
|
D | sch_ingress.c | 25 static struct Qdisc *ingress_leaf(struct Qdisc *sch, unsigned long arg) in ingress_leaf() argument 30 static unsigned long ingress_get(struct Qdisc *sch, u32 classid) in ingress_get() argument 35 static unsigned long ingress_bind_filter(struct Qdisc *sch, in ingress_bind_filter() argument 38 return ingress_get(sch, classid); in ingress_bind_filter() 41 static void ingress_put(struct Qdisc *sch, unsigned long cl) in ingress_put() argument 45 static void ingress_walk(struct Qdisc *sch, struct qdisc_walker *walker) in ingress_walk() argument 49 static struct tcf_proto __rcu **ingress_find_tcf(struct Qdisc *sch, in ingress_find_tcf() argument 52 struct ingress_qdisc_data *p = qdisc_priv(sch); in ingress_find_tcf() 59 static int ingress_enqueue(struct sk_buff *skb, struct Qdisc *sch) in ingress_enqueue() argument 61 struct ingress_qdisc_data *p = qdisc_priv(sch); in ingress_enqueue() [all …]
|
D | sch_sfq.c | 185 static unsigned int sfq_classify(struct sk_buff *skb, struct Qdisc *sch, in sfq_classify() argument 188 struct sfq_sched_data *q = qdisc_priv(sch); in sfq_classify() 193 if (TC_H_MAJ(skb->priority) == sch->handle && in sfq_classify() 315 static unsigned int sfq_drop(struct Qdisc *sch) in sfq_drop() argument 317 struct sfq_sched_data *q = qdisc_priv(sch); in sfq_drop() 333 sch->q.qlen--; in sfq_drop() 334 qdisc_qstats_drop(sch); in sfq_drop() 335 qdisc_qstats_backlog_dec(sch, skb); in sfq_drop() 369 sfq_enqueue(struct sk_buff *skb, struct Qdisc *sch) in sfq_enqueue() argument 371 struct sfq_sched_data *q = qdisc_priv(sch); in sfq_enqueue() [all …]
|
D | sch_gred.c | 92 static inline int gred_wred_mode_check(struct Qdisc *sch) in gred_wred_mode_check() argument 94 struct gred_sched *table = qdisc_priv(sch); in gred_wred_mode_check() 115 struct Qdisc *sch) in gred_backlog() argument 118 return sch->qstats.backlog; in gred_backlog() 152 static int gred_enqueue(struct sk_buff *skb, struct Qdisc *sch) in gred_enqueue() argument 155 struct gred_sched *t = qdisc_priv(sch); in gred_enqueue() 168 if (skb_queue_len(&sch->q) < qdisc_dev(sch)->tx_queue_len) in gred_enqueue() 169 return qdisc_enqueue_tail(skb, sch); in gred_enqueue() 199 gred_backlog(t, q, sch)); in gred_enqueue() 212 qdisc_qstats_overlimit(sch); in gred_enqueue() [all …]
|
D | sch_hhf.c | 263 static enum wdrr_bucket_idx hhf_classify(struct sk_buff *skb, struct Qdisc *sch) in hhf_classify() argument 265 struct hhf_sched_data *q = qdisc_priv(sch); in hhf_classify() 365 static unsigned int hhf_drop(struct Qdisc *sch) in hhf_drop() argument 367 struct hhf_sched_data *q = qdisc_priv(sch); in hhf_drop() 378 sch->q.qlen--; in hhf_drop() 379 qdisc_qstats_drop(sch); in hhf_drop() 380 qdisc_qstats_backlog_dec(sch, skb); in hhf_drop() 388 static int hhf_enqueue(struct sk_buff *skb, struct Qdisc *sch) in hhf_enqueue() argument 390 struct hhf_sched_data *q = qdisc_priv(sch); in hhf_enqueue() 394 idx = hhf_classify(skb, sch); in hhf_enqueue() [all …]
|
D | sch_sfb.c | 279 static int sfb_enqueue(struct sk_buff *skb, struct Qdisc *sch) in sfb_enqueue() argument 282 struct sfb_sched_data *q = qdisc_priv(sch); in sfb_enqueue() 292 if (unlikely(sch->q.qlen >= q->limit)) { in sfb_enqueue() 293 qdisc_qstats_overlimit(sch); in sfb_enqueue() 351 qdisc_qstats_overlimit(sch); in sfb_enqueue() 379 qdisc_qstats_overlimit(sch); in sfb_enqueue() 410 sch->q.qlen++; in sfb_enqueue() 414 qdisc_qstats_drop(sch); in sfb_enqueue() 419 qdisc_drop(skb, sch); in sfb_enqueue() 423 qdisc_qstats_drop(sch); in sfb_enqueue() [all …]
|
D | sch_netem.c | 362 static void tfifo_reset(struct Qdisc *sch) in tfifo_reset() argument 364 struct netem_sched_data *q = qdisc_priv(sch); in tfifo_reset() 377 static void tfifo_enqueue(struct sk_buff *nskb, struct Qdisc *sch) in tfifo_enqueue() argument 379 struct netem_sched_data *q = qdisc_priv(sch); in tfifo_enqueue() 395 sch->q.qlen++; in tfifo_enqueue() 404 static int netem_enqueue(struct sk_buff *skb, struct Qdisc *sch) in netem_enqueue() argument 406 struct netem_sched_data *q = qdisc_priv(sch); in netem_enqueue() 419 qdisc_qstats_drop(sch); /* mark packet */ in netem_enqueue() 424 qdisc_qstats_drop(sch); in netem_enqueue() 441 struct Qdisc *rootq = qdisc_root(sch); in netem_enqueue() [all …]
|
D | sch_fq.c | 296 static struct sk_buff *fq_dequeue_head(struct Qdisc *sch, struct fq_flow *flow) in fq_dequeue_head() argument 304 qdisc_qstats_backlog_dec(sch, skb); in fq_dequeue_head() 305 sch->q.qlen--; in fq_dequeue_head() 368 static int fq_enqueue(struct sk_buff *skb, struct Qdisc *sch) in fq_enqueue() argument 370 struct fq_sched_data *q = qdisc_priv(sch); in fq_enqueue() 373 if (unlikely(sch->q.qlen >= sch->limit)) in fq_enqueue() 374 return qdisc_drop(skb, sch); in fq_enqueue() 379 return qdisc_drop(skb, sch); in fq_enqueue() 385 qdisc_qstats_backlog_inc(sch, skb); in fq_enqueue() 399 sch->q.qlen++; in fq_enqueue() [all …]
|
D | sch_hfsc.c | 879 qdisc_peek_len(struct Qdisc *sch) in qdisc_peek_len() argument 884 skb = sch->ops->peek(sch); in qdisc_peek_len() 886 qdisc_warn_nonwc("qdisc_peek_len", sch); in qdisc_peek_len() 895 hfsc_purge_queue(struct Qdisc *sch, struct hfsc_class *cl) in hfsc_purge_queue() argument 920 hfsc_find_class(u32 classid, struct Qdisc *sch) in hfsc_find_class() argument 922 struct hfsc_sched *q = qdisc_priv(sch); in hfsc_find_class() 969 hfsc_change_class(struct Qdisc *sch, u32 classid, u32 parentid, in hfsc_change_class() argument 972 struct hfsc_sched *q = qdisc_priv(sch); in hfsc_change_class() 1017 spinlock_t *lock = qdisc_root_sleeping_lock(sch); in hfsc_change_class() 1027 sch_tree_lock(sch); in hfsc_change_class() [all …]
|
D | sch_qfq.c | 210 static struct qfq_class *qfq_find_class(struct Qdisc *sch, u32 classid) in qfq_find_class() argument 212 struct qfq_sched *q = qdisc_priv(sch); in qfq_find_class() 387 static int qfq_change_agg(struct Qdisc *sch, struct qfq_class *cl, u32 weight, in qfq_change_agg() argument 390 struct qfq_sched *q = qdisc_priv(sch); in qfq_change_agg() 405 static int qfq_change_class(struct Qdisc *sch, u32 classid, u32 parentid, in qfq_change_class() argument 408 struct qfq_sched *q = qdisc_priv(sch); in qfq_change_class() 442 lmax = psched_mtu(qdisc_dev(sch)); in qfq_change_class() 464 qdisc_root_sleeping_lock(sch), in qfq_change_class() 482 cl->qdisc = qdisc_create_dflt(sch->dev_queue, in qfq_change_class() 490 qdisc_root_sleeping_lock(sch), in qfq_change_class() [all …]
|
D | sch_cbq.c | 217 cbq_classify(struct sk_buff *skb, struct Qdisc *sch, int *qerr) in cbq_classify() argument 219 struct cbq_sched_data *q = qdisc_priv(sch); in cbq_classify() 230 if (TC_H_MAJ(prio ^ sch->handle) == 0 && in cbq_classify() 369 cbq_enqueue(struct sk_buff *skb, struct Qdisc *sch) in cbq_enqueue() argument 371 struct cbq_sched_data *q = qdisc_priv(sch); in cbq_enqueue() 373 struct cbq_class *cl = cbq_classify(skb, sch, &ret); in cbq_enqueue() 380 qdisc_qstats_drop(sch); in cbq_enqueue() 386 cl->q->__parent = sch; in cbq_enqueue() 390 sch->q.qlen++; in cbq_enqueue() 398 qdisc_qstats_drop(sch); in cbq_enqueue() [all …]
|
D | sch_htb.c | 181 static inline struct htb_class *htb_find(u32 handle, struct Qdisc *sch) in htb_find() argument 183 struct htb_sched *q = qdisc_priv(sch); in htb_find() 206 static struct htb_class *htb_classify(struct sk_buff *skb, struct Qdisc *sch, in htb_classify() argument 209 struct htb_sched *q = qdisc_priv(sch); in htb_classify() 219 if (skb->priority == sch->handle) in htb_classify() 221 cl = htb_find(skb->priority, sch); in htb_classify() 244 if (res.classid == sch->handle) in htb_classify() 246 cl = htb_find(res.classid, sch); in htb_classify() 257 cl = htb_find(TC_H_MAKE(TC_H_MAJ(sch->handle), q->defcls), sch); in htb_classify() 572 static int htb_enqueue(struct sk_buff *skb, struct Qdisc *sch) in htb_enqueue() argument [all …]
|
D | sch_teql.c | 80 teql_enqueue(struct sk_buff *skb, struct Qdisc *sch) in teql_enqueue() argument 82 struct net_device *dev = qdisc_dev(sch); in teql_enqueue() 83 struct teql_sched_data *q = qdisc_priv(sch); in teql_enqueue() 90 return qdisc_drop(skb, sch); in teql_enqueue() 94 teql_dequeue(struct Qdisc *sch) in teql_dequeue() argument 96 struct teql_sched_data *dat = qdisc_priv(sch); in teql_dequeue() 108 dat->m->slaves = sch; in teql_dequeue() 112 qdisc_bstats_update(sch, skb); in teql_dequeue() 114 sch->q.qlen = dat->q.qlen + q->q.qlen; in teql_dequeue() 119 teql_peek(struct Qdisc *sch) in teql_peek() argument [all …]
|
D | sch_api.c | 651 void qdisc_class_hash_grow(struct Qdisc *sch, struct Qdisc_class_hash *clhash) in qdisc_class_hash_grow() argument 671 sch_tree_lock(sch); in qdisc_class_hash_grow() 681 sch_tree_unlock(sch); in qdisc_class_hash_grow() 747 void qdisc_tree_decrease_qlen(struct Qdisc *sch, unsigned int n) in qdisc_tree_decrease_qlen() argument 758 while ((parentid = sch->parent)) { in qdisc_tree_decrease_qlen() 762 if (sch->flags & TCQ_F_NOPARENT) in qdisc_tree_decrease_qlen() 765 sch = qdisc_lookup(qdisc_dev(sch), TC_H_MAJ(parentid)); in qdisc_tree_decrease_qlen() 766 if (sch == NULL) { in qdisc_tree_decrease_qlen() 770 cops = sch->ops->cl_ops; in qdisc_tree_decrease_qlen() 772 cl = cops->get(sch, parentid); in qdisc_tree_decrease_qlen() [all …]
|
D | sch_generic.c | 585 struct Qdisc *sch; in qdisc_alloc() local 586 unsigned int size = QDISC_ALIGN(sizeof(*sch)) + ops->priv_size; in qdisc_alloc() 595 sch = (struct Qdisc *) QDISC_ALIGN((unsigned long) p); in qdisc_alloc() 597 if (sch != p) { in qdisc_alloc() 603 sch = (struct Qdisc *) QDISC_ALIGN((unsigned long) p); in qdisc_alloc() 604 sch->padded = (char *) sch - (char *) p; in qdisc_alloc() 606 INIT_LIST_HEAD(&sch->list); in qdisc_alloc() 607 skb_queue_head_init(&sch->q); in qdisc_alloc() 609 spin_lock_init(&sch->busylock); in qdisc_alloc() 610 lockdep_set_class(&sch->busylock, in qdisc_alloc() [all …]
|
D | sch_blackhole.c | 20 static int blackhole_enqueue(struct sk_buff *skb, struct Qdisc *sch) in blackhole_enqueue() argument 22 qdisc_drop(skb, sch); in blackhole_enqueue() 26 static struct sk_buff *blackhole_dequeue(struct Qdisc *sch) in blackhole_dequeue() argument
|
/linux-4.1.27/include/net/ |
D | sch_generic.h | 491 const struct Qdisc *sch) in qdisc_calculate_pkt_len() argument 494 struct qdisc_size_table *stab = rcu_dereference_bh(sch->stab); in qdisc_calculate_pkt_len() 501 static inline int qdisc_enqueue(struct sk_buff *skb, struct Qdisc *sch) in qdisc_enqueue() argument 503 qdisc_calculate_pkt_len(skb, sch); in qdisc_enqueue() 504 return sch->enqueue(skb, sch); in qdisc_enqueue() 507 static inline int qdisc_enqueue_root(struct sk_buff *skb, struct Qdisc *sch) in qdisc_enqueue_root() argument 510 return qdisc_enqueue(skb, sch) & NET_XMIT_MASK; in qdisc_enqueue_root() 525 static inline void qdisc_bstats_update_cpu(struct Qdisc *sch, in qdisc_bstats_update_cpu() argument 529 this_cpu_ptr(sch->cpu_bstats); in qdisc_bstats_update_cpu() 536 static inline void qdisc_bstats_update(struct Qdisc *sch, in qdisc_bstats_update() argument [all …]
|
D | codel.h | 172 const struct Qdisc *sch) in codel_params_init() argument 176 params->mtu = psched_mtu(qdisc_dev(sch)); in codel_params_init() 221 struct Qdisc *sch, in codel_should_drop() argument 235 sch->qstats.backlog -= qdisc_pkt_len(skb); in codel_should_drop() 241 sch->qstats.backlog <= params->mtu) { in codel_should_drop() 259 struct Qdisc *sch); 261 static struct sk_buff *codel_dequeue(struct Qdisc *sch, 267 struct sk_buff *skb = dequeue_func(vars, sch); 276 drop = codel_should_drop(skb, sch, vars, params, stats, now); 304 qdisc_drop(skb, sch); [all …]
|
D | pkt_sched.h | 84 struct Qdisc *fifo_create_dflt(struct Qdisc *sch, struct Qdisc_ops *ops,
|
/linux-4.1.27/net/netfilter/ |
D | xt_sctp.c | 45 const sctp_chunkhdr_t *sch; in match_packet() local 59 sch = skb_header_pointer(skb, offset, sizeof(_sch), &_sch); in match_packet() 60 if (sch == NULL || sch->length == 0) { in match_packet() 68 ++i, offset, sch->type, htons(sch->length), in match_packet() 69 sch->flags); in match_packet() 71 offset += WORD_ROUND(ntohs(sch->length)); in match_packet() 75 if (SCTP_CHUNKMAP_IS_SET(info->chunkmap, sch->type)) { in match_packet() 79 sch->type, sch->flags)) { in match_packet() 86 sch->type, sch->flags)) in match_packet() 87 SCTP_CHUNKMAP_CLEAR(chunkmapcopy, sch->type); in match_packet() [all …]
|
D | nf_conntrack_proto_sctp.c | 189 #define for_each_sctp_chunk(skb, sch, _sch, offset, dataoff, count) \ argument 192 ((sch) = skb_header_pointer((skb), (offset), sizeof(_sch), &(_sch))); \ 193 (offset) += (ntohs((sch)->length) + 3) & ~3, (count)++) 202 sctp_chunkhdr_t _sch, *sch; in do_basic_checks() local 207 for_each_sctp_chunk (skb, sch, _sch, offset, dataoff, count) { in do_basic_checks() 208 pr_debug("Chunk Num: %d Type: %d\n", count, sch->type); in do_basic_checks() 210 if (sch->type == SCTP_CID_INIT || in do_basic_checks() 211 sch->type == SCTP_CID_INIT_ACK || in do_basic_checks() 212 sch->type == SCTP_CID_SHUTDOWN_COMPLETE) in do_basic_checks() 220 if (((sch->type == SCTP_CID_COOKIE_ACK || in do_basic_checks() [all …]
|
/linux-4.1.27/net/netfilter/ipvs/ |
D | ip_vs_proto_sctp.c | 19 sctp_chunkhdr_t _schunkh, *sch; in sctp_conn_schedule() local 28 sch = skb_header_pointer(skb, iph->len + sizeof(sctp_sctphdr_t), in sctp_conn_schedule() 30 if (sch == NULL) { in sctp_conn_schedule() 38 if ((sch->type == SCTP_CID_INIT || sysctl_sloppy_sctp(ipvs)) && in sctp_conn_schedule() 372 sctp_chunkhdr_t _sctpch, *sch; in set_sctp_state() local 384 sch = skb_header_pointer(skb, cofs, sizeof(_sctpch), &_sctpch); in set_sctp_state() 385 if (sch == NULL) in set_sctp_state() 388 chunk_type = sch->type; in set_sctp_state() 400 if ((sch->type == SCTP_CID_COOKIE_ECHO) || in set_sctp_state() 401 (sch->type == SCTP_CID_COOKIE_ACK)) { in set_sctp_state() [all …]
|
D | ip_vs_core.c | 1021 sctp_chunkhdr_t *sch, schunk; in is_sctp_abort() local 1022 sch = skb_header_pointer(skb, nh_len + sizeof(sctp_sctphdr_t), in is_sctp_abort() 1024 if (sch == NULL) in is_sctp_abort() 1026 if (sch->type == SCTP_CID_ABORT) in is_sctp_abort() 1054 sctp_chunkhdr_t *sch, schunk; in is_new_conn() local 1056 sch = skb_header_pointer(skb, iph->len + sizeof(sctp_sctphdr_t), in is_new_conn() 1058 if (sch == NULL) in is_new_conn() 1060 return sch->type == SCTP_CID_INIT; in is_new_conn()
|
/linux-4.1.27/arch/mips/include/asm/netlogic/xlp-hal/ |
D | pic.h | 229 int sch, int vec, int dt, int db, int cpu) in nlm_9xx_pic_write_irt() argument 243 int sch, int vec, int dt, int db, int dte) in nlm_pic_write_irt() argument 248 ((sch & 0x1) << 28) | ((vec & 0x3f) << 20) | in nlm_pic_write_irt() 257 int sch, int vec, int cpu) in nlm_pic_write_irt_direct() argument 260 nlm_9xx_pic_write_irt(base, irt_num, en, nmi, sch, vec, in nlm_pic_write_irt_direct() 263 nlm_pic_write_irt(base, irt_num, en, nmi, sch, vec, 1, in nlm_pic_write_irt_direct()
|
/linux-4.1.27/arch/s390/include/asm/ |
D | qdio.h | 289 u16 sch; member
|
/linux-4.1.27/drivers/net/wireless/iwlwifi/pcie/ |
D | internal.h | 66 u32 sch; member
|
D | rx.c | 982 isr_stats->sch++; in iwl_pcie_irq_handler()
|
D | trans.c | 1905 isr_stats->sch); in iwl_dbgfs_interrupt_read()
|
/linux-4.1.27/drivers/net/wireless/iwlegacy/ |
D | debug.c | 707 il->isr_stats.sch); in il_dbgfs_interrupt_read()
|
D | common.h | 1024 u32 sch; member
|
D | 3945-mac.c | 1465 il->isr_stats.sch++; in il3945_irq_tasklet()
|
D | 4965-mac.c | 4428 il->isr_stats.sch++; in il4965_irq_tasklet()
|
/linux-4.1.27/arch/mips/include/asm/octeon/ |
D | cvmx-mio-defs.h | 2340 uint64_t sch:4; member 2350 uint64_t sch:4;
|
/linux-4.1.27/ |
D | CREDITS | 2057 M: seasons@falcon.sch.bme.hu
|