Searched refs:evt (Results 1 - 200 of 210) sorted by relevance

12

/linux-4.1.27/drivers/clocksource/
H A Dmtk_timer.c68 static void mtk_clkevt_time_stop(struct mtk_clock_event_device *evt, u8 timer) mtk_clkevt_time_stop() argument
72 val = readl(evt->gpt_base + TIMER_CTRL_REG(timer)); mtk_clkevt_time_stop()
73 writel(val & ~TIMER_CTRL_ENABLE, evt->gpt_base + mtk_clkevt_time_stop()
77 static void mtk_clkevt_time_setup(struct mtk_clock_event_device *evt, mtk_clkevt_time_setup() argument
80 writel(delay, evt->gpt_base + TIMER_CMP_REG(timer)); mtk_clkevt_time_setup()
83 static void mtk_clkevt_time_start(struct mtk_clock_event_device *evt, mtk_clkevt_time_start() argument
89 writel(GPT_IRQ_ACK(timer), evt->gpt_base + GPT_IRQ_ACK_REG); mtk_clkevt_time_start()
91 val = readl(evt->gpt_base + TIMER_CTRL_REG(timer)); mtk_clkevt_time_start()
102 evt->gpt_base + TIMER_CTRL_REG(timer)); mtk_clkevt_time_start()
108 struct mtk_clock_event_device *evt = to_mtk_clk(clk); mtk_clkevt_mode() local
110 mtk_clkevt_time_stop(evt, GPT_CLK_EVT); mtk_clkevt_mode()
114 mtk_clkevt_time_setup(evt, evt->ticks_per_jiffy, GPT_CLK_EVT); mtk_clkevt_mode()
115 mtk_clkevt_time_start(evt, true, GPT_CLK_EVT); mtk_clkevt_mode()
131 struct mtk_clock_event_device *evt = to_mtk_clk(clk); mtk_clkevt_next_event() local
133 mtk_clkevt_time_stop(evt, GPT_CLK_EVT); mtk_clkevt_next_event()
134 mtk_clkevt_time_setup(evt, event, GPT_CLK_EVT); mtk_clkevt_next_event()
135 mtk_clkevt_time_start(evt, false, GPT_CLK_EVT); mtk_clkevt_next_event()
142 struct mtk_clock_event_device *evt = dev_id; mtk_timer_interrupt() local
145 writel(GPT_IRQ_ACK(GPT_CLK_EVT), evt->gpt_base + GPT_IRQ_ACK_REG); mtk_timer_interrupt()
146 evt->dev.event_handler(&evt->dev); mtk_timer_interrupt()
151 static void mtk_timer_global_reset(struct mtk_clock_event_device *evt) mtk_timer_global_reset() argument
154 writel(0x0, evt->gpt_base + GPT_IRQ_EN_REG); mtk_timer_global_reset()
156 writel(0x3f, evt->gpt_base + GPT_IRQ_ACK_REG); mtk_timer_global_reset()
160 mtk_timer_setup(struct mtk_clock_event_device *evt, u8 timer, u8 option) mtk_timer_setup() argument
163 evt->gpt_base + TIMER_CTRL_REG(timer)); mtk_timer_setup()
166 evt->gpt_base + TIMER_CLK_REG(timer)); mtk_timer_setup()
168 writel(0x0, evt->gpt_base + TIMER_CMP_REG(timer)); mtk_timer_setup()
171 evt->gpt_base + TIMER_CTRL_REG(timer)); mtk_timer_setup()
174 static void mtk_timer_enable_irq(struct mtk_clock_event_device *evt, u8 timer) mtk_timer_enable_irq() argument
178 val = readl(evt->gpt_base + GPT_IRQ_EN_REG); mtk_timer_enable_irq()
180 evt->gpt_base + GPT_IRQ_EN_REG); mtk_timer_enable_irq()
185 struct mtk_clock_event_device *evt; mtk_timer_init() local
190 evt = kzalloc(sizeof(*evt), GFP_KERNEL); mtk_timer_init()
191 if (!evt) { mtk_timer_init()
196 evt->dev.name = "mtk_tick"; mtk_timer_init()
197 evt->dev.rating = 300; mtk_timer_init()
198 evt->dev.features = CLOCK_EVT_FEAT_PERIODIC | CLOCK_EVT_FEAT_ONESHOT; mtk_timer_init()
199 evt->dev.set_mode = mtk_clkevt_mode; mtk_timer_init()
200 evt->dev.set_next_event = mtk_clkevt_next_event; mtk_timer_init()
201 evt->dev.cpumask = cpu_possible_mask; mtk_timer_init()
203 evt->gpt_base = of_io_request_and_map(node, 0, "mtk-timer"); mtk_timer_init()
204 if (IS_ERR(evt->gpt_base)) { mtk_timer_init()
209 evt->dev.irq = irq_of_parse_and_map(node, 0); mtk_timer_init()
210 if (evt->dev.irq <= 0) { mtk_timer_init()
227 mtk_timer_global_reset(evt); mtk_timer_init()
229 if (request_irq(evt->dev.irq, mtk_timer_interrupt, mtk_timer_init()
230 IRQF_TIMER | IRQF_IRQPOLL, "mtk_timer", evt)) { mtk_timer_init()
231 pr_warn("failed to setup irq %d\n", evt->dev.irq); mtk_timer_init()
235 evt->ticks_per_jiffy = DIV_ROUND_UP(rate, HZ); mtk_timer_init()
238 mtk_timer_setup(evt, GPT_CLK_SRC, TIMER_CTRL_OP_FREERUN); mtk_timer_init()
239 clocksource_mmio_init(evt->gpt_base + TIMER_CNT_REG(GPT_CLK_SRC), mtk_timer_init()
243 mtk_timer_setup(evt, GPT_CLK_EVT, TIMER_CTRL_OP_REPEAT); mtk_timer_init()
244 clockevents_config_and_register(&evt->dev, rate, 0x3, mtk_timer_init()
247 mtk_timer_enable_irq(evt, GPT_CLK_EVT); mtk_timer_init()
256 irq_dispose_mapping(evt->dev.irq); mtk_timer_init()
258 iounmap(evt->gpt_base); mtk_timer_init()
H A Ddummy_timer.c20 struct clock_event_device *evt) dummy_timer_set_mode()
31 struct clock_event_device *evt = raw_cpu_ptr(&dummy_timer_evt); dummy_timer_setup() local
33 evt->name = "dummy_timer"; dummy_timer_setup()
34 evt->features = CLOCK_EVT_FEAT_PERIODIC | dummy_timer_setup()
37 evt->rating = 100; dummy_timer_setup()
38 evt->set_mode = dummy_timer_set_mode; dummy_timer_setup()
39 evt->cpumask = cpumask_of(cpu); dummy_timer_setup()
41 clockevents_register_device(evt); dummy_timer_setup()
19 dummy_timer_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) dummy_timer_set_mode() argument
H A Dbcm2835_timer.c46 struct clock_event_device evt; member in struct:bcm2835_timer
76 struct bcm2835_timer, evt); bcm2835_time_set_next_event()
89 event_handler = ACCESS_ONCE(timer->evt.event_handler); bcm2835_time_interrupt()
91 event_handler(&timer->evt); bcm2835_time_interrupt()
129 timer->evt.name = node->name; bcm2835_timer_init()
130 timer->evt.rating = 300; bcm2835_timer_init()
131 timer->evt.features = CLOCK_EVT_FEAT_ONESHOT; bcm2835_timer_init()
132 timer->evt.set_mode = bcm2835_time_set_mode; bcm2835_timer_init()
133 timer->evt.set_next_event = bcm2835_time_set_next_event; bcm2835_timer_init()
134 timer->evt.cpumask = cpumask_of(0); bcm2835_timer_init()
143 clockevents_config_and_register(&timer->evt, freq, 0xf, 0xffffffff); bcm2835_timer_init()
H A Dexynos_mct.c86 struct clock_event_device evt; member in struct:mct_clock_event_device
284 struct clock_event_device *evt) exynos4_comp_set_next_event()
286 exynos4_mct_comp0_start(evt->mode, cycles); exynos4_comp_set_next_event()
292 struct clock_event_device *evt) exynos4_comp_set_mode()
300 (((unsigned long long) NSEC_PER_SEC / HZ * evt->mult) >> evt->shift); exynos4_comp_set_mode()
322 struct clock_event_device *evt = dev_id; exynos4_mct_comp_isr() local
326 evt->event_handler(evt); exynos4_mct_comp_isr()
384 struct clock_event_device *evt) exynos4_tick_set_next_event()
394 struct clock_event_device *evt) exynos4_tick_set_mode()
404 (((unsigned long long) NSEC_PER_SEC / HZ * evt->mult) >> evt->shift); exynos4_tick_set_mode()
418 struct clock_event_device *evt = &mevt->evt; exynos4_mct_tick_clear() local
425 if (evt->mode != CLOCK_EVT_MODE_PERIODIC) exynos4_mct_tick_clear()
440 struct clock_event_device *evt = &mevt->evt; exynos4_mct_tick_isr() local
444 evt->event_handler(evt); exynos4_mct_tick_isr()
449 static int exynos4_local_timer_setup(struct clock_event_device *evt) exynos4_local_timer_setup() argument
454 mevt = container_of(evt, struct mct_clock_event_device, evt); exynos4_local_timer_setup()
459 evt->name = mevt->name; exynos4_local_timer_setup()
460 evt->cpumask = cpumask_of(cpu); exynos4_local_timer_setup()
461 evt->set_next_event = exynos4_tick_set_next_event; exynos4_local_timer_setup()
462 evt->set_mode = exynos4_tick_set_mode; exynos4_local_timer_setup()
463 evt->features = CLOCK_EVT_FEAT_PERIODIC | CLOCK_EVT_FEAT_ONESHOT; exynos4_local_timer_setup()
464 evt->rating = 450; exynos4_local_timer_setup()
470 if (evt->irq == -1) exynos4_local_timer_setup()
473 irq_force_affinity(evt->irq, cpumask_of(cpu)); exynos4_local_timer_setup()
474 enable_irq(evt->irq); exynos4_local_timer_setup()
478 clockevents_config_and_register(evt, clk_rate / (TICK_BASE_CNT + 1), exynos4_local_timer_setup()
484 static void exynos4_local_timer_stop(struct clock_event_device *evt) exynos4_local_timer_stop() argument
486 evt->set_mode(CLOCK_EVT_MODE_UNUSED, evt); exynos4_local_timer_stop()
488 if (evt->irq != -1) exynos4_local_timer_stop()
489 disable_irq_nosync(evt->irq); exynos4_local_timer_stop()
507 exynos4_local_timer_setup(&mevt->evt); exynos4_mct_cpu_notify()
511 exynos4_local_timer_stop(&mevt->evt); exynos4_mct_cpu_notify()
556 pcpu_mevt->evt.irq = -1; for_each_possible_cpu()
568 pcpu_mevt->evt.irq = mct_irq; for_each_possible_cpu()
577 exynos4_local_timer_setup(&mevt->evt);
283 exynos4_comp_set_next_event(unsigned long cycles, struct clock_event_device *evt) exynos4_comp_set_next_event() argument
291 exynos4_comp_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) exynos4_comp_set_mode() argument
383 exynos4_tick_set_next_event(unsigned long cycles, struct clock_event_device *evt) exynos4_tick_set_next_event() argument
393 exynos4_tick_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) exynos4_tick_set_mode() argument
H A Dqcom-timer.c50 struct clock_event_device *evt = dev_id; msm_timer_interrupt() local
52 if (evt->mode == CLOCK_EVT_MODE_ONESHOT) { msm_timer_interrupt()
57 evt->event_handler(evt); msm_timer_interrupt()
62 struct clock_event_device *evt) msm_timer_set_next_event()
81 struct clock_event_device *evt) msm_timer_set_mode()
122 static int msm_local_timer_setup(struct clock_event_device *evt) msm_local_timer_setup() argument
127 evt->irq = msm_timer_irq; msm_local_timer_setup()
128 evt->name = "msm_timer"; msm_local_timer_setup()
129 evt->features = CLOCK_EVT_FEAT_ONESHOT; msm_local_timer_setup()
130 evt->rating = 200; msm_local_timer_setup()
131 evt->set_mode = msm_timer_set_mode; msm_local_timer_setup()
132 evt->set_next_event = msm_timer_set_next_event; msm_local_timer_setup()
133 evt->cpumask = cpumask_of(cpu); msm_local_timer_setup()
135 clockevents_config_and_register(evt, GPT_HZ, 4, 0xffffffff); msm_local_timer_setup()
138 enable_percpu_irq(evt->irq, IRQ_TYPE_EDGE_RISING); msm_local_timer_setup()
140 err = request_irq(evt->irq, msm_timer_interrupt, msm_local_timer_setup()
142 IRQF_TRIGGER_RISING, "gp_timer", evt); msm_local_timer_setup()
150 static void msm_local_timer_stop(struct clock_event_device *evt) msm_local_timer_stop() argument
152 evt->set_mode(CLOCK_EVT_MODE_UNUSED, evt); msm_local_timer_stop()
153 disable_percpu_irq(evt->irq); msm_local_timer_stop()
61 msm_timer_set_next_event(unsigned long cycles, struct clock_event_device *evt) msm_timer_set_next_event() argument
80 msm_timer_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) msm_timer_set_mode() argument
H A Dtime-armada-370-xp.c160 struct clock_event_device *evt = dev_id; armada_370_xp_timer_interrupt() local
163 evt->event_handler(evt); armada_370_xp_timer_interrupt()
171 static int armada_370_xp_timer_setup(struct clock_event_device *evt) armada_370_xp_timer_setup() argument
182 evt->name = "armada_370_xp_per_cpu_tick", armada_370_xp_timer_setup()
183 evt->features = CLOCK_EVT_FEAT_ONESHOT | armada_370_xp_timer_setup()
185 evt->shift = 32, armada_370_xp_timer_setup()
186 evt->rating = 300, armada_370_xp_timer_setup()
187 evt->set_next_event = armada_370_xp_clkevt_next_event, armada_370_xp_timer_setup()
188 evt->set_mode = armada_370_xp_clkevt_mode, armada_370_xp_timer_setup()
189 evt->irq = armada_370_xp_clkevt_irq; armada_370_xp_timer_setup()
190 evt->cpumask = cpumask_of(cpu); armada_370_xp_timer_setup()
192 clockevents_config_and_register(evt, timer_clk, 1, 0xfffffffe); armada_370_xp_timer_setup()
193 enable_percpu_irq(evt->irq, 0); armada_370_xp_timer_setup()
198 static void armada_370_xp_timer_stop(struct clock_event_device *evt) armada_370_xp_timer_stop() argument
200 evt->set_mode(CLOCK_EVT_MODE_UNUSED, evt); armada_370_xp_timer_stop()
201 disable_percpu_irq(evt->irq); armada_370_xp_timer_stop()
H A Dclps711x-timer.c57 struct clock_event_device *evt = dev_id; clps711x_timer_interrupt() local
59 evt->event_handler(evt); clps711x_timer_interrupt()
65 struct clock_event_device *evt) clps711x_clockevent_set_mode()
64 clps711x_clockevent_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) clps711x_clockevent_set_mode() argument
H A Dmeson6_timer.c91 static int meson6_clkevt_next_event(unsigned long evt, meson6_clkevt_next_event() argument
95 meson6_clkevt_time_setup(CED_ID, evt); meson6_clkevt_next_event()
111 struct clock_event_device *evt = (struct clock_event_device *)dev_id; meson6_timer_interrupt() local
113 evt->event_handler(evt); meson6_timer_interrupt()
H A Darm_arch_timer.c51 struct clock_event_device evt; member in struct:arch_timer
54 #define to_arch_timer(e) container_of(e, struct arch_timer, evt)
141 struct clock_event_device *evt) timer_handler()
145 ctrl = arch_timer_reg_read(access, ARCH_TIMER_REG_CTRL, evt); timer_handler()
148 arch_timer_reg_write(access, ARCH_TIMER_REG_CTRL, ctrl, evt); timer_handler()
149 evt->event_handler(evt); timer_handler()
158 struct clock_event_device *evt = dev_id; arch_timer_handler_virt() local
160 return timer_handler(ARCH_TIMER_VIRT_ACCESS, evt); arch_timer_handler_virt()
165 struct clock_event_device *evt = dev_id; arch_timer_handler_phys() local
167 return timer_handler(ARCH_TIMER_PHYS_ACCESS, evt); arch_timer_handler_phys()
172 struct clock_event_device *evt = dev_id; arch_timer_handler_phys_mem() local
174 return timer_handler(ARCH_TIMER_MEM_PHYS_ACCESS, evt); arch_timer_handler_phys_mem()
179 struct clock_event_device *evt = dev_id; arch_timer_handler_virt_mem() local
181 return timer_handler(ARCH_TIMER_MEM_VIRT_ACCESS, evt); arch_timer_handler_virt_mem()
224 static __always_inline void set_next_event(const int access, unsigned long evt, set_next_event() argument
231 arch_timer_reg_write(access, ARCH_TIMER_REG_TVAL, evt, clk); set_next_event()
235 static int arch_timer_set_next_event_virt(unsigned long evt, arch_timer_set_next_event_virt() argument
238 set_next_event(ARCH_TIMER_VIRT_ACCESS, evt, clk); arch_timer_set_next_event_virt()
242 static int arch_timer_set_next_event_phys(unsigned long evt, arch_timer_set_next_event_phys() argument
245 set_next_event(ARCH_TIMER_PHYS_ACCESS, evt, clk); arch_timer_set_next_event_phys()
249 static int arch_timer_set_next_event_virt_mem(unsigned long evt, arch_timer_set_next_event_virt_mem() argument
252 set_next_event(ARCH_TIMER_MEM_VIRT_ACCESS, evt, clk); arch_timer_set_next_event_virt_mem()
256 static int arch_timer_set_next_event_phys_mem(unsigned long evt, arch_timer_set_next_event_phys_mem() argument
259 set_next_event(ARCH_TIMER_MEM_PHYS_ACCESS, evt, clk); arch_timer_set_next_event_phys_mem()
640 t->evt.irq = irq; arch_timer_mem_register()
641 __arch_timer_setup(ARCH_MEM_TIMER, &t->evt); arch_timer_mem_register()
648 ret = request_irq(irq, func, IRQF_TIMER, "arch_mem_timer", &t->evt); arch_timer_mem_register()
140 timer_handler(const int access, struct clock_event_device *evt) timer_handler() argument
H A Dtimer-keystone.c132 struct clock_event_device *evt = dev_id; keystone_timer_interrupt() local
134 evt->event_handler(evt); keystone_timer_interrupt()
139 struct clock_event_device *evt) keystone_set_next_event()
141 return keystone_timer_config(cycles, evt->mode); keystone_set_next_event()
145 struct clock_event_device *evt) keystone_set_mode()
138 keystone_set_next_event(unsigned long cycles, struct clock_event_device *evt) keystone_set_next_event() argument
144 keystone_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) keystone_set_mode() argument
H A Dvf_pit_timer.c90 struct clock_event_device *evt) pit_set_mode()
94 pit_set_next_event(cycle_per_jiffy, evt); pit_set_mode()
107 struct clock_event_device *evt = dev_id; pit_timer_interrupt() local
117 if (likely(evt->mode == CLOCK_EVT_MODE_ONESHOT)) pit_timer_interrupt()
120 evt->event_handler(evt); pit_timer_interrupt()
89 pit_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) pit_set_mode() argument
H A Ddw_apb_timer.c41 ced_to_dw_apb_ced(struct clock_event_device *evt) ced_to_dw_apb_ced() argument
43 return container_of(evt, struct dw_apb_clock_event_device, ced); ced_to_dw_apb_ced()
89 struct clock_event_device *evt = data; dw_apb_clockevent_irq() local
90 struct dw_apb_clock_event_device *dw_ced = ced_to_dw_apb_ced(evt); dw_apb_clockevent_irq()
92 if (!evt->event_handler) { dw_apb_clockevent_irq()
100 evt->event_handler(evt); dw_apb_clockevent_irq()
114 struct clock_event_device *evt) apbt_set_mode()
118 struct dw_apb_clock_event_device *dw_ced = ced_to_dw_apb_ced(evt); apbt_set_mode()
121 cpumask_first(evt->cpumask), apbt_set_mode()
181 struct clock_event_device *evt) apbt_next_event()
184 struct dw_apb_clock_event_device *dw_ced = ced_to_dw_apb_ced(evt); apbt_next_event()
113 apbt_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) apbt_set_mode() argument
180 apbt_next_event(unsigned long delta, struct clock_event_device *evt) apbt_next_event() argument
H A Dfsl_ftm_timer.c157 struct clock_event_device *evt) ftm_set_mode()
161 ftm_set_next_event(priv->periodic_cyc, evt); ftm_set_mode()
173 struct clock_event_device *evt = dev_id; ftm_evt_interrupt() local
177 if (likely(evt->mode == CLOCK_EVT_MODE_ONESHOT)) { ftm_evt_interrupt()
182 evt->event_handler(evt); ftm_evt_interrupt()
287 freq = __ftm_clk_init(np, "ftm-evt-counter-en", "ftm-evt"); ftm_clk_init()
156 ftm_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) ftm_set_mode() argument
H A Dmxs_timer.c109 static int timrotv1_set_next_event(unsigned long evt, timrotv1_set_next_event() argument
113 __raw_writel(evt, mxs_timrot_base + HW_TIMROT_TIMCOUNTn(0)); timrotv1_set_next_event()
118 static int timrotv2_set_next_event(unsigned long evt, timrotv2_set_next_event() argument
122 __raw_writel(evt, mxs_timrot_base + HW_TIMROT_FIXED_COUNTn(0)); timrotv2_set_next_event()
129 struct clock_event_device *evt = dev_id; mxs_timer_interrupt() local
132 evt->event_handler(evt); mxs_timer_interrupt()
154 struct clock_event_device *evt) mxs_set_mode()
153 mxs_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) mxs_set_mode() argument
H A Dsun4i_timer.c105 static int sun4i_clkevt_next_event(unsigned long evt, sun4i_clkevt_next_event() argument
109 sun4i_clkevt_time_setup(0, evt - TIMER_SYNC_TICKS); sun4i_clkevt_next_event()
126 struct clock_event_device *evt = (struct clock_event_device *)dev_id; sun4i_timer_interrupt() local
129 evt->event_handler(evt); sun4i_timer_interrupt()
H A Dtimer-digicolor.c113 static int digicolor_clkevt_next_event(unsigned long evt, digicolor_clkevt_next_event() argument
117 dc_timer_set_count(ce, evt); digicolor_clkevt_next_event()
136 struct clock_event_device *evt = dev_id; digicolor_timer_interrupt() local
138 evt->event_handler(evt); digicolor_timer_interrupt()
H A Dtimer-integrator-ap.c67 struct clock_event_device *evt = dev_id; integrator_timer_interrupt() local
72 evt->event_handler(evt); integrator_timer_interrupt()
77 static void clkevt_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) clkevt_set_mode() argument
106 static int clkevt_set_next_event(unsigned long next, struct clock_event_device *evt) clkevt_set_next_event() argument
H A Dvt8500_timer.c76 struct clock_event_device *evt) vt8500_timer_set_next_event()
94 struct clock_event_device *evt) vt8500_timer_set_mode()
120 struct clock_event_device *evt = dev_id; vt8500_timer_interrupt() local
122 evt->event_handler(evt); vt8500_timer_interrupt()
75 vt8500_timer_set_next_event(unsigned long cycles, struct clock_event_device *evt) vt8500_timer_set_next_event() argument
93 vt8500_timer_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) vt8500_timer_set_mode() argument
H A Dmips-gic-timer.c24 static int gic_next_event(unsigned long delta, struct clock_event_device *evt) gic_next_event() argument
31 gic_write_cpu_compare(cnt, cpumask_first(evt->cpumask)); gic_next_event()
37 struct clock_event_device *evt) gic_set_clock_mode()
36 gic_set_clock_mode(enum clock_event_mode mode, struct clock_event_device *evt) gic_set_clock_mode() argument
H A Dmoxart_timer.c107 struct clock_event_device *evt = dev_id; moxart_timer_interrupt() local
108 evt->event_handler(evt); moxart_timer_interrupt()
H A Di8253.c109 struct clock_event_device *evt) init_pit_timer()
123 if (evt->mode == CLOCK_EVT_MODE_PERIODIC || init_pit_timer()
124 evt->mode == CLOCK_EVT_MODE_ONESHOT) { init_pit_timer()
148 static int pit_next_event(unsigned long delta, struct clock_event_device *evt) pit_next_event() argument
108 init_pit_timer(enum clock_event_mode mode, struct clock_event_device *evt) init_pit_timer() argument
H A Dmetag_generic.c60 struct clock_event_device *evt) metag_timer_set_mode()
93 struct clock_event_device *evt = this_cpu_ptr(&local_clockevent); metag_timer_interrupt() local
95 evt->event_handler(evt); metag_timer_interrupt()
59 metag_timer_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) metag_timer_set_mode() argument
H A Darm_global_timer.c132 static int gt_clockevent_set_next_event(unsigned long evt, gt_clockevent_set_next_event() argument
135 gt_compare_set(evt, 0); gt_clockevent_set_next_event()
141 struct clock_event_device *evt = dev_id; gt_clockevent_interrupt() local
158 if (evt->mode == CLOCK_EVT_MODE_ONESHOT) gt_clockevent_interrupt()
162 evt->event_handler(evt); gt_clockevent_interrupt()
H A Dasm9260_timer.c114 struct clock_event_device *evt) asm9260_timer_set_next_event()
124 struct clock_event_device *evt) asm9260_timer_set_mode()
159 struct clock_event_device *evt = dev_id; asm9260_timer_interrupt() local
161 evt->event_handler(evt); asm9260_timer_interrupt()
113 asm9260_timer_set_next_event(unsigned long delta, struct clock_event_device *evt) asm9260_timer_set_next_event() argument
123 asm9260_timer_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) asm9260_timer_set_mode() argument
H A Dtegra20_timer.c65 struct clock_event_device *evt) tegra_timer_set_next_event()
76 struct clock_event_device *evt) tegra_timer_set_mode()
151 struct clock_event_device *evt = (struct clock_event_device *)dev_id; tegra_timer_interrupt() local
153 evt->event_handler(evt); tegra_timer_interrupt()
64 tegra_timer_set_next_event(unsigned long cycles, struct clock_event_device *evt) tegra_timer_set_next_event() argument
75 tegra_timer_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) tegra_timer_set_mode() argument
H A Dbcm_kona_timer.c160 struct clock_event_device *evt = &kona_clockevent_timer; kona_timer_interrupt() local
163 evt->event_handler(evt); kona_timer_interrupt()
H A Dsamsung_pwm_timer.c189 struct clock_event_device *evt) samsung_set_next_event()
211 struct clock_event_device *evt) samsung_set_mode()
253 struct clock_event_device *evt = dev_id; samsung_clock_event_isr() local
260 evt->event_handler(evt); samsung_clock_event_isr()
188 samsung_set_next_event(unsigned long cycles, struct clock_event_device *evt) samsung_set_next_event() argument
210 samsung_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) samsung_set_mode() argument
H A Dcadence_ttc_timer.c179 * @evt: Address of clock event instance
184 struct clock_event_device *evt) ttc_set_next_event()
186 struct ttc_timer_clockevent *ttce = to_ttc_timer_clkevent(evt); ttc_set_next_event()
197 * @evt: Address of clock event instance
200 struct clock_event_device *evt) ttc_set_mode()
202 struct ttc_timer_clockevent *ttce = to_ttc_timer_clkevent(evt); ttc_set_mode()
183 ttc_set_next_event(unsigned long cycles, struct clock_event_device *evt) ttc_set_next_event() argument
199 ttc_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) ttc_set_mode() argument
H A Dtimer-u300.c196 struct clock_event_device *evt) u300_set_mode()
199 container_of(evt, struct u300_clockevent_data, cevd); u300_set_mode()
280 struct clock_event_device *evt) u300_set_next_event()
325 struct clock_event_device *evt = &u300_clockevent_data.cevd; u300_timer_interrupt() local
330 evt->event_handler(evt); u300_timer_interrupt()
195 u300_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) u300_set_mode() argument
279 u300_set_next_event(unsigned long cycles, struct clock_event_device *evt) u300_set_next_event() argument
H A Dcs5535-clockevt.c81 struct clock_event_device *evt) mfgpt_set_mode()
91 static int mfgpt_next_event(unsigned long delta, struct clock_event_device *evt) mfgpt_next_event() argument
80 mfgpt_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) mfgpt_set_mode() argument
H A Dtime-efm32.c87 static int efm32_clock_event_set_next_event(unsigned long evt, efm32_clock_event_set_next_event() argument
94 writel_relaxed(evt, ddata->base + TIMERn_CNT); efm32_clock_event_set_next_event()
H A Dnomadik-mtu.c93 static int nmdk_clkevt_next(unsigned long evt, struct clock_event_device *ev) nmdk_clkevt_next() argument
96 writel(evt, mtu_base + MTU_LR(1)); nmdk_clkevt_next()
H A Dtimer-sun5i.c129 static int sun5i_clkevt_next_event(unsigned long evt, sun5i_clkevt_next_event() argument
135 sun5i_clkevt_time_setup(ce, 0, evt - TIMER_SYNC_TICKS); sun5i_clkevt_next_event()
/linux-4.1.27/arch/powerpc/kernel/
H A Dmce.c182 struct machine_check_event evt; machine_check_queue_event() local
184 if (!get_mce_event(&evt, MCE_EVENT_RELEASE)) machine_check_queue_event()
193 memcpy(this_cpu_ptr(&mce_event_queue[index]), &evt, sizeof(evt)); machine_check_queue_event() local
219 void machine_check_print_event_info(struct machine_check_event *evt) machine_check_print_event_info() argument
246 if (evt->version != MCE_V1) { machine_check_print_event_info()
248 evt->version); machine_check_print_event_info()
251 switch (evt->severity) { machine_check_print_event_info()
272 evt->disposition == MCE_DISPOSITION_RECOVERED ? machine_check_print_event_info()
275 evt->initiator == MCE_INITIATOR_CPU ? "CPU" : "Unknown"); machine_check_print_event_info()
276 switch (evt->error_type) { machine_check_print_event_info()
278 subtype = evt->u.ue_error.ue_error_type < machine_check_print_event_info()
280 mc_ue_types[evt->u.ue_error.ue_error_type] machine_check_print_event_info()
283 if (evt->u.ue_error.effective_address_provided) machine_check_print_event_info()
285 level, evt->u.ue_error.effective_address); machine_check_print_event_info()
286 if (evt->u.ue_error.physical_address_provided) machine_check_print_event_info()
288 level, evt->u.ue_error.physical_address); machine_check_print_event_info()
291 subtype = evt->u.slb_error.slb_error_type < machine_check_print_event_info()
293 mc_slb_types[evt->u.slb_error.slb_error_type] machine_check_print_event_info()
296 if (evt->u.slb_error.effective_address_provided) machine_check_print_event_info()
298 level, evt->u.slb_error.effective_address); machine_check_print_event_info()
301 subtype = evt->u.erat_error.erat_error_type < machine_check_print_event_info()
303 mc_erat_types[evt->u.erat_error.erat_error_type] machine_check_print_event_info()
306 if (evt->u.erat_error.effective_address_provided) machine_check_print_event_info()
308 level, evt->u.erat_error.effective_address); machine_check_print_event_info()
311 subtype = evt->u.tlb_error.tlb_error_type < machine_check_print_event_info()
313 mc_tlb_types[evt->u.tlb_error.tlb_error_type] machine_check_print_event_info()
316 if (evt->u.tlb_error.effective_address_provided) machine_check_print_event_info()
318 level, evt->u.tlb_error.effective_address); machine_check_print_event_info()
327 uint64_t get_mce_fault_addr(struct machine_check_event *evt) get_mce_fault_addr() argument
329 switch (evt->error_type) { get_mce_fault_addr()
331 if (evt->u.ue_error.effective_address_provided) get_mce_fault_addr()
332 return evt->u.ue_error.effective_address; get_mce_fault_addr()
335 if (evt->u.slb_error.effective_address_provided) get_mce_fault_addr()
336 return evt->u.slb_error.effective_address; get_mce_fault_addr()
339 if (evt->u.erat_error.effective_address_provided) get_mce_fault_addr()
340 return evt->u.erat_error.effective_address; get_mce_fault_addr()
343 if (evt->u.tlb_error.effective_address_provided) get_mce_fault_addr()
344 return evt->u.tlb_error.effective_address; get_mce_fault_addr()
H A Dtime.c100 static int decrementer_set_next_event(unsigned long evt,
487 struct clock_event_device *evt = this_cpu_ptr(&decrementers); __timer_interrupt() local
500 if (evt->event_handler) __timer_interrupt()
501 evt->event_handler(evt); __timer_interrupt()
852 static int decrementer_set_next_event(unsigned long evt, decrementer_set_next_event() argument
855 __this_cpu_write(decrementers_next_tb, get_tb_or_rtc() + evt); decrementer_set_next_event()
856 set_dec(evt); decrementer_set_next_event()
/linux-4.1.27/arch/blackfin/kernel/
H A Dtime-ts.c129 struct clock_event_device *evt) bfin_gptmr0_set_next_event()
140 struct clock_event_device *evt) bfin_gptmr0_set_mode()
194 struct clock_event_device *evt = dev_id; bfin_gptmr0_interrupt() local
203 evt->event_handler(evt); bfin_gptmr0_interrupt()
223 static void __init bfin_gptmr0_clockevent_init(struct clock_event_device *evt) bfin_gptmr0_clockevent_init() argument
228 evt->mult = div_sc(clock_tick, NSEC_PER_SEC, evt->shift); bfin_gptmr0_clockevent_init()
229 evt->max_delta_ns = clockevent_delta2ns(-1, evt); bfin_gptmr0_clockevent_init()
230 evt->min_delta_ns = clockevent_delta2ns(100, evt); bfin_gptmr0_clockevent_init()
232 evt->cpumask = cpumask_of(0); bfin_gptmr0_clockevent_init()
234 clockevents_register_device(evt); bfin_gptmr0_clockevent_init()
243 struct clock_event_device *evt) bfin_coretmr_set_next_event()
254 struct clock_event_device *evt) bfin_coretmr_set_mode()
306 struct clock_event_device *evt = &per_cpu(coretmr_events, cpu); bfin_coretmr_interrupt() local
309 evt->event_handler(evt); bfin_coretmr_interrupt()
326 struct clock_event_device *evt = &per_cpu(coretmr_events, cpu); bfin_coretmr_clockevent_init() local
329 evt->broadcast = smp_timer_broadcast; bfin_coretmr_clockevent_init()
332 evt->name = "bfin_core_timer"; bfin_coretmr_clockevent_init()
333 evt->rating = 350; bfin_coretmr_clockevent_init()
334 evt->irq = -1; bfin_coretmr_clockevent_init()
335 evt->shift = 32; bfin_coretmr_clockevent_init()
336 evt->features = CLOCK_EVT_FEAT_PERIODIC | CLOCK_EVT_FEAT_ONESHOT; bfin_coretmr_clockevent_init()
337 evt->set_next_event = bfin_coretmr_set_next_event; bfin_coretmr_clockevent_init()
338 evt->set_mode = bfin_coretmr_set_mode; bfin_coretmr_clockevent_init()
341 evt->mult = div_sc(clock_tick, NSEC_PER_SEC, evt->shift); bfin_coretmr_clockevent_init()
342 evt->max_delta_ns = clockevent_delta2ns(-1, evt); bfin_coretmr_clockevent_init()
343 evt->min_delta_ns = clockevent_delta2ns(100, evt); bfin_coretmr_clockevent_init()
345 evt->cpumask = cpumask_of(cpu); bfin_coretmr_clockevent_init()
347 clockevents_register_device(evt); bfin_coretmr_clockevent_init()
128 bfin_gptmr0_set_next_event(unsigned long cycles, struct clock_event_device *evt) bfin_gptmr0_set_next_event() argument
139 bfin_gptmr0_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) bfin_gptmr0_set_mode() argument
242 bfin_coretmr_set_next_event(unsigned long cycles, struct clock_event_device *evt) bfin_coretmr_set_next_event() argument
253 bfin_coretmr_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) bfin_coretmr_set_mode() argument
H A Dearly_printk.c165 u32 evt; init_early_exception_vectors() local
188 for (evt = EVT2; evt <= EVT15; evt += 4) init_early_exception_vectors()
189 bfin_write32(evt, early_trap); init_early_exception_vectors()
/linux-4.1.27/drivers/uwb/
H A Duwbd.c85 * @evt->rc is already referenced and guaranteed to exist. See
164 * @evt: the event to handle
174 * controller in @evt->rc referenced. The reference will be dropped
179 int uwbd_event_handle_urc(struct uwb_event *evt) uwbd_event_handle_urc() argument
187 type = evt->notif.rceb->bEventType; uwbd_event_handle_urc()
188 event = le16_to_cpu(evt->notif.rceb->wEvent); uwbd_event_handle_urc()
189 context = evt->notif.rceb->bEventContext; uwbd_event_handle_urc()
202 result = (*handler)(evt); uwbd_event_handle_urc()
205 dev_err(&evt->rc->uwb_dev.dev, uwbd_event_handle_urc()
211 static void uwbd_event_handle_message(struct uwb_event *evt) uwbd_event_handle_message() argument
216 rc = evt->rc; uwbd_event_handle_message()
218 if (evt->message < 0 || evt->message >= ARRAY_SIZE(uwbd_message_handlers)) { uwbd_event_handle_message()
219 dev_err(&rc->uwb_dev.dev, "UWBD: invalid message type %d\n", evt->message); uwbd_event_handle_message()
223 result = uwbd_message_handlers[evt->message].handler(evt); uwbd_event_handle_message()
226 uwbd_message_handlers[evt->message].name, result); uwbd_event_handle_message()
229 static void uwbd_event_handle(struct uwb_event *evt) uwbd_event_handle() argument
234 rc = evt->rc; uwbd_event_handle()
237 switch (evt->type) { uwbd_event_handle()
239 should_keep = uwbd_event_handle_urc(evt); uwbd_event_handle()
241 kfree(evt->notif.rceb); uwbd_event_handle()
244 uwbd_event_handle_message(evt); uwbd_event_handle()
247 dev_err(&rc->uwb_dev.dev, "UWBD: invalid event type %d\n", evt->type); uwbd_event_handle()
271 struct uwb_event *evt; uwbd() local
286 evt = list_first_entry(&rc->uwbd.event_list, struct uwb_event, list_node); uwbd()
287 list_del(&evt->list_node); uwbd()
289 evt = NULL; uwbd()
292 if (evt) { uwbd()
293 uwbd_event_handle(evt); uwbd()
294 kfree(evt); uwbd()
333 void uwbd_event_queue(struct uwb_event *evt) uwbd_event_queue() argument
335 struct uwb_rc *rc = evt->rc; uwbd_event_queue()
340 list_add(&evt->list_node, &rc->uwbd.event_list); uwbd_event_queue()
343 __uwb_rc_put(evt->rc); uwbd_event_queue()
344 if (evt->type == UWB_EVT_TYPE_NOTIF) uwbd_event_queue()
345 kfree(evt->notif.rceb); uwbd_event_queue()
346 kfree(evt); uwbd_event_queue()
354 struct uwb_event *evt, *nxt; uwbd_flush() local
357 list_for_each_entry_safe(evt, nxt, &rc->uwbd.event_list, list_node) { uwbd_flush()
358 if (evt->rc == rc) { uwbd_flush()
360 list_del(&evt->list_node); uwbd_flush()
361 if (evt->type == UWB_EVT_TYPE_NOTIF) uwbd_flush()
362 kfree(evt->notif.rceb); uwbd_flush()
363 kfree(evt); uwbd_flush()
H A Die-rcv.c29 int uwbd_evt_handle_rc_ie_rcv(struct uwb_event *evt) uwbd_evt_handle_rc_ie_rcv() argument
32 struct device *dev = &evt->rc->uwb_dev.dev; uwbd_evt_handle_rc_ie_rcv()
36 if (evt->notif.size < sizeof(*iercv)) { uwbd_evt_handle_rc_ie_rcv()
39 evt->notif.size, sizeof(*iercv)); uwbd_evt_handle_rc_ie_rcv()
42 iercv = container_of(evt->notif.rceb, struct uwb_rc_evt_ie_rcv, rceb); uwbd_evt_handle_rc_ie_rcv()
H A Dbeacon.c361 static int uwb_verify_beacon(struct uwb_rc *rc, struct uwb_event *evt, uwb_verify_beacon() argument
369 if (evt->notif.size < sizeof(*be) + sizeof(*bf)) { uwb_verify_beacon()
371 "(%zu vs %zu bytes needed)\n", evt->notif.size, uwb_verify_beacon()
393 int uwbd_evt_handle_rc_beacon(struct uwb_event *evt) uwbd_evt_handle_rc_beacon() argument
401 rc = evt->rc; uwbd_evt_handle_rc_beacon()
402 be = container_of(evt->notif.rceb, struct uwb_rc_evt_beacon, rceb); uwbd_evt_handle_rc_beacon()
403 result = uwb_verify_beacon(rc, evt, be); uwbd_evt_handle_rc_beacon()
430 uwb_beacon_print(evt->rc, be, bf); uwbd_evt_handle_rc_beacon()
431 bce = __uwb_beca_add(rc, be, bf, evt->ts_jiffies); uwbd_evt_handle_rc_beacon()
444 bce->ts_jiffies = evt->ts_jiffies; uwbd_evt_handle_rc_beacon()
457 uwbd_dev_onair(evt->rc, bce); uwbd_evt_handle_rc_beacon()
469 int uwbd_evt_handle_rc_beacon_size(struct uwb_event *evt) uwbd_evt_handle_rc_beacon_size() argument
472 struct device *dev = &evt->rc->uwb_dev.dev; uwbd_evt_handle_rc_beacon_size()
476 if (evt->notif.size < sizeof(*bs)) { uwbd_evt_handle_rc_beacon_size()
479 evt->notif.size, sizeof(*bs)); uwbd_evt_handle_rc_beacon_size()
482 bs = container_of(evt->notif.rceb, struct uwb_rc_evt_beacon_size, rceb); uwbd_evt_handle_rc_beacon_size()
500 * @evt: the BP_SLOT_CHANGE notification from the radio controller
506 int uwbd_evt_handle_rc_bp_slot_change(struct uwb_event *evt) uwbd_evt_handle_rc_bp_slot_change() argument
508 struct uwb_rc *rc = evt->rc; uwbd_evt_handle_rc_bp_slot_change()
512 if (evt->notif.size < sizeof(*bpsc)) { uwbd_evt_handle_rc_bp_slot_change()
516 bpsc = container_of(evt->notif.rceb, struct uwb_rc_evt_bp_slot_change, rceb); uwbd_evt_handle_rc_bp_slot_change()
540 int uwbd_evt_handle_rc_bpoie_change(struct uwb_event *evt) uwbd_evt_handle_rc_bpoie_change() argument
543 struct device *dev = &evt->rc->uwb_dev.dev; uwbd_evt_handle_rc_bpoie_change()
550 if (evt->notif.size < sizeof(*bpoiec)) { uwbd_evt_handle_rc_bpoie_change()
553 evt->notif.size, sizeof(*bpoiec)); uwbd_evt_handle_rc_bpoie_change()
556 bpoiec = container_of(evt->notif.rceb, struct uwb_rc_evt_bpoie_change, rceb); uwbd_evt_handle_rc_bpoie_change()
H A Ddrp-avail.c227 int uwbd_evt_get_drp_avail(struct uwb_event *evt, unsigned long *bmp) uwbd_evt_get_drp_avail() argument
229 struct device *dev = &evt->rc->uwb_dev.dev; uwbd_evt_get_drp_avail()
234 if (evt->notif.size < sizeof(*drp_evt)) { uwbd_evt_get_drp_avail()
237 "needed]\n", evt->notif.size, sizeof(*drp_evt)); uwbd_evt_get_drp_avail()
240 drp_evt = container_of(evt->notif.rceb, struct uwb_rc_evt_drp_avail, rceb); uwbd_evt_get_drp_avail()
251 * @evt: Event information (packs the actual event data, which
271 int uwbd_evt_handle_rc_drp_avail(struct uwb_event *evt) uwbd_evt_handle_rc_drp_avail() argument
274 struct uwb_rc *rc = evt->rc; uwbd_evt_handle_rc_drp_avail()
277 result = uwbd_evt_get_drp_avail(evt, bmp); uwbd_evt_handle_rc_drp_avail()
H A Dreset.c323 int uwbd_msg_handle_reset(struct uwb_event *evt) uwbd_msg_handle_reset() argument
325 struct uwb_rc *rc = evt->rc; uwbd_msg_handle_reset()
352 struct uwb_event *evt; uwb_rc_reset_all() local
354 evt = kzalloc(sizeof(struct uwb_event), GFP_ATOMIC); uwb_rc_reset_all()
355 if (unlikely(evt == NULL)) uwb_rc_reset_all()
358 evt->rc = __uwb_rc_get(rc); /* will be put by uwbd's uwbd_event_handle() */ uwb_rc_reset_all()
359 evt->ts_jiffies = jiffies; uwb_rc_reset_all()
360 evt->type = UWB_EVT_TYPE_MSG; uwb_rc_reset_all()
361 evt->message = UWB_EVT_MSG_RESET; uwb_rc_reset_all()
363 uwbd_event_queue(evt); uwb_rc_reset_all()
H A Daddress.c167 struct uwb_rc_evt_dev_addr_mgmt evt; uwb_rc_addr_get() local
184 result = uwb_rc_dev_addr_mgmt(rc, bmOperationType, baAddr, &evt); uwb_rc_addr_get()
188 memcpy(&dev_addr->data, evt.baAddr, uwb_rc_addr_get()
192 memcpy(&mac_addr->data, evt.baAddr, uwb_rc_addr_get()
291 * @evt: the DEV_ADDR_CONFLICT notification from the radio controller
297 int uwbd_evt_handle_rc_dev_addr_conflict(struct uwb_event *evt) uwbd_evt_handle_rc_dev_addr_conflict() argument
299 struct uwb_rc *rc = evt->rc; uwbd_evt_handle_rc_dev_addr_conflict()
H A Ddrp.c775 * @evt: the DRP_IE event from the radio controller
805 int uwbd_evt_handle_rc_drp(struct uwb_event *evt) uwbd_evt_handle_rc_drp() argument
807 struct device *dev = &evt->rc->uwb_dev.dev; uwbd_evt_handle_rc_drp()
808 struct uwb_rc *rc = evt->rc; uwbd_evt_handle_rc_drp()
816 if (evt->notif.size < sizeof(*drp_evt)) { uwbd_evt_handle_rc_drp()
819 evt->notif.size, sizeof(*drp_evt)); uwbd_evt_handle_rc_drp()
822 bytes_left = evt->notif.size - sizeof(*drp_evt); uwbd_evt_handle_rc_drp()
823 drp_evt = container_of(evt->notif.rceb, struct uwb_rc_evt_drp, rceb); uwbd_evt_handle_rc_drp()
H A Dneh.c97 * evt). On reception of the notification/event, the callback (cb) is
108 __le16 evt; member in struct:uwb_rc_neh
232 neh->evt = cpu_to_le16(expected_event); uwb_rc_neh_add()
310 && neh->evt == rceb->wEvent uwb_rc_neh_match()
H A Duwb-internal.h247 int uwbd_msg_handle_reset(struct uwb_event *evt);
254 int uwbd_evt_handle_rc_dev_addr_conflict(struct uwb_event *evt);
H A Dwhc-rc.c259 dev_err(dev, "Can't allocate evt transfer buffer\n"); whcrc_setup_rc_umc()
/linux-4.1.27/drivers/net/wireless/ath/wil6210/
H A Dwmi.c285 struct wmi_ready_event *evt = d; wmi_evt_ready() local
287 wil->fw_version = le32_to_cpu(evt->sw_version); wmi_evt_ready()
288 wil->n_mids = evt->numof_additional_mids; wmi_evt_ready()
291 evt->mac, wil->n_mids); wmi_evt_ready()
388 struct wmi_connect_event *evt = d; wmi_evt_connect() local
398 if (len < sizeof(*evt)) { wmi_evt_connect()
402 if (len != sizeof(*evt) + evt->beacon_ie_len + evt->assoc_req_len + wmi_evt_connect()
403 evt->assoc_resp_len) { wmi_evt_connect()
406 len, (int)sizeof(*evt), evt->beacon_ie_len, wmi_evt_connect()
407 evt->assoc_req_len, evt->assoc_resp_len); wmi_evt_connect()
410 if (evt->cid >= WIL6210_MAX_CID) { wmi_evt_connect()
411 wil_err(wil, "Connect CID invalid : %d\n", evt->cid); wmi_evt_connect()
415 ch = evt->channel + 1; wmi_evt_connect()
417 evt->bssid, ch, evt->cid); wmi_evt_connect()
419 evt->assoc_info, len - sizeof(*evt), true); wmi_evt_connect()
422 assoc_req_ie = &evt->assoc_info[evt->beacon_ie_len + wmi_evt_connect()
424 assoc_req_ielen = evt->assoc_req_len - assoc_req_ie_offset; wmi_evt_connect()
425 if (evt->assoc_req_len <= assoc_req_ie_offset) { wmi_evt_connect()
430 assoc_resp_ie = &evt->assoc_info[evt->beacon_ie_len + wmi_evt_connect()
431 evt->assoc_req_len + wmi_evt_connect()
433 assoc_resp_ielen = evt->assoc_resp_len - assoc_resp_ie_offset; wmi_evt_connect()
434 if (evt->assoc_resp_len <= assoc_resp_ie_offset) { wmi_evt_connect()
446 cfg80211_connect_result(ndev, evt->bssid, wmi_evt_connect()
462 cfg80211_new_sta(ndev, evt->bssid, &sinfo, GFP_KERNEL); wmi_evt_connect()
469 ether_addr_copy(wil->sta[evt->cid].addr, evt->bssid); wmi_evt_connect()
470 wil->sta[evt->cid].status = wil_sta_conn_pending; wmi_evt_connect()
472 wil->pending_connect_cid = evt->cid; wmi_evt_connect()
479 struct wmi_disconnect_event *evt = d; wmi_evt_disconnect() local
480 u16 reason_code = le16_to_cpu(evt->protocol_reason_status); wmi_evt_disconnect()
483 evt->bssid, reason_code, evt->disconnect_reason); wmi_evt_disconnect()
488 wil6210_disconnect(wil, evt->bssid, reason_code, true); wmi_evt_disconnect()
500 struct wmi_eapol_rx_event *evt = d; wmi_evt_eapol_rx() local
501 u16 eapol_len = le16_to_cpu(evt->eapol_len); wmi_evt_eapol_rx()
509 evt->src_mac); wmi_evt_eapol_rx()
511 cid = wil_find_cid(wil, evt->src_mac); wmi_evt_eapol_rx()
528 ether_addr_copy(eth->h_source, evt->src_mac); wmi_evt_eapol_rx()
530 memcpy(skb_put(skb, eapol_len), evt->eapol, eapol_len); wmi_evt_eapol_rx()
564 struct wmi_data_port_open_event *evt = d; wmi_evt_linkup() local
565 u8 cid = evt->cid; wmi_evt_linkup()
582 struct wmi_wbe_link_down_event *evt = d; wmi_evt_linkdown() local
583 u8 cid = evt->cid; wmi_evt_linkdown()
586 cid, le32_to_cpu(evt->reason)); wmi_evt_linkdown()
600 struct wmi_vring_ba_status_event *evt = d; wmi_evt_ba_status() local
604 evt->ringid, wmi_evt_ba_status()
605 evt->status == WMI_BA_AGREED ? "OK" : "N/A", wmi_evt_ba_status()
606 evt->agg_wsize, __le16_to_cpu(evt->ba_timeout), wmi_evt_ba_status()
607 evt->amsdu ? "+" : "-"); wmi_evt_ba_status()
609 if (evt->ringid >= WIL6210_MAX_TX_RINGS) { wmi_evt_ba_status()
610 wil_err(wil, "invalid ring id %d\n", evt->ringid); wmi_evt_ba_status()
614 if (evt->status != WMI_BA_AGREED) { wmi_evt_ba_status()
615 evt->ba_timeout = 0; wmi_evt_ba_status()
616 evt->agg_wsize = 0; wmi_evt_ba_status()
617 evt->amsdu = 0; wmi_evt_ba_status()
620 txdata = &wil->vring_tx_data[evt->ringid]; wmi_evt_ba_status()
622 txdata->agg_timeout = le16_to_cpu(evt->ba_timeout); wmi_evt_ba_status()
623 txdata->agg_wsize = evt->agg_wsize; wmi_evt_ba_status()
624 txdata->agg_amsdu = evt->amsdu; wmi_evt_ba_status()
631 struct wmi_rcp_addba_req_event *evt = d; wmi_evt_addba_rx_req() local
633 wil_addba_rx_request(wil, evt->cidxtid, evt->dialog_token, wmi_evt_addba_rx_req()
634 evt->ba_param_set, evt->ba_timeout, wmi_evt_addba_rx_req()
635 evt->ba_seq_ctrl); wmi_evt_addba_rx_req()
641 struct wmi_delba_event *evt = d; variable in typeref:struct:wmi_delba_event
643 u16 reason = __le16_to_cpu(evt->reason);
648 parse_cidxtid(evt->cidxtid, &cid, &tid);
651 evt->from_initiator ? "originator" : "recipient",
653 if (!evt->from_initiator) {
716 struct pending_wmi_event *evt; wmi_recv_cmd() local
742 wil_err(wil, "Mbox evt not owned by FW?\n"); wmi_recv_cmd()
748 wil_err(wil, "Mbox evt at 0x%08x?\n", wmi_recv_cmd()
753 wil_dbg_wmi(wil, "Mbox evt %04x %04x %04x %02x\n", wmi_recv_cmd()
760 evt = kmalloc(ALIGN(offsetof(struct pending_wmi_event, wmi_recv_cmd()
763 if (!evt) wmi_recv_cmd()
766 evt->event.hdr = hdr; wmi_recv_cmd()
767 cmd = (void *)&evt->event.wmi; wmi_recv_cmd()
775 struct wil6210_mbox_hdr_wmi *wmi = &evt->event.wmi; wmi_recv_cmd()
784 wil_hex_dump_wmi("evt ", DUMP_PREFIX_OFFSET, 16, 1, wmi_recv_cmd()
785 &evt->event.hdr, sizeof(hdr) + len, true); wmi_recv_cmd()
795 list_add_tail(&evt->list, &wil->pending_wmi_ev); wmi_recv_cmd()
874 struct wmi_pcp_started_event evt; wmi_pcp_start() member in struct:__anon7741
897 if (reply.evt.status != WMI_FW_STATUS_SUCCESS) wmi_pcp_start()
1053 struct wmi_listen_started_event evt; wmi_rxon() member in struct:__anon7744
1062 if ((rc == 0) && (reply.evt.status != WMI_FW_STATUS_SUCCESS)) wmi_rxon()
1090 struct wmi_cfg_rx_chain_done_event evt; wmi_rx_chain_add() member in struct:__anon7745
1091 } __packed evt; wmi_rx_chain_add() local
1119 WMI_CFG_RX_CHAIN_DONE_EVENTID, &evt, sizeof(evt), 2000); wmi_rx_chain_add()
1123 vring->hwtail = le32_to_cpu(evt.evt.rx_ring_tail_ptr); wmi_rx_chain_add()
1126 le32_to_cpu(evt.evt.status), vring->hwtail); wmi_rx_chain_add()
1128 if (le32_to_cpu(evt.evt.status) != WMI_CFG_RX_CHAIN_SUCCESS) wmi_rx_chain_add()
1144 struct wmi_temp_sense_done_event evt; wmi_get_temperature() member in struct:__anon7746
1153 *t_bb = le32_to_cpu(reply.evt.baseband_t1000); wmi_get_temperature()
1155 *t_rf = le32_to_cpu(reply.evt.rf_t1000); wmi_get_temperature()
1233 struct wmi_rcp_addba_resp_sent_event evt; wmi_addba_rx_resp() member in struct:__anon7747
1245 if (reply.evt.status) { wmi_addba_rx_resp()
1247 le16_to_cpu(reply.evt.status)); wmi_addba_rx_resp()
1256 struct pending_wmi_event *evt, *t; wmi_event_flush() local
1260 list_for_each_entry_safe(evt, t, &wil->pending_wmi_ev, list) { wmi_event_flush()
1261 list_del(&evt->list); wmi_event_flush()
1262 kfree(evt); wmi_event_flush()
1315 print_hex_dump(KERN_ERR, "evt?? ", DUMP_PREFIX_OFFSET, 16, 1, wmi_event_handle()
1347 struct pending_wmi_event *evt; wmi_event_worker() local
1352 evt = list_entry(lh, struct pending_wmi_event, list); wmi_event_worker()
1353 wmi_event_handle(wil, &evt->event.hdr); wmi_event_worker()
1354 kfree(evt); wmi_event_worker()
H A Drx_reorder.c341 struct wil_back_rx *evt, *t; wil_back_rx_flush() local
347 list_for_each_entry_safe(evt, t, &wil->back_rx_pending, list) { wil_back_rx_flush()
348 list_del(&evt->list); wil_back_rx_flush()
349 kfree(evt); wil_back_rx_flush()
376 struct wil_back_rx *evt; wil_back_rx_worker() local
380 evt = list_entry(lh, struct wil_back_rx, list); wil_back_rx_worker()
382 wil_back_rx_handle(wil, evt); wil_back_rx_worker()
383 kfree(evt); wil_back_rx_worker()
431 struct wil_back_tx *evt; wil_back_tx_worker() local
435 evt = list_entry(lh, struct wil_back_tx, list); wil_back_tx_worker()
437 wil_back_tx_handle(wil, evt); wil_back_tx_worker()
438 kfree(evt); wil_back_tx_worker()
444 struct wil_back_tx *evt, *t; wil_back_tx_flush() local
450 list_for_each_entry_safe(evt, t, &wil->back_tx_pending, list) { wil_back_tx_flush()
451 list_del(&evt->list); wil_back_tx_flush()
452 kfree(evt); wil_back_tx_flush()
H A Dcfg80211.c117 struct wmi_notify_req_done_event evt; wil_cid_fill_sinfo() member in struct:__anon7723
132 cid, le16_to_cpu(reply.evt.bf_mcs), wil_cid_fill_sinfo()
133 le64_to_cpu(reply.evt.tsf), reply.evt.status, wil_cid_fill_sinfo()
134 le32_to_cpu(reply.evt.snr_val), wil_cid_fill_sinfo()
135 reply.evt.sqi, wil_cid_fill_sinfo()
136 le32_to_cpu(reply.evt.tx_tpt), wil_cid_fill_sinfo()
137 le32_to_cpu(reply.evt.tx_goodput), wil_cid_fill_sinfo()
138 le32_to_cpu(reply.evt.rx_goodput), wil_cid_fill_sinfo()
139 le16_to_cpu(reply.evt.my_rx_sector), wil_cid_fill_sinfo()
140 le16_to_cpu(reply.evt.my_tx_sector), wil_cid_fill_sinfo()
141 le16_to_cpu(reply.evt.other_rx_sector), wil_cid_fill_sinfo()
142 le16_to_cpu(reply.evt.other_tx_sector)); wil_cid_fill_sinfo()
156 sinfo->txrate.mcs = le16_to_cpu(reply.evt.bf_mcs); wil_cid_fill_sinfo()
168 sinfo->signal = reply.evt.sqi; wil_cid_fill_sinfo()
528 struct wmi_sw_tx_complete_event evt; wil_cfg80211_mgmt_tx() member in struct:__anon7725
529 } __packed evt; wil_cfg80211_mgmt_tx() local
542 WMI_SW_TX_COMPLETE_EVENTID, &evt, sizeof(evt), 2000); wil_cfg80211_mgmt_tx()
544 tx_status = !evt.evt.status; wil_cfg80211_mgmt_tx()
H A Ddebugfs.c930 struct wmi_notify_req_done_event evt; wil_bf_debugfs_show() member in struct:__anon7727
941 if (rc || is_all_zeros(&reply.evt, sizeof(reply.evt))) wil_bf_debugfs_show()
944 status = le32_to_cpu(reply.evt.status); wil_bf_debugfs_show()
954 le64_to_cpu(reply.evt.tsf), wil_bf_debugfs_show()
955 le16_to_cpu(reply.evt.bf_mcs), wil_bf_debugfs_show()
956 le32_to_cpu(reply.evt.tx_tpt), wil_bf_debugfs_show()
957 reply.evt.sqi, wil_bf_debugfs_show()
959 le16_to_cpu(reply.evt.my_rx_sector), wil_bf_debugfs_show()
960 le16_to_cpu(reply.evt.my_tx_sector), wil_bf_debugfs_show()
961 le16_to_cpu(reply.evt.other_rx_sector), wil_bf_debugfs_show()
962 le16_to_cpu(reply.evt.other_tx_sector), wil_bf_debugfs_show()
963 le32_to_cpu(reply.evt.rx_goodput), wil_bf_debugfs_show()
964 le32_to_cpu(reply.evt.tx_goodput)); wil_bf_debugfs_show()
/linux-4.1.27/arch/xtensa/kernel/
H A Dtime.c56 struct clock_event_device *evt);
58 struct clock_event_device evt; member in struct:ccount_timer
81 struct clock_event_device *evt) ccount_timer_set_mode()
84 container_of(evt, struct ccount_timer, evt); ccount_timer_set_mode()
96 disable_irq(evt->irq); ccount_timer_set_mode()
103 enable_irq(evt->irq); ccount_timer_set_mode()
121 struct clock_event_device *clockevent = &timer->evt; local_timer_setup()
150 setup_irq(this_cpu_ptr(&ccount_timer)->evt.irq, &timer_irqaction); time_init()
161 struct clock_event_device *evt = &this_cpu_ptr(&ccount_timer)->evt; timer_interrupt() local
164 evt->event_handler(evt); timer_interrupt()
80 ccount_timer_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) ccount_timer_set_mode() argument
/linux-4.1.27/drivers/scsi/ibmvscsi/
H A Dibmvfc.c150 * @evt: ibmvfc event struct
153 static void ibmvfc_trc_start(struct ibmvfc_event *evt) ibmvfc_trc_start() argument
155 struct ibmvfc_host *vhost = evt->vhost; ibmvfc_trc_start()
156 struct ibmvfc_cmd *vfc_cmd = &evt->iu.cmd; ibmvfc_trc_start()
157 struct ibmvfc_mad_common *mad = &evt->iu.mad_common; ibmvfc_trc_start()
161 entry->evt = evt; ibmvfc_trc_start()
163 entry->fmt = evt->crq.format; ibmvfc_trc_start()
184 * @evt: ibmvfc event struct
187 static void ibmvfc_trc_end(struct ibmvfc_event *evt) ibmvfc_trc_end() argument
189 struct ibmvfc_host *vhost = evt->vhost; ibmvfc_trc_end()
190 struct ibmvfc_cmd *vfc_cmd = &evt->xfer_iu->cmd; ibmvfc_trc_end()
191 struct ibmvfc_mad_common *mad = &evt->xfer_iu->mad_common; ibmvfc_trc_end()
194 entry->evt = evt; ibmvfc_trc_end()
196 entry->fmt = evt->crq.format; ibmvfc_trc_end()
222 #define ibmvfc_trc_start(evt) do { } while (0)
223 #define ibmvfc_trc_end(evt) do { } while (0)
743 * @evt: ibmvfc event to be checked for validity
749 struct ibmvfc_event *evt) ibmvfc_valid_event()
751 int index = evt - pool->events; ibmvfc_valid_event()
754 if (evt != pool->events + index) /* unaligned */ ibmvfc_valid_event()
761 * @evt: ibmvfc_event to be freed
764 static void ibmvfc_free_event(struct ibmvfc_event *evt) ibmvfc_free_event() argument
766 struct ibmvfc_host *vhost = evt->vhost; ibmvfc_free_event()
769 BUG_ON(!ibmvfc_valid_event(pool, evt)); ibmvfc_free_event()
770 BUG_ON(atomic_inc_return(&evt->free) != 1); ibmvfc_free_event()
771 list_add_tail(&evt->queue, &vhost->free); ibmvfc_free_event()
776 * @evt: ibmvfc event struct
781 static void ibmvfc_scsi_eh_done(struct ibmvfc_event *evt) ibmvfc_scsi_eh_done() argument
783 struct scsi_cmnd *cmnd = evt->cmnd; ibmvfc_scsi_eh_done()
790 if (evt->eh_comp) ibmvfc_scsi_eh_done()
791 complete(evt->eh_comp); ibmvfc_scsi_eh_done()
793 ibmvfc_free_event(evt); ibmvfc_scsi_eh_done()
798 * @evt: ibmvfc event struct
804 static void ibmvfc_fail_request(struct ibmvfc_event *evt, int error_code) ibmvfc_fail_request() argument
806 if (evt->cmnd) { ibmvfc_fail_request()
807 evt->cmnd->result = (error_code << 16); ibmvfc_fail_request()
808 evt->done = ibmvfc_scsi_eh_done; ibmvfc_fail_request()
810 evt->xfer_iu->mad_common.status = cpu_to_be16(IBMVFC_MAD_DRIVER_FAILED); ibmvfc_fail_request()
812 list_del(&evt->queue); ibmvfc_fail_request()
813 del_timer(&evt->timer); ibmvfc_fail_request()
814 ibmvfc_trc_end(evt); ibmvfc_fail_request()
815 evt->done(evt); ibmvfc_fail_request()
828 struct ibmvfc_event *evt, *pos; ibmvfc_purge_requests() local
831 list_for_each_entry_safe(evt, pos, &vhost->sent, queue) ibmvfc_purge_requests()
832 ibmvfc_fail_request(evt, error_code); ibmvfc_purge_requests()
1225 struct ibmvfc_event *evt = &pool->events[i]; ibmvfc_init_event_pool() local
1226 atomic_set(&evt->free, 1); ibmvfc_init_event_pool()
1227 evt->crq.valid = 0x80; ibmvfc_init_event_pool()
1228 evt->crq.ioba = cpu_to_be64(pool->iu_token + (sizeof(*evt->xfer_iu) * i)); ibmvfc_init_event_pool()
1229 evt->xfer_iu = pool->iu_storage + i; ibmvfc_init_event_pool()
1230 evt->vhost = vhost; ibmvfc_init_event_pool()
1231 evt->ext_list = NULL; ibmvfc_init_event_pool()
1232 list_add_tail(&evt->queue, &vhost->free); ibmvfc_init_event_pool()
1274 struct ibmvfc_event *evt; ibmvfc_get_event() local
1277 evt = list_entry(vhost->free.next, struct ibmvfc_event, queue); ibmvfc_get_event()
1278 atomic_set(&evt->free, 0); ibmvfc_get_event()
1279 list_del(&evt->queue); ibmvfc_get_event()
1280 return evt; ibmvfc_get_event()
1286 * @evt: The event
1290 static void ibmvfc_init_event(struct ibmvfc_event *evt, ibmvfc_init_event() argument
1293 evt->cmnd = NULL; ibmvfc_init_event()
1294 evt->sync_iu = NULL; ibmvfc_init_event()
1295 evt->crq.format = format; ibmvfc_init_event()
1296 evt->done = done; ibmvfc_init_event()
1297 evt->eh_comp = NULL; ibmvfc_init_event()
1322 * @evt: ibmvfc event struct
1330 struct ibmvfc_event *evt, ibmvfc_map_sg_data()
1363 if (!evt->ext_list) { ibmvfc_map_sg_data()
1364 evt->ext_list = dma_pool_alloc(vhost->sg_pool, GFP_ATOMIC, ibmvfc_map_sg_data()
1365 &evt->ext_list_token); ibmvfc_map_sg_data()
1367 if (!evt->ext_list) { ibmvfc_map_sg_data()
1375 ibmvfc_map_sg_list(scmd, sg_mapped, evt->ext_list); ibmvfc_map_sg_data()
1377 data->va = cpu_to_be64(evt->ext_list_token); ibmvfc_map_sg_data()
1385 * @evt: struct ibmvfc_event that timed out
1389 static void ibmvfc_timeout(struct ibmvfc_event *evt) ibmvfc_timeout() argument
1391 struct ibmvfc_host *vhost = evt->vhost; ibmvfc_timeout()
1392 dev_err(vhost->dev, "Command timed out (%p). Resetting connection\n", evt); ibmvfc_timeout()
1398 * @evt: event to be sent
1404 static int ibmvfc_send_event(struct ibmvfc_event *evt, ibmvfc_send_event() argument
1407 __be64 *crq_as_u64 = (__be64 *) &evt->crq; ibmvfc_send_event()
1411 *evt->xfer_iu = evt->iu; ibmvfc_send_event()
1412 if (evt->crq.format == IBMVFC_CMD_FORMAT) ibmvfc_send_event()
1413 evt->xfer_iu->cmd.tag = cpu_to_be64((u64)evt); ibmvfc_send_event()
1414 else if (evt->crq.format == IBMVFC_MAD_FORMAT) ibmvfc_send_event()
1415 evt->xfer_iu->mad_common.tag = cpu_to_be64((u64)evt); ibmvfc_send_event()
1419 list_add_tail(&evt->queue, &vhost->sent); ibmvfc_send_event()
1420 init_timer(&evt->timer); ibmvfc_send_event()
1423 evt->timer.data = (unsigned long) evt; ibmvfc_send_event()
1424 evt->timer.expires = jiffies + (timeout * HZ); ibmvfc_send_event()
1425 evt->timer.function = (void (*)(unsigned long))ibmvfc_timeout; ibmvfc_send_event()
1426 add_timer(&evt->timer); ibmvfc_send_event()
1433 list_del(&evt->queue); ibmvfc_send_event()
1434 del_timer(&evt->timer); ibmvfc_send_event()
1444 if (evt->cmnd) ibmvfc_send_event()
1445 scsi_dma_unmap(evt->cmnd); ibmvfc_send_event()
1446 ibmvfc_free_event(evt); ibmvfc_send_event()
1451 if (evt->cmnd) { ibmvfc_send_event()
1452 evt->cmnd->result = DID_ERROR << 16; ibmvfc_send_event()
1453 evt->done = ibmvfc_scsi_eh_done; ibmvfc_send_event()
1455 evt->xfer_iu->mad_common.status = cpu_to_be16(IBMVFC_MAD_CRQ_ERROR); ibmvfc_send_event()
1457 evt->done(evt); ibmvfc_send_event()
1459 ibmvfc_trc_start(evt); ibmvfc_send_event()
1466 * @evt: ibmvfc event to log
1469 static void ibmvfc_log_error(struct ibmvfc_event *evt) ibmvfc_log_error() argument
1471 struct ibmvfc_cmd *vfc_cmd = &evt->xfer_iu->cmd; ibmvfc_log_error()
1472 struct ibmvfc_host *vhost = evt->vhost; ibmvfc_log_error()
1474 struct scsi_cmnd *cmnd = evt->cmnd; ibmvfc_log_error()
1520 * @evt: ibmvfc event to be handled
1524 static void ibmvfc_scsi_done(struct ibmvfc_event *evt) ibmvfc_scsi_done() argument
1526 struct ibmvfc_cmd *vfc_cmd = &evt->xfer_iu->cmd; ibmvfc_scsi_done()
1528 struct scsi_cmnd *cmnd = evt->cmnd; ibmvfc_scsi_done()
1556 ibmvfc_log_error(evt); ibmvfc_scsi_done()
1567 if (evt->eh_comp) ibmvfc_scsi_done()
1568 complete(evt->eh_comp); ibmvfc_scsi_done()
1570 ibmvfc_free_event(evt); ibmvfc_scsi_done()
1617 struct ibmvfc_event *evt; ibmvfc_queuecommand_lck() local
1628 evt = ibmvfc_get_event(vhost); ibmvfc_queuecommand_lck()
1629 ibmvfc_init_event(evt, ibmvfc_scsi_done, IBMVFC_CMD_FORMAT); ibmvfc_queuecommand_lck()
1630 evt->cmnd = cmnd; ibmvfc_queuecommand_lck()
1632 vfc_cmd = &evt->iu.cmd; ibmvfc_queuecommand_lck()
1634 vfc_cmd->resp.va = cpu_to_be64(be64_to_cpu(evt->crq.ioba) + offsetof(struct ibmvfc_cmd, rsp)); ibmvfc_queuecommand_lck()
1650 if (likely(!(rc = ibmvfc_map_sg_data(cmnd, evt, vfc_cmd, vhost->dev)))) ibmvfc_queuecommand_lck()
1651 return ibmvfc_send_event(evt, vhost, 0); ibmvfc_queuecommand_lck()
1653 ibmvfc_free_event(evt); ibmvfc_queuecommand_lck()
1670 * @evt: ibmvfc event struct
1673 static void ibmvfc_sync_completion(struct ibmvfc_event *evt) ibmvfc_sync_completion() argument
1676 if (evt->sync_iu) ibmvfc_sync_completion()
1677 *evt->sync_iu = *evt->xfer_iu; ibmvfc_sync_completion()
1679 complete(&evt->comp); ibmvfc_sync_completion()
1684 * @evt: struct ibmvfc_event
1687 static void ibmvfc_bsg_timeout_done(struct ibmvfc_event *evt) ibmvfc_bsg_timeout_done() argument
1689 struct ibmvfc_host *vhost = evt->vhost; ibmvfc_bsg_timeout_done()
1691 ibmvfc_free_event(evt); ibmvfc_bsg_timeout_done()
1707 struct ibmvfc_event *evt; ibmvfc_bsg_timeout() local
1721 evt = ibmvfc_get_event(vhost); ibmvfc_bsg_timeout()
1722 ibmvfc_init_event(evt, ibmvfc_bsg_timeout_done, IBMVFC_MAD_FORMAT); ibmvfc_bsg_timeout()
1724 tmf = &evt->iu.tmf; ibmvfc_bsg_timeout()
1732 rc = ibmvfc_send_event(evt, vhost, default_timeout); ibmvfc_bsg_timeout()
1760 struct ibmvfc_event *evt; ibmvfc_bsg_plogi() local
1779 evt = ibmvfc_get_event(vhost); ibmvfc_bsg_plogi()
1780 ibmvfc_init_event(evt, ibmvfc_sync_completion, IBMVFC_MAD_FORMAT); ibmvfc_bsg_plogi()
1781 plogi = &evt->iu.plogi; ibmvfc_bsg_plogi()
1787 evt->sync_iu = &rsp_iu; ibmvfc_bsg_plogi()
1788 init_completion(&evt->comp); ibmvfc_bsg_plogi()
1790 rc = ibmvfc_send_event(evt, vhost, default_timeout); ibmvfc_bsg_plogi()
1796 wait_for_completion(&evt->comp); ibmvfc_bsg_plogi()
1802 ibmvfc_free_event(evt); ibmvfc_bsg_plogi()
1821 struct ibmvfc_event *evt; ibmvfc_bsg_request() local
1893 evt = ibmvfc_get_event(vhost); ibmvfc_bsg_request()
1894 ibmvfc_init_event(evt, ibmvfc_sync_completion, IBMVFC_MAD_FORMAT); ibmvfc_bsg_request()
1895 mad = &evt->iu.passthru; ibmvfc_bsg_request()
1902 mad->cmd_ioba.va = cpu_to_be64(be64_to_cpu(evt->crq.ioba) + ibmvfc_bsg_request()
1916 mad->iu.tag = cpu_to_be64((u64)evt); ibmvfc_bsg_request()
1919 evt->sync_iu = &rsp_iu; ibmvfc_bsg_request()
1920 init_completion(&evt->comp); ibmvfc_bsg_request()
1921 rc = ibmvfc_send_event(evt, vhost, 0); ibmvfc_bsg_request()
1929 wait_for_completion(&evt->comp); ibmvfc_bsg_request()
1937 ibmvfc_free_event(evt); ibmvfc_bsg_request()
1966 struct ibmvfc_event *evt = NULL; ibmvfc_reset_device() local
1975 evt = ibmvfc_get_event(vhost); ibmvfc_reset_device()
1976 ibmvfc_init_event(evt, ibmvfc_sync_completion, IBMVFC_CMD_FORMAT); ibmvfc_reset_device()
1978 tmf = &evt->iu.cmd; ibmvfc_reset_device()
1980 tmf->resp.va = cpu_to_be64(be64_to_cpu(evt->crq.ioba) + offsetof(struct ibmvfc_cmd, rsp)); ibmvfc_reset_device()
1990 evt->sync_iu = &rsp_iu; ibmvfc_reset_device()
1992 init_completion(&evt->comp); ibmvfc_reset_device()
1993 rsp_rc = ibmvfc_send_event(evt, vhost, default_timeout); ibmvfc_reset_device()
2004 wait_for_completion(&evt->comp); ibmvfc_reset_device()
2023 ibmvfc_free_event(evt); ibmvfc_reset_device()
2030 * @evt: ibmvfc event struct
2036 static int ibmvfc_match_rport(struct ibmvfc_event *evt, void *rport) ibmvfc_match_rport() argument
2040 if (evt->cmnd) { ibmvfc_match_rport()
2041 cmd_rport = starget_to_rport(scsi_target(evt->cmnd->device)); ibmvfc_match_rport()
2050 * @evt: ibmvfc event struct
2056 static int ibmvfc_match_target(struct ibmvfc_event *evt, void *device) ibmvfc_match_target() argument
2058 if (evt->cmnd && scsi_target(evt->cmnd->device) == device) ibmvfc_match_target()
2065 * @evt: ibmvfc event struct
2071 static int ibmvfc_match_lun(struct ibmvfc_event *evt, void *device) ibmvfc_match_lun() argument
2073 if (evt->cmnd && evt->cmnd->device == device) ibmvfc_match_lun()
2090 struct ibmvfc_event *evt; ibmvfc_wait_for_ops() local
2100 list_for_each_entry(evt, &vhost->sent, queue) { ibmvfc_wait_for_ops()
2101 if (match(evt, device)) { ibmvfc_wait_for_ops()
2102 evt->eh_comp = &comp; ibmvfc_wait_for_ops()
2114 list_for_each_entry(evt, &vhost->sent, queue) { ibmvfc_wait_for_ops()
2115 if (match(evt, device)) { ibmvfc_wait_for_ops()
2116 evt->eh_comp = NULL; ibmvfc_wait_for_ops()
2150 struct ibmvfc_event *evt, *found_evt; ibmvfc_cancel_all() local
2159 list_for_each_entry(evt, &vhost->sent, queue) { ibmvfc_cancel_all()
2160 if (evt->cmnd && evt->cmnd->device == sdev) { ibmvfc_cancel_all()
2161 found_evt = evt; ibmvfc_cancel_all()
2174 evt = ibmvfc_get_event(vhost); ibmvfc_cancel_all()
2175 ibmvfc_init_event(evt, ibmvfc_sync_completion, IBMVFC_MAD_FORMAT); ibmvfc_cancel_all()
2177 tmf = &evt->iu.tmf; ibmvfc_cancel_all()
2193 evt->sync_iu = &rsp; ibmvfc_cancel_all()
2194 init_completion(&evt->comp); ibmvfc_cancel_all()
2195 rsp_rc = ibmvfc_send_event(evt, vhost, default_timeout); ibmvfc_cancel_all()
2210 wait_for_completion(&evt->comp); ibmvfc_cancel_all()
2213 ibmvfc_free_event(evt); ibmvfc_cancel_all()
2235 * @evt: ibmvfc event struct
2241 static int ibmvfc_match_key(struct ibmvfc_event *evt, void *key) ibmvfc_match_key() argument
2245 if (evt->crq.format == IBMVFC_CMD_FORMAT && ibmvfc_match_key()
2246 be32_to_cpu(evt->iu.cmd.cancel_key) == cancel_key) ibmvfc_match_key()
2253 * @evt: ibmvfc event struct
2259 static int ibmvfc_match_evt(struct ibmvfc_event *evt, void *match) ibmvfc_match_evt() argument
2261 if (evt == match) ibmvfc_match_evt()
2281 struct ibmvfc_event *evt, *found_evt; ibmvfc_abort_task_set() local
2290 list_for_each_entry(evt, &vhost->sent, queue) { ibmvfc_abort_task_set()
2291 if (evt->cmnd && evt->cmnd->device == sdev) { ibmvfc_abort_task_set()
2292 found_evt = evt; ibmvfc_abort_task_set()
2305 evt = ibmvfc_get_event(vhost); ibmvfc_abort_task_set()
2306 ibmvfc_init_event(evt, ibmvfc_sync_completion, IBMVFC_CMD_FORMAT); ibmvfc_abort_task_set()
2308 tmf = &evt->iu.cmd; ibmvfc_abort_task_set()
2310 tmf->resp.va = cpu_to_be64(be64_to_cpu(evt->crq.ioba) + offsetof(struct ibmvfc_cmd, rsp)); ibmvfc_abort_task_set()
2320 evt->sync_iu = &rsp_iu; ibmvfc_abort_task_set()
2322 init_completion(&evt->comp); ibmvfc_abort_task_set()
2323 rsp_rc = ibmvfc_send_event(evt, vhost, default_timeout); ibmvfc_abort_task_set()
2334 timeout = wait_for_completion_timeout(&evt->comp, timeout); ibmvfc_abort_task_set()
2353 rc = ibmvfc_wait_for_ops(vhost, evt, ibmvfc_match_evt); ibmvfc_abort_task_set()
2383 ibmvfc_free_event(evt); ibmvfc_abort_task_set()
2724 struct ibmvfc_event *evt = (struct ibmvfc_event *)be64_to_cpu(crq->ioba); ibmvfc_handle_crq() local
2778 if (unlikely(!ibmvfc_valid_event(&vhost->pool, evt))) { ibmvfc_handle_crq()
2784 if (unlikely(atomic_read(&evt->free))) { ibmvfc_handle_crq()
2790 del_timer(&evt->timer); ibmvfc_handle_crq()
2791 list_del(&evt->queue); ibmvfc_handle_crq()
2792 ibmvfc_trc_end(evt); ibmvfc_handle_crq()
2793 evt->done(evt); ibmvfc_handle_crq()
3288 * @evt: ibmvfc event struct
3291 static void ibmvfc_tgt_prli_done(struct ibmvfc_event *evt) ibmvfc_tgt_prli_done() argument
3293 struct ibmvfc_target *tgt = evt->tgt; ibmvfc_tgt_prli_done()
3294 struct ibmvfc_host *vhost = evt->vhost; ibmvfc_tgt_prli_done()
3295 struct ibmvfc_process_login *rsp = &evt->xfer_iu->prli; ibmvfc_tgt_prli_done()
3351 ibmvfc_free_event(evt); ibmvfc_tgt_prli_done()
3364 struct ibmvfc_event *evt; ibmvfc_tgt_send_prli() local
3370 evt = ibmvfc_get_event(vhost); ibmvfc_tgt_send_prli()
3372 ibmvfc_init_event(evt, ibmvfc_tgt_prli_done, IBMVFC_MAD_FORMAT); ibmvfc_tgt_send_prli()
3373 evt->tgt = tgt; ibmvfc_tgt_send_prli()
3374 prli = &evt->iu.prli; ibmvfc_tgt_send_prli()
3386 if (ibmvfc_send_event(evt, vhost, default_timeout)) { ibmvfc_tgt_send_prli()
3396 * @evt: ibmvfc event struct
3399 static void ibmvfc_tgt_plogi_done(struct ibmvfc_event *evt) ibmvfc_tgt_plogi_done() argument
3401 struct ibmvfc_target *tgt = evt->tgt; ibmvfc_tgt_plogi_done()
3402 struct ibmvfc_host *vhost = evt->vhost; ibmvfc_tgt_plogi_done()
3403 struct ibmvfc_port_login *rsp = &evt->xfer_iu->plogi; ibmvfc_tgt_plogi_done()
3447 ibmvfc_free_event(evt); ibmvfc_tgt_plogi_done()
3460 struct ibmvfc_event *evt; ibmvfc_tgt_send_plogi() local
3467 evt = ibmvfc_get_event(vhost); ibmvfc_tgt_send_plogi()
3470 ibmvfc_init_event(evt, ibmvfc_tgt_plogi_done, IBMVFC_MAD_FORMAT); ibmvfc_tgt_send_plogi()
3471 evt->tgt = tgt; ibmvfc_tgt_send_plogi()
3472 plogi = &evt->iu.plogi; ibmvfc_tgt_send_plogi()
3479 if (ibmvfc_send_event(evt, vhost, default_timeout)) { ibmvfc_tgt_send_plogi()
3489 * @evt: ibmvfc event struct
3492 static void ibmvfc_tgt_implicit_logout_done(struct ibmvfc_event *evt) ibmvfc_tgt_implicit_logout_done() argument
3494 struct ibmvfc_target *tgt = evt->tgt; ibmvfc_tgt_implicit_logout_done()
3495 struct ibmvfc_host *vhost = evt->vhost; ibmvfc_tgt_implicit_logout_done()
3496 struct ibmvfc_implicit_logout *rsp = &evt->xfer_iu->implicit_logout; ibmvfc_tgt_implicit_logout_done()
3500 ibmvfc_free_event(evt); ibmvfc_tgt_implicit_logout_done()
3535 struct ibmvfc_event *evt; ibmvfc_tgt_implicit_logout() local
3541 evt = ibmvfc_get_event(vhost); ibmvfc_tgt_implicit_logout()
3543 ibmvfc_init_event(evt, ibmvfc_tgt_implicit_logout_done, IBMVFC_MAD_FORMAT); ibmvfc_tgt_implicit_logout()
3544 evt->tgt = tgt; ibmvfc_tgt_implicit_logout()
3545 mad = &evt->iu.implicit_logout; ibmvfc_tgt_implicit_logout()
3553 if (ibmvfc_send_event(evt, vhost, default_timeout)) { ibmvfc_tgt_implicit_logout()
3585 * @evt: ibmvfc event struct
3588 static void ibmvfc_tgt_adisc_done(struct ibmvfc_event *evt) ibmvfc_tgt_adisc_done() argument
3590 struct ibmvfc_target *tgt = evt->tgt; ibmvfc_tgt_adisc_done()
3591 struct ibmvfc_host *vhost = evt->vhost; ibmvfc_tgt_adisc_done()
3592 struct ibmvfc_passthru_mad *mad = &evt->xfer_iu->passthru; ibmvfc_tgt_adisc_done()
3622 ibmvfc_free_event(evt); ibmvfc_tgt_adisc_done()
3628 * @evt: ibmvfc event struct
3631 static void ibmvfc_init_passthru(struct ibmvfc_event *evt) ibmvfc_init_passthru() argument
3633 struct ibmvfc_passthru_mad *mad = &evt->iu.passthru; ibmvfc_init_passthru()
3639 mad->cmd_ioba.va = cpu_to_be64((u64)be64_to_cpu(evt->crq.ioba) + ibmvfc_init_passthru()
3644 mad->iu.cmd.va = cpu_to_be64((u64)be64_to_cpu(evt->crq.ioba) + ibmvfc_init_passthru()
3648 mad->iu.rsp.va = cpu_to_be64((u64)be64_to_cpu(evt->crq.ioba) + ibmvfc_init_passthru()
3656 * @evt: ibmvfc event struct
3664 static void ibmvfc_tgt_adisc_cancel_done(struct ibmvfc_event *evt) ibmvfc_tgt_adisc_cancel_done() argument
3666 struct ibmvfc_host *vhost = evt->vhost; ibmvfc_tgt_adisc_cancel_done()
3667 struct ibmvfc_target *tgt = evt->tgt; ibmvfc_tgt_adisc_cancel_done()
3671 ibmvfc_free_event(evt); ibmvfc_tgt_adisc_cancel_done()
3687 struct ibmvfc_event *evt; ibmvfc_adisc_timeout() local
3704 evt = ibmvfc_get_event(vhost); ibmvfc_adisc_timeout()
3705 ibmvfc_init_event(evt, ibmvfc_tgt_adisc_cancel_done, IBMVFC_MAD_FORMAT); ibmvfc_adisc_timeout()
3707 evt->tgt = tgt; ibmvfc_adisc_timeout()
3708 tmf = &evt->iu.tmf; ibmvfc_adisc_timeout()
3716 rc = ibmvfc_send_event(evt, vhost, default_timeout); ibmvfc_adisc_timeout()
3743 struct ibmvfc_event *evt; ibmvfc_tgt_adisc() local
3749 evt = ibmvfc_get_event(vhost); ibmvfc_tgt_adisc()
3751 ibmvfc_init_event(evt, ibmvfc_tgt_adisc_done, IBMVFC_MAD_FORMAT); ibmvfc_tgt_adisc()
3752 evt->tgt = tgt; ibmvfc_tgt_adisc()
3754 ibmvfc_init_passthru(evt); ibmvfc_tgt_adisc()
3755 mad = &evt->iu.passthru; ibmvfc_tgt_adisc()
3777 if (ibmvfc_send_event(evt, vhost, IBMVFC_ADISC_PLUS_CANCEL_TIMEOUT)) { ibmvfc_tgt_adisc()
3788 * @evt: ibmvfc event struct
3791 static void ibmvfc_tgt_query_target_done(struct ibmvfc_event *evt) ibmvfc_tgt_query_target_done() argument
3793 struct ibmvfc_target *tgt = evt->tgt; ibmvfc_tgt_query_target_done()
3794 struct ibmvfc_host *vhost = evt->vhost; ibmvfc_tgt_query_target_done()
3795 struct ibmvfc_query_tgt *rsp = &evt->xfer_iu->query_tgt; ibmvfc_tgt_query_target_done()
3835 ibmvfc_free_event(evt); ibmvfc_tgt_query_target_done()
3848 struct ibmvfc_event *evt; ibmvfc_tgt_query_target() local
3854 evt = ibmvfc_get_event(vhost); ibmvfc_tgt_query_target()
3856 evt->tgt = tgt; ibmvfc_tgt_query_target()
3857 ibmvfc_init_event(evt, ibmvfc_tgt_query_target_done, IBMVFC_MAD_FORMAT); ibmvfc_tgt_query_target()
3858 query_tgt = &evt->iu.query_tgt; ibmvfc_tgt_query_target()
3866 if (ibmvfc_send_event(evt, vhost, default_timeout)) { ibmvfc_tgt_query_target()
3942 * @evt: ibmvfc event struct
3945 static void ibmvfc_discover_targets_done(struct ibmvfc_event *evt) ibmvfc_discover_targets_done() argument
3947 struct ibmvfc_host *vhost = evt->vhost; ibmvfc_discover_targets_done()
3948 struct ibmvfc_discover_targets *rsp = &evt->xfer_iu->discover_targets; ibmvfc_discover_targets_done()
3972 ibmvfc_free_event(evt); ibmvfc_discover_targets_done()
3984 struct ibmvfc_event *evt = ibmvfc_get_event(vhost); ibmvfc_discover_targets() local
3986 ibmvfc_init_event(evt, ibmvfc_discover_targets_done, IBMVFC_MAD_FORMAT); ibmvfc_discover_targets()
3987 mad = &evt->iu.discover_targets; ibmvfc_discover_targets()
3997 if (!ibmvfc_send_event(evt, vhost, default_timeout)) ibmvfc_discover_targets()
4005 * @evt: ibmvfc event struct
4008 static void ibmvfc_npiv_login_done(struct ibmvfc_event *evt) ibmvfc_npiv_login_done() argument
4010 struct ibmvfc_host *vhost = evt->vhost; ibmvfc_npiv_login_done()
4011 u32 mad_status = be16_to_cpu(evt->xfer_iu->npiv_login.common.status); ibmvfc_npiv_login_done()
4018 ibmvfc_free_event(evt); ibmvfc_npiv_login_done()
4028 ibmvfc_free_event(evt); ibmvfc_npiv_login_done()
4033 ibmvfc_free_event(evt); ibmvfc_npiv_login_done()
4038 ibmvfc_free_event(evt); ibmvfc_npiv_login_done()
4095 struct ibmvfc_event *evt = ibmvfc_get_event(vhost); ibmvfc_npiv_login() local
4099 ibmvfc_init_event(evt, ibmvfc_npiv_login_done, IBMVFC_MAD_FORMAT); ibmvfc_npiv_login()
4102 mad = &evt->iu.npiv_login; ibmvfc_npiv_login()
4112 if (!ibmvfc_send_event(evt, vhost, default_timeout)) ibmvfc_npiv_login()
4123 static void ibmvfc_npiv_logout_done(struct ibmvfc_event *evt) ibmvfc_npiv_logout_done() argument
4125 struct ibmvfc_host *vhost = evt->vhost; ibmvfc_npiv_logout_done()
4126 u32 mad_status = be16_to_cpu(evt->xfer_iu->npiv_logout.common.status); ibmvfc_npiv_logout_done()
4128 ibmvfc_free_event(evt); ibmvfc_npiv_logout_done()
4158 struct ibmvfc_event *evt; ibmvfc_npiv_logout() local
4160 evt = ibmvfc_get_event(vhost); ibmvfc_npiv_logout()
4161 ibmvfc_init_event(evt, ibmvfc_npiv_logout_done, IBMVFC_MAD_FORMAT); ibmvfc_npiv_logout()
4163 mad = &evt->iu.npiv_logout; ibmvfc_npiv_logout()
4171 if (!ibmvfc_send_event(evt, vhost, default_timeout)) ibmvfc_npiv_logout()
748 ibmvfc_valid_event(struct ibmvfc_event_pool *pool, struct ibmvfc_event *evt) ibmvfc_valid_event() argument
1329 ibmvfc_map_sg_data(struct scsi_cmnd *scmd, struct ibmvfc_event *evt, struct ibmvfc_cmd *vfc_cmd, struct device *dev) ibmvfc_map_sg_data() argument
H A Dibmvscsi.c474 struct srp_event_struct *evt = &pool->events[i]; initialize_event_pool() local
475 memset(&evt->crq, 0x00, sizeof(evt->crq)); initialize_event_pool()
476 atomic_set(&evt->free, 1); initialize_event_pool()
477 evt->crq.valid = 0x80; initialize_event_pool()
478 evt->crq.IU_length = cpu_to_be16(sizeof(*evt->xfer_iu)); initialize_event_pool()
479 evt->crq.IU_data_ptr = cpu_to_be64(pool->iu_token + initialize_event_pool()
480 sizeof(*evt->xfer_iu) * i); initialize_event_pool()
481 evt->xfer_iu = pool->iu_storage + i; initialize_event_pool()
482 evt->hostdata = hostdata; initialize_event_pool()
483 evt->ext_list = NULL; initialize_event_pool()
484 evt->ext_list_token = 0; initialize_event_pool()
523 * @evt: srp_event_struct to be checked for validity
528 struct srp_event_struct *evt) valid_event_struct()
530 int index = evt - pool->events; valid_event_struct()
533 if (evt != pool->events + index) /* unaligned */ valid_event_struct()
541 * @evt: srp_event_struct to be modified
545 struct srp_event_struct *evt) free_event_struct()
547 if (!valid_event_struct(pool, evt)) { free_event_struct()
548 dev_err(evt->hostdata->dev, "Freeing invalid event_struct %p " free_event_struct()
549 "(not in pool %p)\n", evt, pool->events); free_event_struct()
552 if (atomic_inc_return(&evt->free) != 1) { free_event_struct()
553 dev_err(evt->hostdata->dev, "Freeing event_struct %p " free_event_struct()
554 "which is not in use!\n", evt); free_event_struct()
588 * @evt: The event
794 struct srp_event_struct *evt; purge_requests() local
799 evt = list_first_entry(&hostdata->sent, struct srp_event_struct, list); purge_requests()
800 list_del(&evt->list); purge_requests()
801 del_timer(&evt->timer); purge_requests()
804 if (evt->cmnd) { purge_requests()
805 evt->cmnd->result = (error_code << 16); purge_requests()
806 unmap_cmd_data(&evt->iu.srp.cmd, evt, purge_requests()
807 evt->hostdata->dev); purge_requests()
808 if (evt->cmnd_done) purge_requests()
809 evt->cmnd_done(evt->cmnd); purge_requests()
810 } else if (evt->done && evt->crq.format != VIOSRP_MAD_FORMAT && purge_requests()
811 evt->iu.srp.login_req.opcode != SRP_LOGIN_REQ) purge_requests()
812 evt->done(evt); purge_requests()
813 free_event_struct(&evt->hostdata->pool, evt); purge_requests()
1487 struct srp_event_struct *evt; ibmvscsi_eh_abort_handler() local
1514 evt = get_event_struct(&hostdata->pool); ibmvscsi_eh_abort_handler()
1515 if (evt == NULL) { ibmvscsi_eh_abort_handler()
1522 init_event_struct(evt, ibmvscsi_eh_abort_handler()
1527 tsk_mgmt = &evt->iu.srp.tsk_mgmt; ibmvscsi_eh_abort_handler()
1536 evt->sync_srp = &srp_rsp; ibmvscsi_eh_abort_handler()
1538 init_completion(&evt->comp); ibmvscsi_eh_abort_handler()
1539 rsp_rc = ibmvscsi_send_srp_event(evt, hostdata, abort_timeout * 2); ibmvscsi_eh_abort_handler()
1561 wait_for_completion(&evt->comp); ibmvscsi_eh_abort_handler()
1626 struct srp_event_struct *evt; ibmvscsi_eh_device_reset_handler() local
1637 evt = get_event_struct(&hostdata->pool); ibmvscsi_eh_device_reset_handler()
1638 if (evt == NULL) { ibmvscsi_eh_device_reset_handler()
1645 init_event_struct(evt, ibmvscsi_eh_device_reset_handler()
1650 tsk_mgmt = &evt->iu.srp.tsk_mgmt; ibmvscsi_eh_device_reset_handler()
1658 evt->sync_srp = &srp_rsp; ibmvscsi_eh_device_reset_handler()
1660 init_completion(&evt->comp); ibmvscsi_eh_device_reset_handler()
1661 rsp_rc = ibmvscsi_send_srp_event(evt, hostdata, reset_timeout * 2); ibmvscsi_eh_device_reset_handler()
1682 wait_for_completion(&evt->comp); ibmvscsi_eh_device_reset_handler()
527 valid_event_struct(struct event_pool *pool, struct srp_event_struct *evt) valid_event_struct() argument
544 free_event_struct(struct event_pool *pool, struct srp_event_struct *evt) free_event_struct() argument
H A Dibmvfc.h507 struct ibmvfc_event *evt; member in struct:ibmvfc_trace_entry
/linux-4.1.27/arch/x86/xen/
H A Dtime.c278 struct clock_event_device *evt) xen_timerop_set_mode()
298 struct clock_event_device *evt) xen_timerop_set_next_event()
300 WARN_ON(evt->mode != CLOCK_EVT_MODE_ONESHOT); xen_timerop_set_next_event()
330 struct clock_event_device *evt) xen_vcpuop_set_mode()
356 struct clock_event_device *evt) xen_vcpuop_set_next_event()
362 WARN_ON(evt->mode != CLOCK_EVT_MODE_ONESHOT); xen_vcpuop_set_next_event()
393 struct clock_event_device evt; member in struct:xen_clock_event_device
396 static DEFINE_PER_CPU(struct xen_clock_event_device, xen_clock_events) = { .evt.irq = -1 };
400 struct clock_event_device *evt = this_cpu_ptr(&xen_clock_events.evt); xen_timer_interrupt() local
404 if (evt->event_handler) { xen_timer_interrupt()
405 evt->event_handler(evt); xen_timer_interrupt()
416 struct clock_event_device *evt; xen_teardown_timer() local
418 evt = &per_cpu(xen_clock_events, cpu).evt; xen_teardown_timer()
420 if (evt->irq >= 0) { xen_teardown_timer()
421 unbind_from_irqhandler(evt->irq, NULL); xen_teardown_timer()
422 evt->irq = -1; xen_teardown_timer()
429 struct clock_event_device *evt = &xevt->evt; xen_setup_timer() local
432 WARN(evt->irq >= 0, "IRQ%d for CPU%d is already allocated\n", evt->irq, cpu); xen_setup_timer()
433 if (evt->irq >= 0) xen_setup_timer()
446 memcpy(evt, xen_clockevent, sizeof(*evt)); xen_setup_timer()
448 evt->cpumask = cpumask_of(cpu); xen_setup_timer()
449 evt->irq = irq; xen_setup_timer()
455 clockevents_register_device(this_cpu_ptr(&xen_clock_events.evt)); xen_setup_cpu_clockevents()
277 xen_timerop_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) xen_timerop_set_mode() argument
297 xen_timerop_set_next_event(unsigned long delta, struct clock_event_device *evt) xen_timerop_set_next_event() argument
329 xen_vcpuop_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) xen_vcpuop_set_mode() argument
355 xen_vcpuop_set_next_event(unsigned long delta, struct clock_event_device *evt) xen_vcpuop_set_next_event() argument
/linux-4.1.27/arch/m68k/coldfire/
H A Dpit.c46 struct clock_event_device *evt) init_cf_pit_timer()
84 struct clock_event_device *evt) cf_pit_next_event()
105 struct clock_event_device *evt = &cf_pit_clockevent; pit_tick() local
113 evt->event_handler(evt); pit_tick()
45 init_cf_pit_timer(enum clock_event_mode mode, struct clock_event_device *evt) init_cf_pit_timer() argument
83 cf_pit_next_event(unsigned long delta, struct clock_event_device *evt) cf_pit_next_event() argument
/linux-4.1.27/drivers/scsi/csiostor/
H A Dcsio_rnode.c315 "Unhandled rport_type:%d recv in rdev evt " csio_confirm_rnode()
600 * @evt - SM event.
604 csio_rns_uninit(struct csio_rnode *rn, enum csio_rn_ev evt) csio_rns_uninit() argument
609 CSIO_INC_STATS(rn, n_evt_sm[evt]); csio_rns_uninit()
611 switch (evt) { csio_rns_uninit()
625 "in rn state[uninit]\n", csio_rn_flowid(rn), evt); csio_rns_uninit()
631 "in rn state[uninit]\n", csio_rn_flowid(rn), evt); csio_rns_uninit()
640 * @evt - SM event.
644 csio_rns_ready(struct csio_rnode *rn, enum csio_rn_ev evt) csio_rns_ready() argument
649 CSIO_INC_STATS(rn, n_evt_sm[evt]); csio_rns_ready()
651 switch (evt) { csio_rns_ready()
656 "in rn state[ready]\n", csio_rn_flowid(rn), evt, csio_rns_ready()
715 "in rn state[uninit]\n", csio_rn_flowid(rn), evt, csio_rns_ready()
725 * @evt - SM event.
729 csio_rns_offline(struct csio_rnode *rn, enum csio_rn_ev evt) csio_rns_offline() argument
734 CSIO_INC_STATS(rn, n_evt_sm[evt]); csio_rns_offline()
736 switch (evt) { csio_rns_offline()
752 "in rn state[offline]\n", csio_rn_flowid(rn), evt, csio_rns_offline()
774 "in rn state[offline]\n", csio_rn_flowid(rn), evt, csio_rns_offline()
784 * @evt - SM event.
788 csio_rns_disappeared(struct csio_rnode *rn, enum csio_rn_ev evt) csio_rns_disappeared() argument
793 CSIO_INC_STATS(rn, n_evt_sm[evt]); csio_rns_disappeared()
795 switch (evt) { csio_rns_disappeared()
823 evt, rn->nport_id); csio_rns_disappeared()
830 evt, rn->nport_id); csio_rns_disappeared()
871 enum csio_rn_ev evt; csio_rnode_fwevt_handler() local
873 evt = CSIO_FWE_TO_RNFE(fwevt); csio_rnode_fwevt_handler()
874 if (!evt) { csio_rnode_fwevt_handler()
887 csio_post_event(&rn->sm, evt); csio_rnode_fwevt_handler()
H A Dcsio_lnode.c907 * @evt - Given rnode event
917 csio_post_event_rns(struct csio_lnode *ln, enum csio_rn_ev evt) csio_post_event_rns() argument
925 csio_post_event(&rn->sm, evt); csio_post_event_rns()
955 * @evt - Given lnode event
965 csio_post_event_lns(struct csio_lnode *ln, enum csio_ln_ev evt) csio_post_event_lns() argument
970 /* If NPIV lnode, send evt only to that and return */ csio_post_event_lns()
972 csio_post_event(&ln->sm, evt); csio_post_event_lns()
977 /* Traverse children lnodes list and send evt */ csio_post_event_lns()
980 csio_post_event(&cln->sm, evt); csio_post_event_lns()
983 /* Send evt to parent lnode */ csio_post_event_lns()
984 csio_post_event(&ln->sm, evt); csio_post_event_lns()
1089 * @evt - Event to be processed.
1096 csio_lns_uninit(struct csio_lnode *ln, enum csio_ln_ev evt) csio_lns_uninit() argument
1102 CSIO_INC_STATS(ln, n_evt_sm[evt]); csio_lns_uninit()
1103 switch (evt) { csio_lns_uninit()
1133 "ln state[uninit].\n", evt, ln->nport_id); csio_lns_uninit()
1142 * @evt - Event to be processed.
1149 csio_lns_online(struct csio_lnode *ln, enum csio_ln_ev evt) csio_lns_online() argument
1153 CSIO_INC_STATS(ln, n_evt_sm[evt]); csio_lns_online()
1154 switch (evt) { csio_lns_online()
1184 "ln state[uninit].\n", evt, ln->nport_id); csio_lns_online()
1194 * @evt - Event to be processed.
1201 csio_lns_ready(struct csio_lnode *ln, enum csio_ln_ev evt) csio_lns_ready() argument
1205 CSIO_INC_STATS(ln, n_evt_sm[evt]); csio_lns_ready()
1206 switch (evt) { csio_lns_ready()
1210 "in ln state[ready].\n", evt, ln->nport_id); csio_lns_ready()
1258 "ln state[uninit].\n", evt, ln->nport_id); csio_lns_ready()
1268 * @evt - Event to be processed.
1275 csio_lns_offline(struct csio_lnode *ln, enum csio_ln_ev evt) csio_lns_offline() argument
1281 CSIO_INC_STATS(ln, n_evt_sm[evt]); csio_lns_offline()
1282 switch (evt) { csio_lns_offline()
1311 "in ln state[offline].\n", evt, ln->nport_id); csio_lns_offline()
1323 "ln state[offline]\n", evt, ln->nport_id); csio_lns_offline()
1476 enum csio_ln_ev evt; csio_fcoe_fwevt_handler() local
1561 evt = CSIO_FWE_TO_LNE(rdev_wr->event_cause); csio_fcoe_fwevt_handler()
1562 if (evt) { csio_fcoe_fwevt_handler()
1565 "cause:%d flowid:x%x\n", evt, csio_fcoe_fwevt_handler()
1567 csio_post_event(&ln->sm, evt); csio_fcoe_fwevt_handler()
1911 /* Traverse children lnodes list and send evt */ csio_notify_lnodes()
1953 /* Traverse sibling lnodes list and send evt */ csio_disable_lnodes()
H A Dcsio_defs.h96 csio_post_event(void *smp, uint32_t evt) csio_post_event() argument
98 ((struct csio_sm *)smp)->sm_state(smp, evt); csio_post_event()
H A Dcsio_hw.c2325 * @evt - Event
2329 csio_hws_uninit(struct csio_hw *hw, enum csio_hw_ev evt) csio_hws_uninit() argument
2332 hw->cur_evt = evt; csio_hws_uninit()
2333 CSIO_INC_STATS(hw, n_evt_sm[evt]); csio_hws_uninit()
2335 switch (evt) { csio_hws_uninit()
2350 * @evt - Event
2354 csio_hws_configuring(struct csio_hw *hw, enum csio_hw_ev evt) csio_hws_configuring() argument
2357 hw->cur_evt = evt; csio_hws_configuring()
2358 CSIO_INC_STATS(hw, n_evt_sm[evt]); csio_hws_configuring()
2360 switch (evt) { csio_hws_configuring()
2388 * @evt - Event
2392 csio_hws_initializing(struct csio_hw *hw, enum csio_hw_ev evt) csio_hws_initializing() argument
2395 hw->cur_evt = evt; csio_hws_initializing()
2396 CSIO_INC_STATS(hw, n_evt_sm[evt]); csio_hws_initializing()
2398 switch (evt) { csio_hws_initializing()
2426 * @evt - Event
2430 csio_hws_ready(struct csio_hw *hw, enum csio_hw_ev evt) csio_hws_ready() argument
2433 hw->evtflag = evt; csio_hws_ready()
2436 hw->cur_evt = evt; csio_hws_ready()
2437 CSIO_INC_STATS(hw, n_evt_sm[evt]); csio_hws_ready()
2439 switch (evt) { csio_hws_ready()
2447 if (evt == CSIO_HWE_HBA_RESET || csio_hws_ready()
2448 evt == CSIO_HWE_PCIERR_DETECTED) csio_hws_ready()
2475 * @evt - Event
2479 csio_hws_quiescing(struct csio_hw *hw, enum csio_hw_ev evt) csio_hws_quiescing() argument
2482 hw->cur_evt = evt; csio_hws_quiescing()
2483 CSIO_INC_STATS(hw, n_evt_sm[evt]); csio_hws_quiescing()
2485 switch (evt) { csio_hws_quiescing()
2535 * @evt - Event
2539 csio_hws_quiesced(struct csio_hw *hw, enum csio_hw_ev evt) csio_hws_quiesced() argument
2542 hw->cur_evt = evt; csio_hws_quiesced()
2543 CSIO_INC_STATS(hw, n_evt_sm[evt]); csio_hws_quiesced()
2545 switch (evt) { csio_hws_quiesced()
2560 * @evt - Event
2564 csio_hws_resetting(struct csio_hw *hw, enum csio_hw_ev evt) csio_hws_resetting() argument
2567 hw->cur_evt = evt; csio_hws_resetting()
2568 CSIO_INC_STATS(hw, n_evt_sm[evt]); csio_hws_resetting()
2570 switch (evt) { csio_hws_resetting()
2586 * @evt - Event
2590 csio_hws_removing(struct csio_hw *hw, enum csio_hw_ev evt) csio_hws_removing() argument
2593 hw->cur_evt = evt; csio_hws_removing()
2594 CSIO_INC_STATS(hw, n_evt_sm[evt]); csio_hws_removing()
2596 switch (evt) { csio_hws_removing()
2621 * @evt - Event
2625 csio_hws_pcierr(struct csio_hw *hw, enum csio_hw_ev evt) csio_hws_pcierr() argument
2628 hw->cur_evt = evt; csio_hws_pcierr()
2629 CSIO_INC_STATS(hw, n_evt_sm[evt]); csio_hws_pcierr()
2631 switch (evt) { csio_hws_pcierr()
3330 csio_err(hw, "Failed to alloc evt entry, msg type %d len %d\n", csio_enqueue_evt()
3373 csio_err(hw, "Failed to alloc evt entry, msg type %d len %d\n", csio_enqueue_evt_lock()
H A Dcsio_hw.h78 #define CSIO_EXTRA_VECS 2 /* non-data + FW evt */
172 struct list_head list; /* evt queue*/
431 int fwevt_iq_idx; /* FW evt queue */
435 struct list_head evt_free_q; /* freelist of evt
438 struct list_head evt_active_q; /* active evt queue*/
462 uint8_t cur_evt; /* current s/m evt */
463 uint8_t prev_evt; /* Previous s/m evt */
474 uint32_t fwevt_intr_idx; /* FW evt MSIX/interrupt
H A Dcsio_scsi.c703 csio_scsis_uninit(struct csio_ioreq *req, enum csio_scsi_ev evt) csio_scsis_uninit() argument
708 switch (evt) { csio_scsis_uninit()
767 csio_dbg(hw, "Unhandled event:%d sent to req:%p\n", evt, req); csio_scsis_uninit()
773 csio_scsis_io_active(struct csio_ioreq *req, enum csio_scsi_ev evt) csio_scsis_io_active() argument
779 switch (evt) { csio_scsis_io_active()
839 csio_dbg(hw, "Unhandled event:%d sent to req:%p\n", evt, req); csio_scsis_io_active()
845 csio_scsis_tm_active(struct csio_ioreq *req, enum csio_scsi_ev evt) csio_scsis_tm_active() argument
850 switch (evt) { csio_scsis_tm_active()
882 csio_dbg(hw, "Unhandled event:%d sent to req:%p\n", evt, req); csio_scsis_tm_active()
888 csio_scsis_aborting(struct csio_ioreq *req, enum csio_scsi_ev evt) csio_scsis_aborting() argument
893 switch (evt) { csio_scsis_aborting()
979 csio_dbg(hw, "Unhandled event:%d sent to req:%p\n", evt, req); csio_scsis_aborting()
985 csio_scsis_closing(struct csio_ioreq *req, enum csio_scsi_ev evt) csio_scsis_closing() argument
990 switch (evt) { csio_scsis_closing()
1043 csio_dbg(hw, "Unhandled event:%d sent to req:%p\n", evt, req); csio_scsis_closing()
1049 csio_scsis_shost_cmpl_await(struct csio_ioreq *req, enum csio_scsi_ev evt) csio_scsis_shost_cmpl_await() argument
1051 switch (evt) { csio_scsis_shost_cmpl_await()
1077 evt, req); csio_scsis_shost_cmpl_await()
H A Dcsio_init.c385 /* Allocate the FW evt queue */ csio_config_queues()
392 csio_err(hw, "FW evt queue creation failed\n"); csio_config_queues()
1063 /* Post PCI error detected evt to HW s/m csio_pci_error_detected()
1064 * HW s/m handles this evt by quiescing IOs, unregisters rports csio_pci_error_detected()
/linux-4.1.27/arch/mips/ralink/
H A Dcevt-rt3352.c40 struct clock_event_device *evt);
43 struct clock_event_device *evt) systick_next_event()
48 sdev = container_of(evt, struct systick_device, dev); systick_next_event()
91 struct clock_event_device *evt) systick_set_clock_mode()
95 sdev = container_of(evt, struct systick_device, dev); systick_set_clock_mode()
42 systick_next_event(unsigned long delta, struct clock_event_device *evt) systick_next_event() argument
90 systick_set_clock_mode(enum clock_event_mode mode, struct clock_event_device *evt) systick_set_clock_mode() argument
/linux-4.1.27/arch/arm/mach-w90x900/
H A Dtime.c78 static int nuc900_clockevent_setnextevent(unsigned long evt, nuc900_clockevent_setnextevent() argument
83 __raw_writel(evt, REG_TICR0); nuc900_clockevent_setnextevent()
104 struct clock_event_device *evt = &nuc900_clockevent_device; nuc900_timer0_interrupt() local
108 evt->event_handler(evt); nuc900_timer0_interrupt()
/linux-4.1.27/arch/arm/mach-gemini/
H A Dtime.c41 struct clock_event_device *evt) gemini_timer_set_next_event()
63 struct clock_event_device *evt) gemini_timer_set_mode()
112 struct clock_event_device *evt = &gemini_clockevent; gemini_timer_interrupt() local
114 evt->event_handler(evt); gemini_timer_interrupt()
40 gemini_timer_set_next_event(unsigned long cycles, struct clock_event_device *evt) gemini_timer_set_next_event() argument
62 gemini_timer_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) gemini_timer_set_mode() argument
/linux-4.1.27/arch/arm/mach-netx/
H A Dtime.c73 static int netx_set_next_event(unsigned long evt, netx_set_next_event() argument
76 writel(0 - evt, NETX_GPIO_COUNTER_CURRENT(TIMER_CLOCKEVENT)); netx_set_next_event()
93 struct clock_event_device *evt = &netx_clockevent; netx_timer_interrupt() local
98 evt->event_handler(evt); netx_timer_interrupt()
/linux-4.1.27/drivers/staging/i2o/
H A Ddriver.c213 struct i2o_event *evt; i2o_driver_dispatch() local
225 evt = kzalloc(size * 4 + sizeof(*evt), GFP_ATOMIC); i2o_driver_dispatch()
226 if (!evt) i2o_driver_dispatch()
229 evt->size = size; i2o_driver_dispatch()
230 evt->tcntxt = le32_to_cpu(msg->u.s.tcntxt); i2o_driver_dispatch()
231 evt->event_indicator = le32_to_cpu(msg->body[0]); i2o_driver_dispatch()
232 memcpy(&evt->data, &msg->body[1], size * 4); i2o_driver_dispatch()
236 evt->i2o_dev = dev; i2o_driver_dispatch()
240 INIT_WORK(&evt->work, drv->event); i2o_driver_dispatch()
241 queue_work(drv->event_queue, &evt->work); i2o_driver_dispatch()
H A Dexec-osm.c526 struct i2o_event *evt = container_of(work, struct i2o_event, work); i2o_exec_event() local
528 if (likely(evt->i2o_dev)) i2o_exec_event()
530 evt->i2o_dev->lct_data.tid); i2o_exec_event()
531 kfree(evt); i2o_exec_event()
H A Di2o_block.c507 struct i2o_event *evt = container_of(work, struct i2o_event, work); i2o_block_event() local
509 kfree(evt); i2o_block_event()
/linux-4.1.27/arch/tile/kernel/
H A Dtime.c131 struct clock_event_device *evt) tile_timer_set_next_event()
144 struct clock_event_device *evt) tile_timer_set_mode()
165 struct clock_event_device *evt = this_cpu_ptr(&tile_timer); setup_tile_timer() local
168 clockevents_calc_mult_shift(evt, cycles_per_sec, TILE_MINSEC); setup_tile_timer()
169 evt->max_delta_ns = clockevent_delta2ns(MAX_TICK, evt); setup_tile_timer()
172 evt->cpumask = cpumask_of(smp_processor_id()); setup_tile_timer()
178 clockevents_register_device(evt); setup_tile_timer()
185 struct clock_event_device *evt = this_cpu_ptr(&tile_timer); do_timer_interrupt() local
200 evt->event_handler(evt); do_timer_interrupt()
130 tile_timer_set_next_event(unsigned long ticks, struct clock_event_device *evt) tile_timer_set_next_event() argument
143 tile_timer_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) tile_timer_set_mode() argument
/linux-4.1.27/arch/nios2/kernel/
H A Dtime.c49 to_nios2_clkevent(struct clock_event_device *evt) to_nios2_clkevent() argument
51 return container_of(evt, struct nios2_clockevent_dev, ced); to_nios2_clkevent()
155 struct clock_event_device *evt) nios2_timer_set_next_event()
157 struct nios2_clockevent_dev *nios2_ced = to_nios2_clkevent(evt); nios2_timer_set_next_event()
159 nios2_timer_config(&nios2_ced->timer, delta, evt->mode); nios2_timer_set_next_event()
165 struct clock_event_device *evt) nios2_timer_set_mode()
168 struct nios2_clockevent_dev *nios2_ced = to_nios2_clkevent(evt); nios2_timer_set_mode()
189 struct clock_event_device *evt = (struct clock_event_device *) dev_id; timer_interrupt() local
190 struct nios2_clockevent_dev *nios2_ced = to_nios2_clkevent(evt); timer_interrupt()
194 evt->event_handler(evt); timer_interrupt()
154 nios2_timer_set_next_event(unsigned long delta, struct clock_event_device *evt) nios2_timer_set_next_event() argument
164 nios2_timer_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) nios2_timer_set_mode() argument
/linux-4.1.27/arch/arm/common/
H A Dtimer-sp.c120 struct clock_event_device *evt = dev_id; sp804_timer_interrupt() local
125 evt->event_handler(evt); sp804_timer_interrupt()
131 struct clock_event_device *evt) sp804_set_mode()
158 struct clock_event_device *evt) sp804_set_next_event()
185 struct clock_event_device *evt = &sp804_clockevent; __sp804_clockevents_init() local
202 evt->name = name; __sp804_clockevents_init()
203 evt->irq = irq; __sp804_clockevents_init()
204 evt->cpumask = cpu_possible_mask; __sp804_clockevents_init()
209 clockevents_config_and_register(evt, rate, 0xf, 0xffffffff); __sp804_clockevents_init()
130 sp804_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) sp804_set_mode() argument
157 sp804_set_next_event(unsigned long next, struct clock_event_device *evt) sp804_set_next_event() argument
/linux-4.1.27/arch/powerpc/platforms/powernv/
H A Dopal.c289 __be64 evt = 0; opal_notifier_enable() local
294 rc = opal_poll_events(&evt); opal_notifier_enable()
295 if (rc == OPAL_SUCCESS && evt) opal_notifier_enable()
296 opal_do_notifier(be64_to_cpu(evt)); opal_notifier_enable()
401 __be64 evt, len; opal_get_chars() local
405 opal_poll_events(&evt); opal_get_chars()
406 if ((be64_to_cpu(evt) & OPAL_EVENT_CONSOLE_INPUT) == 0) opal_get_chars()
421 __be64 evt; opal_put_chars() local
475 opal_poll_events(&evt); opal_put_chars()
477 (be64_to_cpu(evt) & OPAL_EVENT_CONSOLE_OUTPUT)); opal_put_chars()
484 struct machine_check_event *evt) opal_recover_mce()
487 uint64_t ea = get_mce_fault_addr(evt); opal_recover_mce()
492 } else if (evt->disposition == MCE_DISPOSITION_RECOVERED) { opal_recover_mce()
510 evt->severity == MCE_SEV_ERROR_SYNC) { opal_recover_mce()
523 struct machine_check_event evt; opal_machine_check() local
525 if (!get_mce_event(&evt, MCE_EVENT_RELEASE)) opal_machine_check()
529 if (evt.version != MCE_V1) { opal_machine_check()
531 evt.version); opal_machine_check()
534 machine_check_print_event_info(&evt); opal_machine_check()
536 if (opal_recover_mce(regs, &evt)) opal_machine_check()
563 __be64 evt = 0; opal_handle_hmi_exception() local
574 rc = opal_poll_events(&evt); opal_handle_hmi_exception()
575 if (rc == OPAL_SUCCESS && evt) opal_handle_hmi_exception()
576 opal_do_notifier(be64_to_cpu(evt)); opal_handle_hmi_exception()
483 opal_recover_mce(struct pt_regs *regs, struct machine_check_event *evt) opal_recover_mce() argument
/linux-4.1.27/arch/mips/kernel/
H A Dcevt-txx9.c89 struct clock_event_device *evt) txx9tmr_set_mode()
92 container_of(evt, struct txx9_clock_event_device, cd); txx9tmr_set_mode()
101 __raw_writel(((u64)(NSEC_PER_SEC / HZ) * evt->mult) >> txx9tmr_set_mode()
102 evt->shift, txx9tmr_set_mode()
121 struct clock_event_device *evt) txx9tmr_set_next_event()
124 container_of(evt, struct txx9_clock_event_device, cd); txx9tmr_set_next_event()
88 txx9tmr_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) txx9tmr_set_mode() argument
120 txx9tmr_set_next_event(unsigned long delta, struct clock_event_device *evt) txx9tmr_set_next_event() argument
H A Dcevt-ds1287.c57 struct clock_event_device *evt) ds1287_set_next_event()
63 struct clock_event_device *evt) ds1287_set_mode()
56 ds1287_set_next_event(unsigned long delta, struct clock_event_device *evt) ds1287_set_next_event() argument
62 ds1287_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) ds1287_set_mode() argument
H A Dcevt-gt641xx.c49 struct clock_event_device *evt) gt641xx_timer0_set_next_event()
68 struct clock_event_device *evt) gt641xx_timer0_set_mode()
48 gt641xx_timer0_set_next_event(unsigned long delta, struct clock_event_device *evt) gt641xx_timer0_set_next_event() argument
67 gt641xx_timer0_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) gt641xx_timer0_set_mode() argument
H A Dcevt-r4k.c19 struct clock_event_device *evt) mips_next_event()
32 struct clock_event_device *evt) mips_set_clock_mode()
18 mips_next_event(unsigned long delta, struct clock_event_device *evt) mips_next_event() argument
31 mips_set_clock_mode(enum clock_event_mode mode, struct clock_event_device *evt) mips_set_clock_mode() argument
H A Dcevt-bcm1480.c44 struct clock_event_device *evt) sibyte_set_mode()
43 sibyte_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) sibyte_set_mode() argument
H A Dcevt-sb1250.c42 struct clock_event_device *evt) sibyte_set_mode()
41 sibyte_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) sibyte_set_mode() argument
H A Dperf_event_mipsxx.c341 static void mipsxx_pmu_enable_event(struct hw_perf_event *evt, int idx) mipsxx_pmu_enable_event() argument
347 cpuc->saved_ctrl[idx] = M_PERFCTL_EVENT(evt->event_base & 0xff) | mipsxx_pmu_enable_event()
348 (evt->config_base & M_PERFCTL_CONFIG_MASK) | mipsxx_pmu_enable_event()
/linux-4.1.27/arch/arm/mach-imx/
H A Depit.c97 static int epit_set_next_event(unsigned long evt, epit_set_next_event() argument
104 __raw_writel(tcmp - evt, timer_base + EPITCMPR); epit_set_next_event()
110 struct clock_event_device *evt) epit_set_mode()
163 struct clock_event_device *evt = &clockevent_epit; epit_timer_interrupt() local
167 evt->event_handler(evt); epit_timer_interrupt()
109 epit_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) epit_set_mode() argument
H A Dtime.c153 static int mx1_2_set_next_event(unsigned long evt, mx1_2_set_next_event() argument
158 tcmp = __raw_readl(timer_base + MX1_2_TCN) + evt; mx1_2_set_next_event()
166 static int v2_set_next_event(unsigned long evt, v2_set_next_event() argument
171 tcmp = __raw_readl(timer_base + V2_TCN) + evt; v2_set_next_event()
175 return evt < 0x7fffffff && v2_set_next_event()
191 struct clock_event_device *evt) mxc_set_mode()
256 struct clock_event_device *evt = &clockevent_mxc; mxc_timer_interrupt() local
266 evt->event_handler(evt); mxc_timer_interrupt()
190 mxc_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) mxc_set_mode() argument
/linux-4.1.27/arch/arm64/include/asm/
H A Dpmu.h50 void (*enable)(struct hw_perf_event *evt, int idx);
51 void (*disable)(struct hw_perf_event *evt, int idx);
54 int (*set_event_filter)(struct hw_perf_event *evt,
/linux-4.1.27/drivers/usb/dwc3/
H A Dcore.c122 * @evt: Pointer to event buffer to be freed
125 struct dwc3_event_buffer *evt) dwc3_free_one_event_buffer()
127 dma_free_coherent(dwc->dev, evt->length, evt->buf, evt->dma); dwc3_free_one_event_buffer()
141 struct dwc3_event_buffer *evt; dwc3_alloc_one_event_buffer() local
143 evt = devm_kzalloc(dwc->dev, sizeof(*evt), GFP_KERNEL); dwc3_alloc_one_event_buffer()
144 if (!evt) dwc3_alloc_one_event_buffer()
147 evt->dwc = dwc; dwc3_alloc_one_event_buffer()
148 evt->length = length; dwc3_alloc_one_event_buffer()
149 evt->buf = dma_alloc_coherent(dwc->dev, length, dwc3_alloc_one_event_buffer()
150 &evt->dma, GFP_KERNEL); dwc3_alloc_one_event_buffer()
151 if (!evt->buf) dwc3_alloc_one_event_buffer()
154 return evt; dwc3_alloc_one_event_buffer()
163 struct dwc3_event_buffer *evt; dwc3_free_event_buffers() local
167 evt = dwc->ev_buffs[i]; dwc3_free_event_buffers()
168 if (evt) dwc3_free_event_buffers()
169 dwc3_free_one_event_buffer(dwc, evt); dwc3_free_event_buffers()
195 struct dwc3_event_buffer *evt; dwc3_alloc_event_buffers() local
197 evt = dwc3_alloc_one_event_buffer(dwc, length); dwc3_alloc_event_buffers()
198 if (IS_ERR(evt)) { dwc3_alloc_event_buffers()
200 return PTR_ERR(evt); dwc3_alloc_event_buffers()
202 dwc->ev_buffs[i] = evt; dwc3_alloc_event_buffers()
216 struct dwc3_event_buffer *evt; dwc3_event_buffers_setup() local
220 evt = dwc->ev_buffs[n]; dwc3_event_buffers_setup()
222 evt->buf, (unsigned long long) evt->dma, dwc3_event_buffers_setup()
223 evt->length); dwc3_event_buffers_setup()
225 evt->lpos = 0; dwc3_event_buffers_setup()
228 lower_32_bits(evt->dma)); dwc3_event_buffers_setup()
230 upper_32_bits(evt->dma)); dwc3_event_buffers_setup()
232 DWC3_GEVNTSIZ_SIZE(evt->length)); dwc3_event_buffers_setup()
241 struct dwc3_event_buffer *evt; dwc3_event_buffers_cleanup() local
245 evt = dwc->ev_buffs[n]; dwc3_event_buffers_cleanup()
247 evt->lpos = 0; dwc3_event_buffers_cleanup()
124 dwc3_free_one_event_buffer(struct dwc3 *dwc, struct dwc3_event_buffer *evt) dwc3_free_one_event_buffer() argument
H A Dgadget.c2602 struct dwc3_event_buffer *evt; dwc3_process_event_buf() local
2607 evt = dwc->ev_buffs[buf]; dwc3_process_event_buf()
2608 left = evt->count; dwc3_process_event_buf()
2610 if (!(evt->flags & DWC3_EVENT_PENDING)) dwc3_process_event_buf()
2616 event.raw = *(u32 *) (evt->buf + evt->lpos); dwc3_process_event_buf()
2629 evt->lpos = (evt->lpos + 4) % DWC3_EVENT_BUFFERS_SIZE; dwc3_process_event_buf()
2635 evt->count = 0; dwc3_process_event_buf()
2636 evt->flags &= ~DWC3_EVENT_PENDING; dwc3_process_event_buf()
2666 struct dwc3_event_buffer *evt; dwc3_check_event_buf() local
2670 evt = dwc->ev_buffs[buf]; dwc3_check_event_buf()
2677 evt->count = count; dwc3_check_event_buf()
2678 evt->flags |= DWC3_EVENT_PENDING; dwc3_check_event_buf()
/linux-4.1.27/include/linux/
H A Dsh_intc.h16 #define evt2irq(evt) (((evt) >> 5) - 16)
19 #define evt2irq(evt) (evt)
H A Dclockchips.h108 int (*set_next_event)(unsigned long evt, struct clock_event_device *);
169 extern u64 clockevent_delta2ns(unsigned long latch, struct clock_event_device *evt);
H A Dleds.h339 extern void ledtrig_cpu(enum cpu_led_event evt);
341 static inline void ledtrig_cpu(enum cpu_led_event evt) ledtrig_cpu() argument
/linux-4.1.27/arch/arc/kernel/
H A Dtime.c199 * Note that generic IRQ core could have passed @evt for @dev_id if timer_irq_handler()
202 struct clock_event_device *evt = this_cpu_ptr(&arc_clockevent_device); timer_irq_handler() local
203 int irq_reenable = evt->mode == CLOCK_EVT_MODE_PERIODIC; timer_irq_handler()
212 evt->event_handler(evt); timer_irq_handler()
222 struct clock_event_device *evt = this_cpu_ptr(&arc_clockevent_device); arc_local_timer_setup() local
225 evt->cpumask = cpumask_of(cpu); arc_local_timer_setup()
226 clockevents_config_and_register(evt, arc_get_core_freq(), arc_local_timer_setup()
231 "Timer0 (per-cpu-tick)", evt); arc_local_timer_setup()
/linux-4.1.27/drivers/scsi/lpfc/
H A Dlpfc_nportdisc.c813 void *arg, uint32_t evt) lpfc_disc_illegal()
823 (evt == NLP_EVT_CMPL_REG_LOGIN) && lpfc_disc_illegal()
832 ndlp->nlp_DID, evt, ndlp->nlp_state, ndlp->nlp_rpi, lpfc_disc_illegal()
839 void *arg, uint32_t evt) lpfc_cmpl_plogi_illegal()
850 ndlp->nlp_DID, evt, ndlp->nlp_state, ndlp->nlp_rpi, lpfc_cmpl_plogi_illegal()
860 void *arg, uint32_t evt) lpfc_rcv_plogi_unused_node()
874 void *arg, uint32_t evt) lpfc_rcv_els_unused_node()
882 void *arg, uint32_t evt) lpfc_rcv_logo_unused_node()
897 void *arg, uint32_t evt) lpfc_cmpl_logo_unused_node()
904 void *arg, uint32_t evt) lpfc_device_rm_unused_node()
912 void *arg, uint32_t evt) lpfc_device_recov_unused_node()
919 void *arg, uint32_t evt) lpfc_rcv_plogi_plogi_issue()
970 void *arg, uint32_t evt) lpfc_rcv_prli_plogi_issue()
984 void *arg, uint32_t evt) lpfc_rcv_logo_plogi_issue()
997 void *arg, uint32_t evt) lpfc_rcv_els_plogi_issue()
1006 if (evt == NLP_EVT_RCV_LOGO) { lpfc_rcv_els_plogi_issue()
1028 uint32_t evt) lpfc_cmpl_plogi_plogi_issue()
1175 void *arg, uint32_t evt) lpfc_cmpl_logo_plogi_issue()
1182 struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_cmpl_reglogin_plogi_issue()
1201 void *arg, uint32_t evt) lpfc_device_rm_plogi_issue()
1223 uint32_t evt) lpfc_device_recov_plogi_issue()
1248 void *arg, uint32_t evt) lpfc_rcv_plogi_adisc_issue()
1278 void *arg, uint32_t evt) lpfc_rcv_prli_adisc_issue()
1288 void *arg, uint32_t evt) lpfc_rcv_logo_adisc_issue()
1305 void *arg, uint32_t evt) lpfc_rcv_padisc_adisc_issue()
1317 void *arg, uint32_t evt) lpfc_rcv_prlo_adisc_issue()
1331 void *arg, uint32_t evt) lpfc_cmpl_adisc_adisc_issue()
1387 void *arg, uint32_t evt) lpfc_device_rm_adisc_issue()
1409 uint32_t evt) lpfc_device_recov_adisc_issue()
1436 uint32_t evt) lpfc_rcv_plogi_reglogin_issue()
1448 uint32_t evt) lpfc_rcv_prli_reglogin_issue()
1460 uint32_t evt) lpfc_rcv_logo_reglogin_issue()
1505 uint32_t evt) lpfc_rcv_padisc_reglogin_issue()
1517 uint32_t evt) lpfc_rcv_prlo_reglogin_issue()
1530 uint32_t evt) lpfc_cmpl_reglogin_reglogin_issue()
1591 uint32_t evt) lpfc_device_rm_reglogin_issue()
1610 uint32_t evt) lpfc_device_recov_reglogin_issue()
1632 void *arg, uint32_t evt) lpfc_rcv_plogi_prli_issue()
1644 void *arg, uint32_t evt) lpfc_rcv_prli_prli_issue()
1654 void *arg, uint32_t evt) lpfc_rcv_logo_prli_issue()
1667 void *arg, uint32_t evt) lpfc_rcv_padisc_prli_issue()
1682 void *arg, uint32_t evt) lpfc_rcv_prlo_prli_issue()
1692 void *arg, uint32_t evt) lpfc_cmpl_prli_prli_issue()
1760 * \param evt
1773 void *arg, uint32_t evt) lpfc_device_rm_prli_issue()
1799 * \param evt
1812 uint32_t evt) lpfc_device_recov_prli_issue()
1837 void *arg, uint32_t evt) lpfc_rcv_plogi_logo_issue()
1851 void *arg, uint32_t evt) lpfc_rcv_prli_logo_issue()
1865 void *arg, uint32_t evt) lpfc_rcv_logo_logo_issue()
1879 void *arg, uint32_t evt) lpfc_rcv_padisc_logo_issue()
1893 void *arg, uint32_t evt) lpfc_rcv_prlo_logo_issue()
1907 void *arg, uint32_t evt) lpfc_cmpl_logo_logo_issue()
1922 void *arg, uint32_t evt) lpfc_device_rm_logo_issue()
1936 void *arg, uint32_t evt) lpfc_device_recov_logo_issue()
1948 void *arg, uint32_t evt) lpfc_rcv_plogi_unmap_node()
1958 void *arg, uint32_t evt) lpfc_rcv_prli_unmap_node()
1969 void *arg, uint32_t evt) lpfc_rcv_logo_unmap_node()
1979 void *arg, uint32_t evt) lpfc_rcv_padisc_unmap_node()
1989 void *arg, uint32_t evt) lpfc_rcv_prlo_unmap_node()
2001 uint32_t evt) lpfc_device_recov_unmap_node()
2017 void *arg, uint32_t evt) lpfc_rcv_plogi_mapped_node()
2027 void *arg, uint32_t evt) lpfc_rcv_prli_mapped_node()
2037 void *arg, uint32_t evt) lpfc_rcv_logo_mapped_node()
2048 void *arg, uint32_t evt) lpfc_rcv_padisc_mapped_node()
2058 void *arg, uint32_t evt) lpfc_rcv_prlo_mapped_node()
2076 uint32_t evt) lpfc_device_recov_mapped_node()
2091 void *arg, uint32_t evt) lpfc_rcv_plogi_npr_node()
2117 void *arg, uint32_t evt) lpfc_rcv_prli_npr_node()
2147 void *arg, uint32_t evt) lpfc_rcv_logo_npr_node()
2157 void *arg, uint32_t evt) lpfc_rcv_padisc_npr_node()
2185 void *arg, uint32_t evt) lpfc_rcv_prlo_npr_node()
2214 void *arg, uint32_t evt) lpfc_cmpl_plogi_npr_node()
2235 void *arg, uint32_t evt) lpfc_cmpl_prli_npr_node()
2253 void *arg, uint32_t evt) lpfc_cmpl_logo_npr_node()
2269 void *arg, uint32_t evt) lpfc_cmpl_adisc_npr_node()
2288 void *arg, uint32_t evt) lpfc_cmpl_reglogin_npr_node()
2309 void *arg, uint32_t evt) lpfc_device_rm_npr_node()
2325 void *arg, uint32_t evt) lpfc_device_recov_npr_node()
2531 void *arg, uint32_t evt) lpfc_disc_state_machine()
2543 /* DSM in event <evt> on NPort <nlp_DID> in state <cur_state> */ lpfc_disc_state_machine()
2547 evt, ndlp->nlp_DID, cur_state, ndlp->nlp_flag); lpfc_disc_state_machine()
2550 "DSM in: evt:%d ste:%d did:x%x", lpfc_disc_state_machine()
2551 evt, cur_state, ndlp->nlp_DID); lpfc_disc_state_machine()
2553 func = lpfc_disc_action[(cur_state * NLP_EVT_MAX_EVENT) + evt]; lpfc_disc_state_machine()
2554 rc = (func) (vport, ndlp, arg, evt); lpfc_disc_state_machine()
812 lpfc_disc_illegal(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_disc_illegal() argument
838 lpfc_cmpl_plogi_illegal(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_cmpl_plogi_illegal() argument
859 lpfc_rcv_plogi_unused_node(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_plogi_unused_node() argument
873 lpfc_rcv_els_unused_node(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_els_unused_node() argument
881 lpfc_rcv_logo_unused_node(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_logo_unused_node() argument
896 lpfc_cmpl_logo_unused_node(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_cmpl_logo_unused_node() argument
903 lpfc_device_rm_unused_node(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_device_rm_unused_node() argument
910 lpfc_device_recov_unused_node(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_device_recov_unused_node() argument
918 lpfc_rcv_plogi_plogi_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_plogi_plogi_issue() argument
969 lpfc_rcv_prli_plogi_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_prli_plogi_issue() argument
983 lpfc_rcv_logo_plogi_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_logo_plogi_issue() argument
996 lpfc_rcv_els_plogi_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_els_plogi_issue() argument
1025 lpfc_cmpl_plogi_plogi_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_cmpl_plogi_plogi_issue() argument
1174 lpfc_cmpl_logo_plogi_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_cmpl_logo_plogi_issue() argument
1181 lpfc_cmpl_reglogin_plogi_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_cmpl_reglogin_plogi_issue() argument
1200 lpfc_device_rm_plogi_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_device_rm_plogi_issue() argument
1220 lpfc_device_recov_plogi_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_device_recov_plogi_issue() argument
1247 lpfc_rcv_plogi_adisc_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_plogi_adisc_issue() argument
1277 lpfc_rcv_prli_adisc_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_prli_adisc_issue() argument
1287 lpfc_rcv_logo_adisc_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_logo_adisc_issue() argument
1303 lpfc_rcv_padisc_adisc_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_padisc_adisc_issue() argument
1316 lpfc_rcv_prlo_adisc_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_prlo_adisc_issue() argument
1329 lpfc_cmpl_adisc_adisc_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_cmpl_adisc_adisc_issue() argument
1386 lpfc_device_rm_adisc_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_device_rm_adisc_issue() argument
1406 lpfc_device_recov_adisc_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_device_recov_adisc_issue() argument
1433 lpfc_rcv_plogi_reglogin_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_plogi_reglogin_issue() argument
1445 lpfc_rcv_prli_reglogin_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_prli_reglogin_issue() argument
1457 lpfc_rcv_logo_reglogin_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_logo_reglogin_issue() argument
1502 lpfc_rcv_padisc_reglogin_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_padisc_reglogin_issue() argument
1514 lpfc_rcv_prlo_reglogin_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_prlo_reglogin_issue() argument
1527 lpfc_cmpl_reglogin_reglogin_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_cmpl_reglogin_reglogin_issue() argument
1588 lpfc_device_rm_reglogin_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_device_rm_reglogin_issue() argument
1607 lpfc_device_recov_reglogin_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_device_recov_reglogin_issue() argument
1631 lpfc_rcv_plogi_prli_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_plogi_prli_issue() argument
1643 lpfc_rcv_prli_prli_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_prli_prli_issue() argument
1653 lpfc_rcv_logo_prli_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_logo_prli_issue() argument
1666 lpfc_rcv_padisc_prli_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_padisc_prli_issue() argument
1681 lpfc_rcv_prlo_prli_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_prlo_prli_issue() argument
1691 lpfc_cmpl_prli_prli_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_cmpl_prli_prli_issue() argument
1772 lpfc_device_rm_prli_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_device_rm_prli_issue() argument
1809 lpfc_device_recov_prli_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_device_recov_prli_issue() argument
1836 lpfc_rcv_plogi_logo_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_plogi_logo_issue() argument
1850 lpfc_rcv_prli_logo_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_prli_logo_issue() argument
1864 lpfc_rcv_logo_logo_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_logo_logo_issue() argument
1878 lpfc_rcv_padisc_logo_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_padisc_logo_issue() argument
1892 lpfc_rcv_prlo_logo_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_prlo_logo_issue() argument
1906 lpfc_cmpl_logo_logo_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_cmpl_logo_logo_issue() argument
1921 lpfc_device_rm_logo_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_device_rm_logo_issue() argument
1934 lpfc_device_recov_logo_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_device_recov_logo_issue() argument
1947 lpfc_rcv_plogi_unmap_node(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_plogi_unmap_node() argument
1957 lpfc_rcv_prli_unmap_node(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_prli_unmap_node() argument
1968 lpfc_rcv_logo_unmap_node(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_logo_unmap_node() argument
1978 lpfc_rcv_padisc_unmap_node(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_padisc_unmap_node() argument
1988 lpfc_rcv_prlo_unmap_node(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_prlo_unmap_node() argument
1998 lpfc_device_recov_unmap_node(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_device_recov_unmap_node() argument
2016 lpfc_rcv_plogi_mapped_node(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_plogi_mapped_node() argument
2026 lpfc_rcv_prli_mapped_node(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_prli_mapped_node() argument
2036 lpfc_rcv_logo_mapped_node(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_logo_mapped_node() argument
2046 lpfc_rcv_padisc_mapped_node(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_padisc_mapped_node() argument
2057 lpfc_rcv_prlo_mapped_node(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_prlo_mapped_node() argument
2073 lpfc_device_recov_mapped_node(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_device_recov_mapped_node() argument
2090 lpfc_rcv_plogi_npr_node(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_plogi_npr_node() argument
2116 lpfc_rcv_prli_npr_node(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_prli_npr_node() argument
2146 lpfc_rcv_logo_npr_node(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_logo_npr_node() argument
2156 lpfc_rcv_padisc_npr_node(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_padisc_npr_node() argument
2184 lpfc_rcv_prlo_npr_node(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_prlo_npr_node() argument
2213 lpfc_cmpl_plogi_npr_node(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_cmpl_plogi_npr_node() argument
2234 lpfc_cmpl_prli_npr_node(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_cmpl_prli_npr_node() argument
2252 lpfc_cmpl_logo_npr_node(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_cmpl_logo_npr_node() argument
2268 lpfc_cmpl_adisc_npr_node(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_cmpl_adisc_npr_node() argument
2286 lpfc_cmpl_reglogin_npr_node(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_cmpl_reglogin_npr_node() argument
2308 lpfc_device_rm_npr_node(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_device_rm_npr_node() argument
2324 lpfc_device_recov_npr_node(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_device_recov_npr_node() argument
2530 lpfc_disc_state_machine(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_disc_state_machine() argument
H A Dlpfc_bsg.c102 struct lpfc_bsg_event *evt; member in union:bsg_job_data::__anon9070
790 struct lpfc_bsg_event *evt = container_of(kref, struct lpfc_bsg_event, lpfc_bsg_event_free() local
794 list_del(&evt->node); lpfc_bsg_event_free()
796 while (!list_empty(&evt->events_to_get)) { lpfc_bsg_event_free()
797 ed = list_entry(evt->events_to_get.next, typeof(*ed), node); lpfc_bsg_event_free()
803 while (!list_empty(&evt->events_to_see)) { lpfc_bsg_event_free()
804 ed = list_entry(evt->events_to_see.next, typeof(*ed), node); lpfc_bsg_event_free()
810 kfree(evt->dd_data); lpfc_bsg_event_free()
811 kfree(evt); lpfc_bsg_event_free()
816 * @evt: Pointer to an event structure.
819 lpfc_bsg_event_ref(struct lpfc_bsg_event *evt) lpfc_bsg_event_ref() argument
821 kref_get(&evt->kref); lpfc_bsg_event_ref()
826 * @evt: Pointer to an event structure.
829 lpfc_bsg_event_unref(struct lpfc_bsg_event *evt) lpfc_bsg_event_unref() argument
831 kref_put(&evt->kref, lpfc_bsg_event_free); lpfc_bsg_event_unref()
843 struct lpfc_bsg_event *evt = kzalloc(sizeof(*evt), GFP_KERNEL); lpfc_bsg_event_new() local
845 if (!evt) lpfc_bsg_event_new()
848 INIT_LIST_HEAD(&evt->events_to_get); lpfc_bsg_event_new()
849 INIT_LIST_HEAD(&evt->events_to_see); lpfc_bsg_event_new()
850 evt->type_mask = ev_mask; lpfc_bsg_event_new()
851 evt->req_id = ev_req_id; lpfc_bsg_event_new()
852 evt->reg_id = ev_reg_id; lpfc_bsg_event_new()
853 evt->wait_time_stamp = jiffies; lpfc_bsg_event_new()
854 evt->dd_data = NULL; lpfc_bsg_event_new()
855 init_waitqueue_head(&evt->wq); lpfc_bsg_event_new()
856 kref_init(&evt->kref); lpfc_bsg_event_new()
857 return evt; lpfc_bsg_event_new()
909 struct lpfc_bsg_event *evt; lpfc_bsg_ct_unsol_event() local
954 list_for_each_entry(evt, &phba->ct_ev_waiters, node) { lpfc_bsg_ct_unsol_event()
955 if (!(evt->type_mask & FC_REG_CT_EVENT) || lpfc_bsg_ct_unsol_event()
956 evt->req_id != evt_req_id) lpfc_bsg_ct_unsol_event()
959 lpfc_bsg_event_ref(evt); lpfc_bsg_ct_unsol_event()
964 lpfc_bsg_event_unref(evt); lpfc_bsg_ct_unsol_event()
991 lpfc_bsg_event_unref(evt); lpfc_bsg_ct_unsol_event()
1037 lpfc_bsg_event_unref(evt); lpfc_bsg_ct_unsol_event()
1111 list_add(&evt_dat->node, &evt->events_to_see); lpfc_bsg_ct_unsol_event()
1113 wake_up_interruptible(&evt->wq); lpfc_bsg_ct_unsol_event()
1114 lpfc_bsg_event_unref(evt); lpfc_bsg_ct_unsol_event()
1118 list_move(evt->events_to_see.prev, &evt->events_to_get); lpfc_bsg_ct_unsol_event()
1120 dd_data = (struct bsg_job_data *)evt->dd_data; lpfc_bsg_ct_unsol_event()
1123 lpfc_bsg_event_unref(evt); lpfc_bsg_ct_unsol_event()
1197 struct lpfc_bsg_event *evt; lpfc_bsg_hba_set_event() local
1217 list_for_each_entry(evt, &phba->ct_ev_waiters, node) { lpfc_bsg_hba_set_event()
1218 if (evt->reg_id == event_req->ev_reg_id) { lpfc_bsg_hba_set_event()
1219 lpfc_bsg_event_ref(evt); lpfc_bsg_hba_set_event()
1220 evt->wait_time_stamp = jiffies; lpfc_bsg_hba_set_event()
1221 dd_data = (struct bsg_job_data *)evt->dd_data; lpfc_bsg_hba_set_event()
1227 if (&evt->node == &phba->ct_ev_waiters) { lpfc_bsg_hba_set_event()
1236 evt = lpfc_bsg_event_new(ev_mask, event_req->ev_reg_id, lpfc_bsg_hba_set_event()
1238 if (!evt) { lpfc_bsg_hba_set_event()
1247 dd_data->context_un.evt = evt; lpfc_bsg_hba_set_event()
1248 evt->dd_data = (void *)dd_data; lpfc_bsg_hba_set_event()
1250 list_add(&evt->node, &phba->ct_ev_waiters); lpfc_bsg_hba_set_event()
1251 lpfc_bsg_event_ref(evt); lpfc_bsg_hba_set_event()
1252 evt->wait_time_stamp = jiffies; lpfc_bsg_hba_set_event()
1257 evt->waiting = 1; lpfc_bsg_hba_set_event()
1282 struct lpfc_bsg_event *evt, *evt_next; lpfc_bsg_hba_get_event() local
1302 list_for_each_entry_safe(evt, evt_next, &phba->ct_ev_waiters, node) { lpfc_bsg_hba_get_event()
1303 if (evt->reg_id == event_req->ev_reg_id) { lpfc_bsg_hba_get_event()
1304 if (list_empty(&evt->events_to_get)) lpfc_bsg_hba_get_event()
1306 lpfc_bsg_event_ref(evt); lpfc_bsg_hba_get_event()
1307 evt->wait_time_stamp = jiffies; lpfc_bsg_hba_get_event()
1308 evt_dat = list_entry(evt->events_to_get.prev, lpfc_bsg_hba_get_event()
1349 lpfc_bsg_event_unref(evt); lpfc_bsg_hba_get_event()
2537 struct lpfc_bsg_event *evt; lpfcdiag_loop_get_xri() local
2550 evt = lpfc_bsg_event_new(FC_REG_CT_EVENT, current->pid, lpfcdiag_loop_get_xri()
2552 if (!evt) lpfcdiag_loop_get_xri()
2556 list_add(&evt->node, &phba->ct_ev_waiters); lpfcdiag_loop_get_xri()
2557 lpfc_bsg_event_ref(evt); lpfcdiag_loop_get_xri()
2633 evt->waiting = 1; lpfcdiag_loop_get_xri()
2634 evt->wait_time_stamp = jiffies; lpfcdiag_loop_get_xri()
2636 evt->wq, !list_empty(&evt->events_to_see), lpfcdiag_loop_get_xri()
2639 if (list_empty(&evt->events_to_see)) lpfcdiag_loop_get_xri()
2643 list_move(evt->events_to_see.prev, &evt->events_to_get); lpfcdiag_loop_get_xri()
2645 *rxxri = (list_entry(evt->events_to_get.prev, lpfcdiag_loop_get_xri()
2649 evt->waiting = 0; lpfcdiag_loop_get_xri()
2653 lpfc_bsg_event_unref(evt); /* release ref */ lpfcdiag_loop_get_xri()
2654 lpfc_bsg_event_unref(evt); /* delete */ lpfcdiag_loop_get_xri()
2992 struct lpfc_bsg_event *evt; lpfc_bsg_diag_loopback_run() local
3101 evt = lpfc_bsg_event_new(FC_REG_CT_EVENT, current->pid, lpfc_bsg_diag_loopback_run()
3103 if (!evt) { lpfc_bsg_diag_loopback_run()
3110 list_add(&evt->node, &phba->ct_ev_waiters); lpfc_bsg_diag_loopback_run()
3111 lpfc_bsg_event_ref(evt); lpfc_bsg_diag_loopback_run()
3214 evt->waiting = 1; lpfc_bsg_diag_loopback_run()
3216 evt->wq, !list_empty(&evt->events_to_see), lpfc_bsg_diag_loopback_run()
3219 evt->waiting = 0; lpfc_bsg_diag_loopback_run()
3220 if (list_empty(&evt->events_to_see)) { lpfc_bsg_diag_loopback_run()
3227 list_move(evt->events_to_see.prev, &evt->events_to_get); lpfc_bsg_diag_loopback_run()
3228 evdat = list_entry(evt->events_to_get.prev, lpfc_bsg_diag_loopback_run()
3257 lpfc_bsg_event_unref(evt); /* release ref */ lpfc_bsg_diag_loopback_run()
3258 lpfc_bsg_event_unref(evt); /* delete */ lpfc_bsg_diag_loopback_run()
H A Dlpfc_disc.h49 enum lpfc_work_type evt; member in struct:lpfc_work_evt
H A Dlpfc_hbadisc.c192 evtp->evt = LPFC_EVT_DEV_LOSS; lpfc_dev_loss_tmo_callbk()
436 ret->work_evt.evt = LPFC_EVT_FASTPATH_MGMT_EVT; lpfc_alloc_fast_evt()
444 * @evt: Event object which need to be freed.
451 struct lpfc_fast_path_event *evt) { lpfc_free_fast_evt()
454 kfree(evt); lpfc_free_fast_evt()
547 switch (evtp->evt) { lpfc_work_list_done()
551 free_evt = 0; /* evt is part of ndlp */ lpfc_work_list_done()
796 uint32_t evt) lpfc_workq_post_event()
811 evtp->evt = evt; lpfc_workq_post_event()
450 lpfc_free_fast_evt(struct lpfc_hba *phba, struct lpfc_fast_path_event *evt) lpfc_free_fast_evt() argument
795 lpfc_workq_post_event(struct lpfc_hba *phba, void *arg1, void *arg2, uint32_t evt) lpfc_workq_post_event() argument
H A Dlpfc_debugfs.h416 * lpfc_debug_dump_hba_eq - dump all entries from a fcp work queue's evt queue
/linux-4.1.27/arch/um/kernel/
H A Dtime.c26 struct clock_event_device *evt) itimer_set_mode()
45 struct clock_event_device *evt) itimer_next_event()
25 itimer_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) itimer_set_mode() argument
44 itimer_next_event(unsigned long delta, struct clock_event_device *evt) itimer_next_event() argument
/linux-4.1.27/arch/c6x/kernel/
H A Dsoc.c26 void soc_assert_event(unsigned int evt) soc_assert_event() argument
29 soc_ops.assert_event(evt); soc_assert_event()
H A Dtraps.c302 int evt; process_eexcept() local
306 while ((evt = soc_get_exception()) >= 0) process_eexcept()
307 do_trap(&eexcept_table[evt], regs); process_eexcept()
/linux-4.1.27/arch/arm/mach-ks8695/
H A Dtime.c58 struct clock_event_device *evt) ks8695_set_mode()
82 struct clock_event_device *evt) ks8695_set_next_event()
117 struct clock_event_device *evt = &clockevent_ks8695; ks8695_timer_interrupt() local
119 evt->event_handler(evt); ks8695_timer_interrupt()
57 ks8695_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) ks8695_set_mode() argument
81 ks8695_set_next_event(unsigned long cycles, struct clock_event_device *evt) ks8695_set_next_event() argument
/linux-4.1.27/arch/arm/mach-omap1/
H A Dtime.c121 struct clock_event_device *evt) omap_mpu_set_next_event()
128 struct clock_event_device *evt) omap_mpu_set_mode()
154 struct clock_event_device *evt = &clockevent_mpu_timer1; omap_mpu_timer1_interrupt() local
156 evt->event_handler(evt); omap_mpu_timer1_interrupt()
120 omap_mpu_set_next_event(unsigned long cycles, struct clock_event_device *evt) omap_mpu_set_next_event() argument
127 omap_mpu_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) omap_mpu_set_mode() argument
H A Dtimer32k.c118 struct clock_event_device *evt) omap_32k_timer_set_mode()
144 struct clock_event_device *evt = &clockevent_32k_timer; omap_32k_timer_interrupt() local
147 evt->event_handler(evt); omap_32k_timer_interrupt()
117 omap_32k_timer_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) omap_32k_timer_set_mode() argument
/linux-4.1.27/arch/x86/kernel/
H A Dhpet.c47 struct clock_event_device evt; member in struct:hpet_dev
57 return container_of(evtdev, struct hpet_dev, evt); EVT_TO_HPET_DEV()
229 struct clock_event_device *evt); hpet_reserve_platform_timers()
231 struct clock_event_device *evt); hpet_reserve_platform_timers()
311 struct clock_event_device *evt, int timer) hpet_set_mode()
319 delta = ((uint64_t)(NSEC_PER_SEC/HZ)) * evt->mult; hpet_set_mode()
320 delta >>= evt->shift; hpet_set_mode()
359 struct hpet_dev *hdev = EVT_TO_HPET_DEV(evt); hpet_set_mode()
371 struct clock_event_device *evt, int timer) hpet_next_event()
408 struct clock_event_device *evt) hpet_legacy_set_mode()
410 hpet_set_mode(mode, evt, 0); hpet_legacy_set_mode()
414 struct clock_event_device *evt) hpet_legacy_next_event()
416 return hpet_next_event(delta, evt, 0); hpet_legacy_next_event()
463 struct clock_event_device *evt) hpet_msi_set_mode()
465 struct hpet_dev *hdev = EVT_TO_HPET_DEV(evt); hpet_msi_set_mode()
466 hpet_set_mode(mode, evt, hdev->num); hpet_msi_set_mode()
470 struct clock_event_device *evt) hpet_msi_next_event()
472 struct hpet_dev *hdev = EVT_TO_HPET_DEV(evt); hpet_msi_next_event()
473 return hpet_next_event(delta, evt, hdev->num); hpet_msi_next_event()
504 struct clock_event_device *hevt = &dev->evt; hpet_interrupt_handler()
537 struct clock_event_device *evt = &hdev->evt; init_one_hpet_msi_clockevent() local
548 evt->name = hdev->name; init_one_hpet_msi_clockevent()
550 evt->irq = hdev->irq; init_one_hpet_msi_clockevent()
552 evt->rating = 110; init_one_hpet_msi_clockevent()
553 evt->features = CLOCK_EVT_FEAT_ONESHOT; init_one_hpet_msi_clockevent()
555 evt->features |= CLOCK_EVT_FEAT_PERIODIC; init_one_hpet_msi_clockevent()
557 evt->set_mode = hpet_msi_set_mode; init_one_hpet_msi_clockevent()
558 evt->set_next_event = hpet_msi_next_event; init_one_hpet_msi_clockevent()
559 evt->cpumask = cpumask_of(hdev->cpu); init_one_hpet_msi_clockevent()
561 clockevents_config_and_register(evt, hpet_freq, HPET_MIN_PROG_DELTA, init_one_hpet_msi_clockevent()
310 hpet_set_mode(enum clock_event_mode mode, struct clock_event_device *evt, int timer) hpet_set_mode() argument
370 hpet_next_event(unsigned long delta, struct clock_event_device *evt, int timer) hpet_next_event() argument
407 hpet_legacy_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) hpet_legacy_set_mode() argument
413 hpet_legacy_next_event(unsigned long delta, struct clock_event_device *evt) hpet_legacy_next_event() argument
462 hpet_msi_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) hpet_msi_set_mode() argument
469 hpet_msi_next_event(unsigned long delta, struct clock_event_device *evt) hpet_msi_next_event() argument
/linux-4.1.27/net/tipc/
H A Dsubscr.c73 msg_sect.iov_base = (void *)&sub->evt; subscr_send_event()
75 sub->evt.event = htohl(event, sub->swap); subscr_send_event()
76 sub->evt.found_lower = htohl(found_lower, sub->swap); subscr_send_event()
77 sub->evt.found_upper = htohl(found_upper, sub->swap); subscr_send_event()
78 sub->evt.port.ref = htohl(port_ref, sub->swap); subscr_send_event()
79 sub->evt.port.node = htohl(node, sub->swap); subscr_send_event()
142 subscr_send_event(sub, sub->evt.s.seq.lower, sub->evt.s.seq.upper, subscr_timeout()
207 if (!memcmp(s, &sub->evt.s, sizeof(struct tipc_subscr))) { subscr_cancel()
279 memcpy(&sub->evt.s, s, sizeof(struct tipc_subscr)); subscr_subscribe()
H A Dsubscr.h60 * @evt: template for events generated by subscription
72 struct tipc_event evt; member in struct:tipc_subscription
H A Dbearer.c510 * @evt: the type of event
516 static int tipc_l2_device_event(struct notifier_block *nb, unsigned long evt, tipc_l2_device_event() argument
529 switch (evt) { tipc_l2_device_event()
/linux-4.1.27/arch/arm/plat-iop/
H A Dtime.c120 struct clock_event_device *evt = dev_id; iop_timer_interrupt() local
123 evt->event_handler(evt); iop_timer_interrupt()
/linux-4.1.27/arch/arm/kernel/
H A Dperf_event_v6.c262 unsigned long val, mask, evt, flags; armv6pmu_enable_event() local
270 evt = ARMV6_PMCR_CCOUNT_IEN; armv6pmu_enable_event()
273 evt = (hwc->config_base << ARMV6_PMCR_EVT_COUNT0_SHIFT) | armv6pmu_enable_event()
277 evt = (hwc->config_base << ARMV6_PMCR_EVT_COUNT1_SHIFT) | armv6pmu_enable_event()
291 val |= evt; armv6pmu_enable_event()
409 unsigned long val, mask, evt, flags; armv6pmu_disable_event() local
417 evt = 0; armv6pmu_disable_event()
420 evt = ARMV6_PERFCTR_NOP << ARMV6_PMCR_EVT_COUNT0_SHIFT; armv6pmu_disable_event()
423 evt = ARMV6_PERFCTR_NOP << ARMV6_PMCR_EVT_COUNT1_SHIFT; armv6pmu_disable_event()
437 val |= evt; armv6pmu_disable_event()
444 unsigned long val, mask, flags, evt = 0; armv6mpcore_pmu_disable_event() local
468 val |= evt; armv6mpcore_pmu_disable_event()
H A Dperf_event_xscale.c198 unsigned long val, mask, evt, flags; xscale1pmu_enable_event() local
207 evt = XSCALE1_CCOUNT_INT_EN; xscale1pmu_enable_event()
211 evt = (hwc->config_base << XSCALE1_COUNT0_EVT_SHFT) | xscale1pmu_enable_event()
216 evt = (hwc->config_base << XSCALE1_COUNT1_EVT_SHFT) | xscale1pmu_enable_event()
227 val |= evt; xscale1pmu_enable_event()
234 unsigned long val, mask, evt, flags; xscale1pmu_disable_event() local
243 evt = 0; xscale1pmu_disable_event()
247 evt = XSCALE_PERFCTR_UNUSED << XSCALE1_COUNT0_EVT_SHFT; xscale1pmu_disable_event()
251 evt = XSCALE_PERFCTR_UNUSED << XSCALE1_COUNT1_EVT_SHFT; xscale1pmu_disable_event()
261 val |= evt; xscale1pmu_disable_event()
H A Dsmp_twd.c64 static int twd_set_next_event(unsigned long evt, twd_set_next_event() argument
71 writel_relaxed(evt, twd_base + TWD_TIMER_COUNTER); twd_set_next_event()
233 struct clock_event_device *evt = dev_id; twd_handler() local
236 evt->event_handler(evt); twd_handler()
/linux-4.1.27/arch/arm/mach-lpc32xx/
H A Dtimer.c80 struct clock_event_device *evt = &lpc32xx_clkevt; lpc32xx_timer_interrupt() local
86 evt->event_handler(evt); lpc32xx_timer_interrupt()
/linux-4.1.27/drivers/gpio/
H A Dgpio-kempld.c125 u16 evt, evt_back; kempld_gpio_pincount() local
134 evt = kempld_read16(pld, KEMPLD_GPIO_EVT_LVL_EDGE); kempld_gpio_pincount()
140 return evt ? __ffs(evt) : 16; kempld_gpio_pincount()
/linux-4.1.27/arch/openrisc/kernel/
H A Dtime.c52 struct clock_event_device *evt) openrisc_timer_set_mode()
110 struct clock_event_device *evt = &clockevent_openrisc_timer; timer_interrupt() local
118 evt->event_handler(evt); timer_interrupt()
51 openrisc_timer_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) openrisc_timer_set_mode() argument
/linux-4.1.27/arch/c6x/include/asm/
H A Dsoc.h20 void (*assert_event)(unsigned int evt);
/linux-4.1.27/arch/arm/mach-spear/
H A Dtime.c71 static int clockevent_next_event(unsigned long evt,
164 struct clock_event_device *evt = &clkevt; spear_timer_interrupt() local
168 evt->event_handler(evt); spear_timer_interrupt()
/linux-4.1.27/drivers/gpu/drm/armada/
H A Darmada_drv.c234 struct armada_vbl_event *evt) armada_drm_vbl_event_add()
239 if (list_empty(&evt->node)) { armada_drm_vbl_event_add()
240 list_add_tail(&evt->node, &dcrtc->vbl_list); armada_drm_vbl_event_add()
248 struct armada_vbl_event *evt) armada_drm_vbl_event_remove()
250 if (!list_empty(&evt->node)) { armada_drm_vbl_event_remove()
251 list_del_init(&evt->node); armada_drm_vbl_event_remove()
257 struct armada_vbl_event *evt) armada_drm_vbl_event_remove_unlocked()
262 armada_drm_vbl_event_remove(dcrtc, evt); armada_drm_vbl_event_remove_unlocked()
233 armada_drm_vbl_event_add(struct armada_crtc *dcrtc, struct armada_vbl_event *evt) armada_drm_vbl_event_add() argument
247 armada_drm_vbl_event_remove(struct armada_crtc *dcrtc, struct armada_vbl_event *evt) armada_drm_vbl_event_remove() argument
256 armada_drm_vbl_event_remove_unlocked(struct armada_crtc *dcrtc, struct armada_vbl_event *evt) armada_drm_vbl_event_remove_unlocked() argument
/linux-4.1.27/drivers/devfreq/exynos/
H A Dexynos_ppmu.c30 unsigned int evt) exynos_ppmu_setevent()
32 __raw_writel(evt, ppmu_base + PPMU_BEVTSEL(ch)); exynos_ppmu_setevent()
29 exynos_ppmu_setevent(void __iomem *ppmu_base, unsigned int ch, unsigned int evt) exynos_ppmu_setevent() argument
H A Dexynos_ppmu.h79 unsigned int evt);
/linux-4.1.27/arch/mips/jz4740/
H A Dtime.c86 static int jz4740_clockevent_set_next(unsigned long evt, jz4740_clockevent_set_next() argument
90 jz4740_timer_set_period(TIMER_CLOCKEVENT, evt); jz4740_clockevent_set_next()
/linux-4.1.27/drivers/infiniband/core/
H A Ducm.c252 static int ib_ucm_event_process(struct ib_cm_event *evt, ib_ucm_event_process() argument
257 switch (evt->event) { ib_ucm_event_process()
260 &evt->param.req_rcvd); ib_ucm_event_process()
263 uvt->resp.present |= (evt->param.req_rcvd.alternate_path ? ib_ucm_event_process()
268 &evt->param.rep_rcvd); ib_ucm_event_process()
273 uvt->resp.u.send_status = evt->param.send_status; ib_ucm_event_process()
277 uvt->resp.u.send_status = evt->param.send_status; ib_ucm_event_process()
281 uvt->resp.u.send_status = evt->param.send_status; ib_ucm_event_process()
285 evt->param.mra_rcvd.service_timeout; ib_ucm_event_process()
289 uvt->resp.u.rej_resp.reason = evt->param.rej_rcvd.reason; ib_ucm_event_process()
291 uvt->info_len = evt->param.rej_rcvd.ari_length; ib_ucm_event_process()
292 info = evt->param.rej_rcvd.ari; ib_ucm_event_process()
296 evt->param.lap_rcvd.alternate_path); ib_ucm_event_process()
301 uvt->resp.u.apr_resp.status = evt->param.apr_rcvd.ap_status; ib_ucm_event_process()
303 uvt->info_len = evt->param.apr_rcvd.info_len; ib_ucm_event_process()
304 info = evt->param.apr_rcvd.apr_info; ib_ucm_event_process()
308 evt->param.sidr_req_rcvd.pkey; ib_ucm_event_process()
310 evt->param.sidr_req_rcvd.port; ib_ucm_event_process()
315 &evt->param.sidr_rep_rcvd); ib_ucm_event_process()
317 uvt->info_len = evt->param.sidr_rep_rcvd.info_len; ib_ucm_event_process()
318 info = evt->param.sidr_rep_rcvd.info; ib_ucm_event_process()
321 uvt->resp.u.send_status = evt->param.send_status; ib_ucm_event_process()
326 uvt->data = kmemdup(evt->private_data, uvt->data_len, GFP_KERNEL); ib_ucm_event_process()
H A Duverbs_main.c158 struct ib_uverbs_event *evt, *tmp; ib_uverbs_release_ucq() local
162 list_for_each_entry_safe(evt, tmp, &uobj->comp_list, obj_list) { ib_uverbs_release_ucq()
163 list_del(&evt->list); ib_uverbs_release_ucq()
164 kfree(evt); ib_uverbs_release_ucq()
172 list_for_each_entry_safe(evt, tmp, &uobj->async_list, obj_list) { ib_uverbs_release_ucq()
173 list_del(&evt->list); ib_uverbs_release_ucq()
174 kfree(evt); ib_uverbs_release_ucq()
182 struct ib_uverbs_event *evt, *tmp; ib_uverbs_release_uevent() local
185 list_for_each_entry_safe(evt, tmp, &uobj->event_list, obj_list) { ib_uverbs_release_uevent()
186 list_del(&evt->list); ib_uverbs_release_uevent()
187 kfree(evt); ib_uverbs_release_uevent()
/linux-4.1.27/arch/sparc/kernel/
H A Dtime_32.c113 struct clock_event_device *evt) timer_ce_set_mode()
195 struct clock_event_device *evt) percpu_ce_setup()
197 int cpu = cpumask_first(evt->cpumask); percpu_ce_setup()
215 struct clock_event_device *evt) percpu_ce_set_next_event()
217 int cpu = cpumask_first(evt->cpumask); percpu_ce_set_next_event()
112 timer_ce_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) timer_ce_set_mode() argument
194 percpu_ce_setup(enum clock_event_mode mode, struct clock_event_device *evt) percpu_ce_setup() argument
214 percpu_ce_set_next_event(unsigned long delta, struct clock_event_device *evt) percpu_ce_set_next_event() argument
H A Dtime_64.c686 struct clock_event_device *evt) sparc64_next_event()
692 struct clock_event_device *evt) sparc64_timer_setup()
725 struct clock_event_device *evt = &per_cpu(sparc64_events, cpu); timer_interrupt() local
734 if (unlikely(!evt->event_handler)) { timer_interrupt()
738 evt->event_handler(evt); timer_interrupt()
685 sparc64_next_event(unsigned long delta, struct clock_event_device *evt) sparc64_next_event() argument
691 sparc64_timer_setup(enum clock_event_mode mode, struct clock_event_device *evt) sparc64_timer_setup() argument
/linux-4.1.27/arch/cris/arch-v32/kernel/
H A Dtime.c186 static int crisv32_clkevt_next_event(unsigned long evt, crisv32_clkevt_next_event() argument
194 REG_WR(timer, timer_base, rw_tmr0_div, evt); crisv32_clkevt_next_event()
205 struct clock_event_device *evt = dev_id; crisv32_timer_interrupt() local
225 evt->event_handler(evt); crisv32_timer_interrupt()
/linux-4.1.27/arch/arm/mach-cns3xxx/
H A Dcore.c142 static int cns3xxx_timer_set_next_event(unsigned long evt, cns3xxx_timer_set_next_event() argument
147 writel(evt, cns3xxx_tmr1 + TIMER1_AUTO_RELOAD_OFFSET); cns3xxx_timer_set_next_event()
175 struct clock_event_device *evt = &cns3xxx_tmr1_clockevent; cns3xxx_timer_interrupt() local
183 evt->event_handler(evt); cns3xxx_timer_interrupt()
/linux-4.1.27/arch/mips/loongson/common/cs5536/
H A Dcs5536_mfgpt.c55 struct clock_event_device *evt) init_mfgpt_timer()
68 if (evt->mode == CLOCK_EVT_MODE_PERIODIC || init_mfgpt_timer()
69 evt->mode == CLOCK_EVT_MODE_ONESHOT) init_mfgpt_timer()
54 init_mfgpt_timer(enum clock_event_mode mode, struct clock_event_device *evt) init_mfgpt_timer() argument
/linux-4.1.27/kernel/time/
H A Dclockevents.c36 static u64 cev_delta2ns(unsigned long latch, struct clock_event_device *evt, cev_delta2ns() argument
39 u64 clc = (u64) latch << evt->shift; cev_delta2ns()
42 if (unlikely(!evt->mult)) { cev_delta2ns()
43 evt->mult = 1; cev_delta2ns()
46 rnd = (u64) evt->mult - 1; cev_delta2ns()
52 if ((clc >> evt->shift) != (u64)latch) cev_delta2ns()
75 (!ismax || evt->mult <= (1ULL << evt->shift))) cev_delta2ns()
78 do_div(clc, evt->mult); cev_delta2ns()
87 * @evt: pointer to clock event device descriptor
91 u64 clockevent_delta2ns(unsigned long latch, struct clock_event_device *evt) clockevent_delta2ns() argument
93 return cev_delta2ns(latch, evt, false); clockevent_delta2ns()
H A Dtick-broadcast.c242 struct clock_event_device *evt = td->evtdev; tick_receive_broadcast() local
244 if (!evt) tick_receive_broadcast()
247 if (!evt->event_handler) tick_receive_broadcast()
250 evt->event_handler(evt); tick_receive_broadcast()
/linux-4.1.27/drivers/virtio/
H A Dvirtio_input.c15 struct virtqueue *evt, *sts; member in struct:virtio_input
27 virtqueue_add_inbuf(vi->evt, sg, 1, evtbuf, GFP_ATOMIC); virtinput_queue_evtbuf()
39 while ((event = virtqueue_get_buf(vi->evt, &len)) != NULL) { virtinput_recv_events()
179 vi->evt = vqs[0]; virtinput_init_vqs()
191 size = virtqueue_get_vring_size(vi->evt); virtinput_fill_evt()
196 virtqueue_kick(vi->evt); virtinput_fill_evt()
/linux-4.1.27/arch/metag/kernel/perf/
H A Dperf_event.h81 void (*enable)(struct hw_perf_event *evt, int idx);
82 void (*disable)(struct hw_perf_event *evt, int idx);
/linux-4.1.27/arch/sh/boards/mach-cayman/
H A Dirq.c98 int cayman_irq_demux(int evt) cayman_irq_demux() argument
100 int irq = intc_evt_to_irq[evt]; cayman_irq_demux()
/linux-4.1.27/arch/mn10300/kernel/
H A Dcevt-mn10300.c30 struct clock_event_device *evt) next_event()
45 struct clock_event_device *evt) set_clock_mode()
29 next_event(unsigned long delta, struct clock_event_device *evt) next_event() argument
44 set_clock_mode(enum clock_event_mode mode, struct clock_event_device *evt) set_clock_mode() argument
/linux-4.1.27/arch/microblaze/kernel/
H A Dtimer.c126 struct clock_event_device *evt) xilinx_timer_set_mode()
165 struct clock_event_device *evt = &clockevent_xilinx_timer; timer_interrupt() local
170 evt->event_handler(evt); timer_interrupt()
125 xilinx_timer_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) xilinx_timer_set_mode() argument
/linux-4.1.27/arch/sh/kernel/cpu/sh3/
H A Dex.S57 * location via set_exception_table_{evt,vec}().
/linux-4.1.27/arch/arm/mach-davinci/
H A Dtime.c167 struct clock_event_device *evt = &clockevent_davinci; timer_interrupt() local
169 evt->event_handler(evt); timer_interrupt()
297 struct clock_event_device *evt) davinci_set_next_event()
307 struct clock_event_device *evt) davinci_set_mode()
296 davinci_set_next_event(unsigned long cycles, struct clock_event_device *evt) davinci_set_next_event() argument
306 davinci_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) davinci_set_mode() argument
/linux-4.1.27/drivers/vhost/
H A Dscsi.c212 struct vhost_work vs_event_work; /* evt injection work item */
213 struct llist_head vs_event_list; /* evt injection queue */
567 static void vhost_scsi_free_evt(struct vhost_scsi *vs, struct vhost_scsi_evt *evt) vhost_scsi_free_evt() argument
570 kfree(evt); vhost_scsi_free_evt()
578 struct vhost_scsi_evt *evt; vhost_scsi_allocate_evt() local
585 evt = kzalloc(sizeof(*evt), GFP_KERNEL); vhost_scsi_allocate_evt()
586 if (!evt) { vhost_scsi_allocate_evt()
592 evt->event.event = cpu_to_vhost32(vq, event); vhost_scsi_allocate_evt()
593 evt->event.reason = cpu_to_vhost32(vq, reason); vhost_scsi_allocate_evt()
596 return evt; vhost_scsi_allocate_evt()
614 vhost_scsi_do_evt_work(struct vhost_scsi *vs, struct vhost_scsi_evt *evt) vhost_scsi_do_evt_work() argument
617 struct virtio_scsi_event *event = &evt->event; vhost_scsi_do_evt_work()
668 struct vhost_scsi_evt *evt; vhost_scsi_evt_work() local
674 evt = llist_entry(llnode, struct vhost_scsi_evt, list); vhost_scsi_evt_work()
676 vhost_scsi_do_evt_work(vs, evt); vhost_scsi_evt_work()
677 vhost_scsi_free_evt(vs, evt); vhost_scsi_evt_work()
1265 struct vhost_scsi_evt *evt; vhost_scsi_send_evt() local
1267 evt = vhost_scsi_allocate_evt(vs, event, reason); vhost_scsi_send_evt()
1268 if (!evt) vhost_scsi_send_evt()
1274 * Note: evt->event is zeroed when we allocate it and vhost_scsi_send_evt()
1277 evt->event.lun[0] = 0x01; vhost_scsi_send_evt()
1278 evt->event.lun[1] = tpg->tport_tpgt; vhost_scsi_send_evt()
1280 evt->event.lun[2] = lun->unpacked_lun >> 8 | 0x40 ; vhost_scsi_send_evt()
1281 evt->event.lun[3] = lun->unpacked_lun & 0xFF; vhost_scsi_send_evt()
1284 llist_add(&evt->list, &vs->vs_event_list); vhost_scsi_send_evt()
1625 /* Jobs can re-queue themselves in evt kick handler. Do extra flush. */ vhost_scsi_release()
/linux-4.1.27/drivers/firewire/
H A Dohci.c379 static void log_irqs(struct fw_ohci *ohci, u32 evt) log_irqs() argument
386 !(evt & OHCI1394_busReset)) log_irqs()
389 ohci_notice(ohci, "IRQ %08x%s%s%s%s%s%s%s%s%s%s%s%s%s%s%s\n", evt, log_irqs()
390 evt & OHCI1394_selfIDComplete ? " selfID" : "", log_irqs()
391 evt & OHCI1394_RQPkt ? " AR_req" : "", log_irqs()
392 evt & OHCI1394_RSPkt ? " AR_resp" : "", log_irqs()
393 evt & OHCI1394_reqTxComplete ? " AT_req" : "", log_irqs()
394 evt & OHCI1394_respTxComplete ? " AT_resp" : "", log_irqs()
395 evt & OHCI1394_isochRx ? " IR" : "", log_irqs()
396 evt & OHCI1394_isochTx ? " IT" : "", log_irqs()
397 evt & OHCI1394_postedWriteErr ? " postedWriteErr" : "", log_irqs()
398 evt & OHCI1394_cycleTooLong ? " cycleTooLong" : "", log_irqs()
399 evt & OHCI1394_cycle64Seconds ? " cycle64Seconds" : "", log_irqs()
400 evt & OHCI1394_cycleInconsistent ? " cycleInconsistent" : "", log_irqs()
401 evt & OHCI1394_regAccessFail ? " regAccessFail" : "", log_irqs()
402 evt & OHCI1394_unrecoverableError ? " unrecoverableError" : "", log_irqs()
403 evt & OHCI1394_busReset ? " busReset" : "", log_irqs()
404 evt & ~(OHCI1394_selfIDComplete | OHCI1394_RQPkt | log_irqs()
485 char dir, int speed, u32 *header, int evt) log_ar_at_event()
493 if (unlikely(evt >= ARRAY_SIZE(evts))) log_ar_at_event()
494 evt = 0x1f; log_ar_at_event()
496 if (evt == OHCI1394_evt_bus_reset) { log_ar_at_event()
518 dir, evts[evt], tcodes[tcode]); log_ar_at_event()
522 dir, evts[evt], header[1], header[2]); log_ar_at_event()
528 header[1] >> 16, header[0] >> 16, evts[evt], log_ar_at_event()
535 header[1] >> 16, header[0] >> 16, evts[evt], log_ar_at_event()
818 int evt; handle_ar_packet() local
869 evt = (status >> 16) & 0x1f; handle_ar_packet()
871 p.ack = evt - 16; handle_ar_packet()
876 log_ar_at_event(ohci, 'R', p.speed, p.header, evt); handle_ar_packet()
882 if (evt == OHCI1394_evt_no_status && handle_ar_packet()
899 if (evt == OHCI1394_evt_bus_reset) { handle_ar_packet()
1445 int evt; handle_at_packet() local
1461 evt = le16_to_cpu(last->transfer_status) & 0x1f; handle_at_packet()
1464 log_ar_at_event(ohci, 'T', packet->speed, packet->header, evt); handle_at_packet()
1466 switch (evt) { handle_at_packet()
1499 packet->ack = evt - 0x10; handle_at_packet()
484 log_ar_at_event(struct fw_ohci *ohci, char dir, int speed, u32 *header, int evt) log_ar_at_event() argument
/linux-4.1.27/include/linux/uwb/
H A Dspec.h685 const struct uwb_rc_evt_bp_slot_change *evt) uwb_rc_evt_bp_slot_change_slot_num()
687 return evt->slot_info & 0x7f; uwb_rc_evt_bp_slot_change_slot_num()
691 const struct uwb_rc_evt_bp_slot_change *evt) uwb_rc_evt_bp_slot_change_no_slot()
693 return (evt->slot_info & 0x80) >> 7; uwb_rc_evt_bp_slot_change_no_slot()
719 static inline enum uwb_drp_notif_reason uwb_rc_evt_drp_reason(struct uwb_rc_evt_drp *evt) uwb_rc_evt_drp_reason() argument
721 return evt->reason & 0x0f; uwb_rc_evt_drp_reason()
684 uwb_rc_evt_bp_slot_change_slot_num( const struct uwb_rc_evt_bp_slot_change *evt) uwb_rc_evt_bp_slot_change_slot_num() argument
690 uwb_rc_evt_bp_slot_change_no_slot( const struct uwb_rc_evt_bp_slot_change *evt) uwb_rc_evt_bp_slot_change_no_slot() argument
/linux-4.1.27/arch/arm/mach-ixp4xx/
H A Dcommon.c312 struct clock_event_device *evt = dev_id; ixp4xx_timer_interrupt() local
317 evt->event_handler(evt); ixp4xx_timer_interrupt()
514 static int ixp4xx_set_next_event(unsigned long evt, ixp4xx_set_next_event() argument
519 *IXP4XX_OSRT1 = (evt & ~IXP4XX_OST_RELOAD_MASK) | opts; ixp4xx_set_next_event()
525 struct clock_event_device *evt) ixp4xx_set_mode()
524 ixp4xx_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) ixp4xx_set_mode() argument
/linux-4.1.27/arch/mips/loongson/loongson-3/
H A Dhpet.c82 struct clock_event_device *evt) hpet_set_mode()
133 struct clock_event_device *evt) hpet_next_event()
81 hpet_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) hpet_set_mode() argument
132 hpet_next_event(unsigned long delta, struct clock_event_device *evt) hpet_next_event() argument
/linux-4.1.27/arch/mips/loongson1/common/
H A Dtime.c151 static int ls1x_clockevent_set_next(unsigned long evt, ls1x_clockevent_set_next() argument
155 ls1x_pwmtimer_set_period(evt); ls1x_clockevent_set_next()
/linux-4.1.27/arch/mips/sgi-ip27/
H A Dip27-timer.c54 static int rt_next_event(unsigned long delta, struct clock_event_device *evt) rt_next_event() argument
68 struct clock_event_device *evt) rt_set_mode()
67 rt_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) rt_set_mode() argument
/linux-4.1.27/arch/powerpc/include/asm/
H A Dmce.h195 extern void machine_check_print_event_info(struct machine_check_event *evt);
196 extern uint64_t get_mce_fault_addr(struct machine_check_event *evt);
/linux-4.1.27/arch/c6x/platforms/
H A Dtimer64.c123 struct clock_event_device *evt) next_event()
130 struct clock_event_device *evt) set_clock_mode()
122 next_event(unsigned long delta, struct clock_event_device *evt) next_event() argument
129 set_clock_mode(enum clock_event_mode mode, struct clock_event_device *evt) set_clock_mode() argument
/linux-4.1.27/arch/blackfin/mach-bf609/
H A Ddpm.S102 /* Enable evt 11 in IMASK before idle, otherwise core doesn't wake up. */
/linux-4.1.27/drivers/bluetooth/
H A Dbtusb.c1196 static void btusb_notify(struct hci_dev *hdev, unsigned int evt) btusb_notify() argument
1200 BT_DBG("%s evt %d", hdev->name, evt); btusb_notify()
1808 struct hci_event_hdr *evt = NULL; btusb_setup_intel_patching() local
1862 evt = (struct hci_event_hdr *)(*fw_ptr); btusb_setup_intel_patching()
1863 *fw_ptr += sizeof(*evt); btusb_setup_intel_patching()
1864 remain -= sizeof(*evt); btusb_setup_intel_patching()
1866 if (remain < evt->plen) { btusb_setup_intel_patching()
1867 BT_ERR("%s Intel fw corrupted: invalid evt len", btusb_setup_intel_patching()
1873 *fw_ptr += evt->plen; btusb_setup_intel_patching()
1874 remain -= evt->plen; btusb_setup_intel_patching()
1881 if (!evt || !evt_param || remain < 0) { btusb_setup_intel_patching()
1882 BT_ERR("%s Intel fw corrupted: invalid evt read", hdev->name); btusb_setup_intel_patching()
1887 cmd_param, evt->evt, HCI_INIT_TIMEOUT); btusb_setup_intel_patching()
1898 if (skb->len != evt->plen) { btusb_setup_intel_patching()
1905 if (memcmp(skb->data, evt_param, evt->plen)) { btusb_setup_intel_patching()
2129 struct hci_ev_cmd_complete *evt; inject_cmd_complete() local
2131 skb = bt_skb_alloc(sizeof(*hdr) + sizeof(*evt) + 1, GFP_ATOMIC); inject_cmd_complete()
2136 hdr->evt = HCI_EV_CMD_COMPLETE; inject_cmd_complete()
2137 hdr->plen = sizeof(*evt) + 1; inject_cmd_complete()
2139 evt = (struct hci_ev_cmd_complete *)skb_put(skb, sizeof(*evt)); inject_cmd_complete()
2140 evt->ncmd = 0x01; inject_cmd_complete()
2141 evt->opcode = cpu_to_le16(opcode); inject_cmd_complete()
2166 const struct intel_bootup *evt = ptr; btusb_intel_bootup() local
2168 if (len != sizeof(*evt)) btusb_intel_bootup()
2180 const struct intel_secure_send_result *evt = ptr; btusb_intel_secure_send_result() local
2182 if (len != sizeof(*evt)) btusb_intel_secure_send_result()
2185 if (evt->result) btusb_intel_secure_send_result()
2202 if (skb->len > HCI_EVENT_HDR_SIZE && hdr->evt == 0xff && btusb_recv_event_intel()
H A Dhci_ll.c406 BT_DBG("Event header: evt 0x%2.2x plen %d", eh->evt, eh->plen); ll_recv()
H A Dhci_bcsp.c191 chan = 5; /* BCSP cmd/evt channel */ bcsp_prepare_pkt()
523 hdr.evt = 0xff; bcsp_complete_rx_pkt()
H A Dbtmrvl_main.c58 if (hdr->evt == HCI_EV_CMD_COMPLETE) { btmrvl_check_evtpkt()
/linux-4.1.27/drivers/ptp/
H A Dptp_clock.c291 struct pps_event_time evt; ptp_clock_event() local
304 pps_get_ts(&evt); ptp_clock_event()
305 pps_event(ptp->pps_source, &evt, PTP_PPS_EVENT, NULL); ptp_clock_event()
/linux-4.1.27/arch/blackfin/mach-common/
H A Dsmp.c135 struct clock_event_device *evt = &per_cpu(coretmr_events, cpu); ipi_timer() local
136 evt->event_handler(evt); ipi_timer()
/linux-4.1.27/drivers/net/ethernet/sfc/
H A Dptp.c930 struct efx_ptp_event_rx *evt; efx_ptp_drop_time_expired_events() local
932 evt = list_entry(cursor, struct efx_ptp_event_rx, efx_ptp_drop_time_expired_events()
934 if (time_after(jiffies, evt->expiry)) { efx_ptp_drop_time_expired_events()
935 list_move(&evt->link, &ptp->evt_free_list); efx_ptp_drop_time_expired_events()
967 struct efx_ptp_event_rx *evt; efx_ptp_match_rx() local
969 evt = list_entry(cursor, struct efx_ptp_event_rx, link); efx_ptp_match_rx()
970 if ((evt->seq0 == match->words[0]) && efx_ptp_match_rx()
971 (evt->seq1 == match->words[1])) { efx_ptp_match_rx()
976 timestamps->hwtstamp = evt->hwtimestamp; efx_ptp_match_rx()
980 list_move(&evt->link, &ptp->evt_free_list); efx_ptp_match_rx()
1596 struct efx_ptp_event_rx *evt = NULL; ptp_event_rx() local
1608 evt = list_first_entry(&ptp->evt_free_list, ptp_event_rx()
1610 list_del(&evt->link); ptp_event_rx()
1612 evt->seq0 = EFX_QWORD_FIELD(ptp->evt_frags[2], MCDI_EVENT_DATA); ptp_event_rx()
1613 evt->seq1 = (EFX_QWORD_FIELD(ptp->evt_frags[2], ptp_event_rx()
1619 evt->hwtimestamp = efx->ptp_data->nic_to_kernel_time( ptp_event_rx()
1623 evt->expiry = jiffies + msecs_to_jiffies(PKT_EVENT_LIFETIME_MS); ptp_event_rx()
1624 list_add_tail(&evt->link, &ptp->evt_list); ptp_event_rx()
/linux-4.1.27/drivers/scsi/be2iscsi/
H A Dbe_cmds.c437 struct be_async_event_link_state *evt) beiscsi_async_link_state_process()
439 if ((evt->port_link_status == ASYNC_EVENT_LINK_DOWN) || beiscsi_async_link_state_process()
440 ((evt->port_link_status & ASYNC_EVENT_LOGICAL) && beiscsi_async_link_state_process()
441 (evt->port_fault != BEISCSI_PHY_LINK_FAULT_NONE))) { beiscsi_async_link_state_process()
447 evt->physical_port); beiscsi_async_link_state_process()
451 } else if ((evt->port_link_status & ASYNC_EVENT_LINK_UP) || beiscsi_async_link_state_process()
452 ((evt->port_link_status & ASYNC_EVENT_LOGICAL) && beiscsi_async_link_state_process()
453 (evt->port_fault == BEISCSI_PHY_LINK_FAULT_NONE))) { beiscsi_async_link_state_process()
459 evt->physical_port); beiscsi_async_link_state_process()
474 /* Interpret compl as a async link evt */ beiscsi_process_mcc()
436 beiscsi_async_link_state_process(struct beiscsi_hba *phba, struct be_async_event_link_state *evt) beiscsi_async_link_state_process() argument
H A Dbe_cmds.h741 struct be_async_event_link_state *evt);
/linux-4.1.27/drivers/hv/
H A Dhv.c270 struct clock_event_device *evt) hv_ce_set_next_event()
274 WARN_ON(evt->mode != CLOCK_EVT_MODE_ONESHOT); hv_ce_set_next_event()
283 struct clock_event_device *evt) hv_ce_setmode()
269 hv_ce_set_next_event(unsigned long delta, struct clock_event_device *evt) hv_ce_set_next_event() argument
282 hv_ce_setmode(enum clock_event_mode mode, struct clock_event_device *evt) hv_ce_setmode() argument
/linux-4.1.27/arch/alpha/kernel/
H A Dtime.c116 rtc_ce_set_next_event(unsigned long evt, struct clock_event_device *ce) rtc_ce_set_next_event()
175 qemu_ce_set_next_event(unsigned long evt, struct clock_event_device *ce) qemu_ce_set_next_event()
177 qemu_set_alarm_rel(evt);
115 rtc_ce_set_next_event(unsigned long evt, struct clock_event_device *ce) rtc_ce_set_next_event() argument
173 qemu_ce_set_next_event(unsigned long evt, struct clock_event_device *ce) qemu_ce_set_next_event() argument
/linux-4.1.27/drivers/leds/trigger/
H A Dledtrig-cpu.c43 * @evt: CPU event to be emitted
/linux-4.1.27/drivers/net/ethernet/alteon/
H A Dacenic.h319 u32 evt:8; member in struct:event
321 u32 evt:8;
365 u32 evt:8; member in struct:cmd
367 u32 evt:8;
H A Dacenic.c1671 cmd.evt = C_SET_RX_PRD_IDX; ace_load_std_rx_ring()
1788 cmd.evt = C_SET_RX_JUMBO_PRD_IDX; ace_load_jumbo_rx_ring()
1820 switch (ap->evt_ring[evtcsm].evt) { ace_handle_event()
1891 cmd.evt = C_SET_RX_JUMBO_PRD_IDX; ace_handle_event()
1909 ap->name, ap->evt_ring[evtcsm].evt); ace_handle_event()
2243 cmd.evt = C_CLEAR_STATS; ace_open()
2248 cmd.evt = C_HOST_STATE; ace_open()
2258 cmd.evt = C_SET_PROMISC_MODE; ace_open()
2269 cmd.evt = C_LNK_NEGOTIATION; ace_open()
2302 cmd.evt = C_SET_PROMISC_MODE; ace_close()
2309 cmd.evt = C_HOST_STATE; ace_close()
2354 cmd.evt = C_RESET_JUMBO_RNG; ace_close()
2573 cmd.evt = C_RESET_JUMBO_RNG; ace_change_mtu()
2690 cmd.evt = C_LNK_NEGOTIATION; ace_set_settings()
2736 cmd.evt = C_SET_MAC_ADDR; ace_set_mac_addr()
2752 cmd.evt = C_SET_MULTICAST_MODE; ace_set_multicast_list()
2758 cmd.evt = C_SET_MULTICAST_MODE; ace_set_multicast_list()
2766 cmd.evt = C_SET_PROMISC_MODE; ace_set_multicast_list()
2772 cmd.evt = C_SET_PROMISC_MODE; ace_set_multicast_list()
2786 cmd.evt = C_SET_MULTICAST_MODE; ace_set_multicast_list()
2791 cmd.evt = C_SET_MULTICAST_MODE; ace_set_multicast_list()
/linux-4.1.27/arch/mips/jazz/
H A Dirq.c114 struct clock_event_device *evt) r4030_set_mode()
113 r4030_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) r4030_set_mode() argument
/linux-4.1.27/net/nfc/hci/
H A Dhci.h29 u8 header; /* type -cmd,evt,rsp- + instruction */
/linux-4.1.27/drivers/net/ethernet/emulex/benet/
H A Dbe_cmds.c241 /* Link state evt is a string of bytes; no need for endian swapping */ be_async_link_state_process()
245 struct be_async_event_link_state *evt = be_async_link_state_process() local
257 !(evt->port_link_status & LOGICAL_LINK_STATUS_MASK)) be_async_link_state_process()
265 evt->port_link_status & LINK_STATUS_MASK); be_async_link_state_process()
271 struct be_async_event_misconfig_port *evt = be_async_port_misconfig_event_process() local
273 u32 sfp_mismatch_evt = le32_to_cpu(evt->event_data_word1); be_async_port_misconfig_event_process()
291 /* Grp5 CoS Priority evt */ be_async_grp5_cos_priority_process()
295 struct be_async_event_grp5_cos_priority *evt = be_async_grp5_cos_priority_process() local
298 if (evt->valid) { be_async_grp5_cos_priority_process()
299 adapter->vlan_prio_bmap = evt->available_priority_bmap; be_async_grp5_cos_priority_process()
302 evt->reco_default_priority << VLAN_PRIO_SHIFT; be_async_grp5_cos_priority_process()
306 /* Grp5 QOS Speed evt: qos_link_speed is in units of 10 Mbps */ be_async_grp5_qos_speed_process()
310 struct be_async_event_grp5_qos_link_speed *evt = be_async_grp5_qos_speed_process() local
314 evt->physical_port == adapter->port_num) be_async_grp5_qos_speed_process()
315 adapter->phy.link_speed = le16_to_cpu(evt->qos_link_speed) * 10; be_async_grp5_qos_speed_process()
318 /*Grp5 PVID evt*/ be_async_grp5_pvid_state_process()
322 struct be_async_event_grp5_pvid_state *evt = be_async_grp5_pvid_state_process() local
325 if (evt->enabled) { be_async_grp5_pvid_state_process()
326 adapter->pvid = le16_to_cpu(evt->tag) & VLAN_VID_MASK; be_async_grp5_pvid_state_process()
358 struct be_async_event_qnq *evt = (struct be_async_event_qnq *)cmp; be_async_dbg_evt_process() local
365 if (evt->valid) be_async_dbg_evt_process()
366 adapter->qnq_vid = le16_to_cpu(evt->vlan_tag); be_async_dbg_evt_process()
H A Dbe_hw.h190 u32 evt; member in struct:be_eq_entry
H A Dbe_main.c802 /* Hack to skip HW VLAN tagging needs evt = 1, compl = 0. When this wrb_fill_hdr()
803 * hack is not needed, the evt bit is set while ringing DB. wrb_fill_hdr()
2198 if (eqe->evt == 0) events_get()
2202 eqe->evt = 0; events_get()
/linux-4.1.27/drivers/scsi/
H A Dgdth.c158 u16 idx, gdth_evt_data *evt);
2838 u16 idx, gdth_evt_data *evt) gdth_store_event()
2850 ((evt->size != 0 && ebuffer[elastidx].event_data.size != 0 && gdth_store_event()
2852 (char *)&evt->eu, evt->size)) || gdth_store_event()
2853 (evt->size == 0 && ebuffer[elastidx].event_data.size == 0 && gdth_store_event()
2855 (char *)&evt->event_string)))) { gdth_store_event()
2877 e->event_data = *evt; gdth_store_event()
4059 gdth_ioctl_event evt; ioc_event() local
4063 if (copy_from_user(&evt, arg, sizeof(gdth_ioctl_event))) ioc_event()
4065 ha = gdth_find_ha(evt.ionode); ioc_event()
4069 if (evt.erase == 0xff) { ioc_event()
4070 if (evt.event.event_source == ES_TEST) ioc_event()
4071 evt.event.event_data.size=sizeof(evt.event.event_data.eu.test); ioc_event()
4072 else if (evt.event.event_source == ES_DRIVER) ioc_event()
4073 evt.event.event_data.size=sizeof(evt.event.event_data.eu.driver); ioc_event()
4074 else if (evt.event.event_source == ES_SYNC) ioc_event()
4075 evt.event.event_data.size=sizeof(evt.event.event_data.eu.sync); ioc_event()
4077 evt.event.event_data.size=sizeof(evt.event.event_data.eu.async); ioc_event()
4079 gdth_store_event(ha, evt.event.event_source, evt.event.event_idx, ioc_event()
4080 &evt.event.event_data); ioc_event()
4082 } else if (evt.erase == 0xfe) { ioc_event()
4084 } else if (evt.erase == 0) { ioc_event()
4085 evt.handle = gdth_read_event(ha, evt.handle, &evt.event); ioc_event()
4087 gdth_readapp_event(ha, evt.erase, &evt.event); ioc_event()
4089 if (copy_to_user(arg, &evt, sizeof(gdth_ioctl_event))) ioc_event()
2837 gdth_store_event(gdth_ha_str *ha, u16 source, u16 idx, gdth_evt_data *evt) gdth_store_event() argument
H A Dscsi_lib.c2685 * @evt: event to emit
2689 static void scsi_evt_emit(struct scsi_device *sdev, struct scsi_event *evt) scsi_evt_emit() argument
2694 switch (evt->evt_type) { scsi_evt_emit()
2743 struct scsi_event *evt; scsi_evt_thread() local
2755 evt = list_entry(this, struct scsi_event, node); scsi_evt_thread()
2756 list_del(&evt->node); scsi_evt_thread()
2757 scsi_evt_emit(sdev, evt); scsi_evt_thread()
2758 kfree(evt); scsi_evt_thread()
2766 * @evt: event to send
2770 void sdev_evt_send(struct scsi_device *sdev, struct scsi_event *evt) sdev_evt_send() argument
2778 if (!test_bit(evt->evt_type, sdev->supported_events)) { sdev_evt_send()
2779 kfree(evt); sdev_evt_send()
2785 list_add_tail(&evt->node, &sdev->event_list); sdev_evt_send()
2801 struct scsi_event *evt = kzalloc(sizeof(struct scsi_event), gfpflags); sdev_evt_alloc() local
2802 if (!evt) sdev_evt_alloc()
2805 evt->evt_type = evt_type; sdev_evt_alloc()
2806 INIT_LIST_HEAD(&evt->node); sdev_evt_alloc()
2821 return evt; sdev_evt_alloc()
2836 struct scsi_event *evt = sdev_evt_alloc(evt_type, gfpflags); sdev_evt_send_simple() local
2837 if (!evt) { sdev_evt_send_simple()
2843 sdev_evt_send(sdev, evt); sdev_evt_send_simple()
H A Dgdth_ioctl.h206 u8 evt[EVENT_SIZE]; /* event structure */ member in struct:__anon8932::__anon8933::__anon8937
253 u8 evt[EVENT_SIZE]; /* event structure */ member in struct:__anon8940::__anon8941::__anon8947
H A Dscsi_sysfs.c413 struct scsi_event *evt; scsi_device_dev_release_usercontext() local
415 evt = list_entry(this, struct scsi_event, node); scsi_device_dev_release_usercontext()
416 list_del(&evt->node); scsi_device_dev_release_usercontext()
417 kfree(evt); scsi_device_dev_release_usercontext()
H A Dvirtio_scsi.c51 struct virtio_scsi_event evt; member in union:virtio_scsi_cmd::__anon9695
/linux-4.1.27/drivers/infiniband/hw/cxgb3/
H A Diwch.c205 static void iwch_event_handler(struct t3cdev *tdev, u32 evt, u32 port_id) iwch_event_handler() argument
216 switch (evt) { iwch_event_handler()
/linux-4.1.27/arch/sh/include/asm/
H A Duaccess.h198 static inline void *set_exception_table_evt(unsigned int evt, void *handler) set_exception_table_evt() argument
200 return set_exception_table_vec(evt >> 5, handler); set_exception_table_evt()
/linux-4.1.27/arch/hexagon/kernel/
H A Dtime.c88 static int set_next_event(unsigned long delta, struct clock_event_device *evt) set_next_event() argument
104 struct clock_event_device *evt) set_mode()
103 set_mode(enum clock_event_mode mode, struct clock_event_device *evt) set_mode() argument
/linux-4.1.27/arch/arm/mach-shmobile/
H A Dintc.h293 #define intcs_evt2irq(evt) evt2irq(INTCS_VECT_BASE + (evt))
/linux-4.1.27/arch/arm/mach-omap2/
H A Dtimer.c82 struct clock_event_device *evt = &clockevent_gpt; omap2_gp_timer_interrupt() local
86 evt->event_handler(evt); omap2_gp_timer_interrupt()
97 struct clock_event_device *evt) omap2_gp_timer_set_next_event()
106 struct clock_event_device *evt) omap2_gp_timer_set_mode()
96 omap2_gp_timer_set_next_event(unsigned long cycles, struct clock_event_device *evt) omap2_gp_timer_set_next_event() argument
105 omap2_gp_timer_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) omap2_gp_timer_set_mode() argument
/linux-4.1.27/drivers/isdn/hisax/
H A Dst5481.h501 static inline const char *ST5481_CMD_string(int evt) ST5481_CMD_string() argument
505 switch (evt) { ST5481_CMD_string()
517 sprintf(s, "0x%x", evt); ST5481_CMD_string()
/linux-4.1.27/arch/mips/sni/
H A Dtime.c18 struct clock_event_device *evt) a20r_set_mode()
17 a20r_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) a20r_set_mode() argument
/linux-4.1.27/arch/arm/include/asm/
H A Dpmu.h104 int (*set_event_filter)(struct hw_perf_event *evt,
/linux-4.1.27/arch/x86/platform/uv/
H A Duv_time.c327 struct clock_event_device *evt) uv_rtc_timer_setup()
329 int ced_cpu = cpumask_first(evt->cpumask); uv_rtc_timer_setup()
326 uv_rtc_timer_setup(enum clock_event_mode mode, struct clock_event_device *evt) uv_rtc_timer_setup() argument
/linux-4.1.27/drivers/gpu/ipu-v3/
H A Dipu-dc.c55 #define DC_RL_CH(evt) (8 + ((evt) & ~0x1) * 2)
/linux-4.1.27/arch/x86/kernel/apic/
H A Dapic.c456 struct clock_event_device *evt) lapic_next_event()
463 struct clock_event_device *evt) lapic_next_deadline()
476 struct clock_event_device *evt) lapic_timer_setup()
482 if (evt->features & CLOCK_EVT_FEAT_DUMMY) lapic_timer_setup()
872 struct clock_event_device *evt = &per_cpu(lapic_events, cpu); local_apic_timer_interrupt() local
881 * its possible that when we get here evt->event_handler is NULL. local_apic_timer_interrupt()
885 if (!evt->event_handler) { local_apic_timer_interrupt()
888 lapic_timer_setup(CLOCK_EVT_MODE_SHUTDOWN, evt); local_apic_timer_interrupt()
897 evt->event_handler(evt); local_apic_timer_interrupt()
455 lapic_next_event(unsigned long delta, struct clock_event_device *evt) lapic_next_event() argument
462 lapic_next_deadline(unsigned long delta, struct clock_event_device *evt) lapic_next_deadline() argument
475 lapic_timer_setup(enum clock_event_mode mode, struct clock_event_device *evt) lapic_timer_setup() argument
/linux-4.1.27/drivers/net/ethernet/pasemi/
H A Dpasemi_mac.c1399 int cs_size, i, fill, hdr, cpyhdr, evt; pasemi_mac_queue_csdesc() local
1447 evt = !csring->last_event; pasemi_mac_queue_csdesc()
1448 csring->last_event = evt; pasemi_mac_queue_csdesc()
1452 CTRL_CMD_ETYPE_SET | CTRL_CMD_REG(csring->events[evt]); pasemi_mac_queue_csdesc()
1455 CTRL_CMD_ETYPE_WCLR | CTRL_CMD_REG(csring->events[!evt]); pasemi_mac_queue_csdesc()
1465 CTRL_CMD_ETYPE_WSET | CTRL_CMD_REG(csring->events[evt]); pasemi_mac_queue_csdesc()
1468 CTRL_CMD_ETYPE_CLR | CTRL_CMD_REG(csring->events[!evt]); pasemi_mac_queue_csdesc()
/linux-4.1.27/drivers/scsi/bfa/
H A Dbfa_fcpim.c2797 enum bfa_ioim_event evt = BFA_IOIM_SM_COMP; bfa_ioim_isr() local
2815 evt = BFA_IOIM_SM_DONE; bfa_ioim_isr()
2817 evt = BFA_IOIM_SM_COMP; bfa_ioim_isr()
2826 evt = BFA_IOIM_SM_DONE; bfa_ioim_isr()
2828 evt = BFA_IOIM_SM_COMP; bfa_ioim_isr()
2834 evt = BFA_IOIM_SM_COMP; bfa_ioim_isr()
2840 evt = BFA_IOIM_SM_SQRETRY; bfa_ioim_isr()
2845 evt = BFA_IOIM_SM_FREE; bfa_ioim_isr()
2857 evt = BFA_IOIM_SM_ABORT_COMP; bfa_ioim_isr()
2859 evt = BFA_IOIM_SM_ABORT_DONE; bfa_ioim_isr()
2864 evt = BFA_IOIM_SM_COMP_UTAG; bfa_ioim_isr()
2871 bfa_sm_send_event(ioim, evt); bfa_ioim_isr()
/linux-4.1.27/drivers/scsi/qla2xxx/
H A Dqla_mr.c1711 qlafx00_process_aen(struct scsi_qla_host *vha, struct qla_work_evt *evt) qlafx00_process_aen() argument
1717 aen_data = evt->u.aenfx.evtcode; qlafx00_process_aen()
1719 switch (evt->u.aenfx.evtcode) { qlafx00_process_aen()
1721 if (evt->u.aenfx.mbx[1] == 0) { qlafx00_process_aen()
1722 if (evt->u.aenfx.mbx[2] == 1) { qlafx00_process_aen()
1729 } else if (evt->u.aenfx.mbx[2] == 2) { qlafx00_process_aen()
1730 qlafx00_tgt_detach(vha, evt->u.aenfx.mbx[3]); qlafx00_process_aen()
1732 } else if (evt->u.aenfx.mbx[1] == 0xffff) { qlafx00_process_aen()
1733 if (evt->u.aenfx.mbx[2] == 1) { qlafx00_process_aen()
1737 } else if (evt->u.aenfx.mbx[2] == 2) { qlafx00_process_aen()
1755 evt->u.aenfx.evtcode); qlafx00_process_aen()
/linux-4.1.27/drivers/gpu/drm/sti/
H A Dsti_hqvdp.c708 * @evt: event message
716 int sti_hqvdp_vtg_cb(struct notifier_block *nb, unsigned long evt, void *data) sti_hqvdp_vtg_cb() argument
722 if ((evt != VTG_TOP_FIELD_EVENT) && (evt != VTG_BOTTOM_FIELD_EVENT)) { sti_hqvdp_vtg_cb()
/linux-4.1.27/drivers/net/wan/
H A Ddscc4.c1779 }, *evt; dscc4_rx_irq() local
1781 for (evt = evts; evt->irq_name; evt++) { dscc4_rx_irq()
1782 if (state & evt->mask) { dscc4_rx_irq()
1784 dev->name, evt->irq_name); dscc4_rx_irq()
1785 if (!(state &= ~evt->mask)) dscc4_rx_irq()
/linux-4.1.27/drivers/dma/
H A Dimx-sdma.c475 unsigned long evt, mcu, dsp; sdma_config_ownership() local
480 evt = readl_relaxed(sdma->regs + SDMA_H_EVTOVR); sdma_config_ownership()
490 __clear_bit(channel, &evt); sdma_config_ownership()
492 __set_bit(channel, &evt); sdma_config_ownership()
499 writel_relaxed(evt, sdma->regs + SDMA_H_EVTOVR); sdma_config_ownership()
/linux-4.1.27/net/wireless/
H A Dtrace.h2440 TP_PROTO(struct net_device *netdev, enum nl80211_radar_event evt),
2441 TP_ARGS(netdev, evt),
2444 __field(enum nl80211_radar_event, evt)
2448 __entry->evt = evt;
2451 NETDEV_PR_ARG, __entry->evt)
/linux-4.1.27/drivers/sh/intc/
H A Dcore.c344 * In the case of multi-evt handling and sparse register_intc_controller()
/linux-4.1.27/drivers/net/wireless/cw1200/
H A Dwsm.c934 event->evt.id = WSM_GET32(buf); wsm_event_indication()
935 event->evt.data = WSM_GET32(buf); wsm_event_indication()
938 event->evt.id, event->evt.data); wsm_event_indication()
H A Dsta.c988 switch (event->evt.id) { cw1200_event_handler()
1021 int rcpi_rssi = (int)(event->evt.data & 0xFF); cw1200_event_handler()
/linux-4.1.27/include/uapi/linux/
H A Di2o-dev.h102 /* Event data size = frame size - message header + evt indicator */
/linux-4.1.27/net/nfc/nci/
H A Dhci.c61 u8 header; /* type -cmd,evt,rsp- + instruction */
/linux-4.1.27/drivers/infiniband/hw/ocrdma/
H A Docrdma_hw.c761 struct ocrdma_ae_pvid_mcqe *evt; ocrdma_process_grp5_aync() local
767 evt = (struct ocrdma_ae_pvid_mcqe *)cqe; ocrdma_process_grp5_aync()
768 if ((evt->tag_enabled & OCRDMA_AE_PVID_MCQE_ENABLED_MASK) >> ocrdma_process_grp5_aync()
770 dev->pvid = ((evt->tag_enabled & ocrdma_process_grp5_aync()
796 pr_err("%s(%d) invalid evt code=0x%x\n", __func__, ocrdma_process_acqe()
/linux-4.1.27/drivers/net/wireless/ath/ath6kl/
H A Dwmi.c481 struct wmi_tx_complete_event *evt; ath6kl_wmi_tx_complete_event_rx() local
485 evt = (struct wmi_tx_complete_event *) datap; ath6kl_wmi_tx_complete_event_rx()
488 evt->num_msg, evt->msg_len, evt->msg_type); ath6kl_wmi_tx_complete_event_rx()
490 for (index = 0; index < evt->num_msg; index++) { ath6kl_wmi_tx_complete_event_rx()
1105 "bss info evt - ch %u, snr %d, rssi %d, bssid \"%pM\" " ath6kl_wmi_bssinfo_event_rx()
/linux-4.1.27/include/scsi/
H A Dscsi_device.h403 extern void sdev_evt_send(struct scsi_device *sdev, struct scsi_event *evt);
/linux-4.1.27/arch/s390/kernel/
H A Dtime.c113 struct clock_event_device *evt) s390_next_event()
121 struct clock_event_device *evt) s390_set_mode()
112 s390_next_event(unsigned long delta, struct clock_event_device *evt) s390_next_event() argument
120 s390_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) s390_set_mode() argument
/linux-4.1.27/sound/soc/davinci/
H A Ddavinci-mcasp.c1532 ret = of_property_read_u32(np, "tx-num-evt", &val); davinci_mcasp_set_pdata_from_of()
1536 ret = of_property_read_u32(np, "rx-num-evt", &val); davinci_mcasp_set_pdata_from_of()
/linux-4.1.27/arch/x86/lguest/
H A Dboot.c972 struct clock_event_device *evt) lguest_clockevent_set_next_event()
989 struct clock_event_device *evt) lguest_clockevent_set_mode()
971 lguest_clockevent_set_next_event(unsigned long delta, struct clock_event_device *evt) lguest_clockevent_set_next_event() argument
988 lguest_clockevent_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) lguest_clockevent_set_mode() argument
/linux-4.1.27/include/net/bluetooth/
H A Dhci_core.h389 void (*notify)(struct hci_dev *hdev, unsigned int evt);
/linux-4.1.27/net/bluetooth/
H A Dhci_event.c5156 if (hdr->evt != event) hci_get_cmd_complete()
5161 if (hdr->evt != HCI_EV_CMD_COMPLETE) { hci_get_cmd_complete()
5162 BT_DBG("Last event is not cmd complete (0x%2.2x)", hdr->evt); hci_get_cmd_complete()
5189 u8 status = 0, event = hdr->evt, req_evt = 0; hci_event_packet()

Completed in 4317 milliseconds

12