/linux-4.1.27/drivers/clocksource/ |
H A D | mtk_timer.c | 68 static void mtk_clkevt_time_stop(struct mtk_clock_event_device *evt, u8 timer) mtk_clkevt_time_stop() argument 72 val = readl(evt->gpt_base + TIMER_CTRL_REG(timer)); mtk_clkevt_time_stop() 73 writel(val & ~TIMER_CTRL_ENABLE, evt->gpt_base + mtk_clkevt_time_stop() 77 static void mtk_clkevt_time_setup(struct mtk_clock_event_device *evt, mtk_clkevt_time_setup() argument 80 writel(delay, evt->gpt_base + TIMER_CMP_REG(timer)); mtk_clkevt_time_setup() 83 static void mtk_clkevt_time_start(struct mtk_clock_event_device *evt, mtk_clkevt_time_start() argument 89 writel(GPT_IRQ_ACK(timer), evt->gpt_base + GPT_IRQ_ACK_REG); mtk_clkevt_time_start() 91 val = readl(evt->gpt_base + TIMER_CTRL_REG(timer)); mtk_clkevt_time_start() 102 evt->gpt_base + TIMER_CTRL_REG(timer)); mtk_clkevt_time_start() 108 struct mtk_clock_event_device *evt = to_mtk_clk(clk); mtk_clkevt_mode() local 110 mtk_clkevt_time_stop(evt, GPT_CLK_EVT); mtk_clkevt_mode() 114 mtk_clkevt_time_setup(evt, evt->ticks_per_jiffy, GPT_CLK_EVT); mtk_clkevt_mode() 115 mtk_clkevt_time_start(evt, true, GPT_CLK_EVT); mtk_clkevt_mode() 131 struct mtk_clock_event_device *evt = to_mtk_clk(clk); mtk_clkevt_next_event() local 133 mtk_clkevt_time_stop(evt, GPT_CLK_EVT); mtk_clkevt_next_event() 134 mtk_clkevt_time_setup(evt, event, GPT_CLK_EVT); mtk_clkevt_next_event() 135 mtk_clkevt_time_start(evt, false, GPT_CLK_EVT); mtk_clkevt_next_event() 142 struct mtk_clock_event_device *evt = dev_id; mtk_timer_interrupt() local 145 writel(GPT_IRQ_ACK(GPT_CLK_EVT), evt->gpt_base + GPT_IRQ_ACK_REG); mtk_timer_interrupt() 146 evt->dev.event_handler(&evt->dev); mtk_timer_interrupt() 151 static void mtk_timer_global_reset(struct mtk_clock_event_device *evt) mtk_timer_global_reset() argument 154 writel(0x0, evt->gpt_base + GPT_IRQ_EN_REG); mtk_timer_global_reset() 156 writel(0x3f, evt->gpt_base + GPT_IRQ_ACK_REG); mtk_timer_global_reset() 160 mtk_timer_setup(struct mtk_clock_event_device *evt, u8 timer, u8 option) mtk_timer_setup() argument 163 evt->gpt_base + TIMER_CTRL_REG(timer)); mtk_timer_setup() 166 evt->gpt_base + TIMER_CLK_REG(timer)); mtk_timer_setup() 168 writel(0x0, evt->gpt_base + TIMER_CMP_REG(timer)); mtk_timer_setup() 171 evt->gpt_base + TIMER_CTRL_REG(timer)); mtk_timer_setup() 174 static void mtk_timer_enable_irq(struct mtk_clock_event_device *evt, u8 timer) mtk_timer_enable_irq() argument 178 val = readl(evt->gpt_base + GPT_IRQ_EN_REG); mtk_timer_enable_irq() 180 evt->gpt_base + GPT_IRQ_EN_REG); mtk_timer_enable_irq() 185 struct mtk_clock_event_device *evt; mtk_timer_init() local 190 evt = kzalloc(sizeof(*evt), GFP_KERNEL); mtk_timer_init() 191 if (!evt) { mtk_timer_init() 196 evt->dev.name = "mtk_tick"; mtk_timer_init() 197 evt->dev.rating = 300; mtk_timer_init() 198 evt->dev.features = CLOCK_EVT_FEAT_PERIODIC | CLOCK_EVT_FEAT_ONESHOT; mtk_timer_init() 199 evt->dev.set_mode = mtk_clkevt_mode; mtk_timer_init() 200 evt->dev.set_next_event = mtk_clkevt_next_event; mtk_timer_init() 201 evt->dev.cpumask = cpu_possible_mask; mtk_timer_init() 203 evt->gpt_base = of_io_request_and_map(node, 0, "mtk-timer"); mtk_timer_init() 204 if (IS_ERR(evt->gpt_base)) { mtk_timer_init() 209 evt->dev.irq = irq_of_parse_and_map(node, 0); mtk_timer_init() 210 if (evt->dev.irq <= 0) { mtk_timer_init() 227 mtk_timer_global_reset(evt); mtk_timer_init() 229 if (request_irq(evt->dev.irq, mtk_timer_interrupt, mtk_timer_init() 230 IRQF_TIMER | IRQF_IRQPOLL, "mtk_timer", evt)) { mtk_timer_init() 231 pr_warn("failed to setup irq %d\n", evt->dev.irq); mtk_timer_init() 235 evt->ticks_per_jiffy = DIV_ROUND_UP(rate, HZ); mtk_timer_init() 238 mtk_timer_setup(evt, GPT_CLK_SRC, TIMER_CTRL_OP_FREERUN); mtk_timer_init() 239 clocksource_mmio_init(evt->gpt_base + TIMER_CNT_REG(GPT_CLK_SRC), mtk_timer_init() 243 mtk_timer_setup(evt, GPT_CLK_EVT, TIMER_CTRL_OP_REPEAT); mtk_timer_init() 244 clockevents_config_and_register(&evt->dev, rate, 0x3, mtk_timer_init() 247 mtk_timer_enable_irq(evt, GPT_CLK_EVT); mtk_timer_init() 256 irq_dispose_mapping(evt->dev.irq); mtk_timer_init() 258 iounmap(evt->gpt_base); mtk_timer_init()
|
H A D | dummy_timer.c | 20 struct clock_event_device *evt) dummy_timer_set_mode() 31 struct clock_event_device *evt = raw_cpu_ptr(&dummy_timer_evt); dummy_timer_setup() local 33 evt->name = "dummy_timer"; dummy_timer_setup() 34 evt->features = CLOCK_EVT_FEAT_PERIODIC | dummy_timer_setup() 37 evt->rating = 100; dummy_timer_setup() 38 evt->set_mode = dummy_timer_set_mode; dummy_timer_setup() 39 evt->cpumask = cpumask_of(cpu); dummy_timer_setup() 41 clockevents_register_device(evt); dummy_timer_setup() 19 dummy_timer_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) dummy_timer_set_mode() argument
|
H A D | bcm2835_timer.c | 46 struct clock_event_device evt; member in struct:bcm2835_timer 76 struct bcm2835_timer, evt); bcm2835_time_set_next_event() 89 event_handler = ACCESS_ONCE(timer->evt.event_handler); bcm2835_time_interrupt() 91 event_handler(&timer->evt); bcm2835_time_interrupt() 129 timer->evt.name = node->name; bcm2835_timer_init() 130 timer->evt.rating = 300; bcm2835_timer_init() 131 timer->evt.features = CLOCK_EVT_FEAT_ONESHOT; bcm2835_timer_init() 132 timer->evt.set_mode = bcm2835_time_set_mode; bcm2835_timer_init() 133 timer->evt.set_next_event = bcm2835_time_set_next_event; bcm2835_timer_init() 134 timer->evt.cpumask = cpumask_of(0); bcm2835_timer_init() 143 clockevents_config_and_register(&timer->evt, freq, 0xf, 0xffffffff); bcm2835_timer_init()
|
H A D | exynos_mct.c | 86 struct clock_event_device evt; member in struct:mct_clock_event_device 284 struct clock_event_device *evt) exynos4_comp_set_next_event() 286 exynos4_mct_comp0_start(evt->mode, cycles); exynos4_comp_set_next_event() 292 struct clock_event_device *evt) exynos4_comp_set_mode() 300 (((unsigned long long) NSEC_PER_SEC / HZ * evt->mult) >> evt->shift); exynos4_comp_set_mode() 322 struct clock_event_device *evt = dev_id; exynos4_mct_comp_isr() local 326 evt->event_handler(evt); exynos4_mct_comp_isr() 384 struct clock_event_device *evt) exynos4_tick_set_next_event() 394 struct clock_event_device *evt) exynos4_tick_set_mode() 404 (((unsigned long long) NSEC_PER_SEC / HZ * evt->mult) >> evt->shift); exynos4_tick_set_mode() 418 struct clock_event_device *evt = &mevt->evt; exynos4_mct_tick_clear() local 425 if (evt->mode != CLOCK_EVT_MODE_PERIODIC) exynos4_mct_tick_clear() 440 struct clock_event_device *evt = &mevt->evt; exynos4_mct_tick_isr() local 444 evt->event_handler(evt); exynos4_mct_tick_isr() 449 static int exynos4_local_timer_setup(struct clock_event_device *evt) exynos4_local_timer_setup() argument 454 mevt = container_of(evt, struct mct_clock_event_device, evt); exynos4_local_timer_setup() 459 evt->name = mevt->name; exynos4_local_timer_setup() 460 evt->cpumask = cpumask_of(cpu); exynos4_local_timer_setup() 461 evt->set_next_event = exynos4_tick_set_next_event; exynos4_local_timer_setup() 462 evt->set_mode = exynos4_tick_set_mode; exynos4_local_timer_setup() 463 evt->features = CLOCK_EVT_FEAT_PERIODIC | CLOCK_EVT_FEAT_ONESHOT; exynos4_local_timer_setup() 464 evt->rating = 450; exynos4_local_timer_setup() 470 if (evt->irq == -1) exynos4_local_timer_setup() 473 irq_force_affinity(evt->irq, cpumask_of(cpu)); exynos4_local_timer_setup() 474 enable_irq(evt->irq); exynos4_local_timer_setup() 478 clockevents_config_and_register(evt, clk_rate / (TICK_BASE_CNT + 1), exynos4_local_timer_setup() 484 static void exynos4_local_timer_stop(struct clock_event_device *evt) exynos4_local_timer_stop() argument 486 evt->set_mode(CLOCK_EVT_MODE_UNUSED, evt); exynos4_local_timer_stop() 488 if (evt->irq != -1) exynos4_local_timer_stop() 489 disable_irq_nosync(evt->irq); exynos4_local_timer_stop() 507 exynos4_local_timer_setup(&mevt->evt); exynos4_mct_cpu_notify() 511 exynos4_local_timer_stop(&mevt->evt); exynos4_mct_cpu_notify() 556 pcpu_mevt->evt.irq = -1; for_each_possible_cpu() 568 pcpu_mevt->evt.irq = mct_irq; for_each_possible_cpu() 577 exynos4_local_timer_setup(&mevt->evt); 283 exynos4_comp_set_next_event(unsigned long cycles, struct clock_event_device *evt) exynos4_comp_set_next_event() argument 291 exynos4_comp_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) exynos4_comp_set_mode() argument 383 exynos4_tick_set_next_event(unsigned long cycles, struct clock_event_device *evt) exynos4_tick_set_next_event() argument 393 exynos4_tick_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) exynos4_tick_set_mode() argument
|
H A D | qcom-timer.c | 50 struct clock_event_device *evt = dev_id; msm_timer_interrupt() local 52 if (evt->mode == CLOCK_EVT_MODE_ONESHOT) { msm_timer_interrupt() 57 evt->event_handler(evt); msm_timer_interrupt() 62 struct clock_event_device *evt) msm_timer_set_next_event() 81 struct clock_event_device *evt) msm_timer_set_mode() 122 static int msm_local_timer_setup(struct clock_event_device *evt) msm_local_timer_setup() argument 127 evt->irq = msm_timer_irq; msm_local_timer_setup() 128 evt->name = "msm_timer"; msm_local_timer_setup() 129 evt->features = CLOCK_EVT_FEAT_ONESHOT; msm_local_timer_setup() 130 evt->rating = 200; msm_local_timer_setup() 131 evt->set_mode = msm_timer_set_mode; msm_local_timer_setup() 132 evt->set_next_event = msm_timer_set_next_event; msm_local_timer_setup() 133 evt->cpumask = cpumask_of(cpu); msm_local_timer_setup() 135 clockevents_config_and_register(evt, GPT_HZ, 4, 0xffffffff); msm_local_timer_setup() 138 enable_percpu_irq(evt->irq, IRQ_TYPE_EDGE_RISING); msm_local_timer_setup() 140 err = request_irq(evt->irq, msm_timer_interrupt, msm_local_timer_setup() 142 IRQF_TRIGGER_RISING, "gp_timer", evt); msm_local_timer_setup() 150 static void msm_local_timer_stop(struct clock_event_device *evt) msm_local_timer_stop() argument 152 evt->set_mode(CLOCK_EVT_MODE_UNUSED, evt); msm_local_timer_stop() 153 disable_percpu_irq(evt->irq); msm_local_timer_stop() 61 msm_timer_set_next_event(unsigned long cycles, struct clock_event_device *evt) msm_timer_set_next_event() argument 80 msm_timer_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) msm_timer_set_mode() argument
|
H A D | time-armada-370-xp.c | 160 struct clock_event_device *evt = dev_id; armada_370_xp_timer_interrupt() local 163 evt->event_handler(evt); armada_370_xp_timer_interrupt() 171 static int armada_370_xp_timer_setup(struct clock_event_device *evt) armada_370_xp_timer_setup() argument 182 evt->name = "armada_370_xp_per_cpu_tick", armada_370_xp_timer_setup() 183 evt->features = CLOCK_EVT_FEAT_ONESHOT | armada_370_xp_timer_setup() 185 evt->shift = 32, armada_370_xp_timer_setup() 186 evt->rating = 300, armada_370_xp_timer_setup() 187 evt->set_next_event = armada_370_xp_clkevt_next_event, armada_370_xp_timer_setup() 188 evt->set_mode = armada_370_xp_clkevt_mode, armada_370_xp_timer_setup() 189 evt->irq = armada_370_xp_clkevt_irq; armada_370_xp_timer_setup() 190 evt->cpumask = cpumask_of(cpu); armada_370_xp_timer_setup() 192 clockevents_config_and_register(evt, timer_clk, 1, 0xfffffffe); armada_370_xp_timer_setup() 193 enable_percpu_irq(evt->irq, 0); armada_370_xp_timer_setup() 198 static void armada_370_xp_timer_stop(struct clock_event_device *evt) armada_370_xp_timer_stop() argument 200 evt->set_mode(CLOCK_EVT_MODE_UNUSED, evt); armada_370_xp_timer_stop() 201 disable_percpu_irq(evt->irq); armada_370_xp_timer_stop()
|
H A D | clps711x-timer.c | 57 struct clock_event_device *evt = dev_id; clps711x_timer_interrupt() local 59 evt->event_handler(evt); clps711x_timer_interrupt() 65 struct clock_event_device *evt) clps711x_clockevent_set_mode() 64 clps711x_clockevent_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) clps711x_clockevent_set_mode() argument
|
H A D | meson6_timer.c | 91 static int meson6_clkevt_next_event(unsigned long evt, meson6_clkevt_next_event() argument 95 meson6_clkevt_time_setup(CED_ID, evt); meson6_clkevt_next_event() 111 struct clock_event_device *evt = (struct clock_event_device *)dev_id; meson6_timer_interrupt() local 113 evt->event_handler(evt); meson6_timer_interrupt()
|
H A D | arm_arch_timer.c | 51 struct clock_event_device evt; member in struct:arch_timer 54 #define to_arch_timer(e) container_of(e, struct arch_timer, evt) 141 struct clock_event_device *evt) timer_handler() 145 ctrl = arch_timer_reg_read(access, ARCH_TIMER_REG_CTRL, evt); timer_handler() 148 arch_timer_reg_write(access, ARCH_TIMER_REG_CTRL, ctrl, evt); timer_handler() 149 evt->event_handler(evt); timer_handler() 158 struct clock_event_device *evt = dev_id; arch_timer_handler_virt() local 160 return timer_handler(ARCH_TIMER_VIRT_ACCESS, evt); arch_timer_handler_virt() 165 struct clock_event_device *evt = dev_id; arch_timer_handler_phys() local 167 return timer_handler(ARCH_TIMER_PHYS_ACCESS, evt); arch_timer_handler_phys() 172 struct clock_event_device *evt = dev_id; arch_timer_handler_phys_mem() local 174 return timer_handler(ARCH_TIMER_MEM_PHYS_ACCESS, evt); arch_timer_handler_phys_mem() 179 struct clock_event_device *evt = dev_id; arch_timer_handler_virt_mem() local 181 return timer_handler(ARCH_TIMER_MEM_VIRT_ACCESS, evt); arch_timer_handler_virt_mem() 224 static __always_inline void set_next_event(const int access, unsigned long evt, set_next_event() argument 231 arch_timer_reg_write(access, ARCH_TIMER_REG_TVAL, evt, clk); set_next_event() 235 static int arch_timer_set_next_event_virt(unsigned long evt, arch_timer_set_next_event_virt() argument 238 set_next_event(ARCH_TIMER_VIRT_ACCESS, evt, clk); arch_timer_set_next_event_virt() 242 static int arch_timer_set_next_event_phys(unsigned long evt, arch_timer_set_next_event_phys() argument 245 set_next_event(ARCH_TIMER_PHYS_ACCESS, evt, clk); arch_timer_set_next_event_phys() 249 static int arch_timer_set_next_event_virt_mem(unsigned long evt, arch_timer_set_next_event_virt_mem() argument 252 set_next_event(ARCH_TIMER_MEM_VIRT_ACCESS, evt, clk); arch_timer_set_next_event_virt_mem() 256 static int arch_timer_set_next_event_phys_mem(unsigned long evt, arch_timer_set_next_event_phys_mem() argument 259 set_next_event(ARCH_TIMER_MEM_PHYS_ACCESS, evt, clk); arch_timer_set_next_event_phys_mem() 640 t->evt.irq = irq; arch_timer_mem_register() 641 __arch_timer_setup(ARCH_MEM_TIMER, &t->evt); arch_timer_mem_register() 648 ret = request_irq(irq, func, IRQF_TIMER, "arch_mem_timer", &t->evt); arch_timer_mem_register() 140 timer_handler(const int access, struct clock_event_device *evt) timer_handler() argument
|
H A D | timer-keystone.c | 132 struct clock_event_device *evt = dev_id; keystone_timer_interrupt() local 134 evt->event_handler(evt); keystone_timer_interrupt() 139 struct clock_event_device *evt) keystone_set_next_event() 141 return keystone_timer_config(cycles, evt->mode); keystone_set_next_event() 145 struct clock_event_device *evt) keystone_set_mode() 138 keystone_set_next_event(unsigned long cycles, struct clock_event_device *evt) keystone_set_next_event() argument 144 keystone_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) keystone_set_mode() argument
|
H A D | vf_pit_timer.c | 90 struct clock_event_device *evt) pit_set_mode() 94 pit_set_next_event(cycle_per_jiffy, evt); pit_set_mode() 107 struct clock_event_device *evt = dev_id; pit_timer_interrupt() local 117 if (likely(evt->mode == CLOCK_EVT_MODE_ONESHOT)) pit_timer_interrupt() 120 evt->event_handler(evt); pit_timer_interrupt() 89 pit_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) pit_set_mode() argument
|
H A D | dw_apb_timer.c | 41 ced_to_dw_apb_ced(struct clock_event_device *evt) ced_to_dw_apb_ced() argument 43 return container_of(evt, struct dw_apb_clock_event_device, ced); ced_to_dw_apb_ced() 89 struct clock_event_device *evt = data; dw_apb_clockevent_irq() local 90 struct dw_apb_clock_event_device *dw_ced = ced_to_dw_apb_ced(evt); dw_apb_clockevent_irq() 92 if (!evt->event_handler) { dw_apb_clockevent_irq() 100 evt->event_handler(evt); dw_apb_clockevent_irq() 114 struct clock_event_device *evt) apbt_set_mode() 118 struct dw_apb_clock_event_device *dw_ced = ced_to_dw_apb_ced(evt); apbt_set_mode() 121 cpumask_first(evt->cpumask), apbt_set_mode() 181 struct clock_event_device *evt) apbt_next_event() 184 struct dw_apb_clock_event_device *dw_ced = ced_to_dw_apb_ced(evt); apbt_next_event() 113 apbt_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) apbt_set_mode() argument 180 apbt_next_event(unsigned long delta, struct clock_event_device *evt) apbt_next_event() argument
|
H A D | fsl_ftm_timer.c | 157 struct clock_event_device *evt) ftm_set_mode() 161 ftm_set_next_event(priv->periodic_cyc, evt); ftm_set_mode() 173 struct clock_event_device *evt = dev_id; ftm_evt_interrupt() local 177 if (likely(evt->mode == CLOCK_EVT_MODE_ONESHOT)) { ftm_evt_interrupt() 182 evt->event_handler(evt); ftm_evt_interrupt() 287 freq = __ftm_clk_init(np, "ftm-evt-counter-en", "ftm-evt"); ftm_clk_init() 156 ftm_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) ftm_set_mode() argument
|
H A D | mxs_timer.c | 109 static int timrotv1_set_next_event(unsigned long evt, timrotv1_set_next_event() argument 113 __raw_writel(evt, mxs_timrot_base + HW_TIMROT_TIMCOUNTn(0)); timrotv1_set_next_event() 118 static int timrotv2_set_next_event(unsigned long evt, timrotv2_set_next_event() argument 122 __raw_writel(evt, mxs_timrot_base + HW_TIMROT_FIXED_COUNTn(0)); timrotv2_set_next_event() 129 struct clock_event_device *evt = dev_id; mxs_timer_interrupt() local 132 evt->event_handler(evt); mxs_timer_interrupt() 154 struct clock_event_device *evt) mxs_set_mode() 153 mxs_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) mxs_set_mode() argument
|
H A D | sun4i_timer.c | 105 static int sun4i_clkevt_next_event(unsigned long evt, sun4i_clkevt_next_event() argument 109 sun4i_clkevt_time_setup(0, evt - TIMER_SYNC_TICKS); sun4i_clkevt_next_event() 126 struct clock_event_device *evt = (struct clock_event_device *)dev_id; sun4i_timer_interrupt() local 129 evt->event_handler(evt); sun4i_timer_interrupt()
|
H A D | timer-digicolor.c | 113 static int digicolor_clkevt_next_event(unsigned long evt, digicolor_clkevt_next_event() argument 117 dc_timer_set_count(ce, evt); digicolor_clkevt_next_event() 136 struct clock_event_device *evt = dev_id; digicolor_timer_interrupt() local 138 evt->event_handler(evt); digicolor_timer_interrupt()
|
H A D | timer-integrator-ap.c | 67 struct clock_event_device *evt = dev_id; integrator_timer_interrupt() local 72 evt->event_handler(evt); integrator_timer_interrupt() 77 static void clkevt_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) clkevt_set_mode() argument 106 static int clkevt_set_next_event(unsigned long next, struct clock_event_device *evt) clkevt_set_next_event() argument
|
H A D | vt8500_timer.c | 76 struct clock_event_device *evt) vt8500_timer_set_next_event() 94 struct clock_event_device *evt) vt8500_timer_set_mode() 120 struct clock_event_device *evt = dev_id; vt8500_timer_interrupt() local 122 evt->event_handler(evt); vt8500_timer_interrupt() 75 vt8500_timer_set_next_event(unsigned long cycles, struct clock_event_device *evt) vt8500_timer_set_next_event() argument 93 vt8500_timer_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) vt8500_timer_set_mode() argument
|
H A D | mips-gic-timer.c | 24 static int gic_next_event(unsigned long delta, struct clock_event_device *evt) gic_next_event() argument 31 gic_write_cpu_compare(cnt, cpumask_first(evt->cpumask)); gic_next_event() 37 struct clock_event_device *evt) gic_set_clock_mode() 36 gic_set_clock_mode(enum clock_event_mode mode, struct clock_event_device *evt) gic_set_clock_mode() argument
|
H A D | moxart_timer.c | 107 struct clock_event_device *evt = dev_id; moxart_timer_interrupt() local 108 evt->event_handler(evt); moxart_timer_interrupt()
|
H A D | i8253.c | 109 struct clock_event_device *evt) init_pit_timer() 123 if (evt->mode == CLOCK_EVT_MODE_PERIODIC || init_pit_timer() 124 evt->mode == CLOCK_EVT_MODE_ONESHOT) { init_pit_timer() 148 static int pit_next_event(unsigned long delta, struct clock_event_device *evt) pit_next_event() argument 108 init_pit_timer(enum clock_event_mode mode, struct clock_event_device *evt) init_pit_timer() argument
|
H A D | metag_generic.c | 60 struct clock_event_device *evt) metag_timer_set_mode() 93 struct clock_event_device *evt = this_cpu_ptr(&local_clockevent); metag_timer_interrupt() local 95 evt->event_handler(evt); metag_timer_interrupt() 59 metag_timer_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) metag_timer_set_mode() argument
|
H A D | arm_global_timer.c | 132 static int gt_clockevent_set_next_event(unsigned long evt, gt_clockevent_set_next_event() argument 135 gt_compare_set(evt, 0); gt_clockevent_set_next_event() 141 struct clock_event_device *evt = dev_id; gt_clockevent_interrupt() local 158 if (evt->mode == CLOCK_EVT_MODE_ONESHOT) gt_clockevent_interrupt() 162 evt->event_handler(evt); gt_clockevent_interrupt()
|
H A D | asm9260_timer.c | 114 struct clock_event_device *evt) asm9260_timer_set_next_event() 124 struct clock_event_device *evt) asm9260_timer_set_mode() 159 struct clock_event_device *evt = dev_id; asm9260_timer_interrupt() local 161 evt->event_handler(evt); asm9260_timer_interrupt() 113 asm9260_timer_set_next_event(unsigned long delta, struct clock_event_device *evt) asm9260_timer_set_next_event() argument 123 asm9260_timer_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) asm9260_timer_set_mode() argument
|
H A D | tegra20_timer.c | 65 struct clock_event_device *evt) tegra_timer_set_next_event() 76 struct clock_event_device *evt) tegra_timer_set_mode() 151 struct clock_event_device *evt = (struct clock_event_device *)dev_id; tegra_timer_interrupt() local 153 evt->event_handler(evt); tegra_timer_interrupt() 64 tegra_timer_set_next_event(unsigned long cycles, struct clock_event_device *evt) tegra_timer_set_next_event() argument 75 tegra_timer_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) tegra_timer_set_mode() argument
|
H A D | bcm_kona_timer.c | 160 struct clock_event_device *evt = &kona_clockevent_timer; kona_timer_interrupt() local 163 evt->event_handler(evt); kona_timer_interrupt()
|
H A D | samsung_pwm_timer.c | 189 struct clock_event_device *evt) samsung_set_next_event() 211 struct clock_event_device *evt) samsung_set_mode() 253 struct clock_event_device *evt = dev_id; samsung_clock_event_isr() local 260 evt->event_handler(evt); samsung_clock_event_isr() 188 samsung_set_next_event(unsigned long cycles, struct clock_event_device *evt) samsung_set_next_event() argument 210 samsung_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) samsung_set_mode() argument
|
H A D | cadence_ttc_timer.c | 179 * @evt: Address of clock event instance 184 struct clock_event_device *evt) ttc_set_next_event() 186 struct ttc_timer_clockevent *ttce = to_ttc_timer_clkevent(evt); ttc_set_next_event() 197 * @evt: Address of clock event instance 200 struct clock_event_device *evt) ttc_set_mode() 202 struct ttc_timer_clockevent *ttce = to_ttc_timer_clkevent(evt); ttc_set_mode() 183 ttc_set_next_event(unsigned long cycles, struct clock_event_device *evt) ttc_set_next_event() argument 199 ttc_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) ttc_set_mode() argument
|
H A D | timer-u300.c | 196 struct clock_event_device *evt) u300_set_mode() 199 container_of(evt, struct u300_clockevent_data, cevd); u300_set_mode() 280 struct clock_event_device *evt) u300_set_next_event() 325 struct clock_event_device *evt = &u300_clockevent_data.cevd; u300_timer_interrupt() local 330 evt->event_handler(evt); u300_timer_interrupt() 195 u300_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) u300_set_mode() argument 279 u300_set_next_event(unsigned long cycles, struct clock_event_device *evt) u300_set_next_event() argument
|
H A D | cs5535-clockevt.c | 81 struct clock_event_device *evt) mfgpt_set_mode() 91 static int mfgpt_next_event(unsigned long delta, struct clock_event_device *evt) mfgpt_next_event() argument 80 mfgpt_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) mfgpt_set_mode() argument
|
H A D | time-efm32.c | 87 static int efm32_clock_event_set_next_event(unsigned long evt, efm32_clock_event_set_next_event() argument 94 writel_relaxed(evt, ddata->base + TIMERn_CNT); efm32_clock_event_set_next_event()
|
H A D | nomadik-mtu.c | 93 static int nmdk_clkevt_next(unsigned long evt, struct clock_event_device *ev) nmdk_clkevt_next() argument 96 writel(evt, mtu_base + MTU_LR(1)); nmdk_clkevt_next()
|
H A D | timer-sun5i.c | 129 static int sun5i_clkevt_next_event(unsigned long evt, sun5i_clkevt_next_event() argument 135 sun5i_clkevt_time_setup(ce, 0, evt - TIMER_SYNC_TICKS); sun5i_clkevt_next_event()
|
/linux-4.1.27/arch/powerpc/kernel/ |
H A D | mce.c | 182 struct machine_check_event evt; machine_check_queue_event() local 184 if (!get_mce_event(&evt, MCE_EVENT_RELEASE)) machine_check_queue_event() 193 memcpy(this_cpu_ptr(&mce_event_queue[index]), &evt, sizeof(evt)); machine_check_queue_event() local 219 void machine_check_print_event_info(struct machine_check_event *evt) machine_check_print_event_info() argument 246 if (evt->version != MCE_V1) { machine_check_print_event_info() 248 evt->version); machine_check_print_event_info() 251 switch (evt->severity) { machine_check_print_event_info() 272 evt->disposition == MCE_DISPOSITION_RECOVERED ? machine_check_print_event_info() 275 evt->initiator == MCE_INITIATOR_CPU ? "CPU" : "Unknown"); machine_check_print_event_info() 276 switch (evt->error_type) { machine_check_print_event_info() 278 subtype = evt->u.ue_error.ue_error_type < machine_check_print_event_info() 280 mc_ue_types[evt->u.ue_error.ue_error_type] machine_check_print_event_info() 283 if (evt->u.ue_error.effective_address_provided) machine_check_print_event_info() 285 level, evt->u.ue_error.effective_address); machine_check_print_event_info() 286 if (evt->u.ue_error.physical_address_provided) machine_check_print_event_info() 288 level, evt->u.ue_error.physical_address); machine_check_print_event_info() 291 subtype = evt->u.slb_error.slb_error_type < machine_check_print_event_info() 293 mc_slb_types[evt->u.slb_error.slb_error_type] machine_check_print_event_info() 296 if (evt->u.slb_error.effective_address_provided) machine_check_print_event_info() 298 level, evt->u.slb_error.effective_address); machine_check_print_event_info() 301 subtype = evt->u.erat_error.erat_error_type < machine_check_print_event_info() 303 mc_erat_types[evt->u.erat_error.erat_error_type] machine_check_print_event_info() 306 if (evt->u.erat_error.effective_address_provided) machine_check_print_event_info() 308 level, evt->u.erat_error.effective_address); machine_check_print_event_info() 311 subtype = evt->u.tlb_error.tlb_error_type < machine_check_print_event_info() 313 mc_tlb_types[evt->u.tlb_error.tlb_error_type] machine_check_print_event_info() 316 if (evt->u.tlb_error.effective_address_provided) machine_check_print_event_info() 318 level, evt->u.tlb_error.effective_address); machine_check_print_event_info() 327 uint64_t get_mce_fault_addr(struct machine_check_event *evt) get_mce_fault_addr() argument 329 switch (evt->error_type) { get_mce_fault_addr() 331 if (evt->u.ue_error.effective_address_provided) get_mce_fault_addr() 332 return evt->u.ue_error.effective_address; get_mce_fault_addr() 335 if (evt->u.slb_error.effective_address_provided) get_mce_fault_addr() 336 return evt->u.slb_error.effective_address; get_mce_fault_addr() 339 if (evt->u.erat_error.effective_address_provided) get_mce_fault_addr() 340 return evt->u.erat_error.effective_address; get_mce_fault_addr() 343 if (evt->u.tlb_error.effective_address_provided) get_mce_fault_addr() 344 return evt->u.tlb_error.effective_address; get_mce_fault_addr()
|
H A D | time.c | 100 static int decrementer_set_next_event(unsigned long evt, 487 struct clock_event_device *evt = this_cpu_ptr(&decrementers); __timer_interrupt() local 500 if (evt->event_handler) __timer_interrupt() 501 evt->event_handler(evt); __timer_interrupt() 852 static int decrementer_set_next_event(unsigned long evt, decrementer_set_next_event() argument 855 __this_cpu_write(decrementers_next_tb, get_tb_or_rtc() + evt); decrementer_set_next_event() 856 set_dec(evt); decrementer_set_next_event()
|
/linux-4.1.27/arch/blackfin/kernel/ |
H A D | time-ts.c | 129 struct clock_event_device *evt) bfin_gptmr0_set_next_event() 140 struct clock_event_device *evt) bfin_gptmr0_set_mode() 194 struct clock_event_device *evt = dev_id; bfin_gptmr0_interrupt() local 203 evt->event_handler(evt); bfin_gptmr0_interrupt() 223 static void __init bfin_gptmr0_clockevent_init(struct clock_event_device *evt) bfin_gptmr0_clockevent_init() argument 228 evt->mult = div_sc(clock_tick, NSEC_PER_SEC, evt->shift); bfin_gptmr0_clockevent_init() 229 evt->max_delta_ns = clockevent_delta2ns(-1, evt); bfin_gptmr0_clockevent_init() 230 evt->min_delta_ns = clockevent_delta2ns(100, evt); bfin_gptmr0_clockevent_init() 232 evt->cpumask = cpumask_of(0); bfin_gptmr0_clockevent_init() 234 clockevents_register_device(evt); bfin_gptmr0_clockevent_init() 243 struct clock_event_device *evt) bfin_coretmr_set_next_event() 254 struct clock_event_device *evt) bfin_coretmr_set_mode() 306 struct clock_event_device *evt = &per_cpu(coretmr_events, cpu); bfin_coretmr_interrupt() local 309 evt->event_handler(evt); bfin_coretmr_interrupt() 326 struct clock_event_device *evt = &per_cpu(coretmr_events, cpu); bfin_coretmr_clockevent_init() local 329 evt->broadcast = smp_timer_broadcast; bfin_coretmr_clockevent_init() 332 evt->name = "bfin_core_timer"; bfin_coretmr_clockevent_init() 333 evt->rating = 350; bfin_coretmr_clockevent_init() 334 evt->irq = -1; bfin_coretmr_clockevent_init() 335 evt->shift = 32; bfin_coretmr_clockevent_init() 336 evt->features = CLOCK_EVT_FEAT_PERIODIC | CLOCK_EVT_FEAT_ONESHOT; bfin_coretmr_clockevent_init() 337 evt->set_next_event = bfin_coretmr_set_next_event; bfin_coretmr_clockevent_init() 338 evt->set_mode = bfin_coretmr_set_mode; bfin_coretmr_clockevent_init() 341 evt->mult = div_sc(clock_tick, NSEC_PER_SEC, evt->shift); bfin_coretmr_clockevent_init() 342 evt->max_delta_ns = clockevent_delta2ns(-1, evt); bfin_coretmr_clockevent_init() 343 evt->min_delta_ns = clockevent_delta2ns(100, evt); bfin_coretmr_clockevent_init() 345 evt->cpumask = cpumask_of(cpu); bfin_coretmr_clockevent_init() 347 clockevents_register_device(evt); bfin_coretmr_clockevent_init() 128 bfin_gptmr0_set_next_event(unsigned long cycles, struct clock_event_device *evt) bfin_gptmr0_set_next_event() argument 139 bfin_gptmr0_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) bfin_gptmr0_set_mode() argument 242 bfin_coretmr_set_next_event(unsigned long cycles, struct clock_event_device *evt) bfin_coretmr_set_next_event() argument 253 bfin_coretmr_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) bfin_coretmr_set_mode() argument
|
H A D | early_printk.c | 165 u32 evt; init_early_exception_vectors() local 188 for (evt = EVT2; evt <= EVT15; evt += 4) init_early_exception_vectors() 189 bfin_write32(evt, early_trap); init_early_exception_vectors()
|
/linux-4.1.27/drivers/uwb/ |
H A D | uwbd.c | 85 * @evt->rc is already referenced and guaranteed to exist. See 164 * @evt: the event to handle 174 * controller in @evt->rc referenced. The reference will be dropped 179 int uwbd_event_handle_urc(struct uwb_event *evt) uwbd_event_handle_urc() argument 187 type = evt->notif.rceb->bEventType; uwbd_event_handle_urc() 188 event = le16_to_cpu(evt->notif.rceb->wEvent); uwbd_event_handle_urc() 189 context = evt->notif.rceb->bEventContext; uwbd_event_handle_urc() 202 result = (*handler)(evt); uwbd_event_handle_urc() 205 dev_err(&evt->rc->uwb_dev.dev, uwbd_event_handle_urc() 211 static void uwbd_event_handle_message(struct uwb_event *evt) uwbd_event_handle_message() argument 216 rc = evt->rc; uwbd_event_handle_message() 218 if (evt->message < 0 || evt->message >= ARRAY_SIZE(uwbd_message_handlers)) { uwbd_event_handle_message() 219 dev_err(&rc->uwb_dev.dev, "UWBD: invalid message type %d\n", evt->message); uwbd_event_handle_message() 223 result = uwbd_message_handlers[evt->message].handler(evt); uwbd_event_handle_message() 226 uwbd_message_handlers[evt->message].name, result); uwbd_event_handle_message() 229 static void uwbd_event_handle(struct uwb_event *evt) uwbd_event_handle() argument 234 rc = evt->rc; uwbd_event_handle() 237 switch (evt->type) { uwbd_event_handle() 239 should_keep = uwbd_event_handle_urc(evt); uwbd_event_handle() 241 kfree(evt->notif.rceb); uwbd_event_handle() 244 uwbd_event_handle_message(evt); uwbd_event_handle() 247 dev_err(&rc->uwb_dev.dev, "UWBD: invalid event type %d\n", evt->type); uwbd_event_handle() 271 struct uwb_event *evt; uwbd() local 286 evt = list_first_entry(&rc->uwbd.event_list, struct uwb_event, list_node); uwbd() 287 list_del(&evt->list_node); uwbd() 289 evt = NULL; uwbd() 292 if (evt) { uwbd() 293 uwbd_event_handle(evt); uwbd() 294 kfree(evt); uwbd() 333 void uwbd_event_queue(struct uwb_event *evt) uwbd_event_queue() argument 335 struct uwb_rc *rc = evt->rc; uwbd_event_queue() 340 list_add(&evt->list_node, &rc->uwbd.event_list); uwbd_event_queue() 343 __uwb_rc_put(evt->rc); uwbd_event_queue() 344 if (evt->type == UWB_EVT_TYPE_NOTIF) uwbd_event_queue() 345 kfree(evt->notif.rceb); uwbd_event_queue() 346 kfree(evt); uwbd_event_queue() 354 struct uwb_event *evt, *nxt; uwbd_flush() local 357 list_for_each_entry_safe(evt, nxt, &rc->uwbd.event_list, list_node) { uwbd_flush() 358 if (evt->rc == rc) { uwbd_flush() 360 list_del(&evt->list_node); uwbd_flush() 361 if (evt->type == UWB_EVT_TYPE_NOTIF) uwbd_flush() 362 kfree(evt->notif.rceb); uwbd_flush() 363 kfree(evt); uwbd_flush()
|
H A D | ie-rcv.c | 29 int uwbd_evt_handle_rc_ie_rcv(struct uwb_event *evt) uwbd_evt_handle_rc_ie_rcv() argument 32 struct device *dev = &evt->rc->uwb_dev.dev; uwbd_evt_handle_rc_ie_rcv() 36 if (evt->notif.size < sizeof(*iercv)) { uwbd_evt_handle_rc_ie_rcv() 39 evt->notif.size, sizeof(*iercv)); uwbd_evt_handle_rc_ie_rcv() 42 iercv = container_of(evt->notif.rceb, struct uwb_rc_evt_ie_rcv, rceb); uwbd_evt_handle_rc_ie_rcv()
|
H A D | beacon.c | 361 static int uwb_verify_beacon(struct uwb_rc *rc, struct uwb_event *evt, uwb_verify_beacon() argument 369 if (evt->notif.size < sizeof(*be) + sizeof(*bf)) { uwb_verify_beacon() 371 "(%zu vs %zu bytes needed)\n", evt->notif.size, uwb_verify_beacon() 393 int uwbd_evt_handle_rc_beacon(struct uwb_event *evt) uwbd_evt_handle_rc_beacon() argument 401 rc = evt->rc; uwbd_evt_handle_rc_beacon() 402 be = container_of(evt->notif.rceb, struct uwb_rc_evt_beacon, rceb); uwbd_evt_handle_rc_beacon() 403 result = uwb_verify_beacon(rc, evt, be); uwbd_evt_handle_rc_beacon() 430 uwb_beacon_print(evt->rc, be, bf); uwbd_evt_handle_rc_beacon() 431 bce = __uwb_beca_add(rc, be, bf, evt->ts_jiffies); uwbd_evt_handle_rc_beacon() 444 bce->ts_jiffies = evt->ts_jiffies; uwbd_evt_handle_rc_beacon() 457 uwbd_dev_onair(evt->rc, bce); uwbd_evt_handle_rc_beacon() 469 int uwbd_evt_handle_rc_beacon_size(struct uwb_event *evt) uwbd_evt_handle_rc_beacon_size() argument 472 struct device *dev = &evt->rc->uwb_dev.dev; uwbd_evt_handle_rc_beacon_size() 476 if (evt->notif.size < sizeof(*bs)) { uwbd_evt_handle_rc_beacon_size() 479 evt->notif.size, sizeof(*bs)); uwbd_evt_handle_rc_beacon_size() 482 bs = container_of(evt->notif.rceb, struct uwb_rc_evt_beacon_size, rceb); uwbd_evt_handle_rc_beacon_size() 500 * @evt: the BP_SLOT_CHANGE notification from the radio controller 506 int uwbd_evt_handle_rc_bp_slot_change(struct uwb_event *evt) uwbd_evt_handle_rc_bp_slot_change() argument 508 struct uwb_rc *rc = evt->rc; uwbd_evt_handle_rc_bp_slot_change() 512 if (evt->notif.size < sizeof(*bpsc)) { uwbd_evt_handle_rc_bp_slot_change() 516 bpsc = container_of(evt->notif.rceb, struct uwb_rc_evt_bp_slot_change, rceb); uwbd_evt_handle_rc_bp_slot_change() 540 int uwbd_evt_handle_rc_bpoie_change(struct uwb_event *evt) uwbd_evt_handle_rc_bpoie_change() argument 543 struct device *dev = &evt->rc->uwb_dev.dev; uwbd_evt_handle_rc_bpoie_change() 550 if (evt->notif.size < sizeof(*bpoiec)) { uwbd_evt_handle_rc_bpoie_change() 553 evt->notif.size, sizeof(*bpoiec)); uwbd_evt_handle_rc_bpoie_change() 556 bpoiec = container_of(evt->notif.rceb, struct uwb_rc_evt_bpoie_change, rceb); uwbd_evt_handle_rc_bpoie_change()
|
H A D | drp-avail.c | 227 int uwbd_evt_get_drp_avail(struct uwb_event *evt, unsigned long *bmp) uwbd_evt_get_drp_avail() argument 229 struct device *dev = &evt->rc->uwb_dev.dev; uwbd_evt_get_drp_avail() 234 if (evt->notif.size < sizeof(*drp_evt)) { uwbd_evt_get_drp_avail() 237 "needed]\n", evt->notif.size, sizeof(*drp_evt)); uwbd_evt_get_drp_avail() 240 drp_evt = container_of(evt->notif.rceb, struct uwb_rc_evt_drp_avail, rceb); uwbd_evt_get_drp_avail() 251 * @evt: Event information (packs the actual event data, which 271 int uwbd_evt_handle_rc_drp_avail(struct uwb_event *evt) uwbd_evt_handle_rc_drp_avail() argument 274 struct uwb_rc *rc = evt->rc; uwbd_evt_handle_rc_drp_avail() 277 result = uwbd_evt_get_drp_avail(evt, bmp); uwbd_evt_handle_rc_drp_avail()
|
H A D | reset.c | 323 int uwbd_msg_handle_reset(struct uwb_event *evt) uwbd_msg_handle_reset() argument 325 struct uwb_rc *rc = evt->rc; uwbd_msg_handle_reset() 352 struct uwb_event *evt; uwb_rc_reset_all() local 354 evt = kzalloc(sizeof(struct uwb_event), GFP_ATOMIC); uwb_rc_reset_all() 355 if (unlikely(evt == NULL)) uwb_rc_reset_all() 358 evt->rc = __uwb_rc_get(rc); /* will be put by uwbd's uwbd_event_handle() */ uwb_rc_reset_all() 359 evt->ts_jiffies = jiffies; uwb_rc_reset_all() 360 evt->type = UWB_EVT_TYPE_MSG; uwb_rc_reset_all() 361 evt->message = UWB_EVT_MSG_RESET; uwb_rc_reset_all() 363 uwbd_event_queue(evt); uwb_rc_reset_all()
|
H A D | address.c | 167 struct uwb_rc_evt_dev_addr_mgmt evt; uwb_rc_addr_get() local 184 result = uwb_rc_dev_addr_mgmt(rc, bmOperationType, baAddr, &evt); uwb_rc_addr_get() 188 memcpy(&dev_addr->data, evt.baAddr, uwb_rc_addr_get() 192 memcpy(&mac_addr->data, evt.baAddr, uwb_rc_addr_get() 291 * @evt: the DEV_ADDR_CONFLICT notification from the radio controller 297 int uwbd_evt_handle_rc_dev_addr_conflict(struct uwb_event *evt) uwbd_evt_handle_rc_dev_addr_conflict() argument 299 struct uwb_rc *rc = evt->rc; uwbd_evt_handle_rc_dev_addr_conflict()
|
H A D | drp.c | 775 * @evt: the DRP_IE event from the radio controller 805 int uwbd_evt_handle_rc_drp(struct uwb_event *evt) uwbd_evt_handle_rc_drp() argument 807 struct device *dev = &evt->rc->uwb_dev.dev; uwbd_evt_handle_rc_drp() 808 struct uwb_rc *rc = evt->rc; uwbd_evt_handle_rc_drp() 816 if (evt->notif.size < sizeof(*drp_evt)) { uwbd_evt_handle_rc_drp() 819 evt->notif.size, sizeof(*drp_evt)); uwbd_evt_handle_rc_drp() 822 bytes_left = evt->notif.size - sizeof(*drp_evt); uwbd_evt_handle_rc_drp() 823 drp_evt = container_of(evt->notif.rceb, struct uwb_rc_evt_drp, rceb); uwbd_evt_handle_rc_drp()
|
H A D | neh.c | 97 * evt). On reception of the notification/event, the callback (cb) is 108 __le16 evt; member in struct:uwb_rc_neh 232 neh->evt = cpu_to_le16(expected_event); uwb_rc_neh_add() 310 && neh->evt == rceb->wEvent uwb_rc_neh_match()
|
H A D | uwb-internal.h | 247 int uwbd_msg_handle_reset(struct uwb_event *evt); 254 int uwbd_evt_handle_rc_dev_addr_conflict(struct uwb_event *evt);
|
H A D | whc-rc.c | 259 dev_err(dev, "Can't allocate evt transfer buffer\n"); whcrc_setup_rc_umc()
|
/linux-4.1.27/drivers/net/wireless/ath/wil6210/ |
H A D | wmi.c | 285 struct wmi_ready_event *evt = d; wmi_evt_ready() local 287 wil->fw_version = le32_to_cpu(evt->sw_version); wmi_evt_ready() 288 wil->n_mids = evt->numof_additional_mids; wmi_evt_ready() 291 evt->mac, wil->n_mids); wmi_evt_ready() 388 struct wmi_connect_event *evt = d; wmi_evt_connect() local 398 if (len < sizeof(*evt)) { wmi_evt_connect() 402 if (len != sizeof(*evt) + evt->beacon_ie_len + evt->assoc_req_len + wmi_evt_connect() 403 evt->assoc_resp_len) { wmi_evt_connect() 406 len, (int)sizeof(*evt), evt->beacon_ie_len, wmi_evt_connect() 407 evt->assoc_req_len, evt->assoc_resp_len); wmi_evt_connect() 410 if (evt->cid >= WIL6210_MAX_CID) { wmi_evt_connect() 411 wil_err(wil, "Connect CID invalid : %d\n", evt->cid); wmi_evt_connect() 415 ch = evt->channel + 1; wmi_evt_connect() 417 evt->bssid, ch, evt->cid); wmi_evt_connect() 419 evt->assoc_info, len - sizeof(*evt), true); wmi_evt_connect() 422 assoc_req_ie = &evt->assoc_info[evt->beacon_ie_len + wmi_evt_connect() 424 assoc_req_ielen = evt->assoc_req_len - assoc_req_ie_offset; wmi_evt_connect() 425 if (evt->assoc_req_len <= assoc_req_ie_offset) { wmi_evt_connect() 430 assoc_resp_ie = &evt->assoc_info[evt->beacon_ie_len + wmi_evt_connect() 431 evt->assoc_req_len + wmi_evt_connect() 433 assoc_resp_ielen = evt->assoc_resp_len - assoc_resp_ie_offset; wmi_evt_connect() 434 if (evt->assoc_resp_len <= assoc_resp_ie_offset) { wmi_evt_connect() 446 cfg80211_connect_result(ndev, evt->bssid, wmi_evt_connect() 462 cfg80211_new_sta(ndev, evt->bssid, &sinfo, GFP_KERNEL); wmi_evt_connect() 469 ether_addr_copy(wil->sta[evt->cid].addr, evt->bssid); wmi_evt_connect() 470 wil->sta[evt->cid].status = wil_sta_conn_pending; wmi_evt_connect() 472 wil->pending_connect_cid = evt->cid; wmi_evt_connect() 479 struct wmi_disconnect_event *evt = d; wmi_evt_disconnect() local 480 u16 reason_code = le16_to_cpu(evt->protocol_reason_status); wmi_evt_disconnect() 483 evt->bssid, reason_code, evt->disconnect_reason); wmi_evt_disconnect() 488 wil6210_disconnect(wil, evt->bssid, reason_code, true); wmi_evt_disconnect() 500 struct wmi_eapol_rx_event *evt = d; wmi_evt_eapol_rx() local 501 u16 eapol_len = le16_to_cpu(evt->eapol_len); wmi_evt_eapol_rx() 509 evt->src_mac); wmi_evt_eapol_rx() 511 cid = wil_find_cid(wil, evt->src_mac); wmi_evt_eapol_rx() 528 ether_addr_copy(eth->h_source, evt->src_mac); wmi_evt_eapol_rx() 530 memcpy(skb_put(skb, eapol_len), evt->eapol, eapol_len); wmi_evt_eapol_rx() 564 struct wmi_data_port_open_event *evt = d; wmi_evt_linkup() local 565 u8 cid = evt->cid; wmi_evt_linkup() 582 struct wmi_wbe_link_down_event *evt = d; wmi_evt_linkdown() local 583 u8 cid = evt->cid; wmi_evt_linkdown() 586 cid, le32_to_cpu(evt->reason)); wmi_evt_linkdown() 600 struct wmi_vring_ba_status_event *evt = d; wmi_evt_ba_status() local 604 evt->ringid, wmi_evt_ba_status() 605 evt->status == WMI_BA_AGREED ? "OK" : "N/A", wmi_evt_ba_status() 606 evt->agg_wsize, __le16_to_cpu(evt->ba_timeout), wmi_evt_ba_status() 607 evt->amsdu ? "+" : "-"); wmi_evt_ba_status() 609 if (evt->ringid >= WIL6210_MAX_TX_RINGS) { wmi_evt_ba_status() 610 wil_err(wil, "invalid ring id %d\n", evt->ringid); wmi_evt_ba_status() 614 if (evt->status != WMI_BA_AGREED) { wmi_evt_ba_status() 615 evt->ba_timeout = 0; wmi_evt_ba_status() 616 evt->agg_wsize = 0; wmi_evt_ba_status() 617 evt->amsdu = 0; wmi_evt_ba_status() 620 txdata = &wil->vring_tx_data[evt->ringid]; wmi_evt_ba_status() 622 txdata->agg_timeout = le16_to_cpu(evt->ba_timeout); wmi_evt_ba_status() 623 txdata->agg_wsize = evt->agg_wsize; wmi_evt_ba_status() 624 txdata->agg_amsdu = evt->amsdu; wmi_evt_ba_status() 631 struct wmi_rcp_addba_req_event *evt = d; wmi_evt_addba_rx_req() local 633 wil_addba_rx_request(wil, evt->cidxtid, evt->dialog_token, wmi_evt_addba_rx_req() 634 evt->ba_param_set, evt->ba_timeout, wmi_evt_addba_rx_req() 635 evt->ba_seq_ctrl); wmi_evt_addba_rx_req() 641 struct wmi_delba_event *evt = d; variable in typeref:struct:wmi_delba_event 643 u16 reason = __le16_to_cpu(evt->reason); 648 parse_cidxtid(evt->cidxtid, &cid, &tid); 651 evt->from_initiator ? "originator" : "recipient", 653 if (!evt->from_initiator) { 716 struct pending_wmi_event *evt; wmi_recv_cmd() local 742 wil_err(wil, "Mbox evt not owned by FW?\n"); wmi_recv_cmd() 748 wil_err(wil, "Mbox evt at 0x%08x?\n", wmi_recv_cmd() 753 wil_dbg_wmi(wil, "Mbox evt %04x %04x %04x %02x\n", wmi_recv_cmd() 760 evt = kmalloc(ALIGN(offsetof(struct pending_wmi_event, wmi_recv_cmd() 763 if (!evt) wmi_recv_cmd() 766 evt->event.hdr = hdr; wmi_recv_cmd() 767 cmd = (void *)&evt->event.wmi; wmi_recv_cmd() 775 struct wil6210_mbox_hdr_wmi *wmi = &evt->event.wmi; wmi_recv_cmd() 784 wil_hex_dump_wmi("evt ", DUMP_PREFIX_OFFSET, 16, 1, wmi_recv_cmd() 785 &evt->event.hdr, sizeof(hdr) + len, true); wmi_recv_cmd() 795 list_add_tail(&evt->list, &wil->pending_wmi_ev); wmi_recv_cmd() 874 struct wmi_pcp_started_event evt; wmi_pcp_start() member in struct:__anon7741 897 if (reply.evt.status != WMI_FW_STATUS_SUCCESS) wmi_pcp_start() 1053 struct wmi_listen_started_event evt; wmi_rxon() member in struct:__anon7744 1062 if ((rc == 0) && (reply.evt.status != WMI_FW_STATUS_SUCCESS)) wmi_rxon() 1090 struct wmi_cfg_rx_chain_done_event evt; wmi_rx_chain_add() member in struct:__anon7745 1091 } __packed evt; wmi_rx_chain_add() local 1119 WMI_CFG_RX_CHAIN_DONE_EVENTID, &evt, sizeof(evt), 2000); wmi_rx_chain_add() 1123 vring->hwtail = le32_to_cpu(evt.evt.rx_ring_tail_ptr); wmi_rx_chain_add() 1126 le32_to_cpu(evt.evt.status), vring->hwtail); wmi_rx_chain_add() 1128 if (le32_to_cpu(evt.evt.status) != WMI_CFG_RX_CHAIN_SUCCESS) wmi_rx_chain_add() 1144 struct wmi_temp_sense_done_event evt; wmi_get_temperature() member in struct:__anon7746 1153 *t_bb = le32_to_cpu(reply.evt.baseband_t1000); wmi_get_temperature() 1155 *t_rf = le32_to_cpu(reply.evt.rf_t1000); wmi_get_temperature() 1233 struct wmi_rcp_addba_resp_sent_event evt; wmi_addba_rx_resp() member in struct:__anon7747 1245 if (reply.evt.status) { wmi_addba_rx_resp() 1247 le16_to_cpu(reply.evt.status)); wmi_addba_rx_resp() 1256 struct pending_wmi_event *evt, *t; wmi_event_flush() local 1260 list_for_each_entry_safe(evt, t, &wil->pending_wmi_ev, list) { wmi_event_flush() 1261 list_del(&evt->list); wmi_event_flush() 1262 kfree(evt); wmi_event_flush() 1315 print_hex_dump(KERN_ERR, "evt?? ", DUMP_PREFIX_OFFSET, 16, 1, wmi_event_handle() 1347 struct pending_wmi_event *evt; wmi_event_worker() local 1352 evt = list_entry(lh, struct pending_wmi_event, list); wmi_event_worker() 1353 wmi_event_handle(wil, &evt->event.hdr); wmi_event_worker() 1354 kfree(evt); wmi_event_worker()
|
H A D | rx_reorder.c | 341 struct wil_back_rx *evt, *t; wil_back_rx_flush() local 347 list_for_each_entry_safe(evt, t, &wil->back_rx_pending, list) { wil_back_rx_flush() 348 list_del(&evt->list); wil_back_rx_flush() 349 kfree(evt); wil_back_rx_flush() 376 struct wil_back_rx *evt; wil_back_rx_worker() local 380 evt = list_entry(lh, struct wil_back_rx, list); wil_back_rx_worker() 382 wil_back_rx_handle(wil, evt); wil_back_rx_worker() 383 kfree(evt); wil_back_rx_worker() 431 struct wil_back_tx *evt; wil_back_tx_worker() local 435 evt = list_entry(lh, struct wil_back_tx, list); wil_back_tx_worker() 437 wil_back_tx_handle(wil, evt); wil_back_tx_worker() 438 kfree(evt); wil_back_tx_worker() 444 struct wil_back_tx *evt, *t; wil_back_tx_flush() local 450 list_for_each_entry_safe(evt, t, &wil->back_tx_pending, list) { wil_back_tx_flush() 451 list_del(&evt->list); wil_back_tx_flush() 452 kfree(evt); wil_back_tx_flush()
|
H A D | cfg80211.c | 117 struct wmi_notify_req_done_event evt; wil_cid_fill_sinfo() member in struct:__anon7723 132 cid, le16_to_cpu(reply.evt.bf_mcs), wil_cid_fill_sinfo() 133 le64_to_cpu(reply.evt.tsf), reply.evt.status, wil_cid_fill_sinfo() 134 le32_to_cpu(reply.evt.snr_val), wil_cid_fill_sinfo() 135 reply.evt.sqi, wil_cid_fill_sinfo() 136 le32_to_cpu(reply.evt.tx_tpt), wil_cid_fill_sinfo() 137 le32_to_cpu(reply.evt.tx_goodput), wil_cid_fill_sinfo() 138 le32_to_cpu(reply.evt.rx_goodput), wil_cid_fill_sinfo() 139 le16_to_cpu(reply.evt.my_rx_sector), wil_cid_fill_sinfo() 140 le16_to_cpu(reply.evt.my_tx_sector), wil_cid_fill_sinfo() 141 le16_to_cpu(reply.evt.other_rx_sector), wil_cid_fill_sinfo() 142 le16_to_cpu(reply.evt.other_tx_sector)); wil_cid_fill_sinfo() 156 sinfo->txrate.mcs = le16_to_cpu(reply.evt.bf_mcs); wil_cid_fill_sinfo() 168 sinfo->signal = reply.evt.sqi; wil_cid_fill_sinfo() 528 struct wmi_sw_tx_complete_event evt; wil_cfg80211_mgmt_tx() member in struct:__anon7725 529 } __packed evt; wil_cfg80211_mgmt_tx() local 542 WMI_SW_TX_COMPLETE_EVENTID, &evt, sizeof(evt), 2000); wil_cfg80211_mgmt_tx() 544 tx_status = !evt.evt.status; wil_cfg80211_mgmt_tx()
|
H A D | debugfs.c | 930 struct wmi_notify_req_done_event evt; wil_bf_debugfs_show() member in struct:__anon7727 941 if (rc || is_all_zeros(&reply.evt, sizeof(reply.evt))) wil_bf_debugfs_show() 944 status = le32_to_cpu(reply.evt.status); wil_bf_debugfs_show() 954 le64_to_cpu(reply.evt.tsf), wil_bf_debugfs_show() 955 le16_to_cpu(reply.evt.bf_mcs), wil_bf_debugfs_show() 956 le32_to_cpu(reply.evt.tx_tpt), wil_bf_debugfs_show() 957 reply.evt.sqi, wil_bf_debugfs_show() 959 le16_to_cpu(reply.evt.my_rx_sector), wil_bf_debugfs_show() 960 le16_to_cpu(reply.evt.my_tx_sector), wil_bf_debugfs_show() 961 le16_to_cpu(reply.evt.other_rx_sector), wil_bf_debugfs_show() 962 le16_to_cpu(reply.evt.other_tx_sector), wil_bf_debugfs_show() 963 le32_to_cpu(reply.evt.rx_goodput), wil_bf_debugfs_show() 964 le32_to_cpu(reply.evt.tx_goodput)); wil_bf_debugfs_show()
|
/linux-4.1.27/arch/xtensa/kernel/ |
H A D | time.c | 56 struct clock_event_device *evt); 58 struct clock_event_device evt; member in struct:ccount_timer 81 struct clock_event_device *evt) ccount_timer_set_mode() 84 container_of(evt, struct ccount_timer, evt); ccount_timer_set_mode() 96 disable_irq(evt->irq); ccount_timer_set_mode() 103 enable_irq(evt->irq); ccount_timer_set_mode() 121 struct clock_event_device *clockevent = &timer->evt; local_timer_setup() 150 setup_irq(this_cpu_ptr(&ccount_timer)->evt.irq, &timer_irqaction); time_init() 161 struct clock_event_device *evt = &this_cpu_ptr(&ccount_timer)->evt; timer_interrupt() local 164 evt->event_handler(evt); timer_interrupt() 80 ccount_timer_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) ccount_timer_set_mode() argument
|
/linux-4.1.27/drivers/scsi/ibmvscsi/ |
H A D | ibmvfc.c | 150 * @evt: ibmvfc event struct 153 static void ibmvfc_trc_start(struct ibmvfc_event *evt) ibmvfc_trc_start() argument 155 struct ibmvfc_host *vhost = evt->vhost; ibmvfc_trc_start() 156 struct ibmvfc_cmd *vfc_cmd = &evt->iu.cmd; ibmvfc_trc_start() 157 struct ibmvfc_mad_common *mad = &evt->iu.mad_common; ibmvfc_trc_start() 161 entry->evt = evt; ibmvfc_trc_start() 163 entry->fmt = evt->crq.format; ibmvfc_trc_start() 184 * @evt: ibmvfc event struct 187 static void ibmvfc_trc_end(struct ibmvfc_event *evt) ibmvfc_trc_end() argument 189 struct ibmvfc_host *vhost = evt->vhost; ibmvfc_trc_end() 190 struct ibmvfc_cmd *vfc_cmd = &evt->xfer_iu->cmd; ibmvfc_trc_end() 191 struct ibmvfc_mad_common *mad = &evt->xfer_iu->mad_common; ibmvfc_trc_end() 194 entry->evt = evt; ibmvfc_trc_end() 196 entry->fmt = evt->crq.format; ibmvfc_trc_end() 222 #define ibmvfc_trc_start(evt) do { } while (0) 223 #define ibmvfc_trc_end(evt) do { } while (0) 743 * @evt: ibmvfc event to be checked for validity 749 struct ibmvfc_event *evt) ibmvfc_valid_event() 751 int index = evt - pool->events; ibmvfc_valid_event() 754 if (evt != pool->events + index) /* unaligned */ ibmvfc_valid_event() 761 * @evt: ibmvfc_event to be freed 764 static void ibmvfc_free_event(struct ibmvfc_event *evt) ibmvfc_free_event() argument 766 struct ibmvfc_host *vhost = evt->vhost; ibmvfc_free_event() 769 BUG_ON(!ibmvfc_valid_event(pool, evt)); ibmvfc_free_event() 770 BUG_ON(atomic_inc_return(&evt->free) != 1); ibmvfc_free_event() 771 list_add_tail(&evt->queue, &vhost->free); ibmvfc_free_event() 776 * @evt: ibmvfc event struct 781 static void ibmvfc_scsi_eh_done(struct ibmvfc_event *evt) ibmvfc_scsi_eh_done() argument 783 struct scsi_cmnd *cmnd = evt->cmnd; ibmvfc_scsi_eh_done() 790 if (evt->eh_comp) ibmvfc_scsi_eh_done() 791 complete(evt->eh_comp); ibmvfc_scsi_eh_done() 793 ibmvfc_free_event(evt); ibmvfc_scsi_eh_done() 798 * @evt: ibmvfc event struct 804 static void ibmvfc_fail_request(struct ibmvfc_event *evt, int error_code) ibmvfc_fail_request() argument 806 if (evt->cmnd) { ibmvfc_fail_request() 807 evt->cmnd->result = (error_code << 16); ibmvfc_fail_request() 808 evt->done = ibmvfc_scsi_eh_done; ibmvfc_fail_request() 810 evt->xfer_iu->mad_common.status = cpu_to_be16(IBMVFC_MAD_DRIVER_FAILED); ibmvfc_fail_request() 812 list_del(&evt->queue); ibmvfc_fail_request() 813 del_timer(&evt->timer); ibmvfc_fail_request() 814 ibmvfc_trc_end(evt); ibmvfc_fail_request() 815 evt->done(evt); ibmvfc_fail_request() 828 struct ibmvfc_event *evt, *pos; ibmvfc_purge_requests() local 831 list_for_each_entry_safe(evt, pos, &vhost->sent, queue) ibmvfc_purge_requests() 832 ibmvfc_fail_request(evt, error_code); ibmvfc_purge_requests() 1225 struct ibmvfc_event *evt = &pool->events[i]; ibmvfc_init_event_pool() local 1226 atomic_set(&evt->free, 1); ibmvfc_init_event_pool() 1227 evt->crq.valid = 0x80; ibmvfc_init_event_pool() 1228 evt->crq.ioba = cpu_to_be64(pool->iu_token + (sizeof(*evt->xfer_iu) * i)); ibmvfc_init_event_pool() 1229 evt->xfer_iu = pool->iu_storage + i; ibmvfc_init_event_pool() 1230 evt->vhost = vhost; ibmvfc_init_event_pool() 1231 evt->ext_list = NULL; ibmvfc_init_event_pool() 1232 list_add_tail(&evt->queue, &vhost->free); ibmvfc_init_event_pool() 1274 struct ibmvfc_event *evt; ibmvfc_get_event() local 1277 evt = list_entry(vhost->free.next, struct ibmvfc_event, queue); ibmvfc_get_event() 1278 atomic_set(&evt->free, 0); ibmvfc_get_event() 1279 list_del(&evt->queue); ibmvfc_get_event() 1280 return evt; ibmvfc_get_event() 1286 * @evt: The event 1290 static void ibmvfc_init_event(struct ibmvfc_event *evt, ibmvfc_init_event() argument 1293 evt->cmnd = NULL; ibmvfc_init_event() 1294 evt->sync_iu = NULL; ibmvfc_init_event() 1295 evt->crq.format = format; ibmvfc_init_event() 1296 evt->done = done; ibmvfc_init_event() 1297 evt->eh_comp = NULL; ibmvfc_init_event() 1322 * @evt: ibmvfc event struct 1330 struct ibmvfc_event *evt, ibmvfc_map_sg_data() 1363 if (!evt->ext_list) { ibmvfc_map_sg_data() 1364 evt->ext_list = dma_pool_alloc(vhost->sg_pool, GFP_ATOMIC, ibmvfc_map_sg_data() 1365 &evt->ext_list_token); ibmvfc_map_sg_data() 1367 if (!evt->ext_list) { ibmvfc_map_sg_data() 1375 ibmvfc_map_sg_list(scmd, sg_mapped, evt->ext_list); ibmvfc_map_sg_data() 1377 data->va = cpu_to_be64(evt->ext_list_token); ibmvfc_map_sg_data() 1385 * @evt: struct ibmvfc_event that timed out 1389 static void ibmvfc_timeout(struct ibmvfc_event *evt) ibmvfc_timeout() argument 1391 struct ibmvfc_host *vhost = evt->vhost; ibmvfc_timeout() 1392 dev_err(vhost->dev, "Command timed out (%p). Resetting connection\n", evt); ibmvfc_timeout() 1398 * @evt: event to be sent 1404 static int ibmvfc_send_event(struct ibmvfc_event *evt, ibmvfc_send_event() argument 1407 __be64 *crq_as_u64 = (__be64 *) &evt->crq; ibmvfc_send_event() 1411 *evt->xfer_iu = evt->iu; ibmvfc_send_event() 1412 if (evt->crq.format == IBMVFC_CMD_FORMAT) ibmvfc_send_event() 1413 evt->xfer_iu->cmd.tag = cpu_to_be64((u64)evt); ibmvfc_send_event() 1414 else if (evt->crq.format == IBMVFC_MAD_FORMAT) ibmvfc_send_event() 1415 evt->xfer_iu->mad_common.tag = cpu_to_be64((u64)evt); ibmvfc_send_event() 1419 list_add_tail(&evt->queue, &vhost->sent); ibmvfc_send_event() 1420 init_timer(&evt->timer); ibmvfc_send_event() 1423 evt->timer.data = (unsigned long) evt; ibmvfc_send_event() 1424 evt->timer.expires = jiffies + (timeout * HZ); ibmvfc_send_event() 1425 evt->timer.function = (void (*)(unsigned long))ibmvfc_timeout; ibmvfc_send_event() 1426 add_timer(&evt->timer); ibmvfc_send_event() 1433 list_del(&evt->queue); ibmvfc_send_event() 1434 del_timer(&evt->timer); ibmvfc_send_event() 1444 if (evt->cmnd) ibmvfc_send_event() 1445 scsi_dma_unmap(evt->cmnd); ibmvfc_send_event() 1446 ibmvfc_free_event(evt); ibmvfc_send_event() 1451 if (evt->cmnd) { ibmvfc_send_event() 1452 evt->cmnd->result = DID_ERROR << 16; ibmvfc_send_event() 1453 evt->done = ibmvfc_scsi_eh_done; ibmvfc_send_event() 1455 evt->xfer_iu->mad_common.status = cpu_to_be16(IBMVFC_MAD_CRQ_ERROR); ibmvfc_send_event() 1457 evt->done(evt); ibmvfc_send_event() 1459 ibmvfc_trc_start(evt); ibmvfc_send_event() 1466 * @evt: ibmvfc event to log 1469 static void ibmvfc_log_error(struct ibmvfc_event *evt) ibmvfc_log_error() argument 1471 struct ibmvfc_cmd *vfc_cmd = &evt->xfer_iu->cmd; ibmvfc_log_error() 1472 struct ibmvfc_host *vhost = evt->vhost; ibmvfc_log_error() 1474 struct scsi_cmnd *cmnd = evt->cmnd; ibmvfc_log_error() 1520 * @evt: ibmvfc event to be handled 1524 static void ibmvfc_scsi_done(struct ibmvfc_event *evt) ibmvfc_scsi_done() argument 1526 struct ibmvfc_cmd *vfc_cmd = &evt->xfer_iu->cmd; ibmvfc_scsi_done() 1528 struct scsi_cmnd *cmnd = evt->cmnd; ibmvfc_scsi_done() 1556 ibmvfc_log_error(evt); ibmvfc_scsi_done() 1567 if (evt->eh_comp) ibmvfc_scsi_done() 1568 complete(evt->eh_comp); ibmvfc_scsi_done() 1570 ibmvfc_free_event(evt); ibmvfc_scsi_done() 1617 struct ibmvfc_event *evt; ibmvfc_queuecommand_lck() local 1628 evt = ibmvfc_get_event(vhost); ibmvfc_queuecommand_lck() 1629 ibmvfc_init_event(evt, ibmvfc_scsi_done, IBMVFC_CMD_FORMAT); ibmvfc_queuecommand_lck() 1630 evt->cmnd = cmnd; ibmvfc_queuecommand_lck() 1632 vfc_cmd = &evt->iu.cmd; ibmvfc_queuecommand_lck() 1634 vfc_cmd->resp.va = cpu_to_be64(be64_to_cpu(evt->crq.ioba) + offsetof(struct ibmvfc_cmd, rsp)); ibmvfc_queuecommand_lck() 1650 if (likely(!(rc = ibmvfc_map_sg_data(cmnd, evt, vfc_cmd, vhost->dev)))) ibmvfc_queuecommand_lck() 1651 return ibmvfc_send_event(evt, vhost, 0); ibmvfc_queuecommand_lck() 1653 ibmvfc_free_event(evt); ibmvfc_queuecommand_lck() 1670 * @evt: ibmvfc event struct 1673 static void ibmvfc_sync_completion(struct ibmvfc_event *evt) ibmvfc_sync_completion() argument 1676 if (evt->sync_iu) ibmvfc_sync_completion() 1677 *evt->sync_iu = *evt->xfer_iu; ibmvfc_sync_completion() 1679 complete(&evt->comp); ibmvfc_sync_completion() 1684 * @evt: struct ibmvfc_event 1687 static void ibmvfc_bsg_timeout_done(struct ibmvfc_event *evt) ibmvfc_bsg_timeout_done() argument 1689 struct ibmvfc_host *vhost = evt->vhost; ibmvfc_bsg_timeout_done() 1691 ibmvfc_free_event(evt); ibmvfc_bsg_timeout_done() 1707 struct ibmvfc_event *evt; ibmvfc_bsg_timeout() local 1721 evt = ibmvfc_get_event(vhost); ibmvfc_bsg_timeout() 1722 ibmvfc_init_event(evt, ibmvfc_bsg_timeout_done, IBMVFC_MAD_FORMAT); ibmvfc_bsg_timeout() 1724 tmf = &evt->iu.tmf; ibmvfc_bsg_timeout() 1732 rc = ibmvfc_send_event(evt, vhost, default_timeout); ibmvfc_bsg_timeout() 1760 struct ibmvfc_event *evt; ibmvfc_bsg_plogi() local 1779 evt = ibmvfc_get_event(vhost); ibmvfc_bsg_plogi() 1780 ibmvfc_init_event(evt, ibmvfc_sync_completion, IBMVFC_MAD_FORMAT); ibmvfc_bsg_plogi() 1781 plogi = &evt->iu.plogi; ibmvfc_bsg_plogi() 1787 evt->sync_iu = &rsp_iu; ibmvfc_bsg_plogi() 1788 init_completion(&evt->comp); ibmvfc_bsg_plogi() 1790 rc = ibmvfc_send_event(evt, vhost, default_timeout); ibmvfc_bsg_plogi() 1796 wait_for_completion(&evt->comp); ibmvfc_bsg_plogi() 1802 ibmvfc_free_event(evt); ibmvfc_bsg_plogi() 1821 struct ibmvfc_event *evt; ibmvfc_bsg_request() local 1893 evt = ibmvfc_get_event(vhost); ibmvfc_bsg_request() 1894 ibmvfc_init_event(evt, ibmvfc_sync_completion, IBMVFC_MAD_FORMAT); ibmvfc_bsg_request() 1895 mad = &evt->iu.passthru; ibmvfc_bsg_request() 1902 mad->cmd_ioba.va = cpu_to_be64(be64_to_cpu(evt->crq.ioba) + ibmvfc_bsg_request() 1916 mad->iu.tag = cpu_to_be64((u64)evt); ibmvfc_bsg_request() 1919 evt->sync_iu = &rsp_iu; ibmvfc_bsg_request() 1920 init_completion(&evt->comp); ibmvfc_bsg_request() 1921 rc = ibmvfc_send_event(evt, vhost, 0); ibmvfc_bsg_request() 1929 wait_for_completion(&evt->comp); ibmvfc_bsg_request() 1937 ibmvfc_free_event(evt); ibmvfc_bsg_request() 1966 struct ibmvfc_event *evt = NULL; ibmvfc_reset_device() local 1975 evt = ibmvfc_get_event(vhost); ibmvfc_reset_device() 1976 ibmvfc_init_event(evt, ibmvfc_sync_completion, IBMVFC_CMD_FORMAT); ibmvfc_reset_device() 1978 tmf = &evt->iu.cmd; ibmvfc_reset_device() 1980 tmf->resp.va = cpu_to_be64(be64_to_cpu(evt->crq.ioba) + offsetof(struct ibmvfc_cmd, rsp)); ibmvfc_reset_device() 1990 evt->sync_iu = &rsp_iu; ibmvfc_reset_device() 1992 init_completion(&evt->comp); ibmvfc_reset_device() 1993 rsp_rc = ibmvfc_send_event(evt, vhost, default_timeout); ibmvfc_reset_device() 2004 wait_for_completion(&evt->comp); ibmvfc_reset_device() 2023 ibmvfc_free_event(evt); ibmvfc_reset_device() 2030 * @evt: ibmvfc event struct 2036 static int ibmvfc_match_rport(struct ibmvfc_event *evt, void *rport) ibmvfc_match_rport() argument 2040 if (evt->cmnd) { ibmvfc_match_rport() 2041 cmd_rport = starget_to_rport(scsi_target(evt->cmnd->device)); ibmvfc_match_rport() 2050 * @evt: ibmvfc event struct 2056 static int ibmvfc_match_target(struct ibmvfc_event *evt, void *device) ibmvfc_match_target() argument 2058 if (evt->cmnd && scsi_target(evt->cmnd->device) == device) ibmvfc_match_target() 2065 * @evt: ibmvfc event struct 2071 static int ibmvfc_match_lun(struct ibmvfc_event *evt, void *device) ibmvfc_match_lun() argument 2073 if (evt->cmnd && evt->cmnd->device == device) ibmvfc_match_lun() 2090 struct ibmvfc_event *evt; ibmvfc_wait_for_ops() local 2100 list_for_each_entry(evt, &vhost->sent, queue) { ibmvfc_wait_for_ops() 2101 if (match(evt, device)) { ibmvfc_wait_for_ops() 2102 evt->eh_comp = ∁ ibmvfc_wait_for_ops() 2114 list_for_each_entry(evt, &vhost->sent, queue) { ibmvfc_wait_for_ops() 2115 if (match(evt, device)) { ibmvfc_wait_for_ops() 2116 evt->eh_comp = NULL; ibmvfc_wait_for_ops() 2150 struct ibmvfc_event *evt, *found_evt; ibmvfc_cancel_all() local 2159 list_for_each_entry(evt, &vhost->sent, queue) { ibmvfc_cancel_all() 2160 if (evt->cmnd && evt->cmnd->device == sdev) { ibmvfc_cancel_all() 2161 found_evt = evt; ibmvfc_cancel_all() 2174 evt = ibmvfc_get_event(vhost); ibmvfc_cancel_all() 2175 ibmvfc_init_event(evt, ibmvfc_sync_completion, IBMVFC_MAD_FORMAT); ibmvfc_cancel_all() 2177 tmf = &evt->iu.tmf; ibmvfc_cancel_all() 2193 evt->sync_iu = &rsp; ibmvfc_cancel_all() 2194 init_completion(&evt->comp); ibmvfc_cancel_all() 2195 rsp_rc = ibmvfc_send_event(evt, vhost, default_timeout); ibmvfc_cancel_all() 2210 wait_for_completion(&evt->comp); ibmvfc_cancel_all() 2213 ibmvfc_free_event(evt); ibmvfc_cancel_all() 2235 * @evt: ibmvfc event struct 2241 static int ibmvfc_match_key(struct ibmvfc_event *evt, void *key) ibmvfc_match_key() argument 2245 if (evt->crq.format == IBMVFC_CMD_FORMAT && ibmvfc_match_key() 2246 be32_to_cpu(evt->iu.cmd.cancel_key) == cancel_key) ibmvfc_match_key() 2253 * @evt: ibmvfc event struct 2259 static int ibmvfc_match_evt(struct ibmvfc_event *evt, void *match) ibmvfc_match_evt() argument 2261 if (evt == match) ibmvfc_match_evt() 2281 struct ibmvfc_event *evt, *found_evt; ibmvfc_abort_task_set() local 2290 list_for_each_entry(evt, &vhost->sent, queue) { ibmvfc_abort_task_set() 2291 if (evt->cmnd && evt->cmnd->device == sdev) { ibmvfc_abort_task_set() 2292 found_evt = evt; ibmvfc_abort_task_set() 2305 evt = ibmvfc_get_event(vhost); ibmvfc_abort_task_set() 2306 ibmvfc_init_event(evt, ibmvfc_sync_completion, IBMVFC_CMD_FORMAT); ibmvfc_abort_task_set() 2308 tmf = &evt->iu.cmd; ibmvfc_abort_task_set() 2310 tmf->resp.va = cpu_to_be64(be64_to_cpu(evt->crq.ioba) + offsetof(struct ibmvfc_cmd, rsp)); ibmvfc_abort_task_set() 2320 evt->sync_iu = &rsp_iu; ibmvfc_abort_task_set() 2322 init_completion(&evt->comp); ibmvfc_abort_task_set() 2323 rsp_rc = ibmvfc_send_event(evt, vhost, default_timeout); ibmvfc_abort_task_set() 2334 timeout = wait_for_completion_timeout(&evt->comp, timeout); ibmvfc_abort_task_set() 2353 rc = ibmvfc_wait_for_ops(vhost, evt, ibmvfc_match_evt); ibmvfc_abort_task_set() 2383 ibmvfc_free_event(evt); ibmvfc_abort_task_set() 2724 struct ibmvfc_event *evt = (struct ibmvfc_event *)be64_to_cpu(crq->ioba); ibmvfc_handle_crq() local 2778 if (unlikely(!ibmvfc_valid_event(&vhost->pool, evt))) { ibmvfc_handle_crq() 2784 if (unlikely(atomic_read(&evt->free))) { ibmvfc_handle_crq() 2790 del_timer(&evt->timer); ibmvfc_handle_crq() 2791 list_del(&evt->queue); ibmvfc_handle_crq() 2792 ibmvfc_trc_end(evt); ibmvfc_handle_crq() 2793 evt->done(evt); ibmvfc_handle_crq() 3288 * @evt: ibmvfc event struct 3291 static void ibmvfc_tgt_prli_done(struct ibmvfc_event *evt) ibmvfc_tgt_prli_done() argument 3293 struct ibmvfc_target *tgt = evt->tgt; ibmvfc_tgt_prli_done() 3294 struct ibmvfc_host *vhost = evt->vhost; ibmvfc_tgt_prli_done() 3295 struct ibmvfc_process_login *rsp = &evt->xfer_iu->prli; ibmvfc_tgt_prli_done() 3351 ibmvfc_free_event(evt); ibmvfc_tgt_prli_done() 3364 struct ibmvfc_event *evt; ibmvfc_tgt_send_prli() local 3370 evt = ibmvfc_get_event(vhost); ibmvfc_tgt_send_prli() 3372 ibmvfc_init_event(evt, ibmvfc_tgt_prli_done, IBMVFC_MAD_FORMAT); ibmvfc_tgt_send_prli() 3373 evt->tgt = tgt; ibmvfc_tgt_send_prli() 3374 prli = &evt->iu.prli; ibmvfc_tgt_send_prli() 3386 if (ibmvfc_send_event(evt, vhost, default_timeout)) { ibmvfc_tgt_send_prli() 3396 * @evt: ibmvfc event struct 3399 static void ibmvfc_tgt_plogi_done(struct ibmvfc_event *evt) ibmvfc_tgt_plogi_done() argument 3401 struct ibmvfc_target *tgt = evt->tgt; ibmvfc_tgt_plogi_done() 3402 struct ibmvfc_host *vhost = evt->vhost; ibmvfc_tgt_plogi_done() 3403 struct ibmvfc_port_login *rsp = &evt->xfer_iu->plogi; ibmvfc_tgt_plogi_done() 3447 ibmvfc_free_event(evt); ibmvfc_tgt_plogi_done() 3460 struct ibmvfc_event *evt; ibmvfc_tgt_send_plogi() local 3467 evt = ibmvfc_get_event(vhost); ibmvfc_tgt_send_plogi() 3470 ibmvfc_init_event(evt, ibmvfc_tgt_plogi_done, IBMVFC_MAD_FORMAT); ibmvfc_tgt_send_plogi() 3471 evt->tgt = tgt; ibmvfc_tgt_send_plogi() 3472 plogi = &evt->iu.plogi; ibmvfc_tgt_send_plogi() 3479 if (ibmvfc_send_event(evt, vhost, default_timeout)) { ibmvfc_tgt_send_plogi() 3489 * @evt: ibmvfc event struct 3492 static void ibmvfc_tgt_implicit_logout_done(struct ibmvfc_event *evt) ibmvfc_tgt_implicit_logout_done() argument 3494 struct ibmvfc_target *tgt = evt->tgt; ibmvfc_tgt_implicit_logout_done() 3495 struct ibmvfc_host *vhost = evt->vhost; ibmvfc_tgt_implicit_logout_done() 3496 struct ibmvfc_implicit_logout *rsp = &evt->xfer_iu->implicit_logout; ibmvfc_tgt_implicit_logout_done() 3500 ibmvfc_free_event(evt); ibmvfc_tgt_implicit_logout_done() 3535 struct ibmvfc_event *evt; ibmvfc_tgt_implicit_logout() local 3541 evt = ibmvfc_get_event(vhost); ibmvfc_tgt_implicit_logout() 3543 ibmvfc_init_event(evt, ibmvfc_tgt_implicit_logout_done, IBMVFC_MAD_FORMAT); ibmvfc_tgt_implicit_logout() 3544 evt->tgt = tgt; ibmvfc_tgt_implicit_logout() 3545 mad = &evt->iu.implicit_logout; ibmvfc_tgt_implicit_logout() 3553 if (ibmvfc_send_event(evt, vhost, default_timeout)) { ibmvfc_tgt_implicit_logout() 3585 * @evt: ibmvfc event struct 3588 static void ibmvfc_tgt_adisc_done(struct ibmvfc_event *evt) ibmvfc_tgt_adisc_done() argument 3590 struct ibmvfc_target *tgt = evt->tgt; ibmvfc_tgt_adisc_done() 3591 struct ibmvfc_host *vhost = evt->vhost; ibmvfc_tgt_adisc_done() 3592 struct ibmvfc_passthru_mad *mad = &evt->xfer_iu->passthru; ibmvfc_tgt_adisc_done() 3622 ibmvfc_free_event(evt); ibmvfc_tgt_adisc_done() 3628 * @evt: ibmvfc event struct 3631 static void ibmvfc_init_passthru(struct ibmvfc_event *evt) ibmvfc_init_passthru() argument 3633 struct ibmvfc_passthru_mad *mad = &evt->iu.passthru; ibmvfc_init_passthru() 3639 mad->cmd_ioba.va = cpu_to_be64((u64)be64_to_cpu(evt->crq.ioba) + ibmvfc_init_passthru() 3644 mad->iu.cmd.va = cpu_to_be64((u64)be64_to_cpu(evt->crq.ioba) + ibmvfc_init_passthru() 3648 mad->iu.rsp.va = cpu_to_be64((u64)be64_to_cpu(evt->crq.ioba) + ibmvfc_init_passthru() 3656 * @evt: ibmvfc event struct 3664 static void ibmvfc_tgt_adisc_cancel_done(struct ibmvfc_event *evt) ibmvfc_tgt_adisc_cancel_done() argument 3666 struct ibmvfc_host *vhost = evt->vhost; ibmvfc_tgt_adisc_cancel_done() 3667 struct ibmvfc_target *tgt = evt->tgt; ibmvfc_tgt_adisc_cancel_done() 3671 ibmvfc_free_event(evt); ibmvfc_tgt_adisc_cancel_done() 3687 struct ibmvfc_event *evt; ibmvfc_adisc_timeout() local 3704 evt = ibmvfc_get_event(vhost); ibmvfc_adisc_timeout() 3705 ibmvfc_init_event(evt, ibmvfc_tgt_adisc_cancel_done, IBMVFC_MAD_FORMAT); ibmvfc_adisc_timeout() 3707 evt->tgt = tgt; ibmvfc_adisc_timeout() 3708 tmf = &evt->iu.tmf; ibmvfc_adisc_timeout() 3716 rc = ibmvfc_send_event(evt, vhost, default_timeout); ibmvfc_adisc_timeout() 3743 struct ibmvfc_event *evt; ibmvfc_tgt_adisc() local 3749 evt = ibmvfc_get_event(vhost); ibmvfc_tgt_adisc() 3751 ibmvfc_init_event(evt, ibmvfc_tgt_adisc_done, IBMVFC_MAD_FORMAT); ibmvfc_tgt_adisc() 3752 evt->tgt = tgt; ibmvfc_tgt_adisc() 3754 ibmvfc_init_passthru(evt); ibmvfc_tgt_adisc() 3755 mad = &evt->iu.passthru; ibmvfc_tgt_adisc() 3777 if (ibmvfc_send_event(evt, vhost, IBMVFC_ADISC_PLUS_CANCEL_TIMEOUT)) { ibmvfc_tgt_adisc() 3788 * @evt: ibmvfc event struct 3791 static void ibmvfc_tgt_query_target_done(struct ibmvfc_event *evt) ibmvfc_tgt_query_target_done() argument 3793 struct ibmvfc_target *tgt = evt->tgt; ibmvfc_tgt_query_target_done() 3794 struct ibmvfc_host *vhost = evt->vhost; ibmvfc_tgt_query_target_done() 3795 struct ibmvfc_query_tgt *rsp = &evt->xfer_iu->query_tgt; ibmvfc_tgt_query_target_done() 3835 ibmvfc_free_event(evt); ibmvfc_tgt_query_target_done() 3848 struct ibmvfc_event *evt; ibmvfc_tgt_query_target() local 3854 evt = ibmvfc_get_event(vhost); ibmvfc_tgt_query_target() 3856 evt->tgt = tgt; ibmvfc_tgt_query_target() 3857 ibmvfc_init_event(evt, ibmvfc_tgt_query_target_done, IBMVFC_MAD_FORMAT); ibmvfc_tgt_query_target() 3858 query_tgt = &evt->iu.query_tgt; ibmvfc_tgt_query_target() 3866 if (ibmvfc_send_event(evt, vhost, default_timeout)) { ibmvfc_tgt_query_target() 3942 * @evt: ibmvfc event struct 3945 static void ibmvfc_discover_targets_done(struct ibmvfc_event *evt) ibmvfc_discover_targets_done() argument 3947 struct ibmvfc_host *vhost = evt->vhost; ibmvfc_discover_targets_done() 3948 struct ibmvfc_discover_targets *rsp = &evt->xfer_iu->discover_targets; ibmvfc_discover_targets_done() 3972 ibmvfc_free_event(evt); ibmvfc_discover_targets_done() 3984 struct ibmvfc_event *evt = ibmvfc_get_event(vhost); ibmvfc_discover_targets() local 3986 ibmvfc_init_event(evt, ibmvfc_discover_targets_done, IBMVFC_MAD_FORMAT); ibmvfc_discover_targets() 3987 mad = &evt->iu.discover_targets; ibmvfc_discover_targets() 3997 if (!ibmvfc_send_event(evt, vhost, default_timeout)) ibmvfc_discover_targets() 4005 * @evt: ibmvfc event struct 4008 static void ibmvfc_npiv_login_done(struct ibmvfc_event *evt) ibmvfc_npiv_login_done() argument 4010 struct ibmvfc_host *vhost = evt->vhost; ibmvfc_npiv_login_done() 4011 u32 mad_status = be16_to_cpu(evt->xfer_iu->npiv_login.common.status); ibmvfc_npiv_login_done() 4018 ibmvfc_free_event(evt); ibmvfc_npiv_login_done() 4028 ibmvfc_free_event(evt); ibmvfc_npiv_login_done() 4033 ibmvfc_free_event(evt); ibmvfc_npiv_login_done() 4038 ibmvfc_free_event(evt); ibmvfc_npiv_login_done() 4095 struct ibmvfc_event *evt = ibmvfc_get_event(vhost); ibmvfc_npiv_login() local 4099 ibmvfc_init_event(evt, ibmvfc_npiv_login_done, IBMVFC_MAD_FORMAT); ibmvfc_npiv_login() 4102 mad = &evt->iu.npiv_login; ibmvfc_npiv_login() 4112 if (!ibmvfc_send_event(evt, vhost, default_timeout)) ibmvfc_npiv_login() 4123 static void ibmvfc_npiv_logout_done(struct ibmvfc_event *evt) ibmvfc_npiv_logout_done() argument 4125 struct ibmvfc_host *vhost = evt->vhost; ibmvfc_npiv_logout_done() 4126 u32 mad_status = be16_to_cpu(evt->xfer_iu->npiv_logout.common.status); ibmvfc_npiv_logout_done() 4128 ibmvfc_free_event(evt); ibmvfc_npiv_logout_done() 4158 struct ibmvfc_event *evt; ibmvfc_npiv_logout() local 4160 evt = ibmvfc_get_event(vhost); ibmvfc_npiv_logout() 4161 ibmvfc_init_event(evt, ibmvfc_npiv_logout_done, IBMVFC_MAD_FORMAT); ibmvfc_npiv_logout() 4163 mad = &evt->iu.npiv_logout; ibmvfc_npiv_logout() 4171 if (!ibmvfc_send_event(evt, vhost, default_timeout)) ibmvfc_npiv_logout() 748 ibmvfc_valid_event(struct ibmvfc_event_pool *pool, struct ibmvfc_event *evt) ibmvfc_valid_event() argument 1329 ibmvfc_map_sg_data(struct scsi_cmnd *scmd, struct ibmvfc_event *evt, struct ibmvfc_cmd *vfc_cmd, struct device *dev) ibmvfc_map_sg_data() argument
|
H A D | ibmvscsi.c | 474 struct srp_event_struct *evt = &pool->events[i]; initialize_event_pool() local 475 memset(&evt->crq, 0x00, sizeof(evt->crq)); initialize_event_pool() 476 atomic_set(&evt->free, 1); initialize_event_pool() 477 evt->crq.valid = 0x80; initialize_event_pool() 478 evt->crq.IU_length = cpu_to_be16(sizeof(*evt->xfer_iu)); initialize_event_pool() 479 evt->crq.IU_data_ptr = cpu_to_be64(pool->iu_token + initialize_event_pool() 480 sizeof(*evt->xfer_iu) * i); initialize_event_pool() 481 evt->xfer_iu = pool->iu_storage + i; initialize_event_pool() 482 evt->hostdata = hostdata; initialize_event_pool() 483 evt->ext_list = NULL; initialize_event_pool() 484 evt->ext_list_token = 0; initialize_event_pool() 523 * @evt: srp_event_struct to be checked for validity 528 struct srp_event_struct *evt) valid_event_struct() 530 int index = evt - pool->events; valid_event_struct() 533 if (evt != pool->events + index) /* unaligned */ valid_event_struct() 541 * @evt: srp_event_struct to be modified 545 struct srp_event_struct *evt) free_event_struct() 547 if (!valid_event_struct(pool, evt)) { free_event_struct() 548 dev_err(evt->hostdata->dev, "Freeing invalid event_struct %p " free_event_struct() 549 "(not in pool %p)\n", evt, pool->events); free_event_struct() 552 if (atomic_inc_return(&evt->free) != 1) { free_event_struct() 553 dev_err(evt->hostdata->dev, "Freeing event_struct %p " free_event_struct() 554 "which is not in use!\n", evt); free_event_struct() 588 * @evt: The event 794 struct srp_event_struct *evt; purge_requests() local 799 evt = list_first_entry(&hostdata->sent, struct srp_event_struct, list); purge_requests() 800 list_del(&evt->list); purge_requests() 801 del_timer(&evt->timer); purge_requests() 804 if (evt->cmnd) { purge_requests() 805 evt->cmnd->result = (error_code << 16); purge_requests() 806 unmap_cmd_data(&evt->iu.srp.cmd, evt, purge_requests() 807 evt->hostdata->dev); purge_requests() 808 if (evt->cmnd_done) purge_requests() 809 evt->cmnd_done(evt->cmnd); purge_requests() 810 } else if (evt->done && evt->crq.format != VIOSRP_MAD_FORMAT && purge_requests() 811 evt->iu.srp.login_req.opcode != SRP_LOGIN_REQ) purge_requests() 812 evt->done(evt); purge_requests() 813 free_event_struct(&evt->hostdata->pool, evt); purge_requests() 1487 struct srp_event_struct *evt; ibmvscsi_eh_abort_handler() local 1514 evt = get_event_struct(&hostdata->pool); ibmvscsi_eh_abort_handler() 1515 if (evt == NULL) { ibmvscsi_eh_abort_handler() 1522 init_event_struct(evt, ibmvscsi_eh_abort_handler() 1527 tsk_mgmt = &evt->iu.srp.tsk_mgmt; ibmvscsi_eh_abort_handler() 1536 evt->sync_srp = &srp_rsp; ibmvscsi_eh_abort_handler() 1538 init_completion(&evt->comp); ibmvscsi_eh_abort_handler() 1539 rsp_rc = ibmvscsi_send_srp_event(evt, hostdata, abort_timeout * 2); ibmvscsi_eh_abort_handler() 1561 wait_for_completion(&evt->comp); ibmvscsi_eh_abort_handler() 1626 struct srp_event_struct *evt; ibmvscsi_eh_device_reset_handler() local 1637 evt = get_event_struct(&hostdata->pool); ibmvscsi_eh_device_reset_handler() 1638 if (evt == NULL) { ibmvscsi_eh_device_reset_handler() 1645 init_event_struct(evt, ibmvscsi_eh_device_reset_handler() 1650 tsk_mgmt = &evt->iu.srp.tsk_mgmt; ibmvscsi_eh_device_reset_handler() 1658 evt->sync_srp = &srp_rsp; ibmvscsi_eh_device_reset_handler() 1660 init_completion(&evt->comp); ibmvscsi_eh_device_reset_handler() 1661 rsp_rc = ibmvscsi_send_srp_event(evt, hostdata, reset_timeout * 2); ibmvscsi_eh_device_reset_handler() 1682 wait_for_completion(&evt->comp); ibmvscsi_eh_device_reset_handler() 527 valid_event_struct(struct event_pool *pool, struct srp_event_struct *evt) valid_event_struct() argument 544 free_event_struct(struct event_pool *pool, struct srp_event_struct *evt) free_event_struct() argument
|
H A D | ibmvfc.h | 507 struct ibmvfc_event *evt; member in struct:ibmvfc_trace_entry
|
/linux-4.1.27/arch/x86/xen/ |
H A D | time.c | 278 struct clock_event_device *evt) xen_timerop_set_mode() 298 struct clock_event_device *evt) xen_timerop_set_next_event() 300 WARN_ON(evt->mode != CLOCK_EVT_MODE_ONESHOT); xen_timerop_set_next_event() 330 struct clock_event_device *evt) xen_vcpuop_set_mode() 356 struct clock_event_device *evt) xen_vcpuop_set_next_event() 362 WARN_ON(evt->mode != CLOCK_EVT_MODE_ONESHOT); xen_vcpuop_set_next_event() 393 struct clock_event_device evt; member in struct:xen_clock_event_device 396 static DEFINE_PER_CPU(struct xen_clock_event_device, xen_clock_events) = { .evt.irq = -1 }; 400 struct clock_event_device *evt = this_cpu_ptr(&xen_clock_events.evt); xen_timer_interrupt() local 404 if (evt->event_handler) { xen_timer_interrupt() 405 evt->event_handler(evt); xen_timer_interrupt() 416 struct clock_event_device *evt; xen_teardown_timer() local 418 evt = &per_cpu(xen_clock_events, cpu).evt; xen_teardown_timer() 420 if (evt->irq >= 0) { xen_teardown_timer() 421 unbind_from_irqhandler(evt->irq, NULL); xen_teardown_timer() 422 evt->irq = -1; xen_teardown_timer() 429 struct clock_event_device *evt = &xevt->evt; xen_setup_timer() local 432 WARN(evt->irq >= 0, "IRQ%d for CPU%d is already allocated\n", evt->irq, cpu); xen_setup_timer() 433 if (evt->irq >= 0) xen_setup_timer() 446 memcpy(evt, xen_clockevent, sizeof(*evt)); xen_setup_timer() 448 evt->cpumask = cpumask_of(cpu); xen_setup_timer() 449 evt->irq = irq; xen_setup_timer() 455 clockevents_register_device(this_cpu_ptr(&xen_clock_events.evt)); xen_setup_cpu_clockevents() 277 xen_timerop_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) xen_timerop_set_mode() argument 297 xen_timerop_set_next_event(unsigned long delta, struct clock_event_device *evt) xen_timerop_set_next_event() argument 329 xen_vcpuop_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) xen_vcpuop_set_mode() argument 355 xen_vcpuop_set_next_event(unsigned long delta, struct clock_event_device *evt) xen_vcpuop_set_next_event() argument
|
/linux-4.1.27/arch/m68k/coldfire/ |
H A D | pit.c | 46 struct clock_event_device *evt) init_cf_pit_timer() 84 struct clock_event_device *evt) cf_pit_next_event() 105 struct clock_event_device *evt = &cf_pit_clockevent; pit_tick() local 113 evt->event_handler(evt); pit_tick() 45 init_cf_pit_timer(enum clock_event_mode mode, struct clock_event_device *evt) init_cf_pit_timer() argument 83 cf_pit_next_event(unsigned long delta, struct clock_event_device *evt) cf_pit_next_event() argument
|
/linux-4.1.27/drivers/scsi/csiostor/ |
H A D | csio_rnode.c | 315 "Unhandled rport_type:%d recv in rdev evt " csio_confirm_rnode() 600 * @evt - SM event. 604 csio_rns_uninit(struct csio_rnode *rn, enum csio_rn_ev evt) csio_rns_uninit() argument 609 CSIO_INC_STATS(rn, n_evt_sm[evt]); csio_rns_uninit() 611 switch (evt) { csio_rns_uninit() 625 "in rn state[uninit]\n", csio_rn_flowid(rn), evt); csio_rns_uninit() 631 "in rn state[uninit]\n", csio_rn_flowid(rn), evt); csio_rns_uninit() 640 * @evt - SM event. 644 csio_rns_ready(struct csio_rnode *rn, enum csio_rn_ev evt) csio_rns_ready() argument 649 CSIO_INC_STATS(rn, n_evt_sm[evt]); csio_rns_ready() 651 switch (evt) { csio_rns_ready() 656 "in rn state[ready]\n", csio_rn_flowid(rn), evt, csio_rns_ready() 715 "in rn state[uninit]\n", csio_rn_flowid(rn), evt, csio_rns_ready() 725 * @evt - SM event. 729 csio_rns_offline(struct csio_rnode *rn, enum csio_rn_ev evt) csio_rns_offline() argument 734 CSIO_INC_STATS(rn, n_evt_sm[evt]); csio_rns_offline() 736 switch (evt) { csio_rns_offline() 752 "in rn state[offline]\n", csio_rn_flowid(rn), evt, csio_rns_offline() 774 "in rn state[offline]\n", csio_rn_flowid(rn), evt, csio_rns_offline() 784 * @evt - SM event. 788 csio_rns_disappeared(struct csio_rnode *rn, enum csio_rn_ev evt) csio_rns_disappeared() argument 793 CSIO_INC_STATS(rn, n_evt_sm[evt]); csio_rns_disappeared() 795 switch (evt) { csio_rns_disappeared() 823 evt, rn->nport_id); csio_rns_disappeared() 830 evt, rn->nport_id); csio_rns_disappeared() 871 enum csio_rn_ev evt; csio_rnode_fwevt_handler() local 873 evt = CSIO_FWE_TO_RNFE(fwevt); csio_rnode_fwevt_handler() 874 if (!evt) { csio_rnode_fwevt_handler() 887 csio_post_event(&rn->sm, evt); csio_rnode_fwevt_handler()
|
H A D | csio_lnode.c | 907 * @evt - Given rnode event 917 csio_post_event_rns(struct csio_lnode *ln, enum csio_rn_ev evt) csio_post_event_rns() argument 925 csio_post_event(&rn->sm, evt); csio_post_event_rns() 955 * @evt - Given lnode event 965 csio_post_event_lns(struct csio_lnode *ln, enum csio_ln_ev evt) csio_post_event_lns() argument 970 /* If NPIV lnode, send evt only to that and return */ csio_post_event_lns() 972 csio_post_event(&ln->sm, evt); csio_post_event_lns() 977 /* Traverse children lnodes list and send evt */ csio_post_event_lns() 980 csio_post_event(&cln->sm, evt); csio_post_event_lns() 983 /* Send evt to parent lnode */ csio_post_event_lns() 984 csio_post_event(&ln->sm, evt); csio_post_event_lns() 1089 * @evt - Event to be processed. 1096 csio_lns_uninit(struct csio_lnode *ln, enum csio_ln_ev evt) csio_lns_uninit() argument 1102 CSIO_INC_STATS(ln, n_evt_sm[evt]); csio_lns_uninit() 1103 switch (evt) { csio_lns_uninit() 1133 "ln state[uninit].\n", evt, ln->nport_id); csio_lns_uninit() 1142 * @evt - Event to be processed. 1149 csio_lns_online(struct csio_lnode *ln, enum csio_ln_ev evt) csio_lns_online() argument 1153 CSIO_INC_STATS(ln, n_evt_sm[evt]); csio_lns_online() 1154 switch (evt) { csio_lns_online() 1184 "ln state[uninit].\n", evt, ln->nport_id); csio_lns_online() 1194 * @evt - Event to be processed. 1201 csio_lns_ready(struct csio_lnode *ln, enum csio_ln_ev evt) csio_lns_ready() argument 1205 CSIO_INC_STATS(ln, n_evt_sm[evt]); csio_lns_ready() 1206 switch (evt) { csio_lns_ready() 1210 "in ln state[ready].\n", evt, ln->nport_id); csio_lns_ready() 1258 "ln state[uninit].\n", evt, ln->nport_id); csio_lns_ready() 1268 * @evt - Event to be processed. 1275 csio_lns_offline(struct csio_lnode *ln, enum csio_ln_ev evt) csio_lns_offline() argument 1281 CSIO_INC_STATS(ln, n_evt_sm[evt]); csio_lns_offline() 1282 switch (evt) { csio_lns_offline() 1311 "in ln state[offline].\n", evt, ln->nport_id); csio_lns_offline() 1323 "ln state[offline]\n", evt, ln->nport_id); csio_lns_offline() 1476 enum csio_ln_ev evt; csio_fcoe_fwevt_handler() local 1561 evt = CSIO_FWE_TO_LNE(rdev_wr->event_cause); csio_fcoe_fwevt_handler() 1562 if (evt) { csio_fcoe_fwevt_handler() 1565 "cause:%d flowid:x%x\n", evt, csio_fcoe_fwevt_handler() 1567 csio_post_event(&ln->sm, evt); csio_fcoe_fwevt_handler() 1911 /* Traverse children lnodes list and send evt */ csio_notify_lnodes() 1953 /* Traverse sibling lnodes list and send evt */ csio_disable_lnodes()
|
H A D | csio_defs.h | 96 csio_post_event(void *smp, uint32_t evt) csio_post_event() argument 98 ((struct csio_sm *)smp)->sm_state(smp, evt); csio_post_event()
|
H A D | csio_hw.c | 2325 * @evt - Event 2329 csio_hws_uninit(struct csio_hw *hw, enum csio_hw_ev evt) csio_hws_uninit() argument 2332 hw->cur_evt = evt; csio_hws_uninit() 2333 CSIO_INC_STATS(hw, n_evt_sm[evt]); csio_hws_uninit() 2335 switch (evt) { csio_hws_uninit() 2350 * @evt - Event 2354 csio_hws_configuring(struct csio_hw *hw, enum csio_hw_ev evt) csio_hws_configuring() argument 2357 hw->cur_evt = evt; csio_hws_configuring() 2358 CSIO_INC_STATS(hw, n_evt_sm[evt]); csio_hws_configuring() 2360 switch (evt) { csio_hws_configuring() 2388 * @evt - Event 2392 csio_hws_initializing(struct csio_hw *hw, enum csio_hw_ev evt) csio_hws_initializing() argument 2395 hw->cur_evt = evt; csio_hws_initializing() 2396 CSIO_INC_STATS(hw, n_evt_sm[evt]); csio_hws_initializing() 2398 switch (evt) { csio_hws_initializing() 2426 * @evt - Event 2430 csio_hws_ready(struct csio_hw *hw, enum csio_hw_ev evt) csio_hws_ready() argument 2433 hw->evtflag = evt; csio_hws_ready() 2436 hw->cur_evt = evt; csio_hws_ready() 2437 CSIO_INC_STATS(hw, n_evt_sm[evt]); csio_hws_ready() 2439 switch (evt) { csio_hws_ready() 2447 if (evt == CSIO_HWE_HBA_RESET || csio_hws_ready() 2448 evt == CSIO_HWE_PCIERR_DETECTED) csio_hws_ready() 2475 * @evt - Event 2479 csio_hws_quiescing(struct csio_hw *hw, enum csio_hw_ev evt) csio_hws_quiescing() argument 2482 hw->cur_evt = evt; csio_hws_quiescing() 2483 CSIO_INC_STATS(hw, n_evt_sm[evt]); csio_hws_quiescing() 2485 switch (evt) { csio_hws_quiescing() 2535 * @evt - Event 2539 csio_hws_quiesced(struct csio_hw *hw, enum csio_hw_ev evt) csio_hws_quiesced() argument 2542 hw->cur_evt = evt; csio_hws_quiesced() 2543 CSIO_INC_STATS(hw, n_evt_sm[evt]); csio_hws_quiesced() 2545 switch (evt) { csio_hws_quiesced() 2560 * @evt - Event 2564 csio_hws_resetting(struct csio_hw *hw, enum csio_hw_ev evt) csio_hws_resetting() argument 2567 hw->cur_evt = evt; csio_hws_resetting() 2568 CSIO_INC_STATS(hw, n_evt_sm[evt]); csio_hws_resetting() 2570 switch (evt) { csio_hws_resetting() 2586 * @evt - Event 2590 csio_hws_removing(struct csio_hw *hw, enum csio_hw_ev evt) csio_hws_removing() argument 2593 hw->cur_evt = evt; csio_hws_removing() 2594 CSIO_INC_STATS(hw, n_evt_sm[evt]); csio_hws_removing() 2596 switch (evt) { csio_hws_removing() 2621 * @evt - Event 2625 csio_hws_pcierr(struct csio_hw *hw, enum csio_hw_ev evt) csio_hws_pcierr() argument 2628 hw->cur_evt = evt; csio_hws_pcierr() 2629 CSIO_INC_STATS(hw, n_evt_sm[evt]); csio_hws_pcierr() 2631 switch (evt) { csio_hws_pcierr() 3330 csio_err(hw, "Failed to alloc evt entry, msg type %d len %d\n", csio_enqueue_evt() 3373 csio_err(hw, "Failed to alloc evt entry, msg type %d len %d\n", csio_enqueue_evt_lock()
|
H A D | csio_hw.h | 78 #define CSIO_EXTRA_VECS 2 /* non-data + FW evt */ 172 struct list_head list; /* evt queue*/ 431 int fwevt_iq_idx; /* FW evt queue */ 435 struct list_head evt_free_q; /* freelist of evt 438 struct list_head evt_active_q; /* active evt queue*/ 462 uint8_t cur_evt; /* current s/m evt */ 463 uint8_t prev_evt; /* Previous s/m evt */ 474 uint32_t fwevt_intr_idx; /* FW evt MSIX/interrupt
|
H A D | csio_scsi.c | 703 csio_scsis_uninit(struct csio_ioreq *req, enum csio_scsi_ev evt) csio_scsis_uninit() argument 708 switch (evt) { csio_scsis_uninit() 767 csio_dbg(hw, "Unhandled event:%d sent to req:%p\n", evt, req); csio_scsis_uninit() 773 csio_scsis_io_active(struct csio_ioreq *req, enum csio_scsi_ev evt) csio_scsis_io_active() argument 779 switch (evt) { csio_scsis_io_active() 839 csio_dbg(hw, "Unhandled event:%d sent to req:%p\n", evt, req); csio_scsis_io_active() 845 csio_scsis_tm_active(struct csio_ioreq *req, enum csio_scsi_ev evt) csio_scsis_tm_active() argument 850 switch (evt) { csio_scsis_tm_active() 882 csio_dbg(hw, "Unhandled event:%d sent to req:%p\n", evt, req); csio_scsis_tm_active() 888 csio_scsis_aborting(struct csio_ioreq *req, enum csio_scsi_ev evt) csio_scsis_aborting() argument 893 switch (evt) { csio_scsis_aborting() 979 csio_dbg(hw, "Unhandled event:%d sent to req:%p\n", evt, req); csio_scsis_aborting() 985 csio_scsis_closing(struct csio_ioreq *req, enum csio_scsi_ev evt) csio_scsis_closing() argument 990 switch (evt) { csio_scsis_closing() 1043 csio_dbg(hw, "Unhandled event:%d sent to req:%p\n", evt, req); csio_scsis_closing() 1049 csio_scsis_shost_cmpl_await(struct csio_ioreq *req, enum csio_scsi_ev evt) csio_scsis_shost_cmpl_await() argument 1051 switch (evt) { csio_scsis_shost_cmpl_await() 1077 evt, req); csio_scsis_shost_cmpl_await()
|
H A D | csio_init.c | 385 /* Allocate the FW evt queue */ csio_config_queues() 392 csio_err(hw, "FW evt queue creation failed\n"); csio_config_queues() 1063 /* Post PCI error detected evt to HW s/m csio_pci_error_detected() 1064 * HW s/m handles this evt by quiescing IOs, unregisters rports csio_pci_error_detected()
|
/linux-4.1.27/arch/mips/ralink/ |
H A D | cevt-rt3352.c | 40 struct clock_event_device *evt); 43 struct clock_event_device *evt) systick_next_event() 48 sdev = container_of(evt, struct systick_device, dev); systick_next_event() 91 struct clock_event_device *evt) systick_set_clock_mode() 95 sdev = container_of(evt, struct systick_device, dev); systick_set_clock_mode() 42 systick_next_event(unsigned long delta, struct clock_event_device *evt) systick_next_event() argument 90 systick_set_clock_mode(enum clock_event_mode mode, struct clock_event_device *evt) systick_set_clock_mode() argument
|
/linux-4.1.27/arch/arm/mach-w90x900/ |
H A D | time.c | 78 static int nuc900_clockevent_setnextevent(unsigned long evt, nuc900_clockevent_setnextevent() argument 83 __raw_writel(evt, REG_TICR0); nuc900_clockevent_setnextevent() 104 struct clock_event_device *evt = &nuc900_clockevent_device; nuc900_timer0_interrupt() local 108 evt->event_handler(evt); nuc900_timer0_interrupt()
|
/linux-4.1.27/arch/arm/mach-gemini/ |
H A D | time.c | 41 struct clock_event_device *evt) gemini_timer_set_next_event() 63 struct clock_event_device *evt) gemini_timer_set_mode() 112 struct clock_event_device *evt = &gemini_clockevent; gemini_timer_interrupt() local 114 evt->event_handler(evt); gemini_timer_interrupt() 40 gemini_timer_set_next_event(unsigned long cycles, struct clock_event_device *evt) gemini_timer_set_next_event() argument 62 gemini_timer_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) gemini_timer_set_mode() argument
|
/linux-4.1.27/arch/arm/mach-netx/ |
H A D | time.c | 73 static int netx_set_next_event(unsigned long evt, netx_set_next_event() argument 76 writel(0 - evt, NETX_GPIO_COUNTER_CURRENT(TIMER_CLOCKEVENT)); netx_set_next_event() 93 struct clock_event_device *evt = &netx_clockevent; netx_timer_interrupt() local 98 evt->event_handler(evt); netx_timer_interrupt()
|
/linux-4.1.27/drivers/staging/i2o/ |
H A D | driver.c | 213 struct i2o_event *evt; i2o_driver_dispatch() local 225 evt = kzalloc(size * 4 + sizeof(*evt), GFP_ATOMIC); i2o_driver_dispatch() 226 if (!evt) i2o_driver_dispatch() 229 evt->size = size; i2o_driver_dispatch() 230 evt->tcntxt = le32_to_cpu(msg->u.s.tcntxt); i2o_driver_dispatch() 231 evt->event_indicator = le32_to_cpu(msg->body[0]); i2o_driver_dispatch() 232 memcpy(&evt->data, &msg->body[1], size * 4); i2o_driver_dispatch() 236 evt->i2o_dev = dev; i2o_driver_dispatch() 240 INIT_WORK(&evt->work, drv->event); i2o_driver_dispatch() 241 queue_work(drv->event_queue, &evt->work); i2o_driver_dispatch()
|
H A D | exec-osm.c | 526 struct i2o_event *evt = container_of(work, struct i2o_event, work); i2o_exec_event() local 528 if (likely(evt->i2o_dev)) i2o_exec_event() 530 evt->i2o_dev->lct_data.tid); i2o_exec_event() 531 kfree(evt); i2o_exec_event()
|
H A D | i2o_block.c | 507 struct i2o_event *evt = container_of(work, struct i2o_event, work); i2o_block_event() local 509 kfree(evt); i2o_block_event()
|
/linux-4.1.27/arch/tile/kernel/ |
H A D | time.c | 131 struct clock_event_device *evt) tile_timer_set_next_event() 144 struct clock_event_device *evt) tile_timer_set_mode() 165 struct clock_event_device *evt = this_cpu_ptr(&tile_timer); setup_tile_timer() local 168 clockevents_calc_mult_shift(evt, cycles_per_sec, TILE_MINSEC); setup_tile_timer() 169 evt->max_delta_ns = clockevent_delta2ns(MAX_TICK, evt); setup_tile_timer() 172 evt->cpumask = cpumask_of(smp_processor_id()); setup_tile_timer() 178 clockevents_register_device(evt); setup_tile_timer() 185 struct clock_event_device *evt = this_cpu_ptr(&tile_timer); do_timer_interrupt() local 200 evt->event_handler(evt); do_timer_interrupt() 130 tile_timer_set_next_event(unsigned long ticks, struct clock_event_device *evt) tile_timer_set_next_event() argument 143 tile_timer_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) tile_timer_set_mode() argument
|
/linux-4.1.27/arch/nios2/kernel/ |
H A D | time.c | 49 to_nios2_clkevent(struct clock_event_device *evt) to_nios2_clkevent() argument 51 return container_of(evt, struct nios2_clockevent_dev, ced); to_nios2_clkevent() 155 struct clock_event_device *evt) nios2_timer_set_next_event() 157 struct nios2_clockevent_dev *nios2_ced = to_nios2_clkevent(evt); nios2_timer_set_next_event() 159 nios2_timer_config(&nios2_ced->timer, delta, evt->mode); nios2_timer_set_next_event() 165 struct clock_event_device *evt) nios2_timer_set_mode() 168 struct nios2_clockevent_dev *nios2_ced = to_nios2_clkevent(evt); nios2_timer_set_mode() 189 struct clock_event_device *evt = (struct clock_event_device *) dev_id; timer_interrupt() local 190 struct nios2_clockevent_dev *nios2_ced = to_nios2_clkevent(evt); timer_interrupt() 194 evt->event_handler(evt); timer_interrupt() 154 nios2_timer_set_next_event(unsigned long delta, struct clock_event_device *evt) nios2_timer_set_next_event() argument 164 nios2_timer_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) nios2_timer_set_mode() argument
|
/linux-4.1.27/arch/arm/common/ |
H A D | timer-sp.c | 120 struct clock_event_device *evt = dev_id; sp804_timer_interrupt() local 125 evt->event_handler(evt); sp804_timer_interrupt() 131 struct clock_event_device *evt) sp804_set_mode() 158 struct clock_event_device *evt) sp804_set_next_event() 185 struct clock_event_device *evt = &sp804_clockevent; __sp804_clockevents_init() local 202 evt->name = name; __sp804_clockevents_init() 203 evt->irq = irq; __sp804_clockevents_init() 204 evt->cpumask = cpu_possible_mask; __sp804_clockevents_init() 209 clockevents_config_and_register(evt, rate, 0xf, 0xffffffff); __sp804_clockevents_init() 130 sp804_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) sp804_set_mode() argument 157 sp804_set_next_event(unsigned long next, struct clock_event_device *evt) sp804_set_next_event() argument
|
/linux-4.1.27/arch/powerpc/platforms/powernv/ |
H A D | opal.c | 289 __be64 evt = 0; opal_notifier_enable() local 294 rc = opal_poll_events(&evt); opal_notifier_enable() 295 if (rc == OPAL_SUCCESS && evt) opal_notifier_enable() 296 opal_do_notifier(be64_to_cpu(evt)); opal_notifier_enable() 401 __be64 evt, len; opal_get_chars() local 405 opal_poll_events(&evt); opal_get_chars() 406 if ((be64_to_cpu(evt) & OPAL_EVENT_CONSOLE_INPUT) == 0) opal_get_chars() 421 __be64 evt; opal_put_chars() local 475 opal_poll_events(&evt); opal_put_chars() 477 (be64_to_cpu(evt) & OPAL_EVENT_CONSOLE_OUTPUT)); opal_put_chars() 484 struct machine_check_event *evt) opal_recover_mce() 487 uint64_t ea = get_mce_fault_addr(evt); opal_recover_mce() 492 } else if (evt->disposition == MCE_DISPOSITION_RECOVERED) { opal_recover_mce() 510 evt->severity == MCE_SEV_ERROR_SYNC) { opal_recover_mce() 523 struct machine_check_event evt; opal_machine_check() local 525 if (!get_mce_event(&evt, MCE_EVENT_RELEASE)) opal_machine_check() 529 if (evt.version != MCE_V1) { opal_machine_check() 531 evt.version); opal_machine_check() 534 machine_check_print_event_info(&evt); opal_machine_check() 536 if (opal_recover_mce(regs, &evt)) opal_machine_check() 563 __be64 evt = 0; opal_handle_hmi_exception() local 574 rc = opal_poll_events(&evt); opal_handle_hmi_exception() 575 if (rc == OPAL_SUCCESS && evt) opal_handle_hmi_exception() 576 opal_do_notifier(be64_to_cpu(evt)); opal_handle_hmi_exception() 483 opal_recover_mce(struct pt_regs *regs, struct machine_check_event *evt) opal_recover_mce() argument
|
/linux-4.1.27/arch/mips/kernel/ |
H A D | cevt-txx9.c | 89 struct clock_event_device *evt) txx9tmr_set_mode() 92 container_of(evt, struct txx9_clock_event_device, cd); txx9tmr_set_mode() 101 __raw_writel(((u64)(NSEC_PER_SEC / HZ) * evt->mult) >> txx9tmr_set_mode() 102 evt->shift, txx9tmr_set_mode() 121 struct clock_event_device *evt) txx9tmr_set_next_event() 124 container_of(evt, struct txx9_clock_event_device, cd); txx9tmr_set_next_event() 88 txx9tmr_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) txx9tmr_set_mode() argument 120 txx9tmr_set_next_event(unsigned long delta, struct clock_event_device *evt) txx9tmr_set_next_event() argument
|
H A D | cevt-ds1287.c | 57 struct clock_event_device *evt) ds1287_set_next_event() 63 struct clock_event_device *evt) ds1287_set_mode() 56 ds1287_set_next_event(unsigned long delta, struct clock_event_device *evt) ds1287_set_next_event() argument 62 ds1287_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) ds1287_set_mode() argument
|
H A D | cevt-gt641xx.c | 49 struct clock_event_device *evt) gt641xx_timer0_set_next_event() 68 struct clock_event_device *evt) gt641xx_timer0_set_mode() 48 gt641xx_timer0_set_next_event(unsigned long delta, struct clock_event_device *evt) gt641xx_timer0_set_next_event() argument 67 gt641xx_timer0_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) gt641xx_timer0_set_mode() argument
|
H A D | cevt-r4k.c | 19 struct clock_event_device *evt) mips_next_event() 32 struct clock_event_device *evt) mips_set_clock_mode() 18 mips_next_event(unsigned long delta, struct clock_event_device *evt) mips_next_event() argument 31 mips_set_clock_mode(enum clock_event_mode mode, struct clock_event_device *evt) mips_set_clock_mode() argument
|
H A D | cevt-bcm1480.c | 44 struct clock_event_device *evt) sibyte_set_mode() 43 sibyte_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) sibyte_set_mode() argument
|
H A D | cevt-sb1250.c | 42 struct clock_event_device *evt) sibyte_set_mode() 41 sibyte_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) sibyte_set_mode() argument
|
H A D | perf_event_mipsxx.c | 341 static void mipsxx_pmu_enable_event(struct hw_perf_event *evt, int idx) mipsxx_pmu_enable_event() argument 347 cpuc->saved_ctrl[idx] = M_PERFCTL_EVENT(evt->event_base & 0xff) | mipsxx_pmu_enable_event() 348 (evt->config_base & M_PERFCTL_CONFIG_MASK) | mipsxx_pmu_enable_event()
|
/linux-4.1.27/arch/arm/mach-imx/ |
H A D | epit.c | 97 static int epit_set_next_event(unsigned long evt, epit_set_next_event() argument 104 __raw_writel(tcmp - evt, timer_base + EPITCMPR); epit_set_next_event() 110 struct clock_event_device *evt) epit_set_mode() 163 struct clock_event_device *evt = &clockevent_epit; epit_timer_interrupt() local 167 evt->event_handler(evt); epit_timer_interrupt() 109 epit_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) epit_set_mode() argument
|
H A D | time.c | 153 static int mx1_2_set_next_event(unsigned long evt, mx1_2_set_next_event() argument 158 tcmp = __raw_readl(timer_base + MX1_2_TCN) + evt; mx1_2_set_next_event() 166 static int v2_set_next_event(unsigned long evt, v2_set_next_event() argument 171 tcmp = __raw_readl(timer_base + V2_TCN) + evt; v2_set_next_event() 175 return evt < 0x7fffffff && v2_set_next_event() 191 struct clock_event_device *evt) mxc_set_mode() 256 struct clock_event_device *evt = &clockevent_mxc; mxc_timer_interrupt() local 266 evt->event_handler(evt); mxc_timer_interrupt() 190 mxc_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) mxc_set_mode() argument
|
/linux-4.1.27/arch/arm64/include/asm/ |
H A D | pmu.h | 50 void (*enable)(struct hw_perf_event *evt, int idx); 51 void (*disable)(struct hw_perf_event *evt, int idx); 54 int (*set_event_filter)(struct hw_perf_event *evt,
|
/linux-4.1.27/drivers/usb/dwc3/ |
H A D | core.c | 122 * @evt: Pointer to event buffer to be freed 125 struct dwc3_event_buffer *evt) dwc3_free_one_event_buffer() 127 dma_free_coherent(dwc->dev, evt->length, evt->buf, evt->dma); dwc3_free_one_event_buffer() 141 struct dwc3_event_buffer *evt; dwc3_alloc_one_event_buffer() local 143 evt = devm_kzalloc(dwc->dev, sizeof(*evt), GFP_KERNEL); dwc3_alloc_one_event_buffer() 144 if (!evt) dwc3_alloc_one_event_buffer() 147 evt->dwc = dwc; dwc3_alloc_one_event_buffer() 148 evt->length = length; dwc3_alloc_one_event_buffer() 149 evt->buf = dma_alloc_coherent(dwc->dev, length, dwc3_alloc_one_event_buffer() 150 &evt->dma, GFP_KERNEL); dwc3_alloc_one_event_buffer() 151 if (!evt->buf) dwc3_alloc_one_event_buffer() 154 return evt; dwc3_alloc_one_event_buffer() 163 struct dwc3_event_buffer *evt; dwc3_free_event_buffers() local 167 evt = dwc->ev_buffs[i]; dwc3_free_event_buffers() 168 if (evt) dwc3_free_event_buffers() 169 dwc3_free_one_event_buffer(dwc, evt); dwc3_free_event_buffers() 195 struct dwc3_event_buffer *evt; dwc3_alloc_event_buffers() local 197 evt = dwc3_alloc_one_event_buffer(dwc, length); dwc3_alloc_event_buffers() 198 if (IS_ERR(evt)) { dwc3_alloc_event_buffers() 200 return PTR_ERR(evt); dwc3_alloc_event_buffers() 202 dwc->ev_buffs[i] = evt; dwc3_alloc_event_buffers() 216 struct dwc3_event_buffer *evt; dwc3_event_buffers_setup() local 220 evt = dwc->ev_buffs[n]; dwc3_event_buffers_setup() 222 evt->buf, (unsigned long long) evt->dma, dwc3_event_buffers_setup() 223 evt->length); dwc3_event_buffers_setup() 225 evt->lpos = 0; dwc3_event_buffers_setup() 228 lower_32_bits(evt->dma)); dwc3_event_buffers_setup() 230 upper_32_bits(evt->dma)); dwc3_event_buffers_setup() 232 DWC3_GEVNTSIZ_SIZE(evt->length)); dwc3_event_buffers_setup() 241 struct dwc3_event_buffer *evt; dwc3_event_buffers_cleanup() local 245 evt = dwc->ev_buffs[n]; dwc3_event_buffers_cleanup() 247 evt->lpos = 0; dwc3_event_buffers_cleanup() 124 dwc3_free_one_event_buffer(struct dwc3 *dwc, struct dwc3_event_buffer *evt) dwc3_free_one_event_buffer() argument
|
H A D | gadget.c | 2602 struct dwc3_event_buffer *evt; dwc3_process_event_buf() local 2607 evt = dwc->ev_buffs[buf]; dwc3_process_event_buf() 2608 left = evt->count; dwc3_process_event_buf() 2610 if (!(evt->flags & DWC3_EVENT_PENDING)) dwc3_process_event_buf() 2616 event.raw = *(u32 *) (evt->buf + evt->lpos); dwc3_process_event_buf() 2629 evt->lpos = (evt->lpos + 4) % DWC3_EVENT_BUFFERS_SIZE; dwc3_process_event_buf() 2635 evt->count = 0; dwc3_process_event_buf() 2636 evt->flags &= ~DWC3_EVENT_PENDING; dwc3_process_event_buf() 2666 struct dwc3_event_buffer *evt; dwc3_check_event_buf() local 2670 evt = dwc->ev_buffs[buf]; dwc3_check_event_buf() 2677 evt->count = count; dwc3_check_event_buf() 2678 evt->flags |= DWC3_EVENT_PENDING; dwc3_check_event_buf()
|
/linux-4.1.27/include/linux/ |
H A D | sh_intc.h | 16 #define evt2irq(evt) (((evt) >> 5) - 16) 19 #define evt2irq(evt) (evt)
|
H A D | clockchips.h | 108 int (*set_next_event)(unsigned long evt, struct clock_event_device *); 169 extern u64 clockevent_delta2ns(unsigned long latch, struct clock_event_device *evt);
|
H A D | leds.h | 339 extern void ledtrig_cpu(enum cpu_led_event evt); 341 static inline void ledtrig_cpu(enum cpu_led_event evt) ledtrig_cpu() argument
|
/linux-4.1.27/arch/arc/kernel/ |
H A D | time.c | 199 * Note that generic IRQ core could have passed @evt for @dev_id if timer_irq_handler() 202 struct clock_event_device *evt = this_cpu_ptr(&arc_clockevent_device); timer_irq_handler() local 203 int irq_reenable = evt->mode == CLOCK_EVT_MODE_PERIODIC; timer_irq_handler() 212 evt->event_handler(evt); timer_irq_handler() 222 struct clock_event_device *evt = this_cpu_ptr(&arc_clockevent_device); arc_local_timer_setup() local 225 evt->cpumask = cpumask_of(cpu); arc_local_timer_setup() 226 clockevents_config_and_register(evt, arc_get_core_freq(), arc_local_timer_setup() 231 "Timer0 (per-cpu-tick)", evt); arc_local_timer_setup()
|
/linux-4.1.27/drivers/scsi/lpfc/ |
H A D | lpfc_nportdisc.c | 813 void *arg, uint32_t evt) lpfc_disc_illegal() 823 (evt == NLP_EVT_CMPL_REG_LOGIN) && lpfc_disc_illegal() 832 ndlp->nlp_DID, evt, ndlp->nlp_state, ndlp->nlp_rpi, lpfc_disc_illegal() 839 void *arg, uint32_t evt) lpfc_cmpl_plogi_illegal() 850 ndlp->nlp_DID, evt, ndlp->nlp_state, ndlp->nlp_rpi, lpfc_cmpl_plogi_illegal() 860 void *arg, uint32_t evt) lpfc_rcv_plogi_unused_node() 874 void *arg, uint32_t evt) lpfc_rcv_els_unused_node() 882 void *arg, uint32_t evt) lpfc_rcv_logo_unused_node() 897 void *arg, uint32_t evt) lpfc_cmpl_logo_unused_node() 904 void *arg, uint32_t evt) lpfc_device_rm_unused_node() 912 void *arg, uint32_t evt) lpfc_device_recov_unused_node() 919 void *arg, uint32_t evt) lpfc_rcv_plogi_plogi_issue() 970 void *arg, uint32_t evt) lpfc_rcv_prli_plogi_issue() 984 void *arg, uint32_t evt) lpfc_rcv_logo_plogi_issue() 997 void *arg, uint32_t evt) lpfc_rcv_els_plogi_issue() 1006 if (evt == NLP_EVT_RCV_LOGO) { lpfc_rcv_els_plogi_issue() 1028 uint32_t evt) lpfc_cmpl_plogi_plogi_issue() 1175 void *arg, uint32_t evt) lpfc_cmpl_logo_plogi_issue() 1182 struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_cmpl_reglogin_plogi_issue() 1201 void *arg, uint32_t evt) lpfc_device_rm_plogi_issue() 1223 uint32_t evt) lpfc_device_recov_plogi_issue() 1248 void *arg, uint32_t evt) lpfc_rcv_plogi_adisc_issue() 1278 void *arg, uint32_t evt) lpfc_rcv_prli_adisc_issue() 1288 void *arg, uint32_t evt) lpfc_rcv_logo_adisc_issue() 1305 void *arg, uint32_t evt) lpfc_rcv_padisc_adisc_issue() 1317 void *arg, uint32_t evt) lpfc_rcv_prlo_adisc_issue() 1331 void *arg, uint32_t evt) lpfc_cmpl_adisc_adisc_issue() 1387 void *arg, uint32_t evt) lpfc_device_rm_adisc_issue() 1409 uint32_t evt) lpfc_device_recov_adisc_issue() 1436 uint32_t evt) lpfc_rcv_plogi_reglogin_issue() 1448 uint32_t evt) lpfc_rcv_prli_reglogin_issue() 1460 uint32_t evt) lpfc_rcv_logo_reglogin_issue() 1505 uint32_t evt) lpfc_rcv_padisc_reglogin_issue() 1517 uint32_t evt) lpfc_rcv_prlo_reglogin_issue() 1530 uint32_t evt) lpfc_cmpl_reglogin_reglogin_issue() 1591 uint32_t evt) lpfc_device_rm_reglogin_issue() 1610 uint32_t evt) lpfc_device_recov_reglogin_issue() 1632 void *arg, uint32_t evt) lpfc_rcv_plogi_prli_issue() 1644 void *arg, uint32_t evt) lpfc_rcv_prli_prli_issue() 1654 void *arg, uint32_t evt) lpfc_rcv_logo_prli_issue() 1667 void *arg, uint32_t evt) lpfc_rcv_padisc_prli_issue() 1682 void *arg, uint32_t evt) lpfc_rcv_prlo_prli_issue() 1692 void *arg, uint32_t evt) lpfc_cmpl_prli_prli_issue() 1760 * \param evt 1773 void *arg, uint32_t evt) lpfc_device_rm_prli_issue() 1799 * \param evt 1812 uint32_t evt) lpfc_device_recov_prli_issue() 1837 void *arg, uint32_t evt) lpfc_rcv_plogi_logo_issue() 1851 void *arg, uint32_t evt) lpfc_rcv_prli_logo_issue() 1865 void *arg, uint32_t evt) lpfc_rcv_logo_logo_issue() 1879 void *arg, uint32_t evt) lpfc_rcv_padisc_logo_issue() 1893 void *arg, uint32_t evt) lpfc_rcv_prlo_logo_issue() 1907 void *arg, uint32_t evt) lpfc_cmpl_logo_logo_issue() 1922 void *arg, uint32_t evt) lpfc_device_rm_logo_issue() 1936 void *arg, uint32_t evt) lpfc_device_recov_logo_issue() 1948 void *arg, uint32_t evt) lpfc_rcv_plogi_unmap_node() 1958 void *arg, uint32_t evt) lpfc_rcv_prli_unmap_node() 1969 void *arg, uint32_t evt) lpfc_rcv_logo_unmap_node() 1979 void *arg, uint32_t evt) lpfc_rcv_padisc_unmap_node() 1989 void *arg, uint32_t evt) lpfc_rcv_prlo_unmap_node() 2001 uint32_t evt) lpfc_device_recov_unmap_node() 2017 void *arg, uint32_t evt) lpfc_rcv_plogi_mapped_node() 2027 void *arg, uint32_t evt) lpfc_rcv_prli_mapped_node() 2037 void *arg, uint32_t evt) lpfc_rcv_logo_mapped_node() 2048 void *arg, uint32_t evt) lpfc_rcv_padisc_mapped_node() 2058 void *arg, uint32_t evt) lpfc_rcv_prlo_mapped_node() 2076 uint32_t evt) lpfc_device_recov_mapped_node() 2091 void *arg, uint32_t evt) lpfc_rcv_plogi_npr_node() 2117 void *arg, uint32_t evt) lpfc_rcv_prli_npr_node() 2147 void *arg, uint32_t evt) lpfc_rcv_logo_npr_node() 2157 void *arg, uint32_t evt) lpfc_rcv_padisc_npr_node() 2185 void *arg, uint32_t evt) lpfc_rcv_prlo_npr_node() 2214 void *arg, uint32_t evt) lpfc_cmpl_plogi_npr_node() 2235 void *arg, uint32_t evt) lpfc_cmpl_prli_npr_node() 2253 void *arg, uint32_t evt) lpfc_cmpl_logo_npr_node() 2269 void *arg, uint32_t evt) lpfc_cmpl_adisc_npr_node() 2288 void *arg, uint32_t evt) lpfc_cmpl_reglogin_npr_node() 2309 void *arg, uint32_t evt) lpfc_device_rm_npr_node() 2325 void *arg, uint32_t evt) lpfc_device_recov_npr_node() 2531 void *arg, uint32_t evt) lpfc_disc_state_machine() 2543 /* DSM in event <evt> on NPort <nlp_DID> in state <cur_state> */ lpfc_disc_state_machine() 2547 evt, ndlp->nlp_DID, cur_state, ndlp->nlp_flag); lpfc_disc_state_machine() 2550 "DSM in: evt:%d ste:%d did:x%x", lpfc_disc_state_machine() 2551 evt, cur_state, ndlp->nlp_DID); lpfc_disc_state_machine() 2553 func = lpfc_disc_action[(cur_state * NLP_EVT_MAX_EVENT) + evt]; lpfc_disc_state_machine() 2554 rc = (func) (vport, ndlp, arg, evt); lpfc_disc_state_machine() 812 lpfc_disc_illegal(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_disc_illegal() argument 838 lpfc_cmpl_plogi_illegal(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_cmpl_plogi_illegal() argument 859 lpfc_rcv_plogi_unused_node(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_plogi_unused_node() argument 873 lpfc_rcv_els_unused_node(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_els_unused_node() argument 881 lpfc_rcv_logo_unused_node(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_logo_unused_node() argument 896 lpfc_cmpl_logo_unused_node(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_cmpl_logo_unused_node() argument 903 lpfc_device_rm_unused_node(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_device_rm_unused_node() argument 910 lpfc_device_recov_unused_node(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_device_recov_unused_node() argument 918 lpfc_rcv_plogi_plogi_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_plogi_plogi_issue() argument 969 lpfc_rcv_prli_plogi_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_prli_plogi_issue() argument 983 lpfc_rcv_logo_plogi_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_logo_plogi_issue() argument 996 lpfc_rcv_els_plogi_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_els_plogi_issue() argument 1025 lpfc_cmpl_plogi_plogi_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_cmpl_plogi_plogi_issue() argument 1174 lpfc_cmpl_logo_plogi_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_cmpl_logo_plogi_issue() argument 1181 lpfc_cmpl_reglogin_plogi_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_cmpl_reglogin_plogi_issue() argument 1200 lpfc_device_rm_plogi_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_device_rm_plogi_issue() argument 1220 lpfc_device_recov_plogi_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_device_recov_plogi_issue() argument 1247 lpfc_rcv_plogi_adisc_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_plogi_adisc_issue() argument 1277 lpfc_rcv_prli_adisc_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_prli_adisc_issue() argument 1287 lpfc_rcv_logo_adisc_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_logo_adisc_issue() argument 1303 lpfc_rcv_padisc_adisc_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_padisc_adisc_issue() argument 1316 lpfc_rcv_prlo_adisc_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_prlo_adisc_issue() argument 1329 lpfc_cmpl_adisc_adisc_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_cmpl_adisc_adisc_issue() argument 1386 lpfc_device_rm_adisc_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_device_rm_adisc_issue() argument 1406 lpfc_device_recov_adisc_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_device_recov_adisc_issue() argument 1433 lpfc_rcv_plogi_reglogin_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_plogi_reglogin_issue() argument 1445 lpfc_rcv_prli_reglogin_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_prli_reglogin_issue() argument 1457 lpfc_rcv_logo_reglogin_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_logo_reglogin_issue() argument 1502 lpfc_rcv_padisc_reglogin_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_padisc_reglogin_issue() argument 1514 lpfc_rcv_prlo_reglogin_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_prlo_reglogin_issue() argument 1527 lpfc_cmpl_reglogin_reglogin_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_cmpl_reglogin_reglogin_issue() argument 1588 lpfc_device_rm_reglogin_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_device_rm_reglogin_issue() argument 1607 lpfc_device_recov_reglogin_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_device_recov_reglogin_issue() argument 1631 lpfc_rcv_plogi_prli_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_plogi_prli_issue() argument 1643 lpfc_rcv_prli_prli_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_prli_prli_issue() argument 1653 lpfc_rcv_logo_prli_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_logo_prli_issue() argument 1666 lpfc_rcv_padisc_prli_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_padisc_prli_issue() argument 1681 lpfc_rcv_prlo_prli_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_prlo_prli_issue() argument 1691 lpfc_cmpl_prli_prli_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_cmpl_prli_prli_issue() argument 1772 lpfc_device_rm_prli_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_device_rm_prli_issue() argument 1809 lpfc_device_recov_prli_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_device_recov_prli_issue() argument 1836 lpfc_rcv_plogi_logo_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_plogi_logo_issue() argument 1850 lpfc_rcv_prli_logo_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_prli_logo_issue() argument 1864 lpfc_rcv_logo_logo_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_logo_logo_issue() argument 1878 lpfc_rcv_padisc_logo_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_padisc_logo_issue() argument 1892 lpfc_rcv_prlo_logo_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_prlo_logo_issue() argument 1906 lpfc_cmpl_logo_logo_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_cmpl_logo_logo_issue() argument 1921 lpfc_device_rm_logo_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_device_rm_logo_issue() argument 1934 lpfc_device_recov_logo_issue(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_device_recov_logo_issue() argument 1947 lpfc_rcv_plogi_unmap_node(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_plogi_unmap_node() argument 1957 lpfc_rcv_prli_unmap_node(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_prli_unmap_node() argument 1968 lpfc_rcv_logo_unmap_node(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_logo_unmap_node() argument 1978 lpfc_rcv_padisc_unmap_node(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_padisc_unmap_node() argument 1988 lpfc_rcv_prlo_unmap_node(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_prlo_unmap_node() argument 1998 lpfc_device_recov_unmap_node(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_device_recov_unmap_node() argument 2016 lpfc_rcv_plogi_mapped_node(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_plogi_mapped_node() argument 2026 lpfc_rcv_prli_mapped_node(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_prli_mapped_node() argument 2036 lpfc_rcv_logo_mapped_node(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_logo_mapped_node() argument 2046 lpfc_rcv_padisc_mapped_node(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_padisc_mapped_node() argument 2057 lpfc_rcv_prlo_mapped_node(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_prlo_mapped_node() argument 2073 lpfc_device_recov_mapped_node(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_device_recov_mapped_node() argument 2090 lpfc_rcv_plogi_npr_node(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_plogi_npr_node() argument 2116 lpfc_rcv_prli_npr_node(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_prli_npr_node() argument 2146 lpfc_rcv_logo_npr_node(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_logo_npr_node() argument 2156 lpfc_rcv_padisc_npr_node(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_padisc_npr_node() argument 2184 lpfc_rcv_prlo_npr_node(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_rcv_prlo_npr_node() argument 2213 lpfc_cmpl_plogi_npr_node(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_cmpl_plogi_npr_node() argument 2234 lpfc_cmpl_prli_npr_node(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_cmpl_prli_npr_node() argument 2252 lpfc_cmpl_logo_npr_node(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_cmpl_logo_npr_node() argument 2268 lpfc_cmpl_adisc_npr_node(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_cmpl_adisc_npr_node() argument 2286 lpfc_cmpl_reglogin_npr_node(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_cmpl_reglogin_npr_node() argument 2308 lpfc_device_rm_npr_node(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_device_rm_npr_node() argument 2324 lpfc_device_recov_npr_node(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_device_recov_npr_node() argument 2530 lpfc_disc_state_machine(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) lpfc_disc_state_machine() argument
|
H A D | lpfc_bsg.c | 102 struct lpfc_bsg_event *evt; member in union:bsg_job_data::__anon9070 790 struct lpfc_bsg_event *evt = container_of(kref, struct lpfc_bsg_event, lpfc_bsg_event_free() local 794 list_del(&evt->node); lpfc_bsg_event_free() 796 while (!list_empty(&evt->events_to_get)) { lpfc_bsg_event_free() 797 ed = list_entry(evt->events_to_get.next, typeof(*ed), node); lpfc_bsg_event_free() 803 while (!list_empty(&evt->events_to_see)) { lpfc_bsg_event_free() 804 ed = list_entry(evt->events_to_see.next, typeof(*ed), node); lpfc_bsg_event_free() 810 kfree(evt->dd_data); lpfc_bsg_event_free() 811 kfree(evt); lpfc_bsg_event_free() 816 * @evt: Pointer to an event structure. 819 lpfc_bsg_event_ref(struct lpfc_bsg_event *evt) lpfc_bsg_event_ref() argument 821 kref_get(&evt->kref); lpfc_bsg_event_ref() 826 * @evt: Pointer to an event structure. 829 lpfc_bsg_event_unref(struct lpfc_bsg_event *evt) lpfc_bsg_event_unref() argument 831 kref_put(&evt->kref, lpfc_bsg_event_free); lpfc_bsg_event_unref() 843 struct lpfc_bsg_event *evt = kzalloc(sizeof(*evt), GFP_KERNEL); lpfc_bsg_event_new() local 845 if (!evt) lpfc_bsg_event_new() 848 INIT_LIST_HEAD(&evt->events_to_get); lpfc_bsg_event_new() 849 INIT_LIST_HEAD(&evt->events_to_see); lpfc_bsg_event_new() 850 evt->type_mask = ev_mask; lpfc_bsg_event_new() 851 evt->req_id = ev_req_id; lpfc_bsg_event_new() 852 evt->reg_id = ev_reg_id; lpfc_bsg_event_new() 853 evt->wait_time_stamp = jiffies; lpfc_bsg_event_new() 854 evt->dd_data = NULL; lpfc_bsg_event_new() 855 init_waitqueue_head(&evt->wq); lpfc_bsg_event_new() 856 kref_init(&evt->kref); lpfc_bsg_event_new() 857 return evt; lpfc_bsg_event_new() 909 struct lpfc_bsg_event *evt; lpfc_bsg_ct_unsol_event() local 954 list_for_each_entry(evt, &phba->ct_ev_waiters, node) { lpfc_bsg_ct_unsol_event() 955 if (!(evt->type_mask & FC_REG_CT_EVENT) || lpfc_bsg_ct_unsol_event() 956 evt->req_id != evt_req_id) lpfc_bsg_ct_unsol_event() 959 lpfc_bsg_event_ref(evt); lpfc_bsg_ct_unsol_event() 964 lpfc_bsg_event_unref(evt); lpfc_bsg_ct_unsol_event() 991 lpfc_bsg_event_unref(evt); lpfc_bsg_ct_unsol_event() 1037 lpfc_bsg_event_unref(evt); lpfc_bsg_ct_unsol_event() 1111 list_add(&evt_dat->node, &evt->events_to_see); lpfc_bsg_ct_unsol_event() 1113 wake_up_interruptible(&evt->wq); lpfc_bsg_ct_unsol_event() 1114 lpfc_bsg_event_unref(evt); lpfc_bsg_ct_unsol_event() 1118 list_move(evt->events_to_see.prev, &evt->events_to_get); lpfc_bsg_ct_unsol_event() 1120 dd_data = (struct bsg_job_data *)evt->dd_data; lpfc_bsg_ct_unsol_event() 1123 lpfc_bsg_event_unref(evt); lpfc_bsg_ct_unsol_event() 1197 struct lpfc_bsg_event *evt; lpfc_bsg_hba_set_event() local 1217 list_for_each_entry(evt, &phba->ct_ev_waiters, node) { lpfc_bsg_hba_set_event() 1218 if (evt->reg_id == event_req->ev_reg_id) { lpfc_bsg_hba_set_event() 1219 lpfc_bsg_event_ref(evt); lpfc_bsg_hba_set_event() 1220 evt->wait_time_stamp = jiffies; lpfc_bsg_hba_set_event() 1221 dd_data = (struct bsg_job_data *)evt->dd_data; lpfc_bsg_hba_set_event() 1227 if (&evt->node == &phba->ct_ev_waiters) { lpfc_bsg_hba_set_event() 1236 evt = lpfc_bsg_event_new(ev_mask, event_req->ev_reg_id, lpfc_bsg_hba_set_event() 1238 if (!evt) { lpfc_bsg_hba_set_event() 1247 dd_data->context_un.evt = evt; lpfc_bsg_hba_set_event() 1248 evt->dd_data = (void *)dd_data; lpfc_bsg_hba_set_event() 1250 list_add(&evt->node, &phba->ct_ev_waiters); lpfc_bsg_hba_set_event() 1251 lpfc_bsg_event_ref(evt); lpfc_bsg_hba_set_event() 1252 evt->wait_time_stamp = jiffies; lpfc_bsg_hba_set_event() 1257 evt->waiting = 1; lpfc_bsg_hba_set_event() 1282 struct lpfc_bsg_event *evt, *evt_next; lpfc_bsg_hba_get_event() local 1302 list_for_each_entry_safe(evt, evt_next, &phba->ct_ev_waiters, node) { lpfc_bsg_hba_get_event() 1303 if (evt->reg_id == event_req->ev_reg_id) { lpfc_bsg_hba_get_event() 1304 if (list_empty(&evt->events_to_get)) lpfc_bsg_hba_get_event() 1306 lpfc_bsg_event_ref(evt); lpfc_bsg_hba_get_event() 1307 evt->wait_time_stamp = jiffies; lpfc_bsg_hba_get_event() 1308 evt_dat = list_entry(evt->events_to_get.prev, lpfc_bsg_hba_get_event() 1349 lpfc_bsg_event_unref(evt); lpfc_bsg_hba_get_event() 2537 struct lpfc_bsg_event *evt; lpfcdiag_loop_get_xri() local 2550 evt = lpfc_bsg_event_new(FC_REG_CT_EVENT, current->pid, lpfcdiag_loop_get_xri() 2552 if (!evt) lpfcdiag_loop_get_xri() 2556 list_add(&evt->node, &phba->ct_ev_waiters); lpfcdiag_loop_get_xri() 2557 lpfc_bsg_event_ref(evt); lpfcdiag_loop_get_xri() 2633 evt->waiting = 1; lpfcdiag_loop_get_xri() 2634 evt->wait_time_stamp = jiffies; lpfcdiag_loop_get_xri() 2636 evt->wq, !list_empty(&evt->events_to_see), lpfcdiag_loop_get_xri() 2639 if (list_empty(&evt->events_to_see)) lpfcdiag_loop_get_xri() 2643 list_move(evt->events_to_see.prev, &evt->events_to_get); lpfcdiag_loop_get_xri() 2645 *rxxri = (list_entry(evt->events_to_get.prev, lpfcdiag_loop_get_xri() 2649 evt->waiting = 0; lpfcdiag_loop_get_xri() 2653 lpfc_bsg_event_unref(evt); /* release ref */ lpfcdiag_loop_get_xri() 2654 lpfc_bsg_event_unref(evt); /* delete */ lpfcdiag_loop_get_xri() 2992 struct lpfc_bsg_event *evt; lpfc_bsg_diag_loopback_run() local 3101 evt = lpfc_bsg_event_new(FC_REG_CT_EVENT, current->pid, lpfc_bsg_diag_loopback_run() 3103 if (!evt) { lpfc_bsg_diag_loopback_run() 3110 list_add(&evt->node, &phba->ct_ev_waiters); lpfc_bsg_diag_loopback_run() 3111 lpfc_bsg_event_ref(evt); lpfc_bsg_diag_loopback_run() 3214 evt->waiting = 1; lpfc_bsg_diag_loopback_run() 3216 evt->wq, !list_empty(&evt->events_to_see), lpfc_bsg_diag_loopback_run() 3219 evt->waiting = 0; lpfc_bsg_diag_loopback_run() 3220 if (list_empty(&evt->events_to_see)) { lpfc_bsg_diag_loopback_run() 3227 list_move(evt->events_to_see.prev, &evt->events_to_get); lpfc_bsg_diag_loopback_run() 3228 evdat = list_entry(evt->events_to_get.prev, lpfc_bsg_diag_loopback_run() 3257 lpfc_bsg_event_unref(evt); /* release ref */ lpfc_bsg_diag_loopback_run() 3258 lpfc_bsg_event_unref(evt); /* delete */ lpfc_bsg_diag_loopback_run()
|
H A D | lpfc_disc.h | 49 enum lpfc_work_type evt; member in struct:lpfc_work_evt
|
H A D | lpfc_hbadisc.c | 192 evtp->evt = LPFC_EVT_DEV_LOSS; lpfc_dev_loss_tmo_callbk() 436 ret->work_evt.evt = LPFC_EVT_FASTPATH_MGMT_EVT; lpfc_alloc_fast_evt() 444 * @evt: Event object which need to be freed. 451 struct lpfc_fast_path_event *evt) { lpfc_free_fast_evt() 454 kfree(evt); lpfc_free_fast_evt() 547 switch (evtp->evt) { lpfc_work_list_done() 551 free_evt = 0; /* evt is part of ndlp */ lpfc_work_list_done() 796 uint32_t evt) lpfc_workq_post_event() 811 evtp->evt = evt; lpfc_workq_post_event() 450 lpfc_free_fast_evt(struct lpfc_hba *phba, struct lpfc_fast_path_event *evt) lpfc_free_fast_evt() argument 795 lpfc_workq_post_event(struct lpfc_hba *phba, void *arg1, void *arg2, uint32_t evt) lpfc_workq_post_event() argument
|
H A D | lpfc_debugfs.h | 416 * lpfc_debug_dump_hba_eq - dump all entries from a fcp work queue's evt queue
|
/linux-4.1.27/arch/um/kernel/ |
H A D | time.c | 26 struct clock_event_device *evt) itimer_set_mode() 45 struct clock_event_device *evt) itimer_next_event() 25 itimer_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) itimer_set_mode() argument 44 itimer_next_event(unsigned long delta, struct clock_event_device *evt) itimer_next_event() argument
|
/linux-4.1.27/arch/c6x/kernel/ |
H A D | soc.c | 26 void soc_assert_event(unsigned int evt) soc_assert_event() argument 29 soc_ops.assert_event(evt); soc_assert_event()
|
H A D | traps.c | 302 int evt; process_eexcept() local 306 while ((evt = soc_get_exception()) >= 0) process_eexcept() 307 do_trap(&eexcept_table[evt], regs); process_eexcept()
|
/linux-4.1.27/arch/arm/mach-ks8695/ |
H A D | time.c | 58 struct clock_event_device *evt) ks8695_set_mode() 82 struct clock_event_device *evt) ks8695_set_next_event() 117 struct clock_event_device *evt = &clockevent_ks8695; ks8695_timer_interrupt() local 119 evt->event_handler(evt); ks8695_timer_interrupt() 57 ks8695_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) ks8695_set_mode() argument 81 ks8695_set_next_event(unsigned long cycles, struct clock_event_device *evt) ks8695_set_next_event() argument
|
/linux-4.1.27/arch/arm/mach-omap1/ |
H A D | time.c | 121 struct clock_event_device *evt) omap_mpu_set_next_event() 128 struct clock_event_device *evt) omap_mpu_set_mode() 154 struct clock_event_device *evt = &clockevent_mpu_timer1; omap_mpu_timer1_interrupt() local 156 evt->event_handler(evt); omap_mpu_timer1_interrupt() 120 omap_mpu_set_next_event(unsigned long cycles, struct clock_event_device *evt) omap_mpu_set_next_event() argument 127 omap_mpu_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) omap_mpu_set_mode() argument
|
H A D | timer32k.c | 118 struct clock_event_device *evt) omap_32k_timer_set_mode() 144 struct clock_event_device *evt = &clockevent_32k_timer; omap_32k_timer_interrupt() local 147 evt->event_handler(evt); omap_32k_timer_interrupt() 117 omap_32k_timer_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) omap_32k_timer_set_mode() argument
|
/linux-4.1.27/arch/x86/kernel/ |
H A D | hpet.c | 47 struct clock_event_device evt; member in struct:hpet_dev 57 return container_of(evtdev, struct hpet_dev, evt); EVT_TO_HPET_DEV() 229 struct clock_event_device *evt); hpet_reserve_platform_timers() 231 struct clock_event_device *evt); hpet_reserve_platform_timers() 311 struct clock_event_device *evt, int timer) hpet_set_mode() 319 delta = ((uint64_t)(NSEC_PER_SEC/HZ)) * evt->mult; hpet_set_mode() 320 delta >>= evt->shift; hpet_set_mode() 359 struct hpet_dev *hdev = EVT_TO_HPET_DEV(evt); hpet_set_mode() 371 struct clock_event_device *evt, int timer) hpet_next_event() 408 struct clock_event_device *evt) hpet_legacy_set_mode() 410 hpet_set_mode(mode, evt, 0); hpet_legacy_set_mode() 414 struct clock_event_device *evt) hpet_legacy_next_event() 416 return hpet_next_event(delta, evt, 0); hpet_legacy_next_event() 463 struct clock_event_device *evt) hpet_msi_set_mode() 465 struct hpet_dev *hdev = EVT_TO_HPET_DEV(evt); hpet_msi_set_mode() 466 hpet_set_mode(mode, evt, hdev->num); hpet_msi_set_mode() 470 struct clock_event_device *evt) hpet_msi_next_event() 472 struct hpet_dev *hdev = EVT_TO_HPET_DEV(evt); hpet_msi_next_event() 473 return hpet_next_event(delta, evt, hdev->num); hpet_msi_next_event() 504 struct clock_event_device *hevt = &dev->evt; hpet_interrupt_handler() 537 struct clock_event_device *evt = &hdev->evt; init_one_hpet_msi_clockevent() local 548 evt->name = hdev->name; init_one_hpet_msi_clockevent() 550 evt->irq = hdev->irq; init_one_hpet_msi_clockevent() 552 evt->rating = 110; init_one_hpet_msi_clockevent() 553 evt->features = CLOCK_EVT_FEAT_ONESHOT; init_one_hpet_msi_clockevent() 555 evt->features |= CLOCK_EVT_FEAT_PERIODIC; init_one_hpet_msi_clockevent() 557 evt->set_mode = hpet_msi_set_mode; init_one_hpet_msi_clockevent() 558 evt->set_next_event = hpet_msi_next_event; init_one_hpet_msi_clockevent() 559 evt->cpumask = cpumask_of(hdev->cpu); init_one_hpet_msi_clockevent() 561 clockevents_config_and_register(evt, hpet_freq, HPET_MIN_PROG_DELTA, init_one_hpet_msi_clockevent() 310 hpet_set_mode(enum clock_event_mode mode, struct clock_event_device *evt, int timer) hpet_set_mode() argument 370 hpet_next_event(unsigned long delta, struct clock_event_device *evt, int timer) hpet_next_event() argument 407 hpet_legacy_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) hpet_legacy_set_mode() argument 413 hpet_legacy_next_event(unsigned long delta, struct clock_event_device *evt) hpet_legacy_next_event() argument 462 hpet_msi_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) hpet_msi_set_mode() argument 469 hpet_msi_next_event(unsigned long delta, struct clock_event_device *evt) hpet_msi_next_event() argument
|
/linux-4.1.27/net/tipc/ |
H A D | subscr.c | 73 msg_sect.iov_base = (void *)&sub->evt; subscr_send_event() 75 sub->evt.event = htohl(event, sub->swap); subscr_send_event() 76 sub->evt.found_lower = htohl(found_lower, sub->swap); subscr_send_event() 77 sub->evt.found_upper = htohl(found_upper, sub->swap); subscr_send_event() 78 sub->evt.port.ref = htohl(port_ref, sub->swap); subscr_send_event() 79 sub->evt.port.node = htohl(node, sub->swap); subscr_send_event() 142 subscr_send_event(sub, sub->evt.s.seq.lower, sub->evt.s.seq.upper, subscr_timeout() 207 if (!memcmp(s, &sub->evt.s, sizeof(struct tipc_subscr))) { subscr_cancel() 279 memcpy(&sub->evt.s, s, sizeof(struct tipc_subscr)); subscr_subscribe()
|
H A D | subscr.h | 60 * @evt: template for events generated by subscription 72 struct tipc_event evt; member in struct:tipc_subscription
|
H A D | bearer.c | 510 * @evt: the type of event 516 static int tipc_l2_device_event(struct notifier_block *nb, unsigned long evt, tipc_l2_device_event() argument 529 switch (evt) { tipc_l2_device_event()
|
/linux-4.1.27/arch/arm/plat-iop/ |
H A D | time.c | 120 struct clock_event_device *evt = dev_id; iop_timer_interrupt() local 123 evt->event_handler(evt); iop_timer_interrupt()
|
/linux-4.1.27/arch/arm/kernel/ |
H A D | perf_event_v6.c | 262 unsigned long val, mask, evt, flags; armv6pmu_enable_event() local 270 evt = ARMV6_PMCR_CCOUNT_IEN; armv6pmu_enable_event() 273 evt = (hwc->config_base << ARMV6_PMCR_EVT_COUNT0_SHIFT) | armv6pmu_enable_event() 277 evt = (hwc->config_base << ARMV6_PMCR_EVT_COUNT1_SHIFT) | armv6pmu_enable_event() 291 val |= evt; armv6pmu_enable_event() 409 unsigned long val, mask, evt, flags; armv6pmu_disable_event() local 417 evt = 0; armv6pmu_disable_event() 420 evt = ARMV6_PERFCTR_NOP << ARMV6_PMCR_EVT_COUNT0_SHIFT; armv6pmu_disable_event() 423 evt = ARMV6_PERFCTR_NOP << ARMV6_PMCR_EVT_COUNT1_SHIFT; armv6pmu_disable_event() 437 val |= evt; armv6pmu_disable_event() 444 unsigned long val, mask, flags, evt = 0; armv6mpcore_pmu_disable_event() local 468 val |= evt; armv6mpcore_pmu_disable_event()
|
H A D | perf_event_xscale.c | 198 unsigned long val, mask, evt, flags; xscale1pmu_enable_event() local 207 evt = XSCALE1_CCOUNT_INT_EN; xscale1pmu_enable_event() 211 evt = (hwc->config_base << XSCALE1_COUNT0_EVT_SHFT) | xscale1pmu_enable_event() 216 evt = (hwc->config_base << XSCALE1_COUNT1_EVT_SHFT) | xscale1pmu_enable_event() 227 val |= evt; xscale1pmu_enable_event() 234 unsigned long val, mask, evt, flags; xscale1pmu_disable_event() local 243 evt = 0; xscale1pmu_disable_event() 247 evt = XSCALE_PERFCTR_UNUSED << XSCALE1_COUNT0_EVT_SHFT; xscale1pmu_disable_event() 251 evt = XSCALE_PERFCTR_UNUSED << XSCALE1_COUNT1_EVT_SHFT; xscale1pmu_disable_event() 261 val |= evt; xscale1pmu_disable_event()
|
H A D | smp_twd.c | 64 static int twd_set_next_event(unsigned long evt, twd_set_next_event() argument 71 writel_relaxed(evt, twd_base + TWD_TIMER_COUNTER); twd_set_next_event() 233 struct clock_event_device *evt = dev_id; twd_handler() local 236 evt->event_handler(evt); twd_handler()
|
/linux-4.1.27/arch/arm/mach-lpc32xx/ |
H A D | timer.c | 80 struct clock_event_device *evt = &lpc32xx_clkevt; lpc32xx_timer_interrupt() local 86 evt->event_handler(evt); lpc32xx_timer_interrupt()
|
/linux-4.1.27/drivers/gpio/ |
H A D | gpio-kempld.c | 125 u16 evt, evt_back; kempld_gpio_pincount() local 134 evt = kempld_read16(pld, KEMPLD_GPIO_EVT_LVL_EDGE); kempld_gpio_pincount() 140 return evt ? __ffs(evt) : 16; kempld_gpio_pincount()
|
/linux-4.1.27/arch/openrisc/kernel/ |
H A D | time.c | 52 struct clock_event_device *evt) openrisc_timer_set_mode() 110 struct clock_event_device *evt = &clockevent_openrisc_timer; timer_interrupt() local 118 evt->event_handler(evt); timer_interrupt() 51 openrisc_timer_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) openrisc_timer_set_mode() argument
|
/linux-4.1.27/arch/c6x/include/asm/ |
H A D | soc.h | 20 void (*assert_event)(unsigned int evt);
|
/linux-4.1.27/arch/arm/mach-spear/ |
H A D | time.c | 71 static int clockevent_next_event(unsigned long evt, 164 struct clock_event_device *evt = &clkevt; spear_timer_interrupt() local 168 evt->event_handler(evt); spear_timer_interrupt()
|
/linux-4.1.27/drivers/gpu/drm/armada/ |
H A D | armada_drv.c | 234 struct armada_vbl_event *evt) armada_drm_vbl_event_add() 239 if (list_empty(&evt->node)) { armada_drm_vbl_event_add() 240 list_add_tail(&evt->node, &dcrtc->vbl_list); armada_drm_vbl_event_add() 248 struct armada_vbl_event *evt) armada_drm_vbl_event_remove() 250 if (!list_empty(&evt->node)) { armada_drm_vbl_event_remove() 251 list_del_init(&evt->node); armada_drm_vbl_event_remove() 257 struct armada_vbl_event *evt) armada_drm_vbl_event_remove_unlocked() 262 armada_drm_vbl_event_remove(dcrtc, evt); armada_drm_vbl_event_remove_unlocked() 233 armada_drm_vbl_event_add(struct armada_crtc *dcrtc, struct armada_vbl_event *evt) armada_drm_vbl_event_add() argument 247 armada_drm_vbl_event_remove(struct armada_crtc *dcrtc, struct armada_vbl_event *evt) armada_drm_vbl_event_remove() argument 256 armada_drm_vbl_event_remove_unlocked(struct armada_crtc *dcrtc, struct armada_vbl_event *evt) armada_drm_vbl_event_remove_unlocked() argument
|
/linux-4.1.27/drivers/devfreq/exynos/ |
H A D | exynos_ppmu.c | 30 unsigned int evt) exynos_ppmu_setevent() 32 __raw_writel(evt, ppmu_base + PPMU_BEVTSEL(ch)); exynos_ppmu_setevent() 29 exynos_ppmu_setevent(void __iomem *ppmu_base, unsigned int ch, unsigned int evt) exynos_ppmu_setevent() argument
|
H A D | exynos_ppmu.h | 79 unsigned int evt);
|
/linux-4.1.27/arch/mips/jz4740/ |
H A D | time.c | 86 static int jz4740_clockevent_set_next(unsigned long evt, jz4740_clockevent_set_next() argument 90 jz4740_timer_set_period(TIMER_CLOCKEVENT, evt); jz4740_clockevent_set_next()
|
/linux-4.1.27/drivers/infiniband/core/ |
H A D | ucm.c | 252 static int ib_ucm_event_process(struct ib_cm_event *evt, ib_ucm_event_process() argument 257 switch (evt->event) { ib_ucm_event_process() 260 &evt->param.req_rcvd); ib_ucm_event_process() 263 uvt->resp.present |= (evt->param.req_rcvd.alternate_path ? ib_ucm_event_process() 268 &evt->param.rep_rcvd); ib_ucm_event_process() 273 uvt->resp.u.send_status = evt->param.send_status; ib_ucm_event_process() 277 uvt->resp.u.send_status = evt->param.send_status; ib_ucm_event_process() 281 uvt->resp.u.send_status = evt->param.send_status; ib_ucm_event_process() 285 evt->param.mra_rcvd.service_timeout; ib_ucm_event_process() 289 uvt->resp.u.rej_resp.reason = evt->param.rej_rcvd.reason; ib_ucm_event_process() 291 uvt->info_len = evt->param.rej_rcvd.ari_length; ib_ucm_event_process() 292 info = evt->param.rej_rcvd.ari; ib_ucm_event_process() 296 evt->param.lap_rcvd.alternate_path); ib_ucm_event_process() 301 uvt->resp.u.apr_resp.status = evt->param.apr_rcvd.ap_status; ib_ucm_event_process() 303 uvt->info_len = evt->param.apr_rcvd.info_len; ib_ucm_event_process() 304 info = evt->param.apr_rcvd.apr_info; ib_ucm_event_process() 308 evt->param.sidr_req_rcvd.pkey; ib_ucm_event_process() 310 evt->param.sidr_req_rcvd.port; ib_ucm_event_process() 315 &evt->param.sidr_rep_rcvd); ib_ucm_event_process() 317 uvt->info_len = evt->param.sidr_rep_rcvd.info_len; ib_ucm_event_process() 318 info = evt->param.sidr_rep_rcvd.info; ib_ucm_event_process() 321 uvt->resp.u.send_status = evt->param.send_status; ib_ucm_event_process() 326 uvt->data = kmemdup(evt->private_data, uvt->data_len, GFP_KERNEL); ib_ucm_event_process()
|
H A D | uverbs_main.c | 158 struct ib_uverbs_event *evt, *tmp; ib_uverbs_release_ucq() local 162 list_for_each_entry_safe(evt, tmp, &uobj->comp_list, obj_list) { ib_uverbs_release_ucq() 163 list_del(&evt->list); ib_uverbs_release_ucq() 164 kfree(evt); ib_uverbs_release_ucq() 172 list_for_each_entry_safe(evt, tmp, &uobj->async_list, obj_list) { ib_uverbs_release_ucq() 173 list_del(&evt->list); ib_uverbs_release_ucq() 174 kfree(evt); ib_uverbs_release_ucq() 182 struct ib_uverbs_event *evt, *tmp; ib_uverbs_release_uevent() local 185 list_for_each_entry_safe(evt, tmp, &uobj->event_list, obj_list) { ib_uverbs_release_uevent() 186 list_del(&evt->list); ib_uverbs_release_uevent() 187 kfree(evt); ib_uverbs_release_uevent()
|
/linux-4.1.27/arch/sparc/kernel/ |
H A D | time_32.c | 113 struct clock_event_device *evt) timer_ce_set_mode() 195 struct clock_event_device *evt) percpu_ce_setup() 197 int cpu = cpumask_first(evt->cpumask); percpu_ce_setup() 215 struct clock_event_device *evt) percpu_ce_set_next_event() 217 int cpu = cpumask_first(evt->cpumask); percpu_ce_set_next_event() 112 timer_ce_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) timer_ce_set_mode() argument 194 percpu_ce_setup(enum clock_event_mode mode, struct clock_event_device *evt) percpu_ce_setup() argument 214 percpu_ce_set_next_event(unsigned long delta, struct clock_event_device *evt) percpu_ce_set_next_event() argument
|
H A D | time_64.c | 686 struct clock_event_device *evt) sparc64_next_event() 692 struct clock_event_device *evt) sparc64_timer_setup() 725 struct clock_event_device *evt = &per_cpu(sparc64_events, cpu); timer_interrupt() local 734 if (unlikely(!evt->event_handler)) { timer_interrupt() 738 evt->event_handler(evt); timer_interrupt() 685 sparc64_next_event(unsigned long delta, struct clock_event_device *evt) sparc64_next_event() argument 691 sparc64_timer_setup(enum clock_event_mode mode, struct clock_event_device *evt) sparc64_timer_setup() argument
|
/linux-4.1.27/arch/cris/arch-v32/kernel/ |
H A D | time.c | 186 static int crisv32_clkevt_next_event(unsigned long evt, crisv32_clkevt_next_event() argument 194 REG_WR(timer, timer_base, rw_tmr0_div, evt); crisv32_clkevt_next_event() 205 struct clock_event_device *evt = dev_id; crisv32_timer_interrupt() local 225 evt->event_handler(evt); crisv32_timer_interrupt()
|
/linux-4.1.27/arch/arm/mach-cns3xxx/ |
H A D | core.c | 142 static int cns3xxx_timer_set_next_event(unsigned long evt, cns3xxx_timer_set_next_event() argument 147 writel(evt, cns3xxx_tmr1 + TIMER1_AUTO_RELOAD_OFFSET); cns3xxx_timer_set_next_event() 175 struct clock_event_device *evt = &cns3xxx_tmr1_clockevent; cns3xxx_timer_interrupt() local 183 evt->event_handler(evt); cns3xxx_timer_interrupt()
|
/linux-4.1.27/arch/mips/loongson/common/cs5536/ |
H A D | cs5536_mfgpt.c | 55 struct clock_event_device *evt) init_mfgpt_timer() 68 if (evt->mode == CLOCK_EVT_MODE_PERIODIC || init_mfgpt_timer() 69 evt->mode == CLOCK_EVT_MODE_ONESHOT) init_mfgpt_timer() 54 init_mfgpt_timer(enum clock_event_mode mode, struct clock_event_device *evt) init_mfgpt_timer() argument
|
/linux-4.1.27/kernel/time/ |
H A D | clockevents.c | 36 static u64 cev_delta2ns(unsigned long latch, struct clock_event_device *evt, cev_delta2ns() argument 39 u64 clc = (u64) latch << evt->shift; cev_delta2ns() 42 if (unlikely(!evt->mult)) { cev_delta2ns() 43 evt->mult = 1; cev_delta2ns() 46 rnd = (u64) evt->mult - 1; cev_delta2ns() 52 if ((clc >> evt->shift) != (u64)latch) cev_delta2ns() 75 (!ismax || evt->mult <= (1ULL << evt->shift))) cev_delta2ns() 78 do_div(clc, evt->mult); cev_delta2ns() 87 * @evt: pointer to clock event device descriptor 91 u64 clockevent_delta2ns(unsigned long latch, struct clock_event_device *evt) clockevent_delta2ns() argument 93 return cev_delta2ns(latch, evt, false); clockevent_delta2ns()
|
H A D | tick-broadcast.c | 242 struct clock_event_device *evt = td->evtdev; tick_receive_broadcast() local 244 if (!evt) tick_receive_broadcast() 247 if (!evt->event_handler) tick_receive_broadcast() 250 evt->event_handler(evt); tick_receive_broadcast()
|
/linux-4.1.27/drivers/virtio/ |
H A D | virtio_input.c | 15 struct virtqueue *evt, *sts; member in struct:virtio_input 27 virtqueue_add_inbuf(vi->evt, sg, 1, evtbuf, GFP_ATOMIC); virtinput_queue_evtbuf() 39 while ((event = virtqueue_get_buf(vi->evt, &len)) != NULL) { virtinput_recv_events() 179 vi->evt = vqs[0]; virtinput_init_vqs() 191 size = virtqueue_get_vring_size(vi->evt); virtinput_fill_evt() 196 virtqueue_kick(vi->evt); virtinput_fill_evt()
|
/linux-4.1.27/arch/metag/kernel/perf/ |
H A D | perf_event.h | 81 void (*enable)(struct hw_perf_event *evt, int idx); 82 void (*disable)(struct hw_perf_event *evt, int idx);
|
/linux-4.1.27/arch/sh/boards/mach-cayman/ |
H A D | irq.c | 98 int cayman_irq_demux(int evt) cayman_irq_demux() argument 100 int irq = intc_evt_to_irq[evt]; cayman_irq_demux()
|
/linux-4.1.27/arch/mn10300/kernel/ |
H A D | cevt-mn10300.c | 30 struct clock_event_device *evt) next_event() 45 struct clock_event_device *evt) set_clock_mode() 29 next_event(unsigned long delta, struct clock_event_device *evt) next_event() argument 44 set_clock_mode(enum clock_event_mode mode, struct clock_event_device *evt) set_clock_mode() argument
|
/linux-4.1.27/arch/microblaze/kernel/ |
H A D | timer.c | 126 struct clock_event_device *evt) xilinx_timer_set_mode() 165 struct clock_event_device *evt = &clockevent_xilinx_timer; timer_interrupt() local 170 evt->event_handler(evt); timer_interrupt() 125 xilinx_timer_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) xilinx_timer_set_mode() argument
|
/linux-4.1.27/arch/sh/kernel/cpu/sh3/ |
H A D | ex.S | 57 * location via set_exception_table_{evt,vec}().
|
/linux-4.1.27/arch/arm/mach-davinci/ |
H A D | time.c | 167 struct clock_event_device *evt = &clockevent_davinci; timer_interrupt() local 169 evt->event_handler(evt); timer_interrupt() 297 struct clock_event_device *evt) davinci_set_next_event() 307 struct clock_event_device *evt) davinci_set_mode() 296 davinci_set_next_event(unsigned long cycles, struct clock_event_device *evt) davinci_set_next_event() argument 306 davinci_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) davinci_set_mode() argument
|
/linux-4.1.27/drivers/vhost/ |
H A D | scsi.c | 212 struct vhost_work vs_event_work; /* evt injection work item */ 213 struct llist_head vs_event_list; /* evt injection queue */ 567 static void vhost_scsi_free_evt(struct vhost_scsi *vs, struct vhost_scsi_evt *evt) vhost_scsi_free_evt() argument 570 kfree(evt); vhost_scsi_free_evt() 578 struct vhost_scsi_evt *evt; vhost_scsi_allocate_evt() local 585 evt = kzalloc(sizeof(*evt), GFP_KERNEL); vhost_scsi_allocate_evt() 586 if (!evt) { vhost_scsi_allocate_evt() 592 evt->event.event = cpu_to_vhost32(vq, event); vhost_scsi_allocate_evt() 593 evt->event.reason = cpu_to_vhost32(vq, reason); vhost_scsi_allocate_evt() 596 return evt; vhost_scsi_allocate_evt() 614 vhost_scsi_do_evt_work(struct vhost_scsi *vs, struct vhost_scsi_evt *evt) vhost_scsi_do_evt_work() argument 617 struct virtio_scsi_event *event = &evt->event; vhost_scsi_do_evt_work() 668 struct vhost_scsi_evt *evt; vhost_scsi_evt_work() local 674 evt = llist_entry(llnode, struct vhost_scsi_evt, list); vhost_scsi_evt_work() 676 vhost_scsi_do_evt_work(vs, evt); vhost_scsi_evt_work() 677 vhost_scsi_free_evt(vs, evt); vhost_scsi_evt_work() 1265 struct vhost_scsi_evt *evt; vhost_scsi_send_evt() local 1267 evt = vhost_scsi_allocate_evt(vs, event, reason); vhost_scsi_send_evt() 1268 if (!evt) vhost_scsi_send_evt() 1274 * Note: evt->event is zeroed when we allocate it and vhost_scsi_send_evt() 1277 evt->event.lun[0] = 0x01; vhost_scsi_send_evt() 1278 evt->event.lun[1] = tpg->tport_tpgt; vhost_scsi_send_evt() 1280 evt->event.lun[2] = lun->unpacked_lun >> 8 | 0x40 ; vhost_scsi_send_evt() 1281 evt->event.lun[3] = lun->unpacked_lun & 0xFF; vhost_scsi_send_evt() 1284 llist_add(&evt->list, &vs->vs_event_list); vhost_scsi_send_evt() 1625 /* Jobs can re-queue themselves in evt kick handler. Do extra flush. */ vhost_scsi_release()
|
/linux-4.1.27/drivers/firewire/ |
H A D | ohci.c | 379 static void log_irqs(struct fw_ohci *ohci, u32 evt) log_irqs() argument 386 !(evt & OHCI1394_busReset)) log_irqs() 389 ohci_notice(ohci, "IRQ %08x%s%s%s%s%s%s%s%s%s%s%s%s%s%s%s\n", evt, log_irqs() 390 evt & OHCI1394_selfIDComplete ? " selfID" : "", log_irqs() 391 evt & OHCI1394_RQPkt ? " AR_req" : "", log_irqs() 392 evt & OHCI1394_RSPkt ? " AR_resp" : "", log_irqs() 393 evt & OHCI1394_reqTxComplete ? " AT_req" : "", log_irqs() 394 evt & OHCI1394_respTxComplete ? " AT_resp" : "", log_irqs() 395 evt & OHCI1394_isochRx ? " IR" : "", log_irqs() 396 evt & OHCI1394_isochTx ? " IT" : "", log_irqs() 397 evt & OHCI1394_postedWriteErr ? " postedWriteErr" : "", log_irqs() 398 evt & OHCI1394_cycleTooLong ? " cycleTooLong" : "", log_irqs() 399 evt & OHCI1394_cycle64Seconds ? " cycle64Seconds" : "", log_irqs() 400 evt & OHCI1394_cycleInconsistent ? " cycleInconsistent" : "", log_irqs() 401 evt & OHCI1394_regAccessFail ? " regAccessFail" : "", log_irqs() 402 evt & OHCI1394_unrecoverableError ? " unrecoverableError" : "", log_irqs() 403 evt & OHCI1394_busReset ? " busReset" : "", log_irqs() 404 evt & ~(OHCI1394_selfIDComplete | OHCI1394_RQPkt | log_irqs() 485 char dir, int speed, u32 *header, int evt) log_ar_at_event() 493 if (unlikely(evt >= ARRAY_SIZE(evts))) log_ar_at_event() 494 evt = 0x1f; log_ar_at_event() 496 if (evt == OHCI1394_evt_bus_reset) { log_ar_at_event() 518 dir, evts[evt], tcodes[tcode]); log_ar_at_event() 522 dir, evts[evt], header[1], header[2]); log_ar_at_event() 528 header[1] >> 16, header[0] >> 16, evts[evt], log_ar_at_event() 535 header[1] >> 16, header[0] >> 16, evts[evt], log_ar_at_event() 818 int evt; handle_ar_packet() local 869 evt = (status >> 16) & 0x1f; handle_ar_packet() 871 p.ack = evt - 16; handle_ar_packet() 876 log_ar_at_event(ohci, 'R', p.speed, p.header, evt); handle_ar_packet() 882 if (evt == OHCI1394_evt_no_status && handle_ar_packet() 899 if (evt == OHCI1394_evt_bus_reset) { handle_ar_packet() 1445 int evt; handle_at_packet() local 1461 evt = le16_to_cpu(last->transfer_status) & 0x1f; handle_at_packet() 1464 log_ar_at_event(ohci, 'T', packet->speed, packet->header, evt); handle_at_packet() 1466 switch (evt) { handle_at_packet() 1499 packet->ack = evt - 0x10; handle_at_packet() 484 log_ar_at_event(struct fw_ohci *ohci, char dir, int speed, u32 *header, int evt) log_ar_at_event() argument
|
/linux-4.1.27/include/linux/uwb/ |
H A D | spec.h | 685 const struct uwb_rc_evt_bp_slot_change *evt) uwb_rc_evt_bp_slot_change_slot_num() 687 return evt->slot_info & 0x7f; uwb_rc_evt_bp_slot_change_slot_num() 691 const struct uwb_rc_evt_bp_slot_change *evt) uwb_rc_evt_bp_slot_change_no_slot() 693 return (evt->slot_info & 0x80) >> 7; uwb_rc_evt_bp_slot_change_no_slot() 719 static inline enum uwb_drp_notif_reason uwb_rc_evt_drp_reason(struct uwb_rc_evt_drp *evt) uwb_rc_evt_drp_reason() argument 721 return evt->reason & 0x0f; uwb_rc_evt_drp_reason() 684 uwb_rc_evt_bp_slot_change_slot_num( const struct uwb_rc_evt_bp_slot_change *evt) uwb_rc_evt_bp_slot_change_slot_num() argument 690 uwb_rc_evt_bp_slot_change_no_slot( const struct uwb_rc_evt_bp_slot_change *evt) uwb_rc_evt_bp_slot_change_no_slot() argument
|
/linux-4.1.27/arch/arm/mach-ixp4xx/ |
H A D | common.c | 312 struct clock_event_device *evt = dev_id; ixp4xx_timer_interrupt() local 317 evt->event_handler(evt); ixp4xx_timer_interrupt() 514 static int ixp4xx_set_next_event(unsigned long evt, ixp4xx_set_next_event() argument 519 *IXP4XX_OSRT1 = (evt & ~IXP4XX_OST_RELOAD_MASK) | opts; ixp4xx_set_next_event() 525 struct clock_event_device *evt) ixp4xx_set_mode() 524 ixp4xx_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) ixp4xx_set_mode() argument
|
/linux-4.1.27/arch/mips/loongson/loongson-3/ |
H A D | hpet.c | 82 struct clock_event_device *evt) hpet_set_mode() 133 struct clock_event_device *evt) hpet_next_event() 81 hpet_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) hpet_set_mode() argument 132 hpet_next_event(unsigned long delta, struct clock_event_device *evt) hpet_next_event() argument
|
/linux-4.1.27/arch/mips/loongson1/common/ |
H A D | time.c | 151 static int ls1x_clockevent_set_next(unsigned long evt, ls1x_clockevent_set_next() argument 155 ls1x_pwmtimer_set_period(evt); ls1x_clockevent_set_next()
|
/linux-4.1.27/arch/mips/sgi-ip27/ |
H A D | ip27-timer.c | 54 static int rt_next_event(unsigned long delta, struct clock_event_device *evt) rt_next_event() argument 68 struct clock_event_device *evt) rt_set_mode() 67 rt_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) rt_set_mode() argument
|
/linux-4.1.27/arch/powerpc/include/asm/ |
H A D | mce.h | 195 extern void machine_check_print_event_info(struct machine_check_event *evt); 196 extern uint64_t get_mce_fault_addr(struct machine_check_event *evt);
|
/linux-4.1.27/arch/c6x/platforms/ |
H A D | timer64.c | 123 struct clock_event_device *evt) next_event() 130 struct clock_event_device *evt) set_clock_mode() 122 next_event(unsigned long delta, struct clock_event_device *evt) next_event() argument 129 set_clock_mode(enum clock_event_mode mode, struct clock_event_device *evt) set_clock_mode() argument
|
/linux-4.1.27/arch/blackfin/mach-bf609/ |
H A D | dpm.S | 102 /* Enable evt 11 in IMASK before idle, otherwise core doesn't wake up. */
|
/linux-4.1.27/drivers/bluetooth/ |
H A D | btusb.c | 1196 static void btusb_notify(struct hci_dev *hdev, unsigned int evt) btusb_notify() argument 1200 BT_DBG("%s evt %d", hdev->name, evt); btusb_notify() 1808 struct hci_event_hdr *evt = NULL; btusb_setup_intel_patching() local 1862 evt = (struct hci_event_hdr *)(*fw_ptr); btusb_setup_intel_patching() 1863 *fw_ptr += sizeof(*evt); btusb_setup_intel_patching() 1864 remain -= sizeof(*evt); btusb_setup_intel_patching() 1866 if (remain < evt->plen) { btusb_setup_intel_patching() 1867 BT_ERR("%s Intel fw corrupted: invalid evt len", btusb_setup_intel_patching() 1873 *fw_ptr += evt->plen; btusb_setup_intel_patching() 1874 remain -= evt->plen; btusb_setup_intel_patching() 1881 if (!evt || !evt_param || remain < 0) { btusb_setup_intel_patching() 1882 BT_ERR("%s Intel fw corrupted: invalid evt read", hdev->name); btusb_setup_intel_patching() 1887 cmd_param, evt->evt, HCI_INIT_TIMEOUT); btusb_setup_intel_patching() 1898 if (skb->len != evt->plen) { btusb_setup_intel_patching() 1905 if (memcmp(skb->data, evt_param, evt->plen)) { btusb_setup_intel_patching() 2129 struct hci_ev_cmd_complete *evt; inject_cmd_complete() local 2131 skb = bt_skb_alloc(sizeof(*hdr) + sizeof(*evt) + 1, GFP_ATOMIC); inject_cmd_complete() 2136 hdr->evt = HCI_EV_CMD_COMPLETE; inject_cmd_complete() 2137 hdr->plen = sizeof(*evt) + 1; inject_cmd_complete() 2139 evt = (struct hci_ev_cmd_complete *)skb_put(skb, sizeof(*evt)); inject_cmd_complete() 2140 evt->ncmd = 0x01; inject_cmd_complete() 2141 evt->opcode = cpu_to_le16(opcode); inject_cmd_complete() 2166 const struct intel_bootup *evt = ptr; btusb_intel_bootup() local 2168 if (len != sizeof(*evt)) btusb_intel_bootup() 2180 const struct intel_secure_send_result *evt = ptr; btusb_intel_secure_send_result() local 2182 if (len != sizeof(*evt)) btusb_intel_secure_send_result() 2185 if (evt->result) btusb_intel_secure_send_result() 2202 if (skb->len > HCI_EVENT_HDR_SIZE && hdr->evt == 0xff && btusb_recv_event_intel()
|
H A D | hci_ll.c | 406 BT_DBG("Event header: evt 0x%2.2x plen %d", eh->evt, eh->plen); ll_recv()
|
H A D | hci_bcsp.c | 191 chan = 5; /* BCSP cmd/evt channel */ bcsp_prepare_pkt() 523 hdr.evt = 0xff; bcsp_complete_rx_pkt()
|
H A D | btmrvl_main.c | 58 if (hdr->evt == HCI_EV_CMD_COMPLETE) { btmrvl_check_evtpkt()
|
/linux-4.1.27/drivers/ptp/ |
H A D | ptp_clock.c | 291 struct pps_event_time evt; ptp_clock_event() local 304 pps_get_ts(&evt); ptp_clock_event() 305 pps_event(ptp->pps_source, &evt, PTP_PPS_EVENT, NULL); ptp_clock_event()
|
/linux-4.1.27/arch/blackfin/mach-common/ |
H A D | smp.c | 135 struct clock_event_device *evt = &per_cpu(coretmr_events, cpu); ipi_timer() local 136 evt->event_handler(evt); ipi_timer()
|
/linux-4.1.27/drivers/net/ethernet/sfc/ |
H A D | ptp.c | 930 struct efx_ptp_event_rx *evt; efx_ptp_drop_time_expired_events() local 932 evt = list_entry(cursor, struct efx_ptp_event_rx, efx_ptp_drop_time_expired_events() 934 if (time_after(jiffies, evt->expiry)) { efx_ptp_drop_time_expired_events() 935 list_move(&evt->link, &ptp->evt_free_list); efx_ptp_drop_time_expired_events() 967 struct efx_ptp_event_rx *evt; efx_ptp_match_rx() local 969 evt = list_entry(cursor, struct efx_ptp_event_rx, link); efx_ptp_match_rx() 970 if ((evt->seq0 == match->words[0]) && efx_ptp_match_rx() 971 (evt->seq1 == match->words[1])) { efx_ptp_match_rx() 976 timestamps->hwtstamp = evt->hwtimestamp; efx_ptp_match_rx() 980 list_move(&evt->link, &ptp->evt_free_list); efx_ptp_match_rx() 1596 struct efx_ptp_event_rx *evt = NULL; ptp_event_rx() local 1608 evt = list_first_entry(&ptp->evt_free_list, ptp_event_rx() 1610 list_del(&evt->link); ptp_event_rx() 1612 evt->seq0 = EFX_QWORD_FIELD(ptp->evt_frags[2], MCDI_EVENT_DATA); ptp_event_rx() 1613 evt->seq1 = (EFX_QWORD_FIELD(ptp->evt_frags[2], ptp_event_rx() 1619 evt->hwtimestamp = efx->ptp_data->nic_to_kernel_time( ptp_event_rx() 1623 evt->expiry = jiffies + msecs_to_jiffies(PKT_EVENT_LIFETIME_MS); ptp_event_rx() 1624 list_add_tail(&evt->link, &ptp->evt_list); ptp_event_rx()
|
/linux-4.1.27/drivers/scsi/be2iscsi/ |
H A D | be_cmds.c | 437 struct be_async_event_link_state *evt) beiscsi_async_link_state_process() 439 if ((evt->port_link_status == ASYNC_EVENT_LINK_DOWN) || beiscsi_async_link_state_process() 440 ((evt->port_link_status & ASYNC_EVENT_LOGICAL) && beiscsi_async_link_state_process() 441 (evt->port_fault != BEISCSI_PHY_LINK_FAULT_NONE))) { beiscsi_async_link_state_process() 447 evt->physical_port); beiscsi_async_link_state_process() 451 } else if ((evt->port_link_status & ASYNC_EVENT_LINK_UP) || beiscsi_async_link_state_process() 452 ((evt->port_link_status & ASYNC_EVENT_LOGICAL) && beiscsi_async_link_state_process() 453 (evt->port_fault == BEISCSI_PHY_LINK_FAULT_NONE))) { beiscsi_async_link_state_process() 459 evt->physical_port); beiscsi_async_link_state_process() 474 /* Interpret compl as a async link evt */ beiscsi_process_mcc() 436 beiscsi_async_link_state_process(struct beiscsi_hba *phba, struct be_async_event_link_state *evt) beiscsi_async_link_state_process() argument
|
H A D | be_cmds.h | 741 struct be_async_event_link_state *evt);
|
/linux-4.1.27/drivers/hv/ |
H A D | hv.c | 270 struct clock_event_device *evt) hv_ce_set_next_event() 274 WARN_ON(evt->mode != CLOCK_EVT_MODE_ONESHOT); hv_ce_set_next_event() 283 struct clock_event_device *evt) hv_ce_setmode() 269 hv_ce_set_next_event(unsigned long delta, struct clock_event_device *evt) hv_ce_set_next_event() argument 282 hv_ce_setmode(enum clock_event_mode mode, struct clock_event_device *evt) hv_ce_setmode() argument
|
/linux-4.1.27/arch/alpha/kernel/ |
H A D | time.c | 116 rtc_ce_set_next_event(unsigned long evt, struct clock_event_device *ce) rtc_ce_set_next_event() 175 qemu_ce_set_next_event(unsigned long evt, struct clock_event_device *ce) qemu_ce_set_next_event() 177 qemu_set_alarm_rel(evt); 115 rtc_ce_set_next_event(unsigned long evt, struct clock_event_device *ce) rtc_ce_set_next_event() argument 173 qemu_ce_set_next_event(unsigned long evt, struct clock_event_device *ce) qemu_ce_set_next_event() argument
|
/linux-4.1.27/drivers/leds/trigger/ |
H A D | ledtrig-cpu.c | 43 * @evt: CPU event to be emitted
|
/linux-4.1.27/drivers/net/ethernet/alteon/ |
H A D | acenic.h | 319 u32 evt:8; member in struct:event 321 u32 evt:8; 365 u32 evt:8; member in struct:cmd 367 u32 evt:8;
|
H A D | acenic.c | 1671 cmd.evt = C_SET_RX_PRD_IDX; ace_load_std_rx_ring() 1788 cmd.evt = C_SET_RX_JUMBO_PRD_IDX; ace_load_jumbo_rx_ring() 1820 switch (ap->evt_ring[evtcsm].evt) { ace_handle_event() 1891 cmd.evt = C_SET_RX_JUMBO_PRD_IDX; ace_handle_event() 1909 ap->name, ap->evt_ring[evtcsm].evt); ace_handle_event() 2243 cmd.evt = C_CLEAR_STATS; ace_open() 2248 cmd.evt = C_HOST_STATE; ace_open() 2258 cmd.evt = C_SET_PROMISC_MODE; ace_open() 2269 cmd.evt = C_LNK_NEGOTIATION; ace_open() 2302 cmd.evt = C_SET_PROMISC_MODE; ace_close() 2309 cmd.evt = C_HOST_STATE; ace_close() 2354 cmd.evt = C_RESET_JUMBO_RNG; ace_close() 2573 cmd.evt = C_RESET_JUMBO_RNG; ace_change_mtu() 2690 cmd.evt = C_LNK_NEGOTIATION; ace_set_settings() 2736 cmd.evt = C_SET_MAC_ADDR; ace_set_mac_addr() 2752 cmd.evt = C_SET_MULTICAST_MODE; ace_set_multicast_list() 2758 cmd.evt = C_SET_MULTICAST_MODE; ace_set_multicast_list() 2766 cmd.evt = C_SET_PROMISC_MODE; ace_set_multicast_list() 2772 cmd.evt = C_SET_PROMISC_MODE; ace_set_multicast_list() 2786 cmd.evt = C_SET_MULTICAST_MODE; ace_set_multicast_list() 2791 cmd.evt = C_SET_MULTICAST_MODE; ace_set_multicast_list()
|
/linux-4.1.27/arch/mips/jazz/ |
H A D | irq.c | 114 struct clock_event_device *evt) r4030_set_mode() 113 r4030_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) r4030_set_mode() argument
|
/linux-4.1.27/net/nfc/hci/ |
H A D | hci.h | 29 u8 header; /* type -cmd,evt,rsp- + instruction */
|
/linux-4.1.27/drivers/net/ethernet/emulex/benet/ |
H A D | be_cmds.c | 241 /* Link state evt is a string of bytes; no need for endian swapping */ be_async_link_state_process() 245 struct be_async_event_link_state *evt = be_async_link_state_process() local 257 !(evt->port_link_status & LOGICAL_LINK_STATUS_MASK)) be_async_link_state_process() 265 evt->port_link_status & LINK_STATUS_MASK); be_async_link_state_process() 271 struct be_async_event_misconfig_port *evt = be_async_port_misconfig_event_process() local 273 u32 sfp_mismatch_evt = le32_to_cpu(evt->event_data_word1); be_async_port_misconfig_event_process() 291 /* Grp5 CoS Priority evt */ be_async_grp5_cos_priority_process() 295 struct be_async_event_grp5_cos_priority *evt = be_async_grp5_cos_priority_process() local 298 if (evt->valid) { be_async_grp5_cos_priority_process() 299 adapter->vlan_prio_bmap = evt->available_priority_bmap; be_async_grp5_cos_priority_process() 302 evt->reco_default_priority << VLAN_PRIO_SHIFT; be_async_grp5_cos_priority_process() 306 /* Grp5 QOS Speed evt: qos_link_speed is in units of 10 Mbps */ be_async_grp5_qos_speed_process() 310 struct be_async_event_grp5_qos_link_speed *evt = be_async_grp5_qos_speed_process() local 314 evt->physical_port == adapter->port_num) be_async_grp5_qos_speed_process() 315 adapter->phy.link_speed = le16_to_cpu(evt->qos_link_speed) * 10; be_async_grp5_qos_speed_process() 318 /*Grp5 PVID evt*/ be_async_grp5_pvid_state_process() 322 struct be_async_event_grp5_pvid_state *evt = be_async_grp5_pvid_state_process() local 325 if (evt->enabled) { be_async_grp5_pvid_state_process() 326 adapter->pvid = le16_to_cpu(evt->tag) & VLAN_VID_MASK; be_async_grp5_pvid_state_process() 358 struct be_async_event_qnq *evt = (struct be_async_event_qnq *)cmp; be_async_dbg_evt_process() local 365 if (evt->valid) be_async_dbg_evt_process() 366 adapter->qnq_vid = le16_to_cpu(evt->vlan_tag); be_async_dbg_evt_process()
|
H A D | be_hw.h | 190 u32 evt; member in struct:be_eq_entry
|
H A D | be_main.c | 802 /* Hack to skip HW VLAN tagging needs evt = 1, compl = 0. When this wrb_fill_hdr() 803 * hack is not needed, the evt bit is set while ringing DB. wrb_fill_hdr() 2198 if (eqe->evt == 0) events_get() 2202 eqe->evt = 0; events_get()
|
/linux-4.1.27/drivers/scsi/ |
H A D | gdth.c | 158 u16 idx, gdth_evt_data *evt); 2838 u16 idx, gdth_evt_data *evt) gdth_store_event() 2850 ((evt->size != 0 && ebuffer[elastidx].event_data.size != 0 && gdth_store_event() 2852 (char *)&evt->eu, evt->size)) || gdth_store_event() 2853 (evt->size == 0 && ebuffer[elastidx].event_data.size == 0 && gdth_store_event() 2855 (char *)&evt->event_string)))) { gdth_store_event() 2877 e->event_data = *evt; gdth_store_event() 4059 gdth_ioctl_event evt; ioc_event() local 4063 if (copy_from_user(&evt, arg, sizeof(gdth_ioctl_event))) ioc_event() 4065 ha = gdth_find_ha(evt.ionode); ioc_event() 4069 if (evt.erase == 0xff) { ioc_event() 4070 if (evt.event.event_source == ES_TEST) ioc_event() 4071 evt.event.event_data.size=sizeof(evt.event.event_data.eu.test); ioc_event() 4072 else if (evt.event.event_source == ES_DRIVER) ioc_event() 4073 evt.event.event_data.size=sizeof(evt.event.event_data.eu.driver); ioc_event() 4074 else if (evt.event.event_source == ES_SYNC) ioc_event() 4075 evt.event.event_data.size=sizeof(evt.event.event_data.eu.sync); ioc_event() 4077 evt.event.event_data.size=sizeof(evt.event.event_data.eu.async); ioc_event() 4079 gdth_store_event(ha, evt.event.event_source, evt.event.event_idx, ioc_event() 4080 &evt.event.event_data); ioc_event() 4082 } else if (evt.erase == 0xfe) { ioc_event() 4084 } else if (evt.erase == 0) { ioc_event() 4085 evt.handle = gdth_read_event(ha, evt.handle, &evt.event); ioc_event() 4087 gdth_readapp_event(ha, evt.erase, &evt.event); ioc_event() 4089 if (copy_to_user(arg, &evt, sizeof(gdth_ioctl_event))) ioc_event() 2837 gdth_store_event(gdth_ha_str *ha, u16 source, u16 idx, gdth_evt_data *evt) gdth_store_event() argument
|
H A D | scsi_lib.c | 2685 * @evt: event to emit 2689 static void scsi_evt_emit(struct scsi_device *sdev, struct scsi_event *evt) scsi_evt_emit() argument 2694 switch (evt->evt_type) { scsi_evt_emit() 2743 struct scsi_event *evt; scsi_evt_thread() local 2755 evt = list_entry(this, struct scsi_event, node); scsi_evt_thread() 2756 list_del(&evt->node); scsi_evt_thread() 2757 scsi_evt_emit(sdev, evt); scsi_evt_thread() 2758 kfree(evt); scsi_evt_thread() 2766 * @evt: event to send 2770 void sdev_evt_send(struct scsi_device *sdev, struct scsi_event *evt) sdev_evt_send() argument 2778 if (!test_bit(evt->evt_type, sdev->supported_events)) { sdev_evt_send() 2779 kfree(evt); sdev_evt_send() 2785 list_add_tail(&evt->node, &sdev->event_list); sdev_evt_send() 2801 struct scsi_event *evt = kzalloc(sizeof(struct scsi_event), gfpflags); sdev_evt_alloc() local 2802 if (!evt) sdev_evt_alloc() 2805 evt->evt_type = evt_type; sdev_evt_alloc() 2806 INIT_LIST_HEAD(&evt->node); sdev_evt_alloc() 2821 return evt; sdev_evt_alloc() 2836 struct scsi_event *evt = sdev_evt_alloc(evt_type, gfpflags); sdev_evt_send_simple() local 2837 if (!evt) { sdev_evt_send_simple() 2843 sdev_evt_send(sdev, evt); sdev_evt_send_simple()
|
H A D | gdth_ioctl.h | 206 u8 evt[EVENT_SIZE]; /* event structure */ member in struct:__anon8932::__anon8933::__anon8937 253 u8 evt[EVENT_SIZE]; /* event structure */ member in struct:__anon8940::__anon8941::__anon8947
|
H A D | scsi_sysfs.c | 413 struct scsi_event *evt; scsi_device_dev_release_usercontext() local 415 evt = list_entry(this, struct scsi_event, node); scsi_device_dev_release_usercontext() 416 list_del(&evt->node); scsi_device_dev_release_usercontext() 417 kfree(evt); scsi_device_dev_release_usercontext()
|
H A D | virtio_scsi.c | 51 struct virtio_scsi_event evt; member in union:virtio_scsi_cmd::__anon9695
|
/linux-4.1.27/drivers/infiniband/hw/cxgb3/ |
H A D | iwch.c | 205 static void iwch_event_handler(struct t3cdev *tdev, u32 evt, u32 port_id) iwch_event_handler() argument 216 switch (evt) { iwch_event_handler()
|
/linux-4.1.27/arch/sh/include/asm/ |
H A D | uaccess.h | 198 static inline void *set_exception_table_evt(unsigned int evt, void *handler) set_exception_table_evt() argument 200 return set_exception_table_vec(evt >> 5, handler); set_exception_table_evt()
|
/linux-4.1.27/arch/hexagon/kernel/ |
H A D | time.c | 88 static int set_next_event(unsigned long delta, struct clock_event_device *evt) set_next_event() argument 104 struct clock_event_device *evt) set_mode() 103 set_mode(enum clock_event_mode mode, struct clock_event_device *evt) set_mode() argument
|
/linux-4.1.27/arch/arm/mach-shmobile/ |
H A D | intc.h | 293 #define intcs_evt2irq(evt) evt2irq(INTCS_VECT_BASE + (evt))
|
/linux-4.1.27/arch/arm/mach-omap2/ |
H A D | timer.c | 82 struct clock_event_device *evt = &clockevent_gpt; omap2_gp_timer_interrupt() local 86 evt->event_handler(evt); omap2_gp_timer_interrupt() 97 struct clock_event_device *evt) omap2_gp_timer_set_next_event() 106 struct clock_event_device *evt) omap2_gp_timer_set_mode() 96 omap2_gp_timer_set_next_event(unsigned long cycles, struct clock_event_device *evt) omap2_gp_timer_set_next_event() argument 105 omap2_gp_timer_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) omap2_gp_timer_set_mode() argument
|
/linux-4.1.27/drivers/isdn/hisax/ |
H A D | st5481.h | 501 static inline const char *ST5481_CMD_string(int evt) ST5481_CMD_string() argument 505 switch (evt) { ST5481_CMD_string() 517 sprintf(s, "0x%x", evt); ST5481_CMD_string()
|
/linux-4.1.27/arch/mips/sni/ |
H A D | time.c | 18 struct clock_event_device *evt) a20r_set_mode() 17 a20r_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) a20r_set_mode() argument
|
/linux-4.1.27/arch/arm/include/asm/ |
H A D | pmu.h | 104 int (*set_event_filter)(struct hw_perf_event *evt,
|
/linux-4.1.27/arch/x86/platform/uv/ |
H A D | uv_time.c | 327 struct clock_event_device *evt) uv_rtc_timer_setup() 329 int ced_cpu = cpumask_first(evt->cpumask); uv_rtc_timer_setup() 326 uv_rtc_timer_setup(enum clock_event_mode mode, struct clock_event_device *evt) uv_rtc_timer_setup() argument
|
/linux-4.1.27/drivers/gpu/ipu-v3/ |
H A D | ipu-dc.c | 55 #define DC_RL_CH(evt) (8 + ((evt) & ~0x1) * 2)
|
/linux-4.1.27/arch/x86/kernel/apic/ |
H A D | apic.c | 456 struct clock_event_device *evt) lapic_next_event() 463 struct clock_event_device *evt) lapic_next_deadline() 476 struct clock_event_device *evt) lapic_timer_setup() 482 if (evt->features & CLOCK_EVT_FEAT_DUMMY) lapic_timer_setup() 872 struct clock_event_device *evt = &per_cpu(lapic_events, cpu); local_apic_timer_interrupt() local 881 * its possible that when we get here evt->event_handler is NULL. local_apic_timer_interrupt() 885 if (!evt->event_handler) { local_apic_timer_interrupt() 888 lapic_timer_setup(CLOCK_EVT_MODE_SHUTDOWN, evt); local_apic_timer_interrupt() 897 evt->event_handler(evt); local_apic_timer_interrupt() 455 lapic_next_event(unsigned long delta, struct clock_event_device *evt) lapic_next_event() argument 462 lapic_next_deadline(unsigned long delta, struct clock_event_device *evt) lapic_next_deadline() argument 475 lapic_timer_setup(enum clock_event_mode mode, struct clock_event_device *evt) lapic_timer_setup() argument
|
/linux-4.1.27/drivers/net/ethernet/pasemi/ |
H A D | pasemi_mac.c | 1399 int cs_size, i, fill, hdr, cpyhdr, evt; pasemi_mac_queue_csdesc() local 1447 evt = !csring->last_event; pasemi_mac_queue_csdesc() 1448 csring->last_event = evt; pasemi_mac_queue_csdesc() 1452 CTRL_CMD_ETYPE_SET | CTRL_CMD_REG(csring->events[evt]); pasemi_mac_queue_csdesc() 1455 CTRL_CMD_ETYPE_WCLR | CTRL_CMD_REG(csring->events[!evt]); pasemi_mac_queue_csdesc() 1465 CTRL_CMD_ETYPE_WSET | CTRL_CMD_REG(csring->events[evt]); pasemi_mac_queue_csdesc() 1468 CTRL_CMD_ETYPE_CLR | CTRL_CMD_REG(csring->events[!evt]); pasemi_mac_queue_csdesc()
|
/linux-4.1.27/drivers/scsi/bfa/ |
H A D | bfa_fcpim.c | 2797 enum bfa_ioim_event evt = BFA_IOIM_SM_COMP; bfa_ioim_isr() local 2815 evt = BFA_IOIM_SM_DONE; bfa_ioim_isr() 2817 evt = BFA_IOIM_SM_COMP; bfa_ioim_isr() 2826 evt = BFA_IOIM_SM_DONE; bfa_ioim_isr() 2828 evt = BFA_IOIM_SM_COMP; bfa_ioim_isr() 2834 evt = BFA_IOIM_SM_COMP; bfa_ioim_isr() 2840 evt = BFA_IOIM_SM_SQRETRY; bfa_ioim_isr() 2845 evt = BFA_IOIM_SM_FREE; bfa_ioim_isr() 2857 evt = BFA_IOIM_SM_ABORT_COMP; bfa_ioim_isr() 2859 evt = BFA_IOIM_SM_ABORT_DONE; bfa_ioim_isr() 2864 evt = BFA_IOIM_SM_COMP_UTAG; bfa_ioim_isr() 2871 bfa_sm_send_event(ioim, evt); bfa_ioim_isr()
|
/linux-4.1.27/drivers/scsi/qla2xxx/ |
H A D | qla_mr.c | 1711 qlafx00_process_aen(struct scsi_qla_host *vha, struct qla_work_evt *evt) qlafx00_process_aen() argument 1717 aen_data = evt->u.aenfx.evtcode; qlafx00_process_aen() 1719 switch (evt->u.aenfx.evtcode) { qlafx00_process_aen() 1721 if (evt->u.aenfx.mbx[1] == 0) { qlafx00_process_aen() 1722 if (evt->u.aenfx.mbx[2] == 1) { qlafx00_process_aen() 1729 } else if (evt->u.aenfx.mbx[2] == 2) { qlafx00_process_aen() 1730 qlafx00_tgt_detach(vha, evt->u.aenfx.mbx[3]); qlafx00_process_aen() 1732 } else if (evt->u.aenfx.mbx[1] == 0xffff) { qlafx00_process_aen() 1733 if (evt->u.aenfx.mbx[2] == 1) { qlafx00_process_aen() 1737 } else if (evt->u.aenfx.mbx[2] == 2) { qlafx00_process_aen() 1755 evt->u.aenfx.evtcode); qlafx00_process_aen()
|
/linux-4.1.27/drivers/gpu/drm/sti/ |
H A D | sti_hqvdp.c | 708 * @evt: event message 716 int sti_hqvdp_vtg_cb(struct notifier_block *nb, unsigned long evt, void *data) sti_hqvdp_vtg_cb() argument 722 if ((evt != VTG_TOP_FIELD_EVENT) && (evt != VTG_BOTTOM_FIELD_EVENT)) { sti_hqvdp_vtg_cb()
|
/linux-4.1.27/drivers/net/wan/ |
H A D | dscc4.c | 1779 }, *evt; dscc4_rx_irq() local 1781 for (evt = evts; evt->irq_name; evt++) { dscc4_rx_irq() 1782 if (state & evt->mask) { dscc4_rx_irq() 1784 dev->name, evt->irq_name); dscc4_rx_irq() 1785 if (!(state &= ~evt->mask)) dscc4_rx_irq()
|
/linux-4.1.27/drivers/dma/ |
H A D | imx-sdma.c | 475 unsigned long evt, mcu, dsp; sdma_config_ownership() local 480 evt = readl_relaxed(sdma->regs + SDMA_H_EVTOVR); sdma_config_ownership() 490 __clear_bit(channel, &evt); sdma_config_ownership() 492 __set_bit(channel, &evt); sdma_config_ownership() 499 writel_relaxed(evt, sdma->regs + SDMA_H_EVTOVR); sdma_config_ownership()
|
/linux-4.1.27/net/wireless/ |
H A D | trace.h | 2440 TP_PROTO(struct net_device *netdev, enum nl80211_radar_event evt), 2441 TP_ARGS(netdev, evt), 2444 __field(enum nl80211_radar_event, evt) 2448 __entry->evt = evt; 2451 NETDEV_PR_ARG, __entry->evt)
|
/linux-4.1.27/drivers/sh/intc/ |
H A D | core.c | 344 * In the case of multi-evt handling and sparse register_intc_controller()
|
/linux-4.1.27/drivers/net/wireless/cw1200/ |
H A D | wsm.c | 934 event->evt.id = WSM_GET32(buf); wsm_event_indication() 935 event->evt.data = WSM_GET32(buf); wsm_event_indication() 938 event->evt.id, event->evt.data); wsm_event_indication()
|
H A D | sta.c | 988 switch (event->evt.id) { cw1200_event_handler() 1021 int rcpi_rssi = (int)(event->evt.data & 0xFF); cw1200_event_handler()
|
/linux-4.1.27/include/uapi/linux/ |
H A D | i2o-dev.h | 102 /* Event data size = frame size - message header + evt indicator */
|
/linux-4.1.27/net/nfc/nci/ |
H A D | hci.c | 61 u8 header; /* type -cmd,evt,rsp- + instruction */
|
/linux-4.1.27/drivers/infiniband/hw/ocrdma/ |
H A D | ocrdma_hw.c | 761 struct ocrdma_ae_pvid_mcqe *evt; ocrdma_process_grp5_aync() local 767 evt = (struct ocrdma_ae_pvid_mcqe *)cqe; ocrdma_process_grp5_aync() 768 if ((evt->tag_enabled & OCRDMA_AE_PVID_MCQE_ENABLED_MASK) >> ocrdma_process_grp5_aync() 770 dev->pvid = ((evt->tag_enabled & ocrdma_process_grp5_aync() 796 pr_err("%s(%d) invalid evt code=0x%x\n", __func__, ocrdma_process_acqe()
|
/linux-4.1.27/drivers/net/wireless/ath/ath6kl/ |
H A D | wmi.c | 481 struct wmi_tx_complete_event *evt; ath6kl_wmi_tx_complete_event_rx() local 485 evt = (struct wmi_tx_complete_event *) datap; ath6kl_wmi_tx_complete_event_rx() 488 evt->num_msg, evt->msg_len, evt->msg_type); ath6kl_wmi_tx_complete_event_rx() 490 for (index = 0; index < evt->num_msg; index++) { ath6kl_wmi_tx_complete_event_rx() 1105 "bss info evt - ch %u, snr %d, rssi %d, bssid \"%pM\" " ath6kl_wmi_bssinfo_event_rx()
|
/linux-4.1.27/include/scsi/ |
H A D | scsi_device.h | 403 extern void sdev_evt_send(struct scsi_device *sdev, struct scsi_event *evt);
|
/linux-4.1.27/arch/s390/kernel/ |
H A D | time.c | 113 struct clock_event_device *evt) s390_next_event() 121 struct clock_event_device *evt) s390_set_mode() 112 s390_next_event(unsigned long delta, struct clock_event_device *evt) s390_next_event() argument 120 s390_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) s390_set_mode() argument
|
/linux-4.1.27/sound/soc/davinci/ |
H A D | davinci-mcasp.c | 1532 ret = of_property_read_u32(np, "tx-num-evt", &val); davinci_mcasp_set_pdata_from_of() 1536 ret = of_property_read_u32(np, "rx-num-evt", &val); davinci_mcasp_set_pdata_from_of()
|
/linux-4.1.27/arch/x86/lguest/ |
H A D | boot.c | 972 struct clock_event_device *evt) lguest_clockevent_set_next_event() 989 struct clock_event_device *evt) lguest_clockevent_set_mode() 971 lguest_clockevent_set_next_event(unsigned long delta, struct clock_event_device *evt) lguest_clockevent_set_next_event() argument 988 lguest_clockevent_set_mode(enum clock_event_mode mode, struct clock_event_device *evt) lguest_clockevent_set_mode() argument
|
/linux-4.1.27/include/net/bluetooth/ |
H A D | hci_core.h | 389 void (*notify)(struct hci_dev *hdev, unsigned int evt);
|
/linux-4.1.27/net/bluetooth/ |
H A D | hci_event.c | 5156 if (hdr->evt != event) hci_get_cmd_complete() 5161 if (hdr->evt != HCI_EV_CMD_COMPLETE) { hci_get_cmd_complete() 5162 BT_DBG("Last event is not cmd complete (0x%2.2x)", hdr->evt); hci_get_cmd_complete() 5189 u8 status = 0, event = hdr->evt, req_evt = 0; hci_event_packet()
|