evt 109 arch/alpha/kernel/time.c rtc_ce_set_next_event(unsigned long evt, struct clock_event_device *ce) evt 166 arch/alpha/kernel/time.c qemu_ce_set_next_event(unsigned long evt, struct clock_event_device *ce) evt 168 arch/alpha/kernel/time.c qemu_set_alarm_rel(evt); evt 271 arch/arm/kernel/perf_event_v6.c unsigned long val, mask, evt, flags; evt 279 arch/arm/kernel/perf_event_v6.c evt = ARMV6_PMCR_CCOUNT_IEN; evt 282 arch/arm/kernel/perf_event_v6.c evt = (hwc->config_base << ARMV6_PMCR_EVT_COUNT0_SHIFT) | evt 286 arch/arm/kernel/perf_event_v6.c evt = (hwc->config_base << ARMV6_PMCR_EVT_COUNT1_SHIFT) | evt 300 arch/arm/kernel/perf_event_v6.c val |= evt; evt 422 arch/arm/kernel/perf_event_v6.c unsigned long val, mask, evt, flags; evt 430 arch/arm/kernel/perf_event_v6.c evt = 0; evt 433 arch/arm/kernel/perf_event_v6.c evt = ARMV6_PERFCTR_NOP << ARMV6_PMCR_EVT_COUNT0_SHIFT; evt 436 arch/arm/kernel/perf_event_v6.c evt = ARMV6_PERFCTR_NOP << ARMV6_PMCR_EVT_COUNT1_SHIFT; evt 450 arch/arm/kernel/perf_event_v6.c val |= evt; evt 457 arch/arm/kernel/perf_event_v6.c unsigned long val, mask, flags, evt = 0; evt 481 arch/arm/kernel/perf_event_v6.c val |= evt; evt 206 arch/arm/kernel/perf_event_xscale.c unsigned long val, mask, evt, flags; evt 215 arch/arm/kernel/perf_event_xscale.c evt = XSCALE1_CCOUNT_INT_EN; evt 219 arch/arm/kernel/perf_event_xscale.c evt = (hwc->config_base << XSCALE1_COUNT0_EVT_SHFT) | evt 224 arch/arm/kernel/perf_event_xscale.c evt = (hwc->config_base << XSCALE1_COUNT1_EVT_SHFT) | evt 235 arch/arm/kernel/perf_event_xscale.c val |= evt; evt 242 arch/arm/kernel/perf_event_xscale.c unsigned long val, mask, evt, flags; evt 251 arch/arm/kernel/perf_event_xscale.c evt = 0; evt 255 arch/arm/kernel/perf_event_xscale.c evt = XSCALE_PERFCTR_UNUSED << XSCALE1_COUNT0_EVT_SHFT; evt 259 arch/arm/kernel/perf_event_xscale.c evt = XSCALE_PERFCTR_UNUSED << XSCALE1_COUNT1_EVT_SHFT; evt 269 arch/arm/kernel/perf_event_xscale.c val |= evt; evt 63 arch/arm/kernel/smp_twd.c static int twd_set_next_event(unsigned long evt, evt 70 arch/arm/kernel/smp_twd.c writel_relaxed(evt, twd_base + TWD_TIMER_COUNTER); evt 182 arch/arm/kernel/smp_twd.c struct clock_event_device *evt = dev_id; evt 185 arch/arm/kernel/smp_twd.c evt->event_handler(evt); evt 142 arch/arm/mach-cns3xxx/core.c static int cns3xxx_timer_set_next_event(unsigned long evt, evt 147 arch/arm/mach-cns3xxx/core.c writel(evt, cns3xxx_tmr1 + TIMER1_AUTO_RELOAD_OFFSET); evt 179 arch/arm/mach-cns3xxx/core.c struct clock_event_device *evt = &cns3xxx_tmr1_clockevent; evt 187 arch/arm/mach-cns3xxx/core.c evt->event_handler(evt); evt 159 arch/arm/mach-davinci/time.c struct clock_event_device *evt = &clockevent_davinci; evt 161 arch/arm/mach-davinci/time.c evt->event_handler(evt); evt 289 arch/arm/mach-davinci/time.c struct clock_event_device *evt) evt 298 arch/arm/mach-davinci/time.c static int davinci_shutdown(struct clock_event_device *evt) evt 307 arch/arm/mach-davinci/time.c static int davinci_set_oneshot(struct clock_event_device *evt) evt 316 arch/arm/mach-davinci/time.c static int davinci_set_periodic(struct clock_event_device *evt) evt 73 arch/arm/mach-ep93xx/timer-ep93xx.c struct clock_event_device *evt) evt 90 arch/arm/mach-ep93xx/timer-ep93xx.c static int ep93xx_clkevt_shutdown(struct clock_event_device *evt) evt 110 arch/arm/mach-ep93xx/timer-ep93xx.c struct clock_event_device *evt = dev_id; evt 115 arch/arm/mach-ep93xx/timer-ep93xx.c evt->event_handler(evt); evt 77 arch/arm/mach-iop32x/time.c static int iop_set_periodic(struct clock_event_device *evt) evt 90 arch/arm/mach-iop32x/time.c static int iop_set_oneshot(struct clock_event_device *evt) evt 100 arch/arm/mach-iop32x/time.c static int iop_shutdown(struct clock_event_device *evt) evt 109 arch/arm/mach-iop32x/time.c static int iop_resume(struct clock_event_device *evt) evt 133 arch/arm/mach-iop32x/time.c struct clock_event_device *evt = dev_id; evt 136 arch/arm/mach-iop32x/time.c evt->event_handler(evt); evt 118 arch/arm/mach-mmp/time.c static int timer_set_shutdown(struct clock_event_device *evt) evt 121 arch/arm/mach-omap1/time.c struct clock_event_device *evt) evt 127 arch/arm/mach-omap1/time.c static int omap_mpu_set_oneshot(struct clock_event_device *evt) evt 134 arch/arm/mach-omap1/time.c static int omap_mpu_set_periodic(struct clock_event_device *evt) evt 151 arch/arm/mach-omap1/time.c struct clock_event_device *evt = &clockevent_mpu_timer1; evt 153 arch/arm/mach-omap1/time.c evt->event_handler(evt); evt 117 arch/arm/mach-omap1/timer32k.c static int omap_32k_timer_shutdown(struct clock_event_device *evt) evt 123 arch/arm/mach-omap1/timer32k.c static int omap_32k_timer_set_periodic(struct clock_event_device *evt) evt 143 arch/arm/mach-omap1/timer32k.c struct clock_event_device *evt = &clockevent_32k_timer; evt 146 arch/arm/mach-omap1/timer32k.c evt->event_handler(evt); evt 86 arch/arm/mach-omap2/timer.c struct clock_event_device *evt = &clockevent_gpt; evt 90 arch/arm/mach-omap2/timer.c evt->event_handler(evt); evt 101 arch/arm/mach-omap2/timer.c struct clock_event_device *evt) evt 109 arch/arm/mach-omap2/timer.c static int omap2_gp_timer_shutdown(struct clock_event_device *evt) evt 115 arch/arm/mach-omap2/timer.c static int omap2_gp_timer_set_periodic(struct clock_event_device *evt) evt 69 arch/arm/mach-spear/time.c static int clockevent_next_event(unsigned long evt, evt 96 arch/arm/mach-spear/time.c static inline void timer_shutdown(struct clock_event_device *evt) evt 105 arch/arm/mach-spear/time.c static int spear_shutdown(struct clock_event_device *evt) evt 107 arch/arm/mach-spear/time.c timer_shutdown(evt); evt 112 arch/arm/mach-spear/time.c static int spear_set_oneshot(struct clock_event_device *evt) evt 117 arch/arm/mach-spear/time.c timer_shutdown(evt); evt 126 arch/arm/mach-spear/time.c static int spear_set_periodic(struct clock_event_device *evt) evt 132 arch/arm/mach-spear/time.c timer_shutdown(evt); evt 175 arch/arm/mach-spear/time.c struct clock_event_device *evt = &clkevt; evt 179 arch/arm/mach-spear/time.c evt->event_handler(evt); evt 110 arch/arm/plat-orion/time.c static int orion_clkevt_shutdown(struct clock_event_device *evt) evt 133 arch/arm/plat-orion/time.c static int orion_clkevt_set_periodic(struct clock_event_device *evt) evt 20 arch/c6x/include/asm/soc.h void (*assert_event)(unsigned int evt); evt 23 arch/c6x/kernel/soc.c void soc_assert_event(unsigned int evt) evt 26 arch/c6x/kernel/soc.c soc_ops.assert_event(evt); evt 294 arch/c6x/kernel/traps.c int evt; evt 298 arch/c6x/kernel/traps.c while ((evt = soc_get_exception()) >= 0) evt 299 arch/c6x/kernel/traps.c do_trap(&eexcept_table[evt], regs); evt 120 arch/c6x/platforms/timer64.c struct clock_event_device *evt) evt 126 arch/c6x/platforms/timer64.c static int set_periodic(struct clock_event_device *evt) evt 134 arch/c6x/platforms/timer64.c static int set_oneshot(struct clock_event_device *evt) evt 141 arch/c6x/platforms/timer64.c static int shutdown(struct clock_event_device *evt) evt 75 arch/hexagon/kernel/time.c static int set_next_event(unsigned long delta, struct clock_event_device *evt) evt 46 arch/m68k/coldfire/pit.c static int cf_pit_set_periodic(struct clock_event_device *evt) evt 56 arch/m68k/coldfire/pit.c static int cf_pit_set_oneshot(struct clock_event_device *evt) evt 64 arch/m68k/coldfire/pit.c static int cf_pit_shutdown(struct clock_event_device *evt) evt 76 arch/m68k/coldfire/pit.c struct clock_event_device *evt) evt 100 arch/m68k/coldfire/pit.c struct clock_event_device *evt = &cf_pit_clockevent; evt 108 arch/m68k/coldfire/pit.c evt->event_handler(evt); evt 126 arch/microblaze/kernel/timer.c static int xilinx_timer_shutdown(struct clock_event_device *evt) evt 133 arch/microblaze/kernel/timer.c static int xilinx_timer_set_periodic(struct clock_event_device *evt) evt 158 arch/microblaze/kernel/timer.c struct clock_event_device *evt = &clockevent_xilinx_timer; evt 160 arch/microblaze/kernel/timer.c evt->event_handler(evt); evt 31 arch/mips/kernel/cevt-bcm1480.c static int sibyte_set_periodic(struct clock_event_device *evt) evt 45 arch/mips/kernel/cevt-bcm1480.c static int sibyte_shutdown(struct clock_event_device *evt) evt 44 arch/mips/kernel/cevt-ds1287.c struct clock_event_device *evt) evt 49 arch/mips/kernel/cevt-ds1287.c static int ds1287_shutdown(struct clock_event_device *evt) evt 63 arch/mips/kernel/cevt-ds1287.c static int ds1287_set_periodic(struct clock_event_device *evt) evt 36 arch/mips/kernel/cevt-gt641xx.c struct clock_event_device *evt) evt 54 arch/mips/kernel/cevt-gt641xx.c static int gt641xx_timer0_shutdown(struct clock_event_device *evt) evt 68 arch/mips/kernel/cevt-gt641xx.c static int gt641xx_timer0_set_oneshot(struct clock_event_device *evt) evt 83 arch/mips/kernel/cevt-gt641xx.c static int gt641xx_timer0_set_periodic(struct clock_event_device *evt) evt 19 arch/mips/kernel/cevt-r4k.c struct clock_event_device *evt) evt 29 arch/mips/kernel/cevt-sb1250.c static int sibyte_shutdown(struct clock_event_device *evt) evt 41 arch/mips/kernel/cevt-sb1250.c static int sibyte_set_periodic(struct clock_event_device *evt) evt 88 arch/mips/kernel/cevt-txx9.c static int txx9tmr_set_state_periodic(struct clock_event_device *evt) evt 91 arch/mips/kernel/cevt-txx9.c container_of(evt, struct txx9_clock_event_device, cd); evt 98 arch/mips/kernel/cevt-txx9.c __raw_writel(((u64)(NSEC_PER_SEC / HZ) * evt->mult) >> evt->shift, evt 104 arch/mips/kernel/cevt-txx9.c static int txx9tmr_set_state_oneshot(struct clock_event_device *evt) evt 107 arch/mips/kernel/cevt-txx9.c container_of(evt, struct txx9_clock_event_device, cd); evt 115 arch/mips/kernel/cevt-txx9.c static int txx9tmr_set_state_shutdown(struct clock_event_device *evt) evt 118 arch/mips/kernel/cevt-txx9.c container_of(evt, struct txx9_clock_event_device, cd); evt 126 arch/mips/kernel/cevt-txx9.c static int txx9tmr_tick_resume(struct clock_event_device *evt) evt 129 arch/mips/kernel/cevt-txx9.c container_of(evt, struct txx9_clock_event_device, cd); evt 139 arch/mips/kernel/cevt-txx9.c struct clock_event_device *evt) evt 142 arch/mips/kernel/cevt-txx9.c container_of(evt, struct txx9_clock_event_device, cd); evt 315 arch/mips/kernel/perf_event_mipsxx.c static void mipsxx_pmu_enable_event(struct hw_perf_event *evt, int idx) evt 317 arch/mips/kernel/perf_event_mipsxx.c struct perf_event *event = container_of(evt, struct perf_event, hw); evt 319 arch/mips/kernel/perf_event_mipsxx.c unsigned int range = evt->event_base >> 24; evt 323 arch/mips/kernel/perf_event_mipsxx.c cpuc->saved_ctrl[idx] = M_PERFCTL_EVENT(evt->event_base & 0xff) | evt 324 arch/mips/kernel/perf_event_mipsxx.c (evt->config_base & M_PERFCTL_CONFIG_MASK) | evt 156 arch/mips/loongson32/common/time.c static int ls1x_clockevent_set_next(unsigned long evt, evt 160 arch/mips/loongson32/common/time.c ls1x_pwmtimer_set_period(evt); evt 50 arch/mips/loongson64/common/cs5536/cs5536_mfgpt.c static int mfgpt_timer_set_periodic(struct clock_event_device *evt) evt 62 arch/mips/loongson64/common/cs5536/cs5536_mfgpt.c static int mfgpt_timer_shutdown(struct clock_event_device *evt) evt 64 arch/mips/loongson64/common/cs5536/cs5536_mfgpt.c if (clockevent_state_periodic(evt) || clockevent_state_oneshot(evt)) { evt 85 arch/mips/loongson64/loongson-3/hpet.c static int hpet_set_state_periodic(struct clock_event_device *evt) evt 114 arch/mips/loongson64/loongson-3/hpet.c static int hpet_set_state_shutdown(struct clock_event_device *evt) evt 128 arch/mips/loongson64/loongson-3/hpet.c static int hpet_set_state_oneshot(struct clock_event_device *evt) evt 149 arch/mips/loongson64/loongson-3/hpet.c static int hpet_tick_resume(struct clock_event_device *evt) evt 159 arch/mips/loongson64/loongson-3/hpet.c struct clock_event_device *evt) evt 39 arch/mips/ralink/cevt-rt3352.c static int systick_set_oneshot(struct clock_event_device *evt); evt 40 arch/mips/ralink/cevt-rt3352.c static int systick_shutdown(struct clock_event_device *evt); evt 43 arch/mips/ralink/cevt-rt3352.c struct clock_event_device *evt) evt 48 arch/mips/ralink/cevt-rt3352.c sdev = container_of(evt, struct systick_device, dev); evt 91 arch/mips/ralink/cevt-rt3352.c static int systick_shutdown(struct clock_event_device *evt) evt 95 arch/mips/ralink/cevt-rt3352.c sdev = container_of(evt, struct systick_device, dev); evt 105 arch/mips/ralink/cevt-rt3352.c static int systick_set_oneshot(struct clock_event_device *evt) evt 109 arch/mips/ralink/cevt-rt3352.c sdev = container_of(evt, struct systick_device, dev); evt 41 arch/mips/sgi-ip27/ip27-timer.c static int rt_next_event(unsigned long delta, struct clock_event_device *evt) evt 17 arch/mips/sni/time.c static int a20r_set_periodic(struct clock_event_device *evt) evt 72 arch/nds32/include/asm/pmu.h int (*set_event_filter)(struct hw_perf_event *evt, evt 52 arch/nios2/kernel/time.c to_nios2_clkevent(struct clock_event_device *evt) evt 54 arch/nios2/kernel/time.c return container_of(evt, struct nios2_clockevent_dev, ced); evt 162 arch/nios2/kernel/time.c struct clock_event_device *evt) evt 164 arch/nios2/kernel/time.c struct nios2_clockevent_dev *nios2_ced = to_nios2_clkevent(evt); evt 171 arch/nios2/kernel/time.c static int nios2_timer_shutdown(struct clock_event_device *evt) evt 173 arch/nios2/kernel/time.c struct nios2_clockevent_dev *nios2_ced = to_nios2_clkevent(evt); evt 180 arch/nios2/kernel/time.c static int nios2_timer_set_periodic(struct clock_event_device *evt) evt 183 arch/nios2/kernel/time.c struct nios2_clockevent_dev *nios2_ced = to_nios2_clkevent(evt); evt 191 arch/nios2/kernel/time.c static int nios2_timer_resume(struct clock_event_device *evt) evt 193 arch/nios2/kernel/time.c struct nios2_clockevent_dev *nios2_ced = to_nios2_clkevent(evt); evt 202 arch/nios2/kernel/time.c struct clock_event_device *evt = (struct clock_event_device *) dev_id; evt 203 arch/nios2/kernel/time.c struct nios2_clockevent_dev *nios2_ced = to_nios2_clkevent(evt); evt 207 arch/nios2/kernel/time.c evt->event_handler(evt); evt 68 arch/openrisc/kernel/time.c struct clock_event_device *evt = evt 75 arch/openrisc/kernel/time.c evt->broadcast = tick_broadcast; evt 77 arch/openrisc/kernel/time.c evt->name = "openrisc_timer_clockevent", evt 78 arch/openrisc/kernel/time.c evt->features = CLOCK_EVT_FEAT_ONESHOT, evt 79 arch/openrisc/kernel/time.c evt->rating = 300, evt 80 arch/openrisc/kernel/time.c evt->set_next_event = openrisc_timer_set_next_event, evt 82 arch/openrisc/kernel/time.c evt->cpumask = cpumask_of(cpu); evt 85 arch/openrisc/kernel/time.c clockevents_config_and_register(evt, cpuinfo->clock_frequency, evt 113 arch/openrisc/kernel/time.c struct clock_event_device *evt = evt 122 arch/openrisc/kernel/time.c evt->event_handler(evt); evt 218 arch/powerpc/include/asm/mce.h extern void machine_check_print_event_info(struct machine_check_event *evt, evt 37 arch/powerpc/kernel/mce.c static void machine_check_ue_event(struct machine_check_event *evt); evt 216 arch/powerpc/kernel/mce.c static void machine_check_ue_event(struct machine_check_event *evt) evt 226 arch/powerpc/kernel/mce.c memcpy(this_cpu_ptr(&mce_ue_event_queue[index]), evt, sizeof(*evt)); evt 238 arch/powerpc/kernel/mce.c struct machine_check_event evt; evt 240 arch/powerpc/kernel/mce.c if (!get_mce_event(&evt, MCE_EVENT_RELEASE)) evt 249 arch/powerpc/kernel/mce.c memcpy(this_cpu_ptr(&mce_event_queue[index]), &evt, sizeof(evt)); evt 261 arch/powerpc/kernel/mce.c struct machine_check_event *evt; evt 265 arch/powerpc/kernel/mce.c evt = this_cpu_ptr(&mce_ue_event_queue[index]); evt 275 arch/powerpc/kernel/mce.c if (evt->error_type == MCE_ERROR_TYPE_UE) { evt 276 arch/powerpc/kernel/mce.c if (evt->u.ue_error.ignore_event) { evt 281 arch/powerpc/kernel/mce.c if (evt->u.ue_error.physical_address_provided) { evt 284 arch/powerpc/kernel/mce.c pfn = evt->u.ue_error.physical_address >> evt 303 arch/powerpc/kernel/mce.c struct machine_check_event *evt; evt 313 arch/powerpc/kernel/mce.c evt = this_cpu_ptr(&mce_event_queue[index]); evt 315 arch/powerpc/kernel/mce.c if (evt->error_type == MCE_ERROR_TYPE_UE && evt 316 arch/powerpc/kernel/mce.c evt->u.ue_error.ignore_event) { evt 320 arch/powerpc/kernel/mce.c machine_check_print_event_info(evt, false, false); evt 325 arch/powerpc/kernel/mce.c void machine_check_print_event_info(struct machine_check_event *evt, evt 388 arch/powerpc/kernel/mce.c if (evt->version != MCE_V1) { evt 390 arch/powerpc/kernel/mce.c evt->version); evt 393 arch/powerpc/kernel/mce.c switch (evt->severity) { evt 413 arch/powerpc/kernel/mce.c switch(evt->initiator) { evt 435 arch/powerpc/kernel/mce.c switch (evt->error_type) { evt 438 arch/powerpc/kernel/mce.c subtype = evt->u.ue_error.ue_error_type < evt 440 arch/powerpc/kernel/mce.c mc_ue_types[evt->u.ue_error.ue_error_type] evt 442 arch/powerpc/kernel/mce.c if (evt->u.ue_error.effective_address_provided) evt 443 arch/powerpc/kernel/mce.c ea = evt->u.ue_error.effective_address; evt 444 arch/powerpc/kernel/mce.c if (evt->u.ue_error.physical_address_provided) evt 445 arch/powerpc/kernel/mce.c pa = evt->u.ue_error.physical_address; evt 449 arch/powerpc/kernel/mce.c subtype = evt->u.slb_error.slb_error_type < evt 451 arch/powerpc/kernel/mce.c mc_slb_types[evt->u.slb_error.slb_error_type] evt 453 arch/powerpc/kernel/mce.c if (evt->u.slb_error.effective_address_provided) evt 454 arch/powerpc/kernel/mce.c ea = evt->u.slb_error.effective_address; evt 458 arch/powerpc/kernel/mce.c subtype = evt->u.erat_error.erat_error_type < evt 460 arch/powerpc/kernel/mce.c mc_erat_types[evt->u.erat_error.erat_error_type] evt 462 arch/powerpc/kernel/mce.c if (evt->u.erat_error.effective_address_provided) evt 463 arch/powerpc/kernel/mce.c ea = evt->u.erat_error.effective_address; evt 467 arch/powerpc/kernel/mce.c subtype = evt->u.tlb_error.tlb_error_type < evt 469 arch/powerpc/kernel/mce.c mc_tlb_types[evt->u.tlb_error.tlb_error_type] evt 471 arch/powerpc/kernel/mce.c if (evt->u.tlb_error.effective_address_provided) evt 472 arch/powerpc/kernel/mce.c ea = evt->u.tlb_error.effective_address; evt 476 arch/powerpc/kernel/mce.c subtype = evt->u.user_error.user_error_type < evt 478 arch/powerpc/kernel/mce.c mc_user_types[evt->u.user_error.user_error_type] evt 480 arch/powerpc/kernel/mce.c if (evt->u.user_error.effective_address_provided) evt 481 arch/powerpc/kernel/mce.c ea = evt->u.user_error.effective_address; evt 485 arch/powerpc/kernel/mce.c subtype = evt->u.ra_error.ra_error_type < evt 487 arch/powerpc/kernel/mce.c mc_ra_types[evt->u.ra_error.ra_error_type] evt 489 arch/powerpc/kernel/mce.c if (evt->u.ra_error.effective_address_provided) evt 490 arch/powerpc/kernel/mce.c ea = evt->u.ra_error.effective_address; evt 494 arch/powerpc/kernel/mce.c subtype = evt->u.link_error.link_error_type < evt 496 arch/powerpc/kernel/mce.c mc_link_types[evt->u.link_error.link_error_type] evt 498 arch/powerpc/kernel/mce.c if (evt->u.link_error.effective_address_provided) evt 499 arch/powerpc/kernel/mce.c ea = evt->u.link_error.effective_address; evt 517 arch/powerpc/kernel/mce.c if (ea && evt->srr0 != ea) { evt 527 arch/powerpc/kernel/mce.c level, evt->cpu, sevstr, in_guest ? "Guest" : "Host", evt 529 arch/powerpc/kernel/mce.c evt->disposition == MCE_DISPOSITION_RECOVERED ? evt 534 arch/powerpc/kernel/mce.c level, evt->cpu, current->pid, current->comm, evt 535 arch/powerpc/kernel/mce.c in_guest ? "Guest " : "", evt->srr0, pa_str); evt 538 arch/powerpc/kernel/mce.c level, evt->cpu, evt->srr0, (void *)evt->srr0, pa_str); evt 541 arch/powerpc/kernel/mce.c printk("%sMCE: CPU%d: Initiator %s\n", level, evt->cpu, initiator); evt 543 arch/powerpc/kernel/mce.c subtype = evt->error_class < ARRAY_SIZE(mc_error_class) ? evt 544 arch/powerpc/kernel/mce.c mc_error_class[evt->error_class] : "Unknown"; evt 545 arch/powerpc/kernel/mce.c printk("%sMCE: CPU%d: %s\n", level, evt->cpu, subtype); evt 549 arch/powerpc/kernel/mce.c if (evt->error_type == MCE_ERROR_TYPE_SLB) evt 99 arch/powerpc/kernel/time.c static int decrementer_set_next_event(unsigned long evt, evt 101 arch/powerpc/kernel/time.c static int decrementer_shutdown(struct clock_event_device *evt); evt 567 arch/powerpc/kernel/time.c struct clock_event_device *evt = this_cpu_ptr(&decrementers); evt 618 arch/powerpc/kernel/time.c if (evt->event_handler) evt 619 arch/powerpc/kernel/time.c evt->event_handler(evt); evt 974 arch/powerpc/kernel/time.c static int decrementer_set_next_event(unsigned long evt, evt 977 arch/powerpc/kernel/time.c __this_cpu_write(decrementers_next_tb, get_tb_or_rtc() + evt); evt 978 arch/powerpc/kernel/time.c set_dec(evt); evt 348 arch/powerpc/platforms/powernv/opal.c __be64 evt, len; evt 352 arch/powerpc/platforms/powernv/opal.c opal_poll_events(&evt); evt 353 arch/powerpc/platforms/powernv/opal.c if ((be64_to_cpu(evt) & OPAL_EVENT_CONSOLE_INPUT) == 0) evt 439 arch/powerpc/platforms/powernv/opal.c __be64 evt; evt 448 arch/powerpc/platforms/powernv/opal.c opal_poll_events(&evt); evt 449 arch/powerpc/platforms/powernv/opal.c if (!(be64_to_cpu(evt) & OPAL_EVENT_CONSOLE_OUTPUT)) evt 505 arch/powerpc/platforms/powernv/opal.c struct machine_check_event *evt) evt 513 arch/powerpc/platforms/powernv/opal.c } else if (evt->disposition == MCE_DISPOSITION_RECOVERED) { evt 516 arch/powerpc/platforms/powernv/opal.c } else if (evt->severity == MCE_SEV_FATAL) { evt 522 arch/powerpc/platforms/powernv/opal.c if (!recovered && evt->sync_error) { evt 594 arch/powerpc/platforms/powernv/opal.c struct machine_check_event evt; evt 596 arch/powerpc/platforms/powernv/opal.c if (!get_mce_event(&evt, MCE_EVENT_RELEASE)) evt 600 arch/powerpc/platforms/powernv/opal.c if (evt.version != MCE_V1) { evt 602 arch/powerpc/platforms/powernv/opal.c evt.version); evt 605 arch/powerpc/platforms/powernv/opal.c machine_check_print_event_info(&evt, user_mode(regs), false); evt 607 arch/powerpc/platforms/powernv/opal.c if (opal_recover_mce(regs, &evt)) evt 723 arch/powerpc/platforms/pseries/ras.c static int recover_mce(struct pt_regs *regs, struct machine_check_event *evt) evt 731 arch/powerpc/platforms/pseries/ras.c } else if (evt->disposition == MCE_DISPOSITION_RECOVERED) { evt 734 arch/powerpc/platforms/pseries/ras.c } else if (evt->severity == MCE_SEV_FATAL) { evt 740 arch/powerpc/platforms/pseries/ras.c if (!recovered && evt->sync_error) { evt 783 arch/powerpc/platforms/pseries/ras.c struct machine_check_event evt; evt 785 arch/powerpc/platforms/pseries/ras.c if (!get_mce_event(&evt, MCE_EVENT_RELEASE)) evt 789 arch/powerpc/platforms/pseries/ras.c if (evt.version != MCE_V1) { evt 791 arch/powerpc/platforms/pseries/ras.c evt.version); evt 794 arch/powerpc/platforms/pseries/ras.c machine_check_print_event_info(&evt, user_mode(regs), false); evt 796 arch/powerpc/platforms/pseries/ras.c if (recover_mce(regs, &evt)) evt 148 arch/s390/kernel/time.c struct clock_event_device *evt) evt 95 arch/sh/boards/mach-cayman/irq.c int cayman_irq_demux(int evt) evt 97 arch/sh/boards/mach-cayman/irq.c int irq = intc_evt_to_irq[evt]; evt 146 arch/sh/include/asm/uaccess.h static inline void *set_exception_table_evt(unsigned int evt, void *handler) evt 148 arch/sh/include/asm/uaccess.h return set_exception_table_vec(evt >> 5, handler); evt 105 arch/sparc/kernel/time_32.c static int timer_ce_shutdown(struct clock_event_device *evt) evt 112 arch/sparc/kernel/time_32.c static int timer_ce_set_periodic(struct clock_event_device *evt) evt 186 arch/sparc/kernel/time_32.c static int percpu_ce_shutdown(struct clock_event_device *evt) evt 188 arch/sparc/kernel/time_32.c int cpu = cpumask_first(evt->cpumask); evt 194 arch/sparc/kernel/time_32.c static int percpu_ce_set_periodic(struct clock_event_device *evt) evt 196 arch/sparc/kernel/time_32.c int cpu = cpumask_first(evt->cpumask); evt 203 arch/sparc/kernel/time_32.c struct clock_event_device *evt) evt 205 arch/sparc/kernel/time_32.c int cpu = cpumask_first(evt->cpumask); evt 695 arch/sparc/kernel/time_64.c struct clock_event_device *evt) evt 700 arch/sparc/kernel/time_64.c static int sparc64_timer_shutdown(struct clock_event_device *evt) evt 721 arch/sparc/kernel/time_64.c struct clock_event_device *evt = &per_cpu(sparc64_events, cpu); evt 730 arch/sparc/kernel/time_64.c if (unlikely(!evt->event_handler)) { evt 734 arch/sparc/kernel/time_64.c evt->event_handler(evt); evt 58 arch/um/kernel/time.c static int itimer_shutdown(struct clock_event_device *evt) evt 69 arch/um/kernel/time.c static int itimer_set_periodic(struct clock_event_device *evt) evt 86 arch/um/kernel/time.c struct clock_event_device *evt) evt 101 arch/um/kernel/time.c static int itimer_one_shot(struct clock_event_device *evt) evt 103 arch/um/kernel/time.c return itimer_next_event(0, evt); evt 46 arch/unicore32/kernel/time.c static int puv3_osmr0_shutdown(struct clock_event_device *evt) evt 464 arch/x86/kernel/apic/apic.c struct clock_event_device *evt) evt 471 arch/x86/kernel/apic/apic.c struct clock_event_device *evt) evt 480 arch/x86/kernel/apic/apic.c static int lapic_timer_shutdown(struct clock_event_device *evt) evt 485 arch/x86/kernel/apic/apic.c if (evt->features & CLOCK_EVT_FEAT_DUMMY) evt 496 arch/x86/kernel/apic/apic.c lapic_timer_set_periodic_oneshot(struct clock_event_device *evt, bool oneshot) evt 499 arch/x86/kernel/apic/apic.c if (evt->features & CLOCK_EVT_FEAT_DUMMY) evt 506 arch/x86/kernel/apic/apic.c static int lapic_timer_set_periodic(struct clock_event_device *evt) evt 508 arch/x86/kernel/apic/apic.c return lapic_timer_set_periodic_oneshot(evt, false); evt 511 arch/x86/kernel/apic/apic.c static int lapic_timer_set_oneshot(struct clock_event_device *evt) evt 513 arch/x86/kernel/apic/apic.c return lapic_timer_set_periodic_oneshot(evt, true); evt 1093 arch/x86/kernel/apic/apic.c struct clock_event_device *evt = this_cpu_ptr(&lapic_events); evt 1106 arch/x86/kernel/apic/apic.c if (!evt->event_handler) { evt 1110 arch/x86/kernel/apic/apic.c lapic_timer_shutdown(evt); evt 1119 arch/x86/kernel/apic/apic.c evt->event_handler(evt); evt 24 arch/x86/kernel/hpet.c struct clock_event_device evt; evt 70 arch/x86/kernel/hpet.c struct hpet_channel *clockevent_to_channel(struct clock_event_device *evt) evt 72 arch/x86/kernel/hpet.c return container_of(evt, struct hpet_channel, evt); evt 295 arch/x86/kernel/hpet.c static int hpet_clkevt_set_state_periodic(struct clock_event_device *evt) evt 297 arch/x86/kernel/hpet.c unsigned int channel = clockevent_to_channel(evt)->num; evt 302 arch/x86/kernel/hpet.c delta = ((uint64_t)(NSEC_PER_SEC / HZ)) * evt->mult; evt 303 arch/x86/kernel/hpet.c delta >>= evt->shift; evt 326 arch/x86/kernel/hpet.c static int hpet_clkevt_set_state_oneshot(struct clock_event_device *evt) evt 328 arch/x86/kernel/hpet.c unsigned int channel = clockevent_to_channel(evt)->num; evt 339 arch/x86/kernel/hpet.c static int hpet_clkevt_set_state_shutdown(struct clock_event_device *evt) evt 341 arch/x86/kernel/hpet.c unsigned int channel = clockevent_to_channel(evt)->num; evt 351 arch/x86/kernel/hpet.c static int hpet_clkevt_legacy_resume(struct clock_event_device *evt) evt 359 arch/x86/kernel/hpet.c hpet_clkevt_set_next_event(unsigned long delta, struct clock_event_device *evt) evt 361 arch/x86/kernel/hpet.c unsigned int channel = clockevent_to_channel(evt)->num; evt 398 arch/x86/kernel/hpet.c struct clock_event_device *evt = &hc->evt; evt 400 arch/x86/kernel/hpet.c evt->rating = rating; evt 401 arch/x86/kernel/hpet.c evt->irq = hc->irq; evt 402 arch/x86/kernel/hpet.c evt->name = hc->name; evt 403 arch/x86/kernel/hpet.c evt->cpumask = cpumask_of(hc->cpu); evt 404 arch/x86/kernel/hpet.c evt->set_state_oneshot = hpet_clkevt_set_state_oneshot; evt 405 arch/x86/kernel/hpet.c evt->set_next_event = hpet_clkevt_set_next_event; evt 406 arch/x86/kernel/hpet.c evt->set_state_shutdown = hpet_clkevt_set_state_shutdown; evt 408 arch/x86/kernel/hpet.c evt->features = CLOCK_EVT_FEAT_ONESHOT; evt 410 arch/x86/kernel/hpet.c evt->features |= CLOCK_EVT_FEAT_PERIODIC; evt 411 arch/x86/kernel/hpet.c evt->set_state_periodic = hpet_clkevt_set_state_periodic; evt 425 arch/x86/kernel/hpet.c hc->evt.tick_resume = hpet_clkevt_legacy_resume; evt 455 arch/x86/kernel/hpet.c hc->evt.features |= CLOCK_EVT_FEAT_PERIODIC; evt 456 arch/x86/kernel/hpet.c hc->evt.set_state_periodic = hpet_clkevt_set_state_periodic; evt 461 arch/x86/kernel/hpet.c clockevents_config_and_register(&hc->evt, hpet_freq, evt 463 arch/x86/kernel/hpet.c global_clock_event = &hc->evt; evt 498 arch/x86/kernel/hpet.c static int hpet_clkevt_msi_resume(struct clock_event_device *evt) evt 500 arch/x86/kernel/hpet.c struct hpet_channel *hc = clockevent_to_channel(evt); evt 514 arch/x86/kernel/hpet.c struct clock_event_device *evt = &hc->evt; evt 516 arch/x86/kernel/hpet.c if (!evt->event_handler) { evt 521 arch/x86/kernel/hpet.c evt->event_handler(evt); evt 544 arch/x86/kernel/hpet.c struct clock_event_device *evt = &hc->evt; evt 551 arch/x86/kernel/hpet.c evt->tick_resume = hpet_clkevt_msi_resume; evt 553 arch/x86/kernel/hpet.c clockevents_config_and_register(evt, hpet_freq, HPET_MIN_PROG_DELTA, evt 1091 arch/x86/kernel/hpet.c struct clock_event_device *evt = &hpet_base.channels[0].evt; evt 1094 arch/x86/kernel/hpet.c clc = (uint64_t) evt->mult * NSEC_PER_SEC; evt 1095 arch/x86/kernel/hpet.c clc >>= evt->shift + DEFAULT_RTC_SHIFT; evt 1189 arch/x86/kernel/hpet.c struct clock_event_device *evt = &hpet_base.channels[0].evt; evt 1191 arch/x86/kernel/hpet.c clc = (uint64_t) evt->mult * NSEC_PER_SEC; evt 1193 arch/x86/kernel/hpet.c clc >>= evt->shift; evt 22 arch/x86/platform/uv/uv_time.c static int uv_rtc_shutdown(struct clock_event_device *evt); evt 312 arch/x86/platform/uv/uv_time.c static int uv_rtc_shutdown(struct clock_event_device *evt) evt 314 arch/x86/platform/uv/uv_time.c int ced_cpu = cpumask_first(evt->cpumask); evt 192 arch/x86/xen/time.c static int xen_timerop_shutdown(struct clock_event_device *evt) evt 201 arch/x86/xen/time.c struct clock_event_device *evt) evt 203 arch/x86/xen/time.c WARN_ON(!clockevent_state_oneshot(evt)); evt 232 arch/x86/xen/time.c static int xen_vcpuop_shutdown(struct clock_event_device *evt) evt 245 arch/x86/xen/time.c static int xen_vcpuop_set_oneshot(struct clock_event_device *evt) evt 257 arch/x86/xen/time.c struct clock_event_device *evt) evt 263 arch/x86/xen/time.c WARN_ON(!clockevent_state_oneshot(evt)); evt 298 arch/x86/xen/time.c struct clock_event_device evt; evt 301 arch/x86/xen/time.c static DEFINE_PER_CPU(struct xen_clock_event_device, xen_clock_events) = { .evt.irq = -1 }; evt 305 arch/x86/xen/time.c struct clock_event_device *evt = this_cpu_ptr(&xen_clock_events.evt); evt 309 arch/x86/xen/time.c if (evt->event_handler) { evt 310 arch/x86/xen/time.c evt->event_handler(evt); evt 319 arch/x86/xen/time.c struct clock_event_device *evt; evt 320 arch/x86/xen/time.c evt = &per_cpu(xen_clock_events, cpu).evt; evt 322 arch/x86/xen/time.c if (evt->irq >= 0) { evt 323 arch/x86/xen/time.c unbind_from_irqhandler(evt->irq, NULL); evt 324 arch/x86/xen/time.c evt->irq = -1; evt 331 arch/x86/xen/time.c struct clock_event_device *evt = &xevt->evt; evt 334 arch/x86/xen/time.c WARN(evt->irq >= 0, "IRQ%d for CPU%d is already allocated\n", evt->irq, cpu); evt 335 arch/x86/xen/time.c if (evt->irq >= 0) evt 348 arch/x86/xen/time.c memcpy(evt, xen_clockevent, sizeof(*evt)); evt 350 arch/x86/xen/time.c evt->cpumask = cpumask_of(cpu); evt 351 arch/x86/xen/time.c evt->irq = irq; evt 357 arch/x86/xen/time.c clockevents_register_device(this_cpu_ptr(&xen_clock_events.evt)); evt 56 arch/xtensa/kernel/time.c struct clock_event_device evt; evt 83 arch/xtensa/kernel/time.c static int ccount_timer_shutdown(struct clock_event_device *evt) evt 86 arch/xtensa/kernel/time.c container_of(evt, struct ccount_timer, evt); evt 89 arch/xtensa/kernel/time.c disable_irq_nosync(evt->irq); evt 95 arch/xtensa/kernel/time.c static int ccount_timer_set_oneshot(struct clock_event_device *evt) evt 98 arch/xtensa/kernel/time.c container_of(evt, struct ccount_timer, evt); evt 101 arch/xtensa/kernel/time.c enable_irq(evt->irq); evt 108 arch/xtensa/kernel/time.c .evt = { evt 120 arch/xtensa/kernel/time.c struct clock_event_device *evt = &this_cpu_ptr(&ccount_timer)->evt; evt 123 arch/xtensa/kernel/time.c evt->event_handler(evt); evt 140 arch/xtensa/kernel/time.c struct clock_event_device *clockevent = &timer->evt; evt 202 arch/xtensa/kernel/time.c setup_irq(this_cpu_ptr(&ccount_timer)->evt.irq, &timer_irqaction); evt 58 drivers/bluetooth/btmrvl_main.c if (hdr->evt == HCI_EV_CMD_COMPLETE) { evt 339 drivers/bluetooth/btmtksdio.c if (hdr->evt == 0xe4) evt 340 drivers/bluetooth/btmtksdio.c hdr->evt = HCI_EV_VENDOR; evt 357 drivers/bluetooth/btmtksdio.c if (hdr->evt == HCI_EV_VENDOR) { evt 325 drivers/bluetooth/btmtkuart.c if (hdr->evt == 0xe4) evt 326 drivers/bluetooth/btmtkuart.c hdr->evt = HCI_EV_VENDOR; evt 343 drivers/bluetooth/btmtkuart.c if (hdr->evt == HCI_EV_VENDOR) { evt 281 drivers/bluetooth/btqca.c struct hci_ev_cmd_complete *evt; evt 284 drivers/bluetooth/btqca.c skb = bt_skb_alloc(sizeof(*hdr) + sizeof(*evt) + 1, GFP_KERNEL); evt 289 drivers/bluetooth/btqca.c hdr->evt = HCI_EV_CMD_COMPLETE; evt 290 drivers/bluetooth/btqca.c hdr->plen = sizeof(*evt) + 1; evt 292 drivers/bluetooth/btqca.c evt = skb_put(skb, sizeof(*evt)); evt 293 drivers/bluetooth/btqca.c evt->ncmd = 1; evt 294 drivers/bluetooth/btqca.c evt->opcode = cpu_to_le16(QCA_HCI_CC_OPCODE); evt 1470 drivers/bluetooth/btusb.c static void btusb_notify(struct hci_dev *hdev, unsigned int evt) evt 1474 drivers/bluetooth/btusb.c BT_DBG("%s evt %d", hdev->name, evt); evt 1710 drivers/bluetooth/btusb.c struct hci_event_hdr *evt = NULL; evt 1764 drivers/bluetooth/btusb.c evt = (struct hci_event_hdr *)(*fw_ptr); evt 1765 drivers/bluetooth/btusb.c *fw_ptr += sizeof(*evt); evt 1766 drivers/bluetooth/btusb.c remain -= sizeof(*evt); evt 1768 drivers/bluetooth/btusb.c if (remain < evt->plen) { evt 1774 drivers/bluetooth/btusb.c *fw_ptr += evt->plen; evt 1775 drivers/bluetooth/btusb.c remain -= evt->plen; evt 1782 drivers/bluetooth/btusb.c if (!evt || !evt_param || remain < 0) { evt 1788 drivers/bluetooth/btusb.c cmd_param, evt->evt, HCI_INIT_TIMEOUT); evt 1799 drivers/bluetooth/btusb.c if (skb->len != evt->plen) { evt 1806 drivers/bluetooth/btusb.c if (memcmp(skb->data, evt_param, evt->plen)) { evt 1972 drivers/bluetooth/btusb.c struct hci_ev_cmd_complete *evt; evt 1974 drivers/bluetooth/btusb.c skb = bt_skb_alloc(sizeof(*hdr) + sizeof(*evt) + 1, GFP_KERNEL); evt 1979 drivers/bluetooth/btusb.c hdr->evt = HCI_EV_CMD_COMPLETE; evt 1980 drivers/bluetooth/btusb.c hdr->plen = sizeof(*evt) + 1; evt 1982 drivers/bluetooth/btusb.c evt = skb_put(skb, sizeof(*evt)); evt 1983 drivers/bluetooth/btusb.c evt->ncmd = 0x01; evt 1984 drivers/bluetooth/btusb.c evt->opcode = cpu_to_le16(opcode); evt 2009 drivers/bluetooth/btusb.c const struct intel_bootup *evt = ptr; evt 2011 drivers/bluetooth/btusb.c if (len != sizeof(*evt)) evt 2021 drivers/bluetooth/btusb.c const struct intel_secure_send_result *evt = ptr; evt 2023 drivers/bluetooth/btusb.c if (len != sizeof(*evt)) evt 2026 drivers/bluetooth/btusb.c if (evt->result) evt 2041 drivers/bluetooth/btusb.c if (skb->len > HCI_EVENT_HDR_SIZE && hdr->evt == 0xff && evt 2582 drivers/bluetooth/btusb.c hdr->evt = 0xff; evt 543 drivers/bluetooth/hci_bcsp.c hdr.evt = 0xff; evt 447 drivers/bluetooth/hci_intel.c struct hci_ev_cmd_complete *evt; evt 449 drivers/bluetooth/hci_intel.c skb = bt_skb_alloc(sizeof(*hdr) + sizeof(*evt) + 1, GFP_KERNEL); evt 454 drivers/bluetooth/hci_intel.c hdr->evt = HCI_EV_CMD_COMPLETE; evt 455 drivers/bluetooth/hci_intel.c hdr->plen = sizeof(*evt) + 1; evt 457 drivers/bluetooth/hci_intel.c evt = skb_put(skb, sizeof(*evt)); evt 458 drivers/bluetooth/hci_intel.c evt->ncmd = 0x01; evt 459 drivers/bluetooth/hci_intel.c evt->opcode = cpu_to_le16(opcode); evt 894 drivers/bluetooth/hci_intel.c if (skb->len == 7 && hdr->evt == 0xff && hdr->plen == 0x05 && evt 907 drivers/bluetooth/hci_intel.c } else if (skb->len == 9 && hdr->evt == 0xff && hdr->plen == 0x07 && evt 120 drivers/bluetooth/hci_nokia.c u8 evt; evt 538 drivers/bluetooth/hci_nokia.c struct hci_nokia_neg_evt *evt; evt 542 drivers/bluetooth/hci_nokia.c if (hdr->dlen != sizeof(*evt)) { evt 548 drivers/bluetooth/hci_nokia.c evt = skb_pull(skb, sizeof(*hdr)); evt 550 drivers/bluetooth/hci_nokia.c if (evt->ack != NOKIA_NEG_ACK) { evt 557 drivers/bluetooth/hci_nokia.c btdev->man_id = evt->man_id; evt 558 drivers/bluetooth/hci_nokia.c btdev->ver_id = evt->ver_id; evt 561 drivers/bluetooth/hci_nokia.c evt->baud, evt->sys_clk, evt->man_id, evt->ver_id); evt 895 drivers/bluetooth/hci_qca.c if (hdr->evt == HCI_EV_VENDOR) evt 284 drivers/clocksource/arc_timer.c struct clock_event_device *evt = this_cpu_ptr(&arc_clockevent_device); evt 285 drivers/clocksource/arc_timer.c int irq_reenable = clockevent_state_periodic(evt); evt 299 drivers/clocksource/arc_timer.c evt->event_handler(evt); evt 307 drivers/clocksource/arc_timer.c struct clock_event_device *evt = this_cpu_ptr(&arc_clockevent_device); evt 309 drivers/clocksource/arc_timer.c evt->cpumask = cpumask_of(smp_processor_id()); evt 311 drivers/clocksource/arc_timer.c clockevents_config_and_register(evt, arc_timer_freq, 0, ARC_TIMERN_MAX); evt 327 drivers/clocksource/arc_timer.c struct clock_event_device *evt = this_cpu_ptr(&arc_clockevent_device); evt 344 drivers/clocksource/arc_timer.c "Timer0 (per-cpu-tick)", evt); evt 58 drivers/clocksource/arm_arch_timer.c struct clock_event_device evt; evt 61 drivers/clocksource/arm_arch_timer.c #define to_arch_timer(e) container_of(e, struct arch_timer, evt) evt 384 drivers/clocksource/arm_arch_timer.c static void erratum_set_next_event_tval_generic(const int access, unsigned long evt, evt 395 drivers/clocksource/arm_arch_timer.c cval = evt + arch_counter_get_cntpct(); evt 398 drivers/clocksource/arm_arch_timer.c cval = evt + arch_counter_get_cntvct(); evt 405 drivers/clocksource/arm_arch_timer.c static __maybe_unused int erratum_set_next_event_tval_virt(unsigned long evt, evt 408 drivers/clocksource/arm_arch_timer.c erratum_set_next_event_tval_generic(ARCH_TIMER_VIRT_ACCESS, evt, clk); evt 412 drivers/clocksource/arm_arch_timer.c static __maybe_unused int erratum_set_next_event_tval_phys(unsigned long evt, evt 415 drivers/clocksource/arm_arch_timer.c erratum_set_next_event_tval_generic(ARCH_TIMER_PHYS_ACCESS, evt, clk); evt 624 drivers/clocksource/arm_arch_timer.c struct clock_event_device *evt) evt 628 drivers/clocksource/arm_arch_timer.c ctrl = arch_timer_reg_read(access, ARCH_TIMER_REG_CTRL, evt); evt 631 drivers/clocksource/arm_arch_timer.c arch_timer_reg_write(access, ARCH_TIMER_REG_CTRL, ctrl, evt); evt 632 drivers/clocksource/arm_arch_timer.c evt->event_handler(evt); evt 641 drivers/clocksource/arm_arch_timer.c struct clock_event_device *evt = dev_id; evt 643 drivers/clocksource/arm_arch_timer.c return timer_handler(ARCH_TIMER_VIRT_ACCESS, evt); evt 648 drivers/clocksource/arm_arch_timer.c struct clock_event_device *evt = dev_id; evt 650 drivers/clocksource/arm_arch_timer.c return timer_handler(ARCH_TIMER_PHYS_ACCESS, evt); evt 655 drivers/clocksource/arm_arch_timer.c struct clock_event_device *evt = dev_id; evt 657 drivers/clocksource/arm_arch_timer.c return timer_handler(ARCH_TIMER_MEM_PHYS_ACCESS, evt); evt 662 drivers/clocksource/arm_arch_timer.c struct clock_event_device *evt = dev_id; evt 664 drivers/clocksource/arm_arch_timer.c return timer_handler(ARCH_TIMER_MEM_VIRT_ACCESS, evt); evt 699 drivers/clocksource/arm_arch_timer.c static __always_inline void set_next_event(const int access, unsigned long evt, evt 706 drivers/clocksource/arm_arch_timer.c arch_timer_reg_write(access, ARCH_TIMER_REG_TVAL, evt, clk); evt 710 drivers/clocksource/arm_arch_timer.c static int arch_timer_set_next_event_virt(unsigned long evt, evt 713 drivers/clocksource/arm_arch_timer.c set_next_event(ARCH_TIMER_VIRT_ACCESS, evt, clk); evt 717 drivers/clocksource/arm_arch_timer.c static int arch_timer_set_next_event_phys(unsigned long evt, evt 720 drivers/clocksource/arm_arch_timer.c set_next_event(ARCH_TIMER_PHYS_ACCESS, evt, clk); evt 724 drivers/clocksource/arm_arch_timer.c static int arch_timer_set_next_event_virt_mem(unsigned long evt, evt 727 drivers/clocksource/arm_arch_timer.c set_next_event(ARCH_TIMER_MEM_VIRT_ACCESS, evt, clk); evt 731 drivers/clocksource/arm_arch_timer.c static int arch_timer_set_next_event_phys_mem(unsigned long evt, evt 734 drivers/clocksource/arm_arch_timer.c set_next_event(ARCH_TIMER_MEM_PHYS_ACCESS, evt, clk); evt 1145 drivers/clocksource/arm_arch_timer.c t->evt.irq = irq; evt 1146 drivers/clocksource/arm_arch_timer.c __arch_timer_setup(ARCH_TIMER_TYPE_MEM, &t->evt); evt 1153 drivers/clocksource/arm_arch_timer.c ret = request_irq(irq, func, IRQF_TIMER, "arch_mem_timer", &t->evt); evt 113 drivers/clocksource/arm_global_timer.c static int gt_clockevent_shutdown(struct clock_event_device *evt) evt 124 drivers/clocksource/arm_global_timer.c static int gt_clockevent_set_periodic(struct clock_event_device *evt) evt 130 drivers/clocksource/arm_global_timer.c static int gt_clockevent_set_next_event(unsigned long evt, evt 133 drivers/clocksource/arm_global_timer.c gt_compare_set(evt, 0); evt 139 drivers/clocksource/arm_global_timer.c struct clock_event_device *evt = dev_id; evt 156 drivers/clocksource/arm_global_timer.c if (clockevent_state_oneshot(evt)) evt 160 drivers/clocksource/arm_global_timer.c evt->event_handler(evt); evt 110 drivers/clocksource/asm9260_timer.c struct clock_event_device *evt) evt 119 drivers/clocksource/asm9260_timer.c static inline void __asm9260_timer_shutdown(struct clock_event_device *evt) evt 125 drivers/clocksource/asm9260_timer.c static int asm9260_timer_shutdown(struct clock_event_device *evt) evt 127 drivers/clocksource/asm9260_timer.c __asm9260_timer_shutdown(evt); evt 131 drivers/clocksource/asm9260_timer.c static int asm9260_timer_set_oneshot(struct clock_event_device *evt) evt 133 drivers/clocksource/asm9260_timer.c __asm9260_timer_shutdown(evt); evt 141 drivers/clocksource/asm9260_timer.c static int asm9260_timer_set_periodic(struct clock_event_device *evt) evt 143 drivers/clocksource/asm9260_timer.c __asm9260_timer_shutdown(evt); evt 169 drivers/clocksource/asm9260_timer.c struct clock_event_device *evt = dev_id; evt 171 drivers/clocksource/asm9260_timer.c evt->event_handler(evt); evt 33 drivers/clocksource/bcm2835_timer.c struct clock_event_device evt; evt 48 drivers/clocksource/bcm2835_timer.c struct bcm2835_timer, evt); evt 61 drivers/clocksource/bcm2835_timer.c event_handler = READ_ONCE(timer->evt.event_handler); evt 63 drivers/clocksource/bcm2835_timer.c event_handler(&timer->evt); evt 111 drivers/clocksource/bcm2835_timer.c timer->evt.name = node->name; evt 112 drivers/clocksource/bcm2835_timer.c timer->evt.rating = 300; evt 113 drivers/clocksource/bcm2835_timer.c timer->evt.features = CLOCK_EVT_FEAT_ONESHOT; evt 114 drivers/clocksource/bcm2835_timer.c timer->evt.set_next_event = bcm2835_time_set_next_event; evt 115 drivers/clocksource/bcm2835_timer.c timer->evt.cpumask = cpumask_of(0); evt 127 drivers/clocksource/bcm2835_timer.c clockevents_config_and_register(&timer->evt, freq, 0xf, 0xffffffff); evt 133 drivers/clocksource/bcm_kona_timer.c static int kona_timer_shutdown(struct clock_event_device *evt) evt 156 drivers/clocksource/bcm_kona_timer.c struct clock_event_device *evt = &kona_clockevent_timer; evt 159 drivers/clocksource/bcm_kona_timer.c evt->event_handler(evt); evt 44 drivers/clocksource/clps711x-timer.c struct clock_event_device *evt = dev_id; evt 46 drivers/clocksource/clps711x-timer.c evt->event_handler(evt); evt 18 drivers/clocksource/dummy_timer.c struct clock_event_device *evt = per_cpu_ptr(&dummy_timer_evt, cpu); evt 20 drivers/clocksource/dummy_timer.c evt->name = "dummy_timer"; evt 21 drivers/clocksource/dummy_timer.c evt->features = CLOCK_EVT_FEAT_PERIODIC | evt 24 drivers/clocksource/dummy_timer.c evt->rating = 100; evt 25 drivers/clocksource/dummy_timer.c evt->cpumask = cpumask_of(cpu); evt 27 drivers/clocksource/dummy_timer.c clockevents_register_device(evt); evt 38 drivers/clocksource/dw_apb_timer.c ced_to_dw_apb_ced(struct clock_event_device *evt) evt 40 drivers/clocksource/dw_apb_timer.c return container_of(evt, struct dw_apb_clock_event_device, ced); evt 97 drivers/clocksource/dw_apb_timer.c struct clock_event_device *evt = data; evt 98 drivers/clocksource/dw_apb_timer.c struct dw_apb_clock_event_device *dw_ced = ced_to_dw_apb_ced(evt); evt 100 drivers/clocksource/dw_apb_timer.c if (!evt->event_handler) { evt 108 drivers/clocksource/dw_apb_timer.c evt->event_handler(evt); evt 121 drivers/clocksource/dw_apb_timer.c static int apbt_shutdown(struct clock_event_device *evt) evt 123 drivers/clocksource/dw_apb_timer.c struct dw_apb_clock_event_device *dw_ced = ced_to_dw_apb_ced(evt); evt 127 drivers/clocksource/dw_apb_timer.c cpumask_first(evt->cpumask)); evt 135 drivers/clocksource/dw_apb_timer.c static int apbt_set_oneshot(struct clock_event_device *evt) evt 137 drivers/clocksource/dw_apb_timer.c struct dw_apb_clock_event_device *dw_ced = ced_to_dw_apb_ced(evt); evt 141 drivers/clocksource/dw_apb_timer.c cpumask_first(evt->cpumask)); evt 167 drivers/clocksource/dw_apb_timer.c static int apbt_set_periodic(struct clock_event_device *evt) evt 169 drivers/clocksource/dw_apb_timer.c struct dw_apb_clock_event_device *dw_ced = ced_to_dw_apb_ced(evt); evt 174 drivers/clocksource/dw_apb_timer.c cpumask_first(evt->cpumask)); evt 193 drivers/clocksource/dw_apb_timer.c static int apbt_resume(struct clock_event_device *evt) evt 195 drivers/clocksource/dw_apb_timer.c struct dw_apb_clock_event_device *dw_ced = ced_to_dw_apb_ced(evt); evt 198 drivers/clocksource/dw_apb_timer.c cpumask_first(evt->cpumask)); evt 205 drivers/clocksource/dw_apb_timer.c struct clock_event_device *evt) evt 208 drivers/clocksource/dw_apb_timer.c struct dw_apb_clock_event_device *dw_ced = ced_to_dw_apb_ced(evt); evt 81 drivers/clocksource/exynos_mct.c struct clock_event_device evt; evt 284 drivers/clocksource/exynos_mct.c struct clock_event_device *evt) evt 291 drivers/clocksource/exynos_mct.c static int mct_set_state_shutdown(struct clock_event_device *evt) evt 297 drivers/clocksource/exynos_mct.c static int mct_set_state_periodic(struct clock_event_device *evt) evt 301 drivers/clocksource/exynos_mct.c cycles_per_jiffy = (((unsigned long long)NSEC_PER_SEC / HZ * evt->mult) evt 302 drivers/clocksource/exynos_mct.c >> evt->shift); evt 323 drivers/clocksource/exynos_mct.c struct clock_event_device *evt = dev_id; evt 327 drivers/clocksource/exynos_mct.c evt->event_handler(evt); evt 394 drivers/clocksource/exynos_mct.c struct clock_event_device *evt) evt 398 drivers/clocksource/exynos_mct.c mevt = container_of(evt, struct mct_clock_event_device, evt); evt 403 drivers/clocksource/exynos_mct.c static int set_state_shutdown(struct clock_event_device *evt) evt 407 drivers/clocksource/exynos_mct.c mevt = container_of(evt, struct mct_clock_event_device, evt); evt 413 drivers/clocksource/exynos_mct.c static int set_state_periodic(struct clock_event_device *evt) evt 418 drivers/clocksource/exynos_mct.c mevt = container_of(evt, struct mct_clock_event_device, evt); evt 419 drivers/clocksource/exynos_mct.c cycles_per_jiffy = (((unsigned long long)NSEC_PER_SEC / HZ * evt->mult) evt 420 drivers/clocksource/exynos_mct.c >> evt->shift); evt 429 drivers/clocksource/exynos_mct.c struct clock_event_device *evt = &mevt->evt; evt 436 drivers/clocksource/exynos_mct.c if (!clockevent_state_periodic(&mevt->evt)) evt 441 drivers/clocksource/exynos_mct.c evt->event_handler(evt); evt 450 drivers/clocksource/exynos_mct.c struct clock_event_device *evt = &mevt->evt; evt 455 drivers/clocksource/exynos_mct.c evt->name = mevt->name; evt 456 drivers/clocksource/exynos_mct.c evt->cpumask = cpumask_of(cpu); evt 457 drivers/clocksource/exynos_mct.c evt->set_next_event = exynos4_tick_set_next_event; evt 458 drivers/clocksource/exynos_mct.c evt->set_state_periodic = set_state_periodic; evt 459 drivers/clocksource/exynos_mct.c evt->set_state_shutdown = set_state_shutdown; evt 460 drivers/clocksource/exynos_mct.c evt->set_state_oneshot = set_state_shutdown; evt 461 drivers/clocksource/exynos_mct.c evt->set_state_oneshot_stopped = set_state_shutdown; evt 462 drivers/clocksource/exynos_mct.c evt->tick_resume = set_state_shutdown; evt 463 drivers/clocksource/exynos_mct.c evt->features = CLOCK_EVT_FEAT_PERIODIC | CLOCK_EVT_FEAT_ONESHOT; evt 464 drivers/clocksource/exynos_mct.c evt->rating = 500; /* use value higher than ARM arch timer */ evt 470 drivers/clocksource/exynos_mct.c if (evt->irq == -1) evt 473 drivers/clocksource/exynos_mct.c irq_force_affinity(evt->irq, cpumask_of(cpu)); evt 474 drivers/clocksource/exynos_mct.c enable_irq(evt->irq); evt 478 drivers/clocksource/exynos_mct.c clockevents_config_and_register(evt, clk_rate / (TICK_BASE_CNT + 1), evt 488 drivers/clocksource/exynos_mct.c struct clock_event_device *evt = &mevt->evt; evt 490 drivers/clocksource/exynos_mct.c evt->set_state_shutdown(evt); evt 492 drivers/clocksource/exynos_mct.c if (evt->irq != -1) evt 493 drivers/clocksource/exynos_mct.c disable_irq_nosync(evt->irq); evt 533 drivers/clocksource/exynos_mct.c pcpu_mevt->evt.irq = -1; evt 545 drivers/clocksource/exynos_mct.c pcpu_mevt->evt.irq = mct_irq; evt 567 drivers/clocksource/exynos_mct.c if (pcpu_mevt->evt.irq != -1) { evt 568 drivers/clocksource/exynos_mct.c free_irq(pcpu_mevt->evt.irq, pcpu_mevt); evt 569 drivers/clocksource/exynos_mct.c pcpu_mevt->evt.irq = -1; evt 55 drivers/clocksource/hyperv_timer.c struct clock_event_device *evt) evt 65 drivers/clocksource/hyperv_timer.c static int hv_ce_shutdown(struct clock_event_device *evt) evt 75 drivers/clocksource/hyperv_timer.c static int hv_ce_set_oneshot(struct clock_event_device *evt) evt 111 drivers/clocksource/i8253.c static int pit_shutdown(struct clock_event_device *evt) evt 113 drivers/clocksource/i8253.c if (!clockevent_state_oneshot(evt) && !clockevent_state_periodic(evt)) evt 129 drivers/clocksource/i8253.c static int pit_set_oneshot(struct clock_event_device *evt) evt 137 drivers/clocksource/i8253.c static int pit_set_periodic(struct clock_event_device *evt) evt 155 drivers/clocksource/i8253.c static int pit_next_event(unsigned long delta, struct clock_event_device *evt) evt 55 drivers/clocksource/ingenic-timer.c static inline struct ingenic_tcu *to_ingenic_tcu(struct clock_event_device *evt) evt 57 drivers/clocksource/ingenic-timer.c return container_of(evt, struct ingenic_tcu, cevt); evt 60 drivers/clocksource/ingenic-timer.c static int ingenic_tcu_cevt_set_state_shutdown(struct clock_event_device *evt) evt 62 drivers/clocksource/ingenic-timer.c struct ingenic_tcu *tcu = to_ingenic_tcu(evt); evt 70 drivers/clocksource/ingenic-timer.c struct clock_event_device *evt) evt 72 drivers/clocksource/ingenic-timer.c struct ingenic_tcu *tcu = to_ingenic_tcu(evt); evt 86 drivers/clocksource/ingenic-timer.c struct clock_event_device *evt = dev_id; evt 87 drivers/clocksource/ingenic-timer.c struct ingenic_tcu *tcu = to_ingenic_tcu(evt); evt 91 drivers/clocksource/ingenic-timer.c if (evt->event_handler) evt 92 drivers/clocksource/ingenic-timer.c evt->event_handler(evt); evt 43 drivers/clocksource/mips-gic-timer.c static int gic_next_event(unsigned long delta, struct clock_event_device *evt) evt 45 drivers/clocksource/mips-gic-timer.c int cpu = cpumask_first(evt->cpumask); evt 92 drivers/clocksource/mxs_timer.c static int timrotv1_set_next_event(unsigned long evt, evt 96 drivers/clocksource/mxs_timer.c __raw_writel(evt, mxs_timrot_base + HW_TIMROT_TIMCOUNTn(0)); evt 101 drivers/clocksource/mxs_timer.c static int timrotv2_set_next_event(unsigned long evt, evt 105 drivers/clocksource/mxs_timer.c __raw_writel(evt, mxs_timrot_base + HW_TIMROT_FIXED_COUNTn(0)); evt 112 drivers/clocksource/mxs_timer.c struct clock_event_device *evt = dev_id; evt 115 drivers/clocksource/mxs_timer.c evt->event_handler(evt); evt 147 drivers/clocksource/mxs_timer.c static int mxs_shutdown(struct clock_event_device *evt) evt 154 drivers/clocksource/mxs_timer.c static int mxs_set_oneshot(struct clock_event_device *evt) evt 156 drivers/clocksource/mxs_timer.c if (clockevent_state_oneshot(evt)) evt 88 drivers/clocksource/nomadik-mtu.c static int nmdk_clkevt_next(unsigned long evt, struct clock_event_device *ev) evt 91 drivers/clocksource/nomadik-mtu.c writel(evt, mtu_base + MTU_LR(1)); evt 117 drivers/clocksource/nomadik-mtu.c static int nmdk_clkevt_shutdown(struct clock_event_device *evt) evt 127 drivers/clocksource/nomadik-mtu.c static int nmdk_clkevt_set_oneshot(struct clock_event_device *evt) evt 133 drivers/clocksource/nomadik-mtu.c static int nmdk_clkevt_set_periodic(struct clock_event_device *evt) evt 186 drivers/clocksource/samsung_pwm_timer.c struct clock_event_device *evt) evt 207 drivers/clocksource/samsung_pwm_timer.c static int samsung_shutdown(struct clock_event_device *evt) evt 213 drivers/clocksource/samsung_pwm_timer.c static int samsung_set_periodic(struct clock_event_device *evt) evt 247 drivers/clocksource/samsung_pwm_timer.c struct clock_event_device *evt = dev_id; evt 254 drivers/clocksource/samsung_pwm_timer.c evt->event_handler(evt); evt 126 drivers/clocksource/timer-armada-370-xp.c static int armada_370_xp_clkevt_shutdown(struct clock_event_device *evt) evt 140 drivers/clocksource/timer-armada-370-xp.c static int armada_370_xp_clkevt_set_periodic(struct clock_event_device *evt) evt 162 drivers/clocksource/timer-armada-370-xp.c struct clock_event_device *evt = dev_id; evt 165 drivers/clocksource/timer-armada-370-xp.c evt->event_handler(evt); evt 175 drivers/clocksource/timer-armada-370-xp.c struct clock_event_device *evt = per_cpu_ptr(armada_370_xp_evt, cpu); evt 184 drivers/clocksource/timer-armada-370-xp.c evt->name = "armada_370_xp_per_cpu_tick", evt 185 drivers/clocksource/timer-armada-370-xp.c evt->features = CLOCK_EVT_FEAT_ONESHOT | evt 187 drivers/clocksource/timer-armada-370-xp.c evt->shift = 32, evt 188 drivers/clocksource/timer-armada-370-xp.c evt->rating = 300, evt 189 drivers/clocksource/timer-armada-370-xp.c evt->set_next_event = armada_370_xp_clkevt_next_event, evt 190 drivers/clocksource/timer-armada-370-xp.c evt->set_state_shutdown = armada_370_xp_clkevt_shutdown; evt 191 drivers/clocksource/timer-armada-370-xp.c evt->set_state_periodic = armada_370_xp_clkevt_set_periodic; evt 192 drivers/clocksource/timer-armada-370-xp.c evt->set_state_oneshot = armada_370_xp_clkevt_shutdown; evt 193 drivers/clocksource/timer-armada-370-xp.c evt->tick_resume = armada_370_xp_clkevt_shutdown; evt 194 drivers/clocksource/timer-armada-370-xp.c evt->irq = armada_370_xp_clkevt_irq; evt 195 drivers/clocksource/timer-armada-370-xp.c evt->cpumask = cpumask_of(cpu); evt 197 drivers/clocksource/timer-armada-370-xp.c clockevents_config_and_register(evt, timer_clk, 1, 0xfffffffe); evt 198 drivers/clocksource/timer-armada-370-xp.c enable_percpu_irq(evt->irq, 0); evt 205 drivers/clocksource/timer-armada-370-xp.c struct clock_event_device *evt = per_cpu_ptr(armada_370_xp_evt, cpu); evt 207 drivers/clocksource/timer-armada-370-xp.c evt->set_state_shutdown(evt); evt 208 drivers/clocksource/timer-armada-370-xp.c disable_percpu_irq(evt->irq); evt 121 drivers/clocksource/timer-atcpit100.c static int atcpit100_clkevt_next_event(unsigned long evt, evt 129 drivers/clocksource/timer-atcpit100.c writel(evt, timer_of_base(to) + CH0_REL); evt 135 drivers/clocksource/timer-atcpit100.c static int atcpit100_clkevt_set_periodic(struct clock_event_device *evt) evt 137 drivers/clocksource/timer-atcpit100.c struct timer_of *to = to_timer_of(evt); evt 144 drivers/clocksource/timer-atcpit100.c static int atcpit100_clkevt_shutdown(struct clock_event_device *evt) evt 146 drivers/clocksource/timer-atcpit100.c struct timer_of *to = to_timer_of(evt); evt 152 drivers/clocksource/timer-atcpit100.c static int atcpit100_clkevt_set_oneshot(struct clock_event_device *evt) evt 154 drivers/clocksource/timer-atcpit100.c struct timer_of *to = to_timer_of(evt); evt 166 drivers/clocksource/timer-atcpit100.c struct clock_event_device *evt = (struct clock_event_device *)dev_id; evt 167 drivers/clocksource/timer-atcpit100.c struct timer_of *to = to_timer_of(evt); evt 171 drivers/clocksource/timer-atcpit100.c evt->event_handler(evt); evt 120 drivers/clocksource/timer-atlas7.c static int sirfsoc_timer_shutdown(struct clock_event_device *evt) evt 105 drivers/clocksource/timer-atmel-st.c static int clkevt32k_shutdown(struct clock_event_device *evt) evt 176 drivers/clocksource/timer-cadence-ttc.c struct clock_event_device *evt) evt 178 drivers/clocksource/timer-cadence-ttc.c struct ttc_timer_clockevent *ttce = to_ttc_timer_clkevent(evt); evt 190 drivers/clocksource/timer-cadence-ttc.c static int ttc_shutdown(struct clock_event_device *evt) evt 192 drivers/clocksource/timer-cadence-ttc.c struct ttc_timer_clockevent *ttce = to_ttc_timer_clkevent(evt); evt 202 drivers/clocksource/timer-cadence-ttc.c static int ttc_set_periodic(struct clock_event_device *evt) evt 204 drivers/clocksource/timer-cadence-ttc.c struct ttc_timer_clockevent *ttce = to_ttc_timer_clkevent(evt); evt 212 drivers/clocksource/timer-cadence-ttc.c static int ttc_resume(struct clock_event_device *evt) evt 214 drivers/clocksource/timer-cadence-ttc.c struct ttc_timer_clockevent *ttce = to_ttc_timer_clkevent(evt); evt 76 drivers/clocksource/timer-cs5535.c static int mfgpt_shutdown(struct clock_event_device *evt) evt 82 drivers/clocksource/timer-cs5535.c static int mfgpt_set_periodic(struct clock_event_device *evt) evt 89 drivers/clocksource/timer-cs5535.c static int mfgpt_next_event(unsigned long delta, struct clock_event_device *evt) evt 114 drivers/clocksource/timer-digicolor.c static int digicolor_clkevt_next_event(unsigned long evt, evt 118 drivers/clocksource/timer-digicolor.c dc_timer_set_count(ce, evt); evt 140 drivers/clocksource/timer-digicolor.c struct clock_event_device *evt = dev_id; evt 142 drivers/clocksource/timer-digicolor.c evt->event_handler(evt); evt 86 drivers/clocksource/timer-efm32.c static int efm32_clock_event_set_next_event(unsigned long evt, evt 93 drivers/clocksource/timer-efm32.c writel_relaxed(evt, ddata->base + TIMERn_CNT); evt 141 drivers/clocksource/timer-fsl-ftm.c static int ftm_set_oneshot(struct clock_event_device *evt) evt 147 drivers/clocksource/timer-fsl-ftm.c static int ftm_set_periodic(struct clock_event_device *evt) evt 149 drivers/clocksource/timer-fsl-ftm.c ftm_set_next_event(priv->periodic_cyc, evt); evt 155 drivers/clocksource/timer-fsl-ftm.c struct clock_event_device *evt = dev_id; evt 159 drivers/clocksource/timer-fsl-ftm.c if (likely(clockevent_state_oneshot(evt))) { evt 164 drivers/clocksource/timer-fsl-ftm.c evt->event_handler(evt); evt 111 drivers/clocksource/timer-fttmr010.c static inline struct fttmr010 *to_fttmr010(struct clock_event_device *evt) evt 113 drivers/clocksource/timer-fttmr010.c return container_of(evt, struct fttmr010, clkevt); evt 137 drivers/clocksource/timer-fttmr010.c struct clock_event_device *evt) evt 139 drivers/clocksource/timer-fttmr010.c struct fttmr010 *fttmr010 = to_fttmr010(evt); evt 167 drivers/clocksource/timer-fttmr010.c static int fttmr010_timer_shutdown(struct clock_event_device *evt) evt 169 drivers/clocksource/timer-fttmr010.c struct fttmr010 *fttmr010 = to_fttmr010(evt); evt 180 drivers/clocksource/timer-fttmr010.c static int fttmr010_timer_set_oneshot(struct clock_event_device *evt) evt 182 drivers/clocksource/timer-fttmr010.c struct fttmr010 *fttmr010 = to_fttmr010(evt); evt 207 drivers/clocksource/timer-fttmr010.c static int fttmr010_timer_set_periodic(struct clock_event_device *evt) evt 209 drivers/clocksource/timer-fttmr010.c struct fttmr010 *fttmr010 = to_fttmr010(evt); evt 246 drivers/clocksource/timer-fttmr010.c struct clock_event_device *evt = dev_id; evt 248 drivers/clocksource/timer-fttmr010.c evt->event_handler(evt); evt 81 drivers/clocksource/timer-imx-gpt.c int (*set_next_event)(unsigned long evt, evt 173 drivers/clocksource/timer-imx-gpt.c static int mx1_2_set_next_event(unsigned long evt, evt 179 drivers/clocksource/timer-imx-gpt.c tcmp = readl_relaxed(imxtm->base + MX1_2_TCN) + evt; evt 187 drivers/clocksource/timer-imx-gpt.c static int v2_set_next_event(unsigned long evt, evt 193 drivers/clocksource/timer-imx-gpt.c tcmp = readl_relaxed(imxtm->base + V2_TCN) + evt; evt 197 drivers/clocksource/timer-imx-gpt.c return evt < 0x7fffffff && evt 57 drivers/clocksource/timer-imx-sysctr.c struct clock_event_device *evt) evt 79 drivers/clocksource/timer-imx-sysctr.c static int sysctr_set_state_oneshot(struct clock_event_device *evt) evt 84 drivers/clocksource/timer-imx-sysctr.c static int sysctr_set_state_shutdown(struct clock_event_device *evt) evt 93 drivers/clocksource/timer-imx-sysctr.c struct clock_event_device *evt = dev_id; evt 97 drivers/clocksource/timer-imx-sysctr.c evt->event_handler(evt); evt 84 drivers/clocksource/timer-imx-tpm.c struct clock_event_device *evt) evt 102 drivers/clocksource/timer-imx-tpm.c static int tpm_set_state_oneshot(struct clock_event_device *evt) evt 109 drivers/clocksource/timer-imx-tpm.c static int tpm_set_state_shutdown(struct clock_event_device *evt) evt 118 drivers/clocksource/timer-imx-tpm.c struct clock_event_device *evt = dev_id; evt 122 drivers/clocksource/timer-imx-tpm.c evt->event_handler(evt); evt 60 drivers/clocksource/timer-integrator-ap.c struct clock_event_device *evt = dev_id; evt 65 drivers/clocksource/timer-integrator-ap.c evt->event_handler(evt); evt 70 drivers/clocksource/timer-integrator-ap.c static int clkevt_shutdown(struct clock_event_device *evt) evt 79 drivers/clocksource/timer-integrator-ap.c static int clkevt_set_oneshot(struct clock_event_device *evt) evt 89 drivers/clocksource/timer-integrator-ap.c static int clkevt_set_periodic(struct clock_event_device *evt) evt 103 drivers/clocksource/timer-integrator-ap.c static int clkevt_set_next_event(unsigned long next, struct clock_event_device *evt) evt 73 drivers/clocksource/timer-ixp4xx.c to_ixp4xx_timer(struct clock_event_device *evt) evt 75 drivers/clocksource/timer-ixp4xx.c return container_of(evt, struct ixp4xx_timer, clkevt); evt 96 drivers/clocksource/timer-ixp4xx.c struct clock_event_device *evt = &tmr->clkevt; evt 102 drivers/clocksource/timer-ixp4xx.c evt->event_handler(evt); evt 108 drivers/clocksource/timer-ixp4xx.c struct clock_event_device *evt) evt 110 drivers/clocksource/timer-ixp4xx.c struct ixp4xx_timer *tmr = to_ixp4xx_timer(evt); evt 122 drivers/clocksource/timer-ixp4xx.c static int ixp4xx_shutdown(struct clock_event_device *evt) evt 124 drivers/clocksource/timer-ixp4xx.c struct ixp4xx_timer *tmr = to_ixp4xx_timer(evt); evt 134 drivers/clocksource/timer-ixp4xx.c static int ixp4xx_set_oneshot(struct clock_event_device *evt) evt 136 drivers/clocksource/timer-ixp4xx.c struct ixp4xx_timer *tmr = to_ixp4xx_timer(evt); evt 144 drivers/clocksource/timer-ixp4xx.c static int ixp4xx_set_periodic(struct clock_event_device *evt) evt 146 drivers/clocksource/timer-ixp4xx.c struct ixp4xx_timer *tmr = to_ixp4xx_timer(evt); evt 156 drivers/clocksource/timer-ixp4xx.c static int ixp4xx_resume(struct clock_event_device *evt) evt 158 drivers/clocksource/timer-ixp4xx.c struct ixp4xx_timer *tmr = to_ixp4xx_timer(evt); evt 119 drivers/clocksource/timer-keystone.c struct clock_event_device *evt = dev_id; evt 121 drivers/clocksource/timer-keystone.c evt->event_handler(evt); evt 126 drivers/clocksource/timer-keystone.c struct clock_event_device *evt) evt 131 drivers/clocksource/timer-keystone.c static int keystone_shutdown(struct clock_event_device *evt) evt 137 drivers/clocksource/timer-keystone.c static int keystone_set_periodic(struct clock_event_device *evt) evt 101 drivers/clocksource/timer-meson6.c static int meson6_shutdown(struct clock_event_device *evt) evt 107 drivers/clocksource/timer-meson6.c static int meson6_set_oneshot(struct clock_event_device *evt) evt 114 drivers/clocksource/timer-meson6.c static int meson6_set_periodic(struct clock_event_device *evt) evt 122 drivers/clocksource/timer-meson6.c static int meson6_clkevt_next_event(unsigned long evt, evt 126 drivers/clocksource/timer-meson6.c meson6_clkevt_time_setup(evt); evt 146 drivers/clocksource/timer-meson6.c struct clock_event_device *evt = (struct clock_event_device *)dev_id; evt 148 drivers/clocksource/timer-meson6.c evt->event_handler(evt); evt 56 drivers/clocksource/timer-npcm7xx.c static int npcm7xx_timer_resume(struct clock_event_device *evt) evt 58 drivers/clocksource/timer-npcm7xx.c struct timer_of *to = to_timer_of(evt); evt 68 drivers/clocksource/timer-npcm7xx.c static int npcm7xx_timer_shutdown(struct clock_event_device *evt) evt 70 drivers/clocksource/timer-npcm7xx.c struct timer_of *to = to_timer_of(evt); evt 80 drivers/clocksource/timer-npcm7xx.c static int npcm7xx_timer_oneshot(struct clock_event_device *evt) evt 82 drivers/clocksource/timer-npcm7xx.c struct timer_of *to = to_timer_of(evt); evt 93 drivers/clocksource/timer-npcm7xx.c static int npcm7xx_timer_periodic(struct clock_event_device *evt) evt 95 drivers/clocksource/timer-npcm7xx.c struct timer_of *to = to_timer_of(evt); evt 108 drivers/clocksource/timer-npcm7xx.c static int npcm7xx_clockevent_set_next_event(unsigned long evt, evt 114 drivers/clocksource/timer-npcm7xx.c writel(evt, timer_of_base(to) + NPCM7XX_REG_TICR0); evt 124 drivers/clocksource/timer-npcm7xx.c struct clock_event_device *evt = (struct clock_event_device *)dev_id; evt 125 drivers/clocksource/timer-npcm7xx.c struct timer_of *to = to_timer_of(evt); evt 129 drivers/clocksource/timer-npcm7xx.c evt->event_handler(evt); evt 217 drivers/clocksource/timer-nps.c struct clock_event_device *evt = dev_id; evt 220 drivers/clocksource/timer-nps.c evt->event_handler(evt); evt 227 drivers/clocksource/timer-nps.c struct clock_event_device *evt = this_cpu_ptr(&nps_clockevent_device); evt 229 drivers/clocksource/timer-nps.c evt->cpumask = cpumask_of(smp_processor_id()); evt 231 drivers/clocksource/timer-nps.c clockevents_config_and_register(evt, nps_timer0_freq, 0, ULONG_MAX); evt 61 drivers/clocksource/timer-owl.c static int owl_timer_set_state_shutdown(struct clock_event_device *evt) evt 68 drivers/clocksource/timer-owl.c static int owl_timer_set_state_oneshot(struct clock_event_device *evt) evt 75 drivers/clocksource/timer-owl.c static int owl_timer_tick_resume(struct clock_event_device *evt) evt 80 drivers/clocksource/timer-owl.c static int owl_timer_set_next_event(unsigned long evt, evt 88 drivers/clocksource/timer-owl.c writel(evt, base + OWL_Tx_CMP); evt 107 drivers/clocksource/timer-owl.c struct clock_event_device *evt = (struct clock_event_device *)dev_id; evt 111 drivers/clocksource/timer-owl.c evt->event_handler(evt); evt 87 drivers/clocksource/timer-oxnas-rps.c static int oxnas_rps_timer_shutdown(struct clock_event_device *evt) evt 90 drivers/clocksource/timer-oxnas-rps.c container_of(evt, struct oxnas_rps_timer, clkevent); evt 97 drivers/clocksource/timer-oxnas-rps.c static int oxnas_rps_timer_set_periodic(struct clock_event_device *evt) evt 100 drivers/clocksource/timer-oxnas-rps.c container_of(evt, struct oxnas_rps_timer, clkevent); evt 107 drivers/clocksource/timer-oxnas-rps.c static int oxnas_rps_timer_set_oneshot(struct clock_event_device *evt) evt 110 drivers/clocksource/timer-oxnas-rps.c container_of(evt, struct oxnas_rps_timer, clkevent); evt 118 drivers/clocksource/timer-oxnas-rps.c struct clock_event_device *evt) evt 121 drivers/clocksource/timer-oxnas-rps.c container_of(evt, struct oxnas_rps_timer, clkevent); evt 105 drivers/clocksource/timer-prima2.c static int sirfsoc_timer_shutdown(struct clock_event_device *evt) evt 114 drivers/clocksource/timer-prima2.c static int sirfsoc_timer_set_oneshot(struct clock_event_device *evt) evt 91 drivers/clocksource/timer-pxa.c static int pxa_osmr0_shutdown(struct clock_event_device *evt) evt 39 drivers/clocksource/timer-qcom.c struct clock_event_device *evt = dev_id; evt 41 drivers/clocksource/timer-qcom.c if (clockevent_state_oneshot(evt)) { evt 46 drivers/clocksource/timer-qcom.c evt->event_handler(evt); evt 51 drivers/clocksource/timer-qcom.c struct clock_event_device *evt) evt 69 drivers/clocksource/timer-qcom.c static int msm_timer_shutdown(struct clock_event_device *evt) evt 101 drivers/clocksource/timer-qcom.c struct clock_event_device *evt = per_cpu_ptr(msm_evt, cpu); evt 104 drivers/clocksource/timer-qcom.c evt->irq = msm_timer_irq; evt 105 drivers/clocksource/timer-qcom.c evt->name = "msm_timer"; evt 106 drivers/clocksource/timer-qcom.c evt->features = CLOCK_EVT_FEAT_ONESHOT; evt 107 drivers/clocksource/timer-qcom.c evt->rating = 200; evt 108 drivers/clocksource/timer-qcom.c evt->set_state_shutdown = msm_timer_shutdown; evt 109 drivers/clocksource/timer-qcom.c evt->set_state_oneshot = msm_timer_shutdown; evt 110 drivers/clocksource/timer-qcom.c evt->tick_resume = msm_timer_shutdown; evt 111 drivers/clocksource/timer-qcom.c evt->set_next_event = msm_timer_set_next_event; evt 112 drivers/clocksource/timer-qcom.c evt->cpumask = cpumask_of(cpu); evt 114 drivers/clocksource/timer-qcom.c clockevents_config_and_register(evt, GPT_HZ, 4, 0xffffffff); evt 117 drivers/clocksource/timer-qcom.c enable_percpu_irq(evt->irq, IRQ_TYPE_EDGE_RISING); evt 119 drivers/clocksource/timer-qcom.c err = request_irq(evt->irq, msm_timer_interrupt, evt 121 drivers/clocksource/timer-qcom.c IRQF_TRIGGER_RISING, "gp_timer", evt); evt 131 drivers/clocksource/timer-qcom.c struct clock_event_device *evt = per_cpu_ptr(msm_evt, cpu); evt 133 drivers/clocksource/timer-qcom.c evt->set_state_shutdown(evt); evt 134 drivers/clocksource/timer-qcom.c disable_percpu_irq(evt->irq); evt 67 drivers/clocksource/timer-rda.c static int rda_ostimer_set_state_shutdown(struct clock_event_device *evt) evt 69 drivers/clocksource/timer-rda.c struct timer_of *to = to_timer_of(evt); evt 76 drivers/clocksource/timer-rda.c static int rda_ostimer_set_state_oneshot(struct clock_event_device *evt) evt 78 drivers/clocksource/timer-rda.c struct timer_of *to = to_timer_of(evt); evt 85 drivers/clocksource/timer-rda.c static int rda_ostimer_set_state_periodic(struct clock_event_device *evt) evt 87 drivers/clocksource/timer-rda.c struct timer_of *to = to_timer_of(evt); evt 93 drivers/clocksource/timer-rda.c evt->mult) >> evt->shift; evt 99 drivers/clocksource/timer-rda.c static int rda_ostimer_tick_resume(struct clock_event_device *evt) evt 104 drivers/clocksource/timer-rda.c static int rda_ostimer_set_next_event(unsigned long evt, evt 109 drivers/clocksource/timer-rda.c rda_ostimer_start(timer_of_base(to), false, evt); evt 116 drivers/clocksource/timer-rda.c struct clock_event_device *evt = dev_id; evt 117 drivers/clocksource/timer-rda.c struct timer_of *to = to_timer_of(evt); evt 123 drivers/clocksource/timer-rda.c if (evt->event_handler) evt 124 drivers/clocksource/timer-rda.c evt->event_handler(evt); evt 115 drivers/clocksource/timer-sp804.c struct clock_event_device *evt = dev_id; evt 120 drivers/clocksource/timer-sp804.c evt->event_handler(evt); evt 125 drivers/clocksource/timer-sp804.c static inline void timer_shutdown(struct clock_event_device *evt) evt 130 drivers/clocksource/timer-sp804.c static int sp804_shutdown(struct clock_event_device *evt) evt 132 drivers/clocksource/timer-sp804.c timer_shutdown(evt); evt 136 drivers/clocksource/timer-sp804.c static int sp804_set_periodic(struct clock_event_device *evt) evt 141 drivers/clocksource/timer-sp804.c timer_shutdown(evt); evt 148 drivers/clocksource/timer-sp804.c struct clock_event_device *evt) evt 180 drivers/clocksource/timer-sp804.c struct clock_event_device *evt = &sp804_clockevent; evt 197 drivers/clocksource/timer-sp804.c evt->name = name; evt 198 drivers/clocksource/timer-sp804.c evt->irq = irq; evt 199 drivers/clocksource/timer-sp804.c evt->cpumask = cpu_possible_mask; evt 204 drivers/clocksource/timer-sp804.c clockevents_config_and_register(evt, rate, 0xf, 0xffffffff); evt 126 drivers/clocksource/timer-stm32.c static int stm32_clock_event_set_next_event(unsigned long evt, evt 132 drivers/clocksource/timer-stm32.c next = readl_relaxed(timer_of_base(to) + TIM_CNT) + evt; evt 136 drivers/clocksource/timer-stm32.c if ((next - now) > evt) evt 85 drivers/clocksource/timer-sun4i.c static int sun4i_clkevt_shutdown(struct clock_event_device *evt) evt 87 drivers/clocksource/timer-sun4i.c struct timer_of *to = to_timer_of(evt); evt 94 drivers/clocksource/timer-sun4i.c static int sun4i_clkevt_set_oneshot(struct clock_event_device *evt) evt 96 drivers/clocksource/timer-sun4i.c struct timer_of *to = to_timer_of(evt); evt 104 drivers/clocksource/timer-sun4i.c static int sun4i_clkevt_set_periodic(struct clock_event_device *evt) evt 106 drivers/clocksource/timer-sun4i.c struct timer_of *to = to_timer_of(evt); evt 115 drivers/clocksource/timer-sun4i.c static int sun4i_clkevt_next_event(unsigned long evt, evt 121 drivers/clocksource/timer-sun4i.c sun4i_clkevt_time_setup(timer_of_base(to), 0, evt - TIMER_SYNC_TICKS); evt 134 drivers/clocksource/timer-sun4i.c struct clock_event_device *evt = (struct clock_event_device *)dev_id; evt 135 drivers/clocksource/timer-sun4i.c struct timer_of *to = to_timer_of(evt); evt 138 drivers/clocksource/timer-sun4i.c evt->event_handler(evt); evt 134 drivers/clocksource/timer-sun5i.c static int sun5i_clkevt_next_event(unsigned long evt, evt 140 drivers/clocksource/timer-sun5i.c sun5i_clkevt_time_setup(ce, 0, evt - TIMER_SYNC_TICKS); evt 55 drivers/clocksource/timer-tegra.c struct clock_event_device *evt) evt 57 drivers/clocksource/timer-tegra.c void __iomem *reg_base = timer_of_base(to_timer_of(evt)); evt 73 drivers/clocksource/timer-tegra.c static int tegra_timer_shutdown(struct clock_event_device *evt) evt 75 drivers/clocksource/timer-tegra.c void __iomem *reg_base = timer_of_base(to_timer_of(evt)); evt 82 drivers/clocksource/timer-tegra.c static int tegra_timer_set_periodic(struct clock_event_device *evt) evt 84 drivers/clocksource/timer-tegra.c void __iomem *reg_base = timer_of_base(to_timer_of(evt)); evt 85 drivers/clocksource/timer-tegra.c unsigned long period = timer_of_period(to_timer_of(evt)); evt 95 drivers/clocksource/timer-tegra.c struct clock_event_device *evt = dev_id; evt 96 drivers/clocksource/timer-tegra.c void __iomem *reg_base = timer_of_base(to_timer_of(evt)); evt 99 drivers/clocksource/timer-tegra.c evt->event_handler(evt); evt 104 drivers/clocksource/timer-tegra.c static void tegra_timer_suspend(struct clock_event_device *evt) evt 106 drivers/clocksource/timer-tegra.c void __iomem *reg_base = timer_of_base(to_timer_of(evt)); evt 111 drivers/clocksource/timer-tegra.c static void tegra_timer_resume(struct clock_event_device *evt) evt 190 drivers/clocksource/timer-u300.c static int u300_shutdown(struct clock_event_device *evt) evt 205 drivers/clocksource/timer-u300.c static int u300_set_oneshot(struct clock_event_device *evt) evt 236 drivers/clocksource/timer-u300.c static int u300_set_periodic(struct clock_event_device *evt) evt 239 drivers/clocksource/timer-u300.c container_of(evt, struct u300_clockevent_data, cevd); evt 277 drivers/clocksource/timer-u300.c struct clock_event_device *evt) evt 324 drivers/clocksource/timer-u300.c struct clock_event_device *evt = &u300_clockevent_data.cevd; evt 329 drivers/clocksource/timer-u300.c evt->event_handler(evt); evt 85 drivers/clocksource/timer-vf-pit.c static int pit_shutdown(struct clock_event_device *evt) evt 91 drivers/clocksource/timer-vf-pit.c static int pit_set_periodic(struct clock_event_device *evt) evt 93 drivers/clocksource/timer-vf-pit.c pit_set_next_event(cycle_per_jiffy, evt); evt 99 drivers/clocksource/timer-vf-pit.c struct clock_event_device *evt = dev_id; evt 109 drivers/clocksource/timer-vf-pit.c if (likely(clockevent_state_oneshot(evt))) evt 112 drivers/clocksource/timer-vf-pit.c evt->event_handler(evt); evt 62 drivers/clocksource/timer-vt8500.c struct clock_event_device *evt) evt 79 drivers/clocksource/timer-vt8500.c static int vt8500_shutdown(struct clock_event_device *evt) evt 97 drivers/clocksource/timer-vt8500.c struct clock_event_device *evt = dev_id; evt 99 drivers/clocksource/timer-vt8500.c evt->event_handler(evt); evt 628 drivers/dma/imx-sdma.c unsigned long evt, mcu, dsp; evt 633 drivers/dma/imx-sdma.c evt = readl_relaxed(sdma->regs + SDMA_H_EVTOVR); evt 643 drivers/dma/imx-sdma.c __clear_bit(channel, &evt); evt 645 drivers/dma/imx-sdma.c __set_bit(channel, &evt); evt 652 drivers/dma/imx-sdma.c writel_relaxed(evt, sdma->regs + SDMA_H_EVTOVR); evt 366 drivers/firewire/ohci.c static void log_irqs(struct fw_ohci *ohci, u32 evt) evt 373 drivers/firewire/ohci.c !(evt & OHCI1394_busReset)) evt 376 drivers/firewire/ohci.c ohci_notice(ohci, "IRQ %08x%s%s%s%s%s%s%s%s%s%s%s%s%s%s%s\n", evt, evt 377 drivers/firewire/ohci.c evt & OHCI1394_selfIDComplete ? " selfID" : "", evt 378 drivers/firewire/ohci.c evt & OHCI1394_RQPkt ? " AR_req" : "", evt 379 drivers/firewire/ohci.c evt & OHCI1394_RSPkt ? " AR_resp" : "", evt 380 drivers/firewire/ohci.c evt & OHCI1394_reqTxComplete ? " AT_req" : "", evt 381 drivers/firewire/ohci.c evt & OHCI1394_respTxComplete ? " AT_resp" : "", evt 382 drivers/firewire/ohci.c evt & OHCI1394_isochRx ? " IR" : "", evt 383 drivers/firewire/ohci.c evt & OHCI1394_isochTx ? " IT" : "", evt 384 drivers/firewire/ohci.c evt & OHCI1394_postedWriteErr ? " postedWriteErr" : "", evt 385 drivers/firewire/ohci.c evt & OHCI1394_cycleTooLong ? " cycleTooLong" : "", evt 386 drivers/firewire/ohci.c evt & OHCI1394_cycle64Seconds ? " cycle64Seconds" : "", evt 387 drivers/firewire/ohci.c evt & OHCI1394_cycleInconsistent ? " cycleInconsistent" : "", evt 388 drivers/firewire/ohci.c evt & OHCI1394_regAccessFail ? " regAccessFail" : "", evt 389 drivers/firewire/ohci.c evt & OHCI1394_unrecoverableError ? " unrecoverableError" : "", evt 390 drivers/firewire/ohci.c evt & OHCI1394_busReset ? " busReset" : "", evt 391 drivers/firewire/ohci.c evt & ~(OHCI1394_selfIDComplete | OHCI1394_RQPkt | evt 472 drivers/firewire/ohci.c char dir, int speed, u32 *header, int evt) evt 480 drivers/firewire/ohci.c if (unlikely(evt >= ARRAY_SIZE(evts))) evt 481 drivers/firewire/ohci.c evt = 0x1f; evt 483 drivers/firewire/ohci.c if (evt == OHCI1394_evt_bus_reset) { evt 505 drivers/firewire/ohci.c dir, evts[evt], tcodes[tcode]); evt 509 drivers/firewire/ohci.c dir, evts[evt], header[1], header[2]); evt 515 drivers/firewire/ohci.c header[1] >> 16, header[0] >> 16, evts[evt], evt 522 drivers/firewire/ohci.c header[1] >> 16, header[0] >> 16, evts[evt], evt 803 drivers/firewire/ohci.c int evt; evt 854 drivers/firewire/ohci.c evt = (status >> 16) & 0x1f; evt 856 drivers/firewire/ohci.c p.ack = evt - 16; evt 861 drivers/firewire/ohci.c log_ar_at_event(ohci, 'R', p.speed, p.header, evt); evt 867 drivers/firewire/ohci.c if (evt == OHCI1394_evt_no_status && evt 884 drivers/firewire/ohci.c if (evt == OHCI1394_evt_bus_reset) { evt 1436 drivers/firewire/ohci.c int evt; evt 1452 drivers/firewire/ohci.c evt = le16_to_cpu(last->transfer_status) & 0x1f; evt 1455 drivers/firewire/ohci.c log_ar_at_event(ohci, 'T', packet->speed, packet->header, evt); evt 1457 drivers/firewire/ohci.c switch (evt) { evt 1490 drivers/firewire/ohci.c packet->ack = evt - 0x10; evt 112 drivers/gpio/gpio-kempld.c u16 evt, evt_back; evt 121 drivers/gpio/gpio-kempld.c evt = kempld_read16(pld, KEMPLD_GPIO_EVT_LVL_EDGE); evt 127 drivers/gpio/gpio-kempld.c return evt ? __ffs(evt) : 16; evt 795 drivers/gpu/drm/sti/sti_hqvdp.c static int sti_hqvdp_vtg_cb(struct notifier_block *nb, unsigned long evt, void *data) evt 801 drivers/gpu/drm/sti/sti_hqvdp.c if ((evt != VTG_TOP_FIELD_EVENT) && (evt != VTG_BOTTOM_FIELD_EVENT)) { evt 85 drivers/gpu/drm/xen/xen_drm_front_evtchnl.c struct xendispl_event_page *page = evtchnl->u.evt.page; evt 131 drivers/gpu/drm/xen/xen_drm_front_evtchnl.c page = (unsigned long)evtchnl->u.evt.page; evt 204 drivers/gpu/drm/xen/xen_drm_front_evtchnl.c evtchnl->u.evt.page = (struct xendispl_event_page *)page; evt 254 drivers/gpu/drm/xen/xen_drm_front_evtchnl.c &front_info->evt_pairs[conn].evt, evt 318 drivers/gpu/drm/xen/xen_drm_front_evtchnl.c ret = evtchnl_publish(xbt, &front_info->evt_pairs[conn].evt, evt 368 drivers/gpu/drm/xen/xen_drm_front_evtchnl.c front_info->evt_pairs[i].evt.state = state; evt 382 drivers/gpu/drm/xen/xen_drm_front_evtchnl.c evtchnl_free(front_info, &front_info->evt_pairs[i].evt); evt 61 drivers/gpu/drm/xen/xen_drm_front_evtchnl.h } evt; evt 67 drivers/gpu/drm/xen/xen_drm_front_evtchnl.h struct xen_drm_front_evtchnl evt; evt 46 drivers/gpu/ipu-v3/ipu-dc.c #define DC_RL_CH(evt) (8 + ((evt) & ~0x1) * 2) evt 140 drivers/infiniband/core/uverbs_main.c struct ib_uverbs_event *evt, *tmp; evt 144 drivers/infiniband/core/uverbs_main.c list_for_each_entry_safe(evt, tmp, &uobj->comp_list, obj_list) { evt 145 drivers/infiniband/core/uverbs_main.c list_del(&evt->list); evt 146 drivers/infiniband/core/uverbs_main.c kfree(evt); evt 154 drivers/infiniband/core/uverbs_main.c list_for_each_entry_safe(evt, tmp, &uobj->async_list, obj_list) { evt 155 drivers/infiniband/core/uverbs_main.c list_del(&evt->list); evt 156 drivers/infiniband/core/uverbs_main.c kfree(evt); evt 164 drivers/infiniband/core/uverbs_main.c struct ib_uverbs_event *evt, *tmp; evt 167 drivers/infiniband/core/uverbs_main.c list_for_each_entry_safe(evt, tmp, &uobj->event_list, obj_list) { evt 168 drivers/infiniband/core/uverbs_main.c list_del(&evt->list); evt 169 drivers/infiniband/core/uverbs_main.c kfree(evt); evt 195 drivers/infiniband/hw/cxgb3/iwch.c static void iwch_event_handler(struct t3cdev *tdev, u32 evt, u32 port_id) evt 206 drivers/infiniband/hw/cxgb3/iwch.c switch (evt) { evt 804 drivers/infiniband/hw/ocrdma/ocrdma_hw.c struct ocrdma_ae_pvid_mcqe *evt; evt 810 drivers/infiniband/hw/ocrdma/ocrdma_hw.c evt = (struct ocrdma_ae_pvid_mcqe *)cqe; evt 811 drivers/infiniband/hw/ocrdma/ocrdma_hw.c if ((evt->tag_enabled & OCRDMA_AE_PVID_MCQE_ENABLED_MASK) >> evt 813 drivers/infiniband/hw/ocrdma/ocrdma_hw.c dev->pvid = ((evt->tag_enabled & evt 830 drivers/infiniband/hw/ocrdma/ocrdma_hw.c struct ocrdma_ae_lnkst_mcqe *evt; evt 833 drivers/infiniband/hw/ocrdma/ocrdma_hw.c evt = (struct ocrdma_ae_lnkst_mcqe *)cqe; evt 834 drivers/infiniband/hw/ocrdma/ocrdma_hw.c lstate = ocrdma_get_ae_link_state(evt->speed_state_ptn); evt 1700 drivers/iommu/arm-smmu-v3.c u64 evt[EVTQ_ENT_DWORDS]; evt 1703 drivers/iommu/arm-smmu-v3.c while (!queue_remove_raw(q, evt)) { evt 1704 drivers/iommu/arm-smmu-v3.c u8 id = FIELD_GET(EVTQ_0_ID, evt[0]); evt 1707 drivers/iommu/arm-smmu-v3.c for (i = 0; i < ARRAY_SIZE(evt); ++i) evt 1709 drivers/iommu/arm-smmu-v3.c (unsigned long long)evt[i]); evt 1727 drivers/iommu/arm-smmu-v3.c static void arm_smmu_handle_ppr(struct arm_smmu_device *smmu, u64 *evt) evt 1733 drivers/iommu/arm-smmu-v3.c sid = FIELD_GET(PRIQ_0_SID, evt[0]); evt 1734 drivers/iommu/arm-smmu-v3.c ssv = FIELD_GET(PRIQ_0_SSID_V, evt[0]); evt 1735 drivers/iommu/arm-smmu-v3.c ssid = ssv ? FIELD_GET(PRIQ_0_SSID, evt[0]) : 0; evt 1736 drivers/iommu/arm-smmu-v3.c last = FIELD_GET(PRIQ_0_PRG_LAST, evt[0]); evt 1737 drivers/iommu/arm-smmu-v3.c grpid = FIELD_GET(PRIQ_1_PRG_IDX, evt[1]); evt 1743 drivers/iommu/arm-smmu-v3.c evt[0] & PRIQ_0_PERM_PRIV ? "" : "un", evt 1744 drivers/iommu/arm-smmu-v3.c evt[0] & PRIQ_0_PERM_READ ? "R" : "", evt 1745 drivers/iommu/arm-smmu-v3.c evt[0] & PRIQ_0_PERM_WRITE ? "W" : "", evt 1746 drivers/iommu/arm-smmu-v3.c evt[0] & PRIQ_0_PERM_EXEC ? "X" : "", evt 1747 drivers/iommu/arm-smmu-v3.c evt[1] & PRIQ_1_ADDR_MASK); evt 1770 drivers/iommu/arm-smmu-v3.c u64 evt[PRIQ_ENT_DWORDS]; evt 1773 drivers/iommu/arm-smmu-v3.c while (!queue_remove_raw(q, evt)) evt 1774 drivers/iommu/arm-smmu-v3.c arm_smmu_handle_ppr(smmu, evt); evt 1038 drivers/iommu/iommu.c int iommu_report_device_fault(struct device *dev, struct iommu_fault_event *evt) evt 1045 drivers/iommu/iommu.c if (!param || !evt) evt 1056 drivers/iommu/iommu.c if (evt->fault.type == IOMMU_FAULT_PAGE_REQ && evt 1057 drivers/iommu/iommu.c (evt->fault.prm.flags & IOMMU_FAULT_PAGE_REQUEST_LAST_PAGE)) { evt 1058 drivers/iommu/iommu.c evt_pending = kmemdup(evt, sizeof(struct iommu_fault_event), evt 1069 drivers/iommu/iommu.c ret = fparam->handler(&evt->fault, fparam->data); evt 1087 drivers/iommu/iommu.c struct iommu_fault_event *evt; evt 1112 drivers/iommu/iommu.c list_for_each_entry(evt, ¶m->fault_param->faults, list) { evt 1113 drivers/iommu/iommu.c prm = &evt->fault.prm; evt 1123 drivers/iommu/iommu.c ret = domain->ops->page_response(dev, evt, msg); evt 1124 drivers/iommu/iommu.c list_del(&evt->list); evt 1125 drivers/iommu/iommu.c kfree(evt); evt 564 drivers/iommu/virtio-iommu.c struct viommu_event *evt; evt 567 drivers/iommu/virtio-iommu.c while ((evt = virtqueue_get_buf(vq, &len)) != NULL) { evt 568 drivers/iommu/virtio-iommu.c if (len > sizeof(*evt)) { evt 571 drivers/iommu/virtio-iommu.c len, sizeof(*evt)); evt 572 drivers/iommu/virtio-iommu.c } else if (!(evt->head & VIOMMU_FAULT_RESV_MASK)) { evt 573 drivers/iommu/virtio-iommu.c viommu_fault_handler(viommu, &evt->fault); evt 576 drivers/iommu/virtio-iommu.c sg_init_one(sg, evt, sizeof(*evt)); evt 577 drivers/iommu/virtio-iommu.c ret = virtqueue_add_inbuf(vq, sg, 1, evt, GFP_ATOMIC); evt 297 drivers/leds/trigger/ledtrig-netdev.c unsigned long evt, void *dv) evt 304 drivers/leds/trigger/ledtrig-netdev.c if (evt != NETDEV_UP && evt != NETDEV_DOWN && evt != NETDEV_CHANGE evt 305 drivers/leds/trigger/ledtrig-netdev.c && evt != NETDEV_REGISTER && evt != NETDEV_UNREGISTER evt 306 drivers/leds/trigger/ledtrig-netdev.c && evt != NETDEV_CHANGENAME) evt 310 drivers/leds/trigger/ledtrig-netdev.c (evt == NETDEV_CHANGENAME && !strcmp(dev->name, trigger_data->device_name)) || evt 311 drivers/leds/trigger/ledtrig-netdev.c (evt == NETDEV_REGISTER && !strcmp(dev->name, trigger_data->device_name)))) evt 319 drivers/leds/trigger/ledtrig-netdev.c switch (evt) { evt 1672 drivers/net/ethernet/alteon/acenic.c cmd.evt = C_SET_RX_PRD_IDX; evt 1789 drivers/net/ethernet/alteon/acenic.c cmd.evt = C_SET_RX_JUMBO_PRD_IDX; evt 1821 drivers/net/ethernet/alteon/acenic.c switch (ap->evt_ring[evtcsm].evt) { evt 1892 drivers/net/ethernet/alteon/acenic.c cmd.evt = C_SET_RX_JUMBO_PRD_IDX; evt 1910 drivers/net/ethernet/alteon/acenic.c ap->name, ap->evt_ring[evtcsm].evt); evt 2241 drivers/net/ethernet/alteon/acenic.c cmd.evt = C_CLEAR_STATS; evt 2246 drivers/net/ethernet/alteon/acenic.c cmd.evt = C_HOST_STATE; evt 2256 drivers/net/ethernet/alteon/acenic.c cmd.evt = C_SET_PROMISC_MODE; evt 2267 drivers/net/ethernet/alteon/acenic.c cmd.evt = C_LNK_NEGOTIATION; evt 2300 drivers/net/ethernet/alteon/acenic.c cmd.evt = C_SET_PROMISC_MODE; evt 2307 drivers/net/ethernet/alteon/acenic.c cmd.evt = C_HOST_STATE; evt 2352 drivers/net/ethernet/alteon/acenic.c cmd.evt = C_RESET_JUMBO_RNG; evt 2568 drivers/net/ethernet/alteon/acenic.c cmd.evt = C_RESET_JUMBO_RNG; evt 2688 drivers/net/ethernet/alteon/acenic.c cmd.evt = C_LNK_NEGOTIATION; evt 2734 drivers/net/ethernet/alteon/acenic.c cmd.evt = C_SET_MAC_ADDR; evt 2750 drivers/net/ethernet/alteon/acenic.c cmd.evt = C_SET_MULTICAST_MODE; evt 2756 drivers/net/ethernet/alteon/acenic.c cmd.evt = C_SET_MULTICAST_MODE; evt 2764 drivers/net/ethernet/alteon/acenic.c cmd.evt = C_SET_PROMISC_MODE; evt 2770 drivers/net/ethernet/alteon/acenic.c cmd.evt = C_SET_PROMISC_MODE; evt 2784 drivers/net/ethernet/alteon/acenic.c cmd.evt = C_SET_MULTICAST_MODE; evt 2789 drivers/net/ethernet/alteon/acenic.c cmd.evt = C_SET_MULTICAST_MODE; evt 320 drivers/net/ethernet/alteon/acenic.h u32 evt:8; evt 322 drivers/net/ethernet/alteon/acenic.h u32 evt:8; evt 366 drivers/net/ethernet/alteon/acenic.h u32 evt:8; evt 368 drivers/net/ethernet/alteon/acenic.h u32 evt:8; evt 285 drivers/net/ethernet/emulex/benet/be_cmds.c struct be_async_event_link_state *evt = evt 297 drivers/net/ethernet/emulex/benet/be_cmds.c !(evt->port_link_status & LOGICAL_LINK_STATUS_MASK)) evt 305 drivers/net/ethernet/emulex/benet/be_cmds.c evt->port_link_status & LINK_STATUS_MASK); evt 311 drivers/net/ethernet/emulex/benet/be_cmds.c struct be_async_event_misconfig_port *evt = evt 313 drivers/net/ethernet/emulex/benet/be_cmds.c u32 sfp_misconfig_evt_word1 = le32_to_cpu(evt->event_data_word1); evt 314 drivers/net/ethernet/emulex/benet/be_cmds.c u32 sfp_misconfig_evt_word2 = le32_to_cpu(evt->event_data_word2); evt 369 drivers/net/ethernet/emulex/benet/be_cmds.c struct be_async_event_grp5_cos_priority *evt = evt 372 drivers/net/ethernet/emulex/benet/be_cmds.c if (evt->valid) { evt 373 drivers/net/ethernet/emulex/benet/be_cmds.c adapter->vlan_prio_bmap = evt->available_priority_bmap; evt 375 drivers/net/ethernet/emulex/benet/be_cmds.c evt->reco_default_priority << VLAN_PRIO_SHIFT; evt 383 drivers/net/ethernet/emulex/benet/be_cmds.c struct be_async_event_grp5_qos_link_speed *evt = evt 387 drivers/net/ethernet/emulex/benet/be_cmds.c evt->physical_port == adapter->port_num) evt 388 drivers/net/ethernet/emulex/benet/be_cmds.c adapter->phy.link_speed = le16_to_cpu(evt->qos_link_speed) * 10; evt 395 drivers/net/ethernet/emulex/benet/be_cmds.c struct be_async_event_grp5_pvid_state *evt = evt 398 drivers/net/ethernet/emulex/benet/be_cmds.c if (evt->enabled) { evt 399 drivers/net/ethernet/emulex/benet/be_cmds.c adapter->pvid = le16_to_cpu(evt->tag) & VLAN_VID_MASK; evt 410 drivers/net/ethernet/emulex/benet/be_cmds.c struct be_async_fw_control *evt = (struct be_async_fw_control *)compl; evt 411 drivers/net/ethernet/emulex/benet/be_cmds.c u32 evt_dw1 = le32_to_cpu(evt->event_data_word1); evt 415 drivers/net/ethernet/emulex/benet/be_cmds.c adapter->bmc_filt_mask = le32_to_cpu(evt->event_data_word2); evt 450 drivers/net/ethernet/emulex/benet/be_cmds.c struct be_async_event_qnq *evt = (struct be_async_event_qnq *)cmp; evt 457 drivers/net/ethernet/emulex/benet/be_cmds.c if (evt->valid) evt 458 drivers/net/ethernet/emulex/benet/be_cmds.c adapter->qnq_vid = le16_to_cpu(evt->vlan_tag); evt 206 drivers/net/ethernet/emulex/benet/be_hw.h u32 evt; evt 2791 drivers/net/ethernet/emulex/benet/be_main.c if (eqe->evt == 0) evt 2795 drivers/net/ethernet/emulex/benet/be_main.c eqe->evt = 0; evt 1347 drivers/net/ethernet/pasemi/pasemi_mac.c int cs_size, i, fill, hdr, evt; evt 1394 drivers/net/ethernet/pasemi/pasemi_mac.c evt = !csring->last_event; evt 1395 drivers/net/ethernet/pasemi/pasemi_mac.c csring->last_event = evt; evt 1399 drivers/net/ethernet/pasemi/pasemi_mac.c CTRL_CMD_ETYPE_SET | CTRL_CMD_REG(csring->events[evt]); evt 1402 drivers/net/ethernet/pasemi/pasemi_mac.c CTRL_CMD_ETYPE_WCLR | CTRL_CMD_REG(csring->events[!evt]); evt 1412 drivers/net/ethernet/pasemi/pasemi_mac.c CTRL_CMD_ETYPE_WSET | CTRL_CMD_REG(csring->events[evt]); evt 1415 drivers/net/ethernet/pasemi/pasemi_mac.c CTRL_CMD_ETYPE_CLR | CTRL_CMD_REG(csring->events[!evt]); evt 1165 drivers/net/ethernet/sfc/ptp.c struct efx_ptp_event_rx *evt; evt 1167 drivers/net/ethernet/sfc/ptp.c evt = list_entry(cursor, struct efx_ptp_event_rx, evt 1169 drivers/net/ethernet/sfc/ptp.c if (time_after(jiffies, evt->expiry)) { evt 1170 drivers/net/ethernet/sfc/ptp.c list_move(&evt->link, &ptp->evt_free_list); evt 1202 drivers/net/ethernet/sfc/ptp.c struct efx_ptp_event_rx *evt; evt 1204 drivers/net/ethernet/sfc/ptp.c evt = list_entry(cursor, struct efx_ptp_event_rx, link); evt 1205 drivers/net/ethernet/sfc/ptp.c if ((evt->seq0 == match->words[0]) && evt 1206 drivers/net/ethernet/sfc/ptp.c (evt->seq1 == match->words[1])) { evt 1211 drivers/net/ethernet/sfc/ptp.c timestamps->hwtstamp = evt->hwtimestamp; evt 1215 drivers/net/ethernet/sfc/ptp.c list_move(&evt->link, &ptp->evt_free_list); evt 1862 drivers/net/ethernet/sfc/ptp.c struct efx_ptp_event_rx *evt = NULL; evt 1874 drivers/net/ethernet/sfc/ptp.c evt = list_first_entry(&ptp->evt_free_list, evt 1876 drivers/net/ethernet/sfc/ptp.c list_del(&evt->link); evt 1878 drivers/net/ethernet/sfc/ptp.c evt->seq0 = EFX_QWORD_FIELD(ptp->evt_frags[2], MCDI_EVENT_DATA); evt 1879 drivers/net/ethernet/sfc/ptp.c evt->seq1 = (EFX_QWORD_FIELD(ptp->evt_frags[2], evt 1885 drivers/net/ethernet/sfc/ptp.c evt->hwtimestamp = efx->ptp_data->nic_to_kernel_time( evt 1889 drivers/net/ethernet/sfc/ptp.c evt->expiry = jiffies + msecs_to_jiffies(PKT_EVENT_LIFETIME_MS); evt 1890 drivers/net/ethernet/sfc/ptp.c list_add_tail(&evt->link, &ptp->evt_list); evt 481 drivers/net/wireless/ath/ath6kl/wmi.c struct wmi_tx_complete_event *evt; evt 485 drivers/net/wireless/ath/ath6kl/wmi.c evt = (struct wmi_tx_complete_event *) datap; evt 488 drivers/net/wireless/ath/ath6kl/wmi.c evt->num_msg, evt->msg_len, evt->msg_type); evt 490 drivers/net/wireless/ath/ath6kl/wmi.c for (index = 0; index < evt->num_msg; index++) { evt 451 drivers/net/wireless/ath/wil6210/cfg80211.c struct wmi_notify_req_done_event evt; evt 471 drivers/net/wireless/ath/wil6210/cfg80211.c cid, vif->mid, le16_to_cpu(reply.evt.bf_mcs), evt 472 drivers/net/wireless/ath/wil6210/cfg80211.c le64_to_cpu(reply.evt.tsf), reply.evt.status, evt 473 drivers/net/wireless/ath/wil6210/cfg80211.c reply.evt.rssi, evt 474 drivers/net/wireless/ath/wil6210/cfg80211.c reply.evt.sqi, evt 475 drivers/net/wireless/ath/wil6210/cfg80211.c le32_to_cpu(reply.evt.tx_tpt), evt 476 drivers/net/wireless/ath/wil6210/cfg80211.c le32_to_cpu(reply.evt.tx_goodput), evt 477 drivers/net/wireless/ath/wil6210/cfg80211.c le32_to_cpu(reply.evt.rx_goodput), evt 478 drivers/net/wireless/ath/wil6210/cfg80211.c le16_to_cpu(reply.evt.my_rx_sector), evt 479 drivers/net/wireless/ath/wil6210/cfg80211.c le16_to_cpu(reply.evt.my_tx_sector), evt 480 drivers/net/wireless/ath/wil6210/cfg80211.c le16_to_cpu(reply.evt.other_rx_sector), evt 481 drivers/net/wireless/ath/wil6210/cfg80211.c le16_to_cpu(reply.evt.other_tx_sector), evt 482 drivers/net/wireless/ath/wil6210/cfg80211.c reply.evt.tx_mode); evt 495 drivers/net/wireless/ath/wil6210/cfg80211.c if (wil->use_enhanced_dma_hw && reply.evt.tx_mode != WMI_TX_MODE_DMG) evt 499 drivers/net/wireless/ath/wil6210/cfg80211.c sinfo->txrate.mcs = le16_to_cpu(reply.evt.bf_mcs); evt 502 drivers/net/wireless/ath/wil6210/cfg80211.c wil_tx_cb_mode_to_n_bonded(reply.evt.tx_mode); evt 516 drivers/net/wireless/ath/wil6210/cfg80211.c sinfo->signal = reply.evt.rssi; evt 518 drivers/net/wireless/ath/wil6210/cfg80211.c sinfo->signal = reply.evt.sqi; evt 2835 drivers/net/wireless/ath/wil6210/cfg80211.c struct wmi_get_rf_sector_params_done_event evt; evt 2837 drivers/net/wireless/ath/wil6210/cfg80211.c .evt = {.status = WMI_RF_SECTOR_STATUS_NOT_SUPPORTED_ERROR}, evt 2890 drivers/net/wireless/ath/wil6210/cfg80211.c if (reply.evt.status) { evt 2892 drivers/net/wireless/ath/wil6210/cfg80211.c reply.evt.status); evt 2893 drivers/net/wireless/ath/wil6210/cfg80211.c return wil_rf_sector_status_to_rc(reply.evt.status); evt 2902 drivers/net/wireless/ath/wil6210/cfg80211.c le64_to_cpu(reply.evt.tsf), evt 2915 drivers/net/wireless/ath/wil6210/cfg80211.c si = &reply.evt.sectors_info[i]; evt 2957 drivers/net/wireless/ath/wil6210/cfg80211.c struct wmi_set_rf_sector_params_done_event evt; evt 2959 drivers/net/wireless/ath/wil6210/cfg80211.c .evt = {.status = WMI_RF_SECTOR_STATUS_NOT_SUPPORTED_ERROR}, evt 3051 drivers/net/wireless/ath/wil6210/cfg80211.c return wil_rf_sector_status_to_rc(reply.evt.status); evt 3067 drivers/net/wireless/ath/wil6210/cfg80211.c struct wmi_get_selected_rf_sector_index_done_event evt; evt 3069 drivers/net/wireless/ath/wil6210/cfg80211.c .evt = {.status = WMI_RF_SECTOR_STATUS_NOT_SUPPORTED_ERROR}, evt 3117 drivers/net/wireless/ath/wil6210/cfg80211.c if (reply.evt.status) { evt 3119 drivers/net/wireless/ath/wil6210/cfg80211.c reply.evt.status); evt 3120 drivers/net/wireless/ath/wil6210/cfg80211.c return wil_rf_sector_status_to_rc(reply.evt.status); evt 3129 drivers/net/wireless/ath/wil6210/cfg80211.c le64_to_cpu(reply.evt.tsf), evt 3132 drivers/net/wireless/ath/wil6210/cfg80211.c le16_to_cpu(reply.evt.sector_idx))) evt 3149 drivers/net/wireless/ath/wil6210/cfg80211.c struct wmi_set_selected_rf_sector_index_done_event evt; evt 3151 drivers/net/wireless/ath/wil6210/cfg80211.c .evt = {.status = WMI_RF_SECTOR_STATUS_NOT_SUPPORTED_ERROR}, evt 3166 drivers/net/wireless/ath/wil6210/cfg80211.c return wil_rf_sector_status_to_rc(reply.evt.status); evt 1301 drivers/net/wireless/ath/wil6210/debugfs.c struct wmi_notify_req_done_event evt; evt 1315 drivers/net/wireless/ath/wil6210/debugfs.c if (rc || is_all_zeros(&reply.evt, sizeof(reply.evt))) evt 1318 drivers/net/wireless/ath/wil6210/debugfs.c status = le32_to_cpu(reply.evt.status); evt 1329 drivers/net/wireless/ath/wil6210/debugfs.c le64_to_cpu(reply.evt.tsf), evt 1330 drivers/net/wireless/ath/wil6210/debugfs.c le16_to_cpu(reply.evt.bf_mcs), evt 1331 drivers/net/wireless/ath/wil6210/debugfs.c le32_to_cpu(reply.evt.tx_tpt), evt 1332 drivers/net/wireless/ath/wil6210/debugfs.c reply.evt.sqi, evt 1333 drivers/net/wireless/ath/wil6210/debugfs.c reply.evt.rssi, evt 1335 drivers/net/wireless/ath/wil6210/debugfs.c le16_to_cpu(reply.evt.my_rx_sector), evt 1336 drivers/net/wireless/ath/wil6210/debugfs.c le16_to_cpu(reply.evt.my_tx_sector), evt 1337 drivers/net/wireless/ath/wil6210/debugfs.c le16_to_cpu(reply.evt.other_rx_sector), evt 1338 drivers/net/wireless/ath/wil6210/debugfs.c le16_to_cpu(reply.evt.other_tx_sector), evt 1339 drivers/net/wireless/ath/wil6210/debugfs.c le32_to_cpu(reply.evt.rx_goodput), evt 1340 drivers/net/wireless/ath/wil6210/debugfs.c le32_to_cpu(reply.evt.tx_goodput)); evt 480 drivers/net/wireless/ath/wil6210/main.c struct wmi_disconnect_event evt; evt 587 drivers/net/wireless/ath/wil6210/pcie_bus.c enum wil_platform_event evt; evt 598 drivers/net/wireless/ath/wil6210/pcie_bus.c evt = WIL_PLATFORM_EVT_PRE_SUSPEND; evt 601 drivers/net/wireless/ath/wil6210/pcie_bus.c evt); evt 606 drivers/net/wireless/ath/wil6210/pcie_bus.c evt = WIL_PLATFORM_EVT_POST_SUSPEND; evt 609 drivers/net/wireless/ath/wil6210/pcie_bus.c evt); evt 52 drivers/net/wireless/ath/wil6210/wil_platform.h int (*notify)(void *handle, enum wil_platform_event evt); evt 781 drivers/net/wireless/ath/wil6210/wmi.c struct wmi_ready_event *evt = d; evt 785 drivers/net/wireless/ath/wil6210/wmi.c wil->fw_version, le32_to_cpu(evt->sw_version), evt 786 drivers/net/wireless/ath/wil6210/wmi.c evt->mac, evt->numof_additional_mids); evt 787 drivers/net/wireless/ath/wil6210/wmi.c if (evt->numof_additional_mids + 1 < wil->max_vifs) { evt 797 drivers/net/wireless/ath/wil6210/wmi.c evt->rfc_read_calib_result); evt 798 drivers/net/wireless/ath/wil6210/wmi.c wil->fw_calib_result = evt->rfc_read_calib_result; evt 803 drivers/net/wireless/ath/wil6210/wmi.c evt->max_assoc_sta > 0) { evt 804 drivers/net/wireless/ath/wil6210/wmi.c fw_max_assoc_sta = evt->max_assoc_sta; evt 968 drivers/net/wireless/ath/wil6210/wmi.c struct wmi_connect_event *evt = d; evt 979 drivers/net/wireless/ath/wil6210/wmi.c if (len < sizeof(*evt)) { evt 983 drivers/net/wireless/ath/wil6210/wmi.c if (len != sizeof(*evt) + evt->beacon_ie_len + evt->assoc_req_len + evt 984 drivers/net/wireless/ath/wil6210/wmi.c evt->assoc_resp_len) { evt 987 drivers/net/wireless/ath/wil6210/wmi.c len, (int)sizeof(*evt), evt->beacon_ie_len, evt 988 drivers/net/wireless/ath/wil6210/wmi.c evt->assoc_req_len, evt->assoc_resp_len); evt 991 drivers/net/wireless/ath/wil6210/wmi.c if (evt->cid >= wil->max_assoc_sta) { evt 992 drivers/net/wireless/ath/wil6210/wmi.c wil_err(wil, "Connect CID invalid : %d\n", evt->cid); evt 996 drivers/net/wireless/ath/wil6210/wmi.c ch = evt->channel + 1; evt 998 drivers/net/wireless/ath/wil6210/wmi.c evt->bssid, ch, evt->cid, evt->aid); evt 1000 drivers/net/wireless/ath/wil6210/wmi.c evt->assoc_info, len - sizeof(*evt), true); evt 1003 drivers/net/wireless/ath/wil6210/wmi.c assoc_req_ie = &evt->assoc_info[evt->beacon_ie_len + evt 1005 drivers/net/wireless/ath/wil6210/wmi.c assoc_req_ielen = evt->assoc_req_len - assoc_req_ie_offset; evt 1006 drivers/net/wireless/ath/wil6210/wmi.c if (evt->assoc_req_len <= assoc_req_ie_offset) { evt 1011 drivers/net/wireless/ath/wil6210/wmi.c assoc_resp_ie = &evt->assoc_info[evt->beacon_ie_len + evt 1012 drivers/net/wireless/ath/wil6210/wmi.c evt->assoc_req_len + evt 1014 drivers/net/wireless/ath/wil6210/wmi.c assoc_resp_ielen = evt->assoc_resp_len - assoc_resp_ie_offset; evt 1015 drivers/net/wireless/ath/wil6210/wmi.c if (evt->assoc_resp_len <= assoc_resp_ie_offset) { evt 1023 drivers/net/wireless/ath/wil6210/wmi.c evt->cid); evt 1040 drivers/net/wireless/ath/wil6210/wmi.c if (wil->sta[evt->cid].status != wil_sta_unused) { evt 1042 drivers/net/wireless/ath/wil6210/wmi.c wil->sta[evt->cid].status, evt->cid); evt 1048 drivers/net/wireless/ath/wil6210/wmi.c ether_addr_copy(wil->sta[evt->cid].addr, evt->bssid); evt 1049 drivers/net/wireless/ath/wil6210/wmi.c wil->sta[evt->cid].mid = vif->mid; evt 1050 drivers/net/wireless/ath/wil6210/wmi.c wil->sta[evt->cid].status = wil_sta_conn_pending; evt 1052 drivers/net/wireless/ath/wil6210/wmi.c rc = wil_ring_init_tx(vif, evt->cid); evt 1055 drivers/net/wireless/ath/wil6210/wmi.c evt->cid, rc); evt 1056 drivers/net/wireless/ath/wil6210/wmi.c wmi_disconnect_sta(vif, wil->sta[evt->cid].addr, evt 1059 drivers/net/wireless/ath/wil6210/wmi.c wil_info(wil, "successful connection to CID %d\n", evt->cid); evt 1068 drivers/net/wireless/ath/wil6210/wmi.c cfg80211_connect_result(ndev, evt->bssid, NULL, 0, evt 1077 drivers/net/wireless/ath/wil6210/wmi.c cfg80211_connect_bss(ndev, evt->bssid, vif->bss, evt 1090 drivers/net/wireless/ath/wil6210/wmi.c cfg80211_del_sta(ndev, evt->bssid, GFP_KERNEL); evt 1107 drivers/net/wireless/ath/wil6210/wmi.c cfg80211_new_sta(ndev, evt->bssid, sinfo, GFP_KERNEL); evt 1112 drivers/net/wireless/ath/wil6210/wmi.c evt->cid); evt 1116 drivers/net/wireless/ath/wil6210/wmi.c wil->sta[evt->cid].status = wil_sta_connected; evt 1117 drivers/net/wireless/ath/wil6210/wmi.c wil->sta[evt->cid].aid = evt->aid; evt 1124 drivers/net/wireless/ath/wil6210/wmi.c wil->sta[evt->cid].status = wil_sta_unused; evt 1125 drivers/net/wireless/ath/wil6210/wmi.c wil->sta[evt->cid].mid = U8_MAX; evt 1135 drivers/net/wireless/ath/wil6210/wmi.c struct wmi_disconnect_event *evt = d; evt 1136 drivers/net/wireless/ath/wil6210/wmi.c u16 reason_code = le16_to_cpu(evt->protocol_reason_status); evt 1139 drivers/net/wireless/ath/wil6210/wmi.c evt->bssid, reason_code, evt->disconnect_reason); evt 1151 drivers/net/wireless/ath/wil6210/wmi.c wil6210_disconnect_complete(vif, evt->bssid, reason_code); evt 1162 drivers/net/wireless/ath/wil6210/wmi.c cfg80211_cqm_pktloss_notify(ndev, evt->bssid, 0, evt 1180 drivers/net/wireless/ath/wil6210/wmi.c struct wmi_eapol_rx_event *evt = d; evt 1181 drivers/net/wireless/ath/wil6210/wmi.c u16 eapol_len = le16_to_cpu(evt->eapol_len); evt 1189 drivers/net/wireless/ath/wil6210/wmi.c evt->src_mac, vif->mid); evt 1191 drivers/net/wireless/ath/wil6210/wmi.c cid = wil_find_cid(wil, vif->mid, evt->src_mac); evt 1208 drivers/net/wireless/ath/wil6210/wmi.c ether_addr_copy(eth->h_source, evt->src_mac); evt 1210 drivers/net/wireless/ath/wil6210/wmi.c skb_put_data(skb, evt->eapol, eapol_len); evt 1229 drivers/net/wireless/ath/wil6210/wmi.c struct wmi_ring_en_event *evt = d; evt 1230 drivers/net/wireless/ath/wil6210/wmi.c u8 vri = evt->ring_index; evt 1278 drivers/net/wireless/ath/wil6210/wmi.c struct wmi_ba_status_event *evt = d; evt 1282 drivers/net/wireless/ath/wil6210/wmi.c evt->ringid, evt 1283 drivers/net/wireless/ath/wil6210/wmi.c evt->status == WMI_BA_AGREED ? "OK" : "N/A", evt 1284 drivers/net/wireless/ath/wil6210/wmi.c evt->agg_wsize, __le16_to_cpu(evt->ba_timeout), evt 1285 drivers/net/wireless/ath/wil6210/wmi.c evt->amsdu ? "+" : "-"); evt 1287 drivers/net/wireless/ath/wil6210/wmi.c if (evt->ringid >= WIL6210_MAX_TX_RINGS) { evt 1288 drivers/net/wireless/ath/wil6210/wmi.c wil_err(wil, "invalid ring id %d\n", evt->ringid); evt 1292 drivers/net/wireless/ath/wil6210/wmi.c if (evt->status != WMI_BA_AGREED) { evt 1293 drivers/net/wireless/ath/wil6210/wmi.c evt->ba_timeout = 0; evt 1294 drivers/net/wireless/ath/wil6210/wmi.c evt->agg_wsize = 0; evt 1295 drivers/net/wireless/ath/wil6210/wmi.c evt->amsdu = 0; evt 1298 drivers/net/wireless/ath/wil6210/wmi.c txdata = &wil->ring_tx_data[evt->ringid]; evt 1300 drivers/net/wireless/ath/wil6210/wmi.c txdata->agg_timeout = le16_to_cpu(evt->ba_timeout); evt 1301 drivers/net/wireless/ath/wil6210/wmi.c txdata->agg_wsize = evt->agg_wsize; evt 1302 drivers/net/wireless/ath/wil6210/wmi.c txdata->agg_amsdu = evt->amsdu; evt 1311 drivers/net/wireless/ath/wil6210/wmi.c struct wmi_rcp_addba_req_event *evt = d; evt 1313 drivers/net/wireless/ath/wil6210/wmi.c if (evt->cidxtid != CIDXTID_EXTENDED_CID_TID) { evt 1314 drivers/net/wireless/ath/wil6210/wmi.c parse_cidxtid(evt->cidxtid, &cid, &tid); evt 1316 drivers/net/wireless/ath/wil6210/wmi.c cid = evt->cid; evt 1317 drivers/net/wireless/ath/wil6210/wmi.c tid = evt->tid; evt 1319 drivers/net/wireless/ath/wil6210/wmi.c wil_addba_rx_request(wil, vif->mid, cid, tid, evt->dialog_token, evt 1320 drivers/net/wireless/ath/wil6210/wmi.c evt->ba_param_set, evt->ba_timeout, evt 1321 drivers/net/wireless/ath/wil6210/wmi.c evt->ba_seq_ctrl); evt 1328 drivers/net/wireless/ath/wil6210/wmi.c struct wmi_delba_event *evt = d; evt 1330 drivers/net/wireless/ath/wil6210/wmi.c u16 reason = __le16_to_cpu(evt->reason); evt 1336 drivers/net/wireless/ath/wil6210/wmi.c if (evt->cidxtid != CIDXTID_EXTENDED_CID_TID) { evt 1337 drivers/net/wireless/ath/wil6210/wmi.c parse_cidxtid(evt->cidxtid, &cid, &tid); evt 1339 drivers/net/wireless/ath/wil6210/wmi.c cid = evt->cid; evt 1340 drivers/net/wireless/ath/wil6210/wmi.c tid = evt->tid; evt 1350 drivers/net/wireless/ath/wil6210/wmi.c evt->from_initiator ? "originator" : "recipient", evt 1352 drivers/net/wireless/ath/wil6210/wmi.c if (!evt->from_initiator) { evt 1568 drivers/net/wireless/ath/wil6210/wmi.c struct wmi_link_stats_event *evt = d; evt 1575 drivers/net/wireless/ath/wil6210/wmi.c payload_size = le16_to_cpu(evt->payload_size); evt 1581 drivers/net/wireless/ath/wil6210/wmi.c wmi_link_stats_parse(vif, le64_to_cpu(evt->tsf), evt->has_next, evt 1582 drivers/net/wireless/ath/wil6210/wmi.c evt->payload, payload_size); evt 1896 drivers/net/wireless/ath/wil6210/wmi.c struct pending_wmi_event *evt; evt 1947 drivers/net/wireless/ath/wil6210/wmi.c evt = kmalloc(ALIGN(offsetof(struct pending_wmi_event, evt 1950 drivers/net/wireless/ath/wil6210/wmi.c if (!evt) evt 1953 drivers/net/wireless/ath/wil6210/wmi.c evt->event.hdr = hdr; evt 1954 drivers/net/wireless/ath/wil6210/wmi.c cmd = (void *)&evt->event.wmi; evt 1962 drivers/net/wireless/ath/wil6210/wmi.c struct wmi_cmd_hdr *wmi = &evt->event.wmi; evt 1997 drivers/net/wireless/ath/wil6210/wmi.c &evt->event.hdr, sizeof(hdr) + len, true); evt 2008 drivers/net/wireless/ath/wil6210/wmi.c kfree(evt); evt 2014 drivers/net/wireless/ath/wil6210/wmi.c list_add_tail(&evt->list, &wil->pending_wmi_ev); evt 2120 drivers/net/wireless/ath/wil6210/wmi.c struct wmi_led_cfg_done_event evt; evt 2122 drivers/net/wireless/ath/wil6210/wmi.c .evt = {.status = cpu_to_le32(WMI_FW_STATUS_FAILURE)}, evt 2144 drivers/net/wireless/ath/wil6210/wmi.c if (reply.evt.status) { evt 2146 drivers/net/wireless/ath/wil6210/wmi.c led_id, le32_to_cpu(reply.evt.status)); evt 2165 drivers/net/wireless/ath/wil6210/wmi.c struct wmi_rbufcap_cfg_event evt; evt 2167 drivers/net/wireless/ath/wil6210/wmi.c .evt = {.status = WMI_FW_STATUS_FAILURE}, evt 2176 drivers/net/wireless/ath/wil6210/wmi.c if (reply.evt.status != WMI_FW_STATUS_SUCCESS) { evt 2178 drivers/net/wireless/ath/wil6210/wmi.c reply.evt.status); evt 2207 drivers/net/wireless/ath/wil6210/wmi.c struct wmi_pcp_started_event evt; evt 2209 drivers/net/wireless/ath/wil6210/wmi.c .evt = {.status = WMI_FW_STATUS_FAILURE}, evt 2238 drivers/net/wireless/ath/wil6210/wmi.c if (reply.evt.status != WMI_FW_STATUS_SUCCESS) evt 2352 drivers/net/wireless/ath/wil6210/wmi.c struct wmi_p2p_cfg_done_event evt; evt 2354 drivers/net/wireless/ath/wil6210/wmi.c .evt = {.status = WMI_FW_STATUS_FAILURE}, evt 2361 drivers/net/wireless/ath/wil6210/wmi.c if (!rc && reply.evt.status != WMI_FW_STATUS_SUCCESS) { evt 2362 drivers/net/wireless/ath/wil6210/wmi.c wil_err(wil, "P2P_CFG failed. status %d\n", reply.evt.status); evt 2375 drivers/net/wireless/ath/wil6210/wmi.c struct wmi_listen_started_event evt; evt 2377 drivers/net/wireless/ath/wil6210/wmi.c .evt = {.status = WMI_FW_STATUS_FAILURE}, evt 2384 drivers/net/wireless/ath/wil6210/wmi.c if (!rc && reply.evt.status != WMI_FW_STATUS_SUCCESS) { evt 2386 drivers/net/wireless/ath/wil6210/wmi.c reply.evt.status); evt 2399 drivers/net/wireless/ath/wil6210/wmi.c struct wmi_search_started_event evt; evt 2401 drivers/net/wireless/ath/wil6210/wmi.c .evt = {.status = WMI_FW_STATUS_FAILURE}, evt 2408 drivers/net/wireless/ath/wil6210/wmi.c if (!rc && reply.evt.status != WMI_FW_STATUS_SUCCESS) { evt 2410 drivers/net/wireless/ath/wil6210/wmi.c reply.evt.status); evt 2571 drivers/net/wireless/ath/wil6210/wmi.c struct wmi_listen_started_event evt; evt 2573 drivers/net/wireless/ath/wil6210/wmi.c .evt = {.status = WMI_FW_STATUS_FAILURE}, evt 2583 drivers/net/wireless/ath/wil6210/wmi.c if ((rc == 0) && (reply.evt.status != WMI_FW_STATUS_SUCCESS)) evt 2614 drivers/net/wireless/ath/wil6210/wmi.c struct wmi_cfg_rx_chain_done_event evt; evt 2615 drivers/net/wireless/ath/wil6210/wmi.c } __packed evt; evt 2618 drivers/net/wireless/ath/wil6210/wmi.c memset(&evt, 0, sizeof(evt)); evt 2645 drivers/net/wireless/ath/wil6210/wmi.c WMI_CFG_RX_CHAIN_DONE_EVENTID, &evt, sizeof(evt), 2000); evt 2649 drivers/net/wireless/ath/wil6210/wmi.c if (le32_to_cpu(evt.evt.status) != WMI_CFG_RX_CHAIN_SUCCESS) evt 2652 drivers/net/wireless/ath/wil6210/wmi.c vring->hwtail = le32_to_cpu(evt.evt.rx_ring_tail_ptr); evt 2655 drivers/net/wireless/ath/wil6210/wmi.c le32_to_cpu(evt.evt.status), vring->hwtail); evt 2671 drivers/net/wireless/ath/wil6210/wmi.c struct wmi_temp_sense_done_event evt; evt 2683 drivers/net/wireless/ath/wil6210/wmi.c *t_bb = le32_to_cpu(reply.evt.baseband_t1000); evt 2685 drivers/net/wireless/ath/wil6210/wmi.c *t_rf = le32_to_cpu(reply.evt.rf_t1000); evt 2703 drivers/net/wireless/ath/wil6210/wmi.c struct wmi_temp_sense_all_done_event evt; evt 2712 drivers/net/wireless/ath/wil6210/wmi.c reply.evt.status = WMI_FW_STATUS_FAILURE; evt 2719 drivers/net/wireless/ath/wil6210/wmi.c if (reply.evt.status == WMI_FW_STATUS_FAILURE) { evt 2724 drivers/net/wireless/ath/wil6210/wmi.c memcpy(sense_all_evt, &reply.evt, sizeof(reply.evt)); evt 2741 drivers/net/wireless/ath/wil6210/wmi.c struct wmi_disconnect_event evt; evt 2840 drivers/net/wireless/ath/wil6210/wmi.c struct wmi_rcp_addba_resp_sent_event evt; evt 2842 drivers/net/wireless/ath/wil6210/wmi.c .evt = {.status = cpu_to_le16(WMI_FW_STATUS_FAILURE)}, evt 2864 drivers/net/wireless/ath/wil6210/wmi.c if (reply.evt.status) { evt 2866 drivers/net/wireless/ath/wil6210/wmi.c le16_to_cpu(reply.evt.status)); evt 2896 drivers/net/wireless/ath/wil6210/wmi.c struct wmi_rcp_addba_resp_sent_event evt; evt 2898 drivers/net/wireless/ath/wil6210/wmi.c .evt = {.status = cpu_to_le16(WMI_FW_STATUS_FAILURE)}, evt 2912 drivers/net/wireless/ath/wil6210/wmi.c if (reply.evt.status) { evt 2914 drivers/net/wireless/ath/wil6210/wmi.c le16_to_cpu(reply.evt.status)); evt 2931 drivers/net/wireless/ath/wil6210/wmi.c struct wmi_ps_dev_profile_cfg_event evt; evt 2933 drivers/net/wireless/ath/wil6210/wmi.c .evt = {.status = cpu_to_le32(WMI_PS_CFG_CMD_STATUS_ERROR)}, evt 2946 drivers/net/wireless/ath/wil6210/wmi.c status = le32_to_cpu(reply.evt.status); evt 2966 drivers/net/wireless/ath/wil6210/wmi.c struct wmi_set_mgmt_retry_limit_event evt; evt 2968 drivers/net/wireless/ath/wil6210/wmi.c .evt = {.status = WMI_FW_STATUS_FAILURE}, evt 2983 drivers/net/wireless/ath/wil6210/wmi.c if (reply.evt.status != WMI_FW_STATUS_SUCCESS) { evt 2985 drivers/net/wireless/ath/wil6210/wmi.c reply.evt.status); evt 2998 drivers/net/wireless/ath/wil6210/wmi.c struct wmi_get_mgmt_retry_limit_event evt; evt 3014 drivers/net/wireless/ath/wil6210/wmi.c *retry_short = reply.evt.mgmt_retry_limit; evt 3055 drivers/net/wireless/ath/wil6210/wmi.c struct pending_wmi_event *evt, *t; evt 3061 drivers/net/wireless/ath/wil6210/wmi.c list_for_each_entry_safe(evt, t, &wil->pending_wmi_ev, list) { evt 3062 drivers/net/wireless/ath/wil6210/wmi.c list_del(&evt->list); evt 3063 drivers/net/wireless/ath/wil6210/wmi.c kfree(evt); evt 3092 drivers/net/wireless/ath/wil6210/wmi.c struct wmi_traffic_suspend_event evt; evt 3094 drivers/net/wireless/ath/wil6210/wmi.c .evt = {.status = WMI_TRAFFIC_SUSPEND_REJECTED_LINK_NOT_IDLE}, evt 3133 drivers/net/wireless/ath/wil6210/wmi.c if (reply.evt.status != WMI_TRAFFIC_SUSPEND_APPROVED) { evt 3135 drivers/net/wireless/ath/wil6210/wmi.c suspend_status2name(reply.evt.status)); evt 3138 drivers/net/wireless/ath/wil6210/wmi.c rc = reply.evt.status; evt 3179 drivers/net/wireless/ath/wil6210/wmi.c struct wmi_traffic_resume_event evt; evt 3181 drivers/net/wireless/ath/wil6210/wmi.c .evt = {.status = WMI_TRAFFIC_RESUME_FAILED, evt 3191 drivers/net/wireless/ath/wil6210/wmi.c resume_triggers2string(le32_to_cpu(reply.evt.resume_triggers), string, evt 3194 drivers/net/wireless/ath/wil6210/wmi.c reply.evt.status ? "failed" : "passed", string, evt 3195 drivers/net/wireless/ath/wil6210/wmi.c le32_to_cpu(reply.evt.resume_triggers)); evt 3197 drivers/net/wireless/ath/wil6210/wmi.c return reply.evt.status; evt 3209 drivers/net/wireless/ath/wil6210/wmi.c struct wmi_port_allocated_event evt; evt 3211 drivers/net/wireless/ath/wil6210/wmi.c .evt = {.status = WMI_FW_STATUS_FAILURE}, evt 3245 drivers/net/wireless/ath/wil6210/wmi.c if (reply.evt.status != WMI_FW_STATUS_SUCCESS) { evt 3247 drivers/net/wireless/ath/wil6210/wmi.c reply.evt.status); evt 3262 drivers/net/wireless/ath/wil6210/wmi.c struct wmi_port_deleted_event evt; evt 3264 drivers/net/wireless/ath/wil6210/wmi.c .evt = {.status = WMI_FW_STATUS_FAILURE}, evt 3277 drivers/net/wireless/ath/wil6210/wmi.c if (reply.evt.status != WMI_FW_STATUS_SUCCESS) { evt 3279 drivers/net/wireless/ath/wil6210/wmi.c reply.evt.status); evt 3395 drivers/net/wireless/ath/wil6210/wmi.c struct pending_wmi_event *evt; evt 3400 drivers/net/wireless/ath/wil6210/wmi.c evt = list_entry(lh, struct pending_wmi_event, list); evt 3401 drivers/net/wireless/ath/wil6210/wmi.c wmi_event_handle(wil, &evt->event.hdr); evt 3402 drivers/net/wireless/ath/wil6210/wmi.c kfree(evt); evt 3535 drivers/net/wireless/ath/wil6210/wmi.c struct wmi_start_sched_scan_event evt; evt 3537 drivers/net/wireless/ath/wil6210/wmi.c .evt = {.result = WMI_PNO_REJECT}, evt 3561 drivers/net/wireless/ath/wil6210/wmi.c if (reply.evt.result != WMI_PNO_SUCCESS) { evt 3563 drivers/net/wireless/ath/wil6210/wmi.c reply.evt.result); evt 3576 drivers/net/wireless/ath/wil6210/wmi.c struct wmi_stop_sched_scan_event evt; evt 3578 drivers/net/wireless/ath/wil6210/wmi.c .evt = {.result = WMI_PNO_REJECT}, evt 3590 drivers/net/wireless/ath/wil6210/wmi.c if (reply.evt.result != WMI_PNO_SUCCESS) { evt 3592 drivers/net/wireless/ath/wil6210/wmi.c reply.evt.result); evt 3607 drivers/net/wireless/ath/wil6210/wmi.c struct wmi_sw_tx_complete_event evt; evt 3608 drivers/net/wireless/ath/wil6210/wmi.c } __packed evt = { evt 3609 drivers/net/wireless/ath/wil6210/wmi.c .evt = {.status = WMI_FW_STATUS_FAILURE}, evt 3635 drivers/net/wireless/ath/wil6210/wmi.c WMI_SW_TX_COMPLETE_EVENTID, &evt, sizeof(evt), 2000); evt 3636 drivers/net/wireless/ath/wil6210/wmi.c if (!rc && evt.evt.status != WMI_FW_STATUS_SUCCESS) { evt 3638 drivers/net/wireless/ath/wil6210/wmi.c evt.evt.status); evt 3656 drivers/net/wireless/ath/wil6210/wmi.c struct wmi_sw_tx_complete_event evt; evt 3657 drivers/net/wireless/ath/wil6210/wmi.c } __packed evt = { evt 3658 drivers/net/wireless/ath/wil6210/wmi.c .evt = {.status = WMI_FW_STATUS_FAILURE}, evt 3689 drivers/net/wireless/ath/wil6210/wmi.c WMI_SW_TX_COMPLETE_EVENTID, &evt, sizeof(evt), 2000); evt 3690 drivers/net/wireless/ath/wil6210/wmi.c if (!rc && evt.evt.status != WMI_FW_STATUS_SUCCESS) { evt 3692 drivers/net/wireless/ath/wil6210/wmi.c evt.evt.status); evt 3714 drivers/net/wireless/ath/wil6210/wmi.c struct wmi_tx_status_ring_cfg_done_event evt; evt 3716 drivers/net/wireless/ath/wil6210/wmi.c .evt = {.status = WMI_FW_STATUS_FAILURE}, evt 3730 drivers/net/wireless/ath/wil6210/wmi.c if (reply.evt.status != WMI_FW_STATUS_SUCCESS) { evt 3732 drivers/net/wireless/ath/wil6210/wmi.c reply.evt.status); evt 3736 drivers/net/wireless/ath/wil6210/wmi.c sring->hwtail = le32_to_cpu(reply.evt.ring_tail_ptr); evt 3755 drivers/net/wireless/ath/wil6210/wmi.c struct wmi_cfg_def_rx_offload_done_event evt; evt 3757 drivers/net/wireless/ath/wil6210/wmi.c .evt = {.status = WMI_FW_STATUS_FAILURE}, evt 3768 drivers/net/wireless/ath/wil6210/wmi.c if (reply.evt.status != WMI_FW_STATUS_SUCCESS) { evt 3770 drivers/net/wireless/ath/wil6210/wmi.c reply.evt.status); evt 3795 drivers/net/wireless/ath/wil6210/wmi.c struct wmi_rx_status_ring_cfg_done_event evt; evt 3797 drivers/net/wireless/ath/wil6210/wmi.c .evt = {.status = WMI_FW_STATUS_FAILURE}, evt 3809 drivers/net/wireless/ath/wil6210/wmi.c if (reply.evt.status != WMI_FW_STATUS_SUCCESS) { evt 3811 drivers/net/wireless/ath/wil6210/wmi.c reply.evt.status); evt 3815 drivers/net/wireless/ath/wil6210/wmi.c sring->hwtail = le32_to_cpu(reply.evt.ring_tail_ptr); evt 3836 drivers/net/wireless/ath/wil6210/wmi.c struct wmi_rx_desc_ring_cfg_done_event evt; evt 3838 drivers/net/wireless/ath/wil6210/wmi.c .evt = {.status = WMI_FW_STATUS_FAILURE}, evt 3851 drivers/net/wireless/ath/wil6210/wmi.c if (reply.evt.status != WMI_FW_STATUS_SUCCESS) { evt 3853 drivers/net/wireless/ath/wil6210/wmi.c reply.evt.status); evt 3857 drivers/net/wireless/ath/wil6210/wmi.c ring->hwtail = le32_to_cpu(reply.evt.ring_tail_ptr); evt 3887 drivers/net/wireless/ath/wil6210/wmi.c struct wmi_tx_desc_ring_cfg_done_event evt; evt 3889 drivers/net/wireless/ath/wil6210/wmi.c .evt = {.status = WMI_FW_STATUS_FAILURE}, evt 3901 drivers/net/wireless/ath/wil6210/wmi.c if (reply.evt.status != WMI_FW_STATUS_SUCCESS) { evt 3903 drivers/net/wireless/ath/wil6210/wmi.c reply.evt.status); evt 3908 drivers/net/wireless/ath/wil6210/wmi.c ring->hwtail = le32_to_cpu(reply.evt.ring_tail_ptr); evt 3932 drivers/net/wireless/ath/wil6210/wmi.c struct wmi_rx_desc_ring_cfg_done_event evt; evt 3934 drivers/net/wireless/ath/wil6210/wmi.c .evt = {.status = WMI_FW_STATUS_FAILURE}, evt 3947 drivers/net/wireless/ath/wil6210/wmi.c if (reply.evt.status != WMI_FW_STATUS_SUCCESS) { evt 3949 drivers/net/wireless/ath/wil6210/wmi.c reply.evt.status); evt 3954 drivers/net/wireless/ath/wil6210/wmi.c ring->hwtail = le32_to_cpu(reply.evt.ring_tail_ptr); evt 3973 drivers/net/wireless/ath/wil6210/wmi.c struct wmi_link_stats_config_done_event evt; evt 3975 drivers/net/wireless/ath/wil6210/wmi.c .evt = {.status = WMI_FW_STATUS_FAILURE}, evt 3987 drivers/net/wireless/ath/wil6210/wmi.c if (reply.evt.status != WMI_FW_STATUS_SUCCESS) { evt 3989 drivers/net/wireless/ath/wil6210/wmi.c reply.evt.status); evt 975 drivers/net/wireless/st/cw1200/sta.c switch (event->evt.id) { evt 1008 drivers/net/wireless/st/cw1200/sta.c int rcpi_rssi = (int)(event->evt.data & 0xFF); evt 929 drivers/net/wireless/st/cw1200/wsm.c event->evt.id = WSM_GET32(buf); evt 930 drivers/net/wireless/st/cw1200/wsm.c event->evt.data = WSM_GET32(buf); evt 933 drivers/net/wireless/st/cw1200/wsm.c event->evt.id, event->evt.data); evt 882 drivers/net/wireless/st/cw1200/wsm.h struct wsm_event evt; evt 205 drivers/perf/arm_dsu_pmu.c unsigned long evt = (unsigned long)eattr->var; evt 207 drivers/perf/arm_dsu_pmu.c return test_bit(evt, dsu_pmu->cpmceid_bitmap) ? attr->mode : 0; evt 314 drivers/ptp/ptp_clock.c struct pps_event_time evt; evt 327 drivers/ptp/ptp_clock.c pps_get_ts(&evt); evt 328 drivers/ptp/ptp_clock.c pps_event(ptp->pps_source, &evt, PTP_PPS_EVENT, NULL); evt 354 drivers/regulator/slg51000-regulator.c u8 evt[SLG51000_MAX_EVT_REGISTER][REG_MAX]; evt 360 drivers/regulator/slg51000-regulator.c ret = regmap_bulk_read(regmap, es_reg[i].ereg, evt[i], REG_MAX); evt 390 drivers/regulator/slg51000-regulator.c if (!(evt[i][R2] & SLG51000_IRQ_ILIM_FLAG_MASK) && evt 391 drivers/regulator/slg51000-regulator.c (evt[i][R0] & SLG51000_EVT_ILIM_FLAG_MASK)) { evt 397 drivers/regulator/slg51000-regulator.c if (evt[i][R1] & SLG51000_STA_ILIM_FLAG_MASK) evt 404 drivers/regulator/slg51000-regulator.c if (!(evt[SLG51000_SCTL_EVT][R2] & SLG51000_IRQ_HIGH_TEMP_WARN_MASK) && evt 405 drivers/regulator/slg51000-regulator.c (evt[SLG51000_SCTL_EVT][R0] & SLG51000_EVT_HIGH_TEMP_WARN_MASK)) { evt 407 drivers/regulator/slg51000-regulator.c if (!(evt[i][R1] & SLG51000_STA_ILIM_FLAG_MASK) && evt 408 drivers/regulator/slg51000-regulator.c (evt[i][R1] & SLG51000_STA_VOUT_OK_FLAG_MASK)) { evt 416 drivers/regulator/slg51000-regulator.c if (evt[SLG51000_SCTL_EVT][R1] & evt 365 drivers/scsi/be2iscsi/be_cmds.c struct be_async_event_link_state *evt; evt 367 drivers/scsi/be2iscsi/be_cmds.c evt = (struct be_async_event_link_state *)compl; evt 369 drivers/scsi/be2iscsi/be_cmds.c phba->port_speed = evt->port_speed; evt 374 drivers/scsi/be2iscsi/be_cmds.c if (evt->port_link_status & BE_ASYNC_LINK_UP_MASK) { evt 380 drivers/scsi/be2iscsi/be_cmds.c evt->physical_port, evt->event_tag); evt 385 drivers/scsi/be2iscsi/be_cmds.c evt->physical_port, evt->event_tag); evt 2791 drivers/scsi/bfa/bfa_fcpim.c enum bfa_ioim_event evt = BFA_IOIM_SM_COMP; evt 2809 drivers/scsi/bfa/bfa_fcpim.c evt = BFA_IOIM_SM_DONE; evt 2811 drivers/scsi/bfa/bfa_fcpim.c evt = BFA_IOIM_SM_COMP; evt 2821 drivers/scsi/bfa/bfa_fcpim.c evt = BFA_IOIM_SM_DONE; evt 2823 drivers/scsi/bfa/bfa_fcpim.c evt = BFA_IOIM_SM_COMP; evt 2829 drivers/scsi/bfa/bfa_fcpim.c evt = BFA_IOIM_SM_COMP; evt 2835 drivers/scsi/bfa/bfa_fcpim.c evt = BFA_IOIM_SM_SQRETRY; evt 2840 drivers/scsi/bfa/bfa_fcpim.c evt = BFA_IOIM_SM_FREE; evt 2852 drivers/scsi/bfa/bfa_fcpim.c evt = BFA_IOIM_SM_ABORT_COMP; evt 2854 drivers/scsi/bfa/bfa_fcpim.c evt = BFA_IOIM_SM_ABORT_DONE; evt 2859 drivers/scsi/bfa/bfa_fcpim.c evt = BFA_IOIM_SM_COMP_UTAG; evt 2866 drivers/scsi/bfa/bfa_fcpim.c bfa_sm_send_event(ioim, evt); evt 138 drivers/scsi/bfa/bfad_im.h int evt) evt 153 drivers/scsi/bfa/bfad_im.h entry->aen_type = evt; evt 96 drivers/scsi/csiostor/csio_defs.h csio_post_event(void *smp, uint32_t evt) evt 98 drivers/scsi/csiostor/csio_defs.h ((struct csio_sm *)smp)->sm_state(smp, evt); evt 2780 drivers/scsi/csiostor/csio_hw.c csio_hws_uninit(struct csio_hw *hw, enum csio_hw_ev evt) evt 2783 drivers/scsi/csiostor/csio_hw.c hw->cur_evt = evt; evt 2784 drivers/scsi/csiostor/csio_hw.c CSIO_INC_STATS(hw, n_evt_sm[evt]); evt 2786 drivers/scsi/csiostor/csio_hw.c switch (evt) { evt 2805 drivers/scsi/csiostor/csio_hw.c csio_hws_configuring(struct csio_hw *hw, enum csio_hw_ev evt) evt 2808 drivers/scsi/csiostor/csio_hw.c hw->cur_evt = evt; evt 2809 drivers/scsi/csiostor/csio_hw.c CSIO_INC_STATS(hw, n_evt_sm[evt]); evt 2811 drivers/scsi/csiostor/csio_hw.c switch (evt) { evt 2843 drivers/scsi/csiostor/csio_hw.c csio_hws_initializing(struct csio_hw *hw, enum csio_hw_ev evt) evt 2846 drivers/scsi/csiostor/csio_hw.c hw->cur_evt = evt; evt 2847 drivers/scsi/csiostor/csio_hw.c CSIO_INC_STATS(hw, n_evt_sm[evt]); evt 2849 drivers/scsi/csiostor/csio_hw.c switch (evt) { evt 2881 drivers/scsi/csiostor/csio_hw.c csio_hws_ready(struct csio_hw *hw, enum csio_hw_ev evt) evt 2884 drivers/scsi/csiostor/csio_hw.c hw->evtflag = evt; evt 2887 drivers/scsi/csiostor/csio_hw.c hw->cur_evt = evt; evt 2888 drivers/scsi/csiostor/csio_hw.c CSIO_INC_STATS(hw, n_evt_sm[evt]); evt 2890 drivers/scsi/csiostor/csio_hw.c switch (evt) { evt 2898 drivers/scsi/csiostor/csio_hw.c if (evt == CSIO_HWE_HBA_RESET || evt 2899 drivers/scsi/csiostor/csio_hw.c evt == CSIO_HWE_PCIERR_DETECTED) evt 2930 drivers/scsi/csiostor/csio_hw.c csio_hws_quiescing(struct csio_hw *hw, enum csio_hw_ev evt) evt 2933 drivers/scsi/csiostor/csio_hw.c hw->cur_evt = evt; evt 2934 drivers/scsi/csiostor/csio_hw.c CSIO_INC_STATS(hw, n_evt_sm[evt]); evt 2936 drivers/scsi/csiostor/csio_hw.c switch (evt) { evt 2990 drivers/scsi/csiostor/csio_hw.c csio_hws_quiesced(struct csio_hw *hw, enum csio_hw_ev evt) evt 2993 drivers/scsi/csiostor/csio_hw.c hw->cur_evt = evt; evt 2994 drivers/scsi/csiostor/csio_hw.c CSIO_INC_STATS(hw, n_evt_sm[evt]); evt 2996 drivers/scsi/csiostor/csio_hw.c switch (evt) { evt 3015 drivers/scsi/csiostor/csio_hw.c csio_hws_resetting(struct csio_hw *hw, enum csio_hw_ev evt) evt 3018 drivers/scsi/csiostor/csio_hw.c hw->cur_evt = evt; evt 3019 drivers/scsi/csiostor/csio_hw.c CSIO_INC_STATS(hw, n_evt_sm[evt]); evt 3021 drivers/scsi/csiostor/csio_hw.c switch (evt) { evt 3041 drivers/scsi/csiostor/csio_hw.c csio_hws_removing(struct csio_hw *hw, enum csio_hw_ev evt) evt 3044 drivers/scsi/csiostor/csio_hw.c hw->cur_evt = evt; evt 3045 drivers/scsi/csiostor/csio_hw.c CSIO_INC_STATS(hw, n_evt_sm[evt]); evt 3047 drivers/scsi/csiostor/csio_hw.c switch (evt) { evt 3076 drivers/scsi/csiostor/csio_hw.c csio_hws_pcierr(struct csio_hw *hw, enum csio_hw_ev evt) evt 3079 drivers/scsi/csiostor/csio_hw.c hw->cur_evt = evt; evt 3080 drivers/scsi/csiostor/csio_hw.c CSIO_INC_STATS(hw, n_evt_sm[evt]); evt 3082 drivers/scsi/csiostor/csio_hw.c switch (evt) { evt 935 drivers/scsi/csiostor/csio_lnode.c csio_post_event_rns(struct csio_lnode *ln, enum csio_rn_ev evt) evt 943 drivers/scsi/csiostor/csio_lnode.c csio_post_event(&rn->sm, evt); evt 983 drivers/scsi/csiostor/csio_lnode.c csio_post_event_lns(struct csio_lnode *ln, enum csio_ln_ev evt) evt 990 drivers/scsi/csiostor/csio_lnode.c csio_post_event(&ln->sm, evt); evt 998 drivers/scsi/csiostor/csio_lnode.c csio_post_event(&cln->sm, evt); evt 1002 drivers/scsi/csiostor/csio_lnode.c csio_post_event(&ln->sm, evt); evt 1114 drivers/scsi/csiostor/csio_lnode.c csio_lns_uninit(struct csio_lnode *ln, enum csio_ln_ev evt) evt 1120 drivers/scsi/csiostor/csio_lnode.c CSIO_INC_STATS(ln, n_evt_sm[evt]); evt 1121 drivers/scsi/csiostor/csio_lnode.c switch (evt) { evt 1151 drivers/scsi/csiostor/csio_lnode.c "ln state[uninit].\n", evt, ln->nport_id); evt 1167 drivers/scsi/csiostor/csio_lnode.c csio_lns_online(struct csio_lnode *ln, enum csio_ln_ev evt) evt 1171 drivers/scsi/csiostor/csio_lnode.c CSIO_INC_STATS(ln, n_evt_sm[evt]); evt 1172 drivers/scsi/csiostor/csio_lnode.c switch (evt) { evt 1202 drivers/scsi/csiostor/csio_lnode.c "ln state[uninit].\n", evt, ln->nport_id); evt 1219 drivers/scsi/csiostor/csio_lnode.c csio_lns_ready(struct csio_lnode *ln, enum csio_ln_ev evt) evt 1223 drivers/scsi/csiostor/csio_lnode.c CSIO_INC_STATS(ln, n_evt_sm[evt]); evt 1224 drivers/scsi/csiostor/csio_lnode.c switch (evt) { evt 1228 drivers/scsi/csiostor/csio_lnode.c "in ln state[ready].\n", evt, ln->nport_id); evt 1276 drivers/scsi/csiostor/csio_lnode.c "ln state[uninit].\n", evt, ln->nport_id); evt 1293 drivers/scsi/csiostor/csio_lnode.c csio_lns_offline(struct csio_lnode *ln, enum csio_ln_ev evt) evt 1299 drivers/scsi/csiostor/csio_lnode.c CSIO_INC_STATS(ln, n_evt_sm[evt]); evt 1300 drivers/scsi/csiostor/csio_lnode.c switch (evt) { evt 1329 drivers/scsi/csiostor/csio_lnode.c "in ln state[offline].\n", evt, ln->nport_id); evt 1341 drivers/scsi/csiostor/csio_lnode.c "ln state[offline]\n", evt, ln->nport_id); evt 1494 drivers/scsi/csiostor/csio_lnode.c enum csio_ln_ev evt; evt 1579 drivers/scsi/csiostor/csio_lnode.c evt = CSIO_FWE_TO_LNE(rdev_wr->event_cause); evt 1580 drivers/scsi/csiostor/csio_lnode.c if (evt) { evt 1583 drivers/scsi/csiostor/csio_lnode.c "cause:%d flowid:x%x\n", evt, evt 1585 drivers/scsi/csiostor/csio_lnode.c csio_post_event(&ln->sm, evt); evt 604 drivers/scsi/csiostor/csio_rnode.c csio_rns_uninit(struct csio_rnode *rn, enum csio_rn_ev evt) evt 609 drivers/scsi/csiostor/csio_rnode.c CSIO_INC_STATS(rn, n_evt_sm[evt]); evt 611 drivers/scsi/csiostor/csio_rnode.c switch (evt) { evt 625 drivers/scsi/csiostor/csio_rnode.c "in rn state[uninit]\n", csio_rn_flowid(rn), evt); evt 631 drivers/scsi/csiostor/csio_rnode.c "in rn state[uninit]\n", csio_rn_flowid(rn), evt); evt 644 drivers/scsi/csiostor/csio_rnode.c csio_rns_ready(struct csio_rnode *rn, enum csio_rn_ev evt) evt 649 drivers/scsi/csiostor/csio_rnode.c CSIO_INC_STATS(rn, n_evt_sm[evt]); evt 651 drivers/scsi/csiostor/csio_rnode.c switch (evt) { evt 656 drivers/scsi/csiostor/csio_rnode.c "in rn state[ready]\n", csio_rn_flowid(rn), evt, evt 715 drivers/scsi/csiostor/csio_rnode.c "in rn state[uninit]\n", csio_rn_flowid(rn), evt, evt 729 drivers/scsi/csiostor/csio_rnode.c csio_rns_offline(struct csio_rnode *rn, enum csio_rn_ev evt) evt 734 drivers/scsi/csiostor/csio_rnode.c CSIO_INC_STATS(rn, n_evt_sm[evt]); evt 736 drivers/scsi/csiostor/csio_rnode.c switch (evt) { evt 752 drivers/scsi/csiostor/csio_rnode.c "in rn state[offline]\n", csio_rn_flowid(rn), evt, evt 774 drivers/scsi/csiostor/csio_rnode.c "in rn state[offline]\n", csio_rn_flowid(rn), evt, evt 788 drivers/scsi/csiostor/csio_rnode.c csio_rns_disappeared(struct csio_rnode *rn, enum csio_rn_ev evt) evt 793 drivers/scsi/csiostor/csio_rnode.c CSIO_INC_STATS(rn, n_evt_sm[evt]); evt 795 drivers/scsi/csiostor/csio_rnode.c switch (evt) { evt 823 drivers/scsi/csiostor/csio_rnode.c evt, rn->nport_id); evt 830 drivers/scsi/csiostor/csio_rnode.c evt, rn->nport_id); evt 871 drivers/scsi/csiostor/csio_rnode.c enum csio_rn_ev evt; evt 873 drivers/scsi/csiostor/csio_rnode.c evt = CSIO_FWE_TO_RNFE(fwevt); evt 874 drivers/scsi/csiostor/csio_rnode.c if (!evt) { evt 887 drivers/scsi/csiostor/csio_rnode.c csio_post_event(&rn->sm, evt); evt 703 drivers/scsi/csiostor/csio_scsi.c csio_scsis_uninit(struct csio_ioreq *req, enum csio_scsi_ev evt) evt 708 drivers/scsi/csiostor/csio_scsi.c switch (evt) { evt 767 drivers/scsi/csiostor/csio_scsi.c csio_dbg(hw, "Unhandled event:%d sent to req:%p\n", evt, req); evt 773 drivers/scsi/csiostor/csio_scsi.c csio_scsis_io_active(struct csio_ioreq *req, enum csio_scsi_ev evt) evt 779 drivers/scsi/csiostor/csio_scsi.c switch (evt) { evt 839 drivers/scsi/csiostor/csio_scsi.c csio_dbg(hw, "Unhandled event:%d sent to req:%p\n", evt, req); evt 845 drivers/scsi/csiostor/csio_scsi.c csio_scsis_tm_active(struct csio_ioreq *req, enum csio_scsi_ev evt) evt 850 drivers/scsi/csiostor/csio_scsi.c switch (evt) { evt 882 drivers/scsi/csiostor/csio_scsi.c csio_dbg(hw, "Unhandled event:%d sent to req:%p\n", evt, req); evt 888 drivers/scsi/csiostor/csio_scsi.c csio_scsis_aborting(struct csio_ioreq *req, enum csio_scsi_ev evt) evt 893 drivers/scsi/csiostor/csio_scsi.c switch (evt) { evt 979 drivers/scsi/csiostor/csio_scsi.c csio_dbg(hw, "Unhandled event:%d sent to req:%p\n", evt, req); evt 985 drivers/scsi/csiostor/csio_scsi.c csio_scsis_closing(struct csio_ioreq *req, enum csio_scsi_ev evt) evt 990 drivers/scsi/csiostor/csio_scsi.c switch (evt) { evt 1043 drivers/scsi/csiostor/csio_scsi.c csio_dbg(hw, "Unhandled event:%d sent to req:%p\n", evt, req); evt 1049 drivers/scsi/csiostor/csio_scsi.c csio_scsis_shost_cmpl_await(struct csio_ioreq *req, enum csio_scsi_ev evt) evt 1051 drivers/scsi/csiostor/csio_scsi.c switch (evt) { evt 1077 drivers/scsi/csiostor/csio_scsi.c evt, req); evt 131 drivers/scsi/gdth.c u16 idx, gdth_evt_data *evt); evt 2359 drivers/scsi/gdth.c u16 idx, gdth_evt_data *evt) evt 2370 drivers/scsi/gdth.c ((evt->size != 0 && ebuffer[elastidx].event_data.size != 0 && evt 2372 drivers/scsi/gdth.c (char *)&evt->eu, evt->size)) || evt 2373 drivers/scsi/gdth.c (evt->size == 0 && ebuffer[elastidx].event_data.size == 0 && evt 2375 drivers/scsi/gdth.c (char *)&evt->event_string)))) { evt 2395 drivers/scsi/gdth.c e->event_data = *evt; evt 3449 drivers/scsi/gdth.c gdth_ioctl_event evt; evt 3453 drivers/scsi/gdth.c if (copy_from_user(&evt, arg, sizeof(gdth_ioctl_event))) evt 3455 drivers/scsi/gdth.c ha = gdth_find_ha(evt.ionode); evt 3459 drivers/scsi/gdth.c if (evt.erase == 0xff) { evt 3460 drivers/scsi/gdth.c if (evt.event.event_source == ES_TEST) evt 3461 drivers/scsi/gdth.c evt.event.event_data.size=sizeof(evt.event.event_data.eu.test); evt 3462 drivers/scsi/gdth.c else if (evt.event.event_source == ES_DRIVER) evt 3463 drivers/scsi/gdth.c evt.event.event_data.size=sizeof(evt.event.event_data.eu.driver); evt 3464 drivers/scsi/gdth.c else if (evt.event.event_source == ES_SYNC) evt 3465 drivers/scsi/gdth.c evt.event.event_data.size=sizeof(evt.event.event_data.eu.sync); evt 3467 drivers/scsi/gdth.c evt.event.event_data.size=sizeof(evt.event.event_data.eu.async); evt 3469 drivers/scsi/gdth.c gdth_store_event(ha, evt.event.event_source, evt.event.event_idx, evt 3470 drivers/scsi/gdth.c &evt.event.event_data); evt 3472 drivers/scsi/gdth.c } else if (evt.erase == 0xfe) { evt 3474 drivers/scsi/gdth.c } else if (evt.erase == 0) { evt 3475 drivers/scsi/gdth.c evt.handle = gdth_read_event(ha, evt.handle, &evt.event); evt 3477 drivers/scsi/gdth.c gdth_readapp_event(ha, evt.erase, &evt.event); evt 3479 drivers/scsi/gdth.c if (copy_to_user(arg, &evt, sizeof(gdth_ioctl_event))) evt 145 drivers/scsi/ibmvscsi/ibmvfc.c static void ibmvfc_trc_start(struct ibmvfc_event *evt) evt 147 drivers/scsi/ibmvscsi/ibmvfc.c struct ibmvfc_host *vhost = evt->vhost; evt 148 drivers/scsi/ibmvscsi/ibmvfc.c struct ibmvfc_cmd *vfc_cmd = &evt->iu.cmd; evt 149 drivers/scsi/ibmvscsi/ibmvfc.c struct ibmvfc_mad_common *mad = &evt->iu.mad_common; evt 153 drivers/scsi/ibmvscsi/ibmvfc.c entry->evt = evt; evt 155 drivers/scsi/ibmvscsi/ibmvfc.c entry->fmt = evt->crq.format; evt 179 drivers/scsi/ibmvscsi/ibmvfc.c static void ibmvfc_trc_end(struct ibmvfc_event *evt) evt 181 drivers/scsi/ibmvscsi/ibmvfc.c struct ibmvfc_host *vhost = evt->vhost; evt 182 drivers/scsi/ibmvscsi/ibmvfc.c struct ibmvfc_cmd *vfc_cmd = &evt->xfer_iu->cmd; evt 183 drivers/scsi/ibmvscsi/ibmvfc.c struct ibmvfc_mad_common *mad = &evt->xfer_iu->mad_common; evt 186 drivers/scsi/ibmvscsi/ibmvfc.c entry->evt = evt; evt 188 drivers/scsi/ibmvscsi/ibmvfc.c entry->fmt = evt->crq.format; evt 214 drivers/scsi/ibmvscsi/ibmvfc.c #define ibmvfc_trc_start(evt) do { } while (0) evt 215 drivers/scsi/ibmvscsi/ibmvfc.c #define ibmvfc_trc_end(evt) do { } while (0) evt 740 drivers/scsi/ibmvscsi/ibmvfc.c struct ibmvfc_event *evt) evt 742 drivers/scsi/ibmvscsi/ibmvfc.c int index = evt - pool->events; evt 745 drivers/scsi/ibmvscsi/ibmvfc.c if (evt != pool->events + index) /* unaligned */ evt 755 drivers/scsi/ibmvscsi/ibmvfc.c static void ibmvfc_free_event(struct ibmvfc_event *evt) evt 757 drivers/scsi/ibmvscsi/ibmvfc.c struct ibmvfc_host *vhost = evt->vhost; evt 760 drivers/scsi/ibmvscsi/ibmvfc.c BUG_ON(!ibmvfc_valid_event(pool, evt)); evt 761 drivers/scsi/ibmvscsi/ibmvfc.c BUG_ON(atomic_inc_return(&evt->free) != 1); evt 762 drivers/scsi/ibmvscsi/ibmvfc.c list_add_tail(&evt->queue, &vhost->free); evt 772 drivers/scsi/ibmvscsi/ibmvfc.c static void ibmvfc_scsi_eh_done(struct ibmvfc_event *evt) evt 774 drivers/scsi/ibmvscsi/ibmvfc.c struct scsi_cmnd *cmnd = evt->cmnd; evt 781 drivers/scsi/ibmvscsi/ibmvfc.c if (evt->eh_comp) evt 782 drivers/scsi/ibmvscsi/ibmvfc.c complete(evt->eh_comp); evt 784 drivers/scsi/ibmvscsi/ibmvfc.c ibmvfc_free_event(evt); evt 795 drivers/scsi/ibmvscsi/ibmvfc.c static void ibmvfc_fail_request(struct ibmvfc_event *evt, int error_code) evt 797 drivers/scsi/ibmvscsi/ibmvfc.c if (evt->cmnd) { evt 798 drivers/scsi/ibmvscsi/ibmvfc.c evt->cmnd->result = (error_code << 16); evt 799 drivers/scsi/ibmvscsi/ibmvfc.c evt->done = ibmvfc_scsi_eh_done; evt 801 drivers/scsi/ibmvscsi/ibmvfc.c evt->xfer_iu->mad_common.status = cpu_to_be16(IBMVFC_MAD_DRIVER_FAILED); evt 803 drivers/scsi/ibmvscsi/ibmvfc.c list_del(&evt->queue); evt 804 drivers/scsi/ibmvscsi/ibmvfc.c del_timer(&evt->timer); evt 805 drivers/scsi/ibmvscsi/ibmvfc.c ibmvfc_trc_end(evt); evt 806 drivers/scsi/ibmvscsi/ibmvfc.c evt->done(evt); evt 819 drivers/scsi/ibmvscsi/ibmvfc.c struct ibmvfc_event *evt, *pos; evt 822 drivers/scsi/ibmvscsi/ibmvfc.c list_for_each_entry_safe(evt, pos, &vhost->sent, queue) evt 823 drivers/scsi/ibmvscsi/ibmvfc.c ibmvfc_fail_request(evt, error_code); evt 1216 drivers/scsi/ibmvscsi/ibmvfc.c struct ibmvfc_event *evt = &pool->events[i]; evt 1217 drivers/scsi/ibmvscsi/ibmvfc.c atomic_set(&evt->free, 1); evt 1218 drivers/scsi/ibmvscsi/ibmvfc.c evt->crq.valid = 0x80; evt 1219 drivers/scsi/ibmvscsi/ibmvfc.c evt->crq.ioba = cpu_to_be64(pool->iu_token + (sizeof(*evt->xfer_iu) * i)); evt 1220 drivers/scsi/ibmvscsi/ibmvfc.c evt->xfer_iu = pool->iu_storage + i; evt 1221 drivers/scsi/ibmvscsi/ibmvfc.c evt->vhost = vhost; evt 1222 drivers/scsi/ibmvscsi/ibmvfc.c evt->ext_list = NULL; evt 1223 drivers/scsi/ibmvscsi/ibmvfc.c list_add_tail(&evt->queue, &vhost->free); evt 1265 drivers/scsi/ibmvscsi/ibmvfc.c struct ibmvfc_event *evt; evt 1268 drivers/scsi/ibmvscsi/ibmvfc.c evt = list_entry(vhost->free.next, struct ibmvfc_event, queue); evt 1269 drivers/scsi/ibmvscsi/ibmvfc.c atomic_set(&evt->free, 0); evt 1270 drivers/scsi/ibmvscsi/ibmvfc.c list_del(&evt->queue); evt 1271 drivers/scsi/ibmvscsi/ibmvfc.c return evt; evt 1281 drivers/scsi/ibmvscsi/ibmvfc.c static void ibmvfc_init_event(struct ibmvfc_event *evt, evt 1284 drivers/scsi/ibmvscsi/ibmvfc.c evt->cmnd = NULL; evt 1285 drivers/scsi/ibmvscsi/ibmvfc.c evt->sync_iu = NULL; evt 1286 drivers/scsi/ibmvscsi/ibmvfc.c evt->crq.format = format; evt 1287 drivers/scsi/ibmvscsi/ibmvfc.c evt->done = done; evt 1288 drivers/scsi/ibmvscsi/ibmvfc.c evt->eh_comp = NULL; evt 1321 drivers/scsi/ibmvscsi/ibmvfc.c struct ibmvfc_event *evt, evt 1357 drivers/scsi/ibmvscsi/ibmvfc.c if (!evt->ext_list) { evt 1358 drivers/scsi/ibmvscsi/ibmvfc.c evt->ext_list = dma_pool_alloc(vhost->sg_pool, GFP_ATOMIC, evt 1359 drivers/scsi/ibmvscsi/ibmvfc.c &evt->ext_list_token); evt 1361 drivers/scsi/ibmvscsi/ibmvfc.c if (!evt->ext_list) { evt 1369 drivers/scsi/ibmvscsi/ibmvfc.c ibmvfc_map_sg_list(scmd, sg_mapped, evt->ext_list); evt 1371 drivers/scsi/ibmvscsi/ibmvfc.c data->va = cpu_to_be64(evt->ext_list_token); evt 1385 drivers/scsi/ibmvscsi/ibmvfc.c struct ibmvfc_event *evt = from_timer(evt, t, timer); evt 1386 drivers/scsi/ibmvscsi/ibmvfc.c struct ibmvfc_host *vhost = evt->vhost; evt 1387 drivers/scsi/ibmvscsi/ibmvfc.c dev_err(vhost->dev, "Command timed out (%p). Resetting connection\n", evt); evt 1399 drivers/scsi/ibmvscsi/ibmvfc.c static int ibmvfc_send_event(struct ibmvfc_event *evt, evt 1402 drivers/scsi/ibmvscsi/ibmvfc.c __be64 *crq_as_u64 = (__be64 *) &evt->crq; evt 1406 drivers/scsi/ibmvscsi/ibmvfc.c *evt->xfer_iu = evt->iu; evt 1407 drivers/scsi/ibmvscsi/ibmvfc.c if (evt->crq.format == IBMVFC_CMD_FORMAT) evt 1408 drivers/scsi/ibmvscsi/ibmvfc.c evt->xfer_iu->cmd.tag = cpu_to_be64((u64)evt); evt 1409 drivers/scsi/ibmvscsi/ibmvfc.c else if (evt->crq.format == IBMVFC_MAD_FORMAT) evt 1410 drivers/scsi/ibmvscsi/ibmvfc.c evt->xfer_iu->mad_common.tag = cpu_to_be64((u64)evt); evt 1414 drivers/scsi/ibmvscsi/ibmvfc.c list_add_tail(&evt->queue, &vhost->sent); evt 1415 drivers/scsi/ibmvscsi/ibmvfc.c timer_setup(&evt->timer, ibmvfc_timeout, 0); evt 1418 drivers/scsi/ibmvscsi/ibmvfc.c evt->timer.expires = jiffies + (timeout * HZ); evt 1419 drivers/scsi/ibmvscsi/ibmvfc.c add_timer(&evt->timer); evt 1426 drivers/scsi/ibmvscsi/ibmvfc.c list_del(&evt->queue); evt 1427 drivers/scsi/ibmvscsi/ibmvfc.c del_timer(&evt->timer); evt 1437 drivers/scsi/ibmvscsi/ibmvfc.c if (evt->cmnd) evt 1438 drivers/scsi/ibmvscsi/ibmvfc.c scsi_dma_unmap(evt->cmnd); evt 1439 drivers/scsi/ibmvscsi/ibmvfc.c ibmvfc_free_event(evt); evt 1444 drivers/scsi/ibmvscsi/ibmvfc.c if (evt->cmnd) { evt 1445 drivers/scsi/ibmvscsi/ibmvfc.c evt->cmnd->result = DID_ERROR << 16; evt 1446 drivers/scsi/ibmvscsi/ibmvfc.c evt->done = ibmvfc_scsi_eh_done; evt 1448 drivers/scsi/ibmvscsi/ibmvfc.c evt->xfer_iu->mad_common.status = cpu_to_be16(IBMVFC_MAD_CRQ_ERROR); evt 1450 drivers/scsi/ibmvscsi/ibmvfc.c evt->done(evt); evt 1452 drivers/scsi/ibmvscsi/ibmvfc.c ibmvfc_trc_start(evt); evt 1462 drivers/scsi/ibmvscsi/ibmvfc.c static void ibmvfc_log_error(struct ibmvfc_event *evt) evt 1464 drivers/scsi/ibmvscsi/ibmvfc.c struct ibmvfc_cmd *vfc_cmd = &evt->xfer_iu->cmd; evt 1465 drivers/scsi/ibmvscsi/ibmvfc.c struct ibmvfc_host *vhost = evt->vhost; evt 1467 drivers/scsi/ibmvscsi/ibmvfc.c struct scsi_cmnd *cmnd = evt->cmnd; evt 1517 drivers/scsi/ibmvscsi/ibmvfc.c static void ibmvfc_scsi_done(struct ibmvfc_event *evt) evt 1519 drivers/scsi/ibmvscsi/ibmvfc.c struct ibmvfc_cmd *vfc_cmd = &evt->xfer_iu->cmd; evt 1521 drivers/scsi/ibmvscsi/ibmvfc.c struct scsi_cmnd *cmnd = evt->cmnd; evt 1549 drivers/scsi/ibmvscsi/ibmvfc.c ibmvfc_log_error(evt); evt 1560 drivers/scsi/ibmvscsi/ibmvfc.c if (evt->eh_comp) evt 1561 drivers/scsi/ibmvscsi/ibmvfc.c complete(evt->eh_comp); evt 1563 drivers/scsi/ibmvscsi/ibmvfc.c ibmvfc_free_event(evt); evt 1610 drivers/scsi/ibmvscsi/ibmvfc.c struct ibmvfc_event *evt; evt 1621 drivers/scsi/ibmvscsi/ibmvfc.c evt = ibmvfc_get_event(vhost); evt 1622 drivers/scsi/ibmvscsi/ibmvfc.c ibmvfc_init_event(evt, ibmvfc_scsi_done, IBMVFC_CMD_FORMAT); evt 1623 drivers/scsi/ibmvscsi/ibmvfc.c evt->cmnd = cmnd; evt 1625 drivers/scsi/ibmvscsi/ibmvfc.c vfc_cmd = &evt->iu.cmd; evt 1627 drivers/scsi/ibmvscsi/ibmvfc.c vfc_cmd->resp.va = cpu_to_be64(be64_to_cpu(evt->crq.ioba) + offsetof(struct ibmvfc_cmd, rsp)); evt 1643 drivers/scsi/ibmvscsi/ibmvfc.c if (likely(!(rc = ibmvfc_map_sg_data(cmnd, evt, vfc_cmd, vhost->dev)))) evt 1644 drivers/scsi/ibmvscsi/ibmvfc.c return ibmvfc_send_event(evt, vhost, 0); evt 1646 drivers/scsi/ibmvscsi/ibmvfc.c ibmvfc_free_event(evt); evt 1666 drivers/scsi/ibmvscsi/ibmvfc.c static void ibmvfc_sync_completion(struct ibmvfc_event *evt) evt 1669 drivers/scsi/ibmvscsi/ibmvfc.c if (evt->sync_iu) evt 1670 drivers/scsi/ibmvscsi/ibmvfc.c *evt->sync_iu = *evt->xfer_iu; evt 1672 drivers/scsi/ibmvscsi/ibmvfc.c complete(&evt->comp); evt 1680 drivers/scsi/ibmvscsi/ibmvfc.c static void ibmvfc_bsg_timeout_done(struct ibmvfc_event *evt) evt 1682 drivers/scsi/ibmvscsi/ibmvfc.c struct ibmvfc_host *vhost = evt->vhost; evt 1684 drivers/scsi/ibmvscsi/ibmvfc.c ibmvfc_free_event(evt); evt 1700 drivers/scsi/ibmvscsi/ibmvfc.c struct ibmvfc_event *evt; evt 1714 drivers/scsi/ibmvscsi/ibmvfc.c evt = ibmvfc_get_event(vhost); evt 1715 drivers/scsi/ibmvscsi/ibmvfc.c ibmvfc_init_event(evt, ibmvfc_bsg_timeout_done, IBMVFC_MAD_FORMAT); evt 1717 drivers/scsi/ibmvscsi/ibmvfc.c tmf = &evt->iu.tmf; evt 1725 drivers/scsi/ibmvscsi/ibmvfc.c rc = ibmvfc_send_event(evt, vhost, default_timeout); evt 1753 drivers/scsi/ibmvscsi/ibmvfc.c struct ibmvfc_event *evt; evt 1772 drivers/scsi/ibmvscsi/ibmvfc.c evt = ibmvfc_get_event(vhost); evt 1773 drivers/scsi/ibmvscsi/ibmvfc.c ibmvfc_init_event(evt, ibmvfc_sync_completion, IBMVFC_MAD_FORMAT); evt 1774 drivers/scsi/ibmvscsi/ibmvfc.c plogi = &evt->iu.plogi; evt 1780 drivers/scsi/ibmvscsi/ibmvfc.c evt->sync_iu = &rsp_iu; evt 1781 drivers/scsi/ibmvscsi/ibmvfc.c init_completion(&evt->comp); evt 1783 drivers/scsi/ibmvscsi/ibmvfc.c rc = ibmvfc_send_event(evt, vhost, default_timeout); evt 1789 drivers/scsi/ibmvscsi/ibmvfc.c wait_for_completion(&evt->comp); evt 1795 drivers/scsi/ibmvscsi/ibmvfc.c ibmvfc_free_event(evt); evt 1814 drivers/scsi/ibmvscsi/ibmvfc.c struct ibmvfc_event *evt; evt 1890 drivers/scsi/ibmvscsi/ibmvfc.c evt = ibmvfc_get_event(vhost); evt 1891 drivers/scsi/ibmvscsi/ibmvfc.c ibmvfc_init_event(evt, ibmvfc_sync_completion, IBMVFC_MAD_FORMAT); evt 1892 drivers/scsi/ibmvscsi/ibmvfc.c mad = &evt->iu.passthru; evt 1899 drivers/scsi/ibmvscsi/ibmvfc.c mad->cmd_ioba.va = cpu_to_be64(be64_to_cpu(evt->crq.ioba) + evt 1913 drivers/scsi/ibmvscsi/ibmvfc.c mad->iu.tag = cpu_to_be64((u64)evt); evt 1916 drivers/scsi/ibmvscsi/ibmvfc.c evt->sync_iu = &rsp_iu; evt 1917 drivers/scsi/ibmvscsi/ibmvfc.c init_completion(&evt->comp); evt 1918 drivers/scsi/ibmvscsi/ibmvfc.c rc = ibmvfc_send_event(evt, vhost, 0); evt 1926 drivers/scsi/ibmvscsi/ibmvfc.c wait_for_completion(&evt->comp); evt 1934 drivers/scsi/ibmvscsi/ibmvfc.c ibmvfc_free_event(evt); evt 1964 drivers/scsi/ibmvscsi/ibmvfc.c struct ibmvfc_event *evt = NULL; evt 1973 drivers/scsi/ibmvscsi/ibmvfc.c evt = ibmvfc_get_event(vhost); evt 1974 drivers/scsi/ibmvscsi/ibmvfc.c ibmvfc_init_event(evt, ibmvfc_sync_completion, IBMVFC_CMD_FORMAT); evt 1976 drivers/scsi/ibmvscsi/ibmvfc.c tmf = &evt->iu.cmd; evt 1978 drivers/scsi/ibmvscsi/ibmvfc.c tmf->resp.va = cpu_to_be64(be64_to_cpu(evt->crq.ioba) + offsetof(struct ibmvfc_cmd, rsp)); evt 1988 drivers/scsi/ibmvscsi/ibmvfc.c evt->sync_iu = &rsp_iu; evt 1990 drivers/scsi/ibmvscsi/ibmvfc.c init_completion(&evt->comp); evt 1991 drivers/scsi/ibmvscsi/ibmvfc.c rsp_rc = ibmvfc_send_event(evt, vhost, default_timeout); evt 2002 drivers/scsi/ibmvscsi/ibmvfc.c wait_for_completion(&evt->comp); evt 2021 drivers/scsi/ibmvscsi/ibmvfc.c ibmvfc_free_event(evt); evt 2034 drivers/scsi/ibmvscsi/ibmvfc.c static int ibmvfc_match_rport(struct ibmvfc_event *evt, void *rport) evt 2038 drivers/scsi/ibmvscsi/ibmvfc.c if (evt->cmnd) { evt 2039 drivers/scsi/ibmvscsi/ibmvfc.c cmd_rport = starget_to_rport(scsi_target(evt->cmnd->device)); evt 2054 drivers/scsi/ibmvscsi/ibmvfc.c static int ibmvfc_match_target(struct ibmvfc_event *evt, void *device) evt 2056 drivers/scsi/ibmvscsi/ibmvfc.c if (evt->cmnd && scsi_target(evt->cmnd->device) == device) evt 2069 drivers/scsi/ibmvscsi/ibmvfc.c static int ibmvfc_match_lun(struct ibmvfc_event *evt, void *device) evt 2071 drivers/scsi/ibmvscsi/ibmvfc.c if (evt->cmnd && evt->cmnd->device == device) evt 2088 drivers/scsi/ibmvscsi/ibmvfc.c struct ibmvfc_event *evt; evt 2098 drivers/scsi/ibmvscsi/ibmvfc.c list_for_each_entry(evt, &vhost->sent, queue) { evt 2099 drivers/scsi/ibmvscsi/ibmvfc.c if (match(evt, device)) { evt 2100 drivers/scsi/ibmvscsi/ibmvfc.c evt->eh_comp = ∁ evt 2112 drivers/scsi/ibmvscsi/ibmvfc.c list_for_each_entry(evt, &vhost->sent, queue) { evt 2113 drivers/scsi/ibmvscsi/ibmvfc.c if (match(evt, device)) { evt 2114 drivers/scsi/ibmvscsi/ibmvfc.c evt->eh_comp = NULL; evt 2148 drivers/scsi/ibmvscsi/ibmvfc.c struct ibmvfc_event *evt, *found_evt; evt 2157 drivers/scsi/ibmvscsi/ibmvfc.c list_for_each_entry(evt, &vhost->sent, queue) { evt 2158 drivers/scsi/ibmvscsi/ibmvfc.c if (evt->cmnd && evt->cmnd->device == sdev) { evt 2159 drivers/scsi/ibmvscsi/ibmvfc.c found_evt = evt; evt 2172 drivers/scsi/ibmvscsi/ibmvfc.c evt = ibmvfc_get_event(vhost); evt 2173 drivers/scsi/ibmvscsi/ibmvfc.c ibmvfc_init_event(evt, ibmvfc_sync_completion, IBMVFC_MAD_FORMAT); evt 2175 drivers/scsi/ibmvscsi/ibmvfc.c tmf = &evt->iu.tmf; evt 2191 drivers/scsi/ibmvscsi/ibmvfc.c evt->sync_iu = &rsp; evt 2192 drivers/scsi/ibmvscsi/ibmvfc.c init_completion(&evt->comp); evt 2193 drivers/scsi/ibmvscsi/ibmvfc.c rsp_rc = ibmvfc_send_event(evt, vhost, default_timeout); evt 2208 drivers/scsi/ibmvscsi/ibmvfc.c wait_for_completion(&evt->comp); evt 2211 drivers/scsi/ibmvscsi/ibmvfc.c ibmvfc_free_event(evt); evt 2239 drivers/scsi/ibmvscsi/ibmvfc.c static int ibmvfc_match_key(struct ibmvfc_event *evt, void *key) evt 2243 drivers/scsi/ibmvscsi/ibmvfc.c if (evt->crq.format == IBMVFC_CMD_FORMAT && evt 2244 drivers/scsi/ibmvscsi/ibmvfc.c be32_to_cpu(evt->iu.cmd.cancel_key) == cancel_key) evt 2257 drivers/scsi/ibmvscsi/ibmvfc.c static int ibmvfc_match_evt(struct ibmvfc_event *evt, void *match) evt 2259 drivers/scsi/ibmvscsi/ibmvfc.c if (evt == match) evt 2279 drivers/scsi/ibmvscsi/ibmvfc.c struct ibmvfc_event *evt, *found_evt; evt 2288 drivers/scsi/ibmvscsi/ibmvfc.c list_for_each_entry(evt, &vhost->sent, queue) { evt 2289 drivers/scsi/ibmvscsi/ibmvfc.c if (evt->cmnd && evt->cmnd->device == sdev) { evt 2290 drivers/scsi/ibmvscsi/ibmvfc.c found_evt = evt; evt 2303 drivers/scsi/ibmvscsi/ibmvfc.c evt = ibmvfc_get_event(vhost); evt 2304 drivers/scsi/ibmvscsi/ibmvfc.c ibmvfc_init_event(evt, ibmvfc_sync_completion, IBMVFC_CMD_FORMAT); evt 2306 drivers/scsi/ibmvscsi/ibmvfc.c tmf = &evt->iu.cmd; evt 2308 drivers/scsi/ibmvscsi/ibmvfc.c tmf->resp.va = cpu_to_be64(be64_to_cpu(evt->crq.ioba) + offsetof(struct ibmvfc_cmd, rsp)); evt 2318 drivers/scsi/ibmvscsi/ibmvfc.c evt->sync_iu = &rsp_iu; evt 2320 drivers/scsi/ibmvscsi/ibmvfc.c init_completion(&evt->comp); evt 2321 drivers/scsi/ibmvscsi/ibmvfc.c rsp_rc = ibmvfc_send_event(evt, vhost, default_timeout); evt 2332 drivers/scsi/ibmvscsi/ibmvfc.c timeout = wait_for_completion_timeout(&evt->comp, timeout); evt 2351 drivers/scsi/ibmvscsi/ibmvfc.c rc = ibmvfc_wait_for_ops(vhost, evt, ibmvfc_match_evt); evt 2381 drivers/scsi/ibmvscsi/ibmvfc.c ibmvfc_free_event(evt); evt 2719 drivers/scsi/ibmvscsi/ibmvfc.c struct ibmvfc_event *evt = (struct ibmvfc_event *)be64_to_cpu(crq->ioba); evt 2775 drivers/scsi/ibmvscsi/ibmvfc.c if (unlikely(!ibmvfc_valid_event(&vhost->pool, evt))) { evt 2781 drivers/scsi/ibmvscsi/ibmvfc.c if (unlikely(atomic_read(&evt->free))) { evt 2787 drivers/scsi/ibmvscsi/ibmvfc.c del_timer(&evt->timer); evt 2788 drivers/scsi/ibmvscsi/ibmvfc.c list_del(&evt->queue); evt 2789 drivers/scsi/ibmvscsi/ibmvfc.c ibmvfc_trc_end(evt); evt 2790 drivers/scsi/ibmvscsi/ibmvfc.c evt->done(evt); evt 3287 drivers/scsi/ibmvscsi/ibmvfc.c static void ibmvfc_tgt_prli_done(struct ibmvfc_event *evt) evt 3289 drivers/scsi/ibmvscsi/ibmvfc.c struct ibmvfc_target *tgt = evt->tgt; evt 3290 drivers/scsi/ibmvscsi/ibmvfc.c struct ibmvfc_host *vhost = evt->vhost; evt 3291 drivers/scsi/ibmvscsi/ibmvfc.c struct ibmvfc_process_login *rsp = &evt->xfer_iu->prli; evt 3347 drivers/scsi/ibmvscsi/ibmvfc.c ibmvfc_free_event(evt); evt 3360 drivers/scsi/ibmvscsi/ibmvfc.c struct ibmvfc_event *evt; evt 3366 drivers/scsi/ibmvscsi/ibmvfc.c evt = ibmvfc_get_event(vhost); evt 3368 drivers/scsi/ibmvscsi/ibmvfc.c ibmvfc_init_event(evt, ibmvfc_tgt_prli_done, IBMVFC_MAD_FORMAT); evt 3369 drivers/scsi/ibmvscsi/ibmvfc.c evt->tgt = tgt; evt 3370 drivers/scsi/ibmvscsi/ibmvfc.c prli = &evt->iu.prli; evt 3386 drivers/scsi/ibmvscsi/ibmvfc.c if (ibmvfc_send_event(evt, vhost, default_timeout)) { evt 3399 drivers/scsi/ibmvscsi/ibmvfc.c static void ibmvfc_tgt_plogi_done(struct ibmvfc_event *evt) evt 3401 drivers/scsi/ibmvscsi/ibmvfc.c struct ibmvfc_target *tgt = evt->tgt; evt 3402 drivers/scsi/ibmvscsi/ibmvfc.c struct ibmvfc_host *vhost = evt->vhost; evt 3403 drivers/scsi/ibmvscsi/ibmvfc.c struct ibmvfc_port_login *rsp = &evt->xfer_iu->plogi; evt 3448 drivers/scsi/ibmvscsi/ibmvfc.c ibmvfc_free_event(evt); evt 3461 drivers/scsi/ibmvscsi/ibmvfc.c struct ibmvfc_event *evt; evt 3468 drivers/scsi/ibmvscsi/ibmvfc.c evt = ibmvfc_get_event(vhost); evt 3471 drivers/scsi/ibmvscsi/ibmvfc.c ibmvfc_init_event(evt, ibmvfc_tgt_plogi_done, IBMVFC_MAD_FORMAT); evt 3472 drivers/scsi/ibmvscsi/ibmvfc.c evt->tgt = tgt; evt 3473 drivers/scsi/ibmvscsi/ibmvfc.c plogi = &evt->iu.plogi; evt 3480 drivers/scsi/ibmvscsi/ibmvfc.c if (ibmvfc_send_event(evt, vhost, default_timeout)) { evt 3493 drivers/scsi/ibmvscsi/ibmvfc.c static void ibmvfc_tgt_implicit_logout_done(struct ibmvfc_event *evt) evt 3495 drivers/scsi/ibmvscsi/ibmvfc.c struct ibmvfc_target *tgt = evt->tgt; evt 3496 drivers/scsi/ibmvscsi/ibmvfc.c struct ibmvfc_host *vhost = evt->vhost; evt 3497 drivers/scsi/ibmvscsi/ibmvfc.c struct ibmvfc_implicit_logout *rsp = &evt->xfer_iu->implicit_logout; evt 3501 drivers/scsi/ibmvscsi/ibmvfc.c ibmvfc_free_event(evt); evt 3536 drivers/scsi/ibmvscsi/ibmvfc.c struct ibmvfc_event *evt; evt 3542 drivers/scsi/ibmvscsi/ibmvfc.c evt = ibmvfc_get_event(vhost); evt 3544 drivers/scsi/ibmvscsi/ibmvfc.c ibmvfc_init_event(evt, ibmvfc_tgt_implicit_logout_done, IBMVFC_MAD_FORMAT); evt 3545 drivers/scsi/ibmvscsi/ibmvfc.c evt->tgt = tgt; evt 3546 drivers/scsi/ibmvscsi/ibmvfc.c mad = &evt->iu.implicit_logout; evt 3554 drivers/scsi/ibmvscsi/ibmvfc.c if (ibmvfc_send_event(evt, vhost, default_timeout)) { evt 3587 drivers/scsi/ibmvscsi/ibmvfc.c static void ibmvfc_tgt_adisc_done(struct ibmvfc_event *evt) evt 3589 drivers/scsi/ibmvscsi/ibmvfc.c struct ibmvfc_target *tgt = evt->tgt; evt 3590 drivers/scsi/ibmvscsi/ibmvfc.c struct ibmvfc_host *vhost = evt->vhost; evt 3591 drivers/scsi/ibmvscsi/ibmvfc.c struct ibmvfc_passthru_mad *mad = &evt->xfer_iu->passthru; evt 3621 drivers/scsi/ibmvscsi/ibmvfc.c ibmvfc_free_event(evt); evt 3630 drivers/scsi/ibmvscsi/ibmvfc.c static void ibmvfc_init_passthru(struct ibmvfc_event *evt) evt 3632 drivers/scsi/ibmvscsi/ibmvfc.c struct ibmvfc_passthru_mad *mad = &evt->iu.passthru; evt 3638 drivers/scsi/ibmvscsi/ibmvfc.c mad->cmd_ioba.va = cpu_to_be64((u64)be64_to_cpu(evt->crq.ioba) + evt 3643 drivers/scsi/ibmvscsi/ibmvfc.c mad->iu.cmd.va = cpu_to_be64((u64)be64_to_cpu(evt->crq.ioba) + evt 3647 drivers/scsi/ibmvscsi/ibmvfc.c mad->iu.rsp.va = cpu_to_be64((u64)be64_to_cpu(evt->crq.ioba) + evt 3663 drivers/scsi/ibmvscsi/ibmvfc.c static void ibmvfc_tgt_adisc_cancel_done(struct ibmvfc_event *evt) evt 3665 drivers/scsi/ibmvscsi/ibmvfc.c struct ibmvfc_host *vhost = evt->vhost; evt 3666 drivers/scsi/ibmvscsi/ibmvfc.c struct ibmvfc_target *tgt = evt->tgt; evt 3670 drivers/scsi/ibmvscsi/ibmvfc.c ibmvfc_free_event(evt); evt 3687 drivers/scsi/ibmvscsi/ibmvfc.c struct ibmvfc_event *evt; evt 3704 drivers/scsi/ibmvscsi/ibmvfc.c evt = ibmvfc_get_event(vhost); evt 3705 drivers/scsi/ibmvscsi/ibmvfc.c ibmvfc_init_event(evt, ibmvfc_tgt_adisc_cancel_done, IBMVFC_MAD_FORMAT); evt 3707 drivers/scsi/ibmvscsi/ibmvfc.c evt->tgt = tgt; evt 3708 drivers/scsi/ibmvscsi/ibmvfc.c tmf = &evt->iu.tmf; evt 3716 drivers/scsi/ibmvscsi/ibmvfc.c rc = ibmvfc_send_event(evt, vhost, default_timeout); evt 3743 drivers/scsi/ibmvscsi/ibmvfc.c struct ibmvfc_event *evt; evt 3749 drivers/scsi/ibmvscsi/ibmvfc.c evt = ibmvfc_get_event(vhost); evt 3751 drivers/scsi/ibmvscsi/ibmvfc.c ibmvfc_init_event(evt, ibmvfc_tgt_adisc_done, IBMVFC_MAD_FORMAT); evt 3752 drivers/scsi/ibmvscsi/ibmvfc.c evt->tgt = tgt; evt 3754 drivers/scsi/ibmvscsi/ibmvfc.c ibmvfc_init_passthru(evt); evt 3755 drivers/scsi/ibmvscsi/ibmvfc.c mad = &evt->iu.passthru; evt 3775 drivers/scsi/ibmvscsi/ibmvfc.c if (ibmvfc_send_event(evt, vhost, IBMVFC_ADISC_PLUS_CANCEL_TIMEOUT)) { evt 3789 drivers/scsi/ibmvscsi/ibmvfc.c static void ibmvfc_tgt_query_target_done(struct ibmvfc_event *evt) evt 3791 drivers/scsi/ibmvscsi/ibmvfc.c struct ibmvfc_target *tgt = evt->tgt; evt 3792 drivers/scsi/ibmvscsi/ibmvfc.c struct ibmvfc_host *vhost = evt->vhost; evt 3793 drivers/scsi/ibmvscsi/ibmvfc.c struct ibmvfc_query_tgt *rsp = &evt->xfer_iu->query_tgt; evt 3834 drivers/scsi/ibmvscsi/ibmvfc.c ibmvfc_free_event(evt); evt 3847 drivers/scsi/ibmvscsi/ibmvfc.c struct ibmvfc_event *evt; evt 3853 drivers/scsi/ibmvscsi/ibmvfc.c evt = ibmvfc_get_event(vhost); evt 3855 drivers/scsi/ibmvscsi/ibmvfc.c evt->tgt = tgt; evt 3856 drivers/scsi/ibmvscsi/ibmvfc.c ibmvfc_init_event(evt, ibmvfc_tgt_query_target_done, IBMVFC_MAD_FORMAT); evt 3857 drivers/scsi/ibmvscsi/ibmvfc.c query_tgt = &evt->iu.query_tgt; evt 3865 drivers/scsi/ibmvscsi/ibmvfc.c if (ibmvfc_send_event(evt, vhost, default_timeout)) { evt 3938 drivers/scsi/ibmvscsi/ibmvfc.c static void ibmvfc_discover_targets_done(struct ibmvfc_event *evt) evt 3940 drivers/scsi/ibmvscsi/ibmvfc.c struct ibmvfc_host *vhost = evt->vhost; evt 3941 drivers/scsi/ibmvscsi/ibmvfc.c struct ibmvfc_discover_targets *rsp = &evt->xfer_iu->discover_targets; evt 3965 drivers/scsi/ibmvscsi/ibmvfc.c ibmvfc_free_event(evt); evt 3977 drivers/scsi/ibmvscsi/ibmvfc.c struct ibmvfc_event *evt = ibmvfc_get_event(vhost); evt 3979 drivers/scsi/ibmvscsi/ibmvfc.c ibmvfc_init_event(evt, ibmvfc_discover_targets_done, IBMVFC_MAD_FORMAT); evt 3980 drivers/scsi/ibmvscsi/ibmvfc.c mad = &evt->iu.discover_targets; evt 3990 drivers/scsi/ibmvscsi/ibmvfc.c if (!ibmvfc_send_event(evt, vhost, default_timeout)) evt 4001 drivers/scsi/ibmvscsi/ibmvfc.c static void ibmvfc_npiv_login_done(struct ibmvfc_event *evt) evt 4003 drivers/scsi/ibmvscsi/ibmvfc.c struct ibmvfc_host *vhost = evt->vhost; evt 4004 drivers/scsi/ibmvscsi/ibmvfc.c u32 mad_status = be16_to_cpu(evt->xfer_iu->npiv_login.common.status); evt 4011 drivers/scsi/ibmvscsi/ibmvfc.c ibmvfc_free_event(evt); evt 4021 drivers/scsi/ibmvscsi/ibmvfc.c ibmvfc_free_event(evt); evt 4027 drivers/scsi/ibmvscsi/ibmvfc.c ibmvfc_free_event(evt); evt 4032 drivers/scsi/ibmvscsi/ibmvfc.c ibmvfc_free_event(evt); evt 4089 drivers/scsi/ibmvscsi/ibmvfc.c struct ibmvfc_event *evt = ibmvfc_get_event(vhost); evt 4093 drivers/scsi/ibmvscsi/ibmvfc.c ibmvfc_init_event(evt, ibmvfc_npiv_login_done, IBMVFC_MAD_FORMAT); evt 4096 drivers/scsi/ibmvscsi/ibmvfc.c mad = &evt->iu.npiv_login; evt 4106 drivers/scsi/ibmvscsi/ibmvfc.c if (!ibmvfc_send_event(evt, vhost, default_timeout)) evt 4117 drivers/scsi/ibmvscsi/ibmvfc.c static void ibmvfc_npiv_logout_done(struct ibmvfc_event *evt) evt 4119 drivers/scsi/ibmvscsi/ibmvfc.c struct ibmvfc_host *vhost = evt->vhost; evt 4120 drivers/scsi/ibmvscsi/ibmvfc.c u32 mad_status = be16_to_cpu(evt->xfer_iu->npiv_logout.common.status); evt 4122 drivers/scsi/ibmvscsi/ibmvfc.c ibmvfc_free_event(evt); evt 4152 drivers/scsi/ibmvscsi/ibmvfc.c struct ibmvfc_event *evt; evt 4154 drivers/scsi/ibmvscsi/ibmvfc.c evt = ibmvfc_get_event(vhost); evt 4155 drivers/scsi/ibmvscsi/ibmvfc.c ibmvfc_init_event(evt, ibmvfc_npiv_logout_done, IBMVFC_MAD_FORMAT); evt 4157 drivers/scsi/ibmvscsi/ibmvfc.c mad = &evt->iu.npiv_logout; evt 4165 drivers/scsi/ibmvscsi/ibmvfc.c if (!ibmvfc_send_event(evt, vhost, default_timeout)) evt 499 drivers/scsi/ibmvscsi/ibmvfc.h struct ibmvfc_event *evt; evt 462 drivers/scsi/ibmvscsi/ibmvscsi.c struct srp_event_struct *evt = &pool->events[i]; evt 463 drivers/scsi/ibmvscsi/ibmvscsi.c memset(&evt->crq, 0x00, sizeof(evt->crq)); evt 464 drivers/scsi/ibmvscsi/ibmvscsi.c atomic_set(&evt->free, 1); evt 465 drivers/scsi/ibmvscsi/ibmvscsi.c evt->crq.valid = VIOSRP_CRQ_CMD_RSP; evt 466 drivers/scsi/ibmvscsi/ibmvscsi.c evt->crq.IU_length = cpu_to_be16(sizeof(*evt->xfer_iu)); evt 467 drivers/scsi/ibmvscsi/ibmvscsi.c evt->crq.IU_data_ptr = cpu_to_be64(pool->iu_token + evt 468 drivers/scsi/ibmvscsi/ibmvscsi.c sizeof(*evt->xfer_iu) * i); evt 469 drivers/scsi/ibmvscsi/ibmvscsi.c evt->xfer_iu = pool->iu_storage + i; evt 470 drivers/scsi/ibmvscsi/ibmvscsi.c evt->hostdata = hostdata; evt 471 drivers/scsi/ibmvscsi/ibmvscsi.c evt->ext_list = NULL; evt 472 drivers/scsi/ibmvscsi/ibmvscsi.c evt->ext_list_token = 0; evt 516 drivers/scsi/ibmvscsi/ibmvscsi.c struct srp_event_struct *evt) evt 518 drivers/scsi/ibmvscsi/ibmvscsi.c int index = evt - pool->events; evt 521 drivers/scsi/ibmvscsi/ibmvscsi.c if (evt != pool->events + index) /* unaligned */ evt 533 drivers/scsi/ibmvscsi/ibmvscsi.c struct srp_event_struct *evt) evt 535 drivers/scsi/ibmvscsi/ibmvscsi.c if (!valid_event_struct(pool, evt)) { evt 536 drivers/scsi/ibmvscsi/ibmvscsi.c dev_err(evt->hostdata->dev, "Freeing invalid event_struct %p " evt 537 drivers/scsi/ibmvscsi/ibmvscsi.c "(not in pool %p)\n", evt, pool->events); evt 540 drivers/scsi/ibmvscsi/ibmvscsi.c if (atomic_inc_return(&evt->free) != 1) { evt 541 drivers/scsi/ibmvscsi/ibmvscsi.c dev_err(evt->hostdata->dev, "Freeing event_struct %p " evt 542 drivers/scsi/ibmvscsi/ibmvscsi.c "which is not in use!\n", evt); evt 782 drivers/scsi/ibmvscsi/ibmvscsi.c struct srp_event_struct *evt; evt 787 drivers/scsi/ibmvscsi/ibmvscsi.c evt = list_first_entry(&hostdata->sent, struct srp_event_struct, list); evt 788 drivers/scsi/ibmvscsi/ibmvscsi.c list_del(&evt->list); evt 789 drivers/scsi/ibmvscsi/ibmvscsi.c del_timer(&evt->timer); evt 792 drivers/scsi/ibmvscsi/ibmvscsi.c if (evt->cmnd) { evt 793 drivers/scsi/ibmvscsi/ibmvscsi.c evt->cmnd->result = (error_code << 16); evt 794 drivers/scsi/ibmvscsi/ibmvscsi.c unmap_cmd_data(&evt->iu.srp.cmd, evt, evt 795 drivers/scsi/ibmvscsi/ibmvscsi.c evt->hostdata->dev); evt 796 drivers/scsi/ibmvscsi/ibmvscsi.c if (evt->cmnd_done) evt 797 drivers/scsi/ibmvscsi/ibmvscsi.c evt->cmnd_done(evt->cmnd); evt 798 drivers/scsi/ibmvscsi/ibmvscsi.c } else if (evt->done && evt->crq.format != VIOSRP_MAD_FORMAT && evt 799 drivers/scsi/ibmvscsi/ibmvscsi.c evt->iu.srp.login_req.opcode != SRP_LOGIN_REQ) evt 800 drivers/scsi/ibmvscsi/ibmvscsi.c evt->done(evt); evt 801 drivers/scsi/ibmvscsi/ibmvscsi.c free_event_struct(&evt->hostdata->pool, evt); evt 1473 drivers/scsi/ibmvscsi/ibmvscsi.c struct srp_event_struct *evt; evt 1500 drivers/scsi/ibmvscsi/ibmvscsi.c evt = get_event_struct(&hostdata->pool); evt 1501 drivers/scsi/ibmvscsi/ibmvscsi.c if (evt == NULL) { evt 1508 drivers/scsi/ibmvscsi/ibmvscsi.c init_event_struct(evt, evt 1513 drivers/scsi/ibmvscsi/ibmvscsi.c tsk_mgmt = &evt->iu.srp.tsk_mgmt; evt 1522 drivers/scsi/ibmvscsi/ibmvscsi.c evt->sync_srp = &srp_rsp; evt 1524 drivers/scsi/ibmvscsi/ibmvscsi.c init_completion(&evt->comp); evt 1525 drivers/scsi/ibmvscsi/ibmvscsi.c rsp_rc = ibmvscsi_send_srp_event(evt, hostdata, abort_timeout * 2); evt 1547 drivers/scsi/ibmvscsi/ibmvscsi.c wait_for_completion(&evt->comp); evt 1612 drivers/scsi/ibmvscsi/ibmvscsi.c struct srp_event_struct *evt; evt 1623 drivers/scsi/ibmvscsi/ibmvscsi.c evt = get_event_struct(&hostdata->pool); evt 1624 drivers/scsi/ibmvscsi/ibmvscsi.c if (evt == NULL) { evt 1631 drivers/scsi/ibmvscsi/ibmvscsi.c init_event_struct(evt, evt 1636 drivers/scsi/ibmvscsi/ibmvscsi.c tsk_mgmt = &evt->iu.srp.tsk_mgmt; evt 1644 drivers/scsi/ibmvscsi/ibmvscsi.c evt->sync_srp = &srp_rsp; evt 1646 drivers/scsi/ibmvscsi/ibmvscsi.c init_completion(&evt->comp); evt 1647 drivers/scsi/ibmvscsi/ibmvscsi.c rsp_rc = ibmvscsi_send_srp_event(evt, hostdata, reset_timeout * 2); evt 1668 drivers/scsi/ibmvscsi/ibmvscsi.c wait_for_completion(&evt->comp); evt 106 drivers/scsi/lpfc/lpfc_bsg.c struct lpfc_bsg_event *evt; evt 803 drivers/scsi/lpfc/lpfc_bsg.c struct lpfc_bsg_event *evt = container_of(kref, struct lpfc_bsg_event, evt 807 drivers/scsi/lpfc/lpfc_bsg.c list_del(&evt->node); evt 809 drivers/scsi/lpfc/lpfc_bsg.c while (!list_empty(&evt->events_to_get)) { evt 810 drivers/scsi/lpfc/lpfc_bsg.c ed = list_entry(evt->events_to_get.next, typeof(*ed), node); evt 816 drivers/scsi/lpfc/lpfc_bsg.c while (!list_empty(&evt->events_to_see)) { evt 817 drivers/scsi/lpfc/lpfc_bsg.c ed = list_entry(evt->events_to_see.next, typeof(*ed), node); evt 823 drivers/scsi/lpfc/lpfc_bsg.c kfree(evt->dd_data); evt 824 drivers/scsi/lpfc/lpfc_bsg.c kfree(evt); evt 832 drivers/scsi/lpfc/lpfc_bsg.c lpfc_bsg_event_ref(struct lpfc_bsg_event *evt) evt 834 drivers/scsi/lpfc/lpfc_bsg.c kref_get(&evt->kref); evt 842 drivers/scsi/lpfc/lpfc_bsg.c lpfc_bsg_event_unref(struct lpfc_bsg_event *evt) evt 844 drivers/scsi/lpfc/lpfc_bsg.c kref_put(&evt->kref, lpfc_bsg_event_free); evt 856 drivers/scsi/lpfc/lpfc_bsg.c struct lpfc_bsg_event *evt = kzalloc(sizeof(*evt), GFP_KERNEL); evt 858 drivers/scsi/lpfc/lpfc_bsg.c if (!evt) evt 861 drivers/scsi/lpfc/lpfc_bsg.c INIT_LIST_HEAD(&evt->events_to_get); evt 862 drivers/scsi/lpfc/lpfc_bsg.c INIT_LIST_HEAD(&evt->events_to_see); evt 863 drivers/scsi/lpfc/lpfc_bsg.c evt->type_mask = ev_mask; evt 864 drivers/scsi/lpfc/lpfc_bsg.c evt->req_id = ev_req_id; evt 865 drivers/scsi/lpfc/lpfc_bsg.c evt->reg_id = ev_reg_id; evt 866 drivers/scsi/lpfc/lpfc_bsg.c evt->wait_time_stamp = jiffies; evt 867 drivers/scsi/lpfc/lpfc_bsg.c evt->dd_data = NULL; evt 868 drivers/scsi/lpfc/lpfc_bsg.c init_waitqueue_head(&evt->wq); evt 869 drivers/scsi/lpfc/lpfc_bsg.c kref_init(&evt->kref); evt 870 drivers/scsi/lpfc/lpfc_bsg.c return evt; evt 921 drivers/scsi/lpfc/lpfc_bsg.c struct lpfc_bsg_event *evt; evt 966 drivers/scsi/lpfc/lpfc_bsg.c list_for_each_entry(evt, &phba->ct_ev_waiters, node) { evt 967 drivers/scsi/lpfc/lpfc_bsg.c if (!(evt->type_mask & FC_REG_CT_EVENT) || evt 968 drivers/scsi/lpfc/lpfc_bsg.c evt->req_id != evt_req_id) evt 971 drivers/scsi/lpfc/lpfc_bsg.c lpfc_bsg_event_ref(evt); evt 976 drivers/scsi/lpfc/lpfc_bsg.c lpfc_bsg_event_unref(evt); evt 1003 drivers/scsi/lpfc/lpfc_bsg.c lpfc_bsg_event_unref(evt); evt 1049 drivers/scsi/lpfc/lpfc_bsg.c lpfc_bsg_event_unref(evt); evt 1123 drivers/scsi/lpfc/lpfc_bsg.c list_add(&evt_dat->node, &evt->events_to_see); evt 1125 drivers/scsi/lpfc/lpfc_bsg.c wake_up_interruptible(&evt->wq); evt 1126 drivers/scsi/lpfc/lpfc_bsg.c lpfc_bsg_event_unref(evt); evt 1130 drivers/scsi/lpfc/lpfc_bsg.c list_move(evt->events_to_see.prev, &evt->events_to_get); evt 1132 drivers/scsi/lpfc/lpfc_bsg.c dd_data = (struct bsg_job_data *)evt->dd_data; evt 1135 drivers/scsi/lpfc/lpfc_bsg.c lpfc_bsg_event_unref(evt); evt 1212 drivers/scsi/lpfc/lpfc_bsg.c struct lpfc_bsg_event *evt; evt 1232 drivers/scsi/lpfc/lpfc_bsg.c list_for_each_entry(evt, &phba->ct_ev_waiters, node) { evt 1233 drivers/scsi/lpfc/lpfc_bsg.c if (evt->reg_id == event_req->ev_reg_id) { evt 1234 drivers/scsi/lpfc/lpfc_bsg.c lpfc_bsg_event_ref(evt); evt 1235 drivers/scsi/lpfc/lpfc_bsg.c evt->wait_time_stamp = jiffies; evt 1236 drivers/scsi/lpfc/lpfc_bsg.c dd_data = (struct bsg_job_data *)evt->dd_data; evt 1242 drivers/scsi/lpfc/lpfc_bsg.c if (&evt->node == &phba->ct_ev_waiters) { evt 1251 drivers/scsi/lpfc/lpfc_bsg.c evt = lpfc_bsg_event_new(ev_mask, event_req->ev_reg_id, evt 1253 drivers/scsi/lpfc/lpfc_bsg.c if (!evt) { evt 1262 drivers/scsi/lpfc/lpfc_bsg.c dd_data->context_un.evt = evt; evt 1263 drivers/scsi/lpfc/lpfc_bsg.c evt->dd_data = (void *)dd_data; evt 1265 drivers/scsi/lpfc/lpfc_bsg.c list_add(&evt->node, &phba->ct_ev_waiters); evt 1266 drivers/scsi/lpfc/lpfc_bsg.c lpfc_bsg_event_ref(evt); evt 1267 drivers/scsi/lpfc/lpfc_bsg.c evt->wait_time_stamp = jiffies; evt 1272 drivers/scsi/lpfc/lpfc_bsg.c evt->waiting = 1; evt 1297 drivers/scsi/lpfc/lpfc_bsg.c struct lpfc_bsg_event *evt, *evt_next; evt 1317 drivers/scsi/lpfc/lpfc_bsg.c list_for_each_entry_safe(evt, evt_next, &phba->ct_ev_waiters, node) { evt 1318 drivers/scsi/lpfc/lpfc_bsg.c if (evt->reg_id == event_req->ev_reg_id) { evt 1319 drivers/scsi/lpfc/lpfc_bsg.c if (list_empty(&evt->events_to_get)) evt 1321 drivers/scsi/lpfc/lpfc_bsg.c lpfc_bsg_event_ref(evt); evt 1322 drivers/scsi/lpfc/lpfc_bsg.c evt->wait_time_stamp = jiffies; evt 1323 drivers/scsi/lpfc/lpfc_bsg.c evt_dat = list_entry(evt->events_to_get.prev, evt 1364 drivers/scsi/lpfc/lpfc_bsg.c lpfc_bsg_event_unref(evt); evt 2651 drivers/scsi/lpfc/lpfc_bsg.c struct lpfc_bsg_event *evt; evt 2664 drivers/scsi/lpfc/lpfc_bsg.c evt = lpfc_bsg_event_new(FC_REG_CT_EVENT, current->pid, evt 2666 drivers/scsi/lpfc/lpfc_bsg.c if (!evt) evt 2670 drivers/scsi/lpfc/lpfc_bsg.c list_add(&evt->node, &phba->ct_ev_waiters); evt 2671 drivers/scsi/lpfc/lpfc_bsg.c lpfc_bsg_event_ref(evt); evt 2747 drivers/scsi/lpfc/lpfc_bsg.c evt->waiting = 1; evt 2748 drivers/scsi/lpfc/lpfc_bsg.c evt->wait_time_stamp = jiffies; evt 2750 drivers/scsi/lpfc/lpfc_bsg.c evt->wq, !list_empty(&evt->events_to_see), evt 2753 drivers/scsi/lpfc/lpfc_bsg.c if (list_empty(&evt->events_to_see)) evt 2757 drivers/scsi/lpfc/lpfc_bsg.c list_move(evt->events_to_see.prev, &evt->events_to_get); evt 2759 drivers/scsi/lpfc/lpfc_bsg.c *rxxri = (list_entry(evt->events_to_get.prev, evt 2763 drivers/scsi/lpfc/lpfc_bsg.c evt->waiting = 0; evt 2767 drivers/scsi/lpfc/lpfc_bsg.c lpfc_bsg_event_unref(evt); /* release ref */ evt 2768 drivers/scsi/lpfc/lpfc_bsg.c lpfc_bsg_event_unref(evt); /* delete */ evt 3104 drivers/scsi/lpfc/lpfc_bsg.c struct lpfc_bsg_event *evt; evt 3211 drivers/scsi/lpfc/lpfc_bsg.c evt = lpfc_bsg_event_new(FC_REG_CT_EVENT, current->pid, evt 3213 drivers/scsi/lpfc/lpfc_bsg.c if (!evt) { evt 3220 drivers/scsi/lpfc/lpfc_bsg.c list_add(&evt->node, &phba->ct_ev_waiters); evt 3221 drivers/scsi/lpfc/lpfc_bsg.c lpfc_bsg_event_ref(evt); evt 3324 drivers/scsi/lpfc/lpfc_bsg.c evt->waiting = 1; evt 3326 drivers/scsi/lpfc/lpfc_bsg.c evt->wq, !list_empty(&evt->events_to_see), evt 3329 drivers/scsi/lpfc/lpfc_bsg.c evt->waiting = 0; evt 3330 drivers/scsi/lpfc/lpfc_bsg.c if (list_empty(&evt->events_to_see)) { evt 3337 drivers/scsi/lpfc/lpfc_bsg.c list_move(evt->events_to_see.prev, &evt->events_to_get); evt 3338 drivers/scsi/lpfc/lpfc_bsg.c evdat = list_entry(evt->events_to_get.prev, evt 3367 drivers/scsi/lpfc/lpfc_bsg.c lpfc_bsg_event_unref(evt); /* release ref */ evt 3368 drivers/scsi/lpfc/lpfc_bsg.c lpfc_bsg_event_unref(evt); /* delete */ evt 51 drivers/scsi/lpfc/lpfc_disc.h enum lpfc_work_type evt; evt 3435 drivers/scsi/lpfc/lpfc_els.c evtp->evt = LPFC_EVT_ELS_RETRY; evt 186 drivers/scsi/lpfc/lpfc_hbadisc.c evtp->evt = LPFC_EVT_DEV_LOSS; evt 420 drivers/scsi/lpfc/lpfc_hbadisc.c ret->work_evt.evt = LPFC_EVT_FASTPATH_MGMT_EVT; evt 435 drivers/scsi/lpfc/lpfc_hbadisc.c struct lpfc_fast_path_event *evt) { evt 438 drivers/scsi/lpfc/lpfc_hbadisc.c kfree(evt); evt 532 drivers/scsi/lpfc/lpfc_hbadisc.c switch (evtp->evt) { evt 781 drivers/scsi/lpfc/lpfc_hbadisc.c uint32_t evt) evt 796 drivers/scsi/lpfc/lpfc_hbadisc.c evtp->evt = evt; evt 931 drivers/scsi/lpfc/lpfc_nportdisc.c void *arg, uint32_t evt) evt 940 drivers/scsi/lpfc/lpfc_nportdisc.c (evt == NLP_EVT_CMPL_REG_LOGIN) && evt 948 drivers/scsi/lpfc/lpfc_nportdisc.c ndlp->nlp_DID, evt, ndlp->nlp_state, ndlp->nlp_rpi, evt 955 drivers/scsi/lpfc/lpfc_nportdisc.c void *arg, uint32_t evt) evt 966 drivers/scsi/lpfc/lpfc_nportdisc.c ndlp->nlp_DID, evt, ndlp->nlp_state, ndlp->nlp_rpi, evt 976 drivers/scsi/lpfc/lpfc_nportdisc.c void *arg, uint32_t evt) evt 990 drivers/scsi/lpfc/lpfc_nportdisc.c void *arg, uint32_t evt) evt 998 drivers/scsi/lpfc/lpfc_nportdisc.c void *arg, uint32_t evt) evt 1013 drivers/scsi/lpfc/lpfc_nportdisc.c void *arg, uint32_t evt) evt 1020 drivers/scsi/lpfc/lpfc_nportdisc.c void *arg, uint32_t evt) evt 1028 drivers/scsi/lpfc/lpfc_nportdisc.c void *arg, uint32_t evt) evt 1035 drivers/scsi/lpfc/lpfc_nportdisc.c void *arg, uint32_t evt) evt 1086 drivers/scsi/lpfc/lpfc_nportdisc.c void *arg, uint32_t evt) evt 1100 drivers/scsi/lpfc/lpfc_nportdisc.c void *arg, uint32_t evt) evt 1116 drivers/scsi/lpfc/lpfc_nportdisc.c void *arg, uint32_t evt) evt 1125 drivers/scsi/lpfc/lpfc_nportdisc.c if (evt == NLP_EVT_RCV_LOGO) { evt 1147 drivers/scsi/lpfc/lpfc_nportdisc.c uint32_t evt) evt 1352 drivers/scsi/lpfc/lpfc_nportdisc.c void *arg, uint32_t evt) evt 1359 drivers/scsi/lpfc/lpfc_nportdisc.c struct lpfc_nodelist *ndlp, void *arg, uint32_t evt) evt 1378 drivers/scsi/lpfc/lpfc_nportdisc.c void *arg, uint32_t evt) evt 1400 drivers/scsi/lpfc/lpfc_nportdisc.c uint32_t evt) evt 1425 drivers/scsi/lpfc/lpfc_nportdisc.c void *arg, uint32_t evt) evt 1455 drivers/scsi/lpfc/lpfc_nportdisc.c void *arg, uint32_t evt) evt 1466 drivers/scsi/lpfc/lpfc_nportdisc.c void *arg, uint32_t evt) evt 1483 drivers/scsi/lpfc/lpfc_nportdisc.c void *arg, uint32_t evt) evt 1495 drivers/scsi/lpfc/lpfc_nportdisc.c void *arg, uint32_t evt) evt 1509 drivers/scsi/lpfc/lpfc_nportdisc.c void *arg, uint32_t evt) evt 1565 drivers/scsi/lpfc/lpfc_nportdisc.c void *arg, uint32_t evt) evt 1587 drivers/scsi/lpfc/lpfc_nportdisc.c uint32_t evt) evt 1614 drivers/scsi/lpfc/lpfc_nportdisc.c uint32_t evt) evt 1626 drivers/scsi/lpfc/lpfc_nportdisc.c uint32_t evt) evt 1665 drivers/scsi/lpfc/lpfc_nportdisc.c uint32_t evt) evt 1720 drivers/scsi/lpfc/lpfc_nportdisc.c uint32_t evt) evt 1732 drivers/scsi/lpfc/lpfc_nportdisc.c uint32_t evt) evt 1745 drivers/scsi/lpfc/lpfc_nportdisc.c uint32_t evt) evt 1857 drivers/scsi/lpfc/lpfc_nportdisc.c uint32_t evt) evt 1876 drivers/scsi/lpfc/lpfc_nportdisc.c uint32_t evt) evt 1905 drivers/scsi/lpfc/lpfc_nportdisc.c void *arg, uint32_t evt) evt 1917 drivers/scsi/lpfc/lpfc_nportdisc.c void *arg, uint32_t evt) evt 1929 drivers/scsi/lpfc/lpfc_nportdisc.c void *arg, uint32_t evt) evt 1942 drivers/scsi/lpfc/lpfc_nportdisc.c void *arg, uint32_t evt) evt 1957 drivers/scsi/lpfc/lpfc_nportdisc.c void *arg, uint32_t evt) evt 1967 drivers/scsi/lpfc/lpfc_nportdisc.c void *arg, uint32_t evt) evt 2141 drivers/scsi/lpfc/lpfc_nportdisc.c void *arg, uint32_t evt) evt 2180 drivers/scsi/lpfc/lpfc_nportdisc.c uint32_t evt) evt 2205 drivers/scsi/lpfc/lpfc_nportdisc.c void *arg, uint32_t evt) evt 2219 drivers/scsi/lpfc/lpfc_nportdisc.c void *arg, uint32_t evt) evt 2233 drivers/scsi/lpfc/lpfc_nportdisc.c void *arg, uint32_t evt) evt 2247 drivers/scsi/lpfc/lpfc_nportdisc.c void *arg, uint32_t evt) evt 2261 drivers/scsi/lpfc/lpfc_nportdisc.c void *arg, uint32_t evt) evt 2275 drivers/scsi/lpfc/lpfc_nportdisc.c void *arg, uint32_t evt) evt 2290 drivers/scsi/lpfc/lpfc_nportdisc.c void *arg, uint32_t evt) evt 2307 drivers/scsi/lpfc/lpfc_nportdisc.c void *arg, uint32_t evt) evt 2319 drivers/scsi/lpfc/lpfc_nportdisc.c void *arg, uint32_t evt) evt 2329 drivers/scsi/lpfc/lpfc_nportdisc.c void *arg, uint32_t evt) evt 2343 drivers/scsi/lpfc/lpfc_nportdisc.c void *arg, uint32_t evt) evt 2353 drivers/scsi/lpfc/lpfc_nportdisc.c void *arg, uint32_t evt) evt 2363 drivers/scsi/lpfc/lpfc_nportdisc.c void *arg, uint32_t evt) evt 2375 drivers/scsi/lpfc/lpfc_nportdisc.c uint32_t evt) evt 2392 drivers/scsi/lpfc/lpfc_nportdisc.c void *arg, uint32_t evt) evt 2402 drivers/scsi/lpfc/lpfc_nportdisc.c void *arg, uint32_t evt) evt 2414 drivers/scsi/lpfc/lpfc_nportdisc.c void *arg, uint32_t evt) evt 2425 drivers/scsi/lpfc/lpfc_nportdisc.c void *arg, uint32_t evt) evt 2435 drivers/scsi/lpfc/lpfc_nportdisc.c void *arg, uint32_t evt) evt 2453 drivers/scsi/lpfc/lpfc_nportdisc.c uint32_t evt) evt 2469 drivers/scsi/lpfc/lpfc_nportdisc.c void *arg, uint32_t evt) evt 2495 drivers/scsi/lpfc/lpfc_nportdisc.c void *arg, uint32_t evt) evt 2525 drivers/scsi/lpfc/lpfc_nportdisc.c void *arg, uint32_t evt) evt 2535 drivers/scsi/lpfc/lpfc_nportdisc.c void *arg, uint32_t evt) evt 2563 drivers/scsi/lpfc/lpfc_nportdisc.c void *arg, uint32_t evt) evt 2592 drivers/scsi/lpfc/lpfc_nportdisc.c void *arg, uint32_t evt) evt 2613 drivers/scsi/lpfc/lpfc_nportdisc.c void *arg, uint32_t evt) evt 2631 drivers/scsi/lpfc/lpfc_nportdisc.c void *arg, uint32_t evt) evt 2647 drivers/scsi/lpfc/lpfc_nportdisc.c void *arg, uint32_t evt) evt 2666 drivers/scsi/lpfc/lpfc_nportdisc.c void *arg, uint32_t evt) evt 2690 drivers/scsi/lpfc/lpfc_nportdisc.c void *arg, uint32_t evt) evt 2706 drivers/scsi/lpfc/lpfc_nportdisc.c void *arg, uint32_t evt) evt 2913 drivers/scsi/lpfc/lpfc_nportdisc.c void *arg, uint32_t evt) evt 2932 drivers/scsi/lpfc/lpfc_nportdisc.c evt, ndlp->nlp_DID, cur_state, ndlp->nlp_rpi, evt 2937 drivers/scsi/lpfc/lpfc_nportdisc.c evt, cur_state, ndlp->nlp_DID); evt 2939 drivers/scsi/lpfc/lpfc_nportdisc.c func = lpfc_disc_action[(cur_state * NLP_EVT_MAX_EVENT) + evt]; evt 2940 drivers/scsi/lpfc/lpfc_nportdisc.c rc = (func) (vport, ndlp, arg, evt); evt 3931 drivers/scsi/lpfc/lpfc_scsi.c fast_path_evt->work_evt.evt = evt 1495 drivers/scsi/pmcraid.c static void pmcraid_notify_ioastate(struct pmcraid_instance *pinstance, u32 evt) evt 1497 drivers/scsi/pmcraid.c pinstance->scn.ioa_state = evt; evt 1715 drivers/scsi/qla2xxx/qla_mr.c qlafx00_process_aen(struct scsi_qla_host *vha, struct qla_work_evt *evt) evt 1721 drivers/scsi/qla2xxx/qla_mr.c aen_data = evt->u.aenfx.evtcode; evt 1723 drivers/scsi/qla2xxx/qla_mr.c switch (evt->u.aenfx.evtcode) { evt 1725 drivers/scsi/qla2xxx/qla_mr.c if (evt->u.aenfx.mbx[1] == 0) { evt 1726 drivers/scsi/qla2xxx/qla_mr.c if (evt->u.aenfx.mbx[2] == 1) { evt 1733 drivers/scsi/qla2xxx/qla_mr.c } else if (evt->u.aenfx.mbx[2] == 2) { evt 1734 drivers/scsi/qla2xxx/qla_mr.c qlafx00_tgt_detach(vha, evt->u.aenfx.mbx[3]); evt 1736 drivers/scsi/qla2xxx/qla_mr.c } else if (evt->u.aenfx.mbx[1] == 0xffff) { evt 1737 drivers/scsi/qla2xxx/qla_mr.c if (evt->u.aenfx.mbx[2] == 1) { evt 1741 drivers/scsi/qla2xxx/qla_mr.c } else if (evt->u.aenfx.mbx[2] == 2) { evt 1759 drivers/scsi/qla2xxx/qla_mr.c evt->u.aenfx.evtcode); evt 2363 drivers/scsi/scsi_lib.c static void scsi_evt_emit(struct scsi_device *sdev, struct scsi_event *evt) evt 2368 drivers/scsi/scsi_lib.c switch (evt->evt_type) { evt 2424 drivers/scsi/scsi_lib.c struct scsi_event *evt; evt 2436 drivers/scsi/scsi_lib.c evt = list_entry(this, struct scsi_event, node); evt 2437 drivers/scsi/scsi_lib.c list_del(&evt->node); evt 2438 drivers/scsi/scsi_lib.c scsi_evt_emit(sdev, evt); evt 2439 drivers/scsi/scsi_lib.c kfree(evt); evt 2451 drivers/scsi/scsi_lib.c void sdev_evt_send(struct scsi_device *sdev, struct scsi_event *evt) evt 2459 drivers/scsi/scsi_lib.c if (!test_bit(evt->evt_type, sdev->supported_events)) { evt 2460 drivers/scsi/scsi_lib.c kfree(evt); evt 2466 drivers/scsi/scsi_lib.c list_add_tail(&evt->node, &sdev->event_list); evt 2482 drivers/scsi/scsi_lib.c struct scsi_event *evt = kzalloc(sizeof(struct scsi_event), gfpflags); evt 2483 drivers/scsi/scsi_lib.c if (!evt) evt 2486 drivers/scsi/scsi_lib.c evt->evt_type = evt_type; evt 2487 drivers/scsi/scsi_lib.c INIT_LIST_HEAD(&evt->node); evt 2504 drivers/scsi/scsi_lib.c return evt; evt 2519 drivers/scsi/scsi_lib.c struct scsi_event *evt = sdev_evt_alloc(evt_type, gfpflags); evt 2520 drivers/scsi/scsi_lib.c if (!evt) { evt 2526 drivers/scsi/scsi_lib.c sdev_evt_send(sdev, evt); evt 457 drivers/scsi/scsi_sysfs.c struct scsi_event *evt; evt 459 drivers/scsi/scsi_sysfs.c evt = list_entry(this, struct scsi_event, node); evt 460 drivers/scsi/scsi_sysfs.c list_del(&evt->node); evt 461 drivers/scsi/scsi_sysfs.c kfree(evt); evt 53 drivers/scsi/virtio_scsi.c struct virtio_scsi_event evt; evt 153 drivers/staging/uwb/address.c struct uwb_rc_evt_dev_addr_mgmt evt; evt 170 drivers/staging/uwb/address.c result = uwb_rc_dev_addr_mgmt(rc, bmOperationType, baAddr, &evt); evt 174 drivers/staging/uwb/address.c memcpy(&dev_addr->data, evt.baAddr, evt 178 drivers/staging/uwb/address.c memcpy(&mac_addr->data, evt.baAddr, evt 283 drivers/staging/uwb/address.c int uwbd_evt_handle_rc_dev_addr_conflict(struct uwb_event *evt) evt 285 drivers/staging/uwb/address.c struct uwb_rc *rc = evt->rc; evt 347 drivers/staging/uwb/beacon.c static int uwb_verify_beacon(struct uwb_rc *rc, struct uwb_event *evt, evt 355 drivers/staging/uwb/beacon.c if (evt->notif.size < sizeof(*be) + sizeof(*bf)) { evt 357 drivers/staging/uwb/beacon.c "(%zu vs %zu bytes needed)\n", evt->notif.size, evt 379 drivers/staging/uwb/beacon.c int uwbd_evt_handle_rc_beacon(struct uwb_event *evt) evt 387 drivers/staging/uwb/beacon.c rc = evt->rc; evt 388 drivers/staging/uwb/beacon.c be = container_of(evt->notif.rceb, struct uwb_rc_evt_beacon, rceb); evt 389 drivers/staging/uwb/beacon.c result = uwb_verify_beacon(rc, evt, be); evt 416 drivers/staging/uwb/beacon.c uwb_beacon_print(evt->rc, be, bf); evt 417 drivers/staging/uwb/beacon.c bce = __uwb_beca_add(rc, be, bf, evt->ts_jiffies); evt 430 drivers/staging/uwb/beacon.c bce->ts_jiffies = evt->ts_jiffies; evt 443 drivers/staging/uwb/beacon.c uwbd_dev_onair(evt->rc, bce); evt 455 drivers/staging/uwb/beacon.c int uwbd_evt_handle_rc_beacon_size(struct uwb_event *evt) evt 458 drivers/staging/uwb/beacon.c struct device *dev = &evt->rc->uwb_dev.dev; evt 462 drivers/staging/uwb/beacon.c if (evt->notif.size < sizeof(*bs)) { evt 465 drivers/staging/uwb/beacon.c evt->notif.size, sizeof(*bs)); evt 468 drivers/staging/uwb/beacon.c bs = container_of(evt->notif.rceb, struct uwb_rc_evt_beacon_size, rceb); evt 492 drivers/staging/uwb/beacon.c int uwbd_evt_handle_rc_bp_slot_change(struct uwb_event *evt) evt 494 drivers/staging/uwb/beacon.c struct uwb_rc *rc = evt->rc; evt 498 drivers/staging/uwb/beacon.c if (evt->notif.size < sizeof(*bpsc)) { evt 502 drivers/staging/uwb/beacon.c bpsc = container_of(evt->notif.rceb, struct uwb_rc_evt_bp_slot_change, rceb); evt 526 drivers/staging/uwb/beacon.c int uwbd_evt_handle_rc_bpoie_change(struct uwb_event *evt) evt 529 drivers/staging/uwb/beacon.c struct device *dev = &evt->rc->uwb_dev.dev; evt 536 drivers/staging/uwb/beacon.c if (evt->notif.size < sizeof(*bpoiec)) { evt 539 drivers/staging/uwb/beacon.c evt->notif.size, sizeof(*bpoiec)); evt 542 drivers/staging/uwb/beacon.c bpoiec = container_of(evt->notif.rceb, struct uwb_rc_evt_bpoie_change, rceb); evt 215 drivers/staging/uwb/drp-avail.c int uwbd_evt_get_drp_avail(struct uwb_event *evt, unsigned long *bmp) evt 217 drivers/staging/uwb/drp-avail.c struct device *dev = &evt->rc->uwb_dev.dev; evt 222 drivers/staging/uwb/drp-avail.c if (evt->notif.size < sizeof(*drp_evt)) { evt 225 drivers/staging/uwb/drp-avail.c "needed]\n", evt->notif.size, sizeof(*drp_evt)); evt 228 drivers/staging/uwb/drp-avail.c drp_evt = container_of(evt->notif.rceb, struct uwb_rc_evt_drp_avail, rceb); evt 259 drivers/staging/uwb/drp-avail.c int uwbd_evt_handle_rc_drp_avail(struct uwb_event *evt) evt 262 drivers/staging/uwb/drp-avail.c struct uwb_rc *rc = evt->rc; evt 265 drivers/staging/uwb/drp-avail.c result = uwbd_evt_get_drp_avail(evt, bmp); evt 792 drivers/staging/uwb/drp.c int uwbd_evt_handle_rc_drp(struct uwb_event *evt) evt 794 drivers/staging/uwb/drp.c struct device *dev = &evt->rc->uwb_dev.dev; evt 795 drivers/staging/uwb/drp.c struct uwb_rc *rc = evt->rc; evt 803 drivers/staging/uwb/drp.c if (evt->notif.size < sizeof(*drp_evt)) { evt 806 drivers/staging/uwb/drp.c evt->notif.size, sizeof(*drp_evt)); evt 809 drivers/staging/uwb/drp.c bytes_left = evt->notif.size - sizeof(*drp_evt); evt 810 drivers/staging/uwb/drp.c drp_evt = container_of(evt->notif.rceb, struct uwb_rc_evt_drp, rceb); evt 18 drivers/staging/uwb/ie-rcv.c int uwbd_evt_handle_rc_ie_rcv(struct uwb_event *evt) evt 21 drivers/staging/uwb/ie-rcv.c struct device *dev = &evt->rc->uwb_dev.dev; evt 25 drivers/staging/uwb/ie-rcv.c if (evt->notif.size < sizeof(*iercv)) { evt 28 drivers/staging/uwb/ie-rcv.c evt->notif.size, sizeof(*iercv)); evt 31 drivers/staging/uwb/ie-rcv.c iercv = container_of(evt->notif.rceb, struct uwb_rc_evt_ie_rcv, rceb); evt 671 drivers/staging/uwb/include/spec.h const struct uwb_rc_evt_bp_slot_change *evt) evt 673 drivers/staging/uwb/include/spec.h return evt->slot_info & 0x7f; evt 677 drivers/staging/uwb/include/spec.h const struct uwb_rc_evt_bp_slot_change *evt) evt 679 drivers/staging/uwb/include/spec.h return (evt->slot_info & 0x80) >> 7; evt 705 drivers/staging/uwb/include/spec.h static inline enum uwb_drp_notif_reason uwb_rc_evt_drp_reason(struct uwb_rc_evt_drp *evt) evt 707 drivers/staging/uwb/include/spec.h return evt->reason & 0x0f; evt 94 drivers/staging/uwb/neh.c __le16 evt; evt 216 drivers/staging/uwb/neh.c neh->evt = cpu_to_le16(expected_event); evt 294 drivers/staging/uwb/neh.c && neh->evt == rceb->wEvent evt 309 drivers/staging/uwb/reset.c int uwbd_msg_handle_reset(struct uwb_event *evt) evt 311 drivers/staging/uwb/reset.c struct uwb_rc *rc = evt->rc; evt 338 drivers/staging/uwb/reset.c struct uwb_event *evt; evt 340 drivers/staging/uwb/reset.c evt = kzalloc(sizeof(struct uwb_event), GFP_ATOMIC); evt 341 drivers/staging/uwb/reset.c if (unlikely(evt == NULL)) evt 344 drivers/staging/uwb/reset.c evt->rc = __uwb_rc_get(rc); /* will be put by uwbd's uwbd_event_handle() */ evt 345 drivers/staging/uwb/reset.c evt->ts_jiffies = jiffies; evt 346 drivers/staging/uwb/reset.c evt->type = UWB_EVT_TYPE_MSG; evt 347 drivers/staging/uwb/reset.c evt->message = UWB_EVT_MSG_RESET; evt 349 drivers/staging/uwb/reset.c uwbd_event_queue(evt); evt 234 drivers/staging/uwb/uwb-internal.h int uwbd_msg_handle_reset(struct uwb_event *evt); evt 241 drivers/staging/uwb/uwb-internal.h int uwbd_evt_handle_rc_dev_addr_conflict(struct uwb_event *evt); evt 165 drivers/staging/uwb/uwbd.c int uwbd_event_handle_urc(struct uwb_event *evt) evt 173 drivers/staging/uwb/uwbd.c type = evt->notif.rceb->bEventType; evt 174 drivers/staging/uwb/uwbd.c event = le16_to_cpu(evt->notif.rceb->wEvent); evt 175 drivers/staging/uwb/uwbd.c context = evt->notif.rceb->bEventContext; evt 188 drivers/staging/uwb/uwbd.c result = (*handler)(evt); evt 191 drivers/staging/uwb/uwbd.c dev_err(&evt->rc->uwb_dev.dev, evt 197 drivers/staging/uwb/uwbd.c static void uwbd_event_handle_message(struct uwb_event *evt) evt 202 drivers/staging/uwb/uwbd.c rc = evt->rc; evt 204 drivers/staging/uwb/uwbd.c if (evt->message < 0 || evt->message >= ARRAY_SIZE(uwbd_message_handlers)) { evt 205 drivers/staging/uwb/uwbd.c dev_err(&rc->uwb_dev.dev, "UWBD: invalid message type %d\n", evt->message); evt 209 drivers/staging/uwb/uwbd.c result = uwbd_message_handlers[evt->message].handler(evt); evt 212 drivers/staging/uwb/uwbd.c uwbd_message_handlers[evt->message].name, result); evt 215 drivers/staging/uwb/uwbd.c static void uwbd_event_handle(struct uwb_event *evt) evt 220 drivers/staging/uwb/uwbd.c rc = evt->rc; evt 223 drivers/staging/uwb/uwbd.c switch (evt->type) { evt 225 drivers/staging/uwb/uwbd.c should_keep = uwbd_event_handle_urc(evt); evt 227 drivers/staging/uwb/uwbd.c kfree(evt->notif.rceb); evt 230 drivers/staging/uwb/uwbd.c uwbd_event_handle_message(evt); evt 233 drivers/staging/uwb/uwbd.c dev_err(&rc->uwb_dev.dev, "UWBD: invalid event type %d\n", evt->type); evt 257 drivers/staging/uwb/uwbd.c struct uwb_event *evt; evt 271 drivers/staging/uwb/uwbd.c evt = list_first_entry(&rc->uwbd.event_list, struct uwb_event, list_node); evt 272 drivers/staging/uwb/uwbd.c list_del(&evt->list_node); evt 274 drivers/staging/uwb/uwbd.c evt = NULL; evt 277 drivers/staging/uwb/uwbd.c if (evt) { evt 278 drivers/staging/uwb/uwbd.c uwbd_event_handle(evt); evt 279 drivers/staging/uwb/uwbd.c kfree(evt); evt 322 drivers/staging/uwb/uwbd.c void uwbd_event_queue(struct uwb_event *evt) evt 324 drivers/staging/uwb/uwbd.c struct uwb_rc *rc = evt->rc; evt 329 drivers/staging/uwb/uwbd.c list_add(&evt->list_node, &rc->uwbd.event_list); evt 332 drivers/staging/uwb/uwbd.c __uwb_rc_put(evt->rc); evt 333 drivers/staging/uwb/uwbd.c if (evt->type == UWB_EVT_TYPE_NOTIF) evt 334 drivers/staging/uwb/uwbd.c kfree(evt->notif.rceb); evt 335 drivers/staging/uwb/uwbd.c kfree(evt); evt 343 drivers/staging/uwb/uwbd.c struct uwb_event *evt, *nxt; evt 346 drivers/staging/uwb/uwbd.c list_for_each_entry_safe(evt, nxt, &rc->uwbd.event_list, list_node) { evt 347 drivers/staging/uwb/uwbd.c if (evt->rc == rc) { evt 349 drivers/staging/uwb/uwbd.c list_del(&evt->list_node); evt 350 drivers/staging/uwb/uwbd.c if (evt->type == UWB_EVT_TYPE_NOTIF) evt 351 drivers/staging/uwb/uwbd.c kfree(evt->notif.rceb); evt 352 drivers/staging/uwb/uwbd.c kfree(evt); evt 193 drivers/staging/wilc1000/wilc_hif.c static int handle_scan_done(struct wilc_vif *vif, enum scan_event evt) evt 201 drivers/staging/wilc1000/wilc_hif.c if (evt == SCAN_EVENT_ABORTED) { evt 222 drivers/staging/wilc1000/wilc_hif.c scan_req->scan_result(evt, NULL, scan_req->arg); evt 119 drivers/staging/wilc1000/wilc_hif.h void (*scan_result)(enum scan_event evt, evt 135 drivers/staging/wilc1000/wilc_hif.h void (*conn_result)(enum conn_event evt, u8 status, void *priv_data); evt 328 drivers/usb/dwc3/core.c struct dwc3_event_buffer *evt) evt 330 drivers/usb/dwc3/core.c dma_free_coherent(dwc->sysdev, evt->length, evt->buf, evt->dma); evt 344 drivers/usb/dwc3/core.c struct dwc3_event_buffer *evt; evt 346 drivers/usb/dwc3/core.c evt = devm_kzalloc(dwc->dev, sizeof(*evt), GFP_KERNEL); evt 347 drivers/usb/dwc3/core.c if (!evt) evt 350 drivers/usb/dwc3/core.c evt->dwc = dwc; evt 351 drivers/usb/dwc3/core.c evt->length = length; evt 352 drivers/usb/dwc3/core.c evt->cache = devm_kzalloc(dwc->dev, length, GFP_KERNEL); evt 353 drivers/usb/dwc3/core.c if (!evt->cache) evt 356 drivers/usb/dwc3/core.c evt->buf = dma_alloc_coherent(dwc->sysdev, length, evt 357 drivers/usb/dwc3/core.c &evt->dma, GFP_KERNEL); evt 358 drivers/usb/dwc3/core.c if (!evt->buf) evt 361 drivers/usb/dwc3/core.c return evt; evt 370 drivers/usb/dwc3/core.c struct dwc3_event_buffer *evt; evt 372 drivers/usb/dwc3/core.c evt = dwc->ev_buf; evt 373 drivers/usb/dwc3/core.c if (evt) evt 374 drivers/usb/dwc3/core.c dwc3_free_one_event_buffer(dwc, evt); evt 387 drivers/usb/dwc3/core.c struct dwc3_event_buffer *evt; evt 389 drivers/usb/dwc3/core.c evt = dwc3_alloc_one_event_buffer(dwc, length); evt 390 drivers/usb/dwc3/core.c if (IS_ERR(evt)) { evt 392 drivers/usb/dwc3/core.c return PTR_ERR(evt); evt 394 drivers/usb/dwc3/core.c dwc->ev_buf = evt; evt 407 drivers/usb/dwc3/core.c struct dwc3_event_buffer *evt; evt 409 drivers/usb/dwc3/core.c evt = dwc->ev_buf; evt 410 drivers/usb/dwc3/core.c evt->lpos = 0; evt 412 drivers/usb/dwc3/core.c lower_32_bits(evt->dma)); evt 414 drivers/usb/dwc3/core.c upper_32_bits(evt->dma)); evt 416 drivers/usb/dwc3/core.c DWC3_GEVNTSIZ_SIZE(evt->length)); evt 424 drivers/usb/dwc3/core.c struct dwc3_event_buffer *evt; evt 426 drivers/usb/dwc3/core.c evt = dwc->ev_buf; evt 428 drivers/usb/dwc3/core.c evt->lpos = 0; evt 372 drivers/usb/dwc3/debug.h const union dwc3_event evt = (union dwc3_event) event; evt 374 drivers/usb/dwc3/debug.h if (evt.type.is_devspec) evt 375 drivers/usb/dwc3/debug.h return dwc3_gadget_event_string(str, size, &evt.devt); evt 377 drivers/usb/dwc3/debug.h return dwc3_ep_event_string(str, size, &evt.depevt, ep0state); evt 3197 drivers/usb/dwc3/gadget.c static irqreturn_t dwc3_process_event_buf(struct dwc3_event_buffer *evt) evt 3199 drivers/usb/dwc3/gadget.c struct dwc3 *dwc = evt->dwc; evt 3204 drivers/usb/dwc3/gadget.c left = evt->count; evt 3206 drivers/usb/dwc3/gadget.c if (!(evt->flags & DWC3_EVENT_PENDING)) evt 3212 drivers/usb/dwc3/gadget.c event.raw = *(u32 *) (evt->cache + evt->lpos); evt 3225 drivers/usb/dwc3/gadget.c evt->lpos = (evt->lpos + 4) % evt->length; evt 3229 drivers/usb/dwc3/gadget.c evt->count = 0; evt 3230 drivers/usb/dwc3/gadget.c evt->flags &= ~DWC3_EVENT_PENDING; evt 3248 drivers/usb/dwc3/gadget.c struct dwc3_event_buffer *evt = _evt; evt 3249 drivers/usb/dwc3/gadget.c struct dwc3 *dwc = evt->dwc; evt 3254 drivers/usb/dwc3/gadget.c ret = dwc3_process_event_buf(evt); evt 3260 drivers/usb/dwc3/gadget.c static irqreturn_t dwc3_check_event_buf(struct dwc3_event_buffer *evt) evt 3262 drivers/usb/dwc3/gadget.c struct dwc3 *dwc = evt->dwc; evt 3280 drivers/usb/dwc3/gadget.c if (evt->flags & DWC3_EVENT_PENDING) evt 3288 drivers/usb/dwc3/gadget.c evt->count = count; evt 3289 drivers/usb/dwc3/gadget.c evt->flags |= DWC3_EVENT_PENDING; evt 3296 drivers/usb/dwc3/gadget.c amount = min(count, evt->length - evt->lpos); evt 3297 drivers/usb/dwc3/gadget.c memcpy(evt->cache + evt->lpos, evt->buf + evt->lpos, amount); evt 3300 drivers/usb/dwc3/gadget.c memcpy(evt->cache, evt->buf, count - amount); evt 3309 drivers/usb/dwc3/gadget.c struct dwc3_event_buffer *evt = _evt; evt 3311 drivers/usb/dwc3/gadget.c return dwc3_check_event_buf(evt); evt 651 drivers/usb/host/xhci-dbgcap.c union xhci_trb *evt; evt 746 drivers/usb/host/xhci-dbgcap.c evt = dbc->ring_evt->dequeue; evt 747 drivers/usb/host/xhci-dbgcap.c while ((le32_to_cpu(evt->event_cmd.flags) & TRB_CYCLE) == evt 755 drivers/usb/host/xhci-dbgcap.c trace_xhci_dbc_handle_event(dbc->ring_evt, &evt->generic); evt 757 drivers/usb/host/xhci-dbgcap.c switch (le32_to_cpu(evt->event_cmd.flags) & TRB_TYPE_BITMASK) { evt 759 drivers/usb/host/xhci-dbgcap.c dbc_handle_port_status(xhci, evt); evt 762 drivers/usb/host/xhci-dbgcap.c dbc_handle_xfer_event(xhci, evt); evt 769 drivers/usb/host/xhci-dbgcap.c evt = dbc->ring_evt->dequeue; evt 400 drivers/vhost/scsi.c static void vhost_scsi_free_evt(struct vhost_scsi *vs, struct vhost_scsi_evt *evt) evt 403 drivers/vhost/scsi.c kfree(evt); evt 411 drivers/vhost/scsi.c struct vhost_scsi_evt *evt; evt 418 drivers/vhost/scsi.c evt = kzalloc(sizeof(*evt), GFP_KERNEL); evt 419 drivers/vhost/scsi.c if (!evt) { evt 425 drivers/vhost/scsi.c evt->event.event = cpu_to_vhost32(vq, event); evt 426 drivers/vhost/scsi.c evt->event.reason = cpu_to_vhost32(vq, reason); evt 429 drivers/vhost/scsi.c return evt; evt 447 drivers/vhost/scsi.c vhost_scsi_do_evt_work(struct vhost_scsi *vs, struct vhost_scsi_evt *evt) evt 450 drivers/vhost/scsi.c struct virtio_scsi_event *event = &evt->event; evt 501 drivers/vhost/scsi.c struct vhost_scsi_evt *evt, *t; evt 506 drivers/vhost/scsi.c llist_for_each_entry_safe(evt, t, llnode, list) { evt 507 drivers/vhost/scsi.c vhost_scsi_do_evt_work(vs, evt); evt 508 drivers/vhost/scsi.c vhost_scsi_free_evt(vs, evt); evt 1295 drivers/vhost/scsi.c struct vhost_scsi_evt *evt; evt 1297 drivers/vhost/scsi.c evt = vhost_scsi_allocate_evt(vs, event, reason); evt 1298 drivers/vhost/scsi.c if (!evt) evt 1307 drivers/vhost/scsi.c evt->event.lun[0] = 0x01; evt 1308 drivers/vhost/scsi.c evt->event.lun[1] = tpg->tport_tpgt; evt 1310 drivers/vhost/scsi.c evt->event.lun[2] = lun->unpacked_lun >> 8 | 0x40 ; evt 1311 drivers/vhost/scsi.c evt->event.lun[3] = lun->unpacked_lun & 0xFF; evt 1314 drivers/vhost/scsi.c llist_add(&evt->list, &vs->vs_event_list); evt 16 drivers/virtio/virtio_input.c struct virtqueue *evt, *sts; evt 28 drivers/virtio/virtio_input.c virtqueue_add_inbuf(vi->evt, sg, 1, evtbuf, GFP_ATOMIC); evt 40 drivers/virtio/virtio_input.c while ((event = virtqueue_get_buf(vi->evt, &len)) != NULL) { evt 180 drivers/virtio/virtio_input.c vi->evt = vqs[0]; evt 192 drivers/virtio/virtio_input.c size = virtqueue_get_vring_size(vi->evt); evt 197 drivers/virtio/virtio_input.c virtqueue_kick(vi->evt); evt 102 include/linux/clockchips.h int (*set_next_event)(unsigned long evt, struct clock_event_device *); evt 182 include/linux/clockchips.h extern u64 clockevent_delta2ns(unsigned long latch, struct clock_event_device *evt); evt 307 include/linux/iommu.h struct iommu_fault_event *evt, evt 477 include/linux/iommu.h struct iommu_fault_event *evt); evt 840 include/linux/iommu.h int iommu_report_device_fault(struct device *dev, struct iommu_fault_event *evt) evt 553 include/linux/leds.h extern void ledtrig_cpu(enum cpu_led_event evt); evt 555 include/linux/leds.h static inline void ledtrig_cpu(enum cpu_led_event evt) evt 90 include/linux/perf/arm_pmu.h int (*set_event_filter)(struct hw_perf_event *evt, evt 17 include/linux/sh_intc.h #define evt2irq(evt) (((evt) >> 5) - 16) evt 20 include/linux/sh_intc.h #define evt2irq(evt) (evt) evt 2264 include/net/bluetooth/hci.h __u8 evt; evt 442 include/net/bluetooth/hci_core.h void (*notify)(struct hci_dev *hdev, unsigned int evt); evt 413 include/scsi/scsi_device.h extern void sdev_evt_send(struct scsi_device *sdev, struct scsi_event *evt); evt 3353 kernel/events/core.c struct perf_event **evt, *evt1, *evt2; evt 3362 kernel/events/core.c evt = &evt1; evt 3364 kernel/events/core.c evt = &evt2; evt 3366 kernel/events/core.c evt = &evt1; evt 3368 kernel/events/core.c evt = &evt2; evt 3371 kernel/events/core.c ret = func(*evt, data); evt 3375 kernel/events/core.c *evt = perf_event_groups_next(*evt); evt 32 kernel/time/clockevents.c static u64 cev_delta2ns(unsigned long latch, struct clock_event_device *evt, evt 35 kernel/time/clockevents.c u64 clc = (u64) latch << evt->shift; evt 38 kernel/time/clockevents.c if (WARN_ON(!evt->mult)) evt 39 kernel/time/clockevents.c evt->mult = 1; evt 40 kernel/time/clockevents.c rnd = (u64) evt->mult - 1; evt 46 kernel/time/clockevents.c if ((clc >> evt->shift) != (u64)latch) evt 69 kernel/time/clockevents.c (!ismax || evt->mult <= (1ULL << evt->shift))) evt 72 kernel/time/clockevents.c do_div(clc, evt->mult); evt 85 kernel/time/clockevents.c u64 clockevent_delta2ns(unsigned long latch, struct clock_event_device *evt) evt 87 kernel/time/clockevents.c return cev_delta2ns(latch, evt, false); evt 20 kernel/time/tick-broadcast-hrtimer.c static int bc_shutdown(struct clock_event_device *evt) evt 248 kernel/time/tick-broadcast.c struct clock_event_device *evt = td->evtdev; evt 250 kernel/time/tick-broadcast.c if (!evt) evt 253 kernel/time/tick-broadcast.c if (!evt->event_handler) evt 256 kernel/time/tick-broadcast.c evt->event_handler(evt); evt 170 kernel/watchdog_hld.c struct perf_event *evt; evt 176 kernel/watchdog_hld.c evt = perf_event_create_kernel_counter(wd_attr, cpu, NULL, evt 178 kernel/watchdog_hld.c if (IS_ERR(evt)) { evt 180 kernel/watchdog_hld.c PTR_ERR(evt)); evt 181 kernel/watchdog_hld.c return PTR_ERR(evt); evt 183 kernel/watchdog_hld.c this_cpu_write(watchdog_ev, evt); evt 5793 net/bluetooth/hci_event.c if (hdr->evt != event) evt 5801 net/bluetooth/hci_event.c if (hdr->evt == HCI_EV_CMD_STATUS) evt 5804 net/bluetooth/hci_event.c if (hdr->evt != HCI_EV_CMD_COMPLETE) { evt 5806 net/bluetooth/hci_event.c hdr->evt); evt 5833 net/bluetooth/hci_event.c u8 status = 0, event = hdr->evt, req_evt = 0; evt 712 net/bluetooth/hci_sock.c hdr->evt = HCI_EV_STACK_INTERNAL; evt 19 net/ipv4/tcp_dctcp.h static inline void dctcp_ece_ack_update(struct sock *sk, enum tcp_ca_event evt, evt 22 net/ipv4/tcp_dctcp.h u32 new_ce_state = (evt == CA_EVENT_ECN_IS_CE) ? 1 : 0; evt 605 net/tipc/bearer.c static int tipc_l2_device_event(struct notifier_block *nb, unsigned long evt, evt 616 net/tipc/bearer.c trace_tipc_l2_device_event(dev, b, evt); evt 617 net/tipc/bearer.c switch (evt) { evt 666 net/tipc/group.c struct tipc_event evt; evt 670 net/tipc/group.c memset(&evt, 0, sizeof(evt)); evt 671 net/tipc/group.c evt.event = event; evt 672 net/tipc/group.c evt.found_lower = m->instance; evt 673 net/tipc/group.c evt.found_upper = m->instance; evt 674 net/tipc/group.c evt.port.ref = m->port; evt 675 net/tipc/group.c evt.port.node = m->node; evt 676 net/tipc/group.c evt.s.seq.type = grp->type; evt 677 net/tipc/group.c evt.s.seq.lower = m->instance; evt 678 net/tipc/group.c evt.s.seq.upper = m->instance; evt 681 net/tipc/group.c GROUP_H_SIZE, sizeof(evt), dnode, m->node, evt 691 net/tipc/group.c memcpy(msg_data(hdr), &evt, sizeof(evt)); evt 855 net/tipc/group.c struct tipc_event *evt = (void *)msg_data(hdr); evt 856 net/tipc/group.c u32 instance = evt->found_lower; evt 857 net/tipc/group.c u32 node = evt->port.node; evt 858 net/tipc/group.c u32 port = evt->port.ref; evt 859 net/tipc/group.c int event = evt->event; evt 553 net/tipc/link.c int tipc_link_fsm_evt(struct tipc_link *l, int evt) evt 560 net/tipc/link.c switch (evt) { evt 578 net/tipc/link.c switch (evt) { evt 596 net/tipc/link.c switch (evt) { evt 613 net/tipc/link.c switch (evt) { evt 630 net/tipc/link.c switch (evt) { evt 651 net/tipc/link.c switch (evt) { evt 676 net/tipc/link.c switch (evt) { evt 703 net/tipc/link.c trace_tipc_link_fsm(l->name, old_state, l->state, evt); evt 707 net/tipc/link.c evt, l->state, l->name); evt 708 net/tipc/link.c trace_tipc_link_fsm(l->name, old_state, l->state, evt); evt 96 net/tipc/link.h int tipc_link_fsm_evt(struct tipc_link *l, int evt); evt 288 net/tipc/name_table.c struct tipc_subscr *sb = &sub->evt.s; evt 680 net/tipc/name_table.c struct tipc_subscr *s = &sub->evt.s; evt 709 net/tipc/name_table.c struct tipc_subscr *s = &sub->evt.s; evt 163 net/tipc/node.c static void tipc_node_fsm_evt(struct tipc_node *n, int evt); evt 1144 net/tipc/node.c static void tipc_node_fsm_evt(struct tipc_node *n, int evt) evt 1150 net/tipc/node.c switch (evt) { evt 1169 net/tipc/node.c switch (evt) { evt 1192 net/tipc/node.c switch (evt) { evt 1209 net/tipc/node.c switch (evt) { evt 1228 net/tipc/node.c switch (evt) { evt 1247 net/tipc/node.c switch (evt) { evt 1264 net/tipc/node.c switch (evt) { evt 1285 net/tipc/node.c switch (evt) { evt 1311 net/tipc/node.c trace_tipc_node_fsm(n->peer_id, n->state, state, evt); evt 1316 net/tipc/node.c pr_err("Illegal node fsm evt %x in state %x\n", evt, state); evt 1317 net/tipc/node.c trace_tipc_node_fsm(n->peer_id, n->state, state, evt); evt 45 net/tipc/subscr.c struct tipc_event *evt = &sub->evt; evt 49 net/tipc/subscr.c tipc_evt_write(evt, event, event); evt 50 net/tipc/subscr.c tipc_evt_write(evt, found_lower, found_lower); evt 51 net/tipc/subscr.c tipc_evt_write(evt, found_upper, found_upper); evt 52 net/tipc/subscr.c tipc_evt_write(evt, port.ref, port); evt 53 net/tipc/subscr.c tipc_evt_write(evt, port.node, node); evt 54 net/tipc/subscr.c tipc_topsrv_queue_evt(sub->net, sub->conid, event, evt); evt 80 net/tipc/subscr.c struct tipc_subscr *s = &sub->evt.s; evt 106 net/tipc/subscr.c struct tipc_subscr *s = &sub->evt.s; evt 153 net/tipc/subscr.c memcpy(&sub->evt.s, s, sizeof(*s)); evt 161 net/tipc/subscr.c timeout = tipc_sub_read(&sub->evt.s, timeout); evt 170 net/tipc/subscr.c if (sub->evt.s.timeout != TIPC_WAIT_FOREVER) evt 63 net/tipc/subscr.h struct tipc_event evt; evt 110 net/tipc/topsrv.c struct tipc_event evt; evt 116 net/tipc/topsrv.c static void tipc_topsrv_kern_evt(struct net *net, struct tipc_event *evt); evt 237 net/tipc/topsrv.c if (!s || !memcmp(s, &sub->evt.s, sizeof(*s))) { evt 252 net/tipc/topsrv.c struct tipc_event *evt; evt 262 net/tipc/topsrv.c evt = &e->evt; evt 266 net/tipc/topsrv.c tipc_conn_delete_sub(con, &evt->s); evt 270 net/tipc/topsrv.c iov.iov_base = evt; evt 271 net/tipc/topsrv.c iov.iov_len = sizeof(*evt); evt 276 net/tipc/topsrv.c 1, sizeof(*evt)); evt 284 net/tipc/topsrv.c tipc_topsrv_kern_evt(srv->net, evt); evt 313 net/tipc/topsrv.c u32 event, struct tipc_event *evt) evt 330 net/tipc/topsrv.c memcpy(&e->evt, evt, sizeof(*evt)); evt 598 net/tipc/topsrv.c static void tipc_topsrv_kern_evt(struct net *net, struct tipc_event *evt) evt 600 net/tipc/topsrv.c u32 port = *(u32 *)&evt->s.usr_handle; evt 605 net/tipc/topsrv.c skb = tipc_msg_create(TOP_SRV, 0, INT_H_SIZE, sizeof(*evt), evt 610 net/tipc/topsrv.c memcpy(msg_data(buf_msg(skb)), evt, sizeof(*evt)); evt 48 net/tipc/topsrv.h u32 event, struct tipc_event *evt); evt 362 net/tipc/trace.h TP_PROTO(const char *name, u32 os, u32 ns, int evt), evt 364 net/tipc/trace.h TP_ARGS(name, os, ns, evt), evt 370 net/tipc/trace.h __field(u32, evt) evt 377 net/tipc/trace.h __entry->evt = evt; evt 381 net/tipc/trace.h state_sym(__entry->os), evt_sym(__entry->evt), evt 387 net/tipc/trace.h TP_PROTO(const char *name, u32 os, u32 ns, int evt), \ evt 388 net/tipc/trace.h TP_ARGS(name, os, ns, evt)) evt 395 net/tipc/trace.h unsigned long evt), evt 397 net/tipc/trace.h TP_ARGS(dev, b, evt), evt 402 net/tipc/trace.h __field(unsigned long, evt) evt 411 net/tipc/trace.h __entry->evt = evt; evt 418 net/tipc/trace.h dev_evt_sym(__entry->evt), __get_str(dev_name), evt 2962 net/wireless/trace.h TP_PROTO(struct net_device *netdev, enum nl80211_radar_event evt), evt 2963 net/wireless/trace.h TP_ARGS(netdev, evt), evt 2966 net/wireless/trace.h __field(enum nl80211_radar_event, evt) evt 2970 net/wireless/trace.h __entry->evt = evt; evt 2973 net/wireless/trace.h NETDEV_PR_ARG, __entry->evt) evt 949 samples/vfio-mdev/mtty.c struct eventfd_ctx *evt; evt 951 samples/vfio-mdev/mtty.c evt = eventfd_ctx_fdget(fd); evt 952 samples/vfio-mdev/mtty.c if (IS_ERR(evt)) { evt 953 samples/vfio-mdev/mtty.c ret = PTR_ERR(evt); evt 956 samples/vfio-mdev/mtty.c mdev_state->intx_evtfd = evt; evt 981 samples/vfio-mdev/mtty.c struct eventfd_ctx *evt; evt 989 samples/vfio-mdev/mtty.c evt = eventfd_ctx_fdget(fd); evt 990 samples/vfio-mdev/mtty.c if (IS_ERR(evt)) { evt 991 samples/vfio-mdev/mtty.c ret = PTR_ERR(evt); evt 994 samples/vfio-mdev/mtty.c mdev_state->msi_evtfd = evt; evt 382 sound/xen/xen_snd_front_alsa.c stream->evt_pair->evt.u.evt.substream = substream; evt 578 sound/xen/xen_snd_front_alsa.c struct snd_pcm_substream *substream = evtchnl->u.evt.substream; evt 94 sound/xen/xen_snd_front_evtchnl.c struct xensnd_event_page *page = channel->u.evt.page; evt 155 sound/xen/xen_snd_front_evtchnl.c page = (unsigned long)channel->u.evt.page; evt 191 sound/xen/xen_snd_front_evtchnl.c evtchnl_free(front_info, &front_info->evt_pairs[i].evt); evt 253 sound/xen/xen_snd_front_evtchnl.c channel->u.evt.page = (struct xensnd_event_page *)page; evt 326 sound/xen/xen_snd_front_evtchnl.c &front_info->evt_pairs[index].evt, evt 346 sound/xen/xen_snd_front_evtchnl.c &front_info->evt_pairs[index].evt, evt 420 sound/xen/xen_snd_front_evtchnl.c &front_info->evt_pairs[index].evt, evt 440 sound/xen/xen_snd_front_evtchnl.c &front_info->evt_pairs[index].evt, evt 479 sound/xen/xen_snd_front_evtchnl.c mutex_lock(&evt_pair->evt.ring_io_lock); evt 480 sound/xen/xen_snd_front_evtchnl.c evt_pair->evt.state = state; evt 481 sound/xen/xen_snd_front_evtchnl.c mutex_unlock(&evt_pair->evt.ring_io_lock); evt 490 sound/xen/xen_snd_front_evtchnl.c mutex_lock(&evt_pair->evt.ring_io_lock); evt 491 sound/xen/xen_snd_front_evtchnl.c evt_pair->evt.evt_next_id = 0; evt 492 sound/xen/xen_snd_front_evtchnl.c mutex_unlock(&evt_pair->evt.ring_io_lock); evt 72 sound/xen/xen_snd_front_evtchnl.h } evt; evt 78 sound/xen/xen_snd_front_evtchnl.h struct xen_snd_front_evtchnl evt;