cycles 1252 arch/alpha/kernel/setup.c int cycles, prev_cycles = 1000000; cycles 1264 arch/alpha/kernel/setup.c cycles = read_mem_block(__va(0), stride, size); cycles 1265 arch/alpha/kernel/setup.c if (cycles > prev_cycles * 2) { cycles 1269 arch/alpha/kernel/setup.c size >> 11, prev_cycles, cycles); cycles 1274 arch/alpha/kernel/setup.c prev_cycles = cycles; cycles 162 arch/arc/include/asm/arcregs.h unsigned int pad:8, x1616:8, dsp:4, cycles:2, type:2, ver:8; cycles 164 arch/arc/include/asm/arcregs.h unsigned int ver:8, type:2, cycles:2, dsp:4, x1616:8, pad:8; cycles 45 arch/arm/lib/delay.c static void __timer_delay(unsigned long cycles) cycles 49 arch/arm/lib/delay.c while ((get_cycles() - start) < cycles) cycles 288 arch/arm/mach-davinci/time.c static int davinci_set_next_event(unsigned long cycles, cycles 293 arch/arm/mach-davinci/time.c t->period = cycles; cycles 85 arch/arm/mach-imx/mmdc.c PMU_EVENT_ATTR_STRING(total-cycles, mmdc_pmu_total_cycles, "event=0x00") cycles 86 arch/arm/mach-imx/mmdc.c PMU_EVENT_ATTR_STRING(busy-cycles, mmdc_pmu_busy_cycles, "event=0x01") cycles 120 arch/arm/mach-omap1/time.c static int omap_mpu_set_next_event(unsigned long cycles, cycles 123 arch/arm/mach-omap1/time.c omap_mpu_timer_start(0, cycles, 0); cycles 100 arch/arm/mach-omap2/timer.c static int omap2_gp_timer_set_next_event(unsigned long cycles, cycles 104 arch/arm/mach-omap2/timer.c 0xffffffff - cycles, OMAP_TIMER_POSTED); cycles 433 arch/arm/mach-omap2/vc.c u32 cycles; cycles 438 arch/arm/mach-omap2/vc.c cycles = voltdm->sys_clk.rate / 1000 * time / 1000; cycles 440 arch/arm/mach-omap2/vc.c cycles /= 64; cycles 446 arch/arm/mach-omap2/vc.c if (cycles > 63) { cycles 447 arch/arm/mach-omap2/vc.c cycles /= 4; cycles 452 arch/arm/mach-omap2/vc.c if (cycles > 63) { cycles 453 arch/arm/mach-omap2/vc.c cycles /= 2; cycles 458 arch/arm/mach-omap2/vc.c if (cycles > 63) { cycles 459 arch/arm/mach-omap2/vc.c cycles /= 4; cycles 464 arch/arm/mach-omap2/vc.c if (cycles > 63) { cycles 470 arch/arm/mach-omap2/vc.c cycles++; cycles 473 arch/arm/mach-omap2/vc.c (cycles << OMAP4430_RAMP_UP_COUNT_SHIFT); cycles 157 arch/arm/mach-spear/time.c static int clockevent_next_event(unsigned long cycles, cycles 165 arch/arm/mach-spear/time.c writew(cycles, gpt_base + LOAD(CLKEVT)); cycles 51 arch/arm/plat-omap/counter_32k.c static cycles_t cycles; cycles 59 arch/arm/plat-omap/counter_32k.c last_cycles = cycles; cycles 60 arch/arm/plat-omap/counter_32k.c cycles = sync32k_cnt_reg ? readl_relaxed(sync32k_cnt_reg) : 0; cycles 62 arch/arm/plat-omap/counter_32k.c nsecs = clocksource_cyc2ns(cycles - last_cycles, cycles 26 arch/arm64/lib/delay.c void __delay(unsigned long cycles) cycles 34 arch/arm64/lib/delay.c while ((get_cycles() - start + timer_evt_period) < cycles) cycles 38 arch/arm64/lib/delay.c while ((get_cycles() - start) < cycles) cycles 14 arch/h8300/lib/delay.c void __delay(unsigned long cycles) cycles 17 arch/h8300/lib/delay.c "bne 1b":"=r"(cycles):"0"(cycles)); cycles 11 arch/hexagon/include/asm/delay.h extern void __delay(unsigned long cycles); cycles 206 arch/hexagon/kernel/time.c void __delay(unsigned long cycles) cycles 210 arch/hexagon/kernel/time.c while ((__vmgettime() - start) < cycles) cycles 1369 arch/ia64/include/asm/pal.h cycles : 8, cycles 667 arch/ia64/kernel/palinfo.c pm_info.pal_perf_mon_info_s.cycles, cycles 82 arch/m68k/68000/timers.c u32 cycles; cycles 85 arch/m68k/68000/timers.c cycles = m68328_tick_cnt + TCN; cycles 88 arch/m68k/68000/timers.c return cycles; cycles 125 arch/m68k/coldfire/pit.c u32 cycles; cycles 130 arch/m68k/coldfire/pit.c cycles = pit_cnt; cycles 133 arch/m68k/coldfire/pit.c return cycles + PIT_CYCLES_PER_JIFFY - pcntr; cycles 104 arch/m68k/coldfire/sltimers.c u32 cycles, scnt; cycles 108 arch/m68k/coldfire/sltimers.c cycles = mcfslt_cnt; cycles 110 arch/m68k/coldfire/sltimers.c cycles += mcfslt_cycles_per_jiffy; cycles 116 arch/m68k/coldfire/sltimers.c return cycles + ((mcfslt_cycles_per_jiffy - 1) - scnt); cycles 96 arch/m68k/coldfire/timers.c u32 cycles; cycles 101 arch/m68k/coldfire/timers.c cycles = mcftmr_cnt; cycles 104 arch/m68k/coldfire/timers.c return cycles + tcn; cycles 415 arch/mips/pci/pci-octeon.c unsigned long cycles, pci_clock; cycles 418 arch/mips/pci/pci-octeon.c cycles = read_c0_cvmcount(); cycles 421 arch/mips/pci/pci-octeon.c cycles = read_c0_cvmcount() - cycles; cycles 423 arch/mips/pci/pci-octeon.c (cycles / (mips_hpt_frequency / 1000000)); cycles 11 arch/nios2/lib/delay.c void __delay(unsigned long cycles) cycles 15 arch/nios2/lib/delay.c while ((get_cycles() - start) < cycles) cycles 29 arch/openrisc/lib/delay.c void __delay(unsigned long cycles) cycles 33 arch/openrisc/lib/delay.c while ((get_cycles() - start) < cycles) cycles 50 arch/powerpc/perf/generic-compat-pmu.c GENERIC_EVENT_ATTR(cpu-cycles, PM_CYC); cycles 92 arch/powerpc/perf/hv-gpci-requests.h __count(0x10, 8, cycles) cycles 374 arch/powerpc/perf/power7-pmu.c GENERIC_EVENT_ATTR(cpu-cycles, PM_CYC); cycles 375 arch/powerpc/perf/power7-pmu.c GENERIC_EVENT_ATTR(stalled-cycles-frontend, PM_GCT_NOSLOT_CYC); cycles 376 arch/powerpc/perf/power7-pmu.c GENERIC_EVENT_ATTR(stalled-cycles-backend, PM_CMPLU_STALL); cycles 123 arch/powerpc/perf/power8-pmu.c GENERIC_EVENT_ATTR(cpu-cycles, PM_CYC); cycles 124 arch/powerpc/perf/power8-pmu.c GENERIC_EVENT_ATTR(stalled-cycles-frontend, PM_GCT_NOSLOT_CYC); cycles 125 arch/powerpc/perf/power8-pmu.c GENERIC_EVENT_ATTR(stalled-cycles-backend, PM_CMPLU_STALL); cycles 152 arch/powerpc/perf/power9-pmu.c GENERIC_EVENT_ATTR(cpu-cycles, PM_CYC); cycles 153 arch/powerpc/perf/power9-pmu.c GENERIC_EVENT_ATTR(stalled-cycles-frontend, PM_ICT_NOSLOT_CYC); cycles 154 arch/powerpc/perf/power9-pmu.c GENERIC_EVENT_ATTR(stalled-cycles-backend, PM_CMPLU_STALL); cycles 18 arch/riscv/include/asm/delay.h extern void __delay(unsigned long cycles); cycles 72 arch/riscv/lib/delay.c void __delay(unsigned long cycles) cycles 76 arch/riscv/lib/delay.c while ((unsigned long)(get_cycles() - t0) < cycles) cycles 155 arch/sparc/kernel/time_32.c u64 cycles; cycles 160 arch/sparc/kernel/time_32.c cycles = timer_cs_internal_counter; cycles 165 arch/sparc/kernel/time_32.c cycles *= sparc_config.cs_period; cycles 166 arch/sparc/kernel/time_32.c cycles += offset; cycles 168 arch/sparc/kernel/time_32.c return cycles; cycles 129 arch/sparc/vdso/vclock_gettime.c u64 cycles; cycles 131 arch/sparc/vdso/vclock_gettime.c cycles = vread_tick(); cycles 132 arch/sparc/vdso/vclock_gettime.c v = (cycles - vvar->clock.cycle_last) & vvar->clock.mask; cycles 139 arch/sparc/vdso/vclock_gettime.c u64 cycles; cycles 141 arch/sparc/vdso/vclock_gettime.c cycles = vread_tick_stick(); cycles 142 arch/sparc/vdso/vclock_gettime.c v = (cycles - vvar->clock.cycle_last) & vvar->clock.mask; cycles 1687 arch/x86/events/core.c EVENT_ATTR(cpu-cycles, CPU_CYCLES ); cycles 1693 arch/x86/events/core.c EVENT_ATTR(bus-cycles, BUS_CYCLES ); cycles 1694 arch/x86/events/core.c EVENT_ATTR(stalled-cycles-frontend, STALLED_CYCLES_FRONTEND ); cycles 1695 arch/x86/events/core.c EVENT_ATTR(stalled-cycles-backend, STALLED_CYCLES_BACKEND ); cycles 1696 arch/x86/events/core.c EVENT_ATTR(ref-cycles, REF_CPU_CYCLES ); cycles 4236 arch/x86/events/intel/core.c EVENT_ATTR_STR(cycles-t, cycles_t, "event=0x3c,in_tx=1"); cycles 4237 arch/x86/events/intel/core.c EVENT_ATTR_STR(cycles-ct, cycles_ct, "event=0x3c,in_tx=1,in_tx_cp=1"); cycles 560 arch/x86/events/intel/lbr.c cpuc->lbr_entries[i].cycles = 0; cycles 592 arch/x86/events/intel/lbr.c u16 cycles = 0; cycles 613 arch/x86/events/intel/lbr.c cycles = (info & LBR_INFO_CYCLES); cycles 620 arch/x86/events/intel/lbr.c cycles = ((to >> 48) & LBR_INFO_CYCLES); cycles 654 arch/x86/events/intel/lbr.c cpuc->lbr_entries[out].cycles = cycles; cycles 1110 arch/x86/events/intel/lbr.c e->cycles = info & LBR_INFO_CYCLES; cycles 287 arch/x86/include/asm/vdso/gettimeofday.h u64 vdso_calc_delta(u64 cycles, u64 last, u64 mask, u32 mult) cycles 289 arch/x86/include/asm/vdso/gettimeofday.h if (cycles > last) cycles 290 arch/x86/include/asm/vdso/gettimeofday.h return (cycles - last) * mult; cycles 1231 arch/x86/kernel/tsc.c .cycles = res}; cycles 1269 arch/x86/kernel/tsc.c .cycles = res}; cycles 40 arch/xtensa/include/asm/delay.h unsigned long cycles = (usecs * (ccount_freq >> 15)) >> 5; cycles 43 arch/xtensa/include/asm/delay.h while (((unsigned long)get_ccount()) - start < cycles) cycles 61 arch/xtensa/include/asm/delay.h unsigned long cycles = (nsec * (ccount_freq >> 15)) >> 15; cycles 62 arch/xtensa/include/asm/delay.h __delay(cycles); cycles 212 crypto/tcrypt.c unsigned long cycles = 0; cycles 239 crypto/tcrypt.c cycles += end - start; cycles 243 crypto/tcrypt.c (cycles + 4) / (8 * num_mb), blen); cycles 478 crypto/tcrypt.c unsigned long cycles = 0; cycles 507 crypto/tcrypt.c cycles += end - start; cycles 513 crypto/tcrypt.c (cycles + 4) / 8, blen); cycles 773 crypto/tcrypt.c unsigned long cycles = 0; cycles 800 crypto/tcrypt.c cycles += end - start; cycles 804 crypto/tcrypt.c (cycles + 4) / (8 * num_mb), blen); cycles 962 crypto/tcrypt.c unsigned long cycles = 0; cycles 984 crypto/tcrypt.c cycles += end - start; cycles 992 crypto/tcrypt.c cycles / 8, cycles / (8 * blen)); cycles 1000 crypto/tcrypt.c unsigned long cycles = 0; cycles 1041 crypto/tcrypt.c cycles += end - start; cycles 1049 crypto/tcrypt.c cycles / 8, cycles / (8 * blen)); cycles 1210 crypto/tcrypt.c unsigned long cycles = 0; cycles 1237 crypto/tcrypt.c cycles += end - start; cycles 1241 crypto/tcrypt.c (cycles + 4) / (8 * num_mb), blen); cycles 1445 crypto/tcrypt.c unsigned long cycles = 0; cycles 1478 crypto/tcrypt.c cycles += end - start; cycles 1484 crypto/tcrypt.c (cycles + 4) / 8, blen); cycles 1210 drivers/char/random.c unsigned cycles; cycles 1216 drivers/char/random.c sample.cycles = random_get_entropy(); cycles 1314 drivers/char/random.c cycles_t cycles = random_get_entropy(); cycles 1320 drivers/char/random.c if (cycles == 0) cycles 1321 drivers/char/random.c cycles = get_reg(fast_pool, regs); cycles 1322 drivers/char/random.c c_high = (sizeof(cycles) > 4) ? cycles >> 32 : 0; cycles 1324 drivers/char/random.c fast_pool->pool[0] ^= cycles ^ j_high ^ irq; cycles 1332 drivers/char/random.c add_interrupt_bench(cycles); cycles 61 drivers/clk/ti/clkctrl.c u32 cycles; cycles 107 drivers/clk/ti/clkctrl.c if (time->cycles++ < timeout) { cycles 243 drivers/clocksource/arc_timer.c static void arc_timer_event_setup(unsigned int cycles) cycles 245 drivers/clocksource/arc_timer.c write_aux_reg(ARC_REG_TIMER0_LIMIT, cycles); cycles 261 drivers/clocksource/exynos_mct.c static void exynos4_mct_comp0_start(bool periodic, unsigned long cycles) cycles 270 drivers/clocksource/exynos_mct.c exynos4_mct_write(cycles, EXYNOS4_MCT_G_COMP0_ADD_INCR); cycles 273 drivers/clocksource/exynos_mct.c comp_cycle = exynos4_read_count_64() + cycles; cycles 283 drivers/clocksource/exynos_mct.c static int exynos4_comp_set_next_event(unsigned long cycles, cycles 286 drivers/clocksource/exynos_mct.c exynos4_mct_comp0_start(false, cycles); cycles 365 drivers/clocksource/exynos_mct.c static void exynos4_mct_tick_start(unsigned long cycles, cycles 372 drivers/clocksource/exynos_mct.c tmp = (1 << 31) | cycles; /* MCT_L_UPDATE_ICNTB */ cycles 393 drivers/clocksource/exynos_mct.c static int exynos4_tick_set_next_event(unsigned long cycles, cycles 399 drivers/clocksource/exynos_mct.c exynos4_mct_tick_start(cycles, mevt); cycles 185 drivers/clocksource/samsung_pwm_timer.c static int samsung_set_next_event(unsigned long cycles, cycles 198 drivers/clocksource/samsung_pwm_timer.c if (!cycles) cycles 199 drivers/clocksource/samsung_pwm_timer.c cycles = 1; cycles 201 drivers/clocksource/samsung_pwm_timer.c samsung_time_setup(pwm.event_id, cycles); cycles 89 drivers/clocksource/timer-atlas7.c u64 cycles; cycles 94 drivers/clocksource/timer-atlas7.c cycles = readl_relaxed(sirfsoc_timer_base + SIRFSOC_TIMER_64COUNTER_RLATCHED_HI); cycles 95 drivers/clocksource/timer-atlas7.c cycles = (cycles << 32) | readl_relaxed(sirfsoc_timer_base + SIRFSOC_TIMER_64COUNTER_RLATCHED_LO); cycles 97 drivers/clocksource/timer-atlas7.c return cycles; cycles 108 drivers/clocksource/timer-cadence-ttc.c unsigned long cycles) cycles 117 drivers/clocksource/timer-cadence-ttc.c writel_relaxed(cycles, timer->base_addr + TTC_INTR_VAL_OFFSET); cycles 175 drivers/clocksource/timer-cadence-ttc.c static int ttc_set_next_event(unsigned long cycles, cycles 181 drivers/clocksource/timer-cadence-ttc.c ttc_set_interval(timer, cycles); cycles 139 drivers/clocksource/timer-davinci.c davinci_clockevent_set_next_event_std(unsigned long cycles, cycles 147 drivers/clocksource/timer-davinci.c davinci_clockevent_write(clockevent, DAVINCI_TIMER_REG_PRD12, cycles); cycles 155 drivers/clocksource/timer-davinci.c davinci_clockevent_set_next_event_cmp(unsigned long cycles, cycles 164 drivers/clocksource/timer-davinci.c clockevent->cmp_off, curr_time + cycles); cycles 136 drivers/clocksource/timer-fttmr010.c static int fttmr010_timer_set_next_event(unsigned long cycles, cycles 152 drivers/clocksource/timer-fttmr010.c writel(cycles, fttmr010->base + TIMER1_LOAD); cycles 156 drivers/clocksource/timer-fttmr010.c writel(cr + cycles, fttmr010->base + TIMER1_MATCH1); cycles 107 drivers/clocksource/timer-ixp4xx.c static int ixp4xx_set_next_event(unsigned long cycles, cycles 116 drivers/clocksource/timer-ixp4xx.c __raw_writel((cycles & ~IXP4XX_OST_RELOAD_MASK) | val, cycles 125 drivers/clocksource/timer-keystone.c static int keystone_set_next_event(unsigned long cycles, cycles 128 drivers/clocksource/timer-keystone.c return keystone_timer_config(cycles, TCR_ENAMODE_ONESHOT_MASK); cycles 76 drivers/clocksource/timer-prima2.c u64 cycles; cycles 81 drivers/clocksource/timer-prima2.c cycles = readl_relaxed(sirfsoc_timer_base + SIRFSOC_TIMER_LATCHED_HI); cycles 82 drivers/clocksource/timer-prima2.c cycles = (cycles << 32) | cycles 85 drivers/clocksource/timer-prima2.c return cycles; cycles 50 drivers/clocksource/timer-qcom.c static int msm_timer_set_next_event(unsigned long cycles, cycles 59 drivers/clocksource/timer-qcom.c writel_relaxed(cycles, event_base + TIMER_MATCH_VAL); cycles 35 drivers/clocksource/timer-rda.c static int rda_ostimer_start(void __iomem *base, bool periodic, u64 cycles) cycles 39 drivers/clocksource/timer-rda.c load_l = (u32)cycles; cycles 40 drivers/clocksource/timer-rda.c ctrl = ((cycles >> 32) & 0xffffff); cycles 65 drivers/clocksource/timer-rockchip.c static void rk_timer_update_counter(unsigned long cycles, cycles 68 drivers/clocksource/timer-rockchip.c writel_relaxed(cycles, timer->base + TIMER_LOAD_COUNT0); cycles 77 drivers/clocksource/timer-rockchip.c static inline int rk_timer_set_next_event(unsigned long cycles, cycles 83 drivers/clocksource/timer-rockchip.c rk_timer_update_counter(cycles, timer); cycles 60 drivers/clocksource/timer-sprd.c static void sprd_timer_update_counter(void __iomem *base, unsigned long cycles) cycles 62 drivers/clocksource/timer-sprd.c writel_relaxed(cycles & TIMER_VALUE_LO_MASK, base + TIMER_LOAD_LO); cycles 79 drivers/clocksource/timer-sprd.c static int sprd_timer_set_next_event(unsigned long cycles, cycles 85 drivers/clocksource/timer-sprd.c sprd_timer_update_counter(timer_of_base(to), cycles); cycles 54 drivers/clocksource/timer-tegra.c static int tegra_timer_set_next_event(unsigned long cycles, cycles 68 drivers/clocksource/timer-tegra.c writel_relaxed(TIMER_PTV_EN | (cycles - 1), reg_base + TIMER_PTV); cycles 276 drivers/clocksource/timer-u300.c static int u300_set_next_event(unsigned long cycles, cycles 290 drivers/clocksource/timer-u300.c writel(cycles, u300_timer_base + U300_TIMER_APP_GPT1TC); cycles 61 drivers/clocksource/timer-vt8500.c static int vt8500_timer_set_next_event(unsigned long cycles, cycles 65 drivers/clocksource/timer-vt8500.c u64 alarm = clocksource.read(&clocksource) + cycles; cycles 164 drivers/crypto/qat/qat_common/qat_hal.c unsigned char ae, unsigned int cycles, cycles 174 drivers/crypto/qat/qat_common/qat_hal.c while ((int)cycles > elapsed_cycles && times--) { cycles 465 drivers/dma/moxart-dma.c unsigned int completed_cycles, cycles; cycles 468 drivers/dma/moxart-dma.c cycles = readl(ch->base + REG_OFF_CYCLES); cycles 469 drivers/dma/moxart-dma.c completed_cycles = (ch->desc->dma_cycles - cycles); cycles 731 drivers/firewire/core-transaction.c unsigned int cycles; cycles 734 drivers/firewire/core-transaction.c cycles = card->split_timeout_cycles; cycles 735 drivers/firewire/core-transaction.c cycles += request_timestamp & 0x1fff; cycles 738 drivers/firewire/core-transaction.c timestamp += (cycles / 8000) << 13; cycles 739 drivers/firewire/core-transaction.c timestamp |= cycles % 8000; cycles 1073 drivers/firewire/core-transaction.c unsigned int cycles; cycles 1075 drivers/firewire/core-transaction.c cycles = card->split_timeout_hi * 8000 + (card->split_timeout_lo >> 19); cycles 1078 drivers/firewire/core-transaction.c cycles = clamp(cycles, 800u, 3u * 8000u); cycles 1080 drivers/firewire/core-transaction.c card->split_timeout_cycles = cycles; cycles 1081 drivers/firewire/core-transaction.c card->split_timeout_jiffies = DIV_ROUND_UP(cycles * HZ, 8000); cycles 756 drivers/gpio/gpio-aspeed.c u32 *cycles) cycles 773 drivers/gpio/gpio-aspeed.c *cycles = n + (!!r); cycles 873 drivers/gpio/gpio-aspeed.c u32 cycles; cycles 875 drivers/gpio/gpio-aspeed.c cycles = ioread32(gpio->base + debounce_timers[i]); cycles 876 drivers/gpio/gpio-aspeed.c if (requested_cycles == cycles) cycles 129 drivers/gpu/drm/nouveau/nvkm/subdev/therm/fan.c u32 cycles, cur, prev; cycles 145 drivers/gpu/drm/nouveau/nvkm/subdev/therm/fan.c cycles = 0; cycles 154 drivers/gpu/drm/nouveau/nvkm/subdev/therm/fan.c cycles++; cycles 157 drivers/gpu/drm/nouveau/nvkm/subdev/therm/fan.c } while (cycles < 5 && nvkm_timer_read(tmr) - start < 250000000); cycles 160 drivers/gpu/drm/nouveau/nvkm/subdev/therm/fan.c if (cycles == 5) { cycles 217 drivers/gpu/drm/v3d/v3d_debugfs.c uint32_t cycles; cycles 241 drivers/gpu/drm/v3d/v3d_debugfs.c cycles = V3D_CORE_READ(core, V3D_PCTR_0_PCTR0); cycles 244 drivers/gpu/drm/v3d/v3d_debugfs.c cycles, cycles 245 drivers/gpu/drm/v3d/v3d_debugfs.c cycles / (measure_ms * 1000), cycles 246 drivers/gpu/drm/v3d/v3d_debugfs.c (cycles / (measure_ms * 100)) % 10); cycles 1597 drivers/infiniband/hw/hfi1/hfi.h u32 cycles; cycles 1607 drivers/infiniband/hw/hfi1/hfi.h cycles = len * 8; /* bits */ cycles 1608 drivers/infiniband/hw/hfi1/hfi.h cycles *= 805; cycles 1609 drivers/infiniband/hw/hfi1/hfi.h cycles /= rate; cycles 1611 drivers/infiniband/hw/hfi1/hfi.h return cycles; cycles 373 drivers/input/keyboard/pmic8xxx-keypad.c int bits, rc, cycles; cycles 438 drivers/input/keyboard/pmic8xxx-keypad.c cycles = (row_hold_ns * KEYP_CLOCK_FREQ) / NSEC_PER_SEC; cycles 440 drivers/input/keyboard/pmic8xxx-keypad.c scan_val |= (cycles << KEYP_SCAN_ROW_HOLD_SHIFT); cycles 136 drivers/media/common/b2c2/flexcop-common.h flexcop_dma_index_t dma_idx, u8 cycles); cycles 387 drivers/media/i2c/ov2685.c static inline u32 ov2685_cal_delay(u32 cycles) cycles 389 drivers/media/i2c/ov2685.c return DIV_ROUND_UP(cycles, OV2685_XVCLK_FREQ / 1000 / 1000); cycles 430 drivers/media/i2c/tc358743_regs.h #define SET_FREQ_RANGE_MODE_CYCLES(cycles) (((cycles) - 1) & \ cycles 162 drivers/media/pci/b2c2/flexcop-dma.c flexcop_dma_index_t dma_idx, u8 cycles) cycles 170 drivers/media/pci/b2c2/flexcop-dma.c v.dma_0x4_write.dmatimer = cycles; cycles 2210 drivers/media/platform/rcar_fdp1.c u32 cycles; cycles 2213 drivers/media/platform/rcar_fdp1.c cycles = fdp1_read(fdp1, FD1_CTL_VCYCLE_STAT); cycles 2228 drivers/media/platform/rcar_fdp1.c cycles, cycles/(fdp1->clk_rate/1000)); cycles 290 drivers/media/rc/iguanair.c uint32_t cycles, fours, sevens; cycles 294 drivers/media/rc/iguanair.c cycles = DIV_ROUND_CLOSEST(24000000, carrier * 2) - cycles 302 drivers/media/rc/iguanair.c sevens = (4 - cycles) & 3; cycles 303 drivers/media/rc/iguanair.c fours = (cycles - sevens * 7) / 4; cycles 176 drivers/media/usb/gspca/stv06xx/stv06xx_hdcs.c int cycles, err; cycles 179 drivers/media/usb/gspca/stv06xx/stv06xx_hdcs.c cycles = val * HDCS_CLK_FREQ_MHZ * 257; cycles 187 drivers/media/usb/gspca/stv06xx/stv06xx_hdcs.c rowexp = cycles / rp; cycles 190 drivers/media/usb/gspca/stv06xx/stv06xx_hdcs.c cycles -= rowexp * rp; cycles 195 drivers/media/usb/gspca/stv06xx/stv06xx_hdcs.c srowexp = hdcs->w - (cycles + hdcs->exp.er + 13) / ct; cycles 201 drivers/media/usb/gspca/stv06xx/stv06xx_hdcs.c srowexp = cp - hdcs->exp.er - 6 - cycles; cycles 160 drivers/memory/jz4780-nemc.c uint32_t smcr, val, cycles; cycles 206 drivers/memory/jz4780-nemc.c cycles = jz4780_nemc_ns_to_cycles(nemc, val); cycles 207 drivers/memory/jz4780-nemc.c if (cycles > nemc->soc_info->tas_tah_cycles_max) { cycles 209 drivers/memory/jz4780-nemc.c val, cycles); cycles 213 drivers/memory/jz4780-nemc.c smcr |= cycles << NEMC_SMCR_TAS_SHIFT; cycles 218 drivers/memory/jz4780-nemc.c cycles = jz4780_nemc_ns_to_cycles(nemc, val); cycles 219 drivers/memory/jz4780-nemc.c if (cycles > nemc->soc_info->tas_tah_cycles_max) { cycles 221 drivers/memory/jz4780-nemc.c val, cycles); cycles 225 drivers/memory/jz4780-nemc.c smcr |= cycles << NEMC_SMCR_TAH_SHIFT; cycles 230 drivers/memory/jz4780-nemc.c cycles = jz4780_nemc_ns_to_cycles(nemc, val); cycles 231 drivers/memory/jz4780-nemc.c if (cycles > 31) { cycles 233 drivers/memory/jz4780-nemc.c val, cycles); cycles 237 drivers/memory/jz4780-nemc.c smcr |= convert_tBP_tAW[cycles] << NEMC_SMCR_TBP_SHIFT; cycles 242 drivers/memory/jz4780-nemc.c cycles = jz4780_nemc_ns_to_cycles(nemc, val); cycles 243 drivers/memory/jz4780-nemc.c if (cycles > 31) { cycles 245 drivers/memory/jz4780-nemc.c val, cycles); cycles 249 drivers/memory/jz4780-nemc.c smcr |= convert_tBP_tAW[cycles] << NEMC_SMCR_TAW_SHIFT; cycles 254 drivers/memory/jz4780-nemc.c cycles = jz4780_nemc_ns_to_cycles(nemc, val); cycles 255 drivers/memory/jz4780-nemc.c if (cycles > 63) { cycles 257 drivers/memory/jz4780-nemc.c val, cycles); cycles 261 drivers/memory/jz4780-nemc.c smcr |= cycles << NEMC_SMCR_TSTRV_SHIFT; cycles 64 drivers/memory/pl172.c int cycles; cycles 68 drivers/memory/pl172.c cycles = DIV_ROUND_UP(val * pl172->rate, NSEC_PER_MSEC) - start; cycles 69 drivers/memory/pl172.c if (cycles < 0) { cycles 70 drivers/memory/pl172.c cycles = 0; cycles 71 drivers/memory/pl172.c } else if (cycles > max) { cycles 76 drivers/memory/pl172.c writel(cycles, pl172->base + reg_offset); cycles 338 drivers/misc/altera-stapl/altera-jtag.c s32 cycles, cycles 356 drivers/misc/altera-stapl/altera-jtag.c for (count = 0L; count < cycles; count++) cycles 83 drivers/misc/altera-stapl/altera-jtag.h int altera_wait_cycles(struct altera_state *astate, s32 cycles, cycles 219 drivers/mtd/nand/raw/diskonchip.c static void DoC_Delay(struct doc_priv *doc, unsigned short cycles) cycles 224 drivers/mtd/nand/raw/diskonchip.c for (i = 0; i < cycles; i++) { cycles 1706 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h void t4_pmtx_get_stats(struct adapter *adap, u32 cnt[], u64 cycles[]); cycles 1707 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h void t4_pmrx_get_stats(struct adapter *adap, u32 cnt[], u64 cycles[]); cycles 6060 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c void t4_pmtx_get_stats(struct adapter *adap, u32 cnt[], u64 cycles[]) cycles 6069 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c cycles[i] = t4_read_reg64(adap, PM_TX_STAT_LSB_A); cycles 6074 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c cycles[i] = (((u64)data[0] << 32) | data[1]); cycles 6087 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c void t4_pmrx_get_stats(struct adapter *adap, u32 cnt[], u64 cycles[]) cycles 6096 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c cycles[i] = t4_read_reg64(adap, PM_RX_STAT_LSB_A); cycles 6101 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c cycles[i] = (((u64)data[0] << 32) | data[1]); cycles 180 drivers/net/ethernet/intel/e1000e/ptp.c u64 cycles, ns; cycles 185 drivers/net/ethernet/intel/e1000e/ptp.c cycles = e1000e_read_systim(adapter, sts); cycles 186 drivers/net/ethernet/intel/e1000e/ptp.c ns = timecounter_cyc2time(&adapter->tc, cycles); cycles 2186 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c static u32 mvpp2_cycles_to_usec(u32 cycles, unsigned long clk_hz) cycles 2188 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c u64 tmp = (u64)cycles * USEC_PER_SEC; cycles 44 drivers/net/ethernet/mellanox/mlx4/en_clock.c container_of(tc, struct mlx4_en_dev, cycles); cycles 141 drivers/net/ethernet/mellanox/mlx4/en_clock.c mdev->cycles.mult = neg_adj ? mult - diff : mult + diff; cycles 210 drivers/net/ethernet/mellanox/mlx4/en_clock.c timecounter_init(&mdev->clock, &mdev->cycles, ns); cycles 277 drivers/net/ethernet/mellanox/mlx4/en_clock.c memset(&mdev->cycles, 0, sizeof(mdev->cycles)); cycles 278 drivers/net/ethernet/mellanox/mlx4/en_clock.c mdev->cycles.read = mlx4_en_read_clock; cycles 279 drivers/net/ethernet/mellanox/mlx4/en_clock.c mdev->cycles.mask = CLOCKSOURCE_MASK(48); cycles 280 drivers/net/ethernet/mellanox/mlx4/en_clock.c mdev->cycles.shift = freq_to_shift(dev->caps.hca_core_clock); cycles 281 drivers/net/ethernet/mellanox/mlx4/en_clock.c mdev->cycles.mult = cycles 282 drivers/net/ethernet/mellanox/mlx4/en_clock.c clocksource_khz2mult(1000 * dev->caps.hca_core_clock, mdev->cycles.shift); cycles 283 drivers/net/ethernet/mellanox/mlx4/en_clock.c mdev->nominal_c_mult = mdev->cycles.mult; cycles 286 drivers/net/ethernet/mellanox/mlx4/en_clock.c timecounter_init(&mdev->clock, &mdev->cycles, cycles 1909 drivers/net/ethernet/mellanox/mlx4/main.c u64 cycles; cycles 1921 drivers/net/ethernet/mellanox/mlx4/main.c cycles = (u64) clockhi << 32 | (u64) clocklo; cycles 1923 drivers/net/ethernet/mellanox/mlx4/main.c return cycles; cycles 432 drivers/net/ethernet/mellanox/mlx4/mlx4_en.h struct cyclecounter cycles; cycles 71 drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c struct mlx5_clock *clock = container_of(cc, struct mlx5_clock, cycles); cycles 91 drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c clock_info->cycles = clock->tc.cycle_last; cycles 92 drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c clock_info->mult = clock->cycles.mult; cycles 152 drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c timecounter_init(&clock->tc, &clock->cycles, ns); cycles 167 drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c u64 cycles, ns; cycles 170 drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c cycles = mlx5_read_internal_timer(mdev, sts); cycles 171 drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c ns = timecounter_cyc2time(&clock->tc, cycles); cycles 213 drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c clock->cycles.mult = neg_adj ? clock->nominal_c_mult - diff : cycles 335 drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c cycles_delta = div64_u64(nsec_delta << clock->cycles.shift, cycles 336 drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c clock->cycles.mult); cycles 501 drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c cycles_delta = div64_u64(nsec_delta << clock->cycles.shift, cycles 502 drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c clock->cycles.mult); cycles 529 drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c clock->cycles.read = read_internal_timer; cycles 530 drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c clock->cycles.shift = MLX5_CYCLES_SHIFT; cycles 531 drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c clock->cycles.mult = clocksource_khz2mult(dev_freq, cycles 532 drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c clock->cycles.shift); cycles 533 drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c clock->nominal_c_mult = clock->cycles.mult; cycles 534 drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c clock->cycles.mask = CLOCKSOURCE_MASK(41); cycles 537 drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c timecounter_init(&clock->tc, &clock->cycles, cycles 547 drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c overflow_cycles = div64_u64(~0ULL >> 1, clock->cycles.mult); cycles 548 drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c overflow_cycles = min(overflow_cycles, div_u64(clock->cycles.mask, 3)); cycles 550 drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c ns = cyclecounter_cyc2ns(&clock->cycles, overflow_cycles, cycles 559 drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c mdev->clock_info->cycles = clock->tc.cycle_last; cycles 560 drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c mdev->clock_info->mask = clock->cycles.mask; cycles 562 drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c mdev->clock_info->shift = clock->cycles.shift; cycles 64 drivers/net/ethernet/mellanox/mlxsw/spectrum_ptp.c struct cyclecounter cycles; cycles 98 drivers/net/ethernet/mellanox/mlxsw/spectrum_ptp.c container_of(cc, struct mlxsw_sp_ptp_clock, cycles); cycles 116 drivers/net/ethernet/mellanox/mlxsw/spectrum_ptp.c u64 cycles = (u64) nsec; cycles 118 drivers/net/ethernet/mellanox/mlxsw/spectrum_ptp.c cycles <<= tc->cc->shift; cycles 119 drivers/net/ethernet/mellanox/mlxsw/spectrum_ptp.c cycles = div_u64(cycles, tc->cc->mult); cycles 121 drivers/net/ethernet/mellanox/mlxsw/spectrum_ptp.c return cycles; cycles 128 drivers/net/ethernet/mellanox/mlxsw/spectrum_ptp.c u64 next_sec, next_sec_in_nsec, cycles; cycles 137 drivers/net/ethernet/mellanox/mlxsw/spectrum_ptp.c cycles = mlxsw_sp1_ptp_ns2cycles(&clock->tc, next_sec_in_nsec); cycles 140 drivers/net/ethernet/mellanox/mlxsw/spectrum_ptp.c mlxsw_reg_mtpps_vpin_pack(mtpps_pl, cycles); cycles 173 drivers/net/ethernet/mellanox/mlxsw/spectrum_ptp.c clock->cycles.mult = neg_adj ? clock->nominal_c_mult - diff : cycles 200 drivers/net/ethernet/mellanox/mlxsw/spectrum_ptp.c u64 cycles, nsec; cycles 203 drivers/net/ethernet/mellanox/mlxsw/spectrum_ptp.c cycles = __mlxsw_sp1_ptp_read_frc(clock, sts); cycles 204 drivers/net/ethernet/mellanox/mlxsw/spectrum_ptp.c nsec = timecounter_cyc2time(&clock->tc, cycles); cycles 220 drivers/net/ethernet/mellanox/mlxsw/spectrum_ptp.c timecounter_init(&clock->tc, &clock->cycles, nsec); cycles 262 drivers/net/ethernet/mellanox/mlxsw/spectrum_ptp.c clock->cycles.read = mlxsw_sp1_ptp_read_frc; cycles 263 drivers/net/ethernet/mellanox/mlxsw/spectrum_ptp.c clock->cycles.shift = MLXSW_SP1_PTP_CLOCK_CYCLES_SHIFT; cycles 264 drivers/net/ethernet/mellanox/mlxsw/spectrum_ptp.c clock->cycles.mult = clocksource_khz2mult(MLXSW_SP1_PTP_CLOCK_FREQ_KHZ, cycles 265 drivers/net/ethernet/mellanox/mlxsw/spectrum_ptp.c clock->cycles.shift); cycles 266 drivers/net/ethernet/mellanox/mlxsw/spectrum_ptp.c clock->nominal_c_mult = clock->cycles.mult; cycles 267 drivers/net/ethernet/mellanox/mlxsw/spectrum_ptp.c clock->cycles.mask = CLOCKSOURCE_MASK(MLXSW_SP1_PTP_CLOCK_MASK); cycles 270 drivers/net/ethernet/mellanox/mlxsw/spectrum_ptp.c timecounter_init(&clock->tc, &clock->cycles, cycles 280 drivers/net/ethernet/mellanox/mlxsw/spectrum_ptp.c overflow_cycles = div64_u64(~0ULL >> 1, clock->cycles.mult); cycles 281 drivers/net/ethernet/mellanox/mlxsw/spectrum_ptp.c overflow_cycles = min(overflow_cycles, div_u64(clock->cycles.mask, 3)); cycles 283 drivers/net/ethernet/mellanox/mlxsw/spectrum_ptp.c nsec = cyclecounter_cyc2ns(&clock->cycles, overflow_cycles, 0, &frac); cycles 368 drivers/net/ethernet/mellanox/mlxsw/spectrum_ptp.c int cycles = MLXSW_SP1_PTP_HT_GC_TIMEOUT / MLXSW_SP1_PTP_HT_GC_INTERVAL; cycles 380 drivers/net/ethernet/mellanox/mlxsw/spectrum_ptp.c unmatched->gc_cycle = mlxsw_sp->ptp_state->gc_cycle + cycles; cycles 441 drivers/net/ethernet/xscale/ixp4xx_eth.c int cycles = 0; cycles 457 drivers/net/ethernet/xscale/ixp4xx_eth.c while ((cycles < MAX_MDIO_RETRIES) && cycles 460 drivers/net/ethernet/xscale/ixp4xx_eth.c cycles++; cycles 463 drivers/net/ethernet/xscale/ixp4xx_eth.c if (cycles == MAX_MDIO_RETRIES) { cycles 471 drivers/net/ethernet/xscale/ixp4xx_eth.c phy_id, write ? "write" : "read", cycles); cycles 46 drivers/net/wireless/ath/ath.h u32 cycles; cycles 728 drivers/net/wireless/ath/ath5k/debug.c _struct.cycles > 0 ? \ cycles 729 drivers/net/wireless/ath/ath5k/debug.c _struct._field * 100 / _struct.cycles : 0 cycles 742 drivers/net/wireless/ath/ath5k/debug.c as->last_cc.cycles); cycles 660 drivers/net/wireless/ath/ath5k/mac80211-ops.c if (cc->cycles > 0) { cycles 661 drivers/net/wireless/ath/ath5k/mac80211-ops.c ah->survey.time += cc->cycles / div; cycles 2151 drivers/net/wireless/ath/ath9k/ar9003_phy.c #define PCT(_field) (common->cc_survey._field * 100 / common->cc_survey.cycles) cycles 2152 drivers/net/wireless/ath/ath9k/ar9003_phy.c if (common->cc_survey.cycles) cycles 522 drivers/net/wireless/ath/ath9k/link.c if (cc->cycles > 0) { cycles 527 drivers/net/wireless/ath/ath9k/link.c survey->time += cc->cycles / div; cycles 533 drivers/net/wireless/ath/ath9k/link.c if (cc->cycles < div) cycles 536 drivers/net/wireless/ath/ath9k/link.c if (cc->cycles > 0) cycles 537 drivers/net/wireless/ath/ath9k/link.c ret = cc->rx_busy * 100 / cc->cycles; cycles 144 drivers/net/wireless/ath/hw.c u32 cycles, busy, rx, tx; cycles 151 drivers/net/wireless/ath/hw.c cycles = REG_READ(ah, AR_CCCNT); cycles 166 drivers/net/wireless/ath/hw.c common->cc_ani.cycles += cycles; cycles 171 drivers/net/wireless/ath/hw.c common->cc_survey.cycles += cycles; cycles 183 drivers/net/wireless/ath/hw.c listen_time = (cc->cycles - cc->rx_frame - cc->tx_frame) / cycles 329 drivers/net/wireless/ti/wl12xx/scan.c cfg->cycles = 0; cycles 75 drivers/net/wireless/ti/wl12xx/scan.h u8 cycles; /* maximum number of scan cycles */ cycles 62 drivers/pcmcia/yenta_socket.c #define to_ns(cycles) ((cycles)*120) cycles 253 drivers/perf/arm-cci.c CCI400_CYCLE_EVENT_EXT_ATTR_ENTRY(cycles, 0xff), cycles 300 drivers/perf/arm-cci.c CCI400_CYCLE_EVENT_EXT_ATTR_ENTRY(cycles, 0xff), cycles 441 drivers/perf/arm-ccn.c CCN_EVENT_CYCLES(cycles), cycles 186 drivers/perf/arm_dsu_pmu.c DSU_EVENT_ATTR(cycles, 0x11), cycles 513 drivers/perf/arm_smmuv3_pmu.c SMMU_EVENT_ATTR(cycles, 0); cycles 119 drivers/perf/fsl_imx8_ddr_perf.c IMX8_DDR_PMU_EVENT_ATTR(cycles, EVENT_CYCLES_ID), cycles 137 drivers/perf/fsl_imx8_ddr_perf.c IMX8_DDR_PMU_EVENT_ATTR(read-cycles, 0x2a), cycles 138 drivers/perf/fsl_imx8_ddr_perf.c IMX8_DDR_PMU_EVENT_ATTR(write-cycles, 0x2b), cycles 731 drivers/perf/qcom_l2_pmu.c L2CACHE_EVENT_ATTR(cycles, L2_EVENT_CYCLES), cycles 656 drivers/perf/qcom_l3_pmu.c L3CACHE_EVENT_ATTR(cycles, L3_EVENT_CYCLES), cycles 458 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(pa-barrier-cycles, 0x1d), cycles 75 drivers/ptp/ptp_kvm.c system_counter->cycles = ret; cycles 117 drivers/pwm/pwm-atmel.c unsigned long long cycles = state->period; cycles 120 drivers/pwm/pwm-atmel.c cycles *= clk_get_rate(atmel_pwm->clk); cycles 121 drivers/pwm/pwm-atmel.c do_div(cycles, NSEC_PER_SEC); cycles 123 drivers/pwm/pwm-atmel.c for (*pres = 0; cycles > atmel_pwm->data->cfg.max_period; cycles >>= 1) cycles 131 drivers/pwm/pwm-atmel.c *cprd = cycles; cycles 139 drivers/pwm/pwm-atmel.c unsigned long long cycles = state->duty_cycle; cycles 141 drivers/pwm/pwm-atmel.c cycles *= cprd; cycles 142 drivers/pwm/pwm-atmel.c do_div(cycles, state->period); cycles 143 drivers/pwm/pwm-atmel.c *cdty = cprd - cycles; cycles 96 drivers/pwm/pwm-berlin.c u64 cycles; cycles 98 drivers/pwm/pwm-berlin.c cycles = clk_get_rate(pwm->clk); cycles 99 drivers/pwm/pwm-berlin.c cycles *= period_ns; cycles 100 drivers/pwm/pwm-berlin.c do_div(cycles, NSEC_PER_SEC); cycles 102 drivers/pwm/pwm-berlin.c if (cycles > BERLIN_PWM_MAX_TCNT) { cycles 104 drivers/pwm/pwm-berlin.c cycles >>= 12; // Prescaled by 4096 cycles 106 drivers/pwm/pwm-berlin.c if (cycles > BERLIN_PWM_MAX_TCNT) cycles 110 drivers/pwm/pwm-berlin.c period = cycles; cycles 111 drivers/pwm/pwm-berlin.c cycles *= duty_ns; cycles 112 drivers/pwm/pwm-berlin.c do_div(cycles, period_ns); cycles 113 drivers/pwm/pwm-berlin.c duty = cycles; cycles 34 drivers/siox/siox-bus-gpio.c size_t cycles = max(setbuf_len, getbuf_len); cycles 44 drivers/siox/siox-bus-gpio.c for (i = 0; i < cycles; ++i) { cycles 48 drivers/siox/siox-bus-gpio.c if (i >= cycles - setbuf_len) cycles 49 drivers/siox/siox-bus-gpio.c set = setbuf[i - (cycles - setbuf_len)]; cycles 408 drivers/soc/ixp4xx/ixp4xx-npe.c int cycles = 0; cycles 427 drivers/soc/ixp4xx/ixp4xx-npe.c while ((cycles < MAX_RETRIES) && cycles 430 drivers/soc/ixp4xx/ixp4xx-npe.c cycles++; cycles 433 drivers/soc/ixp4xx/ixp4xx-npe.c if (cycles == MAX_RETRIES) { cycles 439 drivers/soc/ixp4xx/ixp4xx-npe.c debug_msg(npe, "Sending a message took %i cycles\n", cycles); cycles 447 drivers/soc/ixp4xx/ixp4xx-npe.c int cycles = 0, cnt = 0; cycles 451 drivers/soc/ixp4xx/ixp4xx-npe.c while (cycles < MAX_RETRIES) { cycles 458 drivers/soc/ixp4xx/ixp4xx-npe.c cycles++; cycles 471 drivers/soc/ixp4xx/ixp4xx-npe.c if (cycles == MAX_RETRIES) { cycles 477 drivers/soc/ixp4xx/ixp4xx-npe.c debug_msg(npe, "Receiving a message took %i cycles\n", cycles); cycles 488 drivers/staging/media/imx/imx-media-csi.c passthrough_cycles = incc->cycles; cycles 730 drivers/staging/media/imx/imx-media-csi.c if (is_parallel_bus(&priv->upstream_ep) && incc->cycles) { cycles 731 drivers/staging/media/imx/imx-media-csi.c if_fmt.width *= incc->cycles; cycles 732 drivers/staging/media/imx/imx-media-csi.c crop.width *= incc->cycles; cycles 77 drivers/staging/media/imx/imx-media-utils.c .cycles = 2, cycles 75 drivers/staging/media/imx/imx-media.h int cycles; cycles 270 drivers/tty/serial/serial-tegra.c unsigned int cycles) cycles 273 drivers/tty/serial/serial-tegra.c udelay(DIV_ROUND_UP(cycles * 1000000, tup->current_baud * 16)); cycles 74 drivers/watchdog/renesas_wdt.c static void rwdt_wait_cycles(struct rwdt_priv *priv, unsigned int cycles) cycles 78 drivers/watchdog/renesas_wdt.c delay = DIV_ROUND_UP(cycles * 1000000, priv->clk_rate); cycles 184 include/linux/clocksource.h static inline s64 clocksource_cyc2ns(u64 cycles, u32 mult, u32 shift) cycles 186 include/linux/clocksource.h return ((u64) cycles * mult) >> shift; cycles 654 include/linux/mlx5/driver.h struct cyclecounter cycles; cycles 634 include/linux/mtd/rawnand.h #define NAND_OP_ADDR(ncycles, cycles, ns) \ cycles 639 include/linux/mtd/rawnand.h .addrs = cycles, \ cycles 71 include/linux/timecounter.h u64 cycles, u64 mask, u64 *frac) cycles 73 include/linux/timecounter.h u64 ns = (u64) cycles; cycles 235 include/linux/timekeeping.h u64 cycles; cycles 263 include/linux/timekeeping.h u64 cycles; cycles 1177 include/uapi/linux/perf_event.h cycles:16, /* cycle count to last branch */ cycles 444 include/uapi/rdma/mlx5-abi.h __aligned_u64 cycles; cycles 277 kernel/locking/test-ww_mutex.c struct test_cycle *cycles; cycles 281 kernel/locking/test-ww_mutex.c cycles = kmalloc_array(nthreads, sizeof(*cycles), GFP_KERNEL); cycles 282 kernel/locking/test-ww_mutex.c if (!cycles) cycles 286 kernel/locking/test-ww_mutex.c struct test_cycle *cycle = &cycles[n]; cycles 290 kernel/locking/test-ww_mutex.c cycle->b_mutex = &cycles[0].a_mutex; cycles 292 kernel/locking/test-ww_mutex.c cycle->b_mutex = &cycles[n + 1].a_mutex; cycles 295 kernel/locking/test-ww_mutex.c cycle->a_signal = &cycles[last].b_signal; cycles 297 kernel/locking/test-ww_mutex.c cycle->a_signal = &cycles[n - 1].b_signal; cycles 305 kernel/locking/test-ww_mutex.c queue_work(wq, &cycles[n].work); cycles 311 kernel/locking/test-ww_mutex.c struct test_cycle *cycle = &cycles[n]; cycles 323 kernel/locking/test-ww_mutex.c ww_mutex_destroy(&cycles[n].a_mutex); cycles 324 kernel/locking/test-ww_mutex.c kfree(cycles); cycles 70 kernel/time/timecounter.c u64 cycles, u64 mask, u64 frac) cycles 72 kernel/time/timecounter.c u64 ns = (u64) cycles; cycles 378 kernel/time/timekeeping.c static inline u64 timekeeping_cycles_to_ns(const struct tk_read_base *tkr, u64 cycles) cycles 383 kernel/time/timekeeping.c delta = clocksource_delta(cycles, tkr->cycle_last, tkr->mask); cycles 991 kernel/time/timekeeping.c systime_snapshot->cycles = now; cycles 1127 kernel/time/timekeeping.c u64 cycles, now, interval_start; cycles 1153 kernel/time/timekeeping.c cycles = system_counterval.cycles; cycles 1161 kernel/time/timekeeping.c if (!cycle_between(interval_start, cycles, now)) { cycles 1164 kernel/time/timekeeping.c cycles = interval_start; cycles 1175 kernel/time/timekeeping.c system_counterval.cycles); cycles 1177 kernel/time/timekeeping.c system_counterval.cycles); cycles 1197 kernel/time/timekeeping.c !cycle_between(history_begin->cycles, cycles 1198 kernel/time/timekeeping.c system_counterval.cycles, cycles) || cycles 1201 kernel/time/timekeeping.c partial_history_cycles = cycles - system_counterval.cycles; cycles 1202 kernel/time/timekeeping.c total_history_cycles = cycles - history_begin->cycles; cycles 35 lib/vdso/gettimeofday.c u64 vdso_calc_delta(u64 cycles, u64 last, u64 mask, u32 mult) cycles 37 lib/vdso/gettimeofday.c return ((cycles - last) & mask) * mult; cycles 45 lib/vdso/gettimeofday.c u64 cycles, last, sec, ns; cycles 50 lib/vdso/gettimeofday.c cycles = __arch_get_hw_counter(vd->clock_mode); cycles 53 lib/vdso/gettimeofday.c if (unlikely((s64)cycles < 0)) cycles 56 lib/vdso/gettimeofday.c ns += vdso_calc_delta(cycles, last, vd->mask, vd->mult); cycles 17 sound/firewire/amdtp-stream-trace.h TP_PROTO(const struct amdtp_stream *s, u32 cycles, const __be32 *cip_header, unsigned int payload_length, unsigned int data_blocks, unsigned int data_block_counter, unsigned int index), cycles 18 sound/firewire/amdtp-stream-trace.h TP_ARGS(s, cycles, cip_header, payload_length, data_blocks, data_block_counter, index), cycles 34 sound/firewire/amdtp-stream-trace.h __entry->second = cycles / CYCLES_PER_SECOND; cycles 35 sound/firewire/amdtp-stream-trace.h __entry->cycle = cycles % CYCLES_PER_SECOND; cycles 253 sound/soc/intel/haswell/sst-haswell-ipc.h u64 cycles; cycles 1177 tools/include/uapi/linux/perf_event.h cycles:16, /* cycle count to last branch */ cycles 578 tools/perf/builtin-diff.c l = llabs(left->diff.cycles); cycles 579 tools/perf/builtin-diff.c r = llabs(right->diff.cycles); cycles 609 tools/perf/builtin-diff.c bi->cycles = ch->cycles; cycles 697 tools/perf/builtin-diff.c pair->diff.cycles = cycles 1360 tools/perf/builtin-diff.c start_line, end_line, block_he->diff.cycles); cycles 1363 tools/perf/builtin-diff.c bi->start, bi->end, block_he->diff.cycles); cycles 774 tools/perf/builtin-script.c br->entries[i].flags.cycles); cycles 819 tools/perf/builtin-script.c br->entries[i].flags.cycles); cycles 868 tools/perf/builtin-script.c br->entries[i].flags.cycles); cycles 961 tools/perf/builtin-script.c if (en->flags.cycles) { cycles 962 tools/perf/builtin-script.c *total_cycles += en->flags.cycles; cycles 963 tools/perf/builtin-script.c printed += fprintf(fp, " %d cycles [%d]", en->flags.cycles, *total_cycles); cycles 965 tools/perf/builtin-script.c printed += fprintf(fp, " %.2f IPC", (float)insn / en->flags.cycles); cycles 819 tools/perf/util/annotate.c unsigned offset, unsigned cycles, cycles 831 tools/perf/util/annotate.c ch[offset].cycles_aggr += cycles; cycles 833 tools/perf/util/annotate.c if (cycles > ch[offset].cycles_max) cycles 834 tools/perf/util/annotate.c ch[offset].cycles_max = cycles; cycles 837 tools/perf/util/annotate.c if (cycles && cycles < ch[offset].cycles_min) cycles 838 tools/perf/util/annotate.c ch[offset].cycles_min = cycles; cycles 840 tools/perf/util/annotate.c ch[offset].cycles_min = cycles; cycles 848 tools/perf/util/annotate.c ch[offset].cycles = 0; cycles 858 tools/perf/util/annotate.c ch[offset].cycles += cycles; cycles 951 tools/perf/util/annotate.c struct symbol *sym, unsigned cycles) cycles 973 tools/perf/util/annotate.c offset, cycles, cycles 979 tools/perf/util/annotate.c unsigned cycles) cycles 984 tools/perf/util/annotate.c if (!cycles) cycles 1005 tools/perf/util/annotate.c err = symbol__account_cycles(ams->al_addr, saddr, ams->sym, cycles); cycles 1030 tools/perf/util/annotate.c if (n_insn && ch->num && ch->cycles) { cycles 1031 tools/perf/util/annotate.c float ipc = n_insn / ((double)ch->cycles / (double)ch->num); cycles 1047 tools/perf/util/annotate.c notes->hit_cycles += ch->cycles; cycles 1071 tools/perf/util/annotate.c if (ch && ch->cycles) { cycles 1078 tools/perf/util/annotate.c al->cycles = ch->cycles_aggr / ch->num_aggr; cycles 2827 tools/perf/util/annotate.c if (al->ipc == 0.0 && al->cycles == 0) cycles 2872 tools/perf/util/annotate.c if (al->cycles) cycles 2874 tools/perf/util/annotate.c ANNOTATION__CYCLES_WIDTH - 1, al->cycles); cycles 2883 tools/perf/util/annotate.c if (al->cycles) { cycles 2888 tools/perf/util/annotate.c al->cycles, al->cycles_min, cycles 138 tools/perf/util/annotate.h u64 cycles; cycles 234 tools/perf/util/annotate.h u64 cycles; cycles 342 tools/perf/util/annotate.h unsigned cycles); cycles 21 tools/perf/util/branch.h u64 cycles:16; cycles 615 tools/perf/util/callchain.c cursor_node->branch_flags.cycles; cycles 773 tools/perf/util/callchain.c cnode->cycles_count += node->branch_flags.cycles; cycles 1359 tools/perf/util/callchain.c u64 cycles, v = 0; cycles 1361 tools/perf/util/callchain.c cycles = cycles_count / branch_count; cycles 1362 tools/perf/util/callchain.c if (cycles) { cycles 1364 tools/perf/util/callchain.c cycles, cycles 916 tools/perf/util/hist.c sample->weight = bi->flags.cycles ? bi->flags.cycles : 1; cycles 2577 tools/perf/util/hist.c if (bs && bs->nr && bs->entries[0].flags.cycles) { cycles 2597 tools/perf/util/hist.c bi[i].flags.cycles); cycles 1665 tools/perf/util/intel-pt.c .cycles = info & LBR_INFO_CYCLES, cycles 2002 tools/perf/util/machine.c u64 cycles; cycles 2070 tools/perf/util/machine.c iter_cycles = iter->cycles; cycles 2103 tools/perf/util/machine.c iter->cycles = 0; cycles 2106 tools/perf/util/machine.c iter->cycles += be[i].flags.cycles; cycles 499 tools/perf/util/scripting-engines/trace-event-python.c PyLong_FromUnsignedLongLong(br->entries[i].flags.cycles)); cycles 1078 tools/perf/util/session.c (unsigned short)e->flags.cycles, cycles 980 tools/perf/util/sort.c return left->branch_info->flags.cycles - cycles 981 tools/perf/util/sort.c right->branch_info->flags.cycles; cycles 989 tools/perf/util/sort.c if (he->branch_info->flags.cycles == 0) cycles 992 tools/perf/util/sort.c he->branch_info->flags.cycles); cycles 72 tools/perf/util/sort.h s64 cycles; cycles 700 tools/perf/util/stat-shadow.c double smi_num, aperf, cycles, cost = 0.0; cycles 706 tools/perf/util/stat-shadow.c cycles = runtime_stat_avg(st, STAT_CYCLES, ctx, cpu); cycles 708 tools/perf/util/stat-shadow.c if ((cycles == 0) || (aperf == 0)) cycles 712 tools/perf/util/stat-shadow.c cost = (aperf - cycles) / aperf * 100.00; cycles 89 tools/perf/util/stat.c ID(CYCLES_IN_TX, cpu/cycles-t/), cycles 92 tools/perf/util/stat.c ID(CYCLES_IN_TX_CP, cpu/cycles-ct/), cycles 112 tools/perf/util/symbol.h u64 cycles; cycles 113 tools/power/cpupower/bench/benchmark.c " for %lius\n", _round + 1, config->cycles, cycles 125 tools/power/cpupower/bench/benchmark.c for (cycle = 0; cycle < config->cycles; cycle++) { cycles 139 tools/power/cpupower/bench/benchmark.c performance_time / config->cycles); cycles 151 tools/power/cpupower/bench/benchmark.c for (cycle = 0; cycle < config->cycles; cycle++) { cycles 169 tools/power/cpupower/bench/benchmark.c powersave_time / config->cycles); cycles 122 tools/power/cpupower/bench/main.c sscanf(optarg, "%u", &config->cycles); cycles 173 tools/power/cpupower/bench/main.c config->cycles, cycles 130 tools/power/cpupower/bench/parse.c config->cycles = 5; cycles 196 tools/power/cpupower/bench/parse.c sscanf(val, "%u", &config->cycles); cycles 16 tools/power/cpupower/bench/parse.h unsigned int cycles; /* calculation cycles with the same sleep/load time */ cycles 133 tools/power/cpupower/bench/system.c sleep_time += 2 * config->cycles * cycles 135 tools/power/cpupower/bench/system.c load_time += 2 * config->cycles * cycles 59 tools/testing/selftests/powerpc/pmu/ebb/cycles_test.c return test_harness(cycles, "cycles"); cycles 20 tools/virtio/ringtest/main.h static inline void wait_cycles(unsigned long long cycles) cycles 25 tools/virtio/ringtest/main.h while (__rdtsc() - t < cycles) {} cycles 32 tools/virtio/ringtest/main.h static inline void wait_cycles(unsigned long long cycles) cycles 34 tools/virtio/ringtest/main.h asm volatile("0: brctg %0,0b" : : "d" (cycles)); cycles 42 tools/virtio/ringtest/main.h static inline void wait_cycles(unsigned long long cycles)