cycles           1252 arch/alpha/kernel/setup.c 	int cycles, prev_cycles = 1000000;
cycles           1264 arch/alpha/kernel/setup.c 		cycles = read_mem_block(__va(0), stride, size);
cycles           1265 arch/alpha/kernel/setup.c 		if (cycles > prev_cycles * 2) {
cycles           1269 arch/alpha/kernel/setup.c 			       size >> 11, prev_cycles, cycles);
cycles           1274 arch/alpha/kernel/setup.c 		prev_cycles = cycles;
cycles            162 arch/arc/include/asm/arcregs.h 	unsigned int pad:8, x1616:8, dsp:4, cycles:2, type:2, ver:8;
cycles            164 arch/arc/include/asm/arcregs.h 	unsigned int ver:8, type:2, cycles:2, dsp:4, x1616:8, pad:8;
cycles             45 arch/arm/lib/delay.c static void __timer_delay(unsigned long cycles)
cycles             49 arch/arm/lib/delay.c 	while ((get_cycles() - start) < cycles)
cycles            288 arch/arm/mach-davinci/time.c static int davinci_set_next_event(unsigned long cycles,
cycles            293 arch/arm/mach-davinci/time.c 	t->period = cycles;
cycles             85 arch/arm/mach-imx/mmdc.c PMU_EVENT_ATTR_STRING(total-cycles, mmdc_pmu_total_cycles, "event=0x00")
cycles             86 arch/arm/mach-imx/mmdc.c PMU_EVENT_ATTR_STRING(busy-cycles, mmdc_pmu_busy_cycles, "event=0x01")
cycles            120 arch/arm/mach-omap1/time.c static int omap_mpu_set_next_event(unsigned long cycles,
cycles            123 arch/arm/mach-omap1/time.c 	omap_mpu_timer_start(0, cycles, 0);
cycles            100 arch/arm/mach-omap2/timer.c static int omap2_gp_timer_set_next_event(unsigned long cycles,
cycles            104 arch/arm/mach-omap2/timer.c 				   0xffffffff - cycles, OMAP_TIMER_POSTED);
cycles            433 arch/arm/mach-omap2/vc.c 	u32 cycles;
cycles            438 arch/arm/mach-omap2/vc.c 	cycles = voltdm->sys_clk.rate / 1000 * time / 1000;
cycles            440 arch/arm/mach-omap2/vc.c 	cycles /= 64;
cycles            446 arch/arm/mach-omap2/vc.c 	if (cycles > 63) {
cycles            447 arch/arm/mach-omap2/vc.c 		cycles /= 4;
cycles            452 arch/arm/mach-omap2/vc.c 	if (cycles > 63) {
cycles            453 arch/arm/mach-omap2/vc.c 		cycles /= 2;
cycles            458 arch/arm/mach-omap2/vc.c 	if (cycles > 63) {
cycles            459 arch/arm/mach-omap2/vc.c 		cycles /= 4;
cycles            464 arch/arm/mach-omap2/vc.c 	if (cycles > 63) {
cycles            470 arch/arm/mach-omap2/vc.c 	cycles++;
cycles            473 arch/arm/mach-omap2/vc.c 		(cycles << OMAP4430_RAMP_UP_COUNT_SHIFT);
cycles            157 arch/arm/mach-spear/time.c static int clockevent_next_event(unsigned long cycles,
cycles            165 arch/arm/mach-spear/time.c 	writew(cycles, gpt_base + LOAD(CLKEVT));
cycles             51 arch/arm/plat-omap/counter_32k.c static cycles_t cycles;
cycles             59 arch/arm/plat-omap/counter_32k.c 	last_cycles = cycles;
cycles             60 arch/arm/plat-omap/counter_32k.c 	cycles = sync32k_cnt_reg ? readl_relaxed(sync32k_cnt_reg) : 0;
cycles             62 arch/arm/plat-omap/counter_32k.c 	nsecs = clocksource_cyc2ns(cycles - last_cycles,
cycles             26 arch/arm64/lib/delay.c void __delay(unsigned long cycles)
cycles             34 arch/arm64/lib/delay.c 		while ((get_cycles() - start + timer_evt_period) < cycles)
cycles             38 arch/arm64/lib/delay.c 	while ((get_cycles() - start) < cycles)
cycles             14 arch/h8300/lib/delay.c void __delay(unsigned long cycles)
cycles             17 arch/h8300/lib/delay.c 			  "bne 1b":"=r"(cycles):"0"(cycles));
cycles             11 arch/hexagon/include/asm/delay.h extern void __delay(unsigned long cycles);
cycles            206 arch/hexagon/kernel/time.c void __delay(unsigned long cycles)
cycles            210 arch/hexagon/kernel/time.c 	while ((__vmgettime() - start) < cycles)
cycles           1369 arch/ia64/include/asm/pal.h 				cycles		: 8,
cycles            667 arch/ia64/kernel/palinfo.c 		   pm_info.pal_perf_mon_info_s.cycles,
cycles             82 arch/m68k/68000/timers.c 	u32 cycles;
cycles             85 arch/m68k/68000/timers.c 	cycles = m68328_tick_cnt + TCN;
cycles             88 arch/m68k/68000/timers.c 	return cycles;
cycles            125 arch/m68k/coldfire/pit.c 	u32 cycles;
cycles            130 arch/m68k/coldfire/pit.c 	cycles = pit_cnt;
cycles            133 arch/m68k/coldfire/pit.c 	return cycles + PIT_CYCLES_PER_JIFFY - pcntr;
cycles            104 arch/m68k/coldfire/sltimers.c 	u32 cycles, scnt;
cycles            108 arch/m68k/coldfire/sltimers.c 	cycles = mcfslt_cnt;
cycles            110 arch/m68k/coldfire/sltimers.c 		cycles += mcfslt_cycles_per_jiffy;
cycles            116 arch/m68k/coldfire/sltimers.c 	return cycles + ((mcfslt_cycles_per_jiffy - 1) - scnt);
cycles             96 arch/m68k/coldfire/timers.c 	u32 cycles;
cycles            101 arch/m68k/coldfire/timers.c 	cycles = mcftmr_cnt;
cycles            104 arch/m68k/coldfire/timers.c 	return cycles + tcn;
cycles            415 arch/mips/pci/pci-octeon.c 		unsigned long cycles, pci_clock;
cycles            418 arch/mips/pci/pci-octeon.c 		cycles = read_c0_cvmcount();
cycles            421 arch/mips/pci/pci-octeon.c 		cycles = read_c0_cvmcount() - cycles;
cycles            423 arch/mips/pci/pci-octeon.c 			    (cycles / (mips_hpt_frequency / 1000000));
cycles             11 arch/nios2/lib/delay.c void __delay(unsigned long cycles)
cycles             15 arch/nios2/lib/delay.c 	while ((get_cycles() - start) < cycles)
cycles             29 arch/openrisc/lib/delay.c void __delay(unsigned long cycles)
cycles             33 arch/openrisc/lib/delay.c 	while ((get_cycles() - start) < cycles)
cycles             50 arch/powerpc/perf/generic-compat-pmu.c GENERIC_EVENT_ATTR(cpu-cycles,			PM_CYC);
cycles             92 arch/powerpc/perf/hv-gpci-requests.h 	__count(0x10,	8,	cycles)
cycles            374 arch/powerpc/perf/power7-pmu.c GENERIC_EVENT_ATTR(cpu-cycles,			PM_CYC);
cycles            375 arch/powerpc/perf/power7-pmu.c GENERIC_EVENT_ATTR(stalled-cycles-frontend,	PM_GCT_NOSLOT_CYC);
cycles            376 arch/powerpc/perf/power7-pmu.c GENERIC_EVENT_ATTR(stalled-cycles-backend,	PM_CMPLU_STALL);
cycles            123 arch/powerpc/perf/power8-pmu.c GENERIC_EVENT_ATTR(cpu-cycles,			PM_CYC);
cycles            124 arch/powerpc/perf/power8-pmu.c GENERIC_EVENT_ATTR(stalled-cycles-frontend,	PM_GCT_NOSLOT_CYC);
cycles            125 arch/powerpc/perf/power8-pmu.c GENERIC_EVENT_ATTR(stalled-cycles-backend,	PM_CMPLU_STALL);
cycles            152 arch/powerpc/perf/power9-pmu.c GENERIC_EVENT_ATTR(cpu-cycles,			PM_CYC);
cycles            153 arch/powerpc/perf/power9-pmu.c GENERIC_EVENT_ATTR(stalled-cycles-frontend,	PM_ICT_NOSLOT_CYC);
cycles            154 arch/powerpc/perf/power9-pmu.c GENERIC_EVENT_ATTR(stalled-cycles-backend,	PM_CMPLU_STALL);
cycles             18 arch/riscv/include/asm/delay.h extern void __delay(unsigned long cycles);
cycles             72 arch/riscv/lib/delay.c void __delay(unsigned long cycles)
cycles             76 arch/riscv/lib/delay.c 	while ((unsigned long)(get_cycles() - t0) < cycles)
cycles            155 arch/sparc/kernel/time_32.c 	u64 cycles;
cycles            160 arch/sparc/kernel/time_32.c 		cycles = timer_cs_internal_counter;
cycles            165 arch/sparc/kernel/time_32.c 	cycles *= sparc_config.cs_period;
cycles            166 arch/sparc/kernel/time_32.c 	cycles += offset;
cycles            168 arch/sparc/kernel/time_32.c 	return cycles;
cycles            129 arch/sparc/vdso/vclock_gettime.c 	u64 cycles;
cycles            131 arch/sparc/vdso/vclock_gettime.c 	cycles = vread_tick();
cycles            132 arch/sparc/vdso/vclock_gettime.c 	v = (cycles - vvar->clock.cycle_last) & vvar->clock.mask;
cycles            139 arch/sparc/vdso/vclock_gettime.c 	u64 cycles;
cycles            141 arch/sparc/vdso/vclock_gettime.c 	cycles = vread_tick_stick();
cycles            142 arch/sparc/vdso/vclock_gettime.c 	v = (cycles - vvar->clock.cycle_last) & vvar->clock.mask;
cycles           1687 arch/x86/events/core.c EVENT_ATTR(cpu-cycles,			CPU_CYCLES		);
cycles           1693 arch/x86/events/core.c EVENT_ATTR(bus-cycles,			BUS_CYCLES		);
cycles           1694 arch/x86/events/core.c EVENT_ATTR(stalled-cycles-frontend,	STALLED_CYCLES_FRONTEND	);
cycles           1695 arch/x86/events/core.c EVENT_ATTR(stalled-cycles-backend,	STALLED_CYCLES_BACKEND	);
cycles           1696 arch/x86/events/core.c EVENT_ATTR(ref-cycles,			REF_CPU_CYCLES		);
cycles           4236 arch/x86/events/intel/core.c EVENT_ATTR_STR(cycles-t,	cycles_t,	"event=0x3c,in_tx=1");
cycles           4237 arch/x86/events/intel/core.c EVENT_ATTR_STR(cycles-ct,	cycles_ct,	"event=0x3c,in_tx=1,in_tx_cp=1");
cycles            560 arch/x86/events/intel/lbr.c 		cpuc->lbr_entries[i].cycles	= 0;
cycles            592 arch/x86/events/intel/lbr.c 		u16 cycles = 0;
cycles            613 arch/x86/events/intel/lbr.c 			cycles = (info & LBR_INFO_CYCLES);
cycles            620 arch/x86/events/intel/lbr.c 			cycles = ((to >> 48) & LBR_INFO_CYCLES);
cycles            654 arch/x86/events/intel/lbr.c 		cpuc->lbr_entries[out].cycles	 = cycles;
cycles           1110 arch/x86/events/intel/lbr.c 		e->cycles	= info & LBR_INFO_CYCLES;
cycles            287 arch/x86/include/asm/vdso/gettimeofday.h u64 vdso_calc_delta(u64 cycles, u64 last, u64 mask, u32 mult)
cycles            289 arch/x86/include/asm/vdso/gettimeofday.h 	if (cycles > last)
cycles            290 arch/x86/include/asm/vdso/gettimeofday.h 		return (cycles - last) * mult;
cycles           1231 arch/x86/kernel/tsc.c 			.cycles = res};
cycles           1269 arch/x86/kernel/tsc.c 					      .cycles = res};
cycles             40 arch/xtensa/include/asm/delay.h 	unsigned long cycles = (usecs * (ccount_freq >> 15)) >> 5;
cycles             43 arch/xtensa/include/asm/delay.h 	while (((unsigned long)get_ccount()) - start < cycles)
cycles             61 arch/xtensa/include/asm/delay.h 	unsigned long cycles = (nsec * (ccount_freq >> 15)) >> 15;
cycles             62 arch/xtensa/include/asm/delay.h 	__delay(cycles);
cycles            212 crypto/tcrypt.c 	unsigned long cycles = 0;
cycles            239 crypto/tcrypt.c 		cycles += end - start;
cycles            243 crypto/tcrypt.c 		(cycles + 4) / (8 * num_mb), blen);
cycles            478 crypto/tcrypt.c 	unsigned long cycles = 0;
cycles            507 crypto/tcrypt.c 		cycles += end - start;
cycles            513 crypto/tcrypt.c 		       (cycles + 4) / 8, blen);
cycles            773 crypto/tcrypt.c 	unsigned long cycles = 0;
cycles            800 crypto/tcrypt.c 		cycles += end - start;
cycles            804 crypto/tcrypt.c 		(cycles + 4) / (8 * num_mb), blen);
cycles            962 crypto/tcrypt.c 	unsigned long cycles = 0;
cycles            984 crypto/tcrypt.c 		cycles += end - start;
cycles            992 crypto/tcrypt.c 		cycles / 8, cycles / (8 * blen));
cycles           1000 crypto/tcrypt.c 	unsigned long cycles = 0;
cycles           1041 crypto/tcrypt.c 		cycles += end - start;
cycles           1049 crypto/tcrypt.c 		cycles / 8, cycles / (8 * blen));
cycles           1210 crypto/tcrypt.c 	unsigned long cycles = 0;
cycles           1237 crypto/tcrypt.c 		cycles += end - start;
cycles           1241 crypto/tcrypt.c 		(cycles + 4) / (8 * num_mb), blen);
cycles           1445 crypto/tcrypt.c 	unsigned long cycles = 0;
cycles           1478 crypto/tcrypt.c 		cycles += end - start;
cycles           1484 crypto/tcrypt.c 			(cycles + 4) / 8, blen);
cycles           1210 drivers/char/random.c 		unsigned cycles;
cycles           1216 drivers/char/random.c 	sample.cycles = random_get_entropy();
cycles           1314 drivers/char/random.c 	cycles_t		cycles = random_get_entropy();
cycles           1320 drivers/char/random.c 	if (cycles == 0)
cycles           1321 drivers/char/random.c 		cycles = get_reg(fast_pool, regs);
cycles           1322 drivers/char/random.c 	c_high = (sizeof(cycles) > 4) ? cycles >> 32 : 0;
cycles           1324 drivers/char/random.c 	fast_pool->pool[0] ^= cycles ^ j_high ^ irq;
cycles           1332 drivers/char/random.c 	add_interrupt_bench(cycles);
cycles             61 drivers/clk/ti/clkctrl.c 	u32 cycles;
cycles            107 drivers/clk/ti/clkctrl.c 		if (time->cycles++ < timeout) {
cycles            243 drivers/clocksource/arc_timer.c static void arc_timer_event_setup(unsigned int cycles)
cycles            245 drivers/clocksource/arc_timer.c 	write_aux_reg(ARC_REG_TIMER0_LIMIT, cycles);
cycles            261 drivers/clocksource/exynos_mct.c static void exynos4_mct_comp0_start(bool periodic, unsigned long cycles)
cycles            270 drivers/clocksource/exynos_mct.c 		exynos4_mct_write(cycles, EXYNOS4_MCT_G_COMP0_ADD_INCR);
cycles            273 drivers/clocksource/exynos_mct.c 	comp_cycle = exynos4_read_count_64() + cycles;
cycles            283 drivers/clocksource/exynos_mct.c static int exynos4_comp_set_next_event(unsigned long cycles,
cycles            286 drivers/clocksource/exynos_mct.c 	exynos4_mct_comp0_start(false, cycles);
cycles            365 drivers/clocksource/exynos_mct.c static void exynos4_mct_tick_start(unsigned long cycles,
cycles            372 drivers/clocksource/exynos_mct.c 	tmp = (1 << 31) | cycles;	/* MCT_L_UPDATE_ICNTB */
cycles            393 drivers/clocksource/exynos_mct.c static int exynos4_tick_set_next_event(unsigned long cycles,
cycles            399 drivers/clocksource/exynos_mct.c 	exynos4_mct_tick_start(cycles, mevt);
cycles            185 drivers/clocksource/samsung_pwm_timer.c static int samsung_set_next_event(unsigned long cycles,
cycles            198 drivers/clocksource/samsung_pwm_timer.c 	if (!cycles)
cycles            199 drivers/clocksource/samsung_pwm_timer.c 		cycles = 1;
cycles            201 drivers/clocksource/samsung_pwm_timer.c 	samsung_time_setup(pwm.event_id, cycles);
cycles             89 drivers/clocksource/timer-atlas7.c 	u64 cycles;
cycles             94 drivers/clocksource/timer-atlas7.c 	cycles = readl_relaxed(sirfsoc_timer_base + SIRFSOC_TIMER_64COUNTER_RLATCHED_HI);
cycles             95 drivers/clocksource/timer-atlas7.c 	cycles = (cycles << 32) | readl_relaxed(sirfsoc_timer_base + SIRFSOC_TIMER_64COUNTER_RLATCHED_LO);
cycles             97 drivers/clocksource/timer-atlas7.c 	return cycles;
cycles            108 drivers/clocksource/timer-cadence-ttc.c 					unsigned long cycles)
cycles            117 drivers/clocksource/timer-cadence-ttc.c 	writel_relaxed(cycles, timer->base_addr + TTC_INTR_VAL_OFFSET);
cycles            175 drivers/clocksource/timer-cadence-ttc.c static int ttc_set_next_event(unsigned long cycles,
cycles            181 drivers/clocksource/timer-cadence-ttc.c 	ttc_set_interval(timer, cycles);
cycles            139 drivers/clocksource/timer-davinci.c davinci_clockevent_set_next_event_std(unsigned long cycles,
cycles            147 drivers/clocksource/timer-davinci.c 	davinci_clockevent_write(clockevent, DAVINCI_TIMER_REG_PRD12, cycles);
cycles            155 drivers/clocksource/timer-davinci.c davinci_clockevent_set_next_event_cmp(unsigned long cycles,
cycles            164 drivers/clocksource/timer-davinci.c 				 clockevent->cmp_off, curr_time + cycles);
cycles            136 drivers/clocksource/timer-fttmr010.c static int fttmr010_timer_set_next_event(unsigned long cycles,
cycles            152 drivers/clocksource/timer-fttmr010.c 		writel(cycles, fttmr010->base + TIMER1_LOAD);
cycles            156 drivers/clocksource/timer-fttmr010.c 		writel(cr + cycles, fttmr010->base + TIMER1_MATCH1);
cycles            107 drivers/clocksource/timer-ixp4xx.c static int ixp4xx_set_next_event(unsigned long cycles,
cycles            116 drivers/clocksource/timer-ixp4xx.c 	__raw_writel((cycles & ~IXP4XX_OST_RELOAD_MASK) | val,
cycles            125 drivers/clocksource/timer-keystone.c static int keystone_set_next_event(unsigned long cycles,
cycles            128 drivers/clocksource/timer-keystone.c 	return keystone_timer_config(cycles, TCR_ENAMODE_ONESHOT_MASK);
cycles             76 drivers/clocksource/timer-prima2.c 	u64 cycles;
cycles             81 drivers/clocksource/timer-prima2.c 	cycles = readl_relaxed(sirfsoc_timer_base + SIRFSOC_TIMER_LATCHED_HI);
cycles             82 drivers/clocksource/timer-prima2.c 	cycles = (cycles << 32) |
cycles             85 drivers/clocksource/timer-prima2.c 	return cycles;
cycles             50 drivers/clocksource/timer-qcom.c static int msm_timer_set_next_event(unsigned long cycles,
cycles             59 drivers/clocksource/timer-qcom.c 	writel_relaxed(cycles, event_base + TIMER_MATCH_VAL);
cycles             35 drivers/clocksource/timer-rda.c static int rda_ostimer_start(void __iomem *base, bool periodic, u64 cycles)
cycles             39 drivers/clocksource/timer-rda.c 	load_l = (u32)cycles;
cycles             40 drivers/clocksource/timer-rda.c 	ctrl = ((cycles >> 32) & 0xffffff);
cycles             65 drivers/clocksource/timer-rockchip.c static void rk_timer_update_counter(unsigned long cycles,
cycles             68 drivers/clocksource/timer-rockchip.c 	writel_relaxed(cycles, timer->base + TIMER_LOAD_COUNT0);
cycles             77 drivers/clocksource/timer-rockchip.c static inline int rk_timer_set_next_event(unsigned long cycles,
cycles             83 drivers/clocksource/timer-rockchip.c 	rk_timer_update_counter(cycles, timer);
cycles             60 drivers/clocksource/timer-sprd.c static void sprd_timer_update_counter(void __iomem *base, unsigned long cycles)
cycles             62 drivers/clocksource/timer-sprd.c 	writel_relaxed(cycles & TIMER_VALUE_LO_MASK, base + TIMER_LOAD_LO);
cycles             79 drivers/clocksource/timer-sprd.c static int sprd_timer_set_next_event(unsigned long cycles,
cycles             85 drivers/clocksource/timer-sprd.c 	sprd_timer_update_counter(timer_of_base(to), cycles);
cycles             54 drivers/clocksource/timer-tegra.c static int tegra_timer_set_next_event(unsigned long cycles,
cycles             68 drivers/clocksource/timer-tegra.c 	writel_relaxed(TIMER_PTV_EN | (cycles - 1), reg_base + TIMER_PTV);
cycles            276 drivers/clocksource/timer-u300.c static int u300_set_next_event(unsigned long cycles,
cycles            290 drivers/clocksource/timer-u300.c 	writel(cycles, u300_timer_base + U300_TIMER_APP_GPT1TC);
cycles             61 drivers/clocksource/timer-vt8500.c static int vt8500_timer_set_next_event(unsigned long cycles,
cycles             65 drivers/clocksource/timer-vt8500.c 	u64 alarm = clocksource.read(&clocksource) + cycles;
cycles            164 drivers/crypto/qat/qat_common/qat_hal.c 			       unsigned char ae, unsigned int cycles,
cycles            174 drivers/crypto/qat/qat_common/qat_hal.c 	while ((int)cycles > elapsed_cycles && times--) {
cycles            465 drivers/dma/moxart-dma.c 	unsigned int completed_cycles, cycles;
cycles            468 drivers/dma/moxart-dma.c 	cycles = readl(ch->base + REG_OFF_CYCLES);
cycles            469 drivers/dma/moxart-dma.c 	completed_cycles = (ch->desc->dma_cycles - cycles);
cycles            731 drivers/firewire/core-transaction.c 	unsigned int cycles;
cycles            734 drivers/firewire/core-transaction.c 	cycles = card->split_timeout_cycles;
cycles            735 drivers/firewire/core-transaction.c 	cycles += request_timestamp & 0x1fff;
cycles            738 drivers/firewire/core-transaction.c 	timestamp += (cycles / 8000) << 13;
cycles            739 drivers/firewire/core-transaction.c 	timestamp |= cycles % 8000;
cycles           1073 drivers/firewire/core-transaction.c 	unsigned int cycles;
cycles           1075 drivers/firewire/core-transaction.c 	cycles = card->split_timeout_hi * 8000 + (card->split_timeout_lo >> 19);
cycles           1078 drivers/firewire/core-transaction.c 	cycles = clamp(cycles, 800u, 3u * 8000u);
cycles           1080 drivers/firewire/core-transaction.c 	card->split_timeout_cycles = cycles;
cycles           1081 drivers/firewire/core-transaction.c 	card->split_timeout_jiffies = DIV_ROUND_UP(cycles * HZ, 8000);
cycles            756 drivers/gpio/gpio-aspeed.c 		u32 *cycles)
cycles            773 drivers/gpio/gpio-aspeed.c 	*cycles = n + (!!r);
cycles            873 drivers/gpio/gpio-aspeed.c 		u32 cycles;
cycles            875 drivers/gpio/gpio-aspeed.c 		cycles = ioread32(gpio->base + debounce_timers[i]);
cycles            876 drivers/gpio/gpio-aspeed.c 		if (requested_cycles == cycles)
cycles            129 drivers/gpu/drm/nouveau/nvkm/subdev/therm/fan.c 	u32 cycles, cur, prev;
cycles            145 drivers/gpu/drm/nouveau/nvkm/subdev/therm/fan.c 	cycles = 0;
cycles            154 drivers/gpu/drm/nouveau/nvkm/subdev/therm/fan.c 			cycles++;
cycles            157 drivers/gpu/drm/nouveau/nvkm/subdev/therm/fan.c 	} while (cycles < 5 && nvkm_timer_read(tmr) - start < 250000000);
cycles            160 drivers/gpu/drm/nouveau/nvkm/subdev/therm/fan.c 	if (cycles == 5) {
cycles            217 drivers/gpu/drm/v3d/v3d_debugfs.c 	uint32_t cycles;
cycles            241 drivers/gpu/drm/v3d/v3d_debugfs.c 	cycles = V3D_CORE_READ(core, V3D_PCTR_0_PCTR0);
cycles            244 drivers/gpu/drm/v3d/v3d_debugfs.c 		   cycles,
cycles            245 drivers/gpu/drm/v3d/v3d_debugfs.c 		   cycles / (measure_ms * 1000),
cycles            246 drivers/gpu/drm/v3d/v3d_debugfs.c 		   (cycles / (measure_ms * 100)) % 10);
cycles           1597 drivers/infiniband/hw/hfi1/hfi.h 	u32 cycles;
cycles           1607 drivers/infiniband/hw/hfi1/hfi.h 	cycles = len * 8; /* bits */
cycles           1608 drivers/infiniband/hw/hfi1/hfi.h 	cycles *= 805;
cycles           1609 drivers/infiniband/hw/hfi1/hfi.h 	cycles /= rate;
cycles           1611 drivers/infiniband/hw/hfi1/hfi.h 	return cycles;
cycles            373 drivers/input/keyboard/pmic8xxx-keypad.c 	int bits, rc, cycles;
cycles            438 drivers/input/keyboard/pmic8xxx-keypad.c 	cycles = (row_hold_ns * KEYP_CLOCK_FREQ) / NSEC_PER_SEC;
cycles            440 drivers/input/keyboard/pmic8xxx-keypad.c 	scan_val |= (cycles << KEYP_SCAN_ROW_HOLD_SHIFT);
cycles            136 drivers/media/common/b2c2/flexcop-common.h 		flexcop_dma_index_t dma_idx, u8 cycles);
cycles            387 drivers/media/i2c/ov2685.c static inline u32 ov2685_cal_delay(u32 cycles)
cycles            389 drivers/media/i2c/ov2685.c 	return DIV_ROUND_UP(cycles, OV2685_XVCLK_FREQ / 1000 / 1000);
cycles            430 drivers/media/i2c/tc358743_regs.h #define SET_FREQ_RANGE_MODE_CYCLES(cycles)   (((cycles) - 1) & \
cycles            162 drivers/media/pci/b2c2/flexcop-dma.c 		flexcop_dma_index_t dma_idx, u8 cycles)
cycles            170 drivers/media/pci/b2c2/flexcop-dma.c 	v.dma_0x4_write.dmatimer = cycles;
cycles           2210 drivers/media/platform/rcar_fdp1.c 	u32 cycles;
cycles           2213 drivers/media/platform/rcar_fdp1.c 	cycles = fdp1_read(fdp1, FD1_CTL_VCYCLE_STAT);
cycles           2228 drivers/media/platform/rcar_fdp1.c 			cycles, cycles/(fdp1->clk_rate/1000));
cycles            290 drivers/media/rc/iguanair.c 		uint32_t cycles, fours, sevens;
cycles            294 drivers/media/rc/iguanair.c 		cycles = DIV_ROUND_CLOSEST(24000000, carrier * 2) -
cycles            302 drivers/media/rc/iguanair.c 		sevens = (4 - cycles) & 3;
cycles            303 drivers/media/rc/iguanair.c 		fours = (cycles - sevens * 7) / 4;
cycles            176 drivers/media/usb/gspca/stv06xx/stv06xx_hdcs.c 	int cycles, err;
cycles            179 drivers/media/usb/gspca/stv06xx/stv06xx_hdcs.c 	cycles = val * HDCS_CLK_FREQ_MHZ * 257;
cycles            187 drivers/media/usb/gspca/stv06xx/stv06xx_hdcs.c 	rowexp = cycles / rp;
cycles            190 drivers/media/usb/gspca/stv06xx/stv06xx_hdcs.c 	cycles -= rowexp * rp;
cycles            195 drivers/media/usb/gspca/stv06xx/stv06xx_hdcs.c 		srowexp = hdcs->w - (cycles + hdcs->exp.er + 13) / ct;
cycles            201 drivers/media/usb/gspca/stv06xx/stv06xx_hdcs.c 		srowexp = cp - hdcs->exp.er - 6 - cycles;
cycles            160 drivers/memory/jz4780-nemc.c 	uint32_t smcr, val, cycles;
cycles            206 drivers/memory/jz4780-nemc.c 		cycles = jz4780_nemc_ns_to_cycles(nemc, val);
cycles            207 drivers/memory/jz4780-nemc.c 		if (cycles > nemc->soc_info->tas_tah_cycles_max) {
cycles            209 drivers/memory/jz4780-nemc.c 				val, cycles);
cycles            213 drivers/memory/jz4780-nemc.c 		smcr |= cycles << NEMC_SMCR_TAS_SHIFT;
cycles            218 drivers/memory/jz4780-nemc.c 		cycles = jz4780_nemc_ns_to_cycles(nemc, val);
cycles            219 drivers/memory/jz4780-nemc.c 		if (cycles > nemc->soc_info->tas_tah_cycles_max) {
cycles            221 drivers/memory/jz4780-nemc.c 				val, cycles);
cycles            225 drivers/memory/jz4780-nemc.c 		smcr |= cycles << NEMC_SMCR_TAH_SHIFT;
cycles            230 drivers/memory/jz4780-nemc.c 		cycles = jz4780_nemc_ns_to_cycles(nemc, val);
cycles            231 drivers/memory/jz4780-nemc.c 		if (cycles > 31) {
cycles            233 drivers/memory/jz4780-nemc.c 				val, cycles);
cycles            237 drivers/memory/jz4780-nemc.c 		smcr |= convert_tBP_tAW[cycles] << NEMC_SMCR_TBP_SHIFT;
cycles            242 drivers/memory/jz4780-nemc.c 		cycles = jz4780_nemc_ns_to_cycles(nemc, val);
cycles            243 drivers/memory/jz4780-nemc.c 		if (cycles > 31) {
cycles            245 drivers/memory/jz4780-nemc.c 				val, cycles);
cycles            249 drivers/memory/jz4780-nemc.c 		smcr |= convert_tBP_tAW[cycles] << NEMC_SMCR_TAW_SHIFT;
cycles            254 drivers/memory/jz4780-nemc.c 		cycles = jz4780_nemc_ns_to_cycles(nemc, val);
cycles            255 drivers/memory/jz4780-nemc.c 		if (cycles > 63) {
cycles            257 drivers/memory/jz4780-nemc.c 				val, cycles);
cycles            261 drivers/memory/jz4780-nemc.c 		smcr |= cycles << NEMC_SMCR_TSTRV_SHIFT;
cycles             64 drivers/memory/pl172.c 	int cycles;
cycles             68 drivers/memory/pl172.c 		cycles = DIV_ROUND_UP(val * pl172->rate, NSEC_PER_MSEC) - start;
cycles             69 drivers/memory/pl172.c 		if (cycles < 0) {
cycles             70 drivers/memory/pl172.c 			cycles = 0;
cycles             71 drivers/memory/pl172.c 		} else if (cycles > max) {
cycles             76 drivers/memory/pl172.c 		writel(cycles, pl172->base + reg_offset);
cycles            338 drivers/misc/altera-stapl/altera-jtag.c 					s32 cycles,
cycles            356 drivers/misc/altera-stapl/altera-jtag.c 		for (count = 0L; count < cycles; count++)
cycles             83 drivers/misc/altera-stapl/altera-jtag.h int altera_wait_cycles(struct altera_state *astate, s32 cycles,
cycles            219 drivers/mtd/nand/raw/diskonchip.c static void DoC_Delay(struct doc_priv *doc, unsigned short cycles)
cycles            224 drivers/mtd/nand/raw/diskonchip.c 	for (i = 0; i < cycles; i++) {
cycles           1706 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h void t4_pmtx_get_stats(struct adapter *adap, u32 cnt[], u64 cycles[]);
cycles           1707 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h void t4_pmrx_get_stats(struct adapter *adap, u32 cnt[], u64 cycles[]);
cycles           6060 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c void t4_pmtx_get_stats(struct adapter *adap, u32 cnt[], u64 cycles[])
cycles           6069 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 			cycles[i] = t4_read_reg64(adap, PM_TX_STAT_LSB_A);
cycles           6074 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 			cycles[i] = (((u64)data[0] << 32) | data[1]);
cycles           6087 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c void t4_pmrx_get_stats(struct adapter *adap, u32 cnt[], u64 cycles[])
cycles           6096 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 			cycles[i] = t4_read_reg64(adap, PM_RX_STAT_LSB_A);
cycles           6101 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 			cycles[i] = (((u64)data[0] << 32) | data[1]);
cycles            180 drivers/net/ethernet/intel/e1000e/ptp.c 	u64 cycles, ns;
cycles            185 drivers/net/ethernet/intel/e1000e/ptp.c 	cycles = e1000e_read_systim(adapter, sts);
cycles            186 drivers/net/ethernet/intel/e1000e/ptp.c 	ns = timecounter_cyc2time(&adapter->tc, cycles);
cycles           2186 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c static u32 mvpp2_cycles_to_usec(u32 cycles, unsigned long clk_hz)
cycles           2188 drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 	u64 tmp = (u64)cycles * USEC_PER_SEC;
cycles             44 drivers/net/ethernet/mellanox/mlx4/en_clock.c 		container_of(tc, struct mlx4_en_dev, cycles);
cycles            141 drivers/net/ethernet/mellanox/mlx4/en_clock.c 	mdev->cycles.mult = neg_adj ? mult - diff : mult + diff;
cycles            210 drivers/net/ethernet/mellanox/mlx4/en_clock.c 	timecounter_init(&mdev->clock, &mdev->cycles, ns);
cycles            277 drivers/net/ethernet/mellanox/mlx4/en_clock.c 	memset(&mdev->cycles, 0, sizeof(mdev->cycles));
cycles            278 drivers/net/ethernet/mellanox/mlx4/en_clock.c 	mdev->cycles.read = mlx4_en_read_clock;
cycles            279 drivers/net/ethernet/mellanox/mlx4/en_clock.c 	mdev->cycles.mask = CLOCKSOURCE_MASK(48);
cycles            280 drivers/net/ethernet/mellanox/mlx4/en_clock.c 	mdev->cycles.shift = freq_to_shift(dev->caps.hca_core_clock);
cycles            281 drivers/net/ethernet/mellanox/mlx4/en_clock.c 	mdev->cycles.mult =
cycles            282 drivers/net/ethernet/mellanox/mlx4/en_clock.c 		clocksource_khz2mult(1000 * dev->caps.hca_core_clock, mdev->cycles.shift);
cycles            283 drivers/net/ethernet/mellanox/mlx4/en_clock.c 	mdev->nominal_c_mult = mdev->cycles.mult;
cycles            286 drivers/net/ethernet/mellanox/mlx4/en_clock.c 	timecounter_init(&mdev->clock, &mdev->cycles,
cycles           1909 drivers/net/ethernet/mellanox/mlx4/main.c 	u64 cycles;
cycles           1921 drivers/net/ethernet/mellanox/mlx4/main.c 	cycles = (u64) clockhi << 32 | (u64) clocklo;
cycles           1923 drivers/net/ethernet/mellanox/mlx4/main.c 	return cycles;
cycles            432 drivers/net/ethernet/mellanox/mlx4/mlx4_en.h 	struct cyclecounter	cycles;
cycles             71 drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c 	struct mlx5_clock *clock = container_of(cc, struct mlx5_clock, cycles);
cycles             91 drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c 	clock_info->cycles = clock->tc.cycle_last;
cycles             92 drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c 	clock_info->mult   = clock->cycles.mult;
cycles            152 drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c 	timecounter_init(&clock->tc, &clock->cycles, ns);
cycles            167 drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c 	u64 cycles, ns;
cycles            170 drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c 	cycles = mlx5_read_internal_timer(mdev, sts);
cycles            171 drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c 	ns = timecounter_cyc2time(&clock->tc, cycles);
cycles            213 drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c 	clock->cycles.mult = neg_adj ? clock->nominal_c_mult - diff :
cycles            335 drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c 		cycles_delta = div64_u64(nsec_delta << clock->cycles.shift,
cycles            336 drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c 					 clock->cycles.mult);
cycles            501 drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c 		cycles_delta = div64_u64(nsec_delta << clock->cycles.shift,
cycles            502 drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c 					 clock->cycles.mult);
cycles            529 drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c 	clock->cycles.read = read_internal_timer;
cycles            530 drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c 	clock->cycles.shift = MLX5_CYCLES_SHIFT;
cycles            531 drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c 	clock->cycles.mult = clocksource_khz2mult(dev_freq,
cycles            532 drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c 						  clock->cycles.shift);
cycles            533 drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c 	clock->nominal_c_mult = clock->cycles.mult;
cycles            534 drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c 	clock->cycles.mask = CLOCKSOURCE_MASK(41);
cycles            537 drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c 	timecounter_init(&clock->tc, &clock->cycles,
cycles            547 drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c 	overflow_cycles = div64_u64(~0ULL >> 1, clock->cycles.mult);
cycles            548 drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c 	overflow_cycles = min(overflow_cycles, div_u64(clock->cycles.mask, 3));
cycles            550 drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c 	ns = cyclecounter_cyc2ns(&clock->cycles, overflow_cycles,
cycles            559 drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c 		mdev->clock_info->cycles = clock->tc.cycle_last;
cycles            560 drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c 		mdev->clock_info->mask = clock->cycles.mask;
cycles            562 drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c 		mdev->clock_info->shift = clock->cycles.shift;
cycles             64 drivers/net/ethernet/mellanox/mlxsw/spectrum_ptp.c 	struct cyclecounter cycles;
cycles             98 drivers/net/ethernet/mellanox/mlxsw/spectrum_ptp.c 		container_of(cc, struct mlxsw_sp_ptp_clock, cycles);
cycles            116 drivers/net/ethernet/mellanox/mlxsw/spectrum_ptp.c 	u64 cycles = (u64) nsec;
cycles            118 drivers/net/ethernet/mellanox/mlxsw/spectrum_ptp.c 	cycles <<= tc->cc->shift;
cycles            119 drivers/net/ethernet/mellanox/mlxsw/spectrum_ptp.c 	cycles = div_u64(cycles, tc->cc->mult);
cycles            121 drivers/net/ethernet/mellanox/mlxsw/spectrum_ptp.c 	return cycles;
cycles            128 drivers/net/ethernet/mellanox/mlxsw/spectrum_ptp.c 	u64 next_sec, next_sec_in_nsec, cycles;
cycles            137 drivers/net/ethernet/mellanox/mlxsw/spectrum_ptp.c 	cycles = mlxsw_sp1_ptp_ns2cycles(&clock->tc, next_sec_in_nsec);
cycles            140 drivers/net/ethernet/mellanox/mlxsw/spectrum_ptp.c 	mlxsw_reg_mtpps_vpin_pack(mtpps_pl, cycles);
cycles            173 drivers/net/ethernet/mellanox/mlxsw/spectrum_ptp.c 	clock->cycles.mult = neg_adj ? clock->nominal_c_mult - diff :
cycles            200 drivers/net/ethernet/mellanox/mlxsw/spectrum_ptp.c 	u64 cycles, nsec;
cycles            203 drivers/net/ethernet/mellanox/mlxsw/spectrum_ptp.c 	cycles = __mlxsw_sp1_ptp_read_frc(clock, sts);
cycles            204 drivers/net/ethernet/mellanox/mlxsw/spectrum_ptp.c 	nsec = timecounter_cyc2time(&clock->tc, cycles);
cycles            220 drivers/net/ethernet/mellanox/mlxsw/spectrum_ptp.c 	timecounter_init(&clock->tc, &clock->cycles, nsec);
cycles            262 drivers/net/ethernet/mellanox/mlxsw/spectrum_ptp.c 	clock->cycles.read = mlxsw_sp1_ptp_read_frc;
cycles            263 drivers/net/ethernet/mellanox/mlxsw/spectrum_ptp.c 	clock->cycles.shift = MLXSW_SP1_PTP_CLOCK_CYCLES_SHIFT;
cycles            264 drivers/net/ethernet/mellanox/mlxsw/spectrum_ptp.c 	clock->cycles.mult = clocksource_khz2mult(MLXSW_SP1_PTP_CLOCK_FREQ_KHZ,
cycles            265 drivers/net/ethernet/mellanox/mlxsw/spectrum_ptp.c 						  clock->cycles.shift);
cycles            266 drivers/net/ethernet/mellanox/mlxsw/spectrum_ptp.c 	clock->nominal_c_mult = clock->cycles.mult;
cycles            267 drivers/net/ethernet/mellanox/mlxsw/spectrum_ptp.c 	clock->cycles.mask = CLOCKSOURCE_MASK(MLXSW_SP1_PTP_CLOCK_MASK);
cycles            270 drivers/net/ethernet/mellanox/mlxsw/spectrum_ptp.c 	timecounter_init(&clock->tc, &clock->cycles,
cycles            280 drivers/net/ethernet/mellanox/mlxsw/spectrum_ptp.c 	overflow_cycles = div64_u64(~0ULL >> 1, clock->cycles.mult);
cycles            281 drivers/net/ethernet/mellanox/mlxsw/spectrum_ptp.c 	overflow_cycles = min(overflow_cycles, div_u64(clock->cycles.mask, 3));
cycles            283 drivers/net/ethernet/mellanox/mlxsw/spectrum_ptp.c 	nsec = cyclecounter_cyc2ns(&clock->cycles, overflow_cycles, 0, &frac);
cycles            368 drivers/net/ethernet/mellanox/mlxsw/spectrum_ptp.c 	int cycles = MLXSW_SP1_PTP_HT_GC_TIMEOUT / MLXSW_SP1_PTP_HT_GC_INTERVAL;
cycles            380 drivers/net/ethernet/mellanox/mlxsw/spectrum_ptp.c 	unmatched->gc_cycle = mlxsw_sp->ptp_state->gc_cycle + cycles;
cycles            441 drivers/net/ethernet/xscale/ixp4xx_eth.c 	int cycles = 0;
cycles            457 drivers/net/ethernet/xscale/ixp4xx_eth.c 	while ((cycles < MAX_MDIO_RETRIES) &&
cycles            460 drivers/net/ethernet/xscale/ixp4xx_eth.c 		cycles++;
cycles            463 drivers/net/ethernet/xscale/ixp4xx_eth.c 	if (cycles == MAX_MDIO_RETRIES) {
cycles            471 drivers/net/ethernet/xscale/ixp4xx_eth.c 	       phy_id, write ? "write" : "read", cycles);
cycles             46 drivers/net/wireless/ath/ath.h 	u32 cycles;
cycles            728 drivers/net/wireless/ath/ath5k/debug.c 	_struct.cycles > 0 ? \
cycles            729 drivers/net/wireless/ath/ath5k/debug.c 	_struct._field * 100 / _struct.cycles : 0
cycles            742 drivers/net/wireless/ath/ath5k/debug.c 			as->last_cc.cycles);
cycles            660 drivers/net/wireless/ath/ath5k/mac80211-ops.c 	if (cc->cycles > 0) {
cycles            661 drivers/net/wireless/ath/ath5k/mac80211-ops.c 		ah->survey.time += cc->cycles / div;
cycles           2151 drivers/net/wireless/ath/ath9k/ar9003_phy.c #define PCT(_field) (common->cc_survey._field * 100 / common->cc_survey.cycles)
cycles           2152 drivers/net/wireless/ath/ath9k/ar9003_phy.c 	if (common->cc_survey.cycles)
cycles            522 drivers/net/wireless/ath/ath9k/link.c 	if (cc->cycles > 0) {
cycles            527 drivers/net/wireless/ath/ath9k/link.c 		survey->time += cc->cycles / div;
cycles            533 drivers/net/wireless/ath/ath9k/link.c 	if (cc->cycles < div)
cycles            536 drivers/net/wireless/ath/ath9k/link.c 	if (cc->cycles > 0)
cycles            537 drivers/net/wireless/ath/ath9k/link.c 		ret = cc->rx_busy * 100 / cc->cycles;
cycles            144 drivers/net/wireless/ath/hw.c 	u32 cycles, busy, rx, tx;
cycles            151 drivers/net/wireless/ath/hw.c 	cycles = REG_READ(ah, AR_CCCNT);
cycles            166 drivers/net/wireless/ath/hw.c 	common->cc_ani.cycles += cycles;
cycles            171 drivers/net/wireless/ath/hw.c 	common->cc_survey.cycles += cycles;
cycles            183 drivers/net/wireless/ath/hw.c 	listen_time = (cc->cycles - cc->rx_frame - cc->tx_frame) /
cycles            329 drivers/net/wireless/ti/wl12xx/scan.c 	cfg->cycles = 0;
cycles             75 drivers/net/wireless/ti/wl12xx/scan.h 	u8 cycles;       /* maximum number of scan cycles */
cycles             62 drivers/pcmcia/yenta_socket.c #define to_ns(cycles)	((cycles)*120)
cycles            253 drivers/perf/arm-cci.c 	CCI400_CYCLE_EVENT_EXT_ATTR_ENTRY(cycles, 0xff),
cycles            300 drivers/perf/arm-cci.c 	CCI400_CYCLE_EVENT_EXT_ATTR_ENTRY(cycles, 0xff),
cycles            441 drivers/perf/arm-ccn.c 	CCN_EVENT_CYCLES(cycles),
cycles            186 drivers/perf/arm_dsu_pmu.c 	DSU_EVENT_ATTR(cycles, 0x11),
cycles            513 drivers/perf/arm_smmuv3_pmu.c SMMU_EVENT_ATTR(cycles, 0);
cycles            119 drivers/perf/fsl_imx8_ddr_perf.c 	IMX8_DDR_PMU_EVENT_ATTR(cycles, EVENT_CYCLES_ID),
cycles            137 drivers/perf/fsl_imx8_ddr_perf.c 	IMX8_DDR_PMU_EVENT_ATTR(read-cycles, 0x2a),
cycles            138 drivers/perf/fsl_imx8_ddr_perf.c 	IMX8_DDR_PMU_EVENT_ATTR(write-cycles, 0x2b),
cycles            731 drivers/perf/qcom_l2_pmu.c 	L2CACHE_EVENT_ATTR(cycles, L2_EVENT_CYCLES),
cycles            656 drivers/perf/qcom_l3_pmu.c 	L3CACHE_EVENT_ATTR(cycles, L3_EVENT_CYCLES),
cycles            458 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(pa-barrier-cycles,			0x1d),
cycles             75 drivers/ptp/ptp_kvm.c 	system_counter->cycles = ret;
cycles            117 drivers/pwm/pwm-atmel.c 	unsigned long long cycles = state->period;
cycles            120 drivers/pwm/pwm-atmel.c 	cycles *= clk_get_rate(atmel_pwm->clk);
cycles            121 drivers/pwm/pwm-atmel.c 	do_div(cycles, NSEC_PER_SEC);
cycles            123 drivers/pwm/pwm-atmel.c 	for (*pres = 0; cycles > atmel_pwm->data->cfg.max_period; cycles >>= 1)
cycles            131 drivers/pwm/pwm-atmel.c 	*cprd = cycles;
cycles            139 drivers/pwm/pwm-atmel.c 	unsigned long long cycles = state->duty_cycle;
cycles            141 drivers/pwm/pwm-atmel.c 	cycles *= cprd;
cycles            142 drivers/pwm/pwm-atmel.c 	do_div(cycles, state->period);
cycles            143 drivers/pwm/pwm-atmel.c 	*cdty = cprd - cycles;
cycles             96 drivers/pwm/pwm-berlin.c 	u64 cycles;
cycles             98 drivers/pwm/pwm-berlin.c 	cycles = clk_get_rate(pwm->clk);
cycles             99 drivers/pwm/pwm-berlin.c 	cycles *= period_ns;
cycles            100 drivers/pwm/pwm-berlin.c 	do_div(cycles, NSEC_PER_SEC);
cycles            102 drivers/pwm/pwm-berlin.c 	if (cycles > BERLIN_PWM_MAX_TCNT) {
cycles            104 drivers/pwm/pwm-berlin.c 		cycles >>= 12; // Prescaled by 4096
cycles            106 drivers/pwm/pwm-berlin.c 		if (cycles > BERLIN_PWM_MAX_TCNT)
cycles            110 drivers/pwm/pwm-berlin.c 	period = cycles;
cycles            111 drivers/pwm/pwm-berlin.c 	cycles *= duty_ns;
cycles            112 drivers/pwm/pwm-berlin.c 	do_div(cycles, period_ns);
cycles            113 drivers/pwm/pwm-berlin.c 	duty = cycles;
cycles             34 drivers/siox/siox-bus-gpio.c 	size_t cycles = max(setbuf_len, getbuf_len);
cycles             44 drivers/siox/siox-bus-gpio.c 	for (i = 0; i < cycles; ++i) {
cycles             48 drivers/siox/siox-bus-gpio.c 		if (i >= cycles - setbuf_len)
cycles             49 drivers/siox/siox-bus-gpio.c 			set = setbuf[i - (cycles - setbuf_len)];
cycles            408 drivers/soc/ixp4xx/ixp4xx-npe.c 	int cycles = 0;
cycles            427 drivers/soc/ixp4xx/ixp4xx-npe.c 	while ((cycles < MAX_RETRIES) &&
cycles            430 drivers/soc/ixp4xx/ixp4xx-npe.c 		cycles++;
cycles            433 drivers/soc/ixp4xx/ixp4xx-npe.c 	if (cycles == MAX_RETRIES) {
cycles            439 drivers/soc/ixp4xx/ixp4xx-npe.c 	debug_msg(npe, "Sending a message took %i cycles\n", cycles);
cycles            447 drivers/soc/ixp4xx/ixp4xx-npe.c 	int cycles = 0, cnt = 0;
cycles            451 drivers/soc/ixp4xx/ixp4xx-npe.c 	while (cycles < MAX_RETRIES) {
cycles            458 drivers/soc/ixp4xx/ixp4xx-npe.c 			cycles++;
cycles            471 drivers/soc/ixp4xx/ixp4xx-npe.c 	if (cycles == MAX_RETRIES) {
cycles            477 drivers/soc/ixp4xx/ixp4xx-npe.c 	debug_msg(npe, "Receiving a message took %i cycles\n", cycles);
cycles            488 drivers/staging/media/imx/imx-media-csi.c 			passthrough_cycles = incc->cycles;
cycles            730 drivers/staging/media/imx/imx-media-csi.c 	if (is_parallel_bus(&priv->upstream_ep) && incc->cycles) {
cycles            731 drivers/staging/media/imx/imx-media-csi.c 		if_fmt.width *= incc->cycles;
cycles            732 drivers/staging/media/imx/imx-media-csi.c 		crop.width *= incc->cycles;
cycles             77 drivers/staging/media/imx/imx-media-utils.c 		.cycles = 2,
cycles             75 drivers/staging/media/imx/imx-media.h 	int	cycles;
cycles            270 drivers/tty/serial/serial-tegra.c 				       unsigned int cycles)
cycles            273 drivers/tty/serial/serial-tegra.c 		udelay(DIV_ROUND_UP(cycles * 1000000, tup->current_baud * 16));
cycles             74 drivers/watchdog/renesas_wdt.c static void rwdt_wait_cycles(struct rwdt_priv *priv, unsigned int cycles)
cycles             78 drivers/watchdog/renesas_wdt.c 	delay = DIV_ROUND_UP(cycles * 1000000, priv->clk_rate);
cycles            184 include/linux/clocksource.h static inline s64 clocksource_cyc2ns(u64 cycles, u32 mult, u32 shift)
cycles            186 include/linux/clocksource.h 	return ((u64) cycles * mult) >> shift;
cycles            654 include/linux/mlx5/driver.h 	struct cyclecounter        cycles;
cycles            634 include/linux/mtd/rawnand.h #define NAND_OP_ADDR(ncycles, cycles, ns)				\
cycles            639 include/linux/mtd/rawnand.h 			.addrs = cycles,				\
cycles             71 include/linux/timecounter.h 				      u64 cycles, u64 mask, u64 *frac)
cycles             73 include/linux/timecounter.h 	u64 ns = (u64) cycles;
cycles            235 include/linux/timekeeping.h 	u64		cycles;
cycles            263 include/linux/timekeeping.h 	u64			cycles;
cycles           1177 include/uapi/linux/perf_event.h 		cycles:16,  /* cycle count to last branch */
cycles            444 include/uapi/rdma/mlx5-abi.h 	__aligned_u64 cycles;
cycles            277 kernel/locking/test-ww_mutex.c 	struct test_cycle *cycles;
cycles            281 kernel/locking/test-ww_mutex.c 	cycles = kmalloc_array(nthreads, sizeof(*cycles), GFP_KERNEL);
cycles            282 kernel/locking/test-ww_mutex.c 	if (!cycles)
cycles            286 kernel/locking/test-ww_mutex.c 		struct test_cycle *cycle = &cycles[n];
cycles            290 kernel/locking/test-ww_mutex.c 			cycle->b_mutex = &cycles[0].a_mutex;
cycles            292 kernel/locking/test-ww_mutex.c 			cycle->b_mutex = &cycles[n + 1].a_mutex;
cycles            295 kernel/locking/test-ww_mutex.c 			cycle->a_signal = &cycles[last].b_signal;
cycles            297 kernel/locking/test-ww_mutex.c 			cycle->a_signal = &cycles[n - 1].b_signal;
cycles            305 kernel/locking/test-ww_mutex.c 		queue_work(wq, &cycles[n].work);
cycles            311 kernel/locking/test-ww_mutex.c 		struct test_cycle *cycle = &cycles[n];
cycles            323 kernel/locking/test-ww_mutex.c 		ww_mutex_destroy(&cycles[n].a_mutex);
cycles            324 kernel/locking/test-ww_mutex.c 	kfree(cycles);
cycles             70 kernel/time/timecounter.c 			       u64 cycles, u64 mask, u64 frac)
cycles             72 kernel/time/timecounter.c 	u64 ns = (u64) cycles;
cycles            378 kernel/time/timekeeping.c static inline u64 timekeeping_cycles_to_ns(const struct tk_read_base *tkr, u64 cycles)
cycles            383 kernel/time/timekeeping.c 	delta = clocksource_delta(cycles, tkr->cycle_last, tkr->mask);
cycles            991 kernel/time/timekeeping.c 	systime_snapshot->cycles = now;
cycles           1127 kernel/time/timekeeping.c 	u64 cycles, now, interval_start;
cycles           1153 kernel/time/timekeeping.c 		cycles = system_counterval.cycles;
cycles           1161 kernel/time/timekeeping.c 		if (!cycle_between(interval_start, cycles, now)) {
cycles           1164 kernel/time/timekeeping.c 			cycles = interval_start;
cycles           1175 kernel/time/timekeeping.c 						     system_counterval.cycles);
cycles           1177 kernel/time/timekeeping.c 						    system_counterval.cycles);
cycles           1197 kernel/time/timekeeping.c 		    !cycle_between(history_begin->cycles,
cycles           1198 kernel/time/timekeeping.c 				   system_counterval.cycles, cycles) ||
cycles           1201 kernel/time/timekeeping.c 		partial_history_cycles = cycles - system_counterval.cycles;
cycles           1202 kernel/time/timekeeping.c 		total_history_cycles = cycles - history_begin->cycles;
cycles             35 lib/vdso/gettimeofday.c u64 vdso_calc_delta(u64 cycles, u64 last, u64 mask, u32 mult)
cycles             37 lib/vdso/gettimeofday.c 	return ((cycles - last) & mask) * mult;
cycles             45 lib/vdso/gettimeofday.c 	u64 cycles, last, sec, ns;
cycles             50 lib/vdso/gettimeofday.c 		cycles = __arch_get_hw_counter(vd->clock_mode);
cycles             53 lib/vdso/gettimeofday.c 		if (unlikely((s64)cycles < 0))
cycles             56 lib/vdso/gettimeofday.c 		ns += vdso_calc_delta(cycles, last, vd->mask, vd->mult);
cycles             17 sound/firewire/amdtp-stream-trace.h 	TP_PROTO(const struct amdtp_stream *s, u32 cycles, const __be32 *cip_header, unsigned int payload_length, unsigned int data_blocks, unsigned int data_block_counter, unsigned int index),
cycles             18 sound/firewire/amdtp-stream-trace.h 	TP_ARGS(s, cycles, cip_header, payload_length, data_blocks, data_block_counter, index),
cycles             34 sound/firewire/amdtp-stream-trace.h 		__entry->second = cycles / CYCLES_PER_SECOND;
cycles             35 sound/firewire/amdtp-stream-trace.h 		__entry->cycle = cycles % CYCLES_PER_SECOND;
cycles            253 sound/soc/intel/haswell/sst-haswell-ipc.h 	u64 cycles;
cycles           1177 tools/include/uapi/linux/perf_event.h 		cycles:16,  /* cycle count to last branch */
cycles            578 tools/perf/builtin-diff.c 	l = llabs(left->diff.cycles);
cycles            579 tools/perf/builtin-diff.c 	r = llabs(right->diff.cycles);
cycles            609 tools/perf/builtin-diff.c 	bi->cycles = ch->cycles;
cycles            697 tools/perf/builtin-diff.c 		pair->diff.cycles =
cycles           1360 tools/perf/builtin-diff.c 			  start_line, end_line, block_he->diff.cycles);
cycles           1363 tools/perf/builtin-diff.c 			  bi->start, bi->end, block_he->diff.cycles);
cycles            774 tools/perf/builtin-script.c 			br->entries[i].flags.cycles);
cycles            819 tools/perf/builtin-script.c 			br->entries[i].flags.cycles);
cycles            868 tools/perf/builtin-script.c 			br->entries[i].flags.cycles);
cycles            961 tools/perf/builtin-script.c 	if (en->flags.cycles) {
cycles            962 tools/perf/builtin-script.c 		*total_cycles += en->flags.cycles;
cycles            963 tools/perf/builtin-script.c 		printed += fprintf(fp, " %d cycles [%d]", en->flags.cycles, *total_cycles);
cycles            965 tools/perf/builtin-script.c 			printed += fprintf(fp, " %.2f IPC", (float)insn / en->flags.cycles);
cycles            819 tools/perf/util/annotate.c 				    unsigned offset, unsigned cycles,
cycles            831 tools/perf/util/annotate.c 	ch[offset].cycles_aggr += cycles;
cycles            833 tools/perf/util/annotate.c 	if (cycles > ch[offset].cycles_max)
cycles            834 tools/perf/util/annotate.c 		ch[offset].cycles_max = cycles;
cycles            837 tools/perf/util/annotate.c 		if (cycles && cycles < ch[offset].cycles_min)
cycles            838 tools/perf/util/annotate.c 			ch[offset].cycles_min = cycles;
cycles            840 tools/perf/util/annotate.c 		ch[offset].cycles_min = cycles;
cycles            848 tools/perf/util/annotate.c 			ch[offset].cycles = 0;
cycles            858 tools/perf/util/annotate.c 	ch[offset].cycles += cycles;
cycles            951 tools/perf/util/annotate.c 				  struct symbol *sym, unsigned cycles)
cycles            973 tools/perf/util/annotate.c 					offset, cycles,
cycles            979 tools/perf/util/annotate.c 				    unsigned cycles)
cycles            984 tools/perf/util/annotate.c 	if (!cycles)
cycles           1005 tools/perf/util/annotate.c 	err = symbol__account_cycles(ams->al_addr, saddr, ams->sym, cycles);
cycles           1030 tools/perf/util/annotate.c 	if (n_insn && ch->num && ch->cycles) {
cycles           1031 tools/perf/util/annotate.c 		float ipc = n_insn / ((double)ch->cycles / (double)ch->num);
cycles           1047 tools/perf/util/annotate.c 			notes->hit_cycles += ch->cycles;
cycles           1071 tools/perf/util/annotate.c 		if (ch && ch->cycles) {
cycles           1078 tools/perf/util/annotate.c 				al->cycles = ch->cycles_aggr / ch->num_aggr;
cycles           2827 tools/perf/util/annotate.c 			if (al->ipc == 0.0 && al->cycles == 0)
cycles           2872 tools/perf/util/annotate.c 			if (al->cycles)
cycles           2874 tools/perf/util/annotate.c 					   ANNOTATION__CYCLES_WIDTH - 1, al->cycles);
cycles           2883 tools/perf/util/annotate.c 			if (al->cycles) {
cycles           2888 tools/perf/util/annotate.c 					al->cycles, al->cycles_min,
cycles            138 tools/perf/util/annotate.h 	u64			 cycles;
cycles            234 tools/perf/util/annotate.h 	u64	cycles;
cycles            342 tools/perf/util/annotate.h 				    unsigned cycles);
cycles             21 tools/perf/util/branch.h 	u64 cycles:16;
cycles            615 tools/perf/util/callchain.c 					cursor_node->branch_flags.cycles;
cycles            773 tools/perf/util/callchain.c 			cnode->cycles_count += node->branch_flags.cycles;
cycles           1359 tools/perf/util/callchain.c 	u64 cycles, v = 0;
cycles           1361 tools/perf/util/callchain.c 	cycles = cycles_count / branch_count;
cycles           1362 tools/perf/util/callchain.c 	if (cycles) {
cycles           1364 tools/perf/util/callchain.c 				cycles,
cycles            916 tools/perf/util/hist.c 	sample->weight = bi->flags.cycles ? bi->flags.cycles : 1;
cycles           2577 tools/perf/util/hist.c 	if (bs && bs->nr && bs->entries[0].flags.cycles) {
cycles           2597 tools/perf/util/hist.c 					bi[i].flags.cycles);
cycles           1665 tools/perf/util/intel-pt.c 			.cycles		= info & LBR_INFO_CYCLES,
cycles           2002 tools/perf/util/machine.c 	u64 cycles;
cycles           2070 tools/perf/util/machine.c 		iter_cycles = iter->cycles;
cycles           2103 tools/perf/util/machine.c 	iter->cycles = 0;
cycles           2106 tools/perf/util/machine.c 		iter->cycles += be[i].flags.cycles;
cycles            499 tools/perf/util/scripting-engines/trace-event-python.c 		    PyLong_FromUnsignedLongLong(br->entries[i].flags.cycles));
cycles           1078 tools/perf/util/session.c 				(unsigned short)e->flags.cycles,
cycles            980 tools/perf/util/sort.c 	return left->branch_info->flags.cycles -
cycles            981 tools/perf/util/sort.c 		right->branch_info->flags.cycles;
cycles            989 tools/perf/util/sort.c 	if (he->branch_info->flags.cycles == 0)
cycles            992 tools/perf/util/sort.c 			       he->branch_info->flags.cycles);
cycles             72 tools/perf/util/sort.h 		s64	cycles;
cycles            700 tools/perf/util/stat-shadow.c 	double smi_num, aperf, cycles, cost = 0.0;
cycles            706 tools/perf/util/stat-shadow.c 	cycles = runtime_stat_avg(st, STAT_CYCLES, ctx, cpu);
cycles            708 tools/perf/util/stat-shadow.c 	if ((cycles == 0) || (aperf == 0))
cycles            712 tools/perf/util/stat-shadow.c 		cost = (aperf - cycles) / aperf * 100.00;
cycles             89 tools/perf/util/stat.c 	ID(CYCLES_IN_TX,	cpu/cycles-t/),
cycles             92 tools/perf/util/stat.c 	ID(CYCLES_IN_TX_CP,	cpu/cycles-ct/),
cycles            112 tools/perf/util/symbol.h 	u64			cycles;
cycles            113 tools/power/cpupower/bench/benchmark.c 			       " for %lius\n", _round + 1, config->cycles,
cycles            125 tools/power/cpupower/bench/benchmark.c 		for (cycle = 0; cycle < config->cycles; cycle++) {
cycles            139 tools/power/cpupower/bench/benchmark.c 			performance_time / config->cycles);
cycles            151 tools/power/cpupower/bench/benchmark.c 		for (cycle = 0; cycle < config->cycles; cycle++) {
cycles            169 tools/power/cpupower/bench/benchmark.c 			powersave_time / config->cycles);
cycles            122 tools/power/cpupower/bench/main.c 			sscanf(optarg, "%u", &config->cycles);
cycles            173 tools/power/cpupower/bench/main.c 		       config->cycles,
cycles            130 tools/power/cpupower/bench/parse.c 	config->cycles = 5;
cycles            196 tools/power/cpupower/bench/parse.c 			sscanf(val, "%u", &config->cycles);
cycles             16 tools/power/cpupower/bench/parse.h 	unsigned int cycles;	/* calculation cycles with the same sleep/load time */
cycles            133 tools/power/cpupower/bench/system.c 		sleep_time +=  2 * config->cycles *
cycles            135 tools/power/cpupower/bench/system.c 		load_time += 2 * config->cycles *
cycles             59 tools/testing/selftests/powerpc/pmu/ebb/cycles_test.c 	return test_harness(cycles, "cycles");
cycles             20 tools/virtio/ringtest/main.h static inline void wait_cycles(unsigned long long cycles)
cycles             25 tools/virtio/ringtest/main.h 	while (__rdtsc() - t < cycles) {}
cycles             32 tools/virtio/ringtest/main.h static inline void wait_cycles(unsigned long long cycles)
cycles             34 tools/virtio/ringtest/main.h 	asm volatile("0: brctg %0,0b" : : "d" (cycles));
cycles             42 tools/virtio/ringtest/main.h static inline void wait_cycles(unsigned long long cycles)