This source file includes following definitions.
- enabled_wait
- show_idle_count
- show_idle_time
- arch_cpu_idle_time
- arch_cpu_idle_enter
- arch_cpu_idle
- arch_cpu_idle_exit
- arch_cpu_idle_dead
1
2
3
4
5
6
7
8
9
10 #include <linux/kernel.h>
11 #include <linux/kernel_stat.h>
12 #include <linux/kprobes.h>
13 #include <linux/notifier.h>
14 #include <linux/init.h>
15 #include <linux/cpu.h>
16 #include <linux/sched/cputime.h>
17 #include <asm/nmi.h>
18 #include <asm/smp.h>
19 #include "entry.h"
20
21 static DEFINE_PER_CPU(struct s390_idle_data, s390_idle);
22
23 void enabled_wait(void)
24 {
25 struct s390_idle_data *idle = this_cpu_ptr(&s390_idle);
26 unsigned long long idle_time;
27 unsigned long psw_mask;
28
29 trace_hardirqs_on();
30
31
32 psw_mask = PSW_KERNEL_BITS | PSW_MASK_WAIT | PSW_MASK_DAT |
33 PSW_MASK_IO | PSW_MASK_EXT | PSW_MASK_MCHECK;
34 clear_cpu_flag(CIF_NOHZ_DELAY);
35
36
37 psw_idle(idle, psw_mask);
38
39 trace_hardirqs_off();
40
41
42 write_seqcount_begin(&idle->seqcount);
43 idle_time = idle->clock_idle_exit - idle->clock_idle_enter;
44 idle->clock_idle_enter = idle->clock_idle_exit = 0ULL;
45 idle->idle_time += idle_time;
46 idle->idle_count++;
47 account_idle_time(cputime_to_nsecs(idle_time));
48 write_seqcount_end(&idle->seqcount);
49 }
50 NOKPROBE_SYMBOL(enabled_wait);
51
52 static ssize_t show_idle_count(struct device *dev,
53 struct device_attribute *attr, char *buf)
54 {
55 struct s390_idle_data *idle = &per_cpu(s390_idle, dev->id);
56 unsigned long long idle_count;
57 unsigned int seq;
58
59 do {
60 seq = read_seqcount_begin(&idle->seqcount);
61 idle_count = READ_ONCE(idle->idle_count);
62 if (READ_ONCE(idle->clock_idle_enter))
63 idle_count++;
64 } while (read_seqcount_retry(&idle->seqcount, seq));
65 return sprintf(buf, "%llu\n", idle_count);
66 }
67 DEVICE_ATTR(idle_count, 0444, show_idle_count, NULL);
68
69 static ssize_t show_idle_time(struct device *dev,
70 struct device_attribute *attr, char *buf)
71 {
72 unsigned long long now, idle_time, idle_enter, idle_exit, in_idle;
73 struct s390_idle_data *idle = &per_cpu(s390_idle, dev->id);
74 unsigned int seq;
75
76 do {
77 seq = read_seqcount_begin(&idle->seqcount);
78 idle_time = READ_ONCE(idle->idle_time);
79 idle_enter = READ_ONCE(idle->clock_idle_enter);
80 idle_exit = READ_ONCE(idle->clock_idle_exit);
81 } while (read_seqcount_retry(&idle->seqcount, seq));
82 in_idle = 0;
83 now = get_tod_clock();
84 if (idle_enter) {
85 if (idle_exit) {
86 in_idle = idle_exit - idle_enter;
87 } else if (now > idle_enter) {
88 in_idle = now - idle_enter;
89 }
90 }
91 idle_time += in_idle;
92 return sprintf(buf, "%llu\n", idle_time >> 12);
93 }
94 DEVICE_ATTR(idle_time_us, 0444, show_idle_time, NULL);
95
96 u64 arch_cpu_idle_time(int cpu)
97 {
98 struct s390_idle_data *idle = &per_cpu(s390_idle, cpu);
99 unsigned long long now, idle_enter, idle_exit, in_idle;
100 unsigned int seq;
101
102 do {
103 seq = read_seqcount_begin(&idle->seqcount);
104 idle_enter = READ_ONCE(idle->clock_idle_enter);
105 idle_exit = READ_ONCE(idle->clock_idle_exit);
106 } while (read_seqcount_retry(&idle->seqcount, seq));
107 in_idle = 0;
108 now = get_tod_clock();
109 if (idle_enter) {
110 if (idle_exit) {
111 in_idle = idle_exit - idle_enter;
112 } else if (now > idle_enter) {
113 in_idle = now - idle_enter;
114 }
115 }
116 return cputime_to_nsecs(in_idle);
117 }
118
119 void arch_cpu_idle_enter(void)
120 {
121 local_mcck_disable();
122 }
123
124 void arch_cpu_idle(void)
125 {
126 if (!test_cpu_flag(CIF_MCCK_PENDING))
127
128 enabled_wait();
129 local_irq_enable();
130 }
131
132 void arch_cpu_idle_exit(void)
133 {
134 local_mcck_enable();
135 if (test_cpu_flag(CIF_MCCK_PENDING))
136 s390_handle_mcck();
137 }
138
139 void arch_cpu_idle_dead(void)
140 {
141 cpu_die();
142 }