1#include <linux/module.h> 2#include <linux/preempt.h> 3#include <linux/smp.h> 4#include <asm/msr.h> 5 6static void __rdmsr_on_cpu(void *info) 7{ 8 struct msr_info *rv = info; 9 struct msr *reg; 10 int this_cpu = raw_smp_processor_id(); 11 12 if (rv->msrs) 13 reg = per_cpu_ptr(rv->msrs, this_cpu); 14 else 15 reg = &rv->reg; 16 17 rdmsr(rv->msr_no, reg->l, reg->h); 18} 19 20static void __wrmsr_on_cpu(void *info) 21{ 22 struct msr_info *rv = info; 23 struct msr *reg; 24 int this_cpu = raw_smp_processor_id(); 25 26 if (rv->msrs) 27 reg = per_cpu_ptr(rv->msrs, this_cpu); 28 else 29 reg = &rv->reg; 30 31 wrmsr(rv->msr_no, reg->l, reg->h); 32} 33 34int rdmsr_on_cpu(unsigned int cpu, u32 msr_no, u32 *l, u32 *h) 35{ 36 int err; 37 struct msr_info rv; 38 39 memset(&rv, 0, sizeof(rv)); 40 41 rv.msr_no = msr_no; 42 err = smp_call_function_single(cpu, __rdmsr_on_cpu, &rv, 1); 43 *l = rv.reg.l; 44 *h = rv.reg.h; 45 46 return err; 47} 48EXPORT_SYMBOL(rdmsr_on_cpu); 49 50int rdmsrl_on_cpu(unsigned int cpu, u32 msr_no, u64 *q) 51{ 52 int err; 53 struct msr_info rv; 54 55 memset(&rv, 0, sizeof(rv)); 56 57 rv.msr_no = msr_no; 58 err = smp_call_function_single(cpu, __rdmsr_on_cpu, &rv, 1); 59 *q = rv.reg.q; 60 61 return err; 62} 63EXPORT_SYMBOL(rdmsrl_on_cpu); 64 65int wrmsr_on_cpu(unsigned int cpu, u32 msr_no, u32 l, u32 h) 66{ 67 int err; 68 struct msr_info rv; 69 70 memset(&rv, 0, sizeof(rv)); 71 72 rv.msr_no = msr_no; 73 rv.reg.l = l; 74 rv.reg.h = h; 75 err = smp_call_function_single(cpu, __wrmsr_on_cpu, &rv, 1); 76 77 return err; 78} 79EXPORT_SYMBOL(wrmsr_on_cpu); 80 81int wrmsrl_on_cpu(unsigned int cpu, u32 msr_no, u64 q) 82{ 83 int err; 84 struct msr_info rv; 85 86 memset(&rv, 0, sizeof(rv)); 87 88 rv.msr_no = msr_no; 89 rv.reg.q = q; 90 91 err = smp_call_function_single(cpu, __wrmsr_on_cpu, &rv, 1); 92 93 return err; 94} 95EXPORT_SYMBOL(wrmsrl_on_cpu); 96 97static void __rwmsr_on_cpus(const struct cpumask *mask, u32 msr_no, 98 struct msr *msrs, 99 void (*msr_func) (void *info)) 100{ 101 struct msr_info rv; 102 int this_cpu; 103 104 memset(&rv, 0, sizeof(rv)); 105 106 rv.msrs = msrs; 107 rv.msr_no = msr_no; 108 109 this_cpu = get_cpu(); 110 111 if (cpumask_test_cpu(this_cpu, mask)) 112 msr_func(&rv); 113 114 smp_call_function_many(mask, msr_func, &rv, 1); 115 put_cpu(); 116} 117 118/* rdmsr on a bunch of CPUs 119 * 120 * @mask: which CPUs 121 * @msr_no: which MSR 122 * @msrs: array of MSR values 123 * 124 */ 125void rdmsr_on_cpus(const struct cpumask *mask, u32 msr_no, struct msr *msrs) 126{ 127 __rwmsr_on_cpus(mask, msr_no, msrs, __rdmsr_on_cpu); 128} 129EXPORT_SYMBOL(rdmsr_on_cpus); 130 131/* 132 * wrmsr on a bunch of CPUs 133 * 134 * @mask: which CPUs 135 * @msr_no: which MSR 136 * @msrs: array of MSR values 137 * 138 */ 139void wrmsr_on_cpus(const struct cpumask *mask, u32 msr_no, struct msr *msrs) 140{ 141 __rwmsr_on_cpus(mask, msr_no, msrs, __wrmsr_on_cpu); 142} 143EXPORT_SYMBOL(wrmsr_on_cpus); 144 145/* These "safe" variants are slower and should be used when the target MSR 146 may not actually exist. */ 147static void __rdmsr_safe_on_cpu(void *info) 148{ 149 struct msr_info *rv = info; 150 151 rv->err = rdmsr_safe(rv->msr_no, &rv->reg.l, &rv->reg.h); 152} 153 154static void __wrmsr_safe_on_cpu(void *info) 155{ 156 struct msr_info *rv = info; 157 158 rv->err = wrmsr_safe(rv->msr_no, rv->reg.l, rv->reg.h); 159} 160 161int rdmsr_safe_on_cpu(unsigned int cpu, u32 msr_no, u32 *l, u32 *h) 162{ 163 int err; 164 struct msr_info rv; 165 166 memset(&rv, 0, sizeof(rv)); 167 168 rv.msr_no = msr_no; 169 err = smp_call_function_single(cpu, __rdmsr_safe_on_cpu, &rv, 1); 170 *l = rv.reg.l; 171 *h = rv.reg.h; 172 173 return err ? err : rv.err; 174} 175EXPORT_SYMBOL(rdmsr_safe_on_cpu); 176 177int wrmsr_safe_on_cpu(unsigned int cpu, u32 msr_no, u32 l, u32 h) 178{ 179 int err; 180 struct msr_info rv; 181 182 memset(&rv, 0, sizeof(rv)); 183 184 rv.msr_no = msr_no; 185 rv.reg.l = l; 186 rv.reg.h = h; 187 err = smp_call_function_single(cpu, __wrmsr_safe_on_cpu, &rv, 1); 188 189 return err ? err : rv.err; 190} 191EXPORT_SYMBOL(wrmsr_safe_on_cpu); 192 193int wrmsrl_safe_on_cpu(unsigned int cpu, u32 msr_no, u64 q) 194{ 195 int err; 196 struct msr_info rv; 197 198 memset(&rv, 0, sizeof(rv)); 199 200 rv.msr_no = msr_no; 201 rv.reg.q = q; 202 203 err = smp_call_function_single(cpu, __wrmsr_safe_on_cpu, &rv, 1); 204 205 return err ? err : rv.err; 206} 207EXPORT_SYMBOL(wrmsrl_safe_on_cpu); 208 209int rdmsrl_safe_on_cpu(unsigned int cpu, u32 msr_no, u64 *q) 210{ 211 int err; 212 struct msr_info rv; 213 214 memset(&rv, 0, sizeof(rv)); 215 216 rv.msr_no = msr_no; 217 err = smp_call_function_single(cpu, __rdmsr_safe_on_cpu, &rv, 1); 218 *q = rv.reg.q; 219 220 return err ? err : rv.err; 221} 222EXPORT_SYMBOL(rdmsrl_safe_on_cpu); 223 224/* 225 * These variants are significantly slower, but allows control over 226 * the entire 32-bit GPR set. 227 */ 228static void __rdmsr_safe_regs_on_cpu(void *info) 229{ 230 struct msr_regs_info *rv = info; 231 232 rv->err = rdmsr_safe_regs(rv->regs); 233} 234 235static void __wrmsr_safe_regs_on_cpu(void *info) 236{ 237 struct msr_regs_info *rv = info; 238 239 rv->err = wrmsr_safe_regs(rv->regs); 240} 241 242int rdmsr_safe_regs_on_cpu(unsigned int cpu, u32 *regs) 243{ 244 int err; 245 struct msr_regs_info rv; 246 247 rv.regs = regs; 248 rv.err = -EIO; 249 err = smp_call_function_single(cpu, __rdmsr_safe_regs_on_cpu, &rv, 1); 250 251 return err ? err : rv.err; 252} 253EXPORT_SYMBOL(rdmsr_safe_regs_on_cpu); 254 255int wrmsr_safe_regs_on_cpu(unsigned int cpu, u32 *regs) 256{ 257 int err; 258 struct msr_regs_info rv; 259 260 rv.regs = regs; 261 rv.err = -EIO; 262 err = smp_call_function_single(cpu, __wrmsr_safe_regs_on_cpu, &rv, 1); 263 264 return err ? err : rv.err; 265} 266EXPORT_SYMBOL(wrmsr_safe_regs_on_cpu); 267