This source file includes following definitions.
- aer_error_init
- __find_aer_error
- __find_aer_error_by_dev
- __find_pci_bus_ops
- pci_bus_ops_pop
- find_pci_config_dword
- aer_inj_read
- aer_inj_write
- aer_inj_read_config
- aer_inj_write_config
- pci_bus_ops_init
- pci_bus_set_aer_ops
- aer_inject
- aer_inject_write
- aer_inject_init
- aer_inject_exit
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15 #define dev_fmt(fmt) "aer_inject: " fmt
16
17 #include <linux/module.h>
18 #include <linux/init.h>
19 #include <linux/irq.h>
20 #include <linux/miscdevice.h>
21 #include <linux/pci.h>
22 #include <linux/slab.h>
23 #include <linux/fs.h>
24 #include <linux/uaccess.h>
25 #include <linux/stddef.h>
26 #include <linux/device.h>
27
28 #include "portdrv.h"
29
30
31 static bool aer_mask_override;
32 module_param(aer_mask_override, bool, 0);
33
34 struct aer_error_inj {
35 u8 bus;
36 u8 dev;
37 u8 fn;
38 u32 uncor_status;
39 u32 cor_status;
40 u32 header_log0;
41 u32 header_log1;
42 u32 header_log2;
43 u32 header_log3;
44 u32 domain;
45 };
46
47 struct aer_error {
48 struct list_head list;
49 u32 domain;
50 unsigned int bus;
51 unsigned int devfn;
52 int pos_cap_err;
53
54 u32 uncor_status;
55 u32 cor_status;
56 u32 header_log0;
57 u32 header_log1;
58 u32 header_log2;
59 u32 header_log3;
60 u32 root_status;
61 u32 source_id;
62 };
63
64 struct pci_bus_ops {
65 struct list_head list;
66 struct pci_bus *bus;
67 struct pci_ops *ops;
68 };
69
70 static LIST_HEAD(einjected);
71
72 static LIST_HEAD(pci_bus_ops_list);
73
74
75 static DEFINE_SPINLOCK(inject_lock);
76
77 static void aer_error_init(struct aer_error *err, u32 domain,
78 unsigned int bus, unsigned int devfn,
79 int pos_cap_err)
80 {
81 INIT_LIST_HEAD(&err->list);
82 err->domain = domain;
83 err->bus = bus;
84 err->devfn = devfn;
85 err->pos_cap_err = pos_cap_err;
86 }
87
88
89 static struct aer_error *__find_aer_error(u32 domain, unsigned int bus,
90 unsigned int devfn)
91 {
92 struct aer_error *err;
93
94 list_for_each_entry(err, &einjected, list) {
95 if (domain == err->domain &&
96 bus == err->bus &&
97 devfn == err->devfn)
98 return err;
99 }
100 return NULL;
101 }
102
103
104 static struct aer_error *__find_aer_error_by_dev(struct pci_dev *dev)
105 {
106 int domain = pci_domain_nr(dev->bus);
107 if (domain < 0)
108 return NULL;
109 return __find_aer_error(domain, dev->bus->number, dev->devfn);
110 }
111
112
113 static struct pci_ops *__find_pci_bus_ops(struct pci_bus *bus)
114 {
115 struct pci_bus_ops *bus_ops;
116
117 list_for_each_entry(bus_ops, &pci_bus_ops_list, list) {
118 if (bus_ops->bus == bus)
119 return bus_ops->ops;
120 }
121 return NULL;
122 }
123
124 static struct pci_bus_ops *pci_bus_ops_pop(void)
125 {
126 unsigned long flags;
127 struct pci_bus_ops *bus_ops;
128
129 spin_lock_irqsave(&inject_lock, flags);
130 bus_ops = list_first_entry_or_null(&pci_bus_ops_list,
131 struct pci_bus_ops, list);
132 if (bus_ops)
133 list_del(&bus_ops->list);
134 spin_unlock_irqrestore(&inject_lock, flags);
135 return bus_ops;
136 }
137
138 static u32 *find_pci_config_dword(struct aer_error *err, int where,
139 int *prw1cs)
140 {
141 int rw1cs = 0;
142 u32 *target = NULL;
143
144 if (err->pos_cap_err == -1)
145 return NULL;
146
147 switch (where - err->pos_cap_err) {
148 case PCI_ERR_UNCOR_STATUS:
149 target = &err->uncor_status;
150 rw1cs = 1;
151 break;
152 case PCI_ERR_COR_STATUS:
153 target = &err->cor_status;
154 rw1cs = 1;
155 break;
156 case PCI_ERR_HEADER_LOG:
157 target = &err->header_log0;
158 break;
159 case PCI_ERR_HEADER_LOG+4:
160 target = &err->header_log1;
161 break;
162 case PCI_ERR_HEADER_LOG+8:
163 target = &err->header_log2;
164 break;
165 case PCI_ERR_HEADER_LOG+12:
166 target = &err->header_log3;
167 break;
168 case PCI_ERR_ROOT_STATUS:
169 target = &err->root_status;
170 rw1cs = 1;
171 break;
172 case PCI_ERR_ROOT_ERR_SRC:
173 target = &err->source_id;
174 break;
175 }
176 if (prw1cs)
177 *prw1cs = rw1cs;
178 return target;
179 }
180
181 static int aer_inj_read(struct pci_bus *bus, unsigned int devfn, int where,
182 int size, u32 *val)
183 {
184 struct pci_ops *ops, *my_ops;
185 int rv;
186
187 ops = __find_pci_bus_ops(bus);
188 if (!ops)
189 return -1;
190
191 my_ops = bus->ops;
192 bus->ops = ops;
193 rv = ops->read(bus, devfn, where, size, val);
194 bus->ops = my_ops;
195
196 return rv;
197 }
198
199 static int aer_inj_write(struct pci_bus *bus, unsigned int devfn, int where,
200 int size, u32 val)
201 {
202 struct pci_ops *ops, *my_ops;
203 int rv;
204
205 ops = __find_pci_bus_ops(bus);
206 if (!ops)
207 return -1;
208
209 my_ops = bus->ops;
210 bus->ops = ops;
211 rv = ops->write(bus, devfn, where, size, val);
212 bus->ops = my_ops;
213
214 return rv;
215 }
216
217 static int aer_inj_read_config(struct pci_bus *bus, unsigned int devfn,
218 int where, int size, u32 *val)
219 {
220 u32 *sim;
221 struct aer_error *err;
222 unsigned long flags;
223 int domain;
224 int rv;
225
226 spin_lock_irqsave(&inject_lock, flags);
227 if (size != sizeof(u32))
228 goto out;
229 domain = pci_domain_nr(bus);
230 if (domain < 0)
231 goto out;
232 err = __find_aer_error(domain, bus->number, devfn);
233 if (!err)
234 goto out;
235
236 sim = find_pci_config_dword(err, where, NULL);
237 if (sim) {
238 *val = *sim;
239 spin_unlock_irqrestore(&inject_lock, flags);
240 return 0;
241 }
242 out:
243 rv = aer_inj_read(bus, devfn, where, size, val);
244 spin_unlock_irqrestore(&inject_lock, flags);
245 return rv;
246 }
247
248 static int aer_inj_write_config(struct pci_bus *bus, unsigned int devfn,
249 int where, int size, u32 val)
250 {
251 u32 *sim;
252 struct aer_error *err;
253 unsigned long flags;
254 int rw1cs;
255 int domain;
256 int rv;
257
258 spin_lock_irqsave(&inject_lock, flags);
259 if (size != sizeof(u32))
260 goto out;
261 domain = pci_domain_nr(bus);
262 if (domain < 0)
263 goto out;
264 err = __find_aer_error(domain, bus->number, devfn);
265 if (!err)
266 goto out;
267
268 sim = find_pci_config_dword(err, where, &rw1cs);
269 if (sim) {
270 if (rw1cs)
271 *sim ^= val;
272 else
273 *sim = val;
274 spin_unlock_irqrestore(&inject_lock, flags);
275 return 0;
276 }
277 out:
278 rv = aer_inj_write(bus, devfn, where, size, val);
279 spin_unlock_irqrestore(&inject_lock, flags);
280 return rv;
281 }
282
283 static struct pci_ops aer_inj_pci_ops = {
284 .read = aer_inj_read_config,
285 .write = aer_inj_write_config,
286 };
287
288 static void pci_bus_ops_init(struct pci_bus_ops *bus_ops,
289 struct pci_bus *bus,
290 struct pci_ops *ops)
291 {
292 INIT_LIST_HEAD(&bus_ops->list);
293 bus_ops->bus = bus;
294 bus_ops->ops = ops;
295 }
296
297 static int pci_bus_set_aer_ops(struct pci_bus *bus)
298 {
299 struct pci_ops *ops;
300 struct pci_bus_ops *bus_ops;
301 unsigned long flags;
302
303 bus_ops = kmalloc(sizeof(*bus_ops), GFP_KERNEL);
304 if (!bus_ops)
305 return -ENOMEM;
306 ops = pci_bus_set_ops(bus, &aer_inj_pci_ops);
307 spin_lock_irqsave(&inject_lock, flags);
308 if (ops == &aer_inj_pci_ops)
309 goto out;
310 pci_bus_ops_init(bus_ops, bus, ops);
311 list_add(&bus_ops->list, &pci_bus_ops_list);
312 bus_ops = NULL;
313 out:
314 spin_unlock_irqrestore(&inject_lock, flags);
315 kfree(bus_ops);
316 return 0;
317 }
318
319 static int aer_inject(struct aer_error_inj *einj)
320 {
321 struct aer_error *err, *rperr;
322 struct aer_error *err_alloc = NULL, *rperr_alloc = NULL;
323 struct pci_dev *dev, *rpdev;
324 struct pcie_device *edev;
325 struct device *device;
326 unsigned long flags;
327 unsigned int devfn = PCI_DEVFN(einj->dev, einj->fn);
328 int pos_cap_err, rp_pos_cap_err;
329 u32 sever, cor_mask, uncor_mask, cor_mask_orig = 0, uncor_mask_orig = 0;
330 int ret = 0;
331
332 dev = pci_get_domain_bus_and_slot(einj->domain, einj->bus, devfn);
333 if (!dev)
334 return -ENODEV;
335 rpdev = pcie_find_root_port(dev);
336 if (!rpdev) {
337 pci_err(dev, "Root port not found\n");
338 ret = -ENODEV;
339 goto out_put;
340 }
341
342 pos_cap_err = dev->aer_cap;
343 if (!pos_cap_err) {
344 pci_err(dev, "Device doesn't support AER\n");
345 ret = -EPROTONOSUPPORT;
346 goto out_put;
347 }
348 pci_read_config_dword(dev, pos_cap_err + PCI_ERR_UNCOR_SEVER, &sever);
349 pci_read_config_dword(dev, pos_cap_err + PCI_ERR_COR_MASK, &cor_mask);
350 pci_read_config_dword(dev, pos_cap_err + PCI_ERR_UNCOR_MASK,
351 &uncor_mask);
352
353 rp_pos_cap_err = rpdev->aer_cap;
354 if (!rp_pos_cap_err) {
355 pci_err(rpdev, "Root port doesn't support AER\n");
356 ret = -EPROTONOSUPPORT;
357 goto out_put;
358 }
359
360 err_alloc = kzalloc(sizeof(struct aer_error), GFP_KERNEL);
361 if (!err_alloc) {
362 ret = -ENOMEM;
363 goto out_put;
364 }
365 rperr_alloc = kzalloc(sizeof(struct aer_error), GFP_KERNEL);
366 if (!rperr_alloc) {
367 ret = -ENOMEM;
368 goto out_put;
369 }
370
371 if (aer_mask_override) {
372 cor_mask_orig = cor_mask;
373 cor_mask &= !(einj->cor_status);
374 pci_write_config_dword(dev, pos_cap_err + PCI_ERR_COR_MASK,
375 cor_mask);
376
377 uncor_mask_orig = uncor_mask;
378 uncor_mask &= !(einj->uncor_status);
379 pci_write_config_dword(dev, pos_cap_err + PCI_ERR_UNCOR_MASK,
380 uncor_mask);
381 }
382
383 spin_lock_irqsave(&inject_lock, flags);
384
385 err = __find_aer_error_by_dev(dev);
386 if (!err) {
387 err = err_alloc;
388 err_alloc = NULL;
389 aer_error_init(err, einj->domain, einj->bus, devfn,
390 pos_cap_err);
391 list_add(&err->list, &einjected);
392 }
393 err->uncor_status |= einj->uncor_status;
394 err->cor_status |= einj->cor_status;
395 err->header_log0 = einj->header_log0;
396 err->header_log1 = einj->header_log1;
397 err->header_log2 = einj->header_log2;
398 err->header_log3 = einj->header_log3;
399
400 if (!aer_mask_override && einj->cor_status &&
401 !(einj->cor_status & ~cor_mask)) {
402 ret = -EINVAL;
403 pci_warn(dev, "The correctable error(s) is masked by device\n");
404 spin_unlock_irqrestore(&inject_lock, flags);
405 goto out_put;
406 }
407 if (!aer_mask_override && einj->uncor_status &&
408 !(einj->uncor_status & ~uncor_mask)) {
409 ret = -EINVAL;
410 pci_warn(dev, "The uncorrectable error(s) is masked by device\n");
411 spin_unlock_irqrestore(&inject_lock, flags);
412 goto out_put;
413 }
414
415 rperr = __find_aer_error_by_dev(rpdev);
416 if (!rperr) {
417 rperr = rperr_alloc;
418 rperr_alloc = NULL;
419 aer_error_init(rperr, pci_domain_nr(rpdev->bus),
420 rpdev->bus->number, rpdev->devfn,
421 rp_pos_cap_err);
422 list_add(&rperr->list, &einjected);
423 }
424 if (einj->cor_status) {
425 if (rperr->root_status & PCI_ERR_ROOT_COR_RCV)
426 rperr->root_status |= PCI_ERR_ROOT_MULTI_COR_RCV;
427 else
428 rperr->root_status |= PCI_ERR_ROOT_COR_RCV;
429 rperr->source_id &= 0xffff0000;
430 rperr->source_id |= (einj->bus << 8) | devfn;
431 }
432 if (einj->uncor_status) {
433 if (rperr->root_status & PCI_ERR_ROOT_UNCOR_RCV)
434 rperr->root_status |= PCI_ERR_ROOT_MULTI_UNCOR_RCV;
435 if (sever & einj->uncor_status) {
436 rperr->root_status |= PCI_ERR_ROOT_FATAL_RCV;
437 if (!(rperr->root_status & PCI_ERR_ROOT_UNCOR_RCV))
438 rperr->root_status |= PCI_ERR_ROOT_FIRST_FATAL;
439 } else
440 rperr->root_status |= PCI_ERR_ROOT_NONFATAL_RCV;
441 rperr->root_status |= PCI_ERR_ROOT_UNCOR_RCV;
442 rperr->source_id &= 0x0000ffff;
443 rperr->source_id |= ((einj->bus << 8) | devfn) << 16;
444 }
445 spin_unlock_irqrestore(&inject_lock, flags);
446
447 if (aer_mask_override) {
448 pci_write_config_dword(dev, pos_cap_err + PCI_ERR_COR_MASK,
449 cor_mask_orig);
450 pci_write_config_dword(dev, pos_cap_err + PCI_ERR_UNCOR_MASK,
451 uncor_mask_orig);
452 }
453
454 ret = pci_bus_set_aer_ops(dev->bus);
455 if (ret)
456 goto out_put;
457 ret = pci_bus_set_aer_ops(rpdev->bus);
458 if (ret)
459 goto out_put;
460
461 device = pcie_port_find_device(rpdev, PCIE_PORT_SERVICE_AER);
462 if (device) {
463 edev = to_pcie_device(device);
464 if (!get_service_data(edev)) {
465 pci_warn(edev->port, "AER service is not initialized\n");
466 ret = -EPROTONOSUPPORT;
467 goto out_put;
468 }
469 pci_info(edev->port, "Injecting errors %08x/%08x into device %s\n",
470 einj->cor_status, einj->uncor_status, pci_name(dev));
471 local_irq_disable();
472 generic_handle_irq(edev->irq);
473 local_irq_enable();
474 } else {
475 pci_err(rpdev, "AER device not found\n");
476 ret = -ENODEV;
477 }
478 out_put:
479 kfree(err_alloc);
480 kfree(rperr_alloc);
481 pci_dev_put(dev);
482 return ret;
483 }
484
485 static ssize_t aer_inject_write(struct file *filp, const char __user *ubuf,
486 size_t usize, loff_t *off)
487 {
488 struct aer_error_inj einj;
489 int ret;
490
491 if (!capable(CAP_SYS_ADMIN))
492 return -EPERM;
493 if (usize < offsetof(struct aer_error_inj, domain) ||
494 usize > sizeof(einj))
495 return -EINVAL;
496
497 memset(&einj, 0, sizeof(einj));
498 if (copy_from_user(&einj, ubuf, usize))
499 return -EFAULT;
500
501 ret = aer_inject(&einj);
502 return ret ? ret : usize;
503 }
504
505 static const struct file_operations aer_inject_fops = {
506 .write = aer_inject_write,
507 .owner = THIS_MODULE,
508 .llseek = noop_llseek,
509 };
510
511 static struct miscdevice aer_inject_device = {
512 .minor = MISC_DYNAMIC_MINOR,
513 .name = "aer_inject",
514 .fops = &aer_inject_fops,
515 };
516
517 static int __init aer_inject_init(void)
518 {
519 return misc_register(&aer_inject_device);
520 }
521
522 static void __exit aer_inject_exit(void)
523 {
524 struct aer_error *err, *err_next;
525 unsigned long flags;
526 struct pci_bus_ops *bus_ops;
527
528 misc_deregister(&aer_inject_device);
529
530 while ((bus_ops = pci_bus_ops_pop())) {
531 pci_bus_set_ops(bus_ops->bus, bus_ops->ops);
532 kfree(bus_ops);
533 }
534
535 spin_lock_irqsave(&inject_lock, flags);
536 list_for_each_entry_safe(err, err_next, &einjected, list) {
537 list_del(&err->list);
538 kfree(err);
539 }
540 spin_unlock_irqrestore(&inject_lock, flags);
541 }
542
543 module_init(aer_inject_init);
544 module_exit(aer_inject_exit);
545
546 MODULE_DESCRIPTION("PCIe AER software error injector");
547 MODULE_LICENSE("GPL");