This source file includes following definitions.
- irq_gc_noop
- irq_gc_mask_disable_reg
- irq_gc_mask_set_bit
- irq_gc_mask_clr_bit
- irq_gc_unmask_enable_reg
- irq_gc_ack_set_bit
- irq_gc_ack_clr_bit
- irq_gc_mask_disable_and_ack_set
- irq_gc_eoi
- irq_gc_set_wake
- irq_readl_be
- irq_writel_be
- irq_init_generic_chip
- irq_alloc_generic_chip
- irq_gc_init_mask_cache
- __irq_alloc_domain_generic_chips
- __irq_get_domain_generic_chip
- irq_get_domain_generic_chip
- irq_map_generic_chip
- irq_unmap_generic_chip
- irq_setup_generic_chip
- irq_setup_alt_chip
- irq_remove_generic_chip
- irq_gc_get_irq_data
- irq_gc_suspend
- irq_gc_resume
- irq_gc_shutdown
- irq_gc_init_ops
1
2
3
4
5
6
7 #include <linux/io.h>
8 #include <linux/irq.h>
9 #include <linux/slab.h>
10 #include <linux/export.h>
11 #include <linux/irqdomain.h>
12 #include <linux/interrupt.h>
13 #include <linux/kernel_stat.h>
14 #include <linux/syscore_ops.h>
15
16 #include "internals.h"
17
18 static LIST_HEAD(gc_list);
19 static DEFINE_RAW_SPINLOCK(gc_lock);
20
21
22
23
24
25 void irq_gc_noop(struct irq_data *d)
26 {
27 }
28
29
30
31
32
33
34
35
36 void irq_gc_mask_disable_reg(struct irq_data *d)
37 {
38 struct irq_chip_generic *gc = irq_data_get_irq_chip_data(d);
39 struct irq_chip_type *ct = irq_data_get_chip_type(d);
40 u32 mask = d->mask;
41
42 irq_gc_lock(gc);
43 irq_reg_writel(gc, mask, ct->regs.disable);
44 *ct->mask_cache &= ~mask;
45 irq_gc_unlock(gc);
46 }
47
48
49
50
51
52
53
54
55 void irq_gc_mask_set_bit(struct irq_data *d)
56 {
57 struct irq_chip_generic *gc = irq_data_get_irq_chip_data(d);
58 struct irq_chip_type *ct = irq_data_get_chip_type(d);
59 u32 mask = d->mask;
60
61 irq_gc_lock(gc);
62 *ct->mask_cache |= mask;
63 irq_reg_writel(gc, *ct->mask_cache, ct->regs.mask);
64 irq_gc_unlock(gc);
65 }
66 EXPORT_SYMBOL_GPL(irq_gc_mask_set_bit);
67
68
69
70
71
72
73
74
75 void irq_gc_mask_clr_bit(struct irq_data *d)
76 {
77 struct irq_chip_generic *gc = irq_data_get_irq_chip_data(d);
78 struct irq_chip_type *ct = irq_data_get_chip_type(d);
79 u32 mask = d->mask;
80
81 irq_gc_lock(gc);
82 *ct->mask_cache &= ~mask;
83 irq_reg_writel(gc, *ct->mask_cache, ct->regs.mask);
84 irq_gc_unlock(gc);
85 }
86 EXPORT_SYMBOL_GPL(irq_gc_mask_clr_bit);
87
88
89
90
91
92
93
94
95 void irq_gc_unmask_enable_reg(struct irq_data *d)
96 {
97 struct irq_chip_generic *gc = irq_data_get_irq_chip_data(d);
98 struct irq_chip_type *ct = irq_data_get_chip_type(d);
99 u32 mask = d->mask;
100
101 irq_gc_lock(gc);
102 irq_reg_writel(gc, mask, ct->regs.enable);
103 *ct->mask_cache |= mask;
104 irq_gc_unlock(gc);
105 }
106
107
108
109
110
111 void irq_gc_ack_set_bit(struct irq_data *d)
112 {
113 struct irq_chip_generic *gc = irq_data_get_irq_chip_data(d);
114 struct irq_chip_type *ct = irq_data_get_chip_type(d);
115 u32 mask = d->mask;
116
117 irq_gc_lock(gc);
118 irq_reg_writel(gc, mask, ct->regs.ack);
119 irq_gc_unlock(gc);
120 }
121 EXPORT_SYMBOL_GPL(irq_gc_ack_set_bit);
122
123
124
125
126
127 void irq_gc_ack_clr_bit(struct irq_data *d)
128 {
129 struct irq_chip_generic *gc = irq_data_get_irq_chip_data(d);
130 struct irq_chip_type *ct = irq_data_get_chip_type(d);
131 u32 mask = ~d->mask;
132
133 irq_gc_lock(gc);
134 irq_reg_writel(gc, mask, ct->regs.ack);
135 irq_gc_unlock(gc);
136 }
137
138
139
140
141
142
143
144
145
146
147
148
149
150 void irq_gc_mask_disable_and_ack_set(struct irq_data *d)
151 {
152 struct irq_chip_generic *gc = irq_data_get_irq_chip_data(d);
153 struct irq_chip_type *ct = irq_data_get_chip_type(d);
154 u32 mask = d->mask;
155
156 irq_gc_lock(gc);
157 irq_reg_writel(gc, mask, ct->regs.disable);
158 *ct->mask_cache &= ~mask;
159 irq_reg_writel(gc, mask, ct->regs.ack);
160 irq_gc_unlock(gc);
161 }
162
163
164
165
166
167 void irq_gc_eoi(struct irq_data *d)
168 {
169 struct irq_chip_generic *gc = irq_data_get_irq_chip_data(d);
170 struct irq_chip_type *ct = irq_data_get_chip_type(d);
171 u32 mask = d->mask;
172
173 irq_gc_lock(gc);
174 irq_reg_writel(gc, mask, ct->regs.eoi);
175 irq_gc_unlock(gc);
176 }
177
178
179
180
181
182
183
184
185
186
187 int irq_gc_set_wake(struct irq_data *d, unsigned int on)
188 {
189 struct irq_chip_generic *gc = irq_data_get_irq_chip_data(d);
190 u32 mask = d->mask;
191
192 if (!(mask & gc->wake_enabled))
193 return -EINVAL;
194
195 irq_gc_lock(gc);
196 if (on)
197 gc->wake_active |= mask;
198 else
199 gc->wake_active &= ~mask;
200 irq_gc_unlock(gc);
201 return 0;
202 }
203
204 static u32 irq_readl_be(void __iomem *addr)
205 {
206 return ioread32be(addr);
207 }
208
209 static void irq_writel_be(u32 val, void __iomem *addr)
210 {
211 iowrite32be(val, addr);
212 }
213
214 void irq_init_generic_chip(struct irq_chip_generic *gc, const char *name,
215 int num_ct, unsigned int irq_base,
216 void __iomem *reg_base, irq_flow_handler_t handler)
217 {
218 raw_spin_lock_init(&gc->lock);
219 gc->num_ct = num_ct;
220 gc->irq_base = irq_base;
221 gc->reg_base = reg_base;
222 gc->chip_types->chip.name = name;
223 gc->chip_types->handler = handler;
224 }
225
226
227
228
229
230
231
232
233
234
235
236
237 struct irq_chip_generic *
238 irq_alloc_generic_chip(const char *name, int num_ct, unsigned int irq_base,
239 void __iomem *reg_base, irq_flow_handler_t handler)
240 {
241 struct irq_chip_generic *gc;
242 unsigned long sz = sizeof(*gc) + num_ct * sizeof(struct irq_chip_type);
243
244 gc = kzalloc(sz, GFP_KERNEL);
245 if (gc) {
246 irq_init_generic_chip(gc, name, num_ct, irq_base, reg_base,
247 handler);
248 }
249 return gc;
250 }
251 EXPORT_SYMBOL_GPL(irq_alloc_generic_chip);
252
253 static void
254 irq_gc_init_mask_cache(struct irq_chip_generic *gc, enum irq_gc_flags flags)
255 {
256 struct irq_chip_type *ct = gc->chip_types;
257 u32 *mskptr = &gc->mask_cache, mskreg = ct->regs.mask;
258 int i;
259
260 for (i = 0; i < gc->num_ct; i++) {
261 if (flags & IRQ_GC_MASK_CACHE_PER_TYPE) {
262 mskptr = &ct[i].mask_cache_priv;
263 mskreg = ct[i].regs.mask;
264 }
265 ct[i].mask_cache = mskptr;
266 if (flags & IRQ_GC_INIT_MASK_CACHE)
267 *mskptr = irq_reg_readl(gc, mskreg);
268 }
269 }
270
271
272
273
274
275
276
277
278
279
280
281
282 int __irq_alloc_domain_generic_chips(struct irq_domain *d, int irqs_per_chip,
283 int num_ct, const char *name,
284 irq_flow_handler_t handler,
285 unsigned int clr, unsigned int set,
286 enum irq_gc_flags gcflags)
287 {
288 struct irq_domain_chip_generic *dgc;
289 struct irq_chip_generic *gc;
290 int numchips, sz, i;
291 unsigned long flags;
292 void *tmp;
293
294 if (d->gc)
295 return -EBUSY;
296
297 numchips = DIV_ROUND_UP(d->revmap_size, irqs_per_chip);
298 if (!numchips)
299 return -EINVAL;
300
301
302 sz = sizeof(*dgc) + numchips * sizeof(gc);
303 sz += numchips * (sizeof(*gc) + num_ct * sizeof(struct irq_chip_type));
304
305 tmp = dgc = kzalloc(sz, GFP_KERNEL);
306 if (!dgc)
307 return -ENOMEM;
308 dgc->irqs_per_chip = irqs_per_chip;
309 dgc->num_chips = numchips;
310 dgc->irq_flags_to_set = set;
311 dgc->irq_flags_to_clear = clr;
312 dgc->gc_flags = gcflags;
313 d->gc = dgc;
314
315
316 tmp += sizeof(*dgc) + numchips * sizeof(gc);
317 for (i = 0; i < numchips; i++) {
318
319 dgc->gc[i] = gc = tmp;
320 irq_init_generic_chip(gc, name, num_ct, i * irqs_per_chip,
321 NULL, handler);
322
323 gc->domain = d;
324 if (gcflags & IRQ_GC_BE_IO) {
325 gc->reg_readl = &irq_readl_be;
326 gc->reg_writel = &irq_writel_be;
327 }
328
329 raw_spin_lock_irqsave(&gc_lock, flags);
330 list_add_tail(&gc->list, &gc_list);
331 raw_spin_unlock_irqrestore(&gc_lock, flags);
332
333 tmp += sizeof(*gc) + num_ct * sizeof(struct irq_chip_type);
334 }
335 return 0;
336 }
337 EXPORT_SYMBOL_GPL(__irq_alloc_domain_generic_chips);
338
339 static struct irq_chip_generic *
340 __irq_get_domain_generic_chip(struct irq_domain *d, unsigned int hw_irq)
341 {
342 struct irq_domain_chip_generic *dgc = d->gc;
343 int idx;
344
345 if (!dgc)
346 return ERR_PTR(-ENODEV);
347 idx = hw_irq / dgc->irqs_per_chip;
348 if (idx >= dgc->num_chips)
349 return ERR_PTR(-EINVAL);
350 return dgc->gc[idx];
351 }
352
353
354
355
356
357
358 struct irq_chip_generic *
359 irq_get_domain_generic_chip(struct irq_domain *d, unsigned int hw_irq)
360 {
361 struct irq_chip_generic *gc = __irq_get_domain_generic_chip(d, hw_irq);
362
363 return !IS_ERR(gc) ? gc : NULL;
364 }
365 EXPORT_SYMBOL_GPL(irq_get_domain_generic_chip);
366
367
368
369
370
371 static struct lock_class_key irq_nested_lock_class;
372 static struct lock_class_key irq_nested_request_class;
373
374
375
376
377 int irq_map_generic_chip(struct irq_domain *d, unsigned int virq,
378 irq_hw_number_t hw_irq)
379 {
380 struct irq_data *data = irq_domain_get_irq_data(d, virq);
381 struct irq_domain_chip_generic *dgc = d->gc;
382 struct irq_chip_generic *gc;
383 struct irq_chip_type *ct;
384 struct irq_chip *chip;
385 unsigned long flags;
386 int idx;
387
388 gc = __irq_get_domain_generic_chip(d, hw_irq);
389 if (IS_ERR(gc))
390 return PTR_ERR(gc);
391
392 idx = hw_irq % dgc->irqs_per_chip;
393
394 if (test_bit(idx, &gc->unused))
395 return -ENOTSUPP;
396
397 if (test_bit(idx, &gc->installed))
398 return -EBUSY;
399
400 ct = gc->chip_types;
401 chip = &ct->chip;
402
403
404 if (!gc->installed) {
405 raw_spin_lock_irqsave(&gc->lock, flags);
406 irq_gc_init_mask_cache(gc, dgc->gc_flags);
407 raw_spin_unlock_irqrestore(&gc->lock, flags);
408 }
409
410
411 set_bit(idx, &gc->installed);
412
413 if (dgc->gc_flags & IRQ_GC_INIT_NESTED_LOCK)
414 irq_set_lockdep_class(virq, &irq_nested_lock_class,
415 &irq_nested_request_class);
416
417 if (chip->irq_calc_mask)
418 chip->irq_calc_mask(data);
419 else
420 data->mask = 1 << idx;
421
422 irq_domain_set_info(d, virq, hw_irq, chip, gc, ct->handler, NULL, NULL);
423 irq_modify_status(virq, dgc->irq_flags_to_clear, dgc->irq_flags_to_set);
424 return 0;
425 }
426
427 static void irq_unmap_generic_chip(struct irq_domain *d, unsigned int virq)
428 {
429 struct irq_data *data = irq_domain_get_irq_data(d, virq);
430 struct irq_domain_chip_generic *dgc = d->gc;
431 unsigned int hw_irq = data->hwirq;
432 struct irq_chip_generic *gc;
433 int irq_idx;
434
435 gc = irq_get_domain_generic_chip(d, hw_irq);
436 if (!gc)
437 return;
438
439 irq_idx = hw_irq % dgc->irqs_per_chip;
440
441 clear_bit(irq_idx, &gc->installed);
442 irq_domain_set_info(d, virq, hw_irq, &no_irq_chip, NULL, NULL, NULL,
443 NULL);
444
445 }
446
447 struct irq_domain_ops irq_generic_chip_ops = {
448 .map = irq_map_generic_chip,
449 .unmap = irq_unmap_generic_chip,
450 .xlate = irq_domain_xlate_onetwocell,
451 };
452 EXPORT_SYMBOL_GPL(irq_generic_chip_ops);
453
454
455
456
457
458
459
460
461
462
463
464
465
466 void irq_setup_generic_chip(struct irq_chip_generic *gc, u32 msk,
467 enum irq_gc_flags flags, unsigned int clr,
468 unsigned int set)
469 {
470 struct irq_chip_type *ct = gc->chip_types;
471 struct irq_chip *chip = &ct->chip;
472 unsigned int i;
473
474 raw_spin_lock(&gc_lock);
475 list_add_tail(&gc->list, &gc_list);
476 raw_spin_unlock(&gc_lock);
477
478 irq_gc_init_mask_cache(gc, flags);
479
480 for (i = gc->irq_base; msk; msk >>= 1, i++) {
481 if (!(msk & 0x01))
482 continue;
483
484 if (flags & IRQ_GC_INIT_NESTED_LOCK)
485 irq_set_lockdep_class(i, &irq_nested_lock_class,
486 &irq_nested_request_class);
487
488 if (!(flags & IRQ_GC_NO_MASK)) {
489 struct irq_data *d = irq_get_irq_data(i);
490
491 if (chip->irq_calc_mask)
492 chip->irq_calc_mask(d);
493 else
494 d->mask = 1 << (i - gc->irq_base);
495 }
496 irq_set_chip_and_handler(i, chip, ct->handler);
497 irq_set_chip_data(i, gc);
498 irq_modify_status(i, clr, set);
499 }
500 gc->irq_cnt = i - gc->irq_base;
501 }
502 EXPORT_SYMBOL_GPL(irq_setup_generic_chip);
503
504
505
506
507
508
509
510
511 int irq_setup_alt_chip(struct irq_data *d, unsigned int type)
512 {
513 struct irq_chip_generic *gc = irq_data_get_irq_chip_data(d);
514 struct irq_chip_type *ct = gc->chip_types;
515 unsigned int i;
516
517 for (i = 0; i < gc->num_ct; i++, ct++) {
518 if (ct->type & type) {
519 d->chip = &ct->chip;
520 irq_data_to_desc(d)->handle_irq = ct->handler;
521 return 0;
522 }
523 }
524 return -EINVAL;
525 }
526 EXPORT_SYMBOL_GPL(irq_setup_alt_chip);
527
528
529
530
531
532
533
534
535
536
537 void irq_remove_generic_chip(struct irq_chip_generic *gc, u32 msk,
538 unsigned int clr, unsigned int set)
539 {
540 unsigned int i = gc->irq_base;
541
542 raw_spin_lock(&gc_lock);
543 list_del(&gc->list);
544 raw_spin_unlock(&gc_lock);
545
546 for (; msk; msk >>= 1, i++) {
547 if (!(msk & 0x01))
548 continue;
549
550
551 irq_set_handler(i, NULL);
552 irq_set_chip(i, &no_irq_chip);
553 irq_set_chip_data(i, NULL);
554 irq_modify_status(i, clr, set);
555 }
556 }
557 EXPORT_SYMBOL_GPL(irq_remove_generic_chip);
558
559 static struct irq_data *irq_gc_get_irq_data(struct irq_chip_generic *gc)
560 {
561 unsigned int virq;
562
563 if (!gc->domain)
564 return irq_get_irq_data(gc->irq_base);
565
566
567
568
569
570 if (!gc->installed)
571 return NULL;
572
573 virq = irq_find_mapping(gc->domain, gc->irq_base + __ffs(gc->installed));
574 return virq ? irq_get_irq_data(virq) : NULL;
575 }
576
577 #ifdef CONFIG_PM
578 static int irq_gc_suspend(void)
579 {
580 struct irq_chip_generic *gc;
581
582 list_for_each_entry(gc, &gc_list, list) {
583 struct irq_chip_type *ct = gc->chip_types;
584
585 if (ct->chip.irq_suspend) {
586 struct irq_data *data = irq_gc_get_irq_data(gc);
587
588 if (data)
589 ct->chip.irq_suspend(data);
590 }
591
592 if (gc->suspend)
593 gc->suspend(gc);
594 }
595 return 0;
596 }
597
598 static void irq_gc_resume(void)
599 {
600 struct irq_chip_generic *gc;
601
602 list_for_each_entry(gc, &gc_list, list) {
603 struct irq_chip_type *ct = gc->chip_types;
604
605 if (gc->resume)
606 gc->resume(gc);
607
608 if (ct->chip.irq_resume) {
609 struct irq_data *data = irq_gc_get_irq_data(gc);
610
611 if (data)
612 ct->chip.irq_resume(data);
613 }
614 }
615 }
616 #else
617 #define irq_gc_suspend NULL
618 #define irq_gc_resume NULL
619 #endif
620
621 static void irq_gc_shutdown(void)
622 {
623 struct irq_chip_generic *gc;
624
625 list_for_each_entry(gc, &gc_list, list) {
626 struct irq_chip_type *ct = gc->chip_types;
627
628 if (ct->chip.irq_pm_shutdown) {
629 struct irq_data *data = irq_gc_get_irq_data(gc);
630
631 if (data)
632 ct->chip.irq_pm_shutdown(data);
633 }
634 }
635 }
636
637 static struct syscore_ops irq_gc_syscore_ops = {
638 .suspend = irq_gc_suspend,
639 .resume = irq_gc_resume,
640 .shutdown = irq_gc_shutdown,
641 };
642
643 static int __init irq_gc_init_ops(void)
644 {
645 register_syscore_ops(&irq_gc_syscore_ops);
646 return 0;
647 }
648 device_initcall(irq_gc_init_ops);