This source file includes following definitions.
- mips_cpc_default_phys_base
- mips_cpc_phys_base
- mips_cpc_probe
- mips_cpc_lock_other
- mips_cpc_unlock_other
1
2
3
4
5
6
7 #include <linux/errno.h>
8 #include <linux/percpu.h>
9 #include <linux/of.h>
10 #include <linux/of_address.h>
11 #include <linux/spinlock.h>
12
13 #include <asm/mips-cps.h>
14
15 void __iomem *mips_cpc_base;
16
17 static DEFINE_PER_CPU_ALIGNED(spinlock_t, cpc_core_lock);
18
19 static DEFINE_PER_CPU_ALIGNED(unsigned long, cpc_core_lock_flags);
20
21 phys_addr_t __weak mips_cpc_default_phys_base(void)
22 {
23 struct device_node *cpc_node;
24 struct resource res;
25 int err;
26
27 cpc_node = of_find_compatible_node(of_root, NULL, "mti,mips-cpc");
28 if (cpc_node) {
29 err = of_address_to_resource(cpc_node, 0, &res);
30 if (!err)
31 return res.start;
32 }
33
34 return 0;
35 }
36
37
38
39
40
41
42
43
44 static phys_addr_t mips_cpc_phys_base(void)
45 {
46 unsigned long cpc_base;
47
48 if (!mips_cm_present())
49 return 0;
50
51 if (!(read_gcr_cpc_status() & CM_GCR_CPC_STATUS_EX))
52 return 0;
53
54
55 cpc_base = read_gcr_cpc_base();
56 if (cpc_base & CM_GCR_CPC_BASE_CPCEN)
57 return cpc_base & CM_GCR_CPC_BASE_CPCBASE;
58
59
60 cpc_base = mips_cpc_default_phys_base();
61 if (!cpc_base)
62 return cpc_base;
63
64
65 write_gcr_cpc_base(cpc_base | CM_GCR_CPC_BASE_CPCEN);
66 return cpc_base;
67 }
68
69 int mips_cpc_probe(void)
70 {
71 phys_addr_t addr;
72 unsigned int cpu;
73
74 for_each_possible_cpu(cpu)
75 spin_lock_init(&per_cpu(cpc_core_lock, cpu));
76
77 addr = mips_cpc_phys_base();
78 if (!addr)
79 return -ENODEV;
80
81 mips_cpc_base = ioremap_nocache(addr, 0x8000);
82 if (!mips_cpc_base)
83 return -ENXIO;
84
85 return 0;
86 }
87
88 void mips_cpc_lock_other(unsigned int core)
89 {
90 unsigned int curr_core;
91
92 if (mips_cm_revision() >= CM_REV_CM3)
93
94 return;
95
96 preempt_disable();
97 curr_core = cpu_core(¤t_cpu_data);
98 spin_lock_irqsave(&per_cpu(cpc_core_lock, curr_core),
99 per_cpu(cpc_core_lock_flags, curr_core));
100 write_cpc_cl_other(core << __ffs(CPC_Cx_OTHER_CORENUM));
101
102
103
104
105
106 mb();
107 }
108
109 void mips_cpc_unlock_other(void)
110 {
111 unsigned int curr_core;
112
113 if (mips_cm_revision() >= CM_REV_CM3)
114
115 return;
116
117 curr_core = cpu_core(¤t_cpu_data);
118 spin_unlock_irqrestore(&per_cpu(cpc_core_lock, curr_core),
119 per_cpu(cpc_core_lock_flags, curr_core));
120 preempt_enable();
121 }