This source file includes following definitions.
- read_decode_ccm_bcr
- decode_arc_core
- read_arc_build_cfg_regs
- arc_cpu_mumbojumbo
- arc_extn_mumbojumbo
- arc_chk_core_config
- setup_processor
- uboot_arg_invalid
- handle_uboot_args
- setup_arch
- time_init
- customize_machine
- init_late_machine
- show_cpuinfo
- c_start
- c_next
- c_stop
- topology_init
1
2
3
4
5
6 #include <linux/seq_file.h>
7 #include <linux/fs.h>
8 #include <linux/delay.h>
9 #include <linux/root_dev.h>
10 #include <linux/clk.h>
11 #include <linux/clk-provider.h>
12 #include <linux/clocksource.h>
13 #include <linux/console.h>
14 #include <linux/module.h>
15 #include <linux/sizes.h>
16 #include <linux/cpu.h>
17 #include <linux/of_fdt.h>
18 #include <linux/of.h>
19 #include <linux/cache.h>
20 #include <uapi/linux/mount.h>
21 #include <asm/sections.h>
22 #include <asm/arcregs.h>
23 #include <asm/tlb.h>
24 #include <asm/setup.h>
25 #include <asm/page.h>
26 #include <asm/irq.h>
27 #include <asm/unwind.h>
28 #include <asm/mach_desc.h>
29 #include <asm/smp.h>
30
31 #define FIX_PTR(x) __asm__ __volatile__(";" : "+r"(x))
32
33 unsigned int intr_to_DE_cnt;
34
35
36 int __initdata uboot_tag;
37 int __initdata uboot_magic;
38 char __initdata *uboot_arg;
39
40 const struct machine_desc *machine_desc;
41
42 struct task_struct *_current_task[NR_CPUS];
43
44 struct cpuinfo_arc cpuinfo_arc700[NR_CPUS];
45
46 static const struct id_to_str arc_legacy_rel[] = {
47
48 #ifdef CONFIG_ISA_ARCOMPACT
49 { 0x34, "R4.10"},
50 { 0x35, "R4.11"},
51 #else
52 { 0x51, "R2.0" },
53 { 0x52, "R2.1" },
54 { 0x53, "R3.0" },
55 #endif
56 { 0x00, NULL }
57 };
58
59 static const struct id_to_str arc_cpu_rel[] = {
60
61 { 0, "R3.10a"},
62 { 1, "R3.50a"},
63 { 0xFF, NULL }
64 };
65
66 static void read_decode_ccm_bcr(struct cpuinfo_arc *cpu)
67 {
68 if (is_isa_arcompact()) {
69 struct bcr_iccm_arcompact iccm;
70 struct bcr_dccm_arcompact dccm;
71
72 READ_BCR(ARC_REG_ICCM_BUILD, iccm);
73 if (iccm.ver) {
74 cpu->iccm.sz = 4096 << iccm.sz;
75 cpu->iccm.base_addr = iccm.base << 16;
76 }
77
78 READ_BCR(ARC_REG_DCCM_BUILD, dccm);
79 if (dccm.ver) {
80 unsigned long base;
81 cpu->dccm.sz = 2048 << dccm.sz;
82
83 base = read_aux_reg(ARC_REG_DCCM_BASE_BUILD);
84 cpu->dccm.base_addr = base & ~0xF;
85 }
86 } else {
87 struct bcr_iccm_arcv2 iccm;
88 struct bcr_dccm_arcv2 dccm;
89 unsigned long region;
90
91 READ_BCR(ARC_REG_ICCM_BUILD, iccm);
92 if (iccm.ver) {
93 cpu->iccm.sz = 256 << iccm.sz00;
94 if (iccm.sz00 == 0xF && iccm.sz01 > 0)
95 cpu->iccm.sz <<= iccm.sz01;
96
97 region = read_aux_reg(ARC_REG_AUX_ICCM);
98 cpu->iccm.base_addr = region & 0xF0000000;
99 }
100
101 READ_BCR(ARC_REG_DCCM_BUILD, dccm);
102 if (dccm.ver) {
103 cpu->dccm.sz = 256 << dccm.sz0;
104 if (dccm.sz0 == 0xF && dccm.sz1 > 0)
105 cpu->dccm.sz <<= dccm.sz1;
106
107 region = read_aux_reg(ARC_REG_AUX_DCCM);
108 cpu->dccm.base_addr = region & 0xF0000000;
109 }
110 }
111 }
112
113 static void decode_arc_core(struct cpuinfo_arc *cpu)
114 {
115 struct bcr_uarch_build_arcv2 uarch;
116 const struct id_to_str *tbl;
117
118
119
120
121
122
123
124 if (cpu->core.family < 0x54) {
125
126 for (tbl = &arc_legacy_rel[0]; tbl->id != 0; tbl++) {
127 if (cpu->core.family == tbl->id) {
128 cpu->release = tbl->str;
129 break;
130 }
131 }
132
133 if (is_isa_arcompact())
134 cpu->name = "ARC700";
135 else if (tbl->str)
136 cpu->name = "HS38";
137 else
138 cpu->name = cpu->release = "Unknown";
139
140 return;
141 }
142
143
144
145
146
147
148
149 READ_BCR(ARC_REG_MICRO_ARCH_BCR, uarch);
150
151 if (uarch.prod == 4) {
152 cpu->name = "HS48";
153 cpu->extn.dual = 1;
154
155 } else {
156 cpu->name = "HS38";
157 }
158
159 for (tbl = &arc_cpu_rel[0]; tbl->id != 0xFF; tbl++) {
160 if (uarch.maj == tbl->id) {
161 cpu->release = tbl->str;
162 break;
163 }
164 }
165 }
166
167 static void read_arc_build_cfg_regs(void)
168 {
169 struct bcr_timer timer;
170 struct bcr_generic bcr;
171 struct cpuinfo_arc *cpu = &cpuinfo_arc700[smp_processor_id()];
172 struct bcr_isa_arcv2 isa;
173 struct bcr_actionpoint ap;
174
175 FIX_PTR(cpu);
176
177 READ_BCR(AUX_IDENTITY, cpu->core);
178 decode_arc_core(cpu);
179
180 READ_BCR(ARC_REG_TIMERS_BCR, timer);
181 cpu->extn.timer0 = timer.t0;
182 cpu->extn.timer1 = timer.t1;
183 cpu->extn.rtc = timer.rtc;
184
185 cpu->vec_base = read_aux_reg(AUX_INTR_VEC_BASE);
186
187 READ_BCR(ARC_REG_MUL_BCR, cpu->extn_mpy);
188
189
190 read_decode_ccm_bcr(cpu);
191
192 read_decode_mmu_bcr();
193 read_decode_cache_bcr();
194
195 if (is_isa_arcompact()) {
196 struct bcr_fp_arcompact sp, dp;
197 struct bcr_bpu_arcompact bpu;
198
199 READ_BCR(ARC_REG_FP_BCR, sp);
200 READ_BCR(ARC_REG_DPFP_BCR, dp);
201 cpu->extn.fpu_sp = sp.ver ? 1 : 0;
202 cpu->extn.fpu_dp = dp.ver ? 1 : 0;
203
204 READ_BCR(ARC_REG_BPU_BCR, bpu);
205 cpu->bpu.ver = bpu.ver;
206 cpu->bpu.full = bpu.fam ? 1 : 0;
207 if (bpu.ent) {
208 cpu->bpu.num_cache = 256 << (bpu.ent - 1);
209 cpu->bpu.num_pred = 256 << (bpu.ent - 1);
210 }
211 } else {
212 struct bcr_fp_arcv2 spdp;
213 struct bcr_bpu_arcv2 bpu;
214
215 READ_BCR(ARC_REG_FP_V2_BCR, spdp);
216 cpu->extn.fpu_sp = spdp.sp ? 1 : 0;
217 cpu->extn.fpu_dp = spdp.dp ? 1 : 0;
218
219 READ_BCR(ARC_REG_BPU_BCR, bpu);
220 cpu->bpu.ver = bpu.ver;
221 cpu->bpu.full = bpu.ft;
222 cpu->bpu.num_cache = 256 << bpu.bce;
223 cpu->bpu.num_pred = 2048 << bpu.pte;
224 cpu->bpu.ret_stk = 4 << bpu.rse;
225
226
227 if (cpu->extn.dual) {
228 unsigned int exec_ctrl;
229
230 READ_BCR(AUX_EXEC_CTRL, exec_ctrl);
231 cpu->extn.dual_enb = !(exec_ctrl & 1);
232 }
233 }
234
235 READ_BCR(ARC_REG_AP_BCR, ap);
236 if (ap.ver) {
237 cpu->extn.ap_num = 2 << ap.num;
238 cpu->extn.ap_full = !ap.min;
239 }
240
241 READ_BCR(ARC_REG_SMART_BCR, bcr);
242 cpu->extn.smart = bcr.ver ? 1 : 0;
243
244 READ_BCR(ARC_REG_RTT_BCR, bcr);
245 cpu->extn.rtt = bcr.ver ? 1 : 0;
246
247 READ_BCR(ARC_REG_ISA_CFG_BCR, isa);
248
249
250 if (is_isa_arcompact()) {
251 if (!isa.ver)
252 cpu->isa.atomic = IS_ENABLED(CONFIG_ARC_HAS_LLSC);
253 else {
254
255 struct bcr_generic bcr = *(struct bcr_generic *)&isa;
256 cpu->isa.atomic = bcr.info & 1;
257 }
258
259 cpu->isa.be = IS_ENABLED(CONFIG_CPU_BIG_ENDIAN);
260
261
262 if (unlikely(cpu->core.family < 0x34 || cpu->mmu.ver < 3))
263 cpu->name = "ARC750";
264 } else {
265 cpu->isa = isa;
266 }
267 }
268
269 static char *arc_cpu_mumbojumbo(int cpu_id, char *buf, int len)
270 {
271 struct cpuinfo_arc *cpu = &cpuinfo_arc700[cpu_id];
272 struct bcr_identity *core = &cpu->core;
273 char mpy_opt[16];
274 int n = 0;
275
276 FIX_PTR(cpu);
277
278 n += scnprintf(buf + n, len - n,
279 "\nIDENTITY\t: ARCVER [%#02x] ARCNUM [%#02x] CHIPID [%#4x]\n",
280 core->family, core->cpu_id, core->chip_id);
281
282 n += scnprintf(buf + n, len - n, "processor [%d]\t: %s %s (%s ISA) %s%s%s\n",
283 cpu_id, cpu->name, cpu->release,
284 is_isa_arcompact() ? "ARCompact" : "ARCv2",
285 IS_AVAIL1(cpu->isa.be, "[Big-Endian]"),
286 IS_AVAIL3(cpu->extn.dual, cpu->extn.dual_enb, " Dual-Issue "));
287
288 n += scnprintf(buf + n, len - n, "Timers\t\t: %s%s%s%s%s%s\nISA Extn\t: ",
289 IS_AVAIL1(cpu->extn.timer0, "Timer0 "),
290 IS_AVAIL1(cpu->extn.timer1, "Timer1 "),
291 IS_AVAIL2(cpu->extn.rtc, "RTC [UP 64-bit] ", CONFIG_ARC_TIMERS_64BIT),
292 IS_AVAIL2(cpu->extn.gfrc, "GFRC [SMP 64-bit] ", CONFIG_ARC_TIMERS_64BIT));
293
294 if (cpu->extn_mpy.ver) {
295 if (is_isa_arcompact()) {
296 scnprintf(mpy_opt, 16, "mpy");
297 } else {
298
299 int opt = 2;
300
301 if (cpu->extn_mpy.dsp)
302 opt = cpu->extn_mpy.dsp + 6;
303
304 scnprintf(mpy_opt, 16, "mpy[opt %d] ", opt);
305 }
306 }
307
308 n += scnprintf(buf + n, len - n, "%s%s%s%s%s%s%s%s\n",
309 IS_AVAIL2(cpu->isa.atomic, "atomic ", CONFIG_ARC_HAS_LLSC),
310 IS_AVAIL2(cpu->isa.ldd, "ll64 ", CONFIG_ARC_HAS_LL64),
311 IS_AVAIL2(cpu->isa.unalign, "unalign ", CONFIG_ARC_USE_UNALIGNED_MEM_ACCESS),
312 IS_AVAIL1(cpu->extn_mpy.ver, mpy_opt),
313 IS_AVAIL1(cpu->isa.div_rem, "div_rem "));
314
315 if (cpu->bpu.ver) {
316 n += scnprintf(buf + n, len - n,
317 "BPU\t\t: %s%s match, cache:%d, Predict Table:%d Return stk: %d",
318 IS_AVAIL1(cpu->bpu.full, "full"),
319 IS_AVAIL1(!cpu->bpu.full, "partial"),
320 cpu->bpu.num_cache, cpu->bpu.num_pred, cpu->bpu.ret_stk);
321
322 if (is_isa_arcv2()) {
323 struct bcr_lpb lpb;
324
325 READ_BCR(ARC_REG_LPB_BUILD, lpb);
326 if (lpb.ver) {
327 unsigned int ctl;
328 ctl = read_aux_reg(ARC_REG_LPB_CTRL);
329
330 n += scnprintf(buf + n, len - n, " Loop Buffer:%d %s",
331 lpb.entries,
332 IS_DISABLED_RUN(!ctl));
333 }
334 }
335 n += scnprintf(buf + n, len - n, "\n");
336 }
337
338 return buf;
339 }
340
341 static char *arc_extn_mumbojumbo(int cpu_id, char *buf, int len)
342 {
343 int n = 0;
344 struct cpuinfo_arc *cpu = &cpuinfo_arc700[cpu_id];
345
346 FIX_PTR(cpu);
347
348 n += scnprintf(buf + n, len - n, "Vector Table\t: %#x\n", cpu->vec_base);
349
350 if (cpu->extn.fpu_sp || cpu->extn.fpu_dp)
351 n += scnprintf(buf + n, len - n, "FPU\t\t: %s%s\n",
352 IS_AVAIL1(cpu->extn.fpu_sp, "SP "),
353 IS_AVAIL1(cpu->extn.fpu_dp, "DP "));
354
355 if (cpu->extn.ap_num | cpu->extn.smart | cpu->extn.rtt) {
356 n += scnprintf(buf + n, len - n, "DEBUG\t\t: %s%s",
357 IS_AVAIL1(cpu->extn.smart, "smaRT "),
358 IS_AVAIL1(cpu->extn.rtt, "RTT "));
359 if (cpu->extn.ap_num) {
360 n += scnprintf(buf + n, len - n, "ActionPoint %d/%s",
361 cpu->extn.ap_num,
362 cpu->extn.ap_full ? "full":"min");
363 }
364 n += scnprintf(buf + n, len - n, "\n");
365 }
366
367 if (cpu->dccm.sz || cpu->iccm.sz)
368 n += scnprintf(buf + n, len - n, "Extn [CCM]\t: DCCM @ %x, %d KB / ICCM: @ %x, %d KB\n",
369 cpu->dccm.base_addr, TO_KB(cpu->dccm.sz),
370 cpu->iccm.base_addr, TO_KB(cpu->iccm.sz));
371
372 if (is_isa_arcv2()) {
373
374
375 struct bcr_erp erp;
376 READ_BCR(ARC_REG_ERP_BUILD, erp);
377
378 if (erp.ver) {
379 struct ctl_erp ctl;
380 READ_BCR(ARC_REG_ERP_CTRL, ctl);
381
382
383 n += scnprintf(buf + n, len - n, "Extn [ECC]\t: %s%s%s%s%s%s\n",
384 IS_AVAIL3(erp.ic, !ctl.dpi, "IC "),
385 IS_AVAIL3(erp.dc, !ctl.dpd, "DC "),
386 IS_AVAIL3(erp.mmu, !ctl.mpd, "MMU "));
387 }
388 }
389
390 return buf;
391 }
392
393 static void arc_chk_core_config(void)
394 {
395 struct cpuinfo_arc *cpu = &cpuinfo_arc700[smp_processor_id()];
396 int saved = 0, present = 0;
397 char *opt_nm = NULL;
398
399 if (!cpu->extn.timer0)
400 panic("Timer0 is not present!\n");
401
402 if (!cpu->extn.timer1)
403 panic("Timer1 is not present!\n");
404
405 #ifdef CONFIG_ARC_HAS_DCCM
406
407
408
409
410 if ((unsigned int)__arc_dccm_base != cpu->dccm.base_addr)
411 panic("Linux built with incorrect DCCM Base address\n");
412
413 if (CONFIG_ARC_DCCM_SZ * SZ_1K != cpu->dccm.sz)
414 panic("Linux built with incorrect DCCM Size\n");
415 #endif
416
417 #ifdef CONFIG_ARC_HAS_ICCM
418 if (CONFIG_ARC_ICCM_SZ * SZ_1K != cpu->iccm.sz)
419 panic("Linux built with incorrect ICCM Size\n");
420 #endif
421
422
423
424
425
426
427
428 if (is_isa_arcompact()) {
429 opt_nm = "CONFIG_ARC_FPU_SAVE_RESTORE";
430 saved = IS_ENABLED(CONFIG_ARC_FPU_SAVE_RESTORE);
431
432
433 present = cpu->extn.fpu_dp;
434 } else {
435 opt_nm = "CONFIG_ARC_HAS_ACCL_REGS";
436 saved = IS_ENABLED(CONFIG_ARC_HAS_ACCL_REGS);
437
438
439 present = cpu->extn_mpy.dsp | cpu->extn.fpu_sp | cpu->extn.fpu_dp;
440 }
441
442 if (present && !saved)
443 pr_warn("Enable %s for working apps\n", opt_nm);
444 else if (!present && saved)
445 panic("Disable %s, hardware NOT present\n", opt_nm);
446 }
447
448
449
450
451
452
453
454 void setup_processor(void)
455 {
456 char str[512];
457 int cpu_id = smp_processor_id();
458
459 read_arc_build_cfg_regs();
460 arc_init_IRQ();
461
462 pr_info("%s", arc_cpu_mumbojumbo(cpu_id, str, sizeof(str)));
463
464 arc_mmu_init();
465 arc_cache_init();
466
467 pr_info("%s", arc_extn_mumbojumbo(cpu_id, str, sizeof(str)));
468 pr_info("%s", arc_platform_smp_cpuinfo());
469
470 arc_chk_core_config();
471 }
472
473 static inline bool uboot_arg_invalid(unsigned long addr)
474 {
475
476
477
478
479 if (addr < PAGE_OFFSET)
480 return true;
481
482
483 return addr >= (unsigned long)_stext && addr <= (unsigned long)_end;
484 }
485
486 #define IGNORE_ARGS "Ignore U-boot args: "
487
488
489 #define UBOOT_TAG_NONE 0
490 #define UBOOT_TAG_CMDLINE 1
491 #define UBOOT_TAG_DTB 2
492
493 #define UBOOT_MAGIC_VALUE 0
494
495 void __init handle_uboot_args(void)
496 {
497 bool use_embedded_dtb = true;
498 bool append_cmdline = false;
499
500
501 if (uboot_tag != UBOOT_TAG_NONE &&
502 uboot_tag != UBOOT_TAG_CMDLINE &&
503 uboot_tag != UBOOT_TAG_DTB) {
504 pr_warn(IGNORE_ARGS "invalid uboot tag: '%08x'\n", uboot_tag);
505 goto ignore_uboot_args;
506 }
507
508 if (uboot_magic != UBOOT_MAGIC_VALUE) {
509 pr_warn(IGNORE_ARGS "non zero uboot magic\n");
510 goto ignore_uboot_args;
511 }
512
513 if (uboot_tag != UBOOT_TAG_NONE &&
514 uboot_arg_invalid((unsigned long)uboot_arg)) {
515 pr_warn(IGNORE_ARGS "invalid uboot arg: '%px'\n", uboot_arg);
516 goto ignore_uboot_args;
517 }
518
519
520 if (uboot_tag == UBOOT_TAG_DTB) {
521 machine_desc = setup_machine_fdt((void *)uboot_arg);
522
523
524 use_embedded_dtb = !machine_desc;
525 }
526
527 if (uboot_tag == UBOOT_TAG_CMDLINE)
528 append_cmdline = true;
529
530 ignore_uboot_args:
531
532 if (use_embedded_dtb) {
533 machine_desc = setup_machine_fdt(__dtb_start);
534 if (!machine_desc)
535 panic("Embedded DT invalid\n");
536 }
537
538
539
540
541
542 if (append_cmdline) {
543
544 strlcat(boot_command_line, " ", COMMAND_LINE_SIZE);
545 strlcat(boot_command_line, uboot_arg, COMMAND_LINE_SIZE);
546 }
547 }
548
549 void __init setup_arch(char **cmdline_p)
550 {
551 handle_uboot_args();
552
553
554 *cmdline_p = boot_command_line;
555
556
557 parse_early_param();
558
559
560 if (machine_desc->init_early)
561 machine_desc->init_early();
562
563 smp_init_cpus();
564
565 setup_processor();
566 setup_arch_memory();
567
568
569 unflatten_and_copy_device_tree();
570
571
572
573
574 root_mountflags &= ~MS_RDONLY;
575
576 #if defined(CONFIG_VT) && defined(CONFIG_DUMMY_CONSOLE)
577 conswitchp = &dummy_con;
578 #endif
579
580 arc_unwind_init();
581 }
582
583
584
585
586 void __init time_init(void)
587 {
588 of_clk_init(NULL);
589 timer_probe();
590 }
591
592 static int __init customize_machine(void)
593 {
594 if (machine_desc->init_machine)
595 machine_desc->init_machine();
596
597 return 0;
598 }
599 arch_initcall(customize_machine);
600
601 static int __init init_late_machine(void)
602 {
603 if (machine_desc->init_late)
604 machine_desc->init_late();
605
606 return 0;
607 }
608 late_initcall(init_late_machine);
609
610
611
612
613 #define cpu_to_ptr(c) ((void *)(0xFFFF0000 | (unsigned int)(c)))
614 #define ptr_to_cpu(p) (~0xFFFF0000UL & (unsigned int)(p))
615
616 static int show_cpuinfo(struct seq_file *m, void *v)
617 {
618 char *str;
619 int cpu_id = ptr_to_cpu(v);
620 struct device *cpu_dev = get_cpu_device(cpu_id);
621 struct clk *cpu_clk;
622 unsigned long freq = 0;
623
624 if (!cpu_online(cpu_id)) {
625 seq_printf(m, "processor [%d]\t: Offline\n", cpu_id);
626 goto done;
627 }
628
629 str = (char *)__get_free_page(GFP_KERNEL);
630 if (!str)
631 goto done;
632
633 seq_printf(m, arc_cpu_mumbojumbo(cpu_id, str, PAGE_SIZE));
634
635 cpu_clk = clk_get(cpu_dev, NULL);
636 if (IS_ERR(cpu_clk)) {
637 seq_printf(m, "CPU speed \t: Cannot get clock for processor [%d]\n",
638 cpu_id);
639 } else {
640 freq = clk_get_rate(cpu_clk);
641 }
642 if (freq)
643 seq_printf(m, "CPU speed\t: %lu.%02lu Mhz\n",
644 freq / 1000000, (freq / 10000) % 100);
645
646 seq_printf(m, "Bogo MIPS\t: %lu.%02lu\n",
647 loops_per_jiffy / (500000 / HZ),
648 (loops_per_jiffy / (5000 / HZ)) % 100);
649
650 seq_printf(m, arc_mmu_mumbojumbo(cpu_id, str, PAGE_SIZE));
651 seq_printf(m, arc_cache_mumbojumbo(cpu_id, str, PAGE_SIZE));
652 seq_printf(m, arc_extn_mumbojumbo(cpu_id, str, PAGE_SIZE));
653 seq_printf(m, arc_platform_smp_cpuinfo());
654
655 free_page((unsigned long)str);
656 done:
657 seq_printf(m, "\n");
658
659 return 0;
660 }
661
662 static void *c_start(struct seq_file *m, loff_t *pos)
663 {
664
665
666
667
668
669
670 return *pos < nr_cpu_ids ? cpu_to_ptr(*pos) : NULL;
671 }
672
673 static void *c_next(struct seq_file *m, void *v, loff_t *pos)
674 {
675 ++*pos;
676 return c_start(m, pos);
677 }
678
679 static void c_stop(struct seq_file *m, void *v)
680 {
681 }
682
683 const struct seq_operations cpuinfo_op = {
684 .start = c_start,
685 .next = c_next,
686 .stop = c_stop,
687 .show = show_cpuinfo
688 };
689
690 static DEFINE_PER_CPU(struct cpu, cpu_topology);
691
692 static int __init topology_init(void)
693 {
694 int cpu;
695
696 for_each_present_cpu(cpu)
697 register_cpu(&per_cpu(cpu_topology, cpu), cpu);
698
699 return 0;
700 }
701
702 subsys_initcall(topology_init);