blob: f3931e38fac0fb58a9bb6262e0fe7d9da33a6269 [file] [log] [blame]
Tang Yuantian555eae92013-04-09 16:46:26 +08001/*
2 * Copyright 2013 Freescale Semiconductor, Inc.
3 *
4 * This program is free software; you can redistribute it and/or modify
5 * it under the terms of the GNU General Public License version 2 as
6 * published by the Free Software Foundation.
7 *
Tang Yuantian93a17c02015-01-15 14:03:41 +08008 * clock driver for Freescale QorIQ SoCs.
Tang Yuantian555eae92013-04-09 16:46:26 +08009 */
Emil Medvec88b2b62015-01-21 04:03:29 -060010
11#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
12
Scott Wood0dfc86b2015-09-19 23:29:54 -050013#include <linux/clk.h>
Tang Yuantian555eae92013-04-09 16:46:26 +080014#include <linux/clk-provider.h>
Scott Wood0dfc86b2015-09-19 23:29:54 -050015#include <linux/fsl/guts.h>
Tang Yuantian555eae92013-04-09 16:46:26 +080016#include <linux/io.h>
17#include <linux/kernel.h>
18#include <linux/module.h>
Rob Herringc11eede2013-11-10 23:19:08 -060019#include <linux/of_address.h>
Tang Yuantian555eae92013-04-09 16:46:26 +080020#include <linux/of_platform.h>
21#include <linux/of.h>
22#include <linux/slab.h>
23
Scott Wood0dfc86b2015-09-19 23:29:54 -050024#define PLL_DIV1 0
25#define PLL_DIV2 1
26#define PLL_DIV3 2
27#define PLL_DIV4 3
28
29#define PLATFORM_PLL 0
30#define CGA_PLL1 1
31#define CGA_PLL2 2
32#define CGA_PLL3 3
33#define CGA_PLL4 4 /* only on clockgen-1.0, which lacks CGB */
34#define CGB_PLL1 4
35#define CGB_PLL2 5
36
37struct clockgen_pll_div {
38 struct clk *clk;
39 char name[32];
Tang Yuantian555eae92013-04-09 16:46:26 +080040};
41
Scott Wood0dfc86b2015-09-19 23:29:54 -050042struct clockgen_pll {
43 struct clockgen_pll_div div[4];
44};
Tang Yuantian555eae92013-04-09 16:46:26 +080045
Scott Wood0dfc86b2015-09-19 23:29:54 -050046#define CLKSEL_VALID 1
47#define CLKSEL_80PCT 2 /* Only allowed if PLL <= 80% of max cpu freq */
48
49struct clockgen_sourceinfo {
50 u32 flags; /* CLKSEL_xxx */
51 int pll; /* CGx_PLLn */
52 int div; /* PLL_DIVn */
53};
54
55#define NUM_MUX_PARENTS 16
56
57struct clockgen_muxinfo {
58 struct clockgen_sourceinfo clksel[NUM_MUX_PARENTS];
59};
60
61#define NUM_HWACCEL 5
62#define NUM_CMUX 8
63
64struct clockgen;
65
66/*
67 * cmux freq must be >= platform pll.
68 * If not set, cmux freq must be >= platform pll/2
69 */
70#define CG_CMUX_GE_PLAT 1
Scott Wood9e19ca22015-09-19 23:29:55 -050071
Scott Wood0dfc86b2015-09-19 23:29:54 -050072#define CG_PLL_8BIT 2 /* PLLCnGSR[CFG] is 8 bits, not 6 */
Scott Wood9e19ca22015-09-19 23:29:55 -050073#define CG_VER3 4 /* version 3 cg: reg layout different */
74#define CG_LITTLE_ENDIAN 8
Scott Wood0dfc86b2015-09-19 23:29:54 -050075
76struct clockgen_chipinfo {
77 const char *compat, *guts_compat;
78 const struct clockgen_muxinfo *cmux_groups[2];
79 const struct clockgen_muxinfo *hwaccel[NUM_HWACCEL];
80 void (*init_periph)(struct clockgen *cg);
81 int cmux_to_group[NUM_CMUX]; /* -1 terminates if fewer than NUM_CMUX */
82 u32 pll_mask; /* 1 << n bit set if PLL n is valid */
83 u32 flags; /* CG_xxx */
84};
85
86struct clockgen {
87 struct device_node *node;
88 void __iomem *regs;
89 struct clockgen_chipinfo info; /* mutable copy */
Scott Wood80b4ae72017-03-20 10:37:23 +080090 struct clk *sysclk, *coreclk;
Scott Wood0dfc86b2015-09-19 23:29:54 -050091 struct clockgen_pll pll[6];
92 struct clk *cmux[NUM_CMUX];
93 struct clk *hwaccel[NUM_HWACCEL];
94 struct clk *fman[2];
95 struct ccsr_guts __iomem *guts;
96};
97
98static struct clockgen clockgen;
99
Scott Wood9e19ca22015-09-19 23:29:55 -0500100static void cg_out(struct clockgen *cg, u32 val, u32 __iomem *reg)
101{
102 if (cg->info.flags & CG_LITTLE_ENDIAN)
103 iowrite32(val, reg);
104 else
105 iowrite32be(val, reg);
106}
107
108static u32 cg_in(struct clockgen *cg, u32 __iomem *reg)
109{
110 u32 val;
111
112 if (cg->info.flags & CG_LITTLE_ENDIAN)
113 val = ioread32(reg);
114 else
115 val = ioread32be(reg);
116
117 return val;
118}
119
Scott Wood0dfc86b2015-09-19 23:29:54 -0500120static const struct clockgen_muxinfo p2041_cmux_grp1 = {
121 {
122 [0] = { CLKSEL_VALID, CGA_PLL1, PLL_DIV1 },
123 [1] = { CLKSEL_VALID, CGA_PLL1, PLL_DIV2 },
124 [4] = { CLKSEL_VALID, CGA_PLL2, PLL_DIV1 },
125 }
126};
127
128static const struct clockgen_muxinfo p2041_cmux_grp2 = {
129 {
130 [0] = { CLKSEL_VALID, CGA_PLL1, PLL_DIV1 },
Scott Wood2c7693e2015-10-22 23:21:46 -0500131 [4] = { CLKSEL_VALID, CGA_PLL2, PLL_DIV1 },
132 [5] = { CLKSEL_VALID, CGA_PLL2, PLL_DIV2 },
Scott Wood0dfc86b2015-09-19 23:29:54 -0500133 }
134};
135
136static const struct clockgen_muxinfo p5020_cmux_grp1 = {
137 {
138 [0] = { CLKSEL_VALID, CGA_PLL1, PLL_DIV1 },
139 [1] = { CLKSEL_VALID, CGA_PLL1, PLL_DIV2 },
140 [4] = { CLKSEL_VALID | CLKSEL_80PCT, CGA_PLL2, PLL_DIV1 },
141 }
142};
143
144static const struct clockgen_muxinfo p5020_cmux_grp2 = {
145 {
146 [0] = { CLKSEL_VALID | CLKSEL_80PCT, CGA_PLL1, PLL_DIV1 },
147 [4] = { CLKSEL_VALID, CGA_PLL2, PLL_DIV1 },
148 [5] = { CLKSEL_VALID, CGA_PLL2, PLL_DIV2 },
149 }
150};
151
152static const struct clockgen_muxinfo p5040_cmux_grp1 = {
153 {
154 [0] = { CLKSEL_VALID, CGA_PLL1, PLL_DIV1 },
155 [1] = { CLKSEL_VALID, CGA_PLL1, PLL_DIV2 },
156 [4] = { CLKSEL_VALID | CLKSEL_80PCT, CGA_PLL2, PLL_DIV1 },
157 [5] = { CLKSEL_VALID | CLKSEL_80PCT, CGA_PLL2, PLL_DIV2 },
158 }
159};
160
161static const struct clockgen_muxinfo p5040_cmux_grp2 = {
162 {
163 [0] = { CLKSEL_VALID | CLKSEL_80PCT, CGA_PLL1, PLL_DIV1 },
164 [1] = { CLKSEL_VALID | CLKSEL_80PCT, CGA_PLL1, PLL_DIV2 },
165 [4] = { CLKSEL_VALID, CGA_PLL2, PLL_DIV1 },
166 [5] = { CLKSEL_VALID, CGA_PLL2, PLL_DIV2 },
167 }
168};
169
170static const struct clockgen_muxinfo p4080_cmux_grp1 = {
171 {
172 [0] = { CLKSEL_VALID, CGA_PLL1, PLL_DIV1 },
173 [1] = { CLKSEL_VALID, CGA_PLL1, PLL_DIV2 },
174 [4] = { CLKSEL_VALID, CGA_PLL2, PLL_DIV1 },
175 [5] = { CLKSEL_VALID, CGA_PLL2, PLL_DIV2 },
176 [8] = { CLKSEL_VALID | CLKSEL_80PCT, CGA_PLL3, PLL_DIV1 },
177 }
178};
179
180static const struct clockgen_muxinfo p4080_cmux_grp2 = {
181 {
182 [0] = { CLKSEL_VALID | CLKSEL_80PCT, CGA_PLL1, PLL_DIV1 },
183 [8] = { CLKSEL_VALID, CGA_PLL3, PLL_DIV1 },
184 [9] = { CLKSEL_VALID, CGA_PLL3, PLL_DIV2 },
185 [12] = { CLKSEL_VALID, CGA_PLL4, PLL_DIV1 },
186 [13] = { CLKSEL_VALID, CGA_PLL4, PLL_DIV2 },
187 }
188};
189
190static const struct clockgen_muxinfo t1023_cmux = {
191 {
192 [0] = { CLKSEL_VALID, CGA_PLL1, PLL_DIV1 },
193 [1] = { CLKSEL_VALID, CGA_PLL1, PLL_DIV2 },
194 }
195};
196
197static const struct clockgen_muxinfo t1040_cmux = {
198 {
199 [0] = { CLKSEL_VALID, CGA_PLL1, PLL_DIV1 },
200 [1] = { CLKSEL_VALID, CGA_PLL1, PLL_DIV2 },
201 [4] = { CLKSEL_VALID, CGA_PLL2, PLL_DIV1 },
202 [5] = { CLKSEL_VALID, CGA_PLL2, PLL_DIV2 },
203 }
204};
205
206
207static const struct clockgen_muxinfo clockgen2_cmux_cga = {
208 {
209 { CLKSEL_VALID, CGA_PLL1, PLL_DIV1 },
210 { CLKSEL_VALID, CGA_PLL1, PLL_DIV2 },
211 { CLKSEL_VALID, CGA_PLL1, PLL_DIV4 },
212 {},
213 { CLKSEL_VALID, CGA_PLL2, PLL_DIV1 },
214 { CLKSEL_VALID, CGA_PLL2, PLL_DIV2 },
215 { CLKSEL_VALID, CGA_PLL2, PLL_DIV4 },
216 {},
217 { CLKSEL_VALID, CGA_PLL3, PLL_DIV1 },
218 { CLKSEL_VALID, CGA_PLL3, PLL_DIV2 },
219 { CLKSEL_VALID, CGA_PLL3, PLL_DIV4 },
220 },
221};
222
223static const struct clockgen_muxinfo clockgen2_cmux_cga12 = {
224 {
225 { CLKSEL_VALID, CGA_PLL1, PLL_DIV1 },
226 { CLKSEL_VALID, CGA_PLL1, PLL_DIV2 },
227 { CLKSEL_VALID, CGA_PLL1, PLL_DIV4 },
228 {},
229 { CLKSEL_VALID, CGA_PLL2, PLL_DIV1 },
230 { CLKSEL_VALID, CGA_PLL2, PLL_DIV2 },
231 { CLKSEL_VALID, CGA_PLL2, PLL_DIV4 },
232 },
233};
234
235static const struct clockgen_muxinfo clockgen2_cmux_cgb = {
236 {
237 { CLKSEL_VALID, CGB_PLL1, PLL_DIV1 },
238 { CLKSEL_VALID, CGB_PLL1, PLL_DIV2 },
239 { CLKSEL_VALID, CGB_PLL1, PLL_DIV4 },
240 {},
241 { CLKSEL_VALID, CGB_PLL2, PLL_DIV1 },
242 { CLKSEL_VALID, CGB_PLL2, PLL_DIV2 },
243 { CLKSEL_VALID, CGB_PLL2, PLL_DIV4 },
244 },
245};
246
Hou Zhiqiange994412c2015-10-23 16:01:21 +0800247static const struct clockgen_muxinfo ls1043a_hwa1 = {
248 {
249 {},
250 {},
251 { CLKSEL_VALID, CGA_PLL1, PLL_DIV2 },
252 { CLKSEL_VALID, CGA_PLL1, PLL_DIV3 },
253 {},
254 {},
255 { CLKSEL_VALID, CGA_PLL2, PLL_DIV2 },
256 { CLKSEL_VALID, CGA_PLL2, PLL_DIV3 },
257 },
258};
259
260static const struct clockgen_muxinfo ls1043a_hwa2 = {
261 {
262 {},
263 { CLKSEL_VALID, CGA_PLL2, PLL_DIV1 },
264 {},
265 { CLKSEL_VALID, CGA_PLL2, PLL_DIV3 },
266 },
267};
268
Mingkai Hu80e52192016-09-07 11:48:30 +0800269static const struct clockgen_muxinfo ls1046a_hwa1 = {
270 {
271 {},
272 {},
273 { CLKSEL_VALID, CGA_PLL1, PLL_DIV2 },
274 { CLKSEL_VALID, CGA_PLL1, PLL_DIV3 },
275 { CLKSEL_VALID, CGA_PLL1, PLL_DIV4 },
276 { CLKSEL_VALID, PLATFORM_PLL, PLL_DIV1 },
277 { CLKSEL_VALID, CGA_PLL2, PLL_DIV2 },
278 { CLKSEL_VALID, CGA_PLL2, PLL_DIV3 },
279 },
280};
281
282static const struct clockgen_muxinfo ls1046a_hwa2 = {
283 {
284 {},
285 { CLKSEL_VALID, CGA_PLL2, PLL_DIV1 },
286 { CLKSEL_VALID, CGA_PLL2, PLL_DIV2 },
287 { CLKSEL_VALID, CGA_PLL2, PLL_DIV3 },
288 {},
289 {},
290 { CLKSEL_VALID, CGA_PLL1, PLL_DIV2 },
291 },
292};
293
Tang Yuantian44709352016-11-24 10:36:55 +0800294static const struct clockgen_muxinfo ls1012a_cmux = {
295 {
296 [0] = { CLKSEL_VALID, CGA_PLL1, PLL_DIV1 },
297 {},
298 [2] = { CLKSEL_VALID, CGA_PLL1, PLL_DIV2 },
299 }
300};
301
Scott Wood0dfc86b2015-09-19 23:29:54 -0500302static const struct clockgen_muxinfo t1023_hwa1 = {
303 {
304 {},
305 { CLKSEL_VALID, CGA_PLL1, PLL_DIV1 },
306 { CLKSEL_VALID, CGA_PLL1, PLL_DIV2 },
307 { CLKSEL_VALID, CGA_PLL1, PLL_DIV3 },
308 },
309};
310
311static const struct clockgen_muxinfo t1023_hwa2 = {
312 {
313 [6] = { CLKSEL_VALID, CGA_PLL1, PLL_DIV2 },
314 },
315};
316
317static const struct clockgen_muxinfo t2080_hwa1 = {
318 {
319 {},
320 { CLKSEL_VALID, CGA_PLL1, PLL_DIV1 },
321 { CLKSEL_VALID, CGA_PLL1, PLL_DIV2 },
322 { CLKSEL_VALID, CGA_PLL1, PLL_DIV3 },
323 { CLKSEL_VALID, CGA_PLL1, PLL_DIV4 },
324 { CLKSEL_VALID, PLATFORM_PLL, PLL_DIV1 },
325 { CLKSEL_VALID, CGA_PLL2, PLL_DIV2 },
326 { CLKSEL_VALID, CGA_PLL2, PLL_DIV3 },
327 },
328};
329
330static const struct clockgen_muxinfo t2080_hwa2 = {
331 {
332 {},
333 { CLKSEL_VALID, CGA_PLL2, PLL_DIV1 },
334 { CLKSEL_VALID, CGA_PLL2, PLL_DIV2 },
335 { CLKSEL_VALID, CGA_PLL2, PLL_DIV3 },
336 { CLKSEL_VALID, CGA_PLL2, PLL_DIV4 },
337 { CLKSEL_VALID, PLATFORM_PLL, PLL_DIV1 },
338 { CLKSEL_VALID, CGA_PLL1, PLL_DIV2 },
339 { CLKSEL_VALID, CGA_PLL1, PLL_DIV3 },
340 },
341};
342
343static const struct clockgen_muxinfo t4240_hwa1 = {
344 {
345 { CLKSEL_VALID, PLATFORM_PLL, PLL_DIV2 },
346 { CLKSEL_VALID, CGA_PLL1, PLL_DIV1 },
347 { CLKSEL_VALID, CGA_PLL1, PLL_DIV2 },
348 { CLKSEL_VALID, CGA_PLL1, PLL_DIV3 },
349 { CLKSEL_VALID, CGA_PLL1, PLL_DIV4 },
350 {},
351 { CLKSEL_VALID, CGA_PLL2, PLL_DIV2 },
352 { CLKSEL_VALID, CGA_PLL2, PLL_DIV3 },
353 },
354};
355
356static const struct clockgen_muxinfo t4240_hwa4 = {
357 {
358 [2] = { CLKSEL_VALID, CGB_PLL1, PLL_DIV2 },
359 [3] = { CLKSEL_VALID, CGB_PLL1, PLL_DIV3 },
360 [4] = { CLKSEL_VALID, CGB_PLL1, PLL_DIV4 },
361 [5] = { CLKSEL_VALID, PLATFORM_PLL, PLL_DIV1 },
362 [6] = { CLKSEL_VALID, CGB_PLL2, PLL_DIV2 },
363 },
364};
365
366static const struct clockgen_muxinfo t4240_hwa5 = {
367 {
368 [2] = { CLKSEL_VALID, CGB_PLL2, PLL_DIV2 },
369 [3] = { CLKSEL_VALID, CGB_PLL2, PLL_DIV3 },
370 [4] = { CLKSEL_VALID, CGB_PLL2, PLL_DIV4 },
371 [5] = { CLKSEL_VALID, PLATFORM_PLL, PLL_DIV1 },
372 [6] = { CLKSEL_VALID, CGB_PLL1, PLL_DIV2 },
373 [7] = { CLKSEL_VALID, CGB_PLL1, PLL_DIV3 },
374 },
375};
376
377#define RCWSR7_FM1_CLK_SEL 0x40000000
378#define RCWSR7_FM2_CLK_SEL 0x20000000
379#define RCWSR7_HWA_ASYNC_DIV 0x04000000
380
381static void __init p2041_init_periph(struct clockgen *cg)
Tang Yuantian555eae92013-04-09 16:46:26 +0800382{
Scott Wood0dfc86b2015-09-19 23:29:54 -0500383 u32 reg;
384
385 reg = ioread32be(&cg->guts->rcwsr[7]);
386
387 if (reg & RCWSR7_FM1_CLK_SEL)
388 cg->fman[0] = cg->pll[CGA_PLL2].div[PLL_DIV2].clk;
389 else
390 cg->fman[0] = cg->pll[PLATFORM_PLL].div[PLL_DIV2].clk;
391}
392
393static void __init p4080_init_periph(struct clockgen *cg)
394{
395 u32 reg;
396
397 reg = ioread32be(&cg->guts->rcwsr[7]);
398
399 if (reg & RCWSR7_FM1_CLK_SEL)
400 cg->fman[0] = cg->pll[CGA_PLL3].div[PLL_DIV2].clk;
401 else
402 cg->fman[0] = cg->pll[PLATFORM_PLL].div[PLL_DIV2].clk;
403
404 if (reg & RCWSR7_FM2_CLK_SEL)
405 cg->fman[1] = cg->pll[CGA_PLL3].div[PLL_DIV2].clk;
406 else
407 cg->fman[1] = cg->pll[PLATFORM_PLL].div[PLL_DIV2].clk;
408}
409
410static void __init p5020_init_periph(struct clockgen *cg)
411{
412 u32 reg;
413 int div = PLL_DIV2;
414
415 reg = ioread32be(&cg->guts->rcwsr[7]);
416 if (reg & RCWSR7_HWA_ASYNC_DIV)
417 div = PLL_DIV4;
418
419 if (reg & RCWSR7_FM1_CLK_SEL)
420 cg->fman[0] = cg->pll[CGA_PLL2].div[div].clk;
421 else
422 cg->fman[0] = cg->pll[PLATFORM_PLL].div[PLL_DIV2].clk;
423}
424
425static void __init p5040_init_periph(struct clockgen *cg)
426{
427 u32 reg;
428 int div = PLL_DIV2;
429
430 reg = ioread32be(&cg->guts->rcwsr[7]);
431 if (reg & RCWSR7_HWA_ASYNC_DIV)
432 div = PLL_DIV4;
433
434 if (reg & RCWSR7_FM1_CLK_SEL)
435 cg->fman[0] = cg->pll[CGA_PLL3].div[div].clk;
436 else
437 cg->fman[0] = cg->pll[PLATFORM_PLL].div[PLL_DIV2].clk;
438
439 if (reg & RCWSR7_FM2_CLK_SEL)
440 cg->fman[1] = cg->pll[CGA_PLL3].div[div].clk;
441 else
442 cg->fman[1] = cg->pll[PLATFORM_PLL].div[PLL_DIV2].clk;
443}
444
445static void __init t1023_init_periph(struct clockgen *cg)
446{
447 cg->fman[0] = cg->hwaccel[1];
448}
449
450static void __init t1040_init_periph(struct clockgen *cg)
451{
452 cg->fman[0] = cg->pll[PLATFORM_PLL].div[PLL_DIV1].clk;
453}
454
455static void __init t2080_init_periph(struct clockgen *cg)
456{
457 cg->fman[0] = cg->hwaccel[0];
458}
459
460static void __init t4240_init_periph(struct clockgen *cg)
461{
462 cg->fman[0] = cg->hwaccel[3];
463 cg->fman[1] = cg->hwaccel[4];
464}
465
466static const struct clockgen_chipinfo chipinfo[] = {
467 {
468 .compat = "fsl,b4420-clockgen",
469 .guts_compat = "fsl,b4860-device-config",
470 .init_periph = t2080_init_periph,
471 .cmux_groups = {
472 &clockgen2_cmux_cga12, &clockgen2_cmux_cgb
473 },
474 .hwaccel = {
475 &t2080_hwa1
476 },
477 .cmux_to_group = {
478 0, 1, 1, 1, -1
479 },
480 .pll_mask = 0x3f,
481 .flags = CG_PLL_8BIT,
482 },
483 {
484 .compat = "fsl,b4860-clockgen",
485 .guts_compat = "fsl,b4860-device-config",
486 .init_periph = t2080_init_periph,
487 .cmux_groups = {
488 &clockgen2_cmux_cga12, &clockgen2_cmux_cgb
489 },
490 .hwaccel = {
491 &t2080_hwa1
492 },
493 .cmux_to_group = {
494 0, 1, 1, 1, -1
495 },
496 .pll_mask = 0x3f,
497 .flags = CG_PLL_8BIT,
498 },
499 {
500 .compat = "fsl,ls1021a-clockgen",
501 .cmux_groups = {
502 &t1023_cmux
503 },
504 .cmux_to_group = {
505 0, -1
506 },
507 .pll_mask = 0x03,
508 },
509 {
Hou Zhiqiange994412c2015-10-23 16:01:21 +0800510 .compat = "fsl,ls1043a-clockgen",
511 .init_periph = t2080_init_periph,
512 .cmux_groups = {
513 &t1040_cmux
514 },
515 .hwaccel = {
516 &ls1043a_hwa1, &ls1043a_hwa2
517 },
518 .cmux_to_group = {
519 0, -1
520 },
521 .pll_mask = 0x07,
522 .flags = CG_PLL_8BIT,
523 },
524 {
Mingkai Hu80e52192016-09-07 11:48:30 +0800525 .compat = "fsl,ls1046a-clockgen",
526 .init_periph = t2080_init_periph,
527 .cmux_groups = {
528 &t1040_cmux
529 },
530 .hwaccel = {
531 &ls1046a_hwa1, &ls1046a_hwa2
532 },
533 .cmux_to_group = {
534 0, -1
535 },
536 .pll_mask = 0x07,
537 .flags = CG_PLL_8BIT,
538 },
539 {
Tang Yuantian44709352016-11-24 10:36:55 +0800540 .compat = "fsl,ls1012a-clockgen",
541 .cmux_groups = {
542 &ls1012a_cmux
543 },
544 .cmux_to_group = {
545 0, -1
546 },
547 .pll_mask = 0x03,
548 },
549 {
Scott Wood9e19ca22015-09-19 23:29:55 -0500550 .compat = "fsl,ls2080a-clockgen",
551 .cmux_groups = {
552 &clockgen2_cmux_cga12, &clockgen2_cmux_cgb
553 },
554 .cmux_to_group = {
555 0, 0, 1, 1, -1
556 },
557 .pll_mask = 0x37,
558 .flags = CG_VER3 | CG_LITTLE_ENDIAN,
559 },
560 {
Scott Wood0dfc86b2015-09-19 23:29:54 -0500561 .compat = "fsl,p2041-clockgen",
562 .guts_compat = "fsl,qoriq-device-config-1.0",
563 .init_periph = p2041_init_periph,
564 .cmux_groups = {
565 &p2041_cmux_grp1, &p2041_cmux_grp2
566 },
567 .cmux_to_group = {
568 0, 0, 1, 1, -1
569 },
570 .pll_mask = 0x07,
571 },
572 {
573 .compat = "fsl,p3041-clockgen",
574 .guts_compat = "fsl,qoriq-device-config-1.0",
575 .init_periph = p2041_init_periph,
576 .cmux_groups = {
577 &p2041_cmux_grp1, &p2041_cmux_grp2
578 },
579 .cmux_to_group = {
580 0, 0, 1, 1, -1
581 },
582 .pll_mask = 0x07,
583 },
584 {
585 .compat = "fsl,p4080-clockgen",
586 .guts_compat = "fsl,qoriq-device-config-1.0",
587 .init_periph = p4080_init_periph,
588 .cmux_groups = {
589 &p4080_cmux_grp1, &p4080_cmux_grp2
590 },
591 .cmux_to_group = {
592 0, 0, 0, 0, 1, 1, 1, 1
593 },
594 .pll_mask = 0x1f,
595 },
596 {
597 .compat = "fsl,p5020-clockgen",
598 .guts_compat = "fsl,qoriq-device-config-1.0",
599 .init_periph = p5020_init_periph,
600 .cmux_groups = {
601 &p2041_cmux_grp1, &p2041_cmux_grp2
602 },
603 .cmux_to_group = {
604 0, 1, -1
605 },
606 .pll_mask = 0x07,
607 },
608 {
609 .compat = "fsl,p5040-clockgen",
610 .guts_compat = "fsl,p5040-device-config",
611 .init_periph = p5040_init_periph,
612 .cmux_groups = {
613 &p5040_cmux_grp1, &p5040_cmux_grp2
614 },
615 .cmux_to_group = {
616 0, 0, 1, 1, -1
617 },
618 .pll_mask = 0x0f,
619 },
620 {
621 .compat = "fsl,t1023-clockgen",
622 .guts_compat = "fsl,t1023-device-config",
623 .init_periph = t1023_init_periph,
624 .cmux_groups = {
625 &t1023_cmux
626 },
627 .hwaccel = {
628 &t1023_hwa1, &t1023_hwa2
629 },
630 .cmux_to_group = {
631 0, 0, -1
632 },
633 .pll_mask = 0x03,
634 .flags = CG_PLL_8BIT,
635 },
636 {
637 .compat = "fsl,t1040-clockgen",
638 .guts_compat = "fsl,t1040-device-config",
639 .init_periph = t1040_init_periph,
640 .cmux_groups = {
641 &t1040_cmux
642 },
643 .cmux_to_group = {
644 0, 0, 0, 0, -1
645 },
646 .pll_mask = 0x07,
647 .flags = CG_PLL_8BIT,
648 },
649 {
650 .compat = "fsl,t2080-clockgen",
651 .guts_compat = "fsl,t2080-device-config",
652 .init_periph = t2080_init_periph,
653 .cmux_groups = {
654 &clockgen2_cmux_cga12
655 },
656 .hwaccel = {
657 &t2080_hwa1, &t2080_hwa2
658 },
659 .cmux_to_group = {
660 0, -1
661 },
662 .pll_mask = 0x07,
663 .flags = CG_PLL_8BIT,
664 },
665 {
666 .compat = "fsl,t4240-clockgen",
667 .guts_compat = "fsl,t4240-device-config",
668 .init_periph = t4240_init_periph,
669 .cmux_groups = {
670 &clockgen2_cmux_cga, &clockgen2_cmux_cgb
671 },
672 .hwaccel = {
673 &t4240_hwa1, NULL, NULL, &t4240_hwa4, &t4240_hwa5
674 },
675 .cmux_to_group = {
676 0, 0, 1, -1
677 },
678 .pll_mask = 0x3f,
679 .flags = CG_PLL_8BIT,
680 },
681 {},
682};
683
684struct mux_hwclock {
685 struct clk_hw hw;
686 struct clockgen *cg;
687 const struct clockgen_muxinfo *info;
688 u32 __iomem *reg;
689 u8 parent_to_clksel[NUM_MUX_PARENTS];
690 s8 clksel_to_parent[NUM_MUX_PARENTS];
691 int num_parents;
692};
693
694#define to_mux_hwclock(p) container_of(p, struct mux_hwclock, hw)
695#define CLKSEL_MASK 0x78000000
696#define CLKSEL_SHIFT 27
697
698static int mux_set_parent(struct clk_hw *hw, u8 idx)
699{
700 struct mux_hwclock *hwc = to_mux_hwclock(hw);
Tang Yuantian555eae92013-04-09 16:46:26 +0800701 u32 clksel;
702
Scott Wood0dfc86b2015-09-19 23:29:54 -0500703 if (idx >= hwc->num_parents)
704 return -EINVAL;
705
706 clksel = hwc->parent_to_clksel[idx];
Scott Wood9e19ca22015-09-19 23:29:55 -0500707 cg_out(hwc->cg, (clksel << CLKSEL_SHIFT) & CLKSEL_MASK, hwc->reg);
Tang Yuantian555eae92013-04-09 16:46:26 +0800708
709 return 0;
710}
711
Scott Wood0dfc86b2015-09-19 23:29:54 -0500712static u8 mux_get_parent(struct clk_hw *hw)
Tang Yuantian555eae92013-04-09 16:46:26 +0800713{
Scott Wood0dfc86b2015-09-19 23:29:54 -0500714 struct mux_hwclock *hwc = to_mux_hwclock(hw);
Tang Yuantian555eae92013-04-09 16:46:26 +0800715 u32 clksel;
Scott Wood0dfc86b2015-09-19 23:29:54 -0500716 s8 ret;
Tang Yuantian555eae92013-04-09 16:46:26 +0800717
Scott Wood9e19ca22015-09-19 23:29:55 -0500718 clksel = (cg_in(hwc->cg, hwc->reg) & CLKSEL_MASK) >> CLKSEL_SHIFT;
Tang Yuantian555eae92013-04-09 16:46:26 +0800719
Scott Wood0dfc86b2015-09-19 23:29:54 -0500720 ret = hwc->clksel_to_parent[clksel];
721 if (ret < 0) {
722 pr_err("%s: mux at %p has bad clksel\n", __func__, hwc->reg);
723 return 0;
724 }
725
726 return ret;
Tang Yuantian555eae92013-04-09 16:46:26 +0800727}
728
Emil Medve334680d2015-01-21 04:03:27 -0600729static const struct clk_ops cmux_ops = {
Scott Wood0dfc86b2015-09-19 23:29:54 -0500730 .get_parent = mux_get_parent,
731 .set_parent = mux_set_parent,
Tang Yuantian555eae92013-04-09 16:46:26 +0800732};
733
Scott Wood0dfc86b2015-09-19 23:29:54 -0500734/*
735 * Don't allow setting for now, as the clock options haven't been
736 * sanitized for additional restrictions.
737 */
738static const struct clk_ops hwaccel_ops = {
739 .get_parent = mux_get_parent,
740};
741
742static const struct clockgen_pll_div *get_pll_div(struct clockgen *cg,
743 struct mux_hwclock *hwc,
744 int idx)
745{
746 int pll, div;
747
748 if (!(hwc->info->clksel[idx].flags & CLKSEL_VALID))
749 return NULL;
750
751 pll = hwc->info->clksel[idx].pll;
752 div = hwc->info->clksel[idx].div;
753
754 return &cg->pll[pll].div[div];
755}
756
757static struct clk * __init create_mux_common(struct clockgen *cg,
758 struct mux_hwclock *hwc,
759 const struct clk_ops *ops,
760 unsigned long min_rate,
Scott Wood7c1c5412016-10-17 13:42:23 -0500761 unsigned long max_rate,
Scott Wood0dfc86b2015-09-19 23:29:54 -0500762 unsigned long pct80_rate,
763 const char *fmt, int idx)
764{
765 struct clk_init_data init = {};
766 struct clk *clk;
767 const struct clockgen_pll_div *div;
768 const char *parent_names[NUM_MUX_PARENTS];
769 char name[32];
770 int i, j;
771
772 snprintf(name, sizeof(name), fmt, idx);
773
774 for (i = 0, j = 0; i < NUM_MUX_PARENTS; i++) {
775 unsigned long rate;
776
777 hwc->clksel_to_parent[i] = -1;
778
779 div = get_pll_div(cg, hwc, i);
780 if (!div)
781 continue;
782
783 rate = clk_get_rate(div->clk);
784
785 if (hwc->info->clksel[i].flags & CLKSEL_80PCT &&
786 rate > pct80_rate)
787 continue;
788 if (rate < min_rate)
789 continue;
Scott Wood7c1c5412016-10-17 13:42:23 -0500790 if (rate > max_rate)
791 continue;
Scott Wood0dfc86b2015-09-19 23:29:54 -0500792
793 parent_names[j] = div->name;
794 hwc->parent_to_clksel[j] = i;
795 hwc->clksel_to_parent[i] = j;
796 j++;
797 }
798
799 init.name = name;
800 init.ops = ops;
801 init.parent_names = parent_names;
802 init.num_parents = hwc->num_parents = j;
803 init.flags = 0;
804 hwc->hw.init = &init;
805 hwc->cg = cg;
806
807 clk = clk_register(NULL, &hwc->hw);
808 if (IS_ERR(clk)) {
809 pr_err("%s: Couldn't register %s: %ld\n", __func__, name,
810 PTR_ERR(clk));
811 kfree(hwc);
812 return NULL;
813 }
814
815 return clk;
816}
817
818static struct clk * __init create_one_cmux(struct clockgen *cg, int idx)
819{
820 struct mux_hwclock *hwc;
821 const struct clockgen_pll_div *div;
822 unsigned long plat_rate, min_rate;
Scott Wood7c1c5412016-10-17 13:42:23 -0500823 u64 max_rate, pct80_rate;
Scott Wood0dfc86b2015-09-19 23:29:54 -0500824 u32 clksel;
825
826 hwc = kzalloc(sizeof(*hwc), GFP_KERNEL);
827 if (!hwc)
828 return NULL;
829
Tang Yuantian89641932016-08-15 15:28:20 +0800830 if (cg->info.flags & CG_VER3)
831 hwc->reg = cg->regs + 0x70000 + 0x20 * idx;
832 else
833 hwc->reg = cg->regs + 0x20 * idx;
834
Scott Wood0dfc86b2015-09-19 23:29:54 -0500835 hwc->info = cg->info.cmux_groups[cg->info.cmux_to_group[idx]];
836
837 /*
838 * Find the rate for the default clksel, and treat it as the
839 * maximum rated core frequency. If this is an incorrect
840 * assumption, certain clock options (possibly including the
841 * default clksel) may be inappropriately excluded on certain
842 * chips.
843 */
Scott Wood9e19ca22015-09-19 23:29:55 -0500844 clksel = (cg_in(cg, hwc->reg) & CLKSEL_MASK) >> CLKSEL_SHIFT;
Scott Wood0dfc86b2015-09-19 23:29:54 -0500845 div = get_pll_div(cg, hwc, clksel);
Sudip Mukherjee279104e2015-11-23 15:36:50 +0530846 if (!div) {
847 kfree(hwc);
Scott Wood0dfc86b2015-09-19 23:29:54 -0500848 return NULL;
Sudip Mukherjee279104e2015-11-23 15:36:50 +0530849 }
Scott Wood0dfc86b2015-09-19 23:29:54 -0500850
Scott Wood7c1c5412016-10-17 13:42:23 -0500851 max_rate = clk_get_rate(div->clk);
852 pct80_rate = max_rate * 8;
Scott Wood0dfc86b2015-09-19 23:29:54 -0500853 do_div(pct80_rate, 10);
854
855 plat_rate = clk_get_rate(cg->pll[PLATFORM_PLL].div[PLL_DIV1].clk);
856
857 if (cg->info.flags & CG_CMUX_GE_PLAT)
858 min_rate = plat_rate;
859 else
860 min_rate = plat_rate / 2;
861
Scott Wood7c1c5412016-10-17 13:42:23 -0500862 return create_mux_common(cg, hwc, &cmux_ops, min_rate, max_rate,
Scott Wood0dfc86b2015-09-19 23:29:54 -0500863 pct80_rate, "cg-cmux%d", idx);
864}
865
866static struct clk * __init create_one_hwaccel(struct clockgen *cg, int idx)
867{
868 struct mux_hwclock *hwc;
869
870 hwc = kzalloc(sizeof(*hwc), GFP_KERNEL);
871 if (!hwc)
872 return NULL;
873
874 hwc->reg = cg->regs + 0x20 * idx + 0x10;
875 hwc->info = cg->info.hwaccel[idx];
876
Scott Wood7c1c5412016-10-17 13:42:23 -0500877 return create_mux_common(cg, hwc, &hwaccel_ops, 0, ULONG_MAX, 0,
Scott Wood0dfc86b2015-09-19 23:29:54 -0500878 "cg-hwaccel%d", idx);
879}
880
881static void __init create_muxes(struct clockgen *cg)
882{
883 int i;
884
885 for (i = 0; i < ARRAY_SIZE(cg->cmux); i++) {
886 if (cg->info.cmux_to_group[i] < 0)
887 break;
888 if (cg->info.cmux_to_group[i] >=
889 ARRAY_SIZE(cg->info.cmux_groups)) {
890 WARN_ON_ONCE(1);
891 continue;
892 }
893
894 cg->cmux[i] = create_one_cmux(cg, i);
895 }
896
897 for (i = 0; i < ARRAY_SIZE(cg->hwaccel); i++) {
898 if (!cg->info.hwaccel[i])
899 continue;
900
901 cg->hwaccel[i] = create_one_hwaccel(cg, i);
902 }
903}
904
905static void __init clockgen_init(struct device_node *np);
906
Scott Wood80b4ae72017-03-20 10:37:23 +0800907/*
908 * Legacy nodes may get probed before the parent clockgen node.
909 * It is assumed that device trees with legacy nodes will not
910 * contain a "clocks" property -- otherwise the input clocks may
911 * not be initialized at this point.
912 */
Scott Wood0dfc86b2015-09-19 23:29:54 -0500913static void __init legacy_init_clockgen(struct device_node *np)
914{
915 if (!clockgen.node)
916 clockgen_init(of_get_parent(np));
917}
918
919/* Legacy node */
Tang Yuantian555eae92013-04-09 16:46:26 +0800920static void __init core_mux_init(struct device_node *np)
921{
922 struct clk *clk;
Scott Wood0dfc86b2015-09-19 23:29:54 -0500923 struct resource res;
924 int idx, rc;
Tang Yuantian555eae92013-04-09 16:46:26 +0800925
Scott Wood0dfc86b2015-09-19 23:29:54 -0500926 legacy_init_clockgen(np);
Tang Yuantian555eae92013-04-09 16:46:26 +0800927
Scott Wood0dfc86b2015-09-19 23:29:54 -0500928 if (of_address_to_resource(np, 0, &res))
Tang Yuantian555eae92013-04-09 16:46:26 +0800929 return;
Tang Yuantian555eae92013-04-09 16:46:26 +0800930
Scott Wood0dfc86b2015-09-19 23:29:54 -0500931 idx = (res.start & 0xf0) >> 5;
932 clk = clockgen.cmux[idx];
Tang Yuantian555eae92013-04-09 16:46:26 +0800933
934 rc = of_clk_add_provider(np, of_clk_src_simple_get, clk);
935 if (rc) {
Scott Wood0dfc86b2015-09-19 23:29:54 -0500936 pr_err("%s: Couldn't register clk provider for node %s: %d\n",
937 __func__, np->name, rc);
938 return;
Tang Yuantian555eae92013-04-09 16:46:26 +0800939 }
Tang Yuantian555eae92013-04-09 16:46:26 +0800940}
941
Julia Lawall3432a2e2016-04-18 16:55:34 +0200942static struct clk __init
943*sysclk_from_fixed(struct device_node *node, const char *name)
Tang Yuantian555eae92013-04-09 16:46:26 +0800944{
Scott Wood0dfc86b2015-09-19 23:29:54 -0500945 u32 rate;
Tang Yuantian555eae92013-04-09 16:46:26 +0800946
Scott Wood0dfc86b2015-09-19 23:29:54 -0500947 if (of_property_read_u32(node, "clock-frequency", &rate))
948 return ERR_PTR(-ENODEV);
949
Stephen Boydec3f2fc2016-03-01 11:00:19 -0800950 return clk_register_fixed_rate(NULL, name, NULL, 0, rate);
Scott Wood0dfc86b2015-09-19 23:29:54 -0500951}
952
Scott Wood80b4ae72017-03-20 10:37:23 +0800953static struct clk __init *input_clock(const char *name, struct clk *clk)
Scott Wood0dfc86b2015-09-19 23:29:54 -0500954{
Scott Wood80b4ae72017-03-20 10:37:23 +0800955 const char *input_name;
Scott Wood0dfc86b2015-09-19 23:29:54 -0500956
957 /* Register the input clock under the desired name. */
Scott Wood80b4ae72017-03-20 10:37:23 +0800958 input_name = __clk_get_name(clk);
959 clk = clk_register_fixed_factor(NULL, name, input_name,
Scott Wood0dfc86b2015-09-19 23:29:54 -0500960 0, 1, 1);
961 if (IS_ERR(clk))
962 pr_err("%s: Couldn't register %s: %ld\n", __func__, name,
963 PTR_ERR(clk));
964
965 return clk;
966}
967
Scott Wood80b4ae72017-03-20 10:37:23 +0800968static struct clk __init *input_clock_by_name(const char *name,
969 const char *dtname)
970{
971 struct clk *clk;
972
973 clk = of_clk_get_by_name(clockgen.node, dtname);
974 if (IS_ERR(clk))
975 return clk;
976
977 return input_clock(name, clk);
978}
979
980static struct clk __init *input_clock_by_index(const char *name, int idx)
981{
982 struct clk *clk;
983
984 clk = of_clk_get(clockgen.node, 0);
985 if (IS_ERR(clk))
986 return clk;
987
988 return input_clock(name, clk);
989}
990
Scott Wood0dfc86b2015-09-19 23:29:54 -0500991static struct clk * __init create_sysclk(const char *name)
992{
993 struct device_node *sysclk;
994 struct clk *clk;
995
996 clk = sysclk_from_fixed(clockgen.node, name);
997 if (!IS_ERR(clk))
998 return clk;
999
Scott Wood80b4ae72017-03-20 10:37:23 +08001000 clk = input_clock_by_name(name, "sysclk");
1001 if (!IS_ERR(clk))
1002 return clk;
1003
1004 clk = input_clock_by_index(name, 0);
Scott Wood0dfc86b2015-09-19 23:29:54 -05001005 if (!IS_ERR(clk))
1006 return clk;
1007
1008 sysclk = of_get_child_by_name(clockgen.node, "sysclk");
1009 if (sysclk) {
1010 clk = sysclk_from_fixed(sysclk, name);
1011 if (!IS_ERR(clk))
1012 return clk;
1013 }
1014
Scott Wood80b4ae72017-03-20 10:37:23 +08001015 pr_err("%s: No input sysclk\n", __func__);
1016 return NULL;
1017}
1018
1019static struct clk * __init create_coreclk(const char *name)
1020{
1021 struct clk *clk;
1022
1023 clk = input_clock_by_name(name, "coreclk");
1024 if (!IS_ERR(clk))
1025 return clk;
1026
1027 /*
1028 * This indicates a mix of legacy nodes with the new coreclk
1029 * mechanism, which should never happen. If this error occurs,
1030 * don't use the wrong input clock just because coreclk isn't
1031 * ready yet.
1032 */
1033 if (WARN_ON(PTR_ERR(clk) == -EPROBE_DEFER))
1034 return clk;
1035
Scott Wood0dfc86b2015-09-19 23:29:54 -05001036 return NULL;
1037}
1038
1039/* Legacy node */
1040static void __init sysclk_init(struct device_node *node)
1041{
1042 struct clk *clk;
1043
1044 legacy_init_clockgen(node);
1045
1046 clk = clockgen.sysclk;
1047 if (clk)
1048 of_clk_add_provider(node, of_clk_src_simple_get, clk);
1049}
1050
1051#define PLL_KILL BIT(31)
1052
1053static void __init create_one_pll(struct clockgen *cg, int idx)
1054{
1055 u32 __iomem *reg;
1056 u32 mult;
1057 struct clockgen_pll *pll = &cg->pll[idx];
Scott Wood80b4ae72017-03-20 10:37:23 +08001058 const char *input = "cg-sysclk";
Scott Wood0dfc86b2015-09-19 23:29:54 -05001059 int i;
1060
1061 if (!(cg->info.pll_mask & (1 << idx)))
1062 return;
1063
Scott Wood80b4ae72017-03-20 10:37:23 +08001064 if (cg->coreclk && idx != PLATFORM_PLL) {
1065 if (IS_ERR(cg->coreclk))
1066 return;
1067
1068 input = "cg-coreclk";
1069 }
1070
Scott Wood9e19ca22015-09-19 23:29:55 -05001071 if (cg->info.flags & CG_VER3) {
1072 switch (idx) {
1073 case PLATFORM_PLL:
1074 reg = cg->regs + 0x60080;
1075 break;
1076 case CGA_PLL1:
1077 reg = cg->regs + 0x80;
1078 break;
1079 case CGA_PLL2:
1080 reg = cg->regs + 0xa0;
1081 break;
1082 case CGB_PLL1:
1083 reg = cg->regs + 0x10080;
1084 break;
1085 case CGB_PLL2:
1086 reg = cg->regs + 0x100a0;
1087 break;
1088 default:
1089 WARN_ONCE(1, "index %d\n", idx);
1090 return;
1091 }
1092 } else {
1093 if (idx == PLATFORM_PLL)
1094 reg = cg->regs + 0xc00;
1095 else
1096 reg = cg->regs + 0x800 + 0x20 * (idx - 1);
1097 }
Scott Wood0dfc86b2015-09-19 23:29:54 -05001098
1099 /* Get the multiple of PLL */
Scott Wood9e19ca22015-09-19 23:29:55 -05001100 mult = cg_in(cg, reg);
Scott Wood0dfc86b2015-09-19 23:29:54 -05001101
1102 /* Check if this PLL is disabled */
1103 if (mult & PLL_KILL) {
1104 pr_debug("%s(): pll %p disabled\n", __func__, reg);
Tang Yuantian555eae92013-04-09 16:46:26 +08001105 return;
1106 }
1107
Scott Wood9e19ca22015-09-19 23:29:55 -05001108 if ((cg->info.flags & CG_VER3) ||
1109 ((cg->info.flags & CG_PLL_8BIT) && idx != PLATFORM_PLL))
Scott Wood0dfc86b2015-09-19 23:29:54 -05001110 mult = (mult & GENMASK(8, 1)) >> 1;
1111 else
1112 mult = (mult & GENMASK(6, 1)) >> 1;
Tang Yuantian555eae92013-04-09 16:46:26 +08001113
Scott Wood0dfc86b2015-09-19 23:29:54 -05001114 for (i = 0; i < ARRAY_SIZE(pll->div); i++) {
1115 struct clk *clk;
1116
1117 snprintf(pll->div[i].name, sizeof(pll->div[i].name),
1118 "cg-pll%d-div%d", idx, i + 1);
1119
1120 clk = clk_register_fixed_factor(NULL,
Scott Wood80b4ae72017-03-20 10:37:23 +08001121 pll->div[i].name, input, 0, mult, i + 1);
Scott Wood0dfc86b2015-09-19 23:29:54 -05001122 if (IS_ERR(clk)) {
1123 pr_err("%s: %s: register failed %ld\n",
1124 __func__, pll->div[i].name, PTR_ERR(clk));
1125 continue;
1126 }
1127
1128 pll->div[i].clk = clk;
Tang Yuantian555eae92013-04-09 16:46:26 +08001129 }
Scott Wood0dfc86b2015-09-19 23:29:54 -05001130}
Tang Yuantian555eae92013-04-09 16:46:26 +08001131
Scott Wood0dfc86b2015-09-19 23:29:54 -05001132static void __init create_plls(struct clockgen *cg)
1133{
1134 int i;
Tang Yuantian555eae92013-04-09 16:46:26 +08001135
Scott Wood0dfc86b2015-09-19 23:29:54 -05001136 for (i = 0; i < ARRAY_SIZE(cg->pll); i++)
1137 create_one_pll(cg, i);
1138}
1139
1140static void __init legacy_pll_init(struct device_node *np, int idx)
1141{
1142 struct clockgen_pll *pll;
1143 struct clk_onecell_data *onecell_data;
1144 struct clk **subclks;
1145 int count, rc;
1146
1147 legacy_init_clockgen(np);
1148
1149 pll = &clockgen.pll[idx];
Tang Yuantian555eae92013-04-09 16:46:26 +08001150 count = of_property_count_strings(np, "clock-output-names");
Tang Yuantian555eae92013-04-09 16:46:26 +08001151
Scott Wood0dfc86b2015-09-19 23:29:54 -05001152 BUILD_BUG_ON(ARRAY_SIZE(pll->div) < 4);
1153 subclks = kcalloc(4, sizeof(struct clk *), GFP_KERNEL);
Emil Medve8002cab2015-01-21 04:03:26 -06001154 if (!subclks)
Scott Wood0dfc86b2015-09-19 23:29:54 -05001155 return;
Tang Yuantian555eae92013-04-09 16:46:26 +08001156
Emil Medve6ef1cca2015-01-21 04:03:28 -06001157 onecell_data = kmalloc(sizeof(*onecell_data), GFP_KERNEL);
Emil Medve8002cab2015-01-21 04:03:26 -06001158 if (!onecell_data)
Tang Yuantian555eae92013-04-09 16:46:26 +08001159 goto err_clks;
Tang Yuantian555eae92013-04-09 16:46:26 +08001160
Scott Wood0dfc86b2015-09-19 23:29:54 -05001161 if (count <= 3) {
1162 subclks[0] = pll->div[0].clk;
1163 subclks[1] = pll->div[1].clk;
1164 subclks[2] = pll->div[3].clk;
1165 } else {
1166 subclks[0] = pll->div[0].clk;
1167 subclks[1] = pll->div[1].clk;
1168 subclks[2] = pll->div[2].clk;
1169 subclks[3] = pll->div[3].clk;
Tang Yuantian555eae92013-04-09 16:46:26 +08001170 }
1171
1172 onecell_data->clks = subclks;
1173 onecell_data->clk_num = count;
1174
1175 rc = of_clk_add_provider(np, of_clk_src_onecell_get, onecell_data);
1176 if (rc) {
Scott Wood0dfc86b2015-09-19 23:29:54 -05001177 pr_err("%s: Couldn't register clk provider for node %s: %d\n",
1178 __func__, np->name, rc);
Tang Yuantian555eae92013-04-09 16:46:26 +08001179 goto err_cell;
1180 }
1181
1182 return;
1183err_cell:
1184 kfree(onecell_data);
1185err_clks:
1186 kfree(subclks);
Tang Yuantian00fa6e52014-01-21 09:32:45 +08001187}
1188
Scott Wood0dfc86b2015-09-19 23:29:54 -05001189/* Legacy node */
Emil Medvea513b722015-01-21 04:03:31 -06001190static void __init pltfrm_pll_init(struct device_node *np)
1191{
Scott Wood0dfc86b2015-09-19 23:29:54 -05001192 legacy_pll_init(np, PLATFORM_PLL);
1193}
Emil Medvea513b722015-01-21 04:03:31 -06001194
Scott Wood0dfc86b2015-09-19 23:29:54 -05001195/* Legacy node */
1196static void __init core_pll_init(struct device_node *np)
1197{
1198 struct resource res;
1199 int idx;
1200
1201 if (of_address_to_resource(np, 0, &res))
1202 return;
1203
1204 if ((res.start & 0xfff) == 0xc00) {
1205 /*
1206 * ls1021a devtree labels the platform PLL
1207 * with the core PLL compatible
1208 */
1209 pltfrm_pll_init(np);
1210 } else {
1211 idx = (res.start & 0xf0) >> 5;
1212 legacy_pll_init(np, CGA_PLL1 + idx);
1213 }
1214}
1215
1216static struct clk *clockgen_clk_get(struct of_phandle_args *clkspec, void *data)
1217{
1218 struct clockgen *cg = data;
1219 struct clk *clk;
1220 struct clockgen_pll *pll;
1221 u32 type, idx;
1222
1223 if (clkspec->args_count < 2) {
1224 pr_err("%s: insufficient phandle args\n", __func__);
1225 return ERR_PTR(-EINVAL);
1226 }
1227
1228 type = clkspec->args[0];
1229 idx = clkspec->args[1];
1230
1231 switch (type) {
1232 case 0:
1233 if (idx != 0)
1234 goto bad_args;
1235 clk = cg->sysclk;
1236 break;
1237 case 1:
1238 if (idx >= ARRAY_SIZE(cg->cmux))
1239 goto bad_args;
1240 clk = cg->cmux[idx];
1241 break;
1242 case 2:
1243 if (idx >= ARRAY_SIZE(cg->hwaccel))
1244 goto bad_args;
1245 clk = cg->hwaccel[idx];
1246 break;
1247 case 3:
1248 if (idx >= ARRAY_SIZE(cg->fman))
1249 goto bad_args;
1250 clk = cg->fman[idx];
1251 break;
1252 case 4:
1253 pll = &cg->pll[PLATFORM_PLL];
1254 if (idx >= ARRAY_SIZE(pll->div))
1255 goto bad_args;
1256 clk = pll->div[idx].clk;
1257 break;
Scott Wood80b4ae72017-03-20 10:37:23 +08001258 case 5:
1259 if (idx != 0)
1260 goto bad_args;
1261 clk = cg->coreclk;
1262 if (IS_ERR(clk))
1263 clk = NULL;
1264 break;
Scott Wood0dfc86b2015-09-19 23:29:54 -05001265 default:
1266 goto bad_args;
1267 }
1268
1269 if (!clk)
1270 return ERR_PTR(-ENOENT);
1271 return clk;
1272
1273bad_args:
1274 pr_err("%s: Bad phandle args %u %u\n", __func__, type, idx);
1275 return ERR_PTR(-EINVAL);
1276}
1277
1278#ifdef CONFIG_PPC
1279#include <asm/mpc85xx.h>
1280
1281static const u32 a4510_svrs[] __initconst = {
1282 (SVR_P2040 << 8) | 0x10, /* P2040 1.0 */
1283 (SVR_P2040 << 8) | 0x11, /* P2040 1.1 */
1284 (SVR_P2041 << 8) | 0x10, /* P2041 1.0 */
1285 (SVR_P2041 << 8) | 0x11, /* P2041 1.1 */
1286 (SVR_P3041 << 8) | 0x10, /* P3041 1.0 */
1287 (SVR_P3041 << 8) | 0x11, /* P3041 1.1 */
1288 (SVR_P4040 << 8) | 0x20, /* P4040 2.0 */
1289 (SVR_P4080 << 8) | 0x20, /* P4080 2.0 */
1290 (SVR_P5010 << 8) | 0x10, /* P5010 1.0 */
1291 (SVR_P5010 << 8) | 0x20, /* P5010 2.0 */
1292 (SVR_P5020 << 8) | 0x10, /* P5020 1.0 */
1293 (SVR_P5021 << 8) | 0x10, /* P5021 1.0 */
1294 (SVR_P5040 << 8) | 0x10, /* P5040 1.0 */
1295};
1296
1297#define SVR_SECURITY 0x80000 /* The Security (E) bit */
1298
1299static bool __init has_erratum_a4510(void)
1300{
1301 u32 svr = mfspr(SPRN_SVR);
1302 int i;
1303
1304 svr &= ~SVR_SECURITY;
1305
1306 for (i = 0; i < ARRAY_SIZE(a4510_svrs); i++) {
1307 if (svr == a4510_svrs[i])
1308 return true;
1309 }
1310
1311 return false;
1312}
1313#else
1314static bool __init has_erratum_a4510(void)
1315{
1316 return false;
1317}
1318#endif
1319
1320static void __init clockgen_init(struct device_node *np)
1321{
1322 int i, ret;
1323 bool is_old_ls1021a = false;
1324
1325 /* May have already been called by a legacy probe */
1326 if (clockgen.node)
1327 return;
1328
1329 clockgen.node = np;
1330 clockgen.regs = of_iomap(np, 0);
1331 if (!clockgen.regs &&
1332 of_device_is_compatible(of_root, "fsl,ls1021a")) {
1333 /* Compatibility hack for old, broken device trees */
1334 clockgen.regs = ioremap(0x1ee1000, 0x1000);
1335 is_old_ls1021a = true;
1336 }
1337 if (!clockgen.regs) {
Emil Medvea513b722015-01-21 04:03:31 -06001338 pr_err("%s(): %s: of_iomap() failed\n", __func__, np->name);
1339 return;
1340 }
1341
Scott Wood0dfc86b2015-09-19 23:29:54 -05001342 for (i = 0; i < ARRAY_SIZE(chipinfo); i++) {
1343 if (of_device_is_compatible(np, chipinfo[i].compat))
1344 break;
1345 if (is_old_ls1021a &&
1346 !strcmp(chipinfo[i].compat, "fsl,ls1021a-clockgen"))
1347 break;
Emil Medvea513b722015-01-21 04:03:31 -06001348 }
1349
Scott Wood0dfc86b2015-09-19 23:29:54 -05001350 if (i == ARRAY_SIZE(chipinfo)) {
1351 pr_err("%s: unknown clockgen node %s\n", __func__,
1352 np->full_name);
1353 goto err;
Emil Medvea513b722015-01-21 04:03:31 -06001354 }
Scott Wood0dfc86b2015-09-19 23:29:54 -05001355 clockgen.info = chipinfo[i];
Emil Medvea513b722015-01-21 04:03:31 -06001356
Scott Wood0dfc86b2015-09-19 23:29:54 -05001357 if (clockgen.info.guts_compat) {
1358 struct device_node *guts;
Emil Medvea513b722015-01-21 04:03:31 -06001359
Scott Wood0dfc86b2015-09-19 23:29:54 -05001360 guts = of_find_compatible_node(NULL, NULL,
1361 clockgen.info.guts_compat);
1362 if (guts) {
1363 clockgen.guts = of_iomap(guts, 0);
1364 if (!clockgen.guts) {
1365 pr_err("%s: Couldn't map %s regs\n", __func__,
1366 guts->full_name);
1367 }
Emil Medvea513b722015-01-21 04:03:31 -06001368 }
1369
Emil Medvea513b722015-01-21 04:03:31 -06001370 }
1371
Scott Wood0dfc86b2015-09-19 23:29:54 -05001372 if (has_erratum_a4510())
1373 clockgen.info.flags |= CG_CMUX_GE_PLAT;
1374
1375 clockgen.sysclk = create_sysclk("cg-sysclk");
Scott Wood80b4ae72017-03-20 10:37:23 +08001376 clockgen.coreclk = create_coreclk("cg-coreclk");
Scott Wood0dfc86b2015-09-19 23:29:54 -05001377 create_plls(&clockgen);
1378 create_muxes(&clockgen);
1379
1380 if (clockgen.info.init_periph)
1381 clockgen.info.init_periph(&clockgen);
1382
1383 ret = of_clk_add_provider(np, clockgen_clk_get, &clockgen);
1384 if (ret) {
1385 pr_err("%s: Couldn't register clk provider for node %s: %d\n",
1386 __func__, np->name, ret);
Emil Medvea513b722015-01-21 04:03:31 -06001387 }
1388
1389 return;
Scott Wood0dfc86b2015-09-19 23:29:54 -05001390err:
1391 iounmap(clockgen.regs);
1392 clockgen.regs = NULL;
Emil Medvea513b722015-01-21 04:03:31 -06001393}
1394
Scott Wood0dfc86b2015-09-19 23:29:54 -05001395CLK_OF_DECLARE(qoriq_clockgen_1, "fsl,qoriq-clockgen-1.0", clockgen_init);
1396CLK_OF_DECLARE(qoriq_clockgen_2, "fsl,qoriq-clockgen-2.0", clockgen_init);
Tang Yuantian44709352016-11-24 10:36:55 +08001397CLK_OF_DECLARE(qoriq_clockgen_ls1012a, "fsl,ls1012a-clockgen", clockgen_init);
Scott Wood0dfc86b2015-09-19 23:29:54 -05001398CLK_OF_DECLARE(qoriq_clockgen_ls1021a, "fsl,ls1021a-clockgen", clockgen_init);
Hou Zhiqiange994412c2015-10-23 16:01:21 +08001399CLK_OF_DECLARE(qoriq_clockgen_ls1043a, "fsl,ls1043a-clockgen", clockgen_init);
Mingkai Hu80e52192016-09-07 11:48:30 +08001400CLK_OF_DECLARE(qoriq_clockgen_ls1046a, "fsl,ls1046a-clockgen", clockgen_init);
Scott Wood9e19ca22015-09-19 23:29:55 -05001401CLK_OF_DECLARE(qoriq_clockgen_ls2080a, "fsl,ls2080a-clockgen", clockgen_init);
Scott Wood0dfc86b2015-09-19 23:29:54 -05001402
1403/* Legacy nodes */
Kevin Hao66619ac2014-12-03 16:53:53 +08001404CLK_OF_DECLARE(qoriq_sysclk_1, "fsl,qoriq-sysclk-1.0", sysclk_init);
1405CLK_OF_DECLARE(qoriq_sysclk_2, "fsl,qoriq-sysclk-2.0", sysclk_init);
1406CLK_OF_DECLARE(qoriq_core_pll_1, "fsl,qoriq-core-pll-1.0", core_pll_init);
1407CLK_OF_DECLARE(qoriq_core_pll_2, "fsl,qoriq-core-pll-2.0", core_pll_init);
1408CLK_OF_DECLARE(qoriq_core_mux_1, "fsl,qoriq-core-mux-1.0", core_mux_init);
1409CLK_OF_DECLARE(qoriq_core_mux_2, "fsl,qoriq-core-mux-2.0", core_mux_init);
Emil Medvea513b722015-01-21 04:03:31 -06001410CLK_OF_DECLARE(qoriq_pltfrm_pll_1, "fsl,qoriq-platform-pll-1.0", pltfrm_pll_init);
1411CLK_OF_DECLARE(qoriq_pltfrm_pll_2, "fsl,qoriq-platform-pll-2.0", pltfrm_pll_init);