]> git.proxmox.com Git - mirror_ubuntu-eoan-kernel.git/blob - drivers/cpufreq/exynos5440-cpufreq.c
pinctrl: stm32: add 'depends on HAS_IOMEM' to fix unmet dependency
[mirror_ubuntu-eoan-kernel.git] / drivers / cpufreq / exynos5440-cpufreq.c
1 /*
2 * Copyright (c) 2013 Samsung Electronics Co., Ltd.
3 * http://www.samsung.com
4 *
5 * Amit Daniel Kachhap <amit.daniel@samsung.com>
6 *
7 * EXYNOS5440 - CPU frequency scaling support
8 *
9 * This program is free software; you can redistribute it and/or modify
10 * it under the terms of the GNU General Public License version 2 as
11 * published by the Free Software Foundation.
12 */
13
14 #include <linux/clk.h>
15 #include <linux/cpu.h>
16 #include <linux/cpufreq.h>
17 #include <linux/err.h>
18 #include <linux/interrupt.h>
19 #include <linux/io.h>
20 #include <linux/module.h>
21 #include <linux/of_address.h>
22 #include <linux/of_irq.h>
23 #include <linux/pm_opp.h>
24 #include <linux/platform_device.h>
25 #include <linux/slab.h>
26
27 /* Register definitions */
28 #define XMU_DVFS_CTRL 0x0060
29 #define XMU_PMU_P0_7 0x0064
30 #define XMU_C0_3_PSTATE 0x0090
31 #define XMU_P_LIMIT 0x00a0
32 #define XMU_P_STATUS 0x00a4
33 #define XMU_PMUEVTEN 0x00d0
34 #define XMU_PMUIRQEN 0x00d4
35 #define XMU_PMUIRQ 0x00d8
36
37 /* PMU mask and shift definations */
38 #define P_VALUE_MASK 0x7
39
40 #define XMU_DVFS_CTRL_EN_SHIFT 0
41
42 #define P0_7_CPUCLKDEV_SHIFT 21
43 #define P0_7_CPUCLKDEV_MASK 0x7
44 #define P0_7_ATBCLKDEV_SHIFT 18
45 #define P0_7_ATBCLKDEV_MASK 0x7
46 #define P0_7_CSCLKDEV_SHIFT 15
47 #define P0_7_CSCLKDEV_MASK 0x7
48 #define P0_7_CPUEMA_SHIFT 28
49 #define P0_7_CPUEMA_MASK 0xf
50 #define P0_7_L2EMA_SHIFT 24
51 #define P0_7_L2EMA_MASK 0xf
52 #define P0_7_VDD_SHIFT 8
53 #define P0_7_VDD_MASK 0x7f
54 #define P0_7_FREQ_SHIFT 0
55 #define P0_7_FREQ_MASK 0xff
56
57 #define C0_3_PSTATE_VALID_SHIFT 8
58 #define C0_3_PSTATE_CURR_SHIFT 4
59 #define C0_3_PSTATE_NEW_SHIFT 0
60
61 #define PSTATE_CHANGED_EVTEN_SHIFT 0
62
63 #define PSTATE_CHANGED_IRQEN_SHIFT 0
64
65 #define PSTATE_CHANGED_SHIFT 0
66
67 /* some constant values for clock divider calculation */
68 #define CPU_DIV_FREQ_MAX 500
69 #define CPU_DBG_FREQ_MAX 375
70 #define CPU_ATB_FREQ_MAX 500
71
72 #define PMIC_LOW_VOLT 0x30
73 #define PMIC_HIGH_VOLT 0x28
74
75 #define CPUEMA_HIGH 0x2
76 #define CPUEMA_MID 0x4
77 #define CPUEMA_LOW 0x7
78
79 #define L2EMA_HIGH 0x1
80 #define L2EMA_MID 0x3
81 #define L2EMA_LOW 0x4
82
83 #define DIV_TAB_MAX 2
84 /* frequency unit is 20MHZ */
85 #define FREQ_UNIT 20
86 #define MAX_VOLTAGE 1550000 /* In microvolt */
87 #define VOLTAGE_STEP 12500 /* In microvolt */
88
89 #define CPUFREQ_NAME "exynos5440_dvfs"
90 #define DEF_TRANS_LATENCY 100000
91
92 enum cpufreq_level_index {
93 L0, L1, L2, L3, L4,
94 L5, L6, L7, L8, L9,
95 };
96 #define CPUFREQ_LEVEL_END (L7 + 1)
97
98 struct exynos_dvfs_data {
99 void __iomem *base;
100 struct resource *mem;
101 int irq;
102 struct clk *cpu_clk;
103 unsigned int latency;
104 struct cpufreq_frequency_table *freq_table;
105 unsigned int freq_count;
106 struct device *dev;
107 bool dvfs_enabled;
108 struct work_struct irq_work;
109 };
110
111 static struct exynos_dvfs_data *dvfs_info;
112 static DEFINE_MUTEX(cpufreq_lock);
113 static struct cpufreq_freqs freqs;
114
115 static int init_div_table(void)
116 {
117 struct cpufreq_frequency_table *pos, *freq_tbl = dvfs_info->freq_table;
118 unsigned int tmp, clk_div, ema_div, freq, volt_id, idx;
119 struct dev_pm_opp *opp;
120
121 cpufreq_for_each_entry_idx(pos, freq_tbl, idx) {
122 opp = dev_pm_opp_find_freq_exact(dvfs_info->dev,
123 pos->frequency * 1000, true);
124 if (IS_ERR(opp)) {
125 dev_err(dvfs_info->dev,
126 "failed to find valid OPP for %u KHZ\n",
127 pos->frequency);
128 return PTR_ERR(opp);
129 }
130
131 freq = pos->frequency / 1000; /* In MHZ */
132 clk_div = ((freq / CPU_DIV_FREQ_MAX) & P0_7_CPUCLKDEV_MASK)
133 << P0_7_CPUCLKDEV_SHIFT;
134 clk_div |= ((freq / CPU_ATB_FREQ_MAX) & P0_7_ATBCLKDEV_MASK)
135 << P0_7_ATBCLKDEV_SHIFT;
136 clk_div |= ((freq / CPU_DBG_FREQ_MAX) & P0_7_CSCLKDEV_MASK)
137 << P0_7_CSCLKDEV_SHIFT;
138
139 /* Calculate EMA */
140 volt_id = dev_pm_opp_get_voltage(opp);
141
142 volt_id = (MAX_VOLTAGE - volt_id) / VOLTAGE_STEP;
143 if (volt_id < PMIC_HIGH_VOLT) {
144 ema_div = (CPUEMA_HIGH << P0_7_CPUEMA_SHIFT) |
145 (L2EMA_HIGH << P0_7_L2EMA_SHIFT);
146 } else if (volt_id > PMIC_LOW_VOLT) {
147 ema_div = (CPUEMA_LOW << P0_7_CPUEMA_SHIFT) |
148 (L2EMA_LOW << P0_7_L2EMA_SHIFT);
149 } else {
150 ema_div = (CPUEMA_MID << P0_7_CPUEMA_SHIFT) |
151 (L2EMA_MID << P0_7_L2EMA_SHIFT);
152 }
153
154 tmp = (clk_div | ema_div | (volt_id << P0_7_VDD_SHIFT)
155 | ((freq / FREQ_UNIT) << P0_7_FREQ_SHIFT));
156
157 __raw_writel(tmp, dvfs_info->base + XMU_PMU_P0_7 + 4 * idx);
158 dev_pm_opp_put(opp);
159 }
160
161 return 0;
162 }
163
164 static void exynos_enable_dvfs(unsigned int cur_frequency)
165 {
166 unsigned int tmp, cpu;
167 struct cpufreq_frequency_table *freq_table = dvfs_info->freq_table;
168 struct cpufreq_frequency_table *pos;
169 /* Disable DVFS */
170 __raw_writel(0, dvfs_info->base + XMU_DVFS_CTRL);
171
172 /* Enable PSTATE Change Event */
173 tmp = __raw_readl(dvfs_info->base + XMU_PMUEVTEN);
174 tmp |= (1 << PSTATE_CHANGED_EVTEN_SHIFT);
175 __raw_writel(tmp, dvfs_info->base + XMU_PMUEVTEN);
176
177 /* Enable PSTATE Change IRQ */
178 tmp = __raw_readl(dvfs_info->base + XMU_PMUIRQEN);
179 tmp |= (1 << PSTATE_CHANGED_IRQEN_SHIFT);
180 __raw_writel(tmp, dvfs_info->base + XMU_PMUIRQEN);
181
182 /* Set initial performance index */
183 cpufreq_for_each_entry(pos, freq_table)
184 if (pos->frequency == cur_frequency)
185 break;
186
187 if (pos->frequency == CPUFREQ_TABLE_END) {
188 dev_crit(dvfs_info->dev, "Boot up frequency not supported\n");
189 /* Assign the highest frequency */
190 pos = freq_table;
191 cur_frequency = pos->frequency;
192 }
193
194 dev_info(dvfs_info->dev, "Setting dvfs initial frequency = %uKHZ",
195 cur_frequency);
196
197 for (cpu = 0; cpu < CONFIG_NR_CPUS; cpu++) {
198 tmp = __raw_readl(dvfs_info->base + XMU_C0_3_PSTATE + cpu * 4);
199 tmp &= ~(P_VALUE_MASK << C0_3_PSTATE_NEW_SHIFT);
200 tmp |= ((pos - freq_table) << C0_3_PSTATE_NEW_SHIFT);
201 __raw_writel(tmp, dvfs_info->base + XMU_C0_3_PSTATE + cpu * 4);
202 }
203
204 /* Enable DVFS */
205 __raw_writel(1 << XMU_DVFS_CTRL_EN_SHIFT,
206 dvfs_info->base + XMU_DVFS_CTRL);
207 }
208
209 static int exynos_target(struct cpufreq_policy *policy, unsigned int index)
210 {
211 unsigned int tmp;
212 int i;
213 struct cpufreq_frequency_table *freq_table = dvfs_info->freq_table;
214
215 mutex_lock(&cpufreq_lock);
216
217 freqs.old = policy->cur;
218 freqs.new = freq_table[index].frequency;
219
220 cpufreq_freq_transition_begin(policy, &freqs);
221
222 /* Set the target frequency in all C0_3_PSTATE register */
223 for_each_cpu(i, policy->cpus) {
224 tmp = __raw_readl(dvfs_info->base + XMU_C0_3_PSTATE + i * 4);
225 tmp &= ~(P_VALUE_MASK << C0_3_PSTATE_NEW_SHIFT);
226 tmp |= (index << C0_3_PSTATE_NEW_SHIFT);
227
228 __raw_writel(tmp, dvfs_info->base + XMU_C0_3_PSTATE + i * 4);
229 }
230 mutex_unlock(&cpufreq_lock);
231 return 0;
232 }
233
234 static void exynos_cpufreq_work(struct work_struct *work)
235 {
236 unsigned int cur_pstate, index;
237 struct cpufreq_policy *policy = cpufreq_cpu_get(0); /* boot CPU */
238 struct cpufreq_frequency_table *freq_table = dvfs_info->freq_table;
239
240 /* Ensure we can access cpufreq structures */
241 if (unlikely(dvfs_info->dvfs_enabled == false))
242 goto skip_work;
243
244 mutex_lock(&cpufreq_lock);
245 freqs.old = policy->cur;
246
247 cur_pstate = __raw_readl(dvfs_info->base + XMU_P_STATUS);
248 if (cur_pstate >> C0_3_PSTATE_VALID_SHIFT & 0x1)
249 index = (cur_pstate >> C0_3_PSTATE_CURR_SHIFT) & P_VALUE_MASK;
250 else
251 index = (cur_pstate >> C0_3_PSTATE_NEW_SHIFT) & P_VALUE_MASK;
252
253 if (likely(index < dvfs_info->freq_count)) {
254 freqs.new = freq_table[index].frequency;
255 } else {
256 dev_crit(dvfs_info->dev, "New frequency out of range\n");
257 freqs.new = freqs.old;
258 }
259 cpufreq_freq_transition_end(policy, &freqs, 0);
260
261 cpufreq_cpu_put(policy);
262 mutex_unlock(&cpufreq_lock);
263 skip_work:
264 enable_irq(dvfs_info->irq);
265 }
266
267 static irqreturn_t exynos_cpufreq_irq(int irq, void *id)
268 {
269 unsigned int tmp;
270
271 tmp = __raw_readl(dvfs_info->base + XMU_PMUIRQ);
272 if (tmp >> PSTATE_CHANGED_SHIFT & 0x1) {
273 __raw_writel(tmp, dvfs_info->base + XMU_PMUIRQ);
274 disable_irq_nosync(irq);
275 schedule_work(&dvfs_info->irq_work);
276 }
277 return IRQ_HANDLED;
278 }
279
280 static void exynos_sort_descend_freq_table(void)
281 {
282 struct cpufreq_frequency_table *freq_tbl = dvfs_info->freq_table;
283 int i = 0, index;
284 unsigned int tmp_freq;
285 /*
286 * Exynos5440 clock controller state logic expects the cpufreq table to
287 * be in descending order. But the OPP library constructs the table in
288 * ascending order. So to make the table descending we just need to
289 * swap the i element with the N - i element.
290 */
291 for (i = 0; i < dvfs_info->freq_count / 2; i++) {
292 index = dvfs_info->freq_count - i - 1;
293 tmp_freq = freq_tbl[i].frequency;
294 freq_tbl[i].frequency = freq_tbl[index].frequency;
295 freq_tbl[index].frequency = tmp_freq;
296 }
297 }
298
299 static int exynos_cpufreq_cpu_init(struct cpufreq_policy *policy)
300 {
301 policy->clk = dvfs_info->cpu_clk;
302 return cpufreq_generic_init(policy, dvfs_info->freq_table,
303 dvfs_info->latency);
304 }
305
306 static struct cpufreq_driver exynos_driver = {
307 .flags = CPUFREQ_STICKY | CPUFREQ_ASYNC_NOTIFICATION |
308 CPUFREQ_NEED_INITIAL_FREQ_CHECK,
309 .verify = cpufreq_generic_frequency_table_verify,
310 .target_index = exynos_target,
311 .get = cpufreq_generic_get,
312 .init = exynos_cpufreq_cpu_init,
313 .name = CPUFREQ_NAME,
314 .attr = cpufreq_generic_attr,
315 };
316
317 static const struct of_device_id exynos_cpufreq_match[] = {
318 {
319 .compatible = "samsung,exynos5440-cpufreq",
320 },
321 {},
322 };
323 MODULE_DEVICE_TABLE(of, exynos_cpufreq_match);
324
325 static int exynos_cpufreq_probe(struct platform_device *pdev)
326 {
327 int ret = -EINVAL;
328 struct device_node *np;
329 struct resource res;
330 unsigned int cur_frequency;
331
332 np = pdev->dev.of_node;
333 if (!np)
334 return -ENODEV;
335
336 dvfs_info = devm_kzalloc(&pdev->dev, sizeof(*dvfs_info), GFP_KERNEL);
337 if (!dvfs_info) {
338 ret = -ENOMEM;
339 goto err_put_node;
340 }
341
342 dvfs_info->dev = &pdev->dev;
343
344 ret = of_address_to_resource(np, 0, &res);
345 if (ret)
346 goto err_put_node;
347
348 dvfs_info->base = devm_ioremap_resource(dvfs_info->dev, &res);
349 if (IS_ERR(dvfs_info->base)) {
350 ret = PTR_ERR(dvfs_info->base);
351 goto err_put_node;
352 }
353
354 dvfs_info->irq = irq_of_parse_and_map(np, 0);
355 if (!dvfs_info->irq) {
356 dev_err(dvfs_info->dev, "No cpufreq irq found\n");
357 ret = -ENODEV;
358 goto err_put_node;
359 }
360
361 ret = dev_pm_opp_of_add_table(dvfs_info->dev);
362 if (ret) {
363 dev_err(dvfs_info->dev, "failed to init OPP table: %d\n", ret);
364 goto err_put_node;
365 }
366
367 ret = dev_pm_opp_init_cpufreq_table(dvfs_info->dev,
368 &dvfs_info->freq_table);
369 if (ret) {
370 dev_err(dvfs_info->dev,
371 "failed to init cpufreq table: %d\n", ret);
372 goto err_free_opp;
373 }
374 dvfs_info->freq_count = dev_pm_opp_get_opp_count(dvfs_info->dev);
375 exynos_sort_descend_freq_table();
376
377 if (of_property_read_u32(np, "clock-latency", &dvfs_info->latency))
378 dvfs_info->latency = DEF_TRANS_LATENCY;
379
380 dvfs_info->cpu_clk = devm_clk_get(dvfs_info->dev, "armclk");
381 if (IS_ERR(dvfs_info->cpu_clk)) {
382 dev_err(dvfs_info->dev, "Failed to get cpu clock\n");
383 ret = PTR_ERR(dvfs_info->cpu_clk);
384 goto err_free_table;
385 }
386
387 cur_frequency = clk_get_rate(dvfs_info->cpu_clk);
388 if (!cur_frequency) {
389 dev_err(dvfs_info->dev, "Failed to get clock rate\n");
390 ret = -EINVAL;
391 goto err_free_table;
392 }
393 cur_frequency /= 1000;
394
395 INIT_WORK(&dvfs_info->irq_work, exynos_cpufreq_work);
396 ret = devm_request_irq(dvfs_info->dev, dvfs_info->irq,
397 exynos_cpufreq_irq, IRQF_TRIGGER_NONE,
398 CPUFREQ_NAME, dvfs_info);
399 if (ret) {
400 dev_err(dvfs_info->dev, "Failed to register IRQ\n");
401 goto err_free_table;
402 }
403
404 ret = init_div_table();
405 if (ret) {
406 dev_err(dvfs_info->dev, "Failed to initialise div table\n");
407 goto err_free_table;
408 }
409
410 exynos_enable_dvfs(cur_frequency);
411 ret = cpufreq_register_driver(&exynos_driver);
412 if (ret) {
413 dev_err(dvfs_info->dev,
414 "%s: failed to register cpufreq driver\n", __func__);
415 goto err_free_table;
416 }
417
418 of_node_put(np);
419 dvfs_info->dvfs_enabled = true;
420 return 0;
421
422 err_free_table:
423 dev_pm_opp_free_cpufreq_table(dvfs_info->dev, &dvfs_info->freq_table);
424 err_free_opp:
425 dev_pm_opp_of_remove_table(dvfs_info->dev);
426 err_put_node:
427 of_node_put(np);
428 dev_err(&pdev->dev, "%s: failed initialization\n", __func__);
429 return ret;
430 }
431
432 static int exynos_cpufreq_remove(struct platform_device *pdev)
433 {
434 cpufreq_unregister_driver(&exynos_driver);
435 dev_pm_opp_free_cpufreq_table(dvfs_info->dev, &dvfs_info->freq_table);
436 dev_pm_opp_of_remove_table(dvfs_info->dev);
437 return 0;
438 }
439
440 static struct platform_driver exynos_cpufreq_platdrv = {
441 .driver = {
442 .name = "exynos5440-cpufreq",
443 .of_match_table = exynos_cpufreq_match,
444 },
445 .probe = exynos_cpufreq_probe,
446 .remove = exynos_cpufreq_remove,
447 };
448 module_platform_driver(exynos_cpufreq_platdrv);
449
450 MODULE_AUTHOR("Amit Daniel Kachhap <amit.daniel@samsung.com>");
451 MODULE_DESCRIPTION("Exynos5440 cpufreq driver");
452 MODULE_LICENSE("GPL");