.. | .. |
---|
1 | 1 | // SPDX-License-Identifier: GPL-2.0 |
---|
2 | 2 | /* |
---|
| 3 | + * Copyright (C) 2020-2022 MaxLinear, Inc. |
---|
3 | 4 | * Copyright (C) 2020 Intel Corporation. |
---|
4 | | - * Zhu YiXin <yixin.zhu@intel.com> |
---|
5 | | - * Rahul Tanwar <rahul.tanwar@intel.com> |
---|
| 5 | + * Zhu Yixin <yzhu@maxlinear.com> |
---|
| 6 | + * Rahul Tanwar <rtanwar@maxlinear.com> |
---|
6 | 7 | */ |
---|
7 | 8 | #include <linux/clk-provider.h> |
---|
8 | 9 | #include <linux/device.h> |
---|
.. | .. |
---|
24 | 25 | static struct clk_hw *lgm_clk_register_fixed(struct lgm_clk_provider *ctx, |
---|
25 | 26 | const struct lgm_clk_branch *list) |
---|
26 | 27 | { |
---|
27 | | - unsigned long flags; |
---|
28 | 28 | |
---|
29 | | - if (list->div_flags & CLOCK_FLAG_VAL_INIT) { |
---|
30 | | - spin_lock_irqsave(&ctx->lock, flags); |
---|
| 29 | + if (list->div_flags & CLOCK_FLAG_VAL_INIT) |
---|
31 | 30 | lgm_set_clk_val(ctx->membase, list->div_off, list->div_shift, |
---|
32 | 31 | list->div_width, list->div_val); |
---|
33 | | - spin_unlock_irqrestore(&ctx->lock, flags); |
---|
34 | | - } |
---|
35 | 32 | |
---|
36 | 33 | return clk_hw_register_fixed_rate(NULL, list->name, |
---|
37 | 34 | list->parent_data[0].name, |
---|
.. | .. |
---|
41 | 38 | static u8 lgm_clk_mux_get_parent(struct clk_hw *hw) |
---|
42 | 39 | { |
---|
43 | 40 | struct lgm_clk_mux *mux = to_lgm_clk_mux(hw); |
---|
44 | | - unsigned long flags; |
---|
45 | 41 | u32 val; |
---|
46 | 42 | |
---|
47 | | - spin_lock_irqsave(&mux->lock, flags); |
---|
48 | 43 | if (mux->flags & MUX_CLK_SW) |
---|
49 | 44 | val = mux->reg; |
---|
50 | 45 | else |
---|
51 | 46 | val = lgm_get_clk_val(mux->membase, mux->reg, mux->shift, |
---|
52 | 47 | mux->width); |
---|
53 | | - spin_unlock_irqrestore(&mux->lock, flags); |
---|
54 | 48 | return clk_mux_val_to_index(hw, NULL, mux->flags, val); |
---|
55 | 49 | } |
---|
56 | 50 | |
---|
57 | 51 | static int lgm_clk_mux_set_parent(struct clk_hw *hw, u8 index) |
---|
58 | 52 | { |
---|
59 | 53 | struct lgm_clk_mux *mux = to_lgm_clk_mux(hw); |
---|
60 | | - unsigned long flags; |
---|
61 | 54 | u32 val; |
---|
62 | 55 | |
---|
63 | 56 | val = clk_mux_index_to_val(NULL, mux->flags, index); |
---|
64 | | - spin_lock_irqsave(&mux->lock, flags); |
---|
65 | 57 | if (mux->flags & MUX_CLK_SW) |
---|
66 | 58 | mux->reg = val; |
---|
67 | 59 | else |
---|
68 | 60 | lgm_set_clk_val(mux->membase, mux->reg, mux->shift, |
---|
69 | 61 | mux->width, val); |
---|
70 | | - spin_unlock_irqrestore(&mux->lock, flags); |
---|
71 | 62 | |
---|
72 | 63 | return 0; |
---|
73 | 64 | } |
---|
.. | .. |
---|
90 | 81 | lgm_clk_register_mux(struct lgm_clk_provider *ctx, |
---|
91 | 82 | const struct lgm_clk_branch *list) |
---|
92 | 83 | { |
---|
93 | | - unsigned long flags, cflags = list->mux_flags; |
---|
| 84 | + unsigned long cflags = list->mux_flags; |
---|
94 | 85 | struct device *dev = ctx->dev; |
---|
95 | 86 | u8 shift = list->mux_shift; |
---|
96 | 87 | u8 width = list->mux_width; |
---|
.. | .. |
---|
111 | 102 | init.num_parents = list->num_parents; |
---|
112 | 103 | |
---|
113 | 104 | mux->membase = ctx->membase; |
---|
114 | | - mux->lock = ctx->lock; |
---|
115 | 105 | mux->reg = reg; |
---|
116 | 106 | mux->shift = shift; |
---|
117 | 107 | mux->width = width; |
---|
.. | .. |
---|
123 | 113 | if (ret) |
---|
124 | 114 | return ERR_PTR(ret); |
---|
125 | 115 | |
---|
126 | | - if (cflags & CLOCK_FLAG_VAL_INIT) { |
---|
127 | | - spin_lock_irqsave(&mux->lock, flags); |
---|
| 116 | + if (cflags & CLOCK_FLAG_VAL_INIT) |
---|
128 | 117 | lgm_set_clk_val(mux->membase, reg, shift, width, list->mux_val); |
---|
129 | | - spin_unlock_irqrestore(&mux->lock, flags); |
---|
130 | | - } |
---|
131 | 118 | |
---|
132 | 119 | return hw; |
---|
133 | 120 | } |
---|
.. | .. |
---|
136 | 123 | lgm_clk_divider_recalc_rate(struct clk_hw *hw, unsigned long parent_rate) |
---|
137 | 124 | { |
---|
138 | 125 | struct lgm_clk_divider *divider = to_lgm_clk_divider(hw); |
---|
139 | | - unsigned long flags; |
---|
140 | 126 | unsigned int val; |
---|
141 | 127 | |
---|
142 | | - spin_lock_irqsave(÷r->lock, flags); |
---|
143 | 128 | val = lgm_get_clk_val(divider->membase, divider->reg, |
---|
144 | 129 | divider->shift, divider->width); |
---|
145 | | - spin_unlock_irqrestore(÷r->lock, flags); |
---|
146 | 130 | |
---|
147 | 131 | return divider_recalc_rate(hw, parent_rate, val, divider->table, |
---|
148 | 132 | divider->flags, divider->width); |
---|
.. | .. |
---|
163 | 147 | unsigned long prate) |
---|
164 | 148 | { |
---|
165 | 149 | struct lgm_clk_divider *divider = to_lgm_clk_divider(hw); |
---|
166 | | - unsigned long flags; |
---|
167 | 150 | int value; |
---|
168 | 151 | |
---|
169 | 152 | value = divider_get_val(rate, prate, divider->table, |
---|
.. | .. |
---|
171 | 154 | if (value < 0) |
---|
172 | 155 | return value; |
---|
173 | 156 | |
---|
174 | | - spin_lock_irqsave(÷r->lock, flags); |
---|
175 | 157 | lgm_set_clk_val(divider->membase, divider->reg, |
---|
176 | 158 | divider->shift, divider->width, value); |
---|
177 | | - spin_unlock_irqrestore(÷r->lock, flags); |
---|
178 | 159 | |
---|
179 | 160 | return 0; |
---|
180 | 161 | } |
---|
.. | .. |
---|
182 | 163 | static int lgm_clk_divider_enable_disable(struct clk_hw *hw, int enable) |
---|
183 | 164 | { |
---|
184 | 165 | struct lgm_clk_divider *div = to_lgm_clk_divider(hw); |
---|
185 | | - unsigned long flags; |
---|
186 | 166 | |
---|
187 | | - spin_lock_irqsave(&div->lock, flags); |
---|
188 | | - lgm_set_clk_val(div->membase, div->reg, div->shift_gate, |
---|
189 | | - div->width_gate, enable); |
---|
190 | | - spin_unlock_irqrestore(&div->lock, flags); |
---|
| 167 | + if (div->flags != DIV_CLK_NO_MASK) |
---|
| 168 | + lgm_set_clk_val(div->membase, div->reg, div->shift_gate, |
---|
| 169 | + div->width_gate, enable); |
---|
191 | 170 | return 0; |
---|
192 | 171 | } |
---|
193 | 172 | |
---|
.. | .. |
---|
213 | 192 | lgm_clk_register_divider(struct lgm_clk_provider *ctx, |
---|
214 | 193 | const struct lgm_clk_branch *list) |
---|
215 | 194 | { |
---|
216 | | - unsigned long flags, cflags = list->div_flags; |
---|
| 195 | + unsigned long cflags = list->div_flags; |
---|
217 | 196 | struct device *dev = ctx->dev; |
---|
218 | 197 | struct lgm_clk_divider *div; |
---|
219 | 198 | struct clk_init_data init = {}; |
---|
.. | .. |
---|
236 | 215 | init.num_parents = 1; |
---|
237 | 216 | |
---|
238 | 217 | div->membase = ctx->membase; |
---|
239 | | - div->lock = ctx->lock; |
---|
240 | 218 | div->reg = reg; |
---|
241 | 219 | div->shift = shift; |
---|
242 | 220 | div->width = width; |
---|
.. | .. |
---|
251 | 229 | if (ret) |
---|
252 | 230 | return ERR_PTR(ret); |
---|
253 | 231 | |
---|
254 | | - if (cflags & CLOCK_FLAG_VAL_INIT) { |
---|
255 | | - spin_lock_irqsave(&div->lock, flags); |
---|
| 232 | + if (cflags & CLOCK_FLAG_VAL_INIT) |
---|
256 | 233 | lgm_set_clk_val(div->membase, reg, shift, width, list->div_val); |
---|
257 | | - spin_unlock_irqrestore(&div->lock, flags); |
---|
258 | | - } |
---|
259 | 234 | |
---|
260 | 235 | return hw; |
---|
261 | 236 | } |
---|
.. | .. |
---|
264 | 239 | lgm_clk_register_fixed_factor(struct lgm_clk_provider *ctx, |
---|
265 | 240 | const struct lgm_clk_branch *list) |
---|
266 | 241 | { |
---|
267 | | - unsigned long flags; |
---|
268 | 242 | struct clk_hw *hw; |
---|
269 | 243 | |
---|
270 | 244 | hw = clk_hw_register_fixed_factor(ctx->dev, list->name, |
---|
.. | .. |
---|
273 | 247 | if (IS_ERR(hw)) |
---|
274 | 248 | return ERR_CAST(hw); |
---|
275 | 249 | |
---|
276 | | - if (list->div_flags & CLOCK_FLAG_VAL_INIT) { |
---|
277 | | - spin_lock_irqsave(&ctx->lock, flags); |
---|
| 250 | + if (list->div_flags & CLOCK_FLAG_VAL_INIT) |
---|
278 | 251 | lgm_set_clk_val(ctx->membase, list->div_off, list->div_shift, |
---|
279 | 252 | list->div_width, list->div_val); |
---|
280 | | - spin_unlock_irqrestore(&ctx->lock, flags); |
---|
281 | | - } |
---|
282 | 253 | |
---|
283 | 254 | return hw; |
---|
284 | 255 | } |
---|
.. | .. |
---|
286 | 257 | static int lgm_clk_gate_enable(struct clk_hw *hw) |
---|
287 | 258 | { |
---|
288 | 259 | struct lgm_clk_gate *gate = to_lgm_clk_gate(hw); |
---|
289 | | - unsigned long flags; |
---|
290 | 260 | unsigned int reg; |
---|
291 | 261 | |
---|
292 | | - spin_lock_irqsave(&gate->lock, flags); |
---|
293 | 262 | reg = GATE_HW_REG_EN(gate->reg); |
---|
294 | 263 | lgm_set_clk_val(gate->membase, reg, gate->shift, 1, 1); |
---|
295 | | - spin_unlock_irqrestore(&gate->lock, flags); |
---|
296 | 264 | |
---|
297 | 265 | return 0; |
---|
298 | 266 | } |
---|
.. | .. |
---|
300 | 268 | static void lgm_clk_gate_disable(struct clk_hw *hw) |
---|
301 | 269 | { |
---|
302 | 270 | struct lgm_clk_gate *gate = to_lgm_clk_gate(hw); |
---|
303 | | - unsigned long flags; |
---|
304 | 271 | unsigned int reg; |
---|
305 | 272 | |
---|
306 | | - spin_lock_irqsave(&gate->lock, flags); |
---|
307 | 273 | reg = GATE_HW_REG_DIS(gate->reg); |
---|
308 | 274 | lgm_set_clk_val(gate->membase, reg, gate->shift, 1, 1); |
---|
309 | | - spin_unlock_irqrestore(&gate->lock, flags); |
---|
310 | 275 | } |
---|
311 | 276 | |
---|
312 | 277 | static int lgm_clk_gate_is_enabled(struct clk_hw *hw) |
---|
313 | 278 | { |
---|
314 | 279 | struct lgm_clk_gate *gate = to_lgm_clk_gate(hw); |
---|
315 | 280 | unsigned int reg, ret; |
---|
316 | | - unsigned long flags; |
---|
317 | 281 | |
---|
318 | | - spin_lock_irqsave(&gate->lock, flags); |
---|
319 | 282 | reg = GATE_HW_REG_STAT(gate->reg); |
---|
320 | 283 | ret = lgm_get_clk_val(gate->membase, reg, gate->shift, 1); |
---|
321 | | - spin_unlock_irqrestore(&gate->lock, flags); |
---|
322 | 284 | |
---|
323 | 285 | return ret; |
---|
324 | 286 | } |
---|
.. | .. |
---|
333 | 295 | lgm_clk_register_gate(struct lgm_clk_provider *ctx, |
---|
334 | 296 | const struct lgm_clk_branch *list) |
---|
335 | 297 | { |
---|
336 | | - unsigned long flags, cflags = list->gate_flags; |
---|
| 298 | + unsigned long cflags = list->gate_flags; |
---|
337 | 299 | const char *pname = list->parent_data[0].name; |
---|
338 | 300 | struct device *dev = ctx->dev; |
---|
339 | 301 | u8 shift = list->gate_shift; |
---|
.. | .. |
---|
354 | 316 | init.num_parents = pname ? 1 : 0; |
---|
355 | 317 | |
---|
356 | 318 | gate->membase = ctx->membase; |
---|
357 | | - gate->lock = ctx->lock; |
---|
358 | 319 | gate->reg = reg; |
---|
359 | 320 | gate->shift = shift; |
---|
360 | 321 | gate->flags = cflags; |
---|
.. | .. |
---|
366 | 327 | return ERR_PTR(ret); |
---|
367 | 328 | |
---|
368 | 329 | if (cflags & CLOCK_FLAG_VAL_INIT) { |
---|
369 | | - spin_lock_irqsave(&gate->lock, flags); |
---|
370 | 330 | lgm_set_clk_val(gate->membase, reg, shift, 1, list->gate_val); |
---|
371 | | - spin_unlock_irqrestore(&gate->lock, flags); |
---|
372 | 331 | } |
---|
373 | 332 | |
---|
374 | 333 | return hw; |
---|
.. | .. |
---|
396 | 355 | hw = lgm_clk_register_fixed_factor(ctx, list); |
---|
397 | 356 | break; |
---|
398 | 357 | case CLK_TYPE_GATE: |
---|
399 | | - hw = lgm_clk_register_gate(ctx, list); |
---|
| 358 | + if (list->gate_flags & GATE_CLK_HW) { |
---|
| 359 | + hw = lgm_clk_register_gate(ctx, list); |
---|
| 360 | + } else { |
---|
| 361 | + /* |
---|
| 362 | + * GATE_CLKs can be controlled either from |
---|
| 363 | + * CGU clk driver i.e. this driver or directly |
---|
| 364 | + * from power management driver/daemon. It is |
---|
| 365 | + * dependent on the power policy/profile requirements |
---|
| 366 | + * of the end product. To override control of gate |
---|
| 367 | + * clks from this driver, provide NULL for this index |
---|
| 368 | + * of gate clk provider. |
---|
| 369 | + */ |
---|
| 370 | + hw = NULL; |
---|
| 371 | + } |
---|
400 | 372 | break; |
---|
| 373 | + |
---|
401 | 374 | default: |
---|
402 | 375 | dev_err(ctx->dev, "invalid clk type\n"); |
---|
403 | 376 | return -EINVAL; |
---|
.. | .. |
---|
443 | 416 | static int lgm_clk_ddiv_enable(struct clk_hw *hw) |
---|
444 | 417 | { |
---|
445 | 418 | struct lgm_clk_ddiv *ddiv = to_lgm_clk_ddiv(hw); |
---|
446 | | - unsigned long flags; |
---|
447 | 419 | |
---|
448 | | - spin_lock_irqsave(&ddiv->lock, flags); |
---|
449 | 420 | lgm_set_clk_val(ddiv->membase, ddiv->reg, ddiv->shift_gate, |
---|
450 | 421 | ddiv->width_gate, 1); |
---|
451 | | - spin_unlock_irqrestore(&ddiv->lock, flags); |
---|
452 | 422 | return 0; |
---|
453 | 423 | } |
---|
454 | 424 | |
---|
455 | 425 | static void lgm_clk_ddiv_disable(struct clk_hw *hw) |
---|
456 | 426 | { |
---|
457 | 427 | struct lgm_clk_ddiv *ddiv = to_lgm_clk_ddiv(hw); |
---|
458 | | - unsigned long flags; |
---|
459 | 428 | |
---|
460 | | - spin_lock_irqsave(&ddiv->lock, flags); |
---|
461 | 429 | lgm_set_clk_val(ddiv->membase, ddiv->reg, ddiv->shift_gate, |
---|
462 | 430 | ddiv->width_gate, 0); |
---|
463 | | - spin_unlock_irqrestore(&ddiv->lock, flags); |
---|
464 | 431 | } |
---|
465 | 432 | |
---|
466 | 433 | static int |
---|
.. | .. |
---|
497 | 464 | { |
---|
498 | 465 | struct lgm_clk_ddiv *ddiv = to_lgm_clk_ddiv(hw); |
---|
499 | 466 | u32 div, ddiv1, ddiv2; |
---|
500 | | - unsigned long flags; |
---|
501 | 467 | |
---|
502 | 468 | div = DIV_ROUND_CLOSEST_ULL((u64)prate, rate); |
---|
503 | 469 | |
---|
504 | | - spin_lock_irqsave(&ddiv->lock, flags); |
---|
505 | 470 | if (lgm_get_clk_val(ddiv->membase, ddiv->reg, ddiv->shift2, 1)) { |
---|
506 | 471 | div = DIV_ROUND_CLOSEST_ULL((u64)div, 5); |
---|
507 | 472 | div = div * 2; |
---|
508 | 473 | } |
---|
509 | 474 | |
---|
510 | | - if (div <= 0) { |
---|
511 | | - spin_unlock_irqrestore(&ddiv->lock, flags); |
---|
| 475 | + if (div <= 0) |
---|
512 | 476 | return -EINVAL; |
---|
513 | | - } |
---|
514 | 477 | |
---|
515 | | - if (lgm_clk_get_ddiv_val(div, &ddiv1, &ddiv2)) { |
---|
516 | | - spin_unlock_irqrestore(&ddiv->lock, flags); |
---|
| 478 | + if (lgm_clk_get_ddiv_val(div, &ddiv1, &ddiv2)) |
---|
517 | 479 | return -EINVAL; |
---|
518 | | - } |
---|
519 | 480 | |
---|
520 | 481 | lgm_set_clk_val(ddiv->membase, ddiv->reg, ddiv->shift0, ddiv->width0, |
---|
521 | 482 | ddiv1 - 1); |
---|
522 | 483 | |
---|
523 | 484 | lgm_set_clk_val(ddiv->membase, ddiv->reg, ddiv->shift1, ddiv->width1, |
---|
524 | 485 | ddiv2 - 1); |
---|
525 | | - spin_unlock_irqrestore(&ddiv->lock, flags); |
---|
526 | 486 | |
---|
527 | 487 | return 0; |
---|
528 | 488 | } |
---|
.. | .. |
---|
533 | 493 | { |
---|
534 | 494 | struct lgm_clk_ddiv *ddiv = to_lgm_clk_ddiv(hw); |
---|
535 | 495 | u32 div, ddiv1, ddiv2; |
---|
536 | | - unsigned long flags; |
---|
537 | 496 | u64 rate64; |
---|
538 | 497 | |
---|
539 | 498 | div = DIV_ROUND_CLOSEST_ULL((u64)*prate, rate); |
---|
540 | 499 | |
---|
541 | 500 | /* if predivide bit is enabled, modify div by factor of 2.5 */ |
---|
542 | | - spin_lock_irqsave(&ddiv->lock, flags); |
---|
543 | 501 | if (lgm_get_clk_val(ddiv->membase, ddiv->reg, ddiv->shift2, 1)) { |
---|
544 | 502 | div = div * 2; |
---|
545 | 503 | div = DIV_ROUND_CLOSEST_ULL((u64)div, 5); |
---|
546 | 504 | } |
---|
547 | | - spin_unlock_irqrestore(&ddiv->lock, flags); |
---|
548 | 505 | |
---|
549 | 506 | if (div <= 0) |
---|
550 | 507 | return *prate; |
---|
.. | .. |
---|
558 | 515 | do_div(rate64, ddiv2); |
---|
559 | 516 | |
---|
560 | 517 | /* if predivide bit is enabled, modify rounded rate by factor of 2.5 */ |
---|
561 | | - spin_lock_irqsave(&ddiv->lock, flags); |
---|
562 | 518 | if (lgm_get_clk_val(ddiv->membase, ddiv->reg, ddiv->shift2, 1)) { |
---|
563 | 519 | rate64 = rate64 * 2; |
---|
564 | 520 | rate64 = DIV_ROUND_CLOSEST_ULL(rate64, 5); |
---|
565 | 521 | } |
---|
566 | | - spin_unlock_irqrestore(&ddiv->lock, flags); |
---|
567 | 522 | |
---|
568 | 523 | return rate64; |
---|
569 | 524 | } |
---|
.. | .. |
---|
600 | 555 | init.num_parents = 1; |
---|
601 | 556 | |
---|
602 | 557 | ddiv->membase = ctx->membase; |
---|
603 | | - ddiv->lock = ctx->lock; |
---|
604 | 558 | ddiv->reg = list->reg; |
---|
605 | 559 | ddiv->shift0 = list->shift0; |
---|
606 | 560 | ddiv->width0 = list->width0; |
---|