Quoting Xukai Wang (2025-02-17 06:45:17) > diff --git a/drivers/clk/Kconfig b/drivers/clk/Kconfig > index 299bc678ed1b9fcd9110bb8c5937a1bd1ea60e23..f63355ab8adeeee90d29d1a975607f5dc0a58bd6 100644 > --- a/drivers/clk/Kconfig > +++ b/drivers/clk/Kconfig > @@ -464,6 +464,12 @@ config COMMON_CLK_K210 > help > Support for the Canaan Kendryte K210 RISC-V SoC clocks. > > +config COMMON_CLK_K230 > + bool "Clock driver for the Canaan Kendryte K230 SoC" > + depends on OF && (ARCH_CANAAN || COMPILE_TEST) It doesn't depend on OF to build from what I can tell. This should simply be depends on ARCH_CANAAN || COMPILE_TEST right? > + help > + Support for the Canaan Kendryte K230 RISC-V SoC clocks. > + > config COMMON_CLK_SP7021 > tristate "Clock driver for Sunplus SP7021 SoC" > depends on SOC_SP7021 || COMPILE_TEST > diff --git a/drivers/clk/clk-k230.c b/drivers/clk/clk-k230.c > new file mode 100644 > index 0000000000000000000000000000000000000000..017ecad9ec47ed751f75fe5197aa119ae35e05f9 > --- /dev/null > +++ b/drivers/clk/clk-k230.c > @@ -0,0 +1,1347 @@ > +// SPDX-License-Identifier: GPL-2.0-only > +/* > + * Kendryte Canaan K230 Clock Drivers > + * > + * Author: Xukai Wang <kingxukai@xxxxxxxxxxxx> > + * Author: Troy Mitchell <troymitchell988@xxxxxxxxx> > + */ > + > +#include <dt-bindings/clock/canaan,k230-clk.h> Please put dt-bindigns at the end of the include list. > +#include <linux/clk.h> Is this include used? Ideally a clk provider (clk-provider.h) isn't also a clk consumer (clk.h). > +#include <linux/clkdev.h> Hopefully clkdev isn't used. > +#include <linux/clk-provider.h> > +#include <linux/iopoll.h> > +#include <linux/module.h> Is this include used? > +#include <linux/platform_device.h> > +#include <linux/spinlock.h> Need to include mod_devicetable.h > + > +/* PLL control register bits. */ > +#define K230_PLL_BYPASS_ENABLE BIT(19) > +#define K230_PLL_GATE_ENABLE BIT(2) > +#define K230_PLL_GATE_WRITE_ENABLE BIT(18) > +#define K230_PLL_OD_SHIFT 24 > +#define K230_PLL_OD_MASK 0xF > +#define K230_PLL_R_SHIFT 16 > +#define K230_PLL_R_MASK 0x3F > +#define K230_PLL_F_SHIFT 0 > +#define K230_PLL_F_MASK 0x1FFFF > +#define K230_PLL0_OFFSET_BASE 0x00 > +#define K230_PLL1_OFFSET_BASE 0x10 > +#define K230_PLL2_OFFSET_BASE 0x20 > +#define K230_PLL3_OFFSET_BASE 0x30 > +#define K230_PLL_DIV_REG_OFFSET 0x00 > +#define K230_PLL_BYPASS_REG_OFFSET 0x04 > +#define K230_PLL_GATE_REG_OFFSET 0x08 > +#define K230_PLL_LOCK_REG_OFFSET 0x0C > + > +/* PLL lock register bits. */ > +#define K230_PLL_STATUS_MASK BIT(0) > + > +/* K230 CLK registers offset */ > +#define K230_CLK_AUDIO_CLKDIV_OFFSET 0x34 > +#define K230_CLK_PDM_CLKDIV_OFFSET 0x40 > +#define K230_CLK_CODEC_ADC_MCLKDIV_OFFSET 0x38 > +#define K230_CLK_CODEC_DAC_MCLKDIV_OFFSET 0x3c > + > +/* K230 CLK OPS. */ > +#define K230_CLK_OPS_GATE \ > + .enable = k230_clk_enable, \ > + .disable = k230_clk_disable, \ > + .is_enabled = k230_clk_is_enabled > + > +#define K230_CLK_OPS_RATE \ > + .set_rate = k230_clk_set_rate, \ > + .round_rate = k230_clk_round_rate, \ > + .recalc_rate = k230_clk_get_rate > + > +#define K230_CLK_OPS_MUX \ > + .set_parent = k230_clk_set_parent, \ > + .get_parent = k230_clk_get_parent, \ > + .determine_rate = clk_hw_determine_rate_no_reparent > + > +#define K230_CLK_OPS_ID_NONE 0 > +#define K230_CLK_OPS_ID_GATE_ONLY 1 > +#define K230_CLK_OPS_ID_RATE_ONLY 2 > +#define K230_CLK_OPS_ID_RATE_GATE 3 > +#define K230_CLK_OPS_ID_MUX_ONLY 4 > +#define K230_CLK_OPS_ID_MUX_GATE 5 > +#define K230_CLK_OPS_ID_MUX_RATE 6 > +#define K230_CLK_OPS_ID_ALL 7 > +#define K230_CLK_OPS_ID_NUM 8 > + > +/* K230 CLK MACROS */ > +#define K230_CLK_MAX_PARENT_NUM 6 > + > +#define K230_GATE_FORMAT(_reg, _bit, _reverse, _have_gate) \ > + .gate_reg_off = (_reg), \ > + .gate_bit_enable = (_bit), \ > + .gate_bit_reverse = (_reverse), \ > + .have_gate = (_have_gate) > + > +#define K230_RATE_FORMAT(_mul_min, _mul_max, _mul_shift, _mul_mask, \ > + _div_min, _div_max, _div_shift, _div_mask, \ > + _reg, _bit, _method, _reg_c, _bit_c, \ > + _mul_min_c, _mul_max_c, _mul_shift_c, _mul_mask_c, \ > + _have_rate, _have_rate_c) \ > + .rate_mul_min = (_mul_min), \ > + .rate_mul_max = (_mul_max), \ > + .rate_mul_shift = (_mul_shift), \ > + .rate_mul_mask = (_mul_mask), \ > + .rate_mul_min_c = (_mul_min_c), \ > + .rate_mul_max_c = (_mul_max_c), \ > + .rate_mul_shift_c = (_mul_shift_c), \ > + .rate_mul_mask_c = (_mul_mask_c), \ > + .rate_div_min = (_div_min), \ > + .rate_div_max = (_div_max), \ > + .rate_div_shift = (_div_shift), \ > + .rate_div_mask = (_div_mask), \ > + .rate_reg_off = (_reg), \ > + .rate_reg_off_c = (_reg_c), \ > + .rate_write_enable_bit = (_bit), \ > + .rate_write_enable_bit_c = (_bit_c), \ > + .method = (_method), \ > + .have_rate = (_have_rate), \ > + .have_rate_c = (_have_rate_c) > + > +#define K230_MUX_FORMAT(_reg, _shift, _mask, _have_mux) \ > + .mux_reg_off = (_reg), \ > + .mux_reg_shift = (_shift), \ > + .mux_reg_mask = (_mask), \ > + .have_mux = (_have_mux) > + > +#define K230_GATE_FORMAT_ZERO K230_GATE_FORMAT(0, 0, 0, false) > +#define K230_RATE_FORMAT_ZERO K230_RATE_FORMAT(0, 0, 0, 0, 0, 0, \ > + 0, 0, 0, 0, 0, 0, \ > + 0, 0, 0, 0, 0, false, false) > +#define K230_MUX_FORMAT_ZERO K230_MUX_FORMAT(0, 0, 0, false) > + > +struct k230_sysclk; > + > +/* K230 PLLs. */ > +enum k230_pll_id { > + K230_PLL0, K230_PLL1, K230_PLL2, K230_PLL3, K230_PLL_NUM > +}; > + > +struct k230_pll { > + enum k230_pll_id id; > + struct k230_sysclk *ksc; > + void __iomem *div, *bypass, *gate, *lock; > + struct clk_hw hw; > +}; > + > +#define to_k230_pll(_hw) container_of(_hw, struct k230_pll, hw) > + > +struct k230_pll_cfg { > + u32 reg; > + enum k230_pll_id pll_id; > + const char *name; > +}; > + > +/* K230 PLL_DIVS. */ > +struct k230_pll_div { > + struct k230_sysclk *ksc; > + struct clk_hw *hw; > +}; > + > +struct k230_pll_div_cfg { > + const char *parent_name, *name; > + int div; > +}; > + > +enum k230_pll_div_id { > + K230_PLL0_DIV2, > + K230_PLL0_DIV3, > + K230_PLL0_DIV4, > + K230_PLL1_DIV2, > + K230_PLL1_DIV3, > + K230_PLL1_DIV4, > + K230_PLL2_DIV2, > + K230_PLL2_DIV3, > + K230_PLL2_DIV4, > + K230_PLL3_DIV2, > + K230_PLL3_DIV3, > + K230_PLL3_DIV4, > + K230_PLL_DIV_NUM > +}; > + > +enum k230_clk_div_type { > + K230_MUL, > + K230_DIV, > + K230_MUL_DIV, > +}; > + > +/* K230 CLKS. */ > +struct k230_clk { > + int id; > + struct k230_sysclk *ksc; > + struct clk_hw hw; > +}; > + > +#define to_k230_clk(_hw) container_of(_hw, struct k230_clk, hw) > + > +enum k230_clk_parent_type { > + K230_OSC24M, > + K230_PLL, > + K230_PLL_DIV, > + K230_CLK_COMPOSITE, > +}; > + > +struct k230_clk_parent { > + enum k230_clk_parent_type type; > + union { > + enum k230_pll_div_id pll_div_id; > + enum k230_pll_id pll_id; > + int clk_id; > + }; > +}; > + > +struct k230_clk_cfg { > + /* attr */ > + const char *name; > + /* 0-read & write; 1-read only */ > + bool read_only; > + int num_parent; > + struct k230_clk_parent parent[K230_CLK_MAX_PARENT_NUM]; > + bool status; > + int flags; > + > + /* rate reg */ > + u32 rate_reg_off; > + u32 rate_reg_off_c; > + void __iomem *rate_reg; > + void __iomem *rate_reg_c; > + /* rate info*/ > + u32 rate_write_enable_bit; > + u32 rate_write_enable_bit_c; > + enum k230_clk_div_type method; > + bool have_rate; > + bool have_rate_c; > + /* rate mul */ > + u32 rate_mul_min; > + u32 rate_mul_max; > + u32 rate_mul_shift; > + u32 rate_mul_mask; > + /* rate mul-changable */ > + u32 rate_mul_min_c; > + u32 rate_mul_max_c; > + u32 rate_mul_shift_c; > + u32 rate_mul_mask_c; > + /* rate div */ > + u32 rate_div_min; > + u32 rate_div_max; > + u32 rate_div_shift; > + u32 rate_div_mask; > + > + /* gate reg */ > + u32 gate_reg_off; > + void __iomem *gate_reg; > + /* gate info*/ > + bool have_gate; > + u32 gate_bit_enable; > + u32 gate_bit_reverse; > + > + /* mux reg */ > + u32 mux_reg_off; > + void __iomem *mux_reg; > + /* mux info */ > + bool have_mux; > + u32 mux_reg_shift; > + u32 mux_reg_mask; > +}; > + > +/* K230 SYSCLK. */ > +struct k230_sysclk { > + struct platform_device *pdev; > + void __iomem *pll_regs, *regs; > + spinlock_t pll_lock, clk_lock; > + struct k230_pll *plls; > + struct k230_clk *clks; > + struct k230_pll_div *dclks; > +}; > + > +static const struct k230_pll_cfg k230_pll_cfgs[] = { > + [K230_PLL0] = { > + .reg = K230_PLL0_OFFSET_BASE, > + .pll_id = K230_PLL0, > + .name = "pll0", > + }, > + [K230_PLL1] = { > + .reg = K230_PLL1_OFFSET_BASE, > + .pll_id = K230_PLL1, > + .name = "pll1", > + }, > + [K230_PLL2] = { > + .reg = K230_PLL2_OFFSET_BASE, > + .pll_id = K230_PLL2, > + .name = "pll2", > + }, > + [K230_PLL3] = { > + .reg = K230_PLL3_OFFSET_BASE, > + .pll_id = K230_PLL3, > + .name = "pll3", > + }, > +}; > + > +static const struct k230_pll_div_cfg k230_pll_div_cfgs[] = { > + [K230_PLL0_DIV2] = { "pll0", "pll0_div2", 2}, > + [K230_PLL0_DIV3] = { "pll0", "pll0_div3", 3}, > + [K230_PLL0_DIV4] = { "pll0", "pll0_div4", 4}, > + [K230_PLL1_DIV2] = { "pll1", "pll1_div2", 2}, > + [K230_PLL1_DIV3] = { "pll1", "pll1_div3", 3}, > + [K230_PLL1_DIV4] = { "pll1", "pll1_div4", 4}, > + [K230_PLL2_DIV2] = { "pll2", "pll2_div2", 2}, > + [K230_PLL2_DIV3] = { "pll2", "pll2_div3", 3}, > + [K230_PLL2_DIV4] = { "pll2", "pll2_div4", 4}, > + [K230_PLL3_DIV2] = { "pll3", "pll3_div2", 2}, > + [K230_PLL3_DIV3] = { "pll3", "pll3_div3", 3}, > + [K230_PLL3_DIV4] = { "pll3", "pll3_div4", 4}, > +}; > + > +static struct k230_clk_cfg k230_clk_cfgs[] = { > + [K230_CPU0_SRC] = { Can these be named variables instead of array elements? Then we can map quickly to parents instead of going through an intermediate enum. > + .name = "cpu0_src", > + .read_only = false, > + .flags = 0, > + .status = true, > + .num_parent = 1, > + .parent[0] = { > + .type = K230_PLL_DIV, > + .pll_div_id = K230_PLL0_DIV2, > + }, > + K230_RATE_FORMAT(1, 16, 0, 0, > + 16, 16, 1, 0xF, > + 0x0, 31, K230_MUL, 0, 0, > + 0, 0, 0, 0, > + true, false), > + K230_GATE_FORMAT(0, 0, 0, true), > + K230_MUX_FORMAT_ZERO, > + }, > + [K230_CPU0_ACLK] = { > + .name = "cpu0_aclk", > + .read_only = false, > + .flags = 0, > + .status = true, > + .num_parent = 1, > + .parent[0] = { > + .type = K230_CLK_COMPOSITE, > + .clk_id = K230_CPU0_SRC, > + }, > + K230_RATE_FORMAT(1, 1, 0, 0, > + 1, 8, 7, 0x7, > + 0x0, 31, K230_MUL, 0, 0, > + 0, 0, 0, 0, > + true, false), > + K230_GATE_FORMAT_ZERO, > + K230_MUX_FORMAT_ZERO, > + }, > + [K230_CPU0_PLIC] = { > + .name = "cpu0_plic", > + .read_only = false, > + .flags = 0, > + .status = true, > + .num_parent = 1, > + .parent[0] = { > + .type = K230_CLK_COMPOSITE, > + .clk_id = K230_CPU0_SRC, > + }, > + K230_RATE_FORMAT(1, 1, 0, 0, > + 1, 8, 10, 0x7, > + 0x0, 31, K230_DIV, 0, 0, > + 0, 0, 0, 0, > + true, false), > + K230_GATE_FORMAT(0, 9, 0, true), > + K230_MUX_FORMAT_ZERO, > + }, > + [K230_CPU0_NOC_DDRCP4] = { > + .name = "cpu0_noc_ddrcp4", > + .read_only = false, > + .flags = 0, > + .status = true, > + .num_parent = 1, > + .parent[0] = { > + .type = K230_CLK_COMPOSITE, > + .clk_id = K230_CPU0_SRC, > + }, > + K230_RATE_FORMAT_ZERO, > + K230_GATE_FORMAT(0x60, 7, 0, true), > + K230_MUX_FORMAT_ZERO, > + }, > + [K230_CPU0_PCLK] = { > + .name = "cpu0_pclk", > + .read_only = false, > + .flags = 0, > + .status = true, > + .num_parent = 1, > + .parent[0] = { > + .type = K230_PLL_DIV, > + .pll_div_id = K230_PLL0_DIV4, > + }, > + K230_RATE_FORMAT(1, 1, 0, 0, > + 1, 8, 15, 0x7, > + 0x0, 31, K230_DIV, 0, 0, > + 0, 0, 0, 0, > + true, false), > + K230_GATE_FORMAT(0, 13, 0, true), > + K230_MUX_FORMAT_ZERO, > + }, > + [K230_PMU_PCLK] = { > + .name = "pmu_pclk", > + .read_only = false, > + .flags = 0, > + .status = true, > + .num_parent = 1, > + .parent[0] = { > + .type = K230_OSC24M, > + }, > + K230_RATE_FORMAT_ZERO, > + K230_GATE_FORMAT(0x10, 0, 0, true), > + K230_MUX_FORMAT_ZERO, > + }, > + [K230_HS_OSPI_SRC] = { > + .name = "hs_ospi_src", > + .read_only = false, > + .flags = 0, > + .status = true, > + .num_parent = 2, > + .parent[0] = { > + .type = K230_PLL_DIV, > + .pll_div_id = K230_PLL0_DIV2, > + }, > + .parent[1] = { > + .type = K230_PLL_DIV, > + .pll_div_id = K230_PLL2_DIV4, > + }, > + K230_RATE_FORMAT_ZERO, > + K230_GATE_FORMAT(0x18, 24, 0, true), > + K230_MUX_FORMAT(0x20, 18, 0x1, true), > + }, > +}; > + > +#define K230_CLK_NUM ARRAY_SIZE(k230_clk_cfgs) > + > +static void k230_init_pll(void __iomem *regs, enum k230_pll_id pll_id, > + struct k230_pll *pll) > +{ > + void __iomem *base; > + > + pll->id = pll_id; > + base = regs + k230_pll_cfgs[pll_id].reg; > + pll->div = base + K230_PLL_DIV_REG_OFFSET; > + pll->bypass = base + K230_PLL_BYPASS_REG_OFFSET; > + pll->gate = base + K230_PLL_GATE_REG_OFFSET; > + pll->lock = base + K230_PLL_LOCK_REG_OFFSET; > +} > + > +static int k230_pll_prepare(struct clk_hw *hw) > +{ > + struct k230_pll *pll = to_k230_pll(hw); > + u32 reg; > + > + /* wait for PLL lock until it reachs lock status */ s/reachs/reaches/ > + return readl_poll_timeout(pll->lock, reg, > + (reg & K230_PLL_STATUS_MASK) == K230_PLL_STATUS_MASK, > + 400, 0); > +} > + > +static bool k230_pll_hw_is_enabled(struct k230_pll *pll) > +{ > + return (readl(pll->gate) & K230_PLL_GATE_ENABLE) == K230_PLL_GATE_ENABLE; > +} > + > +static void k230_pll_enable_hw(void __iomem *regs, struct k230_pll *pll) > +{ > + u32 reg; > + > + if (k230_pll_hw_is_enabled(pll)) > + return; > + > + /* Set PLL factors */ > + reg = readl(pll->gate); > + reg |= (K230_PLL_GATE_ENABLE | K230_PLL_GATE_WRITE_ENABLE); > + writel(reg, pll->gate); > +} > + > +static int k230_pll_enable(struct clk_hw *hw) > +{ > + struct k230_pll *pll = to_k230_pll(hw); > + struct k230_sysclk *ksc = pll->ksc; > + > + guard(spinlock)(&ksc->pll_lock); > + k230_pll_enable_hw(ksc->regs, pll); > + > + return 0; > +} > + > +static void k230_pll_disable(struct clk_hw *hw) > +{ > + struct k230_pll *pll = to_k230_pll(hw); > + struct k230_sysclk *ksc = pll->ksc; > + u32 reg; > + > + guard(spinlock)(&ksc->pll_lock); > + reg = readl(pll->gate); > + > + reg &= ~(K230_PLL_GATE_ENABLE); > + reg |= (K230_PLL_GATE_WRITE_ENABLE); > + > + writel(reg, pll->gate); > +} > + > +static int k230_pll_is_enabled(struct clk_hw *hw) > +{ > + return k230_pll_hw_is_enabled(to_k230_pll(hw)); > +} > + > +static int k230_pll_init(struct clk_hw *hw) > +{ > + if (k230_pll_is_enabled(hw)) > + return clk_prepare_enable(hw->clk); Is this CLK_IS_CRITICAL? > + > + return 0; > +} > + > +static unsigned long k230_pll_get_rate(struct clk_hw *hw, unsigned long parent_rate) > +{ > + struct k230_pll *pll = to_k230_pll(hw); > + struct k230_sysclk *ksc = pll->ksc; > + u32 reg; > + u32 r, f, od; > + > + reg = readl(pll->bypass); > + if (reg & K230_PLL_BYPASS_ENABLE) > + return parent_rate; > + > + reg = readl(pll->lock); > + if (!(reg & (K230_PLL_STATUS_MASK))) { /* unlocked */ > + dev_err(&ksc->pdev->dev, "%s is unlock.\n", clk_hw_get_name(hw)); > + return 0; > + } > + > + reg = readl(pll->div); > + r = ((reg >> K230_PLL_R_SHIFT) & K230_PLL_R_MASK) + 1; > + f = ((reg >> K230_PLL_F_SHIFT) & K230_PLL_F_MASK) + 1; > + od = ((reg >> K230_PLL_OD_SHIFT) & K230_PLL_OD_MASK) + 1; > + > + return div_u64((u64)parent_rate * f, r * od); Is this mul_u64_u32_div()? > +} > + > +static const struct clk_ops k230_pll_ops = { > + .init = k230_pll_init, > + .prepare = k230_pll_prepare, > + .enable = k230_pll_enable, > + .disable = k230_pll_disable, > + .is_enabled = k230_pll_is_enabled, > + .recalc_rate = k230_pll_get_rate, > +}; > + > +static int k230_register_pll(struct platform_device *pdev, > + struct k230_sysclk *ksc, > + enum k230_pll_id pll_id, > + const char *name, > + int num_parents, > + const struct clk_ops *ops) > +{ > + struct k230_pll *pll = &ksc->plls[pll_id]; > + struct clk_init_data init = {}; > + struct device *dev = &pdev->dev; > + int ret; > + const struct clk_parent_data parent_data[] = { > + { .index = 0, }, > + }; > + > + init.name = name; > + init.parent_data = parent_data; > + init.num_parents = num_parents; > + init.ops = ops; > + > + pll->hw.init = &init; > + pll->ksc = ksc; > + > + ret = devm_clk_hw_register(dev, &pll->hw); > + if (ret) > + return dev_err_probe(dev, ret, "failed to register pll"); Missing newline on printk. > + > + return 0; > +} > + > +static int k230_register_plls(struct platform_device *pdev, struct k230_sysclk *ksc) > +{ > + int i, ret; > + const struct k230_pll_cfg *cfg; > + > + for (i = 0; i < K230_PLL_NUM; i++) { > + cfg = &k230_pll_cfgs[i]; > + > + k230_init_pll(ksc->pll_regs, i, &ksc->plls[i]); > + > + ret = k230_register_pll(pdev, ksc, cfg->pll_id, cfg->name, 1, > + &k230_pll_ops); > + if (ret) > + return dev_err_probe(&pdev->dev, ret, > + "register %s failed\n", cfg->name); > + } > + > + return 0; > +} > + > +static int k230_register_pll_divs(struct platform_device *pdev, struct k230_sysclk *ksc) > +{ > + struct device *dev = &pdev->dev; > + struct clk_hw *hw; > + > + for (int i = 0; i < K230_PLL_DIV_NUM; i++) { > + hw = devm_clk_hw_register_fixed_factor(dev, k230_pll_div_cfgs[i].name, > + k230_pll_div_cfgs[i].parent_name, > + 0, 1, k230_pll_div_cfgs[i].div); > + if (IS_ERR(hw)) > + return PTR_ERR(hw); > + ksc->dclks[i].hw = hw; > + ksc->dclks[i].ksc = ksc; > + } > + > + return 0; > +} > + > +static int k230_clk_enable(struct clk_hw *hw) > +{ > + struct k230_clk *clk = to_k230_clk(hw); > + struct k230_sysclk *ksc = clk->ksc; > + struct k230_clk_cfg *cfg = &k230_clk_cfgs[clk->id]; > + u32 reg; > + > + if (!cfg->have_gate) > + return dev_err_probe(&ksc->pdev->dev, -EINVAL, "This clock doesn't have gate\n"); > + > + guard(spinlock)(&ksc->clk_lock); > + reg = readl(cfg->gate_reg); > + if (cfg->gate_bit_reverse) > + reg &= ~BIT(cfg->gate_bit_enable); > + else > + reg |= BIT(cfg->gate_bit_enable); > + writel(reg, cfg->gate_reg); > + > + return 0; > +} > + > +static void k230_clk_disable(struct clk_hw *hw) > +{ > + struct k230_clk *clk = to_k230_clk(hw); > + struct k230_sysclk *ksc = clk->ksc; > + struct k230_clk_cfg *cfg = &k230_clk_cfgs[clk->id]; > + u32 reg; > + > + if (!cfg->have_gate) { > + dev_err(&ksc->pdev->dev, "This clock doesn't have gate\n"); Why are the clk_ops assigned to this clk then? > + return; > + } > + > + guard(spinlock)(&ksc->clk_lock); > + reg = readl(cfg->gate_reg); > + > + if (cfg->gate_bit_reverse) > + reg |= BIT(cfg->gate_bit_enable); > + else > + reg &= ~BIT(cfg->gate_bit_enable); > + > + writel(reg, cfg->gate_reg); > +} > + > +static int k230_clk_is_enabled(struct clk_hw *hw) > +{ > + struct k230_clk *clk = to_k230_clk(hw); > + struct k230_sysclk *ksc = clk->ksc; > + struct k230_clk_cfg *cfg = &k230_clk_cfgs[clk->id]; > + u32 reg; > + > + if (!cfg->have_gate) > + return dev_err_probe(&ksc->pdev->dev, -EINVAL, "This clock doesn't have gate\n"); Why are the clk_ops assigned to this clk then? Don't use dev_err_probe() here. > + > + guard(spinlock)(&ksc->clk_lock); > + reg = readl(cfg->gate_reg); > + > + /* Check gate bit condition based on configuration and then set ret */ > + if (cfg->gate_bit_reverse) > + return (BIT(cfg->gate_bit_enable) & reg) ? 1 : 0; > + else > + return (BIT(cfg->gate_bit_enable) & ~reg) ? 1 : 0; > +} > + > +static int k230_clk_set_parent(struct clk_hw *hw, u8 index) > +{ > + struct k230_clk *clk = to_k230_clk(hw); > + struct k230_sysclk *ksc = clk->ksc; > + struct k230_clk_cfg *cfg = &k230_clk_cfgs[clk->id]; > + u8 reg; > + > + if (!cfg->have_mux) > + return dev_err_probe(&ksc->pdev->dev, -EINVAL, "This clock doesn't have mux\n"); Why are the clk_ops assigned to this clk then? Don't use dev_err_probe() here. > + > + guard(spinlock)(&ksc->clk_lock); > + reg = (cfg->mux_reg_mask & index) << cfg->mux_reg_shift; > + writeb(reg, cfg->mux_reg); > + > + return 0; > +} > + > +static u8 k230_clk_get_parent(struct clk_hw *hw) > +{ > + struct k230_clk *clk = to_k230_clk(hw); > + struct k230_sysclk *ksc = clk->ksc; > + struct k230_clk_cfg *cfg = &k230_clk_cfgs[clk->id]; > + u8 reg; > + > + if (!cfg->have_mux) > + return dev_err_probe(&ksc->pdev->dev, -EINVAL, "This clock doesn't have mux\n"); All these checks should go away and we should have different clk_ops for different clk types. > + > + guard(spinlock)(&ksc->clk_lock); > + reg = readb(cfg->mux_reg); > + > + return reg; > +} > + > +static unsigned long k230_clk_get_rate(struct clk_hw *hw, > + unsigned long parent_rate) > +{ > + struct k230_clk *clk = to_k230_clk(hw); > + struct k230_sysclk *ksc = clk->ksc; > + struct k230_clk_cfg *cfg = &k230_clk_cfgs[clk->id]; > + u32 mul, div; > + > + if (!cfg->have_rate) /* no divider, return parents' clk */ > + return parent_rate; > + > + guard(spinlock)(&ksc->clk_lock); > + switch (cfg->method) { > + /* > + * K230_MUL: div_mask+1/div_max... > + * K230_DIV: mul_max/div_mask+1 > + * K230_MUL_DIV: mul_mask/div_mask... > + */ > + case K230_MUL: > + div = cfg->rate_div_max; > + mul = (readl(cfg->rate_reg) >> cfg->rate_div_shift) > + & cfg->rate_div_mask; > + mul++; > + break; > + case K230_DIV: > + mul = cfg->rate_mul_max; > + div = (readl(cfg->rate_reg) >> cfg->rate_div_shift) > + & cfg->rate_div_mask; > + div++; > + break; > + case K230_MUL_DIV: > + if (!cfg->have_rate_c) { > + mul = (readl(cfg->rate_reg) >> cfg->rate_mul_shift) > + & cfg->rate_mul_mask; > + div = (readl(cfg->rate_reg) >> cfg->rate_div_shift) > + & cfg->rate_div_mask; > + } else { > + mul = (readl(cfg->rate_reg_c) >> cfg->rate_mul_shift_c) > + & cfg->rate_mul_mask_c; > + div = (readl(cfg->rate_reg_c) >> cfg->rate_div_shift) > + & cfg->rate_div_mask; > + } > + break; > + default: > + return 0; > + } > + > + return div_u64((u64)parent_rate * mul, div); > +} > + > +static int k230_clk_find_approximate(struct k230_clk *clk, > + u32 mul_min, > + u32 mul_max, > + u32 div_min, > + u32 div_max, > + enum k230_clk_div_type method, > + unsigned long rate, > + unsigned long parent_rate, > + u32 *div, > + u32 *mul) > +{ > + long abs_min; > + long abs_current; > + long perfect_divide; > + struct k230_clk_cfg *cfg = &k230_clk_cfgs[clk->id]; > + > + const u32 codec_clk[9] = { > + 2048000, > + 3072000, > + 4096000, > + 6144000, > + 8192000, > + 11289600, > + 12288000, > + 24576000, > + 49152000 > + }; > + > + const u32 codec_div[9][2] = { > + {3125, 16}, > + {3125, 24}, > + {3125, 32}, > + {3125, 48}, > + {3125, 64}, > + {15625, 441}, > + {3125, 96}, > + {3125, 192}, > + {3125, 384} > + }; > + > + const u32 pdm_clk[20] = { > + 128000, > + 192000, > + 256000, > + 384000, > + 512000, > + 768000, > + 1024000, > + 1411200, > + 1536000, > + 2048000, > + 2822400, > + 3072000, > + 4096000, > + 5644800, > + 6144000, > + 8192000, > + 11289600, > + 12288000, > + 24576000, > + 49152000 > + }; > + > + const u32 pdm_div[20][2] = { > + {3125, 1}, > + {6250, 3}, > + {3125, 2}, > + {3125, 3}, > + {3125, 4}, > + {3125, 6}, > + {3125, 8}, > + {125000, 441}, > + {3125, 12}, > + {3125, 16}, > + {62500, 441}, > + {3125, 24}, > + {3125, 32}, > + {31250, 441}, > + {3125, 48}, > + {3125, 64}, > + {15625, 441}, > + {3125, 96}, > + {3125, 192}, > + {3125, 384} > + }; > + > + switch (method) { > + /* only mul can be changeable 1/12,2/12,3/12...*/ > + case K230_MUL: > + perfect_divide = (long)((parent_rate * 1000) / rate); > + abs_min = abs(perfect_divide - > + (long)(((long)div_max * 1000) / (long)mul_min)); > + *mul = mul_min; > + > + for (u32 i = mul_min + 1; i <= mul_max; i++) { > + abs_current = abs(perfect_divide - > + (long)((long)((long)div_max * 1000) / (long)i)); > + if (abs_min > abs_current) { > + abs_min = abs_current; > + *mul = i; > + } > + } > + > + *div = div_max; > + break; > + /* only div can be changeable, 1/1,1/2,1/3...*/ > + case K230_DIV: > + perfect_divide = (long)((parent_rate * 1000) / rate); > + abs_min = abs(perfect_divide - > + (long)(((long)div_min * 1000) / (long)mul_max)); > + *div = div_min; > + > + for (u32 i = div_min + 1; i <= div_max; i++) { > + abs_current = abs(perfect_divide - > + (long)((long)((long)i * 1000) / (long)mul_max)); > + if (abs_min > abs_current) { > + abs_min = abs_current; > + *div = i; > + } > + } > + > + *mul = mul_max; > + break; > + /* mul and div can be changeable. */ > + case K230_MUL_DIV: > + if (cfg->rate_reg_off == K230_CLK_CODEC_ADC_MCLKDIV_OFFSET || > + cfg->rate_reg_off == K230_CLK_CODEC_DAC_MCLKDIV_OFFSET) { > + for (u32 j = 0; j < 9; j++) { > + if (0 == (rate - codec_clk[j])) { > + *div = codec_div[j][0]; > + *mul = codec_div[j][1]; > + } > + } > + } else if (cfg->rate_reg_off == K230_CLK_AUDIO_CLKDIV_OFFSET || > + cfg->rate_reg_off == K230_CLK_PDM_CLKDIV_OFFSET) { > + for (u32 j = 0; j < 20; j++) { > + if (0 == (rate - pdm_clk[j])) { > + *div = pdm_div[j][0]; > + *mul = pdm_div[j][1]; > + } > + } > + } else { > + return -EINVAL; > + } > + break; > + default: > + WARN_ON_ONCE(true); > + return -EPERM; This is impossible? Remove the default case and the compiler will warn if an enum case is missing. > + } > + return 0; > +} > + > +static long k230_clk_round_rate(struct clk_hw *hw, unsigned long rate, unsigned long *parent_rate) > +{ > + struct k230_clk *clk = to_k230_clk(hw); > + struct k230_sysclk *ksc = clk->ksc; > + struct k230_clk_cfg *cfg = &k230_clk_cfgs[clk->id]; > + u32 div = 0, mul = 0; > + > + if (k230_clk_find_approximate(clk, > + cfg->rate_mul_min, cfg->rate_mul_max, > + cfg->rate_div_min, cfg->rate_div_max, > + cfg->method, rate, *parent_rate, &div, &mul)) { > + dev_err(&ksc->pdev->dev, > + "[%s]: clk %s round rate error!\n", > + __func__, > + clk_hw_get_name(hw)); > + return 0; Why isn't an error returned? > + } > + > + return div_u64((u64)(*parent_rate) * mul, div); > +} > + > +static int k230_clk_set_rate(struct clk_hw *hw, unsigned long rate, > + unsigned long parent_rate) > +{ > + struct k230_clk *clk = to_k230_clk(hw); > + struct k230_sysclk *ksc = clk->ksc; > + struct k230_clk_cfg *cfg = &k230_clk_cfgs[clk->id]; > + u32 div = 0, mul = 0, reg = 0, reg_c; > + > + if (!cfg->have_rate || !cfg->rate_reg) > + return dev_err_probe(&ksc->pdev->dev, -EINVAL, "This clock may have no rate\n"); > + > + if (rate > parent_rate || rate == 0 || parent_rate == 0) > + return dev_err_probe(&ksc->pdev->dev, -EINVAL, "rate or parent_rate error\n"); > + > + if (cfg->read_only) > + return dev_err_probe(&ksc->pdev->dev, -EPERM, "This clk rate is read only\n"); > + > + if (k230_clk_find_approximate(clk, > + cfg->rate_mul_min, cfg->rate_mul_max, > + cfg->rate_div_min, cfg->rate_div_max, > + cfg->method, rate, parent_rate, &div, &mul)) > + return dev_err_probe(&ksc->pdev->dev, -EINVAL, "[%s]: clk %s set rate error!\n", > + __func__, clk_hw_get_name(hw)); > + > + guard(spinlock)(&ksc->clk_lock); > + if (!cfg->have_rate_c) { > + reg = readl(cfg->rate_reg); > + reg &= ~((cfg->rate_div_mask) << (cfg->rate_div_shift)); > + > + if (cfg->method == K230_DIV) { > + reg &= ~((cfg->rate_mul_mask) << (cfg->rate_mul_shift)); > + reg |= ((div - 1) & cfg->rate_div_mask) << (cfg->rate_div_shift); > + } else if (cfg->method == K230_MUL) { > + reg |= ((mul - 1) & cfg->rate_div_mask) << (cfg->rate_div_shift); > + } else { > + reg |= (mul & cfg->rate_mul_mask) << (cfg->rate_mul_shift); > + reg |= (div & cfg->rate_div_mask) << (cfg->rate_div_shift); > + } > + reg |= BIT(cfg->rate_write_enable_bit); > + } else { > + reg = readl(cfg->rate_reg); > + reg_c = readl(cfg->rate_reg_c); > + reg &= ~((cfg->rate_div_mask) << (cfg->rate_div_shift)); > + reg_c &= ~((cfg->rate_mul_mask_c) << (cfg->rate_mul_shift_c)); > + reg_c |= BIT(cfg->rate_write_enable_bit_c); > + > + reg_c |= (mul & cfg->rate_mul_mask_c) << (cfg->rate_mul_shift_c); > + reg |= (div & cfg->rate_div_mask) << (cfg->rate_div_shift); > + > + writel(reg_c, cfg->rate_reg_c); > + } > + writel(reg, cfg->rate_reg); > + > + return 0; > +} > + > +static const struct clk_ops k230_clk_ops_arr[K230_CLK_OPS_ID_NUM] = { > + [K230_CLK_OPS_ID_NONE] = { > + /* Sentinel */ > + }, > + [K230_CLK_OPS_ID_GATE_ONLY] = { > + K230_CLK_OPS_GATE, > + }, > + [K230_CLK_OPS_ID_RATE_ONLY] = { > + K230_CLK_OPS_RATE, > + }, > + [K230_CLK_OPS_ID_RATE_GATE] = { > + K230_CLK_OPS_RATE, > + K230_CLK_OPS_GATE, > + }, > + [K230_CLK_OPS_ID_MUX_ONLY] = { > + K230_CLK_OPS_MUX, > + }, > + [K230_CLK_OPS_ID_MUX_GATE] = { > + K230_CLK_OPS_MUX, > + K230_CLK_OPS_GATE, > + }, > + [K230_CLK_OPS_ID_MUX_RATE] = { > + K230_CLK_OPS_MUX, > + K230_CLK_OPS_RATE, > + }, > + [K230_CLK_OPS_ID_ALL] = { > + K230_CLK_OPS_MUX, > + K230_CLK_OPS_RATE, > + K230_CLK_OPS_GATE, > + }, > +}; > + > +static int k230_register_clk(struct platform_device *pdev, > + struct k230_sysclk *ksc, > + int id, > + const struct clk_parent_data *parent_data, > + u8 num_parents, > + unsigned long flags) > +{ > + struct k230_clk *clk = &ksc->clks[id]; > + struct k230_clk_cfg *cfg = &k230_clk_cfgs[id]; > + struct clk_init_data init = {}; > + int clk_id = 0; > + int ret; > + > + if (cfg->have_rate) { > + cfg->rate_reg = (ksc->regs + cfg->rate_reg_off); Drop useless parenthesis please. > + clk_id += K230_CLK_OPS_ID_RATE_ONLY; > + } > + > + if (cfg->have_mux) { > + cfg->mux_reg = (ksc->regs + cfg->mux_reg_off); > + clk_id += K230_CLK_OPS_ID_MUX_ONLY; > + > + /* mux clock doesn't match the case that num_parents less than 2 */ > + if (num_parents < 2) > + return -EINVAL; > + } > + > + if (cfg->have_gate) { > + cfg->gate_reg = (ksc->regs + cfg->gate_reg_off); Drop useless parenthesis please. > + clk_id += K230_CLK_OPS_ID_GATE_ONLY; > + } > + > + if (cfg->have_rate_c) > + cfg->rate_reg_c = (ksc->regs + cfg->rate_reg_off_c); Drop useless parenthesis please. > + > + init.name = k230_clk_cfgs[id].name; > + init.flags = flags; > + init.parent_data = parent_data; > + init.num_parents = num_parents; > + init.ops = &k230_clk_ops_arr[clk_id]; > + > + clk->id = id; > + clk->ksc = ksc; > + clk->hw.init = &init; > + > + ret = devm_clk_hw_register(&pdev->dev, &clk->hw); > + if (ret) > + return dev_err_probe(&pdev->dev, > + ret, > + "register clock %s failed\n", > + k230_clk_cfgs[id].name); > + > + return 0; > +} > + > +static int k230_register_mux_clk(struct platform_device *pdev, > + struct k230_sysclk *ksc, > + struct clk_parent_data *parent_data, > + int num_parent, > + int id) > +{ > + return k230_register_clk(pdev, ksc, id, > + (const struct clk_parent_data *)parent_data, Please remove useless cast. > + num_parent, 0); > +} > + > +static int k230_register_osc24m_child(struct platform_device *pdev, > + struct k230_sysclk *ksc, > + int id) > +{ > + const struct clk_parent_data parent_data = { > + .index = 0, > + }; > + return k230_register_clk(pdev, ksc, id, &parent_data, 1, 0); > +} > + > +static int k230_register_pll_child(struct platform_device *pdev, > + struct k230_sysclk *ksc, > + int id, > + enum k230_pll_id pll_id, > + unsigned long flags) > +{ > + const struct clk_parent_data parent_data = { > + .hw = &ksc->plls[pll_id].hw, > + }; > + return k230_register_clk(pdev, ksc, id, &parent_data, 1, flags); > +} > + > +static int k230_register_pll_div_child(struct platform_device *pdev, > + struct k230_sysclk *ksc, > + int id, > + enum k230_pll_div_id pll_div_id, > + unsigned long flags) > +{ > + const struct clk_parent_data parent_data = { > + .hw = ksc->dclks[pll_div_id].hw, > + }; > + return k230_register_clk(pdev, ksc, id, &parent_data, 1, flags); > +} > + > +static int k230_register_clk_child(struct platform_device *pdev, > + struct k230_sysclk *ksc, > + int id, > + int parent_id) > +{ > + const struct clk_parent_data parent_data = { > + .hw = &ksc->clks[parent_id].hw, > + }; > + return k230_register_clk(pdev, ksc, id, &parent_data, 1, 0); > +} > + > +static int _k230_clk_mux_get_parent_data(struct k230_sysclk *ksc, > + struct k230_clk_parent *pclk, > + struct clk_parent_data *parent_data) > +{ > + switch (pclk->type) { > + case K230_OSC24M: > + parent_data->index = 0; > + break; > + case K230_PLL: > + parent_data->hw = &ksc->plls[pclk->pll_id].hw; > + break; > + case K230_PLL_DIV: > + parent_data->hw = ksc->dclks[pclk->pll_div_id].hw; > + break; > + case K230_CLK_COMPOSITE: > + parent_data->hw = &ksc->clks[pclk->clk_id].hw; > + break; > + default: > + WARN_ON_ONCE(true); > + return -EPERM; > + } > + > + return 0; > +} > + > +static int k230_clk_mux_get_parent_data(struct k230_sysclk *ksc, > + struct k230_clk_cfg *cfg, > + struct clk_parent_data *parent_data, > + int num_parent) > +{ > + int ret; > + struct k230_clk_parent *pclk = cfg->parent; > + > + for (int i = 0; i < num_parent; i++) { > + ret = _k230_clk_mux_get_parent_data(ksc, &pclk[i], &parent_data[i]); > + if (ret) > + return ret; > + } > + > + return 0; > +} > + > +static int k230_register_clks(struct platform_device *pdev, struct k230_sysclk *ksc) > +{ > + struct k230_clk_cfg *cfg; > + struct k230_clk_parent *pclk; > + struct clk_parent_data parent_data[K230_CLK_MAX_PARENT_NUM]; > + int ret, i; > + > + /* > + * Single parent clock: > + * pll0_div2 sons: cpu0_src > + * pll0_div4 sons: cpu0_pclk > + * cpu0_src sons: cpu0_aclk, cpu0_plic, cpu0_noc_ddrcp4, pmu_pclk > + * > + * Mux clock: > + * hs_ospi_src parents: pll0_div2, pll2_div4 > + */ > + for (i = 0; i < K230_CLK_NUM; i++) { > + cfg = &k230_clk_cfgs[i]; > + if (!cfg->status) > + continue; > + > + if (cfg->have_mux) { > + ret = k230_clk_mux_get_parent_data(ksc, cfg, parent_data, > + cfg->num_parent); > + if (ret) > + return dev_err_probe(&pdev->dev, ret, > + "Failed to get parent data\n"); > + > + ret = k230_register_mux_clk(pdev, ksc, parent_data, > + cfg->num_parent, i); > + } else { > + pclk = cfg->parent; > + switch (pclk->type) { > + case K230_OSC24M: > + ret = k230_register_osc24m_child(pdev, ksc, i); > + break; > + case K230_PLL: > + ret = k230_register_pll_child(pdev, ksc, i, > + pclk->pll_id, cfg->flags); > + break; > + case K230_PLL_DIV: > + ret = k230_register_pll_div_child(pdev, ksc, i, > + pclk->pll_div_id, cfg->flags); > + break; > + case K230_CLK_COMPOSITE: > + ret = k230_register_clk_child(pdev, ksc, i, > + pclk->clk_id); > + break; > + default: > + return dev_err_probe(&pdev->dev, -EINVAL, "Invalid type\n"); > + } > + } > + if (ret) > + return dev_err_probe(&pdev->dev, ret, "register child id %d failed\n", i); > + } > + > + return 0; > +} > + > +static struct clk_hw *k230_clk_hw_onecell_get(struct of_phandle_args *clkspec, void *data) > +{ > + struct k230_sysclk *ksc; > + unsigned int idx; > + > + if (clkspec->args_count != 1) > + return ERR_PTR(-EINVAL); > + > + idx = clkspec->args[0]; > + if (idx >= K230_CLK_NUM) > + return ERR_PTR(-EINVAL); > + > + if (!data) > + return ERR_PTR(-EINVAL); > + > + ksc = (struct k230_sysclk *)data; Drop useless cast please. > + > + return &ksc->clks[idx].hw; > +} > + > +static int k230_clk_init_plls(struct platform_device *pdev) > +{ > + int ret; > + struct k230_sysclk *ksc = platform_get_drvdata(pdev); > + > + spin_lock_init(&ksc->pll_lock); > + > + ksc->pll_regs = devm_platform_ioremap_resource(pdev, 0); > + if (!ksc->pll_regs) Does it return an error pointer or NULL on failure? The dev_err_probe() below is incorrect or the if condition above is. > + return dev_err_probe(&pdev->dev, PTR_ERR(ksc->pll_regs), "map registers failed\n"); Doesn't this already print an error? > + > + ret = k230_register_plls(pdev, ksc); > + if (ret) > + return dev_err_probe(&pdev->dev, ret, "register plls failed\n"); > + > + ret = k230_register_pll_divs(pdev, ksc); > + if (ret) > + return dev_err_probe(&pdev->dev, ret, "register pll_divs failed\n"); > + > + for (int i = 0; i < K230_PLL_DIV_NUM; i++) { > + ret = devm_clk_hw_register_clkdev(&pdev->dev, ksc->dclks[i].hw, > + k230_pll_div_cfgs[i].name, NULL); Why is clkdev being used? > + if (ret) > + return dev_err_probe(&pdev->dev, ret, "clock_lookup create failed\n"); > + } > + > + return 0; > +} > + > +static int k230_clk_init_clks(struct platform_device *pdev) > +{ > + int ret; > + struct k230_sysclk *ksc = platform_get_drvdata(pdev); > + > + spin_lock_init(&ksc->clk_lock); > + > + ksc->regs = devm_platform_ioremap_resource(pdev, 1); > + if (!ksc->regs) > + return dev_err_probe(&pdev->dev, PTR_ERR(ksc->regs), "failed to map registers\n"); > + > + ret = k230_register_clks(pdev, ksc); > + if (ret) > + return dev_err_probe(&pdev->dev, ret, "register clock provider failed\n"); > + > + ret = devm_of_clk_add_hw_provider(&pdev->dev, k230_clk_hw_onecell_get, ksc); > + if (ret) > + return dev_err_probe(&pdev->dev, ret, "add clock provider failed\n"); > + > + return 0; > +} > + > +static int k230_clk_probe(struct platform_device *pdev) > +{ > + int ret; > + struct k230_sysclk *ksc; > + > + ksc = devm_kzalloc(&pdev->dev, sizeof(struct k230_sysclk), GFP_KERNEL); > + if (!ksc) > + return -ENOMEM; > + > + ksc->plls = devm_kmalloc_array(&pdev->dev, K230_PLL_NUM, Any reason devm_kcalloc() isn't used? > + sizeof(struct k230_pll), GFP_KERNEL); > + if (!ksc->plls) > + return -ENOMEM; > + > + ksc->dclks = devm_kmalloc_array(&pdev->dev, K230_PLL_DIV_NUM, > + sizeof(struct k230_pll_div), GFP_KERNEL); > + if (!ksc->dclks) > + return -ENOMEM; > + > + ksc->clks = devm_kmalloc_array(&pdev->dev, K230_CLK_NUM, > + sizeof(struct k230_clk), GFP_KERNEL); > + if (!ksc->clks) > + return -ENOMEM; > + > + ksc->pdev = pdev; > + platform_set_drvdata(pdev, ksc); > + > + ret = k230_clk_init_plls(pdev); > + if (ret) > + return dev_err_probe(&pdev->dev, ret, "init plls failed\n"); > + > + ret = k230_clk_init_clks(pdev); > + if (ret) > + return dev_err_probe(&pdev->dev, ret, "init clks failed\n"); > + > + return 0; > +} > + > +static const struct of_device_id k230_clk_ids[] = { > + { .compatible = "canaan,k230-clk" }, > + { /* Sentinel */ }, Nitpick: Please drop last comma so that nothing can come after without causing a compilation error. > +}; > +MODULE_DEVICE_TABLE(of, k230_clk_ids);