|  | // SPDX-License-Identifier: GPL-2.0 | 
|  | /* | 
|  | * RZ/G2L Clock Pulse Generator | 
|  | * | 
|  | * Copyright (C) 2021 Renesas Electronics Corp. | 
|  | * | 
|  | * Based on renesas-cpg-mssr.c | 
|  | * | 
|  | * Copyright (C) 2015 Glider bvba | 
|  | * Copyright (C) 2013 Ideas On Board SPRL | 
|  | * Copyright (C) 2015 Renesas Electronics Corp. | 
|  | */ | 
|  |  | 
|  | #include <linux/bitfield.h> | 
|  | #include <linux/clk.h> | 
|  | #include <linux/clk-provider.h> | 
|  | #include <linux/clk/renesas.h> | 
|  | #include <linux/delay.h> | 
|  | #include <linux/device.h> | 
|  | #include <linux/init.h> | 
|  | #include <linux/iopoll.h> | 
|  | #include <linux/mod_devicetable.h> | 
|  | #include <linux/module.h> | 
|  | #include <linux/of.h> | 
|  | #include <linux/platform_device.h> | 
|  | #include <linux/pm_clock.h> | 
|  | #include <linux/pm_domain.h> | 
|  | #include <linux/reset-controller.h> | 
|  | #include <linux/slab.h> | 
|  | #include <linux/units.h> | 
|  |  | 
|  | #include <dt-bindings/clock/renesas-cpg-mssr.h> | 
|  |  | 
|  | #include "rzg2l-cpg.h" | 
|  |  | 
|  | #ifdef DEBUG | 
|  | #define WARN_DEBUG(x)	WARN_ON(x) | 
|  | #else | 
|  | #define WARN_DEBUG(x)	do { } while (0) | 
|  | #endif | 
|  |  | 
|  | #define GET_SHIFT(val)		((val >> 12) & 0xff) | 
|  | #define GET_WIDTH(val)		((val >> 8) & 0xf) | 
|  |  | 
|  | #define KDIV(val)		((s16)FIELD_GET(GENMASK(31, 16), val)) | 
|  | #define MDIV(val)		FIELD_GET(GENMASK(15, 6), val) | 
|  | #define PDIV(val)		FIELD_GET(GENMASK(5, 0), val) | 
|  | #define SDIV(val)		FIELD_GET(GENMASK(2, 0), val) | 
|  |  | 
|  | #define RZG3S_DIV_P		GENMASK(28, 26) | 
|  | #define RZG3S_DIV_M		GENMASK(25, 22) | 
|  | #define RZG3S_DIV_NI		GENMASK(21, 13) | 
|  | #define RZG3S_DIV_NF		GENMASK(12, 1) | 
|  |  | 
|  | #define CLK_ON_R(reg)		(reg) | 
|  | #define CLK_MON_R(reg)		(0x180 + (reg)) | 
|  | #define CLK_RST_R(reg)		(reg) | 
|  | #define CLK_MRST_R(reg)		(0x180 + (reg)) | 
|  |  | 
|  | #define GET_REG_OFFSET(val)		((val >> 20) & 0xfff) | 
|  | #define GET_REG_SAMPLL_CLK1(val)	((val >> 22) & 0xfff) | 
|  | #define GET_REG_SAMPLL_CLK2(val)	((val >> 12) & 0xfff) | 
|  |  | 
|  | #define CPG_WEN_BIT		BIT(16) | 
|  |  | 
|  | #define MAX_VCLK_FREQ		(148500000) | 
|  |  | 
|  | /** | 
|  | * struct clk_hw_data - clock hardware data | 
|  | * @hw: clock hw | 
|  | * @conf: clock configuration (register offset, shift, width) | 
|  | * @sconf: clock status configuration (register offset, shift, width) | 
|  | * @priv: CPG private data structure | 
|  | */ | 
|  | struct clk_hw_data { | 
|  | struct clk_hw hw; | 
|  | u32 conf; | 
|  | u32 sconf; | 
|  | struct rzg2l_cpg_priv *priv; | 
|  | }; | 
|  |  | 
|  | #define to_clk_hw_data(_hw)	container_of(_hw, struct clk_hw_data, hw) | 
|  |  | 
|  | /** | 
|  | * struct sd_mux_hw_data - SD MUX clock hardware data | 
|  | * @hw_data: clock hw data | 
|  | * @mtable: clock mux table | 
|  | */ | 
|  | struct sd_mux_hw_data { | 
|  | struct clk_hw_data hw_data; | 
|  | const u32 *mtable; | 
|  | }; | 
|  |  | 
|  | #define to_sd_mux_hw_data(_hw)	container_of(_hw, struct sd_mux_hw_data, hw_data) | 
|  |  | 
|  | /** | 
|  | * struct div_hw_data - divider clock hardware data | 
|  | * @hw_data: clock hw data | 
|  | * @dtable: pointer to divider table | 
|  | * @invalid_rate: invalid rate for divider | 
|  | * @max_rate: maximum rate for divider | 
|  | * @width: divider width | 
|  | */ | 
|  | struct div_hw_data { | 
|  | struct clk_hw_data hw_data; | 
|  | const struct clk_div_table *dtable; | 
|  | unsigned long invalid_rate; | 
|  | unsigned long max_rate; | 
|  | u32 width; | 
|  | }; | 
|  |  | 
|  | #define to_div_hw_data(_hw)	container_of(_hw, struct div_hw_data, hw_data) | 
|  |  | 
|  | struct rzg2l_pll5_param { | 
|  | u32 pl5_fracin; | 
|  | u8 pl5_refdiv; | 
|  | u8 pl5_intin; | 
|  | u8 pl5_postdiv1; | 
|  | u8 pl5_postdiv2; | 
|  | u8 pl5_spread; | 
|  | }; | 
|  |  | 
|  | struct rzg2l_pll5_mux_dsi_div_param { | 
|  | u8 clksrc; | 
|  | u8 dsi_div_a; | 
|  | u8 dsi_div_b; | 
|  | }; | 
|  |  | 
|  | /** | 
|  | * struct rzg2l_cpg_priv - Clock Pulse Generator Private Data | 
|  | * | 
|  | * @rcdev: Reset controller entity | 
|  | * @dev: CPG device | 
|  | * @base: CPG register block base address | 
|  | * @rmw_lock: protects register accesses | 
|  | * @clks: Array containing all Core and Module Clocks | 
|  | * @num_core_clks: Number of Core Clocks in clks[] | 
|  | * @num_mod_clks: Number of Module Clocks in clks[] | 
|  | * @num_resets: Number of Module Resets in info->resets[] | 
|  | * @last_dt_core_clk: ID of the last Core Clock exported to DT | 
|  | * @info: Pointer to platform data | 
|  | * @mux_dsi_div_params: pll5 mux and dsi div parameters | 
|  | */ | 
|  | struct rzg2l_cpg_priv { | 
|  | struct reset_controller_dev rcdev; | 
|  | struct device *dev; | 
|  | void __iomem *base; | 
|  | spinlock_t rmw_lock; | 
|  |  | 
|  | struct clk **clks; | 
|  | unsigned int num_core_clks; | 
|  | unsigned int num_mod_clks; | 
|  | unsigned int num_resets; | 
|  | unsigned int last_dt_core_clk; | 
|  |  | 
|  | const struct rzg2l_cpg_info *info; | 
|  |  | 
|  | struct rzg2l_pll5_mux_dsi_div_param mux_dsi_div_params; | 
|  | }; | 
|  |  | 
|  | static void rzg2l_cpg_del_clk_provider(void *data) | 
|  | { | 
|  | of_clk_del_provider(data); | 
|  | } | 
|  |  | 
|  | /* Must be called in atomic context. */ | 
|  | static int rzg2l_cpg_wait_clk_update_done(void __iomem *base, u32 conf) | 
|  | { | 
|  | u32 bitmask = GENMASK(GET_WIDTH(conf) - 1, 0) << GET_SHIFT(conf); | 
|  | u32 off = GET_REG_OFFSET(conf); | 
|  | u32 val; | 
|  |  | 
|  | return readl_poll_timeout_atomic(base + off, val, !(val & bitmask), 10, 200); | 
|  | } | 
|  |  | 
|  | int rzg2l_cpg_sd_clk_mux_notifier(struct notifier_block *nb, unsigned long event, | 
|  | void *data) | 
|  | { | 
|  | struct clk_notifier_data *cnd = data; | 
|  | struct clk_hw *hw = __clk_get_hw(cnd->clk); | 
|  | struct clk_hw_data *clk_hw_data = to_clk_hw_data(hw); | 
|  | struct rzg2l_cpg_priv *priv = clk_hw_data->priv; | 
|  | u32 off = GET_REG_OFFSET(clk_hw_data->conf); | 
|  | u32 shift = GET_SHIFT(clk_hw_data->conf); | 
|  | const u32 clk_src_266 = 3; | 
|  | unsigned long flags; | 
|  | int ret; | 
|  |  | 
|  | if (event != PRE_RATE_CHANGE || (cnd->new_rate / MEGA == 266)) | 
|  | return NOTIFY_DONE; | 
|  |  | 
|  | spin_lock_irqsave(&priv->rmw_lock, flags); | 
|  |  | 
|  | /* | 
|  | * As per the HW manual, we should not directly switch from 533 MHz to | 
|  | * 400 MHz and vice versa. To change the setting from 2’b01 (533 MHz) | 
|  | * to 2’b10 (400 MHz) or vice versa, Switch to 2’b11 (266 MHz) first, | 
|  | * and then switch to the target setting (2’b01 (533 MHz) or 2’b10 | 
|  | * (400 MHz)). | 
|  | * Setting a value of '0' to the SEL_SDHI0_SET or SEL_SDHI1_SET clock | 
|  | * switching register is prohibited. | 
|  | * The clock mux has 3 input clocks(533 MHz, 400 MHz, and 266 MHz), and | 
|  | * the index to value mapping is done by adding 1 to the index. | 
|  | */ | 
|  |  | 
|  | writel((CPG_WEN_BIT | clk_src_266) << shift, priv->base + off); | 
|  |  | 
|  | /* Wait for the update done. */ | 
|  | ret = rzg2l_cpg_wait_clk_update_done(priv->base, clk_hw_data->sconf); | 
|  |  | 
|  | spin_unlock_irqrestore(&priv->rmw_lock, flags); | 
|  |  | 
|  | if (ret) | 
|  | dev_err(priv->dev, "failed to switch to safe clk source\n"); | 
|  |  | 
|  | return notifier_from_errno(ret); | 
|  | } | 
|  |  | 
|  | int rzg3s_cpg_div_clk_notifier(struct notifier_block *nb, unsigned long event, | 
|  | void *data) | 
|  | { | 
|  | struct clk_notifier_data *cnd = data; | 
|  | struct clk_hw *hw = __clk_get_hw(cnd->clk); | 
|  | struct clk_hw_data *clk_hw_data = to_clk_hw_data(hw); | 
|  | struct div_hw_data *div_hw_data = to_div_hw_data(clk_hw_data); | 
|  | struct rzg2l_cpg_priv *priv = clk_hw_data->priv; | 
|  | u32 off = GET_REG_OFFSET(clk_hw_data->conf); | 
|  | u32 shift = GET_SHIFT(clk_hw_data->conf); | 
|  | unsigned long flags; | 
|  | int ret = 0; | 
|  | u32 val; | 
|  |  | 
|  | if (event != PRE_RATE_CHANGE || !div_hw_data->invalid_rate || | 
|  | div_hw_data->invalid_rate % cnd->new_rate) | 
|  | return NOTIFY_DONE; | 
|  |  | 
|  | spin_lock_irqsave(&priv->rmw_lock, flags); | 
|  |  | 
|  | val = readl(priv->base + off); | 
|  | val >>= shift; | 
|  | val &= GENMASK(GET_WIDTH(clk_hw_data->conf) - 1, 0); | 
|  |  | 
|  | /* | 
|  | * There are different constraints for the user of this notifiers as follows: | 
|  | * 1/ SD div cannot be 1 (val == 0) if parent rate is 800MHz | 
|  | * 2/ OCTA / SPI div cannot be 1 (val == 0) if parent rate is 400MHz | 
|  | * As SD can have only one parent having 800MHz and OCTA div can have | 
|  | * only one parent having 400MHz we took into account the parent rate | 
|  | * at the beginning of function (by checking invalid_rate % new_rate). | 
|  | * Now it is time to check the hardware divider and update it accordingly. | 
|  | */ | 
|  | if (!val) { | 
|  | writel((CPG_WEN_BIT | 1) << shift, priv->base + off); | 
|  | /* Wait for the update done. */ | 
|  | ret = rzg2l_cpg_wait_clk_update_done(priv->base, clk_hw_data->sconf); | 
|  | } | 
|  |  | 
|  | spin_unlock_irqrestore(&priv->rmw_lock, flags); | 
|  |  | 
|  | if (ret) | 
|  | dev_err(priv->dev, "Failed to downgrade the div\n"); | 
|  |  | 
|  | return notifier_from_errno(ret); | 
|  | } | 
|  |  | 
|  | static int rzg2l_register_notifier(struct clk_hw *hw, const struct cpg_core_clk *core, | 
|  | struct rzg2l_cpg_priv *priv) | 
|  | { | 
|  | struct notifier_block *nb; | 
|  |  | 
|  | if (!core->notifier) | 
|  | return 0; | 
|  |  | 
|  | nb = devm_kzalloc(priv->dev, sizeof(*nb), GFP_KERNEL); | 
|  | if (!nb) | 
|  | return -ENOMEM; | 
|  |  | 
|  | nb->notifier_call = core->notifier; | 
|  |  | 
|  | return clk_notifier_register(hw->clk, nb); | 
|  | } | 
|  |  | 
|  | static unsigned long rzg3s_div_clk_recalc_rate(struct clk_hw *hw, | 
|  | unsigned long parent_rate) | 
|  | { | 
|  | struct clk_hw_data *clk_hw_data = to_clk_hw_data(hw); | 
|  | struct div_hw_data *div_hw_data = to_div_hw_data(clk_hw_data); | 
|  | struct rzg2l_cpg_priv *priv = clk_hw_data->priv; | 
|  | u32 val; | 
|  |  | 
|  | val = readl(priv->base + GET_REG_OFFSET(clk_hw_data->conf)); | 
|  | val >>= GET_SHIFT(clk_hw_data->conf); | 
|  | val &= GENMASK(GET_WIDTH(clk_hw_data->conf) - 1, 0); | 
|  |  | 
|  | return divider_recalc_rate(hw, parent_rate, val, div_hw_data->dtable, | 
|  | CLK_DIVIDER_ROUND_CLOSEST, div_hw_data->width); | 
|  | } | 
|  |  | 
|  | static int rzg3s_div_clk_determine_rate(struct clk_hw *hw, struct clk_rate_request *req) | 
|  | { | 
|  | struct clk_hw_data *clk_hw_data = to_clk_hw_data(hw); | 
|  | struct div_hw_data *div_hw_data = to_div_hw_data(clk_hw_data); | 
|  |  | 
|  | if (div_hw_data->max_rate && req->rate > div_hw_data->max_rate) | 
|  | req->rate = div_hw_data->max_rate; | 
|  |  | 
|  | return divider_determine_rate(hw, req, div_hw_data->dtable, div_hw_data->width, | 
|  | CLK_DIVIDER_ROUND_CLOSEST); | 
|  | } | 
|  |  | 
|  | static int rzg3s_div_clk_set_rate(struct clk_hw *hw, unsigned long rate, | 
|  | unsigned long parent_rate) | 
|  | { | 
|  | struct clk_hw_data *clk_hw_data = to_clk_hw_data(hw); | 
|  | struct div_hw_data *div_hw_data = to_div_hw_data(clk_hw_data); | 
|  | struct rzg2l_cpg_priv *priv = clk_hw_data->priv; | 
|  | u32 off = GET_REG_OFFSET(clk_hw_data->conf); | 
|  | u32 shift = GET_SHIFT(clk_hw_data->conf); | 
|  | unsigned long flags; | 
|  | u32 val; | 
|  | int ret; | 
|  |  | 
|  | val = divider_get_val(rate, parent_rate, div_hw_data->dtable, div_hw_data->width, | 
|  | CLK_DIVIDER_ROUND_CLOSEST); | 
|  |  | 
|  | spin_lock_irqsave(&priv->rmw_lock, flags); | 
|  | writel((CPG_WEN_BIT | val) << shift, priv->base + off); | 
|  | /* Wait for the update done. */ | 
|  | ret = rzg2l_cpg_wait_clk_update_done(priv->base, clk_hw_data->sconf); | 
|  | spin_unlock_irqrestore(&priv->rmw_lock, flags); | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static const struct clk_ops rzg3s_div_clk_ops = { | 
|  | .recalc_rate = rzg3s_div_clk_recalc_rate, | 
|  | .determine_rate = rzg3s_div_clk_determine_rate, | 
|  | .set_rate = rzg3s_div_clk_set_rate, | 
|  | }; | 
|  |  | 
|  | static struct clk * __init | 
|  | rzg3s_cpg_div_clk_register(const struct cpg_core_clk *core, struct rzg2l_cpg_priv *priv) | 
|  | { | 
|  | struct div_hw_data *div_hw_data; | 
|  | struct clk_init_data init = {}; | 
|  | const struct clk_div_table *clkt; | 
|  | struct clk_hw *clk_hw; | 
|  | const struct clk *parent; | 
|  | const char *parent_name; | 
|  | u32 max = 0; | 
|  | int ret; | 
|  |  | 
|  | parent = priv->clks[core->parent]; | 
|  | if (IS_ERR(parent)) | 
|  | return ERR_CAST(parent); | 
|  |  | 
|  | parent_name = __clk_get_name(parent); | 
|  |  | 
|  | div_hw_data = devm_kzalloc(priv->dev, sizeof(*div_hw_data), GFP_KERNEL); | 
|  | if (!div_hw_data) | 
|  | return ERR_PTR(-ENOMEM); | 
|  |  | 
|  | init.name = core->name; | 
|  | init.flags = core->flag; | 
|  | init.ops = &rzg3s_div_clk_ops; | 
|  | init.parent_names = &parent_name; | 
|  | init.num_parents = 1; | 
|  |  | 
|  | /* Get the maximum divider to retrieve div width. */ | 
|  | for (clkt = core->dtable; clkt->div; clkt++) { | 
|  | if (max < clkt->div) | 
|  | max = clkt->div; | 
|  | } | 
|  |  | 
|  | div_hw_data->hw_data.priv = priv; | 
|  | div_hw_data->hw_data.conf = core->conf; | 
|  | div_hw_data->hw_data.sconf = core->sconf; | 
|  | div_hw_data->dtable = core->dtable; | 
|  | div_hw_data->invalid_rate = core->invalid_rate; | 
|  | div_hw_data->max_rate = core->max_rate; | 
|  | div_hw_data->width = fls(max) - 1; | 
|  |  | 
|  | clk_hw = &div_hw_data->hw_data.hw; | 
|  | clk_hw->init = &init; | 
|  |  | 
|  | ret = devm_clk_hw_register(priv->dev, clk_hw); | 
|  | if (ret) | 
|  | return ERR_PTR(ret); | 
|  |  | 
|  | ret = rzg2l_register_notifier(clk_hw, core, priv); | 
|  | if (ret) { | 
|  | dev_err(priv->dev, "Failed to register notifier for %s\n", | 
|  | core->name); | 
|  | return ERR_PTR(ret); | 
|  | } | 
|  |  | 
|  | return clk_hw->clk; | 
|  | } | 
|  |  | 
|  | static struct clk * __init | 
|  | rzg2l_cpg_div_clk_register(const struct cpg_core_clk *core, | 
|  | struct rzg2l_cpg_priv *priv) | 
|  | { | 
|  | void __iomem *base = priv->base; | 
|  | struct device *dev = priv->dev; | 
|  | const struct clk *parent; | 
|  | const char *parent_name; | 
|  | struct clk_hw *clk_hw; | 
|  |  | 
|  | parent = priv->clks[core->parent]; | 
|  | if (IS_ERR(parent)) | 
|  | return ERR_CAST(parent); | 
|  |  | 
|  | parent_name = __clk_get_name(parent); | 
|  |  | 
|  | if (core->dtable) | 
|  | clk_hw = clk_hw_register_divider_table(dev, core->name, | 
|  | parent_name, 0, | 
|  | base + GET_REG_OFFSET(core->conf), | 
|  | GET_SHIFT(core->conf), | 
|  | GET_WIDTH(core->conf), | 
|  | core->flag, | 
|  | core->dtable, | 
|  | &priv->rmw_lock); | 
|  | else | 
|  | clk_hw = clk_hw_register_divider(dev, core->name, | 
|  | parent_name, 0, | 
|  | base + GET_REG_OFFSET(core->conf), | 
|  | GET_SHIFT(core->conf), | 
|  | GET_WIDTH(core->conf), | 
|  | core->flag, &priv->rmw_lock); | 
|  |  | 
|  | if (IS_ERR(clk_hw)) | 
|  | return ERR_CAST(clk_hw); | 
|  |  | 
|  | return clk_hw->clk; | 
|  | } | 
|  |  | 
|  | static struct clk * __init | 
|  | rzg2l_cpg_mux_clk_register(const struct cpg_core_clk *core, | 
|  | struct rzg2l_cpg_priv *priv) | 
|  | { | 
|  | const struct clk_hw *clk_hw; | 
|  |  | 
|  | clk_hw = devm_clk_hw_register_mux(priv->dev, core->name, | 
|  | core->parent_names, core->num_parents, | 
|  | core->flag, | 
|  | priv->base + GET_REG_OFFSET(core->conf), | 
|  | GET_SHIFT(core->conf), | 
|  | GET_WIDTH(core->conf), | 
|  | core->mux_flags, &priv->rmw_lock); | 
|  | if (IS_ERR(clk_hw)) | 
|  | return ERR_CAST(clk_hw); | 
|  |  | 
|  | return clk_hw->clk; | 
|  | } | 
|  |  | 
|  | static int rzg2l_cpg_sd_clk_mux_set_parent(struct clk_hw *hw, u8 index) | 
|  | { | 
|  | struct clk_hw_data *clk_hw_data = to_clk_hw_data(hw); | 
|  | struct sd_mux_hw_data *sd_mux_hw_data = to_sd_mux_hw_data(clk_hw_data); | 
|  | struct rzg2l_cpg_priv *priv = clk_hw_data->priv; | 
|  | u32 off = GET_REG_OFFSET(clk_hw_data->conf); | 
|  | u32 shift = GET_SHIFT(clk_hw_data->conf); | 
|  | unsigned long flags; | 
|  | u32 val; | 
|  | int ret; | 
|  |  | 
|  | val = clk_mux_index_to_val(sd_mux_hw_data->mtable, CLK_MUX_ROUND_CLOSEST, index); | 
|  |  | 
|  | spin_lock_irqsave(&priv->rmw_lock, flags); | 
|  |  | 
|  | writel((CPG_WEN_BIT | val) << shift, priv->base + off); | 
|  |  | 
|  | /* Wait for the update done. */ | 
|  | ret = rzg2l_cpg_wait_clk_update_done(priv->base, clk_hw_data->sconf); | 
|  |  | 
|  | spin_unlock_irqrestore(&priv->rmw_lock, flags); | 
|  |  | 
|  | if (ret) | 
|  | dev_err(priv->dev, "Failed to switch parent\n"); | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static u8 rzg2l_cpg_sd_clk_mux_get_parent(struct clk_hw *hw) | 
|  | { | 
|  | struct clk_hw_data *clk_hw_data = to_clk_hw_data(hw); | 
|  | struct sd_mux_hw_data *sd_mux_hw_data = to_sd_mux_hw_data(clk_hw_data); | 
|  | struct rzg2l_cpg_priv *priv = clk_hw_data->priv; | 
|  | u32 val; | 
|  |  | 
|  | val = readl(priv->base + GET_REG_OFFSET(clk_hw_data->conf)); | 
|  | val >>= GET_SHIFT(clk_hw_data->conf); | 
|  | val &= GENMASK(GET_WIDTH(clk_hw_data->conf) - 1, 0); | 
|  |  | 
|  | return clk_mux_val_to_index(hw, sd_mux_hw_data->mtable, CLK_MUX_ROUND_CLOSEST, val); | 
|  | } | 
|  |  | 
|  | static const struct clk_ops rzg2l_cpg_sd_clk_mux_ops = { | 
|  | .determine_rate = __clk_mux_determine_rate_closest, | 
|  | .set_parent	= rzg2l_cpg_sd_clk_mux_set_parent, | 
|  | .get_parent	= rzg2l_cpg_sd_clk_mux_get_parent, | 
|  | }; | 
|  |  | 
|  | static struct clk * __init | 
|  | rzg2l_cpg_sd_mux_clk_register(const struct cpg_core_clk *core, | 
|  | struct rzg2l_cpg_priv *priv) | 
|  | { | 
|  | struct sd_mux_hw_data *sd_mux_hw_data; | 
|  | struct clk_init_data init; | 
|  | struct clk_hw *clk_hw; | 
|  | int ret; | 
|  |  | 
|  | sd_mux_hw_data = devm_kzalloc(priv->dev, sizeof(*sd_mux_hw_data), GFP_KERNEL); | 
|  | if (!sd_mux_hw_data) | 
|  | return ERR_PTR(-ENOMEM); | 
|  |  | 
|  | sd_mux_hw_data->hw_data.priv = priv; | 
|  | sd_mux_hw_data->hw_data.conf = core->conf; | 
|  | sd_mux_hw_data->hw_data.sconf = core->sconf; | 
|  | sd_mux_hw_data->mtable = core->mtable; | 
|  |  | 
|  | init.name = core->name; | 
|  | init.ops = &rzg2l_cpg_sd_clk_mux_ops; | 
|  | init.flags = core->flag; | 
|  | init.num_parents = core->num_parents; | 
|  | init.parent_names = core->parent_names; | 
|  |  | 
|  | clk_hw = &sd_mux_hw_data->hw_data.hw; | 
|  | clk_hw->init = &init; | 
|  |  | 
|  | ret = devm_clk_hw_register(priv->dev, clk_hw); | 
|  | if (ret) | 
|  | return ERR_PTR(ret); | 
|  |  | 
|  | ret = rzg2l_register_notifier(clk_hw, core, priv); | 
|  | if (ret) { | 
|  | dev_err(priv->dev, "Failed to register notifier for %s\n", | 
|  | core->name); | 
|  | return ERR_PTR(ret); | 
|  | } | 
|  |  | 
|  | return clk_hw->clk; | 
|  | } | 
|  |  | 
|  | static unsigned long | 
|  | rzg2l_cpg_get_foutpostdiv_rate(struct rzg2l_pll5_param *params, | 
|  | unsigned long rate) | 
|  | { | 
|  | unsigned long foutpostdiv_rate; | 
|  |  | 
|  | params->pl5_intin = rate / MEGA; | 
|  | params->pl5_fracin = div_u64(((u64)rate % MEGA) << 24, MEGA); | 
|  | params->pl5_refdiv = 2; | 
|  | params->pl5_postdiv1 = 1; | 
|  | params->pl5_postdiv2 = 1; | 
|  | params->pl5_spread = 0x16; | 
|  |  | 
|  | foutpostdiv_rate = | 
|  | EXTAL_FREQ_IN_MEGA_HZ * MEGA / params->pl5_refdiv * | 
|  | ((((params->pl5_intin << 24) + params->pl5_fracin)) >> 24) / | 
|  | (params->pl5_postdiv1 * params->pl5_postdiv2); | 
|  |  | 
|  | return foutpostdiv_rate; | 
|  | } | 
|  |  | 
|  | struct dsi_div_hw_data { | 
|  | struct clk_hw hw; | 
|  | u32 conf; | 
|  | unsigned long rate; | 
|  | struct rzg2l_cpg_priv *priv; | 
|  | }; | 
|  |  | 
|  | #define to_dsi_div_hw_data(_hw)	container_of(_hw, struct dsi_div_hw_data, hw) | 
|  |  | 
|  | static unsigned long rzg2l_cpg_dsi_div_recalc_rate(struct clk_hw *hw, | 
|  | unsigned long parent_rate) | 
|  | { | 
|  | struct dsi_div_hw_data *dsi_div = to_dsi_div_hw_data(hw); | 
|  | unsigned long rate = dsi_div->rate; | 
|  |  | 
|  | if (!rate) | 
|  | rate = parent_rate; | 
|  |  | 
|  | return rate; | 
|  | } | 
|  |  | 
|  | static unsigned long rzg2l_cpg_get_vclk_parent_rate(struct clk_hw *hw, | 
|  | unsigned long rate) | 
|  | { | 
|  | struct dsi_div_hw_data *dsi_div = to_dsi_div_hw_data(hw); | 
|  | struct rzg2l_cpg_priv *priv = dsi_div->priv; | 
|  | struct rzg2l_pll5_param params; | 
|  | unsigned long parent_rate; | 
|  |  | 
|  | parent_rate = rzg2l_cpg_get_foutpostdiv_rate(¶ms, rate); | 
|  |  | 
|  | if (priv->mux_dsi_div_params.clksrc) | 
|  | parent_rate /= 2; | 
|  |  | 
|  | return parent_rate; | 
|  | } | 
|  |  | 
|  | static int rzg2l_cpg_dsi_div_determine_rate(struct clk_hw *hw, | 
|  | struct clk_rate_request *req) | 
|  | { | 
|  | if (req->rate > MAX_VCLK_FREQ) | 
|  | req->rate = MAX_VCLK_FREQ; | 
|  |  | 
|  | req->best_parent_rate = rzg2l_cpg_get_vclk_parent_rate(hw, req->rate); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int rzg2l_cpg_dsi_div_set_rate(struct clk_hw *hw, | 
|  | unsigned long rate, | 
|  | unsigned long parent_rate) | 
|  | { | 
|  | struct dsi_div_hw_data *dsi_div = to_dsi_div_hw_data(hw); | 
|  | struct rzg2l_cpg_priv *priv = dsi_div->priv; | 
|  |  | 
|  | /* | 
|  | * MUX -->DIV_DSI_{A,B} -->M3 -->VCLK | 
|  | * | 
|  | * Based on the dot clock, the DSI divider clock sets the divider value, | 
|  | * calculates the pll parameters for generating FOUTPOSTDIV and the clk | 
|  | * source for the MUX and propagates that info to the parents. | 
|  | */ | 
|  |  | 
|  | if (!rate || rate > MAX_VCLK_FREQ) | 
|  | return -EINVAL; | 
|  |  | 
|  | dsi_div->rate = rate; | 
|  | writel(CPG_PL5_SDIV_DIV_DSI_A_WEN | CPG_PL5_SDIV_DIV_DSI_B_WEN | | 
|  | (priv->mux_dsi_div_params.dsi_div_a << 0) | | 
|  | (priv->mux_dsi_div_params.dsi_div_b << 8), | 
|  | priv->base + CPG_PL5_SDIV); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static const struct clk_ops rzg2l_cpg_dsi_div_ops = { | 
|  | .recalc_rate = rzg2l_cpg_dsi_div_recalc_rate, | 
|  | .determine_rate = rzg2l_cpg_dsi_div_determine_rate, | 
|  | .set_rate = rzg2l_cpg_dsi_div_set_rate, | 
|  | }; | 
|  |  | 
|  | static struct clk * __init | 
|  | rzg2l_cpg_dsi_div_clk_register(const struct cpg_core_clk *core, | 
|  | struct rzg2l_cpg_priv *priv) | 
|  | { | 
|  | struct dsi_div_hw_data *clk_hw_data; | 
|  | const struct clk *parent; | 
|  | const char *parent_name; | 
|  | struct clk_init_data init; | 
|  | struct clk_hw *clk_hw; | 
|  | int ret; | 
|  |  | 
|  | parent = priv->clks[core->parent]; | 
|  | if (IS_ERR(parent)) | 
|  | return ERR_CAST(parent); | 
|  |  | 
|  | clk_hw_data = devm_kzalloc(priv->dev, sizeof(*clk_hw_data), GFP_KERNEL); | 
|  | if (!clk_hw_data) | 
|  | return ERR_PTR(-ENOMEM); | 
|  |  | 
|  | clk_hw_data->priv = priv; | 
|  |  | 
|  | parent_name = __clk_get_name(parent); | 
|  | init.name = core->name; | 
|  | init.ops = &rzg2l_cpg_dsi_div_ops; | 
|  | init.flags = CLK_SET_RATE_PARENT; | 
|  | init.parent_names = &parent_name; | 
|  | init.num_parents = 1; | 
|  |  | 
|  | clk_hw = &clk_hw_data->hw; | 
|  | clk_hw->init = &init; | 
|  |  | 
|  | ret = devm_clk_hw_register(priv->dev, clk_hw); | 
|  | if (ret) | 
|  | return ERR_PTR(ret); | 
|  |  | 
|  | return clk_hw->clk; | 
|  | } | 
|  |  | 
|  | struct pll5_mux_hw_data { | 
|  | struct clk_hw hw; | 
|  | u32 conf; | 
|  | unsigned long rate; | 
|  | struct rzg2l_cpg_priv *priv; | 
|  | }; | 
|  |  | 
|  | #define to_pll5_mux_hw_data(_hw)	container_of(_hw, struct pll5_mux_hw_data, hw) | 
|  |  | 
|  | static int rzg2l_cpg_pll5_4_clk_mux_determine_rate(struct clk_hw *hw, | 
|  | struct clk_rate_request *req) | 
|  | { | 
|  | struct clk_hw *parent; | 
|  | struct pll5_mux_hw_data *hwdata = to_pll5_mux_hw_data(hw); | 
|  | struct rzg2l_cpg_priv *priv = hwdata->priv; | 
|  |  | 
|  | parent = clk_hw_get_parent_by_index(hw, priv->mux_dsi_div_params.clksrc); | 
|  | req->best_parent_hw = parent; | 
|  | req->best_parent_rate = req->rate; | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int rzg2l_cpg_pll5_4_clk_mux_set_parent(struct clk_hw *hw, u8 index) | 
|  | { | 
|  | struct pll5_mux_hw_data *hwdata = to_pll5_mux_hw_data(hw); | 
|  | struct rzg2l_cpg_priv *priv = hwdata->priv; | 
|  |  | 
|  | /* | 
|  | * FOUTPOSTDIV--->| | 
|  | *  |             | -->MUX -->DIV_DSIA_B -->M3 -->VCLK | 
|  | *  |--FOUT1PH0-->| | 
|  | * | 
|  | * Based on the dot clock, the DSI divider clock calculates the parent | 
|  | * rate and clk source for the MUX. It propagates that info to | 
|  | * pll5_4_clk_mux which sets the clock source for DSI divider clock. | 
|  | */ | 
|  |  | 
|  | writel(CPG_OTHERFUNC1_REG_RES0_ON_WEN | index, | 
|  | priv->base + CPG_OTHERFUNC1_REG); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static u8 rzg2l_cpg_pll5_4_clk_mux_get_parent(struct clk_hw *hw) | 
|  | { | 
|  | struct pll5_mux_hw_data *hwdata = to_pll5_mux_hw_data(hw); | 
|  | struct rzg2l_cpg_priv *priv = hwdata->priv; | 
|  |  | 
|  | return readl(priv->base + GET_REG_OFFSET(hwdata->conf)); | 
|  | } | 
|  |  | 
|  | static const struct clk_ops rzg2l_cpg_pll5_4_clk_mux_ops = { | 
|  | .determine_rate = rzg2l_cpg_pll5_4_clk_mux_determine_rate, | 
|  | .set_parent	= rzg2l_cpg_pll5_4_clk_mux_set_parent, | 
|  | .get_parent	= rzg2l_cpg_pll5_4_clk_mux_get_parent, | 
|  | }; | 
|  |  | 
|  | static struct clk * __init | 
|  | rzg2l_cpg_pll5_4_mux_clk_register(const struct cpg_core_clk *core, | 
|  | struct rzg2l_cpg_priv *priv) | 
|  | { | 
|  | struct pll5_mux_hw_data *clk_hw_data; | 
|  | struct clk_init_data init; | 
|  | struct clk_hw *clk_hw; | 
|  | int ret; | 
|  |  | 
|  | clk_hw_data = devm_kzalloc(priv->dev, sizeof(*clk_hw_data), GFP_KERNEL); | 
|  | if (!clk_hw_data) | 
|  | return ERR_PTR(-ENOMEM); | 
|  |  | 
|  | clk_hw_data->priv = priv; | 
|  | clk_hw_data->conf = core->conf; | 
|  |  | 
|  | init.name = core->name; | 
|  | init.ops = &rzg2l_cpg_pll5_4_clk_mux_ops; | 
|  | init.flags = CLK_SET_RATE_PARENT; | 
|  | init.num_parents = core->num_parents; | 
|  | init.parent_names = core->parent_names; | 
|  |  | 
|  | clk_hw = &clk_hw_data->hw; | 
|  | clk_hw->init = &init; | 
|  |  | 
|  | ret = devm_clk_hw_register(priv->dev, clk_hw); | 
|  | if (ret) | 
|  | return ERR_PTR(ret); | 
|  |  | 
|  | return clk_hw->clk; | 
|  | } | 
|  |  | 
|  | struct sipll5 { | 
|  | struct clk_hw hw; | 
|  | u32 conf; | 
|  | unsigned long foutpostdiv_rate; | 
|  | struct rzg2l_cpg_priv *priv; | 
|  | }; | 
|  |  | 
|  | #define to_sipll5(_hw)	container_of(_hw, struct sipll5, hw) | 
|  |  | 
|  | static unsigned long rzg2l_cpg_get_vclk_rate(struct clk_hw *hw, | 
|  | unsigned long rate) | 
|  | { | 
|  | struct sipll5 *sipll5 = to_sipll5(hw); | 
|  | struct rzg2l_cpg_priv *priv = sipll5->priv; | 
|  | unsigned long vclk; | 
|  |  | 
|  | vclk = rate / ((1 << priv->mux_dsi_div_params.dsi_div_a) * | 
|  | (priv->mux_dsi_div_params.dsi_div_b + 1)); | 
|  |  | 
|  | if (priv->mux_dsi_div_params.clksrc) | 
|  | vclk /= 2; | 
|  |  | 
|  | return vclk; | 
|  | } | 
|  |  | 
|  | static unsigned long rzg2l_cpg_sipll5_recalc_rate(struct clk_hw *hw, | 
|  | unsigned long parent_rate) | 
|  | { | 
|  | struct sipll5 *sipll5 = to_sipll5(hw); | 
|  | unsigned long pll5_rate = sipll5->foutpostdiv_rate; | 
|  |  | 
|  | if (!pll5_rate) | 
|  | pll5_rate = parent_rate; | 
|  |  | 
|  | return pll5_rate; | 
|  | } | 
|  |  | 
|  | static long rzg2l_cpg_sipll5_round_rate(struct clk_hw *hw, | 
|  | unsigned long rate, | 
|  | unsigned long *parent_rate) | 
|  | { | 
|  | return rate; | 
|  | } | 
|  |  | 
|  | static int rzg2l_cpg_sipll5_set_rate(struct clk_hw *hw, | 
|  | unsigned long rate, | 
|  | unsigned long parent_rate) | 
|  | { | 
|  | struct sipll5 *sipll5 = to_sipll5(hw); | 
|  | struct rzg2l_cpg_priv *priv = sipll5->priv; | 
|  | struct rzg2l_pll5_param params; | 
|  | unsigned long vclk_rate; | 
|  | int ret; | 
|  | u32 val; | 
|  |  | 
|  | /* | 
|  | *  OSC --> PLL5 --> FOUTPOSTDIV-->| | 
|  | *                   |             | -->MUX -->DIV_DSIA_B -->M3 -->VCLK | 
|  | *                   |--FOUT1PH0-->| | 
|  | * | 
|  | * Based on the dot clock, the DSI divider clock calculates the parent | 
|  | * rate and the pll5 parameters for generating FOUTPOSTDIV. It propagates | 
|  | * that info to sipll5 which sets parameters for generating FOUTPOSTDIV. | 
|  | * | 
|  | * OSC --> PLL5 --> FOUTPOSTDIV | 
|  | */ | 
|  |  | 
|  | if (!rate) | 
|  | return -EINVAL; | 
|  |  | 
|  | vclk_rate = rzg2l_cpg_get_vclk_rate(hw, rate); | 
|  | sipll5->foutpostdiv_rate = | 
|  | rzg2l_cpg_get_foutpostdiv_rate(¶ms, vclk_rate); | 
|  |  | 
|  | /* Put PLL5 into standby mode */ | 
|  | writel(CPG_SIPLL5_STBY_RESETB_WEN, priv->base + CPG_SIPLL5_STBY); | 
|  | ret = readl_poll_timeout(priv->base + CPG_SIPLL5_MON, val, | 
|  | !(val & CPG_SIPLL5_MON_PLL5_LOCK), 100, 250000); | 
|  | if (ret) { | 
|  | dev_err(priv->dev, "failed to release pll5 lock"); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | /* Output clock setting 1 */ | 
|  | writel((params.pl5_postdiv1 << 0) | (params.pl5_postdiv2 << 4) | | 
|  | (params.pl5_refdiv << 8), priv->base + CPG_SIPLL5_CLK1); | 
|  |  | 
|  | /* Output clock setting, SSCG modulation value setting 3 */ | 
|  | writel((params.pl5_fracin << 8), priv->base + CPG_SIPLL5_CLK3); | 
|  |  | 
|  | /* Output clock setting 4 */ | 
|  | writel(CPG_SIPLL5_CLK4_RESV_LSB | (params.pl5_intin << 16), | 
|  | priv->base + CPG_SIPLL5_CLK4); | 
|  |  | 
|  | /* Output clock setting 5 */ | 
|  | writel(params.pl5_spread, priv->base + CPG_SIPLL5_CLK5); | 
|  |  | 
|  | /* PLL normal mode setting */ | 
|  | writel(CPG_SIPLL5_STBY_DOWNSPREAD_WEN | CPG_SIPLL5_STBY_SSCG_EN_WEN | | 
|  | CPG_SIPLL5_STBY_RESETB_WEN | CPG_SIPLL5_STBY_RESETB, | 
|  | priv->base + CPG_SIPLL5_STBY); | 
|  |  | 
|  | /* PLL normal mode transition, output clock stability check */ | 
|  | ret = readl_poll_timeout(priv->base + CPG_SIPLL5_MON, val, | 
|  | (val & CPG_SIPLL5_MON_PLL5_LOCK), 100, 250000); | 
|  | if (ret) { | 
|  | dev_err(priv->dev, "failed to lock pll5"); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static const struct clk_ops rzg2l_cpg_sipll5_ops = { | 
|  | .recalc_rate = rzg2l_cpg_sipll5_recalc_rate, | 
|  | .round_rate = rzg2l_cpg_sipll5_round_rate, | 
|  | .set_rate = rzg2l_cpg_sipll5_set_rate, | 
|  | }; | 
|  |  | 
|  | static struct clk * __init | 
|  | rzg2l_cpg_sipll5_register(const struct cpg_core_clk *core, | 
|  | struct rzg2l_cpg_priv *priv) | 
|  | { | 
|  | const struct clk *parent; | 
|  | struct clk_init_data init; | 
|  | const char *parent_name; | 
|  | struct sipll5 *sipll5; | 
|  | struct clk_hw *clk_hw; | 
|  | int ret; | 
|  |  | 
|  | parent = priv->clks[core->parent]; | 
|  | if (IS_ERR(parent)) | 
|  | return ERR_CAST(parent); | 
|  |  | 
|  | sipll5 = devm_kzalloc(priv->dev, sizeof(*sipll5), GFP_KERNEL); | 
|  | if (!sipll5) | 
|  | return ERR_PTR(-ENOMEM); | 
|  |  | 
|  | init.name = core->name; | 
|  | parent_name = __clk_get_name(parent); | 
|  | init.ops = &rzg2l_cpg_sipll5_ops; | 
|  | init.flags = 0; | 
|  | init.parent_names = &parent_name; | 
|  | init.num_parents = 1; | 
|  |  | 
|  | sipll5->hw.init = &init; | 
|  | sipll5->conf = core->conf; | 
|  | sipll5->priv = priv; | 
|  |  | 
|  | writel(CPG_SIPLL5_STBY_SSCG_EN_WEN | CPG_SIPLL5_STBY_RESETB_WEN | | 
|  | CPG_SIPLL5_STBY_RESETB, priv->base + CPG_SIPLL5_STBY); | 
|  |  | 
|  | clk_hw = &sipll5->hw; | 
|  | clk_hw->init = &init; | 
|  |  | 
|  | ret = devm_clk_hw_register(priv->dev, clk_hw); | 
|  | if (ret) | 
|  | return ERR_PTR(ret); | 
|  |  | 
|  | priv->mux_dsi_div_params.clksrc = 1; /* Use clk src 1 for DSI */ | 
|  | priv->mux_dsi_div_params.dsi_div_a = 1; /* Divided by 2 */ | 
|  | priv->mux_dsi_div_params.dsi_div_b = 2; /* Divided by 3 */ | 
|  |  | 
|  | return clk_hw->clk; | 
|  | } | 
|  |  | 
|  | struct pll_clk { | 
|  | struct clk_hw hw; | 
|  | unsigned int conf; | 
|  | unsigned int type; | 
|  | void __iomem *base; | 
|  | struct rzg2l_cpg_priv *priv; | 
|  | }; | 
|  |  | 
|  | #define to_pll(_hw)	container_of(_hw, struct pll_clk, hw) | 
|  |  | 
|  | static unsigned long rzg2l_cpg_pll_clk_recalc_rate(struct clk_hw *hw, | 
|  | unsigned long parent_rate) | 
|  | { | 
|  | struct pll_clk *pll_clk = to_pll(hw); | 
|  | struct rzg2l_cpg_priv *priv = pll_clk->priv; | 
|  | unsigned int val1, val2; | 
|  | u64 rate; | 
|  |  | 
|  | if (pll_clk->type != CLK_TYPE_SAM_PLL) | 
|  | return parent_rate; | 
|  |  | 
|  | val1 = readl(priv->base + GET_REG_SAMPLL_CLK1(pll_clk->conf)); | 
|  | val2 = readl(priv->base + GET_REG_SAMPLL_CLK2(pll_clk->conf)); | 
|  |  | 
|  | rate = mul_u64_u32_shr(parent_rate, (MDIV(val1) << 16) + KDIV(val1), | 
|  | 16 + SDIV(val2)); | 
|  |  | 
|  | return DIV_ROUND_CLOSEST_ULL(rate, PDIV(val1)); | 
|  | } | 
|  |  | 
|  | static const struct clk_ops rzg2l_cpg_pll_ops = { | 
|  | .recalc_rate = rzg2l_cpg_pll_clk_recalc_rate, | 
|  | }; | 
|  |  | 
|  | static unsigned long rzg3s_cpg_pll_clk_recalc_rate(struct clk_hw *hw, | 
|  | unsigned long parent_rate) | 
|  | { | 
|  | struct pll_clk *pll_clk = to_pll(hw); | 
|  | struct rzg2l_cpg_priv *priv = pll_clk->priv; | 
|  | u32 nir, nfr, mr, pr, val; | 
|  | u64 rate; | 
|  |  | 
|  | if (pll_clk->type != CLK_TYPE_G3S_PLL) | 
|  | return parent_rate; | 
|  |  | 
|  | val = readl(priv->base + GET_REG_SAMPLL_CLK1(pll_clk->conf)); | 
|  |  | 
|  | pr = 1 << FIELD_GET(RZG3S_DIV_P, val); | 
|  | /* Hardware interprets values higher than 8 as p = 16. */ | 
|  | if (pr > 8) | 
|  | pr = 16; | 
|  |  | 
|  | mr  = FIELD_GET(RZG3S_DIV_M, val) + 1; | 
|  | nir = FIELD_GET(RZG3S_DIV_NI, val) + 1; | 
|  | nfr = FIELD_GET(RZG3S_DIV_NF, val); | 
|  |  | 
|  | rate = mul_u64_u32_shr(parent_rate, 4096 * nir + nfr, 12); | 
|  |  | 
|  | return DIV_ROUND_CLOSEST_ULL(rate, (mr * pr)); | 
|  | } | 
|  |  | 
|  | static const struct clk_ops rzg3s_cpg_pll_ops = { | 
|  | .recalc_rate = rzg3s_cpg_pll_clk_recalc_rate, | 
|  | }; | 
|  |  | 
|  | static struct clk * __init | 
|  | rzg2l_cpg_pll_clk_register(const struct cpg_core_clk *core, | 
|  | struct rzg2l_cpg_priv *priv, | 
|  | const struct clk_ops *ops) | 
|  | { | 
|  | struct device *dev = priv->dev; | 
|  | const struct clk *parent; | 
|  | struct clk_init_data init; | 
|  | const char *parent_name; | 
|  | struct pll_clk *pll_clk; | 
|  | int ret; | 
|  |  | 
|  | parent = priv->clks[core->parent]; | 
|  | if (IS_ERR(parent)) | 
|  | return ERR_CAST(parent); | 
|  |  | 
|  | pll_clk = devm_kzalloc(dev, sizeof(*pll_clk), GFP_KERNEL); | 
|  | if (!pll_clk) | 
|  | return ERR_PTR(-ENOMEM); | 
|  |  | 
|  | parent_name = __clk_get_name(parent); | 
|  | init.name = core->name; | 
|  | init.ops = ops; | 
|  | init.flags = 0; | 
|  | init.parent_names = &parent_name; | 
|  | init.num_parents = 1; | 
|  |  | 
|  | pll_clk->hw.init = &init; | 
|  | pll_clk->conf = core->conf; | 
|  | pll_clk->base = priv->base; | 
|  | pll_clk->priv = priv; | 
|  | pll_clk->type = core->type; | 
|  |  | 
|  | ret = devm_clk_hw_register(dev, &pll_clk->hw); | 
|  | if (ret) | 
|  | return ERR_PTR(ret); | 
|  |  | 
|  | return pll_clk->hw.clk; | 
|  | } | 
|  |  | 
|  | static struct clk | 
|  | *rzg2l_cpg_clk_src_twocell_get(struct of_phandle_args *clkspec, | 
|  | void *data) | 
|  | { | 
|  | unsigned int clkidx = clkspec->args[1]; | 
|  | struct rzg2l_cpg_priv *priv = data; | 
|  | struct device *dev = priv->dev; | 
|  | const char *type; | 
|  | struct clk *clk; | 
|  |  | 
|  | switch (clkspec->args[0]) { | 
|  | case CPG_CORE: | 
|  | type = "core"; | 
|  | if (clkidx > priv->last_dt_core_clk) { | 
|  | dev_err(dev, "Invalid %s clock index %u\n", type, clkidx); | 
|  | return ERR_PTR(-EINVAL); | 
|  | } | 
|  | clk = priv->clks[clkidx]; | 
|  | break; | 
|  |  | 
|  | case CPG_MOD: | 
|  | type = "module"; | 
|  | if (clkidx >= priv->num_mod_clks) { | 
|  | dev_err(dev, "Invalid %s clock index %u\n", type, | 
|  | clkidx); | 
|  | return ERR_PTR(-EINVAL); | 
|  | } | 
|  | clk = priv->clks[priv->num_core_clks + clkidx]; | 
|  | break; | 
|  |  | 
|  | default: | 
|  | dev_err(dev, "Invalid CPG clock type %u\n", clkspec->args[0]); | 
|  | return ERR_PTR(-EINVAL); | 
|  | } | 
|  |  | 
|  | if (IS_ERR(clk)) | 
|  | dev_err(dev, "Cannot get %s clock %u: %ld", type, clkidx, | 
|  | PTR_ERR(clk)); | 
|  | else | 
|  | dev_dbg(dev, "clock (%u, %u) is %pC at %lu Hz\n", | 
|  | clkspec->args[0], clkspec->args[1], clk, | 
|  | clk_get_rate(clk)); | 
|  | return clk; | 
|  | } | 
|  |  | 
|  | static void __init | 
|  | rzg2l_cpg_register_core_clk(const struct cpg_core_clk *core, | 
|  | const struct rzg2l_cpg_info *info, | 
|  | struct rzg2l_cpg_priv *priv) | 
|  | { | 
|  | struct clk *clk = ERR_PTR(-EOPNOTSUPP), *parent; | 
|  | struct device *dev = priv->dev; | 
|  | unsigned int id = core->id, div = core->div; | 
|  | const char *parent_name; | 
|  | struct clk_hw *clk_hw; | 
|  |  | 
|  | WARN_DEBUG(id >= priv->num_core_clks); | 
|  | WARN_DEBUG(PTR_ERR(priv->clks[id]) != -ENOENT); | 
|  |  | 
|  | if (!core->name) { | 
|  | /* Skip NULLified clock */ | 
|  | return; | 
|  | } | 
|  |  | 
|  | switch (core->type) { | 
|  | case CLK_TYPE_IN: | 
|  | clk = of_clk_get_by_name(priv->dev->of_node, core->name); | 
|  | break; | 
|  | case CLK_TYPE_FF: | 
|  | WARN_DEBUG(core->parent >= priv->num_core_clks); | 
|  | parent = priv->clks[core->parent]; | 
|  | if (IS_ERR(parent)) { | 
|  | clk = parent; | 
|  | goto fail; | 
|  | } | 
|  |  | 
|  | parent_name = __clk_get_name(parent); | 
|  | clk_hw = devm_clk_hw_register_fixed_factor(dev, core->name, parent_name, | 
|  | CLK_SET_RATE_PARENT, | 
|  | core->mult, div); | 
|  | if (IS_ERR(clk_hw)) | 
|  | clk = ERR_CAST(clk_hw); | 
|  | else | 
|  | clk = clk_hw->clk; | 
|  | break; | 
|  | case CLK_TYPE_SAM_PLL: | 
|  | clk = rzg2l_cpg_pll_clk_register(core, priv, &rzg2l_cpg_pll_ops); | 
|  | break; | 
|  | case CLK_TYPE_G3S_PLL: | 
|  | clk = rzg2l_cpg_pll_clk_register(core, priv, &rzg3s_cpg_pll_ops); | 
|  | break; | 
|  | case CLK_TYPE_SIPLL5: | 
|  | clk = rzg2l_cpg_sipll5_register(core, priv); | 
|  | break; | 
|  | case CLK_TYPE_DIV: | 
|  | clk = rzg2l_cpg_div_clk_register(core, priv); | 
|  | break; | 
|  | case CLK_TYPE_G3S_DIV: | 
|  | clk = rzg3s_cpg_div_clk_register(core, priv); | 
|  | break; | 
|  | case CLK_TYPE_MUX: | 
|  | clk = rzg2l_cpg_mux_clk_register(core, priv); | 
|  | break; | 
|  | case CLK_TYPE_SD_MUX: | 
|  | clk = rzg2l_cpg_sd_mux_clk_register(core, priv); | 
|  | break; | 
|  | case CLK_TYPE_PLL5_4_MUX: | 
|  | clk = rzg2l_cpg_pll5_4_mux_clk_register(core, priv); | 
|  | break; | 
|  | case CLK_TYPE_DSI_DIV: | 
|  | clk = rzg2l_cpg_dsi_div_clk_register(core, priv); | 
|  | break; | 
|  | default: | 
|  | goto fail; | 
|  | } | 
|  |  | 
|  | if (IS_ERR_OR_NULL(clk)) | 
|  | goto fail; | 
|  |  | 
|  | dev_dbg(dev, "Core clock %pC at %lu Hz\n", clk, clk_get_rate(clk)); | 
|  | priv->clks[id] = clk; | 
|  | return; | 
|  |  | 
|  | fail: | 
|  | dev_err(dev, "Failed to register %s clock %s: %ld\n", "core", | 
|  | core->name, PTR_ERR(clk)); | 
|  | } | 
|  |  | 
|  | /** | 
|  | * struct mstp_clock - MSTP gating clock | 
|  | * | 
|  | * @hw: handle between common and hardware-specific interfaces | 
|  | * @off: register offset | 
|  | * @bit: ON/MON bit | 
|  | * @enabled: soft state of the clock, if it is coupled with another clock | 
|  | * @priv: CPG/MSTP private data | 
|  | * @sibling: pointer to the other coupled clock | 
|  | */ | 
|  | struct mstp_clock { | 
|  | struct clk_hw hw; | 
|  | u16 off; | 
|  | u8 bit; | 
|  | bool enabled; | 
|  | struct rzg2l_cpg_priv *priv; | 
|  | struct mstp_clock *sibling; | 
|  | }; | 
|  |  | 
|  | #define to_mod_clock(_hw) container_of(_hw, struct mstp_clock, hw) | 
|  |  | 
|  | static int rzg2l_mod_clock_endisable(struct clk_hw *hw, bool enable) | 
|  | { | 
|  | struct mstp_clock *clock = to_mod_clock(hw); | 
|  | struct rzg2l_cpg_priv *priv = clock->priv; | 
|  | unsigned int reg = clock->off; | 
|  | struct device *dev = priv->dev; | 
|  | u32 bitmask = BIT(clock->bit); | 
|  | u32 value; | 
|  | int error; | 
|  |  | 
|  | if (!clock->off) { | 
|  | dev_dbg(dev, "%pC does not support ON/OFF\n",  hw->clk); | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | dev_dbg(dev, "CLK_ON 0x%x/%pC %s\n", CLK_ON_R(reg), hw->clk, | 
|  | enable ? "ON" : "OFF"); | 
|  |  | 
|  | value = bitmask << 16; | 
|  | if (enable) | 
|  | value |= bitmask; | 
|  |  | 
|  | writel(value, priv->base + CLK_ON_R(reg)); | 
|  |  | 
|  | if (!enable) | 
|  | return 0; | 
|  |  | 
|  | if (!priv->info->has_clk_mon_regs) | 
|  | return 0; | 
|  |  | 
|  | error = readl_poll_timeout_atomic(priv->base + CLK_MON_R(reg), value, | 
|  | value & bitmask, 0, 10); | 
|  | if (error) | 
|  | dev_err(dev, "Failed to enable CLK_ON %p\n", | 
|  | priv->base + CLK_ON_R(reg)); | 
|  |  | 
|  | return error; | 
|  | } | 
|  |  | 
|  | static int rzg2l_mod_clock_enable(struct clk_hw *hw) | 
|  | { | 
|  | struct mstp_clock *clock = to_mod_clock(hw); | 
|  |  | 
|  | if (clock->sibling) { | 
|  | struct rzg2l_cpg_priv *priv = clock->priv; | 
|  | unsigned long flags; | 
|  | bool enabled; | 
|  |  | 
|  | spin_lock_irqsave(&priv->rmw_lock, flags); | 
|  | enabled = clock->sibling->enabled; | 
|  | clock->enabled = true; | 
|  | spin_unlock_irqrestore(&priv->rmw_lock, flags); | 
|  | if (enabled) | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | return rzg2l_mod_clock_endisable(hw, true); | 
|  | } | 
|  |  | 
|  | static void rzg2l_mod_clock_disable(struct clk_hw *hw) | 
|  | { | 
|  | struct mstp_clock *clock = to_mod_clock(hw); | 
|  |  | 
|  | if (clock->sibling) { | 
|  | struct rzg2l_cpg_priv *priv = clock->priv; | 
|  | unsigned long flags; | 
|  | bool enabled; | 
|  |  | 
|  | spin_lock_irqsave(&priv->rmw_lock, flags); | 
|  | enabled = clock->sibling->enabled; | 
|  | clock->enabled = false; | 
|  | spin_unlock_irqrestore(&priv->rmw_lock, flags); | 
|  | if (enabled) | 
|  | return; | 
|  | } | 
|  |  | 
|  | rzg2l_mod_clock_endisable(hw, false); | 
|  | } | 
|  |  | 
|  | static int rzg2l_mod_clock_is_enabled(struct clk_hw *hw) | 
|  | { | 
|  | struct mstp_clock *clock = to_mod_clock(hw); | 
|  | struct rzg2l_cpg_priv *priv = clock->priv; | 
|  | u32 bitmask = BIT(clock->bit); | 
|  | u32 value; | 
|  |  | 
|  | if (!clock->off) { | 
|  | dev_dbg(priv->dev, "%pC does not support ON/OFF\n",  hw->clk); | 
|  | return 1; | 
|  | } | 
|  |  | 
|  | if (clock->sibling) | 
|  | return clock->enabled; | 
|  |  | 
|  | if (priv->info->has_clk_mon_regs) | 
|  | value = readl(priv->base + CLK_MON_R(clock->off)); | 
|  | else | 
|  | value = readl(priv->base + clock->off); | 
|  |  | 
|  | return value & bitmask; | 
|  | } | 
|  |  | 
|  | static const struct clk_ops rzg2l_mod_clock_ops = { | 
|  | .enable = rzg2l_mod_clock_enable, | 
|  | .disable = rzg2l_mod_clock_disable, | 
|  | .is_enabled = rzg2l_mod_clock_is_enabled, | 
|  | }; | 
|  |  | 
|  | static struct mstp_clock | 
|  | *rzg2l_mod_clock_get_sibling(struct mstp_clock *clock, | 
|  | struct rzg2l_cpg_priv *priv) | 
|  | { | 
|  | struct clk_hw *hw; | 
|  | unsigned int i; | 
|  |  | 
|  | for (i = 0; i < priv->num_mod_clks; i++) { | 
|  | struct mstp_clock *clk; | 
|  |  | 
|  | if (priv->clks[priv->num_core_clks + i] == ERR_PTR(-ENOENT)) | 
|  | continue; | 
|  |  | 
|  | hw = __clk_get_hw(priv->clks[priv->num_core_clks + i]); | 
|  | clk = to_mod_clock(hw); | 
|  | if (clock->off == clk->off && clock->bit == clk->bit) | 
|  | return clk; | 
|  | } | 
|  |  | 
|  | return NULL; | 
|  | } | 
|  |  | 
|  | static void __init | 
|  | rzg2l_cpg_register_mod_clk(const struct rzg2l_mod_clk *mod, | 
|  | const struct rzg2l_cpg_info *info, | 
|  | struct rzg2l_cpg_priv *priv) | 
|  | { | 
|  | struct mstp_clock *clock = NULL; | 
|  | struct device *dev = priv->dev; | 
|  | unsigned int id = mod->id; | 
|  | struct clk_init_data init; | 
|  | struct clk *parent, *clk; | 
|  | const char *parent_name; | 
|  | unsigned int i; | 
|  | int ret; | 
|  |  | 
|  | WARN_DEBUG(id < priv->num_core_clks); | 
|  | WARN_DEBUG(id >= priv->num_core_clks + priv->num_mod_clks); | 
|  | WARN_DEBUG(mod->parent >= priv->num_core_clks + priv->num_mod_clks); | 
|  | WARN_DEBUG(PTR_ERR(priv->clks[id]) != -ENOENT); | 
|  |  | 
|  | if (!mod->name) { | 
|  | /* Skip NULLified clock */ | 
|  | return; | 
|  | } | 
|  |  | 
|  | parent = priv->clks[mod->parent]; | 
|  | if (IS_ERR(parent)) { | 
|  | clk = parent; | 
|  | goto fail; | 
|  | } | 
|  |  | 
|  | clock = devm_kzalloc(dev, sizeof(*clock), GFP_KERNEL); | 
|  | if (!clock) { | 
|  | clk = ERR_PTR(-ENOMEM); | 
|  | goto fail; | 
|  | } | 
|  |  | 
|  | init.name = mod->name; | 
|  | init.ops = &rzg2l_mod_clock_ops; | 
|  | init.flags = CLK_SET_RATE_PARENT; | 
|  | for (i = 0; i < info->num_crit_mod_clks; i++) | 
|  | if (id == info->crit_mod_clks[i]) { | 
|  | dev_dbg(dev, "CPG %s setting CLK_IS_CRITICAL\n", | 
|  | mod->name); | 
|  | init.flags |= CLK_IS_CRITICAL; | 
|  | break; | 
|  | } | 
|  |  | 
|  | parent_name = __clk_get_name(parent); | 
|  | init.parent_names = &parent_name; | 
|  | init.num_parents = 1; | 
|  |  | 
|  | clock->off = mod->off; | 
|  | clock->bit = mod->bit; | 
|  | clock->priv = priv; | 
|  | clock->hw.init = &init; | 
|  |  | 
|  | ret = devm_clk_hw_register(dev, &clock->hw); | 
|  | if (ret) { | 
|  | clk = ERR_PTR(ret); | 
|  | goto fail; | 
|  | } | 
|  |  | 
|  | clk = clock->hw.clk; | 
|  | dev_dbg(dev, "Module clock %pC at %lu Hz\n", clk, clk_get_rate(clk)); | 
|  | priv->clks[id] = clk; | 
|  |  | 
|  | if (mod->is_coupled) { | 
|  | struct mstp_clock *sibling; | 
|  |  | 
|  | clock->enabled = rzg2l_mod_clock_is_enabled(&clock->hw); | 
|  | sibling = rzg2l_mod_clock_get_sibling(clock, priv); | 
|  | if (sibling) { | 
|  | clock->sibling = sibling; | 
|  | sibling->sibling = clock; | 
|  | } | 
|  | } | 
|  |  | 
|  | return; | 
|  |  | 
|  | fail: | 
|  | dev_err(dev, "Failed to register %s clock %s: %ld\n", "module", | 
|  | mod->name, PTR_ERR(clk)); | 
|  | } | 
|  |  | 
|  | #define rcdev_to_priv(x)	container_of(x, struct rzg2l_cpg_priv, rcdev) | 
|  |  | 
|  | static int rzg2l_cpg_assert(struct reset_controller_dev *rcdev, | 
|  | unsigned long id) | 
|  | { | 
|  | struct rzg2l_cpg_priv *priv = rcdev_to_priv(rcdev); | 
|  | const struct rzg2l_cpg_info *info = priv->info; | 
|  | unsigned int reg = info->resets[id].off; | 
|  | u32 mask = BIT(info->resets[id].bit); | 
|  | s8 monbit = info->resets[id].monbit; | 
|  | u32 value = mask << 16; | 
|  |  | 
|  | dev_dbg(rcdev->dev, "assert id:%ld offset:0x%x\n", id, CLK_RST_R(reg)); | 
|  |  | 
|  | writel(value, priv->base + CLK_RST_R(reg)); | 
|  |  | 
|  | if (info->has_clk_mon_regs) { | 
|  | reg = CLK_MRST_R(reg); | 
|  | } else if (monbit >= 0) { | 
|  | reg = CPG_RST_MON; | 
|  | mask = BIT(monbit); | 
|  | } else { | 
|  | /* Wait for at least one cycle of the RCLK clock (@ ca. 32 kHz) */ | 
|  | udelay(35); | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | return readl_poll_timeout_atomic(priv->base + reg, value, | 
|  | value & mask, 10, 200); | 
|  | } | 
|  |  | 
|  | static int rzg2l_cpg_deassert(struct reset_controller_dev *rcdev, | 
|  | unsigned long id) | 
|  | { | 
|  | struct rzg2l_cpg_priv *priv = rcdev_to_priv(rcdev); | 
|  | const struct rzg2l_cpg_info *info = priv->info; | 
|  | unsigned int reg = info->resets[id].off; | 
|  | u32 mask = BIT(info->resets[id].bit); | 
|  | s8 monbit = info->resets[id].monbit; | 
|  | u32 value = (mask << 16) | mask; | 
|  |  | 
|  | dev_dbg(rcdev->dev, "deassert id:%ld offset:0x%x\n", id, | 
|  | CLK_RST_R(reg)); | 
|  |  | 
|  | writel(value, priv->base + CLK_RST_R(reg)); | 
|  |  | 
|  | if (info->has_clk_mon_regs) { | 
|  | reg = CLK_MRST_R(reg); | 
|  | } else if (monbit >= 0) { | 
|  | reg = CPG_RST_MON; | 
|  | mask = BIT(monbit); | 
|  | } else { | 
|  | /* Wait for at least one cycle of the RCLK clock (@ ca. 32 kHz) */ | 
|  | udelay(35); | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | return readl_poll_timeout_atomic(priv->base + reg, value, | 
|  | !(value & mask), 10, 200); | 
|  | } | 
|  |  | 
|  | static int rzg2l_cpg_reset(struct reset_controller_dev *rcdev, | 
|  | unsigned long id) | 
|  | { | 
|  | int ret; | 
|  |  | 
|  | ret = rzg2l_cpg_assert(rcdev, id); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | return rzg2l_cpg_deassert(rcdev, id); | 
|  | } | 
|  |  | 
|  | static int rzg2l_cpg_status(struct reset_controller_dev *rcdev, | 
|  | unsigned long id) | 
|  | { | 
|  | struct rzg2l_cpg_priv *priv = rcdev_to_priv(rcdev); | 
|  | const struct rzg2l_cpg_info *info = priv->info; | 
|  | s8 monbit = info->resets[id].monbit; | 
|  | unsigned int reg; | 
|  | u32 bitmask; | 
|  |  | 
|  | if (info->has_clk_mon_regs) { | 
|  | reg = CLK_MRST_R(info->resets[id].off); | 
|  | bitmask = BIT(info->resets[id].bit); | 
|  | } else if (monbit >= 0) { | 
|  | reg = CPG_RST_MON; | 
|  | bitmask = BIT(monbit); | 
|  | } else { | 
|  | return -ENOTSUPP; | 
|  | } | 
|  |  | 
|  | return !!(readl(priv->base + reg) & bitmask); | 
|  | } | 
|  |  | 
|  | static const struct reset_control_ops rzg2l_cpg_reset_ops = { | 
|  | .reset = rzg2l_cpg_reset, | 
|  | .assert = rzg2l_cpg_assert, | 
|  | .deassert = rzg2l_cpg_deassert, | 
|  | .status = rzg2l_cpg_status, | 
|  | }; | 
|  |  | 
|  | static int rzg2l_cpg_reset_xlate(struct reset_controller_dev *rcdev, | 
|  | const struct of_phandle_args *reset_spec) | 
|  | { | 
|  | struct rzg2l_cpg_priv *priv = rcdev_to_priv(rcdev); | 
|  | const struct rzg2l_cpg_info *info = priv->info; | 
|  | unsigned int id = reset_spec->args[0]; | 
|  |  | 
|  | if (id >= rcdev->nr_resets || !info->resets[id].off) { | 
|  | dev_err(rcdev->dev, "Invalid reset index %u\n", id); | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | return id; | 
|  | } | 
|  |  | 
|  | static int rzg2l_cpg_reset_controller_register(struct rzg2l_cpg_priv *priv) | 
|  | { | 
|  | priv->rcdev.ops = &rzg2l_cpg_reset_ops; | 
|  | priv->rcdev.of_node = priv->dev->of_node; | 
|  | priv->rcdev.dev = priv->dev; | 
|  | priv->rcdev.of_reset_n_cells = 1; | 
|  | priv->rcdev.of_xlate = rzg2l_cpg_reset_xlate; | 
|  | priv->rcdev.nr_resets = priv->num_resets; | 
|  |  | 
|  | return devm_reset_controller_register(priv->dev, &priv->rcdev); | 
|  | } | 
|  |  | 
|  | static bool rzg2l_cpg_is_pm_clk(struct rzg2l_cpg_priv *priv, | 
|  | const struct of_phandle_args *clkspec) | 
|  | { | 
|  | const struct rzg2l_cpg_info *info = priv->info; | 
|  | unsigned int id; | 
|  | unsigned int i; | 
|  |  | 
|  | if (clkspec->args_count != 2) | 
|  | return false; | 
|  |  | 
|  | if (clkspec->args[0] != CPG_MOD) | 
|  | return false; | 
|  |  | 
|  | id = clkspec->args[1] + info->num_total_core_clks; | 
|  | for (i = 0; i < info->num_no_pm_mod_clks; i++) { | 
|  | if (info->no_pm_mod_clks[i] == id) | 
|  | return false; | 
|  | } | 
|  |  | 
|  | return true; | 
|  | } | 
|  |  | 
|  | /** | 
|  | * struct rzg2l_cpg_pm_domains - RZ/G2L PM domains data structure | 
|  | * @onecell_data: cell data | 
|  | * @domains: generic PM domains | 
|  | */ | 
|  | struct rzg2l_cpg_pm_domains { | 
|  | struct genpd_onecell_data onecell_data; | 
|  | struct generic_pm_domain *domains[]; | 
|  | }; | 
|  |  | 
|  | /** | 
|  | * struct rzg2l_cpg_pd - RZ/G2L power domain data structure | 
|  | * @genpd: generic PM domain | 
|  | * @priv: pointer to CPG private data structure | 
|  | * @conf: CPG PM domain configuration info | 
|  | * @id: RZ/G2L power domain ID | 
|  | */ | 
|  | struct rzg2l_cpg_pd { | 
|  | struct generic_pm_domain genpd; | 
|  | struct rzg2l_cpg_priv *priv; | 
|  | struct rzg2l_cpg_pm_domain_conf conf; | 
|  | u16 id; | 
|  | }; | 
|  |  | 
|  | static int rzg2l_cpg_attach_dev(struct generic_pm_domain *domain, struct device *dev) | 
|  | { | 
|  | struct rzg2l_cpg_pd *pd = container_of(domain, struct rzg2l_cpg_pd, genpd); | 
|  | struct rzg2l_cpg_priv *priv = pd->priv; | 
|  | struct device_node *np = dev->of_node; | 
|  | struct of_phandle_args clkspec; | 
|  | bool once = true; | 
|  | struct clk *clk; | 
|  | int error; | 
|  | int i = 0; | 
|  |  | 
|  | while (!of_parse_phandle_with_args(np, "clocks", "#clock-cells", i, | 
|  | &clkspec)) { | 
|  | if (rzg2l_cpg_is_pm_clk(priv, &clkspec)) { | 
|  | if (once) { | 
|  | once = false; | 
|  | error = pm_clk_create(dev); | 
|  | if (error) { | 
|  | of_node_put(clkspec.np); | 
|  | goto err; | 
|  | } | 
|  | } | 
|  | clk = of_clk_get_from_provider(&clkspec); | 
|  | of_node_put(clkspec.np); | 
|  | if (IS_ERR(clk)) { | 
|  | error = PTR_ERR(clk); | 
|  | goto fail_destroy; | 
|  | } | 
|  |  | 
|  | error = pm_clk_add_clk(dev, clk); | 
|  | if (error) { | 
|  | dev_err(dev, "pm_clk_add_clk failed %d\n", | 
|  | error); | 
|  | goto fail_put; | 
|  | } | 
|  | } else { | 
|  | of_node_put(clkspec.np); | 
|  | } | 
|  | i++; | 
|  | } | 
|  |  | 
|  | return 0; | 
|  |  | 
|  | fail_put: | 
|  | clk_put(clk); | 
|  |  | 
|  | fail_destroy: | 
|  | pm_clk_destroy(dev); | 
|  | err: | 
|  | return error; | 
|  | } | 
|  |  | 
|  | static void rzg2l_cpg_detach_dev(struct generic_pm_domain *unused, struct device *dev) | 
|  | { | 
|  | if (!pm_clk_no_clocks(dev)) | 
|  | pm_clk_destroy(dev); | 
|  | } | 
|  |  | 
|  | static void rzg2l_cpg_genpd_remove(void *data) | 
|  | { | 
|  | struct genpd_onecell_data *celldata = data; | 
|  |  | 
|  | for (unsigned int i = 0; i < celldata->num_domains; i++) | 
|  | pm_genpd_remove(celldata->domains[i]); | 
|  | } | 
|  |  | 
|  | static void rzg2l_cpg_genpd_remove_simple(void *data) | 
|  | { | 
|  | pm_genpd_remove(data); | 
|  | } | 
|  |  | 
|  | static int rzg2l_cpg_power_on(struct generic_pm_domain *domain) | 
|  | { | 
|  | struct rzg2l_cpg_pd *pd = container_of(domain, struct rzg2l_cpg_pd, genpd); | 
|  | struct rzg2l_cpg_reg_conf mstop = pd->conf.mstop; | 
|  | struct rzg2l_cpg_priv *priv = pd->priv; | 
|  |  | 
|  | /* Set MSTOP. */ | 
|  | if (mstop.mask) | 
|  | writel(mstop.mask << 16, priv->base + mstop.off); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int rzg2l_cpg_power_off(struct generic_pm_domain *domain) | 
|  | { | 
|  | struct rzg2l_cpg_pd *pd = container_of(domain, struct rzg2l_cpg_pd, genpd); | 
|  | struct rzg2l_cpg_reg_conf mstop = pd->conf.mstop; | 
|  | struct rzg2l_cpg_priv *priv = pd->priv; | 
|  |  | 
|  | /* Set MSTOP. */ | 
|  | if (mstop.mask) | 
|  | writel(mstop.mask | (mstop.mask << 16), priv->base + mstop.off); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int __init rzg2l_cpg_pd_setup(struct rzg2l_cpg_pd *pd, bool always_on) | 
|  | { | 
|  | struct dev_power_governor *governor; | 
|  |  | 
|  | pd->genpd.flags |= GENPD_FLAG_PM_CLK | GENPD_FLAG_ACTIVE_WAKEUP; | 
|  | pd->genpd.attach_dev = rzg2l_cpg_attach_dev; | 
|  | pd->genpd.detach_dev = rzg2l_cpg_detach_dev; | 
|  | if (always_on) { | 
|  | pd->genpd.flags |= GENPD_FLAG_ALWAYS_ON; | 
|  | governor = &pm_domain_always_on_gov; | 
|  | } else { | 
|  | pd->genpd.power_on = rzg2l_cpg_power_on; | 
|  | pd->genpd.power_off = rzg2l_cpg_power_off; | 
|  | governor = &simple_qos_governor; | 
|  | } | 
|  |  | 
|  | return pm_genpd_init(&pd->genpd, governor, !always_on); | 
|  | } | 
|  |  | 
|  | static int __init rzg2l_cpg_add_clk_domain(struct rzg2l_cpg_priv *priv) | 
|  | { | 
|  | struct device *dev = priv->dev; | 
|  | struct device_node *np = dev->of_node; | 
|  | struct rzg2l_cpg_pd *pd; | 
|  | int ret; | 
|  |  | 
|  | pd = devm_kzalloc(dev, sizeof(*pd), GFP_KERNEL); | 
|  | if (!pd) | 
|  | return -ENOMEM; | 
|  |  | 
|  | pd->genpd.name = np->name; | 
|  | pd->priv = priv; | 
|  | ret = rzg2l_cpg_pd_setup(pd, true); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | ret = devm_add_action_or_reset(dev, rzg2l_cpg_genpd_remove_simple, &pd->genpd); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | return of_genpd_add_provider_simple(np, &pd->genpd); | 
|  | } | 
|  |  | 
|  | static struct generic_pm_domain * | 
|  | rzg2l_cpg_pm_domain_xlate(const struct of_phandle_args *spec, void *data) | 
|  | { | 
|  | struct generic_pm_domain *domain = ERR_PTR(-ENOENT); | 
|  | struct genpd_onecell_data *genpd = data; | 
|  |  | 
|  | if (spec->args_count != 1) | 
|  | return ERR_PTR(-EINVAL); | 
|  |  | 
|  | for (unsigned int i = 0; i < genpd->num_domains; i++) { | 
|  | struct rzg2l_cpg_pd *pd = container_of(genpd->domains[i], struct rzg2l_cpg_pd, | 
|  | genpd); | 
|  |  | 
|  | if (pd->id == spec->args[0]) { | 
|  | domain = &pd->genpd; | 
|  | break; | 
|  | } | 
|  | } | 
|  |  | 
|  | return domain; | 
|  | } | 
|  |  | 
|  | static int __init rzg2l_cpg_add_pm_domains(struct rzg2l_cpg_priv *priv) | 
|  | { | 
|  | const struct rzg2l_cpg_info *info = priv->info; | 
|  | struct device *dev = priv->dev; | 
|  | struct device_node *np = dev->of_node; | 
|  | struct rzg2l_cpg_pm_domains *domains; | 
|  | struct generic_pm_domain *parent; | 
|  | u32 ncells; | 
|  | int ret; | 
|  |  | 
|  | ret = of_property_read_u32(np, "#power-domain-cells", &ncells); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | /* For backward compatibility. */ | 
|  | if (!ncells) | 
|  | return rzg2l_cpg_add_clk_domain(priv); | 
|  |  | 
|  | domains = devm_kzalloc(dev, struct_size(domains, domains, info->num_pm_domains), | 
|  | GFP_KERNEL); | 
|  | if (!domains) | 
|  | return -ENOMEM; | 
|  |  | 
|  | domains->onecell_data.domains = domains->domains; | 
|  | domains->onecell_data.num_domains = info->num_pm_domains; | 
|  | domains->onecell_data.xlate = rzg2l_cpg_pm_domain_xlate; | 
|  |  | 
|  | ret = devm_add_action_or_reset(dev, rzg2l_cpg_genpd_remove, &domains->onecell_data); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | for (unsigned int i = 0; i < info->num_pm_domains; i++) { | 
|  | bool always_on = !!(info->pm_domains[i].flags & RZG2L_PD_F_ALWAYS_ON); | 
|  | struct rzg2l_cpg_pd *pd; | 
|  |  | 
|  | pd = devm_kzalloc(dev, sizeof(*pd), GFP_KERNEL); | 
|  | if (!pd) | 
|  | return -ENOMEM; | 
|  |  | 
|  | pd->genpd.name = info->pm_domains[i].name; | 
|  | pd->conf = info->pm_domains[i].conf; | 
|  | pd->id = info->pm_domains[i].id; | 
|  | pd->priv = priv; | 
|  |  | 
|  | ret = rzg2l_cpg_pd_setup(pd, always_on); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | if (always_on) { | 
|  | ret = rzg2l_cpg_power_on(&pd->genpd); | 
|  | if (ret) | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | domains->domains[i] = &pd->genpd; | 
|  | /* Parent should be on the very first entry of info->pm_domains[]. */ | 
|  | if (!i) { | 
|  | parent = &pd->genpd; | 
|  | continue; | 
|  | } | 
|  |  | 
|  | ret = pm_genpd_add_subdomain(parent, &pd->genpd); | 
|  | if (ret) | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | ret = of_genpd_add_provider_onecell(np, &domains->onecell_data); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int __init rzg2l_cpg_probe(struct platform_device *pdev) | 
|  | { | 
|  | struct device *dev = &pdev->dev; | 
|  | struct device_node *np = dev->of_node; | 
|  | const struct rzg2l_cpg_info *info; | 
|  | struct rzg2l_cpg_priv *priv; | 
|  | unsigned int nclks, i; | 
|  | struct clk **clks; | 
|  | int error; | 
|  |  | 
|  | info = of_device_get_match_data(dev); | 
|  |  | 
|  | priv = devm_kzalloc(dev, sizeof(*priv), GFP_KERNEL); | 
|  | if (!priv) | 
|  | return -ENOMEM; | 
|  |  | 
|  | priv->dev = dev; | 
|  | priv->info = info; | 
|  | spin_lock_init(&priv->rmw_lock); | 
|  |  | 
|  | priv->base = devm_platform_ioremap_resource(pdev, 0); | 
|  | if (IS_ERR(priv->base)) | 
|  | return PTR_ERR(priv->base); | 
|  |  | 
|  | nclks = info->num_total_core_clks + info->num_hw_mod_clks; | 
|  | clks = devm_kmalloc_array(dev, nclks, sizeof(*clks), GFP_KERNEL); | 
|  | if (!clks) | 
|  | return -ENOMEM; | 
|  |  | 
|  | dev_set_drvdata(dev, priv); | 
|  | priv->clks = clks; | 
|  | priv->num_core_clks = info->num_total_core_clks; | 
|  | priv->num_mod_clks = info->num_hw_mod_clks; | 
|  | priv->num_resets = info->num_resets; | 
|  | priv->last_dt_core_clk = info->last_dt_core_clk; | 
|  |  | 
|  | for (i = 0; i < nclks; i++) | 
|  | clks[i] = ERR_PTR(-ENOENT); | 
|  |  | 
|  | for (i = 0; i < info->num_core_clks; i++) | 
|  | rzg2l_cpg_register_core_clk(&info->core_clks[i], info, priv); | 
|  |  | 
|  | for (i = 0; i < info->num_mod_clks; i++) | 
|  | rzg2l_cpg_register_mod_clk(&info->mod_clks[i], info, priv); | 
|  |  | 
|  | error = of_clk_add_provider(np, rzg2l_cpg_clk_src_twocell_get, priv); | 
|  | if (error) | 
|  | return error; | 
|  |  | 
|  | error = devm_add_action_or_reset(dev, rzg2l_cpg_del_clk_provider, np); | 
|  | if (error) | 
|  | return error; | 
|  |  | 
|  | error = rzg2l_cpg_add_pm_domains(priv); | 
|  | if (error) | 
|  | return error; | 
|  |  | 
|  | error = rzg2l_cpg_reset_controller_register(priv); | 
|  | if (error) | 
|  | return error; | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static const struct of_device_id rzg2l_cpg_match[] = { | 
|  | #ifdef CONFIG_CLK_R9A07G043 | 
|  | { | 
|  | .compatible = "renesas,r9a07g043-cpg", | 
|  | .data = &r9a07g043_cpg_info, | 
|  | }, | 
|  | #endif | 
|  | #ifdef CONFIG_CLK_R9A07G044 | 
|  | { | 
|  | .compatible = "renesas,r9a07g044-cpg", | 
|  | .data = &r9a07g044_cpg_info, | 
|  | }, | 
|  | #endif | 
|  | #ifdef CONFIG_CLK_R9A07G054 | 
|  | { | 
|  | .compatible = "renesas,r9a07g054-cpg", | 
|  | .data = &r9a07g054_cpg_info, | 
|  | }, | 
|  | #endif | 
|  | #ifdef CONFIG_CLK_R9A08G045 | 
|  | { | 
|  | .compatible = "renesas,r9a08g045-cpg", | 
|  | .data = &r9a08g045_cpg_info, | 
|  | }, | 
|  | #endif | 
|  | #ifdef CONFIG_CLK_R9A09G011 | 
|  | { | 
|  | .compatible = "renesas,r9a09g011-cpg", | 
|  | .data = &r9a09g011_cpg_info, | 
|  | }, | 
|  | #endif | 
|  | { /* sentinel */ } | 
|  | }; | 
|  |  | 
|  | static struct platform_driver rzg2l_cpg_driver = { | 
|  | .driver		= { | 
|  | .name	= "rzg2l-cpg", | 
|  | .of_match_table = rzg2l_cpg_match, | 
|  | }, | 
|  | }; | 
|  |  | 
|  | static int __init rzg2l_cpg_init(void) | 
|  | { | 
|  | return platform_driver_probe(&rzg2l_cpg_driver, rzg2l_cpg_probe); | 
|  | } | 
|  |  | 
|  | subsys_initcall(rzg2l_cpg_init); | 
|  |  | 
|  | MODULE_DESCRIPTION("Renesas RZ/G2L CPG Driver"); |