mirror of
				git://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
				synced 2025-09-04 20:19:47 +08:00 
			
		
		
		
	 96dc791d0b
			
		
	
	
		96dc791d0b
		
	
	
	
	
		
			
			The RCG CFG/M/N/D register base could be at a different offset than the CMD register, so introduce a cfg_offset to identify the offset with respect to the CMD RCGR register. Signed-off-by: Taniya Das <tdas@codeaurora.org> Signed-off-by: Anu Ramanathan <anur@codeaurora.org> Signed-off-by: Shawn Guo <shawn.guo@linaro.org> Signed-off-by: Vinod Koul <vkoul@kernel.org> Signed-off-by: Stephen Boyd <sboyd@kernel.org>
		
			
				
	
	
		
			1131 lines
		
	
	
		
			27 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			1131 lines
		
	
	
		
			27 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| // SPDX-License-Identifier: GPL-2.0
 | |
| /*
 | |
|  * Copyright (c) 2013, 2018, The Linux Foundation. All rights reserved.
 | |
|  */
 | |
| 
 | |
| #include <linux/kernel.h>
 | |
| #include <linux/bitops.h>
 | |
| #include <linux/err.h>
 | |
| #include <linux/bug.h>
 | |
| #include <linux/export.h>
 | |
| #include <linux/clk-provider.h>
 | |
| #include <linux/delay.h>
 | |
| #include <linux/regmap.h>
 | |
| #include <linux/math64.h>
 | |
| #include <linux/slab.h>
 | |
| 
 | |
| #include <asm/div64.h>
 | |
| 
 | |
| #include "clk-rcg.h"
 | |
| #include "common.h"
 | |
| 
 | |
| #define CMD_REG			0x0
 | |
| #define CMD_UPDATE		BIT(0)
 | |
| #define CMD_ROOT_EN		BIT(1)
 | |
| #define CMD_DIRTY_CFG		BIT(4)
 | |
| #define CMD_DIRTY_N		BIT(5)
 | |
| #define CMD_DIRTY_M		BIT(6)
 | |
| #define CMD_DIRTY_D		BIT(7)
 | |
| #define CMD_ROOT_OFF		BIT(31)
 | |
| 
 | |
| #define CFG_REG			0x4
 | |
| #define CFG_SRC_DIV_SHIFT	0
 | |
| #define CFG_SRC_SEL_SHIFT	8
 | |
| #define CFG_SRC_SEL_MASK	(0x7 << CFG_SRC_SEL_SHIFT)
 | |
| #define CFG_MODE_SHIFT		12
 | |
| #define CFG_MODE_MASK		(0x3 << CFG_MODE_SHIFT)
 | |
| #define CFG_MODE_DUAL_EDGE	(0x2 << CFG_MODE_SHIFT)
 | |
| #define CFG_HW_CLK_CTRL_MASK	BIT(20)
 | |
| 
 | |
| #define M_REG			0x8
 | |
| #define N_REG			0xc
 | |
| #define D_REG			0x10
 | |
| 
 | |
| #define RCG_CFG_OFFSET(rcg)	((rcg)->cmd_rcgr + (rcg)->cfg_off + CFG_REG)
 | |
| #define RCG_M_OFFSET(rcg)	((rcg)->cmd_rcgr + (rcg)->cfg_off + M_REG)
 | |
| #define RCG_N_OFFSET(rcg)	((rcg)->cmd_rcgr + (rcg)->cfg_off + N_REG)
 | |
| #define RCG_D_OFFSET(rcg)	((rcg)->cmd_rcgr + (rcg)->cfg_off + D_REG)
 | |
| 
 | |
| /* Dynamic Frequency Scaling */
 | |
| #define MAX_PERF_LEVEL		8
 | |
| #define SE_CMD_DFSR_OFFSET	0x14
 | |
| #define SE_CMD_DFS_EN		BIT(0)
 | |
| #define SE_PERF_DFSR(level)	(0x1c + 0x4 * (level))
 | |
| #define SE_PERF_M_DFSR(level)	(0x5c + 0x4 * (level))
 | |
| #define SE_PERF_N_DFSR(level)	(0x9c + 0x4 * (level))
 | |
| 
 | |
| enum freq_policy {
 | |
| 	FLOOR,
 | |
| 	CEIL,
 | |
| };
 | |
| 
 | |
| static int clk_rcg2_is_enabled(struct clk_hw *hw)
 | |
| {
 | |
| 	struct clk_rcg2 *rcg = to_clk_rcg2(hw);
 | |
| 	u32 cmd;
 | |
| 	int ret;
 | |
| 
 | |
| 	ret = regmap_read(rcg->clkr.regmap, rcg->cmd_rcgr + CMD_REG, &cmd);
 | |
| 	if (ret)
 | |
| 		return ret;
 | |
| 
 | |
| 	return (cmd & CMD_ROOT_OFF) == 0;
 | |
| }
 | |
| 
 | |
| static u8 clk_rcg2_get_parent(struct clk_hw *hw)
 | |
| {
 | |
| 	struct clk_rcg2 *rcg = to_clk_rcg2(hw);
 | |
| 	int num_parents = clk_hw_get_num_parents(hw);
 | |
| 	u32 cfg;
 | |
| 	int i, ret;
 | |
| 
 | |
| 	ret = regmap_read(rcg->clkr.regmap, RCG_CFG_OFFSET(rcg), &cfg);
 | |
| 	if (ret)
 | |
| 		goto err;
 | |
| 
 | |
| 	cfg &= CFG_SRC_SEL_MASK;
 | |
| 	cfg >>= CFG_SRC_SEL_SHIFT;
 | |
| 
 | |
| 	for (i = 0; i < num_parents; i++)
 | |
| 		if (cfg == rcg->parent_map[i].cfg)
 | |
| 			return i;
 | |
| 
 | |
| err:
 | |
| 	pr_debug("%s: Clock %s has invalid parent, using default.\n",
 | |
| 		 __func__, clk_hw_get_name(hw));
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static int update_config(struct clk_rcg2 *rcg)
 | |
| {
 | |
| 	int count, ret;
 | |
| 	u32 cmd;
 | |
| 	struct clk_hw *hw = &rcg->clkr.hw;
 | |
| 	const char *name = clk_hw_get_name(hw);
 | |
| 
 | |
| 	ret = regmap_update_bits(rcg->clkr.regmap, rcg->cmd_rcgr + CMD_REG,
 | |
| 				 CMD_UPDATE, CMD_UPDATE);
 | |
| 	if (ret)
 | |
| 		return ret;
 | |
| 
 | |
| 	/* Wait for update to take effect */
 | |
| 	for (count = 500; count > 0; count--) {
 | |
| 		ret = regmap_read(rcg->clkr.regmap, rcg->cmd_rcgr + CMD_REG, &cmd);
 | |
| 		if (ret)
 | |
| 			return ret;
 | |
| 		if (!(cmd & CMD_UPDATE))
 | |
| 			return 0;
 | |
| 		udelay(1);
 | |
| 	}
 | |
| 
 | |
| 	WARN(1, "%s: rcg didn't update its configuration.", name);
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static int clk_rcg2_set_parent(struct clk_hw *hw, u8 index)
 | |
| {
 | |
| 	struct clk_rcg2 *rcg = to_clk_rcg2(hw);
 | |
| 	int ret;
 | |
| 	u32 cfg = rcg->parent_map[index].cfg << CFG_SRC_SEL_SHIFT;
 | |
| 
 | |
| 	ret = regmap_update_bits(rcg->clkr.regmap, RCG_CFG_OFFSET(rcg),
 | |
| 				 CFG_SRC_SEL_MASK, cfg);
 | |
| 	if (ret)
 | |
| 		return ret;
 | |
| 
 | |
| 	return update_config(rcg);
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Calculate m/n:d rate
 | |
|  *
 | |
|  *          parent_rate     m
 | |
|  *   rate = ----------- x  ---
 | |
|  *            hid_div       n
 | |
|  */
 | |
| static unsigned long
 | |
| calc_rate(unsigned long rate, u32 m, u32 n, u32 mode, u32 hid_div)
 | |
| {
 | |
| 	if (hid_div) {
 | |
| 		rate *= 2;
 | |
| 		rate /= hid_div + 1;
 | |
| 	}
 | |
| 
 | |
| 	if (mode) {
 | |
| 		u64 tmp = rate;
 | |
| 		tmp *= m;
 | |
| 		do_div(tmp, n);
 | |
| 		rate = tmp;
 | |
| 	}
 | |
| 
 | |
| 	return rate;
 | |
| }
 | |
| 
 | |
| static unsigned long
 | |
| clk_rcg2_recalc_rate(struct clk_hw *hw, unsigned long parent_rate)
 | |
| {
 | |
| 	struct clk_rcg2 *rcg = to_clk_rcg2(hw);
 | |
| 	u32 cfg, hid_div, m = 0, n = 0, mode = 0, mask;
 | |
| 
 | |
| 	regmap_read(rcg->clkr.regmap, RCG_CFG_OFFSET(rcg), &cfg);
 | |
| 
 | |
| 	if (rcg->mnd_width) {
 | |
| 		mask = BIT(rcg->mnd_width) - 1;
 | |
| 		regmap_read(rcg->clkr.regmap, RCG_M_OFFSET(rcg), &m);
 | |
| 		m &= mask;
 | |
| 		regmap_read(rcg->clkr.regmap, RCG_N_OFFSET(rcg), &n);
 | |
| 		n =  ~n;
 | |
| 		n &= mask;
 | |
| 		n += m;
 | |
| 		mode = cfg & CFG_MODE_MASK;
 | |
| 		mode >>= CFG_MODE_SHIFT;
 | |
| 	}
 | |
| 
 | |
| 	mask = BIT(rcg->hid_width) - 1;
 | |
| 	hid_div = cfg >> CFG_SRC_DIV_SHIFT;
 | |
| 	hid_div &= mask;
 | |
| 
 | |
| 	return calc_rate(parent_rate, m, n, mode, hid_div);
 | |
| }
 | |
| 
 | |
| static int _freq_tbl_determine_rate(struct clk_hw *hw, const struct freq_tbl *f,
 | |
| 				    struct clk_rate_request *req,
 | |
| 				    enum freq_policy policy)
 | |
| {
 | |
| 	unsigned long clk_flags, rate = req->rate;
 | |
| 	struct clk_hw *p;
 | |
| 	struct clk_rcg2 *rcg = to_clk_rcg2(hw);
 | |
| 	int index;
 | |
| 
 | |
| 	switch (policy) {
 | |
| 	case FLOOR:
 | |
| 		f = qcom_find_freq_floor(f, rate);
 | |
| 		break;
 | |
| 	case CEIL:
 | |
| 		f = qcom_find_freq(f, rate);
 | |
| 		break;
 | |
| 	default:
 | |
| 		return -EINVAL;
 | |
| 	};
 | |
| 
 | |
| 	if (!f)
 | |
| 		return -EINVAL;
 | |
| 
 | |
| 	index = qcom_find_src_index(hw, rcg->parent_map, f->src);
 | |
| 	if (index < 0)
 | |
| 		return index;
 | |
| 
 | |
| 	clk_flags = clk_hw_get_flags(hw);
 | |
| 	p = clk_hw_get_parent_by_index(hw, index);
 | |
| 	if (clk_flags & CLK_SET_RATE_PARENT) {
 | |
| 		rate = f->freq;
 | |
| 		if (f->pre_div) {
 | |
| 			rate /= 2;
 | |
| 			rate *= f->pre_div + 1;
 | |
| 		}
 | |
| 
 | |
| 		if (f->n) {
 | |
| 			u64 tmp = rate;
 | |
| 			tmp = tmp * f->n;
 | |
| 			do_div(tmp, f->m);
 | |
| 			rate = tmp;
 | |
| 		}
 | |
| 	} else {
 | |
| 		rate =  clk_hw_get_rate(p);
 | |
| 	}
 | |
| 	req->best_parent_hw = p;
 | |
| 	req->best_parent_rate = rate;
 | |
| 	req->rate = f->freq;
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static int clk_rcg2_determine_rate(struct clk_hw *hw,
 | |
| 				   struct clk_rate_request *req)
 | |
| {
 | |
| 	struct clk_rcg2 *rcg = to_clk_rcg2(hw);
 | |
| 
 | |
| 	return _freq_tbl_determine_rate(hw, rcg->freq_tbl, req, CEIL);
 | |
| }
 | |
| 
 | |
| static int clk_rcg2_determine_floor_rate(struct clk_hw *hw,
 | |
| 					 struct clk_rate_request *req)
 | |
| {
 | |
| 	struct clk_rcg2 *rcg = to_clk_rcg2(hw);
 | |
| 
 | |
| 	return _freq_tbl_determine_rate(hw, rcg->freq_tbl, req, FLOOR);
 | |
| }
 | |
| 
 | |
| static int __clk_rcg2_configure(struct clk_rcg2 *rcg, const struct freq_tbl *f)
 | |
| {
 | |
| 	u32 cfg, mask;
 | |
| 	struct clk_hw *hw = &rcg->clkr.hw;
 | |
| 	int ret, index = qcom_find_src_index(hw, rcg->parent_map, f->src);
 | |
| 
 | |
| 	if (index < 0)
 | |
| 		return index;
 | |
| 
 | |
| 	if (rcg->mnd_width && f->n) {
 | |
| 		mask = BIT(rcg->mnd_width) - 1;
 | |
| 		ret = regmap_update_bits(rcg->clkr.regmap,
 | |
| 				RCG_M_OFFSET(rcg), mask, f->m);
 | |
| 		if (ret)
 | |
| 			return ret;
 | |
| 
 | |
| 		ret = regmap_update_bits(rcg->clkr.regmap,
 | |
| 				RCG_N_OFFSET(rcg), mask, ~(f->n - f->m));
 | |
| 		if (ret)
 | |
| 			return ret;
 | |
| 
 | |
| 		ret = regmap_update_bits(rcg->clkr.regmap,
 | |
| 				RCG_D_OFFSET(rcg), mask, ~f->n);
 | |
| 		if (ret)
 | |
| 			return ret;
 | |
| 	}
 | |
| 
 | |
| 	mask = BIT(rcg->hid_width) - 1;
 | |
| 	mask |= CFG_SRC_SEL_MASK | CFG_MODE_MASK | CFG_HW_CLK_CTRL_MASK;
 | |
| 	cfg = f->pre_div << CFG_SRC_DIV_SHIFT;
 | |
| 	cfg |= rcg->parent_map[index].cfg << CFG_SRC_SEL_SHIFT;
 | |
| 	if (rcg->mnd_width && f->n && (f->m != f->n))
 | |
| 		cfg |= CFG_MODE_DUAL_EDGE;
 | |
| 	return regmap_update_bits(rcg->clkr.regmap, RCG_CFG_OFFSET(rcg),
 | |
| 					mask, cfg);
 | |
| }
 | |
| 
 | |
| static int clk_rcg2_configure(struct clk_rcg2 *rcg, const struct freq_tbl *f)
 | |
| {
 | |
| 	int ret;
 | |
| 
 | |
| 	ret = __clk_rcg2_configure(rcg, f);
 | |
| 	if (ret)
 | |
| 		return ret;
 | |
| 
 | |
| 	return update_config(rcg);
 | |
| }
 | |
| 
 | |
| static int __clk_rcg2_set_rate(struct clk_hw *hw, unsigned long rate,
 | |
| 			       enum freq_policy policy)
 | |
| {
 | |
| 	struct clk_rcg2 *rcg = to_clk_rcg2(hw);
 | |
| 	const struct freq_tbl *f;
 | |
| 
 | |
| 	switch (policy) {
 | |
| 	case FLOOR:
 | |
| 		f = qcom_find_freq_floor(rcg->freq_tbl, rate);
 | |
| 		break;
 | |
| 	case CEIL:
 | |
| 		f = qcom_find_freq(rcg->freq_tbl, rate);
 | |
| 		break;
 | |
| 	default:
 | |
| 		return -EINVAL;
 | |
| 	};
 | |
| 
 | |
| 	if (!f)
 | |
| 		return -EINVAL;
 | |
| 
 | |
| 	return clk_rcg2_configure(rcg, f);
 | |
| }
 | |
| 
 | |
| static int clk_rcg2_set_rate(struct clk_hw *hw, unsigned long rate,
 | |
| 			    unsigned long parent_rate)
 | |
| {
 | |
| 	return __clk_rcg2_set_rate(hw, rate, CEIL);
 | |
| }
 | |
| 
 | |
| static int clk_rcg2_set_floor_rate(struct clk_hw *hw, unsigned long rate,
 | |
| 				   unsigned long parent_rate)
 | |
| {
 | |
| 	return __clk_rcg2_set_rate(hw, rate, FLOOR);
 | |
| }
 | |
| 
 | |
| static int clk_rcg2_set_rate_and_parent(struct clk_hw *hw,
 | |
| 		unsigned long rate, unsigned long parent_rate, u8 index)
 | |
| {
 | |
| 	return __clk_rcg2_set_rate(hw, rate, CEIL);
 | |
| }
 | |
| 
 | |
| static int clk_rcg2_set_floor_rate_and_parent(struct clk_hw *hw,
 | |
| 		unsigned long rate, unsigned long parent_rate, u8 index)
 | |
| {
 | |
| 	return __clk_rcg2_set_rate(hw, rate, FLOOR);
 | |
| }
 | |
| 
 | |
| const struct clk_ops clk_rcg2_ops = {
 | |
| 	.is_enabled = clk_rcg2_is_enabled,
 | |
| 	.get_parent = clk_rcg2_get_parent,
 | |
| 	.set_parent = clk_rcg2_set_parent,
 | |
| 	.recalc_rate = clk_rcg2_recalc_rate,
 | |
| 	.determine_rate = clk_rcg2_determine_rate,
 | |
| 	.set_rate = clk_rcg2_set_rate,
 | |
| 	.set_rate_and_parent = clk_rcg2_set_rate_and_parent,
 | |
| };
 | |
| EXPORT_SYMBOL_GPL(clk_rcg2_ops);
 | |
| 
 | |
| const struct clk_ops clk_rcg2_floor_ops = {
 | |
| 	.is_enabled = clk_rcg2_is_enabled,
 | |
| 	.get_parent = clk_rcg2_get_parent,
 | |
| 	.set_parent = clk_rcg2_set_parent,
 | |
| 	.recalc_rate = clk_rcg2_recalc_rate,
 | |
| 	.determine_rate = clk_rcg2_determine_floor_rate,
 | |
| 	.set_rate = clk_rcg2_set_floor_rate,
 | |
| 	.set_rate_and_parent = clk_rcg2_set_floor_rate_and_parent,
 | |
| };
 | |
| EXPORT_SYMBOL_GPL(clk_rcg2_floor_ops);
 | |
| 
 | |
| struct frac_entry {
 | |
| 	int num;
 | |
| 	int den;
 | |
| };
 | |
| 
 | |
| static const struct frac_entry frac_table_675m[] = {	/* link rate of 270M */
 | |
| 	{ 52, 295 },	/* 119 M */
 | |
| 	{ 11, 57 },	/* 130.25 M */
 | |
| 	{ 63, 307 },	/* 138.50 M */
 | |
| 	{ 11, 50 },	/* 148.50 M */
 | |
| 	{ 47, 206 },	/* 154 M */
 | |
| 	{ 31, 100 },	/* 205.25 M */
 | |
| 	{ 107, 269 },	/* 268.50 M */
 | |
| 	{ },
 | |
| };
 | |
| 
 | |
| static struct frac_entry frac_table_810m[] = { /* Link rate of 162M */
 | |
| 	{ 31, 211 },	/* 119 M */
 | |
| 	{ 32, 199 },	/* 130.25 M */
 | |
| 	{ 63, 307 },	/* 138.50 M */
 | |
| 	{ 11, 60 },	/* 148.50 M */
 | |
| 	{ 50, 263 },	/* 154 M */
 | |
| 	{ 31, 120 },	/* 205.25 M */
 | |
| 	{ 119, 359 },	/* 268.50 M */
 | |
| 	{ },
 | |
| };
 | |
| 
 | |
| static int clk_edp_pixel_set_rate(struct clk_hw *hw, unsigned long rate,
 | |
| 			      unsigned long parent_rate)
 | |
| {
 | |
| 	struct clk_rcg2 *rcg = to_clk_rcg2(hw);
 | |
| 	struct freq_tbl f = *rcg->freq_tbl;
 | |
| 	const struct frac_entry *frac;
 | |
| 	int delta = 100000;
 | |
| 	s64 src_rate = parent_rate;
 | |
| 	s64 request;
 | |
| 	u32 mask = BIT(rcg->hid_width) - 1;
 | |
| 	u32 hid_div;
 | |
| 
 | |
| 	if (src_rate == 810000000)
 | |
| 		frac = frac_table_810m;
 | |
| 	else
 | |
| 		frac = frac_table_675m;
 | |
| 
 | |
| 	for (; frac->num; frac++) {
 | |
| 		request = rate;
 | |
| 		request *= frac->den;
 | |
| 		request = div_s64(request, frac->num);
 | |
| 		if ((src_rate < (request - delta)) ||
 | |
| 		    (src_rate > (request + delta)))
 | |
| 			continue;
 | |
| 
 | |
| 		regmap_read(rcg->clkr.regmap, rcg->cmd_rcgr + CFG_REG,
 | |
| 				&hid_div);
 | |
| 		f.pre_div = hid_div;
 | |
| 		f.pre_div >>= CFG_SRC_DIV_SHIFT;
 | |
| 		f.pre_div &= mask;
 | |
| 		f.m = frac->num;
 | |
| 		f.n = frac->den;
 | |
| 
 | |
| 		return clk_rcg2_configure(rcg, &f);
 | |
| 	}
 | |
| 
 | |
| 	return -EINVAL;
 | |
| }
 | |
| 
 | |
| static int clk_edp_pixel_set_rate_and_parent(struct clk_hw *hw,
 | |
| 		unsigned long rate, unsigned long parent_rate, u8 index)
 | |
| {
 | |
| 	/* Parent index is set statically in frequency table */
 | |
| 	return clk_edp_pixel_set_rate(hw, rate, parent_rate);
 | |
| }
 | |
| 
 | |
| static int clk_edp_pixel_determine_rate(struct clk_hw *hw,
 | |
| 					struct clk_rate_request *req)
 | |
| {
 | |
| 	struct clk_rcg2 *rcg = to_clk_rcg2(hw);
 | |
| 	const struct freq_tbl *f = rcg->freq_tbl;
 | |
| 	const struct frac_entry *frac;
 | |
| 	int delta = 100000;
 | |
| 	s64 request;
 | |
| 	u32 mask = BIT(rcg->hid_width) - 1;
 | |
| 	u32 hid_div;
 | |
| 	int index = qcom_find_src_index(hw, rcg->parent_map, f->src);
 | |
| 
 | |
| 	/* Force the correct parent */
 | |
| 	req->best_parent_hw = clk_hw_get_parent_by_index(hw, index);
 | |
| 	req->best_parent_rate = clk_hw_get_rate(req->best_parent_hw);
 | |
| 
 | |
| 	if (req->best_parent_rate == 810000000)
 | |
| 		frac = frac_table_810m;
 | |
| 	else
 | |
| 		frac = frac_table_675m;
 | |
| 
 | |
| 	for (; frac->num; frac++) {
 | |
| 		request = req->rate;
 | |
| 		request *= frac->den;
 | |
| 		request = div_s64(request, frac->num);
 | |
| 		if ((req->best_parent_rate < (request - delta)) ||
 | |
| 		    (req->best_parent_rate > (request + delta)))
 | |
| 			continue;
 | |
| 
 | |
| 		regmap_read(rcg->clkr.regmap, rcg->cmd_rcgr + CFG_REG,
 | |
| 				&hid_div);
 | |
| 		hid_div >>= CFG_SRC_DIV_SHIFT;
 | |
| 		hid_div &= mask;
 | |
| 
 | |
| 		req->rate = calc_rate(req->best_parent_rate,
 | |
| 				      frac->num, frac->den,
 | |
| 				      !!frac->den, hid_div);
 | |
| 		return 0;
 | |
| 	}
 | |
| 
 | |
| 	return -EINVAL;
 | |
| }
 | |
| 
 | |
| const struct clk_ops clk_edp_pixel_ops = {
 | |
| 	.is_enabled = clk_rcg2_is_enabled,
 | |
| 	.get_parent = clk_rcg2_get_parent,
 | |
| 	.set_parent = clk_rcg2_set_parent,
 | |
| 	.recalc_rate = clk_rcg2_recalc_rate,
 | |
| 	.set_rate = clk_edp_pixel_set_rate,
 | |
| 	.set_rate_and_parent = clk_edp_pixel_set_rate_and_parent,
 | |
| 	.determine_rate = clk_edp_pixel_determine_rate,
 | |
| };
 | |
| EXPORT_SYMBOL_GPL(clk_edp_pixel_ops);
 | |
| 
 | |
| static int clk_byte_determine_rate(struct clk_hw *hw,
 | |
| 				   struct clk_rate_request *req)
 | |
| {
 | |
| 	struct clk_rcg2 *rcg = to_clk_rcg2(hw);
 | |
| 	const struct freq_tbl *f = rcg->freq_tbl;
 | |
| 	int index = qcom_find_src_index(hw, rcg->parent_map, f->src);
 | |
| 	unsigned long parent_rate, div;
 | |
| 	u32 mask = BIT(rcg->hid_width) - 1;
 | |
| 	struct clk_hw *p;
 | |
| 
 | |
| 	if (req->rate == 0)
 | |
| 		return -EINVAL;
 | |
| 
 | |
| 	req->best_parent_hw = p = clk_hw_get_parent_by_index(hw, index);
 | |
| 	req->best_parent_rate = parent_rate = clk_hw_round_rate(p, req->rate);
 | |
| 
 | |
| 	div = DIV_ROUND_UP((2 * parent_rate), req->rate) - 1;
 | |
| 	div = min_t(u32, div, mask);
 | |
| 
 | |
| 	req->rate = calc_rate(parent_rate, 0, 0, 0, div);
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static int clk_byte_set_rate(struct clk_hw *hw, unsigned long rate,
 | |
| 			 unsigned long parent_rate)
 | |
| {
 | |
| 	struct clk_rcg2 *rcg = to_clk_rcg2(hw);
 | |
| 	struct freq_tbl f = *rcg->freq_tbl;
 | |
| 	unsigned long div;
 | |
| 	u32 mask = BIT(rcg->hid_width) - 1;
 | |
| 
 | |
| 	div = DIV_ROUND_UP((2 * parent_rate), rate) - 1;
 | |
| 	div = min_t(u32, div, mask);
 | |
| 
 | |
| 	f.pre_div = div;
 | |
| 
 | |
| 	return clk_rcg2_configure(rcg, &f);
 | |
| }
 | |
| 
 | |
| static int clk_byte_set_rate_and_parent(struct clk_hw *hw,
 | |
| 		unsigned long rate, unsigned long parent_rate, u8 index)
 | |
| {
 | |
| 	/* Parent index is set statically in frequency table */
 | |
| 	return clk_byte_set_rate(hw, rate, parent_rate);
 | |
| }
 | |
| 
 | |
| const struct clk_ops clk_byte_ops = {
 | |
| 	.is_enabled = clk_rcg2_is_enabled,
 | |
| 	.get_parent = clk_rcg2_get_parent,
 | |
| 	.set_parent = clk_rcg2_set_parent,
 | |
| 	.recalc_rate = clk_rcg2_recalc_rate,
 | |
| 	.set_rate = clk_byte_set_rate,
 | |
| 	.set_rate_and_parent = clk_byte_set_rate_and_parent,
 | |
| 	.determine_rate = clk_byte_determine_rate,
 | |
| };
 | |
| EXPORT_SYMBOL_GPL(clk_byte_ops);
 | |
| 
 | |
| static int clk_byte2_determine_rate(struct clk_hw *hw,
 | |
| 				    struct clk_rate_request *req)
 | |
| {
 | |
| 	struct clk_rcg2 *rcg = to_clk_rcg2(hw);
 | |
| 	unsigned long parent_rate, div;
 | |
| 	u32 mask = BIT(rcg->hid_width) - 1;
 | |
| 	struct clk_hw *p;
 | |
| 	unsigned long rate = req->rate;
 | |
| 
 | |
| 	if (rate == 0)
 | |
| 		return -EINVAL;
 | |
| 
 | |
| 	p = req->best_parent_hw;
 | |
| 	req->best_parent_rate = parent_rate = clk_hw_round_rate(p, rate);
 | |
| 
 | |
| 	div = DIV_ROUND_UP((2 * parent_rate), rate) - 1;
 | |
| 	div = min_t(u32, div, mask);
 | |
| 
 | |
| 	req->rate = calc_rate(parent_rate, 0, 0, 0, div);
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static int clk_byte2_set_rate(struct clk_hw *hw, unsigned long rate,
 | |
| 			 unsigned long parent_rate)
 | |
| {
 | |
| 	struct clk_rcg2 *rcg = to_clk_rcg2(hw);
 | |
| 	struct freq_tbl f = { 0 };
 | |
| 	unsigned long div;
 | |
| 	int i, num_parents = clk_hw_get_num_parents(hw);
 | |
| 	u32 mask = BIT(rcg->hid_width) - 1;
 | |
| 	u32 cfg;
 | |
| 
 | |
| 	div = DIV_ROUND_UP((2 * parent_rate), rate) - 1;
 | |
| 	div = min_t(u32, div, mask);
 | |
| 
 | |
| 	f.pre_div = div;
 | |
| 
 | |
| 	regmap_read(rcg->clkr.regmap, rcg->cmd_rcgr + CFG_REG, &cfg);
 | |
| 	cfg &= CFG_SRC_SEL_MASK;
 | |
| 	cfg >>= CFG_SRC_SEL_SHIFT;
 | |
| 
 | |
| 	for (i = 0; i < num_parents; i++) {
 | |
| 		if (cfg == rcg->parent_map[i].cfg) {
 | |
| 			f.src = rcg->parent_map[i].src;
 | |
| 			return clk_rcg2_configure(rcg, &f);
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return -EINVAL;
 | |
| }
 | |
| 
 | |
| static int clk_byte2_set_rate_and_parent(struct clk_hw *hw,
 | |
| 		unsigned long rate, unsigned long parent_rate, u8 index)
 | |
| {
 | |
| 	/* Read the hardware to determine parent during set_rate */
 | |
| 	return clk_byte2_set_rate(hw, rate, parent_rate);
 | |
| }
 | |
| 
 | |
| const struct clk_ops clk_byte2_ops = {
 | |
| 	.is_enabled = clk_rcg2_is_enabled,
 | |
| 	.get_parent = clk_rcg2_get_parent,
 | |
| 	.set_parent = clk_rcg2_set_parent,
 | |
| 	.recalc_rate = clk_rcg2_recalc_rate,
 | |
| 	.set_rate = clk_byte2_set_rate,
 | |
| 	.set_rate_and_parent = clk_byte2_set_rate_and_parent,
 | |
| 	.determine_rate = clk_byte2_determine_rate,
 | |
| };
 | |
| EXPORT_SYMBOL_GPL(clk_byte2_ops);
 | |
| 
 | |
| static const struct frac_entry frac_table_pixel[] = {
 | |
| 	{ 3, 8 },
 | |
| 	{ 2, 9 },
 | |
| 	{ 4, 9 },
 | |
| 	{ 1, 1 },
 | |
| 	{ }
 | |
| };
 | |
| 
 | |
| static int clk_pixel_determine_rate(struct clk_hw *hw,
 | |
| 				    struct clk_rate_request *req)
 | |
| {
 | |
| 	unsigned long request, src_rate;
 | |
| 	int delta = 100000;
 | |
| 	const struct frac_entry *frac = frac_table_pixel;
 | |
| 
 | |
| 	for (; frac->num; frac++) {
 | |
| 		request = (req->rate * frac->den) / frac->num;
 | |
| 
 | |
| 		src_rate = clk_hw_round_rate(req->best_parent_hw, request);
 | |
| 		if ((src_rate < (request - delta)) ||
 | |
| 			(src_rate > (request + delta)))
 | |
| 			continue;
 | |
| 
 | |
| 		req->best_parent_rate = src_rate;
 | |
| 		req->rate = (src_rate * frac->num) / frac->den;
 | |
| 		return 0;
 | |
| 	}
 | |
| 
 | |
| 	return -EINVAL;
 | |
| }
 | |
| 
 | |
| static int clk_pixel_set_rate(struct clk_hw *hw, unsigned long rate,
 | |
| 		unsigned long parent_rate)
 | |
| {
 | |
| 	struct clk_rcg2 *rcg = to_clk_rcg2(hw);
 | |
| 	struct freq_tbl f = { 0 };
 | |
| 	const struct frac_entry *frac = frac_table_pixel;
 | |
| 	unsigned long request;
 | |
| 	int delta = 100000;
 | |
| 	u32 mask = BIT(rcg->hid_width) - 1;
 | |
| 	u32 hid_div, cfg;
 | |
| 	int i, num_parents = clk_hw_get_num_parents(hw);
 | |
| 
 | |
| 	regmap_read(rcg->clkr.regmap, rcg->cmd_rcgr + CFG_REG, &cfg);
 | |
| 	cfg &= CFG_SRC_SEL_MASK;
 | |
| 	cfg >>= CFG_SRC_SEL_SHIFT;
 | |
| 
 | |
| 	for (i = 0; i < num_parents; i++)
 | |
| 		if (cfg == rcg->parent_map[i].cfg) {
 | |
| 			f.src = rcg->parent_map[i].src;
 | |
| 			break;
 | |
| 		}
 | |
| 
 | |
| 	for (; frac->num; frac++) {
 | |
| 		request = (rate * frac->den) / frac->num;
 | |
| 
 | |
| 		if ((parent_rate < (request - delta)) ||
 | |
| 			(parent_rate > (request + delta)))
 | |
| 			continue;
 | |
| 
 | |
| 		regmap_read(rcg->clkr.regmap, rcg->cmd_rcgr + CFG_REG,
 | |
| 				&hid_div);
 | |
| 		f.pre_div = hid_div;
 | |
| 		f.pre_div >>= CFG_SRC_DIV_SHIFT;
 | |
| 		f.pre_div &= mask;
 | |
| 		f.m = frac->num;
 | |
| 		f.n = frac->den;
 | |
| 
 | |
| 		return clk_rcg2_configure(rcg, &f);
 | |
| 	}
 | |
| 	return -EINVAL;
 | |
| }
 | |
| 
 | |
| static int clk_pixel_set_rate_and_parent(struct clk_hw *hw, unsigned long rate,
 | |
| 		unsigned long parent_rate, u8 index)
 | |
| {
 | |
| 	return clk_pixel_set_rate(hw, rate, parent_rate);
 | |
| }
 | |
| 
 | |
| const struct clk_ops clk_pixel_ops = {
 | |
| 	.is_enabled = clk_rcg2_is_enabled,
 | |
| 	.get_parent = clk_rcg2_get_parent,
 | |
| 	.set_parent = clk_rcg2_set_parent,
 | |
| 	.recalc_rate = clk_rcg2_recalc_rate,
 | |
| 	.set_rate = clk_pixel_set_rate,
 | |
| 	.set_rate_and_parent = clk_pixel_set_rate_and_parent,
 | |
| 	.determine_rate = clk_pixel_determine_rate,
 | |
| };
 | |
| EXPORT_SYMBOL_GPL(clk_pixel_ops);
 | |
| 
 | |
| static int clk_gfx3d_determine_rate(struct clk_hw *hw,
 | |
| 				    struct clk_rate_request *req)
 | |
| {
 | |
| 	struct clk_rate_request parent_req = { };
 | |
| 	struct clk_hw *p2, *p8, *p9, *xo;
 | |
| 	unsigned long p9_rate;
 | |
| 	int ret;
 | |
| 
 | |
| 	xo = clk_hw_get_parent_by_index(hw, 0);
 | |
| 	if (req->rate == clk_hw_get_rate(xo)) {
 | |
| 		req->best_parent_hw = xo;
 | |
| 		return 0;
 | |
| 	}
 | |
| 
 | |
| 	p9 = clk_hw_get_parent_by_index(hw, 2);
 | |
| 	p2 = clk_hw_get_parent_by_index(hw, 3);
 | |
| 	p8 = clk_hw_get_parent_by_index(hw, 4);
 | |
| 
 | |
| 	/* PLL9 is a fixed rate PLL */
 | |
| 	p9_rate = clk_hw_get_rate(p9);
 | |
| 
 | |
| 	parent_req.rate = req->rate = min(req->rate, p9_rate);
 | |
| 	if (req->rate == p9_rate) {
 | |
| 		req->rate = req->best_parent_rate = p9_rate;
 | |
| 		req->best_parent_hw = p9;
 | |
| 		return 0;
 | |
| 	}
 | |
| 
 | |
| 	if (req->best_parent_hw == p9) {
 | |
| 		/* Are we going back to a previously used rate? */
 | |
| 		if (clk_hw_get_rate(p8) == req->rate)
 | |
| 			req->best_parent_hw = p8;
 | |
| 		else
 | |
| 			req->best_parent_hw = p2;
 | |
| 	} else if (req->best_parent_hw == p8) {
 | |
| 		req->best_parent_hw = p2;
 | |
| 	} else {
 | |
| 		req->best_parent_hw = p8;
 | |
| 	}
 | |
| 
 | |
| 	ret = __clk_determine_rate(req->best_parent_hw, &parent_req);
 | |
| 	if (ret)
 | |
| 		return ret;
 | |
| 
 | |
| 	req->rate = req->best_parent_rate = parent_req.rate;
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static int clk_gfx3d_set_rate_and_parent(struct clk_hw *hw, unsigned long rate,
 | |
| 		unsigned long parent_rate, u8 index)
 | |
| {
 | |
| 	struct clk_rcg2 *rcg = to_clk_rcg2(hw);
 | |
| 	u32 cfg;
 | |
| 	int ret;
 | |
| 
 | |
| 	/* Just mux it, we don't use the division or m/n hardware */
 | |
| 	cfg = rcg->parent_map[index].cfg << CFG_SRC_SEL_SHIFT;
 | |
| 	ret = regmap_write(rcg->clkr.regmap, rcg->cmd_rcgr + CFG_REG, cfg);
 | |
| 	if (ret)
 | |
| 		return ret;
 | |
| 
 | |
| 	return update_config(rcg);
 | |
| }
 | |
| 
 | |
| static int clk_gfx3d_set_rate(struct clk_hw *hw, unsigned long rate,
 | |
| 			      unsigned long parent_rate)
 | |
| {
 | |
| 	/*
 | |
| 	 * We should never get here; clk_gfx3d_determine_rate() should always
 | |
| 	 * make us use a different parent than what we're currently using, so
 | |
| 	 * clk_gfx3d_set_rate_and_parent() should always be called.
 | |
| 	 */
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| const struct clk_ops clk_gfx3d_ops = {
 | |
| 	.is_enabled = clk_rcg2_is_enabled,
 | |
| 	.get_parent = clk_rcg2_get_parent,
 | |
| 	.set_parent = clk_rcg2_set_parent,
 | |
| 	.recalc_rate = clk_rcg2_recalc_rate,
 | |
| 	.set_rate = clk_gfx3d_set_rate,
 | |
| 	.set_rate_and_parent = clk_gfx3d_set_rate_and_parent,
 | |
| 	.determine_rate = clk_gfx3d_determine_rate,
 | |
| };
 | |
| EXPORT_SYMBOL_GPL(clk_gfx3d_ops);
 | |
| 
 | |
| static int clk_rcg2_set_force_enable(struct clk_hw *hw)
 | |
| {
 | |
| 	struct clk_rcg2 *rcg = to_clk_rcg2(hw);
 | |
| 	const char *name = clk_hw_get_name(hw);
 | |
| 	int ret, count;
 | |
| 
 | |
| 	ret = regmap_update_bits(rcg->clkr.regmap, rcg->cmd_rcgr + CMD_REG,
 | |
| 				 CMD_ROOT_EN, CMD_ROOT_EN);
 | |
| 	if (ret)
 | |
| 		return ret;
 | |
| 
 | |
| 	/* wait for RCG to turn ON */
 | |
| 	for (count = 500; count > 0; count--) {
 | |
| 		if (clk_rcg2_is_enabled(hw))
 | |
| 			return 0;
 | |
| 
 | |
| 		udelay(1);
 | |
| 	}
 | |
| 
 | |
| 	pr_err("%s: RCG did not turn on\n", name);
 | |
| 	return -ETIMEDOUT;
 | |
| }
 | |
| 
 | |
| static int clk_rcg2_clear_force_enable(struct clk_hw *hw)
 | |
| {
 | |
| 	struct clk_rcg2 *rcg = to_clk_rcg2(hw);
 | |
| 
 | |
| 	return regmap_update_bits(rcg->clkr.regmap, rcg->cmd_rcgr + CMD_REG,
 | |
| 					CMD_ROOT_EN, 0);
 | |
| }
 | |
| 
 | |
| static int
 | |
| clk_rcg2_shared_force_enable_clear(struct clk_hw *hw, const struct freq_tbl *f)
 | |
| {
 | |
| 	struct clk_rcg2 *rcg = to_clk_rcg2(hw);
 | |
| 	int ret;
 | |
| 
 | |
| 	ret = clk_rcg2_set_force_enable(hw);
 | |
| 	if (ret)
 | |
| 		return ret;
 | |
| 
 | |
| 	ret = clk_rcg2_configure(rcg, f);
 | |
| 	if (ret)
 | |
| 		return ret;
 | |
| 
 | |
| 	return clk_rcg2_clear_force_enable(hw);
 | |
| }
 | |
| 
 | |
| static int clk_rcg2_shared_set_rate(struct clk_hw *hw, unsigned long rate,
 | |
| 				    unsigned long parent_rate)
 | |
| {
 | |
| 	struct clk_rcg2 *rcg = to_clk_rcg2(hw);
 | |
| 	const struct freq_tbl *f;
 | |
| 
 | |
| 	f = qcom_find_freq(rcg->freq_tbl, rate);
 | |
| 	if (!f)
 | |
| 		return -EINVAL;
 | |
| 
 | |
| 	/*
 | |
| 	 * In case clock is disabled, update the CFG, M, N and D registers
 | |
| 	 * and don't hit the update bit of CMD register.
 | |
| 	 */
 | |
| 	if (!__clk_is_enabled(hw->clk))
 | |
| 		return __clk_rcg2_configure(rcg, f);
 | |
| 
 | |
| 	return clk_rcg2_shared_force_enable_clear(hw, f);
 | |
| }
 | |
| 
 | |
| static int clk_rcg2_shared_set_rate_and_parent(struct clk_hw *hw,
 | |
| 		unsigned long rate, unsigned long parent_rate, u8 index)
 | |
| {
 | |
| 	return clk_rcg2_shared_set_rate(hw, rate, parent_rate);
 | |
| }
 | |
| 
 | |
| static int clk_rcg2_shared_enable(struct clk_hw *hw)
 | |
| {
 | |
| 	struct clk_rcg2 *rcg = to_clk_rcg2(hw);
 | |
| 	int ret;
 | |
| 
 | |
| 	/*
 | |
| 	 * Set the update bit because required configuration has already
 | |
| 	 * been written in clk_rcg2_shared_set_rate()
 | |
| 	 */
 | |
| 	ret = clk_rcg2_set_force_enable(hw);
 | |
| 	if (ret)
 | |
| 		return ret;
 | |
| 
 | |
| 	ret = update_config(rcg);
 | |
| 	if (ret)
 | |
| 		return ret;
 | |
| 
 | |
| 	return clk_rcg2_clear_force_enable(hw);
 | |
| }
 | |
| 
 | |
| static void clk_rcg2_shared_disable(struct clk_hw *hw)
 | |
| {
 | |
| 	struct clk_rcg2 *rcg = to_clk_rcg2(hw);
 | |
| 	u32 cfg;
 | |
| 
 | |
| 	/*
 | |
| 	 * Store current configuration as switching to safe source would clear
 | |
| 	 * the SRC and DIV of CFG register
 | |
| 	 */
 | |
| 	regmap_read(rcg->clkr.regmap, rcg->cmd_rcgr + CFG_REG, &cfg);
 | |
| 
 | |
| 	/*
 | |
| 	 * Park the RCG at a safe configuration - sourced off of safe source.
 | |
| 	 * Force enable and disable the RCG while configuring it to safeguard
 | |
| 	 * against any update signal coming from the downstream clock.
 | |
| 	 * The current parent is still prepared and enabled at this point, and
 | |
| 	 * the safe source is always on while application processor subsystem
 | |
| 	 * is online. Therefore, the RCG can safely switch its parent.
 | |
| 	 */
 | |
| 	clk_rcg2_set_force_enable(hw);
 | |
| 
 | |
| 	regmap_write(rcg->clkr.regmap, rcg->cmd_rcgr + CFG_REG,
 | |
| 		     rcg->safe_src_index << CFG_SRC_SEL_SHIFT);
 | |
| 
 | |
| 	update_config(rcg);
 | |
| 
 | |
| 	clk_rcg2_clear_force_enable(hw);
 | |
| 
 | |
| 	/* Write back the stored configuration corresponding to current rate */
 | |
| 	regmap_write(rcg->clkr.regmap, rcg->cmd_rcgr + CFG_REG, cfg);
 | |
| }
 | |
| 
 | |
| const struct clk_ops clk_rcg2_shared_ops = {
 | |
| 	.enable = clk_rcg2_shared_enable,
 | |
| 	.disable = clk_rcg2_shared_disable,
 | |
| 	.get_parent = clk_rcg2_get_parent,
 | |
| 	.set_parent = clk_rcg2_set_parent,
 | |
| 	.recalc_rate = clk_rcg2_recalc_rate,
 | |
| 	.determine_rate = clk_rcg2_determine_rate,
 | |
| 	.set_rate = clk_rcg2_shared_set_rate,
 | |
| 	.set_rate_and_parent = clk_rcg2_shared_set_rate_and_parent,
 | |
| };
 | |
| EXPORT_SYMBOL_GPL(clk_rcg2_shared_ops);
 | |
| 
 | |
| /* Common APIs to be used for DFS based RCGR */
 | |
| static void clk_rcg2_dfs_populate_freq(struct clk_hw *hw, unsigned int l,
 | |
| 				       struct freq_tbl *f)
 | |
| {
 | |
| 	struct clk_rcg2 *rcg = to_clk_rcg2(hw);
 | |
| 	struct clk_hw *p;
 | |
| 	unsigned long prate = 0;
 | |
| 	u32 val, mask, cfg, mode;
 | |
| 	int i, num_parents;
 | |
| 
 | |
| 	regmap_read(rcg->clkr.regmap, rcg->cmd_rcgr + SE_PERF_DFSR(l), &cfg);
 | |
| 
 | |
| 	mask = BIT(rcg->hid_width) - 1;
 | |
| 	f->pre_div = 1;
 | |
| 	if (cfg & mask)
 | |
| 		f->pre_div = cfg & mask;
 | |
| 
 | |
| 	cfg &= CFG_SRC_SEL_MASK;
 | |
| 	cfg >>= CFG_SRC_SEL_SHIFT;
 | |
| 
 | |
| 	num_parents = clk_hw_get_num_parents(hw);
 | |
| 	for (i = 0; i < num_parents; i++) {
 | |
| 		if (cfg == rcg->parent_map[i].cfg) {
 | |
| 			f->src = rcg->parent_map[i].src;
 | |
| 			p = clk_hw_get_parent_by_index(&rcg->clkr.hw, i);
 | |
| 			prate = clk_hw_get_rate(p);
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	mode = cfg & CFG_MODE_MASK;
 | |
| 	mode >>= CFG_MODE_SHIFT;
 | |
| 	if (mode) {
 | |
| 		mask = BIT(rcg->mnd_width) - 1;
 | |
| 		regmap_read(rcg->clkr.regmap, rcg->cmd_rcgr + SE_PERF_M_DFSR(l),
 | |
| 			    &val);
 | |
| 		val &= mask;
 | |
| 		f->m = val;
 | |
| 
 | |
| 		regmap_read(rcg->clkr.regmap, rcg->cmd_rcgr + SE_PERF_N_DFSR(l),
 | |
| 			    &val);
 | |
| 		val = ~val;
 | |
| 		val &= mask;
 | |
| 		val += f->m;
 | |
| 		f->n = val;
 | |
| 	}
 | |
| 
 | |
| 	f->freq = calc_rate(prate, f->m, f->n, mode, f->pre_div);
 | |
| }
 | |
| 
 | |
| static int clk_rcg2_dfs_populate_freq_table(struct clk_rcg2 *rcg)
 | |
| {
 | |
| 	struct freq_tbl *freq_tbl;
 | |
| 	int i;
 | |
| 
 | |
| 	/* Allocate space for 1 extra since table is NULL terminated */
 | |
| 	freq_tbl = kcalloc(MAX_PERF_LEVEL + 1, sizeof(*freq_tbl), GFP_KERNEL);
 | |
| 	if (!freq_tbl)
 | |
| 		return -ENOMEM;
 | |
| 	rcg->freq_tbl = freq_tbl;
 | |
| 
 | |
| 	for (i = 0; i < MAX_PERF_LEVEL; i++)
 | |
| 		clk_rcg2_dfs_populate_freq(&rcg->clkr.hw, i, freq_tbl + i);
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static int clk_rcg2_dfs_determine_rate(struct clk_hw *hw,
 | |
| 				   struct clk_rate_request *req)
 | |
| {
 | |
| 	struct clk_rcg2 *rcg = to_clk_rcg2(hw);
 | |
| 	int ret;
 | |
| 
 | |
| 	if (!rcg->freq_tbl) {
 | |
| 		ret = clk_rcg2_dfs_populate_freq_table(rcg);
 | |
| 		if (ret) {
 | |
| 			pr_err("Failed to update DFS tables for %s\n",
 | |
| 					clk_hw_get_name(hw));
 | |
| 			return ret;
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return clk_rcg2_determine_rate(hw, req);
 | |
| }
 | |
| 
 | |
| static unsigned long
 | |
| clk_rcg2_dfs_recalc_rate(struct clk_hw *hw, unsigned long parent_rate)
 | |
| {
 | |
| 	struct clk_rcg2 *rcg = to_clk_rcg2(hw);
 | |
| 	u32 level, mask, cfg, m = 0, n = 0, mode, pre_div;
 | |
| 
 | |
| 	regmap_read(rcg->clkr.regmap,
 | |
| 		    rcg->cmd_rcgr + SE_CMD_DFSR_OFFSET, &level);
 | |
| 	level &= GENMASK(4, 1);
 | |
| 	level >>= 1;
 | |
| 
 | |
| 	if (rcg->freq_tbl)
 | |
| 		return rcg->freq_tbl[level].freq;
 | |
| 
 | |
| 	/*
 | |
| 	 * Assume that parent_rate is actually the parent because
 | |
| 	 * we can't do any better at figuring it out when the table
 | |
| 	 * hasn't been populated yet. We only populate the table
 | |
| 	 * in determine_rate because we can't guarantee the parents
 | |
| 	 * will be registered with the framework until then.
 | |
| 	 */
 | |
| 	regmap_read(rcg->clkr.regmap, rcg->cmd_rcgr + SE_PERF_DFSR(level),
 | |
| 		    &cfg);
 | |
| 
 | |
| 	mask = BIT(rcg->hid_width) - 1;
 | |
| 	pre_div = 1;
 | |
| 	if (cfg & mask)
 | |
| 		pre_div = cfg & mask;
 | |
| 
 | |
| 	mode = cfg & CFG_MODE_MASK;
 | |
| 	mode >>= CFG_MODE_SHIFT;
 | |
| 	if (mode) {
 | |
| 		mask = BIT(rcg->mnd_width) - 1;
 | |
| 		regmap_read(rcg->clkr.regmap,
 | |
| 			    rcg->cmd_rcgr + SE_PERF_M_DFSR(level), &m);
 | |
| 		m &= mask;
 | |
| 
 | |
| 		regmap_read(rcg->clkr.regmap,
 | |
| 			    rcg->cmd_rcgr + SE_PERF_N_DFSR(level), &n);
 | |
| 		n = ~n;
 | |
| 		n &= mask;
 | |
| 		n += m;
 | |
| 	}
 | |
| 
 | |
| 	return calc_rate(parent_rate, m, n, mode, pre_div);
 | |
| }
 | |
| 
 | |
| static const struct clk_ops clk_rcg2_dfs_ops = {
 | |
| 	.is_enabled = clk_rcg2_is_enabled,
 | |
| 	.get_parent = clk_rcg2_get_parent,
 | |
| 	.determine_rate = clk_rcg2_dfs_determine_rate,
 | |
| 	.recalc_rate = clk_rcg2_dfs_recalc_rate,
 | |
| };
 | |
| 
 | |
| static int clk_rcg2_enable_dfs(const struct clk_rcg_dfs_data *data,
 | |
| 			       struct regmap *regmap)
 | |
| {
 | |
| 	struct clk_rcg2 *rcg = data->rcg;
 | |
| 	struct clk_init_data *init = data->init;
 | |
| 	u32 val;
 | |
| 	int ret;
 | |
| 
 | |
| 	ret = regmap_read(regmap, rcg->cmd_rcgr + SE_CMD_DFSR_OFFSET, &val);
 | |
| 	if (ret)
 | |
| 		return -EINVAL;
 | |
| 
 | |
| 	if (!(val & SE_CMD_DFS_EN))
 | |
| 		return 0;
 | |
| 
 | |
| 	/*
 | |
| 	 * Rate changes with consumer writing a register in
 | |
| 	 * their own I/O region
 | |
| 	 */
 | |
| 	init->flags |= CLK_GET_RATE_NOCACHE;
 | |
| 	init->ops = &clk_rcg2_dfs_ops;
 | |
| 
 | |
| 	rcg->freq_tbl = NULL;
 | |
| 
 | |
| 	pr_debug("DFS registered for clk %s\n", init->name);
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| int qcom_cc_register_rcg_dfs(struct regmap *regmap,
 | |
| 			     const struct clk_rcg_dfs_data *rcgs, size_t len)
 | |
| {
 | |
| 	int i, ret;
 | |
| 
 | |
| 	for (i = 0; i < len; i++) {
 | |
| 		ret = clk_rcg2_enable_dfs(&rcgs[i], regmap);
 | |
| 		if (ret) {
 | |
| 			const char *name = rcgs[i].init->name;
 | |
| 
 | |
| 			pr_err("DFS register failed for clk %s\n", name);
 | |
| 			return ret;
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| EXPORT_SYMBOL_GPL(qcom_cc_register_rcg_dfs);
 |