mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-18 15:16:56 +07:00
8e4829c6f7
Hynix ufs has deviations on hi36xx platform which will result in ufs bursts transfer failures. To fix the problem, the Hynix device must set the register VS_DebugSaveConfigTime to 0x10, which will set time reference for SaveConfigTime is 250 ns. The time reference for SaveConfigTime is 40 ns by default. This patch is necessary to boot on HiKey960 boards that use Hynix UFS chips (H28U62301AMR model: hB8aL1). Cc: Vinayak Holikatti <vinholikatti@gmail.com> Cc: "James E.J. Bottomley" <jejb@linux.vnet.ibm.com> Cc: "Martin K. Petersen" <martin.petersen@oracle.com> Cc: linux-scsi@vger.kernel.org Signed-off-by: Wei Li <liwei213@huawei.com> Signed-off-by: Dmitry Shmidt <dimitrysh@google.com> [jstultz: Forward ported from older code, slight tweak to commit message] Signed-off-by: John Stultz <john.stultz@linaro.org> Signed-off-by: Martin K. Petersen <martin.petersen@oracle.com>
629 lines
18 KiB
C
629 lines
18 KiB
C
/*
|
|
* HiSilicon Hixxxx UFS Driver
|
|
*
|
|
* Copyright (c) 2016-2017 Linaro Ltd.
|
|
* Copyright (c) 2016-2017 HiSilicon Technologies Co., Ltd.
|
|
*
|
|
* Released under the GPLv2 only.
|
|
* SPDX-License-Identifier: GPL-2.0
|
|
*/
|
|
|
|
#include <linux/time.h>
|
|
#include <linux/of.h>
|
|
#include <linux/of_address.h>
|
|
#include <linux/dma-mapping.h>
|
|
#include <linux/platform_device.h>
|
|
#include <linux/reset.h>
|
|
|
|
#include "ufshcd.h"
|
|
#include "ufshcd-pltfrm.h"
|
|
#include "unipro.h"
|
|
#include "ufs-hisi.h"
|
|
#include "ufshci.h"
|
|
#include "ufs_quirks.h"
|
|
|
|
static int ufs_hisi_check_hibern8(struct ufs_hba *hba)
|
|
{
|
|
int err = 0;
|
|
u32 tx_fsm_val_0 = 0;
|
|
u32 tx_fsm_val_1 = 0;
|
|
unsigned long timeout = jiffies + msecs_to_jiffies(HBRN8_POLL_TOUT_MS);
|
|
|
|
do {
|
|
err = ufshcd_dme_get(hba, UIC_ARG_MIB_SEL(MPHY_TX_FSM_STATE, 0),
|
|
&tx_fsm_val_0);
|
|
err |= ufshcd_dme_get(hba,
|
|
UIC_ARG_MIB_SEL(MPHY_TX_FSM_STATE, 1), &tx_fsm_val_1);
|
|
if (err || (tx_fsm_val_0 == TX_FSM_HIBERN8 &&
|
|
tx_fsm_val_1 == TX_FSM_HIBERN8))
|
|
break;
|
|
|
|
/* sleep for max. 200us */
|
|
usleep_range(100, 200);
|
|
} while (time_before(jiffies, timeout));
|
|
|
|
/*
|
|
* we might have scheduled out for long during polling so
|
|
* check the state again.
|
|
*/
|
|
if (time_after(jiffies, timeout)) {
|
|
err = ufshcd_dme_get(hba, UIC_ARG_MIB_SEL(MPHY_TX_FSM_STATE, 0),
|
|
&tx_fsm_val_0);
|
|
err |= ufshcd_dme_get(hba,
|
|
UIC_ARG_MIB_SEL(MPHY_TX_FSM_STATE, 1), &tx_fsm_val_1);
|
|
}
|
|
|
|
if (err) {
|
|
dev_err(hba->dev, "%s: unable to get TX_FSM_STATE, err %d\n",
|
|
__func__, err);
|
|
} else if (tx_fsm_val_0 != TX_FSM_HIBERN8 ||
|
|
tx_fsm_val_1 != TX_FSM_HIBERN8) {
|
|
err = -1;
|
|
dev_err(hba->dev, "%s: invalid TX_FSM_STATE, lane0 = %d, lane1 = %d\n",
|
|
__func__, tx_fsm_val_0, tx_fsm_val_1);
|
|
}
|
|
|
|
return err;
|
|
}
|
|
|
|
static void ufs_hi3660_clk_init(struct ufs_hba *hba)
|
|
{
|
|
struct ufs_hisi_host *host = ufshcd_get_variant(hba);
|
|
|
|
ufs_sys_ctrl_clr_bits(host, BIT_SYSCTRL_REF_CLOCK_EN, PHY_CLK_CTRL);
|
|
if (ufs_sys_ctrl_readl(host, PHY_CLK_CTRL) & BIT_SYSCTRL_REF_CLOCK_EN)
|
|
mdelay(1);
|
|
/* use abb clk */
|
|
ufs_sys_ctrl_clr_bits(host, BIT_UFS_REFCLK_SRC_SEl, UFS_SYSCTRL);
|
|
ufs_sys_ctrl_clr_bits(host, BIT_UFS_REFCLK_ISO_EN, PHY_ISO_EN);
|
|
/* open mphy ref clk */
|
|
ufs_sys_ctrl_set_bits(host, BIT_SYSCTRL_REF_CLOCK_EN, PHY_CLK_CTRL);
|
|
}
|
|
|
|
static void ufs_hi3660_soc_init(struct ufs_hba *hba)
|
|
{
|
|
struct ufs_hisi_host *host = ufshcd_get_variant(hba);
|
|
u32 reg;
|
|
|
|
if (!IS_ERR(host->rst))
|
|
reset_control_assert(host->rst);
|
|
|
|
/* HC_PSW powerup */
|
|
ufs_sys_ctrl_set_bits(host, BIT_UFS_PSW_MTCMOS_EN, PSW_POWER_CTRL);
|
|
udelay(10);
|
|
/* notify PWR ready */
|
|
ufs_sys_ctrl_set_bits(host, BIT_SYSCTRL_PWR_READY, HC_LP_CTRL);
|
|
ufs_sys_ctrl_writel(host, MASK_UFS_DEVICE_RESET | 0,
|
|
UFS_DEVICE_RESET_CTRL);
|
|
|
|
reg = ufs_sys_ctrl_readl(host, PHY_CLK_CTRL);
|
|
reg = (reg & ~MASK_SYSCTRL_CFG_CLOCK_FREQ) | UFS_FREQ_CFG_CLK;
|
|
/* set cfg clk freq */
|
|
ufs_sys_ctrl_writel(host, reg, PHY_CLK_CTRL);
|
|
/* set ref clk freq */
|
|
ufs_sys_ctrl_clr_bits(host, MASK_SYSCTRL_REF_CLOCK_SEL, PHY_CLK_CTRL);
|
|
/* bypass ufs clk gate */
|
|
ufs_sys_ctrl_set_bits(host, MASK_UFS_CLK_GATE_BYPASS,
|
|
CLOCK_GATE_BYPASS);
|
|
ufs_sys_ctrl_set_bits(host, MASK_UFS_SYSCRTL_BYPASS, UFS_SYSCTRL);
|
|
|
|
/* open psw clk */
|
|
ufs_sys_ctrl_set_bits(host, BIT_SYSCTRL_PSW_CLK_EN, PSW_CLK_CTRL);
|
|
/* disable ufshc iso */
|
|
ufs_sys_ctrl_clr_bits(host, BIT_UFS_PSW_ISO_CTRL, PSW_POWER_CTRL);
|
|
/* disable phy iso */
|
|
ufs_sys_ctrl_clr_bits(host, BIT_UFS_PHY_ISO_CTRL, PHY_ISO_EN);
|
|
/* notice iso disable */
|
|
ufs_sys_ctrl_clr_bits(host, BIT_SYSCTRL_LP_ISOL_EN, HC_LP_CTRL);
|
|
|
|
/* disable lp_reset_n */
|
|
ufs_sys_ctrl_set_bits(host, BIT_SYSCTRL_LP_RESET_N, RESET_CTRL_EN);
|
|
mdelay(1);
|
|
|
|
ufs_sys_ctrl_writel(host, MASK_UFS_DEVICE_RESET | BIT_UFS_DEVICE_RESET,
|
|
UFS_DEVICE_RESET_CTRL);
|
|
|
|
msleep(20);
|
|
|
|
/*
|
|
* enable the fix of linereset recovery,
|
|
* and enable rx_reset/tx_rest beat
|
|
* enable ref_clk_en override(bit5) &
|
|
* override value = 1(bit4), with mask
|
|
*/
|
|
ufs_sys_ctrl_writel(host, 0x03300330, UFS_DEVICE_RESET_CTRL);
|
|
|
|
if (!IS_ERR(host->rst))
|
|
reset_control_deassert(host->rst);
|
|
}
|
|
|
|
static int ufs_hisi_link_startup_pre_change(struct ufs_hba *hba)
|
|
{
|
|
int err;
|
|
uint32_t value;
|
|
uint32_t reg;
|
|
|
|
/* Unipro VS_mphy_disable */
|
|
ufshcd_dme_set(hba, UIC_ARG_MIB_SEL(0xD0C1, 0x0), 0x1);
|
|
/* PA_HSSeries */
|
|
ufshcd_dme_set(hba, UIC_ARG_MIB_SEL(0x156A, 0x0), 0x2);
|
|
/* MPHY CBRATESEL */
|
|
ufshcd_dme_set(hba, UIC_ARG_MIB_SEL(0x8114, 0x0), 0x1);
|
|
/* MPHY CBOVRCTRL2 */
|
|
ufshcd_dme_set(hba, UIC_ARG_MIB_SEL(0x8121, 0x0), 0x2D);
|
|
/* MPHY CBOVRCTRL3 */
|
|
ufshcd_dme_set(hba, UIC_ARG_MIB_SEL(0x8122, 0x0), 0x1);
|
|
/* Unipro VS_MphyCfgUpdt */
|
|
ufshcd_dme_set(hba, UIC_ARG_MIB_SEL(0xD085, 0x0), 0x1);
|
|
/* MPHY RXOVRCTRL4 rx0 */
|
|
ufshcd_dme_set(hba, UIC_ARG_MIB_SEL(0x800D, 0x4), 0x58);
|
|
/* MPHY RXOVRCTRL4 rx1 */
|
|
ufshcd_dme_set(hba, UIC_ARG_MIB_SEL(0x800D, 0x5), 0x58);
|
|
/* MPHY RXOVRCTRL5 rx0 */
|
|
ufshcd_dme_set(hba, UIC_ARG_MIB_SEL(0x800E, 0x4), 0xB);
|
|
/* MPHY RXOVRCTRL5 rx1 */
|
|
ufshcd_dme_set(hba, UIC_ARG_MIB_SEL(0x800E, 0x5), 0xB);
|
|
/* MPHY RXSQCONTROL rx0 */
|
|
ufshcd_dme_set(hba, UIC_ARG_MIB_SEL(0x8009, 0x4), 0x1);
|
|
/* MPHY RXSQCONTROL rx1 */
|
|
ufshcd_dme_set(hba, UIC_ARG_MIB_SEL(0x8009, 0x5), 0x1);
|
|
/* Unipro VS_MphyCfgUpdt */
|
|
ufshcd_dme_set(hba, UIC_ARG_MIB_SEL(0xD085, 0x0), 0x1);
|
|
|
|
ufshcd_dme_set(hba, UIC_ARG_MIB_SEL(0x8113, 0x0), 0x1);
|
|
ufshcd_dme_set(hba, UIC_ARG_MIB_SEL(0xD085, 0x0), 0x1);
|
|
|
|
/* Tactive RX */
|
|
ufshcd_dme_set(hba, UIC_ARG_MIB_SEL(0x008F, 0x4), 0x7);
|
|
/* Tactive RX */
|
|
ufshcd_dme_set(hba, UIC_ARG_MIB_SEL(0x008F, 0x5), 0x7);
|
|
|
|
/* Gear3 Synclength */
|
|
ufshcd_dme_set(hba, UIC_ARG_MIB_SEL(0x0095, 0x4), 0x4F);
|
|
/* Gear3 Synclength */
|
|
ufshcd_dme_set(hba, UIC_ARG_MIB_SEL(0x0095, 0x5), 0x4F);
|
|
/* Gear2 Synclength */
|
|
ufshcd_dme_set(hba, UIC_ARG_MIB_SEL(0x0094, 0x4), 0x4F);
|
|
/* Gear2 Synclength */
|
|
ufshcd_dme_set(hba, UIC_ARG_MIB_SEL(0x0094, 0x5), 0x4F);
|
|
/* Gear1 Synclength */
|
|
ufshcd_dme_set(hba, UIC_ARG_MIB_SEL(0x008B, 0x4), 0x4F);
|
|
/* Gear1 Synclength */
|
|
ufshcd_dme_set(hba, UIC_ARG_MIB_SEL(0x008B, 0x5), 0x4F);
|
|
/* Thibernate Tx */
|
|
ufshcd_dme_set(hba, UIC_ARG_MIB_SEL(0x000F, 0x0), 0x5);
|
|
/* Thibernate Tx */
|
|
ufshcd_dme_set(hba, UIC_ARG_MIB_SEL(0x000F, 0x1), 0x5);
|
|
|
|
ufshcd_dme_set(hba, UIC_ARG_MIB_SEL(0xD085, 0x0), 0x1);
|
|
/* Unipro VS_mphy_disable */
|
|
ufshcd_dme_get(hba, UIC_ARG_MIB_SEL(0xD0C1, 0x0), &value);
|
|
if (value != 0x1)
|
|
dev_info(hba->dev,
|
|
"Warring!!! Unipro VS_mphy_disable is 0x%x\n", value);
|
|
|
|
/* Unipro VS_mphy_disable */
|
|
ufshcd_dme_set(hba, UIC_ARG_MIB_SEL(0xD0C1, 0x0), 0x0);
|
|
err = ufs_hisi_check_hibern8(hba);
|
|
if (err)
|
|
dev_err(hba->dev, "ufs_hisi_check_hibern8 error\n");
|
|
|
|
ufshcd_writel(hba, UFS_HCLKDIV_NORMAL_VALUE, UFS_REG_HCLKDIV);
|
|
|
|
/* disable auto H8 */
|
|
reg = ufshcd_readl(hba, REG_AUTO_HIBERNATE_IDLE_TIMER);
|
|
reg = reg & (~UFS_AHIT_AH8ITV_MASK);
|
|
ufshcd_writel(hba, reg, REG_AUTO_HIBERNATE_IDLE_TIMER);
|
|
|
|
/* Unipro PA_Local_TX_LCC_Enable */
|
|
ufshcd_dme_set(hba, UIC_ARG_MIB_SEL(0x155E, 0x0), 0x0);
|
|
/* close Unipro VS_Mk2ExtnSupport */
|
|
ufshcd_dme_set(hba, UIC_ARG_MIB_SEL(0xD0AB, 0x0), 0x0);
|
|
ufshcd_dme_get(hba, UIC_ARG_MIB_SEL(0xD0AB, 0x0), &value);
|
|
if (value != 0) {
|
|
/* Ensure close success */
|
|
dev_info(hba->dev, "WARN: close VS_Mk2ExtnSupport failed\n");
|
|
}
|
|
|
|
return err;
|
|
}
|
|
|
|
static int ufs_hisi_link_startup_post_change(struct ufs_hba *hba)
|
|
{
|
|
struct ufs_hisi_host *host = ufshcd_get_variant(hba);
|
|
|
|
/* Unipro DL_AFC0CreditThreshold */
|
|
ufshcd_dme_set(hba, UIC_ARG_MIB(0x2044), 0x0);
|
|
/* Unipro DL_TC0OutAckThreshold */
|
|
ufshcd_dme_set(hba, UIC_ARG_MIB(0x2045), 0x0);
|
|
/* Unipro DL_TC0TXFCThreshold */
|
|
ufshcd_dme_set(hba, UIC_ARG_MIB(0x2040), 0x9);
|
|
|
|
/* not bypass ufs clk gate */
|
|
ufs_sys_ctrl_clr_bits(host, MASK_UFS_CLK_GATE_BYPASS,
|
|
CLOCK_GATE_BYPASS);
|
|
ufs_sys_ctrl_clr_bits(host, MASK_UFS_SYSCRTL_BYPASS,
|
|
UFS_SYSCTRL);
|
|
|
|
/* select received symbol cnt */
|
|
ufshcd_dme_set(hba, UIC_ARG_MIB(0xd09a), 0x80000000);
|
|
/* reset counter0 and enable */
|
|
ufshcd_dme_set(hba, UIC_ARG_MIB(0xd09c), 0x00000005);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int ufs_hi3660_link_startup_notify(struct ufs_hba *hba,
|
|
enum ufs_notify_change_status status)
|
|
{
|
|
int err = 0;
|
|
|
|
switch (status) {
|
|
case PRE_CHANGE:
|
|
err = ufs_hisi_link_startup_pre_change(hba);
|
|
break;
|
|
case POST_CHANGE:
|
|
err = ufs_hisi_link_startup_post_change(hba);
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
|
|
return err;
|
|
}
|
|
|
|
struct ufs_hisi_dev_params {
|
|
u32 pwm_rx_gear; /* pwm rx gear to work in */
|
|
u32 pwm_tx_gear; /* pwm tx gear to work in */
|
|
u32 hs_rx_gear; /* hs rx gear to work in */
|
|
u32 hs_tx_gear; /* hs tx gear to work in */
|
|
u32 rx_lanes; /* number of rx lanes */
|
|
u32 tx_lanes; /* number of tx lanes */
|
|
u32 rx_pwr_pwm; /* rx pwm working pwr */
|
|
u32 tx_pwr_pwm; /* tx pwm working pwr */
|
|
u32 rx_pwr_hs; /* rx hs working pwr */
|
|
u32 tx_pwr_hs; /* tx hs working pwr */
|
|
u32 hs_rate; /* rate A/B to work in HS */
|
|
u32 desired_working_mode;
|
|
};
|
|
|
|
static int ufs_hisi_get_pwr_dev_param(
|
|
struct ufs_hisi_dev_params *hisi_param,
|
|
struct ufs_pa_layer_attr *dev_max,
|
|
struct ufs_pa_layer_attr *agreed_pwr)
|
|
{
|
|
int min_hisi_gear;
|
|
int min_dev_gear;
|
|
bool is_dev_sup_hs = false;
|
|
bool is_hisi_max_hs = false;
|
|
|
|
if (dev_max->pwr_rx == FASTAUTO_MODE || dev_max->pwr_rx == FAST_MODE)
|
|
is_dev_sup_hs = true;
|
|
|
|
if (hisi_param->desired_working_mode == FAST) {
|
|
is_hisi_max_hs = true;
|
|
min_hisi_gear = min_t(u32, hisi_param->hs_rx_gear,
|
|
hisi_param->hs_tx_gear);
|
|
} else {
|
|
min_hisi_gear = min_t(u32, hisi_param->pwm_rx_gear,
|
|
hisi_param->pwm_tx_gear);
|
|
}
|
|
|
|
/*
|
|
* device doesn't support HS but
|
|
* hisi_param->desired_working_mode is HS,
|
|
* thus device and hisi_param don't agree
|
|
*/
|
|
if (!is_dev_sup_hs && is_hisi_max_hs) {
|
|
pr_err("%s: device not support HS\n", __func__);
|
|
return -ENOTSUPP;
|
|
} else if (is_dev_sup_hs && is_hisi_max_hs) {
|
|
/*
|
|
* since device supports HS, it supports FAST_MODE.
|
|
* since hisi_param->desired_working_mode is also HS
|
|
* then final decision (FAST/FASTAUTO) is done according
|
|
* to hisi_params as it is the restricting factor
|
|
*/
|
|
agreed_pwr->pwr_rx = agreed_pwr->pwr_tx =
|
|
hisi_param->rx_pwr_hs;
|
|
} else {
|
|
/*
|
|
* here hisi_param->desired_working_mode is PWM.
|
|
* it doesn't matter whether device supports HS or PWM,
|
|
* in both cases hisi_param->desired_working_mode will
|
|
* determine the mode
|
|
*/
|
|
agreed_pwr->pwr_rx = agreed_pwr->pwr_tx =
|
|
hisi_param->rx_pwr_pwm;
|
|
}
|
|
|
|
/*
|
|
* we would like tx to work in the minimum number of lanes
|
|
* between device capability and vendor preferences.
|
|
* the same decision will be made for rx
|
|
*/
|
|
agreed_pwr->lane_tx =
|
|
min_t(u32, dev_max->lane_tx, hisi_param->tx_lanes);
|
|
agreed_pwr->lane_rx =
|
|
min_t(u32, dev_max->lane_rx, hisi_param->rx_lanes);
|
|
|
|
/* device maximum gear is the minimum between device rx and tx gears */
|
|
min_dev_gear = min_t(u32, dev_max->gear_rx, dev_max->gear_tx);
|
|
|
|
/*
|
|
* if both device capabilities and vendor pre-defined preferences are
|
|
* both HS or both PWM then set the minimum gear to be the chosen
|
|
* working gear.
|
|
* if one is PWM and one is HS then the one that is PWM get to decide
|
|
* what is the gear, as it is the one that also decided previously what
|
|
* pwr the device will be configured to.
|
|
*/
|
|
if ((is_dev_sup_hs && is_hisi_max_hs) ||
|
|
(!is_dev_sup_hs && !is_hisi_max_hs))
|
|
agreed_pwr->gear_rx = agreed_pwr->gear_tx =
|
|
min_t(u32, min_dev_gear, min_hisi_gear);
|
|
else
|
|
agreed_pwr->gear_rx = agreed_pwr->gear_tx = min_hisi_gear;
|
|
|
|
agreed_pwr->hs_rate = hisi_param->hs_rate;
|
|
|
|
pr_info("ufs final power mode: gear = %d, lane = %d, pwr = %d, rate = %d\n",
|
|
agreed_pwr->gear_rx, agreed_pwr->lane_rx, agreed_pwr->pwr_rx,
|
|
agreed_pwr->hs_rate);
|
|
return 0;
|
|
}
|
|
|
|
static void ufs_hisi_set_dev_cap(struct ufs_hisi_dev_params *hisi_param)
|
|
{
|
|
hisi_param->rx_lanes = UFS_HISI_LIMIT_NUM_LANES_RX;
|
|
hisi_param->tx_lanes = UFS_HISI_LIMIT_NUM_LANES_TX;
|
|
hisi_param->hs_rx_gear = UFS_HISI_LIMIT_HSGEAR_RX;
|
|
hisi_param->hs_tx_gear = UFS_HISI_LIMIT_HSGEAR_TX;
|
|
hisi_param->pwm_rx_gear = UFS_HISI_LIMIT_PWMGEAR_RX;
|
|
hisi_param->pwm_tx_gear = UFS_HISI_LIMIT_PWMGEAR_TX;
|
|
hisi_param->rx_pwr_pwm = UFS_HISI_LIMIT_RX_PWR_PWM;
|
|
hisi_param->tx_pwr_pwm = UFS_HISI_LIMIT_TX_PWR_PWM;
|
|
hisi_param->rx_pwr_hs = UFS_HISI_LIMIT_RX_PWR_HS;
|
|
hisi_param->tx_pwr_hs = UFS_HISI_LIMIT_TX_PWR_HS;
|
|
hisi_param->hs_rate = UFS_HISI_LIMIT_HS_RATE;
|
|
hisi_param->desired_working_mode = UFS_HISI_LIMIT_DESIRED_MODE;
|
|
}
|
|
|
|
static void ufs_hisi_pwr_change_pre_change(struct ufs_hba *hba)
|
|
{
|
|
if (hba->dev_quirks & UFS_DEVICE_QUIRK_HOST_VS_DEBUGSAVECONFIGTIME) {
|
|
pr_info("ufs flash device must set VS_DebugSaveConfigTime 0x10\n");
|
|
/* VS_DebugSaveConfigTime */
|
|
ufshcd_dme_set(hba, UIC_ARG_MIB(0xD0A0), 0x10);
|
|
/* sync length */
|
|
ufshcd_dme_set(hba, UIC_ARG_MIB(0x1556), 0x48);
|
|
}
|
|
|
|
/* update */
|
|
ufshcd_dme_set(hba, UIC_ARG_MIB(0x15A8), 0x1);
|
|
/* PA_TxSkip */
|
|
ufshcd_dme_set(hba, UIC_ARG_MIB(0x155c), 0x0);
|
|
/*PA_PWRModeUserData0 = 8191, default is 0*/
|
|
ufshcd_dme_set(hba, UIC_ARG_MIB(0x15b0), 8191);
|
|
/*PA_PWRModeUserData1 = 65535, default is 0*/
|
|
ufshcd_dme_set(hba, UIC_ARG_MIB(0x15b1), 65535);
|
|
/*PA_PWRModeUserData2 = 32767, default is 0*/
|
|
ufshcd_dme_set(hba, UIC_ARG_MIB(0x15b2), 32767);
|
|
/*DME_FC0ProtectionTimeOutVal = 8191, default is 0*/
|
|
ufshcd_dme_set(hba, UIC_ARG_MIB(0xd041), 8191);
|
|
/*DME_TC0ReplayTimeOutVal = 65535, default is 0*/
|
|
ufshcd_dme_set(hba, UIC_ARG_MIB(0xd042), 65535);
|
|
/*DME_AFC0ReqTimeOutVal = 32767, default is 0*/
|
|
ufshcd_dme_set(hba, UIC_ARG_MIB(0xd043), 32767);
|
|
/*PA_PWRModeUserData3 = 8191, default is 0*/
|
|
ufshcd_dme_set(hba, UIC_ARG_MIB(0x15b3), 8191);
|
|
/*PA_PWRModeUserData4 = 65535, default is 0*/
|
|
ufshcd_dme_set(hba, UIC_ARG_MIB(0x15b4), 65535);
|
|
/*PA_PWRModeUserData5 = 32767, default is 0*/
|
|
ufshcd_dme_set(hba, UIC_ARG_MIB(0x15b5), 32767);
|
|
/*DME_FC1ProtectionTimeOutVal = 8191, default is 0*/
|
|
ufshcd_dme_set(hba, UIC_ARG_MIB(0xd044), 8191);
|
|
/*DME_TC1ReplayTimeOutVal = 65535, default is 0*/
|
|
ufshcd_dme_set(hba, UIC_ARG_MIB(0xd045), 65535);
|
|
/*DME_AFC1ReqTimeOutVal = 32767, default is 0*/
|
|
ufshcd_dme_set(hba, UIC_ARG_MIB(0xd046), 32767);
|
|
}
|
|
|
|
static int ufs_hi3660_pwr_change_notify(struct ufs_hba *hba,
|
|
enum ufs_notify_change_status status,
|
|
struct ufs_pa_layer_attr *dev_max_params,
|
|
struct ufs_pa_layer_attr *dev_req_params)
|
|
{
|
|
struct ufs_hisi_dev_params ufs_hisi_cap;
|
|
int ret = 0;
|
|
|
|
if (!dev_req_params) {
|
|
dev_err(hba->dev,
|
|
"%s: incoming dev_req_params is NULL\n", __func__);
|
|
ret = -EINVAL;
|
|
goto out;
|
|
}
|
|
|
|
switch (status) {
|
|
case PRE_CHANGE:
|
|
ufs_hisi_set_dev_cap(&ufs_hisi_cap);
|
|
ret = ufs_hisi_get_pwr_dev_param(
|
|
&ufs_hisi_cap, dev_max_params, dev_req_params);
|
|
if (ret) {
|
|
dev_err(hba->dev,
|
|
"%s: failed to determine capabilities\n", __func__);
|
|
goto out;
|
|
}
|
|
|
|
ufs_hisi_pwr_change_pre_change(hba);
|
|
break;
|
|
case POST_CHANGE:
|
|
break;
|
|
default:
|
|
ret = -EINVAL;
|
|
break;
|
|
}
|
|
out:
|
|
return ret;
|
|
}
|
|
|
|
static int ufs_hisi_suspend(struct ufs_hba *hba, enum ufs_pm_op pm_op)
|
|
{
|
|
struct ufs_hisi_host *host = ufshcd_get_variant(hba);
|
|
|
|
if (ufshcd_is_runtime_pm(pm_op))
|
|
return 0;
|
|
|
|
if (host->in_suspend) {
|
|
WARN_ON(1);
|
|
return 0;
|
|
}
|
|
|
|
ufs_sys_ctrl_clr_bits(host, BIT_SYSCTRL_REF_CLOCK_EN, PHY_CLK_CTRL);
|
|
udelay(10);
|
|
/* set ref_dig_clk override of PHY PCS to 0 */
|
|
ufs_sys_ctrl_writel(host, 0x00100000, UFS_DEVICE_RESET_CTRL);
|
|
|
|
host->in_suspend = true;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int ufs_hisi_resume(struct ufs_hba *hba, enum ufs_pm_op pm_op)
|
|
{
|
|
struct ufs_hisi_host *host = ufshcd_get_variant(hba);
|
|
|
|
if (!host->in_suspend)
|
|
return 0;
|
|
|
|
/* set ref_dig_clk override of PHY PCS to 1 */
|
|
ufs_sys_ctrl_writel(host, 0x00100010, UFS_DEVICE_RESET_CTRL);
|
|
udelay(10);
|
|
ufs_sys_ctrl_set_bits(host, BIT_SYSCTRL_REF_CLOCK_EN, PHY_CLK_CTRL);
|
|
|
|
host->in_suspend = false;
|
|
return 0;
|
|
}
|
|
|
|
static int ufs_hisi_get_resource(struct ufs_hisi_host *host)
|
|
{
|
|
struct resource *mem_res;
|
|
struct device *dev = host->hba->dev;
|
|
struct platform_device *pdev = to_platform_device(dev);
|
|
|
|
/* get resource of ufs sys ctrl */
|
|
mem_res = platform_get_resource(pdev, IORESOURCE_MEM, 1);
|
|
host->ufs_sys_ctrl = devm_ioremap_resource(dev, mem_res);
|
|
if (IS_ERR(host->ufs_sys_ctrl))
|
|
return PTR_ERR(host->ufs_sys_ctrl);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void ufs_hisi_set_pm_lvl(struct ufs_hba *hba)
|
|
{
|
|
hba->rpm_lvl = UFS_PM_LVL_1;
|
|
hba->spm_lvl = UFS_PM_LVL_3;
|
|
}
|
|
|
|
/**
|
|
* ufs_hisi_init_common
|
|
* @hba: host controller instance
|
|
*/
|
|
static int ufs_hisi_init_common(struct ufs_hba *hba)
|
|
{
|
|
int err = 0;
|
|
struct device *dev = hba->dev;
|
|
struct ufs_hisi_host *host;
|
|
|
|
host = devm_kzalloc(dev, sizeof(*host), GFP_KERNEL);
|
|
if (!host)
|
|
return -ENOMEM;
|
|
|
|
host->hba = hba;
|
|
ufshcd_set_variant(hba, host);
|
|
|
|
host->rst = devm_reset_control_get(dev, "rst");
|
|
|
|
ufs_hisi_set_pm_lvl(hba);
|
|
|
|
err = ufs_hisi_get_resource(host);
|
|
if (err) {
|
|
ufshcd_set_variant(hba, NULL);
|
|
return err;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int ufs_hi3660_init(struct ufs_hba *hba)
|
|
{
|
|
int ret = 0;
|
|
struct device *dev = hba->dev;
|
|
|
|
ret = ufs_hisi_init_common(hba);
|
|
if (ret) {
|
|
dev_err(dev, "%s: ufs common init fail\n", __func__);
|
|
return ret;
|
|
}
|
|
|
|
ufs_hi3660_clk_init(hba);
|
|
|
|
ufs_hi3660_soc_init(hba);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static struct ufs_hba_variant_ops ufs_hba_hisi_vops = {
|
|
.name = "hi3660",
|
|
.init = ufs_hi3660_init,
|
|
.link_startup_notify = ufs_hi3660_link_startup_notify,
|
|
.pwr_change_notify = ufs_hi3660_pwr_change_notify,
|
|
.suspend = ufs_hisi_suspend,
|
|
.resume = ufs_hisi_resume,
|
|
};
|
|
|
|
static int ufs_hisi_probe(struct platform_device *pdev)
|
|
{
|
|
return ufshcd_pltfrm_init(pdev, &ufs_hba_hisi_vops);
|
|
}
|
|
|
|
static int ufs_hisi_remove(struct platform_device *pdev)
|
|
{
|
|
struct ufs_hba *hba = platform_get_drvdata(pdev);
|
|
|
|
ufshcd_remove(hba);
|
|
return 0;
|
|
}
|
|
|
|
static const struct of_device_id ufs_hisi_of_match[] = {
|
|
{ .compatible = "hisilicon,hi3660-ufs" },
|
|
{},
|
|
};
|
|
|
|
MODULE_DEVICE_TABLE(of, ufs_hisi_of_match);
|
|
|
|
static const struct dev_pm_ops ufs_hisi_pm_ops = {
|
|
.suspend = ufshcd_pltfrm_suspend,
|
|
.resume = ufshcd_pltfrm_resume,
|
|
.runtime_suspend = ufshcd_pltfrm_runtime_suspend,
|
|
.runtime_resume = ufshcd_pltfrm_runtime_resume,
|
|
.runtime_idle = ufshcd_pltfrm_runtime_idle,
|
|
};
|
|
|
|
static struct platform_driver ufs_hisi_pltform = {
|
|
.probe = ufs_hisi_probe,
|
|
.remove = ufs_hisi_remove,
|
|
.shutdown = ufshcd_pltfrm_shutdown,
|
|
.driver = {
|
|
.name = "ufshcd-hisi",
|
|
.pm = &ufs_hisi_pm_ops,
|
|
.of_match_table = of_match_ptr(ufs_hisi_of_match),
|
|
},
|
|
};
|
|
module_platform_driver(ufs_hisi_pltform);
|
|
|
|
MODULE_LICENSE("GPL");
|
|
MODULE_ALIAS("platform:ufshcd-hisi");
|
|
MODULE_DESCRIPTION("HiSilicon Hixxxx UFS Driver");
|