508 строки
14 KiB
C
508 строки
14 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
/*
|
|
* Copyright (c) 2018-2020, The Linux Foundation. All rights reserved.
|
|
*
|
|
*/
|
|
|
|
#include <linux/delay.h>
|
|
#include <linux/device.h>
|
|
#include <linux/dma-direction.h>
|
|
#include <linux/dma-mapping.h>
|
|
#include <linux/firmware.h>
|
|
#include <linux/interrupt.h>
|
|
#include <linux/list.h>
|
|
#include <linux/mhi.h>
|
|
#include <linux/module.h>
|
|
#include <linux/random.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/wait.h>
|
|
#include "internal.h"
|
|
|
|
/* Setup RDDM vector table for RDDM transfer and program RXVEC */
|
|
void mhi_rddm_prepare(struct mhi_controller *mhi_cntrl,
|
|
struct image_info *img_info)
|
|
{
|
|
struct mhi_buf *mhi_buf = img_info->mhi_buf;
|
|
struct bhi_vec_entry *bhi_vec = img_info->bhi_vec;
|
|
void __iomem *base = mhi_cntrl->bhie;
|
|
struct device *dev = &mhi_cntrl->mhi_dev->dev;
|
|
u32 sequence_id;
|
|
unsigned int i;
|
|
|
|
for (i = 0; i < img_info->entries - 1; i++, mhi_buf++, bhi_vec++) {
|
|
bhi_vec->dma_addr = mhi_buf->dma_addr;
|
|
bhi_vec->size = mhi_buf->len;
|
|
}
|
|
|
|
dev_dbg(dev, "BHIe programming for RDDM\n");
|
|
|
|
mhi_write_reg(mhi_cntrl, base, BHIE_RXVECADDR_HIGH_OFFS,
|
|
upper_32_bits(mhi_buf->dma_addr));
|
|
|
|
mhi_write_reg(mhi_cntrl, base, BHIE_RXVECADDR_LOW_OFFS,
|
|
lower_32_bits(mhi_buf->dma_addr));
|
|
|
|
mhi_write_reg(mhi_cntrl, base, BHIE_RXVECSIZE_OFFS, mhi_buf->len);
|
|
sequence_id = prandom_u32() & BHIE_RXVECSTATUS_SEQNUM_BMSK;
|
|
|
|
if (unlikely(!sequence_id))
|
|
sequence_id = 1;
|
|
|
|
mhi_write_reg_field(mhi_cntrl, base, BHIE_RXVECDB_OFFS,
|
|
BHIE_RXVECDB_SEQNUM_BMSK, BHIE_RXVECDB_SEQNUM_SHFT,
|
|
sequence_id);
|
|
|
|
dev_dbg(dev, "Address: %p and len: 0x%zx sequence: %u\n",
|
|
&mhi_buf->dma_addr, mhi_buf->len, sequence_id);
|
|
}
|
|
|
|
/* Collect RDDM buffer during kernel panic */
|
|
static int __mhi_download_rddm_in_panic(struct mhi_controller *mhi_cntrl)
|
|
{
|
|
int ret;
|
|
u32 rx_status;
|
|
enum mhi_ee_type ee;
|
|
const u32 delayus = 2000;
|
|
u32 retry = (mhi_cntrl->timeout_ms * 1000) / delayus;
|
|
const u32 rddm_timeout_us = 200000;
|
|
int rddm_retry = rddm_timeout_us / delayus;
|
|
void __iomem *base = mhi_cntrl->bhie;
|
|
struct device *dev = &mhi_cntrl->mhi_dev->dev;
|
|
|
|
dev_dbg(dev, "Entered with pm_state:%s dev_state:%s ee:%s\n",
|
|
to_mhi_pm_state_str(mhi_cntrl->pm_state),
|
|
TO_MHI_STATE_STR(mhi_cntrl->dev_state),
|
|
TO_MHI_EXEC_STR(mhi_cntrl->ee));
|
|
|
|
/*
|
|
* This should only be executing during a kernel panic, we expect all
|
|
* other cores to shutdown while we're collecting RDDM buffer. After
|
|
* returning from this function, we expect the device to reset.
|
|
*
|
|
* Normaly, we read/write pm_state only after grabbing the
|
|
* pm_lock, since we're in a panic, skipping it. Also there is no
|
|
* gurantee that this state change would take effect since
|
|
* we're setting it w/o grabbing pm_lock
|
|
*/
|
|
mhi_cntrl->pm_state = MHI_PM_LD_ERR_FATAL_DETECT;
|
|
/* update should take the effect immediately */
|
|
smp_wmb();
|
|
|
|
/*
|
|
* Make sure device is not already in RDDM. In case the device asserts
|
|
* and a kernel panic follows, device will already be in RDDM.
|
|
* Do not trigger SYS ERR again and proceed with waiting for
|
|
* image download completion.
|
|
*/
|
|
ee = mhi_get_exec_env(mhi_cntrl);
|
|
if (ee != MHI_EE_RDDM) {
|
|
dev_dbg(dev, "Trigger device into RDDM mode using SYS ERR\n");
|
|
mhi_set_mhi_state(mhi_cntrl, MHI_STATE_SYS_ERR);
|
|
|
|
dev_dbg(dev, "Waiting for device to enter RDDM\n");
|
|
while (rddm_retry--) {
|
|
ee = mhi_get_exec_env(mhi_cntrl);
|
|
if (ee == MHI_EE_RDDM)
|
|
break;
|
|
|
|
udelay(delayus);
|
|
}
|
|
|
|
if (rddm_retry <= 0) {
|
|
/* Hardware reset so force device to enter RDDM */
|
|
dev_dbg(dev,
|
|
"Did not enter RDDM, do a host req reset\n");
|
|
mhi_write_reg(mhi_cntrl, mhi_cntrl->regs,
|
|
MHI_SOC_RESET_REQ_OFFSET,
|
|
MHI_SOC_RESET_REQ);
|
|
udelay(delayus);
|
|
}
|
|
|
|
ee = mhi_get_exec_env(mhi_cntrl);
|
|
}
|
|
|
|
dev_dbg(dev, "Waiting for image download completion, current EE: %s\n",
|
|
TO_MHI_EXEC_STR(ee));
|
|
|
|
while (retry--) {
|
|
ret = mhi_read_reg_field(mhi_cntrl, base, BHIE_RXVECSTATUS_OFFS,
|
|
BHIE_RXVECSTATUS_STATUS_BMSK,
|
|
BHIE_RXVECSTATUS_STATUS_SHFT,
|
|
&rx_status);
|
|
if (ret)
|
|
return -EIO;
|
|
|
|
if (rx_status == BHIE_RXVECSTATUS_STATUS_XFER_COMPL)
|
|
return 0;
|
|
|
|
udelay(delayus);
|
|
}
|
|
|
|
ee = mhi_get_exec_env(mhi_cntrl);
|
|
ret = mhi_read_reg(mhi_cntrl, base, BHIE_RXVECSTATUS_OFFS, &rx_status);
|
|
|
|
dev_err(dev, "Did not complete RDDM transfer\n");
|
|
dev_err(dev, "Current EE: %s\n", TO_MHI_EXEC_STR(ee));
|
|
dev_err(dev, "RXVEC_STATUS: 0x%x\n", rx_status);
|
|
|
|
return -EIO;
|
|
}
|
|
|
|
/* Download RDDM image from device */
|
|
int mhi_download_rddm_img(struct mhi_controller *mhi_cntrl, bool in_panic)
|
|
{
|
|
void __iomem *base = mhi_cntrl->bhie;
|
|
u32 rx_status;
|
|
|
|
if (in_panic)
|
|
return __mhi_download_rddm_in_panic(mhi_cntrl);
|
|
|
|
/* Wait for the image download to complete */
|
|
wait_event_timeout(mhi_cntrl->state_event,
|
|
mhi_read_reg_field(mhi_cntrl, base,
|
|
BHIE_RXVECSTATUS_OFFS,
|
|
BHIE_RXVECSTATUS_STATUS_BMSK,
|
|
BHIE_RXVECSTATUS_STATUS_SHFT,
|
|
&rx_status) || rx_status,
|
|
msecs_to_jiffies(mhi_cntrl->timeout_ms));
|
|
|
|
return (rx_status == BHIE_RXVECSTATUS_STATUS_XFER_COMPL) ? 0 : -EIO;
|
|
}
|
|
EXPORT_SYMBOL_GPL(mhi_download_rddm_img);
|
|
|
|
static int mhi_fw_load_amss(struct mhi_controller *mhi_cntrl,
|
|
const struct mhi_buf *mhi_buf)
|
|
{
|
|
void __iomem *base = mhi_cntrl->bhie;
|
|
rwlock_t *pm_lock = &mhi_cntrl->pm_lock;
|
|
u32 tx_status, sequence_id;
|
|
|
|
read_lock_bh(pm_lock);
|
|
if (!MHI_REG_ACCESS_VALID(mhi_cntrl->pm_state)) {
|
|
read_unlock_bh(pm_lock);
|
|
return -EIO;
|
|
}
|
|
|
|
mhi_write_reg(mhi_cntrl, base, BHIE_TXVECADDR_HIGH_OFFS,
|
|
upper_32_bits(mhi_buf->dma_addr));
|
|
|
|
mhi_write_reg(mhi_cntrl, base, BHIE_TXVECADDR_LOW_OFFS,
|
|
lower_32_bits(mhi_buf->dma_addr));
|
|
|
|
mhi_write_reg(mhi_cntrl, base, BHIE_TXVECSIZE_OFFS, mhi_buf->len);
|
|
|
|
sequence_id = prandom_u32() & BHIE_TXVECSTATUS_SEQNUM_BMSK;
|
|
mhi_write_reg_field(mhi_cntrl, base, BHIE_TXVECDB_OFFS,
|
|
BHIE_TXVECDB_SEQNUM_BMSK, BHIE_TXVECDB_SEQNUM_SHFT,
|
|
sequence_id);
|
|
read_unlock_bh(pm_lock);
|
|
|
|
/* Wait for the image download to complete */
|
|
wait_event_timeout(mhi_cntrl->state_event,
|
|
MHI_PM_IN_ERROR_STATE(mhi_cntrl->pm_state) ||
|
|
mhi_read_reg_field(mhi_cntrl, base,
|
|
BHIE_TXVECSTATUS_OFFS,
|
|
BHIE_TXVECSTATUS_STATUS_BMSK,
|
|
BHIE_TXVECSTATUS_STATUS_SHFT,
|
|
&tx_status) || tx_status,
|
|
msecs_to_jiffies(mhi_cntrl->timeout_ms));
|
|
|
|
if (MHI_PM_IN_ERROR_STATE(mhi_cntrl->pm_state))
|
|
return -EIO;
|
|
|
|
return (tx_status == BHIE_TXVECSTATUS_STATUS_XFER_COMPL) ? 0 : -EIO;
|
|
}
|
|
|
|
static int mhi_fw_load_sbl(struct mhi_controller *mhi_cntrl,
|
|
dma_addr_t dma_addr,
|
|
size_t size)
|
|
{
|
|
u32 tx_status, val, session_id;
|
|
int i, ret;
|
|
void __iomem *base = mhi_cntrl->bhi;
|
|
rwlock_t *pm_lock = &mhi_cntrl->pm_lock;
|
|
struct device *dev = &mhi_cntrl->mhi_dev->dev;
|
|
struct {
|
|
char *name;
|
|
u32 offset;
|
|
} error_reg[] = {
|
|
{ "ERROR_CODE", BHI_ERRCODE },
|
|
{ "ERROR_DBG1", BHI_ERRDBG1 },
|
|
{ "ERROR_DBG2", BHI_ERRDBG2 },
|
|
{ "ERROR_DBG3", BHI_ERRDBG3 },
|
|
{ NULL },
|
|
};
|
|
|
|
read_lock_bh(pm_lock);
|
|
if (!MHI_REG_ACCESS_VALID(mhi_cntrl->pm_state)) {
|
|
read_unlock_bh(pm_lock);
|
|
goto invalid_pm_state;
|
|
}
|
|
|
|
dev_dbg(dev, "Starting SBL download via BHI\n");
|
|
mhi_write_reg(mhi_cntrl, base, BHI_STATUS, 0);
|
|
mhi_write_reg(mhi_cntrl, base, BHI_IMGADDR_HIGH,
|
|
upper_32_bits(dma_addr));
|
|
mhi_write_reg(mhi_cntrl, base, BHI_IMGADDR_LOW,
|
|
lower_32_bits(dma_addr));
|
|
mhi_write_reg(mhi_cntrl, base, BHI_IMGSIZE, size);
|
|
session_id = prandom_u32() & BHI_TXDB_SEQNUM_BMSK;
|
|
mhi_write_reg(mhi_cntrl, base, BHI_IMGTXDB, session_id);
|
|
read_unlock_bh(pm_lock);
|
|
|
|
/* Wait for the image download to complete */
|
|
ret = wait_event_timeout(mhi_cntrl->state_event,
|
|
MHI_PM_IN_ERROR_STATE(mhi_cntrl->pm_state) ||
|
|
mhi_read_reg_field(mhi_cntrl, base, BHI_STATUS,
|
|
BHI_STATUS_MASK, BHI_STATUS_SHIFT,
|
|
&tx_status) || tx_status,
|
|
msecs_to_jiffies(mhi_cntrl->timeout_ms));
|
|
if (MHI_PM_IN_ERROR_STATE(mhi_cntrl->pm_state))
|
|
goto invalid_pm_state;
|
|
|
|
if (tx_status == BHI_STATUS_ERROR) {
|
|
dev_err(dev, "Image transfer failed\n");
|
|
read_lock_bh(pm_lock);
|
|
if (MHI_REG_ACCESS_VALID(mhi_cntrl->pm_state)) {
|
|
for (i = 0; error_reg[i].name; i++) {
|
|
ret = mhi_read_reg(mhi_cntrl, base,
|
|
error_reg[i].offset, &val);
|
|
if (ret)
|
|
break;
|
|
dev_err(dev, "Reg: %s value: 0x%x\n",
|
|
error_reg[i].name, val);
|
|
}
|
|
}
|
|
read_unlock_bh(pm_lock);
|
|
goto invalid_pm_state;
|
|
}
|
|
|
|
return (!ret) ? -ETIMEDOUT : 0;
|
|
|
|
invalid_pm_state:
|
|
|
|
return -EIO;
|
|
}
|
|
|
|
void mhi_free_bhie_table(struct mhi_controller *mhi_cntrl,
|
|
struct image_info *image_info)
|
|
{
|
|
int i;
|
|
struct mhi_buf *mhi_buf = image_info->mhi_buf;
|
|
|
|
for (i = 0; i < image_info->entries; i++, mhi_buf++)
|
|
mhi_free_coherent(mhi_cntrl, mhi_buf->len, mhi_buf->buf,
|
|
mhi_buf->dma_addr);
|
|
|
|
kfree(image_info->mhi_buf);
|
|
kfree(image_info);
|
|
}
|
|
|
|
int mhi_alloc_bhie_table(struct mhi_controller *mhi_cntrl,
|
|
struct image_info **image_info,
|
|
size_t alloc_size)
|
|
{
|
|
size_t seg_size = mhi_cntrl->seg_len;
|
|
int segments = DIV_ROUND_UP(alloc_size, seg_size) + 1;
|
|
int i;
|
|
struct image_info *img_info;
|
|
struct mhi_buf *mhi_buf;
|
|
|
|
img_info = kzalloc(sizeof(*img_info), GFP_KERNEL);
|
|
if (!img_info)
|
|
return -ENOMEM;
|
|
|
|
/* Allocate memory for entries */
|
|
img_info->mhi_buf = kcalloc(segments, sizeof(*img_info->mhi_buf),
|
|
GFP_KERNEL);
|
|
if (!img_info->mhi_buf)
|
|
goto error_alloc_mhi_buf;
|
|
|
|
/* Allocate and populate vector table */
|
|
mhi_buf = img_info->mhi_buf;
|
|
for (i = 0; i < segments; i++, mhi_buf++) {
|
|
size_t vec_size = seg_size;
|
|
|
|
/* Vector table is the last entry */
|
|
if (i == segments - 1)
|
|
vec_size = sizeof(struct bhi_vec_entry) * i;
|
|
|
|
mhi_buf->len = vec_size;
|
|
mhi_buf->buf = mhi_alloc_coherent(mhi_cntrl, vec_size,
|
|
&mhi_buf->dma_addr,
|
|
GFP_KERNEL);
|
|
if (!mhi_buf->buf)
|
|
goto error_alloc_segment;
|
|
}
|
|
|
|
img_info->bhi_vec = img_info->mhi_buf[segments - 1].buf;
|
|
img_info->entries = segments;
|
|
*image_info = img_info;
|
|
|
|
return 0;
|
|
|
|
error_alloc_segment:
|
|
for (--i, --mhi_buf; i >= 0; i--, mhi_buf--)
|
|
mhi_free_coherent(mhi_cntrl, mhi_buf->len, mhi_buf->buf,
|
|
mhi_buf->dma_addr);
|
|
|
|
error_alloc_mhi_buf:
|
|
kfree(img_info);
|
|
|
|
return -ENOMEM;
|
|
}
|
|
|
|
static void mhi_firmware_copy(struct mhi_controller *mhi_cntrl,
|
|
const struct firmware *firmware,
|
|
struct image_info *img_info)
|
|
{
|
|
size_t remainder = firmware->size;
|
|
size_t to_cpy;
|
|
const u8 *buf = firmware->data;
|
|
int i = 0;
|
|
struct mhi_buf *mhi_buf = img_info->mhi_buf;
|
|
struct bhi_vec_entry *bhi_vec = img_info->bhi_vec;
|
|
|
|
while (remainder) {
|
|
to_cpy = min(remainder, mhi_buf->len);
|
|
memcpy(mhi_buf->buf, buf, to_cpy);
|
|
bhi_vec->dma_addr = mhi_buf->dma_addr;
|
|
bhi_vec->size = to_cpy;
|
|
|
|
buf += to_cpy;
|
|
remainder -= to_cpy;
|
|
i++;
|
|
bhi_vec++;
|
|
mhi_buf++;
|
|
}
|
|
}
|
|
|
|
void mhi_fw_load_worker(struct work_struct *work)
|
|
{
|
|
struct mhi_controller *mhi_cntrl;
|
|
const struct firmware *firmware = NULL;
|
|
struct image_info *image_info;
|
|
struct device *dev;
|
|
const char *fw_name;
|
|
void *buf;
|
|
dma_addr_t dma_addr;
|
|
size_t size;
|
|
int ret;
|
|
|
|
mhi_cntrl = container_of(work, struct mhi_controller, fw_worker);
|
|
dev = &mhi_cntrl->mhi_dev->dev;
|
|
|
|
dev_dbg(dev, "Waiting for device to enter PBL from: %s\n",
|
|
TO_MHI_EXEC_STR(mhi_cntrl->ee));
|
|
|
|
ret = wait_event_timeout(mhi_cntrl->state_event,
|
|
MHI_IN_PBL(mhi_cntrl->ee) ||
|
|
MHI_PM_IN_ERROR_STATE(mhi_cntrl->pm_state),
|
|
msecs_to_jiffies(mhi_cntrl->timeout_ms));
|
|
|
|
if (!ret || MHI_PM_IN_ERROR_STATE(mhi_cntrl->pm_state)) {
|
|
dev_err(dev, "Device MHI is not in valid state\n");
|
|
return;
|
|
}
|
|
|
|
/* If device is in pass through, do reset to ready state transition */
|
|
if (mhi_cntrl->ee == MHI_EE_PTHRU)
|
|
goto fw_load_ee_pthru;
|
|
|
|
fw_name = (mhi_cntrl->ee == MHI_EE_EDL) ?
|
|
mhi_cntrl->edl_image : mhi_cntrl->fw_image;
|
|
|
|
if (!fw_name || (mhi_cntrl->fbc_download && (!mhi_cntrl->sbl_size ||
|
|
!mhi_cntrl->seg_len))) {
|
|
dev_err(dev,
|
|
"No firmware image defined or !sbl_size || !seg_len\n");
|
|
return;
|
|
}
|
|
|
|
ret = request_firmware(&firmware, fw_name, dev);
|
|
if (ret) {
|
|
dev_err(dev, "Error loading firmware: %d\n", ret);
|
|
return;
|
|
}
|
|
|
|
size = (mhi_cntrl->fbc_download) ? mhi_cntrl->sbl_size : firmware->size;
|
|
|
|
/* SBL size provided is maximum size, not necessarily the image size */
|
|
if (size > firmware->size)
|
|
size = firmware->size;
|
|
|
|
buf = mhi_alloc_coherent(mhi_cntrl, size, &dma_addr, GFP_KERNEL);
|
|
if (!buf) {
|
|
release_firmware(firmware);
|
|
return;
|
|
}
|
|
|
|
/* Download SBL image */
|
|
memcpy(buf, firmware->data, size);
|
|
ret = mhi_fw_load_sbl(mhi_cntrl, dma_addr, size);
|
|
mhi_free_coherent(mhi_cntrl, size, buf, dma_addr);
|
|
|
|
if (!mhi_cntrl->fbc_download || ret || mhi_cntrl->ee == MHI_EE_EDL)
|
|
release_firmware(firmware);
|
|
|
|
/* Error or in EDL mode, we're done */
|
|
if (ret || mhi_cntrl->ee == MHI_EE_EDL)
|
|
return;
|
|
|
|
write_lock_irq(&mhi_cntrl->pm_lock);
|
|
mhi_cntrl->dev_state = MHI_STATE_RESET;
|
|
write_unlock_irq(&mhi_cntrl->pm_lock);
|
|
|
|
/*
|
|
* If we're doing fbc, populate vector tables while
|
|
* device transitioning into MHI READY state
|
|
*/
|
|
if (mhi_cntrl->fbc_download) {
|
|
ret = mhi_alloc_bhie_table(mhi_cntrl, &mhi_cntrl->fbc_image,
|
|
firmware->size);
|
|
if (ret)
|
|
goto error_alloc_fw_table;
|
|
|
|
/* Load the firmware into BHIE vec table */
|
|
mhi_firmware_copy(mhi_cntrl, firmware, mhi_cntrl->fbc_image);
|
|
}
|
|
|
|
fw_load_ee_pthru:
|
|
/* Transitioning into MHI RESET->READY state */
|
|
ret = mhi_ready_state_transition(mhi_cntrl);
|
|
|
|
if (!mhi_cntrl->fbc_download)
|
|
return;
|
|
|
|
if (ret)
|
|
goto error_read;
|
|
|
|
/* Wait for the SBL event */
|
|
ret = wait_event_timeout(mhi_cntrl->state_event,
|
|
mhi_cntrl->ee == MHI_EE_SBL ||
|
|
MHI_PM_IN_ERROR_STATE(mhi_cntrl->pm_state),
|
|
msecs_to_jiffies(mhi_cntrl->timeout_ms));
|
|
|
|
if (!ret || MHI_PM_IN_ERROR_STATE(mhi_cntrl->pm_state)) {
|
|
dev_err(dev, "MHI did not enter SBL\n");
|
|
goto error_read;
|
|
}
|
|
|
|
/* Start full firmware image download */
|
|
image_info = mhi_cntrl->fbc_image;
|
|
ret = mhi_fw_load_amss(mhi_cntrl,
|
|
/* Vector table is the last entry */
|
|
&image_info->mhi_buf[image_info->entries - 1]);
|
|
|
|
release_firmware(firmware);
|
|
|
|
return;
|
|
|
|
error_read:
|
|
mhi_free_bhie_table(mhi_cntrl, mhi_cntrl->fbc_image);
|
|
mhi_cntrl->fbc_image = NULL;
|
|
|
|
error_alloc_fw_table:
|
|
release_firmware(firmware);
|
|
}
|