spi/s3c64xx: Add support DMA engine API
This patch adds to support DMA generic API to transfer raw SPI data. Basiclly the spi driver uses DMA generic API if architecture supports it. Otherwise, uses Samsung specific S3C-PL330 APIs. Signed-off-by: Boojin Kim <boojin.kim@samsung.com> Acked-by: Linus Walleij <linus.walleij@linaro.org> Acked-by: Vinod Koul <vinod.koul@intel.com> Acked-by: Grant Likely <grant.likely@secretlab.ca> Signed-off-by: Kukjin Kim <kgene.kim@samsung.com> Signed-off-by: Vinod Koul <vinod.koul@intel.com>
This commit is contained in:
Родитель
978ce50dd5
Коммит
39d3e8074e
|
@ -171,6 +171,9 @@ struct s3c64xx_spi_driver_data {
|
||||||
unsigned state;
|
unsigned state;
|
||||||
unsigned cur_mode, cur_bpw;
|
unsigned cur_mode, cur_bpw;
|
||||||
unsigned cur_speed;
|
unsigned cur_speed;
|
||||||
|
unsigned rx_ch;
|
||||||
|
unsigned tx_ch;
|
||||||
|
struct samsung_dma_ops *ops;
|
||||||
};
|
};
|
||||||
|
|
||||||
static struct s3c2410_dma_client s3c64xx_spi_dma_client = {
|
static struct s3c2410_dma_client s3c64xx_spi_dma_client = {
|
||||||
|
@ -226,6 +229,38 @@ static void flush_fifo(struct s3c64xx_spi_driver_data *sdd)
|
||||||
writel(val, regs + S3C64XX_SPI_CH_CFG);
|
writel(val, regs + S3C64XX_SPI_CH_CFG);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static void s3c64xx_spi_dma_rxcb(void *data)
|
||||||
|
{
|
||||||
|
struct s3c64xx_spi_driver_data *sdd
|
||||||
|
= (struct s3c64xx_spi_driver_data *)data;
|
||||||
|
unsigned long flags;
|
||||||
|
|
||||||
|
spin_lock_irqsave(&sdd->lock, flags);
|
||||||
|
|
||||||
|
sdd->state &= ~RXBUSY;
|
||||||
|
/* If the other done */
|
||||||
|
if (!(sdd->state & TXBUSY))
|
||||||
|
complete(&sdd->xfer_completion);
|
||||||
|
|
||||||
|
spin_unlock_irqrestore(&sdd->lock, flags);
|
||||||
|
}
|
||||||
|
|
||||||
|
static void s3c64xx_spi_dma_txcb(void *data)
|
||||||
|
{
|
||||||
|
struct s3c64xx_spi_driver_data *sdd
|
||||||
|
= (struct s3c64xx_spi_driver_data *)data;
|
||||||
|
unsigned long flags;
|
||||||
|
|
||||||
|
spin_lock_irqsave(&sdd->lock, flags);
|
||||||
|
|
||||||
|
sdd->state &= ~TXBUSY;
|
||||||
|
/* If the other done */
|
||||||
|
if (!(sdd->state & RXBUSY))
|
||||||
|
complete(&sdd->xfer_completion);
|
||||||
|
|
||||||
|
spin_unlock_irqrestore(&sdd->lock, flags);
|
||||||
|
}
|
||||||
|
|
||||||
static void enable_datapath(struct s3c64xx_spi_driver_data *sdd,
|
static void enable_datapath(struct s3c64xx_spi_driver_data *sdd,
|
||||||
struct spi_device *spi,
|
struct spi_device *spi,
|
||||||
struct spi_transfer *xfer, int dma_mode)
|
struct spi_transfer *xfer, int dma_mode)
|
||||||
|
@ -233,6 +268,7 @@ static void enable_datapath(struct s3c64xx_spi_driver_data *sdd,
|
||||||
struct s3c64xx_spi_info *sci = sdd->cntrlr_info;
|
struct s3c64xx_spi_info *sci = sdd->cntrlr_info;
|
||||||
void __iomem *regs = sdd->regs;
|
void __iomem *regs = sdd->regs;
|
||||||
u32 modecfg, chcfg;
|
u32 modecfg, chcfg;
|
||||||
|
struct samsung_dma_prep_info info;
|
||||||
|
|
||||||
modecfg = readl(regs + S3C64XX_SPI_MODE_CFG);
|
modecfg = readl(regs + S3C64XX_SPI_MODE_CFG);
|
||||||
modecfg &= ~(S3C64XX_SPI_MODE_TXDMA_ON | S3C64XX_SPI_MODE_RXDMA_ON);
|
modecfg &= ~(S3C64XX_SPI_MODE_TXDMA_ON | S3C64XX_SPI_MODE_RXDMA_ON);
|
||||||
|
@ -258,10 +294,14 @@ static void enable_datapath(struct s3c64xx_spi_driver_data *sdd,
|
||||||
chcfg |= S3C64XX_SPI_CH_TXCH_ON;
|
chcfg |= S3C64XX_SPI_CH_TXCH_ON;
|
||||||
if (dma_mode) {
|
if (dma_mode) {
|
||||||
modecfg |= S3C64XX_SPI_MODE_TXDMA_ON;
|
modecfg |= S3C64XX_SPI_MODE_TXDMA_ON;
|
||||||
s3c2410_dma_config(sdd->tx_dmach, sdd->cur_bpw / 8);
|
info.cap = DMA_SLAVE;
|
||||||
s3c2410_dma_enqueue(sdd->tx_dmach, (void *)sdd,
|
info.direction = DMA_TO_DEVICE;
|
||||||
xfer->tx_dma, xfer->len);
|
info.buf = xfer->tx_dma;
|
||||||
s3c2410_dma_ctrl(sdd->tx_dmach, S3C2410_DMAOP_START);
|
info.len = xfer->len;
|
||||||
|
info.fp = s3c64xx_spi_dma_txcb;
|
||||||
|
info.fp_param = sdd;
|
||||||
|
sdd->ops->prepare(sdd->tx_ch, &info);
|
||||||
|
sdd->ops->trigger(sdd->tx_ch);
|
||||||
} else {
|
} else {
|
||||||
switch (sdd->cur_bpw) {
|
switch (sdd->cur_bpw) {
|
||||||
case 32:
|
case 32:
|
||||||
|
@ -293,10 +333,14 @@ static void enable_datapath(struct s3c64xx_spi_driver_data *sdd,
|
||||||
writel(((xfer->len * 8 / sdd->cur_bpw) & 0xffff)
|
writel(((xfer->len * 8 / sdd->cur_bpw) & 0xffff)
|
||||||
| S3C64XX_SPI_PACKET_CNT_EN,
|
| S3C64XX_SPI_PACKET_CNT_EN,
|
||||||
regs + S3C64XX_SPI_PACKET_CNT);
|
regs + S3C64XX_SPI_PACKET_CNT);
|
||||||
s3c2410_dma_config(sdd->rx_dmach, sdd->cur_bpw / 8);
|
info.cap = DMA_SLAVE;
|
||||||
s3c2410_dma_enqueue(sdd->rx_dmach, (void *)sdd,
|
info.direction = DMA_FROM_DEVICE;
|
||||||
xfer->rx_dma, xfer->len);
|
info.buf = xfer->rx_dma;
|
||||||
s3c2410_dma_ctrl(sdd->rx_dmach, S3C2410_DMAOP_START);
|
info.len = xfer->len;
|
||||||
|
info.fp = s3c64xx_spi_dma_rxcb;
|
||||||
|
info.fp_param = sdd;
|
||||||
|
sdd->ops->prepare(sdd->rx_ch, &info);
|
||||||
|
sdd->ops->trigger(sdd->rx_ch);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -482,46 +526,6 @@ static void s3c64xx_spi_config(struct s3c64xx_spi_driver_data *sdd)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
static void s3c64xx_spi_dma_rxcb(struct s3c2410_dma_chan *chan, void *buf_id,
|
|
||||||
int size, enum s3c2410_dma_buffresult res)
|
|
||||||
{
|
|
||||||
struct s3c64xx_spi_driver_data *sdd = buf_id;
|
|
||||||
unsigned long flags;
|
|
||||||
|
|
||||||
spin_lock_irqsave(&sdd->lock, flags);
|
|
||||||
|
|
||||||
if (res == S3C2410_RES_OK)
|
|
||||||
sdd->state &= ~RXBUSY;
|
|
||||||
else
|
|
||||||
dev_err(&sdd->pdev->dev, "DmaAbrtRx-%d\n", size);
|
|
||||||
|
|
||||||
/* If the other done */
|
|
||||||
if (!(sdd->state & TXBUSY))
|
|
||||||
complete(&sdd->xfer_completion);
|
|
||||||
|
|
||||||
spin_unlock_irqrestore(&sdd->lock, flags);
|
|
||||||
}
|
|
||||||
|
|
||||||
static void s3c64xx_spi_dma_txcb(struct s3c2410_dma_chan *chan, void *buf_id,
|
|
||||||
int size, enum s3c2410_dma_buffresult res)
|
|
||||||
{
|
|
||||||
struct s3c64xx_spi_driver_data *sdd = buf_id;
|
|
||||||
unsigned long flags;
|
|
||||||
|
|
||||||
spin_lock_irqsave(&sdd->lock, flags);
|
|
||||||
|
|
||||||
if (res == S3C2410_RES_OK)
|
|
||||||
sdd->state &= ~TXBUSY;
|
|
||||||
else
|
|
||||||
dev_err(&sdd->pdev->dev, "DmaAbrtTx-%d \n", size);
|
|
||||||
|
|
||||||
/* If the other done */
|
|
||||||
if (!(sdd->state & RXBUSY))
|
|
||||||
complete(&sdd->xfer_completion);
|
|
||||||
|
|
||||||
spin_unlock_irqrestore(&sdd->lock, flags);
|
|
||||||
}
|
|
||||||
|
|
||||||
#define XFER_DMAADDR_INVALID DMA_BIT_MASK(32)
|
#define XFER_DMAADDR_INVALID DMA_BIT_MASK(32)
|
||||||
|
|
||||||
static int s3c64xx_spi_map_mssg(struct s3c64xx_spi_driver_data *sdd,
|
static int s3c64xx_spi_map_mssg(struct s3c64xx_spi_driver_data *sdd,
|
||||||
|
@ -696,12 +700,10 @@ static void handle_msg(struct s3c64xx_spi_driver_data *sdd,
|
||||||
if (use_dma) {
|
if (use_dma) {
|
||||||
if (xfer->tx_buf != NULL
|
if (xfer->tx_buf != NULL
|
||||||
&& (sdd->state & TXBUSY))
|
&& (sdd->state & TXBUSY))
|
||||||
s3c2410_dma_ctrl(sdd->tx_dmach,
|
sdd->ops->stop(sdd->tx_ch);
|
||||||
S3C2410_DMAOP_FLUSH);
|
|
||||||
if (xfer->rx_buf != NULL
|
if (xfer->rx_buf != NULL
|
||||||
&& (sdd->state & RXBUSY))
|
&& (sdd->state & RXBUSY))
|
||||||
s3c2410_dma_ctrl(sdd->rx_dmach,
|
sdd->ops->stop(sdd->rx_ch);
|
||||||
S3C2410_DMAOP_FLUSH);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
goto out;
|
goto out;
|
||||||
|
@ -741,24 +743,19 @@ out:
|
||||||
|
|
||||||
static int acquire_dma(struct s3c64xx_spi_driver_data *sdd)
|
static int acquire_dma(struct s3c64xx_spi_driver_data *sdd)
|
||||||
{
|
{
|
||||||
if (s3c2410_dma_request(sdd->rx_dmach,
|
|
||||||
&s3c64xx_spi_dma_client, NULL) < 0) {
|
|
||||||
dev_err(&sdd->pdev->dev, "cannot get RxDMA\n");
|
|
||||||
return 0;
|
|
||||||
}
|
|
||||||
s3c2410_dma_set_buffdone_fn(sdd->rx_dmach, s3c64xx_spi_dma_rxcb);
|
|
||||||
s3c2410_dma_devconfig(sdd->rx_dmach, S3C2410_DMASRC_HW,
|
|
||||||
sdd->sfr_start + S3C64XX_SPI_RX_DATA);
|
|
||||||
|
|
||||||
if (s3c2410_dma_request(sdd->tx_dmach,
|
struct samsung_dma_info info;
|
||||||
&s3c64xx_spi_dma_client, NULL) < 0) {
|
sdd->ops = samsung_dma_get_ops();
|
||||||
dev_err(&sdd->pdev->dev, "cannot get TxDMA\n");
|
|
||||||
s3c2410_dma_free(sdd->rx_dmach, &s3c64xx_spi_dma_client);
|
info.cap = DMA_SLAVE;
|
||||||
return 0;
|
info.client = &s3c64xx_spi_dma_client;
|
||||||
}
|
info.direction = DMA_FROM_DEVICE;
|
||||||
s3c2410_dma_set_buffdone_fn(sdd->tx_dmach, s3c64xx_spi_dma_txcb);
|
info.fifo = sdd->sfr_start + S3C64XX_SPI_RX_DATA;
|
||||||
s3c2410_dma_devconfig(sdd->tx_dmach, S3C2410_DMASRC_MEM,
|
info.width = sdd->cur_bpw / 8;
|
||||||
sdd->sfr_start + S3C64XX_SPI_TX_DATA);
|
sdd->rx_ch = sdd->ops->request(sdd->rx_dmach, &info);
|
||||||
|
info.direction = DMA_TO_DEVICE;
|
||||||
|
info.fifo = sdd->sfr_start + S3C64XX_SPI_TX_DATA;
|
||||||
|
sdd->tx_ch = sdd->ops->request(sdd->tx_dmach, &info);
|
||||||
|
|
||||||
return 1;
|
return 1;
|
||||||
}
|
}
|
||||||
|
@ -799,8 +796,8 @@ static void s3c64xx_spi_work(struct work_struct *work)
|
||||||
spin_unlock_irqrestore(&sdd->lock, flags);
|
spin_unlock_irqrestore(&sdd->lock, flags);
|
||||||
|
|
||||||
/* Free DMA channels */
|
/* Free DMA channels */
|
||||||
s3c2410_dma_free(sdd->tx_dmach, &s3c64xx_spi_dma_client);
|
sdd->ops->release(sdd->rx_ch, &s3c64xx_spi_dma_client);
|
||||||
s3c2410_dma_free(sdd->rx_dmach, &s3c64xx_spi_dma_client);
|
sdd->ops->release(sdd->tx_ch, &s3c64xx_spi_dma_client);
|
||||||
}
|
}
|
||||||
|
|
||||||
static int s3c64xx_spi_transfer(struct spi_device *spi,
|
static int s3c64xx_spi_transfer(struct spi_device *spi,
|
||||||
|
|
Загрузка…
Ссылка в новой задаче