mirror of
https://github.com/nxp-imx/linux-imx.git
synced 2025-09-02 18:06:13 +02:00

-----BEGIN PGP SIGNATURE----- iQIzBAABCAAdFiEEZH8oZUiU471FcZm+ONu9yGCSaT4FAmbisF0ACgkQONu9yGCS aT5Y8xAAqS/rmrC+/qlFvbtAqK+KXLq9BIGvDHW2QHfCyMpSZ6isehVhh64apHE/ /XvJ6a+2iPVp5o52iDTUKzbcDr3Jx/QwhS8Xa/HyQQy1rXIPpJNJb8Vuvkn/B2Cq cPCfTtfPZUUQTd09uAdBhy5NT8hsT2kSVpmSXDnahn9ih8k0tR40udw5Qf7xpWcf HqljbfonLP86mF/SB9m+VhDGF9fekujyb+0iS0OPE+TdvSjKB9ySoeL4PIeTSxrz goZdp9ygAYy8Bks825ztbfQszqIwceHU/xZRaUrGfOOk4A5kwTmbdUQu7ooMc+5F kbpifbewmY1UGn2KTxgj59xCjQ7HLQe+sqacy0/gALzRSajUNyjLn0n4w3UqaJWb pf+gwqHBLgDRfvWctggEdY2ApKgOlM9D7TTpWWB9uv1oR/g3PGfgehZgrMMPgPUw EZ8JiwnITfRaRFiH/vSR3aJKRj6qjb4mX3/U8HgGcACtyFfHgtuI7jzhnX36fRNO FG38bxSUMrJnlohghfBl6zyaruZBMHVaoQzs6MYZ7qrVvCbt3CHivJdaQ85nw0h7 YHa2zYFfT0ztyaSMzWq6JatgI7BZfd8PjobhbRZADBBD39KC8aL8XLoDPnpzWMUY UDlK8n96gOKo0t8ILDWcIisCVGNogcHJlGppC8Fu7ZyKzYsMhN4= =OEL/ -----END PGP SIGNATURE----- Merge tag 'v6.6.51' into lf-6.6.y This is the 6.6.51 stable release * tag 'v6.6.51': (2369 commits) Linux 6.6.51 Bluetooth: hci_sync: Fix UAF on hci_abort_conn_sync Bluetooth: hci_sync: Fix UAF on create_le_conn_complete ... Signed-off-by: Jason Liu <jason.hui.liu@nxp.com> Conflicts: arch/arm64/boot/dts/freescale/imx8mp.dtsi arch/arm64/boot/dts/freescale/imx93.dtsi drivers/dma/fsl-edma-common.c drivers/dma/fsl-edma-common.h drivers/dma/fsl-edma.c drivers/irqchip/irq-imx-irqsteer.c drivers/perf/fsl_imx9_ddr_perf.c drivers/spi/spi-fsl-lpspi.c sound/soc/sof/imx/imx8m.c
1202 lines
29 KiB
C
1202 lines
29 KiB
C
// SPDX-License-Identifier: GPL-2.0+
|
|
//
|
|
// Freescale i.MX7ULP LPSPI driver
|
|
//
|
|
// Copyright 2016 Freescale Semiconductor, Inc.
|
|
// Copyright 2018 NXP Semiconductors
|
|
|
|
#include <linux/clk.h>
|
|
#include <linux/completion.h>
|
|
#include <linux/delay.h>
|
|
#include <linux/dmaengine.h>
|
|
#include <linux/dma-mapping.h>
|
|
#include <linux/err.h>
|
|
#include <linux/interrupt.h>
|
|
#include <linux/io.h>
|
|
#include <linux/irq.h>
|
|
#include <linux/kernel.h>
|
|
#include <linux/module.h>
|
|
#include <linux/of.h>
|
|
#include <linux/pinctrl/consumer.h>
|
|
#include <linux/platform_device.h>
|
|
#include <linux/dma/imx-dma.h>
|
|
#include <linux/pm_runtime.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/spi/spi.h>
|
|
#include <linux/spi/spi_bitbang.h>
|
|
#include <linux/types.h>
|
|
|
|
#define DRIVER_NAME "fsl_lpspi"
|
|
|
|
#define FSL_LPSPI_RPM_TIMEOUT 50 /* 50ms */
|
|
|
|
/* The maximum bytes that edma can transfer once.*/
|
|
#define FSL_LPSPI_MAX_EDMA_BYTES ((1 << 15) - 1)
|
|
|
|
/* i.MX7ULP LPSPI registers */
|
|
#define IMX7ULP_VERID 0x0
|
|
#define IMX7ULP_PARAM 0x4
|
|
#define IMX7ULP_CR 0x10
|
|
#define IMX7ULP_SR 0x14
|
|
#define IMX7ULP_IER 0x18
|
|
#define IMX7ULP_DER 0x1c
|
|
#define IMX7ULP_CFGR0 0x20
|
|
#define IMX7ULP_CFGR1 0x24
|
|
#define IMX7ULP_DMR0 0x30
|
|
#define IMX7ULP_DMR1 0x34
|
|
#define IMX7ULP_CCR 0x40
|
|
#define IMX7ULP_FCR 0x58
|
|
#define IMX7ULP_FSR 0x5c
|
|
#define IMX7ULP_TCR 0x60
|
|
#define IMX7ULP_TDR 0x64
|
|
#define IMX7ULP_RSR 0x70
|
|
#define IMX7ULP_RDR 0x74
|
|
|
|
/* General control register field define */
|
|
#define CR_RRF BIT(9)
|
|
#define CR_RTF BIT(8)
|
|
#define CR_RST BIT(1)
|
|
#define CR_MEN BIT(0)
|
|
#define SR_MBF BIT(24)
|
|
#define SR_TCF BIT(10)
|
|
#define SR_FCF BIT(9)
|
|
#define SR_RDF BIT(1)
|
|
#define SR_TDF BIT(0)
|
|
#define IER_TCIE BIT(10)
|
|
#define IER_FCIE BIT(9)
|
|
#define IER_RDIE BIT(1)
|
|
#define IER_TDIE BIT(0)
|
|
#define DER_RDDE BIT(1)
|
|
#define DER_TDDE BIT(0)
|
|
#define CFGR1_PCSCFG BIT(27)
|
|
#define CFGR1_PINCFG (BIT(24)|BIT(25))
|
|
#define CFGR1_PCSPOL BIT(8)
|
|
#define CFGR1_NOSTALL BIT(3)
|
|
#define CFGR1_HOST BIT(0)
|
|
#define FSR_TXCOUNT (0xFF)
|
|
#define RSR_RXEMPTY BIT(1)
|
|
#define TCR_CPOL BIT(31)
|
|
#define TCR_CPHA BIT(30)
|
|
#define TCR_CONT BIT(21)
|
|
#define TCR_CONTC BIT(20)
|
|
#define TCR_RXMSK BIT(19)
|
|
#define TCR_TXMSK BIT(18)
|
|
|
|
enum fsl_lpspi_devtype {
|
|
IMX7ULP_LPSPI,
|
|
IMX93_LPSPI,
|
|
IMX95_LPSPI,
|
|
};
|
|
|
|
struct fsl_lpspi_devtype_data {
|
|
enum fsl_lpspi_devtype devtype;
|
|
u8 prescale_max;
|
|
};
|
|
|
|
struct lpspi_config {
|
|
u8 bpw;
|
|
u8 chip_select;
|
|
u8 prescale;
|
|
u16 mode;
|
|
u32 speed_hz;
|
|
};
|
|
|
|
struct fsl_lpspi_data {
|
|
struct device *dev;
|
|
void __iomem *base;
|
|
unsigned long base_phys;
|
|
struct clk *clk_ipg;
|
|
struct clk *clk_per;
|
|
bool is_target;
|
|
bool is_only_cs1;
|
|
bool is_first_byte;
|
|
|
|
void *rx_buf;
|
|
const void *tx_buf;
|
|
void (*tx)(struct fsl_lpspi_data *);
|
|
void (*rx)(struct fsl_lpspi_data *);
|
|
|
|
u32 remain;
|
|
u8 watermark;
|
|
u8 txfifosize;
|
|
u8 rxfifosize;
|
|
|
|
struct lpspi_config config;
|
|
struct completion xfer_done;
|
|
|
|
bool target_aborted;
|
|
|
|
/* DMA */
|
|
bool usedma;
|
|
struct completion dma_rx_completion;
|
|
struct completion dma_tx_completion;
|
|
/* DMA for slave*/
|
|
struct spi_transfer *cur_transfer;
|
|
|
|
const struct fsl_lpspi_devtype_data *devtype_data;
|
|
};
|
|
|
|
static inline int is_imx7ulp_lpspi(struct fsl_lpspi_data *d)
|
|
{
|
|
return d->devtype_data->devtype == IMX7ULP_LPSPI;
|
|
};
|
|
|
|
static inline int is_imx93_lpspi(struct fsl_lpspi_data *d)
|
|
{
|
|
return d->devtype_data->devtype == IMX93_LPSPI;
|
|
};
|
|
|
|
static inline int is_imx95_lpspi(struct fsl_lpspi_data *d)
|
|
{
|
|
return d->devtype_data->devtype == IMX95_LPSPI;
|
|
};
|
|
|
|
/*
|
|
* ERR051608 fixed or not:
|
|
* https://www.nxp.com/docs/en/errata/i.MX93_1P87f.pdf
|
|
*/
|
|
static struct fsl_lpspi_devtype_data imx93_lpspi_devtype_data = {
|
|
.devtype = IMX93_LPSPI,
|
|
.prescale_max = 1,
|
|
};
|
|
|
|
static struct fsl_lpspi_devtype_data imx95_lpspi_devtype_data = {
|
|
.devtype = IMX95_LPSPI,
|
|
.prescale_max = 7,
|
|
};
|
|
|
|
static struct fsl_lpspi_devtype_data imx7ulp_lpspi_devtype_data = {
|
|
.devtype = IMX7ULP_LPSPI,
|
|
.prescale_max = 7,
|
|
};
|
|
|
|
/*
|
|
* IMX95, IMX93 have a different edma driver from imx7ulp, so lpspi slave
|
|
* will have different settings according to the edma of different platforms
|
|
* to meet the needs of each platform.
|
|
*/
|
|
static const struct of_device_id fsl_lpspi_dt_ids[] = {
|
|
{ .compatible = "fsl,imx7ulp-spi", .data = &imx7ulp_lpspi_devtype_data,},
|
|
{ .compatible = "fsl,imx93-spi", .data = &imx93_lpspi_devtype_data,},
|
|
{ .compatible = "fsl,imx95-spi", .data = &imx95_lpspi_devtype_data,},
|
|
{ /* sentinel */ }
|
|
};
|
|
MODULE_DEVICE_TABLE(of, fsl_lpspi_dt_ids);
|
|
|
|
#define LPSPI_BUF_RX(type) \
|
|
static void fsl_lpspi_buf_rx_##type(struct fsl_lpspi_data *fsl_lpspi) \
|
|
{ \
|
|
unsigned int val = readl(fsl_lpspi->base + IMX7ULP_RDR); \
|
|
\
|
|
if (fsl_lpspi->rx_buf) { \
|
|
*(type *)fsl_lpspi->rx_buf = val; \
|
|
fsl_lpspi->rx_buf += sizeof(type); \
|
|
} \
|
|
}
|
|
|
|
#define LPSPI_BUF_TX(type) \
|
|
static void fsl_lpspi_buf_tx_##type(struct fsl_lpspi_data *fsl_lpspi) \
|
|
{ \
|
|
type val = 0; \
|
|
\
|
|
if (fsl_lpspi->tx_buf) { \
|
|
val = *(type *)fsl_lpspi->tx_buf; \
|
|
fsl_lpspi->tx_buf += sizeof(type); \
|
|
} \
|
|
\
|
|
fsl_lpspi->remain -= sizeof(type); \
|
|
writel(val, fsl_lpspi->base + IMX7ULP_TDR); \
|
|
}
|
|
|
|
LPSPI_BUF_RX(u8)
|
|
LPSPI_BUF_TX(u8)
|
|
LPSPI_BUF_RX(u16)
|
|
LPSPI_BUF_TX(u16)
|
|
LPSPI_BUF_RX(u32)
|
|
LPSPI_BUF_TX(u32)
|
|
|
|
static void fsl_lpspi_intctrl(struct fsl_lpspi_data *fsl_lpspi,
|
|
unsigned int enable)
|
|
{
|
|
writel(enable, fsl_lpspi->base + IMX7ULP_IER);
|
|
}
|
|
|
|
static int fsl_lpspi_bytes_per_word(const int bpw)
|
|
{
|
|
return DIV_ROUND_UP(bpw, BITS_PER_BYTE);
|
|
}
|
|
|
|
static bool fsl_lpspi_can_dma(struct spi_controller *controller,
|
|
struct spi_device *spi,
|
|
struct spi_transfer *transfer)
|
|
{
|
|
unsigned int bytes_per_word;
|
|
|
|
if (!controller->dma_rx)
|
|
return false;
|
|
|
|
bytes_per_word = fsl_lpspi_bytes_per_word(transfer->bits_per_word);
|
|
|
|
switch (bytes_per_word) {
|
|
case 1:
|
|
case 2:
|
|
case 4:
|
|
break;
|
|
default:
|
|
return false;
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
static int lpspi_prepare_xfer_hardware(struct spi_controller *controller)
|
|
{
|
|
struct fsl_lpspi_data *fsl_lpspi =
|
|
spi_controller_get_devdata(controller);
|
|
int ret;
|
|
|
|
ret = pm_runtime_resume_and_get(fsl_lpspi->dev);
|
|
if (ret < 0) {
|
|
dev_err(fsl_lpspi->dev, "failed to enable clock\n");
|
|
return ret;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int lpspi_unprepare_xfer_hardware(struct spi_controller *controller)
|
|
{
|
|
struct fsl_lpspi_data *fsl_lpspi =
|
|
spi_controller_get_devdata(controller);
|
|
|
|
pm_runtime_mark_last_busy(fsl_lpspi->dev);
|
|
pm_runtime_put_autosuspend(fsl_lpspi->dev);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void fsl_lpspi_write_tx_fifo(struct fsl_lpspi_data *fsl_lpspi)
|
|
{
|
|
u8 txfifo_cnt;
|
|
u32 temp;
|
|
|
|
txfifo_cnt = readl(fsl_lpspi->base + IMX7ULP_FSR) & 0xff;
|
|
|
|
while (txfifo_cnt < fsl_lpspi->txfifosize) {
|
|
if (!fsl_lpspi->remain)
|
|
break;
|
|
fsl_lpspi->tx(fsl_lpspi);
|
|
txfifo_cnt++;
|
|
}
|
|
|
|
if (txfifo_cnt < fsl_lpspi->txfifosize) {
|
|
if (!fsl_lpspi->is_target) {
|
|
temp = readl(fsl_lpspi->base + IMX7ULP_TCR);
|
|
temp &= ~TCR_CONTC;
|
|
writel(temp, fsl_lpspi->base + IMX7ULP_TCR);
|
|
}
|
|
|
|
fsl_lpspi_intctrl(fsl_lpspi, IER_FCIE);
|
|
} else
|
|
fsl_lpspi_intctrl(fsl_lpspi, IER_TDIE);
|
|
}
|
|
|
|
static void fsl_lpspi_read_rx_fifo(struct fsl_lpspi_data *fsl_lpspi)
|
|
{
|
|
while (!(readl(fsl_lpspi->base + IMX7ULP_RSR) & RSR_RXEMPTY))
|
|
fsl_lpspi->rx(fsl_lpspi);
|
|
}
|
|
|
|
static void fsl_lpspi_set_cmd(struct fsl_lpspi_data *fsl_lpspi)
|
|
{
|
|
u32 temp = 0;
|
|
|
|
temp |= fsl_lpspi->config.bpw - 1;
|
|
temp |= (fsl_lpspi->config.mode & 0x3) << 30;
|
|
temp |= (fsl_lpspi->config.chip_select & 0x3) << 24;
|
|
if (!fsl_lpspi->is_target) {
|
|
temp |= fsl_lpspi->config.prescale << 27;
|
|
/*
|
|
* Set TCR_CONT will keep SS asserted after current transfer.
|
|
* For the first transfer, clear TCR_CONTC to assert SS.
|
|
* For subsequent transfer, set TCR_CONTC to keep SS asserted.
|
|
*/
|
|
if (!fsl_lpspi->usedma) {
|
|
temp |= TCR_CONT;
|
|
if (fsl_lpspi->is_first_byte)
|
|
temp &= ~TCR_CONTC;
|
|
else
|
|
temp |= TCR_CONTC;
|
|
}
|
|
}
|
|
writel(temp, fsl_lpspi->base + IMX7ULP_TCR);
|
|
|
|
dev_dbg(fsl_lpspi->dev, "TCR=0x%x\n", temp);
|
|
}
|
|
|
|
static void fsl_lpspi_set_watermark(struct fsl_lpspi_data *fsl_lpspi)
|
|
{
|
|
u32 temp;
|
|
|
|
if (!fsl_lpspi->usedma)
|
|
temp = fsl_lpspi->watermark >> 1 |
|
|
(fsl_lpspi->watermark >> 1) << 16;
|
|
else
|
|
temp = fsl_lpspi->watermark >> 1;
|
|
|
|
writel(temp, fsl_lpspi->base + IMX7ULP_FCR);
|
|
|
|
dev_dbg(fsl_lpspi->dev, "FCR=0x%x\n", temp);
|
|
}
|
|
|
|
static int fsl_lpspi_set_bitrate(struct fsl_lpspi_data *fsl_lpspi)
|
|
{
|
|
struct lpspi_config config = fsl_lpspi->config;
|
|
unsigned int perclk_rate, scldiv, div;
|
|
u8 prescale_max;
|
|
u8 prescale;
|
|
|
|
perclk_rate = clk_get_rate(fsl_lpspi->clk_per);
|
|
if (!perclk_rate) {
|
|
dev_err(fsl_lpspi->dev, "per-clk rate was not set\n");
|
|
return -EINVAL;
|
|
}
|
|
prescale_max = fsl_lpspi->devtype_data->prescale_max;
|
|
|
|
if (!config.speed_hz) {
|
|
dev_err(fsl_lpspi->dev,
|
|
"error: the transmission speed provided is 0!\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (config.speed_hz > perclk_rate / 2) {
|
|
dev_err(fsl_lpspi->dev,
|
|
"per-clk should be at least two times of transfer speed");
|
|
return -EINVAL;
|
|
}
|
|
|
|
div = DIV_ROUND_UP(perclk_rate, config.speed_hz);
|
|
|
|
for (prescale = 0; prescale <= prescale_max; prescale++) {
|
|
scldiv = div / (1 << prescale) - 2;
|
|
if (scldiv < 256) {
|
|
fsl_lpspi->config.prescale = prescale;
|
|
break;
|
|
}
|
|
}
|
|
|
|
if (scldiv >= 256)
|
|
return -EINVAL;
|
|
|
|
writel(scldiv | (scldiv << 8) | ((scldiv >> 1) << 16),
|
|
fsl_lpspi->base + IMX7ULP_CCR);
|
|
|
|
dev_dbg(fsl_lpspi->dev, "perclk=%d, speed=%d, prescale=%d, scldiv=%d\n",
|
|
perclk_rate, config.speed_hz, prescale, scldiv);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int fsl_lpspi_dma_configure(struct spi_controller *controller)
|
|
{
|
|
int ret, dma_burst;
|
|
enum dma_slave_buswidth buswidth;
|
|
struct dma_slave_config rx = {}, tx = {};
|
|
struct fsl_lpspi_data *fsl_lpspi =
|
|
spi_controller_get_devdata(controller);
|
|
|
|
switch (fsl_lpspi_bytes_per_word(fsl_lpspi->config.bpw)) {
|
|
case 4:
|
|
buswidth = DMA_SLAVE_BUSWIDTH_4_BYTES;
|
|
break;
|
|
case 2:
|
|
buswidth = DMA_SLAVE_BUSWIDTH_2_BYTES;
|
|
break;
|
|
case 1:
|
|
buswidth = DMA_SLAVE_BUSWIDTH_1_BYTE;
|
|
break;
|
|
default:
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (fsl_lpspi->is_target && (is_imx95_lpspi(fsl_lpspi) || is_imx93_lpspi(fsl_lpspi))) {
|
|
|
|
/*
|
|
* Dma maxburst should equal to fifo watermark. But when data length <= fifo_size/2
|
|
* dma should burst all data into fifo.
|
|
*/
|
|
if (fsl_lpspi->cur_transfer->len > fsl_lpspi->txfifosize >> 1)
|
|
dma_burst = fsl_lpspi->txfifosize >> 1;
|
|
else
|
|
dma_burst = fsl_lpspi->cur_transfer->len;
|
|
} else
|
|
dma_burst = 1;
|
|
|
|
tx.direction = DMA_MEM_TO_DEV;
|
|
tx.dst_addr = fsl_lpspi->base_phys + IMX7ULP_TDR;
|
|
tx.dst_addr_width = buswidth;
|
|
tx.dst_maxburst = dma_burst;
|
|
ret = dmaengine_slave_config(controller->dma_tx, &tx);
|
|
if (ret) {
|
|
dev_err(fsl_lpspi->dev, "TX dma configuration failed with %d\n",
|
|
ret);
|
|
return ret;
|
|
}
|
|
|
|
rx.direction = DMA_DEV_TO_MEM;
|
|
rx.src_addr = fsl_lpspi->base_phys + IMX7ULP_RDR;
|
|
rx.src_addr_width = buswidth;
|
|
rx.src_maxburst = 1;
|
|
ret = dmaengine_slave_config(controller->dma_rx, &rx);
|
|
if (ret) {
|
|
dev_err(fsl_lpspi->dev, "RX dma configuration failed with %d\n",
|
|
ret);
|
|
return ret;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int fsl_lpspi_config(struct fsl_lpspi_data *fsl_lpspi)
|
|
{
|
|
u32 temp;
|
|
int ret;
|
|
|
|
if (!fsl_lpspi->is_target) {
|
|
ret = fsl_lpspi_set_bitrate(fsl_lpspi);
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
|
|
fsl_lpspi_set_watermark(fsl_lpspi);
|
|
|
|
if (!fsl_lpspi->is_target)
|
|
temp = CFGR1_HOST;
|
|
else
|
|
temp = CFGR1_PINCFG;
|
|
if (fsl_lpspi->config.mode & SPI_CS_HIGH)
|
|
temp |= CFGR1_PCSPOL;
|
|
writel(temp, fsl_lpspi->base + IMX7ULP_CFGR1);
|
|
|
|
temp = readl(fsl_lpspi->base + IMX7ULP_CR);
|
|
temp |= CR_RRF | CR_RTF | CR_MEN;
|
|
writel(temp, fsl_lpspi->base + IMX7ULP_CR);
|
|
|
|
temp = 0;
|
|
if (fsl_lpspi->usedma)
|
|
temp = DER_TDDE | DER_RDDE;
|
|
writel(temp, fsl_lpspi->base + IMX7ULP_DER);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int fsl_lpspi_setup_transfer(struct spi_controller *controller,
|
|
struct spi_device *spi,
|
|
struct spi_transfer *t)
|
|
{
|
|
struct fsl_lpspi_data *fsl_lpspi =
|
|
spi_controller_get_devdata(spi->controller);
|
|
|
|
if (t == NULL)
|
|
return -EINVAL;
|
|
|
|
fsl_lpspi->config.mode = spi->mode;
|
|
fsl_lpspi->config.bpw = t->bits_per_word;
|
|
fsl_lpspi->config.speed_hz = t->speed_hz;
|
|
if (fsl_lpspi->is_only_cs1)
|
|
fsl_lpspi->config.chip_select = 1;
|
|
else
|
|
fsl_lpspi->config.chip_select = spi_get_chipselect(spi, 0);
|
|
|
|
if (!fsl_lpspi->config.speed_hz)
|
|
fsl_lpspi->config.speed_hz = spi->max_speed_hz;
|
|
if (!fsl_lpspi->config.bpw)
|
|
fsl_lpspi->config.bpw = spi->bits_per_word;
|
|
|
|
/* Initialize the functions for transfer */
|
|
if (fsl_lpspi->config.bpw <= 8) {
|
|
fsl_lpspi->rx = fsl_lpspi_buf_rx_u8;
|
|
fsl_lpspi->tx = fsl_lpspi_buf_tx_u8;
|
|
} else if (fsl_lpspi->config.bpw <= 16) {
|
|
fsl_lpspi->rx = fsl_lpspi_buf_rx_u16;
|
|
fsl_lpspi->tx = fsl_lpspi_buf_tx_u16;
|
|
} else {
|
|
fsl_lpspi->rx = fsl_lpspi_buf_rx_u32;
|
|
fsl_lpspi->tx = fsl_lpspi_buf_tx_u32;
|
|
}
|
|
|
|
if (t->len <= fsl_lpspi->txfifosize)
|
|
fsl_lpspi->watermark = t->len;
|
|
else
|
|
fsl_lpspi->watermark = fsl_lpspi->txfifosize;
|
|
|
|
if (fsl_lpspi_can_dma(controller, spi, t))
|
|
fsl_lpspi->usedma = true;
|
|
else
|
|
fsl_lpspi->usedma = false;
|
|
|
|
return fsl_lpspi_config(fsl_lpspi);
|
|
}
|
|
|
|
static int fsl_lpspi_target_abort(struct spi_controller *controller)
|
|
{
|
|
struct fsl_lpspi_data *fsl_lpspi =
|
|
spi_controller_get_devdata(controller);
|
|
|
|
fsl_lpspi->target_aborted = true;
|
|
if (!fsl_lpspi->usedma)
|
|
complete(&fsl_lpspi->xfer_done);
|
|
else {
|
|
complete(&fsl_lpspi->dma_tx_completion);
|
|
complete(&fsl_lpspi->dma_rx_completion);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int fsl_lpspi_wait_for_completion(struct spi_controller *controller)
|
|
{
|
|
struct fsl_lpspi_data *fsl_lpspi =
|
|
spi_controller_get_devdata(controller);
|
|
|
|
if (fsl_lpspi->is_target) {
|
|
if (wait_for_completion_interruptible(&fsl_lpspi->xfer_done) ||
|
|
fsl_lpspi->target_aborted) {
|
|
dev_dbg(fsl_lpspi->dev, "interrupted\n");
|
|
return -EINTR;
|
|
}
|
|
} else {
|
|
if (!wait_for_completion_timeout(&fsl_lpspi->xfer_done, HZ)) {
|
|
dev_dbg(fsl_lpspi->dev, "wait for completion timeout\n");
|
|
return -ETIMEDOUT;
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int fsl_lpspi_reset(struct fsl_lpspi_data *fsl_lpspi)
|
|
{
|
|
u32 temp;
|
|
|
|
if (!fsl_lpspi->usedma) {
|
|
/* Disable all interrupt */
|
|
fsl_lpspi_intctrl(fsl_lpspi, 0);
|
|
}
|
|
|
|
/* W1C for all flags in SR */
|
|
temp = 0x3F << 8;
|
|
writel(temp, fsl_lpspi->base + IMX7ULP_SR);
|
|
|
|
/* Clear FIFO and disable module */
|
|
temp = CR_RRF | CR_RTF;
|
|
writel(temp, fsl_lpspi->base + IMX7ULP_CR);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void fsl_lpspi_dma_rx_callback(void *cookie)
|
|
{
|
|
struct fsl_lpspi_data *fsl_lpspi = (struct fsl_lpspi_data *)cookie;
|
|
|
|
complete(&fsl_lpspi->dma_rx_completion);
|
|
}
|
|
|
|
static void fsl_lpspi_dma_tx_callback(void *cookie)
|
|
{
|
|
struct fsl_lpspi_data *fsl_lpspi = (struct fsl_lpspi_data *)cookie;
|
|
|
|
complete(&fsl_lpspi->dma_tx_completion);
|
|
}
|
|
|
|
static int fsl_lpspi_calculate_timeout(struct fsl_lpspi_data *fsl_lpspi,
|
|
int size)
|
|
{
|
|
unsigned long timeout = 0;
|
|
|
|
/* Time with actual data transfer and CS change delay related to HW */
|
|
timeout = (8 + 4) * size / fsl_lpspi->config.speed_hz;
|
|
|
|
/* Add extra second for scheduler related activities */
|
|
timeout += 1;
|
|
|
|
/* Double calculated timeout */
|
|
return msecs_to_jiffies(2 * timeout * MSEC_PER_SEC);
|
|
}
|
|
|
|
static struct sg_table *fsl_lpspi_allocate_sg_for_target(struct spi_controller *controller,
|
|
enum dma_data_direction dir)
|
|
{
|
|
struct fsl_lpspi_data *fsl_lpspi = spi_controller_get_devdata(controller);
|
|
struct spi_transfer *xfer = fsl_lpspi->cur_transfer;
|
|
struct scatterlist *sg;
|
|
struct sg_table *sgt;
|
|
struct dma_chan *chan;
|
|
size_t bytes;
|
|
const void *buf, *pbuf;
|
|
int i, ret, sg_num, len, tail;
|
|
|
|
/*
|
|
* When lpspi transfer data is not a multiple of edma burst, it means that
|
|
* there is a tail data which edma can not burst the data into fifo. So add
|
|
* an extra sg to help edma to handle the tail data. Edma will automatically
|
|
* reduce burst length to ensure that tail data can be burst to FIFO correctly.
|
|
* Using an extra sg to handle tail data using a lower edma performance but it
|
|
* can ensure other data can be bursted into FIFO using a higher edma performance.
|
|
*/
|
|
len = xfer->len;
|
|
if (len > fsl_lpspi->txfifosize >> 1)
|
|
tail = len % (fsl_lpspi->txfifosize >> 1);
|
|
else
|
|
tail = 0;
|
|
|
|
switch (dir) {
|
|
case DMA_FROM_DEVICE:
|
|
chan = controller->dma_rx;
|
|
buf = xfer->rx_buf;
|
|
sgt = &xfer->rx_sg;
|
|
break;
|
|
case DMA_TO_DEVICE:
|
|
chan = controller->dma_tx;
|
|
buf = xfer->tx_buf;
|
|
sgt = &xfer->tx_sg;
|
|
break;
|
|
default:
|
|
return ERR_PTR(-EINVAL);
|
|
}
|
|
|
|
if (!virt_addr_valid(buf))
|
|
return ERR_PTR(-EINVAL);
|
|
|
|
sg_num = DIV_ROUND_UP(len, PAGE_SIZE);
|
|
if (tail)
|
|
sg_num += 1;
|
|
|
|
sg_free_table(sgt);
|
|
|
|
ret = sg_alloc_table(sgt, sg_num, GFP_KERNEL);
|
|
|
|
if (ret) {
|
|
sg_free_table(sgt);
|
|
return ERR_PTR(ret);
|
|
}
|
|
|
|
pbuf = buf;
|
|
for_each_sg(sgt->sgl, sg, sgt->nents, i) {
|
|
if (i == (sg_num - 1) && tail)
|
|
bytes = min_t(size_t, tail, PAGE_SIZE);
|
|
else
|
|
bytes = min_t(size_t, len - tail, PAGE_SIZE);
|
|
|
|
sg_set_buf(sg, pbuf, bytes);
|
|
|
|
pbuf += bytes;
|
|
len -= bytes;
|
|
}
|
|
|
|
if (WARN_ON(len)) {
|
|
dev_err(&controller->dev, "len = %d but expected 0!\n", len);
|
|
sg_free_table(sgt);
|
|
return ERR_PTR(-EINVAL);
|
|
}
|
|
|
|
sg_num = dma_map_sg(chan->device->dev, sgt->sgl, sgt->nents, dir);
|
|
if (!sg_num) {
|
|
sg_free_table(sgt);
|
|
return ERR_PTR(-ENOMEM);
|
|
}
|
|
return sgt;
|
|
}
|
|
|
|
static int fsl_lpspi_dma_transfer(struct spi_controller *controller,
|
|
struct fsl_lpspi_data *fsl_lpspi,
|
|
struct spi_transfer *transfer)
|
|
{
|
|
struct dma_async_tx_descriptor *desc_tx, *desc_rx;
|
|
unsigned long transfer_timeout;
|
|
unsigned long timeout;
|
|
struct sg_table *tx, *rx;
|
|
int ret;
|
|
|
|
/* Only lpspi slave on imx93 and imx95 need using the special edma configuration */
|
|
if (fsl_lpspi->is_target && (is_imx95_lpspi(fsl_lpspi) || is_imx93_lpspi(fsl_lpspi))) {
|
|
fsl_lpspi->cur_transfer = transfer;
|
|
ret = fsl_lpspi_dma_configure(controller);
|
|
if (ret)
|
|
return ret;
|
|
|
|
rx = fsl_lpspi_allocate_sg_for_target(controller, DMA_FROM_DEVICE);
|
|
if (IS_ERR(rx)) {
|
|
dev_err(&controller->dev, "DMA allocate RX sgtable failed: %ld\n",
|
|
PTR_ERR(rx));
|
|
return PTR_ERR(rx);
|
|
}
|
|
|
|
tx = fsl_lpspi_allocate_sg_for_target(controller, DMA_TO_DEVICE);
|
|
if (IS_ERR(tx)) {
|
|
dev_err(&controller->dev, "DMA allocate TX sgtable failed: %ld\n",
|
|
PTR_ERR(tx));
|
|
return PTR_ERR(tx);
|
|
}
|
|
} else {
|
|
ret = fsl_lpspi_dma_configure(controller);
|
|
if (ret)
|
|
return ret;
|
|
|
|
tx = &transfer->tx_sg;
|
|
rx = &transfer->rx_sg;
|
|
}
|
|
|
|
desc_rx = dmaengine_prep_slave_sg(controller->dma_rx,
|
|
rx->sgl, rx->nents, DMA_DEV_TO_MEM,
|
|
DMA_PREP_INTERRUPT | DMA_CTRL_ACK);
|
|
if (!desc_rx)
|
|
return -EINVAL;
|
|
|
|
desc_rx->callback = fsl_lpspi_dma_rx_callback;
|
|
desc_rx->callback_param = (void *)fsl_lpspi;
|
|
dmaengine_submit(desc_rx);
|
|
reinit_completion(&fsl_lpspi->dma_rx_completion);
|
|
dma_async_issue_pending(controller->dma_rx);
|
|
|
|
desc_tx = dmaengine_prep_slave_sg(controller->dma_tx,
|
|
tx->sgl, tx->nents, DMA_MEM_TO_DEV,
|
|
DMA_PREP_INTERRUPT | DMA_CTRL_ACK);
|
|
if (!desc_tx) {
|
|
dmaengine_terminate_all(controller->dma_tx);
|
|
return -EINVAL;
|
|
}
|
|
|
|
desc_tx->callback = fsl_lpspi_dma_tx_callback;
|
|
desc_tx->callback_param = (void *)fsl_lpspi;
|
|
dmaengine_submit(desc_tx);
|
|
reinit_completion(&fsl_lpspi->dma_tx_completion);
|
|
dma_async_issue_pending(controller->dma_tx);
|
|
|
|
fsl_lpspi->target_aborted = false;
|
|
|
|
if (!fsl_lpspi->is_target) {
|
|
transfer_timeout = fsl_lpspi_calculate_timeout(fsl_lpspi,
|
|
transfer->len);
|
|
|
|
/* Wait eDMA to finish the data transfer.*/
|
|
timeout = wait_for_completion_timeout(&fsl_lpspi->dma_tx_completion,
|
|
transfer_timeout);
|
|
if (!timeout) {
|
|
dev_err(fsl_lpspi->dev, "I/O Error in DMA TX\n");
|
|
dmaengine_terminate_all(controller->dma_tx);
|
|
dmaengine_terminate_all(controller->dma_rx);
|
|
fsl_lpspi_reset(fsl_lpspi);
|
|
return -ETIMEDOUT;
|
|
}
|
|
|
|
timeout = wait_for_completion_timeout(&fsl_lpspi->dma_rx_completion,
|
|
transfer_timeout);
|
|
if (!timeout) {
|
|
dev_err(fsl_lpspi->dev, "I/O Error in DMA RX\n");
|
|
dmaengine_terminate_all(controller->dma_tx);
|
|
dmaengine_terminate_all(controller->dma_rx);
|
|
fsl_lpspi_reset(fsl_lpspi);
|
|
return -ETIMEDOUT;
|
|
}
|
|
} else {
|
|
if (wait_for_completion_interruptible(&fsl_lpspi->dma_tx_completion) ||
|
|
fsl_lpspi->target_aborted) {
|
|
dev_dbg(fsl_lpspi->dev,
|
|
"I/O Error in DMA TX interrupted\n");
|
|
dmaengine_terminate_all(controller->dma_tx);
|
|
dmaengine_terminate_all(controller->dma_rx);
|
|
fsl_lpspi_reset(fsl_lpspi);
|
|
return -EINTR;
|
|
}
|
|
|
|
if (wait_for_completion_interruptible(&fsl_lpspi->dma_rx_completion) ||
|
|
fsl_lpspi->target_aborted) {
|
|
dev_dbg(fsl_lpspi->dev,
|
|
"I/O Error in DMA RX interrupted\n");
|
|
dmaengine_terminate_all(controller->dma_tx);
|
|
dmaengine_terminate_all(controller->dma_rx);
|
|
fsl_lpspi_reset(fsl_lpspi);
|
|
return -EINTR;
|
|
}
|
|
}
|
|
|
|
fsl_lpspi_reset(fsl_lpspi);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void fsl_lpspi_dma_exit(struct spi_controller *controller)
|
|
{
|
|
if (controller->dma_rx) {
|
|
dma_release_channel(controller->dma_rx);
|
|
controller->dma_rx = NULL;
|
|
}
|
|
|
|
if (controller->dma_tx) {
|
|
dma_release_channel(controller->dma_tx);
|
|
controller->dma_tx = NULL;
|
|
}
|
|
}
|
|
|
|
static int fsl_lpspi_dma_init(struct device *dev,
|
|
struct fsl_lpspi_data *fsl_lpspi,
|
|
struct spi_controller *controller)
|
|
{
|
|
int ret;
|
|
|
|
/* Prepare for TX DMA: */
|
|
controller->dma_tx = dma_request_chan(dev, "tx");
|
|
if (IS_ERR(controller->dma_tx)) {
|
|
ret = PTR_ERR(controller->dma_tx);
|
|
dev_dbg(dev, "can't get the TX DMA channel, error %d!\n", ret);
|
|
controller->dma_tx = NULL;
|
|
goto err;
|
|
}
|
|
|
|
/* Prepare for RX DMA: */
|
|
controller->dma_rx = dma_request_chan(dev, "rx");
|
|
if (IS_ERR(controller->dma_rx)) {
|
|
ret = PTR_ERR(controller->dma_rx);
|
|
dev_dbg(dev, "can't get the RX DMA channel, error %d\n", ret);
|
|
controller->dma_rx = NULL;
|
|
goto err;
|
|
}
|
|
|
|
init_completion(&fsl_lpspi->dma_rx_completion);
|
|
init_completion(&fsl_lpspi->dma_tx_completion);
|
|
controller->can_dma = fsl_lpspi_can_dma;
|
|
controller->max_dma_len = FSL_LPSPI_MAX_EDMA_BYTES;
|
|
|
|
return 0;
|
|
err:
|
|
fsl_lpspi_dma_exit(controller);
|
|
return ret;
|
|
}
|
|
|
|
static int fsl_lpspi_pio_transfer(struct spi_controller *controller,
|
|
struct spi_transfer *t)
|
|
{
|
|
struct fsl_lpspi_data *fsl_lpspi =
|
|
spi_controller_get_devdata(controller);
|
|
int ret;
|
|
|
|
fsl_lpspi->tx_buf = t->tx_buf;
|
|
fsl_lpspi->rx_buf = t->rx_buf;
|
|
fsl_lpspi->remain = t->len;
|
|
|
|
reinit_completion(&fsl_lpspi->xfer_done);
|
|
fsl_lpspi->target_aborted = false;
|
|
|
|
fsl_lpspi_write_tx_fifo(fsl_lpspi);
|
|
|
|
ret = fsl_lpspi_wait_for_completion(controller);
|
|
if (ret)
|
|
return ret;
|
|
|
|
fsl_lpspi_reset(fsl_lpspi);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int fsl_lpspi_transfer_one(struct spi_controller *controller,
|
|
struct spi_device *spi,
|
|
struct spi_transfer *t)
|
|
{
|
|
struct fsl_lpspi_data *fsl_lpspi =
|
|
spi_controller_get_devdata(controller);
|
|
int ret;
|
|
|
|
/*
|
|
* Reset FIFO and clear flags when start TO transfer to avoid
|
|
* being affected if there is a previous abnormal transmission.
|
|
*/
|
|
fsl_lpspi_reset(fsl_lpspi);
|
|
|
|
fsl_lpspi->is_first_byte = true;
|
|
ret = fsl_lpspi_setup_transfer(controller, spi, t);
|
|
if (ret < 0)
|
|
return ret;
|
|
|
|
fsl_lpspi_set_cmd(fsl_lpspi);
|
|
fsl_lpspi->is_first_byte = false;
|
|
|
|
if (fsl_lpspi->usedma)
|
|
ret = fsl_lpspi_dma_transfer(controller, fsl_lpspi, t);
|
|
else
|
|
ret = fsl_lpspi_pio_transfer(controller, t);
|
|
if (ret < 0)
|
|
return ret;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static irqreturn_t fsl_lpspi_isr(int irq, void *dev_id)
|
|
{
|
|
u32 temp_SR, temp_IER;
|
|
struct fsl_lpspi_data *fsl_lpspi = dev_id;
|
|
|
|
temp_IER = readl(fsl_lpspi->base + IMX7ULP_IER);
|
|
fsl_lpspi_intctrl(fsl_lpspi, 0);
|
|
temp_SR = readl(fsl_lpspi->base + IMX7ULP_SR);
|
|
|
|
fsl_lpspi_read_rx_fifo(fsl_lpspi);
|
|
|
|
if ((temp_SR & SR_TDF) && (temp_IER & IER_TDIE)) {
|
|
fsl_lpspi_write_tx_fifo(fsl_lpspi);
|
|
return IRQ_HANDLED;
|
|
}
|
|
|
|
if (temp_SR & SR_MBF ||
|
|
readl(fsl_lpspi->base + IMX7ULP_FSR) & FSR_TXCOUNT) {
|
|
writel(SR_FCF, fsl_lpspi->base + IMX7ULP_SR);
|
|
fsl_lpspi_intctrl(fsl_lpspi, IER_FCIE);
|
|
return IRQ_HANDLED;
|
|
}
|
|
|
|
if (temp_SR & SR_FCF && (temp_IER & IER_FCIE)) {
|
|
writel(SR_FCF, fsl_lpspi->base + IMX7ULP_SR);
|
|
complete(&fsl_lpspi->xfer_done);
|
|
return IRQ_HANDLED;
|
|
}
|
|
|
|
return IRQ_NONE;
|
|
}
|
|
|
|
#ifdef CONFIG_PM
|
|
static int fsl_lpspi_runtime_resume(struct device *dev)
|
|
{
|
|
struct spi_controller *controller = dev_get_drvdata(dev);
|
|
struct fsl_lpspi_data *fsl_lpspi;
|
|
int ret;
|
|
|
|
fsl_lpspi = spi_controller_get_devdata(controller);
|
|
|
|
ret = clk_prepare_enable(fsl_lpspi->clk_per);
|
|
if (ret)
|
|
return ret;
|
|
|
|
ret = clk_prepare_enable(fsl_lpspi->clk_ipg);
|
|
if (ret) {
|
|
clk_disable_unprepare(fsl_lpspi->clk_per);
|
|
return ret;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int fsl_lpspi_runtime_suspend(struct device *dev)
|
|
{
|
|
struct spi_controller *controller = dev_get_drvdata(dev);
|
|
struct fsl_lpspi_data *fsl_lpspi;
|
|
|
|
fsl_lpspi = spi_controller_get_devdata(controller);
|
|
|
|
clk_disable_unprepare(fsl_lpspi->clk_per);
|
|
clk_disable_unprepare(fsl_lpspi->clk_ipg);
|
|
|
|
return 0;
|
|
}
|
|
#endif
|
|
|
|
static int fsl_lpspi_init_rpm(struct fsl_lpspi_data *fsl_lpspi)
|
|
{
|
|
struct device *dev = fsl_lpspi->dev;
|
|
|
|
pm_runtime_enable(dev);
|
|
pm_runtime_set_autosuspend_delay(dev, FSL_LPSPI_RPM_TIMEOUT);
|
|
pm_runtime_use_autosuspend(dev);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int fsl_lpspi_probe(struct platform_device *pdev)
|
|
{
|
|
const struct fsl_lpspi_devtype_data *devtype_data;
|
|
struct fsl_lpspi_data *fsl_lpspi;
|
|
struct spi_controller *controller;
|
|
struct resource *res;
|
|
int ret, irq;
|
|
u32 num_cs;
|
|
u32 temp;
|
|
bool is_target;
|
|
|
|
devtype_data = of_device_get_match_data(&pdev->dev);
|
|
if (!devtype_data)
|
|
return -ENODEV;
|
|
|
|
is_target = of_property_read_bool((&pdev->dev)->of_node, "spi-slave");
|
|
if (is_target)
|
|
controller = devm_spi_alloc_target(&pdev->dev,
|
|
sizeof(struct fsl_lpspi_data));
|
|
else
|
|
controller = devm_spi_alloc_host(&pdev->dev,
|
|
sizeof(struct fsl_lpspi_data));
|
|
|
|
if (!controller)
|
|
return -ENOMEM;
|
|
|
|
platform_set_drvdata(pdev, controller);
|
|
|
|
fsl_lpspi = spi_controller_get_devdata(controller);
|
|
fsl_lpspi->dev = &pdev->dev;
|
|
fsl_lpspi->is_target = is_target;
|
|
fsl_lpspi->is_only_cs1 = of_property_read_bool((&pdev->dev)->of_node,
|
|
"fsl,spi-only-use-cs1-sel");
|
|
fsl_lpspi->devtype_data = devtype_data;
|
|
|
|
init_completion(&fsl_lpspi->xfer_done);
|
|
|
|
fsl_lpspi->base = devm_platform_get_and_ioremap_resource(pdev, 0, &res);
|
|
if (IS_ERR(fsl_lpspi->base)) {
|
|
ret = PTR_ERR(fsl_lpspi->base);
|
|
return ret;
|
|
}
|
|
fsl_lpspi->base_phys = res->start;
|
|
|
|
irq = platform_get_irq(pdev, 0);
|
|
if (irq < 0) {
|
|
ret = irq;
|
|
return ret;
|
|
}
|
|
|
|
ret = devm_request_irq(&pdev->dev, irq, fsl_lpspi_isr, 0,
|
|
dev_name(&pdev->dev), fsl_lpspi);
|
|
if (ret) {
|
|
dev_err(&pdev->dev, "can't get irq%d: %d\n", irq, ret);
|
|
return ret;
|
|
}
|
|
|
|
fsl_lpspi->clk_per = devm_clk_get(&pdev->dev, "per");
|
|
if (IS_ERR(fsl_lpspi->clk_per)) {
|
|
ret = PTR_ERR(fsl_lpspi->clk_per);
|
|
return ret;
|
|
}
|
|
|
|
fsl_lpspi->clk_ipg = devm_clk_get(&pdev->dev, "ipg");
|
|
if (IS_ERR(fsl_lpspi->clk_ipg)) {
|
|
ret = PTR_ERR(fsl_lpspi->clk_ipg);
|
|
return ret;
|
|
}
|
|
|
|
/* enable the clock */
|
|
ret = fsl_lpspi_init_rpm(fsl_lpspi);
|
|
if (ret)
|
|
return ret;
|
|
|
|
ret = pm_runtime_get_sync(fsl_lpspi->dev);
|
|
if (ret < 0) {
|
|
dev_err(fsl_lpspi->dev, "failed to enable clock\n");
|
|
goto out_pm_get;
|
|
}
|
|
|
|
temp = readl(fsl_lpspi->base + IMX7ULP_PARAM);
|
|
fsl_lpspi->txfifosize = 1 << (temp & 0x0f);
|
|
fsl_lpspi->rxfifosize = 1 << ((temp >> 8) & 0x0f);
|
|
if (of_property_read_u32((&pdev->dev)->of_node, "num-cs",
|
|
&num_cs)) {
|
|
if (of_device_is_compatible(pdev->dev.of_node, "fsl,imx93-spi"))
|
|
num_cs = ((temp >> 16) & 0xf);
|
|
else
|
|
num_cs = 1;
|
|
}
|
|
|
|
controller->bits_per_word_mask = SPI_BPW_RANGE_MASK(8, 32);
|
|
controller->transfer_one = fsl_lpspi_transfer_one;
|
|
controller->prepare_transfer_hardware = lpspi_prepare_xfer_hardware;
|
|
controller->unprepare_transfer_hardware = lpspi_unprepare_xfer_hardware;
|
|
controller->mode_bits = SPI_CPOL | SPI_CPHA | SPI_CS_HIGH;
|
|
controller->flags = SPI_CONTROLLER_MUST_RX | SPI_CONTROLLER_MUST_TX;
|
|
controller->dev.of_node = pdev->dev.of_node;
|
|
controller->bus_num = pdev->id;
|
|
controller->num_chipselect = num_cs;
|
|
controller->target_abort = fsl_lpspi_target_abort;
|
|
if (!fsl_lpspi->is_target)
|
|
controller->use_gpio_descriptors = true;
|
|
|
|
ret = fsl_lpspi_dma_init(&pdev->dev, fsl_lpspi, controller);
|
|
if (ret == -EPROBE_DEFER)
|
|
goto out_pm_get;
|
|
if (ret < 0)
|
|
dev_warn(&pdev->dev, "dma setup error %d, use pio\n", ret);
|
|
else
|
|
/*
|
|
* disable LPSPI module IRQ when enable DMA mode successfully,
|
|
* to prevent the unexpected LPSPI module IRQ events.
|
|
*/
|
|
disable_irq(irq);
|
|
|
|
ret = devm_spi_register_controller(&pdev->dev, controller);
|
|
if (ret < 0) {
|
|
dev_err_probe(&pdev->dev, ret, "spi_register_controller error\n");
|
|
goto free_dma;
|
|
}
|
|
|
|
pm_runtime_mark_last_busy(fsl_lpspi->dev);
|
|
pm_runtime_put_autosuspend(fsl_lpspi->dev);
|
|
|
|
return 0;
|
|
|
|
free_dma:
|
|
fsl_lpspi_dma_exit(controller);
|
|
out_pm_get:
|
|
pm_runtime_dont_use_autosuspend(fsl_lpspi->dev);
|
|
pm_runtime_put_sync(fsl_lpspi->dev);
|
|
pm_runtime_disable(fsl_lpspi->dev);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static void fsl_lpspi_remove(struct platform_device *pdev)
|
|
{
|
|
struct spi_controller *controller = platform_get_drvdata(pdev);
|
|
struct fsl_lpspi_data *fsl_lpspi =
|
|
spi_controller_get_devdata(controller);
|
|
|
|
fsl_lpspi_dma_exit(controller);
|
|
|
|
pm_runtime_disable(fsl_lpspi->dev);
|
|
}
|
|
|
|
static int __maybe_unused fsl_lpspi_suspend(struct device *dev)
|
|
{
|
|
pinctrl_pm_select_sleep_state(dev);
|
|
return pm_runtime_force_suspend(dev);
|
|
}
|
|
|
|
static int __maybe_unused fsl_lpspi_resume(struct device *dev)
|
|
{
|
|
int ret;
|
|
|
|
ret = pm_runtime_force_resume(dev);
|
|
if (ret) {
|
|
dev_err(dev, "Error in resume: %d\n", ret);
|
|
return ret;
|
|
}
|
|
|
|
pinctrl_pm_select_default_state(dev);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static const struct dev_pm_ops fsl_lpspi_pm_ops = {
|
|
SET_RUNTIME_PM_OPS(fsl_lpspi_runtime_suspend,
|
|
fsl_lpspi_runtime_resume, NULL)
|
|
SET_SYSTEM_SLEEP_PM_OPS(fsl_lpspi_suspend, fsl_lpspi_resume)
|
|
};
|
|
|
|
static struct platform_driver fsl_lpspi_driver = {
|
|
.driver = {
|
|
.name = DRIVER_NAME,
|
|
.of_match_table = fsl_lpspi_dt_ids,
|
|
.pm = &fsl_lpspi_pm_ops,
|
|
},
|
|
.probe = fsl_lpspi_probe,
|
|
.remove_new = fsl_lpspi_remove,
|
|
};
|
|
module_platform_driver(fsl_lpspi_driver);
|
|
|
|
MODULE_DESCRIPTION("LPSPI Controller driver");
|
|
MODULE_AUTHOR("Gao Pan <pandy.gao@nxp.com>");
|
|
MODULE_LICENSE("GPL");
|