mirror of
https://github.com/nxp-imx/linux-imx.git
synced 2025-10-22 23:23:03 +02:00
Revert fsl-edma to the version of next-20230804
The working on the verification of upstream edma driver is still in progress. Let's use the local version first for the rebase. Revert "dmaengine: fsl-edma: integrate v3 support" This reverts commit72f5801a4e
. Revert "dmaengine: fsl-edma: move tcd into struct fsl_dma_chan" This reverts commit7536f8b371
. Revert "dmaengine: fsl-edma: refactor chan_name setup and safety" This reverts commit9b05554c5c
. Revert "dmaengine: fsl-edma: move clearing of register interrupt into setup_irq function" This reverts commitf5b3ba52f3
. Revert "dmaengine: fsl-edma: refactor using devm_clk_get_enabled" This reverts commita9903de3aa
. Revert "dmaengine: fsl-edma: simply ATTR_DSIZE and ATTR_SSIZE by using ffs()" This reverts commitee2dda0646
. Revert "dmaengine: fsl-edma: move common IRQ handler to common.c" This reverts commit79434f9b97
. Revert "dmaengine: fsl-edma: Remove enum edma_version" This reverts commitc26e611433
. Revert "dmaengine: fsl-edma: transition from bool fields to bitmask flags in drvdata" This reverts commit9e006b2439
. Revert "dmaengine: fsl-edma: clean up EXPORT_SYMBOL_GPL in fsl-edma-common.c" This reverts commit66aac8ea0a
.
This commit is contained in:
parent
9d7e24bcc7
commit
17508933c5
|
@ -32,10 +32,8 @@ obj-$(CONFIG_DW_DMAC_CORE) += dw/
|
|||
obj-$(CONFIG_DW_EDMA) += dw-edma/
|
||||
obj-$(CONFIG_EP93XX_DMA) += ep93xx_dma.o
|
||||
obj-$(CONFIG_FSL_DMA) += fsldma.o
|
||||
fsl-edma-objs := fsl-edma-main.o fsl-edma-common.o
|
||||
obj-$(CONFIG_FSL_EDMA) += fsl-edma.o
|
||||
mcf-edma-objs := mcf-edma-main.o fsl-edma-common.o
|
||||
obj-$(CONFIG_MCF_EDMA) += mcf-edma.o
|
||||
obj-$(CONFIG_FSL_EDMA) += fsl-edma.o fsl-edma-common.o
|
||||
obj-$(CONFIG_MCF_EDMA) += mcf-edma.o fsl-edma-common.o
|
||||
obj-$(CONFIG_FSL_QDMA) += fsl-qdma.o
|
||||
obj-$(CONFIG_FSL_RAID) += fsl_raid.o
|
||||
obj-$(CONFIG_HISI_DMA) += hisi_dma.o
|
||||
|
|
|
@ -7,8 +7,6 @@
|
|||
#include <linux/module.h>
|
||||
#include <linux/slab.h>
|
||||
#include <linux/dma-mapping.h>
|
||||
#include <linux/pm_runtime.h>
|
||||
#include <linux/pm_domain.h>
|
||||
|
||||
#include "fsl-edma-common.h"
|
||||
|
||||
|
@ -42,73 +40,14 @@
|
|||
#define EDMA64_ERRH 0x28
|
||||
#define EDMA64_ERRL 0x2c
|
||||
|
||||
void fsl_edma_tx_chan_handler(struct fsl_edma_chan *fsl_chan)
|
||||
{
|
||||
spin_lock(&fsl_chan->vchan.lock);
|
||||
|
||||
if (!fsl_chan->edesc) {
|
||||
/* terminate_all called before */
|
||||
spin_unlock(&fsl_chan->vchan.lock);
|
||||
return;
|
||||
}
|
||||
|
||||
if (!fsl_chan->edesc->iscyclic) {
|
||||
list_del(&fsl_chan->edesc->vdesc.node);
|
||||
vchan_cookie_complete(&fsl_chan->edesc->vdesc);
|
||||
fsl_chan->edesc = NULL;
|
||||
fsl_chan->status = DMA_COMPLETE;
|
||||
fsl_chan->idle = true;
|
||||
} else {
|
||||
vchan_cyclic_callback(&fsl_chan->edesc->vdesc);
|
||||
}
|
||||
|
||||
if (!fsl_chan->edesc)
|
||||
fsl_edma_xfer_desc(fsl_chan);
|
||||
|
||||
spin_unlock(&fsl_chan->vchan.lock);
|
||||
}
|
||||
|
||||
static void fsl_edma3_enable_request(struct fsl_edma_chan *fsl_chan)
|
||||
{
|
||||
u32 val, flags;
|
||||
|
||||
flags = fsl_edma_drvflags(fsl_chan);
|
||||
val = edma_readl_chreg(fsl_chan, ch_sbr);
|
||||
/* Remote/local swapped wrongly on iMX8 QM Audio edma */
|
||||
if (flags & FSL_EDMA_DRV_QUIRK_SWAPPED) {
|
||||
if (!fsl_chan->is_rxchan)
|
||||
val |= EDMA_V3_CH_SBR_RD;
|
||||
else
|
||||
val |= EDMA_V3_CH_SBR_WR;
|
||||
} else {
|
||||
if (fsl_chan->is_rxchan)
|
||||
val |= EDMA_V3_CH_SBR_RD;
|
||||
else
|
||||
val |= EDMA_V3_CH_SBR_WR;
|
||||
}
|
||||
|
||||
if (fsl_chan->is_remote)
|
||||
val &= ~(EDMA_V3_CH_SBR_RD | EDMA_V3_CH_SBR_WR);
|
||||
|
||||
edma_writel_chreg(fsl_chan, val, ch_sbr);
|
||||
|
||||
if (flags & FSL_EDMA_DRV_HAS_CHMUX)
|
||||
edma_writel_chreg(fsl_chan, fsl_chan->srcid, ch_mux);
|
||||
|
||||
val = edma_readl_chreg(fsl_chan, ch_csr);
|
||||
val |= EDMA_V3_CH_CSR_ERQ;
|
||||
edma_writel_chreg(fsl_chan, val, ch_csr);
|
||||
}
|
||||
#define EDMA_TCD 0x1000
|
||||
|
||||
static void fsl_edma_enable_request(struct fsl_edma_chan *fsl_chan)
|
||||
{
|
||||
struct edma_regs *regs = &fsl_chan->edma->regs;
|
||||
u32 ch = fsl_chan->vchan.chan.chan_id;
|
||||
|
||||
if (fsl_edma_drvflags(fsl_chan) & FSL_EDMA_DRV_SPLIT_REG)
|
||||
return fsl_edma3_enable_request(fsl_chan);
|
||||
|
||||
if (fsl_chan->edma->drvdata->flags & FSL_EDMA_DRV_WRAP_IO) {
|
||||
if (fsl_chan->edma->drvdata->version == v1) {
|
||||
edma_writeb(fsl_chan->edma, EDMA_SEEI_SEEI(ch), regs->seei);
|
||||
edma_writeb(fsl_chan->edma, ch, regs->serq);
|
||||
} else {
|
||||
|
@ -120,29 +59,12 @@ static void fsl_edma_enable_request(struct fsl_edma_chan *fsl_chan)
|
|||
}
|
||||
}
|
||||
|
||||
static void fsl_edma3_disable_request(struct fsl_edma_chan *fsl_chan)
|
||||
{
|
||||
u32 val = edma_readl_chreg(fsl_chan, ch_csr);
|
||||
u32 flags;
|
||||
|
||||
flags = fsl_edma_drvflags(fsl_chan);
|
||||
|
||||
if (flags & FSL_EDMA_DRV_HAS_CHMUX)
|
||||
edma_writel_chreg(fsl_chan, 0, ch_mux);
|
||||
|
||||
val &= ~EDMA_V3_CH_CSR_ERQ;
|
||||
edma_writel_chreg(fsl_chan, val, ch_csr);
|
||||
}
|
||||
|
||||
void fsl_edma_disable_request(struct fsl_edma_chan *fsl_chan)
|
||||
{
|
||||
struct edma_regs *regs = &fsl_chan->edma->regs;
|
||||
u32 ch = fsl_chan->vchan.chan.chan_id;
|
||||
|
||||
if (fsl_edma_drvflags(fsl_chan) & FSL_EDMA_DRV_SPLIT_REG)
|
||||
return fsl_edma3_disable_request(fsl_chan);
|
||||
|
||||
if (fsl_chan->edma->drvdata->flags & FSL_EDMA_DRV_WRAP_IO) {
|
||||
if (fsl_chan->edma->drvdata->version == v1) {
|
||||
edma_writeb(fsl_chan->edma, ch, regs->cerq);
|
||||
edma_writeb(fsl_chan->edma, EDMA_CEEI_CEEI(ch), regs->ceei);
|
||||
} else {
|
||||
|
@ -153,6 +75,7 @@ void fsl_edma_disable_request(struct fsl_edma_chan *fsl_chan)
|
|||
iowrite8(EDMA_CEEI_CEEI(ch), regs->ceei);
|
||||
}
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(fsl_edma_disable_request);
|
||||
|
||||
static void mux_configure8(struct fsl_edma_chan *fsl_chan, void __iomem *addr,
|
||||
u32 off, u32 slot, bool enable)
|
||||
|
@ -189,33 +112,36 @@ void fsl_edma_chan_mux(struct fsl_edma_chan *fsl_chan,
|
|||
int endian_diff[4] = {3, 1, -1, -3};
|
||||
u32 dmamux_nr = fsl_chan->edma->drvdata->dmamuxs;
|
||||
|
||||
if (!dmamux_nr)
|
||||
return;
|
||||
|
||||
chans_per_mux = fsl_chan->edma->n_chans / dmamux_nr;
|
||||
ch_off = fsl_chan->vchan.chan.chan_id % chans_per_mux;
|
||||
|
||||
if (fsl_chan->edma->drvdata->flags & FSL_EDMA_DRV_MUX_SWAP)
|
||||
if (fsl_chan->edma->drvdata->mux_swap)
|
||||
ch_off += endian_diff[ch_off % 4];
|
||||
|
||||
muxaddr = fsl_chan->edma->muxbase[ch / chans_per_mux];
|
||||
slot = EDMAMUX_CHCFG_SOURCE(slot);
|
||||
|
||||
if (fsl_chan->edma->drvdata->flags & FSL_EDMA_DRV_CONFIG32)
|
||||
if (fsl_chan->edma->drvdata->version == v3)
|
||||
mux_configure32(fsl_chan, muxaddr, ch_off, slot, enable);
|
||||
else
|
||||
mux_configure8(fsl_chan, muxaddr, ch_off, slot, enable);
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(fsl_edma_chan_mux);
|
||||
|
||||
static unsigned int fsl_edma_get_tcd_attr(enum dma_slave_buswidth addr_width)
|
||||
{
|
||||
u32 val;
|
||||
|
||||
if (addr_width == DMA_SLAVE_BUSWIDTH_UNDEFINED)
|
||||
addr_width = DMA_SLAVE_BUSWIDTH_4_BYTES;
|
||||
|
||||
val = ffs(addr_width) - 1;
|
||||
return val | (val << 8);
|
||||
switch (addr_width) {
|
||||
case 1:
|
||||
return EDMA_TCD_ATTR_SSIZE_8BIT | EDMA_TCD_ATTR_DSIZE_8BIT;
|
||||
case 2:
|
||||
return EDMA_TCD_ATTR_SSIZE_16BIT | EDMA_TCD_ATTR_DSIZE_16BIT;
|
||||
case 4:
|
||||
return EDMA_TCD_ATTR_SSIZE_32BIT | EDMA_TCD_ATTR_DSIZE_32BIT;
|
||||
case 8:
|
||||
return EDMA_TCD_ATTR_SSIZE_64BIT | EDMA_TCD_ATTR_DSIZE_64BIT;
|
||||
default:
|
||||
return EDMA_TCD_ATTR_SSIZE_32BIT | EDMA_TCD_ATTR_DSIZE_32BIT;
|
||||
}
|
||||
}
|
||||
|
||||
void fsl_edma_free_desc(struct virt_dma_desc *vdesc)
|
||||
|
@ -229,6 +155,7 @@ void fsl_edma_free_desc(struct virt_dma_desc *vdesc)
|
|||
fsl_desc->tcd[i].ptcd);
|
||||
kfree(fsl_desc);
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(fsl_edma_free_desc);
|
||||
|
||||
int fsl_edma_terminate_all(struct dma_chan *chan)
|
||||
{
|
||||
|
@ -243,12 +170,9 @@ int fsl_edma_terminate_all(struct dma_chan *chan)
|
|||
vchan_get_all_descriptors(&fsl_chan->vchan, &head);
|
||||
spin_unlock_irqrestore(&fsl_chan->vchan.lock, flags);
|
||||
vchan_dma_desc_free_list(&fsl_chan->vchan, &head);
|
||||
|
||||
if (fsl_edma_drvflags(fsl_chan) & FSL_EDMA_DRV_HAS_PD)
|
||||
pm_runtime_allow(fsl_chan->pd_dev);
|
||||
|
||||
return 0;
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(fsl_edma_terminate_all);
|
||||
|
||||
int fsl_edma_pause(struct dma_chan *chan)
|
||||
{
|
||||
|
@ -264,6 +188,7 @@ int fsl_edma_pause(struct dma_chan *chan)
|
|||
spin_unlock_irqrestore(&fsl_chan->vchan.lock, flags);
|
||||
return 0;
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(fsl_edma_pause);
|
||||
|
||||
int fsl_edma_resume(struct dma_chan *chan)
|
||||
{
|
||||
|
@ -279,6 +204,7 @@ int fsl_edma_resume(struct dma_chan *chan)
|
|||
spin_unlock_irqrestore(&fsl_chan->vchan.lock, flags);
|
||||
return 0;
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(fsl_edma_resume);
|
||||
|
||||
static void fsl_edma_unprep_slave_dma(struct fsl_edma_chan *fsl_chan)
|
||||
{
|
||||
|
@ -339,41 +265,36 @@ int fsl_edma_slave_config(struct dma_chan *chan,
|
|||
|
||||
return 0;
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(fsl_edma_slave_config);
|
||||
|
||||
static size_t fsl_edma_desc_residue(struct fsl_edma_chan *fsl_chan,
|
||||
struct virt_dma_desc *vdesc, bool in_progress)
|
||||
{
|
||||
struct fsl_edma_desc *edesc = fsl_chan->edesc;
|
||||
struct edma_regs *regs = &fsl_chan->edma->regs;
|
||||
u32 ch = fsl_chan->vchan.chan.chan_id;
|
||||
enum dma_transfer_direction dir = edesc->dirn;
|
||||
dma_addr_t cur_addr, dma_addr;
|
||||
size_t len, size;
|
||||
u32 nbytes = 0;
|
||||
int i;
|
||||
|
||||
/* calculate the total size in this desc */
|
||||
for (len = i = 0; i < fsl_chan->edesc->n_tcds; i++) {
|
||||
nbytes = le32_to_cpu(edesc->tcd[i].vtcd->nbytes);
|
||||
if (nbytes & (EDMA_V3_TCD_NBYTES_DMLOE | EDMA_V3_TCD_NBYTES_SMLOE))
|
||||
nbytes = EDMA_V3_TCD_NBYTES_MLOFF_NBYTES(nbytes);
|
||||
len += nbytes * le16_to_cpu(edesc->tcd[i].vtcd->biter);
|
||||
}
|
||||
for (len = i = 0; i < fsl_chan->edesc->n_tcds; i++)
|
||||
len += le32_to_cpu(edesc->tcd[i].vtcd->nbytes)
|
||||
* le16_to_cpu(edesc->tcd[i].vtcd->biter);
|
||||
|
||||
if (!in_progress)
|
||||
return len;
|
||||
|
||||
if (dir == DMA_MEM_TO_DEV)
|
||||
cur_addr = edma_read_tcdreg(fsl_chan, saddr);
|
||||
cur_addr = edma_readl(fsl_chan->edma, ®s->tcd[ch].saddr);
|
||||
else
|
||||
cur_addr = edma_read_tcdreg(fsl_chan, daddr);
|
||||
cur_addr = edma_readl(fsl_chan->edma, ®s->tcd[ch].daddr);
|
||||
|
||||
/* figure out the finished and calculate the residue */
|
||||
for (i = 0; i < fsl_chan->edesc->n_tcds; i++) {
|
||||
nbytes = le32_to_cpu(edesc->tcd[i].vtcd->nbytes);
|
||||
if (nbytes & (EDMA_V3_TCD_NBYTES_DMLOE | EDMA_V3_TCD_NBYTES_SMLOE))
|
||||
nbytes = EDMA_V3_TCD_NBYTES_MLOFF_NBYTES(nbytes);
|
||||
|
||||
size = nbytes * le16_to_cpu(edesc->tcd[i].vtcd->biter);
|
||||
|
||||
size = le32_to_cpu(edesc->tcd[i].vtcd->nbytes)
|
||||
* le16_to_cpu(edesc->tcd[i].vtcd->biter);
|
||||
if (dir == DMA_MEM_TO_DEV)
|
||||
dma_addr = le32_to_cpu(edesc->tcd[i].vtcd->saddr);
|
||||
else
|
||||
|
@ -419,10 +340,14 @@ enum dma_status fsl_edma_tx_status(struct dma_chan *chan,
|
|||
|
||||
return fsl_chan->status;
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(fsl_edma_tx_status);
|
||||
|
||||
static void fsl_edma_set_tcd_regs(struct fsl_edma_chan *fsl_chan,
|
||||
struct fsl_edma_hw_tcd *tcd)
|
||||
{
|
||||
struct fsl_edma_engine *edma = fsl_chan->edma;
|
||||
struct edma_regs *regs = &fsl_chan->edma->regs;
|
||||
u32 ch = fsl_chan->vchan.chan.chan_id;
|
||||
u16 csr = 0;
|
||||
|
||||
/*
|
||||
|
@ -431,22 +356,23 @@ static void fsl_edma_set_tcd_regs(struct fsl_edma_chan *fsl_chan,
|
|||
* big- or little-endian obeying the eDMA engine model endian,
|
||||
* and this is performed from specific edma_write functions
|
||||
*/
|
||||
edma_write_tcdreg(fsl_chan, 0, csr);
|
||||
edma_writew(edma, 0, ®s->tcd[ch].csr);
|
||||
|
||||
edma_write_tcdreg(fsl_chan, tcd->saddr, saddr);
|
||||
edma_write_tcdreg(fsl_chan, tcd->daddr, daddr);
|
||||
edma_writel(edma, (s32)tcd->saddr, ®s->tcd[ch].saddr);
|
||||
edma_writel(edma, (s32)tcd->daddr, ®s->tcd[ch].daddr);
|
||||
|
||||
edma_write_tcdreg(fsl_chan, tcd->attr, attr);
|
||||
edma_write_tcdreg(fsl_chan, tcd->soff, soff);
|
||||
edma_writew(edma, (s16)tcd->attr, ®s->tcd[ch].attr);
|
||||
edma_writew(edma, tcd->soff, ®s->tcd[ch].soff);
|
||||
|
||||
edma_write_tcdreg(fsl_chan, tcd->nbytes, nbytes);
|
||||
edma_write_tcdreg(fsl_chan, tcd->slast, slast);
|
||||
edma_writel(edma, (s32)tcd->nbytes, ®s->tcd[ch].nbytes);
|
||||
edma_writel(edma, (s32)tcd->slast, ®s->tcd[ch].slast);
|
||||
|
||||
edma_write_tcdreg(fsl_chan, tcd->citer, citer);
|
||||
edma_write_tcdreg(fsl_chan, tcd->biter, biter);
|
||||
edma_write_tcdreg(fsl_chan, tcd->doff, doff);
|
||||
edma_writew(edma, (s16)tcd->citer, ®s->tcd[ch].citer);
|
||||
edma_writew(edma, (s16)tcd->biter, ®s->tcd[ch].biter);
|
||||
edma_writew(edma, (s16)tcd->doff, ®s->tcd[ch].doff);
|
||||
|
||||
edma_write_tcdreg(fsl_chan, tcd->dlast_sga, dlast_sga);
|
||||
edma_writel(edma, (s32)tcd->dlast_sga,
|
||||
®s->tcd[ch].dlast_sga);
|
||||
|
||||
if (fsl_chan->is_sw) {
|
||||
csr = le16_to_cpu(tcd->csr);
|
||||
|
@ -454,19 +380,16 @@ static void fsl_edma_set_tcd_regs(struct fsl_edma_chan *fsl_chan,
|
|||
tcd->csr = cpu_to_le16(csr);
|
||||
}
|
||||
|
||||
edma_write_tcdreg(fsl_chan, tcd->csr, csr);
|
||||
edma_writew(edma, (s16)tcd->csr, ®s->tcd[ch].csr);
|
||||
}
|
||||
|
||||
static inline
|
||||
void fsl_edma_fill_tcd(struct fsl_edma_chan *fsl_chan,
|
||||
struct fsl_edma_hw_tcd *tcd, u32 src, u32 dst,
|
||||
void fsl_edma_fill_tcd(struct fsl_edma_hw_tcd *tcd, u32 src, u32 dst,
|
||||
u16 attr, u16 soff, u32 nbytes, u32 slast, u16 citer,
|
||||
u16 biter, u16 doff, u32 dlast_sga, bool major_int,
|
||||
bool disable_req, bool enable_sg)
|
||||
{
|
||||
struct dma_slave_config *cfg = &fsl_chan->cfg;
|
||||
u16 csr = 0;
|
||||
u32 burst;
|
||||
|
||||
/*
|
||||
* eDMA hardware SGs require the TCDs to be stored in little
|
||||
|
@ -481,21 +404,6 @@ void fsl_edma_fill_tcd(struct fsl_edma_chan *fsl_chan,
|
|||
|
||||
tcd->soff = cpu_to_le16(soff);
|
||||
|
||||
if (fsl_chan->is_multi_fifo) {
|
||||
/* set mloff to support multiple fifo */
|
||||
burst = cfg->direction == DMA_DEV_TO_MEM ?
|
||||
cfg->src_addr_width : cfg->dst_addr_width;
|
||||
nbytes |= EDMA_V3_TCD_NBYTES_MLOFF(-(burst * 4));
|
||||
/* enable DMLOE/SMLOE */
|
||||
if (cfg->direction == DMA_MEM_TO_DEV) {
|
||||
nbytes |= EDMA_V3_TCD_NBYTES_DMLOE;
|
||||
nbytes &= ~EDMA_V3_TCD_NBYTES_SMLOE;
|
||||
} else {
|
||||
nbytes |= EDMA_V3_TCD_NBYTES_SMLOE;
|
||||
nbytes &= ~EDMA_V3_TCD_NBYTES_DMLOE;
|
||||
}
|
||||
}
|
||||
|
||||
tcd->nbytes = cpu_to_le32(nbytes);
|
||||
tcd->slast = cpu_to_le32(slast);
|
||||
|
||||
|
@ -514,12 +422,6 @@ void fsl_edma_fill_tcd(struct fsl_edma_chan *fsl_chan,
|
|||
if (enable_sg)
|
||||
csr |= EDMA_TCD_CSR_E_SG;
|
||||
|
||||
if (fsl_chan->is_rxchan)
|
||||
csr |= EDMA_TCD_CSR_ACTIVE;
|
||||
|
||||
if (fsl_chan->is_sw)
|
||||
csr |= EDMA_TCD_CSR_START;
|
||||
|
||||
tcd->csr = cpu_to_le16(csr);
|
||||
}
|
||||
|
||||
|
@ -559,7 +461,6 @@ struct dma_async_tx_descriptor *fsl_edma_prep_dma_cyclic(
|
|||
struct fsl_edma_chan *fsl_chan = to_fsl_edma_chan(chan);
|
||||
struct fsl_edma_desc *fsl_desc;
|
||||
dma_addr_t dma_buf_next;
|
||||
bool major_int = true;
|
||||
int sg_len, i;
|
||||
u32 src_addr, dst_addr, last_sg, nbytes;
|
||||
u16 soff, doff, iter;
|
||||
|
@ -603,28 +504,23 @@ struct dma_async_tx_descriptor *fsl_edma_prep_dma_cyclic(
|
|||
src_addr = dma_buf_next;
|
||||
dst_addr = fsl_chan->dma_dev_addr;
|
||||
soff = fsl_chan->cfg.dst_addr_width;
|
||||
doff = fsl_chan->is_multi_fifo ? 4 : 0;
|
||||
} else if (direction == DMA_DEV_TO_MEM) {
|
||||
doff = 0;
|
||||
} else {
|
||||
src_addr = fsl_chan->dma_dev_addr;
|
||||
dst_addr = dma_buf_next;
|
||||
soff = fsl_chan->is_multi_fifo ? 4 : 0;
|
||||
soff = 0;
|
||||
doff = fsl_chan->cfg.src_addr_width;
|
||||
} else {
|
||||
/* DMA_DEV_TO_DEV */
|
||||
src_addr = fsl_chan->cfg.src_addr;
|
||||
dst_addr = fsl_chan->cfg.dst_addr;
|
||||
soff = doff = 0;
|
||||
major_int = false;
|
||||
}
|
||||
|
||||
fsl_edma_fill_tcd(fsl_chan, fsl_desc->tcd[i].vtcd, src_addr, dst_addr,
|
||||
fsl_edma_fill_tcd(fsl_desc->tcd[i].vtcd, src_addr, dst_addr,
|
||||
fsl_chan->attr, soff, nbytes, 0, iter,
|
||||
iter, doff, last_sg, major_int, false, true);
|
||||
iter, doff, last_sg, true, false, true);
|
||||
dma_buf_next += period_len;
|
||||
}
|
||||
|
||||
return vchan_tx_prep(&fsl_chan->vchan, &fsl_desc->vdesc, flags);
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(fsl_edma_prep_dma_cyclic);
|
||||
|
||||
struct dma_async_tx_descriptor *fsl_edma_prep_slave_sg(
|
||||
struct dma_chan *chan, struct scatterlist *sgl,
|
||||
|
@ -668,51 +564,23 @@ struct dma_async_tx_descriptor *fsl_edma_prep_slave_sg(
|
|||
dst_addr = fsl_chan->dma_dev_addr;
|
||||
soff = fsl_chan->cfg.dst_addr_width;
|
||||
doff = 0;
|
||||
} else if (direction == DMA_DEV_TO_MEM) {
|
||||
} else {
|
||||
src_addr = fsl_chan->dma_dev_addr;
|
||||
dst_addr = sg_dma_address(sg);
|
||||
soff = 0;
|
||||
doff = fsl_chan->cfg.src_addr_width;
|
||||
} else {
|
||||
/* DMA_DEV_TO_DEV */
|
||||
src_addr = fsl_chan->cfg.src_addr;
|
||||
dst_addr = fsl_chan->cfg.dst_addr;
|
||||
soff = 0;
|
||||
doff = 0;
|
||||
}
|
||||
|
||||
/*
|
||||
* Choose the suitable burst length if sg_dma_len is not
|
||||
* multiple of burst length so that the whole transfer length is
|
||||
* multiple of minor loop(burst length).
|
||||
*/
|
||||
if (sg_dma_len(sg) % nbytes) {
|
||||
u32 width = (direction == DMA_DEV_TO_MEM) ? doff : soff;
|
||||
u32 burst = (direction == DMA_DEV_TO_MEM) ?
|
||||
fsl_chan->cfg.src_maxburst :
|
||||
fsl_chan->cfg.dst_maxburst;
|
||||
int j;
|
||||
|
||||
for (j = burst; j > 1; j--) {
|
||||
if (!(sg_dma_len(sg) % (j * width))) {
|
||||
nbytes = j * width;
|
||||
break;
|
||||
}
|
||||
}
|
||||
/* Set burst size as 1 if there's no suitable one */
|
||||
if (j == 1)
|
||||
nbytes = width;
|
||||
}
|
||||
iter = sg_dma_len(sg) / nbytes;
|
||||
if (i < sg_len - 1) {
|
||||
last_sg = fsl_desc->tcd[(i + 1)].ptcd;
|
||||
fsl_edma_fill_tcd(fsl_chan, fsl_desc->tcd[i].vtcd, src_addr,
|
||||
fsl_edma_fill_tcd(fsl_desc->tcd[i].vtcd, src_addr,
|
||||
dst_addr, fsl_chan->attr, soff,
|
||||
nbytes, 0, iter, iter, doff, last_sg,
|
||||
false, false, true);
|
||||
} else {
|
||||
last_sg = 0;
|
||||
fsl_edma_fill_tcd(fsl_chan, fsl_desc->tcd[i].vtcd, src_addr,
|
||||
fsl_edma_fill_tcd(fsl_desc->tcd[i].vtcd, src_addr,
|
||||
dst_addr, fsl_chan->attr, soff,
|
||||
nbytes, 0, iter, iter, doff, last_sg,
|
||||
true, true, false);
|
||||
|
@ -721,6 +589,7 @@ struct dma_async_tx_descriptor *fsl_edma_prep_slave_sg(
|
|||
|
||||
return vchan_tx_prep(&fsl_chan->vchan, &fsl_desc->vdesc, flags);
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(fsl_edma_prep_slave_sg);
|
||||
|
||||
struct dma_async_tx_descriptor *fsl_edma_prep_memcpy(struct dma_chan *chan,
|
||||
dma_addr_t dma_dst, dma_addr_t dma_src,
|
||||
|
@ -737,12 +606,13 @@ struct dma_async_tx_descriptor *fsl_edma_prep_memcpy(struct dma_chan *chan,
|
|||
fsl_chan->is_sw = true;
|
||||
|
||||
/* To match with copy_align and max_seg_size so 1 tcd is enough */
|
||||
fsl_edma_fill_tcd(fsl_chan, fsl_desc->tcd[0].vtcd, dma_src, dma_dst,
|
||||
fsl_edma_get_tcd_attr(DMA_SLAVE_BUSWIDTH_32_BYTES),
|
||||
fsl_edma_fill_tcd(fsl_desc->tcd[0].vtcd, dma_src, dma_dst,
|
||||
EDMA_TCD_ATTR_SSIZE_32BYTE | EDMA_TCD_ATTR_DSIZE_32BYTE,
|
||||
32, len, 0, 1, 1, 32, 0, true, true, false);
|
||||
|
||||
return vchan_tx_prep(&fsl_chan->vchan, &fsl_desc->vdesc, flags);
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(fsl_edma_prep_memcpy);
|
||||
|
||||
void fsl_edma_xfer_desc(struct fsl_edma_chan *fsl_chan)
|
||||
{
|
||||
|
@ -759,6 +629,7 @@ void fsl_edma_xfer_desc(struct fsl_edma_chan *fsl_chan)
|
|||
fsl_chan->status = DMA_IN_PROGRESS;
|
||||
fsl_chan->idle = false;
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(fsl_edma_xfer_desc);
|
||||
|
||||
void fsl_edma_issue_pending(struct dma_chan *chan)
|
||||
{
|
||||
|
@ -778,6 +649,7 @@ void fsl_edma_issue_pending(struct dma_chan *chan)
|
|||
|
||||
spin_unlock_irqrestore(&fsl_chan->vchan.lock, flags);
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(fsl_edma_issue_pending);
|
||||
|
||||
int fsl_edma_alloc_chan_resources(struct dma_chan *chan)
|
||||
{
|
||||
|
@ -788,6 +660,7 @@ int fsl_edma_alloc_chan_resources(struct dma_chan *chan)
|
|||
32, 0);
|
||||
return 0;
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(fsl_edma_alloc_chan_resources);
|
||||
|
||||
void fsl_edma_free_chan_resources(struct dma_chan *chan)
|
||||
{
|
||||
|
@ -810,6 +683,7 @@ void fsl_edma_free_chan_resources(struct dma_chan *chan)
|
|||
fsl_chan->tcd_pool = NULL;
|
||||
fsl_chan->is_sw = false;
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(fsl_edma_free_chan_resources);
|
||||
|
||||
void fsl_edma_cleanup_vchan(struct dma_device *dmadev)
|
||||
{
|
||||
|
@ -821,10 +695,12 @@ void fsl_edma_cleanup_vchan(struct dma_device *dmadev)
|
|||
tasklet_kill(&chan->vchan.task);
|
||||
}
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(fsl_edma_cleanup_vchan);
|
||||
|
||||
/*
|
||||
* On the 32 channels Vybrid/mpc577x edma version, register offsets are
|
||||
* different compared to ColdFire mcf5441x 64 channels edma.
|
||||
* On the 32 channels Vybrid/mpc577x edma version (here called "v1"),
|
||||
* register offsets are different compared to ColdFire mcf5441x 64 channels
|
||||
* edma (here called "v2").
|
||||
*
|
||||
* This function sets up register offsets as per proper declared version
|
||||
* so must be called in xxx_edma_probe() just after setting the
|
||||
|
@ -832,30 +708,41 @@ void fsl_edma_cleanup_vchan(struct dma_device *dmadev)
|
|||
*/
|
||||
void fsl_edma_setup_regs(struct fsl_edma_engine *edma)
|
||||
{
|
||||
bool is64 = !!(edma->drvdata->flags & FSL_EDMA_DRV_EDMA64);
|
||||
|
||||
edma->regs.cr = edma->membase + EDMA_CR;
|
||||
edma->regs.es = edma->membase + EDMA_ES;
|
||||
edma->regs.erql = edma->membase + EDMA_ERQ;
|
||||
edma->regs.eeil = edma->membase + EDMA_EEI;
|
||||
|
||||
edma->regs.serq = edma->membase + (is64 ? EDMA64_SERQ : EDMA_SERQ);
|
||||
edma->regs.cerq = edma->membase + (is64 ? EDMA64_CERQ : EDMA_CERQ);
|
||||
edma->regs.seei = edma->membase + (is64 ? EDMA64_SEEI : EDMA_SEEI);
|
||||
edma->regs.ceei = edma->membase + (is64 ? EDMA64_CEEI : EDMA_CEEI);
|
||||
edma->regs.cint = edma->membase + (is64 ? EDMA64_CINT : EDMA_CINT);
|
||||
edma->regs.cerr = edma->membase + (is64 ? EDMA64_CERR : EDMA_CERR);
|
||||
edma->regs.ssrt = edma->membase + (is64 ? EDMA64_SSRT : EDMA_SSRT);
|
||||
edma->regs.cdne = edma->membase + (is64 ? EDMA64_CDNE : EDMA_CDNE);
|
||||
edma->regs.intl = edma->membase + (is64 ? EDMA64_INTL : EDMA_INTR);
|
||||
edma->regs.errl = edma->membase + (is64 ? EDMA64_ERRL : EDMA_ERR);
|
||||
edma->regs.serq = edma->membase + ((edma->drvdata->version == v2) ?
|
||||
EDMA64_SERQ : EDMA_SERQ);
|
||||
edma->regs.cerq = edma->membase + ((edma->drvdata->version == v2) ?
|
||||
EDMA64_CERQ : EDMA_CERQ);
|
||||
edma->regs.seei = edma->membase + ((edma->drvdata->version == v2) ?
|
||||
EDMA64_SEEI : EDMA_SEEI);
|
||||
edma->regs.ceei = edma->membase + ((edma->drvdata->version == v2) ?
|
||||
EDMA64_CEEI : EDMA_CEEI);
|
||||
edma->regs.cint = edma->membase + ((edma->drvdata->version == v2) ?
|
||||
EDMA64_CINT : EDMA_CINT);
|
||||
edma->regs.cerr = edma->membase + ((edma->drvdata->version == v2) ?
|
||||
EDMA64_CERR : EDMA_CERR);
|
||||
edma->regs.ssrt = edma->membase + ((edma->drvdata->version == v2) ?
|
||||
EDMA64_SSRT : EDMA_SSRT);
|
||||
edma->regs.cdne = edma->membase + ((edma->drvdata->version == v2) ?
|
||||
EDMA64_CDNE : EDMA_CDNE);
|
||||
edma->regs.intl = edma->membase + ((edma->drvdata->version == v2) ?
|
||||
EDMA64_INTL : EDMA_INTR);
|
||||
edma->regs.errl = edma->membase + ((edma->drvdata->version == v2) ?
|
||||
EDMA64_ERRL : EDMA_ERR);
|
||||
|
||||
if (is64) {
|
||||
if (edma->drvdata->version == v2) {
|
||||
edma->regs.erqh = edma->membase + EDMA64_ERQH;
|
||||
edma->regs.eeih = edma->membase + EDMA64_EEIH;
|
||||
edma->regs.errh = edma->membase + EDMA64_ERRH;
|
||||
edma->regs.inth = edma->membase + EDMA64_INTH;
|
||||
}
|
||||
|
||||
edma->regs.tcd = edma->membase + EDMA_TCD;
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(fsl_edma_setup_regs);
|
||||
|
||||
MODULE_LICENSE("GPL v2");
|
||||
|
|
|
@ -29,6 +29,16 @@
|
|||
#define EDMA_TCD_ATTR_DMOD(x) (((x) & GENMASK(4, 0)) << 3)
|
||||
#define EDMA_TCD_ATTR_SSIZE(x) (((x) & GENMASK(2, 0)) << 8)
|
||||
#define EDMA_TCD_ATTR_SMOD(x) (((x) & GENMASK(4, 0)) << 11)
|
||||
#define EDMA_TCD_ATTR_DSIZE_8BIT 0
|
||||
#define EDMA_TCD_ATTR_DSIZE_16BIT BIT(0)
|
||||
#define EDMA_TCD_ATTR_DSIZE_32BIT BIT(1)
|
||||
#define EDMA_TCD_ATTR_DSIZE_64BIT (BIT(0) | BIT(1))
|
||||
#define EDMA_TCD_ATTR_DSIZE_32BYTE (BIT(2) | BIT(0))
|
||||
#define EDMA_TCD_ATTR_SSIZE_8BIT 0
|
||||
#define EDMA_TCD_ATTR_SSIZE_16BIT (EDMA_TCD_ATTR_DSIZE_16BIT << 8)
|
||||
#define EDMA_TCD_ATTR_SSIZE_32BIT (EDMA_TCD_ATTR_DSIZE_32BIT << 8)
|
||||
#define EDMA_TCD_ATTR_SSIZE_64BIT (EDMA_TCD_ATTR_DSIZE_64BIT << 8)
|
||||
#define EDMA_TCD_ATTR_SSIZE_32BYTE (EDMA_TCD_ATTR_DSIZE_32BYTE << 8)
|
||||
|
||||
#define EDMA_TCD_CITER_CITER(x) ((x) & GENMASK(14, 0))
|
||||
#define EDMA_TCD_BITER_BITER(x) ((x) & GENMASK(14, 0))
|
||||
|
@ -42,32 +52,16 @@
|
|||
#define EDMA_TCD_CSR_ACTIVE BIT(6)
|
||||
#define EDMA_TCD_CSR_DONE BIT(7)
|
||||
|
||||
#define EDMA_V3_TCD_NBYTES_MLOFF_NBYTES(x) ((x) & GENMASK(9, 0))
|
||||
#define EDMA_V3_TCD_NBYTES_MLOFF(x) (x << 10)
|
||||
#define EDMA_V3_TCD_NBYTES_DMLOE (1 << 30)
|
||||
#define EDMA_V3_TCD_NBYTES_SMLOE (1 << 31)
|
||||
|
||||
#define EDMAMUX_CHCFG_DIS 0x0
|
||||
#define EDMAMUX_CHCFG_ENBL 0x80
|
||||
#define EDMAMUX_CHCFG_SOURCE(n) ((n) & 0x3F)
|
||||
|
||||
#define DMAMUX_NR 2
|
||||
|
||||
#define EDMA_TCD 0x1000
|
||||
|
||||
#define FSL_EDMA_BUSWIDTHS (BIT(DMA_SLAVE_BUSWIDTH_1_BYTE) | \
|
||||
BIT(DMA_SLAVE_BUSWIDTH_2_BYTES) | \
|
||||
BIT(DMA_SLAVE_BUSWIDTH_4_BYTES) | \
|
||||
BIT(DMA_SLAVE_BUSWIDTH_8_BYTES))
|
||||
|
||||
#define EDMA_V3_CH_SBR_RD BIT(22)
|
||||
#define EDMA_V3_CH_SBR_WR BIT(21)
|
||||
#define EDMA_V3_CH_CSR_ERQ BIT(0)
|
||||
#define EDMA_V3_CH_CSR_EARQ BIT(1)
|
||||
#define EDMA_V3_CH_CSR_EEI BIT(2)
|
||||
#define EDMA_V3_CH_CSR_DONE BIT(30)
|
||||
#define EDMA_V3_CH_CSR_ACTIVE BIT(31)
|
||||
|
||||
enum fsl_edma_pm_state {
|
||||
RUNNING = 0,
|
||||
SUSPENDED,
|
||||
|
@ -87,18 +81,6 @@ struct fsl_edma_hw_tcd {
|
|||
__le16 biter;
|
||||
};
|
||||
|
||||
struct fsl_edma3_ch_reg {
|
||||
__le32 ch_csr;
|
||||
__le32 ch_es;
|
||||
__le32 ch_int;
|
||||
__le32 ch_sbr;
|
||||
__le32 ch_pri;
|
||||
__le32 ch_mux;
|
||||
__le32 ch_mattr; /* edma4, reserved for edma3 */
|
||||
__le32 ch_reserved;
|
||||
struct fsl_edma_hw_tcd tcd;
|
||||
} __packed;
|
||||
|
||||
/*
|
||||
* These are iomem pointers, for both v32 and v64.
|
||||
*/
|
||||
|
@ -121,6 +103,7 @@ struct edma_regs {
|
|||
void __iomem *intl;
|
||||
void __iomem *errh;
|
||||
void __iomem *errl;
|
||||
struct fsl_edma_hw_tcd __iomem *tcd;
|
||||
};
|
||||
|
||||
struct fsl_edma_sw_tcd {
|
||||
|
@ -143,20 +126,7 @@ struct fsl_edma_chan {
|
|||
dma_addr_t dma_dev_addr;
|
||||
u32 dma_dev_size;
|
||||
enum dma_data_direction dma_dir;
|
||||
char chan_name[32];
|
||||
struct fsl_edma_hw_tcd __iomem *tcd;
|
||||
u32 real_count;
|
||||
struct work_struct issue_worker;
|
||||
struct platform_device *pdev;
|
||||
struct device *pd_dev;
|
||||
u32 srcid;
|
||||
struct clk *clk;
|
||||
int priority;
|
||||
int hw_chanid;
|
||||
int txirq;
|
||||
bool is_rxchan;
|
||||
bool is_remote;
|
||||
bool is_multi_fifo;
|
||||
char chan_name[16];
|
||||
};
|
||||
|
||||
struct fsl_edma_desc {
|
||||
|
@ -168,32 +138,17 @@ struct fsl_edma_desc {
|
|||
struct fsl_edma_sw_tcd tcd[];
|
||||
};
|
||||
|
||||
#define FSL_EDMA_DRV_HAS_DMACLK BIT(0)
|
||||
#define FSL_EDMA_DRV_MUX_SWAP BIT(1)
|
||||
#define FSL_EDMA_DRV_CONFIG32 BIT(2)
|
||||
#define FSL_EDMA_DRV_WRAP_IO BIT(3)
|
||||
#define FSL_EDMA_DRV_EDMA64 BIT(4)
|
||||
#define FSL_EDMA_DRV_HAS_PD BIT(5)
|
||||
#define FSL_EDMA_DRV_HAS_CHCLK BIT(6)
|
||||
#define FSL_EDMA_DRV_HAS_CHMUX BIT(7)
|
||||
/* imx8 QM audio edma remote local swapped */
|
||||
#define FSL_EDMA_DRV_QUIRK_SWAPPED BIT(8)
|
||||
/* control and status register is in tcd address space, edma3 reg layout */
|
||||
#define FSL_EDMA_DRV_SPLIT_REG BIT(9)
|
||||
#define FSL_EDMA_DRV_BUS_8BYTE BIT(10)
|
||||
#define FSL_EDMA_DRV_DEV_TO_DEV BIT(11)
|
||||
#define FSL_EDMA_DRV_ALIGN_64BYTE BIT(12)
|
||||
|
||||
#define FSL_EDMA_DRV_EDMA3 (FSL_EDMA_DRV_SPLIT_REG | \
|
||||
FSL_EDMA_DRV_BUS_8BYTE | \
|
||||
FSL_EDMA_DRV_DEV_TO_DEV | \
|
||||
FSL_EDMA_DRV_ALIGN_64BYTE)
|
||||
enum edma_version {
|
||||
v1, /* 32ch, Vybrid, mpc57x, etc */
|
||||
v2, /* 64ch Coldfire */
|
||||
v3, /* 32ch, i.mx7ulp */
|
||||
};
|
||||
|
||||
struct fsl_edma_drvdata {
|
||||
u32 dmamuxs; /* only used before v3 */
|
||||
u32 chreg_off;
|
||||
u32 chreg_space_sz;
|
||||
u32 flags;
|
||||
enum edma_version version;
|
||||
u32 dmamuxs;
|
||||
bool has_dmaclk;
|
||||
bool mux_swap;
|
||||
int (*setup_irq)(struct platform_device *pdev,
|
||||
struct fsl_edma_engine *fsl_edma);
|
||||
};
|
||||
|
@ -204,7 +159,6 @@ struct fsl_edma_engine {
|
|||
void __iomem *muxbase[DMAMUX_NR];
|
||||
struct clk *muxclk[DMAMUX_NR];
|
||||
struct clk *dmaclk;
|
||||
struct clk *chclk;
|
||||
struct mutex fsl_edma_mutex;
|
||||
const struct fsl_edma_drvdata *drvdata;
|
||||
u32 n_chans;
|
||||
|
@ -212,28 +166,9 @@ struct fsl_edma_engine {
|
|||
int errirq;
|
||||
bool big_endian;
|
||||
struct edma_regs regs;
|
||||
u64 chan_masked;
|
||||
struct fsl_edma_chan chans[];
|
||||
};
|
||||
|
||||
#define edma_read_tcdreg(chan, __name) \
|
||||
(sizeof(chan->tcd->__name) == sizeof(u32) ? \
|
||||
edma_readl(chan->edma, &chan->tcd->__name) : \
|
||||
edma_readw(chan->edma, &chan->tcd->__name))
|
||||
|
||||
#define edma_write_tcdreg(chan, val, __name) \
|
||||
(sizeof(chan->tcd->__name) == sizeof(u32) ? \
|
||||
edma_writel(chan->edma, (u32 __force)val, &chan->tcd->__name) : \
|
||||
edma_writew(chan->edma, (u16 __force)val, &chan->tcd->__name))
|
||||
|
||||
#define edma_readl_chreg(chan, __name) \
|
||||
edma_readl(chan->edma, \
|
||||
(void __iomem *)&(container_of(chan->tcd, struct fsl_edma3_ch_reg, tcd)->__name))
|
||||
|
||||
#define edma_writel_chreg(chan, val, __name) \
|
||||
edma_writel(chan->edma, val, \
|
||||
(void __iomem *)&(container_of(chan->tcd, struct fsl_edma3_ch_reg, tcd)->__name))
|
||||
|
||||
/*
|
||||
* R/W functions for big- or little-endian registers:
|
||||
* The eDMA controller's endian is independent of the CPU core's endian.
|
||||
|
@ -248,14 +183,6 @@ static inline u32 edma_readl(struct fsl_edma_engine *edma, void __iomem *addr)
|
|||
return ioread32(addr);
|
||||
}
|
||||
|
||||
static inline u16 edma_readw(struct fsl_edma_engine *edma, void __iomem *addr)
|
||||
{
|
||||
if (edma->big_endian)
|
||||
return ioread16be(addr);
|
||||
else
|
||||
return ioread16(addr);
|
||||
}
|
||||
|
||||
static inline void edma_writeb(struct fsl_edma_engine *edma,
|
||||
u8 val, void __iomem *addr)
|
||||
{
|
||||
|
@ -290,23 +217,11 @@ static inline struct fsl_edma_chan *to_fsl_edma_chan(struct dma_chan *chan)
|
|||
return container_of(chan, struct fsl_edma_chan, vchan.chan);
|
||||
}
|
||||
|
||||
static inline u32 fsl_edma_drvflags(struct fsl_edma_chan *fsl_chan)
|
||||
{
|
||||
return fsl_chan->edma->drvdata->flags;
|
||||
}
|
||||
|
||||
static inline struct fsl_edma_desc *to_fsl_edma_desc(struct virt_dma_desc *vd)
|
||||
{
|
||||
return container_of(vd, struct fsl_edma_desc, vdesc);
|
||||
}
|
||||
|
||||
static inline void fsl_edma_err_chan_handler(struct fsl_edma_chan *fsl_chan)
|
||||
{
|
||||
fsl_chan->status = DMA_ERROR;
|
||||
fsl_chan->idle = true;
|
||||
}
|
||||
|
||||
void fsl_edma_tx_chan_handler(struct fsl_edma_chan *fsl_chan);
|
||||
void fsl_edma_disable_request(struct fsl_edma_chan *fsl_chan);
|
||||
void fsl_edma_chan_mux(struct fsl_edma_chan *fsl_chan,
|
||||
unsigned int slot, bool enable);
|
||||
|
|
|
@ -18,15 +18,9 @@
|
|||
#include <linux/of_irq.h>
|
||||
#include <linux/of_dma.h>
|
||||
#include <linux/dma-mapping.h>
|
||||
#include <linux/pm_runtime.h>
|
||||
#include <linux/pm_domain.h>
|
||||
|
||||
#include "fsl-edma-common.h"
|
||||
|
||||
#define ARGS_RX BIT(0)
|
||||
#define ARGS_REMOTE BIT(1)
|
||||
#define ARGS_MULTI_FIFO BIT(2)
|
||||
|
||||
static void fsl_edma_synchronize(struct dma_chan *chan)
|
||||
{
|
||||
struct fsl_edma_chan *fsl_chan = to_fsl_edma_chan(chan);
|
||||
|
@ -39,6 +33,7 @@ static irqreturn_t fsl_edma_tx_handler(int irq, void *dev_id)
|
|||
struct fsl_edma_engine *fsl_edma = dev_id;
|
||||
unsigned int intr, ch;
|
||||
struct edma_regs *regs = &fsl_edma->regs;
|
||||
struct fsl_edma_chan *fsl_chan;
|
||||
|
||||
intr = edma_readl(fsl_edma, regs->intl);
|
||||
if (!intr)
|
||||
|
@ -47,28 +42,36 @@ static irqreturn_t fsl_edma_tx_handler(int irq, void *dev_id)
|
|||
for (ch = 0; ch < fsl_edma->n_chans; ch++) {
|
||||
if (intr & (0x1 << ch)) {
|
||||
edma_writeb(fsl_edma, EDMA_CINT_CINT(ch), regs->cint);
|
||||
fsl_edma_tx_chan_handler(&fsl_edma->chans[ch]);
|
||||
|
||||
fsl_chan = &fsl_edma->chans[ch];
|
||||
|
||||
spin_lock(&fsl_chan->vchan.lock);
|
||||
|
||||
if (!fsl_chan->edesc) {
|
||||
/* terminate_all called before */
|
||||
spin_unlock(&fsl_chan->vchan.lock);
|
||||
continue;
|
||||
}
|
||||
|
||||
if (!fsl_chan->edesc->iscyclic) {
|
||||
list_del(&fsl_chan->edesc->vdesc.node);
|
||||
vchan_cookie_complete(&fsl_chan->edesc->vdesc);
|
||||
fsl_chan->edesc = NULL;
|
||||
fsl_chan->status = DMA_COMPLETE;
|
||||
fsl_chan->idle = true;
|
||||
} else {
|
||||
vchan_cyclic_callback(&fsl_chan->edesc->vdesc);
|
||||
}
|
||||
|
||||
if (!fsl_chan->edesc)
|
||||
fsl_edma_xfer_desc(fsl_chan);
|
||||
|
||||
spin_unlock(&fsl_chan->vchan.lock);
|
||||
}
|
||||
}
|
||||
return IRQ_HANDLED;
|
||||
}
|
||||
|
||||
static irqreturn_t fsl_edma3_tx_handler(int irq, void *dev_id)
|
||||
{
|
||||
struct fsl_edma_chan *fsl_chan = dev_id;
|
||||
unsigned int intr;
|
||||
|
||||
intr = edma_readl_chreg(fsl_chan, ch_int);
|
||||
if (!intr)
|
||||
return IRQ_HANDLED;
|
||||
|
||||
edma_writel_chreg(fsl_chan, 1, ch_int);
|
||||
|
||||
fsl_edma_tx_chan_handler(fsl_chan);
|
||||
|
||||
return IRQ_HANDLED;
|
||||
}
|
||||
|
||||
static irqreturn_t fsl_edma_err_handler(int irq, void *dev_id)
|
||||
{
|
||||
struct fsl_edma_engine *fsl_edma = dev_id;
|
||||
|
@ -83,7 +86,8 @@ static irqreturn_t fsl_edma_err_handler(int irq, void *dev_id)
|
|||
if (err & (0x1 << ch)) {
|
||||
fsl_edma_disable_request(&fsl_edma->chans[ch]);
|
||||
edma_writeb(fsl_edma, EDMA_CERR_CERR(ch), regs->cerr);
|
||||
fsl_edma_err_chan_handler(&fsl_edma->chans[ch]);
|
||||
fsl_edma->chans[ch].status = DMA_ERROR;
|
||||
fsl_edma->chans[ch].idle = true;
|
||||
}
|
||||
}
|
||||
return IRQ_HANDLED;
|
||||
|
@ -130,58 +134,11 @@ static struct dma_chan *fsl_edma_xlate(struct of_phandle_args *dma_spec,
|
|||
return NULL;
|
||||
}
|
||||
|
||||
static struct dma_chan *fsl_edma3_xlate(struct of_phandle_args *dma_spec,
|
||||
struct of_dma *ofdma)
|
||||
{
|
||||
struct fsl_edma_engine *fsl_edma = ofdma->of_dma_data;
|
||||
struct dma_chan *chan, *_chan;
|
||||
struct fsl_edma_chan *fsl_chan;
|
||||
bool b_chmux;
|
||||
int i;
|
||||
|
||||
if (dma_spec->args_count != 3)
|
||||
return NULL;
|
||||
|
||||
b_chmux = !!(fsl_edma->drvdata->flags & FSL_EDMA_DRV_HAS_CHMUX);
|
||||
|
||||
mutex_lock(&fsl_edma->fsl_edma_mutex);
|
||||
list_for_each_entry_safe(chan, _chan, &fsl_edma->dma_dev.channels,
|
||||
device_node) {
|
||||
|
||||
if (chan->client_count)
|
||||
continue;
|
||||
|
||||
fsl_chan = to_fsl_edma_chan(chan);
|
||||
i = fsl_chan - fsl_edma->chans;
|
||||
|
||||
chan = dma_get_slave_channel(chan);
|
||||
chan->device->privatecnt++;
|
||||
fsl_chan->priority = dma_spec->args[1];
|
||||
fsl_chan->is_rxchan = dma_spec->args[2] & ARGS_RX;
|
||||
fsl_chan->is_remote = dma_spec->args[2] & ARGS_REMOTE;
|
||||
fsl_chan->is_multi_fifo = dma_spec->args[2] & ARGS_MULTI_FIFO;
|
||||
|
||||
if (!b_chmux && i == dma_spec->args[0]) {
|
||||
mutex_unlock(&fsl_edma->fsl_edma_mutex);
|
||||
return chan;
|
||||
} else if (b_chmux && !fsl_chan->srcid) {
|
||||
/* if controller support channel mux, choose a free channel */
|
||||
fsl_chan->srcid = dma_spec->args[0];
|
||||
mutex_unlock(&fsl_edma->fsl_edma_mutex);
|
||||
return chan;
|
||||
}
|
||||
}
|
||||
mutex_unlock(&fsl_edma->fsl_edma_mutex);
|
||||
return NULL;
|
||||
}
|
||||
|
||||
static int
|
||||
fsl_edma_irq_init(struct platform_device *pdev, struct fsl_edma_engine *fsl_edma)
|
||||
{
|
||||
int ret;
|
||||
|
||||
edma_writel(fsl_edma, ~0, fsl_edma->regs.intl);
|
||||
|
||||
fsl_edma->txirq = platform_get_irq_byname(pdev, "edma-tx");
|
||||
if (fsl_edma->txirq < 0)
|
||||
return fsl_edma->txirq;
|
||||
|
@ -216,37 +173,6 @@ fsl_edma_irq_init(struct platform_device *pdev, struct fsl_edma_engine *fsl_edma
|
|||
return 0;
|
||||
}
|
||||
|
||||
static int fsl_edma3_irq_init(struct platform_device *pdev, struct fsl_edma_engine *fsl_edma)
|
||||
{
|
||||
int ret;
|
||||
int i;
|
||||
|
||||
for (i = 0; i < fsl_edma->n_chans; i++) {
|
||||
|
||||
struct fsl_edma_chan *fsl_chan = &fsl_edma->chans[i];
|
||||
|
||||
if (fsl_edma->chan_masked & BIT(i))
|
||||
continue;
|
||||
|
||||
/* request channel irq */
|
||||
fsl_chan->txirq = platform_get_irq(pdev, i);
|
||||
if (fsl_chan->txirq < 0) {
|
||||
dev_err(&pdev->dev, "Can't get chan %d's irq.\n", i);
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
ret = devm_request_irq(&pdev->dev, fsl_chan->txirq,
|
||||
fsl_edma3_tx_handler, IRQF_SHARED,
|
||||
fsl_chan->chan_name, fsl_chan);
|
||||
if (ret) {
|
||||
dev_err(&pdev->dev, "Can't register chan%d's IRQ.\n", i);
|
||||
return -EINVAL;
|
||||
}
|
||||
}
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int
|
||||
fsl_edma2_irq_init(struct platform_device *pdev,
|
||||
struct fsl_edma_engine *fsl_edma)
|
||||
|
@ -254,8 +180,6 @@ fsl_edma2_irq_init(struct platform_device *pdev,
|
|||
int i, ret, irq;
|
||||
int count;
|
||||
|
||||
edma_writel(fsl_edma, ~0, fsl_edma->regs.intl);
|
||||
|
||||
count = platform_irq_count(pdev);
|
||||
dev_dbg(&pdev->dev, "%s Found %d interrupts\r\n", __func__, count);
|
||||
if (count <= 2) {
|
||||
|
@ -273,6 +197,8 @@ fsl_edma2_irq_init(struct platform_device *pdev,
|
|||
if (irq < 0)
|
||||
return -ENXIO;
|
||||
|
||||
sprintf(fsl_edma->chans[i].chan_name, "eDMA2-CH%02d", i);
|
||||
|
||||
/* The last IRQ is for eDMA err */
|
||||
if (i == count - 1)
|
||||
ret = devm_request_irq(&pdev->dev, irq,
|
||||
|
@ -310,110 +236,33 @@ static void fsl_disable_clocks(struct fsl_edma_engine *fsl_edma, int nr_clocks)
|
|||
}
|
||||
|
||||
static struct fsl_edma_drvdata vf610_data = {
|
||||
.version = v1,
|
||||
.dmamuxs = DMAMUX_NR,
|
||||
.flags = FSL_EDMA_DRV_WRAP_IO,
|
||||
.chreg_off = EDMA_TCD,
|
||||
.chreg_space_sz = sizeof(struct fsl_edma_hw_tcd),
|
||||
.setup_irq = fsl_edma_irq_init,
|
||||
};
|
||||
|
||||
static struct fsl_edma_drvdata ls1028a_data = {
|
||||
.version = v1,
|
||||
.dmamuxs = DMAMUX_NR,
|
||||
.flags = FSL_EDMA_DRV_MUX_SWAP | FSL_EDMA_DRV_WRAP_IO,
|
||||
.chreg_off = EDMA_TCD,
|
||||
.chreg_space_sz = sizeof(struct fsl_edma_hw_tcd),
|
||||
.mux_swap = true,
|
||||
.setup_irq = fsl_edma_irq_init,
|
||||
};
|
||||
|
||||
static struct fsl_edma_drvdata imx7ulp_data = {
|
||||
.version = v3,
|
||||
.dmamuxs = 1,
|
||||
.chreg_off = EDMA_TCD,
|
||||
.chreg_space_sz = sizeof(struct fsl_edma_hw_tcd),
|
||||
.flags = FSL_EDMA_DRV_HAS_DMACLK | FSL_EDMA_DRV_CONFIG32,
|
||||
.has_dmaclk = true,
|
||||
.setup_irq = fsl_edma2_irq_init,
|
||||
};
|
||||
|
||||
static struct fsl_edma_drvdata imx8qm_data = {
|
||||
.flags = FSL_EDMA_DRV_HAS_PD | FSL_EDMA_DRV_EDMA3,
|
||||
.chreg_space_sz = 0x10000,
|
||||
.chreg_off = 0x10000,
|
||||
.setup_irq = fsl_edma3_irq_init,
|
||||
};
|
||||
|
||||
static struct fsl_edma_drvdata imx8qm_audio_data = {
|
||||
.flags = FSL_EDMA_DRV_QUIRK_SWAPPED | FSL_EDMA_DRV_HAS_PD | FSL_EDMA_DRV_EDMA3,
|
||||
.chreg_space_sz = 0x10000,
|
||||
.chreg_off = 0x10000,
|
||||
.setup_irq = fsl_edma3_irq_init,
|
||||
};
|
||||
|
||||
static struct fsl_edma_drvdata imx93_data3 = {
|
||||
.flags = FSL_EDMA_DRV_HAS_DMACLK | FSL_EDMA_DRV_EDMA3,
|
||||
.chreg_space_sz = 0x10000,
|
||||
.chreg_off = 0x10000,
|
||||
.setup_irq = fsl_edma3_irq_init,
|
||||
};
|
||||
|
||||
static struct fsl_edma_drvdata imx93_data4 = {
|
||||
.flags = FSL_EDMA_DRV_HAS_CHMUX | FSL_EDMA_DRV_HAS_DMACLK | FSL_EDMA_DRV_EDMA3,
|
||||
.chreg_space_sz = 0x8000,
|
||||
.chreg_off = 0x10000,
|
||||
.setup_irq = fsl_edma3_irq_init,
|
||||
};
|
||||
|
||||
static const struct of_device_id fsl_edma_dt_ids[] = {
|
||||
{ .compatible = "fsl,vf610-edma", .data = &vf610_data},
|
||||
{ .compatible = "fsl,ls1028a-edma", .data = &ls1028a_data},
|
||||
{ .compatible = "fsl,imx7ulp-edma", .data = &imx7ulp_data},
|
||||
{ .compatible = "fsl,imx8qm-edma", .data = &imx8qm_data},
|
||||
{ .compatible = "fsl,imx8qm-adma", .data = &imx8qm_audio_data},
|
||||
{ .compatible = "fsl,imx93-edma3", .data = &imx93_data3},
|
||||
{ .compatible = "fsl,imx93-edma4", .data = &imx93_data4},
|
||||
{ /* sentinel */ }
|
||||
};
|
||||
MODULE_DEVICE_TABLE(of, fsl_edma_dt_ids);
|
||||
|
||||
static int fsl_edma3_attach_pd(struct platform_device *pdev, struct fsl_edma_engine *fsl_edma)
|
||||
{
|
||||
struct fsl_edma_chan *fsl_chan;
|
||||
struct device_link *link;
|
||||
struct device *pd_chan;
|
||||
struct device *dev;
|
||||
int i;
|
||||
|
||||
dev = &pdev->dev;
|
||||
|
||||
for (i = 0; i < fsl_edma->n_chans; i++) {
|
||||
if (fsl_edma->chan_masked & BIT(i))
|
||||
continue;
|
||||
|
||||
fsl_chan = &fsl_edma->chans[i];
|
||||
|
||||
pd_chan = dev_pm_domain_attach_by_id(dev, i);
|
||||
if (IS_ERR_OR_NULL(pd_chan)) {
|
||||
dev_err(dev, "Failed attach pd %d\n", i);
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
link = device_link_add(dev, pd_chan, DL_FLAG_STATELESS |
|
||||
DL_FLAG_PM_RUNTIME |
|
||||
DL_FLAG_RPM_ACTIVE);
|
||||
if (IS_ERR(link)) {
|
||||
dev_err(dev, "Failed to add device_link to %d: %ld\n", i,
|
||||
PTR_ERR(link));
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
fsl_chan->pd_dev = pd_chan;
|
||||
|
||||
pm_runtime_use_autosuspend(fsl_chan->pd_dev);
|
||||
pm_runtime_set_autosuspend_delay(fsl_chan->pd_dev, 200);
|
||||
pm_runtime_set_active(fsl_chan->pd_dev);
|
||||
}
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int fsl_edma_probe(struct platform_device *pdev)
|
||||
{
|
||||
const struct of_device_id *of_id =
|
||||
|
@ -421,7 +270,6 @@ static int fsl_edma_probe(struct platform_device *pdev)
|
|||
struct device_node *np = pdev->dev.of_node;
|
||||
struct fsl_edma_engine *fsl_edma;
|
||||
const struct fsl_edma_drvdata *drvdata = NULL;
|
||||
u32 chan_mask[2] = {0, 0};
|
||||
struct edma_regs *regs;
|
||||
int chans;
|
||||
int ret, i;
|
||||
|
@ -452,42 +300,26 @@ static int fsl_edma_probe(struct platform_device *pdev)
|
|||
if (IS_ERR(fsl_edma->membase))
|
||||
return PTR_ERR(fsl_edma->membase);
|
||||
|
||||
if (!(drvdata->flags & FSL_EDMA_DRV_SPLIT_REG)) {
|
||||
fsl_edma_setup_regs(fsl_edma);
|
||||
regs = &fsl_edma->regs;
|
||||
}
|
||||
fsl_edma_setup_regs(fsl_edma);
|
||||
regs = &fsl_edma->regs;
|
||||
|
||||
if (drvdata->flags & FSL_EDMA_DRV_HAS_DMACLK) {
|
||||
fsl_edma->dmaclk = devm_clk_get_enabled(&pdev->dev, "dma");
|
||||
if (drvdata->has_dmaclk) {
|
||||
fsl_edma->dmaclk = devm_clk_get(&pdev->dev, "dma");
|
||||
if (IS_ERR(fsl_edma->dmaclk)) {
|
||||
dev_err(&pdev->dev, "Missing DMA block clock.\n");
|
||||
return PTR_ERR(fsl_edma->dmaclk);
|
||||
}
|
||||
}
|
||||
|
||||
if (drvdata->flags & FSL_EDMA_DRV_HAS_CHCLK) {
|
||||
fsl_edma->chclk = devm_clk_get_enabled(&pdev->dev, "mp");
|
||||
if (IS_ERR(fsl_edma->chclk)) {
|
||||
dev_err(&pdev->dev, "Missing MP block clock.\n");
|
||||
return PTR_ERR(fsl_edma->chclk);
|
||||
ret = clk_prepare_enable(fsl_edma->dmaclk);
|
||||
if (ret) {
|
||||
dev_err(&pdev->dev, "DMA clk block failed.\n");
|
||||
return ret;
|
||||
}
|
||||
}
|
||||
|
||||
ret = of_property_read_variable_u32_array(np, "dma-channel-mask", chan_mask, 1, 2);
|
||||
|
||||
if (ret > 0) {
|
||||
fsl_edma->chan_masked = chan_mask[1];
|
||||
fsl_edma->chan_masked <<= 32;
|
||||
fsl_edma->chan_masked |= chan_mask[0];
|
||||
}
|
||||
|
||||
for (i = 0; i < fsl_edma->drvdata->dmamuxs; i++) {
|
||||
char clkname[32];
|
||||
|
||||
/* eDMAv3 mux register move to TCD area if ch_mux exist */
|
||||
if (drvdata->flags & FSL_EDMA_DRV_SPLIT_REG)
|
||||
break;
|
||||
|
||||
fsl_edma->muxbase[i] = devm_platform_ioremap_resource(pdev,
|
||||
1 + i);
|
||||
if (IS_ERR(fsl_edma->muxbase[i])) {
|
||||
|
@ -497,32 +329,26 @@ static int fsl_edma_probe(struct platform_device *pdev)
|
|||
}
|
||||
|
||||
sprintf(clkname, "dmamux%d", i);
|
||||
fsl_edma->muxclk[i] = devm_clk_get_enabled(&pdev->dev, clkname);
|
||||
fsl_edma->muxclk[i] = devm_clk_get(&pdev->dev, clkname);
|
||||
if (IS_ERR(fsl_edma->muxclk[i])) {
|
||||
dev_err(&pdev->dev, "Missing DMAMUX block clock.\n");
|
||||
/* on error: disable all previously enabled clks */
|
||||
fsl_disable_clocks(fsl_edma, i);
|
||||
return PTR_ERR(fsl_edma->muxclk[i]);
|
||||
}
|
||||
|
||||
ret = clk_prepare_enable(fsl_edma->muxclk[i]);
|
||||
if (ret)
|
||||
/* on error: disable all previously enabled clks */
|
||||
fsl_disable_clocks(fsl_edma, i);
|
||||
|
||||
}
|
||||
|
||||
fsl_edma->big_endian = of_property_read_bool(np, "big-endian");
|
||||
|
||||
if (drvdata->flags & FSL_EDMA_DRV_HAS_PD) {
|
||||
ret = fsl_edma3_attach_pd(pdev, fsl_edma);
|
||||
if (ret)
|
||||
return ret;
|
||||
}
|
||||
|
||||
INIT_LIST_HEAD(&fsl_edma->dma_dev.channels);
|
||||
for (i = 0; i < fsl_edma->n_chans; i++) {
|
||||
struct fsl_edma_chan *fsl_chan = &fsl_edma->chans[i];
|
||||
int len;
|
||||
|
||||
if (fsl_edma->chan_masked & BIT(i))
|
||||
continue;
|
||||
|
||||
snprintf(fsl_chan->chan_name, sizeof(fsl_chan->chan_name), "%s-CH%02d",
|
||||
dev_name(&pdev->dev), i);
|
||||
|
||||
fsl_chan->edma = fsl_edma;
|
||||
fsl_chan->pm_state = RUNNING;
|
||||
|
@ -530,19 +356,13 @@ static int fsl_edma_probe(struct platform_device *pdev)
|
|||
fsl_chan->idle = true;
|
||||
fsl_chan->dma_dir = DMA_NONE;
|
||||
fsl_chan->vchan.desc_free = fsl_edma_free_desc;
|
||||
|
||||
len = (drvdata->flags & FSL_EDMA_DRV_SPLIT_REG) ?
|
||||
offsetof(struct fsl_edma3_ch_reg, tcd) : 0;
|
||||
fsl_chan->tcd = fsl_edma->membase
|
||||
+ i * drvdata->chreg_space_sz + drvdata->chreg_off + len;
|
||||
|
||||
fsl_chan->pdev = pdev;
|
||||
vchan_init(&fsl_chan->vchan, &fsl_edma->dma_dev);
|
||||
|
||||
edma_write_tcdreg(fsl_chan, 0, csr);
|
||||
edma_writew(fsl_edma, 0x0, ®s->tcd[i].csr);
|
||||
fsl_edma_chan_mux(fsl_chan, 0, false);
|
||||
}
|
||||
|
||||
edma_writel(fsl_edma, ~0, regs->intl);
|
||||
ret = fsl_edma->drvdata->setup_irq(pdev, fsl_edma);
|
||||
if (ret)
|
||||
return ret;
|
||||
|
@ -570,47 +390,33 @@ static int fsl_edma_probe(struct platform_device *pdev)
|
|||
|
||||
fsl_edma->dma_dev.src_addr_widths = FSL_EDMA_BUSWIDTHS;
|
||||
fsl_edma->dma_dev.dst_addr_widths = FSL_EDMA_BUSWIDTHS;
|
||||
|
||||
if (drvdata->flags & FSL_EDMA_DRV_BUS_8BYTE) {
|
||||
fsl_edma->dma_dev.src_addr_widths |= BIT(DMA_SLAVE_BUSWIDTH_8_BYTES);
|
||||
fsl_edma->dma_dev.dst_addr_widths |= BIT(DMA_SLAVE_BUSWIDTH_8_BYTES);
|
||||
}
|
||||
|
||||
fsl_edma->dma_dev.directions = BIT(DMA_DEV_TO_MEM) | BIT(DMA_MEM_TO_DEV);
|
||||
if (drvdata->flags & FSL_EDMA_DRV_DEV_TO_DEV)
|
||||
fsl_edma->dma_dev.directions |= BIT(DMA_DEV_TO_DEV);
|
||||
|
||||
fsl_edma->dma_dev.copy_align = drvdata->flags & FSL_EDMA_DRV_ALIGN_64BYTE ?
|
||||
DMAENGINE_ALIGN_64_BYTES :
|
||||
DMAENGINE_ALIGN_32_BYTES;
|
||||
|
||||
fsl_edma->dma_dev.copy_align = DMAENGINE_ALIGN_32_BYTES;
|
||||
/* Per worst case 'nbytes = 1' take CITER as the max_seg_size */
|
||||
dma_set_max_seg_size(fsl_edma->dma_dev.dev, 0x3fff);
|
||||
|
||||
fsl_edma->dma_dev.residue_granularity = DMA_RESIDUE_GRANULARITY_SEGMENT;
|
||||
|
||||
platform_set_drvdata(pdev, fsl_edma);
|
||||
|
||||
ret = dma_async_device_register(&fsl_edma->dma_dev);
|
||||
if (ret) {
|
||||
dev_err(&pdev->dev,
|
||||
"Can't register Freescale eDMA engine. (%d)\n", ret);
|
||||
fsl_disable_clocks(fsl_edma, fsl_edma->drvdata->dmamuxs);
|
||||
return ret;
|
||||
}
|
||||
|
||||
ret = of_dma_controller_register(np,
|
||||
drvdata->flags & FSL_EDMA_DRV_SPLIT_REG ? fsl_edma3_xlate : fsl_edma_xlate,
|
||||
fsl_edma);
|
||||
ret = of_dma_controller_register(np, fsl_edma_xlate, fsl_edma);
|
||||
if (ret) {
|
||||
dev_err(&pdev->dev,
|
||||
"Can't register Freescale eDMA of_dma. (%d)\n", ret);
|
||||
dma_async_device_unregister(&fsl_edma->dma_dev);
|
||||
fsl_disable_clocks(fsl_edma, fsl_edma->drvdata->dmamuxs);
|
||||
return ret;
|
||||
}
|
||||
|
||||
/* enable round robin arbitration */
|
||||
if (!(drvdata->flags & FSL_EDMA_DRV_SPLIT_REG))
|
||||
edma_writel(fsl_edma, EDMA_CR_ERGA | EDMA_CR_ERCA, regs->cr);
|
||||
edma_writel(fsl_edma, EDMA_CR_ERGA | EDMA_CR_ERCA, regs->cr);
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
@ -663,7 +469,7 @@ static int fsl_edma_resume_early(struct device *dev)
|
|||
for (i = 0; i < fsl_edma->n_chans; i++) {
|
||||
fsl_chan = &fsl_edma->chans[i];
|
||||
fsl_chan->pm_state = RUNNING;
|
||||
edma_write_tcdreg(fsl_chan, 0, csr);
|
||||
edma_writew(fsl_edma, 0x0, ®s->tcd[i].csr);
|
||||
if (fsl_chan->slave_id != 0)
|
||||
fsl_edma_chan_mux(fsl_chan, fsl_chan->slave_id, true);
|
||||
}
|
|
@ -19,6 +19,7 @@ static irqreturn_t mcf_edma_tx_handler(int irq, void *dev_id)
|
|||
struct fsl_edma_engine *mcf_edma = dev_id;
|
||||
struct edma_regs *regs = &mcf_edma->regs;
|
||||
unsigned int ch;
|
||||
struct fsl_edma_chan *mcf_chan;
|
||||
u64 intmap;
|
||||
|
||||
intmap = ioread32(regs->inth);
|
||||
|
@ -30,7 +31,31 @@ static irqreturn_t mcf_edma_tx_handler(int irq, void *dev_id)
|
|||
for (ch = 0; ch < mcf_edma->n_chans; ch++) {
|
||||
if (intmap & BIT(ch)) {
|
||||
iowrite8(EDMA_MASK_CH(ch), regs->cint);
|
||||
fsl_edma_tx_chan_handler(&mcf_edma->chans[ch]);
|
||||
|
||||
mcf_chan = &mcf_edma->chans[ch];
|
||||
|
||||
spin_lock(&mcf_chan->vchan.lock);
|
||||
|
||||
if (!mcf_chan->edesc) {
|
||||
/* terminate_all called before */
|
||||
spin_unlock(&mcf_chan->vchan.lock);
|
||||
continue;
|
||||
}
|
||||
|
||||
if (!mcf_chan->edesc->iscyclic) {
|
||||
list_del(&mcf_chan->edesc->vdesc.node);
|
||||
vchan_cookie_complete(&mcf_chan->edesc->vdesc);
|
||||
mcf_chan->edesc = NULL;
|
||||
mcf_chan->status = DMA_COMPLETE;
|
||||
mcf_chan->idle = true;
|
||||
} else {
|
||||
vchan_cyclic_callback(&mcf_chan->edesc->vdesc);
|
||||
}
|
||||
|
||||
if (!mcf_chan->edesc)
|
||||
fsl_edma_xfer_desc(mcf_chan);
|
||||
|
||||
spin_unlock(&mcf_chan->vchan.lock);
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -51,7 +76,8 @@ static irqreturn_t mcf_edma_err_handler(int irq, void *dev_id)
|
|||
if (err & BIT(ch)) {
|
||||
fsl_edma_disable_request(&mcf_edma->chans[ch]);
|
||||
iowrite8(EDMA_CERR_CERR(ch), regs->cerr);
|
||||
fsl_edma_err_chan_handler(&mcf_edma->chans[ch]);
|
||||
mcf_edma->chans[ch].status = DMA_ERROR;
|
||||
mcf_edma->chans[ch].idle = true;
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -146,7 +172,7 @@ static void mcf_edma_irq_free(struct platform_device *pdev,
|
|||
}
|
||||
|
||||
static struct fsl_edma_drvdata mcf_data = {
|
||||
.flags = FSL_EDMA_DRV_EDMA64,
|
||||
.version = v2,
|
||||
.setup_irq = mcf_edma_irq_init,
|
||||
};
|
||||
|
||||
|
@ -200,9 +226,7 @@ static int mcf_edma_probe(struct platform_device *pdev)
|
|||
mcf_chan->dma_dir = DMA_NONE;
|
||||
mcf_chan->vchan.desc_free = fsl_edma_free_desc;
|
||||
vchan_init(&mcf_chan->vchan, &mcf_edma->dma_dev);
|
||||
mcf_chan->tcd = mcf_edma->membase + EDMA_TCD
|
||||
+ i * sizeof(struct fsl_edma_hw_tcd);
|
||||
iowrite32(0x0, &mcf_chan->tcd->csr);
|
||||
iowrite32(0x0, ®s->tcd[i].csr);
|
||||
}
|
||||
|
||||
iowrite32(~0, regs->inth);
|
Loading…
Reference in New Issue
Block a user