Re: [PATCH v9 12/13] dmaengine: fsl-edma: integrate v3 support

From: Vinod Koul
Date: Fri Jul 07 2023 - 01:44:38 EST


On 20-06-23, 16:12, Frank Li wrote:
> Significant alterations have been made to the EDMA v3's register layout.
> Now, each channel possesses a separate address space, encapsulating all
> channel-related controls and statuses, including IRQs. There are changes
> in bit position definitions as well. However, the fundamental control flow
> remains analogous to the previous versions.
>
> EDMA v3 was utilized in imx8qm, imx93, and will be in forthcoming chips.
>
> Signed-off-by: Frank Li <Frank.Li@xxxxxxx>
> ---
> drivers/dma/fsl-edma-common.c | 158 +++++++++++++++++++---
> drivers/dma/fsl-edma-common.h | 73 +++++++++-
> drivers/dma/fsl-edma-main.c | 247 ++++++++++++++++++++++++++++++++--
> 3 files changed, 454 insertions(+), 24 deletions(-)
>
> diff --git a/drivers/dma/fsl-edma-common.c b/drivers/dma/fsl-edma-common.c
> index eead6a4765f7..bec1f4d6a3e1 100644
> --- a/drivers/dma/fsl-edma-common.c
> +++ b/drivers/dma/fsl-edma-common.c
> @@ -7,6 +7,8 @@
> #include <linux/module.h>
> #include <linux/slab.h>
> #include <linux/dma-mapping.h>
> +#include <linux/pm_runtime.h>
> +#include <linux/pm_domain.h>
>
> #include "fsl-edma-common.h"
>
> @@ -66,11 +68,46 @@ void fsl_edma_tx_chan_handler(struct fsl_edma_chan *fsl_chan)
> spin_unlock(&fsl_chan->vchan.lock);
> }
>
> +static void fsl_edma3_enable_request(struct fsl_edma_chan *fsl_chan)
> +{
> + u32 val, flags;
> +
> + flags = fsl_edma_drvflags(fsl_chan);
> + val = edma_readl_chreg(fsl_chan, ch_sbr);
> + /* Remote/local swapped wrongly on iMX8 QM Audio edma */
> + if (flags & FSL_EDMA_DRV_QUIRK_SWAPPED) {
> + if (!fsl_chan->is_rxchan)
> + val |= EDMA_V3_CH_SBR_RD;
> + else
> + val |= EDMA_V3_CH_SBR_WR;
> + } else {
> + if (fsl_chan->is_rxchan)
> + val |= EDMA_V3_CH_SBR_RD;
> + else
> + val |= EDMA_V3_CH_SBR_WR;
> + }
> +
> + if (fsl_chan->is_remote)
> + val &= ~(EDMA_V3_CH_SBR_RD | EDMA_V3_CH_SBR_WR);
> +
> + edma_writel_chreg(fsl_chan, val, ch_sbr);
> +
> + if (flags & FSL_EDMA_DRV_HAS_CHMUX)
> + edma_writel_chreg(fsl_chan, fsl_chan->srcid, ch_mux);
> +
> + val = edma_readl_chreg(fsl_chan, ch_csr);
> + val |= EDMA_V3_CH_CSR_ERQ;
> + edma_writel_chreg(fsl_chan, val, ch_csr);
> +}
> +
> static void fsl_edma_enable_request(struct fsl_edma_chan *fsl_chan)
> {
> struct edma_regs *regs = &fsl_chan->edma->regs;
> u32 ch = fsl_chan->vchan.chan.chan_id;
>
> + if (fsl_edma_drvflags(fsl_chan) & FSL_EDMA_DRV_SPLIT_REG)
> + return fsl_edma3_enable_request(fsl_chan);
> +
> if (fsl_chan->edma->drvdata->flags & FSL_EDMA_DRV_WRAP_IO) {
> edma_writeb(fsl_chan->edma, EDMA_SEEI_SEEI(ch), regs->seei);
> edma_writeb(fsl_chan->edma, ch, regs->serq);
> @@ -83,11 +120,28 @@ static void fsl_edma_enable_request(struct fsl_edma_chan *fsl_chan)
> }
> }
>
> +static void fsl_edma3_disable_request(struct fsl_edma_chan *fsl_chan)
> +{
> + u32 val = edma_readl_chreg(fsl_chan, ch_csr);
> + u32 flags;
> +
> + flags = fsl_edma_drvflags(fsl_chan);
> +
> + if (flags & FSL_EDMA_DRV_HAS_CHMUX)
> + edma_writel_chreg(fsl_chan, 0, ch_mux);
> +
> + val &= ~EDMA_V3_CH_CSR_ERQ;
> + edma_writel_chreg(fsl_chan, val, ch_csr);
> +}
> +
> void fsl_edma_disable_request(struct fsl_edma_chan *fsl_chan)
> {
> struct edma_regs *regs = &fsl_chan->edma->regs;
> u32 ch = fsl_chan->vchan.chan.chan_id;
>
> + if (fsl_edma_drvflags(fsl_chan) & FSL_EDMA_DRV_SPLIT_REG)
> + return fsl_edma3_disable_request(fsl_chan);
> +
> if (fsl_chan->edma->drvdata->flags & FSL_EDMA_DRV_WRAP_IO) {
> edma_writeb(fsl_chan->edma, ch, regs->cerq);
> edma_writeb(fsl_chan->edma, EDMA_CEEI_CEEI(ch), regs->ceei);
> @@ -135,6 +189,9 @@ void fsl_edma_chan_mux(struct fsl_edma_chan *fsl_chan,
> int endian_diff[4] = {3, 1, -1, -3};
> u32 dmamux_nr = fsl_chan->edma->drvdata->dmamuxs;
>
> + if (!dmamux_nr)
> + return;
> +
> chans_per_mux = fsl_chan->edma->n_chans / dmamux_nr;
> ch_off = fsl_chan->vchan.chan.chan_id % chans_per_mux;
>
> @@ -186,6 +243,10 @@ int fsl_edma_terminate_all(struct dma_chan *chan)
> vchan_get_all_descriptors(&fsl_chan->vchan, &head);
> spin_unlock_irqrestore(&fsl_chan->vchan.lock, flags);
> vchan_dma_desc_free_list(&fsl_chan->vchan, &head);
> +
> + if (fsl_edma_drvflags(fsl_chan) & FSL_EDMA_DRV_HAS_PD)
> + pm_runtime_allow(fsl_chan->pd_dev);
> +
> return 0;
> }
>
> @@ -286,12 +347,16 @@ static size_t fsl_edma_desc_residue(struct fsl_edma_chan *fsl_chan,
> enum dma_transfer_direction dir = edesc->dirn;
> dma_addr_t cur_addr, dma_addr;
> size_t len, size;
> + u32 nbytes = 0;
> int i;
>
> /* calculate the total size in this desc */
> - for (len = i = 0; i < fsl_chan->edesc->n_tcds; i++)
> - len += le32_to_cpu(edesc->tcd[i].vtcd->nbytes)
> - * le16_to_cpu(edesc->tcd[i].vtcd->biter);
> + for (len = i = 0; i < fsl_chan->edesc->n_tcds; i++) {
> + nbytes = le32_to_cpu(edesc->tcd[i].vtcd->nbytes);
> + if (nbytes & (EDMA_V3_TCD_NBYTES_DMLOE | EDMA_V3_TCD_NBYTES_SMLOE))
> + nbytes = EDMA_V3_TCD_NBYTES_MLOFF_NBYTES(nbytes);
> + len += nbytes * le16_to_cpu(edesc->tcd[i].vtcd->biter);
> + }
>
> if (!in_progress)
> return len;
> @@ -303,8 +368,12 @@ static size_t fsl_edma_desc_residue(struct fsl_edma_chan *fsl_chan,
>
> /* figure out the finished and calculate the residue */
> for (i = 0; i < fsl_chan->edesc->n_tcds; i++) {
> - size = le32_to_cpu(edesc->tcd[i].vtcd->nbytes)
> - * le16_to_cpu(edesc->tcd[i].vtcd->biter);
> + nbytes = le32_to_cpu(edesc->tcd[i].vtcd->nbytes);
> + if (nbytes & (EDMA_V3_TCD_NBYTES_DMLOE | EDMA_V3_TCD_NBYTES_SMLOE))
> + nbytes = EDMA_V3_TCD_NBYTES_MLOFF_NBYTES(nbytes);
> +
> + size = nbytes * le16_to_cpu(edesc->tcd[i].vtcd->biter);
> +
> if (dir == DMA_MEM_TO_DEV)
> dma_addr = le32_to_cpu(edesc->tcd[i].vtcd->saddr);
> else
> @@ -389,12 +458,15 @@ static void fsl_edma_set_tcd_regs(struct fsl_edma_chan *fsl_chan,
> }
>
> static inline
> -void fsl_edma_fill_tcd(struct fsl_edma_hw_tcd *tcd, u32 src, u32 dst,
> +void fsl_edma_fill_tcd(struct fsl_edma_chan *fsl_chan,
> + struct fsl_edma_hw_tcd *tcd, u32 src, u32 dst,
> u16 attr, u16 soff, u32 nbytes, u32 slast, u16 citer,
> u16 biter, u16 doff, u32 dlast_sga, bool major_int,
> bool disable_req, bool enable_sg)
> {
> + struct dma_slave_config *cfg = &fsl_chan->cfg;
> u16 csr = 0;
> + u32 burst;
>
> /*
> * eDMA hardware SGs require the TCDs to be stored in little
> @@ -409,6 +481,21 @@ void fsl_edma_fill_tcd(struct fsl_edma_hw_tcd *tcd, u32 src, u32 dst,
>
> tcd->soff = cpu_to_le16(soff);
>
> + if (fsl_chan->is_multi_fifo) {
> + /* set mloff to support multiple fifo */
> + burst = cfg->direction == DMA_DEV_TO_MEM ?
> + cfg->src_addr_width : cfg->dst_addr_width;
> + nbytes |= EDMA_V3_TCD_NBYTES_MLOFF(-(burst * 4));
> + /* enable DMLOE/SMLOE */
> + if (cfg->direction == DMA_MEM_TO_DEV) {
> + nbytes |= EDMA_V3_TCD_NBYTES_DMLOE;
> + nbytes &= ~EDMA_V3_TCD_NBYTES_SMLOE;
> + } else {
> + nbytes |= EDMA_V3_TCD_NBYTES_SMLOE;
> + nbytes &= ~EDMA_V3_TCD_NBYTES_DMLOE;
> + }
> + }
> +
> tcd->nbytes = cpu_to_le32(nbytes);
> tcd->slast = cpu_to_le32(slast);
>
> @@ -427,6 +514,12 @@ void fsl_edma_fill_tcd(struct fsl_edma_hw_tcd *tcd, u32 src, u32 dst,
> if (enable_sg)
> csr |= EDMA_TCD_CSR_E_SG;
>
> + if (fsl_chan->is_rxchan)
> + csr |= EDMA_TCD_CSR_ACTIVE;
> +
> + if (fsl_chan->is_sw)
> + csr |= EDMA_TCD_CSR_START;
> +
> tcd->csr = cpu_to_le16(csr);
> }
>
> @@ -466,6 +559,7 @@ struct dma_async_tx_descriptor *fsl_edma_prep_dma_cyclic(
> struct fsl_edma_chan *fsl_chan = to_fsl_edma_chan(chan);
> struct fsl_edma_desc *fsl_desc;
> dma_addr_t dma_buf_next;
> + bool major_int = true;
> int sg_len, i;
> u32 src_addr, dst_addr, last_sg, nbytes;
> u16 soff, doff, iter;
> @@ -509,17 +603,23 @@ struct dma_async_tx_descriptor *fsl_edma_prep_dma_cyclic(
> src_addr = dma_buf_next;
> dst_addr = fsl_chan->dma_dev_addr;
> soff = fsl_chan->cfg.dst_addr_width;
> - doff = 0;
> - } else {
> + doff = fsl_chan->is_multi_fifo ? 4 : 0;
> + } else if (direction == DMA_DEV_TO_MEM) {
> src_addr = fsl_chan->dma_dev_addr;
> dst_addr = dma_buf_next;
> - soff = 0;
> + soff = fsl_chan->is_multi_fifo ? 4 : 0;
> doff = fsl_chan->cfg.src_addr_width;
> + } else {
> + /* DMA_DEV_TO_DEV */
> + src_addr = fsl_chan->cfg.src_addr;
> + dst_addr = fsl_chan->cfg.dst_addr;
> + soff = doff = 0;
> + major_int = false;
> }
>
> - fsl_edma_fill_tcd(fsl_desc->tcd[i].vtcd, src_addr, dst_addr,
> + fsl_edma_fill_tcd(fsl_chan, fsl_desc->tcd[i].vtcd, src_addr, dst_addr,
> fsl_chan->attr, soff, nbytes, 0, iter,
> - iter, doff, last_sg, true, false, true);
> + iter, doff, last_sg, major_int, false, true);
> dma_buf_next += period_len;
> }
>
> @@ -568,23 +668,51 @@ struct dma_async_tx_descriptor *fsl_edma_prep_slave_sg(
> dst_addr = fsl_chan->dma_dev_addr;
> soff = fsl_chan->cfg.dst_addr_width;
> doff = 0;
> - } else {
> + } else if (direction == DMA_DEV_TO_MEM) {
> src_addr = fsl_chan->dma_dev_addr;
> dst_addr = sg_dma_address(sg);
> soff = 0;
> doff = fsl_chan->cfg.src_addr_width;
> + } else {
> + /* DMA_DEV_TO_DEV */
> + src_addr = fsl_chan->cfg.src_addr;
> + dst_addr = fsl_chan->cfg.dst_addr;
> + soff = 0;
> + doff = 0;
> }
>
> + /*
> + * Choose the suitable burst length if sg_dma_len is not
> + * multiple of burst length so that the whole transfer length is
> + * multiple of minor loop(burst length).
> + */
> + if (sg_dma_len(sg) % nbytes) {
> + u32 width = (direction == DMA_DEV_TO_MEM) ? doff : soff;
> + u32 burst = (direction == DMA_DEV_TO_MEM) ?
> + fsl_chan->cfg.src_maxburst :
> + fsl_chan->cfg.dst_maxburst;
> + int j;
> +
> + for (j = burst; j > 1; j--) {
> + if (!(sg_dma_len(sg) % (j * width))) {
> + nbytes = j * width;
> + break;
> + }
> + }
> + /* Set burst size as 1 if there's no suitable one */
> + if (j == 1)
> + nbytes = width;
> + }
> iter = sg_dma_len(sg) / nbytes;
> if (i < sg_len - 1) {
> last_sg = fsl_desc->tcd[(i + 1)].ptcd;
> - fsl_edma_fill_tcd(fsl_desc->tcd[i].vtcd, src_addr,
> + fsl_edma_fill_tcd(fsl_chan, fsl_desc->tcd[i].vtcd, src_addr,
> dst_addr, fsl_chan->attr, soff,
> nbytes, 0, iter, iter, doff, last_sg,
> false, false, true);
> } else {
> last_sg = 0;
> - fsl_edma_fill_tcd(fsl_desc->tcd[i].vtcd, src_addr,
> + fsl_edma_fill_tcd(fsl_chan, fsl_desc->tcd[i].vtcd, src_addr,
> dst_addr, fsl_chan->attr, soff,
> nbytes, 0, iter, iter, doff, last_sg,
> true, true, false);
> @@ -609,7 +737,7 @@ struct dma_async_tx_descriptor *fsl_edma_prep_memcpy(struct dma_chan *chan,
> fsl_chan->is_sw = true;
>
> /* To match with copy_align and max_seg_size so 1 tcd is enough */
> - fsl_edma_fill_tcd(fsl_desc->tcd[0].vtcd, dma_src, dma_dst,
> + fsl_edma_fill_tcd(fsl_chan, fsl_desc->tcd[0].vtcd, dma_src, dma_dst,
> fsl_edma_get_tcd_attr(DMA_SLAVE_BUSWIDTH_32_BYTES),
> 32, len, 0, 1, 1, 32, 0, true, true, false);
>
> diff --git a/drivers/dma/fsl-edma-common.h b/drivers/dma/fsl-edma-common.h
> index cfc41915eaa1..3cc0cc8fc2d0 100644
> --- a/drivers/dma/fsl-edma-common.h
> +++ b/drivers/dma/fsl-edma-common.h
> @@ -42,6 +42,11 @@
> #define EDMA_TCD_CSR_ACTIVE BIT(6)
> #define EDMA_TCD_CSR_DONE BIT(7)
>
> +#define EDMA_V3_TCD_NBYTES_MLOFF_NBYTES(x) ((x) & GENMASK(9, 0))
> +#define EDMA_V3_TCD_NBYTES_MLOFF(x) (x << 10)
> +#define EDMA_V3_TCD_NBYTES_DMLOE (1 << 30)
> +#define EDMA_V3_TCD_NBYTES_SMLOE (1 << 31)
> +
> #define EDMAMUX_CHCFG_DIS 0x0
> #define EDMAMUX_CHCFG_ENBL 0x80
> #define EDMAMUX_CHCFG_SOURCE(n) ((n) & 0x3F)
> @@ -54,6 +59,15 @@
> BIT(DMA_SLAVE_BUSWIDTH_2_BYTES) | \
> BIT(DMA_SLAVE_BUSWIDTH_4_BYTES) | \
> BIT(DMA_SLAVE_BUSWIDTH_8_BYTES))
> +
> +#define EDMA_V3_CH_SBR_RD BIT(22)
> +#define EDMA_V3_CH_SBR_WR BIT(21)
> +#define EDMA_V3_CH_CSR_ERQ BIT(0)
> +#define EDMA_V3_CH_CSR_EARQ BIT(1)
> +#define EDMA_V3_CH_CSR_EEI BIT(2)
> +#define EDMA_V3_CH_CSR_DONE BIT(30)
> +#define EDMA_V3_CH_CSR_ACTIVE BIT(31)
> +
> enum fsl_edma_pm_state {
> RUNNING = 0,
> SUSPENDED,
> @@ -73,6 +87,18 @@ struct fsl_edma_hw_tcd {
> __le16 biter;
> };
>
> +struct fsl_edma3_ch_reg {
> + __le32 ch_csr;
> + __le32 ch_es;
> + __le32 ch_int;
> + __le32 ch_sbr;
> + __le32 ch_pri;
> + __le32 ch_mux;
> + __le32 ch_mattr; /* edma4, reserved for edma3 */
> + __le32 ch_reserved;
> + struct fsl_edma_hw_tcd tcd;
> +} __packed;
> +
> /*
> * These are iomem pointers, for both v32 and v64.
> */
> @@ -119,6 +145,18 @@ struct fsl_edma_chan {
> enum dma_data_direction dma_dir;
> char chan_name[32];
> struct fsl_edma_hw_tcd __iomem *tcd;
> + u32 real_count;
> + struct work_struct issue_worker;
> + struct platform_device *pdev;
> + struct device *pd_dev;
> + u32 srcid;
> + struct clk *clk;
> + int priority;
> + int hw_chanid;
> + int txirq;
> + bool is_rxchan;
> + bool is_remote;
> + bool is_multi_fifo;
> };
>
> struct fsl_edma_desc {
> @@ -135,8 +173,26 @@ struct fsl_edma_desc {
> #define FSL_EDMA_DRV_CONFIG32 BIT(2)
> #define FSL_EDMA_DRV_WRAP_IO BIT(3)
> #define FSL_EDMA_DRV_EDMA64 BIT(4)
> +#define FSL_EDMA_DRV_HAS_PD BIT(5)
> +#define FSL_EDMA_DRV_HAS_CHCLK BIT(6)
> +#define FSL_EDMA_DRV_HAS_CHMUX BIT(7)
> +/* imx8 QM audio edma remote local swapped */
> +#define FSL_EDMA_DRV_QUIRK_SWAPPED BIT(8)
> +/* control and status register is in tcd address space, edma3 reg layout */
> +#define FSL_EDMA_DRV_SPLIT_REG BIT(9)
> +#define FSL_EDMA_DRV_BUS_8BYTE BIT(10)
> +#define FSL_EDMA_DRV_DEV_TO_DEV BIT(11)
> +#define FSL_EDMA_DRV_ALIGN_64BYTE BIT(12)
> +
> +#define FSL_EDMA_DRV_EDMA3 (FSL_EDMA_DRV_SPLIT_REG | \
> + FSL_EDMA_DRV_BUS_8BYTE | \
> + FSL_EDMA_DRV_DEV_TO_DEV | \
> + FSL_EDMA_DRV_ALIGN_64BYTE)
> +
> struct fsl_edma_drvdata {
> - u32 dmamuxs;
> + u32 dmamuxs; /* only used before v3 */
> + u32 chreg_off;
> + u32 chreg_space_sz;
> u32 flags;
> int (*setup_irq)(struct platform_device *pdev,
> struct fsl_edma_engine *fsl_edma);
> @@ -148,6 +204,7 @@ struct fsl_edma_engine {
> void __iomem *muxbase[DMAMUX_NR];
> struct clk *muxclk[DMAMUX_NR];
> struct clk *dmaclk;
> + struct clk *chclk;
> struct mutex fsl_edma_mutex;
> const struct fsl_edma_drvdata *drvdata;
> u32 n_chans;
> @@ -155,6 +212,7 @@ struct fsl_edma_engine {
> int errirq;
> bool big_endian;
> struct edma_regs regs;
> + u64 chan_masked;
> struct fsl_edma_chan chans[];
> };
>
> @@ -168,6 +226,14 @@ struct fsl_edma_engine {
> edma_writel(chan->edma, (u32 __force)val, &chan->tcd->__name) : \
> edma_writew(chan->edma, (u16 __force)val, &chan->tcd->__name))
>
> +#define edma_readl_chreg(chan, __name) \
> + edma_readl(chan->edma, \
> + (void __iomem *)&(container_of(chan->tcd, struct fsl_edma3_ch_reg, tcd)->__name))
> +
> +#define edma_writel_chreg(chan, val, __name) \
> + edma_writel(chan->edma, val, \
> + (void __iomem *)&(container_of(chan->tcd, struct fsl_edma3_ch_reg, tcd)->__name))
> +
> /*
> * R/W functions for big- or little-endian registers:
> * The eDMA controller's endian is independent of the CPU core's endian.
> @@ -224,6 +290,11 @@ static inline struct fsl_edma_chan *to_fsl_edma_chan(struct dma_chan *chan)
> return container_of(chan, struct fsl_edma_chan, vchan.chan);
> }
>
> +static inline u32 fsl_edma_drvflags(struct fsl_edma_chan *fsl_chan)
> +{
> + return fsl_chan->edma->drvdata->flags;
> +}
> +
> static inline struct fsl_edma_desc *to_fsl_edma_desc(struct virt_dma_desc *vd)
> {
> return container_of(vd, struct fsl_edma_desc, vdesc);
> diff --git a/drivers/dma/fsl-edma-main.c b/drivers/dma/fsl-edma-main.c
> index 8ba705080f5d..b0ddb82862ea 100644
> --- a/drivers/dma/fsl-edma-main.c
> +++ b/drivers/dma/fsl-edma-main.c
> @@ -18,9 +18,15 @@
> #include <linux/of_irq.h>
> #include <linux/of_dma.h>
> #include <linux/dma-mapping.h>
> +#include <linux/pm_runtime.h>
> +#include <linux/pm_domain.h>
>
> #include "fsl-edma-common.h"
>
> +#define ARGS_RX BIT(0)
> +#define ARGS_REMOTE BIT(1)
> +#define ARGS_MULTI_FIFO BIT(2)
> +
> static void fsl_edma_synchronize(struct dma_chan *chan)
> {
> struct fsl_edma_chan *fsl_chan = to_fsl_edma_chan(chan);
> @@ -47,6 +53,22 @@ static irqreturn_t fsl_edma_tx_handler(int irq, void *dev_id)
> return IRQ_HANDLED;
> }
>
> +static irqreturn_t fsl_edma3_tx_handler(int irq, void *dev_id)
> +{
> + struct fsl_edma_chan *fsl_chan = dev_id;
> + unsigned int intr;
> +
> + intr = edma_readl_chreg(fsl_chan, ch_int);
> + if (!intr)
> + return IRQ_HANDLED;
> +
> + edma_writel_chreg(fsl_chan, 1, ch_int);
> +
> + fsl_edma_tx_chan_handler(fsl_chan);
> +
> + return IRQ_HANDLED;
> +}
> +
> static irqreturn_t fsl_edma_err_handler(int irq, void *dev_id)
> {
> struct fsl_edma_engine *fsl_edma = dev_id;
> @@ -108,6 +130,51 @@ static struct dma_chan *fsl_edma_xlate(struct of_phandle_args *dma_spec,
> return NULL;
> }
>
> +static struct dma_chan *fsl_edma3_xlate(struct of_phandle_args *dma_spec,
> + struct of_dma *ofdma)
> +{
> + struct fsl_edma_engine *fsl_edma = ofdma->of_dma_data;
> + struct dma_chan *chan, *_chan;
> + struct fsl_edma_chan *fsl_chan;
> + bool b_chmux;
> + int i;
> +
> + if (dma_spec->args_count != 3)
> + return NULL;
> +
> + b_chmux = !!(fsl_edma->drvdata->flags & FSL_EDMA_DRV_HAS_CHMUX);
> +
> + mutex_lock(&fsl_edma->fsl_edma_mutex);
> + list_for_each_entry_safe(chan, _chan, &fsl_edma->dma_dev.channels,
> + device_node) {
> +
> + if (chan->client_count)
> + continue;
> +
> + fsl_chan = to_fsl_edma_chan(chan);
> + i = fsl_chan - fsl_edma->chans;
> +
> + chan = dma_get_slave_channel(chan);
> + chan->device->privatecnt++;
> + fsl_chan->priority = dma_spec->args[1];
> + fsl_chan->is_rxchan = dma_spec->args[2] & ARGS_RX;
> + fsl_chan->is_remote = dma_spec->args[2] & ARGS_REMOTE;
> + fsl_chan->is_multi_fifo = dma_spec->args[2] & ARGS_MULTI_FIFO;
> +
> + if (!b_chmux && i == dma_spec->args[0]) {
> + mutex_unlock(&fsl_edma->fsl_edma_mutex);
> + return chan;
> + } else if (b_chmux && !fsl_chan->srcid) {
> + /* if controller support channel mux, choose a free channel */
> + fsl_chan->srcid = dma_spec->args[0];
> + mutex_unlock(&fsl_edma->fsl_edma_mutex);
> + return chan;
> + }
> + }
> + mutex_unlock(&fsl_edma->fsl_edma_mutex);
> + return NULL;
> +}
> +
> static int
> fsl_edma_irq_init(struct platform_device *pdev, struct fsl_edma_engine *fsl_edma)
> {
> @@ -149,6 +216,37 @@ fsl_edma_irq_init(struct platform_device *pdev, struct fsl_edma_engine *fsl_edma
> return 0;
> }
>
> +static int fsl_edma3_irq_init(struct platform_device *pdev, struct fsl_edma_engine *fsl_edma)
> +{
> + int ret;
> + int i;
> +
> + for (i = 0; i < fsl_edma->n_chans; i++) {
> +
> + struct fsl_edma_chan *fsl_chan = &fsl_edma->chans[i];
> +
> + if (fsl_edma->chan_masked & BIT(i))
> + continue;
> +
> + /* request channel irq */
> + fsl_chan->txirq = platform_get_irq(pdev, i);
> + if (fsl_chan->txirq < 0) {
> + dev_err(&pdev->dev, "Can't get chan %d's irq.\n", i);
> + return -EINVAL;
> + }
> +
> + ret = devm_request_irq(&pdev->dev, fsl_chan->txirq,
> + fsl_edma3_tx_handler, IRQF_SHARED,
> + fsl_chan->chan_name, fsl_chan);
> + if (ret) {
> + dev_err(&pdev->dev, "Can't register chan%d's IRQ.\n", i);
> + return -EINVAL;
> + }
> + }
> +
> + return 0;
> +}
> +
> static int
> fsl_edma2_irq_init(struct platform_device *pdev,
> struct fsl_edma_engine *fsl_edma)
> @@ -203,29 +301,109 @@ static void fsl_disable_clocks(struct fsl_edma_engine *fsl_edma, int nr_clocks)
> static struct fsl_edma_drvdata vf610_data = {
> .dmamuxs = DMAMUX_NR,
> .flags = FSL_EDMA_DRV_WRAP_IO,
> + .chreg_off = EDMA_TCD,
> + .chreg_space_sz = sizeof(struct fsl_edma_hw_tcd),
> .setup_irq = fsl_edma_irq_init,
> };
>
> static struct fsl_edma_drvdata ls1028a_data = {
> .dmamuxs = DMAMUX_NR,
> .flags = FSL_EDMA_DRV_MUX_SWAP | FSL_EDMA_DRV_WRAP_IO,
> + .chreg_off = EDMA_TCD,
> + .chreg_space_sz = sizeof(struct fsl_edma_hw_tcd),
> .setup_irq = fsl_edma_irq_init,
> };
>
> static struct fsl_edma_drvdata imx7ulp_data = {
> .dmamuxs = 1,
> + .chreg_off = EDMA_TCD,
> + .chreg_space_sz = sizeof(struct fsl_edma_hw_tcd),
> .flags = FSL_EDMA_DRV_HAS_DMACLK | FSL_EDMA_DRV_CONFIG32,
> .setup_irq = fsl_edma2_irq_init,
> };
>
> +static struct fsl_edma_drvdata imx8qm_data = {
> + .flags = FSL_EDMA_DRV_HAS_PD | FSL_EDMA_DRV_EDMA3,
> + .chreg_space_sz = 0x10000,
> + .chreg_off = 0x10000,
> + .setup_irq = fsl_edma3_irq_init,
> +};
> +
> +static struct fsl_edma_drvdata imx8qm_audio_data = {
> + .flags = FSL_EDMA_DRV_QUIRK_SWAPPED | FSL_EDMA_DRV_HAS_PD | FSL_EDMA_DRV_EDMA3,
> + .chreg_space_sz = 0x10000,
> + .chreg_off = 0x10000,
> + .setup_irq = fsl_edma3_irq_init,
> +};
> +
> +static struct fsl_edma_drvdata imx93_data3 = {
> + .flags = FSL_EDMA_DRV_HAS_DMACLK | FSL_EDMA_DRV_EDMA3,
> + .chreg_space_sz = 0x10000,
> + .chreg_off = 0x10000,
> + .setup_irq = fsl_edma3_irq_init,
> +};
> +
> +static struct fsl_edma_drvdata imx93_data4 = {
> + .flags = FSL_EDMA_DRV_HAS_CHMUX | FSL_EDMA_DRV_HAS_DMACLK | FSL_EDMA_DRV_EDMA3,
> + .chreg_space_sz = 0x8000,
> + .chreg_off = 0x10000,
> + .setup_irq = fsl_edma3_irq_init,
> +};
> +
> static const struct of_device_id fsl_edma_dt_ids[] = {
> { .compatible = "fsl,vf610-edma", .data = &vf610_data},
> { .compatible = "fsl,ls1028a-edma", .data = &ls1028a_data},
> { .compatible = "fsl,imx7ulp-edma", .data = &imx7ulp_data},
> + { .compatible = "fsl,imx8qm-edma", .data = &imx8qm_data},
> + { .compatible = "fsl,imx8qm-adma", .data = &imx8qm_audio_data},
> + { .compatible = "fsl,imx93-edma3", .data = &imx93_data3},
> + { .compatible = "fsl,imx93-edma4", .data = &imx93_data4},
> { /* sentinel */ }
> };
> MODULE_DEVICE_TABLE(of, fsl_edma_dt_ids);
>
> +static int fsl_edma3_attach_pd(struct platform_device *pdev, struct fsl_edma_engine *fsl_edma)
> +{
> + struct fsl_edma_chan *fsl_chan;
> + struct device_link *link;
> + struct device *pd_chan;
> + struct device *dev;
> + int i;
> +
> + dev = &pdev->dev;
> +
> + for (i = 0; i < fsl_edma->n_chans; i++) {
> + if (fsl_edma->chan_masked & BIT(i))
> + continue;
> +
> + fsl_chan = &fsl_edma->chans[i];
> +
> + pd_chan = dev_pm_domain_attach_by_id(dev, i);
> + if (IS_ERR_OR_NULL(pd_chan)) {
> + dev_err(dev, "Failed attach pd %d\n", i);
> + return -EINVAL;
> + }
> +
> + link = device_link_add(dev, pd_chan, DL_FLAG_STATELESS |
> + DL_FLAG_PM_RUNTIME |
> + DL_FLAG_RPM_ACTIVE);
> + if (IS_ERR(link)) {
> + dev_err(dev, "Failed to add device_link to %d: %ld\n", i,
> + PTR_ERR(link));
> + return -EINVAL;
> + }
> +
> + fsl_chan->pd_dev = pd_chan;
> +
> + pm_runtime_use_autosuspend(fsl_chan->pd_dev);
> + pm_runtime_set_autosuspend_delay(fsl_chan->pd_dev, 200);
> + pm_runtime_set_active(fsl_chan->pd_dev);
> + //pm_runtime_put_sync_suspend(fsl_chan->pd_dev);

Ouch!

Always remember to run checkpatch on your code before sending!

> + }
> +
> + return 0;
> +}
> +
> static int fsl_edma_probe(struct platform_device *pdev)
> {
> const struct of_device_id *of_id =
> @@ -234,6 +412,7 @@ static int fsl_edma_probe(struct platform_device *pdev)
> struct fsl_edma_engine *fsl_edma;
> const struct fsl_edma_drvdata *drvdata = NULL;
> struct fsl_edma_chan *fsl_chan;
> + u32 chan_mask[2] = {0, 0};
> struct edma_regs *regs;
> int len, chans;
> int ret, i;
> @@ -264,8 +443,10 @@ static int fsl_edma_probe(struct platform_device *pdev)
> if (IS_ERR(fsl_edma->membase))
> return PTR_ERR(fsl_edma->membase);
>
> - fsl_edma_setup_regs(fsl_edma);
> - regs = &fsl_edma->regs;
> + if (!(drvdata->flags & FSL_EDMA_DRV_SPLIT_REG)) {
> + fsl_edma_setup_regs(fsl_edma);
> + regs = &fsl_edma->regs;
> + }
>
> if (drvdata->flags & FSL_EDMA_DRV_HAS_DMACLK) {
> fsl_edma->dmaclk = devm_clk_get_enabled(&pdev->dev, "dma");
> @@ -275,9 +456,29 @@ static int fsl_edma_probe(struct platform_device *pdev)
> }
> }
>
> + if (drvdata->flags & FSL_EDMA_DRV_HAS_CHCLK) {
> + fsl_edma->chclk = devm_clk_get_enabled(&pdev->dev, "mp");
> + if (IS_ERR(fsl_edma->chclk)) {
> + dev_err(&pdev->dev, "Missing MP block clock.\n");
> + return PTR_ERR(fsl_edma->chclk);
> + }
> + }
> +
> + ret = of_property_read_variable_u32_array(np, "dma-channel-mask", chan_mask, 1, 2);
> +
> + if (ret > 0) {
> + fsl_edma->chan_masked = chan_mask[1];
> + fsl_edma->chan_masked <<= 32;
> + fsl_edma->chan_masked |= chan_mask[0];
> + }
> +
> for (i = 0; i < fsl_edma->drvdata->dmamuxs; i++) {
> char clkname[32];
>
> + /* eDMAv3 mux register move to TCD area if ch_mux exist */
> + if (drvdata->flags & FSL_EDMA_DRV_SPLIT_REG)
> + break;
> +
> fsl_edma->muxbase[i] = devm_platform_ioremap_resource(pdev,
> 1 + i);
> if (IS_ERR(fsl_edma->muxbase[i])) {
> @@ -297,9 +498,18 @@ static int fsl_edma_probe(struct platform_device *pdev)
>
> fsl_edma->big_endian = of_property_read_bool(np, "big-endian");
>
> + if (drvdata->flags & FSL_EDMA_DRV_HAS_PD) {
> + ret = fsl_edma3_attach_pd(pdev, fsl_edma);
> + if (ret)
> + return ret;
> + }
> +
> INIT_LIST_HEAD(&fsl_edma->dma_dev.channels);
> for (i = 0; i < fsl_edma->n_chans; i++) {
> - struct fsl_edma_chan *fsl_chan = &fsl_edma->chans[i];
> + fsl_chan = &fsl_edma->chans[i];
> +
> + if (fsl_edma->chan_masked & BIT(i))
> + continue;
>
> snprintf(fsl_chan->chan_name, sizeof(fsl_chan->chan_name), "%s-CH%02d",
> dev_name(&pdev->dev), i);
> @@ -310,8 +520,13 @@ static int fsl_edma_probe(struct platform_device *pdev)
> fsl_chan->idle = true;
> fsl_chan->dma_dir = DMA_NONE;
> fsl_chan->vchan.desc_free = fsl_edma_free_desc;
> - fsl_chan->tcd = fsl_edma->membase + EDMA_TCD
> - + i * sizeof(struct fsl_edma_hw_tcd);
> +
> + len = (drvdata->flags & FSL_EDMA_DRV_SPLIT_REG) ?
> + offsetof(struct fsl_edma3_ch_reg, tcd) : 0;
> + fsl_chan->tcd = fsl_edma->membase
> + + i * drvdata->chreg_space_sz + drvdata->chreg_off + len;
> +
> + fsl_chan->pdev = pdev;
> vchan_init(&fsl_chan->vchan, &fsl_edma->dma_dev);
>
> edma_write_tcdreg(fsl_chan, 0, csr);
> @@ -345,12 +560,25 @@ static int fsl_edma_probe(struct platform_device *pdev)
>
> fsl_edma->dma_dev.src_addr_widths = FSL_EDMA_BUSWIDTHS;
> fsl_edma->dma_dev.dst_addr_widths = FSL_EDMA_BUSWIDTHS;
> +
> + if (drvdata->flags & FSL_EDMA_DRV_BUS_8BYTE) {
> + fsl_edma->dma_dev.src_addr_widths |= BIT(DMA_SLAVE_BUSWIDTH_8_BYTES);
> + fsl_edma->dma_dev.dst_addr_widths |= BIT(DMA_SLAVE_BUSWIDTH_8_BYTES);
> + }
> +
> fsl_edma->dma_dev.directions = BIT(DMA_DEV_TO_MEM) | BIT(DMA_MEM_TO_DEV);
> + if (drvdata->flags & FSL_EDMA_DRV_DEV_TO_DEV)
> + fsl_edma->dma_dev.directions |= BIT(DMA_DEV_TO_DEV);
> +
> + fsl_edma->dma_dev.copy_align = drvdata->flags & FSL_EDMA_DRV_ALIGN_64BYTE ?
> + DMAENGINE_ALIGN_64_BYTES :
> + DMAENGINE_ALIGN_32_BYTES;
>
> - fsl_edma->dma_dev.copy_align = DMAENGINE_ALIGN_32_BYTES;
> /* Per worst case 'nbytes = 1' take CITER as the max_seg_size */
> dma_set_max_seg_size(fsl_edma->dma_dev.dev, 0x3fff);
>
> + fsl_edma->dma_dev.residue_granularity = DMA_RESIDUE_GRANULARITY_SEGMENT;
> +
> platform_set_drvdata(pdev, fsl_edma);
>
> ret = dma_async_device_register(&fsl_edma->dma_dev);
> @@ -360,7 +588,9 @@ static int fsl_edma_probe(struct platform_device *pdev)
> return ret;
> }
>
> - ret = of_dma_controller_register(np, fsl_edma_xlate, fsl_edma);
> + ret = of_dma_controller_register(np,
> + drvdata->flags & FSL_EDMA_DRV_SPLIT_REG ? fsl_edma3_xlate : fsl_edma_xlate,
> + fsl_edma);
> if (ret) {
> dev_err(&pdev->dev,
> "Can't register Freescale eDMA of_dma. (%d)\n", ret);
> @@ -369,7 +599,8 @@ static int fsl_edma_probe(struct platform_device *pdev)
> }
>
> /* enable round robin arbitration */
> - edma_writel(fsl_edma, EDMA_CR_ERGA | EDMA_CR_ERCA, regs->cr);
> + if (!(drvdata->flags & FSL_EDMA_DRV_SPLIT_REG))
> + edma_writel(fsl_edma, EDMA_CR_ERGA | EDMA_CR_ERCA, regs->cr);
>
> return 0;
> }
> --
> 2.34.1

--
~Vinod