The current descriptor layout is:
struct dw_edma_desc *desc
└─ chunk list
└─ burst[]
Creating a DMA descriptor requires at least two kzalloc() calls because
each chunk is allocated as a linked-list node. Since the number of bursts
is already known when the descriptor is created, this linked-list layer is
unnecessary.
Move the burst array directly into struct dw_edma_desc and remove the
struct dw_edma_chunk layer entirely.
Use start_burst and done_burst to track the current bursts, which current
are in the DMA link list.
Signed-off-by: Frank Li <Frank.Li@nxp.com>
---
drivers/dma/dw-edma/dw-edma-core.c | 130 ++++++++++++-------------------------
drivers/dma/dw-edma/dw-edma-core.h | 24 ++++---
2 files changed, 57 insertions(+), 97 deletions(-)
diff --git a/drivers/dma/dw-edma/dw-edma-core.c b/drivers/dma/dw-edma/dw-edma-core.c
index 9e65155fd93d69ddbc8235fad671fad4dc120979..6e7b7ee99aaf6e1c3e354d36ca058813dc95b8dd 100644
--- a/drivers/dma/dw-edma/dw-edma-core.c
+++ b/drivers/dma/dw-edma/dw-edma-core.c
@@ -40,76 +40,45 @@ u64 dw_edma_get_pci_address(struct dw_edma_chan *chan, phys_addr_t cpu_addr)
return cpu_addr;
}
-static struct dw_edma_chunk *dw_edma_alloc_chunk(struct dw_edma_desc *desc, u32 nburst)
-{
- struct dw_edma_chan *chan = desc->chan;
- struct dw_edma_chunk *chunk;
-
- chunk = kzalloc(struct_size(chunk, burst, nburst), GFP_NOWAIT);
- if (unlikely(!chunk))
- return NULL;
-
- chunk->chan = chan;
- /* Toggling change bit (CB) in each chunk, this is a mechanism to
- * inform the eDMA HW block that this is a new linked list ready
- * to be consumed.
- * - Odd chunks originate CB equal to 0
- * - Even chunks originate CB equal to 1
- */
- chunk->cb = !(desc->chunks_alloc % 2);
-
- chunk->nburst = nburst;
-
- list_add_tail(&chunk->list, &desc->chunk_list);
- desc->chunks_alloc++;
-
- return chunk;
-}
-
-static struct dw_edma_desc *dw_edma_alloc_desc(struct dw_edma_chan *chan)
+static struct dw_edma_desc *
+dw_edma_alloc_desc(struct dw_edma_chan *chan, u32 nburst)
{
struct dw_edma_desc *desc;
- desc = kzalloc(sizeof(*desc), GFP_NOWAIT);
+ desc = kzalloc(struct_size(desc, burst, nburst), GFP_NOWAIT);
if (unlikely(!desc))
return NULL;
desc->chan = chan;
-
- INIT_LIST_HEAD(&desc->chunk_list);
+ desc->nburst = nburst;
+ desc->cb = true;
return desc;
}
-static void dw_edma_free_desc(struct dw_edma_desc *desc)
-{
- struct dw_edma_chunk *child, *_next;
-
- /* Remove all the list elements */
- list_for_each_entry_safe(child, _next, &desc->chunk_list, list) {
- list_del(&child->list);
- kfree(child);
- desc->chunks_alloc--;
- }
-
- kfree(desc);
-}
-
static void vchan_free_desc(struct virt_dma_desc *vdesc)
{
- dw_edma_free_desc(vd2dw_edma_desc(vdesc));
+ kfree(vd2dw_edma_desc(vdesc));
}
-static void dw_edma_core_start(struct dw_edma_chunk *chunk, bool first)
+static void dw_edma_core_start(struct dw_edma_desc *desc, bool first)
{
- struct dw_edma_chan *chan = chunk->chan;
+ struct dw_edma_chan *chan = desc->chan;
u32 i = 0;
- for (i = 0; i < chunk->nburst; i++)
- dw_edma_core_ll_data(chan, &chunk->burst[i], i, chunk->cb,
- i == chunk->nburst - 1);
+ for (i = 0; i < desc->nburst; i++) {
+ if (i == chan->ll_max - 1)
+ break;
+
+ dw_edma_core_ll_data(chan, &desc->burst[i + desc->start_burst],
+ i, desc->cb,
+ i == desc->nburst - 1 || i == chan->ll_max - 2);
+ }
- dw_edma_core_ll_link(chan, i, chunk->cb, chan->ll_region.paddr);
+ desc->done_burst = desc->start_burst;
+ desc->start_burst += i;
+
+ dw_edma_core_ll_link(chan, i, desc->cb, chan->ll_region.paddr);
if (first)
dw_edma_core_ch_enable(chan);
@@ -119,7 +88,6 @@ static void dw_edma_core_start(struct dw_edma_chunk *chunk, bool first)
static int dw_edma_start_transfer(struct dw_edma_chan *chan)
{
- struct dw_edma_chunk *child;
struct dw_edma_desc *desc;
struct virt_dma_desc *vd;
@@ -131,16 +99,9 @@ static int dw_edma_start_transfer(struct dw_edma_chan *chan)
if (!desc)
return 0;
- child = list_first_entry_or_null(&desc->chunk_list,
- struct dw_edma_chunk, list);
- if (!child)
- return 0;
+ dw_edma_core_start(desc, !desc->start_burst);
- dw_edma_core_start(child, !desc->xfer_sz);
- desc->xfer_sz += child->xfer_sz;
- list_del(&child->list);
- kfree(child);
- desc->chunks_alloc--;
+ desc->cb = !desc->cb;
return 1;
}
@@ -289,8 +250,10 @@ dw_edma_device_tx_status(struct dma_chan *dchan, dma_cookie_t cookie,
vd = vchan_find_desc(&chan->vc, cookie);
if (vd) {
desc = vd2dw_edma_desc(vd);
- if (desc)
- residue = desc->alloc_sz - desc->xfer_sz;
+
+ residue = desc->alloc_sz;
+ if (desc && desc->done_burst)
+ residue -= desc->burst[desc->done_burst].xfer_sz;
}
spin_unlock_irqrestore(&chan->vc.lock, flags);
@@ -307,7 +270,6 @@ dw_edma_device_transfer(struct dw_edma_transfer *xfer,
struct dw_edma_chan *chan = dchan2dw_edma_chan(xfer->dchan);
enum dma_transfer_direction dir = xfer->direction;
struct scatterlist *sg = NULL;
- struct dw_edma_chunk *chunk = NULL;
struct dw_edma_burst *burst;
struct dw_edma_desc *desc;
u64 src_addr, dst_addr;
@@ -369,10 +331,6 @@ dw_edma_device_transfer(struct dw_edma_transfer *xfer,
return NULL;
}
- desc = dw_edma_alloc_desc(chan);
- if (unlikely(!desc))
- goto err_alloc;
-
if (xfer->type == EDMA_XFER_INTERLEAVED) {
src_addr = xfer->xfer.il->src_start;
dst_addr = xfer->xfer.il->dst_start;
@@ -396,19 +354,15 @@ dw_edma_device_transfer(struct dw_edma_transfer *xfer,
fsz = xfer->xfer.il->frame_size;
}
+ desc = dw_edma_alloc_desc(chan, cnt);
+ if (unlikely(!desc))
+ return NULL;
+
for (i = 0; i < cnt; i++) {
if (xfer->type == EDMA_XFER_SCATTER_GATHER && !sg)
break;
- if (!(i % chan->ll_max)) {
- u32 n = min(cnt - i, chan->ll_max);
-
- chunk = dw_edma_alloc_chunk(desc, n);
- if (unlikely(!chunk))
- goto err_alloc;
- }
-
- burst = chunk->burst + (i % chan->ll_max);
+ burst = desc->burst + i;
if (xfer->type == EDMA_XFER_CYCLIC)
burst->sz = xfer->xfer.cyclic.len;
@@ -417,8 +371,8 @@ dw_edma_device_transfer(struct dw_edma_transfer *xfer,
else if (xfer->type == EDMA_XFER_INTERLEAVED)
burst->sz = xfer->xfer.il->sgl[i % fsz].size;
- chunk->xfer_sz += burst->sz;
desc->alloc_sz += burst->sz;
+ burst->xfer_sz = desc->alloc_sz;
if (dir == DMA_DEV_TO_MEM) {
burst->sar = src_addr;
@@ -473,12 +427,6 @@ dw_edma_device_transfer(struct dw_edma_transfer *xfer,
}
return vchan_tx_prep(&chan->vc, &desc->vd, xfer->flags);
-
-err_alloc:
- if (desc)
- dw_edma_free_desc(desc);
-
- return NULL;
}
static struct dma_async_tx_descriptor *
@@ -551,8 +499,14 @@ static void dw_hdma_set_callback_result(struct virt_dma_desc *vd,
return;
desc = vd2dw_edma_desc(vd);
- if (desc)
- residue = desc->alloc_sz - desc->xfer_sz;
+ residue = desc->alloc_sz;
+
+ if (desc) {
+ if (result == DMA_TRANS_NOERROR)
+ residue -= desc->burst[desc->start_burst - 1].xfer_sz;
+ else if (desc->done_burst)
+ residue -= desc->burst[desc->done_burst - 1].xfer_sz;
+ }
res = &vd->tx_result;
res->result = result;
@@ -571,7 +525,7 @@ static void dw_edma_done_interrupt(struct dw_edma_chan *chan)
switch (chan->request) {
case EDMA_REQ_NONE:
desc = vd2dw_edma_desc(vd);
- if (!desc->chunks_alloc) {
+ if (desc->start_burst >= desc->nburst) {
dw_hdma_set_callback_result(vd,
DMA_TRANS_NOERROR);
list_del(&vd->node);
@@ -936,7 +890,7 @@ int dw_edma_probe(struct dw_edma_chip *chip)
goto err_irq_free;
/* Turn debugfs on */
- dw_edma_core_debugfs_on(dw);
+ //dw_edma_core_debugfs_on(dw);
chip->dw = dw;
diff --git a/drivers/dma/dw-edma/dw-edma-core.h b/drivers/dma/dw-edma/dw-edma-core.h
index 1930c3bce2bf33fdfbf4e8d99002483a4565faed..ba83c42dee5224dccdf34cec6481e9404a607702 100644
--- a/drivers/dma/dw-edma/dw-edma-core.h
+++ b/drivers/dma/dw-edma/dw-edma-core.h
@@ -46,15 +46,8 @@ struct dw_edma_burst {
u64 sar;
u64 dar;
u32 sz;
-};
-
-struct dw_edma_chunk {
- struct list_head list;
- struct dw_edma_chan *chan;
- u8 cb;
+ /* precalulate summary of previous burst total size */
u32 xfer_sz;
- u32 nburst;
- struct dw_edma_burst burst[] __counted_by(nburst);
};
struct dw_edma_desc {
@@ -66,6 +59,12 @@ struct dw_edma_desc {
u32 alloc_sz;
u32 xfer_sz;
+
+ u32 done_burst;
+ u32 start_burst;
+ u8 cb;
+ u32 nburst;
+ struct dw_edma_burst burst[] __counted_by(nburst);
};
struct dw_edma_chan {
@@ -126,7 +125,6 @@ struct dw_edma_core_ops {
void (*ll_link)(struct dw_edma_chan *chan, u32 idx, bool cb, u64 addr);
void (*ch_doorbell)(struct dw_edma_chan *chan);
void (*ch_enable)(struct dw_edma_chan *chan);
-
void (*ch_config)(struct dw_edma_chan *chan);
void (*debugfs_on)(struct dw_edma *dw);
};
@@ -166,6 +164,14 @@ struct dw_edma_chan *dchan2dw_edma_chan(struct dma_chan *dchan)
return vc2dw_edma_chan(to_virt_chan(dchan));
}
+static inline u64 dw_edma_core_get_ll_paddr(struct dw_edma_chan *chan)
+{
+ if (chan->dir == EDMA_DIR_WRITE)
+ return chan->dw->chip->ll_region_wr[chan->id].paddr;
+
+ return chan->dw->chip->ll_region_rd[chan->id].paddr;
+}
+
static inline
void dw_edma_core_off(struct dw_edma *dw)
{
--
2.34.1
On Fri, Dec 12, 2025 at 05:24:50PM -0500, Frank Li wrote:
> The current descriptor layout is:
>
> struct dw_edma_desc *desc
> └─ chunk list
> └─ burst[]
>
> Creating a DMA descriptor requires at least two kzalloc() calls because
> each chunk is allocated as a linked-list node. Since the number of bursts
> is already known when the descriptor is created, this linked-list layer is
> unnecessary.
>
> Move the burst array directly into struct dw_edma_desc and remove the
> struct dw_edma_chunk layer entirely.
>
> Use start_burst and done_burst to track the current bursts, which current
> are in the DMA link list.
>
> Signed-off-by: Frank Li <Frank.Li@nxp.com>
> ---
> drivers/dma/dw-edma/dw-edma-core.c | 130 ++++++++++++-------------------------
> drivers/dma/dw-edma/dw-edma-core.h | 24 ++++---
> 2 files changed, 57 insertions(+), 97 deletions(-)
>
[...]
> static struct dma_async_tx_descriptor *
> @@ -551,8 +499,14 @@ static void dw_hdma_set_callback_result(struct virt_dma_desc *vd,
> return;
>
> desc = vd2dw_edma_desc(vd);
> - if (desc)
> - residue = desc->alloc_sz - desc->xfer_sz;
> + residue = desc->alloc_sz;
Now you dereference desc without checking for NULL.
> +
> + if (desc) {
> + if (result == DMA_TRANS_NOERROR)
> + residue -= desc->burst[desc->start_burst - 1].xfer_sz;
> + else if (desc->done_burst)
> + residue -= desc->burst[desc->done_burst - 1].xfer_sz;
> + }
>
> res = &vd->tx_result;
> res->result = result;
> @@ -571,7 +525,7 @@ static void dw_edma_done_interrupt(struct dw_edma_chan *chan)
> switch (chan->request) {
> case EDMA_REQ_NONE:
> desc = vd2dw_edma_desc(vd);
> - if (!desc->chunks_alloc) {
> + if (desc->start_burst >= desc->nburst) {
> dw_hdma_set_callback_result(vd,
> DMA_TRANS_NOERROR);
> list_del(&vd->node);
> @@ -936,7 +890,7 @@ int dw_edma_probe(struct dw_edma_chip *chip)
> goto err_irq_free;
>
> /* Turn debugfs on */
> - dw_edma_core_debugfs_on(dw);
> + //dw_edma_core_debugfs_on(dw);
debug code?
>
> chip->dw = dw;
>
> diff --git a/drivers/dma/dw-edma/dw-edma-core.h b/drivers/dma/dw-edma/dw-edma-core.h
> index 1930c3bce2bf33fdfbf4e8d99002483a4565faed..ba83c42dee5224dccdf34cec6481e9404a607702 100644
> --- a/drivers/dma/dw-edma/dw-edma-core.h
> +++ b/drivers/dma/dw-edma/dw-edma-core.h
> @@ -46,15 +46,8 @@ struct dw_edma_burst {
> u64 sar;
> u64 dar;
> u32 sz;
> -};
> -
> -struct dw_edma_chunk {
> - struct list_head list;
> - struct dw_edma_chan *chan;
> - u8 cb;
> + /* precalulate summary of previous burst total size */
> u32 xfer_sz;
> - u32 nburst;
> - struct dw_edma_burst burst[] __counted_by(nburst);
> };
>
> struct dw_edma_desc {
> @@ -66,6 +59,12 @@ struct dw_edma_desc {
>
> u32 alloc_sz;
> u32 xfer_sz;
> +
> + u32 done_burst;
> + u32 start_burst;
> + u8 cb;
> + u32 nburst;
> + struct dw_edma_burst burst[] __counted_by(nburst);
> };
>
> struct dw_edma_chan {
> @@ -126,7 +125,6 @@ struct dw_edma_core_ops {
> void (*ll_link)(struct dw_edma_chan *chan, u32 idx, bool cb, u64 addr);
> void (*ch_doorbell)(struct dw_edma_chan *chan);
> void (*ch_enable)(struct dw_edma_chan *chan);
> -
> void (*ch_config)(struct dw_edma_chan *chan);
> void (*debugfs_on)(struct dw_edma *dw);
> };
> @@ -166,6 +164,14 @@ struct dw_edma_chan *dchan2dw_edma_chan(struct dma_chan *dchan)
> return vc2dw_edma_chan(to_virt_chan(dchan));
> }
>
> +static inline u64 dw_edma_core_get_ll_paddr(struct dw_edma_chan *chan)
No need of inline.
- Mani
--
மணிவண்ணன் சதாசிவம்
On Tue, Dec 30, 2025 at 10:31:28PM +0530, Manivannan Sadhasivam wrote:
> On Fri, Dec 12, 2025 at 05:24:50PM -0500, Frank Li wrote:
> > The current descriptor layout is:
> >
> > struct dw_edma_desc *desc
> > └─ chunk list
> > └─ burst[]
> >
> > Creating a DMA descriptor requires at least two kzalloc() calls because
> > each chunk is allocated as a linked-list node. Since the number of bursts
> > is already known when the descriptor is created, this linked-list layer is
> > unnecessary.
> >
> > Move the burst array directly into struct dw_edma_desc and remove the
> > struct dw_edma_chunk layer entirely.
> >
> > Use start_burst and done_burst to track the current bursts, which current
> > are in the DMA link list.
> >
> > Signed-off-by: Frank Li <Frank.Li@nxp.com>
> > ---
> > drivers/dma/dw-edma/dw-edma-core.c | 130 ++++++++++++-------------------------
> > drivers/dma/dw-edma/dw-edma-core.h | 24 ++++---
> > 2 files changed, 57 insertions(+), 97 deletions(-)
> >
>
> [...]
>
> > static struct dma_async_tx_descriptor *
> > @@ -551,8 +499,14 @@ static void dw_hdma_set_callback_result(struct virt_dma_desc *vd,
> > return;
> >
> > desc = vd2dw_edma_desc(vd);
> > - if (desc)
> > - residue = desc->alloc_sz - desc->xfer_sz;
> > + residue = desc->alloc_sz;
>
> Now you dereference desc without checking for NULL.
It is impossible that desc is NULL if vd is not NULL.
static inline
struct dw_edma_desc *vd2dw_edma_desc(struct virt_dma_desc *vd)
{
return container_of(vd, struct dw_edma_desc, vd);
}
Previous check is reduntant.
Frank
>
> > +
> > + if (desc) {
> > + if (result == DMA_TRANS_NOERROR)
> > + residue -= desc->burst[desc->start_burst - 1].xfer_sz;
> > + else if (desc->done_burst)
> > + residue -= desc->burst[desc->done_burst - 1].xfer_sz;
> > + }
> >
> > res = &vd->tx_result;
> > res->result = result;
> > @@ -571,7 +525,7 @@ static void dw_edma_done_interrupt(struct dw_edma_chan *chan)
> > switch (chan->request) {
> > case EDMA_REQ_NONE:
> > desc = vd2dw_edma_desc(vd);
> > - if (!desc->chunks_alloc) {
> > + if (desc->start_burst >= desc->nburst) {
> > dw_hdma_set_callback_result(vd,
> > DMA_TRANS_NOERROR);
> > list_del(&vd->node);
> > @@ -936,7 +890,7 @@ int dw_edma_probe(struct dw_edma_chip *chip)
> > goto err_irq_free;
> >
> > /* Turn debugfs on */
> > - dw_edma_core_debugfs_on(dw);
> > + //dw_edma_core_debugfs_on(dw);
>
> debug code?
Sorry, forget remove it.
Frank
>
> >
> > chip->dw = dw;
> >
> > diff --git a/drivers/dma/dw-edma/dw-edma-core.h b/drivers/dma/dw-edma/dw-edma-core.h
> > index 1930c3bce2bf33fdfbf4e8d99002483a4565faed..ba83c42dee5224dccdf34cec6481e9404a607702 100644
> > --- a/drivers/dma/dw-edma/dw-edma-core.h
> > +++ b/drivers/dma/dw-edma/dw-edma-core.h
> > @@ -46,15 +46,8 @@ struct dw_edma_burst {
> > u64 sar;
> > u64 dar;
> > u32 sz;
> > -};
> > -
> > -struct dw_edma_chunk {
> > - struct list_head list;
> > - struct dw_edma_chan *chan;
> > - u8 cb;
> > + /* precalulate summary of previous burst total size */
> > u32 xfer_sz;
> > - u32 nburst;
> > - struct dw_edma_burst burst[] __counted_by(nburst);
> > };
> >
> > struct dw_edma_desc {
> > @@ -66,6 +59,12 @@ struct dw_edma_desc {
> >
> > u32 alloc_sz;
> > u32 xfer_sz;
> > +
> > + u32 done_burst;
> > + u32 start_burst;
> > + u8 cb;
> > + u32 nburst;
> > + struct dw_edma_burst burst[] __counted_by(nburst);
> > };
> >
> > struct dw_edma_chan {
> > @@ -126,7 +125,6 @@ struct dw_edma_core_ops {
> > void (*ll_link)(struct dw_edma_chan *chan, u32 idx, bool cb, u64 addr);
> > void (*ch_doorbell)(struct dw_edma_chan *chan);
> > void (*ch_enable)(struct dw_edma_chan *chan);
> > -
> > void (*ch_config)(struct dw_edma_chan *chan);
> > void (*debugfs_on)(struct dw_edma *dw);
> > };
> > @@ -166,6 +164,14 @@ struct dw_edma_chan *dchan2dw_edma_chan(struct dma_chan *dchan)
> > return vc2dw_edma_chan(to_virt_chan(dchan));
> > }
> >
> > +static inline u64 dw_edma_core_get_ll_paddr(struct dw_edma_chan *chan)
>
> No need of inline.
>
> - Mani
>
> --
> மணிவண்ணன் சதாசிவம்
© 2016 - 2026 Red Hat, Inc.