Some reserved memory regions might have particular memory setup or
attributes that make them good candidates for heaps.
Let's provide a heap type that will create a new heap for each reserved
memory region flagged as such.
Signed-off-by: Maxime Ripard <mripard@kernel.org>
---
drivers/dma-buf/heaps/Kconfig | 8 +
drivers/dma-buf/heaps/Makefile | 1 +
drivers/dma-buf/heaps/carveout_heap.c | 362 ++++++++++++++++++++++++++++++++++
3 files changed, 371 insertions(+)
diff --git a/drivers/dma-buf/heaps/Kconfig b/drivers/dma-buf/heaps/Kconfig
index a5eef06c422644e8aadaf5aff2bd9a33c49c1ba3..1ce4f6828d8c06bfdd7bc2e5127707f1778586e6 100644
--- a/drivers/dma-buf/heaps/Kconfig
+++ b/drivers/dma-buf/heaps/Kconfig
@@ -1,5 +1,13 @@
+config DMABUF_HEAPS_CARVEOUT
+ bool "DMA-BUF Carveout Heaps"
+ depends on DMABUF_HEAPS
+ help
+ Choose this option to enable the carveout dmabuf heap. The carveout
+ heap is backed by pages from reserved memory regions flagged as
+ exportable. If in doubt, say Y.
+
config DMABUF_HEAPS_SYSTEM
bool "DMA-BUF System Heap"
depends on DMABUF_HEAPS
help
Choose this option to enable the system dmabuf heap. The system heap
diff --git a/drivers/dma-buf/heaps/Makefile b/drivers/dma-buf/heaps/Makefile
index 974467791032ffb8a7aba17b1407d9a19b3f3b44..b734647ad5c84f449106748160258e372f153df2 100644
--- a/drivers/dma-buf/heaps/Makefile
+++ b/drivers/dma-buf/heaps/Makefile
@@ -1,3 +1,4 @@
# SPDX-License-Identifier: GPL-2.0
+obj-$(CONFIG_DMABUF_HEAPS_CARVEOUT) += carveout_heap.o
obj-$(CONFIG_DMABUF_HEAPS_SYSTEM) += system_heap.o
obj-$(CONFIG_DMABUF_HEAPS_CMA) += cma_heap.o
diff --git a/drivers/dma-buf/heaps/carveout_heap.c b/drivers/dma-buf/heaps/carveout_heap.c
new file mode 100644
index 0000000000000000000000000000000000000000..c01abc72c09d4f2a373462fded2856d975a36c8f
--- /dev/null
+++ b/drivers/dma-buf/heaps/carveout_heap.c
@@ -0,0 +1,362 @@
+// SPDX-License-Identifier: GPL-2.0
+
+#include <linux/dma-buf.h>
+#include <linux/dma-heap.h>
+#include <linux/genalloc.h>
+#include <linux/highmem.h>
+#include <linux/of_reserved_mem.h>
+
+struct carveout_heap_priv {
+ struct dma_heap *heap;
+ struct gen_pool *pool;
+};
+
+struct carveout_heap_buffer_priv {
+ struct mutex lock;
+ struct list_head attachments;
+
+ unsigned long num_pages;
+ struct carveout_heap_priv *heap;
+ phys_addr_t paddr;
+ void *vaddr;
+ unsigned int vmap_cnt;
+};
+
+struct carveout_heap_attachment {
+ struct list_head head;
+ struct sg_table table;
+
+ struct device *dev;
+ bool mapped;
+};
+
+static int carveout_heap_attach(struct dma_buf *buf,
+ struct dma_buf_attachment *attachment)
+{
+ struct carveout_heap_buffer_priv *priv = buf->priv;
+ struct carveout_heap_attachment *a;
+ struct sg_table *sgt;
+ unsigned long len = priv->num_pages * PAGE_SIZE;
+ int ret;
+
+ a = kzalloc(sizeof(*a), GFP_KERNEL);
+ if (!a)
+ return -ENOMEM;
+ INIT_LIST_HEAD(&a->head);
+ a->dev = attachment->dev;
+ attachment->priv = a;
+
+ sgt = &a->table;
+ ret = sg_alloc_table(sgt, 1, GFP_KERNEL);
+ if (ret)
+ goto err_cleanup_attach;
+
+ sg_set_buf(sgt->sgl, priv->vaddr, len);
+
+ mutex_lock(&priv->lock);
+ list_add(&a->head, &priv->attachments);
+ mutex_unlock(&priv->lock);
+
+ return 0;
+
+err_cleanup_attach:
+ kfree(a);
+ return ret;
+}
+
+static void carveout_heap_detach(struct dma_buf *dmabuf,
+ struct dma_buf_attachment *attachment)
+{
+ struct carveout_heap_buffer_priv *priv = dmabuf->priv;
+ struct carveout_heap_attachment *a = attachment->priv;
+
+ mutex_lock(&priv->lock);
+ list_del(&a->head);
+ mutex_unlock(&priv->lock);
+
+ sg_free_table(&a->table);
+ kfree(a);
+}
+
+static struct sg_table *
+carveout_heap_map_dma_buf(struct dma_buf_attachment *attachment,
+ enum dma_data_direction direction)
+{
+ struct carveout_heap_attachment *a = attachment->priv;
+ struct sg_table *table = &a->table;
+ int ret;
+
+ ret = dma_map_sgtable(a->dev, table, direction, 0);
+ if (ret)
+ return ERR_PTR(ret);
+
+ a->mapped = true;
+
+ return table;
+}
+
+static void carveout_heap_unmap_dma_buf(struct dma_buf_attachment *attachment,
+ struct sg_table *table,
+ enum dma_data_direction direction)
+{
+ struct carveout_heap_attachment *a = attachment->priv;
+
+ a->mapped = false;
+ dma_unmap_sgtable(a->dev, table, direction, 0);
+}
+
+static int
+carveout_heap_dma_buf_begin_cpu_access(struct dma_buf *dmabuf,
+ enum dma_data_direction direction)
+{
+ struct carveout_heap_buffer_priv *priv = dmabuf->priv;
+ struct carveout_heap_attachment *a;
+ unsigned long len = priv->num_pages * PAGE_SIZE;
+
+ mutex_lock(&priv->lock);
+
+ if (priv->vmap_cnt)
+ invalidate_kernel_vmap_range(priv->vaddr, len);
+
+ list_for_each_entry(a, &priv->attachments, head) {
+ if (!a->mapped)
+ continue;
+
+ dma_sync_sgtable_for_cpu(a->dev, &a->table, direction);
+ }
+
+ mutex_unlock(&priv->lock);
+
+ return 0;
+}
+
+static int
+carveout_heap_dma_buf_end_cpu_access(struct dma_buf *dmabuf,
+ enum dma_data_direction direction)
+{
+ struct carveout_heap_buffer_priv *priv = dmabuf->priv;
+ struct carveout_heap_attachment *a;
+ unsigned long len = priv->num_pages * PAGE_SIZE;
+
+ mutex_lock(&priv->lock);
+
+ if (priv->vmap_cnt)
+ flush_kernel_vmap_range(priv->vaddr, len);
+
+ list_for_each_entry(a, &priv->attachments, head) {
+ if (!a->mapped)
+ continue;
+
+ dma_sync_sgtable_for_device(a->dev, &a->table, direction);
+ }
+
+ mutex_unlock(&priv->lock);
+
+ return 0;
+}
+
+static int carveout_heap_mmap(struct dma_buf *dmabuf,
+ struct vm_area_struct *vma)
+{
+ struct carveout_heap_buffer_priv *priv = dmabuf->priv;
+ unsigned long len = priv->num_pages * PAGE_SIZE;
+
+ return vm_iomap_memory(vma, priv->paddr, len);
+}
+
+static int carveout_heap_vmap(struct dma_buf *dmabuf, struct iosys_map *map)
+{
+ struct carveout_heap_buffer_priv *priv = dmabuf->priv;
+ unsigned long len = priv->num_pages * PAGE_SIZE;
+
+ mutex_lock(&priv->lock);
+
+ if (!priv->vmap_cnt) {
+ void *vaddr = memremap(priv->paddr, len, MEMREMAP_WB);
+
+ if (!vaddr) {
+ mutex_unlock(&priv->lock);
+ return -ENOMEM;
+ }
+
+ priv->vaddr = vaddr;
+ }
+
+ WARN_ON(!priv->vaddr);
+ iosys_map_set_vaddr(map, priv->vaddr);
+ priv->vmap_cnt++;
+
+ mutex_unlock(&priv->lock);
+
+ return 0;
+}
+
+static void carveout_heap_vunmap(struct dma_buf *dmabuf, struct iosys_map *map)
+{
+ struct carveout_heap_buffer_priv *priv = dmabuf->priv;
+
+ mutex_lock(&priv->lock);
+
+ priv->vmap_cnt--;
+ if (!priv->vmap_cnt) {
+ memunmap(priv->vaddr);
+ priv->vaddr = NULL;
+ }
+
+ mutex_unlock(&priv->lock);
+
+ iosys_map_clear(map);
+}
+
+static void carveout_heap_dma_buf_release(struct dma_buf *buf)
+{
+ struct carveout_heap_buffer_priv *buffer_priv = buf->priv;
+ struct carveout_heap_priv *heap_priv = buffer_priv->heap;
+ unsigned long len = buffer_priv->num_pages * PAGE_SIZE;
+
+ gen_pool_free(heap_priv->pool, buffer_priv->paddr, len);
+ kfree(buffer_priv);
+}
+
+static const struct dma_buf_ops carveout_heap_buf_ops = {
+ .attach = carveout_heap_attach,
+ .detach = carveout_heap_detach,
+ .map_dma_buf = carveout_heap_map_dma_buf,
+ .unmap_dma_buf = carveout_heap_unmap_dma_buf,
+ .begin_cpu_access = carveout_heap_dma_buf_begin_cpu_access,
+ .end_cpu_access = carveout_heap_dma_buf_end_cpu_access,
+ .mmap = carveout_heap_mmap,
+ .vmap = carveout_heap_vmap,
+ .vunmap = carveout_heap_vunmap,
+ .release = carveout_heap_dma_buf_release,
+};
+
+static struct dma_buf *carveout_heap_allocate(struct dma_heap *heap,
+ unsigned long len,
+ u32 fd_flags,
+ u64 heap_flags)
+{
+ struct carveout_heap_priv *heap_priv = dma_heap_get_drvdata(heap);
+ struct carveout_heap_buffer_priv *buffer_priv;
+ DEFINE_DMA_BUF_EXPORT_INFO(exp_info);
+ struct dma_buf *buf;
+ phys_addr_t paddr;
+ /* len is guaranteed to be page-aligned by the framework, so we can use it as is. */
+ size_t size = len;
+ int ret;
+
+ buffer_priv = kzalloc(sizeof(*buffer_priv), GFP_KERNEL);
+ if (!buffer_priv)
+ return ERR_PTR(-ENOMEM);
+
+ INIT_LIST_HEAD(&buffer_priv->attachments);
+ mutex_init(&buffer_priv->lock);
+
+ paddr = gen_pool_alloc(heap_priv->pool, size);
+ if (!paddr) {
+ ret = -ENOMEM;
+ goto err_free_buffer_priv;
+ }
+
+ buffer_priv->paddr = paddr;
+ buffer_priv->heap = heap_priv;
+ buffer_priv->num_pages = size >> PAGE_SHIFT;
+
+ /* create the dmabuf */
+ exp_info.exp_name = dma_heap_get_name(heap);
+ exp_info.ops = &carveout_heap_buf_ops;
+ exp_info.size = size;
+ exp_info.flags = fd_flags;
+ exp_info.priv = buffer_priv;
+
+ buf = dma_buf_export(&exp_info);
+ if (IS_ERR(buf)) {
+ ret = PTR_ERR(buf);
+ goto err_free_buffer;
+ }
+
+ return buf;
+
+err_free_buffer:
+ gen_pool_free(heap_priv->pool, paddr, len);
+err_free_buffer_priv:
+ kfree(buffer_priv);
+
+ return ERR_PTR(ret);
+}
+
+static const struct dma_heap_ops carveout_heap_ops = {
+ .allocate = carveout_heap_allocate,
+};
+
+static int __init carveout_heap_setup(struct device_node *node)
+{
+ struct dma_heap_export_info exp_info = {};
+ const struct reserved_mem *rmem;
+ struct carveout_heap_priv *priv;
+ struct dma_heap *heap;
+ struct gen_pool *pool;
+ int ret;
+
+ rmem = of_reserved_mem_lookup(node);
+ if (!rmem)
+ return -EINVAL;
+
+ priv = kzalloc(sizeof(*priv), GFP_KERNEL);
+ if (!priv)
+ return -ENOMEM;
+
+ pool = gen_pool_create(PAGE_SHIFT, NUMA_NO_NODE);
+ if (!pool) {
+ ret = -ENOMEM;
+ goto err_cleanup_heap;
+ }
+ priv->pool = pool;
+
+ ret = gen_pool_add(pool, rmem->base, rmem->size, NUMA_NO_NODE);
+ if (ret)
+ goto err_release_mem_region;
+
+ exp_info.name = node->full_name;
+ exp_info.ops = &carveout_heap_ops;
+ exp_info.priv = priv;
+
+ heap = dma_heap_add(&exp_info);
+ if (IS_ERR(heap)) {
+ ret = PTR_ERR(heap);
+ goto err_release_mem_region;
+ }
+ priv->heap = heap;
+
+ return 0;
+
+err_release_mem_region:
+ gen_pool_destroy(pool);
+err_cleanup_heap:
+ kfree(priv);
+ return ret;
+}
+
+static int __init carveout_heap_init(void)
+{
+ struct device_node *rmem_node;
+ struct device_node *node;
+ int ret;
+
+ rmem_node = of_find_node_by_path("/reserved-memory");
+ if (!rmem_node)
+ return 0;
+
+ for_each_child_of_node(rmem_node, node) {
+ if (!of_device_is_compatible(node, "carved-out"))
+ continue;
+
+ ret = carveout_heap_setup(node);
+ if (ret)
+ return ret;
+ }
+
+ return 0;
+}
+
+module_init(carveout_heap_init);
--
2.49.0
Hi Maxime,
On 6/17/2025 5:55 PM, Maxime Ripard wrote:
> +static void carveout_heap_dma_buf_release(struct dma_buf *buf)
> +{
> + struct carveout_heap_buffer_priv *buffer_priv = buf->priv;
> + struct carveout_heap_priv *heap_priv = buffer_priv->heap;
> + unsigned long len = buffer_priv->num_pages * PAGE_SIZE;
> +
> + gen_pool_free(heap_priv->pool, buffer_priv->paddr, len);
Just checking If clearing of the memory is missed before releasing it to
the free pool. If not, isn't it a leak of data when the heap is being
used by the multiple apps.
BTW, thanks for these patches.
> + kfree(buffer_priv);
> +}
-Charan
Hi,
On Fri, Aug 08, 2025 at 03:46:21PM +0530, Charan Teja Kalla wrote:
> On 6/17/2025 5:55 PM, Maxime Ripard wrote:
> > +static void carveout_heap_dma_buf_release(struct dma_buf *buf)
> > +{
> > + struct carveout_heap_buffer_priv *buffer_priv = buf->priv;
> > + struct carveout_heap_priv *heap_priv = buffer_priv->heap;
> > + unsigned long len = buffer_priv->num_pages * PAGE_SIZE;
> > +
> > + gen_pool_free(heap_priv->pool, buffer_priv->paddr, len);
>
> Just checking If clearing of the memory is missed before releasing it to
> the free pool. If not, isn't it a leak of data when the heap is being
> used by the multiple apps.
>
> BTW, thanks for these patches.
Thanks for the review. Note that we've since moved to another approach
here:
https://lore.kernel.org/r/20250721-dma-buf-ecc-heap-v7-0-031836e1a942@kernel.org
Maxime
On Tue, Jun 17, 2025 at 02:25:41PM +0200, Maxime Ripard wrote:
> Some reserved memory regions might have particular memory setup or
> attributes that make them good candidates for heaps.
>
> Let's provide a heap type that will create a new heap for each reserved
> memory region flagged as such.
>
> Signed-off-by: Maxime Ripard <mripard@kernel.org>
> ---
> drivers/dma-buf/heaps/Kconfig | 8 +
> drivers/dma-buf/heaps/Makefile | 1 +
> drivers/dma-buf/heaps/carveout_heap.c | 362 ++++++++++++++++++++++++++++++++++
> 3 files changed, 371 insertions(+)
>
> diff --git a/drivers/dma-buf/heaps/Kconfig b/drivers/dma-buf/heaps/Kconfig
> index a5eef06c422644e8aadaf5aff2bd9a33c49c1ba3..1ce4f6828d8c06bfdd7bc2e5127707f1778586e6 100644
> --- a/drivers/dma-buf/heaps/Kconfig
> +++ b/drivers/dma-buf/heaps/Kconfig
> @@ -1,5 +1,13 @@
> +config DMABUF_HEAPS_CARVEOUT
> + bool "DMA-BUF Carveout Heaps"
> + depends on DMABUF_HEAPS
> + help
> + Choose this option to enable the carveout dmabuf heap. The carveout
> + heap is backed by pages from reserved memory regions flagged as
> + exportable. If in doubt, say Y.
> +
> config DMABUF_HEAPS_SYSTEM
> bool "DMA-BUF System Heap"
> depends on DMABUF_HEAPS
> help
> Choose this option to enable the system dmabuf heap. The system heap
> diff --git a/drivers/dma-buf/heaps/Makefile b/drivers/dma-buf/heaps/Makefile
> index 974467791032ffb8a7aba17b1407d9a19b3f3b44..b734647ad5c84f449106748160258e372f153df2 100644
> --- a/drivers/dma-buf/heaps/Makefile
> +++ b/drivers/dma-buf/heaps/Makefile
> @@ -1,3 +1,4 @@
> # SPDX-License-Identifier: GPL-2.0
> +obj-$(CONFIG_DMABUF_HEAPS_CARVEOUT) += carveout_heap.o
> obj-$(CONFIG_DMABUF_HEAPS_SYSTEM) += system_heap.o
> obj-$(CONFIG_DMABUF_HEAPS_CMA) += cma_heap.o
> diff --git a/drivers/dma-buf/heaps/carveout_heap.c b/drivers/dma-buf/heaps/carveout_heap.c
> new file mode 100644
> index 0000000000000000000000000000000000000000..c01abc72c09d4f2a373462fded2856d975a36c8f
> --- /dev/null
> +++ b/drivers/dma-buf/heaps/carveout_heap.c
> @@ -0,0 +1,362 @@
> +// SPDX-License-Identifier: GPL-2.0
> +
> +#include <linux/dma-buf.h>
> +#include <linux/dma-heap.h>
> +#include <linux/genalloc.h>
> +#include <linux/highmem.h>
> +#include <linux/of_reserved_mem.h>
> +
> +struct carveout_heap_priv {
> + struct dma_heap *heap;
> + struct gen_pool *pool;
> +};
> +
> +struct carveout_heap_buffer_priv {
> + struct mutex lock;
> + struct list_head attachments;
> +
> + unsigned long num_pages;
> + struct carveout_heap_priv *heap;
> + phys_addr_t paddr;
> + void *vaddr;
> + unsigned int vmap_cnt;
> +};
> +
> +struct carveout_heap_attachment {
> + struct list_head head;
> + struct sg_table table;
> +
> + struct device *dev;
> + bool mapped;
> +};
> +
> +static int carveout_heap_attach(struct dma_buf *buf,
> + struct dma_buf_attachment *attachment)
> +{
> + struct carveout_heap_buffer_priv *priv = buf->priv;
> + struct carveout_heap_attachment *a;
> + struct sg_table *sgt;
> + unsigned long len = priv->num_pages * PAGE_SIZE;
> + int ret;
> +
> + a = kzalloc(sizeof(*a), GFP_KERNEL);
> + if (!a)
> + return -ENOMEM;
> + INIT_LIST_HEAD(&a->head);
> + a->dev = attachment->dev;
> + attachment->priv = a;
> +
> + sgt = &a->table;
> + ret = sg_alloc_table(sgt, 1, GFP_KERNEL);
> + if (ret)
> + goto err_cleanup_attach;
> +
> + sg_set_buf(sgt->sgl, priv->vaddr, len);
> +
> + mutex_lock(&priv->lock);
> + list_add(&a->head, &priv->attachments);
> + mutex_unlock(&priv->lock);
> +
> + return 0;
> +
> +err_cleanup_attach:
> + kfree(a);
> + return ret;
> +}
> +
> +static void carveout_heap_detach(struct dma_buf *dmabuf,
> + struct dma_buf_attachment *attachment)
> +{
> + struct carveout_heap_buffer_priv *priv = dmabuf->priv;
> + struct carveout_heap_attachment *a = attachment->priv;
> +
> + mutex_lock(&priv->lock);
> + list_del(&a->head);
> + mutex_unlock(&priv->lock);
> +
> + sg_free_table(&a->table);
> + kfree(a);
> +}
> +
> +static struct sg_table *
> +carveout_heap_map_dma_buf(struct dma_buf_attachment *attachment,
> + enum dma_data_direction direction)
> +{
> + struct carveout_heap_attachment *a = attachment->priv;
> + struct sg_table *table = &a->table;
> + int ret;
> +
> + ret = dma_map_sgtable(a->dev, table, direction, 0);
> + if (ret)
> + return ERR_PTR(ret);
> +
> + a->mapped = true;
> +
> + return table;
> +}
> +
> +static void carveout_heap_unmap_dma_buf(struct dma_buf_attachment *attachment,
> + struct sg_table *table,
> + enum dma_data_direction direction)
> +{
> + struct carveout_heap_attachment *a = attachment->priv;
> +
> + a->mapped = false;
> + dma_unmap_sgtable(a->dev, table, direction, 0);
> +}
> +
> +static int
> +carveout_heap_dma_buf_begin_cpu_access(struct dma_buf *dmabuf,
> + enum dma_data_direction direction)
> +{
> + struct carveout_heap_buffer_priv *priv = dmabuf->priv;
> + struct carveout_heap_attachment *a;
> + unsigned long len = priv->num_pages * PAGE_SIZE;
> +
> + mutex_lock(&priv->lock);
> +
> + if (priv->vmap_cnt)
> + invalidate_kernel_vmap_range(priv->vaddr, len);
> +
> + list_for_each_entry(a, &priv->attachments, head) {
> + if (!a->mapped)
> + continue;
> +
> + dma_sync_sgtable_for_cpu(a->dev, &a->table, direction);
> + }
> +
> + mutex_unlock(&priv->lock);
> +
> + return 0;
> +}
> +
> +static int
> +carveout_heap_dma_buf_end_cpu_access(struct dma_buf *dmabuf,
> + enum dma_data_direction direction)
> +{
> + struct carveout_heap_buffer_priv *priv = dmabuf->priv;
> + struct carveout_heap_attachment *a;
> + unsigned long len = priv->num_pages * PAGE_SIZE;
> +
> + mutex_lock(&priv->lock);
> +
> + if (priv->vmap_cnt)
> + flush_kernel_vmap_range(priv->vaddr, len);
> +
> + list_for_each_entry(a, &priv->attachments, head) {
> + if (!a->mapped)
> + continue;
> +
> + dma_sync_sgtable_for_device(a->dev, &a->table, direction);
> + }
> +
> + mutex_unlock(&priv->lock);
> +
> + return 0;
> +}
> +
> +static int carveout_heap_mmap(struct dma_buf *dmabuf,
> + struct vm_area_struct *vma)
> +{
> + struct carveout_heap_buffer_priv *priv = dmabuf->priv;
> + unsigned long len = priv->num_pages * PAGE_SIZE;
> +
> + return vm_iomap_memory(vma, priv->paddr, len);
> +}
> +
> +static int carveout_heap_vmap(struct dma_buf *dmabuf, struct iosys_map *map)
> +{
> + struct carveout_heap_buffer_priv *priv = dmabuf->priv;
> + unsigned long len = priv->num_pages * PAGE_SIZE;
> +
> + mutex_lock(&priv->lock);
> +
> + if (!priv->vmap_cnt) {
> + void *vaddr = memremap(priv->paddr, len, MEMREMAP_WB);
> +
> + if (!vaddr) {
> + mutex_unlock(&priv->lock);
> + return -ENOMEM;
> + }
> +
> + priv->vaddr = vaddr;
> + }
> +
> + WARN_ON(!priv->vaddr);
> + iosys_map_set_vaddr(map, priv->vaddr);
> + priv->vmap_cnt++;
> +
> + mutex_unlock(&priv->lock);
> +
> + return 0;
> +}
> +
> +static void carveout_heap_vunmap(struct dma_buf *dmabuf, struct iosys_map *map)
> +{
> + struct carveout_heap_buffer_priv *priv = dmabuf->priv;
> +
> + mutex_lock(&priv->lock);
> +
> + priv->vmap_cnt--;
> + if (!priv->vmap_cnt) {
> + memunmap(priv->vaddr);
> + priv->vaddr = NULL;
> + }
> +
> + mutex_unlock(&priv->lock);
> +
> + iosys_map_clear(map);
> +}
> +
> +static void carveout_heap_dma_buf_release(struct dma_buf *buf)
> +{
> + struct carveout_heap_buffer_priv *buffer_priv = buf->priv;
> + struct carveout_heap_priv *heap_priv = buffer_priv->heap;
> + unsigned long len = buffer_priv->num_pages * PAGE_SIZE;
> +
> + gen_pool_free(heap_priv->pool, buffer_priv->paddr, len);
> + kfree(buffer_priv);
> +}
> +
> +static const struct dma_buf_ops carveout_heap_buf_ops = {
> + .attach = carveout_heap_attach,
> + .detach = carveout_heap_detach,
> + .map_dma_buf = carveout_heap_map_dma_buf,
> + .unmap_dma_buf = carveout_heap_unmap_dma_buf,
> + .begin_cpu_access = carveout_heap_dma_buf_begin_cpu_access,
> + .end_cpu_access = carveout_heap_dma_buf_end_cpu_access,
> + .mmap = carveout_heap_mmap,
> + .vmap = carveout_heap_vmap,
> + .vunmap = carveout_heap_vunmap,
> + .release = carveout_heap_dma_buf_release,
> +};
> +
> +static struct dma_buf *carveout_heap_allocate(struct dma_heap *heap,
> + unsigned long len,
> + u32 fd_flags,
> + u64 heap_flags)
> +{
> + struct carveout_heap_priv *heap_priv = dma_heap_get_drvdata(heap);
> + struct carveout_heap_buffer_priv *buffer_priv;
> + DEFINE_DMA_BUF_EXPORT_INFO(exp_info);
> + struct dma_buf *buf;
> + phys_addr_t paddr;
> + /* len is guaranteed to be page-aligned by the framework, so we can use it as is. */
> + size_t size = len;
> + int ret;
> +
> + buffer_priv = kzalloc(sizeof(*buffer_priv), GFP_KERNEL);
> + if (!buffer_priv)
> + return ERR_PTR(-ENOMEM);
> +
> + INIT_LIST_HEAD(&buffer_priv->attachments);
> + mutex_init(&buffer_priv->lock);
> +
> + paddr = gen_pool_alloc(heap_priv->pool, size);
> + if (!paddr) {
> + ret = -ENOMEM;
> + goto err_free_buffer_priv;
> + }
> +
> + buffer_priv->paddr = paddr;
> + buffer_priv->heap = heap_priv;
> + buffer_priv->num_pages = size >> PAGE_SHIFT;
> +
> + /* create the dmabuf */
> + exp_info.exp_name = dma_heap_get_name(heap);
> + exp_info.ops = &carveout_heap_buf_ops;
> + exp_info.size = size;
> + exp_info.flags = fd_flags;
> + exp_info.priv = buffer_priv;
> +
> + buf = dma_buf_export(&exp_info);
> + if (IS_ERR(buf)) {
> + ret = PTR_ERR(buf);
> + goto err_free_buffer;
> + }
> +
> + return buf;
> +
> +err_free_buffer:
> + gen_pool_free(heap_priv->pool, paddr, len);
> +err_free_buffer_priv:
> + kfree(buffer_priv);
> +
> + return ERR_PTR(ret);
> +}
> +
> +static const struct dma_heap_ops carveout_heap_ops = {
> + .allocate = carveout_heap_allocate,
> +};
> +
> +static int __init carveout_heap_setup(struct device_node *node)
> +{
> + struct dma_heap_export_info exp_info = {};
> + const struct reserved_mem *rmem;
> + struct carveout_heap_priv *priv;
> + struct dma_heap *heap;
> + struct gen_pool *pool;
> + int ret;
> +
> + rmem = of_reserved_mem_lookup(node);
> + if (!rmem)
> + return -EINVAL;
> +
> + priv = kzalloc(sizeof(*priv), GFP_KERNEL);
> + if (!priv)
> + return -ENOMEM;
> +
> + pool = gen_pool_create(PAGE_SHIFT, NUMA_NO_NODE);
> + if (!pool) {
> + ret = -ENOMEM;
> + goto err_cleanup_heap;
> + }
> + priv->pool = pool;
> +
> + ret = gen_pool_add(pool, rmem->base, rmem->size, NUMA_NO_NODE);
> + if (ret)
> + goto err_release_mem_region;
> +
> + exp_info.name = node->full_name;
> + exp_info.ops = &carveout_heap_ops;
> + exp_info.priv = priv;
> +
> + heap = dma_heap_add(&exp_info);
> + if (IS_ERR(heap)) {
> + ret = PTR_ERR(heap);
> + goto err_release_mem_region;
> + }
> + priv->heap = heap;
> +
> + return 0;
> +
> +err_release_mem_region:
> + gen_pool_destroy(pool);
> +err_cleanup_heap:
> + kfree(priv);
> + return ret;
> +}
> +
> +static int __init carveout_heap_init(void)
> +{
> + struct device_node *rmem_node;
> + struct device_node *node;
> + int ret;
> +
> + rmem_node = of_find_node_by_path("/reserved-memory");
> + if (!rmem_node)
> + return 0;
> +
> + for_each_child_of_node(rmem_node, node) {
> + if (!of_device_is_compatible(node, "carved-out"))
> + continue;
> +
> + ret = carveout_heap_setup(node);
> + if (ret)
> + return ret;
> + }
/reserved-memory nodes get a platform_device, so why not make this a
driver?
Rob
On Fri, Jun 27, 2025 at 02:23:47PM -0500, Rob Herring wrote:
> > +static int __init carveout_heap_init(void)
> > +{
> > + struct device_node *rmem_node;
> > + struct device_node *node;
> > + int ret;
> > +
> > + rmem_node = of_find_node_by_path("/reserved-memory");
> > + if (!rmem_node)
> > + return 0;
> > +
> > + for_each_child_of_node(rmem_node, node) {
> > + if (!of_device_is_compatible(node, "carved-out"))
> > + continue;
> > +
> > + ret = carveout_heap_setup(node);
> > + if (ret)
> > + return ret;
> > + }
>
> /reserved-memory nodes get a platform_device, so why not make this a
> driver?
Because I never realised we could :)
Thanks, I'll fix it for the next version
Maxime
© 2016 - 2026 Red Hat, Inc.