drivers/gpu/drm/etnaviv/etnaviv_mmu.c | 6 +++++- 1 file changed, 5 insertions(+), 1 deletion(-)
Etnaviv assumes that GPU page size is 4KiB, yet on some systems, the CPU
page size is 16 KiB. The size of etnaviv buffer objects will be aligned
to CPU page size on kernel side, however, userspace still assumes the
page size is 4KiB and doing allocation with 4KiB page as unit. This
results in softpin(userspace managed per-process address spaces) fails.
Because kernel side BO takes up bigger address space than user space
assumes whenever the size of a BO is not CPU page size aligned.
Insert an error message to help debug when such an issue happen.
Signed-off-by: Sui Jingfeng <sui.jingfeng@linux.dev>
---
For example, when running glmark2-drm:
[kernel space debug log]
etnaviv 0000:03:00.0: Insert bo failed, va: fd38b000, size: 4000
etnaviv 0000:03:00.0: Insert bo failed, va: fd38a000, size: 4000
[user space debug log]
bo->va = 0xfd48c000, bo->size=100000
bo->va = 0xfd38c000, bo->size=100000
bo->va = 0xfd38b000, bo->size=1000 <-- Insert IOVA fails started at here.
bo->va = 0xfd38a000, bo->size=1000
bo->va = 0xfd389000, bo->size=1000
[texture] texture-filter=nearest:MESA: error: etna_cmd_stream_flush:238: submit failed: -28 (No space left on device)
---
drivers/gpu/drm/etnaviv/etnaviv_mmu.c | 6 +++++-
1 file changed, 5 insertions(+), 1 deletion(-)
diff --git a/drivers/gpu/drm/etnaviv/etnaviv_mmu.c b/drivers/gpu/drm/etnaviv/etnaviv_mmu.c
index 1661d589bf3e..682f27b27d59 100644
--- a/drivers/gpu/drm/etnaviv/etnaviv_mmu.c
+++ b/drivers/gpu/drm/etnaviv/etnaviv_mmu.c
@@ -310,8 +310,12 @@ int etnaviv_iommu_map_gem(struct etnaviv_iommu_context *context,
else
ret = etnaviv_iommu_find_iova(context, node,
etnaviv_obj->base.size);
- if (ret < 0)
+ if (ret < 0) {
+ dev_err(context->global->dev,
+ "Insert iova failed, va: %llx, size: %zx\n",
+ va, etnaviv_obj->base.size);
goto unlock;
+ }
mapping->iova = node->start;
ret = etnaviv_iommu_map(context, node->start, sgt,
--
2.34.1
Hi Sui, Am Dienstag, dem 01.10.2024 um 06:17 +0800 schrieb Sui Jingfeng: > Etnaviv assumes that GPU page size is 4KiB, yet on some systems, the CPU > page size is 16 KiB. The size of etnaviv buffer objects will be aligned > to CPU page size on kernel side, however, userspace still assumes the > page size is 4KiB and doing allocation with 4KiB page as unit. This > results in softpin(userspace managed per-process address spaces) fails. > Because kernel side BO takes up bigger address space than user space > assumes whenever the size of a BO is not CPU page size aligned. > Seems we need to track the GPU and CPU allocation sizes separately. Userspace is correct in assuming that the GPU page size is 4K and buffers are aligned to this granule. There should be no need to waste GPU VA space just because the CPU page size is larger than that and we need to overallocate buffers to suit the CPU. > Insert an error message to help debug when such an issue happen. > > Signed-off-by: Sui Jingfeng <sui.jingfeng@linux.dev> > --- > For example, when running glmark2-drm: > > [kernel space debug log] > > etnaviv 0000:03:00.0: Insert bo failed, va: fd38b000, size: 4000 > etnaviv 0000:03:00.0: Insert bo failed, va: fd38a000, size: 4000 > > [user space debug log] > > bo->va = 0xfd48c000, bo->size=100000 > bo->va = 0xfd38c000, bo->size=100000 > bo->va = 0xfd38b000, bo->size=1000 <-- Insert IOVA fails started at here. > bo->va = 0xfd38a000, bo->size=1000 > bo->va = 0xfd389000, bo->size=1000 > > [texture] texture-filter=nearest:MESA: error: etna_cmd_stream_flush:238: submit failed: -28 (No space left on device) > --- > drivers/gpu/drm/etnaviv/etnaviv_mmu.c | 6 +++++- > 1 file changed, 5 insertions(+), 1 deletion(-) > > diff --git a/drivers/gpu/drm/etnaviv/etnaviv_mmu.c b/drivers/gpu/drm/etnaviv/etnaviv_mmu.c > index 1661d589bf3e..682f27b27d59 100644 > --- a/drivers/gpu/drm/etnaviv/etnaviv_mmu.c > +++ b/drivers/gpu/drm/etnaviv/etnaviv_mmu.c > @@ -310,8 +310,12 @@ int etnaviv_iommu_map_gem(struct etnaviv_iommu_context *context, > else > ret = etnaviv_iommu_find_iova(context, node, > etnaviv_obj->base.size); > - if (ret < 0) > + if (ret < 0) { > + dev_err(context->global->dev, > + "Insert iova failed, va: %llx, size: %zx\n", > + va, etnaviv_obj->base.size); As this might happen for a lot of buffers in a single submit and userspace might be unimpressed by the submit failure and keep pushing new submits, this has a potential to spam the logs. Please use dev_err_ratelimited. Other than that, this patch looks good. Regards, Lucas > goto unlock; > + } > > mapping->iova = node->start; > ret = etnaviv_iommu_map(context, node->start, sgt,
Hi, On 2024/10/1 16:27, Lucas Stach wrote: > Hi Sui, > > Am Dienstag, dem 01.10.2024 um 06:17 +0800 schrieb Sui Jingfeng: >> Etnaviv assumes that GPU page size is 4KiB, yet on some systems, the CPU >> page size is 16 KiB. The size of etnaviv buffer objects will be aligned >> to CPU page size on kernel side, however, userspace still assumes the >> page size is 4KiB and doing allocation with 4KiB page as unit. This >> results in softpin(userspace managed per-process address spaces) fails. >> Because kernel side BO takes up bigger address space than user space >> assumes whenever the size of a BO is not CPU page size aligned. >> > Seems we need to track the GPU and CPU allocation sizes separately. The idea is cool and fancy, I have been tried. By adding a 'user_size' member into the struct etnaviv_gem_object, and use this 'user_size'; to track the actual size that user-space thing of. (or in other words, the actual size that potential user allow to use) Using 'user_size' is pin, this partly solve VA address space collision under softpin fashion. This is partly works under my hasty test. But ... > Userspace is correct in assuming that the GPU page size is 4K and > buffers are aligned to this granule. Vivante GPU support 4KB and 64KB GPU page size. > There should be no need to waste GPU VA space We have nearly 4GBGPU VA space, As far as I can see it, we only use a few. So, is it true that we are wealthy about the VA space? > just because the CPU page size is larger than that and we > need to overallocate buffers to suit the CPU. A single CPU page share the same caching property, therefore, I image that asingle VA address range at least should occupy entire room of a single CPU page. Otherwise, it possible that 4 GPUVA share a single CPU page. if each GPUVA mapped with a different caching property from others. This get coherency requirements involved. >> Insert an error message to help debug when such an issue happen. >> >> Signed-off-by: Sui Jingfeng <sui.jingfeng@linux.dev> >> --- >> For example, when running glmark2-drm: >> >> [kernel space debug log] >> >> etnaviv 0000:03:00.0: Insert bo failed, va: fd38b000, size: 4000 >> etnaviv 0000:03:00.0: Insert bo failed, va: fd38a000, size: 4000 >> >> [user space debug log] >> >> bo->va = 0xfd48c000, bo->size=100000 >> bo->va = 0xfd38c000, bo->size=100000 >> bo->va = 0xfd38b000, bo->size=1000 <-- Insert IOVA fails started at here. >> bo->va = 0xfd38a000, bo->size=1000 >> bo->va = 0xfd389000, bo->size=1000 >> >> [texture] texture-filter=nearest:MESA: error: etna_cmd_stream_flush:238: submit failed: -28 (No space left on device) >> --- >> drivers/gpu/drm/etnaviv/etnaviv_mmu.c | 6 +++++- >> 1 file changed, 5 insertions(+), 1 deletion(-) >> >> diff --git a/drivers/gpu/drm/etnaviv/etnaviv_mmu.c b/drivers/gpu/drm/etnaviv/etnaviv_mmu.c >> index 1661d589bf3e..682f27b27d59 100644 >> --- a/drivers/gpu/drm/etnaviv/etnaviv_mmu.c >> +++ b/drivers/gpu/drm/etnaviv/etnaviv_mmu.c >> @@ -310,8 +310,12 @@ int etnaviv_iommu_map_gem(struct etnaviv_iommu_context *context, >> else >> ret = etnaviv_iommu_find_iova(context, node, >> etnaviv_obj->base.size); >> - if (ret < 0) >> + if (ret < 0) { >> + dev_err(context->global->dev, >> + "Insert iova failed, va: %llx, size: %zx\n", >> + va, etnaviv_obj->base.size); > As this might happen for a lot of buffers in a single submit and > userspace might be unimpressed by the submit failure and keep pushing > new submits, this has a potential to spam the logs. Please use > dev_err_ratelimited. Other than that, this patch looks good. > > Regards, > Lucas > >> goto unlock; >> + } >> >> mapping->iova = node->start; >> ret = etnaviv_iommu_map(context, node->start, sgt, -- Best regards, Sui
Am Mittwoch, dem 02.10.2024 um 03:26 +0800 schrieb Sui Jingfeng: > Hi, > > On 2024/10/1 16:27, Lucas Stach wrote: > > Hi Sui, > > > > Am Dienstag, dem 01.10.2024 um 06:17 +0800 schrieb Sui Jingfeng: > > > Etnaviv assumes that GPU page size is 4KiB, yet on some systems, the CPU > > > page size is 16 KiB. The size of etnaviv buffer objects will be aligned > > > to CPU page size on kernel side, however, userspace still assumes the > > > page size is 4KiB and doing allocation with 4KiB page as unit. This > > > results in softpin(userspace managed per-process address spaces) fails. > > > Because kernel side BO takes up bigger address space than user space > > > assumes whenever the size of a BO is not CPU page size aligned. > > > > > Seems we need to track the GPU and CPU allocation sizes separately. > > > The idea is cool and fancy, I have been tried. > > By adding a 'user_size' member into the struct etnaviv_gem_object, > and use this 'user_size'; to track the actual size that user-space > thing of. (or in other words, the actual size that potential user > allow to use) > > Using 'user_size' is pin, this partly solve VA address space collision > under softpin fashion. This is partly works under my hasty test. But ... > > > Userspace is correct in assuming that the GPU page size is 4K and > > buffers are aligned to this granule. > > > Vivante GPU support 4KB and 64KB GPU page size. > 64k is very impractical, as it can't really be mixed freely with 4k page size. If we ever going to support this, then it will be exposed via a userspace queryable param, so userspace will know when we need bigger alignment. > > > There should be no need to waste GPU VA space > > > We have nearly 4GBGPU VA space, As far as I can see it, we only use a few. So, is it true > that we are wealthy about the VA space? > Those GPUs are used within systems that support more physical memory than that. Especially the machines that are using larger page sizes on the CPU are likely to support more physical memory than 4GB. I don't see a reason why we should waste GPU VA space when we can avoid it. It's true that workloads you would run on a GC1000 are unlikely to use more than a fraction of the GPU VA space, but I can easily see more capable GPU cores with compute capabilities running up against the GPU VA space limit. > > > just because the CPU page size is larger than that and we > > need to overallocate buffers to suit the CPU. > > > A single CPU page share the same caching property, therefore, I image that > asingle VA address range at least should occupy entire room of a single CPU > page. > > Otherwise, it possible that 4 GPUVA share a single CPU page. > if each GPUVA mapped with a different caching property from others. > This get coherency requirements involved. > That won't happen. We still allocate the whole 16k page for a BO, so no other BO with different caching flags can share the same CPU page. We just don't map the whole page to the GPU side. Regards, Lucas > > > > Insert an error message to help debug when such an issue happen. > > > > > > Signed-off-by: Sui Jingfeng <sui.jingfeng@linux.dev> > > > --- > > > For example, when running glmark2-drm: > > > > > > [kernel space debug log] > > > > > > etnaviv 0000:03:00.0: Insert bo failed, va: fd38b000, size: 4000 > > > etnaviv 0000:03:00.0: Insert bo failed, va: fd38a000, size: 4000 > > > > > > [user space debug log] > > > > > > bo->va = 0xfd48c000, bo->size=100000 > > > bo->va = 0xfd38c000, bo->size=100000 > > > bo->va = 0xfd38b000, bo->size=1000 <-- Insert IOVA fails started at here. > > > bo->va = 0xfd38a000, bo->size=1000 > > > bo->va = 0xfd389000, bo->size=1000 > > > > > > [texture] texture-filter=nearest:MESA: error: etna_cmd_stream_flush:238: submit failed: -28 (No space left on device) > > > --- > > > drivers/gpu/drm/etnaviv/etnaviv_mmu.c | 6 +++++- > > > 1 file changed, 5 insertions(+), 1 deletion(-) > > > > > > diff --git a/drivers/gpu/drm/etnaviv/etnaviv_mmu.c b/drivers/gpu/drm/etnaviv/etnaviv_mmu.c > > > index 1661d589bf3e..682f27b27d59 100644 > > > --- a/drivers/gpu/drm/etnaviv/etnaviv_mmu.c > > > +++ b/drivers/gpu/drm/etnaviv/etnaviv_mmu.c > > > @@ -310,8 +310,12 @@ int etnaviv_iommu_map_gem(struct etnaviv_iommu_context *context, > > > else > > > ret = etnaviv_iommu_find_iova(context, node, > > > etnaviv_obj->base.size); > > > - if (ret < 0) > > > + if (ret < 0) { > > > + dev_err(context->global->dev, > > > + "Insert iova failed, va: %llx, size: %zx\n", > > > + va, etnaviv_obj->base.size); > > As this might happen for a lot of buffers in a single submit and > > userspace might be unimpressed by the submit failure and keep pushing > > new submits, this has a potential to spam the logs. Please use > > dev_err_ratelimited. Other than that, this patch looks good. > > > > Regards, > > Lucas > > > > > goto unlock; > > > + } > > > > > > mapping->iova = node->start; > > > ret = etnaviv_iommu_map(context, node->start, sgt, >
© 2016 - 2024 Red Hat, Inc.