From: Kent Overstreet <kent.overstreet@linux.dev>
This wrapps all external vmalloc allocation functions with the
alloc_hooks() wrapper, and switches internal allocations to _noprof
variants where appropriate, for the new memory allocation profiling
feature.
Signed-off-by: Kent Overstreet <kent.overstreet@linux.dev>
Signed-off-by: Suren Baghdasaryan <surenb@google.com>
---
drivers/staging/media/atomisp/pci/hmm/hmm.c | 2 +-
include/linux/vmalloc.h | 60 ++++++++++----
kernel/kallsyms_selftest.c | 2 +-
mm/nommu.c | 64 +++++++--------
mm/util.c | 24 +++---
mm/vmalloc.c | 88 ++++++++++-----------
6 files changed, 135 insertions(+), 105 deletions(-)
diff --git a/drivers/staging/media/atomisp/pci/hmm/hmm.c b/drivers/staging/media/atomisp/pci/hmm/hmm.c
index bb12644fd033..3e2899ad8517 100644
--- a/drivers/staging/media/atomisp/pci/hmm/hmm.c
+++ b/drivers/staging/media/atomisp/pci/hmm/hmm.c
@@ -205,7 +205,7 @@ static ia_css_ptr __hmm_alloc(size_t bytes, enum hmm_bo_type type,
}
dev_dbg(atomisp_dev, "pages: 0x%08x (%zu bytes), type: %d, vmalloc %p\n",
- bo->start, bytes, type, vmalloc);
+ bo->start, bytes, type, vmalloc_noprof);
return bo->start;
diff --git a/include/linux/vmalloc.h b/include/linux/vmalloc.h
index 98ea90e90439..e4a631ec430b 100644
--- a/include/linux/vmalloc.h
+++ b/include/linux/vmalloc.h
@@ -2,6 +2,8 @@
#ifndef _LINUX_VMALLOC_H
#define _LINUX_VMALLOC_H
+#include <linux/alloc_tag.h>
+#include <linux/sched.h>
#include <linux/spinlock.h>
#include <linux/init.h>
#include <linux/list.h>
@@ -138,26 +140,54 @@ extern unsigned long vmalloc_nr_pages(void);
static inline unsigned long vmalloc_nr_pages(void) { return 0; }
#endif
-extern void *vmalloc(unsigned long size) __alloc_size(1);
-extern void *vzalloc(unsigned long size) __alloc_size(1);
-extern void *vmalloc_user(unsigned long size) __alloc_size(1);
-extern void *vmalloc_node(unsigned long size, int node) __alloc_size(1);
-extern void *vzalloc_node(unsigned long size, int node) __alloc_size(1);
-extern void *vmalloc_32(unsigned long size) __alloc_size(1);
-extern void *vmalloc_32_user(unsigned long size) __alloc_size(1);
-extern void *__vmalloc(unsigned long size, gfp_t gfp_mask) __alloc_size(1);
-extern void *__vmalloc_node_range(unsigned long size, unsigned long align,
+extern void *vmalloc_noprof(unsigned long size) __alloc_size(1);
+#define vmalloc(...) alloc_hooks(vmalloc_noprof(__VA_ARGS__))
+
+extern void *vzalloc_noprof(unsigned long size) __alloc_size(1);
+#define vzalloc(...) alloc_hooks(vzalloc_noprof(__VA_ARGS__))
+
+extern void *vmalloc_user_noprof(unsigned long size) __alloc_size(1);
+#define vmalloc_user(...) alloc_hooks(vmalloc_user_noprof(__VA_ARGS__))
+
+extern void *vmalloc_node_noprof(unsigned long size, int node) __alloc_size(1);
+#define vmalloc_node(...) alloc_hooks(vmalloc_node_noprof(__VA_ARGS__))
+
+extern void *vzalloc_node_noprof(unsigned long size, int node) __alloc_size(1);
+#define vzalloc_node(...) alloc_hooks(vzalloc_node_noprof(__VA_ARGS__))
+
+extern void *vmalloc_32_noprof(unsigned long size) __alloc_size(1);
+#define vmalloc_32(...) alloc_hooks(vmalloc_32_noprof(__VA_ARGS__))
+
+extern void *vmalloc_32_user_noprof(unsigned long size) __alloc_size(1);
+#define vmalloc_32_user(...) alloc_hooks(vmalloc_32_user_noprof(__VA_ARGS__))
+
+extern void *__vmalloc_noprof(unsigned long size, gfp_t gfp_mask) __alloc_size(1);
+#define __vmalloc(...) alloc_hooks(__vmalloc_noprof(__VA_ARGS__))
+
+extern void *__vmalloc_node_range_noprof(unsigned long size, unsigned long align,
unsigned long start, unsigned long end, gfp_t gfp_mask,
pgprot_t prot, unsigned long vm_flags, int node,
const void *caller) __alloc_size(1);
-void *__vmalloc_node(unsigned long size, unsigned long align, gfp_t gfp_mask,
+#define __vmalloc_node_range(...) alloc_hooks(__vmalloc_node_range_noprof(__VA_ARGS__))
+
+void *__vmalloc_node_noprof(unsigned long size, unsigned long align, gfp_t gfp_mask,
int node, const void *caller) __alloc_size(1);
-void *vmalloc_huge(unsigned long size, gfp_t gfp_mask) __alloc_size(1);
+#define __vmalloc_node(...) alloc_hooks(__vmalloc_node_noprof(__VA_ARGS__))
+
+void *vmalloc_huge_noprof(unsigned long size, gfp_t gfp_mask) __alloc_size(1);
+#define vmalloc_huge(...) alloc_hooks(vmalloc_huge_noprof(__VA_ARGS__))
+
+extern void *__vmalloc_array_noprof(size_t n, size_t size, gfp_t flags) __alloc_size(1, 2);
+#define __vmalloc_array(...) alloc_hooks(__vmalloc_array_noprof(__VA_ARGS__))
+
+extern void *vmalloc_array_noprof(size_t n, size_t size) __alloc_size(1, 2);
+#define vmalloc_array(...) alloc_hooks(vmalloc_array_noprof(__VA_ARGS__))
+
+extern void *__vcalloc_noprof(size_t n, size_t size, gfp_t flags) __alloc_size(1, 2);
+#define __vcalloc(...) alloc_hooks(__vcalloc_noprof(__VA_ARGS__))
-extern void *__vmalloc_array(size_t n, size_t size, gfp_t flags) __alloc_size(1, 2);
-extern void *vmalloc_array(size_t n, size_t size) __alloc_size(1, 2);
-extern void *__vcalloc(size_t n, size_t size, gfp_t flags) __alloc_size(1, 2);
-extern void *vcalloc(size_t n, size_t size) __alloc_size(1, 2);
+extern void *vcalloc_noprof(size_t n, size_t size) __alloc_size(1, 2);
+#define vcalloc(...) alloc_hooks(vcalloc_noprof(__VA_ARGS__))
extern void vfree(const void *addr);
extern void vfree_atomic(const void *addr);
diff --git a/kernel/kallsyms_selftest.c b/kernel/kallsyms_selftest.c
index 8a689b4ff4f9..2f84896a7bcb 100644
--- a/kernel/kallsyms_selftest.c
+++ b/kernel/kallsyms_selftest.c
@@ -82,7 +82,7 @@ static struct test_item test_items[] = {
ITEM_FUNC(kallsyms_test_func_static),
ITEM_FUNC(kallsyms_test_func),
ITEM_FUNC(kallsyms_test_func_weak),
- ITEM_FUNC(vmalloc),
+ ITEM_FUNC(vmalloc_noprof),
ITEM_FUNC(vfree),
#ifdef CONFIG_KALLSYMS_ALL
ITEM_DATA(kallsyms_test_var_bss_static),
diff --git a/mm/nommu.c b/mm/nommu.c
index 5ec8f44e7ce9..69a6f3b4d156 100644
--- a/mm/nommu.c
+++ b/mm/nommu.c
@@ -137,28 +137,28 @@ void vfree(const void *addr)
}
EXPORT_SYMBOL(vfree);
-void *__vmalloc(unsigned long size, gfp_t gfp_mask)
+void *__vmalloc_noprof(unsigned long size, gfp_t gfp_mask)
{
/*
* You can't specify __GFP_HIGHMEM with kmalloc() since kmalloc()
* returns only a logical address.
*/
- return kmalloc(size, (gfp_mask | __GFP_COMP) & ~__GFP_HIGHMEM);
+ return kmalloc_noprof(size, (gfp_mask | __GFP_COMP) & ~__GFP_HIGHMEM);
}
-EXPORT_SYMBOL(__vmalloc);
+EXPORT_SYMBOL(__vmalloc_noprof);
-void *__vmalloc_node_range(unsigned long size, unsigned long align,
+void *__vmalloc_node_range_noprof(unsigned long size, unsigned long align,
unsigned long start, unsigned long end, gfp_t gfp_mask,
pgprot_t prot, unsigned long vm_flags, int node,
const void *caller)
{
- return __vmalloc(size, gfp_mask);
+ return __vmalloc_noprof(size, gfp_mask);
}
-void *__vmalloc_node(unsigned long size, unsigned long align, gfp_t gfp_mask,
+void *__vmalloc_node_noprof(unsigned long size, unsigned long align, gfp_t gfp_mask,
int node, const void *caller)
{
- return __vmalloc(size, gfp_mask);
+ return __vmalloc_noprof(size, gfp_mask);
}
static void *__vmalloc_user_flags(unsigned long size, gfp_t flags)
@@ -179,11 +179,11 @@ static void *__vmalloc_user_flags(unsigned long size, gfp_t flags)
return ret;
}
-void *vmalloc_user(unsigned long size)
+void *vmalloc_user_noprof(unsigned long size)
{
return __vmalloc_user_flags(size, GFP_KERNEL | __GFP_ZERO);
}
-EXPORT_SYMBOL(vmalloc_user);
+EXPORT_SYMBOL(vmalloc_user_noprof);
struct page *vmalloc_to_page(const void *addr)
{
@@ -217,13 +217,13 @@ long vread_iter(struct iov_iter *iter, const char *addr, size_t count)
* For tight control over page level allocator and protection flags
* use __vmalloc() instead.
*/
-void *vmalloc(unsigned long size)
+void *vmalloc_noprof(unsigned long size)
{
- return __vmalloc(size, GFP_KERNEL);
+ return __vmalloc_noprof(size, GFP_KERNEL);
}
-EXPORT_SYMBOL(vmalloc);
+EXPORT_SYMBOL(vmalloc_noprof);
-void *vmalloc_huge(unsigned long size, gfp_t gfp_mask) __weak __alias(__vmalloc);
+void *vmalloc_huge_noprof(unsigned long size, gfp_t gfp_mask) __weak __alias(__vmalloc_noprof);
/*
* vzalloc - allocate virtually contiguous memory with zero fill
@@ -237,14 +237,14 @@ void *vmalloc_huge(unsigned long size, gfp_t gfp_mask) __weak __alias(__vmalloc)
* For tight control over page level allocator and protection flags
* use __vmalloc() instead.
*/
-void *vzalloc(unsigned long size)
+void *vzalloc_noprof(unsigned long size)
{
- return __vmalloc(size, GFP_KERNEL | __GFP_ZERO);
+ return __vmalloc_noprof(size, GFP_KERNEL | __GFP_ZERO);
}
-EXPORT_SYMBOL(vzalloc);
+EXPORT_SYMBOL(vzalloc_noprof);
/**
- * vmalloc_node - allocate memory on a specific node
+ * vmalloc_node_noprof - allocate memory on a specific node
* @size: allocation size
* @node: numa node
*
@@ -254,14 +254,14 @@ EXPORT_SYMBOL(vzalloc);
* For tight control over page level allocator and protection flags
* use __vmalloc() instead.
*/
-void *vmalloc_node(unsigned long size, int node)
+void *vmalloc_node_noprof(unsigned long size, int node)
{
- return vmalloc(size);
+ return vmalloc_noprof(size);
}
-EXPORT_SYMBOL(vmalloc_node);
+EXPORT_SYMBOL(vmalloc_node_noprof);
/**
- * vzalloc_node - allocate memory on a specific node with zero fill
+ * vzalloc_node_noprof - allocate memory on a specific node with zero fill
* @size: allocation size
* @node: numa node
*
@@ -272,27 +272,27 @@ EXPORT_SYMBOL(vmalloc_node);
* For tight control over page level allocator and protection flags
* use __vmalloc() instead.
*/
-void *vzalloc_node(unsigned long size, int node)
+void *vzalloc_node_noprof(unsigned long size, int node)
{
- return vzalloc(size);
+ return vzalloc_noprof(size);
}
-EXPORT_SYMBOL(vzalloc_node);
+EXPORT_SYMBOL(vzalloc_node_noprof);
/**
- * vmalloc_32 - allocate virtually contiguous memory (32bit addressable)
+ * vmalloc_32_noprof - allocate virtually contiguous memory (32bit addressable)
* @size: allocation size
*
* Allocate enough 32bit PA addressable pages to cover @size from the
* page level allocator and map them into contiguous kernel virtual space.
*/
-void *vmalloc_32(unsigned long size)
+void *vmalloc_32_noprof(unsigned long size)
{
- return __vmalloc(size, GFP_KERNEL);
+ return __vmalloc_noprof(size, GFP_KERNEL);
}
-EXPORT_SYMBOL(vmalloc_32);
+EXPORT_SYMBOL(vmalloc_32_noprof);
/**
- * vmalloc_32_user - allocate zeroed virtually contiguous 32bit memory
+ * vmalloc_32_user_noprof - allocate zeroed virtually contiguous 32bit memory
* @size: allocation size
*
* The resulting memory area is 32bit addressable and zeroed so it can be
@@ -301,15 +301,15 @@ EXPORT_SYMBOL(vmalloc_32);
* VM_USERMAP is set on the corresponding VMA so that subsequent calls to
* remap_vmalloc_range() are permissible.
*/
-void *vmalloc_32_user(unsigned long size)
+void *vmalloc_32_user_noprof(unsigned long size)
{
/*
* We'll have to sort out the ZONE_DMA bits for 64-bit,
* but for now this can simply use vmalloc_user() directly.
*/
- return vmalloc_user(size);
+ return vmalloc_user_noprof(size);
}
-EXPORT_SYMBOL(vmalloc_32_user);
+EXPORT_SYMBOL(vmalloc_32_user_noprof);
void *vmap(struct page **pages, unsigned int count, unsigned long flags, pgprot_t prot)
{
diff --git a/mm/util.c b/mm/util.c
index a79dce7546f1..157b5edcba75 100644
--- a/mm/util.c
+++ b/mm/util.c
@@ -656,7 +656,7 @@ void *kvmalloc_node_noprof(size_t size, gfp_t flags, int node)
* about the resulting pointer, and cannot play
* protection games.
*/
- return __vmalloc_node_range(size, 1, VMALLOC_START, VMALLOC_END,
+ return __vmalloc_node_range_noprof(size, 1, VMALLOC_START, VMALLOC_END,
flags, PAGE_KERNEL, VM_ALLOW_HUGE_VMAP,
node, __builtin_return_address(0));
}
@@ -715,12 +715,12 @@ void *kvrealloc_noprof(const void *p, size_t oldsize, size_t newsize, gfp_t flag
EXPORT_SYMBOL(kvrealloc_noprof);
/**
- * __vmalloc_array - allocate memory for a virtually contiguous array.
+ * __vmalloc_array_noprof - allocate memory for a virtually contiguous array.
* @n: number of elements.
* @size: element size.
* @flags: the type of memory to allocate (see kmalloc).
*/
-void *__vmalloc_array(size_t n, size_t size, gfp_t flags)
+void *__vmalloc_array_noprof(size_t n, size_t size, gfp_t flags)
{
size_t bytes;
@@ -728,18 +728,18 @@ void *__vmalloc_array(size_t n, size_t size, gfp_t flags)
return NULL;
return __vmalloc(bytes, flags);
}
-EXPORT_SYMBOL(__vmalloc_array);
+EXPORT_SYMBOL(__vmalloc_array_noprof);
/**
- * vmalloc_array - allocate memory for a virtually contiguous array.
+ * vmalloc_array_noprof - allocate memory for a virtually contiguous array.
* @n: number of elements.
* @size: element size.
*/
-void *vmalloc_array(size_t n, size_t size)
+void *vmalloc_array_noprof(size_t n, size_t size)
{
return __vmalloc_array(n, size, GFP_KERNEL);
}
-EXPORT_SYMBOL(vmalloc_array);
+EXPORT_SYMBOL(vmalloc_array_noprof);
/**
* __vcalloc - allocate and zero memory for a virtually contiguous array.
@@ -747,22 +747,22 @@ EXPORT_SYMBOL(vmalloc_array);
* @size: element size.
* @flags: the type of memory to allocate (see kmalloc).
*/
-void *__vcalloc(size_t n, size_t size, gfp_t flags)
+void *__vcalloc_noprof(size_t n, size_t size, gfp_t flags)
{
return __vmalloc_array(n, size, flags | __GFP_ZERO);
}
-EXPORT_SYMBOL(__vcalloc);
+EXPORT_SYMBOL(__vcalloc_noprof);
/**
- * vcalloc - allocate and zero memory for a virtually contiguous array.
+ * vcalloc_noprof - allocate and zero memory for a virtually contiguous array.
* @n: number of elements.
* @size: element size.
*/
-void *vcalloc(size_t n, size_t size)
+void *vcalloc_noprof(size_t n, size_t size)
{
return __vmalloc_array(n, size, GFP_KERNEL | __GFP_ZERO);
}
-EXPORT_SYMBOL(vcalloc);
+EXPORT_SYMBOL(vcalloc_noprof);
struct anon_vma *folio_anon_vma(struct folio *folio)
{
diff --git a/mm/vmalloc.c b/mm/vmalloc.c
index 22aa63f4ef63..b2f2248d85a9 100644
--- a/mm/vmalloc.c
+++ b/mm/vmalloc.c
@@ -3507,12 +3507,12 @@ vm_area_alloc_pages(gfp_t gfp, int nid,
* but mempolicy wants to alloc memory by interleaving.
*/
if (IS_ENABLED(CONFIG_NUMA) && nid == NUMA_NO_NODE)
- nr = alloc_pages_bulk_array_mempolicy(bulk_gfp,
+ nr = alloc_pages_bulk_array_mempolicy_noprof(bulk_gfp,
nr_pages_request,
pages + nr_allocated);
else
- nr = alloc_pages_bulk_array_node(bulk_gfp, nid,
+ nr = alloc_pages_bulk_array_node_noprof(bulk_gfp, nid,
nr_pages_request,
pages + nr_allocated);
@@ -3542,9 +3542,9 @@ vm_area_alloc_pages(gfp_t gfp, int nid,
break;
if (nid == NUMA_NO_NODE)
- page = alloc_pages(alloc_gfp, order);
+ page = alloc_pages_noprof(alloc_gfp, order);
else
- page = alloc_pages_node(nid, alloc_gfp, order);
+ page = alloc_pages_node_noprof(nid, alloc_gfp, order);
if (unlikely(!page)) {
if (!nofail)
break;
@@ -3601,10 +3601,10 @@ static void *__vmalloc_area_node(struct vm_struct *area, gfp_t gfp_mask,
/* Please note that the recursion is strictly bounded. */
if (array_size > PAGE_SIZE) {
- area->pages = __vmalloc_node(array_size, 1, nested_gfp, node,
+ area->pages = __vmalloc_node_noprof(array_size, 1, nested_gfp, node,
area->caller);
} else {
- area->pages = kmalloc_node(array_size, nested_gfp, node);
+ area->pages = kmalloc_node_noprof(array_size, nested_gfp, node);
}
if (!area->pages) {
@@ -3687,7 +3687,7 @@ static void *__vmalloc_area_node(struct vm_struct *area, gfp_t gfp_mask,
}
/**
- * __vmalloc_node_range - allocate virtually contiguous memory
+ * __vmalloc_node_range_noprof - allocate virtually contiguous memory
* @size: allocation size
* @align: desired alignment
* @start: vm area range start
@@ -3714,7 +3714,7 @@ static void *__vmalloc_area_node(struct vm_struct *area, gfp_t gfp_mask,
*
* Return: the address of the area or %NULL on failure
*/
-void *__vmalloc_node_range(unsigned long size, unsigned long align,
+void *__vmalloc_node_range_noprof(unsigned long size, unsigned long align,
unsigned long start, unsigned long end, gfp_t gfp_mask,
pgprot_t prot, unsigned long vm_flags, int node,
const void *caller)
@@ -3843,7 +3843,7 @@ void *__vmalloc_node_range(unsigned long size, unsigned long align,
}
/**
- * __vmalloc_node - allocate virtually contiguous memory
+ * __vmalloc_node_noprof - allocate virtually contiguous memory
* @size: allocation size
* @align: desired alignment
* @gfp_mask: flags for the page level allocator
@@ -3861,10 +3861,10 @@ void *__vmalloc_node_range(unsigned long size, unsigned long align,
*
* Return: pointer to the allocated memory or %NULL on error
*/
-void *__vmalloc_node(unsigned long size, unsigned long align,
+void *__vmalloc_node_noprof(unsigned long size, unsigned long align,
gfp_t gfp_mask, int node, const void *caller)
{
- return __vmalloc_node_range(size, align, VMALLOC_START, VMALLOC_END,
+ return __vmalloc_node_range_noprof(size, align, VMALLOC_START, VMALLOC_END,
gfp_mask, PAGE_KERNEL, 0, node, caller);
}
/*
@@ -3873,15 +3873,15 @@ void *__vmalloc_node(unsigned long size, unsigned long align,
* than that.
*/
#ifdef CONFIG_TEST_VMALLOC_MODULE
-EXPORT_SYMBOL_GPL(__vmalloc_node);
+EXPORT_SYMBOL_GPL(__vmalloc_node_noprof);
#endif
-void *__vmalloc(unsigned long size, gfp_t gfp_mask)
+void *__vmalloc_noprof(unsigned long size, gfp_t gfp_mask)
{
- return __vmalloc_node(size, 1, gfp_mask, NUMA_NO_NODE,
+ return __vmalloc_node_noprof(size, 1, gfp_mask, NUMA_NO_NODE,
__builtin_return_address(0));
}
-EXPORT_SYMBOL(__vmalloc);
+EXPORT_SYMBOL(__vmalloc_noprof);
/**
* vmalloc - allocate virtually contiguous memory
@@ -3895,12 +3895,12 @@ EXPORT_SYMBOL(__vmalloc);
*
* Return: pointer to the allocated memory or %NULL on error
*/
-void *vmalloc(unsigned long size)
+void *vmalloc_noprof(unsigned long size)
{
- return __vmalloc_node(size, 1, GFP_KERNEL, NUMA_NO_NODE,
+ return __vmalloc_node_noprof(size, 1, GFP_KERNEL, NUMA_NO_NODE,
__builtin_return_address(0));
}
-EXPORT_SYMBOL(vmalloc);
+EXPORT_SYMBOL(vmalloc_noprof);
/**
* vmalloc_huge - allocate virtually contiguous memory, allow huge pages
@@ -3914,16 +3914,16 @@ EXPORT_SYMBOL(vmalloc);
*
* Return: pointer to the allocated memory or %NULL on error
*/
-void *vmalloc_huge(unsigned long size, gfp_t gfp_mask)
+void *vmalloc_huge_noprof(unsigned long size, gfp_t gfp_mask)
{
- return __vmalloc_node_range(size, 1, VMALLOC_START, VMALLOC_END,
+ return __vmalloc_node_range_noprof(size, 1, VMALLOC_START, VMALLOC_END,
gfp_mask, PAGE_KERNEL, VM_ALLOW_HUGE_VMAP,
NUMA_NO_NODE, __builtin_return_address(0));
}
-EXPORT_SYMBOL_GPL(vmalloc_huge);
+EXPORT_SYMBOL_GPL(vmalloc_huge_noprof);
/**
- * vzalloc - allocate virtually contiguous memory with zero fill
+ * vzalloc_noprof - allocate virtually contiguous memory with zero fill
* @size: allocation size
*
* Allocate enough pages to cover @size from the page level
@@ -3935,12 +3935,12 @@ EXPORT_SYMBOL_GPL(vmalloc_huge);
*
* Return: pointer to the allocated memory or %NULL on error
*/
-void *vzalloc(unsigned long size)
+void *vzalloc_noprof(unsigned long size)
{
- return __vmalloc_node(size, 1, GFP_KERNEL | __GFP_ZERO, NUMA_NO_NODE,
+ return __vmalloc_node_noprof(size, 1, GFP_KERNEL | __GFP_ZERO, NUMA_NO_NODE,
__builtin_return_address(0));
}
-EXPORT_SYMBOL(vzalloc);
+EXPORT_SYMBOL(vzalloc_noprof);
/**
* vmalloc_user - allocate zeroed virtually contiguous memory for userspace
@@ -3951,17 +3951,17 @@ EXPORT_SYMBOL(vzalloc);
*
* Return: pointer to the allocated memory or %NULL on error
*/
-void *vmalloc_user(unsigned long size)
+void *vmalloc_user_noprof(unsigned long size)
{
- return __vmalloc_node_range(size, SHMLBA, VMALLOC_START, VMALLOC_END,
+ return __vmalloc_node_range_noprof(size, SHMLBA, VMALLOC_START, VMALLOC_END,
GFP_KERNEL | __GFP_ZERO, PAGE_KERNEL,
VM_USERMAP, NUMA_NO_NODE,
__builtin_return_address(0));
}
-EXPORT_SYMBOL(vmalloc_user);
+EXPORT_SYMBOL(vmalloc_user_noprof);
/**
- * vmalloc_node - allocate memory on a specific node
+ * vmalloc_node_noprof - allocate memory on a specific node
* @size: allocation size
* @node: numa node
*
@@ -3973,15 +3973,15 @@ EXPORT_SYMBOL(vmalloc_user);
*
* Return: pointer to the allocated memory or %NULL on error
*/
-void *vmalloc_node(unsigned long size, int node)
+void *vmalloc_node_noprof(unsigned long size, int node)
{
- return __vmalloc_node(size, 1, GFP_KERNEL, node,
+ return __vmalloc_node_noprof(size, 1, GFP_KERNEL, node,
__builtin_return_address(0));
}
-EXPORT_SYMBOL(vmalloc_node);
+EXPORT_SYMBOL(vmalloc_node_noprof);
/**
- * vzalloc_node - allocate memory on a specific node with zero fill
+ * vzalloc_node_noprof - allocate memory on a specific node with zero fill
* @size: allocation size
* @node: numa node
*
@@ -3991,12 +3991,12 @@ EXPORT_SYMBOL(vmalloc_node);
*
* Return: pointer to the allocated memory or %NULL on error
*/
-void *vzalloc_node(unsigned long size, int node)
+void *vzalloc_node_noprof(unsigned long size, int node)
{
- return __vmalloc_node(size, 1, GFP_KERNEL | __GFP_ZERO, node,
+ return __vmalloc_node_noprof(size, 1, GFP_KERNEL | __GFP_ZERO, node,
__builtin_return_address(0));
}
-EXPORT_SYMBOL(vzalloc_node);
+EXPORT_SYMBOL(vzalloc_node_noprof);
#if defined(CONFIG_64BIT) && defined(CONFIG_ZONE_DMA32)
#define GFP_VMALLOC32 (GFP_DMA32 | GFP_KERNEL)
@@ -4011,7 +4011,7 @@ EXPORT_SYMBOL(vzalloc_node);
#endif
/**
- * vmalloc_32 - allocate virtually contiguous memory (32bit addressable)
+ * vmalloc_32_noprof - allocate virtually contiguous memory (32bit addressable)
* @size: allocation size
*
* Allocate enough 32bit PA addressable pages to cover @size from the
@@ -4019,15 +4019,15 @@ EXPORT_SYMBOL(vzalloc_node);
*
* Return: pointer to the allocated memory or %NULL on error
*/
-void *vmalloc_32(unsigned long size)
+void *vmalloc_32_noprof(unsigned long size)
{
- return __vmalloc_node(size, 1, GFP_VMALLOC32, NUMA_NO_NODE,
+ return __vmalloc_node_noprof(size, 1, GFP_VMALLOC32, NUMA_NO_NODE,
__builtin_return_address(0));
}
-EXPORT_SYMBOL(vmalloc_32);
+EXPORT_SYMBOL(vmalloc_32_noprof);
/**
- * vmalloc_32_user - allocate zeroed virtually contiguous 32bit memory
+ * vmalloc_32_user_noprof - allocate zeroed virtually contiguous 32bit memory
* @size: allocation size
*
* The resulting memory area is 32bit addressable and zeroed so it can be
@@ -4035,14 +4035,14 @@ EXPORT_SYMBOL(vmalloc_32);
*
* Return: pointer to the allocated memory or %NULL on error
*/
-void *vmalloc_32_user(unsigned long size)
+void *vmalloc_32_user_noprof(unsigned long size)
{
- return __vmalloc_node_range(size, SHMLBA, VMALLOC_START, VMALLOC_END,
+ return __vmalloc_node_range_noprof(size, SHMLBA, VMALLOC_START, VMALLOC_END,
GFP_VMALLOC32 | __GFP_ZERO, PAGE_KERNEL,
VM_USERMAP, NUMA_NO_NODE,
__builtin_return_address(0));
}
-EXPORT_SYMBOL(vmalloc_32_user);
+EXPORT_SYMBOL(vmalloc_32_user_noprof);
/*
* Atomically zero bytes in the iterator.
--
2.44.0.291.gc1ea87d7ee-goog
> -void *__vcalloc(size_t n, size_t size, gfp_t flags)
> +void *__vcalloc_noprof(size_t n, size_t size, gfp_t flags)
> {
> return __vmalloc_array(n, size, flags | __GFP_ZERO);
> }
> -EXPORT_SYMBOL(__vcalloc);
> +EXPORT_SYMBOL(__vcalloc_noprof);
__vmalloc_array should instead be __vmalloc_array_noprof. This is because
we would want the more specific tag present in /proc/allocinfo
Hi Suren and Kent,
On Thu, 21 Mar 2024 09:36:52 -0700 Suren Baghdasaryan <surenb@google.com> wrote:
> From: Kent Overstreet <kent.overstreet@linux.dev>
>
> This wrapps all external vmalloc allocation functions with the
> alloc_hooks() wrapper, and switches internal allocations to _noprof
> variants where appropriate, for the new memory allocation profiling
> feature.
I just noticed latest mm-unstable fails running kunit on my machine as below.
'git-bisect' says this is the first commit of the failure.
$ ./tools/testing/kunit/kunit.py run --build_dir ../kunit.out/
[10:59:53] Configuring KUnit Kernel ...
[10:59:53] Building KUnit Kernel ...
Populating config with:
$ make ARCH=um O=../kunit.out/ olddefconfig
Building with:
$ make ARCH=um O=../kunit.out/ --jobs=36
ERROR:root:/usr/bin/ld: arch/um/os-Linux/main.o: in function `__wrap_malloc':
main.c:(.text+0x10b): undefined reference to `vmalloc'
collect2: error: ld returned 1 exit status
Haven't looked into the code yet, but reporting first. May I ask your idea?
Thanks,
SJ
>
> Signed-off-by: Kent Overstreet <kent.overstreet@linux.dev>
> Signed-off-by: Suren Baghdasaryan <surenb@google.com>
> ---
> drivers/staging/media/atomisp/pci/hmm/hmm.c | 2 +-
> include/linux/vmalloc.h | 60 ++++++++++----
> kernel/kallsyms_selftest.c | 2 +-
> mm/nommu.c | 64 +++++++--------
> mm/util.c | 24 +++---
> mm/vmalloc.c | 88 ++++++++++-----------
> 6 files changed, 135 insertions(+), 105 deletions(-)
>
> diff --git a/drivers/staging/media/atomisp/pci/hmm/hmm.c b/drivers/staging/media/atomisp/pci/hmm/hmm.c
> index bb12644fd033..3e2899ad8517 100644
> --- a/drivers/staging/media/atomisp/pci/hmm/hmm.c
> +++ b/drivers/staging/media/atomisp/pci/hmm/hmm.c
> @@ -205,7 +205,7 @@ static ia_css_ptr __hmm_alloc(size_t bytes, enum hmm_bo_type type,
> }
>
> dev_dbg(atomisp_dev, "pages: 0x%08x (%zu bytes), type: %d, vmalloc %p\n",
> - bo->start, bytes, type, vmalloc);
> + bo->start, bytes, type, vmalloc_noprof);
>
> return bo->start;
>
> diff --git a/include/linux/vmalloc.h b/include/linux/vmalloc.h
> index 98ea90e90439..e4a631ec430b 100644
> --- a/include/linux/vmalloc.h
> +++ b/include/linux/vmalloc.h
> @@ -2,6 +2,8 @@
> #ifndef _LINUX_VMALLOC_H
> #define _LINUX_VMALLOC_H
>
> +#include <linux/alloc_tag.h>
> +#include <linux/sched.h>
> #include <linux/spinlock.h>
> #include <linux/init.h>
> #include <linux/list.h>
> @@ -138,26 +140,54 @@ extern unsigned long vmalloc_nr_pages(void);
> static inline unsigned long vmalloc_nr_pages(void) { return 0; }
> #endif
>
> -extern void *vmalloc(unsigned long size) __alloc_size(1);
> -extern void *vzalloc(unsigned long size) __alloc_size(1);
> -extern void *vmalloc_user(unsigned long size) __alloc_size(1);
> -extern void *vmalloc_node(unsigned long size, int node) __alloc_size(1);
> -extern void *vzalloc_node(unsigned long size, int node) __alloc_size(1);
> -extern void *vmalloc_32(unsigned long size) __alloc_size(1);
> -extern void *vmalloc_32_user(unsigned long size) __alloc_size(1);
> -extern void *__vmalloc(unsigned long size, gfp_t gfp_mask) __alloc_size(1);
> -extern void *__vmalloc_node_range(unsigned long size, unsigned long align,
> +extern void *vmalloc_noprof(unsigned long size) __alloc_size(1);
> +#define vmalloc(...) alloc_hooks(vmalloc_noprof(__VA_ARGS__))
> +
> +extern void *vzalloc_noprof(unsigned long size) __alloc_size(1);
> +#define vzalloc(...) alloc_hooks(vzalloc_noprof(__VA_ARGS__))
> +
> +extern void *vmalloc_user_noprof(unsigned long size) __alloc_size(1);
> +#define vmalloc_user(...) alloc_hooks(vmalloc_user_noprof(__VA_ARGS__))
> +
> +extern void *vmalloc_node_noprof(unsigned long size, int node) __alloc_size(1);
> +#define vmalloc_node(...) alloc_hooks(vmalloc_node_noprof(__VA_ARGS__))
> +
> +extern void *vzalloc_node_noprof(unsigned long size, int node) __alloc_size(1);
> +#define vzalloc_node(...) alloc_hooks(vzalloc_node_noprof(__VA_ARGS__))
> +
> +extern void *vmalloc_32_noprof(unsigned long size) __alloc_size(1);
> +#define vmalloc_32(...) alloc_hooks(vmalloc_32_noprof(__VA_ARGS__))
> +
> +extern void *vmalloc_32_user_noprof(unsigned long size) __alloc_size(1);
> +#define vmalloc_32_user(...) alloc_hooks(vmalloc_32_user_noprof(__VA_ARGS__))
> +
> +extern void *__vmalloc_noprof(unsigned long size, gfp_t gfp_mask) __alloc_size(1);
> +#define __vmalloc(...) alloc_hooks(__vmalloc_noprof(__VA_ARGS__))
> +
> +extern void *__vmalloc_node_range_noprof(unsigned long size, unsigned long align,
> unsigned long start, unsigned long end, gfp_t gfp_mask,
> pgprot_t prot, unsigned long vm_flags, int node,
> const void *caller) __alloc_size(1);
> -void *__vmalloc_node(unsigned long size, unsigned long align, gfp_t gfp_mask,
> +#define __vmalloc_node_range(...) alloc_hooks(__vmalloc_node_range_noprof(__VA_ARGS__))
> +
> +void *__vmalloc_node_noprof(unsigned long size, unsigned long align, gfp_t gfp_mask,
> int node, const void *caller) __alloc_size(1);
> -void *vmalloc_huge(unsigned long size, gfp_t gfp_mask) __alloc_size(1);
> +#define __vmalloc_node(...) alloc_hooks(__vmalloc_node_noprof(__VA_ARGS__))
> +
> +void *vmalloc_huge_noprof(unsigned long size, gfp_t gfp_mask) __alloc_size(1);
> +#define vmalloc_huge(...) alloc_hooks(vmalloc_huge_noprof(__VA_ARGS__))
> +
> +extern void *__vmalloc_array_noprof(size_t n, size_t size, gfp_t flags) __alloc_size(1, 2);
> +#define __vmalloc_array(...) alloc_hooks(__vmalloc_array_noprof(__VA_ARGS__))
> +
> +extern void *vmalloc_array_noprof(size_t n, size_t size) __alloc_size(1, 2);
> +#define vmalloc_array(...) alloc_hooks(vmalloc_array_noprof(__VA_ARGS__))
> +
> +extern void *__vcalloc_noprof(size_t n, size_t size, gfp_t flags) __alloc_size(1, 2);
> +#define __vcalloc(...) alloc_hooks(__vcalloc_noprof(__VA_ARGS__))
>
> -extern void *__vmalloc_array(size_t n, size_t size, gfp_t flags) __alloc_size(1, 2);
> -extern void *vmalloc_array(size_t n, size_t size) __alloc_size(1, 2);
> -extern void *__vcalloc(size_t n, size_t size, gfp_t flags) __alloc_size(1, 2);
> -extern void *vcalloc(size_t n, size_t size) __alloc_size(1, 2);
> +extern void *vcalloc_noprof(size_t n, size_t size) __alloc_size(1, 2);
> +#define vcalloc(...) alloc_hooks(vcalloc_noprof(__VA_ARGS__))
>
> extern void vfree(const void *addr);
> extern void vfree_atomic(const void *addr);
> diff --git a/kernel/kallsyms_selftest.c b/kernel/kallsyms_selftest.c
> index 8a689b4ff4f9..2f84896a7bcb 100644
> --- a/kernel/kallsyms_selftest.c
> +++ b/kernel/kallsyms_selftest.c
> @@ -82,7 +82,7 @@ static struct test_item test_items[] = {
> ITEM_FUNC(kallsyms_test_func_static),
> ITEM_FUNC(kallsyms_test_func),
> ITEM_FUNC(kallsyms_test_func_weak),
> - ITEM_FUNC(vmalloc),
> + ITEM_FUNC(vmalloc_noprof),
> ITEM_FUNC(vfree),
> #ifdef CONFIG_KALLSYMS_ALL
> ITEM_DATA(kallsyms_test_var_bss_static),
> diff --git a/mm/nommu.c b/mm/nommu.c
> index 5ec8f44e7ce9..69a6f3b4d156 100644
> --- a/mm/nommu.c
> +++ b/mm/nommu.c
> @@ -137,28 +137,28 @@ void vfree(const void *addr)
> }
> EXPORT_SYMBOL(vfree);
>
> -void *__vmalloc(unsigned long size, gfp_t gfp_mask)
> +void *__vmalloc_noprof(unsigned long size, gfp_t gfp_mask)
> {
> /*
> * You can't specify __GFP_HIGHMEM with kmalloc() since kmalloc()
> * returns only a logical address.
> */
> - return kmalloc(size, (gfp_mask | __GFP_COMP) & ~__GFP_HIGHMEM);
> + return kmalloc_noprof(size, (gfp_mask | __GFP_COMP) & ~__GFP_HIGHMEM);
> }
> -EXPORT_SYMBOL(__vmalloc);
> +EXPORT_SYMBOL(__vmalloc_noprof);
>
> -void *__vmalloc_node_range(unsigned long size, unsigned long align,
> +void *__vmalloc_node_range_noprof(unsigned long size, unsigned long align,
> unsigned long start, unsigned long end, gfp_t gfp_mask,
> pgprot_t prot, unsigned long vm_flags, int node,
> const void *caller)
> {
> - return __vmalloc(size, gfp_mask);
> + return __vmalloc_noprof(size, gfp_mask);
> }
>
> -void *__vmalloc_node(unsigned long size, unsigned long align, gfp_t gfp_mask,
> +void *__vmalloc_node_noprof(unsigned long size, unsigned long align, gfp_t gfp_mask,
> int node, const void *caller)
> {
> - return __vmalloc(size, gfp_mask);
> + return __vmalloc_noprof(size, gfp_mask);
> }
>
> static void *__vmalloc_user_flags(unsigned long size, gfp_t flags)
> @@ -179,11 +179,11 @@ static void *__vmalloc_user_flags(unsigned long size, gfp_t flags)
> return ret;
> }
>
> -void *vmalloc_user(unsigned long size)
> +void *vmalloc_user_noprof(unsigned long size)
> {
> return __vmalloc_user_flags(size, GFP_KERNEL | __GFP_ZERO);
> }
> -EXPORT_SYMBOL(vmalloc_user);
> +EXPORT_SYMBOL(vmalloc_user_noprof);
>
> struct page *vmalloc_to_page(const void *addr)
> {
> @@ -217,13 +217,13 @@ long vread_iter(struct iov_iter *iter, const char *addr, size_t count)
> * For tight control over page level allocator and protection flags
> * use __vmalloc() instead.
> */
> -void *vmalloc(unsigned long size)
> +void *vmalloc_noprof(unsigned long size)
> {
> - return __vmalloc(size, GFP_KERNEL);
> + return __vmalloc_noprof(size, GFP_KERNEL);
> }
> -EXPORT_SYMBOL(vmalloc);
> +EXPORT_SYMBOL(vmalloc_noprof);
>
> -void *vmalloc_huge(unsigned long size, gfp_t gfp_mask) __weak __alias(__vmalloc);
> +void *vmalloc_huge_noprof(unsigned long size, gfp_t gfp_mask) __weak __alias(__vmalloc_noprof);
>
> /*
> * vzalloc - allocate virtually contiguous memory with zero fill
> @@ -237,14 +237,14 @@ void *vmalloc_huge(unsigned long size, gfp_t gfp_mask) __weak __alias(__vmalloc)
> * For tight control over page level allocator and protection flags
> * use __vmalloc() instead.
> */
> -void *vzalloc(unsigned long size)
> +void *vzalloc_noprof(unsigned long size)
> {
> - return __vmalloc(size, GFP_KERNEL | __GFP_ZERO);
> + return __vmalloc_noprof(size, GFP_KERNEL | __GFP_ZERO);
> }
> -EXPORT_SYMBOL(vzalloc);
> +EXPORT_SYMBOL(vzalloc_noprof);
>
> /**
> - * vmalloc_node - allocate memory on a specific node
> + * vmalloc_node_noprof - allocate memory on a specific node
> * @size: allocation size
> * @node: numa node
> *
> @@ -254,14 +254,14 @@ EXPORT_SYMBOL(vzalloc);
> * For tight control over page level allocator and protection flags
> * use __vmalloc() instead.
> */
> -void *vmalloc_node(unsigned long size, int node)
> +void *vmalloc_node_noprof(unsigned long size, int node)
> {
> - return vmalloc(size);
> + return vmalloc_noprof(size);
> }
> -EXPORT_SYMBOL(vmalloc_node);
> +EXPORT_SYMBOL(vmalloc_node_noprof);
>
> /**
> - * vzalloc_node - allocate memory on a specific node with zero fill
> + * vzalloc_node_noprof - allocate memory on a specific node with zero fill
> * @size: allocation size
> * @node: numa node
> *
> @@ -272,27 +272,27 @@ EXPORT_SYMBOL(vmalloc_node);
> * For tight control over page level allocator and protection flags
> * use __vmalloc() instead.
> */
> -void *vzalloc_node(unsigned long size, int node)
> +void *vzalloc_node_noprof(unsigned long size, int node)
> {
> - return vzalloc(size);
> + return vzalloc_noprof(size);
> }
> -EXPORT_SYMBOL(vzalloc_node);
> +EXPORT_SYMBOL(vzalloc_node_noprof);
>
> /**
> - * vmalloc_32 - allocate virtually contiguous memory (32bit addressable)
> + * vmalloc_32_noprof - allocate virtually contiguous memory (32bit addressable)
> * @size: allocation size
> *
> * Allocate enough 32bit PA addressable pages to cover @size from the
> * page level allocator and map them into contiguous kernel virtual space.
> */
> -void *vmalloc_32(unsigned long size)
> +void *vmalloc_32_noprof(unsigned long size)
> {
> - return __vmalloc(size, GFP_KERNEL);
> + return __vmalloc_noprof(size, GFP_KERNEL);
> }
> -EXPORT_SYMBOL(vmalloc_32);
> +EXPORT_SYMBOL(vmalloc_32_noprof);
>
> /**
> - * vmalloc_32_user - allocate zeroed virtually contiguous 32bit memory
> + * vmalloc_32_user_noprof - allocate zeroed virtually contiguous 32bit memory
> * @size: allocation size
> *
> * The resulting memory area is 32bit addressable and zeroed so it can be
> @@ -301,15 +301,15 @@ EXPORT_SYMBOL(vmalloc_32);
> * VM_USERMAP is set on the corresponding VMA so that subsequent calls to
> * remap_vmalloc_range() are permissible.
> */
> -void *vmalloc_32_user(unsigned long size)
> +void *vmalloc_32_user_noprof(unsigned long size)
> {
> /*
> * We'll have to sort out the ZONE_DMA bits for 64-bit,
> * but for now this can simply use vmalloc_user() directly.
> */
> - return vmalloc_user(size);
> + return vmalloc_user_noprof(size);
> }
> -EXPORT_SYMBOL(vmalloc_32_user);
> +EXPORT_SYMBOL(vmalloc_32_user_noprof);
>
> void *vmap(struct page **pages, unsigned int count, unsigned long flags, pgprot_t prot)
> {
> diff --git a/mm/util.c b/mm/util.c
> index a79dce7546f1..157b5edcba75 100644
> --- a/mm/util.c
> +++ b/mm/util.c
> @@ -656,7 +656,7 @@ void *kvmalloc_node_noprof(size_t size, gfp_t flags, int node)
> * about the resulting pointer, and cannot play
> * protection games.
> */
> - return __vmalloc_node_range(size, 1, VMALLOC_START, VMALLOC_END,
> + return __vmalloc_node_range_noprof(size, 1, VMALLOC_START, VMALLOC_END,
> flags, PAGE_KERNEL, VM_ALLOW_HUGE_VMAP,
> node, __builtin_return_address(0));
> }
> @@ -715,12 +715,12 @@ void *kvrealloc_noprof(const void *p, size_t oldsize, size_t newsize, gfp_t flag
> EXPORT_SYMBOL(kvrealloc_noprof);
>
> /**
> - * __vmalloc_array - allocate memory for a virtually contiguous array.
> + * __vmalloc_array_noprof - allocate memory for a virtually contiguous array.
> * @n: number of elements.
> * @size: element size.
> * @flags: the type of memory to allocate (see kmalloc).
> */
> -void *__vmalloc_array(size_t n, size_t size, gfp_t flags)
> +void *__vmalloc_array_noprof(size_t n, size_t size, gfp_t flags)
> {
> size_t bytes;
>
> @@ -728,18 +728,18 @@ void *__vmalloc_array(size_t n, size_t size, gfp_t flags)
> return NULL;
> return __vmalloc(bytes, flags);
> }
> -EXPORT_SYMBOL(__vmalloc_array);
> +EXPORT_SYMBOL(__vmalloc_array_noprof);
>
> /**
> - * vmalloc_array - allocate memory for a virtually contiguous array.
> + * vmalloc_array_noprof - allocate memory for a virtually contiguous array.
> * @n: number of elements.
> * @size: element size.
> */
> -void *vmalloc_array(size_t n, size_t size)
> +void *vmalloc_array_noprof(size_t n, size_t size)
> {
> return __vmalloc_array(n, size, GFP_KERNEL);
> }
> -EXPORT_SYMBOL(vmalloc_array);
> +EXPORT_SYMBOL(vmalloc_array_noprof);
>
> /**
> * __vcalloc - allocate and zero memory for a virtually contiguous array.
> @@ -747,22 +747,22 @@ EXPORT_SYMBOL(vmalloc_array);
> * @size: element size.
> * @flags: the type of memory to allocate (see kmalloc).
> */
> -void *__vcalloc(size_t n, size_t size, gfp_t flags)
> +void *__vcalloc_noprof(size_t n, size_t size, gfp_t flags)
> {
> return __vmalloc_array(n, size, flags | __GFP_ZERO);
> }
> -EXPORT_SYMBOL(__vcalloc);
> +EXPORT_SYMBOL(__vcalloc_noprof);
>
> /**
> - * vcalloc - allocate and zero memory for a virtually contiguous array.
> + * vcalloc_noprof - allocate and zero memory for a virtually contiguous array.
> * @n: number of elements.
> * @size: element size.
> */
> -void *vcalloc(size_t n, size_t size)
> +void *vcalloc_noprof(size_t n, size_t size)
> {
> return __vmalloc_array(n, size, GFP_KERNEL | __GFP_ZERO);
> }
> -EXPORT_SYMBOL(vcalloc);
> +EXPORT_SYMBOL(vcalloc_noprof);
>
> struct anon_vma *folio_anon_vma(struct folio *folio)
> {
> diff --git a/mm/vmalloc.c b/mm/vmalloc.c
> index 22aa63f4ef63..b2f2248d85a9 100644
> --- a/mm/vmalloc.c
> +++ b/mm/vmalloc.c
> @@ -3507,12 +3507,12 @@ vm_area_alloc_pages(gfp_t gfp, int nid,
> * but mempolicy wants to alloc memory by interleaving.
> */
> if (IS_ENABLED(CONFIG_NUMA) && nid == NUMA_NO_NODE)
> - nr = alloc_pages_bulk_array_mempolicy(bulk_gfp,
> + nr = alloc_pages_bulk_array_mempolicy_noprof(bulk_gfp,
> nr_pages_request,
> pages + nr_allocated);
>
> else
> - nr = alloc_pages_bulk_array_node(bulk_gfp, nid,
> + nr = alloc_pages_bulk_array_node_noprof(bulk_gfp, nid,
> nr_pages_request,
> pages + nr_allocated);
>
> @@ -3542,9 +3542,9 @@ vm_area_alloc_pages(gfp_t gfp, int nid,
> break;
>
> if (nid == NUMA_NO_NODE)
> - page = alloc_pages(alloc_gfp, order);
> + page = alloc_pages_noprof(alloc_gfp, order);
> else
> - page = alloc_pages_node(nid, alloc_gfp, order);
> + page = alloc_pages_node_noprof(nid, alloc_gfp, order);
> if (unlikely(!page)) {
> if (!nofail)
> break;
> @@ -3601,10 +3601,10 @@ static void *__vmalloc_area_node(struct vm_struct *area, gfp_t gfp_mask,
>
> /* Please note that the recursion is strictly bounded. */
> if (array_size > PAGE_SIZE) {
> - area->pages = __vmalloc_node(array_size, 1, nested_gfp, node,
> + area->pages = __vmalloc_node_noprof(array_size, 1, nested_gfp, node,
> area->caller);
> } else {
> - area->pages = kmalloc_node(array_size, nested_gfp, node);
> + area->pages = kmalloc_node_noprof(array_size, nested_gfp, node);
> }
>
> if (!area->pages) {
> @@ -3687,7 +3687,7 @@ static void *__vmalloc_area_node(struct vm_struct *area, gfp_t gfp_mask,
> }
>
> /**
> - * __vmalloc_node_range - allocate virtually contiguous memory
> + * __vmalloc_node_range_noprof - allocate virtually contiguous memory
> * @size: allocation size
> * @align: desired alignment
> * @start: vm area range start
> @@ -3714,7 +3714,7 @@ static void *__vmalloc_area_node(struct vm_struct *area, gfp_t gfp_mask,
> *
> * Return: the address of the area or %NULL on failure
> */
> -void *__vmalloc_node_range(unsigned long size, unsigned long align,
> +void *__vmalloc_node_range_noprof(unsigned long size, unsigned long align,
> unsigned long start, unsigned long end, gfp_t gfp_mask,
> pgprot_t prot, unsigned long vm_flags, int node,
> const void *caller)
> @@ -3843,7 +3843,7 @@ void *__vmalloc_node_range(unsigned long size, unsigned long align,
> }
>
> /**
> - * __vmalloc_node - allocate virtually contiguous memory
> + * __vmalloc_node_noprof - allocate virtually contiguous memory
> * @size: allocation size
> * @align: desired alignment
> * @gfp_mask: flags for the page level allocator
> @@ -3861,10 +3861,10 @@ void *__vmalloc_node_range(unsigned long size, unsigned long align,
> *
> * Return: pointer to the allocated memory or %NULL on error
> */
> -void *__vmalloc_node(unsigned long size, unsigned long align,
> +void *__vmalloc_node_noprof(unsigned long size, unsigned long align,
> gfp_t gfp_mask, int node, const void *caller)
> {
> - return __vmalloc_node_range(size, align, VMALLOC_START, VMALLOC_END,
> + return __vmalloc_node_range_noprof(size, align, VMALLOC_START, VMALLOC_END,
> gfp_mask, PAGE_KERNEL, 0, node, caller);
> }
> /*
> @@ -3873,15 +3873,15 @@ void *__vmalloc_node(unsigned long size, unsigned long align,
> * than that.
> */
> #ifdef CONFIG_TEST_VMALLOC_MODULE
> -EXPORT_SYMBOL_GPL(__vmalloc_node);
> +EXPORT_SYMBOL_GPL(__vmalloc_node_noprof);
> #endif
>
> -void *__vmalloc(unsigned long size, gfp_t gfp_mask)
> +void *__vmalloc_noprof(unsigned long size, gfp_t gfp_mask)
> {
> - return __vmalloc_node(size, 1, gfp_mask, NUMA_NO_NODE,
> + return __vmalloc_node_noprof(size, 1, gfp_mask, NUMA_NO_NODE,
> __builtin_return_address(0));
> }
> -EXPORT_SYMBOL(__vmalloc);
> +EXPORT_SYMBOL(__vmalloc_noprof);
>
> /**
> * vmalloc - allocate virtually contiguous memory
> @@ -3895,12 +3895,12 @@ EXPORT_SYMBOL(__vmalloc);
> *
> * Return: pointer to the allocated memory or %NULL on error
> */
> -void *vmalloc(unsigned long size)
> +void *vmalloc_noprof(unsigned long size)
> {
> - return __vmalloc_node(size, 1, GFP_KERNEL, NUMA_NO_NODE,
> + return __vmalloc_node_noprof(size, 1, GFP_KERNEL, NUMA_NO_NODE,
> __builtin_return_address(0));
> }
> -EXPORT_SYMBOL(vmalloc);
> +EXPORT_SYMBOL(vmalloc_noprof);
>
> /**
> * vmalloc_huge - allocate virtually contiguous memory, allow huge pages
> @@ -3914,16 +3914,16 @@ EXPORT_SYMBOL(vmalloc);
> *
> * Return: pointer to the allocated memory or %NULL on error
> */
> -void *vmalloc_huge(unsigned long size, gfp_t gfp_mask)
> +void *vmalloc_huge_noprof(unsigned long size, gfp_t gfp_mask)
> {
> - return __vmalloc_node_range(size, 1, VMALLOC_START, VMALLOC_END,
> + return __vmalloc_node_range_noprof(size, 1, VMALLOC_START, VMALLOC_END,
> gfp_mask, PAGE_KERNEL, VM_ALLOW_HUGE_VMAP,
> NUMA_NO_NODE, __builtin_return_address(0));
> }
> -EXPORT_SYMBOL_GPL(vmalloc_huge);
> +EXPORT_SYMBOL_GPL(vmalloc_huge_noprof);
>
> /**
> - * vzalloc - allocate virtually contiguous memory with zero fill
> + * vzalloc_noprof - allocate virtually contiguous memory with zero fill
> * @size: allocation size
> *
> * Allocate enough pages to cover @size from the page level
> @@ -3935,12 +3935,12 @@ EXPORT_SYMBOL_GPL(vmalloc_huge);
> *
> * Return: pointer to the allocated memory or %NULL on error
> */
> -void *vzalloc(unsigned long size)
> +void *vzalloc_noprof(unsigned long size)
> {
> - return __vmalloc_node(size, 1, GFP_KERNEL | __GFP_ZERO, NUMA_NO_NODE,
> + return __vmalloc_node_noprof(size, 1, GFP_KERNEL | __GFP_ZERO, NUMA_NO_NODE,
> __builtin_return_address(0));
> }
> -EXPORT_SYMBOL(vzalloc);
> +EXPORT_SYMBOL(vzalloc_noprof);
>
> /**
> * vmalloc_user - allocate zeroed virtually contiguous memory for userspace
> @@ -3951,17 +3951,17 @@ EXPORT_SYMBOL(vzalloc);
> *
> * Return: pointer to the allocated memory or %NULL on error
> */
> -void *vmalloc_user(unsigned long size)
> +void *vmalloc_user_noprof(unsigned long size)
> {
> - return __vmalloc_node_range(size, SHMLBA, VMALLOC_START, VMALLOC_END,
> + return __vmalloc_node_range_noprof(size, SHMLBA, VMALLOC_START, VMALLOC_END,
> GFP_KERNEL | __GFP_ZERO, PAGE_KERNEL,
> VM_USERMAP, NUMA_NO_NODE,
> __builtin_return_address(0));
> }
> -EXPORT_SYMBOL(vmalloc_user);
> +EXPORT_SYMBOL(vmalloc_user_noprof);
>
> /**
> - * vmalloc_node - allocate memory on a specific node
> + * vmalloc_node_noprof - allocate memory on a specific node
> * @size: allocation size
> * @node: numa node
> *
> @@ -3973,15 +3973,15 @@ EXPORT_SYMBOL(vmalloc_user);
> *
> * Return: pointer to the allocated memory or %NULL on error
> */
> -void *vmalloc_node(unsigned long size, int node)
> +void *vmalloc_node_noprof(unsigned long size, int node)
> {
> - return __vmalloc_node(size, 1, GFP_KERNEL, node,
> + return __vmalloc_node_noprof(size, 1, GFP_KERNEL, node,
> __builtin_return_address(0));
> }
> -EXPORT_SYMBOL(vmalloc_node);
> +EXPORT_SYMBOL(vmalloc_node_noprof);
>
> /**
> - * vzalloc_node - allocate memory on a specific node with zero fill
> + * vzalloc_node_noprof - allocate memory on a specific node with zero fill
> * @size: allocation size
> * @node: numa node
> *
> @@ -3991,12 +3991,12 @@ EXPORT_SYMBOL(vmalloc_node);
> *
> * Return: pointer to the allocated memory or %NULL on error
> */
> -void *vzalloc_node(unsigned long size, int node)
> +void *vzalloc_node_noprof(unsigned long size, int node)
> {
> - return __vmalloc_node(size, 1, GFP_KERNEL | __GFP_ZERO, node,
> + return __vmalloc_node_noprof(size, 1, GFP_KERNEL | __GFP_ZERO, node,
> __builtin_return_address(0));
> }
> -EXPORT_SYMBOL(vzalloc_node);
> +EXPORT_SYMBOL(vzalloc_node_noprof);
>
> #if defined(CONFIG_64BIT) && defined(CONFIG_ZONE_DMA32)
> #define GFP_VMALLOC32 (GFP_DMA32 | GFP_KERNEL)
> @@ -4011,7 +4011,7 @@ EXPORT_SYMBOL(vzalloc_node);
> #endif
>
> /**
> - * vmalloc_32 - allocate virtually contiguous memory (32bit addressable)
> + * vmalloc_32_noprof - allocate virtually contiguous memory (32bit addressable)
> * @size: allocation size
> *
> * Allocate enough 32bit PA addressable pages to cover @size from the
> @@ -4019,15 +4019,15 @@ EXPORT_SYMBOL(vzalloc_node);
> *
> * Return: pointer to the allocated memory or %NULL on error
> */
> -void *vmalloc_32(unsigned long size)
> +void *vmalloc_32_noprof(unsigned long size)
> {
> - return __vmalloc_node(size, 1, GFP_VMALLOC32, NUMA_NO_NODE,
> + return __vmalloc_node_noprof(size, 1, GFP_VMALLOC32, NUMA_NO_NODE,
> __builtin_return_address(0));
> }
> -EXPORT_SYMBOL(vmalloc_32);
> +EXPORT_SYMBOL(vmalloc_32_noprof);
>
> /**
> - * vmalloc_32_user - allocate zeroed virtually contiguous 32bit memory
> + * vmalloc_32_user_noprof - allocate zeroed virtually contiguous 32bit memory
> * @size: allocation size
> *
> * The resulting memory area is 32bit addressable and zeroed so it can be
> @@ -4035,14 +4035,14 @@ EXPORT_SYMBOL(vmalloc_32);
> *
> * Return: pointer to the allocated memory or %NULL on error
> */
> -void *vmalloc_32_user(unsigned long size)
> +void *vmalloc_32_user_noprof(unsigned long size)
> {
> - return __vmalloc_node_range(size, SHMLBA, VMALLOC_START, VMALLOC_END,
> + return __vmalloc_node_range_noprof(size, SHMLBA, VMALLOC_START, VMALLOC_END,
> GFP_VMALLOC32 | __GFP_ZERO, PAGE_KERNEL,
> VM_USERMAP, NUMA_NO_NODE,
> __builtin_return_address(0));
> }
> -EXPORT_SYMBOL(vmalloc_32_user);
> +EXPORT_SYMBOL(vmalloc_32_user_noprof);
>
> /*
> * Atomically zero bytes in the iterator.
> --
> 2.44.0.291.gc1ea87d7ee-goog
On Sat, Mar 23, 2024 at 6:05 PM SeongJae Park <sj@kernel.org> wrote:
>
> Hi Suren and Kent,
>
> On Thu, 21 Mar 2024 09:36:52 -0700 Suren Baghdasaryan <surenb@google.com> wrote:
>
> > From: Kent Overstreet <kent.overstreet@linux.dev>
> >
> > This wrapps all external vmalloc allocation functions with the
> > alloc_hooks() wrapper, and switches internal allocations to _noprof
> > variants where appropriate, for the new memory allocation profiling
> > feature.
>
> I just noticed latest mm-unstable fails running kunit on my machine as below.
> 'git-bisect' says this is the first commit of the failure.
>
> $ ./tools/testing/kunit/kunit.py run --build_dir ../kunit.out/
> [10:59:53] Configuring KUnit Kernel ...
> [10:59:53] Building KUnit Kernel ...
> Populating config with:
> $ make ARCH=um O=../kunit.out/ olddefconfig
> Building with:
> $ make ARCH=um O=../kunit.out/ --jobs=36
> ERROR:root:/usr/bin/ld: arch/um/os-Linux/main.o: in function `__wrap_malloc':
> main.c:(.text+0x10b): undefined reference to `vmalloc'
> collect2: error: ld returned 1 exit status
>
> Haven't looked into the code yet, but reporting first. May I ask your idea?
Hi SeongJae,
Looks like we missed adding "#include <linux/vmalloc.h>" inside
arch/um/os-Linux/main.c in this patch:
https://lore.kernel.org/all/20240321163705.3067592-2-surenb@google.com/.
I'll be posing fixes for all 0-day issues found over the weekend and
will include a fix for this. In the meantime, to work around it you
can add that include yourself. Please let me know if the issue still
persists after doing that.
Thanks,
Suren.
>
>
> Thanks,
> SJ
>
> >
> > Signed-off-by: Kent Overstreet <kent.overstreet@linux.dev>
> > Signed-off-by: Suren Baghdasaryan <surenb@google.com>
> > ---
> > drivers/staging/media/atomisp/pci/hmm/hmm.c | 2 +-
> > include/linux/vmalloc.h | 60 ++++++++++----
> > kernel/kallsyms_selftest.c | 2 +-
> > mm/nommu.c | 64 +++++++--------
> > mm/util.c | 24 +++---
> > mm/vmalloc.c | 88 ++++++++++-----------
> > 6 files changed, 135 insertions(+), 105 deletions(-)
> >
> > diff --git a/drivers/staging/media/atomisp/pci/hmm/hmm.c b/drivers/staging/media/atomisp/pci/hmm/hmm.c
> > index bb12644fd033..3e2899ad8517 100644
> > --- a/drivers/staging/media/atomisp/pci/hmm/hmm.c
> > +++ b/drivers/staging/media/atomisp/pci/hmm/hmm.c
> > @@ -205,7 +205,7 @@ static ia_css_ptr __hmm_alloc(size_t bytes, enum hmm_bo_type type,
> > }
> >
> > dev_dbg(atomisp_dev, "pages: 0x%08x (%zu bytes), type: %d, vmalloc %p\n",
> > - bo->start, bytes, type, vmalloc);
> > + bo->start, bytes, type, vmalloc_noprof);
> >
> > return bo->start;
> >
> > diff --git a/include/linux/vmalloc.h b/include/linux/vmalloc.h
> > index 98ea90e90439..e4a631ec430b 100644
> > --- a/include/linux/vmalloc.h
> > +++ b/include/linux/vmalloc.h
> > @@ -2,6 +2,8 @@
> > #ifndef _LINUX_VMALLOC_H
> > #define _LINUX_VMALLOC_H
> >
> > +#include <linux/alloc_tag.h>
> > +#include <linux/sched.h>
> > #include <linux/spinlock.h>
> > #include <linux/init.h>
> > #include <linux/list.h>
> > @@ -138,26 +140,54 @@ extern unsigned long vmalloc_nr_pages(void);
> > static inline unsigned long vmalloc_nr_pages(void) { return 0; }
> > #endif
> >
> > -extern void *vmalloc(unsigned long size) __alloc_size(1);
> > -extern void *vzalloc(unsigned long size) __alloc_size(1);
> > -extern void *vmalloc_user(unsigned long size) __alloc_size(1);
> > -extern void *vmalloc_node(unsigned long size, int node) __alloc_size(1);
> > -extern void *vzalloc_node(unsigned long size, int node) __alloc_size(1);
> > -extern void *vmalloc_32(unsigned long size) __alloc_size(1);
> > -extern void *vmalloc_32_user(unsigned long size) __alloc_size(1);
> > -extern void *__vmalloc(unsigned long size, gfp_t gfp_mask) __alloc_size(1);
> > -extern void *__vmalloc_node_range(unsigned long size, unsigned long align,
> > +extern void *vmalloc_noprof(unsigned long size) __alloc_size(1);
> > +#define vmalloc(...) alloc_hooks(vmalloc_noprof(__VA_ARGS__))
> > +
> > +extern void *vzalloc_noprof(unsigned long size) __alloc_size(1);
> > +#define vzalloc(...) alloc_hooks(vzalloc_noprof(__VA_ARGS__))
> > +
> > +extern void *vmalloc_user_noprof(unsigned long size) __alloc_size(1);
> > +#define vmalloc_user(...) alloc_hooks(vmalloc_user_noprof(__VA_ARGS__))
> > +
> > +extern void *vmalloc_node_noprof(unsigned long size, int node) __alloc_size(1);
> > +#define vmalloc_node(...) alloc_hooks(vmalloc_node_noprof(__VA_ARGS__))
> > +
> > +extern void *vzalloc_node_noprof(unsigned long size, int node) __alloc_size(1);
> > +#define vzalloc_node(...) alloc_hooks(vzalloc_node_noprof(__VA_ARGS__))
> > +
> > +extern void *vmalloc_32_noprof(unsigned long size) __alloc_size(1);
> > +#define vmalloc_32(...) alloc_hooks(vmalloc_32_noprof(__VA_ARGS__))
> > +
> > +extern void *vmalloc_32_user_noprof(unsigned long size) __alloc_size(1);
> > +#define vmalloc_32_user(...) alloc_hooks(vmalloc_32_user_noprof(__VA_ARGS__))
> > +
> > +extern void *__vmalloc_noprof(unsigned long size, gfp_t gfp_mask) __alloc_size(1);
> > +#define __vmalloc(...) alloc_hooks(__vmalloc_noprof(__VA_ARGS__))
> > +
> > +extern void *__vmalloc_node_range_noprof(unsigned long size, unsigned long align,
> > unsigned long start, unsigned long end, gfp_t gfp_mask,
> > pgprot_t prot, unsigned long vm_flags, int node,
> > const void *caller) __alloc_size(1);
> > -void *__vmalloc_node(unsigned long size, unsigned long align, gfp_t gfp_mask,
> > +#define __vmalloc_node_range(...) alloc_hooks(__vmalloc_node_range_noprof(__VA_ARGS__))
> > +
> > +void *__vmalloc_node_noprof(unsigned long size, unsigned long align, gfp_t gfp_mask,
> > int node, const void *caller) __alloc_size(1);
> > -void *vmalloc_huge(unsigned long size, gfp_t gfp_mask) __alloc_size(1);
> > +#define __vmalloc_node(...) alloc_hooks(__vmalloc_node_noprof(__VA_ARGS__))
> > +
> > +void *vmalloc_huge_noprof(unsigned long size, gfp_t gfp_mask) __alloc_size(1);
> > +#define vmalloc_huge(...) alloc_hooks(vmalloc_huge_noprof(__VA_ARGS__))
> > +
> > +extern void *__vmalloc_array_noprof(size_t n, size_t size, gfp_t flags) __alloc_size(1, 2);
> > +#define __vmalloc_array(...) alloc_hooks(__vmalloc_array_noprof(__VA_ARGS__))
> > +
> > +extern void *vmalloc_array_noprof(size_t n, size_t size) __alloc_size(1, 2);
> > +#define vmalloc_array(...) alloc_hooks(vmalloc_array_noprof(__VA_ARGS__))
> > +
> > +extern void *__vcalloc_noprof(size_t n, size_t size, gfp_t flags) __alloc_size(1, 2);
> > +#define __vcalloc(...) alloc_hooks(__vcalloc_noprof(__VA_ARGS__))
> >
> > -extern void *__vmalloc_array(size_t n, size_t size, gfp_t flags) __alloc_size(1, 2);
> > -extern void *vmalloc_array(size_t n, size_t size) __alloc_size(1, 2);
> > -extern void *__vcalloc(size_t n, size_t size, gfp_t flags) __alloc_size(1, 2);
> > -extern void *vcalloc(size_t n, size_t size) __alloc_size(1, 2);
> > +extern void *vcalloc_noprof(size_t n, size_t size) __alloc_size(1, 2);
> > +#define vcalloc(...) alloc_hooks(vcalloc_noprof(__VA_ARGS__))
> >
> > extern void vfree(const void *addr);
> > extern void vfree_atomic(const void *addr);
> > diff --git a/kernel/kallsyms_selftest.c b/kernel/kallsyms_selftest.c
> > index 8a689b4ff4f9..2f84896a7bcb 100644
> > --- a/kernel/kallsyms_selftest.c
> > +++ b/kernel/kallsyms_selftest.c
> > @@ -82,7 +82,7 @@ static struct test_item test_items[] = {
> > ITEM_FUNC(kallsyms_test_func_static),
> > ITEM_FUNC(kallsyms_test_func),
> > ITEM_FUNC(kallsyms_test_func_weak),
> > - ITEM_FUNC(vmalloc),
> > + ITEM_FUNC(vmalloc_noprof),
> > ITEM_FUNC(vfree),
> > #ifdef CONFIG_KALLSYMS_ALL
> > ITEM_DATA(kallsyms_test_var_bss_static),
> > diff --git a/mm/nommu.c b/mm/nommu.c
> > index 5ec8f44e7ce9..69a6f3b4d156 100644
> > --- a/mm/nommu.c
> > +++ b/mm/nommu.c
> > @@ -137,28 +137,28 @@ void vfree(const void *addr)
> > }
> > EXPORT_SYMBOL(vfree);
> >
> > -void *__vmalloc(unsigned long size, gfp_t gfp_mask)
> > +void *__vmalloc_noprof(unsigned long size, gfp_t gfp_mask)
> > {
> > /*
> > * You can't specify __GFP_HIGHMEM with kmalloc() since kmalloc()
> > * returns only a logical address.
> > */
> > - return kmalloc(size, (gfp_mask | __GFP_COMP) & ~__GFP_HIGHMEM);
> > + return kmalloc_noprof(size, (gfp_mask | __GFP_COMP) & ~__GFP_HIGHMEM);
> > }
> > -EXPORT_SYMBOL(__vmalloc);
> > +EXPORT_SYMBOL(__vmalloc_noprof);
> >
> > -void *__vmalloc_node_range(unsigned long size, unsigned long align,
> > +void *__vmalloc_node_range_noprof(unsigned long size, unsigned long align,
> > unsigned long start, unsigned long end, gfp_t gfp_mask,
> > pgprot_t prot, unsigned long vm_flags, int node,
> > const void *caller)
> > {
> > - return __vmalloc(size, gfp_mask);
> > + return __vmalloc_noprof(size, gfp_mask);
> > }
> >
> > -void *__vmalloc_node(unsigned long size, unsigned long align, gfp_t gfp_mask,
> > +void *__vmalloc_node_noprof(unsigned long size, unsigned long align, gfp_t gfp_mask,
> > int node, const void *caller)
> > {
> > - return __vmalloc(size, gfp_mask);
> > + return __vmalloc_noprof(size, gfp_mask);
> > }
> >
> > static void *__vmalloc_user_flags(unsigned long size, gfp_t flags)
> > @@ -179,11 +179,11 @@ static void *__vmalloc_user_flags(unsigned long size, gfp_t flags)
> > return ret;
> > }
> >
> > -void *vmalloc_user(unsigned long size)
> > +void *vmalloc_user_noprof(unsigned long size)
> > {
> > return __vmalloc_user_flags(size, GFP_KERNEL | __GFP_ZERO);
> > }
> > -EXPORT_SYMBOL(vmalloc_user);
> > +EXPORT_SYMBOL(vmalloc_user_noprof);
> >
> > struct page *vmalloc_to_page(const void *addr)
> > {
> > @@ -217,13 +217,13 @@ long vread_iter(struct iov_iter *iter, const char *addr, size_t count)
> > * For tight control over page level allocator and protection flags
> > * use __vmalloc() instead.
> > */
> > -void *vmalloc(unsigned long size)
> > +void *vmalloc_noprof(unsigned long size)
> > {
> > - return __vmalloc(size, GFP_KERNEL);
> > + return __vmalloc_noprof(size, GFP_KERNEL);
> > }
> > -EXPORT_SYMBOL(vmalloc);
> > +EXPORT_SYMBOL(vmalloc_noprof);
> >
> > -void *vmalloc_huge(unsigned long size, gfp_t gfp_mask) __weak __alias(__vmalloc);
> > +void *vmalloc_huge_noprof(unsigned long size, gfp_t gfp_mask) __weak __alias(__vmalloc_noprof);
> >
> > /*
> > * vzalloc - allocate virtually contiguous memory with zero fill
> > @@ -237,14 +237,14 @@ void *vmalloc_huge(unsigned long size, gfp_t gfp_mask) __weak __alias(__vmalloc)
> > * For tight control over page level allocator and protection flags
> > * use __vmalloc() instead.
> > */
> > -void *vzalloc(unsigned long size)
> > +void *vzalloc_noprof(unsigned long size)
> > {
> > - return __vmalloc(size, GFP_KERNEL | __GFP_ZERO);
> > + return __vmalloc_noprof(size, GFP_KERNEL | __GFP_ZERO);
> > }
> > -EXPORT_SYMBOL(vzalloc);
> > +EXPORT_SYMBOL(vzalloc_noprof);
> >
> > /**
> > - * vmalloc_node - allocate memory on a specific node
> > + * vmalloc_node_noprof - allocate memory on a specific node
> > * @size: allocation size
> > * @node: numa node
> > *
> > @@ -254,14 +254,14 @@ EXPORT_SYMBOL(vzalloc);
> > * For tight control over page level allocator and protection flags
> > * use __vmalloc() instead.
> > */
> > -void *vmalloc_node(unsigned long size, int node)
> > +void *vmalloc_node_noprof(unsigned long size, int node)
> > {
> > - return vmalloc(size);
> > + return vmalloc_noprof(size);
> > }
> > -EXPORT_SYMBOL(vmalloc_node);
> > +EXPORT_SYMBOL(vmalloc_node_noprof);
> >
> > /**
> > - * vzalloc_node - allocate memory on a specific node with zero fill
> > + * vzalloc_node_noprof - allocate memory on a specific node with zero fill
> > * @size: allocation size
> > * @node: numa node
> > *
> > @@ -272,27 +272,27 @@ EXPORT_SYMBOL(vmalloc_node);
> > * For tight control over page level allocator and protection flags
> > * use __vmalloc() instead.
> > */
> > -void *vzalloc_node(unsigned long size, int node)
> > +void *vzalloc_node_noprof(unsigned long size, int node)
> > {
> > - return vzalloc(size);
> > + return vzalloc_noprof(size);
> > }
> > -EXPORT_SYMBOL(vzalloc_node);
> > +EXPORT_SYMBOL(vzalloc_node_noprof);
> >
> > /**
> > - * vmalloc_32 - allocate virtually contiguous memory (32bit addressable)
> > + * vmalloc_32_noprof - allocate virtually contiguous memory (32bit addressable)
> > * @size: allocation size
> > *
> > * Allocate enough 32bit PA addressable pages to cover @size from the
> > * page level allocator and map them into contiguous kernel virtual space.
> > */
> > -void *vmalloc_32(unsigned long size)
> > +void *vmalloc_32_noprof(unsigned long size)
> > {
> > - return __vmalloc(size, GFP_KERNEL);
> > + return __vmalloc_noprof(size, GFP_KERNEL);
> > }
> > -EXPORT_SYMBOL(vmalloc_32);
> > +EXPORT_SYMBOL(vmalloc_32_noprof);
> >
> > /**
> > - * vmalloc_32_user - allocate zeroed virtually contiguous 32bit memory
> > + * vmalloc_32_user_noprof - allocate zeroed virtually contiguous 32bit memory
> > * @size: allocation size
> > *
> > * The resulting memory area is 32bit addressable and zeroed so it can be
> > @@ -301,15 +301,15 @@ EXPORT_SYMBOL(vmalloc_32);
> > * VM_USERMAP is set on the corresponding VMA so that subsequent calls to
> > * remap_vmalloc_range() are permissible.
> > */
> > -void *vmalloc_32_user(unsigned long size)
> > +void *vmalloc_32_user_noprof(unsigned long size)
> > {
> > /*
> > * We'll have to sort out the ZONE_DMA bits for 64-bit,
> > * but for now this can simply use vmalloc_user() directly.
> > */
> > - return vmalloc_user(size);
> > + return vmalloc_user_noprof(size);
> > }
> > -EXPORT_SYMBOL(vmalloc_32_user);
> > +EXPORT_SYMBOL(vmalloc_32_user_noprof);
> >
> > void *vmap(struct page **pages, unsigned int count, unsigned long flags, pgprot_t prot)
> > {
> > diff --git a/mm/util.c b/mm/util.c
> > index a79dce7546f1..157b5edcba75 100644
> > --- a/mm/util.c
> > +++ b/mm/util.c
> > @@ -656,7 +656,7 @@ void *kvmalloc_node_noprof(size_t size, gfp_t flags, int node)
> > * about the resulting pointer, and cannot play
> > * protection games.
> > */
> > - return __vmalloc_node_range(size, 1, VMALLOC_START, VMALLOC_END,
> > + return __vmalloc_node_range_noprof(size, 1, VMALLOC_START, VMALLOC_END,
> > flags, PAGE_KERNEL, VM_ALLOW_HUGE_VMAP,
> > node, __builtin_return_address(0));
> > }
> > @@ -715,12 +715,12 @@ void *kvrealloc_noprof(const void *p, size_t oldsize, size_t newsize, gfp_t flag
> > EXPORT_SYMBOL(kvrealloc_noprof);
> >
> > /**
> > - * __vmalloc_array - allocate memory for a virtually contiguous array.
> > + * __vmalloc_array_noprof - allocate memory for a virtually contiguous array.
> > * @n: number of elements.
> > * @size: element size.
> > * @flags: the type of memory to allocate (see kmalloc).
> > */
> > -void *__vmalloc_array(size_t n, size_t size, gfp_t flags)
> > +void *__vmalloc_array_noprof(size_t n, size_t size, gfp_t flags)
> > {
> > size_t bytes;
> >
> > @@ -728,18 +728,18 @@ void *__vmalloc_array(size_t n, size_t size, gfp_t flags)
> > return NULL;
> > return __vmalloc(bytes, flags);
> > }
> > -EXPORT_SYMBOL(__vmalloc_array);
> > +EXPORT_SYMBOL(__vmalloc_array_noprof);
> >
> > /**
> > - * vmalloc_array - allocate memory for a virtually contiguous array.
> > + * vmalloc_array_noprof - allocate memory for a virtually contiguous array.
> > * @n: number of elements.
> > * @size: element size.
> > */
> > -void *vmalloc_array(size_t n, size_t size)
> > +void *vmalloc_array_noprof(size_t n, size_t size)
> > {
> > return __vmalloc_array(n, size, GFP_KERNEL);
> > }
> > -EXPORT_SYMBOL(vmalloc_array);
> > +EXPORT_SYMBOL(vmalloc_array_noprof);
> >
> > /**
> > * __vcalloc - allocate and zero memory for a virtually contiguous array.
> > @@ -747,22 +747,22 @@ EXPORT_SYMBOL(vmalloc_array);
> > * @size: element size.
> > * @flags: the type of memory to allocate (see kmalloc).
> > */
> > -void *__vcalloc(size_t n, size_t size, gfp_t flags)
> > +void *__vcalloc_noprof(size_t n, size_t size, gfp_t flags)
> > {
> > return __vmalloc_array(n, size, flags | __GFP_ZERO);
> > }
> > -EXPORT_SYMBOL(__vcalloc);
> > +EXPORT_SYMBOL(__vcalloc_noprof);
> >
> > /**
> > - * vcalloc - allocate and zero memory for a virtually contiguous array.
> > + * vcalloc_noprof - allocate and zero memory for a virtually contiguous array.
> > * @n: number of elements.
> > * @size: element size.
> > */
> > -void *vcalloc(size_t n, size_t size)
> > +void *vcalloc_noprof(size_t n, size_t size)
> > {
> > return __vmalloc_array(n, size, GFP_KERNEL | __GFP_ZERO);
> > }
> > -EXPORT_SYMBOL(vcalloc);
> > +EXPORT_SYMBOL(vcalloc_noprof);
> >
> > struct anon_vma *folio_anon_vma(struct folio *folio)
> > {
> > diff --git a/mm/vmalloc.c b/mm/vmalloc.c
> > index 22aa63f4ef63..b2f2248d85a9 100644
> > --- a/mm/vmalloc.c
> > +++ b/mm/vmalloc.c
> > @@ -3507,12 +3507,12 @@ vm_area_alloc_pages(gfp_t gfp, int nid,
> > * but mempolicy wants to alloc memory by interleaving.
> > */
> > if (IS_ENABLED(CONFIG_NUMA) && nid == NUMA_NO_NODE)
> > - nr = alloc_pages_bulk_array_mempolicy(bulk_gfp,
> > + nr = alloc_pages_bulk_array_mempolicy_noprof(bulk_gfp,
> > nr_pages_request,
> > pages + nr_allocated);
> >
> > else
> > - nr = alloc_pages_bulk_array_node(bulk_gfp, nid,
> > + nr = alloc_pages_bulk_array_node_noprof(bulk_gfp, nid,
> > nr_pages_request,
> > pages + nr_allocated);
> >
> > @@ -3542,9 +3542,9 @@ vm_area_alloc_pages(gfp_t gfp, int nid,
> > break;
> >
> > if (nid == NUMA_NO_NODE)
> > - page = alloc_pages(alloc_gfp, order);
> > + page = alloc_pages_noprof(alloc_gfp, order);
> > else
> > - page = alloc_pages_node(nid, alloc_gfp, order);
> > + page = alloc_pages_node_noprof(nid, alloc_gfp, order);
> > if (unlikely(!page)) {
> > if (!nofail)
> > break;
> > @@ -3601,10 +3601,10 @@ static void *__vmalloc_area_node(struct vm_struct *area, gfp_t gfp_mask,
> >
> > /* Please note that the recursion is strictly bounded. */
> > if (array_size > PAGE_SIZE) {
> > - area->pages = __vmalloc_node(array_size, 1, nested_gfp, node,
> > + area->pages = __vmalloc_node_noprof(array_size, 1, nested_gfp, node,
> > area->caller);
> > } else {
> > - area->pages = kmalloc_node(array_size, nested_gfp, node);
> > + area->pages = kmalloc_node_noprof(array_size, nested_gfp, node);
> > }
> >
> > if (!area->pages) {
> > @@ -3687,7 +3687,7 @@ static void *__vmalloc_area_node(struct vm_struct *area, gfp_t gfp_mask,
> > }
> >
> > /**
> > - * __vmalloc_node_range - allocate virtually contiguous memory
> > + * __vmalloc_node_range_noprof - allocate virtually contiguous memory
> > * @size: allocation size
> > * @align: desired alignment
> > * @start: vm area range start
> > @@ -3714,7 +3714,7 @@ static void *__vmalloc_area_node(struct vm_struct *area, gfp_t gfp_mask,
> > *
> > * Return: the address of the area or %NULL on failure
> > */
> > -void *__vmalloc_node_range(unsigned long size, unsigned long align,
> > +void *__vmalloc_node_range_noprof(unsigned long size, unsigned long align,
> > unsigned long start, unsigned long end, gfp_t gfp_mask,
> > pgprot_t prot, unsigned long vm_flags, int node,
> > const void *caller)
> > @@ -3843,7 +3843,7 @@ void *__vmalloc_node_range(unsigned long size, unsigned long align,
> > }
> >
> > /**
> > - * __vmalloc_node - allocate virtually contiguous memory
> > + * __vmalloc_node_noprof - allocate virtually contiguous memory
> > * @size: allocation size
> > * @align: desired alignment
> > * @gfp_mask: flags for the page level allocator
> > @@ -3861,10 +3861,10 @@ void *__vmalloc_node_range(unsigned long size, unsigned long align,
> > *
> > * Return: pointer to the allocated memory or %NULL on error
> > */
> > -void *__vmalloc_node(unsigned long size, unsigned long align,
> > +void *__vmalloc_node_noprof(unsigned long size, unsigned long align,
> > gfp_t gfp_mask, int node, const void *caller)
> > {
> > - return __vmalloc_node_range(size, align, VMALLOC_START, VMALLOC_END,
> > + return __vmalloc_node_range_noprof(size, align, VMALLOC_START, VMALLOC_END,
> > gfp_mask, PAGE_KERNEL, 0, node, caller);
> > }
> > /*
> > @@ -3873,15 +3873,15 @@ void *__vmalloc_node(unsigned long size, unsigned long align,
> > * than that.
> > */
> > #ifdef CONFIG_TEST_VMALLOC_MODULE
> > -EXPORT_SYMBOL_GPL(__vmalloc_node);
> > +EXPORT_SYMBOL_GPL(__vmalloc_node_noprof);
> > #endif
> >
> > -void *__vmalloc(unsigned long size, gfp_t gfp_mask)
> > +void *__vmalloc_noprof(unsigned long size, gfp_t gfp_mask)
> > {
> > - return __vmalloc_node(size, 1, gfp_mask, NUMA_NO_NODE,
> > + return __vmalloc_node_noprof(size, 1, gfp_mask, NUMA_NO_NODE,
> > __builtin_return_address(0));
> > }
> > -EXPORT_SYMBOL(__vmalloc);
> > +EXPORT_SYMBOL(__vmalloc_noprof);
> >
> > /**
> > * vmalloc - allocate virtually contiguous memory
> > @@ -3895,12 +3895,12 @@ EXPORT_SYMBOL(__vmalloc);
> > *
> > * Return: pointer to the allocated memory or %NULL on error
> > */
> > -void *vmalloc(unsigned long size)
> > +void *vmalloc_noprof(unsigned long size)
> > {
> > - return __vmalloc_node(size, 1, GFP_KERNEL, NUMA_NO_NODE,
> > + return __vmalloc_node_noprof(size, 1, GFP_KERNEL, NUMA_NO_NODE,
> > __builtin_return_address(0));
> > }
> > -EXPORT_SYMBOL(vmalloc);
> > +EXPORT_SYMBOL(vmalloc_noprof);
> >
> > /**
> > * vmalloc_huge - allocate virtually contiguous memory, allow huge pages
> > @@ -3914,16 +3914,16 @@ EXPORT_SYMBOL(vmalloc);
> > *
> > * Return: pointer to the allocated memory or %NULL on error
> > */
> > -void *vmalloc_huge(unsigned long size, gfp_t gfp_mask)
> > +void *vmalloc_huge_noprof(unsigned long size, gfp_t gfp_mask)
> > {
> > - return __vmalloc_node_range(size, 1, VMALLOC_START, VMALLOC_END,
> > + return __vmalloc_node_range_noprof(size, 1, VMALLOC_START, VMALLOC_END,
> > gfp_mask, PAGE_KERNEL, VM_ALLOW_HUGE_VMAP,
> > NUMA_NO_NODE, __builtin_return_address(0));
> > }
> > -EXPORT_SYMBOL_GPL(vmalloc_huge);
> > +EXPORT_SYMBOL_GPL(vmalloc_huge_noprof);
> >
> > /**
> > - * vzalloc - allocate virtually contiguous memory with zero fill
> > + * vzalloc_noprof - allocate virtually contiguous memory with zero fill
> > * @size: allocation size
> > *
> > * Allocate enough pages to cover @size from the page level
> > @@ -3935,12 +3935,12 @@ EXPORT_SYMBOL_GPL(vmalloc_huge);
> > *
> > * Return: pointer to the allocated memory or %NULL on error
> > */
> > -void *vzalloc(unsigned long size)
> > +void *vzalloc_noprof(unsigned long size)
> > {
> > - return __vmalloc_node(size, 1, GFP_KERNEL | __GFP_ZERO, NUMA_NO_NODE,
> > + return __vmalloc_node_noprof(size, 1, GFP_KERNEL | __GFP_ZERO, NUMA_NO_NODE,
> > __builtin_return_address(0));
> > }
> > -EXPORT_SYMBOL(vzalloc);
> > +EXPORT_SYMBOL(vzalloc_noprof);
> >
> > /**
> > * vmalloc_user - allocate zeroed virtually contiguous memory for userspace
> > @@ -3951,17 +3951,17 @@ EXPORT_SYMBOL(vzalloc);
> > *
> > * Return: pointer to the allocated memory or %NULL on error
> > */
> > -void *vmalloc_user(unsigned long size)
> > +void *vmalloc_user_noprof(unsigned long size)
> > {
> > - return __vmalloc_node_range(size, SHMLBA, VMALLOC_START, VMALLOC_END,
> > + return __vmalloc_node_range_noprof(size, SHMLBA, VMALLOC_START, VMALLOC_END,
> > GFP_KERNEL | __GFP_ZERO, PAGE_KERNEL,
> > VM_USERMAP, NUMA_NO_NODE,
> > __builtin_return_address(0));
> > }
> > -EXPORT_SYMBOL(vmalloc_user);
> > +EXPORT_SYMBOL(vmalloc_user_noprof);
> >
> > /**
> > - * vmalloc_node - allocate memory on a specific node
> > + * vmalloc_node_noprof - allocate memory on a specific node
> > * @size: allocation size
> > * @node: numa node
> > *
> > @@ -3973,15 +3973,15 @@ EXPORT_SYMBOL(vmalloc_user);
> > *
> > * Return: pointer to the allocated memory or %NULL on error
> > */
> > -void *vmalloc_node(unsigned long size, int node)
> > +void *vmalloc_node_noprof(unsigned long size, int node)
> > {
> > - return __vmalloc_node(size, 1, GFP_KERNEL, node,
> > + return __vmalloc_node_noprof(size, 1, GFP_KERNEL, node,
> > __builtin_return_address(0));
> > }
> > -EXPORT_SYMBOL(vmalloc_node);
> > +EXPORT_SYMBOL(vmalloc_node_noprof);
> >
> > /**
> > - * vzalloc_node - allocate memory on a specific node with zero fill
> > + * vzalloc_node_noprof - allocate memory on a specific node with zero fill
> > * @size: allocation size
> > * @node: numa node
> > *
> > @@ -3991,12 +3991,12 @@ EXPORT_SYMBOL(vmalloc_node);
> > *
> > * Return: pointer to the allocated memory or %NULL on error
> > */
> > -void *vzalloc_node(unsigned long size, int node)
> > +void *vzalloc_node_noprof(unsigned long size, int node)
> > {
> > - return __vmalloc_node(size, 1, GFP_KERNEL | __GFP_ZERO, node,
> > + return __vmalloc_node_noprof(size, 1, GFP_KERNEL | __GFP_ZERO, node,
> > __builtin_return_address(0));
> > }
> > -EXPORT_SYMBOL(vzalloc_node);
> > +EXPORT_SYMBOL(vzalloc_node_noprof);
> >
> > #if defined(CONFIG_64BIT) && defined(CONFIG_ZONE_DMA32)
> > #define GFP_VMALLOC32 (GFP_DMA32 | GFP_KERNEL)
> > @@ -4011,7 +4011,7 @@ EXPORT_SYMBOL(vzalloc_node);
> > #endif
> >
> > /**
> > - * vmalloc_32 - allocate virtually contiguous memory (32bit addressable)
> > + * vmalloc_32_noprof - allocate virtually contiguous memory (32bit addressable)
> > * @size: allocation size
> > *
> > * Allocate enough 32bit PA addressable pages to cover @size from the
> > @@ -4019,15 +4019,15 @@ EXPORT_SYMBOL(vzalloc_node);
> > *
> > * Return: pointer to the allocated memory or %NULL on error
> > */
> > -void *vmalloc_32(unsigned long size)
> > +void *vmalloc_32_noprof(unsigned long size)
> > {
> > - return __vmalloc_node(size, 1, GFP_VMALLOC32, NUMA_NO_NODE,
> > + return __vmalloc_node_noprof(size, 1, GFP_VMALLOC32, NUMA_NO_NODE,
> > __builtin_return_address(0));
> > }
> > -EXPORT_SYMBOL(vmalloc_32);
> > +EXPORT_SYMBOL(vmalloc_32_noprof);
> >
> > /**
> > - * vmalloc_32_user - allocate zeroed virtually contiguous 32bit memory
> > + * vmalloc_32_user_noprof - allocate zeroed virtually contiguous 32bit memory
> > * @size: allocation size
> > *
> > * The resulting memory area is 32bit addressable and zeroed so it can be
> > @@ -4035,14 +4035,14 @@ EXPORT_SYMBOL(vmalloc_32);
> > *
> > * Return: pointer to the allocated memory or %NULL on error
> > */
> > -void *vmalloc_32_user(unsigned long size)
> > +void *vmalloc_32_user_noprof(unsigned long size)
> > {
> > - return __vmalloc_node_range(size, SHMLBA, VMALLOC_START, VMALLOC_END,
> > + return __vmalloc_node_range_noprof(size, SHMLBA, VMALLOC_START, VMALLOC_END,
> > GFP_VMALLOC32 | __GFP_ZERO, PAGE_KERNEL,
> > VM_USERMAP, NUMA_NO_NODE,
> > __builtin_return_address(0));
> > }
> > -EXPORT_SYMBOL(vmalloc_32_user);
> > +EXPORT_SYMBOL(vmalloc_32_user_noprof);
> >
> > /*
> > * Atomically zero bytes in the iterator.
> > --
> > 2.44.0.291.gc1ea87d7ee-goog
>
> --
> To unsubscribe from this group and stop receiving emails from it, send an email to kernel-team+unsubscribe@android.com.
>
On Mon, 25 Mar 2024 14:56:01 +0000 Suren Baghdasaryan <surenb@google.com> wrote:
> On Sat, Mar 23, 2024 at 6:05 PM SeongJae Park <sj@kernel.org> wrote:
> >
> > Hi Suren and Kent,
> >
> > On Thu, 21 Mar 2024 09:36:52 -0700 Suren Baghdasaryan <surenb@google.com> wrote:
> >
> > > From: Kent Overstreet <kent.overstreet@linux.dev>
> > >
> > > This wrapps all external vmalloc allocation functions with the
> > > alloc_hooks() wrapper, and switches internal allocations to _noprof
> > > variants where appropriate, for the new memory allocation profiling
> > > feature.
> >
> > I just noticed latest mm-unstable fails running kunit on my machine as below.
> > 'git-bisect' says this is the first commit of the failure.
> >
> > $ ./tools/testing/kunit/kunit.py run --build_dir ../kunit.out/
> > [10:59:53] Configuring KUnit Kernel ...
> > [10:59:53] Building KUnit Kernel ...
> > Populating config with:
> > $ make ARCH=um O=../kunit.out/ olddefconfig
> > Building with:
> > $ make ARCH=um O=../kunit.out/ --jobs=36
> > ERROR:root:/usr/bin/ld: arch/um/os-Linux/main.o: in function `__wrap_malloc':
> > main.c:(.text+0x10b): undefined reference to `vmalloc'
> > collect2: error: ld returned 1 exit status
> >
> > Haven't looked into the code yet, but reporting first. May I ask your idea?
>
> Hi SeongJae,
> Looks like we missed adding "#include <linux/vmalloc.h>" inside
> arch/um/os-Linux/main.c in this patch:
> https://lore.kernel.org/all/20240321163705.3067592-2-surenb@google.com/.
> I'll be posing fixes for all 0-day issues found over the weekend and
> will include a fix for this. In the meantime, to work around it you
> can add that include yourself. Please let me know if the issue still
> persists after doing that.
Thank you, Suren. The change made the error message disappears. However, it
introduced another one.
$ git diff
diff --git a/arch/um/os-Linux/main.c b/arch/um/os-Linux/main.c
index c8a42ecbd7a2..8fe274e9f3a4 100644
--- a/arch/um/os-Linux/main.c
+++ b/arch/um/os-Linux/main.c
@@ -16,6 +16,7 @@
#include <kern_util.h>
#include <os.h>
#include <um_malloc.h>
+#include <linux/vmalloc.h>
#define PGD_BOUND (4 * 1024 * 1024)
#define STACKSIZE (8 * 1024 * 1024)
$
$ ./tools/testing/kunit/kunit.py run --build_dir ../kunit.out/
[10:43:13] Configuring KUnit Kernel ...
[10:43:13] Building KUnit Kernel ...
Populating config with:
$ make ARCH=um O=../kunit.out/ olddefconfig
Building with:
$ make ARCH=um O=../kunit.out/ --jobs=36
ERROR:root:In file included from .../arch/um/kernel/asm-offsets.c:1:
.../arch/x86/um/shared/sysdep/kernel-offsets.h:9:6: warning: no previous prototype for ‘foo’ [-Wmissing-prototypes]
9 | void foo(void)
| ^~~
In file included from .../include/linux/alloc_tag.h:8,
from .../include/linux/vmalloc.h:5,
from .../arch/um/os-Linux/main.c:19:
.../include/linux/bug.h:5:10: fatal error: asm/bug.h: No such file or directory
5 | #include <asm/bug.h>
| ^~~~~~~~~~~
compilation terminated.
Thanks,
SJ
[...]
On Mon, Mar 25, 2024 at 10:49 AM SeongJae Park <sj@kernel.org> wrote: > > On Mon, 25 Mar 2024 14:56:01 +0000 Suren Baghdasaryan <surenb@google.com> wrote: > > > On Sat, Mar 23, 2024 at 6:05 PM SeongJae Park <sj@kernel.org> wrote: > > > > > > Hi Suren and Kent, > > > > > > On Thu, 21 Mar 2024 09:36:52 -0700 Suren Baghdasaryan <surenb@google.com> wrote: > > > > > > > From: Kent Overstreet <kent.overstreet@linux.dev> > > > > > > > > This wrapps all external vmalloc allocation functions with the > > > > alloc_hooks() wrapper, and switches internal allocations to _noprof > > > > variants where appropriate, for the new memory allocation profiling > > > > feature. > > > > > > I just noticed latest mm-unstable fails running kunit on my machine as below. > > > 'git-bisect' says this is the first commit of the failure. > > > > > > $ ./tools/testing/kunit/kunit.py run --build_dir ../kunit.out/ > > > [10:59:53] Configuring KUnit Kernel ... > > > [10:59:53] Building KUnit Kernel ... > > > Populating config with: > > > $ make ARCH=um O=../kunit.out/ olddefconfig > > > Building with: > > > $ make ARCH=um O=../kunit.out/ --jobs=36 > > > ERROR:root:/usr/bin/ld: arch/um/os-Linux/main.o: in function `__wrap_malloc': > > > main.c:(.text+0x10b): undefined reference to `vmalloc' > > > collect2: error: ld returned 1 exit status > > > > > > Haven't looked into the code yet, but reporting first. May I ask your idea? > > > > Hi SeongJae, > > Looks like we missed adding "#include <linux/vmalloc.h>" inside > > arch/um/os-Linux/main.c in this patch: > > https://lore.kernel.org/all/20240321163705.3067592-2-surenb@google.com/. > > I'll be posing fixes for all 0-day issues found over the weekend and > > will include a fix for this. In the meantime, to work around it you > > can add that include yourself. Please let me know if the issue still > > persists after doing that. > > Thank you, Suren. The change made the error message disappears. However, it > introduced another one. Ok, let me investigate and I'll try to get a fix for it today evening. Thanks, Suren. > > $ git diff > diff --git a/arch/um/os-Linux/main.c b/arch/um/os-Linux/main.c > index c8a42ecbd7a2..8fe274e9f3a4 100644 > --- a/arch/um/os-Linux/main.c > +++ b/arch/um/os-Linux/main.c > @@ -16,6 +16,7 @@ > #include <kern_util.h> > #include <os.h> > #include <um_malloc.h> > +#include <linux/vmalloc.h> > > #define PGD_BOUND (4 * 1024 * 1024) > #define STACKSIZE (8 * 1024 * 1024) > $ > $ ./tools/testing/kunit/kunit.py run --build_dir ../kunit.out/ > [10:43:13] Configuring KUnit Kernel ... > [10:43:13] Building KUnit Kernel ... > Populating config with: > $ make ARCH=um O=../kunit.out/ olddefconfig > Building with: > $ make ARCH=um O=../kunit.out/ --jobs=36 > ERROR:root:In file included from .../arch/um/kernel/asm-offsets.c:1: > .../arch/x86/um/shared/sysdep/kernel-offsets.h:9:6: warning: no previous prototype for ‘foo’ [-Wmissing-prototypes] > 9 | void foo(void) > | ^~~ > In file included from .../include/linux/alloc_tag.h:8, > from .../include/linux/vmalloc.h:5, > from .../arch/um/os-Linux/main.c:19: > .../include/linux/bug.h:5:10: fatal error: asm/bug.h: No such file or directory > 5 | #include <asm/bug.h> > | ^~~~~~~~~~~ > compilation terminated. > > > Thanks, > SJ > > [...]
On Mon, 25 Mar 2024 10:59:01 -0700 Suren Baghdasaryan <surenb@google.com> wrote: > On Mon, Mar 25, 2024 at 10:49 AM SeongJae Park <sj@kernel.org> wrote: > > > > On Mon, 25 Mar 2024 14:56:01 +0000 Suren Baghdasaryan <surenb@google.com> wrote: > > > > > On Sat, Mar 23, 2024 at 6:05 PM SeongJae Park <sj@kernel.org> wrote: > > > > > > > > Hi Suren and Kent, > > > > > > > > On Thu, 21 Mar 2024 09:36:52 -0700 Suren Baghdasaryan <surenb@google.com> wrote: > > > > > > > > > From: Kent Overstreet <kent.overstreet@linux.dev> > > > > > > > > > > This wrapps all external vmalloc allocation functions with the > > > > > alloc_hooks() wrapper, and switches internal allocations to _noprof > > > > > variants where appropriate, for the new memory allocation profiling > > > > > feature. > > > > > > > > I just noticed latest mm-unstable fails running kunit on my machine as below. > > > > 'git-bisect' says this is the first commit of the failure. > > > > > > > > $ ./tools/testing/kunit/kunit.py run --build_dir ../kunit.out/ > > > > [10:59:53] Configuring KUnit Kernel ... > > > > [10:59:53] Building KUnit Kernel ... > > > > Populating config with: > > > > $ make ARCH=um O=../kunit.out/ olddefconfig > > > > Building with: > > > > $ make ARCH=um O=../kunit.out/ --jobs=36 > > > > ERROR:root:/usr/bin/ld: arch/um/os-Linux/main.o: in function `__wrap_malloc': > > > > main.c:(.text+0x10b): undefined reference to `vmalloc' > > > > collect2: error: ld returned 1 exit status > > > > > > > > Haven't looked into the code yet, but reporting first. May I ask your idea? > > > > > > Hi SeongJae, > > > Looks like we missed adding "#include <linux/vmalloc.h>" inside > > > arch/um/os-Linux/main.c in this patch: > > > https://lore.kernel.org/all/20240321163705.3067592-2-surenb@google.com/. > > > I'll be posing fixes for all 0-day issues found over the weekend and > > > will include a fix for this. In the meantime, to work around it you > > > can add that include yourself. Please let me know if the issue still > > > persists after doing that. > > > > Thank you, Suren. The change made the error message disappears. However, it > > introduced another one. > > Ok, let me investigate and I'll try to get a fix for it today evening. Thank you for this kind reply. Nonetheless, this is not blocking some real thing from me. So, no rush. Plese take your time :) Thanks, SJ > Thanks, > Suren. > > > > > $ git diff > > diff --git a/arch/um/os-Linux/main.c b/arch/um/os-Linux/main.c > > index c8a42ecbd7a2..8fe274e9f3a4 100644 > > --- a/arch/um/os-Linux/main.c > > +++ b/arch/um/os-Linux/main.c > > @@ -16,6 +16,7 @@ > > #include <kern_util.h> > > #include <os.h> > > #include <um_malloc.h> > > +#include <linux/vmalloc.h> > > > > #define PGD_BOUND (4 * 1024 * 1024) > > #define STACKSIZE (8 * 1024 * 1024) > > $ > > $ ./tools/testing/kunit/kunit.py run --build_dir ../kunit.out/ > > [10:43:13] Configuring KUnit Kernel ... > > [10:43:13] Building KUnit Kernel ... > > Populating config with: > > $ make ARCH=um O=../kunit.out/ olddefconfig > > Building with: > > $ make ARCH=um O=../kunit.out/ --jobs=36 > > ERROR:root:In file included from .../arch/um/kernel/asm-offsets.c:1: > > .../arch/x86/um/shared/sysdep/kernel-offsets.h:9:6: warning: no previous prototype for ‘foo’ [-Wmissing-prototypes] > > 9 | void foo(void) > > | ^~~ > > In file included from .../include/linux/alloc_tag.h:8, > > from .../include/linux/vmalloc.h:5, > > from .../arch/um/os-Linux/main.c:19: > > .../include/linux/bug.h:5:10: fatal error: asm/bug.h: No such file or directory > > 5 | #include <asm/bug.h> > > | ^~~~~~~~~~~ > > compilation terminated. > > > > > > Thanks, > > SJ > > > > [...] >
On Mon, Mar 25, 2024 at 11:20 AM SeongJae Park <sj@kernel.org> wrote: > > On Mon, 25 Mar 2024 10:59:01 -0700 Suren Baghdasaryan <surenb@google.com> wrote: > > > On Mon, Mar 25, 2024 at 10:49 AM SeongJae Park <sj@kernel.org> wrote: > > > > > > On Mon, 25 Mar 2024 14:56:01 +0000 Suren Baghdasaryan <surenb@google.com> wrote: > > > > > > > On Sat, Mar 23, 2024 at 6:05 PM SeongJae Park <sj@kernel.org> wrote: > > > > > > > > > > Hi Suren and Kent, > > > > > > > > > > On Thu, 21 Mar 2024 09:36:52 -0700 Suren Baghdasaryan <surenb@google.com> wrote: > > > > > > > > > > > From: Kent Overstreet <kent.overstreet@linux.dev> > > > > > > > > > > > > This wrapps all external vmalloc allocation functions with the > > > > > > alloc_hooks() wrapper, and switches internal allocations to _noprof > > > > > > variants where appropriate, for the new memory allocation profiling > > > > > > feature. > > > > > > > > > > I just noticed latest mm-unstable fails running kunit on my machine as below. > > > > > 'git-bisect' says this is the first commit of the failure. > > > > > > > > > > $ ./tools/testing/kunit/kunit.py run --build_dir ../kunit.out/ > > > > > [10:59:53] Configuring KUnit Kernel ... > > > > > [10:59:53] Building KUnit Kernel ... > > > > > Populating config with: > > > > > $ make ARCH=um O=../kunit.out/ olddefconfig > > > > > Building with: > > > > > $ make ARCH=um O=../kunit.out/ --jobs=36 > > > > > ERROR:root:/usr/bin/ld: arch/um/os-Linux/main.o: in function `__wrap_malloc': > > > > > main.c:(.text+0x10b): undefined reference to `vmalloc' > > > > > collect2: error: ld returned 1 exit status > > > > > > > > > > Haven't looked into the code yet, but reporting first. May I ask your idea? > > > > > > > > Hi SeongJae, > > > > Looks like we missed adding "#include <linux/vmalloc.h>" inside > > > > arch/um/os-Linux/main.c in this patch: > > > > https://lore.kernel.org/all/20240321163705.3067592-2-surenb@google.com/. > > > > I'll be posing fixes for all 0-day issues found over the weekend and > > > > will include a fix for this. In the meantime, to work around it you > > > > can add that include yourself. Please let me know if the issue still > > > > persists after doing that. > > > > > > Thank you, Suren. The change made the error message disappears. However, it > > > introduced another one. > > > > Ok, let me investigate and I'll try to get a fix for it today evening. > > Thank you for this kind reply. Nonetheless, this is not blocking some real > thing from me. So, no rush. Plese take your time :) I posted a fix here: https://lore.kernel.org/all/20240326073750.726636-1-surenb@google.com/ Please let me know if this resolves the issue. Thanks, Suren. > > > Thanks, > SJ > > > Thanks, > > Suren. > > > > > > > > $ git diff > > > diff --git a/arch/um/os-Linux/main.c b/arch/um/os-Linux/main.c > > > index c8a42ecbd7a2..8fe274e9f3a4 100644 > > > --- a/arch/um/os-Linux/main.c > > > +++ b/arch/um/os-Linux/main.c > > > @@ -16,6 +16,7 @@ > > > #include <kern_util.h> > > > #include <os.h> > > > #include <um_malloc.h> > > > +#include <linux/vmalloc.h> > > > > > > #define PGD_BOUND (4 * 1024 * 1024) > > > #define STACKSIZE (8 * 1024 * 1024) > > > $ > > > $ ./tools/testing/kunit/kunit.py run --build_dir ../kunit.out/ > > > [10:43:13] Configuring KUnit Kernel ... > > > [10:43:13] Building KUnit Kernel ... > > > Populating config with: > > > $ make ARCH=um O=../kunit.out/ olddefconfig > > > Building with: > > > $ make ARCH=um O=../kunit.out/ --jobs=36 > > > ERROR:root:In file included from .../arch/um/kernel/asm-offsets.c:1: > > > .../arch/x86/um/shared/sysdep/kernel-offsets.h:9:6: warning: no previous prototype for ‘foo’ [-Wmissing-prototypes] > > > 9 | void foo(void) > > > | ^~~ > > > In file included from .../include/linux/alloc_tag.h:8, > > > from .../include/linux/vmalloc.h:5, > > > from .../arch/um/os-Linux/main.c:19: > > > .../include/linux/bug.h:5:10: fatal error: asm/bug.h: No such file or directory > > > 5 | #include <asm/bug.h> > > > | ^~~~~~~~~~~ > > > compilation terminated. > > > > > > > > > Thanks, > > > SJ > > > > > > [...] > > > > -- > To unsubscribe from this group and stop receiving emails from it, send an email to kernel-team+unsubscribe@android.com. >
On Tue, 26 Mar 2024 00:51:21 -0700 Suren Baghdasaryan <surenb@google.com> wrote: > On Mon, Mar 25, 2024 at 11:20 AM SeongJae Park <sj@kernel.org> wrote: > > > > On Mon, 25 Mar 2024 10:59:01 -0700 Suren Baghdasaryan <surenb@google.com> wrote: > > > > > On Mon, Mar 25, 2024 at 10:49 AM SeongJae Park <sj@kernel.org> wrote: > > > > > > > > On Mon, 25 Mar 2024 14:56:01 +0000 Suren Baghdasaryan <surenb@google.com> wrote: > > > > > > > > > On Sat, Mar 23, 2024 at 6:05 PM SeongJae Park <sj@kernel.org> wrote: > > > > > > > > > > > > Hi Suren and Kent, > > > > > > > > > > > > On Thu, 21 Mar 2024 09:36:52 -0700 Suren Baghdasaryan <surenb@google.com> wrote: > > > > > > > > > > > > > From: Kent Overstreet <kent.overstreet@linux.dev> > > > > > > > > > > > > > > This wrapps all external vmalloc allocation functions with the > > > > > > > alloc_hooks() wrapper, and switches internal allocations to _noprof > > > > > > > variants where appropriate, for the new memory allocation profiling > > > > > > > feature. > > > > > > > > > > > > I just noticed latest mm-unstable fails running kunit on my machine as below. > > > > > > 'git-bisect' says this is the first commit of the failure. > > > > > > > > > > > > $ ./tools/testing/kunit/kunit.py run --build_dir ../kunit.out/ > > > > > > [10:59:53] Configuring KUnit Kernel ... > > > > > > [10:59:53] Building KUnit Kernel ... > > > > > > Populating config with: > > > > > > $ make ARCH=um O=../kunit.out/ olddefconfig > > > > > > Building with: > > > > > > $ make ARCH=um O=../kunit.out/ --jobs=36 > > > > > > ERROR:root:/usr/bin/ld: arch/um/os-Linux/main.o: in function `__wrap_malloc': > > > > > > main.c:(.text+0x10b): undefined reference to `vmalloc' > > > > > > collect2: error: ld returned 1 exit status > > > > > > > > > > > > Haven't looked into the code yet, but reporting first. May I ask your idea? > > > > > > > > > > Hi SeongJae, > > > > > Looks like we missed adding "#include <linux/vmalloc.h>" inside > > > > > arch/um/os-Linux/main.c in this patch: > > > > > https://lore.kernel.org/all/20240321163705.3067592-2-surenb@google.com/. > > > > > I'll be posing fixes for all 0-day issues found over the weekend and > > > > > will include a fix for this. In the meantime, to work around it you > > > > > can add that include yourself. Please let me know if the issue still > > > > > persists after doing that. > > > > > > > > Thank you, Suren. The change made the error message disappears. However, it > > > > introduced another one. > > > > > > Ok, let me investigate and I'll try to get a fix for it today evening. > > > > Thank you for this kind reply. Nonetheless, this is not blocking some real > > thing from me. So, no rush. Plese take your time :) > > I posted a fix here: > https://lore.kernel.org/all/20240326073750.726636-1-surenb@google.com/ > Please let me know if this resolves the issue. I confirmed it is fixing the issue, and replied to the patch with my Tested-by: tag. Thank you for this kind fix, Suren. Thanks, SJ [...]
© 2016 - 2026 Red Hat, Inc.