Address some occurrences of shadowing between the global
variable 'e820' in 'xen/arch/x86/e820.c' and function
parameter names (such as that of 'e820_add_range').
The parameter is removed in those functions whose call chain
ultimately supplies the global variable as a parameter, which
is already visible from their definitions in 'e820.c'.
Signed-off-by: Nicola Vetrini <nicola.vetrini@bugseng.com>
---
Changes in v2:
- Reworked the patch to remove the parameter e820 where possible
and consequential changes to the involved functions.
---
xen/arch/x86/e820.c | 36 ++++++++++-----------
xen/arch/x86/guest/hyperv/hyperv.c | 4 +--
xen/arch/x86/guest/hypervisor.c | 2 +-
xen/arch/x86/guest/xen/xen.c | 4 +--
xen/arch/x86/include/asm/e820.h | 3 +-
xen/arch/x86/include/asm/guest/hypervisor.h | 2 +-
xen/arch/x86/include/asm/pv/shim.h | 2 +-
xen/arch/x86/pv/shim.c | 10 +++---
xen/arch/x86/setup.c | 8 ++---
xen/arch/x86/x86_64/mmconf-fam10h.c | 2 +-
xen/drivers/passthrough/amd/iommu_acpi.c | 2 +-
11 files changed, 37 insertions(+), 38 deletions(-)
diff --git a/xen/arch/x86/e820.c b/xen/arch/x86/e820.c
index 4911e64b8c..875572b23e 100644
--- a/xen/arch/x86/e820.c
+++ b/xen/arch/x86/e820.c
@@ -543,27 +543,27 @@ static void __init machine_specific_memory_setup(struct e820map *raw)
clip_to_limit(top_of_ram, "MTRRs do not cover all of memory.");
}
-/* This function relies on the passed in e820->map[] being sorted. */
-int __init e820_add_range(
- struct e820map *e820, uint64_t s, uint64_t e, uint32_t type)
+/* This function relies on the global e820->map[] being sorted. */
+int __init e820_add_range(uint64_t s, uint64_t e, uint32_t type)
{
unsigned int i;
+ struct e820entry *ei = e820.map;
- for ( i = 0; i < e820->nr_map; ++i )
+ for ( i = 0; i < e820.nr_map; ++i )
{
- uint64_t rs = e820->map[i].addr;
- uint64_t re = rs + e820->map[i].size;
+ uint64_t rs = ei[i].addr;
+ uint64_t re = rs + ei[i].size;
- if ( rs == e && e820->map[i].type == type )
+ if ( rs == e && ei[i].type == type )
{
- e820->map[i].addr = s;
+ ei[i].addr = s;
return 1;
}
- if ( re == s && e820->map[i].type == type &&
- (i + 1 == e820->nr_map || e820->map[i + 1].addr >= e) )
+ if ( re == s && ei[i].type == type &&
+ (i + 1 == e820.nr_map || ei[i + 1].addr >= e) )
{
- e820->map[i].size += e - s;
+ ei[i].size += e - s;
return 1;
}
@@ -574,20 +574,20 @@ int __init e820_add_range(
return 0;
}
- if ( e820->nr_map >= ARRAY_SIZE(e820->map) )
+ if ( e820.nr_map >= ARRAY_SIZE(e820.map) )
{
printk(XENLOG_WARNING "E820: overflow while adding region"
" %"PRIx64"-%"PRIx64"\n", s, e);
return 0;
}
- memmove(e820->map + i + 1, e820->map + i,
- (e820->nr_map - i) * sizeof(*e820->map));
+ memmove(ei + i + 1, ei + i,
+ (e820.nr_map - i) * sizeof(*e820.map));
- e820->nr_map++;
- e820->map[i].addr = s;
- e820->map[i].size = e - s;
- e820->map[i].type = type;
+ e820.nr_map++;
+ ei[i].addr = s;
+ ei[i].size = e - s;
+ ei[i].type = type;
return 1;
}
diff --git a/xen/arch/x86/guest/hyperv/hyperv.c b/xen/arch/x86/guest/hyperv/hyperv.c
index aacc7a6167..912099564e 100644
--- a/xen/arch/x86/guest/hyperv/hyperv.c
+++ b/xen/arch/x86/guest/hyperv/hyperv.c
@@ -187,11 +187,11 @@ static int cf_check ap_setup(void)
return setup_vp_assist();
}
-static void __init cf_check e820_fixup(struct e820map *e820)
+static void __init cf_check e820_fixup(void)
{
uint64_t s = HV_HCALL_MFN << PAGE_SHIFT;
- if ( !e820_add_range(e820, s, s + PAGE_SIZE, E820_RESERVED) )
+ if ( !e820_add_range(s, s + PAGE_SIZE, E820_RESERVED) )
panic("Unable to reserve Hyper-V hypercall range\n");
}
diff --git a/xen/arch/x86/guest/hypervisor.c b/xen/arch/x86/guest/hypervisor.c
index b8549a131a..bf5be97bc3 100644
--- a/xen/arch/x86/guest/hypervisor.c
+++ b/xen/arch/x86/guest/hypervisor.c
@@ -63,7 +63,7 @@ void hypervisor_resume(void)
void __init hypervisor_e820_fixup(struct e820map *e820)
{
if ( ops.e820_fixup )
- ops.e820_fixup(e820);
+ ops.e820_fixup();
}
int hypervisor_flush_tlb(const cpumask_t *mask, const void *va,
diff --git a/xen/arch/x86/guest/xen/xen.c b/xen/arch/x86/guest/xen/xen.c
index f93dfc89f7..139489c666 100644
--- a/xen/arch/x86/guest/xen/xen.c
+++ b/xen/arch/x86/guest/xen/xen.c
@@ -306,10 +306,10 @@ static void cf_check resume(void)
pv_console_init();
}
-static void __init cf_check e820_fixup(struct e820map *e820)
+static void __init cf_check e820_fixup()
{
if ( pv_shim )
- pv_shim_fixup_e820(e820);
+ pv_shim_fixup_e820();
}
static int cf_check flush_tlb(
diff --git a/xen/arch/x86/include/asm/e820.h b/xen/arch/x86/include/asm/e820.h
index 213d5b5dd2..af90085d65 100644
--- a/xen/arch/x86/include/asm/e820.h
+++ b/xen/arch/x86/include/asm/e820.h
@@ -29,8 +29,7 @@ extern int reserve_e820_ram(struct e820map *e820, uint64_t s, uint64_t e);
extern int e820_change_range_type(
struct e820map *e820, uint64_t s, uint64_t e,
uint32_t orig_type, uint32_t new_type);
-extern int e820_add_range(
- struct e820map *, uint64_t s, uint64_t e, uint32_t type);
+extern int e820_add_range(uint64_t s, uint64_t e, uint32_t type);
extern unsigned long init_e820(const char *str, struct e820map *raw);
extern void print_e820_memory_map(const struct e820entry *map,
unsigned int entries);
diff --git a/xen/arch/x86/include/asm/guest/hypervisor.h b/xen/arch/x86/include/asm/guest/hypervisor.h
index 4cffea3866..9c8a893da3 100644
--- a/xen/arch/x86/include/asm/guest/hypervisor.h
+++ b/xen/arch/x86/include/asm/guest/hypervisor.h
@@ -22,7 +22,7 @@ struct hypervisor_ops {
/* Resume from suspension */
void (*resume)(void);
/* Fix up e820 map */
- void (*e820_fixup)(struct e820map *e820);
+ void (*e820_fixup)(void);
/* L0 assisted TLB flush */
int (*flush_tlb)(const cpumask_t *mask, const void *va, unsigned int flags);
};
diff --git a/xen/arch/x86/include/asm/pv/shim.h b/xen/arch/x86/include/asm/pv/shim.h
index 5625b90b72..7bae9ae372 100644
--- a/xen/arch/x86/include/asm/pv/shim.h
+++ b/xen/arch/x86/include/asm/pv/shim.h
@@ -85,7 +85,7 @@ static inline uint64_t pv_shim_mem(uint64_t avail)
ASSERT_UNREACHABLE();
return 0;
}
-static inline void pv_shim_fixup_e820(struct e820map *e820)
+static inline void pv_shim_fixup_e820(void)
{
ASSERT_UNREACHABLE();
}
diff --git a/xen/arch/x86/pv/shim.c b/xen/arch/x86/pv/shim.c
index 4044087119..a8883a1ebd 100644
--- a/xen/arch/x86/pv/shim.c
+++ b/xen/arch/x86/pv/shim.c
@@ -98,17 +98,17 @@ uint64_t pv_shim_mem(uint64_t avail)
return shim_nrpages;
}
-static void __init mark_pfn_as_ram(struct e820map *e820, uint64_t pfn)
+static void __init mark_pfn_as_ram(uint64_t pfn)
{
- if ( !e820_add_range(e820, pfn << PAGE_SHIFT,
+ if ( !e820_add_range(pfn << PAGE_SHIFT,
(pfn << PAGE_SHIFT) + PAGE_SIZE, E820_RAM) &&
- !e820_change_range_type(e820, pfn << PAGE_SHIFT,
+ !e820_change_range_type(&e820, pfn << PAGE_SHIFT,
(pfn << PAGE_SHIFT) + PAGE_SIZE,
E820_RESERVED, E820_RAM) )
panic("Unable to add/change memory type of pfn %#lx to RAM\n", pfn);
}
-void __init pv_shim_fixup_e820(struct e820map *e820)
+void __init pv_shim_fixup_e820(void)
{
uint64_t pfn = 0;
unsigned int i = 0;
@@ -120,7 +120,7 @@ void __init pv_shim_fixup_e820(struct e820map *e820)
rc = xen_hypercall_hvm_get_param(p, &pfn); \
if ( rc ) \
panic("Unable to get " #p "\n"); \
- mark_pfn_as_ram(e820, pfn); \
+ mark_pfn_as_ram(pfn); \
ASSERT(i < ARRAY_SIZE(reserved_pages)); \
reserved_pages[i++].mfn = pfn; \
})
diff --git a/xen/arch/x86/setup.c b/xen/arch/x86/setup.c
index 80ae973d64..03f9a03180 100644
--- a/xen/arch/x86/setup.c
+++ b/xen/arch/x86/setup.c
@@ -686,7 +686,7 @@ static void __init parse_video_info(void)
#endif
}
-static void __init kexec_reserve_area(struct e820map *e820)
+static void __init kexec_reserve_area(void)
{
#ifdef CONFIG_KEXEC
unsigned long kdump_start = kexec_crash_area.start;
@@ -700,7 +700,7 @@ static void __init kexec_reserve_area(struct e820map *e820)
is_reserved = true;
- if ( !reserve_e820_ram(e820, kdump_start, kdump_start + kdump_size) )
+ if ( !reserve_e820_ram(&boot_e820, kdump_start, kdump_start + kdump_size) )
{
printk("Kdump: DISABLED (failed to reserve %luMB (%lukB) at %#lx)"
"\n", kdump_size >> 20, kdump_size >> 10, kdump_start);
@@ -1308,7 +1308,7 @@ void __init noreturn __start_xen(unsigned long mbi_p)
if ( e820.map[i].type == E820_RAM )
nr_pages += e820.map[i].size >> PAGE_SHIFT;
set_kexec_crash_area_size((u64)nr_pages << PAGE_SHIFT);
- kexec_reserve_area(&boot_e820);
+ kexec_reserve_area();
initial_images = mod;
nr_initial_images = mbi->mods_count;
@@ -1495,7 +1495,7 @@ void __init noreturn __start_xen(unsigned long mbi_p)
reserve_e820_ram(&boot_e820, __pa(_stext), __pa(__2M_rwdata_end));
/* Late kexec reservation (dynamic start address). */
- kexec_reserve_area(&boot_e820);
+ kexec_reserve_area();
setup_max_pdx(raw_max_page);
if ( highmem_start )
diff --git a/xen/arch/x86/x86_64/mmconf-fam10h.c b/xen/arch/x86/x86_64/mmconf-fam10h.c
index a834ab3149..36b32eb769 100644
--- a/xen/arch/x86/x86_64/mmconf-fam10h.c
+++ b/xen/arch/x86/x86_64/mmconf-fam10h.c
@@ -135,7 +135,7 @@ static void __init get_fam10h_pci_mmconf_base(void)
return;
out:
- if (e820_add_range(&e820, start, start + SIZE, E820_RESERVED))
+ if (e820_add_range(start, start + SIZE, E820_RESERVED))
fam10h_pci_mmconf_base = start;
}
diff --git a/xen/drivers/passthrough/amd/iommu_acpi.c b/xen/drivers/passthrough/amd/iommu_acpi.c
index 3b577c9b39..db993d6df2 100644
--- a/xen/drivers/passthrough/amd/iommu_acpi.c
+++ b/xen/drivers/passthrough/amd/iommu_acpi.c
@@ -418,7 +418,7 @@ static int __init parse_ivmd_block(const struct acpi_ivrs_memory *ivmd_block)
if ( type == RAM_TYPE_UNKNOWN )
{
- if ( e820_add_range(&e820, addr, addr + PAGE_SIZE,
+ if ( e820_add_range(addr, addr + PAGE_SIZE,
E820_RESERVED) )
continue;
AMD_IOMMU_ERROR("IVMD: page at %lx couldn't be reserved\n",
--
2.34.1
On 08.08.2023 13:08, Nicola Vetrini wrote:
> --- a/xen/arch/x86/e820.c
> +++ b/xen/arch/x86/e820.c
> @@ -543,27 +543,27 @@ static void __init machine_specific_memory_setup(struct e820map *raw)
> clip_to_limit(top_of_ram, "MTRRs do not cover all of memory.");
> }
>
> -/* This function relies on the passed in e820->map[] being sorted. */
> -int __init e820_add_range(
> - struct e820map *e820, uint64_t s, uint64_t e, uint32_t type)
> +/* This function relies on the global e820->map[] being sorted. */
> +int __init e820_add_range(uint64_t s, uint64_t e, uint32_t type)
> {
> unsigned int i;
> + struct e820entry *ei = e820.map;
>
> - for ( i = 0; i < e820->nr_map; ++i )
> + for ( i = 0; i < e820.nr_map; ++i )
> {
> - uint64_t rs = e820->map[i].addr;
> - uint64_t re = rs + e820->map[i].size;
> + uint64_t rs = ei[i].addr;
> + uint64_t re = rs + ei[i].size;
>
> - if ( rs == e && e820->map[i].type == type )
> + if ( rs == e && ei[i].type == type )
> {
> - e820->map[i].addr = s;
> + ei[i].addr = s;
> return 1;
> }
>
> - if ( re == s && e820->map[i].type == type &&
> - (i + 1 == e820->nr_map || e820->map[i + 1].addr >= e) )
> + if ( re == s && ei[i].type == type &&
> + (i + 1 == e820.nr_map || ei[i + 1].addr >= e) )
> {
> - e820->map[i].size += e - s;
> + ei[i].size += e - s;
> return 1;
> }
>
> @@ -574,20 +574,20 @@ int __init e820_add_range(
> return 0;
> }
>
> - if ( e820->nr_map >= ARRAY_SIZE(e820->map) )
> + if ( e820.nr_map >= ARRAY_SIZE(e820.map) )
> {
> printk(XENLOG_WARNING "E820: overflow while adding region"
> " %"PRIx64"-%"PRIx64"\n", s, e);
> return 0;
> }
>
> - memmove(e820->map + i + 1, e820->map + i,
> - (e820->nr_map - i) * sizeof(*e820->map));
> + memmove(ei + i + 1, ei + i,
> + (e820.nr_map - i) * sizeof(*e820.map));
>
> - e820->nr_map++;
> - e820->map[i].addr = s;
> - e820->map[i].size = e - s;
> - e820->map[i].type = type;
> + e820.nr_map++;
> + ei[i].addr = s;
> + ei[i].size = e - s;
> + ei[i].type = type;
>
> return 1;
> }
To be honest this isn't quite what I was hoping for; the many ei[i]. are
(imo) quite a bit harder to read than ei-> would have been (hence my
earlier suggestion to also update that pointer in the for() loop header).
Then again I see there is one use of ei[i + 1], which would likely look
less neat as ei[1].addr when everywhere else we have ei->. So I guess up
to you whether you adjust further; I'll ack either form.
> --- a/xen/arch/x86/guest/hypervisor.c
> +++ b/xen/arch/x86/guest/hypervisor.c
> @@ -63,7 +63,7 @@ void hypervisor_resume(void)
> void __init hypervisor_e820_fixup(struct e820map *e820)
What about this one? The function parameter ...
> {
> if ( ops.e820_fixup )
> - ops.e820_fixup(e820);
> + ops.e820_fixup();
> }
... isn't used anymore, and the sole call site passes &e820.
> --- a/xen/arch/x86/include/asm/e820.h
> +++ b/xen/arch/x86/include/asm/e820.h
> @@ -29,8 +29,7 @@ extern int reserve_e820_ram(struct e820map *e820, uint64_t s, uint64_t e);
> extern int e820_change_range_type(
> struct e820map *e820, uint64_t s, uint64_t e,
> uint32_t orig_type, uint32_t new_type);
And what about this one? None of the other subjects in the series suggest
this is then taken care of in a separate patch (as per the earlier
discussion it indeed doesn't want dealing with right here).
> --- a/xen/arch/x86/setup.c
> +++ b/xen/arch/x86/setup.c
> @@ -686,7 +686,7 @@ static void __init parse_video_info(void)
> #endif
> }
>
> -static void __init kexec_reserve_area(struct e820map *e820)
> +static void __init kexec_reserve_area(void)
> {
> #ifdef CONFIG_KEXEC
> unsigned long kdump_start = kexec_crash_area.start;
> @@ -700,7 +700,7 @@ static void __init kexec_reserve_area(struct e820map *e820)
>
> is_reserved = true;
>
> - if ( !reserve_e820_ram(e820, kdump_start, kdump_start + kdump_size) )
> + if ( !reserve_e820_ram(&boot_e820, kdump_start, kdump_start + kdump_size) )
> {
> printk("Kdump: DISABLED (failed to reserve %luMB (%lukB) at %#lx)"
> "\n", kdump_size >> 20, kdump_size >> 10, kdump_start);
> @@ -1308,7 +1308,7 @@ void __init noreturn __start_xen(unsigned long mbi_p)
> if ( e820.map[i].type == E820_RAM )
> nr_pages += e820.map[i].size >> PAGE_SHIFT;
> set_kexec_crash_area_size((u64)nr_pages << PAGE_SHIFT);
> - kexec_reserve_area(&boot_e820);
> + kexec_reserve_area();
>
> initial_images = mod;
> nr_initial_images = mbi->mods_count;
> @@ -1495,7 +1495,7 @@ void __init noreturn __start_xen(unsigned long mbi_p)
> reserve_e820_ram(&boot_e820, __pa(_stext), __pa(__2M_rwdata_end));
>
> /* Late kexec reservation (dynamic start address). */
> - kexec_reserve_area(&boot_e820);
> + kexec_reserve_area();
>
> setup_max_pdx(raw_max_page);
> if ( highmem_start )
Seeing all the knock-on effects for the add_range() change, I think this
separate adjustment would better have been an independent patch.
Jan
On 08/08/2023 15:46, Jan Beulich wrote:
> On 08.08.2023 13:08, Nicola Vetrini wrote:
>> --- a/xen/arch/x86/e820.c
>> +++ b/xen/arch/x86/e820.c
>> @@ -543,27 +543,27 @@ static void __init
>> machine_specific_memory_setup(struct e820map *raw)
>> clip_to_limit(top_of_ram, "MTRRs do not cover all of
>> memory.");
>> }
>>
>> -/* This function relies on the passed in e820->map[] being sorted. */
>> -int __init e820_add_range(
>> - struct e820map *e820, uint64_t s, uint64_t e, uint32_t type)
>> +/* This function relies on the global e820->map[] being sorted. */
>> +int __init e820_add_range(uint64_t s, uint64_t e, uint32_t type)
>> {
>> unsigned int i;
>> + struct e820entry *ei = e820.map;
>>
>> - for ( i = 0; i < e820->nr_map; ++i )
>> + for ( i = 0; i < e820.nr_map; ++i )
>> {
>> - uint64_t rs = e820->map[i].addr;
>> - uint64_t re = rs + e820->map[i].size;
>> + uint64_t rs = ei[i].addr;
>> + uint64_t re = rs + ei[i].size;
>>
>> - if ( rs == e && e820->map[i].type == type )
>> + if ( rs == e && ei[i].type == type )
>> {
>> - e820->map[i].addr = s;
>> + ei[i].addr = s;
>> return 1;
>> }
>>
>> - if ( re == s && e820->map[i].type == type &&
>> - (i + 1 == e820->nr_map || e820->map[i + 1].addr >= e) )
>> + if ( re == s && ei[i].type == type &&
>> + (i + 1 == e820.nr_map || ei[i + 1].addr >= e) )
>> {
>> - e820->map[i].size += e - s;
>> + ei[i].size += e - s;
>> return 1;
>> }
>>
>> @@ -574,20 +574,20 @@ int __init e820_add_range(
>> return 0;
>> }
>>
>> - if ( e820->nr_map >= ARRAY_SIZE(e820->map) )
>> + if ( e820.nr_map >= ARRAY_SIZE(e820.map) )
>> {
>> printk(XENLOG_WARNING "E820: overflow while adding region"
>> " %"PRIx64"-%"PRIx64"\n", s, e);
>> return 0;
>> }
>>
>> - memmove(e820->map + i + 1, e820->map + i,
>> - (e820->nr_map - i) * sizeof(*e820->map));
>> + memmove(ei + i + 1, ei + i,
>> + (e820.nr_map - i) * sizeof(*e820.map));
>>
>> - e820->nr_map++;
>> - e820->map[i].addr = s;
>> - e820->map[i].size = e - s;
>> - e820->map[i].type = type;
>> + e820.nr_map++;
>> + ei[i].addr = s;
>> + ei[i].size = e - s;
>> + ei[i].type = type;
>>
>> return 1;
>> }
>
> To be honest this isn't quite what I was hoping for; the many ei[i].
> are
> (imo) quite a bit harder to read than ei-> would have been (hence my
> earlier suggestion to also update that pointer in the for() loop
> header).
> Then again I see there is one use of ei[i + 1], which would likely look
> less neat as ei[1].addr when everywhere else we have ei->. So I guess
> up
> to you whether you adjust further; I'll ack either form.
>
I'll leave it as is.
>> --- a/xen/arch/x86/guest/hypervisor.c
>> +++ b/xen/arch/x86/guest/hypervisor.c
>> @@ -63,7 +63,7 @@ void hypervisor_resume(void)
>> void __init hypervisor_e820_fixup(struct e820map *e820)
>
> What about this one? The function parameter ...
>
>> {
>> if ( ops.e820_fixup )
>> - ops.e820_fixup(e820);
>> + ops.e820_fixup();
>> }
>
> ... isn't used anymore, and the sole call site passes &e820.
>
It remained there by accident.
>> --- a/xen/arch/x86/include/asm/e820.h
>> +++ b/xen/arch/x86/include/asm/e820.h
>> @@ -29,8 +29,7 @@ extern int reserve_e820_ram(struct e820map *e820,
>> uint64_t s, uint64_t e);
>> extern int e820_change_range_type(
>> struct e820map *e820, uint64_t s, uint64_t e,
>> uint32_t orig_type, uint32_t new_type);
>
> And what about this one? None of the other subjects in the series
> suggest
> this is then taken care of in a separate patch (as per the earlier
> discussion it indeed doesn't want dealing with right here).
>
I'll mention this detail. While I work on other rules I'll think of a
good way to rename.
>> --- a/xen/arch/x86/setup.c
>> +++ b/xen/arch/x86/setup.c
>> @@ -686,7 +686,7 @@ static void __init parse_video_info(void)
>> #endif
>> }
>>
>> -static void __init kexec_reserve_area(struct e820map *e820)
>> +static void __init kexec_reserve_area(void)
>> {
>> #ifdef CONFIG_KEXEC
>> unsigned long kdump_start = kexec_crash_area.start;
>> @@ -700,7 +700,7 @@ static void __init kexec_reserve_area(struct
>> e820map *e820)
>>
>> is_reserved = true;
>>
>> - if ( !reserve_e820_ram(e820, kdump_start, kdump_start +
>> kdump_size) )
>> + if ( !reserve_e820_ram(&boot_e820, kdump_start, kdump_start +
>> kdump_size) )
>> {
>> printk("Kdump: DISABLED (failed to reserve %luMB (%lukB) at
>> %#lx)"
>> "\n", kdump_size >> 20, kdump_size >> 10,
>> kdump_start);
>> @@ -1308,7 +1308,7 @@ void __init noreturn __start_xen(unsigned long
>> mbi_p)
>> if ( e820.map[i].type == E820_RAM )
>> nr_pages += e820.map[i].size >> PAGE_SHIFT;
>> set_kexec_crash_area_size((u64)nr_pages << PAGE_SHIFT);
>> - kexec_reserve_area(&boot_e820);
>> + kexec_reserve_area();
>>
>> initial_images = mod;
>> nr_initial_images = mbi->mods_count;
>> @@ -1495,7 +1495,7 @@ void __init noreturn __start_xen(unsigned long
>> mbi_p)
>> reserve_e820_ram(&boot_e820, __pa(_stext),
>> __pa(__2M_rwdata_end));
>>
>> /* Late kexec reservation (dynamic start address). */
>> - kexec_reserve_area(&boot_e820);
>> + kexec_reserve_area();
>>
>> setup_max_pdx(raw_max_page);
>> if ( highmem_start )
>
> Seeing all the knock-on effects for the add_range() change, I think
> this
> separate adjustment would better have been an independent patch.
>
> Jan
I can submit it standalone and put together x86/vmsi and delay
--
Nicola Vetrini, BSc
Software Engineer, BUGSENG srl (https://bugseng.com)
© 2016 - 2026 Red Hat, Inc.