mirror of
https://github.com/torvalds/linux.git
synced 2026-03-08 02:24:32 +01:00
arch, mm: consolidate initialization of nodes, zones and memory map
To initialize node, zone and memory map data structures every architecture calls free_area_init() during setup_arch() and passes it an array of zone limits. Beside code duplication it creates "interesting" ordering cases between allocation and initialization of hugetlb and the memory map. Some architectures allocate hugetlb pages very early in setup_arch() in certain cases, some only create hugetlb CMA areas in setup_arch() and sometimes hugetlb allocations happen mm_core_init(). With arch_zone_limits_init() helper available now on all architectures it is no longer necessary to call free_area_init() from architecture setup code. Rather core MM initialization can call arch_zone_limits_init() in a single place. This allows to unify ordering of hugetlb vs memory map allocation and initialization. Remove the call to free_area_init() from architecture specific code and place it in a new mm_core_init_early() function that is called immediately after setup_arch(). After this refactoring it is possible to consolidate hugetlb allocations and eliminate differences in ordering of hugetlb and memory map initialization among different architectures. As the first step of this consolidation move hugetlb_bootmem_alloc() to mm_core_early_init(). Link: https://lkml.kernel.org/r/20260111082105.290734-24-rppt@kernel.org Signed-off-by: Mike Rapoport (Microsoft) <rppt@kernel.org> Cc: Alexander Gordeev <agordeev@linux.ibm.com> Cc: Alex Shi <alexs@kernel.org> Cc: Andreas Larsson <andreas@gaisler.com> Cc: "Borislav Petkov (AMD)" <bp@alien8.de> Cc: Catalin Marinas <catalin.marinas@arm.com> Cc: David Hildenbrand <david@kernel.org> Cc: David S. Miller <davem@davemloft.net> Cc: Dinh Nguyen <dinguyen@kernel.org> Cc: Geert Uytterhoeven <geert@linux-m68k.org> Cc: Guo Ren <guoren@kernel.org> Cc: Heiko Carstens <hca@linux.ibm.com> Cc: Helge Deller <deller@gmx.de> Cc: Huacai Chen <chenhuacai@kernel.org> Cc: Ingo Molnar <mingo@redhat.com> Cc: Johannes Berg <johannes@sipsolutions.net> Cc: John Paul Adrian Glaubitz <glaubitz@physik.fu-berlin.de> Cc: Jonathan Corbet <corbet@lwn.net> Cc: Klara Modin <klarasmodin@gmail.com> Cc: Liam Howlett <liam.howlett@oracle.com> Cc: Lorenzo Stoakes <lorenzo.stoakes@oracle.com> Cc: Magnus Lindholm <linmag7@gmail.com> Cc: Matt Turner <mattst88@gmail.com> Cc: Max Filippov <jcmvbkbc@gmail.com> Cc: Michael Ellerman <mpe@ellerman.id.au> Cc: Michal Hocko <mhocko@suse.com> Cc: Michal Simek <monstr@monstr.eu> Cc: Muchun Song <muchun.song@linux.dev> Cc: Oscar Salvador <osalvador@suse.de> Cc: Palmer Dabbelt <palmer@dabbelt.com> Cc: Pratyush Yadav <pratyush@kernel.org> Cc: Richard Weinberger <richard@nod.at> Cc: "Ritesh Harjani (IBM)" <ritesh.list@gmail.com> Cc: Russell King <linux@armlinux.org.uk> Cc: Stafford Horne <shorne@gmail.com> Cc: Suren Baghdasaryan <surenb@google.com> Cc: Thomas Bogendoerfer <tsbogend@alpha.franken.de> Cc: Thomas Gleixner <tglx@linutronix.de> Cc: Vasily Gorbik <gor@linux.ibm.com> Cc: Vineet Gupta <vgupta@kernel.org> Cc: Vlastimil Babka <vbabka@suse.cz> Cc: Will Deacon <will@kernel.org> Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
This commit is contained in:
parent
2d3c8c5f33
commit
d49004c5f0
35 changed files with 15 additions and 200 deletions
|
|
@ -220,17 +220,10 @@ void __init arch_zone_limits_init(unsigned long *max_zone_pfn)
|
|||
}
|
||||
|
||||
/*
|
||||
* paging_init() sets up the memory map.
|
||||
* paging_init() initializes the kernel's ZERO_PGE.
|
||||
*/
|
||||
void __init paging_init(void)
|
||||
{
|
||||
unsigned long max_zone_pfn[MAX_NR_ZONES] = {0, };
|
||||
|
||||
/* Initialize mem_map[]. */
|
||||
arch_zone_limits_init(max_zone_pfn);
|
||||
free_area_init(max_zone_pfn);
|
||||
|
||||
/* Initialize the kernel's ZERO_PGE. */
|
||||
memset(absolute_pointer(ZERO_PGE), 0, PAGE_SIZE);
|
||||
}
|
||||
|
||||
|
|
|
|||
|
|
@ -102,8 +102,6 @@ void __init arch_zone_limits_init(unsigned long *max_zone_pfn)
|
|||
*/
|
||||
void __init setup_arch_memory(void)
|
||||
{
|
||||
unsigned long max_zone_pfn[MAX_NR_ZONES] = { 0 };
|
||||
|
||||
setup_initial_init_mm(_text, _etext, _edata, _end);
|
||||
|
||||
/* first page of system - kernel .vector starts here */
|
||||
|
|
@ -158,9 +156,6 @@ void __init setup_arch_memory(void)
|
|||
arch_pfn_offset = min(min_low_pfn, min_high_pfn);
|
||||
kmap_init();
|
||||
#endif /* CONFIG_HIGHMEM */
|
||||
|
||||
arch_zone_limits_init(max_zone_pfn);
|
||||
free_area_init(max_zone_pfn);
|
||||
}
|
||||
|
||||
void __init arch_mm_preinit(void)
|
||||
|
|
|
|||
|
|
@ -118,15 +118,6 @@ void __init arch_zone_limits_init(unsigned long *max_zone_pfn)
|
|||
#endif
|
||||
}
|
||||
|
||||
static void __init zone_sizes_init(unsigned long min, unsigned long max_low,
|
||||
unsigned long max_high)
|
||||
{
|
||||
unsigned long max_zone_pfn[MAX_NR_ZONES] = { 0 };
|
||||
|
||||
arch_zone_limits_init(max_zone_pfn);
|
||||
free_area_init(max_zone_pfn);
|
||||
}
|
||||
|
||||
#ifdef CONFIG_HAVE_ARCH_PFN_VALID
|
||||
int pfn_valid(unsigned long pfn)
|
||||
{
|
||||
|
|
@ -222,13 +213,6 @@ void __init bootmem_init(void)
|
|||
* done after the fixed reservations
|
||||
*/
|
||||
sparse_init();
|
||||
|
||||
/*
|
||||
* Now free the memory - free_area_init needs
|
||||
* the sparse mem_map arrays initialized by sparse_init()
|
||||
* for memmap_init_zone(), otherwise all PFNs are invalid.
|
||||
*/
|
||||
zone_sizes_init(min_low_pfn, max_low_pfn, max_pfn);
|
||||
}
|
||||
|
||||
/*
|
||||
|
|
|
|||
|
|
@ -134,7 +134,6 @@ void __init arch_zone_limits_init(unsigned long *max_zone_pfns)
|
|||
|
||||
static void __init dma_limits_init(void)
|
||||
{
|
||||
unsigned long max_zone_pfns[MAX_NR_ZONES] = {0};
|
||||
phys_addr_t __maybe_unused acpi_zone_dma_limit;
|
||||
phys_addr_t __maybe_unused dt_zone_dma_limit;
|
||||
phys_addr_t __maybe_unused dma32_phys_limit =
|
||||
|
|
@ -160,9 +159,6 @@ static void __init dma_limits_init(void)
|
|||
#endif
|
||||
if (!arm64_dma_phys_limit)
|
||||
arm64_dma_phys_limit = PHYS_MASK + 1;
|
||||
|
||||
arch_zone_limits_init(max_zone_pfns);
|
||||
free_area_init(max_zone_pfns);
|
||||
}
|
||||
|
||||
int pfn_is_map_memory(unsigned long pfn)
|
||||
|
|
|
|||
|
|
@ -63,7 +63,6 @@ static void __init csky_memblock_init(void)
|
|||
{
|
||||
unsigned long lowmem_size = PFN_DOWN(LOWMEM_LIMIT - PHYS_OFFSET_OFFSET);
|
||||
unsigned long sseg_size = PFN_DOWN(SSEG_SIZE - PHYS_OFFSET_OFFSET);
|
||||
unsigned long max_zone_pfn[MAX_NR_ZONES] = { 0 };
|
||||
signed long size;
|
||||
|
||||
memblock_reserve(__pa(_start), _end - _start);
|
||||
|
|
@ -101,9 +100,6 @@ static void __init csky_memblock_init(void)
|
|||
memblock_set_current_limit(PFN_PHYS(max_low_pfn));
|
||||
|
||||
dma_contiguous_reserve(0);
|
||||
|
||||
arch_zone_limits_init(max_zone_pfn);
|
||||
free_area_init(max_zone_pfn);
|
||||
}
|
||||
|
||||
void __init setup_arch(char **cmdline_p)
|
||||
|
|
|
|||
|
|
@ -66,20 +66,8 @@ void __init arch_zone_limits_init(unsigned long *max_zone_pfns)
|
|||
max_zone_pfns[ZONE_NORMAL] = max_low_pfn;
|
||||
}
|
||||
|
||||
/*
|
||||
* In order to set up page allocator "nodes",
|
||||
* somebody has to call free_area_init() for UMA.
|
||||
*
|
||||
* In this mode, we only have one pg_data_t
|
||||
* structure: contig_mem_data.
|
||||
*/
|
||||
static void __init paging_init(void)
|
||||
{
|
||||
unsigned long max_zone_pfn[MAX_NR_ZONES] = {0, };
|
||||
|
||||
arch_zone_limits_init(max_zone_pfn);
|
||||
free_area_init(max_zone_pfn); /* sets up the zonelists and mem_map */
|
||||
|
||||
/*
|
||||
* Set the init_mm descriptors "context" value to point to the
|
||||
* initial kernel segment table's physical address.
|
||||
|
|
|
|||
|
|
@ -353,8 +353,6 @@ static inline pte_t pte_swp_clear_exclusive(pte_t pte)
|
|||
return pte;
|
||||
}
|
||||
|
||||
extern void paging_init(void);
|
||||
|
||||
#define pte_none(pte) (!(pte_val(pte) & ~_PAGE_GLOBAL))
|
||||
#define pte_present(pte) (pte_val(pte) & (_PAGE_PRESENT | _PAGE_PROTNONE))
|
||||
#define pte_no_exec(pte) (pte_val(pte) & _PAGE_NO_EXEC)
|
||||
|
|
|
|||
|
|
@ -621,8 +621,6 @@ void __init setup_arch(char **cmdline_p)
|
|||
prefill_possible_map();
|
||||
#endif
|
||||
|
||||
paging_init();
|
||||
|
||||
#ifdef CONFIG_KASAN
|
||||
kasan_init();
|
||||
#endif
|
||||
|
|
|
|||
|
|
@ -68,14 +68,6 @@ void __init arch_zone_limits_init(unsigned long *max_zone_pfns)
|
|||
max_zone_pfns[ZONE_NORMAL] = max_low_pfn;
|
||||
}
|
||||
|
||||
void __init paging_init(void)
|
||||
{
|
||||
unsigned long max_zone_pfns[MAX_NR_ZONES];
|
||||
|
||||
arch_zone_limits_init(max_zone_pfns);
|
||||
free_area_init(max_zone_pfns);
|
||||
}
|
||||
|
||||
void __ref free_initmem(void)
|
||||
{
|
||||
free_initmem_default(POISON_FREE_INITMEM);
|
||||
|
|
|
|||
|
|
@ -69,13 +69,10 @@ void __init paging_init(void)
|
|||
* page_alloc get different views of the world.
|
||||
*/
|
||||
unsigned long end_mem = memory_end & PAGE_MASK;
|
||||
unsigned long max_zone_pfn[MAX_NR_ZONES] = { 0, };
|
||||
|
||||
high_memory = (void *) end_mem;
|
||||
|
||||
empty_zero_page = memblock_alloc_or_panic(PAGE_SIZE, PAGE_SIZE);
|
||||
arch_zone_limits_init(max_zone_pfn);
|
||||
free_area_init(max_zone_pfn);
|
||||
}
|
||||
|
||||
#endif /* CONFIG_MMU */
|
||||
|
|
|
|||
|
|
@ -39,7 +39,6 @@ void __init paging_init(void)
|
|||
pte_t *pg_table;
|
||||
unsigned long address, size;
|
||||
unsigned long next_pgtable;
|
||||
unsigned long max_zone_pfn[MAX_NR_ZONES] = { 0 };
|
||||
int i;
|
||||
|
||||
empty_zero_page = memblock_alloc_or_panic(PAGE_SIZE, PAGE_SIZE);
|
||||
|
|
@ -73,8 +72,6 @@ void __init paging_init(void)
|
|||
}
|
||||
|
||||
current->mm = NULL;
|
||||
arch_zone_limits_init(max_zone_pfn);
|
||||
free_area_init(max_zone_pfn);
|
||||
}
|
||||
|
||||
int cf_tlb_miss(struct pt_regs *regs, int write, int dtlb, int extension_word)
|
||||
|
|
|
|||
|
|
@ -429,7 +429,6 @@ DECLARE_VM_GET_PAGE_PROT
|
|||
*/
|
||||
void __init paging_init(void)
|
||||
{
|
||||
unsigned long max_zone_pfn[MAX_NR_ZONES] = { 0, };
|
||||
unsigned long min_addr, max_addr;
|
||||
unsigned long addr;
|
||||
int i;
|
||||
|
|
@ -511,12 +510,9 @@ void __init paging_init(void)
|
|||
set_fc(USER_DATA);
|
||||
|
||||
#ifdef DEBUG
|
||||
printk ("before free_area_init\n");
|
||||
printk ("before node_set_state\n");
|
||||
#endif
|
||||
for (i = 0; i < m68k_num_memory; i++)
|
||||
if (node_present_pages(i))
|
||||
node_set_state(i, N_NORMAL_MEMORY);
|
||||
|
||||
arch_zone_limits_init(max_zone_pfn);
|
||||
free_area_init(max_zone_pfn);
|
||||
}
|
||||
|
|
|
|||
|
|
@ -41,7 +41,6 @@ void __init paging_init(void)
|
|||
unsigned long address;
|
||||
unsigned long next_pgtable;
|
||||
unsigned long bootmem_end;
|
||||
unsigned long max_zone_pfn[MAX_NR_ZONES] = { 0, };
|
||||
unsigned long size;
|
||||
|
||||
empty_zero_page = memblock_alloc_or_panic(PAGE_SIZE, PAGE_SIZE);
|
||||
|
|
@ -80,14 +79,6 @@ void __init paging_init(void)
|
|||
mmu_emu_init(bootmem_end);
|
||||
|
||||
current->mm = NULL;
|
||||
|
||||
/* memory sizing is a hack stolen from motorola.c.. hope it works for us */
|
||||
arch_zone_limits_init(max_zone_pfn);
|
||||
|
||||
/* I really wish I knew why the following change made things better... -- Sam */
|
||||
free_area_init(max_zone_pfn);
|
||||
|
||||
|
||||
}
|
||||
|
||||
static const pgprot_t protection_map[16] = {
|
||||
|
|
|
|||
|
|
@ -69,22 +69,15 @@ void __init arch_zone_limits_init(unsigned long *max_zone_pfns)
|
|||
*/
|
||||
static void __init paging_init(void)
|
||||
{
|
||||
unsigned long zones_size[MAX_NR_ZONES];
|
||||
int idx;
|
||||
|
||||
/* Setup fixmaps */
|
||||
for (idx = 0; idx < __end_of_fixed_addresses; idx++)
|
||||
clear_fixmap(idx);
|
||||
|
||||
/* Clean every zones */
|
||||
memset(zones_size, 0, sizeof(zones_size));
|
||||
|
||||
#ifdef CONFIG_HIGHMEM
|
||||
highmem_init();
|
||||
#endif
|
||||
arch_zone_limits_init(zones_size);
|
||||
/* We don't have holes in memory map */
|
||||
free_area_init(zones_size);
|
||||
}
|
||||
|
||||
void __init setup_memory(void)
|
||||
|
|
|
|||
|
|
@ -162,11 +162,7 @@ void __init arch_zone_limits_init(unsigned long *max_zone_pfns)
|
|||
|
||||
void __init paging_init(void)
|
||||
{
|
||||
unsigned long zones_size[MAX_NR_ZONES] = {0, };
|
||||
|
||||
pagetable_init();
|
||||
arch_zone_limits_init(zones_size);
|
||||
free_area_init(zones_size);
|
||||
}
|
||||
|
||||
/* All PCI device belongs to logical Node-0 */
|
||||
|
|
|
|||
|
|
@ -417,12 +417,7 @@ void __init arch_zone_limits_init(unsigned long *max_zone_pfns)
|
|||
|
||||
void __init paging_init(void)
|
||||
{
|
||||
unsigned long max_zone_pfns[MAX_NR_ZONES];
|
||||
|
||||
pagetable_init();
|
||||
|
||||
arch_zone_limits_init(max_zone_pfns);
|
||||
free_area_init(max_zone_pfns);
|
||||
}
|
||||
|
||||
#ifdef CONFIG_64BIT
|
||||
|
|
|
|||
|
|
@ -413,9 +413,5 @@ void __init arch_zone_limits_init(unsigned long *max_zone_pfns)
|
|||
|
||||
void __init paging_init(void)
|
||||
{
|
||||
unsigned long zones_size[MAX_NR_ZONES] = {0, };
|
||||
|
||||
pagetable_init();
|
||||
arch_zone_limits_init(zones_size);
|
||||
free_area_init(zones_size);
|
||||
}
|
||||
|
|
|
|||
|
|
@ -51,15 +51,9 @@ void __init arch_zone_limits_init(unsigned long *max_zone_pfns)
|
|||
*/
|
||||
void __init paging_init(void)
|
||||
{
|
||||
unsigned long max_zone_pfn[MAX_NR_ZONES] = { 0 };
|
||||
|
||||
pagetable_init();
|
||||
pgd_current = swapper_pg_dir;
|
||||
|
||||
arch_zone_limits_init(max_zone_pfn);
|
||||
/* pass the memory from the bootmem allocator to the main allocator */
|
||||
free_area_init(max_zone_pfn);
|
||||
|
||||
flush_dcache_range((unsigned long)empty_zero_page,
|
||||
(unsigned long)empty_zero_page + PAGE_SIZE);
|
||||
}
|
||||
|
|
|
|||
|
|
@ -47,14 +47,6 @@ void __init arch_zone_limits_init(unsigned long *max_zone_pfns)
|
|||
max_zone_pfns[ZONE_NORMAL] = max_low_pfn;
|
||||
}
|
||||
|
||||
static void __init zone_sizes_init(void)
|
||||
{
|
||||
unsigned long max_zone_pfn[MAX_NR_ZONES] = { 0 };
|
||||
|
||||
arch_zone_limits_init(max_zone_pfn);
|
||||
free_area_init(max_zone_pfn);
|
||||
}
|
||||
|
||||
extern const char _s_kernel_ro[], _e_kernel_ro[];
|
||||
|
||||
/*
|
||||
|
|
@ -145,8 +137,6 @@ void __init paging_init(void)
|
|||
|
||||
map_ram();
|
||||
|
||||
zone_sizes_init();
|
||||
|
||||
/* self modifying code ;) */
|
||||
/* Since the old TLB miss handler has been running up until now,
|
||||
* the kernel pages are still all RW, so we can still modify the
|
||||
|
|
|
|||
|
|
@ -698,14 +698,6 @@ void __init arch_zone_limits_init(unsigned long *max_zone_pfns)
|
|||
max_zone_pfns[ZONE_NORMAL] = PFN_DOWN(memblock_end_of_DRAM());
|
||||
}
|
||||
|
||||
static void __init parisc_bootmem_free(void)
|
||||
{
|
||||
unsigned long max_zone_pfn[MAX_NR_ZONES] = { 0, };
|
||||
|
||||
arch_zone_limits_init(max_zone_pfn);
|
||||
free_area_init(max_zone_pfn);
|
||||
}
|
||||
|
||||
void __init paging_init(void)
|
||||
{
|
||||
setup_bootmem();
|
||||
|
|
@ -716,7 +708,6 @@ void __init paging_init(void)
|
|||
flush_tlb_all_local(NULL);
|
||||
|
||||
sparse_init();
|
||||
parisc_bootmem_free();
|
||||
}
|
||||
|
||||
static void alloc_btlb(unsigned long start, unsigned long end, int *slot,
|
||||
|
|
|
|||
|
|
@ -237,7 +237,6 @@ void __init arch_zone_limits_init(unsigned long *max_zone_pfns)
|
|||
*/
|
||||
void __init paging_init(void)
|
||||
{
|
||||
unsigned long max_zone_pfns[MAX_NR_ZONES] = { 0 };
|
||||
unsigned long long total_ram = memblock_phys_mem_size();
|
||||
phys_addr_t top_of_ram = memblock_end_of_DRAM();
|
||||
int zone_dma_bits;
|
||||
|
|
@ -269,9 +268,6 @@ void __init paging_init(void)
|
|||
|
||||
zone_dma_limit = DMA_BIT_MASK(zone_dma_bits);
|
||||
|
||||
arch_zone_limits_init(max_zone_pfns);
|
||||
free_area_init(max_zone_pfns);
|
||||
|
||||
mark_nonram_nosave();
|
||||
}
|
||||
|
||||
|
|
|
|||
|
|
@ -87,14 +87,6 @@ void __init arch_zone_limits_init(unsigned long *max_zone_pfns)
|
|||
max_zone_pfns[ZONE_NORMAL] = max_low_pfn;
|
||||
}
|
||||
|
||||
static void __init zone_sizes_init(void)
|
||||
{
|
||||
unsigned long max_zone_pfns[MAX_NR_ZONES] = { 0, };
|
||||
|
||||
arch_zone_limits_init(max_zone_pfns);
|
||||
free_area_init(max_zone_pfns);
|
||||
}
|
||||
|
||||
#if defined(CONFIG_MMU) && defined(CONFIG_DEBUG_VM)
|
||||
|
||||
#define LOG2_SZ_1K ilog2(SZ_1K)
|
||||
|
|
@ -1443,7 +1435,6 @@ void __init misc_mem_init(void)
|
|||
/* The entire VMEMMAP region has been populated. Flush TLB for this region */
|
||||
local_flush_tlb_kernel_range(VMEMMAP_START, VMEMMAP_END);
|
||||
#endif
|
||||
zone_sizes_init();
|
||||
arch_reserve_crashkernel();
|
||||
memblock_dump_all();
|
||||
}
|
||||
|
|
|
|||
|
|
@ -97,14 +97,9 @@ void __init arch_zone_limits_init(unsigned long *max_zone_pfns)
|
|||
*/
|
||||
void __init paging_init(void)
|
||||
{
|
||||
unsigned long max_zone_pfns[MAX_NR_ZONES];
|
||||
|
||||
vmem_map_init();
|
||||
sparse_init();
|
||||
zone_dma_limit = DMA_BIT_MASK(31);
|
||||
memset(max_zone_pfns, 0, sizeof(max_zone_pfns));
|
||||
arch_zone_limits_init(max_zone_pfns);
|
||||
free_area_init(max_zone_pfns);
|
||||
}
|
||||
|
||||
void mark_rodata_ro(void)
|
||||
|
|
|
|||
|
|
@ -271,7 +271,6 @@ void __init arch_zone_limits_init(unsigned long *max_zone_pfns)
|
|||
|
||||
void __init paging_init(void)
|
||||
{
|
||||
unsigned long max_zone_pfns[MAX_NR_ZONES];
|
||||
unsigned long vaddr, end;
|
||||
|
||||
sh_mv.mv_mem_init();
|
||||
|
|
@ -325,10 +324,6 @@ void __init paging_init(void)
|
|||
page_table_range_init(vaddr, end, swapper_pg_dir);
|
||||
|
||||
kmap_coherent_init();
|
||||
|
||||
memset(max_zone_pfns, 0, sizeof(max_zone_pfns));
|
||||
arch_zone_limits_init(max_zone_pfns);
|
||||
free_area_init(max_zone_pfns);
|
||||
}
|
||||
|
||||
unsigned int mem_init_done = 0;
|
||||
|
|
|
|||
|
|
@ -2459,17 +2459,6 @@ void __init paging_init(void)
|
|||
|
||||
kernel_physical_mapping_init();
|
||||
|
||||
{
|
||||
unsigned long max_zone_pfns[MAX_NR_ZONES];
|
||||
|
||||
memset(max_zone_pfns, 0, sizeof(max_zone_pfns));
|
||||
|
||||
max_zone_pfns[ZONE_NORMAL] = end_pfn;
|
||||
|
||||
arch_zone_limits_init(max_zone_pfns);
|
||||
free_area_init(max_zone_pfns);
|
||||
}
|
||||
|
||||
printk("Booting Linux...\n");
|
||||
}
|
||||
|
||||
|
|
|
|||
|
|
@ -970,13 +970,6 @@ void __init srmmu_paging_init(void)
|
|||
flush_tlb_all();
|
||||
|
||||
sparc_context_init(num_contexts);
|
||||
|
||||
{
|
||||
unsigned long max_zone_pfn[MAX_NR_ZONES] = { 0 };
|
||||
|
||||
arch_zone_limits_init(max_zone_pfn);
|
||||
free_area_init(max_zone_pfn);
|
||||
}
|
||||
}
|
||||
|
||||
void mmu_info(struct seq_file *m)
|
||||
|
|
|
|||
|
|
@ -91,16 +91,11 @@ void __init arch_zone_limits_init(unsigned long *max_zone_pfns)
|
|||
|
||||
void __init paging_init(void)
|
||||
{
|
||||
unsigned long max_zone_pfn[MAX_NR_ZONES] = { 0 };
|
||||
|
||||
empty_zero_page = (unsigned long *) memblock_alloc_low(PAGE_SIZE,
|
||||
PAGE_SIZE);
|
||||
if (!empty_zero_page)
|
||||
panic("%s: Failed to allocate %lu bytes align=%lx\n",
|
||||
__func__, PAGE_SIZE, PAGE_SIZE);
|
||||
|
||||
arch_zone_limits_init(max_zone_pfn);
|
||||
free_area_init(max_zone_pfn);
|
||||
}
|
||||
|
||||
/*
|
||||
|
|
|
|||
|
|
@ -1011,16 +1011,6 @@ void __init arch_zone_limits_init(unsigned long *max_zone_pfns)
|
|||
#endif
|
||||
}
|
||||
|
||||
void __init zone_sizes_init(void)
|
||||
{
|
||||
unsigned long max_zone_pfns[MAX_NR_ZONES];
|
||||
|
||||
memset(max_zone_pfns, 0, sizeof(max_zone_pfns));
|
||||
|
||||
arch_zone_limits_init(max_zone_pfns);
|
||||
free_area_init(max_zone_pfns);
|
||||
}
|
||||
|
||||
__visible DEFINE_PER_CPU_ALIGNED(struct tlb_state, cpu_tlbstate) = {
|
||||
.loaded_mm = &init_mm,
|
||||
.next_asid = 1,
|
||||
|
|
|
|||
|
|
@ -655,7 +655,6 @@ void __init paging_init(void)
|
|||
*/
|
||||
olpc_dt_build_devicetree();
|
||||
sparse_init();
|
||||
zone_sizes_init();
|
||||
}
|
||||
|
||||
/*
|
||||
|
|
|
|||
|
|
@ -843,8 +843,6 @@ void __init paging_init(void)
|
|||
*/
|
||||
node_clear_state(0, N_MEMORY);
|
||||
node_clear_state(0, N_NORMAL_MEMORY);
|
||||
|
||||
zone_sizes_init();
|
||||
}
|
||||
|
||||
#define PAGE_UNUSED 0xFD
|
||||
|
|
|
|||
|
|
@ -17,7 +17,6 @@ unsigned long kernel_physical_mapping_init(unsigned long start,
|
|||
unsigned long kernel_physical_mapping_change(unsigned long start,
|
||||
unsigned long end,
|
||||
unsigned long page_size_mask);
|
||||
void zone_sizes_init(void);
|
||||
|
||||
extern int after_bootmem;
|
||||
|
||||
|
|
|
|||
|
|
@ -126,10 +126,6 @@ void __init arch_zone_limits_init(unsigned long *max_zone_pfns)
|
|||
|
||||
void __init zones_init(void)
|
||||
{
|
||||
unsigned long max_zone_pfn[MAX_NR_ZONES] = { 0, };
|
||||
|
||||
arch_zone_limits_init(max_zone_pfn);
|
||||
free_area_init(max_zone_pfn);
|
||||
print_vm_layout();
|
||||
}
|
||||
|
||||
|
|
|
|||
|
|
@ -45,6 +45,7 @@ struct pt_regs;
|
|||
struct folio_batch;
|
||||
|
||||
void arch_mm_preinit(void);
|
||||
void mm_core_init_early(void);
|
||||
void mm_core_init(void);
|
||||
void init_mm_internals(void);
|
||||
|
||||
|
|
@ -3540,7 +3541,7 @@ static inline unsigned long get_num_physpages(void)
|
|||
}
|
||||
|
||||
/*
|
||||
* Using memblock node mappings, an architecture may initialise its
|
||||
* FIXME: Using memblock node mappings, an architecture may initialise its
|
||||
* zones, allocate the backing mem_map and account for memory holes in an
|
||||
* architecture independent manner.
|
||||
*
|
||||
|
|
@ -3555,7 +3556,6 @@ static inline unsigned long get_num_physpages(void)
|
|||
* memblock_add_node(base, size, nid, MEMBLOCK_NONE)
|
||||
* free_area_init(max_zone_pfns);
|
||||
*/
|
||||
void free_area_init(unsigned long *max_zone_pfn);
|
||||
void arch_zone_limits_init(unsigned long *max_zone_pfn);
|
||||
unsigned long node_map_pfn_alignment(void);
|
||||
extern unsigned long absent_pages_in_range(unsigned long start_pfn,
|
||||
|
|
|
|||
|
|
@ -1025,6 +1025,7 @@ void start_kernel(void)
|
|||
page_address_init();
|
||||
pr_notice("%s", linux_banner);
|
||||
setup_arch(&command_line);
|
||||
mm_core_init_early();
|
||||
/* Static keys and static calls are needed by LSMs */
|
||||
jump_label_init();
|
||||
static_call_init();
|
||||
|
|
|
|||
18
mm/mm_init.c
18
mm/mm_init.c
|
|
@ -1807,7 +1807,6 @@ static void __init set_high_memory(void)
|
|||
|
||||
/**
|
||||
* free_area_init - Initialise all pg_data_t and zone data
|
||||
* @max_zone_pfn: an array of max PFNs for each zone
|
||||
*
|
||||
* This will call free_area_init_node() for each active node in the system.
|
||||
* Using the page ranges provided by memblock_set_node(), the size of each
|
||||
|
|
@ -1818,17 +1817,14 @@ static void __init set_high_memory(void)
|
|||
* starts where the previous one ended. For example, ZONE_DMA32 starts
|
||||
* at arch_max_dma_pfn.
|
||||
*/
|
||||
void __init free_area_init(unsigned long *max_zone_pfn)
|
||||
static void __init free_area_init(void)
|
||||
{
|
||||
unsigned long max_zone_pfn[MAX_NR_ZONES] = { 0 };
|
||||
unsigned long start_pfn, end_pfn;
|
||||
int i, nid, zone;
|
||||
bool descending;
|
||||
|
||||
/* Record where the zone boundaries are */
|
||||
memset(arch_zone_lowest_possible_pfn, 0,
|
||||
sizeof(arch_zone_lowest_possible_pfn));
|
||||
memset(arch_zone_highest_possible_pfn, 0,
|
||||
sizeof(arch_zone_highest_possible_pfn));
|
||||
arch_zone_limits_init(max_zone_pfn);
|
||||
|
||||
start_pfn = PHYS_PFN(memblock_start_of_DRAM());
|
||||
descending = arch_has_descending_max_zone_pfns();
|
||||
|
|
@ -2678,13 +2674,19 @@ void __init __weak mem_init(void)
|
|||
{
|
||||
}
|
||||
|
||||
void __init mm_core_init_early(void)
|
||||
{
|
||||
hugetlb_bootmem_alloc();
|
||||
|
||||
free_area_init();
|
||||
}
|
||||
|
||||
/*
|
||||
* Set up kernel memory allocators
|
||||
*/
|
||||
void __init mm_core_init(void)
|
||||
{
|
||||
arch_mm_preinit();
|
||||
hugetlb_bootmem_alloc();
|
||||
|
||||
/* Initializations relying on SMP setup */
|
||||
BUILD_BUG_ON(MAX_ZONELISTS > 2);
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue