Searched refs:PAGES_PER_SECTION (Results 1 – 17 of 17) sorted by relevance
/linux/mm/ |
A D | sparse.c | 206 pfns = min(nr_pages, PAGES_PER_SECTION in subsection_map_init() 245 for (pfn = start; pfn < end; pfn += PAGES_PER_SECTION) { in memory_present() 420 return ALIGN(sizeof(struct page) * PAGES_PER_SECTION, PMD_SIZE); in section_map_size() 426 return PAGE_ALIGN(sizeof(struct page) * PAGES_PER_SECTION); in section_map_size() 526 map = __populate_section_memmap(pfn, PAGES_PER_SECTION, in sparse_init_nid() 596 for (pfn = start_pfn; pfn < end_pfn; pfn += PAGES_PER_SECTION) { in online_mem_sections() 614 for (pfn = start_pfn; pfn < end_pfn; pfn += PAGES_PER_SECTION) { in offline_mem_sections() 648 unsigned long end = (unsigned long)(memmap + PAGES_PER_SECTION); in free_map_bootmem() 706 PAGES_PER_SECTION), GFP_KERNEL, nid); in populate_section_memmap() 721 nr_pages = PAGE_ALIGN(PAGES_PER_SECTION * sizeof(struct page)) in free_map_bootmem() [all …]
|
A D | page_ext.c | 248 table_size = page_ext_size * PAGES_PER_SECTION; in init_section_page_ext() 281 table_size = page_ext_size * PAGES_PER_SECTION; in free_page_ext() 322 for (pfn = start; !fail && pfn < end; pfn += PAGES_PER_SECTION) in online_page_ext() 328 for (pfn = start; pfn < end; pfn += PAGES_PER_SECTION) in online_page_ext() 342 for (pfn = start; pfn < end; pfn += PAGES_PER_SECTION) in offline_page_ext() 396 pfn = ALIGN(pfn + 1, PAGES_PER_SECTION)) { in page_ext_init()
|
A D | bootmem_info.c | 60 mapsize = sizeof(struct page) * PAGES_PER_SECTION; in register_page_bootmem_info_section() 89 register_page_bootmem_memmap(section_nr, memmap, PAGES_PER_SECTION); in register_page_bootmem_info_section() 117 for (; pfn < end_pfn; pfn += PAGES_PER_SECTION) { in register_page_bootmem_info_node()
|
A D | memory_hotplug.c | 240 min_align = PAGES_PER_SECTION; in check_pfn_span() 700 if (!IS_ALIGNED(start_pfn, PAGES_PER_SECTION)) in move_pfn_range_to_zone() 702 if (!IS_ALIGNED(start_pfn + nr_pages, PAGES_PER_SECTION)) in move_pfn_range_to_zone() 922 for (; pfn < end_pfn; pfn += PAGES_PER_SECTION) { in auto_movable_zone_for_pfn() 929 online_pages += PAGES_PER_SECTION; in auto_movable_zone_for_pfn() 1035 if (nr_pages >= PAGES_PER_SECTION) in mhp_init_memmap_on_memory() 1050 if (nr_pages >= PAGES_PER_SECTION) in mhp_deinit_memmap_on_memory() 1051 offline_mem_sections(pfn, ALIGN_DOWN(end_pfn, PAGES_PER_SECTION)); in mhp_deinit_memmap_on_memory() 1079 !IS_ALIGNED(pfn + nr_pages, PAGES_PER_SECTION))) in online_pages() 1605 pfn = sec_end_pfn, sec_end_pfn += PAGES_PER_SECTION) { in test_pages_in_a_zone() [all …]
|
A D | memblock.c | 1973 start = min(start, ALIGN(prev_end, PAGES_PER_SECTION)); in free_unused_memmap() 1998 if (!IS_ALIGNED(prev_end, PAGES_PER_SECTION)) { in free_unused_memmap() 2000 free_memmap(prev_end, ALIGN(prev_end, PAGES_PER_SECTION)); in free_unused_memmap()
|
A D | page_alloc.c | 439 if ((nr_initialised > PAGES_PER_SECTION) && in defer_init() 440 (pfn & (PAGES_PER_SECTION - 1)) == 0) { in defer_init() 479 pfn &= (PAGES_PER_SECTION-1); in pfn_to_bitidx() 2081 unsigned long epfn_align = ALIGN(epfn, PAGES_PER_SECTION); in deferred_init_memmap() 2087 .align = PAGES_PER_SECTION, in deferred_init_memmap() 2088 .min_chunk = PAGES_PER_SECTION, in deferred_init_memmap() 2125 unsigned long nr_pages_needed = ALIGN(1 << order, PAGES_PER_SECTION); in deferred_grow_zone() 2169 if ((first_deferred_pfn ^ spfn) < PAGES_PER_SECTION) in deferred_grow_zone() 6742 end_pfn = round_up(end_pfn, PAGES_PER_SECTION); in memmap_init()
|
/linux/drivers/xen/ |
A D | unpopulated-alloc.c | 23 unsigned int i, alloc_pages = round_up(nr_pages, PAGES_PER_SECTION); in fill_list() 35 PAGES_PER_SECTION * PAGE_SIZE, NULL, NULL); in fill_list()
|
A D | balloon.c | 267 PAGES_PER_SECTION * PAGE_SIZE, NULL, NULL); in additional_memory_resource() 294 balloon_hotplug = round_up(credit, PAGES_PER_SECTION); in reserve_additional_memory()
|
/linux/drivers/base/ |
A D | memory.c | 181 unsigned long nr_pages = PAGES_PER_SECTION * sections_per_block; in memory_block_online() 224 unsigned long nr_pages = PAGES_PER_SECTION * sections_per_block; in memory_block_offline() 402 unsigned long nr_pages = PAGES_PER_SECTION * sections_per_block; in valid_zones_show() 494 unsigned long pages_per_block = PAGES_PER_SECTION * sections_per_block; in probe_store()
|
A D | node.c | 842 pfn = round_down(pfn + PAGES_PER_SECTION, in register_mem_block_under_node_early() 843 PAGES_PER_SECTION) - 1; in register_mem_block_under_node_early()
|
/linux/include/linux/ |
A D | mmzone.h | 1280 #define PAGES_PER_SECTION (1UL << PFN_SECTION_SHIFT) macro 1281 #define PAGE_SECTION_MASK (~(PAGES_PER_SECTION-1)) 1299 #define SECTION_ALIGN_UP(pfn) (((pfn) + PAGES_PER_SECTION - 1) & PAGE_SECTION_MASK)
|
/linux/drivers/virtio/ |
A D | virtio_mem.c | 2087 for (pfn = start_pfn; pfn < end_pfn; pfn += PAGES_PER_SECTION) { in virtio_mem_bbm_offline_remove_and_unplug_bb() 2092 rc = virtio_mem_fake_offline(pfn, PAGES_PER_SECTION); in virtio_mem_bbm_offline_remove_and_unplug_bb() 2120 for (pfn = start_pfn; pfn < end_pfn; pfn += PAGES_PER_SECTION) { in virtio_mem_bbm_offline_remove_and_unplug_bb() 2124 virtio_mem_fake_online(pfn, PAGES_PER_SECTION); in virtio_mem_bbm_offline_remove_and_unplug_bb() 2142 pfn += PAGES_PER_SECTION) { in virtio_mem_bbm_bb_is_offline() 2162 pfn += PAGES_PER_SECTION) { in virtio_mem_bbm_bb_is_movable()
|
/linux/arch/x86/mm/ |
A D | numa.c | 579 if (pfn_align && pfn_align < PAGES_PER_SECTION) { in numa_register_memblks() 582 PFN_PHYS(PAGES_PER_SECTION) >> 20); in numa_register_memblks()
|
A D | init_64.c | 1611 if (end - start < PAGES_PER_SECTION * sizeof(struct page)) in vmemmap_populate()
|
/linux/arch/powerpc/platforms/powernv/ |
A D | memtrace.c | 98 if (IS_ALIGNED(pfn, PAGES_PER_SECTION)) in memtrace_clear_range()
|
/linux/fs/proc/ |
A D | page.c | 33 return round_up(max_pfn, PAGES_PER_SECTION); in get_max_dump_pfn()
|
/linux/arch/arm64/mm/ |
A D | mmu.c | 1540 for (; pfn < end_pfn; pfn += PAGES_PER_SECTION) { in prevent_bootmem_remove_notifier()
|
Completed in 58 milliseconds