Lines Matching refs:start_pfn
577 unsigned long sp, start_pfn; in page_outside_zone_boundaries() local
581 start_pfn = zone->zone_start_pfn; in page_outside_zone_boundaries()
590 start_pfn, start_pfn + sp); in page_outside_zone_boundaries()
1605 unsigned long start_pfn = PFN_DOWN(start); in reserve_bootmem_region() local
1608 for (; start_pfn < end_pfn; start_pfn++) { in reserve_bootmem_region()
1609 if (pfn_valid(start_pfn)) { in reserve_bootmem_region()
1610 struct page *page = pfn_to_page(start_pfn); in reserve_bootmem_region()
1612 init_reserved_page(start_pfn); in reserve_bootmem_region()
1701 unsigned long start_pfn, end_pfn; in __early_pfn_to_nid() local
1707 nid = memblock_search_pfn_nid(pfn, &start_pfn, &end_pfn); in __early_pfn_to_nid()
1709 state->last_start = start_pfn; in __early_pfn_to_nid()
1755 struct page *__pageblock_pfn_to_page(unsigned long start_pfn, in __pageblock_pfn_to_page() argument
1764 if (!pfn_valid(start_pfn) || !pfn_valid(end_pfn)) in __pageblock_pfn_to_page()
1767 start_page = pfn_to_online_page(start_pfn); in __pageblock_pfn_to_page()
1960 deferred_init_maxorder(u64 *i, struct zone *zone, unsigned long *start_pfn, in deferred_init_maxorder() argument
1963 unsigned long mo_pfn = ALIGN(*start_pfn + 1, MAX_ORDER_NR_PAGES); in deferred_init_maxorder()
1964 unsigned long spfn = *start_pfn, epfn = *end_pfn; in deferred_init_maxorder()
1969 for_each_free_mem_pfn_range_in_zone_from(j, zone, start_pfn, end_pfn) { in deferred_init_maxorder()
1972 if (mo_pfn <= *start_pfn) in deferred_init_maxorder()
1976 nr_pages += deferred_init_pages(zone, *start_pfn, t); in deferred_init_maxorder()
1979 *start_pfn = mo_pfn; in deferred_init_maxorder()
2004 deferred_init_memmap_chunk(unsigned long start_pfn, unsigned long end_pfn, in deferred_init_memmap_chunk() argument
2011 deferred_init_mem_pfn_range_in_zone(&i, zone, &spfn, &epfn, start_pfn); in deferred_init_memmap_chunk()
2496 unsigned long start_pfn, unsigned long end_pfn, in move_freepages() argument
2504 for (pfn = start_pfn; pfn <= end_pfn;) { in move_freepages()
2535 unsigned long start_pfn, end_pfn, pfn; in move_freepages_block() local
2541 start_pfn = pfn & ~(pageblock_nr_pages - 1); in move_freepages_block()
2542 end_pfn = start_pfn + pageblock_nr_pages - 1; in move_freepages_block()
2545 if (!zone_spans_pfn(zone, start_pfn)) in move_freepages_block()
2546 start_pfn = pfn; in move_freepages_block()
2550 return move_freepages(zone, start_pfn, end_pfn, migratetype, in move_freepages_block()
6506 unsigned long start_pfn, unsigned long zone_end_pfn, in memmap_init_range() argument
6510 unsigned long pfn, end_pfn = start_pfn + size; in memmap_init_range()
6528 if (start_pfn == altmap->base_pfn) in memmap_init_range()
6529 start_pfn += altmap->reserve; in memmap_init_range()
6534 for (pfn = start_pfn; pfn < end_pfn; ) { in memmap_init_range()
6566 unsigned long start_pfn, in memmap_init_zone_device() argument
6570 unsigned long pfn, end_pfn = start_pfn + nr_pages; in memmap_init_zone_device()
6586 start_pfn = altmap->base_pfn + vmem_altmap_offset(altmap); in memmap_init_zone_device()
6587 nr_pages = end_pfn - start_pfn; in memmap_init_zone_device()
6590 for (pfn = start_pfn; pfn < end_pfn; pfn++) { in memmap_init_zone_device()
6688 unsigned long start_pfn, in memmap_init_zone_range() argument
6696 start_pfn = clamp(start_pfn, zone_start_pfn, zone_end_pfn); in memmap_init_zone_range()
6699 if (start_pfn >= end_pfn) in memmap_init_zone_range()
6702 memmap_init_range(end_pfn - start_pfn, nid, zone_id, start_pfn, in memmap_init_zone_range()
6705 if (*hole_pfn < start_pfn) in memmap_init_zone_range()
6706 init_unavailable_range(*hole_pfn, start_pfn, zone_id, nid); in memmap_init_zone_range()
6713 unsigned long start_pfn, end_pfn; in memmap_init() local
6717 for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, &nid) { in memmap_init()
6726 memmap_init_zone_range(zone, start_pfn, end_pfn, in memmap_init()
7045 unsigned long *start_pfn, unsigned long *end_pfn) in get_pfn_range_for_nid() argument
7050 *start_pfn = -1UL; in get_pfn_range_for_nid()
7054 *start_pfn = min(*start_pfn, this_start_pfn); in get_pfn_range_for_nid()
7058 if (*start_pfn == -1UL) in get_pfn_range_for_nid()
7059 *start_pfn = 0; in get_pfn_range_for_nid()
7165 unsigned long start_pfn, end_pfn; in __absent_pages_in_range() local
7168 for_each_mem_pfn_range(i, nid, &start_pfn, &end_pfn, NULL) { in __absent_pages_in_range()
7169 start_pfn = clamp(start_pfn, range_start_pfn, range_end_pfn); in __absent_pages_in_range()
7171 nr_absent -= end_pfn - start_pfn; in __absent_pages_in_range()
7183 unsigned long __init absent_pages_in_range(unsigned long start_pfn, in absent_pages_in_range() argument
7186 return __absent_pages_in_range(MAX_NUMNODES, start_pfn, end_pfn); in absent_pages_in_range()
7218 unsigned long start_pfn, end_pfn; in zone_absent_pages_in_node() local
7222 start_pfn = clamp(memblock_region_memory_base_pfn(r), in zone_absent_pages_in_node()
7229 nr_absent += end_pfn - start_pfn; in zone_absent_pages_in_node()
7233 nr_absent += end_pfn - start_pfn; in zone_absent_pages_in_node()
7584 unsigned long start_pfn = 0; in free_area_init_node() local
7590 get_pfn_range_for_nid(nid, &start_pfn, &end_pfn); in free_area_init_node()
7593 pgdat->node_start_pfn = start_pfn; in free_area_init_node()
7597 (u64)start_pfn << PAGE_SHIFT, in free_area_init_node()
7599 calculate_node_totalpages(pgdat, start_pfn, end_pfn); in free_area_init_node()
7694 unsigned long start_pfn, end_pfn; in early_calculate_totalpages() local
7697 for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, &nid) { in early_calculate_totalpages()
7698 unsigned long pages = end_pfn - start_pfn; in early_calculate_totalpages()
7825 unsigned long start_pfn, end_pfn; in find_zone_movable_pfns_for_nodes() local
7843 for_each_mem_pfn_range(i, nid, &start_pfn, &end_pfn, NULL) { in find_zone_movable_pfns_for_nodes()
7846 start_pfn = max(start_pfn, zone_movable_pfn[nid]); in find_zone_movable_pfns_for_nodes()
7847 if (start_pfn >= end_pfn) in find_zone_movable_pfns_for_nodes()
7851 if (start_pfn < usable_startpfn) { in find_zone_movable_pfns_for_nodes()
7854 - start_pfn; in find_zone_movable_pfns_for_nodes()
7873 start_pfn = usable_startpfn; in find_zone_movable_pfns_for_nodes()
7881 size_pages = end_pfn - start_pfn; in find_zone_movable_pfns_for_nodes()
7884 zone_movable_pfn[nid] = start_pfn + size_pages; in find_zone_movable_pfns_for_nodes()
7961 unsigned long start_pfn, end_pfn; in free_area_init() local
7971 start_pfn = find_min_pfn_with_active_regions(); in free_area_init()
7983 end_pfn = max(max_zone_pfn[zone], start_pfn); in free_area_init()
7984 arch_zone_lowest_possible_pfn[zone] = start_pfn; in free_area_init()
7987 start_pfn = end_pfn; in free_area_init()
8025 for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, &nid) { in free_area_init()
8027 (u64)start_pfn << PAGE_SHIFT, in free_area_init()
8029 subsection_map_init(start_pfn, end_pfn - start_pfn); in free_area_init()
9167 static int __alloc_contig_pages(unsigned long start_pfn, in __alloc_contig_pages() argument
9170 unsigned long end_pfn = start_pfn + nr_pages; in __alloc_contig_pages()
9172 return alloc_contig_range(start_pfn, end_pfn, MIGRATE_MOVABLE, in __alloc_contig_pages()
9176 static bool pfn_range_valid_contig(struct zone *z, unsigned long start_pfn, in pfn_range_valid_contig() argument
9179 unsigned long i, end_pfn = start_pfn + nr_pages; in pfn_range_valid_contig()
9182 for (i = start_pfn; i < end_pfn; i++) { in pfn_range_valid_contig()
9197 unsigned long start_pfn, unsigned long nr_pages) in zone_spans_last_pfn() argument
9199 unsigned long last_pfn = start_pfn + nr_pages - 1; in zone_spans_last_pfn()
9331 void __offline_isolated_pages(unsigned long start_pfn, unsigned long end_pfn) in __offline_isolated_pages() argument
9333 unsigned long pfn = start_pfn; in __offline_isolated_pages()