/linux/include/linux/ |
A D | oom.h | 13 struct zonelist; 31 struct zonelist *zonelist; member
|
A D | mmzone.h | 782 struct zonelist { struct 822 struct zonelist node_zonelists[MAX_ZONELISTS]; 1195 static inline struct zoneref *first_zones_zonelist(struct zonelist *zonelist, in first_zones_zonelist() argument 1199 return next_zones_zonelist(zonelist->_zonerefs, in first_zones_zonelist() 1242 struct zonelist *zonelist; in movable_only_nodes() local 1255 zonelist = &NODE_DATA(nid)->node_zonelists[ZONELIST_FALLBACK]; in movable_only_nodes() 1256 z = first_zones_zonelist(zonelist, ZONE_NORMAL, nodes); in movable_only_nodes()
|
A D | swap.h | 385 extern unsigned long try_to_free_pages(struct zonelist *zonelist, int order,
|
A D | gfp.h | 512 static inline struct zonelist *node_zonelist(int nid, gfp_t flags) in node_zonelist()
|
/linux/mm/ |
A D | mm_init.c | 34 struct zonelist *zonelist; in mminit_verify_zonelist() local 43 zonelist = &pgdat->node_zonelists[listid]; in mminit_verify_zonelist() 54 for_each_zone_zonelist(zone, z, zonelist, zoneid) in mminit_verify_zonelist()
|
A D | vmscan.c | 3451 static void shrink_zones(struct zonelist *zonelist, struct scan_control *sc) in shrink_zones() argument 3472 for_each_zone_zonelist_nodemask(zone, z, zonelist, in shrink_zones() 3571 static unsigned long do_try_to_free_pages(struct zonelist *zonelist, in do_try_to_free_pages() argument 3588 shrink_zones(zonelist, sc); in do_try_to_free_pages() 3708 static bool throttle_direct_reclaim(gfp_t gfp_mask, struct zonelist *zonelist, in throttle_direct_reclaim() argument 3746 for_each_zone_zonelist_nodemask(zone, z, zonelist, in throttle_direct_reclaim() 3788 unsigned long try_to_free_pages(struct zonelist *zonelist, int order, in try_to_free_pages() argument 3823 nr_reclaimed = do_try_to_free_pages(zonelist, &sc); in try_to_free_pages() 3896 struct zonelist *zonelist = node_zonelist(numa_node_id(), sc.gfp_mask); in try_to_free_mem_cgroup_pages() local 3902 nr_reclaimed = do_try_to_free_pages(zonelist, &sc); in try_to_free_mem_cgroup_pages() [all …]
|
A D | page_alloc.c | 2805 struct zonelist *zonelist = ac->zonelist; in unreserve_highatomic_pageblock() local 4249 .zonelist = ac->zonelist, in __alloc_pages_may_oom() 4503 for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, in should_compact_retry() 4775 for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, in should_reclaim_retry() 4900 struct zoneref *z = first_zones_zonelist(ac->zonelist, in __alloc_pages_slowpath() 5125 ac->zonelist = node_zonelist(preferred_nid, gfp_mask); in prepare_alloc_pages() 5694 struct zonelist *zonelist = node_zonelist(numa_node_id(), GFP_KERNEL); in nr_free_zone_pages() local 5696 for_each_zone_zonelist(zone, z, zonelist, offset) { in nr_free_zone_pages() 9229 struct zonelist *zonelist; in alloc_contig_pages() local 9233 zonelist = node_zonelist(nid, gfp_mask); in alloc_contig_pages() [all …]
|
A D | internal.h | 183 struct zonelist *zonelist; member
|
A D | oom_kill.c | 273 if (!oc->zonelist) in constrained_alloc() 297 for_each_zone_zonelist_nodemask(zone, z, oc->zonelist, in constrained_alloc()
|
A D | mempolicy.c | 1851 struct zonelist *zonelist; in mempolicy_slab_node() local 1853 zonelist = &NODE_DATA(node)->node_zonelists[ZONELIST_FALLBACK]; in mempolicy_slab_node() 1854 z = first_zones_zonelist(zonelist, highest_zoneidx, in mempolicy_slab_node()
|
A D | slab.c | 3099 struct zonelist *zonelist; in fallback_alloc() local 3113 zonelist = node_zonelist(mempolicy_slab_node(), flags); in fallback_alloc() 3120 for_each_zone_zonelist(zone, z, zonelist, highest_zoneidx) { in fallback_alloc()
|
A D | compaction.c | 2252 for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, in compaction_zonelist_suitable() 2575 for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, in try_to_compact_pages()
|
A D | slub.c | 2183 struct zonelist *zonelist; in get_any_partial() local 2214 zonelist = node_zonelist(mempolicy_slab_node(), flags); in get_any_partial() 2215 for_each_zone_zonelist(zone, z, zonelist, highest_zoneidx) { in get_any_partial()
|
A D | hugetlb.c | 1152 struct zonelist *zonelist; in dequeue_huge_page_nodemask() local 1157 zonelist = node_zonelist(nid, gfp_mask); in dequeue_huge_page_nodemask() 1161 for_each_zone_zonelist_nodemask(zone, z, zonelist, gfp_zone(gfp_mask), nmask) { in dequeue_huge_page_nodemask()
|
A D | memcontrol.c | 1592 .zonelist = NULL, in mem_cgroup_out_of_memory()
|
/linux/Documentation/vm/ |
A D | numa.rst | 76 an ordered "zonelist". A zonelist specifies the zones/nodes to visit when a 86 a default Node ordered zonelist. This means it tries to fallback to other zones 91 Linux will attempt to allocate from the first node in the appropriate zonelist 94 nodes' zones in the selected zonelist looking for the first zone in the list 122 zonelist--will not be the node itself. Rather, it will be the node that the
|
/linux/drivers/tty/ |
A D | sysrq.c | 382 .zonelist = node_zonelist(first_memory_node, gfp_mask), in moom_callback()
|
/linux/Documentation/admin-guide/sysctl/ |
A D | vm.rst | 610 In non-NUMA case, a zonelist for GFP_KERNEL is ordered as following. 616 Assume 2 node NUMA and below is zonelist of Node(0)'s GFP_KERNEL::
|
/linux/Documentation/admin-guide/mm/ |
A D | numa_memory_policy.rst | 236 node zonelist.
|
/linux/Documentation/admin-guide/ |
A D | kernel-parameters.txt | 3644 numa_zonelist_order= [KNL, BOOT] Select zonelist order for NUMA.
|