/linux/mm/ |
A D | page_alloc.c | 2993 alloc_flags)) in __rmqueue() 3020 alloc_flags); in rmqueue_bulk() 3998 return alloc_flags; in alloc_flags_nofragment() 4001 return alloc_flags; in alloc_flags_nofragment() 4010 return alloc_flags; in alloc_flags_nofragment() 4014 return alloc_flags; in alloc_flags_nofragment() 4025 return alloc_flags; in gfp_to_alloc_flags_cma() 4239 alloc_flags, ac); in __alloc_pages_cpuset_fallback() 4684 alloc_flags = gfp_to_alloc_flags_cma(gfp_mask, alloc_flags); in gfp_to_alloc_flags() 4686 return alloc_flags; in gfp_to_alloc_flags() [all …]
|
A D | compaction.c | 2155 unsigned int alloc_flags, in __compaction_suitable() argument 2164 watermark = wmark_pages(zone, alloc_flags & ALLOC_WMARK_MASK); in __compaction_suitable() 2170 alloc_flags)) in __compaction_suitable() 2205 unsigned int alloc_flags, in compaction_suitable() argument 2243 int alloc_flags) in compaction_zonelist_suitable() argument 2265 compact_result = __compaction_suitable(zone, order, alloc_flags, in compaction_zonelist_suitable() 2296 ret = compaction_suitable(cc->zone, cc->order, cc->alloc_flags, in compact_zone() 2488 unsigned int alloc_flags, int highest_zoneidx, in compact_zone_order() argument 2499 .alloc_flags = alloc_flags, in compact_zone_order() 2557 unsigned int alloc_flags, const struct alloc_context *ac, in try_to_compact_pages() argument [all …]
|
A D | internal.h | 294 const unsigned int alloc_flags; /* alloc flags of a direct compactor */ member
|
/linux/lib/ |
A D | stackdepot.c | 317 gfp_t alloc_flags, bool can_alloc) in __stack_depot_save() argument 356 alloc_flags &= ~GFP_ZONEMASK; in __stack_depot_save() 357 alloc_flags &= (GFP_ATOMIC | GFP_KERNEL); in __stack_depot_save() 358 alloc_flags |= __GFP_NOWARN; in __stack_depot_save() 359 page = alloc_pages(alloc_flags, STACK_ALLOC_ORDER); in __stack_depot_save() 414 gfp_t alloc_flags) in stack_depot_save() argument 416 return __stack_depot_save(entries, nr_entries, alloc_flags, true); in stack_depot_save()
|
/linux/include/linux/ |
A D | compaction.h | 95 unsigned int order, unsigned int alloc_flags, 100 unsigned int alloc_flags, int highest_zoneidx); 178 int alloc_flags); 190 int alloc_flags, int highest_zoneidx) in compaction_suitable() argument
|
A D | mmzone.h | 949 int highest_zoneidx, unsigned int alloc_flags, 953 unsigned int alloc_flags);
|
/linux/drivers/gpu/drm/amd/amdgpu/ |
A D | amdgpu_amdkfd_gpuvm.c | 420 if (mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_VRAM) { in get_pte_flags() 435 !(mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_VRAM)) in get_pte_flags() 472 mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_WRITABLE ? in kfd_mem_dmamap_userptr() 553 mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_WRITABLE ? in kfd_mem_dmaunmap_userptr() 636 mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_WRITABLE ? in kfd_mem_attach_dmabuf() 1401 u64 alloc_flags; in amdgpu_amdkfd_gpuvm_alloc_memory_of_gpu() local 1414 alloc_flags = 0; in amdgpu_amdkfd_gpuvm_alloc_memory_of_gpu() 1418 alloc_flags = AMDGPU_GEM_CREATE_PREEMPTIBLE; in amdgpu_amdkfd_gpuvm_alloc_memory_of_gpu() 1427 alloc_flags = 0; in amdgpu_amdkfd_gpuvm_alloc_memory_of_gpu() 1453 (*mem)->alloc_flags = flags; in amdgpu_amdkfd_gpuvm_alloc_memory_of_gpu() [all …]
|
A D | amdgpu_amdkfd.h | 75 uint32_t alloc_flags; member
|
/linux/drivers/md/ |
A D | dm-zoned-reclaim.c | 286 int alloc_flags = DMZ_ALLOC_SEQ; in dmz_reclaim_rnd_data() local 292 alloc_flags | DMZ_ALLOC_RECLAIM); in dmz_reclaim_rnd_data() 293 if (!szone && alloc_flags == DMZ_ALLOC_SEQ && dmz_nr_cache_zones(zmd)) { in dmz_reclaim_rnd_data() 294 alloc_flags = DMZ_ALLOC_RND; in dmz_reclaim_rnd_data()
|
A D | dm-zoned-metadata.c | 2070 int alloc_flags = zmd->nr_cache ? DMZ_ALLOC_CACHE : DMZ_ALLOC_RND; in dmz_get_chunk_mapping() local 2085 dzone = dmz_alloc_zone(zmd, 0, alloc_flags); in dmz_get_chunk_mapping() 2182 int alloc_flags = zmd->nr_cache ? DMZ_ALLOC_CACHE : DMZ_ALLOC_RND; in dmz_get_chunk_buffer() local 2191 bzone = dmz_alloc_zone(zmd, 0, alloc_flags); in dmz_get_chunk_buffer()
|
/linux/fs/btrfs/ |
A D | block-group.c | 2546 u64 alloc_flags; in btrfs_inc_block_group_ro() local 2581 alloc_flags = btrfs_get_alloc_profile(fs_info, cache->flags); in btrfs_inc_block_group_ro() 2582 if (alloc_flags != cache->flags) { in btrfs_inc_block_group_ro() 2583 ret = btrfs_chunk_alloc(trans, alloc_flags, in btrfs_inc_block_group_ro() 2601 alloc_flags = btrfs_get_alloc_profile(fs_info, cache->space_info->flags); in btrfs_inc_block_group_ro() 2602 ret = btrfs_chunk_alloc(trans, alloc_flags, CHUNK_ALLOC_FORCE); in btrfs_inc_block_group_ro() 2610 alloc_flags = btrfs_get_alloc_profile(fs_info, cache->flags); in btrfs_inc_block_group_ro() 2612 check_system_chunk(trans, alloc_flags); in btrfs_inc_block_group_ro() 3404 u64 alloc_flags = btrfs_get_alloc_profile(trans->fs_info, type); in btrfs_force_chunk_alloc() local 3406 return btrfs_chunk_alloc(trans, alloc_flags, CHUNK_ALLOC_FORCE); in btrfs_force_chunk_alloc()
|
/linux/drivers/base/regmap/ |
A D | internal.h | 64 gfp_t alloc_flags; member
|
A D | regmap.c | 819 map->alloc_flags = GFP_ATOMIC; in __regmap_init() 821 map->alloc_flags = GFP_KERNEL; in __regmap_init() 2296 wval = kmemdup(val, val_count * val_bytes, map->alloc_flags); in regmap_bulk_write()
|
/linux/drivers/iommu/ |
A D | dma-iommu.c | 576 gfp_t alloc_flags = gfp; in __iommu_dma_alloc_pages() local 580 alloc_flags |= __GFP_NORETRY; in __iommu_dma_alloc_pages() 581 page = alloc_pages_node(nid, alloc_flags, order); in __iommu_dma_alloc_pages()
|
/linux/drivers/gpu/drm/i915/gem/ |
A D | i915_gem_object.h | 60 unsigned alloc_flags);
|
/linux/drivers/net/ethernet/mellanox/mlx5/core/ |
A D | cmd.c | 78 gfp_t alloc_flags = cbk ? GFP_ATOMIC : GFP_KERNEL; in cmd_alloc_ent() local 81 ent = kzalloc(sizeof(*ent), alloc_flags); in cmd_alloc_ent()
|
/linux/arch/s390/kvm/ |
A D | kvm-s390.c | 2679 gfp_t alloc_flags = GFP_KERNEL_ACCOUNT; in kvm_arch_init_vm() local 2702 alloc_flags |= GFP_DMA; in kvm_arch_init_vm() 2705 kvm->arch.sca = (struct bsca_block *) get_zeroed_page(alloc_flags); in kvm_arch_init_vm()
|