/linux/block/ |
A D | blk-lib.c | 27 sector_t nr_sects, gfp_t gfp_mask, int flags, in __blkdev_issue_discard() 133 sector_t nr_sects, gfp_t gfp_mask, unsigned long flags) in blkdev_issue_discard() 167 sector_t nr_sects, gfp_t gfp_mask, struct page *page, in __blkdev_issue_write_same() 228 sector_t nr_sects, gfp_t gfp_mask, in blkdev_issue_write_same() 248 sector_t sector, sector_t nr_sects, gfp_t gfp_mask, in __blkdev_issue_write_zeroes() 304 sector_t sector, sector_t nr_sects, gfp_t gfp_mask, in __blkdev_issue_zero_pages() 360 sector_t nr_sects, gfp_t gfp_mask, struct bio **biop, in __blkdev_issue_zeroout() 394 sector_t nr_sects, gfp_t gfp_mask, unsigned flags) in blkdev_issue_zeroout()
|
A D | blk-map.c | 22 gfp_t gfp_mask) in bio_alloc_map_data() 130 struct iov_iter *iter, gfp_t gfp_mask) in bio_copy_user_iov() 234 gfp_t gfp_mask) in bio_map_user_iov() 340 unsigned int len, gfp_t gfp_mask) in bio_map_kern() 422 unsigned int len, gfp_t gfp_mask, int reading) in bio_copy_kern() 523 const struct iov_iter *iter, gfp_t gfp_mask) in blk_rq_map_user_iov() 567 unsigned long len, gfp_t gfp_mask) in blk_rq_map_user() 626 unsigned int len, gfp_t gfp_mask) in blk_rq_map_kern()
|
A D | blk-crypto.c | 82 const u64 dun[BLK_CRYPTO_DUN_ARRAY_SIZE], gfp_t gfp_mask) in bio_crypt_set_ctx() 106 int __bio_crypt_clone(struct bio *dst, struct bio *src, gfp_t gfp_mask) in __bio_crypt_clone() 297 gfp_t gfp_mask) in __blk_crypto_rq_bio_prep()
|
A D | blk-zoned.c | 191 gfp_t gfp_mask) in blkdev_zone_reset_all_emulated() 238 static int blkdev_zone_reset_all(struct block_device *bdev, gfp_t gfp_mask) in blkdev_zone_reset_all() 267 gfp_t gfp_mask) in blkdev_zone_mgmt()
|
A D | bio-integrity.c | 49 gfp_t gfp_mask, in bio_integrity_alloc() 404 gfp_t gfp_mask) in bio_integrity_clone()
|
/linux/include/linux/ |
A D | gfp.h | 565 __alloc_pages_node(int nid, gfp_t gfp_mask, unsigned int order) in __alloc_pages_node() 587 static inline struct page *alloc_pages_node(int nid, gfp_t gfp_mask, in alloc_pages_node() 602 #define alloc_hugepage_vma(gfp_mask, vma, addr, order) \ argument 605 static inline struct page *alloc_pages(gfp_t gfp_mask, unsigned int order) in alloc_pages() 613 #define alloc_pages_vma(gfp_mask, order, vma, addr, node, false)\ argument 615 #define alloc_hugepage_vma(gfp_mask, vma, addr, order) \ argument 618 #define alloc_page(gfp_mask) alloc_pages(gfp_mask, 0) argument 619 #define alloc_page_vma(gfp_mask, vma, addr) \ argument 629 #define __get_free_page(gfp_mask) \ argument 632 #define __get_dma_pages(gfp_mask, order) \ argument [all …]
|
A D | cpuset.h | 85 static inline bool cpuset_node_allowed(int node, gfp_t gfp_mask) in cpuset_node_allowed() 92 static inline bool __cpuset_zone_allowed(struct zone *z, gfp_t gfp_mask) in __cpuset_zone_allowed() 97 static inline bool cpuset_zone_allowed(struct zone *z, gfp_t gfp_mask) in cpuset_zone_allowed() 226 static inline bool cpuset_node_allowed(int node, gfp_t gfp_mask) in cpuset_node_allowed() 231 static inline bool __cpuset_zone_allowed(struct zone *z, gfp_t gfp_mask) in __cpuset_zone_allowed() 236 static inline bool cpuset_zone_allowed(struct zone *z, gfp_t gfp_mask) in cpuset_zone_allowed()
|
A D | page_owner.h | 28 unsigned short order, gfp_t gfp_mask) in set_page_owner() 59 unsigned int order, gfp_t gfp_mask) in set_page_owner()
|
A D | swap.h | 581 static inline int add_swap_count_continuation(swp_entry_t swp, gfp_t gfp_mask) in add_swap_count_continuation() 604 gfp_t gfp_mask, struct vm_fault *vmf) in swap_cluster_readahead() 609 static inline struct page *swapin_readahead(swp_entry_t swp, gfp_t gfp_mask, in swapin_readahead() 644 gfp_t gfp_mask, void **shadowp) in add_to_swap_cache() 731 static inline void cgroup_throttle_swaprate(struct page *page, gfp_t gfp_mask) in cgroup_throttle_swaprate() 738 static inline void cgroup_throttle_swaprate(struct page *page, gfp_t gfp_mask) in cgroup_throttle_swaprate()
|
/linux/fs/nfs/blocklayout/ |
A D | dev.c | 236 struct pnfs_block_volume *volumes, int idx, gfp_t gfp_mask) in bl_parse_simple() 353 struct pnfs_block_volume *volumes, int idx, gfp_t gfp_mask) in bl_parse_scsi() 402 struct pnfs_block_volume *volumes, int idx, gfp_t gfp_mask) in bl_parse_slice() 418 struct pnfs_block_volume *volumes, int idx, gfp_t gfp_mask) in bl_parse_concat() 447 struct pnfs_block_volume *volumes, int idx, gfp_t gfp_mask) in bl_parse_stripe() 476 struct pnfs_block_volume *volumes, int idx, gfp_t gfp_mask) in bl_parse_deviceid() 497 gfp_t gfp_mask) in bl_alloc_deviceid_node()
|
/linux/mm/ |
A D | mempool.c | 181 gfp_t gfp_mask, int node_id) in mempool_init_node() 262 gfp_t gfp_mask, int node_id) in mempool_create_node() 374 void *mempool_alloc(mempool_t *pool, gfp_t gfp_mask) in mempool_alloc() 508 void *mempool_alloc_slab(gfp_t gfp_mask, void *pool_data) in mempool_alloc_slab() 527 void *mempool_kmalloc(gfp_t gfp_mask, void *pool_data) in mempool_kmalloc() 544 void *mempool_alloc_pages(gfp_t gfp_mask, void *pool_data) in mempool_alloc_pages()
|
A D | page_owner.c | 26 gfp_t gfp_mask; member 152 unsigned short order, gfp_t gfp_mask) in __set_page_owner_handle() 173 gfp_t gfp_mask) in __set_page_owner() 394 gfp_t gfp_mask; in __dump_page_owner() local
|
A D | page_alloc.c | 3912 unsigned int alloc_flags, gfp_t gfp_mask) in zone_watermark_fast() 4517 static bool __need_reclaim(gfp_t gfp_mask) in __need_reclaim() 4543 void fs_reclaim_acquire(gfp_t gfp_mask) in fs_reclaim_acquire() 4560 void fs_reclaim_release(gfp_t gfp_mask) in fs_reclaim_release() 4648 gfp_to_alloc_flags(gfp_t gfp_mask) in gfp_to_alloc_flags() 4726 bool gfp_pfmemalloc_allowed(gfp_t gfp_mask) in gfp_pfmemalloc_allowed() 5425 unsigned long get_zeroed_page(gfp_t gfp_mask) in get_zeroed_page() 5483 gfp_t gfp_mask) in __page_frag_cache_refill() 5513 unsigned int fragsz, gfp_t gfp_mask, in page_frag_alloc_align() 9037 unsigned migratetype, gfp_t gfp_mask) in alloc_contig_range() [all …]
|
A D | swap_state.c | 417 struct page *__read_swap_cache_async(swp_entry_t entry, gfp_t gfp_mask, in __read_swap_cache_async() 521 struct page *read_swap_cache_async(swp_entry_t entry, gfp_t gfp_mask, in read_swap_cache_async() 613 struct page *swap_cluster_readahead(swp_entry_t entry, gfp_t gfp_mask, in swap_cluster_readahead() 788 static struct page *swap_vma_readahead(swp_entry_t fentry, gfp_t gfp_mask, in swap_vma_readahead() 849 struct page *swapin_readahead(swp_entry_t entry, gfp_t gfp_mask, in swapin_readahead()
|
/linux/net/sunrpc/auth_gss/ |
A D | gss_krb5_mech.c | 312 context_derive_keys_des3(struct krb5_ctx *ctx, gfp_t gfp_mask) in context_derive_keys_des3() 357 context_derive_keys_new(struct krb5_ctx *ctx, gfp_t gfp_mask) in context_derive_keys_new() 469 gfp_t gfp_mask) in gss_import_v2_context() 547 gfp_t gfp_mask) in gss_import_sec_context_kerberos()
|
/linux/fs/btrfs/ |
A D | ulist.c | 92 struct ulist *ulist_alloc(gfp_t gfp_mask) in ulist_alloc() 186 int ulist_add(struct ulist *ulist, u64 val, u64 aux, gfp_t gfp_mask) in ulist_add() 192 u64 *old_aux, gfp_t gfp_mask) in ulist_add_merge()
|
/linux/lib/ |
A D | generic-radix-tree.c | 79 static inline struct genradix_node *genradix_alloc_node(gfp_t gfp_mask) in genradix_alloc_node() 105 gfp_t gfp_mask) in __genradix_ptr_alloc() 218 gfp_t gfp_mask) in __genradix_prealloc()
|
A D | scatterlist.c | 149 static struct scatterlist *sg_kmalloc(unsigned int nents, gfp_t gfp_mask) in sg_kmalloc() 284 unsigned int nents_first_chunk, gfp_t gfp_mask, in __sg_alloc_table() 371 int sg_alloc_table(struct sg_table *table, unsigned int nents, gfp_t gfp_mask) in sg_alloc_table() 386 gfp_t gfp_mask) in get_next_sg() 445 unsigned int left_pages, gfp_t gfp_mask) in sg_alloc_append_table_from_pages() 568 gfp_t gfp_mask) in sg_alloc_table_from_pages_segment()
|
/linux/include/linux/sched/ |
A D | mm.h | 213 static inline void fs_reclaim_acquire(gfp_t gfp_mask) { } in fs_reclaim_acquire() 214 static inline void fs_reclaim_release(gfp_t gfp_mask) { } in fs_reclaim_release() 225 static inline void might_alloc(gfp_t gfp_mask) in might_alloc()
|
/linux/drivers/net/ethernet/mellanox/mlx4/ |
A D | icm.c | 99 gfp_t gfp_mask, int node) in mlx4_alloc_icm_pages() 115 int order, gfp_t gfp_mask) in mlx4_alloc_icm_coherent() 133 gfp_t gfp_mask, int coherent) in mlx4_alloc_icm()
|
/linux/drivers/connector/ |
A D | connector.c | 62 gfp_t gfp_mask) in cn_netlink_send_mult() 122 gfp_t gfp_mask) in cn_netlink_send()
|
/linux/kernel/power/ |
A D | snapshot.c | 189 static void *get_image_page(gfp_t gfp_mask, int safe_needed) in get_image_page() 208 static void *__get_safe_page(gfp_t gfp_mask) in __get_safe_page() 220 unsigned long get_safe_page(gfp_t gfp_mask) in get_safe_page() 225 static struct page *alloc_image_page(gfp_t gfp_mask) in alloc_image_page() 295 gfp_t gfp_mask; /* mask for allocating pages */ member 299 static void chain_init(struct chain_allocator *ca, gfp_t gfp_mask, in chain_init() 435 static struct rtree_node *alloc_rtree_node(gfp_t gfp_mask, int safe_needed, in alloc_rtree_node() 461 static int add_rtree_block(struct mem_zone_bm_rtree *zone, gfp_t gfp_mask, in add_rtree_block() 531 static struct mem_zone_bm_rtree *create_zone_bm_rtree(gfp_t gfp_mask, in create_zone_bm_rtree() 620 static int create_mem_extents(struct list_head *list, gfp_t gfp_mask) in create_mem_extents() [all …]
|
/linux/fs/ntfs/ |
A D | malloc.h | 28 static inline void *__ntfs_malloc(unsigned long size, gfp_t gfp_mask) in __ntfs_malloc()
|
/linux/kernel/ |
A D | umh.c | 359 char **envp, gfp_t gfp_mask, in call_usermodehelper_setup() 477 gfp_t gfp_mask = (wait == UMH_NO_WAIT) ? GFP_ATOMIC : GFP_KERNEL; in call_usermodehelper() local
|
/linux/fs/crypto/ |
A D | inline_crypt.c | 243 u64 first_lblk, gfp_t gfp_mask) in fscrypt_set_bio_crypt_ctx() 293 gfp_t gfp_mask) in fscrypt_set_bio_crypt_ctx_bh()
|