/linux/mm/ |
A D | slab.c | 1343 cachep->name, cachep->size, cachep->gfporder); in slab_out_of_memory() 1844 cachep->colour = left / cachep->colour_off; in set_objfreelist_slab_cache() 1878 cachep->colour = left / cachep->colour_off; in set_off_slab_cache() 1894 cachep->colour = left / cachep->colour_off; in set_on_slab_cache() 1972 if (cachep->colour_off < cachep->align) in __kmem_cache_create() 1973 cachep->colour_off = cachep->align; in __kmem_cache_create() 2352 cachep->ctor(objp + obj_offset(cachep)); in cache_init_objs_debug() 2354 cachep, objp + obj_offset(cachep)); in cache_init_objs_debug() 3018 if (cachep->ctor && cachep->flags & SLAB_POISON) in cache_alloc_debugcheck_after() 3602 cachep->object_size, cachep->size, in kmem_cache_alloc_node() [all …]
|
A D | slab.h | 442 struct kmem_cache *cachep; in cache_from_obj() local 448 cachep = virt_to_cache(x); in cache_from_obj() 449 if (WARN(cachep && cachep != s, in cache_from_obj() 451 __func__, s->name, cachep->name)) in cache_from_obj() 452 print_tracking(cachep, x); in cache_from_obj() 453 return cachep; in cache_from_obj() 594 int cache_random_seq_create(struct kmem_cache *cachep, unsigned int count, 596 void cache_random_seq_destroy(struct kmem_cache *cachep); 598 static inline int cache_random_seq_create(struct kmem_cache *cachep, in cache_random_seq_create() argument 603 static inline void cache_random_seq_destroy(struct kmem_cache *cachep) { } in cache_random_seq_destroy() argument
|
A D | slab_common.c | 525 int kmem_cache_shrink(struct kmem_cache *cachep) in kmem_cache_shrink() argument 530 kasan_cache_shrink(cachep); in kmem_cache_shrink() 531 ret = __kmem_cache_shrink(cachep); in kmem_cache_shrink() 992 int cache_random_seq_create(struct kmem_cache *cachep, unsigned int count, in cache_random_seq_create() argument 997 if (count < 2 || cachep->random_seq) in cache_random_seq_create() 1000 cachep->random_seq = kcalloc(count, sizeof(unsigned int), gfp); in cache_random_seq_create() 1001 if (!cachep->random_seq) in cache_random_seq_create() 1007 freelist_randomize(&state, cachep->random_seq, count); in cache_random_seq_create() 1012 void cache_random_seq_destroy(struct kmem_cache *cachep) in cache_random_seq_destroy() argument 1014 kfree(cachep->random_seq); in cache_random_seq_destroy() [all …]
|
A D | slob.c | 630 void *kmem_cache_alloc(struct kmem_cache *cachep, gfp_t flags) in kmem_cache_alloc() argument 632 return slob_alloc_node(cachep, flags, NUMA_NO_NODE); in kmem_cache_alloc() 643 void *kmem_cache_alloc_node(struct kmem_cache *cachep, gfp_t gfp, int node) in kmem_cache_alloc_node() argument 645 return slob_alloc_node(cachep, gfp, node); in kmem_cache_alloc_node()
|
/linux/tools/testing/radix-tree/ |
A D | linux.c | 37 if (cachep->nr_objs) { in kmem_cache_alloc() 39 cachep->nr_objs--; in kmem_cache_alloc() 40 cachep->objs = node->parent; in kmem_cache_alloc() 46 if (cachep->align) in kmem_cache_alloc() 47 posix_memalign(&p, cachep->align, cachep->size); in kmem_cache_alloc() 49 p = malloc(cachep->size); in kmem_cache_alloc() 50 if (cachep->ctor) in kmem_cache_alloc() 51 cachep->ctor(p); in kmem_cache_alloc() 69 if (cachep->nr_objs > 10 || cachep->align) { in kmem_cache_free() 74 cachep->nr_objs++; in kmem_cache_free() [all …]
|
/linux/drivers/scsi/snic/ |
A D | snic_main.c | 833 struct kmem_cache *cachep; in snic_global_data_init() local 866 cachep = kmem_cache_create("snic_req_dfltsgl", len, SNIC_SG_DESC_ALIGN, in snic_global_data_init() 868 if (!cachep) { in snic_global_data_init() 874 snic_glob->req_cache[SNIC_REQ_CACHE_DFLT_SGL] = cachep; in snic_global_data_init() 879 cachep = kmem_cache_create("snic_req_maxsgl", len, SNIC_SG_DESC_ALIGN, in snic_global_data_init() 881 if (!cachep) { in snic_global_data_init() 887 snic_glob->req_cache[SNIC_REQ_CACHE_MAX_SGL] = cachep; in snic_global_data_init() 890 cachep = kmem_cache_create("snic_req_maxsgl", len, SNIC_SG_DESC_ALIGN, in snic_global_data_init() 892 if (!cachep) { in snic_global_data_init() 898 snic_glob->req_cache[SNIC_REQ_TM_CACHE] = cachep; in snic_global_data_init()
|
/linux/tools/testing/radix-tree/linux/ |
A D | slab.h | 20 void *kmem_cache_alloc(struct kmem_cache *cachep, int flags); 21 void kmem_cache_free(struct kmem_cache *cachep, void *objp);
|
/linux/arch/powerpc/mm/ |
A D | hugetlbpage.c | 48 struct kmem_cache *cachep; in __hugepte_alloc() local 54 cachep = PGT_CACHE(PTE_T_ORDER); in __hugepte_alloc() 57 cachep = PGT_CACHE(pdshift - pshift); in __hugepte_alloc() 61 if (!cachep) { in __hugepte_alloc() 66 new = kmem_cache_alloc(cachep, pgtable_gfp_flags(mm, GFP_KERNEL)); in __hugepte_alloc() 97 kmem_cache_free(cachep, new); in __hugepte_alloc()
|
/linux/include/net/ |
A D | inet_hashtables.h | 218 inet_bind_bucket_create(struct kmem_cache *cachep, struct net *net, 221 void inet_bind_bucket_destroy(struct kmem_cache *cachep,
|
/linux/net/ipv4/ |
A D | inet_hashtables.c | 64 struct inet_bind_bucket *inet_bind_bucket_create(struct kmem_cache *cachep, in inet_bind_bucket_create() argument 70 struct inet_bind_bucket *tb = kmem_cache_alloc(cachep, GFP_ATOMIC); in inet_bind_bucket_create() 87 void inet_bind_bucket_destroy(struct kmem_cache *cachep, struct inet_bind_bucket *tb) in inet_bind_bucket_destroy() argument 91 kmem_cache_free(cachep, tb); in inet_bind_bucket_destroy()
|
/linux/Documentation/RCU/ |
A D | rculist_nulls.rst | 125 kmem_cache_free(cachep, obj); 188 obj = kmem_cache_alloc(cachep);
|
/linux/fs/gfs2/ |
A D | glock.c | 1163 struct kmem_cache *cachep; in gfs2_glock_get() local 1175 cachep = gfs2_glock_aspace_cachep; in gfs2_glock_get() 1177 cachep = gfs2_glock_cachep; in gfs2_glock_get() 1178 gl = kmem_cache_alloc(cachep, GFP_NOFS); in gfs2_glock_get() 1187 kmem_cache_free(cachep, gl); in gfs2_glock_get() 1239 kmem_cache_free(cachep, gl); in gfs2_glock_get()
|
/linux/fs/ext4/ |
A D | mballoc.c | 3076 struct kmem_cache *cachep = ext4_groupinfo_caches[cache_index]; in get_groupinfo_cache() local 3078 BUG_ON(!cachep); in get_groupinfo_cache() 3079 return cachep; in get_groupinfo_cache() 3151 meta_group_info[i] = kmem_cache_zalloc(cachep, GFP_NOFS); in ext4_mb_add_groupinfo() 3206 struct kmem_cache *cachep; in ext4_mb_init_backend() local 3265 cachep = get_groupinfo_cache(sb->s_blocksize_bits); in ext4_mb_init_backend() 3267 kmem_cache_free(cachep, ext4_get_group_info(sb, i)); in ext4_mb_init_backend() 3298 struct kmem_cache *cachep; in ext4_groupinfo_create_slab() local 3319 ext4_groupinfo_caches[cache_index] = cachep; in ext4_groupinfo_create_slab() 3322 if (!cachep) { in ext4_groupinfo_create_slab() [all …]
|
/linux/fs/f2fs/ |
A D | f2fs.h | 2612 static inline void *f2fs_kmem_cache_alloc_nofail(struct kmem_cache *cachep, in f2fs_kmem_cache_alloc_nofail() argument 2617 entry = kmem_cache_alloc(cachep, flags); in f2fs_kmem_cache_alloc_nofail() 2619 entry = kmem_cache_alloc(cachep, flags | __GFP_NOFAIL); in f2fs_kmem_cache_alloc_nofail() 2623 static inline void *f2fs_kmem_cache_alloc(struct kmem_cache *cachep, in f2fs_kmem_cache_alloc() argument 2627 return f2fs_kmem_cache_alloc_nofail(cachep, flags); in f2fs_kmem_cache_alloc() 2634 return kmem_cache_alloc(cachep, flags); in f2fs_kmem_cache_alloc()
|