/linux/scripts/coccinelle/api/ |
A D | kfree_mismatch.cocci | 58 kvmalloc_array\)(...) 124 * kvmalloc_array\)(...)@a 134 kvmalloc_array\)(...)
|
/linux/drivers/gpu/drm/ttm/ |
A D | ttm_tt.c | 99 ttm->pages = kvmalloc_array(ttm->num_pages, sizeof(void*), in ttm_tt_alloc_page_directory() 108 ttm->pages = kvmalloc_array(ttm->num_pages, in ttm_dma_tt_alloc_page_directory() 121 ttm->dma_address = kvmalloc_array(ttm->num_pages, in ttm_sg_tt_alloc_page_directory()
|
/linux/drivers/net/ethernet/mellanox/mlx5/core/en/ |
A D | rqt.c | 97 rss_rqns = kvmalloc_array(MLX5E_INDIR_RQT_SIZE, sizeof(*rss_rqns), GFP_KERNEL); in mlx5e_rqt_init_indir() 157 rss_rqns = kvmalloc_array(MLX5E_INDIR_RQT_SIZE, sizeof(*rss_rqns), GFP_KERNEL); in mlx5e_rqt_redirect_indir()
|
/linux/drivers/gpu/drm/etnaviv/ |
A D | etnaviv_gem_submit.c | 476 bos = kvmalloc_array(args->nr_bos, sizeof(*bos), GFP_KERNEL); in etnaviv_ioctl_gem_submit() 477 relocs = kvmalloc_array(args->nr_relocs, sizeof(*relocs), GFP_KERNEL); in etnaviv_ioctl_gem_submit() 478 pmrs = kvmalloc_array(args->nr_pmrs, sizeof(*pmrs), GFP_KERNEL); in etnaviv_ioctl_gem_submit() 479 stream = kvmalloc_array(1, args->stream_size, GFP_KERNEL); in etnaviv_ioctl_gem_submit()
|
A D | etnaviv_gem_prime.c | 121 etnaviv_obj->pages = kvmalloc_array(npages, sizeof(struct page *), GFP_KERNEL); in etnaviv_gem_prime_import_sg_table()
|
/linux/drivers/gpu/drm/nouveau/nvkm/subdev/mmu/ |
A D | mem.c | 194 if (!(mem->mem = kvmalloc_array(size, sizeof(*mem->mem), GFP_KERNEL))) in nvkm_mem_new_host() 196 if (!(mem->dma = kvmalloc_array(size, sizeof(*mem->dma), GFP_KERNEL))) in nvkm_mem_new_host()
|
/linux/lib/ |
A D | bucket_locks.c | 34 tlocks = kvmalloc_array(size, sizeof(spinlock_t), gfp); in __alloc_bucket_spinlocks()
|
/linux/drivers/gpu/drm/nouveau/dispnv50/ |
A D | lut.c | 41 in = kvmalloc_array(1024, sizeof(*in), GFP_KERNEL); in nv50_lut_load()
|
/linux/drivers/xen/ |
A D | gntdev.c | 136 add->grants = kvmalloc_array(count, sizeof(add->grants[0]), in gntdev_alloc_map() 138 add->map_ops = kvmalloc_array(count, sizeof(add->map_ops[0]), in gntdev_alloc_map() 140 add->unmap_ops = kvmalloc_array(count, sizeof(add->unmap_ops[0]), in gntdev_alloc_map() 149 add->kmap_ops = kvmalloc_array(count, sizeof(add->kmap_ops[0]), in gntdev_alloc_map() 151 add->kunmap_ops = kvmalloc_array(count, sizeof(add->kunmap_ops[0]), in gntdev_alloc_map()
|
/linux/tools/virtio/ringtest/ |
A D | ptr_ring.c | 61 #define kvmalloc_array kmalloc_array macro
|
/linux/drivers/gpu/drm/i915/gt/ |
A D | shmem_utils.c | 61 pages = kvmalloc_array(n_pages, sizeof(*pages), GFP_KERNEL); in shmem_pin_map()
|
/linux/drivers/gpu/drm/ |
A D | drm_gem.c | 562 pages = kvmalloc_array(npages, sizeof(struct page *), GFP_KERNEL); in drm_gem_get_pages() 699 objs = kvmalloc_array(count, sizeof(struct drm_gem_object *), in drm_gem_objects_lookup() 706 handles = kvmalloc_array(count, sizeof(u32), GFP_KERNEL); in drm_gem_objects_lookup()
|
/linux/drivers/gpu/drm/amd/amdgpu/ |
A D | amdgpu_mn.c | 175 pfns = kvmalloc_array(npages, sizeof(*pfns), GFP_KERNEL); in amdgpu_hmm_range_get_pages()
|
A D | amdgpu_cs.c | 120 chunk_array = kvmalloc_array(cs->in.num_chunks, sizeof(uint64_t), GFP_KERNEL); in amdgpu_cs_parser_init() 147 p->chunks = kvmalloc_array(p->nchunks, sizeof(struct amdgpu_cs_chunk), in amdgpu_cs_parser_init() 172 p->chunks[i].kdata = kvmalloc_array(size, sizeof(uint32_t), GFP_KERNEL); in amdgpu_cs_parser_init() 543 e->user_pages = kvmalloc_array(bo->tbo.ttm->num_pages, in amdgpu_cs_parser_bos()
|
A D | amdgpu_bo_list.c | 231 info = kvmalloc_array(in->bo_number, info_size, GFP_KERNEL); in amdgpu_bo_create_list_entry_array()
|
/linux/include/linux/ |
A D | slab.h | 752 static inline __alloc_size(1, 2) void *kvmalloc_array(size_t n, size_t size, gfp_t flags) in kvmalloc_array() function 764 return kvmalloc_array(n, size, flags | __GFP_ZERO); in kvcalloc()
|
/linux/drivers/gpu/drm/i915/gem/ |
A D | i915_gem_pages.c | 298 pages = kvmalloc_array(n_pages, sizeof(*pages), GFP_KERNEL); in i915_gem_object_map_page() 328 pfns = kvmalloc_array(n_pfn, sizeof(*pfns), GFP_KERNEL); in i915_gem_object_map_pfn()
|
/linux/drivers/gpu/drm/v3d/ |
A D | v3d_gem.c | 311 job->bo = kvmalloc_array(job->bo_count, in v3d_lookup_bos() 319 handles = kvmalloc_array(job->bo_count, sizeof(u32), GFP_KERNEL); in v3d_lookup_bos() 601 kvmalloc_array(count, in v3d_get_multisync_post_deps()
|
/linux/drivers/dma-buf/ |
A D | st-dma-fence-chain.c | 117 fc->chains = kvmalloc_array(count, sizeof(*fc->chains), in fence_chains_init() 122 fc->fences = kvmalloc_array(count, sizeof(*fc->fences), in fence_chains_init()
|
/linux/drivers/staging/media/ipu3/ |
A D | ipu3-dmamap.c | 42 pages = kvmalloc_array(count, sizeof(*pages), GFP_KERNEL); in imgu_dmamap_alloc_buffer()
|
/linux/drivers/gpu/drm/panfrost/ |
A D | panfrost_drv.c | 147 job->mappings = kvmalloc_array(job->bo_count, in panfrost_lookup_bos() 198 handles = kvmalloc_array(in_fence_count, sizeof(u32), GFP_KERNEL); in panfrost_copy_in_sync()
|
A D | panfrost_mmu.c | 451 bo->sgts = kvmalloc_array(bo->base.base.size / SZ_2M, in panfrost_mmu_map_fault_addr() 459 pages = kvmalloc_array(bo->base.base.size >> PAGE_SHIFT, in panfrost_mmu_map_fault_addr()
|
/linux/drivers/gpu/drm/xen/ |
A D | xen_drm_front_gem.c | 49 xen_obj->pages = kvmalloc_array(xen_obj->num_pages, in gem_alloc_pages_array()
|
/linux/drivers/gpu/drm/virtio/ |
A D | virtgpu_object.c | 186 *ents = kvmalloc_array(*nents, in virtio_gpu_object_shmem_init()
|
/linux/drivers/gpu/drm/vc4/ |
A D | vc4_gem.c | 746 exec->bo = kvmalloc_array(exec->bo_count, in vc4_cl_lookup_bos() 754 handles = kvmalloc_array(exec->bo_count, sizeof(uint32_t), GFP_KERNEL); in vc4_cl_lookup_bos() 855 temp = kvmalloc_array(temp_size, 1, GFP_KERNEL); in vc4_get_bcl()
|