/linux/net/ceph/ |
A D | pagevec.c | 19 set_page_dirty_lock(pages[i]); in ceph_put_page_vector() 20 put_page(pages[i]); in ceph_put_page_vector() 22 kvfree(pages); in ceph_put_page_vector() 31 __free_pages(pages[i], 0); in ceph_release_page_vector() 32 kfree(pages); in ceph_release_page_vector() 41 struct page **pages; in ceph_alloc_page_vector() local 44 pages = kmalloc_array(num_pages, sizeof(*pages), flags); in ceph_alloc_page_vector() 45 if (!pages) in ceph_alloc_page_vector() 49 if (pages[i] == NULL) { in ceph_alloc_page_vector() 50 ceph_release_page_vector(pages, i); in ceph_alloc_page_vector() [all …]
|
/linux/fs/isofs/ |
A D | compress.c | 68 if (!pages[i]) in zisofs_uncompress_block() 122 if (pages[curpage]) { in zisofs_uncompress_block() 175 if (pages[curpage]) { in zisofs_uncompress_block() 309 struct page **pages; in zisofs_readpage() local 336 if (!pages) { in zisofs_readpage() 345 if (pages[i]) { in zisofs_readpage() 347 kmap(pages[i]); in zisofs_readpage() 355 if (pages[i]) { in zisofs_readpage() 359 kunmap(pages[i]); in zisofs_readpage() 362 put_page(pages[i]); in zisofs_readpage() [all …]
|
/linux/mm/ |
A D | percpu-vm.c | 34 static struct page **pages; in pcpu_get_pages() local 39 if (!pages) in pcpu_get_pages() 41 return pages; in pcpu_get_pages() 165 pages[pcpu_page_idx(cpu, i)] = page; in pcpu_unmap_pages() 197 PAGE_KERNEL, pages, PAGE_SHIFT); in __pcpu_map_pages() 279 struct page **pages; in pcpu_populate_chunk() local 281 pages = pcpu_get_pages(); in pcpu_populate_chunk() 282 if (!pages) in pcpu_populate_chunk() 315 struct page **pages; in pcpu_depopulate_chunk() local 322 pages = pcpu_get_pages(); in pcpu_depopulate_chunk() [all …]
|
A D | gup_test.c | 18 put_page(pages[i]); in put_back_pages() 31 put_page(pages[i]); in put_back_pages() 49 page = pages[i]; in verify_dma_pinned() 104 struct page **pages; in __gup_test_ioctl() local 114 if (!pages) in __gup_test_ioctl() 138 pages + i); in __gup_test_ioctl() 146 pages + i); in __gup_test_ioctl() 155 pages + i, NULL); in __gup_test_ioctl() 160 pages + i, NULL); in __gup_test_ioctl() 163 pages + i, NULL); in __gup_test_ioctl() [all …]
|
A D | gup.c | 1116 pages ? &pages[i] : NULL); in __get_user_pages() 1185 if (pages) { in __get_user_pages() 1377 pages += ret; in __get_user_pages_locked() 1423 pages++; in __get_user_pages_locked() 1644 if (pages) { in __get_user_pages_locked() 1646 if (pages[i]) in __get_user_pages_locked() 2650 pages, nr)) in gup_pmd_range() 2683 pages, nr)) in gup_pud_range() 2869 pages); in internal_get_user_pages_fast() 2916 pages); in get_user_pages_fast_only() [all …]
|
/linux/drivers/gpu/drm/i915/gem/selftests/ |
A D | huge_gem_object.c | 12 struct sg_table *pages) in huge_free_pages() argument 24 sg_free_table(pages); in huge_free_pages() 25 kfree(pages); in huge_free_pages() 34 struct sg_table *pages; in huge_get_pages() local 37 pages = kmalloc(sizeof(*pages), GFP); in huge_get_pages() 38 if (!pages) in huge_get_pages() 42 kfree(pages); in huge_get_pages() 46 sg = pages->sgl; in huge_get_pages() 64 src = pages->sgl; in huge_get_pages() 76 huge_free_pages(obj, pages); in huge_get_pages() [all …]
|
/linux/fs/erofs/ |
A D | pcpubuf.c | 15 struct page **pages; member 67 pages = kmalloc_array(nrpages, sizeof(*pages), GFP_KERNEL); in erofs_pcpubuf_growsize() 68 if (!pages) { in erofs_pcpubuf_growsize() 75 if (!pages[i]) { in erofs_pcpubuf_growsize() 77 oldpages = pages; in erofs_pcpubuf_growsize() 84 oldpages = pages; in erofs_pcpubuf_growsize() 91 pcb->pages = pages; in erofs_pcpubuf_growsize() 139 if (!pcb->pages) in erofs_pcpubuf_exit() 143 if (pcb->pages[i]) in erofs_pcpubuf_exit() 145 kfree(pcb->pages); in erofs_pcpubuf_exit() [all …]
|
/linux/drivers/xen/ |
A D | xlate_mmu.c | 71 struct page **pages; member 148 struct page **pages) in xen_xlate_remap_gfn_array() argument 163 data.pages = pages; in xen_xlate_remap_gfn_array() 217 struct page **pages; in xen_xlate_map_ballooned_pages() local 226 pages = kcalloc(nr_pages, sizeof(pages[0]), GFP_KERNEL); in xen_xlate_map_ballooned_pages() 227 if (!pages) in xen_xlate_map_ballooned_pages() 232 kfree(pages); in xen_xlate_map_ballooned_pages() 239 kfree(pages); in xen_xlate_map_ballooned_pages() 253 kfree(pages); in xen_xlate_map_ballooned_pages() 257 kfree(pages); in xen_xlate_map_ballooned_pages() [all …]
|
/linux/kernel/dma/ |
A D | remap.c | 15 return area->pages; in dma_common_find_pages() 22 void *dma_common_pages_remap(struct page **pages, size_t size, in dma_common_pages_remap() argument 27 vaddr = vmap(pages, PAGE_ALIGN(size) >> PAGE_SHIFT, in dma_common_pages_remap() 30 find_vm_area(vaddr)->pages = pages; in dma_common_pages_remap() 42 struct page **pages; in dma_common_contiguous_remap() local 46 pages = kmalloc_array(count, sizeof(struct page *), GFP_KERNEL); in dma_common_contiguous_remap() 47 if (!pages) in dma_common_contiguous_remap() 50 pages[i] = nth_page(page, i); in dma_common_contiguous_remap() 51 vaddr = vmap(pages, count, VM_DMA_COHERENT, prot); in dma_common_contiguous_remap() 52 kfree(pages); in dma_common_contiguous_remap()
|
/linux/drivers/gpu/drm/xen/ |
A D | xen_drm_front_gem.c | 30 struct page **pages; member 56 kvfree(xen_obj->pages); in gem_free_pages_array() 57 xen_obj->pages = NULL; in gem_free_pages_array() 136 if (IS_ERR(xen_obj->pages)) { in gem_create() 138 xen_obj->pages = NULL; in gem_create() 169 if (xen_obj->pages) { in xen_drm_front_gem_free_object_unlocked() 172 xen_obj->pages); in xen_drm_front_gem_free_object_unlocked() 188 return xen_obj->pages; in xen_drm_front_gem_get_pages() 195 if (!xen_obj->pages) in xen_drm_front_gem_get_sg_table() 231 xen_obj->pages); in xen_drm_front_gem_import_sg_table() [all …]
|
/linux/drivers/staging/media/ipu3/ |
A D | ipu3-dmamap.c | 26 __free_page(pages[count]); in imgu_dmamap_free_buffer() 27 kvfree(pages); in imgu_dmamap_free_buffer() 36 struct page **pages; in imgu_dmamap_alloc_buffer() local 42 pages = kvmalloc_array(count, sizeof(*pages), GFP_KERNEL); in imgu_dmamap_alloc_buffer() 44 if (!pages) in imgu_dmamap_alloc_buffer() 77 pages[i++] = page++; in imgu_dmamap_alloc_buffer() 80 return pages; in imgu_dmamap_alloc_buffer() 100 struct page **pages; in imgu_dmamap_alloc() local 113 if (!pages) in imgu_dmamap_alloc() 131 map->pages = pages; in imgu_dmamap_alloc() [all …]
|
/linux/Documentation/admin-guide/mm/ |
A D | hugetlbpage.rst | 32 and surplus huge pages in the pool of huge pages of default size. 48 is the size of the pool of huge pages. 71 pages of all sizes. 82 pages in the kernel's huge page pool. "Persistent" huge pages will be 173 default sized persistent huge pages:: 207 huge pages can grow, if more huge pages than ``/proc/sys/vm/nr_hugepages`` are 215 surplus pages will first be promoted to persistent huge pages. Then, additional 228 of the in-use huge pages to surplus huge pages. This will occur even if 249 pages may exist:: 272 pages size are allowed. [all …]
|
/linux/fs/ramfs/ |
A D | file-nommu.c | 65 struct page *pages; in ramfs_nommu_expand_for_mapping() local 84 pages = alloc_pages(gfp, order); in ramfs_nommu_expand_for_mapping() 85 if (!pages) in ramfs_nommu_expand_for_mapping() 92 split_page(pages, order); in ramfs_nommu_expand_for_mapping() 96 __free_page(pages + loop); in ramfs_nommu_expand_for_mapping() 100 data = page_address(pages); in ramfs_nommu_expand_for_mapping() 124 __free_page(pages + loop++); in ramfs_nommu_expand_for_mapping() 225 if (!pages) in ramfs_nommu_get_unmapped_area() 233 ptr = pages; in ramfs_nommu_get_unmapped_area() 244 ptr = pages; in ramfs_nommu_get_unmapped_area() [all …]
|
/linux/drivers/staging/media/atomisp/pci/hmm/ |
A D | hmm_reserved_pool.c | 53 page_obj[i].page = repool_info->pages[j]; in get_pages_from_reserved_pool() 97 if (unlikely(!pool_info->pages)) { in hmm_reserved_pool_setup() 120 struct page *pages; in hmm_reserved_pool_init() local 146 pages = alloc_pages(GFP_KERNEL | __GFP_NOWARN, order); in hmm_reserved_pool_init() 147 if (unlikely(!pages)) { in hmm_reserved_pool_init() 163 ret = set_pages_uc(pages, blk_pgnr); in hmm_reserved_pool_init() 167 __free_pages(pages, order); in hmm_reserved_pool_init() 172 repool_info->pages[i++] = pages + j; in hmm_reserved_pool_init() 216 ret = set_pages_wb(repool_info->pages[i], 1); in hmm_reserved_pool_exit() 228 __free_pages(repool_info->pages[i], 0); in hmm_reserved_pool_exit() [all …]
|
/linux/net/rds/ |
A D | info.c | 65 struct page **pages; member 122 iter->addr = kmap_atomic(*iter->pages); in rds_info_copy() 127 "bytes %lu\n", *iter->pages, iter->addr, in rds_info_copy() 140 iter->pages++; in rds_info_copy() 166 struct page **pages = NULL; in rds_info_getsockopt() local 191 if (!pages) { in rds_info_getsockopt() 195 ret = pin_user_pages_fast(start, nr_pages, FOLL_WRITE, pages); in rds_info_getsockopt() 214 iter.pages = pages; in rds_info_getsockopt() 237 if (pages) in rds_info_getsockopt() 238 unpin_user_pages(pages, nr_pages); in rds_info_getsockopt() [all …]
|
/linux/fs/squashfs/ |
A D | page_actor.c | 30 if (actor->next_page == actor->pages) in cache_next_page() 42 int pages, int length) in squashfs_page_actor_init() argument 49 actor->length = length ? : pages * PAGE_SIZE; in squashfs_page_actor_init() 51 actor->pages = pages; in squashfs_page_actor_init() 71 return actor->pageaddr = actor->next_page == actor->pages ? NULL : in direct_next_page() 82 int pages, int length) in squashfs_page_actor_init_special() argument 89 actor->length = length ? : pages * PAGE_SIZE; in squashfs_page_actor_init_special() 91 actor->pages = pages; in squashfs_page_actor_init_special()
|
A D | file_direct.c | 22 int pages, struct page **page, int bytes); 36 int i, n, pages, missing_pages, bytes, res = -ENOMEM; in squashfs_readpage_block() local 44 pages = end_index - start_index + 1; in squashfs_readpage_block() 46 page = kmalloc_array(pages, sizeof(void *), GFP_KERNEL); in squashfs_readpage_block() 54 actor = squashfs_page_actor_init_special(page, pages, 0); in squashfs_readpage_block() 84 res = squashfs_read_cache(target_page, block, bsize, pages, in squashfs_readpage_block() 105 pageaddr = kmap_atomic(page[pages - 1]); in squashfs_readpage_block() 111 for (i = 0; i < pages; i++) { in squashfs_readpage_block() 128 for (i = 0; i < pages; i++) { in squashfs_readpage_block() 145 int pages, struct page **page, int bytes) in squashfs_read_cache() argument [all …]
|
/linux/drivers/gpu/drm/i915/gem/ |
A D | i915_gem_pages.c | 14 struct sg_table *pages, in __i915_gem_object_set_pages() argument 31 drm_clflush_sg(pages); in __i915_gem_object_set_pages() 40 obj->mm.pages = pages; in __i915_gem_object_set_pages() 204 pages = fetch_and_zero(&obj->mm.pages); in __i915_gem_object_unset_pages() 206 return pages; in __i915_gem_object_unset_pages() 221 return pages; in __i915_gem_object_unset_pages() 298 pages = kvmalloc_array(n_pages, sizeof(*pages), GFP_KERNEL); in i915_gem_object_map_page() 299 if (!pages) in i915_gem_object_map_page() 305 pages[i++] = page; in i915_gem_object_map_page() 307 if (pages != stack) in i915_gem_object_map_page() [all …]
|
A D | i915_gem_phys.c | 95 struct sg_table *pages) in i915_gem_object_put_pages_phys() argument 97 dma_addr_t dma = sg_dma_address(pages->sgl); in i915_gem_object_put_pages_phys() 98 void *vaddr = sg_page(pages->sgl); in i915_gem_object_put_pages_phys() 100 __i915_gem_object_release_shmem(obj, pages, false); in i915_gem_object_put_pages_phys() 130 sg_free_table(pages); in i915_gem_object_put_pages_phys() 131 kfree(pages); in i915_gem_object_put_pages_phys() 190 struct sg_table *pages; in i915_gem_object_shmem_to_phys() local 193 pages = __i915_gem_object_unset_pages(obj); in i915_gem_object_shmem_to_phys() 202 if (!IS_ERR_OR_NULL(pages)) in i915_gem_object_shmem_to_phys() 203 i915_gem_object_put_pages_shmem(obj, pages); in i915_gem_object_shmem_to_phys() [all …]
|
/linux/Documentation/vm/ |
A D | unevictable-lru.rst | 15 pages. 30 pages and to hide these pages from vmscan. This mechanism is based on a patch 69 (1) We get to "treat unevictable pages just like we treat other pages in the 84 swap-backed pages. This differentiation is only important while the pages are, 192 "cull" such pages that it encounters: that is, it diverts those pages to the 231 prevented the management of the pages on an LRU list, and thus mlocked pages 245 mlocked pages - pages mapped into a VM_LOCKED VMA - are a class of unevictable 301 get_user_pages() will be unable to fault in the pages. That's okay. If pages 348 such as the VDSO page, relay channel pages, etc. These pages 378 these pages for munlocking. Because we don't want to leave pages mlocked, [all …]
|
/linux/arch/arm/plat-omap/ |
A D | sram.c | 65 int pages; in omap_sram_push() local 73 pages = PAGE_ALIGN(size) / PAGE_SIZE; in omap_sram_push() 75 set_memory_rw(base, pages); in omap_sram_push() 79 set_memory_ro(base, pages); in omap_sram_push() 80 set_memory_x(base, pages); in omap_sram_push() 101 int pages; in omap_map_sram() local 125 pages = PAGE_ALIGN(omap_sram_size) / PAGE_SIZE; in omap_map_sram() 127 set_memory_ro(base, pages); in omap_map_sram() 128 set_memory_x(base, pages); in omap_map_sram()
|
/linux/drivers/media/common/videobuf2/ |
A D | frame_vector.c | 107 struct page **pages; in put_vaddr_frames() local 111 pages = frame_vector_pages(vec); in put_vaddr_frames() 117 if (WARN_ON(IS_ERR(pages))) in put_vaddr_frames() 120 unpin_user_pages(pages, vec->nr_frames); in put_vaddr_frames() 139 struct page **pages; in frame_vector_to_pages() local 147 pages = (struct page **)nums; in frame_vector_to_pages() 149 pages[i] = pfn_to_page(nums[i]); in frame_vector_to_pages() 165 struct page **pages; in frame_vector_to_pfns() local 169 pages = (struct page **)(vec->ptrs); in frame_vector_to_pfns() 170 nums = (unsigned long *)pages; in frame_vector_to_pfns() [all …]
|
/linux/drivers/gpu/drm/ |
A D | drm_scatter.c | 51 for (i = 0; i < entry->pages; i++) { in drm_sg_cleanup() 83 unsigned long pages, i, j; in drm_legacy_sg_alloc() local 103 pages = (request->size + PAGE_SIZE - 1) / PAGE_SIZE; in drm_legacy_sg_alloc() 104 DRM_DEBUG("size=%ld pages=%ld\n", request->size, pages); in drm_legacy_sg_alloc() 106 entry->pages = pages; in drm_legacy_sg_alloc() 107 entry->pagelist = kcalloc(pages, sizeof(*entry->pagelist), GFP_KERNEL); in drm_legacy_sg_alloc() 113 entry->busaddr = kcalloc(pages, sizeof(*entry->busaddr), GFP_KERNEL); in drm_legacy_sg_alloc() 120 entry->virtual = vmalloc_32(pages << PAGE_SHIFT); in drm_legacy_sg_alloc() 131 memset(entry->virtual, 0, pages << PAGE_SHIFT); in drm_legacy_sg_alloc() 138 for (i = (unsigned long)entry->virtual, j = 0; j < pages; in drm_legacy_sg_alloc() [all …]
|
/linux/drivers/block/xen-blkback/ |
A D | blkback.c | 252 unmap_data.pages = pages; in free_persistent_gnts() 294 unmap_data.pages = pages; in xen_blkbk_unmap_purged_grants() 647 struct grant_page **pages, in xen_blkbk_unmap_prepare() argument 717 work->pages = req->unmap_pages; in xen_blkbk_unmap_and_respond() 751 pages += batch; in xen_blkbk_unmap() 757 struct grant_page *pages[], in xen_blkbk_map() argument 785 pages[i]->gref); in xen_blkbk_map() 797 &pages[i]->page)) { in xen_blkbk_map() 804 addr = vaddr(pages[i]->page); in xen_blkbk_map() 811 flags, pages[i]->gref, in xen_blkbk_map() [all …]
|
/linux/drivers/net/ethernet/amd/xgbe/ |
A D | xgbe-desc.c | 141 if (ring->rx_hdr_pa.pages) { in xgbe_free_ring() 152 if (ring->rx_buf_pa.pages) { in xgbe_free_ring() 289 struct page *pages = NULL; in xgbe_alloc_pages() local 301 if (pages) in xgbe_alloc_pages() 313 if (!pages) in xgbe_alloc_pages() 320 put_page(pages); in xgbe_alloc_pages() 324 pa->pages = pages; in xgbe_alloc_pages() 336 get_page(pa->pages); in xgbe_set_buffer_data() 349 pa->pages = NULL; in xgbe_set_buffer_data() 490 if (rdata->rx.hdr.pa.pages) in xgbe_unmap_rdata() [all …]
|